From b1bd07ab9e28bf8d58cf1d8207d836c9ca45e57a Mon Sep 17 00:00:00 2001 From: marco Date: Tue, 14 Apr 2020 20:09:40 +0200 Subject: [PATCH] fix wstream --- servers/wstream.json | 4 ++-- servers/wstream.py | 12 +++++++----- 2 files changed, 9 insertions(+), 7 deletions(-) diff --git a/servers/wstream.json b/servers/wstream.json index 07e60564..da667068 100644 --- a/servers/wstream.json +++ b/servers/wstream.json @@ -8,11 +8,11 @@ "patterns": [ { "pattern": "wstream\\.video(?!<)(?:=|/)(?:video[a-zA-Z0-9.?_]*|embed[a-zA-Z0-9]*|)?(?!api|swembedid)(?:-|/|=)?(?:[a-z0-9A-Z]+/)?([a-z0-9A-Z]+)", - "url": "http://116.202.226.34/video.php?file_code=\\1" + "url": "http://wstream.video/video.php?file_code=\\1" }, { "pattern": "wstream\\.video/(api/vcmod/fastredirect/streaming\\.php\\?id=[$0-9]+)", - "url": "http://116.202.226.34/\\1" + "url": "http://wstream.video/\\1" } ], "ignore_urls": [] diff --git a/servers/wstream.py b/servers/wstream.py index 9d7dc41d..464aa6ee 100644 --- a/servers/wstream.py +++ b/servers/wstream.py @@ -17,17 +17,17 @@ headers = [['User-Agent', 'Mozilla/5.0 (Windows NT 6.1; WOW64; rv:54.0) Gecko/20 def test_video_exists(page_url): logger.info("(page_url='%s')" % page_url) - resp = httptools.downloadpage(page_url, headers=headers) + resp = httptools.downloadpage(page_url.replace('wstream.video', '116.202.226.34'), headers=headers, verify=False) global data, real_url data = resp.data page_url = resp.url.replace('wstream.video', '116.202.226.34') if '/streaming.php' in page_url in page_url: - code = httptools.downloadpage(page_url, headers=headers, follow_redirects=False, only_headers=True).headers['location'].split('/')[-1].replace('.html', '') + code = httptools.downloadpage(page_url, headers=headers, follow_redirects=False, only_headers=True, verify=False).headers['location'].split('/')[-1].replace('.html', '') # logger.info('WCODE=' + code) page_url = 'https://116.202.226.34/video.php?file_code=' + code - data = httptools.downloadpage(page_url, headers=headers, follow_redirects=True).data + data = httptools.downloadpage(page_url, headers=headers, follow_redirects=True, verify=False).data real_url = page_url if "Not Found" in data or "File was deleted" in data: @@ -42,7 +42,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= global data,headers page_url = scrapertools.find_single_match(data, r"""
|>)""") + from core import support + support.dbg() if possibleParam and possibleParam[0]: post = {param[0]: param[1] for param in possibleParam if param[0]} if captcha: post['g-recaptcha-response'] = captcha if post: - data = httptools.downloadpage(real_url, headers=headers, post=post, follow_redirects=True).data + data = httptools.downloadpage(real_url, headers=headers, post=post, follow_redirects=True, verify=False).data elif captcha: int_bckup_method() elif captcha: