From 460ba2408092afefe8693d12468100c373ac9018 Mon Sep 17 00:00:00 2001 From: Intel1 <25161862+Intel11@users.noreply.github.com> Date: Mon, 5 Feb 2018 09:27:33 -0500 Subject: [PATCH 01/10] userscloud: fix --- plugin.video.alfa/servers/userscloud.json | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/plugin.video.alfa/servers/userscloud.json b/plugin.video.alfa/servers/userscloud.json index 6a4280a4..f3da7dd7 100755 --- a/plugin.video.alfa/servers/userscloud.json +++ b/plugin.video.alfa/servers/userscloud.json @@ -4,7 +4,7 @@ "ignore_urls": [], "patterns": [ { - "pattern": "userscloud.com/(?:embed-|)([A-z0-9]+)", + "pattern": "userscloud.com/(?:embed-|embed/|)([A-z0-9]+)", "url": "http://userscloud.com/\\1" } ] @@ -43,4 +43,4 @@ } ], "thumbnail": "http://i.imgur.com/u4W2DgA.png?1" -} \ No newline at end of file +} From 7f3d9127e3e1b49c1ea697f64744374c782a11d8 Mon Sep 17 00:00:00 2001 From: Intel1 <25161862+Intel11@users.noreply.github.com> Date: Mon, 5 Feb 2018 09:28:21 -0500 Subject: [PATCH 02/10] userscloud: fix --- plugin.video.alfa/servers/userscloud.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/plugin.video.alfa/servers/userscloud.py b/plugin.video.alfa/servers/userscloud.py index 58a572b6..8b2e4942 100755 --- a/plugin.video.alfa/servers/userscloud.py +++ b/plugin.video.alfa/servers/userscloud.py @@ -19,18 +19,18 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): logger.info("url=" + page_url) - video_urls = [] + unpacked = "" data = httptools.downloadpage(page_url).data packed = scrapertools.find_single_match(data, "function\(p,a,c,k.*?") - unpacked = jsunpack.unpack(packed) + if packed: + unpacked = jsunpack.unpack(packed) media_url = scrapertools.find_single_match(unpacked, 'src"value="([^"]+)') if not media_url: id_ = page_url.rsplit("/", 1)[1] rand = scrapertools.find_single_match(data, 'name="rand" value="([^"]+)"') post = "op=download2&id=%s&rand=%s&referer=%s&method_free=&method_premium=" % (id_, rand, page_url) data = httptools.downloadpage(page_url, post).data - media_url = scrapertools.find_single_match(data, '