From 296829935456425aa696b81ed415dcb441106b80 Mon Sep 17 00:00:00 2001 From: marco <10120390+mac12m99@users.noreply.github.com> Date: Mon, 14 Nov 2022 20:59:32 +0100 Subject: [PATCH] =?UTF-8?q?Streamingcommunity:=20non=20riprodurre=20ultimo?= =?UTF-8?q?=20contenuto=20quando=20il=20selezionato=20uscir=C3=A0=20prossi?= =?UTF-8?q?mamente?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- channels/animesaturn.py | 3 --- channels/streamingcommunity.py | 15 +++++++-------- servers/streamingcommunityws.py | 6 ++++-- 3 files changed, 11 insertions(+), 13 deletions(-) diff --git a/channels/animesaturn.py b/channels/animesaturn.py index ce97cd06..eecc547e 100644 --- a/channels/animesaturn.py +++ b/channels/animesaturn.py @@ -3,10 +3,7 @@ # Canale per AnimeSaturn # ---------------------------------------------------------- -from lib import js2py from core import support -from platformcode import config -from platformcode.logger import debug host = support.config.get_channel_url() __channel__ = 'animesaturn' diff --git a/channels/streamingcommunity.py b/channels/streamingcommunity.py index 8e5a59de..6295113e 100644 --- a/channels/streamingcommunity.py +++ b/channels/streamingcommunity.py @@ -17,8 +17,7 @@ else: host = support.config.get_channel_url() -headers = {} -headers = {'user-agent':httptools.random_useragent(), +headers = {'user-agent': httptools.random_useragent(), 'referer': host + '/browse'} # def getHeaders(forced=False): @@ -112,7 +111,6 @@ def newest(category): return itemlist - def peliculas(item): logger.debug() if item.mainThumb: item.thumbnail = item.mainThumb @@ -138,7 +136,6 @@ def peliculas(item): headers['referer'] = host + '/search' records = requests.post(host + '/api/search', headers=headers, json=payload).json()['records'] - if records and type(records[0]) == list: js = [] for record in records: @@ -152,7 +149,7 @@ def peliculas(item): else: recordlist.append(it) - itlist = [makeItem(i, it, item) for i, it in enumerate(items)] + # itlist = [makeItem(i, it, item) for i, it in enumerate(items)] with futures.ThreadPoolExecutor() as executor: itlist = [executor.submit(makeItem, i, it, item) for i, it in enumerate(items)] @@ -172,6 +169,7 @@ def peliculas(item): support.check_trakt(itemlist) return itemlist + def makeItem(n, it, item): info = httptools.downloadpage(host + '/api/titles/preview/{}'.format(it['id']), headers=headers, post={}).json logger.debug(jsontools.dump(info)) @@ -183,7 +181,6 @@ def makeItem(n, it, item): itm.language = lang itm.year = info['release_date'].split('-')[0] - if itm.contentType == 'movie': # itm.contentType = 'movie' itm.fulltitle = itm.show = itm.contentTitle = title @@ -200,6 +197,7 @@ def makeItem(n, it, item): itm.n = n return itm + def episodios(item): # getHeaders() logger.debug() @@ -238,8 +236,9 @@ def episodios(item): def findvideos(item): support.callAds('https://thaudray.com/5/3523301', host) # Fix for old items in videolibrary - if item.episodeid and not item.episodeid in item.url: + if item.episodeid and item.episodeid not in item.url: item.url += item.episodeid - itemlist = [item.clone(title = channeltools.get_channel_parameters(item.channel)['title'], url=item.url, server='streamingcommunityws')] + + itemlist = [item.clone(title=channeltools.get_channel_parameters(item.channel)['title'], url=item.url, server='streamingcommunityws')] return support.server(item, itemlist=itemlist, referer=False) diff --git a/servers/streamingcommunityws.py b/servers/streamingcommunityws.py index 882a2bcc..8af40950 100755 --- a/servers/streamingcommunityws.py +++ b/servers/streamingcommunityws.py @@ -1,5 +1,4 @@ # -*- coding: utf-8 -*- -from wsgiref import headers from core import httptools, support, filetools from platformcode import logger, config UA = httptools.random_useragent() @@ -8,11 +7,14 @@ UA = httptools.random_useragent() def test_video_exists(page_url): global scws_id logger.debug('page url=', page_url) + scws_id = '' if page_url.isdigit(): scws_id = page_url else: - scws_id = support.match(page_url, patron=r'scws_id[^:]+:(\d+)').match + page = httptools.downloadpage(page_url) + if page.url == page_url: # se non esiste, reindirizza all'ultimo url chiamato esistente + scws_id = support.scrapertools.find_single_match(page.data, r'scws_id[^:]+:(\d+)') if not scws_id: return False, config.get_localized_string(70449) % 'StreamingCommunityWS'