diff --git a/plugin.video.alfa/channels/animejl.json b/plugin.video.alfa/channels/animejl.json
index aae006fd..9fa69fcd 100644
--- a/plugin.video.alfa/channels/animejl.json
+++ b/plugin.video.alfa/channels/animejl.json
@@ -4,7 +4,7 @@
"active": true,
"adult": false,
"language": ["cast", "lat"],
- "thumbnail": "https://www.animejl.net/img/Logo.png",
+ "thumbnail": "https://i.imgur.com/S6foTE9.png",
"banner": "",
"categories": [
"anime"
@@ -27,4 +27,4 @@
"visible": true
}
]
-}
\ No newline at end of file
+}
diff --git a/plugin.video.alfa/channels/animeshd.py b/plugin.video.alfa/channels/animeshd.py
index 18d53b81..12c4778b 100644
--- a/plugin.video.alfa/channels/animeshd.py
+++ b/plugin.video.alfa/channels/animeshd.py
@@ -113,9 +113,10 @@ def lista(item):
patron = 'class="anime">'
patron +='.*?
([^<]+)<\/h2>'
matches = re.compile(patron, re.DOTALL).findall(data)
- context = renumbertools.context(item)
- context2 = autoplay.context
- context.extend(context2)
+ if item.extra != "next":
+ context = renumbertools.context(item)
+ context2 = autoplay.context
+ context.extend(context2)
for scrapedurl, scrapedthumbnail, scrapedtitle in matches:
url = scrapedurl
thumbnail = host + scrapedthumbnail
@@ -133,12 +134,12 @@ def lista(item):
' (.*?)
[\s\S]+?[\s\S]+?
', re.DOTALL).findall(data)
+ itemlist = []
+ for episode, url, thumbnail,season in matches:
+
+ if item.extra == "watch-series":
+ scrapedinfo = season.split(' - ')
+ scrapedtitle = scrapedinfo[0]
+ season = scrapertools.find_single_match(scrapedinfo[1], 'Season (\d+)')
+ episode = scrapertools.find_single_match(episode, 'Episode (\d+)')
+ title = scrapedtitle + " %sx%s" % (season, episode)
+ else:
+ scrapedtitle = season
+ title = scrapedtitle + ' - ' + episode
+ url = urlparse.urljoin(host, url)
+
+ new_item = Item(channel=item.channel, action="findvideos", title=title, url=url, thumbnail=thumbnail,
+ contentSerieName=scrapedtitle,)
+ itemlist.append(new_item)
+ tmdb.set_infoLabels_itemlist(itemlist, seekTmdb=True, idioma_busqueda='en')
+ if url_pagination:
+ url = urlparse.urljoin(host + item.extra, url_pagination)
+ title = ">> Pagina Siguiente"
+ itemlist.append(Item(channel=item.channel, action="novedades_episodios", title=title, url=url, extra=item.extra))
+ return itemlist
+
+
+def novedades_cine(item):
+ logger.info()
+ data = httptools.downloadpage(item.url).data
+ url_pagination = scrapertools.find_single_match(data, "
(.*?)[\s\S]+?[\s\S]+?
', re.DOTALL).findall(data)
+ itemlist = []
+ for episode, url, thumbnail,season in matches:
+ scrapedyear = '-'
+ title = "%s [%s]" % (season, episode)
+ url = urlparse.urljoin(host, url)
+ new_item = Item(channel=item.channel, action="findvideos",title=title, url=url, contentTitle=season, thumbnail=thumbnail,infoLabels={'year':scrapedyear})
+ itemlist.append(new_item)
+ tmdb.set_infoLabels_itemlist(itemlist, seekTmdb=True, idioma_busqueda='en')
+ if url_pagination:
+ url = urlparse.urljoin(host + item.extra, url_pagination)
+ title = ">> Pagina Siguiente"
+ itemlist.append(Item(channel=item.channel, action="novedades_cine", title=title, url=url))
+ return itemlist
+
+def popular(item):
+ logger.info()
+ data = httptools.downloadpage(item.url).data
+ url_pagination = scrapertools.find_single_match(data, "[\s\S]+?[\s\S]+?
Episode (\d+)', re.DOTALL).findall(data)
+ for url, title, episode in matches:
+ url = urlparse.urljoin(host, url)
+ thumbnail = item.thumbnail
+ title = title + " - Ep. " + episode
+ if " Season " in title:
+ scrapedinfo = title.split(' Season ')
+ title = scrapedinfo[0] + " " + infoLabels['season'] + "x" + episode
+ infoLabels['episode'] = episode
+ itemlist.append(Item(channel=item.channel,
+ action="findvideos",
+ title=title,
+ url=url,
+ thumbnail=thumbnail,
+ infoLabels=infoLabels
+ ))
+ tmdb.set_infoLabels_itemlist(itemlist, seekTmdb=True, idioma_busqueda='en')
+ return itemlist
+
+
+def findvideos(item):
+ logger.info()
+ itemlist = []
+ if "-episode-0" in item.url:
+ data1 = httptools.downloadpage(item.url).data
+ if "Page not found" in data1:
+ item.url = item.url.replace("-episode-0", "-episode-1")
+
+ data = re.sub(r"\n|\r|\t|\s{2}|-\s", "", httptools.downloadpage(item.url).data)
+ matches = scrapertools.find_multiple_matches(data, 'data-video="(.*?)"')
+ url = ''
+ urlsub = ''
+ urlsub = scrapertools.find_single_match(data, "&sub=(.*?)&cover")
+ if urlsub != '':
+ urlsub = base64.b64decode(urlsub)
+ urlsub = 'https://sub.movie-series.net' + urlsub
+ for source in matches:
+ if '/streaming.php' in source:
+ new_data = httptools.downloadpage("https:" + source).data
+ url = scrapertools.find_single_match(new_data, "file: '(https://redirector.*?)'")
+ thumbnail= "https://martechforum.com/wp-content/uploads/2015/07/drive-300x300.png"
+ if url == "":
+ source = source.replace("streaming.php", "load.php")
+ elif '/load.php' in source:
+ new_data = httptools.downloadpage("https:" + source).data
+ url = scrapertools.find_single_match(new_data, "file: '(https://[A-z0-9]+.cdnfile.info/.*?)'")
+ thumbnail= "https://vidcloud.icu/img/logo_vid.png"
+ else:
+ url = source
+ thumbnail= ""
+ if "https://redirector." in url or "cdnfile.info" in url:
+ url = url+"|referer=https://vidcloud.icu/"
+
+ if url != "":
+ itemlist.append(Item(channel=item.channel, url=url, title='%s', action='play',plot=item.plot, thumbnail=thumbnail, subtitle=urlsub))
+
+ itemlist = servertools.get_servers_itemlist(itemlist, lambda i: i.title % i.server)
+ if config.get_videolibrary_support() and len(itemlist) > 0 and item.extra == 'film':
+ itemlist.append(Item(channel=item.channel, title="Añadir a la Videoteca", text_color="yellow",
+ action="add_pelicula_to_library", url=item.url, thumbnail = item.thumbnail,
+ contentTitle = item.contentTitle
+ ))
+ # Requerido para AutoPlay
+
+ autoplay.start(itemlist, item)
+
+ return itemlist
diff --git a/plugin.video.alfa/servers/gamovideo.json b/plugin.video.alfa/servers/gamovideo.json
index 46a93854..1ebc6814 100755
--- a/plugin.video.alfa/servers/gamovideo.json
+++ b/plugin.video.alfa/servers/gamovideo.json
@@ -5,7 +5,7 @@
"patterns": [
{
"pattern": "gamovideo.com/(?:embed-|)([a-z0-9]+)",
- "url": "http://gamovideo.com/embed-\\1.html"
+ "url": "http://gamovideo.com/\\1"
}
]
},
@@ -39,4 +39,4 @@
}
],
"thumbnail": "server_gamovideo.png"
-}
\ No newline at end of file
+}
diff --git a/plugin.video.alfa/servers/gamovideo.py b/plugin.video.alfa/servers/gamovideo.py
index 8a5c6e7d..d22255f1 100755
--- a/plugin.video.alfa/servers/gamovideo.py
+++ b/plugin.video.alfa/servers/gamovideo.py
@@ -7,17 +7,14 @@ from core import scrapertools
from lib import jsunpack
from platformcode import logger
-headers = {"User-Agent":"Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:61.0) Gecko/20100101 Firefox/61.0"}
+headers = {"User-Agent":"Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:66.0) Gecko/20100101 Firefox/66.0"}
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
- data = httptools.downloadpage(page_url)
- if data.code==404:
- data = httptools.downloadpage(page_url, headers=headers, add_referer=True)
- data = data.data
+ data = httptools.downloadpage(page_url, headers=headers, cookies=False).data
- if "File was deleted" in data or "Not Found" in data or "File was locked by administrator" in data:
+ if "File was deleted" in data or "