diff --git a/plugin.video.alfa/channels/hentaiid.py b/plugin.video.alfa/channels/hentaiid.py
index fd2f6c07..f6cb51aa 100755
--- a/plugin.video.alfa/channels/hentaiid.py
+++ b/plugin.video.alfa/channels/hentaiid.py
@@ -17,7 +17,6 @@ def mainlist(item):
itemlist = list()
itemlist.append(Item(channel=item.channel, action="series", title="Novedades",
url=urlparse.urljoin(CHANNEL_HOST, "archivos/h2/"), extra="novedades"))
- itemlist.append(Item(channel=item.channel, action="letras", title="Por orden alfabético"))
itemlist.append(Item(channel=item.channel, action="generos", title="Por géneros", url=CHANNEL_HOST))
itemlist.append(Item(channel=item.channel, action="series", title="Sin Censura",
url=urlparse.urljoin(CHANNEL_HOST, "archivos/sin-censura/")))
@@ -25,20 +24,6 @@ def mainlist(item):
url=urlparse.urljoin(CHANNEL_HOST, "archivos/hight-definition/")))
itemlist.append(Item(channel=item.channel, action="series", title="Mejores Hentais",
url=urlparse.urljoin(CHANNEL_HOST, "archivos/ranking-hentai/")))
- itemlist.append(Item(channel=item.channel, action="search", title="Buscar",
- url=urlparse.urljoin(CHANNEL_HOST, "?s=")))
-
- return itemlist
-
-
-def letras(item):
- logger.info()
-
- itemlist = []
-
- for letra in '0ABCDEFGHIJKLMNOPQRSTUVWXYZ':
- itemlist.append(Item(channel=item.channel, action="series", title=letra,
- url=urlparse.urljoin(CHANNEL_HOST, "/?s=letra-%s" % letra.replace("0", "num"))))
return itemlist
@@ -47,49 +32,34 @@ def generos(item):
logger.info()
itemlist = []
- data = httptools.downloadpage(item.url).data
- data = re.sub(r"\n|\r|\t|\s{2}", "", data)
+ data = re.sub(r"\n|\r|\t|\s{2}", "", httptools.downloadpage(item.url).data)
- data = scrapertools.get_match(data, "
' \
- '(.*?)[^<]+
[^<]+
[^<]+
0:
- scrapedurl = match
- scrapedtitle = ">> Pagina Siguiente"
+ if pagination:
+ page = scrapertools.find_single_match(pagination, '>Página\s*(\d+)\s*de\s*\d+<')
+ pattern = 'href="([^"]+)">%s<' % (int(page) + 1)
+ url_page = scrapertools.find_single_match(pagination, pattern)
- itemlist.append(Item(channel=item.channel, action="series", title=scrapedtitle, url=scrapedurl,
- folder=True, viewmode="movies_with_plot"))
+ if url_page:
+ itemlist.append(Item(channel=item.channel, action="series", title=">> Página Siguiente", url=url_page))
return itemlist
@@ -124,9 +89,11 @@ def episodios(item):
logger.info()
itemlist = []
- data = httptools.downloadpage(item.url).data
- data = scrapertools.find_single_match(data, '
(.*?)
')
- patron = '
([^<]+)'
+ data = re.sub(r"\n|\r|\t|\s{2}", "", httptools.downloadpage(item.url).data)
+ pattern = '
Lista de Capítulos
(.*?)
'
+
+ data = scrapertools.find_single_match(data, pattern)
+ patron = '
]+>([^<]+)'
matches = re.compile(patron, re.DOTALL).findall(data)
for scrapedurl, scrapedtitle in matches:
@@ -136,10 +103,9 @@ def episodios(item):
plot = item.plot
# logger.debug("title=[{0}], url=[{1}], thumbnail=[{2}]".format(title, url, thumbnail))
-
itemlist.append(Item(channel=item.channel, action="findvideos", title=title, url=url,
thumbnail=thumbnail, plot=plot, show=item.show, fulltitle="%s %s" % (item.show, title),
- fanart=thumbnail, viewmode="movies_with_plot", folder=True))
+ fanart=thumbnail))
return itemlist
@@ -148,7 +114,8 @@ def findvideos(item):
logger.info()
data = httptools.downloadpage(item.url).data
- patron = '
[^<]+<[iframe|IFRAME].*?[src|SRC]="([^"]+)"'
+
+ patron = '<(?:iframe)?(?:IFRAME)?\s*(?:src)?(?:SRC)?="([^"]+)"'
matches = re.compile(patron, re.DOTALL).findall(data)
for url in matches: