([^<]+)
' # plot matches = re.compile(patron, re.DOTALL).findall(data) - for scrapedthumbnail, scrapedtitle, scrapedurl in matches: - title = "%s" % (scrapedtitle) + for scrapedthumbnail, scrapedurl, scrapedtitle, plot in matches: - itemlist.append(item.clone(channel=item.channel, action="findvideos", title=title, - url=scrapedurl, thumbnail=scrapedthumbnail, plot='', - viewmode="movie_with_plot", folder=True)) + itemlist.append(item.clone(channel=__channel__, action="findvideos", title=scrapedtitle.capitalize(), + url=scrapedurl, thumbnail=scrapedthumbnail, infoLabels={"plot": plot}, fanart=scrapedthumbnail, + viewmode="movie_with_plot", folder=True, contentTitle=scrapedtitle)) # Extrae el paginador paginacion = scrapertools.find_single_match(data, 'Next ›(.*?)
", "", data) - logger.info(data) + # logger.info(data) patron = 'data-lazy-src="([^"]+)".*?' # img patron += '([^<]+).*?' # title @@ -124,11 +112,9 @@ def categorias(item): for scrapedthumbnail, scrapedurl, scrapedtitle, vids in matches: title = "%s (%s)" % (scrapedtitle, vids.title()) - thumbnail = scrapedthumbnail - url = scrapedurl - itemlist.append(Item(channel=item.channel, action="peliculas", fanart=scrapedthumbnail, - title=title, url=url, thumbnail=thumbnail, plot='', - viewmode="movie_with_plot", folder=True)) + itemlist.append(item.clone(channel=__channel__, action="peliculas", fanart=scrapedthumbnail, + title=title, url=scrapedurl, thumbnail=scrapedthumbnail, + viewmode="movie_with_plot", folder=True)) return itemlist @@ -157,27 +143,22 @@ def sub_search(item): data = httptools.downloadpage(item.url).data data = re.sub(r"\n|\r|\t| |
", "", data) - data = httptools.downloadpage(item.url).data - data = re.sub(r"\n|\r|\t| |
", "", data) - patron_todos = '
([^<]+)
' # plot matches = re.compile(patron, re.DOTALL).findall(data) - for scrapedthumbnail, scrapedtitle, scrapedurl in matches: - title = "%s" % (scrapedtitle) - itemlist.append(item.clone(title=title, url=scrapedurl, - action="findvideos", thumbnail=scrapedthumbnail)) + for scrapedthumbnail, scrapedtitle, scrapedurl, plot in matches: + itemlist.append(item.clone(title=scrapedtitle, url=scrapedurl, plot=plot, fanart=scrapedthumbnail, + action="findvideos", thumbnail=scrapedthumbnail)) paginacion = scrapertools.find_single_match( data, "\d+") if paginacion: - itemlist.append(Item(channel=item.channel, action="sub_search", - title="\xc2\xbb Siguiente \xc2\xbb", url=paginacion)) + itemlist.append(item.clone(channel=__channel__, action="sub_search", + title="\xc2\xbb Siguiente \xc2\xbb", url=paginacion)) return itemlist @@ -186,21 +167,15 @@ def findvideos(item): itemlist = [] data = httptools.downloadpage(item.url).data data = re.sub(r"\n|\r|\t|amp;|\s{2}| ", "", data) - logger.info(data) - patron_todos = '' - data = scrapertools.find_single_match(data, patron_todos) - patron = '