# -*- coding: utf-8 -*- #------------------------------------------------------------ import urlparse,urllib2,urllib,re import os, sys from platformcode import config, logger from core import scrapertools from core.item import Item from core import servertools from core import httptools from core import tmdb from core import jsontools host = 'https://www.tnaflix.com' def mainlist(item): logger.info() itemlist = [] itemlist.append( Item(channel=item.channel, title="Nuevas" , action="peliculas", url=host + "/new/?d=all&period=all")) itemlist.append( Item(channel=item.channel, title="Popular" , action="peliculas", url=host + "/popular/?d=all&period=all")) itemlist.append( Item(channel=item.channel, title="Mejor valorado" , action="peliculas", url=host + "/toprated/?d=all&period=month")) itemlist.append( Item(channel=item.channel, title="Canal" , action="catalogo", url=host + "/channels/all/top-rated/1/all")) itemlist.append( Item(channel=item.channel, title="PornStars" , action="categorias", url=host + "/pornstars")) itemlist.append( Item(channel=item.channel, title="Categorias" , action="categorias", url=host + "/categories/")) itemlist.append( Item(channel=item.channel, title="Buscar", action="search")) return itemlist def search(item, texto): logger.info() texto = texto.replace(" ", "+") item.url = host + "/search.php?what=%s&tab=" % texto try: return peliculas(item) except: import sys for line in sys.exc_info(): logger.error("%s" % line) return [] def catalogo(item): logger.info() itemlist = [] data = httptools.downloadpage(item.url).data patron = '
(\d+)
.*?.*?data-original="([^"]+)"' matches = re.compile(patron,re.DOTALL).findall(data) for cantidad,scrapedurl,scrapedtitle,scrapedthumbnail in matches: scrapedurl = urlparse.urljoin(item.url,scrapedurl) title = scrapedtitle + " (" + cantidad + ")" scrapedplot = "" itemlist.append( Item(channel=item.channel, action="peliculas", title=title , url=scrapedurl , thumbnail=scrapedthumbnail , plot=scrapedplot , folder=True) ) next_page_url = scrapertools.find_single_match(data,'') if next_page_url!="": next_page_url = urlparse.urljoin(item.url,next_page_url) itemlist.append( Item(channel=item.channel , action="catalogo" , title="Página Siguiente >>" , text_color="blue", url=next_page_url , folder=True) ) return itemlist def categorias(item): logger.info() itemlist = [] data = httptools.downloadpage(item.url).data data = re.sub(r"\n|\r|\t| |
", "", data) if item.title=="PornStars" : data = scrapertools.get_match(data,' Hall Of Fame Pornstars(.*?)') patron = '
.*?(.*?).*?([^"]+)' matches = re.compile(patron,re.DOTALL).findall(data) for scrapedurl,scrapedthumbnail,cantidad,scrapedtitle in matches: scrapedplot = "" if item.title=="Categorias" : scrapedthumbnail = "http:" + scrapedthumbnail scrapedurl = urlparse.urljoin(item.url,scrapedurl) if item.title=="PornStars" : scrapedurl = urlparse.urljoin(item.url,scrapedurl) + "?section=videos" scrapedtitle = scrapedtitle + " (" + cantidad + ")" itemlist.append( Item(channel=item.channel, action="peliculas", title=scrapedtitle , url=scrapedurl , thumbnail=scrapedthumbnail , plot=scrapedplot , folder=True) ) next_page_url = scrapertools.find_single_match(data,'') if next_page_url!="": next_page_url = urlparse.urljoin(item.url,next_page_url) itemlist.append( Item(channel=item.channel , action="categorias" , title="Página Siguiente >>" , text_color="blue", url=next_page_url , folder=True) ) return itemlist def peliculas(item): logger.info() itemlist = [] data = httptools.downloadpage(item.url).data data = re.sub(r"\n|\r|\t| |
", "", data) patron = '
([^<]+)
' matches = re.compile(patron,re.DOTALL).findall(data) for scrapedurl,scrapedthumbnail,scrapedtitle,duracion in matches: url = urlparse.urljoin(item.url,scrapedurl) title = "[COLOR yellow]" + duracion + "[/COLOR] " + scrapedtitle contentTitle = title thumbnail = scrapedthumbnail plot = "" year = "" itemlist.append( Item(channel=item.channel, action="play" , title=title , url=url, thumbnail=thumbnail, plot=plot, contentTitle = contentTitle, infoLabels={'year':year} )) next_page_url = scrapertools.find_single_match(data,'
') if next_page_url!="": next_page_url = urlparse.urljoin(item.url,next_page_url) itemlist.append( Item(channel=item.channel , action="peliculas" , title="Página Siguiente >>" , text_color="blue", url=next_page_url , folder=True) ) return itemlist def play(item): logger.info() itemlist = [] data = httptools.downloadpage(item.url).data patron = '' matches = scrapertools.find_multiple_matches(data, patron) for url in matches: itemlist.append(Item(channel=item.channel, action="play", title=item.title, fulltitle=item.fulltitle, url=url, thumbnail=item.thumbnail, plot=item.plot, show=item.title, server="directo", folder=False)) return itemlist