Files
addon/channels/titsbox.py
2019-08-28 18:13:01 +02:00

72 lines
2.8 KiB
Python
Executable File

# -*- coding: utf-8 -*-
#------------------------------------------------------------
import urlparse,urllib2,urllib,re
from core import jsontools as json
from platformcode import config, logger
from core import scrapertools
from core.item import Item
from core import servertools
from core import httptools
host = 'https://titsbox.com' #titbox vivud zmovs
url_api = host + "/?ajax=1&type="
def mainlist(item):
logger.info()
itemlist = []
itemlist.append( Item(channel=item.channel, title="Nuevos" , action="lista", url=url_api + "most-recent&page=1"))
itemlist.append( Item(channel=item.channel, title="Mejor valorados" , action="lista", url=url_api + "top-rated&page=1"))
itemlist.append( Item(channel=item.channel, title="Mas vistos" , action="lista", url=url_api + "long&page=1"))
itemlist.append( Item(channel=item.channel, title="Categorias" , action="categorias", url=host))
return itemlist
def categorias(item):
logger.info()
itemlist = []
data = httptools.downloadpage(item.url).data
data = scrapertools.find_single_match(data, '<ul class="sidebar-nav">(.*?)</ul>')
data = re.sub(r"\n|\r|\t|&nbsp;|<br>|<br/>", "", data)
patron = '<a class="category-item" href="([^"]+)">([^"]+)</a>'
matches = re.compile(patron,re.DOTALL).findall(data)
for scrapedurl,scrapedtitle in matches:
url = host + scrapedurl + "?ajax=1&type=most-recent&page=1"
scrapedplot = ""
thumbnail = ""
itemlist.append( Item(channel=item.channel, action="lista", title=scrapedtitle, url=url,
thumbnail=thumbnail , plot=scrapedplot) )
return itemlist
def lista(item):
logger.info()
itemlist = []
data = httptools.downloadpage(item.url).data
data = re.sub(r"\n|\r|\t|&nbsp;|<br>|<br/>", "", data)
JSONData = json.load(data)
for Video in JSONData["data"]:
duration = Video["duration"]
title = Video["videoTitle"]
title = "[COLOR yellow]%s[/COLOR] %s" % (duration,title)
src= Video["src"]
domain=""
thumbnail = src.get('domain', domain) + src.get('pathMedium', domain)+"1.jpg"
url= Video["urls_CDN"]
url= url.get('480', domain)
url = url.replace("/\n/", "/")
plot = ""
itemlist.append( Item(channel=item.channel, action="play", title=title, url=url, thumbnail=thumbnail,
fanart=thumbnail, plot=plot,))
Actual = int(scrapertools.find_single_match(item.url, '&page=([0-9]+)'))
if JSONData["pagesLeft"] - 1 > Actual:
scrapedurl = item.url.replace("&page=" + str(Actual), "&page=" + str(Actual + 1))
itemlist.append(item.clone(action="lista", title="Página Siguiente >>", text_color="blue", url=scrapedurl))
return itemlist
def findvideos(item):
return