* Improvements channel Guardaserie.click, Fastsubita and support * Refactor channels series with support & fix * Refactor channels series / anime with support & fix New server animeworld.biz * Fix videolibrary update
380 lines
12 KiB
Python
380 lines
12 KiB
Python
# -*- coding: utf-8 -*-
|
||
# ------------------------------------------------------------
|
||
# Canale per AnimeSaturn
|
||
# Thanks to 4l3x87
|
||
# ----------------------------------------------------------
|
||
import re
|
||
|
||
import urlparse
|
||
|
||
import channelselector
|
||
from core import httptools, tmdb, support, scrapertools, jsontools
|
||
from core.item import Item
|
||
from core.support import log
|
||
from platformcode import logger, config
|
||
from specials import autoplay, autorenumber
|
||
|
||
__channel__ = "animesaturn"
|
||
host = config.get_setting("channel_host", __channel__)
|
||
headers = [['Referer', host]]
|
||
|
||
IDIOMAS = {'Italiano': 'IT'}
|
||
list_language = IDIOMAS.values()
|
||
list_servers = ['openload', 'fembed', 'animeworld']
|
||
list_quality = ['default', '480p', '720p', '1080p']
|
||
|
||
|
||
def mainlist(item):
|
||
log()
|
||
itemlist = []
|
||
support.menu(itemlist, 'Novità bold', 'ultimiep', "%s/fetch_pages.php?request=episodes" % host, 'tvshow')
|
||
support.menu(itemlist, 'Anime bold', 'lista_anime', "%s/animelist?load_all=1" % host)
|
||
support.menu(itemlist, 'Archivio A-Z submenu', 'list_az', '%s/animelist?load_all=1' % host, args=['tvshow', 'alfabetico'])
|
||
support.menu(itemlist, 'Cerca', 'search', host)
|
||
support.aplay(item, itemlist, list_servers, list_quality)
|
||
support.channel_config(item, itemlist)
|
||
|
||
return itemlist
|
||
|
||
|
||
# ----------------------------------------------------------------------------------------------------------------
|
||
def cleantitle(scrapedtitle):
|
||
scrapedtitle = scrapertools.decodeHtmlentities(scrapedtitle.strip())
|
||
scrapedtitle = scrapedtitle.replace('[HD]', '').replace('’', '\'').replace('×', 'x').replace('"', "'")
|
||
year = scrapertools.find_single_match(scrapedtitle, '\((\d{4})\)')
|
||
if year:
|
||
scrapedtitle = scrapedtitle.replace('(' + year + ')', '')
|
||
|
||
return scrapedtitle.strip()
|
||
|
||
|
||
# ================================================================================================================
|
||
|
||
# ----------------------------------------------------------------------------------------------------------------
|
||
def lista_anime(item):
|
||
log()
|
||
itemlist = []
|
||
|
||
PERPAGE = 15
|
||
|
||
p = 1
|
||
if '{}' in item.url:
|
||
item.url, p = item.url.split('{}')
|
||
p = int(p)
|
||
|
||
if '||' in item.url:
|
||
series = item.url.split('\n\n')
|
||
matches = []
|
||
for i, serie in enumerate(series):
|
||
matches.append(serie.split('||'))
|
||
else:
|
||
# Estrae i contenuti
|
||
patron = r'<a href="([^"]+)"[^>]*?>[^>]*?>(.+?)<'
|
||
matches = support.match(item, patron, headers=headers)[0]
|
||
|
||
scrapedplot = ""
|
||
scrapedthumbnail = ""
|
||
for i, (scrapedurl, scrapedtitle) in enumerate(matches):
|
||
if (p - 1) * PERPAGE > i: continue
|
||
if i >= p * PERPAGE: break
|
||
title = cleantitle(scrapedtitle).replace('(ita)', '(ITA)')
|
||
movie = False
|
||
showtitle = title
|
||
if '(ITA)' in title:
|
||
title = title.replace('(ITA)', '').strip()
|
||
showtitle = title
|
||
else:
|
||
title += ' ' + support.typo('Sub-ITA', '_ [] color kod')
|
||
|
||
infoLabels = {}
|
||
if 'Akira' in title:
|
||
movie = True
|
||
infoLabels['year'] = 1988
|
||
|
||
if 'Dragon Ball Super Movie' in title:
|
||
movie = True
|
||
infoLabels['year'] = 2019
|
||
|
||
itemlist.append(
|
||
Item(channel=item.channel,
|
||
extra=item.extra,
|
||
action="episodios" if movie == False else 'findvideos',
|
||
title=title,
|
||
url=scrapedurl,
|
||
thumbnail=scrapedthumbnail,
|
||
fulltitle=showtitle,
|
||
show=showtitle,
|
||
contentTitle=showtitle,
|
||
plot=scrapedplot,
|
||
contentType='episode' if movie == False else 'movie',
|
||
originalUrl=scrapedurl,
|
||
infoLabels=infoLabels,
|
||
folder=True))
|
||
|
||
tmdb.set_infoLabels_itemlist(itemlist, seekTmdb=True)
|
||
autorenumber.renumber(itemlist)
|
||
|
||
# Paginazione
|
||
if len(matches) >= p * PERPAGE:
|
||
support.nextPage(itemlist, item, next_page=(item.url + '{}' + str(p + 1)))
|
||
|
||
return itemlist
|
||
|
||
|
||
# ================================================================================================================
|
||
|
||
|
||
# ----------------------------------------------------------------------------------------------------------------
|
||
def episodios(item):
|
||
log()
|
||
itemlist = []
|
||
|
||
data = httptools.downloadpage(item.url, headers=headers, ignore_response_code=True).data
|
||
anime_id = scrapertools.find_single_match(data, r'\?anime_id=(\d+)')
|
||
# movie or series
|
||
movie = scrapertools.find_single_match(data, r'\Episodi:</b>\s(\d*)\sMovie')
|
||
|
||
data = httptools.downloadpage(
|
||
host + "/loading_anime?anime_id=" + anime_id,
|
||
headers={
|
||
'X-Requested-With': 'XMLHttpRequest'
|
||
}).data
|
||
|
||
patron = r'<td style="[^"]+"><b><strong" style="[^"]+">(.+?)</b></strong></td>\s*'
|
||
patron += r'<td style="[^"]+"><a href="([^"]+)"'
|
||
matches = scrapertools.find_multiple_matches(data, patron)
|
||
|
||
for scrapedtitle, scrapedurl in matches:
|
||
scrapedtitle = cleantitle(scrapedtitle)
|
||
scrapedtitle = re.sub(r'<[^>]*?>', '', scrapedtitle)
|
||
scrapedtitle = '[B]' + scrapedtitle + '[/B]'
|
||
|
||
itemlist.append(
|
||
Item(
|
||
channel=item.channel,
|
||
action="findvideos",
|
||
contentType="episode",
|
||
title=scrapedtitle,
|
||
url=urlparse.urljoin(host, scrapedurl),
|
||
fulltitle=scrapedtitle,
|
||
show=scrapedtitle,
|
||
plot=item.plot,
|
||
fanart=item.thumbnail,
|
||
thumbnail=item.thumbnail))
|
||
|
||
if ((len(itemlist) == 1 and 'Movie' in itemlist[0].title) or movie) and item.contentType != 'movie':
|
||
item.url = itemlist[0].url
|
||
item.contentType = 'movie'
|
||
return findvideos(item)
|
||
|
||
tmdb.set_infoLabels_itemlist(itemlist, seekTmdb=True)
|
||
autorenumber.renumber(itemlist, item)
|
||
support.videolibrary(itemlist, item, 'bold color kod')
|
||
|
||
return itemlist
|
||
|
||
|
||
# ================================================================================================================
|
||
|
||
# ----------------------------------------------------------------------------------------------------------------
|
||
def findvideos(item):
|
||
log()
|
||
originalItem = item
|
||
|
||
if item.contentType == 'movie':
|
||
episodes = episodios(item)
|
||
if len(episodes) > 0:
|
||
item.url = episodes[0].url
|
||
|
||
itemlist = []
|
||
data = httptools.downloadpage(item.url, headers=headers, ignore_response_code=True).data
|
||
data = re.sub(r'\n|\t|\s+', ' ', data)
|
||
patron = r'<a href="([^"]+)"><div class="downloadestreaming">'
|
||
url = scrapertools.find_single_match(data, patron)
|
||
data = httptools.downloadpage(url, headers=headers, ignore_response_code=True).data
|
||
data = re.sub(r'\n|\t|\s+', ' ', data)
|
||
itemlist = support.server(item, data=data)
|
||
|
||
return itemlist
|
||
|
||
|
||
# ================================================================================================================
|
||
|
||
|
||
# ----------------------------------------------------------------------------------------------------------------
|
||
|
||
def ultimiep(item):
|
||
log()
|
||
itemlist = []
|
||
|
||
p = 1
|
||
if '{}' in item.url:
|
||
item.url, p = item.url.split('{}')
|
||
p = int(p)
|
||
|
||
post = "page=%s" % p if p > 1 else None
|
||
|
||
data = httptools.downloadpage(
|
||
item.url, post=post, headers={
|
||
'X-Requested-With': 'XMLHttpRequest'
|
||
}).data
|
||
|
||
patron = r"""<a href='[^']+'><div class="locandina"><img alt="[^"]+" src="([^"]+)" title="[^"]+" class="grandezza"></div></a>\s*"""
|
||
patron += r"""<a href='([^']+)'><div class="testo">(.+?)</div></a>\s*"""
|
||
patron += r"""<a href='[^']+'><div class="testo2">(.+?)</div></a>"""
|
||
matches = re.compile(patron, re.DOTALL).findall(data)
|
||
|
||
for scrapedthumbnail, scrapedurl, scrapedtitle1, scrapedtitle2 in matches:
|
||
scrapedtitle1 = cleantitle(scrapedtitle1)
|
||
scrapedtitle2 = cleantitle(scrapedtitle2)
|
||
scrapedtitle = scrapedtitle1 + ' - ' + scrapedtitle2 + ''
|
||
|
||
title = scrapedtitle
|
||
showtitle = scrapedtitle
|
||
if '(ITA)' in title:
|
||
title = title.replace('(ITA)', '').strip()
|
||
showtitle = title
|
||
else:
|
||
title += ' ' + support.typo('Sub-ITA', '_ [] color kod')
|
||
|
||
|
||
itemlist.append(
|
||
Item(channel=item.channel,
|
||
contentType="episode",
|
||
action="findvideos",
|
||
title=title,
|
||
url=scrapedurl,
|
||
fulltitle=scrapedtitle1,
|
||
show=showtitle,
|
||
thumbnail=scrapedthumbnail))
|
||
|
||
tmdb.set_infoLabels_itemlist(itemlist, seekTmdb=True)
|
||
|
||
# Pagine
|
||
patronvideos = r'data-page="(\d+)" title="Next">Pagina Successiva'
|
||
next_page = scrapertools.find_single_match(data, patronvideos)
|
||
if next_page:
|
||
support.nextPage(itemlist, item, next_page=(item.url + '{}' + next_page))
|
||
|
||
return itemlist
|
||
|
||
|
||
# ================================================================================================================
|
||
|
||
|
||
# ----------------------------------------------------------------------------------------------------------------
|
||
def newest(categoria):
|
||
log(categoria)
|
||
itemlist = []
|
||
item = Item()
|
||
item.url = host
|
||
item.extra = ''
|
||
try:
|
||
if categoria == "anime":
|
||
item.url = "%s/fetch_pages?request=episodios" % host
|
||
item.action = "ultimiep"
|
||
itemlist = ultimiep(item)
|
||
|
||
if itemlist[-1].action == "ultimiep":
|
||
itemlist.pop()
|
||
|
||
# Continua la ricerca in caso di errore
|
||
except:
|
||
import sys
|
||
for line in sys.exc_info():
|
||
logger.error("{0}".format(line))
|
||
return []
|
||
|
||
return itemlist
|
||
|
||
|
||
# ================================================================================================================
|
||
|
||
# ----------------------------------------------------------------------------------------------------------------
|
||
def search_anime(item, texto):
|
||
log(texto)
|
||
itemlist = []
|
||
|
||
data = httptools.downloadpage(host + "/index.php?search=1&key=%s" % texto).data
|
||
jsondata = jsontools.load(data)
|
||
|
||
for title in jsondata:
|
||
data = str(httptools.downloadpage("%s/templates/header?check=1" % host, post="typeahead=%s" % title).data)
|
||
|
||
if 'Anime non esistente' in data:
|
||
continue
|
||
else:
|
||
title = title.replace('(ita)', '(ITA)')
|
||
showtitle = title
|
||
if '(ITA)' in title:
|
||
title = title.replace('(ITA)', '').strip()
|
||
showtitle = title
|
||
else:
|
||
title += ' ' + support.typo('Sub-ITA', '_ [] color kod')
|
||
|
||
url = "%s/anime/%s" % (host, data)
|
||
|
||
itemlist.append(
|
||
Item(
|
||
channel=item.channel,
|
||
contentType="episode",
|
||
action="episodios",
|
||
title=title,
|
||
url=url,
|
||
fulltitle=title,
|
||
show=showtitle,
|
||
thumbnail=""))
|
||
|
||
return itemlist
|
||
|
||
|
||
# ================================================================================================================
|
||
|
||
# ----------------------------------------------------------------------------------------------------------------
|
||
def search(item, texto):
|
||
log(texto)
|
||
itemlist = []
|
||
|
||
try:
|
||
return search_anime(item, texto)
|
||
|
||
except:
|
||
import sys
|
||
for line in sys.exc_info():
|
||
logger.error("%s" % line)
|
||
return []
|
||
|
||
|
||
# ================================================================================================================
|
||
|
||
# ----------------------------------------------------------------------------------------------------------------
|
||
|
||
|
||
def list_az(item):
|
||
log()
|
||
itemlist = []
|
||
|
||
alphabet = dict()
|
||
|
||
# Articoli
|
||
patron = r'<a href="([^"]+)"[^>]*?>[^>]*?>(.+?)<'
|
||
matches = support.match(item, patron, headers=headers)[0]
|
||
|
||
for i, (scrapedurl, scrapedtitle) in enumerate(matches):
|
||
letter = scrapedtitle[0].upper()
|
||
if letter not in alphabet:
|
||
alphabet[letter] = []
|
||
alphabet[letter].append(scrapedurl + '||' + scrapedtitle)
|
||
|
||
for letter in sorted(alphabet):
|
||
itemlist.append(
|
||
Item(channel=item.channel,
|
||
action="lista_anime",
|
||
url='\n\n'.join(alphabet[letter]),
|
||
title=letter,
|
||
fulltitle=letter))
|
||
|
||
return itemlist
|
||
|
||
# ================================================================================================================
|