.*?
'
- matches = re.compile(patronvideos, re.DOTALL).findall(data)
-
- if len(matches) > 0:
- scrapedurl = urlparse.urljoin(item.url, matches[0])
- itemlist.append(
- Item(channel=item.channel,
- action="peliculas_tv",
- title="[COLOR lightgreen]" + config.get_localized_string(30992) + "[/COLOR]",
- url=scrapedurl,
- thumbnail="http://2.bp.blogspot.com/-fE9tzwmjaeQ/UcM2apxDtjI/AAAAAAAAeeg/WKSGM2TADLM/s1600/pager+old.png",
- folder=True))
-
- return itemlist
-
-def pelis_top100(item):
- logger.info("kod.guardarefilm peliculas")
- itemlist = []
-
- # Carica la pagina
- data = httptools.downloadpage(item.url, headers=headers).data
-
- # Estrae i contenuti
- patron = r'(.*?\(\d+\))'
- matches = re.compile(patron).findall(data)
-
- for scrapedurl, scrapedtitle in matches:
- html = httptools.downloadpage(scrapedurl, headers=headers).data
- start = html.find("")
- end = html.find("
", start)
- scrapedplot = html[start:end]
- scrapedplot = re.sub(r'<[^>]*>', '', scrapedplot)
- scrapedplot = scrapertools.decodeHtmlentities(scrapedplot)
- scrapedtitle = scrapertools.decodeHtmlentities(scrapedtitle)
- scrapedthumbnail = scrapertools.find_single_match(html, r'class="poster-wrapp">'
- patron += r'[^>]+>[^>]+>[^>]+>[^>]+>[^>]+>(.*?)'
- matches = re.compile(patron, re.DOTALL).findall(data)
- for scrapedtitle, scrapedurl in matches:
- scrapedtitle = scrapertools.decodeHtmlentities(scrapedtitle)
- itemlist.append(
- Item(channel=item.channel,
- action="findvideos",
- contentType="episode",
- title=scrapedtitle,
- url=scrapedurl,
- thumbnail=item.thumbnail,
- extra=item.extra,
- fulltitle=item.fulltitle,
- show=item.show))
-
- if config.get_videolibrary_support() and len(itemlist) != 0:
- itemlist.append(
- Item(channel=item.channel,
- title="[COLOR lightblue]%s[/COLOR]" % config.get_localized_string(30161),
- url=item.url,
- action="add_serie_to_library",
- extra="episodios",
- show=item.show))
-
- return itemlist
-
-
-def findvideos(item):
- logger.info("kod.guardarefilm findvideos")
-
- # Carica la pagina
- data = item.url if item.contentType == "episode" else httptools.downloadpage(item.url).data
-
- itemlist = servertools.find_video_items(data=data)
- for videoitem in itemlist:
- videoitem.title = item.title + videoitem.title
- videoitem.fulltitle = item.fulltitle
- videoitem.thumbnail = item.thumbnail
- videoitem.show = item.show
- videoitem.plot = item.plot
- videoitem.channel = item.channel
- videoitem.contentType = item.contentType
-
- return itemlist
diff --git a/channels/guardaseriecc.json b/channels/guardaseriecc.json
deleted file mode 100644
index 1df288e7..00000000
--- a/channels/guardaseriecc.json
+++ /dev/null
@@ -1,36 +0,0 @@
-{
- "id": "guardaseriecc",
- "name": "Guardaserie.cc",
- "active": true,
- "adult": false,
- "language": ["ita"],
- "thumbnail": "https:\/\/raw.githubusercontent.com\/costaplus\/images\/master\/posters\/guardaseriecc.png",
- "bannermenu": "https:\/\/raw.githubusercontent.com\/costaplus\/images\/master\/posters\/guardaseriecc.png",
- "categories": ["tvshow"],
- "settings": [
- {
- "id": "include_in_global_search",
- "type": "bool",
- "label": "Includi ricerca globale",
- "default": false,
- "enabled": false,
- "visible": false
- },
- {
- "id": "include_in_newest_series",
- "type": "bool",
- "label": "Includi in Novità - Serie TV",
- "default": true,
- "enabled": true,
- "visible": true
- },
- {
- "id": "include_in_newest_italiano",
- "type": "bool",
- "label": "Includi in Novità - Italiano",
- "default": true,
- "enabled": true,
- "visible": true
- }
- ]
-}
diff --git a/channels/guardaseriecc.py b/channels/guardaseriecc.py
deleted file mode 100644
index 232f5a4e..00000000
--- a/channels/guardaseriecc.py
+++ /dev/null
@@ -1,263 +0,0 @@
-# -*- coding: utf-8 -*-
-# ------------------------------------------------------------
-# Ringraziamo Icarus crew
-# Canale per guardaseriecc
-#
-# ----------------------------------------------------------
-import re
-
-from core import httptools, scrapertools, servertools
-from core import tmdb
-from core.item import Item
-from lib import unshortenit
-from platformcode import logger, config
-
-__channel__ = 'guardaseriecc'
-host = config.get_channel_url(__channel__)
-headers = [['Referer', host]]
-
-
-def mainlist(item):
- logger.info("Alfa.leserietv mainlist")
- itemlist = [Item(channel=item.channel,
- action="lista_serie",
- title="[COLOR azure]Tutte le serie[/COLOR]",
- url="%s/serietv/" % host,
- thumbnail=thumbnail_lista,
- fanart=FilmFanart),
- Item(channel=item.channel,
- title="[COLOR azure]Categorie[/COLOR]",
- action="categoria",
- url=host,
- thumbnail=thumbnail_categoria,
- fanart=FilmFanart),
- Item(channel=item.channel,
- action="search",
- title="[COLOR orange]Cerca...[/COLOR]",
- thumbnail=thumbnail_cerca,
- fanart=FilmFanart)]
- return itemlist
-
-
-def categoria(item):
- logger.info("[Alfa].[guardareseriecc] [categoria]")
- itemlist = []
-
- patron = '(.*?)'
- data = httptools.downloadpage(item.url, headers=headers).data
- matches = scrapertools.find_multiple_matches(data, patron)
-
- for scrapedurl, scrapedtitle in matches:
- itemlist.append(Item(channel=item.channel,
- action="lista_serie",
- title="[COLOR azure]" + scrapedtitle + "[/COLOR]",
- url=scrapedurl,
- thumbnail=item.thumbnail,
- fulltitle=scrapedtitle,
- show=scrapedtitle, viewmode="movie"))
-
- return itemlist
-
-
-def lista_serie(item):
- logger.info("[Alfa].[guardareseriecc] [lista_serie]")
- itemlist = []
- patron = '[^<]+'
- data = httptools.downloadpage(item.url, headers=headers).data
- matches = re.compile(patron, re.DOTALL).findall(data)
- scrapertools.printMatches(matches)
-
- for scrapedthumbnail, scrapedtitle, scrapedurl in matches:
- scrapedtitle = scrapedtitle.split("(")[0]
- scrapedtitle = scrapertools.decodeHtmlentities(scrapedtitle).strip()
- itemlist.append(
- Item(channel=item.channel,
- action="episodios",
- title="[COLOR azure]" + scrapedtitle + "[/COLOR]",
- url=scrapedurl,
- thumbnail=scrapedthumbnail,
- fulltitle=scrapedtitle,
- show=scrapedtitle, viewmode="movie"))
-
- # Paginazione
- # ===========================================================
- patron = 'class="current">.*? 0:
- itemlist.append(
- Item(channel=item.channel,
- action="lista_serie",
- title="[COLOR lightgreen]" + config.get_localized_string(30992) + "[/COLOR]",
- url=matches,
- thumbnail=thumbnail_successivo,
- folder=True))
-
- tmdb.set_infoLabels_itemlist(itemlist, seekTmdb=True)
- return itemlist
-
-
-def episodios(item):
- logger.info("[Alfa].[guardareseriecc] [stagione]")
- itemlist = []
-
- patron = '.*?'
- data = httptools.downloadpage(item.url, headers=headers).data
- elenco = scrapertools.find_single_match(data, patron)
-
- patron = '.*?Stagioni.*?[^<]+(.*?)'
- data = httptools.downloadpage(item.url, headers=headers).data
- elenco = scrapertools.find_single_match(data, patron, 0)
-
- patron = '(.*?)'
- elenco_link = scrapertools.find_multiple_matches(elenco, patron)
-
- for scrapedurl, scrapedtitle in elenco_link:
- data = httptools.downloadpage(scrapedurl, headers=headers).data
- if 'protectlink' in data:
- urls = scrapertools.find_multiple_matches(data, r''
-
- if nextpage_regex:
- next_page = scrapertools.find_single_match(data, nextpage_regex)
- if next_page != "":
- itemlist.append(
- Item(channel=item.channel,
- action="list_titles",
- title="[COLOR lightgreen]" + config.get_localized_string(30992) + "[/COLOR]",
- url="%s" % next_page,
- extra=item.extra,
- thumbnail="http://2.bp.blogspot.com/-fE9tzwmjaeQ/UcM2apxDtjI/AAAAAAAAeeg/WKSGM2TADLM/s1600/pager+old.png"))
-
- tmdb.set_infoLabels_itemlist(itemlist, seekTmdb=True)
- return itemlist
-
-def search(item, texto):
- logger.info("[guardogratis.py] search")
- item.url = host + "/?s=" + texto
- try:
- if item.extra == "movie":
- return list_titles(item)
- if item.extra == "tvshow":
- return list_titles(item)
- # Continua la ricerca in caso di errore
- except:
- import sys
- for line in sys.exc_info():
- logger.error("%s" % line)
- return []
-
-def categorie(item):
- logger.info("[guardogratis.py] categorie")
- itemlist = []
-
- if item.url == "":
- item.url = host
-
- # Carica la pagina
- data = httptools.downloadpage(item.url, headers=headers).data
- patronvideos = '