From 3a49b8a4429fa19ef70792d4187abe90122247f8 Mon Sep 17 00:00:00 2001 From: danielr460 Date: Sun, 29 Oct 2017 08:54:27 -0500 Subject: [PATCH 01/12] =?UTF-8?q?Funci=C3=B3n=20Play=20eliminaba=20info=20?= =?UTF-8?q?de=20la=20serie.=20Corregido?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- plugin.video.alfa/channels/seriesblanco.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/plugin.video.alfa/channels/seriesblanco.py b/plugin.video.alfa/channels/seriesblanco.py index f105654f..204e9104 100644 --- a/plugin.video.alfa/channels/seriesblanco.py +++ b/plugin.video.alfa/channels/seriesblanco.py @@ -338,7 +338,7 @@ def play(item): else: url = item.url - itemlist = servertools.find_video_items(data=url) + itemlist = servertools.find_video_items(item=item,data=url) titulo = scrapertoolsV2.find_single_match(item.fulltitle, "^(.*?)\s\[.+?$") if titulo: From 11fceffd14a4a5aba7cce5381d472db9e7df7443 Mon Sep 17 00:00:00 2001 From: Intel1 <25161862+Intel11@users.noreply.github.com> Date: Sun, 29 Oct 2017 10:00:39 -0500 Subject: [PATCH 02/12] bajui2: fix --- plugin.video.alfa/channels/bajui2.py | 26 +++++++++++++------------- 1 file changed, 13 insertions(+), 13 deletions(-) diff --git a/plugin.video.alfa/channels/bajui2.py b/plugin.video.alfa/channels/bajui2.py index 7f042be1..efa7a3eb 100755 --- a/plugin.video.alfa/channels/bajui2.py +++ b/plugin.video.alfa/channels/bajui2.py @@ -1,4 +1,4 @@ -# -*- coding: utf-8 -*- +# -*- coding: utf-8 -*- import re import urlparse @@ -13,7 +13,7 @@ def mainlist(item): logger.info() itemlist = [] itemlist.append(Item(channel=item.channel, title="Películas", action="menupeliculas", - url="http://www.bajui2.com/descargas/categoria/2/peliculas", + url="http://www.bajui.org/descargas/categoria/2/peliculas", fanart=item.fanart)) itemlist.append(Item(channel=item.channel, title="Series", action="menuseries", fanart=item.fanart)) @@ -51,13 +51,13 @@ def menuseries(item): logger.info() itemlist = [] itemlist.append(Item(channel=item.channel, title="Series - Novedades", action="peliculas", - url="http://www.bajui2.com/descargas/categoria/3/series", + url="http://www.bajui.org/descargas/categoria/3/series", fanart=item.fanart, viewmode="movie_with_plot")) itemlist.append(Item(channel=item.channel, title="Series - A-Z", action="peliculas", - url="http://www.bajui2.com/descargas/categoria/3/series/orden:nombre", + url="http://www.bajui.org/descargas/categoria/3/series/orden:nombre", fanart=item.fanart, viewmode="movie_with_plot")) itemlist.append(Item(channel=item.channel, title="Series - HD", action="peliculas", - url="http://www.bajui2.com/descargas/subcategoria/11/hd/orden:nombre", + url="http://www.bajui.org/descargas/subcategoria/11/hd/orden:nombre", fanart=item.fanart, viewmode="movie_with_plot")) itemlist.append(Item(channel=item.channel, title="Buscar", action="search", url="", fanart=item.fanart)) @@ -68,10 +68,10 @@ def menudocumentales(item): logger.info() itemlist = [] itemlist.append(Item(channel=item.channel, title="Documentales - Novedades", action="peliculas", - url="http://www.bajui2.com/descargas/categoria/7/docus-y-tv", + url="http://www.bajui.org/descargas/categoria/7/docus-y-tv", fanart=item.fanart, viewmode="movie_with_plot")) itemlist.append(Item(channel=item.channel, title="Documentales - A-Z", action="peliculas", - url="http://www.bajui2.com/descargas/categoria/7/docus-y-tv/orden:nombre", + url="http://www.bajui.org/descargas/categoria/7/docus-y-tv/orden:nombre", fanart=item.fanart, viewmode="movie_with_plot")) itemlist.append(Item(channel=item.channel, title="Buscar", action="search", url="", fanart=item.fanart)) @@ -86,7 +86,7 @@ def search(item, texto, categoria=""): texto = texto.replace(" ", "+") logger.info("categoria: " + categoria + " url: " + url) try: - item.url = "http://www.bajui2.com/descargas/busqueda/%s" + item.url = "http://www.bajui.org/descargas/busqueda/%s" item.url = item.url % texto itemlist.extend(peliculas(item)) return itemlist @@ -118,7 +118,7 @@ def peliculas(item, paginacion=True): scrapedtitle = title scrapedplot = clean_plot(plot) scrapedurl = urlparse.urljoin(item.url, url) - scrapedthumbnail = urlparse.urljoin("http://www.bajui2.com/", thumbnail.replace("_m.jpg", "_g.jpg")) + scrapedthumbnail = urlparse.urljoin("http://www.bajui.org/", thumbnail.replace("_m.jpg", "_g.jpg")) logger.debug("title=[" + scrapedtitle + "], url=[" + scrapedurl + "], thumbnail=[" + scrapedthumbnail + "]") # Añade al listado de XBMC @@ -133,7 +133,7 @@ def peliculas(item, paginacion=True): scrapertools.printMatches(matches) if len(matches) > 0: - scrapedurl = urlparse.urljoin("http://www.bajui2.com/", matches[0]) + scrapedurl = urlparse.urljoin("http://www.bajui.org/", matches[0]) pagitem = Item(channel=item.channel, action="peliculas", title=">> Página siguiente", url=scrapedurl, fanart=item.fanart, viewmode="movie_with_plot") if not paginacion: @@ -197,7 +197,7 @@ def enlaces(item): try: item.thumbnail = scrapertools.get_match(data, '
Date: Sun, 29 Oct 2017 19:50:42 -0500 Subject: [PATCH 03/12] Arreglado serieslan --- plugin.video.alfa/channels/serieslan.py | 40 +++++++++++++++++-------- 1 file changed, 28 insertions(+), 12 deletions(-) diff --git a/plugin.video.alfa/channels/serieslan.py b/plugin.video.alfa/channels/serieslan.py index efdada3e..05d15b36 100644 --- a/plugin.video.alfa/channels/serieslan.py +++ b/plugin.video.alfa/channels/serieslan.py @@ -14,10 +14,7 @@ from channels import autoplay IDIOMAS = {'latino': 'Latino'} list_language = IDIOMAS.values() -list_servers = ['openload', - 'okru', - 'netutv', - 'rapidvideo' +list_servers = ['openload' ] list_quality = ['default'] @@ -49,7 +46,11 @@ def lista(item): patron = '10: + itemlist.append( + Item(channel=item.channel, title="Página Siguiente >>", url=url, action="lista", page=0)) + else: + itemlist.append( + Item(channel=item.channel, title="Página Siguiente >>", url=item.url, action="lista", page=item.page + 1)) tmdb.set_infoLabels(itemlist) return itemlist @@ -93,6 +105,10 @@ def episodios(item): title = "" pat = "/" + if "Mike, Lu & Og"==item.title: + pat="&/" + if "KND" in item.title: + pat="-" # varios episodios en un enlace if len(name.split(pat)) > 1: i = 0 @@ -134,7 +150,7 @@ def findvideos(item): itemlist = [] url_server = "https://openload.co/embed/%s/" - url_api_get_key = "https://serieslan.com/ide.php?i=%s&k=%s" + url_api_get_key = "https://serieslan.com/idx.php?i=%s&k=%s" def txc(key, _str): s = range(256) @@ -157,7 +173,7 @@ def findvideos(item): return res data = httptools.downloadpage(item.url).data - pattern = '
' + pattern = "") if packer != "": From 22827e0f7e5fe4603a83aaa7f2655d699bbd1247 Mon Sep 17 00:00:00 2001 From: Intel1 <25161862+Intel11@users.noreply.github.com> Date: Mon, 30 Oct 2017 10:28:29 -0500 Subject: [PATCH 06/12] Update animemovil.json --- plugin.video.alfa/channels/animemovil.json | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/plugin.video.alfa/channels/animemovil.json b/plugin.video.alfa/channels/animemovil.json index 41ac4634..8f9f296c 100644 --- a/plugin.video.alfa/channels/animemovil.json +++ b/plugin.video.alfa/channels/animemovil.json @@ -3,7 +3,7 @@ "name": "Animemovil", "active": true, "adult": false, - "language": ["cat", "lat"], + "language": ["cast", "lat"], "thumbnail": "https://s1.postimg.org/92ji7stii7/animemovil1.png", "banner": "", "categories": [ From 19101b5310f6473505b1829d544f0793992e8586 Mon Sep 17 00:00:00 2001 From: Intel1 <25161862+Intel11@users.noreply.github.com> Date: Mon, 30 Oct 2017 10:32:27 -0500 Subject: [PATCH 07/12] Delete bajui2.py --- plugin.video.alfa/channels/bajui2.py | 270 --------------------------- 1 file changed, 270 deletions(-) delete mode 100755 plugin.video.alfa/channels/bajui2.py diff --git a/plugin.video.alfa/channels/bajui2.py b/plugin.video.alfa/channels/bajui2.py deleted file mode 100755 index efa7a3eb..00000000 --- a/plugin.video.alfa/channels/bajui2.py +++ /dev/null @@ -1,270 +0,0 @@ -# -*- coding: utf-8 -*- - -import re -import urlparse - -from core import scrapertools -from core import servertools -from core.item import Item -from platformcode import logger - - -def mainlist(item): - logger.info() - itemlist = [] - itemlist.append(Item(channel=item.channel, title="Películas", action="menupeliculas", - url="http://www.bajui.org/descargas/categoria/2/peliculas", - fanart=item.fanart)) - itemlist.append(Item(channel=item.channel, title="Series", action="menuseries", - fanart=item.fanart)) - itemlist.append(Item(channel=item.channel, title="Documentales", action="menudocumentales", - fanart=item.fanart)) - itemlist.append(Item(channel=item.channel, title="Buscar", action="search", - fanart=item.fanart)) - return itemlist - - -def menupeliculas(item): - logger.info() - itemlist = [] - itemlist.append(Item(channel=item.channel, title="Películas - Novedades", action="peliculas", url=item.url, - fanart=item.fanart, viewmode="movie_with_plot")) - itemlist.append( - Item(channel=item.channel, title="Películas - A-Z", action="peliculas", url=item.url + "/orden:nombre", - fanart=item.fanart, viewmode="movie_with_plot")) - - # - data = scrapertools.cache_page(item.url) - data = scrapertools.get_match(data, '') - patron = '([^<]+)' - matches = re.compile(patron, re.DOTALL).findall(data) - for url, title in matches: - scrapedurl = urlparse.urljoin(item.url, url) - itemlist.append(Item(channel=item.channel, title="Películas en " + title, action="peliculas", url=scrapedurl, - fanart=item.fanart, viewmode="movie_with_plot")) - - itemlist.append(Item(channel=item.channel, title="Buscar", action="search", url="", fanart=item.fanart)) - return itemlist - - -def menuseries(item): - logger.info() - itemlist = [] - itemlist.append(Item(channel=item.channel, title="Series - Novedades", action="peliculas", - url="http://www.bajui.org/descargas/categoria/3/series", - fanart=item.fanart, viewmode="movie_with_plot")) - itemlist.append(Item(channel=item.channel, title="Series - A-Z", action="peliculas", - url="http://www.bajui.org/descargas/categoria/3/series/orden:nombre", - fanart=item.fanart, viewmode="movie_with_plot")) - itemlist.append(Item(channel=item.channel, title="Series - HD", action="peliculas", - url="http://www.bajui.org/descargas/subcategoria/11/hd/orden:nombre", - fanart=item.fanart, viewmode="movie_with_plot")) - itemlist.append(Item(channel=item.channel, title="Buscar", action="search", url="", - fanart=item.fanart)) - return itemlist - - -def menudocumentales(item): - logger.info() - itemlist = [] - itemlist.append(Item(channel=item.channel, title="Documentales - Novedades", action="peliculas", - url="http://www.bajui.org/descargas/categoria/7/docus-y-tv", - fanart=item.fanart, viewmode="movie_with_plot")) - itemlist.append(Item(channel=item.channel, title="Documentales - A-Z", action="peliculas", - url="http://www.bajui.org/descargas/categoria/7/docus-y-tv/orden:nombre", - fanart=item.fanart, viewmode="movie_with_plot")) - itemlist.append(Item(channel=item.channel, title="Buscar", action="search", url="", - fanart=item.fanart)) - return itemlist - - -# Al llamarse "search" la función, el launcher pide un texto a buscar y lo añade como parámetro -def search(item, texto, categoria=""): - logger.info(item.url + " search " + texto) - itemlist = [] - url = item.url - texto = texto.replace(" ", "+") - logger.info("categoria: " + categoria + " url: " + url) - try: - item.url = "http://www.bajui.org/descargas/busqueda/%s" - item.url = item.url % texto - itemlist.extend(peliculas(item)) - return itemlist - # Se captura la excepción, para no interrumpir al buscador global si un canal falla - except: - import sys - for line in sys.exc_info(): - logger.error("%s" % line) - return [] - - -def peliculas(item, paginacion=True): - logger.info() - url = item.url - - # Descarga la página - data = scrapertools.cache_page(url) - patron = '
  • 0: - scrapedurl = urlparse.urljoin("http://www.bajui.org/", matches[0]) - pagitem = Item(channel=item.channel, action="peliculas", title=">> Página siguiente", url=scrapedurl, - fanart=item.fanart, viewmode="movie_with_plot") - if not paginacion: - itemlist.extend(peliculas(pagitem)) - else: - itemlist.append(pagitem) - - return itemlist - - -def clean_plot(scrapedplot): - scrapedplot = scrapedplot.replace("\n", "").replace("\r", "") - scrapedplot = re.compile("TÍTULO ORIGINAL[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("AÑO[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("Año[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("DURACIÓN[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("Duración[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("PAIS[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("PAÍS[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("Pais[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("País[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("DIRECTOR[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("DIRECCIÓN[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("Dirección[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("REPARTO[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("Reparto[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("Interpretación[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("GUIÓN[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("Guión[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("MÚSICA[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("Música[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("FOTOGRAFÍA[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("Fotografía[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("PRODUCTORA[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("Producción[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("Montaje[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("Vestuario[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("GÉNERO[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("GENERO[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("Genero[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("Género[^<]+
    ", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("PREMIOS[^<]+
    ", re.DOTALL).sub("", scrapedplot) - - scrapedplot = re.compile("SINOPSIS", re.DOTALL).sub("", scrapedplot) - scrapedplot = re.compile("Sinopsis", re.DOTALL).sub("", scrapedplot) - scrapedplot = scrapertools.htmlclean(scrapedplot) - return scrapedplot - - -def enlaces(item): - logger.info() - itemlist = [] - - data = scrapertools.cache_page(item.url) - - try: - item.plot = scrapertools.get_match(data, '(.*?)') - item.plot = clean_plot(item.plot) - except: - pass - - try: - item.thumbnail = scrapertools.get_match(data, '
    -
  • -
    -
    Enlaces de: - jerobien
    -
    Actualizado: Hace 8 minutos
    - -
    uploaded.combitshare.comfreakshare.comletitbit.netturbobit.netrapidgator.netclz.to
    -
    - ''' - - patron = '
    ]+>Mostrar enlaces
    [^<]+' - patron += '
    (.*?)
    ' - - matches = re.compile(patron, re.DOTALL).findall(data) - scrapertools.printMatches(matches) - logger.debug("matches=" + repr(matches)) - - for thumbnail, usuario, fecha, id, id2, servidores in matches: - # bitshare.comfreakshare.comrapidgator.netturbobit.netmuchshare.netletitbit.netshareflare.netOtros - patronservidores = ' Date: Mon, 30 Oct 2017 10:32:41 -0500 Subject: [PATCH 08/12] Delete bajui2.json --- plugin.video.alfa/channels/bajui2.json | 26 -------------------------- 1 file changed, 26 deletions(-) delete mode 100755 plugin.video.alfa/channels/bajui2.json diff --git a/plugin.video.alfa/channels/bajui2.json b/plugin.video.alfa/channels/bajui2.json deleted file mode 100755 index 2af4b680..00000000 --- a/plugin.video.alfa/channels/bajui2.json +++ /dev/null @@ -1,26 +0,0 @@ -{ - "id": "bajui2", - "name": "Bajui2", - "active": true, - "adult": false, - "language": ["cast"], - "thumbnail": "bajui.png", - "banner": "bajui.png", - "fanart": "bajui.png", - "categories": [ - "movie", - "tvshow", - "documentary", - "vos" - ], - "settings": [ - { - "id": "include_in_global_search", - "type": "bool", - "label": "Incluir en busqueda global", - "default": false, - "enabled": true, - "visible": true - } - ] -} From ae318721ab0de5e12edbcafdaccdc7e509f3ed6d Mon Sep 17 00:00:00 2001 From: Intel1 <25161862+Intel11@users.noreply.github.com> Date: Mon, 30 Oct 2017 10:34:08 -0500 Subject: [PATCH 09/12] Add files via upload --- plugin.video.alfa/channels/bajui.json | 26 +++ plugin.video.alfa/channels/bajui.py | 270 ++++++++++++++++++++++++++ 2 files changed, 296 insertions(+) create mode 100644 plugin.video.alfa/channels/bajui.json create mode 100644 plugin.video.alfa/channels/bajui.py diff --git a/plugin.video.alfa/channels/bajui.json b/plugin.video.alfa/channels/bajui.json new file mode 100644 index 00000000..3d4890d0 --- /dev/null +++ b/plugin.video.alfa/channels/bajui.json @@ -0,0 +1,26 @@ +{ + "id": "bajui", + "name": "Bajui", + "active": true, + "adult": false, + "language": ["cast"], + "thumbnail": "bajui.png", + "banner": "bajui.png", + "fanart": "bajui.png", + "categories": [ + "movie", + "tvshow", + "documentary", + "vos" + ], + "settings": [ + { + "id": "include_in_global_search", + "type": "bool", + "label": "Incluir en busqueda global", + "default": false, + "enabled": true, + "visible": true + } + ] +} diff --git a/plugin.video.alfa/channels/bajui.py b/plugin.video.alfa/channels/bajui.py new file mode 100644 index 00000000..0baf7190 --- /dev/null +++ b/plugin.video.alfa/channels/bajui.py @@ -0,0 +1,270 @@ +# -*- coding: utf-8 -*- + +import re +import urlparse + +from core import scrapertools +from core import servertools +from core.item import Item +from platformcode import logger + + +def mainlist(item): + logger.info() + itemlist = [] + itemlist.append(Item(channel=item.channel, title="Películas", action="menupeliculas", + url="http://www.bajui.org/descargas/categoria/2/peliculas", + fanart=item.fanart)) + itemlist.append(Item(channel=item.channel, title="Series", action="menuseries", + fanart=item.fanart)) + itemlist.append(Item(channel=item.channel, title="Documentales", action="menudocumentales", + fanart=item.fanart)) + itemlist.append(Item(channel=item.channel, title="Buscar", action="search", + fanart=item.fanart)) + return itemlist + + +def menupeliculas(item): + logger.info() + itemlist = [] + itemlist.append(Item(channel=item.channel, title="Películas - Novedades", action="peliculas", url=item.url, + fanart=item.fanart, viewmode="movie_with_plot")) + itemlist.append( + Item(channel=item.channel, title="Películas - A-Z", action="peliculas", url=item.url + "/orden:nombre", + fanart=item.fanart, viewmode="movie_with_plot")) + + # + data = scrapertools.cache_page(item.url) + data = scrapertools.get_match(data, '') + patron = '([^<]+)' + matches = re.compile(patron, re.DOTALL).findall(data) + for url, title in matches: + scrapedurl = urlparse.urljoin(item.url, url) + itemlist.append(Item(channel=item.channel, title="Películas en " + title, action="peliculas", url=scrapedurl, + fanart=item.fanart, viewmode="movie_with_plot")) + + itemlist.append(Item(channel=item.channel, title="Buscar", action="search", url="", fanart=item.fanart)) + return itemlist + + +def menuseries(item): + logger.info() + itemlist = [] + itemlist.append(Item(channel=item.channel, title="Series - Novedades", action="peliculas", + url="http://www.bajui.org/descargas/categoria/3/series", + fanart=item.fanart, viewmode="movie_with_plot")) + itemlist.append(Item(channel=item.channel, title="Series - A-Z", action="peliculas", + url="http://www.bajui.org/descargas/categoria/3/series/orden:nombre", + fanart=item.fanart, viewmode="movie_with_plot")) + itemlist.append(Item(channel=item.channel, title="Series - HD", action="peliculas", + url="http://www.bajui.org/descargas/subcategoria/11/hd/orden:nombre", + fanart=item.fanart, viewmode="movie_with_plot")) + itemlist.append(Item(channel=item.channel, title="Buscar", action="search", url="", + fanart=item.fanart)) + return itemlist + + +def menudocumentales(item): + logger.info() + itemlist = [] + itemlist.append(Item(channel=item.channel, title="Documentales - Novedades", action="peliculas", + url="http://www.bajui.org/descargas/categoria/7/docus-y-tv", + fanart=item.fanart, viewmode="movie_with_plot")) + itemlist.append(Item(channel=item.channel, title="Documentales - A-Z", action="peliculas", + url="http://www.bajui.org/descargas/categoria/7/docus-y-tv/orden:nombre", + fanart=item.fanart, viewmode="movie_with_plot")) + itemlist.append(Item(channel=item.channel, title="Buscar", action="search", url="", + fanart=item.fanart)) + return itemlist + + +# Al llamarse "search" la función, el launcher pide un texto a buscar y lo añade como parámetro +def search(item, texto, categoria=""): + logger.info(item.url + " search " + texto) + itemlist = [] + url = item.url + texto = texto.replace(" ", "+") + logger.info("categoria: " + categoria + " url: " + url) + try: + item.url = "http://www.bajui.org/descargas/busqueda/%s" + item.url = item.url % texto + itemlist.extend(peliculas(item)) + return itemlist + # Se captura la excepción, para no interrumpir al buscador global si un canal falla + except: + import sys + for line in sys.exc_info(): + logger.error("%s" % line) + return [] + + +def peliculas(item, paginacion=True): + logger.info() + url = item.url + + # Descarga la página + data = scrapertools.cache_page(url) + patron = '
  • 0: + scrapedurl = urlparse.urljoin("http://www.bajui.org/", matches[0]) + pagitem = Item(channel=item.channel, action="peliculas", title=">> Página siguiente", url=scrapedurl, + fanart=item.fanart, viewmode="movie_with_plot") + if not paginacion: + itemlist.extend(peliculas(pagitem)) + else: + itemlist.append(pagitem) + + return itemlist + + +def clean_plot(scrapedplot): + scrapedplot = scrapedplot.replace("\n", "").replace("\r", "") + scrapedplot = re.compile("TÍTULO ORIGINAL[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("AÑO[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("Año[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("DURACIÓN[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("Duración[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("PAIS[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("PAÍS[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("Pais[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("País[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("DIRECTOR[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("DIRECCIÓN[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("Dirección[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("REPARTO[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("Reparto[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("Interpretación[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("GUIÓN[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("Guión[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("MÚSICA[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("Música[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("FOTOGRAFÍA[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("Fotografía[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("PRODUCTORA[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("Producción[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("Montaje[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("Vestuario[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("GÉNERO[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("GENERO[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("Genero[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("Género[^<]+
    ", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("PREMIOS[^<]+
    ", re.DOTALL).sub("", scrapedplot) + + scrapedplot = re.compile("SINOPSIS", re.DOTALL).sub("", scrapedplot) + scrapedplot = re.compile("Sinopsis", re.DOTALL).sub("", scrapedplot) + scrapedplot = scrapertools.htmlclean(scrapedplot) + return scrapedplot + + +def enlaces(item): + logger.info() + itemlist = [] + + data = scrapertools.cache_page(item.url) + + try: + item.plot = scrapertools.get_match(data, '(.*?)') + item.plot = clean_plot(item.plot) + except: + pass + + try: + item.thumbnail = scrapertools.get_match(data, '
    +
  • +
    +
    Enlaces de: + jerobien
    +
    Actualizado: Hace 8 minutos
    + +
    uploaded.combitshare.comfreakshare.comletitbit.netturbobit.netrapidgator.netclz.to
    +
    + ''' + + patron = '
    ]+>Mostrar enlaces
    [^<]+' + patron += '
    (.*?)
    ' + + matches = re.compile(patron, re.DOTALL).findall(data) + scrapertools.printMatches(matches) + logger.debug("matches=" + repr(matches)) + + for thumbnail, usuario, fecha, id, id2, servidores in matches: + # bitshare.comfreakshare.comrapidgator.netturbobit.netmuchshare.netletitbit.netshareflare.netOtros + patronservidores = ' Date: Tue, 31 Oct 2017 07:17:04 -0500 Subject: [PATCH 10/12] Update danimados.json --- plugin.video.alfa/channels/danimados.json | 7 ------- 1 file changed, 7 deletions(-) diff --git a/plugin.video.alfa/channels/danimados.json b/plugin.video.alfa/channels/danimados.json index 51d80b24..0bd6230e 100644 --- a/plugin.video.alfa/channels/danimados.json +++ b/plugin.video.alfa/channels/danimados.json @@ -6,13 +6,6 @@ "language": ["lat"], "thumbnail": "https://imgur.com/kU5Lx1S.png", "banner": "https://imgur.com/xG5xqBq.png", - "version": 1, - "changes": [ - { - "date": "24/10/2017", - "description": "Primera version del canal" - } - ], "categories": [ "tvshow" ] From 7b2a3c218129db6da6d78a4f8a6dd226164b2c8c Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Daniel=20Rinc=C3=B3n=20Rodr=C3=ADguez?= Date: Tue, 31 Oct 2017 07:19:55 -0500 Subject: [PATCH 11/12] Update mundiseries.json --- plugin.video.alfa/channels/mundiseries.json | 7 ------- 1 file changed, 7 deletions(-) diff --git a/plugin.video.alfa/channels/mundiseries.json b/plugin.video.alfa/channels/mundiseries.json index 283f80f6..dc01afb0 100644 --- a/plugin.video.alfa/channels/mundiseries.json +++ b/plugin.video.alfa/channels/mundiseries.json @@ -6,13 +6,6 @@ "language": ["cast", "lat"], "thumbnail": "https://imgur.com/GdGMFi1.png", "banner": "https://imgur.com/1bDbYY1.png", - "version": 1, - "changes": [ - { - "date": "23/10/2017", - "description": "Primera versión del canal" - } - ], "categories": [ "tvshow" ] From 99dfa2be58e7f8da058f8d3496b69c2a3f66a38f Mon Sep 17 00:00:00 2001 From: alfa-addon Date: Tue, 31 Oct 2017 20:57:02 -0400 Subject: [PATCH 12/12] v2.3.3 --- plugin.video.alfa/addon.xml | 17 ++++++++--------- 1 file changed, 8 insertions(+), 9 deletions(-) diff --git a/plugin.video.alfa/addon.xml b/plugin.video.alfa/addon.xml index 0468d0e7..613cdb61 100755 --- a/plugin.video.alfa/addon.xml +++ b/plugin.video.alfa/addon.xml @@ -1,5 +1,5 @@  - + @@ -19,14 +19,13 @@ [B]Estos son los cambios para esta versión:[/B] [COLOR green][B]Canales agregados y arreglos[/B][/COLOR] - » cinetux » animemovil - » anitoonstv » cartoonlatino - » seriesblanco » damimados - » mundiseries » serieslan - » cinetux » animemovil - » plusdede » pelisplus - » rapidvideo » flashx - ¤ arreglos internos + » serieslan » animemovil + » mundiseries » bajui + » seriesblanco » descargamix + » miradetodo » pelisgratis + » tvseriesdk » ultrapeliculashd + » gamovideo » flashx + » danimados ¤ arreglos internos [COLOR green]Gracias a [COLOR yellow]Danielr460[/COLOR] por su colaboración en esta versión[/COLOR] Navega con Kodi por páginas web para ver sus videos de manera fácil.