diff --git a/plugin.video.alfa/addon.xml b/plugin.video.alfa/addon.xml index 613cdb61..553ba30b 100755 --- a/plugin.video.alfa/addon.xml +++ b/plugin.video.alfa/addon.xml @@ -1,5 +1,5 @@  - + @@ -19,15 +19,9 @@ [B]Estos son los cambios para esta versión:[/B] [COLOR green][B]Canales agregados y arreglos[/B][/COLOR] - » serieslan » animemovil - » mundiseries » bajui - » seriesblanco » descargamix - » miradetodo » pelisgratis - » tvseriesdk » ultrapeliculashd - » gamovideo » flashx - » danimados ¤ arreglos internos - [COLOR green]Gracias a [COLOR yellow]Danielr460[/COLOR] por su colaboración en esta versión[/COLOR] - + » seriesblanco » hdfull + » gamovideo ¤ arreglos internos + Navega con Kodi por páginas web para ver sus videos de manera fácil. Browse web pages using Kodi Browse web pages using Kodi, you can easily watch their video content. diff --git a/plugin.video.alfa/channels/allpeliculas.py b/plugin.video.alfa/channels/allpeliculas.py index 6f4ca9e1..06175ca6 100644 --- a/plugin.video.alfa/channels/allpeliculas.py +++ b/plugin.video.alfa/channels/allpeliculas.py @@ -1,7 +1,5 @@ # -*- coding: utf-8 -*- -import urlparse - from core import httptools from core import jsontools from core import scrapertools @@ -59,6 +57,7 @@ def colecciones(item): title = scrapedtitle.capitalize() + " (" + scrapedcantidad + ")" itemlist.append(Item(channel = item.channel, action = "listado_colecciones", + page = 1, thumbnail = host + scrapedthumbnail, title = title, url = host + scrapedurl @@ -71,7 +70,7 @@ def listado_colecciones(item): itemlist = [] data = httptools.downloadpage(item.url).data data_url = scrapertools.find_single_match(data, "data_url: '([^']+)") - post = "page=1" + post = "page=%s" %item.page data = httptools.downloadpage(host + data_url, post=post).data patron = 'a href="(/peli[^"]+).*?' patron += 'src="([^"]+).*?' @@ -88,6 +87,16 @@ def listado_colecciones(item): url = host + scrapedurl )) tmdb.set_infoLabels(itemlist) + item.page += 1 + post = "page=%s" %item.page + data = httptools.downloadpage(host + data_url, post=post).data + if len(data) > 50: + itemlist.append(Item(channel = item.channel, + action = "listado_colecciones", + title = "Pagina siguiente>>", + page = item.page, + url = item.url + )) return itemlist @@ -159,6 +168,7 @@ def lista(item): params = jsontools.dump(dict_param) data = httptools.downloadpage(item.url, post=params).data + data = data.replace("","").replace("<\/mark>","") dict_data = jsontools.load(data) for it in dict_data["items"]: @@ -167,7 +177,7 @@ def lista(item): rating = it["imdb"] year = it["year"] url = host + "pelicula/" + it["slug"] - thumb = urlparse.urljoin(host, it["image"]) + thumb = host + it["image"] item.infoLabels['year'] = year itemlist.append(item.clone(action="findvideos", title=title, fulltitle=title, url=url, thumbnail=thumb, plot=plot, context=["buscar_trailer"], contentTitle=title, contentType="movie")) diff --git a/plugin.video.alfa/channels/anitoonstv.py b/plugin.video.alfa/channels/anitoonstv.py index b463b0ca..58efb7c6 100644 --- a/plugin.video.alfa/channels/anitoonstv.py +++ b/plugin.video.alfa/channels/anitoonstv.py @@ -148,15 +148,21 @@ def findvideos(item): itemla = scrapertools.find_multiple_matches(data_vid, '
.+?-(.+?)-(.+?)<\/div><.+? src="(.+?)"') for server, quality, url in itemla: if "Calidad Alta" in quality: - quality = quality.replace("Calidad Alta", "HQ") + quality = "HQ" + if "HQ" in quality: + quality = "HD" if " Calidad media - Carga mas rapido" in quality: - quality = quality.replace(" Calidad media - Carga mas rapido", "360p") + quality = "360p" server = server.lower().strip() - if "ok" == server: + if "ok" in server: server = 'okru' + if "rapid" in server: + server = 'rapidvideo' + if "netu" in server: + server = 'netutv' itemlist.append(item.clone(url=url, action="play", server=server, contentQuality=quality, thumbnail=scrapedthumbnail, plot=scrapedplot, - title="Enlace encontrado en %s: [%s]" % (server.capitalize(), quality))) + title="Enlace encontrado en: %s [%s]" % (server.capitalize(), quality))) autoplay.start(itemlist, item) return itemlist diff --git a/plugin.video.alfa/channels/hdfull.py b/plugin.video.alfa/channels/hdfull.py index 9e4ad2f1..1a311969 100644 --- a/plugin.video.alfa/channels/hdfull.py +++ b/plugin.video.alfa/channels/hdfull.py @@ -1,3 +1,4 @@ + # -*- coding: utf-8 -*- import base64 @@ -15,6 +16,7 @@ from platformcode import platformtools host = "http://hdfull.tv" +A_A = {'User-Agent':'Mozilla/5.0 AppLeWebKit/537.36 (KHTML, like Gecko) Chrome/59.0.3071.115 SaFAri/537.36'} if config.get_setting('hdfulluser', 'hdfull'): account = True else: @@ -28,7 +30,7 @@ def settingCanal(item): def login(): logger.info() - data = agrupa_datos(httptools.downloadpage(host).data) + data = agrupa_datos(httptools.downloadpage(host, headers=A_A).data) patron = "" sid = scrapertools.find_single_match(data, patron) @@ -37,7 +39,7 @@ def login(): 'hdfull') + "&password=" + config.get_setting( 'hdfullpassword', 'hdfull') + "&action=login" - httptools.downloadpage(host, post=post) + httptools.downloadpage(host, post=post, headers=A_A) def mainlist(item): @@ -137,7 +139,7 @@ def menuseries(item): def search(item, texto): logger.info() - data = agrupa_datos(httptools.downloadpage(host).data) + data = agrupa_datos(httptools.downloadpage(host, headers=A_A).data) sid = scrapertools.get_match(data, '.__csrf_magic. value="(sid:[^"]+)"') item.extra = urllib.urlencode({'__csrf_magic': sid}) + '&menu=search&query=' + texto @@ -173,7 +175,7 @@ def items_usuario(item): itemlist = [] ## Carga estados - status = jsontools.load(httptools.downloadpage(host + '/a/status/all').data) + status = jsontools.load(httptools.downloadpage(host + '/a/status/all', headers=A_A).data) ## Fichas usuario url = item.url.split("?")[0] @@ -187,7 +189,7 @@ def items_usuario(item): next_page = url + "?" + post ## Carga las fichas de usuario - data = httptools.downloadpage(url, post=post).data + data = httptools.downloadpage(url, post=post, headers=A_A).data fichas_usuario = jsontools.load(data) for ficha in fichas_usuario: @@ -255,7 +257,7 @@ def listado_series(item): itemlist = [] - data = agrupa_datos(httptools.downloadpage(item.url).data) + data = agrupa_datos(httptools.downloadpage(item.url, headers=A_A).data) patron = '' matches = re.compile(patron, re.DOTALL).findall(data) @@ -275,10 +277,10 @@ def fichas(item): textoidiomas='' infoLabels=dict() ## Carga estados - status = jsontools.load(httptools.downloadpage(host + '/a/status/all').data) + status = jsontools.load(httptools.downloadpage(host + '/a/status/all', headers=A_A).data) if item.title == "Buscar...": - data = agrupa_datos(httptools.downloadpage(item.url, post=item.extra).data) + data = agrupa_datos(httptools.downloadpage(item.url, post=item.extra, headers=A_A).data) s_p = scrapertools.get_match(data, '

(.*?)