([^<]+)<'
matches = scrapertools.find_multiple_matches(data, patron)
@@ -168,11 +166,10 @@ def playlists(item):
if not scrapedthumbnail.startswith("https"):
scrapedthumbnail = "https:%s" % scrapedthumbnail
if videos:
- scrapedtitle = "%s (%s)" % (scrapedtitle, videos)
+ scrapedtitle = "%s [COLOR red](%s)[/COLOR]" % (scrapedtitle, videos)
itemlist.append(item.clone(action="videos", title=scrapedtitle, url=scrapedurl, thumbnail=scrapedthumbnail,
fanart=scrapedthumbnail))
-
- # Extrae la marca de siguiente página
+ #Extrae la marca de siguiente página
next_page = scrapertools.find_single_match(data, '
.*?href="([^"]+)"')
if next_page:
if "go.php?" in next_page:
@@ -187,37 +184,30 @@ def playlists(item):
def videos(item):
logger.info()
itemlist = []
-
- # Descarga la pagina
+ # Descarga la pagina
data = get_data(item.url)
-
action = "play"
if config.get_setting("menu_info", "porntrex"):
action = "menu_info"
# Extrae las entradas
- patron = '\s*([^<]+)<'
+ patron = '.*?([^<]+)
'
matches = scrapertools.find_multiple_matches(data, patron)
- for scrapedurl, scrapedthumbnail, scrapedtitle in matches:
- scrapedtitle = scrapedtitle.strip()
+ for scrapedurl, scrapedtitle, scrapedthumbnail, quality, duration in matches:
if "go.php?" in scrapedurl:
scrapedurl = urllib.unquote(scrapedurl.split("/go.php?u=")[1].split("&")[0])
scrapedthumbnail = urlparse.urljoin(host, scrapedthumbnail)
- itemlist.append(item.clone(action=action, title=scrapedtitle, url=scrapedurl, thumbnail=scrapedthumbnail,
- fanart=scrapedthumbnail))
-
- # Extrae la marca de siguiente página
- next_page = scrapertools.find_single_match(data, '.*?from:(\d+)')
- if next_page:
- if "from=" in item.url:
- next_page = re.sub(r'&from=(\d+)', '&from=%s' % next_page, item.url)
else:
- next_page = "%s?mode=async&function=get_block&block_id=playlist_view_playlist_view&sort_by" \
- "=added2fav_date&&from=%s" % (item.url, next_page)
- itemlist.append(item.clone(action="videos", title=">> Página Siguiente", url=next_page))
-
+ scrapedurl = urlparse.urljoin(host, scrapedurl)
+ if not scrapedthumbnail.startswith("https"):
+ scrapedthumbnail = "https:%s" % scrapedthumbnail
+ if duration:
+ scrapedtitle = "%s - %s" % (duration, scrapedtitle)
+ if '>HD<' in quality:
+ scrapedtitle += " [COLOR red][HD][/COLOR]"
+ itemlist.append(item.clone(action=action, title=scrapedtitle, url=scrapedurl, thumbnail=scrapedthumbnail, contentThumbnail=scrapedthumbnail,
+ fanart=scrapedthumbnail))
return itemlist
-
def play(item):
logger.info()
itemlist = []
@@ -230,7 +220,7 @@ def play(item):
patron = '
Date: Wed, 5 Sep 2018 11:59:33 -0300
Subject: [PATCH 4/4] porntrex fix
---
plugin.video.alfa/channels/porntrex.py | 15 ++++++++++++++-
1 file changed, 14 insertions(+), 1 deletion(-)
diff --git a/plugin.video.alfa/channels/porntrex.py b/plugin.video.alfa/channels/porntrex.py
index b4d4803c..17387bea 100644
--- a/plugin.video.alfa/channels/porntrex.py
+++ b/plugin.video.alfa/channels/porntrex.py
@@ -10,6 +10,7 @@ from core.item import Item
from platformcode import config, logger
host = "https://www.porntrex.com"
+perpage = 20
def mainlist(item):
@@ -183,6 +184,8 @@ def playlists(item):
def videos(item):
logger.info()
+ if not item.indexp:
+ item.indexp = 1
itemlist = []
# Descarga la pagina
data = get_data(item.url)
@@ -190,9 +193,13 @@ def videos(item):
if config.get_setting("menu_info", "porntrex"):
action = "menu_info"
# Extrae las entradas
- patron = '.*?([^<]+)
'
+ patron = '.*?([^<]+)
'
matches = scrapertools.find_multiple_matches(data, patron)
+ count = 0
for scrapedurl, scrapedtitle, scrapedthumbnail, quality, duration in matches:
+ count += 1
+ if count < item.indexp:
+ continue
if "go.php?" in scrapedurl:
scrapedurl = urllib.unquote(scrapedurl.split("/go.php?u=")[1].split("&")[0])
scrapedthumbnail = urlparse.urljoin(host, scrapedthumbnail)
@@ -204,8 +211,14 @@ def videos(item):
scrapedtitle = "%s - %s" % (duration, scrapedtitle)
if '>HD<' in quality:
scrapedtitle += " [COLOR red][HD][/COLOR]"
+ if len(itemlist) >= perpage:
+ break;
itemlist.append(item.clone(action=action, title=scrapedtitle, url=scrapedurl, thumbnail=scrapedthumbnail, contentThumbnail=scrapedthumbnail,
fanart=scrapedthumbnail))
+ #Extrae la marca de siguiente página
+ if item.channel and len(itemlist) >= perpage:
+ itemlist.append( item.clone(title = "Página siguiente >>>", indexp = count + 1) )
+
return itemlist
def play(item):