some fixes

This commit is contained in:
mac12m99
2019-06-06 22:57:32 +02:00
parent f1222d0232
commit 45796c324a
20 changed files with 44 additions and 102 deletions

View File

@@ -7,38 +7,3 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
return turbovid.get_video_url(page_url, premium, user, password, video_password, 'deltabit')
##import time
##import urllib
##
##from core import httptools
##from core import scrapertools
##from platformcode import logger
##
##
##def test_video_exists(page_url):
## logger.info("(page_url='%s')" % page_url)
## data = httptools.downloadpage(page_url).data
## if "Not Found" in data or "File Does not Exist" in data:
## return False, "[deltabit] El fichero no existe o ha sido borrado"
## return True, ""
##
##
##def get_video_url(page_url, premium=False, user="", password="", video_password=""):
#### logger.info("(deltabit page_url='%s')" % page_url)
## video_urls = []
## data = httptools.downloadpage(page_url).data
## data = data.replace('"', "'")
## page_url_post = scrapertools.find_single_match(data, "<Form method='POST' action='([^']+)'>")
## imhuman = "&imhuman=" + scrapertools.find_single_match(data, "name='imhuman' value='([^']+)'").replace(" ", "+")
## post = urllib.urlencode({k: v for k, v in scrapertools.find_multiple_matches(data, "name='([^']+)' value='([^']*)'")}) + imhuman
## time.sleep(6)
## data = httptools.downloadpage(page_url_post, post=post).data
#### logger.info("(data page_url='%s')" % data)
## sources = scrapertools.find_single_match(data, 'sources: \[([^\]]+)\]')
##
## for media_url in scrapertools.find_multiple_matches(sources, '"([^"]+)"'):
## media_url = media_url.replace('https:', 'http:')
## ext = scrapertools.get_filename_from_url(media_url)[-4:]
## video_urls.append(["%s [deltabit]" % (ext), media_url])
## return video_urls

View File

@@ -33,5 +33,5 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
for media_url in scrapertools.find_multiple_matches(sources, '"([^"]+)"'):
media_url = media_url.replace('https:', 'http:')
ext = scrapertools.get_filename_from_url(media_url)[-4:]
video_urls.append(["%s [turbovid]" % (ext), media_url])
video_urls.append(["%s [%s]" % (ext, server), media_url])
return video_urls