Minori informazioni con log disabilitato

This commit is contained in:
Alhaziel01
2020-11-25 17:54:25 +01:00
parent 9f11eac225
commit 4f498a05f8
157 changed files with 720 additions and 726 deletions

View File

@@ -13,7 +13,7 @@ headers = [['User-Agent', 'Mozilla/5.0 (Windows NT 6.1; WOW64; rv:54.0) Gecko/20
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
# page_url = re.sub('akvideo.stream/(?:video/|video\\.php\\?file_code=)?(?:embed-)?([a-zA-Z0-9]+)','akvideo.stream/video/\\1',page_url)
global data
page = httptools.downloadpage(page_url, headers=headers)
@@ -32,18 +32,18 @@ def test_video_exists(page_url):
# ID, code = scrapertools.find_single_match(data, r"""input\D*id=(?:'|")([^'"]+)(?:'|").*?value='([a-z0-9]+)""")
# post = urllib.urlencode({ID: code})
# logger.info('PAGE DATA' + data)
# logger.debug('PAGE DATA' + data)
if "File Not Found" in data:
return False, config.get_localized_string(70449) % "Akvideo"
return True, ""
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info(" url=" + page_url)
logger.debug(" url=" + page_url)
video_urls = []
global data
# logger.info('PAGE DATA' + data)
# logger.debug('PAGE DATA' + data)
# sitekey = scrapertools.find_single_match(data, 'data-sitekey="([^"]+)')
# captcha = platformtools.show_recaptcha(sitekey, page_url) if sitekey else ''
#

View File

@@ -6,7 +6,7 @@ from platformcode import config, logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global data
data = httptools.downloadpage(page_url, cookies=False).data
if 'File you are looking for is not found.' in data:

View File

@@ -6,7 +6,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
if "no longer exists" in data or "to copyright issues" in data:
return False, config.get_localized_string(70449) % "animeid"
@@ -16,7 +16,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
video_urls = []
label, videourl = scrapertools.find_single_match(data, 'label":"([^"]+)".*?file":"([^"]+)')

View File

@@ -9,7 +9,7 @@ from platformcode import logger, config
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
response = httptools.downloadpage(page_url)
if not response.success or "Not Found" in response.data or "File was deleted" in response.data or "is no longer available" in response.data:
return False, config.get_localized_string(70449) % "anonfile"
@@ -17,7 +17,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
data = httptools.downloadpage(page_url).data
patron = 'download-url.*?href="([^"]+)"'

View File

@@ -9,7 +9,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url)
if data.code == 404:
return False, config.get_localized_string(70449) % "ArchiveOrg"
@@ -17,7 +17,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
data = httptools.downloadpage(page_url).data
patron = '<meta property="og:video" content="([^"]+)">'

View File

@@ -9,7 +9,7 @@ except ImportError:
from urllib import urlencode
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
if 'http://' in page_url: # fastids
page_url = httptools.downloadpage(page_url, follow_redirects=False, only_headers=True).headers['location']
@@ -24,7 +24,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("page_url=" + page_url)
logger.debug("page_url=" + page_url)
video_urls = []
@@ -36,18 +36,18 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
if data_pack:
from lib import jsunpack
data = jsunpack.unpack(data_pack)
logger.info("page_url=" + data)
logger.debug("page_url=" + data)
# URL
url = scrapertools.find_single_match(data, r'"src"value="([^"]+)"')
if not url:
url = scrapertools.find_single_match(data, r'file\s*:\s*"([^"]+)"')
logger.info("URL=" + str(url))
logger.debug("URL=" + str(url))
# URL del vídeo
video_urls.append([".mp4" + " [backin]", url])
for video_url in video_urls:
logger.info("%s - %s" % (video_url[0], httptools.get_url_headers(video_url[1])))
logger.debug("%s - %s" % (video_url[0], httptools.get_url_headers(video_url[1])))
return video_urls

View File

@@ -11,7 +11,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global page
page = httptools.downloadpage(page_url)
if not page.success:
@@ -20,7 +20,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
ext = '.mp4'

View File

@@ -10,7 +10,7 @@ headers = {'User-Agent': 'Mozilla/5.0 (Linux; Android 6.0; Nexus 5 Build/MRA58N)
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
if "Archive no Encontrado" in data:
return False, config.get_localized_string(70449) % "bdupload"
@@ -19,7 +19,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
post = ""
patron = '(?s)type="hidden" name="([^"]+)".*?value="([^"]*)"'

View File

@@ -11,7 +11,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url)
if data.code == 404:
return False, config.get_localized_string(70449) % "CinemaUpload"
@@ -19,7 +19,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
data = httptools.downloadpage(page_url).data
data = re.sub(r'\n|\r|\t|&nbsp;|<br>|\s{2,}', "", data)

View File

@@ -22,7 +22,7 @@ excption = False
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = get_data(page_url.replace(".org", ".me"))
if "File Not Found" in data: return False, config.get_localized_string(70449) % "Clicknupload"
@@ -31,7 +31,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
data = get_data(page_url.replace(".org", ".me"))
@@ -51,7 +51,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
media_url = media.rsplit('/', 1)[0] + "/" + url_strip
video_urls.append([scrapertools.get_filename_from_url(media_url)[-4:] + " [clicknupload]", media_url])
for video_url in video_urls:
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -6,7 +6,7 @@ from lib import jsunpack
from platformcode import logger, config
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global data
data = httptools.downloadpage(page_url).data
if "File Not Found" in data or "File was deleted" in data:
@@ -15,7 +15,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
try:

View File

@@ -8,7 +8,7 @@ from lib import jsunpack
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
html = httptools.downloadpage(page_url)
global data
data = html.data
@@ -18,7 +18,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
global data
# data = httptools.downloadpage(page_url).data

View File

@@ -30,7 +30,7 @@ proxy = "https://www.usa-proxy.org/"
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url, headers=GLOBAL_HEADER).data
if "Este es un clip de muestra" in data:
@@ -44,7 +44,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
#page_url='https://www.crunchyroll.com/es-es/one-piece/episode-891-climbing-up-a-waterfall-a-great-journey-through-the-land-of-wanos-sea-zone-786643'
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
if "crunchyroll.com" in page_url:
media_id = page_url.rsplit("-", 1)[1]
@@ -94,7 +94,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
file_sub = ""
video_urls.append(["%s %sp [crunchyroll]" % (filename, quality), media_url, 0, file_sub])
for video_url in video_urls:
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -6,7 +6,7 @@ from platformcode import logger, config
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global response
response = httptools.downloadpage(page_url, cookies=False)
@@ -18,7 +18,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
cookie = {'Cookie': response.headers["set-cookie"]}
data = response.data.replace("\\", "")
@@ -40,5 +40,5 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
stream_url = stream_url_http
video_urls.append(["%sp .%s [dailymotion]" % (calidad, stream_type), stream_url, 0, subtitle])
for video_url in video_urls:
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -7,7 +7,7 @@ from platformcode import logger
# Returns an array of possible video url's from the page_url
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info()
logger.debug()
page_url = correct_url(page_url)
dd1 = httptools.downloadpage("https://api.alldebrid.com/user/login?agent=mySoft&username=%s&password=%s" %(user, password)).data
token = scrapertools.find_single_match(dd1, 'token":"([^"]+)')

View File

@@ -22,7 +22,7 @@ headers = {'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; WOW64; rv:65.0) Gecko/20
# Returns an array of possible video url's from the page_url
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s' , video_password=%s)" % (page_url, video_password))
logger.debug("(page_url='%s' , video_password=%s)" % (page_url, video_password))
page_url = page_url.replace(".nz/embed", ".nz/")
# Se comprueba si existe un token guardado y sino se ejecuta el proceso de autentificación
token_auth = config.get_setting("token", server="realdebrid")
@@ -99,7 +99,7 @@ def get_enlaces(data):
def authentication():
logger.info()
logger.debug()
try:
client_id = "YTWNFBIJEEBP6"

View File

@@ -8,7 +8,7 @@ from platformcode import logger
def get_long_url(short_url):
logger.info("short_url = '%s'" % short_url)
logger.debug("short_url = '%s'" % short_url)
data = httptools.downloadpage(short_url).data
ysmm = scrapertools.find_single_match(data, "var ysmm = '([^']+)';")

View File

@@ -17,7 +17,7 @@ from platformcode import logger
# Obtiene la URL que hay detrás de un enlace a linkbucks
def get_long_url(short_url):
logger.info("(short_url='%s')" % short_url)
logger.debug("(short_url='%s')" % short_url)
request_headers = []
request_headers.append(["User-Agent",
@@ -33,17 +33,17 @@ def get_long_url(short_url):
while True:
for name, value in response_headers:
if name == "set-cookie":
logger.info("Set-Cookie: " + value)
logger.debug("Set-Cookie: " + value)
cookie_name = scrapertools.scrapertools.find_single_match(value, '(.*?)\=.*?\;')
cookie_value = scrapertools.scrapertools.find_single_match(value, '.*?\=(.*?)\;')
request_headers.append(["Cookie", cookie_name + "=" + cookie_value])
body, response_headers = scrapertools.read_body_and_headers(url, headers=request_headers)
logger.info("body=" + body)
logger.debug("body=" + body)
try:
location = scrapertools.scrapertools.find_single_match(body, '<textarea.*?class="caja_des">([^<]+)</textarea>')
logger.info("location=" + location)
logger.debug("location=" + location)
break
except:
n = n + 1

View File

@@ -38,15 +38,15 @@ servers = get_server_list()
def get_long_urls(data):
logger.info()
logger.debug()
patron = '<a href="http://([^"]+)"'
matches = re.compile(patron, re.DOTALL).findall(data)
for short_url in matches:
if short_url.startswith(tuple(servers)):
logger.info(": " + short_url)
logger.debug(": " + short_url)
longurl_data = httptools.downloadpage(
"http://api.longurl.org/v2/expand?url=" + urllib.quote_plus(short_url)).data
logger.info(longurl_data)
logger.debug(longurl_data)
try:
long_url = scrapertools.scrapertools.find_single_match(longurl_data, '<long-url><!\[CDATA\[(.*?)\]\]></long-url>')
except:

View File

@@ -5,9 +5,9 @@ from platformcode import logger
def get_long_url(short_url):
logger.info("(short_url='%s')" % short_url)
logger.debug("(short_url='%s')" % short_url)
location = scrapertools.get_header_from_response(short_url, header_to_get="location")
logger.info("location=" + location)
logger.debug("location=" + location)
return location

View File

@@ -8,7 +8,7 @@ from platformcode import logger, config
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global data, real_url
page = httptools.downloadpage(page_url)
data = page.data.replace('"', "'")
@@ -20,7 +20,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(deltabit page_url='%s')" % page_url)
logger.debug("(deltabit page_url='%s')" % page_url)
global data, real_url
post = {k: v for k, v in scrapertools.find_multiple_matches(data, "name='([^']+)' value='([^']*)'")}
time.sleep(2.5)

View File

@@ -8,7 +8,7 @@ def test_video_exists(page_url):
# Returns an array of possible video url's from the page_url
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = [["%s %s" % (page_url[-4:], config.get_localized_string(30137)), page_url]]

View File

@@ -7,7 +7,7 @@ from platformcode import logger, config
def test_video_exists(page_url):
global data
logger.info('page url=', page_url)
logger.debug('page url=', page_url)
response = httptools.downloadpage(page_url)
if response.code == 404 or 'File you are looking for is not found' in response.data:
@@ -20,7 +20,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
# from core.support import dbg;dbg()
global data
logger.info("URL", page_url)
logger.debug("URL", page_url)
video_urls = list()
host = "https://dood.to"
@@ -28,7 +28,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
new_url = scrapertools.find_single_match(data, r'<iframe src="([^"]+)"')
if new_url:
data = httptools.downloadpage(host + new_url).data
logger.info('DATA', data)
logger.debug('DATA', data)
label = scrapertools.find_single_match(data, r'type:\s*"video/([^"]+)"')

View File

@@ -9,7 +9,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url)
if data.code == 404:
return False, config.get_localized_string(70449) % "Dostream"
@@ -17,7 +17,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
data = httptools.downloadpage(page_url, headers={"Referer":page_url}).data
patron = '"label":"([^"]+)".*?'

View File

@@ -6,7 +6,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
if "no longer exists" in data or "to copyright issues" in data:
return False, "[Downace] El video ha sido borrado"
@@ -18,7 +18,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
video_urls = []
videourl = scrapertools.find_single_match(data, 'controls preload.*?src="([^"]+)')

View File

@@ -19,10 +19,10 @@ from platformcode import logger
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
page_url = page_url.replace("amp;", "")
data = httptools.downloadpage(page_url).data
logger.info("data=" + data)
logger.debug("data=" + data)
video_urls = []
patron = "video_src.*?(http.*?)%22%2C%22video_timestamp"
matches = re.compile(patron, re.DOTALL).findall(data)
@@ -33,5 +33,5 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
videourl = urllib.unquote(videourl)
video_urls.append(["[facebook]", videourl])
for video_url in video_urls:
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -7,7 +7,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url)
if "Object not found" in data.data or "longer exists on our servers" in data.data:
@@ -18,7 +18,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
if "p,a,c,k,e,d" in data:
@@ -38,6 +38,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
except:
pass
for video_url in video_urls:
logger.info(" %s - %s" % (video_url[0], video_url[1]))
logger.debug(" %s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -6,7 +6,7 @@ from core import jsontools
from platformcode import logger, config
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global data
page_url = re.sub('://[^/]+/', '://feurl.com/', page_url)
@@ -16,14 +16,14 @@ def test_video_exists(page_url):
page_url = page_url.replace("/f/","/v/")
page_url = page_url.replace("/v/","/api/source/")
data = httptools.downloadpage(page_url, post={}).json
logger.info(data)
logger.debug(data)
if "Video not found or" in data or "We are encoding this video" in data:
return False, config.get_localized_string(70449) % "Fembed"
return True, ""
def get_video_url(page_url, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
for file in data['data']:
media_url = file['file']

View File

@@ -8,7 +8,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url, follow_redirects=False)
@@ -18,7 +18,7 @@ def test_video_exists(page_url):
return True, ""
def get_video_url(page_url, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
data = httptools.downloadpage(page_url, follow_redirects=False, only_headers=True)
logger.debug(data.headers)

View File

@@ -4,12 +4,12 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
return True, ""
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
return video_urls

View File

@@ -9,7 +9,7 @@ from platformcode import logger, config
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
response = httptools.downloadpage(page_url)
if "File was deleted" in response.data or "is no longer available" in response.data:
return False, config.get_localized_string(70449) % "filepup"
@@ -17,7 +17,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
page_url = page_url.replace("https","http") + "?wmode=transparent"
data = httptools.downloadpage(page_url).data
@@ -36,5 +36,5 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
video_urls.sort(key=lambda x: x[2])
for video_url in video_urls:
video_url[2] = 0
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -6,7 +6,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
if "File was deleted" in data:
@@ -16,7 +16,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
data = httptools.downloadpage(page_url).data
url = scrapertools.find_single_match(data, '(?i)link:\s*"(https://.*?filescdn\.com.*?mp4)"')

View File

@@ -6,7 +6,7 @@ from platformcode import logger
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("( page_url='%s')")
logger.debug("( page_url='%s')")
video_urls = []
itemlist = []
data1 = ''

View File

@@ -23,7 +23,7 @@ flashx_hash_f = ""
flashx_post = ""
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global flashx_data
try:
flashx_data = httptools.downloadpage(page_url, cookies="xfsts=pfp5dj3e6go1l2o1").data
@@ -53,7 +53,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
pfxfx = ""
data = flashx_data
data = data.replace("\n", "")
@@ -70,8 +70,8 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
matches = scrapertools.find_multiple_matches(mfxfx, '(\w+):(\w+)')
for f, v in matches:
pfxfx += f + "=" + v + "&"
logger.info("mfxfxfx1= %s" % js_fxfx)
logger.info("mfxfxfx2= %s" % pfxfx)
logger.debug("mfxfxfx1= %s" % js_fxfx)
logger.debug("mfxfxfx2= %s" % pfxfx)
if pfxfx == "":
pfxfx = "f=fail&fxfx=6"
coding_url = 'https://www.flashx.co/flashx.php?%s' % pfxfx
@@ -119,14 +119,14 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
filetools.write(subtitle, data)
except:
import traceback
logger.info("Error al descargar el subtítulo: " + traceback.format_exc())
logger.debug("Error al descargar el subtítulo: " + traceback.format_exc())
for media_url, label in media_urls:
if not media_url.endswith("png") and not media_url.endswith(".srt"):
video_urls.append(["." + media_url.rsplit('.', 1)[1] + " [flashx]", media_url, 0, subtitle])
for video_url in video_urls:
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
except:
pass

View File

@@ -8,7 +8,7 @@ from platformcode import logger
# Returns an array of possible video url's from the page_url
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
@@ -17,7 +17,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
page_url = scrapertools.get_header_from_response(page_url, header_to_get="location")
# http://www.4shared.com/flash/player.swf?file=http://dc237.4shared.com/img/392975628/ff297d3f/dlink__2Fdownload_2Flj9Qu-tF_3Ftsid_3D20101030-200423-87e3ba9b/preview.flv&d
logger.info("redirect a '%s'" % page_url)
logger.debug("redirect a '%s'" % page_url)
patron = "file\=([^\&]+)\&"
matches = re.compile(patron, re.DOTALL).findall(page_url)
@@ -29,6 +29,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
video_urls.append(["[fourshared]", page_url])
for video_url in video_urls:
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -22,7 +22,7 @@ headers = {"User-Agent":"Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:%s.0) Geck
DATA = ''
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = alfaresolver.get_data(page_url, False)
@@ -46,7 +46,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = DATA
@@ -87,7 +87,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
video_urls.append([scrapertools.get_filename_from_url(mediaurl)[-4:] + " [gamovideo]", mediaurl])
for video_url in video_urls:
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -5,7 +5,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
if '<h2 class="error">Download error</h2>' in data:
return False, "El enlace no es válido<br/>o ha sido borrado de gigasize"
@@ -13,7 +13,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
return video_urls

View File

@@ -7,14 +7,14 @@ from platformcode import logger
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
# Lo extrae a partir de flashvideodownloader.org
if page_url.startswith("http://"):
url = 'http://www.flashvideodownloader.org/download.php?u=' + page_url
else:
url = 'http://www.flashvideodownloader.org/download.php?u=http://video.google.com/videoplay?docid=' + page_url
logger.info("url=" + url)
logger.debug("url=" + url)
data = httptools.downloadpage(url).data
# Extrae el vídeo
@@ -24,6 +24,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
video_urls.append(["[googlevideo]", newmatches[0]])
for video_url in video_urls:
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -20,14 +20,14 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
global data
data = re.sub(r'"|\n|\r|\t|&nbsp;|<br>|\s{2,}', "", data)
# logger.info('GOUN DATA= '+data)
# logger.debug('GOUN DATA= '+data)
packed_data = scrapertools.find_single_match(data, "javascript'>(eval.*?)</script>")
unpacked = jsunpack.unpack(packed_data)
# logger.info('GOUN DATA= '+unpacked)
# logger.debug('GOUN DATA= '+unpacked)
patron = r"sources..([^\]]+)"
matches = re.compile(patron, re.DOTALL).findall(unpacked)
if not matches:

View File

@@ -42,7 +42,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, user="", password="", video_password=""):
logger.info()
logger.debug()
video_urls = []
urls = []
streams =[]

View File

@@ -7,7 +7,7 @@ from platformcode import config, logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url, cookies=False).data
if 'Not found id' in data:
@@ -17,12 +17,12 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info()
logger.debug()
itemlist = []
logger.info(page_url)
logger.debug(page_url)
data = httptools.downloadpage(page_url, post='').data
logger.info(data)
logger.debug(data)
url = base64.b64decode(data)
itemlist.append([".mp4 [HDLoad]", url])

View File

@@ -8,12 +8,12 @@ from lib.fakeMail import Gmailnator
baseUrl = 'https://hdmario.live'
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global page, data
page = httptools.downloadpage(page_url)
data = page.data
logger.info(page.url)
logger.debug(page.url)
if "the page you are looking for could not be found" in data:
return False, config.get_localized_string(70449) % "HDmario"
@@ -54,12 +54,12 @@ def registerOrLogin(page_url):
else:
import random
import string
logger.info('Registrazione automatica in corso')
logger.debug('Registrazione automatica in corso')
mailbox = Gmailnator()
randPsw = ''.join(random.choice(string.ascii_letters + string.digits) for i in range(10))
captcha = httptools.downloadpage(baseUrl + '/captchaInfo').json
logger.info('email: ' + mailbox.address)
logger.info('pass: ' + randPsw)
logger.debug('email: ' + mailbox.address)
logger.debug('pass: ' + randPsw)
reg = platformtools.dialog_register(baseUrl + '/register/', email=True, password=True, email_default=mailbox.address, password_default=randPsw, captcha_img=captcha['captchaUrl'])
if not reg:
return False
@@ -90,7 +90,7 @@ def registerOrLogin(page_url):
else:
platformtools.dialog_ok('HDmario', 'Hai modificato la mail quindi KoD non sarà in grado di effettuare la verifica in autonomia, apri la casella ' + reg['email']
+ ' e clicca sul link. Premi ok quando fatto')
logger.info('Registrazione completata')
logger.debug('Registrazione completata')
return True
@@ -98,7 +98,7 @@ def registerOrLogin(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
global page, data
page_url = page_url.replace('?', '')
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
if 'unconfirmed' in page.url:
id = page_url.split('/')[-1]
@@ -109,7 +109,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
}
httptools.downloadpage(page.url, post=postData)
mail = mailbox.waitForMail()
logger.info(mail)
logger.debug(mail)
if mail:
code = mail.subject.split(' - ')[0]
page = httptools.downloadpage(page_url + '?code=' + code)
@@ -122,12 +122,12 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
page = httptools.downloadpage(page_url)
data = page.data
logger.info(data)
logger.debug(data)
from lib import jsunpack_js2py
unpacked = jsunpack_js2py.unpack(scrapertools.find_single_match(data, '<script type="text/javascript">\n*\s*\n*(eval.*)'))
# p,a,c,k,e,d data -> xhr.setRequestHeader
secureProof = scrapertools.find_single_match(unpacked, """X-Secure-Proof['"]\s*,\s*['"]([^"']+)""")
logger.info('X-Secure-Proof=' + secureProof)
logger.debug('X-Secure-Proof=' + secureProof)
data = httptools.downloadpage(baseUrl + '/pl/' + page_url.split('/')[-1].replace('?', '') + '.m3u8', headers=[['X-Secure-Proof', secureProof]]).data
filetools.write(xbmc.translatePath('special://temp/hdmario.m3u8'), data, 'w')

View File

@@ -20,7 +20,7 @@ from platformcode import logger
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
post = {}
r = re.findall(r'type="hidden" name="(.+?)"\s* value="?(.+?)">', data)
@@ -45,6 +45,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
video_urls.append([scrapertools.get_filename_from_url(mediaurl)[-4:] + " [hugefiles]", mediaurl])
for video_url in video_urls:
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -10,7 +10,7 @@ from platformcode import logger
data = ""
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global data
data = httptools.downloadpage(page_url)
@@ -22,7 +22,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
logger.error(data)
video_urls = []
patron = 'source src="([^"]+)" type="([^"]+)" res=(\d+)'

View File

@@ -6,7 +6,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
if "The file you were looking for could not be found" in data:
return False, config.get_localized_string(70449) % "jawcloud"
@@ -14,7 +14,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
video_urls = []
videourl = scrapertools.find_single_match(data, 'source src="([^"]+)')

View File

@@ -9,7 +9,7 @@ from platformcode import logger
video_urls = []
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
subtitles = ""
response = httptools.downloadpage(page_url)
@@ -21,7 +21,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
media_url = scrapertools.find_single_match(data, '<video src="([^"]+)"')
if media_url:

View File

@@ -7,7 +7,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
page_url = page_url.replace("embed/", "").replace(".html", ".json")
data = httptools.downloadpage(page_url).data
if '"error":"video_not_found"' in data or '"error":"Can\'t find VideoInstance"' in data:
@@ -17,7 +17,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % (page_url))
logger.debug("(page_url='%s')" % (page_url))
video_urls = []
# Carga la página para coger las cookies
@@ -45,6 +45,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
pass
for video_url in video_urls:
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -6,7 +6,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
if "Invalid or Deleted File" in data or "Well, looks like we" in data:
return False, config.get_localized_string(70449) % "Mediafire"
@@ -16,7 +16,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
data = httptools.downloadpage(page_url).data
patron = "DownloadButtonAd-startDownload gbtnSecondary.*?href='([^']+)'"
@@ -27,5 +27,5 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
if len(matches) > 0:
video_urls.append([matches[0][-4:] + " [mediafire]", matches[0]])
for video_url in video_urls:
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -79,7 +79,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
page_url = page_url.replace('/embed#', '/#')
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
# si hay mas de 5 archivos crea un playlist con todos

View File

@@ -10,7 +10,7 @@ from platformcode import logger, config
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global data
data = httptools.downloadpage(page_url).data
@@ -25,7 +25,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
ext = '.mp4'

View File

@@ -15,7 +15,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = re.sub(r"\n|\r|\t|\s{2}", "", httptools.downloadpage(page_url).data)
match = scrapertools.find_single_match(data, "<script type='text/javascript'>(.*?)</script>")
data = jsunpack.unpack(match)
@@ -23,9 +23,9 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
media_url = scrapertools.find_single_match(data, '{type:"video/mp4",src:"([^"]+)"}')
if not media_url:
media_url = scrapertools.find_single_match(data, '"file":"([^"]+)')
logger.info("media_url=" + media_url)
logger.debug("media_url=" + media_url)
video_urls = list()
video_urls.append([scrapertools.get_filename_from_url(media_url)[-4:] + " [mp4upload]", media_url])
for video_url in video_urls:
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -21,7 +21,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info()
logger.debug()
video_urls = []
data = httptools.downloadpage(page_url).data
data = scrapertools.find_single_match(data, 'var srca = \[(.*?)\]')

View File

@@ -10,7 +10,7 @@ import re
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url)
global page_data
page_data = data.data
@@ -21,7 +21,7 @@ def test_video_exists(page_url):
return True, ""
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
global page_data
video_url = scrapertools.find_single_match(decode(page_data), r"'src',\s*'([^']+)")

View File

@@ -23,7 +23,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info()
logger.debug()
video_urls = []
data = httptools.downloadpage(page_url).data
matches = scrapertools.find_multiple_matches(data, 'tracker: "([^"]+)"')

View File

@@ -25,7 +25,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
#Deshabilitamos el server hasta nueva orden
return False, "[netutv] Servidor deshabilitado"
# http://netu.tv/watch_video.php=XX solo contiene una redireccion, ir directamente a http://hqq.tv/player/embed_player.php?vid=XX
@@ -37,7 +37,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
if "hash=" in page_url:

View File

@@ -12,7 +12,7 @@ from platformcode import logger, config
headers = [['User-Agent', 'Mozilla/5.0 (Windows NT 6.1; WOW64; rv:54.0) Gecko/20100101 Firefox/54.0']]
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
if "Not Found" in data or "File was deleted" in data or "The file is being converted" in data or "Please try again later" in data:
return False, config.get_localized_string(70293) % "NowVideo"
@@ -23,7 +23,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
host = 'http://nowvideo.club'
logger.info("(nowvideo page_url='%s')" % page_url)
logger.debug("(nowvideo page_url='%s')" % page_url)
video_urls = []
data = httptools.downloadpage(page_url).data
page_url_post = scrapertools.find_single_match(data, '<Form id="[^"]+" method="POST" action="([^"]+)">')
@@ -31,7 +31,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
imhuman = '&imhuman=' + scrapertools.find_single_match(data, 'name="imhuman" value="([^"]+)"').replace(" ", "+")
post = urllib.urlencode({k: v for k, v in scrapertools.find_multiple_matches(data, 'name="([^"]+)" value="([^"]*)"')}) + imhuman
data = httptools.downloadpage(host + page_url_post, post=post).data
logger.info("nowvideo data page_url2 ='%s'" % data)
logger.debug("nowvideo data page_url2 ='%s'" % data)
headers.append(['Referer', page_url])
post_data = scrapertools.find_single_match(data,"</div>\s*<script>(eval.function.p,a,c,k,e,.*?)\s*</script>")
@@ -46,11 +46,11 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
_headers = urllib.urlencode(dict(headers))
for media_url in media_urls:
#logger.info("nowvideo data page_url2 ='%s'" % media_url)
#logger.debug("nowvideo data page_url2 ='%s'" % media_url)
video_urls.append([" mp4 [nowvideo] ", media_url + '|' + _headers])
for video_url in media_urls:
logger.info("[nowvideo.py] %s - %s" % (video_url[0], video_url[1]))
logger.debug("[nowvideo.py] %s - %s" % (video_url[0], video_url[1]))
return video_urls
@@ -60,7 +60,7 @@ def find_videos(data):
devuelve = []
patronvideos = r"nowvideo.club/(?:play|videos)?([a-z0-9A-Z]+)"
logger.info("[nowvideo.py] find_videos #" + patronvideos + "#")
logger.debug("[nowvideo.py] find_videos #" + patronvideos + "#")
matches = re.compile(patronvideos, re.DOTALL).findall(data)
for match in matches:
@@ -68,10 +68,10 @@ def find_videos(data):
url = 'http://nowvideo.club/%s' % match
if url not in encontrados:
logger.info(" url=" + url)
logger.debug(" url=" + url)
devuelve.append([titulo, url, 'nowvideo'])
encontrados.add(url)
else:
logger.info(" url duplicada=" + url)
logger.debug(" url duplicada=" + url)
return devuelve

View File

@@ -8,7 +8,7 @@ from platformcode import logger, config
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
if "copyrightsRestricted" in data or "COPYRIGHTS_RESTRICTED" in data:
@@ -20,7 +20,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
data = httptools.downloadpage(page_url).data

View File

@@ -21,25 +21,25 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
if config.get_setting("premium", server="onefichier"):
user = config.get_setting("user", server="onefichier")
password = config.get_setting("password", server="onefichier")
url = "https://1fichier.com/login.pl"
logger.info("url=" + url)
logger.debug("url=" + url)
post_parameters = {"mail": user, "pass": password, "lt": "on", "purge": "on", "valider": "Send"}
post = urllib.urlencode(post_parameters)
logger.info("post=" + post)
logger.debug("post=" + post)
data = httptools.downloadpage(url, post=post).data
# logger.info("data="+data)
# logger.debug("data="+data)
cookies = config.get_cookie_data()
logger.info("cookies=" + cookies)
logger.debug("cookies=" + cookies)
# 1fichier.com TRUE / FALSE 1443553315 SID imC3q8MQ7cARw5tkXeWvKyrH493rR=1yvrjhxDAA0T0iEmqRfNF9GXwjrwPHssAQ
sid_cookie_value = scrapertools.find_single_match(cookies, "1fichier.com.*?SID\s+([A-Za-z0-9\+\=]+)")
logger.info("sid_cookie_value=" + sid_cookie_value)
logger.debug("sid_cookie_value=" + sid_cookie_value)
# .1fichier.com TRUE / FALSE 1443553315 SID imC3q8MQ7cARw5tkXeWvKyrH493rR=1yvrjhxDAA0T0iEmqRfNF9GXwjrwPHssAQ
cookie = urllib.urlencode({"SID": sid_cookie_value})
@@ -50,16 +50,16 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
'Mozilla/5.0 (Macintosh; U; Intel Mac OS X 10.6; es-ES; rv:1.9.2.12) Gecko/20101026 Firefox/3.6.12'])
headers.append(['Cookie', cookie])
filename = scrapertools.get_header_from_response(page_url, header_to_get="Content-Disposition")
logger.info("filename=" + filename)
logger.debug("filename=" + filename)
# Construye la URL final para Kodi
location = page_url + "|Cookie=" + cookie
logger.info("location=" + location)
logger.debug("location=" + location)
video_urls = []
video_urls.append([filename[-4:] + " (Premium) [1fichier]", location])
for video_url in video_urls:
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -6,7 +6,7 @@ from platformcode import config, logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global data
data = httptools.downloadpage(page_url).data
@@ -17,8 +17,8 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
global data
# logger.info(data)
# logger.debug(data)
video_urls = support.get_jwplayer_mediaurl(data, 'Onlystream')
return video_urls

View File

@@ -8,6 +8,6 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
return video_urls

View File

@@ -10,7 +10,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url)
if "Object not found" in data.data or "longer exists on our servers" in data.data:
@@ -21,7 +21,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
if "p,a,c,k,e,d" in data:
@@ -44,6 +44,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
except:
pass
for video_url in video_urls:
logger.info(" %s - %s" % (video_url[0], video_url[1]))
logger.debug(" %s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -24,13 +24,13 @@ from core import jsontools
def get_source(url):
logger.info()
logger.debug()
data = httptools.downloadpage(url).data
data = re.sub(r'\n|\r|\t|&nbsp;|<br>|\s{2,}', "", data)
return data
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = get_source(page_url)
if "File was deleted" in data or "File Not Found" in data:
@@ -40,7 +40,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
referer = ''

View File

@@ -10,7 +10,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global data
data = httptools.downloadpage(page_url).data
if "Not Found" in data or "File was deleted" in data:
@@ -19,7 +19,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
ext = 'mp4'

View File

@@ -9,7 +9,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
data = scrapertools.httptools.downloadpage(page_url).data
media_url = scrapertools.find_single_match(data, 'var\s+video_source\s+\=\s+"([^"]+)"')
@@ -24,5 +24,5 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
else:
video_urls.append([scrapertools.get_filename_from_url(media_url)[-4:] + " [sendvid]", media_url])
for video_url in video_urls:
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -5,7 +5,7 @@ from core import httptools, scrapertools
from platformcode import config, logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
@@ -15,22 +15,22 @@ def test_video_exists(page_url):
return True, ""
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
quality ={'MOBILE':1,
'NORMAL':2,
'HD':3}
data = httptools.downloadpage(page_url).data
logger.info('SPEEDVIDEO DATA '+ data)
logger.debug('SPEEDVIDEO DATA '+ data)
media_urls = scrapertools.find_multiple_matches(data, r"file:[^']'([^']+)',\s*label:[^\"]\"([^\"]+)\"")
logger.info("speed video - media urls: %s " % media_urls)
logger.debug("speed video - media urls: %s " % media_urls)
for media_url, label in media_urls:
media_url = httptools.downloadpage(media_url, only_headers=True, follow_redirects=False).headers.get("location", "")
if media_url:
video_urls.append([media_url.split('.')[-1] + ' - ' + label + ' - ' + ' [Speedvideo]', media_url])
logger.info("speed video - media urls: %s " % video_urls)
logger.debug("speed video - media urls: %s " % video_urls)
return sorted(video_urls, key=lambda x: quality[x[0].split(' - ')[1]])

View File

@@ -11,7 +11,7 @@ if sys.version_info[0] >= 3: PY3 = True; unicode = str; unichr = chr; long = int
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global data
referer = {"Referer": page_url}
@@ -25,7 +25,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
possible_url = scrapertools.find_single_match(data, 'innerHTML = "([^"]+)')

View File

@@ -8,7 +8,7 @@ from lib import jsunpack
def test_video_exists(page_url):
global data
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
if "<font color=\"red\"><b>File not found, sorry!" in data:
@@ -17,7 +17,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, video_password):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
from core.support import match
matches = match(data, patron=r'(eval\(function\(p,a,c,k,e,d\).*?)\s+</script>').matches

View File

@@ -9,7 +9,7 @@ from platformcode import config, logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global data
data = httptools.downloadpage(page_url, cookies=False).data
if 'File is no longer available as it expired or has been deleted' in data:
@@ -19,7 +19,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
# data = httptools.downloadpage(page_url).data
global data

View File

@@ -8,7 +8,7 @@ from platformcode import logger, config
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
if "Video not found..." in data or "Video removed due to copyright" in data:
return False, config.get_localized_string(70292) % "Thevid"
@@ -31,5 +31,5 @@ def get_video_url(page_url, user="", password="", video_password=""):
continue
video = "https:" + video
video_urls.append(["mp4 [Thevid]", video])
logger.info("Url: %s" % videos)
logger.debug("Url: %s" % videos)
return video_urls

View File

@@ -9,7 +9,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
if "no longer exists" in data or "to copyright issues" in data:
return False, config.get_localized_string(70449) % "thevideobee"
@@ -17,7 +17,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
video_urls = []
videourl = scrapertools.find_single_match(data, 'src: "([^"]+)')

View File

@@ -4,6 +4,6 @@ from platformcode import logger
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
return video_urls

View File

@@ -12,7 +12,7 @@ from platformcode import logger, config
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
if "Not Found" in data or "File Does not Exist" in data:
return False, config.get_localized_string(70449) % "Turbovid"
@@ -21,7 +21,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password="", server='Turbovid'):
logger.info("(turbovid page_url='%s')" % page_url)
logger.debug("(turbovid page_url='%s')" % page_url)
video_urls = []
data = httptools.downloadpage(page_url).data
data = data.replace('"', "'")
@@ -31,6 +31,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
time.sleep(6)
data = httptools.downloadpage(page_url_post, post=post).data
logger.info("(data page_url='%s')" % data)
logger.debug("(data page_url='%s')" % data)
video_urls = support.get_jwplayer_mediaurl(data, 'Turbovid')
return video_urls

View File

@@ -9,7 +9,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
if "no longer exists" in data or "to copyright issues" in data:
return False, config.get_localized_string(70449) % "tusfiles"
@@ -17,7 +17,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
video_urls = []
videourl = scrapertools.find_single_match(data, 'source src="([^"]+)')

View File

@@ -6,7 +6,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
real_url = page_url.replace("uploaded.to", "uploaded.net")
code = httptools.downloadpage(real_url, only_headers=True).code
@@ -16,36 +16,36 @@ def test_video_exists(page_url):
return True, ""
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
if premium:
#Si no hay almacenada una cookie activa, hacemos login
if check_cookie("uploaded.net", "login") != True:
# Login para conseguir la cookie
logger.info("-------------------------------------------")
logger.info("login")
logger.info("-------------------------------------------")
logger.debug("-------------------------------------------")
logger.debug("login")
logger.debug("-------------------------------------------")
login_url = "http://uploaded.net/io/login"
post = "id=" + user + "&pw=" + password
setcookie = httptools.downloadpage(login_url, post=post, follow_redirects=False,
only_headers=True).headers.get("set-cookie", "")
logger.info("-------------------------------------------")
logger.info("obtiene la url")
logger.info("-------------------------------------------")
logger.debug("-------------------------------------------")
logger.debug("obtiene la url")
logger.debug("-------------------------------------------")
location = httptools.downloadpage(page_url, follow_redirects=False, only_headers=True).headers.get("location",
"")
logger.info("location=" + location)
logger.debug("location=" + location)
#fix descarga no directa
if location == "":
data = httptools.downloadpage(page_url).data
#logger.info("data: %s" % data)
#logger.debug("data: %s" % data)
if "<h1>Premium Download</h1>" in data:
location = scrapertools.find_single_match(data, '<form method="post" action="([^"]+)"')
#logger.info("location: %s" % location)
#logger.debug("location: %s" % location)
elif "Hybrid-Traffic is completely exhausted" in data:
logger.error("Trafico agotado")
@@ -53,13 +53,13 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
logger.error("Cuenta Free")
else:
logger.error("Error Desconocido")
logger.info("-------------------------------------------")
logger.info("obtiene el nombre del fichero")
logger.info("-------------------------------------------")
logger.debug("-------------------------------------------")
logger.debug("obtiene el nombre del fichero")
logger.debug("-------------------------------------------")
try:
content_disposition = httptools.downloadpage(location, post="", follow_redirects=False,
only_headers=True).headers.get("content-disposition", "")
logger.info("content_disposition=" + content_disposition)
logger.debug("content_disposition=" + content_disposition)
if content_disposition != "":
filename = scrapertools.find_single_match(content_disposition, 'filename="([^"]+)"')
extension = filename[-4:]
@@ -74,7 +74,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
video_urls.append([extension + " (Premium) [uploaded.to]", location])
for video_url in video_urls:
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -9,7 +9,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = get_source(page_url)
if "File was deleted" in data or "File Not Found" in data:
@@ -19,7 +19,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info()
logger.debug()
video_urls = []
data = httptools.downloadpage(page_url).data
fid = scrapertools.find_single_match(data, ' name="id" value="([^"]+)"')

View File

@@ -9,7 +9,7 @@ from platformcode import logger, config
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global data
data = httptools.downloadpage(page_url).data
if "<h2>WE ARE SORRY</h2>" in data or '<title>404 Not Found</title>' in data:

View File

@@ -19,7 +19,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
@@ -37,7 +37,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
# Si el enlace es directo de upstream
if "uptobox" not in page_url:
data = httptools.downloadpage(page_url).data
@@ -60,7 +60,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
video_urls = uptobox(page_url, data)
for video_url in video_urls:
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -13,7 +13,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url)
if data.code == 404:
return False, config.get_localized_string(70449) % "upvid"
@@ -23,7 +23,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium = False, user = "", password = "", video_password = ""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
headers = {'referer': page_url}
for i in range(0, 3):

View File

@@ -10,7 +10,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url)
@@ -21,7 +21,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
data = httptools.downloadpage(page_url).data

View File

@@ -7,7 +7,7 @@ from platformcode import logger, config
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
response = httptools.downloadpage(page_url)
@@ -18,7 +18,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
unpacked = ""
data = httptools.downloadpage(page_url).data
@@ -37,6 +37,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
video_urls.append(["%s [userscloud]" % ext, media_url])
for video_url in video_urls:
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -18,7 +18,7 @@ from platformcode import logger, config
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
if "File was deleted" in data or "Page Cannot Be Found" in data or "<title>Video not found" in data:
return False, config.get_localized_string(70449) % "vevio"
@@ -26,7 +26,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
post = {}
post = urllib.urlencode(post)

View File

@@ -13,7 +13,7 @@ from platformcode import logger, config
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
if "We're Sorry" in data:
return False, config.get_localized_string(70292) % "Vidcloud"
@@ -22,7 +22,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
@@ -56,6 +56,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
video_urls.append(["%s [Vidcloud" % ext, media_url])
for video_url in video_urls:
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -11,7 +11,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
if "borrado" in data or "Deleted" in data:
return False, config.get_localized_string(70449) % "videobin"
@@ -19,7 +19,7 @@ def test_video_exists(page_url):
return True, ""
def get_video_url(page_url, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
data = httptools.downloadpage(page_url).data
bloque = scrapertools.find_single_match(data, 'sources:.\[.*?]')

View File

@@ -5,7 +5,7 @@ from platformcode import logger
def get_video_url(page_url, video_password):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
data = httptools.downloadpage(page_url).data
m= scrapertools.find_single_match(data, '<link href="(Br74.*?==.css)"')

View File

@@ -9,7 +9,7 @@ from platformcode import logger
video_urls = []
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
subtitles = ""
response = httptools.downloadpage(page_url)
@@ -21,9 +21,9 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
logger.info("Intel11 %s" %data)
logger.debug("Intel11 %s" %data)
media_url = scrapertools.find_single_match(data, 'file:"([^"]+)')
if media_url:
ext = media_url[-4:]

View File

@@ -11,7 +11,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global data
data = httptools.downloadpage(page_url).data
if "borrado" in data or "Deleted" in data:
@@ -21,7 +21,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
bloque = scrapertools.find_single_match(data, 'sources:.\[.*?]')

View File

@@ -5,7 +5,7 @@ from platformcode import logger, config
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global data
resp = httptools.downloadpage(page_url)
data = resp.data
@@ -15,7 +15,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
global data
video_urls = support.get_jwplayer_mediaurl(data, 'Vidmoly')

View File

@@ -8,7 +8,7 @@ from platformcode import logger, config
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global data
data = httptools.downloadpage(page_url).data
if "Page not found" in data or "File was deleted" in data:
@@ -20,7 +20,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global data
video_urls = []

View File

@@ -8,7 +8,7 @@ from platformcode import logger
id_server = "vidtodo"
response = ""
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global response
response = httptools.downloadpage(page_url)
if not response.success or "Not Found" in response.data:
@@ -19,7 +19,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
data = response.data
packed_data = scrapertools.find_single_match(data, "javascript'>(eval.*?)</script>")
@@ -41,5 +41,5 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
video_urls.append([".mp4 [%s] %s" % (id_server, inf), mp4 % h])
video_urls.append(["RTMP [%s] %s" % (id_server, inf), "%s playpath=%s" % (rtmp, playpath)])
for video_url in video_urls:
logger.info("video_url: %s - %s" % (video_url[0], video_url[1]))
logger.debug("video_url: %s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -6,7 +6,7 @@ from lib import jsunpack
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global data
data = httptools.downloadpage(page_url).data
if "Not Found" in data or "File Does not Exist" in data:
@@ -15,7 +15,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
global data
video_urls = []
code = scrapertools.find_single_match(data, 'name="code" value="([^"]+)')

View File

@@ -20,7 +20,7 @@ from platformcode import logger
def test_video_exists(page_url):
return False, "[Vidup] Servidor Deshabilitado"
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
page = httptools.downloadpage(page_url)
url = page.url
if "Not Found" in page.data or "/404" in url:
@@ -29,7 +29,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
post= {}
post = urllib.urlencode(post)

View File

@@ -5,7 +5,7 @@ from core import scrapertools
from platformcode import logger, config
headers = [['User-Agent', 'Mozilla/5.0']]
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
global data
if "|" in page_url:
@@ -23,7 +23,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
global data
@@ -38,6 +38,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
video_urls.sort(key=lambda x: x[2])
for video_url in video_urls:
video_url[2] = 0
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -7,7 +7,7 @@ from platformcode import config, logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
if '"title":"Video Not Found"' in data:
return False, config.get_localized_string(70449) % "Vimple"
@@ -16,7 +16,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url=%s)" % page_url)
logger.debug("(page_url=%s)" % page_url)
data = httptools.downloadpage(page_url).data
@@ -34,6 +34,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
video_urls.append([scrapertools.get_filename_from_url(media_url)[-4:] + " [vimple.ru]", media_url])
for video_url in video_urls:
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -9,7 +9,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url)
if data.code == 404:
return False, config.get_localized_string(70449) % "Vivo"
@@ -17,7 +17,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
data = httptools.downloadpage(page_url).data
enc_data = scrapertools.find_single_match(data, 'data-stream="([^"]+)')

View File

@@ -17,7 +17,7 @@ from platformcode import config, logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
if not login():
return False, "Falta Ingresar/Actualizar las credenciales en el servidor vk. Configuracion - Preferencias - Ajustes de servidores - Configuración del servidor vk"
data = httptools.downloadpage(page_url).data
@@ -28,7 +28,7 @@ def test_video_exists(page_url):
# Returns an array of possible video url's from the page_url
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
data = httptools.downloadpage(page_url).data
matches = scrapertools.find_multiple_matches(data, '<source src="([^"]+)" type="video/(\w+)')
@@ -37,7 +37,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
video_urls.append([calidad + "p ." + ext + " [vk]", media_url])
video_urls.sort(key=lambda it: int(it[0].split("p ", 1)[0]))
for video_url in video_urls:
logger.info("%s - %s" % (video_url[0], video_url[1]))
logger.debug("%s - %s" % (video_url[0], video_url[1]))
return video_urls

View File

@@ -9,7 +9,7 @@ from platformcode import logger
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
response = httptools.downloadpage(page_url)
if response.code != 200 or "No longer available!" in response.data:
return False, config.get_localized_string(70449) % "vshare"
@@ -18,7 +18,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url = " + page_url)
logger.debug("url = " + page_url)
headers = {"Referer":page_url}
data = httptools.downloadpage(page_url, headers=headers).data
flowplayer = re.search("url: [\"']([^\"']+)", data)

View File

@@ -6,7 +6,7 @@ data = ""
def test_video_exists(page_url):
global data
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
response = httptools.downloadpage(page_url)
if response.code == 404:
@@ -18,5 +18,5 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
global data
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
return support.get_jwplayer_mediaurl(data, 'Vudeo')

View File

@@ -5,7 +5,7 @@ from platformcode import logger, config
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
page = httptools.downloadpage(page_url)
global data
data = page.data
@@ -15,7 +15,7 @@ def test_video_exists(page_url):
def get_video_url(page_url, premium=False, user="", password="", video_password=""):
logger.info("url=" + page_url)
logger.debug("url=" + page_url)
video_urls = []
global data
patron = r'sources:\s*\[\{src:\s*"([^"]+)"'

View File

@@ -19,7 +19,7 @@ payload = {'conn_id': conn_id}
def test_video_exists(page_url):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
data = httptools.downloadpage(page_url).data
if "Not Found" in data or "File was deleted" in data:
return False, config.get_localized_string(70449) % "VVVVID"
@@ -45,7 +45,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password=
# Getting info from Site
json_url = "https://www.vvvvid.it/vvvvid/ondemand/" + show_id + '/season/' +season_id + '/'
json_file = current_session.get(json_url, headers=headers, params=payload).json()
logger.info(json_file['data'])
logger.debug(json_file['data'])
# Search for the correct episode
for episode in json_file['data']:

View File

@@ -5,7 +5,7 @@ from platformcode import logger
def get_video_url(page_url, video_password):
logger.info("(page_url='%s')" % page_url)
logger.debug("(page_url='%s')" % page_url)
video_urls = []
url = page_url.replace("/v/", "/api/source/")
post = "r=&d=watchanimestream.net"

Some files were not shown because too many files have changed in this diff Show More