diff --git a/plugin.video.alfa/servers/auroravid.json b/plugin.video.alfa/servers/auroravid.json
deleted file mode 100755
index 4966911a..00000000
--- a/plugin.video.alfa/servers/auroravid.json
+++ /dev/null
@@ -1,41 +0,0 @@
-{
- "active": true,
- "find_videos": {
- "ignore_urls": [],
- "patterns": [
- {
- "pattern": "(?:embed.|)auroravid.to/(?:video/|embed/\\?v=)([A-z0-9]{13})",
- "url": "http://www.auroravid.to/embed/?v=\\1"
- }
- ]
- },
- "free": true,
- "id": "auroravid",
- "name": "auroravid",
- "settings": [
- {
- "default": false,
- "enabled": true,
- "id": "black_list",
- "label": "@60654",
- "type": "bool",
- "visible": true
- },
- {
- "default": 0,
- "enabled": true,
- "id": "favorites_servers_list",
- "label": "@60655",
- "lvalues": [
- "No",
- "1",
- "2",
- "3",
- "4",
- "5"
- ],
- "type": "list",
- "visible": false
- }
- ]
-}
\ No newline at end of file
diff --git a/plugin.video.alfa/servers/auroravid.py b/plugin.video.alfa/servers/auroravid.py
deleted file mode 100755
index db22d30f..00000000
--- a/plugin.video.alfa/servers/auroravid.py
+++ /dev/null
@@ -1,43 +0,0 @@
-# -*- coding: utf-8 -*-
-
-import re
-
-from core import httptools
-from core import scrapertools
-from platformcode import logger
-
-
-def test_video_exists(page_url):
- logger.info("(page_url='%s')" % page_url)
-
- data = httptools.downloadpage(page_url).data
-
- if "This file no longer exists on our servers" in data:
- return False, "[Auroravid] El fichero ha sido borrado"
-
- elif "is being converted" in data:
- return False, "[Auroravid] El fichero está en proceso todavía"
-
- return True, ""
-
-
-def get_video_url(page_url, premium=False, user="", password="", video_password=""):
- logger.info("(page_url='%s')" % page_url)
- data = httptools.downloadpage(page_url).data
-
- video_urls = []
- videourls = scrapertools.find_multiple_matches(data, 'src\s*:\s*[\'"]([^\'"]+)[\'"]')
- if not videourls:
- videourls = scrapertools.find_multiple_matches(data, '' in data:
- # return False,"The video has been cancelled from Backin.net"
-
- return True, ""
-
-
-def get_video_url(page_url, premium=False, user="", password="", video_password=""):
- logger.info("url=" + page_url)
- video_urls = []
- headers = []
- headers.append(["User-Agent",
- "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_8_5) AppleWebKit/537.17 (KHTML, like Gecko) Chrome/24.0.1312.52 Safari/537.17"])
-
- # First access
- data = scrapertools.cache_page(page_url, headers=headers)
- logger.info("data=" + data)
-
- # URL
- url = scrapertools.find_single_match(data, 'type="video/mp4" src="([^"]+)"')
- logger.info("url=" + url)
-
- # URL del vídeo
- video_urls.append([".mp4" + " [backin]", url])
-
- for video_url in video_urls:
- logger.info("%s - %s" % (video_url[0], video_url[1]))
-
- return video_urls
diff --git a/plugin.video.alfa/servers/bigfile.json b/plugin.video.alfa/servers/bigfile.json
deleted file mode 100755
index a06452f0..00000000
--- a/plugin.video.alfa/servers/bigfile.json
+++ /dev/null
@@ -1,44 +0,0 @@
-{
- "active": true,
- "find_videos": {
- "ignore_urls": [],
- "patterns": [
- {
- "pattern": "bigfile.to/((?:list|file)/[\\w]+)",
- "url": "https://www.bigfile.to/\\1"
- }
- ]
- },
- "free": false,
- "id": "bigfile",
- "name": "bigfile",
- "premium": [
- "realdebrid"
- ],
- "settings": [
- {
- "default": false,
- "enabled": true,
- "id": "black_list",
- "label": "@60654",
- "type": "bool",
- "visible": true
- },
- {
- "default": 0,
- "enabled": true,
- "id": "favorites_servers_list",
- "label": "@60655",
- "lvalues": [
- "No",
- "1",
- "2",
- "3",
- "4",
- "5"
- ],
- "type": "list",
- "visible": false
- }
- ]
-}
\ No newline at end of file
diff --git a/plugin.video.alfa/servers/bigfile.py b/plugin.video.alfa/servers/bigfile.py
deleted file mode 100755
index dbc147ea..00000000
--- a/plugin.video.alfa/servers/bigfile.py
+++ /dev/null
@@ -1,15 +0,0 @@
-# -*- coding: utf-8 -*-
-
-from platformcode import logger
-
-
-def test_video_exists(page_url):
- logger.info("(page_url='%s')" % page_url)
-
- return True, ""
-
-
-def get_video_url(page_url, premium=False, user="", password="", video_password=""):
- logger.info("(page_url='%s')" % page_url)
- video_urls = []
- return video_urls
diff --git a/plugin.video.alfa/servers/streaminto.json b/plugin.video.alfa/servers/streaminto.json
deleted file mode 100755
index 77c0ad5f..00000000
--- a/plugin.video.alfa/servers/streaminto.json
+++ /dev/null
@@ -1,58 +0,0 @@
-{
- "active": true,
- "find_videos": {
- "ignore_urls": [
- "http://streamin.to/embed-theme.html",
- "http://streamin.to/embed-jquery.html",
- "http://streamin.to/embed-s.html",
- "http://streamin.to/embed-images.html",
- "http://streamin.to/embed-faq.html",
- "http://streamin.to/embed-embed.html",
- "http://streamin.to/embed-ri.html",
- "http://streamin.to/embed-d.html",
- "http://streamin.to/embed-css.html",
- "http://streamin.to/embed-js.html",
- "http://streamin.to/embed-player.html",
- "http://streamin.to/embed-cgi.html"
- ],
- "patterns": [
- {
- "pattern": "streamin.to/(?:embed-)?([a-z0-9A-Z]+)",
- "url": "http://streamin.to/embed-\\1.html"
- }
- ]
- },
- "free": true,
- "id": [
- "streaminto",
- "streamin"
- ],
- "name": "streamin",
- "settings": [
- {
- "default": false,
- "enabled": true,
- "id": "black_list",
- "label": "@60654",
- "type": "bool",
- "visible": true
- },
- {
- "default": 0,
- "enabled": true,
- "id": "favorites_servers_list",
- "label": "@60655",
- "lvalues": [
- "No",
- "1",
- "2",
- "3",
- "4",
- "5"
- ],
- "type": "list",
- "visible": false
- }
- ],
- "thumbnail": "server_streaminto.png"
-}
\ No newline at end of file
diff --git a/plugin.video.alfa/servers/streaminto.py b/plugin.video.alfa/servers/streaminto.py
deleted file mode 100755
index 1744b430..00000000
--- a/plugin.video.alfa/servers/streaminto.py
+++ /dev/null
@@ -1,77 +0,0 @@
-# -*- coding: utf-8 -*-
-
-import re
-
-from core import httptools
-from core import scrapertools
-from platformcode import logger
-
-
-def test_video_exists(page_url):
- logger.info("(page_url='%s')" % page_url)
-
- data = httptools.downloadpage(page_url).data
- if "File was deleted" in data:
- return False, "El archivo no existe
en streaminto o ha sido borrado."
- elif "Video is processing now" in data:
- return False, "El archivo está siendo procesado
Prueba dentro de un rato."
- else:
- return True, ""
-
-
-def get_video_url(page_url, premium=False, user="", password="", video_password=""):
- logger.info("url=" + page_url)
-
- data = re.sub(r'\n|\t|\s+', '', httptools.downloadpage(page_url).data)
-
- video_urls = []
- try:
- media_url = scrapertools.get_match(data, """.setup\({file:"([^"]+)",image""")
- except:
- js_data = scrapertools.find_single_match(data, "(eval.function.p,a,c,k,e.*?)")
- js_data = unPack(js_data)
- media_url = scrapertools.get_match(js_data, """.setup\({file:"([^"]+)",image""")
-
- if media_url.endswith("v.mp4"):
- media_url_mp42flv = re.sub(r'/v.mp4$', '/v.flv', media_url)
- video_urls.append(
- [scrapertools.get_filename_from_url(media_url_mp42flv)[-4:] + " [streaminto]", media_url_mp42flv])
- if media_url.endswith("v.flv"):
- media_url_flv2mp4 = re.sub(r'/v.flv$', '/v.mp4', media_url)
- video_urls.append(
- [scrapertools.get_filename_from_url(media_url_flv2mp4)[-4:] + " [streaminto]", media_url_flv2mp4])
- video_urls.append([scrapertools.get_filename_from_url(media_url)[-4:] + " [streaminto]", media_url])
-
- for video_url in video_urls:
- logger.info("%s - %s" % (video_url[0], video_url[1]))
-
- return video_urls
-
-
-def unPack(packed):
- pattern = "}\('(.*)', *(\d+), *(\d+), *'(.*)'\.split\('([^']+)'\)"
- d = [d for d in re.search(pattern, packed, re.DOTALL).groups()]
-
- p = d[0];
- a = int(d[1]);
- c = int(d[2]);
- k = d[3].split(d[4])
-
- if a <= 62:
- toString = "0123456789abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ"
- else:
- toString = """ !"#$%&'()*+,-./0123456789:;<=>?@ABCDEFGHIJKLMNOPQRSTUVWXYZ[]^_`abcdefghijklmnopqrstuvwxyz{|}~"""
-
- def e(c):
- return toString[c] if c < a else toString[c // a] + toString[c % a]
-
- while c > 0:
- c -= 1
- if k[c]:
- x = e(c)
- else:
- x = k[c]
- y = k[c]
- p = re.sub(r"(\b%s\b)" % x, y, p)
-
- return p
diff --git a/plugin.video.alfa/servers/tunepk.json b/plugin.video.alfa/servers/tunepk.json
deleted file mode 100755
index 3d3452ce..00000000
--- a/plugin.video.alfa/servers/tunepk.json
+++ /dev/null
@@ -1,41 +0,0 @@
-{
- "active": true,
- "find_videos": {
- "ignore_urls": [],
- "patterns": [
- {
- "pattern": "tune.pk/player/embed_player.php\\?vid\\=(\\d+)",
- "url": "http://embed.tune.pk/play/\\1?autoplay=no"
- }
- ]
- },
- "free": true,
- "id": "tunepk",
- "name": "tunepk",
- "settings": [
- {
- "default": false,
- "enabled": true,
- "id": "black_list",
- "label": "@60654",
- "type": "bool",
- "visible": true
- },
- {
- "default": 0,
- "enabled": true,
- "id": "favorites_servers_list",
- "label": "@60655",
- "lvalues": [
- "No",
- "1",
- "2",
- "3",
- "4",
- "5"
- ],
- "type": "list",
- "visible": false
- }
- ]
-}
\ No newline at end of file
diff --git a/plugin.video.alfa/servers/tunepk.py b/plugin.video.alfa/servers/tunepk.py
deleted file mode 100755
index cde33744..00000000
--- a/plugin.video.alfa/servers/tunepk.py
+++ /dev/null
@@ -1,32 +0,0 @@
-# -*- coding: utf-8 -*-
-
-import re
-
-from core import scrapertools
-from platformcode import logger
-
-
-# Returns an array of possible video url's from the page_url
-def get_video_url(page_url, premium=False, user="", password="", video_password=""):
- logger.info("(page_url='%s')" % page_url)
-
- video_urls = []
-
- data = scrapertools.cache_page(page_url)
- logger.info(data)
- patron = 'file: "([^"]+)",\s+'
- patron += 'width: "[^"]+",\s+'
- patron += 'height: "[^"]+",\s+'
- patron += 'label : "([^"]+)",\s+'
- patron += 'type : "([^"]+)"'
- matches = re.compile(patron, re.DOTALL).findall(data)
- scrapertools.printMatches(matches)
-
- for url, calidad, formato in matches:
- video_url = ["%s %s [tune.pk]" % (calidad, formato), url]
- video_urls.append(video_url)
-
- for video_url in video_urls:
- logger.info("%s - %s" % (video_url[0], video_url[1]))
-
- return video_urls
diff --git a/plugin.video.alfa/servers/tutv.json b/plugin.video.alfa/servers/tutv.json
deleted file mode 100755
index e7c11eff..00000000
--- a/plugin.video.alfa/servers/tutv.json
+++ /dev/null
@@ -1,45 +0,0 @@
-{
- "active": true,
- "find_videos": {
- "ignore_urls": [],
- "patterns": [
- {
- "pattern": "(http://(?:www.)?tu.tv[^\"]+)",
- "url": "\\1"
- },
- {
- "pattern": "tu.tv/(iframe/\\d+)",
- "url": "http://tu.tv/\\1"
- }
- ]
- },
- "free": true,
- "id": "tutv",
- "name": "tutv",
- "settings": [
- {
- "default": false,
- "enabled": true,
- "id": "black_list",
- "label": "@60654",
- "type": "bool",
- "visible": true
- },
- {
- "default": 0,
- "enabled": true,
- "id": "favorites_servers_list",
- "label": "@60655",
- "lvalues": [
- "No",
- "1",
- "2",
- "3",
- "4",
- "5"
- ],
- "type": "list",
- "visible": false
- }
- ]
-}
\ No newline at end of file
diff --git a/plugin.video.alfa/servers/tutv.py b/plugin.video.alfa/servers/tutv.py
deleted file mode 100755
index 475652ba..00000000
--- a/plugin.video.alfa/servers/tutv.py
+++ /dev/null
@@ -1,56 +0,0 @@
-# -*- coding: utf-8 -*-
-
-import re
-import urllib
-
-from core import scrapertools
-from platformcode import logger
-
-
-# Returns an array of possible video url's from the page_url
-def get_video_url(page_url, premium=False, user="", password="", video_password=""):
- logger.info("(page_url='%s')" % page_url)
-
- # Busca el ID en la URL
- id = extract_id(page_url)
-
- # Si no lo tiene, lo extrae de la página
- if id == "":
- # La descarga
- data = scrapertools.cache_page(page_url)
- patron = ''
- matches = re.compile(patron, re.DOTALL).findall(data)
- if len(matches) > 0:
- id = extract_id(matches[0])
- else:
- id = ""
-
- if id == "":
- id = scrapertools.get_match(page_url, "tu.tv/iframe/(\d+)")
-
- # Descarga el descriptor
- url = "http://tu.tv/visualizacionExterna2.php?web=undefined&codVideo=" + id
- data = scrapertools.cache_page(url)
-
- # Obtiene el enlace al vídeo
- patronvideos = 'urlVideo0=([^\&]+)\&'
- matches = re.compile(patronvideos, re.DOTALL).findall(data)
- # scrapertools.printMatches(matches)
- url = urllib.unquote_plus(matches[0])
- video_urls = [["[tu.tv]", url]]
-
- for video_url in video_urls:
- logger.info("%s - %s" % (video_url[0], video_url[1]))
-
- return video_urls
-
-
-def extract_id(text):
- patron = "xtp\=([a-zA-Z0-9]+)"
- matches = re.compile(patron, re.DOTALL).findall(text)
- if len(matches) > 0:
- devuelve = matches[0]
- else:
- devuelve = ""
-
- return devuelve