From cdb735e3a3a81b8cba5b30f79887e339fe136bd3 Mon Sep 17 00:00:00 2001
From: Intel1 <25161862+Intel11@users.noreply.github.com>
Date: Mon, 5 Feb 2018 09:29:36 -0500
Subject: [PATCH] maxipelis: dominio no existe
---
plugin.video.alfa/channels/maxipelis.py | 152 ------------------------
1 file changed, 152 deletions(-)
delete mode 100644 plugin.video.alfa/channels/maxipelis.py
diff --git a/plugin.video.alfa/channels/maxipelis.py b/plugin.video.alfa/channels/maxipelis.py
deleted file mode 100644
index 8f6b07de..00000000
--- a/plugin.video.alfa/channels/maxipelis.py
+++ /dev/null
@@ -1,152 +0,0 @@
-# -*- coding: utf-8 -*-
-# ------------------------------------------------------------
-# Alfa
-# ------------------------------------------------------------
-import urlparse,urllib2,urllib,re
-import os, sys
-
-from core import jsontools as json
-from core import scrapertools
-from core import servertools
-from core.item import Item
-from platformcode import config, logger
-from core import httptools
-from core import tmdb
-
-host = 'http://www.maxipelis.net'
-
-
-def mainlist(item):
- logger.info()
- itemlist = []
- itemlist.append(Item(channel=item.channel, title="Novedades" , action="peliculas", url=host + "/pelicula"))
-
- itemlist.append(Item(channel=item.channel, title="Categorias" , action="categorias", url=host))
- itemlist.append(Item(channel=item.channel, title="Buscar", action="search"))
- return itemlist
-
-
-def search(item, texto):
- logger.info()
- texto = texto.replace(" ", "+")
- item.url = host + "/?s=%s" % texto
-
- try:
- return sub_search(item)
-
- except:
- import sys
- for line in sys.exc_info():
- logger.error("%s" % line)
- return []
-
-
-def sub_search(item):
- logger.info()
- itemlist = []
- data = httptools.downloadpage(item.url).data
- data = re.sub(r"\n|\r|\t| |
", "", data)
-
- patron = '
(.*?)
' - matches = scrapertools.find_multiple_matches(data, patron) - - for url,img,name,plot in matches: - itemlist.append(item.clone(channel=item.channel, action="findvideos", title=name, url=url, plot=plot, - thumbnail=img)) - - paginacion = scrapertools.find_single_match(data, '