From 0b605cf2154dc45c93106715bd5b3862b24f82f5 Mon Sep 17 00:00:00 2001 From: marco Date: Sat, 19 Sep 2020 16:32:37 +0200 Subject: [PATCH] riscritto logger --- channels/0example.py.txt | 22 ++-- channels/altadefinizione01.py | 12 +-- channels/altadefinizione01_link.py | 12 +-- channels/altadefinizioneclick.py | 8 +- channels/animealtadefinizione.py | 8 +- channels/animeforce.py | 12 +-- channels/animeleggendari.py | 8 +- channels/animesaturn.py | 10 +- channels/animeunity.py | 22 ++-- channels/animeuniverse.py | 8 +- channels/animeworld.py | 10 +- channels/casacinema.py | 12 +-- channels/cb01anime.py | 8 +- channels/cineblog01.py | 20 ++-- channels/cinemalibero.py | 16 +-- channels/cinetecadibologna.py | 6 +- channels/dreamsub.py | 16 +-- channels/dsda.py | 8 +- channels/eurostreaming.py | 12 +-- channels/fastsubita.py | 18 ++-- channels/filmigratis.py | 16 +-- channels/filmpertutti.py | 16 +-- channels/guardaseriecam.py | 6 +- channels/guardaserieclick.py | 20 ++-- channels/guardaserieicu.py | 8 +- channels/hd4me.py | 4 +- channels/ilcorsaronero.py | 6 +- channels/ilgeniodellostreaming.py | 20 ++-- channels/ilgeniodellostreaming_cam.py | 12 +-- channels/italiaserie.py | 18 ++-- channels/la7.py | 2 +- channels/mediasetplay.py | 24 ++--- channels/metalvideo.py | 4 +- channels/mondoserietv.py | 8 +- channels/paramount.py | 8 +- channels/piratestreaming.py | 10 +- channels/polpotv.py | 20 ++-- channels/pufimovies.py | 10 +- channels/raiplay.py | 38 +++---- channels/seriehd.py | 10 +- channels/serietvonline.py | 18 ++-- channels/serietvsubita.py | 22 ++-- channels/serietvu.py | 16 +-- channels/streamingaltadefinizione.py | 4 +- channels/streamingcommunity.py | 20 ++-- channels/tantifilm.py | 22 ++-- channels/toonitalia.py | 10 +- channels/tunein.py | 6 +- channels/vvvvid.py | 10 +- channelselector.py | 20 ++-- core/channeltools.py | 12 +-- core/downloader.py | 28 +++--- core/downloadtools.py | 138 +++++++++++++------------- core/filetools.py | 4 +- core/httptools.py | 24 ++--- core/item.py | 2 +- core/jsontools.py | 18 ++-- core/scraper.py | 12 +-- core/scrapertools.py | 4 +- core/servertools.py | 62 ++++++------ core/support.py | 52 +++++----- core/tmdb.py | 18 ++-- core/trakt_tools.py | 14 +-- core/tvdb.py | 32 +++--- core/videolibrarytools.py | 56 +++++------ core/ziptools.py | 16 +-- default.py | 2 +- lib/arm_chromeos.py | 8 +- lib/generictools.py | 30 +++--- lib/megaserver/client.py | 6 +- lib/sambatools/libsmb.py | 28 +++--- lib/unshortenit.py | 10 +- platformcode/config.py | 10 +- platformcode/download_and_play.py | 92 ++++++++--------- platformcode/envtal.py | 38 +++---- platformcode/launcher.py | 33 +++--- platformcode/logger.py | 122 +++++------------------ platformcode/platformtools.py | 65 ++++++------ platformcode/recaptcha.py | 2 +- platformcode/subtitletools.py | 20 ++-- platformcode/unify.py | 18 ++-- platformcode/updater.py | 44 ++++---- platformcode/xbmc_config_menu.py | 4 +- platformcode/xbmc_info_window.py | 4 +- platformcode/xbmc_videolibrary.py | 110 ++++++++++---------- servers/akvideo.py | 8 +- servers/anavids.py | 2 +- servers/animeid.py | 4 +- servers/anonfile.py | 4 +- servers/archiveorg.py | 4 +- servers/backin.py | 10 +- servers/badshare.py | 4 +- servers/bdupload.py | 4 +- servers/cinemaupload.py | 4 +- servers/clicknupload.py | 6 +- servers/clipwatching.py | 4 +- servers/cloudvideo.py | 4 +- servers/crunchyroll.py | 6 +- servers/dailymotion.py | 6 +- servers/debriders/alldebrid.py | 2 +- servers/debriders/realdebrid.py | 4 +- servers/decrypters/adfly.py | 2 +- servers/decrypters/linkbucks.py | 8 +- servers/decrypters/longurl.py | 6 +- servers/decrypters/safelinking.py | 4 +- servers/deltabit.py | 4 +- servers/directo.py | 2 +- servers/dostream.py | 4 +- servers/downace.py | 4 +- servers/facebook.py | 6 +- servers/fastplay.py | 6 +- servers/fembed.py | 6 +- servers/fex.py | 4 +- servers/filefactory.py | 4 +- servers/filepup.py | 6 +- servers/filescdn.py | 4 +- servers/filesmonster.py | 2 +- servers/flashx.py | 12 +-- servers/fourshared.py | 6 +- servers/gamovideo.py | 6 +- servers/gigasize.py | 4 +- servers/googlevideo.py | 6 +- servers/gounlimited.py | 6 +- servers/gvideo.py | 2 +- servers/hdload.py | 8 +- servers/hdmario.py | 20 ++-- servers/hugefiles.py | 4 +- servers/idtbox.py | 4 +- servers/jawcloud.py | 4 +- servers/jetload.py | 4 +- servers/mailru.py | 6 +- servers/mediafire.py | 6 +- servers/mega.py | 2 +- servers/mixdrop.py | 4 +- servers/mp4upload.py | 6 +- servers/mydaddy.py | 2 +- servers/mystream.py | 4 +- servers/myupload.py | 2 +- servers/netutv.py | 4 +- servers/nowvideo.py | 16 +-- servers/okru.py | 4 +- servers/onefichier.py | 18 ++-- servers/onlystream.py | 6 +- servers/rapidgator.py | 2 +- servers/rcdnme.py | 6 +- servers/rutube.py | 6 +- servers/samaup.py | 4 +- servers/sendvid.py | 4 +- servers/speedvideo.py | 10 +- servers/streamtape.py | 4 +- servers/streamz.py | 2 +- servers/supervideo.py | 4 +- servers/thevid.py | 4 +- servers/thevideobee.py | 4 +- servers/torrent.py | 4 +- servers/turbobit.py | 2 +- servers/turbovid.py | 6 +- servers/tusfiles.py | 4 +- servers/uploadedto.py | 32 +++--- servers/uppom.py | 4 +- servers/upstream.py | 2 +- servers/uptobox.py | 6 +- servers/upvid.py | 4 +- servers/uqload.py | 4 +- servers/userscloud.py | 6 +- servers/vevio.py | 4 +- servers/vidcloud.py | 6 +- servers/videobin.py | 4 +- servers/videomega.py | 2 +- servers/vidfast.py | 6 +- servers/vidlox.py | 4 +- servers/vidmoly.py | 4 +- servers/vidoza.py | 4 +- servers/vidtodo.py | 6 +- servers/vidtome.py | 4 +- servers/vidup.py | 4 +- servers/vimeo.py | 6 +- servers/vimpleru.py | 6 +- servers/vivo.py | 4 +- servers/vk.py | 6 +- servers/vshare.py | 4 +- servers/vudeo.py | 4 +- servers/vupplayer.py | 4 +- servers/vvvvid.py | 4 +- servers/watchanimestream.py | 2 +- servers/watchvideo.py | 6 +- servers/wstream.py | 8 +- servers/youdbox.py | 2 +- servers/yourupload.py | 4 +- servers/youtube.py | 6 +- servers/youwatch.py | 6 +- servers/zippyshare.py | 4 +- service.py | 28 +++--- specials/autoplay.py | 4 +- specials/autorenumber.py | 26 ++--- specials/backup.py | 4 +- specials/checkhost.py | 10 +- specials/community.py | 44 ++++---- specials/downloads.py | 68 ++++++------- specials/elementum_download.py | 18 ++-- specials/favorites.py | 12 +-- specials/filmontv.py | 8 +- specials/filtertools.py | 26 ++--- specials/help.py | 2 +- specials/infoplus.py | 4 +- specials/kodfavorites.py | 50 +++++----- specials/news.py | 36 +++---- specials/resolverdns.py | 10 +- specials/search.py | 20 ++-- specials/setting.py | 46 ++++----- specials/shortcuts.py | 2 +- specials/side_menu.py | 8 +- specials/trailertools.py | 20 ++-- specials/tvmoviedb.py | 46 ++++----- specials/url.py | 6 +- specials/videolibrary.py | 72 +++++++------- tests/test_generic.py | 2 +- updatetvshow.py | 8 +- 218 files changed, 1425 insertions(+), 1511 deletions(-) diff --git a/channels/0example.py.txt b/channels/0example.py.txt index 72e4b75b..7ee29078 100644 --- a/channels/0example.py.txt +++ b/channels/0example.py.txt @@ -76,7 +76,7 @@ headers = [['Referer', host]] @support.menu def mainlist(item): - support.log(item) + support.info(item) # Ordine delle voci # Voce FILM, puoi solo impostare l'url @@ -147,7 +147,7 @@ def mainlist(item): # AVVERTENZE: Se il titolo è trovato nella ricerca TMDB/TVDB/Altro allora le locandine e altre info non saranno quelle recuperate nel sito.!!!! @support.scrape def peliculas(item): - support.log(item) + support.info(item) #support.dbg() # decommentare per attivare web_pdb action = '' @@ -162,7 +162,7 @@ def peliculas(item): @support.scrape def episodios(item): - support.log(item) + support.info(item) #support.dbg() action = '' @@ -179,7 +179,7 @@ def episodios(item): # per genere, per anno, per lettera, per qualità ecc ecc @support.scrape def genres(item): - support.log(item) + support.info(item) #support.dbg() action = '' @@ -199,7 +199,7 @@ def genres(item): # e la ricerca porta i titoli mischiati senza poterli distinguere tra loro # andranno modificate anche le def peliculas e episodios ove occorre def select(item): - support.log('select --->', item) + support.info('select --->', item) #support.dbg() data = httptools.downloadpage(item.url, headers=headers).data # pulizia di data, in caso commentare le prossime 2 righe @@ -207,7 +207,7 @@ def select(item): data = re.sub(r'>\s+<', '> <', data) block = scrapertools.find_single_match(data, r'') if re.findall('', data, re.IGNORECASE): - support.log('select = ### è una serie ###') + support.info('select = ### è una serie ###') return episodios(Item(channel=item.channel, title=item.title, fulltitle=item.fulltitle, @@ -220,7 +220,7 @@ def select(item): ############## Fondo Pagina # da adattare al canale def search(item, text): - support.log('search', item) + support.info('search', item) itemlist = [] text = text.replace(' ', '+') item.url = host + '/index.php?do=search&story=%s&subaction=search' % (text) @@ -233,7 +233,7 @@ def search(item, text): except: import sys for line in sys.exc_info(): - log('search log:', line) + info('search log:', line) return [] @@ -241,7 +241,7 @@ def search(item, text): # inserire newest solo se il sito ha la pagina con le ultime novità/aggiunte # altrimenti NON inserirlo def newest(categoria): - support.log('newest ->', categoria) + support.info('newest ->', categoria) itemlist = [] item = Item() try: @@ -256,7 +256,7 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.log('newest log: ', {0}.format(line)) + support.info('newest log: ', {0}.format(line)) return [] return itemlist @@ -266,5 +266,5 @@ def newest(categoria): # sia per i siti con hdpass #support.server(item, data='', itemlist=[], headers='', AutoPlay=True, CheckLinks=True) def findvideos(item): - support.log('findvideos ->', item) + support.info('findvideos ->', item) return support.server(item, headers=headers) diff --git a/channels/altadefinizione01.py b/channels/altadefinizione01.py index 5be5ac9d..19d2396d 100644 --- a/channels/altadefinizione01.py +++ b/channels/altadefinizione01.py @@ -47,7 +47,7 @@ def mainlist(item): @support.scrape def peliculas(item): - support.log('peliculas',item) + support.info('peliculas', item) ## deflang = 'ITA' action="findvideos" @@ -69,7 +69,7 @@ def peliculas(item): @support.scrape def genres(item): - support.log('genres',item) + support.info('genres',item) if item.args != 'orderalf': action = "peliculas" else: action = 'orderalf' @@ -90,7 +90,7 @@ def genres(item): @support.scrape def orderalf(item): - support.log('orderalf',item) + support.info('orderalf',item) action= 'findvideos' patron = r'(?P[^<]+)[^>]+>[^>]+>[^>]+>[^>]+>[^>]+>\s*<div class="[^"]+" style="background-image:url\((?P<thumb>[^\)]+)[^>]+>[^>]+>[^>]+>[^>]+>[^>]+>[^>]+>\s*(?P<year>\d{4})[^>]+>[^>]+>(?:\s*(?P<duration>\d+)[^>]+>[^>]+>)?\s*(?P<quality>[a-zA-Z]+) [^>]+>[^>]+> (?P<lang>[^>]+) [^>]+>' patronNext = r'<span>\d</span> <a href="([^"]+)">' @@ -43,7 +43,7 @@ def peliculas(item): # =========== def pagina categorie ====================================== @support.scrape def genres(item): - support.log('genres',item) + support.info('genres',item) action = 'peliculas' if item.args == 'genres': @@ -63,7 +63,7 @@ def genres(item): # =========== def per cercare film/serietv ============= #host+/index.php?do=search&story=avatar&subaction=search def search(item, text): - support.log('search', item) + support.info('search', item) itemlist = [] text = text.replace(" ", "+") item.url = host+"/index.php?do=search&story=%s&subaction=search" % (text) @@ -79,7 +79,7 @@ def search(item, text): # =========== def per le novità nel menu principale ============= def newest(categoria): - support.log('newest', categoria) + support.info('newest', categoria) itemlist = [] item = Item() try: @@ -101,5 +101,5 @@ def newest(categoria): return itemlist def findvideos(item): - support.log('findvideos', item) + support.info('findvideos', item) return support.server(item, support.match(item, patron='<ul class="playernav">.*?</ul>', headers=headers).match) diff --git a/channels/altadefinizioneclick.py b/channels/altadefinizioneclick.py index a1d91ebc..90101a51 100644 --- a/channels/altadefinizioneclick.py +++ b/channels/altadefinizioneclick.py @@ -96,7 +96,7 @@ def genres(item): return locals() def search(item, texto): - support.log("search ", texto) + support.info("search ", texto) item.args = 'search' item.url = host + "/?s=" + texto @@ -106,11 +106,11 @@ def search(item, texto): except: import sys for line in sys.exc_info(): - support.logger.error("%s" % line) + support.infoger.error("%s" % line) return [] def newest(categoria): - support.log(categoria) + support.info(categoria) itemlist = [] item = Item() try: @@ -135,7 +135,7 @@ def newest(categoria): return itemlist def findvideos(item): - support.log('findvideos', item) + support.info('findvideos', item) return support.hdpass_get_servers(item) def play(item): diff --git a/channels/animealtadefinizione.py b/channels/animealtadefinizione.py index ee9a5082..ab22ed15 100644 --- a/channels/animealtadefinizione.py +++ b/channels/animealtadefinizione.py @@ -32,7 +32,7 @@ def menu(item): def search(item, texto): - support.log(texto) + support.info(texto) item.search = texto try: return peliculas(item) @@ -40,12 +40,12 @@ def search(item, texto): except: import sys for line in sys.exc_info(): - support.logger.error("%s" % line) + support.infoger.error("%s" % line) return [] def newest(categoria): - support.log(categoria) + support.info(categoria) item = support.Item() try: if categoria == "anime": @@ -56,7 +56,7 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.logger.error("{0}".format(line)) + support.infoger.error("{0}".format(line)) return [] diff --git a/channels/animeforce.py b/channels/animeforce.py index 2faf4252..6477d06a 100644 --- a/channels/animeforce.py +++ b/channels/animeforce.py @@ -36,7 +36,7 @@ def submenu(item): def newest(categoria): - support.log(categoria) + support.info(categoria) itemlist = [] item = support.Item() try: @@ -49,13 +49,13 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.logger.error("{0}".format(line)) + support.infoger.error("{0}".format(line)) return [] return itemlist def search(item, texto): - support.log(texto) + support.info(texto) item.args = 'noorder' item.url = host + '/?s=' + texto + '&cat=6010' item.contentType = 'tvshow' @@ -65,7 +65,7 @@ def search(item, texto): except: import sys for line in sys.exc_info(): - support.logger.error("%s" % line) + support.infoger.error("%s" % line) return [] @@ -114,7 +114,7 @@ def episodios(item): else: patron = r'<a\s*href="(?P<url>[^"]+)"\s*title="(?P<title>[^"]+)"\s*class="btn btn-dark mb-1">' def itemHook(item): - support.log(item) + support.info(item) if item.url.startswith('//'): item.url= 'https:' + item.url elif item.url.startswith('/'): item.url= 'https:/' + item.url return item @@ -123,7 +123,7 @@ def episodios(item): def findvideos(item): - support.log(item) + support.info(item) itemlist = [] if 'adf.ly' in item.url: diff --git a/channels/animeleggendari.py b/channels/animeleggendari.py index 46d7b5c7..da1c6682 100644 --- a/channels/animeleggendari.py +++ b/channels/animeleggendari.py @@ -31,7 +31,7 @@ def mainlist(item): def search(item, texto): - support.log(texto) + support.info(texto) item.url = host + "/?s=" + texto try: @@ -41,7 +41,7 @@ def search(item, texto): except: import sys for line in sys.exc_info(): - support.logger.error("%s" % line) + support.infoger.error("%s" % line) return [] @@ -83,7 +83,7 @@ def peliculas(item): def episodios(item): data = support.match(item, headers=headers, patronBlock=r'entry-content clearfix">(.*?)class="mh-widget mh-posts-2 widget_text').block if not 'pagination clearfix' in data: - support.log('NOT IN DATA') + support.info('NOT IN DATA') patron = r'<iframe.*?src="(?P<url>[^"]+)"' title = item.title def fullItemlistHook(itemlist): @@ -124,7 +124,7 @@ def check(item): return data def findvideos(item): - support.log() + support.info() if item.data: data = item.data else: diff --git a/channels/animesaturn.py b/channels/animesaturn.py index a3870e47..0a579549 100644 --- a/channels/animesaturn.py +++ b/channels/animesaturn.py @@ -62,7 +62,7 @@ def mainlist(item): def search(item, texto): - support.log(texto) + support.info(texto) item.url = host + '/animelist?search=' + texto item.contentType = 'tvshow' try: @@ -71,12 +71,12 @@ def search(item, texto): except: import sys for line in sys.exc_info(): - support.logger.error("%s" % line) + support.infoger.error("%s" % line) return [] def newest(categoria): - support.log() + support.info() itemlist = [] item = support.Item() try: @@ -88,7 +88,7 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.logger.error("{0}".format(line)) + support.infoger.error("{0}".format(line)) return [] return itemlist @@ -163,7 +163,7 @@ def episodios(item): def findvideos(item): - support.log() + support.info() itemlist = [] page_data = '' titles =['Primario', 'Secondario', 'Alternativo Primario', 'Alternativo Secondario'] diff --git a/channels/animeunity.py b/channels/animeunity.py index 63582afa..04f4d11e 100644 --- a/channels/animeunity.py +++ b/channels/animeunity.py @@ -51,7 +51,7 @@ def menu(item): def genres(item): - support.log() + support.info() # support.dbg() itemlist = [] @@ -63,7 +63,7 @@ def genres(item): return support.thumb(itemlist) def years(item): - support.log() + support.info() itemlist = [] from datetime import datetime @@ -77,7 +77,7 @@ def years(item): def search(item, text): - support.log('search', item) + support.info('search', item) if not item.args: item.args = {'title':text} else: @@ -90,12 +90,12 @@ def search(item, text): except: import sys for line in sys.exc_info(): - support.log('search log:', line) + support.info('search log:', line) return [] def newest(categoria): - support.log(categoria) + support.info(categoria) itemlist = [] item = support.Item() item.url = host @@ -109,13 +109,13 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.log(line) + support.info(line) return [] return itemlist def news(item): - support.log() + support.info() item.contentType = 'tvshow' itemlist = [] @@ -140,7 +140,7 @@ def news(item): def peliculas(item): - support.log() + support.info() itemlist = [] page = item.page if item.page else 0 @@ -154,7 +154,7 @@ def peliculas(item): payload = json.dumps(item.args) records = requests.post(host + '/archivio/get-animes', headers=headers, data=payload).json()['records'] # js = [] - # support.log(records) + # support.info(records) # for record in records: # js += record for it in records: @@ -194,7 +194,7 @@ def peliculas(item): return itemlist def episodios(item): - support.log() + support.info() itemlist = [] title = 'Parte ' if item.type.lower() == 'movie' else 'Episodio ' for it in item.episodes: @@ -219,5 +219,5 @@ def episodios(item): def findvideos(item): - support.log() + support.info() return support.server(item,itemlist=[item.clone(title=support.config.get_localized_string(30137), server='directo', action='play')]) \ No newline at end of file diff --git a/channels/animeuniverse.py b/channels/animeuniverse.py index b5639af1..c3b3beb8 100644 --- a/channels/animeuniverse.py +++ b/channels/animeuniverse.py @@ -33,7 +33,7 @@ def menu(item): def search(item, texto): - support.log(texto) + support.info(texto) item.search = texto try: return peliculas(item) @@ -41,12 +41,12 @@ def search(item, texto): except: import sys for line in sys.exc_info(): - support.logger.error("%s" % line) + support.infoger.error("%s" % line) return [] def newest(categoria): - support.log(categoria) + support.info(categoria) item = support.Item() try: if categoria == "anime": @@ -57,7 +57,7 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.logger.error("{0}".format(line)) + support.infoger.error("{0}".format(line)) return [] diff --git a/channels/animeworld.py b/channels/animeworld.py index 4c7045df..ba961f77 100644 --- a/channels/animeworld.py +++ b/channels/animeworld.py @@ -107,7 +107,7 @@ def submenu(item): def newest(categoria): - support.log(categoria) + support.info(categoria) item = support.Item() try: if categoria == "anime": @@ -118,12 +118,12 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.logger.error("{0}".format(line)) + support.infoger.error("{0}".format(line)) return [] def search(item, texto): - support.log(texto) + support.info(texto) if item.search: item.url = host + '/filter?dub=' + item.args + '&keyword=' + texto + '&sort=' else: @@ -136,7 +136,7 @@ def search(item, texto): except: import sys for line in sys.exc_info(): - support.logger.error("%s" % line) + support.infoger.error("%s" % line) return [] @@ -187,7 +187,7 @@ def episodios(item): def findvideos(item): import time - support.log(item) + support.info(item) itemlist = [] urls = [] resp = support.match(get_data(item), headers=headers, patron=r'data-name="(\d+)">([^<]+)<') diff --git a/channels/casacinema.py b/channels/casacinema.py index 5eb7ab31..bf27fed8 100644 --- a/channels/casacinema.py +++ b/channels/casacinema.py @@ -39,17 +39,17 @@ def genres(item): def select(item): item.data = support.match(item).data if 'continua con il video' in item.data.lower(): - support.log('select = ### è un film ###') + support.info('select = ### è un film ###') item.contentType = 'movie' return findvideos(item) else: - support.log('select = ### è una serie ###') + support.info('select = ### è una serie ###') item.contentType = 'tvshow' return episodios(item) def search(item, text): - support.log(text) + support.info(text) text = text.replace(' ', '+') item.url = host + '/?s=' + text item.args = 'search' @@ -59,7 +59,7 @@ def search(item, text): except: import sys for line in sys.exc_info(): - support.log('search log:', line) + support.info('search log:', line) return [] @@ -87,7 +87,7 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.log("%s" % line) + support.info("%s" % line) return [] return itemlist @@ -148,7 +148,7 @@ def findvideos(item): data = '' from lib.unshortenit import unshorten_only for link in links: - support.log('URL=',link) + support.info('URL=',link) url, c = unshorten_only(link.replace('#', 'speedvideo.net')) data += url + '\n' return support.server(item, data) diff --git a/channels/cb01anime.py b/channels/cb01anime.py index b47edc59..81ca8fde 100644 --- a/channels/cb01anime.py +++ b/channels/cb01anime.py @@ -35,12 +35,12 @@ def menu(item): def search(item, texto): - support.log(texto) + support.info(texto) item.url = host + "/?s=" + texto return peliculas(item) def newest(categoria): - support.log(categoria) + support.info(categoria) itemlist = [] item = support.Item() try: @@ -52,7 +52,7 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.logger.error("{0}".format(line)) + support.infoger.error("{0}".format(line)) return [] return itemlist @@ -80,7 +80,7 @@ def check(item): @support.scrape def episodios(item): - support.log('EPISODIOS ', item.data) + support.info('EPISODIOS ', item.data) data = '' matches = item.data season = 1 diff --git a/channels/cineblog01.py b/channels/cineblog01.py index fbc579fe..4a6d655c 100644 --- a/channels/cineblog01.py +++ b/channels/cineblog01.py @@ -55,7 +55,7 @@ def menu(item): def newest(categoria): - support.log(categoria) + support.info(categoria) item = support.Item() try: @@ -71,12 +71,12 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.logger.error("{0}".format(line)) + support.infoger.error("{0}".format(line)) return [] def search(item, text): - support.log(item.url, "search", text) + support.info(item.url, "search", text) try: item.url = item.url + "/?s=" + text.replace(' ', '+') @@ -127,7 +127,7 @@ def peliculas(item): def episodios(item): # support.dbg() data = support.match(item.url, headers=headers).data - support.log(data) + support.info(data) if 'TUTTA LA ' in data: folderUrl = scrapertools.find_single_match(data, r'TUTTA LA \w+\s+(?:–|-)\s+<a href="?([^" ]+)') data = httptools.downloadpage(folderUrl).data @@ -170,14 +170,14 @@ def findvideos(item): def load_links(itemlist, re_txt, desc_txt, quality=""): streaming = scrapertools.find_single_match(data, re_txt).replace('"', '') - support.log('STREAMING', streaming) - support.log('STREAMING=', streaming) + support.info('STREAMING', streaming) + support.info('STREAMING=', streaming) matches = support.match(streaming, patron = r'<td><a.*?href=([^ ]+) [^>]+>([^<]+)<').matches for scrapedurl, scrapedtitle in matches: logger.debug("##### findvideos %s ## %s ## %s ##" % (desc_txt, scrapedurl, scrapedtitle)) itemlist.append(item.clone(action="play", title=scrapedtitle, url=scrapedurl, server=scrapedtitle, quality=quality)) - support.log() + support.info() itemlist = [] @@ -210,7 +210,7 @@ def findvideos(item): def findvid_serie(item): def load_vid_series(html, item, itemlist, blktxt): - support.log('HTML',html) + support.info('HTML',html) # Estrae i contenuti matches = support.match(html, patron=r'<a href=(?:")?([^ "]+)[^>]+>(?!<!--)(.*?)(?:</a>|<img)').matches for url, server in matches: @@ -218,7 +218,7 @@ def findvid_serie(item): if 'swzz' in item.url: item.url = support.swzz_get_url(item) itemlist.append(item) - support.log() + support.info() itemlist = [] @@ -238,5 +238,5 @@ def findvid_serie(item): def play(item): - support.log() + support.info() return servertools.find_video_items(item, data=item.url) diff --git a/channels/cinemalibero.py b/channels/cinemalibero.py index 2fedaf89..1ea7df7f 100644 --- a/channels/cinemalibero.py +++ b/channels/cinemalibero.py @@ -89,13 +89,13 @@ def episodios(item): data=item.data # debugBlock=True if item.args == 'anime': - support.log("Anime :", item) + support.info("Anime :", item) # blacklist = ['Clipwatching', 'Verystream', 'Easybytez', 'Flix555', 'Cloudvideo'] patron = r'<a target=(?P<url>[^>]+>(?P<title>Episodio\s(?P<episode>\d+))(?::)?(?:(?P<title2>[^<]+))?.*?(?:<br|</p))' patronBlock = r'(?:Stagione (?P<season>\d+))?(?:</span><br />|</span></p>|strong></p>)(?P<block>.*?)(?:<div style="margin-left|<span class="txt_dow">)' item.contentType = 'tvshow' else:# item.extra == 'serie': - support.log("Serie :", item) + support.info("Serie :", item) patron = r'(?:>| )(?P<episode>\d+(?:x|×|×)\d+)[;]?[ ]?(?:(?P<title>[^<–-]+)(?P<url>.*?)|(\2[ ])(?:<(\3.*?)))(?:</a><br /|</a></p|$)' patronBlock = r'>(?:[^<]+[Ss]tagione\s|[Ss]tagione [Uu]nica)(?:(?P<lang>iTA|ITA|Sub-ITA|Sub-iTA))?.*?</strong>(?P<block>.+?)(?:<strong|<div class="at-below)' item.contentType = 'tvshow' @@ -118,7 +118,7 @@ def genres(item): def search(item, texto): - support.log(item.url,texto) + support.info(item.url,texto) texto = texto.replace(' ', '+') item.url = host + "/?s=" + texto # item.contentType = 'tv' @@ -129,11 +129,11 @@ def search(item, texto): except: import sys for line in sys.exc_info(): - support.log("%s" % line) + support.info("%s" % line) return [] def newest(categoria): - support.log('newest ->', categoria) + support.info('newest ->', categoria) itemlist = [] item = Item() item.args = 'newest' @@ -147,13 +147,13 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.log('newest log: ', (line)) + support.info('newest log: ', (line)) return [] return itemlist def check(item): - support.log() + support.info() data = support.match(item.url, headers=headers).data if data: blockAnime = support.match(data, patron=r'<div id="container" class="container">(.+?<div style="margin-left)').match @@ -175,6 +175,6 @@ def check(item): return findvideos(item) def findvideos(item): - support.log() + support.info() item.url = item.url.replace('http://rapidcrypt.net/verys/', '').replace('http://rapidcrypt.net/open/', '') #blocca la ricerca return support.server(item, data= item.url) diff --git a/channels/cinetecadibologna.py b/channels/cinetecadibologna.py index a2fd96af..33591e2f 100644 --- a/channels/cinetecadibologna.py +++ b/channels/cinetecadibologna.py @@ -34,7 +34,7 @@ def menu(item): def search(item, text): - support.log(text) + support.info(text) item.args = 'noorder' item.url = host + '/ricerca/type_ALL/ricerca_' + text item.contentType = 'movie' @@ -44,7 +44,7 @@ def search(item, text): except: import sys for line in sys.exc_info(): - support.logger.error("%s" % line) + support.infoger.error("%s" % line) return [] @@ -62,7 +62,7 @@ def peliculas(item): def findvideos(item): - support.log() + support.info() itemlist = [] matches = support.match(item, patron=r'filename: "(.*?)"').matches diff --git a/channels/dreamsub.py b/channels/dreamsub.py index 638a4581..b93b1647 100644 --- a/channels/dreamsub.py +++ b/channels/dreamsub.py @@ -10,7 +10,7 @@ headers = [['Referer', host]] @support.menu def mainlist(item): - support.log(item) + support.info(item) anime = ['/search?typeY=tv', ('Movie', ['/search?typeY=movie', 'peliculas', '', 'movie']), @@ -34,7 +34,7 @@ def menu(item): patronMenu = r'<a class="[^"]+" data-state="[^"]+" (?P<other>[^>]+)>[^>]+></i>[^>]+></i>[^>]+></i>(?P<title>[^>]+)</a>' def itemHook(item): - support.log(item.type) + support.info(item.type) for Type, ID in support.match(item.other, patron=r'data-type="([^"]+)" data-id="([^"]+)"').matches: item.url = host + '/search?' + Type + 'Y=' + ID return item @@ -42,7 +42,7 @@ def menu(item): def search(item, text): - support.log(text) + support.info(text) text = text.replace(' ', '+') item.url = host + '/search/' + text @@ -53,12 +53,12 @@ def search(item, text): except: import sys for line in sys.exc_info(): - support.log('search log:', line) + support.info('search log:', line) return [] def newest(categoria): - support.log(categoria) + support.info(categoria) item = support.Item() try: if categoria == "anime": @@ -69,7 +69,7 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.logger.error("{0}".format(line)) + support.infoger.error("{0}".format(line)) return [] @@ -110,7 +110,7 @@ def episodios(item): def findvideos(item): itemlist = [] - support.log() + support.info() # support.dbg() matches = support.match(item, patron=r'href="([^"]+)"', patronBlock=r'<div style="white-space: (.*?)<div id="main-content"') @@ -124,7 +124,7 @@ def findvideos(item): itemlist.append(item.clone(action="play", title='VVVVID', url=support.match(matches.data, patron=r'(http://www.vvvvid[^"]+)').match, server='vvvvid')) else: # matches.matches.sort() - support.log('VIDEO') + support.info('VIDEO') for url in matches.matches: lang = url.split('/')[-2] if 'ita' in lang.lower(): diff --git a/channels/dsda.py b/channels/dsda.py index 8fc342c1..fdc72fa1 100644 --- a/channels/dsda.py +++ b/channels/dsda.py @@ -33,7 +33,7 @@ def menu(item): return locals() def newest(categoria): - support.log() + support.info() item = Item() try: if categoria == "documentales": @@ -45,12 +45,12 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.logger.error("{0}".format(line)) + support.infoger.error("{0}".format(line)) return [] def search(item, texto): - support.log(texto) + support.info(texto) item.url = host + "/?s=" + texto try: return peliculas(item) @@ -133,7 +133,7 @@ def episodios(item): def findvideos(item): - support.log() + support.info() if item.args == 'raccolta' or item.contentType == 'episode': return support.server(item, item.url) else: diff --git a/channels/eurostreaming.py b/channels/eurostreaming.py index 532b46d7..167fe68a 100644 --- a/channels/eurostreaming.py +++ b/channels/eurostreaming.py @@ -17,7 +17,7 @@ headers = [['Referer', host]] @support.menu def mainlist(item): - support.log() + support.info() tvshow = [] anime = ['/category/anime-cartoni-animati/'] mix = [('Aggiornamenti {bullet bold} {TV}', ['/aggiornamento-episodi/', 'peliculas', 'newest']), @@ -57,7 +57,7 @@ def episodios(item): def search(item, texto): - support.log() + support.info() item.url = "%s/?s=%s" % (host, texto) item.contentType = 'tvshow' @@ -69,12 +69,12 @@ def search(item, texto): except: import sys for line in sys.exc_info(): - support.log(line) + support.info(line) return [] def newest(categoria): - support.log() + support.info() itemlist = [] item = Item() @@ -88,12 +88,12 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.log("{0}".format(line)) + support.info("{0}".format(line)) return [] return itemlist def findvideos(item): - support.log() + support.info() return support.server(item, item.url) diff --git a/channels/fastsubita.py b/channels/fastsubita.py index 0e23e097..088c3a43 100644 --- a/channels/fastsubita.py +++ b/channels/fastsubita.py @@ -18,7 +18,7 @@ """ from core import support, httptools, scrapertools from core.item import Item -from core.support import log +from core.support import info from platformcode import config host = config.get_channel_url() @@ -42,7 +42,7 @@ def mainlist(item): @support.scrape def peliculas(item): - support.log(item) + support.info(item) # support.dbg() deflang = 'Sub-ITA' @@ -67,7 +67,7 @@ def peliculas(item): @support.scrape def episodios(item): - support.log(item) + support.info(item) #support.dbg() deflang = 'Sub-ITA' @@ -82,7 +82,7 @@ def episodios(item): @support.scrape def genres(item): - support.log() + support.info() #support.dbg() action = 'peliculas' @@ -99,7 +99,7 @@ def genres(item): def search(item, text): - support.log('search', item) + support.info('search', item) text = text.replace(' ', '+') item.url = host + '?s=' + text try: @@ -110,12 +110,12 @@ def search(item, text): except: import sys for line in sys.exc_info(): - log('search log:', line) + info('search log:', line) return [] def newest(categoria): - support.log('newest ->', categoria) + support.info('newest ->', categoria) itemlist = [] item = Item() if categoria == 'series': @@ -132,14 +132,14 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.log('newest log: ', line) + support.info('newest log: ', line) return [] return itemlist def findvideos(item): - support.log('findvideos ->', item) + support.info('findvideos ->', item) itemlist = [] patronBlock = '<div class="entry-content">(?P<block>.*)<footer class="entry-footer">' patron = r'<a href="([^"]+)">' diff --git a/channels/filmigratis.py b/channels/filmigratis.py index 4f9335cb..22ec19da 100644 --- a/channels/filmigratis.py +++ b/channels/filmigratis.py @@ -41,7 +41,7 @@ def mainlist(item): @support.scrape def peliculas(item): - support.log() + support.info() if item.args == 'search': action = '' @@ -88,7 +88,7 @@ def peliculas(item): @support.scrape def episodios(item): - support.log() + support.info() action = 'findvideos' patronBlock = r'<div class="row">(?P<block>.*?)<section class="main-content">' @@ -98,7 +98,7 @@ def episodios(item): @support.scrape def genres(item): - support.log() + support.info() if item.contentType == 'movie': action = 'peliculas' @@ -115,7 +115,7 @@ def genres(item): def search(item, text): - support.log('search', item) + support.info('search', item) text = text.replace(' ', '+') item.url = host + '/search/?s=' + text @@ -126,11 +126,11 @@ def search(item, text): except: import sys for line in sys.exc_info(): - support.log('search log:', line) + support.info('search log:', line) return [] def newest(categoria): - support.log('newest ->', categoria) + support.info('newest ->', categoria) itemlist = [] item = Item() try: @@ -146,11 +146,11 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.log({0}.format(line)) + support.info({0}.format(line)) return [] return itemlist def findvideos(item): - support.log() + support.info() return support.server(item) diff --git a/channels/filmpertutti.py b/channels/filmpertutti.py index 6b09c7d4..b3822d9b 100644 --- a/channels/filmpertutti.py +++ b/channels/filmpertutti.py @@ -33,7 +33,7 @@ def mainlist(item): @support.scrape def peliculas(item): - support.log() + support.info() if item.args != 'newest': patronBlock = r'<ul class="posts">(?P<block>.*)<\/ul>' @@ -94,21 +94,21 @@ def genres(item): def select(item): - support.log() + support.info() patron=r'class="taxonomy category" ><span property="name">([^>]+)</span></a><meta property="position" content="2">' block = support.match(item.url, patron=patron,headers=headers).match if block.lower() != 'film': - support.log('select = ### è una serie ###') + support.info('select = ### è una serie ###') item.contentType='tvshow' return episodios(item) else: - support.log('select = ### è un movie ###') + support.info('select = ### è un movie ###') item.contentType='movie' return findvideos(item) def search(item, texto): - support.log() + support.info() item.url = host + "/?s=" + texto item.contentType = 'episode' item.args = 'search' @@ -118,12 +118,12 @@ def search(item, texto): except: import sys for line in sys.exc_info(): - support.log("%s" % line) + support.info("%s" % line) return [] def newest(categoria): - support.log() + support.info() itemlist = [] item = Item() try: @@ -144,7 +144,7 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.log("{0}".format(line)) + support.info("{0}".format(line)) return [] return itemlist diff --git a/channels/guardaseriecam.py b/channels/guardaseriecam.py index 3fb78a62..1dc671de 100644 --- a/channels/guardaseriecam.py +++ b/channels/guardaseriecam.py @@ -11,7 +11,7 @@ # possibilità di miglioramento: inserire menu per genere - lista serie tv e gestire le novità from core import support -from core.support import log +from core.support import info from platformcode import logger, config host = config.get_channel_url() @@ -51,7 +51,7 @@ def episodios(item): def search(item, text): - support.log('search', item) + support.info('search', item) item.contentType = 'tvshow' itemlist = [] text = text.replace(' ', '+') @@ -61,7 +61,7 @@ def search(item, text): except: import sys for line in sys.exc_info(): - log('search log:', line) + info('search log:', line) return [] diff --git a/channels/guardaserieclick.py b/channels/guardaserieclick.py index e81dd22f..1ed39ed4 100644 --- a/channels/guardaserieclick.py +++ b/channels/guardaserieclick.py @@ -17,7 +17,7 @@ from core import support from core.item import Item from platformcode import config -from core.support import log +from core.support import info host = config.get_channel_url() headers = [['Referer', host]] @@ -40,7 +40,7 @@ def mainlist(item): ##@support.scrape ##def peliculas(item): #### import web_pdb; web_pdb.set_trace() -## log('peliculas ->\n', item) +## info('peliculas ->\n', item) ## ## action = 'episodios' ## block = r'(?P<block>.*?)<div\s+class="btn btn-lg btn-default btn-load-other-series">' @@ -77,7 +77,7 @@ def mainlist(item): @support.scrape def peliculas(item): ## import web_pdb; web_pdb.set_trace() - log('peliculas ->\n', item) + info('peliculas ->\n', item) action = 'episodios' blacklist = ['DMCA'] @@ -122,7 +122,7 @@ def peliculas(item): @support.scrape def episodios(item): - log() + info() action = 'findvideos' patron = r'<div class="number-episodes-on-img">\s?\d+.\d+\s?(?:\((?P<lang>[a-zA-Z\-]+)\))?</div>.+?(?:<span class="pull-left bottom-year">(?P<title2>[^<]+)<[^>]+>[^>]+>[^>]+>[^>]+>[^>]+>[^>]+>(?P<plot>[^<]+)<[^>]+>[^>]+>[^>]+>\s?)?<span(?: meta-nextep="[^"]+")? class="[^"]+" meta-serie="(?P<title>[^"]+)" meta-stag="(?P<season>\d+)" meta-ep="(?P<episode>\d+)" meta-embed="(?P<url>[^>]+)">' @@ -137,7 +137,7 @@ def episodios(item): @support.scrape def genres(item): - log() + info() action = 'peliculas' patron = r'<li>\s<a\shref="(?P<url>[^"]+)"[^>]+>(?P<title>[^<]+)</a></li>' @@ -148,7 +148,7 @@ def genres(item): def search(item, text): - log(text) + info(text) item.url = host + "/?s=" + text item.contentType = 'tvshow' item.args = 'search' @@ -158,11 +158,11 @@ def search(item, text): except: import sys for line in sys.exc_info(): - log("%s" % line) + info("%s" % line) return [] def newest(categoria): - log() + info() itemlist = [] item = Item() item.contentType= 'tvshow' @@ -177,12 +177,12 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - log("{0}".format(line)) + info("{0}".format(line)) return [] return itemlist def findvideos(item): - log('--->', item) + info('--->', item) return support.server(item, item.url) diff --git a/channels/guardaserieicu.py b/channels/guardaserieicu.py index fda9c818..42beaf9d 100644 --- a/channels/guardaserieicu.py +++ b/channels/guardaserieicu.py @@ -10,7 +10,7 @@ # possibilità di miglioramento: gestire le novità (sezione Ultimi episodi sul sito) -from core.support import log +from core.support import info from core import support from platformcode import config @@ -40,7 +40,7 @@ def episodios(item): return locals() def search(item, text): - log(text) + info(text) itemlist = [] text = text.replace(' ', '+') item.url = host + "/?s=" + text @@ -50,10 +50,10 @@ def search(item, text): except: import sys for line in sys.exc_info(): - log("%s" % line) + info("%s" % line) return [] def findvideos(item): - support.log('findvideos', item) + support.info('findvideos', item) return support.server(item, headers=headers) \ No newline at end of file diff --git a/channels/hd4me.py b/channels/hd4me.py index 8cd2cbc1..2c3e0659 100644 --- a/channels/hd4me.py +++ b/channels/hd4me.py @@ -42,7 +42,7 @@ def genre(item): def search(item, text): - support.log(text) + support.info(text) item.url = host + '/?s=' + text try: return peliculas(item) @@ -50,7 +50,7 @@ def search(item, text): except: import sys for line in sys.exc_info(): - support.logger.error("search except: %s" % line) + support.infoger.error("search except: %s" % line) return [] diff --git a/channels/ilcorsaronero.py b/channels/ilcorsaronero.py index 6f4a6502..7060b2d1 100644 --- a/channels/ilcorsaronero.py +++ b/channels/ilcorsaronero.py @@ -11,7 +11,7 @@ def findhost(): return url[:-1] if url.endswith('/') else url host = support.config.get_channel_url(findhost) -support.log('HOST',host) +support.info('HOST',host) # host = 'https://ilcorsaronero.xyz' headers = [['Referer', host]] @@ -64,7 +64,7 @@ def peliculas(item): def search(item, text): - support.log(item, text) + support.info(item, text) if 'all' in item.args: item.url += text else: @@ -75,7 +75,7 @@ def search(item, text): except: import sys for line in sys.exc_info(): - support.logger.error("search except: %s" % line) + support.infoger.error("search except: %s" % line) return [] diff --git a/channels/ilgeniodellostreaming.py b/channels/ilgeniodellostreaming.py index 08a88d70..fe5cd59d 100644 --- a/channels/ilgeniodellostreaming.py +++ b/channels/ilgeniodellostreaming.py @@ -6,7 +6,7 @@ from core import support -from core.support import log +from core.support import info from core.item import Item from platformcode import config @@ -15,7 +15,7 @@ headers = [['Referer', host]] @support.menu def mainlist(item): - support.log(item) + support.info(item) film = ['/film/', ('Generi',['', 'genres', 'genres']), @@ -46,7 +46,7 @@ def mainlist(item): @support.scrape def peliculas(item): - log() + info() # debug = True if item.args == 'search': @@ -96,7 +96,7 @@ def peliculas(item): @support.scrape def episodios(item): - log() + info() patronBlock = r'<h1>.*?[ ]?(?:\[(?P<lang>.+?\]))?</h1>.+?<div class="se-a" '\ 'style="display:block"><ul class="episodios">(?P<block>.*?)</ul>'\ @@ -109,7 +109,7 @@ def episodios(item): @support.scrape def genres(item): - log(item) + info(item) action='peliculas' if item.args == 'genres': @@ -125,7 +125,7 @@ def genres(item): return locals() def search(item, text): - log(text) + info(text) itemlist = [] text = text.replace(' ', '+') item.url = host + "/search/" + text @@ -135,12 +135,12 @@ def search(item, text): except: import sys for line in sys.exc_info(): - log("%s" % line) + info("%s" % line) return [] def newest(categoria): - log(categoria) + info(categoria) itemlist = [] item = Item() @@ -161,14 +161,14 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - log("{0}".format(line)) + info("{0}".format(line)) return [] return itemlist def findvideos(item): - log() + info() matches = support.match(item, patron=[r'class="metaframe rptss" src="([^"]+)"',r' href="#option-\d">([^\s]+)\s*([^\s]+)']).matches itemlist = [] list_url = [] diff --git a/channels/ilgeniodellostreaming_cam.py b/channels/ilgeniodellostreaming_cam.py index 8c8d6015..14d48ae6 100644 --- a/channels/ilgeniodellostreaming_cam.py +++ b/channels/ilgeniodellostreaming_cam.py @@ -5,7 +5,7 @@ from core import support -from core.support import log +from core.support import info from core.item import Item from platformcode import config @@ -54,7 +54,7 @@ def genres(item): return locals() def search(item, text): - log(text) + info(text) text = text.replace(' ', '+') item.url = host + "/search/" + text try: @@ -62,12 +62,12 @@ def search(item, text): except: import sys for line in sys.exc_info(): - log("%s" % line) + info("%s" % line) return [] def newest(categoria): - log(categoria) + info(categoria) itemlist = [] item = Item() @@ -81,12 +81,12 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - log("{0}".format(line)) + info("{0}".format(line)) return [] return itemlist def findvideos(item): - log() + info() return support.server(item) \ No newline at end of file diff --git a/channels/italiaserie.py b/channels/italiaserie.py index 02d04680..b56d643b 100644 --- a/channels/italiaserie.py +++ b/channels/italiaserie.py @@ -25,7 +25,7 @@ headers = [['Referer', host]] @support.menu def mainlist(item): - support.log() + support.info() tvshow = ['/category/serie-tv/', ('Aggiornamenti', ['/ultimi-episodi/', 'peliculas', 'update']), @@ -37,7 +37,7 @@ def mainlist(item): @support.scrape def peliculas(item): - support.log() + support.info() action = 'episodios' patron = r'<div class="post-thumb">\s*<a href="(?P<url>[^"]+)" '\ @@ -54,7 +54,7 @@ def peliculas(item): @support.scrape def episodios(item): - support.log() + support.info() patronBlock = r'</i> Stagione (?P<block>(?P<season>\d+)</div> '\ '<div class="su-spoiler-content".*?)<div class="clearfix">' @@ -70,7 +70,7 @@ def episodios(item): @support.scrape def category(item): - support.log() + support.info() action = 'peliculas' patron = r'<li class="cat-item.*?href="(?P<url>[^"]+)".*?>(?P<title>.*?)</a>' @@ -79,7 +79,7 @@ def category(item): def search(item, texto): - support.log("s=", texto) + support.info("s=", texto) item.url = host + "/?s=" + texto item.contentType = 'tvshow' try: @@ -88,12 +88,12 @@ def search(item, texto): except: import sys for line in sys.exc_info(): - support.log("%s" % line) + support.info("%s" % line) return [] def newest(categoria): - support.log("newest", categoria) + support.info("newest", categoria) itemlist = [] item = Item() try: @@ -111,14 +111,14 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.log("{0}".format(line)) + support.info("{0}".format(line)) return [] return itemlist def findvideos(item): - support.log() + support.info() if item.args == 'update': itemlist = [] diff --git a/channels/la7.py b/channels/la7.py index f42f8601..16f7c3c2 100644 --- a/channels/la7.py +++ b/channels/la7.py @@ -117,7 +117,7 @@ def episodios(item): def play(item): - support.log() + support.info() data = support.match(item).data match = support.match(data, patron='/content/entry/data/(.*?).mp4').match if match: diff --git a/channels/mediasetplay.py b/channels/mediasetplay.py index 6d1d7cb3..834aef8e 100644 --- a/channels/mediasetplay.py +++ b/channels/mediasetplay.py @@ -49,7 +49,7 @@ def mainlist(item): def search(item, text): - support.log(text) + support.info(text) item.search = text if not item.args: item.contentType = 'undefined' @@ -60,11 +60,11 @@ def search(item, text): except: import sys for line in sys.exc_info(): - support.logger.error("%s" % line) + support.infoger.error("%s" % line) return itemlist def menu(item): - support.log() + support.info() itemlist = [item.clone(title=support.typo(item.args[0],'bullet bold'), url='', action='peliculas')] if item.url: json = get_from_id(item) @@ -74,7 +74,7 @@ def menu(item): return itemlist def live(item): - support.log() + support.info() itemlist = [] json = current_session.get(item.url).json()['entries'] for it in json: @@ -99,7 +99,7 @@ def live(item): return support.thumb(itemlist, live=True) def peliculas(item): - support.log() + support.info() itemlist = [] contentType = '' json = get_programs(item) @@ -144,7 +144,7 @@ def peliculas(item): return itemlist def epmenu(item): - support.log() + support.info() itemlist = [] entries = current_session.get('https://feed.entertainment.tv.theplatform.eu/f/PR1GhC/mediaset-prod-all-brands?byCustomValue={brandId}{' + item.urls + '}').json()['entries'] for entry in entries: @@ -159,7 +159,7 @@ def epmenu(item): def episodios(item): - support.log() + support.info() itemlist = [] episode = '' @@ -188,12 +188,12 @@ def episodios(item): return sorted(itemlist, key=lambda it: it.title) def findvideos(item): - support.log() + support.info() itemlist = [support.Item(server = 'directo', title = 'Direct', url = item.urls, action = 'play')] return support.server(item, itemlist=itemlist, Download=False) def play(item): - support.log() + support.info() if not item.urls: urls = item.url else: urls = item.urls for url in urls: @@ -209,13 +209,13 @@ def play(item): return support.servertools.find_video_items(item, data=data) def subBrand(json): - support.log() + support.info() subBrandId = current_session.get('https://feed.entertainment.tv.theplatform.eu/f/PR1GhC/mediaset-prod-all-brands?byCustomValue={brandId}{' + json + '}').json()['entries'][-1]['mediasetprogram$subBrandId'] json = current_session.get('https://feed.entertainment.tv.theplatform.eu/f/PR1GhC/mediaset-prod-all-programs?byCustomValue={subBrandId}{' + subBrandId + '}').json()['entries'] return json def get_from_id(item): - support.log() + support.info() json = current_session.get(entry.format(id=item.url)).json() if 'components' in json: id = quote(",".join(json["components"])) @@ -225,7 +225,7 @@ def get_from_id(item): return {} def get_programs(item, ret=[], args={}): - support.log() + support.info() hasMore = False if not args: if item.url: diff --git a/channels/metalvideo.py b/channels/metalvideo.py index 8164dc36..ebcc8298 100644 --- a/channels/metalvideo.py +++ b/channels/metalvideo.py @@ -56,7 +56,7 @@ def findvideos(item): def search(item, text): - support.log(text) + support.info(text) item.url = host + '/search.php?keywords=' + text + '&video-id=' try: return peliculas(item) @@ -64,5 +64,5 @@ def search(item, text): except: import sys for line in sys.exc_info(): - support.logger.error("%s" % line) + support.infoger.error("%s" % line) return [] diff --git a/channels/mondoserietv.py b/channels/mondoserietv.py index 13997cd3..5f4a9de8 100644 --- a/channels/mondoserietv.py +++ b/channels/mondoserietv.py @@ -35,7 +35,7 @@ def mainlist(item): def search(item, text): - support.log(text) + support.info(text) if item.contentType == 'movie' or item.extra == 'movie': action = 'findvideos' else: @@ -48,12 +48,12 @@ def search(item, text): except: import sys for line in sys.exc_info(): - support.logger.error("%s" % line) + support.infoger.error("%s" % line) return [] def newest(categoria): - support.log(categoria) + support.info(categoria) item = support.Item() try: if categoria == "series": @@ -69,7 +69,7 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.logger.error("{0}".format(line)) + support.infoger.error("{0}".format(line)) return [] diff --git a/channels/paramount.py b/channels/paramount.py index bf3e2db7..5500383c 100644 --- a/channels/paramount.py +++ b/channels/paramount.py @@ -27,7 +27,7 @@ def menu(item): def search(item, text): - support.log(text) + support.info(text) item.search = text.replace(' ','+') try: @@ -36,7 +36,7 @@ def search(item, text): except: import sys for line in sys.exc_info(): - support.logger.error("%s" % line) + support.infoger.error("%s" % line) return [] @@ -54,7 +54,7 @@ def live(item): if url not in urls: urls.append(url) info = jsontools.load(support.match(host +'/api/on-air?channelId=' + ch_dict[title]).data) - support.log(info) + support.info(info) plot= '[B]' + info['seriesTitle'] +'[/B]\n' + info['description'] if 'seriesTitle' in info else '' itemlist.append(item.clone(title=support.typo(title,'bold'), contentTitle=title, url=host+url, plot=plot, action='findvideos')) return itemlist @@ -76,7 +76,7 @@ def peliculas(item): for it in data: title = it['meta']['header']['title'] - support.log(title, it) + support.info(title, it) d = it['meta']['date'].split('/') if it['meta']['date'] else ['0000','00','00'] date = int(d[2] + d[1] + d[0]) if item.search.lower() in title.lower() \ diff --git a/channels/piratestreaming.py b/channels/piratestreaming.py index ed2082cf..13c6a29d 100644 --- a/channels/piratestreaming.py +++ b/channels/piratestreaming.py @@ -5,7 +5,7 @@ from core import support -from core.support import config, log +from core.support import config, info host = config.get_channel_url() headers = [['Referer', host]] @@ -23,7 +23,7 @@ def mainlist(item): def search(item, texto): - log(texto) + info(texto) item.url = host + "/?s=" + texto try: return peliculas(item) @@ -31,12 +31,12 @@ def search(item, texto): except: import sys for line in sys.exc_info(): - support.logger.error("%s" % line) + support.infoger.error("%s" % line) return [] def newest(categoria): - support.log(categoria) + support.info(categoria) itemlist = [] item = support.Item() try: @@ -56,7 +56,7 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.logger.error("{0}".format(line)) + support.infoger.error("{0}".format(line)) return [] return itemlist diff --git a/channels/polpotv.py b/channels/polpotv.py index f03b16a6..c5771c7d 100644 --- a/channels/polpotv.py +++ b/channels/polpotv.py @@ -36,7 +36,7 @@ def mainlist(item): return locals() def newest(categoria): - support.log() + support.info() item = Item() if categoria == 'peliculas': item.contentType = 'movie' @@ -47,7 +47,7 @@ def newest(categoria): return peliculas(item) def peliculas(item): - support.log() + support.info() itemlist = [] data = support.match(item.url, headers=headers).data @@ -68,7 +68,7 @@ def peliculas(item): return itemlist def episodios(item): - support.log() + support.info() itemlist = [] data = support.match(item.url, headers=headers).data json_object = jsontools.load(data) @@ -84,7 +84,7 @@ def episodios(item): return itemlist def get_season(item, seas_url, seasonNumber): - support.log() + support.info() itemlist = [] data = support.match(seas_url, headers=headers).data json_object = jsontools.load(data) @@ -98,7 +98,7 @@ def get_season(item, seas_url, seasonNumber): return itemlist[::-1] def search(item, texto): - support.log(item.url, "search", texto) + support.info(item.url, "search", texto) itemlist=[] try: item.url = host + "/api/movies?originalTitle="+texto+"&translations.name=" +texto @@ -118,11 +118,11 @@ def search(item, texto): except: import sys for line in sys.exc_info(): - support.logger.error("%s" % line) + support.infoger.error("%s" % line) return [] def search_movie_by_genre(item): - support.log() + support.info() itemlist = [] data = support.match(item.url, headers=headers).data json_object = jsontools.load(data) @@ -135,7 +135,7 @@ def search_movie_by_genre(item): return support.thumb(itemlist, True) def search_movie_by_year(item): - support.log() + support.info() now = datetime.datetime.now() year = int(now.year) itemlist = [] @@ -150,7 +150,7 @@ def search_movie_by_year(item): return itemlist def findvideos(item): - support.log() + support.info() itemlist = [] try: data = support.match(item.url, headers=headers).data @@ -171,7 +171,7 @@ def findvideos(item): return support.server(item, itemlist=itemlist) def get_itemlist_element(element,item): - support.log() + support.info() itemlist=[] contentSerieName = '' contentTitle ='' diff --git a/channels/pufimovies.py b/channels/pufimovies.py index a9dddb44..962a13c0 100644 --- a/channels/pufimovies.py +++ b/channels/pufimovies.py @@ -38,7 +38,7 @@ def menu(item): def search(item, text): - support.log('search', item) + support.info('search', item) itemlist = [] text = text.replace(' ', '+') item.url = host + '/search/keyword/' + text @@ -52,12 +52,12 @@ def search(item, text): except: import sys for line in sys.exc_info(): - support.log('search log:', line) + support.info('search log:', line) return [] def newest(categoria): - support.log(categoria) + support.info(categoria) itemlist = [] item = support.Item() item.url = host @@ -77,7 +77,7 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.logger.error("%s" % line) + support.infoger.error("%s" % line) return [] return itemlist @@ -109,6 +109,6 @@ def episodios(item): def findvideos(item): - support.log() + support.info() # match = support.match(item, patron='wstream', debug=True) return support.server(item) diff --git a/channels/raiplay.py b/channels/raiplay.py index 3bef9aab..88fe6861 100644 --- a/channels/raiplay.py +++ b/channels/raiplay.py @@ -38,7 +38,7 @@ def mainlist(item): def menu(item): - support.log() + support.info() itemlist = [item.clone(title = support.typo('Tutti','bullet bold'), action = 'peliculas'), item.clone(title = support.typo('Generi','submenu'), args = 'genre', action = 'submenu'), item.clone(title = support.typo('A-Z','submenu'), args = 'az', action = 'submenu'), @@ -48,7 +48,7 @@ def menu(item): def learning(item): - support.log() + support.info() itemlist =[] json = current_session.get(item.url).json()['contents'] for key in json: @@ -58,7 +58,7 @@ def learning(item): def submenu(item): - support.log() + support.info() itemlist = [] json = current_session.get(item.url).json()['contents'][-1]['contents'] if item.args == 'az': @@ -76,7 +76,7 @@ def submenu(item): def replay_menu(item): - support.log() + support.info() import datetime, xbmc # create day and month list @@ -91,14 +91,14 @@ def replay_menu(item): today = datetime.date.today() for d in range(7): day = today - datetime.timedelta(days=d) - support.log(day) + support.info(day) itemlist.append(item.clone(action = 'replay_channels', date = day.strftime("%d-%m-%Y"), title = support.typo(days[int(day.strftime("%w"))] + " " + day.strftime("%d") + " " + months[int(day.strftime("%m"))-1], 'bold'))) return itemlist def replay_channels(item): - support.log() + support.info() itemlist = [] json = current_session.get(item.url).json()['dirette'] for key in json: @@ -108,18 +108,18 @@ def replay_channels(item): def replay(item): - support.log() + support.info() itemlist = [] json = current_session.get(item.url).json()[item.fulltitle][0]['palinsesto'][0]['programmi'] for key in json: - support.log('KEY=',key) + support.info('KEY=',key) if key and key['pathID']: itemlist.append(item.clone(thumbnail = getUrl(key['images']['landscape']), fanart = getUrl(key['images']['landscape']), url = getUrl(key['pathID']), fulltitle = key['name'], show = key['name'], title = support.typo(key['timePublished'], 'color kod bold') + support.typo(' | ' + key['name'], ' bold'), plot = key['testoBreve'], action = 'findvideos')) return itemlist def search(item, text): # support.dbg() - support.log() + support.info() itemlist =[] try: if item.url != host: @@ -136,7 +136,7 @@ def search(item, text): except: import sys for line in sys.exc_info(): - support.logger.error("%s" % line) + support.infoger.error("%s" % line) return [] return itemlist @@ -153,12 +153,12 @@ def Type(item): def live(item): - support.log() + support.info() itemlist =[] info={} json = current_session.get(item.url).json()['dirette'] onAir = current_session.get(onair).json()['on_air'] - support.log(onAir) + support.info(onAir) for key in onAir: channel = key['channel'] info[channel] = {} @@ -174,7 +174,7 @@ def live(item): def peliculas(item): - support.log() + support.info() itemlist = [] keys = [] key_list = [] @@ -222,7 +222,7 @@ def peliculas(item): def select(item): - support.log() + support.info() itemlist = [] json = current_session.get(item.url).json()['blocks'] for key in json: @@ -234,7 +234,7 @@ def select(item): def episodios(item): - support.log() + support.info() itemlist = [] if type(item.url) in [list, dict] and len(item.url) > 1 and ('name' in item.url[0] and 'stagione' not in item.url[0]['name'].lower()): for key in item.url: @@ -276,7 +276,7 @@ def episodios(item): def findvideos(item): - support.log() + support.info() itemlist = [] if item.url.endswith('json'): json = current_session.get(item.url).json() @@ -293,7 +293,7 @@ def findvideos(item): def getUrl(pathId): - support.log() + support.info() url = pathId.replace(" ", "%20") if url.startswith("/raiplay/"): url = url.replace("/raiplay/",host +'/') @@ -315,7 +315,7 @@ def getUrl(pathId): def addinfo(key, item): - support.log() + support.info() info = current_session.get(getUrl(key['info_url'])).json() if not item.search or item.search.lower() in key['name'].lower(): it = item.clone(title = support.typo(key['name'],'bold'), fulltitle = key['name'], show = key['name'], @@ -333,7 +333,7 @@ def addinfo(key, item): def load_episodes(key, item): - support.log() + support.info() itemlist = [] json = current_session.get(getUrl(key['path_id'])).json()['items'] order = 0 diff --git a/channels/seriehd.py b/channels/seriehd.py index c7a6139b..cd8ca5c5 100644 --- a/channels/seriehd.py +++ b/channels/seriehd.py @@ -25,7 +25,7 @@ def mainlist(item): def search(item, texto): - support.log(texto) + support.info(texto) item.contentType = 'tvshow' @@ -36,12 +36,12 @@ def search(item, texto): except: import sys for line in sys.exc_info(): - support.logger.error("%s" % line) + support.infoger.error("%s" % line) return [] def newest(categoria): - support.log(categoria) + support.info(categoria) itemlist = [] item = support.Item() @@ -55,7 +55,7 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.logger.error("{0}".format(line)) + support.infoger.error("{0}".format(line)) return [] return itemlist @@ -134,7 +134,7 @@ def menu(item): def findvideos(item): item.url = item.url.replace('&', '&') - support.log(item) + support.info(item) if item.args == 'last': url = support.match(item, patron = r'<iframe id="iframeVid" width="[^"]+" height="[^"]+" src="([^"]+)" allowfullscreen').match matches = support.match(url,patron=r'<a href="([^"]+)">(\d+)<', patronBlock=r'<h3>EPISODIO</h3><ul>(.*?)</ul>').matches diff --git a/channels/serietvonline.py b/channels/serietvonline.py index 592800d1..993dfbf6 100644 --- a/channels/serietvonline.py +++ b/channels/serietvonline.py @@ -31,7 +31,7 @@ headers = [['Referer', host]] @support.menu def mainlist(item): - support.log() + support.info() film = ['/ultimi-film-aggiunti/', @@ -56,7 +56,7 @@ def mainlist(item): @support.scrape def peliculas(item): - support.log() + support.info() blacklist = ['DMCA', 'Contatti', 'Attenzione NON FARTI OSCURARE', 'Lista Cartoni Animati e Anime'] patronBlock = r'<h1>.+?</h1>(?P<block>.*?)<div class="footer_c">' @@ -105,7 +105,7 @@ def peliculas(item): @support.scrape def episodios(item): - support.log() + support.info() action = 'findvideos' patronBlock = r'<table>(?P<block>.*?)<\/table>' patron = r'<tr><td>(?P<title>.*?)?[ ](?:Parte)?(?P<episode>\d+x\d+|\d+)(?:|[ ]?(?P<title2>.+?)?(?:avi)?)<(?P<url>.*?)</td><tr>' @@ -119,7 +119,7 @@ def episodios(item): def search(item, text): - support.log("CERCA :" ,text, item) + support.info("CERCA :" ,text, item) item.url = "%s/?s=%s" % (host, text) @@ -130,11 +130,11 @@ def search(item, text): except: import sys for line in sys.exc_info(): - support.log("%s" % line) + support.info("%s" % line) return [] def newest(categoria): - support.log(categoria) + support.info(categoria) itemlist = [] item = Item() @@ -153,13 +153,13 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.log("{0}".format(line)) + support.info("{0}".format(line)) return [] return itemlist def findvideos(item): - support.log() + support.info() if item.contentType == 'movie': return support.server(item, headers=headers) else: @@ -173,7 +173,7 @@ def findvideos(item): data = httptools.downloadpage(item.url, headers=headers).data data = re.sub('\n|\t', ' ', data) data = re.sub(r'>\s+<', '> <', data) - #support.log("DATA - HTML:\n", data) + #support.info("DATA - HTML:\n", data) url_video = scrapertools.find_single_match(data, r'<tr><td>(.+?)</td><tr>', -1) url_serie = scrapertools.find_single_match(data, r'<link rel="canonical" href="([^"]+)"\s?/>') goseries = support.typo("Vai alla Serie:", ' bold') diff --git a/channels/serietvsubita.py b/channels/serietvsubita.py index a89d0b0b..ce3f5cab 100644 --- a/channels/serietvsubita.py +++ b/channels/serietvsubita.py @@ -9,7 +9,7 @@ import time from core import httptools, tmdb, scrapertools, support from core.item import Item -from core.support import log +from core.support import info from platformcode import logger, config host = config.get_channel_url() @@ -21,7 +21,7 @@ list_language = IDIOMAS.values() @support.menu def mainlist(item): - log() + info() itemlist = [] tvshowSub = [ ('Novità {bold}',[ '', 'peliculas_tv', '', 'tvshow']), @@ -52,7 +52,7 @@ def cleantitle(scrapedtitle): # ---------------------------------------------------------------------------------------------------------------- def findvideos(item): - log() + info() data = httptools.downloadpage(item.url, headers=headers, ignore_response_code=True).data data = re.sub(r'\n|\t|\s+', ' ', data) # recupero il blocco contenente i link @@ -66,8 +66,8 @@ def findvideos(item): episodio = item.infoLabels['episode'] patron = r'\.\.:: Episodio %s([\s\S]*?)(<div class="post|..:: Episodio)' % episodio - log(patron) - log(blocco) + info(patron) + info(blocco) matches = scrapertools.find_multiple_matches(blocco, patron) if len(matches): @@ -89,7 +89,7 @@ def findvideos(item): # ---------------------------------------------------------------------------------------------------------------- def lista_serie(item): - log() + info() itemlist = [] PERPAGE = 15 @@ -137,7 +137,7 @@ def lista_serie(item): # ---------------------------------------------------------------------------------------------------------------- def episodios(item, itemlist=[]): - log() + info() patron = r'<div class="post-meta">\s*<a href="([^"]+)"\s*title="([^"]+)"\s*class=".*?"></a>.*?' patron += r'<p><a href="([^"]+)">' @@ -212,7 +212,7 @@ def episodios(item, itemlist=[]): # ---------------------------------------------------------------------------------------------------------------- def peliculas_tv(item): - log() + info() itemlist = [] patron = r'<div class="post-meta">\s*<a href="([^"]+)"\s*title="([^"]+)"\s*class=".*?"></a>' @@ -265,7 +265,7 @@ def peliculas_tv(item): # ---------------------------------------------------------------------------------------------------------------- def newest(categoria): - log(categoria) + info(categoria) itemlist = [] item = Item() item.url = host @@ -289,7 +289,7 @@ def newest(categoria): # ---------------------------------------------------------------------------------------------------------------- def search(item, texto): - log(texto) + info(texto) itemlist = [] patron = r'<li class="cat-item cat-item-\d+"><a href="([^"]+)"\s?>([^<]+)</a>' @@ -320,7 +320,7 @@ def search(item, texto): def list_az(item): - log() + info() itemlist = [] alphabet = dict() diff --git a/channels/serietvu.py b/channels/serietvu.py index c1fec70e..2f5834d2 100644 --- a/channels/serietvu.py +++ b/channels/serietvu.py @@ -10,7 +10,7 @@ import re from core import support, httptools, scrapertools from core.item import Item -from core.support import log +from core.support import info from platformcode import config host = config.get_channel_url() @@ -54,8 +54,8 @@ def episodios(item): patron = r'(?:<div class="list (?:active)?")?\s*<a data-id="\d+(?:[ ](?P<lang>[SuUbBiItTaA\-]+))?"(?P<other>[^>]+)>.*?Episodio [0-9]+\s?(?:<br>(?P<title>[^<]+))?.*?Stagione (?P<season>[0-9]+) , Episodio - (?P<episode>[0-9]+).*?<(?P<url>.*?<iframe)' def itemHook(item): for value, season in seasons: - log(value) - log(season) + info(value) + info(season) item.title = item.title.replace(value+'x',season+'x') item.url += '\n' + item.other return item @@ -72,7 +72,7 @@ def genres(item): def search(item, text): - log(text) + info(text) item.url = host + "/?s=" + text try: item.contentType = 'tvshow' @@ -81,12 +81,12 @@ def search(item, text): except: import sys for line in sys.exc_info(): - log("%s" % line) + info("%s" % line) return [] def newest(categoria): - log(categoria) + info(categoria) itemlist = [] item = Item() try: @@ -101,14 +101,14 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - log("{0}".format(line)) + info("{0}".format(line)) return [] return itemlist def findvideos(item): - log(item) + info(item) if item.args != 'update': return support.server(item, data=item.url) else: diff --git a/channels/streamingaltadefinizione.py b/channels/streamingaltadefinizione.py index 78bca72c..65fae73f 100644 --- a/channels/streamingaltadefinizione.py +++ b/channels/streamingaltadefinizione.py @@ -32,14 +32,14 @@ def mainlist(item): def search(item, text): - support.log("[streamingaltadefinizione.py] " + item.url + " search " + text) + support.info("[streamingaltadefinizione.py] " + item.url + " search " + text) item.url = item.url + "/?s=" + text try: return support.dooplay_search(item) except: import sys for line in sys.exc_info(): - support.logger.error("%s" % line) + support.infoger.error("%s" % line) return [] diff --git a/channels/streamingcommunity.py b/channels/streamingcommunity.py index 5eafc07d..60f3d180 100644 --- a/channels/streamingcommunity.py +++ b/channels/streamingcommunity.py @@ -37,7 +37,7 @@ def mainlist(item): def genres(item): - support.log() + support.info() itemlist = [] data = support.scrapertools.decodeHtmlentities(support.match(item).data) args = support.match(data, patronBlock=r'genre-options-json="([^\]]+)\]', patron=r'name"\s*:\s*"([^"]+)').matches @@ -48,7 +48,7 @@ def genres(item): def search(item, text): - support.log('search', item) + support.info('search', item) item.search = text try: @@ -57,12 +57,12 @@ def search(item, text): except: import sys for line in sys.exc_info(): - support.log('search log:', line) + support.info('search log:', line) return [] def newest(category): - support.log(category) + support.info(category) itemlist = [] item = support.Item() item.args = 1 @@ -80,7 +80,7 @@ def newest(category): except: import sys for line in sys.exc_info(): - support.log(line) + support.info(line) return [] return itemlist @@ -88,7 +88,7 @@ def newest(category): def peliculas(item): - support.log() + support.info() itemlist = [] videoType = 'movie' if item.contentType == 'movie' else 'tv' @@ -147,15 +147,15 @@ def peliculas(item): return itemlist def episodios(item): - support.log() + support.info() itemlist = [] js = json.loads(support.match(item.url, patron=r'seasons="([^"]+)').match.replace('"','"')) - support.log(js) + support.info(js) for episodes in js: for it in episodes['episodes']: - support.log(it) + support.info(it) itemlist.append( support.Item(channel=item.channel, title=support.typo(str(episodes['number']) + 'x' + str(it['number']).zfill(2) + ' - ' + it['name'], 'bold'), @@ -174,7 +174,7 @@ def episodios(item): def findvideos(item): - support.log() + support.info() itemlist=[] url = support.match(support.match(item).data.replace('"','"').replace('\\',''), patron=r'video_url"\s*:\s*"([^"]+)"').match playlist = support.match(url.replace('https','http'), patron=r'\./([^.]+)').matches diff --git a/channels/tantifilm.py b/channels/tantifilm.py index 4deb01ec..777e7676 100644 --- a/channels/tantifilm.py +++ b/channels/tantifilm.py @@ -7,7 +7,7 @@ import re from core import scrapertools, httptools, support from core.item import Item -from core.support import log +from core.support import info from platformcode import logger from platformcode import config, unify @@ -24,7 +24,7 @@ player_iframe = r'<iframe src="([^"]+)"[^>]+></iframe>\s?<div class="player' @support.menu def mainlist(item): - log() + info() top = [('Generi', ['', 'category'])] film = ['/film', @@ -65,7 +65,7 @@ def peliculas(item): @support.scrape def episodios(item): - log() + info() if not item.data: data_check = httptools.downloadpage(item.url, headers=headers).data data_check = re.sub('\n|\t', ' ', data_check) @@ -125,7 +125,7 @@ def category(item): def search(item, texto): - log(texto) + info(texto) item.url = host + "/?s=" + texto @@ -166,17 +166,17 @@ def hdpass(item): def findvideos(item): - log() - support.log("ITEMLIST: ", item) + info() + support.info("ITEMLIST: ", item) data = support.match(item.url, headers=headers).data check = support.match(data, patron=r'<div class="category-film">(.*?)</div>').match if 'sub' in check.lower(): item.contentLanguage = 'Sub-ITA' - support.log("CHECK : ", check) + support.info("CHECK : ", check) if 'anime' in check.lower(): item.contentType = 'tvshow' item.data = data - support.log('select = ### è una anime ###') + support.info('select = ### è una anime ###') try: return episodios(item) except: @@ -188,7 +188,7 @@ def findvideos(item): # if 'protectlink' in data: # urls = scrapertools.find_multiple_matches(data, r'<iframe src="[^=]+=(.*?)"') - # support.log("SONO QUI: ", urls) + # support.info("SONO QUI: ", urls) # for url in urls: # url = url.decode('base64') # # tiro via l'ultimo carattere perchè non c'entra @@ -199,7 +199,7 @@ def findvideos(item): # if url: # listurl.add(url) # data += '\n'.join(listurl) - log(data) + info(data) itemlist = [] # support.dbg() @@ -211,7 +211,7 @@ def findvideos(item): if item.otherLinks: urls += support.match(item.otherLinks, patron=r'href="([^"]+)').matches - log('URLS', urls) + info('URLS', urls) for u in urls: if 'hdplayer.casa/series/' in u: urls.remove(u) diff --git a/channels/toonitalia.py b/channels/toonitalia.py index d2b88667..e2b66b55 100644 --- a/channels/toonitalia.py +++ b/channels/toonitalia.py @@ -27,7 +27,7 @@ def mainlist(item): def search(item, texto): - support.log(texto) + support.info(texto) item.args='search' item.contentType='tvshow' item.url = host + '/?s=' + texto @@ -37,12 +37,12 @@ def search(item, texto): except: import sys for line in sys.exc_info(): - support.logger.error("%s" % line) + support.infoger.error("%s" % line) return [] def newest(categoria): - support.log(categoria) + support.info(categoria) item = support.Item() try: item.contentType = 'tvshow' @@ -53,7 +53,7 @@ def newest(categoria): except: import sys for line in sys.exc_info(): - support.logger.error("{0}".format(line)) + support.infoger.error("{0}".format(line)) return [] @@ -86,7 +86,7 @@ def peliculas(item): patron = r'<li ><a href="(?P<url>[^"]+)" title="[^>]+">(?P<title>[^<|\(]+)?(?:\([^\d]*(?P<year>\d+)\))?[^<]*</a>' def itemHook(item): - support.log(item.title) + support.info(item.title) item.title = support.re.sub(' (?:- )?[Ss]erie [Tt][Vv]', '', item.title) if item.args == 'sub': #corregge l'esatta lang per quelle pagine in cui c'è diff --git a/channels/tunein.py b/channels/tunein.py index c3a9bbbb..3780ea9c 100644 --- a/channels/tunein.py +++ b/channels/tunein.py @@ -34,7 +34,7 @@ def mainlist(item): def radio(item): - support.log() + support.info() itemlist = [] data = support.match(item, patron= r'text="(?P<title>[^\("]+)(?:\((?P<location>[^\)]+)\))?" URL="(?P<url>[^"]+)" bitrate="(?P<quality>[^"]+)" reliability="[^"]+" guide_id="[^"]+" subtext="(?P<song>[^"]+)" genre_id="[^"]+" formats="(?P<type>[^"]+)" (?:playing="[^"]+" )?(?:playing_image="[^"]+" )?(?:show_id="[^"]+" )?(?:item="[^"]+" )?image="(?P<thumb>[^"]+)"') if data.matches: @@ -85,7 +85,7 @@ def findvideos(item): def search(item, text): - support.log(text) + support.info(text) item.url = host + '/Search.ashx?query=' +text try: return radio(item) @@ -93,5 +93,5 @@ def search(item, text): except: import sys for line in sys.exc_info(): - support.logger.error("%s" % line) + support.infoger.error("%s" % line) return [] diff --git a/channels/vvvvid.py b/channels/vvvvid.py index 1f15a81f..e14af738 100644 --- a/channels/vvvvid.py +++ b/channels/vvvvid.py @@ -71,7 +71,7 @@ def mainlist(item): def search(item, text): - support.log(text) + support.info(text) itemlist = [] if conn_id: if 'film' in item.url: item.contentType = 'movie' @@ -82,7 +82,7 @@ def search(item, text): except: import sys for line in sys.exc_info(): - support.logger.error("%s" % line) + support.infoger.error("%s" % line) return [] return itemlist @@ -106,7 +106,7 @@ def peliculas(item): itemlist = [] if not item.args: json_file =loadjs(item.url + 'channel/10005/last/') - support.log(json_file) + support.logger.debug(json_file) make_itemlist(itemlist, item, json_file) elif ('=' not in item.args) and ('=' not in item.url): @@ -142,7 +142,7 @@ def episodios(item): show_id = str(json_file['data'][0]['show_id']) season_id = str(json_file['data'][0]['season_id']) episodes = [] - support.log('SEASON ID= ',season_id) + support.info('SEASON ID= ',season_id) for episode in json_file['data']: episodes.append(episode['episodes']) for episode in episodes: @@ -233,7 +233,7 @@ def make_itemlist(itemlist, item, data): def loadjs(url): if '?category' not in url: url += '?full=true' - support.log('Json URL;',url) + support.info('Json URL;',url) json = current_session.get(url, headers=headers, params=payload).json() return json diff --git a/channelselector.py b/channelselector.py index 42b4db51..5c57098e 100644 --- a/channelselector.py +++ b/channelselector.py @@ -9,7 +9,7 @@ addon = config.__settings__ downloadenabled = addon.getSetting('downloadenabled') def getmainlist(view="thumb_"): - logger.log() + logger.info() itemlist = list() if config.dev_mode(): @@ -62,14 +62,14 @@ def getmainlist(view="thumb_"): def getchanneltypes(view="thumb_"): - logger.log() + logger.info() # Category List channel_types = ["movie", "tvshow", "anime", "documentary", "vos", "live", "torrent", "music"] #, "direct" # Channel Language channel_language = auto_filter() - logger.log("channel_language=%s" % channel_language) + logger.info("channel_language=%s" % channel_language) # Build Itemlist itemlist = list() @@ -92,7 +92,7 @@ def getchanneltypes(view="thumb_"): def filterchannels(category, view="thumb_"): from core import channeltools - logger.log('Filter Channels ' + category) + logger.info('Filter Channels ' + category) channelslist = [] @@ -103,17 +103,17 @@ def filterchannels(category, view="thumb_"): appenddisabledchannels = True channel_path = os.path.join(config.get_runtime_path(), 'channels', '*.json') - logger.log("channel_path = %s" % channel_path) + logger.info("channel_path = %s" % channel_path) channel_files = glob.glob(channel_path) - logger.log("channel_files found %s" % (len(channel_files))) + logger.info("channel_files found %s" % (len(channel_files))) # Channel Language channel_language = auto_filter() - logger.log("channel_language=%s" % channel_language) + logger.info("channel_language=%s" % channel_language) for channel_path in channel_files: - logger.log("channel in for = %s" % channel_path) + logger.debug("channel in for = %s" % channel_path) channel = os.path.basename(channel_path).replace(".json", "") @@ -126,7 +126,7 @@ def filterchannels(category, view="thumb_"): # If it's not a channel we skip it if not channel_parameters["channel"]: continue - logger.log("channel_parameters=%s" % repr(channel_parameters)) + logger.debug("channel_parameters=%s" % repr(channel_parameters)) # If you prefer the banner and the channel has it, now change your mind if view == "banner_" and "banner" in channel_parameters: @@ -221,7 +221,7 @@ def get_thumb(thumb_name, view="thumb_"): def set_channel_info(parameters): - logger.log() + logger.info() info = '' language = '' diff --git a/core/channeltools.py b/core/channeltools.py index 528acdfb..90a2dd74 100644 --- a/core/channeltools.py +++ b/core/channeltools.py @@ -15,7 +15,7 @@ default_file = dict() remote_path = 'https://raw.githubusercontent.com/kodiondemand/media/master/' def is_enabled(channel_name): - logger.log("channel_name=" + channel_name) + logger.info("channel_name=" + channel_name) return get_channel_parameters(channel_name)["active"] and get_channel_setting("enabled", channel=channel_name, default=True) @@ -87,7 +87,7 @@ def get_channel_parameters(channel_name): def get_channel_json(channel_name): - # logger.log("channel_name=" + channel_name) + # logger.info("channel_name=" + channel_name) from core import filetools channel_json = None try: @@ -101,9 +101,9 @@ def get_channel_json(channel_name): channel_name + ".json") if filetools.isfile(channel_path): - # logger.log("channel_data=" + channel_path) + # logger.info("channel_data=" + channel_path) channel_json = jsontools.load(filetools.read(channel_path)) - # logger.log("channel_json= %s" % channel_json) + # logger.info("channel_json= %s" % channel_json) except Exception as ex: template = "An exception of type %s occured. Arguments:\n%r" @@ -114,7 +114,7 @@ def get_channel_json(channel_name): def get_channel_controls_settings(channel_name): - # logger.log("channel_name=" + channel_name) + # logger.info("channel_name=" + channel_name) dict_settings = {} # import web_pdb; web_pdb.set_trace() # list_controls = get_channel_json(channel_name).get('settings', list()) @@ -137,7 +137,7 @@ def get_lang(channel_name): if hasattr(channel, 'list_language'): for language in channel.list_language: list_language.append(language) - logger.log(list_language) + logger.info(list_language) else: sub = False langs = [] diff --git a/core/downloader.py b/core/downloader.py index c08e8fdf..b714b474 100644 --- a/core/downloader.py +++ b/core/downloader.py @@ -253,12 +253,12 @@ class Downloader(object): self.file.seek(2 ** 31, 0) except OverflowError: self._seekable = False - logger.log("Cannot do seek() or tell() in files larger than 2GB") + logger.info("Cannot do seek() or tell() in files larger than 2GB") self.__get_download_info__() try: - logger.log("Download started: Parts: %s | Path: %s | File: %s | Size: %s" % (str(len(self._download_info["parts"])), self._pathencode('utf-8'), self._filenameencode('utf-8'), str(self._download_info["size"]))) + logger.info("Download started: Parts: %s | Path: %s | File: %s | Size: %s" % (str(len(self._download_info["parts"])), self._pathencode('utf-8'), self._filenameencode('utf-8'), str(self._download_info["size"]))) except: pass @@ -410,7 +410,7 @@ class Downloader(object): return id == 0 or (len(self.completed_parts) >= id and sorted(self.completed_parts)[id - 1] == id - 1) def __save_file__(self): - logger.log("Thread started: %s" % threading.current_thread().name) + logger.info("Thread started: %s" % threading.current_thread().name) while self._state == self.states.downloading: if not self.pending_parts and not self.download_parts and not self.save_parts: # Download finished @@ -449,7 +449,7 @@ class Downloader(object): self._download_info["parts"][s]["status"] = self.states.stopped self._download_info["parts"][s]["current"] = self._download_info["parts"][s]["start"] - logger.log("Thread stopped: %s" % threading.current_thread().name) + logger.info("Thread stopped: %s" % threading.current_thread().name) def __get_part_id__(self): self._download_lock.acquire() @@ -464,21 +464,21 @@ class Downloader(object): return None def __set_part_connecting__(self, id): - logger.log("ID: %s Establishing connection" % id) + logger.info("ID: %s Establishing connection" % id) self._download_info["parts"][id]["status"] = self.states.connecting def __set_part__error__(self, id): - logger.log("ID: %s Download failed" % id) + logger.info("ID: %s Download failed" % id) self._download_info["parts"][id]["status"] = self.states.error self.pending_parts.add(id) self.download_parts.remove(id) def __set_part__downloading__(self, id): - logger.log("ID: %s Downloading data ..." % id) + logger.info("ID: %s Downloading data ..." % id) self._download_info["parts"][id]["status"] = self.states.downloading def __set_part_completed__(self, id): - logger.log("ID: %s Download finished!" % id) + logger.info("ID: %s Download finished!" % id) self._download_info["parts"][id]["status"] = self.states.saving self.download_parts.remove(id) self.save_parts.add(id) @@ -501,7 +501,7 @@ class Downloader(object): return file def __start_part__(self): - logger.log("Thread Started: %s" % threading.current_thread().name) + logger.info("Thread Started: %s" % threading.current_thread().name) while self._state == self.states.downloading: id = self.__get_part_id__() if id is None: break @@ -528,7 +528,7 @@ class Downloader(object): buffer = connection.read(self._block_size) speed.append(old_div(len(buffer), ((time.time() - start) or 0.001))) except: - logger.log("ID: %s Error downloading data" % id) + logger.info("ID: %s Error downloading data" % id) self._download_info["parts"][id]["status"] = self.states.error self.pending_parts.add(id) self.download_parts.remove(id) @@ -546,7 +546,7 @@ class Downloader(object): if velocidad_minima > speed[-1] and velocidad_minima > speed[-2] and self._download_info["parts"][id]["current"] < self._download_info["parts"][id]["end"]: if connection.fp: connection.fp._sock.close() - logger.log("ID: %s Restarting connection! | Minimum Speed: %.2f %s/s | Speed: %.2f %s/s" % (id, vm[1], vm[2], v[1], v[2])) + logger.info("ID: %s Restarting connection! | Minimum Speed: %.2f %s/s | Speed: %.2f %s/s" % (id, vm[1], vm[2], v[1], v[2])) # file.close() break else: @@ -556,7 +556,7 @@ class Downloader(object): break self.__set_part_stopped__(id) - logger.log("Thread stopped: %s" % threading.current_thread().name) + logger.info("Thread stopped: %s" % threading.current_thread().name) def __update_json(self, started=True): text = filetools.read(self._json_path) @@ -564,10 +564,10 @@ class Downloader(object): if self._json_text != text: self._json_text = text self._json_item = Item().fromjson(text) - logger.log('item loaded') + logger.info('item loaded') progress = int(self.progress) if started and self._json_item.downloadStatus == 0: # stopped - logger.log('Download paused') + logger.info('Download paused') self.stop() elif self._json_item.downloadProgress != progress or not started: params = {"downloadStatus": 4, "downloadComplete": 0, "downloadProgress": progress} diff --git a/core/downloadtools.py b/core/downloadtools.py index b8328145..f0a7d4cf 100644 --- a/core/downloadtools.py +++ b/core/downloadtools.py @@ -97,11 +97,11 @@ def limpia_nombre_excepto_1(s): try: s = unicode(s, "utf-8") except UnicodeError: - # logger.log("no es utf-8") + # logger.info("no es utf-8") try: s = unicode(s, "iso-8859-1") except UnicodeError: - # logger.log("no es iso-8859-1") + # logger.info("no es iso-8859-1") pass # Remove accents s = limpia_nombre_sin_acentos(s) @@ -125,29 +125,29 @@ def limpia_nombre_excepto_2(s): def getfilefromtitle(url, title): # Print in the log what you will discard - logger.log("title=" + title) - logger.log("url=" + url) + logger.info("title=" + title) + logger.info("url=" + url) plataforma = config.get_system_platform() - logger.log("platform=" + plataforma) + logger.info("platform=" + plataforma) # filename = xbmc.makeLegalFilename(title + url[-4:]) from core import scrapertools nombrefichero = title + scrapertools.get_filename_from_url(url)[-4:] - logger.log("filename= %s" % nombrefichero) + logger.info("filename= %s" % nombrefichero) if "videobb" in url or "videozer" in url or "putlocker" in url: nombrefichero = title + ".flv" if "videobam" in url: nombrefichero = title + "." + url.rsplit(".", 1)[1][0:3] - logger.log("filename= %s" % nombrefichero) + logger.info("filename= %s" % nombrefichero) nombrefichero = limpia_nombre_caracteres_especiales(nombrefichero) - logger.log("filename= %s" % nombrefichero) + logger.info("filename= %s" % nombrefichero) fullpath = filetools.join(config.get_setting("downloadpath"), nombrefichero) - logger.log("fullpath= %s" % fullpath) + logger.info("fullpath= %s" % fullpath) if config.is_xbmc() and fullpath.startswith("special://"): import xbmc @@ -162,7 +162,7 @@ def downloadtitle(url, title): def downloadbest(video_urls, title, continuar=False): - logger.log() + logger.info() # Flip it over, to put the highest quality one first (list () is for you to make a copy of) invertida = list(video_urls) @@ -172,9 +172,9 @@ def downloadbest(video_urls, title, continuar=False): # videotitle = elemento[0] url = elemento[1] if not PY3: - logger.log("Downloading option " + title + " " + url.encode('ascii', 'ignore')) + logger.info("Downloading option " + title + " " + url.encode('ascii', 'ignore')) else: - logger.log("Downloading option " + title + " " + url.encode('ascii', 'ignore').decode('utf-8')) + logger.info("Downloading option " + title + " " + url.encode('ascii', 'ignore').decode('utf-8')) # Calculate the file where you should record try: @@ -200,25 +200,25 @@ def downloadbest(video_urls, title, continuar=False): else: # EThe file doesn't even exist if not filetools.exists(fullpath): - logger.log("-> You have not downloaded anything, testing with the following option if there is") + logger.info("-> You have not downloaded anything, testing with the following option if there is") # The file exists else: tamanyo = filetools.getsize(fullpath) # It has size 0 if tamanyo == 0: - logger.log("-> Download a file with size 0, testing with the following option if it exists") + logger.info("-> Download a file with size 0, testing with the following option if it exists") os.remove(fullpath) else: - logger.log("-> Download a file with size %d, he takes it for good" % tamanyo) + logger.info("-> Download a file with size %d, he takes it for good" % tamanyo) return 0 return -2 def downloadfile(url, nombrefichero, headers=None, silent=False, continuar=False, resumir=True, header=''): - logger.log("url= " + url) - logger.log("filename= " + nombrefichero) + logger.info("url= " + url) + logger.info("filename= " + nombrefichero) if headers is None: headers = [] @@ -242,14 +242,14 @@ def downloadfile(url, nombrefichero, headers=None, silent=False, continuar=False nombrefichero = xbmc.makeLegalFilename(nombrefichero) except: pass - logger.log("filename= " + nombrefichero) + logger.info("filename= " + nombrefichero) # The file exists and you want to continue if filetools.exists(nombrefichero) and continuar: f = filetools.file_open(nombrefichero, 'r+b', vfs=VFS) if resumir: exist_size = filetools.getsize(nombrefichero) - logger.log("the file exists, size= %d" % exist_size) + logger.info("the file exists, size= %d" % exist_size) grabado = exist_size f.seek(exist_size) else: @@ -258,13 +258,13 @@ def downloadfile(url, nombrefichero, headers=None, silent=False, continuar=False # the file already exists and you don't want to continue, it aborts elif filetools.exists(nombrefichero) and not continuar: - logger.log("the file exists, it does not download again") + logger.info("the file exists, it does not download again") return -3 # the file does not exist else: exist_size = 0 - logger.log("the file does not exist") + logger.info("the file does not exist") f = filetools.file_open(nombrefichero, 'wb', vfs=VFS) grabado = 0 @@ -285,13 +285,13 @@ def downloadfile(url, nombrefichero, headers=None, silent=False, continuar=False additional_headers = [additional_headers] for additional_header in additional_headers: - logger.log("additional_header: " + additional_header) + logger.info("additional_header: " + additional_header) name = re.findall("(.*?)=.*?", additional_header)[0] value = urllib.parse.unquote_plus(re.findall(".*?=(.*?)$", additional_header)[0]) headers.append([name, value]) url = url.split("|")[0] - logger.log("url=" + url) + logger.info("url=" + url) # Socket timeout at 60 seconds socket.setdefaulttimeout(60) @@ -299,7 +299,7 @@ def downloadfile(url, nombrefichero, headers=None, silent=False, continuar=False h = urllib.request.HTTPHandler(debuglevel=0) request = urllib.request.Request(url) for header in headers: - logger.log("Header= " + header[0] + ": " + header[1]) + logger.info("Header= " + header[0] + ": " + header[1]) request.add_header(header[0], header[1]) if exist_size > 0: @@ -328,12 +328,12 @@ def downloadfile(url, nombrefichero, headers=None, silent=False, continuar=False if exist_size > 0: totalfichero = totalfichero + exist_size - logger.log("Content-Length= %s" % totalfichero) + logger.info("Content-Length= %s" % totalfichero) blocksize = 100 * 1024 bloqueleido = connexion.read(blocksize) - logger.log("Starting downloading the file, blocked= %s" % len(bloqueleido)) + logger.info("Starting downloading the file, blocked= %s" % len(bloqueleido)) maxreintentos = 10 @@ -360,7 +360,7 @@ def downloadfile(url, nombrefichero, headers=None, silent=False, continuar=False tiempofalta = old_div(falta, velocidad) else: tiempofalta = 0 - # logger.log(sec_to_hms(tiempofalta)) + # logger.info(sec_to_hms(tiempofalta)) if not silent: progreso.update(percent, "%.2fMB/%.2fMB (%d%%) %.2f Kb/s %s" % (descargadosmb, totalmb, percent, old_div(velocidad, 1024), @@ -368,14 +368,14 @@ def downloadfile(url, nombrefichero, headers=None, silent=False, continuar=False break except: reintentos += 1 - logger.log("ERROR in block download, retry %d" % reintentos) + logger.info("ERROR in block download, retry %d" % reintentos) import traceback logger.error(traceback.print_exc()) # The user cancels the download try: if progreso.iscanceled(): - logger.log("Download of file canceled") + logger.info("Download of file canceled") f.close() progreso.close() return -1 @@ -384,7 +384,7 @@ def downloadfile(url, nombrefichero, headers=None, silent=False, continuar=False # There was an error in the download if reintentos > maxreintentos: - logger.log("ERROR in the file download") + logger.info("ERROR in the file download") f.close() if not silent: progreso.close() @@ -430,7 +430,7 @@ def downloadfile(url, nombrefichero, headers=None, silent=False, continuar=False except: pass - logger.log("End of file download") + logger.info("End of file download") def downloadfileRTMP(url, nombrefichero, silent): @@ -476,7 +476,7 @@ def downloadfileRTMP(url, nombrefichero, silent): try: rtmpdump_args = [rtmpdump_cmd] + rtmpdump_args + ["-o", nombrefichero] from os import spawnv, P_NOWAIT - logger.log("Initiating file download: %s" % " ".join(rtmpdump_args)) + logger.info("Initiating file download: %s" % " ".join(rtmpdump_args)) rtmpdump_exit = spawnv(P_NOWAIT, rtmpdump_cmd, rtmpdump_args) if not silent: from platformcode import platformtools @@ -488,18 +488,18 @@ def downloadfileRTMP(url, nombrefichero, silent): def downloadfileGzipped(url, pathfichero): - logger.log("url= " + url) + logger.info("url= " + url) nombrefichero = pathfichero - logger.log("filename= " + nombrefichero) + logger.info("filename= " + nombrefichero) import xbmc nombrefichero = xbmc.makeLegalFilename(nombrefichero) - logger.log("filename= " + nombrefichero) + logger.info("filename= " + nombrefichero) patron = "(http://[^/]+)/.+" matches = re.compile(patron, re.DOTALL).findall(url) if len(matches): - logger.log("Main URL: " + matches[0]) + logger.info("Main URL: " + matches[0]) url1 = matches[0] else: url1 = url @@ -546,9 +546,9 @@ def downloadfileGzipped(url, pathfichero): nombre_fichero_base = filetools.basename(nombrefichero) if len(nombre_fichero_base) == 0: - logger.log("Searching for name in the answer Headers") + logger.info("Searching for name in the answer Headers") nombre_base = connexion.headers["Content-Disposition"] - logger.log(nombre_base) + logger.info(nombre_base) patron = 'filename="([^"]+)"' matches = re.compile(patron, re.DOTALL).findall(nombre_base) if len(matches) > 0: @@ -556,7 +556,7 @@ def downloadfileGzipped(url, pathfichero): titulo = GetTitleFromFile(titulo) nombrefichero = filetools.join(pathfichero, titulo) else: - logger.log("Name of the file not found, Placing temporary name: no_name.txt") + logger.info("Name of the file not found, Placing temporary name: no_name.txt") titulo = "no_name.txt" nombrefichero = filetools.join(pathfichero, titulo) totalfichero = int(connexion.headers["Content-Length"]) @@ -564,10 +564,10 @@ def downloadfileGzipped(url, pathfichero): # then f = filetools.file_open(nombrefichero, 'w', vfs=VFS) - logger.log("new file open") + logger.info("new file open") grabado = 0 - logger.log("Content-Length= %s" % totalfichero) + logger.info("Content-Length= %s" % totalfichero) blocksize = 100 * 1024 @@ -580,7 +580,7 @@ def downloadfileGzipped(url, pathfichero): gzipper = gzip.GzipFile(fileobj=compressedstream) bloquedata = gzipper.read() gzipper.close() - logger.log("Starting downloading the file, blocked= %s" % len(bloqueleido)) + logger.info("Starting downloading the file, blocked= %s" % len(bloqueleido)) except: logger.error("ERROR: The file to be downloaded is not compressed with Gzip") f.close() @@ -619,32 +619,32 @@ def downloadfileGzipped(url, pathfichero): tiempofalta = old_div(falta, velocidad) else: tiempofalta = 0 - logger.log(sec_to_hms(tiempofalta)) + logger.info(sec_to_hms(tiempofalta)) progreso.update(percent, "%.2fMB/%.2fMB (%d%%) %.2f Kb/s %s left " % (descargadosmb, totalmb, percent, old_div(velocidad, 1024), sec_to_hms(tiempofalta))) break except: reintentos += 1 - logger.log("ERROR in block download, retry %d" % reintentos) + logger.info("ERROR in block download, retry %d" % reintentos) for line in sys.exc_info(): logger.error("%s" % line) # The user cancels the download if progreso.iscanceled(): - logger.log("Download of file canceled") + logger.info("Download of file canceled") f.close() progreso.close() return -1 # There was an error in the download if reintentos > maxreintentos: - logger.log("ERROR in the file download") + logger.info("ERROR in the file download") f.close() progreso.close() return -2 except: - logger.log("ERROR in the file download") + logger.info("ERROR in the file download") for line in sys.exc_info(): logger.error("%s" % line) f.close() @@ -655,15 +655,15 @@ def downloadfileGzipped(url, pathfichero): # print data progreso.close() - logger.log("End download of the file") + logger.info("End download of the file") return nombrefichero def GetTitleFromFile(title): # Print in the log what you will discard - logger.log("title= " + title) + logger.info("title= " + title) plataforma = config.get_system_platform() - logger.log("plataform= " + plataforma) + logger.info("plataform= " + plataforma) # nombrefichero = xbmc.makeLegalFilename(title + url[-4:]) nombrefichero = title @@ -677,11 +677,11 @@ def sec_to_hms(seconds): def downloadIfNotModifiedSince(url, timestamp): - logger.log("(" + url + "," + time.ctime(timestamp) + ")") + logger.info("(" + url + "," + time.ctime(timestamp) + ")") # Convert date to GMT fecha_formateada = time.strftime("%a, %d %b %Y %H:%M:%S +0000", time.gmtime(timestamp)) - logger.log("Formatted date= %s" % fecha_formateada) + logger.info("Formatted date= %s" % fecha_formateada) # Check if it has changed inicio = time.clock() @@ -702,9 +702,9 @@ def downloadIfNotModifiedSince(url, timestamp): except urllib.error.URLError as e: # If it returns 304 it is that it has not changed if hasattr(e, 'code'): - logger.log("HTTP response code : %d" % e.code) + logger.info("HTTP response code : %d" % e.code) if e.code == 304: - logger.log("It has not changed") + logger.info("It has not changed") updated = False # Grab errors with response code from requested external server else: @@ -713,13 +713,13 @@ def downloadIfNotModifiedSince(url, timestamp): data = "" fin = time.clock() - logger.log("Downloaded in %d seconds " % (fin - inicio + 1)) + logger.info("Downloaded in %d seconds " % (fin - inicio + 1)) return updated, data def download_all_episodes(item, channel, first_episode="", preferred_server="vidspot", filter_language=""): - logger.log("show= " + item.show) + logger.info("show= " + item.show) show_title = item.show # Gets the listing from which it was called @@ -749,9 +749,9 @@ def download_all_episodes(item, channel, first_episode="", preferred_server="vid for episode_item in episode_itemlist: try: - logger.log("episode= " + episode_item.title) + logger.info("episode= " + episode_item.title) episode_title = scrapertools.find_single_match(episode_item.title, r"(\d+x\d+)") - logger.log("episode= " + episode_title) + logger.info("episode= " + episode_title) except: import traceback logger.error(traceback.format_exc()) @@ -815,7 +815,7 @@ def download_all_episodes(item, channel, first_episode="", preferred_server="vid new_mirror_itemlist_4 + new_mirror_itemlist_5 + new_mirror_itemlist_6) for mirror_item in mirrors_itemlist: - logger.log("mirror= " + mirror_item.title) + logger.info("mirror= " + mirror_item.title) if "(Italiano)" in mirror_item.title: idioma = "(Italiano)" @@ -836,11 +836,11 @@ def download_all_episodes(item, channel, first_episode="", preferred_server="vid idioma = "(Desconocido)" codigo_idioma = "desconocido" - logger.log("filter_language=#" + filter_language + "#, codigo_idioma=#" + codigo_idioma + "#") + logger.info("filter_language=#" + filter_language + "#, codigo_idioma=#" + codigo_idioma + "#") if filter_language == "" or (filter_language != "" and filter_language == codigo_idioma): - logger.log("downloading mirror") + logger.info("downloading mirror") else: - logger.log("language " + codigo_idioma + " filtered, skipping") + logger.info("language " + codigo_idioma + " filtered, skipping") continue if hasattr(channel, 'play'): @@ -856,14 +856,14 @@ def download_all_episodes(item, channel, first_episode="", preferred_server="vid # Adds it to the download list if puedes: - logger.log("downloading mirror started...") + logger.info("downloading mirror started...") # The highest quality video is the latest # mediaurl = video_urls[len(video_urls) - 1][1] devuelve = downloadbest(video_urls, show_title + " " + episode_title + " " + idioma + " [" + video_item.server + "]", continuar=False) if devuelve == 0: - logger.log("download ok") + logger.info("download ok") descargado = True break elif devuelve == -1: @@ -874,14 +874,14 @@ def download_all_episodes(item, channel, first_episode="", preferred_server="vid pass return else: - logger.log("download error, try another mirror") + logger.info("download error, try another mirror") continue else: - logger.log("downloading mirror not available... trying next") + logger.info("downloading mirror not available... trying next") if not descargado: - logger.log("UNDOWNLOADED EPISODE " + episode_title) + logger.info("UNDOWNLOADED EPISODE " + episode_title) def episodio_ya_descargado(show_title, episode_title): @@ -889,9 +889,9 @@ def episodio_ya_descargado(show_title, episode_title): ficheros = filetools.listdir(".") for fichero in ficheros: - # logger.log("fichero="+fichero) + # logger.info("fichero="+fichero) if fichero.lower().startswith(show_title.lower()) and scrapertools.find_single_match(fichero, "(\d+x\d+)") == episode_title: - logger.log("found!") + logger.info("found!") return True return False diff --git a/core/filetools.py b/core/filetools.py index 7719d6c7..554a2671 100644 --- a/core/filetools.py +++ b/core/filetools.py @@ -814,7 +814,7 @@ def remove_tags(title): @rtype: str @return: string without tags """ - logger.log() + logger.info() title_without_tags = scrapertools.find_single_match(title, r'\[color .+?\](.+)\[\/color\]') @@ -832,7 +832,7 @@ def remove_smb_credential(path): @return: chain without credentials @rtype: str """ - logger.log() + logger.info() if not scrapertools.find_single_match(path, r'(^\w+:\/\/)'): return path diff --git a/core/httptools.py b/core/httptools.py index d4a2b05d..47d2a9e5 100755 --- a/core/httptools.py +++ b/core/httptools.py @@ -125,11 +125,11 @@ def set_cookies(dict_cookie, clear=True, alfa_s=False): def load_cookies(alfa_s=False): cookies_lock.acquire() if os.path.isfile(cookies_file): - if not alfa_s: logger.log("Reading cookies file") + if not alfa_s: logger.info("Reading cookies file") try: cj.load(cookies_file, ignore_discard=True) except: - if not alfa_s: logger.log("The cookie file exists but is illegible, it is deleted") + if not alfa_s: logger.info("The cookie file exists but is illegible, it is deleted") os.remove(cookies_file) cookies_lock.release() @@ -137,7 +137,7 @@ load_cookies() def save_cookies(alfa_s=False): cookies_lock.acquire() - if not alfa_s: logger.log("Saving cookies...") + if not alfa_s: logger.debug("Saving cookies...") cj.save(cookies_file, ignore_discard=True) cookies_lock.release() @@ -161,7 +161,7 @@ def random_useragent(): def show_infobox(info_dict): - logger.log() + logger.debug() from textwrap import wrap box_items_kodi = {'r_up_corner': u'\u250c', @@ -193,9 +193,9 @@ def show_infobox(info_dict): else: box = box_items - logger.log('%s%s%s' % (box['r_up_corner'], box['fill'] * width, box['l_up_corner'])) - logger.log('%s%s%s' % (box['center'], version.center(width), box['center'])) - logger.log('%s%s%s' % (box['r_center'], box['fill'] * width, box['l_center'])) + logger.debug('%s%s%s' % (box['r_up_corner'], box['fill'] * width, box['l_up_corner'])) + logger.debug('%s%s%s' % (box['center'], version.center(width), box['center'])) + logger.debug('%s%s%s' % (box['r_center'], box['fill'] * width, box['l_center'])) count = 0 for key, value in info_dict: @@ -210,19 +210,19 @@ def show_infobox(info_dict): for line in text: if len(line) < width: line = line.ljust(width, ' ') - logger.log('%s%s%s' % (box['center'], line, box['center'])) + logger.debug('%s%s%s' % (box['center'], line, box['center'])) else: - logger.log('%s%s%s' % (box['center'], text, box['center'])) + logger.debug('%s%s%s' % (box['center'], text, box['center'])) if count < len(info_dict): - logger.log('%s%s%s' % (box['r_center'], box['fill'] * width, box['l_center'])) + logger.debug('%s%s%s' % (box['r_center'], box['fill'] * width, box['l_center'])) else: - logger.log('%s%s%s' % (box['r_dn_corner'], box['fill'] * width, box['l_dn_corner'])) + logger.debug('%s%s%s' % (box['r_dn_corner'], box['fill'] * width, box['l_dn_corner'])) return def downloadpage(url, **opt): - # logger.log() + # logger.info() """ Open a url and return the data obtained diff --git a/core/item.py b/core/item.py index b95e8de3..c8411533 100644 --- a/core/item.py +++ b/core/item.py @@ -298,7 +298,7 @@ class Item(object): def tostring(self, separator=", "): """ Generate a text string with the item's data for the log - Use: logger.log(item.tostring()) + Use: logger.info(item.tostring()) @param separator: string to be used as a separator @type separator: str '""" diff --git a/core/jsontools.py b/core/jsontools.py index d21215e3..b2fed3a6 100644 --- a/core/jsontools.py +++ b/core/jsontools.py @@ -11,24 +11,24 @@ from inspect import stack try: import json except: - logger.log("json included in the interpreter **NOT** available") + logger.info("json included in the interpreter **NOT** available") try: import simplejson as json except: - logger.log("simplejson included in the interpreter **NOT** available") + logger.info("simplejson included in the interpreter **NOT** available") try: from lib import simplejson as json except: - logger.log("simplejson in lib directory **NOT** available") + logger.info("simplejson in lib directory **NOT** available") logger.error("A valid JSON parser was not found") json = None else: - logger.log("Using simplejson in the lib directory") + logger.info("Using simplejson in the lib directory") else: - logger.log("Using simplejson included in the interpreter") + logger.info("Using simplejson included in the interpreter") # ~ else: - # ~ logger.log("Usando json incluido en el interprete") + # ~ logger.info("Usando json incluido en el interprete") import sys PY3 = False @@ -90,7 +90,7 @@ def get_node_from_file(name_file, node, path=None): @return: dict with the node to return @rtype: dict """ - logger.log() + logger.debug() from platformcode import config from core import filetools @@ -129,7 +129,7 @@ def check_to_backup(data, fname, dict_data): @param dict_data: dictionary name @type dict_data: dict """ - logger.log() + logger.debug() if not dict_data: logger.error("Error loading json from file %s" % fname) @@ -161,7 +161,7 @@ def update_node(dict_node, name_file, node, path=None, silent=False): @return json_data @rtype: dict """ - if not silent: logger.log() + if not silent: logger.info() from platformcode import config from core import filetools diff --git a/core/scraper.py b/core/scraper.py index daae7035..2515a77d 100644 --- a/core/scraper.py +++ b/core/scraper.py @@ -61,7 +61,7 @@ def find_and_set_infoLabels(item): # Check if there is a 'code' if scraper_result and item.infoLabels['code']: # correct code - logger.log("Identificador encontrado: %s" % item.infoLabels['code']) + logger.info("Identificador encontrado: %s" % item.infoLabels['code']) scraper.completar_codigos(item) return True elif scraper_result: @@ -71,7 +71,7 @@ def find_and_set_infoLabels(item): # Content not found msg = config.get_localized_string(60228) % title - logger.log(msg) + logger.info(msg) # Show box with other options: if scrapers_disponibles[scraper_actual] in list_opciones_cuadro: list_opciones_cuadro.remove(scrapers_disponibles[scraper_actual]) @@ -95,10 +95,10 @@ def find_and_set_infoLabels(item): elif index == 1: # You have to create a dialog box to enter the data - logger.log("Complete information") + logger.info("Complete information") if cuadro_completar(item): # correct code - logger.log("Identifier found: %s" % str(item.infoLabels['code'])) + logger.info("Identifier found: %s" % str(item.infoLabels['code'])) return True # raise @@ -121,7 +121,7 @@ def find_and_set_infoLabels(item): def cuadro_completar(item): - logger.log() + logger.info() global dict_default dict_default = {} @@ -234,7 +234,7 @@ def get_nfo(item): @rtype: str @return: """ - logger.log() + logger.info() if "infoLabels" in item and "noscrap_id" in item.infoLabels: # Create the xml file with the data obtained from the item since there is no active scraper info_nfo = '<?xml version="1.0" encoding="UTF-8" standalone="yes" ?>' diff --git a/core/scrapertools.py b/core/scrapertools.py index aaa75e15..3df0dab8 100644 --- a/core/scrapertools.py +++ b/core/scrapertools.py @@ -34,7 +34,7 @@ from platformcode import logger def printMatches(matches): i = 0 for match in matches: - logger.log("%d %s" % (i, match)) + logger.info("%d %s" % (i, match)) i = i + 1 @@ -447,7 +447,7 @@ def get_season_and_episode(title): except: pass - logger.log("'" + title + "' -> '" + filename + "'") + logger.info("'" + title + "' -> '" + filename + "'") return filename diff --git a/core/servertools.py b/core/servertools.py index 06180340..36e0ef74 100644 --- a/core/servertools.py +++ b/core/servertools.py @@ -47,7 +47,7 @@ def find_video_items(item=None, data=None): @return: returns the itemlist with the results @rtype: list """ - logger.log() + logger.info() itemlist = [] # Download the page @@ -97,7 +97,7 @@ def get_servers_itemlist(itemlist, fnc=None, sort=False): # Walk the patterns for pattern in server_parameters.get("find_videos", {}).get("patterns", []): - logger.log(pattern["pattern"]) + logger.info(pattern["pattern"]) # Scroll through the results for match in re.compile(pattern["pattern"], re.DOTALL).finditer( "\n".join([item.url.split('|')[0] for item in itemlist if not item.server])): @@ -144,7 +144,7 @@ def findvideos(data, skip=False): return some link. It can also be an integer greater than 1, which would represent the maximum number of links to search. :return: """ - logger.log() + logger.info() devuelve = [] skip = int(skip) servers_list = list(get_servers_list().keys()) @@ -181,7 +181,7 @@ def findvideosbyserver(data, serverid): value = translate_server_name(server_parameters["name"]) , url, serverid, server_parameters.get("thumbnail", "") if value not in devuelve and url not in server_parameters["find_videos"].get("ignore_urls", []): devuelve.append(value) - logger.log(msg) + logger.info(msg) return devuelve @@ -193,7 +193,7 @@ def guess_server_thumbnail(serverid): def get_server_from_url(url): - logger.log() + logger.info() servers_list = list(get_servers_list().keys()) # Run findvideos on each active server @@ -211,7 +211,7 @@ def get_server_from_url(url): for n, pattern in enumerate(server_parameters["find_videos"].get("patterns", [])): msg = "%s\npattern: %s" % (serverid, pattern["pattern"]) if not "pattern_compiled" in pattern: - # logger.log('compiled ' + serverid) + # logger.info('compiled ' + serverid) pattern["pattern_compiled"] = re.compile(pattern["pattern"]) dict_servers_parameters[serverid]["find_videos"]["patterns"][n]["pattern_compiled"] = pattern["pattern_compiled"] # Scroll through the results @@ -224,7 +224,7 @@ def get_server_from_url(url): msg += "\nurl encontrada: %s" % url value = translate_server_name(server_parameters["name"]), url, serverid, server_parameters.get("thumbnail", "") if url not in server_parameters["find_videos"].get("ignore_urls", []): - logger.log(msg) + logger.info(msg) return value return None @@ -247,7 +247,7 @@ def resolve_video_urls_for_playing(server, url, video_password="", muestra_dialo @return: returns the url of the video @rtype: list """ - logger.log("Server: %s, Url: %s" % (server, url)) + logger.info("Server: %s, Url: %s" % (server, url)) server = server.lower() @@ -260,7 +260,7 @@ def resolve_video_urls_for_playing(server, url, video_password="", muestra_dialo if server == "directo" or server == "local": if isinstance(video_password, list): return video_password, len(video_password) > 0, "<br/>".join(error_messages) - logger.log("Server: %s, url is good" % server) + logger.info("Server: %s, url is good" % server) video_urls.append(["%s [%s]" % (urlparse.urlparse(url)[2][-4:], config.get_localized_string(30137)), url]) # Find out the video URL @@ -291,7 +291,7 @@ def resolve_video_urls_for_playing(server, url, video_password="", muestra_dialo priority = int(config.get_setting("resolve_priority")) opciones = sorted(opciones, key=lambda x: orden[priority].index(x)) - logger.log("Available options: %s | %s" % (len(opciones), opciones)) + logger.info("Available options: %s | %s" % (len(opciones), opciones)) else: logger.error("There is no connector for the server %s" % server) error_messages.append(config.get_localized_string(60004) % server) @@ -300,7 +300,7 @@ def resolve_video_urls_for_playing(server, url, video_password="", muestra_dialo # Import the server try: server_module = __import__('servers.%s' % server, None, None, ["servers.%s" % server]) - logger.log("Imported server: %s" % server_module) + logger.info("Imported server: %s" % server_module) except: server_module = None if muestra_dialogo: @@ -311,17 +311,17 @@ def resolve_video_urls_for_playing(server, url, video_password="", muestra_dialo # If it has a function to see if the video exists, check it now if hasattr(server_module, 'test_video_exists'): - logger.log("Invoking a %s.test_video_exists" % server) + logger.info("Invoking a %s.test_video_exists" % server) try: video_exists, message = server_module.test_video_exists(page_url=url) if not video_exists: error_messages.append(message) - logger.log("test_video_exists says video doesn't exist") + logger.info("test_video_exists says video doesn't exist") if muestra_dialogo: progreso.close() else: - logger.log("test_video_exists says the video DOES exist") + logger.info("test_video_exists says the video DOES exist") except: logger.error("Could not verify if the video exists") import traceback @@ -348,7 +348,7 @@ def resolve_video_urls_for_playing(server, url, video_password="", muestra_dialo # Free mode if opcion == "free": try: - logger.log("Invoking a %s.get_video_url" % server) + logger.info("Invoking a %s.get_video_url" % server) response = serverid.get_video_url(page_url=url, video_password=video_password) video_urls.extend(response) except: @@ -360,7 +360,7 @@ def resolve_video_urls_for_playing(server, url, video_password="", muestra_dialo # Premium mode else: try: - logger.log("Invoking a %s.get_video_url" % opcion) + logger.info("Invoking a %s.get_video_url" % opcion) response = serverid.get_video_url(page_url=url, premium=True, user=config.get_setting("user", server=opcion), password=config.get_setting("password", server=opcion), @@ -470,7 +470,7 @@ def get_server_parameters(server): @return: server data @rtype: dict """ - # logger.log("server %s" % server) + # logger.info("server %s" % server) global dict_servers_parameters server = server.split('.')[0] if not server: @@ -520,15 +520,15 @@ def get_server_parameters(server): # def get_server_json(server_name): -# # logger.log("server_name=" + server_name) +# # logger.info("server_name=" + server_name) # try: # server_path = filetools.join(config.get_runtime_path(), "servers", server_name + ".json") # if not filetools.exists(server_path): # server_path = filetools.join(config.get_runtime_path(), "servers", "debriders", server_name + ".json") # -# # logger.log("server_path=" + server_path) +# # logger.info("server_path=" + server_path) # server_json = jsontools.load(filetools.read(server_path)) -# # logger.log("server_json= %s" % server_json) +# # logger.info("server_json= %s" % server_json) # # except Exception as ex: # template = "An exception of type %s occured. Arguments:\n%r" @@ -600,7 +600,7 @@ def get_server_setting(name, server, default=None): if isinstance(dict_file, dict) and 'settings' in dict_file: dict_settings = dict_file['settings'] except EnvironmentError: - logger.log("ERROR when reading the file: %s" % file_settings) + logger.info("ERROR when reading the file: %s" % file_settings) if not dict_settings or name not in dict_settings: # We get controls from the file ../servers/server.json @@ -614,7 +614,7 @@ def get_server_setting(name, server, default=None): dict_file['settings'] = dict_settings # We create the file ../settings/channel_data.json if not filetools.write(file_settings, jsontools.dump(dict_file)): - logger.log("ERROR saving file: %s" % file_settings) + logger.info("ERROR saving file: %s" % file_settings) # We return the value of the local parameter 'name' if it exists, if default is not returned return dict_settings.get(name, default) @@ -636,7 +636,7 @@ def set_server_setting(name, value, server): dict_file = jsontools.load(filetools.read(file_settings)) dict_settings = dict_file.get('settings', {}) except EnvironmentError: - logger.log("ERROR when reading the file: %s" % file_settings) + logger.info("ERROR when reading the file: %s" % file_settings) dict_settings[name] = value @@ -648,7 +648,7 @@ def set_server_setting(name, value, server): # We create the file ../settings/channel_data.json if not filetools.write(file_settings, jsontools.dump(dict_file)): - logger.log("ERROR saving file: %s" % file_settings) + logger.info("ERROR saving file: %s" % file_settings) return None return value @@ -683,7 +683,7 @@ def get_debriders_list(): if server.endswith(".json"): server_parameters = get_server_parameters(server) if server_parameters["active"] == True: - logger.log(server_parameters) + logger.info(server_parameters) server_list[server.split(".")[0]] = server_parameters return server_list @@ -729,7 +729,7 @@ def check_list_links(itemlist, numero='', timeout=3): it = res[0] verificacion = res[1] it.title = verificacion + ' ' + it.title.strip() - logger.log('VERIFICATION= ' + verificacion) + logger.info('VERIFICATION= ' + verificacion) it.alive = verificacion return itemlist @@ -750,7 +750,7 @@ def check_video_link(item, timeout=3): server_module = __import__('servers.%s' % server, None, None, ["servers.%s" % server]) except: server_module = None - logger.log("[check_video_link] Cannot import server! %s" % server) + logger.info("[check_video_link] Cannot import server! %s" % server) return item, NK if hasattr(server_module, 'test_video_exists'): @@ -760,20 +760,20 @@ def check_video_link(item, timeout=3): try: video_exists, message = server_module.test_video_exists(page_url=url) if not video_exists: - logger.log("[check_video_link] Does not exist! %s %s %s" % (message, server, url)) + logger.info("[check_video_link] Does not exist! %s %s %s" % (message, server, url)) resultado = KO else: - logger.log("[check_video_link] check ok %s %s" % (server, url)) + logger.info("[check_video_link] check ok %s %s" % (server, url)) resultado = OK except: - logger.log("[check_video_link] Can't check now! %s %s" % (server, url)) + logger.info("[check_video_link] Can't check now! %s %s" % (server, url)) resultado = NK finally: httptools.HTTPTOOLS_DEFAULT_DOWNLOAD_TIMEOUT = ant_timeout # Restore download time return item, resultado - logger.log("[check_video_link] There is no test_video_exists for server: %s" % server) + logger.info("[check_video_link] There is no test_video_exists for server: %s" % server) return item, NK def translate_server_name(name): diff --git a/core/support.py b/core/support.py index d1c62eaa..805b9918 100755 --- a/core/support.py +++ b/core/support.py @@ -7,13 +7,12 @@ PY3 = False if sys.version_info[0] >= 3: PY3 = True; unicode = str; unichr = chr; long = int if PY3: from concurrent import futures - import urllib.request as urllib from urllib.request import Request, urlopen import urllib.parse as urlparse from urllib.parse import urlencode else: from concurrent_py2 import futures - import urllib, urlparse + import urlparse from urllib2 import Request, urlopen from urllib import urlencode @@ -22,7 +21,7 @@ from core import httptools, scrapertools, servertools, tmdb, channeltools from core.item import Item from lib import unshortenit from platformcode import config -from platformcode.logger import log +from platformcode.logger import info from platformcode import logger from specials import autoplay @@ -35,7 +34,7 @@ def hdpass_get_servers(item): for mir_url, srv in scrapertools.find_multiple_matches(mir, patron_option): mir_url = scrapertools.decodeHtmlentities(mir_url) - log(mir_url) + info(mir_url) it = item.clone(action="play", quality=quality, title=srv, server=srv, url= mir_url) if not servertools.get_server_parameters(srv.lower()): it = hdpass_get_url(it)[0] # do not exists or it's empty ret.append(it) @@ -83,7 +82,7 @@ def color(text, color): def search(channel, item, texto): - log(item.url + " search " + texto) + info(item.url + " search " + texto) if 'findhost' in dir(channel): channel.findhost() item.url = channel.host + "/?s=" + texto try: @@ -162,11 +161,10 @@ def unifyEp(ep): def scrapeBlock(item, args, block, patron, headers, action, pagination, debug, typeContentDict, typeActionDict, blacklist, search, pag, function, lang, sceneTitle): itemlist = [] - log("scrapeBlock qui") if debug: regexDbg(item, patron, headers, block) matches = scrapertools.find_multiple_matches_groups(block, patron) - log('MATCHES =', matches) + logger.debug('MATCHES =', matches) known_keys = ['url', 'title', 'title2', 'season', 'episode', 'thumb', 'quality', 'year', 'plot', 'duration', 'genere', 'rating', 'type', 'lang', 'other', 'size', 'seed'] # Legenda known_keys per i groups nei patron @@ -283,7 +281,7 @@ def scrapeBlock(item, args, block, patron, headers, action, pagination, debug, t try: parsedTitle = guessit(title) title = longtitle = parsedTitle.get('title', '') - log('TITOLO',title) + logger.debug('TITOLO',title) if parsedTitle.get('source'): quality = str(parsedTitle.get('source')) if parsedTitle.get('screen_size'): @@ -317,7 +315,7 @@ def scrapeBlock(item, args, block, patron, headers, action, pagination, debug, t longtitle += s + parsedTitle.get('episode_title') item.contentEpisodeTitle = parsedTitle.get('episode_title') except: - log('Error') + logger.debug('Error') longtitle = typo(longtitle, 'bold') lang1, longtitle = scrapeLang(scraped, lang, longtitle) @@ -404,7 +402,7 @@ def scrape(func): args = func(*args) function = func.__name__ if not 'actLike' in args else args['actLike'] - # log('STACK= ',inspect.stack()[1][3]) + # info('STACK= ',inspect.stack()[1][3]) item = args['item'] @@ -436,7 +434,7 @@ def scrape(func): matches = [] for n in range(2): - log('PATRON= ', patron) + logger.debug('PATRON= ', patron) if not data: page = httptools.downloadpage(item.url, headers=headers, ignore_response_code=True) data = re.sub("='([^']+)'", '="\\1"', page.data) @@ -452,7 +450,7 @@ def scrape(func): blocks = scrapertools.find_multiple_matches_groups(data, patronBlock) block = "" for bl in blocks: - # log(len(blocks),bl) + # info(len(blocks),bl) if 'season' in bl and bl['season']: item.season = bl['season'] blockItemlist, blockMatches = scrapeBlock(item, args, bl['block'], patron, headers, action, pagination, debug, @@ -477,7 +475,7 @@ def scrape(func): # if url may be changed and channel has findhost to update if 'findhost' in func.__globals__ and not itemlist: - log('running findhost ' + func.__module__) + info('running findhost ' + func.__module__) host = func.__globals__['findhost']() parse = list(urlparse.urlparse(item.url)) from core import jsontools @@ -534,7 +532,7 @@ def scrape(func): if config.get_setting('trakt_sync'): from core import trakt_tools trakt_tools.trakt_check(itemlist) - log('scraping time: ', time()-scrapingTime) + logger.debug('scraping time: ', time()-scrapingTime) return itemlist return wrapper @@ -712,15 +710,11 @@ def menuItem(itemlist, filename, title='', action='', url='', contentType='undef def menu(func): def wrapper(*args): - log() args = func(*args) item = args['item'] - log(item.channel + ' start') + logger.debug(item.channel + ' menu start') host = func.__globals__['host'] - list_servers = func.__globals__['list_servers'] if 'list_servers' in func.__globals__ else ['directo'] - list_quality = func.__globals__['list_quality'] if 'list_quality' in func.__globals__ else ['default'] - log('LIST QUALITY', list_quality) filename = func.__module__.split('.')[1] single_search = False # listUrls = ['film', 'filmSub', 'tvshow', 'tvshowSub', 'anime', 'animeSub', 'search', 'top', 'topSub'] @@ -735,7 +729,7 @@ def menu(func): for name in listUrls: dictUrl[name] = args[name] if name in args else None - log(dictUrl[name]) + logger.debug(dictUrl[name]) if name == 'film': title = 'Film' if name == 'tvshow': title = 'Serie TV' if name == 'anime': title = 'Anime' @@ -803,7 +797,7 @@ def menu(func): # Apply auto Thumbnails at the menus thumb(itemlist) - log(item.channel + ' end') + logger.debug(item.channel + ' menu end') return itemlist return wrapper @@ -1066,7 +1060,7 @@ def videolibrary(itemlist, item, typography='', function_level=1, function=''): # Simply add this function to add video library support # Function_level is useful if the function is called by another function. # If the call is direct, leave it blank - log() + info() if item.contentType == 'movie': action = 'add_pelicula_to_library' @@ -1116,7 +1110,7 @@ def videolibrary(itemlist, item, typography='', function_level=1, function=''): def nextPage(itemlist, item, data='', patron='', function_or_level=1, next_page='', resub=[]): # Function_level is useful if the function is called by another function. # If the call is direct, leave it blank - log() + info() action = inspect.stack()[function_or_level][3] if type(function_or_level) == int else function_or_level if next_page == '': next_page = scrapertools.find_single_match(data, patron) @@ -1126,7 +1120,7 @@ def nextPage(itemlist, item, data='', patron='', function_or_level=1, next_page= if 'http' not in next_page: next_page = scrapertools.find_single_match(item.url, 'https?://[a-z0-9.-]+') + (next_page if next_page.startswith('/') else '/' + next_page) next_page = next_page.replace('&', '&') - log('NEXT= ', next_page) + info('NEXT= ', next_page) itemlist.append( Item(channel=item.channel, action = action, @@ -1154,7 +1148,7 @@ def pagination(itemlist, item, page, perpage, function_level=1): def server(item, data='', itemlist=[], headers='', AutoPlay=True, CheckLinks=True, Download=True, patronTag=None, Videolibrary=True): - log() + info() blacklisted_servers = config.get_setting("black_list", server='servers') if not blacklisted_servers: blacklisted_servers = [] if not data and not itemlist: @@ -1167,7 +1161,7 @@ def server(item, data='', itemlist=[], headers='', AutoPlay=True, CheckLinks=Tru def getItem(videoitem): if not servertools.get_server_parameters(videoitem.server.lower()): # do not exists or it's empty findS = servertools.get_server_from_url(videoitem.url) - log(findS) + info(findS) if not findS: if item.channel == 'community': findS= (config.get_localized_string(30137), videoitem.url, 'directo') @@ -1175,7 +1169,7 @@ def server(item, data='', itemlist=[], headers='', AutoPlay=True, CheckLinks=Tru videoitem.url = unshortenit.unshorten_only(videoitem.url)[0] findS = servertools.get_server_from_url(videoitem.url) if not findS: - log(videoitem, 'Non supportato') + info(videoitem, 'Non supportato') return videoitem.server = findS[2] videoitem.title = findS[0] @@ -1328,7 +1322,7 @@ def addQualityTag(item, itemlist, data, patron): folder=False, thumbnail=thumb('info'))) else: - log('nessun tag qualità trovato') + info('nessun tag qualità trovato') def get_jwplayer_mediaurl(data, srvName, onlyHttp=False): video_urls = [] @@ -1419,7 +1413,7 @@ def thumb(item_itemlist_string=None, genre=False, live=False): '_tvshow':['serie','tv', 'fiction']} def autoselect_thumb(item, genre): - log('SPLIT',re.split(r'\.|\{|\}|\[|\]|\(|\)|/| ',item.title.lower())) + info('SPLIT',re.split(r'\.|\{|\}|\[|\]|\(|\)|/| ',item.title.lower())) if genre == False: for thumb, titles in icon_dict.items(): if any(word in re.split(r'\.|\{|\}|\[|\]|\(|\)|/| ',item.title.lower()) for word in search): diff --git a/core/tmdb.py b/core/tmdb.py index efddebd7..5c6e403e 100644 --- a/core/tmdb.py +++ b/core/tmdb.py @@ -87,7 +87,7 @@ create_bd() # The function name is the name of the decorator and receives the function that decorates. def cache_response(fn): - logger.log() + logger.info() # import time # start_time = time.time() @@ -498,7 +498,7 @@ def set_infoLabels_item(item, seekTmdb=True, idioma_busqueda=def_lang, lock=None def find_and_set_infoLabels(item): - logger.log() + logger.info() global otmdb_global tmdb_result = None @@ -908,7 +908,7 @@ class Tmdb(object): cls.dic_generos[idioma][tipo] = {} url = ('http://api.themoviedb.org/3/genre/%s/list?api_key=a1ab8b8669da03637a4b98fa39c39228&language=%s' % (tipo, idioma)) try: - logger.log("[Tmdb.py] Filling in dictionary of genres") + logger.info("[Tmdb.py] Filling in dictionary of genres") resultado = cls.get_json(url) if not isinstance(resultado, dict): @@ -940,7 +940,7 @@ class Tmdb(object): '&language=%s' % (self.busqueda_id, source, self.busqueda_idioma)) buscando = "%s: %s" % (source.capitalize(), self.busqueda_id) - logger.log("[Tmdb.py] Searching %s:\n%s" % (buscando, url)) + logger.info("[Tmdb.py] Searching %s:\n%s" % (buscando, url)) resultado = self.get_json(url) if not isinstance(resultado, dict): resultado = ast.literal_eval(resultado.decode('utf-8')) @@ -985,7 +985,7 @@ class Tmdb(object): url += '&year=%s' % self.busqueda_year buscando = self.busqueda_texto.capitalize() - logger.log("[Tmdb.py] Searching %s on page %s:\n%s" % (buscando, page, url)) + logger.info("[Tmdb.py] Searching %s on page %s:\n%s" % (buscando, page, url)) resultado = self.get_json(url) if not isinstance(resultado, dict): resultado = ast.literal_eval(resultado.decode('utf-8')) @@ -1046,7 +1046,7 @@ class Tmdb(object): url = ('http://api.themoviedb.org/3/%s?api_key=a1ab8b8669da03637a4b98fa39c39228&%s' % (type_search, "&".join(params))) - logger.log("[Tmdb.py] Searcing %s:\n%s" % (type_search, url)) + logger.info("[Tmdb.py] Searcing %s:\n%s" % (type_search, url)) resultado = self.get_json(url) if not isinstance(resultado, dict): resultado = ast.literal_eval(resultado.decode('utf-8')) @@ -1111,7 +1111,7 @@ class Tmdb(object): return True def get_list_resultados(self, num_result=20): - # logger.log("self %s" % str(self)) + # logger.info("self %s" % str(self)) res = [] if num_result <= 0: @@ -1331,7 +1331,7 @@ class Tmdb(object): "&append_to_response=credits" % (self.result["id"], numtemporada, self.busqueda_idioma) buscando = "id_Tmdb: " + str(self.result["id"]) + " season: " + str(numtemporada) + "\nURL: " + url - logger.log("[Tmdb.py] Searcing " + buscando) + logger.info("[Tmdb.py] Searcing " + buscando) try: self.temporada[numtemporada] = self.get_json(url) if not isinstance(self.temporada[numtemporada], dict): @@ -1520,7 +1520,7 @@ class Tmdb(object): items.extend(list(self.get_episodio(ret_infoLabels['season'], episodio).items())) - # logger.log("ret_infoLabels" % ret_infoLabels) + # logger.info("ret_infoLabels" % ret_infoLabels) for k, v in items: if not v: diff --git a/core/trakt_tools.py b/core/trakt_tools.py index ae6cf44f..e5d23e51 100644 --- a/core/trakt_tools.py +++ b/core/trakt_tools.py @@ -128,7 +128,7 @@ def token_trakt(item): def set_trakt_info(item): - logger.log() + logger.info() import xbmcgui # Envia los datos a trakt try: @@ -139,7 +139,7 @@ def set_trakt_info(item): pass def get_trakt_watched(id_type, mediatype, update=False): - logger.log() + logger.info() id_list = [] id_dict = dict() @@ -229,7 +229,7 @@ def trakt_check(itemlist): def get_sync_from_file(): - logger.log() + logger.info() sync_path = os.path.join(config.get_data_path(), 'settings_channels', 'trakt_data.json') trakt_node = {} if os.path.exists(sync_path): @@ -241,7 +241,7 @@ def get_sync_from_file(): def update_trakt_data(mediatype, trakt_data): - logger.log() + logger.info() sync_path = os.path.join(config.get_data_path(), 'settings_channels', 'trakt_data.json') if os.path.exists(sync_path): @@ -251,7 +251,7 @@ def update_trakt_data(mediatype, trakt_data): def ask_install_script(): - logger.log() + logger.info() from platformcode import platformtools @@ -265,7 +265,7 @@ def ask_install_script(): def wait_for_update_trakt(): - logger.log() + logger.info() t = Thread(update_all) t.setDaemon(True) t.start() @@ -274,7 +274,7 @@ def wait_for_update_trakt(): def update_all(): # from core.support import dbg;dbg() from time import sleep - logger.log() + logger.info() sleep(20) while xbmc.Player().isPlaying(): sleep(20) diff --git a/core/tvdb.py b/core/tvdb.py index 984b2c8e..e961e4b9 100644 --- a/core/tvdb.py +++ b/core/tvdb.py @@ -73,8 +73,8 @@ otvdb_global = None def find_and_set_infoLabels(item): - logger.log() - # logger.log("item es %s" % item) + logger.info() + # logger.info("item es %s" % item) p_dialog = None if not item.contentSeason: @@ -368,7 +368,7 @@ class Tvdb(object): @classmethod def __check_token(cls): - # logger.log() + # logger.info() if TOKEN == "": cls.__login() else: @@ -383,7 +383,7 @@ class Tvdb(object): @staticmethod def __login(): - # logger.log() + # logger.info() global TOKEN apikey = "106B699FDC04301C" @@ -413,7 +413,7 @@ class Tvdb(object): @classmethod def __refresh_token(cls): - # logger.log() + # logger.info() global TOKEN is_success = False @@ -512,7 +512,7 @@ class Tvdb(object): ] } """ - logger.log() + logger.info() if id_episode and self.episodes.get(id_episode): return self.episodes.get(id_episode) @@ -582,7 +582,7 @@ class Tvdb(object): } } """ - logger.log() + logger.info() try: url = HOST + "/series/%s/episodes?page=%s" % (_id, page) @@ -600,7 +600,7 @@ class Tvdb(object): else: self.list_episodes[page] = jsontools.load(html) - # logger.log("dict_html %s" % self.list_episodes) + # logger.info("dict_html %s" % self.list_episodes) return self.list_episodes[page] @@ -668,7 +668,7 @@ class Tvdb(object): """ if semaforo: semaforo.acquire() - logger.log() + logger.info() url = HOST + "/episodes/%s" % _id @@ -691,7 +691,7 @@ class Tvdb(object): dict_html = jsontools.load(html) dict_html = dict_html.pop("data") - logger.log("dict_html %s" % dict_html) + logger.info("dict_html %s" % dict_html) self.episodes[_id] = dict_html if semaforo: @@ -722,7 +722,7 @@ class Tvdb(object): "status": "string" } """ - logger.log() + logger.info() try: @@ -743,7 +743,7 @@ class Tvdb(object): req = urllib.request.Request(url, headers=DEFAULT_HEADERS) response = urllib.request.urlopen(req) html = response.read() - logger.log(html) + logger.info(html) response.close() except Exception as ex: @@ -821,7 +821,7 @@ class Tvdb(object): } } """ - logger.log() + logger.info() resultado = {} url = HOST + "/series/%s" % _id @@ -880,7 +880,7 @@ class Tvdb(object): @rtype: dict """ - logger.log() + logger.info() if self.result.get('image_season_%s' % season): return self.result['image_season_%s' % season] @@ -932,7 +932,7 @@ class Tvdb(object): @return: dictionary with actors @rtype: dict """ - logger.log() + logger.info() url = HOST + "/series/%s/actors" % _id DEFAULT_HEADERS["Accept-Language"] = lang @@ -962,7 +962,7 @@ class Tvdb(object): @rtype: list @return: list of results """ - logger.log() + logger.info() list_results = [] # if we have a result and it has seriesName, we already have the info of the series, it is not necessary to search again diff --git a/core/videolibrarytools.py b/core/videolibrarytools.py index 01dc85e7..7efb313b 100644 --- a/core/videolibrarytools.py +++ b/core/videolibrarytools.py @@ -78,7 +78,7 @@ def save_movie(item, silent=False): @rtype fallidos: int @return: the number of failed items or -1 if all failed """ - logger.log() + logger.info() # logger.debug(item.tostring('\n')) insertados = 0 sobreescritos = 0 @@ -144,7 +144,7 @@ def save_movie(item, silent=False): if not path: # Create folder path = filetools.join(MOVIES_PATH, ("%s [%s]" % (base_name, _id)).strip()) - logger.log("Creating movie directory:" + path) + logger.info("Creating movie directory:" + path) if not filetools.mkdir(path): logger.debug("Could not create directory") return 0, 0, -1, path @@ -159,7 +159,7 @@ def save_movie(item, silent=False): if not nfo_exists: # We create .nfo if it doesn't exist - logger.log("Creating .nfo: " + nfo_path) + logger.info("Creating .nfo: " + nfo_path) head_nfo = scraper.get_nfo(item) item_nfo = Item(title=item.contentTitle, channel="videolibrary", action='findvideos', @@ -182,7 +182,7 @@ def save_movie(item, silent=False): if item_nfo and strm_exists: if json_exists: - logger.log("The file exists. Is overwritten") + logger.info("The file exists. Is overwritten") sobreescritos += 1 else: insertados += 1 @@ -209,7 +209,7 @@ def save_movie(item, silent=False): item_nfo.library_urls[item.channel] = item.url if filetools.write(nfo_path, head_nfo + item_nfo.tojson()): - #logger.log("FOLDER_MOVIES : %s" % FOLDER_MOVIES) + #logger.info("FOLDER_MOVIES : %s" % FOLDER_MOVIES) # We update the Kodi video library with the movie if config.is_xbmc() and config.get_setting("videolibrary_kodi") and not silent: from platformcode import xbmc_videolibrary @@ -238,7 +238,7 @@ def update_renumber_options(item, head_nfo, path): json = json_file['TVSHOW_AUTORENUMBER'] if item.fulltitle in json: item.channel_prefs[channel]['TVSHOW_AUTORENUMBER'] = json[item.fulltitle] - logger.log('UPDATED=\n' + str(item.channel_prefs)) + logger.info('UPDATED=\n' + str(item.channel_prefs)) filetools.write(tvshow_path, head_nfo + item.tojson()) def add_renumber_options(item, head_nfo, path): @@ -426,7 +426,7 @@ def save_tvshow(item, episodelist, silent=False): @rtype path: str @return: serial directory """ - logger.log() + logger.info() # logger.debug(item.tostring('\n')) path = "" @@ -483,7 +483,7 @@ def save_tvshow(item, episodelist, silent=False): if not path: path = filetools.join(TVSHOWS_PATH, ("%s [%s]" % (base_name, _id)).strip()) - logger.log("Creating series directory: " + path) + logger.info("Creating series directory: " + path) try: filetools.mkdir(path) except OSError as exception: @@ -493,7 +493,7 @@ def save_tvshow(item, episodelist, silent=False): tvshow_path = filetools.join(path, "tvshow.nfo") if not filetools.exists(tvshow_path): # We create tvshow.nfo, if it does not exist, with the head_nfo, series info and watched episode marks - logger.log("Creating tvshow.nfo: " + tvshow_path) + logger.info("Creating tvshow.nfo: " + tvshow_path) head_nfo = scraper.get_nfo(item) item.infoLabels['mediatype'] = "tvshow" item.infoLabels['title'] = item.contentSerieName @@ -567,11 +567,11 @@ def save_episodes(path, episodelist, serie, silent=False, overwrite=True): @rtype fallidos: int @return: the number of failed episodes """ - logger.log() + logger.info() episodelist = filter_list(episodelist, serie.action, path) # No episode list, nothing to save if not len(episodelist): - logger.log("There is no episode list, we go out without creating strm") + logger.info("There is no episode list, we go out without creating strm") return 0, 0, 0 # process local episodes @@ -586,7 +586,7 @@ def save_episodes(path, episodelist, serie, silent=False, overwrite=True): elif config.get_setting("local_episodes", "videolibrary"): done, local_episodes_path = config_local_episodes_path(path, serie) if done < 0: - logger.log("An issue has occurred while configuring local episodes, going out without creating strm") + logger.info("An issue has occurred while configuring local episodes, going out without creating strm") return 0, 0, done item_nfo.local_episodes_path = local_episodes_path filetools.write(nfo_path, head_nfo + item_nfo.tojson()) @@ -710,7 +710,7 @@ def save_episodes(path, episodelist, serie, silent=False, overwrite=True): # No episode list, nothing to save if not len(new_episodelist): - logger.log("There is no episode list, we go out without creating strm") + logger.info("There is no episode list, we go out without creating strm") return 0, 0, 0 local_episodelist += get_local_content(path) @@ -742,12 +742,12 @@ def save_episodes(path, episodelist, serie, silent=False, overwrite=True): json_path = filetools.join(path, ("%s [%s].json" % (season_episode, e.channel)).lower()) if season_episode in local_episodelist: - logger.log('Skipped: Serie ' + serie.contentSerieName + ' ' + season_episode + ' available as local content') + logger.info('Skipped: Serie ' + serie.contentSerieName + ' ' + season_episode + ' available as local content') continue # check if the episode has been downloaded if filetools.join(path, "%s [downloads].json" % season_episode) in ficheros: - logger.log('INFO: "%s" episode %s has been downloaded, skipping it' % (serie.contentSerieName, season_episode)) + logger.info('INFO: "%s" episode %s has been downloaded, skipping it' % (serie.contentSerieName, season_episode)) continue strm_exists = strm_path in ficheros @@ -800,7 +800,7 @@ def save_episodes(path, episodelist, serie, silent=False, overwrite=True): if filetools.write(json_path, e.tojson()): if not json_exists: - logger.log("Inserted: %s" % json_path) + logger.info("Inserted: %s" % json_path) insertados += 1 # We mark episode as unseen news_in_playcounts[season_episode] = 0 @@ -811,14 +811,14 @@ def save_episodes(path, episodelist, serie, silent=False, overwrite=True): news_in_playcounts[serie.contentSerieName] = 0 else: - logger.log("Overwritten: %s" % json_path) + logger.info("Overwritten: %s" % json_path) sobreescritos += 1 else: - logger.log("Failed: %s" % json_path) + logger.info("Failed: %s" % json_path) fallidos += 1 else: - logger.log("Failed: %s" % json_path) + logger.info("Failed: %s" % json_path) fallidos += 1 if not silent and p_dialog.iscanceled(): @@ -888,7 +888,7 @@ def save_episodes(path, episodelist, serie, silent=False, overwrite=True): def config_local_episodes_path(path, item, silent=False): - logger.log(item) + logger.info(item) from platformcode.xbmc_videolibrary import search_local_path local_episodes_path=search_local_path(item) if not local_episodes_path: @@ -900,11 +900,11 @@ def config_local_episodes_path(path, item, silent=False): platformtools.dialog_ok(config.get_localized_string(30131), config.get_localized_string(80043)) local_episodes_path = platformtools.dialog_browse(0, config.get_localized_string(80046)) if local_episodes_path == '': - logger.log("User has canceled the dialog") + logger.info("User has canceled the dialog") return -2, local_episodes_path elif path in local_episodes_path: platformtools.dialog_ok(config.get_localized_string(30131), config.get_localized_string(80045)) - logger.log("Selected folder is the same of the TV show one") + logger.info("Selected folder is the same of the TV show one") return -2, local_episodes_path if local_episodes_path: @@ -919,7 +919,7 @@ def config_local_episodes_path(path, item, silent=False): def process_local_episodes(local_episodes_path, path): - logger.log() + logger.info() sub_extensions = ['.srt', '.sub', '.sbv', '.ass', '.idx', '.ssa', '.smi'] artwork_extensions = ['.jpg', '.jpeg', '.png'] @@ -958,7 +958,7 @@ def process_local_episodes(local_episodes_path, path): def get_local_content(path): - logger.log() + logger.info() local_episodelist = [] for root, folders, files in filetools.walk(path): @@ -987,7 +987,7 @@ def add_movie(item): @type item: item @param item: item to be saved. """ - logger.log() + logger.info() from platformcode.launcher import set_search_temp; set_search_temp(item) # To disambiguate titles, TMDB is caused to ask for the really desired title @@ -1034,7 +1034,7 @@ def add_tvshow(item, channel=None): @param channel: channel from which the series will be saved. By default, item.from_channel or item.channel will be imported. """ - logger.log("show=#" + item.show + "#") + logger.info("show=#" + item.show + "#") from platformcode.launcher import set_search_temp; set_search_temp(item) if item.channel == "downloads": @@ -1105,7 +1105,7 @@ def add_tvshow(item, channel=None): else: platformtools.dialog_ok(config.get_localized_string(30131), config.get_localized_string(60070) % item.show) - logger.log("%s episodes of series %s have been added to the video library" % (insertados, item.show)) + logger.info("%s episodes of series %s have been added to the video library" % (insertados, item.show)) if config.is_xbmc(): if config.get_setting("sync_trakt_new_tvshow", "videolibrary"): import xbmc @@ -1121,7 +1121,7 @@ def add_tvshow(item, channel=None): def emergency_urls(item, channel=None, path=None, headers={}): - logger.log() + logger.info() import re from servers import torrent try: diff --git a/core/ziptools.py b/core/ziptools.py index 3b3a3683..138c4c1b 100644 --- a/core/ziptools.py +++ b/core/ziptools.py @@ -17,8 +17,8 @@ from core import filetools class ziptools(object): def extract(self, file, dir, folder_to_extract="", overwrite_question=False, backup=False): - logger.log("file= %s" % file) - logger.log("dir= %s" % dir) + logger.info("file= %s" % file) + logger.info("dir= %s" % dir) if not dir.endswith(':') and not filetools.exists(dir): filetools.mkdir(dir) @@ -30,13 +30,13 @@ class ziptools(object): for nameo in zf.namelist(): name = nameo.replace(':', '_').replace('<', '_').replace('>', '_').replace('|', '_').replace('"', '_').replace('?', '_').replace('*', '_') - logger.log("name=%s" % nameo) + logger.info("name=%s" % nameo) if not name.endswith('/'): - logger.log("it's not a directory") + logger.info("it's not a directory") try: (path, filename) = filetools.split(filetools.join(dir, name)) - logger.log("path=%s" % path) - logger.log("name=%s" % name) + logger.info("path=%s" % path) + logger.info("name=%s" % name) if folder_to_extract: if path != filetools.join(dir, folder_to_extract): break @@ -49,7 +49,7 @@ class ziptools(object): else: outfilename = filetools.join(dir, name) - logger.log("outfilename=%s" % outfilename) + logger.info("outfilename=%s" % outfilename) try: if filetools.exists(outfilename) and overwrite_question: from platformcode import platformtools @@ -74,7 +74,7 @@ class ziptools(object): try: zf.close() except: - logger.log("Error closing .zip " + file) + logger.info("Error closing .zip " + file) def _createstructure(self, file, dir): self._makedirs(self._listdirs(file), dir) diff --git a/default.py b/default.py index b5c1474e..7ed2cc72 100644 --- a/default.py +++ b/default.py @@ -9,7 +9,7 @@ import sys import xbmc from platformcode import config, logger -logger.log("init...") +logger.info("init...") librerias = xbmc.translatePath(os.path.join(config.get_runtime_path(), 'lib')) sys.path.insert(0, librerias) diff --git a/lib/arm_chromeos.py b/lib/arm_chromeos.py index 4dddd954..d4808402 100644 --- a/lib/arm_chromeos.py +++ b/lib/arm_chromeos.py @@ -27,7 +27,7 @@ class ChromeOSImage: """ def __init__(self, imgpath): - logger.log('Image Path: ' + imgpath) + logger.info('Image Path: ' + imgpath) """Prepares the image""" self.imgpath = imgpath self.bstream = self.get_bstream(imgpath) @@ -59,7 +59,7 @@ class ChromeOSImage: self.seek_stream(entries_start * lba_size) if not calcsize(part_format) == entry_size: - logger.log('Partition table entries are not 128 bytes long') + logger.info('Partition table entries are not 128 bytes long') return 0 for index in range(1, entries_num + 1): # pylint: disable=unused-variable @@ -71,7 +71,7 @@ class ChromeOSImage: break if not offset: - logger.log('Failed to calculate losetup offset.') + logger.info('Failed to calculate losetup offset.') return 0 return offset @@ -93,7 +93,7 @@ class ChromeOSImage: while True: chunk2 = self.read_stream(chunksize) if not chunk2: - logger.log('File %s not found in the ChromeOS image' % filename) + logger.info('File %s not found in the ChromeOS image' % filename) return False chunk = chunk1 + chunk2 diff --git a/lib/generictools.py b/lib/generictools.py index dd344e09..ffe77898 100644 --- a/lib/generictools.py +++ b/lib/generictools.py @@ -25,7 +25,7 @@ intervenido_sucuri = 'Access Denied - Sucuri Website Firewall' def update_title(item): - logger.log() + logger.info() from core import scraper,support @@ -41,7 +41,7 @@ def update_title(item): The channel must add a method to be able to receive the call from Kodi / Alfa, and be able to call this method: def actualizar_titulos(item): - logger.log() + logger.info() itemlist = [] from lib import generictools from platformcode import launcher @@ -205,7 +205,7 @@ def update_title(item): def refresh_screen(item): - logger.log() + logger.info() """ #### Kodi 18 compatibility #### @@ -239,7 +239,7 @@ def refresh_screen(item): def post_tmdb_listado(item, itemlist): - logger.log() + logger.info() itemlist_fo = [] """ @@ -484,7 +484,7 @@ def post_tmdb_listado(item, itemlist): def post_tmdb_seasons(item, itemlist): - logger.log() + logger.info() """ @@ -644,7 +644,7 @@ def post_tmdb_seasons(item, itemlist): def post_tmdb_episodios(item, itemlist): - logger.log() + logger.info() itemlist_fo = [] """ @@ -995,7 +995,7 @@ def post_tmdb_episodios(item, itemlist): def post_tmdb_findvideos(item, itemlist): - logger.log() + logger.info() """ @@ -1215,7 +1215,7 @@ def post_tmdb_findvideos(item, itemlist): def get_field_from_kodi_DB(item, from_fields='*', files='file'): - logger.log() + logger.info() """ Call to read from the Kodi DB the input fields received (from_fields, by default "*") of the video indicated in Item @@ -1293,7 +1293,7 @@ def get_field_from_kodi_DB(item, from_fields='*', files='file'): def fail_over_newpct1(item, patron, patron2=None, timeout=None): - logger.log() + logger.info() import ast """ @@ -1494,7 +1494,7 @@ def fail_over_newpct1(item, patron, patron2=None, timeout=None): def web_intervenida(item, data, desactivar=True): - logger.log() + logger.info() """ @@ -1577,7 +1577,7 @@ def web_intervenida(item, data, desactivar=True): def regenerate_clones(): - logger.log() + logger.info() import json from core import videolibrarytools @@ -1591,7 +1591,7 @@ def regenerate_clones(): # Find the paths where to leave the control .json file, and the Video Library json_path = filetools.exists(filetools.join(config.get_runtime_path(), 'verify_cached_torrents.json')) if json_path: - logger.log('Previously repaired video library: WE ARE GOING') + logger.info('Previously repaired video library: WE ARE GOING') return False json_path = filetools.join(config.get_runtime_path(), 'verify_cached_torrents.json') filetools.write(json_path, json.dumps({"CINE_verify": True})) # Prevents another simultaneous process from being launched @@ -1631,7 +1631,7 @@ def regenerate_clones(): # Delete the Tvshow.nfo files and check if the .nfo has more than one channel and one is clone Newpct1 for file in files: - # logger.log('file - nfos: ' + file) + # logger.info('file - nfos: ' + file) if 'tvshow.nfo' in file: file_path = filetools.join(root, 'tvshow.nfo') filetools.remove(file_path) @@ -1697,7 +1697,7 @@ def regenerate_clones(): for file in files: file_path = filetools.join(root, file) if '.json' in file: - logger.log('** file: ' + file) + logger.info('** file: ' + file) canal_json = scrapertools.find_single_match(file, r'\[(\w+)\].json') if canal_json not in nfo.library_urls: filetools.remove(file_path) # we delete the .json is a zombie @@ -1740,7 +1740,7 @@ def regenerate_clones(): def dejuice(data): - logger.log() + logger.info() # Method to unobtrusive JuicyCodes data import base64 diff --git a/lib/megaserver/client.py b/lib/megaserver/client.py index 2b96c7a1..eb9fb395 100644 --- a/lib/megaserver/client.py +++ b/lib/megaserver/client.py @@ -47,7 +47,7 @@ class Client(object): t= Thread(target=self._auto_shutdown) t.setDaemon(True) t.start() - logger.log("MEGA Server Started") + logger.info("MEGA Server Started") def _auto_shutdown(self): while self.running: @@ -77,7 +77,7 @@ class Client(object): def stop(self): self.running = False self._server.stop() - logger.log("MEGA Server Stopped") + logger.info("MEGA Server Stopped") def get_play_list(self): if len(self.files) > 1: @@ -105,7 +105,7 @@ class Client(object): return files except: - logger.log(traceback.format_exc()) + logger.info(traceback.format_exc()) pass return files diff --git a/lib/sambatools/libsmb.py b/lib/sambatools/libsmb.py index 5cb83fbd..4a66385d 100644 --- a/lib/sambatools/libsmb.py +++ b/lib/sambatools/libsmb.py @@ -14,7 +14,7 @@ remote = None def parse_url(url): - # logger.log("Url: %s" % url) + # logger.info("Url: %s" % url) url = url.strip() patron = "^smb://(?:([^;\n]+);)?(?:([^:@\n]+)[:|@])?(?:([^@\n]+)@)?([^/]+)/([^/\n]+)([/]?.*?)$" domain, user, password, server_name, share_name, path = re.compile(patron, re.DOTALL).match(url).groups() @@ -27,7 +27,7 @@ def parse_url(url): if path.endswith("/"): path = path[:-1] if not path: path = "/" - # logger.log("Dominio: '%s' |Usuario: '%s' | Password: '%s' | Servidor: '%s' | IP: '%s' | Share Name: '%s' | Path: '%s'" % (domain, user, password, server_name, server_ip, share_name, path)) + # logger.info("Dominio: '%s' |Usuario: '%s' | Password: '%s' | Servidor: '%s' | IP: '%s' | Share Name: '%s' | Path: '%s'" % (domain, user, password, server_name, server_ip, share_name, path)) return server_name, server_ip, share_name, unicode(path, "utf8"), user, password, domain @@ -46,7 +46,7 @@ def get_server_name_ip(server): def connect(url): - # logger.log("Url: %s" % url) + # logger.info("Url: %s" % url) global remote server_name, server_ip, share_name, path, user, password, domain = parse_url(url) @@ -63,7 +63,7 @@ def connect(url): def listdir(url): - logger.log("Url: %s" % url) + logger.info("Url: %s" % url) remote, share_name, path = connect(url) try: files = [f.filename for f in remote.listPath(share_name, path) if not f.filename in [".", ".."]] @@ -73,7 +73,7 @@ def listdir(url): def walk(url, topdown=True, onerror=None): - logger.log("Url: %s" % url) + logger.info("Url: %s" % url) remote, share_name, path = connect(url) try: @@ -103,7 +103,7 @@ def walk(url, topdown=True, onerror=None): def get_attributes(url): - logger.log("Url: %s" % url) + logger.info("Url: %s" % url) remote, share_name, path = connect(url) try: return remote.getAttributes(share_name, path) @@ -112,7 +112,7 @@ def get_attributes(url): def mkdir(url): - logger.log("Url: %s" % url) + logger.info("Url: %s" % url) remote, share_name, path = connect(url) try: remote.createDirectory(share_name, path) @@ -121,12 +121,12 @@ def mkdir(url): def smb_open(url, mode): - logger.log("Url: %s" % url) + logger.info("Url: %s" % url) return SMBFile(url, mode) def isfile(url): - logger.log("Url: %s" % url) + logger.info("Url: %s" % url) remote, share_name, path = connect(url) try: files = [f.filename for f in remote.listPath(share_name, os.path.dirname(path)) if not f.isDirectory] @@ -136,7 +136,7 @@ def isfile(url): def isdir(url): - logger.log("Url: %s" % url) + logger.info("Url: %s" % url) remote, share_name, path = connect(url) try: folders = [f.filename for f in remote.listPath(share_name, os.path.dirname(path)) if f.isDirectory] @@ -146,7 +146,7 @@ def isdir(url): def exists(url): - logger.log("Url: %s" % url) + logger.info("Url: %s" % url) remote, share_name, path = connect(url) try: files = [f.filename for f in remote.listPath(share_name, os.path.dirname(path))] @@ -156,7 +156,7 @@ def exists(url): def remove(url): - logger.log("Url: %s" % url) + logger.info("Url: %s" % url) remote, share_name, path = connect(url) try: remote.deleteFiles(share_name, path) @@ -165,7 +165,7 @@ def remove(url): def rmdir(url): - logger.log("Url: %s" % url) + logger.info("Url: %s" % url) remote, share_name, path = connect(url) try: remote.deleteDirectory(share_name, path) @@ -174,7 +174,7 @@ def rmdir(url): def rename(url, new_name): - logger.log("Url: %s" % url) + logger.info("Url: %s" % url) remote, share_name, path = connect(url) _, _, _, new_name, _, _, _ = parse_url(new_name) try: diff --git a/lib/unshortenit.py b/lib/unshortenit.py index 3ca16285..6752d994 100644 --- a/lib/unshortenit.py +++ b/lib/unshortenit.py @@ -96,7 +96,7 @@ class UnshortenIt(object): if oldUri == uri: break - logger.log(uri) + logger.info(uri) return uri, code @@ -531,12 +531,12 @@ class UnshortenIt(object): r = httptools.downloadpage(uri, timeout=self._timeout, headers=headers, follow_redirects=False) if 'Wait 1 hour' in r.data: uri = '' - logger.log('IP bannato da vcrypt, aspetta un ora') + logger.info('IP bannato da vcrypt, aspetta un ora') else: prev_uri = uri uri = r.headers['location'] if uri == prev_uri: - logger.log('Use Cloudscraper') + logger.info('Use Cloudscraper') uri = httptools.downloadpage(uri, timeout=self._timeout, headers=headers, follow_redirects=False, cf=True).headers['location'] if "4snip" in uri: @@ -593,7 +593,7 @@ class UnshortenIt(object): r = httptools.downloadpage(uri, follow_redirect=True, timeout=self._timeout, cookies=False) if 'get/' in r.url: uri = 'https://linkhub.icu/view/' + re.search('\.\./view/([^"]+)', r.data).group(1) - logger.log(uri) + logger.info(uri) r = httptools.downloadpage(uri, follow_redirect=True, timeout=self._timeout, cookies=False) uri = re.search('<div id="text-url".*\n\s+<a href="([^"]+)', r.data).group(0) return uri, r.code @@ -683,7 +683,7 @@ def findlinks(text): regex = '(?:https?://(?:[\w\d]+\.)?)?(?:' + regex + ')/[a-zA-Z0-9_=/]+' for match in re.findall(regex, text): matches.append(match) - logger.log('matches=' + str(matches)) + logger.info('matches=' + str(matches)) if len(matches) == 1: text += '\n' + unshorten(matches[0])[0] elif matches: diff --git a/platformcode/config.py b/platformcode/config.py index 47a8b9e1..3d160bb7 100644 --- a/platformcode/config.py +++ b/platformcode/config.py @@ -215,23 +215,23 @@ def get_setting(name, channel="", server="", default=None): # Specific channel setting if channel: - # logger.log("get_setting reading channel setting '"+name+"' from channel json") + # logger.info("get_setting reading channel setting '"+name+"' from channel json") from core import channeltools value = channeltools.get_channel_setting(name, channel, default) - # logger.log("get_setting -> '"+repr(value)+"'") + # logger.info("get_setting -> '"+repr(value)+"'") return value # Specific server setting elif server: - # logger.log("get_setting reading server setting '"+name+"' from server json") + # logger.info("get_setting reading server setting '"+name+"' from server json") from core import servertools value = servertools.get_server_setting(name, server, default) - # logger.log("get_setting -> '"+repr(value)+"'") + # logger.info("get_setting -> '"+repr(value)+"'") return value # Global setting else: - # logger.log("get_setting reading main setting '"+name+"'") + # logger.info("get_setting reading main setting '"+name+"'") value = __settings__.getSetting(name) if not value: return default diff --git a/platformcode/download_and_play.py b/platformcode/download_and_play.py index 29d1ced7..24bc68a1 100644 --- a/platformcode/download_and_play.py +++ b/platformcode/download_and_play.py @@ -22,17 +22,17 @@ from platformcode import config, logger # Download a file and start playing while downloading def download_and_play(url, file_name, download_path): # Start thread - logger.log("Active threads " + str(threading.active_count())) - logger.log("" + repr(threading.enumerate())) - logger.log("Starting download thread...") + logger.info("Active threads " + str(threading.active_count())) + logger.info("" + repr(threading.enumerate())) + logger.info("Starting download thread...") download_thread = DownloadThread(url, file_name, download_path) download_thread.start() - logger.log("Download thread started") - logger.log("Active threads " + str(threading.active_count())) - logger.log("" + repr(threading.enumerate())) + logger.info("Download thread started") + logger.info("Active threads " + str(threading.active_count())) + logger.info("" + repr(threading.enumerate())) # Wait - logger.log("Waiting...") + logger.info("Waiting...") while True: cancelled = False @@ -53,7 +53,7 @@ def download_and_play(url, file_name, download_path): dialog.close() - logger.log("End of waiting") + logger.info("End of waiting") # Launch the player player = CustomPlayer() @@ -61,66 +61,66 @@ def download_and_play(url, file_name, download_path): player.PlayStream(download_thread.get_file_name()) # End of playback - logger.log("End of playback") + logger.info("End of playback") if player.is_stopped(): - logger.log("Terminated by user") + logger.info("Terminated by user") break else: if not download_thread.isAlive(): - logger.log("Download has finished") + logger.info("Download has finished") break else: - logger.log("Continua la descarga") + logger.info("Continua la descarga") # When the player finishes, if you continue downloading it for now - logger.log("Download thread alive=" + str(download_thread.isAlive())) + logger.info("Download thread alive=" + str(download_thread.isAlive())) if download_thread.isAlive(): - logger.log("Killing download thread") + logger.info("Killing download thread") download_thread.force_stop() class CustomPlayer(xbmc.Player): def __init__(self, *args, **kwargs): - logger.log() + logger.info() self.actualtime = 0 self.totaltime = 0 self.stopped = False xbmc.Player.__init__(self) def PlayStream(self, url): - logger.log("url=" + url) + logger.info("url=" + url) self.play(url) self.actualtime = 0 self.url = url while self.isPlaying(): self.actualtime = self.getTime() self.totaltime = self.getTotalTime() - logger.log("actualtime=" + str(self.actualtime) + " totaltime=" + str(self.totaltime)) + logger.info("actualtime=" + str(self.actualtime) + " totaltime=" + str(self.totaltime)) xbmc.sleep(3000) def set_download_thread(self, download_thread): - logger.log() + logger.info() self.download_thread = download_thread def force_stop_download_thread(self): - logger.log() + logger.info() if self.download_thread.isAlive(): - logger.log("Killing download thread") + logger.info("Killing download thread") self.download_thread.force_stop() # while self.download_thread.isAlive(): # xbmc.sleep(1000) def onPlayBackStarted(self): - logger.log("PLAYBACK STARTED") + logger.info("PLAYBACK STARTED") def onPlayBackEnded(self): - logger.log("PLAYBACK ENDED") + logger.info("PLAYBACK ENDED") def onPlayBackStopped(self): - logger.log("PLAYBACK STOPPED") + logger.info("PLAYBACK STOPPED") self.stopped = True self.force_stop_download_thread() @@ -131,7 +131,7 @@ class CustomPlayer(xbmc.Player): # Download in background class DownloadThread(threading.Thread): def __init__(self, url, file_name, download_path): - # logger.log(repr(file)) + # logger.info(repr(file)) self.url = url self.download_path = download_path self.file_name = os.path.join(download_path, file_name) @@ -148,16 +148,16 @@ class DownloadThread(threading.Thread): threading.Thread.__init__(self) def run(self): - logger.log("Download starts...") + logger.info("Download starts...") if "megacrypter.com" in self.url: self.download_file_megacrypter() else: self.download_file() - logger.log("Download ends") + logger.info("Download ends") def force_stop(self): - logger.log() + logger.info() force_stop_file = open(self.force_stop_file_name, "w") force_stop_file.write("0") force_stop_file.close() @@ -181,38 +181,38 @@ class DownloadThread(threading.Thread): return self.total_size def download_file_megacrypter(self): - logger.log() + logger.info() comando = "./megacrypter.sh" - logger.log("command= " + comando) + logger.info("command= " + comando) oldcwd = os.getcwd() - logger.log("oldcwd= " + oldcwd) + logger.info("oldcwd= " + oldcwd) cwd = os.path.join(config.get_runtime_path(), "tools") - logger.log("cwd= " + cwd) + logger.info("cwd= " + cwd) os.chdir(cwd) - logger.log("directory changed to= " + os.getcwd()) + logger.info("directory changed to= " + os.getcwd()) - logger.log("destination= " + self.download_path) + logger.info("destination= " + self.download_path) os.system(comando + " '" + self.url + "' \"" + self.download_path + "\"") # p = subprocess.Popen([comando , self.url , self.download_path], cwd=cwd, stdout=subprocess.PIPE , stderr=subprocess.PIPE ) # out, err = p.communicate() - # logger.log("DownloadThread.download_file out="+out) + # logger.info("DownloadThread.download_file out="+out) os.chdir(oldcwd) def download_file(self): - logger.log("Direct download") + logger.info("Direct download") headers = [] # Ensures that the file can be created - logger.log("filename= " + self.file_name) + logger.info("filename= " + self.file_name) self.file_name = xbmc.makeLegalFilename(self.file_name) - logger.log("filename= " + self.file_name) - logger.log("url= " + self.url) + logger.info("filename= " + self.file_name) + logger.info("url= " + self.url) # Create the file existSize = 0 @@ -228,13 +228,13 @@ class DownloadThread(threading.Thread): additional_headers = [additional_headers] for additional_header in additional_headers: - logger.log("additional_header: " + additional_header) + logger.info("additional_header: " + additional_header) name = re.findall("(.*?)=.*?", additional_header)[0] value = urllib.parse.unquote_plus(re.findall(".*?=(.*?)$", additional_header)[0]) headers.append([name, value]) self.url = self.url.split("|")[0] - logger.log("url= " + self.url) + logger.info("url= " + self.url) # Timeout del socket a 60 segundos socket.setdefaulttimeout(60) @@ -243,7 +243,7 @@ class DownloadThread(threading.Thread): h = urllib.request.HTTPHandler(debuglevel=0) request = urllib.request.Request(self.url) for header in headers: - logger.log("Header= " + header[0] + ": " + header[1]) + logger.info("Header= " + header[0] + ": " + header[1]) request.add_header(header[0], header[1]) # Lanza la petición @@ -272,18 +272,18 @@ class DownloadThread(threading.Thread): self.total_size = int(float(totalfichero) / float(1024 * 1024)) - logger.log("Content-Length=%s" % totalfichero) + logger.info("Content-Length=%s" % totalfichero) blocksize = 100 * 1024 bloqueleido = connexion.read(blocksize) - logger.log("Starting file download, blocked= %s" % len(bloqueleido)) + logger.info("Starting file download, blocked= %s" % len(bloqueleido)) maxreintentos = 10 while len(bloqueleido) > 0: try: if os.path.exists(self.force_stop_file_name): - logger.log("Force_stop file detected, download is interrupted") + logger.info("Force_stop file detected, download is interrupted") f.close() xbmc.executebuiltin("Notification(%s,%s,300)" % (config.get_localized_string(60319),config.get_localized_string(60320))) @@ -297,7 +297,7 @@ class DownloadThread(threading.Thread): # except: f.write(bloqueleido) grabado = grabado + len(bloqueleido) - logger.log("grabado=%d de %d" % (grabado, totalfichero)) + logger.info("grabado=%d de %d" % (grabado, totalfichero)) percent = int(float(grabado) * 100 / float(totalfichero)) self.progress = percent totalmb = float(float(totalfichero) / (1024 * 1024)) @@ -323,7 +323,7 @@ class DownloadThread(threading.Thread): except: import sys reintentos = reintentos + 1 - logger.log("ERROR in block download, retry %d" % reintentos) + logger.info("ERROR in block download, retry %d" % reintentos) for line in sys.exc_info(): logger.error("%s" % line) diff --git a/platformcode/envtal.py b/platformcode/envtal.py index 541d06ad..0894d564 100644 --- a/platformcode/envtal.py +++ b/platformcode/envtal.py @@ -25,7 +25,7 @@ from platformcode import logger, config, platformtools def get_environment(): """ Returns the most common OS, Kodi and Alpha environment variables, -    necessary for fault diagnosis + necessary for fault diagnosis """ try: @@ -341,34 +341,31 @@ def list_env(environment={}): if not environment: environment = get_environment() - if environment['debug'] == 'False': - logger.log_enable(True) + logger.info(sep) + logger.info('KoD environment variables: ' + environment['addon_version'] + ' Debug: ' + environment['debug']) + logger.info(sep) - logger.log(sep) - logger.log('KoD environment variables: ' + environment['addon_version'] + ' Debug: ' + environment['debug']) - logger.log(sep) - - logger.log(environment['os_name'] + ' ' + environment['prod_model'] + ' ' + + logger.info(environment['os_name'] + ' ' + environment['prod_model'] + ' ' + environment['os_release'] + ' ' + environment['machine'] + ' ' + environment['architecture'] + ' ' + environment['language']) - logger.log('Kodi ' + environment['num_version'] + ', Vídeo: ' + + logger.info('Kodi ' + environment['num_version'] + ', Vídeo: ' + environment['video_db'] + ', Python ' + environment['python_version']) if environment['cpu_usage']: - logger.log('CPU: ' + environment['cpu_usage']) + logger.info('CPU: ' + environment['cpu_usage']) if environment['mem_total'] or environment['mem_free']: - logger.log('Memory: Total: ' + environment['mem_total'] + ' MB | Disp.: ' + + logger.info('Memory: Total: ' + environment['mem_total'] + ' MB | Disp.: ' + environment['mem_free'] + ' MB | Buffers: ' + str(int(environment['kodi_buffer']) * 3) + ' MB | Buffermode: ' + environment['kodi_bmode'] + ' | Readfactor: ' + environment['kodi_rfactor']) - logger.log('Userdata: ' + environment['userdata_path'] + ' - Free: ' + + logger.info('Userdata: ' + environment['userdata_path'] + ' - Free: ' + environment['userdata_free'].replace('.', ',') + ' GB') - logger.log('Videolibrary: Series/Episodes: ' + environment['videolab_series'] + '/' + + logger.info('Videolibrary: Series/Episodes: ' + environment['videolab_series'] + '/' + environment['videolab_episodios'] + ' - Pelis: ' + environment['videolab_pelis'] + ' - Upd: ' + environment['videolab_update'] + ' - Path: ' + @@ -380,27 +377,24 @@ def list_env(environment={}): # if x == 0: # cliente_alt = cliente.copy() # del cliente_alt['Torrent_opt'] - # logger.log('Torrent: Opt: %s, %s' % (str(cliente['Torrent_opt']), \ + # logger.info('Torrent: Opt: %s, %s' % (str(cliente['Torrent_opt']), \ # str(cliente_alt).replace('{', '').replace('}', '') \ # .replace("'", '').replace('_', ' '))) # elif x == 1 and environment['torrent_error']: - # logger.log('- ' + str(cliente).replace('{', '').replace('}', '') \ + # logger.info('- ' + str(cliente).replace('{', '').replace('}', '') \ # .replace("'", '').replace('_', ' ')) # else: # cliente_alt = cliente.copy() # del cliente_alt['Plug_in'] # cliente_alt['Libre'] = cliente_alt['Libre'].replace('.', ',') + ' GB' - # logger.log('- %s: %s' % (str(cliente['Plug_in']), str(cliente_alt) \ + # logger.info('- %s: %s' % (str(cliente['Plug_in']), str(cliente_alt) \ # .replace('{', '').replace('}', '').replace("'", '') \ # .replace('\\\\', '\\'))) - # logger.log('Proxy: ' + environment['proxy_active']) + # logger.info('Proxy: ' + environment['proxy_active']) - logger.log('LOG Size: ' + environment['log_size'].replace('.', ',') + ' MB') - logger.log(sep) - - if environment['debug'] == 'False': - logger.log_enable(False) + logger.info('LOG Size: ' + environment['log_size'].replace('.', ',') + ' MB') + logger.info(sep) return environment diff --git a/platformcode/launcher.py b/platformcode/launcher.py index 8c0629fe..e9d2f862 100644 --- a/platformcode/launcher.py +++ b/platformcode/launcher.py @@ -19,7 +19,7 @@ def start(): Within this function all calls should go to functions that we want to execute as soon as we open the plugin. """ - logger.log() + logger.info() # config.set_setting('show_once', True) # Test if all the required directories are created config.verify_directories_created() @@ -37,7 +37,8 @@ def start(): updater.showSavedChangelog() def run(item=None): - logger.log() + from core.support import dbg + logger.info() if not item: # Extract item from sys.argv if sys.argv[2]: @@ -76,7 +77,7 @@ def run(item=None): xbmc_videolibrary.ask_set_content(silent=False) config.set_setting('show_once', True) - logger.log(item.tostring()) + logger.info(item.tostring()) try: if not config.get_setting('tmdb_active'): @@ -84,7 +85,7 @@ def run(item=None): # If item has no action, stops here if item.action == "": - logger.log("Item without action") + logger.info("Item without action") return # Action for main menu in channelselector @@ -154,7 +155,7 @@ def run(item=None): channel_file = os.path.join(config.get_runtime_path(), CHANNELS, item.channel + ".py") - logger.log("channel_file= " + channel_file + ' - ' + CHANNELS + ' - ' + item.channel) + logger.info("channel_file= " + channel_file + ' - ' + CHANNELS + ' - ' + item.channel) channel = None @@ -164,7 +165,7 @@ def run(item=None): except ImportError: exec("import " + CHANNELS + "." + item.channel + " as channel") - logger.log("Running channel %s | %s" % (channel.__name__, channel.__file__)) + logger.info("Running channel %s | %s" % (channel.__name__, channel.__file__)) # Special play action if item.action == "play": @@ -174,12 +175,12 @@ def run(item=None): trakt_tools.set_trakt_info(item) except: pass - logger.log("item.action=%s" % item.action.upper()) + logger.info("item.action=%s" % item.action.upper()) # logger.debug("item_toPlay: " + "\n" + item.tostring('\n')) # First checks if channel has a "play" function if hasattr(channel, 'play'): - logger.log("Executing channel 'play' method") + logger.info("Executing channel 'play' method") itemlist = channel.play(item) b_favourite = item.isFavourite # Play should return a list of playable URLS @@ -200,7 +201,7 @@ def run(item=None): # If player don't have a "play" function, not uses the standard play from platformtools else: - logger.log("Executing core 'play' method") + logger.info("Executing core 'play' method") platformtools.play_video(item) # Special action for findvideos, where the plugin looks for known urls @@ -213,7 +214,7 @@ def run(item=None): # If not, uses the generic findvideos function else: - logger.log("No channel 'findvideos' method, " + logger.info("No channel 'findvideos' method, " "executing core method") itemlist = servertools.find_video_items(item) @@ -258,7 +259,7 @@ def run(item=None): else: filetools.remove(temp_search_file) - logger.log("item.action=%s" % item.action.upper()) + logger.info("item.action=%s" % item.action.upper()) from core import channeltools if config.get_setting('last_search'): @@ -279,7 +280,7 @@ def run(item=None): # For all other actions else: # import web_pdb; web_pdb.set_trace() - logger.log("Executing channel '%s' method" % item.action) + logger.info("Executing channel '%s' method" % item.action) itemlist = getattr(channel, item.action)(item) if config.get_setting('trakt_sync'): from core import trakt_tools @@ -360,7 +361,7 @@ def set_search_temp(item): filetools.write(temp_search_file, f) def reorder_itemlist(itemlist): - logger.log() + logger.info() # logger.debug("Inlet itemlist size: %i" % len(itemlist)) new_list = [] @@ -398,7 +399,7 @@ def reorder_itemlist(itemlist): new_list.extend(mod_list) new_list.extend(not_mod_list) - logger.log("Modified Titles:%i |Unmodified:%i" % (modified, not_modified)) + logger.info("Modified Titles:%i |Unmodified:%i" % (modified, not_modified)) if len(new_list) == 0: new_list = itemlist @@ -408,7 +409,7 @@ def reorder_itemlist(itemlist): def limit_itemlist(itemlist): - logger.log() + logger.info() # logger.debug("Inlet itemlist size: %i" % len(itemlist)) try: @@ -441,7 +442,7 @@ def play_from_library(item): itemlist=[] item.fromLibrary = True - logger.log() + logger.info() # logger.debug("item: \n" + item.tostring('\n')) # Try to reproduce an image (this does nothing and also does not give an error) diff --git a/platformcode/logger.py b/platformcode/logger.py index 7748545d..8815fd0a 100644 --- a/platformcode/logger.py +++ b/platformcode/logger.py @@ -2,113 +2,41 @@ # -------------------------------------------------------------------------------- # Logger (kodi) # -------------------------------------------------------------------------------- - -import inspect, sys, os, xbmc +from __future__ import unicode_literals +import inspect, os, xbmc, sys from platformcode import config -PY3 = False -if sys.version_info[0] >= 3: PY3 = True; unicode = str; unichr = chr; long = int - -loggeractive = (config.get_setting("debug") == True) +LOG_FORMAT = '{addname}[{filename}.{function}:{line}]{sep} {message}' +DEBUG_ENABLED = config.get_setting("debug") +DEF_LEVEL = xbmc.LOGINFO if sys.version_info[0] >= 3 else xbmc.LOGNOTICE -def log_enable(active): - global loggeractive - loggeractive = active +def info(*args): + log(*args) -def encode_log(message=""): - - # Unicode to utf8 - if isinstance(message, unicode): - message = message.encode("utf8") - if PY3: message = message.decode("utf8") - - # All encodings to utf8 - elif not PY3 and isinstance(message, str): - message = unicode(message, "utf8", errors="replace").encode("utf8") - - # Bytes encodings to utf8 - elif PY3 and isinstance(message, bytes): - message = message.decode("utf8") - - # Objects to string - else: - message = str(message) - - return message +def debug(*args): + if DEBUG_ENABLED: + log(*args) -def get_caller(message=None): - - if message and isinstance(message, unicode): - message = message.encode("utf8") - if PY3: message = message.decode("utf8") - elif message and PY3 and isinstance(message, bytes): - message = message.decode("utf8") - elif message and not PY3: - message = unicode(message, "utf8", errors="replace").encode("utf8") - elif message: - message = str(message) - - module = inspect.getmodule(inspect.currentframe().f_back.f_back) - - if module == None: - module = "None" - else: - module = module.__name__ - - function = inspect.currentframe().f_back.f_back.f_code.co_name - - if module == "__main__": - module = "kod" - else: - module = "kod." + module - if message: - if module not in message: - if function == "<module>": - return module + " " + message - else: - return module + " [" + function + "] " + message - else: - return message - else: - if function == "<module>": - return module - else: - return module + "." + function +def error(*args): + log("######## ERROR #########", level=xbmc.LOGERROR) + log(*args, level=xbmc.LOGERROR) -def info(texto=""): - if loggeractive: - xbmc.log(get_caller(encode_log(texto)), xbmc.LOGNOTICE) - - -def debug(texto=""): - if loggeractive: - texto = " [" + get_caller() + "] " + encode_log(texto) - - xbmc.log("######## DEBUG #########", xbmc.LOGNOTICE) - xbmc.log(texto, xbmc.LOGNOTICE) - - -def error(texto=""): - texto = " [" + get_caller() + "] " + encode_log(texto) - - xbmc.log("######## ERROR #########", xbmc.LOGERROR) - xbmc.log(texto, xbmc.LOGERROR) - - -def log(*args): - # Function to simplify the log - # Automatically returns File Name and Function Name - if loggeractive: - string = '' - for arg in args: string += ' '+str(arg) - frame = inspect.stack()[1] - filename = frame[0].f_code.co_filename - filename = os.path.basename(filename) - xbmc.log("[" + filename + "] [" + inspect.stack()[1][3] + "] " + string, xbmc.LOGNOTICE) +def log(*args, **kwargs): + msg = '' + for arg in args: msg += ' ' + str(arg) + frame = inspect.currentframe().f_back.f_back + filename = frame.f_code.co_filename + filename = os.path.basename(filename).split('.')[0] + xbmc.log(LOG_FORMAT.format(addname=config.PLUGIN_NAME, + filename=filename, + line=frame.f_lineno, + sep=':' if msg else '', + function=frame.f_code.co_name, + message=msg), kwargs.get('level', DEF_LEVEL)) class WebErrorException(Exception): diff --git a/platformcode/platformtools.py b/platformcode/platformtools.py index 470c3299..6a544f24 100644 --- a/platformcode/platformtools.py +++ b/platformcode/platformtools.py @@ -112,7 +112,7 @@ def dialog_browse(_type, heading, shares="files", mask="", useThumbs=False, trea def itemlist_refresh(): # pos = Item().fromurl(xbmc.getInfoLabel('ListItem.FileNameAndPath')).itemlistPosition - # logger.log('Current position: ' + str(pos)) + # logger.info('Current position: ' + str(pos)) xbmc.executebuiltin("Container.Refresh") # while Item().fromurl(xbmc.getInfoLabel('ListItem.FileNameAndPath')).itemlistPosition != pos: @@ -133,7 +133,7 @@ def render_items(itemlist, parent_item): """ Function used to render itemlist on kodi """ - logger.log('START render_items') + logger.info('START render_items') thumb_type = config.get_setting('video_thumbnail_type') from specials import shortcuts from core import httptools @@ -218,7 +218,7 @@ def render_items(itemlist, parent_item): set_view_mode(itemlist[0], parent_item) xbmcplugin.endOfDirectory(_handle) - logger.log('END render_items') + logger.info('END render_items') def getCurrentView(item=None, parent_item=None): @@ -275,11 +275,11 @@ def set_view_mode(item, parent_item): if content: mode = int(config.get_setting('view_mode_%s' % content).split(',')[-1]) if mode == 0: - logger.log('default mode') + logger.info('default mode') mode = 55 xbmcplugin.setContent(handle=int(sys.argv[1]), content=Type) xbmc.executebuiltin('Container.SetViewMode(%s)' % mode) - logger.log('TYPE: ' + Type + ' - ' + 'CONTENT: ' + content) + logger.info('TYPE: ' + Type + ' - ' + 'CONTENT: ' + content) def set_infolabels(listitem, item, player=False): @@ -499,10 +499,10 @@ def is_playing(): def play_video(item, strm=False, force_direct=False, autoplay=False): - logger.log() + logger.info() logger.debug(item.tostring('\n')) if item.channel == 'downloads': - logger.log("Play local video: %s [%s]" % (item.title, item.url)) + logger.info("Play local video: %s [%s]" % (item.title, item.url)) xlistitem = xbmcgui.ListItem(path=item.url) xlistitem.setArt({"thumb": item.thumbnail}) set_infolabels(xlistitem, item, True) @@ -510,7 +510,7 @@ def play_video(item, strm=False, force_direct=False, autoplay=False): return default_action = config.get_setting("default_action") - logger.log("default_action=%s" % default_action) + logger.info("default_action=%s" % default_action) # Open the selection dialog to see the available options opciones, video_urls, seleccion, salir = get_dialogo_opciones(item, default_action, strm, autoplay) @@ -520,8 +520,8 @@ def play_video(item, strm=False, force_direct=False, autoplay=False): seleccion = get_seleccion(default_action, opciones, seleccion, video_urls) if seleccion < 0: return # Canceled box - logger.log("selection=%d" % seleccion) - logger.log("selection=%s" % opciones[seleccion]) + logger.info("selection=%d" % seleccion) + logger.info("selection=%s" % opciones[seleccion]) # run the available option, jdwonloader, download, favorites, add to the video library ... IF IT IS NOT PLAY salir = set_opcion(item, seleccion, opciones, video_urls) @@ -682,7 +682,7 @@ def alert_unsopported_server(): def handle_wait(time_to_wait, title, text): - logger.log("handle_wait(time_to_wait=%d)" % time_to_wait) + logger.info("handle_wait(time_to_wait=%d)" % time_to_wait) espera = dialog_progress(' ' + title, "") secs = 0 @@ -701,15 +701,15 @@ def handle_wait(time_to_wait, title, text): break if cancelled: - logger.log('Wait canceled') + logger.info('Wait canceled') return False else: - logger.log('Wait finished') + logger.info('Wait finished') return True def get_dialogo_opciones(item, default_action, strm, autoplay): - logger.log() + logger.info() # logger.debug(item.tostring('\n')) from core import servertools @@ -793,7 +793,7 @@ def get_dialogo_opciones(item, default_action, strm, autoplay): def set_opcion(item, seleccion, opciones, video_urls): - logger.log() + logger.info() # logger.debug(item.tostring('\n')) salir = False # You have not chosen anything, most likely because you have given the ESC @@ -843,7 +843,7 @@ def set_opcion(item, seleccion, opciones, video_urls): def get_video_seleccionado(item, seleccion, video_urls): - logger.log() + logger.info() mediaurl = "" view = False wait_time = 0 @@ -869,7 +869,7 @@ def get_video_seleccionado(item, seleccion, video_urls): mpd = True # If there is no mediaurl it is because the video is not there :) - logger.log("mediaurl=" + mediaurl) + logger.info("mediaurl=" + mediaurl) if mediaurl == "": if item.server == "unknown": alert_unsopported_server() @@ -886,7 +886,7 @@ def get_video_seleccionado(item, seleccion, video_urls): def set_player(item, xlistitem, mediaurl, view, strm, nfo_path=None, head_nfo=None, item_nfo=None): - logger.log() + logger.info() # logger.debug("item:\n" + item.tostring('\n')) # Moved del conector "torrent" here if item.server == "torrent": @@ -903,10 +903,10 @@ def set_player(item, xlistitem, mediaurl, view, strm, nfo_path=None, head_nfo=No player_mode = config.get_setting("player_mode") if (player_mode == 3 and mediaurl.startswith("rtmp")) or item.play_from == 'window' or item.nfo: player_mode = 0 elif "megacrypter.com" in mediaurl: player_mode = 3 - logger.log("mediaurl=" + mediaurl) + logger.info("mediaurl=" + mediaurl) if player_mode == 0: - logger.log('Player Mode: Direct') + logger.info('Player Mode: Direct') # Add the listitem to a playlist playlist = xbmc.PlayList(xbmc.PLAYLIST_VIDEO) playlist.clear() @@ -919,24 +919,24 @@ def set_player(item, xlistitem, mediaurl, view, strm, nfo_path=None, head_nfo=No trakt_tools.wait_for_update_trakt() elif player_mode == 1: - logger.log('Player Mode: setResolvedUrl') + logger.info('Player Mode: setResolvedUrl') xlistitem.setPath(mediaurl) xbmcplugin.setResolvedUrl(int(sys.argv[1]), True, xlistitem) xbmc.sleep(2500) elif player_mode == 2: - logger.log('Player Mode: Built-In') + logger.info('Player Mode: Built-In') xbmc.executebuiltin("PlayMedia(" + mediaurl + ")") elif player_mode == 3: - logger.log('Player Mode: Download and Play') + logger.info('Player Mode: Download and Play') from platformcode import download_and_play download_and_play.download_and_play(mediaurl, "download_and_play.tmp", config.get_setting("downloadpath")) return # ALL LOOKING TO REMOVE VIEW if item.subtitle and view: - logger.log("External subtitles: " + item.subtitle) + logger.info("External subtitles: " + item.subtitle) xbmc.sleep(2000) xbmc_player.setSubtitles(item.subtitle) @@ -962,7 +962,7 @@ def torrent_client_installed(show_tuple=False): def play_torrent(item, xlistitem, mediaurl): - logger.log() + logger.info() import time from servers import torrent @@ -1002,9 +1002,6 @@ def play_torrent(item, xlistitem, mediaurl): time.sleep(3) -def log(texto): - xbmc.log(texto, xbmc.LOGNOTICE) - def resume_playback(item, return_played_time=False): class ResumePlayback(xbmcgui.WindowXMLDialog): Close = False @@ -1082,17 +1079,17 @@ def install_inputstream(): # Check if InputStream add-on exists! Addon('inputstream.adaptive') - logger.log('InputStream add-on installed from repo.') + logger.info('InputStream add-on installed from repo.') except RuntimeError: - logger.log('InputStream add-on not installed.') + logger.info('InputStream add-on not installed.') dialog_ok(config.get_localized_string(20000), config.get_localized_string(30126)) return False else: try: Addon('inputstream.adaptive') - logger.log('InputStream add-on is installed and enabled') + logger.info('InputStream add-on is installed and enabled') except: - logger.log('enabling InputStream add-on') + logger.info('enabling InputStream add-on') xbmc.executebuiltin('UpdateLocalAddons') xbmc.executeJSONRPC('{"jsonrpc": "2.0", "id":1, "method": "Addons.SetAddonEnabled", "params": { "addonid": "inputstream.adaptive", "enabled": true }}') return True @@ -1207,13 +1204,13 @@ def best_chromeos_image(devices): # Select the newest version from distutils.version import LooseVersion # pylint: disable=import-error,no-name-in-module,useless-suppression if LooseVersion(device['version']) > LooseVersion(best['version']): - logger.log('%s (%s) is newer than %s (%s)' % (device['hwid'], device['version'], best['hwid'], best['version'])) + logger.info('%s (%s) is newer than %s (%s)' % (device['hwid'], device['version'], best['hwid'], best['version'])) best = device # Select the smallest image (disk space requirement) elif LooseVersion(device['version']) == LooseVersion(best['version']): if int(device['filesize']) + int(device['zipfilesize']) < int(best['filesize']) + int(best['zipfilesize']): - logger.log('%s (%d) is smaller than %s (%d)' % (device['hwid'], int(device['filesize']) + int(device['zipfilesize']), best['hwid'], int(best['filesize']) + int(best['zipfilesize']))) + logger.info('%s (%d) is smaller than %s (%d)' % (device['hwid'], int(device['filesize']) + int(device['zipfilesize']), best['hwid'], int(best['filesize']) + int(best['zipfilesize']))) best = device return best diff --git a/platformcode/recaptcha.py b/platformcode/recaptcha.py index c609227a..c4422143 100644 --- a/platformcode/recaptcha.py +++ b/platformcode/recaptcha.py @@ -65,7 +65,7 @@ class Recaptcha(xbmcgui.WindowXMLDialog): data = httptools.downloadpage(self.url, post=post, headers=self.headers).data from platformcode import logger - logger.log(data) + logger.info(data) self.result = scrapertools.find_single_match(data, '<div class="fbc-verification-token">.*?>([^<]+)<') if self.result: platformtools.dialog_notification("Captcha corretto", "Verifica conclusa") diff --git a/platformcode/subtitletools.py b/platformcode/subtitletools.py index 20d231e7..9a90bd1f 100644 --- a/platformcode/subtitletools.py +++ b/platformcode/subtitletools.py @@ -84,7 +84,7 @@ def regex_tvshow(compare, file, sub=""): def set_Subtitle(): - logger.log() + logger.info() exts = [".srt", ".sub", ".txt", ".smi", ".ssa", ".ass"] subtitle_folder_path = filetools.join(config.get_data_path(), "subtitles") @@ -93,7 +93,7 @@ def set_Subtitle(): if subtitle_type == "2": subtitle_path = config.get_setting("subtitlepath_file") - logger.log("Con subtitulo : " + subtitle_path) + logger.info("Con subtitulo : " + subtitle_path) xbmc.Player().setSubtitles(subtitle_path) else: if subtitle_type == "0": @@ -106,7 +106,7 @@ def set_Subtitle(): long_v = len(subtitle_path) if long_v > 0: if subtitle_path.startswith("http") or subtitle_path[long_v - 4, long] in exts: - logger.log("Con subtitulo : " + subtitle_path) + logger.info("Con subtitulo : " + subtitle_path) xbmc.Player().setSubtitles(subtitle_path) return else: @@ -125,7 +125,7 @@ def set_Subtitle(): Subnames = glob.glob(filetools.join(subtitle_path, "Movies", subtitle_name + "*.??.???")) for Subname in Subnames: if os.path.splitext(Subname)[1] in exts: - logger.log("Con subtitulo : " + filetools.split(Subname)[1]) + logger.info("Con subtitulo : " + filetools.split(Subname)[1]) xbmc.Player().setSubtitles((Subname)) except: logger.error("error al cargar subtitulos") @@ -216,7 +216,7 @@ def searchSubtitle(item): filetools.mkdir(full_path_tvshow) # title_new + ".mp4" full_path_video_new = xbmc.translatePath( filetools.join(full_path_tvshow, "%s %sx%s.mp4" % (tvshow_title, season, episode))) - logger.log(full_path_video_new) + logger.info(full_path_video_new) listitem = xbmcgui.ListItem(title_new, iconImage="DefaultVideo.png", thumbnailImage="") listitem.setInfo("video", {"Title": title_new, "Genre": "Tv shows", "episode": int(episode), "season": int(season), "tvshowtitle": tvshow_title}) @@ -230,7 +230,7 @@ def searchSubtitle(item): try: filetools.copy(path_video_temp, full_path_video_new) copy = True - logger.log("nuevo path =" + full_path_video_new) + logger.info("nuevo path =" + full_path_video_new) time.sleep(2) playlist = xbmc.PlayList(xbmc.PLAYLIST_VIDEO) playlist.clear() @@ -288,7 +288,7 @@ def get_from_subdivx(sub_url): :return: The path to the unzipped subtitle """ - logger.log() + logger.info() sub = '' sub_dir = os.path.join(config.get_data_path(), 'temp_subs') @@ -312,9 +312,9 @@ def get_from_subdivx(sub_url): filetools.write(filename, data_dl) sub = extract_file_online(sub_dir, filename) except: - logger.log('sub invalid') + logger.info('sub invalid') else: - logger.log('sub invalid') + logger.info('sub invalid') return sub @@ -328,7 +328,7 @@ def extract_file_online(path, filename): :return: """ - logger.log() + logger.info() url = "http://online.b1.org/rest/online/upload" diff --git a/platformcode/unify.py b/platformcode/unify.py index 71ccf803..ec90ffd3 100644 --- a/platformcode/unify.py +++ b/platformcode/unify.py @@ -98,7 +98,7 @@ thumb_dict = {"movies": "https://s10.postimg.cc/fxtqzdog9/peliculas.png", def set_genre(string): - # logger.log() + # logger.info() genres_dict = {'accion': ['accion', 'action', 'accion y aventura', 'action & adventure'], 'adultos': ['adultos', 'adultos +', 'adulto'], @@ -140,7 +140,7 @@ def set_genre(string): def remove_format(string): - # logger.log() + # logger.info() # logger.debug('enter remove: %s' % string) string = string.rstrip() string = re.sub(r'(\[|\[\/)(?:color|COLOR|b|B|i|I).*?\]|\[|\]|\(|\)|\:|\.', '', string) @@ -156,7 +156,7 @@ def normalize(string): def simplify(string): - # logger.log() + # logger.info() # logger.debug('enter simplify: %s'%string) string = remove_format(string) string = string.replace('-', ' ').replace('_', ' ') @@ -175,7 +175,7 @@ def simplify(string): def add_languages(title, languages): - # logger.log() + # logger.info() if isinstance(languages, list): for language in languages: @@ -186,7 +186,7 @@ def add_languages(title, languages): def add_info_plot(plot, languages, quality): - # logger.log() + # logger.info() last = '[/I][/B]\n' if languages: @@ -221,7 +221,7 @@ def add_info_plot(plot, languages, quality): def set_color(title, category): - # logger.log() + # logger.info() from core import jsontools styles_path = os.path.join(config.get_runtime_path(), 'resources', 'color_styles.json') @@ -262,7 +262,7 @@ def set_color(title, category): def set_lang(language): - # logger.log() + # logger.info() cast = ['castellano', 'español', 'espanol', 'cast', 'esp', 'espaol', 'es', 'zc', 'spa', 'spanish', 'vc'] ita = ['italiano', 'italian', 'ita', 'it'] @@ -303,7 +303,7 @@ def set_lang(language): def title_format(item): - # logger.log() + # logger.info() lang = False valid = True @@ -567,7 +567,7 @@ def title_format(item): def thumbnail_type(item): - # logger.log() + # logger.info() # Check what type of thumbnail will be used in findvideos, Poster or Logo of the server thumb_type = config.get_setting('video_thumbnail_type') diff --git a/platformcode/updater.py b/platformcode/updater.py index bf1c3265..2fde815e 100644 --- a/platformcode/updater.py +++ b/platformcode/updater.py @@ -34,7 +34,7 @@ changelogFile = "special://profile/addon_data/plugin.video.kod/changelog.txt" def loadCommits(page=1): apiLink = 'https://api.github.com/repos/' + user + '/' + repo + '/commits?sha=' + branch + "&page=" + str(page) - logger.log(apiLink) + logger.info(apiLink) # riprova ogni secondo finchè non riesce (ad esempio per mancanza di connessione) for n in range(10): try: @@ -54,7 +54,7 @@ def loadCommits(page=1): def check(background=False): if not addon.getSetting('addon_update_enabled'): return False, False - logger.log('Cerco aggiornamenti..') + logger.info('Cerco aggiornamenti..') commits = loadCommits() if not commits: return False, False @@ -66,7 +66,7 @@ def check(background=False): localCommitFile = open(os.path.join(addonDir, trackingFile), 'r+') localCommitSha = localCommitFile.read() localCommitSha = localCommitSha.replace('\n', '') # da testare - logger.log('Commit locale: ' + localCommitSha) + logger.info('Commit locale: ' + localCommitSha) updated = False serviceChanged = False @@ -91,7 +91,7 @@ def check(background=False): # evitiamo di applicare i merge commit if 'Merge' in commitJson['commit']['message']: continue - logger.log('aggiornando a ' + commitJson['sha']) + logger.info('aggiornando a ' + commitJson['sha']) # major update if len(commitJson['files']) > 50: @@ -112,7 +112,7 @@ def check(background=False): if file["filename"] == trackingFile: # il file di tracking non si modifica continue else: - logger.log(file["filename"]) + logger.info(file["filename"]) if 'resources/language' in file["filename"]: poFilesChanged = True if 'service.py' in file["filename"]: @@ -156,7 +156,7 @@ def check(background=False): elif changelog: platformtools.dialog_ok(config.get_localized_string(20000), config.get_localized_string(80041) + changelog) else: - logger.log('Nessun nuovo aggiornamento') + logger.info('Nessun nuovo aggiornamento') return updated, serviceChanged @@ -173,7 +173,7 @@ def showSavedChangelog(): def calcCurrHash(): treeHash = githash.tree_hash(addonDir).hexdigest() - logger.log('tree hash: ' + treeHash) + logger.info('tree hash: ' + treeHash) commits = loadCommits() lastCommitSha = commits[0]['sha'] page = 1 @@ -193,7 +193,7 @@ def calcCurrHash(): if found: break else: - logger.log('Non sono riuscito a trovare il commit attuale, scarico lo zip') + logger.info('Non sono riuscito a trovare il commit attuale, scarico lo zip') hash = updateFromZip() # se ha scaricato lo zip si trova di sicuro all'ultimo commit localCommitFile = open(os.path.join(xbmc.translatePath("special://home/addons/"), 'plugin.video.kod', trackingFile), 'w') @@ -228,9 +228,9 @@ def updateFromZip(message=config.get_localized_string(80050)): destpathname = xbmc.translatePath("special://home/addons/") extractedDir = filetools.join(destpathname, "addon-" + branch) - logger.log("remotefilename=%s" % remotefilename) - logger.log("localfilename=%s" % localfilename) - logger.log('extract dir: ' + extractedDir) + logger.info("remotefilename=%s" % remotefilename) + logger.info("localfilename=%s" % localfilename) + logger.info('extract dir: ' + extractedDir) # pulizia preliminare remove(localfilename) @@ -241,24 +241,24 @@ def updateFromZip(message=config.get_localized_string(80050)): lambda nb, bs, fs, url=remotefilename: _pbhook(nb, bs, fs, url, dp)) except Exception as e: platformtools.dialog_ok(config.get_localized_string(20000), config.get_localized_string(80031)) - logger.log('Non sono riuscito a scaricare il file zip') - logger.log(e) + logger.info('Non sono riuscito a scaricare il file zip') + logger.info(e) dp.close() return False # Lo descomprime - logger.log("decompressione...") - logger.log("destpathname=%s" % destpathname) + logger.info("decompressione...") + logger.info("destpathname=%s" % destpathname) if os.path.isfile(localfilename): - logger.log('il file esiste') + logger.info('il file esiste') dp.update(80, config.get_localized_string(20000) + '\n' + config.get_localized_string(80032)) import zipfile try: hash = fixZipGetHash(localfilename) - logger.log(hash) + logger.info(hash) with zipfile.ZipFile(filetools.file_open(localfilename, 'rb', vfs=False)) as zip: size = sum([zinfo.file_size for zinfo in zip.filelist]) @@ -269,7 +269,7 @@ def updateFromZip(message=config.get_localized_string(80050)): dp.update(int(80 + cur_size * 15 / size)) except Exception as e: - logger.log('Non sono riuscito ad estrarre il file zip') + logger.info('Non sono riuscito ad estrarre il file zip') logger.error(e) import traceback logger.error(traceback.print_exc()) @@ -289,7 +289,7 @@ def updateFromZip(message=config.get_localized_string(80050)): rename(extractedDir, 'plugin.video.kod') addonDir = filetools.join(destpathname, 'plugin.video.kod') - logger.log("Cancellando il file zip...") + logger.info("Cancellando il file zip...") remove(localfilename) dp.update(100) @@ -318,7 +318,7 @@ def remove(file): try: os.remove(file) except: - logger.log('File ' + file + ' NON eliminato') + logger.info('File ' + file + ' NON eliminato') def onerror(func, path, exc_info): @@ -345,7 +345,7 @@ def removeTree(dir): try: shutil.rmtree(dir, ignore_errors=False, onerror=onerror) except Exception as e: - logger.log('Cartella ' + dir + ' NON eliminata') + logger.info('Cartella ' + dir + ' NON eliminata') logger.error(e) @@ -353,7 +353,7 @@ def rename(dir1, dir2): try: filetools.rename(dir1, dir2, silent=True, vfs=False) except: - logger.log('cartella ' + dir1 + ' NON rinominata') + logger.info('cartella ' + dir1 + ' NON rinominata') # https://stackoverflow.com/questions/3083235/unzipping-file-results-in-badzipfile-file-is-not-a-zip-file diff --git a/platformcode/xbmc_config_menu.py b/platformcode/xbmc_config_menu.py index c7d91adf..f578e2bd 100644 --- a/platformcode/xbmc_config_menu.py +++ b/platformcode/xbmc_config_menu.py @@ -12,7 +12,7 @@ from past.utils import old_div from core import channeltools, servertools, scrapertools from platformcode import config, logger, platformtools -from core.support import log, dbg, match +from core.support import info, dbg, match class SettingsWindow(xbmcgui.WindowXMLDialog): @@ -141,7 +141,7 @@ class SettingsWindow(xbmcgui.WindowXMLDialog): """ def start(self, list_controls=None, dict_values=None, caption="", callback=None, item=None, custom_button=None, channelpath=None): - log() + info() # Media Path self.mediapath = os.path.join(config.get_runtime_path(), 'resources', 'skins', 'Default', 'media') diff --git a/platformcode/xbmc_info_window.py b/platformcode/xbmc_info_window.py index 22e75e1e..19148fa2 100644 --- a/platformcode/xbmc_info_window.py +++ b/platformcode/xbmc_info_window.py @@ -261,7 +261,7 @@ class InfoWindow(xbmcgui.WindowXMLDialog): return self.return_value def onClick(self, _id): - logger.log("onClick id=" + repr(_id)) + logger.info("onClick id=" + repr(_id)) if _id == ID_BUTTON_PREVIOUS and self.indexList > 0: self.indexList -= 1 self.get_scraper_data(self.listData[self.indexList]) @@ -281,7 +281,7 @@ class InfoWindow(xbmcgui.WindowXMLDialog): self.return_value = None def onAction(self, action): - logger.log("action=" + repr(action.getId())) + logger.info("action=" + repr(action.getId())) action = action.getId() # Find Focus diff --git a/platformcode/xbmc_videolibrary.py b/platformcode/xbmc_videolibrary.py index 2807262d..60518125 100644 --- a/platformcode/xbmc_videolibrary.py +++ b/platformcode/xbmc_videolibrary.py @@ -17,7 +17,7 @@ from xml.dom import minidom def mark_auto_as_watched(item, nfo_path=None, head_nfo=None, item_nfo=None): def mark_as_watched_subThread(item, nfo_path, head_nfo, item_nfo): - logger.log() + logger.info() # logger.debug("item:\n" + item.tostring('\n')) time_limit = time.time() + 30 @@ -99,7 +99,7 @@ def sync_trakt_addon(path_folder): """ Updates the values ​​of episodes seen if """ - logger.log() + logger.info() # if the addon exists we do the search if xbmc.getCondVisibility('System.HasAddon("script.trakt")'): # we import dependencies @@ -225,7 +225,7 @@ def sync_trakt_kodi(silent=True): notificacion = False xbmc.executebuiltin('RunScript(script.trakt,action=sync,silent=%s)' % silent) - logger.log("Synchronization with Trakt started") + logger.info("Synchronization with Trakt started") if notificacion: platformtools.dialog_notification(config.get_localized_string(20000), config.get_localized_string(60045), sound=False, time=2000) @@ -239,7 +239,7 @@ def mark_content_as_watched_on_kodi(item, value=1): @type value: int @param value: > 0 for seen, 0 for not seen """ - logger.log() + logger.info() # logger.debug("item:\n" + item.tostring('\n')) payload_f = '' @@ -311,7 +311,7 @@ def mark_season_as_watched_on_kodi(item, value=1): @type value: int @param value: > 0 for seen, 0 for not seen """ - logger.log() + logger.info() # logger.debug("item:\n" + item.tostring('\n')) # We can only mark the season as seen in the Kodi database if the database is local, in case of sharing database this functionality will not work @@ -345,7 +345,7 @@ def mark_content_as_watched_on_kod(path): @type str: path @param path: content folder to mark """ - logger.log() + logger.info() #logger.debug("path: " + path) FOLDER_MOVIES = config.get_setting("folder_movies") @@ -435,7 +435,7 @@ def get_data(payload): :return: """ import urllib.request, urllib.error - logger.log("payload: %s" % payload) + logger.info("payload: %s" % payload) # Required header for XBMC JSON-RPC calls, otherwise you'll get a 415 HTTP response code - Unsupported media type headers = {'content-type': 'application/json'} @@ -452,7 +452,7 @@ def get_data(payload): response = f.read() f.close() - logger.log("get_data: response %s" % response) + logger.info("get_data: response %s" % response) data = jsontools.load(response) except Exception as ex: template = "An exception of type %s occured. Arguments:\n%r" @@ -468,7 +468,7 @@ def get_data(payload): logger.error("error en xbmc.executeJSONRPC: %s" % message) data = ["error"] - logger.log("data: %s" % data) + logger.info("data: %s" % data) return data @@ -482,7 +482,7 @@ def update(folder_content=config.get_setting("folder_tvshows"), folder=""): @type folder: str @param folder: name of the folder to scan. """ - logger.log(folder) + logger.info(folder) payload = { "jsonrpc": "2.0", @@ -546,7 +546,7 @@ def set_content(content_type, silent=False, custom=False): @type content_type: str ('movie' o 'tvshow') @param content_type: content type to configure, series or movies """ - logger.log() + logger.info() continuar = True msg_text = "" videolibrarypath = config.get_setting("videolibrarypath") @@ -572,7 +572,7 @@ def set_content(content_type, silent=False, custom=False): try: # Install metadata.themoviedb.org xbmc.executebuiltin('InstallAddon(metadata.themoviedb.org)', True) - logger.log("Instalado el Scraper de películas de TheMovieDB") + logger.info("Instalado el Scraper de películas de TheMovieDB") except: pass @@ -626,7 +626,7 @@ def set_content(content_type, silent=False, custom=False): try: # Install metadata.tvdb.com xbmc.executebuiltin('InstallAddon(metadata.tvdb.com)', True) - logger.log("The TVDB series Scraper installed ") + logger.info("The TVDB series Scraper installed ") except: pass @@ -721,7 +721,7 @@ def set_content(content_type, silent=False, custom=False): strScraper = 'metadata.universal' path_settings = xbmc.translatePath("special://profile/addon_data/metadata.universal/settings.xml") if not os.path.exists(path_settings): - logger.log("%s: %s" % (content_type, path_settings + " doesn't exist")) + logger.info("%s: %s" % (content_type, path_settings + " doesn't exist")) return continuar settings_data = filetools.read(path_settings) strSettings = ' '.join(settings_data.split()).replace("> <", "><") @@ -740,7 +740,7 @@ def set_content(content_type, silent=False, custom=False): strScraper = 'metadata.tvshows.themoviedb.org' path_settings = xbmc.translatePath("special://profile/addon_data/metadata.tvshows.themoviedb.org/settings.xml") if not os.path.exists(path_settings): - logger.log("%s: %s" % (content_type, path_settings + " doesn't exist")) + logger.info("%s: %s" % (content_type, path_settings + " doesn't exist")) return continuar settings_data = filetools.read(path_settings) strSettings = ' '.join(settings_data.split()).replace("> <", "><") @@ -750,7 +750,7 @@ def set_content(content_type, silent=False, custom=False): videolibrarypath += sep strPath = videolibrarypath + config.get_setting("folder_tvshows") + sep - logger.log("%s: %s" % (content_type, strPath)) + logger.info("%s: %s" % (content_type, strPath)) # We check if strPath already exists in the DB to avoid duplicates sql = 'SELECT idPath FROM path where strPath="%s"' % strPath nun_records, records = execute_sql_kodi(sql) @@ -792,15 +792,15 @@ def set_content(content_type, silent=False, custom=False): heading = config.get_localized_string(70103) % content_type msg_text = config.get_localized_string(70104) - logger.log("%s: %s" % (heading, msg_text)) + logger.info("%s: %s" % (heading, msg_text)) return continuar def update_db(old_path, new_path, old_movies_folder, new_movies_folder, old_tvshows_folder, new_tvshows_folder, progress): def path_replace(path, old, new): - logger.log() - logger.log('path: ' + path + ', old: ' + old + ', new: ' + new) + logger.info() + logger.info('path: ' + path + ', old: ' + old + ', new: ' + new) if new.startswith("special://") or '://' in new: sep = '/' else: sep = os.sep @@ -811,7 +811,7 @@ def update_db(old_path, new_path, old_movies_folder, new_movies_folder, old_tvsh return path - logger.log() + logger.info() sql_old_path = old_path if sql_old_path.startswith("special://"): @@ -823,10 +823,10 @@ def update_db(old_path, new_path, old_movies_folder, new_movies_folder, old_tvsh if not sql_old_path.endswith(sep): sql_old_path += sep - logger.log('sql_old_path: ' + sql_old_path) + logger.info('sql_old_path: ' + sql_old_path) # search MAIN path in the DB sql = 'SELECT idPath, strPath FROM path where strPath LIKE "%s"' % sql_old_path - logger.log('sql: ' + sql) + logger.info('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) # change main path @@ -834,7 +834,7 @@ def update_db(old_path, new_path, old_movies_folder, new_movies_folder, old_tvsh idPath = records[0][0] strPath = path_replace(records[0][1], old_path, new_path) sql = 'UPDATE path SET strPath="%s" WHERE idPath=%s' % (strPath, idPath) - logger.log('sql: ' + sql) + logger.info('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) else: progress.update(100) @@ -851,7 +851,7 @@ def update_db(old_path, new_path, old_movies_folder, new_movies_folder, old_tvsh # Search Main Sub Folder sql = 'SELECT idPath, strPath FROM path where strPath LIKE "%s"' % sql_old_folder - logger.log('sql: ' + sql) + logger.info('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) # Change Main Sub Folder @@ -860,13 +860,13 @@ def update_db(old_path, new_path, old_movies_folder, new_movies_folder, old_tvsh idPath = record[0] strPath = path_replace(record[1], filetools.join(old_path, OldFolder), filetools.join(new_path, NewFolder)) sql = 'UPDATE path SET strPath="%s" WHERE idPath=%s' % (strPath, idPath) - logger.log('sql: ' + sql) + logger.info('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) # Search if Sub Folder exixt in all paths sql_old_folder += '%' sql = 'SELECT idPath, strPath FROM path where strPath LIKE "%s"' % sql_old_folder - logger.log('sql: ' + sql) + logger.info('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) #Change Sub Folder in all paths @@ -875,7 +875,7 @@ def update_db(old_path, new_path, old_movies_folder, new_movies_folder, old_tvsh idPath = record[0] strPath = path_replace(record[1], filetools.join(old_path, OldFolder), filetools.join(new_path, NewFolder)) sql = 'UPDATE path SET strPath="%s" WHERE idPath=%s' % (strPath, idPath) - logger.log('sql: ' + sql) + logger.info('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) @@ -883,27 +883,27 @@ def update_db(old_path, new_path, old_movies_folder, new_movies_folder, old_tvsh # if is Movie Folder # search and modify in "movie" sql = 'SELECT idMovie, c22 FROM movie where c22 LIKE "%s"' % sql_old_folder - logger.log('sql: ' + sql) + logger.info('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) if records: for record in records: idMovie = record[0] strPath = path_replace(record[1], filetools.join(old_path, OldFolder), filetools.join(new_path, NewFolder)) sql = 'UPDATE movie SET c22="%s" WHERE idMovie=%s' % (strPath, idMovie) - logger.log('sql: ' + sql) + logger.info('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) else: # if is TV Show Folder # search and modify in "episode" sql = 'SELECT idEpisode, c18 FROM episode where c18 LIKE "%s"' % sql_old_folder - logger.log('sql: ' + sql) + logger.info('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) if records: for record in records: idEpisode = record[0] strPath = path_replace(record[1], filetools.join(old_path, OldFolder), filetools.join(new_path, NewFolder)) sql = 'UPDATE episode SET c18="%s" WHERE idEpisode=%s' % (strPath, idEpisode) - logger.log('sql: ' + sql) + logger.info('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) p += 5 progress.update(p, config.get_localized_string(20000) + '\n' + config.get_localized_string(80013)) @@ -928,26 +928,26 @@ def clean(path_list=[]): return path, sep - logger.log() + logger.info() progress = platformtools.dialog_progress_bg(config.get_localized_string(20000), config.get_localized_string(80025)) progress.update(0) # if the path list is empty, clean the entire video library if not path_list: - logger.log('the path list is empty, clean the entire video library') + logger.info('the path list is empty, clean the entire video library') if not config.get_setting("videolibrary_kodi"): sql_path, sep = sql_format(config.get_setting("videolibrarypath")) if not sql_path.endswith(sep): sql_path += sep sql = 'SELECT idPath FROM path where strPath LIKE "%s"' % sql_path - logger.log('sql: ' + sql) + logger.info('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) idPath = records[0][0] sql = 'DELETE from path WHERE idPath=%s' % idPath - logger.log('sql: ' + sql) + logger.info('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) sql = 'DELETE from path WHERE idParentPath=%s' % idPath - logger.log('sql: ' + sql) + logger.info('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) from core import videolibrarytools @@ -961,7 +961,7 @@ def clean(path_list=[]): if filetools.exists(tvshow_nfo): path_list.append(filetools.join(config.get_setting("videolibrarypath"), videolibrarytools.FOLDER_TVSHOWS, folder)) - logger.log('path_list: ' + str(path_list)) + logger.info('path_list: ' + str(path_list)) if path_list: t = float(100) / len(path_list) for i, path in enumerate(path_list): progress.update(int(math.ceil((i + 1) * t))) @@ -971,13 +971,13 @@ def clean(path_list=[]): sql_path, sep = sql_format(path) if filetools.isdir(path) and not sql_path.endswith(sep): sql_path += sep - logger.log('path: ' + path) - logger.log('sql_path: ' + sql_path) + logger.info('path: ' + path) + logger.info('sql_path: ' + sql_path) if filetools.isdir(path): # search movie in the DB sql = 'SELECT idMovie FROM movie where c22 LIKE "%s"' % (sql_path + '%') - logger.log('sql: ' + sql) + logger.info('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) # delete movie if records: @@ -986,7 +986,7 @@ def clean(path_list=[]): continue # search TV show in the DB sql = 'SELECT idShow FROM tvshow_view where strPath LIKE "%s"' % sql_path - logger.log('sql: ' + sql) + logger.info('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) # delete TV show if records: @@ -995,7 +995,7 @@ def clean(path_list=[]): elif config.get_setting("folder_movies") in sql_path: # search movie in the DB sql = 'SELECT idMovie FROM movie where c22 LIKE "%s"' % sql_path - logger.log('sql: ' + sql) + logger.info('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) # delete movie if records: @@ -1004,7 +1004,7 @@ def clean(path_list=[]): else: # search episode in the DB sql = 'SELECT idEpisode FROM episode where c18 LIKE "%s"' % sql_path - logger.log('sql: ' + sql) + logger.info('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) # delete episode if records: @@ -1023,7 +1023,7 @@ def check_db(path): ret = False sql_path = '%' + sep + path.split(sep)[-1] + sep + '%' sql = 'SELECT idShow FROM tvshow_view where strPath LIKE "%s"' % sql_path - logger.log('sql: ' + sql) + logger.info('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) if records: ret = True @@ -1040,7 +1040,7 @@ def execute_sql_kodi(sql): @return: list with the query result @rtype records: list of tuples """ - logger.log() + logger.info() file_db = "" nun_records = 0 records = None @@ -1061,14 +1061,14 @@ def execute_sql_kodi(sql): break if file_db: - logger.log("DB file: %s" % file_db) + logger.info("DB file: %s" % file_db) conn = None try: import sqlite3 conn = sqlite3.connect(file_db) cursor = conn.cursor() - logger.log("Running sql: %s" % sql) + logger.info("Running sql: %s" % sql) cursor.execute(sql) conn.commit() @@ -1082,7 +1082,7 @@ def execute_sql_kodi(sql): nun_records = conn.total_changes conn.close() - logger.log("Query executed. Records: %s" % nun_records) + logger.info("Query executed. Records: %s" % nun_records) except: logger.error("Error executing sql query") @@ -1102,7 +1102,7 @@ def check_sources(new_movies_path='', new_tvshows_path=''): if not path.endswith(sep): path += sep return path - logger.log() + logger.info() new_movies_path = format_path(new_movies_path) new_tvshows_path = format_path(new_tvshows_path) @@ -1132,7 +1132,7 @@ def check_sources(new_movies_path='', new_tvshows_path=''): def update_sources(new='', old=''): - logger.log() + logger.info() if new == old: return SOURCES_PATH = xbmc.translatePath("special://userdata/sources.xml") @@ -1174,9 +1174,9 @@ def update_sources(new='', old=''): # create new path list_path = [p.firstChild.data for p in paths_node] if new in list_path: - logger.log("The path %s already exists in sources.xml" % new) + logger.info("The path %s already exists in sources.xml" % new) return - logger.log("The path %s does not exist in sources.xml" % new) + logger.info("The path %s does not exist in sources.xml" % new) # if the path does not exist we create one source_node = xmldoc.createElement("source") @@ -1215,7 +1215,7 @@ def update_sources(new='', old=''): def ask_set_content(silent=False): - logger.log() + logger.info() logger.debug("videolibrary_kodi %s" % config.get_setting("videolibrary_kodi")) def do_config(custom=False): @@ -1272,7 +1272,7 @@ def ask_set_content(silent=False): def next_ep(item): from core.item import Item - logger.log() + logger.info() item.next_ep = False # check if next file exist @@ -1288,7 +1288,7 @@ def next_ep(item): nextIndex = fileList.index(current_filename) + 1 if nextIndex == 0 or nextIndex == len(fileList): next_file = None else: next_file = fileList[nextIndex] - logger.log('Next File:' + str(next_file)) + logger.info('Next File:' + str(next_file)) # start next episode window afther x time if next_file: diff --git a/servers/akvideo.py b/servers/akvideo.py index bf5416ec..14ff3836 100644 --- a/servers/akvideo.py +++ b/servers/akvideo.py @@ -9,7 +9,7 @@ from platformcode import logger, config headers = [['User-Agent', 'Mozilla/5.0 (Windows NT 6.1; WOW64; rv:54.0) Gecko/20100101 Firefox/54.0']] def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) # page_url = re.sub('akvideo.stream/(?:video/|video\\.php\\?file_code=)?(?:embed-)?([a-zA-Z0-9]+)','akvideo.stream/video/\\1',page_url) global data page = httptools.downloadpage(page_url, headers=headers) @@ -28,18 +28,18 @@ def test_video_exists(page_url): # ID, code = scrapertools.find_single_match(data, r"""input\D*id=(?:'|")([^'"]+)(?:'|").*?value='([a-z0-9]+)""") # post = urllib.urlencode({ID: code}) - # logger.log('PAGE DATA' + data) + # logger.info('PAGE DATA' + data) if "File Not Found" in data: return False, config.get_localized_string(70449) % "Akvideo" return True, "" def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log(" url=" + page_url) + logger.info(" url=" + page_url) video_urls = [] global data - # logger.log('PAGE DATA' + data) + # logger.info('PAGE DATA' + data) # sitekey = scrapertools.find_single_match(data, 'data-sitekey="([^"]+)') # captcha = platformtools.show_recaptcha(sitekey, page_url) if sitekey else '' # diff --git a/servers/anavids.py b/servers/anavids.py index 34462c9f..c4169620 100644 --- a/servers/anavids.py +++ b/servers/anavids.py @@ -6,7 +6,7 @@ from platformcode import config, logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url, cookies=False).data if 'File you are looking for is not found.' in data: diff --git a/servers/animeid.py b/servers/animeid.py index 56eaa3bd..3cec4070 100644 --- a/servers/animeid.py +++ b/servers/animeid.py @@ -6,7 +6,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "no longer exists" in data or "to copyright issues" in data: return False, config.get_localized_string(70449) % "animeid" @@ -16,7 +16,7 @@ def test_video_exists(page_url): def get_video_url(page_url, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data video_urls = [] label, videourl = scrapertools.find_single_match(data, 'label":"([^"]+)".*?file":"([^"]+)') diff --git a/servers/anonfile.py b/servers/anonfile.py index ea1070bb..4191caff 100644 --- a/servers/anonfile.py +++ b/servers/anonfile.py @@ -9,7 +9,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) response = httptools.downloadpage(page_url) if not response.success or "Not Found" in response.data or "File was deleted" in response.data or "is no longer available" in response.data: return False, config.get_localized_string(70449) % "anonfile" @@ -17,7 +17,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] data = httptools.downloadpage(page_url).data patron = 'download-url.*?href="([^"]+)"' diff --git a/servers/archiveorg.py b/servers/archiveorg.py index ee9a51cd..61677b02 100644 --- a/servers/archiveorg.py +++ b/servers/archiveorg.py @@ -9,7 +9,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url) if data.code == 404: return False, config.get_localized_string(70449) % "ArchiveOrg" @@ -17,7 +17,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] data = httptools.downloadpage(page_url).data patron = '<meta property="og:video" content="([^"]+)">' diff --git a/servers/backin.py b/servers/backin.py index fa6c4d9e..c070d5d6 100644 --- a/servers/backin.py +++ b/servers/backin.py @@ -9,7 +9,7 @@ except ImportError: from urllib import urlencode def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) if 'http://' in page_url: # fastids page_url = httptools.downloadpage(page_url, follow_redirects=False, only_headers=True).headers['location'] @@ -24,7 +24,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("page_url=" + page_url) + logger.info("page_url=" + page_url) video_urls = [] @@ -36,18 +36,18 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= if data_pack: from lib import jsunpack data = jsunpack.unpack(data_pack) - logger.log("page_url=" + data) + logger.info("page_url=" + data) # URL url = scrapertools.find_single_match(data, r'"src"value="([^"]+)"') if not url: url = scrapertools.find_single_match(data, r'file\s*:\s*"([^"]+)"') - logger.log("URL=" + str(url)) + logger.info("URL=" + str(url)) # URL del vídeo video_urls.append([".mp4" + " [backin]", url]) for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], httptools.get_url_headers(video_url[1]))) + logger.info("%s - %s" % (video_url[0], httptools.get_url_headers(video_url[1]))) return video_urls diff --git a/servers/badshare.py b/servers/badshare.py index 3e05b115..78c62986 100644 --- a/servers/badshare.py +++ b/servers/badshare.py @@ -11,7 +11,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global page page = httptools.downloadpage(page_url) if not page.success: @@ -20,7 +20,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] ext = '.mp4' diff --git a/servers/bdupload.py b/servers/bdupload.py index 79c40c74..6321dd3a 100644 --- a/servers/bdupload.py +++ b/servers/bdupload.py @@ -10,7 +10,7 @@ headers = {'User-Agent': 'Mozilla/5.0 (Linux; Android 6.0; Nexus 5 Build/MRA58N) def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "Archive no Encontrado" in data: return False, config.get_localized_string(70449) % "bdupload" @@ -19,7 +19,7 @@ def test_video_exists(page_url): def get_video_url(page_url, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data post = "" patron = '(?s)type="hidden" name="([^"]+)".*?value="([^"]*)"' diff --git a/servers/cinemaupload.py b/servers/cinemaupload.py index 8d1af217..c3f05da8 100644 --- a/servers/cinemaupload.py +++ b/servers/cinemaupload.py @@ -11,7 +11,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url) if data.code == 404: return False, config.get_localized_string(70449) % "CinemaUpload" @@ -19,7 +19,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] data = httptools.downloadpage(page_url).data data = re.sub(r'\n|\r|\t| |<br>|\s{2,}', "", data) diff --git a/servers/clicknupload.py b/servers/clicknupload.py index 1f40f6f1..f54c5d5c 100755 --- a/servers/clicknupload.py +++ b/servers/clicknupload.py @@ -22,7 +22,7 @@ excption = False def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = get_data(page_url.replace(".org", ".me")) if "File Not Found" in data: return False, config.get_localized_string(70449) % "Clicknupload" @@ -31,7 +31,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) data = get_data(page_url.replace(".org", ".me")) @@ -51,7 +51,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= media_url = media.rsplit('/', 1)[0] + "/" + url_strip video_urls.append([scrapertools.get_filename_from_url(media_url)[-4:] + " [clicknupload]", media_url]) for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/clipwatching.py b/servers/clipwatching.py index 7653b330..93aa4804 100644 --- a/servers/clipwatching.py +++ b/servers/clipwatching.py @@ -6,7 +6,7 @@ from lib import jsunpack from platformcode import logger, config def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url).data if "File Not Found" in data or "File was deleted" in data: @@ -15,7 +15,7 @@ def test_video_exists(page_url): def get_video_url(page_url, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] try: diff --git a/servers/cloudvideo.py b/servers/cloudvideo.py index 0dd0b0f8..b7885afe 100644 --- a/servers/cloudvideo.py +++ b/servers/cloudvideo.py @@ -8,7 +8,7 @@ from lib import jsunpack def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) html = httptools.downloadpage(page_url) global data data = html.data @@ -18,7 +18,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] global data # data = httptools.downloadpage(page_url).data diff --git a/servers/crunchyroll.py b/servers/crunchyroll.py index 8c9eec30..2d252b3a 100755 --- a/servers/crunchyroll.py +++ b/servers/crunchyroll.py @@ -30,7 +30,7 @@ proxy = "https://www.usa-proxy.org/" def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url, headers=GLOBAL_HEADER).data if "Este es un clip de muestra" in data: @@ -44,7 +44,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): #page_url='https://www.crunchyroll.com/es-es/one-piece/episode-891-climbing-up-a-waterfall-a-great-journey-through-the-land-of-wanos-sea-zone-786643' - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] if "crunchyroll.com" in page_url: media_id = page_url.rsplit("-", 1)[1] @@ -94,7 +94,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= file_sub = "" video_urls.append(["%s %sp [crunchyroll]" % (filename, quality), media_url, 0, file_sub]) for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/dailymotion.py b/servers/dailymotion.py index ad72f92f..7de4b4fd 100644 --- a/servers/dailymotion.py +++ b/servers/dailymotion.py @@ -6,7 +6,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global response response = httptools.downloadpage(page_url, cookies=False) @@ -18,7 +18,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] cookie = {'Cookie': response.headers["set-cookie"]} data = response.data.replace("\\", "") @@ -40,5 +40,5 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= stream_url = stream_url_http video_urls.append(["%sp .%s [dailymotion]" % (calidad, stream_type), stream_url, 0, subtitle]) for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls \ No newline at end of file diff --git a/servers/debriders/alldebrid.py b/servers/debriders/alldebrid.py index db6e3010..6f0bd18a 100644 --- a/servers/debriders/alldebrid.py +++ b/servers/debriders/alldebrid.py @@ -7,7 +7,7 @@ from platformcode import logger # Returns an array of possible video url's from the page_url def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log() + logger.info() page_url = correct_url(page_url) dd1 = httptools.downloadpage("https://api.alldebrid.com/user/login?agent=mySoft&username=%s&password=%s" %(user, password)).data token = scrapertools.find_single_match(dd1, 'token":"([^"]+)') diff --git a/servers/debriders/realdebrid.py b/servers/debriders/realdebrid.py index 558ddbc9..7d8b855d 100755 --- a/servers/debriders/realdebrid.py +++ b/servers/debriders/realdebrid.py @@ -22,7 +22,7 @@ headers = {'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; WOW64; rv:65.0) Gecko/20 # Returns an array of possible video url's from the page_url def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s' , video_password=%s)" % (page_url, video_password)) + logger.info("(page_url='%s' , video_password=%s)" % (page_url, video_password)) page_url = page_url.replace(".nz/embed", ".nz/") # Se comprueba si existe un token guardado y sino se ejecuta el proceso de autentificación token_auth = config.get_setting("token", server="realdebrid") @@ -99,7 +99,7 @@ def get_enlaces(data): def authentication(): - logger.log() + logger.info() try: client_id = "YTWNFBIJEEBP6" diff --git a/servers/decrypters/adfly.py b/servers/decrypters/adfly.py index e9846a53..20b34d18 100755 --- a/servers/decrypters/adfly.py +++ b/servers/decrypters/adfly.py @@ -8,7 +8,7 @@ from platformcode import logger def get_long_url(short_url): - logger.log("short_url = '%s'" % short_url) + logger.info("short_url = '%s'" % short_url) data = httptools.downloadpage(short_url).data ysmm = scrapertools.find_single_match(data, "var ysmm = '([^']+)';") diff --git a/servers/decrypters/linkbucks.py b/servers/decrypters/linkbucks.py index b9af3d30..5b15ec00 100755 --- a/servers/decrypters/linkbucks.py +++ b/servers/decrypters/linkbucks.py @@ -17,7 +17,7 @@ from platformcode import logger # Obtiene la URL que hay detrás de un enlace a linkbucks def get_long_url(short_url): - logger.log("(short_url='%s')" % short_url) + logger.info("(short_url='%s')" % short_url) request_headers = [] request_headers.append(["User-Agent", @@ -33,17 +33,17 @@ def get_long_url(short_url): while True: for name, value in response_headers: if name == "set-cookie": - logger.log("Set-Cookie: " + value) + logger.info("Set-Cookie: " + value) cookie_name = scrapertools.scrapertools.find_single_match(value, '(.*?)\=.*?\;') cookie_value = scrapertools.scrapertools.find_single_match(value, '.*?\=(.*?)\;') request_headers.append(["Cookie", cookie_name + "=" + cookie_value]) body, response_headers = scrapertools.read_body_and_headers(url, headers=request_headers) - logger.log("body=" + body) + logger.info("body=" + body) try: location = scrapertools.scrapertools.find_single_match(body, '<textarea.*?class="caja_des">([^<]+)</textarea>') - logger.log("location=" + location) + logger.info("location=" + location) break except: n = n + 1 diff --git a/servers/decrypters/longurl.py b/servers/decrypters/longurl.py index 18f7cd38..4c6cb012 100755 --- a/servers/decrypters/longurl.py +++ b/servers/decrypters/longurl.py @@ -38,15 +38,15 @@ servers = get_server_list() def get_long_urls(data): - logger.log() + logger.info() patron = '<a href="http://([^"]+)"' matches = re.compile(patron, re.DOTALL).findall(data) for short_url in matches: if short_url.startswith(tuple(servers)): - logger.log(": " + short_url) + logger.info(": " + short_url) longurl_data = httptools.downloadpage( "http://api.longurl.org/v2/expand?url=" + urllib.quote_plus(short_url)).data - logger.log(longurl_data) + logger.info(longurl_data) try: long_url = scrapertools.scrapertools.find_single_match(longurl_data, '<long-url><!\[CDATA\[(.*?)\]\]></long-url>') except: diff --git a/servers/decrypters/safelinking.py b/servers/decrypters/safelinking.py index 0596789b..912177e0 100644 --- a/servers/decrypters/safelinking.py +++ b/servers/decrypters/safelinking.py @@ -5,9 +5,9 @@ from platformcode import logger def get_long_url(short_url): - logger.log("(short_url='%s')" % short_url) + logger.info("(short_url='%s')" % short_url) location = scrapertools.get_header_from_response(short_url, header_to_get="location") - logger.log("location=" + location) + logger.info("location=" + location) return location diff --git a/servers/deltabit.py b/servers/deltabit.py index 2a06ca89..957dca72 100644 --- a/servers/deltabit.py +++ b/servers/deltabit.py @@ -12,7 +12,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url).data.replace('"', "'") if "Not Found" in data or "File Does not Exist" in data: @@ -20,7 +20,7 @@ def test_video_exists(page_url): return True, "" def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(deltabit page_url='%s')" % page_url) + logger.info("(deltabit page_url='%s')" % page_url) video_urls = [] global data diff --git a/servers/directo.py b/servers/directo.py index 508c9f38..1efbe7fc 100644 --- a/servers/directo.py +++ b/servers/directo.py @@ -8,7 +8,7 @@ def test_video_exists(page_url): # Returns an array of possible video url's from the page_url def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [["%s %s" % (page_url[-4:], config.get_localized_string(30137)), page_url]] diff --git a/servers/dostream.py b/servers/dostream.py index 4fbad488..1fdc4883 100644 --- a/servers/dostream.py +++ b/servers/dostream.py @@ -9,7 +9,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url) if data.code == 404: return False, config.get_localized_string(70449) % "Dostream" @@ -17,7 +17,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] data = httptools.downloadpage(page_url, headers={"Referer":page_url}).data patron = '"label":"([^"]+)".*?' diff --git a/servers/downace.py b/servers/downace.py index 333bbb8e..a238f2de 100644 --- a/servers/downace.py +++ b/servers/downace.py @@ -6,7 +6,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "no longer exists" in data or "to copyright issues" in data: return False, "[Downace] El video ha sido borrado" @@ -18,7 +18,7 @@ def test_video_exists(page_url): def get_video_url(page_url, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data video_urls = [] videourl = scrapertools.find_single_match(data, 'controls preload.*?src="([^"]+)') diff --git a/servers/facebook.py b/servers/facebook.py index d60563d1..15721d9a 100755 --- a/servers/facebook.py +++ b/servers/facebook.py @@ -19,10 +19,10 @@ from platformcode import logger def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) page_url = page_url.replace("amp;", "") data = httptools.downloadpage(page_url).data - logger.log("data=" + data) + logger.info("data=" + data) video_urls = [] patron = "video_src.*?(http.*?)%22%2C%22video_timestamp" matches = re.compile(patron, re.DOTALL).findall(data) @@ -33,5 +33,5 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= videourl = urllib.unquote(videourl) video_urls.append(["[facebook]", videourl]) for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/fastplay.py b/servers/fastplay.py index 2e0c7959..74468000 100644 --- a/servers/fastplay.py +++ b/servers/fastplay.py @@ -7,7 +7,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url) if "Object not found" in data.data or "longer exists on our servers" in data.data: @@ -18,7 +18,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "p,a,c,k,e,d" in data: @@ -38,6 +38,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= except: pass for video_url in video_urls: - logger.log(" %s - %s" % (video_url[0], video_url[1])) + logger.info(" %s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/fembed.py b/servers/fembed.py index 74bab813..f11a5d88 100644 --- a/servers/fembed.py +++ b/servers/fembed.py @@ -6,7 +6,7 @@ from core import jsontools from platformcode import logger, config def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global data page_url = re.sub('://[^/]+/', '://feurl.com/', page_url) @@ -16,14 +16,14 @@ def test_video_exists(page_url): page_url = page_url.replace("/f/","/v/") page_url = page_url.replace("/v/","/api/source/") data = httptools.downloadpage(page_url, post={}).json - logger.log(data) + logger.info(data) if "Video not found or" in data or "We are encoding this video" in data: return False, config.get_localized_string(70449) % "Fembed" return True, "" def get_video_url(page_url, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] for file in data['data']: media_url = file['file'] diff --git a/servers/fex.py b/servers/fex.py index 680f3151..d64e6434 100644 --- a/servers/fex.py +++ b/servers/fex.py @@ -8,7 +8,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url, follow_redirects=False) @@ -18,7 +18,7 @@ def test_video_exists(page_url): return True, "" def get_video_url(page_url, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] data = httptools.downloadpage(page_url, follow_redirects=False, only_headers=True) logger.debug(data.headers) diff --git a/servers/filefactory.py b/servers/filefactory.py index 1dbd46f0..75e75a53 100644 --- a/servers/filefactory.py +++ b/servers/filefactory.py @@ -4,12 +4,12 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) return True, "" def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] return video_urls diff --git a/servers/filepup.py b/servers/filepup.py index 96638aec..0636c120 100644 --- a/servers/filepup.py +++ b/servers/filepup.py @@ -9,7 +9,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) response = httptools.downloadpage(page_url) if "File was deleted" in response.data or "is no longer available" in response.data: return False, config.get_localized_string(70449) % "filepup" @@ -17,7 +17,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] page_url = page_url.replace("https","http") + "?wmode=transparent" data = httptools.downloadpage(page_url).data @@ -36,5 +36,5 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls.sort(key=lambda x: x[2]) for video_url in video_urls: video_url[2] = 0 - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/filescdn.py b/servers/filescdn.py index 61af3d86..3a260b47 100644 --- a/servers/filescdn.py +++ b/servers/filescdn.py @@ -6,7 +6,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "File was deleted" in data: @@ -16,7 +16,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] data = httptools.downloadpage(page_url).data url = scrapertools.find_single_match(data, '(?i)link:\s*"(https://.*?filescdn\.com.*?mp4)"') diff --git a/servers/filesmonster.py b/servers/filesmonster.py index 086f8ae3..7a1cb05d 100644 --- a/servers/filesmonster.py +++ b/servers/filesmonster.py @@ -6,7 +6,7 @@ from platformcode import logger def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("( page_url='%s')") + logger.info("( page_url='%s')") video_urls = [] itemlist = [] data1 = '' diff --git a/servers/flashx.py b/servers/flashx.py index 8ba47d51..8ce5d966 100644 --- a/servers/flashx.py +++ b/servers/flashx.py @@ -23,7 +23,7 @@ flashx_hash_f = "" flashx_post = "" def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global flashx_data try: flashx_data = httptools.downloadpage(page_url, cookies="xfsts=pfp5dj3e6go1l2o1").data @@ -53,7 +53,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) pfxfx = "" data = flashx_data data = data.replace("\n", "") @@ -70,8 +70,8 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= matches = scrapertools.find_multiple_matches(mfxfx, '(\w+):(\w+)') for f, v in matches: pfxfx += f + "=" + v + "&" - logger.log("mfxfxfx1= %s" % js_fxfx) - logger.log("mfxfxfx2= %s" % pfxfx) + logger.info("mfxfxfx1= %s" % js_fxfx) + logger.info("mfxfxfx2= %s" % pfxfx) if pfxfx == "": pfxfx = "f=fail&fxfx=6" coding_url = 'https://www.flashx.co/flashx.php?%s' % pfxfx @@ -119,14 +119,14 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= filetools.write(subtitle, data) except: import traceback - logger.log("Error al descargar el subtítulo: " + traceback.format_exc()) + logger.info("Error al descargar el subtítulo: " + traceback.format_exc()) for media_url, label in media_urls: if not media_url.endswith("png") and not media_url.endswith(".srt"): video_urls.append(["." + media_url.rsplit('.', 1)[1] + " [flashx]", media_url, 0, subtitle]) for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) except: pass diff --git a/servers/fourshared.py b/servers/fourshared.py index 0ad76c90..802fa1d3 100644 --- a/servers/fourshared.py +++ b/servers/fourshared.py @@ -8,7 +8,7 @@ from platformcode import logger # Returns an array of possible video url's from the page_url def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] @@ -17,7 +17,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= page_url = scrapertools.get_header_from_response(page_url, header_to_get="location") # http://www.4shared.com/flash/player.swf?file=http://dc237.4shared.com/img/392975628/ff297d3f/dlink__2Fdownload_2Flj9Qu-tF_3Ftsid_3D20101030-200423-87e3ba9b/preview.flv&d - logger.log("redirect a '%s'" % page_url) + logger.info("redirect a '%s'" % page_url) patron = "file\=([^\&]+)\&" matches = re.compile(patron, re.DOTALL).findall(page_url) @@ -29,6 +29,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls.append(["[fourshared]", page_url]) for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/gamovideo.py b/servers/gamovideo.py index 6f3e0fcb..7a4e6f20 100755 --- a/servers/gamovideo.py +++ b/servers/gamovideo.py @@ -22,7 +22,7 @@ headers = {"User-Agent":"Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:%s.0) Geck DATA = '' def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = alfaresolver.get_data(page_url, False) @@ -46,7 +46,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = DATA @@ -87,7 +87,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls.append([scrapertools.get_filename_from_url(mediaurl)[-4:] + " [gamovideo]", mediaurl]) for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/gigasize.py b/servers/gigasize.py index 400d7968..85faa514 100644 --- a/servers/gigasize.py +++ b/servers/gigasize.py @@ -5,7 +5,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if '<h2 class="error">Download error</h2>' in data: return False, "El enlace no es válido<br/>o ha sido borrado de gigasize" @@ -13,7 +13,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] return video_urls diff --git a/servers/googlevideo.py b/servers/googlevideo.py index e324062b..3a411e80 100755 --- a/servers/googlevideo.py +++ b/servers/googlevideo.py @@ -7,14 +7,14 @@ from platformcode import logger def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] # Lo extrae a partir de flashvideodownloader.org if page_url.startswith("http://"): url = 'http://www.flashvideodownloader.org/download.php?u=' + page_url else: url = 'http://www.flashvideodownloader.org/download.php?u=http://video.google.com/videoplay?docid=' + page_url - logger.log("url=" + url) + logger.info("url=" + url) data = httptools.downloadpage(url).data # Extrae el vídeo @@ -24,6 +24,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls.append(["[googlevideo]", newmatches[0]]) for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/gounlimited.py b/servers/gounlimited.py index 5b71a74e..8bba60b4 100644 --- a/servers/gounlimited.py +++ b/servers/gounlimited.py @@ -20,14 +20,14 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] global data data = re.sub(r'"|\n|\r|\t| |<br>|\s{2,}', "", data) - # logger.log('GOUN DATA= '+data) + # logger.info('GOUN DATA= '+data) packed_data = scrapertools.find_single_match(data, "javascript'>(eval.*?)</script>") unpacked = jsunpack.unpack(packed_data) - # logger.log('GOUN DATA= '+unpacked) + # logger.info('GOUN DATA= '+unpacked) patron = r"sources..([^\]]+)" matches = re.compile(patron, re.DOTALL).findall(unpacked) if not matches: diff --git a/servers/gvideo.py b/servers/gvideo.py index adc6e52e..aa9712c4 100644 --- a/servers/gvideo.py +++ b/servers/gvideo.py @@ -42,7 +42,7 @@ def test_video_exists(page_url): def get_video_url(page_url, user="", password="", video_password=""): - logger.log() + logger.info() video_urls = [] urls = [] streams =[] diff --git a/servers/hdload.py b/servers/hdload.py index 9cce3157..30240022 100644 --- a/servers/hdload.py +++ b/servers/hdload.py @@ -7,7 +7,7 @@ from platformcode import config, logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url, cookies=False).data if 'Not found id' in data: @@ -17,12 +17,12 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log() + logger.info() itemlist = [] - logger.log(page_url) + logger.info(page_url) data = httptools.downloadpage(page_url, post='').data - logger.log(data) + logger.info(data) url = base64.b64decode(data) itemlist.append([".mp4 [HDLoad]", url]) diff --git a/servers/hdmario.py b/servers/hdmario.py index 12b6b9aa..27bb27e9 100644 --- a/servers/hdmario.py +++ b/servers/hdmario.py @@ -7,12 +7,12 @@ from platformcode import logger, config, platformtools baseUrl = 'https://hdmario.live' def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global page, data page = httptools.downloadpage(page_url) data = page.data - logger.log(page.url) + logger.info(page.url) if "the page you are looking for could not be found" in data: return False, config.get_localized_string(70449) % "HDmario" @@ -37,13 +37,13 @@ def registerOrLogin(page_url, forced=False): setting.server_config(Item(config='hdmario')) login() else: - logger.log('Registrazione automatica in corso') + logger.info('Registrazione automatica in corso') import random import string randEmail = ''.join(random.choice(string.ascii_letters + string.digits) for i in range(random.randint(9, 14))) + '@gmail.com' randPsw = ''.join(random.choice(string.ascii_letters + string.digits) for i in range(10)) - logger.log('email: ' + randEmail) - logger.log('pass: ' + randPsw) + logger.info('email: ' + randEmail) + logger.info('pass: ' + randPsw) nTry = 0 while nTry < 5: nTry += 1 @@ -59,7 +59,7 @@ def registerOrLogin(page_url, forced=False): break else: platformtools.dialog_ok('HDmario', 'Impossibile registrarsi automaticamente') - logger.log('Registrazione completata') + logger.info('Registrazione completata') global page, data page = httptools.downloadpage(page_url) data = page.data @@ -67,7 +67,7 @@ def registerOrLogin(page_url, forced=False): def get_video_url(page_url, premium=False, user="", password="", video_password=""): global page, data page_url = page_url.replace('?', '') - logger.log("url=" + page_url) + logger.info("url=" + page_url) if 'unconfirmed' in page.url: from lib import onesecmail @@ -79,7 +79,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= } httptools.downloadpage(page.url, post=postData) jsonMail = onesecmail.waitForMail(mail) - logger.log(jsonMail) + logger.info(jsonMail) if jsonMail: code = jsonMail['subject'].split(' - ')[0] page = httptools.downloadpage(page_url + '?code=' + code) @@ -91,12 +91,12 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= if 'Registrati' in data: platformtools.dialog_ok('HDmario', 'Username/password non validi') registerOrLogin(page_url, True) - logger.log(data) + logger.info(data) from lib import jsunpack_js2py unpacked = jsunpack_js2py.unpack(scrapertools.find_single_match(data, '<script type="text/javascript">\n*\s*\n*(eval.*)')) # p,a,c,k,e,d data -> xhr.setRequestHeader secureProof = scrapertools.find_single_match(unpacked, """X-Secure-Proof['"]\s*,\s*['"]([^"']+)""") - logger.log('X-Secure-Proof=' + secureProof) + logger.info('X-Secure-Proof=' + secureProof) data = httptools.downloadpage(baseUrl + '/pl/' + page_url.split('/')[-1].replace('?', '') + '.m3u8', headers=[['X-Secure-Proof', secureProof]]).data filetools.write(xbmc.translatePath('special://temp/hdmario.m3u8'), data, 'w') diff --git a/servers/hugefiles.py b/servers/hugefiles.py index e0bce659..8bc848be 100755 --- a/servers/hugefiles.py +++ b/servers/hugefiles.py @@ -20,7 +20,7 @@ from platformcode import logger def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data post = {} r = re.findall(r'type="hidden" name="(.+?)"\s* value="?(.+?)">', data) @@ -45,6 +45,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls.append([scrapertools.get_filename_from_url(mediaurl)[-4:] + " [hugefiles]", mediaurl]) for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/idtbox.py b/servers/idtbox.py index 221b0c0e..b4b8f0e2 100644 --- a/servers/idtbox.py +++ b/servers/idtbox.py @@ -10,7 +10,7 @@ from platformcode import logger data = "" def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url) @@ -22,7 +22,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) logger.error(data) video_urls = [] patron = 'source src="([^"]+)" type="([^"]+)" res=(\d+)' diff --git a/servers/jawcloud.py b/servers/jawcloud.py index 0f632651..8fde2947 100644 --- a/servers/jawcloud.py +++ b/servers/jawcloud.py @@ -6,7 +6,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "The file you were looking for could not be found" in data: return False, config.get_localized_string(70449) % "jawcloud" @@ -14,7 +14,7 @@ def test_video_exists(page_url): def get_video_url(page_url, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data video_urls = [] videourl = scrapertools.find_single_match(data, 'source src="([^"]+)') diff --git a/servers/jetload.py b/servers/jetload.py index 4fb70dfb..bd3df40d 100644 --- a/servers/jetload.py +++ b/servers/jetload.py @@ -9,7 +9,7 @@ from platformcode import logger video_urls = [] def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) subtitles = "" response = httptools.downloadpage(page_url) @@ -21,7 +21,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] media_url = scrapertools.find_single_match(data, '<video src="([^"]+)"') if media_url: diff --git a/servers/mailru.py b/servers/mailru.py index d43ce4a2..8383103d 100644 --- a/servers/mailru.py +++ b/servers/mailru.py @@ -7,7 +7,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) page_url = page_url.replace("embed/", "").replace(".html", ".json") data = httptools.downloadpage(page_url).data if '"error":"video_not_found"' in data or '"error":"Can\'t find VideoInstance"' in data: @@ -17,7 +17,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % (page_url)) + logger.info("(page_url='%s')" % (page_url)) video_urls = [] # Carga la página para coger las cookies @@ -45,6 +45,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= pass for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/mediafire.py b/servers/mediafire.py index 96f2049a..5b452878 100755 --- a/servers/mediafire.py +++ b/servers/mediafire.py @@ -6,7 +6,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "Invalid or Deleted File" in data or "Well, looks like we" in data: return False, config.get_localized_string(70449) % "Mediafire" @@ -16,7 +16,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] data = httptools.downloadpage(page_url).data patron = "DownloadButtonAd-startDownload gbtnSecondary.*?href='([^']+)'" @@ -27,5 +27,5 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= if len(matches) > 0: video_urls.append([matches[0][-4:] + " [mediafire]", matches[0]]) for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/mega.py b/servers/mega.py index d9526bb9..83d20c24 100755 --- a/servers/mega.py +++ b/servers/mega.py @@ -79,7 +79,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): page_url = page_url.replace('/embed#', '/#') - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] # si hay mas de 5 archivos crea un playlist con todos diff --git a/servers/mixdrop.py b/servers/mixdrop.py index e5b5e019..43b6eb26 100644 --- a/servers/mixdrop.py +++ b/servers/mixdrop.py @@ -10,7 +10,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url).data @@ -25,7 +25,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] ext = '.mp4' diff --git a/servers/mp4upload.py b/servers/mp4upload.py index db80aef6..769d8699 100644 --- a/servers/mp4upload.py +++ b/servers/mp4upload.py @@ -15,7 +15,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = re.sub(r"\n|\r|\t|\s{2}", "", httptools.downloadpage(page_url).data) match = scrapertools.find_single_match(data, "<script type='text/javascript'>(.*?)</script>") data = jsunpack.unpack(match) @@ -23,9 +23,9 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= media_url = scrapertools.find_single_match(data, '{type:"video/mp4",src:"([^"]+)"}') if not media_url: media_url = scrapertools.find_single_match(data, '"file":"([^"]+)') - logger.log("media_url=" + media_url) + logger.info("media_url=" + media_url) video_urls = list() video_urls.append([scrapertools.get_filename_from_url(media_url)[-4:] + " [mp4upload]", media_url]) for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/mydaddy.py b/servers/mydaddy.py index cf65b0ba..1f6ec8e4 100644 --- a/servers/mydaddy.py +++ b/servers/mydaddy.py @@ -21,7 +21,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log() + logger.info() video_urls = [] data = httptools.downloadpage(page_url).data data = scrapertools.find_single_match(data, 'var srca = \[(.*?)\]') diff --git a/servers/mystream.py b/servers/mystream.py index ec8f0a17..5a7bc97a 100644 --- a/servers/mystream.py +++ b/servers/mystream.py @@ -10,7 +10,7 @@ import re def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url) global page_data page_data = data.data @@ -21,7 +21,7 @@ def test_video_exists(page_url): return True, "" def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] global page_data video_url = scrapertools.find_single_match(decode(page_data), r"'src',\s*'([^']+)") diff --git a/servers/myupload.py b/servers/myupload.py index 228fc82b..04eb0724 100644 --- a/servers/myupload.py +++ b/servers/myupload.py @@ -23,7 +23,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log() + logger.info() video_urls = [] data = httptools.downloadpage(page_url).data matches = scrapertools.find_multiple_matches(data, 'tracker: "([^"]+)"') diff --git a/servers/netutv.py b/servers/netutv.py index aa475c41..8e391f6a 100755 --- a/servers/netutv.py +++ b/servers/netutv.py @@ -25,7 +25,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) #Deshabilitamos el server hasta nueva orden return False, "[netutv] Servidor deshabilitado" # http://netu.tv/watch_video.php=XX solo contiene una redireccion, ir directamente a http://hqq.tv/player/embed_player.php?vid=XX @@ -37,7 +37,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] if "hash=" in page_url: diff --git a/servers/nowvideo.py b/servers/nowvideo.py index 21cde3ca..170b3f94 100644 --- a/servers/nowvideo.py +++ b/servers/nowvideo.py @@ -9,7 +9,7 @@ from platformcode import logger, config headers = [['User-Agent', 'Mozilla/5.0 (Windows NT 6.1; WOW64; rv:54.0) Gecko/20100101 Firefox/54.0']] def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "Not Found" in data or "File was deleted" in data or "The file is being converted" in data or "Please try again later" in data: return False, config.get_localized_string(70293) % "NowVideo" @@ -20,7 +20,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): host = 'http://nowvideo.club' - logger.log("(nowvideo page_url='%s')" % page_url) + logger.info("(nowvideo page_url='%s')" % page_url) video_urls = [] data = httptools.downloadpage(page_url).data page_url_post = scrapertools.find_single_match(data, '<Form id="[^"]+" method="POST" action="([^"]+)">') @@ -28,7 +28,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= imhuman = '&imhuman=' + scrapertools.find_single_match(data, 'name="imhuman" value="([^"]+)"').replace(" ", "+") post = urllib.urlencode({k: v for k, v in scrapertools.find_multiple_matches(data, 'name="([^"]+)" value="([^"]*)"')}) + imhuman data = httptools.downloadpage(host + page_url_post, post=post).data - logger.log("nowvideo data page_url2 ='%s'" % data) + logger.info("nowvideo data page_url2 ='%s'" % data) headers.append(['Referer', page_url]) post_data = scrapertools.find_single_match(data,"</div>\s*<script>(eval.function.p,a,c,k,e,.*?)\s*</script>") @@ -43,11 +43,11 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= _headers = urllib.urlencode(dict(headers)) for media_url in media_urls: - #logger.log("nowvideo data page_url2 ='%s'" % media_url) + #logger.info("nowvideo data page_url2 ='%s'" % media_url) video_urls.append([" mp4 [nowvideo] ", media_url + '|' + _headers]) for video_url in media_urls: - logger.log("[nowvideo.py] %s - %s" % (video_url[0], video_url[1])) + logger.info("[nowvideo.py] %s - %s" % (video_url[0], video_url[1])) return video_urls @@ -57,7 +57,7 @@ def find_videos(data): devuelve = [] patronvideos = r"nowvideo.club/(?:play|videos)?([a-z0-9A-Z]+)" - logger.log("[nowvideo.py] find_videos #" + patronvideos + "#") + logger.info("[nowvideo.py] find_videos #" + patronvideos + "#") matches = re.compile(patronvideos, re.DOTALL).findall(data) for match in matches: @@ -65,10 +65,10 @@ def find_videos(data): url = 'http://nowvideo.club/%s' % match if url not in encontrados: - logger.log(" url=" + url) + logger.info(" url=" + url) devuelve.append([titulo, url, 'nowvideo']) encontrados.add(url) else: - logger.log(" url duplicada=" + url) + logger.info(" url duplicada=" + url) return devuelve diff --git a/servers/okru.py b/servers/okru.py index c8e4588b..f8fb8f83 100644 --- a/servers/okru.py +++ b/servers/okru.py @@ -8,7 +8,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "copyrightsRestricted" in data or "COPYRIGHTS_RESTRICTED" in data: @@ -20,7 +20,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] data = httptools.downloadpage(page_url).data diff --git a/servers/onefichier.py b/servers/onefichier.py index 160b2e59..971d954e 100644 --- a/servers/onefichier.py +++ b/servers/onefichier.py @@ -21,25 +21,25 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) if config.get_setting("premium", server="onefichier"): user = config.get_setting("user", server="onefichier") password = config.get_setting("password", server="onefichier") url = "https://1fichier.com/login.pl" - logger.log("url=" + url) + logger.info("url=" + url) post_parameters = {"mail": user, "pass": password, "lt": "on", "purge": "on", "valider": "Send"} post = urllib.urlencode(post_parameters) - logger.log("post=" + post) + logger.info("post=" + post) data = httptools.downloadpage(url, post=post).data - # logger.log("data="+data) + # logger.info("data="+data) cookies = config.get_cookie_data() - logger.log("cookies=" + cookies) + logger.info("cookies=" + cookies) # 1fichier.com TRUE / FALSE 1443553315 SID imC3q8MQ7cARw5tkXeWvKyrH493rR=1yvrjhxDAA0T0iEmqRfNF9GXwjrwPHssAQ sid_cookie_value = scrapertools.find_single_match(cookies, "1fichier.com.*?SID\s+([A-Za-z0-9\+\=]+)") - logger.log("sid_cookie_value=" + sid_cookie_value) + logger.info("sid_cookie_value=" + sid_cookie_value) # .1fichier.com TRUE / FALSE 1443553315 SID imC3q8MQ7cARw5tkXeWvKyrH493rR=1yvrjhxDAA0T0iEmqRfNF9GXwjrwPHssAQ cookie = urllib.urlencode({"SID": sid_cookie_value}) @@ -50,16 +50,16 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= 'Mozilla/5.0 (Macintosh; U; Intel Mac OS X 10.6; es-ES; rv:1.9.2.12) Gecko/20101026 Firefox/3.6.12']) headers.append(['Cookie', cookie]) filename = scrapertools.get_header_from_response(page_url, header_to_get="Content-Disposition") - logger.log("filename=" + filename) + logger.info("filename=" + filename) # Construye la URL final para Kodi location = page_url + "|Cookie=" + cookie - logger.log("location=" + location) + logger.info("location=" + location) video_urls = [] video_urls.append([filename[-4:] + " (Premium) [1fichier]", location]) for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/onlystream.py b/servers/onlystream.py index 76af2680..12d68cd1 100644 --- a/servers/onlystream.py +++ b/servers/onlystream.py @@ -6,7 +6,7 @@ from platformcode import config, logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url, cookies=False).data if 'File you are looking for is not found.' in data: @@ -16,8 +16,8 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) data = httptools.downloadpage(page_url).data - # logger.log(data) + # logger.info(data) video_urls = support.get_jwplayer_mediaurl(data, 'Onlystream') return video_urls diff --git a/servers/rapidgator.py b/servers/rapidgator.py index 70e94d11..ed25c540 100644 --- a/servers/rapidgator.py +++ b/servers/rapidgator.py @@ -8,6 +8,6 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] return video_urls diff --git a/servers/rcdnme.py b/servers/rcdnme.py index 89189e62..d7fc0b2c 100644 --- a/servers/rcdnme.py +++ b/servers/rcdnme.py @@ -10,7 +10,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url) if "Object not found" in data.data or "longer exists on our servers" in data.data: @@ -21,7 +21,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "p,a,c,k,e,d" in data: @@ -44,6 +44,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= except: pass for video_url in video_urls: - logger.log(" %s - %s" % (video_url[0], video_url[1])) + logger.info(" %s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/rutube.py b/servers/rutube.py index 42541cbb..e6414be7 100644 --- a/servers/rutube.py +++ b/servers/rutube.py @@ -24,13 +24,13 @@ from core import jsontools def get_source(url): - logger.log() + logger.info() data = httptools.downloadpage(url).data data = re.sub(r'\n|\r|\t| |<br>|\s{2,}', "", data) return data def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = get_source(page_url) if "File was deleted" in data or "File Not Found" in data: @@ -40,7 +40,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] referer = '' diff --git a/servers/samaup.py b/servers/samaup.py index ea2918e3..cc6285c0 100644 --- a/servers/samaup.py +++ b/servers/samaup.py @@ -10,7 +10,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url).data if "Not Found" in data or "File was deleted" in data: @@ -19,7 +19,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] ext = 'mp4' diff --git a/servers/sendvid.py b/servers/sendvid.py index 62d4d191..ed11f426 100755 --- a/servers/sendvid.py +++ b/servers/sendvid.py @@ -9,7 +9,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] data = scrapertools.httptools.downloadpage(page_url).data media_url = scrapertools.find_single_match(data, 'var\s+video_source\s+\=\s+"([^"]+)"') @@ -24,5 +24,5 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= else: video_urls.append([scrapertools.get_filename_from_url(media_url)[-4:] + " [sendvid]", media_url]) for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/speedvideo.py b/servers/speedvideo.py index a2349547..c9a2e42e 100644 --- a/servers/speedvideo.py +++ b/servers/speedvideo.py @@ -5,7 +5,7 @@ from core import httptools, scrapertools from platformcode import config, logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data @@ -15,22 +15,22 @@ def test_video_exists(page_url): return True, "" def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] quality ={'MOBILE':1, 'NORMAL':2, 'HD':3} data = httptools.downloadpage(page_url).data - logger.log('SPEEDVIDEO DATA '+ data) + logger.info('SPEEDVIDEO DATA '+ data) media_urls = scrapertools.find_multiple_matches(data, r"file:[^']'([^']+)',\s*label:[^\"]\"([^\"]+)\"") - logger.log("speed video - media urls: %s " % media_urls) + logger.info("speed video - media urls: %s " % media_urls) for media_url, label in media_urls: media_url = httptools.downloadpage(media_url, only_headers=True, follow_redirects=False).headers.get("location", "") if media_url: video_urls.append([media_url.split('.')[-1] + ' - ' + label + ' - ' + ' [Speedvideo]', media_url]) - logger.log("speed video - media urls: %s " % video_urls) + logger.info("speed video - media urls: %s " % video_urls) return sorted(video_urls, key=lambda x: quality[x[0].split(' - ')[1]]) diff --git a/servers/streamtape.py b/servers/streamtape.py index 81f3b405..5bc74691 100644 --- a/servers/streamtape.py +++ b/servers/streamtape.py @@ -6,7 +6,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url).data if "Video not found" in data: @@ -15,7 +15,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] global data diff --git a/servers/streamz.py b/servers/streamz.py index 7427a5f6..3d85f57a 100644 --- a/servers/streamz.py +++ b/servers/streamz.py @@ -6,7 +6,7 @@ from platformcode import logger def get_video_url(page_url, video_password): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] url = httptools.downloadpage(page_url).url data = httptools.downloadpage(url).data diff --git a/servers/supervideo.py b/servers/supervideo.py index 861420bb..85dabc79 100644 --- a/servers/supervideo.py +++ b/servers/supervideo.py @@ -9,7 +9,7 @@ from platformcode import config, logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url, cookies=False).data if 'File is no longer available as it expired or has been deleted' in data: @@ -19,7 +19,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] # data = httptools.downloadpage(page_url).data global data diff --git a/servers/thevid.py b/servers/thevid.py index ed182221..e061856b 100644 --- a/servers/thevid.py +++ b/servers/thevid.py @@ -8,7 +8,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "Video not found..." in data or "Video removed due to copyright" in data: return False, config.get_localized_string(70292) % "Thevid" @@ -31,5 +31,5 @@ def get_video_url(page_url, user="", password="", video_password=""): continue video = "https:" + video video_urls.append(["mp4 [Thevid]", video]) - logger.log("Url: %s" % videos) + logger.info("Url: %s" % videos) return video_urls diff --git a/servers/thevideobee.py b/servers/thevideobee.py index b2f2fc75..f319dcfe 100644 --- a/servers/thevideobee.py +++ b/servers/thevideobee.py @@ -9,7 +9,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "no longer exists" in data or "to copyright issues" in data: return False, config.get_localized_string(70449) % "thevideobee" @@ -17,7 +17,7 @@ def test_video_exists(page_url): def get_video_url(page_url, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data video_urls = [] videourl = scrapertools.find_single_match(data, 'src: "([^"]+)') diff --git a/servers/torrent.py b/servers/torrent.py index b724bf17..2b53ff27 100755 --- a/servers/torrent.py +++ b/servers/torrent.py @@ -3,7 +3,7 @@ import re, os, sys, time, requests, xbmc, xbmcaddon from core import filetools, jsontools -from core.support import dbg, log, match +from core.support import dbg, info, match from platformcode import config, platformtools from torrentool.api import Torrent from lib.guessit import guessit @@ -27,7 +27,7 @@ def get_video_url(page_url, premium=False, user='', password='', video_password= if len(torrent_options) == 0: from specials import elementum_download elementum_download.download() - log('server=torrent, the url is the good') + info('server=torrent, the url is the good') if page_url.startswith('magnet:'): video_urls = [['magnet: [torrent]', page_url]] diff --git a/servers/turbobit.py b/servers/turbobit.py index 8872142c..d370f03f 100644 --- a/servers/turbobit.py +++ b/servers/turbobit.py @@ -4,6 +4,6 @@ from platformcode import logger def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] return video_urls diff --git a/servers/turbovid.py b/servers/turbovid.py index 4e5f0bb4..8353df9b 100644 --- a/servers/turbovid.py +++ b/servers/turbovid.py @@ -9,7 +9,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "Not Found" in data or "File Does not Exist" in data: return False, config.get_localized_string(70449) % "Turbovid" @@ -18,7 +18,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password="", server='Turbovid'): - logger.log("(turbovid page_url='%s')" % page_url) + logger.info("(turbovid page_url='%s')" % page_url) video_urls = [] data = httptools.downloadpage(page_url).data data = data.replace('"', "'") @@ -28,6 +28,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= time.sleep(6) data = httptools.downloadpage(page_url_post, post=post).data - logger.log("(data page_url='%s')" % data) + logger.info("(data page_url='%s')" % data) video_urls = support.get_jwplayer_mediaurl(data, 'Turbovid') return video_urls diff --git a/servers/tusfiles.py b/servers/tusfiles.py index 59e25d8a..18cec1ee 100644 --- a/servers/tusfiles.py +++ b/servers/tusfiles.py @@ -9,7 +9,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "no longer exists" in data or "to copyright issues" in data: return False, config.get_localized_string(70449) % "tusfiles" @@ -17,7 +17,7 @@ def test_video_exists(page_url): def get_video_url(page_url, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data video_urls = [] videourl = scrapertools.find_single_match(data, 'source src="([^"]+)') diff --git a/servers/uploadedto.py b/servers/uploadedto.py index 73d624e5..72b83510 100755 --- a/servers/uploadedto.py +++ b/servers/uploadedto.py @@ -6,7 +6,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) real_url = page_url.replace("uploaded.to", "uploaded.net") code = httptools.downloadpage(real_url, only_headers=True).code @@ -16,36 +16,36 @@ def test_video_exists(page_url): return True, "" def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] if premium: #Si no hay almacenada una cookie activa, hacemos login if check_cookie("uploaded.net", "login") != True: # Login para conseguir la cookie - logger.log("-------------------------------------------") - logger.log("login") - logger.log("-------------------------------------------") + logger.info("-------------------------------------------") + logger.info("login") + logger.info("-------------------------------------------") login_url = "http://uploaded.net/io/login" post = "id=" + user + "&pw=" + password setcookie = httptools.downloadpage(login_url, post=post, follow_redirects=False, only_headers=True).headers.get("set-cookie", "") - logger.log("-------------------------------------------") - logger.log("obtiene la url") - logger.log("-------------------------------------------") + logger.info("-------------------------------------------") + logger.info("obtiene la url") + logger.info("-------------------------------------------") location = httptools.downloadpage(page_url, follow_redirects=False, only_headers=True).headers.get("location", "") - logger.log("location=" + location) + logger.info("location=" + location) #fix descarga no directa if location == "": data = httptools.downloadpage(page_url).data - #logger.log("data: %s" % data) + #logger.info("data: %s" % data) if "<h1>Premium Download</h1>" in data: location = scrapertools.find_single_match(data, '<form method="post" action="([^"]+)"') - #logger.log("location: %s" % location) + #logger.info("location: %s" % location) elif "Hybrid-Traffic is completely exhausted" in data: logger.error("Trafico agotado") @@ -53,13 +53,13 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= logger.error("Cuenta Free") else: logger.error("Error Desconocido") - logger.log("-------------------------------------------") - logger.log("obtiene el nombre del fichero") - logger.log("-------------------------------------------") + logger.info("-------------------------------------------") + logger.info("obtiene el nombre del fichero") + logger.info("-------------------------------------------") try: content_disposition = httptools.downloadpage(location, post="", follow_redirects=False, only_headers=True).headers.get("content-disposition", "") - logger.log("content_disposition=" + content_disposition) + logger.info("content_disposition=" + content_disposition) if content_disposition != "": filename = scrapertools.find_single_match(content_disposition, 'filename="([^"]+)"') extension = filename[-4:] @@ -74,7 +74,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls.append([extension + " (Premium) [uploaded.to]", location]) for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/uppom.py b/servers/uppom.py index e69be294..12b19d30 100644 --- a/servers/uppom.py +++ b/servers/uppom.py @@ -9,7 +9,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = get_source(page_url) if "File was deleted" in data or "File Not Found" in data: @@ -19,7 +19,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log() + logger.info() video_urls = [] data = httptools.downloadpage(page_url).data fid = scrapertools.find_single_match(data, ' name="id" value="([^"]+)"') diff --git a/servers/upstream.py b/servers/upstream.py index 2dc77041..f254ebdc 100644 --- a/servers/upstream.py +++ b/servers/upstream.py @@ -9,7 +9,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url).data if "<h2>WE ARE SORRY</h2>" in data or '<title>404 Not Found' in data: diff --git a/servers/uptobox.py b/servers/uptobox.py index 5e8f0c26..52cc08aa 100755 --- a/servers/uptobox.py +++ b/servers/uptobox.py @@ -19,7 +19,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data @@ -37,7 +37,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) # Si el enlace es directo de upstream if "uptobox" not in page_url: data = httptools.downloadpage(page_url).data @@ -60,7 +60,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls = uptobox(page_url, data) for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/upvid.py b/servers/upvid.py index 79577e8e..4523d5dc 100644 --- a/servers/upvid.py +++ b/servers/upvid.py @@ -13,7 +13,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url) if data.code == 404: return False, config.get_localized_string(70449) % "upvid" @@ -23,7 +23,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium = False, user = "", password = "", video_password = ""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] headers = {'referer': page_url} for i in range(0, 3): diff --git a/servers/uqload.py b/servers/uqload.py index 12828a21..3c254960 100644 --- a/servers/uqload.py +++ b/servers/uqload.py @@ -10,7 +10,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url) @@ -21,7 +21,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] data = httptools.downloadpage(page_url).data diff --git a/servers/userscloud.py b/servers/userscloud.py index 163bf240..8988cdb9 100644 --- a/servers/userscloud.py +++ b/servers/userscloud.py @@ -7,7 +7,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) response = httptools.downloadpage(page_url) @@ -18,7 +18,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] unpacked = "" data = httptools.downloadpage(page_url).data @@ -37,6 +37,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls.append(["%s [userscloud]" % ext, media_url]) for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/vevio.py b/servers/vevio.py index f0e8d92a..f0c1a6c0 100644 --- a/servers/vevio.py +++ b/servers/vevio.py @@ -18,7 +18,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "File was deleted" in data or "Page Cannot Be Found" in data or "Video not found" in data: return False, config.get_localized_string(70449) % "vevio" @@ -26,7 +26,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] post = {} post = urllib.urlencode(post) diff --git a/servers/vidcloud.py b/servers/vidcloud.py index 0fc875bc..4a28eb22 100644 --- a/servers/vidcloud.py +++ b/servers/vidcloud.py @@ -13,7 +13,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "We're Sorry" in data: return False, config.get_localized_string(70292) % "Vidcloud" @@ -22,7 +22,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] @@ -56,6 +56,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls.append(["%s [Vidcloud" % ext, media_url]) for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/videobin.py b/servers/videobin.py index dd1dd075..c2d0125a 100644 --- a/servers/videobin.py +++ b/servers/videobin.py @@ -11,7 +11,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "borrado" in data or "Deleted" in data: return False, config.get_localized_string(70449) % "videobin" @@ -19,7 +19,7 @@ def test_video_exists(page_url): return True, "" def get_video_url(page_url, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] data = httptools.downloadpage(page_url).data bloque = scrapertools.find_single_match(data, 'sources:.\[.*?]') diff --git a/servers/videomega.py b/servers/videomega.py index b2ec79b3..de473b1d 100644 --- a/servers/videomega.py +++ b/servers/videomega.py @@ -5,7 +5,7 @@ from platformcode import logger def get_video_url(page_url, video_password): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] data = httptools.downloadpage(page_url).data m= scrapertools.find_single_match(data, '<link href="(Br74.*?==.css)"') diff --git a/servers/vidfast.py b/servers/vidfast.py index c31f1fd9..382cb0bb 100644 --- a/servers/vidfast.py +++ b/servers/vidfast.py @@ -9,7 +9,7 @@ from platformcode import logger video_urls = [] def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) subtitles = "" response = httptools.downloadpage(page_url) @@ -21,9 +21,9 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] - logger.log("Intel11 %s" %data) + logger.info("Intel11 %s" %data) media_url = scrapertools.find_single_match(data, 'file:"([^"]+)') if media_url: ext = media_url[-4:] diff --git a/servers/vidlox.py b/servers/vidlox.py index 12ce9d3c..7cbb684e 100644 --- a/servers/vidlox.py +++ b/servers/vidlox.py @@ -11,7 +11,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url).data if "borrado" in data or "Deleted" in data: @@ -21,7 +21,7 @@ def test_video_exists(page_url): def get_video_url(page_url, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] bloque = scrapertools.find_single_match(data, 'sources:.\[.*?]') diff --git a/servers/vidmoly.py b/servers/vidmoly.py index 681f00b5..6252edff 100644 --- a/servers/vidmoly.py +++ b/servers/vidmoly.py @@ -5,7 +5,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global data resp = httptools.downloadpage(page_url) data = resp.data @@ -15,7 +15,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) global data video_urls = support.get_jwplayer_mediaurl(data, 'Vidmoly') diff --git a/servers/vidoza.py b/servers/vidoza.py index 76d37ae0..c7a39de0 100644 --- a/servers/vidoza.py +++ b/servers/vidoza.py @@ -8,7 +8,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url).data if "Page not found" in data or "File was deleted" in data: @@ -20,7 +20,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global data video_urls = [] diff --git a/servers/vidtodo.py b/servers/vidtodo.py index 94d611ec..ba08b6b9 100755 --- a/servers/vidtodo.py +++ b/servers/vidtodo.py @@ -8,7 +8,7 @@ from platformcode import logger id_server = "vidtodo" response = "" def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global response response = httptools.downloadpage(page_url) if not response.success or "Not Found" in response.data: @@ -19,7 +19,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] data = response.data packed_data = scrapertools.find_single_match(data, "javascript'>(eval.*?)</script>") @@ -41,5 +41,5 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls.append([".mp4 [%s] %s" % (id_server, inf), mp4 % h]) video_urls.append(["RTMP [%s] %s" % (id_server, inf), "%s playpath=%s" % (rtmp, playpath)]) for video_url in video_urls: - logger.log("video_url: %s - %s" % (video_url[0], video_url[1])) + logger.info("video_url: %s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/vidtome.py b/servers/vidtome.py index 062cf4df..9f1eeafc 100644 --- a/servers/vidtome.py +++ b/servers/vidtome.py @@ -6,7 +6,7 @@ from lib import jsunpack def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url).data if "Not Found" in data or "File Does not Exist" in data: @@ -15,7 +15,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) global data video_urls = [] code = scrapertools.find_single_match(data, 'name="code" value="([^"]+)') diff --git a/servers/vidup.py b/servers/vidup.py index 3da01527..c300d778 100755 --- a/servers/vidup.py +++ b/servers/vidup.py @@ -20,7 +20,7 @@ from platformcode import logger def test_video_exists(page_url): return False, "[Vidup] Servidor Deshabilitado" - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) page = httptools.downloadpage(page_url) url = page.url if "Not Found" in page.data or "/404" in url: @@ -29,7 +29,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] post= {} post = urllib.urlencode(post) diff --git a/servers/vimeo.py b/servers/vimeo.py index 7fd880ac..5308551f 100644 --- a/servers/vimeo.py +++ b/servers/vimeo.py @@ -5,7 +5,7 @@ from core import scrapertools from platformcode import logger, config headers = [['User-Agent', 'Mozilla/5.0']] def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global data if "|" in page_url: @@ -23,7 +23,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] global data @@ -38,6 +38,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls.sort(key=lambda x: x[2]) for video_url in video_urls: video_url[2] = 0 - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/vimpleru.py b/servers/vimpleru.py index 7588aecc..dc21c4aa 100644 --- a/servers/vimpleru.py +++ b/servers/vimpleru.py @@ -7,7 +7,7 @@ from platformcode import config, logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if '"title":"Video Not Found"' in data: return False, config.get_localized_string(70449) % "Vimple" @@ -16,7 +16,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url=%s)" % page_url) + logger.info("(page_url=%s)" % page_url) data = httptools.downloadpage(page_url).data @@ -34,6 +34,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls.append([scrapertools.get_filename_from_url(media_url)[-4:] + " [vimple.ru]", media_url]) for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/vivo.py b/servers/vivo.py index 559e0e60..4c7b73a3 100644 --- a/servers/vivo.py +++ b/servers/vivo.py @@ -9,7 +9,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url) if data.code == 404: return False, config.get_localized_string(70449) % "Vivo" @@ -17,7 +17,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] data = httptools.downloadpage(page_url).data enc_data = scrapertools.find_single_match(data, 'data-stream="([^"]+)') diff --git a/servers/vk.py b/servers/vk.py index a4ec514a..0468cf8a 100755 --- a/servers/vk.py +++ b/servers/vk.py @@ -17,7 +17,7 @@ from platformcode import config, logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) if not login(): return False, "Falta Ingresar/Actualizar las credenciales en el servidor vk. Configuracion - Preferencias - Ajustes de servidores - Configuración del servidor vk" data = httptools.downloadpage(page_url).data @@ -28,7 +28,7 @@ def test_video_exists(page_url): # Returns an array of possible video url's from the page_url def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] data = httptools.downloadpage(page_url).data matches = scrapertools.find_multiple_matches(data, '<source src="([^"]+)" type="video/(\w+)') @@ -37,7 +37,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls.append([calidad + "p ." + ext + " [vk]", media_url]) video_urls.sort(key=lambda it: int(it[0].split("p ", 1)[0])) for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/vshare.py b/servers/vshare.py index a0ec42c9..20f1a000 100644 --- a/servers/vshare.py +++ b/servers/vshare.py @@ -9,7 +9,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) response = httptools.downloadpage(page_url) if response.code != 200 or "No longer available!" in response.data: return False, config.get_localized_string(70449) % "vshare" @@ -18,7 +18,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url = " + page_url) + logger.info("url = " + page_url) headers = {"Referer":page_url} data = httptools.downloadpage(page_url, headers=headers).data flowplayer = re.search("url: [\"']([^\"']+)", data) diff --git a/servers/vudeo.py b/servers/vudeo.py index e2b6bd72..b5986b72 100644 --- a/servers/vudeo.py +++ b/servers/vudeo.py @@ -6,7 +6,7 @@ data = "" def test_video_exists(page_url): global data - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) response = httptools.downloadpage(page_url) if response.code == 404: @@ -18,5 +18,5 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): global data - logger.log("url=" + page_url) + logger.info("url=" + page_url) return support.get_jwplayer_mediaurl(data, 'vudeo') diff --git a/servers/vupplayer.py b/servers/vupplayer.py index dcf187e7..66238660 100644 --- a/servers/vupplayer.py +++ b/servers/vupplayer.py @@ -5,7 +5,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) page = httptools.downloadpage(page_url) global data data = page.data @@ -15,7 +15,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] global data patron = r'sources:\s*\[\{src:\s*"([^"]+)"' diff --git a/servers/vvvvid.py b/servers/vvvvid.py index b3818ed6..53f0568c 100644 --- a/servers/vvvvid.py +++ b/servers/vvvvid.py @@ -19,7 +19,7 @@ payload = {'conn_id': conn_id} def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "Not Found" in data or "File was deleted" in data: return False, config.get_localized_string(70449) % "VVVVID" @@ -45,7 +45,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= # Getting info from Site json_url = "https://www.vvvvid.it/vvvvid/ondemand/" + show_id + '/season/' +season_id + '/' json_file = current_session.get(json_url, headers=headers, params=payload).json() - logger.log(json_file['data']) + logger.info(json_file['data']) # Search for the correct episode for episode in json_file['data']: diff --git a/servers/watchanimestream.py b/servers/watchanimestream.py index 72b41db1..3df92201 100644 --- a/servers/watchanimestream.py +++ b/servers/watchanimestream.py @@ -5,7 +5,7 @@ from platformcode import logger def get_video_url(page_url, video_password): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] url = page_url.replace("/v/", "/api/source/") post = "r=&d=watchanimestream.net" diff --git a/servers/watchvideo.py b/servers/watchvideo.py index b8d57574..81ce9abe 100644 --- a/servers/watchvideo.py +++ b/servers/watchvideo.py @@ -7,7 +7,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url).data if "Not Found" in data or "File was deleted" in data: @@ -16,7 +16,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("url=" + page_url) + logger.info("url=" + page_url) video_urls = [] media_urls = scrapertools.find_multiple_matches(data, 'file:"([^"]+)"') if not media_urls: @@ -34,5 +34,5 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls.append(["%s [watchvideo]" % (ext), media_url]) video_urls.reverse() for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/wstream.py b/servers/wstream.py index 2d162269..57643a51 100644 --- a/servers/wstream.py +++ b/servers/wstream.py @@ -19,7 +19,7 @@ def test_video_exists(page_url): headers = [['User-Agent', 'Mozilla/5.0 (Windows NT 6.1; WOW64; rv:54.0) Gecko/20100101 Firefox/54.0'], ['Host', scrapertools.get_domain_from_url(page_url)]] - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) if 'wstream' in page_url: resp = httptools.downloadpage(page_url.replace(headers[1][1], real_host), headers=headers, verify=False) else: @@ -31,7 +31,7 @@ def test_video_exists(page_url): page_url = resp.url.replace(headers[1][1], real_host) if '/streaming.php' in page_url in page_url: code = httptools.downloadpage(page_url, headers=headers, follow_redirects=False, only_headers=True, verify=False).headers['location'].split('/')[-1].replace('.html', '') - # logger.log('WCODE=' + code) + # logger.info('WCODE=' + code) page_url = 'https://' + real_host + '/video.php?file_code=' + code data = httptools.downloadpage(page_url, headers=headers, follow_redirects=True, verify=False).data @@ -45,7 +45,7 @@ def test_video_exists(page_url): page_url = 'https://' + real_host + '/video.php?file_code=' + scrapertools.find_single_match(dec, "src='([^']+)").split('/')[-1].replace('.html','') headers = [['User-Agent', 'Mozilla/5.0 (Windows NT 6.1; WOW64; rv:54.0) Gecko/20100101 Firefox/54.0'],['Host', 'wstream.video']] new_data = httptools.downloadpage(page_url, headers=headers, follow_redirects=True, verify=False).data - logger.log('NEW DATA: \n' + new_data) + logger.info('NEW DATA: \n' + new_data) if new_data: data = new_data @@ -93,7 +93,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= except: pass - logger.log("[Wstream] url=" + page_url) + logger.info("[Wstream] url=" + page_url) video_urls = [] global data, real_url, headers diff --git a/servers/youdbox.py b/servers/youdbox.py index 00f0de74..a58912b4 100644 --- a/servers/youdbox.py +++ b/servers/youdbox.py @@ -5,7 +5,7 @@ from platformcode import logger def get_video_url(page_url, video_password): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] data = httptools.downloadpage(page_url).data url = scrapertools.find_single_match(data, '<source src="([^"]+)"') diff --git a/servers/yourupload.py b/servers/yourupload.py index f2410141..7625c0a3 100755 --- a/servers/yourupload.py +++ b/servers/yourupload.py @@ -6,7 +6,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url).data @@ -17,7 +17,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] referer = {'Referer': page_url} diff --git a/servers/youtube.py b/servers/youtube.py index 57caa68e..f142696b 100644 --- a/servers/youtube.py +++ b/servers/youtube.py @@ -6,7 +6,7 @@ from platformcode import config, logger, platformtools name = 'plugin.video.youtube' def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data @@ -18,12 +18,12 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): import xbmc from xbmcaddon import Addon - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] if not page_url.startswith("http"): page_url = "http://www.youtube.com/watch?v=%s" % page_url - logger.log(" page_url->'%s'" % page_url) + logger.info(" page_url->'%s'" % page_url) video_id = scrapertools.find_single_match(page_url, '(?:v=|embed/)([A-z0-9_-]{11})') inputstream = platformtools.install_inputstream() diff --git a/servers/youwatch.py b/servers/youwatch.py index 3e7064ae..82f461c3 100644 --- a/servers/youwatch.py +++ b/servers/youwatch.py @@ -6,7 +6,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "File Not Found" in data: return False, config.get_localized_string(70449) % "Youwatch" @@ -20,7 +20,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data url_redirect = scrapertools.find_single_match(data, '<iframe src="([^"]+)"') @@ -31,6 +31,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls = [[scrapertools.get_filename_from_url(url)[-4:] + " [youwatch]", video_url]] for video_url in video_urls: - logger.log("%s - %s" % (video_url[0], video_url[1])) + logger.info("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/zippyshare.py b/servers/zippyshare.py index 0d61f447..65c45a70 100755 --- a/servers/zippyshare.py +++ b/servers/zippyshare.py @@ -29,7 +29,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.log("(page_url='%s')" % page_url) + logger.info("(page_url='%s')" % page_url) video_urls = [] data = httptools.downloadpage(page_url).data @@ -43,5 +43,5 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= mediaurl = '%s%s' % (domain, url) extension = "." + mediaurl.split('.')[-1] video_urls.append([extension + " [zippyshare]", mediaurl]) - logger.log("url=%s" %video_urls) + logger.info("url=%s" %video_urls) return video_urls diff --git a/service.py b/service.py index 4dee0dda..4156cda4 100644 --- a/service.py +++ b/service.py @@ -19,7 +19,7 @@ from servers import torrent def update(path, p_dialog, i, t, serie, overwrite): - logger.log("Updating " + path) + logger.info("Updating " + path) insertados_total = 0 nfo_file = xbmc.translatePath(filetools.join(path, 'tvshow.nfo')) @@ -51,7 +51,7 @@ def update(path, p_dialog, i, t, serie, overwrite): p_dialog.update(int(math.ceil((i + 1) * t)), heading, config.get_localized_string(60389) % (serie.contentSerieName, serie.channel.capitalize())) try: pathchannels = filetools.join(config.get_runtime_path(), "channels", serie.channel + '.py') - logger.log("loading channel: " + pathchannels + " " + + logger.info("loading channel: " + pathchannels + " " + serie.channel) if serie.library_filter_show: @@ -104,7 +104,7 @@ def update(path, p_dialog, i, t, serie, overwrite): def check_for_update(overwrite=True): - logger.log("Update Series...") + logger.info("Update Series...") p_dialog = None serie_actualizada = False update_when_finished = False @@ -136,7 +136,7 @@ def check_for_update(overwrite=True): filetools.write(tvshow_file, head_nfo + serie.tojson()) path = filetools.dirname(tvshow_file) - logger.log("serie=" + serie.contentSerieName) + logger.info("serie=" + serie.contentSerieName) p_dialog.update(int(math.ceil((i + 1) * t)), heading, serie.contentSerieName) #Verificamos el estado del serie.library_playcounts de la Serie por si está incompleto @@ -248,7 +248,7 @@ def check_for_update(overwrite=True): p_dialog.close() else: - logger.log("Not update the video library, it is disabled") + logger.info("Not update the video library, it is disabled") except Exception as ex: logger.error("An error occurred while updating the series") @@ -278,7 +278,7 @@ def viewmodeMonitor(): if content: defaultMode = int(config.get_setting('view_mode_%s' % content).split(',')[-1]) if currentMode != defaultMode: - logger.log('viewmode changed: ' + currentModeName + '-' + str(currentMode) + ' - content: ' + content) + logger.info('viewmode changed: ' + currentModeName + '-' + str(currentMode) + ' - content: ' + content) config.set_setting('view_mode_%s' % content, currentModeName + ', ' + str(currentMode)) except: logger.error(traceback.print_exc()) @@ -318,7 +318,7 @@ class AddonMonitor(xbmc.Monitor): super(AddonMonitor, self).__init__() def onSettingsChanged(self): - logger.log('settings changed') + logger.debug('settings changed') settings_post = config.get_all_settings_addon() from platformcode import xbmc_videolibrary @@ -359,11 +359,11 @@ class AddonMonitor(xbmc.Monitor): self.settings_pre = settings_post def onScreensaverActivated(self): - logger.log('screensaver activated, un-scheduling screen-on jobs') + logger.info('screensaver activated, un-scheduling screen-on jobs') schedule.clear('screenOn') def onScreensaverDeactivated(self): - logger.log('screensaver deactivated, re-scheduling screen-on jobs') + logger.info('screensaver deactivated, re-scheduling screen-on jobs') self.scheduleScreenOnJobs() def scheduleUpdater(self): @@ -371,7 +371,7 @@ class AddonMonitor(xbmc.Monitor): updaterCheck() self.updaterPeriod = config.get_setting('addon_update_timer') schedule.every(self.updaterPeriod).hours.do(updaterCheck).tag('updater') - logger.log('scheduled updater every ' + str(self.updaterPeriod) + ' hours') + logger.info('scheduled updater every ' + str(self.updaterPeriod) + ' hours') def scheduleVideolibrary(self): self.update_setting = config.get_setting("update", "videolibrary") @@ -379,23 +379,23 @@ class AddonMonitor(xbmc.Monitor): if self.update_setting == 2 or self.update_setting == 3: self.update_hour = config.get_setting("everyday_delay", "videolibrary") * 4 schedule.every().day.at(str(self.update_hour).zfill(2) + ':00').do(run_threaded, check_for_update, (False,)).tag('videolibrary') - logger.log('scheduled videolibrary at ' + str(self.update_hour).zfill(2) + ':00') + logger.info('scheduled videolibrary at ' + str(self.update_hour).zfill(2) + ':00') def scheduleScreenOnJobs(self): schedule.every().second.do(viewmodeMonitor).tag('screenOn') schedule.every().second.do(torrent.elementum_monitor).tag('screenOn') def onDPMSActivated(self): - logger.log('DPMS activated, un-scheduling screen-on jobs') + logger.info('DPMS activated, un-scheduling screen-on jobs') schedule.clear('screenOn') def onDPMSDeactivated(self): - logger.log('DPMS deactivated, re-scheduling screen-on jobs') + logger.info('DPMS deactivated, re-scheduling screen-on jobs') self.scheduleScreenOnJobs() if __name__ == "__main__": - logger.log('Starting KoD service') + logger.info('Starting KoD service') monitor = AddonMonitor() # mark as stopped all downloads (if we are here, probably kodi just started) diff --git a/specials/autoplay.py b/specials/autoplay.py index 79e9c9a1..8ce72c55 100644 --- a/specials/autoplay.py +++ b/specials/autoplay.py @@ -29,7 +29,7 @@ def start(itemlist, item): if item.global_search: return itemlist - logger.log() + logger.info() global PLAYED PLAYED = False @@ -274,7 +274,7 @@ def start(itemlist, item): def play_multi_channel(item, itemlist): - logger.log() + logger.info() start(itemlist, item) diff --git a/specials/autorenumber.py b/specials/autorenumber.py index 2086d9ea..528b41b6 100644 --- a/specials/autorenumber.py +++ b/specials/autorenumber.py @@ -33,7 +33,7 @@ except: xbmcgui = None import re, base64, json, inspect from core import jsontools, tvdb, scrapertools, filetools -from core.support import typo, log +from core.support import typo, info from platformcode import config, platformtools TAG_TVSHOW_RENUMERATE = "TVSHOW_AUTORENUMBER" @@ -49,7 +49,7 @@ TAG_TYPE = "Type" def renumber(itemlist, item='', typography=''): - log() + info() dict_series = load(itemlist[0]) if len(itemlist) > 0 else {} if item: @@ -103,7 +103,7 @@ def renumber(itemlist, item='', typography=''): def config_item(item, itemlist=[], typography='', active=False): - log() + info() # Configurazione Automatica, Tenta la numerazione Automatica degli episodi title = item.fulltitle.rstrip() @@ -140,7 +140,7 @@ def config_item(item, itemlist=[], typography='', active=False): def semiautomatic_config_item(item): - log() + info() # Configurazione Semi Automatica, utile in caso la numerazione automatica fallisca tvdb.find_and_set_infoLabels(item) @@ -254,7 +254,7 @@ def renumeration (itemlist, item, typography, dict_series, ID, Season, Episode, def manual_renumeration(item, modify=False): - log() + info() _list = [] if item.from_channel: item.channel = item.from_channel title = item.fulltitle.rstrip() @@ -342,7 +342,7 @@ def manual_renumeration(item, modify=False): def delete_renumeration(item): - log() + info() if item.from_channel: item.channel = item.from_channel title = item.fulltitle.rstrip() @@ -352,7 +352,7 @@ def delete_renumeration(item): def make_list(itemlist, item, typography, dict_series, ID, Season, Episode, Mode, title): - log() + info() exist = True item.infoLabels['tvdb_id'] = ID tvdb.set_infoLabels_item(item) @@ -478,7 +478,7 @@ def make_list(itemlist, item, typography, dict_series, ID, Season, Episode, Mode def check(item): - log() + info() dict_series = load(item) title = item.fulltitle.rstrip() if title in dict_series: title = dict_series[title] @@ -493,7 +493,7 @@ def error(itemlist): def find_episodes(item): - log() + info() ch = __import__('channels.' + item.channel, fromlist=["channels.%s" % item.channel]) itemlist = ch.episodios(item) return itemlist @@ -501,7 +501,7 @@ def find_episodes(item): def RepresentsInt(s): # Controllo Numro Stagione - log() + info() try: int(s) return True @@ -537,7 +537,7 @@ def select_type(item): def filename(item): - log() + info() name_file = item.channel + "_data.json" path = filetools.join(config.get_data_path(), "settings_channels") fname = filetools.join(path, name_file) @@ -546,7 +546,7 @@ def filename(item): def load(item): - log() + info() try: json_file = open(filename(item), "r").read() json = jsontools.load(json_file)[TAG_TVSHOW_RENUMERATE] @@ -558,7 +558,7 @@ def load(item): def write(item, json): - log() + info() json_file = open(filename(item), "r").read() js = jsontools.load(json_file) js[TAG_TVSHOW_RENUMERATE] = json diff --git a/specials/backup.py b/specials/backup.py index f4248065..c6405d45 100644 --- a/specials/backup.py +++ b/specials/backup.py @@ -18,7 +18,7 @@ videolibrary_tvshows_path = u'' + videolibrarytools.TVSHOWS_PATH def export_videolibrary(item): - logger.log() + logger.info() zip_file_folder = platformtools.dialog_browse(3, config.get_localized_string(80002)) if zip_file_folder == "": @@ -51,7 +51,7 @@ def export_videolibrary(item): def import_videolibrary(item): - logger.log() + logger.info() zip_file = u'' + platformtools.dialog_browse(1, config.get_localized_string(80005), mask=".zip") if zip_file == "": diff --git a/specials/checkhost.py b/specials/checkhost.py index d431fb3b..31ea9187 100644 --- a/specials/checkhost.py +++ b/specials/checkhost.py @@ -37,7 +37,7 @@ class Kdicc(): self.view_msg = view_msg self.lst_site_check_dns = lst_site_check_dns self.urls = [] - #logger.log("check #### INIZIO INIT#### ") + #logger.info("check #### INIZIO INIT#### ") def check_Ip(self): """ @@ -248,7 +248,7 @@ def check_channels(inutile=''): This is because it can happen that at any time the connection may have problems. If it does, check it relative writing of the file is interrupted with a warning message """ - logger.log() + logger.info() folderJson = xbmc.translatePath(xbmcaddon.Addon().getAddonInfo('path')).decode('utf-8') fileJson = 'channels.json' @@ -264,7 +264,7 @@ def check_channels(inutile=''): # to get an idea of ​​the timing # useful only if you control all channels # for channels with error 522 about 40 seconds are lost ... - logger.log("check #### INIZIO #### channel - host :%s - %s " % (chann, host)) + logger.info("check #### INIZIO #### channel - host :%s - %s " % (chann, host)) rslt = Kdicc(lst_urls = [host]).http_Resp() @@ -288,10 +288,10 @@ def check_channels(inutile=''): # risultato[chann] = 'Errore Sconosciuto - '+str(rslt['code']) +' - '+ host risultato[chann] = host - logger.log("check #### FINE #### rslt :%s " % (rslt)) + logger.info("check #### FINE #### rslt :%s " % (rslt)) fileJson_test = 'channels-test.json' # I write the updated file with open(folderJson+'/'+fileJson_test, 'w') as f: data = json.dump(risultato, f, sort_keys=True, indent=4) - logger.log(data) + logger.info(data) diff --git a/specials/community.py b/specials/community.py index 711eef44..f31c28d1 100644 --- a/specials/community.py +++ b/specials/community.py @@ -26,7 +26,7 @@ list_quality = ['SD', '720', '1080', '4k'] tmdb_api = 'a1ab8b8669da03637a4b98fa39c39228' def mainlist(item): - support.log() + support.info() path = filetools.join(config.get_data_path(), 'community_channels.json') if not filetools.exists(path): @@ -38,7 +38,7 @@ def mainlist(item): def show_channels(item): - support.log() + support.info() itemlist = [] # add context menu @@ -78,7 +78,7 @@ def show_channels(item): def show_menu(item): - support.log() + support.info() itemlist = [] @@ -114,12 +114,12 @@ def show_menu(item): if 'channel_name' in json and not 'disable_search' in json and 'search' not in json: itemlist += get_search_menu(item, json, channel_name=json['channel_name']) - support.log('PAGINATION:', disable_pagination) + support.info('PAGINATION:', disable_pagination) return itemlist def search(item, text): - support.log(text) + support.info(text) itemlist = [] if item.custom_search: @@ -164,7 +164,7 @@ def global_search(item, text): def peliculas(item, json='', key='', itemlist=[]): item.plot = item.thumb = item.fanart ='' - support.log('PAGINATION:', item.disable_pagination) + support.info('PAGINATION:', item.disable_pagination) if not json: key = item.key json = load_json(item)[key] @@ -233,7 +233,7 @@ def peliculas(item, json='', key='', itemlist=[]): def get_seasons(item): - support.log() + support.info() itemlist = [] infoLabels = item.infoLabels json = item.url if type(item.url) == dict else item.url @@ -273,7 +273,7 @@ def get_seasons(item): def episodios(item, json ='', key='', itemlist =[]): - support.log() + support.info() infoLabels = item.infoLabels itm=item @@ -384,7 +384,7 @@ def episodios(item, json ='', key='', itemlist =[]): # Find Servers def findvideos(item): - support.log() + support.info() itemlist = [] if 'links' in item.url: json = item.url['links'] @@ -404,7 +404,7 @@ def findvideos(item): ################################ Menu ################################ def get_menu(item, json, key, itemlist=[]): - support.log() + support.info() json = json[key] for option in json: title = option['title'] if 'title' in option else json[option] if 'search' not in option else '' @@ -439,7 +439,7 @@ def get_menu(item, json, key, itemlist=[]): def get_sub_menu(item, json, key, itemlist=[]): - support.log() + support.info() json = json[key] search = False if item.menu: @@ -478,7 +478,7 @@ def get_sub_menu(item, json, key, itemlist=[]): def get_search_menu(item, json='', itemlist=[], channel_name=''): - support.log() + support.info() if 'title' in json: title = json['title'] elif channel_name: @@ -504,7 +504,7 @@ def get_search_menu(item, json='', itemlist=[], channel_name=''): def submenu(item, json, key, itemlist = [], filter_list = []): - support.log(item) + support.info(item) import sys if sys.version_info[0] >= 3: from concurrent import futures @@ -575,7 +575,7 @@ def filter_thread(filter, key, item, description): if id: thumbnail = 'http://image.tmdb.org/t/p/original' + results['profile_path'] if results['profile_path'] else item.thumbnail json_file = httptools.downloadpage('http://api.themoviedb.org/3/person/'+ str(id) + '?api_key=' + tmdb_api + '&language=en', use_requests=True).data - support.log(json_file) + support.info(json_file) plot += jsontools.load(json_file)['biography'] if description: @@ -609,7 +609,7 @@ def filter_thread(filter, key, item, description): # for load json from item or url def load_json(item, no_order=False): - support.log() + support.info() if type(item) == Item: url = item.url filterkey = item.filterkey @@ -634,7 +634,7 @@ def load_json(item, no_order=False): # Load Channels json and check that the paths and channel titles are correct def load_and_check(item): - support.log() + support.info() path = filetools.join(config.get_data_path(), 'community_channels.json') file = open(path, "r") json = jsontools.load(file.read()) @@ -655,7 +655,7 @@ def load_and_check(item): # set extra values def set_extra_values(item, json, path): - support.log() + support.info() ret = Item() for key in json: if key == 'quality': @@ -701,7 +701,7 @@ def set_extra_values(item, json, path): # format titles def set_title(title, language='', quality=''): - support.log() + support.info() t = support.match(title, patron=r'\{([^\}]+)\}').match if 'bold' not in t: t += ' bold' @@ -722,7 +722,7 @@ def set_title(title, language='', quality=''): # for relative path def relative(key, json, path): - support.log() + support.info() ret = '' if key in json: if key in ['thumbnail', 'poster']: @@ -734,7 +734,7 @@ def relative(key, json, path): def pagination(item, itemlist = []): - support.log() + support.info() import json itlist = [] @@ -775,7 +775,7 @@ def pagination(item, itemlist = []): return itlist def add_channel(item): - support.log() + support.info() channel_to_add = {} json_file = '' result = platformtools.dialog_select(config.get_localized_string(70676), [config.get_localized_string(70678), config.get_localized_string(70679)]) @@ -828,7 +828,7 @@ def add_channel(item): return def remove_channel(item): - support.log() + support.info() path = filetools.join(config.get_data_path(), 'community_channels.json') diff --git a/specials/downloads.py b/specials/downloads.py index 3e668621..aa719830 100644 --- a/specials/downloads.py +++ b/specials/downloads.py @@ -18,7 +18,7 @@ from core import filetools, jsontools, scraper, scrapertools, servertools, video from core.downloader import Downloader from core.item import Item from platformcode import config, logger, platformtools -from core.support import log, dbg, typo +from core.support import info, dbg, typo from servers import torrent kb = '0xFF65B3DA' @@ -40,7 +40,7 @@ extensions_list = ['.aaf', '.3gp', '.asf', '.avi', '.flv', '.mpeg', '.m1v', '.m2 def mainlist(item): - log() + info() itemlist = [] # File list @@ -143,7 +143,7 @@ def settings(item): def browser(item): - log() + info() itemlist = [] for file in filetools.listdir(item.url): @@ -177,7 +177,7 @@ def del_dir(item): def clean_all(item): - log() + info() stop_all() removeFiles = False if platformtools.dialog_yesno(config.get_localized_string(20000), config.get_localized_string(30300)): @@ -203,7 +203,7 @@ def reload(item): def stop_all(item=None): - log() + info() for fichero in sorted(filetools.listdir(DOWNLOAD_LIST_PATH)): if fichero.endswith(".json"): @@ -221,7 +221,7 @@ def stop_all(item=None): def clean_ready(item): - log() + info() for fichero in sorted(filetools.listdir(DOWNLOAD_LIST_PATH)): if fichero.endswith(".json"): download_item = Item().fromjson(filetools.read(filetools.join(DOWNLOAD_LIST_PATH, fichero))) @@ -233,7 +233,7 @@ def clean_ready(item): def restart_error(item): - log() + info() for fichero in sorted(filetools.listdir(DOWNLOAD_LIST_PATH)): if fichero.endswith(".json"): download_item = Item().fromjson(filetools.read(filetools.join(DOWNLOAD_LIST_PATH, fichero))) @@ -272,7 +272,7 @@ def download_all_background(item): def menu(item): - log(item) + info(item) if item.downloadServer: servidor = item.downloadServer.get("server", "Auto") else: @@ -312,7 +312,7 @@ def menu(item): # Show Dialog seleccion = platformtools.dialog_select(config.get_localized_string(30163), opciones) - logger.log('SELECTION: '+ op[seleccion]) + logger.info('SELECTION: '+ op[seleccion]) # -1 is cancel if seleccion == -1: return @@ -368,7 +368,7 @@ def menu(item): def move_to_libray(item): - log() + info() if item.contentType == 'movie': FOLDER = FOLDER_MOVIES @@ -422,7 +422,7 @@ def move_to_libray(item): if filename.startswith(name) and (filename.endswith('.strm') or (filename.endswith('.json') and 'downloads' not in filename)): clean = True file_path = filetools.join(config.get_setting("videolibrarypath"), FOLDER, path_title, File) - log('Delete File:', str(file_path)) + info('Delete File:', str(file_path)) filetools.remove(file_path) if file_path.endswith('.strm'): file_strm_path = file_path @@ -578,7 +578,7 @@ def sort_method(item): def download_from_url(url, item): - log("Attempting to download:", url) + info("Attempting to download:", url) if url.lower().split('|')[0].endswith(".m3u8") or url.lower().startswith("rtmp"): save_server_statistics(item.server, 0, False) platformtools.dialog_notification('m3u8 Download',config.get_localized_string(60364), sound=False) @@ -607,22 +607,22 @@ def download_from_url(url, item): update_json(item.path, {"downloadUrl": d.download_url, "downloadStatus": STATUS_CODES.downloading, "downloadSize": d.size[0], "downloadProgress": d.progress, "downloadCompleted": d.downloaded[0], "downloadFilename": file}) - d.start_dialog(config.get_localized_string(60332)) + d.start_diainfo(config.get_localized_string(60332)) # Download stopped. We get the state: # Download failed if d.state == d.states.error: - log("Error trying to download", url) + info("Error trying to download", url) status = STATUS_CODES.error # Download has stopped elif d.state == d.states.stopped: - log("Stop download") + info("Stop download") status = STATUS_CODES.canceled # Download is complete elif d.state == d.states.completed: - log("Downloaded correctly") + info("Downloaded correctly") status = STATUS_CODES.completed if (item.downloadSize and item.downloadSize != d.size[0]) or d.size[0] < 5000000: # if size don't correspond or file is too little (gounlimited for example send a little video to say the server is overloaded) @@ -638,7 +638,7 @@ def download_from_url(url, item): def download_from_server(item): - log(item.tostring()) + info(item.tostring()) unsupported_servers = ["torrent"] if item.contentChannel == 'local': @@ -668,11 +668,11 @@ def download_from_server(item): item.video_urls = itemlist if not item.server: item.server = "directo" else: - log("There is nothing to reproduce") + info("There is nothing to reproduce") return {"downloadStatus": STATUS_CODES.error} finally: progreso.close() - log("contentAction: %s | contentChannel: %s | server: %s | url: %s" % (item.contentAction, item.contentChannel, item.server, item.url)) + info("contentAction: %s | contentChannel: %s | server: %s | url: %s" % (item.contentAction, item.contentChannel, item.server, item.url)) if item.server == 'torrent': import xbmcgui @@ -692,11 +692,11 @@ def download_from_server(item): # If it is not available, we go out if not puedes: - log("The video is NOT available") + info("The video is NOT available") return {"downloadStatus": STATUS_CODES.error} else: - log("YES Video is available") + info("YES Video is available") result = {} @@ -717,14 +717,14 @@ def download_from_server(item): def download_from_best_server(item): - log("contentAction: %s | contentChannel: %s | url: %s" % (item.contentAction, item.contentChannel, item.url)) + info("contentAction: %s | contentChannel: %s | url: %s" % (item.contentAction, item.contentChannel, item.url)) result = {"downloadStatus": STATUS_CODES.error} progreso = platformtools.dialog_progress_bg(config.get_localized_string(30101), config.get_localized_string(70179)) try: if item.downloadItemlist: - log('using cached servers') + info('using cached servers') play_items = [Item().fromurl(i) for i in item.downloadItemlist] else: if item.contentChannel in ['community', 'videolibrary']: @@ -773,11 +773,11 @@ def download_from_best_server(item): def select_server(item): if item.server: return "Auto" - log("contentAction: %s | contentChannel: %s | url: %s" % (item.contentAction, item.contentChannel, item.url)) + info("contentAction: %s | contentChannel: %s | url: %s" % (item.contentAction, item.contentChannel, item.url)) progreso = platformtools.dialog_progress_bg(config.get_localized_string(30101), config.get_localized_string(70179)) try: if item.downloadItemlist: - log('using cached servers') + info('using cached servers') play_items = [Item().fromurl(i) for i in item.downloadItemlist] else: if item.contentChannel in ['community', 'videolibrary']: @@ -819,7 +819,7 @@ def select_server(item): def start_download(item): - log("contentAction: %s | contentChannel: %s | url: %s" % (item.contentAction, item.contentChannel, item.url)) + info("contentAction: %s | contentChannel: %s | url: %s" % (item.contentAction, item.contentChannel, item.url)) # We already have a server, we just need to download if item.contentAction == "play": ret = download_from_server(item) @@ -839,7 +839,7 @@ def start_download(item): def get_episodes(item): - log("contentAction: %s | contentChannel: %s | contentType: %s" % (item.contentAction, item.contentChannel, item.contentType)) + info("contentAction: %s | contentChannel: %s | contentType: %s" % (item.contentAction, item.contentChannel, item.contentType)) if 'dlseason' in item: season = True @@ -916,7 +916,7 @@ def get_episodes(item): # Any other result is not worth it, we ignore it else: - log("Omitiendo item no válido:", episode.tostring()) + info("Omitiendo item no válido:", episode.tostring()) # Any other result is not worth it, we ignore it... itemlist = videolibrarytools.filter_list(itemlist) @@ -925,7 +925,7 @@ def get_episodes(item): def write_json(item): - log() + info() channel = item.from_channel if item.from_channel else item.channel item.action = "menu" @@ -970,7 +970,7 @@ def save_download(item): def save_download_background(item): - log() + info() # Menu contextual if item.from_action and item.from_channel: item.channel = item.from_channel @@ -1019,7 +1019,7 @@ def save_download_background(item): def save_download_videolibrary(item): - log() + info() show_disclaimer() item.contentChannel = 'videolibrary' item.channel = "downloads" @@ -1028,7 +1028,7 @@ def save_download_videolibrary(item): def save_download_video(item): - log("contentAction: %s | contentChannel: %s | contentTitle: %s" % (item.contentAction, item.contentChannel, item.contentTitle)) + info("contentAction: %s | contentChannel: %s | contentTitle: %s" % (item.contentAction, item.contentChannel, item.contentTitle)) set_movie_title(item) @@ -1043,7 +1043,7 @@ def save_download_video(item): def save_download_movie(item): - log("contentAction: %s | contentChannel: %s | contentTitle: %s" % ( item.contentAction, item.contentChannel, item.contentTitle)) + info("contentAction: %s | contentChannel: %s | contentTitle: %s" % ( item.contentAction, item.contentChannel, item.contentTitle)) progreso = platformtools.dialog_progress_bg(config.get_localized_string(30101), config.get_localized_string(70191)) @@ -1077,7 +1077,7 @@ def save_download_movie(item): def save_download_tvshow(item): - log("contentAction: %s | contentChannel: %s | contentType: %s | contentSerieName: %s" % (item.contentAction, item.contentChannel, item.contentType, item.contentSerieName)) + info("contentAction: %s | contentChannel: %s | contentType: %s | contentSerieName: %s" % (item.contentAction, item.contentChannel, item.contentType, item.contentSerieName)) progreso = platformtools.dialog_progress_bg(config.get_localized_string(30101), config.get_localized_string(70188)) try: item.show = item.fulltitle diff --git a/specials/elementum_download.py b/specials/elementum_download.py index bef65515..ba871785 100644 --- a/specials/elementum_download.py +++ b/specials/elementum_download.py @@ -25,9 +25,9 @@ def download(item=None): if platformtools.dialog_yesno(config.get_localized_string(70784), config.get_localized_string(70782)): pform = get_platform() url = support.match(elementum_url, patronBlock=r'<div class="release-entry">(.*?)<!-- /.release-body -->', patron=r'<a href="([a-zA-Z0-9/\.-]+%s.zip)' % pform).match - support.log('OS:', pform) - support.log('Extract IN:', elementum_path) - support.log('URL:', url) + support.info('OS:', pform) + support.info('Extract IN:', elementum_path) + support.info('URL:', url) if url: downloadtools.downloadfile(host + url, filename) extract() @@ -38,19 +38,19 @@ def download(item=None): def extract(): import zipfile from platformcode.updater import fixZipGetHash - support.log('Estraggo Elementum in:', elementum_path) + support.info('Estraggo Elementum in:', elementum_path) try: hash = fixZipGetHash(filename) - support.log(hash) + support.info(hash) with zipfile.ZipFile(filetools.file_open(filename, 'rb', vfs=False)) as zip_ref: zip_ref.extractall(xbmc.translatePath(addon_path)) except Exception as e: - support.log('Non sono riuscito ad estrarre il file zip') - support.logger.error(e) + support.info('Non sono riuscito ad estrarre il file zip') + support.infoger.error(e) import traceback - support.logger.error(traceback.print_exc()) + support.infoger.error(traceback.print_exc()) def setting(): @@ -68,7 +68,7 @@ def setting(): __settings__.setSetting('do_not_disturb', 'true') Continue = False except: - support.log('RIPROVO') + support.info('RIPROVO') xbmc.sleep(100) else: if not filetools.exists(elementum_path): diff --git a/specials/favorites.py b/specials/favorites.py index 8e7650cb..b6fe9720 100644 --- a/specials/favorites.py +++ b/specials/favorites.py @@ -24,7 +24,7 @@ except: def mainlist(item): - logger.log() + logger.info() itemlist = [] for name, thumb, data in read_favourites(): @@ -81,7 +81,7 @@ def save_favourites(favourites_list): def addFavourite(item): - logger.log() + logger.info() # logger.debug(item.tostring('\n')) # If you get here through the context menu, you must retrieve the action and channel parameters @@ -100,7 +100,7 @@ def addFavourite(item): def delFavourite(item): - logger.log() + logger.info() # logger.debug(item.tostring('\n')) if item.from_title: @@ -118,7 +118,7 @@ def delFavourite(item): def renameFavourite(item): - logger.log() + logger.info() # logger.debug(item.tostring('\n')) # Find the item we want to rename in favorites.xml @@ -137,7 +137,7 @@ def renameFavourite(item): ################################################## # Features to migrate old favorites (.txt) def readbookmark(filepath): - logger.log() + logger.info() import urllib bookmarkfile = filetools.file_open(filepath) @@ -230,6 +230,6 @@ try: if config.get_setting("bookmarkpath") != "": check_bookmark(config.get_setting("bookmarkpath")) else: - logger.log("No path to old version favorites") + logger.info("No path to old version favorites") except: pass diff --git a/specials/filmontv.py b/specials/filmontv.py index c9973ed0..cbc9dc39 100644 --- a/specials/filmontv.py +++ b/specials/filmontv.py @@ -15,7 +15,7 @@ blacklisted_genres = ['attualita', 'scienza', 'religione', 'cucina', 'notiziario def mainlist(item): - support.log() + support.info() itemlist = [Item(title=support.typo('Film in onda oggi', 'bold'), channel=item.channel, action='category', contentType='movie', thumbnail=support.thumb('movie')), Item(title=support.typo('Serie Tv in onda oggi', 'bold'), channel=item.channel, action='peliculas', contentType='tvshow', thumbnail=support.thumb('tvshow')), @@ -31,13 +31,13 @@ def getEpg(): archiveName = fileName + '.gz' xmlName = fileName + '.xml' if not filetools.exists(xmlName): - support.log('downloading epg') + support.info('downloading epg') # cancello quelli vecchi for f in glob.glob(support.config.get_temp_file('guidatv-') + '*'): filetools.remove(f, silent=True) # inmemory = io.BytesIO(httptools.downloadpage(host).data) downloadtools.downloadfile(host, archiveName) - support.log('opening gzip and writing xml') + support.info('opening gzip and writing xml') with gzip.GzipFile(fileobj=filetools.file_open(archiveName, mode='rb', vfs=False)) as f: guide = f.read().decode('utf-8') guide = guide.replace('\n', ' ').replace('><', '>\n<') @@ -194,7 +194,7 @@ def listaCanali(item): thumbnail = None skip = False # return itemlist - # logger.log([i.title for i in itemlist]) + # logger.info([i.title for i in itemlist]) f.close() return sorted(itemlist, key=lambda x: x.title) diff --git a/specials/filtertools.py b/specials/filtertools.py index 6ea735d5..c03ffe28 100644 --- a/specials/filtertools.py +++ b/specials/filtertools.py @@ -234,7 +234,7 @@ def get_link(list_item, item, list_language, list_quality=None, global_filter_la @return: Item list @rtype: list[Item] """ - logger.log() + logger.info() # if the required fields are None we leave if list_item is None or item is None: @@ -274,7 +274,7 @@ def get_links(list_item, item, list_language, list_quality=None, global_filter_l @return: lista de Item @rtype: list[Item] """ - logger.log() + logger.info() # if the required fields are None we leave @@ -362,7 +362,7 @@ def no_filter(item): @return: lista de enlaces @rtype: list[Item] """ - logger.log() + logger.info() itemlist = [] for i in item.list_item_all: @@ -384,7 +384,7 @@ def mainlist(channel, list_language, list_quality): @return: Item list @rtype: list[Item] """ - logger.log() + logger.info() itemlist = [] dict_series = jsontools.get_node_from_file(channel, TAG_TVSHOW_FILTER) @@ -425,8 +425,8 @@ def config_item(item): @param item: item @type item: Item """ - logger.log() - logger.log("item %s" % item.tostring()) + logger.info() + logger.info("item %s" % item.tostring()) # WE GET THE JSON DATA dict_series = jsontools.get_node_from_file(item.from_channel, TAG_TVSHOW_FILTER) @@ -448,8 +448,8 @@ def config_item(item): else: lang_selected = dict_series.get(tvshow, {}).get(TAG_LANGUAGE, default_lang) list_quality = dict_series.get(tvshow, {}).get(TAG_QUALITY_ALLOWED, [x.lower() for x in item.list_quality]) - # logger.log("lang selected {}".format(lang_selected)) - # logger.log("list quality {}".format(list_quality)) + # logger.info("lang selected {}".format(lang_selected)) + # logger.info("list quality {}".format(list_quality)) active = True custom_button = {'visible': False} @@ -516,7 +516,7 @@ def config_item(item): def delete(item, dict_values): - logger.log() + logger.info() if item: dict_series = jsontools.get_node_from_file(item.from_channel, TAG_TVSHOW_FILTER) @@ -554,7 +554,7 @@ def save(item, dict_data_saved): @param dict_data_saved: dictionary with saved data @type dict_data_saved: dict """ - logger.log() + logger.info() if item and dict_data_saved: logger.debug('item: %s\ndatos: %s' % (item.tostring(), dict_data_saved)) @@ -564,7 +564,7 @@ def save(item, dict_data_saved): dict_series = jsontools.get_node_from_file(item.from_channel, TAG_TVSHOW_FILTER) tvshow = item.show.strip().lower() - logger.log("Data is updated") + logger.info("Data is updated") list_quality = [] for _id, value in list(dict_data_saved.items()): @@ -599,7 +599,7 @@ def save_from_context(item): @param item: item @type item: item """ - logger.log() + logger.info() dict_series = jsontools.get_node_from_file(item.from_channel, TAG_TVSHOW_FILTER) tvshow = item.show.strip().lower() @@ -630,7 +630,7 @@ def delete_from_context(item): @param item: item @type item: item """ - logger.log() + logger.info() # We come from get_links and no result has been obtained, in context menu and we delete if item.to_channel != "": diff --git a/specials/help.py b/specials/help.py index 8ce57173..53d9cac8 100644 --- a/specials/help.py +++ b/specials/help.py @@ -37,7 +37,7 @@ if config.is_xbmc(): def mainlist(item): - logger.log() + logger.info() itemlist = [] if config.is_xbmc(): diff --git a/specials/infoplus.py b/specials/infoplus.py index 4cf93103..3209f82d 100644 --- a/specials/infoplus.py +++ b/specials/infoplus.py @@ -432,7 +432,7 @@ class main(xbmcgui.WindowDialog): else: for button, title, id, poster in self.idps: - logger.log('INFOS',button, title, id, poster) + logger.info('INFOS',button, title, id, poster) if control == button: new_item=self.item.clone(title=title, infoLabels={'tmdb_id':id}, contentType=self.item.contentType) self.close() @@ -602,7 +602,7 @@ class GlobalSearch(xbmcgui.WindowXMLDialog): def globalSearch(item, infoLabels, org_title=False): - logger.log() + logger.info() if item.contentType == "movie": cat = ["movie"] else: cat = ["serie"] cat += ["infoPlus"] diff --git a/specials/kodfavorites.py b/specials/kodfavorites.py index bfed74b0..97274fcb 100644 --- a/specials/kodfavorites.py +++ b/specials/kodfavorites.py @@ -126,7 +126,7 @@ class KodfavouritesData(object): # ============================ def addFavourite(item): - logger.log() + logger.info() alfav = KodfavouritesData() # If you get here through the context menu, you must retrieve the action and channel parameters @@ -179,7 +179,7 @@ def addFavourite(item): # ==================== def mainlist(item): - logger.log() + logger.info() alfav = KodfavouritesData() item.category = get_name_from_filename(os.path.basename(alfav.user_favorites_file)) @@ -216,7 +216,7 @@ def mainlist(item): def mostrar_perfil(item): - logger.log() + logger.info() alfav = KodfavouritesData() itemlist = [] @@ -289,7 +289,7 @@ def _crea_perfil(alfav): # Profile and link management def crear_perfil(item): - logger.log() + logger.info() alfav = KodfavouritesData() if not _crea_perfil(alfav): return False @@ -299,7 +299,7 @@ def crear_perfil(item): def editar_perfil_titulo(item): - logger.log() + logger.info() alfav = KodfavouritesData() if not alfav.user_favorites[item.i_perfil]: return False @@ -316,7 +316,7 @@ def editar_perfil_titulo(item): def eliminar_perfil(item): - logger.log() + logger.info() alfav = KodfavouritesData() if not alfav.user_favorites[item.i_perfil]: return False @@ -332,7 +332,7 @@ def eliminar_perfil(item): def acciones_enlace(item): - logger.log() + logger.info() acciones = [config.get_localized_string(70620), config.get_localized_string(70621), config.get_localized_string(70622), config.get_localized_string(70623), config.get_localized_string(70624), config.get_localized_string(70548), config.get_localized_string(70625), @@ -364,7 +364,7 @@ def acciones_enlace(item): def editar_enlace_titulo(item): - logger.log() + logger.info() alfav = KodfavouritesData() if not alfav.user_favorites[item.i_perfil]: return False @@ -386,7 +386,7 @@ def editar_enlace_titulo(item): def editar_enlace_color(item): - logger.log() + logger.info() alfav = KodfavouritesData() if not alfav.user_favorites[item.i_perfil]: return False @@ -410,7 +410,7 @@ def editar_enlace_color(item): def editar_enlace_thumbnail(item): - logger.log() + logger.info() alfav = KodfavouritesData() if not alfav.user_favorites[item.i_perfil]: return False @@ -470,7 +470,7 @@ def editar_enlace_thumbnail(item): def editar_enlace_carpeta(item): - logger.log() + logger.info() alfav = KodfavouritesData() if not alfav.user_favorites[item.i_perfil]: return False @@ -489,7 +489,7 @@ def editar_enlace_carpeta(item): def editar_enlace_lista(item): - logger.log() + logger.info() alfav = KodfavouritesData() if not alfav.user_favorites[item.i_perfil]: return False @@ -527,7 +527,7 @@ def editar_enlace_lista(item): def eliminar_enlace(item): - logger.log() + logger.info() alfav = KodfavouritesData() if not alfav.user_favorites[item.i_perfil]: return False @@ -542,7 +542,7 @@ def eliminar_enlace(item): # Move profiles and links (up, down, top, bottom) def mover_perfil(item): - logger.log() + logger.info() alfav = KodfavouritesData() alfav.user_favorites = _mover_item(alfav.user_favorites, item.i_perfil, item.direccion) @@ -552,7 +552,7 @@ def mover_perfil(item): return True def mover_enlace(item): - logger.log() + logger.info() alfav = KodfavouritesData() if not alfav.user_favorites[item.i_perfil]: return False @@ -597,7 +597,7 @@ def _mover_item(lista, i_selected, direccion): # ------------------------------------------ def mainlist_listas(item): - logger.log() + logger.info() itemlist = [] item.category = 'Listas' @@ -619,7 +619,7 @@ def mainlist_listas(item): def acciones_lista(item): - logger.log() + logger.info() acciones = [config.get_localized_string(70604), config.get_localized_string(70629), config.get_localized_string(70605), config.get_localized_string(70606), config.get_localized_string(70607)] @@ -641,7 +641,7 @@ def acciones_lista(item): def activar_lista(item): - logger.log() + logger.info() fullfilename = os.path.join(config.get_data_path(), item.lista) if not os.path.exists(fullfilename): @@ -659,7 +659,7 @@ def activar_lista(item): def renombrar_lista(item): - logger.log() + logger.info() fullfilename_current = os.path.join(config.get_data_path(), item.lista) if not os.path.exists(fullfilename_current): @@ -695,7 +695,7 @@ def renombrar_lista(item): def eliminar_lista(item): - logger.log() + logger.info() fullfilename = os.path.join(config.get_data_path(), item.lista) if not os.path.exists(fullfilename): @@ -714,7 +714,7 @@ def eliminar_lista(item): def informacion_lista(item): - logger.log() + logger.info() fullfilename = os.path.join(config.get_data_path(), item.lista) if not os.path.exists(fullfilename): @@ -741,7 +741,7 @@ def informacion_lista(item): def compartir_lista(item): - logger.log() + logger.info() fullfilename = os.path.join(config.get_data_path(), item.lista) if not os.path.exists(fullfilename): @@ -796,7 +796,7 @@ def compartir_lista(item): def acciones_nueva_lista(item): - logger.log() + logger.info() acciones = [config.get_localized_string(70651), config.get_localized_string(70652), @@ -830,7 +830,7 @@ def acciones_nueva_lista(item): def crear_lista(item): - logger.log() + logger.info() titulo = platformtools.dialog_input(default='', heading=config.get_localized_string(70612)) if titulo is None or titulo == '': @@ -853,7 +853,7 @@ def crear_lista(item): def descargar_lista(item, url): - logger.log() + logger.info() from core import httptools, scrapertools if 'tinyupload.com/' in url: diff --git a/specials/news.py b/specials/news.py index b22ddc9b..ea3096fc 100644 --- a/specials/news.py +++ b/specials/news.py @@ -42,7 +42,7 @@ menu_settings_path = os.path.join(config.get_data_path(), "settings_channels", ' def mainlist(item): - logger.log() + logger.info() itemlist = [] # list_canales, any_active = get_channels_list() @@ -128,7 +128,7 @@ def set_category_context(item): def get_channels_list(): - logger.log() + logger.info() ## import web_pdb; web_pdb.set_trace() ## list_canales = {'peliculas': [], '4k': [], 'terror': [], 'infantiles': [], 'series': [], 'anime': [], ## 'castellano': [], 'latino':[], 'italiano':[], 'torrent':[], 'documentales': []} @@ -166,14 +166,14 @@ def get_channels_list(): return list_canales, any_active def set_cache(item): - logger.log() + logger.info() item.mode = 'set_cache' t = Thread(target=novedades, args=[item]) t.start() #t.join() def get_from_cache(item): - logger.log() + logger.info() itemlist=[] cache_node = jsontools.get_node_from_file('menu_cache_data.json', 'cached') first=item.last @@ -198,7 +198,7 @@ def get_from_cache(item): return itemlist def add_menu_items(item, itemlist): - logger.log() + logger.info() menu_icon = get_thumb('menu.png') menu = Item(channel="channelselector", action="getmainlist", viewmode="movie", thumbnail=menu_icon, title='Menu') @@ -215,7 +215,7 @@ def add_menu_items(item, itemlist): return itemlist def novedades(item): - logger.log() + logger.info() global list_newest threads = [] @@ -231,7 +231,7 @@ def novedades(item): return get_from_cache(item) multithread = config.get_setting("multithread", "news") - logger.log("multithread= " + str(multithread)) + logger.info("multithread= " + str(multithread)) if not multithread: if platformtools.dialog_yesno(config.get_localized_string(60515), @@ -270,7 +270,7 @@ def novedades(item): # if progreso.iscanceled(): # progreso.close() - # logger.log("Búsqueda cancelada") + # logger.info("Búsqueda cancelada") # return itemlist # Modo Multi Thread @@ -284,7 +284,7 @@ def novedades(item): # Modo single Thread else: if mode == 'normal': - logger.log("Obteniendo novedades de channel_id=" + channel_id) + logger.info("Obteniendo novedades de channel_id=" + channel_id) progreso.update(percentage, "", config.get_localized_string(60520) % channel_title) get_newest(channel_id, item.extra) @@ -304,7 +304,7 @@ def novedades(item): logger.debug(mensaje) if progreso.iscanceled(): - logger.log("Busqueda de novedades cancelada") + logger.info("Busqueda de novedades cancelada") break time.sleep(0.5) @@ -312,7 +312,7 @@ def novedades(item): if mode == 'normal': mensaje = config.get_localized_string(60522) % (len(list_newest), time.time() - start_time) progreso.update(100, mensaje) - logger.log(mensaje) + logger.info(mensaje) start_time = time.time() # logger.debug(start_time) @@ -345,7 +345,7 @@ def novedades(item): def get_newest(channel_id, categoria): - logger.log("channel_id=" + channel_id + ", categoria=" + categoria) + logger.info("channel_id=" + channel_id + ", categoria=" + categoria) global list_newest global list_newest_tourl @@ -366,9 +366,9 @@ def get_newest(channel_id, categoria): if not puede: return - logger.log("running channel " + modulo.__name__ + " " + modulo.__file__) + logger.info("running channel " + modulo.__name__ + " " + modulo.__file__) list_result = modulo.newest(categoria) - logger.log("canal= %s %d resultados" % (channel_id, len(list_result))) + logger.info("canal= %s %d resultados" % (channel_id, len(list_result))) exist=False if os.path.exists(menu_cache_path): cache_node = jsontools.get_node_from_file('menu_cache_data.json', 'cached') @@ -377,7 +377,7 @@ def get_newest(channel_id, categoria): cache_node = {} # logger.debug('cache node: %s' % cache_node) for item in list_result: - # logger.log("item="+item.tostring()) + # logger.info("item="+item.tostring()) item.channel = channel_id list_newest.append(item) list_newest_tourl.append(item.tourl()) @@ -393,7 +393,7 @@ def get_newest(channel_id, categoria): def get_title(item): - # support.log("ITEM NEWEST ->", item) + # support.info("ITEM NEWEST ->", item) # item.contentSerieName c'è anche se è un film if item.contentSerieName and item.contentType != 'movie': # Si es una serie title = item.contentSerieName @@ -446,7 +446,7 @@ def no_group(list_result_canal): global channels_id_name for i in list_result_canal: - # support.log("NO GROUP i -> ", i) + # support.info("NO GROUP i -> ", i) canale = channels_id_name[i.channel] canale = canale # to differentiate it from the color of the other items i.title = get_title(i) + " [" + canale + "]" @@ -532,7 +532,7 @@ def group_by_content(list_result_canal): def show_channels(item): - logger.log() + logger.info() global channels_id_name channels_id_name = item.extra itemlist = [] diff --git a/specials/resolverdns.py b/specials/resolverdns.py index 73c9dceb..2b3bc896 100644 --- a/specials/resolverdns.py +++ b/specials/resolverdns.py @@ -66,13 +66,13 @@ class CipherSuiteAdapter(host_header_ssl.HostHeaderSSLAdapter): try: self.cur.execute('select ip from dnscache where domain=?', (domain,)) ip = self.cur.fetchall()[0][0] - logger.log('Cache DNS: ' + domain + ' = ' + str(ip)) + logger.info('Cache DNS: ' + domain + ' = ' + str(ip)) except: pass if not ip: # not cached try: ip = doh.query(domain)[0] - logger.log('Query DoH: ' + domain + ' = ' + str(ip)) + logger.info('Query DoH: ' + domain + ' = ' + str(ip)) self.writeToCache(domain, ip) except Exception: logger.error('Failed to resolve hostname, fallback to normal dns') @@ -133,8 +133,8 @@ class CipherSuiteAdapter(host_header_ssl.HostHeaderSSLAdapter): try: ret = super(CipherSuiteAdapter, self).send(request, **kwargs) except Exception as e: - logger.log('Request for ' + domain + ' with ip ' + ip + ' failed') - logger.log(e) + logger.info('Request for ' + domain + ' with ip ' + ip + ' failed') + logger.info(e) # if 'SSLError' in str(e): # # disabilito # config.set_setting("resolver_dns", False) @@ -143,7 +143,7 @@ class CipherSuiteAdapter(host_header_ssl.HostHeaderSSLAdapter): # else: tryFlush = True if tryFlush and not flushedDns: # re-request ips and update cache - logger.log('Flushing dns cache for ' + domain) + logger.info('Flushing dns cache for ' + domain) return self.flushDns(request, domain, **kwargs) ret.url = realUrl else: diff --git a/specials/search.py b/specials/search.py index a19f41c9..6052b655 100644 --- a/specials/search.py +++ b/specials/search.py @@ -31,7 +31,7 @@ def_lang = info_language[config.get_setting("info_language", "videolibrary")] def mainlist(item): - logger.log() + logger.info() itemlist = [Item(channel=item.channel, title=config.get_localized_string(70276), action='new_search', mode='all', thumbnail=get_thumb("search.png")), Item(channel=item.channel, title=config.get_localized_string(70741) % config.get_localized_string(30122), action='new_search', mode='movie', thumbnail=get_thumb("search_movie.png")), @@ -48,7 +48,7 @@ def mainlist(item): def sub_menu(item): - logger.log() + logger.info() itemlist = [Item(channel=item.channel, action='genres_menu', title=config.get_localized_string(70306), mode='movie', thumbnail=get_thumb("movie_genre.png")), Item(channel=item.channel, action='years_menu', title=config.get_localized_string(70742), mode='movie', thumbnail=get_thumb("movie_year.png")), @@ -66,7 +66,7 @@ def sub_menu(item): def saved_search(item): - logger.log() + logger.info() itemlist = list() saved_searches_list = get_saved_searches() @@ -93,7 +93,7 @@ def saved_search(item): def new_search(item): - logger.log() + logger.info() temp_search_file = config.get_temp_file('temp-search') if filetools.isfile(temp_search_file): @@ -163,7 +163,7 @@ def new_search(item): def channel_search(item): - logger.log(item) + logger.debug(item) start = time.time() searching = list() @@ -389,14 +389,14 @@ def get_servers(item, module_dict): def get_info(itemlist): - logger.log() + logger.info() tmdb.set_infoLabels_itemlist(itemlist, True, forced=True) return itemlist def get_channels(item): - logger.log() + logger.info() channels_list = list() title_list = list() @@ -717,7 +717,7 @@ def discover_list(item): def from_context(item): - logger.log() + logger.info() select = setting_channel_new(item) @@ -738,7 +738,7 @@ def from_context(item): def set_context(itemlist): - logger.log() + logger.info() for elem in itemlist: elem.context = [{"title": config.get_localized_string(60412), @@ -755,7 +755,7 @@ def set_context(itemlist): def get_from_temp(item): - logger.log() + logger.info() n = 30 nTotal = len(item.itemlist) diff --git a/specials/setting.py b/specials/setting.py index 8926e254..895ffc27 100644 --- a/specials/setting.py +++ b/specials/setting.py @@ -21,7 +21,7 @@ CHANNELNAME = "setting" AUTOSTART = config.is_autorun_enabled() def mainlist(item): - logger.log() + logger.info() itemlist = list() itemlist.append(Item(channel=CHANNELNAME, title=config.get_localized_string(60535), action="settings", folder=False, @@ -62,7 +62,7 @@ def mainlist(item): def menu_channels(item): - logger.log() + logger.info() itemlist = list() itemlist.append(Item(channel=CHANNELNAME, title=config.get_localized_string(60545), action="conf_tools", folder=False, @@ -98,15 +98,15 @@ def channel_config(item): def autostart(item): # item required launcher.py line 265 if config.enable_disable_autorun(AUTOSTART): - logger.log('AUTOSTART ENABLED') + logger.info('AUTOSTART ENABLED') # xbmcgui.Dialog().ok(config.get_localized_string(20000), config.get_localized_string(70709)) else: - logger.log('AUTOSTART ENABLED') + logger.info('AUTOSTART ENABLED') # xbmcgui.Dialog().ok(config.get_localized_string(20000), config.get_localized_string(70710)) # def setting_torrent(item): -# logger.log() +# logger.info() # LIBTORRENT_PATH = config.get_setting("libtorrent_path", server="torrent", default="") # LIBTORRENT_ERROR = config.get_setting("libtorrent_error", server="torrent", default="") @@ -243,7 +243,7 @@ def autostart(item): # item required launcher.py line 265 # config.set_setting("magnet2torrent", dict_data_saved["magnet2torrent"], server="torrent") def menu_servers(item): - logger.log() + logger.info() itemlist = list() itemlist.append(Item(channel=CHANNELNAME, title=config.get_localized_string(60550), action="servers_blacklist", folder=False, @@ -272,7 +272,7 @@ def menu_servers(item): for server in sorted(server_list): server_parameters = servertools.get_server_parameters(server) - logger.log(server_parameters) + logger.info(server_parameters) if server_parameters["has_settings"] and [x for x in server_parameters["settings"] if x["id"] not in ["black_list", "white_list"]]: itemlist.append( Item(channel=CHANNELNAME, title=". " + config.get_localized_string(60553) % server_parameters["name"], @@ -406,7 +406,7 @@ def cb_servers_favorites(server_names, dict_values): i += 1 c = 1 - logger.log(dict_favorites) + logger.info(dict_favorites) favorites_servers_list = [] while c in dict_favorites: favorites_servers_list.append(dict_favorites[c]) @@ -424,7 +424,7 @@ def settings(item): def submenu_tools(item): - logger.log() + logger.info() itemlist = list() # Custom tools @@ -465,7 +465,7 @@ def submenu_tools(item): def check_quickfixes(item): - logger.log() + logger.info() if not config.dev_mode(): from platformcode import updater @@ -476,7 +476,7 @@ def check_quickfixes(item): # def update_quasar(item): -# logger.log() +# logger.info() # from platformcode import custom_code, platformtools # stat = False @@ -488,7 +488,7 @@ def check_quickfixes(item): def conf_tools(item): - logger.log() + logger.info() # Enable or disable channels if item.extra == "channels_onoff": @@ -590,14 +590,14 @@ def conf_tools(item): action="", folder=False, thumbnail=channel.thumbnail)) continue - # logger.log(channel.channel + " SALTADO!") + # logger.info(channel.channel + " SALTADO!") # The json file settings of the channel are loaded file_settings = os.path.join(config.get_data_path(), "settings_channels", channel.channel + "_data.json") dict_settings = {} dict_file = {} if filetools.exists(file_settings): - # logger.log(channel.channel + " Has _data.json file") + # logger.info(channel.channel + " Has _data.json file") channeljson_exists = True # We get saved settings from ../settings/channel_data.json try: @@ -607,7 +607,7 @@ def conf_tools(item): except EnvironmentError: logger.error("ERROR when reading the file: %s" % file_settings) else: - # logger.log(channel.channel + " No _data.json file") + # logger.info(channel.channel + " No _data.json file") channeljson_exists = False if channeljson_exists: @@ -627,7 +627,7 @@ def conf_tools(item): # Default settings are loaded list_controls, default_settings = channeltools.get_channel_controls_settings( channel.channel) - # logger.log(channel.title + " | Default: %s" % default_settings) + # logger.info(channel.title + " | Default: %s" % default_settings) except: import traceback logger.error(channel.title + config.get_localized_string(60570) % traceback.format_exc()) @@ -649,7 +649,7 @@ def conf_tools(item): list_status = config.get_localized_string(60571) else: - # logger.log(channel.channel + " - NO correction needed!") + # logger.info(channel.channel + " - NO correction needed!") needsfix = False # If the channel status has been set it is added to the list @@ -748,7 +748,7 @@ def channel_status(item, dict_values): for k in dict_values: if k == "all_channels": - logger.log("All channels | Selected state: %s" % dict_values[k]) + logger.info("All channels | Selected state: %s" % dict_values[k]) if dict_values[k] != 0: excluded_channels = ['url', 'search', 'videolibrary', 'setting', @@ -789,9 +789,9 @@ def channel_status(item, dict_values): continue else: - logger.log("Channel: %s | State: %s" % (k, dict_values[k])) + logger.info("Channel: %s | State: %s" % (k, dict_values[k])) config.set_setting("enabled", dict_values[k], k) - logger.log("the value is like %s " % config.get_setting("enabled", k)) + logger.info("the value is like %s " % config.get_setting("enabled", k)) platformtools.itemlist_update(Item(channel=CHANNELNAME, action="mainlist")) @@ -879,7 +879,7 @@ def restore_tools(item): def report_menu(item): - logger.log('URL: ' + item.url) + logger.info('URL: ' + item.url) from channelselector import get_thumb @@ -952,7 +952,7 @@ def report_menu(item): def activate_debug(item): - logger.log(item.extra) + logger.info(item.extra) from platformcode import platformtools #Enable / disable DEBUB option in settings.xml @@ -1232,7 +1232,7 @@ def report_send(item, description='', fatal=False): continue status = True # Upload operation completed successfully - logger.log('Report created: ' + str(item.url)) # The URL of the user report is saved + logger.info('Report created: ' + str(item.url)) # The URL of the user report is saved # if fatal: # For future use, for logger.crash # platformtools.dialog_ok('KoD CREATED ERROR report', 'Report it in the forum by adding FATAL ERROR and this URL: ', '[COLOR gold]%s[/COLOR]' % item.url, pastebin_one_use_msg) # else: # Report URL passed to user diff --git a/specials/shortcuts.py b/specials/shortcuts.py index 7be05e2e..c8a05776 100644 --- a/specials/shortcuts.py +++ b/specials/shortcuts.py @@ -126,7 +126,7 @@ def SettingOnPosition(item): xbmc.executebuiltin('Addon.OpenSettings(plugin.video.kod)') category = item.category if item.category else 0 setting = item.setting if item.setting else 0 - logger.log('SETTING= ' + str(setting)) + logger.info('SETTING= ' + str(setting)) xbmc.executebuiltin('SetFocus(%i)' % (category - 100)) xbmc.executebuiltin('SetFocus(%i)' % (setting - 80)) diff --git a/specials/side_menu.py b/specials/side_menu.py index 3d0d765f..b4d4e962 100644 --- a/specials/side_menu.py +++ b/specials/side_menu.py @@ -43,7 +43,7 @@ def set_menu_settings(item): jsontools.update_node(menu_node, 'menu_settings_data.json', "menu") def check_user_home(item): - logger.log() + logger.info() if os.path.exists(menu_settings_path): menu_node = jsontools.get_node_from_file('menu_settings_data.json', 'menu') if 'user_home' in menu_node: @@ -55,7 +55,7 @@ def check_user_home(item): return item def set_custom_start(item): - logger.log() + logger.info() if os.path.exists(menu_settings_path): menu_node = jsontools.get_node_from_file('menu_settings_data.json', 'menu') else: @@ -69,7 +69,7 @@ def set_custom_start(item): jsontools.update_node(menu_node, 'menu_settings_data.json', "menu") def get_start_page(): - logger.log() + logger.info() dictCategory = { config.get_localized_string(70137): 'peliculas', @@ -356,7 +356,7 @@ class Main(xbmcgui.WindowXMLDialog): self.focus -= 1 def run_action(self, item): - logger.log() + logger.info() if item.menu != True: self.close() xbmc.executebuiltin("Container.update(%s)"%launcher.run(item)) diff --git a/specials/trailertools.py b/specials/trailertools.py index 096a1c2a..ce8b253f 100644 --- a/specials/trailertools.py +++ b/specials/trailertools.py @@ -43,7 +43,7 @@ else: def buscartrailer(item, trailers=[]): - logger.log() + logger.info() # List of actions if run from context menu if item.action == "manual_search" and item.contextual: @@ -78,8 +78,8 @@ def buscartrailer(item, trailers=[]): item.year = item.infoLabels['year'] - logger.log("Search: %s" % item.contentTitle) - logger.log("Year: %s" % item.year) + logger.info("Search: %s" % item.contentTitle) + logger.info("Year: %s" % item.year) if item.infoLabels['trailer'] and not trailers: url = item.infoLabels['trailer'] if "youtube" in url: @@ -122,7 +122,7 @@ def buscartrailer(item, trailers=[]): def manual_search(item): - logger.log() + logger.info() texto = platformtools.dialog_input(default=item.contentTitle, heading=config.get_localized_string(30112)) if texto is not None: if item.extra == "mymovies": @@ -134,7 +134,7 @@ def manual_search(item): def tmdb_trailers(item, tipo="movie"): - logger.log() + logger.info() from core.tmdb import Tmdb itemlist = [] @@ -153,7 +153,7 @@ def tmdb_trailers(item, tipo="movie"): def youtube_search(item): - logger.log() + logger.info() itemlist = [] title = item.contentTitle if item.extra != "youtube": @@ -192,7 +192,7 @@ def youtube_search(item): def mymovies_search(item): - logger.log() + logger.info() import json title = item.contentTitle @@ -216,7 +216,7 @@ def mymovies_search(item): def search_links_mymovies(item): - logger.log() + logger.info() trailer_url = match(item, patron=r'<li class="bottone_playlist"[^>]+><a href="([^"]+)"').match itemlist = [] data = httptools.downloadpage(item.url).data @@ -236,7 +236,7 @@ def search_links_mymovies(item): def filmaffinity_search(item): - logger.log() + logger.info() if item.filmaffinity: item.url = item.filmaffinity @@ -284,7 +284,7 @@ def filmaffinity_search(item): def search_links_filmaff(item): - logger.log() + logger.info() itemlist = [] data = httptools.downloadpage(item.url).data diff --git a/specials/tvmoviedb.py b/specials/tvmoviedb.py index 02e1b14c..b0ab08db 100644 --- a/specials/tvmoviedb.py +++ b/specials/tvmoviedb.py @@ -1,11 +1,11 @@ # -*- coding: utf-8 -*- import re -from core import support, filetools, jsontools, trakt_tools +from core import filetools, jsontools, trakt_tools from core.tmdb import Tmdb from core.scrapertools import htmlclean, decodeHtmlentities -from core.support import thumb, typo, match, Item, dbg +from core.support import thumb, typo, match, Item from platformcode import config, platformtools -from platformcode.logger import log, error +from platformcode.logger import info, error info_language = ["de", "en", "es", "fr", "it", "pt"] # from videolibrary.json def_lang = info_language[config.get_setting("info_language", "videolibrary")] @@ -28,7 +28,7 @@ default_fan = filetools.join(config.get_runtime_path(), "fanart.jpg") def mainlist(item): - log() + info() itemlist = [ # TMDB # item.clone(title=typo(config.get_localized_string(70021), 'bold'), action=""), @@ -55,7 +55,7 @@ def configuracion(item): return ret def search_star(item): - log() + info() itemlist = [] item.type='movie' @@ -97,7 +97,7 @@ def search_(item): def searcing(item): - log() + info() new_item = Item(title=item.contentTitle, text=item.contentTitle.replace("+", " "), mode=item.contentType, infoLabels=item.infoLabels) @@ -625,7 +625,7 @@ def indices_tmdb(item): def filter(item): - log() + info() from datetime import datetime list_controls = [] @@ -705,7 +705,7 @@ def filtered(item, values): def musica_movie(item): - log() + info() itemlist = [] data = match(item).data matches = match(data, patron=r'<td class="left">([^<]+)<br><small>([^<]+)</small>.*?<td>(\d+:\d+).*?<p id="([^"]+)"').matches @@ -729,7 +729,7 @@ def list_imdb(item): url = 'http://www.imdb.com/search/title?' + item.url # data = httptools.downloadpage(url, headers=headers, replace_headers=True).data data = match(url, headers=headers).data - log(data) + info(data) # data = re.sub(r"\n|\r|\t| ", "", data) # data = re.sub(r"\s{2}", " ", data) @@ -812,7 +812,7 @@ def list_imdb(item): def filter_imdb(item): - log() + info() from datetime import datetime list_controls = [] @@ -1392,7 +1392,7 @@ def indices_imdb(item): # def filter_fa(item): -# log() +# info() # from datetime import datetime # list_controls = [] @@ -1494,7 +1494,7 @@ def indices_imdb(item): # def login_fa(): -# log() +# info() # try: # user = config.get_setting("usuariofa", "tvmoviedb") @@ -1519,7 +1519,7 @@ def indices_imdb(item): # userid = scrapertools.find_single_match(data, 'id-user=(\d+)') # if userid: # config.set_setting("userid", userid, "tvmoviedb") -# log("Login correcto") +# info("Login correcto") # return True, "" # except: # import traceback @@ -1644,7 +1644,7 @@ def indices_imdb(item): # def votar_fa(item): # # Window to select the vote -# log() +# info() # list_controls = [] # valores = {} @@ -1920,7 +1920,7 @@ def acciones_trakt(item): ratings = [] try: for i, entry in enumerate(data): - log('ENTRY:',entry) + info('ENTRY:',entry) if i <= item.pagina: continue # try: entry = entry[item.args] # except: pass @@ -1996,7 +1996,7 @@ def acciones_trakt(item): def order_list(item): - log() + info() list_controls = [] valores1 = ['rating', 'added', 'title', 'released', 'runtime', 'popularity', 'percentage', 'votes'] @@ -2338,7 +2338,7 @@ def indices_mal(item): matches = match("https://myanimelist.net/anime.php", cookies=False, patronBlock=patronBlock, patron=patron).matches for url, title in matches: genero = title.split(" (", 1)[0] - log(url_base, genero) + info(url_base, genero) thumbnail = url_base + genero.lower().replace(" ", "%20") if genero in ["Hentai", "Yaoi", "Yuri"] and not adult_mal: continue @@ -2487,7 +2487,7 @@ def detail_staff(item): patron_bio = r'<?<div class="spaceit_pad">(.*?)</td>' bio = match(data, patron=patron_bio).match bio = htmlclean(bio.replace("</div>", "\n")) - log(bio) + info(bio) infoLabels = {'plot': bio} if not "No voice acting roles" in data: itemlist.append(Item(channel=item.channel, title=typo(config.get_localized_string(70374),'bold bullet'), action="", thumbnail=item.thumbnail, infoLabels=infoLabels)) @@ -2625,7 +2625,7 @@ def info_anidb(item, itemlist, url): def filter_mal(item): - log() + info() list_controls = [] valores = {} @@ -2703,7 +2703,7 @@ def callback_mal(item, values): def musica_anime(item): # List available anime and songs similar to the anime title - log() + info() itemlist = [] data = match("http://www.freeanimemusic.org/song_search.php", post=item.post).data @@ -2738,7 +2738,7 @@ def musica_anime(item): def login_mal(from_list=False): - log() + info() from core import httptools from base64 import b64decode as bdec @@ -2767,7 +2767,7 @@ def login_mal(from_list=False): else: if generic: return False, config.get_localized_string(70393), user - log("Correct login") + info("Correct login") return True, "", user except: import traceback @@ -2799,7 +2799,7 @@ def cuenta_mal(item): def items_mal(item): # Scraper for personal lists - log() + info() itemlist = [] data = match(item.url).data diff --git a/specials/url.py b/specials/url.py index d808747e..04e1544e 100644 --- a/specials/url.py +++ b/specials/url.py @@ -1,13 +1,13 @@ # -*- coding: utf-8 -*- from core import servertools -from core.support import match, log +from core.support import match, info from core.item import Item from platformcode import config, logger def mainlist(item): - log() + info() itemlist = [] itemlist.append(Item(channel=item.channel, action="search", title=config.get_localized_string(60089), thumbnail=item.thumbnail, args='server')) @@ -19,7 +19,7 @@ def mainlist(item): # When the function "search" is called, the launcher asks for a text to search for and adds it as a parameter def search(item, text): - log(text) + info(text) if not text.startswith("http"): text = "http://" + text diff --git a/specials/videolibrary.py b/specials/videolibrary.py index 252e945c..00ef56e5 100644 --- a/specials/videolibrary.py +++ b/specials/videolibrary.py @@ -21,7 +21,7 @@ else: def mainlist(item): - logger.log() + logger.info() itemlist = [Item(channel=item.channel, action="list_movies", title=config.get_localized_string(60509), category=config.get_localized_string(70270), thumbnail=thumb("videolibrary_movie")), @@ -38,7 +38,7 @@ def channel_config(item): def list_movies(item, silent=False): - logger.log() + logger.info() itemlist = [] movies_path = [] for root, folders, files in filetools.walk(videolibrarytools.MOVIES_PATH): @@ -64,7 +64,7 @@ def list_movies(item, silent=False): def list_tvshows(item): from time import time start = time() - logger.log() + logger.info() itemlist = [] lista = [] tvshows_path = [] @@ -89,7 +89,7 @@ def list_tvshows(item): title=typo(config.get_localized_string(70269), 'bold color kod'), folder=False), Item(channel=item.channel, action="configure_update_videolibrary", thumbnail=item.thumbnail, title=typo(config.get_localized_string(60599), 'bold color kod'), lista=lista, folder=False)] - logger.log('TEMPO= ' + str(time() - start)) + logger.info('TEMPO= ' + str(time() - start)) return itemlist @@ -189,7 +189,7 @@ def get_results(nfo_path, root, Type, local=False): # Contextual menu: Mark as seen / not seen visto = item.library_playcounts.get(item.contentTitle, 0) item.infoLabels["playcount"] = visto - logger.log('item\n' + str(item)) + logger.info('item\n' + str(item)) if visto > 0: seen_text = config.get_localized_string(60020) counter = 0 @@ -261,7 +261,7 @@ def configure_update_videolibrary(item): def get_seasons(item): - logger.log() + logger.info() # logger.debug("item:\n" + item.tostring('\n')) itemlist = [] dict_temp = {} @@ -320,7 +320,7 @@ def get_seasons(item): def get_episodes(item): - logger.log() + logger.info() # logger.debug("item:\n" + item.tostring('\n')) itemlist = [] @@ -384,7 +384,7 @@ def get_episodes(item): def findvideos(item): from specials import autoplay - logger.log() + logger.info() # logger.debug("item:\n" + item.tostring('\n')) videolibrarytools.check_renumber_options(item) itemlist = [] @@ -537,7 +537,7 @@ def findvideos(item): def play(item): - logger.log() + logger.info() # logger.debug("item:\n" + item.tostring('\n')) if not item.contentChannel == "local": @@ -577,7 +577,7 @@ def play(item): def update_videolibrary(item=''): - logger.log() + logger.info() # Update active series by overwriting import service @@ -597,17 +597,17 @@ def update_videolibrary(item=''): def move_videolibrary(current_path, new_path, current_movies_folder, new_movies_folder, current_tvshows_folder, new_tvshows_folder): - logger.log() + logger.info() backup_current_path = current_path backup_new_path = new_path - logger.log('current_path: ' + current_path) - logger.log('new_path: ' + new_path) - logger.log('current_movies_folder: ' + current_movies_folder) - logger.log('new_movies_folder: ' + new_movies_folder) - logger.log('current_tvshows_folder: ' + current_tvshows_folder) - logger.log('new_tvshows_folder: ' + new_tvshows_folder) + logger.info('current_path: ' + current_path) + logger.info('new_path: ' + new_path) + logger.info('current_movies_folder: ' + current_movies_folder) + logger.info('new_movies_folder: ' + new_movies_folder) + logger.info('current_tvshows_folder: ' + current_tvshows_folder) + logger.info('new_tvshows_folder: ' + new_tvshows_folder) notify = False progress = platformtools.dialog_progress_bg(config.get_localized_string(20000), config.get_localized_string(80011)) @@ -618,14 +618,14 @@ def move_videolibrary(current_path, new_path, current_movies_folder, new_movies_ current_tvshows_path = u'' + filetools.join(current_path, current_tvshows_folder) new_tvshows_path = u'' + filetools.join(new_path, new_tvshows_folder) - logger.log('current_movies_path: ' + current_movies_path) - logger.log('new_movies_path: ' + new_movies_path) - logger.log('current_tvshows_path: ' + current_tvshows_path) - logger.log('new_tvshows_path: ' + new_tvshows_path) + logger.info('current_movies_path: ' + current_movies_path) + logger.info('new_movies_path: ' + new_movies_path) + logger.info('current_tvshows_path: ' + current_tvshows_path) + logger.info('new_tvshows_path: ' + new_tvshows_path) from platformcode import xbmc_videolibrary movies_path, tvshows_path = xbmc_videolibrary.check_sources(new_movies_path, new_tvshows_path) - logger.log('check_sources: ' + str(movies_path) + ', ' + str(tvshows_path)) + logger.info('check_sources: ' + str(movies_path) + ', ' + str(tvshows_path)) if movies_path or tvshows_path: if not movies_path: filetools.rmdir(new_movies_path) @@ -670,7 +670,7 @@ def move_videolibrary(current_path, new_path, current_movies_folder, new_movies_ def delete_videolibrary(item): - logger.log() + logger.info() if not platformtools.dialog_yesno(config.get_localized_string(20000), config.get_localized_string(80037)): return @@ -696,7 +696,7 @@ def delete_videolibrary(item): # context menu methods def update_tvshow(item): - logger.log() + logger.info() # logger.debug("item:\n" + item.tostring('\n')) heading = config.get_localized_string(60037) @@ -722,11 +722,11 @@ def update_tvshow(item): def add_local_episodes(item): - logger.log() + logger.info() done, local_episodes_path = videolibrarytools.config_local_episodes_path(item.path, item, silent=True) if done < 0: - logger.log("An issue has occurred while configuring local episodes") + logger.info("An issue has occurred while configuring local episodes") elif local_episodes_path: nfo_path = filetools.join(item.path, "tvshow.nfo") head_nfo, item_nfo = videolibrarytools.read_nfo(nfo_path) @@ -741,7 +741,7 @@ def add_local_episodes(item): def remove_local_episodes(item): - logger.log() + logger.info() nfo_path = filetools.join(item.path, "tvshow.nfo") head_nfo, item_nfo = videolibrarytools.read_nfo(nfo_path) @@ -759,7 +759,7 @@ def remove_local_episodes(item): def verify_playcount_series(item, path): - logger.log() + logger.info() """ This method reviews and repairs the PlayCount of a series that has become out of sync with the actual list of episodes in its folder. Entries for missing episodes, seasons, or series are created with the "not seen" mark. Later it is sent to verify the counters of Seasons and Series @@ -822,7 +822,7 @@ def verify_playcount_series(item, path): def mark_content_as_watched2(item): - logger.log() + logger.info() # logger.debug("item:\n" + item.tostring('\n')) if filetools.isfile(item.nfo): head_nfo, it = videolibrarytools.read_nfo(item.nfo) @@ -860,7 +860,7 @@ def mark_content_as_watched2(item): def mark_content_as_watched(item): - logger.log() + logger.info() #logger.debug("item:\n" + item.tostring('\n')) if filetools.exists(item.nfo): @@ -898,7 +898,7 @@ def mark_content_as_watched(item): def mark_season_as_watched(item): - logger.log() + logger.info() # logger.debug("item:\n" + item.tostring('\n')) # Get dictionary of marked episodes @@ -951,7 +951,7 @@ def mark_season_as_watched(item): def mark_tvshow_as_updatable(item, silent=False): - logger.log() + logger.info() head_nfo, it = videolibrarytools.read_nfo(item.nfo) it.active = item.active filetools.write(item.nfo, head_nfo + it.tojson()) @@ -984,7 +984,7 @@ def delete(item): elif platformtools.dialog_yesno(heading, config.get_localized_string(70081) % filetools.basename(_item.path)): filetools.rmdirtree(_item.path) - logger.log("All links removed") + logger.info("All links removed") xbmc.sleep(1000) platformtools.itemlist_refresh() @@ -1055,7 +1055,7 @@ def delete(item): filetools.write(item.nfo, head_nfo + item_nfo.tojson()) msg_txt = config.get_localized_string(70087) % (num_enlaces, canal) - logger.log(msg_txt) + logger.info(msg_txt) platformtools.dialog_notification(heading, msg_txt) platformtools.itemlist_refresh() @@ -1065,7 +1065,7 @@ def delete(item): def check_season_playcount(item, season): - logger.log() + logger.info() if season: episodios_temporada = 0 @@ -1087,7 +1087,7 @@ def check_season_playcount(item, season): def check_tvshow_playcount(item, season): - logger.log() + logger.info() if season: temporadas_serie = 0 temporadas_vistas_serie = 0 diff --git a/tests/test_generic.py b/tests/test_generic.py index a488fd0b..35418ac1 100644 --- a/tests/test_generic.py +++ b/tests/test_generic.py @@ -224,7 +224,7 @@ class GenericChannelMenuItemTest(unittest.TestCase): break for resIt in self.itemlist: - logger.log(resIt.title + ' -> ' + resIt.url) + logger.info(resIt.title + ' -> ' + resIt.url) self.assertLess(len(resIt.fulltitle), 110, 'channel ' + self.ch + ' -> ' + self.title + ' might contain wrong titles<br>' + resIt.fulltitle) if resIt.url: diff --git a/updatetvshow.py b/updatetvshow.py index 6ab933c9..7227ec72 100644 --- a/updatetvshow.py +++ b/updatetvshow.py @@ -49,7 +49,7 @@ def search_paths(Id): def execute_sql(sql): - logger.log() + logger.info() file_db = "" records = None @@ -69,14 +69,14 @@ def execute_sql(sql): break if file_db: - logger.log("DB file: %s" % file_db) + logger.info("DB file: %s" % file_db) conn = None try: import sqlite3 conn = sqlite3.connect(file_db) cursor = conn.cursor() - logger.log("Running sql: %s" % sql) + logger.info("Running sql: %s" % sql) cursor.execute(sql) conn.commit() @@ -86,7 +86,7 @@ def execute_sql(sql): records = [] conn.close() - logger.log("Query executed. Records: %s" % nun_records) + logger.info("Query executed. Records: %s" % nun_records) except: logger.error("Error executing sql query")