From 4f498a05f8dc8df267912fc0965b0c11305992a9 Mon Sep 17 00:00:00 2001 From: Alhaziel01 Date: Wed, 25 Nov 2020 17:54:25 +0100 Subject: [PATCH] Minori informazioni con log disabilitato --- channels/altadefinizione01_link.py | 2 +- channels/cineblog01.py | 12 ++-- channels/film4k.py | 2 +- channels/guardaseriecam.py | 4 +- channels/netfreex.py | 2 +- channels/vedohd.py | 5 +- channels/youtube_channel.py | 14 ++-- channelselector.py | 16 ++--- core/autoplay.py | 4 +- core/channeltools.py | 14 ++-- core/downloader.py | 2 +- core/filetools.py | 4 +- core/filtertools.py | 26 +++---- core/httptools.py | 2 +- core/jsontools.py | 8 +-- core/scraper.py | 8 +-- core/scrapertools.py | 4 +- core/servertools.py | 30 ++++---- core/support.py | 12 ++-- core/tmdb.py | 18 ++--- core/trakt_tools.py | 14 ++-- core/tvdb.py | 28 ++++---- core/videolibrarytools.py | 56 +++++++-------- core/ziptools.py | 16 ++--- lib/arm_chromeos.py | 8 +-- lib/doh.py | 2 +- lib/generictools.py | 30 ++++---- platformcode/autorenumber.py | 14 ++-- platformcode/globalsearch.py | 20 +++--- platformcode/launcher.py | 29 ++++---- platformcode/platformtools.py | 36 +++++----- platformcode/recaptcha.py | 2 +- platformcode/shortcuts.py | 2 +- platformcode/side_menu.py | 8 +-- platformcode/subtitletools.py | 14 ++-- platformcode/xbmc_info_window.py | 4 +- platformcode/xbmc_videolibrary.py | 112 ++++++++++++++--------------- servers/akvideo.py | 8 +-- servers/anavids.py | 2 +- servers/animeid.py | 4 +- servers/anonfile.py | 4 +- servers/archiveorg.py | 4 +- servers/backin.py | 10 +-- servers/badshare.py | 4 +- servers/bdupload.py | 4 +- servers/cinemaupload.py | 4 +- servers/clicknupload.py | 6 +- servers/clipwatching.py | 4 +- servers/cloudvideo.py | 4 +- servers/crunchyroll.py | 6 +- servers/dailymotion.py | 6 +- servers/debriders/alldebrid.py | 2 +- servers/debriders/realdebrid.py | 4 +- servers/decrypters/adfly.py | 2 +- servers/decrypters/linkbucks.py | 8 +-- servers/decrypters/longurl.py | 6 +- servers/decrypters/safelinking.py | 4 +- servers/deltabit.py | 4 +- servers/directo.py | 2 +- servers/doodstream.py | 6 +- servers/dostream.py | 4 +- servers/downace.py | 4 +- servers/facebook.py | 6 +- servers/fastplay.py | 6 +- servers/fembed.py | 6 +- servers/fex.py | 4 +- servers/filefactory.py | 4 +- servers/filepup.py | 6 +- servers/filescdn.py | 4 +- servers/filesmonster.py | 2 +- servers/flashx.py | 12 ++-- servers/fourshared.py | 6 +- servers/gamovideo.py | 6 +- servers/gigasize.py | 4 +- servers/googlevideo.py | 6 +- servers/gounlimited.py | 6 +- servers/gvideo.py | 2 +- servers/hdload.py | 8 +-- servers/hdmario.py | 20 +++--- servers/hugefiles.py | 4 +- servers/idtbox.py | 4 +- servers/jawcloud.py | 4 +- servers/jetload.py | 4 +- servers/mailru.py | 6 +- servers/mediafire.py | 6 +- servers/mega.py | 2 +- servers/mixdrop.py | 4 +- servers/mp4upload.py | 6 +- servers/mydaddy.py | 2 +- servers/mystream.py | 4 +- servers/myupload.py | 2 +- servers/netutv.py | 4 +- servers/nowvideo.py | 16 ++--- servers/okru.py | 4 +- servers/onefichier.py | 18 ++--- servers/onlystream.py | 6 +- servers/rapidgator.py | 2 +- servers/rcdnme.py | 6 +- servers/rutube.py | 6 +- servers/samaup.py | 4 +- servers/sendvid.py | 4 +- servers/speedvideo.py | 10 +-- servers/streamtape.py | 4 +- servers/streamz.py | 4 +- servers/supervideo.py | 4 +- servers/thevid.py | 4 +- servers/thevideobee.py | 4 +- servers/turbobit.py | 2 +- servers/turbovid.py | 6 +- servers/tusfiles.py | 4 +- servers/uploadedto.py | 32 ++++----- servers/uppom.py | 4 +- servers/upstream.py | 2 +- servers/uptobox.py | 6 +- servers/upvid.py | 4 +- servers/uqload.py | 4 +- servers/userscloud.py | 6 +- servers/vevio.py | 4 +- servers/vidcloud.py | 6 +- servers/videobin.py | 4 +- servers/videomega.py | 2 +- servers/vidfast.py | 6 +- servers/vidlox.py | 4 +- servers/vidmoly.py | 4 +- servers/vidoza.py | 4 +- servers/vidtodo.py | 6 +- servers/vidtome.py | 4 +- servers/vidup.py | 4 +- servers/vimeo.py | 6 +- servers/vimpleru.py | 6 +- servers/vivo.py | 4 +- servers/vk.py | 6 +- servers/vshare.py | 4 +- servers/vudeo.py | 4 +- servers/vupplayer.py | 4 +- servers/vvvvid.py | 4 +- servers/watchanimestream.py | 2 +- servers/watchvideo.py | 6 +- servers/wstream.py | 8 +-- servers/youdbox.py | 2 +- servers/yourupload.py | 4 +- servers/youtube.py | 6 +- servers/youwatch.py | 6 +- servers/zippyshare.py | 4 +- service.py | 25 ++++--- specials/downloads.py | 2 +- specials/favorites.py | 12 ++-- specials/filmontv.py | 10 +-- specials/help.py | 2 +- specials/kodfavorites.py | 50 ++++++------- specials/news.py | 32 ++++----- specials/search.py | 18 ++--- specials/setting.py | 30 ++++---- specials/trailertools.py | 20 +++--- specials/tvmoviedb.py | 40 +++++------ specials/videolibrary.py | 47 ++++++------ updatetvshow.py | 8 +-- 157 files changed, 720 insertions(+), 726 deletions(-) diff --git a/channels/altadefinizione01_link.py b/channels/altadefinizione01_link.py index 17f2a409..e76ca1e0 100644 --- a/channels/altadefinizione01_link.py +++ b/channels/altadefinizione01_link.py @@ -73,7 +73,7 @@ def search(item, text): except: import sys for line in sys.exc_info(): - logger.info("%s mainlist search log: %s" % (__channel__, line)) + logger.error("%s" % line) return [] # =========== def per le novità nel menu principale ============= diff --git a/channels/cineblog01.py b/channels/cineblog01.py index 8ddea810..3137da60 100644 --- a/channels/cineblog01.py +++ b/channels/cineblog01.py @@ -76,7 +76,7 @@ def newest(categoria): def search(item, text): - logger.info(item, "search", text) + logger.info("search", text) if item.contentType == 'tvshow': item.url = host + '/serietv/' else: item.url = host try: @@ -189,14 +189,14 @@ def findvideos(item): def load_links(itemlist, re_txt, desc_txt, quality=""): streaming = scrapertools.find_single_match(data, re_txt).replace('"', '') - support.info('STREAMING', streaming) - support.info('STREAMING=', streaming) + logger.debug('STREAMING', streaming) + logger.debug('STREAMING=', streaming) matches = support.match(streaming, patron = r']+>([^<]+)<').matches for scrapedurl, scrapedtitle in matches: logger.debug("##### findvideos %s ## %s ## %s ##" % (desc_txt, scrapedurl, scrapedtitle)) itemlist.append(item.clone(action="play", title=scrapedtitle, url=scrapedurl, server=scrapedtitle, quality=quality)) - support.info() + logger.debug() itemlist = [] @@ -228,12 +228,12 @@ def findvideos(item): def findvid_serie(item): - support.info() + logger.debug() data = re.sub(r'((?:

|)?[^\d]*\d*(?:×|×)[0-9]+[^<]+)', '', item.other) return support.server(item, data=data) def play(item): - support.info() + logger.debug() return servertools.find_video_items(item, data=item.url) diff --git a/channels/film4k.py b/channels/film4k.py index cc605718..3e82f3b3 100644 --- a/channels/film4k.py +++ b/channels/film4k.py @@ -29,7 +29,7 @@ def mainlist(item): def search(item, text): - logger.info() + logger.info('search', text) item.url = item.url + "/?s=" + text try: return support.dooplay_search(item) diff --git a/channels/guardaseriecam.py b/channels/guardaseriecam.py index 1dc671de..cdab2da3 100644 --- a/channels/guardaseriecam.py +++ b/channels/guardaseriecam.py @@ -51,7 +51,7 @@ def episodios(item): def search(item, text): - support.info('search', item) + support.info('search', text) item.contentType = 'tvshow' itemlist = [] text = text.replace(' ', '+') @@ -66,5 +66,5 @@ def search(item, text): def findvideos(item): - logger.info("[guardaserie_live] findvideos") + logger.debug() return support.server(item, item.url) \ No newline at end of file diff --git a/channels/netfreex.py b/channels/netfreex.py index c40dd8cc..442145e7 100644 --- a/channels/netfreex.py +++ b/channels/netfreex.py @@ -29,7 +29,7 @@ def mainlist(item): def search(item, text): - logger.info() + logger.info('search', text) item.url = item.url + "/?s=" + text try: return support.dooplay_search(item) diff --git a/channels/vedohd.py b/channels/vedohd.py index 619135de..ffb81a21 100644 --- a/channels/vedohd.py +++ b/channels/vedohd.py @@ -30,7 +30,7 @@ def mainlist(item): def search(item, text): - logger.info("[vedohd.py] " + item.url + " search " + text) + logger.info("search",text) item.url = item.url + "/?s=" + text return support.dooplay_search(item, blacklist) @@ -44,7 +44,6 @@ def findvideos(item): itemlist = [] for link in support.dooplay_get_links(item, host): if link['title'] != 'Trailer': - logger.info(link['title']) server, quality = scrapertools.find_single_match(link['title'], '([^ ]+) ?(HD|3D)?') if quality: title = server + " [COLOR blue][" + quality + "][/COLOR]" @@ -63,7 +62,7 @@ def menu(item): def play(item): - logger.info("[vedohd.py] play") + logger.debug() data = support.swzz_get_url(item) diff --git a/channels/youtube_channel.py b/channels/youtube_channel.py index 66409814..80d57a85 100644 --- a/channels/youtube_channel.py +++ b/channels/youtube_channel.py @@ -14,15 +14,15 @@ YOUTUBE_V3_API_KEY = "AIzaSyCjsmBT0JZy1RT-PLwB-Zkfba87sa2inyI" def youtube_api_call(method, parameters): - logger.info("method=" + method + ", parameters=" + repr(parameters)) + logger.debug("method=" + method + ", parameters=" + repr(parameters)) encoded_parameters = urllib.urlencode(parameters) url = "https://www.googleapis.com/youtube/v3/" + method + "?" + encoded_parameters + "&key=" + YOUTUBE_V3_API_KEY; - logger.info("url=" + url) + logger.debug("url=" + url) data = httptools.downloadpage(url).data - logger.info("data=" + data) + logger.debug("data=" + data) json_object = jsontools.load(data) @@ -51,13 +51,13 @@ def youtube_get_playlist_items(playlist_id, pageToken=""): # Show all YouTube playlists for the selected channel def playlists(item, channel_id, pageToken=""): - logger.info() + logger.debug() itemlist = [] json_object = youtube_get_user_playlists(channel_id, pageToken) for entry in json_object["items"]: - logger.info("entry=" + repr(entry)) + logger.debug("entry=" + repr(entry)) title = entry["snippet"]["title"] plot = entry["snippet"]["description"] @@ -85,13 +85,13 @@ def latest_videos(item, channel_id): # Show all YouTube videos for the selected playlist def videos(item, pageToken=""): - logger.info() + logger.debug() itemlist = [] json_object = youtube_get_playlist_items(item.url, pageToken) for entry in json_object["items"]: - logger.info("entry=" + repr(entry)) + logger.debug("entry=" + repr(entry)) title = entry["snippet"]["title"] plot = entry["snippet"]["description"] diff --git a/channelselector.py b/channelselector.py index 9f810eff..649772c5 100644 --- a/channelselector.py +++ b/channelselector.py @@ -9,7 +9,7 @@ downloadenabled = addon.getSetting('downloadenabled') def getmainlist(view="thumb_"): - logger.info() + logger.debug() itemlist = list() if config.dev_mode(): @@ -62,14 +62,14 @@ def getmainlist(view="thumb_"): def getchanneltypes(view="thumb_"): - logger.info() + logger.debug() # Category List channel_types = ["movie", "tvshow", "anime", "documentary", "vos", "live", "torrent", "music"] #, "direct" # Channel Language channel_language = auto_filter() - logger.info("channel_language=%s" % channel_language) + logger.debug("channel_language=%s" % channel_language) # Build Itemlist itemlist = list() @@ -92,7 +92,7 @@ def getchanneltypes(view="thumb_"): def filterchannels(category, view="thumb_"): from core import channeltools - logger.info('Filter Channels ' + category) + logger.debug('Filter Channels ' + category) channelslist = [] @@ -103,14 +103,14 @@ def filterchannels(category, view="thumb_"): appenddisabledchannels = True channel_path = os.path.join(config.get_runtime_path(), 'channels', '*.json') - logger.info("channel_path = %s" % channel_path) + logger.debug("channel_path = %s" % channel_path) channel_files = glob.glob(channel_path) - logger.info("channel_files found %s" % (len(channel_files))) + logger.debug("channel_files found %s" % (len(channel_files))) # Channel Language channel_language = auto_filter() - logger.info("channel_language=%s" % channel_language) + logger.debug("channel_language=%s" % channel_language) for channel_path in channel_files: logger.debug("channel in for = %s" % channel_path) @@ -221,7 +221,7 @@ def get_thumb(thumb_name, view="thumb_"): def set_channel_info(parameters): - logger.info() + logger.debug() info = '' language = '' diff --git a/core/autoplay.py b/core/autoplay.py index f6707ffb..a8d141d6 100644 --- a/core/autoplay.py +++ b/core/autoplay.py @@ -29,7 +29,7 @@ def start(itemlist, item): if item.global_search: return itemlist - logger.info() + logger.debug() global PLAYED PLAYED = False @@ -274,7 +274,7 @@ def start(itemlist, item): def play_multi_channel(item, itemlist): - logger.info() + logger.debug() start(itemlist, item) diff --git a/core/channeltools.py b/core/channeltools.py index 90a2dd74..ff5a0442 100644 --- a/core/channeltools.py +++ b/core/channeltools.py @@ -15,7 +15,7 @@ default_file = dict() remote_path = 'https://raw.githubusercontent.com/kodiondemand/media/master/' def is_enabled(channel_name): - logger.info("channel_name=" + channel_name) + logger.debug("channel_name=" + channel_name) return get_channel_parameters(channel_name)["active"] and get_channel_setting("enabled", channel=channel_name, default=True) @@ -27,7 +27,7 @@ def get_channel_parameters(channel_name): if channel_name not in dict_channels_parameters: try: channel_parameters = get_channel_json(channel_name) - # logger.debug(channel_parameters) + logger.debug(channel_parameters) if channel_parameters: # name and default changes channel_parameters["title"] = channel_parameters.pop("name") + (' [DEPRECATED]' if 'deprecated' in channel_parameters and channel_parameters['deprecated'] else '') @@ -87,7 +87,7 @@ def get_channel_parameters(channel_name): def get_channel_json(channel_name): - # logger.info("channel_name=" + channel_name) + logger.debug("channel_name=" + channel_name) from core import filetools channel_json = None try: @@ -101,9 +101,9 @@ def get_channel_json(channel_name): channel_name + ".json") if filetools.isfile(channel_path): - # logger.info("channel_data=" + channel_path) + logger.debug("channel_data=" + channel_path) channel_json = jsontools.load(filetools.read(channel_path)) - # logger.info("channel_json= %s" % channel_json) + logger.debug("channel_json= %s" % channel_json) except Exception as ex: template = "An exception of type %s occured. Arguments:\n%r" @@ -114,7 +114,7 @@ def get_channel_json(channel_name): def get_channel_controls_settings(channel_name): - # logger.info("channel_name=" + channel_name) + logger.debug("channel_name=" + channel_name) dict_settings = {} # import web_pdb; web_pdb.set_trace() # list_controls = get_channel_json(channel_name).get('settings', list()) @@ -137,7 +137,7 @@ def get_lang(channel_name): if hasattr(channel, 'list_language'): for language in channel.list_language: list_language.append(language) - logger.info(list_language) + logger.debug(list_language) else: sub = False langs = [] diff --git a/core/downloader.py b/core/downloader.py index b714b474..919f9469 100644 --- a/core/downloader.py +++ b/core/downloader.py @@ -253,7 +253,7 @@ class Downloader(object): self.file.seek(2 ** 31, 0) except OverflowError: self._seekable = False - logger.info("Cannot do seek() or tell() in files larger than 2GB") + logger.error("Cannot do seek() or tell() in files larger than 2GB") self.__get_download_info__() diff --git a/core/filetools.py b/core/filetools.py index 554a2671..481f015c 100644 --- a/core/filetools.py +++ b/core/filetools.py @@ -814,7 +814,7 @@ def remove_tags(title): @rtype: str @return: string without tags """ - logger.info() + logger.debug() title_without_tags = scrapertools.find_single_match(title, r'\[color .+?\](.+)\[\/color\]') @@ -832,7 +832,7 @@ def remove_smb_credential(path): @return: chain without credentials @rtype: str """ - logger.info() + logger.debug() if not scrapertools.find_single_match(path, r'(^\w+:\/\/)'): return path diff --git a/core/filtertools.py b/core/filtertools.py index c03ffe28..09f016db 100644 --- a/core/filtertools.py +++ b/core/filtertools.py @@ -234,7 +234,7 @@ def get_link(list_item, item, list_language, list_quality=None, global_filter_la @return: Item list @rtype: list[Item] """ - logger.info() + logger.debug() # if the required fields are None we leave if list_item is None or item is None: @@ -274,7 +274,7 @@ def get_links(list_item, item, list_language, list_quality=None, global_filter_l @return: lista de Item @rtype: list[Item] """ - logger.info() + logger.debug() # if the required fields are None we leave @@ -362,7 +362,7 @@ def no_filter(item): @return: lista de enlaces @rtype: list[Item] """ - logger.info() + logger.debug() itemlist = [] for i in item.list_item_all: @@ -384,7 +384,7 @@ def mainlist(channel, list_language, list_quality): @return: Item list @rtype: list[Item] """ - logger.info() + logger.debug() itemlist = [] dict_series = jsontools.get_node_from_file(channel, TAG_TVSHOW_FILTER) @@ -425,8 +425,8 @@ def config_item(item): @param item: item @type item: Item """ - logger.info() - logger.info("item %s" % item.tostring()) + logger.debug() + logger.debug("item %s" % item.tostring()) # WE GET THE JSON DATA dict_series = jsontools.get_node_from_file(item.from_channel, TAG_TVSHOW_FILTER) @@ -448,8 +448,8 @@ def config_item(item): else: lang_selected = dict_series.get(tvshow, {}).get(TAG_LANGUAGE, default_lang) list_quality = dict_series.get(tvshow, {}).get(TAG_QUALITY_ALLOWED, [x.lower() for x in item.list_quality]) - # logger.info("lang selected {}".format(lang_selected)) - # logger.info("list quality {}".format(list_quality)) + # logger.debug("lang selected {}".format(lang_selected)) + # logger.debug("list quality {}".format(list_quality)) active = True custom_button = {'visible': False} @@ -516,7 +516,7 @@ def config_item(item): def delete(item, dict_values): - logger.info() + logger.debug() if item: dict_series = jsontools.get_node_from_file(item.from_channel, TAG_TVSHOW_FILTER) @@ -554,7 +554,7 @@ def save(item, dict_data_saved): @param dict_data_saved: dictionary with saved data @type dict_data_saved: dict """ - logger.info() + logger.debug() if item and dict_data_saved: logger.debug('item: %s\ndatos: %s' % (item.tostring(), dict_data_saved)) @@ -564,7 +564,7 @@ def save(item, dict_data_saved): dict_series = jsontools.get_node_from_file(item.from_channel, TAG_TVSHOW_FILTER) tvshow = item.show.strip().lower() - logger.info("Data is updated") + logger.debug("Data is updated") list_quality = [] for _id, value in list(dict_data_saved.items()): @@ -599,7 +599,7 @@ def save_from_context(item): @param item: item @type item: item """ - logger.info() + logger.debug() dict_series = jsontools.get_node_from_file(item.from_channel, TAG_TVSHOW_FILTER) tvshow = item.show.strip().lower() @@ -630,7 +630,7 @@ def delete_from_context(item): @param item: item @type item: item """ - logger.info() + logger.debug() # We come from get_links and no result has been obtained, in context menu and we delete if item.to_channel != "": diff --git a/core/httptools.py b/core/httptools.py index a06eb9ac..ae7de722 100755 --- a/core/httptools.py +++ b/core/httptools.py @@ -449,7 +449,7 @@ def downloadpage(url, **opt): if not 'api.themoviedb' in url and not opt.get('alfa_s', False): show_infobox(info_dict) - + if not config.get_setting("debug"): logger.info('Page URL:',url) return type('HTTPResponse', (), response) def fill_fields_pre(url, opt, proxy_data, file_name): diff --git a/core/jsontools.py b/core/jsontools.py index b2fed3a6..6bfe960f 100644 --- a/core/jsontools.py +++ b/core/jsontools.py @@ -11,22 +11,22 @@ from inspect import stack try: import json except: - logger.info("json included in the interpreter **NOT** available") + logger.error("json included in the interpreter **NOT** available") try: import simplejson as json except: - logger.info("simplejson included in the interpreter **NOT** available") + logger.error("simplejson included in the interpreter **NOT** available") try: from lib import simplejson as json except: - logger.info("simplejson in lib directory **NOT** available") + logger.error("simplejson in lib directory **NOT** available") logger.error("A valid JSON parser was not found") json = None else: logger.info("Using simplejson in the lib directory") else: - logger.info("Using simplejson included in the interpreter") + logger.error("Using simplejson included in the interpreter") # ~ else: # ~ logger.info("Usando json incluido en el interprete") diff --git a/core/scraper.py b/core/scraper.py index 3fd90d5e..d5aa9ac2 100644 --- a/core/scraper.py +++ b/core/scraper.py @@ -62,7 +62,7 @@ def find_and_set_infoLabels(item): # Check if there is a 'code' if scraper_result and item.infoLabels['code']: # correct code - logger.info("Identificador encontrado: %s" % item.infoLabels['code']) + logger.debug("Identificador encontrado: %s" % item.infoLabels['code']) scraper.completar_codigos(item) return True elif scraper_result: @@ -72,7 +72,7 @@ def find_and_set_infoLabels(item): # Content not found msg = config.get_localized_string(60228) % title - logger.info(msg) + logger.debug(msg) # Show box with other options: item = platformtools.dialog_info(item, scraper_actual) if item.exit: @@ -83,7 +83,7 @@ def find_and_set_infoLabels(item): def cuadro_completar(item): - logger.info() + logger.debug() global dict_default dict_default = {} @@ -196,7 +196,7 @@ def get_nfo(item): @rtype: str @return: """ - logger.info() + logger.debug() if "infoLabels" in item and "noscrap_id" in item.infoLabels: # Create the xml file with the data obtained from the item since there is no active scraper info_nfo = '' diff --git a/core/scrapertools.py b/core/scrapertools.py index 3df0dab8..aa926c8b 100644 --- a/core/scrapertools.py +++ b/core/scrapertools.py @@ -34,7 +34,7 @@ from platformcode import logger def printMatches(matches): i = 0 for match in matches: - logger.info("%d %s" % (i, match)) + logger.debug("%d %s" % (i, match)) i = i + 1 @@ -447,7 +447,7 @@ def get_season_and_episode(title): except: pass - logger.info("'" + title + "' -> '" + filename + "'") + logger.debug("'" + title + "' -> '" + filename + "'") return filename diff --git a/core/servertools.py b/core/servertools.py index 19e688f9..8017f6c7 100644 --- a/core/servertools.py +++ b/core/servertools.py @@ -47,7 +47,7 @@ def find_video_items(item=None, data=None): @return: returns the itemlist with the results @rtype: list """ - logger.info() + logger.debug() itemlist = [] # Download the page @@ -97,7 +97,7 @@ def get_servers_itemlist(itemlist, fnc=None, sort=False): # Walk the patterns for pattern in server_parameters.get("find_videos", {}).get("patterns", []): - logger.info(pattern["pattern"]) + logger.debug(pattern["pattern"]) # Scroll through the results for match in re.compile(pattern["pattern"], re.DOTALL).finditer( "\n".join([item.url.split('|')[0] for item in itemlist if not item.server])): @@ -144,7 +144,7 @@ def findvideos(data, skip=False): return some link. It can also be an integer greater than 1, which would represent the maximum number of links to search. :return: """ - logger.info() + logger.debug() devuelve = [] skip = int(skip) servers_list = list(get_servers_list().keys()) @@ -181,7 +181,7 @@ def findvideosbyserver(data, serverid): value = translate_server_name(server_parameters["name"]) , url, serverid, server_parameters.get("thumbnail", "") if value not in devuelve and url not in server_parameters["find_videos"].get("ignore_urls", []): devuelve.append(value) - logger.info(msg) + logger.debug(msg) return devuelve @@ -193,7 +193,7 @@ def guess_server_thumbnail(serverid): def get_server_from_url(url): - logger.info() + logger.debug() servers_list = list(get_servers_list().keys()) # Run findvideos on each active server @@ -211,7 +211,7 @@ def get_server_from_url(url): for n, pattern in enumerate(server_parameters["find_videos"].get("patterns", [])): msg = "%s\npattern: %s" % (serverid, pattern["pattern"]) if not "pattern_compiled" in pattern: - # logger.info('compiled ' + serverid) + # logger.debug('compiled ' + serverid) pattern["pattern_compiled"] = re.compile(pattern["pattern"]) dict_servers_parameters[serverid]["find_videos"]["patterns"][n]["pattern_compiled"] = pattern["pattern_compiled"] # Scroll through the results @@ -224,7 +224,7 @@ def get_server_from_url(url): msg += "\nurl encontrada: %s" % url value = translate_server_name(server_parameters["name"]), url, serverid, server_parameters.get("thumbnail", "") if url not in server_parameters["find_videos"].get("ignore_urls", []): - logger.info(msg) + logger.debug(msg) return value return None @@ -616,7 +616,7 @@ def get_server_setting(name, server, default=None): dict_file['settings'] = dict_settings # We create the file ../settings/channel_data.json if not filetools.write(file_settings, jsontools.dump(dict_file)): - logger.info("ERROR saving file: %s" % file_settings) + logger.error("ERROR saving file: %s" % file_settings) # We return the value of the local parameter 'name' if it exists, if default is not returned return dict_settings.get(name, default) @@ -638,7 +638,7 @@ def set_server_setting(name, value, server): dict_file = jsontools.load(filetools.read(file_settings)) dict_settings = dict_file.get('settings', {}) except EnvironmentError: - logger.info("ERROR when reading the file: %s" % file_settings) + logger.error("ERROR when reading the file: %s" % file_settings) dict_settings[name] = value @@ -650,7 +650,7 @@ def set_server_setting(name, value, server): # We create the file ../settings/channel_data.json if not filetools.write(file_settings, jsontools.dump(dict_file)): - logger.info("ERROR saving file: %s" % file_settings) + logger.error("ERROR saving file: %s" % file_settings) return None return value @@ -752,7 +752,7 @@ def check_video_link(item, timeout=3): server_module = __import__('servers.%s' % server, None, None, ["servers.%s" % server]) except: server_module = None - logger.info("[check_video_link] Cannot import server! %s" % server) + logger.error("[check_video_link] Cannot import server! %s" % server) return item, NK if hasattr(server_module, 'test_video_exists'): @@ -762,20 +762,20 @@ def check_video_link(item, timeout=3): try: video_exists, message = server_module.test_video_exists(page_url=url) if not video_exists: - logger.info("[check_video_link] Does not exist! %s %s %s" % (message, server, url)) + logger.error("[check_video_link] Does not exist! %s %s %s" % (message, server, url)) resultado = KO else: - logger.info("[check_video_link] check ok %s %s" % (server, url)) + logger.debug("[check_video_link] check ok %s %s" % (server, url)) resultado = OK except: - logger.info("[check_video_link] Can't check now! %s %s" % (server, url)) + logger.error("[check_video_link] Can't check now! %s %s" % (server, url)) resultado = NK finally: httptools.HTTPTOOLS_DEFAULT_DOWNLOAD_TIMEOUT = ant_timeout # Restore download time return item, resultado - logger.info("[check_video_link] There is no test_video_exists for server: %s" % server) + logger.debug("[check_video_link] There is no test_video_exists for server: %s" % server) return item, NK def translate_server_name(name): diff --git a/core/support.py b/core/support.py index 693c6319..99c83171 100755 --- a/core/support.py +++ b/core/support.py @@ -33,7 +33,7 @@ def hdpass_get_servers(item): for mir_url, srv in scrapertools.find_multiple_matches(mir, patron_option): mir_url = scrapertools.decodeHtmlentities(mir_url) - info(mir_url) + logger.debug(mir_url) it = item.clone(action="play", quality=quality, title=srv, server=srv, url= mir_url) if not servertools.get_server_parameters(srv.lower()): it = hdpass_get_url(it)[0] # do not exists or it's empty ret.append(it) @@ -1022,7 +1022,7 @@ def videolibrary(itemlist, item, typography='', function_level=1, function=''): # Simply add this function to add video library support # Function_level is useful if the function is called by another function. # If the call is direct, leave it blank - info() + logger.debug() if item.contentType == 'movie': action = 'add_pelicula_to_library' @@ -1073,7 +1073,7 @@ def videolibrary(itemlist, item, typography='', function_level=1, function=''): def nextPage(itemlist, item, data='', patron='', function_or_level=1, next_page='', resub=[]): # Function_level is useful if the function is called by another function. # If the call is direct, leave it blank - info() + logger.debug() action = inspect.stack()[function_or_level][3] if type(function_or_level) == int else function_or_level if next_page == '': next_page = scrapertools.find_single_match(data, patron) @@ -1083,7 +1083,7 @@ def nextPage(itemlist, item, data='', patron='', function_or_level=1, next_page= if 'http' not in next_page: next_page = scrapertools.find_single_match(item.url, 'https?://[a-z0-9.-]+') + (next_page if next_page.startswith('/') else '/' + next_page) next_page = next_page.replace('&', '&') - info('NEXT= ', next_page) + logger.debug('NEXT= ', next_page) itemlist.append( item.clone(channel=item.channel, action = action, @@ -1110,7 +1110,7 @@ def pagination(itemlist, item, page, perpage, function_level=1): def server(item, data='', itemlist=[], headers='', AutoPlay=True, CheckLinks=True, Download=True, patronTag=None, Videolibrary=True): - info() + logger.debug() blacklisted_servers = config.get_setting("black_list", server='servers') if not blacklisted_servers: blacklisted_servers = [] if not data and not itemlist: @@ -1375,7 +1375,7 @@ def thumb(item_itemlist_string=None, genre=False, live=False): '_tvshow':['serie','tv', 'fiction']} def autoselect_thumb(item, genre): - info('SPLIT',re.split(r'\.|\{|\}|\[|\]|\(|\)|/| ',item.title.lower())) + logger.debug('SPLIT',re.split(r'\.|\{|\}|\[|\]|\(|\)|/| ',item.title.lower())) if genre == False: for thumb, titles in icon_dict.items(): if any(word in re.split(r'\.|\{|\}|\[|\]|\(|\)|/| ',item.title.lower()) for word in search): diff --git a/core/tmdb.py b/core/tmdb.py index e5a6ab5e..0b6c709e 100644 --- a/core/tmdb.py +++ b/core/tmdb.py @@ -87,7 +87,7 @@ create_bd() # The function name is the name of the decorator and receives the function that decorates. def cache_response(fn): - logger.info() + logger.debug() # import time # start_time = time.time() @@ -495,7 +495,7 @@ def set_infoLabels_item(item, seekTmdb=True, idioma_busqueda=def_lang, lock=None def find_and_set_infoLabels(item): - logger.info() + logger.debug() global otmdb_global tmdb_result = None @@ -902,7 +902,7 @@ class Tmdb(object): cls.dic_generos[idioma][tipo] = {} url = ('http://api.themoviedb.org/3/genre/%s/list?api_key=a1ab8b8669da03637a4b98fa39c39228&language=%s' % (tipo, idioma)) try: - logger.info("[Tmdb.py] Filling in dictionary of genres") + logger.debug("[Tmdb.py] Filling in dictionary of genres") resultado = cls.get_json(url) if not isinstance(resultado, dict): @@ -934,7 +934,7 @@ class Tmdb(object): '&language=%s' % (self.busqueda_id, source, self.busqueda_idioma)) buscando = "%s: %s" % (source.capitalize(), self.busqueda_id) - logger.info("[Tmdb.py] Searching %s:\n%s" % (buscando, url)) + logger.debug("[Tmdb.py] Searching %s:\n%s" % (buscando, url)) resultado = self.get_json(url) if not isinstance(resultado, dict): resultado = ast.literal_eval(resultado.decode('utf-8')) @@ -981,7 +981,7 @@ class Tmdb(object): url += '&year=%s' % self.busqueda_year buscando = self.busqueda_texto.capitalize() - logger.info("[Tmdb.py] Searching %s on page %s:\n%s" % (buscando, page, url)) + logger.debug("[Tmdb.py] Searching %s on page %s:\n%s" % (buscando, page, url)) resultado = self.get_json(url) if not isinstance(resultado, dict): resultado = ast.literal_eval(resultado.decode('utf-8')) @@ -1042,7 +1042,7 @@ class Tmdb(object): url = ('http://api.themoviedb.org/3/%s?api_key=a1ab8b8669da03637a4b98fa39c39228&%s' % (type_search, "&".join(params))) - logger.info("[Tmdb.py] Searcing %s:\n%s" % (type_search, url)) + logger.debug("[Tmdb.py] Searcing %s:\n%s" % (type_search, url)) resultado = self.get_json(url, cache=False) if not isinstance(resultado, dict): resultado = ast.literal_eval(resultado.decode('utf-8')) @@ -1107,7 +1107,7 @@ class Tmdb(object): return True def get_list_resultados(self, num_result=20): - # logger.info("self %s" % str(self)) + # logger.debug("self %s" % str(self)) res = [] if num_result <= 0: @@ -1327,7 +1327,7 @@ class Tmdb(object): "&append_to_response=credits" % (self.result["id"], numtemporada, self.busqueda_idioma) buscando = "id_Tmdb: " + str(self.result["id"]) + " season: " + str(numtemporada) + "\nURL: " + url - logger.info("[Tmdb.py] Searcing " + buscando) + logger.debug("[Tmdb.py] Searcing " + buscando) try: self.temporada[numtemporada] = self.get_json(url) if not isinstance(self.temporada[numtemporada], dict): @@ -1516,7 +1516,7 @@ class Tmdb(object): items.extend(list(self.get_episodio(ret_infoLabels['season'], episodio).items())) - # logger.info("ret_infoLabels" % ret_infoLabels) + # logger.debug("ret_infoLabels" % ret_infoLabels) for k, v in items: if not v: diff --git a/core/trakt_tools.py b/core/trakt_tools.py index e5d23e51..639c8803 100644 --- a/core/trakt_tools.py +++ b/core/trakt_tools.py @@ -128,7 +128,7 @@ def token_trakt(item): def set_trakt_info(item): - logger.info() + logger.debug() import xbmcgui # Envia los datos a trakt try: @@ -139,7 +139,7 @@ def set_trakt_info(item): pass def get_trakt_watched(id_type, mediatype, update=False): - logger.info() + logger.debug() id_list = [] id_dict = dict() @@ -229,7 +229,7 @@ def trakt_check(itemlist): def get_sync_from_file(): - logger.info() + logger.debug() sync_path = os.path.join(config.get_data_path(), 'settings_channels', 'trakt_data.json') trakt_node = {} if os.path.exists(sync_path): @@ -241,7 +241,7 @@ def get_sync_from_file(): def update_trakt_data(mediatype, trakt_data): - logger.info() + logger.debug() sync_path = os.path.join(config.get_data_path(), 'settings_channels', 'trakt_data.json') if os.path.exists(sync_path): @@ -251,7 +251,7 @@ def update_trakt_data(mediatype, trakt_data): def ask_install_script(): - logger.info() + logger.debug() from platformcode import platformtools @@ -265,7 +265,7 @@ def ask_install_script(): def wait_for_update_trakt(): - logger.info() + logger.debug() t = Thread(update_all) t.setDaemon(True) t.start() @@ -274,7 +274,7 @@ def wait_for_update_trakt(): def update_all(): # from core.support import dbg;dbg() from time import sleep - logger.info() + logger.debug() sleep(20) while xbmc.Player().isPlaying(): sleep(20) diff --git a/core/tvdb.py b/core/tvdb.py index 5e5f6f6b..eb635d36 100644 --- a/core/tvdb.py +++ b/core/tvdb.py @@ -77,9 +77,9 @@ otvdb_global = None def find_and_set_infoLabels(item): - logger.info() + logger.debug() # from core.support import dbg;dbg() - # logger.info("item es %s" % item) + # logger.debug("item es %s" % item) p_dialog = None if not item.contentSeason: @@ -382,7 +382,7 @@ class Tvdb(object): @classmethod def __check_token(cls): - # logger.info() + # logger.debug() if TOKEN == "": cls.__login() else: @@ -397,7 +397,7 @@ class Tvdb(object): @staticmethod def __login(): - # logger.info() + # logger.debug() global TOKEN apikey = "106B699FDC04301C" @@ -423,7 +423,7 @@ class Tvdb(object): @classmethod def __refresh_token(cls): - # logger.info() + # logger.debug() global TOKEN is_success = False @@ -521,7 +521,7 @@ class Tvdb(object): ] } """ - logger.info() + logger.debug() if id_episode and self.episodes.get(id_episode): return self.episodes.get(id_episode) @@ -589,7 +589,7 @@ class Tvdb(object): } } """ - logger.info() + logger.debug() url = HOST + "/series/%s/episodes?page=%s" % (_id, page) @@ -662,7 +662,7 @@ class Tvdb(object): """ if semaforo: semaforo.acquire() - logger.info() + logger.debug() url = HOST + "/episodes/%s" % _id @@ -681,7 +681,7 @@ class Tvdb(object): else: dict_html = req.json() - # logger.info("dict_html %s" % dict_html) + # logger.debug("dict_html %s" % dict_html) self.episodes[_id] = dict_html.pop("data") if 'Error' not in dict_html else {} if semaforo: @@ -712,7 +712,7 @@ class Tvdb(object): "status": "string" } """ - logger.info() + logger.debug() params = {} if name: @@ -802,7 +802,7 @@ class Tvdb(object): } } """ - logger.info() + logger.debug() resultado = {} url = HOST + "/series/%s" % _id @@ -855,7 +855,7 @@ class Tvdb(object): @rtype: dict """ - logger.info() + logger.debug() if self.result.get('image_season_%s' % season): return self.result['image_season_%s' % season] @@ -909,7 +909,7 @@ class Tvdb(object): @return: dictionary with actors @rtype: dict """ - logger.info() + logger.debug() url = HOST + "/series/%s/actors" % _id DEFAULT_HEADERS["Accept-Language"] = lang @@ -942,7 +942,7 @@ class Tvdb(object): @rtype: list @return: list of results """ - logger.info() + logger.debug() list_results = [] # if we have a result and it has seriesName, we already have the info of the series, it is not necessary to search again diff --git a/core/videolibrarytools.py b/core/videolibrarytools.py index 36ade9ed..a11e75ca 100644 --- a/core/videolibrarytools.py +++ b/core/videolibrarytools.py @@ -78,7 +78,7 @@ def save_movie(item, silent=False): @rtype fallidos: int @return: the number of failed items or -1 if all failed """ - logger.info() + logger.debug() # logger.debug(item.tostring('\n')) insertados = 0 sobreescritos = 0 @@ -144,7 +144,7 @@ def save_movie(item, silent=False): if not path: # Create folder path = filetools.join(MOVIES_PATH, ("%s [%s]" % (base_name, _id)).strip()) - logger.info("Creating movie directory:" + path) + logger.debug("Creating movie directory:" + path) if not filetools.mkdir(path): logger.debug("Could not create directory") return 0, 0, -1, path @@ -159,7 +159,7 @@ def save_movie(item, silent=False): if not nfo_exists: # We create .nfo if it doesn't exist - logger.info("Creating .nfo: " + nfo_path) + logger.debug("Creating .nfo: " + nfo_path) head_nfo = scraper.get_nfo(item) item_nfo = Item(title=item.contentTitle, channel="videolibrary", action='findvideos', @@ -182,7 +182,7 @@ def save_movie(item, silent=False): if item_nfo and strm_exists: if json_exists: - logger.info("The file exists. Is overwritten") + logger.debug("The file exists. Is overwritten") sobreescritos += 1 else: insertados += 1 @@ -209,7 +209,7 @@ def save_movie(item, silent=False): item_nfo.library_urls[item.channel] = item.url if filetools.write(nfo_path, head_nfo + item_nfo.tojson()): - #logger.info("FOLDER_MOVIES : %s" % FOLDER_MOVIES) + #logger.debug("FOLDER_MOVIES : %s" % FOLDER_MOVIES) # We update the Kodi video library with the movie if config.is_xbmc() and config.get_setting("videolibrary_kodi") and not silent: from platformcode import xbmc_videolibrary @@ -238,7 +238,7 @@ def update_renumber_options(item, head_nfo, path): json = json_file['TVSHOW_AUTORENUMBER'] if item.fulltitle in json: item.channel_prefs[channel]['TVSHOW_AUTORENUMBER'] = json[item.fulltitle] - logger.info('UPDATED=\n' + str(item.channel_prefs)) + logger.debug('UPDATED=\n' + str(item.channel_prefs)) filetools.write(tvshow_path, head_nfo + item.tojson()) def add_renumber_options(item, head_nfo, path): @@ -426,7 +426,7 @@ def save_tvshow(item, episodelist, silent=False): @rtype path: str @return: serial directory """ - logger.info() + logger.debug() # logger.debug(item.tostring('\n')) path = "" @@ -486,7 +486,7 @@ def save_tvshow(item, episodelist, silent=False): if not path: path = filetools.join(TVSHOWS_PATH, ("%s [%s]" % (base_name, _id)).strip()) - logger.info("Creating series directory: " + path) + logger.debug("Creating series directory: " + path) try: filetools.mkdir(path) except OSError as exception: @@ -496,7 +496,7 @@ def save_tvshow(item, episodelist, silent=False): tvshow_path = filetools.join(path, "tvshow.nfo") if not filetools.exists(tvshow_path): # We create tvshow.nfo, if it does not exist, with the head_nfo, series info and watched episode marks - logger.info("Creating tvshow.nfo: " + tvshow_path) + logger.debug("Creating tvshow.nfo: " + tvshow_path) head_nfo = scraper.get_nfo(item) item.infoLabels['mediatype'] = "tvshow" item.infoLabels['title'] = item.contentSerieName @@ -570,11 +570,11 @@ def save_episodes(path, episodelist, serie, silent=False, overwrite=True): @rtype fallidos: int @return: the number of failed episodes """ - logger.info() + logger.debug() episodelist = filter_list(episodelist, serie.action, path) # No episode list, nothing to save if not len(episodelist): - logger.info("There is no episode list, we go out without creating strm") + logger.debug("There is no episode list, we go out without creating strm") return 0, 0, 0 # process local episodes @@ -589,7 +589,7 @@ def save_episodes(path, episodelist, serie, silent=False, overwrite=True): elif config.get_setting("local_episodes", "videolibrary"): done, local_episodes_path = config_local_episodes_path(path, serie) if done < 0: - logger.info("An issue has occurred while configuring local episodes, going out without creating strm") + logger.debug("An issue has occurred while configuring local episodes, going out without creating strm") return 0, 0, done item_nfo.local_episodes_path = local_episodes_path filetools.write(nfo_path, head_nfo + item_nfo.tojson()) @@ -713,7 +713,7 @@ def save_episodes(path, episodelist, serie, silent=False, overwrite=True): # No episode list, nothing to save if not len(new_episodelist): - logger.info("There is no episode list, we go out without creating strm") + logger.debug("There is no episode list, we go out without creating strm") return 0, 0, 0 local_episodelist += get_local_content(path) @@ -745,12 +745,12 @@ def save_episodes(path, episodelist, serie, silent=False, overwrite=True): json_path = filetools.join(path, ("%s [%s].json" % (season_episode, e.channel)).lower()) if season_episode in local_episodelist: - logger.info('Skipped: Serie ' + serie.contentSerieName + ' ' + season_episode + ' available as local content') + logger.debug('Skipped: Serie ' + serie.contentSerieName + ' ' + season_episode + ' available as local content') continue # check if the episode has been downloaded if filetools.join(path, "%s [downloads].json" % season_episode) in ficheros: - logger.info('INFO: "%s" episode %s has been downloaded, skipping it' % (serie.contentSerieName, season_episode)) + logger.debug('INFO: "%s" episode %s has been downloaded, skipping it' % (serie.contentSerieName, season_episode)) continue strm_exists = strm_path in ficheros @@ -806,7 +806,7 @@ def save_episodes(path, episodelist, serie, silent=False, overwrite=True): if filetools.write(json_path, e.tojson()): if not json_exists: - logger.info("Inserted: %s" % json_path) + logger.debug("Inserted: %s" % json_path) insertados += 1 # We mark episode as unseen news_in_playcounts[season_episode] = 0 @@ -817,14 +817,14 @@ def save_episodes(path, episodelist, serie, silent=False, overwrite=True): news_in_playcounts[serie.contentSerieName] = 0 else: - logger.info("Overwritten: %s" % json_path) + logger.debug("Overwritten: %s" % json_path) sobreescritos += 1 else: - logger.info("Failed: %s" % json_path) + logger.debug("Failed: %s" % json_path) fallidos += 1 else: - logger.info("Failed: %s" % json_path) + logger.debug("Failed: %s" % json_path) fallidos += 1 if not silent and p_dialog.iscanceled(): @@ -894,7 +894,7 @@ def save_episodes(path, episodelist, serie, silent=False, overwrite=True): def config_local_episodes_path(path, item, silent=False): - logger.info(item) + logger.debug(item) from platformcode.xbmc_videolibrary import search_local_path local_episodes_path=search_local_path(item) if not local_episodes_path: @@ -906,11 +906,11 @@ def config_local_episodes_path(path, item, silent=False): platformtools.dialog_ok(config.get_localized_string(30131), config.get_localized_string(80043)) local_episodes_path = platformtools.dialog_browse(0, config.get_localized_string(80046)) if local_episodes_path == '': - logger.info("User has canceled the dialog") + logger.debug("User has canceled the dialog") return -2, local_episodes_path elif path in local_episodes_path: platformtools.dialog_ok(config.get_localized_string(30131), config.get_localized_string(80045)) - logger.info("Selected folder is the same of the TV show one") + logger.debug("Selected folder is the same of the TV show one") return -2, local_episodes_path if local_episodes_path: @@ -925,7 +925,7 @@ def config_local_episodes_path(path, item, silent=False): def process_local_episodes(local_episodes_path, path): - logger.info() + logger.debug() sub_extensions = ['.srt', '.sub', '.sbv', '.ass', '.idx', '.ssa', '.smi'] artwork_extensions = ['.jpg', '.jpeg', '.png'] @@ -964,7 +964,7 @@ def process_local_episodes(local_episodes_path, path): def get_local_content(path): - logger.info() + logger.debug() local_episodelist = [] for root, folders, files in filetools.walk(path): @@ -993,7 +993,7 @@ def add_movie(item): @type item: item @param item: item to be saved. """ - logger.info() + logger.debug() from platformcode.launcher import set_search_temp; set_search_temp(item) # To disambiguate titles, TMDB is caused to ask for the really desired title @@ -1040,7 +1040,7 @@ def add_tvshow(item, channel=None): @param channel: channel from which the series will be saved. By default, item.from_channel or item.channel will be imported. """ - logger.info("show=#" + item.show + "#") + logger.debug("show=#" + item.show + "#") from platformcode.launcher import set_search_temp; set_search_temp(item) if item.channel == "downloads": @@ -1117,7 +1117,7 @@ def add_tvshow(item, channel=None): else: platformtools.dialog_ok(config.get_localized_string(30131), config.get_localized_string(60070) % item.show) - logger.info("%s episodes of series %s have been added to the video library" % (insertados, item.show)) + logger.debug("%s episodes of series %s have been added to the video library" % (insertados, item.show)) if config.is_xbmc(): if config.get_setting("sync_trakt_new_tvshow", "videolibrary"): import xbmc @@ -1133,7 +1133,7 @@ def add_tvshow(item, channel=None): def emergency_urls(item, channel=None, path=None, headers={}): - logger.info() + logger.debug() import re from servers import torrent try: diff --git a/core/ziptools.py b/core/ziptools.py index 138c4c1b..f7f73db6 100644 --- a/core/ziptools.py +++ b/core/ziptools.py @@ -17,8 +17,8 @@ from core import filetools class ziptools(object): def extract(self, file, dir, folder_to_extract="", overwrite_question=False, backup=False): - logger.info("file= %s" % file) - logger.info("dir= %s" % dir) + logger.debug("file= %s" % file) + logger.debug("dir= %s" % dir) if not dir.endswith(':') and not filetools.exists(dir): filetools.mkdir(dir) @@ -30,13 +30,13 @@ class ziptools(object): for nameo in zf.namelist(): name = nameo.replace(':', '_').replace('<', '_').replace('>', '_').replace('|', '_').replace('"', '_').replace('?', '_').replace('*', '_') - logger.info("name=%s" % nameo) + logger.debug("name=%s" % nameo) if not name.endswith('/'): - logger.info("it's not a directory") + logger.debug("it's not a directory") try: (path, filename) = filetools.split(filetools.join(dir, name)) - logger.info("path=%s" % path) - logger.info("name=%s" % name) + logger.debug("path=%s" % path) + logger.debug("name=%s" % name) if folder_to_extract: if path != filetools.join(dir, folder_to_extract): break @@ -49,7 +49,7 @@ class ziptools(object): else: outfilename = filetools.join(dir, name) - logger.info("outfilename=%s" % outfilename) + logger.debug("outfilename=%s" % outfilename) try: if filetools.exists(outfilename) and overwrite_question: from platformcode import platformtools @@ -74,7 +74,7 @@ class ziptools(object): try: zf.close() except: - logger.info("Error closing .zip " + file) + logger.error("Error closing .zip " + file) def _createstructure(self, file, dir): self._makedirs(self._listdirs(file), dir) diff --git a/lib/arm_chromeos.py b/lib/arm_chromeos.py index d4808402..b33d1dfd 100644 --- a/lib/arm_chromeos.py +++ b/lib/arm_chromeos.py @@ -27,7 +27,7 @@ class ChromeOSImage: """ def __init__(self, imgpath): - logger.info('Image Path: ' + imgpath) + logger.debug('Image Path: ' + imgpath) """Prepares the image""" self.imgpath = imgpath self.bstream = self.get_bstream(imgpath) @@ -59,7 +59,7 @@ class ChromeOSImage: self.seek_stream(entries_start * lba_size) if not calcsize(part_format) == entry_size: - logger.info('Partition table entries are not 128 bytes long') + logger.debug('Partition table entries are not 128 bytes long') return 0 for index in range(1, entries_num + 1): # pylint: disable=unused-variable @@ -71,7 +71,7 @@ class ChromeOSImage: break if not offset: - logger.info('Failed to calculate losetup offset.') + logger.debug('Failed to calculate losetup offset.') return 0 return offset @@ -93,7 +93,7 @@ class ChromeOSImage: while True: chunk2 = self.read_stream(chunksize) if not chunk2: - logger.info('File %s not found in the ChromeOS image' % filename) + logger.debug('File %s not found in the ChromeOS image' % filename) return False chunk = chunk1 + chunk2 diff --git a/lib/doh.py b/lib/doh.py index 5375caa5..14ce62a6 100644 --- a/lib/doh.py +++ b/lib/doh.py @@ -47,7 +47,7 @@ def query(name, type='A', server=DOH_SERVER, path="/dns-query", fallback=True): else: retval = [] except Exception as ex: - logger.info("Exception occurred: '%s'" % ex) + logger.error("Exception occurred: '%s'" % ex) if retval is None and fallback: if type == 'A': diff --git a/lib/generictools.py b/lib/generictools.py index a4757b59..44789305 100644 --- a/lib/generictools.py +++ b/lib/generictools.py @@ -25,7 +25,7 @@ intervenido_sucuri = 'Access Denied - Sucuri Website Firewall' def update_title(item): - logger.info() + logger.debug() from core import scraper,support @@ -41,7 +41,7 @@ def update_title(item): The channel must add a method to be able to receive the call from Kodi / Alfa, and be able to call this method: def actualizar_titulos(item): - logger.info() + logger.debug() itemlist = [] from lib import generictools from platformcode import launcher @@ -206,7 +206,7 @@ def update_title(item): def refresh_screen(item): - logger.info() + logger.debug() """ #### Kodi 18 compatibility #### @@ -240,7 +240,7 @@ def refresh_screen(item): def post_tmdb_listado(item, itemlist): - logger.info() + logger.debug() itemlist_fo = [] """ @@ -485,7 +485,7 @@ def post_tmdb_listado(item, itemlist): def post_tmdb_seasons(item, itemlist): - logger.info() + logger.debug() """ @@ -645,7 +645,7 @@ def post_tmdb_seasons(item, itemlist): def post_tmdb_episodios(item, itemlist): - logger.info() + logger.debug() itemlist_fo = [] """ @@ -996,7 +996,7 @@ def post_tmdb_episodios(item, itemlist): def post_tmdb_findvideos(item, itemlist): - logger.info() + logger.debug() """ @@ -1216,7 +1216,7 @@ def post_tmdb_findvideos(item, itemlist): def get_field_from_kodi_DB(item, from_fields='*', files='file'): - logger.info() + logger.debug() """ Call to read from the Kodi DB the input fields received (from_fields, by default "*") of the video indicated in Item @@ -1294,7 +1294,7 @@ def get_field_from_kodi_DB(item, from_fields='*', files='file'): def fail_over_newpct1(item, patron, patron2=None, timeout=None): - logger.info() + logger.debug() import ast """ @@ -1495,7 +1495,7 @@ def fail_over_newpct1(item, patron, patron2=None, timeout=None): def web_intervenida(item, data, desactivar=True): - logger.info() + logger.debug() """ @@ -1578,7 +1578,7 @@ def web_intervenida(item, data, desactivar=True): def regenerate_clones(): - logger.info() + logger.debug() import json from core import videolibrarytools @@ -1592,7 +1592,7 @@ def regenerate_clones(): # Find the paths where to leave the control .json file, and the Video Library json_path = filetools.exists(filetools.join(config.get_runtime_path(), 'verify_cached_torrents.json')) if json_path: - logger.info('Previously repaired video library: WE ARE GOING') + logger.debug('Previously repaired video library: WE ARE GOING') return False json_path = filetools.join(config.get_runtime_path(), 'verify_cached_torrents.json') filetools.write(json_path, json.dumps({"CINE_verify": True})) # Prevents another simultaneous process from being launched @@ -1632,7 +1632,7 @@ def regenerate_clones(): # Delete the Tvshow.nfo files and check if the .nfo has more than one channel and one is clone Newpct1 for file in files: - # logger.info('file - nfos: ' + file) + # logger.debug('file - nfos: ' + file) if 'tvshow.nfo' in file: file_path = filetools.join(root, 'tvshow.nfo') filetools.remove(file_path) @@ -1698,7 +1698,7 @@ def regenerate_clones(): for file in files: file_path = filetools.join(root, file) if '.json' in file: - logger.info('** file: ' + file) + logger.debug('** file: ' + file) canal_json = scrapertools.find_single_match(file, r'\[(\w+)\].json') if canal_json not in nfo.library_urls: filetools.remove(file_path) # we delete the .json is a zombie @@ -1741,7 +1741,7 @@ def regenerate_clones(): def dejuice(data): - logger.info() + logger.debug() # Method to unobtrusive JuicyCodes data import base64 diff --git a/platformcode/autorenumber.py b/platformcode/autorenumber.py index 8c530384..cf5b2d27 100644 --- a/platformcode/autorenumber.py +++ b/platformcode/autorenumber.py @@ -25,14 +25,14 @@ TYPE = "Type" # helper Functions def check(item): - logger.info() + logger.debug() dict_series = load(item) title = item.fulltitle.rstrip() if title in dict_series: title = dict_series[title] return True if ID in title and EPISODE in title else False def filename(item): - logger.info() + logger.debug() name_file = item.channel + "_data.json" path = filetools.join(config.get_data_path(), "settings_channels") fname = filetools.join(path, name_file) @@ -40,7 +40,7 @@ def filename(item): def load(item): - logger.info() + logger.debug() try: json_file = open(filename(item), "r").read() json = jsontools.load(json_file)[TVSHOW_RENUMERATE] @@ -52,7 +52,7 @@ def load(item): def write(item, json): - logger.info() + logger.debug() json_file = open(filename(item), "r").read() js = jsontools.load(json_file) js[TVSHOW_RENUMERATE] = json @@ -71,7 +71,7 @@ def b64(json, mode = 'encode'): def RepresentsInt(s): # Controllo Numro Stagione - logger.info() + logger.debug() try: int(s) return True @@ -79,7 +79,7 @@ def RepresentsInt(s): return False def find_episodes(item): - logger.info() + logger.debug() ch = __import__('channels.' + item.channel, fromlist=["channels.%s" % item.channel]) itemlist = ch.episodios(item) return itemlist @@ -705,7 +705,7 @@ class SelectreNumerationWindow(xbmcgui.WindowXMLDialog): items.append(item) self.seasons[item.getLabel()] = '%sx%s' % (item.getProperty('season'), item.getProperty('episode')) self.items = items - logger.info('SELF',self.seasons) + logger.debug('SELF',self.seasons) def addseasons(self): seasonlist = [] diff --git a/platformcode/globalsearch.py b/platformcode/globalsearch.py index 43aaa72e..bb20e575 100644 --- a/platformcode/globalsearch.py +++ b/platformcode/globalsearch.py @@ -59,7 +59,7 @@ SERVERLIST = 300 class SearchWindow(xbmcgui.WindowXML): def start(self, item): - logger.info() + logger.debug() self.exit = False self.item = item self.lastSearch() @@ -81,7 +81,7 @@ class SearchWindow(xbmcgui.WindowXML): self.doModal() def lastSearch(self): - logger.info() + logger.debug() if not self.item.text: if config.get_setting('last_search'): last_search = channeltools.get_channel_setting('Last_searched', 'search', '') else: last_search = '' @@ -89,7 +89,7 @@ class SearchWindow(xbmcgui.WindowXML): if self.item.text: channeltools.set_channel_setting('Last_searched', self.item.text, 'search') def select(self): - logger.info() + logger.debug() self.PROGRESS.setVisible(False) items = [] if self.persons: @@ -122,7 +122,7 @@ class SearchWindow(xbmcgui.WindowXML): self.NORESULTS.setVisible(True) def actors(self): - logger.info() + logger.debug() self.PROGRESS.setVisible(False) items = [] @@ -174,7 +174,7 @@ class SearchWindow(xbmcgui.WindowXML): self.NORESULTS.setVisible(True) def get_channels(self): - logger.info() + logger.debug() channels_list = [] all_channels = channelselector.filterchannels('all') @@ -196,12 +196,12 @@ class SearchWindow(xbmcgui.WindowXML): if config.get_setting("include_in_global_search", channel) and ch_param.get("active", False): channels_list.append(channel) - logger.info('search in channels:',channels_list) + logger.debug('search in channels:',channels_list) return channels_list def getModule(self, channel): - logger.info() + logger.debug() try: module = __import__('channels.%s' % channel, fromlist=["channels.%s" % channel]) mainlist = getattr(module, 'mainlist')(Item(channel=channel, global_search=True)) @@ -233,7 +233,7 @@ class SearchWindow(xbmcgui.WindowXML): executor.submit(self.get_channel_results, self.item, self.moduleDict, searchAction) def get_channel_results(self, item, module_dict, search_action): - logger.info() + logger.debug() channel = search_action.channel results = [] valid = [] @@ -266,7 +266,7 @@ class SearchWindow(xbmcgui.WindowXML): else: self.update(channel, valid + other) def makeItem(self, item): - logger.info() + logger.debug() thumb = item.thumbnail if item.thumbnail else 'Infoplus/' + item.contentType.replace('show','') it = xbmcgui.ListItem(item.title) it.setProperty('thumb', thumb) @@ -282,7 +282,7 @@ class SearchWindow(xbmcgui.WindowXML): return it def update(self, channel, results): - logger.info('Search on channel', channel) + logger.debug('Search on channel', channel) if results: channelParams = channeltools.get_channel_parameters(channel) name = channelParams['title'] diff --git a/platformcode/launcher.py b/platformcode/launcher.py index bd30e037..6bba9076 100644 --- a/platformcode/launcher.py +++ b/platformcode/launcher.py @@ -19,7 +19,7 @@ def start(): Within this function all calls should go to functions that we want to execute as soon as we open the plugin. """ - logger.info() + logger.debug() # config.set_setting('show_once', True) # Test if all the required directories are created config.verify_directories_created() @@ -37,7 +37,7 @@ def start(): updater.showSavedChangelog() def run(item=None): - logger.info() + logger.debug() if not item: # Extract item from sys.argv if sys.argv[2]: @@ -94,7 +94,7 @@ def run(item=None): # If item has no action, stops here if item.action == "": - logger.info("Item without action") + logger.debug("Item without action") return # Action for main menu in channelselector @@ -193,7 +193,7 @@ def run(item=None): channel_file = os.path.join(config.get_runtime_path(), CHANNELS, item.channel + ".py") - logger.info("channel_file= " + channel_file + ' - ' + CHANNELS + ' - ' + item.channel) + logger.debug("channel_file= " + channel_file + ' - ' + CHANNELS + ' - ' + item.channel) channel = None @@ -213,12 +213,12 @@ def run(item=None): trakt_tools.set_trakt_info(item) except: pass - logger.info("item.action=%s" % item.action.upper()) + logger.debug("item.action=%s" % item.action.upper()) # logger.debug("item_toPlay: " + "\n" + item.tostring('\n')) # First checks if channel has a "play" function if hasattr(channel, 'play'): - logger.info("Executing channel 'play' method") + logger.debug("Executing channel 'play' method") itemlist = channel.play(item) b_favourite = item.isFavourite # Play should return a list of playable URLS @@ -239,7 +239,7 @@ def run(item=None): # If player don't have a "play" function, not uses the standard play from platformtools else: - logger.info("Executing core 'play' method") + logger.debug("Executing core 'play' method") platformtools.play_video(item) # Special action for findvideos, where the plugin looks for known urls @@ -252,8 +252,7 @@ def run(item=None): # If not, uses the generic findvideos function else: - logger.info("No channel 'findvideos' method, " - "executing core method") + logger.debug("No channel 'findvideos' method, " "executing core method") itemlist = servertools.find_video_items(item) if config.get_setting("max_links", "videolibrary") != 0: @@ -297,7 +296,7 @@ def run(item=None): else: filetools.remove(temp_search_file) - logger.info("item.action=%s" % item.action.upper()) + logger.debug("item.action=%s" % item.action.upper()) from core import channeltools if config.get_setting('last_search'): @@ -318,7 +317,7 @@ def run(item=None): # For all other actions else: # import web_pdb; web_pdb.set_trace() - logger.info("Executing channel '%s' method" % item.action) + logger.debug("Executing channel '%s' method" % item.action) itemlist = getattr(channel, item.action)(item) if config.get_setting('trakt_sync'): from core import trakt_tools @@ -399,7 +398,7 @@ def set_search_temp(item): filetools.write(temp_search_file, f) def reorder_itemlist(itemlist): - logger.info() + logger.debug() # logger.debug("Inlet itemlist size: %i" % len(itemlist)) new_list = [] @@ -437,7 +436,7 @@ def reorder_itemlist(itemlist): new_list.extend(mod_list) new_list.extend(not_mod_list) - logger.info("Modified Titles:%i |Unmodified:%i" % (modified, not_modified)) + logger.debug("Modified Titles:%i |Unmodified:%i" % (modified, not_modified)) if len(new_list) == 0: new_list = itemlist @@ -447,7 +446,7 @@ def reorder_itemlist(itemlist): def limit_itemlist(itemlist): - logger.info() + logger.debug() # logger.debug("Inlet itemlist size: %i" % len(itemlist)) try: @@ -480,7 +479,7 @@ def play_from_library(item): itemlist=[] item.fromLibrary = True - logger.info() + logger.debug() # logger.debug("item: \n" + item.tostring('\n')) # Try to reproduce an image (this does nothing and also does not give an error) diff --git a/platformcode/platformtools.py b/platformcode/platformtools.py index be2e432a..d70a04ff 100644 --- a/platformcode/platformtools.py +++ b/platformcode/platformtools.py @@ -206,7 +206,7 @@ def render_items(itemlist, parent_item): """ Function used to render itemlist on kodi """ - logger.info('START render_items') + logger.debug('START render_items') thumb_type = config.get_setting('video_thumbnail_type') from platformcode import shortcuts # from core import httptools @@ -291,7 +291,7 @@ def render_items(itemlist, parent_item): set_view_mode(itemlist[0], parent_item) xbmcplugin.endOfDirectory(_handle) - logger.info('END render_items') + logger.debug('END render_items') def getCurrentView(item=None, parent_item=None): @@ -348,11 +348,11 @@ def set_view_mode(item, parent_item): if content: mode = int(config.get_setting('view_mode_%s' % content).split(',')[-1]) if mode == 0: - logger.info('default mode') + logger.debug('default mode') mode = 55 xbmcplugin.setContent(handle=int(sys.argv[1]), content=Type) xbmc.executebuiltin('Container.SetViewMode(%s)' % mode) - logger.info('TYPE: ' + Type + ' - ' + 'CONTENT: ' + content) + logger.debug('TYPE: ' + Type + ' - ' + 'CONTENT: ' + content) def set_infolabels(listitem, item, player=False): @@ -572,10 +572,10 @@ def is_playing(): def play_video(item, strm=False, force_direct=False, autoplay=False): - logger.info() + logger.debug() logger.debug(item.tostring('\n')) if item.channel == 'downloads': - logger.info("Play local video: %s [%s]" % (item.title, item.url)) + logger.debug("Play local video: %s [%s]" % (item.title, item.url)) xlistitem = xbmcgui.ListItem(path=item.url) xlistitem.setArt({"thumb": item.thumbnail}) set_infolabels(xlistitem, item, True) @@ -583,7 +583,7 @@ def play_video(item, strm=False, force_direct=False, autoplay=False): return default_action = config.get_setting("default_action") - logger.info("default_action=%s" % default_action) + logger.debug("default_action=%s" % default_action) # Open the selection dialog to see the available options opciones, video_urls, seleccion, salir = get_dialogo_opciones(item, default_action, strm, autoplay) @@ -593,8 +593,8 @@ def play_video(item, strm=False, force_direct=False, autoplay=False): seleccion = get_seleccion(default_action, opciones, seleccion, video_urls) if seleccion < 0: return # Canceled box - logger.info("selection=%d" % seleccion) - logger.info("selection=%s" % opciones[seleccion]) + logger.debug("selection=%d" % seleccion) + logger.debug("selection=%s" % opciones[seleccion]) # run the available option, jdwonloader, download, favorites, add to the video library ... IF IT IS NOT PLAY salir = set_opcion(item, seleccion, opciones, video_urls) @@ -755,7 +755,7 @@ def alert_unsopported_server(): def handle_wait(time_to_wait, title, text): - logger.info("handle_wait(time_to_wait=%d)" % time_to_wait) + logger.debug("handle_wait(time_to_wait=%d)" % time_to_wait) espera = dialog_progress(' ' + title, "") secs = 0 @@ -774,15 +774,15 @@ def handle_wait(time_to_wait, title, text): break if cancelled: - logger.info('Wait canceled') + logger.debug('Wait canceled') return False else: - logger.info('Wait finished') + logger.debug('Wait finished') return True def get_dialogo_opciones(item, default_action, strm, autoplay): - logger.info() + logger.debug() # logger.debug(item.tostring('\n')) from core import servertools @@ -866,7 +866,7 @@ def get_dialogo_opciones(item, default_action, strm, autoplay): def set_opcion(item, seleccion, opciones, video_urls): - logger.info() + logger.debug() # logger.debug(item.tostring('\n')) salir = False # You have not chosen anything, most likely because you have given the ESC @@ -916,7 +916,7 @@ def set_opcion(item, seleccion, opciones, video_urls): def get_video_seleccionado(item, seleccion, video_urls): - logger.info() + logger.debug() mediaurl = "" view = False wait_time = 0 @@ -942,7 +942,7 @@ def get_video_seleccionado(item, seleccion, video_urls): mpd = True # If there is no mediaurl it is because the video is not there :) - logger.info("mediaurl=" + mediaurl) + logger.debug("mediaurl=" + mediaurl) if mediaurl == "": if item.server == "unknown": alert_unsopported_server() @@ -959,7 +959,7 @@ def get_video_seleccionado(item, seleccion, video_urls): def set_player(item, xlistitem, mediaurl, view, strm, nfo_path=None, head_nfo=None, item_nfo=None): - logger.info() + logger.debug() # logger.debug("item:\n" + item.tostring('\n')) # Moved del conector "torrent" here if item.server == "torrent": @@ -1046,7 +1046,7 @@ def torrent_client_installed(show_tuple=False): def play_torrent(item, xlistitem, mediaurl): - logger.info() + logger.debug() import time from servers import torrent diff --git a/platformcode/recaptcha.py b/platformcode/recaptcha.py index c4422143..d7265d43 100644 --- a/platformcode/recaptcha.py +++ b/platformcode/recaptcha.py @@ -65,7 +65,7 @@ class Recaptcha(xbmcgui.WindowXMLDialog): data = httptools.downloadpage(self.url, post=post, headers=self.headers).data from platformcode import logger - logger.info(data) + logger.debug(data) self.result = scrapertools.find_single_match(data, '

.*?>([^<]+)<') if self.result: platformtools.dialog_notification("Captcha corretto", "Verifica conclusa") diff --git a/platformcode/shortcuts.py b/platformcode/shortcuts.py index d5effcf2..e2b6eb80 100644 --- a/platformcode/shortcuts.py +++ b/platformcode/shortcuts.py @@ -126,7 +126,7 @@ def SettingOnPosition(item): xbmc.executebuiltin('Addon.OpenSettings(plugin.video.kod)') category = item.category if item.category else 0 setting = item.setting if item.setting else 0 - logger.info('SETTING= ' + str(setting)) + logger.debug('SETTING= ' + str(setting)) xbmc.executebuiltin('SetFocus(%i)' % (category - 100)) xbmc.executebuiltin('SetFocus(%i)' % (setting - 80)) diff --git a/platformcode/side_menu.py b/platformcode/side_menu.py index bef22c2a..fe46c936 100644 --- a/platformcode/side_menu.py +++ b/platformcode/side_menu.py @@ -43,7 +43,7 @@ def set_menu_settings(item): jsontools.update_node(menu_node, 'menu_settings_data.json', "menu") def check_user_home(item): - logger.info() + logger.debug() if os.path.exists(menu_settings_path): menu_node = jsontools.get_node_from_file('menu_settings_data.json', 'menu') if 'user_home' in menu_node: @@ -55,7 +55,7 @@ def check_user_home(item): return item def set_custom_start(item): - logger.info() + logger.debug() if os.path.exists(menu_settings_path): menu_node = jsontools.get_node_from_file('menu_settings_data.json', 'menu') else: @@ -69,7 +69,7 @@ def set_custom_start(item): jsontools.update_node(menu_node, 'menu_settings_data.json', "menu") def get_start_page(): - logger.info() + logger.debug() dictCategory = { config.get_localized_string(70137): 'peliculas', @@ -355,7 +355,7 @@ class Main(xbmcgui.WindowXMLDialog): self.focus -= 1 def run_action(self, item): - logger.info() + logger.debug() if item.menu != True: self.close() xbmc.executebuiltin("Container.update(%s)"%launcher.run(item)) diff --git a/platformcode/subtitletools.py b/platformcode/subtitletools.py index 9a90bd1f..2442eca1 100644 --- a/platformcode/subtitletools.py +++ b/platformcode/subtitletools.py @@ -84,7 +84,7 @@ def regex_tvshow(compare, file, sub=""): def set_Subtitle(): - logger.info() + logger.debug() exts = [".srt", ".sub", ".txt", ".smi", ".ssa", ".ass"] subtitle_folder_path = filetools.join(config.get_data_path(), "subtitles") @@ -216,7 +216,7 @@ def searchSubtitle(item): filetools.mkdir(full_path_tvshow) # title_new + ".mp4" full_path_video_new = xbmc.translatePath( filetools.join(full_path_tvshow, "%s %sx%s.mp4" % (tvshow_title, season, episode))) - logger.info(full_path_video_new) + logger.debug(full_path_video_new) listitem = xbmcgui.ListItem(title_new, iconImage="DefaultVideo.png", thumbnailImage="") listitem.setInfo("video", {"Title": title_new, "Genre": "Tv shows", "episode": int(episode), "season": int(season), "tvshowtitle": tvshow_title}) @@ -230,7 +230,7 @@ def searchSubtitle(item): try: filetools.copy(path_video_temp, full_path_video_new) copy = True - logger.info("nuevo path =" + full_path_video_new) + logger.debug("nuevo path =" + full_path_video_new) time.sleep(2) playlist = xbmc.PlayList(xbmc.PLAYLIST_VIDEO) playlist.clear() @@ -288,7 +288,7 @@ def get_from_subdivx(sub_url): :return: The path to the unzipped subtitle """ - logger.info() + logger.debug() sub = '' sub_dir = os.path.join(config.get_data_path(), 'temp_subs') @@ -312,9 +312,9 @@ def get_from_subdivx(sub_url): filetools.write(filename, data_dl) sub = extract_file_online(sub_dir, filename) except: - logger.info('sub invalid') + logger.debug('sub invalid') else: - logger.info('sub invalid') + logger.debug('sub invalid') return sub @@ -328,7 +328,7 @@ def extract_file_online(path, filename): :return: """ - logger.info() + logger.debug() url = "http://online.b1.org/rest/online/upload" diff --git a/platformcode/xbmc_info_window.py b/platformcode/xbmc_info_window.py index 6f3a994e..ac588928 100644 --- a/platformcode/xbmc_info_window.py +++ b/platformcode/xbmc_info_window.py @@ -32,7 +32,7 @@ class InfoWindow(xbmcgui.WindowXMLDialog): self.scraper = scraper self.doModal() - logger.info('RESPONSE',self.response) + logger.debug('RESPONSE',self.response) return self.response def make_items(self, i, result): @@ -52,7 +52,7 @@ class InfoWindow(xbmcgui.WindowXMLDialog): self.setCoordinateResolution(2) with futures.ThreadPoolExecutor() as executor: for i, result in enumerate(self.results): - logger.info(result) + logger.debug(result) if ('seriesName' in result and result['seriesName']) or ('name' in result and result['name']) or ('title' in result and result['title']): self.items += [executor.submit(self.make_items, i, result).result()] self.items.sort(key=lambda it: int(it.getProperty('position'))) diff --git a/platformcode/xbmc_videolibrary.py b/platformcode/xbmc_videolibrary.py index ea0c7745..96d9c6da 100644 --- a/platformcode/xbmc_videolibrary.py +++ b/platformcode/xbmc_videolibrary.py @@ -22,7 +22,7 @@ from xml.dom import minidom def mark_auto_as_watched(item, nfo_path=None, head_nfo=None, item_nfo=None): def mark_as_watched_subThread(item, nfo_path, head_nfo, item_nfo): - logger.info() + logger.debug() # logger.debug("item:\n" + item.tostring('\n')) time_limit = time.time() + 30 @@ -53,7 +53,7 @@ def mark_auto_as_watched(item, nfo_path=None, head_nfo=None, item_nfo=None): # Mark as Watched if actual_time > mark_time and not marked: - logger.debug("Marked as Watched") + logger.info("Marked as Watched") item.playcount = 1 marked = True show_server = False @@ -104,7 +104,7 @@ def sync_trakt_addon(path_folder): """ Updates the values ​​of episodes seen if """ - logger.info() + logger.debug() # if the addon exists we do the search if xbmc.getCondVisibility('System.HasAddon("script.trakt")'): # we import dependencies @@ -230,7 +230,7 @@ def sync_trakt_kodi(silent=True): notificacion = False xbmc.executebuiltin('RunScript(script.trakt,action=sync,silent=%s)' % silent) - logger.info("Synchronization with Trakt started") + logger.debug("Synchronization with Trakt started") if notificacion: platformtools.dialog_notification(config.get_localized_string(20000), config.get_localized_string(60045), sound=False, time=2000) @@ -244,7 +244,7 @@ def mark_content_as_watched_on_kodi(item, value=1): @type value: int @param value: > 0 for seen, 0 for not seen """ - logger.info() + logger.debug() # logger.debug("item:\n" + item.tostring('\n')) payload_f = '' @@ -316,7 +316,7 @@ def mark_season_as_watched_on_kodi(item, value=1): @type value: int @param value: > 0 for seen, 0 for not seen """ - logger.info() + logger.debug() # logger.debug("item:\n" + item.tostring('\n')) # We can only mark the season as seen in the Kodi database if the database is local, in case of sharing database this functionality will not work @@ -350,7 +350,7 @@ def mark_content_as_watched_on_kod(path): @type str: path @param path: content folder to mark """ - logger.info() + logger.debug() #logger.debug("path: " + path) FOLDER_MOVIES = config.get_setting("folder_movies") @@ -443,7 +443,7 @@ def get_data(payload): import urllib.request as urllib except ImportError: import urllib - logger.info("payload: %s" % payload) + logger.debug("payload: %s" % payload) # Required header for XBMC JSON-RPC calls, otherwise you'll get a 415 HTTP response code - Unsupported media type headers = {'content-type': 'application/json'} @@ -460,7 +460,7 @@ def get_data(payload): response = f.read() f.close() - logger.info("get_data: response %s" % response) + logger.debug("get_data: response %s" % response) data = jsontools.load(response) except Exception as ex: template = "An exception of type %s occured. Arguments:\n%r" @@ -476,7 +476,7 @@ def get_data(payload): logger.error("error en xbmc.executeJSONRPC: %s" % message) data = ["error"] - logger.info("data: %s" % data) + logger.debug("data: %s" % data) return data @@ -490,7 +490,7 @@ def update(folder_content=config.get_setting("folder_tvshows"), folder=""): @type folder: str @param folder: name of the folder to scan. """ - logger.info(folder) + logger.debug(folder) payload = { "jsonrpc": "2.0", @@ -554,7 +554,7 @@ def set_content(content_type, silent=False, custom=False): @type content_type: str ('movie' o 'tvshow') @param content_type: content type to configure, series or movies """ - logger.info() + logger.debug() continuar = True msg_text = "" videolibrarypath = config.get_setting("videolibrarypath") @@ -580,7 +580,7 @@ def set_content(content_type, silent=False, custom=False): try: # Install metadata.themoviedb.org xbmc.executebuiltin('InstallAddon(metadata.themoviedb.org)', True) - logger.info("Instalado el Scraper de películas de TheMovieDB") + logger.debug("Instalado el Scraper de películas de TheMovieDB") except: pass @@ -634,7 +634,7 @@ def set_content(content_type, silent=False, custom=False): try: # Install metadata.tvdb.com xbmc.executebuiltin('InstallAddon(metadata.tvdb.com)', True) - logger.info("The TVDB series Scraper installed ") + logger.debug("The TVDB series Scraper installed ") except: pass @@ -729,7 +729,7 @@ def set_content(content_type, silent=False, custom=False): strScraper = 'metadata.universal' path_settings = xbmc.translatePath("special://profile/addon_data/metadata.universal/settings.xml") if not os.path.exists(path_settings): - logger.info("%s: %s" % (content_type, path_settings + " doesn't exist")) + logger.debug("%s: %s" % (content_type, path_settings + " doesn't exist")) return continuar settings_data = filetools.read(path_settings) strSettings = ' '.join(settings_data.split()).replace("> <", "><") @@ -748,7 +748,7 @@ def set_content(content_type, silent=False, custom=False): strScraper = 'metadata.tvshows.themoviedb.org' path_settings = xbmc.translatePath("special://profile/addon_data/metadata.tvshows.themoviedb.org/settings.xml") if not os.path.exists(path_settings): - logger.info("%s: %s" % (content_type, path_settings + " doesn't exist")) + logger.debug("%s: %s" % (content_type, path_settings + " doesn't exist")) return continuar settings_data = filetools.read(path_settings) strSettings = ' '.join(settings_data.split()).replace("> <", "><") @@ -758,7 +758,7 @@ def set_content(content_type, silent=False, custom=False): videolibrarypath += sep strPath = videolibrarypath + config.get_setting("folder_tvshows") + sep - logger.info("%s: %s" % (content_type, strPath)) + logger.debug("%s: %s" % (content_type, strPath)) # We check if strPath already exists in the DB to avoid duplicates sql = 'SELECT idPath FROM path where strPath="%s"' % strPath nun_records, records = execute_sql_kodi(sql) @@ -800,15 +800,15 @@ def set_content(content_type, silent=False, custom=False): heading = config.get_localized_string(70103) % content_type msg_text = config.get_localized_string(70104) - logger.info("%s: %s" % (heading, msg_text)) + logger.debug("%s: %s" % (heading, msg_text)) return continuar def update_db(old_path, new_path, old_movies_folder, new_movies_folder, old_tvshows_folder, new_tvshows_folder, progress): def path_replace(path, old, new): - logger.info() - logger.info('path: ' + path + ', old: ' + old + ', new: ' + new) + logger.debug() + logger.debug('path: ' + path + ', old: ' + old + ', new: ' + new) if new.startswith("special://") or '://' in new: sep = '/' else: sep = os.sep @@ -819,7 +819,7 @@ def update_db(old_path, new_path, old_movies_folder, new_movies_folder, old_tvsh return path - logger.info() + logger.debug() sql_old_path = old_path if sql_old_path.startswith("special://"): @@ -831,10 +831,10 @@ def update_db(old_path, new_path, old_movies_folder, new_movies_folder, old_tvsh if not sql_old_path.endswith(sep): sql_old_path += sep - logger.info('sql_old_path: ' + sql_old_path) + logger.debug('sql_old_path: ' + sql_old_path) # search MAIN path in the DB sql = 'SELECT idPath, strPath FROM path where strPath LIKE "%s"' % sql_old_path - logger.info('sql: ' + sql) + logger.debug('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) # change main path @@ -842,7 +842,7 @@ def update_db(old_path, new_path, old_movies_folder, new_movies_folder, old_tvsh idPath = records[0][0] strPath = path_replace(records[0][1], old_path, new_path) sql = 'UPDATE path SET strPath="%s" WHERE idPath=%s' % (strPath, idPath) - logger.info('sql: ' + sql) + logger.debug('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) else: progress.update(100) @@ -859,7 +859,7 @@ def update_db(old_path, new_path, old_movies_folder, new_movies_folder, old_tvsh # Search Main Sub Folder sql = 'SELECT idPath, strPath FROM path where strPath LIKE "%s"' % sql_old_folder - logger.info('sql: ' + sql) + logger.debug('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) # Change Main Sub Folder @@ -868,13 +868,13 @@ def update_db(old_path, new_path, old_movies_folder, new_movies_folder, old_tvsh idPath = record[0] strPath = path_replace(record[1], filetools.join(old_path, OldFolder), filetools.join(new_path, NewFolder)) sql = 'UPDATE path SET strPath="%s" WHERE idPath=%s' % (strPath, idPath) - logger.info('sql: ' + sql) + logger.debug('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) # Search if Sub Folder exixt in all paths sql_old_folder += '%' sql = 'SELECT idPath, strPath FROM path where strPath LIKE "%s"' % sql_old_folder - logger.info('sql: ' + sql) + logger.debug('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) #Change Sub Folder in all paths @@ -883,7 +883,7 @@ def update_db(old_path, new_path, old_movies_folder, new_movies_folder, old_tvsh idPath = record[0] strPath = path_replace(record[1], filetools.join(old_path, OldFolder), filetools.join(new_path, NewFolder)) sql = 'UPDATE path SET strPath="%s" WHERE idPath=%s' % (strPath, idPath) - logger.info('sql: ' + sql) + logger.debug('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) @@ -891,27 +891,27 @@ def update_db(old_path, new_path, old_movies_folder, new_movies_folder, old_tvsh # if is Movie Folder # search and modify in "movie" sql = 'SELECT idMovie, c22 FROM movie where c22 LIKE "%s"' % sql_old_folder - logger.info('sql: ' + sql) + logger.debug('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) if records: for record in records: idMovie = record[0] strPath = path_replace(record[1], filetools.join(old_path, OldFolder), filetools.join(new_path, NewFolder)) sql = 'UPDATE movie SET c22="%s" WHERE idMovie=%s' % (strPath, idMovie) - logger.info('sql: ' + sql) + logger.debug('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) else: # if is TV Show Folder # search and modify in "episode" sql = 'SELECT idEpisode, c18 FROM episode where c18 LIKE "%s"' % sql_old_folder - logger.info('sql: ' + sql) + logger.debug('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) if records: for record in records: idEpisode = record[0] strPath = path_replace(record[1], filetools.join(old_path, OldFolder), filetools.join(new_path, NewFolder)) sql = 'UPDATE episode SET c18="%s" WHERE idEpisode=%s' % (strPath, idEpisode) - logger.info('sql: ' + sql) + logger.debug('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) p += 5 progress.update(p, config.get_localized_string(20000) + '\n' + config.get_localized_string(80013)) @@ -936,26 +936,26 @@ def clean(path_list=[]): return path, sep - logger.info() + logger.debug() progress = platformtools.dialog_progress_bg(config.get_localized_string(20000), config.get_localized_string(80025)) progress.update(0) # if the path list is empty, clean the entire video library if not path_list: - logger.info('the path list is empty, clean the entire video library') + logger.debug('the path list is empty, clean the entire video library') if not config.get_setting("videolibrary_kodi"): sql_path, sep = sql_format(config.get_setting("videolibrarypath")) if not sql_path.endswith(sep): sql_path += sep sql = 'SELECT idPath FROM path where strPath LIKE "%s"' % sql_path - logger.info('sql: ' + sql) + logger.debug('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) idPath = records[0][0] sql = 'DELETE from path WHERE idPath=%s' % idPath - logger.info('sql: ' + sql) + logger.debug('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) sql = 'DELETE from path WHERE idParentPath=%s' % idPath - logger.info('sql: ' + sql) + logger.debug('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) from core import videolibrarytools @@ -969,7 +969,7 @@ def clean(path_list=[]): if filetools.exists(tvshow_nfo): path_list.append(filetools.join(config.get_setting("videolibrarypath"), videolibrarytools.FOLDER_TVSHOWS, folder)) - logger.info('path_list: ' + str(path_list)) + logger.debug('path_list: ' + str(path_list)) if path_list: t = float(100) / len(path_list) for i, path in enumerate(path_list): progress.update(int(math.ceil((i + 1) * t))) @@ -979,13 +979,13 @@ def clean(path_list=[]): sql_path, sep = sql_format(path) if filetools.isdir(path) and not sql_path.endswith(sep): sql_path += sep - logger.info('path: ' + path) - logger.info('sql_path: ' + sql_path) + logger.debug('path: ' + path) + logger.debug('sql_path: ' + sql_path) if filetools.isdir(path): # search movie in the DB sql = 'SELECT idMovie FROM movie where c22 LIKE "%s"' % (sql_path + '%') - logger.info('sql: ' + sql) + logger.debug('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) # delete movie if records: @@ -994,7 +994,7 @@ def clean(path_list=[]): continue # search TV show in the DB sql = 'SELECT idShow FROM tvshow_view where strPath LIKE "%s"' % sql_path - logger.info('sql: ' + sql) + logger.debug('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) # delete TV show if records: @@ -1003,7 +1003,7 @@ def clean(path_list=[]): elif config.get_setting("folder_movies") in sql_path: # search movie in the DB sql = 'SELECT idMovie FROM movie where c22 LIKE "%s"' % sql_path - logger.info('sql: ' + sql) + logger.debug('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) # delete movie if records: @@ -1012,7 +1012,7 @@ def clean(path_list=[]): else: # search episode in the DB sql = 'SELECT idEpisode FROM episode where c18 LIKE "%s"' % sql_path - logger.info('sql: ' + sql) + logger.debug('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) # delete episode if records: @@ -1031,7 +1031,7 @@ def check_db(path): ret = False sql_path = '%' + sep + path.split(sep)[-1] + sep + '%' sql = 'SELECT idShow FROM tvshow_view where strPath LIKE "%s"' % sql_path - logger.info('sql: ' + sql) + logger.debug('sql: ' + sql) nun_records, records = execute_sql_kodi(sql) if records: ret = True @@ -1048,7 +1048,7 @@ def execute_sql_kodi(sql): @return: list with the query result @rtype records: list of tuples """ - logger.info() + logger.debug() file_db = "" nun_records = 0 records = None @@ -1069,14 +1069,14 @@ def execute_sql_kodi(sql): break if file_db: - logger.info("DB file: %s" % file_db) + logger.debug("DB file: %s" % file_db) conn = None try: import sqlite3 conn = sqlite3.connect(file_db) cursor = conn.cursor() - logger.info("Running sql: %s" % sql) + logger.debug("Running sql: %s" % sql) cursor.execute(sql) conn.commit() @@ -1090,7 +1090,7 @@ def execute_sql_kodi(sql): nun_records = conn.total_changes conn.close() - logger.info("Query executed. Records: %s" % nun_records) + logger.debug("Query executed. Records: %s" % nun_records) except: logger.error("Error executing sql query") @@ -1110,7 +1110,7 @@ def check_sources(new_movies_path='', new_tvshows_path=''): if not path.endswith(sep): path += sep return path - logger.info() + logger.debug() new_movies_path = format_path(new_movies_path) new_tvshows_path = format_path(new_tvshows_path) @@ -1140,7 +1140,7 @@ def check_sources(new_movies_path='', new_tvshows_path=''): def update_sources(new='', old=''): - logger.info() + logger.debug() if new == old: return SOURCES_PATH = xbmc.translatePath("special://userdata/sources.xml") @@ -1182,9 +1182,9 @@ def update_sources(new='', old=''): # create new path list_path = [p.firstChild.data for p in paths_node] if new in list_path: - logger.info("The path %s already exists in sources.xml" % new) + logger.debug("The path %s already exists in sources.xml" % new) return - logger.info("The path %s does not exist in sources.xml" % new) + logger.debug("The path %s does not exist in sources.xml" % new) # if the path does not exist we create one source_node = xmldoc.createElement("source") @@ -1223,7 +1223,7 @@ def update_sources(new='', old=''): def ask_set_content(silent=False): - logger.info() + logger.debug() logger.debug("videolibrary_kodi %s" % config.get_setting("videolibrary_kodi")) def do_config(custom=False): @@ -1280,7 +1280,7 @@ def ask_set_content(silent=False): def next_ep(item): from core.item import Item - logger.info() + logger.debug() item.next_ep = False # check if next file exist @@ -1296,7 +1296,7 @@ def next_ep(item): nextIndex = fileList.index(current_filename) + 1 if nextIndex == 0 or nextIndex == len(fileList): next_file = None else: next_file = fileList[nextIndex] - logger.info('Next File:' + str(next_file)) + logger.debug('Next File:' + str(next_file)) # start next episode window afther x time if next_file: diff --git a/servers/akvideo.py b/servers/akvideo.py index 346b3ff3..cabb9b02 100644 --- a/servers/akvideo.py +++ b/servers/akvideo.py @@ -13,7 +13,7 @@ headers = [['User-Agent', 'Mozilla/5.0 (Windows NT 6.1; WOW64; rv:54.0) Gecko/20 def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) # page_url = re.sub('akvideo.stream/(?:video/|video\\.php\\?file_code=)?(?:embed-)?([a-zA-Z0-9]+)','akvideo.stream/video/\\1',page_url) global data page = httptools.downloadpage(page_url, headers=headers) @@ -32,18 +32,18 @@ def test_video_exists(page_url): # ID, code = scrapertools.find_single_match(data, r"""input\D*id=(?:'|")([^'"]+)(?:'|").*?value='([a-z0-9]+)""") # post = urllib.urlencode({ID: code}) - # logger.info('PAGE DATA' + data) + # logger.debug('PAGE DATA' + data) if "File Not Found" in data: return False, config.get_localized_string(70449) % "Akvideo" return True, "" def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info(" url=" + page_url) + logger.debug(" url=" + page_url) video_urls = [] global data - # logger.info('PAGE DATA' + data) + # logger.debug('PAGE DATA' + data) # sitekey = scrapertools.find_single_match(data, 'data-sitekey="([^"]+)') # captcha = platformtools.show_recaptcha(sitekey, page_url) if sitekey else '' # diff --git a/servers/anavids.py b/servers/anavids.py index c4169620..dd506dd3 100644 --- a/servers/anavids.py +++ b/servers/anavids.py @@ -6,7 +6,7 @@ from platformcode import config, logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url, cookies=False).data if 'File you are looking for is not found.' in data: diff --git a/servers/animeid.py b/servers/animeid.py index 3cec4070..1eb8414a 100644 --- a/servers/animeid.py +++ b/servers/animeid.py @@ -6,7 +6,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "no longer exists" in data or "to copyright issues" in data: return False, config.get_localized_string(70449) % "animeid" @@ -16,7 +16,7 @@ def test_video_exists(page_url): def get_video_url(page_url, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data video_urls = [] label, videourl = scrapertools.find_single_match(data, 'label":"([^"]+)".*?file":"([^"]+)') diff --git a/servers/anonfile.py b/servers/anonfile.py index 4191caff..9d48396b 100644 --- a/servers/anonfile.py +++ b/servers/anonfile.py @@ -9,7 +9,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) response = httptools.downloadpage(page_url) if not response.success or "Not Found" in response.data or "File was deleted" in response.data or "is no longer available" in response.data: return False, config.get_localized_string(70449) % "anonfile" @@ -17,7 +17,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] data = httptools.downloadpage(page_url).data patron = 'download-url.*?href="([^"]+)"' diff --git a/servers/archiveorg.py b/servers/archiveorg.py index 61677b02..df81e381 100644 --- a/servers/archiveorg.py +++ b/servers/archiveorg.py @@ -9,7 +9,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url) if data.code == 404: return False, config.get_localized_string(70449) % "ArchiveOrg" @@ -17,7 +17,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] data = httptools.downloadpage(page_url).data patron = '' diff --git a/servers/backin.py b/servers/backin.py index c070d5d6..8214a944 100644 --- a/servers/backin.py +++ b/servers/backin.py @@ -9,7 +9,7 @@ except ImportError: from urllib import urlencode def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) if 'http://' in page_url: # fastids page_url = httptools.downloadpage(page_url, follow_redirects=False, only_headers=True).headers['location'] @@ -24,7 +24,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("page_url=" + page_url) + logger.debug("page_url=" + page_url) video_urls = [] @@ -36,18 +36,18 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= if data_pack: from lib import jsunpack data = jsunpack.unpack(data_pack) - logger.info("page_url=" + data) + logger.debug("page_url=" + data) # URL url = scrapertools.find_single_match(data, r'"src"value="([^"]+)"') if not url: url = scrapertools.find_single_match(data, r'file\s*:\s*"([^"]+)"') - logger.info("URL=" + str(url)) + logger.debug("URL=" + str(url)) # URL del vídeo video_urls.append([".mp4" + " [backin]", url]) for video_url in video_urls: - logger.info("%s - %s" % (video_url[0], httptools.get_url_headers(video_url[1]))) + logger.debug("%s - %s" % (video_url[0], httptools.get_url_headers(video_url[1]))) return video_urls diff --git a/servers/badshare.py b/servers/badshare.py index 78c62986..7e713a9a 100644 --- a/servers/badshare.py +++ b/servers/badshare.py @@ -11,7 +11,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) global page page = httptools.downloadpage(page_url) if not page.success: @@ -20,7 +20,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] ext = '.mp4' diff --git a/servers/bdupload.py b/servers/bdupload.py index 6321dd3a..cd72b590 100644 --- a/servers/bdupload.py +++ b/servers/bdupload.py @@ -10,7 +10,7 @@ headers = {'User-Agent': 'Mozilla/5.0 (Linux; Android 6.0; Nexus 5 Build/MRA58N) def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "Archive no Encontrado" in data: return False, config.get_localized_string(70449) % "bdupload" @@ -19,7 +19,7 @@ def test_video_exists(page_url): def get_video_url(page_url, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data post = "" patron = '(?s)type="hidden" name="([^"]+)".*?value="([^"]*)"' diff --git a/servers/cinemaupload.py b/servers/cinemaupload.py index c3f05da8..a4d75d56 100644 --- a/servers/cinemaupload.py +++ b/servers/cinemaupload.py @@ -11,7 +11,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url) if data.code == 404: return False, config.get_localized_string(70449) % "CinemaUpload" @@ -19,7 +19,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] data = httptools.downloadpage(page_url).data data = re.sub(r'\n|\r|\t| |
|\s{2,}', "", data) diff --git a/servers/clicknupload.py b/servers/clicknupload.py index f54c5d5c..a2f95930 100755 --- a/servers/clicknupload.py +++ b/servers/clicknupload.py @@ -22,7 +22,7 @@ excption = False def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = get_data(page_url.replace(".org", ".me")) if "File Not Found" in data: return False, config.get_localized_string(70449) % "Clicknupload" @@ -31,7 +31,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) data = get_data(page_url.replace(".org", ".me")) @@ -51,7 +51,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= media_url = media.rsplit('/', 1)[0] + "/" + url_strip video_urls.append([scrapertools.get_filename_from_url(media_url)[-4:] + " [clicknupload]", media_url]) for video_url in video_urls: - logger.info("%s - %s" % (video_url[0], video_url[1])) + logger.debug("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/clipwatching.py b/servers/clipwatching.py index 93aa4804..6ac3ec78 100644 --- a/servers/clipwatching.py +++ b/servers/clipwatching.py @@ -6,7 +6,7 @@ from lib import jsunpack from platformcode import logger, config def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url).data if "File Not Found" in data or "File was deleted" in data: @@ -15,7 +15,7 @@ def test_video_exists(page_url): def get_video_url(page_url, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] try: diff --git a/servers/cloudvideo.py b/servers/cloudvideo.py index b7885afe..36e299fc 100644 --- a/servers/cloudvideo.py +++ b/servers/cloudvideo.py @@ -8,7 +8,7 @@ from lib import jsunpack def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) html = httptools.downloadpage(page_url) global data data = html.data @@ -18,7 +18,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] global data # data = httptools.downloadpage(page_url).data diff --git a/servers/crunchyroll.py b/servers/crunchyroll.py index 2d252b3a..31384209 100755 --- a/servers/crunchyroll.py +++ b/servers/crunchyroll.py @@ -30,7 +30,7 @@ proxy = "https://www.usa-proxy.org/" def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url, headers=GLOBAL_HEADER).data if "Este es un clip de muestra" in data: @@ -44,7 +44,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): #page_url='https://www.crunchyroll.com/es-es/one-piece/episode-891-climbing-up-a-waterfall-a-great-journey-through-the-land-of-wanos-sea-zone-786643' - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] if "crunchyroll.com" in page_url: media_id = page_url.rsplit("-", 1)[1] @@ -94,7 +94,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= file_sub = "" video_urls.append(["%s %sp [crunchyroll]" % (filename, quality), media_url, 0, file_sub]) for video_url in video_urls: - logger.info("%s - %s" % (video_url[0], video_url[1])) + logger.debug("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/dailymotion.py b/servers/dailymotion.py index 7de4b4fd..b6d3ddbc 100644 --- a/servers/dailymotion.py +++ b/servers/dailymotion.py @@ -6,7 +6,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) global response response = httptools.downloadpage(page_url, cookies=False) @@ -18,7 +18,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] cookie = {'Cookie': response.headers["set-cookie"]} data = response.data.replace("\\", "") @@ -40,5 +40,5 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= stream_url = stream_url_http video_urls.append(["%sp .%s [dailymotion]" % (calidad, stream_type), stream_url, 0, subtitle]) for video_url in video_urls: - logger.info("%s - %s" % (video_url[0], video_url[1])) + logger.debug("%s - %s" % (video_url[0], video_url[1])) return video_urls \ No newline at end of file diff --git a/servers/debriders/alldebrid.py b/servers/debriders/alldebrid.py index 6f0bd18a..3818fffd 100644 --- a/servers/debriders/alldebrid.py +++ b/servers/debriders/alldebrid.py @@ -7,7 +7,7 @@ from platformcode import logger # Returns an array of possible video url's from the page_url def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info() + logger.debug() page_url = correct_url(page_url) dd1 = httptools.downloadpage("https://api.alldebrid.com/user/login?agent=mySoft&username=%s&password=%s" %(user, password)).data token = scrapertools.find_single_match(dd1, 'token":"([^"]+)') diff --git a/servers/debriders/realdebrid.py b/servers/debriders/realdebrid.py index 7d8b855d..d8e6d4f5 100755 --- a/servers/debriders/realdebrid.py +++ b/servers/debriders/realdebrid.py @@ -22,7 +22,7 @@ headers = {'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; WOW64; rv:65.0) Gecko/20 # Returns an array of possible video url's from the page_url def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url='%s' , video_password=%s)" % (page_url, video_password)) + logger.debug("(page_url='%s' , video_password=%s)" % (page_url, video_password)) page_url = page_url.replace(".nz/embed", ".nz/") # Se comprueba si existe un token guardado y sino se ejecuta el proceso de autentificación token_auth = config.get_setting("token", server="realdebrid") @@ -99,7 +99,7 @@ def get_enlaces(data): def authentication(): - logger.info() + logger.debug() try: client_id = "YTWNFBIJEEBP6" diff --git a/servers/decrypters/adfly.py b/servers/decrypters/adfly.py index 20b34d18..8acac50a 100755 --- a/servers/decrypters/adfly.py +++ b/servers/decrypters/adfly.py @@ -8,7 +8,7 @@ from platformcode import logger def get_long_url(short_url): - logger.info("short_url = '%s'" % short_url) + logger.debug("short_url = '%s'" % short_url) data = httptools.downloadpage(short_url).data ysmm = scrapertools.find_single_match(data, "var ysmm = '([^']+)';") diff --git a/servers/decrypters/linkbucks.py b/servers/decrypters/linkbucks.py index 5b15ec00..e1a7b521 100755 --- a/servers/decrypters/linkbucks.py +++ b/servers/decrypters/linkbucks.py @@ -17,7 +17,7 @@ from platformcode import logger # Obtiene la URL que hay detrás de un enlace a linkbucks def get_long_url(short_url): - logger.info("(short_url='%s')" % short_url) + logger.debug("(short_url='%s')" % short_url) request_headers = [] request_headers.append(["User-Agent", @@ -33,17 +33,17 @@ def get_long_url(short_url): while True: for name, value in response_headers: if name == "set-cookie": - logger.info("Set-Cookie: " + value) + logger.debug("Set-Cookie: " + value) cookie_name = scrapertools.scrapertools.find_single_match(value, '(.*?)\=.*?\;') cookie_value = scrapertools.scrapertools.find_single_match(value, '.*?\=(.*?)\;') request_headers.append(["Cookie", cookie_name + "=" + cookie_value]) body, response_headers = scrapertools.read_body_and_headers(url, headers=request_headers) - logger.info("body=" + body) + logger.debug("body=" + body) try: location = scrapertools.scrapertools.find_single_match(body, '([^<]+)') - logger.info("location=" + location) + logger.debug("location=" + location) break except: n = n + 1 diff --git a/servers/decrypters/longurl.py b/servers/decrypters/longurl.py index 4c6cb012..b434f95b 100755 --- a/servers/decrypters/longurl.py +++ b/servers/decrypters/longurl.py @@ -38,15 +38,15 @@ servers = get_server_list() def get_long_urls(data): - logger.info() + logger.debug() patron = 'Download error' in data: return False, "El enlace no es válido
o ha sido borrado de gigasize" @@ -13,7 +13,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] return video_urls diff --git a/servers/googlevideo.py b/servers/googlevideo.py index 3a411e80..8f857fde 100755 --- a/servers/googlevideo.py +++ b/servers/googlevideo.py @@ -7,14 +7,14 @@ from platformcode import logger def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] # Lo extrae a partir de flashvideodownloader.org if page_url.startswith("http://"): url = 'http://www.flashvideodownloader.org/download.php?u=' + page_url else: url = 'http://www.flashvideodownloader.org/download.php?u=http://video.google.com/videoplay?docid=' + page_url - logger.info("url=" + url) + logger.debug("url=" + url) data = httptools.downloadpage(url).data # Extrae el vídeo @@ -24,6 +24,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls.append(["[googlevideo]", newmatches[0]]) for video_url in video_urls: - logger.info("%s - %s" % (video_url[0], video_url[1])) + logger.debug("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/gounlimited.py b/servers/gounlimited.py index 8bba60b4..53642c2f 100644 --- a/servers/gounlimited.py +++ b/servers/gounlimited.py @@ -20,14 +20,14 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] global data data = re.sub(r'"|\n|\r|\t| |
|\s{2,}', "", data) - # logger.info('GOUN DATA= '+data) + # logger.debug('GOUN DATA= '+data) packed_data = scrapertools.find_single_match(data, "javascript'>(eval.*?)") unpacked = jsunpack.unpack(packed_data) - # logger.info('GOUN DATA= '+unpacked) + # logger.debug('GOUN DATA= '+unpacked) patron = r"sources..([^\]]+)" matches = re.compile(patron, re.DOTALL).findall(unpacked) if not matches: diff --git a/servers/gvideo.py b/servers/gvideo.py index aa9712c4..edaaf8ac 100644 --- a/servers/gvideo.py +++ b/servers/gvideo.py @@ -42,7 +42,7 @@ def test_video_exists(page_url): def get_video_url(page_url, user="", password="", video_password=""): - logger.info() + logger.debug() video_urls = [] urls = [] streams =[] diff --git a/servers/hdload.py b/servers/hdload.py index 30240022..35392d6a 100644 --- a/servers/hdload.py +++ b/servers/hdload.py @@ -7,7 +7,7 @@ from platformcode import config, logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url, cookies=False).data if 'Not found id' in data: @@ -17,12 +17,12 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info() + logger.debug() itemlist = [] - logger.info(page_url) + logger.debug(page_url) data = httptools.downloadpage(page_url, post='').data - logger.info(data) + logger.debug(data) url = base64.b64decode(data) itemlist.append([".mp4 [HDLoad]", url]) diff --git a/servers/hdmario.py b/servers/hdmario.py index bf50a3cb..1eea0d75 100644 --- a/servers/hdmario.py +++ b/servers/hdmario.py @@ -8,12 +8,12 @@ from lib.fakeMail import Gmailnator baseUrl = 'https://hdmario.live' def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) global page, data page = httptools.downloadpage(page_url) data = page.data - logger.info(page.url) + logger.debug(page.url) if "the page you are looking for could not be found" in data: return False, config.get_localized_string(70449) % "HDmario" @@ -54,12 +54,12 @@ def registerOrLogin(page_url): else: import random import string - logger.info('Registrazione automatica in corso') + logger.debug('Registrazione automatica in corso') mailbox = Gmailnator() randPsw = ''.join(random.choice(string.ascii_letters + string.digits) for i in range(10)) captcha = httptools.downloadpage(baseUrl + '/captchaInfo').json - logger.info('email: ' + mailbox.address) - logger.info('pass: ' + randPsw) + logger.debug('email: ' + mailbox.address) + logger.debug('pass: ' + randPsw) reg = platformtools.dialog_register(baseUrl + '/register/', email=True, password=True, email_default=mailbox.address, password_default=randPsw, captcha_img=captcha['captchaUrl']) if not reg: return False @@ -90,7 +90,7 @@ def registerOrLogin(page_url): else: platformtools.dialog_ok('HDmario', 'Hai modificato la mail quindi KoD non sarà in grado di effettuare la verifica in autonomia, apri la casella ' + reg['email'] + ' e clicca sul link. Premi ok quando fatto') - logger.info('Registrazione completata') + logger.debug('Registrazione completata') return True @@ -98,7 +98,7 @@ def registerOrLogin(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): global page, data page_url = page_url.replace('?', '') - logger.info("url=" + page_url) + logger.debug("url=" + page_url) if 'unconfirmed' in page.url: id = page_url.split('/')[-1] @@ -109,7 +109,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= } httptools.downloadpage(page.url, post=postData) mail = mailbox.waitForMail() - logger.info(mail) + logger.debug(mail) if mail: code = mail.subject.split(' - ')[0] page = httptools.downloadpage(page_url + '?code=' + code) @@ -122,12 +122,12 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= page = httptools.downloadpage(page_url) data = page.data - logger.info(data) + logger.debug(data) from lib import jsunpack_js2py unpacked = jsunpack_js2py.unpack(scrapertools.find_single_match(data, '") data = jsunpack.unpack(match) @@ -23,9 +23,9 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= media_url = scrapertools.find_single_match(data, '{type:"video/mp4",src:"([^"]+)"}') if not media_url: media_url = scrapertools.find_single_match(data, '"file":"([^"]+)') - logger.info("media_url=" + media_url) + logger.debug("media_url=" + media_url) video_urls = list() video_urls.append([scrapertools.get_filename_from_url(media_url)[-4:] + " [mp4upload]", media_url]) for video_url in video_urls: - logger.info("%s - %s" % (video_url[0], video_url[1])) + logger.debug("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/mydaddy.py b/servers/mydaddy.py index 1f6ec8e4..d09dc5da 100644 --- a/servers/mydaddy.py +++ b/servers/mydaddy.py @@ -21,7 +21,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info() + logger.debug() video_urls = [] data = httptools.downloadpage(page_url).data data = scrapertools.find_single_match(data, 'var srca = \[(.*?)\]') diff --git a/servers/mystream.py b/servers/mystream.py index 5a7bc97a..4da94a91 100644 --- a/servers/mystream.py +++ b/servers/mystream.py @@ -10,7 +10,7 @@ import re def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url) global page_data page_data = data.data @@ -21,7 +21,7 @@ def test_video_exists(page_url): return True, "" def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] global page_data video_url = scrapertools.find_single_match(decode(page_data), r"'src',\s*'([^']+)") diff --git a/servers/myupload.py b/servers/myupload.py index 04eb0724..6b7eaf18 100644 --- a/servers/myupload.py +++ b/servers/myupload.py @@ -23,7 +23,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info() + logger.debug() video_urls = [] data = httptools.downloadpage(page_url).data matches = scrapertools.find_multiple_matches(data, 'tracker: "([^"]+)"') diff --git a/servers/netutv.py b/servers/netutv.py index 8e391f6a..1ae2f344 100755 --- a/servers/netutv.py +++ b/servers/netutv.py @@ -25,7 +25,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) #Deshabilitamos el server hasta nueva orden return False, "[netutv] Servidor deshabilitado" # http://netu.tv/watch_video.php=XX solo contiene una redireccion, ir directamente a http://hqq.tv/player/embed_player.php?vid=XX @@ -37,7 +37,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] if "hash=" in page_url: diff --git a/servers/nowvideo.py b/servers/nowvideo.py index 95271e63..4623f7aa 100644 --- a/servers/nowvideo.py +++ b/servers/nowvideo.py @@ -12,7 +12,7 @@ from platformcode import logger, config headers = [['User-Agent', 'Mozilla/5.0 (Windows NT 6.1; WOW64; rv:54.0) Gecko/20100101 Firefox/54.0']] def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "Not Found" in data or "File was deleted" in data or "The file is being converted" in data or "Please try again later" in data: return False, config.get_localized_string(70293) % "NowVideo" @@ -23,7 +23,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): host = 'http://nowvideo.club' - logger.info("(nowvideo page_url='%s')" % page_url) + logger.debug("(nowvideo page_url='%s')" % page_url) video_urls = [] data = httptools.downloadpage(page_url).data page_url_post = scrapertools.find_single_match(data, '
') @@ -31,7 +31,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= imhuman = '&imhuman=' + scrapertools.find_single_match(data, 'name="imhuman" value="([^"]+)"').replace(" ", "+") post = urllib.urlencode({k: v for k, v in scrapertools.find_multiple_matches(data, 'name="([^"]+)" value="([^"]*)"')}) + imhuman data = httptools.downloadpage(host + page_url_post, post=post).data - logger.info("nowvideo data page_url2 ='%s'" % data) + logger.debug("nowvideo data page_url2 ='%s'" % data) headers.append(['Referer', page_url]) post_data = scrapertools.find_single_match(data,"
\s*") @@ -46,11 +46,11 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= _headers = urllib.urlencode(dict(headers)) for media_url in media_urls: - #logger.info("nowvideo data page_url2 ='%s'" % media_url) + #logger.debug("nowvideo data page_url2 ='%s'" % media_url) video_urls.append([" mp4 [nowvideo] ", media_url + '|' + _headers]) for video_url in media_urls: - logger.info("[nowvideo.py] %s - %s" % (video_url[0], video_url[1])) + logger.debug("[nowvideo.py] %s - %s" % (video_url[0], video_url[1])) return video_urls @@ -60,7 +60,7 @@ def find_videos(data): devuelve = [] patronvideos = r"nowvideo.club/(?:play|videos)?([a-z0-9A-Z]+)" - logger.info("[nowvideo.py] find_videos #" + patronvideos + "#") + logger.debug("[nowvideo.py] find_videos #" + patronvideos + "#") matches = re.compile(patronvideos, re.DOTALL).findall(data) for match in matches: @@ -68,10 +68,10 @@ def find_videos(data): url = 'http://nowvideo.club/%s' % match if url not in encontrados: - logger.info(" url=" + url) + logger.debug(" url=" + url) devuelve.append([titulo, url, 'nowvideo']) encontrados.add(url) else: - logger.info(" url duplicada=" + url) + logger.debug(" url duplicada=" + url) return devuelve diff --git a/servers/okru.py b/servers/okru.py index f8fb8f83..c18c422c 100644 --- a/servers/okru.py +++ b/servers/okru.py @@ -8,7 +8,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "copyrightsRestricted" in data or "COPYRIGHTS_RESTRICTED" in data: @@ -20,7 +20,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] data = httptools.downloadpage(page_url).data diff --git a/servers/onefichier.py b/servers/onefichier.py index 971d954e..54c678e6 100644 --- a/servers/onefichier.py +++ b/servers/onefichier.py @@ -21,25 +21,25 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) if config.get_setting("premium", server="onefichier"): user = config.get_setting("user", server="onefichier") password = config.get_setting("password", server="onefichier") url = "https://1fichier.com/login.pl" - logger.info("url=" + url) + logger.debug("url=" + url) post_parameters = {"mail": user, "pass": password, "lt": "on", "purge": "on", "valider": "Send"} post = urllib.urlencode(post_parameters) - logger.info("post=" + post) + logger.debug("post=" + post) data = httptools.downloadpage(url, post=post).data - # logger.info("data="+data) + # logger.debug("data="+data) cookies = config.get_cookie_data() - logger.info("cookies=" + cookies) + logger.debug("cookies=" + cookies) # 1fichier.com TRUE / FALSE 1443553315 SID imC3q8MQ7cARw5tkXeWvKyrH493rR=1yvrjhxDAA0T0iEmqRfNF9GXwjrwPHssAQ sid_cookie_value = scrapertools.find_single_match(cookies, "1fichier.com.*?SID\s+([A-Za-z0-9\+\=]+)") - logger.info("sid_cookie_value=" + sid_cookie_value) + logger.debug("sid_cookie_value=" + sid_cookie_value) # .1fichier.com TRUE / FALSE 1443553315 SID imC3q8MQ7cARw5tkXeWvKyrH493rR=1yvrjhxDAA0T0iEmqRfNF9GXwjrwPHssAQ cookie = urllib.urlencode({"SID": sid_cookie_value}) @@ -50,16 +50,16 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= 'Mozilla/5.0 (Macintosh; U; Intel Mac OS X 10.6; es-ES; rv:1.9.2.12) Gecko/20101026 Firefox/3.6.12']) headers.append(['Cookie', cookie]) filename = scrapertools.get_header_from_response(page_url, header_to_get="Content-Disposition") - logger.info("filename=" + filename) + logger.debug("filename=" + filename) # Construye la URL final para Kodi location = page_url + "|Cookie=" + cookie - logger.info("location=" + location) + logger.debug("location=" + location) video_urls = [] video_urls.append([filename[-4:] + " (Premium) [1fichier]", location]) for video_url in video_urls: - logger.info("%s - %s" % (video_url[0], video_url[1])) + logger.debug("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/onlystream.py b/servers/onlystream.py index 2deef1a7..214f4e12 100644 --- a/servers/onlystream.py +++ b/servers/onlystream.py @@ -6,7 +6,7 @@ from platformcode import config, logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url).data @@ -17,8 +17,8 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) global data - # logger.info(data) + # logger.debug(data) video_urls = support.get_jwplayer_mediaurl(data, 'Onlystream') return video_urls diff --git a/servers/rapidgator.py b/servers/rapidgator.py index ed25c540..fdc262e3 100644 --- a/servers/rapidgator.py +++ b/servers/rapidgator.py @@ -8,6 +8,6 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] return video_urls diff --git a/servers/rcdnme.py b/servers/rcdnme.py index d7fc0b2c..fbb09582 100644 --- a/servers/rcdnme.py +++ b/servers/rcdnme.py @@ -10,7 +10,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url) if "Object not found" in data.data or "longer exists on our servers" in data.data: @@ -21,7 +21,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "p,a,c,k,e,d" in data: @@ -44,6 +44,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= except: pass for video_url in video_urls: - logger.info(" %s - %s" % (video_url[0], video_url[1])) + logger.debug(" %s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/rutube.py b/servers/rutube.py index e6414be7..aa153868 100644 --- a/servers/rutube.py +++ b/servers/rutube.py @@ -24,13 +24,13 @@ from core import jsontools def get_source(url): - logger.info() + logger.debug() data = httptools.downloadpage(url).data data = re.sub(r'\n|\r|\t| |
|\s{2,}', "", data) return data def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = get_source(page_url) if "File was deleted" in data or "File Not Found" in data: @@ -40,7 +40,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] referer = '' diff --git a/servers/samaup.py b/servers/samaup.py index cc6285c0..5d168a9c 100644 --- a/servers/samaup.py +++ b/servers/samaup.py @@ -10,7 +10,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url).data if "Not Found" in data or "File was deleted" in data: @@ -19,7 +19,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] ext = 'mp4' diff --git a/servers/sendvid.py b/servers/sendvid.py index ed11f426..7cdd94b4 100755 --- a/servers/sendvid.py +++ b/servers/sendvid.py @@ -9,7 +9,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] data = scrapertools.httptools.downloadpage(page_url).data media_url = scrapertools.find_single_match(data, 'var\s+video_source\s+\=\s+"([^"]+)"') @@ -24,5 +24,5 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= else: video_urls.append([scrapertools.get_filename_from_url(media_url)[-4:] + " [sendvid]", media_url]) for video_url in video_urls: - logger.info("%s - %s" % (video_url[0], video_url[1])) + logger.debug("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/speedvideo.py b/servers/speedvideo.py index c9a2e42e..4f4d50a2 100644 --- a/servers/speedvideo.py +++ b/servers/speedvideo.py @@ -5,7 +5,7 @@ from core import httptools, scrapertools from platformcode import config, logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data @@ -15,22 +15,22 @@ def test_video_exists(page_url): return True, "" def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] quality ={'MOBILE':1, 'NORMAL':2, 'HD':3} data = httptools.downloadpage(page_url).data - logger.info('SPEEDVIDEO DATA '+ data) + logger.debug('SPEEDVIDEO DATA '+ data) media_urls = scrapertools.find_multiple_matches(data, r"file:[^']'([^']+)',\s*label:[^\"]\"([^\"]+)\"") - logger.info("speed video - media urls: %s " % media_urls) + logger.debug("speed video - media urls: %s " % media_urls) for media_url, label in media_urls: media_url = httptools.downloadpage(media_url, only_headers=True, follow_redirects=False).headers.get("location", "") if media_url: video_urls.append([media_url.split('.')[-1] + ' - ' + label + ' - ' + ' [Speedvideo]', media_url]) - logger.info("speed video - media urls: %s " % video_urls) + logger.debug("speed video - media urls: %s " % video_urls) return sorted(video_urls, key=lambda x: quality[x[0].split(' - ')[1]]) diff --git a/servers/streamtape.py b/servers/streamtape.py index a2b4b4a2..05fbb07b 100644 --- a/servers/streamtape.py +++ b/servers/streamtape.py @@ -11,7 +11,7 @@ if sys.version_info[0] >= 3: PY3 = True; unicode = str; unichr = chr; long = int def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) global data referer = {"Referer": page_url} @@ -25,7 +25,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] possible_url = scrapertools.find_single_match(data, 'innerHTML = "([^"]+)') diff --git a/servers/streamz.py b/servers/streamz.py index 5540558d..ac0fb568 100644 --- a/servers/streamz.py +++ b/servers/streamz.py @@ -8,7 +8,7 @@ from lib import jsunpack def test_video_exists(page_url): global data - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "File not found, sorry!" in data: @@ -17,7 +17,7 @@ def test_video_exists(page_url): def get_video_url(page_url, video_password): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] from core.support import match matches = match(data, patron=r'(eval\(function\(p,a,c,k,e,d\).*?)\s+').matches diff --git a/servers/supervideo.py b/servers/supervideo.py index 1665152d..2ab9ae1a 100644 --- a/servers/supervideo.py +++ b/servers/supervideo.py @@ -9,7 +9,7 @@ from platformcode import config, logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url, cookies=False).data if 'File is no longer available as it expired or has been deleted' in data: @@ -19,7 +19,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] # data = httptools.downloadpage(page_url).data global data diff --git a/servers/thevid.py b/servers/thevid.py index e061856b..13ee9582 100644 --- a/servers/thevid.py +++ b/servers/thevid.py @@ -8,7 +8,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "Video not found..." in data or "Video removed due to copyright" in data: return False, config.get_localized_string(70292) % "Thevid" @@ -31,5 +31,5 @@ def get_video_url(page_url, user="", password="", video_password=""): continue video = "https:" + video video_urls.append(["mp4 [Thevid]", video]) - logger.info("Url: %s" % videos) + logger.debug("Url: %s" % videos) return video_urls diff --git a/servers/thevideobee.py b/servers/thevideobee.py index f319dcfe..001d3471 100644 --- a/servers/thevideobee.py +++ b/servers/thevideobee.py @@ -9,7 +9,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "no longer exists" in data or "to copyright issues" in data: return False, config.get_localized_string(70449) % "thevideobee" @@ -17,7 +17,7 @@ def test_video_exists(page_url): def get_video_url(page_url, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data video_urls = [] videourl = scrapertools.find_single_match(data, 'src: "([^"]+)') diff --git a/servers/turbobit.py b/servers/turbobit.py index d370f03f..9f254dc9 100644 --- a/servers/turbobit.py +++ b/servers/turbobit.py @@ -4,6 +4,6 @@ from platformcode import logger def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] return video_urls diff --git a/servers/turbovid.py b/servers/turbovid.py index 4bb6b993..d1f1e19c 100644 --- a/servers/turbovid.py +++ b/servers/turbovid.py @@ -12,7 +12,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "Not Found" in data or "File Does not Exist" in data: return False, config.get_localized_string(70449) % "Turbovid" @@ -21,7 +21,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password="", server='Turbovid'): - logger.info("(turbovid page_url='%s')" % page_url) + logger.debug("(turbovid page_url='%s')" % page_url) video_urls = [] data = httptools.downloadpage(page_url).data data = data.replace('"', "'") @@ -31,6 +31,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= time.sleep(6) data = httptools.downloadpage(page_url_post, post=post).data - logger.info("(data page_url='%s')" % data) + logger.debug("(data page_url='%s')" % data) video_urls = support.get_jwplayer_mediaurl(data, 'Turbovid') return video_urls diff --git a/servers/tusfiles.py b/servers/tusfiles.py index 18cec1ee..6390c208 100644 --- a/servers/tusfiles.py +++ b/servers/tusfiles.py @@ -9,7 +9,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "no longer exists" in data or "to copyright issues" in data: return False, config.get_localized_string(70449) % "tusfiles" @@ -17,7 +17,7 @@ def test_video_exists(page_url): def get_video_url(page_url, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data video_urls = [] videourl = scrapertools.find_single_match(data, 'source src="([^"]+)') diff --git a/servers/uploadedto.py b/servers/uploadedto.py index 72b83510..37529ab7 100755 --- a/servers/uploadedto.py +++ b/servers/uploadedto.py @@ -6,7 +6,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) real_url = page_url.replace("uploaded.to", "uploaded.net") code = httptools.downloadpage(real_url, only_headers=True).code @@ -16,36 +16,36 @@ def test_video_exists(page_url): return True, "" def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] if premium: #Si no hay almacenada una cookie activa, hacemos login if check_cookie("uploaded.net", "login") != True: # Login para conseguir la cookie - logger.info("-------------------------------------------") - logger.info("login") - logger.info("-------------------------------------------") + logger.debug("-------------------------------------------") + logger.debug("login") + logger.debug("-------------------------------------------") login_url = "http://uploaded.net/io/login" post = "id=" + user + "&pw=" + password setcookie = httptools.downloadpage(login_url, post=post, follow_redirects=False, only_headers=True).headers.get("set-cookie", "") - logger.info("-------------------------------------------") - logger.info("obtiene la url") - logger.info("-------------------------------------------") + logger.debug("-------------------------------------------") + logger.debug("obtiene la url") + logger.debug("-------------------------------------------") location = httptools.downloadpage(page_url, follow_redirects=False, only_headers=True).headers.get("location", "") - logger.info("location=" + location) + logger.debug("location=" + location) #fix descarga no directa if location == "": data = httptools.downloadpage(page_url).data - #logger.info("data: %s" % data) + #logger.debug("data: %s" % data) if "

Premium Download

" in data: location = scrapertools.find_single_match(data, 'WE ARE SORRY" in data or '404 Not Found' in data: diff --git a/servers/uptobox.py b/servers/uptobox.py index 52cc08aa..cda39d75 100755 --- a/servers/uptobox.py +++ b/servers/uptobox.py @@ -19,7 +19,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data @@ -37,7 +37,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) # Si el enlace es directo de upstream if "uptobox" not in page_url: data = httptools.downloadpage(page_url).data @@ -60,7 +60,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls = uptobox(page_url, data) for video_url in video_urls: - logger.info("%s - %s" % (video_url[0], video_url[1])) + logger.debug("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/upvid.py b/servers/upvid.py index 4523d5dc..e1de4583 100644 --- a/servers/upvid.py +++ b/servers/upvid.py @@ -13,7 +13,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url) if data.code == 404: return False, config.get_localized_string(70449) % "upvid" @@ -23,7 +23,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium = False, user = "", password = "", video_password = ""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] headers = {'referer': page_url} for i in range(0, 3): diff --git a/servers/uqload.py b/servers/uqload.py index 3c254960..46ec18ce 100644 --- a/servers/uqload.py +++ b/servers/uqload.py @@ -10,7 +10,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url) @@ -21,7 +21,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] data = httptools.downloadpage(page_url).data diff --git a/servers/userscloud.py b/servers/userscloud.py index 8988cdb9..d1a85478 100644 --- a/servers/userscloud.py +++ b/servers/userscloud.py @@ -7,7 +7,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) response = httptools.downloadpage(page_url) @@ -18,7 +18,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] unpacked = "" data = httptools.downloadpage(page_url).data @@ -37,6 +37,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls.append(["%s [userscloud]" % ext, media_url]) for video_url in video_urls: - logger.info("%s - %s" % (video_url[0], video_url[1])) + logger.debug("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/vevio.py b/servers/vevio.py index f0c1a6c0..e244e0a0 100644 --- a/servers/vevio.py +++ b/servers/vevio.py @@ -18,7 +18,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "File was deleted" in data or "Page Cannot Be Found" in data or "Video not found" in data: return False, config.get_localized_string(70449) % "vevio" @@ -26,7 +26,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] post = {} post = urllib.urlencode(post) diff --git a/servers/vidcloud.py b/servers/vidcloud.py index 4a28eb22..137e2950 100644 --- a/servers/vidcloud.py +++ b/servers/vidcloud.py @@ -13,7 +13,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "We're Sorry" in data: return False, config.get_localized_string(70292) % "Vidcloud" @@ -22,7 +22,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] @@ -56,6 +56,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls.append(["%s [Vidcloud" % ext, media_url]) for video_url in video_urls: - logger.info("%s - %s" % (video_url[0], video_url[1])) + logger.debug("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/videobin.py b/servers/videobin.py index c2d0125a..fc07160a 100644 --- a/servers/videobin.py +++ b/servers/videobin.py @@ -11,7 +11,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "borrado" in data or "Deleted" in data: return False, config.get_localized_string(70449) % "videobin" @@ -19,7 +19,7 @@ def test_video_exists(page_url): return True, "" def get_video_url(page_url, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] data = httptools.downloadpage(page_url).data bloque = scrapertools.find_single_match(data, 'sources:.\[.*?]') diff --git a/servers/videomega.py b/servers/videomega.py index de473b1d..c78b1bc1 100644 --- a/servers/videomega.py +++ b/servers/videomega.py @@ -5,7 +5,7 @@ from platformcode import logger def get_video_url(page_url, video_password): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] data = httptools.downloadpage(page_url).data m= scrapertools.find_single_match(data, '<link href="(Br74.*?==.css)"') diff --git a/servers/vidfast.py b/servers/vidfast.py index 382cb0bb..9ffb07b6 100644 --- a/servers/vidfast.py +++ b/servers/vidfast.py @@ -9,7 +9,7 @@ from platformcode import logger video_urls = [] def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) subtitles = "" response = httptools.downloadpage(page_url) @@ -21,9 +21,9 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] - logger.info("Intel11 %s" %data) + logger.debug("Intel11 %s" %data) media_url = scrapertools.find_single_match(data, 'file:"([^"]+)') if media_url: ext = media_url[-4:] diff --git a/servers/vidlox.py b/servers/vidlox.py index 7cbb684e..7c5336af 100644 --- a/servers/vidlox.py +++ b/servers/vidlox.py @@ -11,7 +11,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url).data if "borrado" in data or "Deleted" in data: @@ -21,7 +21,7 @@ def test_video_exists(page_url): def get_video_url(page_url, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] bloque = scrapertools.find_single_match(data, 'sources:.\[.*?]') diff --git a/servers/vidmoly.py b/servers/vidmoly.py index 6252edff..7332b2cd 100644 --- a/servers/vidmoly.py +++ b/servers/vidmoly.py @@ -5,7 +5,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) global data resp = httptools.downloadpage(page_url) data = resp.data @@ -15,7 +15,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) global data video_urls = support.get_jwplayer_mediaurl(data, 'Vidmoly') diff --git a/servers/vidoza.py b/servers/vidoza.py index c7a39de0..04743d0b 100644 --- a/servers/vidoza.py +++ b/servers/vidoza.py @@ -8,7 +8,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url).data if "Page not found" in data or "File was deleted" in data: @@ -20,7 +20,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) global data video_urls = [] diff --git a/servers/vidtodo.py b/servers/vidtodo.py index ba08b6b9..f438dbe5 100755 --- a/servers/vidtodo.py +++ b/servers/vidtodo.py @@ -8,7 +8,7 @@ from platformcode import logger id_server = "vidtodo" response = "" def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) global response response = httptools.downloadpage(page_url) if not response.success or "Not Found" in response.data: @@ -19,7 +19,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] data = response.data packed_data = scrapertools.find_single_match(data, "javascript'>(eval.*?)</script>") @@ -41,5 +41,5 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls.append([".mp4 [%s] %s" % (id_server, inf), mp4 % h]) video_urls.append(["RTMP [%s] %s" % (id_server, inf), "%s playpath=%s" % (rtmp, playpath)]) for video_url in video_urls: - logger.info("video_url: %s - %s" % (video_url[0], video_url[1])) + logger.debug("video_url: %s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/vidtome.py b/servers/vidtome.py index 9f1eeafc..bf5db104 100644 --- a/servers/vidtome.py +++ b/servers/vidtome.py @@ -6,7 +6,7 @@ from lib import jsunpack def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url).data if "Not Found" in data or "File Does not Exist" in data: @@ -15,7 +15,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) global data video_urls = [] code = scrapertools.find_single_match(data, 'name="code" value="([^"]+)') diff --git a/servers/vidup.py b/servers/vidup.py index c300d778..b3ff4939 100755 --- a/servers/vidup.py +++ b/servers/vidup.py @@ -20,7 +20,7 @@ from platformcode import logger def test_video_exists(page_url): return False, "[Vidup] Servidor Deshabilitado" - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) page = httptools.downloadpage(page_url) url = page.url if "Not Found" in page.data or "/404" in url: @@ -29,7 +29,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] post= {} post = urllib.urlencode(post) diff --git a/servers/vimeo.py b/servers/vimeo.py index 5308551f..b3f3077f 100644 --- a/servers/vimeo.py +++ b/servers/vimeo.py @@ -5,7 +5,7 @@ from core import scrapertools from platformcode import logger, config headers = [['User-Agent', 'Mozilla/5.0']] def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) global data if "|" in page_url: @@ -23,7 +23,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] global data @@ -38,6 +38,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls.sort(key=lambda x: x[2]) for video_url in video_urls: video_url[2] = 0 - logger.info("%s - %s" % (video_url[0], video_url[1])) + logger.debug("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/vimpleru.py b/servers/vimpleru.py index dc21c4aa..f06848df 100644 --- a/servers/vimpleru.py +++ b/servers/vimpleru.py @@ -7,7 +7,7 @@ from platformcode import config, logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if '"title":"Video Not Found"' in data: return False, config.get_localized_string(70449) % "Vimple" @@ -16,7 +16,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url=%s)" % page_url) + logger.debug("(page_url=%s)" % page_url) data = httptools.downloadpage(page_url).data @@ -34,6 +34,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls.append([scrapertools.get_filename_from_url(media_url)[-4:] + " [vimple.ru]", media_url]) for video_url in video_urls: - logger.info("%s - %s" % (video_url[0], video_url[1])) + logger.debug("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/vivo.py b/servers/vivo.py index 4c7b73a3..115faaf1 100644 --- a/servers/vivo.py +++ b/servers/vivo.py @@ -9,7 +9,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url) if data.code == 404: return False, config.get_localized_string(70449) % "Vivo" @@ -17,7 +17,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] data = httptools.downloadpage(page_url).data enc_data = scrapertools.find_single_match(data, 'data-stream="([^"]+)') diff --git a/servers/vk.py b/servers/vk.py index 0468cf8a..5fdbee13 100755 --- a/servers/vk.py +++ b/servers/vk.py @@ -17,7 +17,7 @@ from platformcode import config, logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) if not login(): return False, "Falta Ingresar/Actualizar las credenciales en el servidor vk. Configuracion - Preferencias - Ajustes de servidores - Configuración del servidor vk" data = httptools.downloadpage(page_url).data @@ -28,7 +28,7 @@ def test_video_exists(page_url): # Returns an array of possible video url's from the page_url def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] data = httptools.downloadpage(page_url).data matches = scrapertools.find_multiple_matches(data, '<source src="([^"]+)" type="video/(\w+)') @@ -37,7 +37,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls.append([calidad + "p ." + ext + " [vk]", media_url]) video_urls.sort(key=lambda it: int(it[0].split("p ", 1)[0])) for video_url in video_urls: - logger.info("%s - %s" % (video_url[0], video_url[1])) + logger.debug("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/vshare.py b/servers/vshare.py index 20f1a000..a07425ba 100644 --- a/servers/vshare.py +++ b/servers/vshare.py @@ -9,7 +9,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) response = httptools.downloadpage(page_url) if response.code != 200 or "No longer available!" in response.data: return False, config.get_localized_string(70449) % "vshare" @@ -18,7 +18,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url = " + page_url) + logger.debug("url = " + page_url) headers = {"Referer":page_url} data = httptools.downloadpage(page_url, headers=headers).data flowplayer = re.search("url: [\"']([^\"']+)", data) diff --git a/servers/vudeo.py b/servers/vudeo.py index 4768e8ea..c8ad45f3 100644 --- a/servers/vudeo.py +++ b/servers/vudeo.py @@ -6,7 +6,7 @@ data = "" def test_video_exists(page_url): global data - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) response = httptools.downloadpage(page_url) if response.code == 404: @@ -18,5 +18,5 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): global data - logger.info("url=" + page_url) + logger.debug("url=" + page_url) return support.get_jwplayer_mediaurl(data, 'Vudeo') diff --git a/servers/vupplayer.py b/servers/vupplayer.py index 2c8da556..b1eaf15e 100644 --- a/servers/vupplayer.py +++ b/servers/vupplayer.py @@ -5,7 +5,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) page = httptools.downloadpage(page_url) global data data = page.data @@ -15,7 +15,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] global data patron = r'sources:\s*\[\{src:\s*"([^"]+)"' diff --git a/servers/vvvvid.py b/servers/vvvvid.py index 53f0568c..50635aa1 100644 --- a/servers/vvvvid.py +++ b/servers/vvvvid.py @@ -19,7 +19,7 @@ payload = {'conn_id': conn_id} def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "Not Found" in data or "File was deleted" in data: return False, config.get_localized_string(70449) % "VVVVID" @@ -45,7 +45,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= # Getting info from Site json_url = "https://www.vvvvid.it/vvvvid/ondemand/" + show_id + '/season/' +season_id + '/' json_file = current_session.get(json_url, headers=headers, params=payload).json() - logger.info(json_file['data']) + logger.debug(json_file['data']) # Search for the correct episode for episode in json_file['data']: diff --git a/servers/watchanimestream.py b/servers/watchanimestream.py index 3df92201..54093ff3 100644 --- a/servers/watchanimestream.py +++ b/servers/watchanimestream.py @@ -5,7 +5,7 @@ from platformcode import logger def get_video_url(page_url, video_password): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] url = page_url.replace("/v/", "/api/source/") post = "r=&d=watchanimestream.net" diff --git a/servers/watchvideo.py b/servers/watchvideo.py index 81ce9abe..b7db7671 100644 --- a/servers/watchvideo.py +++ b/servers/watchvideo.py @@ -7,7 +7,7 @@ from platformcode import logger, config def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url).data if "Not Found" in data or "File was deleted" in data: @@ -16,7 +16,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("url=" + page_url) + logger.debug("url=" + page_url) video_urls = [] media_urls = scrapertools.find_multiple_matches(data, 'file:"([^"]+)"') if not media_urls: @@ -34,5 +34,5 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls.append(["%s [watchvideo]" % (ext), media_url]) video_urls.reverse() for video_url in video_urls: - logger.info("%s - %s" % (video_url[0], video_url[1])) + logger.debug("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/wstream.py b/servers/wstream.py index 57643a51..c0eb6098 100644 --- a/servers/wstream.py +++ b/servers/wstream.py @@ -19,7 +19,7 @@ def test_video_exists(page_url): headers = [['User-Agent', 'Mozilla/5.0 (Windows NT 6.1; WOW64; rv:54.0) Gecko/20100101 Firefox/54.0'], ['Host', scrapertools.get_domain_from_url(page_url)]] - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) if 'wstream' in page_url: resp = httptools.downloadpage(page_url.replace(headers[1][1], real_host), headers=headers, verify=False) else: @@ -31,7 +31,7 @@ def test_video_exists(page_url): page_url = resp.url.replace(headers[1][1], real_host) if '/streaming.php' in page_url in page_url: code = httptools.downloadpage(page_url, headers=headers, follow_redirects=False, only_headers=True, verify=False).headers['location'].split('/')[-1].replace('.html', '') - # logger.info('WCODE=' + code) + # logger.debug('WCODE=' + code) page_url = 'https://' + real_host + '/video.php?file_code=' + code data = httptools.downloadpage(page_url, headers=headers, follow_redirects=True, verify=False).data @@ -45,7 +45,7 @@ def test_video_exists(page_url): page_url = 'https://' + real_host + '/video.php?file_code=' + scrapertools.find_single_match(dec, "src='([^']+)").split('/')[-1].replace('.html','') headers = [['User-Agent', 'Mozilla/5.0 (Windows NT 6.1; WOW64; rv:54.0) Gecko/20100101 Firefox/54.0'],['Host', 'wstream.video']] new_data = httptools.downloadpage(page_url, headers=headers, follow_redirects=True, verify=False).data - logger.info('NEW DATA: \n' + new_data) + logger.debug('NEW DATA: \n' + new_data) if new_data: data = new_data @@ -93,7 +93,7 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= except: pass - logger.info("[Wstream] url=" + page_url) + logger.debug("[Wstream] url=" + page_url) video_urls = [] global data, real_url, headers diff --git a/servers/youdbox.py b/servers/youdbox.py index a58912b4..a371e7d3 100644 --- a/servers/youdbox.py +++ b/servers/youdbox.py @@ -5,7 +5,7 @@ from platformcode import logger def get_video_url(page_url, video_password): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] data = httptools.downloadpage(page_url).data url = scrapertools.find_single_match(data, '<source src="([^"]+)"') diff --git a/servers/yourupload.py b/servers/yourupload.py index 7625c0a3..6959c681 100755 --- a/servers/yourupload.py +++ b/servers/yourupload.py @@ -6,7 +6,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) global data data = httptools.downloadpage(page_url).data @@ -17,7 +17,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] referer = {'Referer': page_url} diff --git a/servers/youtube.py b/servers/youtube.py index 2ad0a8b8..be8c2934 100644 --- a/servers/youtube.py +++ b/servers/youtube.py @@ -6,7 +6,7 @@ from platformcode import config, logger, platformtools name = 'plugin.video.youtube' def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data @@ -18,12 +18,12 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): import xbmc from xbmcaddon import Addon - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] if not page_url.startswith("http"): page_url = "http://www.youtube.com/watch?v=%s" % page_url - logger.info(" page_url->'%s'" % page_url) + logger.debug(" page_url->'%s'" % page_url) video_id = scrapertools.find_single_match(page_url, '(?:v=|embed/)([A-z0-9_-]{11})') inputstream = platformtools.install_inputstream() diff --git a/servers/youwatch.py b/servers/youwatch.py index 82f461c3..5bc49f55 100644 --- a/servers/youwatch.py +++ b/servers/youwatch.py @@ -6,7 +6,7 @@ from platformcode import logger def test_video_exists(page_url): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data if "File Not Found" in data: return False, config.get_localized_string(70449) % "Youwatch" @@ -20,7 +20,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) data = httptools.downloadpage(page_url).data url_redirect = scrapertools.find_single_match(data, '<iframe src="([^"]+)"') @@ -31,6 +31,6 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= video_urls = [[scrapertools.get_filename_from_url(url)[-4:] + " [youwatch]", video_url]] for video_url in video_urls: - logger.info("%s - %s" % (video_url[0], video_url[1])) + logger.debug("%s - %s" % (video_url[0], video_url[1])) return video_urls diff --git a/servers/zippyshare.py b/servers/zippyshare.py index 65c45a70..aec47843 100755 --- a/servers/zippyshare.py +++ b/servers/zippyshare.py @@ -29,7 +29,7 @@ def test_video_exists(page_url): def get_video_url(page_url, premium=False, user="", password="", video_password=""): - logger.info("(page_url='%s')" % page_url) + logger.debug("(page_url='%s')" % page_url) video_urls = [] data = httptools.downloadpage(page_url).data @@ -43,5 +43,5 @@ def get_video_url(page_url, premium=False, user="", password="", video_password= mediaurl = '%s%s' % (domain, url) extension = "." + mediaurl.split('.')[-1] video_urls.append([extension + " [zippyshare]", mediaurl]) - logger.info("url=%s" %video_urls) + logger.debug("url=%s" %video_urls) return video_urls diff --git a/service.py b/service.py index 36806865..5ab5e931 100644 --- a/service.py +++ b/service.py @@ -25,7 +25,7 @@ from servers import torrent def update(path, p_dialog, i, t, serie, overwrite): - logger.info("Updating " + path) + logger.debug("Updating " + path) insertados_total = 0 nfo_file = xbmc.translatePath(filetools.join(path, 'tvshow.nfo')) @@ -57,8 +57,7 @@ def update(path, p_dialog, i, t, serie, overwrite): p_dialog.update(int(math.ceil((i + 1) * t)), heading, config.get_localized_string(60389) % (serie.contentSerieName, serie.channel.capitalize())) try: pathchannels = filetools.join(config.get_runtime_path(), "channels", serie.channel + '.py') - logger.info("loading channel: " + pathchannels + " " + - serie.channel) + logger.debug("loading channel: " + pathchannels + " " + serie.channel) if serie.library_filter_show: serie.show = serie.library_filter_show.get(serie.channel, serie.contentSerieName) @@ -110,7 +109,7 @@ def update(path, p_dialog, i, t, serie, overwrite): def check_for_update(overwrite=True): - logger.info("Update Series...") + logger.debug("Update Series...") p_dialog = None serie_actualizada = False update_when_finished = False @@ -142,7 +141,7 @@ def check_for_update(overwrite=True): filetools.write(tvshow_file, head_nfo + serie.tojson()) path = filetools.dirname(tvshow_file) - logger.info("serie=" + serie.contentSerieName) + logger.debug("serie=" + serie.contentSerieName) p_dialog.update(int(math.ceil((i + 1) * t)), heading, serie.contentSerieName) #Verificamos el estado del serie.library_playcounts de la Serie por si está incompleto @@ -254,7 +253,7 @@ def check_for_update(overwrite=True): p_dialog.close() else: - logger.info("Not update the video library, it is disabled") + logger.debug("Not update the video library, it is disabled") except Exception as ex: logger.error("An error occurred while updating the series") @@ -284,7 +283,7 @@ def viewmodeMonitor(): if content: defaultMode = int(config.get_setting('view_mode_%s' % content).split(',')[-1]) if currentMode != defaultMode: - logger.info('viewmode changed: ' + currentModeName + '-' + str(currentMode) + ' - content: ' + content) + logger.debug('viewmode changed: ' + currentModeName + '-' + str(currentMode) + ' - content: ' + content) config.set_setting('view_mode_%s' % content, currentModeName + ', ' + str(currentMode)) except: logger.error(traceback.print_exc()) @@ -370,11 +369,11 @@ class AddonMonitor(xbmc.Monitor): self.settings_pre = settings_post def onScreensaverActivated(self): - logger.info('screensaver activated, un-scheduling screen-on jobs') + logger.debug('screensaver activated, un-scheduling screen-on jobs') schedule.clear('screenOn') def onScreensaverDeactivated(self): - logger.info('screensaver deactivated, re-scheduling screen-on jobs') + logger.debug('screensaver deactivated, re-scheduling screen-on jobs') self.scheduleScreenOnJobs() def scheduleUpdater(self): @@ -382,7 +381,7 @@ class AddonMonitor(xbmc.Monitor): updaterCheck() self.updaterPeriod = config.get_setting('addon_update_timer') schedule.every(self.updaterPeriod).hours.do(updaterCheck).tag('updater') - logger.info('scheduled updater every ' + str(self.updaterPeriod) + ' hours') + logger.debug('scheduled updater every ' + str(self.updaterPeriod) + ' hours') def scheduleVideolibrary(self): self.update_setting = config.get_setting("update", "videolibrary") @@ -390,18 +389,18 @@ class AddonMonitor(xbmc.Monitor): if self.update_setting == 2 or self.update_setting == 3: self.update_hour = config.get_setting("everyday_delay", "videolibrary") * 4 schedule.every().day.at(str(self.update_hour).zfill(2) + ':00').do(run_threaded, check_for_update, (False,)).tag('videolibrary') - logger.info('scheduled videolibrary at ' + str(self.update_hour).zfill(2) + ':00') + logger.debug('scheduled videolibrary at ' + str(self.update_hour).zfill(2) + ':00') def scheduleScreenOnJobs(self): schedule.every().second.do(viewmodeMonitor).tag('screenOn') schedule.every().second.do(torrent.elementum_monitor).tag('screenOn') def onDPMSActivated(self): - logger.info('DPMS activated, un-scheduling screen-on jobs') + logger.debug('DPMS activated, un-scheduling screen-on jobs') schedule.clear('screenOn') def onDPMSDeactivated(self): - logger.info('DPMS deactivated, re-scheduling screen-on jobs') + logger.debug('DPMS deactivated, re-scheduling screen-on jobs') self.scheduleScreenOnJobs() diff --git a/specials/downloads.py b/specials/downloads.py index c10ba045..45dd43a4 100644 --- a/specials/downloads.py +++ b/specials/downloads.py @@ -311,7 +311,7 @@ def menu(item): # Show Dialog seleccion = platformtools.dialog_select(config.get_localized_string(30163), opciones) - logger.info('SELECTION: '+ op[seleccion]) + logger.debug('SELECTION: '+ op[seleccion]) # -1 is cancel if seleccion == -1: return diff --git a/specials/favorites.py b/specials/favorites.py index d7c885bc..60634bb5 100644 --- a/specials/favorites.py +++ b/specials/favorites.py @@ -24,7 +24,7 @@ except: def mainlist(item): - logger.info() + logger.debug() itemlist = [] for name, thumb, data in read_favourites(): @@ -81,7 +81,7 @@ def save_favourites(favourites_list): def addFavourite(item): - logger.info() + logger.debug() # logger.debug(item.tostring('\n')) # If you get here through the context menu, you must retrieve the action and channel parameters @@ -100,7 +100,7 @@ def addFavourite(item): def delFavourite(item): - logger.info() + logger.debug() # logger.debug(item.tostring('\n')) if item.from_title: @@ -118,7 +118,7 @@ def delFavourite(item): def renameFavourite(item): - logger.info() + logger.debug() # logger.debug(item.tostring('\n')) # Find the item we want to rename in favorites.xml @@ -137,7 +137,7 @@ def renameFavourite(item): ################################################## # Features to migrate old favorites (.txt) def readbookmark(filepath): - logger.info() + logger.debug() try: import urllib.parse as urllib except ImportError: @@ -233,6 +233,6 @@ try: if config.get_setting("bookmarkpath") != "": check_bookmark(config.get_setting("bookmarkpath")) else: - logger.info("No path to old version favorites") + logger.debug("No path to old version favorites") except: pass diff --git a/specials/filmontv.py b/specials/filmontv.py index 5def299e..0a44c78c 100644 --- a/specials/filmontv.py +++ b/specials/filmontv.py @@ -18,7 +18,7 @@ TIMEOUT_TOTAL = 60 def mainlist(item): - logger.info(" mainlist") + logger.debug(" mainlist") itemlist = [#Item(channel="search", action='discover_list', title=config.get_localized_string(70309), #search_type='list', list_type='movie/now_playing', # thumbnail=get_thumb("now_playing.png")), @@ -91,7 +91,7 @@ def server_config(item): return platformtools.show_channel_settings(channelpath=filetools.join(config.get_runtime_path(), "specials", item.config)) def now_on_misc_film(item): - logger.info("filmontv tvoggi") + logger.debug("filmontv tvoggi") itemlist = [] # Carica la pagina @@ -126,7 +126,7 @@ def now_on_misc_film(item): return itemlist def now_on_misc(item): - logger.info("filmontv tvoggi") + logger.debug("filmontv tvoggi") itemlist = [] # Carica la pagina @@ -161,7 +161,7 @@ def now_on_misc(item): return itemlist def now_on_tv(item): - logger.info("filmontv tvoggi") + logger.debug("filmontv tvoggi") itemlist = [] # Carica la pagina @@ -195,7 +195,7 @@ def now_on_tv(item): return itemlist def primafila(item): - logger.info("filmontv tvoggi") + logger.debug("filmontv tvoggi") itemlist = [] # Carica la pagina diff --git a/specials/help.py b/specials/help.py index 53d9cac8..f90cfe96 100644 --- a/specials/help.py +++ b/specials/help.py @@ -37,7 +37,7 @@ if config.is_xbmc(): def mainlist(item): - logger.info() + logger.debug() itemlist = [] if config.is_xbmc(): diff --git a/specials/kodfavorites.py b/specials/kodfavorites.py index 97274fcb..af732369 100644 --- a/specials/kodfavorites.py +++ b/specials/kodfavorites.py @@ -126,7 +126,7 @@ class KodfavouritesData(object): # ============================ def addFavourite(item): - logger.info() + logger.debug() alfav = KodfavouritesData() # If you get here through the context menu, you must retrieve the action and channel parameters @@ -179,7 +179,7 @@ def addFavourite(item): # ==================== def mainlist(item): - logger.info() + logger.debug() alfav = KodfavouritesData() item.category = get_name_from_filename(os.path.basename(alfav.user_favorites_file)) @@ -216,7 +216,7 @@ def mainlist(item): def mostrar_perfil(item): - logger.info() + logger.debug() alfav = KodfavouritesData() itemlist = [] @@ -289,7 +289,7 @@ def _crea_perfil(alfav): # Profile and link management def crear_perfil(item): - logger.info() + logger.debug() alfav = KodfavouritesData() if not _crea_perfil(alfav): return False @@ -299,7 +299,7 @@ def crear_perfil(item): def editar_perfil_titulo(item): - logger.info() + logger.debug() alfav = KodfavouritesData() if not alfav.user_favorites[item.i_perfil]: return False @@ -316,7 +316,7 @@ def editar_perfil_titulo(item): def eliminar_perfil(item): - logger.info() + logger.debug() alfav = KodfavouritesData() if not alfav.user_favorites[item.i_perfil]: return False @@ -332,7 +332,7 @@ def eliminar_perfil(item): def acciones_enlace(item): - logger.info() + logger.debug() acciones = [config.get_localized_string(70620), config.get_localized_string(70621), config.get_localized_string(70622), config.get_localized_string(70623), config.get_localized_string(70624), config.get_localized_string(70548), config.get_localized_string(70625), @@ -364,7 +364,7 @@ def acciones_enlace(item): def editar_enlace_titulo(item): - logger.info() + logger.debug() alfav = KodfavouritesData() if not alfav.user_favorites[item.i_perfil]: return False @@ -386,7 +386,7 @@ def editar_enlace_titulo(item): def editar_enlace_color(item): - logger.info() + logger.debug() alfav = KodfavouritesData() if not alfav.user_favorites[item.i_perfil]: return False @@ -410,7 +410,7 @@ def editar_enlace_color(item): def editar_enlace_thumbnail(item): - logger.info() + logger.debug() alfav = KodfavouritesData() if not alfav.user_favorites[item.i_perfil]: return False @@ -470,7 +470,7 @@ def editar_enlace_thumbnail(item): def editar_enlace_carpeta(item): - logger.info() + logger.debug() alfav = KodfavouritesData() if not alfav.user_favorites[item.i_perfil]: return False @@ -489,7 +489,7 @@ def editar_enlace_carpeta(item): def editar_enlace_lista(item): - logger.info() + logger.debug() alfav = KodfavouritesData() if not alfav.user_favorites[item.i_perfil]: return False @@ -527,7 +527,7 @@ def editar_enlace_lista(item): def eliminar_enlace(item): - logger.info() + logger.debug() alfav = KodfavouritesData() if not alfav.user_favorites[item.i_perfil]: return False @@ -542,7 +542,7 @@ def eliminar_enlace(item): # Move profiles and links (up, down, top, bottom) def mover_perfil(item): - logger.info() + logger.debug() alfav = KodfavouritesData() alfav.user_favorites = _mover_item(alfav.user_favorites, item.i_perfil, item.direccion) @@ -552,7 +552,7 @@ def mover_perfil(item): return True def mover_enlace(item): - logger.info() + logger.debug() alfav = KodfavouritesData() if not alfav.user_favorites[item.i_perfil]: return False @@ -597,7 +597,7 @@ def _mover_item(lista, i_selected, direccion): # ------------------------------------------ def mainlist_listas(item): - logger.info() + logger.debug() itemlist = [] item.category = 'Listas' @@ -619,7 +619,7 @@ def mainlist_listas(item): def acciones_lista(item): - logger.info() + logger.debug() acciones = [config.get_localized_string(70604), config.get_localized_string(70629), config.get_localized_string(70605), config.get_localized_string(70606), config.get_localized_string(70607)] @@ -641,7 +641,7 @@ def acciones_lista(item): def activar_lista(item): - logger.info() + logger.debug() fullfilename = os.path.join(config.get_data_path(), item.lista) if not os.path.exists(fullfilename): @@ -659,7 +659,7 @@ def activar_lista(item): def renombrar_lista(item): - logger.info() + logger.debug() fullfilename_current = os.path.join(config.get_data_path(), item.lista) if not os.path.exists(fullfilename_current): @@ -695,7 +695,7 @@ def renombrar_lista(item): def eliminar_lista(item): - logger.info() + logger.debug() fullfilename = os.path.join(config.get_data_path(), item.lista) if not os.path.exists(fullfilename): @@ -714,7 +714,7 @@ def eliminar_lista(item): def informacion_lista(item): - logger.info() + logger.debug() fullfilename = os.path.join(config.get_data_path(), item.lista) if not os.path.exists(fullfilename): @@ -741,7 +741,7 @@ def informacion_lista(item): def compartir_lista(item): - logger.info() + logger.debug() fullfilename = os.path.join(config.get_data_path(), item.lista) if not os.path.exists(fullfilename): @@ -796,7 +796,7 @@ def compartir_lista(item): def acciones_nueva_lista(item): - logger.info() + logger.debug() acciones = [config.get_localized_string(70651), config.get_localized_string(70652), @@ -830,7 +830,7 @@ def acciones_nueva_lista(item): def crear_lista(item): - logger.info() + logger.debug() titulo = platformtools.dialog_input(default='', heading=config.get_localized_string(70612)) if titulo is None or titulo == '': @@ -853,7 +853,7 @@ def crear_lista(item): def descargar_lista(item, url): - logger.info() + logger.debug() from core import httptools, scrapertools if 'tinyupload.com/' in url: diff --git a/specials/news.py b/specials/news.py index eecbf8f3..17e2289e 100644 --- a/specials/news.py +++ b/specials/news.py @@ -42,7 +42,7 @@ menu_settings_path = os.path.join(config.get_data_path(), "settings_channels", ' def mainlist(item): - logger.info() + logger.debug() itemlist = [] # list_canales, any_active = get_channels_list() @@ -128,7 +128,7 @@ def set_category_context(item): def get_channels_list(): - logger.info() + logger.debug() ## import web_pdb; web_pdb.set_trace() ## list_canales = {'peliculas': [], '4k': [], 'terror': [], 'infantiles': [], 'series': [], 'anime': [], ## 'castellano': [], 'latino':[], 'italiano':[], 'torrent':[], 'documentales': []} @@ -166,14 +166,14 @@ def get_channels_list(): return list_canales, any_active def set_cache(item): - logger.info() + logger.debug() item.mode = 'set_cache' t = Thread(target=novedades, args=[item]) t.start() #t.join() def get_from_cache(item): - logger.info() + logger.debug() itemlist=[] cache_node = jsontools.get_node_from_file('menu_cache_data.json', 'cached') first=item.last @@ -198,7 +198,7 @@ def get_from_cache(item): return itemlist def add_menu_items(item, itemlist): - logger.info() + logger.debug() menu_icon = get_thumb('menu.png') menu = Item(channel="channelselector", action="getmainlist", viewmode="movie", thumbnail=menu_icon, title='Menu') @@ -215,7 +215,7 @@ def add_menu_items(item, itemlist): return itemlist def novedades(item): - logger.info() + logger.debug() global list_newest threads = [] @@ -231,7 +231,7 @@ def novedades(item): return get_from_cache(item) multithread = config.get_setting("multithread", "news") - logger.info("multithread= " + str(multithread)) + logger.debug("multithread= " + str(multithread)) if not multithread: if platformtools.dialog_yesno(config.get_localized_string(60515), @@ -270,7 +270,7 @@ def novedades(item): # if progreso.iscanceled(): # progreso.close() - # logger.info("Búsqueda cancelada") + # logger.debug("Búsqueda cancelada") # return itemlist # Modo Multi Thread @@ -284,7 +284,7 @@ def novedades(item): # Modo single Thread else: if mode == 'normal': - logger.info("Obteniendo novedades de channel_id=" + channel_id) + logger.debug("Obteniendo novedades de channel_id=" + channel_id) progreso.update(percentage, "", config.get_localized_string(60520) % channel_title) get_newest(channel_id, item.extra) @@ -304,7 +304,7 @@ def novedades(item): logger.debug(mensaje) if progreso.iscanceled(): - logger.info("Busqueda de novedades cancelada") + logger.debug("Busqueda de novedades cancelada") break time.sleep(0.5) @@ -312,7 +312,7 @@ def novedades(item): if mode == 'normal': mensaje = config.get_localized_string(60522) % (len(list_newest), time.time() - start_time) progreso.update(100, mensaje) - logger.info(mensaje) + logger.debug(mensaje) start_time = time.time() # logger.debug(start_time) @@ -345,7 +345,7 @@ def novedades(item): def get_newest(channel_id, categoria): - logger.info("channel_id=" + channel_id + ", categoria=" + categoria) + logger.debug("channel_id=" + channel_id + ", categoria=" + categoria) global list_newest global list_newest_tourl @@ -366,9 +366,9 @@ def get_newest(channel_id, categoria): if not puede: return - logger.info("running channel " + modulo.__name__ + " " + modulo.__file__) + logger.debug("running channel " + modulo.__name__ + " " + modulo.__file__) list_result = modulo.newest(categoria) - logger.info("canal= %s %d resultados" % (channel_id, len(list_result))) + logger.debug("canal= %s %d resultados" % (channel_id, len(list_result))) exist=False if os.path.exists(menu_cache_path): cache_node = jsontools.get_node_from_file('menu_cache_data.json', 'cached') @@ -377,7 +377,7 @@ def get_newest(channel_id, categoria): cache_node = {} # logger.debug('cache node: %s' % cache_node) for item in list_result: - # logger.info("item="+item.tostring()) + # logger.debug("item="+item.tostring()) item.channel = channel_id list_newest.append(item) list_newest_tourl.append(item.tourl()) @@ -532,7 +532,7 @@ def group_by_content(list_result_canal): def show_channels(item): - logger.info() + logger.debug() global channels_id_name channels_id_name = item.extra itemlist = [] diff --git a/specials/search.py b/specials/search.py index b657b8b2..1968dc9b 100644 --- a/specials/search.py +++ b/specials/search.py @@ -36,7 +36,7 @@ def_lang = info_language[config.get_setting("info_language", "videolibrary")] def mainlist(item): - logger.info() + logger.debug() itemlist = [Item(channel=item.channel, title=config.get_localized_string(70276), action='new_search', mode='all', thumbnail=get_thumb("search.png")), Item(channel=item.channel, title=config.get_localized_string(70741) % config.get_localized_string(30122), action='new_search', mode='movie', thumbnail=get_thumb("search_movie.png")), @@ -53,7 +53,7 @@ def mainlist(item): def sub_menu(item): - logger.info() + logger.debug() itemlist = [Item(channel=item.channel, action='genres_menu', title=config.get_localized_string(70306), mode='movie', thumbnail=get_thumb("movie_genre.png")), Item(channel=item.channel, action='years_menu', title=config.get_localized_string(70742), mode='movie', thumbnail=get_thumb("movie_year.png")), @@ -71,7 +71,7 @@ def sub_menu(item): def saved_search(item): - logger.info() + logger.debug() itemlist = list() saved_searches_list = get_saved_searches() @@ -98,7 +98,7 @@ def saved_search(item): def new_search(item): - logger.info() + logger.debug() temp_search_file = config.get_temp_file('temp-search') if filetools.isfile(temp_search_file): @@ -394,14 +394,14 @@ def get_servers(item, module_dict): def get_info(itemlist): - logger.info() + logger.debug() tmdb.set_infoLabels_itemlist(itemlist, True, forced=True) return itemlist def get_channels(item): - logger.info() + logger.debug() channels_list = list() title_list = list() @@ -723,7 +723,7 @@ def discover_list(item): def from_context(item): - logger.info() + logger.debug() select = setting_channel_new(item) @@ -744,7 +744,7 @@ def from_context(item): def set_context(itemlist): - logger.info() + logger.debug() for elem in itemlist: elem.context = [{"title": config.get_localized_string(60412), @@ -761,7 +761,7 @@ def set_context(itemlist): def get_from_temp(item): - logger.info() + logger.debug() n = 30 nTotal = len(item.itemlist) diff --git a/specials/setting.py b/specials/setting.py index cf7b9fa0..cba4ac44 100644 --- a/specials/setting.py +++ b/specials/setting.py @@ -20,7 +20,7 @@ CHANNELNAME = "setting" def menu_channels(item): - logger.info() + logger.debug() itemlist = list() itemlist.append(Item(channel=CHANNELNAME, title=config.get_localized_string(60545), action="conf_tools", folder=False, @@ -55,7 +55,7 @@ def channel_config(item): # def setting_torrent(item): -# logger.info() +# logger.debug() # LIBTORRENT_PATH = config.get_setting("libtorrent_path", server="torrent", default="") # LIBTORRENT_ERROR = config.get_setting("libtorrent_error", server="torrent", default="") @@ -192,7 +192,7 @@ def channel_config(item): # config.set_setting("magnet2torrent", dict_data_saved["magnet2torrent"], server="torrent") def menu_servers(item): - logger.info() + logger.debug() itemlist = list() itemlist.append(Item(channel=CHANNELNAME, title=config.get_localized_string(60550), action="servers_blacklist", folder=False, @@ -221,7 +221,7 @@ def menu_servers(item): for server in sorted(server_list): server_parameters = servertools.get_server_parameters(server) - logger.info(server_parameters) + logger.debug(server_parameters) if server_parameters["has_settings"] and [x for x in server_parameters["settings"] if x["id"] not in ["black_list", "white_list"]]: itemlist.append( Item(channel=CHANNELNAME, title=". " + config.get_localized_string(60553) % server_parameters["name"], @@ -355,7 +355,7 @@ def cb_servers_favorites(server_names, dict_values): i += 1 c = 1 - logger.info(dict_favorites) + logger.debug(dict_favorites) favorites_servers_list = [] while c in dict_favorites: favorites_servers_list.append(dict_favorites[c]) @@ -373,7 +373,7 @@ def settings(item): def submenu_tools(item): - logger.info() + logger.debug() itemlist = list() # Custom tools @@ -414,7 +414,7 @@ def submenu_tools(item): def check_quickfixes(item): - logger.info() + logger.debug() if not config.dev_mode(): from platformcode import updater @@ -425,7 +425,7 @@ def check_quickfixes(item): # def update_quasar(item): -# logger.info() +# logger.debug() # from platformcode import custom_code, platformtools # stat = False @@ -437,7 +437,7 @@ def check_quickfixes(item): def conf_tools(item): - logger.info() + logger.debug() # Enable or disable channels if item.extra == "channels_onoff": @@ -539,14 +539,14 @@ def conf_tools(item): action="", folder=False, thumbnail=channel.thumbnail)) continue - # logger.info(channel.channel + " SALTADO!") + # logger.debug(channel.channel + " SALTADO!") # The json file settings of the channel are loaded file_settings = os.path.join(config.get_data_path(), "settings_channels", channel.channel + "_data.json") dict_settings = {} dict_file = {} if filetools.exists(file_settings): - # logger.info(channel.channel + " Has _data.json file") + # logger.debug(channel.channel + " Has _data.json file") channeljson_exists = True # We get saved settings from ../settings/channel_data.json try: @@ -556,7 +556,7 @@ def conf_tools(item): except EnvironmentError: logger.error("ERROR when reading the file: %s" % file_settings) else: - # logger.info(channel.channel + " No _data.json file") + # logger.debug(channel.channel + " No _data.json file") channeljson_exists = False if channeljson_exists: @@ -576,7 +576,7 @@ def conf_tools(item): # Default settings are loaded list_controls, default_settings = channeltools.get_channel_controls_settings( channel.channel) - # logger.info(channel.title + " | Default: %s" % default_settings) + # logger.debug(channel.title + " | Default: %s" % default_settings) except: import traceback logger.error(channel.title + config.get_localized_string(60570) % traceback.format_exc()) @@ -598,7 +598,7 @@ def conf_tools(item): list_status = config.get_localized_string(60571) else: - # logger.info(channel.channel + " - NO correction needed!") + # logger.debug(channel.channel + " - NO correction needed!") needsfix = False # If the channel status has been set it is added to the list @@ -828,7 +828,7 @@ def restore_tools(item): def report_menu(item): - logger.info('URL: ' + item.url) + logger.debug('URL: ' + item.url) from channelselector import get_thumb diff --git a/specials/trailertools.py b/specials/trailertools.py index cb23706b..e2230931 100644 --- a/specials/trailertools.py +++ b/specials/trailertools.py @@ -43,7 +43,7 @@ else: def buscartrailer(item, trailers=[]): - logger.info() + logger.debug() # List of actions if run from context menu if item.action == "manual_search" and item.contextual: @@ -78,8 +78,8 @@ def buscartrailer(item, trailers=[]): item.year = item.infoLabels['year'] - logger.info("Search: %s" % item.contentTitle) - logger.info("Year: %s" % item.year) + logger.debug("Search: %s" % item.contentTitle) + logger.debug("Year: %s" % item.year) if item.infoLabels['trailer'] and not trailers: url = item.infoLabels['trailer'] if "youtube" in url: @@ -122,7 +122,7 @@ def buscartrailer(item, trailers=[]): def manual_search(item): - logger.info() + logger.debug() texto = platformtools.dialog_input(default=item.contentTitle, heading=config.get_localized_string(30112)) if texto is not None: if item.extra == "mymovies": @@ -134,7 +134,7 @@ def manual_search(item): def tmdb_trailers(item, tipo="movie"): - logger.info() + logger.debug() from core.tmdb import Tmdb itemlist = [] @@ -153,7 +153,7 @@ def tmdb_trailers(item, tipo="movie"): def youtube_search(item): - logger.info() + logger.debug() itemlist = [] title = item.contentTitle if item.extra != "youtube": @@ -192,7 +192,7 @@ def youtube_search(item): def mymovies_search(item): - logger.info() + logger.debug() import json title = item.contentTitle @@ -216,7 +216,7 @@ def mymovies_search(item): def search_links_mymovies(item): - logger.info() + logger.debug() trailer_url = match(item, patron=r'<li class="bottone_playlist"[^>]+><a href="([^"]+)"').match itemlist = [] data = httptools.downloadpage(item.url).data @@ -236,7 +236,7 @@ def search_links_mymovies(item): def filmaffinity_search(item): - logger.info() + logger.debug() if item.filmaffinity: item.url = item.filmaffinity @@ -284,7 +284,7 @@ def filmaffinity_search(item): def search_links_filmaff(item): - logger.info() + logger.debug() itemlist = [] data = httptools.downloadpage(item.url).data diff --git a/specials/tvmoviedb.py b/specials/tvmoviedb.py index a181f044..6041fe42 100644 --- a/specials/tvmoviedb.py +++ b/specials/tvmoviedb.py @@ -28,7 +28,7 @@ default_fan = filetools.join(config.get_runtime_path(), "fanart.jpg") def mainlist(item): - logger.info() + logger.debug() itemlist = [ # TMDB # item.clone(title=typo(config.get_localized_string(70021), 'bold'), action=""), @@ -55,7 +55,7 @@ def configuracion(item): return ret def search_star(item): - logger.info() + logger.debug() itemlist = [] item.type='movie' @@ -97,7 +97,7 @@ def search_(item): def searcing(item): - logger.info() + logger.debug() new_item = Item(title=item.contentTitle, text=item.contentTitle.replace("+", " "), mode=item.contentType, infoLabels=item.infoLabels) @@ -625,7 +625,7 @@ def indices_tmdb(item): def filter(item): - logger.info() + logger.debug() from datetime import datetime list_controls = [] @@ -705,7 +705,7 @@ def filtered(item, values): def musica_movie(item): - logger.info() + logger.debug() itemlist = [] data = match(item).data matches = match(data, patron=r'<td class="left">([^<]+)<br><small>([^<]+)</small>.*?<td>(\d+:\d+).*?<p id="([^"]+)"').matches @@ -729,7 +729,7 @@ def list_imdb(item): url = 'http://www.imdb.com/search/title?' + item.url # data = httptools.downloadpage(url, headers=headers, replace_headers=True).data data = match(url, headers=headers).data - logger.info(data) + logger.debug(data) # data = re.sub(r"\n|\r|\t| ", "", data) # data = re.sub(r"\s{2}", " ", data) @@ -812,7 +812,7 @@ def list_imdb(item): def filter_imdb(item): - logger.info() + logger.debug() from datetime import datetime list_controls = [] @@ -1392,7 +1392,7 @@ def indices_imdb(item): # def filter_fa(item): -# logger.info() +# logger.debug() # from datetime import datetime # list_controls = [] @@ -1494,7 +1494,7 @@ def indices_imdb(item): # def login_fa(): -# logger.info() +# logger.debug() # try: # user = config.get_setting("usuariofa", "tvmoviedb") @@ -1519,7 +1519,7 @@ def indices_imdb(item): # userid = scrapertools.find_single_match(data, 'id-user=(\d+)') # if userid: # config.set_setting("userid", userid, "tvmoviedb") -# logger.info("Login correcto") +# logger.debug("Login correcto") # return True, "" # except: # import traceback @@ -1644,7 +1644,7 @@ def indices_imdb(item): # def votar_fa(item): # # Window to select the vote -# logger.info() +# logger.debug() # list_controls = [] # valores = {} @@ -1921,7 +1921,7 @@ def acciones_trakt(item): ratings = [] try: for i, entry in enumerate(data): - logger.info('ENTRY:',entry) + logger.debug('ENTRY:',entry) if i <= item.pagina: continue # try: entry = entry[item.args] # except: pass @@ -1997,7 +1997,7 @@ def acciones_trakt(item): def order_list(item): - logger.info() + logger.debug() list_controls = [] valores1 = ['rating', 'added', 'title', 'released', 'runtime', 'popularity', 'percentage', 'votes'] @@ -2339,7 +2339,7 @@ def indices_mal(item): matches = match("https://myanimelist.net/anime.php", cookies=False, patronBlock=patronBlock, patron=patron).matches for url, title in matches: genero = title.split(" (", 1)[0] - logger.info(url_base, genero) + logger.debug(url_base, genero) thumbnail = url_base + genero.lower().replace(" ", "%20") if genero in ["Hentai", "Yaoi", "Yuri"] and not adult_mal: continue @@ -2488,7 +2488,7 @@ def detail_staff(item): patron_bio = r'<?<div class="spaceit_pad">(.*?)</td>' bio = match(data, patron=patron_bio).match bio = htmlclean(bio.replace("</div>", "\n")) - logger.info(bio) + logger.debug(bio) infoLabels = {'plot': bio} if not "No voice acting roles" in data: itemlist.append(Item(channel=item.channel, title=typo(config.get_localized_string(70374),'bold bullet'), action="", thumbnail=item.thumbnail, infoLabels=infoLabels)) @@ -2626,7 +2626,7 @@ def info_anidb(item, itemlist, url): def filter_mal(item): - logger.info() + logger.debug() list_controls = [] valores = {} @@ -2704,7 +2704,7 @@ def callback_mal(item, values): def musica_anime(item): # List available anime and songs similar to the anime title - logger.info() + logger.debug() itemlist = [] data = match("http://www.freeanimemusic.org/song_search.php", post=item.post).data @@ -2739,7 +2739,7 @@ def musica_anime(item): def login_mal(from_list=False): - logger.info() + logger.debug() from core import httptools from base64 import b64decode as bdec @@ -2768,7 +2768,7 @@ def login_mal(from_list=False): else: if generic: return False, config.get_localized_string(70393), user - logger.info("Correct login") + logger.debug("Correct login") return True, "", user except: import traceback @@ -2800,7 +2800,7 @@ def cuenta_mal(item): def items_mal(item): # Scraper for personal lists - logger.info() + logger.debug() itemlist = [] data = match(item.url).data diff --git a/specials/videolibrary.py b/specials/videolibrary.py index a46fcd6a..b543b98e 100644 --- a/specials/videolibrary.py +++ b/specials/videolibrary.py @@ -20,7 +20,7 @@ else: def mainlist(item): - logger.info() + logger.debug() itemlist = [Item(channel=item.channel, action="list_movies", title=config.get_localized_string(60509), category=config.get_localized_string(70270), thumbnail=thumb("videolibrary_movie")), @@ -37,7 +37,7 @@ def channel_config(item): def list_movies(item, silent=False): - logger.info() + logger.debug() itemlist = [] movies_path = [] for root, folders, files in filetools.walk(videolibrarytools.MOVIES_PATH): @@ -61,9 +61,7 @@ def list_movies(item, silent=False): def list_tvshows(item): - from time import time - start = time() - logger.info() + logger.debug() itemlist = [] lista = [] tvshows_path = [] @@ -88,7 +86,6 @@ def list_tvshows(item): title=typo(config.get_localized_string(70269), 'bold color kod'), folder=False), Item(channel=item.channel, action="configure_update_videolibrary", thumbnail=item.thumbnail, title=typo(config.get_localized_string(60599), 'bold color kod'), lista=lista, folder=False)] - logger.info('TEMPO= ' + str(time() - start)) return itemlist @@ -188,7 +185,7 @@ def get_results(nfo_path, root, Type, local=False): # Contextual menu: Mark as seen / not seen visto = item.library_playcounts.get(item.contentTitle, 0) item.infoLabels["playcount"] = visto - logger.info('item\n' + str(item)) + logger.debug('item\n' + str(item)) if visto > 0: seen_text = config.get_localized_string(60020) counter = 0 @@ -260,7 +257,7 @@ def configure_update_videolibrary(item): def get_seasons(item): - logger.info() + logger.debug() # logger.debug("item:\n" + item.tostring('\n')) itemlist = [] dict_temp = {} @@ -319,7 +316,7 @@ def get_seasons(item): def get_episodes(item): - logger.info() + logger.debug() # logger.debug("item:\n" + item.tostring('\n')) itemlist = [] @@ -383,7 +380,7 @@ def get_episodes(item): def findvideos(item): from core import autoplay - logger.info() + logger.debug() # logger.debug("item:\n" + item.tostring('\n')) videolibrarytools.check_renumber_options(item) itemlist = [] @@ -535,7 +532,7 @@ def findvideos(item): def play(item): - logger.info() + logger.debug() # logger.debug("item:\n" + item.tostring('\n')) if not item.contentChannel == "local": @@ -575,7 +572,7 @@ def play(item): def update_videolibrary(item=''): - logger.info() + logger.debug() # Update active series by overwriting import service @@ -595,7 +592,7 @@ def update_videolibrary(item=''): def move_videolibrary(current_path, new_path, current_movies_folder, new_movies_folder, current_tvshows_folder, new_tvshows_folder): - logger.info() + logger.debug() backup_current_path = current_path backup_new_path = new_path @@ -668,7 +665,7 @@ def move_videolibrary(current_path, new_path, current_movies_folder, new_movies_ def delete_videolibrary(item): - logger.info() + logger.debug() if not platformtools.dialog_yesno(config.get_localized_string(20000), config.get_localized_string(80037)): return @@ -694,7 +691,7 @@ def delete_videolibrary(item): # context menu methods def update_tvshow(item): - logger.info() + logger.debug() # logger.debug("item:\n" + item.tostring('\n')) heading = config.get_localized_string(60037) @@ -720,11 +717,11 @@ def update_tvshow(item): def add_local_episodes(item): - logger.info() + logger.debug() done, local_episodes_path = videolibrarytools.config_local_episodes_path(item.path, item, silent=True) if done < 0: - logger.info("An issue has occurred while configuring local episodes") + logger.debug("An issue has occurred while configuring local episodes") elif local_episodes_path: nfo_path = filetools.join(item.path, "tvshow.nfo") head_nfo, item_nfo = videolibrarytools.read_nfo(nfo_path) @@ -739,7 +736,7 @@ def add_local_episodes(item): def remove_local_episodes(item): - logger.info() + logger.debug() nfo_path = filetools.join(item.path, "tvshow.nfo") head_nfo, item_nfo = videolibrarytools.read_nfo(nfo_path) @@ -757,7 +754,7 @@ def remove_local_episodes(item): def verify_playcount_series(item, path): - logger.info() + logger.debug() """ This method reviews and repairs the PlayCount of a series that has become out of sync with the actual list of episodes in its folder. Entries for missing episodes, seasons, or series are created with the "not seen" mark. Later it is sent to verify the counters of Seasons and Series @@ -820,7 +817,7 @@ def verify_playcount_series(item, path): def mark_content_as_watched2(item): - logger.info() + logger.debug() # logger.debug("item:\n" + item.tostring('\n')) if filetools.isfile(item.nfo): head_nfo, it = videolibrarytools.read_nfo(item.nfo) @@ -858,7 +855,7 @@ def mark_content_as_watched2(item): def mark_content_as_watched(item): - logger.info() + logger.debug() #logger.debug("item:\n" + item.tostring('\n')) if filetools.exists(item.nfo): @@ -896,7 +893,7 @@ def mark_content_as_watched(item): def mark_season_as_watched(item): - logger.info() + logger.debug() # logger.debug("item:\n" + item.tostring('\n')) # Get dictionary of marked episodes @@ -949,7 +946,7 @@ def mark_season_as_watched(item): def mark_tvshow_as_updatable(item, silent=False): - logger.info() + logger.debug() head_nfo, it = videolibrarytools.read_nfo(item.nfo) it.active = item.active filetools.write(item.nfo, head_nfo + it.tojson()) @@ -1063,7 +1060,7 @@ def delete(item): def check_season_playcount(item, season): - logger.info() + logger.debug() if season: episodios_temporada = 0 @@ -1085,7 +1082,7 @@ def check_season_playcount(item, season): def check_tvshow_playcount(item, season): - logger.info() + logger.debug() if season: temporadas_serie = 0 temporadas_vistas_serie = 0 diff --git a/updatetvshow.py b/updatetvshow.py index 7227ec72..1b39f173 100644 --- a/updatetvshow.py +++ b/updatetvshow.py @@ -49,7 +49,7 @@ def search_paths(Id): def execute_sql(sql): - logger.info() + logger.debug() file_db = "" records = None @@ -69,14 +69,14 @@ def execute_sql(sql): break if file_db: - logger.info("DB file: %s" % file_db) + logger.debug("DB file: %s" % file_db) conn = None try: import sqlite3 conn = sqlite3.connect(file_db) cursor = conn.cursor() - logger.info("Running sql: %s" % sql) + logger.debug("Running sql: %s" % sql) cursor.execute(sql) conn.commit() @@ -86,7 +86,7 @@ def execute_sql(sql): records = [] conn.close() - logger.info("Query executed. Records: %s" % nun_records) + logger.debug("Query executed. Records: %s" % nun_records) except: logger.error("Error executing sql query")