migliorato resolverdns (con supporto cloudscraper)

This commit is contained in:
marco
2019-12-30 20:59:49 +01:00
parent 7cb35cb36c
commit 47a912b407
3 changed files with 36 additions and 39 deletions

View File

@@ -8,8 +8,6 @@ import sys
import xbmc
from platformcode import config, logger
import ssl
logger.info(ssl.OPENSSL_VERSION)
logger.info("init...")

View File

@@ -1,15 +1,14 @@
# -*- coding: utf-8 -*-
import os
import ssl
import urlparse
import xbmc
from core import jsontools
from lib.requests_toolbelt.adapters import host_header_ssl
from lib import requests, doh
from lib import cloudscraper
from lib import doh
from platformcode import logger, config
import re
try:
import _sqlite3 as sql
except:
@@ -38,41 +37,32 @@ class CustomContext(ssl.SSLContext):
_context=self)
class CipherSuiteAdapter(host_header_ssl.HostHeaderSSLAdapter):
class CipherSuiteAdapter(host_header_ssl.HostHeaderSSLAdapter, cloudscraper.CipherSuiteAdapter):
def __init__(self, hostname, *args, **kwargs):
self.ssl_context = kwargs.pop('ssl_context', None)
self.cipherSuite = kwargs.pop('cipherSuite', None)
self.cipherSuite = kwargs.get('cipherSuite', None)
self.hostname = hostname
if not self.ssl_context:
self.ssl_context = CustomContext(ssl.PROTOCOL_TLS, hostname)
self.ssl_context.set_ciphers(self.cipherSuite)
super(CipherSuiteAdapter, self).__init__(**kwargs)
def init_poolmanager(self, *args, **kwargs):
self.ssl_context = CustomContext(ssl.PROTOCOL_TLS, hostname)
self.ssl_context.set_ciphers(self.cipherSuite)
self.ssl_context.options |= (ssl.OP_NO_SSLv2 | ssl.OP_NO_SSLv3 | ssl.OP_NO_TLSv1 | ssl.OP_NO_TLSv1_1)
kwargs['ssl_context'] = self.ssl_context
return super(CipherSuiteAdapter, self).init_poolmanager(*args, **kwargs)
def proxy_manager_for(self, *args, **kwargs):
kwargs['ssl_context'] = self.ssl_context
return super(CipherSuiteAdapter, self).proxy_manager_for(*args, **kwargs)
cloudscraper.CipherSuiteAdapter.__init__(self, *args, **kwargs)
class session(requests.Session):
def __init__(self):
class session(cloudscraper.CloudScraper):
def __init__(self, *args, **kwargs):
self.conn = sql.connect(db)
self.cur = self.conn.cursor()
super(session, self).__init__()
super(session, self).__init__(*args, **kwargs)
def getIp(self, domain):
import time
t = time.time()
ip = None
try:
self.cur.execute('select ip from dnscache where domain=?', (domain,))
ip = self.cur.fetchall()[0][0]
logger.info('Cache DNS: ' + domain + ' = ' + str(ip))
except:
pass
if not ip: # not cached
@@ -84,8 +74,7 @@ class session(requests.Session):
logger.error('Failed to resolve hostname, fallback to normal dns')
import traceback
logger.error(traceback.print_exc())
ip = [domain]
logger.info('tempo getIP: ' + str(time.time()-t))
ip = domain
return ip
def writeToCache(self, domain, ip):
@@ -105,12 +94,11 @@ class session(requests.Session):
return self.request(method, realUrl, flushedDns=True, **kwargs)
def request(self, method, url, headers=None, flushedDns=False, **kwargs):
import time
t = time.time()
protocol, domain, port, resource = re.match(r'^(http[s]?:\/\/)?([^:\/\s]+)(?::([^\/]*))?([^\s]*)$', url, flags=re.IGNORECASE).groups()
self.mount('https://', CipherSuiteAdapter(domain, cipherSuite="ALL"))
realUrl = url
parse = urlparse.urlparse(url)
domain = headers['Host'] if headers and 'Host' in headers.keys() else parse.netloc
ip = self.getIp(domain)
self.mount('https://', CipherSuiteAdapter(domain, cipherSuite=self.adapters['https://'].cipherSuite))
realUrl = url
if headers:
headers["Host"] = domain
@@ -118,14 +106,22 @@ class session(requests.Session):
headers = {"Host": domain}
ret = None
tryFlush = False
url = protocol + ip + (':' + port if port else '') + resource
parse = list(parse)
parse[1] = ip
url = urlparse.urlunparse(parse)
allow_redirects = kwargs.get('allow_redirects', True)
if 'allow_redirects' in kwargs:
del kwargs['allow_redirects']
try:
ret = super(session, self).request(method, url, headers=headers, allow_redirects=False, **kwargs)
newUrl = ret.headers.get('Location', realUrl)
newUrl = urlparse.urlparse(ret.headers.get('Location', realUrl))
if not newUrl.netloc and not newUrl.scheme:
newUrl = list(newUrl)
newUrl[0] = 'https://'
newUrl[1] = domain
newUrl = urlparse.urlunparse(newUrl)
if allow_redirects:
redirectN = 0
while newUrl != realUrl and redirectN < self.max_redirects:
@@ -141,5 +137,4 @@ class session(requests.Session):
logger.info('Flushing dns cache for ' + domain)
return self.flushDns(method, realUrl, domain, **kwargs)
logger.info('tempo dns: ' + str(time.time()-t))
return ret

View File

@@ -181,8 +181,12 @@ def channel_search(item):
channel_list, channel_titles = get_channels(item)
from lib import cloudscraper
session = cloudscraper.create_scraper()
if config.get_setting('resolver_dns'):
from specials import resolverdns
session = resolverdns.session()
else:
from lib import cloudscraper
session = cloudscraper.create_scraper()
searching += channel_list
searching_titles += channel_titles