X-Git-Url: https://jfr.im/git/erebus.git/blobdiff_plain/b91b84fabab3a5dbbc50cae53a54dcedaa96d903..3cec5bdc665d310ac628e80b007dd2d71ae7d7bd:/modules/urls.py diff --git a/modules/urls.py b/modules/urls.py index 5a8b4a6..65f9ca1 100644 --- a/modules/urls.py +++ b/modules/urls.py @@ -247,9 +247,28 @@ def _do_request(url, try_aia=False): - the HTTPResponse object, or a string on error. Empty string -> no response. - and a flag indicating whether AIA was used """ - request = urllib2.Request(url, headers={'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/115.0.0.0 Safari/537.36', 'Sec-Ch-Ua': '"Chromium";v="116", "Not)A;Brand";v="24", "Google Chrome";v="116"', 'Sec-Ch-Ua-Mobile': '?0', 'Sec-Ch-Ua-Platform': '"Linux"', 'Sec-Fetch-Dest': 'document', 'Sec-Fetch-Mode': 'navigate', 'Sec-Fetch-Site': 'same-origin', 'Sec-Fetch-User': '?1', 'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,image/apng,*/*;q=0.8,application/signed-exchange;v=b3;q=0.7', 'Accept-Language': 'en-US,en;q=0.9', 'Cache-Control': 'no-cache', 'Pragma': 'no-cache', 'Upgrade-Insecure-Requests': '1'}) + try: + request = urllib2.Request(url, headers={ + 'accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,image/apng,*/*;q=0.8,application/signed-exchange;v=b3;q=0.7', + 'accept-language': 'en-US,en;q=0.9', + 'cache-control': 'max-age=0', + 'sec-ch-ua': '"Chromium";v="122", "Not(A:Brand";v="24", "Google Chrome";v="122"', + 'sec-ch-ua-mobile': '?0', + 'sec-ch-ua-platform': '"Linux"', + 'sec-fetch-dest': 'document', + 'sec-fetch-mode': 'navigate', + 'sec-fetch-site': 'none', + 'sec-fetch-user': '?1', + 'upgrade-insecure-requests': '1', + 'user-agent': 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/122.0.0.0 Safari/537.36' + }) + except ValueError: + return '', False if try_aia: - opener = urllib2.build_opener(urllib2.HTTPSHandler(context=aia_session.ssl_context_from_url(url)), SmartRedirectHandler()) + try: + opener = urllib2.build_opener(urllib2.HTTPSHandler(context=aia_session.ssl_context_from_url(url)), SmartRedirectHandler()) + except aia.AIAError as e: + return 'Request error: %s.%s: %s' % (e.__module__, e.__class__.__name__, e.args[0]), True else: opener = urllib2.build_opener(SmartRedirectHandler()) @@ -317,9 +336,12 @@ def goturl(url): except Exception as e: output.append('Error reading response body: %s %r' % (type(e).__name__, e.args)) else: - if c_len is not None and len(responsebody) != int(c_len): - output.append("[actual %s; Content-Length %s] " % (_humanize_bytes(len(responsebody)), _humanize_bytes(c_len))) - else: + if c_len is not None and len(responsebody) != int(c_len): # did we read a different amount than Content-Length? + if response.read(1): # there's more data, we just aren't reading it + output.append("[read %s; Content-Length %s] " % (_humanize_bytes(len(responsebody)), _humanize_bytes(c_len))) + else: + output.append("[actual %s; Content-Length %s] " % (_humanize_bytes(len(responsebody)), _humanize_bytes(c_len))) + else: # Content-Length = amount read output.append("[%s] " % (_humanize_bytes(len(responsebody)))) try: soup = BeautifulSoup(responsebody, from_encoding=c_charset) @@ -333,12 +355,13 @@ def goturl(url): return ''.join(output) url_regex = ( - re.compile(r'https?://(?:[^/\s.]+\.)+[^/\s.]+(?:/\S+)?'), + re.compile(r'https?://(?:[^/\s.]+\.)+[a-z0-9-]+(?::\d{1,5})?(?:/[^\s\]>)}]+)?', re.I), ) other_regexes = ( (lambda x: '', (re.compile(r"""https?://(?:www\.)?(?:twitter|x)\.com/""", re.I),)), # skip twitter (lambda x: '', (re.compile(r"""https?://(?:www\.)?reddit\.com/""", re.I),)), # skip new-reddit (lambda x: '', (re.compile(r"""https?://jfr\.im/git/""", re.I),)), # skip my gitweb + (lambda x: '', (re.compile(r"""https?://(?:www\.)?wunderground\.com/""", re.I),)), # skip wunderground, they time us out ) regexes = other_regexes + ( (goturl, url_regex),