Compare commits

...

4 Commits

3 changed files with 94 additions and 93 deletions

View File

@ -36,7 +36,7 @@ try:
) )
except NameError: except NameError:
compat_str, compat_basestring, compat_chr = ( compat_str, compat_basestring, compat_chr = (
str, str, chr str, (str, bytes), chr
) )
# casefold # casefold

View File

@ -1678,9 +1678,7 @@ def random_user_agent():
std_headers = { std_headers = {
'User-Agent': random_user_agent(), 'User-Agent': random_user_agent(),
'Accept-Charset': 'ISO-8859-1,utf-8;q=0.7,*;q=0.7',
'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8', 'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8',
'Accept-Encoding': 'gzip, deflate',
'Accept-Language': 'en-us,en;q=0.5', 'Accept-Language': 'en-us,en;q=0.5',
} }
@ -1826,11 +1824,11 @@ def write_json_file(obj, fn):
if sys.version_info < (3, 0) and sys.platform != 'win32': if sys.version_info < (3, 0) and sys.platform != 'win32':
encoding = get_filesystem_encoding() encoding = get_filesystem_encoding()
# os.path.basename returns a bytes object, but NamedTemporaryFile # os.path.basename returns a bytes object, but NamedTemporaryFile
# will fail if the filename contains non ascii characters unless we # will fail if the filename contains non-ascii characters unless we
# use a unicode object # use a unicode object
path_basename = lambda f: os.path.basename(fn).decode(encoding) path_basename = lambda f: os.path.basename(f).decode(encoding)
# the same for os.path.dirname # the same for os.path.dirname
path_dirname = lambda f: os.path.dirname(fn).decode(encoding) path_dirname = lambda f: os.path.dirname(f).decode(encoding)
else: else:
path_basename = os.path.basename path_basename = os.path.basename
path_dirname = os.path.dirname path_dirname = os.path.dirname
@ -1894,10 +1892,10 @@ else:
return f return f
return None return None
# On python2.6 the xml.etree.ElementTree.Element methods don't support # On python2.6 the xml.etree.ElementTree.Element methods don't support
# the namespace parameter # the namespace parameter
def xpath_with_ns(path, ns_map): def xpath_with_ns(path, ns_map):
components = [c.split(':') for c in path.split('/')] components = [c.split(':') for c in path.split('/')]
replaced = [] replaced = []
@ -1914,7 +1912,7 @@ def xpath_element(node, xpath, name=None, fatal=False, default=NO_DEFAULT):
def _find_xpath(xpath): def _find_xpath(xpath):
return node.find(compat_xpath(xpath)) return node.find(compat_xpath(xpath))
if isinstance(xpath, (str, compat_str)): if isinstance(xpath, compat_basestring):
n = _find_xpath(xpath) n = _find_xpath(xpath)
else: else:
for xp in xpath: for xp in xpath:
@ -2235,7 +2233,7 @@ def _htmlentity_transform(entity_with_semicolon):
def unescapeHTML(s): def unescapeHTML(s):
if s is None: if s is None:
return None return None
assert type(s) == compat_str assert isinstance(s, compat_str)
return re.sub( return re.sub(
r'&([^&;]+;)', lambda m: _htmlentity_transform(m.group(1)), s) r'&([^&;]+;)', lambda m: _htmlentity_transform(m.group(1)), s)
@ -2262,39 +2260,32 @@ def get_subprocess_encoding():
return encoding return encoding
# Jython assumes filenames are Unicode strings though reported as Python 2.x compatible
if sys.version_info < (3, 0) and not sys.platform.startswith('java'):
def encodeFilename(s, for_subprocess=False): def encodeFilename(s, for_subprocess=False):
""" """
@param s The name of the file @param s The name of the file
""" """
assert type(s) == compat_str
# Python 3 has a Unicode API
if sys.version_info >= (3, 0):
return s
# Pass '' directly to use Unicode APIs on Windows 2000 and up # Pass '' directly to use Unicode APIs on Windows 2000 and up
# (Detecting Windows NT 4 is tricky because 'major >= 4' would # (Detecting Windows NT 4 is tricky because 'major >= 4' would
# match Windows 9x series as well. Besides, NT 4 is obsolete.) # match Windows 9x series as well. Besides, NT 4 is obsolete.)
if not for_subprocess and sys.platform == 'win32' and sys.getwindowsversion()[0] >= 5: if (not for_subprocess
and sys.platform == 'win32'
and sys.getwindowsversion()[0] >= 5
and isinstance(s, compat_str)):
return s return s
# Jython assumes filenames are Unicode strings though reported as Python 2.x compatible return _encode_compat_str(s, get_subprocess_encoding(), 'ignore')
if sys.platform.startswith('java'):
return s
return s.encode(get_subprocess_encoding(), 'ignore')
def decodeFilename(b, for_subprocess=False): def decodeFilename(b, for_subprocess=False):
return _decode_compat_str(b, get_subprocess_encoding(), 'ignore')
if sys.version_info >= (3, 0): else:
return b
if not isinstance(b, bytes): # Python 3 has a Unicode API
return b encodeFilename = decodeFilename = lambda *s, **k: s[0]
return b.decode(get_subprocess_encoding(), 'ignore')
def encodeArgument(s): def encodeArgument(s):
@ -2313,11 +2304,7 @@ def decodeArgument(b):
def decodeOption(optval): def decodeOption(optval):
if optval is None: if optval is None:
return optval return optval
if isinstance(optval, bytes): return _decode_compat_str(optval)
optval = optval.decode(preferredencoding())
assert isinstance(optval, compat_str)
return optval
def formatSeconds(secs): def formatSeconds(secs):
@ -2363,7 +2350,7 @@ def make_HTTPS_handler(params, **kwargs):
if sys.version_info < (3, 2): if sys.version_info < (3, 2):
return YoutubeDLHTTPSHandler(params, **kwargs) return YoutubeDLHTTPSHandler(params, **kwargs)
else: # Python < 3.4 else: # Python3 < 3.4
context = ssl.SSLContext(ssl.PROTOCOL_TLSv1) context = ssl.SSLContext(ssl.PROTOCOL_TLSv1)
context.verify_mode = (ssl.CERT_NONE context.verify_mode = (ssl.CERT_NONE
if opts_no_check_certificate if opts_no_check_certificate
@ -2597,7 +2584,7 @@ def handle_youtubedl_headers(headers):
filtered_headers = headers filtered_headers = headers
if 'Youtubedl-no-compression' in filtered_headers: if 'Youtubedl-no-compression' in filtered_headers:
filtered_headers = dict((k, v) for k, v in filtered_headers.items() if k.lower() != 'accept-encoding') filtered_headers = filter_dict(filtered_headers, cndn=lambda k, _: k.lower() != 'accept-encoding')
del filtered_headers['Youtubedl-no-compression'] del filtered_headers['Youtubedl-no-compression']
return filtered_headers return filtered_headers
@ -2735,6 +2722,13 @@ class YoutubeDLHandler(compat_urllib_request.HTTPHandler):
if h.capitalize() not in req.headers: if h.capitalize() not in req.headers:
req.add_header(h, v) req.add_header(h, v)
# Similarly, 'Accept-encoding'
if 'Accept-encoding' not in req.headers:
req.add_header(
'Accept-Encoding', join_nonempty(
'gzip', 'deflate', brotli and 'br', ncompress and 'compress',
delim=', '))
req.headers = handle_youtubedl_headers(req.headers) req.headers = handle_youtubedl_headers(req.headers)
if sys.version_info < (2, 7): if sys.version_info < (2, 7):
@ -2818,8 +2812,7 @@ class YoutubeDLHandler(compat_urllib_request.HTTPHandler):
location_escaped = escape_url(location_fixed) location_escaped = escape_url(location_fixed)
if location != location_escaped: if location != location_escaped:
del resp.headers['Location'] del resp.headers['Location']
# if sys.version_info < (3, 0): if not isinstance(location_escaped, str): # Py 2 case
if not isinstance(location_escaped, str):
location_escaped = location_escaped.encode('utf-8') location_escaped = location_escaped.encode('utf-8')
resp.headers['Location'] = location_escaped resp.headers['Location'] = location_escaped
return resp return resp
@ -3086,8 +3079,7 @@ class YoutubeDLRedirectHandler(compat_urllib_request.HTTPRedirectHandler):
# On python 2 urlh.geturl() may sometimes return redirect URL # On python 2 urlh.geturl() may sometimes return redirect URL
# as a byte string instead of unicode. This workaround forces # as a byte string instead of unicode. This workaround forces
# it to return unicode. # it to return unicode.
if sys.version_info[0] < 3: newurl = _decode_compat_str(newurl)
newurl = compat_str(newurl)
# Be conciliant with URIs containing a space. This is mainly # Be conciliant with URIs containing a space. This is mainly
# redundant with the more complete encoding done in http_error_302(), # redundant with the more complete encoding done in http_error_302(),
@ -3115,9 +3107,7 @@ class YoutubeDLRedirectHandler(compat_urllib_request.HTTPRedirectHandler):
new_data = None new_data = None
remove_headers.extend(['Content-Length', 'Content-Type']) remove_headers.extend(['Content-Length', 'Content-Type'])
# NB: don't use dict comprehension for python 2.6 compatibility new_headers = filter_dict(req.headers, cndn=lambda k, _: k.title() not in remove_headers)
new_headers = dict((k, v) for k, v in req.headers.items()
if k.title() not in remove_headers)
return compat_urllib_request.Request( return compat_urllib_request.Request(
newurl, headers=new_headers, origin_req_host=req.origin_req_host, newurl, headers=new_headers, origin_req_host=req.origin_req_host,
@ -3333,11 +3323,7 @@ class DateRange(object):
def platform_name(): def platform_name():
""" Returns the platform name as a compat_str """ """ Returns the platform name as a compat_str """
res = platform.platform() res = platform.platform()
if isinstance(res, bytes): return _decode_compat_str(res)
res = res.decode(preferredencoding())
assert isinstance(res, compat_str)
return res
def _windows_write_string(s, out): def _windows_write_string(s, out):
@ -3418,7 +3404,7 @@ def _windows_write_string(s, out):
def write_string(s, out=None, encoding=None): def write_string(s, out=None, encoding=None):
if out is None: if out is None:
out = sys.stderr out = sys.stderr
assert type(s) == compat_str assert isinstance(s, compat_str)
if sys.platform == 'win32' and encoding is None and hasattr(out, 'fileno'): if sys.platform == 'win32' and encoding is None and hasattr(out, 'fileno'):
if _windows_write_string(s, out): if _windows_write_string(s, out):
@ -3567,9 +3553,8 @@ def shell_quote(args):
quoted_args = [] quoted_args = []
encoding = get_filesystem_encoding() encoding = get_filesystem_encoding()
for a in args: for a in args:
if isinstance(a, bytes):
# We may get a filename encoded with 'encodeFilename' # We may get a filename encoded with 'encodeFilename'
a = a.decode(encoding) a = _decode_compat_str(a, encoding)
quoted_args.append(compat_shlex_quote(a)) quoted_args.append(compat_shlex_quote(a))
return ' '.join(quoted_args) return ' '.join(quoted_args)
@ -3733,8 +3718,9 @@ def parse_resolution(s):
def parse_bitrate(s): def parse_bitrate(s):
if not isinstance(s, compat_str): s = txt_or_none(s)
return if not s:
return None
mobj = re.search(r'\b(\d+)\s*kbps', s) mobj = re.search(r'\b(\d+)\s*kbps', s)
if mobj: if mobj:
return int(mobj.group(1)) return int(mobj.group(1))
@ -3822,18 +3808,17 @@ def base_url(url):
def urljoin(base, path): def urljoin(base, path):
if isinstance(path, bytes): path = _decode_compat_str(path, encoding='utf-8', or_none=True)
path = path.decode('utf-8') if not path:
if not isinstance(path, compat_str) or not path:
return None return None
if re.match(r'^(?:[a-zA-Z][a-zA-Z0-9+-.]*:)?//', path): if re.match(r'^(?:[a-zA-Z][a-zA-Z0-9+-.]*:)?//', path):
return path return path
if isinstance(base, bytes): base = _decode_compat_str(base, encoding='utf-8', or_none=True)
base = base.decode('utf-8') if not base:
if not isinstance(base, compat_str) or not re.match(
r'^(?:https?:)?//', base):
return None return None
return compat_urllib_parse.urljoin(base, path) return (
re.match(r'^(?:https?:)?//', base)
and compat_urllib_parse.urljoin(base, path))
class HEADRequest(compat_urllib_request.Request): class HEADRequest(compat_urllib_request.Request):
@ -3998,8 +3983,7 @@ def get_exe_version(exe, args=['--version'],
stdout=subprocess.PIPE, stderr=subprocess.STDOUT)) stdout=subprocess.PIPE, stderr=subprocess.STDOUT))
except OSError: except OSError:
return False return False
if isinstance(out, bytes): # Python 2.x out = _decode_compat_str(out, 'ascii', 'ignore')
out = out.decode('ascii', 'ignore')
return detect_exe_version(out, version_re, unrecognized) return detect_exe_version(out, version_re, unrecognized)
@ -4218,8 +4202,8 @@ def lowercase_escape(s):
def escape_rfc3986(s): def escape_rfc3986(s):
"""Escape non-ASCII characters as suggested by RFC 3986""" """Escape non-ASCII characters as suggested by RFC 3986"""
if sys.version_info < (3, 0) and isinstance(s, compat_str): if sys.version_info < (3, 0):
s = s.encode('utf-8') s = _encode_compat_str(s, 'utf-8')
# ensure unicode: after quoting, it can always be converted # ensure unicode: after quoting, it can always be converted
return compat_str(compat_urllib_parse.quote(s, b"%/;:@&=+$,!~*'()?#[]")) return compat_str(compat_urllib_parse.quote(s, b"%/;:@&=+$,!~*'()?#[]"))
@ -4242,8 +4226,7 @@ def parse_qs(url, **kwargs):
def read_batch_urls(batch_fd): def read_batch_urls(batch_fd):
def fixup(url): def fixup(url):
if not isinstance(url, compat_str): url = _decode_compat_str(url, 'utf-8', 'replace')
url = url.decode('utf-8', 'replace')
BOM_UTF8 = '\xef\xbb\xbf' BOM_UTF8 = '\xef\xbb\xbf'
if url.startswith(BOM_UTF8): if url.startswith(BOM_UTF8):
url = url[len(BOM_UTF8):] url = url[len(BOM_UTF8):]
@ -4305,10 +4288,8 @@ def _multipart_encode_impl(data, boundary):
out = b'' out = b''
for k, v in data.items(): for k, v in data.items():
out += b'--' + boundary.encode('ascii') + b'\r\n' out += b'--' + boundary.encode('ascii') + b'\r\n'
if isinstance(k, compat_str): k = _encode_compat_str(k, 'utf-8')
k = k.encode('utf-8') v = _encode_compat_str(v, 'utf-8')
if isinstance(v, compat_str):
v = v.encode('utf-8')
# RFC 2047 requires non-ASCII field names to be encoded, while RFC 7578 # RFC 2047 requires non-ASCII field names to be encoded, while RFC 7578
# suggests sending UTF-8 directly. Firefox sends UTF-8, too # suggests sending UTF-8 directly. Firefox sends UTF-8, too
content = b'Content-Disposition: form-data; name="' + k + b'"\r\n\r\n' + v + b'\r\n' content = b'Content-Disposition: form-data; name="' + k + b'"\r\n\r\n' + v + b'\r\n'
@ -4399,6 +4380,11 @@ def try_get(src, getter, expected_type=None):
return v return v
def filter_dict(dct, cndn=lambda _, v: v is not None):
# NB: don't use dict comprehension for python 2.6 compatibility
return dict((k, v) for k, v in dct.items() if cndn(k, v))
def merge_dicts(*dicts, **kwargs): def merge_dicts(*dicts, **kwargs):
""" """
Merge the `dict`s in `dicts` using the first valid value for each key. Merge the `dict`s in `dicts` using the first valid value for each key.
@ -4435,8 +4421,26 @@ def merge_dicts(*dicts, **kwargs):
return merged return merged
def encode_compat_str(string, encoding=preferredencoding(), errors='strict'): # very poor choice of name, as if Python string encodings weren't confusing enough
return string if isinstance(string, compat_str) else compat_str(string, encoding, errors) def encode_compat_str(s, encoding=preferredencoding(), errors='strict'):
assert isinstance(s, compat_basestring)
return s if isinstance(s, compat_str) else compat_str(s, encoding, errors)
# what it could have been
def _decode_compat_str(s, encoding=preferredencoding(), errors='strict', or_none=False):
if not or_none:
assert isinstance(s, compat_basestring)
return (
s if isinstance(s, compat_str)
else compat_str(s, encoding, errors) if isinstance(s, compat_basestring)
else None)
# the real encode_compat_str, but only for internal use
def _encode_compat_str(s, encoding=preferredencoding(), errors='strict'):
assert isinstance(s, compat_basestring)
return s.encode(encoding, errors) if isinstance(s, compat_str) else s
US_RATINGS = { US_RATINGS = {
@ -4459,8 +4463,10 @@ TV_PARENTAL_GUIDELINES = {
def parse_age_limit(s): def parse_age_limit(s):
if type(s) == int: if not isinstance(s, bool):
return s if 0 <= s <= 21 else None age = int_or_none(s)
if age is not None:
return age if 0 <= age <= 21 else None
if not isinstance(s, compat_basestring): if not isinstance(s, compat_basestring):
return None return None
m = re.match(r'^(?P<age>\d{1,2})\+?$', s) m = re.match(r'^(?P<age>\d{1,2})\+?$', s)
@ -4637,12 +4643,7 @@ def args_to_str(args):
def error_to_compat_str(err): def error_to_compat_str(err):
err_str = str(err) return _decode_compat_str(str(err))
# On python 2 error byte string must be decoded with proper
# encoding rather than ascii
if sys.version_info[0] < 3:
err_str = err_str.decode(preferredencoding())
return err_str
def mimetype2ext(mt): def mimetype2ext(mt):