2012-03-24 19:07:37 -06:00
|
|
|
#!/usr/bin/env python
|
|
|
|
# -*- coding: utf-8 -*-
|
|
|
|
|
|
|
|
import gzip
|
2012-11-27 16:09:17 -07:00
|
|
|
import io
|
2012-12-20 05:13:24 -07:00
|
|
|
import json
|
2012-03-24 19:07:37 -06:00
|
|
|
import locale
|
|
|
|
import os
|
|
|
|
import re
|
|
|
|
import sys
|
2013-01-03 07:39:55 -07:00
|
|
|
import traceback
|
2012-03-24 19:07:37 -06:00
|
|
|
import zlib
|
|
|
|
import email.utils
|
2012-05-01 09:01:51 -06:00
|
|
|
import json
|
2012-03-24 19:07:37 -06:00
|
|
|
|
2012-11-27 15:54:09 -07:00
|
|
|
try:
|
2012-11-27 18:04:46 -07:00
|
|
|
import urllib.request as compat_urllib_request
|
2012-11-27 15:54:09 -07:00
|
|
|
except ImportError: # Python 2
|
2012-11-27 18:04:46 -07:00
|
|
|
import urllib2 as compat_urllib_request
|
2012-11-27 15:54:09 -07:00
|
|
|
|
|
|
|
try:
|
2012-11-27 18:04:46 -07:00
|
|
|
import urllib.error as compat_urllib_error
|
2012-11-27 15:54:09 -07:00
|
|
|
except ImportError: # Python 2
|
2012-11-27 18:04:46 -07:00
|
|
|
import urllib2 as compat_urllib_error
|
2012-11-27 15:54:09 -07:00
|
|
|
|
|
|
|
try:
|
2012-11-27 18:04:46 -07:00
|
|
|
import urllib.parse as compat_urllib_parse
|
2012-11-27 15:54:09 -07:00
|
|
|
except ImportError: # Python 2
|
2012-11-27 18:04:46 -07:00
|
|
|
import urllib as compat_urllib_parse
|
2012-11-27 15:54:09 -07:00
|
|
|
|
2012-11-27 20:51:27 -07:00
|
|
|
try:
|
|
|
|
from urllib.parse import urlparse as compat_urllib_parse_urlparse
|
|
|
|
except ImportError: # Python 2
|
|
|
|
from urlparse import urlparse as compat_urllib_parse_urlparse
|
|
|
|
|
2012-11-27 15:54:09 -07:00
|
|
|
try:
|
2012-11-27 18:04:46 -07:00
|
|
|
import http.cookiejar as compat_cookiejar
|
2012-11-27 15:54:09 -07:00
|
|
|
except ImportError: # Python 2
|
2012-11-27 18:04:46 -07:00
|
|
|
import cookielib as compat_cookiejar
|
2012-11-27 15:54:09 -07:00
|
|
|
|
2012-11-27 16:02:55 -07:00
|
|
|
try:
|
2012-11-27 18:04:46 -07:00
|
|
|
import html.entities as compat_html_entities
|
2012-11-27 16:17:12 -07:00
|
|
|
except ImportError: # Python 2
|
2012-11-27 18:04:46 -07:00
|
|
|
import htmlentitydefs as compat_html_entities
|
2012-11-27 16:02:55 -07:00
|
|
|
|
2012-11-27 16:06:28 -07:00
|
|
|
try:
|
2012-11-27 18:04:46 -07:00
|
|
|
import html.parser as compat_html_parser
|
2012-11-27 16:17:12 -07:00
|
|
|
except ImportError: # Python 2
|
2012-11-27 18:04:46 -07:00
|
|
|
import HTMLParser as compat_html_parser
|
2012-11-27 16:06:28 -07:00
|
|
|
|
2012-11-27 16:13:00 -07:00
|
|
|
try:
|
2012-11-27 18:04:46 -07:00
|
|
|
import http.client as compat_http_client
|
2012-11-27 16:17:12 -07:00
|
|
|
except ImportError: # Python 2
|
2012-11-27 18:04:46 -07:00
|
|
|
import httplib as compat_http_client
|
2012-11-27 16:13:00 -07:00
|
|
|
|
2012-12-16 04:29:03 -07:00
|
|
|
try:
|
|
|
|
from subprocess import DEVNULL
|
|
|
|
compat_subprocess_get_DEVNULL = lambda: DEVNULL
|
|
|
|
except ImportError:
|
|
|
|
compat_subprocess_get_DEVNULL = lambda: open(os.path.devnull, 'w')
|
|
|
|
|
2012-11-27 16:17:12 -07:00
|
|
|
try:
|
2012-11-27 18:04:46 -07:00
|
|
|
from urllib.parse import parse_qs as compat_parse_qs
|
2012-11-27 16:17:12 -07:00
|
|
|
except ImportError: # Python 2
|
2012-11-27 18:04:46 -07:00
|
|
|
# HACK: The following is the correct parse_qs implementation from cpython 3's stdlib.
|
|
|
|
# Python 2's version is apparently totally broken
|
|
|
|
def _unquote(string, encoding='utf-8', errors='replace'):
|
|
|
|
if string == '':
|
|
|
|
return string
|
|
|
|
res = string.split('%')
|
|
|
|
if len(res) == 1:
|
|
|
|
return string
|
|
|
|
if encoding is None:
|
|
|
|
encoding = 'utf-8'
|
|
|
|
if errors is None:
|
|
|
|
errors = 'replace'
|
|
|
|
# pct_sequence: contiguous sequence of percent-encoded bytes, decoded
|
|
|
|
pct_sequence = b''
|
|
|
|
string = res[0]
|
|
|
|
for item in res[1:]:
|
|
|
|
try:
|
|
|
|
if not item:
|
|
|
|
raise ValueError
|
|
|
|
pct_sequence += item[:2].decode('hex')
|
|
|
|
rest = item[2:]
|
|
|
|
if not rest:
|
|
|
|
# This segment was just a single percent-encoded character.
|
|
|
|
# May be part of a sequence of code units, so delay decoding.
|
|
|
|
# (Stored in pct_sequence).
|
|
|
|
continue
|
|
|
|
except ValueError:
|
|
|
|
rest = '%' + item
|
|
|
|
# Encountered non-percent-encoded characters. Flush the current
|
|
|
|
# pct_sequence.
|
|
|
|
string += pct_sequence.decode(encoding, errors) + rest
|
|
|
|
pct_sequence = b''
|
|
|
|
if pct_sequence:
|
|
|
|
# Flush the final pct_sequence
|
|
|
|
string += pct_sequence.decode(encoding, errors)
|
|
|
|
return string
|
|
|
|
|
|
|
|
def _parse_qsl(qs, keep_blank_values=False, strict_parsing=False,
|
|
|
|
encoding='utf-8', errors='replace'):
|
|
|
|
qs, _coerce_result = qs, unicode
|
|
|
|
pairs = [s2 for s1 in qs.split('&') for s2 in s1.split(';')]
|
|
|
|
r = []
|
|
|
|
for name_value in pairs:
|
|
|
|
if not name_value and not strict_parsing:
|
|
|
|
continue
|
|
|
|
nv = name_value.split('=', 1)
|
|
|
|
if len(nv) != 2:
|
|
|
|
if strict_parsing:
|
|
|
|
raise ValueError("bad query field: %r" % (name_value,))
|
|
|
|
# Handle case of a control-name with no equal sign
|
|
|
|
if keep_blank_values:
|
|
|
|
nv.append('')
|
|
|
|
else:
|
|
|
|
continue
|
|
|
|
if len(nv[1]) or keep_blank_values:
|
|
|
|
name = nv[0].replace('+', ' ')
|
|
|
|
name = _unquote(name, encoding=encoding, errors=errors)
|
|
|
|
name = _coerce_result(name)
|
|
|
|
value = nv[1].replace('+', ' ')
|
|
|
|
value = _unquote(value, encoding=encoding, errors=errors)
|
|
|
|
value = _coerce_result(value)
|
|
|
|
r.append((name, value))
|
|
|
|
return r
|
|
|
|
|
|
|
|
def compat_parse_qs(qs, keep_blank_values=False, strict_parsing=False,
|
|
|
|
encoding='utf-8', errors='replace'):
|
|
|
|
parsed_result = {}
|
|
|
|
pairs = _parse_qsl(qs, keep_blank_values, strict_parsing,
|
|
|
|
encoding=encoding, errors=errors)
|
|
|
|
for name, value in pairs:
|
|
|
|
if name in parsed_result:
|
|
|
|
parsed_result[name].append(value)
|
|
|
|
else:
|
|
|
|
parsed_result[name] = [value]
|
|
|
|
return parsed_result
|
2012-11-27 16:13:00 -07:00
|
|
|
|
2012-11-27 16:02:55 -07:00
|
|
|
try:
|
2012-11-27 18:04:46 -07:00
|
|
|
compat_str = unicode # Python 2
|
2012-11-27 16:02:55 -07:00
|
|
|
except NameError:
|
2012-11-27 18:04:46 -07:00
|
|
|
compat_str = str
|
2012-11-27 16:02:55 -07:00
|
|
|
|
|
|
|
try:
|
2012-11-27 18:04:46 -07:00
|
|
|
compat_chr = unichr # Python 2
|
2012-11-27 16:02:55 -07:00
|
|
|
except NameError:
|
2012-11-27 18:04:46 -07:00
|
|
|
compat_chr = chr
|
2012-11-27 16:02:55 -07:00
|
|
|
|
|
|
|
std_headers = {
|
2012-11-27 18:04:46 -07:00
|
|
|
'User-Agent': 'Mozilla/5.0 (X11; Linux x86_64; rv:10.0) Gecko/20100101 Firefox/10.0',
|
|
|
|
'Accept-Charset': 'ISO-8859-1,utf-8;q=0.7,*;q=0.7',
|
|
|
|
'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8',
|
|
|
|
'Accept-Encoding': 'gzip, deflate',
|
|
|
|
'Accept-Language': 'en-us,en;q=0.5',
|
2012-11-27 16:02:55 -07:00
|
|
|
}
|
2012-12-30 10:22:36 -07:00
|
|
|
|
2012-03-24 19:07:37 -06:00
|
|
|
def preferredencoding():
|
2012-11-27 18:04:46 -07:00
|
|
|
"""Get preferred encoding.
|
2012-03-24 19:07:37 -06:00
|
|
|
|
2012-11-27 18:04:46 -07:00
|
|
|
Returns the best encoding scheme for the system, based on
|
|
|
|
locale.getpreferredencoding() and some further tweaks.
|
|
|
|
"""
|
|
|
|
try:
|
|
|
|
pref = locale.getpreferredencoding()
|
|
|
|
u'TEST'.encode(pref)
|
|
|
|
except:
|
|
|
|
pref = 'UTF-8'
|
2012-07-01 10:21:27 -06:00
|
|
|
|
2012-11-27 18:04:46 -07:00
|
|
|
return pref
|
2012-03-24 19:07:37 -06:00
|
|
|
|
2012-11-27 16:46:21 -07:00
|
|
|
if sys.version_info < (3,0):
|
2012-11-27 18:04:46 -07:00
|
|
|
def compat_print(s):
|
|
|
|
print(s.encode(preferredencoding(), 'xmlcharrefreplace'))
|
2012-11-27 16:46:21 -07:00
|
|
|
else:
|
2012-11-27 18:04:46 -07:00
|
|
|
def compat_print(s):
|
|
|
|
assert type(s) == type(u'')
|
|
|
|
print(s)
|
2012-03-24 19:07:37 -06:00
|
|
|
|
2012-12-20 05:13:24 -07:00
|
|
|
# In Python 2.x, json.dump expects a bytestream.
|
|
|
|
# In Python 3.x, it writes to a character stream
|
|
|
|
if sys.version_info < (3,0):
|
|
|
|
def write_json_file(obj, fn):
|
|
|
|
with open(fn, 'wb') as f:
|
|
|
|
json.dump(obj, f)
|
|
|
|
else:
|
|
|
|
def write_json_file(obj, fn):
|
|
|
|
with open(fn, 'w', encoding='utf-8') as f:
|
|
|
|
json.dump(obj, f)
|
|
|
|
|
2012-03-24 19:07:37 -06:00
|
|
|
def htmlentity_transform(matchobj):
|
2012-11-27 18:04:46 -07:00
|
|
|
"""Transforms an HTML entity to a character.
|
|
|
|
|
|
|
|
This function receives a match object and is intended to be used with
|
|
|
|
the re.sub() function.
|
|
|
|
"""
|
|
|
|
entity = matchobj.group(1)
|
|
|
|
|
|
|
|
# Known non-numeric HTML entity
|
|
|
|
if entity in compat_html_entities.name2codepoint:
|
|
|
|
return compat_chr(compat_html_entities.name2codepoint[entity])
|
|
|
|
|
|
|
|
mobj = re.match(u'(?u)#(x?\\d+)', entity)
|
|
|
|
if mobj is not None:
|
|
|
|
numstr = mobj.group(1)
|
|
|
|
if numstr.startswith(u'x'):
|
|
|
|
base = 16
|
|
|
|
numstr = u'0%s' % numstr
|
|
|
|
else:
|
|
|
|
base = 10
|
|
|
|
return compat_chr(int(numstr, base))
|
|
|
|
|
|
|
|
# Unknown entity in name, return its literal representation
|
|
|
|
return (u'&%s;' % entity)
|
2012-03-24 19:07:37 -06:00
|
|
|
|
2012-11-27 16:06:28 -07:00
|
|
|
compat_html_parser.locatestarttagend = re.compile(r"""<[a-zA-Z][-.a-zA-Z0-9:_]*(?:\s+(?:(?<=['"\s])[^\s/>][^\s/=>]*(?:\s*=+\s*(?:'[^']*'|"[^"]*"|(?!['"])[^>\s]*))?\s*)*)?\s*""", re.VERBOSE) # backport bugfix
|
2012-12-19 07:21:14 -07:00
|
|
|
class AttrParser(compat_html_parser.HTMLParser):
|
|
|
|
"""Modified HTMLParser that isolates a tag with the specified attribute"""
|
|
|
|
def __init__(self, attribute, value):
|
|
|
|
self.attribute = attribute
|
|
|
|
self.value = value
|
2012-11-27 18:04:46 -07:00
|
|
|
self.result = None
|
|
|
|
self.started = False
|
|
|
|
self.depth = {}
|
|
|
|
self.html = None
|
|
|
|
self.watch_startpos = False
|
|
|
|
self.error_count = 0
|
|
|
|
compat_html_parser.HTMLParser.__init__(self)
|
|
|
|
|
|
|
|
def error(self, message):
|
|
|
|
if self.error_count > 10 or self.started:
|
|
|
|
raise compat_html_parser.HTMLParseError(message, self.getpos())
|
|
|
|
self.rawdata = '\n'.join(self.html.split('\n')[self.getpos()[0]:]) # skip one line
|
|
|
|
self.error_count += 1
|
|
|
|
self.goahead(1)
|
|
|
|
|
|
|
|
def loads(self, html):
|
|
|
|
self.html = html
|
|
|
|
self.feed(html)
|
|
|
|
self.close()
|
|
|
|
|
|
|
|
def handle_starttag(self, tag, attrs):
|
|
|
|
attrs = dict(attrs)
|
|
|
|
if self.started:
|
|
|
|
self.find_startpos(None)
|
2012-12-19 07:21:14 -07:00
|
|
|
if self.attribute in attrs and attrs[self.attribute] == self.value:
|
2012-11-27 18:04:46 -07:00
|
|
|
self.result = [tag]
|
|
|
|
self.started = True
|
|
|
|
self.watch_startpos = True
|
|
|
|
if self.started:
|
|
|
|
if not tag in self.depth: self.depth[tag] = 0
|
|
|
|
self.depth[tag] += 1
|
|
|
|
|
|
|
|
def handle_endtag(self, tag):
|
|
|
|
if self.started:
|
|
|
|
if tag in self.depth: self.depth[tag] -= 1
|
|
|
|
if self.depth[self.result[0]] == 0:
|
|
|
|
self.started = False
|
|
|
|
self.result.append(self.getpos())
|
|
|
|
|
|
|
|
def find_startpos(self, x):
|
|
|
|
"""Needed to put the start position of the result (self.result[1])
|
|
|
|
after the opening tag with the requested id"""
|
|
|
|
if self.watch_startpos:
|
|
|
|
self.watch_startpos = False
|
|
|
|
self.result.append(self.getpos())
|
|
|
|
handle_entityref = handle_charref = handle_data = handle_comment = \
|
|
|
|
handle_decl = handle_pi = unknown_decl = find_startpos
|
|
|
|
|
|
|
|
def get_result(self):
|
|
|
|
if self.result is None:
|
|
|
|
return None
|
|
|
|
if len(self.result) != 3:
|
|
|
|
return None
|
|
|
|
lines = self.html.split('\n')
|
|
|
|
lines = lines[self.result[1][0]-1:self.result[2][0]]
|
|
|
|
lines[0] = lines[0][self.result[1][1]:]
|
|
|
|
if len(lines) == 1:
|
|
|
|
lines[-1] = lines[-1][:self.result[2][1]-self.result[1][1]]
|
|
|
|
lines[-1] = lines[-1][:self.result[2][1]]
|
|
|
|
return '\n'.join(lines).strip()
|
2013-02-01 09:29:50 -07:00
|
|
|
# Hack for https://github.com/rg3/youtube-dl/issues/662
|
|
|
|
if sys.version_info < (2, 7, 3):
|
|
|
|
AttrParser.parse_endtag = (lambda self, i:
|
|
|
|
i + len("</scr'+'ipt>")
|
|
|
|
if self.rawdata[i:].startswith("</scr'+'ipt>")
|
|
|
|
else compat_html_parser.HTMLParser.parse_endtag(self, i))
|
2012-04-10 16:22:51 -06:00
|
|
|
|
|
|
|
def get_element_by_id(id, html):
|
2012-12-19 07:21:14 -07:00
|
|
|
"""Return the content of the tag with the specified ID in the passed HTML document"""
|
|
|
|
return get_element_by_attribute("id", id, html)
|
|
|
|
|
|
|
|
def get_element_by_attribute(attribute, value, html):
|
|
|
|
"""Return the content of the tag with the specified attribute in the passed HTML document"""
|
|
|
|
parser = AttrParser(attribute, value)
|
2012-11-27 18:04:46 -07:00
|
|
|
try:
|
|
|
|
parser.loads(html)
|
|
|
|
except compat_html_parser.HTMLParseError:
|
|
|
|
pass
|
|
|
|
return parser.get_result()
|
2012-04-10 16:22:51 -06:00
|
|
|
|
|
|
|
|
|
|
|
def clean_html(html):
|
2012-11-27 18:04:46 -07:00
|
|
|
"""Clean an HTML snippet into a readable string"""
|
|
|
|
# Newline vs <br />
|
|
|
|
html = html.replace('\n', ' ')
|
2012-12-20 08:30:55 -07:00
|
|
|
html = re.sub(r'\s*<\s*br\s*/?\s*>\s*', '\n', html)
|
|
|
|
html = re.sub(r'<\s*/\s*p\s*>\s*<\s*p[^>]*>', '\n', html)
|
2012-11-27 18:04:46 -07:00
|
|
|
# Strip html tags
|
|
|
|
html = re.sub('<.*?>', '', html)
|
|
|
|
# Replace html entities
|
|
|
|
html = unescapeHTML(html)
|
|
|
|
return html
|
2012-04-10 16:22:51 -06:00
|
|
|
|
|
|
|
|
2012-03-24 19:07:37 -06:00
|
|
|
def sanitize_open(filename, open_mode):
|
2012-11-27 18:04:46 -07:00
|
|
|
"""Try to open the given filename, and slightly tweak it if this fails.
|
|
|
|
|
|
|
|
Attempts to open the given filename. If this fails, it tries to change
|
|
|
|
the filename slightly, step by step, until it's either able to open it
|
|
|
|
or it fails and raises a final exception, like the standard open()
|
|
|
|
function.
|
|
|
|
|
|
|
|
It returns the tuple (stream, definitive_file_name).
|
|
|
|
"""
|
|
|
|
try:
|
|
|
|
if filename == u'-':
|
|
|
|
if sys.platform == 'win32':
|
|
|
|
import msvcrt
|
|
|
|
msvcrt.setmode(sys.stdout.fileno(), os.O_BINARY)
|
|
|
|
return (sys.stdout, filename)
|
|
|
|
stream = open(encodeFilename(filename), open_mode)
|
|
|
|
return (stream, filename)
|
|
|
|
except (IOError, OSError) as err:
|
|
|
|
# In case of error, try to remove win32 forbidden chars
|
|
|
|
filename = re.sub(u'[/<>:"\\|\\\\?\\*]', u'#', filename)
|
|
|
|
|
|
|
|
# An exception here should be caught in the caller
|
|
|
|
stream = open(encodeFilename(filename), open_mode)
|
|
|
|
return (stream, filename)
|
2012-03-24 19:07:37 -06:00
|
|
|
|
|
|
|
|
|
|
|
def timeconvert(timestr):
|
2012-11-27 18:04:46 -07:00
|
|
|
"""Convert RFC 2822 defined time string into system timestamp"""
|
|
|
|
timestamp = None
|
|
|
|
timetuple = email.utils.parsedate_tz(timestr)
|
|
|
|
if timetuple is not None:
|
|
|
|
timestamp = email.utils.mktime_tz(timetuple)
|
|
|
|
return timestamp
|
2012-11-26 15:58:46 -07:00
|
|
|
|
2012-12-03 07:36:24 -07:00
|
|
|
def sanitize_filename(s, restricted=False, is_id=False):
|
2012-11-27 18:04:46 -07:00
|
|
|
"""Sanitizes a string so it could be used as part of a filename.
|
|
|
|
If restricted is set, use a stricter subset of allowed characters.
|
2012-12-03 07:36:24 -07:00
|
|
|
Set is_id if this is not an arbitrary string, but an ID that should be kept if possible
|
2012-11-27 18:04:46 -07:00
|
|
|
"""
|
|
|
|
def replace_insane(char):
|
|
|
|
if char == '?' or ord(char) < 32 or ord(char) == 127:
|
|
|
|
return ''
|
|
|
|
elif char == '"':
|
|
|
|
return '' if restricted else '\''
|
|
|
|
elif char == ':':
|
|
|
|
return '_-' if restricted else ' -'
|
|
|
|
elif char in '\\/|*<>':
|
|
|
|
return '_'
|
2012-11-28 04:59:27 -07:00
|
|
|
if restricted and (char in '!&\'()[]{}$;`^,#' or char.isspace()):
|
2012-11-27 18:04:46 -07:00
|
|
|
return '_'
|
|
|
|
if restricted and ord(char) > 127:
|
|
|
|
return '_'
|
|
|
|
return char
|
|
|
|
|
|
|
|
result = u''.join(map(replace_insane, s))
|
2012-12-03 07:36:24 -07:00
|
|
|
if not is_id:
|
|
|
|
while '__' in result:
|
|
|
|
result = result.replace('__', '_')
|
|
|
|
result = result.strip('_')
|
|
|
|
# Common case of "Foreign band name - English song title"
|
|
|
|
if restricted and result.startswith('-_'):
|
|
|
|
result = result[2:]
|
|
|
|
if not result:
|
|
|
|
result = '_'
|
2012-11-27 18:04:46 -07:00
|
|
|
return result
|
2012-03-24 19:07:37 -06:00
|
|
|
|
|
|
|
def orderedSet(iterable):
|
2012-11-27 18:04:46 -07:00
|
|
|
""" Remove all duplicates from the input iterable """
|
|
|
|
res = []
|
|
|
|
for el in iterable:
|
|
|
|
if el not in res:
|
|
|
|
res.append(el)
|
|
|
|
return res
|
2012-03-24 19:07:37 -06:00
|
|
|
|
|
|
|
def unescapeHTML(s):
|
2012-11-27 18:04:46 -07:00
|
|
|
"""
|
|
|
|
@param s a string
|
|
|
|
"""
|
|
|
|
assert type(s) == type(u'')
|
2012-03-24 19:07:37 -06:00
|
|
|
|
2012-11-27 18:04:46 -07:00
|
|
|
result = re.sub(u'(?u)&(.+?);', htmlentity_transform, s)
|
|
|
|
return result
|
2012-03-24 19:07:37 -06:00
|
|
|
|
|
|
|
def encodeFilename(s):
|
2012-11-27 18:04:46 -07:00
|
|
|
"""
|
|
|
|
@param s The name of the file
|
|
|
|
"""
|
2012-03-24 19:07:37 -06:00
|
|
|
|
2012-11-27 18:04:46 -07:00
|
|
|
assert type(s) == type(u'')
|
2012-03-24 19:07:37 -06:00
|
|
|
|
2012-11-27 18:04:46 -07:00
|
|
|
# Python 3 has a Unicode API
|
|
|
|
if sys.version_info >= (3, 0):
|
|
|
|
return s
|
2012-11-27 16:56:20 -07:00
|
|
|
|
2012-11-27 18:04:46 -07:00
|
|
|
if sys.platform == 'win32' and sys.getwindowsversion()[0] >= 5:
|
|
|
|
# Pass u'' directly to use Unicode APIs on Windows 2000 and up
|
|
|
|
# (Detecting Windows NT 4 is tricky because 'major >= 4' would
|
|
|
|
# match Windows 9x series as well. Besides, NT 4 is obsolete.)
|
|
|
|
return s
|
|
|
|
else:
|
2013-01-19 17:48:05 -07:00
|
|
|
encoding = sys.getfilesystemencoding()
|
|
|
|
if encoding is None:
|
|
|
|
encoding = 'utf-8'
|
|
|
|
return s.encode(encoding, 'ignore')
|
2012-03-24 19:07:37 -06:00
|
|
|
|
2013-01-01 12:27:53 -07:00
|
|
|
|
|
|
|
class ExtractorError(Exception):
|
|
|
|
"""Error during info extraction."""
|
|
|
|
def __init__(self, msg, tb=None):
|
2013-01-03 07:39:55 -07:00
|
|
|
""" tb, if given, is the original traceback (so that it can be printed out). """
|
2013-01-01 12:27:53 -07:00
|
|
|
super(ExtractorError, self).__init__(msg)
|
|
|
|
self.traceback = tb
|
|
|
|
|
2013-01-03 07:39:55 -07:00
|
|
|
def format_traceback(self):
|
|
|
|
if self.traceback is None:
|
|
|
|
return None
|
|
|
|
return u''.join(traceback.format_tb(self.traceback))
|
|
|
|
|
2013-01-01 12:27:53 -07:00
|
|
|
|
2012-03-24 19:07:37 -06:00
|
|
|
class DownloadError(Exception):
|
2012-11-27 18:04:46 -07:00
|
|
|
"""Download Error exception.
|
2012-03-24 19:07:37 -06:00
|
|
|
|
2012-11-27 18:04:46 -07:00
|
|
|
This exception may be thrown by FileDownloader objects if they are not
|
|
|
|
configured to continue on errors. They will contain the appropriate
|
|
|
|
error message.
|
|
|
|
"""
|
|
|
|
pass
|
2012-03-24 19:07:37 -06:00
|
|
|
|
|
|
|
|
|
|
|
class SameFileError(Exception):
|
2012-11-27 18:04:46 -07:00
|
|
|
"""Same File exception.
|
2012-03-24 19:07:37 -06:00
|
|
|
|
2012-11-27 18:04:46 -07:00
|
|
|
This exception will be thrown by FileDownloader objects if they detect
|
|
|
|
multiple files would have to be downloaded to the same file on disk.
|
|
|
|
"""
|
|
|
|
pass
|
2012-03-24 19:07:37 -06:00
|
|
|
|
|
|
|
|
|
|
|
class PostProcessingError(Exception):
|
2012-11-27 18:04:46 -07:00
|
|
|
"""Post Processing exception.
|
2012-03-24 19:07:37 -06:00
|
|
|
|
2012-11-27 18:04:46 -07:00
|
|
|
This exception may be raised by PostProcessor's .run() method to
|
|
|
|
indicate an error in the postprocessing task.
|
|
|
|
"""
|
2013-01-12 07:07:59 -07:00
|
|
|
def __init__(self, msg):
|
|
|
|
self.msg = msg
|
2012-03-24 19:07:37 -06:00
|
|
|
|
|
|
|
class MaxDownloadsReached(Exception):
|
2012-11-27 18:04:46 -07:00
|
|
|
""" --max-downloads limit has been reached. """
|
|
|
|
pass
|
2012-03-24 19:07:37 -06:00
|
|
|
|
|
|
|
|
|
|
|
class UnavailableVideoError(Exception):
|
2012-11-27 18:04:46 -07:00
|
|
|
"""Unavailable Format exception.
|
2012-03-24 19:07:37 -06:00
|
|
|
|
2012-11-27 18:04:46 -07:00
|
|
|
This exception will be thrown when a video is requested
|
|
|
|
in a format that is not available for that video.
|
|
|
|
"""
|
|
|
|
pass
|
2012-03-24 19:07:37 -06:00
|
|
|
|
|
|
|
|
|
|
|
class ContentTooShortError(Exception):
|
2012-11-27 18:04:46 -07:00
|
|
|
"""Content Too Short exception.
|
2012-03-24 19:07:37 -06:00
|
|
|
|
2012-11-27 18:04:46 -07:00
|
|
|
This exception may be raised by FileDownloader objects when a file they
|
|
|
|
download is too small for what the server announced first, indicating
|
|
|
|
the connection was probably interrupted.
|
|
|
|
"""
|
|
|
|
# Both in bytes
|
|
|
|
downloaded = None
|
|
|
|
expected = None
|
2012-03-24 19:07:37 -06:00
|
|
|
|
2012-11-27 18:04:46 -07:00
|
|
|
def __init__(self, downloaded, expected):
|
|
|
|
self.downloaded = downloaded
|
|
|
|
self.expected = expected
|
2012-03-24 19:07:37 -06:00
|
|
|
|
2012-11-27 15:54:09 -07:00
|
|
|
class YoutubeDLHandler(compat_urllib_request.HTTPHandler):
|
2012-11-27 18:04:46 -07:00
|
|
|
"""Handler for HTTP requests and responses.
|
|
|
|
|
|
|
|
This class, when installed with an OpenerDirector, automatically adds
|
|
|
|
the standard headers to every HTTP request and handles gzipped and
|
|
|
|
deflated responses from web servers. If compression is to be avoided in
|
|
|
|
a particular request, the original request in the program code only has
|
|
|
|
to include the HTTP header "Youtubedl-No-Compression", which will be
|
|
|
|
removed before making the real request.
|
|
|
|
|
|
|
|
Part of this code was copied from:
|
|
|
|
|
|
|
|
http://techknack.net/python-urllib2-handlers/
|
|
|
|
|
|
|
|
Andrew Rowls, the author of that code, agreed to release it to the
|
|
|
|
public domain.
|
|
|
|
"""
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def deflate(data):
|
|
|
|
try:
|
|
|
|
return zlib.decompress(data, -zlib.MAX_WBITS)
|
|
|
|
except zlib.error:
|
|
|
|
return zlib.decompress(data)
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def addinfourl_wrapper(stream, headers, url, code):
|
|
|
|
if hasattr(compat_urllib_request.addinfourl, 'getcode'):
|
|
|
|
return compat_urllib_request.addinfourl(stream, headers, url, code)
|
|
|
|
ret = compat_urllib_request.addinfourl(stream, headers, url)
|
|
|
|
ret.code = code
|
|
|
|
return ret
|
|
|
|
|
|
|
|
def http_request(self, req):
|
2013-01-12 10:38:23 -07:00
|
|
|
for h,v in std_headers.items():
|
2012-11-27 18:04:46 -07:00
|
|
|
if h in req.headers:
|
|
|
|
del req.headers[h]
|
2013-01-12 10:38:23 -07:00
|
|
|
req.add_header(h, v)
|
2012-11-27 18:04:46 -07:00
|
|
|
if 'Youtubedl-no-compression' in req.headers:
|
|
|
|
if 'Accept-encoding' in req.headers:
|
|
|
|
del req.headers['Accept-encoding']
|
|
|
|
del req.headers['Youtubedl-no-compression']
|
2013-01-12 08:49:13 -07:00
|
|
|
if 'Youtubedl-user-agent' in req.headers:
|
2013-01-12 10:38:23 -07:00
|
|
|
if 'User-agent' in req.headers:
|
|
|
|
del req.headers['User-agent']
|
|
|
|
req.headers['User-agent'] = req.headers['Youtubedl-user-agent']
|
2013-01-12 08:49:13 -07:00
|
|
|
del req.headers['Youtubedl-user-agent']
|
2012-11-27 18:04:46 -07:00
|
|
|
return req
|
|
|
|
|
|
|
|
def http_response(self, req, resp):
|
|
|
|
old_resp = resp
|
|
|
|
# gzip
|
|
|
|
if resp.headers.get('Content-encoding', '') == 'gzip':
|
|
|
|
gz = gzip.GzipFile(fileobj=io.BytesIO(resp.read()), mode='r')
|
|
|
|
resp = self.addinfourl_wrapper(gz, old_resp.headers, old_resp.url, old_resp.code)
|
|
|
|
resp.msg = old_resp.msg
|
|
|
|
# deflate
|
|
|
|
if resp.headers.get('Content-encoding', '') == 'deflate':
|
|
|
|
gz = io.BytesIO(self.deflate(resp.read()))
|
|
|
|
resp = self.addinfourl_wrapper(gz, old_resp.headers, old_resp.url, old_resp.code)
|
|
|
|
resp.msg = old_resp.msg
|
|
|
|
return resp
|
2012-12-06 16:39:44 -07:00
|
|
|
|
|
|
|
https_request = http_request
|
|
|
|
https_response = http_response
|