mirror of
https://github.com/yt-dlp/yt-dlp.git
synced 2024-11-22 02:15:12 +00:00
merged pep8_whitespace
This commit is contained in:
commit
ef53099e35
1 changed files with 51 additions and 31 deletions
72
youtube-dl
72
youtube-dl
|
@ -201,6 +201,7 @@ def preferredencoding():
|
|||
yield pref
|
||||
return yield_preferredencoding().next()
|
||||
|
||||
|
||||
def htmlentity_transform(matchobj):
|
||||
"""Transforms an HTML entity to a Unicode character.
|
||||
|
||||
|
@ -227,11 +228,13 @@ def htmlentity_transform(matchobj):
|
|||
# Unknown entity in name, return its literal representation
|
||||
return (u'&%s;' % entity)
|
||||
|
||||
|
||||
def sanitize_title(utitle):
|
||||
"""Sanitizes a video title so it could be used as part of a filename."""
|
||||
utitle = re.sub(ur'(?u)&(.+?);', htmlentity_transform, utitle)
|
||||
return utitle.replace(unicode(os.sep), u'%')
|
||||
|
||||
|
||||
def sanitize_open(filename, open_mode):
|
||||
"""Try to open the given filename, and slightly tweak it if this fails.
|
||||
|
||||
|
@ -258,13 +261,15 @@ def sanitize_open(filename, open_mode):
|
|||
stream = open(filename, open_mode)
|
||||
return (stream, filename)
|
||||
|
||||
|
||||
def timeconvert(timestr):
|
||||
"""Convert RFC 2822 defined time string into system timestamp"""
|
||||
timestamp = None
|
||||
timetuple = email.utils.parsedate_tz(timestr)
|
||||
if timetuple is not None:
|
||||
timestamp = email.utils.mktime_tz(timetuple)
|
||||
return timestamp
|
||||
"""Convert RFC 2822 defined time string into system timestamp"""
|
||||
timestamp = None
|
||||
timetuple = email.utils.parsedate_tz(timestr)
|
||||
if timetuple is not None:
|
||||
timestamp = email.utils.mktime_tz(timetuple)
|
||||
return timestamp
|
||||
|
||||
|
||||
class DownloadError(Exception):
|
||||
"""Download Error exception.
|
||||
|
@ -275,6 +280,7 @@ class DownloadError(Exception):
|
|||
"""
|
||||
pass
|
||||
|
||||
|
||||
class SameFileError(Exception):
|
||||
"""Same File exception.
|
||||
|
||||
|
@ -283,6 +289,7 @@ class SameFileError(Exception):
|
|||
"""
|
||||
pass
|
||||
|
||||
|
||||
class PostProcessingError(Exception):
|
||||
"""Post Processing exception.
|
||||
|
||||
|
@ -291,6 +298,7 @@ class PostProcessingError(Exception):
|
|||
"""
|
||||
pass
|
||||
|
||||
|
||||
class UnavailableVideoError(Exception):
|
||||
"""Unavailable Format exception.
|
||||
|
||||
|
@ -299,6 +307,7 @@ class UnavailableVideoError(Exception):
|
|||
"""
|
||||
pass
|
||||
|
||||
|
||||
class ContentTooShortError(Exception):
|
||||
"""Content Too Short exception.
|
||||
|
||||
|
@ -314,6 +323,7 @@ class ContentTooShortError(Exception):
|
|||
self.downloaded = downloaded
|
||||
self.expected = expected
|
||||
|
||||
|
||||
class YoutubeDLHandler(urllib2.HTTPHandler):
|
||||
"""Handler for HTTP requests and responses.
|
||||
|
||||
|
@ -326,7 +336,7 @@ class YoutubeDLHandler(urllib2.HTTPHandler):
|
|||
|
||||
Part of this code was copied from:
|
||||
|
||||
http://techknack.net/python-urllib2-handlers/
|
||||
http://techknack.net/python-urllib2-handlers/
|
||||
|
||||
Andrew Rowls, the author of that code, agreed to release it to the
|
||||
public domain.
|
||||
|
@ -372,6 +382,7 @@ class YoutubeDLHandler(urllib2.HTTPHandler):
|
|||
resp.msg = old_resp.msg
|
||||
return resp
|
||||
|
||||
|
||||
class FileDownloader(object):
|
||||
"""File Downloader class.
|
||||
|
||||
|
@ -465,7 +476,7 @@ class FileDownloader(object):
|
|||
else:
|
||||
exponent = long(math.log(bytes, 1024.0))
|
||||
suffix = 'bkMGTPEZY'[exponent]
|
||||
converted = float(bytes) / float(1024**exponent)
|
||||
converted = float(bytes) / float(1024 ** exponent)
|
||||
return '%.2f%s' % (converted, suffix)
|
||||
|
||||
@staticmethod
|
||||
|
@ -617,7 +628,7 @@ class FileDownloader(object):
|
|||
if filetime is None:
|
||||
return
|
||||
try:
|
||||
os.utime(filename,(time.time(), filetime))
|
||||
os.utime(filename, (time.time(), filetime))
|
||||
except:
|
||||
pass
|
||||
|
||||
|
@ -859,7 +870,7 @@ class FileDownloader(object):
|
|||
# Request parameters in case of being able to resume
|
||||
if self.params.get('continuedl', False) and resume_len != 0:
|
||||
self.report_resuming_byte(resume_len)
|
||||
request.add_header('Range','bytes=%d-' % resume_len)
|
||||
request.add_header('Range', 'bytes=%d-' % resume_len)
|
||||
open_mode = 'ab'
|
||||
|
||||
count = 0
|
||||
|
@ -885,7 +896,7 @@ class FileDownloader(object):
|
|||
else:
|
||||
# Examine the reported length
|
||||
if (content_length is not None and
|
||||
(resume_len - 100 < long(content_length) < resume_len + 100)):
|
||||
(resume_len - 100 < long(content_length) < resume_len + 100)):
|
||||
# The file had already been fully downloaded.
|
||||
# Explanation to the above condition: in issue #175 it was revealed that
|
||||
# YouTube sometimes adds or removes a few bytes from the end of the file,
|
||||
|
@ -967,6 +978,7 @@ class FileDownloader(object):
|
|||
|
||||
return True
|
||||
|
||||
|
||||
class InfoExtractor(object):
|
||||
"""Information Extractor class.
|
||||
|
||||
|
@ -1038,6 +1050,7 @@ class InfoExtractor(object):
|
|||
"""Real extraction process. Redefine in subclasses."""
|
||||
pass
|
||||
|
||||
|
||||
class YoutubeIE(InfoExtractor):
|
||||
"""Information extractor for youtube.com."""
|
||||
|
||||
|
@ -1192,7 +1205,7 @@ class YoutubeIE(InfoExtractor):
|
|||
self.report_video_info_webpage_download(video_id)
|
||||
for el_type in ['&el=embedded', '&el=detailpage', '&el=vevo', '']:
|
||||
video_info_url = ('http://www.youtube.com/get_video_info?&video_id=%s%s&ps=default&eurl=&gl=US&hl=en'
|
||||
% (video_id, el_type))
|
||||
% (video_id, el_type))
|
||||
request = urllib2.Request(video_info_url)
|
||||
try:
|
||||
video_info_webpage = urllib2.urlopen(request).read()
|
||||
|
@ -1561,6 +1574,7 @@ class DailymotionIE(InfoExtractor):
|
|||
except UnavailableVideoError:
|
||||
self._downloader.trouble(u'\nERROR: unable to download video')
|
||||
|
||||
|
||||
class GoogleIE(InfoExtractor):
|
||||
"""Information extractor for video.google.com."""
|
||||
|
||||
|
@ -1654,7 +1668,6 @@ class GoogleIE(InfoExtractor):
|
|||
else: # we need something to pass to process_info
|
||||
video_thumbnail = ''
|
||||
|
||||
|
||||
try:
|
||||
# Process video information
|
||||
self._downloader.process_info({
|
||||
|
@ -1854,7 +1867,8 @@ class YahooIE(InfoExtractor):
|
|||
self._downloader.trouble(u'ERROR: unable to extract video description')
|
||||
return
|
||||
video_description = mobj.group(1).decode('utf-8')
|
||||
if not video_description: video_description = 'No description available.'
|
||||
if not video_description:
|
||||
video_description = 'No description available.'
|
||||
|
||||
# Extract video height and width
|
||||
mobj = re.search(r'<meta name="video_height" content="([0-9]+)" />', webpage)
|
||||
|
@ -1875,8 +1889,8 @@ class YahooIE(InfoExtractor):
|
|||
yv_lg = 'R0xx6idZnW2zlrKP8xxAIR' # not sure what this represents
|
||||
yv_bitrate = '700' # according to Wikipedia this is hard-coded
|
||||
request = urllib2.Request('http://cosmos.bcst.yahoo.com/up/yep/process/getPlaylistFOP.php?node_id=' + video_id +
|
||||
'&tech=flash&mode=playlist&lg=' + yv_lg + '&bitrate=' + yv_bitrate + '&vidH=' + yv_video_height +
|
||||
'&vidW=' + yv_video_width + '&swf=as3&rd=video.yahoo.com&tk=null&adsupported=v1,v2,&eventid=1301797')
|
||||
'&tech=flash&mode=playlist&lg=' + yv_lg + '&bitrate=' + yv_bitrate + '&vidH=' + yv_video_height +
|
||||
'&vidW=' + yv_video_width + '&swf=as3&rd=video.yahoo.com&tk=null&adsupported=v1,v2,&eventid=1301797')
|
||||
try:
|
||||
self.report_download_webpage(video_id)
|
||||
webpage = urllib2.urlopen(request).read()
|
||||
|
@ -2085,11 +2099,11 @@ class GenericIE(InfoExtractor):
|
|||
return
|
||||
|
||||
video_url = urllib.unquote(mobj.group(1))
|
||||
video_id = os.path.basename(video_url)
|
||||
video_id = os.path.basename(video_url)
|
||||
|
||||
# here's a fun little line of code for you:
|
||||
video_extension = os.path.splitext(video_id)[1][1:]
|
||||
video_id = os.path.splitext(video_id)[0]
|
||||
video_id = os.path.splitext(video_id)[0]
|
||||
|
||||
# it's tempting to parse this further, but you would
|
||||
# have to take into account all the variations like
|
||||
|
@ -2162,7 +2176,7 @@ class YoutubeSearchIE(InfoExtractor):
|
|||
|
||||
prefix, query = query.split(':')
|
||||
prefix = prefix[8:]
|
||||
query = query.encode('utf-8')
|
||||
query = query.encode('utf-8')
|
||||
if prefix == '':
|
||||
self._download_n_results(query, 1)
|
||||
return
|
||||
|
@ -2176,7 +2190,7 @@ class YoutubeSearchIE(InfoExtractor):
|
|||
self._downloader.trouble(u'ERROR: invalid download number %s for query "%s"' % (n, query))
|
||||
return
|
||||
elif n > self._max_youtube_results:
|
||||
self._downloader.to_stderr(u'WARNING: ytsearch returns max %i results (you requested %i)' % (self._max_youtube_results, n))
|
||||
self._downloader.to_stderr(u'WARNING: ytsearch returns max %i results (you requested %i)' % (self._max_youtube_results, n))
|
||||
n = self._max_youtube_results
|
||||
self._download_n_results(query, n)
|
||||
return
|
||||
|
@ -2220,6 +2234,7 @@ class YoutubeSearchIE(InfoExtractor):
|
|||
|
||||
pagenum = pagenum + 1
|
||||
|
||||
|
||||
class GoogleSearchIE(InfoExtractor):
|
||||
"""Information Extractor for Google Video search queries."""
|
||||
_VALID_QUERY = r'gvsearch(\d+|all)?:[\s\S]+'
|
||||
|
@ -2253,7 +2268,7 @@ class GoogleSearchIE(InfoExtractor):
|
|||
|
||||
prefix, query = query.split(':')
|
||||
prefix = prefix[8:]
|
||||
query = query.encode('utf-8')
|
||||
query = query.encode('utf-8')
|
||||
if prefix == '':
|
||||
self._download_n_results(query, 1)
|
||||
return
|
||||
|
@ -2267,7 +2282,7 @@ class GoogleSearchIE(InfoExtractor):
|
|||
self._downloader.trouble(u'ERROR: invalid download number %s for query "%s"' % (n, query))
|
||||
return
|
||||
elif n > self._max_google_results:
|
||||
self._downloader.to_stderr(u'WARNING: gvsearch returns max %i results (you requested %i)' % (self._max_google_results, n))
|
||||
self._downloader.to_stderr(u'WARNING: gvsearch returns max %i results (you requested %i)' % (self._max_google_results, n))
|
||||
n = self._max_google_results
|
||||
self._download_n_results(query, n)
|
||||
return
|
||||
|
@ -2311,6 +2326,7 @@ class GoogleSearchIE(InfoExtractor):
|
|||
|
||||
pagenum = pagenum + 1
|
||||
|
||||
|
||||
class YahooSearchIE(InfoExtractor):
|
||||
"""Information Extractor for Yahoo! Video search queries."""
|
||||
_VALID_QUERY = r'yvsearch(\d+|all)?:[\s\S]+'
|
||||
|
@ -2344,7 +2360,7 @@ class YahooSearchIE(InfoExtractor):
|
|||
|
||||
prefix, query = query.split(':')
|
||||
prefix = prefix[8:]
|
||||
query = query.encode('utf-8')
|
||||
query = query.encode('utf-8')
|
||||
if prefix == '':
|
||||
self._download_n_results(query, 1)
|
||||
return
|
||||
|
@ -2358,7 +2374,7 @@ class YahooSearchIE(InfoExtractor):
|
|||
self._downloader.trouble(u'ERROR: invalid download number %s for query "%s"' % (n, query))
|
||||
return
|
||||
elif n > self._max_yahoo_results:
|
||||
self._downloader.to_stderr(u'WARNING: yvsearch returns max %i results (you requested %i)' % (self._max_yahoo_results, n))
|
||||
self._downloader.to_stderr(u'WARNING: yvsearch returns max %i results (you requested %i)' % (self._max_yahoo_results, n))
|
||||
n = self._max_yahoo_results
|
||||
self._download_n_results(query, n)
|
||||
return
|
||||
|
@ -2402,6 +2418,7 @@ class YahooSearchIE(InfoExtractor):
|
|||
|
||||
pagenum = pagenum + 1
|
||||
|
||||
|
||||
class YoutubePlaylistIE(InfoExtractor):
|
||||
"""Information Extractor for YouTube playlists."""
|
||||
|
||||
|
@ -2478,6 +2495,7 @@ class YoutubePlaylistIE(InfoExtractor):
|
|||
self._youtube_ie.extract('http://www.youtube.com/watch?v=%s' % id)
|
||||
return
|
||||
|
||||
|
||||
class YoutubeUserIE(InfoExtractor):
|
||||
"""Information Extractor for YouTube users."""
|
||||
|
||||
|
@ -2499,7 +2517,7 @@ class YoutubeUserIE(InfoExtractor):
|
|||
def report_download_page(self, username, start_index):
|
||||
"""Report attempt to download user page."""
|
||||
self._downloader.to_screen(u'[youtube] user %s: Downloading video ids from %d to %d' %
|
||||
(username, start_index, start_index + self._GDATA_PAGE_SIZE))
|
||||
(username, start_index, start_index + self._GDATA_PAGE_SIZE))
|
||||
|
||||
def _real_initialize(self):
|
||||
self._youtube_ie.initialize()
|
||||
|
@ -2563,7 +2581,7 @@ class YoutubeUserIE(InfoExtractor):
|
|||
video_ids = video_ids[playliststart:playlistend]
|
||||
|
||||
self._downloader.to_screen("[youtube] user %s: Collected %d video ids (downloading %d of them)" %
|
||||
(username, all_ids_count, len(video_ids)))
|
||||
(username, all_ids_count, len(video_ids)))
|
||||
|
||||
for video_id in video_ids:
|
||||
self._youtube_ie.extract('http://www.youtube.com/watch?v=%s' % video_id)
|
||||
|
@ -2648,6 +2666,7 @@ class DepositFilesIE(InfoExtractor):
|
|||
except UnavailableVideoError, err:
|
||||
self._downloader.trouble(u'ERROR: unable to download file')
|
||||
|
||||
|
||||
class FacebookIE(InfoExtractor):
|
||||
"""Information Extractor for Facebook"""
|
||||
|
||||
|
@ -2989,6 +3008,7 @@ class PostProcessor(object):
|
|||
"""
|
||||
return information # by default, do nothing
|
||||
|
||||
|
||||
class FFmpegExtractAudioPP(PostProcessor):
|
||||
|
||||
def __init__(self, downloader=None, preferredcodec=None):
|
||||
|
|
Loading…
Reference in a new issue