mirror of
https://github.com/yt-dlp/yt-dlp.git
synced 2024-11-26 02:55:17 +00:00
various py3 fixes; all tests green on 3.3
This commit is contained in:
parent
5e5ddcfbcf
commit
80d3177e5c
2 changed files with 10 additions and 10 deletions
|
@ -453,9 +453,9 @@ def process_info(self, info_dict):
|
||||||
self.trouble(u'ERROR: No JSON encoder found. Update to Python 2.6+, setup a json module, or leave out --write-info-json.')
|
self.trouble(u'ERROR: No JSON encoder found. Update to Python 2.6+, setup a json module, or leave out --write-info-json.')
|
||||||
return
|
return
|
||||||
try:
|
try:
|
||||||
infof = open(encodeFilename(infofn), 'wb')
|
infof = open(encodeFilename(infofn), 'w')
|
||||||
try:
|
try:
|
||||||
json_info_dict = dict((k,v) for k,v in info_dict.iteritems() if not k in ('urlhandle',))
|
json_info_dict = dict((k, info_dict[k]) for k in info_dict if not k in ['urlhandle'])
|
||||||
json.dump(json_info_dict, infof)
|
json.dump(json_info_dict, infof)
|
||||||
finally:
|
finally:
|
||||||
infof.close()
|
infof.close()
|
||||||
|
|
|
@ -272,7 +272,7 @@ def _real_initialize(self):
|
||||||
request = compat_urllib_request.Request(self._LOGIN_URL, compat_urllib_parse.urlencode(login_form))
|
request = compat_urllib_request.Request(self._LOGIN_URL, compat_urllib_parse.urlencode(login_form))
|
||||||
try:
|
try:
|
||||||
self.report_login()
|
self.report_login()
|
||||||
login_results = compat_urllib_request.urlopen(request).read()
|
login_results = compat_urllib_request.urlopen(request).read().decode('utf-8')
|
||||||
if re.search(r'(?i)<form[^>]* name="loginForm"', login_results) is not None:
|
if re.search(r'(?i)<form[^>]* name="loginForm"', login_results) is not None:
|
||||||
self._downloader.to_stderr(u'WARNING: unable to log in: bad username or password')
|
self._downloader.to_stderr(u'WARNING: unable to log in: bad username or password')
|
||||||
return
|
return
|
||||||
|
@ -288,7 +288,7 @@ def _real_initialize(self):
|
||||||
request = compat_urllib_request.Request(self._AGE_URL, compat_urllib_parse.urlencode(age_form))
|
request = compat_urllib_request.Request(self._AGE_URL, compat_urllib_parse.urlencode(age_form))
|
||||||
try:
|
try:
|
||||||
self.report_age_confirmation()
|
self.report_age_confirmation()
|
||||||
age_results = compat_urllib_request.urlopen(request).read()
|
age_results = compat_urllib_request.urlopen(request).read().decode('utf-8')
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
||||||
self._downloader.trouble(u'ERROR: unable to confirm age: %s' % compat_str(err))
|
self._downloader.trouble(u'ERROR: unable to confirm age: %s' % compat_str(err))
|
||||||
return
|
return
|
||||||
|
@ -399,7 +399,7 @@ def _real_extract(self, url):
|
||||||
self.report_video_subtitles_download(video_id)
|
self.report_video_subtitles_download(video_id)
|
||||||
request = compat_urllib_request.Request('http://video.google.com/timedtext?hl=en&type=list&v=%s' % video_id)
|
request = compat_urllib_request.Request('http://video.google.com/timedtext?hl=en&type=list&v=%s' % video_id)
|
||||||
try:
|
try:
|
||||||
srt_list = compat_urllib_request.urlopen(request).read()
|
srt_list = compat_urllib_request.urlopen(request).read().decode('utf-8')
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
||||||
raise Trouble(u'WARNING: unable to download video subtitles: %s' % compat_str(err))
|
raise Trouble(u'WARNING: unable to download video subtitles: %s' % compat_str(err))
|
||||||
srt_lang_list = re.findall(r'name="([^"]*)"[^>]+lang_code="([\w\-]+)"', srt_list)
|
srt_lang_list = re.findall(r'name="([^"]*)"[^>]+lang_code="([\w\-]+)"', srt_list)
|
||||||
|
@ -416,14 +416,14 @@ def _real_extract(self, url):
|
||||||
raise Trouble(u'WARNING: no closed captions found in the specified language')
|
raise Trouble(u'WARNING: no closed captions found in the specified language')
|
||||||
request = compat_urllib_request.Request('http://www.youtube.com/api/timedtext?lang=%s&name=%s&v=%s' % (srt_lang, srt_lang_list[srt_lang], video_id))
|
request = compat_urllib_request.Request('http://www.youtube.com/api/timedtext?lang=%s&name=%s&v=%s' % (srt_lang, srt_lang_list[srt_lang], video_id))
|
||||||
try:
|
try:
|
||||||
srt_xml = compat_urllib_request.urlopen(request).read()
|
srt_xml = compat_urllib_request.urlopen(request).read().decode('utf-8')
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
||||||
raise Trouble(u'WARNING: unable to download video subtitles: %s' % compat_str(err))
|
raise Trouble(u'WARNING: unable to download video subtitles: %s' % compat_str(err))
|
||||||
if not srt_xml:
|
if not srt_xml:
|
||||||
raise Trouble(u'WARNING: unable to download video subtitles')
|
raise Trouble(u'WARNING: unable to download video subtitles')
|
||||||
video_subtitles = self._closed_captions_xml_to_srt(srt_xml.decode('utf-8'))
|
video_subtitles = self._closed_captions_xml_to_srt(srt_xml)
|
||||||
except Trouble as trouble:
|
except Trouble as trouble:
|
||||||
self._downloader.trouble(trouble[0])
|
self._downloader.trouble(str(trouble))
|
||||||
|
|
||||||
if 'length_seconds' not in video_info:
|
if 'length_seconds' not in video_info:
|
||||||
self._downloader.trouble(u'WARNING: unable to extract video duration')
|
self._downloader.trouble(u'WARNING: unable to extract video duration')
|
||||||
|
@ -1715,7 +1715,7 @@ def _real_extract(self, url):
|
||||||
url = self._TEMPLATE_URL % (playlist_access, playlist_prefix, playlist_id, pagenum)
|
url = self._TEMPLATE_URL % (playlist_access, playlist_prefix, playlist_id, pagenum)
|
||||||
request = compat_urllib_request.Request(url)
|
request = compat_urllib_request.Request(url)
|
||||||
try:
|
try:
|
||||||
page = compat_urllib_request.urlopen(request).read().decode('utf8')
|
page = compat_urllib_request.urlopen(request).read().decode('utf-8')
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
||||||
self._downloader.trouble(u'ERROR: unable to download webpage: %s' % compat_str(err))
|
self._downloader.trouble(u'ERROR: unable to download webpage: %s' % compat_str(err))
|
||||||
return
|
return
|
||||||
|
@ -1844,7 +1844,7 @@ def _real_extract(self, url):
|
||||||
request = compat_urllib_request.Request(self._GDATA_URL % (username, self._GDATA_PAGE_SIZE, start_index))
|
request = compat_urllib_request.Request(self._GDATA_URL % (username, self._GDATA_PAGE_SIZE, start_index))
|
||||||
|
|
||||||
try:
|
try:
|
||||||
page = compat_urllib_request.urlopen(request).read()
|
page = compat_urllib_request.urlopen(request).read().decode('utf-8')
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
||||||
self._downloader.trouble(u'ERROR: unable to download webpage: %s' % compat_str(err))
|
self._downloader.trouble(u'ERROR: unable to download webpage: %s' % compat_str(err))
|
||||||
return
|
return
|
||||||
|
|
Loading…
Reference in a new issue