0
0
Fork 0
mirror of https://github.com/yt-dlp/yt-dlp.git synced 2024-11-04 23:35:04 +00:00

More calls to trouble changed to report_error

This commit is contained in:
Jaime Marquínez Ferrándiz 2013-04-23 11:31:37 +02:00
parent 9edb0916f4
commit 613bf66939
2 changed files with 28 additions and 28 deletions

View file

@ -406,10 +406,10 @@ def prepare_filename(self, info_dict):
filename = self.params['outtmpl'] % template_dict filename = self.params['outtmpl'] % template_dict
return filename return filename
except KeyError as err: except KeyError as err:
self.trouble(u'ERROR: Erroneous output template') self.report_error(u'Erroneous output template')
return None return None
except ValueError as err: except ValueError as err:
self.trouble(u'ERROR: Insufficient system charset ' + repr(preferredencoding())) self.report_error(u'Insufficient system charset ' + repr(preferredencoding()))
return None return None
def _match_entry(self, info_dict): def _match_entry(self, info_dict):
@ -468,16 +468,16 @@ def extract_info(self, url, download = True, ie_name = None):
results.append(self.process_ie_result(ie_result, download)) results.append(self.process_ie_result(ie_result, download))
return results return results
except ExtractorError as de: # An error we somewhat expected except ExtractorError as de: # An error we somewhat expected
self.trouble(u'ERROR: ' + compat_str(de), de.format_traceback()) self.report_error(compat_str(de), de.format_traceback())
break break
except Exception as e: except Exception as e:
if self.params.get('ignoreerrors', False): if self.params.get('ignoreerrors', False):
self.trouble(u'ERROR: ' + compat_str(e), tb=compat_str(traceback.format_exc())) self.report_error(compat_str(e), tb=compat_str(traceback.format_exc()))
break break
else: else:
raise raise
if not suitable_found: if not suitable_found:
self.trouble(u'ERROR: no suitable InfoExtractor: %s' % url) self.report_error(u'no suitable InfoExtractor: %s' % url)
def process_ie_result(self, ie_result, download = True): def process_ie_result(self, ie_result, download = True):
""" """
@ -636,7 +636,7 @@ def process_info(self, info_dict):
with io.open(encodeFilename(sub_filename), 'w', encoding='utf-8') as subfile: with io.open(encodeFilename(sub_filename), 'w', encoding='utf-8') as subfile:
subfile.write(sub) subfile.write(sub)
except (OSError, IOError): except (OSError, IOError):
self.trouble(u'ERROR: Cannot write subtitles file ' + descfn) self.report_error(u'Cannot write subtitles file ' + descfn)
return return
if self.params.get('onlysubtitles', False): if self.params.get('onlysubtitles', False):
return return
@ -683,7 +683,7 @@ def download(self, url_list):
#It also downloads the videos #It also downloads the videos
videos = self.extract_info(url) videos = self.extract_info(url)
except UnavailableVideoError: except UnavailableVideoError:
self.trouble(u'\nERROR: unable to download video') self.report_error(u'unable to download video')
except MaxDownloadsReached: except MaxDownloadsReached:
self.to_screen(u'[info] Maximum number of downloaded files reached.') self.to_screen(u'[info] Maximum number of downloaded files reached.')
raise raise

View file

@ -1224,7 +1224,7 @@ def grep_webpage(self, url, regex, regexFlags, matchTuples):
for (i, key, err) in matchTuples: for (i, key, err) in matchTuples:
if mobj.group(i) is None: if mobj.group(i) is None:
self._downloader.trouble(err) self._downloader.report_error(err)
return return
else: else:
info[key] = mobj.group(i) info[key] = mobj.group(i)
@ -1238,7 +1238,7 @@ def extractLiveStream(self, url):
r'src="(.*?/videothek_js.*?\.js)', r'src="(.*?/videothek_js.*?\.js)',
0, 0,
[ [
(1, 'url', u'ERROR: Invalid URL: %s' % url) (1, 'url', u'Invalid URL: %s' % url)
] ]
) )
http_host = url.split('/')[2] http_host = url.split('/')[2]
@ -1250,9 +1250,9 @@ def extractLiveStream(self, url):
'(rtmp://.*?)\'', '(rtmp://.*?)\'',
re.DOTALL, re.DOTALL,
[ [
(1, 'path', u'ERROR: could not extract video path: %s' % url), (1, 'path', u'could not extract video path: %s' % url),
(2, 'player', u'ERROR: could not extract video player: %s' % url), (2, 'player', u'could not extract video player: %s' % url),
(3, 'url', u'ERROR: could not extract video url: %s' % url) (3, 'url', u'could not extract video url: %s' % url)
] ]
) )
video_url = u'%s/%s' % (info.get('url'), info.get('path')) video_url = u'%s/%s' % (info.get('url'), info.get('path'))
@ -1264,7 +1264,7 @@ def extractPlus7Stream(self, url):
r'param name="movie".*?videorefFileUrl=(http[^\'"&]*)', r'param name="movie".*?videorefFileUrl=(http[^\'"&]*)',
0, 0,
[ [
(1, 'url', u'ERROR: Invalid URL: %s' % url) (1, 'url', u'Invalid URL: %s' % url)
] ]
) )
next_url = compat_urllib_parse.unquote(info.get('url')) next_url = compat_urllib_parse.unquote(info.get('url'))
@ -1273,7 +1273,7 @@ def extractPlus7Stream(self, url):
r'<video lang="%s" ref="(http[^\'"&]*)' % video_lang, r'<video lang="%s" ref="(http[^\'"&]*)' % video_lang,
0, 0,
[ [
(1, 'url', u'ERROR: Could not find <video> tag: %s' % url) (1, 'url', u'Could not find <video> tag: %s' % url)
] ]
) )
next_url = compat_urllib_parse.unquote(info.get('url')) next_url = compat_urllib_parse.unquote(info.get('url'))
@ -1286,10 +1286,10 @@ def extractPlus7Stream(self, url):
'<url quality="hd">(.*?)</url>', '<url quality="hd">(.*?)</url>',
re.DOTALL, re.DOTALL,
[ [
(1, 'id', u'ERROR: could not extract video id: %s' % url), (1, 'id', u'could not extract video id: %s' % url),
(2, 'title', u'ERROR: could not extract video title: %s' % url), (2, 'title', u'could not extract video title: %s' % url),
(3, 'date', u'ERROR: could not extract video date: %s' % url), (3, 'date', u'could not extract video date: %s' % url),
(4, 'url', u'ERROR: could not extract video url: %s' % url) (4, 'url', u'could not extract video url: %s' % url)
] ]
) )
@ -1518,7 +1518,7 @@ def _get_n_results(self, query, n):
api_response = json.loads(data)['data'] api_response = json.loads(data)['data']
if not 'items' in api_response: if not 'items' in api_response:
self._downloader.trouble(u'[youtube] No video results') self._downloader.report_error(u'[youtube] No video results')
return return
new_ids = list(video['id'] for video in api_response['items']) new_ids = list(video['id'] for video in api_response['items'])
@ -2787,7 +2787,7 @@ def report_resolve(self, video_id):
def _real_extract(self, url): def _real_extract(self, url):
mobj = re.match(self._VALID_URL, url) mobj = re.match(self._VALID_URL, url)
if mobj is None: if mobj is None:
self._downloader.trouble(u'ERROR: invalid URL: %s' % url) self._downloader.report_error(u'invalid URL: %s' % url)
return return
# extract uploader (which is in the url) # extract uploader (which is in the url)
@ -2805,14 +2805,14 @@ def _real_extract(self, url):
info_json_bytes = compat_urllib_request.urlopen(request).read() info_json_bytes = compat_urllib_request.urlopen(request).read()
info_json = info_json_bytes.decode('utf-8') info_json = info_json_bytes.decode('utf-8')
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err: except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
self._downloader.trouble(u'ERROR: unable to download video webpage: %s' % compat_str(err)) self._downloader.report_error(u'unable to download video webpage: %s' % compat_str(err))
return return
videos = [] videos = []
info = json.loads(info_json) info = json.loads(info_json)
if 'errors' in info: if 'errors' in info:
for err in info['errors']: for err in info['errors']:
self._downloader.trouble(u'ERROR: unable to download video webpage: %s' % compat_str(err['error_message'])) self._downloader.report_error(u'unable to download video webpage: %s' % compat_str(err['error_message']))
return return
for track in info['tracks']: for track in info['tracks']:
@ -2825,7 +2825,7 @@ def _real_extract(self, url):
stream_json_bytes = compat_urllib_request.urlopen(request).read() stream_json_bytes = compat_urllib_request.urlopen(request).read()
stream_json = stream_json_bytes.decode('utf-8') stream_json = stream_json_bytes.decode('utf-8')
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err: except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
self._downloader.trouble(u'ERROR: unable to download stream definitions: %s' % compat_str(err)) self._downloader.report_error(u'unable to download stream definitions: %s' % compat_str(err))
return return
streams = json.loads(stream_json) streams = json.loads(stream_json)
@ -3173,7 +3173,7 @@ def _real_extract(self, url):
format = ext + '-' + rendition.attrib['width'] + 'x' + rendition.attrib['height'] + '_' + rendition.attrib['bitrate'] format = ext + '-' + rendition.attrib['width'] + 'x' + rendition.attrib['height'] + '_' + rendition.attrib['bitrate']
video_url = rendition.find('./src').text video_url = rendition.find('./src').text
except KeyError: except KeyError:
self._downloader.trouble('Invalid rendition field.') self._downloader.report_error('Invalid rendition field.')
return return
info = { info = {
@ -3620,7 +3620,7 @@ def _real_extract(self, url):
if not m: if not m:
m = re.search(r'<title>(?P<title>[^<]+?)</title>', webpage) m = re.search(r'<title>(?P<title>[^<]+?)</title>', webpage)
if not m: if not m:
self._downloader.trouble(u'Cannot find video title') self._downloader.report_error(u'Cannot find video title')
title = clean_html(m.group('title')) title = clean_html(m.group('title'))
m = re.search(r'<meta property="og:description" content="(?P<desc>.*?)"', webpage) m = re.search(r'<meta property="og:description" content="(?P<desc>.*?)"', webpage)
@ -3726,7 +3726,7 @@ def _real_extract(self, url):
else: else:
ext = 'flv' ext = 'flv'
else: else:
self._downloader.trouble(u'ERROR: Cannot find video url for %s' % video_id) self._downloader.report_error(u'Cannot find video url for %s' % video_id)
return return
_title = r"""<title>(.*)</title>""" _title = r"""<title>(.*)</title>"""
@ -4246,7 +4246,7 @@ class LiveLeakIE(InfoExtractor):
def _real_extract(self, url): def _real_extract(self, url):
mobj = re.match(self._VALID_URL, url) mobj = re.match(self._VALID_URL, url)
if mobj is None: if mobj is None:
self._downloader.trouble(u'ERROR: invalid URL: %s' % url) self._downloader.report_error(u'invalid URL: %s' % url)
return return
video_id = mobj.group('video_id') video_id = mobj.group('video_id')
@ -4261,7 +4261,7 @@ def _real_extract(self, url):
m = re.search(r'<meta property="og:title" content="(?P<title>.*?)"', webpage) m = re.search(r'<meta property="og:title" content="(?P<title>.*?)"', webpage)
if not m: if not m:
self._downloader.trouble(u'Cannot find video title') self._downloader.report_error(u'Cannot find video title')
title = unescapeHTML(m.group('title')).replace('LiveLeak.com -', '').strip() title = unescapeHTML(m.group('title')).replace('LiveLeak.com -', '').strip()
m = re.search(r'<meta property="og:description" content="(?P<desc>.*?)"', webpage) m = re.search(r'<meta property="og:description" content="(?P<desc>.*?)"', webpage)