mirror of
https://github.com/yt-dlp/yt-dlp.git
synced 2024-11-26 02:55:17 +00:00
Merge branch 'master' of github.com:rg3/youtube-dl
This commit is contained in:
commit
a7c0f8602e
10 changed files with 106 additions and 77 deletions
19
README.md
19
README.md
|
@ -46,15 +46,16 @@ ## Filesystem Options:
|
||||||
-A, --auto-number number downloaded files starting from 00000
|
-A, --auto-number number downloaded files starting from 00000
|
||||||
-o, --output TEMPLATE output filename template. Use %(title)s to get the
|
-o, --output TEMPLATE output filename template. Use %(title)s to get the
|
||||||
title, %(uploader)s for the uploader name,
|
title, %(uploader)s for the uploader name,
|
||||||
%(autonumber)s to get an automatically incremented
|
%(uploader_id)s for the uploader nickname if
|
||||||
number, %(ext)s for the filename extension,
|
different, %(autonumber)s to get an automatically
|
||||||
%(upload_date)s for the upload date (YYYYMMDD),
|
incremented number, %(ext)s for the filename
|
||||||
%(extractor)s for the provider (youtube, metacafe,
|
extension, %(upload_date)s for the upload date
|
||||||
etc), %(id)s for the video id and %% for a literal
|
(YYYYMMDD), %(extractor)s for the provider
|
||||||
percent. Use - to output to stdout. Can also be
|
(youtube, metacafe, etc), %(id)s for the video id
|
||||||
used to download to a different directory, for
|
and %% for a literal percent. Use - to output to
|
||||||
example with -o '/my/downloads/%(uploader)s/%(title
|
stdout. Can also be used to download to a different
|
||||||
)s-%(id)s.%(ext)s' .
|
directory, for example with -o '/my/downloads/%(upl
|
||||||
|
oader)s/%(title)s-%(id)s.%(ext)s' .
|
||||||
--restrict-filenames Restrict filenames to only ASCII characters, and
|
--restrict-filenames Restrict filenames to only ASCII characters, and
|
||||||
avoid "&" and spaces in filenames
|
avoid "&" and spaces in filenames
|
||||||
-a, --batch-file FILE file containing URLs to download ('-' for stdin)
|
-a, --batch-file FILE file containing URLs to download ('-' for stdin)
|
||||||
|
|
|
@ -6,6 +6,7 @@
|
||||||
import json
|
import json
|
||||||
import unittest
|
import unittest
|
||||||
import sys
|
import sys
|
||||||
|
import hashlib
|
||||||
import socket
|
import socket
|
||||||
|
|
||||||
# Allow direct execution
|
# Allow direct execution
|
||||||
|
@ -24,12 +25,15 @@
|
||||||
proxy_handler = compat_urllib_request.ProxyHandler()
|
proxy_handler = compat_urllib_request.ProxyHandler()
|
||||||
opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
|
opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
|
||||||
compat_urllib_request.install_opener(opener)
|
compat_urllib_request.install_opener(opener)
|
||||||
socket.setdefaulttimeout(300) # 5 minutes should be enough (famous last words)
|
|
||||||
|
|
||||||
class FileDownloader(youtube_dl.FileDownloader):
|
class FileDownloader(youtube_dl.FileDownloader):
|
||||||
def __init__(self, *args, **kwargs):
|
def __init__(self, *args, **kwargs):
|
||||||
youtube_dl.FileDownloader.__init__(self, *args, **kwargs)
|
|
||||||
self.to_stderr = self.to_screen
|
self.to_stderr = self.to_screen
|
||||||
|
self.processed_info_dicts = []
|
||||||
|
return youtube_dl.FileDownloader.__init__(self, *args, **kwargs)
|
||||||
|
def process_info(self, info_dict):
|
||||||
|
self.processed_info_dicts.append(info_dict)
|
||||||
|
return youtube_dl.FileDownloader.process_info(self, info_dict)
|
||||||
|
|
||||||
def _file_md5(fn):
|
def _file_md5(fn):
|
||||||
with open(fn, 'rb') as f:
|
with open(fn, 'rb') as f:
|
||||||
|
@ -40,6 +44,7 @@ def _file_md5(fn):
|
||||||
with io.open(PARAMETERS_FILE, encoding='utf-8') as pf:
|
with io.open(PARAMETERS_FILE, encoding='utf-8') as pf:
|
||||||
parameters = json.load(pf)
|
parameters = json.load(pf)
|
||||||
|
|
||||||
|
|
||||||
class TestDownload(unittest.TestCase):
|
class TestDownload(unittest.TestCase):
|
||||||
def setUp(self):
|
def setUp(self):
|
||||||
self.parameters = parameters
|
self.parameters = parameters
|
||||||
|
@ -68,18 +73,28 @@ def test_template(self):
|
||||||
if 'skip' in test_case:
|
if 'skip' in test_case:
|
||||||
print('Skipping: {0}'.format(test_case['skip']))
|
print('Skipping: {0}'.format(test_case['skip']))
|
||||||
return
|
return
|
||||||
|
|
||||||
params = dict(self.parameters) # Duplicate it locally
|
params = dict(self.parameters) # Duplicate it locally
|
||||||
for p in test_case.get('params', {}):
|
for p in test_case.get('params', {}):
|
||||||
params[p] = test_case['params'][p]
|
params[p] = test_case['params'][p]
|
||||||
|
|
||||||
fd = FileDownloader(params)
|
fd = FileDownloader(params)
|
||||||
fd.add_info_extractor(ie())
|
fd.add_info_extractor(ie())
|
||||||
for ien in test_case.get('add_ie', []):
|
for ien in test_case.get('add_ie', []):
|
||||||
fd.add_info_extractor(getattr(youtube_dl.InfoExtractors, ien + 'IE')())
|
fd.add_info_extractor(getattr(youtube_dl.InfoExtractors, ien + 'IE')())
|
||||||
fd.download([test_case['url']])
|
fd.download([test_case['url']])
|
||||||
|
|
||||||
self.assertTrue(os.path.exists(test_case['file']))
|
self.assertTrue(os.path.exists(test_case['file']))
|
||||||
if 'md5' in test_case:
|
if 'md5' in test_case:
|
||||||
md5_for_file = _file_md5(test_case['file'])
|
md5_for_file = _file_md5(test_case['file'])
|
||||||
self.assertEqual(md5_for_file, test_case['md5'])
|
self.assertEqual(md5_for_file, test_case['md5'])
|
||||||
|
info_dict = fd.processed_info_dicts[0]
|
||||||
|
for (info_field, value) in test_case.get('info_dict', {}).items():
|
||||||
|
if value.startswith('md5:'):
|
||||||
|
md5_info_value = hashlib.md5(info_dict.get(info_field, '')).hexdigest()
|
||||||
|
self.assertEqual(value[3:], md5_info_value)
|
||||||
|
else:
|
||||||
|
self.assertEqual(value, info_dict.get(info_field))
|
||||||
|
|
||||||
return test_template
|
return test_template
|
||||||
|
|
||||||
|
|
|
@ -3,7 +3,6 @@
|
||||||
|
|
||||||
import json
|
import json
|
||||||
import os
|
import os
|
||||||
import socket
|
|
||||||
import sys
|
import sys
|
||||||
import unittest
|
import unittest
|
||||||
|
|
||||||
|
@ -22,7 +21,6 @@
|
||||||
proxy_handler = compat_urllib_request.ProxyHandler()
|
proxy_handler = compat_urllib_request.ProxyHandler()
|
||||||
opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
|
opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
|
||||||
compat_urllib_request.install_opener(opener)
|
compat_urllib_request.install_opener(opener)
|
||||||
socket.setdefaulttimeout(300) # 5 minutes should be enough (famous last words)
|
|
||||||
|
|
||||||
class FileDownloader(youtube_dl.FileDownloader):
|
class FileDownloader(youtube_dl.FileDownloader):
|
||||||
def __init__(self, *args, **kwargs):
|
def __init__(self, *args, **kwargs):
|
||||||
|
|
|
@ -2,7 +2,6 @@
|
||||||
|
|
||||||
import sys
|
import sys
|
||||||
import unittest
|
import unittest
|
||||||
import socket
|
|
||||||
import json
|
import json
|
||||||
|
|
||||||
# Allow direct execution
|
# Allow direct execution
|
||||||
|
@ -22,7 +21,6 @@
|
||||||
proxy_handler = compat_urllib_request.ProxyHandler()
|
proxy_handler = compat_urllib_request.ProxyHandler()
|
||||||
opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
|
opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
|
||||||
compat_urllib_request.install_opener(opener)
|
compat_urllib_request.install_opener(opener)
|
||||||
socket.setdefaulttimeout(300) # 5 minutes should be enough (famous last words)
|
|
||||||
|
|
||||||
class FakeDownloader(object):
|
class FakeDownloader(object):
|
||||||
def __init__(self):
|
def __init__(self):
|
||||||
|
|
|
@ -2,7 +2,6 @@
|
||||||
|
|
||||||
import sys
|
import sys
|
||||||
import unittest
|
import unittest
|
||||||
import socket
|
|
||||||
import json
|
import json
|
||||||
import io
|
import io
|
||||||
import hashlib
|
import hashlib
|
||||||
|
@ -24,7 +23,6 @@
|
||||||
proxy_handler = compat_urllib_request.ProxyHandler()
|
proxy_handler = compat_urllib_request.ProxyHandler()
|
||||||
opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
|
opener = compat_urllib_request.build_opener(proxy_handler, cookie_processor, YoutubeDLHandler())
|
||||||
compat_urllib_request.install_opener(opener)
|
compat_urllib_request.install_opener(opener)
|
||||||
socket.setdefaulttimeout(300) # 5 minutes should be enough (famous last words)
|
|
||||||
|
|
||||||
class FakeDownloader(object):
|
class FakeDownloader(object):
|
||||||
def __init__(self):
|
def __init__(self):
|
||||||
|
|
|
@ -2,7 +2,14 @@
|
||||||
{
|
{
|
||||||
"name": "Youtube",
|
"name": "Youtube",
|
||||||
"url": "http://www.youtube.com/watch?v=BaW_jenozKc",
|
"url": "http://www.youtube.com/watch?v=BaW_jenozKc",
|
||||||
"file": "BaW_jenozKc.mp4"
|
"file": "BaW_jenozKc.mp4",
|
||||||
|
"info_dict": {
|
||||||
|
"title": "youtube-dl test video \"'/\\ä↭𝕐",
|
||||||
|
"uploader": "Philipp Hagemeister",
|
||||||
|
"uploader_id": "phihag",
|
||||||
|
"upload_date": "20121002",
|
||||||
|
"description": "test chars: \"'/\\ä↭𝕐\n\nThis is a test video for youtube-dl.\n\nFor more information, contact phihag@phihag.de ."
|
||||||
|
}
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"name": "Dailymotion",
|
"name": "Dailymotion",
|
||||||
|
@ -30,9 +37,16 @@
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"name": "Vimeo",
|
"name": "Vimeo",
|
||||||
"md5": "60540a4ec7cc378ec84b919c0aed5023",
|
"md5": "8879b6cc097e987f02484baf890129e5",
|
||||||
"url": "http://vimeo.com/14160053",
|
"url": "http://vimeo.com/56015672",
|
||||||
"file": "14160053.mp4"
|
"file": "56015672.mp4",
|
||||||
|
"info_dict": {
|
||||||
|
"title": "youtube-dl test video - ★ \" ' 幸 / \\ ä ↭ 𝕐",
|
||||||
|
"uploader": "Filippo Valsorda",
|
||||||
|
"uploader_id": "user7108434",
|
||||||
|
"upload_date": "20121220",
|
||||||
|
"description": "This is a test case for youtube-dl.\nFor more information, see github.com/rg3/youtube-dl\nTest chars: ★ \" ' 幸 / \\ ä ↭ 𝕐"
|
||||||
|
}
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"name": "Soundcloud",
|
"name": "Soundcloud",
|
||||||
|
@ -81,7 +95,8 @@
|
||||||
"name": "Escapist",
|
"name": "Escapist",
|
||||||
"url": "http://www.escapistmagazine.com/videos/view/the-escapist-presents/6618-Breaking-Down-Baldurs-Gate",
|
"url": "http://www.escapistmagazine.com/videos/view/the-escapist-presents/6618-Breaking-Down-Baldurs-Gate",
|
||||||
"file": "6618-Breaking-Down-Baldurs-Gate.flv",
|
"file": "6618-Breaking-Down-Baldurs-Gate.flv",
|
||||||
"md5": "c6793dbda81388f4264c1ba18684a74d"
|
"md5": "c6793dbda81388f4264c1ba18684a74d",
|
||||||
|
"skip": "Fails with timeout on Travis"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"name": "GooglePlus",
|
"name": "GooglePlus",
|
||||||
|
|
|
@ -433,11 +433,8 @@ def process_info(self, info_dict):
|
||||||
try:
|
try:
|
||||||
srtfn = filename.rsplit('.', 1)[0] + u'.srt'
|
srtfn = filename.rsplit('.', 1)[0] + u'.srt'
|
||||||
self.report_writesubtitles(srtfn)
|
self.report_writesubtitles(srtfn)
|
||||||
srtfile = open(encodeFilename(srtfn), 'wb')
|
with io.open(encodeFilename(srtfn), 'w', encoding='utf-8') as srtfile:
|
||||||
try:
|
srtfile.write(info_dict['subtitles'])
|
||||||
srtfile.write(info_dict['subtitles'].encode('utf-8'))
|
|
||||||
finally:
|
|
||||||
srtfile.close()
|
|
||||||
except (OSError, IOError):
|
except (OSError, IOError):
|
||||||
self.trouble(u'ERROR: Cannot write subtitles file ' + descfn)
|
self.trouble(u'ERROR: Cannot write subtitles file ' + descfn)
|
||||||
return
|
return
|
||||||
|
|
|
@ -32,7 +32,7 @@ class InfoExtractor(object):
|
||||||
|
|
||||||
id: Video identifier.
|
id: Video identifier.
|
||||||
url: Final video URL.
|
url: Final video URL.
|
||||||
uploader: Nickname of the video uploader, unescaped.
|
uploader: Full name of the video uploader, unescaped.
|
||||||
upload_date: Video upload date (YYYYMMDD).
|
upload_date: Video upload date (YYYYMMDD).
|
||||||
title: Video title, unescaped.
|
title: Video title, unescaped.
|
||||||
ext: Video filename extension.
|
ext: Video filename extension.
|
||||||
|
@ -42,6 +42,7 @@ class InfoExtractor(object):
|
||||||
format: The video format, defaults to ext (used for --get-format)
|
format: The video format, defaults to ext (used for --get-format)
|
||||||
thumbnail: Full URL to a video thumbnail image.
|
thumbnail: Full URL to a video thumbnail image.
|
||||||
description: One-line video description.
|
description: One-line video description.
|
||||||
|
uploader_id: Nickname or id of the video uploader.
|
||||||
player_url: SWF Player URL (used for rtmpdump).
|
player_url: SWF Player URL (used for rtmpdump).
|
||||||
subtitles: The .srt file contents.
|
subtitles: The .srt file contents.
|
||||||
urlhandle: [internal] The urlHandle to be used to download the file,
|
urlhandle: [internal] The urlHandle to be used to download the file,
|
||||||
|
@ -219,6 +220,34 @@ def _closed_captions_xml_to_srt(self, xml_string):
|
||||||
srt += caption + '\n\n'
|
srt += caption + '\n\n'
|
||||||
return srt
|
return srt
|
||||||
|
|
||||||
|
def _extract_subtitles(self, video_id):
|
||||||
|
self.report_video_subtitles_download(video_id)
|
||||||
|
request = compat_urllib_request.Request('http://video.google.com/timedtext?hl=en&type=list&v=%s' % video_id)
|
||||||
|
try:
|
||||||
|
srt_list = compat_urllib_request.urlopen(request).read().decode('utf-8')
|
||||||
|
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
||||||
|
return (u'WARNING: unable to download video subtitles: %s' % compat_str(err), None)
|
||||||
|
srt_lang_list = re.findall(r'name="([^"]*)"[^>]+lang_code="([\w\-]+)"', srt_list)
|
||||||
|
srt_lang_list = dict((l[1], l[0]) for l in srt_lang_list)
|
||||||
|
if not srt_lang_list:
|
||||||
|
return (u'WARNING: video has no closed captions', None)
|
||||||
|
if self._downloader.params.get('subtitleslang', False):
|
||||||
|
srt_lang = self._downloader.params.get('subtitleslang')
|
||||||
|
elif 'en' in srt_lang_list:
|
||||||
|
srt_lang = 'en'
|
||||||
|
else:
|
||||||
|
srt_lang = list(srt_lang_list.keys())[0]
|
||||||
|
if not srt_lang in srt_lang_list:
|
||||||
|
return (u'WARNING: no closed captions found in the specified language', None)
|
||||||
|
request = compat_urllib_request.Request('http://www.youtube.com/api/timedtext?lang=%s&name=%s&v=%s' % (srt_lang, srt_lang_list[srt_lang], video_id))
|
||||||
|
try:
|
||||||
|
srt_xml = compat_urllib_request.urlopen(request).read().decode('utf-8')
|
||||||
|
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
||||||
|
return (u'WARNING: unable to download video subtitles: %s' % compat_str(err), None)
|
||||||
|
if not srt_xml:
|
||||||
|
return (u'WARNING: unable to download video subtitles', None)
|
||||||
|
return (None, self._closed_captions_xml_to_srt(srt_xml))
|
||||||
|
|
||||||
def _print_formats(self, formats):
|
def _print_formats(self, formats):
|
||||||
print('Available formats:')
|
print('Available formats:')
|
||||||
for x in formats:
|
for x in formats:
|
||||||
|
@ -356,10 +385,18 @@ def _real_extract(self, url):
|
||||||
|
|
||||||
# uploader
|
# uploader
|
||||||
if 'author' not in video_info:
|
if 'author' not in video_info:
|
||||||
self._downloader.trouble(u'ERROR: unable to extract uploader nickname')
|
self._downloader.trouble(u'ERROR: unable to extract uploader name')
|
||||||
return
|
return
|
||||||
video_uploader = compat_urllib_parse.unquote_plus(video_info['author'][0])
|
video_uploader = compat_urllib_parse.unquote_plus(video_info['author'][0])
|
||||||
|
|
||||||
|
# uploader_id
|
||||||
|
video_uploader_id = None
|
||||||
|
mobj = re.search(r'<link itemprop="url" href="http://www.youtube.com/user/([^"]+)">', video_webpage)
|
||||||
|
if mobj is not None:
|
||||||
|
video_uploader_id = mobj.group(1)
|
||||||
|
else:
|
||||||
|
self._downloader.trouble(u'WARNING: unable to extract uploader nickname')
|
||||||
|
|
||||||
# title
|
# title
|
||||||
if 'title' not in video_info:
|
if 'title' not in video_info:
|
||||||
self._downloader.trouble(u'ERROR: unable to extract video title')
|
self._downloader.trouble(u'ERROR: unable to extract video title')
|
||||||
|
@ -395,35 +432,9 @@ def _real_extract(self, url):
|
||||||
# closed captions
|
# closed captions
|
||||||
video_subtitles = None
|
video_subtitles = None
|
||||||
if self._downloader.params.get('writesubtitles', False):
|
if self._downloader.params.get('writesubtitles', False):
|
||||||
try:
|
(srt_error, video_subtitles) = self._extract_subtitles(video_id)
|
||||||
self.report_video_subtitles_download(video_id)
|
if srt_error:
|
||||||
request = compat_urllib_request.Request('http://video.google.com/timedtext?hl=en&type=list&v=%s' % video_id)
|
self._downloader.trouble(srt_error)
|
||||||
try:
|
|
||||||
srt_list = compat_urllib_request.urlopen(request).read().decode('utf-8')
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
raise Trouble(u'WARNING: unable to download video subtitles: %s' % compat_str(err))
|
|
||||||
srt_lang_list = re.findall(r'name="([^"]*)"[^>]+lang_code="([\w\-]+)"', srt_list)
|
|
||||||
srt_lang_list = dict((l[1], l[0]) for l in srt_lang_list)
|
|
||||||
if not srt_lang_list:
|
|
||||||
raise Trouble(u'WARNING: video has no closed captions')
|
|
||||||
if self._downloader.params.get('subtitleslang', False):
|
|
||||||
srt_lang = self._downloader.params.get('subtitleslang')
|
|
||||||
elif 'en' in srt_lang_list:
|
|
||||||
srt_lang = 'en'
|
|
||||||
else:
|
|
||||||
srt_lang = srt_lang_list.keys()[0]
|
|
||||||
if not srt_lang in srt_lang_list:
|
|
||||||
raise Trouble(u'WARNING: no closed captions found in the specified language')
|
|
||||||
request = compat_urllib_request.Request('http://www.youtube.com/api/timedtext?lang=%s&name=%s&v=%s' % (srt_lang, srt_lang_list[srt_lang], video_id))
|
|
||||||
try:
|
|
||||||
srt_xml = compat_urllib_request.urlopen(request).read().decode('utf-8')
|
|
||||||
except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
|
|
||||||
raise Trouble(u'WARNING: unable to download video subtitles: %s' % compat_str(err))
|
|
||||||
if not srt_xml:
|
|
||||||
raise Trouble(u'WARNING: unable to download video subtitles')
|
|
||||||
video_subtitles = self._closed_captions_xml_to_srt(srt_xml)
|
|
||||||
except Trouble as trouble:
|
|
||||||
self._downloader.trouble(str(trouble))
|
|
||||||
|
|
||||||
if 'length_seconds' not in video_info:
|
if 'length_seconds' not in video_info:
|
||||||
self._downloader.trouble(u'WARNING: unable to extract video duration')
|
self._downloader.trouble(u'WARNING: unable to extract video duration')
|
||||||
|
@ -443,7 +454,7 @@ def _real_extract(self, url):
|
||||||
elif 'url_encoded_fmt_stream_map' in video_info and len(video_info['url_encoded_fmt_stream_map']) >= 1:
|
elif 'url_encoded_fmt_stream_map' in video_info and len(video_info['url_encoded_fmt_stream_map']) >= 1:
|
||||||
url_data_strs = video_info['url_encoded_fmt_stream_map'][0].split(',')
|
url_data_strs = video_info['url_encoded_fmt_stream_map'][0].split(',')
|
||||||
url_data = [compat_parse_qs(uds) for uds in url_data_strs]
|
url_data = [compat_parse_qs(uds) for uds in url_data_strs]
|
||||||
url_data = filter(lambda ud: 'itag' in ud and 'url' in ud, url_data)
|
url_data = [ud for ud in url_data if 'itag' in ud and 'url' in ud]
|
||||||
url_map = dict((ud['itag'][0], ud['url'][0] + '&signature=' + ud['sig'][0]) for ud in url_data)
|
url_map = dict((ud['itag'][0], ud['url'][0] + '&signature=' + ud['sig'][0]) for ud in url_data)
|
||||||
|
|
||||||
format_limit = self._downloader.params.get('format_limit', None)
|
format_limit = self._downloader.params.get('format_limit', None)
|
||||||
|
@ -493,6 +504,7 @@ def _real_extract(self, url):
|
||||||
'id': video_id,
|
'id': video_id,
|
||||||
'url': video_real_url,
|
'url': video_real_url,
|
||||||
'uploader': video_uploader,
|
'uploader': video_uploader,
|
||||||
|
'uploader_id': video_uploader_id,
|
||||||
'upload_date': upload_date,
|
'upload_date': upload_date,
|
||||||
'title': video_title,
|
'title': video_title,
|
||||||
'ext': video_extension,
|
'ext': video_extension,
|
||||||
|
@ -992,8 +1004,9 @@ def _real_extract(self, url, new_video=True):
|
||||||
# Extract title
|
# Extract title
|
||||||
video_title = config["video"]["title"]
|
video_title = config["video"]["title"]
|
||||||
|
|
||||||
# Extract uploader
|
# Extract uploader and uploader_id
|
||||||
video_uploader = config["video"]["owner"]["name"]
|
video_uploader = config["video"]["owner"]["name"]
|
||||||
|
video_uploader_id = config["video"]["owner"]["url"].split('/')[-1]
|
||||||
|
|
||||||
# Extract video thumbnail
|
# Extract video thumbnail
|
||||||
video_thumbnail = config["video"]["thumbnail"]
|
video_thumbnail = config["video"]["thumbnail"]
|
||||||
|
@ -1005,9 +1018,9 @@ def _real_extract(self, url, new_video=True):
|
||||||
|
|
||||||
# Extract upload date
|
# Extract upload date
|
||||||
video_upload_date = None
|
video_upload_date = None
|
||||||
mobj = re.search(r'<span id="clip-date" style="display:none">[^:]*: (.*?)( \([^\(]*\))?</span>', webpage)
|
mobj = re.search(r'<meta itemprop="dateCreated" content="(\d{4})-(\d{2})-(\d{2})T', webpage)
|
||||||
if mobj is not None:
|
if mobj is not None:
|
||||||
video_upload_date = mobj.group(1)
|
video_upload_date = mobj.group(1) + mobj.group(2) + mobj.group(3)
|
||||||
|
|
||||||
# Vimeo specific: extract request signature and timestamp
|
# Vimeo specific: extract request signature and timestamp
|
||||||
sig = config['request']['signature']
|
sig = config['request']['signature']
|
||||||
|
@ -1045,6 +1058,7 @@ def _real_extract(self, url, new_video=True):
|
||||||
'id': video_id,
|
'id': video_id,
|
||||||
'url': video_url,
|
'url': video_url,
|
||||||
'uploader': video_uploader,
|
'uploader': video_uploader,
|
||||||
|
'uploader_id': video_uploader_id,
|
||||||
'upload_date': video_upload_date,
|
'upload_date': video_upload_date,
|
||||||
'title': video_title,
|
'title': video_title,
|
||||||
'ext': video_extension,
|
'ext': video_extension,
|
||||||
|
@ -2113,7 +2127,7 @@ def _real_extract(self, url):
|
||||||
video_description = video_info.get('description', 'No description available.')
|
video_description = video_info.get('description', 'No description available.')
|
||||||
|
|
||||||
url_map = video_info['video_urls']
|
url_map = video_info['video_urls']
|
||||||
if len(url_map.keys()) > 0:
|
if len(list(url_map.keys())) > 0:
|
||||||
# Decide which formats to download
|
# Decide which formats to download
|
||||||
req_format = self._downloader.params.get('format', None)
|
req_format = self._downloader.params.get('format', None)
|
||||||
format_limit = self._downloader.params.get('format_limit', None)
|
format_limit = self._downloader.params.get('format_limit', None)
|
||||||
|
@ -2973,7 +2987,7 @@ def _real_extract(self, url):
|
||||||
if file_url is not None:
|
if file_url is not None:
|
||||||
break # got it!
|
break # got it!
|
||||||
else:
|
else:
|
||||||
if req_format not in formats.keys():
|
if req_format not in list(formats.keys()):
|
||||||
self._downloader.trouble(u'ERROR: format is not available')
|
self._downloader.trouble(u'ERROR: format is not available')
|
||||||
return
|
return
|
||||||
|
|
||||||
|
@ -3272,7 +3286,7 @@ def _real_extract(self, url):
|
||||||
seed = config['data'][0]['seed']
|
seed = config['data'][0]['seed']
|
||||||
|
|
||||||
format = self._downloader.params.get('format', None)
|
format = self._downloader.params.get('format', None)
|
||||||
supported_format = config['data'][0]['streamfileids'].keys()
|
supported_format = list(config['data'][0]['streamfileids'].keys())
|
||||||
|
|
||||||
if format is None or format == 'best':
|
if format is None or format == 'best':
|
||||||
if 'hd2' in supported_format:
|
if 'hd2' in supported_format:
|
||||||
|
|
|
@ -307,7 +307,7 @@ def _find_term_columns():
|
||||||
action='store_true', dest='autonumber',
|
action='store_true', dest='autonumber',
|
||||||
help='number downloaded files starting from 00000', default=False)
|
help='number downloaded files starting from 00000', default=False)
|
||||||
filesystem.add_option('-o', '--output',
|
filesystem.add_option('-o', '--output',
|
||||||
dest='outtmpl', metavar='TEMPLATE', help='output filename template. Use %(title)s to get the title, %(uploader)s for the uploader name, %(autonumber)s to get an automatically incremented number, %(ext)s for the filename extension, %(upload_date)s for the upload date (YYYYMMDD), %(extractor)s for the provider (youtube, metacafe, etc), %(id)s for the video id and %% for a literal percent. Use - to output to stdout. Can also be used to download to a different directory, for example with -o \'/my/downloads/%(uploader)s/%(title)s-%(id)s.%(ext)s\' .')
|
dest='outtmpl', metavar='TEMPLATE', help='output filename template. Use %(title)s to get the title, %(uploader)s for the uploader name, %(uploader_id)s for the uploader nickname if different, %(autonumber)s to get an automatically incremented number, %(ext)s for the filename extension, %(upload_date)s for the upload date (YYYYMMDD), %(extractor)s for the provider (youtube, metacafe, etc), %(id)s for the video id and %% for a literal percent. Use - to output to stdout. Can also be used to download to a different directory, for example with -o \'/my/downloads/%(uploader)s/%(title)s-%(id)s.%(ext)s\' .')
|
||||||
filesystem.add_option('--restrict-filenames',
|
filesystem.add_option('--restrict-filenames',
|
||||||
action='store_true', dest='restrictfilenames',
|
action='store_true', dest='restrictfilenames',
|
||||||
help='Restrict filenames to only ASCII characters, and avoid "&" and spaces in filenames', default=False)
|
help='Restrict filenames to only ASCII characters, and avoid "&" and spaces in filenames', default=False)
|
||||||
|
@ -454,8 +454,8 @@ def _real_main():
|
||||||
if opts.list_extractors:
|
if opts.list_extractors:
|
||||||
for ie in extractors:
|
for ie in extractors:
|
||||||
print(ie.IE_NAME + (' (CURRENTLY BROKEN)' if not ie._WORKING else ''))
|
print(ie.IE_NAME + (' (CURRENTLY BROKEN)' if not ie._WORKING else ''))
|
||||||
matchedUrls = filter(lambda url: ie.suitable(url), all_urls)
|
matchedUrls = [url for url in all_urls if ie.suitable(url)]
|
||||||
all_urls = filter(lambda url: url not in matchedUrls, all_urls)
|
all_urls = [url for url in all_urls if url not in matchedUrls]
|
||||||
for mu in matchedUrls:
|
for mu in matchedUrls:
|
||||||
print(u' ' + mu)
|
print(u' ' + mu)
|
||||||
sys.exit(0)
|
sys.exit(0)
|
||||||
|
|
|
@ -298,7 +298,8 @@ def clean_html(html):
|
||||||
"""Clean an HTML snippet into a readable string"""
|
"""Clean an HTML snippet into a readable string"""
|
||||||
# Newline vs <br />
|
# Newline vs <br />
|
||||||
html = html.replace('\n', ' ')
|
html = html.replace('\n', ' ')
|
||||||
html = re.sub('\s*<\s*br\s*/?\s*>\s*', '\n', html)
|
html = re.sub(r'\s*<\s*br\s*/?\s*>\s*', '\n', html)
|
||||||
|
html = re.sub(r'<\s*/\s*p\s*>\s*<\s*p[^>]*>', '\n', html)
|
||||||
# Strip html tags
|
# Strip html tags
|
||||||
html = re.sub('<.*?>', '', html)
|
html = re.sub('<.*?>', '', html)
|
||||||
# Replace html entities
|
# Replace html entities
|
||||||
|
@ -465,14 +466,6 @@ def __init__(self, downloaded, expected):
|
||||||
self.downloaded = downloaded
|
self.downloaded = downloaded
|
||||||
self.expected = expected
|
self.expected = expected
|
||||||
|
|
||||||
|
|
||||||
class Trouble(Exception):
|
|
||||||
"""Trouble helper exception
|
|
||||||
|
|
||||||
This is an exception to be handled with
|
|
||||||
FileDownloader.trouble
|
|
||||||
"""
|
|
||||||
|
|
||||||
class YoutubeDLHandler(compat_urllib_request.HTTPHandler):
|
class YoutubeDLHandler(compat_urllib_request.HTTPHandler):
|
||||||
"""Handler for HTTP requests and responses.
|
"""Handler for HTTP requests and responses.
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue