0
0
Fork 0
mirror of https://github.com/yt-dlp/yt-dlp.git synced 2024-11-09 00:15:06 +00:00
yt-dlp/yt_dlp/extractor/twitcasting.py

307 lines
12 KiB
Python
Raw Normal View History

import base64
import itertools
2018-10-26 18:40:44 +00:00
import re
2021-01-01 12:26:37 +00:00
from .common import InfoExtractor
from ..dependencies import websockets
2021-01-01 12:26:37 +00:00
from ..utils import (
ExtractorError,
UserNotLive,
clean_html,
2021-01-01 12:26:37 +00:00
float_or_none,
get_element_by_class,
get_element_by_id,
parse_duration,
qualities,
2021-01-01 12:26:37 +00:00
str_to_int,
traverse_obj,
try_get,
2021-01-01 12:26:37 +00:00
unified_timestamp,
urlencode_postdata,
urljoin,
2021-01-01 12:26:37 +00:00
)
2018-10-26 18:40:44 +00:00
class TwitCastingIE(InfoExtractor):
_VALID_URL = r'https?://(?:[^/?#]+\.)?twitcasting\.tv/(?P<uploader_id>[^/?#]+)/(?:movie|twplayer)/(?P<id>\d+)'
_M3U8_HEADERS = {
'Origin': 'https://twitcasting.tv',
'Referer': 'https://twitcasting.tv/',
}
_TESTS = [{
2018-10-26 18:40:44 +00:00
'url': 'https://twitcasting.tv/ivetesangalo/movie/2357609',
'md5': '745243cad58c4681dc752490f7540d7f',
'info_dict': {
'id': '2357609',
'ext': 'mp4',
'title': 'Live #2357609',
2018-10-26 18:40:44 +00:00
'uploader_id': 'ivetesangalo',
2021-01-01 12:26:37 +00:00
'description': 'Twitter Oficial da cantora brasileira Ivete Sangalo.',
2018-10-26 18:40:44 +00:00
'thumbnail': r're:^https?://.*\.jpg$',
2021-01-01 12:26:37 +00:00
'upload_date': '20110822',
'timestamp': 1313978424,
2021-01-01 12:26:37 +00:00
'duration': 32,
'view_count': int,
},
'params': {
'skip_download': True,
},
}, {
'url': 'https://twitcasting.tv/mttbernardini/movie/3689740',
'info_dict': {
'id': '3689740',
'ext': 'mp4',
'title': 'Live playing something #3689740',
'uploader_id': 'mttbernardini',
'description': 'md5:1dc7efa2f1ab932fcd119265cebeec69',
'thumbnail': r're:^https?://.*\.jpg$',
'upload_date': '20120211',
'timestamp': 1328995624,
2021-01-01 12:26:37 +00:00
'duration': 681,
'view_count': int,
},
'params': {
'skip_download': True,
'videopassword': 'abc',
},
}, {
'url': 'https://twitcasting.tv/loft_heaven/movie/685979292',
'info_dict': {
'id': '685979292',
'ext': 'mp4',
'title': '【無料配信】南波一海のhear/here “ナタリー望月哲さんに聞く編集と「渋谷系狂騒曲」”',
'uploader_id': 'loft_heaven',
'description': 'md5:3a0c7b53019df987ce545c935538bacf',
'upload_date': '20210604',
'timestamp': 1622802114,
'thumbnail': r're:^https?://.*\.jpg$',
'duration': 6964,
'view_count': int,
},
'params': {
'skip_download': True,
},
}]
2018-10-26 18:40:44 +00:00
def _parse_data_movie_playlist(self, dmp, video_id):
# attempt 1: parse as JSON directly
try:
return self._parse_json(dmp, video_id)
except ExtractorError:
pass
# attempt 2: decode reversed base64
decoded = base64.b64decode(dmp[::-1])
return self._parse_json(decoded, video_id)
2018-10-26 18:40:44 +00:00
def _real_extract(self, url):
uploader_id, video_id = self._match_valid_url(url).groups()
2018-10-26 18:40:44 +00:00
webpage, urlh = self._download_webpage_handle(url, video_id)
video_password = self.get_param('videopassword')
request_data = None
if video_password:
request_data = urlencode_postdata({
'password': video_password,
**self._hidden_inputs(webpage),
}, encoding='utf-8')
webpage, urlh = self._download_webpage_handle(
url, video_id, data=request_data,
headers={'Origin': 'https://twitcasting.tv'},
note='Trying video password')
if urlh.url != url and request_data:
webpage = self._download_webpage(
urlh.url, video_id, data=request_data,
headers={'Origin': 'https://twitcasting.tv'},
note='Retrying authentication')
# has to check here as the first request can contain password input form even if the password is correct
if re.search(r'<form\s+method="POST">\s*<input\s+[^>]+?name="password"', webpage):
raise ExtractorError('This video is protected by a password, use the --video-password option', expected=True)
2018-10-26 18:40:44 +00:00
title = (clean_html(get_element_by_id('movietitle', webpage))
or self._html_search_meta(['og:title', 'twitter:title'], webpage, fatal=True))
video_js_data = try_get(
webpage,
lambda x: self._parse_data_movie_playlist(self._search_regex(
r'data-movie-playlist=\'([^\']+?)\'',
x, 'movie playlist', default=None), video_id)['2'], list)
thumbnail = traverse_obj(video_js_data, (0, 'thumbnailUrl')) or self._og_search_thumbnail(webpage)
2021-01-01 12:26:37 +00:00
description = clean_html(get_element_by_id(
'authorcomment', webpage)) or self._html_search_meta(
['description', 'og:description', 'twitter:description'], webpage)
duration = (try_get(video_js_data, lambda x: sum(float_or_none(y.get('duration')) for y in x) / 1000)
or parse_duration(clean_html(get_element_by_class('tw-player-duration-time', webpage))))
2021-01-01 12:26:37 +00:00
view_count = str_to_int(self._search_regex(
(r'Total\s*:\s*Views\s*([\d,]+)', r'総視聴者\s*:\s*([\d,]+)\s*</'), webpage, 'views', None))
2021-01-01 12:26:37 +00:00
timestamp = unified_timestamp(self._search_regex(
r'data-toggle="true"[^>]+datetime="([^"]+)"',
webpage, 'datetime', None))
stream_server_data = self._download_json(
'https://twitcasting.tv/streamserver.php?target=%s&mode=client' % uploader_id, video_id,
'Downloading live info', fatal=False)
is_live = any(f'data-{x}' in webpage for x in ['is-onlive="true"', 'live-type="live"', 'status="online"'])
if not traverse_obj(stream_server_data, 'llfmp4') and is_live:
self.raise_login_required(method='cookies')
base_dict = {
2018-10-26 18:40:44 +00:00
'title': title,
'description': description,
'thumbnail': thumbnail,
2021-01-01 12:26:37 +00:00
'timestamp': timestamp,
2018-10-26 18:40:44 +00:00
'uploader_id': uploader_id,
2021-01-01 12:26:37 +00:00
'duration': duration,
'view_count': view_count,
'is_live': is_live,
2018-10-26 18:40:44 +00:00
}
def find_dmu(x):
data_movie_url = self._search_regex(
r'data-movie-url=(["\'])(?P<url>(?:(?!\1).)+)\1',
x, 'm3u8 url', group='url', default=None)
if data_movie_url:
return [data_movie_url]
m3u8_urls = (try_get(webpage, find_dmu, list)
or traverse_obj(video_js_data, (..., 'source', 'url'))
or ([f'https://twitcasting.tv/{uploader_id}/metastream.m3u8'] if is_live else None))
if not m3u8_urls:
raise ExtractorError('Failed to get m3u8 playlist')
if is_live:
m3u8_url = m3u8_urls[0]
formats = self._extract_m3u8_formats(
m3u8_url, video_id, ext='mp4', m3u8_id='hls',
live=True, headers=self._M3U8_HEADERS)
if traverse_obj(stream_server_data, ('hls', 'source')):
formats.extend(self._extract_m3u8_formats(
m3u8_url, video_id, ext='mp4', m3u8_id='source',
live=True, query={'mode': 'source'},
note='Downloading source quality m3u8',
headers=self._M3U8_HEADERS, fatal=False))
if websockets:
qq = qualities(['base', 'mobilesource', 'main'])
streams = traverse_obj(stream_server_data, ('llfmp4', 'streams')) or {}
for mode, ws_url in streams.items():
formats.append({
'url': ws_url,
'format_id': 'ws-%s' % mode,
'ext': 'mp4',
'quality': qq(mode),
'source_preference': -10,
# TwitCasting simply sends moof atom directly over WS
'protocol': 'websocket_frag',
})
infodict = {
2022-11-17 05:10:03 +00:00
'formats': formats,
'_format_sort_fields': ('source', ),
}
elif len(m3u8_urls) == 1:
formats = self._extract_m3u8_formats(
m3u8_urls[0], video_id, 'mp4', headers=self._M3U8_HEADERS)
infodict = {
# No problem here since there's only one manifest
'formats': formats,
'http_headers': self._M3U8_HEADERS,
}
else:
infodict = {
'_type': 'multi_video',
'entries': [{
'id': f'{video_id}-{num}',
'url': m3u8_url,
'ext': 'mp4',
# Requesting the manifests here will cause download to fail.
# So use ffmpeg instead. See: https://github.com/yt-dlp/yt-dlp/issues/382
'protocol': 'm3u8',
'http_headers': self._M3U8_HEADERS,
**base_dict,
} for (num, m3u8_url) in enumerate(m3u8_urls)],
}
return {
'id': video_id,
**base_dict,
**infodict,
}
class TwitCastingLiveIE(InfoExtractor):
_VALID_URL = r'https?://(?:[^/?#]+\.)?twitcasting\.tv/(?P<id>[^/?#]+)/?(?:[#?]|$)'
_TESTS = [{
'url': 'https://twitcasting.tv/ivetesangalo',
'only_matching': True,
}, {
'url': 'https://twitcasting.tv/c:unusedlive',
'expected_exception': 'UserNotLive',
}]
def _real_extract(self, url):
uploader_id = self._match_id(url)
self.to_screen(
'Downloading live video of user {0}. '
'Pass "https://twitcasting.tv/{0}/show" to download the history'.format(uploader_id))
is_live = traverse_obj(self._download_json(
f'https://frontendapi.twitcasting.tv/watch/user/{uploader_id}',
uploader_id, 'Checking live status', data=b'', fatal=False), ('is_live', {bool}))
if is_live is False: # only raise here if API response was as expected
raise UserNotLive(video_id=uploader_id)
# Use /show/ page so that password-protected and members-only livestreams can be found
webpage = self._download_webpage(
f'https://twitcasting.tv/{uploader_id}/show/', uploader_id, 'Downloading live history')
is_live = is_live or self._search_regex(
r'(?s)(<span\s*class="tw-movie-thumbnail2-badge"\s*data-status="live">\s*LIVE)',
webpage, 'is live?', default=False)
# Current live is always the first match
current_live = self._search_regex(
r'(?s)<a\s+class="tw-movie-thumbnail2"\s+href="/[^/"]+/movie/(?P<video_id>\d+)"',
webpage, 'current live ID', default=None, group='video_id')
if not is_live or not current_live:
raise UserNotLive(video_id=uploader_id)
return self.url_result(f'https://twitcasting.tv/{uploader_id}/movie/{current_live}', TwitCastingIE)
class TwitCastingUserIE(InfoExtractor):
_VALID_URL = r'https?://(?:[^/?#]+\.)?twitcasting\.tv/(?P<id>[^/?#]+)/(:?show|archive)/?(?:[#?]|$)'
_TESTS = [{
'url': 'https://twitcasting.tv/natsuiromatsuri/archive/',
'info_dict': {
'id': 'natsuiromatsuri',
'title': 'natsuiromatsuri - Live History',
},
'playlist_mincount': 235,
}, {
'url': 'https://twitcasting.tv/noriyukicas/show',
'only_matching': True,
}]
def _entries(self, uploader_id):
base_url = next_url = 'https://twitcasting.tv/%s/show' % uploader_id
for page_num in itertools.count(1):
webpage = self._download_webpage(
next_url, uploader_id, query={'filter': 'watchable'}, note='Downloading page %d' % page_num)
matches = re.finditer(
r'(?s)<a\s+class="tw-movie-thumbnail2"\s+href="(?P<url>/[^/"]+/movie/\d+)"', webpage)
for mobj in matches:
yield self.url_result(urljoin(base_url, mobj.group('url')))
next_url = self._search_regex(
r'<a href="(/%s/show/%d-\d+)[?"]' % (re.escape(uploader_id), page_num),
webpage, 'next url', default=None)
next_url = urljoin(base_url, next_url)
if not next_url:
return
def _real_extract(self, url):
uploader_id = self._match_id(url)
return self.playlist_result(
self._entries(uploader_id), uploader_id, '%s - Live History' % uploader_id)