import functools import urllib.parse import hashlib from .common import InfoExtractor from ..utils import ( ExtractorError, OnDemandPagedList, int_or_none, mimetype2ext, qualities, traverse_obj, unified_timestamp, ) class IwaraIE(InfoExtractor): IE_NAME = 'iwara' _VALID_URL = r'https?://(?:www\.|ecchi\.)?iwara\.tv/videos?/(?P[a-zA-Z0-9]+)' _TESTS = [{ # this video cannot be played because of migration 'only_matching': True, 'url': 'https://www.iwara.tv/video/k2ayoueezfkx6gvq', 'info_dict': { 'id': 'k2ayoueezfkx6gvq', 'ext': 'mp4', 'age_limit': 18, 'title': 'Defeat of Irybelda - アイリベルダの敗北', 'description': 'md5:70278abebe706647a8b4cb04cf23e0d3', 'uploader': 'Inwerwm', 'uploader_id': 'inwerwm', 'tags': 'count:1', 'like_count': 6133, 'view_count': 1050343, 'comment_count': 1, 'timestamp': 1677843869, 'modified_timestamp': 1679056362, }, }, { 'url': 'https://iwara.tv/video/1ywe1sbkqwumpdxz5/', 'md5': '20691ce1473ec2766c0788e14c60ce66', 'info_dict': { 'id': '1ywe1sbkqwumpdxz5', 'ext': 'mp4', 'age_limit': 18, 'title': 'Aponia 阿波尼亚SEX Party Tonight 手动脱衣 大奶 裸腿', 'description': 'md5:0c4c310f2e0592d68b9f771d348329ca', 'uploader': '龙也zZZ', 'uploader_id': 'user792540', 'tags': [ 'uncategorized' ], 'like_count': 1809, 'view_count': 25156, 'comment_count': 1, 'timestamp': 1678732213, 'modified_timestamp': 1679110271, }, }] def _extract_formats(self, video_id, fileurl): up = urllib.parse.urlparse(fileurl) q = urllib.parse.parse_qs(up.query) paths = up.path.rstrip('/').split('/') # https://github.com/yt-dlp/yt-dlp/issues/6549#issuecomment-1473771047 x_version = hashlib.sha1('_'.join((paths[-1], q['expires'][0], '5nFp9kmbNnHdAFhaqMvt')).encode()).hexdigest() preference = qualities(['preview', '360', '540', 'Source']) files = self._download_json(fileurl, video_id, headers={'X-Version': x_version}) for fmt in files: yield traverse_obj(fmt, { 'format_id': 'name', 'url': ('src', ('view', 'download'), {self._proto_relative_url}), 'ext': ('type', {mimetype2ext}), 'quality': ('name', {preference}), 'height': ('name', {int_or_none}), }, get_all=False) def _real_extract(self, url): video_id = self._match_id(url) video_data = self._download_json(f'https://api.iwara.tv/video/{video_id}', video_id, expected_status=lambda x: True) errmsg = video_data.get('message') # at this point we can actually get uploaded user info, but do we need it? if errmsg == 'errors.privateVideo': self.raise_login_required('Private video. Login if you have permissions to watch') elif errmsg: raise ExtractorError(f'Iwara says: {errmsg}') if not video_data.get('fileUrl'): if video_data.get('embedUrl'): return self.url_result(video_data.get('embedUrl')) raise ExtractorError('This video is unplayable', expected=True) return { 'id': video_id, 'age_limit': 18 if video_data.get('rating') == 'ecchi' else 0, # ecchi is 'sexy' in Japanese **traverse_obj(video_data, { 'title': 'title', 'description': 'body', 'uploader': ('user', 'name'), 'uploader_id': ('user', 'username'), 'tags': ('tags', ..., 'id'), 'like_count': 'numLikes', 'view_count': 'numViews', 'comment_count': 'numComments', 'timestamp': ('createdAt', {unified_timestamp}), 'modified_timestamp': ('updatedAt', {unified_timestamp}), 'thumbnail': ('file', 'id', {str}, { lambda x: f'https://files.iwara.tv/image/thumbnail/{x}/thumbnail-00.jpg'}), }), 'formats': list(self._extract_formats(video_id, video_data.get('fileUrl'))), } class IwaraUserIE(InfoExtractor): _VALID_URL = r'https?://(?:www\.)?iwara\.tv/profile/(?P[^/?#&]+)' IE_NAME = 'iwara:user' _PER_PAGE = 32 _TESTS = [{ 'url': 'https://iwara.tv/profile/user792540/videos', 'info_dict': { 'id': 'user792540', }, 'playlist_mincount': 80, }, { 'url': 'https://iwara.tv/profile/theblackbirdcalls/videos', 'info_dict': { 'id': 'theblackbirdcalls', }, 'playlist_mincount': 723, }, { 'url': 'https://iwara.tv/profile/user792540', 'only_matching': True, }, { 'url': 'https://iwara.tv/profile/theblackbirdcalls', 'only_matching': True, }] def _entries(self, playlist_id, user_id, page): videos = self._download_json( 'https://api.iwara.tv/videos', playlist_id, note=f'Downloading page {page}', query={ 'page': page, 'sort': 'date', 'user': user_id, 'limit': self._PER_PAGE, }) for x in traverse_obj(videos, ('results', ..., 'id')): yield self.url_result(f'https://iwara.tv/video/{x}') def _real_extract(self, url): playlist_id = self._match_id(url) user_info = self._download_json( f'https://api.iwara.tv/profile/{playlist_id}', playlist_id, note='Requesting user info') user_id = traverse_obj(user_info, ('user', 'id')) return self.playlist_result( OnDemandPagedList( functools.partial(self._entries, playlist_id, user_id), self._PER_PAGE), playlist_id, traverse_obj(user_info, ('user', 'name'))) class IwaraPlaylistIE(InfoExtractor): # the ID is an UUID but I don't think it's necessary to write concrete regex _VALID_URL = r'https?://(?:www\.)?iwara\.tv/playlist/(?P[0-9a-f-]+)' IE_NAME = 'iwara:playlist' _PER_PAGE = 32 _TESTS = [{ 'url': 'https://iwara.tv/playlist/458e5486-36a4-4ac0-b233-7e9eef01025f', 'info_dict': { 'id': '458e5486-36a4-4ac0-b233-7e9eef01025f', }, 'playlist_mincount': 3, }] def _entries(self, playlist_id, first_page, page): videos = self._download_json( 'https://api.iwara.tv/videos', playlist_id, f'Downloading page {page}', query={'page': page, 'limit': self._PER_PAGE}) if page else first_page for x in traverse_obj(videos, ('results', ..., 'id')): yield self.url_result(f'https://iwara.tv/video/{x}') def _real_extract(self, url): playlist_id = self._match_id(url) page_0 = self._download_json( f'https://api.iwara.tv/playlist/{playlist_id}?page=0&limit={self._PER_PAGE}', playlist_id, note='Requesting playlist info') return self.playlist_result( OnDemandPagedList( functools.partial(self._entries, playlist_id, page_0), self._PER_PAGE), playlist_id, traverse_obj(page_0, ('title', 'name')))