mirror of
https://github.com/yt-dlp/yt-dlp.git
synced 2024-11-15 05:33:05 +00:00
Compare commits
No commits in common. "6db9c4d57d033fb22c94a2e6f1ecf0207e700b4c" and "ae72962643aa43b1e1037293b226acc71fb34297" have entirely different histories.
6db9c4d57d
...
ae72962643
@ -196,7 +196,15 @@ def expect_dict(self, got_dict, expected_dict):
|
||||
|
||||
def sanitize_got_info_dict(got_dict):
|
||||
IGNORED_FIELDS = (
|
||||
*YoutubeDL._format_fields,
|
||||
# Format keys
|
||||
'url', 'manifest_url', 'format', 'format_id', 'format_note', 'width', 'height', 'resolution',
|
||||
'dynamic_range', 'tbr', 'abr', 'acodec', 'asr', 'vbr', 'fps', 'vcodec', 'container', 'filesize',
|
||||
'filesize_approx', 'player_url', 'protocol', 'fragment_base_url', 'fragments', 'preference',
|
||||
'language', 'language_preference', 'quality', 'source_preference', 'http_headers',
|
||||
'stretched_ratio', 'no_resume', 'has_drm', 'downloader_options',
|
||||
|
||||
# RTMP formats
|
||||
'page_url', 'app', 'play_path', 'tc_url', 'flash_version', 'rtmp_live', 'rtmp_conn', 'rtmp_protocol', 'rtmp_real_time',
|
||||
|
||||
# Lists
|
||||
'formats', 'thumbnails', 'subtitles', 'automatic_captions', 'comments', 'entries',
|
||||
|
@ -931,7 +931,7 @@ class TestYoutubeDL(unittest.TestCase):
|
||||
res = get_videos()
|
||||
self.assertEqual(res, ['1', '2'])
|
||||
|
||||
def f(v, incomplete):
|
||||
def f(v):
|
||||
if v['id'] == '1':
|
||||
return None
|
||||
else:
|
||||
|
@ -513,16 +513,6 @@ class YoutubeDL(object):
|
||||
'track_number', 'disc_number', 'release_year',
|
||||
))
|
||||
|
||||
_format_fields = {
|
||||
# NB: Keep in sync with the docstring of extractor/common.py
|
||||
'url', 'manifest_url', 'ext', 'format', 'format_id', 'format_note',
|
||||
'width', 'height', 'resolution', 'dynamic_range', 'tbr', 'abr', 'acodec', 'asr',
|
||||
'vbr', 'fps', 'vcodec', 'container', 'filesize', 'filesize_approx',
|
||||
'player_url', 'protocol', 'fragment_base_url', 'fragments', 'is_from_start',
|
||||
'preference', 'language', 'language_preference', 'quality', 'source_preference',
|
||||
'http_headers', 'stretched_ratio', 'no_resume', 'has_drm', 'downloader_options',
|
||||
'page_url', 'app', 'play_path', 'tc_url', 'flash_version', 'rtmp_live', 'rtmp_conn', 'rtmp_protocol', 'rtmp_real_time'
|
||||
}
|
||||
_format_selection_exts = {
|
||||
'audio': {'m4a', 'mp3', 'ogg', 'aac'},
|
||||
'video': {'mp4', 'flv', 'webm', '3gp'},
|
||||
@ -2551,7 +2541,7 @@ class YoutubeDL(object):
|
||||
|
||||
info_dict, _ = self.pre_process(info_dict)
|
||||
|
||||
if self._match_entry(info_dict, incomplete=self._format_fields) is not None:
|
||||
if self._match_entry(info_dict) is not None:
|
||||
return info_dict
|
||||
|
||||
self.post_extract(info_dict)
|
||||
|
@ -13,7 +13,6 @@ from ..compat import (
|
||||
)
|
||||
from ..postprocessor.ffmpeg import FFmpegPostProcessor, EXT_TO_OUT_FORMATS
|
||||
from ..utils import (
|
||||
classproperty,
|
||||
cli_option,
|
||||
cli_valueless_option,
|
||||
cli_bool_option,
|
||||
@ -74,23 +73,17 @@ class ExternalFD(FragmentFD):
|
||||
def get_basename(cls):
|
||||
return cls.__name__[:-2].lower()
|
||||
|
||||
@classproperty
|
||||
def EXE_NAME(cls):
|
||||
return cls.get_basename()
|
||||
|
||||
@property
|
||||
def exe(self):
|
||||
return self.EXE_NAME
|
||||
return self.get_basename()
|
||||
|
||||
@classmethod
|
||||
def available(cls, path=None):
|
||||
path = check_executable(
|
||||
cls.EXE_NAME if path in (None, cls.get_basename()) else path,
|
||||
[cls.AVAILABLE_OPT])
|
||||
if not path:
|
||||
return False
|
||||
cls.exe = path
|
||||
return path
|
||||
path = check_executable(path or cls.get_basename(), [cls.AVAILABLE_OPT])
|
||||
if path:
|
||||
cls.exe = path
|
||||
return path
|
||||
return False
|
||||
|
||||
@classmethod
|
||||
def supports(cls, info_dict):
|
||||
@ -113,7 +106,7 @@ class ExternalFD(FragmentFD):
|
||||
|
||||
def _configuration_args(self, keys=None, *args, **kwargs):
|
||||
return _configuration_args(
|
||||
self.get_basename(), self.params.get('external_downloader_args'), self.EXE_NAME,
|
||||
self.get_basename(), self.params.get('external_downloader_args'), self.get_basename(),
|
||||
keys, *args, **kwargs)
|
||||
|
||||
def _call_downloader(self, tmpfilename, info_dict):
|
||||
@ -313,7 +306,10 @@ class Aria2cFD(ExternalFD):
|
||||
|
||||
class HttpieFD(ExternalFD):
|
||||
AVAILABLE_OPT = '--version'
|
||||
EXE_NAME = 'http'
|
||||
|
||||
@classmethod
|
||||
def available(cls, path=None):
|
||||
return super().available(path or 'http')
|
||||
|
||||
def _make_cmd(self, tmpfilename, info_dict):
|
||||
cmd = ['http', '--download', '--output', tmpfilename, info_dict['url']]
|
||||
@ -514,13 +510,11 @@ class AVconvFD(FFmpegFD):
|
||||
pass
|
||||
|
||||
|
||||
_BY_NAME = {
|
||||
klass.get_basename(): klass
|
||||
_BY_NAME = dict(
|
||||
(klass.get_basename(), klass)
|
||||
for name, klass in globals().items()
|
||||
if name.endswith('FD') and name not in ('ExternalFD', 'FragmentFD')
|
||||
}
|
||||
|
||||
_BY_EXE = {klass.EXE_NAME: klass for klass in _BY_NAME.values()}
|
||||
)
|
||||
|
||||
|
||||
def list_external_downloaders():
|
||||
@ -532,4 +526,4 @@ def get_external_downloader(external_downloader):
|
||||
downloader . """
|
||||
# Drop .exe extension on Windows
|
||||
bn = os.path.splitext(os.path.basename(external_downloader))[0]
|
||||
return _BY_NAME.get(bn, _BY_EXE.get(bn))
|
||||
return _BY_NAME.get(bn)
|
||||
|
@ -212,6 +212,7 @@ class FC2LiveIE(InfoExtractor):
|
||||
'Accept': '*/*',
|
||||
'User-Agent': std_headers['User-Agent'],
|
||||
})
|
||||
ws.__enter__()
|
||||
|
||||
self.write_debug('[debug] Sending HLS server request')
|
||||
|
||||
|
@ -1,32 +1,55 @@
|
||||
# coding: utf-8
|
||||
from __future__ import unicode_literals
|
||||
|
||||
import re
|
||||
import json
|
||||
import uuid
|
||||
import random
|
||||
import urllib.parse
|
||||
import re
|
||||
|
||||
from .common import InfoExtractor
|
||||
from ..compat import compat_str
|
||||
from ..compat import (
|
||||
compat_kwargs,
|
||||
compat_str,
|
||||
compat_urlparse,
|
||||
compat_urllib_request,
|
||||
)
|
||||
from ..utils import (
|
||||
ExtractorError,
|
||||
int_or_none,
|
||||
strip_or_none,
|
||||
try_get,
|
||||
smuggle_url,
|
||||
unsmuggle_url,
|
||||
url_or_none,
|
||||
)
|
||||
|
||||
|
||||
class ViuBaseIE(InfoExtractor):
|
||||
def _call_api(self, path, *args, headers={}, **kwargs):
|
||||
def _real_initialize(self):
|
||||
viu_auth_res = self._request_webpage(
|
||||
'https://www.viu.com/api/apps/v2/authenticate', None,
|
||||
'Requesting Viu auth', query={
|
||||
'acct': 'test',
|
||||
'appid': 'viu_desktop',
|
||||
'fmt': 'json',
|
||||
'iid': 'guest',
|
||||
'languageid': 'default',
|
||||
'platform': 'desktop',
|
||||
'userid': 'guest',
|
||||
'useridtype': 'guest',
|
||||
'ver': '1.0'
|
||||
}, headers=self.geo_verification_headers())
|
||||
self._auth_token = viu_auth_res.info()['X-VIU-AUTH']
|
||||
|
||||
def _call_api(self, path, *args, **kwargs):
|
||||
headers = self.geo_verification_headers()
|
||||
headers.update({
|
||||
'X-VIU-AUTH': self._auth_token
|
||||
})
|
||||
headers.update(kwargs.get('headers', {}))
|
||||
kwargs['headers'] = headers
|
||||
response = self._download_json(
|
||||
f'https://www.viu.com/api/{path}', *args, **kwargs,
|
||||
headers={**self.geo_verification_headers(), **headers})['response']
|
||||
'https://www.viu.com/api/' + path, *args,
|
||||
**compat_kwargs(kwargs))['response']
|
||||
if response.get('status') != 'success':
|
||||
raise ExtractorError(f'{self.IE_NAME} said: {response["message"]}', expected=True)
|
||||
raise ExtractorError('%s said: %s' % (
|
||||
self.IE_NAME, response['message']), expected=True)
|
||||
return response
|
||||
|
||||
|
||||
@ -78,7 +101,6 @@ class ViuIE(ViuBaseIE):
|
||||
tdirforwhole = video_data.get('tdirforwhole')
|
||||
# #EXT-X-BYTERANGE is not supported by native hls downloader
|
||||
# and ffmpeg (#10955)
|
||||
# FIXME: It is supported in yt-dlp
|
||||
# hls_file = video_data.get('hlsfile')
|
||||
hls_file = video_data.get('jwhlsfile')
|
||||
if url_path and tdirforwhole and hls_file:
|
||||
@ -205,63 +227,42 @@ class ViuOTTIE(InfoExtractor):
|
||||
'zh-cn': 2,
|
||||
'en-us': 3,
|
||||
}
|
||||
|
||||
_user_token = None
|
||||
_auth_codes = {}
|
||||
_user_info = None
|
||||
|
||||
def _detect_error(self, response):
|
||||
code = try_get(response, lambda x: x['status']['code'])
|
||||
if code and code > 0:
|
||||
code = response.get('status', {}).get('code')
|
||||
if code > 0:
|
||||
message = try_get(response, lambda x: x['status']['message'])
|
||||
raise ExtractorError(f'{self.IE_NAME} said: {message} ({code})', expected=True)
|
||||
return response.get('data') or {}
|
||||
raise ExtractorError('%s said: %s (%s)' % (
|
||||
self.IE_NAME, message, code), expected=True)
|
||||
return response['data']
|
||||
|
||||
def _raise_login_required(self):
|
||||
raise ExtractorError(
|
||||
'This video requires login. '
|
||||
'Specify --username and --password or --netrc (machine: %s) '
|
||||
'to provide account credentials.' % self._NETRC_MACHINE,
|
||||
expected=True)
|
||||
|
||||
def _login(self, country_code, video_id):
|
||||
if self._user_token is None:
|
||||
if not self._user_info:
|
||||
username, password = self._get_login_info()
|
||||
if username is None:
|
||||
if username is None or password is None:
|
||||
return
|
||||
headers = {
|
||||
'Authorization': f'Bearer {self._auth_codes[country_code]}',
|
||||
'Content-Type': 'application/json'
|
||||
}
|
||||
data = self._download_json(
|
||||
'https://api-gateway-global.viu.com/api/account/validate',
|
||||
video_id, 'Validating email address', headers=headers,
|
||||
data=json.dumps({
|
||||
'principal': username,
|
||||
'provider': 'email'
|
||||
}).encode())
|
||||
if not data.get('exists'):
|
||||
raise ExtractorError('Invalid email address')
|
||||
|
||||
data = self._download_json(
|
||||
'https://api-gateway-global.viu.com/api/auth/login',
|
||||
video_id, 'Logging in', headers=headers,
|
||||
compat_urllib_request.Request(
|
||||
'https://www.viu.com/ott/%s/index.php' % country_code, method='POST'),
|
||||
video_id, 'Logging in', errnote=False, fatal=False,
|
||||
query={'r': 'user/login'},
|
||||
data=json.dumps({
|
||||
'email': username,
|
||||
'username': username,
|
||||
'password': password,
|
||||
'provider': 'email',
|
||||
'platform_flag_label': 'web',
|
||||
}).encode())
|
||||
self._detect_error(data)
|
||||
self._user_token = data.get('identity')
|
||||
# need to update with valid user's token else will throw an error again
|
||||
self._auth_codes[country_code] = data.get('token')
|
||||
return self._user_token
|
||||
self._user_info = self._detect_error(data)['user']
|
||||
|
||||
def _get_token(self, country_code, video_id):
|
||||
rand = ''.join(random.choice('0123456789') for _ in range(10))
|
||||
return self._download_json(
|
||||
f'https://api-gateway-global.viu.com/api/auth/token?v={rand}000', video_id,
|
||||
headers={'Content-Type': 'application/json'}, note='Getting bearer token',
|
||||
data=json.dumps({
|
||||
'countryCode': country_code.upper(),
|
||||
'platform': 'browser',
|
||||
'platformFlagLabel': 'web',
|
||||
'language': 'en',
|
||||
'uuid': str(uuid.uuid4()),
|
||||
'carrierId': '0'
|
||||
}).encode('utf-8'))['token']
|
||||
return self._user_info
|
||||
|
||||
def _real_extract(self, url):
|
||||
url, idata = unsmuggle_url(url, {})
|
||||
@ -278,16 +279,16 @@ class ViuOTTIE(InfoExtractor):
|
||||
query['area_id'] = area_id
|
||||
|
||||
product_data = self._download_json(
|
||||
f'http://www.viu.com/ott/{country_code}/index.php', video_id,
|
||||
'http://www.viu.com/ott/%s/index.php' % country_code, video_id,
|
||||
'Downloading video info', query=query)['data']
|
||||
|
||||
video_data = product_data.get('current_product')
|
||||
if not video_data:
|
||||
self.raise_geo_restricted()
|
||||
raise ExtractorError('This video is not available in your region.', expected=True)
|
||||
|
||||
series_id = video_data.get('series_id')
|
||||
if self._yes_playlist(series_id, video_id, idata):
|
||||
series = product_data.get('series') or {}
|
||||
series = product_data.get('series', {})
|
||||
product = series.get('product')
|
||||
if product:
|
||||
entries = []
|
||||
@ -295,10 +296,14 @@ class ViuOTTIE(InfoExtractor):
|
||||
item_id = entry.get('product_id')
|
||||
if not item_id:
|
||||
continue
|
||||
item_id = compat_str(item_id)
|
||||
entries.append(self.url_result(
|
||||
smuggle_url(f'http://www.viu.com/ott/{country_code}/{lang_code}/vod/{item_id}/',
|
||||
{'force_noplaylist': True}),
|
||||
ViuOTTIE, str(item_id), entry.get('synopsis', '').strip()))
|
||||
smuggle_url(
|
||||
'http://www.viu.com/ott/%s/%s/vod/%s/' % (country_code, lang_code, item_id),
|
||||
{'force_noplaylist': True}), # prevent infinite recursion
|
||||
'ViuOTT',
|
||||
item_id,
|
||||
entry.get('synopsis', '').strip()))
|
||||
|
||||
return self.playlist_result(entries, series_id, series.get('name'), series.get('description'))
|
||||
|
||||
@ -307,65 +312,69 @@ class ViuOTTIE(InfoExtractor):
|
||||
'ccs_product_id': video_data['ccs_product_id'],
|
||||
'language_flag_id': self._LANGUAGE_FLAG.get(lang_code.lower()) or '3',
|
||||
}
|
||||
|
||||
def download_playback():
|
||||
stream_data = self._download_json(
|
||||
'https://api-gateway-global.viu.com/api/playback/distribute',
|
||||
video_id=video_id, query=query, fatal=False, note='Downloading stream info',
|
||||
headers={
|
||||
'Authorization': f'Bearer {self._auth_codes[country_code]}',
|
||||
'Referer': url,
|
||||
'Origin': url
|
||||
})
|
||||
return self._detect_error(stream_data).get('stream')
|
||||
|
||||
if not self._auth_codes.get(country_code):
|
||||
self._auth_codes[country_code] = self._get_token(country_code, video_id)
|
||||
|
||||
stream_data = None
|
||||
headers = {
|
||||
'Referer': url,
|
||||
'Origin': url,
|
||||
}
|
||||
try:
|
||||
stream_data = download_playback()
|
||||
stream_data = self._download_json(
|
||||
'https://d1k2us671qcoau.cloudfront.net/distribute_web_%s.php' % country_code,
|
||||
video_id, 'Downloading stream info', query=query, headers=headers)
|
||||
stream_data = self._detect_error(stream_data)['stream']
|
||||
except (ExtractorError, KeyError):
|
||||
token = self._login(country_code, video_id)
|
||||
if token is not None:
|
||||
query['identity'] = token
|
||||
else:
|
||||
# preview is limited to 3min for non-members. But we can try to bypass it
|
||||
duration_limit, query['duration'] = True, '180'
|
||||
try:
|
||||
stream_data = download_playback()
|
||||
except (ExtractorError, KeyError):
|
||||
if token is not None:
|
||||
raise
|
||||
self.raise_login_required(method='password')
|
||||
stream_data = None
|
||||
if video_data.get('user_level', 0) > 0:
|
||||
user = self._login(country_code, video_id)
|
||||
if user:
|
||||
query['identity'] = user['identity']
|
||||
stream_data = self._download_json(
|
||||
'https://d1k2us671qcoau.cloudfront.net/distribute_web_%s.php' % country_code,
|
||||
video_id, 'Downloading stream info', query=query, headers=headers)
|
||||
stream_data = self._detect_error(stream_data).get('stream')
|
||||
else:
|
||||
# preview is limited to 3min for non-members
|
||||
# try to bypass the duration limit
|
||||
duration_limit = True
|
||||
query['duration'] = '180'
|
||||
stream_data = self._download_json(
|
||||
'https://d1k2us671qcoau.cloudfront.net/distribute_web_%s.php' % country_code,
|
||||
video_id, 'Downloading stream info', query=query, headers=headers)
|
||||
try:
|
||||
stream_data = self._detect_error(stream_data)['stream']
|
||||
except (ExtractorError, KeyError): # if still not working, give up
|
||||
self._raise_login_required()
|
||||
|
||||
if not stream_data:
|
||||
raise ExtractorError('Cannot get stream info', expected=True)
|
||||
|
||||
stream_sizes = stream_data.get('size', {})
|
||||
formats = []
|
||||
for vid_format, stream_url in (stream_data.get('url') or {}).items():
|
||||
height = int(self._search_regex(r's(\d+)p', vid_format, 'height', default=None))
|
||||
for vid_format, stream_url in stream_data.get('url', {}).items():
|
||||
height = int_or_none(self._search_regex(
|
||||
r's(\d+)p', vid_format, 'height', default=None))
|
||||
|
||||
# bypass preview duration limit
|
||||
if duration_limit:
|
||||
stream_url = urllib.parse.urlparse(stream_url)
|
||||
stream_url = compat_urlparse.urlparse(stream_url)
|
||||
query = dict(compat_urlparse.parse_qsl(stream_url.query, keep_blank_values=True))
|
||||
time_duration = int_or_none(video_data.get('time_duration'))
|
||||
query.update({
|
||||
'duration': video_data.get('time_duration') or '9999999',
|
||||
'duration': time_duration if time_duration > 0 else '9999999',
|
||||
'duration_start': '0',
|
||||
})
|
||||
stream_url = stream_url._replace(query=urllib.parse.urlencode(dict(
|
||||
urllib.parse.parse_qsl(stream_url.query, keep_blank_values=True)))).geturl()
|
||||
stream_url = stream_url._replace(query=compat_urlparse.urlencode(query)).geturl()
|
||||
|
||||
formats.append({
|
||||
'format_id': vid_format,
|
||||
'url': stream_url,
|
||||
'height': height,
|
||||
'ext': 'mp4',
|
||||
'filesize': try_get(stream_data, lambda x: x['size'][vid_format], int)
|
||||
'filesize': int_or_none(stream_sizes.get(vid_format))
|
||||
})
|
||||
self._sort_formats(formats)
|
||||
|
||||
subtitles = {}
|
||||
for sub in video_data.get('subtitle') or []:
|
||||
for sub in video_data.get('subtitle', []):
|
||||
sub_url = sub.get('url')
|
||||
if not sub_url:
|
||||
continue
|
||||
@ -374,16 +383,17 @@ class ViuOTTIE(InfoExtractor):
|
||||
'ext': 'srt',
|
||||
})
|
||||
|
||||
title = strip_or_none(video_data.get('synopsis'))
|
||||
title = video_data['synopsis'].strip()
|
||||
|
||||
return {
|
||||
'id': video_id,
|
||||
'title': title,
|
||||
'description': video_data.get('description'),
|
||||
'series': try_get(product_data, lambda x: x['series']['name']),
|
||||
'series': product_data.get('series', {}).get('name'),
|
||||
'episode': title,
|
||||
'episode_number': int_or_none(video_data.get('number')),
|
||||
'duration': int_or_none(stream_data.get('duration')),
|
||||
'thumbnail': url_or_none(video_data.get('cover_image_url')),
|
||||
'thumbnail': video_data.get('cover_image_url'),
|
||||
'formats': formats,
|
||||
'subtitles': subtitles,
|
||||
}
|
||||
|
@ -465,18 +465,19 @@ def create_parser():
|
||||
metavar='COUNT', dest='max_views', default=None, type=int,
|
||||
help=optparse.SUPPRESS_HELP)
|
||||
selection.add_option(
|
||||
'--match-filters',
|
||||
metavar='FILTER', dest='match_filter', action='append',
|
||||
'--match-filter',
|
||||
metavar='FILTER', dest='match_filter', default=None,
|
||||
help=(
|
||||
'Generic video filter. Any field (see "OUTPUT TEMPLATE") can be compared with a '
|
||||
'number or a string using the operators defined in "Filtering formats". '
|
||||
'You can also simply specify a field to match if the field is present, '
|
||||
'use "!field" to check if the field is not present, and "&" to check multiple conditions. '
|
||||
'Use a "\\" to escape "&" or quotes if needed. If used multiple times, '
|
||||
'the filter matches if atleast one of the conditions are met. Eg: --match-filter '
|
||||
'!is_live --match-filter "like_count>?100 & description~=\'(?i)\\bcats \\& dogs\\b\'" '
|
||||
'matches only videos that are not live OR those that have a like count more than 100 '
|
||||
'(or the like field is not available) and also has a description '
|
||||
'You can also simply specify a field to match if the field is present '
|
||||
'and "!field" to check if the field is not present. In addition, '
|
||||
'Python style regular expression matching can be done using "~=", '
|
||||
'and multiple filters can be checked with "&". '
|
||||
'Use a "\\" to escape "&" or quotes if needed. Eg: --match-filter '
|
||||
'"!is_live & like_count>?100 & description~=\'(?i)\\bcats \\& dogs\\b\'" '
|
||||
'matches only videos that are not live, has a like count more than 100 '
|
||||
'(or the like field is not available), and also has a description '
|
||||
'that contains the phrase "cats & dogs" (ignoring case)'))
|
||||
selection.add_option(
|
||||
'--no-match-filter',
|
||||
|
@ -1,18 +1,13 @@
|
||||
from __future__ import unicode_literals
|
||||
|
||||
import functools
|
||||
import itertools
|
||||
import json
|
||||
import os
|
||||
import time
|
||||
import urllib.error
|
||||
|
||||
from ..compat import compat_str
|
||||
from ..utils import (
|
||||
_configuration_args,
|
||||
encodeFilename,
|
||||
network_exceptions,
|
||||
PostProcessingError,
|
||||
sanitized_Request,
|
||||
write_string,
|
||||
)
|
||||
|
||||
@ -68,7 +63,7 @@ class PostProcessor(metaclass=PostProcessorMetaClass):
|
||||
@classmethod
|
||||
def pp_key(cls):
|
||||
name = cls.__name__[:-2]
|
||||
return name[6:] if name[:6].lower() == 'ffmpeg' else name
|
||||
return compat_str(name[6:]) if name[:6].lower() == 'ffmpeg' else name
|
||||
|
||||
def to_screen(self, text, prefix=True, *args, **kwargs):
|
||||
tag = '[%s] ' % self.PP_NAME if prefix else ''
|
||||
@ -185,28 +180,6 @@ class PostProcessor(metaclass=PostProcessorMetaClass):
|
||||
progress_template.get('postprocess-title') or 'yt-dlp %(progress._default_template)s',
|
||||
progress_dict))
|
||||
|
||||
def _download_json(self, url, *, expected_http_errors=(404,)):
|
||||
# While this is not an extractor, it behaves similar to one and
|
||||
# so obey extractor_retries and sleep_interval_requests
|
||||
max_retries = self.get_param('extractor_retries', 3)
|
||||
sleep_interval = self.get_param('sleep_interval_requests') or 0
|
||||
|
||||
self.write_debug(f'{self.PP_NAME} query: {url}')
|
||||
for retries in itertools.count():
|
||||
try:
|
||||
rsp = self._downloader.urlopen(sanitized_Request(url))
|
||||
return json.loads(rsp.read().decode(rsp.info().get_param('charset') or 'utf-8'))
|
||||
except network_exceptions as e:
|
||||
if isinstance(e, urllib.error.HTTPError) and e.code in expected_http_errors:
|
||||
return None
|
||||
if retries < max_retries:
|
||||
self.report_warning(f'{e}. Retrying...')
|
||||
if sleep_interval > 0:
|
||||
self.to_screen(f'Sleeping {sleep_interval} seconds ...')
|
||||
time.sleep(sleep_interval)
|
||||
continue
|
||||
raise PostProcessingError(f'Unable to communicate with {self.PP_NAME} API: {e}')
|
||||
|
||||
|
||||
class AudioConversionError(PostProcessingError):
|
||||
pass
|
||||
|
@ -1,9 +1,12 @@
|
||||
from hashlib import sha256
|
||||
import itertools
|
||||
import json
|
||||
import re
|
||||
import time
|
||||
|
||||
from .ffmpeg import FFmpegPostProcessor
|
||||
from ..compat import compat_urllib_parse_urlencode
|
||||
from ..compat import compat_urllib_parse_urlencode, compat_HTTPError
|
||||
from ..utils import PostProcessingError, network_exceptions, sanitized_Request
|
||||
|
||||
|
||||
class SponsorBlockPP(FFmpegPostProcessor):
|
||||
@ -91,7 +94,28 @@ class SponsorBlockPP(FFmpegPostProcessor):
|
||||
'categories': json.dumps(self._categories),
|
||||
'actionTypes': json.dumps(['skip', 'poi'])
|
||||
})
|
||||
for d in self._download_json(url) or []:
|
||||
self.write_debug(f'SponsorBlock query: {url}')
|
||||
for d in self._get_json(url):
|
||||
if d['videoID'] == video_id:
|
||||
return d['segments']
|
||||
return []
|
||||
|
||||
def _get_json(self, url):
|
||||
# While this is not an extractor, it behaves similar to one and
|
||||
# so obey extractor_retries and sleep_interval_requests
|
||||
max_retries = self.get_param('extractor_retries', 3)
|
||||
sleep_interval = self.get_param('sleep_interval_requests') or 0
|
||||
for retries in itertools.count():
|
||||
try:
|
||||
rsp = self._downloader.urlopen(sanitized_Request(url))
|
||||
return json.loads(rsp.read().decode(rsp.info().get_param('charset') or 'utf-8'))
|
||||
except network_exceptions as e:
|
||||
if isinstance(e, compat_HTTPError) and e.code == 404:
|
||||
return []
|
||||
if retries < max_retries:
|
||||
self.report_warning(f'{e}. Retrying...')
|
||||
if sleep_interval > 0:
|
||||
self.to_screen(f'Sleeping {sleep_interval} seconds ...')
|
||||
time.sleep(sleep_interval)
|
||||
continue
|
||||
raise PostProcessingError(f'Unable to communicate with SponsorBlock API: {e}')
|
||||
|
@ -3545,11 +3545,6 @@ def _match_one(filter_part, dct, incomplete):
|
||||
'=': operator.eq,
|
||||
}
|
||||
|
||||
if isinstance(incomplete, bool):
|
||||
is_incomplete = lambda _: incomplete
|
||||
else:
|
||||
is_incomplete = lambda k: k in incomplete
|
||||
|
||||
operator_rex = re.compile(r'''(?x)\s*
|
||||
(?P<key>[a-z_]+)
|
||||
\s*(?P<negation>!\s*)?(?P<op>%s)(?P<none_inclusive>\s*\?)?\s*
|
||||
@ -3588,7 +3583,7 @@ def _match_one(filter_part, dct, incomplete):
|
||||
if numeric_comparison is not None and m['op'] in STRING_OPERATORS:
|
||||
raise ValueError('Operator %s only supports string values!' % m['op'])
|
||||
if actual_value is None:
|
||||
return is_incomplete(m['key']) or m['none_inclusive']
|
||||
return incomplete or m['none_inclusive']
|
||||
return op(actual_value, comparison_value if numeric_comparison is None else numeric_comparison)
|
||||
|
||||
UNARY_OPERATORS = {
|
||||
@ -3603,7 +3598,7 @@ def _match_one(filter_part, dct, incomplete):
|
||||
if m:
|
||||
op = UNARY_OPERATORS[m.group('op')]
|
||||
actual_value = dct.get(m.group('key'))
|
||||
if is_incomplete(m.group('key')) and actual_value is None:
|
||||
if incomplete and actual_value is None:
|
||||
return True
|
||||
return op(actual_value)
|
||||
|
||||
@ -3611,29 +3606,24 @@ def _match_one(filter_part, dct, incomplete):
|
||||
|
||||
|
||||
def match_str(filter_str, dct, incomplete=False):
|
||||
""" Filter a dictionary with a simple string syntax.
|
||||
@returns Whether the filter passes
|
||||
@param incomplete Set of keys that is expected to be missing from dct.
|
||||
Can be True/False to indicate all/none of the keys may be missing.
|
||||
All conditions on incomplete keys pass if the key is missing
|
||||
""" Filter a dictionary with a simple string syntax. Returns True (=passes filter) or false
|
||||
When incomplete, all conditions passes on missing fields
|
||||
"""
|
||||
return all(
|
||||
_match_one(filter_part.replace(r'\&', '&'), dct, incomplete)
|
||||
for filter_part in re.split(r'(?<!\\)&', filter_str))
|
||||
|
||||
|
||||
def match_filter_func(filters):
|
||||
if not filters:
|
||||
def match_filter_func(filter_str):
|
||||
if filter_str is None:
|
||||
return None
|
||||
filters = variadic(filters)
|
||||
|
||||
def _match_func(info_dict, *args, **kwargs):
|
||||
if any(match_str(f, info_dict, *args, **kwargs) for f in filters):
|
||||
if match_str(filter_str, info_dict, *args, **kwargs):
|
||||
return None
|
||||
else:
|
||||
video_title = info_dict.get('title') or info_dict.get('id') or 'video'
|
||||
filter_str = ') | ('.join(map(str.strip, filters))
|
||||
return f'{video_title} does not pass filter ({filter_str}), skipping ..'
|
||||
video_title = info_dict.get('title', info_dict.get('id', 'video'))
|
||||
return '%s does not pass filter %s, skipping ..' % (video_title, filter_str)
|
||||
return _match_func
|
||||
|
||||
|
||||
@ -5444,18 +5434,15 @@ class Config:
|
||||
class WebSocketsWrapper():
|
||||
"""Wraps websockets module to use in non-async scopes"""
|
||||
|
||||
def __init__(self, url, headers=None, connect=True):
|
||||
def __init__(self, url, headers=None):
|
||||
self.loop = asyncio.events.new_event_loop()
|
||||
self.conn = compat_websockets.connect(
|
||||
url, extra_headers=headers, ping_interval=None,
|
||||
close_timeout=float('inf'), loop=self.loop, ping_timeout=float('inf'))
|
||||
if connect:
|
||||
self.__enter__()
|
||||
atexit.register(self.__exit__, None, None, None)
|
||||
|
||||
def __enter__(self):
|
||||
if not self.pool:
|
||||
self.pool = self.run_with_loop(self.conn.__aenter__(), self.loop)
|
||||
self.pool = self.run_with_loop(self.conn.__aenter__(), self.loop)
|
||||
return self
|
||||
|
||||
def send(self, *args):
|
||||
@ -5515,11 +5502,3 @@ has_websockets = bool(compat_websockets)
|
||||
def merge_headers(*dicts):
|
||||
"""Merge dicts of http headers case insensitively, prioritizing the latter ones"""
|
||||
return {k.title(): v for k, v in itertools.chain.from_iterable(map(dict.items, dicts))}
|
||||
|
||||
|
||||
class classproperty:
|
||||
def __init__(self, f):
|
||||
self.f = f
|
||||
|
||||
def __get__(self, _, cls):
|
||||
return self.f(cls)
|
||||
|
Loading…
Reference in New Issue
Block a user