[cleanup] Refactor some code

pull/606/head
pukkandan 3 years ago
parent d74a58a186
commit dbf5416a20

@ -126,6 +126,7 @@ from .extractor import (
) )
from .extractor.openload import PhantomJSwrapper from .extractor.openload import PhantomJSwrapper
from .downloader import ( from .downloader import (
FFmpegFD,
get_suitable_downloader, get_suitable_downloader,
shorten_protocol_name shorten_protocol_name
) )
@ -2690,20 +2691,15 @@ class YoutubeDL(object):
info_dict['__real_download'] = False info_dict['__real_download'] = False
_protocols = set(determine_protocol(f) for f in requested_formats) _protocols = set(determine_protocol(f) for f in requested_formats)
if len(_protocols) == 1: if len(_protocols) == 1: # All requested formats have same protocol
info_dict['protocol'] = _protocols.pop() info_dict['protocol'] = _protocols.pop()
directly_mergable = ( directly_mergable = FFmpegFD.can_merge_formats(info_dict)
'no-direct-merge' not in self.params.get('compat_opts', []) if dl_filename is not None:
and info_dict.get('protocol') is not None # All requested formats have same protocol pass
and not self.params.get('allow_unplayable_formats') elif (directly_mergable and get_suitable_downloader(info_dict, self.params) == FFmpegFD):
and get_suitable_downloader(info_dict, self.params).__name__ == 'FFmpegFD') info_dict['url'] = '\n'.join(f['url'] for f in requested_formats)
if directly_mergable: success, real_download = self.dl(temp_filename, info_dict)
info_dict['url'] = requested_formats[0]['url'] info_dict['__real_download'] = real_download
# Treat it as a single download
dl_filename = existing_file(full_filename, temp_filename)
if dl_filename is None:
success, real_download = self.dl(temp_filename, info_dict)
info_dict['__real_download'] = real_download
else: else:
downloaded = [] downloaded = []
merger = FFmpegMergerPP(self) merger = FFmpegMergerPP(self)
@ -2717,28 +2713,25 @@ class YoutubeDL(object):
'You have requested merging of multiple formats but ffmpeg is not installed. ' 'You have requested merging of multiple formats but ffmpeg is not installed. '
'The formats won\'t be merged.') 'The formats won\'t be merged.')
if dl_filename is None: for f in requested_formats:
for f in requested_formats: new_info = dict(info_dict)
new_info = dict(info_dict) del new_info['requested_formats']
del new_info['requested_formats'] new_info.update(f)
new_info.update(f) fname = prepend_extension(temp_filename, 'f%s' % f['format_id'], new_info['ext'])
fname = prepend_extension( if not self._ensure_dir_exists(fname):
self.prepare_filename(new_info, 'temp'), return
'f%s' % f['format_id'], new_info['ext']) downloaded.append(fname)
if not self._ensure_dir_exists(fname): partial_success, real_download = self.dl(fname, new_info)
return info_dict['__real_download'] = info_dict['__real_download'] or real_download
downloaded.append(fname) success = success and partial_success
partial_success, real_download = self.dl(fname, new_info) if merger.available and not self.params.get('allow_unplayable_formats'):
info_dict['__real_download'] = info_dict['__real_download'] or real_download info_dict['__postprocessors'].append(merger)
success = success and partial_success info_dict['__files_to_merge'] = downloaded
if merger.available and not self.params.get('allow_unplayable_formats'): # Even if there were no downloads, it is being merged only now
info_dict['__postprocessors'].append(merger) info_dict['__real_download'] = True
info_dict['__files_to_merge'] = downloaded else:
# Even if there were no downloads, it is being merged only now for file in downloaded:
info_dict['__real_download'] = True files_to_move[file] = None
else:
for file in downloaded:
files_to_move[file] = None
else: else:
# Just a single file # Just a single file
dl_filename = existing_file(full_filename, temp_filename) dl_filename = existing_file(full_filename, temp_filename)

@ -3,17 +3,19 @@ from __future__ import unicode_literals
from ..compat import compat_str from ..compat import compat_str
from ..utils import ( from ..utils import (
determine_protocol, determine_protocol,
NO_DEFAULT
) )
def _get_real_downloader(info_dict, protocol=None, *args, **kwargs): def get_suitable_downloader(info_dict, params={}, default=NO_DEFAULT, protocol=None):
info_dict['protocol'] = determine_protocol(info_dict)
info_copy = info_dict.copy() info_copy = info_dict.copy()
if protocol: if protocol:
info_copy['protocol'] = protocol info_copy['protocol'] = protocol
return get_suitable_downloader(info_copy, *args, **kwargs) return _get_suitable_downloader(info_copy, params, default)
# Some of these require _get_real_downloader # Some of these require get_suitable_downloader
from .common import FileDownloader from .common import FileDownloader
from .dash import DashSegmentsFD from .dash import DashSegmentsFD
from .f4m import F4mFD from .f4m import F4mFD
@ -69,14 +71,15 @@ def shorten_protocol_name(proto, simplify=False):
return short_protocol_names.get(proto, proto) return short_protocol_names.get(proto, proto)
def get_suitable_downloader(info_dict, params={}, default=HttpFD): def _get_suitable_downloader(info_dict, params, default):
"""Get the downloader class that can handle the info dict.""" """Get the downloader class that can handle the info dict."""
protocol = determine_protocol(info_dict) if default is NO_DEFAULT:
info_dict['protocol'] = protocol default = HttpFD
# if (info_dict.get('start_time') or info_dict.get('end_time')) and not info_dict.get('requested_formats') and FFmpegFD.can_download(info_dict): # if (info_dict.get('start_time') or info_dict.get('end_time')) and not info_dict.get('requested_formats') and FFmpegFD.can_download(info_dict):
# return FFmpegFD # return FFmpegFD
protocol = info_dict['protocol']
downloaders = params.get('external_downloader') downloaders = params.get('external_downloader')
external_downloader = ( external_downloader = (
downloaders if isinstance(downloaders, compat_str) or downloaders is None downloaders if isinstance(downloaders, compat_str) or downloaders is None
@ -94,7 +97,7 @@ def get_suitable_downloader(info_dict, params={}, default=HttpFD):
return FFmpegFD return FFmpegFD
elif external_downloader == 'native': elif external_downloader == 'native':
return HlsFD return HlsFD
elif _get_real_downloader(info_dict, 'm3u8_frag_urls', params, None): elif get_suitable_downloader(info_dict, params, None, protocol='m3u8_frag_urls'):
return HlsFD return HlsFD
elif params.get('hls_prefer_native') is True: elif params.get('hls_prefer_native') is True:
return HlsFD return HlsFD

@ -1,6 +1,6 @@
from __future__ import unicode_literals from __future__ import unicode_literals
from ..downloader import _get_real_downloader from ..downloader import get_suitable_downloader
from .fragment import FragmentFD from .fragment import FragmentFD
from ..utils import urljoin from ..utils import urljoin
@ -15,11 +15,14 @@ class DashSegmentsFD(FragmentFD):
FD_NAME = 'dashsegments' FD_NAME = 'dashsegments'
def real_download(self, filename, info_dict): def real_download(self, filename, info_dict):
if info_dict.get('is_live'):
self.report_error('Live DASH videos are not supported')
fragment_base_url = info_dict.get('fragment_base_url') fragment_base_url = info_dict.get('fragment_base_url')
fragments = info_dict['fragments'][:1] if self.params.get( fragments = info_dict['fragments'][:1] if self.params.get(
'test', False) else info_dict['fragments'] 'test', False) else info_dict['fragments']
real_downloader = _get_real_downloader(info_dict, 'dash_frag_urls', self.params, None) real_downloader = get_suitable_downloader(info_dict, self.params, None, protocol='dash_frag_urls')
ctx = { ctx = {
'filename': filename, 'filename': filename,
@ -54,9 +57,6 @@ class DashSegmentsFD(FragmentFD):
info_copy = info_dict.copy() info_copy = info_dict.copy()
info_copy['fragments'] = fragments_to_download info_copy['fragments'] = fragments_to_download
fd = real_downloader(self.ydl, self.params) fd = real_downloader(self.ydl, self.params)
# TODO: Make progress updates work without hooking twice
# for ph in self._progress_hooks:
# fd.add_progress_hook(ph)
return fd.real_download(filename, info_copy) return fd.real_download(filename, info_copy)
return self.download_and_append_fragments(ctx, fragments_to_download, info_dict) return self.download_and_append_fragments(ctx, fragments_to_download, info_dict)

@ -345,12 +345,22 @@ class FFmpegFD(ExternalFD):
@classmethod @classmethod
def available(cls, path=None): def available(cls, path=None):
# TODO: Fix path for ffmpeg # TODO: Fix path for ffmpeg
# Fixme: This may be wrong when --ffmpeg-location is used
return FFmpegPostProcessor().available return FFmpegPostProcessor().available
def on_process_started(self, proc, stdin): def on_process_started(self, proc, stdin):
""" Override this in subclasses """ """ Override this in subclasses """
pass pass
@classmethod
def can_merge_formats(cls, info_dict, params={}):
return (
info_dict.get('requested_formats')
and info_dict.get('protocol')
and not params.get('allow_unplayable_formats')
and 'no-direct-merge' not in params.get('compat_opts', [])
and cls.can_download(info_dict))
def _call_downloader(self, tmpfilename, info_dict): def _call_downloader(self, tmpfilename, info_dict):
urls = [f['url'] for f in info_dict.get('requested_formats', [])] or [info_dict['url']] urls = [f['url'] for f in info_dict.get('requested_formats', [])] or [info_dict['url']]
ffpp = FFmpegPostProcessor(downloader=self) ffpp = FFmpegPostProcessor(downloader=self)

@ -4,7 +4,7 @@ import re
import io import io
import binascii import binascii
from ..downloader import _get_real_downloader from ..downloader import get_suitable_downloader
from .fragment import FragmentFD, can_decrypt_frag from .fragment import FragmentFD, can_decrypt_frag
from .external import FFmpegFD from .external import FFmpegFD
@ -80,16 +80,13 @@ class HlsFD(FragmentFD):
fd = FFmpegFD(self.ydl, self.params) fd = FFmpegFD(self.ydl, self.params)
self.report_warning( self.report_warning(
'%s detected unsupported features; extraction will be delegated to %s' % (self.FD_NAME, fd.get_basename())) '%s detected unsupported features; extraction will be delegated to %s' % (self.FD_NAME, fd.get_basename()))
# TODO: Make progress updates work without hooking twice
# for ph in self._progress_hooks:
# fd.add_progress_hook(ph)
return fd.real_download(filename, info_dict) return fd.real_download(filename, info_dict)
is_webvtt = info_dict['ext'] == 'vtt' is_webvtt = info_dict['ext'] == 'vtt'
if is_webvtt: if is_webvtt:
real_downloader = None # Packing the fragments is not currently supported for external downloader real_downloader = None # Packing the fragments is not currently supported for external downloader
else: else:
real_downloader = _get_real_downloader(info_dict, 'm3u8_frag_urls', self.params, None) real_downloader = get_suitable_downloader(info_dict, self.params, None, protocol='m3u8_frag_urls')
if real_downloader and not real_downloader.supports_manifest(s): if real_downloader and not real_downloader.supports_manifest(s):
real_downloader = None real_downloader = None
if real_downloader: if real_downloader:

@ -4,7 +4,7 @@ from __future__ import unicode_literals
import threading import threading
from .common import FileDownloader from .common import FileDownloader
from ..downloader import _get_real_downloader from ..downloader import get_suitable_downloader
from ..extractor.niconico import NiconicoIE from ..extractor.niconico import NiconicoIE
from ..compat import compat_urllib_request from ..compat import compat_urllib_request
@ -20,7 +20,7 @@ class NiconicoDmcFD(FileDownloader):
ie = NiconicoIE(self.ydl) ie = NiconicoIE(self.ydl)
info_dict, heartbeat_info_dict = ie._get_heartbeat_info(info_dict) info_dict, heartbeat_info_dict = ie._get_heartbeat_info(info_dict)
fd = _get_real_downloader(info_dict, params=self.params)(self.ydl, self.params) fd = get_suitable_downloader(info_dict, params=self.params)(self.ydl, self.params)
success = download_complete = False success = download_complete = False
timer = [None] timer = [None]

@ -1298,7 +1298,7 @@ class InfoExtractor(object):
# JSON-LD may be malformed and thus `fatal` should be respected. # JSON-LD may be malformed and thus `fatal` should be respected.
# At the same time `default` may be passed that assumes `fatal=False` # At the same time `default` may be passed that assumes `fatal=False`
# for _search_regex. Let's simulate the same behavior here as well. # for _search_regex. Let's simulate the same behavior here as well.
fatal = kwargs.get('fatal', True) if default == NO_DEFAULT else False fatal = kwargs.get('fatal', True) if default is NO_DEFAULT else False
json_ld = [] json_ld = []
for mobj in json_ld_list: for mobj in json_ld_list:
json_ld_item = self._parse_json( json_ld_item = self._parse_json(
@ -1522,7 +1522,7 @@ class InfoExtractor(object):
'size': {'type': 'combined', 'same_limit': True, 'field': ('filesize', 'fs_approx')}, 'size': {'type': 'combined', 'same_limit': True, 'field': ('filesize', 'fs_approx')},
'ext': {'type': 'combined', 'field': ('vext', 'aext')}, 'ext': {'type': 'combined', 'field': ('vext', 'aext')},
'res': {'type': 'multiple', 'field': ('height', 'width'), 'res': {'type': 'multiple', 'field': ('height', 'width'),
'function': lambda it: (lambda l: min(l) if l else 0)(tuple(filter(None, it)))}, 'function': lambda it: (lambda l: min(l) if l else 0)(tuple(filter(None, it)))},
# Most of these exist only for compatibility reasons # Most of these exist only for compatibility reasons
'dimension': {'type': 'alias', 'field': 'res'}, 'dimension': {'type': 'alias', 'field': 'res'},

@ -706,9 +706,8 @@ def parseOpts(overrideArguments=None):
callback_kwargs={ callback_kwargs={
'allowed_keys': 'http|ftp|m3u8|dash|rtsp|rtmp|mms', 'allowed_keys': 'http|ftp|m3u8|dash|rtsp|rtmp|mms',
'default_key': 'default', 'default_key': 'default',
'process': lambda x: x.strip() 'process': str.strip
}, }, help=(
help=(
'Name or path of the external downloader to use (optionally) prefixed by ' 'Name or path of the external downloader to use (optionally) prefixed by '
'the protocols (http, ftp, m3u8, dash, rstp, rtmp, mms) to use it for. ' 'the protocols (http, ftp, m3u8, dash, rstp, rtmp, mms) to use it for. '
'Currently supports native, %s (Recommended: aria2c). ' 'Currently supports native, %s (Recommended: aria2c). '
@ -724,8 +723,7 @@ def parseOpts(overrideArguments=None):
'allowed_keys': '|'.join(list_external_downloaders()), 'allowed_keys': '|'.join(list_external_downloaders()),
'default_key': 'default', 'default_key': 'default',
'process': compat_shlex_split 'process': compat_shlex_split
}, }, help=(
help=(
'Give these arguments to the external downloader. ' 'Give these arguments to the external downloader. '
'Specify the downloader name and the arguments separated by a colon ":". ' 'Specify the downloader name and the arguments separated by a colon ":". '
'You can use this option multiple times to give different arguments to different downloaders ' 'You can use this option multiple times to give different arguments to different downloaders '
@ -944,8 +942,7 @@ def parseOpts(overrideArguments=None):
callback_kwargs={ callback_kwargs={
'allowed_keys': '|'.join(OUTTMPL_TYPES.keys()), 'allowed_keys': '|'.join(OUTTMPL_TYPES.keys()),
'default_key': 'default' 'default_key': 'default'
}, }, help='Output filename template; see "OUTPUT TEMPLATE" for details')
help='Output filename template; see "OUTPUT TEMPLATE" for details')
filesystem.add_option( filesystem.add_option(
'--output-na-placeholder', '--output-na-placeholder',
dest='outtmpl_na_placeholder', metavar='TEXT', default='NA', dest='outtmpl_na_placeholder', metavar='TEXT', default='NA',
@ -1191,8 +1188,7 @@ def parseOpts(overrideArguments=None):
'allowed_keys': r'\w+(?:\+\w+)?', 'default_key': 'default-compat', 'allowed_keys': r'\w+(?:\+\w+)?', 'default_key': 'default-compat',
'process': compat_shlex_split, 'process': compat_shlex_split,
'multiple_keys': False 'multiple_keys': False
}, }, help=(
help=(
'Give these arguments to the postprocessors. ' 'Give these arguments to the postprocessors. '
'Specify the postprocessor/executable name and the arguments separated by a colon ":" ' 'Specify the postprocessor/executable name and the arguments separated by a colon ":" '
'to give the argument to the specified postprocessor/executable. Supported PP are: ' 'to give the argument to the specified postprocessor/executable. Supported PP are: '
@ -1385,8 +1381,7 @@ def parseOpts(overrideArguments=None):
'multiple_keys': False, 'multiple_keys': False,
'process': lambda val: dict( 'process': lambda val: dict(
_extractor_arg_parser(*arg.split('=', 1)) for arg in val.split(';')) _extractor_arg_parser(*arg.split('=', 1)) for arg in val.split(';'))
}, }, help=(
help=(
'Pass these arguments to the extractor. See "EXTRACTOR ARGUMENTS" for details. ' 'Pass these arguments to the extractor. See "EXTRACTOR ARGUMENTS" for details. '
'You can use this option multiple times to give arguments for different extractors')) 'You can use this option multiple times to give arguments for different extractors'))
extractor.add_option( extractor.add_option(

@ -4544,7 +4544,7 @@ def parse_codecs(codecs_str):
if not codecs_str: if not codecs_str:
return {} return {}
split_codecs = list(filter(None, map( split_codecs = list(filter(None, map(
lambda str: str.strip(), codecs_str.strip().strip(',').split(',')))) str.strip, codecs_str.strip().strip(',').split(','))))
vcodec, acodec = None, None vcodec, acodec = None, None
for full_codec in split_codecs: for full_codec in split_codecs:
codec = full_codec.split('.')[0] codec = full_codec.split('.')[0]
@ -6246,7 +6246,7 @@ def traverse_obj(
# TODO: Write tests # TODO: Write tests
''' '''
if not casesense: if not casesense:
_lower = lambda k: k.lower() if isinstance(k, str) else k _lower = lambda k: (k.lower() if isinstance(k, str) else k)
path_list = (map(_lower, variadic(path)) for path in path_list) path_list = (map(_lower, variadic(path)) for path in path_list)
def _traverse_obj(obj, path, _current_depth=0): def _traverse_obj(obj, path, _current_depth=0):

Loading…
Cancel
Save