]> jfr.im git - yt-dlp.git/blame_incremental - youtube_dl/YoutubeDL.py
Allow iterators for playlist result entries
[yt-dlp.git] / youtube_dl / YoutubeDL.py
... / ...
CommitLineData
1#!/usr/bin/env python
2# -*- coding: utf-8 -*-
3
4from __future__ import absolute_import, unicode_literals
5
6import collections
7import datetime
8import errno
9import io
10import itertools
11import json
12import locale
13import os
14import platform
15import re
16import shutil
17import subprocess
18import socket
19import sys
20import time
21import traceback
22
23if os.name == 'nt':
24 import ctypes
25
26from .compat import (
27 compat_cookiejar,
28 compat_expanduser,
29 compat_http_client,
30 compat_str,
31 compat_urllib_error,
32 compat_urllib_request,
33)
34from .utils import (
35 escape_url,
36 ContentTooShortError,
37 date_from_str,
38 DateRange,
39 DEFAULT_OUTTMPL,
40 determine_ext,
41 DownloadError,
42 encodeFilename,
43 ExtractorError,
44 format_bytes,
45 formatSeconds,
46 get_term_width,
47 locked_file,
48 make_HTTPS_handler,
49 MaxDownloadsReached,
50 PagedList,
51 PostProcessingError,
52 platform_name,
53 preferredencoding,
54 SameFileError,
55 sanitize_filename,
56 subtitles_filename,
57 takewhile_inclusive,
58 UnavailableVideoError,
59 url_basename,
60 write_json_file,
61 write_string,
62 YoutubeDLHandler,
63 prepend_extension,
64 args_to_str,
65)
66from .cache import Cache
67from .extractor import get_info_extractor, gen_extractors
68from .downloader import get_suitable_downloader
69from .downloader.rtmp import rtmpdump_version
70from .postprocessor import FFmpegMergerPP, FFmpegPostProcessor
71from .version import __version__
72
73
74class YoutubeDL(object):
75 """YoutubeDL class.
76
77 YoutubeDL objects are the ones responsible of downloading the
78 actual video file and writing it to disk if the user has requested
79 it, among some other tasks. In most cases there should be one per
80 program. As, given a video URL, the downloader doesn't know how to
81 extract all the needed information, task that InfoExtractors do, it
82 has to pass the URL to one of them.
83
84 For this, YoutubeDL objects have a method that allows
85 InfoExtractors to be registered in a given order. When it is passed
86 a URL, the YoutubeDL object handles it to the first InfoExtractor it
87 finds that reports being able to handle it. The InfoExtractor extracts
88 all the information about the video or videos the URL refers to, and
89 YoutubeDL process the extracted information, possibly using a File
90 Downloader to download the video.
91
92 YoutubeDL objects accept a lot of parameters. In order not to saturate
93 the object constructor with arguments, it receives a dictionary of
94 options instead. These options are available through the params
95 attribute for the InfoExtractors to use. The YoutubeDL also
96 registers itself as the downloader in charge for the InfoExtractors
97 that are added to it, so this is a "mutual registration".
98
99 Available options:
100
101 username: Username for authentication purposes.
102 password: Password for authentication purposes.
103 videopassword: Password for acces a video.
104 usenetrc: Use netrc for authentication instead.
105 verbose: Print additional info to stdout.
106 quiet: Do not print messages to stdout.
107 no_warnings: Do not print out anything for warnings.
108 forceurl: Force printing final URL.
109 forcetitle: Force printing title.
110 forceid: Force printing ID.
111 forcethumbnail: Force printing thumbnail URL.
112 forcedescription: Force printing description.
113 forcefilename: Force printing final filename.
114 forceduration: Force printing duration.
115 forcejson: Force printing info_dict as JSON.
116 dump_single_json: Force printing the info_dict of the whole playlist
117 (or video) as a single JSON line.
118 simulate: Do not download the video files.
119 format: Video format code.
120 format_limit: Highest quality format to try.
121 outtmpl: Template for output names.
122 restrictfilenames: Do not allow "&" and spaces in file names
123 ignoreerrors: Do not stop on download errors.
124 nooverwrites: Prevent overwriting files.
125 playliststart: Playlist item to start at.
126 playlistend: Playlist item to end at.
127 matchtitle: Download only matching titles.
128 rejecttitle: Reject downloads for matching titles.
129 logger: Log messages to a logging.Logger instance.
130 logtostderr: Log messages to stderr instead of stdout.
131 writedescription: Write the video description to a .description file
132 writeinfojson: Write the video description to a .info.json file
133 writeannotations: Write the video annotations to a .annotations.xml file
134 writethumbnail: Write the thumbnail image to a file
135 writesubtitles: Write the video subtitles to a file
136 writeautomaticsub: Write the automatic subtitles to a file
137 allsubtitles: Downloads all the subtitles of the video
138 (requires writesubtitles or writeautomaticsub)
139 listsubtitles: Lists all available subtitles for the video
140 subtitlesformat: Subtitle format [srt/sbv/vtt] (default=srt)
141 subtitleslangs: List of languages of the subtitles to download
142 keepvideo: Keep the video file after post-processing
143 daterange: A DateRange object, download only if the upload_date is in the range.
144 skip_download: Skip the actual download of the video file
145 cachedir: Location of the cache files in the filesystem.
146 False to disable filesystem cache.
147 noplaylist: Download single video instead of a playlist if in doubt.
148 age_limit: An integer representing the user's age in years.
149 Unsuitable videos for the given age are skipped.
150 min_views: An integer representing the minimum view count the video
151 must have in order to not be skipped.
152 Videos without view count information are always
153 downloaded. None for no limit.
154 max_views: An integer representing the maximum view count.
155 Videos that are more popular than that are not
156 downloaded.
157 Videos without view count information are always
158 downloaded. None for no limit.
159 download_archive: File name of a file where all downloads are recorded.
160 Videos already present in the file are not downloaded
161 again.
162 cookiefile: File name where cookies should be read from and dumped to.
163 nocheckcertificate:Do not verify SSL certificates
164 prefer_insecure: Use HTTP instead of HTTPS to retrieve information.
165 At the moment, this is only supported by YouTube.
166 proxy: URL of the proxy server to use
167 socket_timeout: Time to wait for unresponsive hosts, in seconds
168 bidi_workaround: Work around buggy terminals without bidirectional text
169 support, using fridibi
170 debug_printtraffic:Print out sent and received HTTP traffic
171 include_ads: Download ads as well
172 default_search: Prepend this string if an input url is not valid.
173 'auto' for elaborate guessing
174 encoding: Use this encoding instead of the system-specified.
175 extract_flat: Do not resolve URLs, return the immediate result.
176 Pass in 'in_playlist' to only show this behavior for
177 playlist items.
178
179 The following parameters are not used by YoutubeDL itself, they are used by
180 the FileDownloader:
181 nopart, updatetime, buffersize, ratelimit, min_filesize, max_filesize, test,
182 noresizebuffer, retries, continuedl, noprogress, consoletitle
183
184 The following options are used by the post processors:
185 prefer_ffmpeg: If True, use ffmpeg instead of avconv if both are available,
186 otherwise prefer avconv.
187 exec_cmd: Arbitrary command to run after downloading
188 """
189
190 params = None
191 _ies = []
192 _pps = []
193 _download_retcode = None
194 _num_downloads = None
195 _screen_file = None
196
197 def __init__(self, params=None, auto_init=True):
198 """Create a FileDownloader object with the given options."""
199 if params is None:
200 params = {}
201 self._ies = []
202 self._ies_instances = {}
203 self._pps = []
204 self._progress_hooks = []
205 self._download_retcode = 0
206 self._num_downloads = 0
207 self._screen_file = [sys.stdout, sys.stderr][params.get('logtostderr', False)]
208 self._err_file = sys.stderr
209 self.params = params
210 self.cache = Cache(self)
211
212 if params.get('bidi_workaround', False):
213 try:
214 import pty
215 master, slave = pty.openpty()
216 width = get_term_width()
217 if width is None:
218 width_args = []
219 else:
220 width_args = ['-w', str(width)]
221 sp_kwargs = dict(
222 stdin=subprocess.PIPE,
223 stdout=slave,
224 stderr=self._err_file)
225 try:
226 self._output_process = subprocess.Popen(
227 ['bidiv'] + width_args, **sp_kwargs
228 )
229 except OSError:
230 self._output_process = subprocess.Popen(
231 ['fribidi', '-c', 'UTF-8'] + width_args, **sp_kwargs)
232 self._output_channel = os.fdopen(master, 'rb')
233 except OSError as ose:
234 if ose.errno == 2:
235 self.report_warning('Could not find fribidi executable, ignoring --bidi-workaround . Make sure that fribidi is an executable file in one of the directories in your $PATH.')
236 else:
237 raise
238
239 if (sys.version_info >= (3,) and sys.platform != 'win32' and
240 sys.getfilesystemencoding() in ['ascii', 'ANSI_X3.4-1968']
241 and not params.get('restrictfilenames', False)):
242 # On Python 3, the Unicode filesystem API will throw errors (#1474)
243 self.report_warning(
244 'Assuming --restrict-filenames since file system encoding '
245 'cannot encode all characters. '
246 'Set the LC_ALL environment variable to fix this.')
247 self.params['restrictfilenames'] = True
248
249 if '%(stitle)s' in self.params.get('outtmpl', ''):
250 self.report_warning('%(stitle)s is deprecated. Use the %(title)s and the --restrict-filenames flag(which also secures %(uploader)s et al) instead.')
251
252 self._setup_opener()
253
254 if auto_init:
255 self.print_debug_header()
256 self.add_default_info_extractors()
257
258 def warn_if_short_id(self, argv):
259 # short YouTube ID starting with dash?
260 idxs = [
261 i for i, a in enumerate(argv)
262 if re.match(r'^-[0-9A-Za-z_-]{10}$', a)]
263 if idxs:
264 correct_argv = (
265 ['youtube-dl'] +
266 [a for i, a in enumerate(argv) if i not in idxs] +
267 ['--'] + [argv[i] for i in idxs]
268 )
269 self.report_warning(
270 'Long argument string detected. '
271 'Use -- to separate parameters and URLs, like this:\n%s\n' %
272 args_to_str(correct_argv))
273
274 def add_info_extractor(self, ie):
275 """Add an InfoExtractor object to the end of the list."""
276 self._ies.append(ie)
277 self._ies_instances[ie.ie_key()] = ie
278 ie.set_downloader(self)
279
280 def get_info_extractor(self, ie_key):
281 """
282 Get an instance of an IE with name ie_key, it will try to get one from
283 the _ies list, if there's no instance it will create a new one and add
284 it to the extractor list.
285 """
286 ie = self._ies_instances.get(ie_key)
287 if ie is None:
288 ie = get_info_extractor(ie_key)()
289 self.add_info_extractor(ie)
290 return ie
291
292 def add_default_info_extractors(self):
293 """
294 Add the InfoExtractors returned by gen_extractors to the end of the list
295 """
296 for ie in gen_extractors():
297 self.add_info_extractor(ie)
298
299 def add_post_processor(self, pp):
300 """Add a PostProcessor object to the end of the chain."""
301 self._pps.append(pp)
302 pp.set_downloader(self)
303
304 def add_progress_hook(self, ph):
305 """Add the progress hook (currently only for the file downloader)"""
306 self._progress_hooks.append(ph)
307
308 def _bidi_workaround(self, message):
309 if not hasattr(self, '_output_channel'):
310 return message
311
312 assert hasattr(self, '_output_process')
313 assert isinstance(message, compat_str)
314 line_count = message.count('\n') + 1
315 self._output_process.stdin.write((message + '\n').encode('utf-8'))
316 self._output_process.stdin.flush()
317 res = ''.join(self._output_channel.readline().decode('utf-8')
318 for _ in range(line_count))
319 return res[:-len('\n')]
320
321 def to_screen(self, message, skip_eol=False):
322 """Print message to stdout if not in quiet mode."""
323 return self.to_stdout(message, skip_eol, check_quiet=True)
324
325 def _write_string(self, s, out=None):
326 write_string(s, out=out, encoding=self.params.get('encoding'))
327
328 def to_stdout(self, message, skip_eol=False, check_quiet=False):
329 """Print message to stdout if not in quiet mode."""
330 if self.params.get('logger'):
331 self.params['logger'].debug(message)
332 elif not check_quiet or not self.params.get('quiet', False):
333 message = self._bidi_workaround(message)
334 terminator = ['\n', ''][skip_eol]
335 output = message + terminator
336
337 self._write_string(output, self._screen_file)
338
339 def to_stderr(self, message):
340 """Print message to stderr."""
341 assert isinstance(message, compat_str)
342 if self.params.get('logger'):
343 self.params['logger'].error(message)
344 else:
345 message = self._bidi_workaround(message)
346 output = message + '\n'
347 self._write_string(output, self._err_file)
348
349 def to_console_title(self, message):
350 if not self.params.get('consoletitle', False):
351 return
352 if os.name == 'nt' and ctypes.windll.kernel32.GetConsoleWindow():
353 # c_wchar_p() might not be necessary if `message` is
354 # already of type unicode()
355 ctypes.windll.kernel32.SetConsoleTitleW(ctypes.c_wchar_p(message))
356 elif 'TERM' in os.environ:
357 self._write_string('\033]0;%s\007' % message, self._screen_file)
358
359 def save_console_title(self):
360 if not self.params.get('consoletitle', False):
361 return
362 if 'TERM' in os.environ:
363 # Save the title on stack
364 self._write_string('\033[22;0t', self._screen_file)
365
366 def restore_console_title(self):
367 if not self.params.get('consoletitle', False):
368 return
369 if 'TERM' in os.environ:
370 # Restore the title from stack
371 self._write_string('\033[23;0t', self._screen_file)
372
373 def __enter__(self):
374 self.save_console_title()
375 return self
376
377 def __exit__(self, *args):
378 self.restore_console_title()
379
380 if self.params.get('cookiefile') is not None:
381 self.cookiejar.save()
382
383 def trouble(self, message=None, tb=None):
384 """Determine action to take when a download problem appears.
385
386 Depending on if the downloader has been configured to ignore
387 download errors or not, this method may throw an exception or
388 not when errors are found, after printing the message.
389
390 tb, if given, is additional traceback information.
391 """
392 if message is not None:
393 self.to_stderr(message)
394 if self.params.get('verbose'):
395 if tb is None:
396 if sys.exc_info()[0]: # if .trouble has been called from an except block
397 tb = ''
398 if hasattr(sys.exc_info()[1], 'exc_info') and sys.exc_info()[1].exc_info[0]:
399 tb += ''.join(traceback.format_exception(*sys.exc_info()[1].exc_info))
400 tb += compat_str(traceback.format_exc())
401 else:
402 tb_data = traceback.format_list(traceback.extract_stack())
403 tb = ''.join(tb_data)
404 self.to_stderr(tb)
405 if not self.params.get('ignoreerrors', False):
406 if sys.exc_info()[0] and hasattr(sys.exc_info()[1], 'exc_info') and sys.exc_info()[1].exc_info[0]:
407 exc_info = sys.exc_info()[1].exc_info
408 else:
409 exc_info = sys.exc_info()
410 raise DownloadError(message, exc_info)
411 self._download_retcode = 1
412
413 def report_warning(self, message):
414 '''
415 Print the message to stderr, it will be prefixed with 'WARNING:'
416 If stderr is a tty file the 'WARNING:' will be colored
417 '''
418 if self.params.get('logger') is not None:
419 self.params['logger'].warning(message)
420 else:
421 if self.params.get('no_warnings'):
422 return
423 if self._err_file.isatty() and os.name != 'nt':
424 _msg_header = '\033[0;33mWARNING:\033[0m'
425 else:
426 _msg_header = 'WARNING:'
427 warning_message = '%s %s' % (_msg_header, message)
428 self.to_stderr(warning_message)
429
430 def report_error(self, message, tb=None):
431 '''
432 Do the same as trouble, but prefixes the message with 'ERROR:', colored
433 in red if stderr is a tty file.
434 '''
435 if self._err_file.isatty() and os.name != 'nt':
436 _msg_header = '\033[0;31mERROR:\033[0m'
437 else:
438 _msg_header = 'ERROR:'
439 error_message = '%s %s' % (_msg_header, message)
440 self.trouble(error_message, tb)
441
442 def report_file_already_downloaded(self, file_name):
443 """Report file has already been fully downloaded."""
444 try:
445 self.to_screen('[download] %s has already been downloaded' % file_name)
446 except UnicodeEncodeError:
447 self.to_screen('[download] The file has already been downloaded')
448
449 def prepare_filename(self, info_dict):
450 """Generate the output filename."""
451 try:
452 template_dict = dict(info_dict)
453
454 template_dict['epoch'] = int(time.time())
455 autonumber_size = self.params.get('autonumber_size')
456 if autonumber_size is None:
457 autonumber_size = 5
458 autonumber_templ = '%0' + str(autonumber_size) + 'd'
459 template_dict['autonumber'] = autonumber_templ % self._num_downloads
460 if template_dict.get('playlist_index') is not None:
461 template_dict['playlist_index'] = '%0*d' % (len(str(template_dict['n_entries'])), template_dict['playlist_index'])
462 if template_dict.get('resolution') is None:
463 if template_dict.get('width') and template_dict.get('height'):
464 template_dict['resolution'] = '%dx%d' % (template_dict['width'], template_dict['height'])
465 elif template_dict.get('height'):
466 template_dict['resolution'] = '%sp' % template_dict['height']
467 elif template_dict.get('width'):
468 template_dict['resolution'] = '?x%d' % template_dict['width']
469
470 sanitize = lambda k, v: sanitize_filename(
471 compat_str(v),
472 restricted=self.params.get('restrictfilenames'),
473 is_id=(k == 'id'))
474 template_dict = dict((k, sanitize(k, v))
475 for k, v in template_dict.items()
476 if v is not None)
477 template_dict = collections.defaultdict(lambda: 'NA', template_dict)
478
479 outtmpl = self.params.get('outtmpl', DEFAULT_OUTTMPL)
480 tmpl = compat_expanduser(outtmpl)
481 filename = tmpl % template_dict
482 return filename
483 except ValueError as err:
484 self.report_error('Error in output template: ' + str(err) + ' (encoding: ' + repr(preferredencoding()) + ')')
485 return None
486
487 def _match_entry(self, info_dict):
488 """ Returns None iff the file should be downloaded """
489
490 video_title = info_dict.get('title', info_dict.get('id', 'video'))
491 if 'title' in info_dict:
492 # This can happen when we're just evaluating the playlist
493 title = info_dict['title']
494 matchtitle = self.params.get('matchtitle', False)
495 if matchtitle:
496 if not re.search(matchtitle, title, re.IGNORECASE):
497 return '"' + title + '" title did not match pattern "' + matchtitle + '"'
498 rejecttitle = self.params.get('rejecttitle', False)
499 if rejecttitle:
500 if re.search(rejecttitle, title, re.IGNORECASE):
501 return '"' + title + '" title matched reject pattern "' + rejecttitle + '"'
502 date = info_dict.get('upload_date', None)
503 if date is not None:
504 dateRange = self.params.get('daterange', DateRange())
505 if date not in dateRange:
506 return '%s upload date is not in range %s' % (date_from_str(date).isoformat(), dateRange)
507 view_count = info_dict.get('view_count', None)
508 if view_count is not None:
509 min_views = self.params.get('min_views')
510 if min_views is not None and view_count < min_views:
511 return 'Skipping %s, because it has not reached minimum view count (%d/%d)' % (video_title, view_count, min_views)
512 max_views = self.params.get('max_views')
513 if max_views is not None and view_count > max_views:
514 return 'Skipping %s, because it has exceeded the maximum view count (%d/%d)' % (video_title, view_count, max_views)
515 age_limit = self.params.get('age_limit')
516 if age_limit is not None:
517 actual_age_limit = info_dict.get('age_limit')
518 if actual_age_limit is None:
519 actual_age_limit = 0
520 if age_limit < actual_age_limit:
521 return 'Skipping "' + title + '" because it is age restricted'
522 if self.in_download_archive(info_dict):
523 return '%s has already been recorded in archive' % video_title
524 return None
525
526 @staticmethod
527 def add_extra_info(info_dict, extra_info):
528 '''Set the keys from extra_info in info dict if they are missing'''
529 for key, value in extra_info.items():
530 info_dict.setdefault(key, value)
531
532 def extract_info(self, url, download=True, ie_key=None, extra_info={},
533 process=True):
534 '''
535 Returns a list with a dictionary for each video we find.
536 If 'download', also downloads the videos.
537 extra_info is a dict containing the extra values to add to each result
538 '''
539
540 if ie_key:
541 ies = [self.get_info_extractor(ie_key)]
542 else:
543 ies = self._ies
544
545 for ie in ies:
546 if not ie.suitable(url):
547 continue
548
549 if not ie.working():
550 self.report_warning('The program functionality for this site has been marked as broken, '
551 'and will probably not work.')
552
553 try:
554 ie_result = ie.extract(url)
555 if ie_result is None: # Finished already (backwards compatibility; listformats and friends should be moved here)
556 break
557 if isinstance(ie_result, list):
558 # Backwards compatibility: old IE result format
559 ie_result = {
560 '_type': 'compat_list',
561 'entries': ie_result,
562 }
563 self.add_default_extra_info(ie_result, ie, url)
564 if process:
565 return self.process_ie_result(ie_result, download, extra_info)
566 else:
567 return ie_result
568 except ExtractorError as de: # An error we somewhat expected
569 self.report_error(compat_str(de), de.format_traceback())
570 break
571 except MaxDownloadsReached:
572 raise
573 except Exception as e:
574 if self.params.get('ignoreerrors', False):
575 self.report_error(compat_str(e), tb=compat_str(traceback.format_exc()))
576 break
577 else:
578 raise
579 else:
580 self.report_error('no suitable InfoExtractor for URL %s' % url)
581
582 def add_default_extra_info(self, ie_result, ie, url):
583 self.add_extra_info(ie_result, {
584 'extractor': ie.IE_NAME,
585 'webpage_url': url,
586 'webpage_url_basename': url_basename(url),
587 'extractor_key': ie.ie_key(),
588 })
589
590 def process_ie_result(self, ie_result, download=True, extra_info={}):
591 """
592 Take the result of the ie(may be modified) and resolve all unresolved
593 references (URLs, playlist items).
594
595 It will also download the videos if 'download'.
596 Returns the resolved ie_result.
597 """
598
599 result_type = ie_result.get('_type', 'video')
600
601 if result_type in ('url', 'url_transparent'):
602 extract_flat = self.params.get('extract_flat', False)
603 if ((extract_flat == 'in_playlist' and 'playlist' in extra_info) or
604 extract_flat is True):
605 if self.params.get('forcejson', False):
606 self.to_stdout(json.dumps(ie_result))
607 return ie_result
608
609 if result_type == 'video':
610 self.add_extra_info(ie_result, extra_info)
611 return self.process_video_result(ie_result, download=download)
612 elif result_type == 'url':
613 # We have to add extra_info to the results because it may be
614 # contained in a playlist
615 return self.extract_info(ie_result['url'],
616 download,
617 ie_key=ie_result.get('ie_key'),
618 extra_info=extra_info)
619 elif result_type == 'url_transparent':
620 # Use the information from the embedding page
621 info = self.extract_info(
622 ie_result['url'], ie_key=ie_result.get('ie_key'),
623 extra_info=extra_info, download=False, process=False)
624
625 def make_result(embedded_info):
626 new_result = ie_result.copy()
627 for f in ('_type', 'url', 'ext', 'player_url', 'formats',
628 'entries', 'ie_key', 'duration',
629 'subtitles', 'annotations', 'format',
630 'thumbnail', 'thumbnails'):
631 if f in new_result:
632 del new_result[f]
633 if f in embedded_info:
634 new_result[f] = embedded_info[f]
635 return new_result
636 new_result = make_result(info)
637
638 assert new_result.get('_type') != 'url_transparent'
639 if new_result.get('_type') == 'compat_list':
640 new_result['entries'] = [
641 make_result(e) for e in new_result['entries']]
642
643 return self.process_ie_result(
644 new_result, download=download, extra_info=extra_info)
645 elif result_type == 'playlist' or result_type == 'multi_video':
646 # We process each entry in the playlist
647 playlist = ie_result.get('title', None) or ie_result.get('id', None)
648 self.to_screen('[download] Downloading playlist: %s' % playlist)
649
650 playlist_results = []
651
652 playliststart = self.params.get('playliststart', 1) - 1
653 playlistend = self.params.get('playlistend', None)
654 # For backwards compatibility, interpret -1 as whole list
655 if playlistend == -1:
656 playlistend = None
657
658 ie_entries = ie_result['entries']
659 if isinstance(ie_entries, list):
660 n_all_entries = len(ie_entries)
661 entries = ie_entries[playliststart:playlistend]
662 n_entries = len(entries)
663 self.to_screen(
664 "[%s] playlist %s: Collected %d video ids (downloading %d of them)" %
665 (ie_result['extractor'], playlist, n_all_entries, n_entries))
666 elif isinstance(ie_entries, PagedList):
667 entries = ie_entries.getslice(
668 playliststart, playlistend)
669 n_entries = len(entries)
670 self.to_screen(
671 "[%s] playlist %s: Downloading %d videos" %
672 (ie_result['extractor'], playlist, n_entries))
673 else: # iterable
674 entries = list(itertools.islice(
675 ie_entries, playliststart, playlistend))
676 n_entries = len(entries)
677 self.to_screen(
678 "[%s] playlist %s: Downloading %d videos" %
679 (ie_result['extractor'], playlist, n_entries))
680
681 for i, entry in enumerate(entries, 1):
682 self.to_screen('[download] Downloading video #%s of %s' % (i, n_entries))
683 extra = {
684 'n_entries': n_entries,
685 'playlist': playlist,
686 'playlist_id': ie_result.get('id'),
687 'playlist_title': ie_result.get('title'),
688 'playlist_index': i + playliststart,
689 'extractor': ie_result['extractor'],
690 'webpage_url': ie_result['webpage_url'],
691 'webpage_url_basename': url_basename(ie_result['webpage_url']),
692 'extractor_key': ie_result['extractor_key'],
693 }
694
695 reason = self._match_entry(entry)
696 if reason is not None:
697 self.to_screen('[download] ' + reason)
698 continue
699
700 entry_result = self.process_ie_result(entry,
701 download=download,
702 extra_info=extra)
703 playlist_results.append(entry_result)
704 ie_result['entries'] = playlist_results
705 return ie_result
706 elif result_type == 'compat_list':
707 self.report_warning(
708 'Extractor %s returned a compat_list result. '
709 'It needs to be updated.' % ie_result.get('extractor'))
710
711 def _fixup(r):
712 self.add_extra_info(
713 r,
714 {
715 'extractor': ie_result['extractor'],
716 'webpage_url': ie_result['webpage_url'],
717 'webpage_url_basename': url_basename(ie_result['webpage_url']),
718 'extractor_key': ie_result['extractor_key'],
719 }
720 )
721 return r
722 ie_result['entries'] = [
723 self.process_ie_result(_fixup(r), download, extra_info)
724 for r in ie_result['entries']
725 ]
726 return ie_result
727 else:
728 raise Exception('Invalid result type: %s' % result_type)
729
730 def select_format(self, format_spec, available_formats):
731 if format_spec == 'best' or format_spec is None:
732 return available_formats[-1]
733 elif format_spec == 'worst':
734 return available_formats[0]
735 elif format_spec == 'bestaudio':
736 audio_formats = [
737 f for f in available_formats
738 if f.get('vcodec') == 'none']
739 if audio_formats:
740 return audio_formats[-1]
741 elif format_spec == 'worstaudio':
742 audio_formats = [
743 f for f in available_formats
744 if f.get('vcodec') == 'none']
745 if audio_formats:
746 return audio_formats[0]
747 elif format_spec == 'bestvideo':
748 video_formats = [
749 f for f in available_formats
750 if f.get('acodec') == 'none']
751 if video_formats:
752 return video_formats[-1]
753 elif format_spec == 'worstvideo':
754 video_formats = [
755 f for f in available_formats
756 if f.get('acodec') == 'none']
757 if video_formats:
758 return video_formats[0]
759 else:
760 extensions = ['mp4', 'flv', 'webm', '3gp', 'm4a']
761 if format_spec in extensions:
762 filter_f = lambda f: f['ext'] == format_spec
763 else:
764 filter_f = lambda f: f['format_id'] == format_spec
765 matches = list(filter(filter_f, available_formats))
766 if matches:
767 return matches[-1]
768 return None
769
770 def process_video_result(self, info_dict, download=True):
771 assert info_dict.get('_type', 'video') == 'video'
772
773 if 'id' not in info_dict:
774 raise ExtractorError('Missing "id" field in extractor result')
775 if 'title' not in info_dict:
776 raise ExtractorError('Missing "title" field in extractor result')
777
778 if 'playlist' not in info_dict:
779 # It isn't part of a playlist
780 info_dict['playlist'] = None
781 info_dict['playlist_index'] = None
782
783 thumbnails = info_dict.get('thumbnails')
784 if thumbnails:
785 thumbnails.sort(key=lambda t: (
786 t.get('width'), t.get('height'), t.get('url')))
787 for t in thumbnails:
788 if 'width' in t and 'height' in t:
789 t['resolution'] = '%dx%d' % (t['width'], t['height'])
790
791 if thumbnails and 'thumbnail' not in info_dict:
792 info_dict['thumbnail'] = thumbnails[-1]['url']
793
794 if 'display_id' not in info_dict and 'id' in info_dict:
795 info_dict['display_id'] = info_dict['id']
796
797 if info_dict.get('upload_date') is None and info_dict.get('timestamp') is not None:
798 # Working around negative timestamps in Windows
799 # (see http://bugs.python.org/issue1646728)
800 if info_dict['timestamp'] < 0 and os.name == 'nt':
801 info_dict['timestamp'] = 0
802 upload_date = datetime.datetime.utcfromtimestamp(
803 info_dict['timestamp'])
804 info_dict['upload_date'] = upload_date.strftime('%Y%m%d')
805
806 # This extractors handle format selection themselves
807 if info_dict['extractor'] in ['Youku']:
808 if download:
809 self.process_info(info_dict)
810 return info_dict
811
812 # We now pick which formats have to be downloaded
813 if info_dict.get('formats') is None:
814 # There's only one format available
815 formats = [info_dict]
816 else:
817 formats = info_dict['formats']
818
819 if not formats:
820 raise ExtractorError('No video formats found!')
821
822 # We check that all the formats have the format and format_id fields
823 for i, format in enumerate(formats):
824 if 'url' not in format:
825 raise ExtractorError('Missing "url" key in result (index %d)' % i)
826
827 if format.get('format_id') is None:
828 format['format_id'] = compat_str(i)
829 if format.get('format') is None:
830 format['format'] = '{id} - {res}{note}'.format(
831 id=format['format_id'],
832 res=self.format_resolution(format),
833 note=' ({0})'.format(format['format_note']) if format.get('format_note') is not None else '',
834 )
835 # Automatically determine file extension if missing
836 if 'ext' not in format:
837 format['ext'] = determine_ext(format['url']).lower()
838
839 format_limit = self.params.get('format_limit', None)
840 if format_limit:
841 formats = list(takewhile_inclusive(
842 lambda f: f['format_id'] != format_limit, formats
843 ))
844
845 # TODO Central sorting goes here
846
847 if formats[0] is not info_dict:
848 # only set the 'formats' fields if the original info_dict list them
849 # otherwise we end up with a circular reference, the first (and unique)
850 # element in the 'formats' field in info_dict is info_dict itself,
851 # wich can't be exported to json
852 info_dict['formats'] = formats
853 if self.params.get('listformats', None):
854 self.list_formats(info_dict)
855 return
856
857 req_format = self.params.get('format')
858 if req_format is None:
859 req_format = 'best'
860 formats_to_download = []
861 # The -1 is for supporting YoutubeIE
862 if req_format in ('-1', 'all'):
863 formats_to_download = formats
864 else:
865 for rfstr in req_format.split(','):
866 # We can accept formats requested in the format: 34/5/best, we pick
867 # the first that is available, starting from left
868 req_formats = rfstr.split('/')
869 for rf in req_formats:
870 if re.match(r'.+?\+.+?', rf) is not None:
871 # Two formats have been requested like '137+139'
872 format_1, format_2 = rf.split('+')
873 formats_info = (self.select_format(format_1, formats),
874 self.select_format(format_2, formats))
875 if all(formats_info):
876 # The first format must contain the video and the
877 # second the audio
878 if formats_info[0].get('vcodec') == 'none':
879 self.report_error('The first format must '
880 'contain the video, try using '
881 '"-f %s+%s"' % (format_2, format_1))
882 return
883 selected_format = {
884 'requested_formats': formats_info,
885 'format': rf,
886 'ext': formats_info[0]['ext'],
887 }
888 else:
889 selected_format = None
890 else:
891 selected_format = self.select_format(rf, formats)
892 if selected_format is not None:
893 formats_to_download.append(selected_format)
894 break
895 if not formats_to_download:
896 raise ExtractorError('requested format not available',
897 expected=True)
898
899 if download:
900 if len(formats_to_download) > 1:
901 self.to_screen('[info] %s: downloading video in %s formats' % (info_dict['id'], len(formats_to_download)))
902 for format in formats_to_download:
903 new_info = dict(info_dict)
904 new_info.update(format)
905 self.process_info(new_info)
906 # We update the info dict with the best quality format (backwards compatibility)
907 info_dict.update(formats_to_download[-1])
908 return info_dict
909
910 def process_info(self, info_dict):
911 """Process a single resolved IE result."""
912
913 assert info_dict.get('_type', 'video') == 'video'
914
915 max_downloads = self.params.get('max_downloads')
916 if max_downloads is not None:
917 if self._num_downloads >= int(max_downloads):
918 raise MaxDownloadsReached()
919
920 info_dict['fulltitle'] = info_dict['title']
921 if len(info_dict['title']) > 200:
922 info_dict['title'] = info_dict['title'][:197] + '...'
923
924 # Keep for backwards compatibility
925 info_dict['stitle'] = info_dict['title']
926
927 if 'format' not in info_dict:
928 info_dict['format'] = info_dict['ext']
929
930 reason = self._match_entry(info_dict)
931 if reason is not None:
932 self.to_screen('[download] ' + reason)
933 return
934
935 self._num_downloads += 1
936
937 filename = self.prepare_filename(info_dict)
938
939 # Forced printings
940 if self.params.get('forcetitle', False):
941 self.to_stdout(info_dict['fulltitle'])
942 if self.params.get('forceid', False):
943 self.to_stdout(info_dict['id'])
944 if self.params.get('forceurl', False):
945 # For RTMP URLs, also include the playpath
946 self.to_stdout(info_dict['url'] + info_dict.get('play_path', ''))
947 if self.params.get('forcethumbnail', False) and info_dict.get('thumbnail') is not None:
948 self.to_stdout(info_dict['thumbnail'])
949 if self.params.get('forcedescription', False) and info_dict.get('description') is not None:
950 self.to_stdout(info_dict['description'])
951 if self.params.get('forcefilename', False) and filename is not None:
952 self.to_stdout(filename)
953 if self.params.get('forceduration', False) and info_dict.get('duration') is not None:
954 self.to_stdout(formatSeconds(info_dict['duration']))
955 if self.params.get('forceformat', False):
956 self.to_stdout(info_dict['format'])
957 if self.params.get('forcejson', False):
958 info_dict['_filename'] = filename
959 self.to_stdout(json.dumps(info_dict))
960 if self.params.get('dump_single_json', False):
961 info_dict['_filename'] = filename
962
963 # Do nothing else if in simulate mode
964 if self.params.get('simulate', False):
965 return
966
967 if filename is None:
968 return
969
970 try:
971 dn = os.path.dirname(encodeFilename(filename))
972 if dn and not os.path.exists(dn):
973 os.makedirs(dn)
974 except (OSError, IOError) as err:
975 self.report_error('unable to create directory ' + compat_str(err))
976 return
977
978 if self.params.get('writedescription', False):
979 descfn = filename + '.description'
980 if self.params.get('nooverwrites', False) and os.path.exists(encodeFilename(descfn)):
981 self.to_screen('[info] Video description is already present')
982 else:
983 try:
984 self.to_screen('[info] Writing video description to: ' + descfn)
985 with io.open(encodeFilename(descfn), 'w', encoding='utf-8') as descfile:
986 descfile.write(info_dict['description'])
987 except (KeyError, TypeError):
988 self.report_warning('There\'s no description to write.')
989 except (OSError, IOError):
990 self.report_error('Cannot write description file ' + descfn)
991 return
992
993 if self.params.get('writeannotations', False):
994 annofn = filename + '.annotations.xml'
995 if self.params.get('nooverwrites', False) and os.path.exists(encodeFilename(annofn)):
996 self.to_screen('[info] Video annotations are already present')
997 else:
998 try:
999 self.to_screen('[info] Writing video annotations to: ' + annofn)
1000 with io.open(encodeFilename(annofn), 'w', encoding='utf-8') as annofile:
1001 annofile.write(info_dict['annotations'])
1002 except (KeyError, TypeError):
1003 self.report_warning('There are no annotations to write.')
1004 except (OSError, IOError):
1005 self.report_error('Cannot write annotations file: ' + annofn)
1006 return
1007
1008 subtitles_are_requested = any([self.params.get('writesubtitles', False),
1009 self.params.get('writeautomaticsub')])
1010
1011 if subtitles_are_requested and 'subtitles' in info_dict and info_dict['subtitles']:
1012 # subtitles download errors are already managed as troubles in relevant IE
1013 # that way it will silently go on when used with unsupporting IE
1014 subtitles = info_dict['subtitles']
1015 sub_format = self.params.get('subtitlesformat', 'srt')
1016 for sub_lang in subtitles.keys():
1017 sub = subtitles[sub_lang]
1018 if sub is None:
1019 continue
1020 try:
1021 sub_filename = subtitles_filename(filename, sub_lang, sub_format)
1022 if self.params.get('nooverwrites', False) and os.path.exists(encodeFilename(sub_filename)):
1023 self.to_screen('[info] Video subtitle %s.%s is already_present' % (sub_lang, sub_format))
1024 else:
1025 self.to_screen('[info] Writing video subtitles to: ' + sub_filename)
1026 with io.open(encodeFilename(sub_filename), 'w', encoding='utf-8') as subfile:
1027 subfile.write(sub)
1028 except (OSError, IOError):
1029 self.report_error('Cannot write subtitles file ' + sub_filename)
1030 return
1031
1032 if self.params.get('writeinfojson', False):
1033 infofn = os.path.splitext(filename)[0] + '.info.json'
1034 if self.params.get('nooverwrites', False) and os.path.exists(encodeFilename(infofn)):
1035 self.to_screen('[info] Video description metadata is already present')
1036 else:
1037 self.to_screen('[info] Writing video description metadata as JSON to: ' + infofn)
1038 try:
1039 write_json_file(info_dict, infofn)
1040 except (OSError, IOError):
1041 self.report_error('Cannot write metadata to JSON file ' + infofn)
1042 return
1043
1044 if self.params.get('writethumbnail', False):
1045 if info_dict.get('thumbnail') is not None:
1046 thumb_format = determine_ext(info_dict['thumbnail'], 'jpg')
1047 thumb_filename = os.path.splitext(filename)[0] + '.' + thumb_format
1048 if self.params.get('nooverwrites', False) and os.path.exists(encodeFilename(thumb_filename)):
1049 self.to_screen('[%s] %s: Thumbnail is already present' %
1050 (info_dict['extractor'], info_dict['id']))
1051 else:
1052 self.to_screen('[%s] %s: Downloading thumbnail ...' %
1053 (info_dict['extractor'], info_dict['id']))
1054 try:
1055 uf = self.urlopen(info_dict['thumbnail'])
1056 with open(thumb_filename, 'wb') as thumbf:
1057 shutil.copyfileobj(uf, thumbf)
1058 self.to_screen('[%s] %s: Writing thumbnail to: %s' %
1059 (info_dict['extractor'], info_dict['id'], thumb_filename))
1060 except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
1061 self.report_warning('Unable to download thumbnail "%s": %s' %
1062 (info_dict['thumbnail'], compat_str(err)))
1063
1064 if not self.params.get('skip_download', False):
1065 if self.params.get('nooverwrites', False) and os.path.exists(encodeFilename(filename)):
1066 success = True
1067 else:
1068 try:
1069 def dl(name, info):
1070 fd = get_suitable_downloader(info)(self, self.params)
1071 for ph in self._progress_hooks:
1072 fd.add_progress_hook(ph)
1073 if self.params.get('verbose'):
1074 self.to_stdout('[debug] Invoking downloader on %r' % info.get('url'))
1075 return fd.download(name, info)
1076 if info_dict.get('requested_formats') is not None:
1077 downloaded = []
1078 success = True
1079 merger = FFmpegMergerPP(self, not self.params.get('keepvideo'))
1080 if not merger._executable:
1081 postprocessors = []
1082 self.report_warning('You have requested multiple '
1083 'formats but ffmpeg or avconv are not installed.'
1084 ' The formats won\'t be merged')
1085 else:
1086 postprocessors = [merger]
1087 for f in info_dict['requested_formats']:
1088 new_info = dict(info_dict)
1089 new_info.update(f)
1090 fname = self.prepare_filename(new_info)
1091 fname = prepend_extension(fname, 'f%s' % f['format_id'])
1092 downloaded.append(fname)
1093 partial_success = dl(fname, new_info)
1094 success = success and partial_success
1095 info_dict['__postprocessors'] = postprocessors
1096 info_dict['__files_to_merge'] = downloaded
1097 else:
1098 # Just a single file
1099 success = dl(filename, info_dict)
1100 except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
1101 self.report_error('unable to download video data: %s' % str(err))
1102 return
1103 except (OSError, IOError) as err:
1104 raise UnavailableVideoError(err)
1105 except (ContentTooShortError, ) as err:
1106 self.report_error('content too short (expected %s bytes and served %s)' % (err.expected, err.downloaded))
1107 return
1108
1109 if success:
1110 try:
1111 self.post_process(filename, info_dict)
1112 except (PostProcessingError) as err:
1113 self.report_error('postprocessing: %s' % str(err))
1114 return
1115
1116 self.record_download_archive(info_dict)
1117
1118 def download(self, url_list):
1119 """Download a given list of URLs."""
1120 outtmpl = self.params.get('outtmpl', DEFAULT_OUTTMPL)
1121 if (len(url_list) > 1 and
1122 '%' not in outtmpl
1123 and self.params.get('max_downloads') != 1):
1124 raise SameFileError(outtmpl)
1125
1126 for url in url_list:
1127 try:
1128 # It also downloads the videos
1129 res = self.extract_info(url)
1130 except UnavailableVideoError:
1131 self.report_error('unable to download video')
1132 except MaxDownloadsReached:
1133 self.to_screen('[info] Maximum number of downloaded files reached.')
1134 raise
1135 else:
1136 if self.params.get('dump_single_json', False):
1137 self.to_stdout(json.dumps(res))
1138
1139 return self._download_retcode
1140
1141 def download_with_info_file(self, info_filename):
1142 with io.open(info_filename, 'r', encoding='utf-8') as f:
1143 info = json.load(f)
1144 try:
1145 self.process_ie_result(info, download=True)
1146 except DownloadError:
1147 webpage_url = info.get('webpage_url')
1148 if webpage_url is not None:
1149 self.report_warning('The info failed to download, trying with "%s"' % webpage_url)
1150 return self.download([webpage_url])
1151 else:
1152 raise
1153 return self._download_retcode
1154
1155 def post_process(self, filename, ie_info):
1156 """Run all the postprocessors on the given file."""
1157 info = dict(ie_info)
1158 info['filepath'] = filename
1159 keep_video = None
1160 pps_chain = []
1161 if ie_info.get('__postprocessors') is not None:
1162 pps_chain.extend(ie_info['__postprocessors'])
1163 pps_chain.extend(self._pps)
1164 for pp in pps_chain:
1165 try:
1166 keep_video_wish, new_info = pp.run(info)
1167 if keep_video_wish is not None:
1168 if keep_video_wish:
1169 keep_video = keep_video_wish
1170 elif keep_video is None:
1171 # No clear decision yet, let IE decide
1172 keep_video = keep_video_wish
1173 except PostProcessingError as e:
1174 self.report_error(e.msg)
1175 if keep_video is False and not self.params.get('keepvideo', False):
1176 try:
1177 self.to_screen('Deleting original file %s (pass -k to keep)' % filename)
1178 os.remove(encodeFilename(filename))
1179 except (IOError, OSError):
1180 self.report_warning('Unable to remove downloaded video file')
1181
1182 def _make_archive_id(self, info_dict):
1183 # Future-proof against any change in case
1184 # and backwards compatibility with prior versions
1185 extractor = info_dict.get('extractor_key')
1186 if extractor is None:
1187 if 'id' in info_dict:
1188 extractor = info_dict.get('ie_key') # key in a playlist
1189 if extractor is None:
1190 return None # Incomplete video information
1191 return extractor.lower() + ' ' + info_dict['id']
1192
1193 def in_download_archive(self, info_dict):
1194 fn = self.params.get('download_archive')
1195 if fn is None:
1196 return False
1197
1198 vid_id = self._make_archive_id(info_dict)
1199 if vid_id is None:
1200 return False # Incomplete video information
1201
1202 try:
1203 with locked_file(fn, 'r', encoding='utf-8') as archive_file:
1204 for line in archive_file:
1205 if line.strip() == vid_id:
1206 return True
1207 except IOError as ioe:
1208 if ioe.errno != errno.ENOENT:
1209 raise
1210 return False
1211
1212 def record_download_archive(self, info_dict):
1213 fn = self.params.get('download_archive')
1214 if fn is None:
1215 return
1216 vid_id = self._make_archive_id(info_dict)
1217 assert vid_id
1218 with locked_file(fn, 'a', encoding='utf-8') as archive_file:
1219 archive_file.write(vid_id + '\n')
1220
1221 @staticmethod
1222 def format_resolution(format, default='unknown'):
1223 if format.get('vcodec') == 'none':
1224 return 'audio only'
1225 if format.get('resolution') is not None:
1226 return format['resolution']
1227 if format.get('height') is not None:
1228 if format.get('width') is not None:
1229 res = '%sx%s' % (format['width'], format['height'])
1230 else:
1231 res = '%sp' % format['height']
1232 elif format.get('width') is not None:
1233 res = '?x%d' % format['width']
1234 else:
1235 res = default
1236 return res
1237
1238 def _format_note(self, fdict):
1239 res = ''
1240 if fdict.get('ext') in ['f4f', 'f4m']:
1241 res += '(unsupported) '
1242 if fdict.get('format_note') is not None:
1243 res += fdict['format_note'] + ' '
1244 if fdict.get('tbr') is not None:
1245 res += '%4dk ' % fdict['tbr']
1246 if fdict.get('container') is not None:
1247 if res:
1248 res += ', '
1249 res += '%s container' % fdict['container']
1250 if (fdict.get('vcodec') is not None and
1251 fdict.get('vcodec') != 'none'):
1252 if res:
1253 res += ', '
1254 res += fdict['vcodec']
1255 if fdict.get('vbr') is not None:
1256 res += '@'
1257 elif fdict.get('vbr') is not None and fdict.get('abr') is not None:
1258 res += 'video@'
1259 if fdict.get('vbr') is not None:
1260 res += '%4dk' % fdict['vbr']
1261 if fdict.get('fps') is not None:
1262 res += ', %sfps' % fdict['fps']
1263 if fdict.get('acodec') is not None:
1264 if res:
1265 res += ', '
1266 if fdict['acodec'] == 'none':
1267 res += 'video only'
1268 else:
1269 res += '%-5s' % fdict['acodec']
1270 elif fdict.get('abr') is not None:
1271 if res:
1272 res += ', '
1273 res += 'audio'
1274 if fdict.get('abr') is not None:
1275 res += '@%3dk' % fdict['abr']
1276 if fdict.get('asr') is not None:
1277 res += ' (%5dHz)' % fdict['asr']
1278 if fdict.get('filesize') is not None:
1279 if res:
1280 res += ', '
1281 res += format_bytes(fdict['filesize'])
1282 elif fdict.get('filesize_approx') is not None:
1283 if res:
1284 res += ', '
1285 res += '~' + format_bytes(fdict['filesize_approx'])
1286 return res
1287
1288 def list_formats(self, info_dict):
1289 def line(format, idlen=20):
1290 return (('%-' + compat_str(idlen + 1) + 's%-10s%-12s%s') % (
1291 format['format_id'],
1292 format['ext'],
1293 self.format_resolution(format),
1294 self._format_note(format),
1295 ))
1296
1297 formats = info_dict.get('formats', [info_dict])
1298 idlen = max(len('format code'),
1299 max(len(f['format_id']) for f in formats))
1300 formats_s = [line(f, idlen) for f in formats]
1301 if len(formats) > 1:
1302 formats_s[0] += (' ' if self._format_note(formats[0]) else '') + '(worst)'
1303 formats_s[-1] += (' ' if self._format_note(formats[-1]) else '') + '(best)'
1304
1305 header_line = line({
1306 'format_id': 'format code', 'ext': 'extension',
1307 'resolution': 'resolution', 'format_note': 'note'}, idlen=idlen)
1308 self.to_screen('[info] Available formats for %s:\n%s\n%s' %
1309 (info_dict['id'], header_line, '\n'.join(formats_s)))
1310
1311 def urlopen(self, req):
1312 """ Start an HTTP download """
1313
1314 # According to RFC 3986, URLs can not contain non-ASCII characters, however this is not
1315 # always respected by websites, some tend to give out URLs with non percent-encoded
1316 # non-ASCII characters (see telemb.py, ard.py [#3412])
1317 # urllib chokes on URLs with non-ASCII characters (see http://bugs.python.org/issue3991)
1318 # To work around aforementioned issue we will replace request's original URL with
1319 # percent-encoded one
1320 req_is_string = isinstance(req, basestring if sys.version_info < (3, 0) else compat_str)
1321 url = req if req_is_string else req.get_full_url()
1322 url_escaped = escape_url(url)
1323
1324 # Substitute URL if any change after escaping
1325 if url != url_escaped:
1326 if req_is_string:
1327 req = url_escaped
1328 else:
1329 req = compat_urllib_request.Request(
1330 url_escaped, data=req.data, headers=req.headers,
1331 origin_req_host=req.origin_req_host, unverifiable=req.unverifiable)
1332
1333 return self._opener.open(req, timeout=self._socket_timeout)
1334
1335 def print_debug_header(self):
1336 if not self.params.get('verbose'):
1337 return
1338
1339 if type('') is not compat_str:
1340 # Python 2.6 on SLES11 SP1 (https://github.com/rg3/youtube-dl/issues/3326)
1341 self.report_warning(
1342 'Your Python is broken! Update to a newer and supported version')
1343
1344 stdout_encoding = getattr(
1345 sys.stdout, 'encoding', 'missing (%s)' % type(sys.stdout).__name__)
1346 encoding_str = (
1347 '[debug] Encodings: locale %s, fs %s, out %s, pref %s\n' % (
1348 locale.getpreferredencoding(),
1349 sys.getfilesystemencoding(),
1350 stdout_encoding,
1351 self.get_encoding()))
1352 write_string(encoding_str, encoding=None)
1353
1354 self._write_string('[debug] youtube-dl version ' + __version__ + '\n')
1355 try:
1356 sp = subprocess.Popen(
1357 ['git', 'rev-parse', '--short', 'HEAD'],
1358 stdout=subprocess.PIPE, stderr=subprocess.PIPE,
1359 cwd=os.path.dirname(os.path.abspath(__file__)))
1360 out, err = sp.communicate()
1361 out = out.decode().strip()
1362 if re.match('[0-9a-f]+', out):
1363 self._write_string('[debug] Git HEAD: ' + out + '\n')
1364 except:
1365 try:
1366 sys.exc_clear()
1367 except:
1368 pass
1369 self._write_string('[debug] Python version %s - %s\n' % (
1370 platform.python_version(), platform_name()))
1371
1372 exe_versions = FFmpegPostProcessor.get_versions()
1373 exe_versions['rtmpdump'] = rtmpdump_version()
1374 exe_str = ', '.join(
1375 '%s %s' % (exe, v)
1376 for exe, v in sorted(exe_versions.items())
1377 if v
1378 )
1379 if not exe_str:
1380 exe_str = 'none'
1381 self._write_string('[debug] exe versions: %s\n' % exe_str)
1382
1383 proxy_map = {}
1384 for handler in self._opener.handlers:
1385 if hasattr(handler, 'proxies'):
1386 proxy_map.update(handler.proxies)
1387 self._write_string('[debug] Proxy map: ' + compat_str(proxy_map) + '\n')
1388
1389 def _setup_opener(self):
1390 timeout_val = self.params.get('socket_timeout')
1391 self._socket_timeout = 600 if timeout_val is None else float(timeout_val)
1392
1393 opts_cookiefile = self.params.get('cookiefile')
1394 opts_proxy = self.params.get('proxy')
1395
1396 if opts_cookiefile is None:
1397 self.cookiejar = compat_cookiejar.CookieJar()
1398 else:
1399 self.cookiejar = compat_cookiejar.MozillaCookieJar(
1400 opts_cookiefile)
1401 if os.access(opts_cookiefile, os.R_OK):
1402 self.cookiejar.load()
1403
1404 cookie_processor = compat_urllib_request.HTTPCookieProcessor(
1405 self.cookiejar)
1406 if opts_proxy is not None:
1407 if opts_proxy == '':
1408 proxies = {}
1409 else:
1410 proxies = {'http': opts_proxy, 'https': opts_proxy}
1411 else:
1412 proxies = compat_urllib_request.getproxies()
1413 # Set HTTPS proxy to HTTP one if given (https://github.com/rg3/youtube-dl/issues/805)
1414 if 'http' in proxies and 'https' not in proxies:
1415 proxies['https'] = proxies['http']
1416 proxy_handler = compat_urllib_request.ProxyHandler(proxies)
1417
1418 debuglevel = 1 if self.params.get('debug_printtraffic') else 0
1419 https_handler = make_HTTPS_handler(
1420 self.params.get('nocheckcertificate', False), debuglevel=debuglevel)
1421 ydlh = YoutubeDLHandler(debuglevel=debuglevel)
1422 opener = compat_urllib_request.build_opener(
1423 https_handler, proxy_handler, cookie_processor, ydlh)
1424 # Delete the default user-agent header, which would otherwise apply in
1425 # cases where our custom HTTP handler doesn't come into play
1426 # (See https://github.com/rg3/youtube-dl/issues/1309 for details)
1427 opener.addheaders = []
1428 self._opener = opener
1429
1430 def encode(self, s):
1431 if isinstance(s, bytes):
1432 return s # Already encoded
1433
1434 try:
1435 return s.encode(self.get_encoding())
1436 except UnicodeEncodeError as err:
1437 err.reason = err.reason + '. Check your system encoding configuration or use the --encoding option.'
1438 raise
1439
1440 def get_encoding(self):
1441 encoding = self.params.get('encoding')
1442 if encoding is None:
1443 encoding = preferredencoding()
1444 return encoding