]> jfr.im git - yt-dlp.git/blob - yt_dlp/extractor/common.py
[cleanup] Use `_html_extract_title`
[yt-dlp.git] / yt_dlp / extractor / common.py
1 # coding: utf-8
2 from __future__ import unicode_literals
3
4 import base64
5 import collections
6 import hashlib
7 import itertools
8 import json
9 import netrc
10 import os
11 import random
12 import re
13 import sys
14 import time
15 import math
16
17 from ..compat import (
18 compat_cookiejar_Cookie,
19 compat_cookies_SimpleCookie,
20 compat_etree_Element,
21 compat_etree_fromstring,
22 compat_expanduser,
23 compat_getpass,
24 compat_http_client,
25 compat_os_name,
26 compat_str,
27 compat_urllib_error,
28 compat_urllib_parse_unquote,
29 compat_urllib_parse_urlencode,
30 compat_urllib_request,
31 compat_urlparse,
32 compat_xml_parse_error,
33 )
34 from ..downloader import FileDownloader
35 from ..downloader.f4m import (
36 get_base_url,
37 remove_encrypted_media,
38 )
39 from ..utils import (
40 age_restricted,
41 base_url,
42 bug_reports_message,
43 clean_html,
44 compiled_regex_type,
45 determine_ext,
46 determine_protocol,
47 dict_get,
48 encode_data_uri,
49 error_to_compat_str,
50 extract_attributes,
51 ExtractorError,
52 filter_dict,
53 fix_xml_ampersands,
54 float_or_none,
55 format_field,
56 GeoRestrictedError,
57 GeoUtils,
58 int_or_none,
59 join_nonempty,
60 js_to_json,
61 JSON_LD_RE,
62 mimetype2ext,
63 network_exceptions,
64 NO_DEFAULT,
65 orderedSet,
66 parse_bitrate,
67 parse_codecs,
68 parse_duration,
69 parse_iso8601,
70 parse_m3u8_attributes,
71 parse_resolution,
72 RegexNotFoundError,
73 sanitize_filename,
74 sanitized_Request,
75 str_or_none,
76 str_to_int,
77 strip_or_none,
78 traverse_obj,
79 try_get,
80 unescapeHTML,
81 UnsupportedError,
82 unified_strdate,
83 unified_timestamp,
84 update_Request,
85 update_url_query,
86 url_basename,
87 url_or_none,
88 urljoin,
89 variadic,
90 xpath_element,
91 xpath_text,
92 xpath_with_ns,
93 )
94
95
96 class InfoExtractor(object):
97 """Information Extractor class.
98
99 Information extractors are the classes that, given a URL, extract
100 information about the video (or videos) the URL refers to. This
101 information includes the real video URL, the video title, author and
102 others. The information is stored in a dictionary which is then
103 passed to the YoutubeDL. The YoutubeDL processes this
104 information possibly downloading the video to the file system, among
105 other possible outcomes.
106
107 The type field determines the type of the result.
108 By far the most common value (and the default if _type is missing) is
109 "video", which indicates a single video.
110
111 For a video, the dictionaries must include the following fields:
112
113 id: Video identifier.
114 title: Video title, unescaped.
115
116 Additionally, it must contain either a formats entry or a url one:
117
118 formats: A list of dictionaries for each format available, ordered
119 from worst to best quality.
120
121 Potential fields:
122 * url The mandatory URL representing the media:
123 for plain file media - HTTP URL of this file,
124 for RTMP - RTMP URL,
125 for HLS - URL of the M3U8 media playlist,
126 for HDS - URL of the F4M manifest,
127 for DASH
128 - HTTP URL to plain file media (in case of
129 unfragmented media)
130 - URL of the MPD manifest or base URL
131 representing the media if MPD manifest
132 is parsed from a string (in case of
133 fragmented media)
134 for MSS - URL of the ISM manifest.
135 * manifest_url
136 The URL of the manifest file in case of
137 fragmented media:
138 for HLS - URL of the M3U8 master playlist,
139 for HDS - URL of the F4M manifest,
140 for DASH - URL of the MPD manifest,
141 for MSS - URL of the ISM manifest.
142 * ext Will be calculated from URL if missing
143 * format A human-readable description of the format
144 ("mp4 container with h264/opus").
145 Calculated from the format_id, width, height.
146 and format_note fields if missing.
147 * format_id A short description of the format
148 ("mp4_h264_opus" or "19").
149 Technically optional, but strongly recommended.
150 * format_note Additional info about the format
151 ("3D" or "DASH video")
152 * width Width of the video, if known
153 * height Height of the video, if known
154 * resolution Textual description of width and height
155 * dynamic_range The dynamic range of the video. One of:
156 "SDR" (None), "HDR10", "HDR10+, "HDR12", "HLG, "DV"
157 * tbr Average bitrate of audio and video in KBit/s
158 * abr Average audio bitrate in KBit/s
159 * acodec Name of the audio codec in use
160 * asr Audio sampling rate in Hertz
161 * vbr Average video bitrate in KBit/s
162 * fps Frame rate
163 * vcodec Name of the video codec in use
164 * container Name of the container format
165 * filesize The number of bytes, if known in advance
166 * filesize_approx An estimate for the number of bytes
167 * player_url SWF Player URL (used for rtmpdump).
168 * protocol The protocol that will be used for the actual
169 download, lower-case. One of "http", "https" or
170 one of the protocols defined in downloader.PROTOCOL_MAP
171 * fragment_base_url
172 Base URL for fragments. Each fragment's path
173 value (if present) will be relative to
174 this URL.
175 * fragments A list of fragments of a fragmented media.
176 Each fragment entry must contain either an url
177 or a path. If an url is present it should be
178 considered by a client. Otherwise both path and
179 fragment_base_url must be present. Here is
180 the list of all potential fields:
181 * "url" - fragment's URL
182 * "path" - fragment's path relative to
183 fragment_base_url
184 * "duration" (optional, int or float)
185 * "filesize" (optional, int)
186 * is_from_start Is a live format that can be downloaded
187 from the start. Boolean
188 * preference Order number of this format. If this field is
189 present and not None, the formats get sorted
190 by this field, regardless of all other values.
191 -1 for default (order by other properties),
192 -2 or smaller for less than default.
193 < -1000 to hide the format (if there is
194 another one which is strictly better)
195 * language Language code, e.g. "de" or "en-US".
196 * language_preference Is this in the language mentioned in
197 the URL?
198 10 if it's what the URL is about,
199 -1 for default (don't know),
200 -10 otherwise, other values reserved for now.
201 * quality Order number of the video quality of this
202 format, irrespective of the file format.
203 -1 for default (order by other properties),
204 -2 or smaller for less than default.
205 * source_preference Order number for this video source
206 (quality takes higher priority)
207 -1 for default (order by other properties),
208 -2 or smaller for less than default.
209 * http_headers A dictionary of additional HTTP headers
210 to add to the request.
211 * stretched_ratio If given and not 1, indicates that the
212 video's pixels are not square.
213 width : height ratio as float.
214 * no_resume The server does not support resuming the
215 (HTTP or RTMP) download. Boolean.
216 * has_drm The format has DRM and cannot be downloaded. Boolean
217 * downloader_options A dictionary of downloader options as
218 described in FileDownloader
219 RTMP formats can also have the additional fields: page_url,
220 app, play_path, tc_url, flash_version, rtmp_live, rtmp_conn,
221 rtmp_protocol, rtmp_real_time
222
223 url: Final video URL.
224 ext: Video filename extension.
225 format: The video format, defaults to ext (used for --get-format)
226 player_url: SWF Player URL (used for rtmpdump).
227
228 The following fields are optional:
229
230 direct: True if a direct video file was given (must only be set by GenericIE)
231 alt_title: A secondary title of the video.
232 display_id An alternative identifier for the video, not necessarily
233 unique, but available before title. Typically, id is
234 something like "4234987", title "Dancing naked mole rats",
235 and display_id "dancing-naked-mole-rats"
236 thumbnails: A list of dictionaries, with the following entries:
237 * "id" (optional, string) - Thumbnail format ID
238 * "url"
239 * "preference" (optional, int) - quality of the image
240 * "width" (optional, int)
241 * "height" (optional, int)
242 * "resolution" (optional, string "{width}x{height}",
243 deprecated)
244 * "filesize" (optional, int)
245 * "http_headers" (dict) - HTTP headers for the request
246 thumbnail: Full URL to a video thumbnail image.
247 description: Full video description.
248 uploader: Full name of the video uploader.
249 license: License name the video is licensed under.
250 creator: The creator of the video.
251 timestamp: UNIX timestamp of the moment the video was uploaded
252 upload_date: Video upload date in UTC (YYYYMMDD).
253 If not explicitly set, calculated from timestamp
254 release_timestamp: UNIX timestamp of the moment the video was released.
255 If it is not clear whether to use timestamp or this, use the former
256 release_date: The date (YYYYMMDD) when the video was released in UTC.
257 If not explicitly set, calculated from release_timestamp
258 modified_timestamp: UNIX timestamp of the moment the video was last modified.
259 modified_date: The date (YYYYMMDD) when the video was last modified in UTC.
260 If not explicitly set, calculated from modified_timestamp
261 uploader_id: Nickname or id of the video uploader.
262 uploader_url: Full URL to a personal webpage of the video uploader.
263 channel: Full name of the channel the video is uploaded on.
264 Note that channel fields may or may not repeat uploader
265 fields. This depends on a particular extractor.
266 channel_id: Id of the channel.
267 channel_url: Full URL to a channel webpage.
268 channel_follower_count: Number of followers of the channel.
269 location: Physical location where the video was filmed.
270 subtitles: The available subtitles as a dictionary in the format
271 {tag: subformats}. "tag" is usually a language code, and
272 "subformats" is a list sorted from lower to higher
273 preference, each element is a dictionary with the "ext"
274 entry and one of:
275 * "data": The subtitles file contents
276 * "url": A URL pointing to the subtitles file
277 It can optionally also have:
278 * "name": Name or description of the subtitles
279 * "http_headers": A dictionary of additional HTTP headers
280 to add to the request.
281 "ext" will be calculated from URL if missing
282 automatic_captions: Like 'subtitles'; contains automatically generated
283 captions instead of normal subtitles
284 duration: Length of the video in seconds, as an integer or float.
285 view_count: How many users have watched the video on the platform.
286 like_count: Number of positive ratings of the video
287 dislike_count: Number of negative ratings of the video
288 repost_count: Number of reposts of the video
289 average_rating: Average rating give by users, the scale used depends on the webpage
290 comment_count: Number of comments on the video
291 comments: A list of comments, each with one or more of the following
292 properties (all but one of text or html optional):
293 * "author" - human-readable name of the comment author
294 * "author_id" - user ID of the comment author
295 * "author_thumbnail" - The thumbnail of the comment author
296 * "id" - Comment ID
297 * "html" - Comment as HTML
298 * "text" - Plain text of the comment
299 * "timestamp" - UNIX timestamp of comment
300 * "parent" - ID of the comment this one is replying to.
301 Set to "root" to indicate that this is a
302 comment to the original video.
303 * "like_count" - Number of positive ratings of the comment
304 * "dislike_count" - Number of negative ratings of the comment
305 * "is_favorited" - Whether the comment is marked as
306 favorite by the video uploader
307 * "author_is_uploader" - Whether the comment is made by
308 the video uploader
309 age_limit: Age restriction for the video, as an integer (years)
310 webpage_url: The URL to the video webpage, if given to yt-dlp it
311 should allow to get the same result again. (It will be set
312 by YoutubeDL if it's missing)
313 categories: A list of categories that the video falls in, for example
314 ["Sports", "Berlin"]
315 tags: A list of tags assigned to the video, e.g. ["sweden", "pop music"]
316 cast: A list of the video cast
317 is_live: True, False, or None (=unknown). Whether this video is a
318 live stream that goes on instead of a fixed-length video.
319 was_live: True, False, or None (=unknown). Whether this video was
320 originally a live stream.
321 live_status: 'is_live', 'is_upcoming', 'was_live', 'not_live' or None (=unknown)
322 If absent, automatically set from is_live, was_live
323 start_time: Time in seconds where the reproduction should start, as
324 specified in the URL.
325 end_time: Time in seconds where the reproduction should end, as
326 specified in the URL.
327 chapters: A list of dictionaries, with the following entries:
328 * "start_time" - The start time of the chapter in seconds
329 * "end_time" - The end time of the chapter in seconds
330 * "title" (optional, string)
331 playable_in_embed: Whether this video is allowed to play in embedded
332 players on other sites. Can be True (=always allowed),
333 False (=never allowed), None (=unknown), or a string
334 specifying the criteria for embedability (Eg: 'whitelist')
335 availability: Under what condition the video is available. One of
336 'private', 'premium_only', 'subscriber_only', 'needs_auth',
337 'unlisted' or 'public'. Use 'InfoExtractor._availability'
338 to set it
339 __post_extractor: A function to be called just before the metadata is
340 written to either disk, logger or console. The function
341 must return a dict which will be added to the info_dict.
342 This is usefull for additional information that is
343 time-consuming to extract. Note that the fields thus
344 extracted will not be available to output template and
345 match_filter. So, only "comments" and "comment_count" are
346 currently allowed to be extracted via this method.
347
348 The following fields should only be used when the video belongs to some logical
349 chapter or section:
350
351 chapter: Name or title of the chapter the video belongs to.
352 chapter_number: Number of the chapter the video belongs to, as an integer.
353 chapter_id: Id of the chapter the video belongs to, as a unicode string.
354
355 The following fields should only be used when the video is an episode of some
356 series, programme or podcast:
357
358 series: Title of the series or programme the video episode belongs to.
359 series_id: Id of the series or programme the video episode belongs to, as a unicode string.
360 season: Title of the season the video episode belongs to.
361 season_number: Number of the season the video episode belongs to, as an integer.
362 season_id: Id of the season the video episode belongs to, as a unicode string.
363 episode: Title of the video episode. Unlike mandatory video title field,
364 this field should denote the exact title of the video episode
365 without any kind of decoration.
366 episode_number: Number of the video episode within a season, as an integer.
367 episode_id: Id of the video episode, as a unicode string.
368
369 The following fields should only be used when the media is a track or a part of
370 a music album:
371
372 track: Title of the track.
373 track_number: Number of the track within an album or a disc, as an integer.
374 track_id: Id of the track (useful in case of custom indexing, e.g. 6.iii),
375 as a unicode string.
376 artist: Artist(s) of the track.
377 genre: Genre(s) of the track.
378 album: Title of the album the track belongs to.
379 album_type: Type of the album (e.g. "Demo", "Full-length", "Split", "Compilation", etc).
380 album_artist: List of all artists appeared on the album (e.g.
381 "Ash Borer / Fell Voices" or "Various Artists", useful for splits
382 and compilations).
383 disc_number: Number of the disc or other physical medium the track belongs to,
384 as an integer.
385 release_year: Year (YYYY) when the album was released.
386 composer: Composer of the piece
387
388 Unless mentioned otherwise, the fields should be Unicode strings.
389
390 Unless mentioned otherwise, None is equivalent to absence of information.
391
392
393 _type "playlist" indicates multiple videos.
394 There must be a key "entries", which is a list, an iterable, or a PagedList
395 object, each element of which is a valid dictionary by this specification.
396
397 Additionally, playlists can have "id", "title", and any other relevent
398 attributes with the same semantics as videos (see above).
399
400 It can also have the following optional fields:
401
402 playlist_count: The total number of videos in a playlist. If not given,
403 YoutubeDL tries to calculate it from "entries"
404
405
406 _type "multi_video" indicates that there are multiple videos that
407 form a single show, for examples multiple acts of an opera or TV episode.
408 It must have an entries key like a playlist and contain all the keys
409 required for a video at the same time.
410
411
412 _type "url" indicates that the video must be extracted from another
413 location, possibly by a different extractor. Its only required key is:
414 "url" - the next URL to extract.
415 The key "ie_key" can be set to the class name (minus the trailing "IE",
416 e.g. "Youtube") if the extractor class is known in advance.
417 Additionally, the dictionary may have any properties of the resolved entity
418 known in advance, for example "title" if the title of the referred video is
419 known ahead of time.
420
421
422 _type "url_transparent" entities have the same specification as "url", but
423 indicate that the given additional information is more precise than the one
424 associated with the resolved URL.
425 This is useful when a site employs a video service that hosts the video and
426 its technical metadata, but that video service does not embed a useful
427 title, description etc.
428
429
430 Subclasses of this should define a _VALID_URL regexp and, re-define the
431 _real_extract() and (optionally) _real_initialize() methods.
432 Probably, they should also be added to the list of extractors.
433
434 Subclasses may also override suitable() if necessary, but ensure the function
435 signature is preserved and that this function imports everything it needs
436 (except other extractors), so that lazy_extractors works correctly.
437
438 To support username + password (or netrc) login, the extractor must define a
439 _NETRC_MACHINE and re-define _perform_login(username, password) and
440 (optionally) _initialize_pre_login() methods. The _perform_login method will
441 be called between _initialize_pre_login and _real_initialize if credentials
442 are passed by the user. In cases where it is necessary to have the login
443 process as part of the extraction rather than initialization, _perform_login
444 can be left undefined.
445
446 _GEO_BYPASS attribute may be set to False in order to disable
447 geo restriction bypass mechanisms for a particular extractor.
448 Though it won't disable explicit geo restriction bypass based on
449 country code provided with geo_bypass_country.
450
451 _GEO_COUNTRIES attribute may contain a list of presumably geo unrestricted
452 countries for this extractor. One of these countries will be used by
453 geo restriction bypass mechanism right away in order to bypass
454 geo restriction, of course, if the mechanism is not disabled.
455
456 _GEO_IP_BLOCKS attribute may contain a list of presumably geo unrestricted
457 IP blocks in CIDR notation for this extractor. One of these IP blocks
458 will be used by geo restriction bypass mechanism similarly
459 to _GEO_COUNTRIES.
460
461 The _WORKING attribute should be set to False for broken IEs
462 in order to warn the users and skip the tests.
463 """
464
465 _ready = False
466 _downloader = None
467 _x_forwarded_for_ip = None
468 _GEO_BYPASS = True
469 _GEO_COUNTRIES = None
470 _GEO_IP_BLOCKS = None
471 _WORKING = True
472 _NETRC_MACHINE = None
473 IE_DESC = None
474
475 _LOGIN_HINTS = {
476 'any': 'Use --cookies, --cookies-from-browser, --username and --password, or --netrc to provide account credentials',
477 'cookies': (
478 'Use --cookies-from-browser or --cookies for the authentication. '
479 'See https://github.com/ytdl-org/youtube-dl#how-do-i-pass-cookies-to-youtube-dl for how to manually pass cookies'),
480 'password': 'Use --username and --password, or --netrc to provide account credentials',
481 }
482
483 def __init__(self, downloader=None):
484 """Constructor. Receives an optional downloader (a YoutubeDL instance).
485 If a downloader is not passed during initialization,
486 it must be set using "set_downloader()" before "extract()" is called"""
487 self._ready = False
488 self._x_forwarded_for_ip = None
489 self._printed_messages = set()
490 self.set_downloader(downloader)
491
492 @classmethod
493 def _match_valid_url(cls, url):
494 # This does not use has/getattr intentionally - we want to know whether
495 # we have cached the regexp for *this* class, whereas getattr would also
496 # match the superclass
497 if '_VALID_URL_RE' not in cls.__dict__:
498 if '_VALID_URL' not in cls.__dict__:
499 cls._VALID_URL = cls._make_valid_url()
500 cls._VALID_URL_RE = re.compile(cls._VALID_URL)
501 return cls._VALID_URL_RE.match(url)
502
503 @classmethod
504 def suitable(cls, url):
505 """Receives a URL and returns True if suitable for this IE."""
506 # This function must import everything it needs (except other extractors),
507 # so that lazy_extractors works correctly
508 return cls._match_valid_url(url) is not None
509
510 @classmethod
511 def _match_id(cls, url):
512 return cls._match_valid_url(url).group('id')
513
514 @classmethod
515 def get_temp_id(cls, url):
516 try:
517 return cls._match_id(url)
518 except (IndexError, AttributeError):
519 return None
520
521 @classmethod
522 def working(cls):
523 """Getter method for _WORKING."""
524 return cls._WORKING
525
526 @classmethod
527 def supports_login(cls):
528 return bool(cls._NETRC_MACHINE)
529
530 def initialize(self):
531 """Initializes an instance (authentication, etc)."""
532 self._printed_messages = set()
533 self._initialize_geo_bypass({
534 'countries': self._GEO_COUNTRIES,
535 'ip_blocks': self._GEO_IP_BLOCKS,
536 })
537 if not self._ready:
538 self._initialize_pre_login()
539 if self.supports_login():
540 username, password = self._get_login_info()
541 if username:
542 self._perform_login(username, password)
543 elif self.get_param('username') and False not in (self.IE_DESC, self._NETRC_MACHINE):
544 self.report_warning(f'Login with password is not supported for this website. {self._LOGIN_HINTS["cookies"]}')
545 self._real_initialize()
546 self._ready = True
547
548 def _initialize_geo_bypass(self, geo_bypass_context):
549 """
550 Initialize geo restriction bypass mechanism.
551
552 This method is used to initialize geo bypass mechanism based on faking
553 X-Forwarded-For HTTP header. A random country from provided country list
554 is selected and a random IP belonging to this country is generated. This
555 IP will be passed as X-Forwarded-For HTTP header in all subsequent
556 HTTP requests.
557
558 This method will be used for initial geo bypass mechanism initialization
559 during the instance initialization with _GEO_COUNTRIES and
560 _GEO_IP_BLOCKS.
561
562 You may also manually call it from extractor's code if geo bypass
563 information is not available beforehand (e.g. obtained during
564 extraction) or due to some other reason. In this case you should pass
565 this information in geo bypass context passed as first argument. It may
566 contain following fields:
567
568 countries: List of geo unrestricted countries (similar
569 to _GEO_COUNTRIES)
570 ip_blocks: List of geo unrestricted IP blocks in CIDR notation
571 (similar to _GEO_IP_BLOCKS)
572
573 """
574 if not self._x_forwarded_for_ip:
575
576 # Geo bypass mechanism is explicitly disabled by user
577 if not self.get_param('geo_bypass', True):
578 return
579
580 if not geo_bypass_context:
581 geo_bypass_context = {}
582
583 # Backward compatibility: previously _initialize_geo_bypass
584 # expected a list of countries, some 3rd party code may still use
585 # it this way
586 if isinstance(geo_bypass_context, (list, tuple)):
587 geo_bypass_context = {
588 'countries': geo_bypass_context,
589 }
590
591 # The whole point of geo bypass mechanism is to fake IP
592 # as X-Forwarded-For HTTP header based on some IP block or
593 # country code.
594
595 # Path 1: bypassing based on IP block in CIDR notation
596
597 # Explicit IP block specified by user, use it right away
598 # regardless of whether extractor is geo bypassable or not
599 ip_block = self.get_param('geo_bypass_ip_block', None)
600
601 # Otherwise use random IP block from geo bypass context but only
602 # if extractor is known as geo bypassable
603 if not ip_block:
604 ip_blocks = geo_bypass_context.get('ip_blocks')
605 if self._GEO_BYPASS and ip_blocks:
606 ip_block = random.choice(ip_blocks)
607
608 if ip_block:
609 self._x_forwarded_for_ip = GeoUtils.random_ipv4(ip_block)
610 self._downloader.write_debug(
611 '[debug] Using fake IP %s as X-Forwarded-For' % self._x_forwarded_for_ip)
612 return
613
614 # Path 2: bypassing based on country code
615
616 # Explicit country code specified by user, use it right away
617 # regardless of whether extractor is geo bypassable or not
618 country = self.get_param('geo_bypass_country', None)
619
620 # Otherwise use random country code from geo bypass context but
621 # only if extractor is known as geo bypassable
622 if not country:
623 countries = geo_bypass_context.get('countries')
624 if self._GEO_BYPASS and countries:
625 country = random.choice(countries)
626
627 if country:
628 self._x_forwarded_for_ip = GeoUtils.random_ipv4(country)
629 self._downloader.write_debug(
630 'Using fake IP %s (%s) as X-Forwarded-For' % (self._x_forwarded_for_ip, country.upper()))
631
632 def extract(self, url):
633 """Extracts URL information and returns it in list of dicts."""
634 try:
635 for _ in range(2):
636 try:
637 self.initialize()
638 self.write_debug('Extracting URL: %s' % url)
639 ie_result = self._real_extract(url)
640 if ie_result is None:
641 return None
642 if self._x_forwarded_for_ip:
643 ie_result['__x_forwarded_for_ip'] = self._x_forwarded_for_ip
644 subtitles = ie_result.get('subtitles')
645 if (subtitles and 'live_chat' in subtitles
646 and 'no-live-chat' in self.get_param('compat_opts', [])):
647 del subtitles['live_chat']
648 return ie_result
649 except GeoRestrictedError as e:
650 if self.__maybe_fake_ip_and_retry(e.countries):
651 continue
652 raise
653 except UnsupportedError:
654 raise
655 except ExtractorError as e:
656 kwargs = {
657 'video_id': e.video_id or self.get_temp_id(url),
658 'ie': self.IE_NAME,
659 'tb': e.traceback or sys.exc_info()[2],
660 'expected': e.expected,
661 'cause': e.cause
662 }
663 if hasattr(e, 'countries'):
664 kwargs['countries'] = e.countries
665 raise type(e)(e.orig_msg, **kwargs)
666 except compat_http_client.IncompleteRead as e:
667 raise ExtractorError('A network error has occurred.', cause=e, expected=True, video_id=self.get_temp_id(url))
668 except (KeyError, StopIteration) as e:
669 raise ExtractorError('An extractor error has occurred.', cause=e, video_id=self.get_temp_id(url))
670
671 def __maybe_fake_ip_and_retry(self, countries):
672 if (not self.get_param('geo_bypass_country', None)
673 and self._GEO_BYPASS
674 and self.get_param('geo_bypass', True)
675 and not self._x_forwarded_for_ip
676 and countries):
677 country_code = random.choice(countries)
678 self._x_forwarded_for_ip = GeoUtils.random_ipv4(country_code)
679 if self._x_forwarded_for_ip:
680 self.report_warning(
681 'Video is geo restricted. Retrying extraction with fake IP %s (%s) as X-Forwarded-For.'
682 % (self._x_forwarded_for_ip, country_code.upper()))
683 return True
684 return False
685
686 def set_downloader(self, downloader):
687 """Sets a YoutubeDL instance as the downloader for this IE."""
688 self._downloader = downloader
689
690 def _initialize_pre_login(self):
691 """ Intialization before login. Redefine in subclasses."""
692 pass
693
694 def _perform_login(self, username, password):
695 """ Login with username and password. Redefine in subclasses."""
696 pass
697
698 def _real_initialize(self):
699 """Real initialization process. Redefine in subclasses."""
700 pass
701
702 def _real_extract(self, url):
703 """Real extraction process. Redefine in subclasses."""
704 raise NotImplementedError('This method must be implemented by subclasses')
705
706 @classmethod
707 def ie_key(cls):
708 """A string for getting the InfoExtractor with get_info_extractor"""
709 return cls.__name__[:-2]
710
711 @property
712 def IE_NAME(self):
713 return compat_str(type(self).__name__[:-2])
714
715 @staticmethod
716 def __can_accept_status_code(err, expected_status):
717 assert isinstance(err, compat_urllib_error.HTTPError)
718 if expected_status is None:
719 return False
720 elif callable(expected_status):
721 return expected_status(err.code) is True
722 else:
723 return err.code in variadic(expected_status)
724
725 def _request_webpage(self, url_or_request, video_id, note=None, errnote=None, fatal=True, data=None, headers={}, query={}, expected_status=None):
726 """
727 Return the response handle.
728
729 See _download_webpage docstring for arguments specification.
730 """
731 if not self._downloader._first_webpage_request:
732 sleep_interval = self.get_param('sleep_interval_requests') or 0
733 if sleep_interval > 0:
734 self.to_screen('Sleeping %s seconds ...' % sleep_interval)
735 time.sleep(sleep_interval)
736 else:
737 self._downloader._first_webpage_request = False
738
739 if note is None:
740 self.report_download_webpage(video_id)
741 elif note is not False:
742 if video_id is None:
743 self.to_screen('%s' % (note,))
744 else:
745 self.to_screen('%s: %s' % (video_id, note))
746
747 # Some sites check X-Forwarded-For HTTP header in order to figure out
748 # the origin of the client behind proxy. This allows bypassing geo
749 # restriction by faking this header's value to IP that belongs to some
750 # geo unrestricted country. We will do so once we encounter any
751 # geo restriction error.
752 if self._x_forwarded_for_ip:
753 if 'X-Forwarded-For' not in headers:
754 headers['X-Forwarded-For'] = self._x_forwarded_for_ip
755
756 if isinstance(url_or_request, compat_urllib_request.Request):
757 url_or_request = update_Request(
758 url_or_request, data=data, headers=headers, query=query)
759 else:
760 if query:
761 url_or_request = update_url_query(url_or_request, query)
762 if data is not None or headers:
763 url_or_request = sanitized_Request(url_or_request, data, headers)
764 try:
765 return self._downloader.urlopen(url_or_request)
766 except network_exceptions as err:
767 if isinstance(err, compat_urllib_error.HTTPError):
768 if self.__can_accept_status_code(err, expected_status):
769 # Retain reference to error to prevent file object from
770 # being closed before it can be read. Works around the
771 # effects of <https://bugs.python.org/issue15002>
772 # introduced in Python 3.4.1.
773 err.fp._error = err
774 return err.fp
775
776 if errnote is False:
777 return False
778 if errnote is None:
779 errnote = 'Unable to download webpage'
780
781 errmsg = '%s: %s' % (errnote, error_to_compat_str(err))
782 if fatal:
783 raise ExtractorError(errmsg, cause=err)
784 else:
785 self.report_warning(errmsg)
786 return False
787
788 def _download_webpage_handle(self, url_or_request, video_id, note=None, errnote=None, fatal=True, encoding=None, data=None, headers={}, query={}, expected_status=None):
789 """
790 Return a tuple (page content as string, URL handle).
791
792 See _download_webpage docstring for arguments specification.
793 """
794 # Strip hashes from the URL (#1038)
795 if isinstance(url_or_request, (compat_str, str)):
796 url_or_request = url_or_request.partition('#')[0]
797
798 urlh = self._request_webpage(url_or_request, video_id, note, errnote, fatal, data=data, headers=headers, query=query, expected_status=expected_status)
799 if urlh is False:
800 assert not fatal
801 return False
802 content = self._webpage_read_content(urlh, url_or_request, video_id, note, errnote, fatal, encoding=encoding)
803 return (content, urlh)
804
805 @staticmethod
806 def _guess_encoding_from_content(content_type, webpage_bytes):
807 m = re.match(r'[a-zA-Z0-9_.-]+/[a-zA-Z0-9_.-]+\s*;\s*charset=(.+)', content_type)
808 if m:
809 encoding = m.group(1)
810 else:
811 m = re.search(br'<meta[^>]+charset=[\'"]?([^\'")]+)[ /\'">]',
812 webpage_bytes[:1024])
813 if m:
814 encoding = m.group(1).decode('ascii')
815 elif webpage_bytes.startswith(b'\xff\xfe'):
816 encoding = 'utf-16'
817 else:
818 encoding = 'utf-8'
819
820 return encoding
821
822 def __check_blocked(self, content):
823 first_block = content[:512]
824 if ('<title>Access to this site is blocked</title>' in content
825 and 'Websense' in first_block):
826 msg = 'Access to this webpage has been blocked by Websense filtering software in your network.'
827 blocked_iframe = self._html_search_regex(
828 r'<iframe src="([^"]+)"', content,
829 'Websense information URL', default=None)
830 if blocked_iframe:
831 msg += ' Visit %s for more details' % blocked_iframe
832 raise ExtractorError(msg, expected=True)
833 if '<title>The URL you requested has been blocked</title>' in first_block:
834 msg = (
835 'Access to this webpage has been blocked by Indian censorship. '
836 'Use a VPN or proxy server (with --proxy) to route around it.')
837 block_msg = self._html_search_regex(
838 r'</h1><p>(.*?)</p>',
839 content, 'block message', default=None)
840 if block_msg:
841 msg += ' (Message: "%s")' % block_msg.replace('\n', ' ')
842 raise ExtractorError(msg, expected=True)
843 if ('<title>TTK :: Доступ к ресурсу ограничен</title>' in content
844 and 'blocklist.rkn.gov.ru' in content):
845 raise ExtractorError(
846 'Access to this webpage has been blocked by decision of the Russian government. '
847 'Visit http://blocklist.rkn.gov.ru/ for a block reason.',
848 expected=True)
849
850 def _webpage_read_content(self, urlh, url_or_request, video_id, note=None, errnote=None, fatal=True, prefix=None, encoding=None):
851 content_type = urlh.headers.get('Content-Type', '')
852 webpage_bytes = urlh.read()
853 if prefix is not None:
854 webpage_bytes = prefix + webpage_bytes
855 if not encoding:
856 encoding = self._guess_encoding_from_content(content_type, webpage_bytes)
857 if self.get_param('dump_intermediate_pages', False):
858 self.to_screen('Dumping request to ' + urlh.geturl())
859 dump = base64.b64encode(webpage_bytes).decode('ascii')
860 self._downloader.to_screen(dump)
861 if self.get_param('write_pages', False):
862 basen = '%s_%s' % (video_id, urlh.geturl())
863 trim_length = self.get_param('trim_file_name') or 240
864 if len(basen) > trim_length:
865 h = '___' + hashlib.md5(basen.encode('utf-8')).hexdigest()
866 basen = basen[:trim_length - len(h)] + h
867 raw_filename = basen + '.dump'
868 filename = sanitize_filename(raw_filename, restricted=True)
869 self.to_screen('Saving request to ' + filename)
870 # Working around MAX_PATH limitation on Windows (see
871 # http://msdn.microsoft.com/en-us/library/windows/desktop/aa365247(v=vs.85).aspx)
872 if compat_os_name == 'nt':
873 absfilepath = os.path.abspath(filename)
874 if len(absfilepath) > 259:
875 filename = '\\\\?\\' + absfilepath
876 with open(filename, 'wb') as outf:
877 outf.write(webpage_bytes)
878
879 try:
880 content = webpage_bytes.decode(encoding, 'replace')
881 except LookupError:
882 content = webpage_bytes.decode('utf-8', 'replace')
883
884 self.__check_blocked(content)
885
886 return content
887
888 def _download_webpage(
889 self, url_or_request, video_id, note=None, errnote=None,
890 fatal=True, tries=1, timeout=5, encoding=None, data=None,
891 headers={}, query={}, expected_status=None):
892 """
893 Return the data of the page as a string.
894
895 Arguments:
896 url_or_request -- plain text URL as a string or
897 a compat_urllib_request.Requestobject
898 video_id -- Video/playlist/item identifier (string)
899
900 Keyword arguments:
901 note -- note printed before downloading (string)
902 errnote -- note printed in case of an error (string)
903 fatal -- flag denoting whether error should be considered fatal,
904 i.e. whether it should cause ExtractionError to be raised,
905 otherwise a warning will be reported and extraction continued
906 tries -- number of tries
907 timeout -- sleep interval between tries
908 encoding -- encoding for a page content decoding, guessed automatically
909 when not explicitly specified
910 data -- POST data (bytes)
911 headers -- HTTP headers (dict)
912 query -- URL query (dict)
913 expected_status -- allows to accept failed HTTP requests (non 2xx
914 status code) by explicitly specifying a set of accepted status
915 codes. Can be any of the following entities:
916 - an integer type specifying an exact failed status code to
917 accept
918 - a list or a tuple of integer types specifying a list of
919 failed status codes to accept
920 - a callable accepting an actual failed status code and
921 returning True if it should be accepted
922 Note that this argument does not affect success status codes (2xx)
923 which are always accepted.
924 """
925
926 success = False
927 try_count = 0
928 while success is False:
929 try:
930 res = self._download_webpage_handle(
931 url_or_request, video_id, note, errnote, fatal,
932 encoding=encoding, data=data, headers=headers, query=query,
933 expected_status=expected_status)
934 success = True
935 except compat_http_client.IncompleteRead as e:
936 try_count += 1
937 if try_count >= tries:
938 raise e
939 self._sleep(timeout, video_id)
940 if res is False:
941 return res
942 else:
943 content, _ = res
944 return content
945
946 def _download_xml_handle(
947 self, url_or_request, video_id, note='Downloading XML',
948 errnote='Unable to download XML', transform_source=None,
949 fatal=True, encoding=None, data=None, headers={}, query={},
950 expected_status=None):
951 """
952 Return a tuple (xml as an compat_etree_Element, URL handle).
953
954 See _download_webpage docstring for arguments specification.
955 """
956 res = self._download_webpage_handle(
957 url_or_request, video_id, note, errnote, fatal=fatal,
958 encoding=encoding, data=data, headers=headers, query=query,
959 expected_status=expected_status)
960 if res is False:
961 return res
962 xml_string, urlh = res
963 return self._parse_xml(
964 xml_string, video_id, transform_source=transform_source,
965 fatal=fatal), urlh
966
967 def _download_xml(
968 self, url_or_request, video_id,
969 note='Downloading XML', errnote='Unable to download XML',
970 transform_source=None, fatal=True, encoding=None,
971 data=None, headers={}, query={}, expected_status=None):
972 """
973 Return the xml as an compat_etree_Element.
974
975 See _download_webpage docstring for arguments specification.
976 """
977 res = self._download_xml_handle(
978 url_or_request, video_id, note=note, errnote=errnote,
979 transform_source=transform_source, fatal=fatal, encoding=encoding,
980 data=data, headers=headers, query=query,
981 expected_status=expected_status)
982 return res if res is False else res[0]
983
984 def _parse_xml(self, xml_string, video_id, transform_source=None, fatal=True):
985 if transform_source:
986 xml_string = transform_source(xml_string)
987 try:
988 return compat_etree_fromstring(xml_string.encode('utf-8'))
989 except compat_xml_parse_error as ve:
990 errmsg = '%s: Failed to parse XML ' % video_id
991 if fatal:
992 raise ExtractorError(errmsg, cause=ve)
993 else:
994 self.report_warning(errmsg + str(ve))
995
996 def _download_json_handle(
997 self, url_or_request, video_id, note='Downloading JSON metadata',
998 errnote='Unable to download JSON metadata', transform_source=None,
999 fatal=True, encoding=None, data=None, headers={}, query={},
1000 expected_status=None):
1001 """
1002 Return a tuple (JSON object, URL handle).
1003
1004 See _download_webpage docstring for arguments specification.
1005 """
1006 res = self._download_webpage_handle(
1007 url_or_request, video_id, note, errnote, fatal=fatal,
1008 encoding=encoding, data=data, headers=headers, query=query,
1009 expected_status=expected_status)
1010 if res is False:
1011 return res
1012 json_string, urlh = res
1013 return self._parse_json(
1014 json_string, video_id, transform_source=transform_source,
1015 fatal=fatal), urlh
1016
1017 def _download_json(
1018 self, url_or_request, video_id, note='Downloading JSON metadata',
1019 errnote='Unable to download JSON metadata', transform_source=None,
1020 fatal=True, encoding=None, data=None, headers={}, query={},
1021 expected_status=None):
1022 """
1023 Return the JSON object as a dict.
1024
1025 See _download_webpage docstring for arguments specification.
1026 """
1027 res = self._download_json_handle(
1028 url_or_request, video_id, note=note, errnote=errnote,
1029 transform_source=transform_source, fatal=fatal, encoding=encoding,
1030 data=data, headers=headers, query=query,
1031 expected_status=expected_status)
1032 return res if res is False else res[0]
1033
1034 def _parse_json(self, json_string, video_id, transform_source=None, fatal=True):
1035 if transform_source:
1036 json_string = transform_source(json_string)
1037 try:
1038 return json.loads(json_string, strict=False)
1039 except ValueError as ve:
1040 errmsg = '%s: Failed to parse JSON ' % video_id
1041 if fatal:
1042 raise ExtractorError(errmsg, cause=ve)
1043 else:
1044 self.report_warning(errmsg + str(ve))
1045
1046 def _parse_socket_response_as_json(self, data, video_id, transform_source=None, fatal=True):
1047 return self._parse_json(
1048 data[data.find('{'):data.rfind('}') + 1],
1049 video_id, transform_source, fatal)
1050
1051 def _download_socket_json_handle(
1052 self, url_or_request, video_id, note='Polling socket',
1053 errnote='Unable to poll socket', transform_source=None,
1054 fatal=True, encoding=None, data=None, headers={}, query={},
1055 expected_status=None):
1056 """
1057 Return a tuple (JSON object, URL handle).
1058
1059 See _download_webpage docstring for arguments specification.
1060 """
1061 res = self._download_webpage_handle(
1062 url_or_request, video_id, note, errnote, fatal=fatal,
1063 encoding=encoding, data=data, headers=headers, query=query,
1064 expected_status=expected_status)
1065 if res is False:
1066 return res
1067 webpage, urlh = res
1068 return self._parse_socket_response_as_json(
1069 webpage, video_id, transform_source=transform_source,
1070 fatal=fatal), urlh
1071
1072 def _download_socket_json(
1073 self, url_or_request, video_id, note='Polling socket',
1074 errnote='Unable to poll socket', transform_source=None,
1075 fatal=True, encoding=None, data=None, headers={}, query={},
1076 expected_status=None):
1077 """
1078 Return the JSON object as a dict.
1079
1080 See _download_webpage docstring for arguments specification.
1081 """
1082 res = self._download_socket_json_handle(
1083 url_or_request, video_id, note=note, errnote=errnote,
1084 transform_source=transform_source, fatal=fatal, encoding=encoding,
1085 data=data, headers=headers, query=query,
1086 expected_status=expected_status)
1087 return res if res is False else res[0]
1088
1089 def report_warning(self, msg, video_id=None, *args, only_once=False, **kwargs):
1090 idstr = format_field(video_id, template='%s: ')
1091 msg = f'[{self.IE_NAME}] {idstr}{msg}'
1092 if only_once:
1093 if f'WARNING: {msg}' in self._printed_messages:
1094 return
1095 self._printed_messages.add(f'WARNING: {msg}')
1096 self._downloader.report_warning(msg, *args, **kwargs)
1097
1098 def to_screen(self, msg, *args, **kwargs):
1099 """Print msg to screen, prefixing it with '[ie_name]'"""
1100 self._downloader.to_screen('[%s] %s' % (self.IE_NAME, msg), *args, **kwargs)
1101
1102 def write_debug(self, msg, *args, **kwargs):
1103 self._downloader.write_debug('[%s] %s' % (self.IE_NAME, msg), *args, **kwargs)
1104
1105 def get_param(self, name, default=None, *args, **kwargs):
1106 if self._downloader:
1107 return self._downloader.params.get(name, default, *args, **kwargs)
1108 return default
1109
1110 def report_drm(self, video_id, partial=False):
1111 self.raise_no_formats('This video is DRM protected', expected=True, video_id=video_id)
1112
1113 def report_extraction(self, id_or_name):
1114 """Report information extraction."""
1115 self.to_screen('%s: Extracting information' % id_or_name)
1116
1117 def report_download_webpage(self, video_id):
1118 """Report webpage download."""
1119 self.to_screen('%s: Downloading webpage' % video_id)
1120
1121 def report_age_confirmation(self):
1122 """Report attempt to confirm age."""
1123 self.to_screen('Confirming age')
1124
1125 def report_login(self):
1126 """Report attempt to log in."""
1127 self.to_screen('Logging in')
1128
1129 def raise_login_required(
1130 self, msg='This video is only available for registered users',
1131 metadata_available=False, method=NO_DEFAULT):
1132 if metadata_available and (
1133 self.get_param('ignore_no_formats_error') or self.get_param('wait_for_video')):
1134 self.report_warning(msg)
1135 return
1136 if method is NO_DEFAULT:
1137 method = 'any' if self.supports_login() else 'cookies'
1138 if method is not None:
1139 assert method in self._LOGIN_HINTS, 'Invalid login method'
1140 msg = '%s. %s' % (msg, self._LOGIN_HINTS[method])
1141 raise ExtractorError(msg, expected=True)
1142
1143 def raise_geo_restricted(
1144 self, msg='This video is not available from your location due to geo restriction',
1145 countries=None, metadata_available=False):
1146 if metadata_available and (
1147 self.get_param('ignore_no_formats_error') or self.get_param('wait_for_video')):
1148 self.report_warning(msg)
1149 else:
1150 raise GeoRestrictedError(msg, countries=countries)
1151
1152 def raise_no_formats(self, msg, expected=False, video_id=None):
1153 if expected and (
1154 self.get_param('ignore_no_formats_error') or self.get_param('wait_for_video')):
1155 self.report_warning(msg, video_id)
1156 elif isinstance(msg, ExtractorError):
1157 raise msg
1158 else:
1159 raise ExtractorError(msg, expected=expected, video_id=video_id)
1160
1161 # Methods for following #608
1162 @staticmethod
1163 def url_result(url, ie=None, video_id=None, video_title=None, *, url_transparent=False, **kwargs):
1164 """Returns a URL that points to a page that should be processed"""
1165 if ie is not None:
1166 kwargs['ie_key'] = ie if isinstance(ie, str) else ie.ie_key()
1167 if video_id is not None:
1168 kwargs['id'] = video_id
1169 if video_title is not None:
1170 kwargs['title'] = video_title
1171 return {
1172 **kwargs,
1173 '_type': 'url_transparent' if url_transparent else 'url',
1174 'url': url,
1175 }
1176
1177 def playlist_from_matches(self, matches, playlist_id=None, playlist_title=None, getter=None, ie=None, video_kwargs=None, **kwargs):
1178 urls = (self.url_result(self._proto_relative_url(m), ie, **(video_kwargs or {}))
1179 for m in orderedSet(map(getter, matches) if getter else matches))
1180 return self.playlist_result(urls, playlist_id, playlist_title, **kwargs)
1181
1182 @staticmethod
1183 def playlist_result(entries, playlist_id=None, playlist_title=None, playlist_description=None, *, multi_video=False, **kwargs):
1184 """Returns a playlist"""
1185 if playlist_id:
1186 kwargs['id'] = playlist_id
1187 if playlist_title:
1188 kwargs['title'] = playlist_title
1189 if playlist_description is not None:
1190 kwargs['description'] = playlist_description
1191 return {
1192 **kwargs,
1193 '_type': 'multi_video' if multi_video else 'playlist',
1194 'entries': entries,
1195 }
1196
1197 def _search_regex(self, pattern, string, name, default=NO_DEFAULT, fatal=True, flags=0, group=None):
1198 """
1199 Perform a regex search on the given string, using a single or a list of
1200 patterns returning the first matching group.
1201 In case of failure return a default value or raise a WARNING or a
1202 RegexNotFoundError, depending on fatal, specifying the field name.
1203 """
1204 if isinstance(pattern, (str, compat_str, compiled_regex_type)):
1205 mobj = re.search(pattern, string, flags)
1206 else:
1207 for p in pattern:
1208 mobj = re.search(p, string, flags)
1209 if mobj:
1210 break
1211
1212 _name = self._downloader._format_err(name, self._downloader.Styles.EMPHASIS)
1213
1214 if mobj:
1215 if group is None:
1216 # return the first matching group
1217 return next(g for g in mobj.groups() if g is not None)
1218 elif isinstance(group, (list, tuple)):
1219 return tuple(mobj.group(g) for g in group)
1220 else:
1221 return mobj.group(group)
1222 elif default is not NO_DEFAULT:
1223 return default
1224 elif fatal:
1225 raise RegexNotFoundError('Unable to extract %s' % _name)
1226 else:
1227 self.report_warning('unable to extract %s' % _name + bug_reports_message())
1228 return None
1229
1230 def _html_search_regex(self, pattern, string, name, default=NO_DEFAULT, fatal=True, flags=0, group=None):
1231 """
1232 Like _search_regex, but strips HTML tags and unescapes entities.
1233 """
1234 res = self._search_regex(pattern, string, name, default, fatal, flags, group)
1235 if res:
1236 return clean_html(res).strip()
1237 else:
1238 return res
1239
1240 def _get_netrc_login_info(self, netrc_machine=None):
1241 username = None
1242 password = None
1243 netrc_machine = netrc_machine or self._NETRC_MACHINE
1244
1245 if self.get_param('usenetrc', False):
1246 try:
1247 netrc_file = compat_expanduser(self.get_param('netrc_location') or '~')
1248 if os.path.isdir(netrc_file):
1249 netrc_file = os.path.join(netrc_file, '.netrc')
1250 info = netrc.netrc(file=netrc_file).authenticators(netrc_machine)
1251 if info is not None:
1252 username = info[0]
1253 password = info[2]
1254 else:
1255 raise netrc.NetrcParseError(
1256 'No authenticators for %s' % netrc_machine)
1257 except (IOError, netrc.NetrcParseError) as err:
1258 self.report_warning(
1259 'parsing .netrc: %s' % error_to_compat_str(err))
1260
1261 return username, password
1262
1263 def _get_login_info(self, username_option='username', password_option='password', netrc_machine=None):
1264 """
1265 Get the login info as (username, password)
1266 First look for the manually specified credentials using username_option
1267 and password_option as keys in params dictionary. If no such credentials
1268 available look in the netrc file using the netrc_machine or _NETRC_MACHINE
1269 value.
1270 If there's no info available, return (None, None)
1271 """
1272
1273 # Attempt to use provided username and password or .netrc data
1274 username = self.get_param(username_option)
1275 if username is not None:
1276 password = self.get_param(password_option)
1277 else:
1278 username, password = self._get_netrc_login_info(netrc_machine)
1279
1280 return username, password
1281
1282 def _get_tfa_info(self, note='two-factor verification code'):
1283 """
1284 Get the two-factor authentication info
1285 TODO - asking the user will be required for sms/phone verify
1286 currently just uses the command line option
1287 If there's no info available, return None
1288 """
1289
1290 tfa = self.get_param('twofactor')
1291 if tfa is not None:
1292 return tfa
1293
1294 return compat_getpass('Type %s and press [Return]: ' % note)
1295
1296 # Helper functions for extracting OpenGraph info
1297 @staticmethod
1298 def _og_regexes(prop):
1299 content_re = r'content=(?:"([^"]+?)"|\'([^\']+?)\'|\s*([^\s"\'=<>`]+?))'
1300 property_re = (r'(?:name|property)=(?:\'og%(sep)s%(prop)s\'|"og%(sep)s%(prop)s"|\s*og%(sep)s%(prop)s\b)'
1301 % {'prop': re.escape(prop), 'sep': '(?:&#x3A;|[:-])'})
1302 template = r'<meta[^>]+?%s[^>]+?%s'
1303 return [
1304 template % (property_re, content_re),
1305 template % (content_re, property_re),
1306 ]
1307
1308 @staticmethod
1309 def _meta_regex(prop):
1310 return r'''(?isx)<meta
1311 (?=[^>]+(?:itemprop|name|property|id|http-equiv)=(["\']?)%s\1)
1312 [^>]+?content=(["\'])(?P<content>.*?)\2''' % re.escape(prop)
1313
1314 def _og_search_property(self, prop, html, name=None, **kargs):
1315 prop = variadic(prop)
1316 if name is None:
1317 name = 'OpenGraph %s' % prop[0]
1318 og_regexes = []
1319 for p in prop:
1320 og_regexes.extend(self._og_regexes(p))
1321 escaped = self._search_regex(og_regexes, html, name, flags=re.DOTALL, **kargs)
1322 if escaped is None:
1323 return None
1324 return unescapeHTML(escaped)
1325
1326 def _og_search_thumbnail(self, html, **kargs):
1327 return self._og_search_property('image', html, 'thumbnail URL', fatal=False, **kargs)
1328
1329 def _og_search_description(self, html, **kargs):
1330 return self._og_search_property('description', html, fatal=False, **kargs)
1331
1332 def _og_search_title(self, html, *, fatal=False, **kargs):
1333 return self._og_search_property('title', html, fatal=fatal, **kargs)
1334
1335 def _og_search_video_url(self, html, name='video url', secure=True, **kargs):
1336 regexes = self._og_regexes('video') + self._og_regexes('video:url')
1337 if secure:
1338 regexes = self._og_regexes('video:secure_url') + regexes
1339 return self._html_search_regex(regexes, html, name, **kargs)
1340
1341 def _og_search_url(self, html, **kargs):
1342 return self._og_search_property('url', html, **kargs)
1343
1344 def _html_extract_title(self, html, name='title', *, fatal=False, **kwargs):
1345 return self._html_search_regex(r'(?s)<title>([^<]+)</title>', html, name, fatal=fatal, **kwargs)
1346
1347 def _html_search_meta(self, name, html, display_name=None, fatal=False, **kwargs):
1348 name = variadic(name)
1349 if display_name is None:
1350 display_name = name[0]
1351 return self._html_search_regex(
1352 [self._meta_regex(n) for n in name],
1353 html, display_name, fatal=fatal, group='content', **kwargs)
1354
1355 def _dc_search_uploader(self, html):
1356 return self._html_search_meta('dc.creator', html, 'uploader')
1357
1358 def _rta_search(self, html):
1359 # See http://www.rtalabel.org/index.php?content=howtofaq#single
1360 if re.search(r'(?ix)<meta\s+name="rating"\s+'
1361 r' content="RTA-5042-1996-1400-1577-RTA"',
1362 html):
1363 return 18
1364 return 0
1365
1366 def _media_rating_search(self, html):
1367 # See http://www.tjg-designs.com/WP/metadata-code-examples-adding-metadata-to-your-web-pages/
1368 rating = self._html_search_meta('rating', html)
1369
1370 if not rating:
1371 return None
1372
1373 RATING_TABLE = {
1374 'safe for kids': 0,
1375 'general': 8,
1376 '14 years': 14,
1377 'mature': 17,
1378 'restricted': 19,
1379 }
1380 return RATING_TABLE.get(rating.lower())
1381
1382 def _family_friendly_search(self, html):
1383 # See http://schema.org/VideoObject
1384 family_friendly = self._html_search_meta(
1385 'isFamilyFriendly', html, default=None)
1386
1387 if not family_friendly:
1388 return None
1389
1390 RATING_TABLE = {
1391 '1': 0,
1392 'true': 0,
1393 '0': 18,
1394 'false': 18,
1395 }
1396 return RATING_TABLE.get(family_friendly.lower())
1397
1398 def _twitter_search_player(self, html):
1399 return self._html_search_meta('twitter:player', html,
1400 'twitter card player')
1401
1402 def _search_json_ld(self, html, video_id, expected_type=None, **kwargs):
1403 json_ld_list = list(re.finditer(JSON_LD_RE, html))
1404 default = kwargs.get('default', NO_DEFAULT)
1405 # JSON-LD may be malformed and thus `fatal` should be respected.
1406 # At the same time `default` may be passed that assumes `fatal=False`
1407 # for _search_regex. Let's simulate the same behavior here as well.
1408 fatal = kwargs.get('fatal', True) if default is NO_DEFAULT else False
1409 json_ld = []
1410 for mobj in json_ld_list:
1411 json_ld_item = self._parse_json(
1412 mobj.group('json_ld'), video_id, fatal=fatal)
1413 if not json_ld_item:
1414 continue
1415 if isinstance(json_ld_item, dict):
1416 json_ld.append(json_ld_item)
1417 elif isinstance(json_ld_item, (list, tuple)):
1418 json_ld.extend(json_ld_item)
1419 if json_ld:
1420 json_ld = self._json_ld(json_ld, video_id, fatal=fatal, expected_type=expected_type)
1421 if json_ld:
1422 return json_ld
1423 if default is not NO_DEFAULT:
1424 return default
1425 elif fatal:
1426 raise RegexNotFoundError('Unable to extract JSON-LD')
1427 else:
1428 self.report_warning('unable to extract JSON-LD %s' % bug_reports_message())
1429 return {}
1430
1431 def _json_ld(self, json_ld, video_id, fatal=True, expected_type=None):
1432 if isinstance(json_ld, compat_str):
1433 json_ld = self._parse_json(json_ld, video_id, fatal=fatal)
1434 if not json_ld:
1435 return {}
1436 info = {}
1437 if not isinstance(json_ld, (list, tuple, dict)):
1438 return info
1439 if isinstance(json_ld, dict):
1440 json_ld = [json_ld]
1441
1442 INTERACTION_TYPE_MAP = {
1443 'CommentAction': 'comment',
1444 'AgreeAction': 'like',
1445 'DisagreeAction': 'dislike',
1446 'LikeAction': 'like',
1447 'DislikeAction': 'dislike',
1448 'ListenAction': 'view',
1449 'WatchAction': 'view',
1450 'ViewAction': 'view',
1451 }
1452
1453 def extract_interaction_type(e):
1454 interaction_type = e.get('interactionType')
1455 if isinstance(interaction_type, dict):
1456 interaction_type = interaction_type.get('@type')
1457 return str_or_none(interaction_type)
1458
1459 def extract_interaction_statistic(e):
1460 interaction_statistic = e.get('interactionStatistic')
1461 if isinstance(interaction_statistic, dict):
1462 interaction_statistic = [interaction_statistic]
1463 if not isinstance(interaction_statistic, list):
1464 return
1465 for is_e in interaction_statistic:
1466 if not isinstance(is_e, dict):
1467 continue
1468 if is_e.get('@type') != 'InteractionCounter':
1469 continue
1470 interaction_type = extract_interaction_type(is_e)
1471 if not interaction_type:
1472 continue
1473 # For interaction count some sites provide string instead of
1474 # an integer (as per spec) with non digit characters (e.g. ",")
1475 # so extracting count with more relaxed str_to_int
1476 interaction_count = str_to_int(is_e.get('userInteractionCount'))
1477 if interaction_count is None:
1478 continue
1479 count_kind = INTERACTION_TYPE_MAP.get(interaction_type.split('/')[-1])
1480 if not count_kind:
1481 continue
1482 count_key = '%s_count' % count_kind
1483 if info.get(count_key) is not None:
1484 continue
1485 info[count_key] = interaction_count
1486
1487 def extract_chapter_information(e):
1488 chapters = [{
1489 'title': part.get('name'),
1490 'start_time': part.get('startOffset'),
1491 'end_time': part.get('endOffset'),
1492 } for part in variadic(e.get('hasPart') or []) if part.get('@type') == 'Clip']
1493 for idx, (last_c, current_c, next_c) in enumerate(zip(
1494 [{'end_time': 0}] + chapters, chapters, chapters[1:])):
1495 current_c['end_time'] = current_c['end_time'] or next_c['start_time']
1496 current_c['start_time'] = current_c['start_time'] or last_c['end_time']
1497 if None in current_c.values():
1498 self.report_warning(f'Chapter {idx} contains broken data. Not extracting chapters')
1499 return
1500 if chapters:
1501 chapters[-1]['end_time'] = chapters[-1]['end_time'] or info['duration']
1502 info['chapters'] = chapters
1503
1504 def extract_video_object(e):
1505 assert e['@type'] == 'VideoObject'
1506 author = e.get('author')
1507 info.update({
1508 'url': url_or_none(e.get('contentUrl')),
1509 'title': unescapeHTML(e.get('name')),
1510 'description': unescapeHTML(e.get('description')),
1511 'thumbnails': [{'url': url_or_none(url)}
1512 for url in variadic(traverse_obj(e, 'thumbnailUrl', 'thumbnailURL'))],
1513 'duration': parse_duration(e.get('duration')),
1514 'timestamp': unified_timestamp(e.get('uploadDate')),
1515 # author can be an instance of 'Organization' or 'Person' types.
1516 # both types can have 'name' property(inherited from 'Thing' type). [1]
1517 # however some websites are using 'Text' type instead.
1518 # 1. https://schema.org/VideoObject
1519 'uploader': author.get('name') if isinstance(author, dict) else author if isinstance(author, compat_str) else None,
1520 'filesize': float_or_none(e.get('contentSize')),
1521 'tbr': int_or_none(e.get('bitrate')),
1522 'width': int_or_none(e.get('width')),
1523 'height': int_or_none(e.get('height')),
1524 'view_count': int_or_none(e.get('interactionCount')),
1525 })
1526 extract_interaction_statistic(e)
1527 extract_chapter_information(e)
1528
1529 def traverse_json_ld(json_ld, at_top_level=True):
1530 for e in json_ld:
1531 if at_top_level and '@context' not in e:
1532 continue
1533 if at_top_level and set(e.keys()) == {'@context', '@graph'}:
1534 traverse_json_ld(variadic(e['@graph'], allowed_types=(dict,)), at_top_level=False)
1535 break
1536 item_type = e.get('@type')
1537 if expected_type is not None and expected_type != item_type:
1538 continue
1539 rating = traverse_obj(e, ('aggregateRating', 'ratingValue'), expected_type=float_or_none)
1540 if rating is not None:
1541 info['average_rating'] = rating
1542 if item_type in ('TVEpisode', 'Episode'):
1543 episode_name = unescapeHTML(e.get('name'))
1544 info.update({
1545 'episode': episode_name,
1546 'episode_number': int_or_none(e.get('episodeNumber')),
1547 'description': unescapeHTML(e.get('description')),
1548 })
1549 if not info.get('title') and episode_name:
1550 info['title'] = episode_name
1551 part_of_season = e.get('partOfSeason')
1552 if isinstance(part_of_season, dict) and part_of_season.get('@type') in ('TVSeason', 'Season', 'CreativeWorkSeason'):
1553 info.update({
1554 'season': unescapeHTML(part_of_season.get('name')),
1555 'season_number': int_or_none(part_of_season.get('seasonNumber')),
1556 })
1557 part_of_series = e.get('partOfSeries') or e.get('partOfTVSeries')
1558 if isinstance(part_of_series, dict) and part_of_series.get('@type') in ('TVSeries', 'Series', 'CreativeWorkSeries'):
1559 info['series'] = unescapeHTML(part_of_series.get('name'))
1560 elif item_type == 'Movie':
1561 info.update({
1562 'title': unescapeHTML(e.get('name')),
1563 'description': unescapeHTML(e.get('description')),
1564 'duration': parse_duration(e.get('duration')),
1565 'timestamp': unified_timestamp(e.get('dateCreated')),
1566 })
1567 elif item_type in ('Article', 'NewsArticle'):
1568 info.update({
1569 'timestamp': parse_iso8601(e.get('datePublished')),
1570 'title': unescapeHTML(e.get('headline')),
1571 'description': unescapeHTML(e.get('articleBody') or e.get('description')),
1572 })
1573 if traverse_obj(e, ('video', 0, '@type')) == 'VideoObject':
1574 extract_video_object(e['video'][0])
1575 elif item_type == 'VideoObject':
1576 extract_video_object(e)
1577 if expected_type is None:
1578 continue
1579 else:
1580 break
1581 video = e.get('video')
1582 if isinstance(video, dict) and video.get('@type') == 'VideoObject':
1583 extract_video_object(video)
1584 if expected_type is None:
1585 continue
1586 else:
1587 break
1588 traverse_json_ld(json_ld)
1589
1590 return filter_dict(info)
1591
1592 def _search_nextjs_data(self, webpage, video_id, *, transform_source=None, fatal=True, **kw):
1593 return self._parse_json(
1594 self._search_regex(
1595 r'(?s)<script[^>]+id=[\'"]__NEXT_DATA__[\'"][^>]*>([^<]+)</script>',
1596 webpage, 'next.js data', fatal=fatal, **kw),
1597 video_id, transform_source=transform_source, fatal=fatal)
1598
1599 def _search_nuxt_data(self, webpage, video_id, context_name='__NUXT__'):
1600 ''' Parses Nuxt.js metadata. This works as long as the function __NUXT__ invokes is a pure function. '''
1601 # not all website do this, but it can be changed
1602 # https://stackoverflow.com/questions/67463109/how-to-change-or-hide-nuxt-and-nuxt-keyword-in-page-source
1603 rectx = re.escape(context_name)
1604 js, arg_keys, arg_vals = self._search_regex(
1605 (r'<script>window\.%s=\(function\((?P<arg_keys>.*?)\)\{return\s(?P<js>\{.*?\})\}\((?P<arg_vals>.+?)\)\);?</script>' % rectx,
1606 r'%s\(.*?\(function\((?P<arg_keys>.*?)\)\{return\s(?P<js>\{.*?\})\}\((?P<arg_vals>.*?)\)' % rectx),
1607 webpage, context_name, group=['js', 'arg_keys', 'arg_vals'])
1608
1609 args = dict(zip(arg_keys.split(','), arg_vals.split(',')))
1610
1611 for key, val in args.items():
1612 if val in ('undefined', 'void 0'):
1613 args[key] = 'null'
1614
1615 return self._parse_json(js_to_json(js, args), video_id)['data'][0]
1616
1617 @staticmethod
1618 def _hidden_inputs(html):
1619 html = re.sub(r'<!--(?:(?!<!--).)*-->', '', html)
1620 hidden_inputs = {}
1621 for input in re.findall(r'(?i)(<input[^>]+>)', html):
1622 attrs = extract_attributes(input)
1623 if not input:
1624 continue
1625 if attrs.get('type') not in ('hidden', 'submit'):
1626 continue
1627 name = attrs.get('name') or attrs.get('id')
1628 value = attrs.get('value')
1629 if name and value is not None:
1630 hidden_inputs[name] = value
1631 return hidden_inputs
1632
1633 def _form_hidden_inputs(self, form_id, html):
1634 form = self._search_regex(
1635 r'(?is)<form[^>]+?id=(["\'])%s\1[^>]*>(?P<form>.+?)</form>' % form_id,
1636 html, '%s form' % form_id, group='form')
1637 return self._hidden_inputs(form)
1638
1639 class FormatSort:
1640 regex = r' *((?P<reverse>\+)?(?P<field>[a-zA-Z0-9_]+)((?P<separator>[~:])(?P<limit>.*?))?)? *$'
1641
1642 default = ('hidden', 'aud_or_vid', 'hasvid', 'ie_pref', 'lang', 'quality',
1643 'res', 'fps', 'hdr:12', 'codec:vp9.2', 'size', 'br', 'asr',
1644 'proto', 'ext', 'hasaud', 'source', 'id') # These must not be aliases
1645 ytdl_default = ('hasaud', 'lang', 'quality', 'tbr', 'filesize', 'vbr',
1646 'height', 'width', 'proto', 'vext', 'abr', 'aext',
1647 'fps', 'fs_approx', 'source', 'id')
1648
1649 settings = {
1650 'vcodec': {'type': 'ordered', 'regex': True,
1651 'order': ['av0?1', 'vp0?9.2', 'vp0?9', '[hx]265|he?vc?', '[hx]264|avc', 'vp0?8', 'mp4v|h263', 'theora', '', None, 'none']},
1652 'acodec': {'type': 'ordered', 'regex': True,
1653 'order': ['[af]lac', 'wav|aiff', 'opus', 'vorbis|ogg', 'aac', 'mp?4a?', 'mp3', 'e-?a?c-?3', 'ac-?3', 'dts', '', None, 'none']},
1654 'hdr': {'type': 'ordered', 'regex': True, 'field': 'dynamic_range',
1655 'order': ['dv', '(hdr)?12', r'(hdr)?10\+', '(hdr)?10', 'hlg', '', 'sdr', None]},
1656 'proto': {'type': 'ordered', 'regex': True, 'field': 'protocol',
1657 'order': ['(ht|f)tps', '(ht|f)tp$', 'm3u8.*', '.*dash', 'websocket_frag', 'rtmpe?', '', 'mms|rtsp', 'ws|websocket', 'f4']},
1658 'vext': {'type': 'ordered', 'field': 'video_ext',
1659 'order': ('mp4', 'webm', 'flv', '', 'none'),
1660 'order_free': ('webm', 'mp4', 'flv', '', 'none')},
1661 'aext': {'type': 'ordered', 'field': 'audio_ext',
1662 'order': ('m4a', 'aac', 'mp3', 'ogg', 'opus', 'webm', '', 'none'),
1663 'order_free': ('opus', 'ogg', 'webm', 'm4a', 'mp3', 'aac', '', 'none')},
1664 'hidden': {'visible': False, 'forced': True, 'type': 'extractor', 'max': -1000},
1665 'aud_or_vid': {'visible': False, 'forced': True, 'type': 'multiple',
1666 'field': ('vcodec', 'acodec'),
1667 'function': lambda it: int(any(v != 'none' for v in it))},
1668 'ie_pref': {'priority': True, 'type': 'extractor'},
1669 'hasvid': {'priority': True, 'field': 'vcodec', 'type': 'boolean', 'not_in_list': ('none',)},
1670 'hasaud': {'field': 'acodec', 'type': 'boolean', 'not_in_list': ('none',)},
1671 'lang': {'convert': 'float', 'field': 'language_preference', 'default': -1},
1672 'quality': {'convert': 'float', 'default': -1},
1673 'filesize': {'convert': 'bytes'},
1674 'fs_approx': {'convert': 'bytes', 'field': 'filesize_approx'},
1675 'id': {'convert': 'string', 'field': 'format_id'},
1676 'height': {'convert': 'float_none'},
1677 'width': {'convert': 'float_none'},
1678 'fps': {'convert': 'float_none'},
1679 'tbr': {'convert': 'float_none'},
1680 'vbr': {'convert': 'float_none'},
1681 'abr': {'convert': 'float_none'},
1682 'asr': {'convert': 'float_none'},
1683 'source': {'convert': 'float', 'field': 'source_preference', 'default': -1},
1684
1685 'codec': {'type': 'combined', 'field': ('vcodec', 'acodec')},
1686 'br': {'type': 'combined', 'field': ('tbr', 'vbr', 'abr'), 'same_limit': True},
1687 'size': {'type': 'combined', 'same_limit': True, 'field': ('filesize', 'fs_approx')},
1688 'ext': {'type': 'combined', 'field': ('vext', 'aext')},
1689 'res': {'type': 'multiple', 'field': ('height', 'width'),
1690 'function': lambda it: (lambda l: min(l) if l else 0)(tuple(filter(None, it)))},
1691
1692 # For compatibility with youtube-dl
1693 'format_id': {'type': 'alias', 'field': 'id'},
1694 'preference': {'type': 'alias', 'field': 'ie_pref'},
1695 'language_preference': {'type': 'alias', 'field': 'lang'},
1696 'source_preference': {'type': 'alias', 'field': 'source'},
1697 'protocol': {'type': 'alias', 'field': 'proto'},
1698 'filesize_approx': {'type': 'alias', 'field': 'fs_approx'},
1699
1700 # Deprecated
1701 'dimension': {'type': 'alias', 'field': 'res', 'deprecated': True},
1702 'resolution': {'type': 'alias', 'field': 'res', 'deprecated': True},
1703 'extension': {'type': 'alias', 'field': 'ext', 'deprecated': True},
1704 'bitrate': {'type': 'alias', 'field': 'br', 'deprecated': True},
1705 'total_bitrate': {'type': 'alias', 'field': 'tbr', 'deprecated': True},
1706 'video_bitrate': {'type': 'alias', 'field': 'vbr', 'deprecated': True},
1707 'audio_bitrate': {'type': 'alias', 'field': 'abr', 'deprecated': True},
1708 'framerate': {'type': 'alias', 'field': 'fps', 'deprecated': True},
1709 'filesize_estimate': {'type': 'alias', 'field': 'size', 'deprecated': True},
1710 'samplerate': {'type': 'alias', 'field': 'asr', 'deprecated': True},
1711 'video_ext': {'type': 'alias', 'field': 'vext', 'deprecated': True},
1712 'audio_ext': {'type': 'alias', 'field': 'aext', 'deprecated': True},
1713 'video_codec': {'type': 'alias', 'field': 'vcodec', 'deprecated': True},
1714 'audio_codec': {'type': 'alias', 'field': 'acodec', 'deprecated': True},
1715 'video': {'type': 'alias', 'field': 'hasvid', 'deprecated': True},
1716 'has_video': {'type': 'alias', 'field': 'hasvid', 'deprecated': True},
1717 'audio': {'type': 'alias', 'field': 'hasaud', 'deprecated': True},
1718 'has_audio': {'type': 'alias', 'field': 'hasaud', 'deprecated': True},
1719 'extractor': {'type': 'alias', 'field': 'ie_pref', 'deprecated': True},
1720 'extractor_preference': {'type': 'alias', 'field': 'ie_pref', 'deprecated': True},
1721 }
1722
1723 def __init__(self, ie, field_preference):
1724 self._order = []
1725 self.ydl = ie._downloader
1726 self.evaluate_params(self.ydl.params, field_preference)
1727 if ie.get_param('verbose'):
1728 self.print_verbose_info(self.ydl.write_debug)
1729
1730 def _get_field_setting(self, field, key):
1731 if field not in self.settings:
1732 if key in ('forced', 'priority'):
1733 return False
1734 self.ydl.deprecation_warning(
1735 f'Using arbitrary fields ({field}) for format sorting is deprecated '
1736 'and may be removed in a future version')
1737 self.settings[field] = {}
1738 propObj = self.settings[field]
1739 if key not in propObj:
1740 type = propObj.get('type')
1741 if key == 'field':
1742 default = 'preference' if type == 'extractor' else (field,) if type in ('combined', 'multiple') else field
1743 elif key == 'convert':
1744 default = 'order' if type == 'ordered' else 'float_string' if field else 'ignore'
1745 else:
1746 default = {'type': 'field', 'visible': True, 'order': [], 'not_in_list': (None,)}.get(key, None)
1747 propObj[key] = default
1748 return propObj[key]
1749
1750 def _resolve_field_value(self, field, value, convertNone=False):
1751 if value is None:
1752 if not convertNone:
1753 return None
1754 else:
1755 value = value.lower()
1756 conversion = self._get_field_setting(field, 'convert')
1757 if conversion == 'ignore':
1758 return None
1759 if conversion == 'string':
1760 return value
1761 elif conversion == 'float_none':
1762 return float_or_none(value)
1763 elif conversion == 'bytes':
1764 return FileDownloader.parse_bytes(value)
1765 elif conversion == 'order':
1766 order_list = (self._use_free_order and self._get_field_setting(field, 'order_free')) or self._get_field_setting(field, 'order')
1767 use_regex = self._get_field_setting(field, 'regex')
1768 list_length = len(order_list)
1769 empty_pos = order_list.index('') if '' in order_list else list_length + 1
1770 if use_regex and value is not None:
1771 for i, regex in enumerate(order_list):
1772 if regex and re.match(regex, value):
1773 return list_length - i
1774 return list_length - empty_pos # not in list
1775 else: # not regex or value = None
1776 return list_length - (order_list.index(value) if value in order_list else empty_pos)
1777 else:
1778 if value.isnumeric():
1779 return float(value)
1780 else:
1781 self.settings[field]['convert'] = 'string'
1782 return value
1783
1784 def evaluate_params(self, params, sort_extractor):
1785 self._use_free_order = params.get('prefer_free_formats', False)
1786 self._sort_user = params.get('format_sort', [])
1787 self._sort_extractor = sort_extractor
1788
1789 def add_item(field, reverse, closest, limit_text):
1790 field = field.lower()
1791 if field in self._order:
1792 return
1793 self._order.append(field)
1794 limit = self._resolve_field_value(field, limit_text)
1795 data = {
1796 'reverse': reverse,
1797 'closest': False if limit is None else closest,
1798 'limit_text': limit_text,
1799 'limit': limit}
1800 if field in self.settings:
1801 self.settings[field].update(data)
1802 else:
1803 self.settings[field] = data
1804
1805 sort_list = (
1806 tuple(field for field in self.default if self._get_field_setting(field, 'forced'))
1807 + (tuple() if params.get('format_sort_force', False)
1808 else tuple(field for field in self.default if self._get_field_setting(field, 'priority')))
1809 + tuple(self._sort_user) + tuple(sort_extractor) + self.default)
1810
1811 for item in sort_list:
1812 match = re.match(self.regex, item)
1813 if match is None:
1814 raise ExtractorError('Invalid format sort string "%s" given by extractor' % item)
1815 field = match.group('field')
1816 if field is None:
1817 continue
1818 if self._get_field_setting(field, 'type') == 'alias':
1819 alias, field = field, self._get_field_setting(field, 'field')
1820 if self._get_field_setting(alias, 'deprecated'):
1821 self.ydl.deprecation_warning(
1822 f'Format sorting alias {alias} is deprecated '
1823 f'and may be removed in a future version. Please use {field} instead')
1824 reverse = match.group('reverse') is not None
1825 closest = match.group('separator') == '~'
1826 limit_text = match.group('limit')
1827
1828 has_limit = limit_text is not None
1829 has_multiple_fields = self._get_field_setting(field, 'type') == 'combined'
1830 has_multiple_limits = has_limit and has_multiple_fields and not self._get_field_setting(field, 'same_limit')
1831
1832 fields = self._get_field_setting(field, 'field') if has_multiple_fields else (field,)
1833 limits = limit_text.split(':') if has_multiple_limits else (limit_text,) if has_limit else tuple()
1834 limit_count = len(limits)
1835 for (i, f) in enumerate(fields):
1836 add_item(f, reverse, closest,
1837 limits[i] if i < limit_count
1838 else limits[0] if has_limit and not has_multiple_limits
1839 else None)
1840
1841 def print_verbose_info(self, write_debug):
1842 if self._sort_user:
1843 write_debug('Sort order given by user: %s' % ', '.join(self._sort_user))
1844 if self._sort_extractor:
1845 write_debug('Sort order given by extractor: %s' % ', '.join(self._sort_extractor))
1846 write_debug('Formats sorted by: %s' % ', '.join(['%s%s%s' % (
1847 '+' if self._get_field_setting(field, 'reverse') else '', field,
1848 '%s%s(%s)' % ('~' if self._get_field_setting(field, 'closest') else ':',
1849 self._get_field_setting(field, 'limit_text'),
1850 self._get_field_setting(field, 'limit'))
1851 if self._get_field_setting(field, 'limit_text') is not None else '')
1852 for field in self._order if self._get_field_setting(field, 'visible')]))
1853
1854 def _calculate_field_preference_from_value(self, format, field, type, value):
1855 reverse = self._get_field_setting(field, 'reverse')
1856 closest = self._get_field_setting(field, 'closest')
1857 limit = self._get_field_setting(field, 'limit')
1858
1859 if type == 'extractor':
1860 maximum = self._get_field_setting(field, 'max')
1861 if value is None or (maximum is not None and value >= maximum):
1862 value = -1
1863 elif type == 'boolean':
1864 in_list = self._get_field_setting(field, 'in_list')
1865 not_in_list = self._get_field_setting(field, 'not_in_list')
1866 value = 0 if ((in_list is None or value in in_list) and (not_in_list is None or value not in not_in_list)) else -1
1867 elif type == 'ordered':
1868 value = self._resolve_field_value(field, value, True)
1869
1870 # try to convert to number
1871 val_num = float_or_none(value, default=self._get_field_setting(field, 'default'))
1872 is_num = self._get_field_setting(field, 'convert') != 'string' and val_num is not None
1873 if is_num:
1874 value = val_num
1875
1876 return ((-10, 0) if value is None
1877 else (1, value, 0) if not is_num # if a field has mixed strings and numbers, strings are sorted higher
1878 else (0, -abs(value - limit), value - limit if reverse else limit - value) if closest
1879 else (0, value, 0) if not reverse and (limit is None or value <= limit)
1880 else (0, -value, 0) if limit is None or (reverse and value == limit) or value > limit
1881 else (-1, value, 0))
1882
1883 def _calculate_field_preference(self, format, field):
1884 type = self._get_field_setting(field, 'type') # extractor, boolean, ordered, field, multiple
1885 get_value = lambda f: format.get(self._get_field_setting(f, 'field'))
1886 if type == 'multiple':
1887 type = 'field' # Only 'field' is allowed in multiple for now
1888 actual_fields = self._get_field_setting(field, 'field')
1889
1890 value = self._get_field_setting(field, 'function')(get_value(f) for f in actual_fields)
1891 else:
1892 value = get_value(field)
1893 return self._calculate_field_preference_from_value(format, field, type, value)
1894
1895 def calculate_preference(self, format):
1896 # Determine missing protocol
1897 if not format.get('protocol'):
1898 format['protocol'] = determine_protocol(format)
1899
1900 # Determine missing ext
1901 if not format.get('ext') and 'url' in format:
1902 format['ext'] = determine_ext(format['url'])
1903 if format.get('vcodec') == 'none':
1904 format['audio_ext'] = format['ext'] if format.get('acodec') != 'none' else 'none'
1905 format['video_ext'] = 'none'
1906 else:
1907 format['video_ext'] = format['ext']
1908 format['audio_ext'] = 'none'
1909 # if format.get('preference') is None and format.get('ext') in ('f4f', 'f4m'): # Not supported?
1910 # format['preference'] = -1000
1911
1912 # Determine missing bitrates
1913 if format.get('tbr') is None:
1914 if format.get('vbr') is not None and format.get('abr') is not None:
1915 format['tbr'] = format.get('vbr', 0) + format.get('abr', 0)
1916 else:
1917 if format.get('vcodec') != 'none' and format.get('vbr') is None:
1918 format['vbr'] = format.get('tbr') - format.get('abr', 0)
1919 if format.get('acodec') != 'none' and format.get('abr') is None:
1920 format['abr'] = format.get('tbr') - format.get('vbr', 0)
1921
1922 return tuple(self._calculate_field_preference(format, field) for field in self._order)
1923
1924 def _sort_formats(self, formats, field_preference=[]):
1925 if not formats:
1926 return
1927 format_sort = self.FormatSort(self, field_preference)
1928 formats.sort(key=lambda f: format_sort.calculate_preference(f))
1929
1930 def _check_formats(self, formats, video_id):
1931 if formats:
1932 formats[:] = filter(
1933 lambda f: self._is_valid_url(
1934 f['url'], video_id,
1935 item='%s video format' % f.get('format_id') if f.get('format_id') else 'video'),
1936 formats)
1937
1938 @staticmethod
1939 def _remove_duplicate_formats(formats):
1940 format_urls = set()
1941 unique_formats = []
1942 for f in formats:
1943 if f['url'] not in format_urls:
1944 format_urls.add(f['url'])
1945 unique_formats.append(f)
1946 formats[:] = unique_formats
1947
1948 def _is_valid_url(self, url, video_id, item='video', headers={}):
1949 url = self._proto_relative_url(url, scheme='http:')
1950 # For now assume non HTTP(S) URLs always valid
1951 if not (url.startswith('http://') or url.startswith('https://')):
1952 return True
1953 try:
1954 self._request_webpage(url, video_id, 'Checking %s URL' % item, headers=headers)
1955 return True
1956 except ExtractorError as e:
1957 self.to_screen(
1958 '%s: %s URL is invalid, skipping: %s'
1959 % (video_id, item, error_to_compat_str(e.cause)))
1960 return False
1961
1962 def http_scheme(self):
1963 """ Either "http:" or "https:", depending on the user's preferences """
1964 return (
1965 'http:'
1966 if self.get_param('prefer_insecure', False)
1967 else 'https:')
1968
1969 def _proto_relative_url(self, url, scheme=None):
1970 if url is None:
1971 return url
1972 if url.startswith('//'):
1973 if scheme is None:
1974 scheme = self.http_scheme()
1975 return scheme + url
1976 else:
1977 return url
1978
1979 def _sleep(self, timeout, video_id, msg_template=None):
1980 if msg_template is None:
1981 msg_template = '%(video_id)s: Waiting for %(timeout)s seconds'
1982 msg = msg_template % {'video_id': video_id, 'timeout': timeout}
1983 self.to_screen(msg)
1984 time.sleep(timeout)
1985
1986 def _extract_f4m_formats(self, manifest_url, video_id, preference=None, quality=None, f4m_id=None,
1987 transform_source=lambda s: fix_xml_ampersands(s).strip(),
1988 fatal=True, m3u8_id=None, data=None, headers={}, query={}):
1989 manifest = self._download_xml(
1990 manifest_url, video_id, 'Downloading f4m manifest',
1991 'Unable to download f4m manifest',
1992 # Some manifests may be malformed, e.g. prosiebensat1 generated manifests
1993 # (see https://github.com/ytdl-org/youtube-dl/issues/6215#issuecomment-121704244)
1994 transform_source=transform_source,
1995 fatal=fatal, data=data, headers=headers, query=query)
1996
1997 if manifest is False:
1998 return []
1999
2000 return self._parse_f4m_formats(
2001 manifest, manifest_url, video_id, preference=preference, quality=quality, f4m_id=f4m_id,
2002 transform_source=transform_source, fatal=fatal, m3u8_id=m3u8_id)
2003
2004 def _parse_f4m_formats(self, manifest, manifest_url, video_id, preference=None, quality=None, f4m_id=None,
2005 transform_source=lambda s: fix_xml_ampersands(s).strip(),
2006 fatal=True, m3u8_id=None):
2007 if not isinstance(manifest, compat_etree_Element) and not fatal:
2008 return []
2009
2010 # currently yt-dlp cannot decode the playerVerificationChallenge as Akamai uses Adobe Alchemy
2011 akamai_pv = manifest.find('{http://ns.adobe.com/f4m/1.0}pv-2.0')
2012 if akamai_pv is not None and ';' in akamai_pv.text:
2013 playerVerificationChallenge = akamai_pv.text.split(';')[0]
2014 if playerVerificationChallenge.strip() != '':
2015 return []
2016
2017 formats = []
2018 manifest_version = '1.0'
2019 media_nodes = manifest.findall('{http://ns.adobe.com/f4m/1.0}media')
2020 if not media_nodes:
2021 manifest_version = '2.0'
2022 media_nodes = manifest.findall('{http://ns.adobe.com/f4m/2.0}media')
2023 # Remove unsupported DRM protected media from final formats
2024 # rendition (see https://github.com/ytdl-org/youtube-dl/issues/8573).
2025 media_nodes = remove_encrypted_media(media_nodes)
2026 if not media_nodes:
2027 return formats
2028
2029 manifest_base_url = get_base_url(manifest)
2030
2031 bootstrap_info = xpath_element(
2032 manifest, ['{http://ns.adobe.com/f4m/1.0}bootstrapInfo', '{http://ns.adobe.com/f4m/2.0}bootstrapInfo'],
2033 'bootstrap info', default=None)
2034
2035 vcodec = None
2036 mime_type = xpath_text(
2037 manifest, ['{http://ns.adobe.com/f4m/1.0}mimeType', '{http://ns.adobe.com/f4m/2.0}mimeType'],
2038 'base URL', default=None)
2039 if mime_type and mime_type.startswith('audio/'):
2040 vcodec = 'none'
2041
2042 for i, media_el in enumerate(media_nodes):
2043 tbr = int_or_none(media_el.attrib.get('bitrate'))
2044 width = int_or_none(media_el.attrib.get('width'))
2045 height = int_or_none(media_el.attrib.get('height'))
2046 format_id = join_nonempty(f4m_id, tbr or i)
2047 # If <bootstrapInfo> is present, the specified f4m is a
2048 # stream-level manifest, and only set-level manifests may refer to
2049 # external resources. See section 11.4 and section 4 of F4M spec
2050 if bootstrap_info is None:
2051 media_url = None
2052 # @href is introduced in 2.0, see section 11.6 of F4M spec
2053 if manifest_version == '2.0':
2054 media_url = media_el.attrib.get('href')
2055 if media_url is None:
2056 media_url = media_el.attrib.get('url')
2057 if not media_url:
2058 continue
2059 manifest_url = (
2060 media_url if media_url.startswith('http://') or media_url.startswith('https://')
2061 else ((manifest_base_url or '/'.join(manifest_url.split('/')[:-1])) + '/' + media_url))
2062 # If media_url is itself a f4m manifest do the recursive extraction
2063 # since bitrates in parent manifest (this one) and media_url manifest
2064 # may differ leading to inability to resolve the format by requested
2065 # bitrate in f4m downloader
2066 ext = determine_ext(manifest_url)
2067 if ext == 'f4m':
2068 f4m_formats = self._extract_f4m_formats(
2069 manifest_url, video_id, preference=preference, quality=quality, f4m_id=f4m_id,
2070 transform_source=transform_source, fatal=fatal)
2071 # Sometimes stream-level manifest contains single media entry that
2072 # does not contain any quality metadata (e.g. http://matchtv.ru/#live-player).
2073 # At the same time parent's media entry in set-level manifest may
2074 # contain it. We will copy it from parent in such cases.
2075 if len(f4m_formats) == 1:
2076 f = f4m_formats[0]
2077 f.update({
2078 'tbr': f.get('tbr') or tbr,
2079 'width': f.get('width') or width,
2080 'height': f.get('height') or height,
2081 'format_id': f.get('format_id') if not tbr else format_id,
2082 'vcodec': vcodec,
2083 })
2084 formats.extend(f4m_formats)
2085 continue
2086 elif ext == 'm3u8':
2087 formats.extend(self._extract_m3u8_formats(
2088 manifest_url, video_id, 'mp4', preference=preference,
2089 quality=quality, m3u8_id=m3u8_id, fatal=fatal))
2090 continue
2091 formats.append({
2092 'format_id': format_id,
2093 'url': manifest_url,
2094 'manifest_url': manifest_url,
2095 'ext': 'flv' if bootstrap_info is not None else None,
2096 'protocol': 'f4m',
2097 'tbr': tbr,
2098 'width': width,
2099 'height': height,
2100 'vcodec': vcodec,
2101 'preference': preference,
2102 'quality': quality,
2103 })
2104 return formats
2105
2106 def _m3u8_meta_format(self, m3u8_url, ext=None, preference=None, quality=None, m3u8_id=None):
2107 return {
2108 'format_id': join_nonempty(m3u8_id, 'meta'),
2109 'url': m3u8_url,
2110 'ext': ext,
2111 'protocol': 'm3u8',
2112 'preference': preference - 100 if preference else -100,
2113 'quality': quality,
2114 'resolution': 'multiple',
2115 'format_note': 'Quality selection URL',
2116 }
2117
2118 def _report_ignoring_subs(self, name):
2119 self.report_warning(bug_reports_message(
2120 f'Ignoring subtitle tracks found in the {name} manifest; '
2121 'if any subtitle tracks are missing,'
2122 ), only_once=True)
2123
2124 def _extract_m3u8_formats(self, *args, **kwargs):
2125 fmts, subs = self._extract_m3u8_formats_and_subtitles(*args, **kwargs)
2126 if subs:
2127 self._report_ignoring_subs('HLS')
2128 return fmts
2129
2130 def _extract_m3u8_formats_and_subtitles(
2131 self, m3u8_url, video_id, ext=None, entry_protocol='m3u8_native',
2132 preference=None, quality=None, m3u8_id=None, note=None,
2133 errnote=None, fatal=True, live=False, data=None, headers={},
2134 query={}):
2135
2136 res = self._download_webpage_handle(
2137 m3u8_url, video_id,
2138 note='Downloading m3u8 information' if note is None else note,
2139 errnote='Failed to download m3u8 information' if errnote is None else errnote,
2140 fatal=fatal, data=data, headers=headers, query=query)
2141
2142 if res is False:
2143 return [], {}
2144
2145 m3u8_doc, urlh = res
2146 m3u8_url = urlh.geturl()
2147
2148 return self._parse_m3u8_formats_and_subtitles(
2149 m3u8_doc, m3u8_url, ext=ext, entry_protocol=entry_protocol,
2150 preference=preference, quality=quality, m3u8_id=m3u8_id,
2151 note=note, errnote=errnote, fatal=fatal, live=live, data=data,
2152 headers=headers, query=query, video_id=video_id)
2153
2154 def _parse_m3u8_formats_and_subtitles(
2155 self, m3u8_doc, m3u8_url=None, ext=None, entry_protocol='m3u8_native',
2156 preference=None, quality=None, m3u8_id=None, live=False, note=None,
2157 errnote=None, fatal=True, data=None, headers={}, query={},
2158 video_id=None):
2159 formats, subtitles = [], {}
2160
2161 has_drm = re.search('|'.join([
2162 r'#EXT-X-FAXS-CM:', # Adobe Flash Access
2163 r'#EXT-X-(?:SESSION-)?KEY:.*?URI="skd://', # Apple FairPlay
2164 ]), m3u8_doc)
2165
2166 def format_url(url):
2167 return url if re.match(r'^https?://', url) else compat_urlparse.urljoin(m3u8_url, url)
2168
2169 if self.get_param('hls_split_discontinuity', False):
2170 def _extract_m3u8_playlist_indices(manifest_url=None, m3u8_doc=None):
2171 if not m3u8_doc:
2172 if not manifest_url:
2173 return []
2174 m3u8_doc = self._download_webpage(
2175 manifest_url, video_id, fatal=fatal, data=data, headers=headers,
2176 note=False, errnote='Failed to download m3u8 playlist information')
2177 if m3u8_doc is False:
2178 return []
2179 return range(1 + sum(line.startswith('#EXT-X-DISCONTINUITY') for line in m3u8_doc.splitlines()))
2180
2181 else:
2182 def _extract_m3u8_playlist_indices(*args, **kwargs):
2183 return [None]
2184
2185 # References:
2186 # 1. https://tools.ietf.org/html/draft-pantos-http-live-streaming-21
2187 # 2. https://github.com/ytdl-org/youtube-dl/issues/12211
2188 # 3. https://github.com/ytdl-org/youtube-dl/issues/18923
2189
2190 # We should try extracting formats only from master playlists [1, 4.3.4],
2191 # i.e. playlists that describe available qualities. On the other hand
2192 # media playlists [1, 4.3.3] should be returned as is since they contain
2193 # just the media without qualities renditions.
2194 # Fortunately, master playlist can be easily distinguished from media
2195 # playlist based on particular tags availability. As of [1, 4.3.3, 4.3.4]
2196 # master playlist tags MUST NOT appear in a media playlist and vice versa.
2197 # As of [1, 4.3.3.1] #EXT-X-TARGETDURATION tag is REQUIRED for every
2198 # media playlist and MUST NOT appear in master playlist thus we can
2199 # clearly detect media playlist with this criterion.
2200
2201 if '#EXT-X-TARGETDURATION' in m3u8_doc: # media playlist, return as is
2202 formats = [{
2203 'format_id': join_nonempty(m3u8_id, idx),
2204 'format_index': idx,
2205 'url': m3u8_url or encode_data_uri(m3u8_doc.encode('utf-8'), 'application/x-mpegurl'),
2206 'ext': ext,
2207 'protocol': entry_protocol,
2208 'preference': preference,
2209 'quality': quality,
2210 'has_drm': has_drm,
2211 } for idx in _extract_m3u8_playlist_indices(m3u8_doc=m3u8_doc)]
2212
2213 return formats, subtitles
2214
2215 groups = {}
2216 last_stream_inf = {}
2217
2218 def extract_media(x_media_line):
2219 media = parse_m3u8_attributes(x_media_line)
2220 # As per [1, 4.3.4.1] TYPE, GROUP-ID and NAME are REQUIRED
2221 media_type, group_id, name = media.get('TYPE'), media.get('GROUP-ID'), media.get('NAME')
2222 if not (media_type and group_id and name):
2223 return
2224 groups.setdefault(group_id, []).append(media)
2225 # <https://tools.ietf.org/html/rfc8216#section-4.3.4.1>
2226 if media_type == 'SUBTITLES':
2227 # According to RFC 8216 §4.3.4.2.1, URI is REQUIRED in the
2228 # EXT-X-MEDIA tag if the media type is SUBTITLES.
2229 # However, lack of URI has been spotted in the wild.
2230 # e.g. NebulaIE; see https://github.com/yt-dlp/yt-dlp/issues/339
2231 if not media.get('URI'):
2232 return
2233 url = format_url(media['URI'])
2234 sub_info = {
2235 'url': url,
2236 'ext': determine_ext(url),
2237 }
2238 if sub_info['ext'] == 'm3u8':
2239 # Per RFC 8216 §3.1, the only possible subtitle format m3u8
2240 # files may contain is WebVTT:
2241 # <https://tools.ietf.org/html/rfc8216#section-3.1>
2242 sub_info['ext'] = 'vtt'
2243 sub_info['protocol'] = 'm3u8_native'
2244 lang = media.get('LANGUAGE') or 'und'
2245 subtitles.setdefault(lang, []).append(sub_info)
2246 if media_type not in ('VIDEO', 'AUDIO'):
2247 return
2248 media_url = media.get('URI')
2249 if media_url:
2250 manifest_url = format_url(media_url)
2251 formats.extend({
2252 'format_id': join_nonempty(m3u8_id, group_id, name, idx),
2253 'format_note': name,
2254 'format_index': idx,
2255 'url': manifest_url,
2256 'manifest_url': m3u8_url,
2257 'language': media.get('LANGUAGE'),
2258 'ext': ext,
2259 'protocol': entry_protocol,
2260 'preference': preference,
2261 'quality': quality,
2262 'vcodec': 'none' if media_type == 'AUDIO' else None,
2263 } for idx in _extract_m3u8_playlist_indices(manifest_url))
2264
2265 def build_stream_name():
2266 # Despite specification does not mention NAME attribute for
2267 # EXT-X-STREAM-INF tag it still sometimes may be present (see [1]
2268 # or vidio test in TestInfoExtractor.test_parse_m3u8_formats)
2269 # 1. http://www.vidio.com/watch/165683-dj_ambred-booyah-live-2015
2270 stream_name = last_stream_inf.get('NAME')
2271 if stream_name:
2272 return stream_name
2273 # If there is no NAME in EXT-X-STREAM-INF it will be obtained
2274 # from corresponding rendition group
2275 stream_group_id = last_stream_inf.get('VIDEO')
2276 if not stream_group_id:
2277 return
2278 stream_group = groups.get(stream_group_id)
2279 if not stream_group:
2280 return stream_group_id
2281 rendition = stream_group[0]
2282 return rendition.get('NAME') or stream_group_id
2283
2284 # parse EXT-X-MEDIA tags before EXT-X-STREAM-INF in order to have the
2285 # chance to detect video only formats when EXT-X-STREAM-INF tags
2286 # precede EXT-X-MEDIA tags in HLS manifest such as [3].
2287 for line in m3u8_doc.splitlines():
2288 if line.startswith('#EXT-X-MEDIA:'):
2289 extract_media(line)
2290
2291 for line in m3u8_doc.splitlines():
2292 if line.startswith('#EXT-X-STREAM-INF:'):
2293 last_stream_inf = parse_m3u8_attributes(line)
2294 elif line.startswith('#') or not line.strip():
2295 continue
2296 else:
2297 tbr = float_or_none(
2298 last_stream_inf.get('AVERAGE-BANDWIDTH')
2299 or last_stream_inf.get('BANDWIDTH'), scale=1000)
2300 manifest_url = format_url(line.strip())
2301
2302 for idx in _extract_m3u8_playlist_indices(manifest_url):
2303 format_id = [m3u8_id, None, idx]
2304 # Bandwidth of live streams may differ over time thus making
2305 # format_id unpredictable. So it's better to keep provided
2306 # format_id intact.
2307 if not live:
2308 stream_name = build_stream_name()
2309 format_id[1] = stream_name or '%d' % (tbr or len(formats))
2310 f = {
2311 'format_id': join_nonempty(*format_id),
2312 'format_index': idx,
2313 'url': manifest_url,
2314 'manifest_url': m3u8_url,
2315 'tbr': tbr,
2316 'ext': ext,
2317 'fps': float_or_none(last_stream_inf.get('FRAME-RATE')),
2318 'protocol': entry_protocol,
2319 'preference': preference,
2320 'quality': quality,
2321 }
2322 resolution = last_stream_inf.get('RESOLUTION')
2323 if resolution:
2324 mobj = re.search(r'(?P<width>\d+)[xX](?P<height>\d+)', resolution)
2325 if mobj:
2326 f['width'] = int(mobj.group('width'))
2327 f['height'] = int(mobj.group('height'))
2328 # Unified Streaming Platform
2329 mobj = re.search(
2330 r'audio.*?(?:%3D|=)(\d+)(?:-video.*?(?:%3D|=)(\d+))?', f['url'])
2331 if mobj:
2332 abr, vbr = mobj.groups()
2333 abr, vbr = float_or_none(abr, 1000), float_or_none(vbr, 1000)
2334 f.update({
2335 'vbr': vbr,
2336 'abr': abr,
2337 })
2338 codecs = parse_codecs(last_stream_inf.get('CODECS'))
2339 f.update(codecs)
2340 audio_group_id = last_stream_inf.get('AUDIO')
2341 # As per [1, 4.3.4.1.1] any EXT-X-STREAM-INF tag which
2342 # references a rendition group MUST have a CODECS attribute.
2343 # However, this is not always respected, for example, [2]
2344 # contains EXT-X-STREAM-INF tag which references AUDIO
2345 # rendition group but does not have CODECS and despite
2346 # referencing an audio group it represents a complete
2347 # (with audio and video) format. So, for such cases we will
2348 # ignore references to rendition groups and treat them
2349 # as complete formats.
2350 if audio_group_id and codecs and f.get('vcodec') != 'none':
2351 audio_group = groups.get(audio_group_id)
2352 if audio_group and audio_group[0].get('URI'):
2353 # TODO: update acodec for audio only formats with
2354 # the same GROUP-ID
2355 f['acodec'] = 'none'
2356 if not f.get('ext'):
2357 f['ext'] = 'm4a' if f.get('vcodec') == 'none' else 'mp4'
2358 formats.append(f)
2359
2360 # for DailyMotion
2361 progressive_uri = last_stream_inf.get('PROGRESSIVE-URI')
2362 if progressive_uri:
2363 http_f = f.copy()
2364 del http_f['manifest_url']
2365 http_f.update({
2366 'format_id': f['format_id'].replace('hls-', 'http-'),
2367 'protocol': 'http',
2368 'url': progressive_uri,
2369 })
2370 formats.append(http_f)
2371
2372 last_stream_inf = {}
2373 return formats, subtitles
2374
2375 def _extract_m3u8_vod_duration(
2376 self, m3u8_vod_url, video_id, note=None, errnote=None, data=None, headers={}, query={}):
2377
2378 m3u8_vod = self._download_webpage(
2379 m3u8_vod_url, video_id,
2380 note='Downloading m3u8 VOD manifest' if note is None else note,
2381 errnote='Failed to download VOD manifest' if errnote is None else errnote,
2382 fatal=False, data=data, headers=headers, query=query)
2383
2384 return self._parse_m3u8_vod_duration(m3u8_vod or '', video_id)
2385
2386 def _parse_m3u8_vod_duration(self, m3u8_vod, video_id):
2387 if '#EXT-X-PLAYLIST-TYPE:VOD' not in m3u8_vod:
2388 return None
2389
2390 return int(sum(
2391 float(line[len('#EXTINF:'):].split(',')[0])
2392 for line in m3u8_vod.splitlines() if line.startswith('#EXTINF:'))) or None
2393
2394 @staticmethod
2395 def _xpath_ns(path, namespace=None):
2396 if not namespace:
2397 return path
2398 out = []
2399 for c in path.split('/'):
2400 if not c or c == '.':
2401 out.append(c)
2402 else:
2403 out.append('{%s}%s' % (namespace, c))
2404 return '/'.join(out)
2405
2406 def _extract_smil_formats_and_subtitles(self, smil_url, video_id, fatal=True, f4m_params=None, transform_source=None):
2407 smil = self._download_smil(smil_url, video_id, fatal=fatal, transform_source=transform_source)
2408
2409 if smil is False:
2410 assert not fatal
2411 return [], {}
2412
2413 namespace = self._parse_smil_namespace(smil)
2414
2415 fmts = self._parse_smil_formats(
2416 smil, smil_url, video_id, namespace=namespace, f4m_params=f4m_params)
2417 subs = self._parse_smil_subtitles(
2418 smil, namespace=namespace)
2419
2420 return fmts, subs
2421
2422 def _extract_smil_formats(self, *args, **kwargs):
2423 fmts, subs = self._extract_smil_formats_and_subtitles(*args, **kwargs)
2424 if subs:
2425 self._report_ignoring_subs('SMIL')
2426 return fmts
2427
2428 def _extract_smil_info(self, smil_url, video_id, fatal=True, f4m_params=None):
2429 smil = self._download_smil(smil_url, video_id, fatal=fatal)
2430 if smil is False:
2431 return {}
2432 return self._parse_smil(smil, smil_url, video_id, f4m_params=f4m_params)
2433
2434 def _download_smil(self, smil_url, video_id, fatal=True, transform_source=None):
2435 return self._download_xml(
2436 smil_url, video_id, 'Downloading SMIL file',
2437 'Unable to download SMIL file', fatal=fatal, transform_source=transform_source)
2438
2439 def _parse_smil(self, smil, smil_url, video_id, f4m_params=None):
2440 namespace = self._parse_smil_namespace(smil)
2441
2442 formats = self._parse_smil_formats(
2443 smil, smil_url, video_id, namespace=namespace, f4m_params=f4m_params)
2444 subtitles = self._parse_smil_subtitles(smil, namespace=namespace)
2445
2446 video_id = os.path.splitext(url_basename(smil_url))[0]
2447 title = None
2448 description = None
2449 upload_date = None
2450 for meta in smil.findall(self._xpath_ns('./head/meta', namespace)):
2451 name = meta.attrib.get('name')
2452 content = meta.attrib.get('content')
2453 if not name or not content:
2454 continue
2455 if not title and name == 'title':
2456 title = content
2457 elif not description and name in ('description', 'abstract'):
2458 description = content
2459 elif not upload_date and name == 'date':
2460 upload_date = unified_strdate(content)
2461
2462 thumbnails = [{
2463 'id': image.get('type'),
2464 'url': image.get('src'),
2465 'width': int_or_none(image.get('width')),
2466 'height': int_or_none(image.get('height')),
2467 } for image in smil.findall(self._xpath_ns('.//image', namespace)) if image.get('src')]
2468
2469 return {
2470 'id': video_id,
2471 'title': title or video_id,
2472 'description': description,
2473 'upload_date': upload_date,
2474 'thumbnails': thumbnails,
2475 'formats': formats,
2476 'subtitles': subtitles,
2477 }
2478
2479 def _parse_smil_namespace(self, smil):
2480 return self._search_regex(
2481 r'(?i)^{([^}]+)?}smil$', smil.tag, 'namespace', default=None)
2482
2483 def _parse_smil_formats(self, smil, smil_url, video_id, namespace=None, f4m_params=None, transform_rtmp_url=None):
2484 base = smil_url
2485 for meta in smil.findall(self._xpath_ns('./head/meta', namespace)):
2486 b = meta.get('base') or meta.get('httpBase')
2487 if b:
2488 base = b
2489 break
2490
2491 formats = []
2492 rtmp_count = 0
2493 http_count = 0
2494 m3u8_count = 0
2495 imgs_count = 0
2496
2497 srcs = set()
2498 media = smil.findall(self._xpath_ns('.//video', namespace)) + smil.findall(self._xpath_ns('.//audio', namespace))
2499 for medium in media:
2500 src = medium.get('src')
2501 if not src or src in srcs:
2502 continue
2503 srcs.add(src)
2504
2505 bitrate = float_or_none(medium.get('system-bitrate') or medium.get('systemBitrate'), 1000)
2506 filesize = int_or_none(medium.get('size') or medium.get('fileSize'))
2507 width = int_or_none(medium.get('width'))
2508 height = int_or_none(medium.get('height'))
2509 proto = medium.get('proto')
2510 ext = medium.get('ext')
2511 src_ext = determine_ext(src)
2512 streamer = medium.get('streamer') or base
2513
2514 if proto == 'rtmp' or streamer.startswith('rtmp'):
2515 rtmp_count += 1
2516 formats.append({
2517 'url': streamer,
2518 'play_path': src,
2519 'ext': 'flv',
2520 'format_id': 'rtmp-%d' % (rtmp_count if bitrate is None else bitrate),
2521 'tbr': bitrate,
2522 'filesize': filesize,
2523 'width': width,
2524 'height': height,
2525 })
2526 if transform_rtmp_url:
2527 streamer, src = transform_rtmp_url(streamer, src)
2528 formats[-1].update({
2529 'url': streamer,
2530 'play_path': src,
2531 })
2532 continue
2533
2534 src_url = src if src.startswith('http') else compat_urlparse.urljoin(base, src)
2535 src_url = src_url.strip()
2536
2537 if proto == 'm3u8' or src_ext == 'm3u8':
2538 m3u8_formats = self._extract_m3u8_formats(
2539 src_url, video_id, ext or 'mp4', m3u8_id='hls', fatal=False)
2540 if len(m3u8_formats) == 1:
2541 m3u8_count += 1
2542 m3u8_formats[0].update({
2543 'format_id': 'hls-%d' % (m3u8_count if bitrate is None else bitrate),
2544 'tbr': bitrate,
2545 'width': width,
2546 'height': height,
2547 })
2548 formats.extend(m3u8_formats)
2549 elif src_ext == 'f4m':
2550 f4m_url = src_url
2551 if not f4m_params:
2552 f4m_params = {
2553 'hdcore': '3.2.0',
2554 'plugin': 'flowplayer-3.2.0.1',
2555 }
2556 f4m_url += '&' if '?' in f4m_url else '?'
2557 f4m_url += compat_urllib_parse_urlencode(f4m_params)
2558 formats.extend(self._extract_f4m_formats(f4m_url, video_id, f4m_id='hds', fatal=False))
2559 elif src_ext == 'mpd':
2560 formats.extend(self._extract_mpd_formats(
2561 src_url, video_id, mpd_id='dash', fatal=False))
2562 elif re.search(r'\.ism/[Mm]anifest', src_url):
2563 formats.extend(self._extract_ism_formats(
2564 src_url, video_id, ism_id='mss', fatal=False))
2565 elif src_url.startswith('http') and self._is_valid_url(src, video_id):
2566 http_count += 1
2567 formats.append({
2568 'url': src_url,
2569 'ext': ext or src_ext or 'flv',
2570 'format_id': 'http-%d' % (bitrate or http_count),
2571 'tbr': bitrate,
2572 'filesize': filesize,
2573 'width': width,
2574 'height': height,
2575 })
2576
2577 for medium in smil.findall(self._xpath_ns('.//imagestream', namespace)):
2578 src = medium.get('src')
2579 if not src or src in srcs:
2580 continue
2581 srcs.add(src)
2582
2583 imgs_count += 1
2584 formats.append({
2585 'format_id': 'imagestream-%d' % (imgs_count),
2586 'url': src,
2587 'ext': mimetype2ext(medium.get('type')),
2588 'acodec': 'none',
2589 'vcodec': 'none',
2590 'width': int_or_none(medium.get('width')),
2591 'height': int_or_none(medium.get('height')),
2592 'format_note': 'SMIL storyboards',
2593 })
2594
2595 return formats
2596
2597 def _parse_smil_subtitles(self, smil, namespace=None, subtitles_lang='en'):
2598 urls = []
2599 subtitles = {}
2600 for num, textstream in enumerate(smil.findall(self._xpath_ns('.//textstream', namespace))):
2601 src = textstream.get('src')
2602 if not src or src in urls:
2603 continue
2604 urls.append(src)
2605 ext = textstream.get('ext') or mimetype2ext(textstream.get('type')) or determine_ext(src)
2606 lang = textstream.get('systemLanguage') or textstream.get('systemLanguageName') or textstream.get('lang') or subtitles_lang
2607 subtitles.setdefault(lang, []).append({
2608 'url': src,
2609 'ext': ext,
2610 })
2611 return subtitles
2612
2613 def _extract_xspf_playlist(self, xspf_url, playlist_id, fatal=True):
2614 xspf = self._download_xml(
2615 xspf_url, playlist_id, 'Downloading xpsf playlist',
2616 'Unable to download xspf manifest', fatal=fatal)
2617 if xspf is False:
2618 return []
2619 return self._parse_xspf(
2620 xspf, playlist_id, xspf_url=xspf_url,
2621 xspf_base_url=base_url(xspf_url))
2622
2623 def _parse_xspf(self, xspf_doc, playlist_id, xspf_url=None, xspf_base_url=None):
2624 NS_MAP = {
2625 'xspf': 'http://xspf.org/ns/0/',
2626 's1': 'http://static.streamone.nl/player/ns/0',
2627 }
2628
2629 entries = []
2630 for track in xspf_doc.findall(xpath_with_ns('./xspf:trackList/xspf:track', NS_MAP)):
2631 title = xpath_text(
2632 track, xpath_with_ns('./xspf:title', NS_MAP), 'title', default=playlist_id)
2633 description = xpath_text(
2634 track, xpath_with_ns('./xspf:annotation', NS_MAP), 'description')
2635 thumbnail = xpath_text(
2636 track, xpath_with_ns('./xspf:image', NS_MAP), 'thumbnail')
2637 duration = float_or_none(
2638 xpath_text(track, xpath_with_ns('./xspf:duration', NS_MAP), 'duration'), 1000)
2639
2640 formats = []
2641 for location in track.findall(xpath_with_ns('./xspf:location', NS_MAP)):
2642 format_url = urljoin(xspf_base_url, location.text)
2643 if not format_url:
2644 continue
2645 formats.append({
2646 'url': format_url,
2647 'manifest_url': xspf_url,
2648 'format_id': location.get(xpath_with_ns('s1:label', NS_MAP)),
2649 'width': int_or_none(location.get(xpath_with_ns('s1:width', NS_MAP))),
2650 'height': int_or_none(location.get(xpath_with_ns('s1:height', NS_MAP))),
2651 })
2652 self._sort_formats(formats)
2653
2654 entries.append({
2655 'id': playlist_id,
2656 'title': title,
2657 'description': description,
2658 'thumbnail': thumbnail,
2659 'duration': duration,
2660 'formats': formats,
2661 })
2662 return entries
2663
2664 def _extract_mpd_formats(self, *args, **kwargs):
2665 fmts, subs = self._extract_mpd_formats_and_subtitles(*args, **kwargs)
2666 if subs:
2667 self._report_ignoring_subs('DASH')
2668 return fmts
2669
2670 def _extract_mpd_formats_and_subtitles(
2671 self, mpd_url, video_id, mpd_id=None, note=None, errnote=None,
2672 fatal=True, data=None, headers={}, query={}):
2673 res = self._download_xml_handle(
2674 mpd_url, video_id,
2675 note='Downloading MPD manifest' if note is None else note,
2676 errnote='Failed to download MPD manifest' if errnote is None else errnote,
2677 fatal=fatal, data=data, headers=headers, query=query)
2678 if res is False:
2679 return [], {}
2680 mpd_doc, urlh = res
2681 if mpd_doc is None:
2682 return [], {}
2683 mpd_base_url = base_url(urlh.geturl())
2684
2685 return self._parse_mpd_formats_and_subtitles(
2686 mpd_doc, mpd_id, mpd_base_url, mpd_url)
2687
2688 def _parse_mpd_formats(self, *args, **kwargs):
2689 fmts, subs = self._parse_mpd_formats_and_subtitles(*args, **kwargs)
2690 if subs:
2691 self._report_ignoring_subs('DASH')
2692 return fmts
2693
2694 def _parse_mpd_formats_and_subtitles(
2695 self, mpd_doc, mpd_id=None, mpd_base_url='', mpd_url=None):
2696 """
2697 Parse formats from MPD manifest.
2698 References:
2699 1. MPEG-DASH Standard, ISO/IEC 23009-1:2014(E),
2700 http://standards.iso.org/ittf/PubliclyAvailableStandards/c065274_ISO_IEC_23009-1_2014.zip
2701 2. https://en.wikipedia.org/wiki/Dynamic_Adaptive_Streaming_over_HTTP
2702 """
2703 if not self.get_param('dynamic_mpd', True):
2704 if mpd_doc.get('type') == 'dynamic':
2705 return [], {}
2706
2707 namespace = self._search_regex(r'(?i)^{([^}]+)?}MPD$', mpd_doc.tag, 'namespace', default=None)
2708
2709 def _add_ns(path):
2710 return self._xpath_ns(path, namespace)
2711
2712 def is_drm_protected(element):
2713 return element.find(_add_ns('ContentProtection')) is not None
2714
2715 def extract_multisegment_info(element, ms_parent_info):
2716 ms_info = ms_parent_info.copy()
2717
2718 # As per [1, 5.3.9.2.2] SegmentList and SegmentTemplate share some
2719 # common attributes and elements. We will only extract relevant
2720 # for us.
2721 def extract_common(source):
2722 segment_timeline = source.find(_add_ns('SegmentTimeline'))
2723 if segment_timeline is not None:
2724 s_e = segment_timeline.findall(_add_ns('S'))
2725 if s_e:
2726 ms_info['total_number'] = 0
2727 ms_info['s'] = []
2728 for s in s_e:
2729 r = int(s.get('r', 0))
2730 ms_info['total_number'] += 1 + r
2731 ms_info['s'].append({
2732 't': int(s.get('t', 0)),
2733 # @d is mandatory (see [1, 5.3.9.6.2, Table 17, page 60])
2734 'd': int(s.attrib['d']),
2735 'r': r,
2736 })
2737 start_number = source.get('startNumber')
2738 if start_number:
2739 ms_info['start_number'] = int(start_number)
2740 timescale = source.get('timescale')
2741 if timescale:
2742 ms_info['timescale'] = int(timescale)
2743 segment_duration = source.get('duration')
2744 if segment_duration:
2745 ms_info['segment_duration'] = float(segment_duration)
2746
2747 def extract_Initialization(source):
2748 initialization = source.find(_add_ns('Initialization'))
2749 if initialization is not None:
2750 ms_info['initialization_url'] = initialization.attrib['sourceURL']
2751
2752 segment_list = element.find(_add_ns('SegmentList'))
2753 if segment_list is not None:
2754 extract_common(segment_list)
2755 extract_Initialization(segment_list)
2756 segment_urls_e = segment_list.findall(_add_ns('SegmentURL'))
2757 if segment_urls_e:
2758 ms_info['segment_urls'] = [segment.attrib['media'] for segment in segment_urls_e]
2759 else:
2760 segment_template = element.find(_add_ns('SegmentTemplate'))
2761 if segment_template is not None:
2762 extract_common(segment_template)
2763 media = segment_template.get('media')
2764 if media:
2765 ms_info['media'] = media
2766 initialization = segment_template.get('initialization')
2767 if initialization:
2768 ms_info['initialization'] = initialization
2769 else:
2770 extract_Initialization(segment_template)
2771 return ms_info
2772
2773 mpd_duration = parse_duration(mpd_doc.get('mediaPresentationDuration'))
2774 formats, subtitles = [], {}
2775 stream_numbers = collections.defaultdict(int)
2776 for period in mpd_doc.findall(_add_ns('Period')):
2777 period_duration = parse_duration(period.get('duration')) or mpd_duration
2778 period_ms_info = extract_multisegment_info(period, {
2779 'start_number': 1,
2780 'timescale': 1,
2781 })
2782 for adaptation_set in period.findall(_add_ns('AdaptationSet')):
2783 adaption_set_ms_info = extract_multisegment_info(adaptation_set, period_ms_info)
2784 for representation in adaptation_set.findall(_add_ns('Representation')):
2785 representation_attrib = adaptation_set.attrib.copy()
2786 representation_attrib.update(representation.attrib)
2787 # According to [1, 5.3.7.2, Table 9, page 41], @mimeType is mandatory
2788 mime_type = representation_attrib['mimeType']
2789 content_type = representation_attrib.get('contentType', mime_type.split('/')[0])
2790
2791 codecs = parse_codecs(representation_attrib.get('codecs', ''))
2792 if content_type not in ('video', 'audio', 'text'):
2793 if mime_type == 'image/jpeg':
2794 content_type = mime_type
2795 elif codecs['vcodec'] != 'none':
2796 content_type = 'video'
2797 elif codecs['acodec'] != 'none':
2798 content_type = 'audio'
2799 elif codecs.get('tcodec', 'none') != 'none':
2800 content_type = 'text'
2801 elif mimetype2ext(mime_type) in ('tt', 'dfxp', 'ttml', 'xml', 'json'):
2802 content_type = 'text'
2803 else:
2804 self.report_warning('Unknown MIME type %s in DASH manifest' % mime_type)
2805 continue
2806
2807 base_url = ''
2808 for element in (representation, adaptation_set, period, mpd_doc):
2809 base_url_e = element.find(_add_ns('BaseURL'))
2810 if base_url_e is not None:
2811 base_url = base_url_e.text + base_url
2812 if re.match(r'^https?://', base_url):
2813 break
2814 if mpd_base_url and base_url.startswith('/'):
2815 base_url = compat_urlparse.urljoin(mpd_base_url, base_url)
2816 elif mpd_base_url and not re.match(r'^https?://', base_url):
2817 if not mpd_base_url.endswith('/'):
2818 mpd_base_url += '/'
2819 base_url = mpd_base_url + base_url
2820 representation_id = representation_attrib.get('id')
2821 lang = representation_attrib.get('lang')
2822 url_el = representation.find(_add_ns('BaseURL'))
2823 filesize = int_or_none(url_el.attrib.get('{http://youtube.com/yt/2012/10/10}contentLength') if url_el is not None else None)
2824 bandwidth = int_or_none(representation_attrib.get('bandwidth'))
2825 if representation_id is not None:
2826 format_id = representation_id
2827 else:
2828 format_id = content_type
2829 if mpd_id:
2830 format_id = mpd_id + '-' + format_id
2831 if content_type in ('video', 'audio'):
2832 f = {
2833 'format_id': format_id,
2834 'manifest_url': mpd_url,
2835 'ext': mimetype2ext(mime_type),
2836 'width': int_or_none(representation_attrib.get('width')),
2837 'height': int_or_none(representation_attrib.get('height')),
2838 'tbr': float_or_none(bandwidth, 1000),
2839 'asr': int_or_none(representation_attrib.get('audioSamplingRate')),
2840 'fps': int_or_none(representation_attrib.get('frameRate')),
2841 'language': lang if lang not in ('mul', 'und', 'zxx', 'mis') else None,
2842 'format_note': 'DASH %s' % content_type,
2843 'filesize': filesize,
2844 'container': mimetype2ext(mime_type) + '_dash',
2845 **codecs
2846 }
2847 elif content_type == 'text':
2848 f = {
2849 'ext': mimetype2ext(mime_type),
2850 'manifest_url': mpd_url,
2851 'filesize': filesize,
2852 }
2853 elif content_type == 'image/jpeg':
2854 # See test case in VikiIE
2855 # https://www.viki.com/videos/1175236v-choosing-spouse-by-lottery-episode-1
2856 f = {
2857 'format_id': format_id,
2858 'ext': 'mhtml',
2859 'manifest_url': mpd_url,
2860 'format_note': 'DASH storyboards (jpeg)',
2861 'acodec': 'none',
2862 'vcodec': 'none',
2863 }
2864 if is_drm_protected(adaptation_set) or is_drm_protected(representation):
2865 f['has_drm'] = True
2866 representation_ms_info = extract_multisegment_info(representation, adaption_set_ms_info)
2867
2868 def prepare_template(template_name, identifiers):
2869 tmpl = representation_ms_info[template_name]
2870 # First of, % characters outside $...$ templates
2871 # must be escaped by doubling for proper processing
2872 # by % operator string formatting used further (see
2873 # https://github.com/ytdl-org/youtube-dl/issues/16867).
2874 t = ''
2875 in_template = False
2876 for c in tmpl:
2877 t += c
2878 if c == '$':
2879 in_template = not in_template
2880 elif c == '%' and not in_template:
2881 t += c
2882 # Next, $...$ templates are translated to their
2883 # %(...) counterparts to be used with % operator
2884 if representation_id is not None:
2885 t = t.replace('$RepresentationID$', representation_id)
2886 t = re.sub(r'\$(%s)\$' % '|'.join(identifiers), r'%(\1)d', t)
2887 t = re.sub(r'\$(%s)%%([^$]+)\$' % '|'.join(identifiers), r'%(\1)\2', t)
2888 t.replace('$$', '$')
2889 return t
2890
2891 # @initialization is a regular template like @media one
2892 # so it should be handled just the same way (see
2893 # https://github.com/ytdl-org/youtube-dl/issues/11605)
2894 if 'initialization' in representation_ms_info:
2895 initialization_template = prepare_template(
2896 'initialization',
2897 # As per [1, 5.3.9.4.2, Table 15, page 54] $Number$ and
2898 # $Time$ shall not be included for @initialization thus
2899 # only $Bandwidth$ remains
2900 ('Bandwidth', ))
2901 representation_ms_info['initialization_url'] = initialization_template % {
2902 'Bandwidth': bandwidth,
2903 }
2904
2905 def location_key(location):
2906 return 'url' if re.match(r'^https?://', location) else 'path'
2907
2908 if 'segment_urls' not in representation_ms_info and 'media' in representation_ms_info:
2909
2910 media_template = prepare_template('media', ('Number', 'Bandwidth', 'Time'))
2911 media_location_key = location_key(media_template)
2912
2913 # As per [1, 5.3.9.4.4, Table 16, page 55] $Number$ and $Time$
2914 # can't be used at the same time
2915 if '%(Number' in media_template and 's' not in representation_ms_info:
2916 segment_duration = None
2917 if 'total_number' not in representation_ms_info and 'segment_duration' in representation_ms_info:
2918 segment_duration = float_or_none(representation_ms_info['segment_duration'], representation_ms_info['timescale'])
2919 representation_ms_info['total_number'] = int(math.ceil(
2920 float_or_none(period_duration, segment_duration, default=0)))
2921 representation_ms_info['fragments'] = [{
2922 media_location_key: media_template % {
2923 'Number': segment_number,
2924 'Bandwidth': bandwidth,
2925 },
2926 'duration': segment_duration,
2927 } for segment_number in range(
2928 representation_ms_info['start_number'],
2929 representation_ms_info['total_number'] + representation_ms_info['start_number'])]
2930 else:
2931 # $Number*$ or $Time$ in media template with S list available
2932 # Example $Number*$: http://www.svtplay.se/klipp/9023742/stopptid-om-bjorn-borg
2933 # Example $Time$: https://play.arkena.com/embed/avp/v2/player/media/b41dda37-d8e7-4d3f-b1b5-9a9db578bdfe/1/129411
2934 representation_ms_info['fragments'] = []
2935 segment_time = 0
2936 segment_d = None
2937 segment_number = representation_ms_info['start_number']
2938
2939 def add_segment_url():
2940 segment_url = media_template % {
2941 'Time': segment_time,
2942 'Bandwidth': bandwidth,
2943 'Number': segment_number,
2944 }
2945 representation_ms_info['fragments'].append({
2946 media_location_key: segment_url,
2947 'duration': float_or_none(segment_d, representation_ms_info['timescale']),
2948 })
2949
2950 for num, s in enumerate(representation_ms_info['s']):
2951 segment_time = s.get('t') or segment_time
2952 segment_d = s['d']
2953 add_segment_url()
2954 segment_number += 1
2955 for r in range(s.get('r', 0)):
2956 segment_time += segment_d
2957 add_segment_url()
2958 segment_number += 1
2959 segment_time += segment_d
2960 elif 'segment_urls' in representation_ms_info and 's' in representation_ms_info:
2961 # No media template
2962 # Example: https://www.youtube.com/watch?v=iXZV5uAYMJI
2963 # or any YouTube dashsegments video
2964 fragments = []
2965 segment_index = 0
2966 timescale = representation_ms_info['timescale']
2967 for s in representation_ms_info['s']:
2968 duration = float_or_none(s['d'], timescale)
2969 for r in range(s.get('r', 0) + 1):
2970 segment_uri = representation_ms_info['segment_urls'][segment_index]
2971 fragments.append({
2972 location_key(segment_uri): segment_uri,
2973 'duration': duration,
2974 })
2975 segment_index += 1
2976 representation_ms_info['fragments'] = fragments
2977 elif 'segment_urls' in representation_ms_info:
2978 # Segment URLs with no SegmentTimeline
2979 # Example: https://www.seznam.cz/zpravy/clanek/cesko-zasahne-vitr-o-sile-vichrice-muze-byt-i-zivotu-nebezpecny-39091
2980 # https://github.com/ytdl-org/youtube-dl/pull/14844
2981 fragments = []
2982 segment_duration = float_or_none(
2983 representation_ms_info['segment_duration'],
2984 representation_ms_info['timescale']) if 'segment_duration' in representation_ms_info else None
2985 for segment_url in representation_ms_info['segment_urls']:
2986 fragment = {
2987 location_key(segment_url): segment_url,
2988 }
2989 if segment_duration:
2990 fragment['duration'] = segment_duration
2991 fragments.append(fragment)
2992 representation_ms_info['fragments'] = fragments
2993 # If there is a fragments key available then we correctly recognized fragmented media.
2994 # Otherwise we will assume unfragmented media with direct access. Technically, such
2995 # assumption is not necessarily correct since we may simply have no support for
2996 # some forms of fragmented media renditions yet, but for now we'll use this fallback.
2997 if 'fragments' in representation_ms_info:
2998 f.update({
2999 # NB: mpd_url may be empty when MPD manifest is parsed from a string
3000 'url': mpd_url or base_url,
3001 'fragment_base_url': base_url,
3002 'fragments': [],
3003 'protocol': 'http_dash_segments' if mime_type != 'image/jpeg' else 'mhtml',
3004 })
3005 if 'initialization_url' in representation_ms_info:
3006 initialization_url = representation_ms_info['initialization_url']
3007 if not f.get('url'):
3008 f['url'] = initialization_url
3009 f['fragments'].append({location_key(initialization_url): initialization_url})
3010 f['fragments'].extend(representation_ms_info['fragments'])
3011 if not period_duration:
3012 period_duration = try_get(
3013 representation_ms_info,
3014 lambda r: sum(frag['duration'] for frag in r['fragments']), float)
3015 else:
3016 # Assuming direct URL to unfragmented media.
3017 f['url'] = base_url
3018 if content_type in ('video', 'audio', 'image/jpeg'):
3019 f['manifest_stream_number'] = stream_numbers[f['url']]
3020 stream_numbers[f['url']] += 1
3021 formats.append(f)
3022 elif content_type == 'text':
3023 subtitles.setdefault(lang or 'und', []).append(f)
3024
3025 return formats, subtitles
3026
3027 def _extract_ism_formats(self, *args, **kwargs):
3028 fmts, subs = self._extract_ism_formats_and_subtitles(*args, **kwargs)
3029 if subs:
3030 self._report_ignoring_subs('ISM')
3031 return fmts
3032
3033 def _extract_ism_formats_and_subtitles(self, ism_url, video_id, ism_id=None, note=None, errnote=None, fatal=True, data=None, headers={}, query={}):
3034 res = self._download_xml_handle(
3035 ism_url, video_id,
3036 note='Downloading ISM manifest' if note is None else note,
3037 errnote='Failed to download ISM manifest' if errnote is None else errnote,
3038 fatal=fatal, data=data, headers=headers, query=query)
3039 if res is False:
3040 return [], {}
3041 ism_doc, urlh = res
3042 if ism_doc is None:
3043 return [], {}
3044
3045 return self._parse_ism_formats_and_subtitles(ism_doc, urlh.geturl(), ism_id)
3046
3047 def _parse_ism_formats_and_subtitles(self, ism_doc, ism_url, ism_id=None):
3048 """
3049 Parse formats from ISM manifest.
3050 References:
3051 1. [MS-SSTR]: Smooth Streaming Protocol,
3052 https://msdn.microsoft.com/en-us/library/ff469518.aspx
3053 """
3054 if ism_doc.get('IsLive') == 'TRUE':
3055 return [], {}
3056
3057 duration = int(ism_doc.attrib['Duration'])
3058 timescale = int_or_none(ism_doc.get('TimeScale')) or 10000000
3059
3060 formats = []
3061 subtitles = {}
3062 for stream in ism_doc.findall('StreamIndex'):
3063 stream_type = stream.get('Type')
3064 if stream_type not in ('video', 'audio', 'text'):
3065 continue
3066 url_pattern = stream.attrib['Url']
3067 stream_timescale = int_or_none(stream.get('TimeScale')) or timescale
3068 stream_name = stream.get('Name')
3069 stream_language = stream.get('Language', 'und')
3070 for track in stream.findall('QualityLevel'):
3071 fourcc = track.get('FourCC') or ('AACL' if track.get('AudioTag') == '255' else None)
3072 # TODO: add support for WVC1 and WMAP
3073 if fourcc not in ('H264', 'AVC1', 'AACL', 'TTML'):
3074 self.report_warning('%s is not a supported codec' % fourcc)
3075 continue
3076 tbr = int(track.attrib['Bitrate']) // 1000
3077 # [1] does not mention Width and Height attributes. However,
3078 # they're often present while MaxWidth and MaxHeight are
3079 # missing, so should be used as fallbacks
3080 width = int_or_none(track.get('MaxWidth') or track.get('Width'))
3081 height = int_or_none(track.get('MaxHeight') or track.get('Height'))
3082 sampling_rate = int_or_none(track.get('SamplingRate'))
3083
3084 track_url_pattern = re.sub(r'{[Bb]itrate}', track.attrib['Bitrate'], url_pattern)
3085 track_url_pattern = compat_urlparse.urljoin(ism_url, track_url_pattern)
3086
3087 fragments = []
3088 fragment_ctx = {
3089 'time': 0,
3090 }
3091 stream_fragments = stream.findall('c')
3092 for stream_fragment_index, stream_fragment in enumerate(stream_fragments):
3093 fragment_ctx['time'] = int_or_none(stream_fragment.get('t')) or fragment_ctx['time']
3094 fragment_repeat = int_or_none(stream_fragment.get('r')) or 1
3095 fragment_ctx['duration'] = int_or_none(stream_fragment.get('d'))
3096 if not fragment_ctx['duration']:
3097 try:
3098 next_fragment_time = int(stream_fragment[stream_fragment_index + 1].attrib['t'])
3099 except IndexError:
3100 next_fragment_time = duration
3101 fragment_ctx['duration'] = (next_fragment_time - fragment_ctx['time']) / fragment_repeat
3102 for _ in range(fragment_repeat):
3103 fragments.append({
3104 'url': re.sub(r'{start[ _]time}', compat_str(fragment_ctx['time']), track_url_pattern),
3105 'duration': fragment_ctx['duration'] / stream_timescale,
3106 })
3107 fragment_ctx['time'] += fragment_ctx['duration']
3108
3109 if stream_type == 'text':
3110 subtitles.setdefault(stream_language, []).append({
3111 'ext': 'ismt',
3112 'protocol': 'ism',
3113 'url': ism_url,
3114 'manifest_url': ism_url,
3115 'fragments': fragments,
3116 '_download_params': {
3117 'stream_type': stream_type,
3118 'duration': duration,
3119 'timescale': stream_timescale,
3120 'fourcc': fourcc,
3121 'language': stream_language,
3122 'codec_private_data': track.get('CodecPrivateData'),
3123 }
3124 })
3125 elif stream_type in ('video', 'audio'):
3126 formats.append({
3127 'format_id': join_nonempty(ism_id, stream_name, tbr),
3128 'url': ism_url,
3129 'manifest_url': ism_url,
3130 'ext': 'ismv' if stream_type == 'video' else 'isma',
3131 'width': width,
3132 'height': height,
3133 'tbr': tbr,
3134 'asr': sampling_rate,
3135 'vcodec': 'none' if stream_type == 'audio' else fourcc,
3136 'acodec': 'none' if stream_type == 'video' else fourcc,
3137 'protocol': 'ism',
3138 'fragments': fragments,
3139 'has_drm': ism_doc.find('Protection') is not None,
3140 '_download_params': {
3141 'stream_type': stream_type,
3142 'duration': duration,
3143 'timescale': stream_timescale,
3144 'width': width or 0,
3145 'height': height or 0,
3146 'fourcc': fourcc,
3147 'language': stream_language,
3148 'codec_private_data': track.get('CodecPrivateData'),
3149 'sampling_rate': sampling_rate,
3150 'channels': int_or_none(track.get('Channels', 2)),
3151 'bits_per_sample': int_or_none(track.get('BitsPerSample', 16)),
3152 'nal_unit_length_field': int_or_none(track.get('NALUnitLengthField', 4)),
3153 },
3154 })
3155 return formats, subtitles
3156
3157 def _parse_html5_media_entries(self, base_url, webpage, video_id, m3u8_id=None, m3u8_entry_protocol='m3u8_native', mpd_id=None, preference=None, quality=None):
3158 def absolute_url(item_url):
3159 return urljoin(base_url, item_url)
3160
3161 def parse_content_type(content_type):
3162 if not content_type:
3163 return {}
3164 ctr = re.search(r'(?P<mimetype>[^/]+/[^;]+)(?:;\s*codecs="?(?P<codecs>[^"]+))?', content_type)
3165 if ctr:
3166 mimetype, codecs = ctr.groups()
3167 f = parse_codecs(codecs)
3168 f['ext'] = mimetype2ext(mimetype)
3169 return f
3170 return {}
3171
3172 def _media_formats(src, cur_media_type, type_info={}):
3173 full_url = absolute_url(src)
3174 ext = type_info.get('ext') or determine_ext(full_url)
3175 if ext == 'm3u8':
3176 is_plain_url = False
3177 formats = self._extract_m3u8_formats(
3178 full_url, video_id, ext='mp4',
3179 entry_protocol=m3u8_entry_protocol, m3u8_id=m3u8_id,
3180 preference=preference, quality=quality, fatal=False)
3181 elif ext == 'mpd':
3182 is_plain_url = False
3183 formats = self._extract_mpd_formats(
3184 full_url, video_id, mpd_id=mpd_id, fatal=False)
3185 else:
3186 is_plain_url = True
3187 formats = [{
3188 'url': full_url,
3189 'vcodec': 'none' if cur_media_type == 'audio' else None,
3190 }]
3191 return is_plain_url, formats
3192
3193 entries = []
3194 # amp-video and amp-audio are very similar to their HTML5 counterparts
3195 # so we wll include them right here (see
3196 # https://www.ampproject.org/docs/reference/components/amp-video)
3197 # For dl8-* tags see https://delight-vr.com/documentation/dl8-video/
3198 _MEDIA_TAG_NAME_RE = r'(?:(?:amp|dl8(?:-live)?)-)?(video|audio)'
3199 media_tags = [(media_tag, media_tag_name, media_type, '')
3200 for media_tag, media_tag_name, media_type
3201 in re.findall(r'(?s)(<(%s)[^>]*/>)' % _MEDIA_TAG_NAME_RE, webpage)]
3202 media_tags.extend(re.findall(
3203 # We only allow video|audio followed by a whitespace or '>'.
3204 # Allowing more characters may end up in significant slow down (see
3205 # https://github.com/ytdl-org/youtube-dl/issues/11979, example URL:
3206 # http://www.porntrex.com/maps/videositemap.xml).
3207 r'(?s)(<(?P<tag>%s)(?:\s+[^>]*)?>)(.*?)</(?P=tag)>' % _MEDIA_TAG_NAME_RE, webpage))
3208 for media_tag, _, media_type, media_content in media_tags:
3209 media_info = {
3210 'formats': [],
3211 'subtitles': {},
3212 }
3213 media_attributes = extract_attributes(media_tag)
3214 src = strip_or_none(media_attributes.get('src'))
3215 if src:
3216 _, formats = _media_formats(src, media_type)
3217 media_info['formats'].extend(formats)
3218 media_info['thumbnail'] = absolute_url(media_attributes.get('poster'))
3219 if media_content:
3220 for source_tag in re.findall(r'<source[^>]+>', media_content):
3221 s_attr = extract_attributes(source_tag)
3222 # data-video-src and data-src are non standard but seen
3223 # several times in the wild
3224 src = strip_or_none(dict_get(s_attr, ('src', 'data-video-src', 'data-src')))
3225 if not src:
3226 continue
3227 f = parse_content_type(s_attr.get('type'))
3228 is_plain_url, formats = _media_formats(src, media_type, f)
3229 if is_plain_url:
3230 # width, height, res, label and title attributes are
3231 # all not standard but seen several times in the wild
3232 labels = [
3233 s_attr.get(lbl)
3234 for lbl in ('label', 'title')
3235 if str_or_none(s_attr.get(lbl))
3236 ]
3237 width = int_or_none(s_attr.get('width'))
3238 height = (int_or_none(s_attr.get('height'))
3239 or int_or_none(s_attr.get('res')))
3240 if not width or not height:
3241 for lbl in labels:
3242 resolution = parse_resolution(lbl)
3243 if not resolution:
3244 continue
3245 width = width or resolution.get('width')
3246 height = height or resolution.get('height')
3247 for lbl in labels:
3248 tbr = parse_bitrate(lbl)
3249 if tbr:
3250 break
3251 else:
3252 tbr = None
3253 f.update({
3254 'width': width,
3255 'height': height,
3256 'tbr': tbr,
3257 'format_id': s_attr.get('label') or s_attr.get('title'),
3258 })
3259 f.update(formats[0])
3260 media_info['formats'].append(f)
3261 else:
3262 media_info['formats'].extend(formats)
3263 for track_tag in re.findall(r'<track[^>]+>', media_content):
3264 track_attributes = extract_attributes(track_tag)
3265 kind = track_attributes.get('kind')
3266 if not kind or kind in ('subtitles', 'captions'):
3267 src = strip_or_none(track_attributes.get('src'))
3268 if not src:
3269 continue
3270 lang = track_attributes.get('srclang') or track_attributes.get('lang') or track_attributes.get('label')
3271 media_info['subtitles'].setdefault(lang, []).append({
3272 'url': absolute_url(src),
3273 })
3274 for f in media_info['formats']:
3275 f.setdefault('http_headers', {})['Referer'] = base_url
3276 if media_info['formats'] or media_info['subtitles']:
3277 entries.append(media_info)
3278 return entries
3279
3280 def _extract_akamai_formats(self, *args, **kwargs):
3281 fmts, subs = self._extract_akamai_formats_and_subtitles(*args, **kwargs)
3282 if subs:
3283 self._report_ignoring_subs('akamai')
3284 return fmts
3285
3286 def _extract_akamai_formats_and_subtitles(self, manifest_url, video_id, hosts={}):
3287 signed = 'hdnea=' in manifest_url
3288 if not signed:
3289 # https://learn.akamai.com/en-us/webhelp/media-services-on-demand/stream-packaging-user-guide/GUID-BE6C0F73-1E06-483B-B0EA-57984B91B7F9.html
3290 manifest_url = re.sub(
3291 r'(?:b=[\d,-]+|(?:__a__|attributes)=off|__b__=\d+)&?',
3292 '', manifest_url).strip('?')
3293
3294 formats = []
3295 subtitles = {}
3296
3297 hdcore_sign = 'hdcore=3.7.0'
3298 f4m_url = re.sub(r'(https?://[^/]+)/i/', r'\1/z/', manifest_url).replace('/master.m3u8', '/manifest.f4m')
3299 hds_host = hosts.get('hds')
3300 if hds_host:
3301 f4m_url = re.sub(r'(https?://)[^/]+', r'\1' + hds_host, f4m_url)
3302 if 'hdcore=' not in f4m_url:
3303 f4m_url += ('&' if '?' in f4m_url else '?') + hdcore_sign
3304 f4m_formats = self._extract_f4m_formats(
3305 f4m_url, video_id, f4m_id='hds', fatal=False)
3306 for entry in f4m_formats:
3307 entry.update({'extra_param_to_segment_url': hdcore_sign})
3308 formats.extend(f4m_formats)
3309
3310 m3u8_url = re.sub(r'(https?://[^/]+)/z/', r'\1/i/', manifest_url).replace('/manifest.f4m', '/master.m3u8')
3311 hls_host = hosts.get('hls')
3312 if hls_host:
3313 m3u8_url = re.sub(r'(https?://)[^/]+', r'\1' + hls_host, m3u8_url)
3314 m3u8_formats, m3u8_subtitles = self._extract_m3u8_formats_and_subtitles(
3315 m3u8_url, video_id, 'mp4', 'm3u8_native',
3316 m3u8_id='hls', fatal=False)
3317 formats.extend(m3u8_formats)
3318 subtitles = self._merge_subtitles(subtitles, m3u8_subtitles)
3319
3320 http_host = hosts.get('http')
3321 if http_host and m3u8_formats and not signed:
3322 REPL_REGEX = r'https?://[^/]+/i/([^,]+),([^/]+),([^/]+)\.csmil/.+'
3323 qualities = re.match(REPL_REGEX, m3u8_url).group(2).split(',')
3324 qualities_length = len(qualities)
3325 if len(m3u8_formats) in (qualities_length, qualities_length + 1):
3326 i = 0
3327 for f in m3u8_formats:
3328 if f['vcodec'] != 'none':
3329 for protocol in ('http', 'https'):
3330 http_f = f.copy()
3331 del http_f['manifest_url']
3332 http_url = re.sub(
3333 REPL_REGEX, protocol + r'://%s/\g<1>%s\3' % (http_host, qualities[i]), f['url'])
3334 http_f.update({
3335 'format_id': http_f['format_id'].replace('hls-', protocol + '-'),
3336 'url': http_url,
3337 'protocol': protocol,
3338 })
3339 formats.append(http_f)
3340 i += 1
3341
3342 return formats, subtitles
3343
3344 def _extract_wowza_formats(self, url, video_id, m3u8_entry_protocol='m3u8_native', skip_protocols=[]):
3345 query = compat_urlparse.urlparse(url).query
3346 url = re.sub(r'/(?:manifest|playlist|jwplayer)\.(?:m3u8|f4m|mpd|smil)', '', url)
3347 mobj = re.search(
3348 r'(?:(?:http|rtmp|rtsp)(?P<s>s)?:)?(?P<url>//[^?]+)', url)
3349 url_base = mobj.group('url')
3350 http_base_url = '%s%s:%s' % ('http', mobj.group('s') or '', url_base)
3351 formats = []
3352
3353 def manifest_url(manifest):
3354 m_url = '%s/%s' % (http_base_url, manifest)
3355 if query:
3356 m_url += '?%s' % query
3357 return m_url
3358
3359 if 'm3u8' not in skip_protocols:
3360 formats.extend(self._extract_m3u8_formats(
3361 manifest_url('playlist.m3u8'), video_id, 'mp4',
3362 m3u8_entry_protocol, m3u8_id='hls', fatal=False))
3363 if 'f4m' not in skip_protocols:
3364 formats.extend(self._extract_f4m_formats(
3365 manifest_url('manifest.f4m'),
3366 video_id, f4m_id='hds', fatal=False))
3367 if 'dash' not in skip_protocols:
3368 formats.extend(self._extract_mpd_formats(
3369 manifest_url('manifest.mpd'),
3370 video_id, mpd_id='dash', fatal=False))
3371 if re.search(r'(?:/smil:|\.smil)', url_base):
3372 if 'smil' not in skip_protocols:
3373 rtmp_formats = self._extract_smil_formats(
3374 manifest_url('jwplayer.smil'),
3375 video_id, fatal=False)
3376 for rtmp_format in rtmp_formats:
3377 rtsp_format = rtmp_format.copy()
3378 rtsp_format['url'] = '%s/%s' % (rtmp_format['url'], rtmp_format['play_path'])
3379 del rtsp_format['play_path']
3380 del rtsp_format['ext']
3381 rtsp_format.update({
3382 'url': rtsp_format['url'].replace('rtmp://', 'rtsp://'),
3383 'format_id': rtmp_format['format_id'].replace('rtmp', 'rtsp'),
3384 'protocol': 'rtsp',
3385 })
3386 formats.extend([rtmp_format, rtsp_format])
3387 else:
3388 for protocol in ('rtmp', 'rtsp'):
3389 if protocol not in skip_protocols:
3390 formats.append({
3391 'url': '%s:%s' % (protocol, url_base),
3392 'format_id': protocol,
3393 'protocol': protocol,
3394 })
3395 return formats
3396
3397 def _find_jwplayer_data(self, webpage, video_id=None, transform_source=js_to_json):
3398 mobj = re.search(
3399 r'(?s)jwplayer\((?P<quote>[\'"])[^\'" ]+(?P=quote)\)(?!</script>).*?\.setup\s*\((?P<options>[^)]+)\)',
3400 webpage)
3401 if mobj:
3402 try:
3403 jwplayer_data = self._parse_json(mobj.group('options'),
3404 video_id=video_id,
3405 transform_source=transform_source)
3406 except ExtractorError:
3407 pass
3408 else:
3409 if isinstance(jwplayer_data, dict):
3410 return jwplayer_data
3411
3412 def _extract_jwplayer_data(self, webpage, video_id, *args, **kwargs):
3413 jwplayer_data = self._find_jwplayer_data(
3414 webpage, video_id, transform_source=js_to_json)
3415 return self._parse_jwplayer_data(
3416 jwplayer_data, video_id, *args, **kwargs)
3417
3418 def _parse_jwplayer_data(self, jwplayer_data, video_id=None, require_title=True,
3419 m3u8_id=None, mpd_id=None, rtmp_params=None, base_url=None):
3420 # JWPlayer backward compatibility: flattened playlists
3421 # https://github.com/jwplayer/jwplayer/blob/v7.4.3/src/js/api/config.js#L81-L96
3422 if 'playlist' not in jwplayer_data:
3423 jwplayer_data = {'playlist': [jwplayer_data]}
3424
3425 entries = []
3426
3427 # JWPlayer backward compatibility: single playlist item
3428 # https://github.com/jwplayer/jwplayer/blob/v7.7.0/src/js/playlist/playlist.js#L10
3429 if not isinstance(jwplayer_data['playlist'], list):
3430 jwplayer_data['playlist'] = [jwplayer_data['playlist']]
3431
3432 for video_data in jwplayer_data['playlist']:
3433 # JWPlayer backward compatibility: flattened sources
3434 # https://github.com/jwplayer/jwplayer/blob/v7.4.3/src/js/playlist/item.js#L29-L35
3435 if 'sources' not in video_data:
3436 video_data['sources'] = [video_data]
3437
3438 this_video_id = video_id or video_data['mediaid']
3439
3440 formats = self._parse_jwplayer_formats(
3441 video_data['sources'], video_id=this_video_id, m3u8_id=m3u8_id,
3442 mpd_id=mpd_id, rtmp_params=rtmp_params, base_url=base_url)
3443
3444 subtitles = {}
3445 tracks = video_data.get('tracks')
3446 if tracks and isinstance(tracks, list):
3447 for track in tracks:
3448 if not isinstance(track, dict):
3449 continue
3450 track_kind = track.get('kind')
3451 if not track_kind or not isinstance(track_kind, compat_str):
3452 continue
3453 if track_kind.lower() not in ('captions', 'subtitles'):
3454 continue
3455 track_url = urljoin(base_url, track.get('file'))
3456 if not track_url:
3457 continue
3458 subtitles.setdefault(track.get('label') or 'en', []).append({
3459 'url': self._proto_relative_url(track_url)
3460 })
3461
3462 entry = {
3463 'id': this_video_id,
3464 'title': unescapeHTML(video_data['title'] if require_title else video_data.get('title')),
3465 'description': clean_html(video_data.get('description')),
3466 'thumbnail': urljoin(base_url, self._proto_relative_url(video_data.get('image'))),
3467 'timestamp': int_or_none(video_data.get('pubdate')),
3468 'duration': float_or_none(jwplayer_data.get('duration') or video_data.get('duration')),
3469 'subtitles': subtitles,
3470 }
3471 # https://github.com/jwplayer/jwplayer/blob/master/src/js/utils/validator.js#L32
3472 if len(formats) == 1 and re.search(r'^(?:http|//).*(?:youtube\.com|youtu\.be)/.+', formats[0]['url']):
3473 entry.update({
3474 '_type': 'url_transparent',
3475 'url': formats[0]['url'],
3476 })
3477 else:
3478 self._sort_formats(formats)
3479 entry['formats'] = formats
3480 entries.append(entry)
3481 if len(entries) == 1:
3482 return entries[0]
3483 else:
3484 return self.playlist_result(entries)
3485
3486 def _parse_jwplayer_formats(self, jwplayer_sources_data, video_id=None,
3487 m3u8_id=None, mpd_id=None, rtmp_params=None, base_url=None):
3488 urls = []
3489 formats = []
3490 for source in jwplayer_sources_data:
3491 if not isinstance(source, dict):
3492 continue
3493 source_url = urljoin(
3494 base_url, self._proto_relative_url(source.get('file')))
3495 if not source_url or source_url in urls:
3496 continue
3497 urls.append(source_url)
3498 source_type = source.get('type') or ''
3499 ext = mimetype2ext(source_type) or determine_ext(source_url)
3500 if source_type == 'hls' or ext == 'm3u8':
3501 formats.extend(self._extract_m3u8_formats(
3502 source_url, video_id, 'mp4', entry_protocol='m3u8_native',
3503 m3u8_id=m3u8_id, fatal=False))
3504 elif source_type == 'dash' or ext == 'mpd':
3505 formats.extend(self._extract_mpd_formats(
3506 source_url, video_id, mpd_id=mpd_id, fatal=False))
3507 elif ext == 'smil':
3508 formats.extend(self._extract_smil_formats(
3509 source_url, video_id, fatal=False))
3510 # https://github.com/jwplayer/jwplayer/blob/master/src/js/providers/default.js#L67
3511 elif source_type.startswith('audio') or ext in (
3512 'oga', 'aac', 'mp3', 'mpeg', 'vorbis'):
3513 formats.append({
3514 'url': source_url,
3515 'vcodec': 'none',
3516 'ext': ext,
3517 })
3518 else:
3519 height = int_or_none(source.get('height'))
3520 if height is None:
3521 # Often no height is provided but there is a label in
3522 # format like "1080p", "720p SD", or 1080.
3523 height = int_or_none(self._search_regex(
3524 r'^(\d{3,4})[pP]?(?:\b|$)', compat_str(source.get('label') or ''),
3525 'height', default=None))
3526 a_format = {
3527 'url': source_url,
3528 'width': int_or_none(source.get('width')),
3529 'height': height,
3530 'tbr': int_or_none(source.get('bitrate')),
3531 'ext': ext,
3532 }
3533 if source_url.startswith('rtmp'):
3534 a_format['ext'] = 'flv'
3535 # See com/longtailvideo/jwplayer/media/RTMPMediaProvider.as
3536 # of jwplayer.flash.swf
3537 rtmp_url_parts = re.split(
3538 r'((?:mp4|mp3|flv):)', source_url, 1)
3539 if len(rtmp_url_parts) == 3:
3540 rtmp_url, prefix, play_path = rtmp_url_parts
3541 a_format.update({
3542 'url': rtmp_url,
3543 'play_path': prefix + play_path,
3544 })
3545 if rtmp_params:
3546 a_format.update(rtmp_params)
3547 formats.append(a_format)
3548 return formats
3549
3550 def _live_title(self, name):
3551 self._downloader.deprecation_warning('yt_dlp.InfoExtractor._live_title is deprecated and does not work as expected')
3552 return name
3553
3554 def _int(self, v, name, fatal=False, **kwargs):
3555 res = int_or_none(v, **kwargs)
3556 if res is None:
3557 msg = 'Failed to extract %s: Could not parse value %r' % (name, v)
3558 if fatal:
3559 raise ExtractorError(msg)
3560 else:
3561 self.report_warning(msg)
3562 return res
3563
3564 def _float(self, v, name, fatal=False, **kwargs):
3565 res = float_or_none(v, **kwargs)
3566 if res is None:
3567 msg = 'Failed to extract %s: Could not parse value %r' % (name, v)
3568 if fatal:
3569 raise ExtractorError(msg)
3570 else:
3571 self.report_warning(msg)
3572 return res
3573
3574 def _set_cookie(self, domain, name, value, expire_time=None, port=None,
3575 path='/', secure=False, discard=False, rest={}, **kwargs):
3576 cookie = compat_cookiejar_Cookie(
3577 0, name, value, port, port is not None, domain, True,
3578 domain.startswith('.'), path, True, secure, expire_time,
3579 discard, None, None, rest)
3580 self._downloader.cookiejar.set_cookie(cookie)
3581
3582 def _get_cookies(self, url):
3583 """ Return a compat_cookies_SimpleCookie with the cookies for the url """
3584 req = sanitized_Request(url)
3585 self._downloader.cookiejar.add_cookie_header(req)
3586 return compat_cookies_SimpleCookie(req.get_header('Cookie'))
3587
3588 def _apply_first_set_cookie_header(self, url_handle, cookie):
3589 """
3590 Apply first Set-Cookie header instead of the last. Experimental.
3591
3592 Some sites (e.g. [1-3]) may serve two cookies under the same name
3593 in Set-Cookie header and expect the first (old) one to be set rather
3594 than second (new). However, as of RFC6265 the newer one cookie
3595 should be set into cookie store what actually happens.
3596 We will workaround this issue by resetting the cookie to
3597 the first one manually.
3598 1. https://new.vk.com/
3599 2. https://github.com/ytdl-org/youtube-dl/issues/9841#issuecomment-227871201
3600 3. https://learning.oreilly.com/
3601 """
3602 for header, cookies in url_handle.headers.items():
3603 if header.lower() != 'set-cookie':
3604 continue
3605 if sys.version_info[0] >= 3:
3606 cookies = cookies.encode('iso-8859-1')
3607 cookies = cookies.decode('utf-8')
3608 cookie_value = re.search(
3609 r'%s=(.+?);.*?\b[Dd]omain=(.+?)(?:[,;]|$)' % cookie, cookies)
3610 if cookie_value:
3611 value, domain = cookie_value.groups()
3612 self._set_cookie(domain, cookie, value)
3613 break
3614
3615 def get_testcases(self, include_onlymatching=False):
3616 t = getattr(self, '_TEST', None)
3617 if t:
3618 assert not hasattr(self, '_TESTS'), \
3619 '%s has _TEST and _TESTS' % type(self).__name__
3620 tests = [t]
3621 else:
3622 tests = getattr(self, '_TESTS', [])
3623 for t in tests:
3624 if not include_onlymatching and t.get('only_matching', False):
3625 continue
3626 t['name'] = type(self).__name__[:-len('IE')]
3627 yield t
3628
3629 def is_suitable(self, age_limit):
3630 """ Test whether the extractor is generally suitable for the given
3631 age limit (i.e. pornographic sites are not, all others usually are) """
3632
3633 any_restricted = False
3634 for tc in self.get_testcases(include_onlymatching=False):
3635 if tc.get('playlist', []):
3636 tc = tc['playlist'][0]
3637 is_restricted = age_restricted(
3638 tc.get('info_dict', {}).get('age_limit'), age_limit)
3639 if not is_restricted:
3640 return True
3641 any_restricted = any_restricted or is_restricted
3642 return not any_restricted
3643
3644 def extract_subtitles(self, *args, **kwargs):
3645 if (self.get_param('writesubtitles', False)
3646 or self.get_param('listsubtitles')):
3647 return self._get_subtitles(*args, **kwargs)
3648 return {}
3649
3650 def _get_subtitles(self, *args, **kwargs):
3651 raise NotImplementedError('This method must be implemented by subclasses')
3652
3653 def extract_comments(self, *args, **kwargs):
3654 if not self.get_param('getcomments'):
3655 return None
3656 generator = self._get_comments(*args, **kwargs)
3657
3658 def extractor():
3659 comments = []
3660 interrupted = True
3661 try:
3662 while True:
3663 comments.append(next(generator))
3664 except StopIteration:
3665 interrupted = False
3666 except KeyboardInterrupt:
3667 self.to_screen('Interrupted by user')
3668 except Exception as e:
3669 if self.get_param('ignoreerrors') is not True:
3670 raise
3671 self._downloader.report_error(e)
3672 comment_count = len(comments)
3673 self.to_screen(f'Extracted {comment_count} comments')
3674 return {
3675 'comments': comments,
3676 'comment_count': None if interrupted else comment_count
3677 }
3678 return extractor
3679
3680 def _get_comments(self, *args, **kwargs):
3681 raise NotImplementedError('This method must be implemented by subclasses')
3682
3683 @staticmethod
3684 def _merge_subtitle_items(subtitle_list1, subtitle_list2):
3685 """ Merge subtitle items for one language. Items with duplicated URLs/data
3686 will be dropped. """
3687 list1_data = set([item.get('url') or item['data'] for item in subtitle_list1])
3688 ret = list(subtitle_list1)
3689 ret.extend([item for item in subtitle_list2 if (item.get('url') or item['data']) not in list1_data])
3690 return ret
3691
3692 @classmethod
3693 def _merge_subtitles(cls, *dicts, target=None):
3694 """ Merge subtitle dictionaries, language by language. """
3695 if target is None:
3696 target = {}
3697 for d in dicts:
3698 for lang, subs in d.items():
3699 target[lang] = cls._merge_subtitle_items(target.get(lang, []), subs)
3700 return target
3701
3702 def extract_automatic_captions(self, *args, **kwargs):
3703 if (self.get_param('writeautomaticsub', False)
3704 or self.get_param('listsubtitles')):
3705 return self._get_automatic_captions(*args, **kwargs)
3706 return {}
3707
3708 def _get_automatic_captions(self, *args, **kwargs):
3709 raise NotImplementedError('This method must be implemented by subclasses')
3710
3711 def mark_watched(self, *args, **kwargs):
3712 if not self.get_param('mark_watched', False):
3713 return
3714 if (self.supports_login() and self._get_login_info()[0] is not None
3715 or self.get_param('cookiefile') or self.get_param('cookiesfrombrowser')):
3716 self._mark_watched(*args, **kwargs)
3717
3718 def _mark_watched(self, *args, **kwargs):
3719 raise NotImplementedError('This method must be implemented by subclasses')
3720
3721 def geo_verification_headers(self):
3722 headers = {}
3723 geo_verification_proxy = self.get_param('geo_verification_proxy')
3724 if geo_verification_proxy:
3725 headers['Ytdl-request-proxy'] = geo_verification_proxy
3726 return headers
3727
3728 def _generic_id(self, url):
3729 return compat_urllib_parse_unquote(os.path.splitext(url.rstrip('/').split('/')[-1])[0])
3730
3731 def _generic_title(self, url):
3732 return compat_urllib_parse_unquote(os.path.splitext(url_basename(url))[0])
3733
3734 @staticmethod
3735 def _availability(is_private=None, needs_premium=None, needs_subscription=None, needs_auth=None, is_unlisted=None):
3736 all_known = all(map(
3737 lambda x: x is not None,
3738 (is_private, needs_premium, needs_subscription, needs_auth, is_unlisted)))
3739 return (
3740 'private' if is_private
3741 else 'premium_only' if needs_premium
3742 else 'subscriber_only' if needs_subscription
3743 else 'needs_auth' if needs_auth
3744 else 'unlisted' if is_unlisted
3745 else 'public' if all_known
3746 else None)
3747
3748 def _configuration_arg(self, key, default=NO_DEFAULT, *, ie_key=None, casesense=False):
3749 '''
3750 @returns A list of values for the extractor argument given by "key"
3751 or "default" if no such key is present
3752 @param default The default value to return when the key is not present (default: [])
3753 @param casesense When false, the values are converted to lower case
3754 '''
3755 val = traverse_obj(
3756 self._downloader.params, ('extractor_args', (ie_key or self.ie_key()).lower(), key))
3757 if val is None:
3758 return [] if default is NO_DEFAULT else default
3759 return list(val) if casesense else [x.lower() for x in val]
3760
3761 def _yes_playlist(self, playlist_id, video_id, smuggled_data=None, *, playlist_label='playlist', video_label='video'):
3762 if not playlist_id or not video_id:
3763 return not video_id
3764
3765 no_playlist = (smuggled_data or {}).get('force_noplaylist')
3766 if no_playlist is not None:
3767 return not no_playlist
3768
3769 video_id = '' if video_id is True else f' {video_id}'
3770 playlist_id = '' if playlist_id is True else f' {playlist_id}'
3771 if self.get_param('noplaylist'):
3772 self.to_screen(f'Downloading just the {video_label}{video_id} because of --no-playlist')
3773 return False
3774 self.to_screen(f'Downloading {playlist_label}{playlist_id} - add --no-playlist to download just the {video_label}{video_id}')
3775 return True
3776
3777
3778 class SearchInfoExtractor(InfoExtractor):
3779 """
3780 Base class for paged search queries extractors.
3781 They accept URLs in the format _SEARCH_KEY(|all|[0-9]):{query}
3782 Instances should define _SEARCH_KEY and optionally _MAX_RESULTS
3783 """
3784
3785 _MAX_RESULTS = float('inf')
3786
3787 @classmethod
3788 def _make_valid_url(cls):
3789 return r'%s(?P<prefix>|[1-9][0-9]*|all):(?P<query>[\s\S]+)' % cls._SEARCH_KEY
3790
3791 def _real_extract(self, query):
3792 prefix, query = self._match_valid_url(query).group('prefix', 'query')
3793 if prefix == '':
3794 return self._get_n_results(query, 1)
3795 elif prefix == 'all':
3796 return self._get_n_results(query, self._MAX_RESULTS)
3797 else:
3798 n = int(prefix)
3799 if n <= 0:
3800 raise ExtractorError('invalid download number %s for query "%s"' % (n, query))
3801 elif n > self._MAX_RESULTS:
3802 self.report_warning('%s returns max %i results (you requested %i)' % (self._SEARCH_KEY, self._MAX_RESULTS, n))
3803 n = self._MAX_RESULTS
3804 return self._get_n_results(query, n)
3805
3806 def _get_n_results(self, query, n):
3807 """Get a specified number of results for a query.
3808 Either this function or _search_results must be overridden by subclasses """
3809 return self.playlist_result(
3810 itertools.islice(self._search_results(query), 0, None if n == float('inf') else n),
3811 query, query)
3812
3813 def _search_results(self, query):
3814 """Returns an iterator of search results"""
3815 raise NotImplementedError('This method must be implemented by subclasses')
3816
3817 @property
3818 def SEARCH_KEY(self):
3819 return self._SEARCH_KEY