]> jfr.im git - yt-dlp.git/blame - yt_dlp/extractor/common.py
Add option `--use-extractors`
[yt-dlp.git] / yt_dlp / extractor / common.py
CommitLineData
d6983cb4 1import base64
234416e4 2import collections
ac668111 3import getpass
3ec05685 4import hashlib
54007a45 5import http.client
6import http.cookiejar
7import http.cookies
cc16383f 8import itertools
3d3538e4 9import json
f8271158 10import math
4094b6e3 11import netrc
d6983cb4 12import os
773f291d 13import random
6929b41a 14import re
d6983cb4 15import sys
4094b6e3 16import time
8f97a15d 17import types
14f25df2 18import urllib.parse
ac668111 19import urllib.request
f8271158 20import xml.etree.ElementTree
d6983cb4 21
6929b41a 22from ..compat import functools # isort: split
14f25df2 23from ..compat import compat_etree_fromstring, compat_expanduser, compat_os_name
eb8a4433 24from ..downloader import FileDownloader
f8271158 25from ..downloader.f4m import get_base_url, remove_encrypted_media
8c25f81b 26from ..utils import (
8f97a15d 27 IDENTITY,
f8271158 28 JSON_LD_RE,
29 NO_DEFAULT,
30 ExtractorError,
31 GeoRestrictedError,
32 GeoUtils,
b7c47b74 33 LenientJSONDecoder,
f8271158 34 RegexNotFoundError,
be5c1ae8 35 RetryManager,
f8271158 36 UnsupportedError,
05900629 37 age_restricted,
02dc0a36 38 base_url,
08f2a92c 39 bug_reports_message,
82d02080 40 classproperty,
d6983cb4 41 clean_html,
70f0f5a8 42 determine_ext,
46b18f23 43 determine_protocol,
d493f15c 44 dict_get,
42676437 45 encode_data_uri,
9b9c5355 46 error_to_compat_str,
46b18f23 47 extract_attributes,
90137ca4 48 filter_dict,
97f4aecf 49 fix_xml_ampersands,
b14f3a4c 50 float_or_none,
b868936c 51 format_field,
31bb8d3f 52 int_or_none,
34921b43 53 join_nonempty,
a4a554a7 54 js_to_json,
46b18f23 55 mimetype2ext,
3158150c 56 network_exceptions,
46b18f23 57 orderedSet,
d493f15c 58 parse_bitrate,
46b18f23
JH
59 parse_codecs,
60 parse_duration,
4ca2a3cf 61 parse_iso8601,
46b18f23 62 parse_m3u8_attributes,
d493f15c 63 parse_resolution,
46b18f23 64 sanitize_filename,
8f97a15d 65 sanitize_url,
b868936c 66 sanitized_Request,
d493f15c 67 str_or_none,
ce5b9040 68 str_to_int,
f856816b 69 strip_or_none,
5d3a0e79 70 traverse_obj,
47046464 71 try_call,
ffa89477 72 try_get,
f38de77f 73 unescapeHTML,
647eab45 74 unified_strdate,
6b3a3098 75 unified_timestamp,
46b18f23 76 update_Request,
09d02ea4 77 update_url_query,
a107193e 78 url_basename,
bebef109 79 url_or_none,
b868936c 80 urljoin,
6606817a 81 variadic,
a6571f10 82 xpath_element,
8d6765cf
S
83 xpath_text,
84 xpath_with_ns,
d6983cb4 85)
c342041f 86
d6983cb4 87
86e5f3ed 88class InfoExtractor:
d6983cb4
PH
89 """Information Extractor class.
90
91 Information extractors are the classes that, given a URL, extract
92 information about the video (or videos) the URL refers to. This
93 information includes the real video URL, the video title, author and
94 others. The information is stored in a dictionary which is then
5d380852 95 passed to the YoutubeDL. The YoutubeDL processes this
d6983cb4
PH
96 information possibly downloading the video to the file system, among
97 other possible outcomes.
98
cf0649f8 99 The type field determines the type of the result.
fed5d032
PH
100 By far the most common value (and the default if _type is missing) is
101 "video", which indicates a single video.
102
103 For a video, the dictionaries must include the following fields:
d6983cb4
PH
104
105 id: Video identifier.
d4736fdb 106 title: Video title, unescaped. Set to an empty string if video has
107 no title as opposed to "None" which signifies that the
108 extractor failed to obtain a title
d67b0b15 109
f49d89ee 110 Additionally, it must contain either a formats entry or a url one:
d67b0b15 111
f49d89ee
PH
112 formats: A list of dictionaries for each format available, ordered
113 from worst to best quality.
114
115 Potential fields:
c790e93a
S
116 * url The mandatory URL representing the media:
117 for plain file media - HTTP URL of this file,
118 for RTMP - RTMP URL,
119 for HLS - URL of the M3U8 media playlist,
120 for HDS - URL of the F4M manifest,
79d2077e
S
121 for DASH
122 - HTTP URL to plain file media (in case of
123 unfragmented media)
124 - URL of the MPD manifest or base URL
125 representing the media if MPD manifest
8ed7a233 126 is parsed from a string (in case of
79d2077e 127 fragmented media)
c790e93a 128 for MSS - URL of the ISM manifest.
86f4d14f
S
129 * manifest_url
130 The URL of the manifest file in case of
c790e93a
S
131 fragmented media:
132 for HLS - URL of the M3U8 master playlist,
133 for HDS - URL of the F4M manifest,
134 for DASH - URL of the MPD manifest,
135 for MSS - URL of the ISM manifest.
a44ca5a4 136 * manifest_stream_number (For internal use only)
137 The index of the stream in the manifest file
10952eb2 138 * ext Will be calculated from URL if missing
d67b0b15
PH
139 * format A human-readable description of the format
140 ("mp4 container with h264/opus").
141 Calculated from the format_id, width, height.
142 and format_note fields if missing.
143 * format_id A short description of the format
5d4f3985
PH
144 ("mp4_h264_opus" or "19").
145 Technically optional, but strongly recommended.
d67b0b15
PH
146 * format_note Additional info about the format
147 ("3D" or "DASH video")
148 * width Width of the video, if known
149 * height Height of the video, if known
f49d89ee 150 * resolution Textual description of width and height
176f1866 151 * dynamic_range The dynamic range of the video. One of:
152 "SDR" (None), "HDR10", "HDR10+, "HDR12", "HLG, "DV"
7217e148 153 * tbr Average bitrate of audio and video in KBit/s
d67b0b15
PH
154 * abr Average audio bitrate in KBit/s
155 * acodec Name of the audio codec in use
dd27fd17 156 * asr Audio sampling rate in Hertz
b8ed0f15 157 * audio_channels Number of audio channels
d67b0b15 158 * vbr Average video bitrate in KBit/s
fbb21cf5 159 * fps Frame rate
d67b0b15 160 * vcodec Name of the video codec in use
1394ce65 161 * container Name of the container format
d67b0b15 162 * filesize The number of bytes, if known in advance
9732d77e 163 * filesize_approx An estimate for the number of bytes
d67b0b15 164 * player_url SWF Player URL (used for rtmpdump).
c7deaa4c 165 * protocol The protocol that will be used for the actual
adbc4ec4
THD
166 download, lower-case. One of "http", "https" or
167 one of the protocols defined in downloader.PROTOCOL_MAP
c58c2d63
S
168 * fragment_base_url
169 Base URL for fragments. Each fragment's path
170 value (if present) will be relative to
171 this URL.
172 * fragments A list of fragments of a fragmented media.
173 Each fragment entry must contain either an url
174 or a path. If an url is present it should be
175 considered by a client. Otherwise both path and
176 fragment_base_url must be present. Here is
177 the list of all potential fields:
178 * "url" - fragment's URL
179 * "path" - fragment's path relative to
180 fragment_base_url
a0d5077c
S
181 * "duration" (optional, int or float)
182 * "filesize" (optional, int)
adbc4ec4
THD
183 * is_from_start Is a live format that can be downloaded
184 from the start. Boolean
f49d89ee 185 * preference Order number of this format. If this field is
08d13955 186 present and not None, the formats get sorted
38d63d84 187 by this field, regardless of all other values.
f49d89ee
PH
188 -1 for default (order by other properties),
189 -2 or smaller for less than default.
e65566a9
PH
190 < -1000 to hide the format (if there is
191 another one which is strictly better)
32f90364
PH
192 * language Language code, e.g. "de" or "en-US".
193 * language_preference Is this in the language mentioned in
194 the URL?
aff2f4f4
PH
195 10 if it's what the URL is about,
196 -1 for default (don't know),
197 -10 otherwise, other values reserved for now.
5d73273f
PH
198 * quality Order number of the video quality of this
199 format, irrespective of the file format.
200 -1 for default (order by other properties),
201 -2 or smaller for less than default.
c64ed2a3
PH
202 * source_preference Order number for this video source
203 (quality takes higher priority)
204 -1 for default (order by other properties),
205 -2 or smaller for less than default.
d769be6c
PH
206 * http_headers A dictionary of additional HTTP headers
207 to add to the request.
6271f1ca 208 * stretched_ratio If given and not 1, indicates that the
3dee7826
PH
209 video's pixels are not square.
210 width : height ratio as float.
211 * no_resume The server does not support resuming the
212 (HTTP or RTMP) download. Boolean.
88acdbc2 213 * has_drm The format has DRM and cannot be downloaded. Boolean
0a5a191a 214 * downloader_options A dictionary of downloader options
215 (For internal use only)
216 * http_chunk_size Chunk size for HTTP downloads
217 * ffmpeg_args Extra arguments for ffmpeg downloader
3b1fe47d 218 RTMP formats can also have the additional fields: page_url,
219 app, play_path, tc_url, flash_version, rtmp_live, rtmp_conn,
220 rtmp_protocol, rtmp_real_time
3dee7826 221
c0ba0f48 222 url: Final video URL.
d6983cb4 223 ext: Video filename extension.
d67b0b15
PH
224 format: The video format, defaults to ext (used for --get-format)
225 player_url: SWF Player URL (used for rtmpdump).
2f5865cc 226
d6983cb4
PH
227 The following fields are optional:
228
08d30158 229 direct: True if a direct video file was given (must only be set by GenericIE)
f5e43bc6 230 alt_title: A secondary title of the video.
0afef30b
PH
231 display_id An alternative identifier for the video, not necessarily
232 unique, but available before title. Typically, id is
233 something like "4234987", title "Dancing naked mole rats",
234 and display_id "dancing-naked-mole-rats"
d5519808 235 thumbnails: A list of dictionaries, with the following entries:
cfb56d1a 236 * "id" (optional, string) - Thumbnail format ID
d5519808 237 * "url"
cfb56d1a 238 * "preference" (optional, int) - quality of the image
d5519808
PH
239 * "width" (optional, int)
240 * "height" (optional, int)
5e1c39ac 241 * "resolution" (optional, string "{width}x{height}",
d5519808 242 deprecated)
2de624fd 243 * "filesize" (optional, int)
297e9952 244 * "http_headers" (dict) - HTTP headers for the request
d6983cb4 245 thumbnail: Full URL to a video thumbnail image.
f5e43bc6 246 description: Full video description.
d6983cb4 247 uploader: Full name of the video uploader.
2bc0c46f 248 license: License name the video is licensed under.
8a92e51c 249 creator: The creator of the video.
10db0d2f 250 timestamp: UNIX timestamp of the moment the video was uploaded
ae6a1b95 251 upload_date: Video upload date in UTC (YYYYMMDD).
f0d785d3 252 If not explicitly set, calculated from timestamp
253 release_timestamp: UNIX timestamp of the moment the video was released.
254 If it is not clear whether to use timestamp or this, use the former
ae6a1b95 255 release_date: The date (YYYYMMDD) when the video was released in UTC.
f0d785d3 256 If not explicitly set, calculated from release_timestamp
257 modified_timestamp: UNIX timestamp of the moment the video was last modified.
ae6a1b95 258 modified_date: The date (YYYYMMDD) when the video was last modified in UTC.
f0d785d3 259 If not explicitly set, calculated from modified_timestamp
d6983cb4 260 uploader_id: Nickname or id of the video uploader.
7bcd2830 261 uploader_url: Full URL to a personal webpage of the video uploader.
6f1f59f3 262 channel: Full name of the channel the video is uploaded on.
0e7b8d3e 263 Note that channel fields may or may not repeat uploader
6f1f59f3
S
264 fields. This depends on a particular extractor.
265 channel_id: Id of the channel.
266 channel_url: Full URL to a channel webpage.
6c73052c 267 channel_follower_count: Number of followers of the channel.
da9ec3b9 268 location: Physical location where the video was filmed.
a504ced0 269 subtitles: The available subtitles as a dictionary in the format
4606c34e
YCH
270 {tag: subformats}. "tag" is usually a language code, and
271 "subformats" is a list sorted from lower to higher
272 preference, each element is a dictionary with the "ext"
273 entry and one of:
a504ced0 274 * "data": The subtitles file contents
10952eb2 275 * "url": A URL pointing to the subtitles file
2412044c 276 It can optionally also have:
277 * "name": Name or description of the subtitles
08d30158 278 * "http_headers": A dictionary of additional HTTP headers
297e9952 279 to add to the request.
4bba3716 280 "ext" will be calculated from URL if missing
e167860c 281 automatic_captions: Like 'subtitles'; contains automatically generated
282 captions instead of normal subtitles
62d231c0 283 duration: Length of the video in seconds, as an integer or float.
f3d29461 284 view_count: How many users have watched the video on the platform.
19e3dfc9
PH
285 like_count: Number of positive ratings of the video
286 dislike_count: Number of negative ratings of the video
02835c6b 287 repost_count: Number of reposts of the video
2d30521a 288 average_rating: Average rating give by users, the scale used depends on the webpage
19e3dfc9 289 comment_count: Number of comments on the video
dd622d7c
PH
290 comments: A list of comments, each with one or more of the following
291 properties (all but one of text or html optional):
292 * "author" - human-readable name of the comment author
293 * "author_id" - user ID of the comment author
a1c5d2ca 294 * "author_thumbnail" - The thumbnail of the comment author
dd622d7c
PH
295 * "id" - Comment ID
296 * "html" - Comment as HTML
297 * "text" - Plain text of the comment
298 * "timestamp" - UNIX timestamp of comment
299 * "parent" - ID of the comment this one is replying to.
300 Set to "root" to indicate that this is a
301 comment to the original video.
a1c5d2ca
M
302 * "like_count" - Number of positive ratings of the comment
303 * "dislike_count" - Number of negative ratings of the comment
304 * "is_favorited" - Whether the comment is marked as
305 favorite by the video uploader
306 * "author_is_uploader" - Whether the comment is made by
307 the video uploader
8dbe9899 308 age_limit: Age restriction for the video, as an integer (years)
7a5c1cfe 309 webpage_url: The URL to the video webpage, if given to yt-dlp it
9103bbc5
JMF
310 should allow to get the same result again. (It will be set
311 by YoutubeDL if it's missing)
ad3bc6ac
PH
312 categories: A list of categories that the video falls in, for example
313 ["Sports", "Berlin"]
864f24bd 314 tags: A list of tags assigned to the video, e.g. ["sweden", "pop music"]
d0fb4bd1 315 cast: A list of the video cast
7267bd53
PH
316 is_live: True, False, or None (=unknown). Whether this video is a
317 live stream that goes on instead of a fixed-length video.
f76ede8e 318 was_live: True, False, or None (=unknown). Whether this video was
319 originally a live stream.
0647d925 320 live_status: None (=unknown), 'is_live', 'is_upcoming', 'was_live', 'not_live',
e325a21a 321 or 'post_live' (was live, but VOD is not yet processed)
ae30b840 322 If absent, automatically set from is_live, was_live
7c80519c 323 start_time: Time in seconds where the reproduction should start, as
10952eb2 324 specified in the URL.
297a564b 325 end_time: Time in seconds where the reproduction should end, as
10952eb2 326 specified in the URL.
55949fed 327 chapters: A list of dictionaries, with the following entries:
328 * "start_time" - The start time of the chapter in seconds
329 * "end_time" - The end time of the chapter in seconds
330 * "title" (optional, string)
6cfda058 331 playable_in_embed: Whether this video is allowed to play in embedded
332 players on other sites. Can be True (=always allowed),
333 False (=never allowed), None (=unknown), or a string
62b58c09 334 specifying the criteria for embedability; e.g. 'whitelist'
c224251a
M
335 availability: Under what condition the video is available. One of
336 'private', 'premium_only', 'subscriber_only', 'needs_auth',
337 'unlisted' or 'public'. Use 'InfoExtractor._availability'
338 to set it
1e8fe57e 339 _old_archive_ids: A list of old archive ids needed for backward compatibility
277d6ff5 340 __post_extractor: A function to be called just before the metadata is
341 written to either disk, logger or console. The function
342 must return a dict which will be added to the info_dict.
343 This is usefull for additional information that is
344 time-consuming to extract. Note that the fields thus
345 extracted will not be available to output template and
346 match_filter. So, only "comments" and "comment_count" are
347 currently allowed to be extracted via this method.
d6983cb4 348
7109903e
S
349 The following fields should only be used when the video belongs to some logical
350 chapter or section:
351
352 chapter: Name or title of the chapter the video belongs to.
27bfd4e5
S
353 chapter_number: Number of the chapter the video belongs to, as an integer.
354 chapter_id: Id of the chapter the video belongs to, as a unicode string.
7109903e
S
355
356 The following fields should only be used when the video is an episode of some
8d76bdf1 357 series, programme or podcast:
7109903e
S
358
359 series: Title of the series or programme the video episode belongs to.
9ac24e23 360 series_id: Id of the series or programme the video episode belongs to, as a unicode string.
7109903e 361 season: Title of the season the video episode belongs to.
27bfd4e5
S
362 season_number: Number of the season the video episode belongs to, as an integer.
363 season_id: Id of the season the video episode belongs to, as a unicode string.
7109903e
S
364 episode: Title of the video episode. Unlike mandatory video title field,
365 this field should denote the exact title of the video episode
366 without any kind of decoration.
27bfd4e5
S
367 episode_number: Number of the video episode within a season, as an integer.
368 episode_id: Id of the video episode, as a unicode string.
7109903e 369
7a93ab5f
S
370 The following fields should only be used when the media is a track or a part of
371 a music album:
372
373 track: Title of the track.
374 track_number: Number of the track within an album or a disc, as an integer.
375 track_id: Id of the track (useful in case of custom indexing, e.g. 6.iii),
376 as a unicode string.
377 artist: Artist(s) of the track.
378 genre: Genre(s) of the track.
379 album: Title of the album the track belongs to.
380 album_type: Type of the album (e.g. "Demo", "Full-length", "Split", "Compilation", etc).
381 album_artist: List of all artists appeared on the album (e.g.
382 "Ash Borer / Fell Voices" or "Various Artists", useful for splits
383 and compilations).
384 disc_number: Number of the disc or other physical medium the track belongs to,
385 as an integer.
386 release_year: Year (YYYY) when the album was released.
8bcd4048 387 composer: Composer of the piece
7a93ab5f 388
3975b4d2 389 The following fields should only be set for clips that should be cut from the original video:
390
391 section_start: Start time of the section in seconds
392 section_end: End time of the section in seconds
393
45e8a04e 394 The following fields should only be set for storyboards:
395 rows: Number of rows in each storyboard fragment, as an integer
396 columns: Number of columns in each storyboard fragment, as an integer
397
deefc05b 398 Unless mentioned otherwise, the fields should be Unicode strings.
d6983cb4 399
d838b1bd
PH
400 Unless mentioned otherwise, None is equivalent to absence of information.
401
fed5d032
PH
402
403 _type "playlist" indicates multiple videos.
b82f815f
PH
404 There must be a key "entries", which is a list, an iterable, or a PagedList
405 object, each element of which is a valid dictionary by this specification.
fed5d032 406
962ffcf8 407 Additionally, playlists can have "id", "title", and any other relevant
b60419c5 408 attributes with the same semantics as videos (see above).
fed5d032 409
f0d785d3 410 It can also have the following optional fields:
411
412 playlist_count: The total number of videos in a playlist. If not given,
413 YoutubeDL tries to calculate it from "entries"
414
fed5d032
PH
415
416 _type "multi_video" indicates that there are multiple videos that
417 form a single show, for examples multiple acts of an opera or TV episode.
418 It must have an entries key like a playlist and contain all the keys
419 required for a video at the same time.
420
421
422 _type "url" indicates that the video must be extracted from another
423 location, possibly by a different extractor. Its only required key is:
424 "url" - the next URL to extract.
f58766ce
PH
425 The key "ie_key" can be set to the class name (minus the trailing "IE",
426 e.g. "Youtube") if the extractor class is known in advance.
427 Additionally, the dictionary may have any properties of the resolved entity
428 known in advance, for example "title" if the title of the referred video is
fed5d032
PH
429 known ahead of time.
430
431
432 _type "url_transparent" entities have the same specification as "url", but
433 indicate that the given additional information is more precise than the one
434 associated with the resolved URL.
435 This is useful when a site employs a video service that hosts the video and
436 its technical metadata, but that video service does not embed a useful
437 title, description etc.
438
439
8f97a15d 440 Subclasses of this should also be added to the list of extractors and
441 should define a _VALID_URL regexp and, re-define the _real_extract() and
442 (optionally) _real_initialize() methods.
d6983cb4 443
e6f21b3d 444 Subclasses may also override suitable() if necessary, but ensure the function
445 signature is preserved and that this function imports everything it needs
52efa4b3 446 (except other extractors), so that lazy_extractors works correctly.
447
8f97a15d 448 Subclasses can define a list of _EMBED_REGEX, which will be searched for in
449 the HTML of Generic webpages. It may also override _extract_embed_urls
450 or _extract_from_webpage as necessary. While these are normally classmethods,
451 _extract_from_webpage is allowed to be an instance method.
452
453 _extract_from_webpage may raise self.StopExtraction() to stop further
454 processing of the webpage and obtain exclusive rights to it. This is useful
62b58c09
L
455 when the extractor cannot reliably be matched using just the URL,
456 e.g. invidious/peertube instances
8f97a15d 457
458 Embed-only extractors can be defined by setting _VALID_URL = False.
459
52efa4b3 460 To support username + password (or netrc) login, the extractor must define a
461 _NETRC_MACHINE and re-define _perform_login(username, password) and
462 (optionally) _initialize_pre_login() methods. The _perform_login method will
463 be called between _initialize_pre_login and _real_initialize if credentials
464 are passed by the user. In cases where it is necessary to have the login
465 process as part of the extraction rather than initialization, _perform_login
466 can be left undefined.
e6f21b3d 467
4248dad9 468 _GEO_BYPASS attribute may be set to False in order to disable
773f291d
S
469 geo restriction bypass mechanisms for a particular extractor.
470 Though it won't disable explicit geo restriction bypass based on
504f20dd 471 country code provided with geo_bypass_country.
4248dad9
S
472
473 _GEO_COUNTRIES attribute may contain a list of presumably geo unrestricted
474 countries for this extractor. One of these countries will be used by
475 geo restriction bypass mechanism right away in order to bypass
504f20dd 476 geo restriction, of course, if the mechanism is not disabled.
773f291d 477
5f95927a
S
478 _GEO_IP_BLOCKS attribute may contain a list of presumably geo unrestricted
479 IP blocks in CIDR notation for this extractor. One of these IP blocks
480 will be used by geo restriction bypass mechanism similarly
504f20dd 481 to _GEO_COUNTRIES.
3ccdde8c 482
fe7866d0 483 The _ENABLED attribute should be set to False for IEs that
484 are disabled by default and must be explicitly enabled.
485
e6f21b3d 486 The _WORKING attribute should be set to False for broken IEs
d6983cb4
PH
487 in order to warn the users and skip the tests.
488 """
489
490 _ready = False
491 _downloader = None
773f291d 492 _x_forwarded_for_ip = None
4248dad9
S
493 _GEO_BYPASS = True
494 _GEO_COUNTRIES = None
5f95927a 495 _GEO_IP_BLOCKS = None
d6983cb4 496 _WORKING = True
fe7866d0 497 _ENABLED = True
52efa4b3 498 _NETRC_MACHINE = None
231025c4 499 IE_DESC = None
8dcce6a8 500 SEARCH_KEY = None
8f97a15d 501 _VALID_URL = None
502 _EMBED_REGEX = []
d6983cb4 503
8dcce6a8 504 def _login_hint(self, method=NO_DEFAULT, netrc=None):
505 password_hint = f'--username and --password, or --netrc ({netrc or self._NETRC_MACHINE}) to provide account credentials'
506 return {
507 None: '',
508 'any': f'Use --cookies, --cookies-from-browser, {password_hint}',
509 'password': f'Use {password_hint}',
510 'cookies': (
511 'Use --cookies-from-browser or --cookies for the authentication. '
512 'See https://github.com/ytdl-org/youtube-dl#how-do-i-pass-cookies-to-youtube-dl for how to manually pass cookies'),
513 }[method if method is not NO_DEFAULT else 'any' if self.supports_login() else 'cookies']
9d5d4d64 514
d6983cb4 515 def __init__(self, downloader=None):
49a57e70 516 """Constructor. Receives an optional downloader (a YoutubeDL instance).
517 If a downloader is not passed during initialization,
518 it must be set using "set_downloader()" before "extract()" is called"""
d6983cb4 519 self._ready = False
773f291d 520 self._x_forwarded_for_ip = None
28f436ba 521 self._printed_messages = set()
d6983cb4
PH
522 self.set_downloader(downloader)
523
524 @classmethod
5ad28e7f 525 def _match_valid_url(cls, url):
8f97a15d 526 if cls._VALID_URL is False:
527 return None
79cb2577
PH
528 # This does not use has/getattr intentionally - we want to know whether
529 # we have cached the regexp for *this* class, whereas getattr would also
530 # match the superclass
531 if '_VALID_URL_RE' not in cls.__dict__:
532 cls._VALID_URL_RE = re.compile(cls._VALID_URL)
5ad28e7f 533 return cls._VALID_URL_RE.match(url)
534
535 @classmethod
536 def suitable(cls, url):
537 """Receives a URL and returns True if suitable for this IE."""
3fb4e21b 538 # This function must import everything it needs (except other extractors),
539 # so that lazy_extractors works correctly
5ad28e7f 540 return cls._match_valid_url(url) is not None
d6983cb4 541
ed9266db
PH
542 @classmethod
543 def _match_id(cls, url):
5ad28e7f 544 return cls._match_valid_url(url).group('id')
ed9266db 545
1151c407 546 @classmethod
547 def get_temp_id(cls, url):
548 try:
549 return cls._match_id(url)
550 except (IndexError, AttributeError):
551 return None
552
d6983cb4
PH
553 @classmethod
554 def working(cls):
555 """Getter method for _WORKING."""
556 return cls._WORKING
557
52efa4b3 558 @classmethod
559 def supports_login(cls):
560 return bool(cls._NETRC_MACHINE)
561
d6983cb4
PH
562 def initialize(self):
563 """Initializes an instance (authentication, etc)."""
28f436ba 564 self._printed_messages = set()
5f95927a
S
565 self._initialize_geo_bypass({
566 'countries': self._GEO_COUNTRIES,
567 'ip_blocks': self._GEO_IP_BLOCKS,
568 })
4248dad9 569 if not self._ready:
52efa4b3 570 self._initialize_pre_login()
571 if self.supports_login():
572 username, password = self._get_login_info()
573 if username:
574 self._perform_login(username, password)
575 elif self.get_param('username') and False not in (self.IE_DESC, self._NETRC_MACHINE):
8dcce6a8 576 self.report_warning(f'Login with password is not supported for this website. {self._login_hint("cookies")}')
4248dad9
S
577 self._real_initialize()
578 self._ready = True
579
5f95927a 580 def _initialize_geo_bypass(self, geo_bypass_context):
e39b5d4a
S
581 """
582 Initialize geo restriction bypass mechanism.
583
584 This method is used to initialize geo bypass mechanism based on faking
585 X-Forwarded-For HTTP header. A random country from provided country list
dc0a869e 586 is selected and a random IP belonging to this country is generated. This
e39b5d4a
S
587 IP will be passed as X-Forwarded-For HTTP header in all subsequent
588 HTTP requests.
e39b5d4a
S
589
590 This method will be used for initial geo bypass mechanism initialization
5f95927a
S
591 during the instance initialization with _GEO_COUNTRIES and
592 _GEO_IP_BLOCKS.
e39b5d4a 593
5f95927a 594 You may also manually call it from extractor's code if geo bypass
e39b5d4a 595 information is not available beforehand (e.g. obtained during
5f95927a
S
596 extraction) or due to some other reason. In this case you should pass
597 this information in geo bypass context passed as first argument. It may
598 contain following fields:
599
600 countries: List of geo unrestricted countries (similar
601 to _GEO_COUNTRIES)
602 ip_blocks: List of geo unrestricted IP blocks in CIDR notation
603 (similar to _GEO_IP_BLOCKS)
604
e39b5d4a 605 """
773f291d 606 if not self._x_forwarded_for_ip:
5f95927a
S
607
608 # Geo bypass mechanism is explicitly disabled by user
a06916d9 609 if not self.get_param('geo_bypass', True):
5f95927a
S
610 return
611
612 if not geo_bypass_context:
613 geo_bypass_context = {}
614
615 # Backward compatibility: previously _initialize_geo_bypass
616 # expected a list of countries, some 3rd party code may still use
617 # it this way
618 if isinstance(geo_bypass_context, (list, tuple)):
619 geo_bypass_context = {
620 'countries': geo_bypass_context,
621 }
622
623 # The whole point of geo bypass mechanism is to fake IP
624 # as X-Forwarded-For HTTP header based on some IP block or
625 # country code.
626
627 # Path 1: bypassing based on IP block in CIDR notation
628
629 # Explicit IP block specified by user, use it right away
630 # regardless of whether extractor is geo bypassable or not
a06916d9 631 ip_block = self.get_param('geo_bypass_ip_block', None)
5f95927a
S
632
633 # Otherwise use random IP block from geo bypass context but only
634 # if extractor is known as geo bypassable
635 if not ip_block:
636 ip_blocks = geo_bypass_context.get('ip_blocks')
637 if self._GEO_BYPASS and ip_blocks:
638 ip_block = random.choice(ip_blocks)
639
640 if ip_block:
641 self._x_forwarded_for_ip = GeoUtils.random_ipv4(ip_block)
8a82af35 642 self.write_debug(f'Using fake IP {self._x_forwarded_for_ip} as X-Forwarded-For')
5f95927a
S
643 return
644
645 # Path 2: bypassing based on country code
646
647 # Explicit country code specified by user, use it right away
648 # regardless of whether extractor is geo bypassable or not
a06916d9 649 country = self.get_param('geo_bypass_country', None)
5f95927a
S
650
651 # Otherwise use random country code from geo bypass context but
652 # only if extractor is known as geo bypassable
653 if not country:
654 countries = geo_bypass_context.get('countries')
655 if self._GEO_BYPASS and countries:
656 country = random.choice(countries)
657
658 if country:
659 self._x_forwarded_for_ip = GeoUtils.random_ipv4(country)
0760b0a7 660 self._downloader.write_debug(
86e5f3ed 661 f'Using fake IP {self._x_forwarded_for_ip} ({country.upper()}) as X-Forwarded-For')
d6983cb4
PH
662
663 def extract(self, url):
664 """Extracts URL information and returns it in list of dicts."""
3a5bcd03 665 try:
773f291d
S
666 for _ in range(2):
667 try:
668 self.initialize()
a06916d9 669 self.write_debug('Extracting URL: %s' % url)
0016b84e 670 ie_result = self._real_extract(url)
07cce701 671 if ie_result is None:
672 return None
0016b84e
S
673 if self._x_forwarded_for_ip:
674 ie_result['__x_forwarded_for_ip'] = self._x_forwarded_for_ip
b79f9e30 675 subtitles = ie_result.get('subtitles') or {}
676 if 'no-live-chat' in self.get_param('compat_opts'):
677 for lang in ('live_chat', 'comments', 'danmaku'):
678 subtitles.pop(lang, None)
0016b84e 679 return ie_result
773f291d 680 except GeoRestrictedError as e:
4248dad9
S
681 if self.__maybe_fake_ip_and_retry(e.countries):
682 continue
773f291d 683 raise
0db3bae8 684 except UnsupportedError:
685 raise
1151c407 686 except ExtractorError as e:
0db3bae8 687 kwargs = {
688 'video_id': e.video_id or self.get_temp_id(url),
689 'ie': self.IE_NAME,
b69fd25c 690 'tb': e.traceback or sys.exc_info()[2],
0db3bae8 691 'expected': e.expected,
692 'cause': e.cause
693 }
694 if hasattr(e, 'countries'):
695 kwargs['countries'] = e.countries
7265a219 696 raise type(e)(e.orig_msg, **kwargs)
ac668111 697 except http.client.IncompleteRead as e:
1151c407 698 raise ExtractorError('A network error has occurred.', cause=e, expected=True, video_id=self.get_temp_id(url))
9650885b 699 except (KeyError, StopIteration) as e:
1151c407 700 raise ExtractorError('An extractor error has occurred.', cause=e, video_id=self.get_temp_id(url))
d6983cb4 701
4248dad9 702 def __maybe_fake_ip_and_retry(self, countries):
a06916d9 703 if (not self.get_param('geo_bypass_country', None)
3089bc74 704 and self._GEO_BYPASS
a06916d9 705 and self.get_param('geo_bypass', True)
3089bc74
S
706 and not self._x_forwarded_for_ip
707 and countries):
eea0716c
S
708 country_code = random.choice(countries)
709 self._x_forwarded_for_ip = GeoUtils.random_ipv4(country_code)
4248dad9
S
710 if self._x_forwarded_for_ip:
711 self.report_warning(
eea0716c
S
712 'Video is geo restricted. Retrying extraction with fake IP %s (%s) as X-Forwarded-For.'
713 % (self._x_forwarded_for_ip, country_code.upper()))
4248dad9
S
714 return True
715 return False
716
d6983cb4 717 def set_downloader(self, downloader):
08d30158 718 """Sets a YoutubeDL instance as the downloader for this IE."""
d6983cb4
PH
719 self._downloader = downloader
720
9809740b 721 @property
722 def cache(self):
723 return self._downloader.cache
724
725 @property
726 def cookiejar(self):
727 return self._downloader.cookiejar
728
52efa4b3 729 def _initialize_pre_login(self):
962ffcf8 730 """ Initialization before login. Redefine in subclasses."""
52efa4b3 731 pass
732
733 def _perform_login(self, username, password):
734 """ Login with username and password. Redefine in subclasses."""
735 pass
736
d6983cb4
PH
737 def _real_initialize(self):
738 """Real initialization process. Redefine in subclasses."""
739 pass
740
741 def _real_extract(self, url):
742 """Real extraction process. Redefine in subclasses."""
08d30158 743 raise NotImplementedError('This method must be implemented by subclasses')
d6983cb4 744
56c73665
JMF
745 @classmethod
746 def ie_key(cls):
747 """A string for getting the InfoExtractor with get_info_extractor"""
3fb4e21b 748 return cls.__name__[:-2]
56c73665 749
82d02080 750 @classproperty
751 def IE_NAME(cls):
752 return cls.__name__[:-2]
d6983cb4 753
d391b7e2
S
754 @staticmethod
755 def __can_accept_status_code(err, expected_status):
ac668111 756 assert isinstance(err, urllib.error.HTTPError)
d391b7e2
S
757 if expected_status is None:
758 return False
d391b7e2
S
759 elif callable(expected_status):
760 return expected_status(err.code) is True
761 else:
6606817a 762 return err.code in variadic(expected_status)
d391b7e2 763
c043c246 764 def _create_request(self, url_or_request, data=None, headers=None, query=None):
ac668111 765 if isinstance(url_or_request, urllib.request.Request):
09d02ea4 766 return update_Request(url_or_request, data=data, headers=headers, query=query)
767 if query:
768 url_or_request = update_url_query(url_or_request, query)
c043c246 769 return sanitized_Request(url_or_request, data, headers or {})
f95b9dee 770
c043c246 771 def _request_webpage(self, url_or_request, video_id, note=None, errnote=None, fatal=True, data=None, headers=None, query=None, expected_status=None):
d391b7e2
S
772 """
773 Return the response handle.
774
775 See _download_webpage docstring for arguments specification.
776 """
1cf376f5 777 if not self._downloader._first_webpage_request:
49a57e70 778 sleep_interval = self.get_param('sleep_interval_requests') or 0
1cf376f5 779 if sleep_interval > 0:
5ef7d9bd 780 self.to_screen('Sleeping %s seconds ...' % sleep_interval)
1cf376f5 781 time.sleep(sleep_interval)
782 else:
783 self._downloader._first_webpage_request = False
784
d6983cb4
PH
785 if note is None:
786 self.report_download_webpage(video_id)
787 elif note is not False:
7cc3570e 788 if video_id is None:
86e5f3ed 789 self.to_screen(str(note))
7cc3570e 790 else:
86e5f3ed 791 self.to_screen(f'{video_id}: {note}')
2132edaa
S
792
793 # Some sites check X-Forwarded-For HTTP header in order to figure out
794 # the origin of the client behind proxy. This allows bypassing geo
795 # restriction by faking this header's value to IP that belongs to some
796 # geo unrestricted country. We will do so once we encounter any
797 # geo restriction error.
798 if self._x_forwarded_for_ip:
c043c246 799 headers = (headers or {}).copy()
800 headers.setdefault('X-Forwarded-For', self._x_forwarded_for_ip)
2132edaa 801
d6983cb4 802 try:
f95b9dee 803 return self._downloader.urlopen(self._create_request(url_or_request, data, headers, query))
3158150c 804 except network_exceptions as err:
ac668111 805 if isinstance(err, urllib.error.HTTPError):
d391b7e2 806 if self.__can_accept_status_code(err, expected_status):
95e42d73
XDG
807 # Retain reference to error to prevent file object from
808 # being closed before it can be read. Works around the
809 # effects of <https://bugs.python.org/issue15002>
810 # introduced in Python 3.4.1.
811 err.fp._error = err
d391b7e2
S
812 return err.fp
813
aa94a6d3
PH
814 if errnote is False:
815 return False
d6983cb4 816 if errnote is None:
f1a9d64e 817 errnote = 'Unable to download webpage'
7f8b2714 818
86e5f3ed 819 errmsg = f'{errnote}: {error_to_compat_str(err)}'
7cc3570e 820 if fatal:
497d2fab 821 raise ExtractorError(errmsg, cause=err)
7cc3570e 822 else:
6a39ee13 823 self.report_warning(errmsg)
7cc3570e 824 return False
d6983cb4 825
1890fc63 826 def _download_webpage_handle(self, url_or_request, video_id, note=None, errnote=None, fatal=True,
827 encoding=None, data=None, headers={}, query={}, expected_status=None):
d391b7e2
S
828 """
829 Return a tuple (page content as string, URL handle).
830
617f658b 831 Arguments:
832 url_or_request -- plain text URL as a string or
ac668111 833 a urllib.request.Request object
617f658b 834 video_id -- Video/playlist/item identifier (string)
835
836 Keyword arguments:
837 note -- note printed before downloading (string)
838 errnote -- note printed in case of an error (string)
839 fatal -- flag denoting whether error should be considered fatal,
840 i.e. whether it should cause ExtractionError to be raised,
841 otherwise a warning will be reported and extraction continued
842 encoding -- encoding for a page content decoding, guessed automatically
843 when not explicitly specified
844 data -- POST data (bytes)
845 headers -- HTTP headers (dict)
846 query -- URL query (dict)
847 expected_status -- allows to accept failed HTTP requests (non 2xx
848 status code) by explicitly specifying a set of accepted status
849 codes. Can be any of the following entities:
850 - an integer type specifying an exact failed status code to
851 accept
852 - a list or a tuple of integer types specifying a list of
853 failed status codes to accept
854 - a callable accepting an actual failed status code and
855 returning True if it should be accepted
856 Note that this argument does not affect success status codes (2xx)
857 which are always accepted.
d391b7e2 858 """
617f658b 859
b9d3e163 860 # Strip hashes from the URL (#1038)
14f25df2 861 if isinstance(url_or_request, str):
b9d3e163
PH
862 url_or_request = url_or_request.partition('#')[0]
863
d391b7e2 864 urlh = self._request_webpage(url_or_request, video_id, note, errnote, fatal, data=data, headers=headers, query=query, expected_status=expected_status)
7cc3570e
PH
865 if urlh is False:
866 assert not fatal
867 return False
c9a77969 868 content = self._webpage_read_content(urlh, url_or_request, video_id, note, errnote, fatal, encoding=encoding)
23be51d8
PH
869 return (content, urlh)
870
c9a77969
YCH
871 @staticmethod
872 def _guess_encoding_from_content(content_type, webpage_bytes):
d6983cb4
PH
873 m = re.match(r'[a-zA-Z0-9_.-]+/[a-zA-Z0-9_.-]+\s*;\s*charset=(.+)', content_type)
874 if m:
875 encoding = m.group(1)
876 else:
0d75ae2c 877 m = re.search(br'<meta[^>]+charset=[\'"]?([^\'")]+)[ /\'">]',
f143d86a
PH
878 webpage_bytes[:1024])
879 if m:
880 encoding = m.group(1).decode('ascii')
b60016e8
PH
881 elif webpage_bytes.startswith(b'\xff\xfe'):
882 encoding = 'utf-16'
f143d86a
PH
883 else:
884 encoding = 'utf-8'
c9a77969
YCH
885
886 return encoding
887
4457823d
S
888 def __check_blocked(self, content):
889 first_block = content[:512]
3089bc74
S
890 if ('<title>Access to this site is blocked</title>' in content
891 and 'Websense' in first_block):
4457823d
S
892 msg = 'Access to this webpage has been blocked by Websense filtering software in your network.'
893 blocked_iframe = self._html_search_regex(
894 r'<iframe src="([^"]+)"', content,
895 'Websense information URL', default=None)
896 if blocked_iframe:
897 msg += ' Visit %s for more details' % blocked_iframe
898 raise ExtractorError(msg, expected=True)
899 if '<title>The URL you requested has been blocked</title>' in first_block:
900 msg = (
901 'Access to this webpage has been blocked by Indian censorship. '
902 'Use a VPN or proxy server (with --proxy) to route around it.')
903 block_msg = self._html_search_regex(
904 r'</h1><p>(.*?)</p>',
905 content, 'block message', default=None)
906 if block_msg:
907 msg += ' (Message: "%s")' % block_msg.replace('\n', ' ')
908 raise ExtractorError(msg, expected=True)
3089bc74
S
909 if ('<title>TTK :: Доступ к ресурсу ограничен</title>' in content
910 and 'blocklist.rkn.gov.ru' in content):
4457823d
S
911 raise ExtractorError(
912 'Access to this webpage has been blocked by decision of the Russian government. '
913 'Visit http://blocklist.rkn.gov.ru/ for a block reason.',
914 expected=True)
915
f95b9dee 916 def _request_dump_filename(self, url, video_id):
917 basen = f'{video_id}_{url}'
918 trim_length = self.get_param('trim_file_name') or 240
919 if len(basen) > trim_length:
920 h = '___' + hashlib.md5(basen.encode('utf-8')).hexdigest()
921 basen = basen[:trim_length - len(h)] + h
922 filename = sanitize_filename(f'{basen}.dump', restricted=True)
923 # Working around MAX_PATH limitation on Windows (see
924 # http://msdn.microsoft.com/en-us/library/windows/desktop/aa365247(v=vs.85).aspx)
925 if compat_os_name == 'nt':
926 absfilepath = os.path.abspath(filename)
927 if len(absfilepath) > 259:
928 filename = fR'\\?\{absfilepath}'
929 return filename
930
931 def __decode_webpage(self, webpage_bytes, encoding, headers):
932 if not encoding:
933 encoding = self._guess_encoding_from_content(headers.get('Content-Type', ''), webpage_bytes)
934 try:
935 return webpage_bytes.decode(encoding, 'replace')
936 except LookupError:
937 return webpage_bytes.decode('utf-8', 'replace')
938
c9a77969 939 def _webpage_read_content(self, urlh, url_or_request, video_id, note=None, errnote=None, fatal=True, prefix=None, encoding=None):
c9a77969
YCH
940 webpage_bytes = urlh.read()
941 if prefix is not None:
942 webpage_bytes = prefix + webpage_bytes
a06916d9 943 if self.get_param('dump_intermediate_pages', False):
f610dbb0 944 self.to_screen('Dumping request to ' + urlh.geturl())
d6983cb4
PH
945 dump = base64.b64encode(webpage_bytes).decode('ascii')
946 self._downloader.to_screen(dump)
f95b9dee 947 if self.get_param('write_pages'):
e121e3ce 948 filename = self._request_dump_filename(urlh.geturl(), video_id)
f95b9dee 949 self.to_screen(f'Saving request to {filename}')
d41e6efc
PH
950 with open(filename, 'wb') as outf:
951 outf.write(webpage_bytes)
952
f95b9dee 953 content = self.__decode_webpage(webpage_bytes, encoding, urlh.headers)
4457823d 954 self.__check_blocked(content)
2410c43d 955
23be51d8 956 return content
d6983cb4 957
6edf2808 958 def __print_error(self, errnote, fatal, video_id, err):
959 if fatal:
c6e07cf1 960 raise ExtractorError(f'{video_id}: {errnote}', cause=err)
6edf2808 961 elif errnote:
c6e07cf1 962 self.report_warning(f'{video_id}: {errnote}: {err}')
6edf2808 963
964 def _parse_xml(self, xml_string, video_id, transform_source=None, fatal=True, errnote=None):
e2b38da9
PH
965 if transform_source:
966 xml_string = transform_source(xml_string)
e01c3d2e
S
967 try:
968 return compat_etree_fromstring(xml_string.encode('utf-8'))
f9934b96 969 except xml.etree.ElementTree.ParseError as ve:
6edf2808 970 self.__print_error('Failed to parse XML' if errnote is None else errnote, fatal, video_id, ve)
267ed0c5 971
6edf2808 972 def _parse_json(self, json_string, video_id, transform_source=None, fatal=True, errnote=None, **parser_kwargs):
3d3538e4 973 try:
b7c47b74 974 return json.loads(
975 json_string, cls=LenientJSONDecoder, strict=False, transform_source=transform_source, **parser_kwargs)
3d3538e4 976 except ValueError as ve:
6edf2808 977 self.__print_error('Failed to parse JSON' if errnote is None else errnote, fatal, video_id, ve)
3d3538e4 978
6edf2808 979 def _parse_socket_response_as_json(self, data, *args, **kwargs):
980 return self._parse_json(data[data.find('{'):data.rfind('}') + 1], *args, **kwargs)
adddc50c 981
617f658b 982 def __create_download_methods(name, parser, note, errnote, return_value):
983
6edf2808 984 def parse(ie, content, *args, errnote=errnote, **kwargs):
617f658b 985 if parser is None:
986 return content
6edf2808 987 if errnote is False:
988 kwargs['errnote'] = errnote
617f658b 989 # parser is fetched by name so subclasses can override it
990 return getattr(ie, parser)(content, *args, **kwargs)
991
c4910024 992 def download_handle(self, url_or_request, video_id, note=note, errnote=errnote, transform_source=None,
993 fatal=True, encoding=None, data=None, headers={}, query={}, expected_status=None):
994 res = self._download_webpage_handle(
995 url_or_request, video_id, note=note, errnote=errnote, fatal=fatal, encoding=encoding,
996 data=data, headers=headers, query=query, expected_status=expected_status)
617f658b 997 if res is False:
998 return res
999 content, urlh = res
6edf2808 1000 return parse(self, content, video_id, transform_source=transform_source, fatal=fatal, errnote=errnote), urlh
617f658b 1001
f95b9dee 1002 def download_content(self, url_or_request, video_id, note=note, errnote=errnote, transform_source=None,
c4910024 1003 fatal=True, encoding=None, data=None, headers={}, query={}, expected_status=None):
f95b9dee 1004 if self.get_param('load_pages'):
1005 url_or_request = self._create_request(url_or_request, data, headers, query)
1006 filename = self._request_dump_filename(url_or_request.full_url, video_id)
1007 self.to_screen(f'Loading request from {filename}')
1008 try:
1009 with open(filename, 'rb') as dumpf:
1010 webpage_bytes = dumpf.read()
1011 except OSError as e:
1012 self.report_warning(f'Unable to load request from disk: {e}')
1013 else:
1014 content = self.__decode_webpage(webpage_bytes, encoding, url_or_request.headers)
6edf2808 1015 return parse(self, content, video_id, transform_source=transform_source, fatal=fatal, errnote=errnote)
c4910024 1016 kwargs = {
1017 'note': note,
1018 'errnote': errnote,
1019 'transform_source': transform_source,
1020 'fatal': fatal,
1021 'encoding': encoding,
1022 'data': data,
1023 'headers': headers,
1024 'query': query,
1025 'expected_status': expected_status,
1026 }
617f658b 1027 if parser is None:
c4910024 1028 kwargs.pop('transform_source')
617f658b 1029 # The method is fetched by name so subclasses can override _download_..._handle
c4910024 1030 res = getattr(self, download_handle.__name__)(url_or_request, video_id, **kwargs)
617f658b 1031 return res if res is False else res[0]
1032
1033 def impersonate(func, name, return_value):
1034 func.__name__, func.__qualname__ = name, f'InfoExtractor.{name}'
1035 func.__doc__ = f'''
1036 @param transform_source Apply this transformation before parsing
1037 @returns {return_value}
1038
1039 See _download_webpage_handle docstring for other arguments specification
1040 '''
1041
1042 impersonate(download_handle, f'_download_{name}_handle', f'({return_value}, URL handle)')
1043 impersonate(download_content, f'_download_{name}', f'{return_value}')
1044 return download_handle, download_content
1045
1046 _download_xml_handle, _download_xml = __create_download_methods(
1047 'xml', '_parse_xml', 'Downloading XML', 'Unable to download XML', 'xml as an xml.etree.ElementTree.Element')
1048 _download_json_handle, _download_json = __create_download_methods(
1049 'json', '_parse_json', 'Downloading JSON metadata', 'Unable to download JSON metadata', 'JSON object as a dict')
1050 _download_socket_json_handle, _download_socket_json = __create_download_methods(
1051 'socket_json', '_parse_socket_response_as_json', 'Polling socket', 'Unable to poll socket', 'JSON object as a dict')
1052 __download_webpage = __create_download_methods('webpage', None, None, None, 'data of the page as a string')[1]
adddc50c 1053
617f658b 1054 def _download_webpage(
1055 self, url_or_request, video_id, note=None, errnote=None,
1056 fatal=True, tries=1, timeout=NO_DEFAULT, *args, **kwargs):
adddc50c 1057 """
617f658b 1058 Return the data of the page as a string.
adddc50c 1059
617f658b 1060 Keyword arguments:
1061 tries -- number of tries
1062 timeout -- sleep interval between tries
1063
1064 See _download_webpage_handle docstring for other arguments specification.
adddc50c 1065 """
617f658b 1066
1067 R''' # NB: These are unused; should they be deprecated?
1068 if tries != 1:
1069 self._downloader.deprecation_warning('tries argument is deprecated in InfoExtractor._download_webpage')
1070 if timeout is NO_DEFAULT:
1071 timeout = 5
1072 else:
1073 self._downloader.deprecation_warning('timeout argument is deprecated in InfoExtractor._download_webpage')
1074 '''
1075
1076 try_count = 0
1077 while True:
1078 try:
1079 return self.__download_webpage(url_or_request, video_id, note, errnote, None, fatal, *args, **kwargs)
ac668111 1080 except http.client.IncompleteRead as e:
617f658b 1081 try_count += 1
1082 if try_count >= tries:
1083 raise e
1084 self._sleep(timeout, video_id)
adddc50c 1085
28f436ba 1086 def report_warning(self, msg, video_id=None, *args, only_once=False, **kwargs):
a70635b8 1087 idstr = format_field(video_id, None, '%s: ')
28f436ba 1088 msg = f'[{self.IE_NAME}] {idstr}{msg}'
1089 if only_once:
1090 if f'WARNING: {msg}' in self._printed_messages:
1091 return
1092 self._printed_messages.add(f'WARNING: {msg}')
1093 self._downloader.report_warning(msg, *args, **kwargs)
f45f96f8 1094
a06916d9 1095 def to_screen(self, msg, *args, **kwargs):
d6983cb4 1096 """Print msg to screen, prefixing it with '[ie_name]'"""
86e5f3ed 1097 self._downloader.to_screen(f'[{self.IE_NAME}] {msg}', *args, **kwargs)
a06916d9 1098
1099 def write_debug(self, msg, *args, **kwargs):
86e5f3ed 1100 self._downloader.write_debug(f'[{self.IE_NAME}] {msg}', *args, **kwargs)
a06916d9 1101
1102 def get_param(self, name, default=None, *args, **kwargs):
1103 if self._downloader:
1104 return self._downloader.params.get(name, default, *args, **kwargs)
1105 return default
d6983cb4 1106
88acdbc2 1107 def report_drm(self, video_id, partial=False):
1108 self.raise_no_formats('This video is DRM protected', expected=True, video_id=video_id)
1109
d6983cb4
PH
1110 def report_extraction(self, id_or_name):
1111 """Report information extraction."""
f1a9d64e 1112 self.to_screen('%s: Extracting information' % id_or_name)
d6983cb4
PH
1113
1114 def report_download_webpage(self, video_id):
1115 """Report webpage download."""
f1a9d64e 1116 self.to_screen('%s: Downloading webpage' % video_id)
d6983cb4
PH
1117
1118 def report_age_confirmation(self):
1119 """Report attempt to confirm age."""
f1a9d64e 1120 self.to_screen('Confirming age')
d6983cb4 1121
fc79158d
JMF
1122 def report_login(self):
1123 """Report attempt to log in."""
f1a9d64e 1124 self.to_screen('Logging in')
fc79158d 1125
b7da73eb 1126 def raise_login_required(
9d5d4d64 1127 self, msg='This video is only available for registered users',
52efa4b3 1128 metadata_available=False, method=NO_DEFAULT):
f2ebc5c7 1129 if metadata_available and (
1130 self.get_param('ignore_no_formats_error') or self.get_param('wait_for_video')):
b7da73eb 1131 self.report_warning(msg)
7265a219 1132 return
a70635b8 1133 msg += format_field(self._login_hint(method), None, '. %s')
46890374 1134 raise ExtractorError(msg, expected=True)
43e7d3c9 1135
b7da73eb 1136 def raise_geo_restricted(
1137 self, msg='This video is not available from your location due to geo restriction',
1138 countries=None, metadata_available=False):
f2ebc5c7 1139 if metadata_available and (
1140 self.get_param('ignore_no_formats_error') or self.get_param('wait_for_video')):
b7da73eb 1141 self.report_warning(msg)
1142 else:
1143 raise GeoRestrictedError(msg, countries=countries)
1144
1145 def raise_no_formats(self, msg, expected=False, video_id=None):
f2ebc5c7 1146 if expected and (
1147 self.get_param('ignore_no_formats_error') or self.get_param('wait_for_video')):
b7da73eb 1148 self.report_warning(msg, video_id)
68f5867c
L
1149 elif isinstance(msg, ExtractorError):
1150 raise msg
b7da73eb 1151 else:
1152 raise ExtractorError(msg, expected=expected, video_id=video_id)
c430802e 1153
5f6a1245 1154 # Methods for following #608
c0d0b01f 1155 @staticmethod
311b6615 1156 def url_result(url, ie=None, video_id=None, video_title=None, *, url_transparent=False, **kwargs):
10952eb2 1157 """Returns a URL that points to a page that should be processed"""
311b6615 1158 if ie is not None:
1159 kwargs['ie_key'] = ie if isinstance(ie, str) else ie.ie_key()
7012b23c 1160 if video_id is not None:
311b6615 1161 kwargs['id'] = video_id
830d53bf 1162 if video_title is not None:
311b6615 1163 kwargs['title'] = video_title
1164 return {
1165 **kwargs,
1166 '_type': 'url_transparent' if url_transparent else 'url',
1167 'url': url,
1168 }
1169
8f97a15d 1170 @classmethod
1171 def playlist_from_matches(cls, matches, playlist_id=None, playlist_title=None,
1172 getter=IDENTITY, ie=None, video_kwargs=None, **kwargs):
1173 return cls.playlist_result(
1174 (cls.url_result(m, ie, **(video_kwargs or {})) for m in orderedSet(map(getter, matches), lazy=True)),
1175 playlist_id, playlist_title, **kwargs)
46b18f23 1176
c0d0b01f 1177 @staticmethod
311b6615 1178 def playlist_result(entries, playlist_id=None, playlist_title=None, playlist_description=None, *, multi_video=False, **kwargs):
d6983cb4 1179 """Returns a playlist"""
d6983cb4 1180 if playlist_id:
311b6615 1181 kwargs['id'] = playlist_id
d6983cb4 1182 if playlist_title:
311b6615 1183 kwargs['title'] = playlist_title
ecc97af3 1184 if playlist_description is not None:
311b6615 1185 kwargs['description'] = playlist_description
1186 return {
1187 **kwargs,
1188 '_type': 'multi_video' if multi_video else 'playlist',
1189 'entries': entries,
1190 }
d6983cb4 1191
c342041f 1192 def _search_regex(self, pattern, string, name, default=NO_DEFAULT, fatal=True, flags=0, group=None):
d6983cb4
PH
1193 """
1194 Perform a regex search on the given string, using a single or a list of
1195 patterns returning the first matching group.
1196 In case of failure return a default value or raise a WARNING or a
55b3e45b 1197 RegexNotFoundError, depending on fatal, specifying the field name.
d6983cb4 1198 """
61d3665d 1199 if string is None:
1200 mobj = None
77f90330 1201 elif isinstance(pattern, (str, re.Pattern)):
d6983cb4
PH
1202 mobj = re.search(pattern, string, flags)
1203 else:
1204 for p in pattern:
1205 mobj = re.search(p, string, flags)
c3415d1b
PH
1206 if mobj:
1207 break
d6983cb4 1208
ec11a9f4 1209 _name = self._downloader._format_err(name, self._downloader.Styles.EMPHASIS)
d6983cb4
PH
1210
1211 if mobj:
711ede6e
PH
1212 if group is None:
1213 # return the first matching group
1214 return next(g for g in mobj.groups() if g is not None)
198f7ea8 1215 elif isinstance(group, (list, tuple)):
1216 return tuple(mobj.group(g) for g in group)
711ede6e
PH
1217 else:
1218 return mobj.group(group)
c342041f 1219 elif default is not NO_DEFAULT:
d6983cb4
PH
1220 return default
1221 elif fatal:
f1a9d64e 1222 raise RegexNotFoundError('Unable to extract %s' % _name)
d6983cb4 1223 else:
6a39ee13 1224 self.report_warning('unable to extract %s' % _name + bug_reports_message())
d6983cb4
PH
1225 return None
1226
f0bc6e20 1227 def _search_json(self, start_pattern, string, name, video_id, *, end_pattern='',
1228 contains_pattern='(?s:.+)', fatal=True, default=NO_DEFAULT, **kwargs):
b7c47b74 1229 """Searches string for the JSON object specified by start_pattern"""
1230 # NB: end_pattern is only used to reduce the size of the initial match
f0bc6e20 1231 if default is NO_DEFAULT:
1232 default, has_default = {}, False
1233 else:
1234 fatal, has_default = False, True
1235
1236 json_string = self._search_regex(
1237 rf'{start_pattern}\s*(?P<json>{{\s*{contains_pattern}\s*}})\s*{end_pattern}',
1238 string, name, group='json', fatal=fatal, default=None if has_default else NO_DEFAULT)
1239 if not json_string:
1240 return default
1241
1242 _name = self._downloader._format_err(name, self._downloader.Styles.EMPHASIS)
1243 try:
1244 return self._parse_json(json_string, video_id, ignore_extra=True, **kwargs)
1245 except ExtractorError as e:
1246 if fatal:
1247 raise ExtractorError(
1248 f'Unable to extract {_name} - Failed to parse JSON', cause=e.cause, video_id=video_id)
1249 elif not has_default:
1250 self.report_warning(
1251 f'Unable to extract {_name} - Failed to parse JSON: {e}', video_id=video_id)
1252 return default
b7c47b74 1253
c342041f 1254 def _html_search_regex(self, pattern, string, name, default=NO_DEFAULT, fatal=True, flags=0, group=None):
d6983cb4
PH
1255 """
1256 Like _search_regex, but strips HTML tags and unescapes entities.
1257 """
711ede6e 1258 res = self._search_regex(pattern, string, name, default, fatal, flags, group)
d6983cb4
PH
1259 if res:
1260 return clean_html(res).strip()
1261 else:
1262 return res
1263
2118fdd1
RA
1264 def _get_netrc_login_info(self, netrc_machine=None):
1265 username = None
1266 password = None
1267 netrc_machine = netrc_machine or self._NETRC_MACHINE
1268
a06916d9 1269 if self.get_param('usenetrc', False):
2118fdd1 1270 try:
0001fcb5 1271 netrc_file = compat_expanduser(self.get_param('netrc_location') or '~')
1272 if os.path.isdir(netrc_file):
1273 netrc_file = os.path.join(netrc_file, '.netrc')
1274 info = netrc.netrc(file=netrc_file).authenticators(netrc_machine)
2118fdd1
RA
1275 if info is not None:
1276 username = info[0]
1277 password = info[2]
1278 else:
dcce092e
S
1279 raise netrc.NetrcParseError(
1280 'No authenticators for %s' % netrc_machine)
86e5f3ed 1281 except (OSError, netrc.NetrcParseError) as err:
6a39ee13 1282 self.report_warning(
dcce092e 1283 'parsing .netrc: %s' % error_to_compat_str(err))
2118fdd1 1284
dcce092e 1285 return username, password
2118fdd1 1286
1b6712ab 1287 def _get_login_info(self, username_option='username', password_option='password', netrc_machine=None):
fc79158d 1288 """
cf0649f8 1289 Get the login info as (username, password)
32443dd3
S
1290 First look for the manually specified credentials using username_option
1291 and password_option as keys in params dictionary. If no such credentials
1292 available look in the netrc file using the netrc_machine or _NETRC_MACHINE
1293 value.
fc79158d
JMF
1294 If there's no info available, return (None, None)
1295 """
fc79158d
JMF
1296
1297 # Attempt to use provided username and password or .netrc data
a06916d9 1298 username = self.get_param(username_option)
1299 if username is not None:
1300 password = self.get_param(password_option)
2118fdd1 1301 else:
1b6712ab 1302 username, password = self._get_netrc_login_info(netrc_machine)
5f6a1245 1303
2133565c 1304 return username, password
fc79158d 1305
e64b7569 1306 def _get_tfa_info(self, note='two-factor verification code'):
83317f69 1307 """
1308 Get the two-factor authentication info
1309 TODO - asking the user will be required for sms/phone verify
1310 currently just uses the command line option
1311 If there's no info available, return None
1312 """
83317f69 1313
a06916d9 1314 tfa = self.get_param('twofactor')
1315 if tfa is not None:
1316 return tfa
83317f69 1317
ac668111 1318 return getpass.getpass('Type %s and press [Return]: ' % note)
83317f69 1319
46720279
JMF
1320 # Helper functions for extracting OpenGraph info
1321 @staticmethod
ab2d5247 1322 def _og_regexes(prop):
448ef1f3 1323 content_re = r'content=(?:"([^"]+?)"|\'([^\']+?)\'|\s*([^\s"\'=<>`]+?))'
fbfde1c3
F
1324 property_re = (r'(?:name|property)=(?:\'og%(sep)s%(prop)s\'|"og%(sep)s%(prop)s"|\s*og%(sep)s%(prop)s\b)'
1325 % {'prop': re.escape(prop), 'sep': '(?:&#x3A;|[:-])'})
78fb87b2 1326 template = r'<meta[^>]+?%s[^>]+?%s'
ab2d5247 1327 return [
78fb87b2
JMF
1328 template % (property_re, content_re),
1329 template % (content_re, property_re),
ab2d5247 1330 ]
46720279 1331
864f24bd
S
1332 @staticmethod
1333 def _meta_regex(prop):
1334 return r'''(?isx)<meta
8b9848ac 1335 (?=[^>]+(?:itemprop|name|property|id|http-equiv)=(["\']?)%s\1)
864f24bd
S
1336 [^>]+?content=(["\'])(?P<content>.*?)\2''' % re.escape(prop)
1337
3c4e6d83 1338 def _og_search_property(self, prop, html, name=None, **kargs):
6606817a 1339 prop = variadic(prop)
46720279 1340 if name is None:
b070564e
S
1341 name = 'OpenGraph %s' % prop[0]
1342 og_regexes = []
1343 for p in prop:
1344 og_regexes.extend(self._og_regexes(p))
1345 escaped = self._search_regex(og_regexes, html, name, flags=re.DOTALL, **kargs)
eb0a8398
PH
1346 if escaped is None:
1347 return None
1348 return unescapeHTML(escaped)
46720279
JMF
1349
1350 def _og_search_thumbnail(self, html, **kargs):
10952eb2 1351 return self._og_search_property('image', html, 'thumbnail URL', fatal=False, **kargs)
46720279
JMF
1352
1353 def _og_search_description(self, html, **kargs):
1354 return self._og_search_property('description', html, fatal=False, **kargs)
1355
04f3fd2c 1356 def _og_search_title(self, html, *, fatal=False, **kargs):
1357 return self._og_search_property('title', html, fatal=fatal, **kargs)
46720279 1358
8ffa13e0 1359 def _og_search_video_url(self, html, name='video url', secure=True, **kargs):
a3681973
PH
1360 regexes = self._og_regexes('video') + self._og_regexes('video:url')
1361 if secure:
1362 regexes = self._og_regexes('video:secure_url') + regexes
8ffa13e0 1363 return self._html_search_regex(regexes, html, name, **kargs)
46720279 1364
78338f71
JMF
1365 def _og_search_url(self, html, **kargs):
1366 return self._og_search_property('url', html, **kargs)
1367
04f3fd2c 1368 def _html_extract_title(self, html, name='title', *, fatal=False, **kwargs):
21633673 1369 return self._html_search_regex(r'(?s)<title\b[^>]*>([^<]+)</title>', html, name, fatal=fatal, **kwargs)
77cc7c6e 1370
40c696e5 1371 def _html_search_meta(self, name, html, display_name=None, fatal=False, **kwargs):
6606817a 1372 name = variadic(name)
59040888 1373 if display_name is None:
88d9f6c0 1374 display_name = name[0]
59040888 1375 return self._html_search_regex(
88d9f6c0 1376 [self._meta_regex(n) for n in name],
711ede6e 1377 html, display_name, fatal=fatal, group='content', **kwargs)
59040888
PH
1378
1379 def _dc_search_uploader(self, html):
1380 return self._html_search_meta('dc.creator', html, 'uploader')
1381
8f97a15d 1382 @staticmethod
1383 def _rta_search(html):
8dbe9899
PH
1384 # See http://www.rtalabel.org/index.php?content=howtofaq#single
1385 if re.search(r'(?ix)<meta\s+name="rating"\s+'
1386 r' content="RTA-5042-1996-1400-1577-RTA"',
1387 html):
1388 return 18
8f97a15d 1389
1390 # And then there are the jokers who advertise that they use RTA, but actually don't.
1391 AGE_LIMIT_MARKERS = [
1392 r'Proudly Labeled <a href="http://www\.rtalabel\.org/" title="Restricted to Adults">RTA</a>',
1393 ]
1394 if any(re.search(marker, html) for marker in AGE_LIMIT_MARKERS):
1395 return 18
8dbe9899
PH
1396 return 0
1397
59040888
PH
1398 def _media_rating_search(self, html):
1399 # See http://www.tjg-designs.com/WP/metadata-code-examples-adding-metadata-to-your-web-pages/
1400 rating = self._html_search_meta('rating', html)
1401
1402 if not rating:
1403 return None
1404
1405 RATING_TABLE = {
1406 'safe for kids': 0,
1407 'general': 8,
1408 '14 years': 14,
1409 'mature': 17,
1410 'restricted': 19,
1411 }
d800609c 1412 return RATING_TABLE.get(rating.lower())
59040888 1413
69319969 1414 def _family_friendly_search(self, html):
6ca7732d 1415 # See http://schema.org/VideoObject
ac8491fc
S
1416 family_friendly = self._html_search_meta(
1417 'isFamilyFriendly', html, default=None)
69319969
NJ
1418
1419 if not family_friendly:
1420 return None
1421
1422 RATING_TABLE = {
1423 '1': 0,
1424 'true': 0,
1425 '0': 18,
1426 'false': 18,
1427 }
d800609c 1428 return RATING_TABLE.get(family_friendly.lower())
69319969 1429
0c708f11
JMF
1430 def _twitter_search_player(self, html):
1431 return self._html_search_meta('twitter:player', html,
9e1a5b84 1432 'twitter card player')
0c708f11 1433
0c36dc00 1434 def _yield_json_ld(self, html, video_id, *, fatal=True, default=NO_DEFAULT):
1435 """Yield all json ld objects in the html"""
1436 if default is not NO_DEFAULT:
1437 fatal = False
1438 for mobj in re.finditer(JSON_LD_RE, html):
1439 json_ld_item = self._parse_json(mobj.group('json_ld'), video_id, fatal=fatal)
1440 for json_ld in variadic(json_ld_item):
1441 if isinstance(json_ld, dict):
1442 yield json_ld
1443
1444 def _search_json_ld(self, html, video_id, expected_type=None, *, fatal=True, default=NO_DEFAULT):
1445 """Search for a video in any json ld in the html"""
1446 if default is not NO_DEFAULT:
1447 fatal = False
1448 info = self._json_ld(
1449 list(self._yield_json_ld(html, video_id, fatal=fatal, default=default)),
1450 video_id, fatal=fatal, expected_type=expected_type)
1451 if info:
1452 return info
4433bb02
S
1453 if default is not NO_DEFAULT:
1454 return default
1455 elif fatal:
1456 raise RegexNotFoundError('Unable to extract JSON-LD')
1457 else:
6a39ee13 1458 self.report_warning('unable to extract JSON-LD %s' % bug_reports_message())
4433bb02 1459 return {}
4ca2a3cf 1460
95b31e26 1461 def _json_ld(self, json_ld, video_id, fatal=True, expected_type=None):
14f25df2 1462 if isinstance(json_ld, str):
4ca2a3cf
S
1463 json_ld = self._parse_json(json_ld, video_id, fatal=fatal)
1464 if not json_ld:
1465 return {}
1466 info = {}
46933a15
S
1467 if not isinstance(json_ld, (list, tuple, dict)):
1468 return info
1469 if isinstance(json_ld, dict):
1470 json_ld = [json_ld]
bae14048 1471
e7e4a6e0
S
1472 INTERACTION_TYPE_MAP = {
1473 'CommentAction': 'comment',
1474 'AgreeAction': 'like',
1475 'DisagreeAction': 'dislike',
1476 'LikeAction': 'like',
1477 'DislikeAction': 'dislike',
1478 'ListenAction': 'view',
1479 'WatchAction': 'view',
1480 'ViewAction': 'view',
1481 }
1482
f3c0c773 1483 def is_type(e, *expected_types):
1484 type = variadic(traverse_obj(e, '@type'))
1485 return any(x in type for x in expected_types)
1486
29f7c58a 1487 def extract_interaction_type(e):
1488 interaction_type = e.get('interactionType')
1489 if isinstance(interaction_type, dict):
1490 interaction_type = interaction_type.get('@type')
1491 return str_or_none(interaction_type)
1492
e7e4a6e0
S
1493 def extract_interaction_statistic(e):
1494 interaction_statistic = e.get('interactionStatistic')
29f7c58a 1495 if isinstance(interaction_statistic, dict):
1496 interaction_statistic = [interaction_statistic]
e7e4a6e0
S
1497 if not isinstance(interaction_statistic, list):
1498 return
1499 for is_e in interaction_statistic:
f3c0c773 1500 if not is_type(is_e, 'InteractionCounter'):
e7e4a6e0 1501 continue
29f7c58a 1502 interaction_type = extract_interaction_type(is_e)
1503 if not interaction_type:
e7e4a6e0 1504 continue
ce5b9040
S
1505 # For interaction count some sites provide string instead of
1506 # an integer (as per spec) with non digit characters (e.g. ",")
1507 # so extracting count with more relaxed str_to_int
1508 interaction_count = str_to_int(is_e.get('userInteractionCount'))
e7e4a6e0
S
1509 if interaction_count is None:
1510 continue
1511 count_kind = INTERACTION_TYPE_MAP.get(interaction_type.split('/')[-1])
1512 if not count_kind:
1513 continue
1514 count_key = '%s_count' % count_kind
1515 if info.get(count_key) is not None:
1516 continue
1517 info[count_key] = interaction_count
1518
f5225737 1519 def extract_chapter_information(e):
1520 chapters = [{
1521 'title': part.get('name'),
1522 'start_time': part.get('startOffset'),
1523 'end_time': part.get('endOffset'),
85553414 1524 } for part in variadic(e.get('hasPart') or []) if part.get('@type') == 'Clip']
f5225737 1525 for idx, (last_c, current_c, next_c) in enumerate(zip(
1526 [{'end_time': 0}] + chapters, chapters, chapters[1:])):
1527 current_c['end_time'] = current_c['end_time'] or next_c['start_time']
1528 current_c['start_time'] = current_c['start_time'] or last_c['end_time']
1529 if None in current_c.values():
1530 self.report_warning(f'Chapter {idx} contains broken data. Not extracting chapters')
1531 return
1532 if chapters:
1533 chapters[-1]['end_time'] = chapters[-1]['end_time'] or info['duration']
1534 info['chapters'] = chapters
1535
bae14048 1536 def extract_video_object(e):
f3c0c773 1537 assert is_type(e, 'VideoObject')
f7ad7160 1538 author = e.get('author')
bae14048 1539 info.update({
0c36dc00 1540 'url': url_or_none(e.get('contentUrl')),
bae14048
S
1541 'title': unescapeHTML(e.get('name')),
1542 'description': unescapeHTML(e.get('description')),
eb2333bc 1543 'thumbnails': [{'url': unescapeHTML(url)}
21633673 1544 for url in variadic(traverse_obj(e, 'thumbnailUrl', 'thumbnailURL'))
1545 if url_or_none(url)],
bae14048
S
1546 'duration': parse_duration(e.get('duration')),
1547 'timestamp': unified_timestamp(e.get('uploadDate')),
f7ad7160 1548 # author can be an instance of 'Organization' or 'Person' types.
1549 # both types can have 'name' property(inherited from 'Thing' type). [1]
1550 # however some websites are using 'Text' type instead.
1551 # 1. https://schema.org/VideoObject
14f25df2 1552 'uploader': author.get('name') if isinstance(author, dict) else author if isinstance(author, str) else None,
56ba69e4 1553 'filesize': int_or_none(float_or_none(e.get('contentSize'))),
bae14048
S
1554 'tbr': int_or_none(e.get('bitrate')),
1555 'width': int_or_none(e.get('width')),
1556 'height': int_or_none(e.get('height')),
33a81c2c 1557 'view_count': int_or_none(e.get('interactionCount')),
bae14048 1558 })
e7e4a6e0 1559 extract_interaction_statistic(e)
f5225737 1560 extract_chapter_information(e)
bae14048 1561
d5c32548
ZM
1562 def traverse_json_ld(json_ld, at_top_level=True):
1563 for e in json_ld:
1564 if at_top_level and '@context' not in e:
1565 continue
1566 if at_top_level and set(e.keys()) == {'@context', '@graph'}:
1567 traverse_json_ld(variadic(e['@graph'], allowed_types=(dict,)), at_top_level=False)
1568 break
f3c0c773 1569 if expected_type is not None and not is_type(e, expected_type):
4433bb02 1570 continue
8f122fa0 1571 rating = traverse_obj(e, ('aggregateRating', 'ratingValue'), expected_type=float_or_none)
1572 if rating is not None:
1573 info['average_rating'] = rating
f3c0c773 1574 if is_type(e, 'TVEpisode', 'Episode'):
440863ad 1575 episode_name = unescapeHTML(e.get('name'))
46933a15 1576 info.update({
440863ad 1577 'episode': episode_name,
46933a15
S
1578 'episode_number': int_or_none(e.get('episodeNumber')),
1579 'description': unescapeHTML(e.get('description')),
1580 })
440863ad
S
1581 if not info.get('title') and episode_name:
1582 info['title'] = episode_name
46933a15 1583 part_of_season = e.get('partOfSeason')
f3c0c773 1584 if is_type(part_of_season, 'TVSeason', 'Season', 'CreativeWorkSeason'):
458fd30f
S
1585 info.update({
1586 'season': unescapeHTML(part_of_season.get('name')),
1587 'season_number': int_or_none(part_of_season.get('seasonNumber')),
1588 })
d16b3c66 1589 part_of_series = e.get('partOfSeries') or e.get('partOfTVSeries')
f3c0c773 1590 if is_type(part_of_series, 'TVSeries', 'Series', 'CreativeWorkSeries'):
46933a15 1591 info['series'] = unescapeHTML(part_of_series.get('name'))
f3c0c773 1592 elif is_type(e, 'Movie'):
391256dc
S
1593 info.update({
1594 'title': unescapeHTML(e.get('name')),
1595 'description': unescapeHTML(e.get('description')),
1596 'duration': parse_duration(e.get('duration')),
1597 'timestamp': unified_timestamp(e.get('dateCreated')),
1598 })
f3c0c773 1599 elif is_type(e, 'Article', 'NewsArticle'):
46933a15
S
1600 info.update({
1601 'timestamp': parse_iso8601(e.get('datePublished')),
1602 'title': unescapeHTML(e.get('headline')),
d5c32548 1603 'description': unescapeHTML(e.get('articleBody') or e.get('description')),
46933a15 1604 })
f3c0c773 1605 if is_type(traverse_obj(e, ('video', 0)), 'VideoObject'):
2edb38e8 1606 extract_video_object(e['video'][0])
f3c0c773 1607 elif is_type(traverse_obj(e, ('subjectOf', 0)), 'VideoObject'):
e50c3500 1608 extract_video_object(e['subjectOf'][0])
f3c0c773 1609 elif is_type(e, 'VideoObject'):
bae14048 1610 extract_video_object(e)
4433bb02
S
1611 if expected_type is None:
1612 continue
1613 else:
1614 break
c69701c6 1615 video = e.get('video')
f3c0c773 1616 if is_type(video, 'VideoObject'):
c69701c6 1617 extract_video_object(video)
4433bb02
S
1618 if expected_type is None:
1619 continue
1620 else:
1621 break
d5c32548
ZM
1622 traverse_json_ld(json_ld)
1623
90137ca4 1624 return filter_dict(info)
4ca2a3cf 1625
135dfa2c 1626 def _search_nextjs_data(self, webpage, video_id, *, transform_source=None, fatal=True, **kw):
f98709af
LL
1627 return self._parse_json(
1628 self._search_regex(
1629 r'(?s)<script[^>]+id=[\'"]__NEXT_DATA__[\'"][^>]*>([^<]+)</script>',
135dfa2c 1630 webpage, 'next.js data', fatal=fatal, **kw),
1631 video_id, transform_source=transform_source, fatal=fatal)
f98709af 1632
8072ef2b 1633 def _search_nuxt_data(self, webpage, video_id, context_name='__NUXT__', *, fatal=True, traverse=('data', 0)):
1634 """Parses Nuxt.js metadata. This works as long as the function __NUXT__ invokes is a pure function"""
66f4c04e 1635 rectx = re.escape(context_name)
8072ef2b 1636 FUNCTION_RE = r'\(function\((?P<arg_keys>.*?)\){return\s+(?P<js>{.*?})\s*;?\s*}\((?P<arg_vals>.*?)\)'
66f4c04e 1637 js, arg_keys, arg_vals = self._search_regex(
8072ef2b 1638 (rf'<script>\s*window\.{rectx}={FUNCTION_RE}\s*\)\s*;?\s*</script>', rf'{rectx}\(.*?{FUNCTION_RE}'),
1639 webpage, context_name, group=('js', 'arg_keys', 'arg_vals'), fatal=fatal)
66f4c04e
THD
1640
1641 args = dict(zip(arg_keys.split(','), arg_vals.split(',')))
1642
1643 for key, val in args.items():
1644 if val in ('undefined', 'void 0'):
1645 args[key] = 'null'
1646
8072ef2b 1647 ret = self._parse_json(js, video_id, transform_source=functools.partial(js_to_json, vars=args), fatal=fatal)
1648 return traverse_obj(ret, traverse) or {}
66f4c04e 1649
27713812 1650 @staticmethod
f8da79f8 1651 def _hidden_inputs(html):
586f1cc5 1652 html = re.sub(r'<!--(?:(?!<!--).)*-->', '', html)
201ea3ee 1653 hidden_inputs = {}
c8498368
S
1654 for input in re.findall(r'(?i)(<input[^>]+>)', html):
1655 attrs = extract_attributes(input)
1656 if not input:
201ea3ee 1657 continue
c8498368 1658 if attrs.get('type') not in ('hidden', 'submit'):
201ea3ee 1659 continue
c8498368
S
1660 name = attrs.get('name') or attrs.get('id')
1661 value = attrs.get('value')
1662 if name and value is not None:
1663 hidden_inputs[name] = value
201ea3ee 1664 return hidden_inputs
27713812 1665
cf61d96d
S
1666 def _form_hidden_inputs(self, form_id, html):
1667 form = self._search_regex(
73eb13df 1668 r'(?is)<form[^>]+?id=(["\'])%s\1[^>]*>(?P<form>.+?)</form>' % form_id,
cf61d96d
S
1669 html, '%s form' % form_id, group='form')
1670 return self._hidden_inputs(form)
1671
eb8a4433 1672 class FormatSort:
b050d210 1673 regex = r' *((?P<reverse>\+)?(?P<field>[a-zA-Z0-9_]+)((?P<separator>[~:])(?P<limit>.*?))?)? *$'
eb8a4433 1674
8326b00a 1675 default = ('hidden', 'aud_or_vid', 'hasvid', 'ie_pref', 'lang', 'quality',
7e798d72 1676 'res', 'fps', 'hdr:12', 'vcodec:vp9.2', 'channels', 'acodec',
1677 'size', 'br', 'asr', 'proto', 'ext', 'hasaud', 'source', 'id') # These must not be aliases
198e3a04 1678 ytdl_default = ('hasaud', 'lang', 'quality', 'tbr', 'filesize', 'vbr',
53ed7066 1679 'height', 'width', 'proto', 'vext', 'abr', 'aext',
f304da8a 1680 'fps', 'fs_approx', 'source', 'id')
eb8a4433 1681
1682 settings = {
1683 'vcodec': {'type': 'ordered', 'regex': True,
155d2b48 1684 'order': ['av0?1', 'vp0?9.2', 'vp0?9', '[hx]265|he?vc?', '[hx]264|avc', 'vp0?8', 'mp4v|h263', 'theora', '', None, 'none']},
eb8a4433 1685 'acodec': {'type': 'ordered', 'regex': True,
a10aa588 1686 'order': ['[af]lac', 'wav|aiff', 'opus', 'vorbis|ogg', 'aac', 'mp?4a?', 'mp3', 'e-?a?c-?3', 'ac-?3', 'dts', '', None, 'none']},
176f1866 1687 'hdr': {'type': 'ordered', 'regex': True, 'field': 'dynamic_range',
1688 'order': ['dv', '(hdr)?12', r'(hdr)?10\+', '(hdr)?10', 'hlg', '', 'sdr', None]},
f137c99e 1689 'proto': {'type': 'ordered', 'regex': True, 'field': 'protocol',
f304da8a 1690 'order': ['(ht|f)tps', '(ht|f)tp$', 'm3u8.*', '.*dash', 'websocket_frag', 'rtmpe?', '', 'mms|rtsp', 'ws|websocket', 'f4']},
eb8a4433 1691 'vext': {'type': 'ordered', 'field': 'video_ext',
91ebc640 1692 'order': ('mp4', 'webm', 'flv', '', 'none'),
eb8a4433 1693 'order_free': ('webm', 'mp4', 'flv', '', 'none')},
1694 'aext': {'type': 'ordered', 'field': 'audio_ext',
1695 'order': ('m4a', 'aac', 'mp3', 'ogg', 'opus', 'webm', '', 'none'),
1696 'order_free': ('opus', 'ogg', 'webm', 'm4a', 'mp3', 'aac', '', 'none')},
1697 'hidden': {'visible': False, 'forced': True, 'type': 'extractor', 'max': -1000},
f5510afe 1698 'aud_or_vid': {'visible': False, 'forced': True, 'type': 'multiple',
8326b00a 1699 'field': ('vcodec', 'acodec'),
1700 'function': lambda it: int(any(v != 'none' for v in it))},
f983b875 1701 'ie_pref': {'priority': True, 'type': 'extractor'},
63be1aab 1702 'hasvid': {'priority': True, 'field': 'vcodec', 'type': 'boolean', 'not_in_list': ('none',)},
1703 'hasaud': {'field': 'acodec', 'type': 'boolean', 'not_in_list': ('none',)},
10beccc9 1704 'lang': {'convert': 'float', 'field': 'language_preference', 'default': -1},
1705 'quality': {'convert': 'float', 'default': -1},
eb8a4433 1706 'filesize': {'convert': 'bytes'},
f137c99e 1707 'fs_approx': {'convert': 'bytes', 'field': 'filesize_approx'},
1708 'id': {'convert': 'string', 'field': 'format_id'},
eb8a4433 1709 'height': {'convert': 'float_none'},
1710 'width': {'convert': 'float_none'},
1711 'fps': {'convert': 'float_none'},
b8ed0f15 1712 'channels': {'convert': 'float_none', 'field': 'audio_channels'},
eb8a4433 1713 'tbr': {'convert': 'float_none'},
1714 'vbr': {'convert': 'float_none'},
1715 'abr': {'convert': 'float_none'},
1716 'asr': {'convert': 'float_none'},
10beccc9 1717 'source': {'convert': 'float', 'field': 'source_preference', 'default': -1},
63be1aab 1718
eb8a4433 1719 'codec': {'type': 'combined', 'field': ('vcodec', 'acodec')},
63be1aab 1720 'br': {'type': 'combined', 'field': ('tbr', 'vbr', 'abr'), 'same_limit': True},
1721 'size': {'type': 'combined', 'same_limit': True, 'field': ('filesize', 'fs_approx')},
1722 'ext': {'type': 'combined', 'field': ('vext', 'aext')},
f5510afe 1723 'res': {'type': 'multiple', 'field': ('height', 'width'),
dbf5416a 1724 'function': lambda it: (lambda l: min(l) if l else 0)(tuple(filter(None, it)))},
63be1aab 1725
b8ed0f15 1726 # Actual field names
19188702 1727 'format_id': {'type': 'alias', 'field': 'id'},
1728 'preference': {'type': 'alias', 'field': 'ie_pref'},
1729 'language_preference': {'type': 'alias', 'field': 'lang'},
63be1aab 1730 'source_preference': {'type': 'alias', 'field': 'source'},
08d30158 1731 'protocol': {'type': 'alias', 'field': 'proto'},
63be1aab 1732 'filesize_approx': {'type': 'alias', 'field': 'fs_approx'},
b8ed0f15 1733 'audio_channels': {'type': 'alias', 'field': 'channels'},
08d30158 1734
1735 # Deprecated
1736 'dimension': {'type': 'alias', 'field': 'res', 'deprecated': True},
1737 'resolution': {'type': 'alias', 'field': 'res', 'deprecated': True},
1738 'extension': {'type': 'alias', 'field': 'ext', 'deprecated': True},
1739 'bitrate': {'type': 'alias', 'field': 'br', 'deprecated': True},
1740 'total_bitrate': {'type': 'alias', 'field': 'tbr', 'deprecated': True},
1741 'video_bitrate': {'type': 'alias', 'field': 'vbr', 'deprecated': True},
1742 'audio_bitrate': {'type': 'alias', 'field': 'abr', 'deprecated': True},
1743 'framerate': {'type': 'alias', 'field': 'fps', 'deprecated': True},
1744 'filesize_estimate': {'type': 'alias', 'field': 'size', 'deprecated': True},
1745 'samplerate': {'type': 'alias', 'field': 'asr', 'deprecated': True},
1746 'video_ext': {'type': 'alias', 'field': 'vext', 'deprecated': True},
1747 'audio_ext': {'type': 'alias', 'field': 'aext', 'deprecated': True},
1748 'video_codec': {'type': 'alias', 'field': 'vcodec', 'deprecated': True},
1749 'audio_codec': {'type': 'alias', 'field': 'acodec', 'deprecated': True},
1750 'video': {'type': 'alias', 'field': 'hasvid', 'deprecated': True},
1751 'has_video': {'type': 'alias', 'field': 'hasvid', 'deprecated': True},
1752 'audio': {'type': 'alias', 'field': 'hasaud', 'deprecated': True},
1753 'has_audio': {'type': 'alias', 'field': 'hasaud', 'deprecated': True},
1754 'extractor': {'type': 'alias', 'field': 'ie_pref', 'deprecated': True},
1755 'extractor_preference': {'type': 'alias', 'field': 'ie_pref', 'deprecated': True},
63be1aab 1756 }
eb8a4433 1757
f304da8a 1758 def __init__(self, ie, field_preference):
1759 self._order = []
1760 self.ydl = ie._downloader
1761 self.evaluate_params(self.ydl.params, field_preference)
1762 if ie.get_param('verbose'):
1763 self.print_verbose_info(self.ydl.write_debug)
eb8a4433 1764
1765 def _get_field_setting(self, field, key):
1766 if field not in self.settings:
ee8dd27a 1767 if key in ('forced', 'priority'):
1768 return False
1769 self.ydl.deprecation_warning(
1770 f'Using arbitrary fields ({field}) for format sorting is deprecated '
1771 'and may be removed in a future version')
eb8a4433 1772 self.settings[field] = {}
1773 propObj = self.settings[field]
1774 if key not in propObj:
1775 type = propObj.get('type')
1776 if key == 'field':
1777 default = 'preference' if type == 'extractor' else (field,) if type in ('combined', 'multiple') else field
1778 elif key == 'convert':
1779 default = 'order' if type == 'ordered' else 'float_string' if field else 'ignore'
4bcc7bd1 1780 else:
f5510afe 1781 default = {'type': 'field', 'visible': True, 'order': [], 'not_in_list': (None,)}.get(key, None)
eb8a4433 1782 propObj[key] = default
1783 return propObj[key]
1784
1785 def _resolve_field_value(self, field, value, convertNone=False):
1786 if value is None:
1787 if not convertNone:
1788 return None
4bcc7bd1 1789 else:
eb8a4433 1790 value = value.lower()
1791 conversion = self._get_field_setting(field, 'convert')
1792 if conversion == 'ignore':
1793 return None
1794 if conversion == 'string':
1795 return value
1796 elif conversion == 'float_none':
1797 return float_or_none(value)
1798 elif conversion == 'bytes':
1799 return FileDownloader.parse_bytes(value)
1800 elif conversion == 'order':
da9be05e 1801 order_list = (self._use_free_order and self._get_field_setting(field, 'order_free')) or self._get_field_setting(field, 'order')
eb8a4433 1802 use_regex = self._get_field_setting(field, 'regex')
1803 list_length = len(order_list)
1804 empty_pos = order_list.index('') if '' in order_list else list_length + 1
1805 if use_regex and value is not None:
da9be05e 1806 for i, regex in enumerate(order_list):
eb8a4433 1807 if regex and re.match(regex, value):
1808 return list_length - i
1809 return list_length - empty_pos # not in list
1810 else: # not regex or value = None
1811 return list_length - (order_list.index(value) if value in order_list else empty_pos)
1812 else:
1813 if value.isnumeric():
1814 return float(value)
4bcc7bd1 1815 else:
eb8a4433 1816 self.settings[field]['convert'] = 'string'
1817 return value
1818
1819 def evaluate_params(self, params, sort_extractor):
1820 self._use_free_order = params.get('prefer_free_formats', False)
1821 self._sort_user = params.get('format_sort', [])
1822 self._sort_extractor = sort_extractor
1823
1824 def add_item(field, reverse, closest, limit_text):
1825 field = field.lower()
1826 if field in self._order:
1827 return
1828 self._order.append(field)
1829 limit = self._resolve_field_value(field, limit_text)
1830 data = {
1831 'reverse': reverse,
1832 'closest': False if limit is None else closest,
1833 'limit_text': limit_text,
1834 'limit': limit}
1835 if field in self.settings:
1836 self.settings[field].update(data)
1837 else:
1838 self.settings[field] = data
1839
1840 sort_list = (
1841 tuple(field for field in self.default if self._get_field_setting(field, 'forced'))
1842 + (tuple() if params.get('format_sort_force', False)
1843 else tuple(field for field in self.default if self._get_field_setting(field, 'priority')))
1844 + tuple(self._sort_user) + tuple(sort_extractor) + self.default)
1845
1846 for item in sort_list:
1847 match = re.match(self.regex, item)
1848 if match is None:
1849 raise ExtractorError('Invalid format sort string "%s" given by extractor' % item)
1850 field = match.group('field')
1851 if field is None:
1852 continue
1853 if self._get_field_setting(field, 'type') == 'alias':
ee8dd27a 1854 alias, field = field, self._get_field_setting(field, 'field')
08d30158 1855 if self._get_field_setting(alias, 'deprecated'):
19188702 1856 self.ydl.deprecation_warning(
1857 f'Format sorting alias {alias} is deprecated '
1858 f'and may be removed in a future version. Please use {field} instead')
eb8a4433 1859 reverse = match.group('reverse') is not None
b050d210 1860 closest = match.group('separator') == '~'
eb8a4433 1861 limit_text = match.group('limit')
1862
1863 has_limit = limit_text is not None
1864 has_multiple_fields = self._get_field_setting(field, 'type') == 'combined'
1865 has_multiple_limits = has_limit and has_multiple_fields and not self._get_field_setting(field, 'same_limit')
1866
1867 fields = self._get_field_setting(field, 'field') if has_multiple_fields else (field,)
b5ae35ee 1868 limits = limit_text.split(':') if has_multiple_limits else (limit_text,) if has_limit else tuple()
eb8a4433 1869 limit_count = len(limits)
1870 for (i, f) in enumerate(fields):
1871 add_item(f, reverse, closest,
1872 limits[i] if i < limit_count
1873 else limits[0] if has_limit and not has_multiple_limits
1874 else None)
1875
0760b0a7 1876 def print_verbose_info(self, write_debug):
b31fdeed 1877 if self._sort_user:
0760b0a7 1878 write_debug('Sort order given by user: %s' % ', '.join(self._sort_user))
eb8a4433 1879 if self._sort_extractor:
0760b0a7 1880 write_debug('Sort order given by extractor: %s' % ', '.join(self._sort_extractor))
1881 write_debug('Formats sorted by: %s' % ', '.join(['%s%s%s' % (
eb8a4433 1882 '+' if self._get_field_setting(field, 'reverse') else '', field,
1883 '%s%s(%s)' % ('~' if self._get_field_setting(field, 'closest') else ':',
1884 self._get_field_setting(field, 'limit_text'),
1885 self._get_field_setting(field, 'limit'))
1886 if self._get_field_setting(field, 'limit_text') is not None else '')
1887 for field in self._order if self._get_field_setting(field, 'visible')]))
1888
1889 def _calculate_field_preference_from_value(self, format, field, type, value):
1890 reverse = self._get_field_setting(field, 'reverse')
1891 closest = self._get_field_setting(field, 'closest')
1892 limit = self._get_field_setting(field, 'limit')
1893
1894 if type == 'extractor':
1895 maximum = self._get_field_setting(field, 'max')
1896 if value is None or (maximum is not None and value >= maximum):
f983b875 1897 value = -1
eb8a4433 1898 elif type == 'boolean':
1899 in_list = self._get_field_setting(field, 'in_list')
1900 not_in_list = self._get_field_setting(field, 'not_in_list')
1901 value = 0 if ((in_list is None or value in in_list) and (not_in_list is None or value not in not_in_list)) else -1
1902 elif type == 'ordered':
1903 value = self._resolve_field_value(field, value, True)
1904
1905 # try to convert to number
6a04a74e 1906 val_num = float_or_none(value, default=self._get_field_setting(field, 'default'))
eb8a4433 1907 is_num = self._get_field_setting(field, 'convert') != 'string' and val_num is not None
1908 if is_num:
1909 value = val_num
1910
1911 return ((-10, 0) if value is None
1912 else (1, value, 0) if not is_num # if a field has mixed strings and numbers, strings are sorted higher
1913 else (0, -abs(value - limit), value - limit if reverse else limit - value) if closest
1914 else (0, value, 0) if not reverse and (limit is None or value <= limit)
1915 else (0, -value, 0) if limit is None or (reverse and value == limit) or value > limit
1916 else (-1, value, 0))
1917
1918 def _calculate_field_preference(self, format, field):
1919 type = self._get_field_setting(field, 'type') # extractor, boolean, ordered, field, multiple
1920 get_value = lambda f: format.get(self._get_field_setting(f, 'field'))
1921 if type == 'multiple':
1922 type = 'field' # Only 'field' is allowed in multiple for now
1923 actual_fields = self._get_field_setting(field, 'field')
1924
f5510afe 1925 value = self._get_field_setting(field, 'function')(get_value(f) for f in actual_fields)
eb8a4433 1926 else:
1927 value = get_value(field)
1928 return self._calculate_field_preference_from_value(format, field, type, value)
1929
1930 def calculate_preference(self, format):
1931 # Determine missing protocol
1932 if not format.get('protocol'):
1933 format['protocol'] = determine_protocol(format)
1934
1935 # Determine missing ext
1936 if not format.get('ext') and 'url' in format:
1937 format['ext'] = determine_ext(format['url'])
1938 if format.get('vcodec') == 'none':
8326b00a 1939 format['audio_ext'] = format['ext'] if format.get('acodec') != 'none' else 'none'
eb8a4433 1940 format['video_ext'] = 'none'
1941 else:
1942 format['video_ext'] = format['ext']
1943 format['audio_ext'] = 'none'
1944 # if format.get('preference') is None and format.get('ext') in ('f4f', 'f4m'): # Not supported?
1945 # format['preference'] = -1000
1946
1947 # Determine missing bitrates
1948 if format.get('tbr') is None:
1949 if format.get('vbr') is not None and format.get('abr') is not None:
1950 format['tbr'] = format.get('vbr', 0) + format.get('abr', 0)
1951 else:
b5ae35ee 1952 if format.get('vcodec') != 'none' and format.get('vbr') is None:
eb8a4433 1953 format['vbr'] = format.get('tbr') - format.get('abr', 0)
b5ae35ee 1954 if format.get('acodec') != 'none' and format.get('abr') is None:
eb8a4433 1955 format['abr'] = format.get('tbr') - format.get('vbr', 0)
1956
1957 return tuple(self._calculate_field_preference(format, field) for field in self._order)
1958
1959 def _sort_formats(self, formats, field_preference=[]):
1960 if not formats:
88acdbc2 1961 return
1d485a1a 1962 formats.sort(key=self.FormatSort(self, field_preference).calculate_preference)
59040888 1963
96a53167
S
1964 def _check_formats(self, formats, video_id):
1965 if formats:
1966 formats[:] = filter(
1967 lambda f: self._is_valid_url(
1968 f['url'], video_id,
1969 item='%s video format' % f.get('format_id') if f.get('format_id') else 'video'),
1970 formats)
1971
f5bdb444
S
1972 @staticmethod
1973 def _remove_duplicate_formats(formats):
1974 format_urls = set()
1975 unique_formats = []
1976 for f in formats:
1977 if f['url'] not in format_urls:
1978 format_urls.add(f['url'])
1979 unique_formats.append(f)
1980 formats[:] = unique_formats
1981
45024183 1982 def _is_valid_url(self, url, video_id, item='video', headers={}):
2f0f6578
S
1983 url = self._proto_relative_url(url, scheme='http:')
1984 # For now assume non HTTP(S) URLs always valid
1985 if not (url.startswith('http://') or url.startswith('https://')):
1986 return True
96a53167 1987 try:
45024183 1988 self._request_webpage(url, video_id, 'Checking %s URL' % item, headers=headers)
96a53167 1989 return True
8bdd16b4 1990 except ExtractorError as e:
25e911a9 1991 self.to_screen(
8bdd16b4 1992 '%s: %s URL is invalid, skipping: %s'
1993 % (video_id, item, error_to_compat_str(e.cause)))
25e911a9 1994 return False
96a53167 1995
20991253 1996 def http_scheme(self):
1ede5b24 1997 """ Either "http:" or "https:", depending on the user's preferences """
20991253
PH
1998 return (
1999 'http:'
a06916d9 2000 if self.get_param('prefer_insecure', False)
20991253
PH
2001 else 'https:')
2002
57c7411f 2003 def _proto_relative_url(self, url, scheme=None):
8f97a15d 2004 scheme = scheme or self.http_scheme()
2005 assert scheme.endswith(':')
2006 return sanitize_url(url, scheme=scheme[:-1])
57c7411f 2007
4094b6e3
PH
2008 def _sleep(self, timeout, video_id, msg_template=None):
2009 if msg_template is None:
f1a9d64e 2010 msg_template = '%(video_id)s: Waiting for %(timeout)s seconds'
4094b6e3
PH
2011 msg = msg_template % {'video_id': video_id, 'timeout': timeout}
2012 self.to_screen(msg)
2013 time.sleep(timeout)
2014
f983b875 2015 def _extract_f4m_formats(self, manifest_url, video_id, preference=None, quality=None, f4m_id=None,
4de61310 2016 transform_source=lambda s: fix_xml_ampersands(s).strip(),
7360c06f 2017 fatal=True, m3u8_id=None, data=None, headers={}, query={}):
a076c1f9 2018 res = self._download_xml_handle(
f036a632 2019 manifest_url, video_id, 'Downloading f4m manifest',
97f4aecf
S
2020 'Unable to download f4m manifest',
2021 # Some manifests may be malformed, e.g. prosiebensat1 generated manifests
067aa17e 2022 # (see https://github.com/ytdl-org/youtube-dl/issues/6215#issuecomment-121704244)
4de61310 2023 transform_source=transform_source,
7360c06f 2024 fatal=fatal, data=data, headers=headers, query=query)
a076c1f9 2025 if res is False:
8d29e47f 2026 return []
31bb8d3f 2027
a076c1f9
E
2028 manifest, urlh = res
2029 manifest_url = urlh.geturl()
2030
0fdbb332 2031 return self._parse_f4m_formats(
f983b875 2032 manifest, manifest_url, video_id, preference=preference, quality=quality, f4m_id=f4m_id,
448bb5f3 2033 transform_source=transform_source, fatal=fatal, m3u8_id=m3u8_id)
0fdbb332 2034
f983b875 2035 def _parse_f4m_formats(self, manifest, manifest_url, video_id, preference=None, quality=None, f4m_id=None,
0fdbb332 2036 transform_source=lambda s: fix_xml_ampersands(s).strip(),
448bb5f3 2037 fatal=True, m3u8_id=None):
f9934b96 2038 if not isinstance(manifest, xml.etree.ElementTree.Element) and not fatal:
d9eb580a
S
2039 return []
2040
7a5c1cfe 2041 # currently yt-dlp cannot decode the playerVerificationChallenge as Akamai uses Adobe Alchemy
fb72ec58 2042 akamai_pv = manifest.find('{http://ns.adobe.com/f4m/1.0}pv-2.0')
2043 if akamai_pv is not None and ';' in akamai_pv.text:
2044 playerVerificationChallenge = akamai_pv.text.split(';')[0]
2045 if playerVerificationChallenge.strip() != '':
2046 return []
2047
31bb8d3f 2048 formats = []
7a47d07c 2049 manifest_version = '1.0'
b2527359 2050 media_nodes = manifest.findall('{http://ns.adobe.com/f4m/1.0}media')
34e48bed 2051 if not media_nodes:
7a47d07c 2052 manifest_version = '2.0'
34e48bed 2053 media_nodes = manifest.findall('{http://ns.adobe.com/f4m/2.0}media')
b22ca762 2054 # Remove unsupported DRM protected media from final formats
067aa17e 2055 # rendition (see https://github.com/ytdl-org/youtube-dl/issues/8573).
b22ca762
S
2056 media_nodes = remove_encrypted_media(media_nodes)
2057 if not media_nodes:
2058 return formats
48107c19
S
2059
2060 manifest_base_url = get_base_url(manifest)
0a5685b2 2061
a6571f10 2062 bootstrap_info = xpath_element(
0a5685b2
YCH
2063 manifest, ['{http://ns.adobe.com/f4m/1.0}bootstrapInfo', '{http://ns.adobe.com/f4m/2.0}bootstrapInfo'],
2064 'bootstrap info', default=None)
2065
edd6074c
RA
2066 vcodec = None
2067 mime_type = xpath_text(
2068 manifest, ['{http://ns.adobe.com/f4m/1.0}mimeType', '{http://ns.adobe.com/f4m/2.0}mimeType'],
2069 'base URL', default=None)
2070 if mime_type and mime_type.startswith('audio/'):
2071 vcodec = 'none'
2072
b2527359 2073 for i, media_el in enumerate(media_nodes):
77b8b4e6
S
2074 tbr = int_or_none(media_el.attrib.get('bitrate'))
2075 width = int_or_none(media_el.attrib.get('width'))
2076 height = int_or_none(media_el.attrib.get('height'))
34921b43 2077 format_id = join_nonempty(f4m_id, tbr or i)
448bb5f3
YCH
2078 # If <bootstrapInfo> is present, the specified f4m is a
2079 # stream-level manifest, and only set-level manifests may refer to
2080 # external resources. See section 11.4 and section 4 of F4M spec
2081 if bootstrap_info is None:
2082 media_url = None
2083 # @href is introduced in 2.0, see section 11.6 of F4M spec
2084 if manifest_version == '2.0':
2085 media_url = media_el.attrib.get('href')
2086 if media_url is None:
2087 media_url = media_el.attrib.get('url')
31c746e5
S
2088 if not media_url:
2089 continue
cc357c4d
S
2090 manifest_url = (
2091 media_url if media_url.startswith('http://') or media_url.startswith('https://')
48107c19 2092 else ((manifest_base_url or '/'.join(manifest_url.split('/')[:-1])) + '/' + media_url))
70f0f5a8
S
2093 # If media_url is itself a f4m manifest do the recursive extraction
2094 # since bitrates in parent manifest (this one) and media_url manifest
2095 # may differ leading to inability to resolve the format by requested
2096 # bitrate in f4m downloader
240b6045
YCH
2097 ext = determine_ext(manifest_url)
2098 if ext == 'f4m':
77b8b4e6 2099 f4m_formats = self._extract_f4m_formats(
f983b875 2100 manifest_url, video_id, preference=preference, quality=quality, f4m_id=f4m_id,
77b8b4e6
S
2101 transform_source=transform_source, fatal=fatal)
2102 # Sometimes stream-level manifest contains single media entry that
2103 # does not contain any quality metadata (e.g. http://matchtv.ru/#live-player).
2104 # At the same time parent's media entry in set-level manifest may
2105 # contain it. We will copy it from parent in such cases.
2106 if len(f4m_formats) == 1:
2107 f = f4m_formats[0]
2108 f.update({
2109 'tbr': f.get('tbr') or tbr,
2110 'width': f.get('width') or width,
2111 'height': f.get('height') or height,
2112 'format_id': f.get('format_id') if not tbr else format_id,
edd6074c 2113 'vcodec': vcodec,
77b8b4e6
S
2114 })
2115 formats.extend(f4m_formats)
70f0f5a8 2116 continue
240b6045
YCH
2117 elif ext == 'm3u8':
2118 formats.extend(self._extract_m3u8_formats(
2119 manifest_url, video_id, 'mp4', preference=preference,
f983b875 2120 quality=quality, m3u8_id=m3u8_id, fatal=fatal))
240b6045 2121 continue
31bb8d3f 2122 formats.append({
77b8b4e6 2123 'format_id': format_id,
31bb8d3f 2124 'url': manifest_url,
30d0b549 2125 'manifest_url': manifest_url,
a6571f10 2126 'ext': 'flv' if bootstrap_info is not None else None,
187ee66c 2127 'protocol': 'f4m',
b2527359 2128 'tbr': tbr,
77b8b4e6
S
2129 'width': width,
2130 'height': height,
edd6074c 2131 'vcodec': vcodec,
60ca389c 2132 'preference': preference,
f983b875 2133 'quality': quality,
31bb8d3f 2134 })
31bb8d3f
JMF
2135 return formats
2136
f983b875 2137 def _m3u8_meta_format(self, m3u8_url, ext=None, preference=None, quality=None, m3u8_id=None):
16da9bbc 2138 return {
34921b43 2139 'format_id': join_nonempty(m3u8_id, 'meta'),
704df56d
PH
2140 'url': m3u8_url,
2141 'ext': ext,
2142 'protocol': 'm3u8',
37768f92 2143 'preference': preference - 100 if preference else -100,
f983b875 2144 'quality': quality,
704df56d
PH
2145 'resolution': 'multiple',
2146 'format_note': 'Quality selection URL',
16da9bbc
YCH
2147 }
2148
b5ae35ee 2149 def _report_ignoring_subs(self, name):
2150 self.report_warning(bug_reports_message(
2151 f'Ignoring subtitle tracks found in the {name} manifest; '
2152 'if any subtitle tracks are missing,'
2153 ), only_once=True)
2154
a0c3b2d5
F
2155 def _extract_m3u8_formats(self, *args, **kwargs):
2156 fmts, subs = self._extract_m3u8_formats_and_subtitles(*args, **kwargs)
2157 if subs:
b5ae35ee 2158 self._report_ignoring_subs('HLS')
a0c3b2d5
F
2159 return fmts
2160
2161 def _extract_m3u8_formats_and_subtitles(
177877c5 2162 self, m3u8_url, video_id, ext=None, entry_protocol='m3u8_native',
a0c3b2d5
F
2163 preference=None, quality=None, m3u8_id=None, note=None,
2164 errnote=None, fatal=True, live=False, data=None, headers={},
2165 query={}):
2166
dbd82a1d 2167 res = self._download_webpage_handle(
81515ad9 2168 m3u8_url, video_id,
37a3bb66 2169 note='Downloading m3u8 information' if note is None else note,
2170 errnote='Failed to download m3u8 information' if errnote is None else errnote,
7360c06f 2171 fatal=fatal, data=data, headers=headers, query=query)
cb252080 2172
dbd82a1d 2173 if res is False:
a0c3b2d5 2174 return [], {}
cb252080 2175
dbd82a1d 2176 m3u8_doc, urlh = res
37113045 2177 m3u8_url = urlh.geturl()
9cdffeeb 2178
a0c3b2d5 2179 return self._parse_m3u8_formats_and_subtitles(
cb252080 2180 m3u8_doc, m3u8_url, ext=ext, entry_protocol=entry_protocol,
310c2ed2 2181 preference=preference, quality=quality, m3u8_id=m3u8_id,
2182 note=note, errnote=errnote, fatal=fatal, live=live, data=data,
2183 headers=headers, query=query, video_id=video_id)
cb252080 2184
a0c3b2d5 2185 def _parse_m3u8_formats_and_subtitles(
42676437 2186 self, m3u8_doc, m3u8_url=None, ext=None, entry_protocol='m3u8_native',
a0c3b2d5
F
2187 preference=None, quality=None, m3u8_id=None, live=False, note=None,
2188 errnote=None, fatal=True, data=None, headers={}, query={},
2189 video_id=None):
60755938 2190 formats, subtitles = [], {}
a0c3b2d5 2191
6b993ca7 2192 has_drm = re.search('|'.join([
2193 r'#EXT-X-FAXS-CM:', # Adobe Flash Access
2194 r'#EXT-X-(?:SESSION-)?KEY:.*?URI="skd://', # Apple FairPlay
2195 ]), m3u8_doc)
a0c3b2d5 2196
60755938 2197 def format_url(url):
14f25df2 2198 return url if re.match(r'^https?://', url) else urllib.parse.urljoin(m3u8_url, url)
60755938 2199
2200 if self.get_param('hls_split_discontinuity', False):
2201 def _extract_m3u8_playlist_indices(manifest_url=None, m3u8_doc=None):
2202 if not m3u8_doc:
2203 if not manifest_url:
2204 return []
2205 m3u8_doc = self._download_webpage(
2206 manifest_url, video_id, fatal=fatal, data=data, headers=headers,
2207 note=False, errnote='Failed to download m3u8 playlist information')
2208 if m3u8_doc is False:
2209 return []
2210 return range(1 + sum(line.startswith('#EXT-X-DISCONTINUITY') for line in m3u8_doc.splitlines()))
0def7587 2211
60755938 2212 else:
2213 def _extract_m3u8_playlist_indices(*args, **kwargs):
2214 return [None]
310c2ed2 2215
cb252080
S
2216 # References:
2217 # 1. https://tools.ietf.org/html/draft-pantos-http-live-streaming-21
067aa17e
S
2218 # 2. https://github.com/ytdl-org/youtube-dl/issues/12211
2219 # 3. https://github.com/ytdl-org/youtube-dl/issues/18923
cb252080
S
2220
2221 # We should try extracting formats only from master playlists [1, 4.3.4],
2222 # i.e. playlists that describe available qualities. On the other hand
2223 # media playlists [1, 4.3.3] should be returned as is since they contain
2224 # just the media without qualities renditions.
9cdffeeb 2225 # Fortunately, master playlist can be easily distinguished from media
cb252080 2226 # playlist based on particular tags availability. As of [1, 4.3.3, 4.3.4]
a0566bbf 2227 # master playlist tags MUST NOT appear in a media playlist and vice versa.
cb252080
S
2228 # As of [1, 4.3.3.1] #EXT-X-TARGETDURATION tag is REQUIRED for every
2229 # media playlist and MUST NOT appear in master playlist thus we can
2230 # clearly detect media playlist with this criterion.
2231
9cdffeeb 2232 if '#EXT-X-TARGETDURATION' in m3u8_doc: # media playlist, return as is
60755938 2233 formats = [{
34921b43 2234 'format_id': join_nonempty(m3u8_id, idx),
60755938 2235 'format_index': idx,
42676437 2236 'url': m3u8_url or encode_data_uri(m3u8_doc.encode('utf-8'), 'application/x-mpegurl'),
60755938 2237 'ext': ext,
2238 'protocol': entry_protocol,
2239 'preference': preference,
2240 'quality': quality,
88acdbc2 2241 'has_drm': has_drm,
60755938 2242 } for idx in _extract_m3u8_playlist_indices(m3u8_doc=m3u8_doc)]
310c2ed2 2243
a0c3b2d5 2244 return formats, subtitles
cb252080
S
2245
2246 groups = {}
2247 last_stream_inf = {}
2248
2249 def extract_media(x_media_line):
2250 media = parse_m3u8_attributes(x_media_line)
2251 # As per [1, 4.3.4.1] TYPE, GROUP-ID and NAME are REQUIRED
2252 media_type, group_id, name = media.get('TYPE'), media.get('GROUP-ID'), media.get('NAME')
2253 if not (media_type and group_id and name):
2254 return
2255 groups.setdefault(group_id, []).append(media)
a0c3b2d5
F
2256 # <https://tools.ietf.org/html/rfc8216#section-4.3.4.1>
2257 if media_type == 'SUBTITLES':
3907333c 2258 # According to RFC 8216 §4.3.4.2.1, URI is REQUIRED in the
2259 # EXT-X-MEDIA tag if the media type is SUBTITLES.
2260 # However, lack of URI has been spotted in the wild.
2261 # e.g. NebulaIE; see https://github.com/yt-dlp/yt-dlp/issues/339
2262 if not media.get('URI'):
2263 return
a0c3b2d5
F
2264 url = format_url(media['URI'])
2265 sub_info = {
2266 'url': url,
2267 'ext': determine_ext(url),
2268 }
4a2f19ab
F
2269 if sub_info['ext'] == 'm3u8':
2270 # Per RFC 8216 §3.1, the only possible subtitle format m3u8
2271 # files may contain is WebVTT:
2272 # <https://tools.ietf.org/html/rfc8216#section-3.1>
2273 sub_info['ext'] = 'vtt'
2274 sub_info['protocol'] = 'm3u8_native'
37a3bb66 2275 lang = media.get('LANGUAGE') or 'und'
a0c3b2d5 2276 subtitles.setdefault(lang, []).append(sub_info)
cb252080
S
2277 if media_type not in ('VIDEO', 'AUDIO'):
2278 return
2279 media_url = media.get('URI')
2280 if media_url:
310c2ed2 2281 manifest_url = format_url(media_url)
60755938 2282 formats.extend({
34921b43 2283 'format_id': join_nonempty(m3u8_id, group_id, name, idx),
60755938 2284 'format_note': name,
2285 'format_index': idx,
2286 'url': manifest_url,
2287 'manifest_url': m3u8_url,
2288 'language': media.get('LANGUAGE'),
2289 'ext': ext,
2290 'protocol': entry_protocol,
2291 'preference': preference,
2292 'quality': quality,
2293 'vcodec': 'none' if media_type == 'AUDIO' else None,
2294 } for idx in _extract_m3u8_playlist_indices(manifest_url))
cb252080
S
2295
2296 def build_stream_name():
2297 # Despite specification does not mention NAME attribute for
3019cb0c
S
2298 # EXT-X-STREAM-INF tag it still sometimes may be present (see [1]
2299 # or vidio test in TestInfoExtractor.test_parse_m3u8_formats)
ddd258f9 2300 # 1. http://www.vidio.com/watch/165683-dj_ambred-booyah-live-2015
cb252080
S
2301 stream_name = last_stream_inf.get('NAME')
2302 if stream_name:
2303 return stream_name
2304 # If there is no NAME in EXT-X-STREAM-INF it will be obtained
2305 # from corresponding rendition group
2306 stream_group_id = last_stream_inf.get('VIDEO')
2307 if not stream_group_id:
2308 return
2309 stream_group = groups.get(stream_group_id)
2310 if not stream_group:
2311 return stream_group_id
2312 rendition = stream_group[0]
2313 return rendition.get('NAME') or stream_group_id
2314
379306ef 2315 # parse EXT-X-MEDIA tags before EXT-X-STREAM-INF in order to have the
2bfc1d9d
RA
2316 # chance to detect video only formats when EXT-X-STREAM-INF tags
2317 # precede EXT-X-MEDIA tags in HLS manifest such as [3].
2318 for line in m3u8_doc.splitlines():
2319 if line.startswith('#EXT-X-MEDIA:'):
2320 extract_media(line)
2321
704df56d
PH
2322 for line in m3u8_doc.splitlines():
2323 if line.startswith('#EXT-X-STREAM-INF:'):
cb252080 2324 last_stream_inf = parse_m3u8_attributes(line)
704df56d
PH
2325 elif line.startswith('#') or not line.strip():
2326 continue
2327 else:
9c99bef7 2328 tbr = float_or_none(
3089bc74
S
2329 last_stream_inf.get('AVERAGE-BANDWIDTH')
2330 or last_stream_inf.get('BANDWIDTH'), scale=1000)
30d0b549 2331 manifest_url = format_url(line.strip())
5ef62fc4 2332
60755938 2333 for idx in _extract_m3u8_playlist_indices(manifest_url):
2334 format_id = [m3u8_id, None, idx]
310c2ed2 2335 # Bandwidth of live streams may differ over time thus making
2336 # format_id unpredictable. So it's better to keep provided
2337 # format_id intact.
2338 if not live:
60755938 2339 stream_name = build_stream_name()
34921b43 2340 format_id[1] = stream_name or '%d' % (tbr or len(formats))
310c2ed2 2341 f = {
34921b43 2342 'format_id': join_nonempty(*format_id),
60755938 2343 'format_index': idx,
310c2ed2 2344 'url': manifest_url,
2345 'manifest_url': m3u8_url,
2346 'tbr': tbr,
2347 'ext': ext,
2348 'fps': float_or_none(last_stream_inf.get('FRAME-RATE')),
2349 'protocol': entry_protocol,
2350 'preference': preference,
2351 'quality': quality,
2352 }
2353 resolution = last_stream_inf.get('RESOLUTION')
2354 if resolution:
2355 mobj = re.search(r'(?P<width>\d+)[xX](?P<height>\d+)', resolution)
2356 if mobj:
2357 f['width'] = int(mobj.group('width'))
2358 f['height'] = int(mobj.group('height'))
2359 # Unified Streaming Platform
2360 mobj = re.search(
2361 r'audio.*?(?:%3D|=)(\d+)(?:-video.*?(?:%3D|=)(\d+))?', f['url'])
2362 if mobj:
2363 abr, vbr = mobj.groups()
2364 abr, vbr = float_or_none(abr, 1000), float_or_none(vbr, 1000)
2365 f.update({
2366 'vbr': vbr,
2367 'abr': abr,
2368 })
2369 codecs = parse_codecs(last_stream_inf.get('CODECS'))
2370 f.update(codecs)
2371 audio_group_id = last_stream_inf.get('AUDIO')
2372 # As per [1, 4.3.4.1.1] any EXT-X-STREAM-INF tag which
2373 # references a rendition group MUST have a CODECS attribute.
62b58c09 2374 # However, this is not always respected. E.g. [2]
310c2ed2 2375 # contains EXT-X-STREAM-INF tag which references AUDIO
2376 # rendition group but does not have CODECS and despite
2377 # referencing an audio group it represents a complete
2378 # (with audio and video) format. So, for such cases we will
2379 # ignore references to rendition groups and treat them
2380 # as complete formats.
2381 if audio_group_id and codecs and f.get('vcodec') != 'none':
2382 audio_group = groups.get(audio_group_id)
2383 if audio_group and audio_group[0].get('URI'):
2384 # TODO: update acodec for audio only formats with
2385 # the same GROUP-ID
2386 f['acodec'] = 'none'
fc21af50 2387 if not f.get('ext'):
2388 f['ext'] = 'm4a' if f.get('vcodec') == 'none' else 'mp4'
310c2ed2 2389 formats.append(f)
2390
2391 # for DailyMotion
2392 progressive_uri = last_stream_inf.get('PROGRESSIVE-URI')
2393 if progressive_uri:
2394 http_f = f.copy()
2395 del http_f['manifest_url']
2396 http_f.update({
2397 'format_id': f['format_id'].replace('hls-', 'http-'),
2398 'protocol': 'http',
2399 'url': progressive_uri,
2400 })
2401 formats.append(http_f)
5ef62fc4 2402
cb252080 2403 last_stream_inf = {}
a0c3b2d5 2404 return formats, subtitles
704df56d 2405
3cf4b91d
C
2406 def _extract_m3u8_vod_duration(
2407 self, m3u8_vod_url, video_id, note=None, errnote=None, data=None, headers={}, query={}):
2408
2409 m3u8_vod = self._download_webpage(
2410 m3u8_vod_url, video_id,
2411 note='Downloading m3u8 VOD manifest' if note is None else note,
2412 errnote='Failed to download VOD manifest' if errnote is None else errnote,
2413 fatal=False, data=data, headers=headers, query=query)
2414
2415 return self._parse_m3u8_vod_duration(m3u8_vod or '', video_id)
2416
2417 def _parse_m3u8_vod_duration(self, m3u8_vod, video_id):
2418 if '#EXT-X-PLAYLIST-TYPE:VOD' not in m3u8_vod:
2419 return None
2420
2421 return int(sum(
2422 float(line[len('#EXTINF:'):].split(',')[0])
2423 for line in m3u8_vod.splitlines() if line.startswith('#EXTINF:'))) or None
2424
a107193e
S
2425 @staticmethod
2426 def _xpath_ns(path, namespace=None):
2427 if not namespace:
2428 return path
2429 out = []
2430 for c in path.split('/'):
2431 if not c or c == '.':
2432 out.append(c)
2433 else:
2434 out.append('{%s}%s' % (namespace, c))
2435 return '/'.join(out)
2436
da1c94ee 2437 def _extract_smil_formats_and_subtitles(self, smil_url, video_id, fatal=True, f4m_params=None, transform_source=None):
a076c1f9
E
2438 res = self._download_smil(smil_url, video_id, fatal=fatal, transform_source=transform_source)
2439 if res is False:
995029a1 2440 assert not fatal
774a46c5 2441 return [], {}
e89a2aab 2442
a076c1f9
E
2443 smil, urlh = res
2444 smil_url = urlh.geturl()
2445
17712eeb 2446 namespace = self._parse_smil_namespace(smil)
a107193e 2447
da1c94ee 2448 fmts = self._parse_smil_formats(
a107193e 2449 smil, smil_url, video_id, namespace=namespace, f4m_params=f4m_params)
da1c94ee
F
2450 subs = self._parse_smil_subtitles(
2451 smil, namespace=namespace)
2452
2453 return fmts, subs
2454
2455 def _extract_smil_formats(self, *args, **kwargs):
2456 fmts, subs = self._extract_smil_formats_and_subtitles(*args, **kwargs)
2457 if subs:
b5ae35ee 2458 self._report_ignoring_subs('SMIL')
da1c94ee 2459 return fmts
a107193e
S
2460
2461 def _extract_smil_info(self, smil_url, video_id, fatal=True, f4m_params=None):
a076c1f9
E
2462 res = self._download_smil(smil_url, video_id, fatal=fatal)
2463 if res is False:
a107193e 2464 return {}
a076c1f9
E
2465
2466 smil, urlh = res
2467 smil_url = urlh.geturl()
2468
a107193e
S
2469 return self._parse_smil(smil, smil_url, video_id, f4m_params=f4m_params)
2470
09f572fb 2471 def _download_smil(self, smil_url, video_id, fatal=True, transform_source=None):
a076c1f9 2472 return self._download_xml_handle(
a107193e 2473 smil_url, video_id, 'Downloading SMIL file',
09f572fb 2474 'Unable to download SMIL file', fatal=fatal, transform_source=transform_source)
a107193e
S
2475
2476 def _parse_smil(self, smil, smil_url, video_id, f4m_params=None):
17712eeb 2477 namespace = self._parse_smil_namespace(smil)
a107193e
S
2478
2479 formats = self._parse_smil_formats(
2480 smil, smil_url, video_id, namespace=namespace, f4m_params=f4m_params)
2481 subtitles = self._parse_smil_subtitles(smil, namespace=namespace)
2482
2483 video_id = os.path.splitext(url_basename(smil_url))[0]
2484 title = None
2485 description = None
647eab45 2486 upload_date = None
a107193e
S
2487 for meta in smil.findall(self._xpath_ns('./head/meta', namespace)):
2488 name = meta.attrib.get('name')
2489 content = meta.attrib.get('content')
2490 if not name or not content:
2491 continue
2492 if not title and name == 'title':
2493 title = content
2494 elif not description and name in ('description', 'abstract'):
2495 description = content
647eab45
S
2496 elif not upload_date and name == 'date':
2497 upload_date = unified_strdate(content)
a107193e 2498
1e5bcdec
S
2499 thumbnails = [{
2500 'id': image.get('type'),
2501 'url': image.get('src'),
2502 'width': int_or_none(image.get('width')),
2503 'height': int_or_none(image.get('height')),
2504 } for image in smil.findall(self._xpath_ns('.//image', namespace)) if image.get('src')]
2505
a107193e
S
2506 return {
2507 'id': video_id,
2508 'title': title or video_id,
2509 'description': description,
647eab45 2510 'upload_date': upload_date,
1e5bcdec 2511 'thumbnails': thumbnails,
a107193e
S
2512 'formats': formats,
2513 'subtitles': subtitles,
2514 }
2515
17712eeb
S
2516 def _parse_smil_namespace(self, smil):
2517 return self._search_regex(
2518 r'(?i)^{([^}]+)?}smil$', smil.tag, 'namespace', default=None)
2519
f877c6ae 2520 def _parse_smil_formats(self, smil, smil_url, video_id, namespace=None, f4m_params=None, transform_rtmp_url=None):
a107193e
S
2521 base = smil_url
2522 for meta in smil.findall(self._xpath_ns('./head/meta', namespace)):
2523 b = meta.get('base') or meta.get('httpBase')
2524 if b:
2525 base = b
2526 break
e89a2aab
S
2527
2528 formats = []
2529 rtmp_count = 0
a107193e 2530 http_count = 0
7f32e5dc 2531 m3u8_count = 0
9359f3d4 2532 imgs_count = 0
a107193e 2533
9359f3d4 2534 srcs = set()
ad96b4c8
YCH
2535 media = smil.findall(self._xpath_ns('.//video', namespace)) + smil.findall(self._xpath_ns('.//audio', namespace))
2536 for medium in media:
2537 src = medium.get('src')
81e1c4e2 2538 if not src or src in srcs:
a107193e 2539 continue
9359f3d4 2540 srcs.add(src)
a107193e 2541
ad96b4c8
YCH
2542 bitrate = float_or_none(medium.get('system-bitrate') or medium.get('systemBitrate'), 1000)
2543 filesize = int_or_none(medium.get('size') or medium.get('fileSize'))
2544 width = int_or_none(medium.get('width'))
2545 height = int_or_none(medium.get('height'))
2546 proto = medium.get('proto')
2547 ext = medium.get('ext')
a107193e 2548 src_ext = determine_ext(src)
ad96b4c8 2549 streamer = medium.get('streamer') or base
a107193e
S
2550
2551 if proto == 'rtmp' or streamer.startswith('rtmp'):
2552 rtmp_count += 1
2553 formats.append({
2554 'url': streamer,
2555 'play_path': src,
2556 'ext': 'flv',
2557 'format_id': 'rtmp-%d' % (rtmp_count if bitrate is None else bitrate),
2558 'tbr': bitrate,
2559 'filesize': filesize,
2560 'width': width,
2561 'height': height,
2562 })
f877c6ae
YCH
2563 if transform_rtmp_url:
2564 streamer, src = transform_rtmp_url(streamer, src)
2565 formats[-1].update({
2566 'url': streamer,
2567 'play_path': src,
2568 })
a107193e
S
2569 continue
2570
14f25df2 2571 src_url = src if src.startswith('http') else urllib.parse.urljoin(base, src)
c349456e 2572 src_url = src_url.strip()
a107193e
S
2573
2574 if proto == 'm3u8' or src_ext == 'm3u8':
7f32e5dc 2575 m3u8_formats = self._extract_m3u8_formats(
2576 src_url, video_id, ext or 'mp4', m3u8_id='hls', fatal=False)
2577 if len(m3u8_formats) == 1:
2578 m3u8_count += 1
2579 m3u8_formats[0].update({
2580 'format_id': 'hls-%d' % (m3u8_count if bitrate is None else bitrate),
2581 'tbr': bitrate,
2582 'width': width,
2583 'height': height,
2584 })
2585 formats.extend(m3u8_formats)
bd21ead2 2586 elif src_ext == 'f4m':
a107193e
S
2587 f4m_url = src_url
2588 if not f4m_params:
2589 f4m_params = {
2590 'hdcore': '3.2.0',
2591 'plugin': 'flowplayer-3.2.0.1',
2592 }
2593 f4m_url += '&' if '?' in f4m_url else '?'
14f25df2 2594 f4m_url += urllib.parse.urlencode(f4m_params)
7e5edcfd 2595 formats.extend(self._extract_f4m_formats(f4m_url, video_id, f4m_id='hds', fatal=False))
bd21ead2
RA
2596 elif src_ext == 'mpd':
2597 formats.extend(self._extract_mpd_formats(
2598 src_url, video_id, mpd_id='dash', fatal=False))
2599 elif re.search(r'\.ism/[Mm]anifest', src_url):
2600 formats.extend(self._extract_ism_formats(
2601 src_url, video_id, ism_id='mss', fatal=False))
2602 elif src_url.startswith('http') and self._is_valid_url(src, video_id):
a107193e
S
2603 http_count += 1
2604 formats.append({
2605 'url': src_url,
2606 'ext': ext or src_ext or 'flv',
2607 'format_id': 'http-%d' % (bitrate or http_count),
2608 'tbr': bitrate,
2609 'filesize': filesize,
2610 'width': width,
2611 'height': height,
2612 })
63757032 2613
9359f3d4
F
2614 for medium in smil.findall(self._xpath_ns('.//imagestream', namespace)):
2615 src = medium.get('src')
2616 if not src or src in srcs:
2617 continue
2618 srcs.add(src)
2619
2620 imgs_count += 1
2621 formats.append({
2622 'format_id': 'imagestream-%d' % (imgs_count),
2623 'url': src,
2624 'ext': mimetype2ext(medium.get('type')),
2625 'acodec': 'none',
2626 'vcodec': 'none',
2627 'width': int_or_none(medium.get('width')),
2628 'height': int_or_none(medium.get('height')),
2629 'format_note': 'SMIL storyboards',
2630 })
2631
e89a2aab
S
2632 return formats
2633
ce00af87 2634 def _parse_smil_subtitles(self, smil, namespace=None, subtitles_lang='en'):
d413095f 2635 urls = []
a107193e
S
2636 subtitles = {}
2637 for num, textstream in enumerate(smil.findall(self._xpath_ns('.//textstream', namespace))):
2638 src = textstream.get('src')
d413095f 2639 if not src or src in urls:
a107193e 2640 continue
d413095f 2641 urls.append(src)
df634be2 2642 ext = textstream.get('ext') or mimetype2ext(textstream.get('type')) or determine_ext(src)
03bc7237 2643 lang = textstream.get('systemLanguage') or textstream.get('systemLanguageName') or textstream.get('lang') or subtitles_lang
a107193e
S
2644 subtitles.setdefault(lang, []).append({
2645 'url': src,
2646 'ext': ext,
2647 })
2648 return subtitles
63757032 2649
47a5cb77 2650 def _extract_xspf_playlist(self, xspf_url, playlist_id, fatal=True):
a076c1f9 2651 res = self._download_xml_handle(
47a5cb77 2652 xspf_url, playlist_id, 'Downloading xpsf playlist',
942acef5 2653 'Unable to download xspf manifest', fatal=fatal)
a076c1f9 2654 if res is False:
942acef5 2655 return []
a076c1f9
E
2656
2657 xspf, urlh = res
2658 xspf_url = urlh.geturl()
2659
47a5cb77
S
2660 return self._parse_xspf(
2661 xspf, playlist_id, xspf_url=xspf_url,
2662 xspf_base_url=base_url(xspf_url))
8d6765cf 2663
47a5cb77 2664 def _parse_xspf(self, xspf_doc, playlist_id, xspf_url=None, xspf_base_url=None):
8d6765cf
S
2665 NS_MAP = {
2666 'xspf': 'http://xspf.org/ns/0/',
2667 's1': 'http://static.streamone.nl/player/ns/0',
2668 }
2669
2670 entries = []
47a5cb77 2671 for track in xspf_doc.findall(xpath_with_ns('./xspf:trackList/xspf:track', NS_MAP)):
8d6765cf 2672 title = xpath_text(
98044462 2673 track, xpath_with_ns('./xspf:title', NS_MAP), 'title', default=playlist_id)
8d6765cf
S
2674 description = xpath_text(
2675 track, xpath_with_ns('./xspf:annotation', NS_MAP), 'description')
2676 thumbnail = xpath_text(
2677 track, xpath_with_ns('./xspf:image', NS_MAP), 'thumbnail')
2678 duration = float_or_none(
2679 xpath_text(track, xpath_with_ns('./xspf:duration', NS_MAP), 'duration'), 1000)
2680
47a5cb77
S
2681 formats = []
2682 for location in track.findall(xpath_with_ns('./xspf:location', NS_MAP)):
2683 format_url = urljoin(xspf_base_url, location.text)
2684 if not format_url:
2685 continue
2686 formats.append({
2687 'url': format_url,
2688 'manifest_url': xspf_url,
2689 'format_id': location.get(xpath_with_ns('s1:label', NS_MAP)),
2690 'width': int_or_none(location.get(xpath_with_ns('s1:width', NS_MAP))),
2691 'height': int_or_none(location.get(xpath_with_ns('s1:height', NS_MAP))),
2692 })
8d6765cf
S
2693 self._sort_formats(formats)
2694
2695 entries.append({
2696 'id': playlist_id,
2697 'title': title,
2698 'description': description,
2699 'thumbnail': thumbnail,
2700 'duration': duration,
2701 'formats': formats,
2702 })
2703 return entries
2704
171e59ed
F
2705 def _extract_mpd_formats(self, *args, **kwargs):
2706 fmts, subs = self._extract_mpd_formats_and_subtitles(*args, **kwargs)
2707 if subs:
b5ae35ee 2708 self._report_ignoring_subs('DASH')
171e59ed
F
2709 return fmts
2710
2711 def _extract_mpd_formats_and_subtitles(
2712 self, mpd_url, video_id, mpd_id=None, note=None, errnote=None,
2713 fatal=True, data=None, headers={}, query={}):
47a5cb77 2714 res = self._download_xml_handle(
1bac3455 2715 mpd_url, video_id,
37a3bb66 2716 note='Downloading MPD manifest' if note is None else note,
2717 errnote='Failed to download MPD manifest' if errnote is None else errnote,
7360c06f 2718 fatal=fatal, data=data, headers=headers, query=query)
1bac3455 2719 if res is False:
171e59ed 2720 return [], {}
47a5cb77 2721 mpd_doc, urlh = res
c25720ef 2722 if mpd_doc is None:
171e59ed 2723 return [], {}
779da8e3
E
2724
2725 # We could have been redirected to a new url when we retrieved our mpd file.
2726 mpd_url = urlh.geturl()
2727 mpd_base_url = base_url(mpd_url)
1bac3455 2728
171e59ed 2729 return self._parse_mpd_formats_and_subtitles(
545cc85d 2730 mpd_doc, mpd_id, mpd_base_url, mpd_url)
2d2fa82d 2731
171e59ed
F
2732 def _parse_mpd_formats(self, *args, **kwargs):
2733 fmts, subs = self._parse_mpd_formats_and_subtitles(*args, **kwargs)
2734 if subs:
b5ae35ee 2735 self._report_ignoring_subs('DASH')
171e59ed
F
2736 return fmts
2737
2738 def _parse_mpd_formats_and_subtitles(
2739 self, mpd_doc, mpd_id=None, mpd_base_url='', mpd_url=None):
f0948348
S
2740 """
2741 Parse formats from MPD manifest.
2742 References:
2743 1. MPEG-DASH Standard, ISO/IEC 23009-1:2014(E),
2744 http://standards.iso.org/ittf/PubliclyAvailableStandards/c065274_ISO_IEC_23009-1_2014.zip
2745 2. https://en.wikipedia.org/wiki/Dynamic_Adaptive_Streaming_over_HTTP
2746 """
a06916d9 2747 if not self.get_param('dynamic_mpd', True):
78895bd3 2748 if mpd_doc.get('type') == 'dynamic':
171e59ed 2749 return [], {}
2d2fa82d 2750
91cb6b50 2751 namespace = self._search_regex(r'(?i)^{([^}]+)?}MPD$', mpd_doc.tag, 'namespace', default=None)
f14be228 2752
2753 def _add_ns(path):
2754 return self._xpath_ns(path, namespace)
2755
675d0016 2756 def is_drm_protected(element):
2757 return element.find(_add_ns('ContentProtection')) is not None
2758
1bac3455 2759 def extract_multisegment_info(element, ms_parent_info):
2760 ms_info = ms_parent_info.copy()
b4c1d6e8
S
2761
2762 # As per [1, 5.3.9.2.2] SegmentList and SegmentTemplate share some
2763 # common attributes and elements. We will only extract relevant
2764 # for us.
2765 def extract_common(source):
2766 segment_timeline = source.find(_add_ns('SegmentTimeline'))
2767 if segment_timeline is not None:
2768 s_e = segment_timeline.findall(_add_ns('S'))
2769 if s_e:
2770 ms_info['total_number'] = 0
2771 ms_info['s'] = []
2772 for s in s_e:
2773 r = int(s.get('r', 0))
2774 ms_info['total_number'] += 1 + r
2775 ms_info['s'].append({
2776 't': int(s.get('t', 0)),
2777 # @d is mandatory (see [1, 5.3.9.6.2, Table 17, page 60])
2778 'd': int(s.attrib['d']),
2779 'r': r,
2780 })
2781 start_number = source.get('startNumber')
2782 if start_number:
2783 ms_info['start_number'] = int(start_number)
2784 timescale = source.get('timescale')
2785 if timescale:
2786 ms_info['timescale'] = int(timescale)
2787 segment_duration = source.get('duration')
2788 if segment_duration:
48504785 2789 ms_info['segment_duration'] = float(segment_duration)
b4c1d6e8
S
2790
2791 def extract_Initialization(source):
2792 initialization = source.find(_add_ns('Initialization'))
2793 if initialization is not None:
2794 ms_info['initialization_url'] = initialization.attrib['sourceURL']
2795
f14be228 2796 segment_list = element.find(_add_ns('SegmentList'))
1bac3455 2797 if segment_list is not None:
b4c1d6e8
S
2798 extract_common(segment_list)
2799 extract_Initialization(segment_list)
f14be228 2800 segment_urls_e = segment_list.findall(_add_ns('SegmentURL'))
1bac3455 2801 if segment_urls_e:
2802 ms_info['segment_urls'] = [segment.attrib['media'] for segment in segment_urls_e]
1bac3455 2803 else:
f14be228 2804 segment_template = element.find(_add_ns('SegmentTemplate'))
1bac3455 2805 if segment_template is not None:
b4c1d6e8 2806 extract_common(segment_template)
e228616c
S
2807 media = segment_template.get('media')
2808 if media:
2809 ms_info['media'] = media
1bac3455 2810 initialization = segment_template.get('initialization')
2811 if initialization:
e228616c 2812 ms_info['initialization'] = initialization
1bac3455 2813 else:
b4c1d6e8 2814 extract_Initialization(segment_template)
1bac3455 2815 return ms_info
b323e170 2816
1bac3455 2817 mpd_duration = parse_duration(mpd_doc.get('mediaPresentationDuration'))
6251555f 2818 formats, subtitles = [], {}
234416e4 2819 stream_numbers = collections.defaultdict(int)
f14be228 2820 for period in mpd_doc.findall(_add_ns('Period')):
1bac3455 2821 period_duration = parse_duration(period.get('duration')) or mpd_duration
2822 period_ms_info = extract_multisegment_info(period, {
2823 'start_number': 1,
2824 'timescale': 1,
2825 })
f14be228 2826 for adaptation_set in period.findall(_add_ns('AdaptationSet')):
1bac3455 2827 adaption_set_ms_info = extract_multisegment_info(adaptation_set, period_ms_info)
f14be228 2828 for representation in adaptation_set.findall(_add_ns('Representation')):
1bac3455 2829 representation_attrib = adaptation_set.attrib.copy()
2830 representation_attrib.update(representation.attrib)
f0948348 2831 # According to [1, 5.3.7.2, Table 9, page 41], @mimeType is mandatory
a6c8b759 2832 mime_type = representation_attrib['mimeType']
171e59ed
F
2833 content_type = representation_attrib.get('contentType', mime_type.split('/')[0])
2834
21633673 2835 codec_str = representation_attrib.get('codecs', '')
2836 # Some kind of binary subtitle found in some youtube livestreams
2837 if mime_type == 'application/x-rawcc':
2838 codecs = {'scodec': codec_str}
2839 else:
2840 codecs = parse_codecs(codec_str)
be2fc5b2 2841 if content_type not in ('video', 'audio', 'text'):
2842 if mime_type == 'image/jpeg':
a8731fcc 2843 content_type = mime_type
21633673 2844 elif codecs.get('vcodec', 'none') != 'none':
4afa3ec4 2845 content_type = 'video'
21633673 2846 elif codecs.get('acodec', 'none') != 'none':
4afa3ec4 2847 content_type = 'audio'
3fe75fdc 2848 elif codecs.get('scodec', 'none') != 'none':
be2fc5b2 2849 content_type = 'text'
6993f78d 2850 elif mimetype2ext(mime_type) in ('tt', 'dfxp', 'ttml', 'xml', 'json'):
2851 content_type = 'text'
cdb19aa4 2852 else:
be2fc5b2 2853 self.report_warning('Unknown MIME type %s in DASH manifest' % mime_type)
2854 continue
2855
2856 base_url = ''
2857 for element in (representation, adaptation_set, period, mpd_doc):
2858 base_url_e = element.find(_add_ns('BaseURL'))
47046464 2859 if try_call(lambda: base_url_e.text) is not None:
be2fc5b2 2860 base_url = base_url_e.text + base_url
2861 if re.match(r'^https?://', base_url):
2862 break
f9cc0161 2863 if mpd_base_url and base_url.startswith('/'):
14f25df2 2864 base_url = urllib.parse.urljoin(mpd_base_url, base_url)
f9cc0161
D
2865 elif mpd_base_url and not re.match(r'^https?://', base_url):
2866 if not mpd_base_url.endswith('/'):
be2fc5b2 2867 mpd_base_url += '/'
2868 base_url = mpd_base_url + base_url
2869 representation_id = representation_attrib.get('id')
2870 lang = representation_attrib.get('lang')
2871 url_el = representation.find(_add_ns('BaseURL'))
2872 filesize = int_or_none(url_el.attrib.get('{http://youtube.com/yt/2012/10/10}contentLength') if url_el is not None else None)
2873 bandwidth = int_or_none(representation_attrib.get('bandwidth'))
2874 if representation_id is not None:
2875 format_id = representation_id
2876 else:
2877 format_id = content_type
2878 if mpd_id:
2879 format_id = mpd_id + '-' + format_id
2880 if content_type in ('video', 'audio'):
2881 f = {
2882 'format_id': format_id,
2883 'manifest_url': mpd_url,
2884 'ext': mimetype2ext(mime_type),
2885 'width': int_or_none(representation_attrib.get('width')),
2886 'height': int_or_none(representation_attrib.get('height')),
2887 'tbr': float_or_none(bandwidth, 1000),
2888 'asr': int_or_none(representation_attrib.get('audioSamplingRate')),
2889 'fps': int_or_none(representation_attrib.get('frameRate')),
2890 'language': lang if lang not in ('mul', 'und', 'zxx', 'mis') else None,
2891 'format_note': 'DASH %s' % content_type,
2892 'filesize': filesize,
2893 'container': mimetype2ext(mime_type) + '_dash',
4afa3ec4 2894 **codecs
be2fc5b2 2895 }
be2fc5b2 2896 elif content_type == 'text':
2897 f = {
2898 'ext': mimetype2ext(mime_type),
2899 'manifest_url': mpd_url,
2900 'filesize': filesize,
2901 }
2902 elif content_type == 'image/jpeg':
2903 # See test case in VikiIE
2904 # https://www.viki.com/videos/1175236v-choosing-spouse-by-lottery-episode-1
2905 f = {
2906 'format_id': format_id,
2907 'ext': 'mhtml',
2908 'manifest_url': mpd_url,
2909 'format_note': 'DASH storyboards (jpeg)',
2910 'acodec': 'none',
2911 'vcodec': 'none',
2912 }
88acdbc2 2913 if is_drm_protected(adaptation_set) or is_drm_protected(representation):
2914 f['has_drm'] = True
be2fc5b2 2915 representation_ms_info = extract_multisegment_info(representation, adaption_set_ms_info)
2916
2917 def prepare_template(template_name, identifiers):
2918 tmpl = representation_ms_info[template_name]
2919 # First of, % characters outside $...$ templates
2920 # must be escaped by doubling for proper processing
2921 # by % operator string formatting used further (see
2922 # https://github.com/ytdl-org/youtube-dl/issues/16867).
2923 t = ''
2924 in_template = False
2925 for c in tmpl:
2926 t += c
2927 if c == '$':
2928 in_template = not in_template
2929 elif c == '%' and not in_template:
eca1f0d1 2930 t += c
be2fc5b2 2931 # Next, $...$ templates are translated to their
2932 # %(...) counterparts to be used with % operator
2933 if representation_id is not None:
2934 t = t.replace('$RepresentationID$', representation_id)
2935 t = re.sub(r'\$(%s)\$' % '|'.join(identifiers), r'%(\1)d', t)
2936 t = re.sub(r'\$(%s)%%([^$]+)\$' % '|'.join(identifiers), r'%(\1)\2', t)
2937 t.replace('$$', '$')
2938 return t
2939
2940 # @initialization is a regular template like @media one
2941 # so it should be handled just the same way (see
2942 # https://github.com/ytdl-org/youtube-dl/issues/11605)
2943 if 'initialization' in representation_ms_info:
2944 initialization_template = prepare_template(
2945 'initialization',
2946 # As per [1, 5.3.9.4.2, Table 15, page 54] $Number$ and
2947 # $Time$ shall not be included for @initialization thus
2948 # only $Bandwidth$ remains
2949 ('Bandwidth', ))
2950 representation_ms_info['initialization_url'] = initialization_template % {
2951 'Bandwidth': bandwidth,
2952 }
2953
2954 def location_key(location):
2955 return 'url' if re.match(r'^https?://', location) else 'path'
2956
2957 if 'segment_urls' not in representation_ms_info and 'media' in representation_ms_info:
2958
2959 media_template = prepare_template('media', ('Number', 'Bandwidth', 'Time'))
2960 media_location_key = location_key(media_template)
2961
2962 # As per [1, 5.3.9.4.4, Table 16, page 55] $Number$ and $Time$
2963 # can't be used at the same time
2964 if '%(Number' in media_template and 's' not in representation_ms_info:
2965 segment_duration = None
2966 if 'total_number' not in representation_ms_info and 'segment_duration' in representation_ms_info:
2967 segment_duration = float_or_none(representation_ms_info['segment_duration'], representation_ms_info['timescale'])
ffa89477 2968 representation_ms_info['total_number'] = int(math.ceil(
2969 float_or_none(period_duration, segment_duration, default=0)))
be2fc5b2 2970 representation_ms_info['fragments'] = [{
2971 media_location_key: media_template % {
2972 'Number': segment_number,
2973 'Bandwidth': bandwidth,
2974 },
2975 'duration': segment_duration,
2976 } for segment_number in range(
2977 representation_ms_info['start_number'],
2978 representation_ms_info['total_number'] + representation_ms_info['start_number'])]
2979 else:
2980 # $Number*$ or $Time$ in media template with S list available
2981 # Example $Number*$: http://www.svtplay.se/klipp/9023742/stopptid-om-bjorn-borg
2982 # Example $Time$: https://play.arkena.com/embed/avp/v2/player/media/b41dda37-d8e7-4d3f-b1b5-9a9db578bdfe/1/129411
2983 representation_ms_info['fragments'] = []
2984 segment_time = 0
2985 segment_d = None
2986 segment_number = representation_ms_info['start_number']
2987
2988 def add_segment_url():
2989 segment_url = media_template % {
2990 'Time': segment_time,
2991 'Bandwidth': bandwidth,
2992 'Number': segment_number,
2993 }
2994 representation_ms_info['fragments'].append({
2995 media_location_key: segment_url,
2996 'duration': float_or_none(segment_d, representation_ms_info['timescale']),
2997 })
2998
2999 for num, s in enumerate(representation_ms_info['s']):
3000 segment_time = s.get('t') or segment_time
3001 segment_d = s['d']
3002 add_segment_url()
3003 segment_number += 1
3004 for r in range(s.get('r', 0)):
3005 segment_time += segment_d
f0948348 3006 add_segment_url()
b4c1d6e8 3007 segment_number += 1
be2fc5b2 3008 segment_time += segment_d
3009 elif 'segment_urls' in representation_ms_info and 's' in representation_ms_info:
62b58c09
L
3010 # No media template,
3011 # e.g. https://www.youtube.com/watch?v=iXZV5uAYMJI
be2fc5b2 3012 # or any YouTube dashsegments video
3013 fragments = []
3014 segment_index = 0
3015 timescale = representation_ms_info['timescale']
3016 for s in representation_ms_info['s']:
3017 duration = float_or_none(s['d'], timescale)
3018 for r in range(s.get('r', 0) + 1):
3019 segment_uri = representation_ms_info['segment_urls'][segment_index]
3020 fragments.append({
3021 location_key(segment_uri): segment_uri,
3022 'duration': duration,
3023 })
3024 segment_index += 1
3025 representation_ms_info['fragments'] = fragments
3026 elif 'segment_urls' in representation_ms_info:
3027 # Segment URLs with no SegmentTimeline
62b58c09 3028 # E.g. https://www.seznam.cz/zpravy/clanek/cesko-zasahne-vitr-o-sile-vichrice-muze-byt-i-zivotu-nebezpecny-39091
be2fc5b2 3029 # https://github.com/ytdl-org/youtube-dl/pull/14844
3030 fragments = []
3031 segment_duration = float_or_none(
3032 representation_ms_info['segment_duration'],
3033 representation_ms_info['timescale']) if 'segment_duration' in representation_ms_info else None
3034 for segment_url in representation_ms_info['segment_urls']:
3035 fragment = {
3036 location_key(segment_url): segment_url,
3037 }
3038 if segment_duration:
3039 fragment['duration'] = segment_duration
3040 fragments.append(fragment)
3041 representation_ms_info['fragments'] = fragments
3042 # If there is a fragments key available then we correctly recognized fragmented media.
3043 # Otherwise we will assume unfragmented media with direct access. Technically, such
3044 # assumption is not necessarily correct since we may simply have no support for
3045 # some forms of fragmented media renditions yet, but for now we'll use this fallback.
3046 if 'fragments' in representation_ms_info:
3047 f.update({
3048 # NB: mpd_url may be empty when MPD manifest is parsed from a string
3049 'url': mpd_url or base_url,
3050 'fragment_base_url': base_url,
3051 'fragments': [],
3052 'protocol': 'http_dash_segments' if mime_type != 'image/jpeg' else 'mhtml',
3053 })
3054 if 'initialization_url' in representation_ms_info:
3055 initialization_url = representation_ms_info['initialization_url']
3056 if not f.get('url'):
3057 f['url'] = initialization_url
3058 f['fragments'].append({location_key(initialization_url): initialization_url})
3059 f['fragments'].extend(representation_ms_info['fragments'])
ffa89477 3060 if not period_duration:
3061 period_duration = try_get(
3062 representation_ms_info,
3063 lambda r: sum(frag['duration'] for frag in r['fragments']), float)
17b598d3 3064 else:
be2fc5b2 3065 # Assuming direct URL to unfragmented media.
3066 f['url'] = base_url
234416e4 3067 if content_type in ('video', 'audio', 'image/jpeg'):
3068 f['manifest_stream_number'] = stream_numbers[f['url']]
3069 stream_numbers[f['url']] += 1
be2fc5b2 3070 formats.append(f)
3071 elif content_type == 'text':
3072 subtitles.setdefault(lang or 'und', []).append(f)
3073
171e59ed 3074 return formats, subtitles
17b598d3 3075
fd76a142
F
3076 def _extract_ism_formats(self, *args, **kwargs):
3077 fmts, subs = self._extract_ism_formats_and_subtitles(*args, **kwargs)
3078 if subs:
b5ae35ee 3079 self._report_ignoring_subs('ISM')
fd76a142
F
3080 return fmts
3081
3082 def _extract_ism_formats_and_subtitles(self, ism_url, video_id, ism_id=None, note=None, errnote=None, fatal=True, data=None, headers={}, query={}):
47a5cb77 3083 res = self._download_xml_handle(
b2758123 3084 ism_url, video_id,
37a3bb66 3085 note='Downloading ISM manifest' if note is None else note,
3086 errnote='Failed to download ISM manifest' if errnote is None else errnote,
7360c06f 3087 fatal=fatal, data=data, headers=headers, query=query)
b2758123 3088 if res is False:
fd76a142 3089 return [], {}
47a5cb77 3090 ism_doc, urlh = res
13b08034 3091 if ism_doc is None:
fd76a142 3092 return [], {}
b2758123 3093
fd76a142 3094 return self._parse_ism_formats_and_subtitles(ism_doc, urlh.geturl(), ism_id)
b2758123 3095
fd76a142 3096 def _parse_ism_formats_and_subtitles(self, ism_doc, ism_url, ism_id=None):
76d5a363
S
3097 """
3098 Parse formats from ISM manifest.
3099 References:
3100 1. [MS-SSTR]: Smooth Streaming Protocol,
3101 https://msdn.microsoft.com/en-us/library/ff469518.aspx
3102 """
06869367 3103 if ism_doc.get('IsLive') == 'TRUE':
fd76a142 3104 return [], {}
b2758123 3105
b2758123
RA
3106 duration = int(ism_doc.attrib['Duration'])
3107 timescale = int_or_none(ism_doc.get('TimeScale')) or 10000000
3108
3109 formats = []
fd76a142 3110 subtitles = {}
b2758123
RA
3111 for stream in ism_doc.findall('StreamIndex'):
3112 stream_type = stream.get('Type')
fd76a142 3113 if stream_type not in ('video', 'audio', 'text'):
b2758123
RA
3114 continue
3115 url_pattern = stream.attrib['Url']
3116 stream_timescale = int_or_none(stream.get('TimeScale')) or timescale
3117 stream_name = stream.get('Name')
fd76a142 3118 stream_language = stream.get('Language', 'und')
b2758123 3119 for track in stream.findall('QualityLevel'):
e2efe599 3120 fourcc = track.get('FourCC') or ('AACL' if track.get('AudioTag') == '255' else None)
b2758123 3121 # TODO: add support for WVC1 and WMAP
66a1b864 3122 if fourcc not in ('H264', 'AVC1', 'AACL', 'TTML'):
b2758123
RA
3123 self.report_warning('%s is not a supported codec' % fourcc)
3124 continue
3125 tbr = int(track.attrib['Bitrate']) // 1000
76d5a363
S
3126 # [1] does not mention Width and Height attributes. However,
3127 # they're often present while MaxWidth and MaxHeight are
3128 # missing, so should be used as fallbacks
3129 width = int_or_none(track.get('MaxWidth') or track.get('Width'))
3130 height = int_or_none(track.get('MaxHeight') or track.get('Height'))
b2758123
RA
3131 sampling_rate = int_or_none(track.get('SamplingRate'))
3132
3133 track_url_pattern = re.sub(r'{[Bb]itrate}', track.attrib['Bitrate'], url_pattern)
14f25df2 3134 track_url_pattern = urllib.parse.urljoin(ism_url, track_url_pattern)
b2758123
RA
3135
3136 fragments = []
3137 fragment_ctx = {
3138 'time': 0,
3139 }
3140 stream_fragments = stream.findall('c')
3141 for stream_fragment_index, stream_fragment in enumerate(stream_fragments):
3142 fragment_ctx['time'] = int_or_none(stream_fragment.get('t')) or fragment_ctx['time']
3143 fragment_repeat = int_or_none(stream_fragment.get('r')) or 1
3144 fragment_ctx['duration'] = int_or_none(stream_fragment.get('d'))
3145 if not fragment_ctx['duration']:
3146 try:
3147 next_fragment_time = int(stream_fragment[stream_fragment_index + 1].attrib['t'])
3148 except IndexError:
3149 next_fragment_time = duration
1616f9b4 3150 fragment_ctx['duration'] = (next_fragment_time - fragment_ctx['time']) / fragment_repeat
b2758123
RA
3151 for _ in range(fragment_repeat):
3152 fragments.append({
14f25df2 3153 'url': re.sub(r'{start[ _]time}', str(fragment_ctx['time']), track_url_pattern),
b2758123
RA
3154 'duration': fragment_ctx['duration'] / stream_timescale,
3155 })
3156 fragment_ctx['time'] += fragment_ctx['duration']
3157
fd76a142
F
3158 if stream_type == 'text':
3159 subtitles.setdefault(stream_language, []).append({
3160 'ext': 'ismt',
3161 'protocol': 'ism',
3162 'url': ism_url,
3163 'manifest_url': ism_url,
3164 'fragments': fragments,
3165 '_download_params': {
3166 'stream_type': stream_type,
3167 'duration': duration,
3168 'timescale': stream_timescale,
3169 'fourcc': fourcc,
3170 'language': stream_language,
3171 'codec_private_data': track.get('CodecPrivateData'),
3172 }
3173 })
3174 elif stream_type in ('video', 'audio'):
3175 formats.append({
34921b43 3176 'format_id': join_nonempty(ism_id, stream_name, tbr),
fd76a142
F
3177 'url': ism_url,
3178 'manifest_url': ism_url,
3179 'ext': 'ismv' if stream_type == 'video' else 'isma',
3180 'width': width,
3181 'height': height,
3182 'tbr': tbr,
3183 'asr': sampling_rate,
3184 'vcodec': 'none' if stream_type == 'audio' else fourcc,
3185 'acodec': 'none' if stream_type == 'video' else fourcc,
3186 'protocol': 'ism',
3187 'fragments': fragments,
88acdbc2 3188 'has_drm': ism_doc.find('Protection') is not None,
fd76a142
F
3189 '_download_params': {
3190 'stream_type': stream_type,
3191 'duration': duration,
3192 'timescale': stream_timescale,
3193 'width': width or 0,
3194 'height': height or 0,
3195 'fourcc': fourcc,
3196 'language': stream_language,
3197 'codec_private_data': track.get('CodecPrivateData'),
3198 'sampling_rate': sampling_rate,
3199 'channels': int_or_none(track.get('Channels', 2)),
3200 'bits_per_sample': int_or_none(track.get('BitsPerSample', 16)),
3201 'nal_unit_length_field': int_or_none(track.get('NALUnitLengthField', 4)),
3202 },
3203 })
3204 return formats, subtitles
b2758123 3205
079a7cfc 3206 def _parse_html5_media_entries(self, base_url, webpage, video_id, m3u8_id=None, m3u8_entry_protocol='m3u8_native', mpd_id=None, preference=None, quality=None):
6780154e
S
3207 def absolute_url(item_url):
3208 return urljoin(base_url, item_url)
59bbe491 3209
3210 def parse_content_type(content_type):
3211 if not content_type:
3212 return {}
3213 ctr = re.search(r'(?P<mimetype>[^/]+/[^;]+)(?:;\s*codecs="?(?P<codecs>[^"]+))?', content_type)
3214 if ctr:
3215 mimetype, codecs = ctr.groups()
3216 f = parse_codecs(codecs)
3217 f['ext'] = mimetype2ext(mimetype)
3218 return f
3219 return {}
3220
222a2308
L
3221 def _media_formats(src, cur_media_type, type_info=None):
3222 type_info = type_info or {}
520251c0 3223 full_url = absolute_url(src)
82889d4a 3224 ext = type_info.get('ext') or determine_ext(full_url)
87a449c1 3225 if ext == 'm3u8':
520251c0
YCH
3226 is_plain_url = False
3227 formats = self._extract_m3u8_formats(
ad120ae1 3228 full_url, video_id, ext='mp4',
eeb0a956 3229 entry_protocol=m3u8_entry_protocol, m3u8_id=m3u8_id,
f983b875 3230 preference=preference, quality=quality, fatal=False)
87a449c1
S
3231 elif ext == 'mpd':
3232 is_plain_url = False
3233 formats = self._extract_mpd_formats(
b359e977 3234 full_url, video_id, mpd_id=mpd_id, fatal=False)
520251c0
YCH
3235 else:
3236 is_plain_url = True
3237 formats = [{
3238 'url': full_url,
3239 'vcodec': 'none' if cur_media_type == 'audio' else None,
222a2308 3240 'ext': ext,
520251c0
YCH
3241 }]
3242 return is_plain_url, formats
3243
59bbe491 3244 entries = []
4328ddf8 3245 # amp-video and amp-audio are very similar to their HTML5 counterparts
962ffcf8 3246 # so we will include them right here (see
4328ddf8 3247 # https://www.ampproject.org/docs/reference/components/amp-video)
29f7c58a 3248 # For dl8-* tags see https://delight-vr.com/documentation/dl8-video/
3249 _MEDIA_TAG_NAME_RE = r'(?:(?:amp|dl8(?:-live)?)-)?(video|audio)'
3250 media_tags = [(media_tag, media_tag_name, media_type, '')
3251 for media_tag, media_tag_name, media_type
3252 in re.findall(r'(?s)(<(%s)[^>]*/>)' % _MEDIA_TAG_NAME_RE, webpage)]
2aec7256
S
3253 media_tags.extend(re.findall(
3254 # We only allow video|audio followed by a whitespace or '>'.
3255 # Allowing more characters may end up in significant slow down (see
62b58c09
L
3256 # https://github.com/ytdl-org/youtube-dl/issues/11979,
3257 # e.g. http://www.porntrex.com/maps/videositemap.xml).
29f7c58a 3258 r'(?s)(<(?P<tag>%s)(?:\s+[^>]*)?>)(.*?)</(?P=tag)>' % _MEDIA_TAG_NAME_RE, webpage))
3259 for media_tag, _, media_type, media_content in media_tags:
59bbe491 3260 media_info = {
3261 'formats': [],
3262 'subtitles': {},
3263 }
3264 media_attributes = extract_attributes(media_tag)
f856816b 3265 src = strip_or_none(media_attributes.get('src'))
59bbe491 3266 if src:
222a2308
L
3267 f = parse_content_type(media_attributes.get('type'))
3268 _, formats = _media_formats(src, media_type, f)
520251c0 3269 media_info['formats'].extend(formats)
6780154e 3270 media_info['thumbnail'] = absolute_url(media_attributes.get('poster'))
59bbe491 3271 if media_content:
3272 for source_tag in re.findall(r'<source[^>]+>', media_content):
d493f15c
S
3273 s_attr = extract_attributes(source_tag)
3274 # data-video-src and data-src are non standard but seen
3275 # several times in the wild
f856816b 3276 src = strip_or_none(dict_get(s_attr, ('src', 'data-video-src', 'data-src')))
59bbe491 3277 if not src:
3278 continue
d493f15c 3279 f = parse_content_type(s_attr.get('type'))
868f79db 3280 is_plain_url, formats = _media_formats(src, media_type, f)
520251c0 3281 if is_plain_url:
d493f15c
S
3282 # width, height, res, label and title attributes are
3283 # all not standard but seen several times in the wild
3284 labels = [
3285 s_attr.get(lbl)
3286 for lbl in ('label', 'title')
3287 if str_or_none(s_attr.get(lbl))
3288 ]
3289 width = int_or_none(s_attr.get('width'))
3089bc74
S
3290 height = (int_or_none(s_attr.get('height'))
3291 or int_or_none(s_attr.get('res')))
d493f15c
S
3292 if not width or not height:
3293 for lbl in labels:
3294 resolution = parse_resolution(lbl)
3295 if not resolution:
3296 continue
3297 width = width or resolution.get('width')
3298 height = height or resolution.get('height')
3299 for lbl in labels:
3300 tbr = parse_bitrate(lbl)
3301 if tbr:
3302 break
3303 else:
3304 tbr = None
1ed45499 3305 f.update({
d493f15c
S
3306 'width': width,
3307 'height': height,
3308 'tbr': tbr,
3309 'format_id': s_attr.get('label') or s_attr.get('title'),
1ed45499 3310 })
520251c0
YCH
3311 f.update(formats[0])
3312 media_info['formats'].append(f)
3313 else:
3314 media_info['formats'].extend(formats)
59bbe491 3315 for track_tag in re.findall(r'<track[^>]+>', media_content):
3316 track_attributes = extract_attributes(track_tag)
3317 kind = track_attributes.get('kind')
5968d7d2 3318 if not kind or kind in ('subtitles', 'captions'):
f856816b 3319 src = strip_or_none(track_attributes.get('src'))
59bbe491 3320 if not src:
3321 continue
3322 lang = track_attributes.get('srclang') or track_attributes.get('lang') or track_attributes.get('label')
3323 media_info['subtitles'].setdefault(lang, []).append({
3324 'url': absolute_url(src),
3325 })
5e8e2fa5
S
3326 for f in media_info['formats']:
3327 f.setdefault('http_headers', {})['Referer'] = base_url
5968d7d2 3328 if media_info['formats'] or media_info['subtitles']:
59bbe491 3329 entries.append(media_info)
3330 return entries
3331
f6a1d69a
F
3332 def _extract_akamai_formats(self, *args, **kwargs):
3333 fmts, subs = self._extract_akamai_formats_and_subtitles(*args, **kwargs)
3334 if subs:
b5ae35ee 3335 self._report_ignoring_subs('akamai')
f6a1d69a
F
3336 return fmts
3337
3338 def _extract_akamai_formats_and_subtitles(self, manifest_url, video_id, hosts={}):
29f7c58a 3339 signed = 'hdnea=' in manifest_url
3340 if not signed:
3341 # https://learn.akamai.com/en-us/webhelp/media-services-on-demand/stream-packaging-user-guide/GUID-BE6C0F73-1E06-483B-B0EA-57984B91B7F9.html
3342 manifest_url = re.sub(
3343 r'(?:b=[\d,-]+|(?:__a__|attributes)=off|__b__=\d+)&?',
3344 '', manifest_url).strip('?')
3345
c7c43a93 3346 formats = []
f6a1d69a 3347 subtitles = {}
70c5802b 3348
e71a4509 3349 hdcore_sign = 'hdcore=3.7.0'
ff6f9a67 3350 f4m_url = re.sub(r'(https?://[^/]+)/i/', r'\1/z/', manifest_url).replace('/master.m3u8', '/manifest.f4m')
c4251b9a
RA
3351 hds_host = hosts.get('hds')
3352 if hds_host:
3353 f4m_url = re.sub(r'(https?://)[^/]+', r'\1' + hds_host, f4m_url)
e71a4509
RA
3354 if 'hdcore=' not in f4m_url:
3355 f4m_url += ('&' if '?' in f4m_url else '?') + hdcore_sign
3356 f4m_formats = self._extract_f4m_formats(
3357 f4m_url, video_id, f4m_id='hds', fatal=False)
3358 for entry in f4m_formats:
3359 entry.update({'extra_param_to_segment_url': hdcore_sign})
3360 formats.extend(f4m_formats)
70c5802b 3361
c4251b9a
RA
3362 m3u8_url = re.sub(r'(https?://[^/]+)/z/', r'\1/i/', manifest_url).replace('/manifest.f4m', '/master.m3u8')
3363 hls_host = hosts.get('hls')
3364 if hls_host:
3365 m3u8_url = re.sub(r'(https?://)[^/]+', r'\1' + hls_host, m3u8_url)
f6a1d69a 3366 m3u8_formats, m3u8_subtitles = self._extract_m3u8_formats_and_subtitles(
c7c43a93 3367 m3u8_url, video_id, 'mp4', 'm3u8_native',
29f7c58a 3368 m3u8_id='hls', fatal=False)
3369 formats.extend(m3u8_formats)
f6a1d69a 3370 subtitles = self._merge_subtitles(subtitles, m3u8_subtitles)
70c5802b 3371
3372 http_host = hosts.get('http')
29f7c58a 3373 if http_host and m3u8_formats and not signed:
3374 REPL_REGEX = r'https?://[^/]+/i/([^,]+),([^/]+),([^/]+)\.csmil/.+'
70c5802b 3375 qualities = re.match(REPL_REGEX, m3u8_url).group(2).split(',')
3376 qualities_length = len(qualities)
29f7c58a 3377 if len(m3u8_formats) in (qualities_length, qualities_length + 1):
70c5802b 3378 i = 0
29f7c58a 3379 for f in m3u8_formats:
3380 if f['vcodec'] != 'none':
70c5802b 3381 for protocol in ('http', 'https'):
3382 http_f = f.copy()
3383 del http_f['manifest_url']
3384 http_url = re.sub(
86e5f3ed 3385 REPL_REGEX, protocol + fr'://{http_host}/\g<1>{qualities[i]}\3', f['url'])
70c5802b 3386 http_f.update({
3387 'format_id': http_f['format_id'].replace('hls-', protocol + '-'),
3388 'url': http_url,
3389 'protocol': protocol,
3390 })
29f7c58a 3391 formats.append(http_f)
70c5802b 3392 i += 1
70c5802b 3393
f6a1d69a 3394 return formats, subtitles
c7c43a93 3395
6ad02195 3396 def _extract_wowza_formats(self, url, video_id, m3u8_entry_protocol='m3u8_native', skip_protocols=[]):
14f25df2 3397 query = urllib.parse.urlparse(url).query
6ad02195 3398 url = re.sub(r'/(?:manifest|playlist|jwplayer)\.(?:m3u8|f4m|mpd|smil)', '', url)
240f2622
S
3399 mobj = re.search(
3400 r'(?:(?:http|rtmp|rtsp)(?P<s>s)?:)?(?P<url>//[^?]+)', url)
3401 url_base = mobj.group('url')
3402 http_base_url = '%s%s:%s' % ('http', mobj.group('s') or '', url_base)
6ad02195 3403 formats = []
044eeb14
S
3404
3405 def manifest_url(manifest):
86e5f3ed 3406 m_url = f'{http_base_url}/{manifest}'
044eeb14
S
3407 if query:
3408 m_url += '?%s' % query
3409 return m_url
3410
6ad02195
RA
3411 if 'm3u8' not in skip_protocols:
3412 formats.extend(self._extract_m3u8_formats(
044eeb14 3413 manifest_url('playlist.m3u8'), video_id, 'mp4',
6ad02195
RA
3414 m3u8_entry_protocol, m3u8_id='hls', fatal=False))
3415 if 'f4m' not in skip_protocols:
3416 formats.extend(self._extract_f4m_formats(
044eeb14 3417 manifest_url('manifest.f4m'),
6ad02195 3418 video_id, f4m_id='hds', fatal=False))
0384932e
RA
3419 if 'dash' not in skip_protocols:
3420 formats.extend(self._extract_mpd_formats(
044eeb14 3421 manifest_url('manifest.mpd'),
0384932e 3422 video_id, mpd_id='dash', fatal=False))
6ad02195 3423 if re.search(r'(?:/smil:|\.smil)', url_base):
6ad02195
RA
3424 if 'smil' not in skip_protocols:
3425 rtmp_formats = self._extract_smil_formats(
044eeb14 3426 manifest_url('jwplayer.smil'),
6ad02195
RA
3427 video_id, fatal=False)
3428 for rtmp_format in rtmp_formats:
3429 rtsp_format = rtmp_format.copy()
3430 rtsp_format['url'] = '%s/%s' % (rtmp_format['url'], rtmp_format['play_path'])
3431 del rtsp_format['play_path']
3432 del rtsp_format['ext']
3433 rtsp_format.update({
3434 'url': rtsp_format['url'].replace('rtmp://', 'rtsp://'),
3435 'format_id': rtmp_format['format_id'].replace('rtmp', 'rtsp'),
3436 'protocol': 'rtsp',
3437 })
3438 formats.extend([rtmp_format, rtsp_format])
3439 else:
3440 for protocol in ('rtmp', 'rtsp'):
3441 if protocol not in skip_protocols:
3442 formats.append({
86e5f3ed 3443 'url': f'{protocol}:{url_base}',
6ad02195
RA
3444 'format_id': protocol,
3445 'protocol': protocol,
3446 })
3447 return formats
3448
c73e330e 3449 def _find_jwplayer_data(self, webpage, video_id=None, transform_source=js_to_json):
a4a554a7 3450 mobj = re.search(
ac9c69ac 3451 r'(?s)jwplayer\((?P<quote>[\'"])[^\'" ]+(?P=quote)\)(?!</script>).*?\.setup\s*\((?P<options>[^)]+)\)',
a4a554a7
YCH
3452 webpage)
3453 if mobj:
c73e330e
RU
3454 try:
3455 jwplayer_data = self._parse_json(mobj.group('options'),
3456 video_id=video_id,
3457 transform_source=transform_source)
3458 except ExtractorError:
3459 pass
3460 else:
3461 if isinstance(jwplayer_data, dict):
3462 return jwplayer_data
a4a554a7
YCH
3463
3464 def _extract_jwplayer_data(self, webpage, video_id, *args, **kwargs):
c73e330e
RU
3465 jwplayer_data = self._find_jwplayer_data(
3466 webpage, video_id, transform_source=js_to_json)
a4a554a7
YCH
3467 return self._parse_jwplayer_data(
3468 jwplayer_data, video_id, *args, **kwargs)
3469
3470 def _parse_jwplayer_data(self, jwplayer_data, video_id=None, require_title=True,
3471 m3u8_id=None, mpd_id=None, rtmp_params=None, base_url=None):
3472 # JWPlayer backward compatibility: flattened playlists
3473 # https://github.com/jwplayer/jwplayer/blob/v7.4.3/src/js/api/config.js#L81-L96
3474 if 'playlist' not in jwplayer_data:
3475 jwplayer_data = {'playlist': [jwplayer_data]}
3476
3477 entries = []
3478
3479 # JWPlayer backward compatibility: single playlist item
3480 # https://github.com/jwplayer/jwplayer/blob/v7.7.0/src/js/playlist/playlist.js#L10
3481 if not isinstance(jwplayer_data['playlist'], list):
3482 jwplayer_data['playlist'] = [jwplayer_data['playlist']]
3483
3484 for video_data in jwplayer_data['playlist']:
3485 # JWPlayer backward compatibility: flattened sources
3486 # https://github.com/jwplayer/jwplayer/blob/v7.4.3/src/js/playlist/item.js#L29-L35
3487 if 'sources' not in video_data:
3488 video_data['sources'] = [video_data]
3489
3490 this_video_id = video_id or video_data['mediaid']
3491
1a2192cb
S
3492 formats = self._parse_jwplayer_formats(
3493 video_data['sources'], video_id=this_video_id, m3u8_id=m3u8_id,
3494 mpd_id=mpd_id, rtmp_params=rtmp_params, base_url=base_url)
a4a554a7
YCH
3495
3496 subtitles = {}
3497 tracks = video_data.get('tracks')
3498 if tracks and isinstance(tracks, list):
3499 for track in tracks:
96a2daa1
S
3500 if not isinstance(track, dict):
3501 continue
f4b74272 3502 track_kind = track.get('kind')
14f25df2 3503 if not track_kind or not isinstance(track_kind, str):
f4b74272
S
3504 continue
3505 if track_kind.lower() not in ('captions', 'subtitles'):
a4a554a7
YCH
3506 continue
3507 track_url = urljoin(base_url, track.get('file'))
3508 if not track_url:
3509 continue
3510 subtitles.setdefault(track.get('label') or 'en', []).append({
3511 'url': self._proto_relative_url(track_url)
3512 })
3513
50d808f5 3514 entry = {
a4a554a7 3515 'id': this_video_id,
50d808f5 3516 'title': unescapeHTML(video_data['title'] if require_title else video_data.get('title')),
f81dd65b 3517 'description': clean_html(video_data.get('description')),
6945b9e7 3518 'thumbnail': urljoin(base_url, self._proto_relative_url(video_data.get('image'))),
a4a554a7
YCH
3519 'timestamp': int_or_none(video_data.get('pubdate')),
3520 'duration': float_or_none(jwplayer_data.get('duration') or video_data.get('duration')),
3521 'subtitles': subtitles,
50d808f5
RA
3522 }
3523 # https://github.com/jwplayer/jwplayer/blob/master/src/js/utils/validator.js#L32
3524 if len(formats) == 1 and re.search(r'^(?:http|//).*(?:youtube\.com|youtu\.be)/.+', formats[0]['url']):
3525 entry.update({
3526 '_type': 'url_transparent',
3527 'url': formats[0]['url'],
3528 })
3529 else:
3530 self._sort_formats(formats)
3531 entry['formats'] = formats
3532 entries.append(entry)
a4a554a7
YCH
3533 if len(entries) == 1:
3534 return entries[0]
3535 else:
3536 return self.playlist_result(entries)
3537
ed0cf9b3
S
3538 def _parse_jwplayer_formats(self, jwplayer_sources_data, video_id=None,
3539 m3u8_id=None, mpd_id=None, rtmp_params=None, base_url=None):
bf1b87cd 3540 urls = []
ed0cf9b3 3541 formats = []
1a2192cb 3542 for source in jwplayer_sources_data:
0a268c6e
S
3543 if not isinstance(source, dict):
3544 continue
6945b9e7
RA
3545 source_url = urljoin(
3546 base_url, self._proto_relative_url(source.get('file')))
3547 if not source_url or source_url in urls:
bf1b87cd
RA
3548 continue
3549 urls.append(source_url)
ed0cf9b3
S
3550 source_type = source.get('type') or ''
3551 ext = mimetype2ext(source_type) or determine_ext(source_url)
3552 if source_type == 'hls' or ext == 'm3u8':
3553 formats.extend(self._extract_m3u8_formats(
0236cd0d
S
3554 source_url, video_id, 'mp4', entry_protocol='m3u8_native',
3555 m3u8_id=m3u8_id, fatal=False))
0d9c48de 3556 elif source_type == 'dash' or ext == 'mpd':
ed0cf9b3
S
3557 formats.extend(self._extract_mpd_formats(
3558 source_url, video_id, mpd_id=mpd_id, fatal=False))
b51dc9db
S
3559 elif ext == 'smil':
3560 formats.extend(self._extract_smil_formats(
3561 source_url, video_id, fatal=False))
ed0cf9b3 3562 # https://github.com/jwplayer/jwplayer/blob/master/src/js/providers/default.js#L67
0236cd0d
S
3563 elif source_type.startswith('audio') or ext in (
3564 'oga', 'aac', 'mp3', 'mpeg', 'vorbis'):
ed0cf9b3
S
3565 formats.append({
3566 'url': source_url,
3567 'vcodec': 'none',
3568 'ext': ext,
3569 })
3570 else:
3571 height = int_or_none(source.get('height'))
3572 if height is None:
3573 # Often no height is provided but there is a label in
0236cd0d 3574 # format like "1080p", "720p SD", or 1080.
ed0cf9b3 3575 height = int_or_none(self._search_regex(
14f25df2 3576 r'^(\d{3,4})[pP]?(?:\b|$)', str(source.get('label') or ''),
ed0cf9b3
S
3577 'height', default=None))
3578 a_format = {
3579 'url': source_url,
3580 'width': int_or_none(source.get('width')),
3581 'height': height,
0236cd0d 3582 'tbr': int_or_none(source.get('bitrate')),
ed0cf9b3
S
3583 'ext': ext,
3584 }
3585 if source_url.startswith('rtmp'):
3586 a_format['ext'] = 'flv'
ed0cf9b3
S
3587 # See com/longtailvideo/jwplayer/media/RTMPMediaProvider.as
3588 # of jwplayer.flash.swf
3589 rtmp_url_parts = re.split(
3590 r'((?:mp4|mp3|flv):)', source_url, 1)
3591 if len(rtmp_url_parts) == 3:
3592 rtmp_url, prefix, play_path = rtmp_url_parts
3593 a_format.update({
3594 'url': rtmp_url,
3595 'play_path': prefix + play_path,
3596 })
3597 if rtmp_params:
3598 a_format.update(rtmp_params)
3599 formats.append(a_format)
3600 return formats
3601
f4b1c7ad 3602 def _live_title(self, name):
39ca3b5c 3603 self._downloader.deprecation_warning('yt_dlp.InfoExtractor._live_title is deprecated and does not work as expected')
3604 return name
f4b1c7ad 3605
b14f3a4c
PH
3606 def _int(self, v, name, fatal=False, **kwargs):
3607 res = int_or_none(v, **kwargs)
b14f3a4c 3608 if res is None:
86e5f3ed 3609 msg = f'Failed to extract {name}: Could not parse value {v!r}'
b14f3a4c
PH
3610 if fatal:
3611 raise ExtractorError(msg)
3612 else:
6a39ee13 3613 self.report_warning(msg)
b14f3a4c
PH
3614 return res
3615
3616 def _float(self, v, name, fatal=False, **kwargs):
3617 res = float_or_none(v, **kwargs)
3618 if res is None:
86e5f3ed 3619 msg = f'Failed to extract {name}: Could not parse value {v!r}'
b14f3a4c
PH
3620 if fatal:
3621 raise ExtractorError(msg)
3622 else:
6a39ee13 3623 self.report_warning(msg)
b14f3a4c
PH
3624 return res
3625
40e41780
TF
3626 def _set_cookie(self, domain, name, value, expire_time=None, port=None,
3627 path='/', secure=False, discard=False, rest={}, **kwargs):
ac668111 3628 cookie = http.cookiejar.Cookie(
4ed2d7b7 3629 0, name, value, port, port is not None, domain, True,
40e41780
TF
3630 domain.startswith('.'), path, True, secure, expire_time,
3631 discard, None, None, rest)
9809740b 3632 self.cookiejar.set_cookie(cookie)
42939b61 3633
799207e8 3634 def _get_cookies(self, url):
ac668111 3635 """ Return a http.cookies.SimpleCookie with the cookies for the url """
3636 return http.cookies.SimpleCookie(self._downloader._calc_cookies(url))
799207e8 3637
e3c1266f 3638 def _apply_first_set_cookie_header(self, url_handle, cookie):
ce2fe4c0
S
3639 """
3640 Apply first Set-Cookie header instead of the last. Experimental.
3641
3642 Some sites (e.g. [1-3]) may serve two cookies under the same name
3643 in Set-Cookie header and expect the first (old) one to be set rather
3644 than second (new). However, as of RFC6265 the newer one cookie
3645 should be set into cookie store what actually happens.
3646 We will workaround this issue by resetting the cookie to
3647 the first one manually.
3648 1. https://new.vk.com/
3649 2. https://github.com/ytdl-org/youtube-dl/issues/9841#issuecomment-227871201
3650 3. https://learning.oreilly.com/
3651 """
e3c1266f
S
3652 for header, cookies in url_handle.headers.items():
3653 if header.lower() != 'set-cookie':
3654 continue
cfb0511d 3655 cookies = cookies.encode('iso-8859-1').decode('utf-8')
e3c1266f
S
3656 cookie_value = re.search(
3657 r'%s=(.+?);.*?\b[Dd]omain=(.+?)(?:[,;]|$)' % cookie, cookies)
3658 if cookie_value:
3659 value, domain = cookie_value.groups()
3660 self._set_cookie(domain, cookie, value)
3661 break
3662
82d02080 3663 @classmethod
3664 def get_testcases(cls, include_onlymatching=False):
3665 t = getattr(cls, '_TEST', None)
05900629 3666 if t:
82d02080 3667 assert not hasattr(cls, '_TESTS'), f'{cls.ie_key()}IE has _TEST and _TESTS'
05900629
PH
3668 tests = [t]
3669 else:
82d02080 3670 tests = getattr(cls, '_TESTS', [])
05900629
PH
3671 for t in tests:
3672 if not include_onlymatching and t.get('only_matching', False):
3673 continue
82d02080 3674 t['name'] = cls.ie_key()
05900629
PH
3675 yield t
3676
f2e8dbcc 3677 @classmethod
3678 def get_webpage_testcases(cls):
3679 tests = getattr(cls, '_WEBPAGE_TESTS', [])
3680 for t in tests:
3681 t['name'] = cls.ie_key()
3682 return tests
3683
24146491 3684 @classproperty
3685 def age_limit(cls):
3686 """Get age limit from the testcases"""
3687 return max(traverse_obj(
f2e8dbcc 3688 (*cls.get_testcases(include_onlymatching=False), *cls.get_webpage_testcases()),
24146491 3689 (..., (('playlist', 0), None), 'info_dict', 'age_limit')) or [0])
3690
82d02080 3691 @classmethod
3692 def is_suitable(cls, age_limit):
24146491 3693 """Test whether the extractor is generally suitable for the given age limit"""
3694 return not age_restricted(cls.age_limit, age_limit)
05900629 3695
82d02080 3696 @classmethod
3697 def description(cls, *, markdown=True, search_examples=None):
8dcce6a8 3698 """Description of the extractor"""
3699 desc = ''
82d02080 3700 if cls._NETRC_MACHINE:
8dcce6a8 3701 if markdown:
82d02080 3702 desc += f' [<abbr title="netrc machine"><em>{cls._NETRC_MACHINE}</em></abbr>]'
8dcce6a8 3703 else:
82d02080 3704 desc += f' [{cls._NETRC_MACHINE}]'
3705 if cls.IE_DESC is False:
8dcce6a8 3706 desc += ' [HIDDEN]'
82d02080 3707 elif cls.IE_DESC:
3708 desc += f' {cls.IE_DESC}'
3709 if cls.SEARCH_KEY:
3710 desc += f'; "{cls.SEARCH_KEY}:" prefix'
8dcce6a8 3711 if search_examples:
3712 _COUNTS = ('', '5', '10', 'all')
62b58c09 3713 desc += f' (e.g. "{cls.SEARCH_KEY}{random.choice(_COUNTS)}:{random.choice(search_examples)}")'
82d02080 3714 if not cls.working():
8dcce6a8 3715 desc += ' (**Currently broken**)' if markdown else ' (Currently broken)'
3716
82d02080 3717 name = f' - **{cls.IE_NAME}**' if markdown else cls.IE_NAME
8dcce6a8 3718 return f'{name}:{desc}' if desc else name
3719
a504ced0 3720 def extract_subtitles(self, *args, **kwargs):
a06916d9 3721 if (self.get_param('writesubtitles', False)
3722 or self.get_param('listsubtitles')):
9868ea49
JMF
3723 return self._get_subtitles(*args, **kwargs)
3724 return {}
a504ced0
JMF
3725
3726 def _get_subtitles(self, *args, **kwargs):
611c1dd9 3727 raise NotImplementedError('This method must be implemented by subclasses')
a504ced0 3728
a2160aa4 3729 def extract_comments(self, *args, **kwargs):
3730 if not self.get_param('getcomments'):
3731 return None
3732 generator = self._get_comments(*args, **kwargs)
3733
3734 def extractor():
3735 comments = []
d2b2fca5 3736 interrupted = True
a2160aa4 3737 try:
3738 while True:
3739 comments.append(next(generator))
a2160aa4 3740 except StopIteration:
3741 interrupted = False
d2b2fca5 3742 except KeyboardInterrupt:
3743 self.to_screen('Interrupted by user')
3744 except Exception as e:
3745 if self.get_param('ignoreerrors') is not True:
3746 raise
3747 self._downloader.report_error(e)
a2160aa4 3748 comment_count = len(comments)
3749 self.to_screen(f'Extracted {comment_count} comments')
3750 return {
3751 'comments': comments,
3752 'comment_count': None if interrupted else comment_count
3753 }
3754 return extractor
3755
3756 def _get_comments(self, *args, **kwargs):
3757 raise NotImplementedError('This method must be implemented by subclasses')
3758
912e0b7e
YCH
3759 @staticmethod
3760 def _merge_subtitle_items(subtitle_list1, subtitle_list2):
a825ffbf 3761 """ Merge subtitle items for one language. Items with duplicated URLs/data
912e0b7e 3762 will be dropped. """
86e5f3ed 3763 list1_data = {(item.get('url'), item.get('data')) for item in subtitle_list1}
912e0b7e 3764 ret = list(subtitle_list1)
a44ca5a4 3765 ret.extend(item for item in subtitle_list2 if (item.get('url'), item.get('data')) not in list1_data)
912e0b7e
YCH
3766 return ret
3767
3768 @classmethod
46890374 3769 def _merge_subtitles(cls, *dicts, target=None):
19bb3920 3770 """ Merge subtitle dictionaries, language by language. """
19bb3920
F
3771 if target is None:
3772 target = {}
3773 for d in dicts:
3774 for lang, subs in d.items():
3775 target[lang] = cls._merge_subtitle_items(target.get(lang, []), subs)
3776 return target
912e0b7e 3777
360e1ca5 3778 def extract_automatic_captions(self, *args, **kwargs):
a06916d9 3779 if (self.get_param('writeautomaticsub', False)
3780 or self.get_param('listsubtitles')):
9868ea49
JMF
3781 return self._get_automatic_captions(*args, **kwargs)
3782 return {}
360e1ca5
JMF
3783
3784 def _get_automatic_captions(self, *args, **kwargs):
611c1dd9 3785 raise NotImplementedError('This method must be implemented by subclasses')
360e1ca5 3786
2762dbb1 3787 @functools.cached_property
24146491 3788 def _cookies_passed(self):
3789 """Whether cookies have been passed to YoutubeDL"""
3790 return self.get_param('cookiefile') is not None or self.get_param('cookiesfrombrowser') is not None
3791
d77ab8e2 3792 def mark_watched(self, *args, **kwargs):
1813a6cc 3793 if not self.get_param('mark_watched', False):
3794 return
24146491 3795 if self.supports_login() and self._get_login_info()[0] is not None or self._cookies_passed:
d77ab8e2
S
3796 self._mark_watched(*args, **kwargs)
3797
3798 def _mark_watched(self, *args, **kwargs):
3799 raise NotImplementedError('This method must be implemented by subclasses')
3800
38cce791
YCH
3801 def geo_verification_headers(self):
3802 headers = {}
a06916d9 3803 geo_verification_proxy = self.get_param('geo_verification_proxy')
38cce791
YCH
3804 if geo_verification_proxy:
3805 headers['Ytdl-request-proxy'] = geo_verification_proxy
3806 return headers
3807
8f97a15d 3808 @staticmethod
3809 def _generic_id(url):
14f25df2 3810 return urllib.parse.unquote(os.path.splitext(url.rstrip('/').split('/')[-1])[0])
98763ee3 3811
8f97a15d 3812 @staticmethod
3813 def _generic_title(url):
14f25df2 3814 return urllib.parse.unquote(os.path.splitext(url_basename(url))[0])
98763ee3 3815
c224251a 3816 @staticmethod
b0089e89 3817 def _availability(is_private=None, needs_premium=None, needs_subscription=None, needs_auth=None, is_unlisted=None):
c224251a
M
3818 all_known = all(map(
3819 lambda x: x is not None,
3820 (is_private, needs_premium, needs_subscription, needs_auth, is_unlisted)))
3821 return (
3822 'private' if is_private
3823 else 'premium_only' if needs_premium
3824 else 'subscriber_only' if needs_subscription
3825 else 'needs_auth' if needs_auth
3826 else 'unlisted' if is_unlisted
3827 else 'public' if all_known
3828 else None)
3829
d43de682 3830 def _configuration_arg(self, key, default=NO_DEFAULT, *, ie_key=None, casesense=False):
4bb6b02f 3831 '''
3832 @returns A list of values for the extractor argument given by "key"
3833 or "default" if no such key is present
3834 @param default The default value to return when the key is not present (default: [])
3835 @param casesense When false, the values are converted to lower case
3836 '''
3837 val = traverse_obj(
d43de682 3838 self._downloader.params, ('extractor_args', (ie_key or self.ie_key()).lower(), key))
4bb6b02f 3839 if val is None:
3840 return [] if default is NO_DEFAULT else default
3841 return list(val) if casesense else [x.lower() for x in val]
5d3a0e79 3842
f40ee5e9 3843 def _yes_playlist(self, playlist_id, video_id, smuggled_data=None, *, playlist_label='playlist', video_label='video'):
3844 if not playlist_id or not video_id:
3845 return not video_id
3846
3847 no_playlist = (smuggled_data or {}).get('force_noplaylist')
3848 if no_playlist is not None:
3849 return not no_playlist
3850
3851 video_id = '' if video_id is True else f' {video_id}'
3852 playlist_id = '' if playlist_id is True else f' {playlist_id}'
3853 if self.get_param('noplaylist'):
3854 self.to_screen(f'Downloading just the {video_label}{video_id} because of --no-playlist')
3855 return False
3856 self.to_screen(f'Downloading {playlist_label}{playlist_id} - add --no-playlist to download just the {video_label}{video_id}')
3857 return True
3858
be5c1ae8 3859 def _error_or_warning(self, err, _count=None, _retries=0, *, fatal=True):
3860 RetryManager.report_retry(err, _count or int(fatal), _retries, info=self.to_screen, warn=self.report_warning,
3861 sleep_func=self.get_param('retry_sleep_functions', {}).get('extractor'))
3862
3863 def RetryManager(self, **kwargs):
3864 return RetryManager(self.get_param('extractor_retries', 3), self._error_or_warning, **kwargs)
3865
8f97a15d 3866 @classmethod
3867 def extract_from_webpage(cls, ydl, url, webpage):
3868 ie = (cls if isinstance(cls._extract_from_webpage, types.MethodType)
3869 else ydl.get_info_extractor(cls.ie_key()))
f2e8dbcc 3870 for info in ie._extract_from_webpage(url, webpage) or []:
3871 # url = None since we do not want to set (webpage/original)_url
3872 ydl.add_default_extra_info(info, ie, None)
3873 yield info
8f97a15d 3874
3875 @classmethod
3876 def _extract_from_webpage(cls, url, webpage):
3877 for embed_url in orderedSet(
3878 cls._extract_embed_urls(url, webpage) or [], lazy=True):
3879 yield cls.url_result(embed_url, cls)
3880
3881 @classmethod
3882 def _extract_embed_urls(cls, url, webpage):
3883 """@returns all the embed urls on the webpage"""
3884 if '_EMBED_URL_RE' not in cls.__dict__:
3885 assert isinstance(cls._EMBED_REGEX, (list, tuple))
3886 for idx, regex in enumerate(cls._EMBED_REGEX):
3887 assert regex.count('(?P<url>') == 1, \
3888 f'{cls.__name__}._EMBED_REGEX[{idx}] must have exactly 1 url group\n\t{regex}'
3889 cls._EMBED_URL_RE = tuple(map(re.compile, cls._EMBED_REGEX))
3890
3891 for regex in cls._EMBED_URL_RE:
3892 for mobj in regex.finditer(webpage):
3893 embed_url = urllib.parse.urljoin(url, unescapeHTML(mobj.group('url')))
3894 if cls._VALID_URL is False or cls.suitable(embed_url):
3895 yield embed_url
3896
3897 class StopExtraction(Exception):
3898 pass
3899
bfd973ec 3900 @classmethod
3901 def _extract_url(cls, webpage): # TODO: Remove
3902 """Only for compatibility with some older extractors"""
3903 return next(iter(cls._extract_embed_urls(None, webpage) or []), None)
3904
8dbe9899 3905
d6983cb4
PH
3906class SearchInfoExtractor(InfoExtractor):
3907 """
3908 Base class for paged search queries extractors.
10952eb2 3909 They accept URLs in the format _SEARCH_KEY(|all|[0-9]):{query}
96565c7e 3910 Instances should define _SEARCH_KEY and optionally _MAX_RESULTS
d6983cb4
PH
3911 """
3912
96565c7e 3913 _MAX_RESULTS = float('inf')
3914
8f97a15d 3915 @classproperty
3916 def _VALID_URL(cls):
d6983cb4
PH
3917 return r'%s(?P<prefix>|[1-9][0-9]*|all):(?P<query>[\s\S]+)' % cls._SEARCH_KEY
3918
d6983cb4 3919 def _real_extract(self, query):
2c4aaadd 3920 prefix, query = self._match_valid_url(query).group('prefix', 'query')
d6983cb4
PH
3921 if prefix == '':
3922 return self._get_n_results(query, 1)
3923 elif prefix == 'all':
3924 return self._get_n_results(query, self._MAX_RESULTS)
3925 else:
3926 n = int(prefix)
3927 if n <= 0:
86e5f3ed 3928 raise ExtractorError(f'invalid download number {n} for query "{query}"')
d6983cb4 3929 elif n > self._MAX_RESULTS:
6a39ee13 3930 self.report_warning('%s returns max %i results (you requested %i)' % (self._SEARCH_KEY, self._MAX_RESULTS, n))
d6983cb4
PH
3931 n = self._MAX_RESULTS
3932 return self._get_n_results(query, n)
3933
3934 def _get_n_results(self, query, n):
cc16383f 3935 """Get a specified number of results for a query.
3936 Either this function or _search_results must be overridden by subclasses """
3937 return self.playlist_result(
3938 itertools.islice(self._search_results(query), 0, None if n == float('inf') else n),
3939 query, query)
3940
3941 def _search_results(self, query):
3942 """Returns an iterator of search results"""
611c1dd9 3943 raise NotImplementedError('This method must be implemented by subclasses')
0f818663 3944
82d02080 3945 @classproperty
3946 def SEARCH_KEY(cls):
3947 return cls._SEARCH_KEY
fe7866d0 3948
3949
3950class UnsupportedURLIE(InfoExtractor):
3951 _VALID_URL = '.*'
3952 _ENABLED = False
3953 IE_DESC = False
3954
3955 def _real_extract(self, url):
3956 raise UnsupportedError(url)