]> jfr.im git - yt-dlp.git/blame - yt_dlp/extractor/common.py
[extractor/youtube:tab] Let `approximate_date` return timestamp
[yt-dlp.git] / yt_dlp / extractor / common.py
CommitLineData
d6983cb4 1import base64
234416e4 2import collections
ac668111 3import getpass
3ec05685 4import hashlib
54007a45 5import http.client
6import http.cookiejar
7import http.cookies
2314b4d8 8import inspect
cc16383f 9import itertools
3d3538e4 10import json
f8271158 11import math
4094b6e3 12import netrc
d6983cb4 13import os
773f291d 14import random
6929b41a 15import re
d6983cb4 16import sys
4094b6e3 17import time
8f97a15d 18import types
14f25df2 19import urllib.parse
ac668111 20import urllib.request
f8271158 21import xml.etree.ElementTree
d6983cb4 22
6929b41a 23from ..compat import functools # isort: split
14f25df2 24from ..compat import compat_etree_fromstring, compat_expanduser, compat_os_name
8817a80d 25from ..cookies import LenientSimpleCookie
eb8a4433 26from ..downloader import FileDownloader
f8271158 27from ..downloader.f4m import get_base_url, remove_encrypted_media
8c25f81b 28from ..utils import (
8f97a15d 29 IDENTITY,
f8271158 30 JSON_LD_RE,
31 NO_DEFAULT,
32 ExtractorError,
33 GeoRestrictedError,
34 GeoUtils,
b7c47b74 35 LenientJSONDecoder,
f8271158 36 RegexNotFoundError,
be5c1ae8 37 RetryManager,
f8271158 38 UnsupportedError,
05900629 39 age_restricted,
02dc0a36 40 base_url,
08f2a92c 41 bug_reports_message,
82d02080 42 classproperty,
d6983cb4 43 clean_html,
70f0f5a8 44 determine_ext,
46b18f23 45 determine_protocol,
d493f15c 46 dict_get,
42676437 47 encode_data_uri,
9b9c5355 48 error_to_compat_str,
46b18f23 49 extract_attributes,
90137ca4 50 filter_dict,
97f4aecf 51 fix_xml_ampersands,
b14f3a4c 52 float_or_none,
b868936c 53 format_field,
31bb8d3f 54 int_or_none,
34921b43 55 join_nonempty,
a4a554a7 56 js_to_json,
46b18f23 57 mimetype2ext,
3158150c 58 network_exceptions,
46b18f23 59 orderedSet,
d493f15c 60 parse_bitrate,
46b18f23
JH
61 parse_codecs,
62 parse_duration,
4ca2a3cf 63 parse_iso8601,
46b18f23 64 parse_m3u8_attributes,
d493f15c 65 parse_resolution,
46b18f23 66 sanitize_filename,
8f97a15d 67 sanitize_url,
b868936c 68 sanitized_Request,
ade1fa70 69 smuggle_url,
d493f15c 70 str_or_none,
ce5b9040 71 str_to_int,
f856816b 72 strip_or_none,
5d3a0e79 73 traverse_obj,
47046464 74 try_call,
ffa89477 75 try_get,
f38de77f 76 unescapeHTML,
647eab45 77 unified_strdate,
6b3a3098 78 unified_timestamp,
46b18f23 79 update_Request,
09d02ea4 80 update_url_query,
a107193e 81 url_basename,
bebef109 82 url_or_none,
b868936c 83 urljoin,
6606817a 84 variadic,
a6571f10 85 xpath_element,
8d6765cf
S
86 xpath_text,
87 xpath_with_ns,
d6983cb4 88)
c342041f 89
d6983cb4 90
86e5f3ed 91class InfoExtractor:
d6983cb4
PH
92 """Information Extractor class.
93
94 Information extractors are the classes that, given a URL, extract
95 information about the video (or videos) the URL refers to. This
96 information includes the real video URL, the video title, author and
97 others. The information is stored in a dictionary which is then
5d380852 98 passed to the YoutubeDL. The YoutubeDL processes this
d6983cb4
PH
99 information possibly downloading the video to the file system, among
100 other possible outcomes.
101
cf0649f8 102 The type field determines the type of the result.
fed5d032
PH
103 By far the most common value (and the default if _type is missing) is
104 "video", which indicates a single video.
105
106 For a video, the dictionaries must include the following fields:
d6983cb4
PH
107
108 id: Video identifier.
d4736fdb 109 title: Video title, unescaped. Set to an empty string if video has
110 no title as opposed to "None" which signifies that the
111 extractor failed to obtain a title
d67b0b15 112
f49d89ee 113 Additionally, it must contain either a formats entry or a url one:
d67b0b15 114
f49d89ee
PH
115 formats: A list of dictionaries for each format available, ordered
116 from worst to best quality.
117
118 Potential fields:
c790e93a
S
119 * url The mandatory URL representing the media:
120 for plain file media - HTTP URL of this file,
121 for RTMP - RTMP URL,
122 for HLS - URL of the M3U8 media playlist,
123 for HDS - URL of the F4M manifest,
79d2077e
S
124 for DASH
125 - HTTP URL to plain file media (in case of
126 unfragmented media)
127 - URL of the MPD manifest or base URL
128 representing the media if MPD manifest
8ed7a233 129 is parsed from a string (in case of
79d2077e 130 fragmented media)
c790e93a 131 for MSS - URL of the ISM manifest.
86f4d14f
S
132 * manifest_url
133 The URL of the manifest file in case of
c790e93a
S
134 fragmented media:
135 for HLS - URL of the M3U8 master playlist,
136 for HDS - URL of the F4M manifest,
137 for DASH - URL of the MPD manifest,
138 for MSS - URL of the ISM manifest.
a44ca5a4 139 * manifest_stream_number (For internal use only)
140 The index of the stream in the manifest file
10952eb2 141 * ext Will be calculated from URL if missing
d67b0b15
PH
142 * format A human-readable description of the format
143 ("mp4 container with h264/opus").
144 Calculated from the format_id, width, height.
145 and format_note fields if missing.
146 * format_id A short description of the format
5d4f3985
PH
147 ("mp4_h264_opus" or "19").
148 Technically optional, but strongly recommended.
d67b0b15
PH
149 * format_note Additional info about the format
150 ("3D" or "DASH video")
151 * width Width of the video, if known
152 * height Height of the video, if known
f49d89ee 153 * resolution Textual description of width and height
176f1866 154 * dynamic_range The dynamic range of the video. One of:
155 "SDR" (None), "HDR10", "HDR10+, "HDR12", "HLG, "DV"
7217e148 156 * tbr Average bitrate of audio and video in KBit/s
d67b0b15
PH
157 * abr Average audio bitrate in KBit/s
158 * acodec Name of the audio codec in use
dd27fd17 159 * asr Audio sampling rate in Hertz
b8ed0f15 160 * audio_channels Number of audio channels
d67b0b15 161 * vbr Average video bitrate in KBit/s
fbb21cf5 162 * fps Frame rate
d67b0b15 163 * vcodec Name of the video codec in use
1394ce65 164 * container Name of the container format
d67b0b15 165 * filesize The number of bytes, if known in advance
9732d77e 166 * filesize_approx An estimate for the number of bytes
d67b0b15 167 * player_url SWF Player URL (used for rtmpdump).
c7deaa4c 168 * protocol The protocol that will be used for the actual
adbc4ec4
THD
169 download, lower-case. One of "http", "https" or
170 one of the protocols defined in downloader.PROTOCOL_MAP
c58c2d63
S
171 * fragment_base_url
172 Base URL for fragments. Each fragment's path
173 value (if present) will be relative to
174 this URL.
175 * fragments A list of fragments of a fragmented media.
176 Each fragment entry must contain either an url
177 or a path. If an url is present it should be
178 considered by a client. Otherwise both path and
179 fragment_base_url must be present. Here is
180 the list of all potential fields:
181 * "url" - fragment's URL
182 * "path" - fragment's path relative to
183 fragment_base_url
a0d5077c
S
184 * "duration" (optional, int or float)
185 * "filesize" (optional, int)
adbc4ec4
THD
186 * is_from_start Is a live format that can be downloaded
187 from the start. Boolean
f49d89ee 188 * preference Order number of this format. If this field is
08d13955 189 present and not None, the formats get sorted
38d63d84 190 by this field, regardless of all other values.
f49d89ee
PH
191 -1 for default (order by other properties),
192 -2 or smaller for less than default.
e65566a9
PH
193 < -1000 to hide the format (if there is
194 another one which is strictly better)
32f90364
PH
195 * language Language code, e.g. "de" or "en-US".
196 * language_preference Is this in the language mentioned in
197 the URL?
aff2f4f4
PH
198 10 if it's what the URL is about,
199 -1 for default (don't know),
200 -10 otherwise, other values reserved for now.
5d73273f
PH
201 * quality Order number of the video quality of this
202 format, irrespective of the file format.
203 -1 for default (order by other properties),
204 -2 or smaller for less than default.
c64ed2a3
PH
205 * source_preference Order number for this video source
206 (quality takes higher priority)
207 -1 for default (order by other properties),
208 -2 or smaller for less than default.
d769be6c
PH
209 * http_headers A dictionary of additional HTTP headers
210 to add to the request.
6271f1ca 211 * stretched_ratio If given and not 1, indicates that the
3dee7826
PH
212 video's pixels are not square.
213 width : height ratio as float.
214 * no_resume The server does not support resuming the
215 (HTTP or RTMP) download. Boolean.
88acdbc2 216 * has_drm The format has DRM and cannot be downloaded. Boolean
0a5a191a 217 * downloader_options A dictionary of downloader options
218 (For internal use only)
219 * http_chunk_size Chunk size for HTTP downloads
220 * ffmpeg_args Extra arguments for ffmpeg downloader
3b1fe47d 221 RTMP formats can also have the additional fields: page_url,
222 app, play_path, tc_url, flash_version, rtmp_live, rtmp_conn,
223 rtmp_protocol, rtmp_real_time
3dee7826 224
c0ba0f48 225 url: Final video URL.
d6983cb4 226 ext: Video filename extension.
d67b0b15
PH
227 format: The video format, defaults to ext (used for --get-format)
228 player_url: SWF Player URL (used for rtmpdump).
2f5865cc 229
d6983cb4
PH
230 The following fields are optional:
231
08d30158 232 direct: True if a direct video file was given (must only be set by GenericIE)
f5e43bc6 233 alt_title: A secondary title of the video.
0afef30b
PH
234 display_id An alternative identifier for the video, not necessarily
235 unique, but available before title. Typically, id is
236 something like "4234987", title "Dancing naked mole rats",
237 and display_id "dancing-naked-mole-rats"
d5519808 238 thumbnails: A list of dictionaries, with the following entries:
cfb56d1a 239 * "id" (optional, string) - Thumbnail format ID
d5519808 240 * "url"
cfb56d1a 241 * "preference" (optional, int) - quality of the image
d5519808
PH
242 * "width" (optional, int)
243 * "height" (optional, int)
5e1c39ac 244 * "resolution" (optional, string "{width}x{height}",
d5519808 245 deprecated)
2de624fd 246 * "filesize" (optional, int)
297e9952 247 * "http_headers" (dict) - HTTP headers for the request
d6983cb4 248 thumbnail: Full URL to a video thumbnail image.
f5e43bc6 249 description: Full video description.
d6983cb4 250 uploader: Full name of the video uploader.
2bc0c46f 251 license: License name the video is licensed under.
8a92e51c 252 creator: The creator of the video.
10db0d2f 253 timestamp: UNIX timestamp of the moment the video was uploaded
ae6a1b95 254 upload_date: Video upload date in UTC (YYYYMMDD).
f0d785d3 255 If not explicitly set, calculated from timestamp
256 release_timestamp: UNIX timestamp of the moment the video was released.
257 If it is not clear whether to use timestamp or this, use the former
ae6a1b95 258 release_date: The date (YYYYMMDD) when the video was released in UTC.
f0d785d3 259 If not explicitly set, calculated from release_timestamp
260 modified_timestamp: UNIX timestamp of the moment the video was last modified.
ae6a1b95 261 modified_date: The date (YYYYMMDD) when the video was last modified in UTC.
f0d785d3 262 If not explicitly set, calculated from modified_timestamp
d6983cb4 263 uploader_id: Nickname or id of the video uploader.
7bcd2830 264 uploader_url: Full URL to a personal webpage of the video uploader.
6f1f59f3 265 channel: Full name of the channel the video is uploaded on.
0e7b8d3e 266 Note that channel fields may or may not repeat uploader
6f1f59f3
S
267 fields. This depends on a particular extractor.
268 channel_id: Id of the channel.
269 channel_url: Full URL to a channel webpage.
6c73052c 270 channel_follower_count: Number of followers of the channel.
da9ec3b9 271 location: Physical location where the video was filmed.
a504ced0 272 subtitles: The available subtitles as a dictionary in the format
4606c34e
YCH
273 {tag: subformats}. "tag" is usually a language code, and
274 "subformats" is a list sorted from lower to higher
275 preference, each element is a dictionary with the "ext"
276 entry and one of:
a504ced0 277 * "data": The subtitles file contents
10952eb2 278 * "url": A URL pointing to the subtitles file
2412044c 279 It can optionally also have:
280 * "name": Name or description of the subtitles
08d30158 281 * "http_headers": A dictionary of additional HTTP headers
297e9952 282 to add to the request.
4bba3716 283 "ext" will be calculated from URL if missing
e167860c 284 automatic_captions: Like 'subtitles'; contains automatically generated
285 captions instead of normal subtitles
62d231c0 286 duration: Length of the video in seconds, as an integer or float.
f3d29461 287 view_count: How many users have watched the video on the platform.
867c66ff 288 concurrent_view_count: How many users are currently watching the video on the platform.
19e3dfc9
PH
289 like_count: Number of positive ratings of the video
290 dislike_count: Number of negative ratings of the video
02835c6b 291 repost_count: Number of reposts of the video
2d30521a 292 average_rating: Average rating give by users, the scale used depends on the webpage
19e3dfc9 293 comment_count: Number of comments on the video
dd622d7c
PH
294 comments: A list of comments, each with one or more of the following
295 properties (all but one of text or html optional):
296 * "author" - human-readable name of the comment author
297 * "author_id" - user ID of the comment author
a1c5d2ca 298 * "author_thumbnail" - The thumbnail of the comment author
dd622d7c
PH
299 * "id" - Comment ID
300 * "html" - Comment as HTML
301 * "text" - Plain text of the comment
302 * "timestamp" - UNIX timestamp of comment
303 * "parent" - ID of the comment this one is replying to.
304 Set to "root" to indicate that this is a
305 comment to the original video.
a1c5d2ca
M
306 * "like_count" - Number of positive ratings of the comment
307 * "dislike_count" - Number of negative ratings of the comment
308 * "is_favorited" - Whether the comment is marked as
309 favorite by the video uploader
310 * "author_is_uploader" - Whether the comment is made by
311 the video uploader
8dbe9899 312 age_limit: Age restriction for the video, as an integer (years)
7a5c1cfe 313 webpage_url: The URL to the video webpage, if given to yt-dlp it
9103bbc5
JMF
314 should allow to get the same result again. (It will be set
315 by YoutubeDL if it's missing)
ad3bc6ac
PH
316 categories: A list of categories that the video falls in, for example
317 ["Sports", "Berlin"]
864f24bd 318 tags: A list of tags assigned to the video, e.g. ["sweden", "pop music"]
d0fb4bd1 319 cast: A list of the video cast
7267bd53
PH
320 is_live: True, False, or None (=unknown). Whether this video is a
321 live stream that goes on instead of a fixed-length video.
f76ede8e 322 was_live: True, False, or None (=unknown). Whether this video was
323 originally a live stream.
0647d925 324 live_status: None (=unknown), 'is_live', 'is_upcoming', 'was_live', 'not_live',
e325a21a 325 or 'post_live' (was live, but VOD is not yet processed)
ae30b840 326 If absent, automatically set from is_live, was_live
7c80519c 327 start_time: Time in seconds where the reproduction should start, as
10952eb2 328 specified in the URL.
297a564b 329 end_time: Time in seconds where the reproduction should end, as
10952eb2 330 specified in the URL.
55949fed 331 chapters: A list of dictionaries, with the following entries:
332 * "start_time" - The start time of the chapter in seconds
333 * "end_time" - The end time of the chapter in seconds
334 * "title" (optional, string)
6cfda058 335 playable_in_embed: Whether this video is allowed to play in embedded
336 players on other sites. Can be True (=always allowed),
337 False (=never allowed), None (=unknown), or a string
62b58c09 338 specifying the criteria for embedability; e.g. 'whitelist'
c224251a
M
339 availability: Under what condition the video is available. One of
340 'private', 'premium_only', 'subscriber_only', 'needs_auth',
341 'unlisted' or 'public'. Use 'InfoExtractor._availability'
342 to set it
1e8fe57e 343 _old_archive_ids: A list of old archive ids needed for backward compatibility
277d6ff5 344 __post_extractor: A function to be called just before the metadata is
345 written to either disk, logger or console. The function
346 must return a dict which will be added to the info_dict.
347 This is usefull for additional information that is
348 time-consuming to extract. Note that the fields thus
349 extracted will not be available to output template and
350 match_filter. So, only "comments" and "comment_count" are
351 currently allowed to be extracted via this method.
d6983cb4 352
7109903e
S
353 The following fields should only be used when the video belongs to some logical
354 chapter or section:
355
356 chapter: Name or title of the chapter the video belongs to.
27bfd4e5
S
357 chapter_number: Number of the chapter the video belongs to, as an integer.
358 chapter_id: Id of the chapter the video belongs to, as a unicode string.
7109903e
S
359
360 The following fields should only be used when the video is an episode of some
8d76bdf1 361 series, programme or podcast:
7109903e
S
362
363 series: Title of the series or programme the video episode belongs to.
9ac24e23 364 series_id: Id of the series or programme the video episode belongs to, as a unicode string.
7109903e 365 season: Title of the season the video episode belongs to.
27bfd4e5
S
366 season_number: Number of the season the video episode belongs to, as an integer.
367 season_id: Id of the season the video episode belongs to, as a unicode string.
7109903e
S
368 episode: Title of the video episode. Unlike mandatory video title field,
369 this field should denote the exact title of the video episode
370 without any kind of decoration.
27bfd4e5
S
371 episode_number: Number of the video episode within a season, as an integer.
372 episode_id: Id of the video episode, as a unicode string.
7109903e 373
7a93ab5f
S
374 The following fields should only be used when the media is a track or a part of
375 a music album:
376
377 track: Title of the track.
378 track_number: Number of the track within an album or a disc, as an integer.
379 track_id: Id of the track (useful in case of custom indexing, e.g. 6.iii),
380 as a unicode string.
381 artist: Artist(s) of the track.
382 genre: Genre(s) of the track.
383 album: Title of the album the track belongs to.
384 album_type: Type of the album (e.g. "Demo", "Full-length", "Split", "Compilation", etc).
385 album_artist: List of all artists appeared on the album (e.g.
386 "Ash Borer / Fell Voices" or "Various Artists", useful for splits
387 and compilations).
388 disc_number: Number of the disc or other physical medium the track belongs to,
389 as an integer.
390 release_year: Year (YYYY) when the album was released.
8bcd4048 391 composer: Composer of the piece
7a93ab5f 392
3975b4d2 393 The following fields should only be set for clips that should be cut from the original video:
394
395 section_start: Start time of the section in seconds
396 section_end: End time of the section in seconds
397
45e8a04e 398 The following fields should only be set for storyboards:
399 rows: Number of rows in each storyboard fragment, as an integer
400 columns: Number of columns in each storyboard fragment, as an integer
401
deefc05b 402 Unless mentioned otherwise, the fields should be Unicode strings.
d6983cb4 403
d838b1bd
PH
404 Unless mentioned otherwise, None is equivalent to absence of information.
405
fed5d032
PH
406
407 _type "playlist" indicates multiple videos.
b82f815f
PH
408 There must be a key "entries", which is a list, an iterable, or a PagedList
409 object, each element of which is a valid dictionary by this specification.
fed5d032 410
962ffcf8 411 Additionally, playlists can have "id", "title", and any other relevant
b60419c5 412 attributes with the same semantics as videos (see above).
fed5d032 413
f0d785d3 414 It can also have the following optional fields:
415
416 playlist_count: The total number of videos in a playlist. If not given,
417 YoutubeDL tries to calculate it from "entries"
418
fed5d032
PH
419
420 _type "multi_video" indicates that there are multiple videos that
421 form a single show, for examples multiple acts of an opera or TV episode.
422 It must have an entries key like a playlist and contain all the keys
423 required for a video at the same time.
424
425
426 _type "url" indicates that the video must be extracted from another
427 location, possibly by a different extractor. Its only required key is:
428 "url" - the next URL to extract.
f58766ce
PH
429 The key "ie_key" can be set to the class name (minus the trailing "IE",
430 e.g. "Youtube") if the extractor class is known in advance.
431 Additionally, the dictionary may have any properties of the resolved entity
432 known in advance, for example "title" if the title of the referred video is
fed5d032
PH
433 known ahead of time.
434
435
436 _type "url_transparent" entities have the same specification as "url", but
437 indicate that the given additional information is more precise than the one
438 associated with the resolved URL.
439 This is useful when a site employs a video service that hosts the video and
440 its technical metadata, but that video service does not embed a useful
441 title, description etc.
442
443
8f97a15d 444 Subclasses of this should also be added to the list of extractors and
445 should define a _VALID_URL regexp and, re-define the _real_extract() and
446 (optionally) _real_initialize() methods.
d6983cb4 447
e6f21b3d 448 Subclasses may also override suitable() if necessary, but ensure the function
449 signature is preserved and that this function imports everything it needs
52efa4b3 450 (except other extractors), so that lazy_extractors works correctly.
451
8f97a15d 452 Subclasses can define a list of _EMBED_REGEX, which will be searched for in
453 the HTML of Generic webpages. It may also override _extract_embed_urls
454 or _extract_from_webpage as necessary. While these are normally classmethods,
455 _extract_from_webpage is allowed to be an instance method.
456
457 _extract_from_webpage may raise self.StopExtraction() to stop further
458 processing of the webpage and obtain exclusive rights to it. This is useful
62b58c09
L
459 when the extractor cannot reliably be matched using just the URL,
460 e.g. invidious/peertube instances
8f97a15d 461
462 Embed-only extractors can be defined by setting _VALID_URL = False.
463
52efa4b3 464 To support username + password (or netrc) login, the extractor must define a
465 _NETRC_MACHINE and re-define _perform_login(username, password) and
466 (optionally) _initialize_pre_login() methods. The _perform_login method will
467 be called between _initialize_pre_login and _real_initialize if credentials
468 are passed by the user. In cases where it is necessary to have the login
469 process as part of the extraction rather than initialization, _perform_login
470 can be left undefined.
e6f21b3d 471
4248dad9 472 _GEO_BYPASS attribute may be set to False in order to disable
773f291d
S
473 geo restriction bypass mechanisms for a particular extractor.
474 Though it won't disable explicit geo restriction bypass based on
504f20dd 475 country code provided with geo_bypass_country.
4248dad9
S
476
477 _GEO_COUNTRIES attribute may contain a list of presumably geo unrestricted
478 countries for this extractor. One of these countries will be used by
479 geo restriction bypass mechanism right away in order to bypass
504f20dd 480 geo restriction, of course, if the mechanism is not disabled.
773f291d 481
5f95927a
S
482 _GEO_IP_BLOCKS attribute may contain a list of presumably geo unrestricted
483 IP blocks in CIDR notation for this extractor. One of these IP blocks
484 will be used by geo restriction bypass mechanism similarly
504f20dd 485 to _GEO_COUNTRIES.
3ccdde8c 486
fe7866d0 487 The _ENABLED attribute should be set to False for IEs that
488 are disabled by default and must be explicitly enabled.
489
e6f21b3d 490 The _WORKING attribute should be set to False for broken IEs
d6983cb4
PH
491 in order to warn the users and skip the tests.
492 """
493
494 _ready = False
495 _downloader = None
773f291d 496 _x_forwarded_for_ip = None
4248dad9
S
497 _GEO_BYPASS = True
498 _GEO_COUNTRIES = None
5f95927a 499 _GEO_IP_BLOCKS = None
d6983cb4 500 _WORKING = True
fe7866d0 501 _ENABLED = True
52efa4b3 502 _NETRC_MACHINE = None
231025c4 503 IE_DESC = None
8dcce6a8 504 SEARCH_KEY = None
8f97a15d 505 _VALID_URL = None
506 _EMBED_REGEX = []
d6983cb4 507
8dcce6a8 508 def _login_hint(self, method=NO_DEFAULT, netrc=None):
509 password_hint = f'--username and --password, or --netrc ({netrc or self._NETRC_MACHINE}) to provide account credentials'
510 return {
511 None: '',
512 'any': f'Use --cookies, --cookies-from-browser, {password_hint}',
513 'password': f'Use {password_hint}',
514 'cookies': (
515 'Use --cookies-from-browser or --cookies for the authentication. '
17ffed18 516 'See https://github.com/yt-dlp/yt-dlp/wiki/FAQ#how-do-i-pass-cookies-to-yt-dlp for how to manually pass cookies'),
8dcce6a8 517 }[method if method is not NO_DEFAULT else 'any' if self.supports_login() else 'cookies']
9d5d4d64 518
d6983cb4 519 def __init__(self, downloader=None):
49a57e70 520 """Constructor. Receives an optional downloader (a YoutubeDL instance).
521 If a downloader is not passed during initialization,
522 it must be set using "set_downloader()" before "extract()" is called"""
d6983cb4 523 self._ready = False
773f291d 524 self._x_forwarded_for_ip = None
28f436ba 525 self._printed_messages = set()
d6983cb4
PH
526 self.set_downloader(downloader)
527
528 @classmethod
5ad28e7f 529 def _match_valid_url(cls, url):
8f97a15d 530 if cls._VALID_URL is False:
531 return None
79cb2577
PH
532 # This does not use has/getattr intentionally - we want to know whether
533 # we have cached the regexp for *this* class, whereas getattr would also
534 # match the superclass
535 if '_VALID_URL_RE' not in cls.__dict__:
536 cls._VALID_URL_RE = re.compile(cls._VALID_URL)
5ad28e7f 537 return cls._VALID_URL_RE.match(url)
538
539 @classmethod
540 def suitable(cls, url):
541 """Receives a URL and returns True if suitable for this IE."""
3fb4e21b 542 # This function must import everything it needs (except other extractors),
543 # so that lazy_extractors works correctly
5ad28e7f 544 return cls._match_valid_url(url) is not None
d6983cb4 545
ed9266db
PH
546 @classmethod
547 def _match_id(cls, url):
5ad28e7f 548 return cls._match_valid_url(url).group('id')
ed9266db 549
1151c407 550 @classmethod
551 def get_temp_id(cls, url):
552 try:
553 return cls._match_id(url)
554 except (IndexError, AttributeError):
555 return None
556
d6983cb4
PH
557 @classmethod
558 def working(cls):
559 """Getter method for _WORKING."""
560 return cls._WORKING
561
52efa4b3 562 @classmethod
563 def supports_login(cls):
564 return bool(cls._NETRC_MACHINE)
565
d6983cb4
PH
566 def initialize(self):
567 """Initializes an instance (authentication, etc)."""
28f436ba 568 self._printed_messages = set()
5f95927a
S
569 self._initialize_geo_bypass({
570 'countries': self._GEO_COUNTRIES,
571 'ip_blocks': self._GEO_IP_BLOCKS,
572 })
4248dad9 573 if not self._ready:
52efa4b3 574 self._initialize_pre_login()
575 if self.supports_login():
576 username, password = self._get_login_info()
577 if username:
578 self._perform_login(username, password)
579 elif self.get_param('username') and False not in (self.IE_DESC, self._NETRC_MACHINE):
8dcce6a8 580 self.report_warning(f'Login with password is not supported for this website. {self._login_hint("cookies")}')
4248dad9
S
581 self._real_initialize()
582 self._ready = True
583
5f95927a 584 def _initialize_geo_bypass(self, geo_bypass_context):
e39b5d4a
S
585 """
586 Initialize geo restriction bypass mechanism.
587
588 This method is used to initialize geo bypass mechanism based on faking
589 X-Forwarded-For HTTP header. A random country from provided country list
dc0a869e 590 is selected and a random IP belonging to this country is generated. This
e39b5d4a
S
591 IP will be passed as X-Forwarded-For HTTP header in all subsequent
592 HTTP requests.
e39b5d4a
S
593
594 This method will be used for initial geo bypass mechanism initialization
5f95927a
S
595 during the instance initialization with _GEO_COUNTRIES and
596 _GEO_IP_BLOCKS.
e39b5d4a 597
5f95927a 598 You may also manually call it from extractor's code if geo bypass
e39b5d4a 599 information is not available beforehand (e.g. obtained during
5f95927a
S
600 extraction) or due to some other reason. In this case you should pass
601 this information in geo bypass context passed as first argument. It may
602 contain following fields:
603
604 countries: List of geo unrestricted countries (similar
605 to _GEO_COUNTRIES)
606 ip_blocks: List of geo unrestricted IP blocks in CIDR notation
607 (similar to _GEO_IP_BLOCKS)
608
e39b5d4a 609 """
773f291d 610 if not self._x_forwarded_for_ip:
5f95927a
S
611
612 # Geo bypass mechanism is explicitly disabled by user
a06916d9 613 if not self.get_param('geo_bypass', True):
5f95927a
S
614 return
615
616 if not geo_bypass_context:
617 geo_bypass_context = {}
618
619 # Backward compatibility: previously _initialize_geo_bypass
620 # expected a list of countries, some 3rd party code may still use
621 # it this way
622 if isinstance(geo_bypass_context, (list, tuple)):
623 geo_bypass_context = {
624 'countries': geo_bypass_context,
625 }
626
627 # The whole point of geo bypass mechanism is to fake IP
628 # as X-Forwarded-For HTTP header based on some IP block or
629 # country code.
630
631 # Path 1: bypassing based on IP block in CIDR notation
632
633 # Explicit IP block specified by user, use it right away
634 # regardless of whether extractor is geo bypassable or not
a06916d9 635 ip_block = self.get_param('geo_bypass_ip_block', None)
5f95927a
S
636
637 # Otherwise use random IP block from geo bypass context but only
638 # if extractor is known as geo bypassable
639 if not ip_block:
640 ip_blocks = geo_bypass_context.get('ip_blocks')
641 if self._GEO_BYPASS and ip_blocks:
642 ip_block = random.choice(ip_blocks)
643
644 if ip_block:
645 self._x_forwarded_for_ip = GeoUtils.random_ipv4(ip_block)
8a82af35 646 self.write_debug(f'Using fake IP {self._x_forwarded_for_ip} as X-Forwarded-For')
5f95927a
S
647 return
648
649 # Path 2: bypassing based on country code
650
651 # Explicit country code specified by user, use it right away
652 # regardless of whether extractor is geo bypassable or not
a06916d9 653 country = self.get_param('geo_bypass_country', None)
5f95927a
S
654
655 # Otherwise use random country code from geo bypass context but
656 # only if extractor is known as geo bypassable
657 if not country:
658 countries = geo_bypass_context.get('countries')
659 if self._GEO_BYPASS and countries:
660 country = random.choice(countries)
661
662 if country:
663 self._x_forwarded_for_ip = GeoUtils.random_ipv4(country)
0760b0a7 664 self._downloader.write_debug(
86e5f3ed 665 f'Using fake IP {self._x_forwarded_for_ip} ({country.upper()}) as X-Forwarded-For')
d6983cb4
PH
666
667 def extract(self, url):
668 """Extracts URL information and returns it in list of dicts."""
3a5bcd03 669 try:
773f291d
S
670 for _ in range(2):
671 try:
672 self.initialize()
a06916d9 673 self.write_debug('Extracting URL: %s' % url)
0016b84e 674 ie_result = self._real_extract(url)
07cce701 675 if ie_result is None:
676 return None
0016b84e
S
677 if self._x_forwarded_for_ip:
678 ie_result['__x_forwarded_for_ip'] = self._x_forwarded_for_ip
b79f9e30 679 subtitles = ie_result.get('subtitles') or {}
680 if 'no-live-chat' in self.get_param('compat_opts'):
681 for lang in ('live_chat', 'comments', 'danmaku'):
682 subtitles.pop(lang, None)
0016b84e 683 return ie_result
773f291d 684 except GeoRestrictedError as e:
4248dad9
S
685 if self.__maybe_fake_ip_and_retry(e.countries):
686 continue
773f291d 687 raise
0db3bae8 688 except UnsupportedError:
689 raise
1151c407 690 except ExtractorError as e:
0db3bae8 691 kwargs = {
692 'video_id': e.video_id or self.get_temp_id(url),
693 'ie': self.IE_NAME,
b69fd25c 694 'tb': e.traceback or sys.exc_info()[2],
0db3bae8 695 'expected': e.expected,
696 'cause': e.cause
697 }
698 if hasattr(e, 'countries'):
699 kwargs['countries'] = e.countries
7265a219 700 raise type(e)(e.orig_msg, **kwargs)
ac668111 701 except http.client.IncompleteRead as e:
1151c407 702 raise ExtractorError('A network error has occurred.', cause=e, expected=True, video_id=self.get_temp_id(url))
9650885b 703 except (KeyError, StopIteration) as e:
1151c407 704 raise ExtractorError('An extractor error has occurred.', cause=e, video_id=self.get_temp_id(url))
d6983cb4 705
4248dad9 706 def __maybe_fake_ip_and_retry(self, countries):
a06916d9 707 if (not self.get_param('geo_bypass_country', None)
3089bc74 708 and self._GEO_BYPASS
a06916d9 709 and self.get_param('geo_bypass', True)
3089bc74
S
710 and not self._x_forwarded_for_ip
711 and countries):
eea0716c
S
712 country_code = random.choice(countries)
713 self._x_forwarded_for_ip = GeoUtils.random_ipv4(country_code)
4248dad9
S
714 if self._x_forwarded_for_ip:
715 self.report_warning(
eea0716c
S
716 'Video is geo restricted. Retrying extraction with fake IP %s (%s) as X-Forwarded-For.'
717 % (self._x_forwarded_for_ip, country_code.upper()))
4248dad9
S
718 return True
719 return False
720
d6983cb4 721 def set_downloader(self, downloader):
08d30158 722 """Sets a YoutubeDL instance as the downloader for this IE."""
d6983cb4
PH
723 self._downloader = downloader
724
9809740b 725 @property
726 def cache(self):
727 return self._downloader.cache
728
729 @property
730 def cookiejar(self):
731 return self._downloader.cookiejar
732
52efa4b3 733 def _initialize_pre_login(self):
962ffcf8 734 """ Initialization before login. Redefine in subclasses."""
52efa4b3 735 pass
736
737 def _perform_login(self, username, password):
738 """ Login with username and password. Redefine in subclasses."""
739 pass
740
d6983cb4
PH
741 def _real_initialize(self):
742 """Real initialization process. Redefine in subclasses."""
743 pass
744
745 def _real_extract(self, url):
746 """Real extraction process. Redefine in subclasses."""
08d30158 747 raise NotImplementedError('This method must be implemented by subclasses')
d6983cb4 748
56c73665
JMF
749 @classmethod
750 def ie_key(cls):
751 """A string for getting the InfoExtractor with get_info_extractor"""
3fb4e21b 752 return cls.__name__[:-2]
56c73665 753
82d02080 754 @classproperty
755 def IE_NAME(cls):
756 return cls.__name__[:-2]
d6983cb4 757
d391b7e2
S
758 @staticmethod
759 def __can_accept_status_code(err, expected_status):
ac668111 760 assert isinstance(err, urllib.error.HTTPError)
d391b7e2
S
761 if expected_status is None:
762 return False
d391b7e2
S
763 elif callable(expected_status):
764 return expected_status(err.code) is True
765 else:
6606817a 766 return err.code in variadic(expected_status)
d391b7e2 767
c043c246 768 def _create_request(self, url_or_request, data=None, headers=None, query=None):
ac668111 769 if isinstance(url_or_request, urllib.request.Request):
09d02ea4 770 return update_Request(url_or_request, data=data, headers=headers, query=query)
771 if query:
772 url_or_request = update_url_query(url_or_request, query)
c043c246 773 return sanitized_Request(url_or_request, data, headers or {})
f95b9dee 774
c043c246 775 def _request_webpage(self, url_or_request, video_id, note=None, errnote=None, fatal=True, data=None, headers=None, query=None, expected_status=None):
d391b7e2
S
776 """
777 Return the response handle.
778
779 See _download_webpage docstring for arguments specification.
780 """
1cf376f5 781 if not self._downloader._first_webpage_request:
49a57e70 782 sleep_interval = self.get_param('sleep_interval_requests') or 0
1cf376f5 783 if sleep_interval > 0:
5ef7d9bd 784 self.to_screen('Sleeping %s seconds ...' % sleep_interval)
1cf376f5 785 time.sleep(sleep_interval)
786 else:
787 self._downloader._first_webpage_request = False
788
d6983cb4
PH
789 if note is None:
790 self.report_download_webpage(video_id)
791 elif note is not False:
7cc3570e 792 if video_id is None:
86e5f3ed 793 self.to_screen(str(note))
7cc3570e 794 else:
86e5f3ed 795 self.to_screen(f'{video_id}: {note}')
2132edaa
S
796
797 # Some sites check X-Forwarded-For HTTP header in order to figure out
798 # the origin of the client behind proxy. This allows bypassing geo
799 # restriction by faking this header's value to IP that belongs to some
800 # geo unrestricted country. We will do so once we encounter any
801 # geo restriction error.
802 if self._x_forwarded_for_ip:
c043c246 803 headers = (headers or {}).copy()
804 headers.setdefault('X-Forwarded-For', self._x_forwarded_for_ip)
2132edaa 805
d6983cb4 806 try:
f95b9dee 807 return self._downloader.urlopen(self._create_request(url_or_request, data, headers, query))
3158150c 808 except network_exceptions as err:
ac668111 809 if isinstance(err, urllib.error.HTTPError):
d391b7e2 810 if self.__can_accept_status_code(err, expected_status):
95e42d73
XDG
811 # Retain reference to error to prevent file object from
812 # being closed before it can be read. Works around the
813 # effects of <https://bugs.python.org/issue15002>
814 # introduced in Python 3.4.1.
815 err.fp._error = err
d391b7e2
S
816 return err.fp
817
aa94a6d3
PH
818 if errnote is False:
819 return False
d6983cb4 820 if errnote is None:
f1a9d64e 821 errnote = 'Unable to download webpage'
7f8b2714 822
86e5f3ed 823 errmsg = f'{errnote}: {error_to_compat_str(err)}'
7cc3570e 824 if fatal:
497d2fab 825 raise ExtractorError(errmsg, cause=err)
7cc3570e 826 else:
6a39ee13 827 self.report_warning(errmsg)
7cc3570e 828 return False
d6983cb4 829
1890fc63 830 def _download_webpage_handle(self, url_or_request, video_id, note=None, errnote=None, fatal=True,
831 encoding=None, data=None, headers={}, query={}, expected_status=None):
d391b7e2
S
832 """
833 Return a tuple (page content as string, URL handle).
834
617f658b 835 Arguments:
836 url_or_request -- plain text URL as a string or
ac668111 837 a urllib.request.Request object
617f658b 838 video_id -- Video/playlist/item identifier (string)
839
840 Keyword arguments:
841 note -- note printed before downloading (string)
842 errnote -- note printed in case of an error (string)
843 fatal -- flag denoting whether error should be considered fatal,
844 i.e. whether it should cause ExtractionError to be raised,
845 otherwise a warning will be reported and extraction continued
846 encoding -- encoding for a page content decoding, guessed automatically
847 when not explicitly specified
848 data -- POST data (bytes)
849 headers -- HTTP headers (dict)
850 query -- URL query (dict)
851 expected_status -- allows to accept failed HTTP requests (non 2xx
852 status code) by explicitly specifying a set of accepted status
853 codes. Can be any of the following entities:
854 - an integer type specifying an exact failed status code to
855 accept
856 - a list or a tuple of integer types specifying a list of
857 failed status codes to accept
858 - a callable accepting an actual failed status code and
859 returning True if it should be accepted
860 Note that this argument does not affect success status codes (2xx)
861 which are always accepted.
d391b7e2 862 """
617f658b 863
b9d3e163 864 # Strip hashes from the URL (#1038)
14f25df2 865 if isinstance(url_or_request, str):
b9d3e163
PH
866 url_or_request = url_or_request.partition('#')[0]
867
d391b7e2 868 urlh = self._request_webpage(url_or_request, video_id, note, errnote, fatal, data=data, headers=headers, query=query, expected_status=expected_status)
7cc3570e
PH
869 if urlh is False:
870 assert not fatal
871 return False
c9a77969 872 content = self._webpage_read_content(urlh, url_or_request, video_id, note, errnote, fatal, encoding=encoding)
23be51d8
PH
873 return (content, urlh)
874
c9a77969
YCH
875 @staticmethod
876 def _guess_encoding_from_content(content_type, webpage_bytes):
d6983cb4
PH
877 m = re.match(r'[a-zA-Z0-9_.-]+/[a-zA-Z0-9_.-]+\s*;\s*charset=(.+)', content_type)
878 if m:
879 encoding = m.group(1)
880 else:
0d75ae2c 881 m = re.search(br'<meta[^>]+charset=[\'"]?([^\'")]+)[ /\'">]',
f143d86a
PH
882 webpage_bytes[:1024])
883 if m:
884 encoding = m.group(1).decode('ascii')
b60016e8
PH
885 elif webpage_bytes.startswith(b'\xff\xfe'):
886 encoding = 'utf-16'
f143d86a
PH
887 else:
888 encoding = 'utf-8'
c9a77969
YCH
889
890 return encoding
891
4457823d
S
892 def __check_blocked(self, content):
893 first_block = content[:512]
3089bc74
S
894 if ('<title>Access to this site is blocked</title>' in content
895 and 'Websense' in first_block):
4457823d
S
896 msg = 'Access to this webpage has been blocked by Websense filtering software in your network.'
897 blocked_iframe = self._html_search_regex(
898 r'<iframe src="([^"]+)"', content,
899 'Websense information URL', default=None)
900 if blocked_iframe:
901 msg += ' Visit %s for more details' % blocked_iframe
902 raise ExtractorError(msg, expected=True)
903 if '<title>The URL you requested has been blocked</title>' in first_block:
904 msg = (
905 'Access to this webpage has been blocked by Indian censorship. '
906 'Use a VPN or proxy server (with --proxy) to route around it.')
907 block_msg = self._html_search_regex(
908 r'</h1><p>(.*?)</p>',
909 content, 'block message', default=None)
910 if block_msg:
911 msg += ' (Message: "%s")' % block_msg.replace('\n', ' ')
912 raise ExtractorError(msg, expected=True)
3089bc74
S
913 if ('<title>TTK :: Доступ к ресурсу ограничен</title>' in content
914 and 'blocklist.rkn.gov.ru' in content):
4457823d
S
915 raise ExtractorError(
916 'Access to this webpage has been blocked by decision of the Russian government. '
917 'Visit http://blocklist.rkn.gov.ru/ for a block reason.',
918 expected=True)
919
f95b9dee 920 def _request_dump_filename(self, url, video_id):
921 basen = f'{video_id}_{url}'
922 trim_length = self.get_param('trim_file_name') or 240
923 if len(basen) > trim_length:
924 h = '___' + hashlib.md5(basen.encode('utf-8')).hexdigest()
925 basen = basen[:trim_length - len(h)] + h
926 filename = sanitize_filename(f'{basen}.dump', restricted=True)
927 # Working around MAX_PATH limitation on Windows (see
928 # http://msdn.microsoft.com/en-us/library/windows/desktop/aa365247(v=vs.85).aspx)
929 if compat_os_name == 'nt':
930 absfilepath = os.path.abspath(filename)
931 if len(absfilepath) > 259:
932 filename = fR'\\?\{absfilepath}'
933 return filename
934
935 def __decode_webpage(self, webpage_bytes, encoding, headers):
936 if not encoding:
937 encoding = self._guess_encoding_from_content(headers.get('Content-Type', ''), webpage_bytes)
938 try:
939 return webpage_bytes.decode(encoding, 'replace')
940 except LookupError:
941 return webpage_bytes.decode('utf-8', 'replace')
942
c9a77969 943 def _webpage_read_content(self, urlh, url_or_request, video_id, note=None, errnote=None, fatal=True, prefix=None, encoding=None):
c9a77969
YCH
944 webpage_bytes = urlh.read()
945 if prefix is not None:
946 webpage_bytes = prefix + webpage_bytes
a06916d9 947 if self.get_param('dump_intermediate_pages', False):
f610dbb0 948 self.to_screen('Dumping request to ' + urlh.geturl())
d6983cb4
PH
949 dump = base64.b64encode(webpage_bytes).decode('ascii')
950 self._downloader.to_screen(dump)
f95b9dee 951 if self.get_param('write_pages'):
e121e3ce 952 filename = self._request_dump_filename(urlh.geturl(), video_id)
f95b9dee 953 self.to_screen(f'Saving request to {filename}')
d41e6efc
PH
954 with open(filename, 'wb') as outf:
955 outf.write(webpage_bytes)
956
f95b9dee 957 content = self.__decode_webpage(webpage_bytes, encoding, urlh.headers)
4457823d 958 self.__check_blocked(content)
2410c43d 959
23be51d8 960 return content
d6983cb4 961
6edf2808 962 def __print_error(self, errnote, fatal, video_id, err):
963 if fatal:
c6e07cf1 964 raise ExtractorError(f'{video_id}: {errnote}', cause=err)
6edf2808 965 elif errnote:
c6e07cf1 966 self.report_warning(f'{video_id}: {errnote}: {err}')
6edf2808 967
968 def _parse_xml(self, xml_string, video_id, transform_source=None, fatal=True, errnote=None):
e2b38da9
PH
969 if transform_source:
970 xml_string = transform_source(xml_string)
e01c3d2e
S
971 try:
972 return compat_etree_fromstring(xml_string.encode('utf-8'))
f9934b96 973 except xml.etree.ElementTree.ParseError as ve:
6edf2808 974 self.__print_error('Failed to parse XML' if errnote is None else errnote, fatal, video_id, ve)
267ed0c5 975
6edf2808 976 def _parse_json(self, json_string, video_id, transform_source=None, fatal=True, errnote=None, **parser_kwargs):
3d3538e4 977 try:
b7c47b74 978 return json.loads(
979 json_string, cls=LenientJSONDecoder, strict=False, transform_source=transform_source, **parser_kwargs)
3d3538e4 980 except ValueError as ve:
6edf2808 981 self.__print_error('Failed to parse JSON' if errnote is None else errnote, fatal, video_id, ve)
3d3538e4 982
6edf2808 983 def _parse_socket_response_as_json(self, data, *args, **kwargs):
984 return self._parse_json(data[data.find('{'):data.rfind('}') + 1], *args, **kwargs)
adddc50c 985
617f658b 986 def __create_download_methods(name, parser, note, errnote, return_value):
987
6edf2808 988 def parse(ie, content, *args, errnote=errnote, **kwargs):
617f658b 989 if parser is None:
990 return content
6edf2808 991 if errnote is False:
992 kwargs['errnote'] = errnote
617f658b 993 # parser is fetched by name so subclasses can override it
994 return getattr(ie, parser)(content, *args, **kwargs)
995
c4910024 996 def download_handle(self, url_or_request, video_id, note=note, errnote=errnote, transform_source=None,
997 fatal=True, encoding=None, data=None, headers={}, query={}, expected_status=None):
998 res = self._download_webpage_handle(
999 url_or_request, video_id, note=note, errnote=errnote, fatal=fatal, encoding=encoding,
1000 data=data, headers=headers, query=query, expected_status=expected_status)
617f658b 1001 if res is False:
1002 return res
1003 content, urlh = res
6edf2808 1004 return parse(self, content, video_id, transform_source=transform_source, fatal=fatal, errnote=errnote), urlh
617f658b 1005
f95b9dee 1006 def download_content(self, url_or_request, video_id, note=note, errnote=errnote, transform_source=None,
c4910024 1007 fatal=True, encoding=None, data=None, headers={}, query={}, expected_status=None):
f95b9dee 1008 if self.get_param('load_pages'):
1009 url_or_request = self._create_request(url_or_request, data, headers, query)
1010 filename = self._request_dump_filename(url_or_request.full_url, video_id)
1011 self.to_screen(f'Loading request from {filename}')
1012 try:
1013 with open(filename, 'rb') as dumpf:
1014 webpage_bytes = dumpf.read()
1015 except OSError as e:
1016 self.report_warning(f'Unable to load request from disk: {e}')
1017 else:
1018 content = self.__decode_webpage(webpage_bytes, encoding, url_or_request.headers)
6edf2808 1019 return parse(self, content, video_id, transform_source=transform_source, fatal=fatal, errnote=errnote)
c4910024 1020 kwargs = {
1021 'note': note,
1022 'errnote': errnote,
1023 'transform_source': transform_source,
1024 'fatal': fatal,
1025 'encoding': encoding,
1026 'data': data,
1027 'headers': headers,
1028 'query': query,
1029 'expected_status': expected_status,
1030 }
617f658b 1031 if parser is None:
c4910024 1032 kwargs.pop('transform_source')
617f658b 1033 # The method is fetched by name so subclasses can override _download_..._handle
c4910024 1034 res = getattr(self, download_handle.__name__)(url_or_request, video_id, **kwargs)
617f658b 1035 return res if res is False else res[0]
1036
1037 def impersonate(func, name, return_value):
1038 func.__name__, func.__qualname__ = name, f'InfoExtractor.{name}'
1039 func.__doc__ = f'''
1040 @param transform_source Apply this transformation before parsing
1041 @returns {return_value}
1042
1043 See _download_webpage_handle docstring for other arguments specification
1044 '''
1045
1046 impersonate(download_handle, f'_download_{name}_handle', f'({return_value}, URL handle)')
1047 impersonate(download_content, f'_download_{name}', f'{return_value}')
1048 return download_handle, download_content
1049
1050 _download_xml_handle, _download_xml = __create_download_methods(
1051 'xml', '_parse_xml', 'Downloading XML', 'Unable to download XML', 'xml as an xml.etree.ElementTree.Element')
1052 _download_json_handle, _download_json = __create_download_methods(
1053 'json', '_parse_json', 'Downloading JSON metadata', 'Unable to download JSON metadata', 'JSON object as a dict')
1054 _download_socket_json_handle, _download_socket_json = __create_download_methods(
1055 'socket_json', '_parse_socket_response_as_json', 'Polling socket', 'Unable to poll socket', 'JSON object as a dict')
1056 __download_webpage = __create_download_methods('webpage', None, None, None, 'data of the page as a string')[1]
adddc50c 1057
617f658b 1058 def _download_webpage(
1059 self, url_or_request, video_id, note=None, errnote=None,
1060 fatal=True, tries=1, timeout=NO_DEFAULT, *args, **kwargs):
adddc50c 1061 """
617f658b 1062 Return the data of the page as a string.
adddc50c 1063
617f658b 1064 Keyword arguments:
1065 tries -- number of tries
1066 timeout -- sleep interval between tries
1067
1068 See _download_webpage_handle docstring for other arguments specification.
adddc50c 1069 """
617f658b 1070
1071 R''' # NB: These are unused; should they be deprecated?
1072 if tries != 1:
1073 self._downloader.deprecation_warning('tries argument is deprecated in InfoExtractor._download_webpage')
1074 if timeout is NO_DEFAULT:
1075 timeout = 5
1076 else:
1077 self._downloader.deprecation_warning('timeout argument is deprecated in InfoExtractor._download_webpage')
1078 '''
1079
1080 try_count = 0
1081 while True:
1082 try:
1083 return self.__download_webpage(url_or_request, video_id, note, errnote, None, fatal, *args, **kwargs)
ac668111 1084 except http.client.IncompleteRead as e:
617f658b 1085 try_count += 1
1086 if try_count >= tries:
1087 raise e
1088 self._sleep(timeout, video_id)
adddc50c 1089
28f436ba 1090 def report_warning(self, msg, video_id=None, *args, only_once=False, **kwargs):
a70635b8 1091 idstr = format_field(video_id, None, '%s: ')
28f436ba 1092 msg = f'[{self.IE_NAME}] {idstr}{msg}'
1093 if only_once:
1094 if f'WARNING: {msg}' in self._printed_messages:
1095 return
1096 self._printed_messages.add(f'WARNING: {msg}')
1097 self._downloader.report_warning(msg, *args, **kwargs)
f45f96f8 1098
a06916d9 1099 def to_screen(self, msg, *args, **kwargs):
d6983cb4 1100 """Print msg to screen, prefixing it with '[ie_name]'"""
86e5f3ed 1101 self._downloader.to_screen(f'[{self.IE_NAME}] {msg}', *args, **kwargs)
a06916d9 1102
1103 def write_debug(self, msg, *args, **kwargs):
86e5f3ed 1104 self._downloader.write_debug(f'[{self.IE_NAME}] {msg}', *args, **kwargs)
a06916d9 1105
1106 def get_param(self, name, default=None, *args, **kwargs):
1107 if self._downloader:
1108 return self._downloader.params.get(name, default, *args, **kwargs)
1109 return default
d6983cb4 1110
88acdbc2 1111 def report_drm(self, video_id, partial=False):
1112 self.raise_no_formats('This video is DRM protected', expected=True, video_id=video_id)
1113
d6983cb4
PH
1114 def report_extraction(self, id_or_name):
1115 """Report information extraction."""
f1a9d64e 1116 self.to_screen('%s: Extracting information' % id_or_name)
d6983cb4
PH
1117
1118 def report_download_webpage(self, video_id):
1119 """Report webpage download."""
f1a9d64e 1120 self.to_screen('%s: Downloading webpage' % video_id)
d6983cb4
PH
1121
1122 def report_age_confirmation(self):
1123 """Report attempt to confirm age."""
f1a9d64e 1124 self.to_screen('Confirming age')
d6983cb4 1125
fc79158d
JMF
1126 def report_login(self):
1127 """Report attempt to log in."""
f1a9d64e 1128 self.to_screen('Logging in')
fc79158d 1129
b7da73eb 1130 def raise_login_required(
9d5d4d64 1131 self, msg='This video is only available for registered users',
52efa4b3 1132 metadata_available=False, method=NO_DEFAULT):
f2ebc5c7 1133 if metadata_available and (
1134 self.get_param('ignore_no_formats_error') or self.get_param('wait_for_video')):
b7da73eb 1135 self.report_warning(msg)
7265a219 1136 return
a70635b8 1137 msg += format_field(self._login_hint(method), None, '. %s')
46890374 1138 raise ExtractorError(msg, expected=True)
43e7d3c9 1139
b7da73eb 1140 def raise_geo_restricted(
1141 self, msg='This video is not available from your location due to geo restriction',
1142 countries=None, metadata_available=False):
f2ebc5c7 1143 if metadata_available and (
1144 self.get_param('ignore_no_formats_error') or self.get_param('wait_for_video')):
b7da73eb 1145 self.report_warning(msg)
1146 else:
1147 raise GeoRestrictedError(msg, countries=countries)
1148
1149 def raise_no_formats(self, msg, expected=False, video_id=None):
f2ebc5c7 1150 if expected and (
1151 self.get_param('ignore_no_formats_error') or self.get_param('wait_for_video')):
b7da73eb 1152 self.report_warning(msg, video_id)
68f5867c
L
1153 elif isinstance(msg, ExtractorError):
1154 raise msg
b7da73eb 1155 else:
1156 raise ExtractorError(msg, expected=expected, video_id=video_id)
c430802e 1157
5f6a1245 1158 # Methods for following #608
c0d0b01f 1159 @staticmethod
311b6615 1160 def url_result(url, ie=None, video_id=None, video_title=None, *, url_transparent=False, **kwargs):
10952eb2 1161 """Returns a URL that points to a page that should be processed"""
311b6615 1162 if ie is not None:
1163 kwargs['ie_key'] = ie if isinstance(ie, str) else ie.ie_key()
7012b23c 1164 if video_id is not None:
311b6615 1165 kwargs['id'] = video_id
830d53bf 1166 if video_title is not None:
311b6615 1167 kwargs['title'] = video_title
1168 return {
1169 **kwargs,
1170 '_type': 'url_transparent' if url_transparent else 'url',
1171 'url': url,
1172 }
1173
8f97a15d 1174 @classmethod
1175 def playlist_from_matches(cls, matches, playlist_id=None, playlist_title=None,
1176 getter=IDENTITY, ie=None, video_kwargs=None, **kwargs):
1177 return cls.playlist_result(
1178 (cls.url_result(m, ie, **(video_kwargs or {})) for m in orderedSet(map(getter, matches), lazy=True)),
1179 playlist_id, playlist_title, **kwargs)
46b18f23 1180
c0d0b01f 1181 @staticmethod
311b6615 1182 def playlist_result(entries, playlist_id=None, playlist_title=None, playlist_description=None, *, multi_video=False, **kwargs):
d6983cb4 1183 """Returns a playlist"""
d6983cb4 1184 if playlist_id:
311b6615 1185 kwargs['id'] = playlist_id
d6983cb4 1186 if playlist_title:
311b6615 1187 kwargs['title'] = playlist_title
ecc97af3 1188 if playlist_description is not None:
311b6615 1189 kwargs['description'] = playlist_description
1190 return {
1191 **kwargs,
1192 '_type': 'multi_video' if multi_video else 'playlist',
1193 'entries': entries,
1194 }
d6983cb4 1195
c342041f 1196 def _search_regex(self, pattern, string, name, default=NO_DEFAULT, fatal=True, flags=0, group=None):
d6983cb4
PH
1197 """
1198 Perform a regex search on the given string, using a single or a list of
1199 patterns returning the first matching group.
1200 In case of failure return a default value or raise a WARNING or a
55b3e45b 1201 RegexNotFoundError, depending on fatal, specifying the field name.
d6983cb4 1202 """
61d3665d 1203 if string is None:
1204 mobj = None
77f90330 1205 elif isinstance(pattern, (str, re.Pattern)):
d6983cb4
PH
1206 mobj = re.search(pattern, string, flags)
1207 else:
1208 for p in pattern:
1209 mobj = re.search(p, string, flags)
c3415d1b
PH
1210 if mobj:
1211 break
d6983cb4 1212
ec11a9f4 1213 _name = self._downloader._format_err(name, self._downloader.Styles.EMPHASIS)
d6983cb4
PH
1214
1215 if mobj:
711ede6e
PH
1216 if group is None:
1217 # return the first matching group
1218 return next(g for g in mobj.groups() if g is not None)
198f7ea8 1219 elif isinstance(group, (list, tuple)):
1220 return tuple(mobj.group(g) for g in group)
711ede6e
PH
1221 else:
1222 return mobj.group(group)
c342041f 1223 elif default is not NO_DEFAULT:
d6983cb4
PH
1224 return default
1225 elif fatal:
f1a9d64e 1226 raise RegexNotFoundError('Unable to extract %s' % _name)
d6983cb4 1227 else:
6a39ee13 1228 self.report_warning('unable to extract %s' % _name + bug_reports_message())
d6983cb4
PH
1229 return None
1230
f0bc6e20 1231 def _search_json(self, start_pattern, string, name, video_id, *, end_pattern='',
8b7fb8b6 1232 contains_pattern=r'{(?s:.+)}', fatal=True, default=NO_DEFAULT, **kwargs):
b7c47b74 1233 """Searches string for the JSON object specified by start_pattern"""
1234 # NB: end_pattern is only used to reduce the size of the initial match
f0bc6e20 1235 if default is NO_DEFAULT:
1236 default, has_default = {}, False
1237 else:
1238 fatal, has_default = False, True
1239
1240 json_string = self._search_regex(
8b7fb8b6 1241 rf'(?:{start_pattern})\s*(?P<json>{contains_pattern})\s*(?:{end_pattern})',
f0bc6e20 1242 string, name, group='json', fatal=fatal, default=None if has_default else NO_DEFAULT)
1243 if not json_string:
1244 return default
1245
1246 _name = self._downloader._format_err(name, self._downloader.Styles.EMPHASIS)
1247 try:
1248 return self._parse_json(json_string, video_id, ignore_extra=True, **kwargs)
1249 except ExtractorError as e:
1250 if fatal:
1251 raise ExtractorError(
1252 f'Unable to extract {_name} - Failed to parse JSON', cause=e.cause, video_id=video_id)
1253 elif not has_default:
1254 self.report_warning(
1255 f'Unable to extract {_name} - Failed to parse JSON: {e}', video_id=video_id)
1256 return default
b7c47b74 1257
c342041f 1258 def _html_search_regex(self, pattern, string, name, default=NO_DEFAULT, fatal=True, flags=0, group=None):
d6983cb4
PH
1259 """
1260 Like _search_regex, but strips HTML tags and unescapes entities.
1261 """
711ede6e 1262 res = self._search_regex(pattern, string, name, default, fatal, flags, group)
d6983cb4
PH
1263 if res:
1264 return clean_html(res).strip()
1265 else:
1266 return res
1267
2118fdd1
RA
1268 def _get_netrc_login_info(self, netrc_machine=None):
1269 username = None
1270 password = None
1271 netrc_machine = netrc_machine or self._NETRC_MACHINE
1272
a06916d9 1273 if self.get_param('usenetrc', False):
2118fdd1 1274 try:
0001fcb5 1275 netrc_file = compat_expanduser(self.get_param('netrc_location') or '~')
1276 if os.path.isdir(netrc_file):
1277 netrc_file = os.path.join(netrc_file, '.netrc')
1278 info = netrc.netrc(file=netrc_file).authenticators(netrc_machine)
2118fdd1
RA
1279 if info is not None:
1280 username = info[0]
1281 password = info[2]
1282 else:
dcce092e
S
1283 raise netrc.NetrcParseError(
1284 'No authenticators for %s' % netrc_machine)
86e5f3ed 1285 except (OSError, netrc.NetrcParseError) as err:
6a39ee13 1286 self.report_warning(
dcce092e 1287 'parsing .netrc: %s' % error_to_compat_str(err))
2118fdd1 1288
dcce092e 1289 return username, password
2118fdd1 1290
1b6712ab 1291 def _get_login_info(self, username_option='username', password_option='password', netrc_machine=None):
fc79158d 1292 """
cf0649f8 1293 Get the login info as (username, password)
32443dd3
S
1294 First look for the manually specified credentials using username_option
1295 and password_option as keys in params dictionary. If no such credentials
1296 available look in the netrc file using the netrc_machine or _NETRC_MACHINE
1297 value.
fc79158d
JMF
1298 If there's no info available, return (None, None)
1299 """
fc79158d
JMF
1300
1301 # Attempt to use provided username and password or .netrc data
a06916d9 1302 username = self.get_param(username_option)
1303 if username is not None:
1304 password = self.get_param(password_option)
2118fdd1 1305 else:
1b6712ab 1306 username, password = self._get_netrc_login_info(netrc_machine)
5f6a1245 1307
2133565c 1308 return username, password
fc79158d 1309
e64b7569 1310 def _get_tfa_info(self, note='two-factor verification code'):
83317f69 1311 """
1312 Get the two-factor authentication info
1313 TODO - asking the user will be required for sms/phone verify
1314 currently just uses the command line option
1315 If there's no info available, return None
1316 """
83317f69 1317
a06916d9 1318 tfa = self.get_param('twofactor')
1319 if tfa is not None:
1320 return tfa
83317f69 1321
ac668111 1322 return getpass.getpass('Type %s and press [Return]: ' % note)
83317f69 1323
46720279
JMF
1324 # Helper functions for extracting OpenGraph info
1325 @staticmethod
ab2d5247 1326 def _og_regexes(prop):
448ef1f3 1327 content_re = r'content=(?:"([^"]+?)"|\'([^\']+?)\'|\s*([^\s"\'=<>`]+?))'
fbfde1c3
F
1328 property_re = (r'(?:name|property)=(?:\'og%(sep)s%(prop)s\'|"og%(sep)s%(prop)s"|\s*og%(sep)s%(prop)s\b)'
1329 % {'prop': re.escape(prop), 'sep': '(?:&#x3A;|[:-])'})
78fb87b2 1330 template = r'<meta[^>]+?%s[^>]+?%s'
ab2d5247 1331 return [
78fb87b2
JMF
1332 template % (property_re, content_re),
1333 template % (content_re, property_re),
ab2d5247 1334 ]
46720279 1335
864f24bd
S
1336 @staticmethod
1337 def _meta_regex(prop):
1338 return r'''(?isx)<meta
8b9848ac 1339 (?=[^>]+(?:itemprop|name|property|id|http-equiv)=(["\']?)%s\1)
864f24bd
S
1340 [^>]+?content=(["\'])(?P<content>.*?)\2''' % re.escape(prop)
1341
3c4e6d83 1342 def _og_search_property(self, prop, html, name=None, **kargs):
6606817a 1343 prop = variadic(prop)
46720279 1344 if name is None:
b070564e
S
1345 name = 'OpenGraph %s' % prop[0]
1346 og_regexes = []
1347 for p in prop:
1348 og_regexes.extend(self._og_regexes(p))
1349 escaped = self._search_regex(og_regexes, html, name, flags=re.DOTALL, **kargs)
eb0a8398
PH
1350 if escaped is None:
1351 return None
1352 return unescapeHTML(escaped)
46720279
JMF
1353
1354 def _og_search_thumbnail(self, html, **kargs):
10952eb2 1355 return self._og_search_property('image', html, 'thumbnail URL', fatal=False, **kargs)
46720279
JMF
1356
1357 def _og_search_description(self, html, **kargs):
1358 return self._og_search_property('description', html, fatal=False, **kargs)
1359
04f3fd2c 1360 def _og_search_title(self, html, *, fatal=False, **kargs):
1361 return self._og_search_property('title', html, fatal=fatal, **kargs)
46720279 1362
8ffa13e0 1363 def _og_search_video_url(self, html, name='video url', secure=True, **kargs):
a3681973
PH
1364 regexes = self._og_regexes('video') + self._og_regexes('video:url')
1365 if secure:
1366 regexes = self._og_regexes('video:secure_url') + regexes
8ffa13e0 1367 return self._html_search_regex(regexes, html, name, **kargs)
46720279 1368
78338f71
JMF
1369 def _og_search_url(self, html, **kargs):
1370 return self._og_search_property('url', html, **kargs)
1371
04f3fd2c 1372 def _html_extract_title(self, html, name='title', *, fatal=False, **kwargs):
21633673 1373 return self._html_search_regex(r'(?s)<title\b[^>]*>([^<]+)</title>', html, name, fatal=fatal, **kwargs)
77cc7c6e 1374
40c696e5 1375 def _html_search_meta(self, name, html, display_name=None, fatal=False, **kwargs):
6606817a 1376 name = variadic(name)
59040888 1377 if display_name is None:
88d9f6c0 1378 display_name = name[0]
59040888 1379 return self._html_search_regex(
88d9f6c0 1380 [self._meta_regex(n) for n in name],
711ede6e 1381 html, display_name, fatal=fatal, group='content', **kwargs)
59040888
PH
1382
1383 def _dc_search_uploader(self, html):
1384 return self._html_search_meta('dc.creator', html, 'uploader')
1385
8f97a15d 1386 @staticmethod
1387 def _rta_search(html):
8dbe9899
PH
1388 # See http://www.rtalabel.org/index.php?content=howtofaq#single
1389 if re.search(r'(?ix)<meta\s+name="rating"\s+'
1390 r' content="RTA-5042-1996-1400-1577-RTA"',
1391 html):
1392 return 18
8f97a15d 1393
1394 # And then there are the jokers who advertise that they use RTA, but actually don't.
1395 AGE_LIMIT_MARKERS = [
1396 r'Proudly Labeled <a href="http://www\.rtalabel\.org/" title="Restricted to Adults">RTA</a>',
1397 ]
1398 if any(re.search(marker, html) for marker in AGE_LIMIT_MARKERS):
1399 return 18
8dbe9899
PH
1400 return 0
1401
59040888
PH
1402 def _media_rating_search(self, html):
1403 # See http://www.tjg-designs.com/WP/metadata-code-examples-adding-metadata-to-your-web-pages/
1404 rating = self._html_search_meta('rating', html)
1405
1406 if not rating:
1407 return None
1408
1409 RATING_TABLE = {
1410 'safe for kids': 0,
1411 'general': 8,
1412 '14 years': 14,
1413 'mature': 17,
1414 'restricted': 19,
1415 }
d800609c 1416 return RATING_TABLE.get(rating.lower())
59040888 1417
69319969 1418 def _family_friendly_search(self, html):
6ca7732d 1419 # See http://schema.org/VideoObject
ac8491fc
S
1420 family_friendly = self._html_search_meta(
1421 'isFamilyFriendly', html, default=None)
69319969
NJ
1422
1423 if not family_friendly:
1424 return None
1425
1426 RATING_TABLE = {
1427 '1': 0,
1428 'true': 0,
1429 '0': 18,
1430 'false': 18,
1431 }
d800609c 1432 return RATING_TABLE.get(family_friendly.lower())
69319969 1433
0c708f11
JMF
1434 def _twitter_search_player(self, html):
1435 return self._html_search_meta('twitter:player', html,
9e1a5b84 1436 'twitter card player')
0c708f11 1437
0c36dc00 1438 def _yield_json_ld(self, html, video_id, *, fatal=True, default=NO_DEFAULT):
1439 """Yield all json ld objects in the html"""
1440 if default is not NO_DEFAULT:
1441 fatal = False
1442 for mobj in re.finditer(JSON_LD_RE, html):
1443 json_ld_item = self._parse_json(mobj.group('json_ld'), video_id, fatal=fatal)
1444 for json_ld in variadic(json_ld_item):
1445 if isinstance(json_ld, dict):
1446 yield json_ld
1447
1448 def _search_json_ld(self, html, video_id, expected_type=None, *, fatal=True, default=NO_DEFAULT):
1449 """Search for a video in any json ld in the html"""
1450 if default is not NO_DEFAULT:
1451 fatal = False
1452 info = self._json_ld(
1453 list(self._yield_json_ld(html, video_id, fatal=fatal, default=default)),
1454 video_id, fatal=fatal, expected_type=expected_type)
1455 if info:
1456 return info
4433bb02
S
1457 if default is not NO_DEFAULT:
1458 return default
1459 elif fatal:
1460 raise RegexNotFoundError('Unable to extract JSON-LD')
1461 else:
6a39ee13 1462 self.report_warning('unable to extract JSON-LD %s' % bug_reports_message())
4433bb02 1463 return {}
4ca2a3cf 1464
95b31e26 1465 def _json_ld(self, json_ld, video_id, fatal=True, expected_type=None):
14f25df2 1466 if isinstance(json_ld, str):
4ca2a3cf
S
1467 json_ld = self._parse_json(json_ld, video_id, fatal=fatal)
1468 if not json_ld:
1469 return {}
1470 info = {}
bae14048 1471
e7e4a6e0
S
1472 INTERACTION_TYPE_MAP = {
1473 'CommentAction': 'comment',
1474 'AgreeAction': 'like',
1475 'DisagreeAction': 'dislike',
1476 'LikeAction': 'like',
1477 'DislikeAction': 'dislike',
1478 'ListenAction': 'view',
1479 'WatchAction': 'view',
1480 'ViewAction': 'view',
1481 }
1482
f3c0c773 1483 def is_type(e, *expected_types):
1484 type = variadic(traverse_obj(e, '@type'))
1485 return any(x in type for x in expected_types)
1486
29f7c58a 1487 def extract_interaction_type(e):
1488 interaction_type = e.get('interactionType')
1489 if isinstance(interaction_type, dict):
1490 interaction_type = interaction_type.get('@type')
1491 return str_or_none(interaction_type)
1492
e7e4a6e0
S
1493 def extract_interaction_statistic(e):
1494 interaction_statistic = e.get('interactionStatistic')
29f7c58a 1495 if isinstance(interaction_statistic, dict):
1496 interaction_statistic = [interaction_statistic]
e7e4a6e0
S
1497 if not isinstance(interaction_statistic, list):
1498 return
1499 for is_e in interaction_statistic:
f3c0c773 1500 if not is_type(is_e, 'InteractionCounter'):
e7e4a6e0 1501 continue
29f7c58a 1502 interaction_type = extract_interaction_type(is_e)
1503 if not interaction_type:
e7e4a6e0 1504 continue
ce5b9040
S
1505 # For interaction count some sites provide string instead of
1506 # an integer (as per spec) with non digit characters (e.g. ",")
1507 # so extracting count with more relaxed str_to_int
1508 interaction_count = str_to_int(is_e.get('userInteractionCount'))
e7e4a6e0
S
1509 if interaction_count is None:
1510 continue
1511 count_kind = INTERACTION_TYPE_MAP.get(interaction_type.split('/')[-1])
1512 if not count_kind:
1513 continue
1514 count_key = '%s_count' % count_kind
1515 if info.get(count_key) is not None:
1516 continue
1517 info[count_key] = interaction_count
1518
f5225737 1519 def extract_chapter_information(e):
1520 chapters = [{
1521 'title': part.get('name'),
1522 'start_time': part.get('startOffset'),
1523 'end_time': part.get('endOffset'),
85553414 1524 } for part in variadic(e.get('hasPart') or []) if part.get('@type') == 'Clip']
f5225737 1525 for idx, (last_c, current_c, next_c) in enumerate(zip(
1526 [{'end_time': 0}] + chapters, chapters, chapters[1:])):
1527 current_c['end_time'] = current_c['end_time'] or next_c['start_time']
1528 current_c['start_time'] = current_c['start_time'] or last_c['end_time']
1529 if None in current_c.values():
1530 self.report_warning(f'Chapter {idx} contains broken data. Not extracting chapters')
1531 return
1532 if chapters:
1533 chapters[-1]['end_time'] = chapters[-1]['end_time'] or info['duration']
1534 info['chapters'] = chapters
1535
bae14048 1536 def extract_video_object(e):
f7ad7160 1537 author = e.get('author')
bae14048 1538 info.update({
0c36dc00 1539 'url': url_or_none(e.get('contentUrl')),
0f60ba6e 1540 'ext': mimetype2ext(e.get('encodingFormat')),
bae14048
S
1541 'title': unescapeHTML(e.get('name')),
1542 'description': unescapeHTML(e.get('description')),
eb2333bc 1543 'thumbnails': [{'url': unescapeHTML(url)}
21633673 1544 for url in variadic(traverse_obj(e, 'thumbnailUrl', 'thumbnailURL'))
1545 if url_or_none(url)],
bae14048
S
1546 'duration': parse_duration(e.get('duration')),
1547 'timestamp': unified_timestamp(e.get('uploadDate')),
f7ad7160 1548 # author can be an instance of 'Organization' or 'Person' types.
1549 # both types can have 'name' property(inherited from 'Thing' type). [1]
1550 # however some websites are using 'Text' type instead.
1551 # 1. https://schema.org/VideoObject
14f25df2 1552 'uploader': author.get('name') if isinstance(author, dict) else author if isinstance(author, str) else None,
0f60ba6e 1553 'artist': traverse_obj(e, ('byArtist', 'name'), expected_type=str),
56ba69e4 1554 'filesize': int_or_none(float_or_none(e.get('contentSize'))),
bae14048
S
1555 'tbr': int_or_none(e.get('bitrate')),
1556 'width': int_or_none(e.get('width')),
1557 'height': int_or_none(e.get('height')),
33a81c2c 1558 'view_count': int_or_none(e.get('interactionCount')),
0f60ba6e 1559 'tags': try_call(lambda: e.get('keywords').split(',')),
bae14048 1560 })
0f60ba6e 1561 if is_type(e, 'AudioObject'):
1562 info.update({
1563 'vcodec': 'none',
1564 'abr': int_or_none(e.get('bitrate')),
1565 })
e7e4a6e0 1566 extract_interaction_statistic(e)
f5225737 1567 extract_chapter_information(e)
bae14048 1568
d5c32548 1569 def traverse_json_ld(json_ld, at_top_level=True):
1d55ebab
SS
1570 for e in variadic(json_ld):
1571 if not isinstance(e, dict):
1572 continue
d5c32548
ZM
1573 if at_top_level and '@context' not in e:
1574 continue
1575 if at_top_level and set(e.keys()) == {'@context', '@graph'}:
1d55ebab 1576 traverse_json_ld(e['@graph'], at_top_level=False)
d5c32548 1577 break
f3c0c773 1578 if expected_type is not None and not is_type(e, expected_type):
4433bb02 1579 continue
8f122fa0 1580 rating = traverse_obj(e, ('aggregateRating', 'ratingValue'), expected_type=float_or_none)
1581 if rating is not None:
1582 info['average_rating'] = rating
f3c0c773 1583 if is_type(e, 'TVEpisode', 'Episode'):
440863ad 1584 episode_name = unescapeHTML(e.get('name'))
46933a15 1585 info.update({
440863ad 1586 'episode': episode_name,
46933a15
S
1587 'episode_number': int_or_none(e.get('episodeNumber')),
1588 'description': unescapeHTML(e.get('description')),
1589 })
440863ad
S
1590 if not info.get('title') and episode_name:
1591 info['title'] = episode_name
46933a15 1592 part_of_season = e.get('partOfSeason')
f3c0c773 1593 if is_type(part_of_season, 'TVSeason', 'Season', 'CreativeWorkSeason'):
458fd30f
S
1594 info.update({
1595 'season': unescapeHTML(part_of_season.get('name')),
1596 'season_number': int_or_none(part_of_season.get('seasonNumber')),
1597 })
d16b3c66 1598 part_of_series = e.get('partOfSeries') or e.get('partOfTVSeries')
f3c0c773 1599 if is_type(part_of_series, 'TVSeries', 'Series', 'CreativeWorkSeries'):
46933a15 1600 info['series'] = unescapeHTML(part_of_series.get('name'))
f3c0c773 1601 elif is_type(e, 'Movie'):
391256dc
S
1602 info.update({
1603 'title': unescapeHTML(e.get('name')),
1604 'description': unescapeHTML(e.get('description')),
1605 'duration': parse_duration(e.get('duration')),
1606 'timestamp': unified_timestamp(e.get('dateCreated')),
1607 })
f3c0c773 1608 elif is_type(e, 'Article', 'NewsArticle'):
46933a15
S
1609 info.update({
1610 'timestamp': parse_iso8601(e.get('datePublished')),
1611 'title': unescapeHTML(e.get('headline')),
d5c32548 1612 'description': unescapeHTML(e.get('articleBody') or e.get('description')),
46933a15 1613 })
f3c0c773 1614 if is_type(traverse_obj(e, ('video', 0)), 'VideoObject'):
2edb38e8 1615 extract_video_object(e['video'][0])
f3c0c773 1616 elif is_type(traverse_obj(e, ('subjectOf', 0)), 'VideoObject'):
e50c3500 1617 extract_video_object(e['subjectOf'][0])
0f60ba6e 1618 elif is_type(e, 'VideoObject', 'AudioObject'):
bae14048 1619 extract_video_object(e)
4433bb02
S
1620 if expected_type is None:
1621 continue
1622 else:
1623 break
c69701c6 1624 video = e.get('video')
f3c0c773 1625 if is_type(video, 'VideoObject'):
c69701c6 1626 extract_video_object(video)
4433bb02
S
1627 if expected_type is None:
1628 continue
1629 else:
1630 break
d5c32548 1631
1d55ebab 1632 traverse_json_ld(json_ld)
90137ca4 1633 return filter_dict(info)
4ca2a3cf 1634
135dfa2c 1635 def _search_nextjs_data(self, webpage, video_id, *, transform_source=None, fatal=True, **kw):
f98709af
LL
1636 return self._parse_json(
1637 self._search_regex(
1638 r'(?s)<script[^>]+id=[\'"]__NEXT_DATA__[\'"][^>]*>([^<]+)</script>',
135dfa2c 1639 webpage, 'next.js data', fatal=fatal, **kw),
1640 video_id, transform_source=transform_source, fatal=fatal)
f98709af 1641
8072ef2b 1642 def _search_nuxt_data(self, webpage, video_id, context_name='__NUXT__', *, fatal=True, traverse=('data', 0)):
1643 """Parses Nuxt.js metadata. This works as long as the function __NUXT__ invokes is a pure function"""
66f4c04e 1644 rectx = re.escape(context_name)
8072ef2b 1645 FUNCTION_RE = r'\(function\((?P<arg_keys>.*?)\){return\s+(?P<js>{.*?})\s*;?\s*}\((?P<arg_vals>.*?)\)'
66f4c04e 1646 js, arg_keys, arg_vals = self._search_regex(
8072ef2b 1647 (rf'<script>\s*window\.{rectx}={FUNCTION_RE}\s*\)\s*;?\s*</script>', rf'{rectx}\(.*?{FUNCTION_RE}'),
1648 webpage, context_name, group=('js', 'arg_keys', 'arg_vals'), fatal=fatal)
66f4c04e
THD
1649
1650 args = dict(zip(arg_keys.split(','), arg_vals.split(',')))
1651
1652 for key, val in args.items():
1653 if val in ('undefined', 'void 0'):
1654 args[key] = 'null'
1655
8072ef2b 1656 ret = self._parse_json(js, video_id, transform_source=functools.partial(js_to_json, vars=args), fatal=fatal)
1657 return traverse_obj(ret, traverse) or {}
66f4c04e 1658
27713812 1659 @staticmethod
f8da79f8 1660 def _hidden_inputs(html):
586f1cc5 1661 html = re.sub(r'<!--(?:(?!<!--).)*-->', '', html)
201ea3ee 1662 hidden_inputs = {}
c8498368
S
1663 for input in re.findall(r'(?i)(<input[^>]+>)', html):
1664 attrs = extract_attributes(input)
1665 if not input:
201ea3ee 1666 continue
c8498368 1667 if attrs.get('type') not in ('hidden', 'submit'):
201ea3ee 1668 continue
c8498368
S
1669 name = attrs.get('name') or attrs.get('id')
1670 value = attrs.get('value')
1671 if name and value is not None:
1672 hidden_inputs[name] = value
201ea3ee 1673 return hidden_inputs
27713812 1674
cf61d96d
S
1675 def _form_hidden_inputs(self, form_id, html):
1676 form = self._search_regex(
73eb13df 1677 r'(?is)<form[^>]+?id=(["\'])%s\1[^>]*>(?P<form>.+?)</form>' % form_id,
cf61d96d
S
1678 html, '%s form' % form_id, group='form')
1679 return self._hidden_inputs(form)
1680
eb8a4433 1681 class FormatSort:
b050d210 1682 regex = r' *((?P<reverse>\+)?(?P<field>[a-zA-Z0-9_]+)((?P<separator>[~:])(?P<limit>.*?))?)? *$'
eb8a4433 1683
8326b00a 1684 default = ('hidden', 'aud_or_vid', 'hasvid', 'ie_pref', 'lang', 'quality',
7e798d72 1685 'res', 'fps', 'hdr:12', 'vcodec:vp9.2', 'channels', 'acodec',
1686 'size', 'br', 'asr', 'proto', 'ext', 'hasaud', 'source', 'id') # These must not be aliases
198e3a04 1687 ytdl_default = ('hasaud', 'lang', 'quality', 'tbr', 'filesize', 'vbr',
53ed7066 1688 'height', 'width', 'proto', 'vext', 'abr', 'aext',
f304da8a 1689 'fps', 'fs_approx', 'source', 'id')
eb8a4433 1690
1691 settings = {
1692 'vcodec': {'type': 'ordered', 'regex': True,
155d2b48 1693 'order': ['av0?1', 'vp0?9.2', 'vp0?9', '[hx]265|he?vc?', '[hx]264|avc', 'vp0?8', 'mp4v|h263', 'theora', '', None, 'none']},
eb8a4433 1694 'acodec': {'type': 'ordered', 'regex': True,
a10aa588 1695 'order': ['[af]lac', 'wav|aiff', 'opus', 'vorbis|ogg', 'aac', 'mp?4a?', 'mp3', 'e-?a?c-?3', 'ac-?3', 'dts', '', None, 'none']},
176f1866 1696 'hdr': {'type': 'ordered', 'regex': True, 'field': 'dynamic_range',
1697 'order': ['dv', '(hdr)?12', r'(hdr)?10\+', '(hdr)?10', 'hlg', '', 'sdr', None]},
f137c99e 1698 'proto': {'type': 'ordered', 'regex': True, 'field': 'protocol',
f304da8a 1699 'order': ['(ht|f)tps', '(ht|f)tp$', 'm3u8.*', '.*dash', 'websocket_frag', 'rtmpe?', '', 'mms|rtsp', 'ws|websocket', 'f4']},
eb8a4433 1700 'vext': {'type': 'ordered', 'field': 'video_ext',
91ebc640 1701 'order': ('mp4', 'webm', 'flv', '', 'none'),
eb8a4433 1702 'order_free': ('webm', 'mp4', 'flv', '', 'none')},
1703 'aext': {'type': 'ordered', 'field': 'audio_ext',
1704 'order': ('m4a', 'aac', 'mp3', 'ogg', 'opus', 'webm', '', 'none'),
f2e9fa3e 1705 'order_free': ('ogg', 'opus', 'webm', 'mp3', 'm4a', 'aac', '', 'none')},
eb8a4433 1706 'hidden': {'visible': False, 'forced': True, 'type': 'extractor', 'max': -1000},
f5510afe 1707 'aud_or_vid': {'visible': False, 'forced': True, 'type': 'multiple',
8326b00a 1708 'field': ('vcodec', 'acodec'),
1709 'function': lambda it: int(any(v != 'none' for v in it))},
f983b875 1710 'ie_pref': {'priority': True, 'type': 'extractor'},
63be1aab 1711 'hasvid': {'priority': True, 'field': 'vcodec', 'type': 'boolean', 'not_in_list': ('none',)},
1712 'hasaud': {'field': 'acodec', 'type': 'boolean', 'not_in_list': ('none',)},
10beccc9 1713 'lang': {'convert': 'float', 'field': 'language_preference', 'default': -1},
1714 'quality': {'convert': 'float', 'default': -1},
eb8a4433 1715 'filesize': {'convert': 'bytes'},
f137c99e 1716 'fs_approx': {'convert': 'bytes', 'field': 'filesize_approx'},
1717 'id': {'convert': 'string', 'field': 'format_id'},
eb8a4433 1718 'height': {'convert': 'float_none'},
1719 'width': {'convert': 'float_none'},
1720 'fps': {'convert': 'float_none'},
b8ed0f15 1721 'channels': {'convert': 'float_none', 'field': 'audio_channels'},
eb8a4433 1722 'tbr': {'convert': 'float_none'},
1723 'vbr': {'convert': 'float_none'},
1724 'abr': {'convert': 'float_none'},
1725 'asr': {'convert': 'float_none'},
10beccc9 1726 'source': {'convert': 'float', 'field': 'source_preference', 'default': -1},
63be1aab 1727
eb8a4433 1728 'codec': {'type': 'combined', 'field': ('vcodec', 'acodec')},
63be1aab 1729 'br': {'type': 'combined', 'field': ('tbr', 'vbr', 'abr'), 'same_limit': True},
1730 'size': {'type': 'combined', 'same_limit': True, 'field': ('filesize', 'fs_approx')},
1731 'ext': {'type': 'combined', 'field': ('vext', 'aext')},
f5510afe 1732 'res': {'type': 'multiple', 'field': ('height', 'width'),
dbf5416a 1733 'function': lambda it: (lambda l: min(l) if l else 0)(tuple(filter(None, it)))},
63be1aab 1734
b8ed0f15 1735 # Actual field names
19188702 1736 'format_id': {'type': 'alias', 'field': 'id'},
1737 'preference': {'type': 'alias', 'field': 'ie_pref'},
1738 'language_preference': {'type': 'alias', 'field': 'lang'},
63be1aab 1739 'source_preference': {'type': 'alias', 'field': 'source'},
08d30158 1740 'protocol': {'type': 'alias', 'field': 'proto'},
63be1aab 1741 'filesize_approx': {'type': 'alias', 'field': 'fs_approx'},
b8ed0f15 1742 'audio_channels': {'type': 'alias', 'field': 'channels'},
08d30158 1743
1744 # Deprecated
1745 'dimension': {'type': 'alias', 'field': 'res', 'deprecated': True},
1746 'resolution': {'type': 'alias', 'field': 'res', 'deprecated': True},
1747 'extension': {'type': 'alias', 'field': 'ext', 'deprecated': True},
1748 'bitrate': {'type': 'alias', 'field': 'br', 'deprecated': True},
1749 'total_bitrate': {'type': 'alias', 'field': 'tbr', 'deprecated': True},
1750 'video_bitrate': {'type': 'alias', 'field': 'vbr', 'deprecated': True},
1751 'audio_bitrate': {'type': 'alias', 'field': 'abr', 'deprecated': True},
1752 'framerate': {'type': 'alias', 'field': 'fps', 'deprecated': True},
1753 'filesize_estimate': {'type': 'alias', 'field': 'size', 'deprecated': True},
1754 'samplerate': {'type': 'alias', 'field': 'asr', 'deprecated': True},
1755 'video_ext': {'type': 'alias', 'field': 'vext', 'deprecated': True},
1756 'audio_ext': {'type': 'alias', 'field': 'aext', 'deprecated': True},
1757 'video_codec': {'type': 'alias', 'field': 'vcodec', 'deprecated': True},
1758 'audio_codec': {'type': 'alias', 'field': 'acodec', 'deprecated': True},
1759 'video': {'type': 'alias', 'field': 'hasvid', 'deprecated': True},
1760 'has_video': {'type': 'alias', 'field': 'hasvid', 'deprecated': True},
1761 'audio': {'type': 'alias', 'field': 'hasaud', 'deprecated': True},
1762 'has_audio': {'type': 'alias', 'field': 'hasaud', 'deprecated': True},
1763 'extractor': {'type': 'alias', 'field': 'ie_pref', 'deprecated': True},
1764 'extractor_preference': {'type': 'alias', 'field': 'ie_pref', 'deprecated': True},
63be1aab 1765 }
eb8a4433 1766
f304da8a 1767 def __init__(self, ie, field_preference):
1768 self._order = []
1769 self.ydl = ie._downloader
1770 self.evaluate_params(self.ydl.params, field_preference)
1771 if ie.get_param('verbose'):
1772 self.print_verbose_info(self.ydl.write_debug)
eb8a4433 1773
1774 def _get_field_setting(self, field, key):
1775 if field not in self.settings:
ee8dd27a 1776 if key in ('forced', 'priority'):
1777 return False
da4db748 1778 self.ydl.deprecated_feature(f'Using arbitrary fields ({field}) for format sorting is '
1779 'deprecated and may be removed in a future version')
eb8a4433 1780 self.settings[field] = {}
1781 propObj = self.settings[field]
1782 if key not in propObj:
1783 type = propObj.get('type')
1784 if key == 'field':
1785 default = 'preference' if type == 'extractor' else (field,) if type in ('combined', 'multiple') else field
1786 elif key == 'convert':
1787 default = 'order' if type == 'ordered' else 'float_string' if field else 'ignore'
4bcc7bd1 1788 else:
f5510afe 1789 default = {'type': 'field', 'visible': True, 'order': [], 'not_in_list': (None,)}.get(key, None)
eb8a4433 1790 propObj[key] = default
1791 return propObj[key]
1792
1793 def _resolve_field_value(self, field, value, convertNone=False):
1794 if value is None:
1795 if not convertNone:
1796 return None
4bcc7bd1 1797 else:
eb8a4433 1798 value = value.lower()
1799 conversion = self._get_field_setting(field, 'convert')
1800 if conversion == 'ignore':
1801 return None
1802 if conversion == 'string':
1803 return value
1804 elif conversion == 'float_none':
1805 return float_or_none(value)
1806 elif conversion == 'bytes':
1807 return FileDownloader.parse_bytes(value)
1808 elif conversion == 'order':
da9be05e 1809 order_list = (self._use_free_order and self._get_field_setting(field, 'order_free')) or self._get_field_setting(field, 'order')
eb8a4433 1810 use_regex = self._get_field_setting(field, 'regex')
1811 list_length = len(order_list)
1812 empty_pos = order_list.index('') if '' in order_list else list_length + 1
1813 if use_regex and value is not None:
da9be05e 1814 for i, regex in enumerate(order_list):
eb8a4433 1815 if regex and re.match(regex, value):
1816 return list_length - i
1817 return list_length - empty_pos # not in list
1818 else: # not regex or value = None
1819 return list_length - (order_list.index(value) if value in order_list else empty_pos)
1820 else:
1821 if value.isnumeric():
1822 return float(value)
4bcc7bd1 1823 else:
eb8a4433 1824 self.settings[field]['convert'] = 'string'
1825 return value
1826
1827 def evaluate_params(self, params, sort_extractor):
1828 self._use_free_order = params.get('prefer_free_formats', False)
1829 self._sort_user = params.get('format_sort', [])
1830 self._sort_extractor = sort_extractor
1831
1832 def add_item(field, reverse, closest, limit_text):
1833 field = field.lower()
1834 if field in self._order:
1835 return
1836 self._order.append(field)
1837 limit = self._resolve_field_value(field, limit_text)
1838 data = {
1839 'reverse': reverse,
1840 'closest': False if limit is None else closest,
1841 'limit_text': limit_text,
1842 'limit': limit}
1843 if field in self.settings:
1844 self.settings[field].update(data)
1845 else:
1846 self.settings[field] = data
1847
1848 sort_list = (
1849 tuple(field for field in self.default if self._get_field_setting(field, 'forced'))
1850 + (tuple() if params.get('format_sort_force', False)
1851 else tuple(field for field in self.default if self._get_field_setting(field, 'priority')))
1852 + tuple(self._sort_user) + tuple(sort_extractor) + self.default)
1853
1854 for item in sort_list:
1855 match = re.match(self.regex, item)
1856 if match is None:
1857 raise ExtractorError('Invalid format sort string "%s" given by extractor' % item)
1858 field = match.group('field')
1859 if field is None:
1860 continue
1861 if self._get_field_setting(field, 'type') == 'alias':
ee8dd27a 1862 alias, field = field, self._get_field_setting(field, 'field')
08d30158 1863 if self._get_field_setting(alias, 'deprecated'):
da4db748 1864 self.ydl.deprecated_feature(f'Format sorting alias {alias} is deprecated and may '
a057779d 1865 f'be removed in a future version. Please use {field} instead')
eb8a4433 1866 reverse = match.group('reverse') is not None
b050d210 1867 closest = match.group('separator') == '~'
eb8a4433 1868 limit_text = match.group('limit')
1869
1870 has_limit = limit_text is not None
1871 has_multiple_fields = self._get_field_setting(field, 'type') == 'combined'
1872 has_multiple_limits = has_limit and has_multiple_fields and not self._get_field_setting(field, 'same_limit')
1873
1874 fields = self._get_field_setting(field, 'field') if has_multiple_fields else (field,)
b5ae35ee 1875 limits = limit_text.split(':') if has_multiple_limits else (limit_text,) if has_limit else tuple()
eb8a4433 1876 limit_count = len(limits)
1877 for (i, f) in enumerate(fields):
1878 add_item(f, reverse, closest,
1879 limits[i] if i < limit_count
1880 else limits[0] if has_limit and not has_multiple_limits
1881 else None)
1882
0760b0a7 1883 def print_verbose_info(self, write_debug):
b31fdeed 1884 if self._sort_user:
0760b0a7 1885 write_debug('Sort order given by user: %s' % ', '.join(self._sort_user))
eb8a4433 1886 if self._sort_extractor:
0760b0a7 1887 write_debug('Sort order given by extractor: %s' % ', '.join(self._sort_extractor))
1888 write_debug('Formats sorted by: %s' % ', '.join(['%s%s%s' % (
eb8a4433 1889 '+' if self._get_field_setting(field, 'reverse') else '', field,
1890 '%s%s(%s)' % ('~' if self._get_field_setting(field, 'closest') else ':',
1891 self._get_field_setting(field, 'limit_text'),
1892 self._get_field_setting(field, 'limit'))
1893 if self._get_field_setting(field, 'limit_text') is not None else '')
1894 for field in self._order if self._get_field_setting(field, 'visible')]))
1895
1896 def _calculate_field_preference_from_value(self, format, field, type, value):
1897 reverse = self._get_field_setting(field, 'reverse')
1898 closest = self._get_field_setting(field, 'closest')
1899 limit = self._get_field_setting(field, 'limit')
1900
1901 if type == 'extractor':
1902 maximum = self._get_field_setting(field, 'max')
1903 if value is None or (maximum is not None and value >= maximum):
f983b875 1904 value = -1
eb8a4433 1905 elif type == 'boolean':
1906 in_list = self._get_field_setting(field, 'in_list')
1907 not_in_list = self._get_field_setting(field, 'not_in_list')
1908 value = 0 if ((in_list is None or value in in_list) and (not_in_list is None or value not in not_in_list)) else -1
1909 elif type == 'ordered':
1910 value = self._resolve_field_value(field, value, True)
1911
1912 # try to convert to number
6a04a74e 1913 val_num = float_or_none(value, default=self._get_field_setting(field, 'default'))
eb8a4433 1914 is_num = self._get_field_setting(field, 'convert') != 'string' and val_num is not None
1915 if is_num:
1916 value = val_num
1917
1918 return ((-10, 0) if value is None
1919 else (1, value, 0) if not is_num # if a field has mixed strings and numbers, strings are sorted higher
1920 else (0, -abs(value - limit), value - limit if reverse else limit - value) if closest
1921 else (0, value, 0) if not reverse and (limit is None or value <= limit)
1922 else (0, -value, 0) if limit is None or (reverse and value == limit) or value > limit
1923 else (-1, value, 0))
1924
1925 def _calculate_field_preference(self, format, field):
1926 type = self._get_field_setting(field, 'type') # extractor, boolean, ordered, field, multiple
1927 get_value = lambda f: format.get(self._get_field_setting(f, 'field'))
1928 if type == 'multiple':
1929 type = 'field' # Only 'field' is allowed in multiple for now
1930 actual_fields = self._get_field_setting(field, 'field')
1931
f5510afe 1932 value = self._get_field_setting(field, 'function')(get_value(f) for f in actual_fields)
eb8a4433 1933 else:
1934 value = get_value(field)
1935 return self._calculate_field_preference_from_value(format, field, type, value)
1936
1937 def calculate_preference(self, format):
1938 # Determine missing protocol
1939 if not format.get('protocol'):
1940 format['protocol'] = determine_protocol(format)
1941
1942 # Determine missing ext
1943 if not format.get('ext') and 'url' in format:
1944 format['ext'] = determine_ext(format['url'])
1945 if format.get('vcodec') == 'none':
8326b00a 1946 format['audio_ext'] = format['ext'] if format.get('acodec') != 'none' else 'none'
eb8a4433 1947 format['video_ext'] = 'none'
1948 else:
1949 format['video_ext'] = format['ext']
1950 format['audio_ext'] = 'none'
1951 # if format.get('preference') is None and format.get('ext') in ('f4f', 'f4m'): # Not supported?
1952 # format['preference'] = -1000
1953
1954 # Determine missing bitrates
1955 if format.get('tbr') is None:
1956 if format.get('vbr') is not None and format.get('abr') is not None:
1957 format['tbr'] = format.get('vbr', 0) + format.get('abr', 0)
1958 else:
b5ae35ee 1959 if format.get('vcodec') != 'none' and format.get('vbr') is None:
eb8a4433 1960 format['vbr'] = format.get('tbr') - format.get('abr', 0)
b5ae35ee 1961 if format.get('acodec') != 'none' and format.get('abr') is None:
eb8a4433 1962 format['abr'] = format.get('tbr') - format.get('vbr', 0)
1963
1964 return tuple(self._calculate_field_preference(format, field) for field in self._order)
1965
1966 def _sort_formats(self, formats, field_preference=[]):
1967 if not formats:
88acdbc2 1968 return
1d485a1a 1969 formats.sort(key=self.FormatSort(self, field_preference).calculate_preference)
59040888 1970
96a53167
S
1971 def _check_formats(self, formats, video_id):
1972 if formats:
1973 formats[:] = filter(
1974 lambda f: self._is_valid_url(
1975 f['url'], video_id,
1976 item='%s video format' % f.get('format_id') if f.get('format_id') else 'video'),
1977 formats)
1978
f5bdb444
S
1979 @staticmethod
1980 def _remove_duplicate_formats(formats):
1981 format_urls = set()
1982 unique_formats = []
1983 for f in formats:
1984 if f['url'] not in format_urls:
1985 format_urls.add(f['url'])
1986 unique_formats.append(f)
1987 formats[:] = unique_formats
1988
45024183 1989 def _is_valid_url(self, url, video_id, item='video', headers={}):
2f0f6578
S
1990 url = self._proto_relative_url(url, scheme='http:')
1991 # For now assume non HTTP(S) URLs always valid
1992 if not (url.startswith('http://') or url.startswith('https://')):
1993 return True
96a53167 1994 try:
45024183 1995 self._request_webpage(url, video_id, 'Checking %s URL' % item, headers=headers)
96a53167 1996 return True
8bdd16b4 1997 except ExtractorError as e:
25e911a9 1998 self.to_screen(
8bdd16b4 1999 '%s: %s URL is invalid, skipping: %s'
2000 % (video_id, item, error_to_compat_str(e.cause)))
25e911a9 2001 return False
96a53167 2002
20991253 2003 def http_scheme(self):
1ede5b24 2004 """ Either "http:" or "https:", depending on the user's preferences """
20991253
PH
2005 return (
2006 'http:'
a06916d9 2007 if self.get_param('prefer_insecure', False)
20991253
PH
2008 else 'https:')
2009
57c7411f 2010 def _proto_relative_url(self, url, scheme=None):
8f97a15d 2011 scheme = scheme or self.http_scheme()
2012 assert scheme.endswith(':')
2013 return sanitize_url(url, scheme=scheme[:-1])
57c7411f 2014
4094b6e3
PH
2015 def _sleep(self, timeout, video_id, msg_template=None):
2016 if msg_template is None:
f1a9d64e 2017 msg_template = '%(video_id)s: Waiting for %(timeout)s seconds'
4094b6e3
PH
2018 msg = msg_template % {'video_id': video_id, 'timeout': timeout}
2019 self.to_screen(msg)
2020 time.sleep(timeout)
2021
f983b875 2022 def _extract_f4m_formats(self, manifest_url, video_id, preference=None, quality=None, f4m_id=None,
4de61310 2023 transform_source=lambda s: fix_xml_ampersands(s).strip(),
7360c06f 2024 fatal=True, m3u8_id=None, data=None, headers={}, query={}):
a076c1f9 2025 res = self._download_xml_handle(
f036a632 2026 manifest_url, video_id, 'Downloading f4m manifest',
97f4aecf
S
2027 'Unable to download f4m manifest',
2028 # Some manifests may be malformed, e.g. prosiebensat1 generated manifests
067aa17e 2029 # (see https://github.com/ytdl-org/youtube-dl/issues/6215#issuecomment-121704244)
4de61310 2030 transform_source=transform_source,
7360c06f 2031 fatal=fatal, data=data, headers=headers, query=query)
a076c1f9 2032 if res is False:
8d29e47f 2033 return []
31bb8d3f 2034
a076c1f9
E
2035 manifest, urlh = res
2036 manifest_url = urlh.geturl()
2037
0fdbb332 2038 return self._parse_f4m_formats(
f983b875 2039 manifest, manifest_url, video_id, preference=preference, quality=quality, f4m_id=f4m_id,
448bb5f3 2040 transform_source=transform_source, fatal=fatal, m3u8_id=m3u8_id)
0fdbb332 2041
f983b875 2042 def _parse_f4m_formats(self, manifest, manifest_url, video_id, preference=None, quality=None, f4m_id=None,
0fdbb332 2043 transform_source=lambda s: fix_xml_ampersands(s).strip(),
448bb5f3 2044 fatal=True, m3u8_id=None):
f9934b96 2045 if not isinstance(manifest, xml.etree.ElementTree.Element) and not fatal:
d9eb580a
S
2046 return []
2047
7a5c1cfe 2048 # currently yt-dlp cannot decode the playerVerificationChallenge as Akamai uses Adobe Alchemy
fb72ec58 2049 akamai_pv = manifest.find('{http://ns.adobe.com/f4m/1.0}pv-2.0')
2050 if akamai_pv is not None and ';' in akamai_pv.text:
2051 playerVerificationChallenge = akamai_pv.text.split(';')[0]
2052 if playerVerificationChallenge.strip() != '':
2053 return []
2054
31bb8d3f 2055 formats = []
7a47d07c 2056 manifest_version = '1.0'
b2527359 2057 media_nodes = manifest.findall('{http://ns.adobe.com/f4m/1.0}media')
34e48bed 2058 if not media_nodes:
7a47d07c 2059 manifest_version = '2.0'
34e48bed 2060 media_nodes = manifest.findall('{http://ns.adobe.com/f4m/2.0}media')
b22ca762 2061 # Remove unsupported DRM protected media from final formats
067aa17e 2062 # rendition (see https://github.com/ytdl-org/youtube-dl/issues/8573).
b22ca762
S
2063 media_nodes = remove_encrypted_media(media_nodes)
2064 if not media_nodes:
2065 return formats
48107c19
S
2066
2067 manifest_base_url = get_base_url(manifest)
0a5685b2 2068
a6571f10 2069 bootstrap_info = xpath_element(
0a5685b2
YCH
2070 manifest, ['{http://ns.adobe.com/f4m/1.0}bootstrapInfo', '{http://ns.adobe.com/f4m/2.0}bootstrapInfo'],
2071 'bootstrap info', default=None)
2072
edd6074c
RA
2073 vcodec = None
2074 mime_type = xpath_text(
2075 manifest, ['{http://ns.adobe.com/f4m/1.0}mimeType', '{http://ns.adobe.com/f4m/2.0}mimeType'],
2076 'base URL', default=None)
2077 if mime_type and mime_type.startswith('audio/'):
2078 vcodec = 'none'
2079
b2527359 2080 for i, media_el in enumerate(media_nodes):
77b8b4e6
S
2081 tbr = int_or_none(media_el.attrib.get('bitrate'))
2082 width = int_or_none(media_el.attrib.get('width'))
2083 height = int_or_none(media_el.attrib.get('height'))
34921b43 2084 format_id = join_nonempty(f4m_id, tbr or i)
448bb5f3
YCH
2085 # If <bootstrapInfo> is present, the specified f4m is a
2086 # stream-level manifest, and only set-level manifests may refer to
2087 # external resources. See section 11.4 and section 4 of F4M spec
2088 if bootstrap_info is None:
2089 media_url = None
2090 # @href is introduced in 2.0, see section 11.6 of F4M spec
2091 if manifest_version == '2.0':
2092 media_url = media_el.attrib.get('href')
2093 if media_url is None:
2094 media_url = media_el.attrib.get('url')
31c746e5
S
2095 if not media_url:
2096 continue
cc357c4d
S
2097 manifest_url = (
2098 media_url if media_url.startswith('http://') or media_url.startswith('https://')
48107c19 2099 else ((manifest_base_url or '/'.join(manifest_url.split('/')[:-1])) + '/' + media_url))
70f0f5a8
S
2100 # If media_url is itself a f4m manifest do the recursive extraction
2101 # since bitrates in parent manifest (this one) and media_url manifest
2102 # may differ leading to inability to resolve the format by requested
2103 # bitrate in f4m downloader
240b6045
YCH
2104 ext = determine_ext(manifest_url)
2105 if ext == 'f4m':
77b8b4e6 2106 f4m_formats = self._extract_f4m_formats(
f983b875 2107 manifest_url, video_id, preference=preference, quality=quality, f4m_id=f4m_id,
77b8b4e6
S
2108 transform_source=transform_source, fatal=fatal)
2109 # Sometimes stream-level manifest contains single media entry that
2110 # does not contain any quality metadata (e.g. http://matchtv.ru/#live-player).
2111 # At the same time parent's media entry in set-level manifest may
2112 # contain it. We will copy it from parent in such cases.
2113 if len(f4m_formats) == 1:
2114 f = f4m_formats[0]
2115 f.update({
2116 'tbr': f.get('tbr') or tbr,
2117 'width': f.get('width') or width,
2118 'height': f.get('height') or height,
2119 'format_id': f.get('format_id') if not tbr else format_id,
edd6074c 2120 'vcodec': vcodec,
77b8b4e6
S
2121 })
2122 formats.extend(f4m_formats)
70f0f5a8 2123 continue
240b6045
YCH
2124 elif ext == 'm3u8':
2125 formats.extend(self._extract_m3u8_formats(
2126 manifest_url, video_id, 'mp4', preference=preference,
f983b875 2127 quality=quality, m3u8_id=m3u8_id, fatal=fatal))
240b6045 2128 continue
31bb8d3f 2129 formats.append({
77b8b4e6 2130 'format_id': format_id,
31bb8d3f 2131 'url': manifest_url,
30d0b549 2132 'manifest_url': manifest_url,
a6571f10 2133 'ext': 'flv' if bootstrap_info is not None else None,
187ee66c 2134 'protocol': 'f4m',
b2527359 2135 'tbr': tbr,
77b8b4e6
S
2136 'width': width,
2137 'height': height,
edd6074c 2138 'vcodec': vcodec,
60ca389c 2139 'preference': preference,
f983b875 2140 'quality': quality,
31bb8d3f 2141 })
31bb8d3f
JMF
2142 return formats
2143
f983b875 2144 def _m3u8_meta_format(self, m3u8_url, ext=None, preference=None, quality=None, m3u8_id=None):
16da9bbc 2145 return {
34921b43 2146 'format_id': join_nonempty(m3u8_id, 'meta'),
704df56d
PH
2147 'url': m3u8_url,
2148 'ext': ext,
2149 'protocol': 'm3u8',
37768f92 2150 'preference': preference - 100 if preference else -100,
f983b875 2151 'quality': quality,
704df56d
PH
2152 'resolution': 'multiple',
2153 'format_note': 'Quality selection URL',
16da9bbc
YCH
2154 }
2155
b5ae35ee 2156 def _report_ignoring_subs(self, name):
2157 self.report_warning(bug_reports_message(
2158 f'Ignoring subtitle tracks found in the {name} manifest; '
2159 'if any subtitle tracks are missing,'
2160 ), only_once=True)
2161
a0c3b2d5
F
2162 def _extract_m3u8_formats(self, *args, **kwargs):
2163 fmts, subs = self._extract_m3u8_formats_and_subtitles(*args, **kwargs)
2164 if subs:
b5ae35ee 2165 self._report_ignoring_subs('HLS')
a0c3b2d5
F
2166 return fmts
2167
2168 def _extract_m3u8_formats_and_subtitles(
177877c5 2169 self, m3u8_url, video_id, ext=None, entry_protocol='m3u8_native',
a0c3b2d5
F
2170 preference=None, quality=None, m3u8_id=None, note=None,
2171 errnote=None, fatal=True, live=False, data=None, headers={},
2172 query={}):
2173
dbd82a1d 2174 res = self._download_webpage_handle(
81515ad9 2175 m3u8_url, video_id,
37a3bb66 2176 note='Downloading m3u8 information' if note is None else note,
2177 errnote='Failed to download m3u8 information' if errnote is None else errnote,
7360c06f 2178 fatal=fatal, data=data, headers=headers, query=query)
cb252080 2179
dbd82a1d 2180 if res is False:
a0c3b2d5 2181 return [], {}
cb252080 2182
dbd82a1d 2183 m3u8_doc, urlh = res
37113045 2184 m3u8_url = urlh.geturl()
9cdffeeb 2185
a0c3b2d5 2186 return self._parse_m3u8_formats_and_subtitles(
cb252080 2187 m3u8_doc, m3u8_url, ext=ext, entry_protocol=entry_protocol,
310c2ed2 2188 preference=preference, quality=quality, m3u8_id=m3u8_id,
2189 note=note, errnote=errnote, fatal=fatal, live=live, data=data,
2190 headers=headers, query=query, video_id=video_id)
cb252080 2191
a0c3b2d5 2192 def _parse_m3u8_formats_and_subtitles(
42676437 2193 self, m3u8_doc, m3u8_url=None, ext=None, entry_protocol='m3u8_native',
a0c3b2d5
F
2194 preference=None, quality=None, m3u8_id=None, live=False, note=None,
2195 errnote=None, fatal=True, data=None, headers={}, query={},
2196 video_id=None):
60755938 2197 formats, subtitles = [], {}
a0c3b2d5 2198
6b993ca7 2199 has_drm = re.search('|'.join([
2200 r'#EXT-X-FAXS-CM:', # Adobe Flash Access
2201 r'#EXT-X-(?:SESSION-)?KEY:.*?URI="skd://', # Apple FairPlay
2202 ]), m3u8_doc)
a0c3b2d5 2203
60755938 2204 def format_url(url):
14f25df2 2205 return url if re.match(r'^https?://', url) else urllib.parse.urljoin(m3u8_url, url)
60755938 2206
2207 if self.get_param('hls_split_discontinuity', False):
2208 def _extract_m3u8_playlist_indices(manifest_url=None, m3u8_doc=None):
2209 if not m3u8_doc:
2210 if not manifest_url:
2211 return []
2212 m3u8_doc = self._download_webpage(
2213 manifest_url, video_id, fatal=fatal, data=data, headers=headers,
2214 note=False, errnote='Failed to download m3u8 playlist information')
2215 if m3u8_doc is False:
2216 return []
2217 return range(1 + sum(line.startswith('#EXT-X-DISCONTINUITY') for line in m3u8_doc.splitlines()))
0def7587 2218
60755938 2219 else:
2220 def _extract_m3u8_playlist_indices(*args, **kwargs):
2221 return [None]
310c2ed2 2222
cb252080
S
2223 # References:
2224 # 1. https://tools.ietf.org/html/draft-pantos-http-live-streaming-21
067aa17e
S
2225 # 2. https://github.com/ytdl-org/youtube-dl/issues/12211
2226 # 3. https://github.com/ytdl-org/youtube-dl/issues/18923
cb252080
S
2227
2228 # We should try extracting formats only from master playlists [1, 4.3.4],
2229 # i.e. playlists that describe available qualities. On the other hand
2230 # media playlists [1, 4.3.3] should be returned as is since they contain
2231 # just the media without qualities renditions.
9cdffeeb 2232 # Fortunately, master playlist can be easily distinguished from media
cb252080 2233 # playlist based on particular tags availability. As of [1, 4.3.3, 4.3.4]
a0566bbf 2234 # master playlist tags MUST NOT appear in a media playlist and vice versa.
cb252080
S
2235 # As of [1, 4.3.3.1] #EXT-X-TARGETDURATION tag is REQUIRED for every
2236 # media playlist and MUST NOT appear in master playlist thus we can
2237 # clearly detect media playlist with this criterion.
2238
9cdffeeb 2239 if '#EXT-X-TARGETDURATION' in m3u8_doc: # media playlist, return as is
60755938 2240 formats = [{
34921b43 2241 'format_id': join_nonempty(m3u8_id, idx),
60755938 2242 'format_index': idx,
42676437 2243 'url': m3u8_url or encode_data_uri(m3u8_doc.encode('utf-8'), 'application/x-mpegurl'),
60755938 2244 'ext': ext,
2245 'protocol': entry_protocol,
2246 'preference': preference,
2247 'quality': quality,
88acdbc2 2248 'has_drm': has_drm,
60755938 2249 } for idx in _extract_m3u8_playlist_indices(m3u8_doc=m3u8_doc)]
310c2ed2 2250
a0c3b2d5 2251 return formats, subtitles
cb252080
S
2252
2253 groups = {}
2254 last_stream_inf = {}
2255
2256 def extract_media(x_media_line):
2257 media = parse_m3u8_attributes(x_media_line)
2258 # As per [1, 4.3.4.1] TYPE, GROUP-ID and NAME are REQUIRED
2259 media_type, group_id, name = media.get('TYPE'), media.get('GROUP-ID'), media.get('NAME')
2260 if not (media_type and group_id and name):
2261 return
2262 groups.setdefault(group_id, []).append(media)
a0c3b2d5
F
2263 # <https://tools.ietf.org/html/rfc8216#section-4.3.4.1>
2264 if media_type == 'SUBTITLES':
3907333c 2265 # According to RFC 8216 §4.3.4.2.1, URI is REQUIRED in the
2266 # EXT-X-MEDIA tag if the media type is SUBTITLES.
2267 # However, lack of URI has been spotted in the wild.
2268 # e.g. NebulaIE; see https://github.com/yt-dlp/yt-dlp/issues/339
2269 if not media.get('URI'):
2270 return
a0c3b2d5
F
2271 url = format_url(media['URI'])
2272 sub_info = {
2273 'url': url,
2274 'ext': determine_ext(url),
2275 }
4a2f19ab
F
2276 if sub_info['ext'] == 'm3u8':
2277 # Per RFC 8216 §3.1, the only possible subtitle format m3u8
2278 # files may contain is WebVTT:
2279 # <https://tools.ietf.org/html/rfc8216#section-3.1>
2280 sub_info['ext'] = 'vtt'
2281 sub_info['protocol'] = 'm3u8_native'
37a3bb66 2282 lang = media.get('LANGUAGE') or 'und'
a0c3b2d5 2283 subtitles.setdefault(lang, []).append(sub_info)
cb252080
S
2284 if media_type not in ('VIDEO', 'AUDIO'):
2285 return
2286 media_url = media.get('URI')
2287 if media_url:
310c2ed2 2288 manifest_url = format_url(media_url)
60755938 2289 formats.extend({
34921b43 2290 'format_id': join_nonempty(m3u8_id, group_id, name, idx),
60755938 2291 'format_note': name,
2292 'format_index': idx,
2293 'url': manifest_url,
2294 'manifest_url': m3u8_url,
2295 'language': media.get('LANGUAGE'),
2296 'ext': ext,
2297 'protocol': entry_protocol,
2298 'preference': preference,
2299 'quality': quality,
2300 'vcodec': 'none' if media_type == 'AUDIO' else None,
2301 } for idx in _extract_m3u8_playlist_indices(manifest_url))
cb252080
S
2302
2303 def build_stream_name():
2304 # Despite specification does not mention NAME attribute for
3019cb0c
S
2305 # EXT-X-STREAM-INF tag it still sometimes may be present (see [1]
2306 # or vidio test in TestInfoExtractor.test_parse_m3u8_formats)
ddd258f9 2307 # 1. http://www.vidio.com/watch/165683-dj_ambred-booyah-live-2015
cb252080
S
2308 stream_name = last_stream_inf.get('NAME')
2309 if stream_name:
2310 return stream_name
2311 # If there is no NAME in EXT-X-STREAM-INF it will be obtained
2312 # from corresponding rendition group
2313 stream_group_id = last_stream_inf.get('VIDEO')
2314 if not stream_group_id:
2315 return
2316 stream_group = groups.get(stream_group_id)
2317 if not stream_group:
2318 return stream_group_id
2319 rendition = stream_group[0]
2320 return rendition.get('NAME') or stream_group_id
2321
379306ef 2322 # parse EXT-X-MEDIA tags before EXT-X-STREAM-INF in order to have the
2bfc1d9d
RA
2323 # chance to detect video only formats when EXT-X-STREAM-INF tags
2324 # precede EXT-X-MEDIA tags in HLS manifest such as [3].
2325 for line in m3u8_doc.splitlines():
2326 if line.startswith('#EXT-X-MEDIA:'):
2327 extract_media(line)
2328
704df56d
PH
2329 for line in m3u8_doc.splitlines():
2330 if line.startswith('#EXT-X-STREAM-INF:'):
cb252080 2331 last_stream_inf = parse_m3u8_attributes(line)
704df56d
PH
2332 elif line.startswith('#') or not line.strip():
2333 continue
2334 else:
9c99bef7 2335 tbr = float_or_none(
3089bc74
S
2336 last_stream_inf.get('AVERAGE-BANDWIDTH')
2337 or last_stream_inf.get('BANDWIDTH'), scale=1000)
30d0b549 2338 manifest_url = format_url(line.strip())
5ef62fc4 2339
60755938 2340 for idx in _extract_m3u8_playlist_indices(manifest_url):
2341 format_id = [m3u8_id, None, idx]
310c2ed2 2342 # Bandwidth of live streams may differ over time thus making
2343 # format_id unpredictable. So it's better to keep provided
2344 # format_id intact.
2345 if not live:
60755938 2346 stream_name = build_stream_name()
34921b43 2347 format_id[1] = stream_name or '%d' % (tbr or len(formats))
310c2ed2 2348 f = {
34921b43 2349 'format_id': join_nonempty(*format_id),
60755938 2350 'format_index': idx,
310c2ed2 2351 'url': manifest_url,
2352 'manifest_url': m3u8_url,
2353 'tbr': tbr,
2354 'ext': ext,
2355 'fps': float_or_none(last_stream_inf.get('FRAME-RATE')),
2356 'protocol': entry_protocol,
2357 'preference': preference,
2358 'quality': quality,
2359 }
2360 resolution = last_stream_inf.get('RESOLUTION')
2361 if resolution:
2362 mobj = re.search(r'(?P<width>\d+)[xX](?P<height>\d+)', resolution)
2363 if mobj:
2364 f['width'] = int(mobj.group('width'))
2365 f['height'] = int(mobj.group('height'))
2366 # Unified Streaming Platform
2367 mobj = re.search(
2368 r'audio.*?(?:%3D|=)(\d+)(?:-video.*?(?:%3D|=)(\d+))?', f['url'])
2369 if mobj:
2370 abr, vbr = mobj.groups()
2371 abr, vbr = float_or_none(abr, 1000), float_or_none(vbr, 1000)
2372 f.update({
2373 'vbr': vbr,
2374 'abr': abr,
2375 })
2376 codecs = parse_codecs(last_stream_inf.get('CODECS'))
2377 f.update(codecs)
2378 audio_group_id = last_stream_inf.get('AUDIO')
2379 # As per [1, 4.3.4.1.1] any EXT-X-STREAM-INF tag which
2380 # references a rendition group MUST have a CODECS attribute.
62b58c09 2381 # However, this is not always respected. E.g. [2]
310c2ed2 2382 # contains EXT-X-STREAM-INF tag which references AUDIO
2383 # rendition group but does not have CODECS and despite
2384 # referencing an audio group it represents a complete
2385 # (with audio and video) format. So, for such cases we will
2386 # ignore references to rendition groups and treat them
2387 # as complete formats.
2388 if audio_group_id and codecs and f.get('vcodec') != 'none':
2389 audio_group = groups.get(audio_group_id)
2390 if audio_group and audio_group[0].get('URI'):
2391 # TODO: update acodec for audio only formats with
2392 # the same GROUP-ID
2393 f['acodec'] = 'none'
fc21af50 2394 if not f.get('ext'):
2395 f['ext'] = 'm4a' if f.get('vcodec') == 'none' else 'mp4'
310c2ed2 2396 formats.append(f)
2397
2398 # for DailyMotion
2399 progressive_uri = last_stream_inf.get('PROGRESSIVE-URI')
2400 if progressive_uri:
2401 http_f = f.copy()
2402 del http_f['manifest_url']
2403 http_f.update({
2404 'format_id': f['format_id'].replace('hls-', 'http-'),
2405 'protocol': 'http',
2406 'url': progressive_uri,
2407 })
2408 formats.append(http_f)
5ef62fc4 2409
cb252080 2410 last_stream_inf = {}
a0c3b2d5 2411 return formats, subtitles
704df56d 2412
3cf4b91d
C
2413 def _extract_m3u8_vod_duration(
2414 self, m3u8_vod_url, video_id, note=None, errnote=None, data=None, headers={}, query={}):
2415
2416 m3u8_vod = self._download_webpage(
2417 m3u8_vod_url, video_id,
2418 note='Downloading m3u8 VOD manifest' if note is None else note,
2419 errnote='Failed to download VOD manifest' if errnote is None else errnote,
2420 fatal=False, data=data, headers=headers, query=query)
2421
2422 return self._parse_m3u8_vod_duration(m3u8_vod or '', video_id)
2423
2424 def _parse_m3u8_vod_duration(self, m3u8_vod, video_id):
2425 if '#EXT-X-PLAYLIST-TYPE:VOD' not in m3u8_vod:
2426 return None
2427
2428 return int(sum(
2429 float(line[len('#EXTINF:'):].split(',')[0])
2430 for line in m3u8_vod.splitlines() if line.startswith('#EXTINF:'))) or None
2431
a107193e
S
2432 @staticmethod
2433 def _xpath_ns(path, namespace=None):
2434 if not namespace:
2435 return path
2436 out = []
2437 for c in path.split('/'):
2438 if not c or c == '.':
2439 out.append(c)
2440 else:
2441 out.append('{%s}%s' % (namespace, c))
2442 return '/'.join(out)
2443
da1c94ee 2444 def _extract_smil_formats_and_subtitles(self, smil_url, video_id, fatal=True, f4m_params=None, transform_source=None):
a076c1f9
E
2445 res = self._download_smil(smil_url, video_id, fatal=fatal, transform_source=transform_source)
2446 if res is False:
995029a1 2447 assert not fatal
774a46c5 2448 return [], {}
e89a2aab 2449
a076c1f9
E
2450 smil, urlh = res
2451 smil_url = urlh.geturl()
2452
17712eeb 2453 namespace = self._parse_smil_namespace(smil)
a107193e 2454
da1c94ee 2455 fmts = self._parse_smil_formats(
a107193e 2456 smil, smil_url, video_id, namespace=namespace, f4m_params=f4m_params)
da1c94ee
F
2457 subs = self._parse_smil_subtitles(
2458 smil, namespace=namespace)
2459
2460 return fmts, subs
2461
2462 def _extract_smil_formats(self, *args, **kwargs):
2463 fmts, subs = self._extract_smil_formats_and_subtitles(*args, **kwargs)
2464 if subs:
b5ae35ee 2465 self._report_ignoring_subs('SMIL')
da1c94ee 2466 return fmts
a107193e
S
2467
2468 def _extract_smil_info(self, smil_url, video_id, fatal=True, f4m_params=None):
a076c1f9
E
2469 res = self._download_smil(smil_url, video_id, fatal=fatal)
2470 if res is False:
a107193e 2471 return {}
a076c1f9
E
2472
2473 smil, urlh = res
2474 smil_url = urlh.geturl()
2475
a107193e
S
2476 return self._parse_smil(smil, smil_url, video_id, f4m_params=f4m_params)
2477
09f572fb 2478 def _download_smil(self, smil_url, video_id, fatal=True, transform_source=None):
a076c1f9 2479 return self._download_xml_handle(
a107193e 2480 smil_url, video_id, 'Downloading SMIL file',
09f572fb 2481 'Unable to download SMIL file', fatal=fatal, transform_source=transform_source)
a107193e
S
2482
2483 def _parse_smil(self, smil, smil_url, video_id, f4m_params=None):
17712eeb 2484 namespace = self._parse_smil_namespace(smil)
a107193e
S
2485
2486 formats = self._parse_smil_formats(
2487 smil, smil_url, video_id, namespace=namespace, f4m_params=f4m_params)
2488 subtitles = self._parse_smil_subtitles(smil, namespace=namespace)
2489
2490 video_id = os.path.splitext(url_basename(smil_url))[0]
2491 title = None
2492 description = None
647eab45 2493 upload_date = None
a107193e
S
2494 for meta in smil.findall(self._xpath_ns('./head/meta', namespace)):
2495 name = meta.attrib.get('name')
2496 content = meta.attrib.get('content')
2497 if not name or not content:
2498 continue
2499 if not title and name == 'title':
2500 title = content
2501 elif not description and name in ('description', 'abstract'):
2502 description = content
647eab45
S
2503 elif not upload_date and name == 'date':
2504 upload_date = unified_strdate(content)
a107193e 2505
1e5bcdec
S
2506 thumbnails = [{
2507 'id': image.get('type'),
2508 'url': image.get('src'),
2509 'width': int_or_none(image.get('width')),
2510 'height': int_or_none(image.get('height')),
2511 } for image in smil.findall(self._xpath_ns('.//image', namespace)) if image.get('src')]
2512
a107193e
S
2513 return {
2514 'id': video_id,
2515 'title': title or video_id,
2516 'description': description,
647eab45 2517 'upload_date': upload_date,
1e5bcdec 2518 'thumbnails': thumbnails,
a107193e
S
2519 'formats': formats,
2520 'subtitles': subtitles,
2521 }
2522
17712eeb
S
2523 def _parse_smil_namespace(self, smil):
2524 return self._search_regex(
2525 r'(?i)^{([^}]+)?}smil$', smil.tag, 'namespace', default=None)
2526
f877c6ae 2527 def _parse_smil_formats(self, smil, smil_url, video_id, namespace=None, f4m_params=None, transform_rtmp_url=None):
a107193e
S
2528 base = smil_url
2529 for meta in smil.findall(self._xpath_ns('./head/meta', namespace)):
2530 b = meta.get('base') or meta.get('httpBase')
2531 if b:
2532 base = b
2533 break
e89a2aab
S
2534
2535 formats = []
2536 rtmp_count = 0
a107193e 2537 http_count = 0
7f32e5dc 2538 m3u8_count = 0
9359f3d4 2539 imgs_count = 0
a107193e 2540
9359f3d4 2541 srcs = set()
ad96b4c8
YCH
2542 media = smil.findall(self._xpath_ns('.//video', namespace)) + smil.findall(self._xpath_ns('.//audio', namespace))
2543 for medium in media:
2544 src = medium.get('src')
81e1c4e2 2545 if not src or src in srcs:
a107193e 2546 continue
9359f3d4 2547 srcs.add(src)
a107193e 2548
ad96b4c8
YCH
2549 bitrate = float_or_none(medium.get('system-bitrate') or medium.get('systemBitrate'), 1000)
2550 filesize = int_or_none(medium.get('size') or medium.get('fileSize'))
2551 width = int_or_none(medium.get('width'))
2552 height = int_or_none(medium.get('height'))
2553 proto = medium.get('proto')
2554 ext = medium.get('ext')
a107193e 2555 src_ext = determine_ext(src)
ad96b4c8 2556 streamer = medium.get('streamer') or base
a107193e
S
2557
2558 if proto == 'rtmp' or streamer.startswith('rtmp'):
2559 rtmp_count += 1
2560 formats.append({
2561 'url': streamer,
2562 'play_path': src,
2563 'ext': 'flv',
2564 'format_id': 'rtmp-%d' % (rtmp_count if bitrate is None else bitrate),
2565 'tbr': bitrate,
2566 'filesize': filesize,
2567 'width': width,
2568 'height': height,
2569 })
f877c6ae
YCH
2570 if transform_rtmp_url:
2571 streamer, src = transform_rtmp_url(streamer, src)
2572 formats[-1].update({
2573 'url': streamer,
2574 'play_path': src,
2575 })
a107193e
S
2576 continue
2577
14f25df2 2578 src_url = src if src.startswith('http') else urllib.parse.urljoin(base, src)
c349456e 2579 src_url = src_url.strip()
a107193e
S
2580
2581 if proto == 'm3u8' or src_ext == 'm3u8':
7f32e5dc 2582 m3u8_formats = self._extract_m3u8_formats(
2583 src_url, video_id, ext or 'mp4', m3u8_id='hls', fatal=False)
2584 if len(m3u8_formats) == 1:
2585 m3u8_count += 1
2586 m3u8_formats[0].update({
2587 'format_id': 'hls-%d' % (m3u8_count if bitrate is None else bitrate),
2588 'tbr': bitrate,
2589 'width': width,
2590 'height': height,
2591 })
2592 formats.extend(m3u8_formats)
bd21ead2 2593 elif src_ext == 'f4m':
a107193e
S
2594 f4m_url = src_url
2595 if not f4m_params:
2596 f4m_params = {
2597 'hdcore': '3.2.0',
2598 'plugin': 'flowplayer-3.2.0.1',
2599 }
2600 f4m_url += '&' if '?' in f4m_url else '?'
14f25df2 2601 f4m_url += urllib.parse.urlencode(f4m_params)
7e5edcfd 2602 formats.extend(self._extract_f4m_formats(f4m_url, video_id, f4m_id='hds', fatal=False))
bd21ead2
RA
2603 elif src_ext == 'mpd':
2604 formats.extend(self._extract_mpd_formats(
2605 src_url, video_id, mpd_id='dash', fatal=False))
2606 elif re.search(r'\.ism/[Mm]anifest', src_url):
2607 formats.extend(self._extract_ism_formats(
2608 src_url, video_id, ism_id='mss', fatal=False))
2609 elif src_url.startswith('http') and self._is_valid_url(src, video_id):
a107193e
S
2610 http_count += 1
2611 formats.append({
2612 'url': src_url,
2613 'ext': ext or src_ext or 'flv',
2614 'format_id': 'http-%d' % (bitrate or http_count),
2615 'tbr': bitrate,
2616 'filesize': filesize,
2617 'width': width,
2618 'height': height,
2619 })
63757032 2620
9359f3d4
F
2621 for medium in smil.findall(self._xpath_ns('.//imagestream', namespace)):
2622 src = medium.get('src')
2623 if not src or src in srcs:
2624 continue
2625 srcs.add(src)
2626
2627 imgs_count += 1
2628 formats.append({
2629 'format_id': 'imagestream-%d' % (imgs_count),
2630 'url': src,
2631 'ext': mimetype2ext(medium.get('type')),
2632 'acodec': 'none',
2633 'vcodec': 'none',
2634 'width': int_or_none(medium.get('width')),
2635 'height': int_or_none(medium.get('height')),
2636 'format_note': 'SMIL storyboards',
2637 })
2638
e89a2aab
S
2639 return formats
2640
ce00af87 2641 def _parse_smil_subtitles(self, smil, namespace=None, subtitles_lang='en'):
d413095f 2642 urls = []
a107193e
S
2643 subtitles = {}
2644 for num, textstream in enumerate(smil.findall(self._xpath_ns('.//textstream', namespace))):
2645 src = textstream.get('src')
d413095f 2646 if not src or src in urls:
a107193e 2647 continue
d413095f 2648 urls.append(src)
df634be2 2649 ext = textstream.get('ext') or mimetype2ext(textstream.get('type')) or determine_ext(src)
03bc7237 2650 lang = textstream.get('systemLanguage') or textstream.get('systemLanguageName') or textstream.get('lang') or subtitles_lang
a107193e
S
2651 subtitles.setdefault(lang, []).append({
2652 'url': src,
2653 'ext': ext,
2654 })
2655 return subtitles
63757032 2656
47a5cb77 2657 def _extract_xspf_playlist(self, xspf_url, playlist_id, fatal=True):
a076c1f9 2658 res = self._download_xml_handle(
47a5cb77 2659 xspf_url, playlist_id, 'Downloading xpsf playlist',
942acef5 2660 'Unable to download xspf manifest', fatal=fatal)
a076c1f9 2661 if res is False:
942acef5 2662 return []
a076c1f9
E
2663
2664 xspf, urlh = res
2665 xspf_url = urlh.geturl()
2666
47a5cb77
S
2667 return self._parse_xspf(
2668 xspf, playlist_id, xspf_url=xspf_url,
2669 xspf_base_url=base_url(xspf_url))
8d6765cf 2670
47a5cb77 2671 def _parse_xspf(self, xspf_doc, playlist_id, xspf_url=None, xspf_base_url=None):
8d6765cf
S
2672 NS_MAP = {
2673 'xspf': 'http://xspf.org/ns/0/',
2674 's1': 'http://static.streamone.nl/player/ns/0',
2675 }
2676
2677 entries = []
47a5cb77 2678 for track in xspf_doc.findall(xpath_with_ns('./xspf:trackList/xspf:track', NS_MAP)):
8d6765cf 2679 title = xpath_text(
98044462 2680 track, xpath_with_ns('./xspf:title', NS_MAP), 'title', default=playlist_id)
8d6765cf
S
2681 description = xpath_text(
2682 track, xpath_with_ns('./xspf:annotation', NS_MAP), 'description')
2683 thumbnail = xpath_text(
2684 track, xpath_with_ns('./xspf:image', NS_MAP), 'thumbnail')
2685 duration = float_or_none(
2686 xpath_text(track, xpath_with_ns('./xspf:duration', NS_MAP), 'duration'), 1000)
2687
47a5cb77
S
2688 formats = []
2689 for location in track.findall(xpath_with_ns('./xspf:location', NS_MAP)):
2690 format_url = urljoin(xspf_base_url, location.text)
2691 if not format_url:
2692 continue
2693 formats.append({
2694 'url': format_url,
2695 'manifest_url': xspf_url,
2696 'format_id': location.get(xpath_with_ns('s1:label', NS_MAP)),
2697 'width': int_or_none(location.get(xpath_with_ns('s1:width', NS_MAP))),
2698 'height': int_or_none(location.get(xpath_with_ns('s1:height', NS_MAP))),
2699 })
8d6765cf
S
2700 self._sort_formats(formats)
2701
2702 entries.append({
2703 'id': playlist_id,
2704 'title': title,
2705 'description': description,
2706 'thumbnail': thumbnail,
2707 'duration': duration,
2708 'formats': formats,
2709 })
2710 return entries
2711
171e59ed
F
2712 def _extract_mpd_formats(self, *args, **kwargs):
2713 fmts, subs = self._extract_mpd_formats_and_subtitles(*args, **kwargs)
2714 if subs:
b5ae35ee 2715 self._report_ignoring_subs('DASH')
171e59ed
F
2716 return fmts
2717
2718 def _extract_mpd_formats_and_subtitles(
2719 self, mpd_url, video_id, mpd_id=None, note=None, errnote=None,
2720 fatal=True, data=None, headers={}, query={}):
47a5cb77 2721 res = self._download_xml_handle(
1bac3455 2722 mpd_url, video_id,
37a3bb66 2723 note='Downloading MPD manifest' if note is None else note,
2724 errnote='Failed to download MPD manifest' if errnote is None else errnote,
7360c06f 2725 fatal=fatal, data=data, headers=headers, query=query)
1bac3455 2726 if res is False:
171e59ed 2727 return [], {}
47a5cb77 2728 mpd_doc, urlh = res
c25720ef 2729 if mpd_doc is None:
171e59ed 2730 return [], {}
779da8e3
E
2731
2732 # We could have been redirected to a new url when we retrieved our mpd file.
2733 mpd_url = urlh.geturl()
2734 mpd_base_url = base_url(mpd_url)
1bac3455 2735
171e59ed 2736 return self._parse_mpd_formats_and_subtitles(
545cc85d 2737 mpd_doc, mpd_id, mpd_base_url, mpd_url)
2d2fa82d 2738
171e59ed
F
2739 def _parse_mpd_formats(self, *args, **kwargs):
2740 fmts, subs = self._parse_mpd_formats_and_subtitles(*args, **kwargs)
2741 if subs:
b5ae35ee 2742 self._report_ignoring_subs('DASH')
171e59ed
F
2743 return fmts
2744
2745 def _parse_mpd_formats_and_subtitles(
2746 self, mpd_doc, mpd_id=None, mpd_base_url='', mpd_url=None):
f0948348
S
2747 """
2748 Parse formats from MPD manifest.
2749 References:
2750 1. MPEG-DASH Standard, ISO/IEC 23009-1:2014(E),
2751 http://standards.iso.org/ittf/PubliclyAvailableStandards/c065274_ISO_IEC_23009-1_2014.zip
2752 2. https://en.wikipedia.org/wiki/Dynamic_Adaptive_Streaming_over_HTTP
2753 """
a06916d9 2754 if not self.get_param('dynamic_mpd', True):
78895bd3 2755 if mpd_doc.get('type') == 'dynamic':
171e59ed 2756 return [], {}
2d2fa82d 2757
91cb6b50 2758 namespace = self._search_regex(r'(?i)^{([^}]+)?}MPD$', mpd_doc.tag, 'namespace', default=None)
f14be228 2759
2760 def _add_ns(path):
2761 return self._xpath_ns(path, namespace)
2762
675d0016 2763 def is_drm_protected(element):
2764 return element.find(_add_ns('ContentProtection')) is not None
2765
1bac3455 2766 def extract_multisegment_info(element, ms_parent_info):
2767 ms_info = ms_parent_info.copy()
b4c1d6e8
S
2768
2769 # As per [1, 5.3.9.2.2] SegmentList and SegmentTemplate share some
2770 # common attributes and elements. We will only extract relevant
2771 # for us.
2772 def extract_common(source):
2773 segment_timeline = source.find(_add_ns('SegmentTimeline'))
2774 if segment_timeline is not None:
2775 s_e = segment_timeline.findall(_add_ns('S'))
2776 if s_e:
2777 ms_info['total_number'] = 0
2778 ms_info['s'] = []
2779 for s in s_e:
2780 r = int(s.get('r', 0))
2781 ms_info['total_number'] += 1 + r
2782 ms_info['s'].append({
2783 't': int(s.get('t', 0)),
2784 # @d is mandatory (see [1, 5.3.9.6.2, Table 17, page 60])
2785 'd': int(s.attrib['d']),
2786 'r': r,
2787 })
2788 start_number = source.get('startNumber')
2789 if start_number:
2790 ms_info['start_number'] = int(start_number)
2791 timescale = source.get('timescale')
2792 if timescale:
2793 ms_info['timescale'] = int(timescale)
2794 segment_duration = source.get('duration')
2795 if segment_duration:
48504785 2796 ms_info['segment_duration'] = float(segment_duration)
b4c1d6e8
S
2797
2798 def extract_Initialization(source):
2799 initialization = source.find(_add_ns('Initialization'))
2800 if initialization is not None:
2801 ms_info['initialization_url'] = initialization.attrib['sourceURL']
2802
f14be228 2803 segment_list = element.find(_add_ns('SegmentList'))
1bac3455 2804 if segment_list is not None:
b4c1d6e8
S
2805 extract_common(segment_list)
2806 extract_Initialization(segment_list)
f14be228 2807 segment_urls_e = segment_list.findall(_add_ns('SegmentURL'))
1bac3455 2808 if segment_urls_e:
2809 ms_info['segment_urls'] = [segment.attrib['media'] for segment in segment_urls_e]
1bac3455 2810 else:
f14be228 2811 segment_template = element.find(_add_ns('SegmentTemplate'))
1bac3455 2812 if segment_template is not None:
b4c1d6e8 2813 extract_common(segment_template)
e228616c
S
2814 media = segment_template.get('media')
2815 if media:
2816 ms_info['media'] = media
1bac3455 2817 initialization = segment_template.get('initialization')
2818 if initialization:
e228616c 2819 ms_info['initialization'] = initialization
1bac3455 2820 else:
b4c1d6e8 2821 extract_Initialization(segment_template)
1bac3455 2822 return ms_info
b323e170 2823
1bac3455 2824 mpd_duration = parse_duration(mpd_doc.get('mediaPresentationDuration'))
6251555f 2825 formats, subtitles = [], {}
234416e4 2826 stream_numbers = collections.defaultdict(int)
f14be228 2827 for period in mpd_doc.findall(_add_ns('Period')):
1bac3455 2828 period_duration = parse_duration(period.get('duration')) or mpd_duration
2829 period_ms_info = extract_multisegment_info(period, {
2830 'start_number': 1,
2831 'timescale': 1,
2832 })
f14be228 2833 for adaptation_set in period.findall(_add_ns('AdaptationSet')):
1bac3455 2834 adaption_set_ms_info = extract_multisegment_info(adaptation_set, period_ms_info)
f14be228 2835 for representation in adaptation_set.findall(_add_ns('Representation')):
1bac3455 2836 representation_attrib = adaptation_set.attrib.copy()
2837 representation_attrib.update(representation.attrib)
f0948348 2838 # According to [1, 5.3.7.2, Table 9, page 41], @mimeType is mandatory
a6c8b759 2839 mime_type = representation_attrib['mimeType']
171e59ed
F
2840 content_type = representation_attrib.get('contentType', mime_type.split('/')[0])
2841
21633673 2842 codec_str = representation_attrib.get('codecs', '')
2843 # Some kind of binary subtitle found in some youtube livestreams
2844 if mime_type == 'application/x-rawcc':
2845 codecs = {'scodec': codec_str}
2846 else:
2847 codecs = parse_codecs(codec_str)
be2fc5b2 2848 if content_type not in ('video', 'audio', 'text'):
2849 if mime_type == 'image/jpeg':
a8731fcc 2850 content_type = mime_type
21633673 2851 elif codecs.get('vcodec', 'none') != 'none':
4afa3ec4 2852 content_type = 'video'
21633673 2853 elif codecs.get('acodec', 'none') != 'none':
4afa3ec4 2854 content_type = 'audio'
3fe75fdc 2855 elif codecs.get('scodec', 'none') != 'none':
be2fc5b2 2856 content_type = 'text'
6993f78d 2857 elif mimetype2ext(mime_type) in ('tt', 'dfxp', 'ttml', 'xml', 'json'):
2858 content_type = 'text'
cdb19aa4 2859 else:
be2fc5b2 2860 self.report_warning('Unknown MIME type %s in DASH manifest' % mime_type)
2861 continue
2862
2863 base_url = ''
2864 for element in (representation, adaptation_set, period, mpd_doc):
2865 base_url_e = element.find(_add_ns('BaseURL'))
47046464 2866 if try_call(lambda: base_url_e.text) is not None:
be2fc5b2 2867 base_url = base_url_e.text + base_url
2868 if re.match(r'^https?://', base_url):
2869 break
f9cc0161 2870 if mpd_base_url and base_url.startswith('/'):
14f25df2 2871 base_url = urllib.parse.urljoin(mpd_base_url, base_url)
f9cc0161
D
2872 elif mpd_base_url and not re.match(r'^https?://', base_url):
2873 if not mpd_base_url.endswith('/'):
be2fc5b2 2874 mpd_base_url += '/'
2875 base_url = mpd_base_url + base_url
2876 representation_id = representation_attrib.get('id')
2877 lang = representation_attrib.get('lang')
2878 url_el = representation.find(_add_ns('BaseURL'))
2879 filesize = int_or_none(url_el.attrib.get('{http://youtube.com/yt/2012/10/10}contentLength') if url_el is not None else None)
2880 bandwidth = int_or_none(representation_attrib.get('bandwidth'))
2881 if representation_id is not None:
2882 format_id = representation_id
2883 else:
2884 format_id = content_type
2885 if mpd_id:
2886 format_id = mpd_id + '-' + format_id
2887 if content_type in ('video', 'audio'):
2888 f = {
2889 'format_id': format_id,
2890 'manifest_url': mpd_url,
2891 'ext': mimetype2ext(mime_type),
2892 'width': int_or_none(representation_attrib.get('width')),
2893 'height': int_or_none(representation_attrib.get('height')),
2894 'tbr': float_or_none(bandwidth, 1000),
2895 'asr': int_or_none(representation_attrib.get('audioSamplingRate')),
2896 'fps': int_or_none(representation_attrib.get('frameRate')),
2897 'language': lang if lang not in ('mul', 'und', 'zxx', 'mis') else None,
2898 'format_note': 'DASH %s' % content_type,
2899 'filesize': filesize,
2900 'container': mimetype2ext(mime_type) + '_dash',
4afa3ec4 2901 **codecs
be2fc5b2 2902 }
be2fc5b2 2903 elif content_type == 'text':
2904 f = {
2905 'ext': mimetype2ext(mime_type),
2906 'manifest_url': mpd_url,
2907 'filesize': filesize,
2908 }
2909 elif content_type == 'image/jpeg':
2910 # See test case in VikiIE
2911 # https://www.viki.com/videos/1175236v-choosing-spouse-by-lottery-episode-1
2912 f = {
2913 'format_id': format_id,
2914 'ext': 'mhtml',
2915 'manifest_url': mpd_url,
2916 'format_note': 'DASH storyboards (jpeg)',
2917 'acodec': 'none',
2918 'vcodec': 'none',
2919 }
88acdbc2 2920 if is_drm_protected(adaptation_set) or is_drm_protected(representation):
2921 f['has_drm'] = True
be2fc5b2 2922 representation_ms_info = extract_multisegment_info(representation, adaption_set_ms_info)
2923
2924 def prepare_template(template_name, identifiers):
2925 tmpl = representation_ms_info[template_name]
0cb0fdbb 2926 if representation_id is not None:
2927 tmpl = tmpl.replace('$RepresentationID$', representation_id)
be2fc5b2 2928 # First of, % characters outside $...$ templates
2929 # must be escaped by doubling for proper processing
2930 # by % operator string formatting used further (see
2931 # https://github.com/ytdl-org/youtube-dl/issues/16867).
2932 t = ''
2933 in_template = False
2934 for c in tmpl:
2935 t += c
2936 if c == '$':
2937 in_template = not in_template
2938 elif c == '%' and not in_template:
eca1f0d1 2939 t += c
be2fc5b2 2940 # Next, $...$ templates are translated to their
2941 # %(...) counterparts to be used with % operator
be2fc5b2 2942 t = re.sub(r'\$(%s)\$' % '|'.join(identifiers), r'%(\1)d', t)
2943 t = re.sub(r'\$(%s)%%([^$]+)\$' % '|'.join(identifiers), r'%(\1)\2', t)
2944 t.replace('$$', '$')
2945 return t
2946
2947 # @initialization is a regular template like @media one
2948 # so it should be handled just the same way (see
2949 # https://github.com/ytdl-org/youtube-dl/issues/11605)
2950 if 'initialization' in representation_ms_info:
2951 initialization_template = prepare_template(
2952 'initialization',
2953 # As per [1, 5.3.9.4.2, Table 15, page 54] $Number$ and
2954 # $Time$ shall not be included for @initialization thus
2955 # only $Bandwidth$ remains
2956 ('Bandwidth', ))
2957 representation_ms_info['initialization_url'] = initialization_template % {
2958 'Bandwidth': bandwidth,
2959 }
2960
2961 def location_key(location):
2962 return 'url' if re.match(r'^https?://', location) else 'path'
2963
2964 if 'segment_urls' not in representation_ms_info and 'media' in representation_ms_info:
2965
2966 media_template = prepare_template('media', ('Number', 'Bandwidth', 'Time'))
2967 media_location_key = location_key(media_template)
2968
2969 # As per [1, 5.3.9.4.4, Table 16, page 55] $Number$ and $Time$
2970 # can't be used at the same time
2971 if '%(Number' in media_template and 's' not in representation_ms_info:
2972 segment_duration = None
2973 if 'total_number' not in representation_ms_info and 'segment_duration' in representation_ms_info:
2974 segment_duration = float_or_none(representation_ms_info['segment_duration'], representation_ms_info['timescale'])
ffa89477 2975 representation_ms_info['total_number'] = int(math.ceil(
2976 float_or_none(period_duration, segment_duration, default=0)))
be2fc5b2 2977 representation_ms_info['fragments'] = [{
2978 media_location_key: media_template % {
2979 'Number': segment_number,
2980 'Bandwidth': bandwidth,
2981 },
2982 'duration': segment_duration,
2983 } for segment_number in range(
2984 representation_ms_info['start_number'],
2985 representation_ms_info['total_number'] + representation_ms_info['start_number'])]
2986 else:
2987 # $Number*$ or $Time$ in media template with S list available
2988 # Example $Number*$: http://www.svtplay.se/klipp/9023742/stopptid-om-bjorn-borg
2989 # Example $Time$: https://play.arkena.com/embed/avp/v2/player/media/b41dda37-d8e7-4d3f-b1b5-9a9db578bdfe/1/129411
2990 representation_ms_info['fragments'] = []
2991 segment_time = 0
2992 segment_d = None
2993 segment_number = representation_ms_info['start_number']
2994
2995 def add_segment_url():
2996 segment_url = media_template % {
2997 'Time': segment_time,
2998 'Bandwidth': bandwidth,
2999 'Number': segment_number,
3000 }
3001 representation_ms_info['fragments'].append({
3002 media_location_key: segment_url,
3003 'duration': float_or_none(segment_d, representation_ms_info['timescale']),
3004 })
3005
3006 for num, s in enumerate(representation_ms_info['s']):
3007 segment_time = s.get('t') or segment_time
3008 segment_d = s['d']
3009 add_segment_url()
3010 segment_number += 1
3011 for r in range(s.get('r', 0)):
3012 segment_time += segment_d
f0948348 3013 add_segment_url()
b4c1d6e8 3014 segment_number += 1
be2fc5b2 3015 segment_time += segment_d
3016 elif 'segment_urls' in representation_ms_info and 's' in representation_ms_info:
62b58c09
L
3017 # No media template,
3018 # e.g. https://www.youtube.com/watch?v=iXZV5uAYMJI
be2fc5b2 3019 # or any YouTube dashsegments video
3020 fragments = []
3021 segment_index = 0
3022 timescale = representation_ms_info['timescale']
3023 for s in representation_ms_info['s']:
3024 duration = float_or_none(s['d'], timescale)
3025 for r in range(s.get('r', 0) + 1):
3026 segment_uri = representation_ms_info['segment_urls'][segment_index]
3027 fragments.append({
3028 location_key(segment_uri): segment_uri,
3029 'duration': duration,
3030 })
3031 segment_index += 1
3032 representation_ms_info['fragments'] = fragments
3033 elif 'segment_urls' in representation_ms_info:
3034 # Segment URLs with no SegmentTimeline
62b58c09 3035 # E.g. https://www.seznam.cz/zpravy/clanek/cesko-zasahne-vitr-o-sile-vichrice-muze-byt-i-zivotu-nebezpecny-39091
be2fc5b2 3036 # https://github.com/ytdl-org/youtube-dl/pull/14844
3037 fragments = []
3038 segment_duration = float_or_none(
3039 representation_ms_info['segment_duration'],
3040 representation_ms_info['timescale']) if 'segment_duration' in representation_ms_info else None
3041 for segment_url in representation_ms_info['segment_urls']:
3042 fragment = {
3043 location_key(segment_url): segment_url,
3044 }
3045 if segment_duration:
3046 fragment['duration'] = segment_duration
3047 fragments.append(fragment)
3048 representation_ms_info['fragments'] = fragments
3049 # If there is a fragments key available then we correctly recognized fragmented media.
3050 # Otherwise we will assume unfragmented media with direct access. Technically, such
3051 # assumption is not necessarily correct since we may simply have no support for
3052 # some forms of fragmented media renditions yet, but for now we'll use this fallback.
3053 if 'fragments' in representation_ms_info:
3054 f.update({
3055 # NB: mpd_url may be empty when MPD manifest is parsed from a string
3056 'url': mpd_url or base_url,
3057 'fragment_base_url': base_url,
3058 'fragments': [],
3059 'protocol': 'http_dash_segments' if mime_type != 'image/jpeg' else 'mhtml',
3060 })
3061 if 'initialization_url' in representation_ms_info:
3062 initialization_url = representation_ms_info['initialization_url']
3063 if not f.get('url'):
3064 f['url'] = initialization_url
3065 f['fragments'].append({location_key(initialization_url): initialization_url})
3066 f['fragments'].extend(representation_ms_info['fragments'])
ffa89477 3067 if not period_duration:
3068 period_duration = try_get(
3069 representation_ms_info,
3070 lambda r: sum(frag['duration'] for frag in r['fragments']), float)
17b598d3 3071 else:
be2fc5b2 3072 # Assuming direct URL to unfragmented media.
3073 f['url'] = base_url
234416e4 3074 if content_type in ('video', 'audio', 'image/jpeg'):
3075 f['manifest_stream_number'] = stream_numbers[f['url']]
3076 stream_numbers[f['url']] += 1
be2fc5b2 3077 formats.append(f)
3078 elif content_type == 'text':
3079 subtitles.setdefault(lang or 'und', []).append(f)
3080
171e59ed 3081 return formats, subtitles
17b598d3 3082
fd76a142
F
3083 def _extract_ism_formats(self, *args, **kwargs):
3084 fmts, subs = self._extract_ism_formats_and_subtitles(*args, **kwargs)
3085 if subs:
b5ae35ee 3086 self._report_ignoring_subs('ISM')
fd76a142
F
3087 return fmts
3088
3089 def _extract_ism_formats_and_subtitles(self, ism_url, video_id, ism_id=None, note=None, errnote=None, fatal=True, data=None, headers={}, query={}):
47a5cb77 3090 res = self._download_xml_handle(
b2758123 3091 ism_url, video_id,
37a3bb66 3092 note='Downloading ISM manifest' if note is None else note,
3093 errnote='Failed to download ISM manifest' if errnote is None else errnote,
7360c06f 3094 fatal=fatal, data=data, headers=headers, query=query)
b2758123 3095 if res is False:
fd76a142 3096 return [], {}
47a5cb77 3097 ism_doc, urlh = res
13b08034 3098 if ism_doc is None:
fd76a142 3099 return [], {}
b2758123 3100
fd76a142 3101 return self._parse_ism_formats_and_subtitles(ism_doc, urlh.geturl(), ism_id)
b2758123 3102
fd76a142 3103 def _parse_ism_formats_and_subtitles(self, ism_doc, ism_url, ism_id=None):
76d5a363
S
3104 """
3105 Parse formats from ISM manifest.
3106 References:
3107 1. [MS-SSTR]: Smooth Streaming Protocol,
3108 https://msdn.microsoft.com/en-us/library/ff469518.aspx
3109 """
06869367 3110 if ism_doc.get('IsLive') == 'TRUE':
fd76a142 3111 return [], {}
b2758123 3112
b2758123
RA
3113 duration = int(ism_doc.attrib['Duration'])
3114 timescale = int_or_none(ism_doc.get('TimeScale')) or 10000000
3115
3116 formats = []
fd76a142 3117 subtitles = {}
b2758123
RA
3118 for stream in ism_doc.findall('StreamIndex'):
3119 stream_type = stream.get('Type')
fd76a142 3120 if stream_type not in ('video', 'audio', 'text'):
b2758123
RA
3121 continue
3122 url_pattern = stream.attrib['Url']
3123 stream_timescale = int_or_none(stream.get('TimeScale')) or timescale
3124 stream_name = stream.get('Name')
fd76a142 3125 stream_language = stream.get('Language', 'und')
b2758123 3126 for track in stream.findall('QualityLevel'):
81b6102d 3127 KNOWN_TAGS = {'255': 'AACL', '65534': 'EC-3'}
3128 fourcc = track.get('FourCC') or KNOWN_TAGS.get(track.get('AudioTag'))
b2758123 3129 # TODO: add support for WVC1 and WMAP
81b6102d 3130 if fourcc not in ('H264', 'AVC1', 'AACL', 'TTML', 'EC-3'):
b2758123
RA
3131 self.report_warning('%s is not a supported codec' % fourcc)
3132 continue
3133 tbr = int(track.attrib['Bitrate']) // 1000
76d5a363
S
3134 # [1] does not mention Width and Height attributes. However,
3135 # they're often present while MaxWidth and MaxHeight are
3136 # missing, so should be used as fallbacks
3137 width = int_or_none(track.get('MaxWidth') or track.get('Width'))
3138 height = int_or_none(track.get('MaxHeight') or track.get('Height'))
b2758123
RA
3139 sampling_rate = int_or_none(track.get('SamplingRate'))
3140
3141 track_url_pattern = re.sub(r'{[Bb]itrate}', track.attrib['Bitrate'], url_pattern)
14f25df2 3142 track_url_pattern = urllib.parse.urljoin(ism_url, track_url_pattern)
b2758123
RA
3143
3144 fragments = []
3145 fragment_ctx = {
3146 'time': 0,
3147 }
3148 stream_fragments = stream.findall('c')
3149 for stream_fragment_index, stream_fragment in enumerate(stream_fragments):
3150 fragment_ctx['time'] = int_or_none(stream_fragment.get('t')) or fragment_ctx['time']
3151 fragment_repeat = int_or_none(stream_fragment.get('r')) or 1
3152 fragment_ctx['duration'] = int_or_none(stream_fragment.get('d'))
3153 if not fragment_ctx['duration']:
3154 try:
3155 next_fragment_time = int(stream_fragment[stream_fragment_index + 1].attrib['t'])
3156 except IndexError:
3157 next_fragment_time = duration
1616f9b4 3158 fragment_ctx['duration'] = (next_fragment_time - fragment_ctx['time']) / fragment_repeat
b2758123
RA
3159 for _ in range(fragment_repeat):
3160 fragments.append({
14f25df2 3161 'url': re.sub(r'{start[ _]time}', str(fragment_ctx['time']), track_url_pattern),
b2758123
RA
3162 'duration': fragment_ctx['duration'] / stream_timescale,
3163 })
3164 fragment_ctx['time'] += fragment_ctx['duration']
3165
fd76a142
F
3166 if stream_type == 'text':
3167 subtitles.setdefault(stream_language, []).append({
3168 'ext': 'ismt',
3169 'protocol': 'ism',
3170 'url': ism_url,
3171 'manifest_url': ism_url,
3172 'fragments': fragments,
3173 '_download_params': {
3174 'stream_type': stream_type,
3175 'duration': duration,
3176 'timescale': stream_timescale,
3177 'fourcc': fourcc,
3178 'language': stream_language,
3179 'codec_private_data': track.get('CodecPrivateData'),
3180 }
3181 })
3182 elif stream_type in ('video', 'audio'):
3183 formats.append({
34921b43 3184 'format_id': join_nonempty(ism_id, stream_name, tbr),
fd76a142
F
3185 'url': ism_url,
3186 'manifest_url': ism_url,
3187 'ext': 'ismv' if stream_type == 'video' else 'isma',
3188 'width': width,
3189 'height': height,
3190 'tbr': tbr,
3191 'asr': sampling_rate,
3192 'vcodec': 'none' if stream_type == 'audio' else fourcc,
3193 'acodec': 'none' if stream_type == 'video' else fourcc,
3194 'protocol': 'ism',
3195 'fragments': fragments,
88acdbc2 3196 'has_drm': ism_doc.find('Protection') is not None,
fd76a142
F
3197 '_download_params': {
3198 'stream_type': stream_type,
3199 'duration': duration,
3200 'timescale': stream_timescale,
3201 'width': width or 0,
3202 'height': height or 0,
3203 'fourcc': fourcc,
3204 'language': stream_language,
3205 'codec_private_data': track.get('CodecPrivateData'),
3206 'sampling_rate': sampling_rate,
3207 'channels': int_or_none(track.get('Channels', 2)),
3208 'bits_per_sample': int_or_none(track.get('BitsPerSample', 16)),
3209 'nal_unit_length_field': int_or_none(track.get('NALUnitLengthField', 4)),
3210 },
3211 })
3212 return formats, subtitles
b2758123 3213
079a7cfc 3214 def _parse_html5_media_entries(self, base_url, webpage, video_id, m3u8_id=None, m3u8_entry_protocol='m3u8_native', mpd_id=None, preference=None, quality=None):
6780154e
S
3215 def absolute_url(item_url):
3216 return urljoin(base_url, item_url)
59bbe491 3217
3218 def parse_content_type(content_type):
3219 if not content_type:
3220 return {}
3221 ctr = re.search(r'(?P<mimetype>[^/]+/[^;]+)(?:;\s*codecs="?(?P<codecs>[^"]+))?', content_type)
3222 if ctr:
3223 mimetype, codecs = ctr.groups()
3224 f = parse_codecs(codecs)
3225 f['ext'] = mimetype2ext(mimetype)
3226 return f
3227 return {}
3228
222a2308
L
3229 def _media_formats(src, cur_media_type, type_info=None):
3230 type_info = type_info or {}
520251c0 3231 full_url = absolute_url(src)
82889d4a 3232 ext = type_info.get('ext') or determine_ext(full_url)
87a449c1 3233 if ext == 'm3u8':
520251c0
YCH
3234 is_plain_url = False
3235 formats = self._extract_m3u8_formats(
ad120ae1 3236 full_url, video_id, ext='mp4',
eeb0a956 3237 entry_protocol=m3u8_entry_protocol, m3u8_id=m3u8_id,
f983b875 3238 preference=preference, quality=quality, fatal=False)
87a449c1
S
3239 elif ext == 'mpd':
3240 is_plain_url = False
3241 formats = self._extract_mpd_formats(
b359e977 3242 full_url, video_id, mpd_id=mpd_id, fatal=False)
520251c0
YCH
3243 else:
3244 is_plain_url = True
3245 formats = [{
3246 'url': full_url,
3247 'vcodec': 'none' if cur_media_type == 'audio' else None,
222a2308 3248 'ext': ext,
520251c0
YCH
3249 }]
3250 return is_plain_url, formats
3251
59bbe491 3252 entries = []
4328ddf8 3253 # amp-video and amp-audio are very similar to their HTML5 counterparts
962ffcf8 3254 # so we will include them right here (see
4328ddf8 3255 # https://www.ampproject.org/docs/reference/components/amp-video)
29f7c58a 3256 # For dl8-* tags see https://delight-vr.com/documentation/dl8-video/
3257 _MEDIA_TAG_NAME_RE = r'(?:(?:amp|dl8(?:-live)?)-)?(video|audio)'
3258 media_tags = [(media_tag, media_tag_name, media_type, '')
3259 for media_tag, media_tag_name, media_type
3260 in re.findall(r'(?s)(<(%s)[^>]*/>)' % _MEDIA_TAG_NAME_RE, webpage)]
2aec7256
S
3261 media_tags.extend(re.findall(
3262 # We only allow video|audio followed by a whitespace or '>'.
3263 # Allowing more characters may end up in significant slow down (see
62b58c09
L
3264 # https://github.com/ytdl-org/youtube-dl/issues/11979,
3265 # e.g. http://www.porntrex.com/maps/videositemap.xml).
29f7c58a 3266 r'(?s)(<(?P<tag>%s)(?:\s+[^>]*)?>)(.*?)</(?P=tag)>' % _MEDIA_TAG_NAME_RE, webpage))
3267 for media_tag, _, media_type, media_content in media_tags:
59bbe491 3268 media_info = {
3269 'formats': [],
3270 'subtitles': {},
3271 }
3272 media_attributes = extract_attributes(media_tag)
bfbecd11 3273 src = strip_or_none(dict_get(media_attributes, ('src', 'data-video-src', 'data-src', 'data-source')))
59bbe491 3274 if src:
222a2308
L
3275 f = parse_content_type(media_attributes.get('type'))
3276 _, formats = _media_formats(src, media_type, f)
520251c0 3277 media_info['formats'].extend(formats)
6780154e 3278 media_info['thumbnail'] = absolute_url(media_attributes.get('poster'))
59bbe491 3279 if media_content:
3280 for source_tag in re.findall(r'<source[^>]+>', media_content):
d493f15c
S
3281 s_attr = extract_attributes(source_tag)
3282 # data-video-src and data-src are non standard but seen
3283 # several times in the wild
bfbecd11 3284 src = strip_or_none(dict_get(s_attr, ('src', 'data-video-src', 'data-src', 'data-source')))
59bbe491 3285 if not src:
3286 continue
d493f15c 3287 f = parse_content_type(s_attr.get('type'))
868f79db 3288 is_plain_url, formats = _media_formats(src, media_type, f)
520251c0 3289 if is_plain_url:
d493f15c
S
3290 # width, height, res, label and title attributes are
3291 # all not standard but seen several times in the wild
3292 labels = [
3293 s_attr.get(lbl)
3294 for lbl in ('label', 'title')
3295 if str_or_none(s_attr.get(lbl))
3296 ]
3297 width = int_or_none(s_attr.get('width'))
3089bc74
S
3298 height = (int_or_none(s_attr.get('height'))
3299 or int_or_none(s_attr.get('res')))
d493f15c
S
3300 if not width or not height:
3301 for lbl in labels:
3302 resolution = parse_resolution(lbl)
3303 if not resolution:
3304 continue
3305 width = width or resolution.get('width')
3306 height = height or resolution.get('height')
3307 for lbl in labels:
3308 tbr = parse_bitrate(lbl)
3309 if tbr:
3310 break
3311 else:
3312 tbr = None
1ed45499 3313 f.update({
d493f15c
S
3314 'width': width,
3315 'height': height,
3316 'tbr': tbr,
3317 'format_id': s_attr.get('label') or s_attr.get('title'),
1ed45499 3318 })
520251c0
YCH
3319 f.update(formats[0])
3320 media_info['formats'].append(f)
3321 else:
3322 media_info['formats'].extend(formats)
59bbe491 3323 for track_tag in re.findall(r'<track[^>]+>', media_content):
3324 track_attributes = extract_attributes(track_tag)
3325 kind = track_attributes.get('kind')
5968d7d2 3326 if not kind or kind in ('subtitles', 'captions'):
f856816b 3327 src = strip_or_none(track_attributes.get('src'))
59bbe491 3328 if not src:
3329 continue
3330 lang = track_attributes.get('srclang') or track_attributes.get('lang') or track_attributes.get('label')
3331 media_info['subtitles'].setdefault(lang, []).append({
3332 'url': absolute_url(src),
3333 })
5e8e2fa5
S
3334 for f in media_info['formats']:
3335 f.setdefault('http_headers', {})['Referer'] = base_url
5968d7d2 3336 if media_info['formats'] or media_info['subtitles']:
59bbe491 3337 entries.append(media_info)
3338 return entries
3339
f6a1d69a
F
3340 def _extract_akamai_formats(self, *args, **kwargs):
3341 fmts, subs = self._extract_akamai_formats_and_subtitles(*args, **kwargs)
3342 if subs:
b5ae35ee 3343 self._report_ignoring_subs('akamai')
f6a1d69a
F
3344 return fmts
3345
3346 def _extract_akamai_formats_and_subtitles(self, manifest_url, video_id, hosts={}):
29f7c58a 3347 signed = 'hdnea=' in manifest_url
3348 if not signed:
3349 # https://learn.akamai.com/en-us/webhelp/media-services-on-demand/stream-packaging-user-guide/GUID-BE6C0F73-1E06-483B-B0EA-57984B91B7F9.html
3350 manifest_url = re.sub(
3351 r'(?:b=[\d,-]+|(?:__a__|attributes)=off|__b__=\d+)&?',
3352 '', manifest_url).strip('?')
3353
c7c43a93 3354 formats = []
f6a1d69a 3355 subtitles = {}
70c5802b 3356
e71a4509 3357 hdcore_sign = 'hdcore=3.7.0'
ff6f9a67 3358 f4m_url = re.sub(r'(https?://[^/]+)/i/', r'\1/z/', manifest_url).replace('/master.m3u8', '/manifest.f4m')
c4251b9a
RA
3359 hds_host = hosts.get('hds')
3360 if hds_host:
3361 f4m_url = re.sub(r'(https?://)[^/]+', r'\1' + hds_host, f4m_url)
e71a4509
RA
3362 if 'hdcore=' not in f4m_url:
3363 f4m_url += ('&' if '?' in f4m_url else '?') + hdcore_sign
3364 f4m_formats = self._extract_f4m_formats(
3365 f4m_url, video_id, f4m_id='hds', fatal=False)
3366 for entry in f4m_formats:
3367 entry.update({'extra_param_to_segment_url': hdcore_sign})
3368 formats.extend(f4m_formats)
70c5802b 3369
c4251b9a
RA
3370 m3u8_url = re.sub(r'(https?://[^/]+)/z/', r'\1/i/', manifest_url).replace('/manifest.f4m', '/master.m3u8')
3371 hls_host = hosts.get('hls')
3372 if hls_host:
3373 m3u8_url = re.sub(r'(https?://)[^/]+', r'\1' + hls_host, m3u8_url)
f6a1d69a 3374 m3u8_formats, m3u8_subtitles = self._extract_m3u8_formats_and_subtitles(
c7c43a93 3375 m3u8_url, video_id, 'mp4', 'm3u8_native',
29f7c58a 3376 m3u8_id='hls', fatal=False)
3377 formats.extend(m3u8_formats)
f6a1d69a 3378 subtitles = self._merge_subtitles(subtitles, m3u8_subtitles)
70c5802b 3379
3380 http_host = hosts.get('http')
29f7c58a 3381 if http_host and m3u8_formats and not signed:
3382 REPL_REGEX = r'https?://[^/]+/i/([^,]+),([^/]+),([^/]+)\.csmil/.+'
70c5802b 3383 qualities = re.match(REPL_REGEX, m3u8_url).group(2).split(',')
3384 qualities_length = len(qualities)
29f7c58a 3385 if len(m3u8_formats) in (qualities_length, qualities_length + 1):
70c5802b 3386 i = 0
29f7c58a 3387 for f in m3u8_formats:
3388 if f['vcodec'] != 'none':
70c5802b 3389 for protocol in ('http', 'https'):
3390 http_f = f.copy()
3391 del http_f['manifest_url']
3392 http_url = re.sub(
86e5f3ed 3393 REPL_REGEX, protocol + fr'://{http_host}/\g<1>{qualities[i]}\3', f['url'])
70c5802b 3394 http_f.update({
3395 'format_id': http_f['format_id'].replace('hls-', protocol + '-'),
3396 'url': http_url,
3397 'protocol': protocol,
3398 })
29f7c58a 3399 formats.append(http_f)
70c5802b 3400 i += 1
70c5802b 3401
f6a1d69a 3402 return formats, subtitles
c7c43a93 3403
6ad02195 3404 def _extract_wowza_formats(self, url, video_id, m3u8_entry_protocol='m3u8_native', skip_protocols=[]):
14f25df2 3405 query = urllib.parse.urlparse(url).query
6ad02195 3406 url = re.sub(r'/(?:manifest|playlist|jwplayer)\.(?:m3u8|f4m|mpd|smil)', '', url)
240f2622
S
3407 mobj = re.search(
3408 r'(?:(?:http|rtmp|rtsp)(?P<s>s)?:)?(?P<url>//[^?]+)', url)
3409 url_base = mobj.group('url')
3410 http_base_url = '%s%s:%s' % ('http', mobj.group('s') or '', url_base)
6ad02195 3411 formats = []
044eeb14
S
3412
3413 def manifest_url(manifest):
86e5f3ed 3414 m_url = f'{http_base_url}/{manifest}'
044eeb14
S
3415 if query:
3416 m_url += '?%s' % query
3417 return m_url
3418
6ad02195
RA
3419 if 'm3u8' not in skip_protocols:
3420 formats.extend(self._extract_m3u8_formats(
044eeb14 3421 manifest_url('playlist.m3u8'), video_id, 'mp4',
6ad02195
RA
3422 m3u8_entry_protocol, m3u8_id='hls', fatal=False))
3423 if 'f4m' not in skip_protocols:
3424 formats.extend(self._extract_f4m_formats(
044eeb14 3425 manifest_url('manifest.f4m'),
6ad02195 3426 video_id, f4m_id='hds', fatal=False))
0384932e
RA
3427 if 'dash' not in skip_protocols:
3428 formats.extend(self._extract_mpd_formats(
044eeb14 3429 manifest_url('manifest.mpd'),
0384932e 3430 video_id, mpd_id='dash', fatal=False))
6ad02195 3431 if re.search(r'(?:/smil:|\.smil)', url_base):
6ad02195
RA
3432 if 'smil' not in skip_protocols:
3433 rtmp_formats = self._extract_smil_formats(
044eeb14 3434 manifest_url('jwplayer.smil'),
6ad02195
RA
3435 video_id, fatal=False)
3436 for rtmp_format in rtmp_formats:
3437 rtsp_format = rtmp_format.copy()
3438 rtsp_format['url'] = '%s/%s' % (rtmp_format['url'], rtmp_format['play_path'])
3439 del rtsp_format['play_path']
3440 del rtsp_format['ext']
3441 rtsp_format.update({
3442 'url': rtsp_format['url'].replace('rtmp://', 'rtsp://'),
3443 'format_id': rtmp_format['format_id'].replace('rtmp', 'rtsp'),
3444 'protocol': 'rtsp',
3445 })
3446 formats.extend([rtmp_format, rtsp_format])
3447 else:
3448 for protocol in ('rtmp', 'rtsp'):
3449 if protocol not in skip_protocols:
3450 formats.append({
86e5f3ed 3451 'url': f'{protocol}:{url_base}',
6ad02195
RA
3452 'format_id': protocol,
3453 'protocol': protocol,
3454 })
3455 return formats
3456
c73e330e 3457 def _find_jwplayer_data(self, webpage, video_id=None, transform_source=js_to_json):
a4a554a7 3458 mobj = re.search(
ac9c69ac 3459 r'(?s)jwplayer\((?P<quote>[\'"])[^\'" ]+(?P=quote)\)(?!</script>).*?\.setup\s*\((?P<options>[^)]+)\)',
a4a554a7
YCH
3460 webpage)
3461 if mobj:
c73e330e
RU
3462 try:
3463 jwplayer_data = self._parse_json(mobj.group('options'),
3464 video_id=video_id,
3465 transform_source=transform_source)
3466 except ExtractorError:
3467 pass
3468 else:
3469 if isinstance(jwplayer_data, dict):
3470 return jwplayer_data
a4a554a7
YCH
3471
3472 def _extract_jwplayer_data(self, webpage, video_id, *args, **kwargs):
c73e330e
RU
3473 jwplayer_data = self._find_jwplayer_data(
3474 webpage, video_id, transform_source=js_to_json)
a4a554a7
YCH
3475 return self._parse_jwplayer_data(
3476 jwplayer_data, video_id, *args, **kwargs)
3477
3478 def _parse_jwplayer_data(self, jwplayer_data, video_id=None, require_title=True,
3479 m3u8_id=None, mpd_id=None, rtmp_params=None, base_url=None):
3480 # JWPlayer backward compatibility: flattened playlists
3481 # https://github.com/jwplayer/jwplayer/blob/v7.4.3/src/js/api/config.js#L81-L96
3482 if 'playlist' not in jwplayer_data:
3483 jwplayer_data = {'playlist': [jwplayer_data]}
3484
3485 entries = []
3486
3487 # JWPlayer backward compatibility: single playlist item
3488 # https://github.com/jwplayer/jwplayer/blob/v7.7.0/src/js/playlist/playlist.js#L10
3489 if not isinstance(jwplayer_data['playlist'], list):
3490 jwplayer_data['playlist'] = [jwplayer_data['playlist']]
3491
3492 for video_data in jwplayer_data['playlist']:
3493 # JWPlayer backward compatibility: flattened sources
3494 # https://github.com/jwplayer/jwplayer/blob/v7.4.3/src/js/playlist/item.js#L29-L35
3495 if 'sources' not in video_data:
3496 video_data['sources'] = [video_data]
3497
3498 this_video_id = video_id or video_data['mediaid']
3499
1a2192cb
S
3500 formats = self._parse_jwplayer_formats(
3501 video_data['sources'], video_id=this_video_id, m3u8_id=m3u8_id,
3502 mpd_id=mpd_id, rtmp_params=rtmp_params, base_url=base_url)
a4a554a7
YCH
3503
3504 subtitles = {}
3505 tracks = video_data.get('tracks')
3506 if tracks and isinstance(tracks, list):
3507 for track in tracks:
96a2daa1
S
3508 if not isinstance(track, dict):
3509 continue
f4b74272 3510 track_kind = track.get('kind')
14f25df2 3511 if not track_kind or not isinstance(track_kind, str):
f4b74272
S
3512 continue
3513 if track_kind.lower() not in ('captions', 'subtitles'):
a4a554a7
YCH
3514 continue
3515 track_url = urljoin(base_url, track.get('file'))
3516 if not track_url:
3517 continue
3518 subtitles.setdefault(track.get('label') or 'en', []).append({
3519 'url': self._proto_relative_url(track_url)
3520 })
3521
50d808f5 3522 entry = {
a4a554a7 3523 'id': this_video_id,
50d808f5 3524 'title': unescapeHTML(video_data['title'] if require_title else video_data.get('title')),
f81dd65b 3525 'description': clean_html(video_data.get('description')),
6945b9e7 3526 'thumbnail': urljoin(base_url, self._proto_relative_url(video_data.get('image'))),
a4a554a7
YCH
3527 'timestamp': int_or_none(video_data.get('pubdate')),
3528 'duration': float_or_none(jwplayer_data.get('duration') or video_data.get('duration')),
3529 'subtitles': subtitles,
50d808f5
RA
3530 }
3531 # https://github.com/jwplayer/jwplayer/blob/master/src/js/utils/validator.js#L32
3532 if len(formats) == 1 and re.search(r'^(?:http|//).*(?:youtube\.com|youtu\.be)/.+', formats[0]['url']):
3533 entry.update({
3534 '_type': 'url_transparent',
3535 'url': formats[0]['url'],
3536 })
3537 else:
3538 self._sort_formats(formats)
3539 entry['formats'] = formats
3540 entries.append(entry)
a4a554a7
YCH
3541 if len(entries) == 1:
3542 return entries[0]
3543 else:
3544 return self.playlist_result(entries)
3545
ed0cf9b3
S
3546 def _parse_jwplayer_formats(self, jwplayer_sources_data, video_id=None,
3547 m3u8_id=None, mpd_id=None, rtmp_params=None, base_url=None):
bf1b87cd 3548 urls = []
ed0cf9b3 3549 formats = []
1a2192cb 3550 for source in jwplayer_sources_data:
0a268c6e
S
3551 if not isinstance(source, dict):
3552 continue
6945b9e7
RA
3553 source_url = urljoin(
3554 base_url, self._proto_relative_url(source.get('file')))
3555 if not source_url or source_url in urls:
bf1b87cd
RA
3556 continue
3557 urls.append(source_url)
ed0cf9b3
S
3558 source_type = source.get('type') or ''
3559 ext = mimetype2ext(source_type) or determine_ext(source_url)
3560 if source_type == 'hls' or ext == 'm3u8':
3561 formats.extend(self._extract_m3u8_formats(
0236cd0d
S
3562 source_url, video_id, 'mp4', entry_protocol='m3u8_native',
3563 m3u8_id=m3u8_id, fatal=False))
0d9c48de 3564 elif source_type == 'dash' or ext == 'mpd':
ed0cf9b3
S
3565 formats.extend(self._extract_mpd_formats(
3566 source_url, video_id, mpd_id=mpd_id, fatal=False))
b51dc9db
S
3567 elif ext == 'smil':
3568 formats.extend(self._extract_smil_formats(
3569 source_url, video_id, fatal=False))
ed0cf9b3 3570 # https://github.com/jwplayer/jwplayer/blob/master/src/js/providers/default.js#L67
0236cd0d
S
3571 elif source_type.startswith('audio') or ext in (
3572 'oga', 'aac', 'mp3', 'mpeg', 'vorbis'):
ed0cf9b3
S
3573 formats.append({
3574 'url': source_url,
3575 'vcodec': 'none',
3576 'ext': ext,
3577 })
3578 else:
3579 height = int_or_none(source.get('height'))
3580 if height is None:
3581 # Often no height is provided but there is a label in
0236cd0d 3582 # format like "1080p", "720p SD", or 1080.
ed0cf9b3 3583 height = int_or_none(self._search_regex(
14f25df2 3584 r'^(\d{3,4})[pP]?(?:\b|$)', str(source.get('label') or ''),
ed0cf9b3
S
3585 'height', default=None))
3586 a_format = {
3587 'url': source_url,
3588 'width': int_or_none(source.get('width')),
3589 'height': height,
d3a3d7f0 3590 'tbr': int_or_none(source.get('bitrate'), scale=1000),
3591 'filesize': int_or_none(source.get('filesize')),
ed0cf9b3
S
3592 'ext': ext,
3593 }
3594 if source_url.startswith('rtmp'):
3595 a_format['ext'] = 'flv'
ed0cf9b3
S
3596 # See com/longtailvideo/jwplayer/media/RTMPMediaProvider.as
3597 # of jwplayer.flash.swf
3598 rtmp_url_parts = re.split(
3599 r'((?:mp4|mp3|flv):)', source_url, 1)
3600 if len(rtmp_url_parts) == 3:
3601 rtmp_url, prefix, play_path = rtmp_url_parts
3602 a_format.update({
3603 'url': rtmp_url,
3604 'play_path': prefix + play_path,
3605 })
3606 if rtmp_params:
3607 a_format.update(rtmp_params)
3608 formats.append(a_format)
3609 return formats
3610
f4b1c7ad 3611 def _live_title(self, name):
39ca3b5c 3612 self._downloader.deprecation_warning('yt_dlp.InfoExtractor._live_title is deprecated and does not work as expected')
3613 return name
f4b1c7ad 3614
b14f3a4c
PH
3615 def _int(self, v, name, fatal=False, **kwargs):
3616 res = int_or_none(v, **kwargs)
b14f3a4c 3617 if res is None:
86e5f3ed 3618 msg = f'Failed to extract {name}: Could not parse value {v!r}'
b14f3a4c
PH
3619 if fatal:
3620 raise ExtractorError(msg)
3621 else:
6a39ee13 3622 self.report_warning(msg)
b14f3a4c
PH
3623 return res
3624
3625 def _float(self, v, name, fatal=False, **kwargs):
3626 res = float_or_none(v, **kwargs)
3627 if res is None:
86e5f3ed 3628 msg = f'Failed to extract {name}: Could not parse value {v!r}'
b14f3a4c
PH
3629 if fatal:
3630 raise ExtractorError(msg)
3631 else:
6a39ee13 3632 self.report_warning(msg)
b14f3a4c
PH
3633 return res
3634
40e41780
TF
3635 def _set_cookie(self, domain, name, value, expire_time=None, port=None,
3636 path='/', secure=False, discard=False, rest={}, **kwargs):
ac668111 3637 cookie = http.cookiejar.Cookie(
4ed2d7b7 3638 0, name, value, port, port is not None, domain, True,
40e41780
TF
3639 domain.startswith('.'), path, True, secure, expire_time,
3640 discard, None, None, rest)
9809740b 3641 self.cookiejar.set_cookie(cookie)
42939b61 3642
799207e8 3643 def _get_cookies(self, url):
ac668111 3644 """ Return a http.cookies.SimpleCookie with the cookies for the url """
8817a80d 3645 return LenientSimpleCookie(self._downloader._calc_cookies(url))
799207e8 3646
e3c1266f 3647 def _apply_first_set_cookie_header(self, url_handle, cookie):
ce2fe4c0
S
3648 """
3649 Apply first Set-Cookie header instead of the last. Experimental.
3650
3651 Some sites (e.g. [1-3]) may serve two cookies under the same name
3652 in Set-Cookie header and expect the first (old) one to be set rather
3653 than second (new). However, as of RFC6265 the newer one cookie
3654 should be set into cookie store what actually happens.
3655 We will workaround this issue by resetting the cookie to
3656 the first one manually.
3657 1. https://new.vk.com/
3658 2. https://github.com/ytdl-org/youtube-dl/issues/9841#issuecomment-227871201
3659 3. https://learning.oreilly.com/
3660 """
e3c1266f
S
3661 for header, cookies in url_handle.headers.items():
3662 if header.lower() != 'set-cookie':
3663 continue
cfb0511d 3664 cookies = cookies.encode('iso-8859-1').decode('utf-8')
e3c1266f
S
3665 cookie_value = re.search(
3666 r'%s=(.+?);.*?\b[Dd]omain=(.+?)(?:[,;]|$)' % cookie, cookies)
3667 if cookie_value:
3668 value, domain = cookie_value.groups()
3669 self._set_cookie(domain, cookie, value)
3670 break
3671
82d02080 3672 @classmethod
3673 def get_testcases(cls, include_onlymatching=False):
3674 t = getattr(cls, '_TEST', None)
05900629 3675 if t:
82d02080 3676 assert not hasattr(cls, '_TESTS'), f'{cls.ie_key()}IE has _TEST and _TESTS'
05900629
PH
3677 tests = [t]
3678 else:
82d02080 3679 tests = getattr(cls, '_TESTS', [])
05900629
PH
3680 for t in tests:
3681 if not include_onlymatching and t.get('only_matching', False):
3682 continue
82d02080 3683 t['name'] = cls.ie_key()
05900629
PH
3684 yield t
3685
f2e8dbcc 3686 @classmethod
3687 def get_webpage_testcases(cls):
3688 tests = getattr(cls, '_WEBPAGE_TESTS', [])
3689 for t in tests:
3690 t['name'] = cls.ie_key()
3691 return tests
3692
24146491 3693 @classproperty
3694 def age_limit(cls):
3695 """Get age limit from the testcases"""
3696 return max(traverse_obj(
f2e8dbcc 3697 (*cls.get_testcases(include_onlymatching=False), *cls.get_webpage_testcases()),
24146491 3698 (..., (('playlist', 0), None), 'info_dict', 'age_limit')) or [0])
3699
82d02080 3700 @classmethod
3701 def is_suitable(cls, age_limit):
24146491 3702 """Test whether the extractor is generally suitable for the given age limit"""
3703 return not age_restricted(cls.age_limit, age_limit)
05900629 3704
82d02080 3705 @classmethod
3706 def description(cls, *, markdown=True, search_examples=None):
8dcce6a8 3707 """Description of the extractor"""
3708 desc = ''
82d02080 3709 if cls._NETRC_MACHINE:
8dcce6a8 3710 if markdown:
82d02080 3711 desc += f' [<abbr title="netrc machine"><em>{cls._NETRC_MACHINE}</em></abbr>]'
8dcce6a8 3712 else:
82d02080 3713 desc += f' [{cls._NETRC_MACHINE}]'
3714 if cls.IE_DESC is False:
8dcce6a8 3715 desc += ' [HIDDEN]'
82d02080 3716 elif cls.IE_DESC:
3717 desc += f' {cls.IE_DESC}'
3718 if cls.SEARCH_KEY:
3719 desc += f'; "{cls.SEARCH_KEY}:" prefix'
8dcce6a8 3720 if search_examples:
3721 _COUNTS = ('', '5', '10', 'all')
62b58c09 3722 desc += f' (e.g. "{cls.SEARCH_KEY}{random.choice(_COUNTS)}:{random.choice(search_examples)}")'
82d02080 3723 if not cls.working():
8dcce6a8 3724 desc += ' (**Currently broken**)' if markdown else ' (Currently broken)'
3725
82d02080 3726 name = f' - **{cls.IE_NAME}**' if markdown else cls.IE_NAME
8dcce6a8 3727 return f'{name}:{desc}' if desc else name
3728
a504ced0 3729 def extract_subtitles(self, *args, **kwargs):
a06916d9 3730 if (self.get_param('writesubtitles', False)
3731 or self.get_param('listsubtitles')):
9868ea49
JMF
3732 return self._get_subtitles(*args, **kwargs)
3733 return {}
a504ced0
JMF
3734
3735 def _get_subtitles(self, *args, **kwargs):
611c1dd9 3736 raise NotImplementedError('This method must be implemented by subclasses')
a504ced0 3737
a2160aa4 3738 def extract_comments(self, *args, **kwargs):
3739 if not self.get_param('getcomments'):
3740 return None
3741 generator = self._get_comments(*args, **kwargs)
3742
3743 def extractor():
3744 comments = []
d2b2fca5 3745 interrupted = True
a2160aa4 3746 try:
3747 while True:
3748 comments.append(next(generator))
a2160aa4 3749 except StopIteration:
3750 interrupted = False
d2b2fca5 3751 except KeyboardInterrupt:
3752 self.to_screen('Interrupted by user')
3753 except Exception as e:
3754 if self.get_param('ignoreerrors') is not True:
3755 raise
3756 self._downloader.report_error(e)
a2160aa4 3757 comment_count = len(comments)
3758 self.to_screen(f'Extracted {comment_count} comments')
3759 return {
3760 'comments': comments,
3761 'comment_count': None if interrupted else comment_count
3762 }
3763 return extractor
3764
3765 def _get_comments(self, *args, **kwargs):
3766 raise NotImplementedError('This method must be implemented by subclasses')
3767
912e0b7e
YCH
3768 @staticmethod
3769 def _merge_subtitle_items(subtitle_list1, subtitle_list2):
a825ffbf 3770 """ Merge subtitle items for one language. Items with duplicated URLs/data
912e0b7e 3771 will be dropped. """
86e5f3ed 3772 list1_data = {(item.get('url'), item.get('data')) for item in subtitle_list1}
912e0b7e 3773 ret = list(subtitle_list1)
a44ca5a4 3774 ret.extend(item for item in subtitle_list2 if (item.get('url'), item.get('data')) not in list1_data)
912e0b7e
YCH
3775 return ret
3776
3777 @classmethod
46890374 3778 def _merge_subtitles(cls, *dicts, target=None):
19bb3920 3779 """ Merge subtitle dictionaries, language by language. """
19bb3920
F
3780 if target is None:
3781 target = {}
3782 for d in dicts:
3783 for lang, subs in d.items():
3784 target[lang] = cls._merge_subtitle_items(target.get(lang, []), subs)
3785 return target
912e0b7e 3786
360e1ca5 3787 def extract_automatic_captions(self, *args, **kwargs):
a06916d9 3788 if (self.get_param('writeautomaticsub', False)
3789 or self.get_param('listsubtitles')):
9868ea49
JMF
3790 return self._get_automatic_captions(*args, **kwargs)
3791 return {}
360e1ca5
JMF
3792
3793 def _get_automatic_captions(self, *args, **kwargs):
611c1dd9 3794 raise NotImplementedError('This method must be implemented by subclasses')
360e1ca5 3795
2762dbb1 3796 @functools.cached_property
24146491 3797 def _cookies_passed(self):
3798 """Whether cookies have been passed to YoutubeDL"""
3799 return self.get_param('cookiefile') is not None or self.get_param('cookiesfrombrowser') is not None
3800
d77ab8e2 3801 def mark_watched(self, *args, **kwargs):
1813a6cc 3802 if not self.get_param('mark_watched', False):
3803 return
24146491 3804 if self.supports_login() and self._get_login_info()[0] is not None or self._cookies_passed:
d77ab8e2
S
3805 self._mark_watched(*args, **kwargs)
3806
3807 def _mark_watched(self, *args, **kwargs):
3808 raise NotImplementedError('This method must be implemented by subclasses')
3809
38cce791
YCH
3810 def geo_verification_headers(self):
3811 headers = {}
a06916d9 3812 geo_verification_proxy = self.get_param('geo_verification_proxy')
38cce791
YCH
3813 if geo_verification_proxy:
3814 headers['Ytdl-request-proxy'] = geo_verification_proxy
3815 return headers
3816
8f97a15d 3817 @staticmethod
3818 def _generic_id(url):
14f25df2 3819 return urllib.parse.unquote(os.path.splitext(url.rstrip('/').split('/')[-1])[0])
98763ee3 3820
8f97a15d 3821 @staticmethod
3822 def _generic_title(url):
14f25df2 3823 return urllib.parse.unquote(os.path.splitext(url_basename(url))[0])
98763ee3 3824
c224251a 3825 @staticmethod
b0089e89 3826 def _availability(is_private=None, needs_premium=None, needs_subscription=None, needs_auth=None, is_unlisted=None):
c224251a
M
3827 all_known = all(map(
3828 lambda x: x is not None,
3829 (is_private, needs_premium, needs_subscription, needs_auth, is_unlisted)))
3830 return (
3831 'private' if is_private
3832 else 'premium_only' if needs_premium
3833 else 'subscriber_only' if needs_subscription
3834 else 'needs_auth' if needs_auth
3835 else 'unlisted' if is_unlisted
3836 else 'public' if all_known
3837 else None)
3838
d43de682 3839 def _configuration_arg(self, key, default=NO_DEFAULT, *, ie_key=None, casesense=False):
4bb6b02f 3840 '''
3841 @returns A list of values for the extractor argument given by "key"
3842 or "default" if no such key is present
3843 @param default The default value to return when the key is not present (default: [])
3844 @param casesense When false, the values are converted to lower case
3845 '''
5225df50 3846 ie_key = ie_key if isinstance(ie_key, str) else (ie_key or self).ie_key()
3847 val = traverse_obj(self._downloader.params, ('extractor_args', ie_key.lower(), key))
4bb6b02f 3848 if val is None:
3849 return [] if default is NO_DEFAULT else default
3850 return list(val) if casesense else [x.lower() for x in val]
5d3a0e79 3851
f40ee5e9 3852 def _yes_playlist(self, playlist_id, video_id, smuggled_data=None, *, playlist_label='playlist', video_label='video'):
3853 if not playlist_id or not video_id:
3854 return not video_id
3855
3856 no_playlist = (smuggled_data or {}).get('force_noplaylist')
3857 if no_playlist is not None:
3858 return not no_playlist
3859
3860 video_id = '' if video_id is True else f' {video_id}'
3861 playlist_id = '' if playlist_id is True else f' {playlist_id}'
3862 if self.get_param('noplaylist'):
3863 self.to_screen(f'Downloading just the {video_label}{video_id} because of --no-playlist')
3864 return False
3865 self.to_screen(f'Downloading {playlist_label}{playlist_id} - add --no-playlist to download just the {video_label}{video_id}')
3866 return True
3867
be5c1ae8 3868 def _error_or_warning(self, err, _count=None, _retries=0, *, fatal=True):
8ca48a1a 3869 RetryManager.report_retry(
3870 err, _count or int(fatal), _retries,
3871 info=self.to_screen, warn=self.report_warning, error=None if fatal else self.report_warning,
3872 sleep_func=self.get_param('retry_sleep_functions', {}).get('extractor'))
be5c1ae8 3873
3874 def RetryManager(self, **kwargs):
3875 return RetryManager(self.get_param('extractor_retries', 3), self._error_or_warning, **kwargs)
3876
ade1fa70 3877 def _extract_generic_embeds(self, url, *args, info_dict={}, note='Extracting generic embeds', **kwargs):
3878 display_id = traverse_obj(info_dict, 'display_id', 'id')
3879 self.to_screen(f'{format_field(display_id, None, "%s: ")}{note}')
3880 return self._downloader.get_info_extractor('Generic')._extract_embeds(
3881 smuggle_url(url, {'block_ies': [self.ie_key()]}), *args, **kwargs)
3882
8f97a15d 3883 @classmethod
3884 def extract_from_webpage(cls, ydl, url, webpage):
3885 ie = (cls if isinstance(cls._extract_from_webpage, types.MethodType)
3886 else ydl.get_info_extractor(cls.ie_key()))
f2e8dbcc 3887 for info in ie._extract_from_webpage(url, webpage) or []:
3888 # url = None since we do not want to set (webpage/original)_url
3889 ydl.add_default_extra_info(info, ie, None)
3890 yield info
8f97a15d 3891
3892 @classmethod
3893 def _extract_from_webpage(cls, url, webpage):
3894 for embed_url in orderedSet(
3895 cls._extract_embed_urls(url, webpage) or [], lazy=True):
d2c8aadf 3896 yield cls.url_result(embed_url, None if cls._VALID_URL is False else cls)
8f97a15d 3897
3898 @classmethod
3899 def _extract_embed_urls(cls, url, webpage):
3900 """@returns all the embed urls on the webpage"""
3901 if '_EMBED_URL_RE' not in cls.__dict__:
3902 assert isinstance(cls._EMBED_REGEX, (list, tuple))
3903 for idx, regex in enumerate(cls._EMBED_REGEX):
3904 assert regex.count('(?P<url>') == 1, \
3905 f'{cls.__name__}._EMBED_REGEX[{idx}] must have exactly 1 url group\n\t{regex}'
3906 cls._EMBED_URL_RE = tuple(map(re.compile, cls._EMBED_REGEX))
3907
3908 for regex in cls._EMBED_URL_RE:
3909 for mobj in regex.finditer(webpage):
3910 embed_url = urllib.parse.urljoin(url, unescapeHTML(mobj.group('url')))
3911 if cls._VALID_URL is False or cls.suitable(embed_url):
3912 yield embed_url
3913
3914 class StopExtraction(Exception):
3915 pass
3916
bfd973ec 3917 @classmethod
3918 def _extract_url(cls, webpage): # TODO: Remove
3919 """Only for compatibility with some older extractors"""
3920 return next(iter(cls._extract_embed_urls(None, webpage) or []), None)
3921
2314b4d8 3922 @classmethod
3923 def __init_subclass__(cls, *, plugin_name=None, **kwargs):
3924 if plugin_name:
3925 mro = inspect.getmro(cls)
3926 super_class = cls.__wrapped__ = mro[mro.index(cls) + 1]
3927 cls.IE_NAME, cls.ie_key = f'{super_class.IE_NAME}+{plugin_name}', super_class.ie_key
3928 while getattr(super_class, '__wrapped__', None):
3929 super_class = super_class.__wrapped__
3930 setattr(sys.modules[super_class.__module__], super_class.__name__, cls)
3931
3932 return super().__init_subclass__(**kwargs)
3933
8dbe9899 3934
d6983cb4
PH
3935class SearchInfoExtractor(InfoExtractor):
3936 """
3937 Base class for paged search queries extractors.
10952eb2 3938 They accept URLs in the format _SEARCH_KEY(|all|[0-9]):{query}
96565c7e 3939 Instances should define _SEARCH_KEY and optionally _MAX_RESULTS
d6983cb4
PH
3940 """
3941
96565c7e 3942 _MAX_RESULTS = float('inf')
3943
8f97a15d 3944 @classproperty
3945 def _VALID_URL(cls):
d6983cb4
PH
3946 return r'%s(?P<prefix>|[1-9][0-9]*|all):(?P<query>[\s\S]+)' % cls._SEARCH_KEY
3947
d6983cb4 3948 def _real_extract(self, query):
2c4aaadd 3949 prefix, query = self._match_valid_url(query).group('prefix', 'query')
d6983cb4
PH
3950 if prefix == '':
3951 return self._get_n_results(query, 1)
3952 elif prefix == 'all':
3953 return self._get_n_results(query, self._MAX_RESULTS)
3954 else:
3955 n = int(prefix)
3956 if n <= 0:
86e5f3ed 3957 raise ExtractorError(f'invalid download number {n} for query "{query}"')
d6983cb4 3958 elif n > self._MAX_RESULTS:
6a39ee13 3959 self.report_warning('%s returns max %i results (you requested %i)' % (self._SEARCH_KEY, self._MAX_RESULTS, n))
d6983cb4
PH
3960 n = self._MAX_RESULTS
3961 return self._get_n_results(query, n)
3962
3963 def _get_n_results(self, query, n):
cc16383f 3964 """Get a specified number of results for a query.
3965 Either this function or _search_results must be overridden by subclasses """
3966 return self.playlist_result(
3967 itertools.islice(self._search_results(query), 0, None if n == float('inf') else n),
3968 query, query)
3969
3970 def _search_results(self, query):
3971 """Returns an iterator of search results"""
611c1dd9 3972 raise NotImplementedError('This method must be implemented by subclasses')
0f818663 3973
82d02080 3974 @classproperty
3975 def SEARCH_KEY(cls):
3976 return cls._SEARCH_KEY
fe7866d0 3977
3978
3979class UnsupportedURLIE(InfoExtractor):
3980 _VALID_URL = '.*'
3981 _ENABLED = False
3982 IE_DESC = False
3983
3984 def _real_extract(self, url):
3985 raise UnsupportedError(url)