]> jfr.im git - yt-dlp.git/blame - yt_dlp/extractor/common.py
Add option `--netrc-cmd` (#6682)
[yt-dlp.git] / yt_dlp / extractor / common.py
CommitLineData
d6983cb4 1import base64
234416e4 2import collections
ac668111 3import getpass
3ec05685 4import hashlib
54007a45 5import http.client
6import http.cookiejar
7import http.cookies
2314b4d8 8import inspect
cc16383f 9import itertools
3d3538e4 10import json
f8271158 11import math
4094b6e3 12import netrc
d6983cb4 13import os
773f291d 14import random
6929b41a 15import re
db3ad8a6 16import subprocess
d6983cb4 17import sys
4094b6e3 18import time
8f97a15d 19import types
14f25df2 20import urllib.parse
ac668111 21import urllib.request
f8271158 22import xml.etree.ElementTree
d6983cb4 23
6929b41a 24from ..compat import functools # isort: split
14f25df2 25from ..compat import compat_etree_fromstring, compat_expanduser, compat_os_name
8817a80d 26from ..cookies import LenientSimpleCookie
f8271158 27from ..downloader.f4m import get_base_url, remove_encrypted_media
8c25f81b 28from ..utils import (
8f97a15d 29 IDENTITY,
f8271158 30 JSON_LD_RE,
31 NO_DEFAULT,
32 ExtractorError,
d0d74b71 33 FormatSorter,
f8271158 34 GeoRestrictedError,
35 GeoUtils,
cb73b846 36 HEADRequest,
b7c47b74 37 LenientJSONDecoder,
db3ad8a6 38 Popen,
f8271158 39 RegexNotFoundError,
be5c1ae8 40 RetryManager,
f8271158 41 UnsupportedError,
05900629 42 age_restricted,
02dc0a36 43 base_url,
08f2a92c 44 bug_reports_message,
82d02080 45 classproperty,
d6983cb4 46 clean_html,
d0d74b71 47 deprecation_warning,
70f0f5a8 48 determine_ext,
d493f15c 49 dict_get,
42676437 50 encode_data_uri,
9b9c5355 51 error_to_compat_str,
46b18f23 52 extract_attributes,
90137ca4 53 filter_dict,
97f4aecf 54 fix_xml_ampersands,
b14f3a4c 55 float_or_none,
b868936c 56 format_field,
31bb8d3f 57 int_or_none,
34921b43 58 join_nonempty,
a4a554a7 59 js_to_json,
46b18f23 60 mimetype2ext,
3158150c 61 network_exceptions,
46b18f23 62 orderedSet,
d493f15c 63 parse_bitrate,
46b18f23
JH
64 parse_codecs,
65 parse_duration,
4ca2a3cf 66 parse_iso8601,
46b18f23 67 parse_m3u8_attributes,
d493f15c 68 parse_resolution,
46b18f23 69 sanitize_filename,
8f97a15d 70 sanitize_url,
b868936c 71 sanitized_Request,
ade1fa70 72 smuggle_url,
d493f15c 73 str_or_none,
ce5b9040 74 str_to_int,
db3ad8a6 75 netrc_from_content,
f856816b 76 strip_or_none,
5d3a0e79 77 traverse_obj,
71df9b7f 78 truncate_string,
47046464 79 try_call,
ffa89477 80 try_get,
f38de77f 81 unescapeHTML,
647eab45 82 unified_strdate,
6b3a3098 83 unified_timestamp,
46b18f23 84 update_Request,
09d02ea4 85 update_url_query,
a107193e 86 url_basename,
bebef109 87 url_or_none,
7e68567e 88 urlhandle_detect_ext,
b868936c 89 urljoin,
6606817a 90 variadic,
a6571f10 91 xpath_element,
8d6765cf
S
92 xpath_text,
93 xpath_with_ns,
d6983cb4 94)
c342041f 95
d6983cb4 96
86e5f3ed 97class InfoExtractor:
d6983cb4
PH
98 """Information Extractor class.
99
100 Information extractors are the classes that, given a URL, extract
101 information about the video (or videos) the URL refers to. This
102 information includes the real video URL, the video title, author and
103 others. The information is stored in a dictionary which is then
5d380852 104 passed to the YoutubeDL. The YoutubeDL processes this
d6983cb4
PH
105 information possibly downloading the video to the file system, among
106 other possible outcomes.
107
cf0649f8 108 The type field determines the type of the result.
fed5d032
PH
109 By far the most common value (and the default if _type is missing) is
110 "video", which indicates a single video.
111
112 For a video, the dictionaries must include the following fields:
d6983cb4
PH
113
114 id: Video identifier.
d4736fdb 115 title: Video title, unescaped. Set to an empty string if video has
116 no title as opposed to "None" which signifies that the
117 extractor failed to obtain a title
d67b0b15 118
f49d89ee 119 Additionally, it must contain either a formats entry or a url one:
d67b0b15 120
f49d89ee
PH
121 formats: A list of dictionaries for each format available, ordered
122 from worst to best quality.
123
124 Potential fields:
c790e93a
S
125 * url The mandatory URL representing the media:
126 for plain file media - HTTP URL of this file,
127 for RTMP - RTMP URL,
128 for HLS - URL of the M3U8 media playlist,
129 for HDS - URL of the F4M manifest,
79d2077e
S
130 for DASH
131 - HTTP URL to plain file media (in case of
132 unfragmented media)
133 - URL of the MPD manifest or base URL
134 representing the media if MPD manifest
8ed7a233 135 is parsed from a string (in case of
79d2077e 136 fragmented media)
c790e93a 137 for MSS - URL of the ISM manifest.
f34804b2 138 * request_data Data to send in POST request to the URL
86f4d14f
S
139 * manifest_url
140 The URL of the manifest file in case of
c790e93a
S
141 fragmented media:
142 for HLS - URL of the M3U8 master playlist,
143 for HDS - URL of the F4M manifest,
144 for DASH - URL of the MPD manifest,
145 for MSS - URL of the ISM manifest.
a44ca5a4 146 * manifest_stream_number (For internal use only)
147 The index of the stream in the manifest file
10952eb2 148 * ext Will be calculated from URL if missing
d67b0b15
PH
149 * format A human-readable description of the format
150 ("mp4 container with h264/opus").
151 Calculated from the format_id, width, height.
152 and format_note fields if missing.
153 * format_id A short description of the format
5d4f3985
PH
154 ("mp4_h264_opus" or "19").
155 Technically optional, but strongly recommended.
d67b0b15
PH
156 * format_note Additional info about the format
157 ("3D" or "DASH video")
158 * width Width of the video, if known
159 * height Height of the video, if known
105bfd90 160 * aspect_ratio Aspect ratio of the video, if known
161 Automatically calculated from width and height
f49d89ee 162 * resolution Textual description of width and height
105bfd90 163 Automatically calculated from width and height
176f1866 164 * dynamic_range The dynamic range of the video. One of:
165 "SDR" (None), "HDR10", "HDR10+, "HDR12", "HLG, "DV"
7217e148 166 * tbr Average bitrate of audio and video in KBit/s
d67b0b15
PH
167 * abr Average audio bitrate in KBit/s
168 * acodec Name of the audio codec in use
dd27fd17 169 * asr Audio sampling rate in Hertz
b8ed0f15 170 * audio_channels Number of audio channels
d67b0b15 171 * vbr Average video bitrate in KBit/s
fbb21cf5 172 * fps Frame rate
d67b0b15 173 * vcodec Name of the video codec in use
1394ce65 174 * container Name of the container format
d67b0b15 175 * filesize The number of bytes, if known in advance
9732d77e 176 * filesize_approx An estimate for the number of bytes
d67b0b15 177 * player_url SWF Player URL (used for rtmpdump).
c7deaa4c 178 * protocol The protocol that will be used for the actual
adbc4ec4
THD
179 download, lower-case. One of "http", "https" or
180 one of the protocols defined in downloader.PROTOCOL_MAP
c58c2d63
S
181 * fragment_base_url
182 Base URL for fragments. Each fragment's path
183 value (if present) will be relative to
184 this URL.
185 * fragments A list of fragments of a fragmented media.
186 Each fragment entry must contain either an url
187 or a path. If an url is present it should be
188 considered by a client. Otherwise both path and
189 fragment_base_url must be present. Here is
190 the list of all potential fields:
191 * "url" - fragment's URL
192 * "path" - fragment's path relative to
193 fragment_base_url
a0d5077c
S
194 * "duration" (optional, int or float)
195 * "filesize" (optional, int)
adbc4ec4
THD
196 * is_from_start Is a live format that can be downloaded
197 from the start. Boolean
f49d89ee 198 * preference Order number of this format. If this field is
08d13955 199 present and not None, the formats get sorted
38d63d84 200 by this field, regardless of all other values.
f49d89ee
PH
201 -1 for default (order by other properties),
202 -2 or smaller for less than default.
e65566a9
PH
203 < -1000 to hide the format (if there is
204 another one which is strictly better)
32f90364
PH
205 * language Language code, e.g. "de" or "en-US".
206 * language_preference Is this in the language mentioned in
207 the URL?
aff2f4f4
PH
208 10 if it's what the URL is about,
209 -1 for default (don't know),
210 -10 otherwise, other values reserved for now.
5d73273f
PH
211 * quality Order number of the video quality of this
212 format, irrespective of the file format.
213 -1 for default (order by other properties),
214 -2 or smaller for less than default.
c64ed2a3
PH
215 * source_preference Order number for this video source
216 (quality takes higher priority)
217 -1 for default (order by other properties),
218 -2 or smaller for less than default.
d769be6c
PH
219 * http_headers A dictionary of additional HTTP headers
220 to add to the request.
6271f1ca 221 * stretched_ratio If given and not 1, indicates that the
3dee7826
PH
222 video's pixels are not square.
223 width : height ratio as float.
224 * no_resume The server does not support resuming the
225 (HTTP or RTMP) download. Boolean.
88acdbc2 226 * has_drm The format has DRM and cannot be downloaded. Boolean
7e68567e 227 * extra_param_to_segment_url A query string to append to each
228 fragment's URL, or to update each existing query string
229 with. Only applied by the native HLS/DASH downloaders.
230 * hls_aes A dictionary of HLS AES-128 decryption information
231 used by the native HLS downloader to override the
232 values in the media playlist when an '#EXT-X-KEY' tag
233 is present in the playlist:
234 * uri The URI from which the key will be downloaded
235 * key The key (as hex) used to decrypt fragments.
236 If `key` is given, any key URI will be ignored
237 * iv The IV (as hex) used to decrypt fragments
0a5a191a 238 * downloader_options A dictionary of downloader options
239 (For internal use only)
240 * http_chunk_size Chunk size for HTTP downloads
241 * ffmpeg_args Extra arguments for ffmpeg downloader
3b1fe47d 242 RTMP formats can also have the additional fields: page_url,
243 app, play_path, tc_url, flash_version, rtmp_live, rtmp_conn,
244 rtmp_protocol, rtmp_real_time
3dee7826 245
c0ba0f48 246 url: Final video URL.
d6983cb4 247 ext: Video filename extension.
d67b0b15
PH
248 format: The video format, defaults to ext (used for --get-format)
249 player_url: SWF Player URL (used for rtmpdump).
2f5865cc 250
d6983cb4
PH
251 The following fields are optional:
252
08d30158 253 direct: True if a direct video file was given (must only be set by GenericIE)
f5e43bc6 254 alt_title: A secondary title of the video.
0afef30b
PH
255 display_id An alternative identifier for the video, not necessarily
256 unique, but available before title. Typically, id is
257 something like "4234987", title "Dancing naked mole rats",
258 and display_id "dancing-naked-mole-rats"
d5519808 259 thumbnails: A list of dictionaries, with the following entries:
cfb56d1a 260 * "id" (optional, string) - Thumbnail format ID
d5519808 261 * "url"
cfb56d1a 262 * "preference" (optional, int) - quality of the image
d5519808
PH
263 * "width" (optional, int)
264 * "height" (optional, int)
5e1c39ac 265 * "resolution" (optional, string "{width}x{height}",
d5519808 266 deprecated)
2de624fd 267 * "filesize" (optional, int)
297e9952 268 * "http_headers" (dict) - HTTP headers for the request
d6983cb4 269 thumbnail: Full URL to a video thumbnail image.
f5e43bc6 270 description: Full video description.
d6983cb4 271 uploader: Full name of the video uploader.
2bc0c46f 272 license: License name the video is licensed under.
8a92e51c 273 creator: The creator of the video.
10db0d2f 274 timestamp: UNIX timestamp of the moment the video was uploaded
ae6a1b95 275 upload_date: Video upload date in UTC (YYYYMMDD).
f0d785d3 276 If not explicitly set, calculated from timestamp
277 release_timestamp: UNIX timestamp of the moment the video was released.
278 If it is not clear whether to use timestamp or this, use the former
ae6a1b95 279 release_date: The date (YYYYMMDD) when the video was released in UTC.
f0d785d3 280 If not explicitly set, calculated from release_timestamp
281 modified_timestamp: UNIX timestamp of the moment the video was last modified.
ae6a1b95 282 modified_date: The date (YYYYMMDD) when the video was last modified in UTC.
f0d785d3 283 If not explicitly set, calculated from modified_timestamp
d6983cb4 284 uploader_id: Nickname or id of the video uploader.
7bcd2830 285 uploader_url: Full URL to a personal webpage of the video uploader.
6f1f59f3 286 channel: Full name of the channel the video is uploaded on.
0e7b8d3e 287 Note that channel fields may or may not repeat uploader
6f1f59f3
S
288 fields. This depends on a particular extractor.
289 channel_id: Id of the channel.
290 channel_url: Full URL to a channel webpage.
6c73052c 291 channel_follower_count: Number of followers of the channel.
8213ce28 292 channel_is_verified: Whether the channel is verified on the platform.
da9ec3b9 293 location: Physical location where the video was filmed.
a504ced0 294 subtitles: The available subtitles as a dictionary in the format
4606c34e
YCH
295 {tag: subformats}. "tag" is usually a language code, and
296 "subformats" is a list sorted from lower to higher
297 preference, each element is a dictionary with the "ext"
298 entry and one of:
a504ced0 299 * "data": The subtitles file contents
10952eb2 300 * "url": A URL pointing to the subtitles file
2412044c 301 It can optionally also have:
302 * "name": Name or description of the subtitles
08d30158 303 * "http_headers": A dictionary of additional HTTP headers
297e9952 304 to add to the request.
4bba3716 305 "ext" will be calculated from URL if missing
e167860c 306 automatic_captions: Like 'subtitles'; contains automatically generated
307 captions instead of normal subtitles
62d231c0 308 duration: Length of the video in seconds, as an integer or float.
f3d29461 309 view_count: How many users have watched the video on the platform.
867c66ff 310 concurrent_view_count: How many users are currently watching the video on the platform.
19e3dfc9
PH
311 like_count: Number of positive ratings of the video
312 dislike_count: Number of negative ratings of the video
02835c6b 313 repost_count: Number of reposts of the video
2d30521a 314 average_rating: Average rating give by users, the scale used depends on the webpage
19e3dfc9 315 comment_count: Number of comments on the video
dd622d7c
PH
316 comments: A list of comments, each with one or more of the following
317 properties (all but one of text or html optional):
318 * "author" - human-readable name of the comment author
319 * "author_id" - user ID of the comment author
a1c5d2ca 320 * "author_thumbnail" - The thumbnail of the comment author
c35448b7 321 * "author_url" - The url to the comment author's page
322 * "author_is_verified" - Whether the author is verified
323 on the platform
324 * "author_is_uploader" - Whether the comment is made by
325 the video uploader
dd622d7c
PH
326 * "id" - Comment ID
327 * "html" - Comment as HTML
328 * "text" - Plain text of the comment
329 * "timestamp" - UNIX timestamp of comment
330 * "parent" - ID of the comment this one is replying to.
331 Set to "root" to indicate that this is a
332 comment to the original video.
a1c5d2ca
M
333 * "like_count" - Number of positive ratings of the comment
334 * "dislike_count" - Number of negative ratings of the comment
335 * "is_favorited" - Whether the comment is marked as
336 favorite by the video uploader
c35448b7 337 * "is_pinned" - Whether the comment is pinned to
338 the top of the comments
8dbe9899 339 age_limit: Age restriction for the video, as an integer (years)
7a5c1cfe 340 webpage_url: The URL to the video webpage, if given to yt-dlp it
9103bbc5
JMF
341 should allow to get the same result again. (It will be set
342 by YoutubeDL if it's missing)
ad3bc6ac
PH
343 categories: A list of categories that the video falls in, for example
344 ["Sports", "Berlin"]
864f24bd 345 tags: A list of tags assigned to the video, e.g. ["sweden", "pop music"]
d0fb4bd1 346 cast: A list of the video cast
7267bd53
PH
347 is_live: True, False, or None (=unknown). Whether this video is a
348 live stream that goes on instead of a fixed-length video.
f76ede8e 349 was_live: True, False, or None (=unknown). Whether this video was
350 originally a live stream.
0647d925 351 live_status: None (=unknown), 'is_live', 'is_upcoming', 'was_live', 'not_live',
e325a21a 352 or 'post_live' (was live, but VOD is not yet processed)
ae30b840 353 If absent, automatically set from is_live, was_live
7c80519c 354 start_time: Time in seconds where the reproduction should start, as
10952eb2 355 specified in the URL.
297a564b 356 end_time: Time in seconds where the reproduction should end, as
10952eb2 357 specified in the URL.
55949fed 358 chapters: A list of dictionaries, with the following entries:
359 * "start_time" - The start time of the chapter in seconds
360 * "end_time" - The end time of the chapter in seconds
361 * "title" (optional, string)
5caf30db
A
362 heatmap: A list of dictionaries, with the following entries:
363 * "start_time" - The start time of the data point in seconds
364 * "end_time" - The end time of the data point in seconds
365 * "value" - The normalized value of the data point (float between 0 and 1)
6cfda058 366 playable_in_embed: Whether this video is allowed to play in embedded
367 players on other sites. Can be True (=always allowed),
368 False (=never allowed), None (=unknown), or a string
62b58c09 369 specifying the criteria for embedability; e.g. 'whitelist'
c224251a
M
370 availability: Under what condition the video is available. One of
371 'private', 'premium_only', 'subscriber_only', 'needs_auth',
372 'unlisted' or 'public'. Use 'InfoExtractor._availability'
373 to set it
1e8fe57e 374 _old_archive_ids: A list of old archive ids needed for backward compatibility
784320c9 375 _format_sort_fields: A list of fields to use for sorting formats
277d6ff5 376 __post_extractor: A function to be called just before the metadata is
377 written to either disk, logger or console. The function
378 must return a dict which will be added to the info_dict.
379 This is usefull for additional information that is
380 time-consuming to extract. Note that the fields thus
381 extracted will not be available to output template and
382 match_filter. So, only "comments" and "comment_count" are
383 currently allowed to be extracted via this method.
d6983cb4 384
7109903e
S
385 The following fields should only be used when the video belongs to some logical
386 chapter or section:
387
388 chapter: Name or title of the chapter the video belongs to.
27bfd4e5
S
389 chapter_number: Number of the chapter the video belongs to, as an integer.
390 chapter_id: Id of the chapter the video belongs to, as a unicode string.
7109903e
S
391
392 The following fields should only be used when the video is an episode of some
8d76bdf1 393 series, programme or podcast:
7109903e
S
394
395 series: Title of the series or programme the video episode belongs to.
9ac24e23 396 series_id: Id of the series or programme the video episode belongs to, as a unicode string.
7109903e 397 season: Title of the season the video episode belongs to.
27bfd4e5
S
398 season_number: Number of the season the video episode belongs to, as an integer.
399 season_id: Id of the season the video episode belongs to, as a unicode string.
7109903e
S
400 episode: Title of the video episode. Unlike mandatory video title field,
401 this field should denote the exact title of the video episode
402 without any kind of decoration.
27bfd4e5
S
403 episode_number: Number of the video episode within a season, as an integer.
404 episode_id: Id of the video episode, as a unicode string.
7109903e 405
7a93ab5f
S
406 The following fields should only be used when the media is a track or a part of
407 a music album:
408
409 track: Title of the track.
410 track_number: Number of the track within an album or a disc, as an integer.
411 track_id: Id of the track (useful in case of custom indexing, e.g. 6.iii),
412 as a unicode string.
413 artist: Artist(s) of the track.
414 genre: Genre(s) of the track.
415 album: Title of the album the track belongs to.
416 album_type: Type of the album (e.g. "Demo", "Full-length", "Split", "Compilation", etc).
417 album_artist: List of all artists appeared on the album (e.g.
418 "Ash Borer / Fell Voices" or "Various Artists", useful for splits
419 and compilations).
420 disc_number: Number of the disc or other physical medium the track belongs to,
421 as an integer.
422 release_year: Year (YYYY) when the album was released.
8bcd4048 423 composer: Composer of the piece
7a93ab5f 424
3975b4d2 425 The following fields should only be set for clips that should be cut from the original video:
426
427 section_start: Start time of the section in seconds
428 section_end: End time of the section in seconds
429
45e8a04e 430 The following fields should only be set for storyboards:
431 rows: Number of rows in each storyboard fragment, as an integer
432 columns: Number of columns in each storyboard fragment, as an integer
433
deefc05b 434 Unless mentioned otherwise, the fields should be Unicode strings.
d6983cb4 435
d838b1bd
PH
436 Unless mentioned otherwise, None is equivalent to absence of information.
437
fed5d032
PH
438
439 _type "playlist" indicates multiple videos.
b82f815f
PH
440 There must be a key "entries", which is a list, an iterable, or a PagedList
441 object, each element of which is a valid dictionary by this specification.
fed5d032 442
962ffcf8 443 Additionally, playlists can have "id", "title", and any other relevant
b60419c5 444 attributes with the same semantics as videos (see above).
fed5d032 445
f0d785d3 446 It can also have the following optional fields:
447
448 playlist_count: The total number of videos in a playlist. If not given,
449 YoutubeDL tries to calculate it from "entries"
450
fed5d032
PH
451
452 _type "multi_video" indicates that there are multiple videos that
453 form a single show, for examples multiple acts of an opera or TV episode.
454 It must have an entries key like a playlist and contain all the keys
455 required for a video at the same time.
456
457
458 _type "url" indicates that the video must be extracted from another
459 location, possibly by a different extractor. Its only required key is:
460 "url" - the next URL to extract.
f58766ce
PH
461 The key "ie_key" can be set to the class name (minus the trailing "IE",
462 e.g. "Youtube") if the extractor class is known in advance.
463 Additionally, the dictionary may have any properties of the resolved entity
464 known in advance, for example "title" if the title of the referred video is
fed5d032
PH
465 known ahead of time.
466
467
468 _type "url_transparent" entities have the same specification as "url", but
469 indicate that the given additional information is more precise than the one
470 associated with the resolved URL.
471 This is useful when a site employs a video service that hosts the video and
472 its technical metadata, but that video service does not embed a useful
473 title, description etc.
474
475
8f97a15d 476 Subclasses of this should also be added to the list of extractors and
477 should define a _VALID_URL regexp and, re-define the _real_extract() and
478 (optionally) _real_initialize() methods.
d6983cb4 479
e6f21b3d 480 Subclasses may also override suitable() if necessary, but ensure the function
481 signature is preserved and that this function imports everything it needs
52efa4b3 482 (except other extractors), so that lazy_extractors works correctly.
483
8f97a15d 484 Subclasses can define a list of _EMBED_REGEX, which will be searched for in
485 the HTML of Generic webpages. It may also override _extract_embed_urls
486 or _extract_from_webpage as necessary. While these are normally classmethods,
487 _extract_from_webpage is allowed to be an instance method.
488
489 _extract_from_webpage may raise self.StopExtraction() to stop further
490 processing of the webpage and obtain exclusive rights to it. This is useful
62b58c09
L
491 when the extractor cannot reliably be matched using just the URL,
492 e.g. invidious/peertube instances
8f97a15d 493
494 Embed-only extractors can be defined by setting _VALID_URL = False.
495
52efa4b3 496 To support username + password (or netrc) login, the extractor must define a
497 _NETRC_MACHINE and re-define _perform_login(username, password) and
498 (optionally) _initialize_pre_login() methods. The _perform_login method will
499 be called between _initialize_pre_login and _real_initialize if credentials
500 are passed by the user. In cases where it is necessary to have the login
501 process as part of the extraction rather than initialization, _perform_login
502 can be left undefined.
e6f21b3d 503
4248dad9 504 _GEO_BYPASS attribute may be set to False in order to disable
773f291d
S
505 geo restriction bypass mechanisms for a particular extractor.
506 Though it won't disable explicit geo restriction bypass based on
504f20dd 507 country code provided with geo_bypass_country.
4248dad9
S
508
509 _GEO_COUNTRIES attribute may contain a list of presumably geo unrestricted
510 countries for this extractor. One of these countries will be used by
511 geo restriction bypass mechanism right away in order to bypass
504f20dd 512 geo restriction, of course, if the mechanism is not disabled.
773f291d 513
5f95927a
S
514 _GEO_IP_BLOCKS attribute may contain a list of presumably geo unrestricted
515 IP blocks in CIDR notation for this extractor. One of these IP blocks
516 will be used by geo restriction bypass mechanism similarly
504f20dd 517 to _GEO_COUNTRIES.
3ccdde8c 518
fe7866d0 519 The _ENABLED attribute should be set to False for IEs that
520 are disabled by default and must be explicitly enabled.
521
e6f21b3d 522 The _WORKING attribute should be set to False for broken IEs
d6983cb4
PH
523 in order to warn the users and skip the tests.
524 """
525
526 _ready = False
527 _downloader = None
773f291d 528 _x_forwarded_for_ip = None
4248dad9
S
529 _GEO_BYPASS = True
530 _GEO_COUNTRIES = None
5f95927a 531 _GEO_IP_BLOCKS = None
d6983cb4 532 _WORKING = True
fe7866d0 533 _ENABLED = True
52efa4b3 534 _NETRC_MACHINE = None
231025c4 535 IE_DESC = None
8dcce6a8 536 SEARCH_KEY = None
8f97a15d 537 _VALID_URL = None
538 _EMBED_REGEX = []
d6983cb4 539
8dcce6a8 540 def _login_hint(self, method=NO_DEFAULT, netrc=None):
db3ad8a6 541 password_hint = f'--username and --password, --netrc-cmd, or --netrc ({netrc or self._NETRC_MACHINE}) to provide account credentials'
8dcce6a8 542 return {
543 None: '',
544 'any': f'Use --cookies, --cookies-from-browser, {password_hint}',
545 'password': f'Use {password_hint}',
546 'cookies': (
547 'Use --cookies-from-browser or --cookies for the authentication. '
17ffed18 548 'See https://github.com/yt-dlp/yt-dlp/wiki/FAQ#how-do-i-pass-cookies-to-yt-dlp for how to manually pass cookies'),
8dcce6a8 549 }[method if method is not NO_DEFAULT else 'any' if self.supports_login() else 'cookies']
9d5d4d64 550
d6983cb4 551 def __init__(self, downloader=None):
49a57e70 552 """Constructor. Receives an optional downloader (a YoutubeDL instance).
553 If a downloader is not passed during initialization,
554 it must be set using "set_downloader()" before "extract()" is called"""
d6983cb4 555 self._ready = False
773f291d 556 self._x_forwarded_for_ip = None
28f436ba 557 self._printed_messages = set()
d6983cb4
PH
558 self.set_downloader(downloader)
559
560 @classmethod
5ad28e7f 561 def _match_valid_url(cls, url):
8f97a15d 562 if cls._VALID_URL is False:
563 return None
79cb2577
PH
564 # This does not use has/getattr intentionally - we want to know whether
565 # we have cached the regexp for *this* class, whereas getattr would also
566 # match the superclass
567 if '_VALID_URL_RE' not in cls.__dict__:
568 cls._VALID_URL_RE = re.compile(cls._VALID_URL)
5ad28e7f 569 return cls._VALID_URL_RE.match(url)
570
571 @classmethod
572 def suitable(cls, url):
573 """Receives a URL and returns True if suitable for this IE."""
3fb4e21b 574 # This function must import everything it needs (except other extractors),
575 # so that lazy_extractors works correctly
5ad28e7f 576 return cls._match_valid_url(url) is not None
d6983cb4 577
ed9266db
PH
578 @classmethod
579 def _match_id(cls, url):
5ad28e7f 580 return cls._match_valid_url(url).group('id')
ed9266db 581
1151c407 582 @classmethod
583 def get_temp_id(cls, url):
584 try:
585 return cls._match_id(url)
586 except (IndexError, AttributeError):
587 return None
588
d6983cb4
PH
589 @classmethod
590 def working(cls):
591 """Getter method for _WORKING."""
592 return cls._WORKING
593
52efa4b3 594 @classmethod
595 def supports_login(cls):
596 return bool(cls._NETRC_MACHINE)
597
d6983cb4
PH
598 def initialize(self):
599 """Initializes an instance (authentication, etc)."""
28f436ba 600 self._printed_messages = set()
5f95927a
S
601 self._initialize_geo_bypass({
602 'countries': self._GEO_COUNTRIES,
603 'ip_blocks': self._GEO_IP_BLOCKS,
604 })
4248dad9 605 if not self._ready:
52efa4b3 606 self._initialize_pre_login()
607 if self.supports_login():
608 username, password = self._get_login_info()
609 if username:
610 self._perform_login(username, password)
611 elif self.get_param('username') and False not in (self.IE_DESC, self._NETRC_MACHINE):
8dcce6a8 612 self.report_warning(f'Login with password is not supported for this website. {self._login_hint("cookies")}')
4248dad9
S
613 self._real_initialize()
614 self._ready = True
615
5f95927a 616 def _initialize_geo_bypass(self, geo_bypass_context):
e39b5d4a
S
617 """
618 Initialize geo restriction bypass mechanism.
619
620 This method is used to initialize geo bypass mechanism based on faking
621 X-Forwarded-For HTTP header. A random country from provided country list
dc0a869e 622 is selected and a random IP belonging to this country is generated. This
e39b5d4a
S
623 IP will be passed as X-Forwarded-For HTTP header in all subsequent
624 HTTP requests.
e39b5d4a
S
625
626 This method will be used for initial geo bypass mechanism initialization
5f95927a
S
627 during the instance initialization with _GEO_COUNTRIES and
628 _GEO_IP_BLOCKS.
e39b5d4a 629
5f95927a 630 You may also manually call it from extractor's code if geo bypass
e39b5d4a 631 information is not available beforehand (e.g. obtained during
5f95927a
S
632 extraction) or due to some other reason. In this case you should pass
633 this information in geo bypass context passed as first argument. It may
634 contain following fields:
635
636 countries: List of geo unrestricted countries (similar
637 to _GEO_COUNTRIES)
638 ip_blocks: List of geo unrestricted IP blocks in CIDR notation
639 (similar to _GEO_IP_BLOCKS)
640
e39b5d4a 641 """
773f291d 642 if not self._x_forwarded_for_ip:
5f95927a
S
643
644 # Geo bypass mechanism is explicitly disabled by user
a06916d9 645 if not self.get_param('geo_bypass', True):
5f95927a
S
646 return
647
648 if not geo_bypass_context:
649 geo_bypass_context = {}
650
651 # Backward compatibility: previously _initialize_geo_bypass
652 # expected a list of countries, some 3rd party code may still use
653 # it this way
654 if isinstance(geo_bypass_context, (list, tuple)):
655 geo_bypass_context = {
656 'countries': geo_bypass_context,
657 }
658
659 # The whole point of geo bypass mechanism is to fake IP
660 # as X-Forwarded-For HTTP header based on some IP block or
661 # country code.
662
663 # Path 1: bypassing based on IP block in CIDR notation
664
665 # Explicit IP block specified by user, use it right away
666 # regardless of whether extractor is geo bypassable or not
a06916d9 667 ip_block = self.get_param('geo_bypass_ip_block', None)
5f95927a
S
668
669 # Otherwise use random IP block from geo bypass context but only
670 # if extractor is known as geo bypassable
671 if not ip_block:
672 ip_blocks = geo_bypass_context.get('ip_blocks')
673 if self._GEO_BYPASS and ip_blocks:
674 ip_block = random.choice(ip_blocks)
675
676 if ip_block:
677 self._x_forwarded_for_ip = GeoUtils.random_ipv4(ip_block)
8a82af35 678 self.write_debug(f'Using fake IP {self._x_forwarded_for_ip} as X-Forwarded-For')
5f95927a
S
679 return
680
681 # Path 2: bypassing based on country code
682
683 # Explicit country code specified by user, use it right away
684 # regardless of whether extractor is geo bypassable or not
a06916d9 685 country = self.get_param('geo_bypass_country', None)
5f95927a
S
686
687 # Otherwise use random country code from geo bypass context but
688 # only if extractor is known as geo bypassable
689 if not country:
690 countries = geo_bypass_context.get('countries')
691 if self._GEO_BYPASS and countries:
692 country = random.choice(countries)
693
694 if country:
695 self._x_forwarded_for_ip = GeoUtils.random_ipv4(country)
0760b0a7 696 self._downloader.write_debug(
86e5f3ed 697 f'Using fake IP {self._x_forwarded_for_ip} ({country.upper()}) as X-Forwarded-For')
d6983cb4
PH
698
699 def extract(self, url):
700 """Extracts URL information and returns it in list of dicts."""
3a5bcd03 701 try:
773f291d
S
702 for _ in range(2):
703 try:
704 self.initialize()
71df9b7f 705 self.to_screen('Extracting URL: %s' % (
706 url if self.get_param('verbose') else truncate_string(url, 100, 20)))
0016b84e 707 ie_result = self._real_extract(url)
07cce701 708 if ie_result is None:
709 return None
0016b84e
S
710 if self._x_forwarded_for_ip:
711 ie_result['__x_forwarded_for_ip'] = self._x_forwarded_for_ip
b79f9e30 712 subtitles = ie_result.get('subtitles') or {}
713 if 'no-live-chat' in self.get_param('compat_opts'):
714 for lang in ('live_chat', 'comments', 'danmaku'):
715 subtitles.pop(lang, None)
0016b84e 716 return ie_result
773f291d 717 except GeoRestrictedError as e:
4248dad9
S
718 if self.__maybe_fake_ip_and_retry(e.countries):
719 continue
773f291d 720 raise
0db3bae8 721 except UnsupportedError:
722 raise
1151c407 723 except ExtractorError as e:
9bcfe33b 724 e.video_id = e.video_id or self.get_temp_id(url),
725 e.ie = e.ie or self.IE_NAME,
726 e.traceback = e.traceback or sys.exc_info()[2]
727 raise
ac668111 728 except http.client.IncompleteRead as e:
1151c407 729 raise ExtractorError('A network error has occurred.', cause=e, expected=True, video_id=self.get_temp_id(url))
9650885b 730 except (KeyError, StopIteration) as e:
1151c407 731 raise ExtractorError('An extractor error has occurred.', cause=e, video_id=self.get_temp_id(url))
d6983cb4 732
4248dad9 733 def __maybe_fake_ip_and_retry(self, countries):
a06916d9 734 if (not self.get_param('geo_bypass_country', None)
3089bc74 735 and self._GEO_BYPASS
a06916d9 736 and self.get_param('geo_bypass', True)
3089bc74
S
737 and not self._x_forwarded_for_ip
738 and countries):
eea0716c
S
739 country_code = random.choice(countries)
740 self._x_forwarded_for_ip = GeoUtils.random_ipv4(country_code)
4248dad9
S
741 if self._x_forwarded_for_ip:
742 self.report_warning(
eea0716c
S
743 'Video is geo restricted. Retrying extraction with fake IP %s (%s) as X-Forwarded-For.'
744 % (self._x_forwarded_for_ip, country_code.upper()))
4248dad9
S
745 return True
746 return False
747
d6983cb4 748 def set_downloader(self, downloader):
08d30158 749 """Sets a YoutubeDL instance as the downloader for this IE."""
d6983cb4
PH
750 self._downloader = downloader
751
9809740b 752 @property
753 def cache(self):
754 return self._downloader.cache
755
756 @property
757 def cookiejar(self):
758 return self._downloader.cookiejar
759
52efa4b3 760 def _initialize_pre_login(self):
962ffcf8 761 """ Initialization before login. Redefine in subclasses."""
52efa4b3 762 pass
763
764 def _perform_login(self, username, password):
765 """ Login with username and password. Redefine in subclasses."""
766 pass
767
d6983cb4
PH
768 def _real_initialize(self):
769 """Real initialization process. Redefine in subclasses."""
770 pass
771
772 def _real_extract(self, url):
773 """Real extraction process. Redefine in subclasses."""
08d30158 774 raise NotImplementedError('This method must be implemented by subclasses')
d6983cb4 775
56c73665
JMF
776 @classmethod
777 def ie_key(cls):
778 """A string for getting the InfoExtractor with get_info_extractor"""
3fb4e21b 779 return cls.__name__[:-2]
56c73665 780
82d02080 781 @classproperty
782 def IE_NAME(cls):
783 return cls.__name__[:-2]
d6983cb4 784
d391b7e2
S
785 @staticmethod
786 def __can_accept_status_code(err, expected_status):
ac668111 787 assert isinstance(err, urllib.error.HTTPError)
d391b7e2
S
788 if expected_status is None:
789 return False
d391b7e2
S
790 elif callable(expected_status):
791 return expected_status(err.code) is True
792 else:
6606817a 793 return err.code in variadic(expected_status)
d391b7e2 794
c043c246 795 def _create_request(self, url_or_request, data=None, headers=None, query=None):
ac668111 796 if isinstance(url_or_request, urllib.request.Request):
09d02ea4 797 return update_Request(url_or_request, data=data, headers=headers, query=query)
798 if query:
799 url_or_request = update_url_query(url_or_request, query)
c043c246 800 return sanitized_Request(url_or_request, data, headers or {})
f95b9dee 801
c043c246 802 def _request_webpage(self, url_or_request, video_id, note=None, errnote=None, fatal=True, data=None, headers=None, query=None, expected_status=None):
d391b7e2
S
803 """
804 Return the response handle.
805
806 See _download_webpage docstring for arguments specification.
807 """
1cf376f5 808 if not self._downloader._first_webpage_request:
49a57e70 809 sleep_interval = self.get_param('sleep_interval_requests') or 0
1cf376f5 810 if sleep_interval > 0:
5ef7d9bd 811 self.to_screen('Sleeping %s seconds ...' % sleep_interval)
1cf376f5 812 time.sleep(sleep_interval)
813 else:
814 self._downloader._first_webpage_request = False
815
d6983cb4
PH
816 if note is None:
817 self.report_download_webpage(video_id)
818 elif note is not False:
7cc3570e 819 if video_id is None:
86e5f3ed 820 self.to_screen(str(note))
7cc3570e 821 else:
86e5f3ed 822 self.to_screen(f'{video_id}: {note}')
2132edaa
S
823
824 # Some sites check X-Forwarded-For HTTP header in order to figure out
825 # the origin of the client behind proxy. This allows bypassing geo
826 # restriction by faking this header's value to IP that belongs to some
827 # geo unrestricted country. We will do so once we encounter any
828 # geo restriction error.
829 if self._x_forwarded_for_ip:
c043c246 830 headers = (headers or {}).copy()
831 headers.setdefault('X-Forwarded-For', self._x_forwarded_for_ip)
2132edaa 832
d6983cb4 833 try:
f95b9dee 834 return self._downloader.urlopen(self._create_request(url_or_request, data, headers, query))
3158150c 835 except network_exceptions as err:
ac668111 836 if isinstance(err, urllib.error.HTTPError):
d391b7e2 837 if self.__can_accept_status_code(err, expected_status):
95e42d73
XDG
838 # Retain reference to error to prevent file object from
839 # being closed before it can be read. Works around the
840 # effects of <https://bugs.python.org/issue15002>
841 # introduced in Python 3.4.1.
842 err.fp._error = err
d391b7e2
S
843 return err.fp
844
aa94a6d3
PH
845 if errnote is False:
846 return False
d6983cb4 847 if errnote is None:
f1a9d64e 848 errnote = 'Unable to download webpage'
7f8b2714 849
86e5f3ed 850 errmsg = f'{errnote}: {error_to_compat_str(err)}'
7cc3570e 851 if fatal:
497d2fab 852 raise ExtractorError(errmsg, cause=err)
7cc3570e 853 else:
6a39ee13 854 self.report_warning(errmsg)
7cc3570e 855 return False
d6983cb4 856
1890fc63 857 def _download_webpage_handle(self, url_or_request, video_id, note=None, errnote=None, fatal=True,
858 encoding=None, data=None, headers={}, query={}, expected_status=None):
d391b7e2
S
859 """
860 Return a tuple (page content as string, URL handle).
861
617f658b 862 Arguments:
863 url_or_request -- plain text URL as a string or
ac668111 864 a urllib.request.Request object
617f658b 865 video_id -- Video/playlist/item identifier (string)
866
867 Keyword arguments:
868 note -- note printed before downloading (string)
869 errnote -- note printed in case of an error (string)
870 fatal -- flag denoting whether error should be considered fatal,
871 i.e. whether it should cause ExtractionError to be raised,
872 otherwise a warning will be reported and extraction continued
873 encoding -- encoding for a page content decoding, guessed automatically
874 when not explicitly specified
875 data -- POST data (bytes)
876 headers -- HTTP headers (dict)
877 query -- URL query (dict)
878 expected_status -- allows to accept failed HTTP requests (non 2xx
879 status code) by explicitly specifying a set of accepted status
880 codes. Can be any of the following entities:
881 - an integer type specifying an exact failed status code to
882 accept
883 - a list or a tuple of integer types specifying a list of
884 failed status codes to accept
885 - a callable accepting an actual failed status code and
886 returning True if it should be accepted
887 Note that this argument does not affect success status codes (2xx)
888 which are always accepted.
d391b7e2 889 """
617f658b 890
b9d3e163 891 # Strip hashes from the URL (#1038)
14f25df2 892 if isinstance(url_or_request, str):
b9d3e163
PH
893 url_or_request = url_or_request.partition('#')[0]
894
d391b7e2 895 urlh = self._request_webpage(url_or_request, video_id, note, errnote, fatal, data=data, headers=headers, query=query, expected_status=expected_status)
7cc3570e
PH
896 if urlh is False:
897 assert not fatal
898 return False
c9a77969 899 content = self._webpage_read_content(urlh, url_or_request, video_id, note, errnote, fatal, encoding=encoding)
23be51d8
PH
900 return (content, urlh)
901
c9a77969
YCH
902 @staticmethod
903 def _guess_encoding_from_content(content_type, webpage_bytes):
d6983cb4
PH
904 m = re.match(r'[a-zA-Z0-9_.-]+/[a-zA-Z0-9_.-]+\s*;\s*charset=(.+)', content_type)
905 if m:
906 encoding = m.group(1)
907 else:
0d75ae2c 908 m = re.search(br'<meta[^>]+charset=[\'"]?([^\'")]+)[ /\'">]',
f143d86a
PH
909 webpage_bytes[:1024])
910 if m:
911 encoding = m.group(1).decode('ascii')
b60016e8
PH
912 elif webpage_bytes.startswith(b'\xff\xfe'):
913 encoding = 'utf-16'
f143d86a
PH
914 else:
915 encoding = 'utf-8'
c9a77969
YCH
916
917 return encoding
918
4457823d
S
919 def __check_blocked(self, content):
920 first_block = content[:512]
3089bc74
S
921 if ('<title>Access to this site is blocked</title>' in content
922 and 'Websense' in first_block):
4457823d
S
923 msg = 'Access to this webpage has been blocked by Websense filtering software in your network.'
924 blocked_iframe = self._html_search_regex(
925 r'<iframe src="([^"]+)"', content,
926 'Websense information URL', default=None)
927 if blocked_iframe:
928 msg += ' Visit %s for more details' % blocked_iframe
929 raise ExtractorError(msg, expected=True)
930 if '<title>The URL you requested has been blocked</title>' in first_block:
931 msg = (
932 'Access to this webpage has been blocked by Indian censorship. '
933 'Use a VPN or proxy server (with --proxy) to route around it.')
934 block_msg = self._html_search_regex(
935 r'</h1><p>(.*?)</p>',
936 content, 'block message', default=None)
937 if block_msg:
938 msg += ' (Message: "%s")' % block_msg.replace('\n', ' ')
939 raise ExtractorError(msg, expected=True)
3089bc74
S
940 if ('<title>TTK :: Доступ к ресурсу ограничен</title>' in content
941 and 'blocklist.rkn.gov.ru' in content):
4457823d
S
942 raise ExtractorError(
943 'Access to this webpage has been blocked by decision of the Russian government. '
944 'Visit http://blocklist.rkn.gov.ru/ for a block reason.',
945 expected=True)
946
f95b9dee 947 def _request_dump_filename(self, url, video_id):
948 basen = f'{video_id}_{url}'
949 trim_length = self.get_param('trim_file_name') or 240
950 if len(basen) > trim_length:
951 h = '___' + hashlib.md5(basen.encode('utf-8')).hexdigest()
952 basen = basen[:trim_length - len(h)] + h
953 filename = sanitize_filename(f'{basen}.dump', restricted=True)
954 # Working around MAX_PATH limitation on Windows (see
955 # http://msdn.microsoft.com/en-us/library/windows/desktop/aa365247(v=vs.85).aspx)
956 if compat_os_name == 'nt':
957 absfilepath = os.path.abspath(filename)
958 if len(absfilepath) > 259:
959 filename = fR'\\?\{absfilepath}'
960 return filename
961
962 def __decode_webpage(self, webpage_bytes, encoding, headers):
963 if not encoding:
964 encoding = self._guess_encoding_from_content(headers.get('Content-Type', ''), webpage_bytes)
965 try:
966 return webpage_bytes.decode(encoding, 'replace')
967 except LookupError:
968 return webpage_bytes.decode('utf-8', 'replace')
969
c9a77969 970 def _webpage_read_content(self, urlh, url_or_request, video_id, note=None, errnote=None, fatal=True, prefix=None, encoding=None):
c9a77969
YCH
971 webpage_bytes = urlh.read()
972 if prefix is not None:
973 webpage_bytes = prefix + webpage_bytes
a06916d9 974 if self.get_param('dump_intermediate_pages', False):
f610dbb0 975 self.to_screen('Dumping request to ' + urlh.geturl())
d6983cb4
PH
976 dump = base64.b64encode(webpage_bytes).decode('ascii')
977 self._downloader.to_screen(dump)
f95b9dee 978 if self.get_param('write_pages'):
e121e3ce 979 filename = self._request_dump_filename(urlh.geturl(), video_id)
f95b9dee 980 self.to_screen(f'Saving request to {filename}')
d41e6efc
PH
981 with open(filename, 'wb') as outf:
982 outf.write(webpage_bytes)
983
f95b9dee 984 content = self.__decode_webpage(webpage_bytes, encoding, urlh.headers)
4457823d 985 self.__check_blocked(content)
2410c43d 986
23be51d8 987 return content
d6983cb4 988
6edf2808 989 def __print_error(self, errnote, fatal, video_id, err):
990 if fatal:
c6e07cf1 991 raise ExtractorError(f'{video_id}: {errnote}', cause=err)
6edf2808 992 elif errnote:
c6e07cf1 993 self.report_warning(f'{video_id}: {errnote}: {err}')
6edf2808 994
995 def _parse_xml(self, xml_string, video_id, transform_source=None, fatal=True, errnote=None):
e2b38da9
PH
996 if transform_source:
997 xml_string = transform_source(xml_string)
e01c3d2e
S
998 try:
999 return compat_etree_fromstring(xml_string.encode('utf-8'))
f9934b96 1000 except xml.etree.ElementTree.ParseError as ve:
6edf2808 1001 self.__print_error('Failed to parse XML' if errnote is None else errnote, fatal, video_id, ve)
267ed0c5 1002
6edf2808 1003 def _parse_json(self, json_string, video_id, transform_source=None, fatal=True, errnote=None, **parser_kwargs):
3d3538e4 1004 try:
b7c47b74 1005 return json.loads(
1006 json_string, cls=LenientJSONDecoder, strict=False, transform_source=transform_source, **parser_kwargs)
3d3538e4 1007 except ValueError as ve:
6edf2808 1008 self.__print_error('Failed to parse JSON' if errnote is None else errnote, fatal, video_id, ve)
3d3538e4 1009
6edf2808 1010 def _parse_socket_response_as_json(self, data, *args, **kwargs):
1011 return self._parse_json(data[data.find('{'):data.rfind('}') + 1], *args, **kwargs)
adddc50c 1012
617f658b 1013 def __create_download_methods(name, parser, note, errnote, return_value):
1014
6edf2808 1015 def parse(ie, content, *args, errnote=errnote, **kwargs):
617f658b 1016 if parser is None:
1017 return content
6edf2808 1018 if errnote is False:
1019 kwargs['errnote'] = errnote
617f658b 1020 # parser is fetched by name so subclasses can override it
1021 return getattr(ie, parser)(content, *args, **kwargs)
1022
c4910024 1023 def download_handle(self, url_or_request, video_id, note=note, errnote=errnote, transform_source=None,
1024 fatal=True, encoding=None, data=None, headers={}, query={}, expected_status=None):
1025 res = self._download_webpage_handle(
1026 url_or_request, video_id, note=note, errnote=errnote, fatal=fatal, encoding=encoding,
1027 data=data, headers=headers, query=query, expected_status=expected_status)
617f658b 1028 if res is False:
1029 return res
1030 content, urlh = res
6edf2808 1031 return parse(self, content, video_id, transform_source=transform_source, fatal=fatal, errnote=errnote), urlh
617f658b 1032
f95b9dee 1033 def download_content(self, url_or_request, video_id, note=note, errnote=errnote, transform_source=None,
c4910024 1034 fatal=True, encoding=None, data=None, headers={}, query={}, expected_status=None):
f95b9dee 1035 if self.get_param('load_pages'):
1036 url_or_request = self._create_request(url_or_request, data, headers, query)
1037 filename = self._request_dump_filename(url_or_request.full_url, video_id)
1038 self.to_screen(f'Loading request from {filename}')
1039 try:
1040 with open(filename, 'rb') as dumpf:
1041 webpage_bytes = dumpf.read()
1042 except OSError as e:
1043 self.report_warning(f'Unable to load request from disk: {e}')
1044 else:
1045 content = self.__decode_webpage(webpage_bytes, encoding, url_or_request.headers)
6edf2808 1046 return parse(self, content, video_id, transform_source=transform_source, fatal=fatal, errnote=errnote)
c4910024 1047 kwargs = {
1048 'note': note,
1049 'errnote': errnote,
1050 'transform_source': transform_source,
1051 'fatal': fatal,
1052 'encoding': encoding,
1053 'data': data,
1054 'headers': headers,
1055 'query': query,
1056 'expected_status': expected_status,
1057 }
617f658b 1058 if parser is None:
c4910024 1059 kwargs.pop('transform_source')
617f658b 1060 # The method is fetched by name so subclasses can override _download_..._handle
c4910024 1061 res = getattr(self, download_handle.__name__)(url_or_request, video_id, **kwargs)
617f658b 1062 return res if res is False else res[0]
1063
1064 def impersonate(func, name, return_value):
1065 func.__name__, func.__qualname__ = name, f'InfoExtractor.{name}'
1066 func.__doc__ = f'''
1067 @param transform_source Apply this transformation before parsing
1068 @returns {return_value}
1069
1070 See _download_webpage_handle docstring for other arguments specification
1071 '''
1072
1073 impersonate(download_handle, f'_download_{name}_handle', f'({return_value}, URL handle)')
1074 impersonate(download_content, f'_download_{name}', f'{return_value}')
1075 return download_handle, download_content
1076
1077 _download_xml_handle, _download_xml = __create_download_methods(
1078 'xml', '_parse_xml', 'Downloading XML', 'Unable to download XML', 'xml as an xml.etree.ElementTree.Element')
1079 _download_json_handle, _download_json = __create_download_methods(
1080 'json', '_parse_json', 'Downloading JSON metadata', 'Unable to download JSON metadata', 'JSON object as a dict')
1081 _download_socket_json_handle, _download_socket_json = __create_download_methods(
1082 'socket_json', '_parse_socket_response_as_json', 'Polling socket', 'Unable to poll socket', 'JSON object as a dict')
1083 __download_webpage = __create_download_methods('webpage', None, None, None, 'data of the page as a string')[1]
adddc50c 1084
617f658b 1085 def _download_webpage(
1086 self, url_or_request, video_id, note=None, errnote=None,
1087 fatal=True, tries=1, timeout=NO_DEFAULT, *args, **kwargs):
adddc50c 1088 """
617f658b 1089 Return the data of the page as a string.
adddc50c 1090
617f658b 1091 Keyword arguments:
1092 tries -- number of tries
1093 timeout -- sleep interval between tries
1094
1095 See _download_webpage_handle docstring for other arguments specification.
adddc50c 1096 """
617f658b 1097
1098 R''' # NB: These are unused; should they be deprecated?
1099 if tries != 1:
1100 self._downloader.deprecation_warning('tries argument is deprecated in InfoExtractor._download_webpage')
1101 if timeout is NO_DEFAULT:
1102 timeout = 5
1103 else:
1104 self._downloader.deprecation_warning('timeout argument is deprecated in InfoExtractor._download_webpage')
1105 '''
1106
1107 try_count = 0
1108 while True:
1109 try:
1110 return self.__download_webpage(url_or_request, video_id, note, errnote, None, fatal, *args, **kwargs)
ac668111 1111 except http.client.IncompleteRead as e:
617f658b 1112 try_count += 1
1113 if try_count >= tries:
1114 raise e
1115 self._sleep(timeout, video_id)
adddc50c 1116
28f436ba 1117 def report_warning(self, msg, video_id=None, *args, only_once=False, **kwargs):
a70635b8 1118 idstr = format_field(video_id, None, '%s: ')
28f436ba 1119 msg = f'[{self.IE_NAME}] {idstr}{msg}'
1120 if only_once:
1121 if f'WARNING: {msg}' in self._printed_messages:
1122 return
1123 self._printed_messages.add(f'WARNING: {msg}')
1124 self._downloader.report_warning(msg, *args, **kwargs)
f45f96f8 1125
a06916d9 1126 def to_screen(self, msg, *args, **kwargs):
d6983cb4 1127 """Print msg to screen, prefixing it with '[ie_name]'"""
86e5f3ed 1128 self._downloader.to_screen(f'[{self.IE_NAME}] {msg}', *args, **kwargs)
a06916d9 1129
1130 def write_debug(self, msg, *args, **kwargs):
86e5f3ed 1131 self._downloader.write_debug(f'[{self.IE_NAME}] {msg}', *args, **kwargs)
a06916d9 1132
1133 def get_param(self, name, default=None, *args, **kwargs):
1134 if self._downloader:
1135 return self._downloader.params.get(name, default, *args, **kwargs)
1136 return default
d6983cb4 1137
d5d1df8a 1138 def report_drm(self, video_id, partial=NO_DEFAULT):
1139 if partial is not NO_DEFAULT:
1140 self._downloader.deprecation_warning('InfoExtractor.report_drm no longer accepts the argument partial')
88acdbc2 1141 self.raise_no_formats('This video is DRM protected', expected=True, video_id=video_id)
1142
d6983cb4
PH
1143 def report_extraction(self, id_or_name):
1144 """Report information extraction."""
f1a9d64e 1145 self.to_screen('%s: Extracting information' % id_or_name)
d6983cb4
PH
1146
1147 def report_download_webpage(self, video_id):
1148 """Report webpage download."""
f1a9d64e 1149 self.to_screen('%s: Downloading webpage' % video_id)
d6983cb4
PH
1150
1151 def report_age_confirmation(self):
1152 """Report attempt to confirm age."""
f1a9d64e 1153 self.to_screen('Confirming age')
d6983cb4 1154
fc79158d
JMF
1155 def report_login(self):
1156 """Report attempt to log in."""
f1a9d64e 1157 self.to_screen('Logging in')
fc79158d 1158
b7da73eb 1159 def raise_login_required(
9d5d4d64 1160 self, msg='This video is only available for registered users',
52efa4b3 1161 metadata_available=False, method=NO_DEFAULT):
f2ebc5c7 1162 if metadata_available and (
1163 self.get_param('ignore_no_formats_error') or self.get_param('wait_for_video')):
b7da73eb 1164 self.report_warning(msg)
7265a219 1165 return
a70635b8 1166 msg += format_field(self._login_hint(method), None, '. %s')
46890374 1167 raise ExtractorError(msg, expected=True)
43e7d3c9 1168
b7da73eb 1169 def raise_geo_restricted(
1170 self, msg='This video is not available from your location due to geo restriction',
1171 countries=None, metadata_available=False):
f2ebc5c7 1172 if metadata_available and (
1173 self.get_param('ignore_no_formats_error') or self.get_param('wait_for_video')):
b7da73eb 1174 self.report_warning(msg)
1175 else:
1176 raise GeoRestrictedError(msg, countries=countries)
1177
1178 def raise_no_formats(self, msg, expected=False, video_id=None):
f2ebc5c7 1179 if expected and (
1180 self.get_param('ignore_no_formats_error') or self.get_param('wait_for_video')):
b7da73eb 1181 self.report_warning(msg, video_id)
68f5867c
L
1182 elif isinstance(msg, ExtractorError):
1183 raise msg
b7da73eb 1184 else:
1185 raise ExtractorError(msg, expected=expected, video_id=video_id)
c430802e 1186
5f6a1245 1187 # Methods for following #608
c0d0b01f 1188 @staticmethod
311b6615 1189 def url_result(url, ie=None, video_id=None, video_title=None, *, url_transparent=False, **kwargs):
10952eb2 1190 """Returns a URL that points to a page that should be processed"""
311b6615 1191 if ie is not None:
1192 kwargs['ie_key'] = ie if isinstance(ie, str) else ie.ie_key()
7012b23c 1193 if video_id is not None:
311b6615 1194 kwargs['id'] = video_id
830d53bf 1195 if video_title is not None:
311b6615 1196 kwargs['title'] = video_title
1197 return {
1198 **kwargs,
1199 '_type': 'url_transparent' if url_transparent else 'url',
1200 'url': url,
1201 }
1202
8f97a15d 1203 @classmethod
1204 def playlist_from_matches(cls, matches, playlist_id=None, playlist_title=None,
1205 getter=IDENTITY, ie=None, video_kwargs=None, **kwargs):
1206 return cls.playlist_result(
1207 (cls.url_result(m, ie, **(video_kwargs or {})) for m in orderedSet(map(getter, matches), lazy=True)),
1208 playlist_id, playlist_title, **kwargs)
46b18f23 1209
c0d0b01f 1210 @staticmethod
311b6615 1211 def playlist_result(entries, playlist_id=None, playlist_title=None, playlist_description=None, *, multi_video=False, **kwargs):
d6983cb4 1212 """Returns a playlist"""
d6983cb4 1213 if playlist_id:
311b6615 1214 kwargs['id'] = playlist_id
d6983cb4 1215 if playlist_title:
311b6615 1216 kwargs['title'] = playlist_title
ecc97af3 1217 if playlist_description is not None:
311b6615 1218 kwargs['description'] = playlist_description
1219 return {
1220 **kwargs,
1221 '_type': 'multi_video' if multi_video else 'playlist',
1222 'entries': entries,
1223 }
d6983cb4 1224
c342041f 1225 def _search_regex(self, pattern, string, name, default=NO_DEFAULT, fatal=True, flags=0, group=None):
d6983cb4
PH
1226 """
1227 Perform a regex search on the given string, using a single or a list of
1228 patterns returning the first matching group.
1229 In case of failure return a default value or raise a WARNING or a
55b3e45b 1230 RegexNotFoundError, depending on fatal, specifying the field name.
d6983cb4 1231 """
61d3665d 1232 if string is None:
1233 mobj = None
77f90330 1234 elif isinstance(pattern, (str, re.Pattern)):
d6983cb4
PH
1235 mobj = re.search(pattern, string, flags)
1236 else:
1237 for p in pattern:
1238 mobj = re.search(p, string, flags)
c3415d1b
PH
1239 if mobj:
1240 break
d6983cb4 1241
ec11a9f4 1242 _name = self._downloader._format_err(name, self._downloader.Styles.EMPHASIS)
d6983cb4
PH
1243
1244 if mobj:
711ede6e
PH
1245 if group is None:
1246 # return the first matching group
1247 return next(g for g in mobj.groups() if g is not None)
198f7ea8 1248 elif isinstance(group, (list, tuple)):
1249 return tuple(mobj.group(g) for g in group)
711ede6e
PH
1250 else:
1251 return mobj.group(group)
c342041f 1252 elif default is not NO_DEFAULT:
d6983cb4
PH
1253 return default
1254 elif fatal:
f1a9d64e 1255 raise RegexNotFoundError('Unable to extract %s' % _name)
d6983cb4 1256 else:
6a39ee13 1257 self.report_warning('unable to extract %s' % _name + bug_reports_message())
d6983cb4
PH
1258 return None
1259
f0bc6e20 1260 def _search_json(self, start_pattern, string, name, video_id, *, end_pattern='',
8b7fb8b6 1261 contains_pattern=r'{(?s:.+)}', fatal=True, default=NO_DEFAULT, **kwargs):
b7c47b74 1262 """Searches string for the JSON object specified by start_pattern"""
1263 # NB: end_pattern is only used to reduce the size of the initial match
f0bc6e20 1264 if default is NO_DEFAULT:
1265 default, has_default = {}, False
1266 else:
1267 fatal, has_default = False, True
1268
1269 json_string = self._search_regex(
8b7fb8b6 1270 rf'(?:{start_pattern})\s*(?P<json>{contains_pattern})\s*(?:{end_pattern})',
f0bc6e20 1271 string, name, group='json', fatal=fatal, default=None if has_default else NO_DEFAULT)
1272 if not json_string:
1273 return default
1274
1275 _name = self._downloader._format_err(name, self._downloader.Styles.EMPHASIS)
1276 try:
1277 return self._parse_json(json_string, video_id, ignore_extra=True, **kwargs)
1278 except ExtractorError as e:
1279 if fatal:
1280 raise ExtractorError(
1281 f'Unable to extract {_name} - Failed to parse JSON', cause=e.cause, video_id=video_id)
1282 elif not has_default:
1283 self.report_warning(
1284 f'Unable to extract {_name} - Failed to parse JSON: {e}', video_id=video_id)
1285 return default
b7c47b74 1286
c342041f 1287 def _html_search_regex(self, pattern, string, name, default=NO_DEFAULT, fatal=True, flags=0, group=None):
d6983cb4
PH
1288 """
1289 Like _search_regex, but strips HTML tags and unescapes entities.
1290 """
711ede6e 1291 res = self._search_regex(pattern, string, name, default, fatal, flags, group)
08e29b9f 1292 if isinstance(res, tuple):
edfc7725 1293 return tuple(map(clean_html, res))
1294 return clean_html(res)
d6983cb4 1295
2118fdd1 1296 def _get_netrc_login_info(self, netrc_machine=None):
2118fdd1
RA
1297 netrc_machine = netrc_machine or self._NETRC_MACHINE
1298
db3ad8a6
ND
1299 cmd = self.get_param('netrc_cmd', '').format(netrc_machine)
1300 if cmd:
1301 self.to_screen(f'Executing command: {cmd}')
1302 stdout, _, ret = Popen.run(cmd, text=True, shell=True, stdout=subprocess.PIPE)
1303 if ret != 0:
1304 raise OSError(f'Command returned error code {ret}')
1305 info = netrc_from_content(stdout).authenticators(netrc_machine)
2118fdd1 1306
db3ad8a6
ND
1307 elif self.get_param('usenetrc', False):
1308 netrc_file = compat_expanduser(self.get_param('netrc_location') or '~')
1309 if os.path.isdir(netrc_file):
1310 netrc_file = os.path.join(netrc_file, '.netrc')
1311 info = netrc.netrc(netrc_file).authenticators(netrc_machine)
1312
1313 else:
1314 return None, None
1315 if not info:
1316 raise netrc.NetrcParseError(f'No authenticators for {netrc_machine}')
1317 return info[0], info[2]
2118fdd1 1318
1b6712ab 1319 def _get_login_info(self, username_option='username', password_option='password', netrc_machine=None):
fc79158d 1320 """
cf0649f8 1321 Get the login info as (username, password)
32443dd3
S
1322 First look for the manually specified credentials using username_option
1323 and password_option as keys in params dictionary. If no such credentials
db3ad8a6
ND
1324 are available try the netrc_cmd if it is defined or look in the
1325 netrc file using the netrc_machine or _NETRC_MACHINE value.
fc79158d
JMF
1326 If there's no info available, return (None, None)
1327 """
fc79158d 1328
a06916d9 1329 username = self.get_param(username_option)
1330 if username is not None:
1331 password = self.get_param(password_option)
2118fdd1 1332 else:
db3ad8a6
ND
1333 try:
1334 username, password = self._get_netrc_login_info(netrc_machine)
1335 except (OSError, netrc.NetrcParseError) as err:
1336 self.report_warning(f'Failed to parse .netrc: {err}')
1337 return None, None
2133565c 1338 return username, password
fc79158d 1339
e64b7569 1340 def _get_tfa_info(self, note='two-factor verification code'):
83317f69 1341 """
1342 Get the two-factor authentication info
1343 TODO - asking the user will be required for sms/phone verify
1344 currently just uses the command line option
1345 If there's no info available, return None
1346 """
83317f69 1347
a06916d9 1348 tfa = self.get_param('twofactor')
1349 if tfa is not None:
1350 return tfa
83317f69 1351
ac668111 1352 return getpass.getpass('Type %s and press [Return]: ' % note)
83317f69 1353
46720279
JMF
1354 # Helper functions for extracting OpenGraph info
1355 @staticmethod
ab2d5247 1356 def _og_regexes(prop):
45b2ee6f 1357 content_re = r'content=(?:"([^"]+?)"|\'([^\']+?)\'|\s*([^\s"\'=<>`]+?)(?=\s|/?>))'
fbfde1c3
F
1358 property_re = (r'(?:name|property)=(?:\'og%(sep)s%(prop)s\'|"og%(sep)s%(prop)s"|\s*og%(sep)s%(prop)s\b)'
1359 % {'prop': re.escape(prop), 'sep': '(?:&#x3A;|[:-])'})
78fb87b2 1360 template = r'<meta[^>]+?%s[^>]+?%s'
ab2d5247 1361 return [
78fb87b2
JMF
1362 template % (property_re, content_re),
1363 template % (content_re, property_re),
ab2d5247 1364 ]
46720279 1365
864f24bd
S
1366 @staticmethod
1367 def _meta_regex(prop):
1368 return r'''(?isx)<meta
8b9848ac 1369 (?=[^>]+(?:itemprop|name|property|id|http-equiv)=(["\']?)%s\1)
864f24bd
S
1370 [^>]+?content=(["\'])(?P<content>.*?)\2''' % re.escape(prop)
1371
3c4e6d83 1372 def _og_search_property(self, prop, html, name=None, **kargs):
6606817a 1373 prop = variadic(prop)
46720279 1374 if name is None:
b070564e
S
1375 name = 'OpenGraph %s' % prop[0]
1376 og_regexes = []
1377 for p in prop:
1378 og_regexes.extend(self._og_regexes(p))
1379 escaped = self._search_regex(og_regexes, html, name, flags=re.DOTALL, **kargs)
eb0a8398
PH
1380 if escaped is None:
1381 return None
1382 return unescapeHTML(escaped)
46720279
JMF
1383
1384 def _og_search_thumbnail(self, html, **kargs):
10952eb2 1385 return self._og_search_property('image', html, 'thumbnail URL', fatal=False, **kargs)
46720279
JMF
1386
1387 def _og_search_description(self, html, **kargs):
1388 return self._og_search_property('description', html, fatal=False, **kargs)
1389
04f3fd2c 1390 def _og_search_title(self, html, *, fatal=False, **kargs):
1391 return self._og_search_property('title', html, fatal=fatal, **kargs)
46720279 1392
8ffa13e0 1393 def _og_search_video_url(self, html, name='video url', secure=True, **kargs):
a3681973
PH
1394 regexes = self._og_regexes('video') + self._og_regexes('video:url')
1395 if secure:
1396 regexes = self._og_regexes('video:secure_url') + regexes
8ffa13e0 1397 return self._html_search_regex(regexes, html, name, **kargs)
46720279 1398
78338f71
JMF
1399 def _og_search_url(self, html, **kargs):
1400 return self._og_search_property('url', html, **kargs)
1401
04f3fd2c 1402 def _html_extract_title(self, html, name='title', *, fatal=False, **kwargs):
21633673 1403 return self._html_search_regex(r'(?s)<title\b[^>]*>([^<]+)</title>', html, name, fatal=fatal, **kwargs)
77cc7c6e 1404
40c696e5 1405 def _html_search_meta(self, name, html, display_name=None, fatal=False, **kwargs):
6606817a 1406 name = variadic(name)
59040888 1407 if display_name is None:
88d9f6c0 1408 display_name = name[0]
59040888 1409 return self._html_search_regex(
88d9f6c0 1410 [self._meta_regex(n) for n in name],
711ede6e 1411 html, display_name, fatal=fatal, group='content', **kwargs)
59040888
PH
1412
1413 def _dc_search_uploader(self, html):
1414 return self._html_search_meta('dc.creator', html, 'uploader')
1415
8f97a15d 1416 @staticmethod
1417 def _rta_search(html):
8dbe9899
PH
1418 # See http://www.rtalabel.org/index.php?content=howtofaq#single
1419 if re.search(r'(?ix)<meta\s+name="rating"\s+'
1420 r' content="RTA-5042-1996-1400-1577-RTA"',
1421 html):
1422 return 18
8f97a15d 1423
1424 # And then there are the jokers who advertise that they use RTA, but actually don't.
1425 AGE_LIMIT_MARKERS = [
1426 r'Proudly Labeled <a href="http://www\.rtalabel\.org/" title="Restricted to Adults">RTA</a>',
32a84bcf
SS
1427 r'>[^<]*you acknowledge you are at least (\d+) years old',
1428 r'>\s*(?:18\s+U(?:\.S\.C\.|SC)\s+)?(?:§+\s*)?2257\b',
8f97a15d 1429 ]
32a84bcf
SS
1430
1431 age_limit = 0
1432 for marker in AGE_LIMIT_MARKERS:
1433 mobj = re.search(marker, html)
1434 if mobj:
1435 age_limit = max(age_limit, int(traverse_obj(mobj, 1, default=18)))
1436 return age_limit
8dbe9899 1437
59040888
PH
1438 def _media_rating_search(self, html):
1439 # See http://www.tjg-designs.com/WP/metadata-code-examples-adding-metadata-to-your-web-pages/
1440 rating = self._html_search_meta('rating', html)
1441
1442 if not rating:
1443 return None
1444
1445 RATING_TABLE = {
1446 'safe for kids': 0,
1447 'general': 8,
1448 '14 years': 14,
1449 'mature': 17,
1450 'restricted': 19,
1451 }
d800609c 1452 return RATING_TABLE.get(rating.lower())
59040888 1453
69319969 1454 def _family_friendly_search(self, html):
6ca7732d 1455 # See http://schema.org/VideoObject
ac8491fc
S
1456 family_friendly = self._html_search_meta(
1457 'isFamilyFriendly', html, default=None)
69319969
NJ
1458
1459 if not family_friendly:
1460 return None
1461
1462 RATING_TABLE = {
1463 '1': 0,
1464 'true': 0,
1465 '0': 18,
1466 'false': 18,
1467 }
d800609c 1468 return RATING_TABLE.get(family_friendly.lower())
69319969 1469
0c708f11
JMF
1470 def _twitter_search_player(self, html):
1471 return self._html_search_meta('twitter:player', html,
9e1a5b84 1472 'twitter card player')
0c708f11 1473
0c36dc00 1474 def _yield_json_ld(self, html, video_id, *, fatal=True, default=NO_DEFAULT):
1475 """Yield all json ld objects in the html"""
1476 if default is not NO_DEFAULT:
1477 fatal = False
1478 for mobj in re.finditer(JSON_LD_RE, html):
1479 json_ld_item = self._parse_json(mobj.group('json_ld'), video_id, fatal=fatal)
1480 for json_ld in variadic(json_ld_item):
1481 if isinstance(json_ld, dict):
1482 yield json_ld
1483
1484 def _search_json_ld(self, html, video_id, expected_type=None, *, fatal=True, default=NO_DEFAULT):
1485 """Search for a video in any json ld in the html"""
1486 if default is not NO_DEFAULT:
1487 fatal = False
1488 info = self._json_ld(
1489 list(self._yield_json_ld(html, video_id, fatal=fatal, default=default)),
1490 video_id, fatal=fatal, expected_type=expected_type)
1491 if info:
1492 return info
4433bb02
S
1493 if default is not NO_DEFAULT:
1494 return default
1495 elif fatal:
1496 raise RegexNotFoundError('Unable to extract JSON-LD')
1497 else:
6a39ee13 1498 self.report_warning('unable to extract JSON-LD %s' % bug_reports_message())
4433bb02 1499 return {}
4ca2a3cf 1500
95b31e26 1501 def _json_ld(self, json_ld, video_id, fatal=True, expected_type=None):
14f25df2 1502 if isinstance(json_ld, str):
4ca2a3cf
S
1503 json_ld = self._parse_json(json_ld, video_id, fatal=fatal)
1504 if not json_ld:
1505 return {}
1506 info = {}
bae14048 1507
e7e4a6e0
S
1508 INTERACTION_TYPE_MAP = {
1509 'CommentAction': 'comment',
1510 'AgreeAction': 'like',
1511 'DisagreeAction': 'dislike',
1512 'LikeAction': 'like',
1513 'DislikeAction': 'dislike',
1514 'ListenAction': 'view',
1515 'WatchAction': 'view',
1516 'ViewAction': 'view',
1517 }
1518
f3c0c773 1519 def is_type(e, *expected_types):
1520 type = variadic(traverse_obj(e, '@type'))
1521 return any(x in type for x in expected_types)
1522
29f7c58a 1523 def extract_interaction_type(e):
1524 interaction_type = e.get('interactionType')
1525 if isinstance(interaction_type, dict):
1526 interaction_type = interaction_type.get('@type')
1527 return str_or_none(interaction_type)
1528
e7e4a6e0
S
1529 def extract_interaction_statistic(e):
1530 interaction_statistic = e.get('interactionStatistic')
29f7c58a 1531 if isinstance(interaction_statistic, dict):
1532 interaction_statistic = [interaction_statistic]
e7e4a6e0
S
1533 if not isinstance(interaction_statistic, list):
1534 return
1535 for is_e in interaction_statistic:
f3c0c773 1536 if not is_type(is_e, 'InteractionCounter'):
e7e4a6e0 1537 continue
29f7c58a 1538 interaction_type = extract_interaction_type(is_e)
1539 if not interaction_type:
e7e4a6e0 1540 continue
ce5b9040
S
1541 # For interaction count some sites provide string instead of
1542 # an integer (as per spec) with non digit characters (e.g. ",")
1543 # so extracting count with more relaxed str_to_int
1544 interaction_count = str_to_int(is_e.get('userInteractionCount'))
e7e4a6e0
S
1545 if interaction_count is None:
1546 continue
1547 count_kind = INTERACTION_TYPE_MAP.get(interaction_type.split('/')[-1])
1548 if not count_kind:
1549 continue
1550 count_key = '%s_count' % count_kind
1551 if info.get(count_key) is not None:
1552 continue
1553 info[count_key] = interaction_count
1554
f5225737 1555 def extract_chapter_information(e):
1556 chapters = [{
1557 'title': part.get('name'),
1558 'start_time': part.get('startOffset'),
1559 'end_time': part.get('endOffset'),
85553414 1560 } for part in variadic(e.get('hasPart') or []) if part.get('@type') == 'Clip']
f5225737 1561 for idx, (last_c, current_c, next_c) in enumerate(zip(
1562 [{'end_time': 0}] + chapters, chapters, chapters[1:])):
1563 current_c['end_time'] = current_c['end_time'] or next_c['start_time']
1564 current_c['start_time'] = current_c['start_time'] or last_c['end_time']
1565 if None in current_c.values():
1566 self.report_warning(f'Chapter {idx} contains broken data. Not extracting chapters')
1567 return
1568 if chapters:
1569 chapters[-1]['end_time'] = chapters[-1]['end_time'] or info['duration']
1570 info['chapters'] = chapters
1571
bae14048 1572 def extract_video_object(e):
f7ad7160 1573 author = e.get('author')
bae14048 1574 info.update({
0c36dc00 1575 'url': url_or_none(e.get('contentUrl')),
0f60ba6e 1576 'ext': mimetype2ext(e.get('encodingFormat')),
bae14048
S
1577 'title': unescapeHTML(e.get('name')),
1578 'description': unescapeHTML(e.get('description')),
eb2333bc 1579 'thumbnails': [{'url': unescapeHTML(url)}
21633673 1580 for url in variadic(traverse_obj(e, 'thumbnailUrl', 'thumbnailURL'))
1581 if url_or_none(url)],
bae14048
S
1582 'duration': parse_duration(e.get('duration')),
1583 'timestamp': unified_timestamp(e.get('uploadDate')),
f7ad7160 1584 # author can be an instance of 'Organization' or 'Person' types.
1585 # both types can have 'name' property(inherited from 'Thing' type). [1]
1586 # however some websites are using 'Text' type instead.
1587 # 1. https://schema.org/VideoObject
14f25df2 1588 'uploader': author.get('name') if isinstance(author, dict) else author if isinstance(author, str) else None,
0f60ba6e 1589 'artist': traverse_obj(e, ('byArtist', 'name'), expected_type=str),
56ba69e4 1590 'filesize': int_or_none(float_or_none(e.get('contentSize'))),
bae14048
S
1591 'tbr': int_or_none(e.get('bitrate')),
1592 'width': int_or_none(e.get('width')),
1593 'height': int_or_none(e.get('height')),
33a81c2c 1594 'view_count': int_or_none(e.get('interactionCount')),
0f60ba6e 1595 'tags': try_call(lambda: e.get('keywords').split(',')),
bae14048 1596 })
0f60ba6e 1597 if is_type(e, 'AudioObject'):
1598 info.update({
1599 'vcodec': 'none',
1600 'abr': int_or_none(e.get('bitrate')),
1601 })
e7e4a6e0 1602 extract_interaction_statistic(e)
f5225737 1603 extract_chapter_information(e)
bae14048 1604
d5c32548 1605 def traverse_json_ld(json_ld, at_top_level=True):
1d55ebab
SS
1606 for e in variadic(json_ld):
1607 if not isinstance(e, dict):
1608 continue
d5c32548
ZM
1609 if at_top_level and '@context' not in e:
1610 continue
1611 if at_top_level and set(e.keys()) == {'@context', '@graph'}:
1d55ebab 1612 traverse_json_ld(e['@graph'], at_top_level=False)
c13a301a 1613 continue
f3c0c773 1614 if expected_type is not None and not is_type(e, expected_type):
4433bb02 1615 continue
8f122fa0 1616 rating = traverse_obj(e, ('aggregateRating', 'ratingValue'), expected_type=float_or_none)
1617 if rating is not None:
1618 info['average_rating'] = rating
f3c0c773 1619 if is_type(e, 'TVEpisode', 'Episode'):
440863ad 1620 episode_name = unescapeHTML(e.get('name'))
46933a15 1621 info.update({
440863ad 1622 'episode': episode_name,
46933a15
S
1623 'episode_number': int_or_none(e.get('episodeNumber')),
1624 'description': unescapeHTML(e.get('description')),
1625 })
440863ad
S
1626 if not info.get('title') and episode_name:
1627 info['title'] = episode_name
46933a15 1628 part_of_season = e.get('partOfSeason')
f3c0c773 1629 if is_type(part_of_season, 'TVSeason', 'Season', 'CreativeWorkSeason'):
458fd30f
S
1630 info.update({
1631 'season': unescapeHTML(part_of_season.get('name')),
1632 'season_number': int_or_none(part_of_season.get('seasonNumber')),
1633 })
d16b3c66 1634 part_of_series = e.get('partOfSeries') or e.get('partOfTVSeries')
f3c0c773 1635 if is_type(part_of_series, 'TVSeries', 'Series', 'CreativeWorkSeries'):
46933a15 1636 info['series'] = unescapeHTML(part_of_series.get('name'))
f3c0c773 1637 elif is_type(e, 'Movie'):
391256dc
S
1638 info.update({
1639 'title': unescapeHTML(e.get('name')),
1640 'description': unescapeHTML(e.get('description')),
1641 'duration': parse_duration(e.get('duration')),
1642 'timestamp': unified_timestamp(e.get('dateCreated')),
1643 })
f3c0c773 1644 elif is_type(e, 'Article', 'NewsArticle'):
46933a15
S
1645 info.update({
1646 'timestamp': parse_iso8601(e.get('datePublished')),
1647 'title': unescapeHTML(e.get('headline')),
d5c32548 1648 'description': unescapeHTML(e.get('articleBody') or e.get('description')),
46933a15 1649 })
f3c0c773 1650 if is_type(traverse_obj(e, ('video', 0)), 'VideoObject'):
2edb38e8 1651 extract_video_object(e['video'][0])
f3c0c773 1652 elif is_type(traverse_obj(e, ('subjectOf', 0)), 'VideoObject'):
e50c3500 1653 extract_video_object(e['subjectOf'][0])
0f60ba6e 1654 elif is_type(e, 'VideoObject', 'AudioObject'):
bae14048 1655 extract_video_object(e)
4433bb02
S
1656 if expected_type is None:
1657 continue
1658 else:
1659 break
c69701c6 1660 video = e.get('video')
f3c0c773 1661 if is_type(video, 'VideoObject'):
c69701c6 1662 extract_video_object(video)
4433bb02
S
1663 if expected_type is None:
1664 continue
1665 else:
1666 break
d5c32548 1667
1d55ebab 1668 traverse_json_ld(json_ld)
90137ca4 1669 return filter_dict(info)
4ca2a3cf 1670
135dfa2c 1671 def _search_nextjs_data(self, webpage, video_id, *, transform_source=None, fatal=True, **kw):
f98709af
LL
1672 return self._parse_json(
1673 self._search_regex(
1674 r'(?s)<script[^>]+id=[\'"]__NEXT_DATA__[\'"][^>]*>([^<]+)</script>',
135dfa2c 1675 webpage, 'next.js data', fatal=fatal, **kw),
1676 video_id, transform_source=transform_source, fatal=fatal)
f98709af 1677
8072ef2b 1678 def _search_nuxt_data(self, webpage, video_id, context_name='__NUXT__', *, fatal=True, traverse=('data', 0)):
1679 """Parses Nuxt.js metadata. This works as long as the function __NUXT__ invokes is a pure function"""
66f4c04e 1680 rectx = re.escape(context_name)
8072ef2b 1681 FUNCTION_RE = r'\(function\((?P<arg_keys>.*?)\){return\s+(?P<js>{.*?})\s*;?\s*}\((?P<arg_vals>.*?)\)'
66f4c04e 1682 js, arg_keys, arg_vals = self._search_regex(
8072ef2b 1683 (rf'<script>\s*window\.{rectx}={FUNCTION_RE}\s*\)\s*;?\s*</script>', rf'{rectx}\(.*?{FUNCTION_RE}'),
f7fc8d39 1684 webpage, context_name, group=('js', 'arg_keys', 'arg_vals'),
1685 default=NO_DEFAULT if fatal else (None, None, None))
1686 if js is None:
1687 return {}
66f4c04e 1688
b23167e7
L
1689 args = dict(zip(arg_keys.split(','), map(json.dumps, self._parse_json(
1690 f'[{arg_vals}]', video_id, transform_source=js_to_json, fatal=fatal) or ())))
66f4c04e 1691
8072ef2b 1692 ret = self._parse_json(js, video_id, transform_source=functools.partial(js_to_json, vars=args), fatal=fatal)
1693 return traverse_obj(ret, traverse) or {}
66f4c04e 1694
27713812 1695 @staticmethod
f8da79f8 1696 def _hidden_inputs(html):
586f1cc5 1697 html = re.sub(r'<!--(?:(?!<!--).)*-->', '', html)
201ea3ee 1698 hidden_inputs = {}
c8498368
S
1699 for input in re.findall(r'(?i)(<input[^>]+>)', html):
1700 attrs = extract_attributes(input)
1701 if not input:
201ea3ee 1702 continue
c8498368 1703 if attrs.get('type') not in ('hidden', 'submit'):
201ea3ee 1704 continue
c8498368
S
1705 name = attrs.get('name') or attrs.get('id')
1706 value = attrs.get('value')
1707 if name and value is not None:
1708 hidden_inputs[name] = value
201ea3ee 1709 return hidden_inputs
27713812 1710
cf61d96d
S
1711 def _form_hidden_inputs(self, form_id, html):
1712 form = self._search_regex(
73eb13df 1713 r'(?is)<form[^>]+?id=(["\'])%s\1[^>]*>(?P<form>.+?)</form>' % form_id,
cf61d96d
S
1714 html, '%s form' % form_id, group='form')
1715 return self._hidden_inputs(form)
1716
d0d74b71 1717 @classproperty(cache=True)
1718 def FormatSort(cls):
1719 class FormatSort(FormatSorter):
1720 def __init__(ie, *args, **kwargs):
1721 super().__init__(ie._downloader, *args, **kwargs)
eb8a4433 1722
d0d74b71 1723 deprecation_warning(
1724 'yt_dlp.InfoExtractor.FormatSort is deprecated and may be removed in the future. '
1725 'Use yt_dlp.utils.FormatSorter instead')
1726 return FormatSort
eb8a4433 1727
1728 def _sort_formats(self, formats, field_preference=[]):
9f14daf2 1729 if not field_preference:
1730 self._downloader.deprecation_warning(
1731 'yt_dlp.InfoExtractor._sort_formats is deprecated and is no longer required')
1732 return
1733 self._downloader.deprecation_warning(
1734 'yt_dlp.InfoExtractor._sort_formats is deprecated and no longer works as expected. '
1735 'Return _format_sort_fields in the info_dict instead')
1736 if formats:
784320c9 1737 formats[0]['__sort_fields'] = field_preference
59040888 1738
96a53167
S
1739 def _check_formats(self, formats, video_id):
1740 if formats:
1741 formats[:] = filter(
1742 lambda f: self._is_valid_url(
1743 f['url'], video_id,
1744 item='%s video format' % f.get('format_id') if f.get('format_id') else 'video'),
1745 formats)
1746
f5bdb444
S
1747 @staticmethod
1748 def _remove_duplicate_formats(formats):
1749 format_urls = set()
1750 unique_formats = []
1751 for f in formats:
1752 if f['url'] not in format_urls:
1753 format_urls.add(f['url'])
1754 unique_formats.append(f)
1755 formats[:] = unique_formats
1756
45024183 1757 def _is_valid_url(self, url, video_id, item='video', headers={}):
2f0f6578
S
1758 url = self._proto_relative_url(url, scheme='http:')
1759 # For now assume non HTTP(S) URLs always valid
1760 if not (url.startswith('http://') or url.startswith('https://')):
1761 return True
96a53167 1762 try:
45024183 1763 self._request_webpage(url, video_id, 'Checking %s URL' % item, headers=headers)
96a53167 1764 return True
8bdd16b4 1765 except ExtractorError as e:
25e911a9 1766 self.to_screen(
8bdd16b4 1767 '%s: %s URL is invalid, skipping: %s'
1768 % (video_id, item, error_to_compat_str(e.cause)))
25e911a9 1769 return False
96a53167 1770
20991253 1771 def http_scheme(self):
1ede5b24 1772 """ Either "http:" or "https:", depending on the user's preferences """
20991253
PH
1773 return (
1774 'http:'
a06916d9 1775 if self.get_param('prefer_insecure', False)
20991253
PH
1776 else 'https:')
1777
57c7411f 1778 def _proto_relative_url(self, url, scheme=None):
8f97a15d 1779 scheme = scheme or self.http_scheme()
1780 assert scheme.endswith(':')
1781 return sanitize_url(url, scheme=scheme[:-1])
57c7411f 1782
4094b6e3
PH
1783 def _sleep(self, timeout, video_id, msg_template=None):
1784 if msg_template is None:
f1a9d64e 1785 msg_template = '%(video_id)s: Waiting for %(timeout)s seconds'
4094b6e3
PH
1786 msg = msg_template % {'video_id': video_id, 'timeout': timeout}
1787 self.to_screen(msg)
1788 time.sleep(timeout)
1789
f983b875 1790 def _extract_f4m_formats(self, manifest_url, video_id, preference=None, quality=None, f4m_id=None,
4de61310 1791 transform_source=lambda s: fix_xml_ampersands(s).strip(),
7360c06f 1792 fatal=True, m3u8_id=None, data=None, headers={}, query={}):
0b5546c7 1793 if self.get_param('ignore_no_formats_error'):
1794 fatal = False
1795
a076c1f9 1796 res = self._download_xml_handle(
f036a632 1797 manifest_url, video_id, 'Downloading f4m manifest',
97f4aecf
S
1798 'Unable to download f4m manifest',
1799 # Some manifests may be malformed, e.g. prosiebensat1 generated manifests
067aa17e 1800 # (see https://github.com/ytdl-org/youtube-dl/issues/6215#issuecomment-121704244)
4de61310 1801 transform_source=transform_source,
7360c06f 1802 fatal=fatal, data=data, headers=headers, query=query)
a076c1f9 1803 if res is False:
8d29e47f 1804 return []
31bb8d3f 1805
a076c1f9
E
1806 manifest, urlh = res
1807 manifest_url = urlh.geturl()
1808
0fdbb332 1809 return self._parse_f4m_formats(
f983b875 1810 manifest, manifest_url, video_id, preference=preference, quality=quality, f4m_id=f4m_id,
448bb5f3 1811 transform_source=transform_source, fatal=fatal, m3u8_id=m3u8_id)
0fdbb332 1812
f983b875 1813 def _parse_f4m_formats(self, manifest, manifest_url, video_id, preference=None, quality=None, f4m_id=None,
0fdbb332 1814 transform_source=lambda s: fix_xml_ampersands(s).strip(),
448bb5f3 1815 fatal=True, m3u8_id=None):
f9934b96 1816 if not isinstance(manifest, xml.etree.ElementTree.Element) and not fatal:
d9eb580a
S
1817 return []
1818
7a5c1cfe 1819 # currently yt-dlp cannot decode the playerVerificationChallenge as Akamai uses Adobe Alchemy
fb72ec58 1820 akamai_pv = manifest.find('{http://ns.adobe.com/f4m/1.0}pv-2.0')
1821 if akamai_pv is not None and ';' in akamai_pv.text:
1822 playerVerificationChallenge = akamai_pv.text.split(';')[0]
1823 if playerVerificationChallenge.strip() != '':
1824 return []
1825
31bb8d3f 1826 formats = []
7a47d07c 1827 manifest_version = '1.0'
b2527359 1828 media_nodes = manifest.findall('{http://ns.adobe.com/f4m/1.0}media')
34e48bed 1829 if not media_nodes:
7a47d07c 1830 manifest_version = '2.0'
34e48bed 1831 media_nodes = manifest.findall('{http://ns.adobe.com/f4m/2.0}media')
b22ca762 1832 # Remove unsupported DRM protected media from final formats
067aa17e 1833 # rendition (see https://github.com/ytdl-org/youtube-dl/issues/8573).
b22ca762
S
1834 media_nodes = remove_encrypted_media(media_nodes)
1835 if not media_nodes:
1836 return formats
48107c19
S
1837
1838 manifest_base_url = get_base_url(manifest)
0a5685b2 1839
a6571f10 1840 bootstrap_info = xpath_element(
0a5685b2
YCH
1841 manifest, ['{http://ns.adobe.com/f4m/1.0}bootstrapInfo', '{http://ns.adobe.com/f4m/2.0}bootstrapInfo'],
1842 'bootstrap info', default=None)
1843
edd6074c
RA
1844 vcodec = None
1845 mime_type = xpath_text(
1846 manifest, ['{http://ns.adobe.com/f4m/1.0}mimeType', '{http://ns.adobe.com/f4m/2.0}mimeType'],
1847 'base URL', default=None)
1848 if mime_type and mime_type.startswith('audio/'):
1849 vcodec = 'none'
1850
b2527359 1851 for i, media_el in enumerate(media_nodes):
77b8b4e6
S
1852 tbr = int_or_none(media_el.attrib.get('bitrate'))
1853 width = int_or_none(media_el.attrib.get('width'))
1854 height = int_or_none(media_el.attrib.get('height'))
34921b43 1855 format_id = join_nonempty(f4m_id, tbr or i)
448bb5f3
YCH
1856 # If <bootstrapInfo> is present, the specified f4m is a
1857 # stream-level manifest, and only set-level manifests may refer to
1858 # external resources. See section 11.4 and section 4 of F4M spec
1859 if bootstrap_info is None:
1860 media_url = None
1861 # @href is introduced in 2.0, see section 11.6 of F4M spec
1862 if manifest_version == '2.0':
1863 media_url = media_el.attrib.get('href')
1864 if media_url is None:
1865 media_url = media_el.attrib.get('url')
31c746e5
S
1866 if not media_url:
1867 continue
cc357c4d
S
1868 manifest_url = (
1869 media_url if media_url.startswith('http://') or media_url.startswith('https://')
48107c19 1870 else ((manifest_base_url or '/'.join(manifest_url.split('/')[:-1])) + '/' + media_url))
70f0f5a8
S
1871 # If media_url is itself a f4m manifest do the recursive extraction
1872 # since bitrates in parent manifest (this one) and media_url manifest
1873 # may differ leading to inability to resolve the format by requested
1874 # bitrate in f4m downloader
240b6045
YCH
1875 ext = determine_ext(manifest_url)
1876 if ext == 'f4m':
77b8b4e6 1877 f4m_formats = self._extract_f4m_formats(
f983b875 1878 manifest_url, video_id, preference=preference, quality=quality, f4m_id=f4m_id,
77b8b4e6
S
1879 transform_source=transform_source, fatal=fatal)
1880 # Sometimes stream-level manifest contains single media entry that
1881 # does not contain any quality metadata (e.g. http://matchtv.ru/#live-player).
1882 # At the same time parent's media entry in set-level manifest may
1883 # contain it. We will copy it from parent in such cases.
1884 if len(f4m_formats) == 1:
1885 f = f4m_formats[0]
1886 f.update({
1887 'tbr': f.get('tbr') or tbr,
1888 'width': f.get('width') or width,
1889 'height': f.get('height') or height,
1890 'format_id': f.get('format_id') if not tbr else format_id,
edd6074c 1891 'vcodec': vcodec,
77b8b4e6
S
1892 })
1893 formats.extend(f4m_formats)
70f0f5a8 1894 continue
240b6045
YCH
1895 elif ext == 'm3u8':
1896 formats.extend(self._extract_m3u8_formats(
1897 manifest_url, video_id, 'mp4', preference=preference,
f983b875 1898 quality=quality, m3u8_id=m3u8_id, fatal=fatal))
240b6045 1899 continue
31bb8d3f 1900 formats.append({
77b8b4e6 1901 'format_id': format_id,
31bb8d3f 1902 'url': manifest_url,
30d0b549 1903 'manifest_url': manifest_url,
a6571f10 1904 'ext': 'flv' if bootstrap_info is not None else None,
187ee66c 1905 'protocol': 'f4m',
b2527359 1906 'tbr': tbr,
77b8b4e6
S
1907 'width': width,
1908 'height': height,
edd6074c 1909 'vcodec': vcodec,
60ca389c 1910 'preference': preference,
f983b875 1911 'quality': quality,
31bb8d3f 1912 })
31bb8d3f
JMF
1913 return formats
1914
f983b875 1915 def _m3u8_meta_format(self, m3u8_url, ext=None, preference=None, quality=None, m3u8_id=None):
16da9bbc 1916 return {
34921b43 1917 'format_id': join_nonempty(m3u8_id, 'meta'),
704df56d
PH
1918 'url': m3u8_url,
1919 'ext': ext,
1920 'protocol': 'm3u8',
37768f92 1921 'preference': preference - 100 if preference else -100,
f983b875 1922 'quality': quality,
704df56d
PH
1923 'resolution': 'multiple',
1924 'format_note': 'Quality selection URL',
16da9bbc
YCH
1925 }
1926
b5ae35ee 1927 def _report_ignoring_subs(self, name):
1928 self.report_warning(bug_reports_message(
1929 f'Ignoring subtitle tracks found in the {name} manifest; '
1930 'if any subtitle tracks are missing,'
1931 ), only_once=True)
1932
a0c3b2d5
F
1933 def _extract_m3u8_formats(self, *args, **kwargs):
1934 fmts, subs = self._extract_m3u8_formats_and_subtitles(*args, **kwargs)
1935 if subs:
b5ae35ee 1936 self._report_ignoring_subs('HLS')
a0c3b2d5
F
1937 return fmts
1938
1939 def _extract_m3u8_formats_and_subtitles(
177877c5 1940 self, m3u8_url, video_id, ext=None, entry_protocol='m3u8_native',
a0c3b2d5
F
1941 preference=None, quality=None, m3u8_id=None, note=None,
1942 errnote=None, fatal=True, live=False, data=None, headers={},
1943 query={}):
1944
0b5546c7 1945 if self.get_param('ignore_no_formats_error'):
1946 fatal = False
1947
71df9b7f 1948 if not m3u8_url:
1949 if errnote is not False:
1950 errnote = errnote or 'Failed to obtain m3u8 URL'
1951 if fatal:
1952 raise ExtractorError(errnote, video_id=video_id)
1953 self.report_warning(f'{errnote}{bug_reports_message()}')
1954 return [], {}
1955
dbd82a1d 1956 res = self._download_webpage_handle(
81515ad9 1957 m3u8_url, video_id,
37a3bb66 1958 note='Downloading m3u8 information' if note is None else note,
1959 errnote='Failed to download m3u8 information' if errnote is None else errnote,
7360c06f 1960 fatal=fatal, data=data, headers=headers, query=query)
cb252080 1961
dbd82a1d 1962 if res is False:
a0c3b2d5 1963 return [], {}
cb252080 1964
dbd82a1d 1965 m3u8_doc, urlh = res
37113045 1966 m3u8_url = urlh.geturl()
9cdffeeb 1967
a0c3b2d5 1968 return self._parse_m3u8_formats_and_subtitles(
cb252080 1969 m3u8_doc, m3u8_url, ext=ext, entry_protocol=entry_protocol,
310c2ed2 1970 preference=preference, quality=quality, m3u8_id=m3u8_id,
1971 note=note, errnote=errnote, fatal=fatal, live=live, data=data,
1972 headers=headers, query=query, video_id=video_id)
cb252080 1973
a0c3b2d5 1974 def _parse_m3u8_formats_and_subtitles(
42676437 1975 self, m3u8_doc, m3u8_url=None, ext=None, entry_protocol='m3u8_native',
a0c3b2d5
F
1976 preference=None, quality=None, m3u8_id=None, live=False, note=None,
1977 errnote=None, fatal=True, data=None, headers={}, query={},
1978 video_id=None):
60755938 1979 formats, subtitles = [], {}
a0c3b2d5 1980
6b993ca7 1981 has_drm = re.search('|'.join([
1982 r'#EXT-X-FAXS-CM:', # Adobe Flash Access
1983 r'#EXT-X-(?:SESSION-)?KEY:.*?URI="skd://', # Apple FairPlay
1984 ]), m3u8_doc)
a0c3b2d5 1985
60755938 1986 def format_url(url):
14f25df2 1987 return url if re.match(r'^https?://', url) else urllib.parse.urljoin(m3u8_url, url)
60755938 1988
1989 if self.get_param('hls_split_discontinuity', False):
1990 def _extract_m3u8_playlist_indices(manifest_url=None, m3u8_doc=None):
1991 if not m3u8_doc:
1992 if not manifest_url:
1993 return []
1994 m3u8_doc = self._download_webpage(
1995 manifest_url, video_id, fatal=fatal, data=data, headers=headers,
1996 note=False, errnote='Failed to download m3u8 playlist information')
1997 if m3u8_doc is False:
1998 return []
1999 return range(1 + sum(line.startswith('#EXT-X-DISCONTINUITY') for line in m3u8_doc.splitlines()))
0def7587 2000
60755938 2001 else:
2002 def _extract_m3u8_playlist_indices(*args, **kwargs):
2003 return [None]
310c2ed2 2004
cb252080
S
2005 # References:
2006 # 1. https://tools.ietf.org/html/draft-pantos-http-live-streaming-21
067aa17e
S
2007 # 2. https://github.com/ytdl-org/youtube-dl/issues/12211
2008 # 3. https://github.com/ytdl-org/youtube-dl/issues/18923
cb252080
S
2009
2010 # We should try extracting formats only from master playlists [1, 4.3.4],
2011 # i.e. playlists that describe available qualities. On the other hand
2012 # media playlists [1, 4.3.3] should be returned as is since they contain
2013 # just the media without qualities renditions.
9cdffeeb 2014 # Fortunately, master playlist can be easily distinguished from media
cb252080 2015 # playlist based on particular tags availability. As of [1, 4.3.3, 4.3.4]
a0566bbf 2016 # master playlist tags MUST NOT appear in a media playlist and vice versa.
cb252080
S
2017 # As of [1, 4.3.3.1] #EXT-X-TARGETDURATION tag is REQUIRED for every
2018 # media playlist and MUST NOT appear in master playlist thus we can
2019 # clearly detect media playlist with this criterion.
2020
9cdffeeb 2021 if '#EXT-X-TARGETDURATION' in m3u8_doc: # media playlist, return as is
60755938 2022 formats = [{
34921b43 2023 'format_id': join_nonempty(m3u8_id, idx),
60755938 2024 'format_index': idx,
42676437 2025 'url': m3u8_url or encode_data_uri(m3u8_doc.encode('utf-8'), 'application/x-mpegurl'),
60755938 2026 'ext': ext,
2027 'protocol': entry_protocol,
2028 'preference': preference,
2029 'quality': quality,
88acdbc2 2030 'has_drm': has_drm,
60755938 2031 } for idx in _extract_m3u8_playlist_indices(m3u8_doc=m3u8_doc)]
310c2ed2 2032
a0c3b2d5 2033 return formats, subtitles
cb252080
S
2034
2035 groups = {}
2036 last_stream_inf = {}
2037
2038 def extract_media(x_media_line):
2039 media = parse_m3u8_attributes(x_media_line)
2040 # As per [1, 4.3.4.1] TYPE, GROUP-ID and NAME are REQUIRED
2041 media_type, group_id, name = media.get('TYPE'), media.get('GROUP-ID'), media.get('NAME')
2042 if not (media_type and group_id and name):
2043 return
2044 groups.setdefault(group_id, []).append(media)
a0c3b2d5
F
2045 # <https://tools.ietf.org/html/rfc8216#section-4.3.4.1>
2046 if media_type == 'SUBTITLES':
3907333c 2047 # According to RFC 8216 §4.3.4.2.1, URI is REQUIRED in the
2048 # EXT-X-MEDIA tag if the media type is SUBTITLES.
2049 # However, lack of URI has been spotted in the wild.
2050 # e.g. NebulaIE; see https://github.com/yt-dlp/yt-dlp/issues/339
2051 if not media.get('URI'):
2052 return
a0c3b2d5
F
2053 url = format_url(media['URI'])
2054 sub_info = {
2055 'url': url,
2056 'ext': determine_ext(url),
2057 }
4a2f19ab
F
2058 if sub_info['ext'] == 'm3u8':
2059 # Per RFC 8216 §3.1, the only possible subtitle format m3u8
2060 # files may contain is WebVTT:
2061 # <https://tools.ietf.org/html/rfc8216#section-3.1>
2062 sub_info['ext'] = 'vtt'
2063 sub_info['protocol'] = 'm3u8_native'
37a3bb66 2064 lang = media.get('LANGUAGE') or 'und'
a0c3b2d5 2065 subtitles.setdefault(lang, []).append(sub_info)
cb252080
S
2066 if media_type not in ('VIDEO', 'AUDIO'):
2067 return
2068 media_url = media.get('URI')
2069 if media_url:
310c2ed2 2070 manifest_url = format_url(media_url)
60755938 2071 formats.extend({
34921b43 2072 'format_id': join_nonempty(m3u8_id, group_id, name, idx),
60755938 2073 'format_note': name,
2074 'format_index': idx,
2075 'url': manifest_url,
2076 'manifest_url': m3u8_url,
2077 'language': media.get('LANGUAGE'),
2078 'ext': ext,
2079 'protocol': entry_protocol,
2080 'preference': preference,
2081 'quality': quality,
43a3eaf9 2082 'has_drm': has_drm,
60755938 2083 'vcodec': 'none' if media_type == 'AUDIO' else None,
2084 } for idx in _extract_m3u8_playlist_indices(manifest_url))
cb252080
S
2085
2086 def build_stream_name():
2087 # Despite specification does not mention NAME attribute for
3019cb0c
S
2088 # EXT-X-STREAM-INF tag it still sometimes may be present (see [1]
2089 # or vidio test in TestInfoExtractor.test_parse_m3u8_formats)
ddd258f9 2090 # 1. http://www.vidio.com/watch/165683-dj_ambred-booyah-live-2015
cb252080
S
2091 stream_name = last_stream_inf.get('NAME')
2092 if stream_name:
2093 return stream_name
2094 # If there is no NAME in EXT-X-STREAM-INF it will be obtained
2095 # from corresponding rendition group
2096 stream_group_id = last_stream_inf.get('VIDEO')
2097 if not stream_group_id:
2098 return
2099 stream_group = groups.get(stream_group_id)
2100 if not stream_group:
2101 return stream_group_id
2102 rendition = stream_group[0]
2103 return rendition.get('NAME') or stream_group_id
2104
379306ef 2105 # parse EXT-X-MEDIA tags before EXT-X-STREAM-INF in order to have the
2bfc1d9d
RA
2106 # chance to detect video only formats when EXT-X-STREAM-INF tags
2107 # precede EXT-X-MEDIA tags in HLS manifest such as [3].
2108 for line in m3u8_doc.splitlines():
2109 if line.startswith('#EXT-X-MEDIA:'):
2110 extract_media(line)
2111
704df56d
PH
2112 for line in m3u8_doc.splitlines():
2113 if line.startswith('#EXT-X-STREAM-INF:'):
cb252080 2114 last_stream_inf = parse_m3u8_attributes(line)
704df56d
PH
2115 elif line.startswith('#') or not line.strip():
2116 continue
2117 else:
9c99bef7 2118 tbr = float_or_none(
3089bc74
S
2119 last_stream_inf.get('AVERAGE-BANDWIDTH')
2120 or last_stream_inf.get('BANDWIDTH'), scale=1000)
30d0b549 2121 manifest_url = format_url(line.strip())
5ef62fc4 2122
60755938 2123 for idx in _extract_m3u8_playlist_indices(manifest_url):
2124 format_id = [m3u8_id, None, idx]
310c2ed2 2125 # Bandwidth of live streams may differ over time thus making
2126 # format_id unpredictable. So it's better to keep provided
2127 # format_id intact.
2128 if not live:
60755938 2129 stream_name = build_stream_name()
34921b43 2130 format_id[1] = stream_name or '%d' % (tbr or len(formats))
310c2ed2 2131 f = {
34921b43 2132 'format_id': join_nonempty(*format_id),
60755938 2133 'format_index': idx,
310c2ed2 2134 'url': manifest_url,
2135 'manifest_url': m3u8_url,
2136 'tbr': tbr,
2137 'ext': ext,
2138 'fps': float_or_none(last_stream_inf.get('FRAME-RATE')),
2139 'protocol': entry_protocol,
2140 'preference': preference,
2141 'quality': quality,
43a3eaf9 2142 'has_drm': has_drm,
310c2ed2 2143 }
2144 resolution = last_stream_inf.get('RESOLUTION')
2145 if resolution:
2146 mobj = re.search(r'(?P<width>\d+)[xX](?P<height>\d+)', resolution)
2147 if mobj:
2148 f['width'] = int(mobj.group('width'))
2149 f['height'] = int(mobj.group('height'))
2150 # Unified Streaming Platform
2151 mobj = re.search(
2152 r'audio.*?(?:%3D|=)(\d+)(?:-video.*?(?:%3D|=)(\d+))?', f['url'])
2153 if mobj:
2154 abr, vbr = mobj.groups()
2155 abr, vbr = float_or_none(abr, 1000), float_or_none(vbr, 1000)
2156 f.update({
2157 'vbr': vbr,
2158 'abr': abr,
2159 })
2160 codecs = parse_codecs(last_stream_inf.get('CODECS'))
2161 f.update(codecs)
2162 audio_group_id = last_stream_inf.get('AUDIO')
2163 # As per [1, 4.3.4.1.1] any EXT-X-STREAM-INF tag which
2164 # references a rendition group MUST have a CODECS attribute.
62b58c09 2165 # However, this is not always respected. E.g. [2]
310c2ed2 2166 # contains EXT-X-STREAM-INF tag which references AUDIO
2167 # rendition group but does not have CODECS and despite
2168 # referencing an audio group it represents a complete
2169 # (with audio and video) format. So, for such cases we will
2170 # ignore references to rendition groups and treat them
2171 # as complete formats.
2172 if audio_group_id and codecs and f.get('vcodec') != 'none':
2173 audio_group = groups.get(audio_group_id)
2174 if audio_group and audio_group[0].get('URI'):
2175 # TODO: update acodec for audio only formats with
2176 # the same GROUP-ID
2177 f['acodec'] = 'none'
fc21af50 2178 if not f.get('ext'):
2179 f['ext'] = 'm4a' if f.get('vcodec') == 'none' else 'mp4'
310c2ed2 2180 formats.append(f)
2181
2182 # for DailyMotion
2183 progressive_uri = last_stream_inf.get('PROGRESSIVE-URI')
2184 if progressive_uri:
2185 http_f = f.copy()
2186 del http_f['manifest_url']
2187 http_f.update({
2188 'format_id': f['format_id'].replace('hls-', 'http-'),
2189 'protocol': 'http',
2190 'url': progressive_uri,
2191 })
2192 formats.append(http_f)
5ef62fc4 2193
cb252080 2194 last_stream_inf = {}
a0c3b2d5 2195 return formats, subtitles
704df56d 2196
3cf4b91d
C
2197 def _extract_m3u8_vod_duration(
2198 self, m3u8_vod_url, video_id, note=None, errnote=None, data=None, headers={}, query={}):
2199
2200 m3u8_vod = self._download_webpage(
2201 m3u8_vod_url, video_id,
2202 note='Downloading m3u8 VOD manifest' if note is None else note,
2203 errnote='Failed to download VOD manifest' if errnote is None else errnote,
2204 fatal=False, data=data, headers=headers, query=query)
2205
2206 return self._parse_m3u8_vod_duration(m3u8_vod or '', video_id)
2207
2208 def _parse_m3u8_vod_duration(self, m3u8_vod, video_id):
5ab3534d 2209 if '#EXT-X-ENDLIST' not in m3u8_vod:
3cf4b91d
C
2210 return None
2211
2212 return int(sum(
2213 float(line[len('#EXTINF:'):].split(',')[0])
2214 for line in m3u8_vod.splitlines() if line.startswith('#EXTINF:'))) or None
2215
5ab3534d 2216 def _extract_mpd_vod_duration(
2217 self, mpd_url, video_id, note=None, errnote=None, data=None, headers={}, query={}):
2218
2219 mpd_doc = self._download_xml(
2220 mpd_url, video_id,
2221 note='Downloading MPD VOD manifest' if note is None else note,
2222 errnote='Failed to download VOD manifest' if errnote is None else errnote,
2223 fatal=False, data=data, headers=headers, query=query) or {}
2224 return int_or_none(parse_duration(mpd_doc.get('mediaPresentationDuration')))
2225
a107193e
S
2226 @staticmethod
2227 def _xpath_ns(path, namespace=None):
2228 if not namespace:
2229 return path
2230 out = []
2231 for c in path.split('/'):
2232 if not c or c == '.':
2233 out.append(c)
2234 else:
2235 out.append('{%s}%s' % (namespace, c))
2236 return '/'.join(out)
2237
da1c94ee 2238 def _extract_smil_formats_and_subtitles(self, smil_url, video_id, fatal=True, f4m_params=None, transform_source=None):
0b5546c7 2239 if self.get_param('ignore_no_formats_error'):
2240 fatal = False
2241
a076c1f9
E
2242 res = self._download_smil(smil_url, video_id, fatal=fatal, transform_source=transform_source)
2243 if res is False:
995029a1 2244 assert not fatal
774a46c5 2245 return [], {}
e89a2aab 2246
a076c1f9
E
2247 smil, urlh = res
2248 smil_url = urlh.geturl()
2249
17712eeb 2250 namespace = self._parse_smil_namespace(smil)
a107193e 2251
da1c94ee 2252 fmts = self._parse_smil_formats(
a107193e 2253 smil, smil_url, video_id, namespace=namespace, f4m_params=f4m_params)
da1c94ee
F
2254 subs = self._parse_smil_subtitles(
2255 smil, namespace=namespace)
2256
2257 return fmts, subs
2258
2259 def _extract_smil_formats(self, *args, **kwargs):
2260 fmts, subs = self._extract_smil_formats_and_subtitles(*args, **kwargs)
2261 if subs:
b5ae35ee 2262 self._report_ignoring_subs('SMIL')
da1c94ee 2263 return fmts
a107193e
S
2264
2265 def _extract_smil_info(self, smil_url, video_id, fatal=True, f4m_params=None):
a076c1f9
E
2266 res = self._download_smil(smil_url, video_id, fatal=fatal)
2267 if res is False:
a107193e 2268 return {}
a076c1f9
E
2269
2270 smil, urlh = res
2271 smil_url = urlh.geturl()
2272
a107193e
S
2273 return self._parse_smil(smil, smil_url, video_id, f4m_params=f4m_params)
2274
09f572fb 2275 def _download_smil(self, smil_url, video_id, fatal=True, transform_source=None):
a076c1f9 2276 return self._download_xml_handle(
a107193e 2277 smil_url, video_id, 'Downloading SMIL file',
09f572fb 2278 'Unable to download SMIL file', fatal=fatal, transform_source=transform_source)
a107193e
S
2279
2280 def _parse_smil(self, smil, smil_url, video_id, f4m_params=None):
17712eeb 2281 namespace = self._parse_smil_namespace(smil)
a107193e
S
2282
2283 formats = self._parse_smil_formats(
2284 smil, smil_url, video_id, namespace=namespace, f4m_params=f4m_params)
2285 subtitles = self._parse_smil_subtitles(smil, namespace=namespace)
2286
2287 video_id = os.path.splitext(url_basename(smil_url))[0]
2288 title = None
2289 description = None
647eab45 2290 upload_date = None
a107193e
S
2291 for meta in smil.findall(self._xpath_ns('./head/meta', namespace)):
2292 name = meta.attrib.get('name')
2293 content = meta.attrib.get('content')
2294 if not name or not content:
2295 continue
2296 if not title and name == 'title':
2297 title = content
2298 elif not description and name in ('description', 'abstract'):
2299 description = content
647eab45
S
2300 elif not upload_date and name == 'date':
2301 upload_date = unified_strdate(content)
a107193e 2302
1e5bcdec
S
2303 thumbnails = [{
2304 'id': image.get('type'),
2305 'url': image.get('src'),
2306 'width': int_or_none(image.get('width')),
2307 'height': int_or_none(image.get('height')),
2308 } for image in smil.findall(self._xpath_ns('.//image', namespace)) if image.get('src')]
2309
a107193e
S
2310 return {
2311 'id': video_id,
2312 'title': title or video_id,
2313 'description': description,
647eab45 2314 'upload_date': upload_date,
1e5bcdec 2315 'thumbnails': thumbnails,
a107193e
S
2316 'formats': formats,
2317 'subtitles': subtitles,
2318 }
2319
17712eeb
S
2320 def _parse_smil_namespace(self, smil):
2321 return self._search_regex(
2322 r'(?i)^{([^}]+)?}smil$', smil.tag, 'namespace', default=None)
2323
f877c6ae 2324 def _parse_smil_formats(self, smil, smil_url, video_id, namespace=None, f4m_params=None, transform_rtmp_url=None):
a107193e
S
2325 base = smil_url
2326 for meta in smil.findall(self._xpath_ns('./head/meta', namespace)):
2327 b = meta.get('base') or meta.get('httpBase')
2328 if b:
2329 base = b
2330 break
e89a2aab
S
2331
2332 formats = []
2333 rtmp_count = 0
a107193e 2334 http_count = 0
7f32e5dc 2335 m3u8_count = 0
9359f3d4 2336 imgs_count = 0
a107193e 2337
9359f3d4 2338 srcs = set()
ad96b4c8
YCH
2339 media = smil.findall(self._xpath_ns('.//video', namespace)) + smil.findall(self._xpath_ns('.//audio', namespace))
2340 for medium in media:
2341 src = medium.get('src')
81e1c4e2 2342 if not src or src in srcs:
a107193e 2343 continue
9359f3d4 2344 srcs.add(src)
a107193e 2345
ad96b4c8
YCH
2346 bitrate = float_or_none(medium.get('system-bitrate') or medium.get('systemBitrate'), 1000)
2347 filesize = int_or_none(medium.get('size') or medium.get('fileSize'))
2348 width = int_or_none(medium.get('width'))
2349 height = int_or_none(medium.get('height'))
2350 proto = medium.get('proto')
2351 ext = medium.get('ext')
cb73b846 2352 src_ext = determine_ext(src, default_ext=None) or ext or urlhandle_detect_ext(
2353 self._request_webpage(HEADRequest(src), video_id, note='Requesting extension info', fatal=False))
ad96b4c8 2354 streamer = medium.get('streamer') or base
a107193e
S
2355
2356 if proto == 'rtmp' or streamer.startswith('rtmp'):
2357 rtmp_count += 1
2358 formats.append({
2359 'url': streamer,
2360 'play_path': src,
2361 'ext': 'flv',
2362 'format_id': 'rtmp-%d' % (rtmp_count if bitrate is None else bitrate),
2363 'tbr': bitrate,
2364 'filesize': filesize,
2365 'width': width,
2366 'height': height,
2367 })
f877c6ae
YCH
2368 if transform_rtmp_url:
2369 streamer, src = transform_rtmp_url(streamer, src)
2370 formats[-1].update({
2371 'url': streamer,
2372 'play_path': src,
2373 })
a107193e
S
2374 continue
2375
14f25df2 2376 src_url = src if src.startswith('http') else urllib.parse.urljoin(base, src)
c349456e 2377 src_url = src_url.strip()
a107193e
S
2378
2379 if proto == 'm3u8' or src_ext == 'm3u8':
7f32e5dc 2380 m3u8_formats = self._extract_m3u8_formats(
2381 src_url, video_id, ext or 'mp4', m3u8_id='hls', fatal=False)
2382 if len(m3u8_formats) == 1:
2383 m3u8_count += 1
2384 m3u8_formats[0].update({
2385 'format_id': 'hls-%d' % (m3u8_count if bitrate is None else bitrate),
2386 'tbr': bitrate,
2387 'width': width,
2388 'height': height,
2389 })
2390 formats.extend(m3u8_formats)
bd21ead2 2391 elif src_ext == 'f4m':
a107193e
S
2392 f4m_url = src_url
2393 if not f4m_params:
2394 f4m_params = {
2395 'hdcore': '3.2.0',
2396 'plugin': 'flowplayer-3.2.0.1',
2397 }
2398 f4m_url += '&' if '?' in f4m_url else '?'
14f25df2 2399 f4m_url += urllib.parse.urlencode(f4m_params)
7e5edcfd 2400 formats.extend(self._extract_f4m_formats(f4m_url, video_id, f4m_id='hds', fatal=False))
bd21ead2
RA
2401 elif src_ext == 'mpd':
2402 formats.extend(self._extract_mpd_formats(
2403 src_url, video_id, mpd_id='dash', fatal=False))
2404 elif re.search(r'\.ism/[Mm]anifest', src_url):
2405 formats.extend(self._extract_ism_formats(
2406 src_url, video_id, ism_id='mss', fatal=False))
2407 elif src_url.startswith('http') and self._is_valid_url(src, video_id):
a107193e
S
2408 http_count += 1
2409 formats.append({
2410 'url': src_url,
2411 'ext': ext or src_ext or 'flv',
2412 'format_id': 'http-%d' % (bitrate or http_count),
2413 'tbr': bitrate,
2414 'filesize': filesize,
2415 'width': width,
2416 'height': height,
2417 })
63757032 2418
9359f3d4
F
2419 for medium in smil.findall(self._xpath_ns('.//imagestream', namespace)):
2420 src = medium.get('src')
2421 if not src or src in srcs:
2422 continue
2423 srcs.add(src)
2424
2425 imgs_count += 1
2426 formats.append({
2427 'format_id': 'imagestream-%d' % (imgs_count),
2428 'url': src,
2429 'ext': mimetype2ext(medium.get('type')),
2430 'acodec': 'none',
2431 'vcodec': 'none',
2432 'width': int_or_none(medium.get('width')),
2433 'height': int_or_none(medium.get('height')),
2434 'format_note': 'SMIL storyboards',
2435 })
2436
e89a2aab
S
2437 return formats
2438
ce00af87 2439 def _parse_smil_subtitles(self, smil, namespace=None, subtitles_lang='en'):
d413095f 2440 urls = []
a107193e
S
2441 subtitles = {}
2442 for num, textstream in enumerate(smil.findall(self._xpath_ns('.//textstream', namespace))):
2443 src = textstream.get('src')
d413095f 2444 if not src or src in urls:
a107193e 2445 continue
d413095f 2446 urls.append(src)
df634be2 2447 ext = textstream.get('ext') or mimetype2ext(textstream.get('type')) or determine_ext(src)
03bc7237 2448 lang = textstream.get('systemLanguage') or textstream.get('systemLanguageName') or textstream.get('lang') or subtitles_lang
a107193e
S
2449 subtitles.setdefault(lang, []).append({
2450 'url': src,
2451 'ext': ext,
2452 })
2453 return subtitles
63757032 2454
47a5cb77 2455 def _extract_xspf_playlist(self, xspf_url, playlist_id, fatal=True):
a076c1f9 2456 res = self._download_xml_handle(
47a5cb77 2457 xspf_url, playlist_id, 'Downloading xpsf playlist',
942acef5 2458 'Unable to download xspf manifest', fatal=fatal)
a076c1f9 2459 if res is False:
942acef5 2460 return []
a076c1f9
E
2461
2462 xspf, urlh = res
2463 xspf_url = urlh.geturl()
2464
47a5cb77
S
2465 return self._parse_xspf(
2466 xspf, playlist_id, xspf_url=xspf_url,
2467 xspf_base_url=base_url(xspf_url))
8d6765cf 2468
47a5cb77 2469 def _parse_xspf(self, xspf_doc, playlist_id, xspf_url=None, xspf_base_url=None):
8d6765cf
S
2470 NS_MAP = {
2471 'xspf': 'http://xspf.org/ns/0/',
2472 's1': 'http://static.streamone.nl/player/ns/0',
2473 }
2474
2475 entries = []
47a5cb77 2476 for track in xspf_doc.findall(xpath_with_ns('./xspf:trackList/xspf:track', NS_MAP)):
8d6765cf 2477 title = xpath_text(
98044462 2478 track, xpath_with_ns('./xspf:title', NS_MAP), 'title', default=playlist_id)
8d6765cf
S
2479 description = xpath_text(
2480 track, xpath_with_ns('./xspf:annotation', NS_MAP), 'description')
2481 thumbnail = xpath_text(
2482 track, xpath_with_ns('./xspf:image', NS_MAP), 'thumbnail')
2483 duration = float_or_none(
2484 xpath_text(track, xpath_with_ns('./xspf:duration', NS_MAP), 'duration'), 1000)
2485
47a5cb77
S
2486 formats = []
2487 for location in track.findall(xpath_with_ns('./xspf:location', NS_MAP)):
2488 format_url = urljoin(xspf_base_url, location.text)
2489 if not format_url:
2490 continue
2491 formats.append({
2492 'url': format_url,
2493 'manifest_url': xspf_url,
2494 'format_id': location.get(xpath_with_ns('s1:label', NS_MAP)),
2495 'width': int_or_none(location.get(xpath_with_ns('s1:width', NS_MAP))),
2496 'height': int_or_none(location.get(xpath_with_ns('s1:height', NS_MAP))),
2497 })
8d6765cf
S
2498
2499 entries.append({
2500 'id': playlist_id,
2501 'title': title,
2502 'description': description,
2503 'thumbnail': thumbnail,
2504 'duration': duration,
2505 'formats': formats,
2506 })
2507 return entries
2508
171e59ed
F
2509 def _extract_mpd_formats(self, *args, **kwargs):
2510 fmts, subs = self._extract_mpd_formats_and_subtitles(*args, **kwargs)
2511 if subs:
b5ae35ee 2512 self._report_ignoring_subs('DASH')
171e59ed
F
2513 return fmts
2514
2515 def _extract_mpd_formats_and_subtitles(
2516 self, mpd_url, video_id, mpd_id=None, note=None, errnote=None,
2517 fatal=True, data=None, headers={}, query={}):
0b5546c7 2518
2519 if self.get_param('ignore_no_formats_error'):
2520 fatal = False
2521
47a5cb77 2522 res = self._download_xml_handle(
1bac3455 2523 mpd_url, video_id,
37a3bb66 2524 note='Downloading MPD manifest' if note is None else note,
2525 errnote='Failed to download MPD manifest' if errnote is None else errnote,
7360c06f 2526 fatal=fatal, data=data, headers=headers, query=query)
1bac3455 2527 if res is False:
171e59ed 2528 return [], {}
47a5cb77 2529 mpd_doc, urlh = res
c25720ef 2530 if mpd_doc is None:
171e59ed 2531 return [], {}
779da8e3
E
2532
2533 # We could have been redirected to a new url when we retrieved our mpd file.
2534 mpd_url = urlh.geturl()
2535 mpd_base_url = base_url(mpd_url)
1bac3455 2536
171e59ed 2537 return self._parse_mpd_formats_and_subtitles(
545cc85d 2538 mpd_doc, mpd_id, mpd_base_url, mpd_url)
2d2fa82d 2539
171e59ed
F
2540 def _parse_mpd_formats(self, *args, **kwargs):
2541 fmts, subs = self._parse_mpd_formats_and_subtitles(*args, **kwargs)
2542 if subs:
b5ae35ee 2543 self._report_ignoring_subs('DASH')
171e59ed
F
2544 return fmts
2545
2546 def _parse_mpd_formats_and_subtitles(
2547 self, mpd_doc, mpd_id=None, mpd_base_url='', mpd_url=None):
f0948348
S
2548 """
2549 Parse formats from MPD manifest.
2550 References:
2551 1. MPEG-DASH Standard, ISO/IEC 23009-1:2014(E),
2552 http://standards.iso.org/ittf/PubliclyAvailableStandards/c065274_ISO_IEC_23009-1_2014.zip
2553 2. https://en.wikipedia.org/wiki/Dynamic_Adaptive_Streaming_over_HTTP
2554 """
a06916d9 2555 if not self.get_param('dynamic_mpd', True):
78895bd3 2556 if mpd_doc.get('type') == 'dynamic':
171e59ed 2557 return [], {}
2d2fa82d 2558
91cb6b50 2559 namespace = self._search_regex(r'(?i)^{([^}]+)?}MPD$', mpd_doc.tag, 'namespace', default=None)
f14be228 2560
2561 def _add_ns(path):
2562 return self._xpath_ns(path, namespace)
2563
675d0016 2564 def is_drm_protected(element):
2565 return element.find(_add_ns('ContentProtection')) is not None
2566
1bac3455 2567 def extract_multisegment_info(element, ms_parent_info):
2568 ms_info = ms_parent_info.copy()
b4c1d6e8
S
2569
2570 # As per [1, 5.3.9.2.2] SegmentList and SegmentTemplate share some
2571 # common attributes and elements. We will only extract relevant
2572 # for us.
2573 def extract_common(source):
2574 segment_timeline = source.find(_add_ns('SegmentTimeline'))
2575 if segment_timeline is not None:
2576 s_e = segment_timeline.findall(_add_ns('S'))
2577 if s_e:
2578 ms_info['total_number'] = 0
2579 ms_info['s'] = []
2580 for s in s_e:
2581 r = int(s.get('r', 0))
2582 ms_info['total_number'] += 1 + r
2583 ms_info['s'].append({
2584 't': int(s.get('t', 0)),
2585 # @d is mandatory (see [1, 5.3.9.6.2, Table 17, page 60])
2586 'd': int(s.attrib['d']),
2587 'r': r,
2588 })
2589 start_number = source.get('startNumber')
2590 if start_number:
2591 ms_info['start_number'] = int(start_number)
2592 timescale = source.get('timescale')
2593 if timescale:
2594 ms_info['timescale'] = int(timescale)
2595 segment_duration = source.get('duration')
2596 if segment_duration:
48504785 2597 ms_info['segment_duration'] = float(segment_duration)
b4c1d6e8
S
2598
2599 def extract_Initialization(source):
2600 initialization = source.find(_add_ns('Initialization'))
2601 if initialization is not None:
2602 ms_info['initialization_url'] = initialization.attrib['sourceURL']
2603
f14be228 2604 segment_list = element.find(_add_ns('SegmentList'))
1bac3455 2605 if segment_list is not None:
b4c1d6e8
S
2606 extract_common(segment_list)
2607 extract_Initialization(segment_list)
f14be228 2608 segment_urls_e = segment_list.findall(_add_ns('SegmentURL'))
1bac3455 2609 if segment_urls_e:
2610 ms_info['segment_urls'] = [segment.attrib['media'] for segment in segment_urls_e]
1bac3455 2611 else:
f14be228 2612 segment_template = element.find(_add_ns('SegmentTemplate'))
1bac3455 2613 if segment_template is not None:
b4c1d6e8 2614 extract_common(segment_template)
e228616c
S
2615 media = segment_template.get('media')
2616 if media:
2617 ms_info['media'] = media
1bac3455 2618 initialization = segment_template.get('initialization')
2619 if initialization:
e228616c 2620 ms_info['initialization'] = initialization
1bac3455 2621 else:
b4c1d6e8 2622 extract_Initialization(segment_template)
1bac3455 2623 return ms_info
b323e170 2624
1bac3455 2625 mpd_duration = parse_duration(mpd_doc.get('mediaPresentationDuration'))
6251555f 2626 formats, subtitles = [], {}
234416e4 2627 stream_numbers = collections.defaultdict(int)
f14be228 2628 for period in mpd_doc.findall(_add_ns('Period')):
1bac3455 2629 period_duration = parse_duration(period.get('duration')) or mpd_duration
2630 period_ms_info = extract_multisegment_info(period, {
2631 'start_number': 1,
2632 'timescale': 1,
2633 })
f14be228 2634 for adaptation_set in period.findall(_add_ns('AdaptationSet')):
1bac3455 2635 adaption_set_ms_info = extract_multisegment_info(adaptation_set, period_ms_info)
f14be228 2636 for representation in adaptation_set.findall(_add_ns('Representation')):
1bac3455 2637 representation_attrib = adaptation_set.attrib.copy()
2638 representation_attrib.update(representation.attrib)
f0948348 2639 # According to [1, 5.3.7.2, Table 9, page 41], @mimeType is mandatory
a6c8b759 2640 mime_type = representation_attrib['mimeType']
171e59ed
F
2641 content_type = representation_attrib.get('contentType', mime_type.split('/')[0])
2642
21633673 2643 codec_str = representation_attrib.get('codecs', '')
2644 # Some kind of binary subtitle found in some youtube livestreams
2645 if mime_type == 'application/x-rawcc':
2646 codecs = {'scodec': codec_str}
2647 else:
2648 codecs = parse_codecs(codec_str)
be2fc5b2 2649 if content_type not in ('video', 'audio', 'text'):
2650 if mime_type == 'image/jpeg':
a8731fcc 2651 content_type = mime_type
21633673 2652 elif codecs.get('vcodec', 'none') != 'none':
4afa3ec4 2653 content_type = 'video'
21633673 2654 elif codecs.get('acodec', 'none') != 'none':
4afa3ec4 2655 content_type = 'audio'
3fe75fdc 2656 elif codecs.get('scodec', 'none') != 'none':
be2fc5b2 2657 content_type = 'text'
6993f78d 2658 elif mimetype2ext(mime_type) in ('tt', 'dfxp', 'ttml', 'xml', 'json'):
2659 content_type = 'text'
cdb19aa4 2660 else:
be2fc5b2 2661 self.report_warning('Unknown MIME type %s in DASH manifest' % mime_type)
2662 continue
2663
2664 base_url = ''
2665 for element in (representation, adaptation_set, period, mpd_doc):
2666 base_url_e = element.find(_add_ns('BaseURL'))
47046464 2667 if try_call(lambda: base_url_e.text) is not None:
be2fc5b2 2668 base_url = base_url_e.text + base_url
2669 if re.match(r'^https?://', base_url):
2670 break
f9cc0161 2671 if mpd_base_url and base_url.startswith('/'):
14f25df2 2672 base_url = urllib.parse.urljoin(mpd_base_url, base_url)
f9cc0161
D
2673 elif mpd_base_url and not re.match(r'^https?://', base_url):
2674 if not mpd_base_url.endswith('/'):
be2fc5b2 2675 mpd_base_url += '/'
2676 base_url = mpd_base_url + base_url
2677 representation_id = representation_attrib.get('id')
2678 lang = representation_attrib.get('lang')
2679 url_el = representation.find(_add_ns('BaseURL'))
2680 filesize = int_or_none(url_el.attrib.get('{http://youtube.com/yt/2012/10/10}contentLength') if url_el is not None else None)
2681 bandwidth = int_or_none(representation_attrib.get('bandwidth'))
2682 if representation_id is not None:
2683 format_id = representation_id
2684 else:
2685 format_id = content_type
2686 if mpd_id:
2687 format_id = mpd_id + '-' + format_id
2688 if content_type in ('video', 'audio'):
2689 f = {
2690 'format_id': format_id,
2691 'manifest_url': mpd_url,
2692 'ext': mimetype2ext(mime_type),
2693 'width': int_or_none(representation_attrib.get('width')),
2694 'height': int_or_none(representation_attrib.get('height')),
2695 'tbr': float_or_none(bandwidth, 1000),
2696 'asr': int_or_none(representation_attrib.get('audioSamplingRate')),
2697 'fps': int_or_none(representation_attrib.get('frameRate')),
2698 'language': lang if lang not in ('mul', 'und', 'zxx', 'mis') else None,
2699 'format_note': 'DASH %s' % content_type,
2700 'filesize': filesize,
2701 'container': mimetype2ext(mime_type) + '_dash',
4afa3ec4 2702 **codecs
be2fc5b2 2703 }
be2fc5b2 2704 elif content_type == 'text':
2705 f = {
2706 'ext': mimetype2ext(mime_type),
2707 'manifest_url': mpd_url,
2708 'filesize': filesize,
2709 }
2710 elif content_type == 'image/jpeg':
2711 # See test case in VikiIE
2712 # https://www.viki.com/videos/1175236v-choosing-spouse-by-lottery-episode-1
2713 f = {
2714 'format_id': format_id,
2715 'ext': 'mhtml',
2716 'manifest_url': mpd_url,
2717 'format_note': 'DASH storyboards (jpeg)',
2718 'acodec': 'none',
2719 'vcodec': 'none',
2720 }
88acdbc2 2721 if is_drm_protected(adaptation_set) or is_drm_protected(representation):
2722 f['has_drm'] = True
be2fc5b2 2723 representation_ms_info = extract_multisegment_info(representation, adaption_set_ms_info)
2724
2725 def prepare_template(template_name, identifiers):
2726 tmpl = representation_ms_info[template_name]
0cb0fdbb 2727 if representation_id is not None:
2728 tmpl = tmpl.replace('$RepresentationID$', representation_id)
be2fc5b2 2729 # First of, % characters outside $...$ templates
2730 # must be escaped by doubling for proper processing
2731 # by % operator string formatting used further (see
2732 # https://github.com/ytdl-org/youtube-dl/issues/16867).
2733 t = ''
2734 in_template = False
2735 for c in tmpl:
2736 t += c
2737 if c == '$':
2738 in_template = not in_template
2739 elif c == '%' and not in_template:
eca1f0d1 2740 t += c
be2fc5b2 2741 # Next, $...$ templates are translated to their
2742 # %(...) counterparts to be used with % operator
be2fc5b2 2743 t = re.sub(r'\$(%s)\$' % '|'.join(identifiers), r'%(\1)d', t)
2744 t = re.sub(r'\$(%s)%%([^$]+)\$' % '|'.join(identifiers), r'%(\1)\2', t)
2745 t.replace('$$', '$')
2746 return t
2747
2748 # @initialization is a regular template like @media one
2749 # so it should be handled just the same way (see
2750 # https://github.com/ytdl-org/youtube-dl/issues/11605)
2751 if 'initialization' in representation_ms_info:
2752 initialization_template = prepare_template(
2753 'initialization',
2754 # As per [1, 5.3.9.4.2, Table 15, page 54] $Number$ and
2755 # $Time$ shall not be included for @initialization thus
2756 # only $Bandwidth$ remains
2757 ('Bandwidth', ))
2758 representation_ms_info['initialization_url'] = initialization_template % {
2759 'Bandwidth': bandwidth,
2760 }
2761
2762 def location_key(location):
2763 return 'url' if re.match(r'^https?://', location) else 'path'
2764
2765 if 'segment_urls' not in representation_ms_info and 'media' in representation_ms_info:
2766
2767 media_template = prepare_template('media', ('Number', 'Bandwidth', 'Time'))
2768 media_location_key = location_key(media_template)
2769
2770 # As per [1, 5.3.9.4.4, Table 16, page 55] $Number$ and $Time$
2771 # can't be used at the same time
2772 if '%(Number' in media_template and 's' not in representation_ms_info:
2773 segment_duration = None
2774 if 'total_number' not in representation_ms_info and 'segment_duration' in representation_ms_info:
2775 segment_duration = float_or_none(representation_ms_info['segment_duration'], representation_ms_info['timescale'])
ffa89477 2776 representation_ms_info['total_number'] = int(math.ceil(
2777 float_or_none(period_duration, segment_duration, default=0)))
be2fc5b2 2778 representation_ms_info['fragments'] = [{
2779 media_location_key: media_template % {
2780 'Number': segment_number,
2781 'Bandwidth': bandwidth,
2782 },
2783 'duration': segment_duration,
2784 } for segment_number in range(
2785 representation_ms_info['start_number'],
2786 representation_ms_info['total_number'] + representation_ms_info['start_number'])]
2787 else:
2788 # $Number*$ or $Time$ in media template with S list available
2789 # Example $Number*$: http://www.svtplay.se/klipp/9023742/stopptid-om-bjorn-borg
2790 # Example $Time$: https://play.arkena.com/embed/avp/v2/player/media/b41dda37-d8e7-4d3f-b1b5-9a9db578bdfe/1/129411
2791 representation_ms_info['fragments'] = []
2792 segment_time = 0
2793 segment_d = None
2794 segment_number = representation_ms_info['start_number']
2795
2796 def add_segment_url():
2797 segment_url = media_template % {
2798 'Time': segment_time,
2799 'Bandwidth': bandwidth,
2800 'Number': segment_number,
2801 }
2802 representation_ms_info['fragments'].append({
2803 media_location_key: segment_url,
2804 'duration': float_or_none(segment_d, representation_ms_info['timescale']),
2805 })
2806
2807 for num, s in enumerate(representation_ms_info['s']):
2808 segment_time = s.get('t') or segment_time
2809 segment_d = s['d']
2810 add_segment_url()
2811 segment_number += 1
2812 for r in range(s.get('r', 0)):
2813 segment_time += segment_d
f0948348 2814 add_segment_url()
b4c1d6e8 2815 segment_number += 1
be2fc5b2 2816 segment_time += segment_d
2817 elif 'segment_urls' in representation_ms_info and 's' in representation_ms_info:
62b58c09
L
2818 # No media template,
2819 # e.g. https://www.youtube.com/watch?v=iXZV5uAYMJI
be2fc5b2 2820 # or any YouTube dashsegments video
2821 fragments = []
2822 segment_index = 0
2823 timescale = representation_ms_info['timescale']
2824 for s in representation_ms_info['s']:
2825 duration = float_or_none(s['d'], timescale)
2826 for r in range(s.get('r', 0) + 1):
2827 segment_uri = representation_ms_info['segment_urls'][segment_index]
2828 fragments.append({
2829 location_key(segment_uri): segment_uri,
2830 'duration': duration,
2831 })
2832 segment_index += 1
2833 representation_ms_info['fragments'] = fragments
2834 elif 'segment_urls' in representation_ms_info:
2835 # Segment URLs with no SegmentTimeline
62b58c09 2836 # E.g. https://www.seznam.cz/zpravy/clanek/cesko-zasahne-vitr-o-sile-vichrice-muze-byt-i-zivotu-nebezpecny-39091
be2fc5b2 2837 # https://github.com/ytdl-org/youtube-dl/pull/14844
2838 fragments = []
2839 segment_duration = float_or_none(
2840 representation_ms_info['segment_duration'],
2841 representation_ms_info['timescale']) if 'segment_duration' in representation_ms_info else None
2842 for segment_url in representation_ms_info['segment_urls']:
2843 fragment = {
2844 location_key(segment_url): segment_url,
2845 }
2846 if segment_duration:
2847 fragment['duration'] = segment_duration
2848 fragments.append(fragment)
2849 representation_ms_info['fragments'] = fragments
2850 # If there is a fragments key available then we correctly recognized fragmented media.
2851 # Otherwise we will assume unfragmented media with direct access. Technically, such
2852 # assumption is not necessarily correct since we may simply have no support for
2853 # some forms of fragmented media renditions yet, but for now we'll use this fallback.
2854 if 'fragments' in representation_ms_info:
2855 f.update({
2856 # NB: mpd_url may be empty when MPD manifest is parsed from a string
2857 'url': mpd_url or base_url,
2858 'fragment_base_url': base_url,
2859 'fragments': [],
2860 'protocol': 'http_dash_segments' if mime_type != 'image/jpeg' else 'mhtml',
2861 })
2862 if 'initialization_url' in representation_ms_info:
2863 initialization_url = representation_ms_info['initialization_url']
2864 if not f.get('url'):
2865 f['url'] = initialization_url
2866 f['fragments'].append({location_key(initialization_url): initialization_url})
2867 f['fragments'].extend(representation_ms_info['fragments'])
ffa89477 2868 if not period_duration:
2869 period_duration = try_get(
2870 representation_ms_info,
2871 lambda r: sum(frag['duration'] for frag in r['fragments']), float)
17b598d3 2872 else:
be2fc5b2 2873 # Assuming direct URL to unfragmented media.
2874 f['url'] = base_url
234416e4 2875 if content_type in ('video', 'audio', 'image/jpeg'):
2876 f['manifest_stream_number'] = stream_numbers[f['url']]
2877 stream_numbers[f['url']] += 1
be2fc5b2 2878 formats.append(f)
2879 elif content_type == 'text':
2880 subtitles.setdefault(lang or 'und', []).append(f)
2881
171e59ed 2882 return formats, subtitles
17b598d3 2883
fd76a142
F
2884 def _extract_ism_formats(self, *args, **kwargs):
2885 fmts, subs = self._extract_ism_formats_and_subtitles(*args, **kwargs)
2886 if subs:
b5ae35ee 2887 self._report_ignoring_subs('ISM')
fd76a142
F
2888 return fmts
2889
2890 def _extract_ism_formats_and_subtitles(self, ism_url, video_id, ism_id=None, note=None, errnote=None, fatal=True, data=None, headers={}, query={}):
0b5546c7 2891 if self.get_param('ignore_no_formats_error'):
2892 fatal = False
2893
47a5cb77 2894 res = self._download_xml_handle(
b2758123 2895 ism_url, video_id,
37a3bb66 2896 note='Downloading ISM manifest' if note is None else note,
2897 errnote='Failed to download ISM manifest' if errnote is None else errnote,
7360c06f 2898 fatal=fatal, data=data, headers=headers, query=query)
b2758123 2899 if res is False:
fd76a142 2900 return [], {}
47a5cb77 2901 ism_doc, urlh = res
13b08034 2902 if ism_doc is None:
fd76a142 2903 return [], {}
b2758123 2904
fd76a142 2905 return self._parse_ism_formats_and_subtitles(ism_doc, urlh.geturl(), ism_id)
b2758123 2906
fd76a142 2907 def _parse_ism_formats_and_subtitles(self, ism_doc, ism_url, ism_id=None):
76d5a363
S
2908 """
2909 Parse formats from ISM manifest.
2910 References:
2911 1. [MS-SSTR]: Smooth Streaming Protocol,
2912 https://msdn.microsoft.com/en-us/library/ff469518.aspx
2913 """
06869367 2914 if ism_doc.get('IsLive') == 'TRUE':
fd76a142 2915 return [], {}
b2758123 2916
b2758123
RA
2917 duration = int(ism_doc.attrib['Duration'])
2918 timescale = int_or_none(ism_doc.get('TimeScale')) or 10000000
2919
2920 formats = []
fd76a142 2921 subtitles = {}
b2758123
RA
2922 for stream in ism_doc.findall('StreamIndex'):
2923 stream_type = stream.get('Type')
fd76a142 2924 if stream_type not in ('video', 'audio', 'text'):
b2758123
RA
2925 continue
2926 url_pattern = stream.attrib['Url']
2927 stream_timescale = int_or_none(stream.get('TimeScale')) or timescale
2928 stream_name = stream.get('Name')
fd76a142 2929 stream_language = stream.get('Language', 'und')
b2758123 2930 for track in stream.findall('QualityLevel'):
81b6102d 2931 KNOWN_TAGS = {'255': 'AACL', '65534': 'EC-3'}
2932 fourcc = track.get('FourCC') or KNOWN_TAGS.get(track.get('AudioTag'))
b2758123 2933 # TODO: add support for WVC1 and WMAP
81b6102d 2934 if fourcc not in ('H264', 'AVC1', 'AACL', 'TTML', 'EC-3'):
b2758123
RA
2935 self.report_warning('%s is not a supported codec' % fourcc)
2936 continue
2937 tbr = int(track.attrib['Bitrate']) // 1000
76d5a363
S
2938 # [1] does not mention Width and Height attributes. However,
2939 # they're often present while MaxWidth and MaxHeight are
2940 # missing, so should be used as fallbacks
2941 width = int_or_none(track.get('MaxWidth') or track.get('Width'))
2942 height = int_or_none(track.get('MaxHeight') or track.get('Height'))
b2758123
RA
2943 sampling_rate = int_or_none(track.get('SamplingRate'))
2944
2945 track_url_pattern = re.sub(r'{[Bb]itrate}', track.attrib['Bitrate'], url_pattern)
14f25df2 2946 track_url_pattern = urllib.parse.urljoin(ism_url, track_url_pattern)
b2758123
RA
2947
2948 fragments = []
2949 fragment_ctx = {
2950 'time': 0,
2951 }
2952 stream_fragments = stream.findall('c')
2953 for stream_fragment_index, stream_fragment in enumerate(stream_fragments):
2954 fragment_ctx['time'] = int_or_none(stream_fragment.get('t')) or fragment_ctx['time']
2955 fragment_repeat = int_or_none(stream_fragment.get('r')) or 1
2956 fragment_ctx['duration'] = int_or_none(stream_fragment.get('d'))
2957 if not fragment_ctx['duration']:
2958 try:
2959 next_fragment_time = int(stream_fragment[stream_fragment_index + 1].attrib['t'])
2960 except IndexError:
2961 next_fragment_time = duration
1616f9b4 2962 fragment_ctx['duration'] = (next_fragment_time - fragment_ctx['time']) / fragment_repeat
b2758123
RA
2963 for _ in range(fragment_repeat):
2964 fragments.append({
14f25df2 2965 'url': re.sub(r'{start[ _]time}', str(fragment_ctx['time']), track_url_pattern),
b2758123
RA
2966 'duration': fragment_ctx['duration'] / stream_timescale,
2967 })
2968 fragment_ctx['time'] += fragment_ctx['duration']
2969
fd76a142
F
2970 if stream_type == 'text':
2971 subtitles.setdefault(stream_language, []).append({
2972 'ext': 'ismt',
2973 'protocol': 'ism',
2974 'url': ism_url,
2975 'manifest_url': ism_url,
2976 'fragments': fragments,
2977 '_download_params': {
2978 'stream_type': stream_type,
2979 'duration': duration,
2980 'timescale': stream_timescale,
2981 'fourcc': fourcc,
2982 'language': stream_language,
2983 'codec_private_data': track.get('CodecPrivateData'),
2984 }
2985 })
2986 elif stream_type in ('video', 'audio'):
2987 formats.append({
34921b43 2988 'format_id': join_nonempty(ism_id, stream_name, tbr),
fd76a142
F
2989 'url': ism_url,
2990 'manifest_url': ism_url,
2991 'ext': 'ismv' if stream_type == 'video' else 'isma',
2992 'width': width,
2993 'height': height,
2994 'tbr': tbr,
2995 'asr': sampling_rate,
2996 'vcodec': 'none' if stream_type == 'audio' else fourcc,
2997 'acodec': 'none' if stream_type == 'video' else fourcc,
2998 'protocol': 'ism',
2999 'fragments': fragments,
88acdbc2 3000 'has_drm': ism_doc.find('Protection') is not None,
f68434cc 3001 'language': stream_language,
3002 'audio_channels': int_or_none(track.get('Channels')),
fd76a142
F
3003 '_download_params': {
3004 'stream_type': stream_type,
3005 'duration': duration,
3006 'timescale': stream_timescale,
3007 'width': width or 0,
3008 'height': height or 0,
3009 'fourcc': fourcc,
3010 'language': stream_language,
3011 'codec_private_data': track.get('CodecPrivateData'),
3012 'sampling_rate': sampling_rate,
3013 'channels': int_or_none(track.get('Channels', 2)),
3014 'bits_per_sample': int_or_none(track.get('BitsPerSample', 16)),
3015 'nal_unit_length_field': int_or_none(track.get('NALUnitLengthField', 4)),
3016 },
3017 })
3018 return formats, subtitles
b2758123 3019
079a7cfc 3020 def _parse_html5_media_entries(self, base_url, webpage, video_id, m3u8_id=None, m3u8_entry_protocol='m3u8_native', mpd_id=None, preference=None, quality=None):
6780154e
S
3021 def absolute_url(item_url):
3022 return urljoin(base_url, item_url)
59bbe491 3023
3024 def parse_content_type(content_type):
3025 if not content_type:
3026 return {}
3027 ctr = re.search(r'(?P<mimetype>[^/]+/[^;]+)(?:;\s*codecs="?(?P<codecs>[^"]+))?', content_type)
3028 if ctr:
3029 mimetype, codecs = ctr.groups()
3030 f = parse_codecs(codecs)
3031 f['ext'] = mimetype2ext(mimetype)
3032 return f
3033 return {}
3034
222a2308
L
3035 def _media_formats(src, cur_media_type, type_info=None):
3036 type_info = type_info or {}
520251c0 3037 full_url = absolute_url(src)
82889d4a 3038 ext = type_info.get('ext') or determine_ext(full_url)
87a449c1 3039 if ext == 'm3u8':
520251c0
YCH
3040 is_plain_url = False
3041 formats = self._extract_m3u8_formats(
ad120ae1 3042 full_url, video_id, ext='mp4',
eeb0a956 3043 entry_protocol=m3u8_entry_protocol, m3u8_id=m3u8_id,
f983b875 3044 preference=preference, quality=quality, fatal=False)
87a449c1
S
3045 elif ext == 'mpd':
3046 is_plain_url = False
3047 formats = self._extract_mpd_formats(
b359e977 3048 full_url, video_id, mpd_id=mpd_id, fatal=False)
520251c0
YCH
3049 else:
3050 is_plain_url = True
3051 formats = [{
3052 'url': full_url,
3053 'vcodec': 'none' if cur_media_type == 'audio' else None,
222a2308 3054 'ext': ext,
520251c0
YCH
3055 }]
3056 return is_plain_url, formats
3057
59bbe491 3058 entries = []
4328ddf8 3059 # amp-video and amp-audio are very similar to their HTML5 counterparts
962ffcf8 3060 # so we will include them right here (see
4328ddf8 3061 # https://www.ampproject.org/docs/reference/components/amp-video)
29f7c58a 3062 # For dl8-* tags see https://delight-vr.com/documentation/dl8-video/
3063 _MEDIA_TAG_NAME_RE = r'(?:(?:amp|dl8(?:-live)?)-)?(video|audio)'
3064 media_tags = [(media_tag, media_tag_name, media_type, '')
3065 for media_tag, media_tag_name, media_type
3066 in re.findall(r'(?s)(<(%s)[^>]*/>)' % _MEDIA_TAG_NAME_RE, webpage)]
2aec7256
S
3067 media_tags.extend(re.findall(
3068 # We only allow video|audio followed by a whitespace or '>'.
3069 # Allowing more characters may end up in significant slow down (see
62b58c09
L
3070 # https://github.com/ytdl-org/youtube-dl/issues/11979,
3071 # e.g. http://www.porntrex.com/maps/videositemap.xml).
29f7c58a 3072 r'(?s)(<(?P<tag>%s)(?:\s+[^>]*)?>)(.*?)</(?P=tag)>' % _MEDIA_TAG_NAME_RE, webpage))
3073 for media_tag, _, media_type, media_content in media_tags:
59bbe491 3074 media_info = {
3075 'formats': [],
3076 'subtitles': {},
3077 }
3078 media_attributes = extract_attributes(media_tag)
bfbecd11 3079 src = strip_or_none(dict_get(media_attributes, ('src', 'data-video-src', 'data-src', 'data-source')))
59bbe491 3080 if src:
222a2308
L
3081 f = parse_content_type(media_attributes.get('type'))
3082 _, formats = _media_formats(src, media_type, f)
520251c0 3083 media_info['formats'].extend(formats)
6780154e 3084 media_info['thumbnail'] = absolute_url(media_attributes.get('poster'))
59bbe491 3085 if media_content:
3086 for source_tag in re.findall(r'<source[^>]+>', media_content):
d493f15c
S
3087 s_attr = extract_attributes(source_tag)
3088 # data-video-src and data-src are non standard but seen
3089 # several times in the wild
bfbecd11 3090 src = strip_or_none(dict_get(s_attr, ('src', 'data-video-src', 'data-src', 'data-source')))
59bbe491 3091 if not src:
3092 continue
d493f15c 3093 f = parse_content_type(s_attr.get('type'))
868f79db 3094 is_plain_url, formats = _media_formats(src, media_type, f)
520251c0 3095 if is_plain_url:
d493f15c
S
3096 # width, height, res, label and title attributes are
3097 # all not standard but seen several times in the wild
3098 labels = [
3099 s_attr.get(lbl)
3100 for lbl in ('label', 'title')
3101 if str_or_none(s_attr.get(lbl))
3102 ]
3103 width = int_or_none(s_attr.get('width'))
3089bc74
S
3104 height = (int_or_none(s_attr.get('height'))
3105 or int_or_none(s_attr.get('res')))
d493f15c
S
3106 if not width or not height:
3107 for lbl in labels:
3108 resolution = parse_resolution(lbl)
3109 if not resolution:
3110 continue
3111 width = width or resolution.get('width')
3112 height = height or resolution.get('height')
3113 for lbl in labels:
3114 tbr = parse_bitrate(lbl)
3115 if tbr:
3116 break
3117 else:
3118 tbr = None
1ed45499 3119 f.update({
d493f15c
S
3120 'width': width,
3121 'height': height,
3122 'tbr': tbr,
3123 'format_id': s_attr.get('label') or s_attr.get('title'),
1ed45499 3124 })
520251c0
YCH
3125 f.update(formats[0])
3126 media_info['formats'].append(f)
3127 else:
3128 media_info['formats'].extend(formats)
59bbe491 3129 for track_tag in re.findall(r'<track[^>]+>', media_content):
3130 track_attributes = extract_attributes(track_tag)
3131 kind = track_attributes.get('kind')
5968d7d2 3132 if not kind or kind in ('subtitles', 'captions'):
f856816b 3133 src = strip_or_none(track_attributes.get('src'))
59bbe491 3134 if not src:
3135 continue
3136 lang = track_attributes.get('srclang') or track_attributes.get('lang') or track_attributes.get('label')
3137 media_info['subtitles'].setdefault(lang, []).append({
3138 'url': absolute_url(src),
3139 })
5e8e2fa5
S
3140 for f in media_info['formats']:
3141 f.setdefault('http_headers', {})['Referer'] = base_url
5968d7d2 3142 if media_info['formats'] or media_info['subtitles']:
59bbe491 3143 entries.append(media_info)
3144 return entries
3145
f6a1d69a
F
3146 def _extract_akamai_formats(self, *args, **kwargs):
3147 fmts, subs = self._extract_akamai_formats_and_subtitles(*args, **kwargs)
3148 if subs:
b5ae35ee 3149 self._report_ignoring_subs('akamai')
f6a1d69a
F
3150 return fmts
3151
3152 def _extract_akamai_formats_and_subtitles(self, manifest_url, video_id, hosts={}):
29f7c58a 3153 signed = 'hdnea=' in manifest_url
3154 if not signed:
3155 # https://learn.akamai.com/en-us/webhelp/media-services-on-demand/stream-packaging-user-guide/GUID-BE6C0F73-1E06-483B-B0EA-57984B91B7F9.html
3156 manifest_url = re.sub(
3157 r'(?:b=[\d,-]+|(?:__a__|attributes)=off|__b__=\d+)&?',
3158 '', manifest_url).strip('?')
3159
c7c43a93 3160 formats = []
f6a1d69a 3161 subtitles = {}
70c5802b 3162
e71a4509 3163 hdcore_sign = 'hdcore=3.7.0'
ff6f9a67 3164 f4m_url = re.sub(r'(https?://[^/]+)/i/', r'\1/z/', manifest_url).replace('/master.m3u8', '/manifest.f4m')
c4251b9a
RA
3165 hds_host = hosts.get('hds')
3166 if hds_host:
3167 f4m_url = re.sub(r'(https?://)[^/]+', r'\1' + hds_host, f4m_url)
e71a4509
RA
3168 if 'hdcore=' not in f4m_url:
3169 f4m_url += ('&' if '?' in f4m_url else '?') + hdcore_sign
3170 f4m_formats = self._extract_f4m_formats(
3171 f4m_url, video_id, f4m_id='hds', fatal=False)
3172 for entry in f4m_formats:
3173 entry.update({'extra_param_to_segment_url': hdcore_sign})
3174 formats.extend(f4m_formats)
70c5802b 3175
c4251b9a
RA
3176 m3u8_url = re.sub(r'(https?://[^/]+)/z/', r'\1/i/', manifest_url).replace('/manifest.f4m', '/master.m3u8')
3177 hls_host = hosts.get('hls')
3178 if hls_host:
3179 m3u8_url = re.sub(r'(https?://)[^/]+', r'\1' + hls_host, m3u8_url)
f6a1d69a 3180 m3u8_formats, m3u8_subtitles = self._extract_m3u8_formats_and_subtitles(
c7c43a93 3181 m3u8_url, video_id, 'mp4', 'm3u8_native',
29f7c58a 3182 m3u8_id='hls', fatal=False)
3183 formats.extend(m3u8_formats)
f6a1d69a 3184 subtitles = self._merge_subtitles(subtitles, m3u8_subtitles)
70c5802b 3185
3186 http_host = hosts.get('http')
29f7c58a 3187 if http_host and m3u8_formats and not signed:
3188 REPL_REGEX = r'https?://[^/]+/i/([^,]+),([^/]+),([^/]+)\.csmil/.+'
70c5802b 3189 qualities = re.match(REPL_REGEX, m3u8_url).group(2).split(',')
3190 qualities_length = len(qualities)
29f7c58a 3191 if len(m3u8_formats) in (qualities_length, qualities_length + 1):
70c5802b 3192 i = 0
29f7c58a 3193 for f in m3u8_formats:
3194 if f['vcodec'] != 'none':
70c5802b 3195 for protocol in ('http', 'https'):
3196 http_f = f.copy()
3197 del http_f['manifest_url']
3198 http_url = re.sub(
86e5f3ed 3199 REPL_REGEX, protocol + fr'://{http_host}/\g<1>{qualities[i]}\3', f['url'])
70c5802b 3200 http_f.update({
3201 'format_id': http_f['format_id'].replace('hls-', protocol + '-'),
3202 'url': http_url,
3203 'protocol': protocol,
3204 })
29f7c58a 3205 formats.append(http_f)
70c5802b 3206 i += 1
70c5802b 3207
f6a1d69a 3208 return formats, subtitles
c7c43a93 3209
6ad02195 3210 def _extract_wowza_formats(self, url, video_id, m3u8_entry_protocol='m3u8_native', skip_protocols=[]):
14f25df2 3211 query = urllib.parse.urlparse(url).query
6ad02195 3212 url = re.sub(r'/(?:manifest|playlist|jwplayer)\.(?:m3u8|f4m|mpd|smil)', '', url)
240f2622
S
3213 mobj = re.search(
3214 r'(?:(?:http|rtmp|rtsp)(?P<s>s)?:)?(?P<url>//[^?]+)', url)
3215 url_base = mobj.group('url')
3216 http_base_url = '%s%s:%s' % ('http', mobj.group('s') or '', url_base)
6ad02195 3217 formats = []
044eeb14
S
3218
3219 def manifest_url(manifest):
86e5f3ed 3220 m_url = f'{http_base_url}/{manifest}'
044eeb14
S
3221 if query:
3222 m_url += '?%s' % query
3223 return m_url
3224
6ad02195
RA
3225 if 'm3u8' not in skip_protocols:
3226 formats.extend(self._extract_m3u8_formats(
044eeb14 3227 manifest_url('playlist.m3u8'), video_id, 'mp4',
6ad02195
RA
3228 m3u8_entry_protocol, m3u8_id='hls', fatal=False))
3229 if 'f4m' not in skip_protocols:
3230 formats.extend(self._extract_f4m_formats(
044eeb14 3231 manifest_url('manifest.f4m'),
6ad02195 3232 video_id, f4m_id='hds', fatal=False))
0384932e
RA
3233 if 'dash' not in skip_protocols:
3234 formats.extend(self._extract_mpd_formats(
044eeb14 3235 manifest_url('manifest.mpd'),
0384932e 3236 video_id, mpd_id='dash', fatal=False))
6ad02195 3237 if re.search(r'(?:/smil:|\.smil)', url_base):
6ad02195
RA
3238 if 'smil' not in skip_protocols:
3239 rtmp_formats = self._extract_smil_formats(
044eeb14 3240 manifest_url('jwplayer.smil'),
6ad02195
RA
3241 video_id, fatal=False)
3242 for rtmp_format in rtmp_formats:
3243 rtsp_format = rtmp_format.copy()
3244 rtsp_format['url'] = '%s/%s' % (rtmp_format['url'], rtmp_format['play_path'])
3245 del rtsp_format['play_path']
3246 del rtsp_format['ext']
3247 rtsp_format.update({
3248 'url': rtsp_format['url'].replace('rtmp://', 'rtsp://'),
3249 'format_id': rtmp_format['format_id'].replace('rtmp', 'rtsp'),
3250 'protocol': 'rtsp',
3251 })
3252 formats.extend([rtmp_format, rtsp_format])
3253 else:
3254 for protocol in ('rtmp', 'rtsp'):
3255 if protocol not in skip_protocols:
3256 formats.append({
86e5f3ed 3257 'url': f'{protocol}:{url_base}',
6ad02195
RA
3258 'format_id': protocol,
3259 'protocol': protocol,
3260 })
3261 return formats
3262
c73e330e 3263 def _find_jwplayer_data(self, webpage, video_id=None, transform_source=js_to_json):
a4a554a7 3264 mobj = re.search(
32a84bcf 3265 r'''(?s)jwplayer\s*\(\s*(?P<q>'|")(?!(?P=q)).+(?P=q)\s*\)(?!</script>).*?\.\s*setup\s*\(\s*(?P<options>(?:\([^)]*\)|[^)])+)\s*\)''',
a4a554a7
YCH
3266 webpage)
3267 if mobj:
c73e330e
RU
3268 try:
3269 jwplayer_data = self._parse_json(mobj.group('options'),
3270 video_id=video_id,
3271 transform_source=transform_source)
3272 except ExtractorError:
3273 pass
3274 else:
3275 if isinstance(jwplayer_data, dict):
3276 return jwplayer_data
a4a554a7
YCH
3277
3278 def _extract_jwplayer_data(self, webpage, video_id, *args, **kwargs):
c73e330e
RU
3279 jwplayer_data = self._find_jwplayer_data(
3280 webpage, video_id, transform_source=js_to_json)
a4a554a7
YCH
3281 return self._parse_jwplayer_data(
3282 jwplayer_data, video_id, *args, **kwargs)
3283
3284 def _parse_jwplayer_data(self, jwplayer_data, video_id=None, require_title=True,
3285 m3u8_id=None, mpd_id=None, rtmp_params=None, base_url=None):
a4a554a7 3286 entries = []
32a84bcf
SS
3287 if not isinstance(jwplayer_data, dict):
3288 return entries
a4a554a7 3289
32a84bcf
SS
3290 playlist_items = jwplayer_data.get('playlist')
3291 # JWPlayer backward compatibility: single playlist item/flattened playlists
a4a554a7 3292 # https://github.com/jwplayer/jwplayer/blob/v7.7.0/src/js/playlist/playlist.js#L10
32a84bcf
SS
3293 # https://github.com/jwplayer/jwplayer/blob/v7.4.3/src/js/api/config.js#L81-L96
3294 if not isinstance(playlist_items, list):
3295 playlist_items = (playlist_items or jwplayer_data, )
a4a554a7 3296
32a84bcf
SS
3297 for video_data in playlist_items:
3298 if not isinstance(video_data, dict):
3299 continue
a4a554a7
YCH
3300 # JWPlayer backward compatibility: flattened sources
3301 # https://github.com/jwplayer/jwplayer/blob/v7.4.3/src/js/playlist/item.js#L29-L35
3302 if 'sources' not in video_data:
3303 video_data['sources'] = [video_data]
3304
3305 this_video_id = video_id or video_data['mediaid']
3306
1a2192cb
S
3307 formats = self._parse_jwplayer_formats(
3308 video_data['sources'], video_id=this_video_id, m3u8_id=m3u8_id,
3309 mpd_id=mpd_id, rtmp_params=rtmp_params, base_url=base_url)
a4a554a7
YCH
3310
3311 subtitles = {}
3312 tracks = video_data.get('tracks')
3313 if tracks and isinstance(tracks, list):
3314 for track in tracks:
96a2daa1
S
3315 if not isinstance(track, dict):
3316 continue
f4b74272 3317 track_kind = track.get('kind')
14f25df2 3318 if not track_kind or not isinstance(track_kind, str):
f4b74272
S
3319 continue
3320 if track_kind.lower() not in ('captions', 'subtitles'):
a4a554a7
YCH
3321 continue
3322 track_url = urljoin(base_url, track.get('file'))
3323 if not track_url:
3324 continue
3325 subtitles.setdefault(track.get('label') or 'en', []).append({
3326 'url': self._proto_relative_url(track_url)
3327 })
3328
50d808f5 3329 entry = {
a4a554a7 3330 'id': this_video_id,
50d808f5 3331 'title': unescapeHTML(video_data['title'] if require_title else video_data.get('title')),
f81dd65b 3332 'description': clean_html(video_data.get('description')),
6945b9e7 3333 'thumbnail': urljoin(base_url, self._proto_relative_url(video_data.get('image'))),
a4a554a7
YCH
3334 'timestamp': int_or_none(video_data.get('pubdate')),
3335 'duration': float_or_none(jwplayer_data.get('duration') or video_data.get('duration')),
3336 'subtitles': subtitles,
32a84bcf
SS
3337 'alt_title': clean_html(video_data.get('subtitle')), # attributes used e.g. by Tele5 ...
3338 'genre': clean_html(video_data.get('genre')),
3339 'channel': clean_html(dict_get(video_data, ('category', 'channel'))),
3340 'season_number': int_or_none(video_data.get('season')),
3341 'episode_number': int_or_none(video_data.get('episode')),
3342 'release_year': int_or_none(video_data.get('releasedate')),
3343 'age_limit': int_or_none(video_data.get('age_restriction')),
50d808f5
RA
3344 }
3345 # https://github.com/jwplayer/jwplayer/blob/master/src/js/utils/validator.js#L32
3346 if len(formats) == 1 and re.search(r'^(?:http|//).*(?:youtube\.com|youtu\.be)/.+', formats[0]['url']):
3347 entry.update({
3348 '_type': 'url_transparent',
3349 'url': formats[0]['url'],
3350 })
3351 else:
50d808f5
RA
3352 entry['formats'] = formats
3353 entries.append(entry)
a4a554a7
YCH
3354 if len(entries) == 1:
3355 return entries[0]
3356 else:
3357 return self.playlist_result(entries)
3358
ed0cf9b3
S
3359 def _parse_jwplayer_formats(self, jwplayer_sources_data, video_id=None,
3360 m3u8_id=None, mpd_id=None, rtmp_params=None, base_url=None):
32a84bcf 3361 urls = set()
ed0cf9b3 3362 formats = []
1a2192cb 3363 for source in jwplayer_sources_data:
0a268c6e
S
3364 if not isinstance(source, dict):
3365 continue
6945b9e7
RA
3366 source_url = urljoin(
3367 base_url, self._proto_relative_url(source.get('file')))
3368 if not source_url or source_url in urls:
bf1b87cd 3369 continue
32a84bcf 3370 urls.add(source_url)
ed0cf9b3
S
3371 source_type = source.get('type') or ''
3372 ext = mimetype2ext(source_type) or determine_ext(source_url)
32a84bcf 3373 if source_type == 'hls' or ext == 'm3u8' or 'format=m3u8-aapl' in source_url:
ed0cf9b3 3374 formats.extend(self._extract_m3u8_formats(
0236cd0d
S
3375 source_url, video_id, 'mp4', entry_protocol='m3u8_native',
3376 m3u8_id=m3u8_id, fatal=False))
32a84bcf 3377 elif source_type == 'dash' or ext == 'mpd' or 'format=mpd-time-csf' in source_url:
ed0cf9b3
S
3378 formats.extend(self._extract_mpd_formats(
3379 source_url, video_id, mpd_id=mpd_id, fatal=False))
b51dc9db
S
3380 elif ext == 'smil':
3381 formats.extend(self._extract_smil_formats(
3382 source_url, video_id, fatal=False))
ed0cf9b3 3383 # https://github.com/jwplayer/jwplayer/blob/master/src/js/providers/default.js#L67
0236cd0d
S
3384 elif source_type.startswith('audio') or ext in (
3385 'oga', 'aac', 'mp3', 'mpeg', 'vorbis'):
ed0cf9b3
S
3386 formats.append({
3387 'url': source_url,
3388 'vcodec': 'none',
3389 'ext': ext,
3390 })
3391 else:
32a84bcf 3392 format_id = str_or_none(source.get('label'))
ed0cf9b3 3393 height = int_or_none(source.get('height'))
32a84bcf 3394 if height is None and format_id:
ed0cf9b3 3395 # Often no height is provided but there is a label in
0236cd0d 3396 # format like "1080p", "720p SD", or 1080.
32a84bcf 3397 height = parse_resolution(format_id).get('height')
ed0cf9b3
S
3398 a_format = {
3399 'url': source_url,
3400 'width': int_or_none(source.get('width')),
3401 'height': height,
d3a3d7f0 3402 'tbr': int_or_none(source.get('bitrate'), scale=1000),
3403 'filesize': int_or_none(source.get('filesize')),
ed0cf9b3 3404 'ext': ext,
32a84bcf 3405 'format_id': format_id
ed0cf9b3
S
3406 }
3407 if source_url.startswith('rtmp'):
3408 a_format['ext'] = 'flv'
ed0cf9b3
S
3409 # See com/longtailvideo/jwplayer/media/RTMPMediaProvider.as
3410 # of jwplayer.flash.swf
3411 rtmp_url_parts = re.split(
3412 r'((?:mp4|mp3|flv):)', source_url, 1)
3413 if len(rtmp_url_parts) == 3:
3414 rtmp_url, prefix, play_path = rtmp_url_parts
3415 a_format.update({
3416 'url': rtmp_url,
3417 'play_path': prefix + play_path,
3418 })
3419 if rtmp_params:
3420 a_format.update(rtmp_params)
3421 formats.append(a_format)
3422 return formats
3423
f4b1c7ad 3424 def _live_title(self, name):
39ca3b5c 3425 self._downloader.deprecation_warning('yt_dlp.InfoExtractor._live_title is deprecated and does not work as expected')
3426 return name
f4b1c7ad 3427
b14f3a4c
PH
3428 def _int(self, v, name, fatal=False, **kwargs):
3429 res = int_or_none(v, **kwargs)
b14f3a4c 3430 if res is None:
86e5f3ed 3431 msg = f'Failed to extract {name}: Could not parse value {v!r}'
b14f3a4c
PH
3432 if fatal:
3433 raise ExtractorError(msg)
3434 else:
6a39ee13 3435 self.report_warning(msg)
b14f3a4c
PH
3436 return res
3437
3438 def _float(self, v, name, fatal=False, **kwargs):
3439 res = float_or_none(v, **kwargs)
3440 if res is None:
86e5f3ed 3441 msg = f'Failed to extract {name}: Could not parse value {v!r}'
b14f3a4c
PH
3442 if fatal:
3443 raise ExtractorError(msg)
3444 else:
6a39ee13 3445 self.report_warning(msg)
b14f3a4c
PH
3446 return res
3447
40e41780
TF
3448 def _set_cookie(self, domain, name, value, expire_time=None, port=None,
3449 path='/', secure=False, discard=False, rest={}, **kwargs):
ac668111 3450 cookie = http.cookiejar.Cookie(
4ed2d7b7 3451 0, name, value, port, port is not None, domain, True,
40e41780
TF
3452 domain.startswith('.'), path, True, secure, expire_time,
3453 discard, None, None, rest)
9809740b 3454 self.cookiejar.set_cookie(cookie)
42939b61 3455
799207e8 3456 def _get_cookies(self, url):
ac668111 3457 """ Return a http.cookies.SimpleCookie with the cookies for the url """
b87e01c1 3458 return LenientSimpleCookie(self._downloader.cookiejar.get_cookie_header(url))
799207e8 3459
e3c1266f 3460 def _apply_first_set_cookie_header(self, url_handle, cookie):
ce2fe4c0
S
3461 """
3462 Apply first Set-Cookie header instead of the last. Experimental.
3463
3464 Some sites (e.g. [1-3]) may serve two cookies under the same name
3465 in Set-Cookie header and expect the first (old) one to be set rather
3466 than second (new). However, as of RFC6265 the newer one cookie
3467 should be set into cookie store what actually happens.
3468 We will workaround this issue by resetting the cookie to
3469 the first one manually.
3470 1. https://new.vk.com/
3471 2. https://github.com/ytdl-org/youtube-dl/issues/9841#issuecomment-227871201
3472 3. https://learning.oreilly.com/
3473 """
e3c1266f
S
3474 for header, cookies in url_handle.headers.items():
3475 if header.lower() != 'set-cookie':
3476 continue
cfb0511d 3477 cookies = cookies.encode('iso-8859-1').decode('utf-8')
e3c1266f
S
3478 cookie_value = re.search(
3479 r'%s=(.+?);.*?\b[Dd]omain=(.+?)(?:[,;]|$)' % cookie, cookies)
3480 if cookie_value:
3481 value, domain = cookie_value.groups()
3482 self._set_cookie(domain, cookie, value)
3483 break
3484
82d02080 3485 @classmethod
3486 def get_testcases(cls, include_onlymatching=False):
6368e2e6 3487 # Do not look in super classes
3488 t = vars(cls).get('_TEST')
05900629 3489 if t:
82d02080 3490 assert not hasattr(cls, '_TESTS'), f'{cls.ie_key()}IE has _TEST and _TESTS'
05900629
PH
3491 tests = [t]
3492 else:
6368e2e6 3493 tests = vars(cls).get('_TESTS', [])
05900629
PH
3494 for t in tests:
3495 if not include_onlymatching and t.get('only_matching', False):
3496 continue
82d02080 3497 t['name'] = cls.ie_key()
05900629 3498 yield t
e756f45b
M
3499 if getattr(cls, '__wrapped__', None):
3500 yield from cls.__wrapped__.get_testcases(include_onlymatching)
05900629 3501
f2e8dbcc 3502 @classmethod
3503 def get_webpage_testcases(cls):
6368e2e6 3504 tests = vars(cls).get('_WEBPAGE_TESTS', [])
f2e8dbcc 3505 for t in tests:
3506 t['name'] = cls.ie_key()
e756f45b
M
3507 yield t
3508 if getattr(cls, '__wrapped__', None):
3509 yield from cls.__wrapped__.get_webpage_testcases()
f2e8dbcc 3510
6368e2e6 3511 @classproperty(cache=True)
24146491 3512 def age_limit(cls):
3513 """Get age limit from the testcases"""
3514 return max(traverse_obj(
f2e8dbcc 3515 (*cls.get_testcases(include_onlymatching=False), *cls.get_webpage_testcases()),
24146491 3516 (..., (('playlist', 0), None), 'info_dict', 'age_limit')) or [0])
3517
171a31db 3518 @classproperty(cache=True)
3519 def _RETURN_TYPE(cls):
3520 """What the extractor returns: "video", "playlist", "any", or None (Unknown)"""
3521 tests = tuple(cls.get_testcases(include_onlymatching=False))
3522 if not tests:
3523 return None
3524 elif not any(k.startswith('playlist') for test in tests for k in test):
3525 return 'video'
3526 elif all(any(k.startswith('playlist') for k in test) for test in tests):
3527 return 'playlist'
3528 return 'any'
3529
3530 @classmethod
3531 def is_single_video(cls, url):
3532 """Returns whether the URL is of a single video, None if unknown"""
baa922b5 3533 if cls.suitable(url):
3534 return {'video': True, 'playlist': False}.get(cls._RETURN_TYPE)
171a31db 3535
82d02080 3536 @classmethod
3537 def is_suitable(cls, age_limit):
24146491 3538 """Test whether the extractor is generally suitable for the given age limit"""
3539 return not age_restricted(cls.age_limit, age_limit)
05900629 3540
82d02080 3541 @classmethod
3542 def description(cls, *, markdown=True, search_examples=None):
8dcce6a8 3543 """Description of the extractor"""
3544 desc = ''
82d02080 3545 if cls._NETRC_MACHINE:
8dcce6a8 3546 if markdown:
5b28cef7 3547 desc += f' [*{cls._NETRC_MACHINE}*](## "netrc machine")'
8dcce6a8 3548 else:
82d02080 3549 desc += f' [{cls._NETRC_MACHINE}]'
3550 if cls.IE_DESC is False:
8dcce6a8 3551 desc += ' [HIDDEN]'
82d02080 3552 elif cls.IE_DESC:
3553 desc += f' {cls.IE_DESC}'
3554 if cls.SEARCH_KEY:
08e29b9f 3555 desc += f'{";" if cls.IE_DESC else ""} "{cls.SEARCH_KEY}:" prefix'
8dcce6a8 3556 if search_examples:
3557 _COUNTS = ('', '5', '10', 'all')
62b58c09 3558 desc += f' (e.g. "{cls.SEARCH_KEY}{random.choice(_COUNTS)}:{random.choice(search_examples)}")'
82d02080 3559 if not cls.working():
8dcce6a8 3560 desc += ' (**Currently broken**)' if markdown else ' (Currently broken)'
3561
46d09f87 3562 # Escape emojis. Ref: https://github.com/github/markup/issues/1153
3563 name = (' - **%s**' % re.sub(r':(\w+:)', ':\u200B\\g<1>', cls.IE_NAME)) if markdown else cls.IE_NAME
8dcce6a8 3564 return f'{name}:{desc}' if desc else name
3565
a504ced0 3566 def extract_subtitles(self, *args, **kwargs):
a06916d9 3567 if (self.get_param('writesubtitles', False)
3568 or self.get_param('listsubtitles')):
9868ea49
JMF
3569 return self._get_subtitles(*args, **kwargs)
3570 return {}
a504ced0
JMF
3571
3572 def _get_subtitles(self, *args, **kwargs):
611c1dd9 3573 raise NotImplementedError('This method must be implemented by subclasses')
a504ced0 3574
0cf643b2
M
3575 class CommentsDisabled(Exception):
3576 """Raise in _get_comments if comments are disabled for the video"""
3577
a2160aa4 3578 def extract_comments(self, *args, **kwargs):
3579 if not self.get_param('getcomments'):
3580 return None
3581 generator = self._get_comments(*args, **kwargs)
3582
3583 def extractor():
3584 comments = []
d2b2fca5 3585 interrupted = True
a2160aa4 3586 try:
3587 while True:
3588 comments.append(next(generator))
a2160aa4 3589 except StopIteration:
3590 interrupted = False
d2b2fca5 3591 except KeyboardInterrupt:
3592 self.to_screen('Interrupted by user')
0cf643b2
M
3593 except self.CommentsDisabled:
3594 return {'comments': None, 'comment_count': None}
d2b2fca5 3595 except Exception as e:
3596 if self.get_param('ignoreerrors') is not True:
3597 raise
3598 self._downloader.report_error(e)
a2160aa4 3599 comment_count = len(comments)
3600 self.to_screen(f'Extracted {comment_count} comments')
3601 return {
3602 'comments': comments,
3603 'comment_count': None if interrupted else comment_count
3604 }
3605 return extractor
3606
3607 def _get_comments(self, *args, **kwargs):
3608 raise NotImplementedError('This method must be implemented by subclasses')
3609
912e0b7e
YCH
3610 @staticmethod
3611 def _merge_subtitle_items(subtitle_list1, subtitle_list2):
a825ffbf 3612 """ Merge subtitle items for one language. Items with duplicated URLs/data
912e0b7e 3613 will be dropped. """
86e5f3ed 3614 list1_data = {(item.get('url'), item.get('data')) for item in subtitle_list1}
912e0b7e 3615 ret = list(subtitle_list1)
a44ca5a4 3616 ret.extend(item for item in subtitle_list2 if (item.get('url'), item.get('data')) not in list1_data)
912e0b7e
YCH
3617 return ret
3618
3619 @classmethod
46890374 3620 def _merge_subtitles(cls, *dicts, target=None):
19bb3920 3621 """ Merge subtitle dictionaries, language by language. """
19bb3920
F
3622 if target is None:
3623 target = {}
3624 for d in dicts:
3625 for lang, subs in d.items():
3626 target[lang] = cls._merge_subtitle_items(target.get(lang, []), subs)
3627 return target
912e0b7e 3628
360e1ca5 3629 def extract_automatic_captions(self, *args, **kwargs):
a06916d9 3630 if (self.get_param('writeautomaticsub', False)
3631 or self.get_param('listsubtitles')):
9868ea49
JMF
3632 return self._get_automatic_captions(*args, **kwargs)
3633 return {}
360e1ca5
JMF
3634
3635 def _get_automatic_captions(self, *args, **kwargs):
611c1dd9 3636 raise NotImplementedError('This method must be implemented by subclasses')
360e1ca5 3637
2762dbb1 3638 @functools.cached_property
24146491 3639 def _cookies_passed(self):
3640 """Whether cookies have been passed to YoutubeDL"""
3641 return self.get_param('cookiefile') is not None or self.get_param('cookiesfrombrowser') is not None
3642
d77ab8e2 3643 def mark_watched(self, *args, **kwargs):
1813a6cc 3644 if not self.get_param('mark_watched', False):
3645 return
24146491 3646 if self.supports_login() and self._get_login_info()[0] is not None or self._cookies_passed:
d77ab8e2
S
3647 self._mark_watched(*args, **kwargs)
3648
3649 def _mark_watched(self, *args, **kwargs):
3650 raise NotImplementedError('This method must be implemented by subclasses')
3651
38cce791
YCH
3652 def geo_verification_headers(self):
3653 headers = {}
a06916d9 3654 geo_verification_proxy = self.get_param('geo_verification_proxy')
38cce791
YCH
3655 if geo_verification_proxy:
3656 headers['Ytdl-request-proxy'] = geo_verification_proxy
3657 return headers
3658
8f97a15d 3659 @staticmethod
3660 def _generic_id(url):
14f25df2 3661 return urllib.parse.unquote(os.path.splitext(url.rstrip('/').split('/')[-1])[0])
98763ee3 3662
62b8dac4 3663 def _generic_title(self, url='', webpage='', *, default=None):
3664 return (self._og_search_title(webpage, default=None)
3665 or self._html_extract_title(webpage, default=None)
3666 or urllib.parse.unquote(os.path.splitext(url_basename(url))[0])
3667 or default)
98763ee3 3668
22ccd542 3669 def _extract_chapters_helper(self, chapter_list, start_function, title_function, duration, strict=True):
3670 if not duration:
3671 return
3672 chapter_list = [{
3673 'start_time': start_function(chapter),
3674 'title': title_function(chapter),
3675 } for chapter in chapter_list or []]
84ffeb7d 3676 if strict:
3677 warn = self.report_warning
3678 else:
3679 warn = self.write_debug
22ccd542 3680 chapter_list.sort(key=lambda c: c['start_time'] or 0)
3681
3682 chapters = [{'start_time': 0}]
3683 for idx, chapter in enumerate(chapter_list):
3684 if chapter['start_time'] is None:
84ffeb7d 3685 warn(f'Incomplete chapter {idx}')
22ccd542 3686 elif chapters[-1]['start_time'] <= chapter['start_time'] <= duration:
3687 chapters.append(chapter)
3688 elif chapter not in chapters:
84ffeb7d 3689 issue = (f'{chapter["start_time"]} > {duration}' if chapter['start_time'] > duration
3690 else f'{chapter["start_time"]} < {chapters[-1]["start_time"]}')
3691 warn(f'Invalid start time ({issue}) for chapter "{chapter["title"]}"')
22ccd542 3692 return chapters[1:]
3693
3694 def _extract_chapters_from_description(self, description, duration):
3695 duration_re = r'(?:\d+:)?\d{1,2}:\d{2}'
3696 sep_re = r'(?m)^\s*(%s)\b\W*\s(%s)\s*$'
3697 return self._extract_chapters_helper(
3698 re.findall(sep_re % (duration_re, r'.+?'), description or ''),
3699 start_function=lambda x: parse_duration(x[0]), title_function=lambda x: x[1],
3700 duration=duration, strict=False) or self._extract_chapters_helper(
3701 re.findall(sep_re % (r'.+?', duration_re), description or ''),
3702 start_function=lambda x: parse_duration(x[1]), title_function=lambda x: x[0],
3703 duration=duration, strict=False)
3704
c224251a 3705 @staticmethod
b0089e89 3706 def _availability(is_private=None, needs_premium=None, needs_subscription=None, needs_auth=None, is_unlisted=None):
c224251a
M
3707 all_known = all(map(
3708 lambda x: x is not None,
3709 (is_private, needs_premium, needs_subscription, needs_auth, is_unlisted)))
3710 return (
3711 'private' if is_private
3712 else 'premium_only' if needs_premium
3713 else 'subscriber_only' if needs_subscription
3714 else 'needs_auth' if needs_auth
3715 else 'unlisted' if is_unlisted
3716 else 'public' if all_known
3717 else None)
3718
d43de682 3719 def _configuration_arg(self, key, default=NO_DEFAULT, *, ie_key=None, casesense=False):
4bb6b02f 3720 '''
3721 @returns A list of values for the extractor argument given by "key"
3722 or "default" if no such key is present
3723 @param default The default value to return when the key is not present (default: [])
3724 @param casesense When false, the values are converted to lower case
3725 '''
5225df50 3726 ie_key = ie_key if isinstance(ie_key, str) else (ie_key or self).ie_key()
3727 val = traverse_obj(self._downloader.params, ('extractor_args', ie_key.lower(), key))
4bb6b02f 3728 if val is None:
3729 return [] if default is NO_DEFAULT else default
3730 return list(val) if casesense else [x.lower() for x in val]
5d3a0e79 3731
f40ee5e9 3732 def _yes_playlist(self, playlist_id, video_id, smuggled_data=None, *, playlist_label='playlist', video_label='video'):
3733 if not playlist_id or not video_id:
3734 return not video_id
3735
3736 no_playlist = (smuggled_data or {}).get('force_noplaylist')
3737 if no_playlist is not None:
3738 return not no_playlist
3739
3740 video_id = '' if video_id is True else f' {video_id}'
3741 playlist_id = '' if playlist_id is True else f' {playlist_id}'
3742 if self.get_param('noplaylist'):
3743 self.to_screen(f'Downloading just the {video_label}{video_id} because of --no-playlist')
3744 return False
3745 self.to_screen(f'Downloading {playlist_label}{playlist_id} - add --no-playlist to download just the {video_label}{video_id}')
3746 return True
3747
be5c1ae8 3748 def _error_or_warning(self, err, _count=None, _retries=0, *, fatal=True):
8ca48a1a 3749 RetryManager.report_retry(
3750 err, _count or int(fatal), _retries,
3751 info=self.to_screen, warn=self.report_warning, error=None if fatal else self.report_warning,
3752 sleep_func=self.get_param('retry_sleep_functions', {}).get('extractor'))
be5c1ae8 3753
3754 def RetryManager(self, **kwargs):
3755 return RetryManager(self.get_param('extractor_retries', 3), self._error_or_warning, **kwargs)
3756
ade1fa70 3757 def _extract_generic_embeds(self, url, *args, info_dict={}, note='Extracting generic embeds', **kwargs):
3758 display_id = traverse_obj(info_dict, 'display_id', 'id')
3759 self.to_screen(f'{format_field(display_id, None, "%s: ")}{note}')
3760 return self._downloader.get_info_extractor('Generic')._extract_embeds(
3761 smuggle_url(url, {'block_ies': [self.ie_key()]}), *args, **kwargs)
3762
8f97a15d 3763 @classmethod
3764 def extract_from_webpage(cls, ydl, url, webpage):
3765 ie = (cls if isinstance(cls._extract_from_webpage, types.MethodType)
3766 else ydl.get_info_extractor(cls.ie_key()))
f2e8dbcc 3767 for info in ie._extract_from_webpage(url, webpage) or []:
3768 # url = None since we do not want to set (webpage/original)_url
3769 ydl.add_default_extra_info(info, ie, None)
3770 yield info
8f97a15d 3771
3772 @classmethod
3773 def _extract_from_webpage(cls, url, webpage):
3774 for embed_url in orderedSet(
3775 cls._extract_embed_urls(url, webpage) or [], lazy=True):
d2c8aadf 3776 yield cls.url_result(embed_url, None if cls._VALID_URL is False else cls)
8f97a15d 3777
3778 @classmethod
3779 def _extract_embed_urls(cls, url, webpage):
3780 """@returns all the embed urls on the webpage"""
3781 if '_EMBED_URL_RE' not in cls.__dict__:
3782 assert isinstance(cls._EMBED_REGEX, (list, tuple))
3783 for idx, regex in enumerate(cls._EMBED_REGEX):
3784 assert regex.count('(?P<url>') == 1, \
3785 f'{cls.__name__}._EMBED_REGEX[{idx}] must have exactly 1 url group\n\t{regex}'
3786 cls._EMBED_URL_RE = tuple(map(re.compile, cls._EMBED_REGEX))
3787
3788 for regex in cls._EMBED_URL_RE:
3789 for mobj in regex.finditer(webpage):
3790 embed_url = urllib.parse.urljoin(url, unescapeHTML(mobj.group('url')))
3791 if cls._VALID_URL is False or cls.suitable(embed_url):
3792 yield embed_url
3793
3794 class StopExtraction(Exception):
3795 pass
3796
bfd973ec 3797 @classmethod
3798 def _extract_url(cls, webpage): # TODO: Remove
3799 """Only for compatibility with some older extractors"""
3800 return next(iter(cls._extract_embed_urls(None, webpage) or []), None)
3801
2314b4d8 3802 @classmethod
3803 def __init_subclass__(cls, *, plugin_name=None, **kwargs):
3804 if plugin_name:
3805 mro = inspect.getmro(cls)
3806 super_class = cls.__wrapped__ = mro[mro.index(cls) + 1]
e756f45b
M
3807 cls.PLUGIN_NAME, cls.ie_key = plugin_name, super_class.ie_key
3808 cls.IE_NAME = f'{super_class.IE_NAME}+{plugin_name}'
2314b4d8 3809 while getattr(super_class, '__wrapped__', None):
3810 super_class = super_class.__wrapped__
3811 setattr(sys.modules[super_class.__module__], super_class.__name__, cls)
e756f45b 3812 _PLUGIN_OVERRIDES[super_class].append(cls)
2314b4d8 3813
3814 return super().__init_subclass__(**kwargs)
3815
8dbe9899 3816
d6983cb4
PH
3817class SearchInfoExtractor(InfoExtractor):
3818 """
3819 Base class for paged search queries extractors.
10952eb2 3820 They accept URLs in the format _SEARCH_KEY(|all|[0-9]):{query}
96565c7e 3821 Instances should define _SEARCH_KEY and optionally _MAX_RESULTS
d6983cb4
PH
3822 """
3823
96565c7e 3824 _MAX_RESULTS = float('inf')
171a31db 3825 _RETURN_TYPE = 'playlist'
96565c7e 3826
8f97a15d 3827 @classproperty
3828 def _VALID_URL(cls):
d6983cb4
PH
3829 return r'%s(?P<prefix>|[1-9][0-9]*|all):(?P<query>[\s\S]+)' % cls._SEARCH_KEY
3830
d6983cb4 3831 def _real_extract(self, query):
2c4aaadd 3832 prefix, query = self._match_valid_url(query).group('prefix', 'query')
d6983cb4
PH
3833 if prefix == '':
3834 return self._get_n_results(query, 1)
3835 elif prefix == 'all':
3836 return self._get_n_results(query, self._MAX_RESULTS)
3837 else:
3838 n = int(prefix)
3839 if n <= 0:
86e5f3ed 3840 raise ExtractorError(f'invalid download number {n} for query "{query}"')
d6983cb4 3841 elif n > self._MAX_RESULTS:
6a39ee13 3842 self.report_warning('%s returns max %i results (you requested %i)' % (self._SEARCH_KEY, self._MAX_RESULTS, n))
d6983cb4
PH
3843 n = self._MAX_RESULTS
3844 return self._get_n_results(query, n)
3845
3846 def _get_n_results(self, query, n):
cc16383f 3847 """Get a specified number of results for a query.
3848 Either this function or _search_results must be overridden by subclasses """
3849 return self.playlist_result(
3850 itertools.islice(self._search_results(query), 0, None if n == float('inf') else n),
3851 query, query)
3852
3853 def _search_results(self, query):
3854 """Returns an iterator of search results"""
611c1dd9 3855 raise NotImplementedError('This method must be implemented by subclasses')
0f818663 3856
82d02080 3857 @classproperty
3858 def SEARCH_KEY(cls):
3859 return cls._SEARCH_KEY
fe7866d0 3860
3861
3862class UnsupportedURLIE(InfoExtractor):
3863 _VALID_URL = '.*'
3864 _ENABLED = False
3865 IE_DESC = False
3866
3867 def _real_extract(self, url):
3868 raise UnsupportedError(url)
e756f45b
M
3869
3870
3871_PLUGIN_OVERRIDES = collections.defaultdict(list)