]> jfr.im git - yt-dlp.git/blame - yt_dlp/extractor/common.py
[extractor/generic] Remove HEAD request
[yt-dlp.git] / yt_dlp / extractor / common.py
CommitLineData
d6983cb4 1import base64
234416e4 2import collections
ac668111 3import getpass
3ec05685 4import hashlib
54007a45 5import http.client
6import http.cookiejar
7import http.cookies
cc16383f 8import itertools
3d3538e4 9import json
f8271158 10import math
4094b6e3 11import netrc
d6983cb4 12import os
773f291d 13import random
d6983cb4 14import sys
4094b6e3 15import time
14f25df2 16import urllib.parse
ac668111 17import urllib.request
f8271158 18import xml.etree.ElementTree
d6983cb4 19
c487cf00 20from ..compat import functools, re # isort: split
14f25df2 21from ..compat import compat_etree_fromstring, compat_expanduser, compat_os_name
eb8a4433 22from ..downloader import FileDownloader
f8271158 23from ..downloader.f4m import get_base_url, remove_encrypted_media
8c25f81b 24from ..utils import (
f8271158 25 JSON_LD_RE,
26 NO_DEFAULT,
27 ExtractorError,
28 GeoRestrictedError,
29 GeoUtils,
b7c47b74 30 LenientJSONDecoder,
f8271158 31 RegexNotFoundError,
32 UnsupportedError,
05900629 33 age_restricted,
02dc0a36 34 base_url,
08f2a92c 35 bug_reports_message,
82d02080 36 classproperty,
d6983cb4 37 clean_html,
70f0f5a8 38 determine_ext,
46b18f23 39 determine_protocol,
d493f15c 40 dict_get,
42676437 41 encode_data_uri,
9b9c5355 42 error_to_compat_str,
46b18f23 43 extract_attributes,
90137ca4 44 filter_dict,
97f4aecf 45 fix_xml_ampersands,
b14f3a4c 46 float_or_none,
b868936c 47 format_field,
31bb8d3f 48 int_or_none,
34921b43 49 join_nonempty,
a4a554a7 50 js_to_json,
46b18f23 51 mimetype2ext,
3158150c 52 network_exceptions,
46b18f23 53 orderedSet,
d493f15c 54 parse_bitrate,
46b18f23
JH
55 parse_codecs,
56 parse_duration,
4ca2a3cf 57 parse_iso8601,
46b18f23 58 parse_m3u8_attributes,
d493f15c 59 parse_resolution,
46b18f23 60 sanitize_filename,
b868936c 61 sanitized_Request,
d493f15c 62 str_or_none,
ce5b9040 63 str_to_int,
f856816b 64 strip_or_none,
5d3a0e79 65 traverse_obj,
47046464 66 try_call,
ffa89477 67 try_get,
f38de77f 68 unescapeHTML,
647eab45 69 unified_strdate,
6b3a3098 70 unified_timestamp,
46b18f23 71 update_Request,
09d02ea4 72 update_url_query,
a107193e 73 url_basename,
bebef109 74 url_or_none,
b868936c 75 urljoin,
6606817a 76 variadic,
a6571f10 77 xpath_element,
8d6765cf
S
78 xpath_text,
79 xpath_with_ns,
d6983cb4 80)
c342041f 81
d6983cb4 82
86e5f3ed 83class InfoExtractor:
d6983cb4
PH
84 """Information Extractor class.
85
86 Information extractors are the classes that, given a URL, extract
87 information about the video (or videos) the URL refers to. This
88 information includes the real video URL, the video title, author and
89 others. The information is stored in a dictionary which is then
5d380852 90 passed to the YoutubeDL. The YoutubeDL processes this
d6983cb4
PH
91 information possibly downloading the video to the file system, among
92 other possible outcomes.
93
cf0649f8 94 The type field determines the type of the result.
fed5d032
PH
95 By far the most common value (and the default if _type is missing) is
96 "video", which indicates a single video.
97
98 For a video, the dictionaries must include the following fields:
d6983cb4
PH
99
100 id: Video identifier.
d4736fdb 101 title: Video title, unescaped. Set to an empty string if video has
102 no title as opposed to "None" which signifies that the
103 extractor failed to obtain a title
d67b0b15 104
f49d89ee 105 Additionally, it must contain either a formats entry or a url one:
d67b0b15 106
f49d89ee
PH
107 formats: A list of dictionaries for each format available, ordered
108 from worst to best quality.
109
110 Potential fields:
c790e93a
S
111 * url The mandatory URL representing the media:
112 for plain file media - HTTP URL of this file,
113 for RTMP - RTMP URL,
114 for HLS - URL of the M3U8 media playlist,
115 for HDS - URL of the F4M manifest,
79d2077e
S
116 for DASH
117 - HTTP URL to plain file media (in case of
118 unfragmented media)
119 - URL of the MPD manifest or base URL
120 representing the media if MPD manifest
8ed7a233 121 is parsed from a string (in case of
79d2077e 122 fragmented media)
c790e93a 123 for MSS - URL of the ISM manifest.
86f4d14f
S
124 * manifest_url
125 The URL of the manifest file in case of
c790e93a
S
126 fragmented media:
127 for HLS - URL of the M3U8 master playlist,
128 for HDS - URL of the F4M manifest,
129 for DASH - URL of the MPD manifest,
130 for MSS - URL of the ISM manifest.
a44ca5a4 131 * manifest_stream_number (For internal use only)
132 The index of the stream in the manifest file
10952eb2 133 * ext Will be calculated from URL if missing
d67b0b15
PH
134 * format A human-readable description of the format
135 ("mp4 container with h264/opus").
136 Calculated from the format_id, width, height.
137 and format_note fields if missing.
138 * format_id A short description of the format
5d4f3985
PH
139 ("mp4_h264_opus" or "19").
140 Technically optional, but strongly recommended.
d67b0b15
PH
141 * format_note Additional info about the format
142 ("3D" or "DASH video")
143 * width Width of the video, if known
144 * height Height of the video, if known
f49d89ee 145 * resolution Textual description of width and height
176f1866 146 * dynamic_range The dynamic range of the video. One of:
147 "SDR" (None), "HDR10", "HDR10+, "HDR12", "HLG, "DV"
7217e148 148 * tbr Average bitrate of audio and video in KBit/s
d67b0b15
PH
149 * abr Average audio bitrate in KBit/s
150 * acodec Name of the audio codec in use
dd27fd17 151 * asr Audio sampling rate in Hertz
d67b0b15 152 * vbr Average video bitrate in KBit/s
fbb21cf5 153 * fps Frame rate
d67b0b15 154 * vcodec Name of the video codec in use
1394ce65 155 * container Name of the container format
d67b0b15 156 * filesize The number of bytes, if known in advance
9732d77e 157 * filesize_approx An estimate for the number of bytes
d67b0b15 158 * player_url SWF Player URL (used for rtmpdump).
c7deaa4c 159 * protocol The protocol that will be used for the actual
adbc4ec4
THD
160 download, lower-case. One of "http", "https" or
161 one of the protocols defined in downloader.PROTOCOL_MAP
c58c2d63
S
162 * fragment_base_url
163 Base URL for fragments. Each fragment's path
164 value (if present) will be relative to
165 this URL.
166 * fragments A list of fragments of a fragmented media.
167 Each fragment entry must contain either an url
168 or a path. If an url is present it should be
169 considered by a client. Otherwise both path and
170 fragment_base_url must be present. Here is
171 the list of all potential fields:
172 * "url" - fragment's URL
173 * "path" - fragment's path relative to
174 fragment_base_url
a0d5077c
S
175 * "duration" (optional, int or float)
176 * "filesize" (optional, int)
adbc4ec4
THD
177 * is_from_start Is a live format that can be downloaded
178 from the start. Boolean
f49d89ee 179 * preference Order number of this format. If this field is
08d13955 180 present and not None, the formats get sorted
38d63d84 181 by this field, regardless of all other values.
f49d89ee
PH
182 -1 for default (order by other properties),
183 -2 or smaller for less than default.
e65566a9
PH
184 < -1000 to hide the format (if there is
185 another one which is strictly better)
32f90364
PH
186 * language Language code, e.g. "de" or "en-US".
187 * language_preference Is this in the language mentioned in
188 the URL?
aff2f4f4
PH
189 10 if it's what the URL is about,
190 -1 for default (don't know),
191 -10 otherwise, other values reserved for now.
5d73273f
PH
192 * quality Order number of the video quality of this
193 format, irrespective of the file format.
194 -1 for default (order by other properties),
195 -2 or smaller for less than default.
c64ed2a3
PH
196 * source_preference Order number for this video source
197 (quality takes higher priority)
198 -1 for default (order by other properties),
199 -2 or smaller for less than default.
d769be6c
PH
200 * http_headers A dictionary of additional HTTP headers
201 to add to the request.
6271f1ca 202 * stretched_ratio If given and not 1, indicates that the
3dee7826
PH
203 video's pixels are not square.
204 width : height ratio as float.
205 * no_resume The server does not support resuming the
206 (HTTP or RTMP) download. Boolean.
88acdbc2 207 * has_drm The format has DRM and cannot be downloaded. Boolean
0a5a191a 208 * downloader_options A dictionary of downloader options
209 (For internal use only)
210 * http_chunk_size Chunk size for HTTP downloads
211 * ffmpeg_args Extra arguments for ffmpeg downloader
3b1fe47d 212 RTMP formats can also have the additional fields: page_url,
213 app, play_path, tc_url, flash_version, rtmp_live, rtmp_conn,
214 rtmp_protocol, rtmp_real_time
3dee7826 215
c0ba0f48 216 url: Final video URL.
d6983cb4 217 ext: Video filename extension.
d67b0b15
PH
218 format: The video format, defaults to ext (used for --get-format)
219 player_url: SWF Player URL (used for rtmpdump).
2f5865cc 220
d6983cb4
PH
221 The following fields are optional:
222
08d30158 223 direct: True if a direct video file was given (must only be set by GenericIE)
f5e43bc6 224 alt_title: A secondary title of the video.
0afef30b
PH
225 display_id An alternative identifier for the video, not necessarily
226 unique, but available before title. Typically, id is
227 something like "4234987", title "Dancing naked mole rats",
228 and display_id "dancing-naked-mole-rats"
d5519808 229 thumbnails: A list of dictionaries, with the following entries:
cfb56d1a 230 * "id" (optional, string) - Thumbnail format ID
d5519808 231 * "url"
cfb56d1a 232 * "preference" (optional, int) - quality of the image
d5519808
PH
233 * "width" (optional, int)
234 * "height" (optional, int)
5e1c39ac 235 * "resolution" (optional, string "{width}x{height}",
d5519808 236 deprecated)
2de624fd 237 * "filesize" (optional, int)
297e9952 238 * "http_headers" (dict) - HTTP headers for the request
d6983cb4 239 thumbnail: Full URL to a video thumbnail image.
f5e43bc6 240 description: Full video description.
d6983cb4 241 uploader: Full name of the video uploader.
2bc0c46f 242 license: License name the video is licensed under.
8a92e51c 243 creator: The creator of the video.
10db0d2f 244 timestamp: UNIX timestamp of the moment the video was uploaded
ae6a1b95 245 upload_date: Video upload date in UTC (YYYYMMDD).
f0d785d3 246 If not explicitly set, calculated from timestamp
247 release_timestamp: UNIX timestamp of the moment the video was released.
248 If it is not clear whether to use timestamp or this, use the former
ae6a1b95 249 release_date: The date (YYYYMMDD) when the video was released in UTC.
f0d785d3 250 If not explicitly set, calculated from release_timestamp
251 modified_timestamp: UNIX timestamp of the moment the video was last modified.
ae6a1b95 252 modified_date: The date (YYYYMMDD) when the video was last modified in UTC.
f0d785d3 253 If not explicitly set, calculated from modified_timestamp
d6983cb4 254 uploader_id: Nickname or id of the video uploader.
7bcd2830 255 uploader_url: Full URL to a personal webpage of the video uploader.
6f1f59f3 256 channel: Full name of the channel the video is uploaded on.
0e7b8d3e 257 Note that channel fields may or may not repeat uploader
6f1f59f3
S
258 fields. This depends on a particular extractor.
259 channel_id: Id of the channel.
260 channel_url: Full URL to a channel webpage.
6c73052c 261 channel_follower_count: Number of followers of the channel.
da9ec3b9 262 location: Physical location where the video was filmed.
a504ced0 263 subtitles: The available subtitles as a dictionary in the format
4606c34e
YCH
264 {tag: subformats}. "tag" is usually a language code, and
265 "subformats" is a list sorted from lower to higher
266 preference, each element is a dictionary with the "ext"
267 entry and one of:
a504ced0 268 * "data": The subtitles file contents
10952eb2 269 * "url": A URL pointing to the subtitles file
2412044c 270 It can optionally also have:
271 * "name": Name or description of the subtitles
08d30158 272 * "http_headers": A dictionary of additional HTTP headers
297e9952 273 to add to the request.
4bba3716 274 "ext" will be calculated from URL if missing
e167860c 275 automatic_captions: Like 'subtitles'; contains automatically generated
276 captions instead of normal subtitles
62d231c0 277 duration: Length of the video in seconds, as an integer or float.
f3d29461 278 view_count: How many users have watched the video on the platform.
19e3dfc9
PH
279 like_count: Number of positive ratings of the video
280 dislike_count: Number of negative ratings of the video
02835c6b 281 repost_count: Number of reposts of the video
2d30521a 282 average_rating: Average rating give by users, the scale used depends on the webpage
19e3dfc9 283 comment_count: Number of comments on the video
dd622d7c
PH
284 comments: A list of comments, each with one or more of the following
285 properties (all but one of text or html optional):
286 * "author" - human-readable name of the comment author
287 * "author_id" - user ID of the comment author
a1c5d2ca 288 * "author_thumbnail" - The thumbnail of the comment author
dd622d7c
PH
289 * "id" - Comment ID
290 * "html" - Comment as HTML
291 * "text" - Plain text of the comment
292 * "timestamp" - UNIX timestamp of comment
293 * "parent" - ID of the comment this one is replying to.
294 Set to "root" to indicate that this is a
295 comment to the original video.
a1c5d2ca
M
296 * "like_count" - Number of positive ratings of the comment
297 * "dislike_count" - Number of negative ratings of the comment
298 * "is_favorited" - Whether the comment is marked as
299 favorite by the video uploader
300 * "author_is_uploader" - Whether the comment is made by
301 the video uploader
8dbe9899 302 age_limit: Age restriction for the video, as an integer (years)
7a5c1cfe 303 webpage_url: The URL to the video webpage, if given to yt-dlp it
9103bbc5
JMF
304 should allow to get the same result again. (It will be set
305 by YoutubeDL if it's missing)
ad3bc6ac
PH
306 categories: A list of categories that the video falls in, for example
307 ["Sports", "Berlin"]
864f24bd 308 tags: A list of tags assigned to the video, e.g. ["sweden", "pop music"]
d0fb4bd1 309 cast: A list of the video cast
7267bd53
PH
310 is_live: True, False, or None (=unknown). Whether this video is a
311 live stream that goes on instead of a fixed-length video.
f76ede8e 312 was_live: True, False, or None (=unknown). Whether this video was
313 originally a live stream.
3dbb2a9d 314 live_status: 'is_live', 'is_upcoming', 'was_live', 'not_live' or None (=unknown)
ae30b840 315 If absent, automatically set from is_live, was_live
7c80519c 316 start_time: Time in seconds where the reproduction should start, as
10952eb2 317 specified in the URL.
297a564b 318 end_time: Time in seconds where the reproduction should end, as
10952eb2 319 specified in the URL.
55949fed 320 chapters: A list of dictionaries, with the following entries:
321 * "start_time" - The start time of the chapter in seconds
322 * "end_time" - The end time of the chapter in seconds
323 * "title" (optional, string)
6cfda058 324 playable_in_embed: Whether this video is allowed to play in embedded
325 players on other sites. Can be True (=always allowed),
326 False (=never allowed), None (=unknown), or a string
c224251a
M
327 specifying the criteria for embedability (Eg: 'whitelist')
328 availability: Under what condition the video is available. One of
329 'private', 'premium_only', 'subscriber_only', 'needs_auth',
330 'unlisted' or 'public'. Use 'InfoExtractor._availability'
331 to set it
277d6ff5 332 __post_extractor: A function to be called just before the metadata is
333 written to either disk, logger or console. The function
334 must return a dict which will be added to the info_dict.
335 This is usefull for additional information that is
336 time-consuming to extract. Note that the fields thus
337 extracted will not be available to output template and
338 match_filter. So, only "comments" and "comment_count" are
339 currently allowed to be extracted via this method.
d6983cb4 340
7109903e
S
341 The following fields should only be used when the video belongs to some logical
342 chapter or section:
343
344 chapter: Name or title of the chapter the video belongs to.
27bfd4e5
S
345 chapter_number: Number of the chapter the video belongs to, as an integer.
346 chapter_id: Id of the chapter the video belongs to, as a unicode string.
7109903e
S
347
348 The following fields should only be used when the video is an episode of some
8d76bdf1 349 series, programme or podcast:
7109903e
S
350
351 series: Title of the series or programme the video episode belongs to.
9ac24e23 352 series_id: Id of the series or programme the video episode belongs to, as a unicode string.
7109903e 353 season: Title of the season the video episode belongs to.
27bfd4e5
S
354 season_number: Number of the season the video episode belongs to, as an integer.
355 season_id: Id of the season the video episode belongs to, as a unicode string.
7109903e
S
356 episode: Title of the video episode. Unlike mandatory video title field,
357 this field should denote the exact title of the video episode
358 without any kind of decoration.
27bfd4e5
S
359 episode_number: Number of the video episode within a season, as an integer.
360 episode_id: Id of the video episode, as a unicode string.
7109903e 361
7a93ab5f
S
362 The following fields should only be used when the media is a track or a part of
363 a music album:
364
365 track: Title of the track.
366 track_number: Number of the track within an album or a disc, as an integer.
367 track_id: Id of the track (useful in case of custom indexing, e.g. 6.iii),
368 as a unicode string.
369 artist: Artist(s) of the track.
370 genre: Genre(s) of the track.
371 album: Title of the album the track belongs to.
372 album_type: Type of the album (e.g. "Demo", "Full-length", "Split", "Compilation", etc).
373 album_artist: List of all artists appeared on the album (e.g.
374 "Ash Borer / Fell Voices" or "Various Artists", useful for splits
375 and compilations).
376 disc_number: Number of the disc or other physical medium the track belongs to,
377 as an integer.
378 release_year: Year (YYYY) when the album was released.
8bcd4048 379 composer: Composer of the piece
7a93ab5f 380
3975b4d2 381 The following fields should only be set for clips that should be cut from the original video:
382
383 section_start: Start time of the section in seconds
384 section_end: End time of the section in seconds
385
deefc05b 386 Unless mentioned otherwise, the fields should be Unicode strings.
d6983cb4 387
d838b1bd
PH
388 Unless mentioned otherwise, None is equivalent to absence of information.
389
fed5d032
PH
390
391 _type "playlist" indicates multiple videos.
b82f815f
PH
392 There must be a key "entries", which is a list, an iterable, or a PagedList
393 object, each element of which is a valid dictionary by this specification.
fed5d032 394
962ffcf8 395 Additionally, playlists can have "id", "title", and any other relevant
b60419c5 396 attributes with the same semantics as videos (see above).
fed5d032 397
f0d785d3 398 It can also have the following optional fields:
399
400 playlist_count: The total number of videos in a playlist. If not given,
401 YoutubeDL tries to calculate it from "entries"
402
fed5d032
PH
403
404 _type "multi_video" indicates that there are multiple videos that
405 form a single show, for examples multiple acts of an opera or TV episode.
406 It must have an entries key like a playlist and contain all the keys
407 required for a video at the same time.
408
409
410 _type "url" indicates that the video must be extracted from another
411 location, possibly by a different extractor. Its only required key is:
412 "url" - the next URL to extract.
f58766ce
PH
413 The key "ie_key" can be set to the class name (minus the trailing "IE",
414 e.g. "Youtube") if the extractor class is known in advance.
415 Additionally, the dictionary may have any properties of the resolved entity
416 known in advance, for example "title" if the title of the referred video is
fed5d032
PH
417 known ahead of time.
418
419
420 _type "url_transparent" entities have the same specification as "url", but
421 indicate that the given additional information is more precise than the one
422 associated with the resolved URL.
423 This is useful when a site employs a video service that hosts the video and
424 its technical metadata, but that video service does not embed a useful
425 title, description etc.
426
427
08d30158 428 Subclasses of this should define a _VALID_URL regexp and, re-define the
429 _real_extract() and (optionally) _real_initialize() methods.
d6983cb4
PH
430 Probably, they should also be added to the list of extractors.
431
e6f21b3d 432 Subclasses may also override suitable() if necessary, but ensure the function
433 signature is preserved and that this function imports everything it needs
52efa4b3 434 (except other extractors), so that lazy_extractors works correctly.
435
436 To support username + password (or netrc) login, the extractor must define a
437 _NETRC_MACHINE and re-define _perform_login(username, password) and
438 (optionally) _initialize_pre_login() methods. The _perform_login method will
439 be called between _initialize_pre_login and _real_initialize if credentials
440 are passed by the user. In cases where it is necessary to have the login
441 process as part of the extraction rather than initialization, _perform_login
442 can be left undefined.
e6f21b3d 443
4248dad9 444 _GEO_BYPASS attribute may be set to False in order to disable
773f291d
S
445 geo restriction bypass mechanisms for a particular extractor.
446 Though it won't disable explicit geo restriction bypass based on
504f20dd 447 country code provided with geo_bypass_country.
4248dad9
S
448
449 _GEO_COUNTRIES attribute may contain a list of presumably geo unrestricted
450 countries for this extractor. One of these countries will be used by
451 geo restriction bypass mechanism right away in order to bypass
504f20dd 452 geo restriction, of course, if the mechanism is not disabled.
773f291d 453
5f95927a
S
454 _GEO_IP_BLOCKS attribute may contain a list of presumably geo unrestricted
455 IP blocks in CIDR notation for this extractor. One of these IP blocks
456 will be used by geo restriction bypass mechanism similarly
504f20dd 457 to _GEO_COUNTRIES.
3ccdde8c 458
e6f21b3d 459 The _WORKING attribute should be set to False for broken IEs
d6983cb4
PH
460 in order to warn the users and skip the tests.
461 """
462
463 _ready = False
464 _downloader = None
773f291d 465 _x_forwarded_for_ip = None
4248dad9
S
466 _GEO_BYPASS = True
467 _GEO_COUNTRIES = None
5f95927a 468 _GEO_IP_BLOCKS = None
d6983cb4 469 _WORKING = True
52efa4b3 470 _NETRC_MACHINE = None
231025c4 471 IE_DESC = None
8dcce6a8 472 SEARCH_KEY = None
d6983cb4 473
8dcce6a8 474 def _login_hint(self, method=NO_DEFAULT, netrc=None):
475 password_hint = f'--username and --password, or --netrc ({netrc or self._NETRC_MACHINE}) to provide account credentials'
476 return {
477 None: '',
478 'any': f'Use --cookies, --cookies-from-browser, {password_hint}',
479 'password': f'Use {password_hint}',
480 'cookies': (
481 'Use --cookies-from-browser or --cookies for the authentication. '
482 'See https://github.com/ytdl-org/youtube-dl#how-do-i-pass-cookies-to-youtube-dl for how to manually pass cookies'),
483 }[method if method is not NO_DEFAULT else 'any' if self.supports_login() else 'cookies']
9d5d4d64 484
d6983cb4 485 def __init__(self, downloader=None):
49a57e70 486 """Constructor. Receives an optional downloader (a YoutubeDL instance).
487 If a downloader is not passed during initialization,
488 it must be set using "set_downloader()" before "extract()" is called"""
d6983cb4 489 self._ready = False
773f291d 490 self._x_forwarded_for_ip = None
28f436ba 491 self._printed_messages = set()
d6983cb4
PH
492 self.set_downloader(downloader)
493
494 @classmethod
5ad28e7f 495 def _match_valid_url(cls, url):
79cb2577
PH
496 # This does not use has/getattr intentionally - we want to know whether
497 # we have cached the regexp for *this* class, whereas getattr would also
498 # match the superclass
499 if '_VALID_URL_RE' not in cls.__dict__:
2c4aaadd 500 if '_VALID_URL' not in cls.__dict__:
501 cls._VALID_URL = cls._make_valid_url()
79cb2577 502 cls._VALID_URL_RE = re.compile(cls._VALID_URL)
5ad28e7f 503 return cls._VALID_URL_RE.match(url)
504
505 @classmethod
506 def suitable(cls, url):
507 """Receives a URL and returns True if suitable for this IE."""
3fb4e21b 508 # This function must import everything it needs (except other extractors),
509 # so that lazy_extractors works correctly
5ad28e7f 510 return cls._match_valid_url(url) is not None
d6983cb4 511
ed9266db
PH
512 @classmethod
513 def _match_id(cls, url):
5ad28e7f 514 return cls._match_valid_url(url).group('id')
ed9266db 515
1151c407 516 @classmethod
517 def get_temp_id(cls, url):
518 try:
519 return cls._match_id(url)
520 except (IndexError, AttributeError):
521 return None
522
d6983cb4
PH
523 @classmethod
524 def working(cls):
525 """Getter method for _WORKING."""
526 return cls._WORKING
527
52efa4b3 528 @classmethod
529 def supports_login(cls):
530 return bool(cls._NETRC_MACHINE)
531
d6983cb4
PH
532 def initialize(self):
533 """Initializes an instance (authentication, etc)."""
28f436ba 534 self._printed_messages = set()
5f95927a
S
535 self._initialize_geo_bypass({
536 'countries': self._GEO_COUNTRIES,
537 'ip_blocks': self._GEO_IP_BLOCKS,
538 })
4248dad9 539 if not self._ready:
52efa4b3 540 self._initialize_pre_login()
541 if self.supports_login():
542 username, password = self._get_login_info()
543 if username:
544 self._perform_login(username, password)
545 elif self.get_param('username') and False not in (self.IE_DESC, self._NETRC_MACHINE):
8dcce6a8 546 self.report_warning(f'Login with password is not supported for this website. {self._login_hint("cookies")}')
4248dad9
S
547 self._real_initialize()
548 self._ready = True
549
5f95927a 550 def _initialize_geo_bypass(self, geo_bypass_context):
e39b5d4a
S
551 """
552 Initialize geo restriction bypass mechanism.
553
554 This method is used to initialize geo bypass mechanism based on faking
555 X-Forwarded-For HTTP header. A random country from provided country list
dc0a869e 556 is selected and a random IP belonging to this country is generated. This
e39b5d4a
S
557 IP will be passed as X-Forwarded-For HTTP header in all subsequent
558 HTTP requests.
e39b5d4a
S
559
560 This method will be used for initial geo bypass mechanism initialization
5f95927a
S
561 during the instance initialization with _GEO_COUNTRIES and
562 _GEO_IP_BLOCKS.
e39b5d4a 563
5f95927a 564 You may also manually call it from extractor's code if geo bypass
e39b5d4a 565 information is not available beforehand (e.g. obtained during
5f95927a
S
566 extraction) or due to some other reason. In this case you should pass
567 this information in geo bypass context passed as first argument. It may
568 contain following fields:
569
570 countries: List of geo unrestricted countries (similar
571 to _GEO_COUNTRIES)
572 ip_blocks: List of geo unrestricted IP blocks in CIDR notation
573 (similar to _GEO_IP_BLOCKS)
574
e39b5d4a 575 """
773f291d 576 if not self._x_forwarded_for_ip:
5f95927a
S
577
578 # Geo bypass mechanism is explicitly disabled by user
a06916d9 579 if not self.get_param('geo_bypass', True):
5f95927a
S
580 return
581
582 if not geo_bypass_context:
583 geo_bypass_context = {}
584
585 # Backward compatibility: previously _initialize_geo_bypass
586 # expected a list of countries, some 3rd party code may still use
587 # it this way
588 if isinstance(geo_bypass_context, (list, tuple)):
589 geo_bypass_context = {
590 'countries': geo_bypass_context,
591 }
592
593 # The whole point of geo bypass mechanism is to fake IP
594 # as X-Forwarded-For HTTP header based on some IP block or
595 # country code.
596
597 # Path 1: bypassing based on IP block in CIDR notation
598
599 # Explicit IP block specified by user, use it right away
600 # regardless of whether extractor is geo bypassable or not
a06916d9 601 ip_block = self.get_param('geo_bypass_ip_block', None)
5f95927a
S
602
603 # Otherwise use random IP block from geo bypass context but only
604 # if extractor is known as geo bypassable
605 if not ip_block:
606 ip_blocks = geo_bypass_context.get('ip_blocks')
607 if self._GEO_BYPASS and ip_blocks:
608 ip_block = random.choice(ip_blocks)
609
610 if ip_block:
611 self._x_forwarded_for_ip = GeoUtils.random_ipv4(ip_block)
8a82af35 612 self.write_debug(f'Using fake IP {self._x_forwarded_for_ip} as X-Forwarded-For')
5f95927a
S
613 return
614
615 # Path 2: bypassing based on country code
616
617 # Explicit country code specified by user, use it right away
618 # regardless of whether extractor is geo bypassable or not
a06916d9 619 country = self.get_param('geo_bypass_country', None)
5f95927a
S
620
621 # Otherwise use random country code from geo bypass context but
622 # only if extractor is known as geo bypassable
623 if not country:
624 countries = geo_bypass_context.get('countries')
625 if self._GEO_BYPASS and countries:
626 country = random.choice(countries)
627
628 if country:
629 self._x_forwarded_for_ip = GeoUtils.random_ipv4(country)
0760b0a7 630 self._downloader.write_debug(
86e5f3ed 631 f'Using fake IP {self._x_forwarded_for_ip} ({country.upper()}) as X-Forwarded-For')
d6983cb4
PH
632
633 def extract(self, url):
634 """Extracts URL information and returns it in list of dicts."""
3a5bcd03 635 try:
773f291d
S
636 for _ in range(2):
637 try:
638 self.initialize()
a06916d9 639 self.write_debug('Extracting URL: %s' % url)
0016b84e 640 ie_result = self._real_extract(url)
07cce701 641 if ie_result is None:
642 return None
0016b84e
S
643 if self._x_forwarded_for_ip:
644 ie_result['__x_forwarded_for_ip'] = self._x_forwarded_for_ip
53ed7066 645 subtitles = ie_result.get('subtitles')
646 if (subtitles and 'live_chat' in subtitles
a06916d9 647 and 'no-live-chat' in self.get_param('compat_opts', [])):
53ed7066 648 del subtitles['live_chat']
0016b84e 649 return ie_result
773f291d 650 except GeoRestrictedError as e:
4248dad9
S
651 if self.__maybe_fake_ip_and_retry(e.countries):
652 continue
773f291d 653 raise
0db3bae8 654 except UnsupportedError:
655 raise
1151c407 656 except ExtractorError as e:
0db3bae8 657 kwargs = {
658 'video_id': e.video_id or self.get_temp_id(url),
659 'ie': self.IE_NAME,
b69fd25c 660 'tb': e.traceback or sys.exc_info()[2],
0db3bae8 661 'expected': e.expected,
662 'cause': e.cause
663 }
664 if hasattr(e, 'countries'):
665 kwargs['countries'] = e.countries
7265a219 666 raise type(e)(e.orig_msg, **kwargs)
ac668111 667 except http.client.IncompleteRead as e:
1151c407 668 raise ExtractorError('A network error has occurred.', cause=e, expected=True, video_id=self.get_temp_id(url))
9650885b 669 except (KeyError, StopIteration) as e:
1151c407 670 raise ExtractorError('An extractor error has occurred.', cause=e, video_id=self.get_temp_id(url))
d6983cb4 671
4248dad9 672 def __maybe_fake_ip_and_retry(self, countries):
a06916d9 673 if (not self.get_param('geo_bypass_country', None)
3089bc74 674 and self._GEO_BYPASS
a06916d9 675 and self.get_param('geo_bypass', True)
3089bc74
S
676 and not self._x_forwarded_for_ip
677 and countries):
eea0716c
S
678 country_code = random.choice(countries)
679 self._x_forwarded_for_ip = GeoUtils.random_ipv4(country_code)
4248dad9
S
680 if self._x_forwarded_for_ip:
681 self.report_warning(
eea0716c
S
682 'Video is geo restricted. Retrying extraction with fake IP %s (%s) as X-Forwarded-For.'
683 % (self._x_forwarded_for_ip, country_code.upper()))
4248dad9
S
684 return True
685 return False
686
d6983cb4 687 def set_downloader(self, downloader):
08d30158 688 """Sets a YoutubeDL instance as the downloader for this IE."""
d6983cb4
PH
689 self._downloader = downloader
690
9809740b 691 @property
692 def cache(self):
693 return self._downloader.cache
694
695 @property
696 def cookiejar(self):
697 return self._downloader.cookiejar
698
52efa4b3 699 def _initialize_pre_login(self):
962ffcf8 700 """ Initialization before login. Redefine in subclasses."""
52efa4b3 701 pass
702
703 def _perform_login(self, username, password):
704 """ Login with username and password. Redefine in subclasses."""
705 pass
706
d6983cb4
PH
707 def _real_initialize(self):
708 """Real initialization process. Redefine in subclasses."""
709 pass
710
711 def _real_extract(self, url):
712 """Real extraction process. Redefine in subclasses."""
08d30158 713 raise NotImplementedError('This method must be implemented by subclasses')
d6983cb4 714
56c73665
JMF
715 @classmethod
716 def ie_key(cls):
717 """A string for getting the InfoExtractor with get_info_extractor"""
3fb4e21b 718 return cls.__name__[:-2]
56c73665 719
82d02080 720 @classproperty
721 def IE_NAME(cls):
722 return cls.__name__[:-2]
d6983cb4 723
d391b7e2
S
724 @staticmethod
725 def __can_accept_status_code(err, expected_status):
ac668111 726 assert isinstance(err, urllib.error.HTTPError)
d391b7e2
S
727 if expected_status is None:
728 return False
d391b7e2
S
729 elif callable(expected_status):
730 return expected_status(err.code) is True
731 else:
6606817a 732 return err.code in variadic(expected_status)
d391b7e2 733
c043c246 734 def _create_request(self, url_or_request, data=None, headers=None, query=None):
ac668111 735 if isinstance(url_or_request, urllib.request.Request):
09d02ea4 736 return update_Request(url_or_request, data=data, headers=headers, query=query)
737 if query:
738 url_or_request = update_url_query(url_or_request, query)
c043c246 739 return sanitized_Request(url_or_request, data, headers or {})
f95b9dee 740
c043c246 741 def _request_webpage(self, url_or_request, video_id, note=None, errnote=None, fatal=True, data=None, headers=None, query=None, expected_status=None):
d391b7e2
S
742 """
743 Return the response handle.
744
745 See _download_webpage docstring for arguments specification.
746 """
1cf376f5 747 if not self._downloader._first_webpage_request:
49a57e70 748 sleep_interval = self.get_param('sleep_interval_requests') or 0
1cf376f5 749 if sleep_interval > 0:
5ef7d9bd 750 self.to_screen('Sleeping %s seconds ...' % sleep_interval)
1cf376f5 751 time.sleep(sleep_interval)
752 else:
753 self._downloader._first_webpage_request = False
754
d6983cb4
PH
755 if note is None:
756 self.report_download_webpage(video_id)
757 elif note is not False:
7cc3570e 758 if video_id is None:
86e5f3ed 759 self.to_screen(str(note))
7cc3570e 760 else:
86e5f3ed 761 self.to_screen(f'{video_id}: {note}')
2132edaa
S
762
763 # Some sites check X-Forwarded-For HTTP header in order to figure out
764 # the origin of the client behind proxy. This allows bypassing geo
765 # restriction by faking this header's value to IP that belongs to some
766 # geo unrestricted country. We will do so once we encounter any
767 # geo restriction error.
768 if self._x_forwarded_for_ip:
c043c246 769 headers = (headers or {}).copy()
770 headers.setdefault('X-Forwarded-For', self._x_forwarded_for_ip)
2132edaa 771
d6983cb4 772 try:
f95b9dee 773 return self._downloader.urlopen(self._create_request(url_or_request, data, headers, query))
3158150c 774 except network_exceptions as err:
ac668111 775 if isinstance(err, urllib.error.HTTPError):
d391b7e2 776 if self.__can_accept_status_code(err, expected_status):
95e42d73
XDG
777 # Retain reference to error to prevent file object from
778 # being closed before it can be read. Works around the
779 # effects of <https://bugs.python.org/issue15002>
780 # introduced in Python 3.4.1.
781 err.fp._error = err
d391b7e2
S
782 return err.fp
783
aa94a6d3
PH
784 if errnote is False:
785 return False
d6983cb4 786 if errnote is None:
f1a9d64e 787 errnote = 'Unable to download webpage'
7f8b2714 788
86e5f3ed 789 errmsg = f'{errnote}: {error_to_compat_str(err)}'
7cc3570e 790 if fatal:
497d2fab 791 raise ExtractorError(errmsg, cause=err)
7cc3570e 792 else:
6a39ee13 793 self.report_warning(errmsg)
7cc3570e 794 return False
d6983cb4 795
1890fc63 796 def _download_webpage_handle(self, url_or_request, video_id, note=None, errnote=None, fatal=True,
797 encoding=None, data=None, headers={}, query={}, expected_status=None):
d391b7e2
S
798 """
799 Return a tuple (page content as string, URL handle).
800
617f658b 801 Arguments:
802 url_or_request -- plain text URL as a string or
ac668111 803 a urllib.request.Request object
617f658b 804 video_id -- Video/playlist/item identifier (string)
805
806 Keyword arguments:
807 note -- note printed before downloading (string)
808 errnote -- note printed in case of an error (string)
809 fatal -- flag denoting whether error should be considered fatal,
810 i.e. whether it should cause ExtractionError to be raised,
811 otherwise a warning will be reported and extraction continued
812 encoding -- encoding for a page content decoding, guessed automatically
813 when not explicitly specified
814 data -- POST data (bytes)
815 headers -- HTTP headers (dict)
816 query -- URL query (dict)
817 expected_status -- allows to accept failed HTTP requests (non 2xx
818 status code) by explicitly specifying a set of accepted status
819 codes. Can be any of the following entities:
820 - an integer type specifying an exact failed status code to
821 accept
822 - a list or a tuple of integer types specifying a list of
823 failed status codes to accept
824 - a callable accepting an actual failed status code and
825 returning True if it should be accepted
826 Note that this argument does not affect success status codes (2xx)
827 which are always accepted.
d391b7e2 828 """
617f658b 829
b9d3e163 830 # Strip hashes from the URL (#1038)
14f25df2 831 if isinstance(url_or_request, str):
b9d3e163
PH
832 url_or_request = url_or_request.partition('#')[0]
833
d391b7e2 834 urlh = self._request_webpage(url_or_request, video_id, note, errnote, fatal, data=data, headers=headers, query=query, expected_status=expected_status)
7cc3570e
PH
835 if urlh is False:
836 assert not fatal
837 return False
c9a77969 838 content = self._webpage_read_content(urlh, url_or_request, video_id, note, errnote, fatal, encoding=encoding)
23be51d8
PH
839 return (content, urlh)
840
c9a77969
YCH
841 @staticmethod
842 def _guess_encoding_from_content(content_type, webpage_bytes):
d6983cb4
PH
843 m = re.match(r'[a-zA-Z0-9_.-]+/[a-zA-Z0-9_.-]+\s*;\s*charset=(.+)', content_type)
844 if m:
845 encoding = m.group(1)
846 else:
0d75ae2c 847 m = re.search(br'<meta[^>]+charset=[\'"]?([^\'")]+)[ /\'">]',
f143d86a
PH
848 webpage_bytes[:1024])
849 if m:
850 encoding = m.group(1).decode('ascii')
b60016e8
PH
851 elif webpage_bytes.startswith(b'\xff\xfe'):
852 encoding = 'utf-16'
f143d86a
PH
853 else:
854 encoding = 'utf-8'
c9a77969
YCH
855
856 return encoding
857
4457823d
S
858 def __check_blocked(self, content):
859 first_block = content[:512]
3089bc74
S
860 if ('<title>Access to this site is blocked</title>' in content
861 and 'Websense' in first_block):
4457823d
S
862 msg = 'Access to this webpage has been blocked by Websense filtering software in your network.'
863 blocked_iframe = self._html_search_regex(
864 r'<iframe src="([^"]+)"', content,
865 'Websense information URL', default=None)
866 if blocked_iframe:
867 msg += ' Visit %s for more details' % blocked_iframe
868 raise ExtractorError(msg, expected=True)
869 if '<title>The URL you requested has been blocked</title>' in first_block:
870 msg = (
871 'Access to this webpage has been blocked by Indian censorship. '
872 'Use a VPN or proxy server (with --proxy) to route around it.')
873 block_msg = self._html_search_regex(
874 r'</h1><p>(.*?)</p>',
875 content, 'block message', default=None)
876 if block_msg:
877 msg += ' (Message: "%s")' % block_msg.replace('\n', ' ')
878 raise ExtractorError(msg, expected=True)
3089bc74
S
879 if ('<title>TTK :: Доступ к ресурсу ограничен</title>' in content
880 and 'blocklist.rkn.gov.ru' in content):
4457823d
S
881 raise ExtractorError(
882 'Access to this webpage has been blocked by decision of the Russian government. '
883 'Visit http://blocklist.rkn.gov.ru/ for a block reason.',
884 expected=True)
885
f95b9dee 886 def _request_dump_filename(self, url, video_id):
887 basen = f'{video_id}_{url}'
888 trim_length = self.get_param('trim_file_name') or 240
889 if len(basen) > trim_length:
890 h = '___' + hashlib.md5(basen.encode('utf-8')).hexdigest()
891 basen = basen[:trim_length - len(h)] + h
892 filename = sanitize_filename(f'{basen}.dump', restricted=True)
893 # Working around MAX_PATH limitation on Windows (see
894 # http://msdn.microsoft.com/en-us/library/windows/desktop/aa365247(v=vs.85).aspx)
895 if compat_os_name == 'nt':
896 absfilepath = os.path.abspath(filename)
897 if len(absfilepath) > 259:
898 filename = fR'\\?\{absfilepath}'
899 return filename
900
901 def __decode_webpage(self, webpage_bytes, encoding, headers):
902 if not encoding:
903 encoding = self._guess_encoding_from_content(headers.get('Content-Type', ''), webpage_bytes)
904 try:
905 return webpage_bytes.decode(encoding, 'replace')
906 except LookupError:
907 return webpage_bytes.decode('utf-8', 'replace')
908
c9a77969 909 def _webpage_read_content(self, urlh, url_or_request, video_id, note=None, errnote=None, fatal=True, prefix=None, encoding=None):
c9a77969
YCH
910 webpage_bytes = urlh.read()
911 if prefix is not None:
912 webpage_bytes = prefix + webpage_bytes
a06916d9 913 if self.get_param('dump_intermediate_pages', False):
f610dbb0 914 self.to_screen('Dumping request to ' + urlh.geturl())
d6983cb4
PH
915 dump = base64.b64encode(webpage_bytes).decode('ascii')
916 self._downloader.to_screen(dump)
f95b9dee 917 if self.get_param('write_pages'):
e121e3ce 918 filename = self._request_dump_filename(urlh.geturl(), video_id)
f95b9dee 919 self.to_screen(f'Saving request to {filename}')
d41e6efc
PH
920 with open(filename, 'wb') as outf:
921 outf.write(webpage_bytes)
922
f95b9dee 923 content = self.__decode_webpage(webpage_bytes, encoding, urlh.headers)
4457823d 924 self.__check_blocked(content)
2410c43d 925
23be51d8 926 return content
d6983cb4 927
e01c3d2e 928 def _parse_xml(self, xml_string, video_id, transform_source=None, fatal=True):
e2b38da9
PH
929 if transform_source:
930 xml_string = transform_source(xml_string)
e01c3d2e
S
931 try:
932 return compat_etree_fromstring(xml_string.encode('utf-8'))
f9934b96 933 except xml.etree.ElementTree.ParseError as ve:
e01c3d2e
S
934 errmsg = '%s: Failed to parse XML ' % video_id
935 if fatal:
936 raise ExtractorError(errmsg, cause=ve)
937 else:
938 self.report_warning(errmsg + str(ve))
267ed0c5 939
b7c47b74 940 def _parse_json(self, json_string, video_id, transform_source=None, fatal=True, **parser_kwargs):
3d3538e4 941 try:
b7c47b74 942 return json.loads(
943 json_string, cls=LenientJSONDecoder, strict=False, transform_source=transform_source, **parser_kwargs)
3d3538e4 944 except ValueError as ve:
1890fc63 945 errmsg = f'{video_id}: Failed to parse JSON'
e7b6d122
PH
946 if fatal:
947 raise ExtractorError(errmsg, cause=ve)
948 else:
e121e3ce 949 self.report_warning(f'{errmsg}: {ve}')
3d3538e4 950
adddc50c 951 def _parse_socket_response_as_json(self, data, video_id, transform_source=None, fatal=True):
952 return self._parse_json(
953 data[data.find('{'):data.rfind('}') + 1],
954 video_id, transform_source, fatal)
955
617f658b 956 def __create_download_methods(name, parser, note, errnote, return_value):
957
958 def parse(ie, content, *args, **kwargs):
959 if parser is None:
960 return content
961 # parser is fetched by name so subclasses can override it
962 return getattr(ie, parser)(content, *args, **kwargs)
963
c4910024 964 def download_handle(self, url_or_request, video_id, note=note, errnote=errnote, transform_source=None,
965 fatal=True, encoding=None, data=None, headers={}, query={}, expected_status=None):
966 res = self._download_webpage_handle(
967 url_or_request, video_id, note=note, errnote=errnote, fatal=fatal, encoding=encoding,
968 data=data, headers=headers, query=query, expected_status=expected_status)
617f658b 969 if res is False:
970 return res
971 content, urlh = res
c4910024 972 return parse(self, content, video_id, transform_source=transform_source, fatal=fatal), urlh
617f658b 973
f95b9dee 974 def download_content(self, url_or_request, video_id, note=note, errnote=errnote, transform_source=None,
c4910024 975 fatal=True, encoding=None, data=None, headers={}, query={}, expected_status=None):
f95b9dee 976 if self.get_param('load_pages'):
977 url_or_request = self._create_request(url_or_request, data, headers, query)
978 filename = self._request_dump_filename(url_or_request.full_url, video_id)
979 self.to_screen(f'Loading request from {filename}')
980 try:
981 with open(filename, 'rb') as dumpf:
982 webpage_bytes = dumpf.read()
983 except OSError as e:
984 self.report_warning(f'Unable to load request from disk: {e}')
985 else:
986 content = self.__decode_webpage(webpage_bytes, encoding, url_or_request.headers)
987 return parse(self, content, video_id, transform_source, fatal)
c4910024 988 kwargs = {
989 'note': note,
990 'errnote': errnote,
991 'transform_source': transform_source,
992 'fatal': fatal,
993 'encoding': encoding,
994 'data': data,
995 'headers': headers,
996 'query': query,
997 'expected_status': expected_status,
998 }
617f658b 999 if parser is None:
c4910024 1000 kwargs.pop('transform_source')
617f658b 1001 # The method is fetched by name so subclasses can override _download_..._handle
c4910024 1002 res = getattr(self, download_handle.__name__)(url_or_request, video_id, **kwargs)
617f658b 1003 return res if res is False else res[0]
1004
1005 def impersonate(func, name, return_value):
1006 func.__name__, func.__qualname__ = name, f'InfoExtractor.{name}'
1007 func.__doc__ = f'''
1008 @param transform_source Apply this transformation before parsing
1009 @returns {return_value}
1010
1011 See _download_webpage_handle docstring for other arguments specification
1012 '''
1013
1014 impersonate(download_handle, f'_download_{name}_handle', f'({return_value}, URL handle)')
1015 impersonate(download_content, f'_download_{name}', f'{return_value}')
1016 return download_handle, download_content
1017
1018 _download_xml_handle, _download_xml = __create_download_methods(
1019 'xml', '_parse_xml', 'Downloading XML', 'Unable to download XML', 'xml as an xml.etree.ElementTree.Element')
1020 _download_json_handle, _download_json = __create_download_methods(
1021 'json', '_parse_json', 'Downloading JSON metadata', 'Unable to download JSON metadata', 'JSON object as a dict')
1022 _download_socket_json_handle, _download_socket_json = __create_download_methods(
1023 'socket_json', '_parse_socket_response_as_json', 'Polling socket', 'Unable to poll socket', 'JSON object as a dict')
1024 __download_webpage = __create_download_methods('webpage', None, None, None, 'data of the page as a string')[1]
adddc50c 1025
617f658b 1026 def _download_webpage(
1027 self, url_or_request, video_id, note=None, errnote=None,
1028 fatal=True, tries=1, timeout=NO_DEFAULT, *args, **kwargs):
adddc50c 1029 """
617f658b 1030 Return the data of the page as a string.
adddc50c 1031
617f658b 1032 Keyword arguments:
1033 tries -- number of tries
1034 timeout -- sleep interval between tries
1035
1036 See _download_webpage_handle docstring for other arguments specification.
adddc50c 1037 """
617f658b 1038
1039 R''' # NB: These are unused; should they be deprecated?
1040 if tries != 1:
1041 self._downloader.deprecation_warning('tries argument is deprecated in InfoExtractor._download_webpage')
1042 if timeout is NO_DEFAULT:
1043 timeout = 5
1044 else:
1045 self._downloader.deprecation_warning('timeout argument is deprecated in InfoExtractor._download_webpage')
1046 '''
1047
1048 try_count = 0
1049 while True:
1050 try:
1051 return self.__download_webpage(url_or_request, video_id, note, errnote, None, fatal, *args, **kwargs)
ac668111 1052 except http.client.IncompleteRead as e:
617f658b 1053 try_count += 1
1054 if try_count >= tries:
1055 raise e
1056 self._sleep(timeout, video_id)
adddc50c 1057
28f436ba 1058 def report_warning(self, msg, video_id=None, *args, only_once=False, **kwargs):
a70635b8 1059 idstr = format_field(video_id, None, '%s: ')
28f436ba 1060 msg = f'[{self.IE_NAME}] {idstr}{msg}'
1061 if only_once:
1062 if f'WARNING: {msg}' in self._printed_messages:
1063 return
1064 self._printed_messages.add(f'WARNING: {msg}')
1065 self._downloader.report_warning(msg, *args, **kwargs)
f45f96f8 1066
a06916d9 1067 def to_screen(self, msg, *args, **kwargs):
d6983cb4 1068 """Print msg to screen, prefixing it with '[ie_name]'"""
86e5f3ed 1069 self._downloader.to_screen(f'[{self.IE_NAME}] {msg}', *args, **kwargs)
a06916d9 1070
1071 def write_debug(self, msg, *args, **kwargs):
86e5f3ed 1072 self._downloader.write_debug(f'[{self.IE_NAME}] {msg}', *args, **kwargs)
a06916d9 1073
1074 def get_param(self, name, default=None, *args, **kwargs):
1075 if self._downloader:
1076 return self._downloader.params.get(name, default, *args, **kwargs)
1077 return default
d6983cb4 1078
88acdbc2 1079 def report_drm(self, video_id, partial=False):
1080 self.raise_no_formats('This video is DRM protected', expected=True, video_id=video_id)
1081
d6983cb4
PH
1082 def report_extraction(self, id_or_name):
1083 """Report information extraction."""
f1a9d64e 1084 self.to_screen('%s: Extracting information' % id_or_name)
d6983cb4
PH
1085
1086 def report_download_webpage(self, video_id):
1087 """Report webpage download."""
f1a9d64e 1088 self.to_screen('%s: Downloading webpage' % video_id)
d6983cb4
PH
1089
1090 def report_age_confirmation(self):
1091 """Report attempt to confirm age."""
f1a9d64e 1092 self.to_screen('Confirming age')
d6983cb4 1093
fc79158d
JMF
1094 def report_login(self):
1095 """Report attempt to log in."""
f1a9d64e 1096 self.to_screen('Logging in')
fc79158d 1097
b7da73eb 1098 def raise_login_required(
9d5d4d64 1099 self, msg='This video is only available for registered users',
52efa4b3 1100 metadata_available=False, method=NO_DEFAULT):
f2ebc5c7 1101 if metadata_available and (
1102 self.get_param('ignore_no_formats_error') or self.get_param('wait_for_video')):
b7da73eb 1103 self.report_warning(msg)
7265a219 1104 return
a70635b8 1105 msg += format_field(self._login_hint(method), None, '. %s')
46890374 1106 raise ExtractorError(msg, expected=True)
43e7d3c9 1107
b7da73eb 1108 def raise_geo_restricted(
1109 self, msg='This video is not available from your location due to geo restriction',
1110 countries=None, metadata_available=False):
f2ebc5c7 1111 if metadata_available and (
1112 self.get_param('ignore_no_formats_error') or self.get_param('wait_for_video')):
b7da73eb 1113 self.report_warning(msg)
1114 else:
1115 raise GeoRestrictedError(msg, countries=countries)
1116
1117 def raise_no_formats(self, msg, expected=False, video_id=None):
f2ebc5c7 1118 if expected and (
1119 self.get_param('ignore_no_formats_error') or self.get_param('wait_for_video')):
b7da73eb 1120 self.report_warning(msg, video_id)
68f5867c
L
1121 elif isinstance(msg, ExtractorError):
1122 raise msg
b7da73eb 1123 else:
1124 raise ExtractorError(msg, expected=expected, video_id=video_id)
c430802e 1125
5f6a1245 1126 # Methods for following #608
c0d0b01f 1127 @staticmethod
311b6615 1128 def url_result(url, ie=None, video_id=None, video_title=None, *, url_transparent=False, **kwargs):
10952eb2 1129 """Returns a URL that points to a page that should be processed"""
311b6615 1130 if ie is not None:
1131 kwargs['ie_key'] = ie if isinstance(ie, str) else ie.ie_key()
7012b23c 1132 if video_id is not None:
311b6615 1133 kwargs['id'] = video_id
830d53bf 1134 if video_title is not None:
311b6615 1135 kwargs['title'] = video_title
1136 return {
1137 **kwargs,
1138 '_type': 'url_transparent' if url_transparent else 'url',
1139 'url': url,
1140 }
1141
27231526
ZM
1142 def playlist_from_matches(self, matches, playlist_id=None, playlist_title=None, getter=None, ie=None, video_kwargs=None, **kwargs):
1143 urls = (self.url_result(self._proto_relative_url(m), ie, **(video_kwargs or {}))
311b6615 1144 for m in orderedSet(map(getter, matches) if getter else matches))
1145 return self.playlist_result(urls, playlist_id, playlist_title, **kwargs)
46b18f23 1146
c0d0b01f 1147 @staticmethod
311b6615 1148 def playlist_result(entries, playlist_id=None, playlist_title=None, playlist_description=None, *, multi_video=False, **kwargs):
d6983cb4 1149 """Returns a playlist"""
d6983cb4 1150 if playlist_id:
311b6615 1151 kwargs['id'] = playlist_id
d6983cb4 1152 if playlist_title:
311b6615 1153 kwargs['title'] = playlist_title
ecc97af3 1154 if playlist_description is not None:
311b6615 1155 kwargs['description'] = playlist_description
1156 return {
1157 **kwargs,
1158 '_type': 'multi_video' if multi_video else 'playlist',
1159 'entries': entries,
1160 }
d6983cb4 1161
c342041f 1162 def _search_regex(self, pattern, string, name, default=NO_DEFAULT, fatal=True, flags=0, group=None):
d6983cb4
PH
1163 """
1164 Perform a regex search on the given string, using a single or a list of
1165 patterns returning the first matching group.
1166 In case of failure return a default value or raise a WARNING or a
55b3e45b 1167 RegexNotFoundError, depending on fatal, specifying the field name.
d6983cb4 1168 """
61d3665d 1169 if string is None:
1170 mobj = None
77f90330 1171 elif isinstance(pattern, (str, re.Pattern)):
d6983cb4
PH
1172 mobj = re.search(pattern, string, flags)
1173 else:
1174 for p in pattern:
1175 mobj = re.search(p, string, flags)
c3415d1b
PH
1176 if mobj:
1177 break
d6983cb4 1178
ec11a9f4 1179 _name = self._downloader._format_err(name, self._downloader.Styles.EMPHASIS)
d6983cb4
PH
1180
1181 if mobj:
711ede6e
PH
1182 if group is None:
1183 # return the first matching group
1184 return next(g for g in mobj.groups() if g is not None)
198f7ea8 1185 elif isinstance(group, (list, tuple)):
1186 return tuple(mobj.group(g) for g in group)
711ede6e
PH
1187 else:
1188 return mobj.group(group)
c342041f 1189 elif default is not NO_DEFAULT:
d6983cb4
PH
1190 return default
1191 elif fatal:
f1a9d64e 1192 raise RegexNotFoundError('Unable to extract %s' % _name)
d6983cb4 1193 else:
6a39ee13 1194 self.report_warning('unable to extract %s' % _name + bug_reports_message())
d6983cb4
PH
1195 return None
1196
f0bc6e20 1197 def _search_json(self, start_pattern, string, name, video_id, *, end_pattern='',
1198 contains_pattern='(?s:.+)', fatal=True, default=NO_DEFAULT, **kwargs):
b7c47b74 1199 """Searches string for the JSON object specified by start_pattern"""
1200 # NB: end_pattern is only used to reduce the size of the initial match
f0bc6e20 1201 if default is NO_DEFAULT:
1202 default, has_default = {}, False
1203 else:
1204 fatal, has_default = False, True
1205
1206 json_string = self._search_regex(
1207 rf'{start_pattern}\s*(?P<json>{{\s*{contains_pattern}\s*}})\s*{end_pattern}',
1208 string, name, group='json', fatal=fatal, default=None if has_default else NO_DEFAULT)
1209 if not json_string:
1210 return default
1211
1212 _name = self._downloader._format_err(name, self._downloader.Styles.EMPHASIS)
1213 try:
1214 return self._parse_json(json_string, video_id, ignore_extra=True, **kwargs)
1215 except ExtractorError as e:
1216 if fatal:
1217 raise ExtractorError(
1218 f'Unable to extract {_name} - Failed to parse JSON', cause=e.cause, video_id=video_id)
1219 elif not has_default:
1220 self.report_warning(
1221 f'Unable to extract {_name} - Failed to parse JSON: {e}', video_id=video_id)
1222 return default
b7c47b74 1223
c342041f 1224 def _html_search_regex(self, pattern, string, name, default=NO_DEFAULT, fatal=True, flags=0, group=None):
d6983cb4
PH
1225 """
1226 Like _search_regex, but strips HTML tags and unescapes entities.
1227 """
711ede6e 1228 res = self._search_regex(pattern, string, name, default, fatal, flags, group)
d6983cb4
PH
1229 if res:
1230 return clean_html(res).strip()
1231 else:
1232 return res
1233
2118fdd1
RA
1234 def _get_netrc_login_info(self, netrc_machine=None):
1235 username = None
1236 password = None
1237 netrc_machine = netrc_machine or self._NETRC_MACHINE
1238
a06916d9 1239 if self.get_param('usenetrc', False):
2118fdd1 1240 try:
0001fcb5 1241 netrc_file = compat_expanduser(self.get_param('netrc_location') or '~')
1242 if os.path.isdir(netrc_file):
1243 netrc_file = os.path.join(netrc_file, '.netrc')
1244 info = netrc.netrc(file=netrc_file).authenticators(netrc_machine)
2118fdd1
RA
1245 if info is not None:
1246 username = info[0]
1247 password = info[2]
1248 else:
dcce092e
S
1249 raise netrc.NetrcParseError(
1250 'No authenticators for %s' % netrc_machine)
86e5f3ed 1251 except (OSError, netrc.NetrcParseError) as err:
6a39ee13 1252 self.report_warning(
dcce092e 1253 'parsing .netrc: %s' % error_to_compat_str(err))
2118fdd1 1254
dcce092e 1255 return username, password
2118fdd1 1256
1b6712ab 1257 def _get_login_info(self, username_option='username', password_option='password', netrc_machine=None):
fc79158d 1258 """
cf0649f8 1259 Get the login info as (username, password)
32443dd3
S
1260 First look for the manually specified credentials using username_option
1261 and password_option as keys in params dictionary. If no such credentials
1262 available look in the netrc file using the netrc_machine or _NETRC_MACHINE
1263 value.
fc79158d
JMF
1264 If there's no info available, return (None, None)
1265 """
fc79158d
JMF
1266
1267 # Attempt to use provided username and password or .netrc data
a06916d9 1268 username = self.get_param(username_option)
1269 if username is not None:
1270 password = self.get_param(password_option)
2118fdd1 1271 else:
1b6712ab 1272 username, password = self._get_netrc_login_info(netrc_machine)
5f6a1245 1273
2133565c 1274 return username, password
fc79158d 1275
e64b7569 1276 def _get_tfa_info(self, note='two-factor verification code'):
83317f69 1277 """
1278 Get the two-factor authentication info
1279 TODO - asking the user will be required for sms/phone verify
1280 currently just uses the command line option
1281 If there's no info available, return None
1282 """
83317f69 1283
a06916d9 1284 tfa = self.get_param('twofactor')
1285 if tfa is not None:
1286 return tfa
83317f69 1287
ac668111 1288 return getpass.getpass('Type %s and press [Return]: ' % note)
83317f69 1289
46720279
JMF
1290 # Helper functions for extracting OpenGraph info
1291 @staticmethod
ab2d5247 1292 def _og_regexes(prop):
448ef1f3 1293 content_re = r'content=(?:"([^"]+?)"|\'([^\']+?)\'|\s*([^\s"\'=<>`]+?))'
fbfde1c3
F
1294 property_re = (r'(?:name|property)=(?:\'og%(sep)s%(prop)s\'|"og%(sep)s%(prop)s"|\s*og%(sep)s%(prop)s\b)'
1295 % {'prop': re.escape(prop), 'sep': '(?:&#x3A;|[:-])'})
78fb87b2 1296 template = r'<meta[^>]+?%s[^>]+?%s'
ab2d5247 1297 return [
78fb87b2
JMF
1298 template % (property_re, content_re),
1299 template % (content_re, property_re),
ab2d5247 1300 ]
46720279 1301
864f24bd
S
1302 @staticmethod
1303 def _meta_regex(prop):
1304 return r'''(?isx)<meta
8b9848ac 1305 (?=[^>]+(?:itemprop|name|property|id|http-equiv)=(["\']?)%s\1)
864f24bd
S
1306 [^>]+?content=(["\'])(?P<content>.*?)\2''' % re.escape(prop)
1307
3c4e6d83 1308 def _og_search_property(self, prop, html, name=None, **kargs):
6606817a 1309 prop = variadic(prop)
46720279 1310 if name is None:
b070564e
S
1311 name = 'OpenGraph %s' % prop[0]
1312 og_regexes = []
1313 for p in prop:
1314 og_regexes.extend(self._og_regexes(p))
1315 escaped = self._search_regex(og_regexes, html, name, flags=re.DOTALL, **kargs)
eb0a8398
PH
1316 if escaped is None:
1317 return None
1318 return unescapeHTML(escaped)
46720279
JMF
1319
1320 def _og_search_thumbnail(self, html, **kargs):
10952eb2 1321 return self._og_search_property('image', html, 'thumbnail URL', fatal=False, **kargs)
46720279
JMF
1322
1323 def _og_search_description(self, html, **kargs):
1324 return self._og_search_property('description', html, fatal=False, **kargs)
1325
04f3fd2c 1326 def _og_search_title(self, html, *, fatal=False, **kargs):
1327 return self._og_search_property('title', html, fatal=fatal, **kargs)
46720279 1328
8ffa13e0 1329 def _og_search_video_url(self, html, name='video url', secure=True, **kargs):
a3681973
PH
1330 regexes = self._og_regexes('video') + self._og_regexes('video:url')
1331 if secure:
1332 regexes = self._og_regexes('video:secure_url') + regexes
8ffa13e0 1333 return self._html_search_regex(regexes, html, name, **kargs)
46720279 1334
78338f71
JMF
1335 def _og_search_url(self, html, **kargs):
1336 return self._og_search_property('url', html, **kargs)
1337
04f3fd2c 1338 def _html_extract_title(self, html, name='title', *, fatal=False, **kwargs):
21633673 1339 return self._html_search_regex(r'(?s)<title\b[^>]*>([^<]+)</title>', html, name, fatal=fatal, **kwargs)
77cc7c6e 1340
40c696e5 1341 def _html_search_meta(self, name, html, display_name=None, fatal=False, **kwargs):
6606817a 1342 name = variadic(name)
59040888 1343 if display_name is None:
88d9f6c0 1344 display_name = name[0]
59040888 1345 return self._html_search_regex(
88d9f6c0 1346 [self._meta_regex(n) for n in name],
711ede6e 1347 html, display_name, fatal=fatal, group='content', **kwargs)
59040888
PH
1348
1349 def _dc_search_uploader(self, html):
1350 return self._html_search_meta('dc.creator', html, 'uploader')
1351
8dbe9899
PH
1352 def _rta_search(self, html):
1353 # See http://www.rtalabel.org/index.php?content=howtofaq#single
1354 if re.search(r'(?ix)<meta\s+name="rating"\s+'
1355 r' content="RTA-5042-1996-1400-1577-RTA"',
1356 html):
1357 return 18
1358 return 0
1359
59040888
PH
1360 def _media_rating_search(self, html):
1361 # See http://www.tjg-designs.com/WP/metadata-code-examples-adding-metadata-to-your-web-pages/
1362 rating = self._html_search_meta('rating', html)
1363
1364 if not rating:
1365 return None
1366
1367 RATING_TABLE = {
1368 'safe for kids': 0,
1369 'general': 8,
1370 '14 years': 14,
1371 'mature': 17,
1372 'restricted': 19,
1373 }
d800609c 1374 return RATING_TABLE.get(rating.lower())
59040888 1375
69319969 1376 def _family_friendly_search(self, html):
6ca7732d 1377 # See http://schema.org/VideoObject
ac8491fc
S
1378 family_friendly = self._html_search_meta(
1379 'isFamilyFriendly', html, default=None)
69319969
NJ
1380
1381 if not family_friendly:
1382 return None
1383
1384 RATING_TABLE = {
1385 '1': 0,
1386 'true': 0,
1387 '0': 18,
1388 'false': 18,
1389 }
d800609c 1390 return RATING_TABLE.get(family_friendly.lower())
69319969 1391
0c708f11
JMF
1392 def _twitter_search_player(self, html):
1393 return self._html_search_meta('twitter:player', html,
9e1a5b84 1394 'twitter card player')
0c708f11 1395
0c36dc00 1396 def _yield_json_ld(self, html, video_id, *, fatal=True, default=NO_DEFAULT):
1397 """Yield all json ld objects in the html"""
1398 if default is not NO_DEFAULT:
1399 fatal = False
1400 for mobj in re.finditer(JSON_LD_RE, html):
1401 json_ld_item = self._parse_json(mobj.group('json_ld'), video_id, fatal=fatal)
1402 for json_ld in variadic(json_ld_item):
1403 if isinstance(json_ld, dict):
1404 yield json_ld
1405
1406 def _search_json_ld(self, html, video_id, expected_type=None, *, fatal=True, default=NO_DEFAULT):
1407 """Search for a video in any json ld in the html"""
1408 if default is not NO_DEFAULT:
1409 fatal = False
1410 info = self._json_ld(
1411 list(self._yield_json_ld(html, video_id, fatal=fatal, default=default)),
1412 video_id, fatal=fatal, expected_type=expected_type)
1413 if info:
1414 return info
4433bb02
S
1415 if default is not NO_DEFAULT:
1416 return default
1417 elif fatal:
1418 raise RegexNotFoundError('Unable to extract JSON-LD')
1419 else:
6a39ee13 1420 self.report_warning('unable to extract JSON-LD %s' % bug_reports_message())
4433bb02 1421 return {}
4ca2a3cf 1422
95b31e26 1423 def _json_ld(self, json_ld, video_id, fatal=True, expected_type=None):
14f25df2 1424 if isinstance(json_ld, str):
4ca2a3cf
S
1425 json_ld = self._parse_json(json_ld, video_id, fatal=fatal)
1426 if not json_ld:
1427 return {}
1428 info = {}
46933a15
S
1429 if not isinstance(json_ld, (list, tuple, dict)):
1430 return info
1431 if isinstance(json_ld, dict):
1432 json_ld = [json_ld]
bae14048 1433
e7e4a6e0
S
1434 INTERACTION_TYPE_MAP = {
1435 'CommentAction': 'comment',
1436 'AgreeAction': 'like',
1437 'DisagreeAction': 'dislike',
1438 'LikeAction': 'like',
1439 'DislikeAction': 'dislike',
1440 'ListenAction': 'view',
1441 'WatchAction': 'view',
1442 'ViewAction': 'view',
1443 }
1444
f3c0c773 1445 def is_type(e, *expected_types):
1446 type = variadic(traverse_obj(e, '@type'))
1447 return any(x in type for x in expected_types)
1448
29f7c58a 1449 def extract_interaction_type(e):
1450 interaction_type = e.get('interactionType')
1451 if isinstance(interaction_type, dict):
1452 interaction_type = interaction_type.get('@type')
1453 return str_or_none(interaction_type)
1454
e7e4a6e0
S
1455 def extract_interaction_statistic(e):
1456 interaction_statistic = e.get('interactionStatistic')
29f7c58a 1457 if isinstance(interaction_statistic, dict):
1458 interaction_statistic = [interaction_statistic]
e7e4a6e0
S
1459 if not isinstance(interaction_statistic, list):
1460 return
1461 for is_e in interaction_statistic:
f3c0c773 1462 if not is_type(is_e, 'InteractionCounter'):
e7e4a6e0 1463 continue
29f7c58a 1464 interaction_type = extract_interaction_type(is_e)
1465 if not interaction_type:
e7e4a6e0 1466 continue
ce5b9040
S
1467 # For interaction count some sites provide string instead of
1468 # an integer (as per spec) with non digit characters (e.g. ",")
1469 # so extracting count with more relaxed str_to_int
1470 interaction_count = str_to_int(is_e.get('userInteractionCount'))
e7e4a6e0
S
1471 if interaction_count is None:
1472 continue
1473 count_kind = INTERACTION_TYPE_MAP.get(interaction_type.split('/')[-1])
1474 if not count_kind:
1475 continue
1476 count_key = '%s_count' % count_kind
1477 if info.get(count_key) is not None:
1478 continue
1479 info[count_key] = interaction_count
1480
f5225737 1481 def extract_chapter_information(e):
1482 chapters = [{
1483 'title': part.get('name'),
1484 'start_time': part.get('startOffset'),
1485 'end_time': part.get('endOffset'),
85553414 1486 } for part in variadic(e.get('hasPart') or []) if part.get('@type') == 'Clip']
f5225737 1487 for idx, (last_c, current_c, next_c) in enumerate(zip(
1488 [{'end_time': 0}] + chapters, chapters, chapters[1:])):
1489 current_c['end_time'] = current_c['end_time'] or next_c['start_time']
1490 current_c['start_time'] = current_c['start_time'] or last_c['end_time']
1491 if None in current_c.values():
1492 self.report_warning(f'Chapter {idx} contains broken data. Not extracting chapters')
1493 return
1494 if chapters:
1495 chapters[-1]['end_time'] = chapters[-1]['end_time'] or info['duration']
1496 info['chapters'] = chapters
1497
bae14048 1498 def extract_video_object(e):
f3c0c773 1499 assert is_type(e, 'VideoObject')
f7ad7160 1500 author = e.get('author')
bae14048 1501 info.update({
0c36dc00 1502 'url': url_or_none(e.get('contentUrl')),
bae14048
S
1503 'title': unescapeHTML(e.get('name')),
1504 'description': unescapeHTML(e.get('description')),
21633673 1505 'thumbnails': [{'url': url}
1506 for url in variadic(traverse_obj(e, 'thumbnailUrl', 'thumbnailURL'))
1507 if url_or_none(url)],
bae14048
S
1508 'duration': parse_duration(e.get('duration')),
1509 'timestamp': unified_timestamp(e.get('uploadDate')),
f7ad7160 1510 # author can be an instance of 'Organization' or 'Person' types.
1511 # both types can have 'name' property(inherited from 'Thing' type). [1]
1512 # however some websites are using 'Text' type instead.
1513 # 1. https://schema.org/VideoObject
14f25df2 1514 'uploader': author.get('name') if isinstance(author, dict) else author if isinstance(author, str) else None,
56ba69e4 1515 'filesize': int_or_none(float_or_none(e.get('contentSize'))),
bae14048
S
1516 'tbr': int_or_none(e.get('bitrate')),
1517 'width': int_or_none(e.get('width')),
1518 'height': int_or_none(e.get('height')),
33a81c2c 1519 'view_count': int_or_none(e.get('interactionCount')),
bae14048 1520 })
e7e4a6e0 1521 extract_interaction_statistic(e)
f5225737 1522 extract_chapter_information(e)
bae14048 1523
d5c32548
ZM
1524 def traverse_json_ld(json_ld, at_top_level=True):
1525 for e in json_ld:
1526 if at_top_level and '@context' not in e:
1527 continue
1528 if at_top_level and set(e.keys()) == {'@context', '@graph'}:
1529 traverse_json_ld(variadic(e['@graph'], allowed_types=(dict,)), at_top_level=False)
1530 break
f3c0c773 1531 if expected_type is not None and not is_type(e, expected_type):
4433bb02 1532 continue
8f122fa0 1533 rating = traverse_obj(e, ('aggregateRating', 'ratingValue'), expected_type=float_or_none)
1534 if rating is not None:
1535 info['average_rating'] = rating
f3c0c773 1536 if is_type(e, 'TVEpisode', 'Episode'):
440863ad 1537 episode_name = unescapeHTML(e.get('name'))
46933a15 1538 info.update({
440863ad 1539 'episode': episode_name,
46933a15
S
1540 'episode_number': int_or_none(e.get('episodeNumber')),
1541 'description': unescapeHTML(e.get('description')),
1542 })
440863ad
S
1543 if not info.get('title') and episode_name:
1544 info['title'] = episode_name
46933a15 1545 part_of_season = e.get('partOfSeason')
f3c0c773 1546 if is_type(part_of_season, 'TVSeason', 'Season', 'CreativeWorkSeason'):
458fd30f
S
1547 info.update({
1548 'season': unescapeHTML(part_of_season.get('name')),
1549 'season_number': int_or_none(part_of_season.get('seasonNumber')),
1550 })
d16b3c66 1551 part_of_series = e.get('partOfSeries') or e.get('partOfTVSeries')
f3c0c773 1552 if is_type(part_of_series, 'TVSeries', 'Series', 'CreativeWorkSeries'):
46933a15 1553 info['series'] = unescapeHTML(part_of_series.get('name'))
f3c0c773 1554 elif is_type(e, 'Movie'):
391256dc
S
1555 info.update({
1556 'title': unescapeHTML(e.get('name')),
1557 'description': unescapeHTML(e.get('description')),
1558 'duration': parse_duration(e.get('duration')),
1559 'timestamp': unified_timestamp(e.get('dateCreated')),
1560 })
f3c0c773 1561 elif is_type(e, 'Article', 'NewsArticle'):
46933a15
S
1562 info.update({
1563 'timestamp': parse_iso8601(e.get('datePublished')),
1564 'title': unescapeHTML(e.get('headline')),
d5c32548 1565 'description': unescapeHTML(e.get('articleBody') or e.get('description')),
46933a15 1566 })
f3c0c773 1567 if is_type(traverse_obj(e, ('video', 0)), 'VideoObject'):
2edb38e8 1568 extract_video_object(e['video'][0])
f3c0c773 1569 elif is_type(traverse_obj(e, ('subjectOf', 0)), 'VideoObject'):
e50c3500 1570 extract_video_object(e['subjectOf'][0])
f3c0c773 1571 elif is_type(e, 'VideoObject'):
bae14048 1572 extract_video_object(e)
4433bb02
S
1573 if expected_type is None:
1574 continue
1575 else:
1576 break
c69701c6 1577 video = e.get('video')
f3c0c773 1578 if is_type(video, 'VideoObject'):
c69701c6 1579 extract_video_object(video)
4433bb02
S
1580 if expected_type is None:
1581 continue
1582 else:
1583 break
d5c32548
ZM
1584 traverse_json_ld(json_ld)
1585
90137ca4 1586 return filter_dict(info)
4ca2a3cf 1587
135dfa2c 1588 def _search_nextjs_data(self, webpage, video_id, *, transform_source=None, fatal=True, **kw):
f98709af
LL
1589 return self._parse_json(
1590 self._search_regex(
1591 r'(?s)<script[^>]+id=[\'"]__NEXT_DATA__[\'"][^>]*>([^<]+)</script>',
135dfa2c 1592 webpage, 'next.js data', fatal=fatal, **kw),
1593 video_id, transform_source=transform_source, fatal=fatal)
f98709af 1594
8072ef2b 1595 def _search_nuxt_data(self, webpage, video_id, context_name='__NUXT__', *, fatal=True, traverse=('data', 0)):
1596 """Parses Nuxt.js metadata. This works as long as the function __NUXT__ invokes is a pure function"""
66f4c04e 1597 rectx = re.escape(context_name)
8072ef2b 1598 FUNCTION_RE = r'\(function\((?P<arg_keys>.*?)\){return\s+(?P<js>{.*?})\s*;?\s*}\((?P<arg_vals>.*?)\)'
66f4c04e 1599 js, arg_keys, arg_vals = self._search_regex(
8072ef2b 1600 (rf'<script>\s*window\.{rectx}={FUNCTION_RE}\s*\)\s*;?\s*</script>', rf'{rectx}\(.*?{FUNCTION_RE}'),
1601 webpage, context_name, group=('js', 'arg_keys', 'arg_vals'), fatal=fatal)
66f4c04e
THD
1602
1603 args = dict(zip(arg_keys.split(','), arg_vals.split(',')))
1604
1605 for key, val in args.items():
1606 if val in ('undefined', 'void 0'):
1607 args[key] = 'null'
1608
8072ef2b 1609 ret = self._parse_json(js, video_id, transform_source=functools.partial(js_to_json, vars=args), fatal=fatal)
1610 return traverse_obj(ret, traverse) or {}
66f4c04e 1611
27713812 1612 @staticmethod
f8da79f8 1613 def _hidden_inputs(html):
586f1cc5 1614 html = re.sub(r'<!--(?:(?!<!--).)*-->', '', html)
201ea3ee 1615 hidden_inputs = {}
c8498368
S
1616 for input in re.findall(r'(?i)(<input[^>]+>)', html):
1617 attrs = extract_attributes(input)
1618 if not input:
201ea3ee 1619 continue
c8498368 1620 if attrs.get('type') not in ('hidden', 'submit'):
201ea3ee 1621 continue
c8498368
S
1622 name = attrs.get('name') or attrs.get('id')
1623 value = attrs.get('value')
1624 if name and value is not None:
1625 hidden_inputs[name] = value
201ea3ee 1626 return hidden_inputs
27713812 1627
cf61d96d
S
1628 def _form_hidden_inputs(self, form_id, html):
1629 form = self._search_regex(
73eb13df 1630 r'(?is)<form[^>]+?id=(["\'])%s\1[^>]*>(?P<form>.+?)</form>' % form_id,
cf61d96d
S
1631 html, '%s form' % form_id, group='form')
1632 return self._hidden_inputs(form)
1633
eb8a4433 1634 class FormatSort:
b050d210 1635 regex = r' *((?P<reverse>\+)?(?P<field>[a-zA-Z0-9_]+)((?P<separator>[~:])(?P<limit>.*?))?)? *$'
eb8a4433 1636
8326b00a 1637 default = ('hidden', 'aud_or_vid', 'hasvid', 'ie_pref', 'lang', 'quality',
176f1866 1638 'res', 'fps', 'hdr:12', 'codec:vp9.2', 'size', 'br', 'asr',
f304da8a 1639 'proto', 'ext', 'hasaud', 'source', 'id') # These must not be aliases
198e3a04 1640 ytdl_default = ('hasaud', 'lang', 'quality', 'tbr', 'filesize', 'vbr',
53ed7066 1641 'height', 'width', 'proto', 'vext', 'abr', 'aext',
f304da8a 1642 'fps', 'fs_approx', 'source', 'id')
eb8a4433 1643
1644 settings = {
1645 'vcodec': {'type': 'ordered', 'regex': True,
155d2b48 1646 'order': ['av0?1', 'vp0?9.2', 'vp0?9', '[hx]265|he?vc?', '[hx]264|avc', 'vp0?8', 'mp4v|h263', 'theora', '', None, 'none']},
eb8a4433 1647 'acodec': {'type': 'ordered', 'regex': True,
a10aa588 1648 'order': ['[af]lac', 'wav|aiff', 'opus', 'vorbis|ogg', 'aac', 'mp?4a?', 'mp3', 'e-?a?c-?3', 'ac-?3', 'dts', '', None, 'none']},
176f1866 1649 'hdr': {'type': 'ordered', 'regex': True, 'field': 'dynamic_range',
1650 'order': ['dv', '(hdr)?12', r'(hdr)?10\+', '(hdr)?10', 'hlg', '', 'sdr', None]},
f137c99e 1651 'proto': {'type': 'ordered', 'regex': True, 'field': 'protocol',
f304da8a 1652 'order': ['(ht|f)tps', '(ht|f)tp$', 'm3u8.*', '.*dash', 'websocket_frag', 'rtmpe?', '', 'mms|rtsp', 'ws|websocket', 'f4']},
eb8a4433 1653 'vext': {'type': 'ordered', 'field': 'video_ext',
91ebc640 1654 'order': ('mp4', 'webm', 'flv', '', 'none'),
eb8a4433 1655 'order_free': ('webm', 'mp4', 'flv', '', 'none')},
1656 'aext': {'type': 'ordered', 'field': 'audio_ext',
1657 'order': ('m4a', 'aac', 'mp3', 'ogg', 'opus', 'webm', '', 'none'),
1658 'order_free': ('opus', 'ogg', 'webm', 'm4a', 'mp3', 'aac', '', 'none')},
1659 'hidden': {'visible': False, 'forced': True, 'type': 'extractor', 'max': -1000},
f5510afe 1660 'aud_or_vid': {'visible': False, 'forced': True, 'type': 'multiple',
8326b00a 1661 'field': ('vcodec', 'acodec'),
1662 'function': lambda it: int(any(v != 'none' for v in it))},
f983b875 1663 'ie_pref': {'priority': True, 'type': 'extractor'},
63be1aab 1664 'hasvid': {'priority': True, 'field': 'vcodec', 'type': 'boolean', 'not_in_list': ('none',)},
1665 'hasaud': {'field': 'acodec', 'type': 'boolean', 'not_in_list': ('none',)},
10beccc9 1666 'lang': {'convert': 'float', 'field': 'language_preference', 'default': -1},
1667 'quality': {'convert': 'float', 'default': -1},
eb8a4433 1668 'filesize': {'convert': 'bytes'},
f137c99e 1669 'fs_approx': {'convert': 'bytes', 'field': 'filesize_approx'},
1670 'id': {'convert': 'string', 'field': 'format_id'},
eb8a4433 1671 'height': {'convert': 'float_none'},
1672 'width': {'convert': 'float_none'},
1673 'fps': {'convert': 'float_none'},
1674 'tbr': {'convert': 'float_none'},
1675 'vbr': {'convert': 'float_none'},
1676 'abr': {'convert': 'float_none'},
1677 'asr': {'convert': 'float_none'},
10beccc9 1678 'source': {'convert': 'float', 'field': 'source_preference', 'default': -1},
63be1aab 1679
eb8a4433 1680 'codec': {'type': 'combined', 'field': ('vcodec', 'acodec')},
63be1aab 1681 'br': {'type': 'combined', 'field': ('tbr', 'vbr', 'abr'), 'same_limit': True},
1682 'size': {'type': 'combined', 'same_limit': True, 'field': ('filesize', 'fs_approx')},
1683 'ext': {'type': 'combined', 'field': ('vext', 'aext')},
f5510afe 1684 'res': {'type': 'multiple', 'field': ('height', 'width'),
dbf5416a 1685 'function': lambda it: (lambda l: min(l) if l else 0)(tuple(filter(None, it)))},
63be1aab 1686
19188702 1687 # For compatibility with youtube-dl
1688 'format_id': {'type': 'alias', 'field': 'id'},
1689 'preference': {'type': 'alias', 'field': 'ie_pref'},
1690 'language_preference': {'type': 'alias', 'field': 'lang'},
63be1aab 1691 'source_preference': {'type': 'alias', 'field': 'source'},
08d30158 1692 'protocol': {'type': 'alias', 'field': 'proto'},
63be1aab 1693 'filesize_approx': {'type': 'alias', 'field': 'fs_approx'},
08d30158 1694
1695 # Deprecated
1696 'dimension': {'type': 'alias', 'field': 'res', 'deprecated': True},
1697 'resolution': {'type': 'alias', 'field': 'res', 'deprecated': True},
1698 'extension': {'type': 'alias', 'field': 'ext', 'deprecated': True},
1699 'bitrate': {'type': 'alias', 'field': 'br', 'deprecated': True},
1700 'total_bitrate': {'type': 'alias', 'field': 'tbr', 'deprecated': True},
1701 'video_bitrate': {'type': 'alias', 'field': 'vbr', 'deprecated': True},
1702 'audio_bitrate': {'type': 'alias', 'field': 'abr', 'deprecated': True},
1703 'framerate': {'type': 'alias', 'field': 'fps', 'deprecated': True},
1704 'filesize_estimate': {'type': 'alias', 'field': 'size', 'deprecated': True},
1705 'samplerate': {'type': 'alias', 'field': 'asr', 'deprecated': True},
1706 'video_ext': {'type': 'alias', 'field': 'vext', 'deprecated': True},
1707 'audio_ext': {'type': 'alias', 'field': 'aext', 'deprecated': True},
1708 'video_codec': {'type': 'alias', 'field': 'vcodec', 'deprecated': True},
1709 'audio_codec': {'type': 'alias', 'field': 'acodec', 'deprecated': True},
1710 'video': {'type': 'alias', 'field': 'hasvid', 'deprecated': True},
1711 'has_video': {'type': 'alias', 'field': 'hasvid', 'deprecated': True},
1712 'audio': {'type': 'alias', 'field': 'hasaud', 'deprecated': True},
1713 'has_audio': {'type': 'alias', 'field': 'hasaud', 'deprecated': True},
1714 'extractor': {'type': 'alias', 'field': 'ie_pref', 'deprecated': True},
1715 'extractor_preference': {'type': 'alias', 'field': 'ie_pref', 'deprecated': True},
63be1aab 1716 }
eb8a4433 1717
f304da8a 1718 def __init__(self, ie, field_preference):
1719 self._order = []
1720 self.ydl = ie._downloader
1721 self.evaluate_params(self.ydl.params, field_preference)
1722 if ie.get_param('verbose'):
1723 self.print_verbose_info(self.ydl.write_debug)
eb8a4433 1724
1725 def _get_field_setting(self, field, key):
1726 if field not in self.settings:
ee8dd27a 1727 if key in ('forced', 'priority'):
1728 return False
1729 self.ydl.deprecation_warning(
1730 f'Using arbitrary fields ({field}) for format sorting is deprecated '
1731 'and may be removed in a future version')
eb8a4433 1732 self.settings[field] = {}
1733 propObj = self.settings[field]
1734 if key not in propObj:
1735 type = propObj.get('type')
1736 if key == 'field':
1737 default = 'preference' if type == 'extractor' else (field,) if type in ('combined', 'multiple') else field
1738 elif key == 'convert':
1739 default = 'order' if type == 'ordered' else 'float_string' if field else 'ignore'
4bcc7bd1 1740 else:
f5510afe 1741 default = {'type': 'field', 'visible': True, 'order': [], 'not_in_list': (None,)}.get(key, None)
eb8a4433 1742 propObj[key] = default
1743 return propObj[key]
1744
1745 def _resolve_field_value(self, field, value, convertNone=False):
1746 if value is None:
1747 if not convertNone:
1748 return None
4bcc7bd1 1749 else:
eb8a4433 1750 value = value.lower()
1751 conversion = self._get_field_setting(field, 'convert')
1752 if conversion == 'ignore':
1753 return None
1754 if conversion == 'string':
1755 return value
1756 elif conversion == 'float_none':
1757 return float_or_none(value)
1758 elif conversion == 'bytes':
1759 return FileDownloader.parse_bytes(value)
1760 elif conversion == 'order':
da9be05e 1761 order_list = (self._use_free_order and self._get_field_setting(field, 'order_free')) or self._get_field_setting(field, 'order')
eb8a4433 1762 use_regex = self._get_field_setting(field, 'regex')
1763 list_length = len(order_list)
1764 empty_pos = order_list.index('') if '' in order_list else list_length + 1
1765 if use_regex and value is not None:
da9be05e 1766 for i, regex in enumerate(order_list):
eb8a4433 1767 if regex and re.match(regex, value):
1768 return list_length - i
1769 return list_length - empty_pos # not in list
1770 else: # not regex or value = None
1771 return list_length - (order_list.index(value) if value in order_list else empty_pos)
1772 else:
1773 if value.isnumeric():
1774 return float(value)
4bcc7bd1 1775 else:
eb8a4433 1776 self.settings[field]['convert'] = 'string'
1777 return value
1778
1779 def evaluate_params(self, params, sort_extractor):
1780 self._use_free_order = params.get('prefer_free_formats', False)
1781 self._sort_user = params.get('format_sort', [])
1782 self._sort_extractor = sort_extractor
1783
1784 def add_item(field, reverse, closest, limit_text):
1785 field = field.lower()
1786 if field in self._order:
1787 return
1788 self._order.append(field)
1789 limit = self._resolve_field_value(field, limit_text)
1790 data = {
1791 'reverse': reverse,
1792 'closest': False if limit is None else closest,
1793 'limit_text': limit_text,
1794 'limit': limit}
1795 if field in self.settings:
1796 self.settings[field].update(data)
1797 else:
1798 self.settings[field] = data
1799
1800 sort_list = (
1801 tuple(field for field in self.default if self._get_field_setting(field, 'forced'))
1802 + (tuple() if params.get('format_sort_force', False)
1803 else tuple(field for field in self.default if self._get_field_setting(field, 'priority')))
1804 + tuple(self._sort_user) + tuple(sort_extractor) + self.default)
1805
1806 for item in sort_list:
1807 match = re.match(self.regex, item)
1808 if match is None:
1809 raise ExtractorError('Invalid format sort string "%s" given by extractor' % item)
1810 field = match.group('field')
1811 if field is None:
1812 continue
1813 if self._get_field_setting(field, 'type') == 'alias':
ee8dd27a 1814 alias, field = field, self._get_field_setting(field, 'field')
08d30158 1815 if self._get_field_setting(alias, 'deprecated'):
19188702 1816 self.ydl.deprecation_warning(
1817 f'Format sorting alias {alias} is deprecated '
1818 f'and may be removed in a future version. Please use {field} instead')
eb8a4433 1819 reverse = match.group('reverse') is not None
b050d210 1820 closest = match.group('separator') == '~'
eb8a4433 1821 limit_text = match.group('limit')
1822
1823 has_limit = limit_text is not None
1824 has_multiple_fields = self._get_field_setting(field, 'type') == 'combined'
1825 has_multiple_limits = has_limit and has_multiple_fields and not self._get_field_setting(field, 'same_limit')
1826
1827 fields = self._get_field_setting(field, 'field') if has_multiple_fields else (field,)
b5ae35ee 1828 limits = limit_text.split(':') if has_multiple_limits else (limit_text,) if has_limit else tuple()
eb8a4433 1829 limit_count = len(limits)
1830 for (i, f) in enumerate(fields):
1831 add_item(f, reverse, closest,
1832 limits[i] if i < limit_count
1833 else limits[0] if has_limit and not has_multiple_limits
1834 else None)
1835
0760b0a7 1836 def print_verbose_info(self, write_debug):
b31fdeed 1837 if self._sort_user:
0760b0a7 1838 write_debug('Sort order given by user: %s' % ', '.join(self._sort_user))
eb8a4433 1839 if self._sort_extractor:
0760b0a7 1840 write_debug('Sort order given by extractor: %s' % ', '.join(self._sort_extractor))
1841 write_debug('Formats sorted by: %s' % ', '.join(['%s%s%s' % (
eb8a4433 1842 '+' if self._get_field_setting(field, 'reverse') else '', field,
1843 '%s%s(%s)' % ('~' if self._get_field_setting(field, 'closest') else ':',
1844 self._get_field_setting(field, 'limit_text'),
1845 self._get_field_setting(field, 'limit'))
1846 if self._get_field_setting(field, 'limit_text') is not None else '')
1847 for field in self._order if self._get_field_setting(field, 'visible')]))
1848
1849 def _calculate_field_preference_from_value(self, format, field, type, value):
1850 reverse = self._get_field_setting(field, 'reverse')
1851 closest = self._get_field_setting(field, 'closest')
1852 limit = self._get_field_setting(field, 'limit')
1853
1854 if type == 'extractor':
1855 maximum = self._get_field_setting(field, 'max')
1856 if value is None or (maximum is not None and value >= maximum):
f983b875 1857 value = -1
eb8a4433 1858 elif type == 'boolean':
1859 in_list = self._get_field_setting(field, 'in_list')
1860 not_in_list = self._get_field_setting(field, 'not_in_list')
1861 value = 0 if ((in_list is None or value in in_list) and (not_in_list is None or value not in not_in_list)) else -1
1862 elif type == 'ordered':
1863 value = self._resolve_field_value(field, value, True)
1864
1865 # try to convert to number
6a04a74e 1866 val_num = float_or_none(value, default=self._get_field_setting(field, 'default'))
eb8a4433 1867 is_num = self._get_field_setting(field, 'convert') != 'string' and val_num is not None
1868 if is_num:
1869 value = val_num
1870
1871 return ((-10, 0) if value is None
1872 else (1, value, 0) if not is_num # if a field has mixed strings and numbers, strings are sorted higher
1873 else (0, -abs(value - limit), value - limit if reverse else limit - value) if closest
1874 else (0, value, 0) if not reverse and (limit is None or value <= limit)
1875 else (0, -value, 0) if limit is None or (reverse and value == limit) or value > limit
1876 else (-1, value, 0))
1877
1878 def _calculate_field_preference(self, format, field):
1879 type = self._get_field_setting(field, 'type') # extractor, boolean, ordered, field, multiple
1880 get_value = lambda f: format.get(self._get_field_setting(f, 'field'))
1881 if type == 'multiple':
1882 type = 'field' # Only 'field' is allowed in multiple for now
1883 actual_fields = self._get_field_setting(field, 'field')
1884
f5510afe 1885 value = self._get_field_setting(field, 'function')(get_value(f) for f in actual_fields)
eb8a4433 1886 else:
1887 value = get_value(field)
1888 return self._calculate_field_preference_from_value(format, field, type, value)
1889
1890 def calculate_preference(self, format):
1891 # Determine missing protocol
1892 if not format.get('protocol'):
1893 format['protocol'] = determine_protocol(format)
1894
1895 # Determine missing ext
1896 if not format.get('ext') and 'url' in format:
1897 format['ext'] = determine_ext(format['url'])
1898 if format.get('vcodec') == 'none':
8326b00a 1899 format['audio_ext'] = format['ext'] if format.get('acodec') != 'none' else 'none'
eb8a4433 1900 format['video_ext'] = 'none'
1901 else:
1902 format['video_ext'] = format['ext']
1903 format['audio_ext'] = 'none'
1904 # if format.get('preference') is None and format.get('ext') in ('f4f', 'f4m'): # Not supported?
1905 # format['preference'] = -1000
1906
1907 # Determine missing bitrates
1908 if format.get('tbr') is None:
1909 if format.get('vbr') is not None and format.get('abr') is not None:
1910 format['tbr'] = format.get('vbr', 0) + format.get('abr', 0)
1911 else:
b5ae35ee 1912 if format.get('vcodec') != 'none' and format.get('vbr') is None:
eb8a4433 1913 format['vbr'] = format.get('tbr') - format.get('abr', 0)
b5ae35ee 1914 if format.get('acodec') != 'none' and format.get('abr') is None:
eb8a4433 1915 format['abr'] = format.get('tbr') - format.get('vbr', 0)
1916
1917 return tuple(self._calculate_field_preference(format, field) for field in self._order)
1918
1919 def _sort_formats(self, formats, field_preference=[]):
1920 if not formats:
88acdbc2 1921 return
1d485a1a 1922 formats.sort(key=self.FormatSort(self, field_preference).calculate_preference)
59040888 1923
96a53167
S
1924 def _check_formats(self, formats, video_id):
1925 if formats:
1926 formats[:] = filter(
1927 lambda f: self._is_valid_url(
1928 f['url'], video_id,
1929 item='%s video format' % f.get('format_id') if f.get('format_id') else 'video'),
1930 formats)
1931
f5bdb444
S
1932 @staticmethod
1933 def _remove_duplicate_formats(formats):
1934 format_urls = set()
1935 unique_formats = []
1936 for f in formats:
1937 if f['url'] not in format_urls:
1938 format_urls.add(f['url'])
1939 unique_formats.append(f)
1940 formats[:] = unique_formats
1941
45024183 1942 def _is_valid_url(self, url, video_id, item='video', headers={}):
2f0f6578
S
1943 url = self._proto_relative_url(url, scheme='http:')
1944 # For now assume non HTTP(S) URLs always valid
1945 if not (url.startswith('http://') or url.startswith('https://')):
1946 return True
96a53167 1947 try:
45024183 1948 self._request_webpage(url, video_id, 'Checking %s URL' % item, headers=headers)
96a53167 1949 return True
8bdd16b4 1950 except ExtractorError as e:
25e911a9 1951 self.to_screen(
8bdd16b4 1952 '%s: %s URL is invalid, skipping: %s'
1953 % (video_id, item, error_to_compat_str(e.cause)))
25e911a9 1954 return False
96a53167 1955
20991253 1956 def http_scheme(self):
1ede5b24 1957 """ Either "http:" or "https:", depending on the user's preferences """
20991253
PH
1958 return (
1959 'http:'
a06916d9 1960 if self.get_param('prefer_insecure', False)
20991253
PH
1961 else 'https:')
1962
57c7411f
PH
1963 def _proto_relative_url(self, url, scheme=None):
1964 if url is None:
1965 return url
1966 if url.startswith('//'):
1967 if scheme is None:
1968 scheme = self.http_scheme()
1969 return scheme + url
1970 else:
1971 return url
1972
4094b6e3
PH
1973 def _sleep(self, timeout, video_id, msg_template=None):
1974 if msg_template is None:
f1a9d64e 1975 msg_template = '%(video_id)s: Waiting for %(timeout)s seconds'
4094b6e3
PH
1976 msg = msg_template % {'video_id': video_id, 'timeout': timeout}
1977 self.to_screen(msg)
1978 time.sleep(timeout)
1979
f983b875 1980 def _extract_f4m_formats(self, manifest_url, video_id, preference=None, quality=None, f4m_id=None,
4de61310 1981 transform_source=lambda s: fix_xml_ampersands(s).strip(),
7360c06f 1982 fatal=True, m3u8_id=None, data=None, headers={}, query={}):
a076c1f9 1983 res = self._download_xml_handle(
f036a632 1984 manifest_url, video_id, 'Downloading f4m manifest',
97f4aecf
S
1985 'Unable to download f4m manifest',
1986 # Some manifests may be malformed, e.g. prosiebensat1 generated manifests
067aa17e 1987 # (see https://github.com/ytdl-org/youtube-dl/issues/6215#issuecomment-121704244)
4de61310 1988 transform_source=transform_source,
7360c06f 1989 fatal=fatal, data=data, headers=headers, query=query)
a076c1f9 1990 if res is False:
8d29e47f 1991 return []
31bb8d3f 1992
a076c1f9
E
1993 manifest, urlh = res
1994 manifest_url = urlh.geturl()
1995
0fdbb332 1996 return self._parse_f4m_formats(
f983b875 1997 manifest, manifest_url, video_id, preference=preference, quality=quality, f4m_id=f4m_id,
448bb5f3 1998 transform_source=transform_source, fatal=fatal, m3u8_id=m3u8_id)
0fdbb332 1999
f983b875 2000 def _parse_f4m_formats(self, manifest, manifest_url, video_id, preference=None, quality=None, f4m_id=None,
0fdbb332 2001 transform_source=lambda s: fix_xml_ampersands(s).strip(),
448bb5f3 2002 fatal=True, m3u8_id=None):
f9934b96 2003 if not isinstance(manifest, xml.etree.ElementTree.Element) and not fatal:
d9eb580a
S
2004 return []
2005
7a5c1cfe 2006 # currently yt-dlp cannot decode the playerVerificationChallenge as Akamai uses Adobe Alchemy
fb72ec58 2007 akamai_pv = manifest.find('{http://ns.adobe.com/f4m/1.0}pv-2.0')
2008 if akamai_pv is not None and ';' in akamai_pv.text:
2009 playerVerificationChallenge = akamai_pv.text.split(';')[0]
2010 if playerVerificationChallenge.strip() != '':
2011 return []
2012
31bb8d3f 2013 formats = []
7a47d07c 2014 manifest_version = '1.0'
b2527359 2015 media_nodes = manifest.findall('{http://ns.adobe.com/f4m/1.0}media')
34e48bed 2016 if not media_nodes:
7a47d07c 2017 manifest_version = '2.0'
34e48bed 2018 media_nodes = manifest.findall('{http://ns.adobe.com/f4m/2.0}media')
b22ca762 2019 # Remove unsupported DRM protected media from final formats
067aa17e 2020 # rendition (see https://github.com/ytdl-org/youtube-dl/issues/8573).
b22ca762
S
2021 media_nodes = remove_encrypted_media(media_nodes)
2022 if not media_nodes:
2023 return formats
48107c19
S
2024
2025 manifest_base_url = get_base_url(manifest)
0a5685b2 2026
a6571f10 2027 bootstrap_info = xpath_element(
0a5685b2
YCH
2028 manifest, ['{http://ns.adobe.com/f4m/1.0}bootstrapInfo', '{http://ns.adobe.com/f4m/2.0}bootstrapInfo'],
2029 'bootstrap info', default=None)
2030
edd6074c
RA
2031 vcodec = None
2032 mime_type = xpath_text(
2033 manifest, ['{http://ns.adobe.com/f4m/1.0}mimeType', '{http://ns.adobe.com/f4m/2.0}mimeType'],
2034 'base URL', default=None)
2035 if mime_type and mime_type.startswith('audio/'):
2036 vcodec = 'none'
2037
b2527359 2038 for i, media_el in enumerate(media_nodes):
77b8b4e6
S
2039 tbr = int_or_none(media_el.attrib.get('bitrate'))
2040 width = int_or_none(media_el.attrib.get('width'))
2041 height = int_or_none(media_el.attrib.get('height'))
34921b43 2042 format_id = join_nonempty(f4m_id, tbr or i)
448bb5f3
YCH
2043 # If <bootstrapInfo> is present, the specified f4m is a
2044 # stream-level manifest, and only set-level manifests may refer to
2045 # external resources. See section 11.4 and section 4 of F4M spec
2046 if bootstrap_info is None:
2047 media_url = None
2048 # @href is introduced in 2.0, see section 11.6 of F4M spec
2049 if manifest_version == '2.0':
2050 media_url = media_el.attrib.get('href')
2051 if media_url is None:
2052 media_url = media_el.attrib.get('url')
31c746e5
S
2053 if not media_url:
2054 continue
cc357c4d
S
2055 manifest_url = (
2056 media_url if media_url.startswith('http://') or media_url.startswith('https://')
48107c19 2057 else ((manifest_base_url or '/'.join(manifest_url.split('/')[:-1])) + '/' + media_url))
70f0f5a8
S
2058 # If media_url is itself a f4m manifest do the recursive extraction
2059 # since bitrates in parent manifest (this one) and media_url manifest
2060 # may differ leading to inability to resolve the format by requested
2061 # bitrate in f4m downloader
240b6045
YCH
2062 ext = determine_ext(manifest_url)
2063 if ext == 'f4m':
77b8b4e6 2064 f4m_formats = self._extract_f4m_formats(
f983b875 2065 manifest_url, video_id, preference=preference, quality=quality, f4m_id=f4m_id,
77b8b4e6
S
2066 transform_source=transform_source, fatal=fatal)
2067 # Sometimes stream-level manifest contains single media entry that
2068 # does not contain any quality metadata (e.g. http://matchtv.ru/#live-player).
2069 # At the same time parent's media entry in set-level manifest may
2070 # contain it. We will copy it from parent in such cases.
2071 if len(f4m_formats) == 1:
2072 f = f4m_formats[0]
2073 f.update({
2074 'tbr': f.get('tbr') or tbr,
2075 'width': f.get('width') or width,
2076 'height': f.get('height') or height,
2077 'format_id': f.get('format_id') if not tbr else format_id,
edd6074c 2078 'vcodec': vcodec,
77b8b4e6
S
2079 })
2080 formats.extend(f4m_formats)
70f0f5a8 2081 continue
240b6045
YCH
2082 elif ext == 'm3u8':
2083 formats.extend(self._extract_m3u8_formats(
2084 manifest_url, video_id, 'mp4', preference=preference,
f983b875 2085 quality=quality, m3u8_id=m3u8_id, fatal=fatal))
240b6045 2086 continue
31bb8d3f 2087 formats.append({
77b8b4e6 2088 'format_id': format_id,
31bb8d3f 2089 'url': manifest_url,
30d0b549 2090 'manifest_url': manifest_url,
a6571f10 2091 'ext': 'flv' if bootstrap_info is not None else None,
187ee66c 2092 'protocol': 'f4m',
b2527359 2093 'tbr': tbr,
77b8b4e6
S
2094 'width': width,
2095 'height': height,
edd6074c 2096 'vcodec': vcodec,
60ca389c 2097 'preference': preference,
f983b875 2098 'quality': quality,
31bb8d3f 2099 })
31bb8d3f
JMF
2100 return formats
2101
f983b875 2102 def _m3u8_meta_format(self, m3u8_url, ext=None, preference=None, quality=None, m3u8_id=None):
16da9bbc 2103 return {
34921b43 2104 'format_id': join_nonempty(m3u8_id, 'meta'),
704df56d
PH
2105 'url': m3u8_url,
2106 'ext': ext,
2107 'protocol': 'm3u8',
37768f92 2108 'preference': preference - 100 if preference else -100,
f983b875 2109 'quality': quality,
704df56d
PH
2110 'resolution': 'multiple',
2111 'format_note': 'Quality selection URL',
16da9bbc
YCH
2112 }
2113
b5ae35ee 2114 def _report_ignoring_subs(self, name):
2115 self.report_warning(bug_reports_message(
2116 f'Ignoring subtitle tracks found in the {name} manifest; '
2117 'if any subtitle tracks are missing,'
2118 ), only_once=True)
2119
a0c3b2d5
F
2120 def _extract_m3u8_formats(self, *args, **kwargs):
2121 fmts, subs = self._extract_m3u8_formats_and_subtitles(*args, **kwargs)
2122 if subs:
b5ae35ee 2123 self._report_ignoring_subs('HLS')
a0c3b2d5
F
2124 return fmts
2125
2126 def _extract_m3u8_formats_and_subtitles(
177877c5 2127 self, m3u8_url, video_id, ext=None, entry_protocol='m3u8_native',
a0c3b2d5
F
2128 preference=None, quality=None, m3u8_id=None, note=None,
2129 errnote=None, fatal=True, live=False, data=None, headers={},
2130 query={}):
2131
dbd82a1d 2132 res = self._download_webpage_handle(
81515ad9 2133 m3u8_url, video_id,
37a3bb66 2134 note='Downloading m3u8 information' if note is None else note,
2135 errnote='Failed to download m3u8 information' if errnote is None else errnote,
7360c06f 2136 fatal=fatal, data=data, headers=headers, query=query)
cb252080 2137
dbd82a1d 2138 if res is False:
a0c3b2d5 2139 return [], {}
cb252080 2140
dbd82a1d 2141 m3u8_doc, urlh = res
37113045 2142 m3u8_url = urlh.geturl()
9cdffeeb 2143
a0c3b2d5 2144 return self._parse_m3u8_formats_and_subtitles(
cb252080 2145 m3u8_doc, m3u8_url, ext=ext, entry_protocol=entry_protocol,
310c2ed2 2146 preference=preference, quality=quality, m3u8_id=m3u8_id,
2147 note=note, errnote=errnote, fatal=fatal, live=live, data=data,
2148 headers=headers, query=query, video_id=video_id)
cb252080 2149
a0c3b2d5 2150 def _parse_m3u8_formats_and_subtitles(
42676437 2151 self, m3u8_doc, m3u8_url=None, ext=None, entry_protocol='m3u8_native',
a0c3b2d5
F
2152 preference=None, quality=None, m3u8_id=None, live=False, note=None,
2153 errnote=None, fatal=True, data=None, headers={}, query={},
2154 video_id=None):
60755938 2155 formats, subtitles = [], {}
a0c3b2d5 2156
6b993ca7 2157 has_drm = re.search('|'.join([
2158 r'#EXT-X-FAXS-CM:', # Adobe Flash Access
2159 r'#EXT-X-(?:SESSION-)?KEY:.*?URI="skd://', # Apple FairPlay
2160 ]), m3u8_doc)
a0c3b2d5 2161
60755938 2162 def format_url(url):
14f25df2 2163 return url if re.match(r'^https?://', url) else urllib.parse.urljoin(m3u8_url, url)
60755938 2164
2165 if self.get_param('hls_split_discontinuity', False):
2166 def _extract_m3u8_playlist_indices(manifest_url=None, m3u8_doc=None):
2167 if not m3u8_doc:
2168 if not manifest_url:
2169 return []
2170 m3u8_doc = self._download_webpage(
2171 manifest_url, video_id, fatal=fatal, data=data, headers=headers,
2172 note=False, errnote='Failed to download m3u8 playlist information')
2173 if m3u8_doc is False:
2174 return []
2175 return range(1 + sum(line.startswith('#EXT-X-DISCONTINUITY') for line in m3u8_doc.splitlines()))
0def7587 2176
60755938 2177 else:
2178 def _extract_m3u8_playlist_indices(*args, **kwargs):
2179 return [None]
310c2ed2 2180
cb252080
S
2181 # References:
2182 # 1. https://tools.ietf.org/html/draft-pantos-http-live-streaming-21
067aa17e
S
2183 # 2. https://github.com/ytdl-org/youtube-dl/issues/12211
2184 # 3. https://github.com/ytdl-org/youtube-dl/issues/18923
cb252080
S
2185
2186 # We should try extracting formats only from master playlists [1, 4.3.4],
2187 # i.e. playlists that describe available qualities. On the other hand
2188 # media playlists [1, 4.3.3] should be returned as is since they contain
2189 # just the media without qualities renditions.
9cdffeeb 2190 # Fortunately, master playlist can be easily distinguished from media
cb252080 2191 # playlist based on particular tags availability. As of [1, 4.3.3, 4.3.4]
a0566bbf 2192 # master playlist tags MUST NOT appear in a media playlist and vice versa.
cb252080
S
2193 # As of [1, 4.3.3.1] #EXT-X-TARGETDURATION tag is REQUIRED for every
2194 # media playlist and MUST NOT appear in master playlist thus we can
2195 # clearly detect media playlist with this criterion.
2196
9cdffeeb 2197 if '#EXT-X-TARGETDURATION' in m3u8_doc: # media playlist, return as is
60755938 2198 formats = [{
34921b43 2199 'format_id': join_nonempty(m3u8_id, idx),
60755938 2200 'format_index': idx,
42676437 2201 'url': m3u8_url or encode_data_uri(m3u8_doc.encode('utf-8'), 'application/x-mpegurl'),
60755938 2202 'ext': ext,
2203 'protocol': entry_protocol,
2204 'preference': preference,
2205 'quality': quality,
88acdbc2 2206 'has_drm': has_drm,
60755938 2207 } for idx in _extract_m3u8_playlist_indices(m3u8_doc=m3u8_doc)]
310c2ed2 2208
a0c3b2d5 2209 return formats, subtitles
cb252080
S
2210
2211 groups = {}
2212 last_stream_inf = {}
2213
2214 def extract_media(x_media_line):
2215 media = parse_m3u8_attributes(x_media_line)
2216 # As per [1, 4.3.4.1] TYPE, GROUP-ID and NAME are REQUIRED
2217 media_type, group_id, name = media.get('TYPE'), media.get('GROUP-ID'), media.get('NAME')
2218 if not (media_type and group_id and name):
2219 return
2220 groups.setdefault(group_id, []).append(media)
a0c3b2d5
F
2221 # <https://tools.ietf.org/html/rfc8216#section-4.3.4.1>
2222 if media_type == 'SUBTITLES':
3907333c 2223 # According to RFC 8216 §4.3.4.2.1, URI is REQUIRED in the
2224 # EXT-X-MEDIA tag if the media type is SUBTITLES.
2225 # However, lack of URI has been spotted in the wild.
2226 # e.g. NebulaIE; see https://github.com/yt-dlp/yt-dlp/issues/339
2227 if not media.get('URI'):
2228 return
a0c3b2d5
F
2229 url = format_url(media['URI'])
2230 sub_info = {
2231 'url': url,
2232 'ext': determine_ext(url),
2233 }
4a2f19ab
F
2234 if sub_info['ext'] == 'm3u8':
2235 # Per RFC 8216 §3.1, the only possible subtitle format m3u8
2236 # files may contain is WebVTT:
2237 # <https://tools.ietf.org/html/rfc8216#section-3.1>
2238 sub_info['ext'] = 'vtt'
2239 sub_info['protocol'] = 'm3u8_native'
37a3bb66 2240 lang = media.get('LANGUAGE') or 'und'
a0c3b2d5 2241 subtitles.setdefault(lang, []).append(sub_info)
cb252080
S
2242 if media_type not in ('VIDEO', 'AUDIO'):
2243 return
2244 media_url = media.get('URI')
2245 if media_url:
310c2ed2 2246 manifest_url = format_url(media_url)
60755938 2247 formats.extend({
34921b43 2248 'format_id': join_nonempty(m3u8_id, group_id, name, idx),
60755938 2249 'format_note': name,
2250 'format_index': idx,
2251 'url': manifest_url,
2252 'manifest_url': m3u8_url,
2253 'language': media.get('LANGUAGE'),
2254 'ext': ext,
2255 'protocol': entry_protocol,
2256 'preference': preference,
2257 'quality': quality,
2258 'vcodec': 'none' if media_type == 'AUDIO' else None,
2259 } for idx in _extract_m3u8_playlist_indices(manifest_url))
cb252080
S
2260
2261 def build_stream_name():
2262 # Despite specification does not mention NAME attribute for
3019cb0c
S
2263 # EXT-X-STREAM-INF tag it still sometimes may be present (see [1]
2264 # or vidio test in TestInfoExtractor.test_parse_m3u8_formats)
ddd258f9 2265 # 1. http://www.vidio.com/watch/165683-dj_ambred-booyah-live-2015
cb252080
S
2266 stream_name = last_stream_inf.get('NAME')
2267 if stream_name:
2268 return stream_name
2269 # If there is no NAME in EXT-X-STREAM-INF it will be obtained
2270 # from corresponding rendition group
2271 stream_group_id = last_stream_inf.get('VIDEO')
2272 if not stream_group_id:
2273 return
2274 stream_group = groups.get(stream_group_id)
2275 if not stream_group:
2276 return stream_group_id
2277 rendition = stream_group[0]
2278 return rendition.get('NAME') or stream_group_id
2279
379306ef 2280 # parse EXT-X-MEDIA tags before EXT-X-STREAM-INF in order to have the
2bfc1d9d
RA
2281 # chance to detect video only formats when EXT-X-STREAM-INF tags
2282 # precede EXT-X-MEDIA tags in HLS manifest such as [3].
2283 for line in m3u8_doc.splitlines():
2284 if line.startswith('#EXT-X-MEDIA:'):
2285 extract_media(line)
2286
704df56d
PH
2287 for line in m3u8_doc.splitlines():
2288 if line.startswith('#EXT-X-STREAM-INF:'):
cb252080 2289 last_stream_inf = parse_m3u8_attributes(line)
704df56d
PH
2290 elif line.startswith('#') or not line.strip():
2291 continue
2292 else:
9c99bef7 2293 tbr = float_or_none(
3089bc74
S
2294 last_stream_inf.get('AVERAGE-BANDWIDTH')
2295 or last_stream_inf.get('BANDWIDTH'), scale=1000)
30d0b549 2296 manifest_url = format_url(line.strip())
5ef62fc4 2297
60755938 2298 for idx in _extract_m3u8_playlist_indices(manifest_url):
2299 format_id = [m3u8_id, None, idx]
310c2ed2 2300 # Bandwidth of live streams may differ over time thus making
2301 # format_id unpredictable. So it's better to keep provided
2302 # format_id intact.
2303 if not live:
60755938 2304 stream_name = build_stream_name()
34921b43 2305 format_id[1] = stream_name or '%d' % (tbr or len(formats))
310c2ed2 2306 f = {
34921b43 2307 'format_id': join_nonempty(*format_id),
60755938 2308 'format_index': idx,
310c2ed2 2309 'url': manifest_url,
2310 'manifest_url': m3u8_url,
2311 'tbr': tbr,
2312 'ext': ext,
2313 'fps': float_or_none(last_stream_inf.get('FRAME-RATE')),
2314 'protocol': entry_protocol,
2315 'preference': preference,
2316 'quality': quality,
2317 }
2318 resolution = last_stream_inf.get('RESOLUTION')
2319 if resolution:
2320 mobj = re.search(r'(?P<width>\d+)[xX](?P<height>\d+)', resolution)
2321 if mobj:
2322 f['width'] = int(mobj.group('width'))
2323 f['height'] = int(mobj.group('height'))
2324 # Unified Streaming Platform
2325 mobj = re.search(
2326 r'audio.*?(?:%3D|=)(\d+)(?:-video.*?(?:%3D|=)(\d+))?', f['url'])
2327 if mobj:
2328 abr, vbr = mobj.groups()
2329 abr, vbr = float_or_none(abr, 1000), float_or_none(vbr, 1000)
2330 f.update({
2331 'vbr': vbr,
2332 'abr': abr,
2333 })
2334 codecs = parse_codecs(last_stream_inf.get('CODECS'))
2335 f.update(codecs)
2336 audio_group_id = last_stream_inf.get('AUDIO')
2337 # As per [1, 4.3.4.1.1] any EXT-X-STREAM-INF tag which
2338 # references a rendition group MUST have a CODECS attribute.
2339 # However, this is not always respected, for example, [2]
2340 # contains EXT-X-STREAM-INF tag which references AUDIO
2341 # rendition group but does not have CODECS and despite
2342 # referencing an audio group it represents a complete
2343 # (with audio and video) format. So, for such cases we will
2344 # ignore references to rendition groups and treat them
2345 # as complete formats.
2346 if audio_group_id and codecs and f.get('vcodec') != 'none':
2347 audio_group = groups.get(audio_group_id)
2348 if audio_group and audio_group[0].get('URI'):
2349 # TODO: update acodec for audio only formats with
2350 # the same GROUP-ID
2351 f['acodec'] = 'none'
fc21af50 2352 if not f.get('ext'):
2353 f['ext'] = 'm4a' if f.get('vcodec') == 'none' else 'mp4'
310c2ed2 2354 formats.append(f)
2355
2356 # for DailyMotion
2357 progressive_uri = last_stream_inf.get('PROGRESSIVE-URI')
2358 if progressive_uri:
2359 http_f = f.copy()
2360 del http_f['manifest_url']
2361 http_f.update({
2362 'format_id': f['format_id'].replace('hls-', 'http-'),
2363 'protocol': 'http',
2364 'url': progressive_uri,
2365 })
2366 formats.append(http_f)
5ef62fc4 2367
cb252080 2368 last_stream_inf = {}
a0c3b2d5 2369 return formats, subtitles
704df56d 2370
3cf4b91d
C
2371 def _extract_m3u8_vod_duration(
2372 self, m3u8_vod_url, video_id, note=None, errnote=None, data=None, headers={}, query={}):
2373
2374 m3u8_vod = self._download_webpage(
2375 m3u8_vod_url, video_id,
2376 note='Downloading m3u8 VOD manifest' if note is None else note,
2377 errnote='Failed to download VOD manifest' if errnote is None else errnote,
2378 fatal=False, data=data, headers=headers, query=query)
2379
2380 return self._parse_m3u8_vod_duration(m3u8_vod or '', video_id)
2381
2382 def _parse_m3u8_vod_duration(self, m3u8_vod, video_id):
2383 if '#EXT-X-PLAYLIST-TYPE:VOD' not in m3u8_vod:
2384 return None
2385
2386 return int(sum(
2387 float(line[len('#EXTINF:'):].split(',')[0])
2388 for line in m3u8_vod.splitlines() if line.startswith('#EXTINF:'))) or None
2389
a107193e
S
2390 @staticmethod
2391 def _xpath_ns(path, namespace=None):
2392 if not namespace:
2393 return path
2394 out = []
2395 for c in path.split('/'):
2396 if not c or c == '.':
2397 out.append(c)
2398 else:
2399 out.append('{%s}%s' % (namespace, c))
2400 return '/'.join(out)
2401
da1c94ee 2402 def _extract_smil_formats_and_subtitles(self, smil_url, video_id, fatal=True, f4m_params=None, transform_source=None):
a076c1f9
E
2403 res = self._download_smil(smil_url, video_id, fatal=fatal, transform_source=transform_source)
2404 if res is False:
995029a1 2405 assert not fatal
774a46c5 2406 return [], {}
e89a2aab 2407
a076c1f9
E
2408 smil, urlh = res
2409 smil_url = urlh.geturl()
2410
17712eeb 2411 namespace = self._parse_smil_namespace(smil)
a107193e 2412
da1c94ee 2413 fmts = self._parse_smil_formats(
a107193e 2414 smil, smil_url, video_id, namespace=namespace, f4m_params=f4m_params)
da1c94ee
F
2415 subs = self._parse_smil_subtitles(
2416 smil, namespace=namespace)
2417
2418 return fmts, subs
2419
2420 def _extract_smil_formats(self, *args, **kwargs):
2421 fmts, subs = self._extract_smil_formats_and_subtitles(*args, **kwargs)
2422 if subs:
b5ae35ee 2423 self._report_ignoring_subs('SMIL')
da1c94ee 2424 return fmts
a107193e
S
2425
2426 def _extract_smil_info(self, smil_url, video_id, fatal=True, f4m_params=None):
a076c1f9
E
2427 res = self._download_smil(smil_url, video_id, fatal=fatal)
2428 if res is False:
a107193e 2429 return {}
a076c1f9
E
2430
2431 smil, urlh = res
2432 smil_url = urlh.geturl()
2433
a107193e
S
2434 return self._parse_smil(smil, smil_url, video_id, f4m_params=f4m_params)
2435
09f572fb 2436 def _download_smil(self, smil_url, video_id, fatal=True, transform_source=None):
a076c1f9 2437 return self._download_xml_handle(
a107193e 2438 smil_url, video_id, 'Downloading SMIL file',
09f572fb 2439 'Unable to download SMIL file', fatal=fatal, transform_source=transform_source)
a107193e
S
2440
2441 def _parse_smil(self, smil, smil_url, video_id, f4m_params=None):
17712eeb 2442 namespace = self._parse_smil_namespace(smil)
a107193e
S
2443
2444 formats = self._parse_smil_formats(
2445 smil, smil_url, video_id, namespace=namespace, f4m_params=f4m_params)
2446 subtitles = self._parse_smil_subtitles(smil, namespace=namespace)
2447
2448 video_id = os.path.splitext(url_basename(smil_url))[0]
2449 title = None
2450 description = None
647eab45 2451 upload_date = None
a107193e
S
2452 for meta in smil.findall(self._xpath_ns('./head/meta', namespace)):
2453 name = meta.attrib.get('name')
2454 content = meta.attrib.get('content')
2455 if not name or not content:
2456 continue
2457 if not title and name == 'title':
2458 title = content
2459 elif not description and name in ('description', 'abstract'):
2460 description = content
647eab45
S
2461 elif not upload_date and name == 'date':
2462 upload_date = unified_strdate(content)
a107193e 2463
1e5bcdec
S
2464 thumbnails = [{
2465 'id': image.get('type'),
2466 'url': image.get('src'),
2467 'width': int_or_none(image.get('width')),
2468 'height': int_or_none(image.get('height')),
2469 } for image in smil.findall(self._xpath_ns('.//image', namespace)) if image.get('src')]
2470
a107193e
S
2471 return {
2472 'id': video_id,
2473 'title': title or video_id,
2474 'description': description,
647eab45 2475 'upload_date': upload_date,
1e5bcdec 2476 'thumbnails': thumbnails,
a107193e
S
2477 'formats': formats,
2478 'subtitles': subtitles,
2479 }
2480
17712eeb
S
2481 def _parse_smil_namespace(self, smil):
2482 return self._search_regex(
2483 r'(?i)^{([^}]+)?}smil$', smil.tag, 'namespace', default=None)
2484
f877c6ae 2485 def _parse_smil_formats(self, smil, smil_url, video_id, namespace=None, f4m_params=None, transform_rtmp_url=None):
a107193e
S
2486 base = smil_url
2487 for meta in smil.findall(self._xpath_ns('./head/meta', namespace)):
2488 b = meta.get('base') or meta.get('httpBase')
2489 if b:
2490 base = b
2491 break
e89a2aab
S
2492
2493 formats = []
2494 rtmp_count = 0
a107193e 2495 http_count = 0
7f32e5dc 2496 m3u8_count = 0
9359f3d4 2497 imgs_count = 0
a107193e 2498
9359f3d4 2499 srcs = set()
ad96b4c8
YCH
2500 media = smil.findall(self._xpath_ns('.//video', namespace)) + smil.findall(self._xpath_ns('.//audio', namespace))
2501 for medium in media:
2502 src = medium.get('src')
81e1c4e2 2503 if not src or src in srcs:
a107193e 2504 continue
9359f3d4 2505 srcs.add(src)
a107193e 2506
ad96b4c8
YCH
2507 bitrate = float_or_none(medium.get('system-bitrate') or medium.get('systemBitrate'), 1000)
2508 filesize = int_or_none(medium.get('size') or medium.get('fileSize'))
2509 width = int_or_none(medium.get('width'))
2510 height = int_or_none(medium.get('height'))
2511 proto = medium.get('proto')
2512 ext = medium.get('ext')
a107193e 2513 src_ext = determine_ext(src)
ad96b4c8 2514 streamer = medium.get('streamer') or base
a107193e
S
2515
2516 if proto == 'rtmp' or streamer.startswith('rtmp'):
2517 rtmp_count += 1
2518 formats.append({
2519 'url': streamer,
2520 'play_path': src,
2521 'ext': 'flv',
2522 'format_id': 'rtmp-%d' % (rtmp_count if bitrate is None else bitrate),
2523 'tbr': bitrate,
2524 'filesize': filesize,
2525 'width': width,
2526 'height': height,
2527 })
f877c6ae
YCH
2528 if transform_rtmp_url:
2529 streamer, src = transform_rtmp_url(streamer, src)
2530 formats[-1].update({
2531 'url': streamer,
2532 'play_path': src,
2533 })
a107193e
S
2534 continue
2535
14f25df2 2536 src_url = src if src.startswith('http') else urllib.parse.urljoin(base, src)
c349456e 2537 src_url = src_url.strip()
a107193e
S
2538
2539 if proto == 'm3u8' or src_ext == 'm3u8':
7f32e5dc 2540 m3u8_formats = self._extract_m3u8_formats(
2541 src_url, video_id, ext or 'mp4', m3u8_id='hls', fatal=False)
2542 if len(m3u8_formats) == 1:
2543 m3u8_count += 1
2544 m3u8_formats[0].update({
2545 'format_id': 'hls-%d' % (m3u8_count if bitrate is None else bitrate),
2546 'tbr': bitrate,
2547 'width': width,
2548 'height': height,
2549 })
2550 formats.extend(m3u8_formats)
bd21ead2 2551 elif src_ext == 'f4m':
a107193e
S
2552 f4m_url = src_url
2553 if not f4m_params:
2554 f4m_params = {
2555 'hdcore': '3.2.0',
2556 'plugin': 'flowplayer-3.2.0.1',
2557 }
2558 f4m_url += '&' if '?' in f4m_url else '?'
14f25df2 2559 f4m_url += urllib.parse.urlencode(f4m_params)
7e5edcfd 2560 formats.extend(self._extract_f4m_formats(f4m_url, video_id, f4m_id='hds', fatal=False))
bd21ead2
RA
2561 elif src_ext == 'mpd':
2562 formats.extend(self._extract_mpd_formats(
2563 src_url, video_id, mpd_id='dash', fatal=False))
2564 elif re.search(r'\.ism/[Mm]anifest', src_url):
2565 formats.extend(self._extract_ism_formats(
2566 src_url, video_id, ism_id='mss', fatal=False))
2567 elif src_url.startswith('http') and self._is_valid_url(src, video_id):
a107193e
S
2568 http_count += 1
2569 formats.append({
2570 'url': src_url,
2571 'ext': ext or src_ext or 'flv',
2572 'format_id': 'http-%d' % (bitrate or http_count),
2573 'tbr': bitrate,
2574 'filesize': filesize,
2575 'width': width,
2576 'height': height,
2577 })
63757032 2578
9359f3d4
F
2579 for medium in smil.findall(self._xpath_ns('.//imagestream', namespace)):
2580 src = medium.get('src')
2581 if not src or src in srcs:
2582 continue
2583 srcs.add(src)
2584
2585 imgs_count += 1
2586 formats.append({
2587 'format_id': 'imagestream-%d' % (imgs_count),
2588 'url': src,
2589 'ext': mimetype2ext(medium.get('type')),
2590 'acodec': 'none',
2591 'vcodec': 'none',
2592 'width': int_or_none(medium.get('width')),
2593 'height': int_or_none(medium.get('height')),
2594 'format_note': 'SMIL storyboards',
2595 })
2596
e89a2aab
S
2597 return formats
2598
ce00af87 2599 def _parse_smil_subtitles(self, smil, namespace=None, subtitles_lang='en'):
d413095f 2600 urls = []
a107193e
S
2601 subtitles = {}
2602 for num, textstream in enumerate(smil.findall(self._xpath_ns('.//textstream', namespace))):
2603 src = textstream.get('src')
d413095f 2604 if not src or src in urls:
a107193e 2605 continue
d413095f 2606 urls.append(src)
df634be2 2607 ext = textstream.get('ext') or mimetype2ext(textstream.get('type')) or determine_ext(src)
03bc7237 2608 lang = textstream.get('systemLanguage') or textstream.get('systemLanguageName') or textstream.get('lang') or subtitles_lang
a107193e
S
2609 subtitles.setdefault(lang, []).append({
2610 'url': src,
2611 'ext': ext,
2612 })
2613 return subtitles
63757032 2614
47a5cb77 2615 def _extract_xspf_playlist(self, xspf_url, playlist_id, fatal=True):
a076c1f9 2616 res = self._download_xml_handle(
47a5cb77 2617 xspf_url, playlist_id, 'Downloading xpsf playlist',
942acef5 2618 'Unable to download xspf manifest', fatal=fatal)
a076c1f9 2619 if res is False:
942acef5 2620 return []
a076c1f9
E
2621
2622 xspf, urlh = res
2623 xspf_url = urlh.geturl()
2624
47a5cb77
S
2625 return self._parse_xspf(
2626 xspf, playlist_id, xspf_url=xspf_url,
2627 xspf_base_url=base_url(xspf_url))
8d6765cf 2628
47a5cb77 2629 def _parse_xspf(self, xspf_doc, playlist_id, xspf_url=None, xspf_base_url=None):
8d6765cf
S
2630 NS_MAP = {
2631 'xspf': 'http://xspf.org/ns/0/',
2632 's1': 'http://static.streamone.nl/player/ns/0',
2633 }
2634
2635 entries = []
47a5cb77 2636 for track in xspf_doc.findall(xpath_with_ns('./xspf:trackList/xspf:track', NS_MAP)):
8d6765cf 2637 title = xpath_text(
98044462 2638 track, xpath_with_ns('./xspf:title', NS_MAP), 'title', default=playlist_id)
8d6765cf
S
2639 description = xpath_text(
2640 track, xpath_with_ns('./xspf:annotation', NS_MAP), 'description')
2641 thumbnail = xpath_text(
2642 track, xpath_with_ns('./xspf:image', NS_MAP), 'thumbnail')
2643 duration = float_or_none(
2644 xpath_text(track, xpath_with_ns('./xspf:duration', NS_MAP), 'duration'), 1000)
2645
47a5cb77
S
2646 formats = []
2647 for location in track.findall(xpath_with_ns('./xspf:location', NS_MAP)):
2648 format_url = urljoin(xspf_base_url, location.text)
2649 if not format_url:
2650 continue
2651 formats.append({
2652 'url': format_url,
2653 'manifest_url': xspf_url,
2654 'format_id': location.get(xpath_with_ns('s1:label', NS_MAP)),
2655 'width': int_or_none(location.get(xpath_with_ns('s1:width', NS_MAP))),
2656 'height': int_or_none(location.get(xpath_with_ns('s1:height', NS_MAP))),
2657 })
8d6765cf
S
2658 self._sort_formats(formats)
2659
2660 entries.append({
2661 'id': playlist_id,
2662 'title': title,
2663 'description': description,
2664 'thumbnail': thumbnail,
2665 'duration': duration,
2666 'formats': formats,
2667 })
2668 return entries
2669
171e59ed
F
2670 def _extract_mpd_formats(self, *args, **kwargs):
2671 fmts, subs = self._extract_mpd_formats_and_subtitles(*args, **kwargs)
2672 if subs:
b5ae35ee 2673 self._report_ignoring_subs('DASH')
171e59ed
F
2674 return fmts
2675
2676 def _extract_mpd_formats_and_subtitles(
2677 self, mpd_url, video_id, mpd_id=None, note=None, errnote=None,
2678 fatal=True, data=None, headers={}, query={}):
47a5cb77 2679 res = self._download_xml_handle(
1bac3455 2680 mpd_url, video_id,
37a3bb66 2681 note='Downloading MPD manifest' if note is None else note,
2682 errnote='Failed to download MPD manifest' if errnote is None else errnote,
7360c06f 2683 fatal=fatal, data=data, headers=headers, query=query)
1bac3455 2684 if res is False:
171e59ed 2685 return [], {}
47a5cb77 2686 mpd_doc, urlh = res
c25720ef 2687 if mpd_doc is None:
171e59ed 2688 return [], {}
779da8e3
E
2689
2690 # We could have been redirected to a new url when we retrieved our mpd file.
2691 mpd_url = urlh.geturl()
2692 mpd_base_url = base_url(mpd_url)
1bac3455 2693
171e59ed 2694 return self._parse_mpd_formats_and_subtitles(
545cc85d 2695 mpd_doc, mpd_id, mpd_base_url, mpd_url)
2d2fa82d 2696
171e59ed
F
2697 def _parse_mpd_formats(self, *args, **kwargs):
2698 fmts, subs = self._parse_mpd_formats_and_subtitles(*args, **kwargs)
2699 if subs:
b5ae35ee 2700 self._report_ignoring_subs('DASH')
171e59ed
F
2701 return fmts
2702
2703 def _parse_mpd_formats_and_subtitles(
2704 self, mpd_doc, mpd_id=None, mpd_base_url='', mpd_url=None):
f0948348
S
2705 """
2706 Parse formats from MPD manifest.
2707 References:
2708 1. MPEG-DASH Standard, ISO/IEC 23009-1:2014(E),
2709 http://standards.iso.org/ittf/PubliclyAvailableStandards/c065274_ISO_IEC_23009-1_2014.zip
2710 2. https://en.wikipedia.org/wiki/Dynamic_Adaptive_Streaming_over_HTTP
2711 """
a06916d9 2712 if not self.get_param('dynamic_mpd', True):
78895bd3 2713 if mpd_doc.get('type') == 'dynamic':
171e59ed 2714 return [], {}
2d2fa82d 2715
91cb6b50 2716 namespace = self._search_regex(r'(?i)^{([^}]+)?}MPD$', mpd_doc.tag, 'namespace', default=None)
f14be228 2717
2718 def _add_ns(path):
2719 return self._xpath_ns(path, namespace)
2720
675d0016 2721 def is_drm_protected(element):
2722 return element.find(_add_ns('ContentProtection')) is not None
2723
1bac3455 2724 def extract_multisegment_info(element, ms_parent_info):
2725 ms_info = ms_parent_info.copy()
b4c1d6e8
S
2726
2727 # As per [1, 5.3.9.2.2] SegmentList and SegmentTemplate share some
2728 # common attributes and elements. We will only extract relevant
2729 # for us.
2730 def extract_common(source):
2731 segment_timeline = source.find(_add_ns('SegmentTimeline'))
2732 if segment_timeline is not None:
2733 s_e = segment_timeline.findall(_add_ns('S'))
2734 if s_e:
2735 ms_info['total_number'] = 0
2736 ms_info['s'] = []
2737 for s in s_e:
2738 r = int(s.get('r', 0))
2739 ms_info['total_number'] += 1 + r
2740 ms_info['s'].append({
2741 't': int(s.get('t', 0)),
2742 # @d is mandatory (see [1, 5.3.9.6.2, Table 17, page 60])
2743 'd': int(s.attrib['d']),
2744 'r': r,
2745 })
2746 start_number = source.get('startNumber')
2747 if start_number:
2748 ms_info['start_number'] = int(start_number)
2749 timescale = source.get('timescale')
2750 if timescale:
2751 ms_info['timescale'] = int(timescale)
2752 segment_duration = source.get('duration')
2753 if segment_duration:
48504785 2754 ms_info['segment_duration'] = float(segment_duration)
b4c1d6e8
S
2755
2756 def extract_Initialization(source):
2757 initialization = source.find(_add_ns('Initialization'))
2758 if initialization is not None:
2759 ms_info['initialization_url'] = initialization.attrib['sourceURL']
2760
f14be228 2761 segment_list = element.find(_add_ns('SegmentList'))
1bac3455 2762 if segment_list is not None:
b4c1d6e8
S
2763 extract_common(segment_list)
2764 extract_Initialization(segment_list)
f14be228 2765 segment_urls_e = segment_list.findall(_add_ns('SegmentURL'))
1bac3455 2766 if segment_urls_e:
2767 ms_info['segment_urls'] = [segment.attrib['media'] for segment in segment_urls_e]
1bac3455 2768 else:
f14be228 2769 segment_template = element.find(_add_ns('SegmentTemplate'))
1bac3455 2770 if segment_template is not None:
b4c1d6e8 2771 extract_common(segment_template)
e228616c
S
2772 media = segment_template.get('media')
2773 if media:
2774 ms_info['media'] = media
1bac3455 2775 initialization = segment_template.get('initialization')
2776 if initialization:
e228616c 2777 ms_info['initialization'] = initialization
1bac3455 2778 else:
b4c1d6e8 2779 extract_Initialization(segment_template)
1bac3455 2780 return ms_info
b323e170 2781
1bac3455 2782 mpd_duration = parse_duration(mpd_doc.get('mediaPresentationDuration'))
6251555f 2783 formats, subtitles = [], {}
234416e4 2784 stream_numbers = collections.defaultdict(int)
f14be228 2785 for period in mpd_doc.findall(_add_ns('Period')):
1bac3455 2786 period_duration = parse_duration(period.get('duration')) or mpd_duration
2787 period_ms_info = extract_multisegment_info(period, {
2788 'start_number': 1,
2789 'timescale': 1,
2790 })
f14be228 2791 for adaptation_set in period.findall(_add_ns('AdaptationSet')):
1bac3455 2792 adaption_set_ms_info = extract_multisegment_info(adaptation_set, period_ms_info)
f14be228 2793 for representation in adaptation_set.findall(_add_ns('Representation')):
1bac3455 2794 representation_attrib = adaptation_set.attrib.copy()
2795 representation_attrib.update(representation.attrib)
f0948348 2796 # According to [1, 5.3.7.2, Table 9, page 41], @mimeType is mandatory
a6c8b759 2797 mime_type = representation_attrib['mimeType']
171e59ed
F
2798 content_type = representation_attrib.get('contentType', mime_type.split('/')[0])
2799
21633673 2800 codec_str = representation_attrib.get('codecs', '')
2801 # Some kind of binary subtitle found in some youtube livestreams
2802 if mime_type == 'application/x-rawcc':
2803 codecs = {'scodec': codec_str}
2804 else:
2805 codecs = parse_codecs(codec_str)
be2fc5b2 2806 if content_type not in ('video', 'audio', 'text'):
2807 if mime_type == 'image/jpeg':
a8731fcc 2808 content_type = mime_type
21633673 2809 elif codecs.get('vcodec', 'none') != 'none':
4afa3ec4 2810 content_type = 'video'
21633673 2811 elif codecs.get('acodec', 'none') != 'none':
4afa3ec4 2812 content_type = 'audio'
3fe75fdc 2813 elif codecs.get('scodec', 'none') != 'none':
be2fc5b2 2814 content_type = 'text'
6993f78d 2815 elif mimetype2ext(mime_type) in ('tt', 'dfxp', 'ttml', 'xml', 'json'):
2816 content_type = 'text'
cdb19aa4 2817 else:
be2fc5b2 2818 self.report_warning('Unknown MIME type %s in DASH manifest' % mime_type)
2819 continue
2820
2821 base_url = ''
2822 for element in (representation, adaptation_set, period, mpd_doc):
2823 base_url_e = element.find(_add_ns('BaseURL'))
47046464 2824 if try_call(lambda: base_url_e.text) is not None:
be2fc5b2 2825 base_url = base_url_e.text + base_url
2826 if re.match(r'^https?://', base_url):
2827 break
f9cc0161 2828 if mpd_base_url and base_url.startswith('/'):
14f25df2 2829 base_url = urllib.parse.urljoin(mpd_base_url, base_url)
f9cc0161
D
2830 elif mpd_base_url and not re.match(r'^https?://', base_url):
2831 if not mpd_base_url.endswith('/'):
be2fc5b2 2832 mpd_base_url += '/'
2833 base_url = mpd_base_url + base_url
2834 representation_id = representation_attrib.get('id')
2835 lang = representation_attrib.get('lang')
2836 url_el = representation.find(_add_ns('BaseURL'))
2837 filesize = int_or_none(url_el.attrib.get('{http://youtube.com/yt/2012/10/10}contentLength') if url_el is not None else None)
2838 bandwidth = int_or_none(representation_attrib.get('bandwidth'))
2839 if representation_id is not None:
2840 format_id = representation_id
2841 else:
2842 format_id = content_type
2843 if mpd_id:
2844 format_id = mpd_id + '-' + format_id
2845 if content_type in ('video', 'audio'):
2846 f = {
2847 'format_id': format_id,
2848 'manifest_url': mpd_url,
2849 'ext': mimetype2ext(mime_type),
2850 'width': int_or_none(representation_attrib.get('width')),
2851 'height': int_or_none(representation_attrib.get('height')),
2852 'tbr': float_or_none(bandwidth, 1000),
2853 'asr': int_or_none(representation_attrib.get('audioSamplingRate')),
2854 'fps': int_or_none(representation_attrib.get('frameRate')),
2855 'language': lang if lang not in ('mul', 'und', 'zxx', 'mis') else None,
2856 'format_note': 'DASH %s' % content_type,
2857 'filesize': filesize,
2858 'container': mimetype2ext(mime_type) + '_dash',
4afa3ec4 2859 **codecs
be2fc5b2 2860 }
be2fc5b2 2861 elif content_type == 'text':
2862 f = {
2863 'ext': mimetype2ext(mime_type),
2864 'manifest_url': mpd_url,
2865 'filesize': filesize,
2866 }
2867 elif content_type == 'image/jpeg':
2868 # See test case in VikiIE
2869 # https://www.viki.com/videos/1175236v-choosing-spouse-by-lottery-episode-1
2870 f = {
2871 'format_id': format_id,
2872 'ext': 'mhtml',
2873 'manifest_url': mpd_url,
2874 'format_note': 'DASH storyboards (jpeg)',
2875 'acodec': 'none',
2876 'vcodec': 'none',
2877 }
88acdbc2 2878 if is_drm_protected(adaptation_set) or is_drm_protected(representation):
2879 f['has_drm'] = True
be2fc5b2 2880 representation_ms_info = extract_multisegment_info(representation, adaption_set_ms_info)
2881
2882 def prepare_template(template_name, identifiers):
2883 tmpl = representation_ms_info[template_name]
2884 # First of, % characters outside $...$ templates
2885 # must be escaped by doubling for proper processing
2886 # by % operator string formatting used further (see
2887 # https://github.com/ytdl-org/youtube-dl/issues/16867).
2888 t = ''
2889 in_template = False
2890 for c in tmpl:
2891 t += c
2892 if c == '$':
2893 in_template = not in_template
2894 elif c == '%' and not in_template:
eca1f0d1 2895 t += c
be2fc5b2 2896 # Next, $...$ templates are translated to their
2897 # %(...) counterparts to be used with % operator
2898 if representation_id is not None:
2899 t = t.replace('$RepresentationID$', representation_id)
2900 t = re.sub(r'\$(%s)\$' % '|'.join(identifiers), r'%(\1)d', t)
2901 t = re.sub(r'\$(%s)%%([^$]+)\$' % '|'.join(identifiers), r'%(\1)\2', t)
2902 t.replace('$$', '$')
2903 return t
2904
2905 # @initialization is a regular template like @media one
2906 # so it should be handled just the same way (see
2907 # https://github.com/ytdl-org/youtube-dl/issues/11605)
2908 if 'initialization' in representation_ms_info:
2909 initialization_template = prepare_template(
2910 'initialization',
2911 # As per [1, 5.3.9.4.2, Table 15, page 54] $Number$ and
2912 # $Time$ shall not be included for @initialization thus
2913 # only $Bandwidth$ remains
2914 ('Bandwidth', ))
2915 representation_ms_info['initialization_url'] = initialization_template % {
2916 'Bandwidth': bandwidth,
2917 }
2918
2919 def location_key(location):
2920 return 'url' if re.match(r'^https?://', location) else 'path'
2921
2922 if 'segment_urls' not in representation_ms_info and 'media' in representation_ms_info:
2923
2924 media_template = prepare_template('media', ('Number', 'Bandwidth', 'Time'))
2925 media_location_key = location_key(media_template)
2926
2927 # As per [1, 5.3.9.4.4, Table 16, page 55] $Number$ and $Time$
2928 # can't be used at the same time
2929 if '%(Number' in media_template and 's' not in representation_ms_info:
2930 segment_duration = None
2931 if 'total_number' not in representation_ms_info and 'segment_duration' in representation_ms_info:
2932 segment_duration = float_or_none(representation_ms_info['segment_duration'], representation_ms_info['timescale'])
ffa89477 2933 representation_ms_info['total_number'] = int(math.ceil(
2934 float_or_none(period_duration, segment_duration, default=0)))
be2fc5b2 2935 representation_ms_info['fragments'] = [{
2936 media_location_key: media_template % {
2937 'Number': segment_number,
2938 'Bandwidth': bandwidth,
2939 },
2940 'duration': segment_duration,
2941 } for segment_number in range(
2942 representation_ms_info['start_number'],
2943 representation_ms_info['total_number'] + representation_ms_info['start_number'])]
2944 else:
2945 # $Number*$ or $Time$ in media template with S list available
2946 # Example $Number*$: http://www.svtplay.se/klipp/9023742/stopptid-om-bjorn-borg
2947 # Example $Time$: https://play.arkena.com/embed/avp/v2/player/media/b41dda37-d8e7-4d3f-b1b5-9a9db578bdfe/1/129411
2948 representation_ms_info['fragments'] = []
2949 segment_time = 0
2950 segment_d = None
2951 segment_number = representation_ms_info['start_number']
2952
2953 def add_segment_url():
2954 segment_url = media_template % {
2955 'Time': segment_time,
2956 'Bandwidth': bandwidth,
2957 'Number': segment_number,
2958 }
2959 representation_ms_info['fragments'].append({
2960 media_location_key: segment_url,
2961 'duration': float_or_none(segment_d, representation_ms_info['timescale']),
2962 })
2963
2964 for num, s in enumerate(representation_ms_info['s']):
2965 segment_time = s.get('t') or segment_time
2966 segment_d = s['d']
2967 add_segment_url()
2968 segment_number += 1
2969 for r in range(s.get('r', 0)):
2970 segment_time += segment_d
f0948348 2971 add_segment_url()
b4c1d6e8 2972 segment_number += 1
be2fc5b2 2973 segment_time += segment_d
2974 elif 'segment_urls' in representation_ms_info and 's' in representation_ms_info:
2975 # No media template
2976 # Example: https://www.youtube.com/watch?v=iXZV5uAYMJI
2977 # or any YouTube dashsegments video
2978 fragments = []
2979 segment_index = 0
2980 timescale = representation_ms_info['timescale']
2981 for s in representation_ms_info['s']:
2982 duration = float_or_none(s['d'], timescale)
2983 for r in range(s.get('r', 0) + 1):
2984 segment_uri = representation_ms_info['segment_urls'][segment_index]
2985 fragments.append({
2986 location_key(segment_uri): segment_uri,
2987 'duration': duration,
2988 })
2989 segment_index += 1
2990 representation_ms_info['fragments'] = fragments
2991 elif 'segment_urls' in representation_ms_info:
2992 # Segment URLs with no SegmentTimeline
2993 # Example: https://www.seznam.cz/zpravy/clanek/cesko-zasahne-vitr-o-sile-vichrice-muze-byt-i-zivotu-nebezpecny-39091
2994 # https://github.com/ytdl-org/youtube-dl/pull/14844
2995 fragments = []
2996 segment_duration = float_or_none(
2997 representation_ms_info['segment_duration'],
2998 representation_ms_info['timescale']) if 'segment_duration' in representation_ms_info else None
2999 for segment_url in representation_ms_info['segment_urls']:
3000 fragment = {
3001 location_key(segment_url): segment_url,
3002 }
3003 if segment_duration:
3004 fragment['duration'] = segment_duration
3005 fragments.append(fragment)
3006 representation_ms_info['fragments'] = fragments
3007 # If there is a fragments key available then we correctly recognized fragmented media.
3008 # Otherwise we will assume unfragmented media with direct access. Technically, such
3009 # assumption is not necessarily correct since we may simply have no support for
3010 # some forms of fragmented media renditions yet, but for now we'll use this fallback.
3011 if 'fragments' in representation_ms_info:
3012 f.update({
3013 # NB: mpd_url may be empty when MPD manifest is parsed from a string
3014 'url': mpd_url or base_url,
3015 'fragment_base_url': base_url,
3016 'fragments': [],
3017 'protocol': 'http_dash_segments' if mime_type != 'image/jpeg' else 'mhtml',
3018 })
3019 if 'initialization_url' in representation_ms_info:
3020 initialization_url = representation_ms_info['initialization_url']
3021 if not f.get('url'):
3022 f['url'] = initialization_url
3023 f['fragments'].append({location_key(initialization_url): initialization_url})
3024 f['fragments'].extend(representation_ms_info['fragments'])
ffa89477 3025 if not period_duration:
3026 period_duration = try_get(
3027 representation_ms_info,
3028 lambda r: sum(frag['duration'] for frag in r['fragments']), float)
17b598d3 3029 else:
be2fc5b2 3030 # Assuming direct URL to unfragmented media.
3031 f['url'] = base_url
234416e4 3032 if content_type in ('video', 'audio', 'image/jpeg'):
3033 f['manifest_stream_number'] = stream_numbers[f['url']]
3034 stream_numbers[f['url']] += 1
be2fc5b2 3035 formats.append(f)
3036 elif content_type == 'text':
3037 subtitles.setdefault(lang or 'und', []).append(f)
3038
171e59ed 3039 return formats, subtitles
17b598d3 3040
fd76a142
F
3041 def _extract_ism_formats(self, *args, **kwargs):
3042 fmts, subs = self._extract_ism_formats_and_subtitles(*args, **kwargs)
3043 if subs:
b5ae35ee 3044 self._report_ignoring_subs('ISM')
fd76a142
F
3045 return fmts
3046
3047 def _extract_ism_formats_and_subtitles(self, ism_url, video_id, ism_id=None, note=None, errnote=None, fatal=True, data=None, headers={}, query={}):
47a5cb77 3048 res = self._download_xml_handle(
b2758123 3049 ism_url, video_id,
37a3bb66 3050 note='Downloading ISM manifest' if note is None else note,
3051 errnote='Failed to download ISM manifest' if errnote is None else errnote,
7360c06f 3052 fatal=fatal, data=data, headers=headers, query=query)
b2758123 3053 if res is False:
fd76a142 3054 return [], {}
47a5cb77 3055 ism_doc, urlh = res
13b08034 3056 if ism_doc is None:
fd76a142 3057 return [], {}
b2758123 3058
fd76a142 3059 return self._parse_ism_formats_and_subtitles(ism_doc, urlh.geturl(), ism_id)
b2758123 3060
fd76a142 3061 def _parse_ism_formats_and_subtitles(self, ism_doc, ism_url, ism_id=None):
76d5a363
S
3062 """
3063 Parse formats from ISM manifest.
3064 References:
3065 1. [MS-SSTR]: Smooth Streaming Protocol,
3066 https://msdn.microsoft.com/en-us/library/ff469518.aspx
3067 """
06869367 3068 if ism_doc.get('IsLive') == 'TRUE':
fd76a142 3069 return [], {}
b2758123 3070
b2758123
RA
3071 duration = int(ism_doc.attrib['Duration'])
3072 timescale = int_or_none(ism_doc.get('TimeScale')) or 10000000
3073
3074 formats = []
fd76a142 3075 subtitles = {}
b2758123
RA
3076 for stream in ism_doc.findall('StreamIndex'):
3077 stream_type = stream.get('Type')
fd76a142 3078 if stream_type not in ('video', 'audio', 'text'):
b2758123
RA
3079 continue
3080 url_pattern = stream.attrib['Url']
3081 stream_timescale = int_or_none(stream.get('TimeScale')) or timescale
3082 stream_name = stream.get('Name')
fd76a142 3083 stream_language = stream.get('Language', 'und')
b2758123 3084 for track in stream.findall('QualityLevel'):
e2efe599 3085 fourcc = track.get('FourCC') or ('AACL' if track.get('AudioTag') == '255' else None)
b2758123 3086 # TODO: add support for WVC1 and WMAP
66a1b864 3087 if fourcc not in ('H264', 'AVC1', 'AACL', 'TTML'):
b2758123
RA
3088 self.report_warning('%s is not a supported codec' % fourcc)
3089 continue
3090 tbr = int(track.attrib['Bitrate']) // 1000
76d5a363
S
3091 # [1] does not mention Width and Height attributes. However,
3092 # they're often present while MaxWidth and MaxHeight are
3093 # missing, so should be used as fallbacks
3094 width = int_or_none(track.get('MaxWidth') or track.get('Width'))
3095 height = int_or_none(track.get('MaxHeight') or track.get('Height'))
b2758123
RA
3096 sampling_rate = int_or_none(track.get('SamplingRate'))
3097
3098 track_url_pattern = re.sub(r'{[Bb]itrate}', track.attrib['Bitrate'], url_pattern)
14f25df2 3099 track_url_pattern = urllib.parse.urljoin(ism_url, track_url_pattern)
b2758123
RA
3100
3101 fragments = []
3102 fragment_ctx = {
3103 'time': 0,
3104 }
3105 stream_fragments = stream.findall('c')
3106 for stream_fragment_index, stream_fragment in enumerate(stream_fragments):
3107 fragment_ctx['time'] = int_or_none(stream_fragment.get('t')) or fragment_ctx['time']
3108 fragment_repeat = int_or_none(stream_fragment.get('r')) or 1
3109 fragment_ctx['duration'] = int_or_none(stream_fragment.get('d'))
3110 if not fragment_ctx['duration']:
3111 try:
3112 next_fragment_time = int(stream_fragment[stream_fragment_index + 1].attrib['t'])
3113 except IndexError:
3114 next_fragment_time = duration
1616f9b4 3115 fragment_ctx['duration'] = (next_fragment_time - fragment_ctx['time']) / fragment_repeat
b2758123
RA
3116 for _ in range(fragment_repeat):
3117 fragments.append({
14f25df2 3118 'url': re.sub(r'{start[ _]time}', str(fragment_ctx['time']), track_url_pattern),
b2758123
RA
3119 'duration': fragment_ctx['duration'] / stream_timescale,
3120 })
3121 fragment_ctx['time'] += fragment_ctx['duration']
3122
fd76a142
F
3123 if stream_type == 'text':
3124 subtitles.setdefault(stream_language, []).append({
3125 'ext': 'ismt',
3126 'protocol': 'ism',
3127 'url': ism_url,
3128 'manifest_url': ism_url,
3129 'fragments': fragments,
3130 '_download_params': {
3131 'stream_type': stream_type,
3132 'duration': duration,
3133 'timescale': stream_timescale,
3134 'fourcc': fourcc,
3135 'language': stream_language,
3136 'codec_private_data': track.get('CodecPrivateData'),
3137 }
3138 })
3139 elif stream_type in ('video', 'audio'):
3140 formats.append({
34921b43 3141 'format_id': join_nonempty(ism_id, stream_name, tbr),
fd76a142
F
3142 'url': ism_url,
3143 'manifest_url': ism_url,
3144 'ext': 'ismv' if stream_type == 'video' else 'isma',
3145 'width': width,
3146 'height': height,
3147 'tbr': tbr,
3148 'asr': sampling_rate,
3149 'vcodec': 'none' if stream_type == 'audio' else fourcc,
3150 'acodec': 'none' if stream_type == 'video' else fourcc,
3151 'protocol': 'ism',
3152 'fragments': fragments,
88acdbc2 3153 'has_drm': ism_doc.find('Protection') is not None,
fd76a142
F
3154 '_download_params': {
3155 'stream_type': stream_type,
3156 'duration': duration,
3157 'timescale': stream_timescale,
3158 'width': width or 0,
3159 'height': height or 0,
3160 'fourcc': fourcc,
3161 'language': stream_language,
3162 'codec_private_data': track.get('CodecPrivateData'),
3163 'sampling_rate': sampling_rate,
3164 'channels': int_or_none(track.get('Channels', 2)),
3165 'bits_per_sample': int_or_none(track.get('BitsPerSample', 16)),
3166 'nal_unit_length_field': int_or_none(track.get('NALUnitLengthField', 4)),
3167 },
3168 })
3169 return formats, subtitles
b2758123 3170
079a7cfc 3171 def _parse_html5_media_entries(self, base_url, webpage, video_id, m3u8_id=None, m3u8_entry_protocol='m3u8_native', mpd_id=None, preference=None, quality=None):
6780154e
S
3172 def absolute_url(item_url):
3173 return urljoin(base_url, item_url)
59bbe491 3174
3175 def parse_content_type(content_type):
3176 if not content_type:
3177 return {}
3178 ctr = re.search(r'(?P<mimetype>[^/]+/[^;]+)(?:;\s*codecs="?(?P<codecs>[^"]+))?', content_type)
3179 if ctr:
3180 mimetype, codecs = ctr.groups()
3181 f = parse_codecs(codecs)
3182 f['ext'] = mimetype2ext(mimetype)
3183 return f
3184 return {}
3185
222a2308
L
3186 def _media_formats(src, cur_media_type, type_info=None):
3187 type_info = type_info or {}
520251c0 3188 full_url = absolute_url(src)
82889d4a 3189 ext = type_info.get('ext') or determine_ext(full_url)
87a449c1 3190 if ext == 'm3u8':
520251c0
YCH
3191 is_plain_url = False
3192 formats = self._extract_m3u8_formats(
ad120ae1 3193 full_url, video_id, ext='mp4',
eeb0a956 3194 entry_protocol=m3u8_entry_protocol, m3u8_id=m3u8_id,
f983b875 3195 preference=preference, quality=quality, fatal=False)
87a449c1
S
3196 elif ext == 'mpd':
3197 is_plain_url = False
3198 formats = self._extract_mpd_formats(
b359e977 3199 full_url, video_id, mpd_id=mpd_id, fatal=False)
520251c0
YCH
3200 else:
3201 is_plain_url = True
3202 formats = [{
3203 'url': full_url,
3204 'vcodec': 'none' if cur_media_type == 'audio' else None,
222a2308 3205 'ext': ext,
520251c0
YCH
3206 }]
3207 return is_plain_url, formats
3208
59bbe491 3209 entries = []
4328ddf8 3210 # amp-video and amp-audio are very similar to their HTML5 counterparts
962ffcf8 3211 # so we will include them right here (see
4328ddf8 3212 # https://www.ampproject.org/docs/reference/components/amp-video)
29f7c58a 3213 # For dl8-* tags see https://delight-vr.com/documentation/dl8-video/
3214 _MEDIA_TAG_NAME_RE = r'(?:(?:amp|dl8(?:-live)?)-)?(video|audio)'
3215 media_tags = [(media_tag, media_tag_name, media_type, '')
3216 for media_tag, media_tag_name, media_type
3217 in re.findall(r'(?s)(<(%s)[^>]*/>)' % _MEDIA_TAG_NAME_RE, webpage)]
2aec7256
S
3218 media_tags.extend(re.findall(
3219 # We only allow video|audio followed by a whitespace or '>'.
3220 # Allowing more characters may end up in significant slow down (see
067aa17e 3221 # https://github.com/ytdl-org/youtube-dl/issues/11979, example URL:
2aec7256 3222 # http://www.porntrex.com/maps/videositemap.xml).
29f7c58a 3223 r'(?s)(<(?P<tag>%s)(?:\s+[^>]*)?>)(.*?)</(?P=tag)>' % _MEDIA_TAG_NAME_RE, webpage))
3224 for media_tag, _, media_type, media_content in media_tags:
59bbe491 3225 media_info = {
3226 'formats': [],
3227 'subtitles': {},
3228 }
3229 media_attributes = extract_attributes(media_tag)
f856816b 3230 src = strip_or_none(media_attributes.get('src'))
59bbe491 3231 if src:
222a2308
L
3232 f = parse_content_type(media_attributes.get('type'))
3233 _, formats = _media_formats(src, media_type, f)
520251c0 3234 media_info['formats'].extend(formats)
6780154e 3235 media_info['thumbnail'] = absolute_url(media_attributes.get('poster'))
59bbe491 3236 if media_content:
3237 for source_tag in re.findall(r'<source[^>]+>', media_content):
d493f15c
S
3238 s_attr = extract_attributes(source_tag)
3239 # data-video-src and data-src are non standard but seen
3240 # several times in the wild
f856816b 3241 src = strip_or_none(dict_get(s_attr, ('src', 'data-video-src', 'data-src')))
59bbe491 3242 if not src:
3243 continue
d493f15c 3244 f = parse_content_type(s_attr.get('type'))
868f79db 3245 is_plain_url, formats = _media_formats(src, media_type, f)
520251c0 3246 if is_plain_url:
d493f15c
S
3247 # width, height, res, label and title attributes are
3248 # all not standard but seen several times in the wild
3249 labels = [
3250 s_attr.get(lbl)
3251 for lbl in ('label', 'title')
3252 if str_or_none(s_attr.get(lbl))
3253 ]
3254 width = int_or_none(s_attr.get('width'))
3089bc74
S
3255 height = (int_or_none(s_attr.get('height'))
3256 or int_or_none(s_attr.get('res')))
d493f15c
S
3257 if not width or not height:
3258 for lbl in labels:
3259 resolution = parse_resolution(lbl)
3260 if not resolution:
3261 continue
3262 width = width or resolution.get('width')
3263 height = height or resolution.get('height')
3264 for lbl in labels:
3265 tbr = parse_bitrate(lbl)
3266 if tbr:
3267 break
3268 else:
3269 tbr = None
1ed45499 3270 f.update({
d493f15c
S
3271 'width': width,
3272 'height': height,
3273 'tbr': tbr,
3274 'format_id': s_attr.get('label') or s_attr.get('title'),
1ed45499 3275 })
520251c0
YCH
3276 f.update(formats[0])
3277 media_info['formats'].append(f)
3278 else:
3279 media_info['formats'].extend(formats)
59bbe491 3280 for track_tag in re.findall(r'<track[^>]+>', media_content):
3281 track_attributes = extract_attributes(track_tag)
3282 kind = track_attributes.get('kind')
5968d7d2 3283 if not kind or kind in ('subtitles', 'captions'):
f856816b 3284 src = strip_or_none(track_attributes.get('src'))
59bbe491 3285 if not src:
3286 continue
3287 lang = track_attributes.get('srclang') or track_attributes.get('lang') or track_attributes.get('label')
3288 media_info['subtitles'].setdefault(lang, []).append({
3289 'url': absolute_url(src),
3290 })
5e8e2fa5
S
3291 for f in media_info['formats']:
3292 f.setdefault('http_headers', {})['Referer'] = base_url
5968d7d2 3293 if media_info['formats'] or media_info['subtitles']:
59bbe491 3294 entries.append(media_info)
3295 return entries
3296
f6a1d69a
F
3297 def _extract_akamai_formats(self, *args, **kwargs):
3298 fmts, subs = self._extract_akamai_formats_and_subtitles(*args, **kwargs)
3299 if subs:
b5ae35ee 3300 self._report_ignoring_subs('akamai')
f6a1d69a
F
3301 return fmts
3302
3303 def _extract_akamai_formats_and_subtitles(self, manifest_url, video_id, hosts={}):
29f7c58a 3304 signed = 'hdnea=' in manifest_url
3305 if not signed:
3306 # https://learn.akamai.com/en-us/webhelp/media-services-on-demand/stream-packaging-user-guide/GUID-BE6C0F73-1E06-483B-B0EA-57984B91B7F9.html
3307 manifest_url = re.sub(
3308 r'(?:b=[\d,-]+|(?:__a__|attributes)=off|__b__=\d+)&?',
3309 '', manifest_url).strip('?')
3310
c7c43a93 3311 formats = []
f6a1d69a 3312 subtitles = {}
70c5802b 3313
e71a4509 3314 hdcore_sign = 'hdcore=3.7.0'
ff6f9a67 3315 f4m_url = re.sub(r'(https?://[^/]+)/i/', r'\1/z/', manifest_url).replace('/master.m3u8', '/manifest.f4m')
c4251b9a
RA
3316 hds_host = hosts.get('hds')
3317 if hds_host:
3318 f4m_url = re.sub(r'(https?://)[^/]+', r'\1' + hds_host, f4m_url)
e71a4509
RA
3319 if 'hdcore=' not in f4m_url:
3320 f4m_url += ('&' if '?' in f4m_url else '?') + hdcore_sign
3321 f4m_formats = self._extract_f4m_formats(
3322 f4m_url, video_id, f4m_id='hds', fatal=False)
3323 for entry in f4m_formats:
3324 entry.update({'extra_param_to_segment_url': hdcore_sign})
3325 formats.extend(f4m_formats)
70c5802b 3326
c4251b9a
RA
3327 m3u8_url = re.sub(r'(https?://[^/]+)/z/', r'\1/i/', manifest_url).replace('/manifest.f4m', '/master.m3u8')
3328 hls_host = hosts.get('hls')
3329 if hls_host:
3330 m3u8_url = re.sub(r'(https?://)[^/]+', r'\1' + hls_host, m3u8_url)
f6a1d69a 3331 m3u8_formats, m3u8_subtitles = self._extract_m3u8_formats_and_subtitles(
c7c43a93 3332 m3u8_url, video_id, 'mp4', 'm3u8_native',
29f7c58a 3333 m3u8_id='hls', fatal=False)
3334 formats.extend(m3u8_formats)
f6a1d69a 3335 subtitles = self._merge_subtitles(subtitles, m3u8_subtitles)
70c5802b 3336
3337 http_host = hosts.get('http')
29f7c58a 3338 if http_host and m3u8_formats and not signed:
3339 REPL_REGEX = r'https?://[^/]+/i/([^,]+),([^/]+),([^/]+)\.csmil/.+'
70c5802b 3340 qualities = re.match(REPL_REGEX, m3u8_url).group(2).split(',')
3341 qualities_length = len(qualities)
29f7c58a 3342 if len(m3u8_formats) in (qualities_length, qualities_length + 1):
70c5802b 3343 i = 0
29f7c58a 3344 for f in m3u8_formats:
3345 if f['vcodec'] != 'none':
70c5802b 3346 for protocol in ('http', 'https'):
3347 http_f = f.copy()
3348 del http_f['manifest_url']
3349 http_url = re.sub(
86e5f3ed 3350 REPL_REGEX, protocol + fr'://{http_host}/\g<1>{qualities[i]}\3', f['url'])
70c5802b 3351 http_f.update({
3352 'format_id': http_f['format_id'].replace('hls-', protocol + '-'),
3353 'url': http_url,
3354 'protocol': protocol,
3355 })
29f7c58a 3356 formats.append(http_f)
70c5802b 3357 i += 1
70c5802b 3358
f6a1d69a 3359 return formats, subtitles
c7c43a93 3360
6ad02195 3361 def _extract_wowza_formats(self, url, video_id, m3u8_entry_protocol='m3u8_native', skip_protocols=[]):
14f25df2 3362 query = urllib.parse.urlparse(url).query
6ad02195 3363 url = re.sub(r'/(?:manifest|playlist|jwplayer)\.(?:m3u8|f4m|mpd|smil)', '', url)
240f2622
S
3364 mobj = re.search(
3365 r'(?:(?:http|rtmp|rtsp)(?P<s>s)?:)?(?P<url>//[^?]+)', url)
3366 url_base = mobj.group('url')
3367 http_base_url = '%s%s:%s' % ('http', mobj.group('s') or '', url_base)
6ad02195 3368 formats = []
044eeb14
S
3369
3370 def manifest_url(manifest):
86e5f3ed 3371 m_url = f'{http_base_url}/{manifest}'
044eeb14
S
3372 if query:
3373 m_url += '?%s' % query
3374 return m_url
3375
6ad02195
RA
3376 if 'm3u8' not in skip_protocols:
3377 formats.extend(self._extract_m3u8_formats(
044eeb14 3378 manifest_url('playlist.m3u8'), video_id, 'mp4',
6ad02195
RA
3379 m3u8_entry_protocol, m3u8_id='hls', fatal=False))
3380 if 'f4m' not in skip_protocols:
3381 formats.extend(self._extract_f4m_formats(
044eeb14 3382 manifest_url('manifest.f4m'),
6ad02195 3383 video_id, f4m_id='hds', fatal=False))
0384932e
RA
3384 if 'dash' not in skip_protocols:
3385 formats.extend(self._extract_mpd_formats(
044eeb14 3386 manifest_url('manifest.mpd'),
0384932e 3387 video_id, mpd_id='dash', fatal=False))
6ad02195 3388 if re.search(r'(?:/smil:|\.smil)', url_base):
6ad02195
RA
3389 if 'smil' not in skip_protocols:
3390 rtmp_formats = self._extract_smil_formats(
044eeb14 3391 manifest_url('jwplayer.smil'),
6ad02195
RA
3392 video_id, fatal=False)
3393 for rtmp_format in rtmp_formats:
3394 rtsp_format = rtmp_format.copy()
3395 rtsp_format['url'] = '%s/%s' % (rtmp_format['url'], rtmp_format['play_path'])
3396 del rtsp_format['play_path']
3397 del rtsp_format['ext']
3398 rtsp_format.update({
3399 'url': rtsp_format['url'].replace('rtmp://', 'rtsp://'),
3400 'format_id': rtmp_format['format_id'].replace('rtmp', 'rtsp'),
3401 'protocol': 'rtsp',
3402 })
3403 formats.extend([rtmp_format, rtsp_format])
3404 else:
3405 for protocol in ('rtmp', 'rtsp'):
3406 if protocol not in skip_protocols:
3407 formats.append({
86e5f3ed 3408 'url': f'{protocol}:{url_base}',
6ad02195
RA
3409 'format_id': protocol,
3410 'protocol': protocol,
3411 })
3412 return formats
3413
c73e330e 3414 def _find_jwplayer_data(self, webpage, video_id=None, transform_source=js_to_json):
a4a554a7 3415 mobj = re.search(
ac9c69ac 3416 r'(?s)jwplayer\((?P<quote>[\'"])[^\'" ]+(?P=quote)\)(?!</script>).*?\.setup\s*\((?P<options>[^)]+)\)',
a4a554a7
YCH
3417 webpage)
3418 if mobj:
c73e330e
RU
3419 try:
3420 jwplayer_data = self._parse_json(mobj.group('options'),
3421 video_id=video_id,
3422 transform_source=transform_source)
3423 except ExtractorError:
3424 pass
3425 else:
3426 if isinstance(jwplayer_data, dict):
3427 return jwplayer_data
a4a554a7
YCH
3428
3429 def _extract_jwplayer_data(self, webpage, video_id, *args, **kwargs):
c73e330e
RU
3430 jwplayer_data = self._find_jwplayer_data(
3431 webpage, video_id, transform_source=js_to_json)
a4a554a7
YCH
3432 return self._parse_jwplayer_data(
3433 jwplayer_data, video_id, *args, **kwargs)
3434
3435 def _parse_jwplayer_data(self, jwplayer_data, video_id=None, require_title=True,
3436 m3u8_id=None, mpd_id=None, rtmp_params=None, base_url=None):
3437 # JWPlayer backward compatibility: flattened playlists
3438 # https://github.com/jwplayer/jwplayer/blob/v7.4.3/src/js/api/config.js#L81-L96
3439 if 'playlist' not in jwplayer_data:
3440 jwplayer_data = {'playlist': [jwplayer_data]}
3441
3442 entries = []
3443
3444 # JWPlayer backward compatibility: single playlist item
3445 # https://github.com/jwplayer/jwplayer/blob/v7.7.0/src/js/playlist/playlist.js#L10
3446 if not isinstance(jwplayer_data['playlist'], list):
3447 jwplayer_data['playlist'] = [jwplayer_data['playlist']]
3448
3449 for video_data in jwplayer_data['playlist']:
3450 # JWPlayer backward compatibility: flattened sources
3451 # https://github.com/jwplayer/jwplayer/blob/v7.4.3/src/js/playlist/item.js#L29-L35
3452 if 'sources' not in video_data:
3453 video_data['sources'] = [video_data]
3454
3455 this_video_id = video_id or video_data['mediaid']
3456
1a2192cb
S
3457 formats = self._parse_jwplayer_formats(
3458 video_data['sources'], video_id=this_video_id, m3u8_id=m3u8_id,
3459 mpd_id=mpd_id, rtmp_params=rtmp_params, base_url=base_url)
a4a554a7
YCH
3460
3461 subtitles = {}
3462 tracks = video_data.get('tracks')
3463 if tracks and isinstance(tracks, list):
3464 for track in tracks:
96a2daa1
S
3465 if not isinstance(track, dict):
3466 continue
f4b74272 3467 track_kind = track.get('kind')
14f25df2 3468 if not track_kind or not isinstance(track_kind, str):
f4b74272
S
3469 continue
3470 if track_kind.lower() not in ('captions', 'subtitles'):
a4a554a7
YCH
3471 continue
3472 track_url = urljoin(base_url, track.get('file'))
3473 if not track_url:
3474 continue
3475 subtitles.setdefault(track.get('label') or 'en', []).append({
3476 'url': self._proto_relative_url(track_url)
3477 })
3478
50d808f5 3479 entry = {
a4a554a7 3480 'id': this_video_id,
50d808f5 3481 'title': unescapeHTML(video_data['title'] if require_title else video_data.get('title')),
f81dd65b 3482 'description': clean_html(video_data.get('description')),
6945b9e7 3483 'thumbnail': urljoin(base_url, self._proto_relative_url(video_data.get('image'))),
a4a554a7
YCH
3484 'timestamp': int_or_none(video_data.get('pubdate')),
3485 'duration': float_or_none(jwplayer_data.get('duration') or video_data.get('duration')),
3486 'subtitles': subtitles,
50d808f5
RA
3487 }
3488 # https://github.com/jwplayer/jwplayer/blob/master/src/js/utils/validator.js#L32
3489 if len(formats) == 1 and re.search(r'^(?:http|//).*(?:youtube\.com|youtu\.be)/.+', formats[0]['url']):
3490 entry.update({
3491 '_type': 'url_transparent',
3492 'url': formats[0]['url'],
3493 })
3494 else:
3495 self._sort_formats(formats)
3496 entry['formats'] = formats
3497 entries.append(entry)
a4a554a7
YCH
3498 if len(entries) == 1:
3499 return entries[0]
3500 else:
3501 return self.playlist_result(entries)
3502
ed0cf9b3
S
3503 def _parse_jwplayer_formats(self, jwplayer_sources_data, video_id=None,
3504 m3u8_id=None, mpd_id=None, rtmp_params=None, base_url=None):
bf1b87cd 3505 urls = []
ed0cf9b3 3506 formats = []
1a2192cb 3507 for source in jwplayer_sources_data:
0a268c6e
S
3508 if not isinstance(source, dict):
3509 continue
6945b9e7
RA
3510 source_url = urljoin(
3511 base_url, self._proto_relative_url(source.get('file')))
3512 if not source_url or source_url in urls:
bf1b87cd
RA
3513 continue
3514 urls.append(source_url)
ed0cf9b3
S
3515 source_type = source.get('type') or ''
3516 ext = mimetype2ext(source_type) or determine_ext(source_url)
3517 if source_type == 'hls' or ext == 'm3u8':
3518 formats.extend(self._extract_m3u8_formats(
0236cd0d
S
3519 source_url, video_id, 'mp4', entry_protocol='m3u8_native',
3520 m3u8_id=m3u8_id, fatal=False))
0d9c48de 3521 elif source_type == 'dash' or ext == 'mpd':
ed0cf9b3
S
3522 formats.extend(self._extract_mpd_formats(
3523 source_url, video_id, mpd_id=mpd_id, fatal=False))
b51dc9db
S
3524 elif ext == 'smil':
3525 formats.extend(self._extract_smil_formats(
3526 source_url, video_id, fatal=False))
ed0cf9b3 3527 # https://github.com/jwplayer/jwplayer/blob/master/src/js/providers/default.js#L67
0236cd0d
S
3528 elif source_type.startswith('audio') or ext in (
3529 'oga', 'aac', 'mp3', 'mpeg', 'vorbis'):
ed0cf9b3
S
3530 formats.append({
3531 'url': source_url,
3532 'vcodec': 'none',
3533 'ext': ext,
3534 })
3535 else:
3536 height = int_or_none(source.get('height'))
3537 if height is None:
3538 # Often no height is provided but there is a label in
0236cd0d 3539 # format like "1080p", "720p SD", or 1080.
ed0cf9b3 3540 height = int_or_none(self._search_regex(
14f25df2 3541 r'^(\d{3,4})[pP]?(?:\b|$)', str(source.get('label') or ''),
ed0cf9b3
S
3542 'height', default=None))
3543 a_format = {
3544 'url': source_url,
3545 'width': int_or_none(source.get('width')),
3546 'height': height,
0236cd0d 3547 'tbr': int_or_none(source.get('bitrate')),
ed0cf9b3
S
3548 'ext': ext,
3549 }
3550 if source_url.startswith('rtmp'):
3551 a_format['ext'] = 'flv'
ed0cf9b3
S
3552 # See com/longtailvideo/jwplayer/media/RTMPMediaProvider.as
3553 # of jwplayer.flash.swf
3554 rtmp_url_parts = re.split(
3555 r'((?:mp4|mp3|flv):)', source_url, 1)
3556 if len(rtmp_url_parts) == 3:
3557 rtmp_url, prefix, play_path = rtmp_url_parts
3558 a_format.update({
3559 'url': rtmp_url,
3560 'play_path': prefix + play_path,
3561 })
3562 if rtmp_params:
3563 a_format.update(rtmp_params)
3564 formats.append(a_format)
3565 return formats
3566
f4b1c7ad 3567 def _live_title(self, name):
39ca3b5c 3568 self._downloader.deprecation_warning('yt_dlp.InfoExtractor._live_title is deprecated and does not work as expected')
3569 return name
f4b1c7ad 3570
b14f3a4c
PH
3571 def _int(self, v, name, fatal=False, **kwargs):
3572 res = int_or_none(v, **kwargs)
b14f3a4c 3573 if res is None:
86e5f3ed 3574 msg = f'Failed to extract {name}: Could not parse value {v!r}'
b14f3a4c
PH
3575 if fatal:
3576 raise ExtractorError(msg)
3577 else:
6a39ee13 3578 self.report_warning(msg)
b14f3a4c
PH
3579 return res
3580
3581 def _float(self, v, name, fatal=False, **kwargs):
3582 res = float_or_none(v, **kwargs)
3583 if res is None:
86e5f3ed 3584 msg = f'Failed to extract {name}: Could not parse value {v!r}'
b14f3a4c
PH
3585 if fatal:
3586 raise ExtractorError(msg)
3587 else:
6a39ee13 3588 self.report_warning(msg)
b14f3a4c
PH
3589 return res
3590
40e41780
TF
3591 def _set_cookie(self, domain, name, value, expire_time=None, port=None,
3592 path='/', secure=False, discard=False, rest={}, **kwargs):
ac668111 3593 cookie = http.cookiejar.Cookie(
4ed2d7b7 3594 0, name, value, port, port is not None, domain, True,
40e41780
TF
3595 domain.startswith('.'), path, True, secure, expire_time,
3596 discard, None, None, rest)
9809740b 3597 self.cookiejar.set_cookie(cookie)
42939b61 3598
799207e8 3599 def _get_cookies(self, url):
ac668111 3600 """ Return a http.cookies.SimpleCookie with the cookies for the url """
3601 return http.cookies.SimpleCookie(self._downloader._calc_cookies(url))
799207e8 3602
e3c1266f 3603 def _apply_first_set_cookie_header(self, url_handle, cookie):
ce2fe4c0
S
3604 """
3605 Apply first Set-Cookie header instead of the last. Experimental.
3606
3607 Some sites (e.g. [1-3]) may serve two cookies under the same name
3608 in Set-Cookie header and expect the first (old) one to be set rather
3609 than second (new). However, as of RFC6265 the newer one cookie
3610 should be set into cookie store what actually happens.
3611 We will workaround this issue by resetting the cookie to
3612 the first one manually.
3613 1. https://new.vk.com/
3614 2. https://github.com/ytdl-org/youtube-dl/issues/9841#issuecomment-227871201
3615 3. https://learning.oreilly.com/
3616 """
e3c1266f
S
3617 for header, cookies in url_handle.headers.items():
3618 if header.lower() != 'set-cookie':
3619 continue
cfb0511d 3620 cookies = cookies.encode('iso-8859-1').decode('utf-8')
e3c1266f
S
3621 cookie_value = re.search(
3622 r'%s=(.+?);.*?\b[Dd]omain=(.+?)(?:[,;]|$)' % cookie, cookies)
3623 if cookie_value:
3624 value, domain = cookie_value.groups()
3625 self._set_cookie(domain, cookie, value)
3626 break
3627
82d02080 3628 @classmethod
3629 def get_testcases(cls, include_onlymatching=False):
3630 t = getattr(cls, '_TEST', None)
05900629 3631 if t:
82d02080 3632 assert not hasattr(cls, '_TESTS'), f'{cls.ie_key()}IE has _TEST and _TESTS'
05900629
PH
3633 tests = [t]
3634 else:
82d02080 3635 tests = getattr(cls, '_TESTS', [])
05900629
PH
3636 for t in tests:
3637 if not include_onlymatching and t.get('only_matching', False):
3638 continue
82d02080 3639 t['name'] = cls.ie_key()
05900629
PH
3640 yield t
3641
24146491 3642 @classproperty
3643 def age_limit(cls):
3644 """Get age limit from the testcases"""
3645 return max(traverse_obj(
3646 tuple(cls.get_testcases(include_onlymatching=False)),
3647 (..., (('playlist', 0), None), 'info_dict', 'age_limit')) or [0])
3648
82d02080 3649 @classmethod
3650 def is_suitable(cls, age_limit):
24146491 3651 """Test whether the extractor is generally suitable for the given age limit"""
3652 return not age_restricted(cls.age_limit, age_limit)
05900629 3653
82d02080 3654 @classmethod
3655 def description(cls, *, markdown=True, search_examples=None):
8dcce6a8 3656 """Description of the extractor"""
3657 desc = ''
82d02080 3658 if cls._NETRC_MACHINE:
8dcce6a8 3659 if markdown:
82d02080 3660 desc += f' [<abbr title="netrc machine"><em>{cls._NETRC_MACHINE}</em></abbr>]'
8dcce6a8 3661 else:
82d02080 3662 desc += f' [{cls._NETRC_MACHINE}]'
3663 if cls.IE_DESC is False:
8dcce6a8 3664 desc += ' [HIDDEN]'
82d02080 3665 elif cls.IE_DESC:
3666 desc += f' {cls.IE_DESC}'
3667 if cls.SEARCH_KEY:
3668 desc += f'; "{cls.SEARCH_KEY}:" prefix'
8dcce6a8 3669 if search_examples:
3670 _COUNTS = ('', '5', '10', 'all')
82d02080 3671 desc += f' (Example: "{cls.SEARCH_KEY}{random.choice(_COUNTS)}:{random.choice(search_examples)}")'
3672 if not cls.working():
8dcce6a8 3673 desc += ' (**Currently broken**)' if markdown else ' (Currently broken)'
3674
82d02080 3675 name = f' - **{cls.IE_NAME}**' if markdown else cls.IE_NAME
8dcce6a8 3676 return f'{name}:{desc}' if desc else name
3677
a504ced0 3678 def extract_subtitles(self, *args, **kwargs):
a06916d9 3679 if (self.get_param('writesubtitles', False)
3680 or self.get_param('listsubtitles')):
9868ea49
JMF
3681 return self._get_subtitles(*args, **kwargs)
3682 return {}
a504ced0
JMF
3683
3684 def _get_subtitles(self, *args, **kwargs):
611c1dd9 3685 raise NotImplementedError('This method must be implemented by subclasses')
a504ced0 3686
a2160aa4 3687 def extract_comments(self, *args, **kwargs):
3688 if not self.get_param('getcomments'):
3689 return None
3690 generator = self._get_comments(*args, **kwargs)
3691
3692 def extractor():
3693 comments = []
d2b2fca5 3694 interrupted = True
a2160aa4 3695 try:
3696 while True:
3697 comments.append(next(generator))
a2160aa4 3698 except StopIteration:
3699 interrupted = False
d2b2fca5 3700 except KeyboardInterrupt:
3701 self.to_screen('Interrupted by user')
3702 except Exception as e:
3703 if self.get_param('ignoreerrors') is not True:
3704 raise
3705 self._downloader.report_error(e)
a2160aa4 3706 comment_count = len(comments)
3707 self.to_screen(f'Extracted {comment_count} comments')
3708 return {
3709 'comments': comments,
3710 'comment_count': None if interrupted else comment_count
3711 }
3712 return extractor
3713
3714 def _get_comments(self, *args, **kwargs):
3715 raise NotImplementedError('This method must be implemented by subclasses')
3716
912e0b7e
YCH
3717 @staticmethod
3718 def _merge_subtitle_items(subtitle_list1, subtitle_list2):
a825ffbf 3719 """ Merge subtitle items for one language. Items with duplicated URLs/data
912e0b7e 3720 will be dropped. """
86e5f3ed 3721 list1_data = {(item.get('url'), item.get('data')) for item in subtitle_list1}
912e0b7e 3722 ret = list(subtitle_list1)
a44ca5a4 3723 ret.extend(item for item in subtitle_list2 if (item.get('url'), item.get('data')) not in list1_data)
912e0b7e
YCH
3724 return ret
3725
3726 @classmethod
46890374 3727 def _merge_subtitles(cls, *dicts, target=None):
19bb3920 3728 """ Merge subtitle dictionaries, language by language. """
19bb3920
F
3729 if target is None:
3730 target = {}
3731 for d in dicts:
3732 for lang, subs in d.items():
3733 target[lang] = cls._merge_subtitle_items(target.get(lang, []), subs)
3734 return target
912e0b7e 3735
360e1ca5 3736 def extract_automatic_captions(self, *args, **kwargs):
a06916d9 3737 if (self.get_param('writeautomaticsub', False)
3738 or self.get_param('listsubtitles')):
9868ea49
JMF
3739 return self._get_automatic_captions(*args, **kwargs)
3740 return {}
360e1ca5
JMF
3741
3742 def _get_automatic_captions(self, *args, **kwargs):
611c1dd9 3743 raise NotImplementedError('This method must be implemented by subclasses')
360e1ca5 3744
2762dbb1 3745 @functools.cached_property
24146491 3746 def _cookies_passed(self):
3747 """Whether cookies have been passed to YoutubeDL"""
3748 return self.get_param('cookiefile') is not None or self.get_param('cookiesfrombrowser') is not None
3749
d77ab8e2 3750 def mark_watched(self, *args, **kwargs):
1813a6cc 3751 if not self.get_param('mark_watched', False):
3752 return
24146491 3753 if self.supports_login() and self._get_login_info()[0] is not None or self._cookies_passed:
d77ab8e2
S
3754 self._mark_watched(*args, **kwargs)
3755
3756 def _mark_watched(self, *args, **kwargs):
3757 raise NotImplementedError('This method must be implemented by subclasses')
3758
38cce791
YCH
3759 def geo_verification_headers(self):
3760 headers = {}
a06916d9 3761 geo_verification_proxy = self.get_param('geo_verification_proxy')
38cce791
YCH
3762 if geo_verification_proxy:
3763 headers['Ytdl-request-proxy'] = geo_verification_proxy
3764 return headers
3765
98763ee3 3766 def _generic_id(self, url):
14f25df2 3767 return urllib.parse.unquote(os.path.splitext(url.rstrip('/').split('/')[-1])[0])
98763ee3
YCH
3768
3769 def _generic_title(self, url):
14f25df2 3770 return urllib.parse.unquote(os.path.splitext(url_basename(url))[0])
98763ee3 3771
c224251a 3772 @staticmethod
b0089e89 3773 def _availability(is_private=None, needs_premium=None, needs_subscription=None, needs_auth=None, is_unlisted=None):
c224251a
M
3774 all_known = all(map(
3775 lambda x: x is not None,
3776 (is_private, needs_premium, needs_subscription, needs_auth, is_unlisted)))
3777 return (
3778 'private' if is_private
3779 else 'premium_only' if needs_premium
3780 else 'subscriber_only' if needs_subscription
3781 else 'needs_auth' if needs_auth
3782 else 'unlisted' if is_unlisted
3783 else 'public' if all_known
3784 else None)
3785
d43de682 3786 def _configuration_arg(self, key, default=NO_DEFAULT, *, ie_key=None, casesense=False):
4bb6b02f 3787 '''
3788 @returns A list of values for the extractor argument given by "key"
3789 or "default" if no such key is present
3790 @param default The default value to return when the key is not present (default: [])
3791 @param casesense When false, the values are converted to lower case
3792 '''
3793 val = traverse_obj(
d43de682 3794 self._downloader.params, ('extractor_args', (ie_key or self.ie_key()).lower(), key))
4bb6b02f 3795 if val is None:
3796 return [] if default is NO_DEFAULT else default
3797 return list(val) if casesense else [x.lower() for x in val]
5d3a0e79 3798
f40ee5e9 3799 def _yes_playlist(self, playlist_id, video_id, smuggled_data=None, *, playlist_label='playlist', video_label='video'):
3800 if not playlist_id or not video_id:
3801 return not video_id
3802
3803 no_playlist = (smuggled_data or {}).get('force_noplaylist')
3804 if no_playlist is not None:
3805 return not no_playlist
3806
3807 video_id = '' if video_id is True else f' {video_id}'
3808 playlist_id = '' if playlist_id is True else f' {playlist_id}'
3809 if self.get_param('noplaylist'):
3810 self.to_screen(f'Downloading just the {video_label}{video_id} because of --no-playlist')
3811 return False
3812 self.to_screen(f'Downloading {playlist_label}{playlist_id} - add --no-playlist to download just the {video_label}{video_id}')
3813 return True
3814
8dbe9899 3815
d6983cb4
PH
3816class SearchInfoExtractor(InfoExtractor):
3817 """
3818 Base class for paged search queries extractors.
10952eb2 3819 They accept URLs in the format _SEARCH_KEY(|all|[0-9]):{query}
96565c7e 3820 Instances should define _SEARCH_KEY and optionally _MAX_RESULTS
d6983cb4
PH
3821 """
3822
96565c7e 3823 _MAX_RESULTS = float('inf')
3824
d6983cb4
PH
3825 @classmethod
3826 def _make_valid_url(cls):
3827 return r'%s(?P<prefix>|[1-9][0-9]*|all):(?P<query>[\s\S]+)' % cls._SEARCH_KEY
3828
d6983cb4 3829 def _real_extract(self, query):
2c4aaadd 3830 prefix, query = self._match_valid_url(query).group('prefix', 'query')
d6983cb4
PH
3831 if prefix == '':
3832 return self._get_n_results(query, 1)
3833 elif prefix == 'all':
3834 return self._get_n_results(query, self._MAX_RESULTS)
3835 else:
3836 n = int(prefix)
3837 if n <= 0:
86e5f3ed 3838 raise ExtractorError(f'invalid download number {n} for query "{query}"')
d6983cb4 3839 elif n > self._MAX_RESULTS:
6a39ee13 3840 self.report_warning('%s returns max %i results (you requested %i)' % (self._SEARCH_KEY, self._MAX_RESULTS, n))
d6983cb4
PH
3841 n = self._MAX_RESULTS
3842 return self._get_n_results(query, n)
3843
3844 def _get_n_results(self, query, n):
cc16383f 3845 """Get a specified number of results for a query.
3846 Either this function or _search_results must be overridden by subclasses """
3847 return self.playlist_result(
3848 itertools.islice(self._search_results(query), 0, None if n == float('inf') else n),
3849 query, query)
3850
3851 def _search_results(self, query):
3852 """Returns an iterator of search results"""
611c1dd9 3853 raise NotImplementedError('This method must be implemented by subclasses')
0f818663 3854
82d02080 3855 @classproperty
3856 def SEARCH_KEY(cls):
3857 return cls._SEARCH_KEY