]> jfr.im git - yt-dlp.git/blob - youtube_dl/extractor/common.py
59ad455c17a4bbc59044fa1febfa0f0f0a9f5a48
[yt-dlp.git] / youtube_dl / extractor / common.py
1 # coding: utf-8
2 from __future__ import unicode_literals
3
4 import base64
5 import datetime
6 import hashlib
7 import json
8 import netrc
9 import os
10 import random
11 import re
12 import socket
13 import sys
14 import time
15 import math
16
17 from ..compat import (
18 compat_cookiejar,
19 compat_cookies,
20 compat_etree_Element,
21 compat_etree_fromstring,
22 compat_getpass,
23 compat_integer_types,
24 compat_http_client,
25 compat_os_name,
26 compat_str,
27 compat_urllib_error,
28 compat_urllib_parse_unquote,
29 compat_urllib_parse_urlencode,
30 compat_urllib_request,
31 compat_urlparse,
32 compat_xml_parse_error,
33 )
34 from ..downloader.f4m import (
35 get_base_url,
36 remove_encrypted_media,
37 )
38 from ..utils import (
39 NO_DEFAULT,
40 age_restricted,
41 base_url,
42 bug_reports_message,
43 clean_html,
44 compiled_regex_type,
45 determine_ext,
46 determine_protocol,
47 dict_get,
48 error_to_compat_str,
49 ExtractorError,
50 extract_attributes,
51 fix_xml_ampersands,
52 float_or_none,
53 GeoRestrictedError,
54 GeoUtils,
55 int_or_none,
56 js_to_json,
57 JSON_LD_RE,
58 mimetype2ext,
59 orderedSet,
60 parse_bitrate,
61 parse_codecs,
62 parse_duration,
63 parse_iso8601,
64 parse_m3u8_attributes,
65 parse_resolution,
66 RegexNotFoundError,
67 sanitized_Request,
68 sanitize_filename,
69 str_or_none,
70 unescapeHTML,
71 unified_strdate,
72 unified_timestamp,
73 update_Request,
74 update_url_query,
75 urljoin,
76 url_basename,
77 url_or_none,
78 xpath_element,
79 xpath_text,
80 xpath_with_ns,
81 )
82
83
84 class InfoExtractor(object):
85 """Information Extractor class.
86
87 Information extractors are the classes that, given a URL, extract
88 information about the video (or videos) the URL refers to. This
89 information includes the real video URL, the video title, author and
90 others. The information is stored in a dictionary which is then
91 passed to the YoutubeDL. The YoutubeDL processes this
92 information possibly downloading the video to the file system, among
93 other possible outcomes.
94
95 The type field determines the type of the result.
96 By far the most common value (and the default if _type is missing) is
97 "video", which indicates a single video.
98
99 For a video, the dictionaries must include the following fields:
100
101 id: Video identifier.
102 title: Video title, unescaped.
103
104 Additionally, it must contain either a formats entry or a url one:
105
106 formats: A list of dictionaries for each format available, ordered
107 from worst to best quality.
108
109 Potential fields:
110 * url The mandatory URL representing the media:
111 for plain file media - HTTP URL of this file,
112 for RTMP - RTMP URL,
113 for HLS - URL of the M3U8 media playlist,
114 for HDS - URL of the F4M manifest,
115 for DASH
116 - HTTP URL to plain file media (in case of
117 unfragmented media)
118 - URL of the MPD manifest or base URL
119 representing the media if MPD manifest
120 is parsed froma string (in case of
121 fragmented media)
122 for MSS - URL of the ISM manifest.
123 * manifest_url
124 The URL of the manifest file in case of
125 fragmented media:
126 for HLS - URL of the M3U8 master playlist,
127 for HDS - URL of the F4M manifest,
128 for DASH - URL of the MPD manifest,
129 for MSS - URL of the ISM manifest.
130 * ext Will be calculated from URL if missing
131 * format A human-readable description of the format
132 ("mp4 container with h264/opus").
133 Calculated from the format_id, width, height.
134 and format_note fields if missing.
135 * format_id A short description of the format
136 ("mp4_h264_opus" or "19").
137 Technically optional, but strongly recommended.
138 * format_note Additional info about the format
139 ("3D" or "DASH video")
140 * width Width of the video, if known
141 * height Height of the video, if known
142 * resolution Textual description of width and height
143 * tbr Average bitrate of audio and video in KBit/s
144 * abr Average audio bitrate in KBit/s
145 * acodec Name of the audio codec in use
146 * asr Audio sampling rate in Hertz
147 * vbr Average video bitrate in KBit/s
148 * fps Frame rate
149 * vcodec Name of the video codec in use
150 * container Name of the container format
151 * filesize The number of bytes, if known in advance
152 * filesize_approx An estimate for the number of bytes
153 * player_url SWF Player URL (used for rtmpdump).
154 * protocol The protocol that will be used for the actual
155 download, lower-case.
156 "http", "https", "rtsp", "rtmp", "rtmpe",
157 "m3u8", "m3u8_native" or "http_dash_segments".
158 * fragment_base_url
159 Base URL for fragments. Each fragment's path
160 value (if present) will be relative to
161 this URL.
162 * fragments A list of fragments of a fragmented media.
163 Each fragment entry must contain either an url
164 or a path. If an url is present it should be
165 considered by a client. Otherwise both path and
166 fragment_base_url must be present. Here is
167 the list of all potential fields:
168 * "url" - fragment's URL
169 * "path" - fragment's path relative to
170 fragment_base_url
171 * "duration" (optional, int or float)
172 * "filesize" (optional, int)
173 * preference Order number of this format. If this field is
174 present and not None, the formats get sorted
175 by this field, regardless of all other values.
176 -1 for default (order by other properties),
177 -2 or smaller for less than default.
178 < -1000 to hide the format (if there is
179 another one which is strictly better)
180 * language Language code, e.g. "de" or "en-US".
181 * language_preference Is this in the language mentioned in
182 the URL?
183 10 if it's what the URL is about,
184 -1 for default (don't know),
185 -10 otherwise, other values reserved for now.
186 * quality Order number of the video quality of this
187 format, irrespective of the file format.
188 -1 for default (order by other properties),
189 -2 or smaller for less than default.
190 * source_preference Order number for this video source
191 (quality takes higher priority)
192 -1 for default (order by other properties),
193 -2 or smaller for less than default.
194 * http_headers A dictionary of additional HTTP headers
195 to add to the request.
196 * stretched_ratio If given and not 1, indicates that the
197 video's pixels are not square.
198 width : height ratio as float.
199 * no_resume The server does not support resuming the
200 (HTTP or RTMP) download. Boolean.
201 * downloader_options A dictionary of downloader options as
202 described in FileDownloader
203
204 url: Final video URL.
205 ext: Video filename extension.
206 format: The video format, defaults to ext (used for --get-format)
207 player_url: SWF Player URL (used for rtmpdump).
208
209 The following fields are optional:
210
211 alt_title: A secondary title of the video.
212 display_id An alternative identifier for the video, not necessarily
213 unique, but available before title. Typically, id is
214 something like "4234987", title "Dancing naked mole rats",
215 and display_id "dancing-naked-mole-rats"
216 thumbnails: A list of dictionaries, with the following entries:
217 * "id" (optional, string) - Thumbnail format ID
218 * "url"
219 * "preference" (optional, int) - quality of the image
220 * "width" (optional, int)
221 * "height" (optional, int)
222 * "resolution" (optional, string "{width}x{height"},
223 deprecated)
224 * "filesize" (optional, int)
225 thumbnail: Full URL to a video thumbnail image.
226 description: Full video description.
227 uploader: Full name of the video uploader.
228 license: License name the video is licensed under.
229 creator: The creator of the video.
230 release_date: The date (YYYYMMDD) when the video was released.
231 timestamp: UNIX timestamp of the moment the video became available.
232 upload_date: Video upload date (YYYYMMDD).
233 If not explicitly set, calculated from timestamp.
234 uploader_id: Nickname or id of the video uploader.
235 uploader_url: Full URL to a personal webpage of the video uploader.
236 channel: Full name of the channel the video is uploaded on.
237 Note that channel fields may or may not repeat uploader
238 fields. This depends on a particular extractor.
239 channel_id: Id of the channel.
240 channel_url: Full URL to a channel webpage.
241 location: Physical location where the video was filmed.
242 subtitles: The available subtitles as a dictionary in the format
243 {tag: subformats}. "tag" is usually a language code, and
244 "subformats" is a list sorted from lower to higher
245 preference, each element is a dictionary with the "ext"
246 entry and one of:
247 * "data": The subtitles file contents
248 * "url": A URL pointing to the subtitles file
249 "ext" will be calculated from URL if missing
250 automatic_captions: Like 'subtitles', used by the YoutubeIE for
251 automatically generated captions
252 duration: Length of the video in seconds, as an integer or float.
253 view_count: How many users have watched the video on the platform.
254 like_count: Number of positive ratings of the video
255 dislike_count: Number of negative ratings of the video
256 repost_count: Number of reposts of the video
257 average_rating: Average rating give by users, the scale used depends on the webpage
258 comment_count: Number of comments on the video
259 comments: A list of comments, each with one or more of the following
260 properties (all but one of text or html optional):
261 * "author" - human-readable name of the comment author
262 * "author_id" - user ID of the comment author
263 * "id" - Comment ID
264 * "html" - Comment as HTML
265 * "text" - Plain text of the comment
266 * "timestamp" - UNIX timestamp of comment
267 * "parent" - ID of the comment this one is replying to.
268 Set to "root" to indicate that this is a
269 comment to the original video.
270 age_limit: Age restriction for the video, as an integer (years)
271 webpage_url: The URL to the video webpage, if given to youtube-dl it
272 should allow to get the same result again. (It will be set
273 by YoutubeDL if it's missing)
274 categories: A list of categories that the video falls in, for example
275 ["Sports", "Berlin"]
276 tags: A list of tags assigned to the video, e.g. ["sweden", "pop music"]
277 is_live: True, False, or None (=unknown). Whether this video is a
278 live stream that goes on instead of a fixed-length video.
279 start_time: Time in seconds where the reproduction should start, as
280 specified in the URL.
281 end_time: Time in seconds where the reproduction should end, as
282 specified in the URL.
283 chapters: A list of dictionaries, with the following entries:
284 * "start_time" - The start time of the chapter in seconds
285 * "end_time" - The end time of the chapter in seconds
286 * "title" (optional, string)
287
288 The following fields should only be used when the video belongs to some logical
289 chapter or section:
290
291 chapter: Name or title of the chapter the video belongs to.
292 chapter_number: Number of the chapter the video belongs to, as an integer.
293 chapter_id: Id of the chapter the video belongs to, as a unicode string.
294
295 The following fields should only be used when the video is an episode of some
296 series, programme or podcast:
297
298 series: Title of the series or programme the video episode belongs to.
299 season: Title of the season the video episode belongs to.
300 season_number: Number of the season the video episode belongs to, as an integer.
301 season_id: Id of the season the video episode belongs to, as a unicode string.
302 episode: Title of the video episode. Unlike mandatory video title field,
303 this field should denote the exact title of the video episode
304 without any kind of decoration.
305 episode_number: Number of the video episode within a season, as an integer.
306 episode_id: Id of the video episode, as a unicode string.
307
308 The following fields should only be used when the media is a track or a part of
309 a music album:
310
311 track: Title of the track.
312 track_number: Number of the track within an album or a disc, as an integer.
313 track_id: Id of the track (useful in case of custom indexing, e.g. 6.iii),
314 as a unicode string.
315 artist: Artist(s) of the track.
316 genre: Genre(s) of the track.
317 album: Title of the album the track belongs to.
318 album_type: Type of the album (e.g. "Demo", "Full-length", "Split", "Compilation", etc).
319 album_artist: List of all artists appeared on the album (e.g.
320 "Ash Borer / Fell Voices" or "Various Artists", useful for splits
321 and compilations).
322 disc_number: Number of the disc or other physical medium the track belongs to,
323 as an integer.
324 release_year: Year (YYYY) when the album was released.
325
326 Unless mentioned otherwise, the fields should be Unicode strings.
327
328 Unless mentioned otherwise, None is equivalent to absence of information.
329
330
331 _type "playlist" indicates multiple videos.
332 There must be a key "entries", which is a list, an iterable, or a PagedList
333 object, each element of which is a valid dictionary by this specification.
334
335 Additionally, playlists can have "id", "title", "description", "uploader",
336 "uploader_id", "uploader_url" attributes with the same semantics as videos
337 (see above).
338
339
340 _type "multi_video" indicates that there are multiple videos that
341 form a single show, for examples multiple acts of an opera or TV episode.
342 It must have an entries key like a playlist and contain all the keys
343 required for a video at the same time.
344
345
346 _type "url" indicates that the video must be extracted from another
347 location, possibly by a different extractor. Its only required key is:
348 "url" - the next URL to extract.
349 The key "ie_key" can be set to the class name (minus the trailing "IE",
350 e.g. "Youtube") if the extractor class is known in advance.
351 Additionally, the dictionary may have any properties of the resolved entity
352 known in advance, for example "title" if the title of the referred video is
353 known ahead of time.
354
355
356 _type "url_transparent" entities have the same specification as "url", but
357 indicate that the given additional information is more precise than the one
358 associated with the resolved URL.
359 This is useful when a site employs a video service that hosts the video and
360 its technical metadata, but that video service does not embed a useful
361 title, description etc.
362
363
364 Subclasses of this one should re-define the _real_initialize() and
365 _real_extract() methods and define a _VALID_URL regexp.
366 Probably, they should also be added to the list of extractors.
367
368 _GEO_BYPASS attribute may be set to False in order to disable
369 geo restriction bypass mechanisms for a particular extractor.
370 Though it won't disable explicit geo restriction bypass based on
371 country code provided with geo_bypass_country.
372
373 _GEO_COUNTRIES attribute may contain a list of presumably geo unrestricted
374 countries for this extractor. One of these countries will be used by
375 geo restriction bypass mechanism right away in order to bypass
376 geo restriction, of course, if the mechanism is not disabled.
377
378 _GEO_IP_BLOCKS attribute may contain a list of presumably geo unrestricted
379 IP blocks in CIDR notation for this extractor. One of these IP blocks
380 will be used by geo restriction bypass mechanism similarly
381 to _GEO_COUNTRIES.
382
383 Finally, the _WORKING attribute should be set to False for broken IEs
384 in order to warn the users and skip the tests.
385 """
386
387 _ready = False
388 _downloader = None
389 _x_forwarded_for_ip = None
390 _GEO_BYPASS = True
391 _GEO_COUNTRIES = None
392 _GEO_IP_BLOCKS = None
393 _WORKING = True
394
395 def __init__(self, downloader=None):
396 """Constructor. Receives an optional downloader."""
397 self._ready = False
398 self._x_forwarded_for_ip = None
399 self.set_downloader(downloader)
400
401 @classmethod
402 def suitable(cls, url):
403 """Receives a URL and returns True if suitable for this IE."""
404
405 # This does not use has/getattr intentionally - we want to know whether
406 # we have cached the regexp for *this* class, whereas getattr would also
407 # match the superclass
408 if '_VALID_URL_RE' not in cls.__dict__:
409 cls._VALID_URL_RE = re.compile(cls._VALID_URL)
410 return cls._VALID_URL_RE.match(url) is not None
411
412 @classmethod
413 def _match_id(cls, url):
414 if '_VALID_URL_RE' not in cls.__dict__:
415 cls._VALID_URL_RE = re.compile(cls._VALID_URL)
416 m = cls._VALID_URL_RE.match(url)
417 assert m
418 return compat_str(m.group('id'))
419
420 @classmethod
421 def working(cls):
422 """Getter method for _WORKING."""
423 return cls._WORKING
424
425 def initialize(self):
426 """Initializes an instance (authentication, etc)."""
427 self._initialize_geo_bypass({
428 'countries': self._GEO_COUNTRIES,
429 'ip_blocks': self._GEO_IP_BLOCKS,
430 })
431 if not self._ready:
432 self._real_initialize()
433 self._ready = True
434
435 def _initialize_geo_bypass(self, geo_bypass_context):
436 """
437 Initialize geo restriction bypass mechanism.
438
439 This method is used to initialize geo bypass mechanism based on faking
440 X-Forwarded-For HTTP header. A random country from provided country list
441 is selected and a random IP belonging to this country is generated. This
442 IP will be passed as X-Forwarded-For HTTP header in all subsequent
443 HTTP requests.
444
445 This method will be used for initial geo bypass mechanism initialization
446 during the instance initialization with _GEO_COUNTRIES and
447 _GEO_IP_BLOCKS.
448
449 You may also manually call it from extractor's code if geo bypass
450 information is not available beforehand (e.g. obtained during
451 extraction) or due to some other reason. In this case you should pass
452 this information in geo bypass context passed as first argument. It may
453 contain following fields:
454
455 countries: List of geo unrestricted countries (similar
456 to _GEO_COUNTRIES)
457 ip_blocks: List of geo unrestricted IP blocks in CIDR notation
458 (similar to _GEO_IP_BLOCKS)
459
460 """
461 if not self._x_forwarded_for_ip:
462
463 # Geo bypass mechanism is explicitly disabled by user
464 if not self._downloader.params.get('geo_bypass', True):
465 return
466
467 if not geo_bypass_context:
468 geo_bypass_context = {}
469
470 # Backward compatibility: previously _initialize_geo_bypass
471 # expected a list of countries, some 3rd party code may still use
472 # it this way
473 if isinstance(geo_bypass_context, (list, tuple)):
474 geo_bypass_context = {
475 'countries': geo_bypass_context,
476 }
477
478 # The whole point of geo bypass mechanism is to fake IP
479 # as X-Forwarded-For HTTP header based on some IP block or
480 # country code.
481
482 # Path 1: bypassing based on IP block in CIDR notation
483
484 # Explicit IP block specified by user, use it right away
485 # regardless of whether extractor is geo bypassable or not
486 ip_block = self._downloader.params.get('geo_bypass_ip_block', None)
487
488 # Otherwise use random IP block from geo bypass context but only
489 # if extractor is known as geo bypassable
490 if not ip_block:
491 ip_blocks = geo_bypass_context.get('ip_blocks')
492 if self._GEO_BYPASS and ip_blocks:
493 ip_block = random.choice(ip_blocks)
494
495 if ip_block:
496 self._x_forwarded_for_ip = GeoUtils.random_ipv4(ip_block)
497 if self._downloader.params.get('verbose', False):
498 self._downloader.to_screen(
499 '[debug] Using fake IP %s as X-Forwarded-For.'
500 % self._x_forwarded_for_ip)
501 return
502
503 # Path 2: bypassing based on country code
504
505 # Explicit country code specified by user, use it right away
506 # regardless of whether extractor is geo bypassable or not
507 country = self._downloader.params.get('geo_bypass_country', None)
508
509 # Otherwise use random country code from geo bypass context but
510 # only if extractor is known as geo bypassable
511 if not country:
512 countries = geo_bypass_context.get('countries')
513 if self._GEO_BYPASS and countries:
514 country = random.choice(countries)
515
516 if country:
517 self._x_forwarded_for_ip = GeoUtils.random_ipv4(country)
518 if self._downloader.params.get('verbose', False):
519 self._downloader.to_screen(
520 '[debug] Using fake IP %s (%s) as X-Forwarded-For.'
521 % (self._x_forwarded_for_ip, country.upper()))
522
523 def extract(self, url):
524 """Extracts URL information and returns it in list of dicts."""
525 try:
526 for _ in range(2):
527 try:
528 self.initialize()
529 ie_result = self._real_extract(url)
530 if self._x_forwarded_for_ip:
531 ie_result['__x_forwarded_for_ip'] = self._x_forwarded_for_ip
532 return ie_result
533 except GeoRestrictedError as e:
534 if self.__maybe_fake_ip_and_retry(e.countries):
535 continue
536 raise
537 except ExtractorError:
538 raise
539 except compat_http_client.IncompleteRead as e:
540 raise ExtractorError('A network error has occurred.', cause=e, expected=True)
541 except (KeyError, StopIteration) as e:
542 raise ExtractorError('An extractor error has occurred.', cause=e)
543
544 def __maybe_fake_ip_and_retry(self, countries):
545 if (not self._downloader.params.get('geo_bypass_country', None) and
546 self._GEO_BYPASS and
547 self._downloader.params.get('geo_bypass', True) and
548 not self._x_forwarded_for_ip and
549 countries):
550 country_code = random.choice(countries)
551 self._x_forwarded_for_ip = GeoUtils.random_ipv4(country_code)
552 if self._x_forwarded_for_ip:
553 self.report_warning(
554 'Video is geo restricted. Retrying extraction with fake IP %s (%s) as X-Forwarded-For.'
555 % (self._x_forwarded_for_ip, country_code.upper()))
556 return True
557 return False
558
559 def set_downloader(self, downloader):
560 """Sets the downloader for this IE."""
561 self._downloader = downloader
562
563 def _real_initialize(self):
564 """Real initialization process. Redefine in subclasses."""
565 pass
566
567 def _real_extract(self, url):
568 """Real extraction process. Redefine in subclasses."""
569 pass
570
571 @classmethod
572 def ie_key(cls):
573 """A string for getting the InfoExtractor with get_info_extractor"""
574 return compat_str(cls.__name__[:-2])
575
576 @property
577 def IE_NAME(self):
578 return compat_str(type(self).__name__[:-2])
579
580 @staticmethod
581 def __can_accept_status_code(err, expected_status):
582 assert isinstance(err, compat_urllib_error.HTTPError)
583 if expected_status is None:
584 return False
585 if isinstance(expected_status, compat_integer_types):
586 return err.code == expected_status
587 elif isinstance(expected_status, (list, tuple)):
588 return err.code in expected_status
589 elif callable(expected_status):
590 return expected_status(err.code) is True
591 else:
592 assert False
593
594 def _request_webpage(self, url_or_request, video_id, note=None, errnote=None, fatal=True, data=None, headers={}, query={}, expected_status=None):
595 """
596 Return the response handle.
597
598 See _download_webpage docstring for arguments specification.
599 """
600 if note is None:
601 self.report_download_webpage(video_id)
602 elif note is not False:
603 if video_id is None:
604 self.to_screen('%s' % (note,))
605 else:
606 self.to_screen('%s: %s' % (video_id, note))
607
608 # Some sites check X-Forwarded-For HTTP header in order to figure out
609 # the origin of the client behind proxy. This allows bypassing geo
610 # restriction by faking this header's value to IP that belongs to some
611 # geo unrestricted country. We will do so once we encounter any
612 # geo restriction error.
613 if self._x_forwarded_for_ip:
614 if 'X-Forwarded-For' not in headers:
615 headers['X-Forwarded-For'] = self._x_forwarded_for_ip
616
617 if isinstance(url_or_request, compat_urllib_request.Request):
618 url_or_request = update_Request(
619 url_or_request, data=data, headers=headers, query=query)
620 else:
621 if query:
622 url_or_request = update_url_query(url_or_request, query)
623 if data is not None or headers:
624 url_or_request = sanitized_Request(url_or_request, data, headers)
625 try:
626 return self._downloader.urlopen(url_or_request)
627 except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
628 if isinstance(err, compat_urllib_error.HTTPError):
629 if self.__can_accept_status_code(err, expected_status):
630 # Retain reference to error to prevent file object from
631 # being closed before it can be read. Works around the
632 # effects of <https://bugs.python.org/issue15002>
633 # introduced in Python 3.4.1.
634 err.fp._error = err
635 return err.fp
636
637 if errnote is False:
638 return False
639 if errnote is None:
640 errnote = 'Unable to download webpage'
641
642 errmsg = '%s: %s' % (errnote, error_to_compat_str(err))
643 if fatal:
644 raise ExtractorError(errmsg, sys.exc_info()[2], cause=err)
645 else:
646 self._downloader.report_warning(errmsg)
647 return False
648
649 def _download_webpage_handle(self, url_or_request, video_id, note=None, errnote=None, fatal=True, encoding=None, data=None, headers={}, query={}, expected_status=None):
650 """
651 Return a tuple (page content as string, URL handle).
652
653 See _download_webpage docstring for arguments specification.
654 """
655 # Strip hashes from the URL (#1038)
656 if isinstance(url_or_request, (compat_str, str)):
657 url_or_request = url_or_request.partition('#')[0]
658
659 urlh = self._request_webpage(url_or_request, video_id, note, errnote, fatal, data=data, headers=headers, query=query, expected_status=expected_status)
660 if urlh is False:
661 assert not fatal
662 return False
663 content = self._webpage_read_content(urlh, url_or_request, video_id, note, errnote, fatal, encoding=encoding)
664 return (content, urlh)
665
666 @staticmethod
667 def _guess_encoding_from_content(content_type, webpage_bytes):
668 m = re.match(r'[a-zA-Z0-9_.-]+/[a-zA-Z0-9_.-]+\s*;\s*charset=(.+)', content_type)
669 if m:
670 encoding = m.group(1)
671 else:
672 m = re.search(br'<meta[^>]+charset=[\'"]?([^\'")]+)[ /\'">]',
673 webpage_bytes[:1024])
674 if m:
675 encoding = m.group(1).decode('ascii')
676 elif webpage_bytes.startswith(b'\xff\xfe'):
677 encoding = 'utf-16'
678 else:
679 encoding = 'utf-8'
680
681 return encoding
682
683 def __check_blocked(self, content):
684 first_block = content[:512]
685 if ('<title>Access to this site is blocked</title>' in content and
686 'Websense' in first_block):
687 msg = 'Access to this webpage has been blocked by Websense filtering software in your network.'
688 blocked_iframe = self._html_search_regex(
689 r'<iframe src="([^"]+)"', content,
690 'Websense information URL', default=None)
691 if blocked_iframe:
692 msg += ' Visit %s for more details' % blocked_iframe
693 raise ExtractorError(msg, expected=True)
694 if '<title>The URL you requested has been blocked</title>' in first_block:
695 msg = (
696 'Access to this webpage has been blocked by Indian censorship. '
697 'Use a VPN or proxy server (with --proxy) to route around it.')
698 block_msg = self._html_search_regex(
699 r'</h1><p>(.*?)</p>',
700 content, 'block message', default=None)
701 if block_msg:
702 msg += ' (Message: "%s")' % block_msg.replace('\n', ' ')
703 raise ExtractorError(msg, expected=True)
704 if ('<title>TTK :: Доступ к ресурсу ограничен</title>' in content and
705 'blocklist.rkn.gov.ru' in content):
706 raise ExtractorError(
707 'Access to this webpage has been blocked by decision of the Russian government. '
708 'Visit http://blocklist.rkn.gov.ru/ for a block reason.',
709 expected=True)
710
711 def _webpage_read_content(self, urlh, url_or_request, video_id, note=None, errnote=None, fatal=True, prefix=None, encoding=None):
712 content_type = urlh.headers.get('Content-Type', '')
713 webpage_bytes = urlh.read()
714 if prefix is not None:
715 webpage_bytes = prefix + webpage_bytes
716 if not encoding:
717 encoding = self._guess_encoding_from_content(content_type, webpage_bytes)
718 if self._downloader.params.get('dump_intermediate_pages', False):
719 self.to_screen('Dumping request to ' + urlh.geturl())
720 dump = base64.b64encode(webpage_bytes).decode('ascii')
721 self._downloader.to_screen(dump)
722 if self._downloader.params.get('write_pages', False):
723 basen = '%s_%s' % (video_id, urlh.geturl())
724 if len(basen) > 240:
725 h = '___' + hashlib.md5(basen.encode('utf-8')).hexdigest()
726 basen = basen[:240 - len(h)] + h
727 raw_filename = basen + '.dump'
728 filename = sanitize_filename(raw_filename, restricted=True)
729 self.to_screen('Saving request to ' + filename)
730 # Working around MAX_PATH limitation on Windows (see
731 # http://msdn.microsoft.com/en-us/library/windows/desktop/aa365247(v=vs.85).aspx)
732 if compat_os_name == 'nt':
733 absfilepath = os.path.abspath(filename)
734 if len(absfilepath) > 259:
735 filename = '\\\\?\\' + absfilepath
736 with open(filename, 'wb') as outf:
737 outf.write(webpage_bytes)
738
739 try:
740 content = webpage_bytes.decode(encoding, 'replace')
741 except LookupError:
742 content = webpage_bytes.decode('utf-8', 'replace')
743
744 self.__check_blocked(content)
745
746 return content
747
748 def _download_webpage(
749 self, url_or_request, video_id, note=None, errnote=None,
750 fatal=True, tries=1, timeout=5, encoding=None, data=None,
751 headers={}, query={}, expected_status=None):
752 """
753 Return the data of the page as a string.
754
755 Arguments:
756 url_or_request -- plain text URL as a string or
757 a compat_urllib_request.Requestobject
758 video_id -- Video/playlist/item identifier (string)
759
760 Keyword arguments:
761 note -- note printed before downloading (string)
762 errnote -- note printed in case of an error (string)
763 fatal -- flag denoting whether error should be considered fatal,
764 i.e. whether it should cause ExtractionError to be raised,
765 otherwise a warning will be reported and extraction continued
766 tries -- number of tries
767 timeout -- sleep interval between tries
768 encoding -- encoding for a page content decoding, guessed automatically
769 when not explicitly specified
770 data -- POST data (bytes)
771 headers -- HTTP headers (dict)
772 query -- URL query (dict)
773 expected_status -- allows to accept failed HTTP requests (non 2xx
774 status code) by explicitly specifying a set of accepted status
775 codes. Can be any of the following entities:
776 - an integer type specifying an exact failed status code to
777 accept
778 - a list or a tuple of integer types specifying a list of
779 failed status codes to accept
780 - a callable accepting an actual failed status code and
781 returning True if it should be accepted
782 Note that this argument does not affect success status codes (2xx)
783 which are always accepted.
784 """
785
786 success = False
787 try_count = 0
788 while success is False:
789 try:
790 res = self._download_webpage_handle(
791 url_or_request, video_id, note, errnote, fatal,
792 encoding=encoding, data=data, headers=headers, query=query,
793 expected_status=expected_status)
794 success = True
795 except compat_http_client.IncompleteRead as e:
796 try_count += 1
797 if try_count >= tries:
798 raise e
799 self._sleep(timeout, video_id)
800 if res is False:
801 return res
802 else:
803 content, _ = res
804 return content
805
806 def _download_xml_handle(
807 self, url_or_request, video_id, note='Downloading XML',
808 errnote='Unable to download XML', transform_source=None,
809 fatal=True, encoding=None, data=None, headers={}, query={},
810 expected_status=None):
811 """
812 Return a tuple (xml as an compat_etree_Element, URL handle).
813
814 See _download_webpage docstring for arguments specification.
815 """
816 res = self._download_webpage_handle(
817 url_or_request, video_id, note, errnote, fatal=fatal,
818 encoding=encoding, data=data, headers=headers, query=query,
819 expected_status=expected_status)
820 if res is False:
821 return res
822 xml_string, urlh = res
823 return self._parse_xml(
824 xml_string, video_id, transform_source=transform_source,
825 fatal=fatal), urlh
826
827 def _download_xml(
828 self, url_or_request, video_id,
829 note='Downloading XML', errnote='Unable to download XML',
830 transform_source=None, fatal=True, encoding=None,
831 data=None, headers={}, query={}, expected_status=None):
832 """
833 Return the xml as an compat_etree_Element.
834
835 See _download_webpage docstring for arguments specification.
836 """
837 res = self._download_xml_handle(
838 url_or_request, video_id, note=note, errnote=errnote,
839 transform_source=transform_source, fatal=fatal, encoding=encoding,
840 data=data, headers=headers, query=query,
841 expected_status=expected_status)
842 return res if res is False else res[0]
843
844 def _parse_xml(self, xml_string, video_id, transform_source=None, fatal=True):
845 if transform_source:
846 xml_string = transform_source(xml_string)
847 try:
848 return compat_etree_fromstring(xml_string.encode('utf-8'))
849 except compat_xml_parse_error as ve:
850 errmsg = '%s: Failed to parse XML ' % video_id
851 if fatal:
852 raise ExtractorError(errmsg, cause=ve)
853 else:
854 self.report_warning(errmsg + str(ve))
855
856 def _download_json_handle(
857 self, url_or_request, video_id, note='Downloading JSON metadata',
858 errnote='Unable to download JSON metadata', transform_source=None,
859 fatal=True, encoding=None, data=None, headers={}, query={},
860 expected_status=None):
861 """
862 Return a tuple (JSON object, URL handle).
863
864 See _download_webpage docstring for arguments specification.
865 """
866 res = self._download_webpage_handle(
867 url_or_request, video_id, note, errnote, fatal=fatal,
868 encoding=encoding, data=data, headers=headers, query=query,
869 expected_status=expected_status)
870 if res is False:
871 return res
872 json_string, urlh = res
873 return self._parse_json(
874 json_string, video_id, transform_source=transform_source,
875 fatal=fatal), urlh
876
877 def _download_json(
878 self, url_or_request, video_id, note='Downloading JSON metadata',
879 errnote='Unable to download JSON metadata', transform_source=None,
880 fatal=True, encoding=None, data=None, headers={}, query={},
881 expected_status=None):
882 """
883 Return the JSON object as a dict.
884
885 See _download_webpage docstring for arguments specification.
886 """
887 res = self._download_json_handle(
888 url_or_request, video_id, note=note, errnote=errnote,
889 transform_source=transform_source, fatal=fatal, encoding=encoding,
890 data=data, headers=headers, query=query,
891 expected_status=expected_status)
892 return res if res is False else res[0]
893
894 def _parse_json(self, json_string, video_id, transform_source=None, fatal=True):
895 if transform_source:
896 json_string = transform_source(json_string)
897 try:
898 return json.loads(json_string)
899 except ValueError as ve:
900 errmsg = '%s: Failed to parse JSON ' % video_id
901 if fatal:
902 raise ExtractorError(errmsg, cause=ve)
903 else:
904 self.report_warning(errmsg + str(ve))
905
906 def report_warning(self, msg, video_id=None):
907 idstr = '' if video_id is None else '%s: ' % video_id
908 self._downloader.report_warning(
909 '[%s] %s%s' % (self.IE_NAME, idstr, msg))
910
911 def to_screen(self, msg):
912 """Print msg to screen, prefixing it with '[ie_name]'"""
913 self._downloader.to_screen('[%s] %s' % (self.IE_NAME, msg))
914
915 def report_extraction(self, id_or_name):
916 """Report information extraction."""
917 self.to_screen('%s: Extracting information' % id_or_name)
918
919 def report_download_webpage(self, video_id):
920 """Report webpage download."""
921 self.to_screen('%s: Downloading webpage' % video_id)
922
923 def report_age_confirmation(self):
924 """Report attempt to confirm age."""
925 self.to_screen('Confirming age')
926
927 def report_login(self):
928 """Report attempt to log in."""
929 self.to_screen('Logging in')
930
931 @staticmethod
932 def raise_login_required(msg='This video is only available for registered users'):
933 raise ExtractorError(
934 '%s. Use --username and --password or --netrc to provide account credentials.' % msg,
935 expected=True)
936
937 @staticmethod
938 def raise_geo_restricted(msg='This video is not available from your location due to geo restriction', countries=None):
939 raise GeoRestrictedError(msg, countries=countries)
940
941 # Methods for following #608
942 @staticmethod
943 def url_result(url, ie=None, video_id=None, video_title=None):
944 """Returns a URL that points to a page that should be processed"""
945 # TODO: ie should be the class used for getting the info
946 video_info = {'_type': 'url',
947 'url': url,
948 'ie_key': ie}
949 if video_id is not None:
950 video_info['id'] = video_id
951 if video_title is not None:
952 video_info['title'] = video_title
953 return video_info
954
955 def playlist_from_matches(self, matches, playlist_id=None, playlist_title=None, getter=None, ie=None):
956 urls = orderedSet(
957 self.url_result(self._proto_relative_url(getter(m) if getter else m), ie)
958 for m in matches)
959 return self.playlist_result(
960 urls, playlist_id=playlist_id, playlist_title=playlist_title)
961
962 @staticmethod
963 def playlist_result(entries, playlist_id=None, playlist_title=None, playlist_description=None):
964 """Returns a playlist"""
965 video_info = {'_type': 'playlist',
966 'entries': entries}
967 if playlist_id:
968 video_info['id'] = playlist_id
969 if playlist_title:
970 video_info['title'] = playlist_title
971 if playlist_description:
972 video_info['description'] = playlist_description
973 return video_info
974
975 def _search_regex(self, pattern, string, name, default=NO_DEFAULT, fatal=True, flags=0, group=None):
976 """
977 Perform a regex search on the given string, using a single or a list of
978 patterns returning the first matching group.
979 In case of failure return a default value or raise a WARNING or a
980 RegexNotFoundError, depending on fatal, specifying the field name.
981 """
982 if isinstance(pattern, (str, compat_str, compiled_regex_type)):
983 mobj = re.search(pattern, string, flags)
984 else:
985 for p in pattern:
986 mobj = re.search(p, string, flags)
987 if mobj:
988 break
989
990 if not self._downloader.params.get('no_color') and compat_os_name != 'nt' and sys.stderr.isatty():
991 _name = '\033[0;34m%s\033[0m' % name
992 else:
993 _name = name
994
995 if mobj:
996 if group is None:
997 # return the first matching group
998 return next(g for g in mobj.groups() if g is not None)
999 else:
1000 return mobj.group(group)
1001 elif default is not NO_DEFAULT:
1002 return default
1003 elif fatal:
1004 raise RegexNotFoundError('Unable to extract %s' % _name)
1005 else:
1006 self._downloader.report_warning('unable to extract %s' % _name + bug_reports_message())
1007 return None
1008
1009 def _html_search_regex(self, pattern, string, name, default=NO_DEFAULT, fatal=True, flags=0, group=None):
1010 """
1011 Like _search_regex, but strips HTML tags and unescapes entities.
1012 """
1013 res = self._search_regex(pattern, string, name, default, fatal, flags, group)
1014 if res:
1015 return clean_html(res).strip()
1016 else:
1017 return res
1018
1019 def _get_netrc_login_info(self, netrc_machine=None):
1020 username = None
1021 password = None
1022 netrc_machine = netrc_machine or self._NETRC_MACHINE
1023
1024 if self._downloader.params.get('usenetrc', False):
1025 try:
1026 info = netrc.netrc().authenticators(netrc_machine)
1027 if info is not None:
1028 username = info[0]
1029 password = info[2]
1030 else:
1031 raise netrc.NetrcParseError(
1032 'No authenticators for %s' % netrc_machine)
1033 except (IOError, netrc.NetrcParseError) as err:
1034 self._downloader.report_warning(
1035 'parsing .netrc: %s' % error_to_compat_str(err))
1036
1037 return username, password
1038
1039 def _get_login_info(self, username_option='username', password_option='password', netrc_machine=None):
1040 """
1041 Get the login info as (username, password)
1042 First look for the manually specified credentials using username_option
1043 and password_option as keys in params dictionary. If no such credentials
1044 available look in the netrc file using the netrc_machine or _NETRC_MACHINE
1045 value.
1046 If there's no info available, return (None, None)
1047 """
1048 if self._downloader is None:
1049 return (None, None)
1050
1051 downloader_params = self._downloader.params
1052
1053 # Attempt to use provided username and password or .netrc data
1054 if downloader_params.get(username_option) is not None:
1055 username = downloader_params[username_option]
1056 password = downloader_params[password_option]
1057 else:
1058 username, password = self._get_netrc_login_info(netrc_machine)
1059
1060 return username, password
1061
1062 def _get_tfa_info(self, note='two-factor verification code'):
1063 """
1064 Get the two-factor authentication info
1065 TODO - asking the user will be required for sms/phone verify
1066 currently just uses the command line option
1067 If there's no info available, return None
1068 """
1069 if self._downloader is None:
1070 return None
1071 downloader_params = self._downloader.params
1072
1073 if downloader_params.get('twofactor') is not None:
1074 return downloader_params['twofactor']
1075
1076 return compat_getpass('Type %s and press [Return]: ' % note)
1077
1078 # Helper functions for extracting OpenGraph info
1079 @staticmethod
1080 def _og_regexes(prop):
1081 content_re = r'content=(?:"([^"]+?)"|\'([^\']+?)\'|\s*([^\s"\'=<>`]+?))'
1082 property_re = (r'(?:name|property)=(?:\'og[:-]%(prop)s\'|"og[:-]%(prop)s"|\s*og[:-]%(prop)s\b)'
1083 % {'prop': re.escape(prop)})
1084 template = r'<meta[^>]+?%s[^>]+?%s'
1085 return [
1086 template % (property_re, content_re),
1087 template % (content_re, property_re),
1088 ]
1089
1090 @staticmethod
1091 def _meta_regex(prop):
1092 return r'''(?isx)<meta
1093 (?=[^>]+(?:itemprop|name|property|id|http-equiv)=(["\']?)%s\1)
1094 [^>]+?content=(["\'])(?P<content>.*?)\2''' % re.escape(prop)
1095
1096 def _og_search_property(self, prop, html, name=None, **kargs):
1097 if not isinstance(prop, (list, tuple)):
1098 prop = [prop]
1099 if name is None:
1100 name = 'OpenGraph %s' % prop[0]
1101 og_regexes = []
1102 for p in prop:
1103 og_regexes.extend(self._og_regexes(p))
1104 escaped = self._search_regex(og_regexes, html, name, flags=re.DOTALL, **kargs)
1105 if escaped is None:
1106 return None
1107 return unescapeHTML(escaped)
1108
1109 def _og_search_thumbnail(self, html, **kargs):
1110 return self._og_search_property('image', html, 'thumbnail URL', fatal=False, **kargs)
1111
1112 def _og_search_description(self, html, **kargs):
1113 return self._og_search_property('description', html, fatal=False, **kargs)
1114
1115 def _og_search_title(self, html, **kargs):
1116 return self._og_search_property('title', html, **kargs)
1117
1118 def _og_search_video_url(self, html, name='video url', secure=True, **kargs):
1119 regexes = self._og_regexes('video') + self._og_regexes('video:url')
1120 if secure:
1121 regexes = self._og_regexes('video:secure_url') + regexes
1122 return self._html_search_regex(regexes, html, name, **kargs)
1123
1124 def _og_search_url(self, html, **kargs):
1125 return self._og_search_property('url', html, **kargs)
1126
1127 def _html_search_meta(self, name, html, display_name=None, fatal=False, **kwargs):
1128 if not isinstance(name, (list, tuple)):
1129 name = [name]
1130 if display_name is None:
1131 display_name = name[0]
1132 return self._html_search_regex(
1133 [self._meta_regex(n) for n in name],
1134 html, display_name, fatal=fatal, group='content', **kwargs)
1135
1136 def _dc_search_uploader(self, html):
1137 return self._html_search_meta('dc.creator', html, 'uploader')
1138
1139 def _rta_search(self, html):
1140 # See http://www.rtalabel.org/index.php?content=howtofaq#single
1141 if re.search(r'(?ix)<meta\s+name="rating"\s+'
1142 r' content="RTA-5042-1996-1400-1577-RTA"',
1143 html):
1144 return 18
1145 return 0
1146
1147 def _media_rating_search(self, html):
1148 # See http://www.tjg-designs.com/WP/metadata-code-examples-adding-metadata-to-your-web-pages/
1149 rating = self._html_search_meta('rating', html)
1150
1151 if not rating:
1152 return None
1153
1154 RATING_TABLE = {
1155 'safe for kids': 0,
1156 'general': 8,
1157 '14 years': 14,
1158 'mature': 17,
1159 'restricted': 19,
1160 }
1161 return RATING_TABLE.get(rating.lower())
1162
1163 def _family_friendly_search(self, html):
1164 # See http://schema.org/VideoObject
1165 family_friendly = self._html_search_meta(
1166 'isFamilyFriendly', html, default=None)
1167
1168 if not family_friendly:
1169 return None
1170
1171 RATING_TABLE = {
1172 '1': 0,
1173 'true': 0,
1174 '0': 18,
1175 'false': 18,
1176 }
1177 return RATING_TABLE.get(family_friendly.lower())
1178
1179 def _twitter_search_player(self, html):
1180 return self._html_search_meta('twitter:player', html,
1181 'twitter card player')
1182
1183 def _search_json_ld(self, html, video_id, expected_type=None, **kwargs):
1184 json_ld = self._search_regex(
1185 JSON_LD_RE, html, 'JSON-LD', group='json_ld', **kwargs)
1186 default = kwargs.get('default', NO_DEFAULT)
1187 if not json_ld:
1188 return default if default is not NO_DEFAULT else {}
1189 # JSON-LD may be malformed and thus `fatal` should be respected.
1190 # At the same time `default` may be passed that assumes `fatal=False`
1191 # for _search_regex. Let's simulate the same behavior here as well.
1192 fatal = kwargs.get('fatal', True) if default == NO_DEFAULT else False
1193 return self._json_ld(json_ld, video_id, fatal=fatal, expected_type=expected_type)
1194
1195 def _json_ld(self, json_ld, video_id, fatal=True, expected_type=None):
1196 if isinstance(json_ld, compat_str):
1197 json_ld = self._parse_json(json_ld, video_id, fatal=fatal)
1198 if not json_ld:
1199 return {}
1200 info = {}
1201 if not isinstance(json_ld, (list, tuple, dict)):
1202 return info
1203 if isinstance(json_ld, dict):
1204 json_ld = [json_ld]
1205
1206 INTERACTION_TYPE_MAP = {
1207 'CommentAction': 'comment',
1208 'AgreeAction': 'like',
1209 'DisagreeAction': 'dislike',
1210 'LikeAction': 'like',
1211 'DislikeAction': 'dislike',
1212 'ListenAction': 'view',
1213 'WatchAction': 'view',
1214 'ViewAction': 'view',
1215 }
1216
1217 def extract_interaction_statistic(e):
1218 interaction_statistic = e.get('interactionStatistic')
1219 if not isinstance(interaction_statistic, list):
1220 return
1221 for is_e in interaction_statistic:
1222 if not isinstance(is_e, dict):
1223 continue
1224 if is_e.get('@type') != 'InteractionCounter':
1225 continue
1226 interaction_type = is_e.get('interactionType')
1227 if not isinstance(interaction_type, compat_str):
1228 continue
1229 interaction_count = int_or_none(is_e.get('userInteractionCount'))
1230 if interaction_count is None:
1231 continue
1232 count_kind = INTERACTION_TYPE_MAP.get(interaction_type.split('/')[-1])
1233 if not count_kind:
1234 continue
1235 count_key = '%s_count' % count_kind
1236 if info.get(count_key) is not None:
1237 continue
1238 info[count_key] = interaction_count
1239
1240 def extract_video_object(e):
1241 assert e['@type'] == 'VideoObject'
1242 info.update({
1243 'url': url_or_none(e.get('contentUrl')),
1244 'title': unescapeHTML(e.get('name')),
1245 'description': unescapeHTML(e.get('description')),
1246 'thumbnail': url_or_none(e.get('thumbnailUrl') or e.get('thumbnailURL')),
1247 'duration': parse_duration(e.get('duration')),
1248 'timestamp': unified_timestamp(e.get('uploadDate')),
1249 'filesize': float_or_none(e.get('contentSize')),
1250 'tbr': int_or_none(e.get('bitrate')),
1251 'width': int_or_none(e.get('width')),
1252 'height': int_or_none(e.get('height')),
1253 'view_count': int_or_none(e.get('interactionCount')),
1254 })
1255 extract_interaction_statistic(e)
1256
1257 for e in json_ld:
1258 if isinstance(e.get('@context'), compat_str) and re.match(r'^https?://schema.org/?$', e.get('@context')):
1259 item_type = e.get('@type')
1260 if expected_type is not None and expected_type != item_type:
1261 return info
1262 if item_type in ('TVEpisode', 'Episode'):
1263 episode_name = unescapeHTML(e.get('name'))
1264 info.update({
1265 'episode': episode_name,
1266 'episode_number': int_or_none(e.get('episodeNumber')),
1267 'description': unescapeHTML(e.get('description')),
1268 })
1269 if not info.get('title') and episode_name:
1270 info['title'] = episode_name
1271 part_of_season = e.get('partOfSeason')
1272 if isinstance(part_of_season, dict) and part_of_season.get('@type') in ('TVSeason', 'Season', 'CreativeWorkSeason'):
1273 info.update({
1274 'season': unescapeHTML(part_of_season.get('name')),
1275 'season_number': int_or_none(part_of_season.get('seasonNumber')),
1276 })
1277 part_of_series = e.get('partOfSeries') or e.get('partOfTVSeries')
1278 if isinstance(part_of_series, dict) and part_of_series.get('@type') in ('TVSeries', 'Series', 'CreativeWorkSeries'):
1279 info['series'] = unescapeHTML(part_of_series.get('name'))
1280 elif item_type == 'Movie':
1281 info.update({
1282 'title': unescapeHTML(e.get('name')),
1283 'description': unescapeHTML(e.get('description')),
1284 'duration': parse_duration(e.get('duration')),
1285 'timestamp': unified_timestamp(e.get('dateCreated')),
1286 })
1287 elif item_type in ('Article', 'NewsArticle'):
1288 info.update({
1289 'timestamp': parse_iso8601(e.get('datePublished')),
1290 'title': unescapeHTML(e.get('headline')),
1291 'description': unescapeHTML(e.get('articleBody')),
1292 })
1293 elif item_type == 'VideoObject':
1294 extract_video_object(e)
1295 continue
1296 video = e.get('video')
1297 if isinstance(video, dict) and video.get('@type') == 'VideoObject':
1298 extract_video_object(video)
1299 break
1300 return dict((k, v) for k, v in info.items() if v is not None)
1301
1302 @staticmethod
1303 def _hidden_inputs(html):
1304 html = re.sub(r'<!--(?:(?!<!--).)*-->', '', html)
1305 hidden_inputs = {}
1306 for input in re.findall(r'(?i)(<input[^>]+>)', html):
1307 attrs = extract_attributes(input)
1308 if not input:
1309 continue
1310 if attrs.get('type') not in ('hidden', 'submit'):
1311 continue
1312 name = attrs.get('name') or attrs.get('id')
1313 value = attrs.get('value')
1314 if name and value is not None:
1315 hidden_inputs[name] = value
1316 return hidden_inputs
1317
1318 def _form_hidden_inputs(self, form_id, html):
1319 form = self._search_regex(
1320 r'(?is)<form[^>]+?id=(["\'])%s\1[^>]*>(?P<form>.+?)</form>' % form_id,
1321 html, '%s form' % form_id, group='form')
1322 return self._hidden_inputs(form)
1323
1324 def _sort_formats(self, formats, field_preference=None):
1325 if not formats:
1326 raise ExtractorError('No video formats found')
1327
1328 for f in formats:
1329 # Automatically determine tbr when missing based on abr and vbr (improves
1330 # formats sorting in some cases)
1331 if 'tbr' not in f and f.get('abr') is not None and f.get('vbr') is not None:
1332 f['tbr'] = f['abr'] + f['vbr']
1333
1334 def _formats_key(f):
1335 # TODO remove the following workaround
1336 from ..utils import determine_ext
1337 if not f.get('ext') and 'url' in f:
1338 f['ext'] = determine_ext(f['url'])
1339
1340 if isinstance(field_preference, (list, tuple)):
1341 return tuple(
1342 f.get(field)
1343 if f.get(field) is not None
1344 else ('' if field == 'format_id' else -1)
1345 for field in field_preference)
1346
1347 preference = f.get('preference')
1348 if preference is None:
1349 preference = 0
1350 if f.get('ext') in ['f4f', 'f4m']: # Not yet supported
1351 preference -= 0.5
1352
1353 protocol = f.get('protocol') or determine_protocol(f)
1354 proto_preference = 0 if protocol in ['http', 'https'] else (-0.5 if protocol == 'rtsp' else -0.1)
1355
1356 if f.get('vcodec') == 'none': # audio only
1357 preference -= 50
1358 if self._downloader.params.get('prefer_free_formats'):
1359 ORDER = ['aac', 'mp3', 'm4a', 'webm', 'ogg', 'opus']
1360 else:
1361 ORDER = ['webm', 'opus', 'ogg', 'mp3', 'aac', 'm4a']
1362 ext_preference = 0
1363 try:
1364 audio_ext_preference = ORDER.index(f['ext'])
1365 except ValueError:
1366 audio_ext_preference = -1
1367 else:
1368 if f.get('acodec') == 'none': # video only
1369 preference -= 40
1370 if self._downloader.params.get('prefer_free_formats'):
1371 ORDER = ['flv', 'mp4', 'webm']
1372 else:
1373 ORDER = ['webm', 'flv', 'mp4']
1374 try:
1375 ext_preference = ORDER.index(f['ext'])
1376 except ValueError:
1377 ext_preference = -1
1378 audio_ext_preference = 0
1379
1380 return (
1381 preference,
1382 f.get('language_preference') if f.get('language_preference') is not None else -1,
1383 f.get('quality') if f.get('quality') is not None else -1,
1384 f.get('tbr') if f.get('tbr') is not None else -1,
1385 f.get('filesize') if f.get('filesize') is not None else -1,
1386 f.get('vbr') if f.get('vbr') is not None else -1,
1387 f.get('height') if f.get('height') is not None else -1,
1388 f.get('width') if f.get('width') is not None else -1,
1389 proto_preference,
1390 ext_preference,
1391 f.get('abr') if f.get('abr') is not None else -1,
1392 audio_ext_preference,
1393 f.get('fps') if f.get('fps') is not None else -1,
1394 f.get('filesize_approx') if f.get('filesize_approx') is not None else -1,
1395 f.get('source_preference') if f.get('source_preference') is not None else -1,
1396 f.get('format_id') if f.get('format_id') is not None else '',
1397 )
1398 formats.sort(key=_formats_key)
1399
1400 def _check_formats(self, formats, video_id):
1401 if formats:
1402 formats[:] = filter(
1403 lambda f: self._is_valid_url(
1404 f['url'], video_id,
1405 item='%s video format' % f.get('format_id') if f.get('format_id') else 'video'),
1406 formats)
1407
1408 @staticmethod
1409 def _remove_duplicate_formats(formats):
1410 format_urls = set()
1411 unique_formats = []
1412 for f in formats:
1413 if f['url'] not in format_urls:
1414 format_urls.add(f['url'])
1415 unique_formats.append(f)
1416 formats[:] = unique_formats
1417
1418 def _is_valid_url(self, url, video_id, item='video', headers={}):
1419 url = self._proto_relative_url(url, scheme='http:')
1420 # For now assume non HTTP(S) URLs always valid
1421 if not (url.startswith('http://') or url.startswith('https://')):
1422 return True
1423 try:
1424 self._request_webpage(url, video_id, 'Checking %s URL' % item, headers=headers)
1425 return True
1426 except ExtractorError as e:
1427 if isinstance(e.cause, compat_urllib_error.URLError):
1428 self.to_screen(
1429 '%s: %s URL is invalid, skipping' % (video_id, item))
1430 return False
1431 raise
1432
1433 def http_scheme(self):
1434 """ Either "http:" or "https:", depending on the user's preferences """
1435 return (
1436 'http:'
1437 if self._downloader.params.get('prefer_insecure', False)
1438 else 'https:')
1439
1440 def _proto_relative_url(self, url, scheme=None):
1441 if url is None:
1442 return url
1443 if url.startswith('//'):
1444 if scheme is None:
1445 scheme = self.http_scheme()
1446 return scheme + url
1447 else:
1448 return url
1449
1450 def _sleep(self, timeout, video_id, msg_template=None):
1451 if msg_template is None:
1452 msg_template = '%(video_id)s: Waiting for %(timeout)s seconds'
1453 msg = msg_template % {'video_id': video_id, 'timeout': timeout}
1454 self.to_screen(msg)
1455 time.sleep(timeout)
1456
1457 def _extract_f4m_formats(self, manifest_url, video_id, preference=None, f4m_id=None,
1458 transform_source=lambda s: fix_xml_ampersands(s).strip(),
1459 fatal=True, m3u8_id=None):
1460 manifest = self._download_xml(
1461 manifest_url, video_id, 'Downloading f4m manifest',
1462 'Unable to download f4m manifest',
1463 # Some manifests may be malformed, e.g. prosiebensat1 generated manifests
1464 # (see https://github.com/ytdl-org/youtube-dl/issues/6215#issuecomment-121704244)
1465 transform_source=transform_source,
1466 fatal=fatal)
1467
1468 if manifest is False:
1469 return []
1470
1471 return self._parse_f4m_formats(
1472 manifest, manifest_url, video_id, preference=preference, f4m_id=f4m_id,
1473 transform_source=transform_source, fatal=fatal, m3u8_id=m3u8_id)
1474
1475 def _parse_f4m_formats(self, manifest, manifest_url, video_id, preference=None, f4m_id=None,
1476 transform_source=lambda s: fix_xml_ampersands(s).strip(),
1477 fatal=True, m3u8_id=None):
1478 if not isinstance(manifest, compat_etree_Element) and not fatal:
1479 return []
1480
1481 # currently youtube-dl cannot decode the playerVerificationChallenge as Akamai uses Adobe Alchemy
1482 akamai_pv = manifest.find('{http://ns.adobe.com/f4m/1.0}pv-2.0')
1483 if akamai_pv is not None and ';' in akamai_pv.text:
1484 playerVerificationChallenge = akamai_pv.text.split(';')[0]
1485 if playerVerificationChallenge.strip() != '':
1486 return []
1487
1488 formats = []
1489 manifest_version = '1.0'
1490 media_nodes = manifest.findall('{http://ns.adobe.com/f4m/1.0}media')
1491 if not media_nodes:
1492 manifest_version = '2.0'
1493 media_nodes = manifest.findall('{http://ns.adobe.com/f4m/2.0}media')
1494 # Remove unsupported DRM protected media from final formats
1495 # rendition (see https://github.com/ytdl-org/youtube-dl/issues/8573).
1496 media_nodes = remove_encrypted_media(media_nodes)
1497 if not media_nodes:
1498 return formats
1499
1500 manifest_base_url = get_base_url(manifest)
1501
1502 bootstrap_info = xpath_element(
1503 manifest, ['{http://ns.adobe.com/f4m/1.0}bootstrapInfo', '{http://ns.adobe.com/f4m/2.0}bootstrapInfo'],
1504 'bootstrap info', default=None)
1505
1506 vcodec = None
1507 mime_type = xpath_text(
1508 manifest, ['{http://ns.adobe.com/f4m/1.0}mimeType', '{http://ns.adobe.com/f4m/2.0}mimeType'],
1509 'base URL', default=None)
1510 if mime_type and mime_type.startswith('audio/'):
1511 vcodec = 'none'
1512
1513 for i, media_el in enumerate(media_nodes):
1514 tbr = int_or_none(media_el.attrib.get('bitrate'))
1515 width = int_or_none(media_el.attrib.get('width'))
1516 height = int_or_none(media_el.attrib.get('height'))
1517 format_id = '-'.join(filter(None, [f4m_id, compat_str(i if tbr is None else tbr)]))
1518 # If <bootstrapInfo> is present, the specified f4m is a
1519 # stream-level manifest, and only set-level manifests may refer to
1520 # external resources. See section 11.4 and section 4 of F4M spec
1521 if bootstrap_info is None:
1522 media_url = None
1523 # @href is introduced in 2.0, see section 11.6 of F4M spec
1524 if manifest_version == '2.0':
1525 media_url = media_el.attrib.get('href')
1526 if media_url is None:
1527 media_url = media_el.attrib.get('url')
1528 if not media_url:
1529 continue
1530 manifest_url = (
1531 media_url if media_url.startswith('http://') or media_url.startswith('https://')
1532 else ((manifest_base_url or '/'.join(manifest_url.split('/')[:-1])) + '/' + media_url))
1533 # If media_url is itself a f4m manifest do the recursive extraction
1534 # since bitrates in parent manifest (this one) and media_url manifest
1535 # may differ leading to inability to resolve the format by requested
1536 # bitrate in f4m downloader
1537 ext = determine_ext(manifest_url)
1538 if ext == 'f4m':
1539 f4m_formats = self._extract_f4m_formats(
1540 manifest_url, video_id, preference=preference, f4m_id=f4m_id,
1541 transform_source=transform_source, fatal=fatal)
1542 # Sometimes stream-level manifest contains single media entry that
1543 # does not contain any quality metadata (e.g. http://matchtv.ru/#live-player).
1544 # At the same time parent's media entry in set-level manifest may
1545 # contain it. We will copy it from parent in such cases.
1546 if len(f4m_formats) == 1:
1547 f = f4m_formats[0]
1548 f.update({
1549 'tbr': f.get('tbr') or tbr,
1550 'width': f.get('width') or width,
1551 'height': f.get('height') or height,
1552 'format_id': f.get('format_id') if not tbr else format_id,
1553 'vcodec': vcodec,
1554 })
1555 formats.extend(f4m_formats)
1556 continue
1557 elif ext == 'm3u8':
1558 formats.extend(self._extract_m3u8_formats(
1559 manifest_url, video_id, 'mp4', preference=preference,
1560 m3u8_id=m3u8_id, fatal=fatal))
1561 continue
1562 formats.append({
1563 'format_id': format_id,
1564 'url': manifest_url,
1565 'manifest_url': manifest_url,
1566 'ext': 'flv' if bootstrap_info is not None else None,
1567 'protocol': 'f4m',
1568 'tbr': tbr,
1569 'width': width,
1570 'height': height,
1571 'vcodec': vcodec,
1572 'preference': preference,
1573 })
1574 return formats
1575
1576 def _m3u8_meta_format(self, m3u8_url, ext=None, preference=None, m3u8_id=None):
1577 return {
1578 'format_id': '-'.join(filter(None, [m3u8_id, 'meta'])),
1579 'url': m3u8_url,
1580 'ext': ext,
1581 'protocol': 'm3u8',
1582 'preference': preference - 100 if preference else -100,
1583 'resolution': 'multiple',
1584 'format_note': 'Quality selection URL',
1585 }
1586
1587 def _extract_m3u8_formats(self, m3u8_url, video_id, ext=None,
1588 entry_protocol='m3u8', preference=None,
1589 m3u8_id=None, note=None, errnote=None,
1590 fatal=True, live=False):
1591 res = self._download_webpage_handle(
1592 m3u8_url, video_id,
1593 note=note or 'Downloading m3u8 information',
1594 errnote=errnote or 'Failed to download m3u8 information',
1595 fatal=fatal)
1596
1597 if res is False:
1598 return []
1599
1600 m3u8_doc, urlh = res
1601 m3u8_url = urlh.geturl()
1602
1603 return self._parse_m3u8_formats(
1604 m3u8_doc, m3u8_url, ext=ext, entry_protocol=entry_protocol,
1605 preference=preference, m3u8_id=m3u8_id, live=live)
1606
1607 def _parse_m3u8_formats(self, m3u8_doc, m3u8_url, ext=None,
1608 entry_protocol='m3u8', preference=None,
1609 m3u8_id=None, live=False):
1610 if '#EXT-X-FAXS-CM:' in m3u8_doc: # Adobe Flash Access
1611 return []
1612
1613 if re.search(r'#EXT-X-SESSION-KEY:.*?URI="skd://', m3u8_doc): # Apple FairPlay
1614 return []
1615
1616 formats = []
1617
1618 format_url = lambda u: (
1619 u
1620 if re.match(r'^https?://', u)
1621 else compat_urlparse.urljoin(m3u8_url, u))
1622
1623 # References:
1624 # 1. https://tools.ietf.org/html/draft-pantos-http-live-streaming-21
1625 # 2. https://github.com/ytdl-org/youtube-dl/issues/12211
1626 # 3. https://github.com/ytdl-org/youtube-dl/issues/18923
1627
1628 # We should try extracting formats only from master playlists [1, 4.3.4],
1629 # i.e. playlists that describe available qualities. On the other hand
1630 # media playlists [1, 4.3.3] should be returned as is since they contain
1631 # just the media without qualities renditions.
1632 # Fortunately, master playlist can be easily distinguished from media
1633 # playlist based on particular tags availability. As of [1, 4.3.3, 4.3.4]
1634 # master playlist tags MUST NOT appear in a media playist and vice versa.
1635 # As of [1, 4.3.3.1] #EXT-X-TARGETDURATION tag is REQUIRED for every
1636 # media playlist and MUST NOT appear in master playlist thus we can
1637 # clearly detect media playlist with this criterion.
1638
1639 if '#EXT-X-TARGETDURATION' in m3u8_doc: # media playlist, return as is
1640 return [{
1641 'url': m3u8_url,
1642 'format_id': m3u8_id,
1643 'ext': ext,
1644 'protocol': entry_protocol,
1645 'preference': preference,
1646 }]
1647
1648 groups = {}
1649 last_stream_inf = {}
1650
1651 def extract_media(x_media_line):
1652 media = parse_m3u8_attributes(x_media_line)
1653 # As per [1, 4.3.4.1] TYPE, GROUP-ID and NAME are REQUIRED
1654 media_type, group_id, name = media.get('TYPE'), media.get('GROUP-ID'), media.get('NAME')
1655 if not (media_type and group_id and name):
1656 return
1657 groups.setdefault(group_id, []).append(media)
1658 if media_type not in ('VIDEO', 'AUDIO'):
1659 return
1660 media_url = media.get('URI')
1661 if media_url:
1662 format_id = []
1663 for v in (m3u8_id, group_id, name):
1664 if v:
1665 format_id.append(v)
1666 f = {
1667 'format_id': '-'.join(format_id),
1668 'url': format_url(media_url),
1669 'manifest_url': m3u8_url,
1670 'language': media.get('LANGUAGE'),
1671 'ext': ext,
1672 'protocol': entry_protocol,
1673 'preference': preference,
1674 }
1675 if media_type == 'AUDIO':
1676 f['vcodec'] = 'none'
1677 formats.append(f)
1678
1679 def build_stream_name():
1680 # Despite specification does not mention NAME attribute for
1681 # EXT-X-STREAM-INF tag it still sometimes may be present (see [1]
1682 # or vidio test in TestInfoExtractor.test_parse_m3u8_formats)
1683 # 1. http://www.vidio.com/watch/165683-dj_ambred-booyah-live-2015
1684 stream_name = last_stream_inf.get('NAME')
1685 if stream_name:
1686 return stream_name
1687 # If there is no NAME in EXT-X-STREAM-INF it will be obtained
1688 # from corresponding rendition group
1689 stream_group_id = last_stream_inf.get('VIDEO')
1690 if not stream_group_id:
1691 return
1692 stream_group = groups.get(stream_group_id)
1693 if not stream_group:
1694 return stream_group_id
1695 rendition = stream_group[0]
1696 return rendition.get('NAME') or stream_group_id
1697
1698 # parse EXT-X-MEDIA tags before EXT-X-STREAM-INF in order to have the
1699 # chance to detect video only formats when EXT-X-STREAM-INF tags
1700 # precede EXT-X-MEDIA tags in HLS manifest such as [3].
1701 for line in m3u8_doc.splitlines():
1702 if line.startswith('#EXT-X-MEDIA:'):
1703 extract_media(line)
1704
1705 for line in m3u8_doc.splitlines():
1706 if line.startswith('#EXT-X-STREAM-INF:'):
1707 last_stream_inf = parse_m3u8_attributes(line)
1708 elif line.startswith('#') or not line.strip():
1709 continue
1710 else:
1711 tbr = float_or_none(
1712 last_stream_inf.get('AVERAGE-BANDWIDTH') or
1713 last_stream_inf.get('BANDWIDTH'), scale=1000)
1714 format_id = []
1715 if m3u8_id:
1716 format_id.append(m3u8_id)
1717 stream_name = build_stream_name()
1718 # Bandwidth of live streams may differ over time thus making
1719 # format_id unpredictable. So it's better to keep provided
1720 # format_id intact.
1721 if not live:
1722 format_id.append(stream_name if stream_name else '%d' % (tbr if tbr else len(formats)))
1723 manifest_url = format_url(line.strip())
1724 f = {
1725 'format_id': '-'.join(format_id),
1726 'url': manifest_url,
1727 'manifest_url': m3u8_url,
1728 'tbr': tbr,
1729 'ext': ext,
1730 'fps': float_or_none(last_stream_inf.get('FRAME-RATE')),
1731 'protocol': entry_protocol,
1732 'preference': preference,
1733 }
1734 resolution = last_stream_inf.get('RESOLUTION')
1735 if resolution:
1736 mobj = re.search(r'(?P<width>\d+)[xX](?P<height>\d+)', resolution)
1737 if mobj:
1738 f['width'] = int(mobj.group('width'))
1739 f['height'] = int(mobj.group('height'))
1740 # Unified Streaming Platform
1741 mobj = re.search(
1742 r'audio.*?(?:%3D|=)(\d+)(?:-video.*?(?:%3D|=)(\d+))?', f['url'])
1743 if mobj:
1744 abr, vbr = mobj.groups()
1745 abr, vbr = float_or_none(abr, 1000), float_or_none(vbr, 1000)
1746 f.update({
1747 'vbr': vbr,
1748 'abr': abr,
1749 })
1750 codecs = parse_codecs(last_stream_inf.get('CODECS'))
1751 f.update(codecs)
1752 audio_group_id = last_stream_inf.get('AUDIO')
1753 # As per [1, 4.3.4.1.1] any EXT-X-STREAM-INF tag which
1754 # references a rendition group MUST have a CODECS attribute.
1755 # However, this is not always respected, for example, [2]
1756 # contains EXT-X-STREAM-INF tag which references AUDIO
1757 # rendition group but does not have CODECS and despite
1758 # referencing an audio group it represents a complete
1759 # (with audio and video) format. So, for such cases we will
1760 # ignore references to rendition groups and treat them
1761 # as complete formats.
1762 if audio_group_id and codecs and f.get('vcodec') != 'none':
1763 audio_group = groups.get(audio_group_id)
1764 if audio_group and audio_group[0].get('URI'):
1765 # TODO: update acodec for audio only formats with
1766 # the same GROUP-ID
1767 f['acodec'] = 'none'
1768 formats.append(f)
1769 last_stream_inf = {}
1770 return formats
1771
1772 @staticmethod
1773 def _xpath_ns(path, namespace=None):
1774 if not namespace:
1775 return path
1776 out = []
1777 for c in path.split('/'):
1778 if not c or c == '.':
1779 out.append(c)
1780 else:
1781 out.append('{%s}%s' % (namespace, c))
1782 return '/'.join(out)
1783
1784 def _extract_smil_formats(self, smil_url, video_id, fatal=True, f4m_params=None, transform_source=None):
1785 smil = self._download_smil(smil_url, video_id, fatal=fatal, transform_source=transform_source)
1786
1787 if smil is False:
1788 assert not fatal
1789 return []
1790
1791 namespace = self._parse_smil_namespace(smil)
1792
1793 return self._parse_smil_formats(
1794 smil, smil_url, video_id, namespace=namespace, f4m_params=f4m_params)
1795
1796 def _extract_smil_info(self, smil_url, video_id, fatal=True, f4m_params=None):
1797 smil = self._download_smil(smil_url, video_id, fatal=fatal)
1798 if smil is False:
1799 return {}
1800 return self._parse_smil(smil, smil_url, video_id, f4m_params=f4m_params)
1801
1802 def _download_smil(self, smil_url, video_id, fatal=True, transform_source=None):
1803 return self._download_xml(
1804 smil_url, video_id, 'Downloading SMIL file',
1805 'Unable to download SMIL file', fatal=fatal, transform_source=transform_source)
1806
1807 def _parse_smil(self, smil, smil_url, video_id, f4m_params=None):
1808 namespace = self._parse_smil_namespace(smil)
1809
1810 formats = self._parse_smil_formats(
1811 smil, smil_url, video_id, namespace=namespace, f4m_params=f4m_params)
1812 subtitles = self._parse_smil_subtitles(smil, namespace=namespace)
1813
1814 video_id = os.path.splitext(url_basename(smil_url))[0]
1815 title = None
1816 description = None
1817 upload_date = None
1818 for meta in smil.findall(self._xpath_ns('./head/meta', namespace)):
1819 name = meta.attrib.get('name')
1820 content = meta.attrib.get('content')
1821 if not name or not content:
1822 continue
1823 if not title and name == 'title':
1824 title = content
1825 elif not description and name in ('description', 'abstract'):
1826 description = content
1827 elif not upload_date and name == 'date':
1828 upload_date = unified_strdate(content)
1829
1830 thumbnails = [{
1831 'id': image.get('type'),
1832 'url': image.get('src'),
1833 'width': int_or_none(image.get('width')),
1834 'height': int_or_none(image.get('height')),
1835 } for image in smil.findall(self._xpath_ns('.//image', namespace)) if image.get('src')]
1836
1837 return {
1838 'id': video_id,
1839 'title': title or video_id,
1840 'description': description,
1841 'upload_date': upload_date,
1842 'thumbnails': thumbnails,
1843 'formats': formats,
1844 'subtitles': subtitles,
1845 }
1846
1847 def _parse_smil_namespace(self, smil):
1848 return self._search_regex(
1849 r'(?i)^{([^}]+)?}smil$', smil.tag, 'namespace', default=None)
1850
1851 def _parse_smil_formats(self, smil, smil_url, video_id, namespace=None, f4m_params=None, transform_rtmp_url=None):
1852 base = smil_url
1853 for meta in smil.findall(self._xpath_ns('./head/meta', namespace)):
1854 b = meta.get('base') or meta.get('httpBase')
1855 if b:
1856 base = b
1857 break
1858
1859 formats = []
1860 rtmp_count = 0
1861 http_count = 0
1862 m3u8_count = 0
1863
1864 srcs = []
1865 media = smil.findall(self._xpath_ns('.//video', namespace)) + smil.findall(self._xpath_ns('.//audio', namespace))
1866 for medium in media:
1867 src = medium.get('src')
1868 if not src or src in srcs:
1869 continue
1870 srcs.append(src)
1871
1872 bitrate = float_or_none(medium.get('system-bitrate') or medium.get('systemBitrate'), 1000)
1873 filesize = int_or_none(medium.get('size') or medium.get('fileSize'))
1874 width = int_or_none(medium.get('width'))
1875 height = int_or_none(medium.get('height'))
1876 proto = medium.get('proto')
1877 ext = medium.get('ext')
1878 src_ext = determine_ext(src)
1879 streamer = medium.get('streamer') or base
1880
1881 if proto == 'rtmp' or streamer.startswith('rtmp'):
1882 rtmp_count += 1
1883 formats.append({
1884 'url': streamer,
1885 'play_path': src,
1886 'ext': 'flv',
1887 'format_id': 'rtmp-%d' % (rtmp_count if bitrate is None else bitrate),
1888 'tbr': bitrate,
1889 'filesize': filesize,
1890 'width': width,
1891 'height': height,
1892 })
1893 if transform_rtmp_url:
1894 streamer, src = transform_rtmp_url(streamer, src)
1895 formats[-1].update({
1896 'url': streamer,
1897 'play_path': src,
1898 })
1899 continue
1900
1901 src_url = src if src.startswith('http') else compat_urlparse.urljoin(base, src)
1902 src_url = src_url.strip()
1903
1904 if proto == 'm3u8' or src_ext == 'm3u8':
1905 m3u8_formats = self._extract_m3u8_formats(
1906 src_url, video_id, ext or 'mp4', m3u8_id='hls', fatal=False)
1907 if len(m3u8_formats) == 1:
1908 m3u8_count += 1
1909 m3u8_formats[0].update({
1910 'format_id': 'hls-%d' % (m3u8_count if bitrate is None else bitrate),
1911 'tbr': bitrate,
1912 'width': width,
1913 'height': height,
1914 })
1915 formats.extend(m3u8_formats)
1916 elif src_ext == 'f4m':
1917 f4m_url = src_url
1918 if not f4m_params:
1919 f4m_params = {
1920 'hdcore': '3.2.0',
1921 'plugin': 'flowplayer-3.2.0.1',
1922 }
1923 f4m_url += '&' if '?' in f4m_url else '?'
1924 f4m_url += compat_urllib_parse_urlencode(f4m_params)
1925 formats.extend(self._extract_f4m_formats(f4m_url, video_id, f4m_id='hds', fatal=False))
1926 elif src_ext == 'mpd':
1927 formats.extend(self._extract_mpd_formats(
1928 src_url, video_id, mpd_id='dash', fatal=False))
1929 elif re.search(r'\.ism/[Mm]anifest', src_url):
1930 formats.extend(self._extract_ism_formats(
1931 src_url, video_id, ism_id='mss', fatal=False))
1932 elif src_url.startswith('http') and self._is_valid_url(src, video_id):
1933 http_count += 1
1934 formats.append({
1935 'url': src_url,
1936 'ext': ext or src_ext or 'flv',
1937 'format_id': 'http-%d' % (bitrate or http_count),
1938 'tbr': bitrate,
1939 'filesize': filesize,
1940 'width': width,
1941 'height': height,
1942 })
1943
1944 return formats
1945
1946 def _parse_smil_subtitles(self, smil, namespace=None, subtitles_lang='en'):
1947 urls = []
1948 subtitles = {}
1949 for num, textstream in enumerate(smil.findall(self._xpath_ns('.//textstream', namespace))):
1950 src = textstream.get('src')
1951 if not src or src in urls:
1952 continue
1953 urls.append(src)
1954 ext = textstream.get('ext') or mimetype2ext(textstream.get('type')) or determine_ext(src)
1955 lang = textstream.get('systemLanguage') or textstream.get('systemLanguageName') or textstream.get('lang') or subtitles_lang
1956 subtitles.setdefault(lang, []).append({
1957 'url': src,
1958 'ext': ext,
1959 })
1960 return subtitles
1961
1962 def _extract_xspf_playlist(self, xspf_url, playlist_id, fatal=True):
1963 xspf = self._download_xml(
1964 xspf_url, playlist_id, 'Downloading xpsf playlist',
1965 'Unable to download xspf manifest', fatal=fatal)
1966 if xspf is False:
1967 return []
1968 return self._parse_xspf(
1969 xspf, playlist_id, xspf_url=xspf_url,
1970 xspf_base_url=base_url(xspf_url))
1971
1972 def _parse_xspf(self, xspf_doc, playlist_id, xspf_url=None, xspf_base_url=None):
1973 NS_MAP = {
1974 'xspf': 'http://xspf.org/ns/0/',
1975 's1': 'http://static.streamone.nl/player/ns/0',
1976 }
1977
1978 entries = []
1979 for track in xspf_doc.findall(xpath_with_ns('./xspf:trackList/xspf:track', NS_MAP)):
1980 title = xpath_text(
1981 track, xpath_with_ns('./xspf:title', NS_MAP), 'title', default=playlist_id)
1982 description = xpath_text(
1983 track, xpath_with_ns('./xspf:annotation', NS_MAP), 'description')
1984 thumbnail = xpath_text(
1985 track, xpath_with_ns('./xspf:image', NS_MAP), 'thumbnail')
1986 duration = float_or_none(
1987 xpath_text(track, xpath_with_ns('./xspf:duration', NS_MAP), 'duration'), 1000)
1988
1989 formats = []
1990 for location in track.findall(xpath_with_ns('./xspf:location', NS_MAP)):
1991 format_url = urljoin(xspf_base_url, location.text)
1992 if not format_url:
1993 continue
1994 formats.append({
1995 'url': format_url,
1996 'manifest_url': xspf_url,
1997 'format_id': location.get(xpath_with_ns('s1:label', NS_MAP)),
1998 'width': int_or_none(location.get(xpath_with_ns('s1:width', NS_MAP))),
1999 'height': int_or_none(location.get(xpath_with_ns('s1:height', NS_MAP))),
2000 })
2001 self._sort_formats(formats)
2002
2003 entries.append({
2004 'id': playlist_id,
2005 'title': title,
2006 'description': description,
2007 'thumbnail': thumbnail,
2008 'duration': duration,
2009 'formats': formats,
2010 })
2011 return entries
2012
2013 def _extract_mpd_formats(self, mpd_url, video_id, mpd_id=None, note=None, errnote=None, fatal=True, formats_dict={}):
2014 res = self._download_xml_handle(
2015 mpd_url, video_id,
2016 note=note or 'Downloading MPD manifest',
2017 errnote=errnote or 'Failed to download MPD manifest',
2018 fatal=fatal)
2019 if res is False:
2020 return []
2021 mpd_doc, urlh = res
2022 if mpd_doc is None:
2023 return []
2024 mpd_base_url = base_url(urlh.geturl())
2025
2026 return self._parse_mpd_formats(
2027 mpd_doc, mpd_id=mpd_id, mpd_base_url=mpd_base_url,
2028 formats_dict=formats_dict, mpd_url=mpd_url)
2029
2030 def _parse_mpd_formats(self, mpd_doc, mpd_id=None, mpd_base_url='', formats_dict={}, mpd_url=None):
2031 """
2032 Parse formats from MPD manifest.
2033 References:
2034 1. MPEG-DASH Standard, ISO/IEC 23009-1:2014(E),
2035 http://standards.iso.org/ittf/PubliclyAvailableStandards/c065274_ISO_IEC_23009-1_2014.zip
2036 2. https://en.wikipedia.org/wiki/Dynamic_Adaptive_Streaming_over_HTTP
2037 """
2038 if mpd_doc.get('type') == 'dynamic':
2039 return []
2040
2041 namespace = self._search_regex(r'(?i)^{([^}]+)?}MPD$', mpd_doc.tag, 'namespace', default=None)
2042
2043 def _add_ns(path):
2044 return self._xpath_ns(path, namespace)
2045
2046 def is_drm_protected(element):
2047 return element.find(_add_ns('ContentProtection')) is not None
2048
2049 def extract_multisegment_info(element, ms_parent_info):
2050 ms_info = ms_parent_info.copy()
2051
2052 # As per [1, 5.3.9.2.2] SegmentList and SegmentTemplate share some
2053 # common attributes and elements. We will only extract relevant
2054 # for us.
2055 def extract_common(source):
2056 segment_timeline = source.find(_add_ns('SegmentTimeline'))
2057 if segment_timeline is not None:
2058 s_e = segment_timeline.findall(_add_ns('S'))
2059 if s_e:
2060 ms_info['total_number'] = 0
2061 ms_info['s'] = []
2062 for s in s_e:
2063 r = int(s.get('r', 0))
2064 ms_info['total_number'] += 1 + r
2065 ms_info['s'].append({
2066 't': int(s.get('t', 0)),
2067 # @d is mandatory (see [1, 5.3.9.6.2, Table 17, page 60])
2068 'd': int(s.attrib['d']),
2069 'r': r,
2070 })
2071 start_number = source.get('startNumber')
2072 if start_number:
2073 ms_info['start_number'] = int(start_number)
2074 timescale = source.get('timescale')
2075 if timescale:
2076 ms_info['timescale'] = int(timescale)
2077 segment_duration = source.get('duration')
2078 if segment_duration:
2079 ms_info['segment_duration'] = float(segment_duration)
2080
2081 def extract_Initialization(source):
2082 initialization = source.find(_add_ns('Initialization'))
2083 if initialization is not None:
2084 ms_info['initialization_url'] = initialization.attrib['sourceURL']
2085
2086 segment_list = element.find(_add_ns('SegmentList'))
2087 if segment_list is not None:
2088 extract_common(segment_list)
2089 extract_Initialization(segment_list)
2090 segment_urls_e = segment_list.findall(_add_ns('SegmentURL'))
2091 if segment_urls_e:
2092 ms_info['segment_urls'] = [segment.attrib['media'] for segment in segment_urls_e]
2093 else:
2094 segment_template = element.find(_add_ns('SegmentTemplate'))
2095 if segment_template is not None:
2096 extract_common(segment_template)
2097 media = segment_template.get('media')
2098 if media:
2099 ms_info['media'] = media
2100 initialization = segment_template.get('initialization')
2101 if initialization:
2102 ms_info['initialization'] = initialization
2103 else:
2104 extract_Initialization(segment_template)
2105 return ms_info
2106
2107 mpd_duration = parse_duration(mpd_doc.get('mediaPresentationDuration'))
2108 formats = []
2109 for period in mpd_doc.findall(_add_ns('Period')):
2110 period_duration = parse_duration(period.get('duration')) or mpd_duration
2111 period_ms_info = extract_multisegment_info(period, {
2112 'start_number': 1,
2113 'timescale': 1,
2114 })
2115 for adaptation_set in period.findall(_add_ns('AdaptationSet')):
2116 if is_drm_protected(adaptation_set):
2117 continue
2118 adaption_set_ms_info = extract_multisegment_info(adaptation_set, period_ms_info)
2119 for representation in adaptation_set.findall(_add_ns('Representation')):
2120 if is_drm_protected(representation):
2121 continue
2122 representation_attrib = adaptation_set.attrib.copy()
2123 representation_attrib.update(representation.attrib)
2124 # According to [1, 5.3.7.2, Table 9, page 41], @mimeType is mandatory
2125 mime_type = representation_attrib['mimeType']
2126 content_type = mime_type.split('/')[0]
2127 if content_type == 'text':
2128 # TODO implement WebVTT downloading
2129 pass
2130 elif content_type in ('video', 'audio'):
2131 base_url = ''
2132 for element in (representation, adaptation_set, period, mpd_doc):
2133 base_url_e = element.find(_add_ns('BaseURL'))
2134 if base_url_e is not None:
2135 base_url = base_url_e.text + base_url
2136 if re.match(r'^https?://', base_url):
2137 break
2138 if mpd_base_url and not re.match(r'^https?://', base_url):
2139 if not mpd_base_url.endswith('/') and not base_url.startswith('/'):
2140 mpd_base_url += '/'
2141 base_url = mpd_base_url + base_url
2142 representation_id = representation_attrib.get('id')
2143 lang = representation_attrib.get('lang')
2144 url_el = representation.find(_add_ns('BaseURL'))
2145 filesize = int_or_none(url_el.attrib.get('{http://youtube.com/yt/2012/10/10}contentLength') if url_el is not None else None)
2146 bandwidth = int_or_none(representation_attrib.get('bandwidth'))
2147 f = {
2148 'format_id': '%s-%s' % (mpd_id, representation_id) if mpd_id else representation_id,
2149 'manifest_url': mpd_url,
2150 'ext': mimetype2ext(mime_type),
2151 'width': int_or_none(representation_attrib.get('width')),
2152 'height': int_or_none(representation_attrib.get('height')),
2153 'tbr': float_or_none(bandwidth, 1000),
2154 'asr': int_or_none(representation_attrib.get('audioSamplingRate')),
2155 'fps': int_or_none(representation_attrib.get('frameRate')),
2156 'language': lang if lang not in ('mul', 'und', 'zxx', 'mis') else None,
2157 'format_note': 'DASH %s' % content_type,
2158 'filesize': filesize,
2159 'container': mimetype2ext(mime_type) + '_dash',
2160 }
2161 f.update(parse_codecs(representation_attrib.get('codecs')))
2162 representation_ms_info = extract_multisegment_info(representation, adaption_set_ms_info)
2163
2164 def prepare_template(template_name, identifiers):
2165 tmpl = representation_ms_info[template_name]
2166 # First of, % characters outside $...$ templates
2167 # must be escaped by doubling for proper processing
2168 # by % operator string formatting used further (see
2169 # https://github.com/ytdl-org/youtube-dl/issues/16867).
2170 t = ''
2171 in_template = False
2172 for c in tmpl:
2173 t += c
2174 if c == '$':
2175 in_template = not in_template
2176 elif c == '%' and not in_template:
2177 t += c
2178 # Next, $...$ templates are translated to their
2179 # %(...) counterparts to be used with % operator
2180 t = t.replace('$RepresentationID$', representation_id)
2181 t = re.sub(r'\$(%s)\$' % '|'.join(identifiers), r'%(\1)d', t)
2182 t = re.sub(r'\$(%s)%%([^$]+)\$' % '|'.join(identifiers), r'%(\1)\2', t)
2183 t.replace('$$', '$')
2184 return t
2185
2186 # @initialization is a regular template like @media one
2187 # so it should be handled just the same way (see
2188 # https://github.com/ytdl-org/youtube-dl/issues/11605)
2189 if 'initialization' in representation_ms_info:
2190 initialization_template = prepare_template(
2191 'initialization',
2192 # As per [1, 5.3.9.4.2, Table 15, page 54] $Number$ and
2193 # $Time$ shall not be included for @initialization thus
2194 # only $Bandwidth$ remains
2195 ('Bandwidth', ))
2196 representation_ms_info['initialization_url'] = initialization_template % {
2197 'Bandwidth': bandwidth,
2198 }
2199
2200 def location_key(location):
2201 return 'url' if re.match(r'^https?://', location) else 'path'
2202
2203 if 'segment_urls' not in representation_ms_info and 'media' in representation_ms_info:
2204
2205 media_template = prepare_template('media', ('Number', 'Bandwidth', 'Time'))
2206 media_location_key = location_key(media_template)
2207
2208 # As per [1, 5.3.9.4.4, Table 16, page 55] $Number$ and $Time$
2209 # can't be used at the same time
2210 if '%(Number' in media_template and 's' not in representation_ms_info:
2211 segment_duration = None
2212 if 'total_number' not in representation_ms_info and 'segment_duration' in representation_ms_info:
2213 segment_duration = float_or_none(representation_ms_info['segment_duration'], representation_ms_info['timescale'])
2214 representation_ms_info['total_number'] = int(math.ceil(float(period_duration) / segment_duration))
2215 representation_ms_info['fragments'] = [{
2216 media_location_key: media_template % {
2217 'Number': segment_number,
2218 'Bandwidth': bandwidth,
2219 },
2220 'duration': segment_duration,
2221 } for segment_number in range(
2222 representation_ms_info['start_number'],
2223 representation_ms_info['total_number'] + representation_ms_info['start_number'])]
2224 else:
2225 # $Number*$ or $Time$ in media template with S list available
2226 # Example $Number*$: http://www.svtplay.se/klipp/9023742/stopptid-om-bjorn-borg
2227 # Example $Time$: https://play.arkena.com/embed/avp/v2/player/media/b41dda37-d8e7-4d3f-b1b5-9a9db578bdfe/1/129411
2228 representation_ms_info['fragments'] = []
2229 segment_time = 0
2230 segment_d = None
2231 segment_number = representation_ms_info['start_number']
2232
2233 def add_segment_url():
2234 segment_url = media_template % {
2235 'Time': segment_time,
2236 'Bandwidth': bandwidth,
2237 'Number': segment_number,
2238 }
2239 representation_ms_info['fragments'].append({
2240 media_location_key: segment_url,
2241 'duration': float_or_none(segment_d, representation_ms_info['timescale']),
2242 })
2243
2244 for num, s in enumerate(representation_ms_info['s']):
2245 segment_time = s.get('t') or segment_time
2246 segment_d = s['d']
2247 add_segment_url()
2248 segment_number += 1
2249 for r in range(s.get('r', 0)):
2250 segment_time += segment_d
2251 add_segment_url()
2252 segment_number += 1
2253 segment_time += segment_d
2254 elif 'segment_urls' in representation_ms_info and 's' in representation_ms_info:
2255 # No media template
2256 # Example: https://www.youtube.com/watch?v=iXZV5uAYMJI
2257 # or any YouTube dashsegments video
2258 fragments = []
2259 segment_index = 0
2260 timescale = representation_ms_info['timescale']
2261 for s in representation_ms_info['s']:
2262 duration = float_or_none(s['d'], timescale)
2263 for r in range(s.get('r', 0) + 1):
2264 segment_uri = representation_ms_info['segment_urls'][segment_index]
2265 fragments.append({
2266 location_key(segment_uri): segment_uri,
2267 'duration': duration,
2268 })
2269 segment_index += 1
2270 representation_ms_info['fragments'] = fragments
2271 elif 'segment_urls' in representation_ms_info:
2272 # Segment URLs with no SegmentTimeline
2273 # Example: https://www.seznam.cz/zpravy/clanek/cesko-zasahne-vitr-o-sile-vichrice-muze-byt-i-zivotu-nebezpecny-39091
2274 # https://github.com/ytdl-org/youtube-dl/pull/14844
2275 fragments = []
2276 segment_duration = float_or_none(
2277 representation_ms_info['segment_duration'],
2278 representation_ms_info['timescale']) if 'segment_duration' in representation_ms_info else None
2279 for segment_url in representation_ms_info['segment_urls']:
2280 fragment = {
2281 location_key(segment_url): segment_url,
2282 }
2283 if segment_duration:
2284 fragment['duration'] = segment_duration
2285 fragments.append(fragment)
2286 representation_ms_info['fragments'] = fragments
2287 # If there is a fragments key available then we correctly recognized fragmented media.
2288 # Otherwise we will assume unfragmented media with direct access. Technically, such
2289 # assumption is not necessarily correct since we may simply have no support for
2290 # some forms of fragmented media renditions yet, but for now we'll use this fallback.
2291 if 'fragments' in representation_ms_info:
2292 f.update({
2293 # NB: mpd_url may be empty when MPD manifest is parsed from a string
2294 'url': mpd_url or base_url,
2295 'fragment_base_url': base_url,
2296 'fragments': [],
2297 'protocol': 'http_dash_segments',
2298 })
2299 if 'initialization_url' in representation_ms_info:
2300 initialization_url = representation_ms_info['initialization_url']
2301 if not f.get('url'):
2302 f['url'] = initialization_url
2303 f['fragments'].append({location_key(initialization_url): initialization_url})
2304 f['fragments'].extend(representation_ms_info['fragments'])
2305 else:
2306 # Assuming direct URL to unfragmented media.
2307 f['url'] = base_url
2308
2309 # According to [1, 5.3.5.2, Table 7, page 35] @id of Representation
2310 # is not necessarily unique within a Period thus formats with
2311 # the same `format_id` are quite possible. There are numerous examples
2312 # of such manifests (see https://github.com/ytdl-org/youtube-dl/issues/15111,
2313 # https://github.com/ytdl-org/youtube-dl/issues/13919)
2314 full_info = formats_dict.get(representation_id, {}).copy()
2315 full_info.update(f)
2316 formats.append(full_info)
2317 else:
2318 self.report_warning('Unknown MIME type %s in DASH manifest' % mime_type)
2319 return formats
2320
2321 def _extract_ism_formats(self, ism_url, video_id, ism_id=None, note=None, errnote=None, fatal=True):
2322 res = self._download_xml_handle(
2323 ism_url, video_id,
2324 note=note or 'Downloading ISM manifest',
2325 errnote=errnote or 'Failed to download ISM manifest',
2326 fatal=fatal)
2327 if res is False:
2328 return []
2329 ism_doc, urlh = res
2330
2331 return self._parse_ism_formats(ism_doc, urlh.geturl(), ism_id)
2332
2333 def _parse_ism_formats(self, ism_doc, ism_url, ism_id=None):
2334 """
2335 Parse formats from ISM manifest.
2336 References:
2337 1. [MS-SSTR]: Smooth Streaming Protocol,
2338 https://msdn.microsoft.com/en-us/library/ff469518.aspx
2339 """
2340 if ism_doc.get('IsLive') == 'TRUE' or ism_doc.find('Protection') is not None:
2341 return []
2342
2343 duration = int(ism_doc.attrib['Duration'])
2344 timescale = int_or_none(ism_doc.get('TimeScale')) or 10000000
2345
2346 formats = []
2347 for stream in ism_doc.findall('StreamIndex'):
2348 stream_type = stream.get('Type')
2349 if stream_type not in ('video', 'audio'):
2350 continue
2351 url_pattern = stream.attrib['Url']
2352 stream_timescale = int_or_none(stream.get('TimeScale')) or timescale
2353 stream_name = stream.get('Name')
2354 for track in stream.findall('QualityLevel'):
2355 fourcc = track.get('FourCC', 'AACL' if track.get('AudioTag') == '255' else None)
2356 # TODO: add support for WVC1 and WMAP
2357 if fourcc not in ('H264', 'AVC1', 'AACL'):
2358 self.report_warning('%s is not a supported codec' % fourcc)
2359 continue
2360 tbr = int(track.attrib['Bitrate']) // 1000
2361 # [1] does not mention Width and Height attributes. However,
2362 # they're often present while MaxWidth and MaxHeight are
2363 # missing, so should be used as fallbacks
2364 width = int_or_none(track.get('MaxWidth') or track.get('Width'))
2365 height = int_or_none(track.get('MaxHeight') or track.get('Height'))
2366 sampling_rate = int_or_none(track.get('SamplingRate'))
2367
2368 track_url_pattern = re.sub(r'{[Bb]itrate}', track.attrib['Bitrate'], url_pattern)
2369 track_url_pattern = compat_urlparse.urljoin(ism_url, track_url_pattern)
2370
2371 fragments = []
2372 fragment_ctx = {
2373 'time': 0,
2374 }
2375 stream_fragments = stream.findall('c')
2376 for stream_fragment_index, stream_fragment in enumerate(stream_fragments):
2377 fragment_ctx['time'] = int_or_none(stream_fragment.get('t')) or fragment_ctx['time']
2378 fragment_repeat = int_or_none(stream_fragment.get('r')) or 1
2379 fragment_ctx['duration'] = int_or_none(stream_fragment.get('d'))
2380 if not fragment_ctx['duration']:
2381 try:
2382 next_fragment_time = int(stream_fragment[stream_fragment_index + 1].attrib['t'])
2383 except IndexError:
2384 next_fragment_time = duration
2385 fragment_ctx['duration'] = (next_fragment_time - fragment_ctx['time']) / fragment_repeat
2386 for _ in range(fragment_repeat):
2387 fragments.append({
2388 'url': re.sub(r'{start[ _]time}', compat_str(fragment_ctx['time']), track_url_pattern),
2389 'duration': fragment_ctx['duration'] / stream_timescale,
2390 })
2391 fragment_ctx['time'] += fragment_ctx['duration']
2392
2393 format_id = []
2394 if ism_id:
2395 format_id.append(ism_id)
2396 if stream_name:
2397 format_id.append(stream_name)
2398 format_id.append(compat_str(tbr))
2399
2400 formats.append({
2401 'format_id': '-'.join(format_id),
2402 'url': ism_url,
2403 'manifest_url': ism_url,
2404 'ext': 'ismv' if stream_type == 'video' else 'isma',
2405 'width': width,
2406 'height': height,
2407 'tbr': tbr,
2408 'asr': sampling_rate,
2409 'vcodec': 'none' if stream_type == 'audio' else fourcc,
2410 'acodec': 'none' if stream_type == 'video' else fourcc,
2411 'protocol': 'ism',
2412 'fragments': fragments,
2413 '_download_params': {
2414 'duration': duration,
2415 'timescale': stream_timescale,
2416 'width': width or 0,
2417 'height': height or 0,
2418 'fourcc': fourcc,
2419 'codec_private_data': track.get('CodecPrivateData'),
2420 'sampling_rate': sampling_rate,
2421 'channels': int_or_none(track.get('Channels', 2)),
2422 'bits_per_sample': int_or_none(track.get('BitsPerSample', 16)),
2423 'nal_unit_length_field': int_or_none(track.get('NALUnitLengthField', 4)),
2424 },
2425 })
2426 return formats
2427
2428 def _parse_html5_media_entries(self, base_url, webpage, video_id, m3u8_id=None, m3u8_entry_protocol='m3u8', mpd_id=None, preference=None):
2429 def absolute_url(item_url):
2430 return urljoin(base_url, item_url)
2431
2432 def parse_content_type(content_type):
2433 if not content_type:
2434 return {}
2435 ctr = re.search(r'(?P<mimetype>[^/]+/[^;]+)(?:;\s*codecs="?(?P<codecs>[^"]+))?', content_type)
2436 if ctr:
2437 mimetype, codecs = ctr.groups()
2438 f = parse_codecs(codecs)
2439 f['ext'] = mimetype2ext(mimetype)
2440 return f
2441 return {}
2442
2443 def _media_formats(src, cur_media_type, type_info={}):
2444 full_url = absolute_url(src)
2445 ext = type_info.get('ext') or determine_ext(full_url)
2446 if ext == 'm3u8':
2447 is_plain_url = False
2448 formats = self._extract_m3u8_formats(
2449 full_url, video_id, ext='mp4',
2450 entry_protocol=m3u8_entry_protocol, m3u8_id=m3u8_id,
2451 preference=preference, fatal=False)
2452 elif ext == 'mpd':
2453 is_plain_url = False
2454 formats = self._extract_mpd_formats(
2455 full_url, video_id, mpd_id=mpd_id, fatal=False)
2456 else:
2457 is_plain_url = True
2458 formats = [{
2459 'url': full_url,
2460 'vcodec': 'none' if cur_media_type == 'audio' else None,
2461 }]
2462 return is_plain_url, formats
2463
2464 entries = []
2465 # amp-video and amp-audio are very similar to their HTML5 counterparts
2466 # so we wll include them right here (see
2467 # https://www.ampproject.org/docs/reference/components/amp-video)
2468 media_tags = [(media_tag, media_type, '')
2469 for media_tag, media_type
2470 in re.findall(r'(?s)(<(?:amp-)?(video|audio)[^>]*/>)', webpage)]
2471 media_tags.extend(re.findall(
2472 # We only allow video|audio followed by a whitespace or '>'.
2473 # Allowing more characters may end up in significant slow down (see
2474 # https://github.com/ytdl-org/youtube-dl/issues/11979, example URL:
2475 # http://www.porntrex.com/maps/videositemap.xml).
2476 r'(?s)(<(?P<tag>(?:amp-)?(?:video|audio))(?:\s+[^>]*)?>)(.*?)</(?P=tag)>', webpage))
2477 for media_tag, media_type, media_content in media_tags:
2478 media_info = {
2479 'formats': [],
2480 'subtitles': {},
2481 }
2482 media_attributes = extract_attributes(media_tag)
2483 src = media_attributes.get('src')
2484 if src:
2485 _, formats = _media_formats(src, media_type)
2486 media_info['formats'].extend(formats)
2487 media_info['thumbnail'] = absolute_url(media_attributes.get('poster'))
2488 if media_content:
2489 for source_tag in re.findall(r'<source[^>]+>', media_content):
2490 s_attr = extract_attributes(source_tag)
2491 # data-video-src and data-src are non standard but seen
2492 # several times in the wild
2493 src = dict_get(s_attr, ('src', 'data-video-src', 'data-src'))
2494 if not src:
2495 continue
2496 f = parse_content_type(s_attr.get('type'))
2497 is_plain_url, formats = _media_formats(src, media_type, f)
2498 if is_plain_url:
2499 # width, height, res, label and title attributes are
2500 # all not standard but seen several times in the wild
2501 labels = [
2502 s_attr.get(lbl)
2503 for lbl in ('label', 'title')
2504 if str_or_none(s_attr.get(lbl))
2505 ]
2506 width = int_or_none(s_attr.get('width'))
2507 height = (int_or_none(s_attr.get('height')) or
2508 int_or_none(s_attr.get('res')))
2509 if not width or not height:
2510 for lbl in labels:
2511 resolution = parse_resolution(lbl)
2512 if not resolution:
2513 continue
2514 width = width or resolution.get('width')
2515 height = height or resolution.get('height')
2516 for lbl in labels:
2517 tbr = parse_bitrate(lbl)
2518 if tbr:
2519 break
2520 else:
2521 tbr = None
2522 f.update({
2523 'width': width,
2524 'height': height,
2525 'tbr': tbr,
2526 'format_id': s_attr.get('label') or s_attr.get('title'),
2527 })
2528 f.update(formats[0])
2529 media_info['formats'].append(f)
2530 else:
2531 media_info['formats'].extend(formats)
2532 for track_tag in re.findall(r'<track[^>]+>', media_content):
2533 track_attributes = extract_attributes(track_tag)
2534 kind = track_attributes.get('kind')
2535 if not kind or kind in ('subtitles', 'captions'):
2536 src = track_attributes.get('src')
2537 if not src:
2538 continue
2539 lang = track_attributes.get('srclang') or track_attributes.get('lang') or track_attributes.get('label')
2540 media_info['subtitles'].setdefault(lang, []).append({
2541 'url': absolute_url(src),
2542 })
2543 for f in media_info['formats']:
2544 f.setdefault('http_headers', {})['Referer'] = base_url
2545 if media_info['formats'] or media_info['subtitles']:
2546 entries.append(media_info)
2547 return entries
2548
2549 def _extract_akamai_formats(self, manifest_url, video_id, hosts={}):
2550 formats = []
2551 hdcore_sign = 'hdcore=3.7.0'
2552 f4m_url = re.sub(r'(https?://[^/]+)/i/', r'\1/z/', manifest_url).replace('/master.m3u8', '/manifest.f4m')
2553 hds_host = hosts.get('hds')
2554 if hds_host:
2555 f4m_url = re.sub(r'(https?://)[^/]+', r'\1' + hds_host, f4m_url)
2556 if 'hdcore=' not in f4m_url:
2557 f4m_url += ('&' if '?' in f4m_url else '?') + hdcore_sign
2558 f4m_formats = self._extract_f4m_formats(
2559 f4m_url, video_id, f4m_id='hds', fatal=False)
2560 for entry in f4m_formats:
2561 entry.update({'extra_param_to_segment_url': hdcore_sign})
2562 formats.extend(f4m_formats)
2563 m3u8_url = re.sub(r'(https?://[^/]+)/z/', r'\1/i/', manifest_url).replace('/manifest.f4m', '/master.m3u8')
2564 hls_host = hosts.get('hls')
2565 if hls_host:
2566 m3u8_url = re.sub(r'(https?://)[^/]+', r'\1' + hls_host, m3u8_url)
2567 formats.extend(self._extract_m3u8_formats(
2568 m3u8_url, video_id, 'mp4', 'm3u8_native',
2569 m3u8_id='hls', fatal=False))
2570 return formats
2571
2572 def _extract_wowza_formats(self, url, video_id, m3u8_entry_protocol='m3u8_native', skip_protocols=[]):
2573 query = compat_urlparse.urlparse(url).query
2574 url = re.sub(r'/(?:manifest|playlist|jwplayer)\.(?:m3u8|f4m|mpd|smil)', '', url)
2575 mobj = re.search(
2576 r'(?:(?:http|rtmp|rtsp)(?P<s>s)?:)?(?P<url>//[^?]+)', url)
2577 url_base = mobj.group('url')
2578 http_base_url = '%s%s:%s' % ('http', mobj.group('s') or '', url_base)
2579 formats = []
2580
2581 def manifest_url(manifest):
2582 m_url = '%s/%s' % (http_base_url, manifest)
2583 if query:
2584 m_url += '?%s' % query
2585 return m_url
2586
2587 if 'm3u8' not in skip_protocols:
2588 formats.extend(self._extract_m3u8_formats(
2589 manifest_url('playlist.m3u8'), video_id, 'mp4',
2590 m3u8_entry_protocol, m3u8_id='hls', fatal=False))
2591 if 'f4m' not in skip_protocols:
2592 formats.extend(self._extract_f4m_formats(
2593 manifest_url('manifest.f4m'),
2594 video_id, f4m_id='hds', fatal=False))
2595 if 'dash' not in skip_protocols:
2596 formats.extend(self._extract_mpd_formats(
2597 manifest_url('manifest.mpd'),
2598 video_id, mpd_id='dash', fatal=False))
2599 if re.search(r'(?:/smil:|\.smil)', url_base):
2600 if 'smil' not in skip_protocols:
2601 rtmp_formats = self._extract_smil_formats(
2602 manifest_url('jwplayer.smil'),
2603 video_id, fatal=False)
2604 for rtmp_format in rtmp_formats:
2605 rtsp_format = rtmp_format.copy()
2606 rtsp_format['url'] = '%s/%s' % (rtmp_format['url'], rtmp_format['play_path'])
2607 del rtsp_format['play_path']
2608 del rtsp_format['ext']
2609 rtsp_format.update({
2610 'url': rtsp_format['url'].replace('rtmp://', 'rtsp://'),
2611 'format_id': rtmp_format['format_id'].replace('rtmp', 'rtsp'),
2612 'protocol': 'rtsp',
2613 })
2614 formats.extend([rtmp_format, rtsp_format])
2615 else:
2616 for protocol in ('rtmp', 'rtsp'):
2617 if protocol not in skip_protocols:
2618 formats.append({
2619 'url': '%s:%s' % (protocol, url_base),
2620 'format_id': protocol,
2621 'protocol': protocol,
2622 })
2623 return formats
2624
2625 def _find_jwplayer_data(self, webpage, video_id=None, transform_source=js_to_json):
2626 mobj = re.search(
2627 r'(?s)jwplayer\((?P<quote>[\'"])[^\'" ]+(?P=quote)\)(?!</script>).*?\.setup\s*\((?P<options>[^)]+)\)',
2628 webpage)
2629 if mobj:
2630 try:
2631 jwplayer_data = self._parse_json(mobj.group('options'),
2632 video_id=video_id,
2633 transform_source=transform_source)
2634 except ExtractorError:
2635 pass
2636 else:
2637 if isinstance(jwplayer_data, dict):
2638 return jwplayer_data
2639
2640 def _extract_jwplayer_data(self, webpage, video_id, *args, **kwargs):
2641 jwplayer_data = self._find_jwplayer_data(
2642 webpage, video_id, transform_source=js_to_json)
2643 return self._parse_jwplayer_data(
2644 jwplayer_data, video_id, *args, **kwargs)
2645
2646 def _parse_jwplayer_data(self, jwplayer_data, video_id=None, require_title=True,
2647 m3u8_id=None, mpd_id=None, rtmp_params=None, base_url=None):
2648 # JWPlayer backward compatibility: flattened playlists
2649 # https://github.com/jwplayer/jwplayer/blob/v7.4.3/src/js/api/config.js#L81-L96
2650 if 'playlist' not in jwplayer_data:
2651 jwplayer_data = {'playlist': [jwplayer_data]}
2652
2653 entries = []
2654
2655 # JWPlayer backward compatibility: single playlist item
2656 # https://github.com/jwplayer/jwplayer/blob/v7.7.0/src/js/playlist/playlist.js#L10
2657 if not isinstance(jwplayer_data['playlist'], list):
2658 jwplayer_data['playlist'] = [jwplayer_data['playlist']]
2659
2660 for video_data in jwplayer_data['playlist']:
2661 # JWPlayer backward compatibility: flattened sources
2662 # https://github.com/jwplayer/jwplayer/blob/v7.4.3/src/js/playlist/item.js#L29-L35
2663 if 'sources' not in video_data:
2664 video_data['sources'] = [video_data]
2665
2666 this_video_id = video_id or video_data['mediaid']
2667
2668 formats = self._parse_jwplayer_formats(
2669 video_data['sources'], video_id=this_video_id, m3u8_id=m3u8_id,
2670 mpd_id=mpd_id, rtmp_params=rtmp_params, base_url=base_url)
2671
2672 subtitles = {}
2673 tracks = video_data.get('tracks')
2674 if tracks and isinstance(tracks, list):
2675 for track in tracks:
2676 if not isinstance(track, dict):
2677 continue
2678 track_kind = track.get('kind')
2679 if not track_kind or not isinstance(track_kind, compat_str):
2680 continue
2681 if track_kind.lower() not in ('captions', 'subtitles'):
2682 continue
2683 track_url = urljoin(base_url, track.get('file'))
2684 if not track_url:
2685 continue
2686 subtitles.setdefault(track.get('label') or 'en', []).append({
2687 'url': self._proto_relative_url(track_url)
2688 })
2689
2690 entry = {
2691 'id': this_video_id,
2692 'title': unescapeHTML(video_data['title'] if require_title else video_data.get('title')),
2693 'description': video_data.get('description'),
2694 'thumbnail': urljoin(base_url, self._proto_relative_url(video_data.get('image'))),
2695 'timestamp': int_or_none(video_data.get('pubdate')),
2696 'duration': float_or_none(jwplayer_data.get('duration') or video_data.get('duration')),
2697 'subtitles': subtitles,
2698 }
2699 # https://github.com/jwplayer/jwplayer/blob/master/src/js/utils/validator.js#L32
2700 if len(formats) == 1 and re.search(r'^(?:http|//).*(?:youtube\.com|youtu\.be)/.+', formats[0]['url']):
2701 entry.update({
2702 '_type': 'url_transparent',
2703 'url': formats[0]['url'],
2704 })
2705 else:
2706 self._sort_formats(formats)
2707 entry['formats'] = formats
2708 entries.append(entry)
2709 if len(entries) == 1:
2710 return entries[0]
2711 else:
2712 return self.playlist_result(entries)
2713
2714 def _parse_jwplayer_formats(self, jwplayer_sources_data, video_id=None,
2715 m3u8_id=None, mpd_id=None, rtmp_params=None, base_url=None):
2716 urls = []
2717 formats = []
2718 for source in jwplayer_sources_data:
2719 if not isinstance(source, dict):
2720 continue
2721 source_url = urljoin(
2722 base_url, self._proto_relative_url(source.get('file')))
2723 if not source_url or source_url in urls:
2724 continue
2725 urls.append(source_url)
2726 source_type = source.get('type') or ''
2727 ext = mimetype2ext(source_type) or determine_ext(source_url)
2728 if source_type == 'hls' or ext == 'm3u8':
2729 formats.extend(self._extract_m3u8_formats(
2730 source_url, video_id, 'mp4', entry_protocol='m3u8_native',
2731 m3u8_id=m3u8_id, fatal=False))
2732 elif source_type == 'dash' or ext == 'mpd':
2733 formats.extend(self._extract_mpd_formats(
2734 source_url, video_id, mpd_id=mpd_id, fatal=False))
2735 elif ext == 'smil':
2736 formats.extend(self._extract_smil_formats(
2737 source_url, video_id, fatal=False))
2738 # https://github.com/jwplayer/jwplayer/blob/master/src/js/providers/default.js#L67
2739 elif source_type.startswith('audio') or ext in (
2740 'oga', 'aac', 'mp3', 'mpeg', 'vorbis'):
2741 formats.append({
2742 'url': source_url,
2743 'vcodec': 'none',
2744 'ext': ext,
2745 })
2746 else:
2747 height = int_or_none(source.get('height'))
2748 if height is None:
2749 # Often no height is provided but there is a label in
2750 # format like "1080p", "720p SD", or 1080.
2751 height = int_or_none(self._search_regex(
2752 r'^(\d{3,4})[pP]?(?:\b|$)', compat_str(source.get('label') or ''),
2753 'height', default=None))
2754 a_format = {
2755 'url': source_url,
2756 'width': int_or_none(source.get('width')),
2757 'height': height,
2758 'tbr': int_or_none(source.get('bitrate')),
2759 'ext': ext,
2760 }
2761 if source_url.startswith('rtmp'):
2762 a_format['ext'] = 'flv'
2763 # See com/longtailvideo/jwplayer/media/RTMPMediaProvider.as
2764 # of jwplayer.flash.swf
2765 rtmp_url_parts = re.split(
2766 r'((?:mp4|mp3|flv):)', source_url, 1)
2767 if len(rtmp_url_parts) == 3:
2768 rtmp_url, prefix, play_path = rtmp_url_parts
2769 a_format.update({
2770 'url': rtmp_url,
2771 'play_path': prefix + play_path,
2772 })
2773 if rtmp_params:
2774 a_format.update(rtmp_params)
2775 formats.append(a_format)
2776 return formats
2777
2778 def _live_title(self, name):
2779 """ Generate the title for a live video """
2780 now = datetime.datetime.now()
2781 now_str = now.strftime('%Y-%m-%d %H:%M')
2782 return name + ' ' + now_str
2783
2784 def _int(self, v, name, fatal=False, **kwargs):
2785 res = int_or_none(v, **kwargs)
2786 if 'get_attr' in kwargs:
2787 print(getattr(v, kwargs['get_attr']))
2788 if res is None:
2789 msg = 'Failed to extract %s: Could not parse value %r' % (name, v)
2790 if fatal:
2791 raise ExtractorError(msg)
2792 else:
2793 self._downloader.report_warning(msg)
2794 return res
2795
2796 def _float(self, v, name, fatal=False, **kwargs):
2797 res = float_or_none(v, **kwargs)
2798 if res is None:
2799 msg = 'Failed to extract %s: Could not parse value %r' % (name, v)
2800 if fatal:
2801 raise ExtractorError(msg)
2802 else:
2803 self._downloader.report_warning(msg)
2804 return res
2805
2806 def _set_cookie(self, domain, name, value, expire_time=None, port=None,
2807 path='/', secure=False, discard=False, rest={}, **kwargs):
2808 cookie = compat_cookiejar.Cookie(
2809 0, name, value, port, port is not None, domain, True,
2810 domain.startswith('.'), path, True, secure, expire_time,
2811 discard, None, None, rest)
2812 self._downloader.cookiejar.set_cookie(cookie)
2813
2814 def _get_cookies(self, url):
2815 """ Return a compat_cookies.SimpleCookie with the cookies for the url """
2816 req = sanitized_Request(url)
2817 self._downloader.cookiejar.add_cookie_header(req)
2818 return compat_cookies.SimpleCookie(req.get_header('Cookie'))
2819
2820 def get_testcases(self, include_onlymatching=False):
2821 t = getattr(self, '_TEST', None)
2822 if t:
2823 assert not hasattr(self, '_TESTS'), \
2824 '%s has _TEST and _TESTS' % type(self).__name__
2825 tests = [t]
2826 else:
2827 tests = getattr(self, '_TESTS', [])
2828 for t in tests:
2829 if not include_onlymatching and t.get('only_matching', False):
2830 continue
2831 t['name'] = type(self).__name__[:-len('IE')]
2832 yield t
2833
2834 def is_suitable(self, age_limit):
2835 """ Test whether the extractor is generally suitable for the given
2836 age limit (i.e. pornographic sites are not, all others usually are) """
2837
2838 any_restricted = False
2839 for tc in self.get_testcases(include_onlymatching=False):
2840 if tc.get('playlist', []):
2841 tc = tc['playlist'][0]
2842 is_restricted = age_restricted(
2843 tc.get('info_dict', {}).get('age_limit'), age_limit)
2844 if not is_restricted:
2845 return True
2846 any_restricted = any_restricted or is_restricted
2847 return not any_restricted
2848
2849 def extract_subtitles(self, *args, **kwargs):
2850 if (self._downloader.params.get('writesubtitles', False) or
2851 self._downloader.params.get('listsubtitles')):
2852 return self._get_subtitles(*args, **kwargs)
2853 return {}
2854
2855 def _get_subtitles(self, *args, **kwargs):
2856 raise NotImplementedError('This method must be implemented by subclasses')
2857
2858 @staticmethod
2859 def _merge_subtitle_items(subtitle_list1, subtitle_list2):
2860 """ Merge subtitle items for one language. Items with duplicated URLs
2861 will be dropped. """
2862 list1_urls = set([item['url'] for item in subtitle_list1])
2863 ret = list(subtitle_list1)
2864 ret.extend([item for item in subtitle_list2 if item['url'] not in list1_urls])
2865 return ret
2866
2867 @classmethod
2868 def _merge_subtitles(cls, subtitle_dict1, subtitle_dict2):
2869 """ Merge two subtitle dictionaries, language by language. """
2870 ret = dict(subtitle_dict1)
2871 for lang in subtitle_dict2:
2872 ret[lang] = cls._merge_subtitle_items(subtitle_dict1.get(lang, []), subtitle_dict2[lang])
2873 return ret
2874
2875 def extract_automatic_captions(self, *args, **kwargs):
2876 if (self._downloader.params.get('writeautomaticsub', False) or
2877 self._downloader.params.get('listsubtitles')):
2878 return self._get_automatic_captions(*args, **kwargs)
2879 return {}
2880
2881 def _get_automatic_captions(self, *args, **kwargs):
2882 raise NotImplementedError('This method must be implemented by subclasses')
2883
2884 def mark_watched(self, *args, **kwargs):
2885 if (self._downloader.params.get('mark_watched', False) and
2886 (self._get_login_info()[0] is not None or
2887 self._downloader.params.get('cookiefile') is not None)):
2888 self._mark_watched(*args, **kwargs)
2889
2890 def _mark_watched(self, *args, **kwargs):
2891 raise NotImplementedError('This method must be implemented by subclasses')
2892
2893 def geo_verification_headers(self):
2894 headers = {}
2895 geo_verification_proxy = self._downloader.params.get('geo_verification_proxy')
2896 if geo_verification_proxy:
2897 headers['Ytdl-request-proxy'] = geo_verification_proxy
2898 return headers
2899
2900 def _generic_id(self, url):
2901 return compat_urllib_parse_unquote(os.path.splitext(url.rstrip('/').split('/')[-1])[0])
2902
2903 def _generic_title(self, url):
2904 return compat_urllib_parse_unquote(os.path.splitext(url_basename(url))[0])
2905
2906
2907 class SearchInfoExtractor(InfoExtractor):
2908 """
2909 Base class for paged search queries extractors.
2910 They accept URLs in the format _SEARCH_KEY(|all|[0-9]):{query}
2911 Instances should define _SEARCH_KEY and _MAX_RESULTS.
2912 """
2913
2914 @classmethod
2915 def _make_valid_url(cls):
2916 return r'%s(?P<prefix>|[1-9][0-9]*|all):(?P<query>[\s\S]+)' % cls._SEARCH_KEY
2917
2918 @classmethod
2919 def suitable(cls, url):
2920 return re.match(cls._make_valid_url(), url) is not None
2921
2922 def _real_extract(self, query):
2923 mobj = re.match(self._make_valid_url(), query)
2924 if mobj is None:
2925 raise ExtractorError('Invalid search query "%s"' % query)
2926
2927 prefix = mobj.group('prefix')
2928 query = mobj.group('query')
2929 if prefix == '':
2930 return self._get_n_results(query, 1)
2931 elif prefix == 'all':
2932 return self._get_n_results(query, self._MAX_RESULTS)
2933 else:
2934 n = int(prefix)
2935 if n <= 0:
2936 raise ExtractorError('invalid download number %s for query "%s"' % (n, query))
2937 elif n > self._MAX_RESULTS:
2938 self._downloader.report_warning('%s returns max %i results (you requested %i)' % (self._SEARCH_KEY, self._MAX_RESULTS, n))
2939 n = self._MAX_RESULTS
2940 return self._get_n_results(query, n)
2941
2942 def _get_n_results(self, query, n):
2943 """Get a specified number of results for a query"""
2944 raise NotImplementedError('This method must be implemented by subclasses')
2945
2946 @property
2947 def SEARCH_KEY(self):
2948 return self._SEARCH_KEY