]> jfr.im git - yt-dlp.git/blame_incremental - youtube_dlc/extractor/youtube.py
[version] update
[yt-dlp.git] / youtube_dlc / extractor / youtube.py
... / ...
CommitLineData
1# coding: utf-8
2
3from __future__ import unicode_literals
4
5
6import itertools
7import json
8import os.path
9import random
10import re
11import time
12import traceback
13
14from .common import InfoExtractor, SearchInfoExtractor
15from ..jsinterp import JSInterpreter
16from ..swfinterp import SWFInterpreter
17from ..compat import (
18 compat_chr,
19 compat_HTTPError,
20 compat_kwargs,
21 compat_parse_qs,
22 compat_urllib_parse_unquote,
23 compat_urllib_parse_unquote_plus,
24 compat_urllib_parse_urlencode,
25 compat_urllib_parse_urlparse,
26 compat_urlparse,
27 compat_str,
28)
29from ..utils import (
30 bool_or_none,
31 clean_html,
32 error_to_compat_str,
33 ExtractorError,
34 float_or_none,
35 get_element_by_id,
36 int_or_none,
37 mimetype2ext,
38 parse_codecs,
39 parse_count,
40 parse_duration,
41 remove_quotes,
42 remove_start,
43 smuggle_url,
44 str_or_none,
45 str_to_int,
46 try_get,
47 unescapeHTML,
48 unified_strdate,
49 unsmuggle_url,
50 update_url_query,
51 uppercase_escape,
52 url_or_none,
53 urlencode_postdata,
54 urljoin,
55)
56
57
58class YoutubeBaseInfoExtractor(InfoExtractor):
59 """Provide base functions for Youtube extractors"""
60 _LOGIN_URL = 'https://accounts.google.com/ServiceLogin'
61 _TWOFACTOR_URL = 'https://accounts.google.com/signin/challenge'
62
63 _LOOKUP_URL = 'https://accounts.google.com/_/signin/sl/lookup'
64 _CHALLENGE_URL = 'https://accounts.google.com/_/signin/sl/challenge'
65 _TFA_URL = 'https://accounts.google.com/_/signin/challenge?hl=en&TL={0}'
66
67 _RESERVED_NAMES = (
68 r'embed|e|watch_popup|channel|c|user|playlist|watch|w|v|movies|results|shared|'
69 r'storefront|oops|index|account|reporthistory|t/terms|about|upload|signin|logout|'
70 r'feed/(?:watch_later|history|subscriptions|library|trending|recommended)')
71
72 _NETRC_MACHINE = 'youtube'
73 # If True it will raise an error if no login info is provided
74 _LOGIN_REQUIRED = False
75
76 _PLAYLIST_ID_RE = r'(?:(?:PL|LL|EC|UU|FL|RD|UL|TL|PU|OLAK5uy_)[0-9A-Za-z-_]{10,}|RDMM|WL|LL|LM)'
77
78 def _set_language(self):
79 self._set_cookie(
80 '.youtube.com', 'PREF', 'f1=50000000&f6=8&hl=en',
81 # YouTube sets the expire time to about two months
82 expire_time=time.time() + 2 * 30 * 24 * 3600)
83
84 def _ids_to_results(self, ids):
85 return [
86 self.url_result(vid_id, 'Youtube', video_id=vid_id)
87 for vid_id in ids]
88
89 def _login(self):
90 """
91 Attempt to log in to YouTube.
92 True is returned if successful or skipped.
93 False is returned if login failed.
94
95 If _LOGIN_REQUIRED is set and no authentication was provided, an error is raised.
96 """
97 username, password = self._get_login_info()
98 # No authentication to be performed
99 if username is None:
100 if self._LOGIN_REQUIRED and self._downloader.params.get('cookiefile') is None:
101 raise ExtractorError('No login info available, needed for using %s.' % self.IE_NAME, expected=True)
102 if self._downloader.params.get('cookiefile') and False: # TODO remove 'and False' later - too many people using outdated cookies and open issues, remind them.
103 self.to_screen('[Cookies] Reminder - Make sure to always use up to date cookies!')
104 return True
105
106 login_page = self._download_webpage(
107 self._LOGIN_URL, None,
108 note='Downloading login page',
109 errnote='unable to fetch login page', fatal=False)
110 if login_page is False:
111 return
112
113 login_form = self._hidden_inputs(login_page)
114
115 def req(url, f_req, note, errnote):
116 data = login_form.copy()
117 data.update({
118 'pstMsg': 1,
119 'checkConnection': 'youtube',
120 'checkedDomains': 'youtube',
121 'hl': 'en',
122 'deviceinfo': '[null,null,null,[],null,"US",null,null,[],"GlifWebSignIn",null,[null,null,[]]]',
123 'f.req': json.dumps(f_req),
124 'flowName': 'GlifWebSignIn',
125 'flowEntry': 'ServiceLogin',
126 # TODO: reverse actual botguard identifier generation algo
127 'bgRequest': '["identifier",""]',
128 })
129 return self._download_json(
130 url, None, note=note, errnote=errnote,
131 transform_source=lambda s: re.sub(r'^[^[]*', '', s),
132 fatal=False,
133 data=urlencode_postdata(data), headers={
134 'Content-Type': 'application/x-www-form-urlencoded;charset=utf-8',
135 'Google-Accounts-XSRF': 1,
136 })
137
138 def warn(message):
139 self._downloader.report_warning(message)
140
141 lookup_req = [
142 username,
143 None, [], None, 'US', None, None, 2, False, True,
144 [
145 None, None,
146 [2, 1, None, 1,
147 'https://accounts.google.com/ServiceLogin?passive=true&continue=https%3A%2F%2Fwww.youtube.com%2Fsignin%3Fnext%3D%252F%26action_handle_signin%3Dtrue%26hl%3Den%26app%3Ddesktop%26feature%3Dsign_in_button&hl=en&service=youtube&uilel=3&requestPath=%2FServiceLogin&Page=PasswordSeparationSignIn',
148 None, [], 4],
149 1, [None, None, []], None, None, None, True
150 ],
151 username,
152 ]
153
154 lookup_results = req(
155 self._LOOKUP_URL, lookup_req,
156 'Looking up account info', 'Unable to look up account info')
157
158 if lookup_results is False:
159 return False
160
161 user_hash = try_get(lookup_results, lambda x: x[0][2], compat_str)
162 if not user_hash:
163 warn('Unable to extract user hash')
164 return False
165
166 challenge_req = [
167 user_hash,
168 None, 1, None, [1, None, None, None, [password, None, True]],
169 [
170 None, None, [2, 1, None, 1, 'https://accounts.google.com/ServiceLogin?passive=true&continue=https%3A%2F%2Fwww.youtube.com%2Fsignin%3Fnext%3D%252F%26action_handle_signin%3Dtrue%26hl%3Den%26app%3Ddesktop%26feature%3Dsign_in_button&hl=en&service=youtube&uilel=3&requestPath=%2FServiceLogin&Page=PasswordSeparationSignIn', None, [], 4],
171 1, [None, None, []], None, None, None, True
172 ]]
173
174 challenge_results = req(
175 self._CHALLENGE_URL, challenge_req,
176 'Logging in', 'Unable to log in')
177
178 if challenge_results is False:
179 return
180
181 login_res = try_get(challenge_results, lambda x: x[0][5], list)
182 if login_res:
183 login_msg = try_get(login_res, lambda x: x[5], compat_str)
184 warn(
185 'Unable to login: %s' % 'Invalid password'
186 if login_msg == 'INCORRECT_ANSWER_ENTERED' else login_msg)
187 return False
188
189 res = try_get(challenge_results, lambda x: x[0][-1], list)
190 if not res:
191 warn('Unable to extract result entry')
192 return False
193
194 login_challenge = try_get(res, lambda x: x[0][0], list)
195 if login_challenge:
196 challenge_str = try_get(login_challenge, lambda x: x[2], compat_str)
197 if challenge_str == 'TWO_STEP_VERIFICATION':
198 # SEND_SUCCESS - TFA code has been successfully sent to phone
199 # QUOTA_EXCEEDED - reached the limit of TFA codes
200 status = try_get(login_challenge, lambda x: x[5], compat_str)
201 if status == 'QUOTA_EXCEEDED':
202 warn('Exceeded the limit of TFA codes, try later')
203 return False
204
205 tl = try_get(challenge_results, lambda x: x[1][2], compat_str)
206 if not tl:
207 warn('Unable to extract TL')
208 return False
209
210 tfa_code = self._get_tfa_info('2-step verification code')
211
212 if not tfa_code:
213 warn(
214 'Two-factor authentication required. Provide it either interactively or with --twofactor <code>'
215 '(Note that only TOTP (Google Authenticator App) codes work at this time.)')
216 return False
217
218 tfa_code = remove_start(tfa_code, 'G-')
219
220 tfa_req = [
221 user_hash, None, 2, None,
222 [
223 9, None, None, None, None, None, None, None,
224 [None, tfa_code, True, 2]
225 ]]
226
227 tfa_results = req(
228 self._TFA_URL.format(tl), tfa_req,
229 'Submitting TFA code', 'Unable to submit TFA code')
230
231 if tfa_results is False:
232 return False
233
234 tfa_res = try_get(tfa_results, lambda x: x[0][5], list)
235 if tfa_res:
236 tfa_msg = try_get(tfa_res, lambda x: x[5], compat_str)
237 warn(
238 'Unable to finish TFA: %s' % 'Invalid TFA code'
239 if tfa_msg == 'INCORRECT_ANSWER_ENTERED' else tfa_msg)
240 return False
241
242 check_cookie_url = try_get(
243 tfa_results, lambda x: x[0][-1][2], compat_str)
244 else:
245 CHALLENGES = {
246 'LOGIN_CHALLENGE': "This device isn't recognized. For your security, Google wants to make sure it's really you.",
247 'USERNAME_RECOVERY': 'Please provide additional information to aid in the recovery process.',
248 'REAUTH': "There is something unusual about your activity. For your security, Google wants to make sure it's really you.",
249 }
250 challenge = CHALLENGES.get(
251 challenge_str,
252 '%s returned error %s.' % (self.IE_NAME, challenge_str))
253 warn('%s\nGo to https://accounts.google.com/, login and solve a challenge.' % challenge)
254 return False
255 else:
256 check_cookie_url = try_get(res, lambda x: x[2], compat_str)
257
258 if not check_cookie_url:
259 warn('Unable to extract CheckCookie URL')
260 return False
261
262 check_cookie_results = self._download_webpage(
263 check_cookie_url, None, 'Checking cookie', fatal=False)
264
265 if check_cookie_results is False:
266 return False
267
268 if 'https://myaccount.google.com/' not in check_cookie_results:
269 warn('Unable to log in')
270 return False
271
272 return True
273
274 def _download_webpage_handle(self, *args, **kwargs):
275 query = kwargs.get('query', {}).copy()
276 kwargs['query'] = query
277 return super(YoutubeBaseInfoExtractor, self)._download_webpage_handle(
278 *args, **compat_kwargs(kwargs))
279
280 def _get_yt_initial_data(self, video_id, webpage):
281 config = self._search_regex(
282 (r'window\["ytInitialData"\]\s*=\s*(.*?)(?<=});',
283 r'var\s+ytInitialData\s*=\s*(.*?)(?<=});'),
284 webpage, 'ytInitialData', default=None)
285 if config:
286 return self._parse_json(
287 uppercase_escape(config), video_id, fatal=False)
288
289 def _real_initialize(self):
290 if self._downloader is None:
291 return
292 self._set_language()
293 if not self._login():
294 return
295
296 _DEFAULT_API_DATA = {
297 'context': {
298 'client': {
299 'clientName': 'WEB',
300 'clientVersion': '2.20201021.03.00',
301 }
302 },
303 }
304
305 _YT_INITIAL_DATA_RE = r'(?:window\s*\[\s*["\']ytInitialData["\']\s*\]|ytInitialData)\s*=\s*({.+?})\s*;'
306 _YT_INITIAL_PLAYER_RESPONSE_RE = r'ytInitialPlayerResponse\s*=\s*({.+?})\s*;'
307 _YT_INITIAL_BOUNDARY_RE = r'(?:var\s+meta|</script|\n)'
308
309 def _call_api(self, ep, query, video_id):
310 data = self._DEFAULT_API_DATA.copy()
311 data.update(query)
312
313 response = self._download_json(
314 'https://www.youtube.com/youtubei/v1/%s' % ep, video_id=video_id,
315 note='Downloading API JSON', errnote='Unable to download API page',
316 data=json.dumps(data).encode('utf8'),
317 headers={'content-type': 'application/json'},
318 query={'key': 'AIzaSyAO_FJ2SlqU8Q4STEHLGCilw_Y9_11qcW8'})
319
320 return response
321
322 def _extract_yt_initial_data(self, video_id, webpage):
323 return self._parse_json(
324 self._search_regex(
325 (r'%s\s*%s' % (self._YT_INITIAL_DATA_RE, self._YT_INITIAL_BOUNDARY_RE),
326 self._YT_INITIAL_DATA_RE), webpage, 'yt initial data'),
327 video_id)
328
329 def _extract_ytcfg(self, video_id, webpage):
330 return self._parse_json(
331 self._search_regex(
332 r'ytcfg\.set\s*\(\s*({.+?})\s*\)\s*;', webpage, 'ytcfg',
333 default='{}'), video_id, fatal=False)
334
335
336class YoutubeIE(YoutubeBaseInfoExtractor):
337 IE_DESC = 'YouTube.com'
338 _VALID_URL = r"""(?x)^
339 (
340 (?:https?://|//) # http(s):// or protocol-independent URL
341 (?:(?:(?:(?:\w+\.)?[yY][oO][uU][tT][uU][bB][eE](?:-nocookie|kids)?\.com/|
342 (?:www\.)?deturl\.com/www\.youtube\.com/|
343 (?:www\.)?pwnyoutube\.com/|
344 (?:www\.)?hooktube\.com/|
345 (?:www\.)?yourepeat\.com/|
346 tube\.majestyc\.net/|
347 # Invidious instances taken from https://github.com/omarroth/invidious/wiki/Invidious-Instances
348 (?:(?:www|dev)\.)?invidio\.us/|
349 (?:(?:www|no)\.)?invidiou\.sh/|
350 (?:(?:www|fi)\.)?invidious\.snopyta\.org/|
351 (?:www\.)?invidious\.kabi\.tk/|
352 (?:www\.)?invidious\.13ad\.de/|
353 (?:www\.)?invidious\.mastodon\.host/|
354 (?:www\.)?invidious\.zapashcanon\.fr/|
355 (?:www\.)?invidious\.kavin\.rocks/|
356 (?:www\.)?invidious\.tube/|
357 (?:www\.)?invidiou\.site/|
358 (?:www\.)?invidious\.site/|
359 (?:www\.)?invidious\.xyz/|
360 (?:www\.)?invidious\.nixnet\.xyz/|
361 (?:www\.)?invidious\.drycat\.fr/|
362 (?:www\.)?tube\.poal\.co/|
363 (?:www\.)?tube\.connect\.cafe/|
364 (?:www\.)?vid\.wxzm\.sx/|
365 (?:www\.)?vid\.mint\.lgbt/|
366 (?:www\.)?yewtu\.be/|
367 (?:www\.)?yt\.elukerio\.org/|
368 (?:www\.)?yt\.lelux\.fi/|
369 (?:www\.)?invidious\.ggc-project\.de/|
370 (?:www\.)?yt\.maisputain\.ovh/|
371 (?:www\.)?invidious\.13ad\.de/|
372 (?:www\.)?invidious\.toot\.koeln/|
373 (?:www\.)?invidious\.fdn\.fr/|
374 (?:www\.)?watch\.nettohikari\.com/|
375 (?:www\.)?kgg2m7yk5aybusll\.onion/|
376 (?:www\.)?qklhadlycap4cnod\.onion/|
377 (?:www\.)?axqzx4s6s54s32yentfqojs3x5i7faxza6xo3ehd4bzzsg2ii4fv2iid\.onion/|
378 (?:www\.)?c7hqkpkpemu6e7emz5b4vyz7idjgdvgaaa3dyimmeojqbgpea3xqjoid\.onion/|
379 (?:www\.)?fz253lmuao3strwbfbmx46yu7acac2jz27iwtorgmbqlkurlclmancad\.onion/|
380 (?:www\.)?invidious\.l4qlywnpwqsluw65ts7md3khrivpirse744un3x7mlskqauz5pyuzgqd\.onion/|
381 (?:www\.)?owxfohz4kjyv25fvlqilyxast7inivgiktls3th44jhk3ej3i7ya\.b32\.i2p/|
382 (?:www\.)?4l2dgddgsrkf2ous66i6seeyi6etzfgrue332grh2n7madpwopotugyd\.onion/|
383 youtube\.googleapis\.com/) # the various hostnames, with wildcard subdomains
384 (?:.*?\#/)? # handle anchor (#/) redirect urls
385 (?: # the various things that can precede the ID:
386 (?:(?:v|embed|e)/(?!videoseries)) # v/ or embed/ or e/
387 |(?: # or the v= param in all its forms
388 (?:(?:watch|movie)(?:_popup)?(?:\.php)?/?)? # preceding watch(_popup|.php) or nothing (like /?v=xxxx)
389 (?:\?|\#!?) # the params delimiter ? or # or #!
390 (?:.*?[&;])?? # any other preceding param (like /?s=tuff&v=xxxx or ?s=tuff&amp;v=V36LpHqtcDY)
391 v=
392 )
393 ))
394 |(?:
395 youtu\.be| # just youtu.be/xxxx
396 vid\.plus| # or vid.plus/xxxx
397 zwearz\.com/watch| # or zwearz.com/watch/xxxx
398 )/
399 |(?:www\.)?cleanvideosearch\.com/media/action/yt/watch\?videoId=
400 )
401 )? # all until now is optional -> you can pass the naked ID
402 (?P<id>[0-9A-Za-z_-]{11}) # here is it! the YouTube video ID
403 (?!.*?\blist=
404 (?:
405 %(playlist_id)s| # combined list/video URLs are handled by the playlist IE
406 WL # WL are handled by the watch later IE
407 )
408 )
409 (?(1).+)? # if we found the ID, everything can follow
410 $""" % {'playlist_id': YoutubeBaseInfoExtractor._PLAYLIST_ID_RE}
411 _NEXT_URL_RE = r'[\?&]next_url=([^&]+)'
412 _PLAYER_INFO_RE = (
413 r'/(?P<id>[a-zA-Z0-9_-]{8,})/player_ias\.vflset(?:/[a-zA-Z]{2,3}_[a-zA-Z]{2,3})?/base\.(?P<ext>[a-z]+)$',
414 r'\b(?P<id>vfl[a-zA-Z0-9_-]+)\b.*?\.(?P<ext>[a-z]+)$',
415 )
416 _formats = {
417 '5': {'ext': 'flv', 'width': 400, 'height': 240, 'acodec': 'mp3', 'abr': 64, 'vcodec': 'h263'},
418 '6': {'ext': 'flv', 'width': 450, 'height': 270, 'acodec': 'mp3', 'abr': 64, 'vcodec': 'h263'},
419 '13': {'ext': '3gp', 'acodec': 'aac', 'vcodec': 'mp4v'},
420 '17': {'ext': '3gp', 'width': 176, 'height': 144, 'acodec': 'aac', 'abr': 24, 'vcodec': 'mp4v'},
421 '18': {'ext': 'mp4', 'width': 640, 'height': 360, 'acodec': 'aac', 'abr': 96, 'vcodec': 'h264'},
422 '22': {'ext': 'mp4', 'width': 1280, 'height': 720, 'acodec': 'aac', 'abr': 192, 'vcodec': 'h264'},
423 '34': {'ext': 'flv', 'width': 640, 'height': 360, 'acodec': 'aac', 'abr': 128, 'vcodec': 'h264'},
424 '35': {'ext': 'flv', 'width': 854, 'height': 480, 'acodec': 'aac', 'abr': 128, 'vcodec': 'h264'},
425 # itag 36 videos are either 320x180 (BaW_jenozKc) or 320x240 (__2ABJjxzNo), abr varies as well
426 '36': {'ext': '3gp', 'width': 320, 'acodec': 'aac', 'vcodec': 'mp4v'},
427 '37': {'ext': 'mp4', 'width': 1920, 'height': 1080, 'acodec': 'aac', 'abr': 192, 'vcodec': 'h264'},
428 '38': {'ext': 'mp4', 'width': 4096, 'height': 3072, 'acodec': 'aac', 'abr': 192, 'vcodec': 'h264'},
429 '43': {'ext': 'webm', 'width': 640, 'height': 360, 'acodec': 'vorbis', 'abr': 128, 'vcodec': 'vp8'},
430 '44': {'ext': 'webm', 'width': 854, 'height': 480, 'acodec': 'vorbis', 'abr': 128, 'vcodec': 'vp8'},
431 '45': {'ext': 'webm', 'width': 1280, 'height': 720, 'acodec': 'vorbis', 'abr': 192, 'vcodec': 'vp8'},
432 '46': {'ext': 'webm', 'width': 1920, 'height': 1080, 'acodec': 'vorbis', 'abr': 192, 'vcodec': 'vp8'},
433 '59': {'ext': 'mp4', 'width': 854, 'height': 480, 'acodec': 'aac', 'abr': 128, 'vcodec': 'h264'},
434 '78': {'ext': 'mp4', 'width': 854, 'height': 480, 'acodec': 'aac', 'abr': 128, 'vcodec': 'h264'},
435
436
437 # 3D videos
438 '82': {'ext': 'mp4', 'height': 360, 'format_note': '3D', 'acodec': 'aac', 'abr': 128, 'vcodec': 'h264', 'preference': -20},
439 '83': {'ext': 'mp4', 'height': 480, 'format_note': '3D', 'acodec': 'aac', 'abr': 128, 'vcodec': 'h264', 'preference': -20},
440 '84': {'ext': 'mp4', 'height': 720, 'format_note': '3D', 'acodec': 'aac', 'abr': 192, 'vcodec': 'h264', 'preference': -20},
441 '85': {'ext': 'mp4', 'height': 1080, 'format_note': '3D', 'acodec': 'aac', 'abr': 192, 'vcodec': 'h264', 'preference': -20},
442 '100': {'ext': 'webm', 'height': 360, 'format_note': '3D', 'acodec': 'vorbis', 'abr': 128, 'vcodec': 'vp8', 'preference': -20},
443 '101': {'ext': 'webm', 'height': 480, 'format_note': '3D', 'acodec': 'vorbis', 'abr': 192, 'vcodec': 'vp8', 'preference': -20},
444 '102': {'ext': 'webm', 'height': 720, 'format_note': '3D', 'acodec': 'vorbis', 'abr': 192, 'vcodec': 'vp8', 'preference': -20},
445
446 # Apple HTTP Live Streaming
447 '91': {'ext': 'mp4', 'height': 144, 'format_note': 'HLS', 'acodec': 'aac', 'abr': 48, 'vcodec': 'h264', 'preference': -10},
448 '92': {'ext': 'mp4', 'height': 240, 'format_note': 'HLS', 'acodec': 'aac', 'abr': 48, 'vcodec': 'h264', 'preference': -10},
449 '93': {'ext': 'mp4', 'height': 360, 'format_note': 'HLS', 'acodec': 'aac', 'abr': 128, 'vcodec': 'h264', 'preference': -10},
450 '94': {'ext': 'mp4', 'height': 480, 'format_note': 'HLS', 'acodec': 'aac', 'abr': 128, 'vcodec': 'h264', 'preference': -10},
451 '95': {'ext': 'mp4', 'height': 720, 'format_note': 'HLS', 'acodec': 'aac', 'abr': 256, 'vcodec': 'h264', 'preference': -10},
452 '96': {'ext': 'mp4', 'height': 1080, 'format_note': 'HLS', 'acodec': 'aac', 'abr': 256, 'vcodec': 'h264', 'preference': -10},
453 '132': {'ext': 'mp4', 'height': 240, 'format_note': 'HLS', 'acodec': 'aac', 'abr': 48, 'vcodec': 'h264', 'preference': -10},
454 '151': {'ext': 'mp4', 'height': 72, 'format_note': 'HLS', 'acodec': 'aac', 'abr': 24, 'vcodec': 'h264', 'preference': -10},
455
456 # DASH mp4 video
457 '133': {'ext': 'mp4', 'height': 240, 'format_note': 'DASH video', 'vcodec': 'h264'},
458 '134': {'ext': 'mp4', 'height': 360, 'format_note': 'DASH video', 'vcodec': 'h264'},
459 '135': {'ext': 'mp4', 'height': 480, 'format_note': 'DASH video', 'vcodec': 'h264'},
460 '136': {'ext': 'mp4', 'height': 720, 'format_note': 'DASH video', 'vcodec': 'h264'},
461 '137': {'ext': 'mp4', 'height': 1080, 'format_note': 'DASH video', 'vcodec': 'h264'},
462 '138': {'ext': 'mp4', 'format_note': 'DASH video', 'vcodec': 'h264'}, # Height can vary (https://github.com/ytdl-org/youtube-dl/issues/4559)
463 '160': {'ext': 'mp4', 'height': 144, 'format_note': 'DASH video', 'vcodec': 'h264'},
464 '212': {'ext': 'mp4', 'height': 480, 'format_note': 'DASH video', 'vcodec': 'h264'},
465 '264': {'ext': 'mp4', 'height': 1440, 'format_note': 'DASH video', 'vcodec': 'h264'},
466 '298': {'ext': 'mp4', 'height': 720, 'format_note': 'DASH video', 'vcodec': 'h264', 'fps': 60},
467 '299': {'ext': 'mp4', 'height': 1080, 'format_note': 'DASH video', 'vcodec': 'h264', 'fps': 60},
468 '266': {'ext': 'mp4', 'height': 2160, 'format_note': 'DASH video', 'vcodec': 'h264'},
469
470 # Dash mp4 audio
471 '139': {'ext': 'm4a', 'format_note': 'DASH audio', 'acodec': 'aac', 'abr': 48, 'container': 'm4a_dash'},
472 '140': {'ext': 'm4a', 'format_note': 'DASH audio', 'acodec': 'aac', 'abr': 128, 'container': 'm4a_dash'},
473 '141': {'ext': 'm4a', 'format_note': 'DASH audio', 'acodec': 'aac', 'abr': 256, 'container': 'm4a_dash'},
474 '256': {'ext': 'm4a', 'format_note': 'DASH audio', 'acodec': 'aac', 'container': 'm4a_dash'},
475 '258': {'ext': 'm4a', 'format_note': 'DASH audio', 'acodec': 'aac', 'container': 'm4a_dash'},
476 '325': {'ext': 'm4a', 'format_note': 'DASH audio', 'acodec': 'dtse', 'container': 'm4a_dash'},
477 '328': {'ext': 'm4a', 'format_note': 'DASH audio', 'acodec': 'ec-3', 'container': 'm4a_dash'},
478
479 # Dash webm
480 '167': {'ext': 'webm', 'height': 360, 'width': 640, 'format_note': 'DASH video', 'container': 'webm', 'vcodec': 'vp8'},
481 '168': {'ext': 'webm', 'height': 480, 'width': 854, 'format_note': 'DASH video', 'container': 'webm', 'vcodec': 'vp8'},
482 '169': {'ext': 'webm', 'height': 720, 'width': 1280, 'format_note': 'DASH video', 'container': 'webm', 'vcodec': 'vp8'},
483 '170': {'ext': 'webm', 'height': 1080, 'width': 1920, 'format_note': 'DASH video', 'container': 'webm', 'vcodec': 'vp8'},
484 '218': {'ext': 'webm', 'height': 480, 'width': 854, 'format_note': 'DASH video', 'container': 'webm', 'vcodec': 'vp8'},
485 '219': {'ext': 'webm', 'height': 480, 'width': 854, 'format_note': 'DASH video', 'container': 'webm', 'vcodec': 'vp8'},
486 '278': {'ext': 'webm', 'height': 144, 'format_note': 'DASH video', 'container': 'webm', 'vcodec': 'vp9'},
487 '242': {'ext': 'webm', 'height': 240, 'format_note': 'DASH video', 'vcodec': 'vp9'},
488 '243': {'ext': 'webm', 'height': 360, 'format_note': 'DASH video', 'vcodec': 'vp9'},
489 '244': {'ext': 'webm', 'height': 480, 'format_note': 'DASH video', 'vcodec': 'vp9'},
490 '245': {'ext': 'webm', 'height': 480, 'format_note': 'DASH video', 'vcodec': 'vp9'},
491 '246': {'ext': 'webm', 'height': 480, 'format_note': 'DASH video', 'vcodec': 'vp9'},
492 '247': {'ext': 'webm', 'height': 720, 'format_note': 'DASH video', 'vcodec': 'vp9'},
493 '248': {'ext': 'webm', 'height': 1080, 'format_note': 'DASH video', 'vcodec': 'vp9'},
494 '271': {'ext': 'webm', 'height': 1440, 'format_note': 'DASH video', 'vcodec': 'vp9'},
495 # itag 272 videos are either 3840x2160 (e.g. RtoitU2A-3E) or 7680x4320 (sLprVF6d7Ug)
496 '272': {'ext': 'webm', 'height': 2160, 'format_note': 'DASH video', 'vcodec': 'vp9'},
497 '302': {'ext': 'webm', 'height': 720, 'format_note': 'DASH video', 'vcodec': 'vp9', 'fps': 60},
498 '303': {'ext': 'webm', 'height': 1080, 'format_note': 'DASH video', 'vcodec': 'vp9', 'fps': 60},
499 '308': {'ext': 'webm', 'height': 1440, 'format_note': 'DASH video', 'vcodec': 'vp9', 'fps': 60},
500 '313': {'ext': 'webm', 'height': 2160, 'format_note': 'DASH video', 'vcodec': 'vp9'},
501 '315': {'ext': 'webm', 'height': 2160, 'format_note': 'DASH video', 'vcodec': 'vp9', 'fps': 60},
502
503 # Dash webm audio
504 '171': {'ext': 'webm', 'acodec': 'vorbis', 'format_note': 'DASH audio', 'abr': 128},
505 '172': {'ext': 'webm', 'acodec': 'vorbis', 'format_note': 'DASH audio', 'abr': 256},
506
507 # Dash webm audio with opus inside
508 '249': {'ext': 'webm', 'format_note': 'DASH audio', 'acodec': 'opus', 'abr': 50},
509 '250': {'ext': 'webm', 'format_note': 'DASH audio', 'acodec': 'opus', 'abr': 70},
510 '251': {'ext': 'webm', 'format_note': 'DASH audio', 'acodec': 'opus', 'abr': 160},
511
512 # RTMP (unnamed)
513 '_rtmp': {'protocol': 'rtmp'},
514
515 # av01 video only formats sometimes served with "unknown" codecs
516 '394': {'acodec': 'none', 'vcodec': 'av01.0.05M.08'},
517 '395': {'acodec': 'none', 'vcodec': 'av01.0.05M.08'},
518 '396': {'acodec': 'none', 'vcodec': 'av01.0.05M.08'},
519 '397': {'acodec': 'none', 'vcodec': 'av01.0.05M.08'},
520 }
521 _SUBTITLE_FORMATS = ('json3', 'srv1', 'srv2', 'srv3', 'ttml', 'vtt')
522
523 _GEO_BYPASS = False
524
525 IE_NAME = 'youtube'
526 _TESTS = [
527 {
528 'url': 'https://www.youtube.com/watch?v=BaW_jenozKc&t=1s&end=9',
529 'info_dict': {
530 'id': 'BaW_jenozKc',
531 'ext': 'mp4',
532 'title': 'youtube-dl test video "\'/\\ä↭𝕐',
533 'uploader': 'Philipp Hagemeister',
534 'uploader_id': 'phihag',
535 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/phihag',
536 'channel_id': 'UCLqxVugv74EIW3VWh2NOa3Q',
537 'channel_url': r're:https?://(?:www\.)?youtube\.com/channel/UCLqxVugv74EIW3VWh2NOa3Q',
538 'upload_date': '20121002',
539 'description': 'test chars: "\'/\\ä↭𝕐\ntest URL: https://github.com/rg3/youtube-dl/issues/1892\n\nThis is a test video for youtube-dl.\n\nFor more information, contact phihag@phihag.de .',
540 'categories': ['Science & Technology'],
541 'tags': ['youtube-dl'],
542 'duration': 10,
543 'view_count': int,
544 'like_count': int,
545 'dislike_count': int,
546 'start_time': 1,
547 'end_time': 9,
548 }
549 },
550 {
551 'url': '//www.YouTube.com/watch?v=yZIXLfi8CZQ',
552 'note': 'Embed-only video (#1746)',
553 'info_dict': {
554 'id': 'yZIXLfi8CZQ',
555 'ext': 'mp4',
556 'upload_date': '20120608',
557 'title': 'Principal Sexually Assaults A Teacher - Episode 117 - 8th June 2012',
558 'description': 'md5:09b78bd971f1e3e289601dfba15ca4f7',
559 'uploader': 'SET India',
560 'uploader_id': 'setindia',
561 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/setindia',
562 'age_limit': 18,
563 }
564 },
565 {
566 'url': 'https://www.youtube.com/watch?v=BaW_jenozKc&v=yZIXLfi8CZQ',
567 'note': 'Use the first video ID in the URL',
568 'info_dict': {
569 'id': 'BaW_jenozKc',
570 'ext': 'mp4',
571 'title': 'youtube-dl test video "\'/\\ä↭𝕐',
572 'uploader': 'Philipp Hagemeister',
573 'uploader_id': 'phihag',
574 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/phihag',
575 'upload_date': '20121002',
576 'description': 'test chars: "\'/\\ä↭𝕐\ntest URL: https://github.com/rg3/youtube-dl/issues/1892\n\nThis is a test video for youtube-dl.\n\nFor more information, contact phihag@phihag.de .',
577 'categories': ['Science & Technology'],
578 'tags': ['youtube-dl'],
579 'duration': 10,
580 'view_count': int,
581 'like_count': int,
582 'dislike_count': int,
583 },
584 'params': {
585 'skip_download': True,
586 },
587 },
588 {
589 'url': 'https://www.youtube.com/watch?v=a9LDPn-MO4I',
590 'note': '256k DASH audio (format 141) via DASH manifest',
591 'info_dict': {
592 'id': 'a9LDPn-MO4I',
593 'ext': 'm4a',
594 'upload_date': '20121002',
595 'uploader_id': '8KVIDEO',
596 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/8KVIDEO',
597 'description': '',
598 'uploader': '8KVIDEO',
599 'title': 'UHDTV TEST 8K VIDEO.mp4'
600 },
601 'params': {
602 'youtube_include_dash_manifest': True,
603 'format': '141',
604 },
605 'skip': 'format 141 not served anymore',
606 },
607 # DASH manifest with encrypted signature
608 {
609 'url': 'https://www.youtube.com/watch?v=IB3lcPjvWLA',
610 'info_dict': {
611 'id': 'IB3lcPjvWLA',
612 'ext': 'm4a',
613 'title': 'Afrojack, Spree Wilson - The Spark (Official Music Video) ft. Spree Wilson',
614 'description': 'md5:8f5e2b82460520b619ccac1f509d43bf',
615 'duration': 244,
616 'uploader': 'AfrojackVEVO',
617 'uploader_id': 'AfrojackVEVO',
618 'upload_date': '20131011',
619 },
620 'params': {
621 'youtube_include_dash_manifest': True,
622 'format': '141/bestaudio[ext=m4a]',
623 },
624 },
625 # Controversy video
626 {
627 'url': 'https://www.youtube.com/watch?v=T4XJQO3qol8',
628 'info_dict': {
629 'id': 'T4XJQO3qol8',
630 'ext': 'mp4',
631 'duration': 219,
632 'upload_date': '20100909',
633 'uploader': 'Amazing Atheist',
634 'uploader_id': 'TheAmazingAtheist',
635 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/TheAmazingAtheist',
636 'title': 'Burning Everyone\'s Koran',
637 'description': 'SUBSCRIBE: http://www.youtube.com/saturninefilms\n\nEven Obama has taken a stand against freedom on this issue: http://www.huffingtonpost.com/2010/09/09/obama-gma-interview-quran_n_710282.html',
638 }
639 },
640 # Normal age-gate video (embed allowed)
641 {
642 'url': 'https://youtube.com/watch?v=HtVdAasjOgU',
643 'info_dict': {
644 'id': 'HtVdAasjOgU',
645 'ext': 'mp4',
646 'title': 'The Witcher 3: Wild Hunt - The Sword Of Destiny Trailer',
647 'description': r're:(?s).{100,}About the Game\n.*?The Witcher 3: Wild Hunt.{100,}',
648 'duration': 142,
649 'uploader': 'The Witcher',
650 'uploader_id': 'WitcherGame',
651 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/WitcherGame',
652 'upload_date': '20140605',
653 'age_limit': 18,
654 },
655 },
656 # video_info is None (https://github.com/ytdl-org/youtube-dl/issues/4421)
657 # YouTube Red ad is not captured for creator
658 {
659 'url': '__2ABJjxzNo',
660 'info_dict': {
661 'id': '__2ABJjxzNo',
662 'ext': 'mp4',
663 'duration': 266,
664 'upload_date': '20100430',
665 'uploader_id': 'deadmau5',
666 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/deadmau5',
667 'creator': 'Dada Life, deadmau5',
668 'description': 'md5:12c56784b8032162bb936a5f76d55360',
669 'uploader': 'deadmau5',
670 'title': 'Deadmau5 - Some Chords (HD)',
671 'alt_title': 'This Machine Kills Some Chords',
672 },
673 'expected_warnings': [
674 'DASH manifest missing',
675 ]
676 },
677 # Olympics (https://github.com/ytdl-org/youtube-dl/issues/4431)
678 {
679 'url': 'lqQg6PlCWgI',
680 'info_dict': {
681 'id': 'lqQg6PlCWgI',
682 'ext': 'mp4',
683 'duration': 6085,
684 'upload_date': '20150827',
685 'uploader_id': 'olympic',
686 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/olympic',
687 'description': 'HO09 - Women - GER-AUS - Hockey - 31 July 2012 - London 2012 Olympic Games',
688 'uploader': 'Olympic',
689 'title': 'Hockey - Women - GER-AUS - London 2012 Olympic Games',
690 },
691 'params': {
692 'skip_download': 'requires avconv',
693 }
694 },
695 # Non-square pixels
696 {
697 'url': 'https://www.youtube.com/watch?v=_b-2C3KPAM0',
698 'info_dict': {
699 'id': '_b-2C3KPAM0',
700 'ext': 'mp4',
701 'stretched_ratio': 16 / 9.,
702 'duration': 85,
703 'upload_date': '20110310',
704 'uploader_id': 'AllenMeow',
705 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/AllenMeow',
706 'description': 'made by Wacom from Korea | 字幕&加油添醋 by TY\'s Allen | 感謝heylisa00cavey1001同學熱情提供梗及翻譯',
707 'uploader': '孫ᄋᄅ',
708 'title': '[A-made] 變態妍字幕版 太妍 我就是這樣的人',
709 },
710 },
711 # url_encoded_fmt_stream_map is empty string
712 {
713 'url': 'qEJwOuvDf7I',
714 'info_dict': {
715 'id': 'qEJwOuvDf7I',
716 'ext': 'webm',
717 'title': 'Обсуждение судебной практики по выборам 14 сентября 2014 года в Санкт-Петербурге',
718 'description': '',
719 'upload_date': '20150404',
720 'uploader_id': 'spbelect',
721 'uploader': 'Наблюдатели Петербурга',
722 },
723 'params': {
724 'skip_download': 'requires avconv',
725 },
726 'skip': 'This live event has ended.',
727 },
728 # Extraction from multiple DASH manifests (https://github.com/ytdl-org/youtube-dl/pull/6097)
729 {
730 'url': 'https://www.youtube.com/watch?v=FIl7x6_3R5Y',
731 'info_dict': {
732 'id': 'FIl7x6_3R5Y',
733 'ext': 'webm',
734 'title': 'md5:7b81415841e02ecd4313668cde88737a',
735 'description': 'md5:116377fd2963b81ec4ce64b542173306',
736 'duration': 220,
737 'upload_date': '20150625',
738 'uploader_id': 'dorappi2000',
739 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/dorappi2000',
740 'uploader': 'dorappi2000',
741 'formats': 'mincount:31',
742 },
743 'skip': 'not actual anymore',
744 },
745 # DASH manifest with segment_list
746 {
747 'url': 'https://www.youtube.com/embed/CsmdDsKjzN8',
748 'md5': '8ce563a1d667b599d21064e982ab9e31',
749 'info_dict': {
750 'id': 'CsmdDsKjzN8',
751 'ext': 'mp4',
752 'upload_date': '20150501', # According to '<meta itemprop="datePublished"', but in other places it's 20150510
753 'uploader': 'Airtek',
754 'description': 'Retransmisión en directo de la XVIII media maratón de Zaragoza.',
755 'uploader_id': 'UCzTzUmjXxxacNnL8I3m4LnQ',
756 'title': 'Retransmisión XVIII Media maratón Zaragoza 2015',
757 },
758 'params': {
759 'youtube_include_dash_manifest': True,
760 'format': '135', # bestvideo
761 },
762 'skip': 'This live event has ended.',
763 },
764 {
765 # Multifeed videos (multiple cameras), URL is for Main Camera
766 'url': 'https://www.youtube.com/watch?v=jqWvoWXjCVs',
767 'info_dict': {
768 'id': 'jqWvoWXjCVs',
769 'title': 'teamPGP: Rocket League Noob Stream',
770 'description': 'md5:dc7872fb300e143831327f1bae3af010',
771 },
772 'playlist': [{
773 'info_dict': {
774 'id': 'jqWvoWXjCVs',
775 'ext': 'mp4',
776 'title': 'teamPGP: Rocket League Noob Stream (Main Camera)',
777 'description': 'md5:dc7872fb300e143831327f1bae3af010',
778 'duration': 7335,
779 'upload_date': '20150721',
780 'uploader': 'Beer Games Beer',
781 'uploader_id': 'beergamesbeer',
782 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/beergamesbeer',
783 'license': 'Standard YouTube License',
784 },
785 }, {
786 'info_dict': {
787 'id': '6h8e8xoXJzg',
788 'ext': 'mp4',
789 'title': 'teamPGP: Rocket League Noob Stream (kreestuh)',
790 'description': 'md5:dc7872fb300e143831327f1bae3af010',
791 'duration': 7337,
792 'upload_date': '20150721',
793 'uploader': 'Beer Games Beer',
794 'uploader_id': 'beergamesbeer',
795 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/beergamesbeer',
796 'license': 'Standard YouTube License',
797 },
798 }, {
799 'info_dict': {
800 'id': 'PUOgX5z9xZw',
801 'ext': 'mp4',
802 'title': 'teamPGP: Rocket League Noob Stream (grizzle)',
803 'description': 'md5:dc7872fb300e143831327f1bae3af010',
804 'duration': 7337,
805 'upload_date': '20150721',
806 'uploader': 'Beer Games Beer',
807 'uploader_id': 'beergamesbeer',
808 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/beergamesbeer',
809 'license': 'Standard YouTube License',
810 },
811 }, {
812 'info_dict': {
813 'id': 'teuwxikvS5k',
814 'ext': 'mp4',
815 'title': 'teamPGP: Rocket League Noob Stream (zim)',
816 'description': 'md5:dc7872fb300e143831327f1bae3af010',
817 'duration': 7334,
818 'upload_date': '20150721',
819 'uploader': 'Beer Games Beer',
820 'uploader_id': 'beergamesbeer',
821 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/beergamesbeer',
822 'license': 'Standard YouTube License',
823 },
824 }],
825 'params': {
826 'skip_download': True,
827 },
828 'skip': 'This video is not available.',
829 },
830 {
831 # Multifeed video with comma in title (see https://github.com/ytdl-org/youtube-dl/issues/8536)
832 'url': 'https://www.youtube.com/watch?v=gVfLd0zydlo',
833 'info_dict': {
834 'id': 'gVfLd0zydlo',
835 'title': 'DevConf.cz 2016 Day 2 Workshops 1 14:00 - 15:30',
836 },
837 'playlist_count': 2,
838 'skip': 'Not multifeed anymore',
839 },
840 {
841 'url': 'https://vid.plus/FlRa-iH7PGw',
842 'only_matching': True,
843 },
844 {
845 'url': 'https://zwearz.com/watch/9lWxNJF-ufM/electra-woman-dyna-girl-official-trailer-grace-helbig.html',
846 'only_matching': True,
847 },
848 {
849 # Title with JS-like syntax "};" (see https://github.com/ytdl-org/youtube-dl/issues/7468)
850 # Also tests cut-off URL expansion in video description (see
851 # https://github.com/ytdl-org/youtube-dl/issues/1892,
852 # https://github.com/ytdl-org/youtube-dl/issues/8164)
853 'url': 'https://www.youtube.com/watch?v=lsguqyKfVQg',
854 'info_dict': {
855 'id': 'lsguqyKfVQg',
856 'ext': 'mp4',
857 'title': '{dark walk}; Loki/AC/Dishonored; collab w/Elflover21',
858 'alt_title': 'Dark Walk - Position Music',
859 'description': 'md5:8085699c11dc3f597ce0410b0dcbb34a',
860 'duration': 133,
861 'upload_date': '20151119',
862 'uploader_id': 'IronSoulElf',
863 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/IronSoulElf',
864 'uploader': 'IronSoulElf',
865 'creator': 'Todd Haberman, Daniel Law Heath and Aaron Kaplan',
866 'track': 'Dark Walk - Position Music',
867 'artist': 'Todd Haberman, Daniel Law Heath and Aaron Kaplan',
868 'album': 'Position Music - Production Music Vol. 143 - Dark Walk',
869 },
870 'params': {
871 'skip_download': True,
872 },
873 },
874 {
875 # Tags with '};' (see https://github.com/ytdl-org/youtube-dl/issues/7468)
876 'url': 'https://www.youtube.com/watch?v=Ms7iBXnlUO8',
877 'only_matching': True,
878 },
879 {
880 # Video with yt:stretch=17:0
881 'url': 'https://www.youtube.com/watch?v=Q39EVAstoRM',
882 'info_dict': {
883 'id': 'Q39EVAstoRM',
884 'ext': 'mp4',
885 'title': 'Clash Of Clans#14 Dicas De Ataque Para CV 4',
886 'description': 'md5:ee18a25c350637c8faff806845bddee9',
887 'upload_date': '20151107',
888 'uploader_id': 'UCCr7TALkRbo3EtFzETQF1LA',
889 'uploader': 'CH GAMER DROID',
890 },
891 'params': {
892 'skip_download': True,
893 },
894 'skip': 'This video does not exist.',
895 },
896 {
897 # Video licensed under Creative Commons
898 'url': 'https://www.youtube.com/watch?v=M4gD1WSo5mA',
899 'info_dict': {
900 'id': 'M4gD1WSo5mA',
901 'ext': 'mp4',
902 'title': 'md5:e41008789470fc2533a3252216f1c1d1',
903 'description': 'md5:a677553cf0840649b731a3024aeff4cc',
904 'duration': 721,
905 'upload_date': '20150127',
906 'uploader_id': 'BerkmanCenter',
907 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/BerkmanCenter',
908 'uploader': 'The Berkman Klein Center for Internet & Society',
909 'license': 'Creative Commons Attribution license (reuse allowed)',
910 },
911 'params': {
912 'skip_download': True,
913 },
914 },
915 {
916 # Channel-like uploader_url
917 'url': 'https://www.youtube.com/watch?v=eQcmzGIKrzg',
918 'info_dict': {
919 'id': 'eQcmzGIKrzg',
920 'ext': 'mp4',
921 'title': 'Democratic Socialism and Foreign Policy | Bernie Sanders',
922 'description': 'md5:dda0d780d5a6e120758d1711d062a867',
923 'duration': 4060,
924 'upload_date': '20151119',
925 'uploader': 'Bernie Sanders',
926 'uploader_id': 'UCH1dpzjCEiGAt8CXkryhkZg',
927 'uploader_url': r're:https?://(?:www\.)?youtube\.com/channel/UCH1dpzjCEiGAt8CXkryhkZg',
928 'license': 'Creative Commons Attribution license (reuse allowed)',
929 },
930 'params': {
931 'skip_download': True,
932 },
933 },
934 {
935 'url': 'https://www.youtube.com/watch?feature=player_embedded&amp;amp;v=V36LpHqtcDY',
936 'only_matching': True,
937 },
938 {
939 # YouTube Red paid video (https://github.com/ytdl-org/youtube-dl/issues/10059)
940 'url': 'https://www.youtube.com/watch?v=i1Ko8UG-Tdo',
941 'only_matching': True,
942 },
943 {
944 # Rental video preview
945 'url': 'https://www.youtube.com/watch?v=yYr8q0y5Jfg',
946 'info_dict': {
947 'id': 'uGpuVWrhIzE',
948 'ext': 'mp4',
949 'title': 'Piku - Trailer',
950 'description': 'md5:c36bd60c3fd6f1954086c083c72092eb',
951 'upload_date': '20150811',
952 'uploader': 'FlixMatrix',
953 'uploader_id': 'FlixMatrixKaravan',
954 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/FlixMatrixKaravan',
955 'license': 'Standard YouTube License',
956 },
957 'params': {
958 'skip_download': True,
959 },
960 'skip': 'This video is not available.',
961 },
962 {
963 # YouTube Red video with episode data
964 'url': 'https://www.youtube.com/watch?v=iqKdEhx-dD4',
965 'info_dict': {
966 'id': 'iqKdEhx-dD4',
967 'ext': 'mp4',
968 'title': 'Isolation - Mind Field (Ep 1)',
969 'description': 'md5:46a29be4ceffa65b92d277b93f463c0f',
970 'duration': 2085,
971 'upload_date': '20170118',
972 'uploader': 'Vsauce',
973 'uploader_id': 'Vsauce',
974 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/Vsauce',
975 'series': 'Mind Field',
976 'season_number': 1,
977 'episode_number': 1,
978 },
979 'params': {
980 'skip_download': True,
981 },
982 'expected_warnings': [
983 'Skipping DASH manifest',
984 ],
985 },
986 {
987 # The following content has been identified by the YouTube community
988 # as inappropriate or offensive to some audiences.
989 'url': 'https://www.youtube.com/watch?v=6SJNVb0GnPI',
990 'info_dict': {
991 'id': '6SJNVb0GnPI',
992 'ext': 'mp4',
993 'title': 'Race Differences in Intelligence',
994 'description': 'md5:5d161533167390427a1f8ee89a1fc6f1',
995 'duration': 965,
996 'upload_date': '20140124',
997 'uploader': 'New Century Foundation',
998 'uploader_id': 'UCEJYpZGqgUob0zVVEaLhvVg',
999 'uploader_url': r're:https?://(?:www\.)?youtube\.com/channel/UCEJYpZGqgUob0zVVEaLhvVg',
1000 },
1001 'params': {
1002 'skip_download': True,
1003 },
1004 },
1005 {
1006 # itag 212
1007 'url': '1t24XAntNCY',
1008 'only_matching': True,
1009 },
1010 {
1011 # geo restricted to JP
1012 'url': 'sJL6WA-aGkQ',
1013 'only_matching': True,
1014 },
1015 {
1016 'url': 'https://invidio.us/watch?v=BaW_jenozKc',
1017 'only_matching': True,
1018 },
1019 {
1020 # DRM protected
1021 'url': 'https://www.youtube.com/watch?v=s7_qI6_mIXc',
1022 'only_matching': True,
1023 },
1024 {
1025 # Video with unsupported adaptive stream type formats
1026 'url': 'https://www.youtube.com/watch?v=Z4Vy8R84T1U',
1027 'info_dict': {
1028 'id': 'Z4Vy8R84T1U',
1029 'ext': 'mp4',
1030 'title': 'saman SMAN 53 Jakarta(Sancety) opening COFFEE4th at SMAN 53 Jakarta',
1031 'description': 'md5:d41d8cd98f00b204e9800998ecf8427e',
1032 'duration': 433,
1033 'upload_date': '20130923',
1034 'uploader': 'Amelia Putri Harwita',
1035 'uploader_id': 'UCpOxM49HJxmC1qCalXyB3_Q',
1036 'uploader_url': r're:https?://(?:www\.)?youtube\.com/channel/UCpOxM49HJxmC1qCalXyB3_Q',
1037 'formats': 'maxcount:10',
1038 },
1039 'params': {
1040 'skip_download': True,
1041 'youtube_include_dash_manifest': False,
1042 },
1043 'skip': 'not actual anymore',
1044 },
1045 {
1046 # Youtube Music Auto-generated description
1047 'url': 'https://music.youtube.com/watch?v=MgNrAu2pzNs',
1048 'info_dict': {
1049 'id': 'MgNrAu2pzNs',
1050 'ext': 'mp4',
1051 'title': 'Voyeur Girl',
1052 'description': 'md5:7ae382a65843d6df2685993e90a8628f',
1053 'upload_date': '20190312',
1054 'uploader': 'Stephen - Topic',
1055 'uploader_id': 'UC-pWHpBjdGG69N9mM2auIAA',
1056 'artist': 'Stephen',
1057 'track': 'Voyeur Girl',
1058 'album': 'it\'s too much love to know my dear',
1059 'release_date': '20190313',
1060 'release_year': 2019,
1061 },
1062 'params': {
1063 'skip_download': True,
1064 },
1065 },
1066 {
1067 'url': 'https://www.youtubekids.com/watch?v=3b8nCWDgZ6Q',
1068 'only_matching': True,
1069 },
1070 {
1071 # invalid -> valid video id redirection
1072 'url': 'DJztXj2GPfl',
1073 'info_dict': {
1074 'id': 'DJztXj2GPfk',
1075 'ext': 'mp4',
1076 'title': 'Panjabi MC - Mundian To Bach Ke (The Dictator Soundtrack)',
1077 'description': 'md5:bf577a41da97918e94fa9798d9228825',
1078 'upload_date': '20090125',
1079 'uploader': 'Prochorowka',
1080 'uploader_id': 'Prochorowka',
1081 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/Prochorowka',
1082 'artist': 'Panjabi MC',
1083 'track': 'Beware of the Boys (Mundian to Bach Ke) - Motivo Hi-Lectro Remix',
1084 'album': 'Beware of the Boys (Mundian To Bach Ke)',
1085 },
1086 'params': {
1087 'skip_download': True,
1088 },
1089 },
1090 {
1091 # empty description results in an empty string
1092 'url': 'https://www.youtube.com/watch?v=x41yOUIvK2k',
1093 'info_dict': {
1094 'id': 'x41yOUIvK2k',
1095 'ext': 'mp4',
1096 'title': 'IMG 3456',
1097 'description': '',
1098 'upload_date': '20170613',
1099 'uploader_id': 'ElevageOrVert',
1100 'uploader': 'ElevageOrVert',
1101 },
1102 'params': {
1103 'skip_download': True,
1104 },
1105 },
1106 {
1107 # with '};' inside yt initial data (see [1])
1108 # see [2] for an example with '};' inside ytInitialPlayerResponse
1109 # 1. https://github.com/ytdl-org/youtube-dl/issues/27093
1110 # 2. https://github.com/ytdl-org/youtube-dl/issues/27216
1111 'url': 'https://www.youtube.com/watch?v=CHqg6qOn4no',
1112 'info_dict': {
1113 'id': 'CHqg6qOn4no',
1114 'ext': 'mp4',
1115 'title': 'Part 77 Sort a list of simple types in c#',
1116 'description': 'md5:b8746fa52e10cdbf47997903f13b20dc',
1117 'upload_date': '20130831',
1118 'uploader_id': 'kudvenkat',
1119 'uploader': 'kudvenkat',
1120 },
1121 'params': {
1122 'skip_download': True,
1123 },
1124 },
1125 {
1126 # another example of '};' in ytInitialData
1127 'url': 'https://www.youtube.com/watch?v=gVfgbahppCY',
1128 'only_matching': True,
1129 },
1130 {
1131 'url': 'https://www.youtube.com/watch_popup?v=63RmMXCd_bQ',
1132 'only_matching': True,
1133 },
1134 ]
1135
1136 def __init__(self, *args, **kwargs):
1137 super(YoutubeIE, self).__init__(*args, **kwargs)
1138 self._player_cache = {}
1139
1140 def report_video_info_webpage_download(self, video_id):
1141 """Report attempt to download video info webpage."""
1142 self.to_screen('%s: Downloading video info webpage' % video_id)
1143
1144 def report_information_extraction(self, video_id):
1145 """Report attempt to extract video information."""
1146 self.to_screen('%s: Extracting video information' % video_id)
1147
1148 def report_unavailable_format(self, video_id, format):
1149 """Report extracted video URL."""
1150 self.to_screen('%s: Format %s not available' % (video_id, format))
1151
1152 def report_rtmp_download(self):
1153 """Indicate the download will use the RTMP protocol."""
1154 self.to_screen('RTMP download detected')
1155
1156 def _signature_cache_id(self, example_sig):
1157 """ Return a string representation of a signature """
1158 return '.'.join(compat_str(len(part)) for part in example_sig.split('.'))
1159
1160 @classmethod
1161 def _extract_player_info(cls, player_url):
1162 for player_re in cls._PLAYER_INFO_RE:
1163 id_m = re.search(player_re, player_url)
1164 if id_m:
1165 break
1166 else:
1167 raise ExtractorError('Cannot identify player %r' % player_url)
1168 return id_m.group('ext'), id_m.group('id')
1169
1170 def _extract_signature_function(self, video_id, player_url, example_sig):
1171 player_type, player_id = self._extract_player_info(player_url)
1172
1173 # Read from filesystem cache
1174 func_id = '%s_%s_%s' % (
1175 player_type, player_id, self._signature_cache_id(example_sig))
1176 assert os.path.basename(func_id) == func_id
1177
1178 cache_spec = self._downloader.cache.load('youtube-sigfuncs', func_id)
1179 if cache_spec is not None:
1180 return lambda s: ''.join(s[i] for i in cache_spec)
1181
1182 download_note = (
1183 'Downloading player %s' % player_url
1184 if self._downloader.params.get('verbose') else
1185 'Downloading %s player %s' % (player_type, player_id)
1186 )
1187 if player_type == 'js':
1188 code = self._download_webpage(
1189 player_url, video_id,
1190 note=download_note,
1191 errnote='Download of %s failed' % player_url)
1192 res = self._parse_sig_js(code)
1193 elif player_type == 'swf':
1194 urlh = self._request_webpage(
1195 player_url, video_id,
1196 note=download_note,
1197 errnote='Download of %s failed' % player_url)
1198 code = urlh.read()
1199 res = self._parse_sig_swf(code)
1200 else:
1201 assert False, 'Invalid player type %r' % player_type
1202
1203 test_string = ''.join(map(compat_chr, range(len(example_sig))))
1204 cache_res = res(test_string)
1205 cache_spec = [ord(c) for c in cache_res]
1206
1207 self._downloader.cache.store('youtube-sigfuncs', func_id, cache_spec)
1208 return res
1209
1210 def _print_sig_code(self, func, example_sig):
1211 def gen_sig_code(idxs):
1212 def _genslice(start, end, step):
1213 starts = '' if start == 0 else str(start)
1214 ends = (':%d' % (end + step)) if end + step >= 0 else ':'
1215 steps = '' if step == 1 else (':%d' % step)
1216 return 's[%s%s%s]' % (starts, ends, steps)
1217
1218 step = None
1219 # Quelch pyflakes warnings - start will be set when step is set
1220 start = '(Never used)'
1221 for i, prev in zip(idxs[1:], idxs[:-1]):
1222 if step is not None:
1223 if i - prev == step:
1224 continue
1225 yield _genslice(start, prev, step)
1226 step = None
1227 continue
1228 if i - prev in [-1, 1]:
1229 step = i - prev
1230 start = prev
1231 continue
1232 else:
1233 yield 's[%d]' % prev
1234 if step is None:
1235 yield 's[%d]' % i
1236 else:
1237 yield _genslice(start, i, step)
1238
1239 test_string = ''.join(map(compat_chr, range(len(example_sig))))
1240 cache_res = func(test_string)
1241 cache_spec = [ord(c) for c in cache_res]
1242 expr_code = ' + '.join(gen_sig_code(cache_spec))
1243 signature_id_tuple = '(%s)' % (
1244 ', '.join(compat_str(len(p)) for p in example_sig.split('.')))
1245 code = ('if tuple(len(p) for p in s.split(\'.\')) == %s:\n'
1246 ' return %s\n') % (signature_id_tuple, expr_code)
1247 self.to_screen('Extracted signature function:\n' + code)
1248
1249 def _parse_sig_js(self, jscode):
1250 funcname = self._search_regex(
1251 (r'\b[cs]\s*&&\s*[adf]\.set\([^,]+\s*,\s*encodeURIComponent\s*\(\s*(?P<sig>[a-zA-Z0-9$]+)\(',
1252 r'\b[a-zA-Z0-9]+\s*&&\s*[a-zA-Z0-9]+\.set\([^,]+\s*,\s*encodeURIComponent\s*\(\s*(?P<sig>[a-zA-Z0-9$]+)\(',
1253 r'(?:\b|[^a-zA-Z0-9$])(?P<sig>[a-zA-Z0-9$]{2})\s*=\s*function\(\s*a\s*\)\s*{\s*a\s*=\s*a\.split\(\s*""\s*\)',
1254 r'(?P<sig>[a-zA-Z0-9$]+)\s*=\s*function\(\s*a\s*\)\s*{\s*a\s*=\s*a\.split\(\s*""\s*\)',
1255 # Obsolete patterns
1256 r'(["\'])signature\1\s*,\s*(?P<sig>[a-zA-Z0-9$]+)\(',
1257 r'\.sig\|\|(?P<sig>[a-zA-Z0-9$]+)\(',
1258 r'yt\.akamaized\.net/\)\s*\|\|\s*.*?\s*[cs]\s*&&\s*[adf]\.set\([^,]+\s*,\s*(?:encodeURIComponent\s*\()?\s*(?P<sig>[a-zA-Z0-9$]+)\(',
1259 r'\b[cs]\s*&&\s*[adf]\.set\([^,]+\s*,\s*(?P<sig>[a-zA-Z0-9$]+)\(',
1260 r'\b[a-zA-Z0-9]+\s*&&\s*[a-zA-Z0-9]+\.set\([^,]+\s*,\s*(?P<sig>[a-zA-Z0-9$]+)\(',
1261 r'\bc\s*&&\s*a\.set\([^,]+\s*,\s*\([^)]*\)\s*\(\s*(?P<sig>[a-zA-Z0-9$]+)\(',
1262 r'\bc\s*&&\s*[a-zA-Z0-9]+\.set\([^,]+\s*,\s*\([^)]*\)\s*\(\s*(?P<sig>[a-zA-Z0-9$]+)\(',
1263 r'\bc\s*&&\s*[a-zA-Z0-9]+\.set\([^,]+\s*,\s*\([^)]*\)\s*\(\s*(?P<sig>[a-zA-Z0-9$]+)\('),
1264 jscode, 'Initial JS player signature function name', group='sig')
1265
1266 jsi = JSInterpreter(jscode)
1267 initial_function = jsi.extract_function(funcname)
1268 return lambda s: initial_function([s])
1269
1270 def _parse_sig_swf(self, file_contents):
1271 swfi = SWFInterpreter(file_contents)
1272 TARGET_CLASSNAME = 'SignatureDecipher'
1273 searched_class = swfi.extract_class(TARGET_CLASSNAME)
1274 initial_function = swfi.extract_function(searched_class, 'decipher')
1275 return lambda s: initial_function([s])
1276
1277 def _decrypt_signature(self, s, video_id, player_url, age_gate=False):
1278 """Turn the encrypted s field into a working signature"""
1279
1280 if player_url is None:
1281 raise ExtractorError('Cannot decrypt signature without player_url')
1282
1283 if player_url.startswith('//'):
1284 player_url = 'https:' + player_url
1285 elif not re.match(r'https?://', player_url):
1286 player_url = compat_urlparse.urljoin(
1287 'https://www.youtube.com', player_url)
1288 try:
1289 player_id = (player_url, self._signature_cache_id(s))
1290 if player_id not in self._player_cache:
1291 func = self._extract_signature_function(
1292 video_id, player_url, s
1293 )
1294 self._player_cache[player_id] = func
1295 func = self._player_cache[player_id]
1296 if self._downloader.params.get('youtube_print_sig_code'):
1297 self._print_sig_code(func, s)
1298 return func(s)
1299 except Exception as e:
1300 tb = traceback.format_exc()
1301 raise ExtractorError(
1302 'Signature extraction failed: ' + tb, cause=e)
1303
1304 def _get_subtitles(self, video_id, webpage, has_live_chat_replay):
1305 try:
1306 subs_doc = self._download_xml(
1307 'https://video.google.com/timedtext?hl=en&type=list&v=%s' % video_id,
1308 video_id, note=False)
1309 except ExtractorError as err:
1310 self._downloader.report_warning('unable to download video subtitles: %s' % error_to_compat_str(err))
1311 return {}
1312
1313 sub_lang_list = {}
1314 for track in subs_doc.findall('track'):
1315 lang = track.attrib['lang_code']
1316 if lang in sub_lang_list:
1317 continue
1318 sub_formats = []
1319 for ext in self._SUBTITLE_FORMATS:
1320 params = compat_urllib_parse_urlencode({
1321 'lang': lang,
1322 'v': video_id,
1323 'fmt': ext,
1324 'name': track.attrib['name'].encode('utf-8'),
1325 })
1326 sub_formats.append({
1327 'url': 'https://www.youtube.com/api/timedtext?' + params,
1328 'ext': ext,
1329 })
1330 sub_lang_list[lang] = sub_formats
1331 if has_live_chat_replay:
1332 sub_lang_list['live_chat'] = [
1333 {
1334 'video_id': video_id,
1335 'ext': 'json',
1336 'protocol': 'youtube_live_chat_replay',
1337 },
1338 ]
1339 if not sub_lang_list:
1340 self._downloader.report_warning('video doesn\'t have subtitles')
1341 return {}
1342 return sub_lang_list
1343
1344 def _get_ytplayer_config(self, video_id, webpage):
1345 patterns = (
1346 # User data may contain arbitrary character sequences that may affect
1347 # JSON extraction with regex, e.g. when '};' is contained the second
1348 # regex won't capture the whole JSON. Yet working around by trying more
1349 # concrete regex first keeping in mind proper quoted string handling
1350 # to be implemented in future that will replace this workaround (see
1351 # https://github.com/ytdl-org/youtube-dl/issues/7468,
1352 # https://github.com/ytdl-org/youtube-dl/pull/7599)
1353 r';ytplayer\.config\s*=\s*({.+?});ytplayer',
1354 r';ytplayer\.config\s*=\s*({.+?});',
1355 )
1356 config = self._search_regex(
1357 patterns, webpage, 'ytplayer.config', default=None)
1358 if config:
1359 return self._parse_json(
1360 uppercase_escape(config), video_id, fatal=False)
1361
1362 def _get_automatic_captions(self, video_id, player_response, player_config):
1363 """We need the webpage for getting the captions url, pass it as an
1364 argument to speed up the process."""
1365 self.to_screen('%s: Looking for automatic captions' % video_id)
1366 err_msg = 'Couldn\'t find automatic captions for %s' % video_id
1367 if not (player_response or player_config):
1368 self._downloader.report_warning(err_msg)
1369 return {}
1370 try:
1371 args = player_config.get('args') if player_config else {}
1372 caption_url = args.get('ttsurl')
1373 if caption_url:
1374 timestamp = args['timestamp']
1375 # We get the available subtitles
1376 list_params = compat_urllib_parse_urlencode({
1377 'type': 'list',
1378 'tlangs': 1,
1379 'asrs': 1,
1380 })
1381 list_url = caption_url + '&' + list_params
1382 caption_list = self._download_xml(list_url, video_id)
1383 original_lang_node = caption_list.find('track')
1384 if original_lang_node is None:
1385 self._downloader.report_warning('Video doesn\'t have automatic captions')
1386 return {}
1387 original_lang = original_lang_node.attrib['lang_code']
1388 caption_kind = original_lang_node.attrib.get('kind', '')
1389
1390 sub_lang_list = {}
1391 for lang_node in caption_list.findall('target'):
1392 sub_lang = lang_node.attrib['lang_code']
1393 sub_formats = []
1394 for ext in self._SUBTITLE_FORMATS:
1395 params = compat_urllib_parse_urlencode({
1396 'lang': original_lang,
1397 'tlang': sub_lang,
1398 'fmt': ext,
1399 'ts': timestamp,
1400 'kind': caption_kind,
1401 })
1402 sub_formats.append({
1403 'url': caption_url + '&' + params,
1404 'ext': ext,
1405 })
1406 sub_lang_list[sub_lang] = sub_formats
1407 return sub_lang_list
1408
1409 def make_captions(sub_url, sub_langs):
1410 parsed_sub_url = compat_urllib_parse_urlparse(sub_url)
1411 caption_qs = compat_parse_qs(parsed_sub_url.query)
1412 captions = {}
1413 for sub_lang in sub_langs:
1414 sub_formats = []
1415 for ext in self._SUBTITLE_FORMATS:
1416 caption_qs.update({
1417 'tlang': [sub_lang],
1418 'fmt': [ext],
1419 })
1420 sub_url = compat_urlparse.urlunparse(parsed_sub_url._replace(
1421 query=compat_urllib_parse_urlencode(caption_qs, True)))
1422 sub_formats.append({
1423 'url': sub_url,
1424 'ext': ext,
1425 })
1426 captions[sub_lang] = sub_formats
1427 return captions
1428
1429 # New captions format as of 22.06.2017
1430 if player_response:
1431 renderer = player_response['captions']['playerCaptionsTracklistRenderer']
1432 base_url = renderer['captionTracks'][0]['baseUrl']
1433 sub_lang_list = []
1434 for lang in renderer['translationLanguages']:
1435 lang_code = lang.get('languageCode')
1436 if lang_code:
1437 sub_lang_list.append(lang_code)
1438 return make_captions(base_url, sub_lang_list)
1439
1440 # Some videos don't provide ttsurl but rather caption_tracks and
1441 # caption_translation_languages (e.g. 20LmZk1hakA)
1442 # Does not used anymore as of 22.06.2017
1443 caption_tracks = args['caption_tracks']
1444 caption_translation_languages = args['caption_translation_languages']
1445 caption_url = compat_parse_qs(caption_tracks.split(',')[0])['u'][0]
1446 sub_lang_list = []
1447 for lang in caption_translation_languages.split(','):
1448 lang_qs = compat_parse_qs(compat_urllib_parse_unquote_plus(lang))
1449 sub_lang = lang_qs.get('lc', [None])[0]
1450 if sub_lang:
1451 sub_lang_list.append(sub_lang)
1452 return make_captions(caption_url, sub_lang_list)
1453 # An extractor error can be raise by the download process if there are
1454 # no automatic captions but there are subtitles
1455 except (KeyError, IndexError, ExtractorError):
1456 self._downloader.report_warning(err_msg)
1457 return {}
1458
1459 def _mark_watched(self, video_id, video_info, player_response):
1460 playback_url = url_or_none(try_get(
1461 player_response,
1462 lambda x: x['playbackTracking']['videostatsPlaybackUrl']['baseUrl']) or try_get(
1463 video_info, lambda x: x['videostats_playback_base_url'][0]))
1464 if not playback_url:
1465 return
1466 parsed_playback_url = compat_urlparse.urlparse(playback_url)
1467 qs = compat_urlparse.parse_qs(parsed_playback_url.query)
1468
1469 # cpn generation algorithm is reverse engineered from base.js.
1470 # In fact it works even with dummy cpn.
1471 CPN_ALPHABET = 'abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789-_'
1472 cpn = ''.join((CPN_ALPHABET[random.randint(0, 256) & 63] for _ in range(0, 16)))
1473
1474 qs.update({
1475 'ver': ['2'],
1476 'cpn': [cpn],
1477 })
1478 playback_url = compat_urlparse.urlunparse(
1479 parsed_playback_url._replace(query=compat_urllib_parse_urlencode(qs, True)))
1480
1481 self._download_webpage(
1482 playback_url, video_id, 'Marking watched',
1483 'Unable to mark watched', fatal=False)
1484
1485 @staticmethod
1486 def _extract_urls(webpage):
1487 # Embedded YouTube player
1488 entries = [
1489 unescapeHTML(mobj.group('url'))
1490 for mobj in re.finditer(r'''(?x)
1491 (?:
1492 <iframe[^>]+?src=|
1493 data-video-url=|
1494 <embed[^>]+?src=|
1495 embedSWF\(?:\s*|
1496 <object[^>]+data=|
1497 new\s+SWFObject\(
1498 )
1499 (["\'])
1500 (?P<url>(?:https?:)?//(?:www\.)?youtube(?:-nocookie)?\.com/
1501 (?:embed|v|p)/[0-9A-Za-z_-]{11}.*?)
1502 \1''', webpage)]
1503
1504 # lazyYT YouTube embed
1505 entries.extend(list(map(
1506 unescapeHTML,
1507 re.findall(r'class="lazyYT" data-youtube-id="([^"]+)"', webpage))))
1508
1509 # Wordpress "YouTube Video Importer" plugin
1510 matches = re.findall(r'''(?x)<div[^>]+
1511 class=(?P<q1>[\'"])[^\'"]*\byvii_single_video_player\b[^\'"]*(?P=q1)[^>]+
1512 data-video_id=(?P<q2>[\'"])([^\'"]+)(?P=q2)''', webpage)
1513 entries.extend(m[-1] for m in matches)
1514
1515 return entries
1516
1517 @staticmethod
1518 def _extract_url(webpage):
1519 urls = YoutubeIE._extract_urls(webpage)
1520 return urls[0] if urls else None
1521
1522 @classmethod
1523 def extract_id(cls, url):
1524 mobj = re.match(cls._VALID_URL, url, re.VERBOSE)
1525 if mobj is None:
1526 raise ExtractorError('Invalid URL: %s' % url)
1527 video_id = mobj.group(2)
1528 return video_id
1529
1530 def _extract_chapters_from_json(self, webpage, video_id, duration):
1531 if not webpage:
1532 return
1533 data = self._extract_yt_initial_data(video_id, webpage)
1534 if not data or not isinstance(data, dict):
1535 return
1536 chapters_list = try_get(
1537 data,
1538 lambda x: x['playerOverlays']
1539 ['playerOverlayRenderer']
1540 ['decoratedPlayerBarRenderer']
1541 ['decoratedPlayerBarRenderer']
1542 ['playerBar']
1543 ['chapteredPlayerBarRenderer']
1544 ['chapters'],
1545 list)
1546 if not chapters_list:
1547 return
1548
1549 def chapter_time(chapter):
1550 return float_or_none(
1551 try_get(
1552 chapter,
1553 lambda x: x['chapterRenderer']['timeRangeStartMillis'],
1554 int),
1555 scale=1000)
1556 chapters = []
1557 for next_num, chapter in enumerate(chapters_list, start=1):
1558 start_time = chapter_time(chapter)
1559 if start_time is None:
1560 continue
1561 end_time = (chapter_time(chapters_list[next_num])
1562 if next_num < len(chapters_list) else duration)
1563 if end_time is None:
1564 continue
1565 title = try_get(
1566 chapter, lambda x: x['chapterRenderer']['title']['simpleText'],
1567 compat_str)
1568 chapters.append({
1569 'start_time': start_time,
1570 'end_time': end_time,
1571 'title': title,
1572 })
1573 return chapters
1574
1575 @staticmethod
1576 def _extract_chapters_from_description(description, duration):
1577 if not description:
1578 return None
1579 chapter_lines = re.findall(
1580 r'(?:^|<br\s*/>)([^<]*<a[^>]+onclick=["\']yt\.www\.watch\.player\.seekTo[^>]+>(\d{1,2}:\d{1,2}(?::\d{1,2})?)</a>[^>]*)(?=$|<br\s*/>)',
1581 description)
1582 if not chapter_lines:
1583 return None
1584 chapters = []
1585 for next_num, (chapter_line, time_point) in enumerate(
1586 chapter_lines, start=1):
1587 start_time = parse_duration(time_point)
1588 if start_time is None:
1589 continue
1590 if start_time > duration:
1591 break
1592 end_time = (duration if next_num == len(chapter_lines)
1593 else parse_duration(chapter_lines[next_num][1]))
1594 if end_time is None:
1595 continue
1596 if end_time > duration:
1597 end_time = duration
1598 if start_time > end_time:
1599 break
1600 chapter_title = re.sub(
1601 r'<a[^>]+>[^<]+</a>', '', chapter_line).strip(' \t-')
1602 chapter_title = re.sub(r'\s+', ' ', chapter_title)
1603 chapters.append({
1604 'start_time': start_time,
1605 'end_time': end_time,
1606 'title': chapter_title,
1607 })
1608 return chapters
1609
1610 def _extract_chapters(self, webpage, description, video_id, duration):
1611 return (self._extract_chapters_from_json(webpage, video_id, duration)
1612 or self._extract_chapters_from_description(description, duration))
1613
1614 def _real_extract(self, url):
1615 url, smuggled_data = unsmuggle_url(url, {})
1616
1617 proto = (
1618 'http' if self._downloader.params.get('prefer_insecure', False)
1619 else 'https')
1620
1621 start_time = None
1622 end_time = None
1623 parsed_url = compat_urllib_parse_urlparse(url)
1624 for component in [parsed_url.fragment, parsed_url.query]:
1625 query = compat_parse_qs(component)
1626 if start_time is None and 't' in query:
1627 start_time = parse_duration(query['t'][0])
1628 if start_time is None and 'start' in query:
1629 start_time = parse_duration(query['start'][0])
1630 if end_time is None and 'end' in query:
1631 end_time = parse_duration(query['end'][0])
1632
1633 # Extract original video URL from URL with redirection, like age verification, using next_url parameter
1634 mobj = re.search(self._NEXT_URL_RE, url)
1635 if mobj:
1636 url = proto + '://www.youtube.com/' + compat_urllib_parse_unquote(mobj.group(1)).lstrip('/')
1637 video_id = self.extract_id(url)
1638
1639 # Get video webpage
1640 url = proto + '://www.youtube.com/watch?v=%s&gl=US&hl=en&has_verified=1&bpctr=9999999999' % video_id
1641 video_webpage, urlh = self._download_webpage_handle(url, video_id)
1642
1643 qs = compat_parse_qs(compat_urllib_parse_urlparse(urlh.geturl()).query)
1644 video_id = qs.get('v', [None])[0] or video_id
1645
1646 # Attempt to extract SWF player URL
1647 mobj = re.search(r'swfConfig.*?"(https?:\\/\\/.*?watch.*?-.*?\.swf)"', video_webpage)
1648 if mobj is not None:
1649 player_url = re.sub(r'\\(.)', r'\1', mobj.group(1))
1650 else:
1651 player_url = None
1652
1653 dash_mpds = []
1654
1655 def add_dash_mpd(video_info):
1656 dash_mpd = video_info.get('dashmpd')
1657 if dash_mpd and dash_mpd[0] not in dash_mpds:
1658 dash_mpds.append(dash_mpd[0])
1659
1660 def add_dash_mpd_pr(pl_response):
1661 dash_mpd = url_or_none(try_get(
1662 pl_response, lambda x: x['streamingData']['dashManifestUrl'],
1663 compat_str))
1664 if dash_mpd and dash_mpd not in dash_mpds:
1665 dash_mpds.append(dash_mpd)
1666
1667 is_live = None
1668 view_count = None
1669
1670 def extract_view_count(v_info):
1671 return int_or_none(try_get(v_info, lambda x: x['view_count'][0]))
1672
1673 def extract_player_response(player_response, video_id):
1674 pl_response = str_or_none(player_response)
1675 if not pl_response:
1676 return
1677 pl_response = self._parse_json(pl_response, video_id, fatal=False)
1678 if isinstance(pl_response, dict):
1679 add_dash_mpd_pr(pl_response)
1680 return pl_response
1681
1682 def extract_embedded_config(embed_webpage, video_id):
1683 embedded_config = self._search_regex(
1684 r'setConfig\(({.*})\);',
1685 embed_webpage, 'ytInitialData', default=None)
1686 if embedded_config:
1687 return embedded_config
1688
1689 video_info = {}
1690 player_response = {}
1691 ytplayer_config = None
1692 embed_webpage = None
1693
1694 # Get video info
1695 if (self._og_search_property('restrictions:age', video_webpage, default=None) == '18+'
1696 or re.search(r'player-age-gate-content">', video_webpage) is not None):
1697 cookie_keys = self._get_cookies('https://www.youtube.com').keys()
1698 age_gate = True
1699 # We simulate the access to the video from www.youtube.com/v/{video_id}
1700 # this can be viewed without login into Youtube
1701 url = proto + '://www.youtube.com/embed/%s' % video_id
1702 embed_webpage = self._download_webpage(url, video_id, 'Downloading embed webpage')
1703 ext = extract_embedded_config(embed_webpage, video_id)
1704 # playabilityStatus = re.search(r'{\\\"status\\\":\\\"(?P<playabilityStatus>[^\"]+)\\\"', ext)
1705 playable_in_embed = re.search(r'{\\\"playableInEmbed\\\":(?P<playableinEmbed>[^\,]+)', ext)
1706 if not playable_in_embed:
1707 self.to_screen('Could not determine whether playabale in embed for video %s' % video_id)
1708 playable_in_embed = ''
1709 else:
1710 playable_in_embed = playable_in_embed.group('playableinEmbed')
1711 # check if video is only playable on youtube in other words not playable in embed - if so it requires auth (cookies)
1712 # if re.search(r'player-unavailable">', embed_webpage) is not None:
1713 if playable_in_embed == 'false':
1714 '''
1715 # TODO apply this patch when Support for Python 2.6(!) and above drops
1716 if ({'VISITOR_INFO1_LIVE', 'HSID', 'SSID', 'SID'} <= cookie_keys
1717 or {'VISITOR_INFO1_LIVE', '__Secure-3PSID', 'LOGIN_INFO'} <= cookie_keys):
1718 '''
1719 if (set(('VISITOR_INFO1_LIVE', 'HSID', 'SSID', 'SID')) <= set(cookie_keys)
1720 or set(('VISITOR_INFO1_LIVE', '__Secure-3PSID', 'LOGIN_INFO')) <= set(cookie_keys)):
1721 age_gate = False
1722 # Try looking directly into the video webpage
1723 ytplayer_config = self._get_ytplayer_config(video_id, video_webpage)
1724 if ytplayer_config:
1725 args = ytplayer_config.get("args")
1726 if args is not None:
1727 if args.get('url_encoded_fmt_stream_map') or args.get('hlsvp'):
1728 # Convert to the same format returned by compat_parse_qs
1729 video_info = dict((k, [v]) for k, v in args.items())
1730 add_dash_mpd(video_info)
1731 # Rental video is not rented but preview is available (e.g.
1732 # https://www.youtube.com/watch?v=yYr8q0y5Jfg,
1733 # https://github.com/ytdl-org/youtube-dl/issues/10532)
1734 if not video_info and args.get('ypc_vid'):
1735 return self.url_result(
1736 args['ypc_vid'], YoutubeIE.ie_key(), video_id=args['ypc_vid'])
1737 if args.get('livestream') == '1' or args.get('live_playback') == 1:
1738 is_live = True
1739 if not player_response:
1740 player_response = extract_player_response(args.get('player_response'), video_id)
1741 elif not player_response:
1742 player_response = ytplayer_config
1743 if not video_info or self._downloader.params.get('youtube_include_dash_manifest', True):
1744 add_dash_mpd_pr(player_response)
1745 else:
1746 raise ExtractorError('Video is age restricted and only playable on Youtube. Requires cookies!', expected=True)
1747 else:
1748 data = compat_urllib_parse_urlencode({
1749 'video_id': video_id,
1750 'eurl': 'https://youtube.googleapis.com/v/' + video_id,
1751 'sts': self._search_regex(
1752 r'"sts"\s*:\s*(\d+)', embed_webpage, 'sts', default=''),
1753 })
1754 video_info_url = proto + '://www.youtube.com/get_video_info?' + data
1755 try:
1756 video_info_webpage = self._download_webpage(
1757 video_info_url, video_id,
1758 note='Refetching age-gated info webpage',
1759 errnote='unable to download video info webpage')
1760 except ExtractorError:
1761 video_info_webpage = None
1762 if video_info_webpage:
1763 video_info = compat_parse_qs(video_info_webpage)
1764 pl_response = video_info.get('player_response', [None])[0]
1765 player_response = extract_player_response(pl_response, video_id)
1766 add_dash_mpd(video_info)
1767 view_count = extract_view_count(video_info)
1768 else:
1769 age_gate = False
1770 # Try looking directly into the video webpage
1771 ytplayer_config = self._get_ytplayer_config(video_id, video_webpage)
1772 if ytplayer_config:
1773 args = ytplayer_config.get('args', {})
1774 if args.get('url_encoded_fmt_stream_map') or args.get('hlsvp'):
1775 # Convert to the same format returned by compat_parse_qs
1776 video_info = dict((k, [v]) for k, v in args.items())
1777 add_dash_mpd(video_info)
1778 # Rental video is not rented but preview is available (e.g.
1779 # https://www.youtube.com/watch?v=yYr8q0y5Jfg,
1780 # https://github.com/ytdl-org/youtube-dl/issues/10532)
1781 if not video_info and args.get('ypc_vid'):
1782 return self.url_result(
1783 args['ypc_vid'], YoutubeIE.ie_key(), video_id=args['ypc_vid'])
1784 if args.get('livestream') == '1' or args.get('live_playback') == 1:
1785 is_live = True
1786 if not player_response:
1787 player_response = extract_player_response(args.get('player_response'), video_id)
1788 if not video_info or self._downloader.params.get('youtube_include_dash_manifest', True):
1789 add_dash_mpd_pr(player_response)
1790
1791 if not video_info and not player_response:
1792 player_response = extract_player_response(
1793 self._search_regex(
1794 (r'%s\s*%s' % (self._YT_INITIAL_PLAYER_RESPONSE_RE, self._YT_INITIAL_BOUNDARY_RE),
1795 self._YT_INITIAL_PLAYER_RESPONSE_RE), video_webpage,
1796 'initial player response', default='{}'),
1797 video_id)
1798
1799 def extract_unavailable_message():
1800 messages = []
1801 for tag, kind in (('h1', 'message'), ('div', 'submessage')):
1802 msg = self._html_search_regex(
1803 r'(?s)<{tag}[^>]+id=["\']unavailable-{kind}["\'][^>]*>(.+?)</{tag}>'.format(tag=tag, kind=kind),
1804 video_webpage, 'unavailable %s' % kind, default=None)
1805 if msg:
1806 messages.append(msg)
1807 if messages:
1808 return '\n'.join(messages)
1809
1810 if not video_info and not player_response:
1811 unavailable_message = extract_unavailable_message()
1812 if not unavailable_message:
1813 unavailable_message = 'Unable to extract video data'
1814 raise ExtractorError(
1815 'YouTube said: %s' % unavailable_message, expected=True, video_id=video_id)
1816
1817 if not isinstance(video_info, dict):
1818 video_info = {}
1819
1820 video_details = try_get(
1821 player_response, lambda x: x['videoDetails'], dict) or {}
1822
1823 microformat = try_get(
1824 player_response, lambda x: x['microformat']['playerMicroformatRenderer'], dict) or {}
1825
1826 video_title = video_info.get('title', [None])[0] or video_details.get('title')
1827 if not video_title:
1828 self._downloader.report_warning('Unable to extract video title')
1829 video_title = '_'
1830
1831 description_original = video_description = get_element_by_id("eow-description", video_webpage)
1832 if video_description:
1833
1834 def replace_url(m):
1835 redir_url = compat_urlparse.urljoin(url, m.group(1))
1836 parsed_redir_url = compat_urllib_parse_urlparse(redir_url)
1837 if re.search(r'^(?:www\.)?(?:youtube(?:-nocookie)?\.com|youtu\.be)$', parsed_redir_url.netloc) and parsed_redir_url.path == '/redirect':
1838 qs = compat_parse_qs(parsed_redir_url.query)
1839 q = qs.get('q')
1840 if q and q[0]:
1841 return q[0]
1842 return redir_url
1843
1844 description_original = video_description = re.sub(r'''(?x)
1845 <a\s+
1846 (?:[a-zA-Z-]+="[^"]*"\s+)*?
1847 (?:title|href)="([^"]+)"\s+
1848 (?:[a-zA-Z-]+="[^"]*"\s+)*?
1849 class="[^"]*"[^>]*>
1850 [^<]+\.{3}\s*
1851 </a>
1852 ''', replace_url, video_description)
1853 video_description = clean_html(video_description)
1854 else:
1855 video_description = video_details.get('shortDescription')
1856 if video_description is None:
1857 video_description = self._html_search_meta('description', video_webpage)
1858
1859 if not smuggled_data.get('force_singlefeed', False):
1860 if not self._downloader.params.get('noplaylist'):
1861 multifeed_metadata_list = try_get(
1862 player_response,
1863 lambda x: x['multicamera']['playerLegacyMulticameraRenderer']['metadataList'],
1864 compat_str) or try_get(
1865 video_info, lambda x: x['multifeed_metadata_list'][0], compat_str)
1866 if multifeed_metadata_list:
1867 entries = []
1868 feed_ids = []
1869 for feed in multifeed_metadata_list.split(','):
1870 # Unquote should take place before split on comma (,) since textual
1871 # fields may contain comma as well (see
1872 # https://github.com/ytdl-org/youtube-dl/issues/8536)
1873 feed_data = compat_parse_qs(compat_urllib_parse_unquote_plus(feed))
1874
1875 def feed_entry(name):
1876 return try_get(feed_data, lambda x: x[name][0], compat_str)
1877
1878 feed_id = feed_entry('id')
1879 if not feed_id:
1880 continue
1881 feed_title = feed_entry('title')
1882 title = video_title
1883 if feed_title:
1884 title += ' (%s)' % feed_title
1885 entries.append({
1886 '_type': 'url_transparent',
1887 'ie_key': 'Youtube',
1888 'url': smuggle_url(
1889 '%s://www.youtube.com/watch?v=%s' % (proto, feed_data['id'][0]),
1890 {'force_singlefeed': True}),
1891 'title': title,
1892 })
1893 feed_ids.append(feed_id)
1894 self.to_screen(
1895 'Downloading multifeed video (%s) - add --no-playlist to just download video %s'
1896 % (', '.join(feed_ids), video_id))
1897 return self.playlist_result(entries, video_id, video_title, video_description)
1898 else:
1899 self.to_screen('Downloading just video %s because of --no-playlist' % video_id)
1900
1901 if view_count is None:
1902 view_count = extract_view_count(video_info)
1903 if view_count is None and video_details:
1904 view_count = int_or_none(video_details.get('viewCount'))
1905 if view_count is None and microformat:
1906 view_count = int_or_none(microformat.get('viewCount'))
1907
1908 if is_live is None:
1909 is_live = bool_or_none(video_details.get('isLive'))
1910
1911 has_live_chat_replay = False
1912 if not is_live:
1913 yt_initial_data = self._get_yt_initial_data(video_id, video_webpage)
1914 try:
1915 yt_initial_data['contents']['twoColumnWatchNextResults']['conversationBar']['liveChatRenderer']['continuations'][0]['reloadContinuationData']['continuation']
1916 has_live_chat_replay = True
1917 except (KeyError, IndexError, TypeError):
1918 pass
1919
1920 # Check for "rental" videos
1921 if 'ypc_video_rental_bar_text' in video_info and 'author' not in video_info:
1922 raise ExtractorError('"rental" videos not supported. See https://github.com/ytdl-org/youtube-dl/issues/359 for more information.', expected=True)
1923
1924 def _extract_filesize(media_url):
1925 return int_or_none(self._search_regex(
1926 r'\bclen[=/](\d+)', media_url, 'filesize', default=None))
1927
1928 streaming_formats = try_get(player_response, lambda x: x['streamingData']['formats'], list) or []
1929 streaming_formats.extend(try_get(player_response, lambda x: x['streamingData']['adaptiveFormats'], list) or [])
1930
1931 if 'conn' in video_info and video_info['conn'][0].startswith('rtmp'):
1932 self.report_rtmp_download()
1933 formats = [{
1934 'format_id': '_rtmp',
1935 'protocol': 'rtmp',
1936 'url': video_info['conn'][0],
1937 'player_url': player_url,
1938 }]
1939 elif not is_live and (streaming_formats or len(video_info.get('url_encoded_fmt_stream_map', [''])[0]) >= 1 or len(video_info.get('adaptive_fmts', [''])[0]) >= 1):
1940 encoded_url_map = video_info.get('url_encoded_fmt_stream_map', [''])[0] + ',' + video_info.get('adaptive_fmts', [''])[0]
1941 if 'rtmpe%3Dyes' in encoded_url_map:
1942 raise ExtractorError('rtmpe downloads are not supported, see https://github.com/ytdl-org/youtube-dl/issues/343 for more information.', expected=True)
1943 formats = []
1944 formats_spec = {}
1945 fmt_list = video_info.get('fmt_list', [''])[0]
1946 if fmt_list:
1947 for fmt in fmt_list.split(','):
1948 spec = fmt.split('/')
1949 if len(spec) > 1:
1950 width_height = spec[1].split('x')
1951 if len(width_height) == 2:
1952 formats_spec[spec[0]] = {
1953 'resolution': spec[1],
1954 'width': int_or_none(width_height[0]),
1955 'height': int_or_none(width_height[1]),
1956 }
1957 for fmt in streaming_formats:
1958 itag = str_or_none(fmt.get('itag'))
1959 if not itag:
1960 continue
1961 quality = fmt.get('quality')
1962 quality_label = fmt.get('qualityLabel') or quality
1963 formats_spec[itag] = {
1964 'asr': int_or_none(fmt.get('audioSampleRate')),
1965 'filesize': int_or_none(fmt.get('contentLength')),
1966 'format_note': quality_label,
1967 'fps': int_or_none(fmt.get('fps')),
1968 'height': int_or_none(fmt.get('height')),
1969 # bitrate for itag 43 is always 2147483647
1970 'tbr': float_or_none(fmt.get('averageBitrate') or fmt.get('bitrate'), 1000) if itag != '43' else None,
1971 'width': int_or_none(fmt.get('width')),
1972 }
1973
1974 for fmt in streaming_formats:
1975 if fmt.get('drmFamilies') or fmt.get('drm_families'):
1976 continue
1977 url = url_or_none(fmt.get('url'))
1978
1979 if not url:
1980 cipher = fmt.get('cipher') or fmt.get('signatureCipher')
1981 if not cipher:
1982 continue
1983 url_data = compat_parse_qs(cipher)
1984 url = url_or_none(try_get(url_data, lambda x: x['url'][0], compat_str))
1985 if not url:
1986 continue
1987 else:
1988 cipher = None
1989 url_data = compat_parse_qs(compat_urllib_parse_urlparse(url).query)
1990
1991 stream_type = int_or_none(try_get(url_data, lambda x: x['stream_type'][0]))
1992 # Unsupported FORMAT_STREAM_TYPE_OTF
1993 if stream_type == 3:
1994 continue
1995
1996 format_id = fmt.get('itag') or url_data['itag'][0]
1997 if not format_id:
1998 continue
1999 format_id = compat_str(format_id)
2000
2001 if cipher:
2002 if 's' in url_data or self._downloader.params.get('youtube_include_dash_manifest', True):
2003 ASSETS_RE = (
2004 r'<script[^>]+\bsrc=("[^"]+")[^>]+\bname=["\']player_ias/base',
2005 r'"jsUrl"\s*:\s*("[^"]+")',
2006 r'"assets":.+?"js":\s*("[^"]+")')
2007 jsplayer_url_json = self._search_regex(
2008 ASSETS_RE,
2009 embed_webpage if age_gate else video_webpage,
2010 'JS player URL (1)', default=None)
2011 if not jsplayer_url_json and not age_gate:
2012 # We need the embed website after all
2013 if embed_webpage is None:
2014 embed_url = proto + '://www.youtube.com/embed/%s' % video_id
2015 embed_webpage = self._download_webpage(
2016 embed_url, video_id, 'Downloading embed webpage')
2017 jsplayer_url_json = self._search_regex(
2018 ASSETS_RE, embed_webpage, 'JS player URL')
2019
2020 player_url = json.loads(jsplayer_url_json)
2021 if player_url is None:
2022 player_url_json = self._search_regex(
2023 r'ytplayer\.config.*?"url"\s*:\s*("[^"]+")',
2024 video_webpage, 'age gate player URL')
2025 player_url = json.loads(player_url_json)
2026
2027 if 'sig' in url_data:
2028 url += '&signature=' + url_data['sig'][0]
2029 elif 's' in url_data:
2030 encrypted_sig = url_data['s'][0]
2031
2032 if self._downloader.params.get('verbose'):
2033 if player_url is None:
2034 player_desc = 'unknown'
2035 else:
2036 player_type, player_version = self._extract_player_info(player_url)
2037 player_desc = '%s player %s' % ('flash' if player_type == 'swf' else 'html5', player_version)
2038 parts_sizes = self._signature_cache_id(encrypted_sig)
2039 self.to_screen('{%s} signature length %s, %s' %
2040 (format_id, parts_sizes, player_desc))
2041
2042 signature = self._decrypt_signature(
2043 encrypted_sig, video_id, player_url, age_gate)
2044 sp = try_get(url_data, lambda x: x['sp'][0], compat_str) or 'signature'
2045 url += '&%s=%s' % (sp, signature)
2046 if 'ratebypass' not in url:
2047 url += '&ratebypass=yes'
2048
2049 dct = {
2050 'format_id': format_id,
2051 'url': url,
2052 'player_url': player_url,
2053 }
2054 if format_id in self._formats:
2055 dct.update(self._formats[format_id])
2056 if format_id in formats_spec:
2057 dct.update(formats_spec[format_id])
2058
2059 # Some itags are not included in DASH manifest thus corresponding formats will
2060 # lack metadata (see https://github.com/ytdl-org/youtube-dl/pull/5993).
2061 # Trying to extract metadata from url_encoded_fmt_stream_map entry.
2062 mobj = re.search(r'^(?P<width>\d+)[xX](?P<height>\d+)$', url_data.get('size', [''])[0])
2063 width, height = (int(mobj.group('width')), int(mobj.group('height'))) if mobj else (None, None)
2064
2065 if width is None:
2066 width = int_or_none(fmt.get('width'))
2067 if height is None:
2068 height = int_or_none(fmt.get('height'))
2069
2070 filesize = int_or_none(url_data.get(
2071 'clen', [None])[0]) or _extract_filesize(url)
2072
2073 quality = url_data.get('quality', [None])[0] or fmt.get('quality')
2074 quality_label = url_data.get('quality_label', [None])[0] or fmt.get('qualityLabel')
2075
2076 tbr = (float_or_none(url_data.get('bitrate', [None])[0], 1000)
2077 or float_or_none(fmt.get('bitrate'), 1000)) if format_id != '43' else None
2078 fps = int_or_none(url_data.get('fps', [None])[0]) or int_or_none(fmt.get('fps'))
2079
2080 more_fields = {
2081 'filesize': filesize,
2082 'tbr': tbr,
2083 'width': width,
2084 'height': height,
2085 'fps': fps,
2086 'format_note': quality_label or quality,
2087 }
2088 for key, value in more_fields.items():
2089 if value:
2090 dct[key] = value
2091 type_ = url_data.get('type', [None])[0] or fmt.get('mimeType')
2092 if type_:
2093 type_split = type_.split(';')
2094 kind_ext = type_split[0].split('/')
2095 if len(kind_ext) == 2:
2096 kind, _ = kind_ext
2097 dct['ext'] = mimetype2ext(type_split[0])
2098 if kind in ('audio', 'video'):
2099 codecs = None
2100 for mobj in re.finditer(
2101 r'(?P<key>[a-zA-Z_-]+)=(?P<quote>["\']?)(?P<val>.+?)(?P=quote)(?:;|$)', type_):
2102 if mobj.group('key') == 'codecs':
2103 codecs = mobj.group('val')
2104 break
2105 if codecs:
2106 dct.update(parse_codecs(codecs))
2107 if dct.get('acodec') == 'none' or dct.get('vcodec') == 'none':
2108 dct['downloader_options'] = {
2109 # Youtube throttles chunks >~10M
2110 'http_chunk_size': 10485760,
2111 }
2112 formats.append(dct)
2113 else:
2114 manifest_url = (
2115 url_or_none(try_get(
2116 player_response,
2117 lambda x: x['streamingData']['hlsManifestUrl'],
2118 compat_str))
2119 or url_or_none(try_get(
2120 video_info, lambda x: x['hlsvp'][0], compat_str)))
2121 if manifest_url:
2122 formats = []
2123 m3u8_formats = self._extract_m3u8_formats(
2124 manifest_url, video_id, 'mp4', fatal=False)
2125 for a_format in m3u8_formats:
2126 itag = self._search_regex(
2127 r'/itag/(\d+)/', a_format['url'], 'itag', default=None)
2128 if itag:
2129 a_format['format_id'] = itag
2130 if itag in self._formats:
2131 dct = self._formats[itag].copy()
2132 dct.update(a_format)
2133 a_format = dct
2134 a_format['player_url'] = player_url
2135 # Accept-Encoding header causes failures in live streams on Youtube and Youtube Gaming
2136 a_format.setdefault('http_headers', {})['Youtubedl-no-compression'] = 'True'
2137 if self._downloader.params.get('youtube_include_hls_manifest', True):
2138 formats.append(a_format)
2139 else:
2140 error_message = extract_unavailable_message()
2141 if not error_message:
2142 reason_list = try_get(
2143 player_response,
2144 lambda x: x['playabilityStatus']['errorScreen']['playerErrorMessageRenderer']['subreason']['runs'],
2145 list) or []
2146 for reason in reason_list:
2147 if not isinstance(reason, dict):
2148 continue
2149 reason_text = try_get(reason, lambda x: x['text'], compat_str)
2150 if reason_text:
2151 if not error_message:
2152 error_message = ''
2153 error_message += reason_text
2154 if error_message:
2155 error_message = clean_html(error_message)
2156 if not error_message:
2157 error_message = clean_html(try_get(
2158 player_response, lambda x: x['playabilityStatus']['reason'],
2159 compat_str))
2160 if not error_message:
2161 error_message = clean_html(
2162 try_get(video_info, lambda x: x['reason'][0], compat_str))
2163 if error_message:
2164 raise ExtractorError(error_message, expected=True)
2165 raise ExtractorError('no conn, hlsvp, hlsManifestUrl or url_encoded_fmt_stream_map information found in video info')
2166
2167 # uploader
2168 video_uploader = try_get(
2169 video_info, lambda x: x['author'][0],
2170 compat_str) or str_or_none(video_details.get('author'))
2171 if video_uploader:
2172 video_uploader = compat_urllib_parse_unquote_plus(video_uploader)
2173 else:
2174 self._downloader.report_warning('unable to extract uploader name')
2175
2176 # uploader_id
2177 video_uploader_id = None
2178 video_uploader_url = None
2179 mobj = re.search(
2180 r'<link itemprop="url" href="(?P<uploader_url>https?://www\.youtube\.com/(?:user|channel)/(?P<uploader_id>[^"]+))">',
2181 video_webpage)
2182 if mobj is not None:
2183 video_uploader_id = mobj.group('uploader_id')
2184 video_uploader_url = mobj.group('uploader_url')
2185 else:
2186 owner_profile_url = url_or_none(microformat.get('ownerProfileUrl'))
2187 if owner_profile_url:
2188 video_uploader_id = self._search_regex(
2189 r'(?:user|channel)/([^/]+)', owner_profile_url, 'uploader id',
2190 default=None)
2191 video_uploader_url = owner_profile_url
2192
2193 channel_id = (
2194 str_or_none(video_details.get('channelId'))
2195 or self._html_search_meta(
2196 'channelId', video_webpage, 'channel id', default=None)
2197 or self._search_regex(
2198 r'data-channel-external-id=(["\'])(?P<id>(?:(?!\1).)+)\1',
2199 video_webpage, 'channel id', default=None, group='id'))
2200 channel_url = 'http://www.youtube.com/channel/%s' % channel_id if channel_id else None
2201
2202 thumbnails = []
2203 thumbnails_list = try_get(
2204 video_details, lambda x: x['thumbnail']['thumbnails'], list) or []
2205 for t in thumbnails_list:
2206 if not isinstance(t, dict):
2207 continue
2208 thumbnail_url = url_or_none(t.get('url'))
2209 if not thumbnail_url:
2210 continue
2211 thumbnails.append({
2212 'url': thumbnail_url,
2213 'width': int_or_none(t.get('width')),
2214 'height': int_or_none(t.get('height')),
2215 })
2216
2217 if not thumbnails:
2218 video_thumbnail = None
2219 # We try first to get a high quality image:
2220 m_thumb = re.search(r'<span itemprop="thumbnail".*?href="(.*?)">',
2221 video_webpage, re.DOTALL)
2222 if m_thumb is not None:
2223 video_thumbnail = m_thumb.group(1)
2224 thumbnail_url = try_get(video_info, lambda x: x['thumbnail_url'][0], compat_str)
2225 if thumbnail_url:
2226 video_thumbnail = compat_urllib_parse_unquote_plus(thumbnail_url)
2227 if video_thumbnail:
2228 thumbnails.append({'url': video_thumbnail})
2229
2230 # upload date
2231 upload_date = self._html_search_meta(
2232 'datePublished', video_webpage, 'upload date', default=None)
2233 if not upload_date:
2234 upload_date = self._search_regex(
2235 [r'(?s)id="eow-date.*?>(.*?)</span>',
2236 r'(?:id="watch-uploader-info".*?>.*?|["\']simpleText["\']\s*:\s*["\'])(?:Published|Uploaded|Streamed live|Started) on (.+?)[<"\']'],
2237 video_webpage, 'upload date', default=None)
2238 if not upload_date:
2239 upload_date = microformat.get('publishDate') or microformat.get('uploadDate')
2240 upload_date = unified_strdate(upload_date)
2241
2242 video_license = self._html_search_regex(
2243 r'<h4[^>]+class="title"[^>]*>\s*License\s*</h4>\s*<ul[^>]*>\s*<li>(.+?)</li',
2244 video_webpage, 'license', default=None)
2245
2246 m_music = re.search(
2247 r'''(?x)
2248 <h4[^>]+class="title"[^>]*>\s*Music\s*</h4>\s*
2249 <ul[^>]*>\s*
2250 <li>(?P<title>.+?)
2251 by (?P<creator>.+?)
2252 (?:
2253 \(.+?\)|
2254 <a[^>]*
2255 (?:
2256 \bhref=["\']/red[^>]*>| # drop possible
2257 >\s*Listen ad-free with YouTube Red # YouTube Red ad
2258 )
2259 .*?
2260 )?</li
2261 ''',
2262 video_webpage)
2263 if m_music:
2264 video_alt_title = remove_quotes(unescapeHTML(m_music.group('title')))
2265 video_creator = clean_html(m_music.group('creator'))
2266 else:
2267 video_alt_title = video_creator = None
2268
2269 def extract_meta(field):
2270 return self._html_search_regex(
2271 r'<h4[^>]+class="title"[^>]*>\s*%s\s*</h4>\s*<ul[^>]*>\s*<li>(.+?)</li>\s*' % field,
2272 video_webpage, field, default=None)
2273
2274 track = extract_meta('Song')
2275 artist = extract_meta('Artist')
2276 album = extract_meta('Album')
2277
2278 # Youtube Music Auto-generated description
2279 release_date = release_year = None
2280 if video_description:
2281 mobj = re.search(r'(?s)(?P<track>[^·\n]+)·(?P<artist>[^\n]+)\n+(?P<album>[^\n]+)(?:.+?℗\s*(?P<release_year>\d{4})(?!\d))?(?:.+?Released on\s*:\s*(?P<release_date>\d{4}-\d{2}-\d{2}))?(.+?\nArtist\s*:\s*(?P<clean_artist>[^\n]+))?.+\nAuto-generated by YouTube\.\s*$', video_description)
2282 if mobj:
2283 if not track:
2284 track = mobj.group('track').strip()
2285 if not artist:
2286 artist = mobj.group('clean_artist') or ', '.join(a.strip() for a in mobj.group('artist').split('·'))
2287 if not album:
2288 album = mobj.group('album'.strip())
2289 release_year = mobj.group('release_year')
2290 release_date = mobj.group('release_date')
2291 if release_date:
2292 release_date = release_date.replace('-', '')
2293 if not release_year:
2294 release_year = int(release_date[:4])
2295 if release_year:
2296 release_year = int(release_year)
2297
2298 yt_initial_data = self._extract_yt_initial_data(video_id, video_webpage)
2299 contents = try_get(yt_initial_data, lambda x: x['contents']['twoColumnWatchNextResults']['results']['results']['contents'], list) or []
2300 for content in contents:
2301 rows = try_get(content, lambda x: x['videoSecondaryInfoRenderer']['metadataRowContainer']['metadataRowContainerRenderer']['rows'], list) or []
2302 multiple_songs = False
2303 for row in rows:
2304 if try_get(row, lambda x: x['metadataRowRenderer']['hasDividerLine']) is True:
2305 multiple_songs = True
2306 break
2307 for row in rows:
2308 mrr = row.get('metadataRowRenderer') or {}
2309 mrr_title = try_get(
2310 mrr, lambda x: x['title']['simpleText'], compat_str)
2311 mrr_contents = try_get(
2312 mrr, lambda x: x['contents'][0], dict) or {}
2313 mrr_contents_text = try_get(mrr_contents, [lambda x: x['simpleText'], lambda x: x['runs'][0]['text']], compat_str)
2314 if not (mrr_title and mrr_contents_text):
2315 continue
2316 if mrr_title == 'License':
2317 video_license = mrr_contents_text
2318 elif not multiple_songs:
2319 if mrr_title == 'Album':
2320 album = mrr_contents_text
2321 elif mrr_title == 'Artist':
2322 artist = mrr_contents_text
2323 elif mrr_title == 'Song':
2324 track = mrr_contents_text
2325
2326 m_episode = re.search(
2327 r'<div[^>]+id="watch7-headline"[^>]*>\s*<span[^>]*>.*?>(?P<series>[^<]+)</a></b>\s*S(?P<season>\d+)\s*•\s*E(?P<episode>\d+)</span>',
2328 video_webpage)
2329 if m_episode:
2330 series = unescapeHTML(m_episode.group('series'))
2331 season_number = int(m_episode.group('season'))
2332 episode_number = int(m_episode.group('episode'))
2333 else:
2334 series = season_number = episode_number = None
2335
2336 m_cat_container = self._search_regex(
2337 r'(?s)<h4[^>]*>\s*Category\s*</h4>\s*<ul[^>]*>(.*?)</ul>',
2338 video_webpage, 'categories', default=None)
2339 category = None
2340 if m_cat_container:
2341 category = self._html_search_regex(
2342 r'(?s)<a[^<]+>(.*?)</a>', m_cat_container, 'category',
2343 default=None)
2344 if not category:
2345 category = try_get(
2346 microformat, lambda x: x['category'], compat_str)
2347 video_categories = None if category is None else [category]
2348
2349 video_tags = [
2350 unescapeHTML(m.group('content'))
2351 for m in re.finditer(self._meta_regex('og:video:tag'), video_webpage)]
2352 if not video_tags:
2353 video_tags = try_get(video_details, lambda x: x['keywords'], list)
2354
2355 def _extract_count(count_name):
2356 return str_to_int(self._search_regex(
2357 (r'-%s-button[^>]+><span[^>]+class="yt-uix-button-content"[^>]*>([\d,]+)</span>' % re.escape(count_name),
2358 r'["\']label["\']\s*:\s*["\']([\d,.]+)\s+%ss["\']' % re.escape(count_name)),
2359 video_webpage, count_name, default=None))
2360
2361 like_count = _extract_count('like')
2362 dislike_count = _extract_count('dislike')
2363
2364 if view_count is None:
2365 view_count = str_to_int(self._search_regex(
2366 r'<[^>]+class=["\']watch-view-count[^>]+>\s*([\d,\s]+)', video_webpage,
2367 'view count', default=None))
2368
2369 average_rating = (
2370 float_or_none(video_details.get('averageRating'))
2371 or try_get(video_info, lambda x: float_or_none(x['avg_rating'][0])))
2372
2373 # subtitles
2374 video_subtitles = self.extract_subtitles(
2375 video_id, video_webpage, has_live_chat_replay)
2376 automatic_captions = self.extract_automatic_captions(video_id, player_response, ytplayer_config)
2377
2378 video_duration = try_get(
2379 video_info, lambda x: int_or_none(x['length_seconds'][0]))
2380 if not video_duration:
2381 video_duration = int_or_none(video_details.get('lengthSeconds'))
2382 if not video_duration:
2383 video_duration = parse_duration(self._html_search_meta(
2384 'duration', video_webpage, 'video duration'))
2385
2386 # Get Subscriber Count of channel
2387 subscriber_count = parse_count(self._search_regex(
2388 r'"text":"([\d\.]+\w?) subscribers"',
2389 video_webpage,
2390 'subscriber count',
2391 default=None
2392 ))
2393
2394 # annotations
2395 video_annotations = None
2396 if self._downloader.params.get('writeannotations', False):
2397 xsrf_token = None
2398 ytcfg = self._extract_ytcfg(video_id, video_webpage)
2399 if ytcfg:
2400 xsrf_token = try_get(ytcfg, lambda x: x['XSRF_TOKEN'], compat_str)
2401 if not xsrf_token:
2402 xsrf_token = self._search_regex(
2403 r'([\'"])XSRF_TOKEN\1\s*:\s*([\'"])(?P<xsrf_token>(?:(?!\2).)+)\2',
2404 video_webpage, 'xsrf token', group='xsrf_token', fatal=False)
2405 invideo_url = try_get(
2406 player_response, lambda x: x['annotations'][0]['playerAnnotationsUrlsRenderer']['invideoUrl'], compat_str)
2407 if xsrf_token and invideo_url:
2408 xsrf_field_name = None
2409 if ytcfg:
2410 xsrf_field_name = try_get(ytcfg, lambda x: x['XSRF_FIELD_NAME'], compat_str)
2411 if not xsrf_field_name:
2412 xsrf_field_name = self._search_regex(
2413 r'([\'"])XSRF_FIELD_NAME\1\s*:\s*([\'"])(?P<xsrf_field_name>\w+)\2',
2414 video_webpage, 'xsrf field name',
2415 group='xsrf_field_name', default='session_token')
2416 video_annotations = self._download_webpage(
2417 self._proto_relative_url(invideo_url),
2418 video_id, note='Downloading annotations',
2419 errnote='Unable to download video annotations', fatal=False,
2420 data=urlencode_postdata({xsrf_field_name: xsrf_token}))
2421
2422 chapters = self._extract_chapters(video_webpage, description_original, video_id, video_duration)
2423
2424 # Look for the DASH manifest
2425 if self._downloader.params.get('youtube_include_dash_manifest', True):
2426 dash_mpd_fatal = True
2427 for mpd_url in dash_mpds:
2428 dash_formats = {}
2429 try:
2430 def decrypt_sig(mobj):
2431 s = mobj.group(1)
2432 dec_s = self._decrypt_signature(s, video_id, player_url, age_gate)
2433 return '/signature/%s' % dec_s
2434
2435 mpd_url = re.sub(r'/s/([a-fA-F0-9\.]+)', decrypt_sig, mpd_url)
2436
2437 for df in self._extract_mpd_formats(
2438 mpd_url, video_id, fatal=dash_mpd_fatal,
2439 formats_dict=self._formats):
2440 if not df.get('filesize'):
2441 df['filesize'] = _extract_filesize(df['url'])
2442 # Do not overwrite DASH format found in some previous DASH manifest
2443 if df['format_id'] not in dash_formats:
2444 dash_formats[df['format_id']] = df
2445 # Additional DASH manifests may end up in HTTP Error 403 therefore
2446 # allow them to fail without bug report message if we already have
2447 # some DASH manifest succeeded. This is temporary workaround to reduce
2448 # burst of bug reports until we figure out the reason and whether it
2449 # can be fixed at all.
2450 dash_mpd_fatal = False
2451 except (ExtractorError, KeyError) as e:
2452 self.report_warning(
2453 'Skipping DASH manifest: %r' % e, video_id)
2454 if dash_formats:
2455 # Remove the formats we found through non-DASH, they
2456 # contain less info and it can be wrong, because we use
2457 # fixed values (for example the resolution). See
2458 # https://github.com/ytdl-org/youtube-dl/issues/5774 for an
2459 # example.
2460 formats = [f for f in formats if f['format_id'] not in dash_formats.keys()]
2461 formats.extend(dash_formats.values())
2462
2463 # Check for malformed aspect ratio
2464 stretched_m = re.search(
2465 r'<meta\s+property="og:video:tag".*?content="yt:stretch=(?P<w>[0-9]+):(?P<h>[0-9]+)">',
2466 video_webpage)
2467 if stretched_m:
2468 w = float(stretched_m.group('w'))
2469 h = float(stretched_m.group('h'))
2470 # yt:stretch may hold invalid ratio data (e.g. for Q39EVAstoRM ratio is 17:0).
2471 # We will only process correct ratios.
2472 if w > 0 and h > 0:
2473 ratio = w / h
2474 for f in formats:
2475 if f.get('vcodec') != 'none':
2476 f['stretched_ratio'] = ratio
2477
2478 if not formats:
2479 if 'reason' in video_info:
2480 if 'The uploader has not made this video available in your country.' in video_info['reason']:
2481 regions_allowed = self._html_search_meta(
2482 'regionsAllowed', video_webpage, default=None)
2483 countries = regions_allowed.split(',') if regions_allowed else None
2484 self.raise_geo_restricted(
2485 msg=video_info['reason'][0], countries=countries)
2486 reason = video_info['reason'][0]
2487 if 'Invalid parameters' in reason:
2488 unavailable_message = extract_unavailable_message()
2489 if unavailable_message:
2490 reason = unavailable_message
2491 raise ExtractorError(
2492 'YouTube said: %s' % reason,
2493 expected=True, video_id=video_id)
2494 if video_info.get('license_info') or try_get(player_response, lambda x: x['streamingData']['licenseInfos']):
2495 raise ExtractorError('This video is DRM protected.', expected=True)
2496
2497 self._sort_formats(formats)
2498
2499 self.mark_watched(video_id, video_info, player_response)
2500
2501 return {
2502 'id': video_id,
2503 'uploader': video_uploader,
2504 'uploader_id': video_uploader_id,
2505 'uploader_url': video_uploader_url,
2506 'channel_id': channel_id,
2507 'channel_url': channel_url,
2508 'upload_date': upload_date,
2509 'license': video_license,
2510 'creator': video_creator or artist,
2511 'title': video_title,
2512 'alt_title': video_alt_title or track,
2513 'thumbnails': thumbnails,
2514 'description': video_description,
2515 'categories': video_categories,
2516 'tags': video_tags,
2517 'subtitles': video_subtitles,
2518 'automatic_captions': automatic_captions,
2519 'duration': video_duration,
2520 'age_limit': 18 if age_gate else 0,
2521 'annotations': video_annotations,
2522 'chapters': chapters,
2523 'webpage_url': proto + '://www.youtube.com/watch?v=%s' % video_id,
2524 'view_count': view_count,
2525 'like_count': like_count,
2526 'dislike_count': dislike_count,
2527 'average_rating': average_rating,
2528 'formats': formats,
2529 'is_live': is_live,
2530 'start_time': start_time,
2531 'end_time': end_time,
2532 'series': series,
2533 'season_number': season_number,
2534 'episode_number': episode_number,
2535 'track': track,
2536 'artist': artist,
2537 'album': album,
2538 'release_date': release_date,
2539 'release_year': release_year,
2540 'subscriber_count': subscriber_count,
2541 }
2542
2543
2544class YoutubeTabIE(YoutubeBaseInfoExtractor):
2545 IE_DESC = 'YouTube.com tab'
2546 _VALID_URL = r'''(?x)
2547 https?://
2548 (?:\w+\.)?
2549 (?:
2550 youtube(?:kids)?\.com|
2551 invidio\.us
2552 )/
2553 (?:
2554 (?:channel|c|user)/|
2555 (?P<not_channel>
2556 feed/|
2557 (?:playlist|watch)\?.*?\blist=
2558 )|
2559 (?!(?:%s)\b) # Direct URLs
2560 )
2561 (?P<id>[^/?\#&]+)
2562 ''' % YoutubeBaseInfoExtractor._RESERVED_NAMES
2563 IE_NAME = 'youtube:tab'
2564
2565 _TESTS = [{
2566 # playlists, multipage
2567 'url': 'https://www.youtube.com/c/ИгорьКлейнер/playlists?view=1&flow=grid',
2568 'playlist_mincount': 94,
2569 'info_dict': {
2570 'id': 'UCqj7Cz7revf5maW9g5pgNcg',
2571 'title': 'Игорь Клейнер - Playlists',
2572 'description': 'md5:be97ee0f14ee314f1f002cf187166ee2',
2573 },
2574 }, {
2575 # playlists, multipage, different order
2576 'url': 'https://www.youtube.com/user/igorkle1/playlists?view=1&sort=dd',
2577 'playlist_mincount': 94,
2578 'info_dict': {
2579 'id': 'UCqj7Cz7revf5maW9g5pgNcg',
2580 'title': 'Игорь Клейнер - Playlists',
2581 'description': 'md5:be97ee0f14ee314f1f002cf187166ee2',
2582 },
2583 }, {
2584 # playlists, singlepage
2585 'url': 'https://www.youtube.com/user/ThirstForScience/playlists',
2586 'playlist_mincount': 4,
2587 'info_dict': {
2588 'id': 'UCAEtajcuhQ6an9WEzY9LEMQ',
2589 'title': 'ThirstForScience - Playlists',
2590 'description': 'md5:609399d937ea957b0f53cbffb747a14c',
2591 }
2592 }, {
2593 'url': 'https://www.youtube.com/c/ChristophLaimer/playlists',
2594 'only_matching': True,
2595 }, {
2596 # basic, single video playlist
2597 'url': 'https://www.youtube.com/playlist?list=PL4lCao7KL_QFVb7Iudeipvc2BCavECqzc',
2598 'info_dict': {
2599 'uploader_id': 'UCmlqkdCBesrv2Lak1mF_MxA',
2600 'uploader': 'Sergey M.',
2601 'id': 'PL4lCao7KL_QFVb7Iudeipvc2BCavECqzc',
2602 'title': 'youtube-dl public playlist',
2603 },
2604 'playlist_count': 1,
2605 }, {
2606 # empty playlist
2607 'url': 'https://www.youtube.com/playlist?list=PL4lCao7KL_QFodcLWhDpGCYnngnHtQ-Xf',
2608 'info_dict': {
2609 'uploader_id': 'UCmlqkdCBesrv2Lak1mF_MxA',
2610 'uploader': 'Sergey M.',
2611 'id': 'PL4lCao7KL_QFodcLWhDpGCYnngnHtQ-Xf',
2612 'title': 'youtube-dl empty playlist',
2613 },
2614 'playlist_count': 0,
2615 }, {
2616 # Home tab
2617 'url': 'https://www.youtube.com/channel/UCKfVa3S1e4PHvxWcwyMMg8w/featured',
2618 'info_dict': {
2619 'id': 'UCKfVa3S1e4PHvxWcwyMMg8w',
2620 'title': 'lex will - Home',
2621 'description': 'md5:2163c5d0ff54ed5f598d6a7e6211e488',
2622 },
2623 'playlist_mincount': 2,
2624 }, {
2625 # Videos tab
2626 'url': 'https://www.youtube.com/channel/UCKfVa3S1e4PHvxWcwyMMg8w/videos',
2627 'info_dict': {
2628 'id': 'UCKfVa3S1e4PHvxWcwyMMg8w',
2629 'title': 'lex will - Videos',
2630 'description': 'md5:2163c5d0ff54ed5f598d6a7e6211e488',
2631 },
2632 'playlist_mincount': 975,
2633 }, {
2634 # Videos tab, sorted by popular
2635 'url': 'https://www.youtube.com/channel/UCKfVa3S1e4PHvxWcwyMMg8w/videos?view=0&sort=p&flow=grid',
2636 'info_dict': {
2637 'id': 'UCKfVa3S1e4PHvxWcwyMMg8w',
2638 'title': 'lex will - Videos',
2639 'description': 'md5:2163c5d0ff54ed5f598d6a7e6211e488',
2640 },
2641 'playlist_mincount': 199,
2642 }, {
2643 # Playlists tab
2644 'url': 'https://www.youtube.com/channel/UCKfVa3S1e4PHvxWcwyMMg8w/playlists',
2645 'info_dict': {
2646 'id': 'UCKfVa3S1e4PHvxWcwyMMg8w',
2647 'title': 'lex will - Playlists',
2648 'description': 'md5:2163c5d0ff54ed5f598d6a7e6211e488',
2649 },
2650 'playlist_mincount': 17,
2651 }, {
2652 # Community tab
2653 'url': 'https://www.youtube.com/channel/UCKfVa3S1e4PHvxWcwyMMg8w/community',
2654 'info_dict': {
2655 'id': 'UCKfVa3S1e4PHvxWcwyMMg8w',
2656 'title': 'lex will - Community',
2657 'description': 'md5:2163c5d0ff54ed5f598d6a7e6211e488',
2658 },
2659 'playlist_mincount': 18,
2660 }, {
2661 # Channels tab
2662 'url': 'https://www.youtube.com/channel/UCKfVa3S1e4PHvxWcwyMMg8w/channels',
2663 'info_dict': {
2664 'id': 'UCKfVa3S1e4PHvxWcwyMMg8w',
2665 'title': 'lex will - Channels',
2666 'description': 'md5:2163c5d0ff54ed5f598d6a7e6211e488',
2667 },
2668 'playlist_mincount': 138,
2669 }, {
2670 'url': 'https://invidio.us/channel/UCmlqkdCBesrv2Lak1mF_MxA',
2671 'only_matching': True,
2672 }, {
2673 'url': 'https://www.youtubekids.com/channel/UCmlqkdCBesrv2Lak1mF_MxA',
2674 'only_matching': True,
2675 }, {
2676 'url': 'https://music.youtube.com/channel/UCmlqkdCBesrv2Lak1mF_MxA',
2677 'only_matching': True,
2678 }, {
2679 'note': 'Playlist with deleted videos (#651). As a bonus, the video #51 is also twice in this list.',
2680 'url': 'https://www.youtube.com/playlist?list=PLwP_SiAcdui0KVebT0mU9Apz359a4ubsC',
2681 'info_dict': {
2682 'title': '29C3: Not my department',
2683 'id': 'PLwP_SiAcdui0KVebT0mU9Apz359a4ubsC',
2684 'uploader': 'Christiaan008',
2685 'uploader_id': 'UCEPzS1rYsrkqzSLNp76nrcg',
2686 },
2687 'playlist_count': 96,
2688 }, {
2689 'note': 'Large playlist',
2690 'url': 'https://www.youtube.com/playlist?list=UUBABnxM4Ar9ten8Mdjj1j0Q',
2691 'info_dict': {
2692 'title': 'Uploads from Cauchemar',
2693 'id': 'UUBABnxM4Ar9ten8Mdjj1j0Q',
2694 'uploader': 'Cauchemar',
2695 'uploader_id': 'UCBABnxM4Ar9ten8Mdjj1j0Q',
2696 },
2697 'playlist_mincount': 1123,
2698 }, {
2699 # even larger playlist, 8832 videos
2700 'url': 'http://www.youtube.com/user/NASAgovVideo/videos',
2701 'only_matching': True,
2702 }, {
2703 'note': 'Buggy playlist: the webpage has a "Load more" button but it doesn\'t have more videos',
2704 'url': 'https://www.youtube.com/playlist?list=UUXw-G3eDE9trcvY2sBMM_aA',
2705 'info_dict': {
2706 'title': 'Uploads from Interstellar Movie',
2707 'id': 'UUXw-G3eDE9trcvY2sBMM_aA',
2708 'uploader': 'Interstellar Movie',
2709 'uploader_id': 'UCXw-G3eDE9trcvY2sBMM_aA',
2710 },
2711 'playlist_mincount': 21,
2712 }, {
2713 # https://github.com/ytdl-org/youtube-dl/issues/21844
2714 'url': 'https://www.youtube.com/playlist?list=PLzH6n4zXuckpfMu_4Ff8E7Z1behQks5ba',
2715 'info_dict': {
2716 'title': 'Data Analysis with Dr Mike Pound',
2717 'id': 'PLzH6n4zXuckpfMu_4Ff8E7Z1behQks5ba',
2718 'uploader_id': 'UC9-y-6csu5WGm29I7JiwpnA',
2719 'uploader': 'Computerphile',
2720 },
2721 'playlist_mincount': 11,
2722 }, {
2723 'url': 'https://invidio.us/playlist?list=PL4lCao7KL_QFVb7Iudeipvc2BCavECqzc',
2724 'only_matching': True,
2725 }, {
2726 # Playlist URL that does not actually serve a playlist
2727 'url': 'https://www.youtube.com/watch?v=FqZTN594JQw&list=PLMYEtVRpaqY00V9W81Cwmzp6N6vZqfUKD4',
2728 'info_dict': {
2729 'id': 'FqZTN594JQw',
2730 'ext': 'webm',
2731 'title': "Smiley's People 01 detective, Adventure Series, Action",
2732 'uploader': 'STREEM',
2733 'uploader_id': 'UCyPhqAZgwYWZfxElWVbVJng',
2734 'uploader_url': r're:https?://(?:www\.)?youtube\.com/channel/UCyPhqAZgwYWZfxElWVbVJng',
2735 'upload_date': '20150526',
2736 'license': 'Standard YouTube License',
2737 'description': 'md5:507cdcb5a49ac0da37a920ece610be80',
2738 'categories': ['People & Blogs'],
2739 'tags': list,
2740 'view_count': int,
2741 'like_count': int,
2742 'dislike_count': int,
2743 },
2744 'params': {
2745 'skip_download': True,
2746 },
2747 'skip': 'This video is not available.',
2748 'add_ie': [YoutubeIE.ie_key()],
2749 }, {
2750 'url': 'https://www.youtubekids.com/watch?v=Agk7R8I8o5U&list=PUZ6jURNr1WQZCNHF0ao-c0g',
2751 'only_matching': True,
2752 }, {
2753 'url': 'https://www.youtube.com/watch?v=MuAGGZNfUkU&list=RDMM',
2754 'only_matching': True,
2755 }, {
2756 'url': 'https://www.youtube.com/channel/UCoMdktPbSTixAyNGwb-UYkQ/live',
2757 'info_dict': {
2758 'id': '9Auq9mYxFEE',
2759 'ext': 'mp4',
2760 'title': 'Watch Sky News live',
2761 'uploader': 'Sky News',
2762 'uploader_id': 'skynews',
2763 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/skynews',
2764 'upload_date': '20191102',
2765 'description': 'md5:78de4e1c2359d0ea3ed829678e38b662',
2766 'categories': ['News & Politics'],
2767 'tags': list,
2768 'like_count': int,
2769 'dislike_count': int,
2770 },
2771 'params': {
2772 'skip_download': True,
2773 },
2774 }, {
2775 'url': 'https://www.youtube.com/user/TheYoungTurks/live',
2776 'info_dict': {
2777 'id': 'a48o2S1cPoo',
2778 'ext': 'mp4',
2779 'title': 'The Young Turks - Live Main Show',
2780 'uploader': 'The Young Turks',
2781 'uploader_id': 'TheYoungTurks',
2782 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/TheYoungTurks',
2783 'upload_date': '20150715',
2784 'license': 'Standard YouTube License',
2785 'description': 'md5:438179573adcdff3c97ebb1ee632b891',
2786 'categories': ['News & Politics'],
2787 'tags': ['Cenk Uygur (TV Program Creator)', 'The Young Turks (Award-Winning Work)', 'Talk Show (TV Genre)'],
2788 'like_count': int,
2789 'dislike_count': int,
2790 },
2791 'params': {
2792 'skip_download': True,
2793 },
2794 'only_matching': True,
2795 }, {
2796 'url': 'https://www.youtube.com/channel/UC1yBKRuGpC1tSM73A0ZjYjQ/live',
2797 'only_matching': True,
2798 }, {
2799 'url': 'https://www.youtube.com/c/CommanderVideoHq/live',
2800 'only_matching': True,
2801 }, {
2802 'url': 'https://www.youtube.com/feed/trending',
2803 'only_matching': True,
2804 }, {
2805 # needs auth
2806 'url': 'https://www.youtube.com/feed/library',
2807 'only_matching': True,
2808 }, {
2809 # needs auth
2810 'url': 'https://www.youtube.com/feed/history',
2811 'only_matching': True,
2812 }, {
2813 # needs auth
2814 'url': 'https://www.youtube.com/feed/subscriptions',
2815 'only_matching': True,
2816 }, {
2817 # needs auth
2818 'url': 'https://www.youtube.com/feed/watch_later',
2819 'only_matching': True,
2820 }, {
2821 # no longer available?
2822 'url': 'https://www.youtube.com/feed/recommended',
2823 'only_matching': True,
2824 }, {
2825 # inline playlist with not always working continuations
2826 'url': 'https://www.youtube.com/watch?v=UC6u0Tct-Fo&list=PL36D642111D65BE7C',
2827 'only_matching': True,
2828 }, {
2829 'url': 'https://www.youtube.com/course?list=ECUl4u3cNGP61MdtwGTqZA0MreSaDybji8',
2830 'only_matching': True,
2831 }, {
2832 'url': 'https://www.youtube.com/course',
2833 'only_matching': True,
2834 }, {
2835 'url': 'https://www.youtube.com/zsecurity',
2836 'only_matching': True,
2837 }, {
2838 'url': 'http://www.youtube.com/NASAgovVideo/videos',
2839 'only_matching': True,
2840 }, {
2841 'url': 'https://www.youtube.com/TheYoungTurks/live',
2842 'only_matching': True,
2843 }]
2844
2845 @classmethod
2846 def suitable(cls, url):
2847 return False if YoutubeIE.suitable(url) else super(
2848 YoutubeTabIE, cls).suitable(url)
2849
2850 def _extract_channel_id(self, webpage):
2851 channel_id = self._html_search_meta(
2852 'channelId', webpage, 'channel id', default=None)
2853 if channel_id:
2854 return channel_id
2855 channel_url = self._html_search_meta(
2856 ('og:url', 'al:ios:url', 'al:android:url', 'al:web:url',
2857 'twitter:url', 'twitter:app:url:iphone', 'twitter:app:url:ipad',
2858 'twitter:app:url:googleplay'), webpage, 'channel url')
2859 return self._search_regex(
2860 r'https?://(?:www\.)?youtube\.com/channel/([^/?#&])+',
2861 channel_url, 'channel id')
2862
2863 @staticmethod
2864 def _extract_grid_item_renderer(item):
2865 for item_kind in ('Playlist', 'Video', 'Channel'):
2866 renderer = item.get('grid%sRenderer' % item_kind)
2867 if renderer:
2868 return renderer
2869
2870 def _extract_video(self, renderer):
2871 video_id = renderer.get('videoId')
2872 title = try_get(
2873 renderer,
2874 (lambda x: x['title']['runs'][0]['text'],
2875 lambda x: x['title']['simpleText']), compat_str)
2876 description = try_get(
2877 renderer, lambda x: x['descriptionSnippet']['runs'][0]['text'],
2878 compat_str)
2879 duration = parse_duration(try_get(
2880 renderer, lambda x: x['lengthText']['simpleText'], compat_str))
2881 view_count_text = try_get(
2882 renderer, lambda x: x['viewCountText']['simpleText'], compat_str) or ''
2883 view_count = str_to_int(self._search_regex(
2884 r'^([\d,]+)', re.sub(r'\s', '', view_count_text),
2885 'view count', default=None))
2886 uploader = try_get(
2887 renderer, lambda x: x['ownerText']['runs'][0]['text'], compat_str)
2888 return {
2889 '_type': 'url_transparent',
2890 'ie_key': YoutubeIE.ie_key(),
2891 'id': video_id,
2892 'url': video_id,
2893 'title': title,
2894 'description': description,
2895 'duration': duration,
2896 'view_count': view_count,
2897 'uploader': uploader,
2898 }
2899
2900 def _grid_entries(self, grid_renderer):
2901 for item in grid_renderer['items']:
2902 if not isinstance(item, dict):
2903 continue
2904 renderer = self._extract_grid_item_renderer(item)
2905 if not isinstance(renderer, dict):
2906 continue
2907 title = try_get(
2908 renderer, lambda x: x['title']['runs'][0]['text'], compat_str)
2909 # playlist
2910 playlist_id = renderer.get('playlistId')
2911 if playlist_id:
2912 yield self.url_result(
2913 'https://www.youtube.com/playlist?list=%s' % playlist_id,
2914 ie=YoutubeTabIE.ie_key(), video_id=playlist_id,
2915 video_title=title)
2916 # video
2917 video_id = renderer.get('videoId')
2918 if video_id:
2919 yield self._extract_video(renderer)
2920 # channel
2921 channel_id = renderer.get('channelId')
2922 if channel_id:
2923 title = try_get(
2924 renderer, lambda x: x['title']['simpleText'], compat_str)
2925 yield self.url_result(
2926 'https://www.youtube.com/channel/%s' % channel_id,
2927 ie=YoutubeTabIE.ie_key(), video_title=title)
2928
2929 def _shelf_entries_from_content(self, shelf_renderer):
2930 content = shelf_renderer.get('content')
2931 if not isinstance(content, dict):
2932 return
2933 renderer = content.get('gridRenderer')
2934 if renderer:
2935 # TODO: add support for nested playlists so each shelf is processed
2936 # as separate playlist
2937 # TODO: this includes only first N items
2938 for entry in self._grid_entries(renderer):
2939 yield entry
2940 renderer = content.get('horizontalListRenderer')
2941 if renderer:
2942 # TODO
2943 pass
2944
2945 def _shelf_entries(self, shelf_renderer, skip_channels=False):
2946 ep = try_get(
2947 shelf_renderer, lambda x: x['endpoint']['commandMetadata']['webCommandMetadata']['url'],
2948 compat_str)
2949 shelf_url = urljoin('https://www.youtube.com', ep)
2950 if shelf_url:
2951 # Skipping links to another channels, note that checking for
2952 # endpoint.commandMetadata.webCommandMetadata.webPageTypwebPageType == WEB_PAGE_TYPE_CHANNEL
2953 # will not work
2954 if skip_channels and '/channels?' in shelf_url:
2955 return
2956 title = try_get(
2957 shelf_renderer, lambda x: x['title']['runs'][0]['text'], compat_str)
2958 yield self.url_result(shelf_url, video_title=title)
2959 # Shelf may not contain shelf URL, fallback to extraction from content
2960 for entry in self._shelf_entries_from_content(shelf_renderer):
2961 yield entry
2962
2963 def _playlist_entries(self, video_list_renderer):
2964 for content in video_list_renderer['contents']:
2965 if not isinstance(content, dict):
2966 continue
2967 renderer = content.get('playlistVideoRenderer') or content.get('playlistPanelVideoRenderer')
2968 if not isinstance(renderer, dict):
2969 continue
2970 video_id = renderer.get('videoId')
2971 if not video_id:
2972 continue
2973 yield self._extract_video(renderer)
2974
2975 r""" # Not needed in the new implementation
2976 def _itemSection_entries(self, item_sect_renderer):
2977 for content in item_sect_renderer['contents']:
2978 if not isinstance(content, dict):
2979 continue
2980 renderer = content.get('videoRenderer', {})
2981 if not isinstance(renderer, dict):
2982 continue
2983 video_id = renderer.get('videoId')
2984 if not video_id:
2985 continue
2986 yield self._extract_video(renderer)
2987 """
2988
2989 def _rich_entries(self, rich_grid_renderer):
2990 renderer = try_get(
2991 rich_grid_renderer, lambda x: x['content']['videoRenderer'], dict) or {}
2992 video_id = renderer.get('videoId')
2993 if not video_id:
2994 return
2995 yield self._extract_video(renderer)
2996
2997 def _video_entry(self, video_renderer):
2998 video_id = video_renderer.get('videoId')
2999 if video_id:
3000 return self._extract_video(video_renderer)
3001
3002 def _post_thread_entries(self, post_thread_renderer):
3003 post_renderer = try_get(
3004 post_thread_renderer, lambda x: x['post']['backstagePostRenderer'], dict)
3005 if not post_renderer:
3006 return
3007 # video attachment
3008 video_renderer = try_get(
3009 post_renderer, lambda x: x['backstageAttachment']['videoRenderer'], dict)
3010 video_id = None
3011 if video_renderer:
3012 entry = self._video_entry(video_renderer)
3013 if entry:
3014 yield entry
3015 # inline video links
3016 runs = try_get(post_renderer, lambda x: x['contentText']['runs'], list) or []
3017 for run in runs:
3018 if not isinstance(run, dict):
3019 continue
3020 ep_url = try_get(
3021 run, lambda x: x['navigationEndpoint']['urlEndpoint']['url'], compat_str)
3022 if not ep_url:
3023 continue
3024 if not YoutubeIE.suitable(ep_url):
3025 continue
3026 ep_video_id = YoutubeIE._match_id(ep_url)
3027 if video_id == ep_video_id:
3028 continue
3029 yield self.url_result(ep_url, ie=YoutubeIE.ie_key(), video_id=video_id)
3030
3031 def _post_thread_continuation_entries(self, post_thread_continuation):
3032 contents = post_thread_continuation.get('contents')
3033 if not isinstance(contents, list):
3034 return
3035 for content in contents:
3036 renderer = content.get('backstagePostThreadRenderer')
3037 if not isinstance(renderer, dict):
3038 continue
3039 for entry in self._post_thread_entries(renderer):
3040 yield entry
3041
3042 @staticmethod
3043 def _build_continuation_query(continuation, ctp=None):
3044 query = {
3045 'ctoken': continuation,
3046 'continuation': continuation,
3047 }
3048 if ctp:
3049 query['itct'] = ctp
3050 return query
3051
3052 @staticmethod
3053 def _extract_next_continuation_data(renderer):
3054 next_continuation = try_get(
3055 renderer, lambda x: x['continuations'][0]['nextContinuationData'], dict)
3056 if not next_continuation:
3057 return
3058 continuation = next_continuation.get('continuation')
3059 if not continuation:
3060 return
3061 ctp = next_continuation.get('clickTrackingParams')
3062 return YoutubeTabIE._build_continuation_query(continuation, ctp)
3063
3064 @classmethod
3065 def _extract_continuation(cls, renderer):
3066 next_continuation = cls._extract_next_continuation_data(renderer)
3067 if next_continuation:
3068 return next_continuation
3069 contents = renderer.get('contents')
3070 if not isinstance(contents, list):
3071 return
3072 for content in contents:
3073 if not isinstance(content, dict):
3074 continue
3075 continuation_ep = try_get(
3076 content, lambda x: x['continuationItemRenderer']['continuationEndpoint'],
3077 dict)
3078 if not continuation_ep:
3079 continue
3080 continuation = try_get(
3081 continuation_ep, lambda x: x['continuationCommand']['token'], compat_str)
3082 if not continuation:
3083 continue
3084 ctp = continuation_ep.get('clickTrackingParams')
3085 return YoutubeTabIE._build_continuation_query(continuation, ctp)
3086
3087 def _entries(self, tab, identity_token):
3088
3089 def extract_entries(parent_renderer): # this needs to called again for continuation to work with feeds
3090 contents = try_get(parent_renderer, lambda x: x['contents'], list) or []
3091 for content in contents:
3092 if not isinstance(content, dict):
3093 continue
3094 is_renderer = try_get(content, lambda x: x['itemSectionRenderer'], dict)
3095 if not is_renderer:
3096 renderer = content.get('richItemRenderer')
3097 if renderer:
3098 for entry in self._rich_entries(renderer):
3099 yield entry
3100 continuation_list[0] = self._extract_continuation(parent_renderer)
3101 continue
3102 isr_contents = try_get(is_renderer, lambda x: x['contents'], list) or []
3103 for isr_content in isr_contents:
3104 if not isinstance(isr_content, dict):
3105 continue
3106 renderer = isr_content.get('playlistVideoListRenderer')
3107 if renderer:
3108 for entry in self._playlist_entries(renderer):
3109 yield entry
3110 continuation_list[0] = self._extract_continuation(renderer)
3111 continue
3112 renderer = isr_content.get('gridRenderer')
3113 if renderer:
3114 for entry in self._grid_entries(renderer):
3115 yield entry
3116 continuation_list[0] = self._extract_continuation(renderer)
3117 continue
3118 renderer = isr_content.get('shelfRenderer')
3119 if renderer:
3120 is_channels_tab = tab.get('title') == 'Channels'
3121 for entry in self._shelf_entries(renderer, not is_channels_tab):
3122 yield entry
3123 continue
3124 renderer = isr_content.get('backstagePostThreadRenderer')
3125 if renderer:
3126 for entry in self._post_thread_entries(renderer):
3127 yield entry
3128 continuation_list[0] = self._extract_continuation(renderer)
3129 continue
3130 renderer = isr_content.get('videoRenderer')
3131 if renderer:
3132 entry = self._video_entry(renderer)
3133 if entry:
3134 yield entry
3135
3136 if not continuation_list[0]:
3137 continuation_list[0] = self._extract_continuation(is_renderer)
3138
3139 if not continuation_list[0]:
3140 continuation_list[0] = self._extract_continuation(parent_renderer)
3141
3142 continuation_list = [None] # Python 2 doesnot support nonlocal
3143 tab_content = try_get(tab, lambda x: x['content'], dict)
3144 if not tab_content:
3145 return
3146 parent_renderer = (
3147 try_get(tab_content, lambda x: x['sectionListRenderer'], dict)
3148 or try_get(tab_content, lambda x: x['richGridRenderer'], dict) or {})
3149 for entry in extract_entries(parent_renderer):
3150 yield entry
3151 continuation = continuation_list[0]
3152
3153 headers = {
3154 'x-youtube-client-name': '1',
3155 'x-youtube-client-version': '2.20201112.04.01',
3156 }
3157 if identity_token:
3158 headers['x-youtube-identity-token'] = identity_token
3159
3160 for page_num in itertools.count(1):
3161 if not continuation:
3162 break
3163 count = 0
3164 retries = 3
3165 while count <= retries:
3166 try:
3167 # Downloading page may result in intermittent 5xx HTTP error
3168 # that is usually worked around with a retry
3169 browse = self._download_json(
3170 'https://www.youtube.com/browse_ajax', None,
3171 'Downloading page %d%s'
3172 % (page_num, ' (retry #%d)' % count if count else ''),
3173 headers=headers, query=continuation)
3174 break
3175 except ExtractorError as e:
3176 if isinstance(e.cause, compat_HTTPError) and e.cause.code in (500, 503):
3177 count += 1
3178 if count <= retries:
3179 continue
3180 raise
3181 if not browse:
3182 break
3183 response = try_get(browse, lambda x: x[1]['response'], dict)
3184 if not response:
3185 break
3186
3187 continuation_contents = try_get(
3188 response, lambda x: x['continuationContents'], dict)
3189 if continuation_contents:
3190 continuation_renderer = continuation_contents.get('playlistVideoListContinuation')
3191 if continuation_renderer:
3192 for entry in self._playlist_entries(continuation_renderer):
3193 yield entry
3194 continuation = self._extract_continuation(continuation_renderer)
3195 continue
3196 continuation_renderer = continuation_contents.get('gridContinuation')
3197 if continuation_renderer:
3198 for entry in self._grid_entries(continuation_renderer):
3199 yield entry
3200 continuation = self._extract_continuation(continuation_renderer)
3201 continue
3202 continuation_renderer = continuation_contents.get('itemSectionContinuation')
3203 if continuation_renderer:
3204 for entry in self._post_thread_continuation_entries(continuation_renderer):
3205 yield entry
3206 continuation = self._extract_continuation(continuation_renderer)
3207 continue
3208 continuation_renderer = continuation_contents.get('sectionListContinuation') # for feeds
3209 if continuation_renderer:
3210 continuation_list = [None]
3211 for entry in extract_entries(continuation_renderer):
3212 yield entry
3213 continuation = continuation_list[0]
3214 continue
3215
3216 continuation_items = try_get(
3217 response, lambda x: x['onResponseReceivedActions'][0]['appendContinuationItemsAction']['continuationItems'], list)
3218 if continuation_items:
3219 continuation_item = continuation_items[0]
3220 if not isinstance(continuation_item, dict):
3221 continue
3222 renderer = continuation_item.get('playlistVideoRenderer') or continuation_item.get('itemSectionRenderer')
3223 if renderer:
3224 video_list_renderer = {'contents': continuation_items}
3225 for entry in self._playlist_entries(video_list_renderer):
3226 yield entry
3227 continuation = self._extract_continuation(video_list_renderer)
3228 continue
3229 break
3230
3231 @staticmethod
3232 def _extract_selected_tab(tabs):
3233 for tab in tabs:
3234 if try_get(tab, lambda x: x['tabRenderer']['selected'], bool):
3235 return tab['tabRenderer']
3236 else:
3237 raise ExtractorError('Unable to find selected tab')
3238
3239 @staticmethod
3240 def _extract_uploader(data):
3241 uploader = {}
3242 sidebar_renderer = try_get(
3243 data, lambda x: x['sidebar']['playlistSidebarRenderer']['items'], list)
3244 if sidebar_renderer:
3245 for item in sidebar_renderer:
3246 if not isinstance(item, dict):
3247 continue
3248 renderer = item.get('playlistSidebarSecondaryInfoRenderer')
3249 if not isinstance(renderer, dict):
3250 continue
3251 owner = try_get(
3252 renderer, lambda x: x['videoOwner']['videoOwnerRenderer']['title']['runs'][0], dict)
3253 if owner:
3254 uploader['uploader'] = owner.get('text')
3255 uploader['uploader_id'] = try_get(
3256 owner, lambda x: x['navigationEndpoint']['browseEndpoint']['browseId'], compat_str)
3257 uploader['uploader_url'] = urljoin(
3258 'https://www.youtube.com/',
3259 try_get(owner, lambda x: x['navigationEndpoint']['browseEndpoint']['canonicalBaseUrl'], compat_str))
3260 return uploader
3261
3262 def _extract_from_tabs(self, item_id, webpage, data, tabs, identity_token):
3263 selected_tab = self._extract_selected_tab(tabs)
3264 renderer = try_get(
3265 data, lambda x: x['metadata']['channelMetadataRenderer'], dict)
3266 playlist_id = title = description = None
3267 if renderer:
3268 channel_title = renderer.get('title') or item_id
3269 tab_title = selected_tab.get('title')
3270 title = channel_title or item_id
3271 if tab_title:
3272 title += ' - %s' % tab_title
3273 description = renderer.get('description')
3274 playlist_id = renderer.get('externalId')
3275 renderer = try_get(
3276 data, lambda x: x['metadata']['playlistMetadataRenderer'], dict)
3277 if renderer:
3278 title = renderer.get('title')
3279 description = None
3280 playlist_id = item_id
3281 if playlist_id is None:
3282 playlist_id = item_id
3283 if title is None:
3284 title = "Youtube " + playlist_id.title()
3285 playlist = self.playlist_result(
3286 self._entries(selected_tab, identity_token),
3287 playlist_id=playlist_id, playlist_title=title,
3288 playlist_description=description)
3289 playlist.update(self._extract_uploader(data))
3290 return playlist
3291
3292 def _extract_from_playlist(self, item_id, url, data, playlist):
3293 title = playlist.get('title') or try_get(
3294 data, lambda x: x['titleText']['simpleText'], compat_str)
3295 playlist_id = playlist.get('playlistId') or item_id
3296 # Inline playlist rendition continuation does not always work
3297 # at Youtube side, so delegating regular tab-based playlist URL
3298 # processing whenever possible.
3299 playlist_url = urljoin(url, try_get(
3300 playlist, lambda x: x['endpoint']['commandMetadata']['webCommandMetadata']['url'],
3301 compat_str))
3302 if playlist_url and playlist_url != url:
3303 return self.url_result(
3304 playlist_url, ie=YoutubeTabIE.ie_key(), video_id=playlist_id,
3305 video_title=title)
3306 return self.playlist_result(
3307 self._playlist_entries(playlist), playlist_id=playlist_id,
3308 playlist_title=title)
3309
3310 @staticmethod
3311 def _extract_alerts(data):
3312 for alert_dict in try_get(data, lambda x: x['alerts'], list) or []:
3313 if not isinstance(alert_dict, dict):
3314 continue
3315 for renderer in alert_dict:
3316 alert = alert_dict[renderer]
3317 alert_type = alert.get('type')
3318 if not alert_type:
3319 continue
3320 message = try_get(alert, lambda x: x['text']['simpleText'], compat_str)
3321 if message:
3322 yield alert_type, message
3323 for run in try_get(alert, lambda x: x['text']['runs'], list) or []:
3324 message = try_get(run, lambda x: x['text'], compat_str)
3325 if message:
3326 yield alert_type, message
3327
3328 def _extract_identity_token(self, webpage, item_id):
3329 ytcfg = self._extract_ytcfg(item_id, webpage)
3330 if ytcfg:
3331 token = try_get(ytcfg, lambda x: x['ID_TOKEN'], compat_str)
3332 if token:
3333 return token
3334 return self._search_regex(
3335 r'\bID_TOKEN["\']\s*:\s*["\'](.+?)["\']', webpage,
3336 'identity token', default=None)
3337
3338 def _real_extract(self, url):
3339 item_id = self._match_id(url)
3340 url = compat_urlparse.urlunparse(
3341 compat_urlparse.urlparse(url)._replace(netloc='www.youtube.com'))
3342 is_home = re.match(r'(?P<pre>%s)(?P<post>/?(?![^#?]).*$)' % self._VALID_URL, url)
3343 if is_home is not None and is_home.group('not_channel') is None and item_id != 'feed':
3344 self._downloader.report_warning(
3345 'A channel/user page was given. All the channel\'s videos will be downloaded. '
3346 'To download only the videos in the home page, add a "/featured" to the URL')
3347 url = '%s/videos%s' % (is_home.group('pre'), is_home.group('post') or '')
3348
3349 # Handle both video/playlist URLs
3350 qs = compat_urlparse.parse_qs(compat_urlparse.urlparse(url).query)
3351 video_id = qs.get('v', [None])[0]
3352 playlist_id = qs.get('list', [None])[0]
3353
3354 if is_home is not None and is_home.group('not_channel') is not None and is_home.group('not_channel').startswith('watch') and not video_id:
3355 if playlist_id:
3356 self._downloader.report_warning('%s is not a valid Youtube URL. Trying to download playlist %s' % (url, playlist_id))
3357 url = 'https://www.youtube.com/playlist?list=%s' % playlist_id
3358 # return self.url_result(playlist_id, ie=YoutubePlaylistIE.ie_key())
3359 else:
3360 raise ExtractorError('Unable to recognize tab page')
3361 if video_id and playlist_id:
3362 if self._downloader.params.get('noplaylist'):
3363 self.to_screen('Downloading just video %s because of --no-playlist' % video_id)
3364 return self.url_result(video_id, ie=YoutubeIE.ie_key(), video_id=video_id)
3365 self.to_screen('Downloading playlist %s - add --no-playlist to just download video %s' % (playlist_id, video_id))
3366
3367 webpage = self._download_webpage(url, item_id)
3368 identity_token = self._extract_identity_token(webpage, item_id)
3369 data = self._extract_yt_initial_data(item_id, webpage)
3370 for alert_type, alert_message in self._extract_alerts(data):
3371 self._downloader.report_warning('YouTube said: %s - %s' % (alert_type, alert_message))
3372 tabs = try_get(
3373 data, lambda x: x['contents']['twoColumnBrowseResultsRenderer']['tabs'], list)
3374 if tabs:
3375 return self._extract_from_tabs(item_id, webpage, data, tabs, identity_token)
3376 playlist = try_get(
3377 data, lambda x: x['contents']['twoColumnWatchNextResults']['playlist']['playlist'], dict)
3378 if playlist:
3379 return self._extract_from_playlist(item_id, url, data, playlist)
3380 # Fallback to video extraction if no playlist alike page is recognized.
3381 # First check for the current video then try the v attribute of URL query.
3382 video_id = try_get(
3383 data, lambda x: x['currentVideoEndpoint']['watchEndpoint']['videoId'],
3384 compat_str) or video_id
3385 if video_id:
3386 return self.url_result(video_id, ie=YoutubeIE.ie_key(), video_id=video_id)
3387 # Failed to recognize
3388 raise ExtractorError('Unable to recognize tab page')
3389
3390
3391class YoutubePlaylistIE(InfoExtractor):
3392 IE_DESC = 'YouTube.com playlists'
3393 _VALID_URL = r'''(?x)(?:
3394 (?:https?://)?
3395 (?:\w+\.)?
3396 (?:
3397 (?:
3398 youtube(?:kids)?\.com|
3399 invidio\.us
3400 )
3401 /.*?\?.*?\blist=
3402 )?
3403 (?P<id>%(playlist_id)s)
3404 )''' % {'playlist_id': YoutubeBaseInfoExtractor._PLAYLIST_ID_RE}
3405 IE_NAME = 'youtube:playlist'
3406 _TESTS = [{
3407 'note': 'issue #673',
3408 'url': 'PLBB231211A4F62143',
3409 'info_dict': {
3410 'title': '[OLD]Team Fortress 2 (Class-based LP)',
3411 'id': 'PLBB231211A4F62143',
3412 'uploader': 'Wickydoo',
3413 'uploader_id': 'UCKSpbfbl5kRQpTdL7kMc-1Q',
3414 },
3415 'playlist_mincount': 29,
3416 }, {
3417 'url': 'PLtPgu7CB4gbY9oDN3drwC3cMbJggS7dKl',
3418 'info_dict': {
3419 'title': 'YDL_safe_search',
3420 'id': 'PLtPgu7CB4gbY9oDN3drwC3cMbJggS7dKl',
3421 },
3422 'playlist_count': 2,
3423 'skip': 'This playlist is private',
3424 }, {
3425 'note': 'embedded',
3426 'url': 'https://www.youtube.com/embed/videoseries?list=PL6IaIsEjSbf96XFRuNccS_RuEXwNdsoEu',
3427 'playlist_count': 4,
3428 'info_dict': {
3429 'title': 'JODA15',
3430 'id': 'PL6IaIsEjSbf96XFRuNccS_RuEXwNdsoEu',
3431 'uploader': 'milan',
3432 'uploader_id': 'UCEI1-PVPcYXjB73Hfelbmaw',
3433 }
3434 }, {
3435 'url': 'http://www.youtube.com/embed/_xDOZElKyNU?list=PLsyOSbh5bs16vubvKePAQ1x3PhKavfBIl',
3436 'playlist_mincount': 982,
3437 'info_dict': {
3438 'title': '2018 Chinese New Singles (11/6 updated)',
3439 'id': 'PLsyOSbh5bs16vubvKePAQ1x3PhKavfBIl',
3440 'uploader': 'LBK',
3441 'uploader_id': 'UC21nz3_MesPLqtDqwdvnoxA',
3442 }
3443 }, {
3444 'url': 'TLGGrESM50VT6acwMjAyMjAxNw',
3445 'only_matching': True,
3446 }, {
3447 # music album playlist
3448 'url': 'OLAK5uy_m4xAFdmMC5rX3Ji3g93pQe3hqLZw_9LhM',
3449 'only_matching': True,
3450 }]
3451
3452 @classmethod
3453 def suitable(cls, url):
3454 return False if YoutubeTabIE.suitable(url) else super(
3455 YoutubePlaylistIE, cls).suitable(url)
3456
3457 def _real_extract(self, url):
3458 playlist_id = self._match_id(url)
3459 qs = compat_urlparse.parse_qs(compat_urlparse.urlparse(url).query)
3460 if not qs:
3461 qs = {'list': playlist_id}
3462 return self.url_result(
3463 update_url_query('https://www.youtube.com/playlist', qs),
3464 ie=YoutubeTabIE.ie_key(), video_id=playlist_id)
3465
3466
3467class YoutubeYtBeIE(InfoExtractor):
3468 IE_DESC = 'youtu.be'
3469 _VALID_URL = r'https?://youtu\.be/(?P<id>[0-9A-Za-z_-]{11})/*?.*?\blist=(?P<playlist_id>%(playlist_id)s)' % {'playlist_id': YoutubeBaseInfoExtractor._PLAYLIST_ID_RE}
3470 _TESTS = [{
3471 'url': 'https://youtu.be/yeWKywCrFtk?list=PL2qgrgXsNUG5ig9cat4ohreBjYLAPC0J5',
3472 'info_dict': {
3473 'id': 'yeWKywCrFtk',
3474 'ext': 'mp4',
3475 'title': 'Small Scale Baler and Braiding Rugs',
3476 'uploader': 'Backus-Page House Museum',
3477 'uploader_id': 'backuspagemuseum',
3478 'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/backuspagemuseum',
3479 'upload_date': '20161008',
3480 'description': 'md5:800c0c78d5eb128500bffd4f0b4f2e8a',
3481 'categories': ['Nonprofits & Activism'],
3482 'tags': list,
3483 'like_count': int,
3484 'dislike_count': int,
3485 },
3486 'params': {
3487 'noplaylist': True,
3488 'skip_download': True,
3489 },
3490 }, {
3491 'url': 'https://youtu.be/uWyaPkt-VOI?list=PL9D9FC436B881BA21',
3492 'only_matching': True,
3493 }]
3494
3495 def _real_extract(self, url):
3496 mobj = re.match(self._VALID_URL, url)
3497 video_id = mobj.group('id')
3498 playlist_id = mobj.group('playlist_id')
3499 return self.url_result(
3500 update_url_query('https://www.youtube.com/watch', {
3501 'v': video_id,
3502 'list': playlist_id,
3503 'feature': 'youtu.be',
3504 }), ie=YoutubeTabIE.ie_key(), video_id=playlist_id)
3505
3506
3507class YoutubeYtUserIE(InfoExtractor):
3508 IE_DESC = 'YouTube.com user videos, URL or "ytuser" keyword'
3509 _VALID_URL = r'ytuser:(?P<id>.+)'
3510 _TESTS = [{
3511 'url': 'ytuser:phihag',
3512 'only_matching': True,
3513 }]
3514
3515 def _real_extract(self, url):
3516 user_id = self._match_id(url)
3517 return self.url_result(
3518 'https://www.youtube.com/user/%s' % user_id,
3519 ie=YoutubeTabIE.ie_key(), video_id=user_id)
3520
3521
3522class YoutubeFavouritesIE(YoutubeBaseInfoExtractor):
3523 IE_NAME = 'youtube:favorites'
3524 IE_DESC = 'YouTube.com liked videos, ":ytfav" for short (requires authentication)'
3525 _VALID_URL = r':ytfav(?:ou?rite)?s?'
3526 _LOGIN_REQUIRED = True
3527 _TESTS = [{
3528 'url': ':ytfav',
3529 'only_matching': True,
3530 }, {
3531 'url': ':ytfavorites',
3532 'only_matching': True,
3533 }]
3534
3535 def _real_extract(self, url):
3536 return self.url_result(
3537 'https://www.youtube.com/playlist?list=LL',
3538 ie=YoutubeTabIE.ie_key())
3539
3540
3541class YoutubeSearchIE(SearchInfoExtractor, YoutubeBaseInfoExtractor):
3542 IE_DESC = 'YouTube.com searches'
3543 # there doesn't appear to be a real limit, for example if you search for
3544 # 'python' you get more than 8.000.000 results
3545 _MAX_RESULTS = float('inf')
3546 IE_NAME = 'youtube:search'
3547 _SEARCH_KEY = 'ytsearch'
3548 _SEARCH_PARAMS = None
3549 _TESTS = []
3550
3551 def _entries(self, query, n):
3552 data = {
3553 'context': {
3554 'client': {
3555 'clientName': 'WEB',
3556 'clientVersion': '2.20201021.03.00',
3557 }
3558 },
3559 'query': query,
3560 }
3561 if self._SEARCH_PARAMS:
3562 data['params'] = self._SEARCH_PARAMS
3563 total = 0
3564 for page_num in itertools.count(1):
3565 search = self._download_json(
3566 'https://www.youtube.com/youtubei/v1/search?key=AIzaSyAO_FJ2SlqU8Q4STEHLGCilw_Y9_11qcW8',
3567 video_id='query "%s"' % query,
3568 note='Downloading page %s' % page_num,
3569 errnote='Unable to download API page', fatal=False,
3570 data=json.dumps(data).encode('utf8'),
3571 headers={'content-type': 'application/json'})
3572 if not search:
3573 break
3574 slr_contents = try_get(
3575 search,
3576 (lambda x: x['contents']['twoColumnSearchResultsRenderer']['primaryContents']['sectionListRenderer']['contents'],
3577 lambda x: x['onResponseReceivedCommands'][0]['appendContinuationItemsAction']['continuationItems']),
3578 list)
3579 if not slr_contents:
3580 break
3581
3582 isr_contents = []
3583 continuation_token = None
3584 # Youtube sometimes adds promoted content to searches,
3585 # changing the index location of videos and token.
3586 # So we search through all entries till we find them.
3587 for index, isr in enumerate(slr_contents):
3588 if not isr_contents:
3589 isr_contents = try_get(
3590 slr_contents,
3591 (lambda x: x[index]['itemSectionRenderer']['contents']),
3592 list)
3593 for content in isr_contents:
3594 if content.get('videoRenderer') is not None:
3595 break
3596 else:
3597 isr_contents = []
3598
3599 if continuation_token is None:
3600 continuation_token = try_get(
3601 slr_contents,
3602 lambda x: x[index]['continuationItemRenderer']['continuationEndpoint']['continuationCommand'][
3603 'token'],
3604 compat_str)
3605 if continuation_token is not None and isr_contents:
3606 break
3607
3608 if not isr_contents:
3609 break
3610 for content in isr_contents:
3611 if not isinstance(content, dict):
3612 continue
3613 video = content.get('videoRenderer')
3614 if not isinstance(video, dict):
3615 continue
3616 video_id = video.get('videoId')
3617 if not video_id:
3618 continue
3619 title = try_get(video, lambda x: x['title']['runs'][0]['text'], compat_str)
3620 description = try_get(video, lambda x: x['descriptionSnippet']['runs'][0]['text'], compat_str)
3621 duration = parse_duration(try_get(video, lambda x: x['lengthText']['simpleText'], compat_str))
3622 view_count_text = try_get(video, lambda x: x['viewCountText']['simpleText'], compat_str) or ''
3623 view_count = int_or_none(self._search_regex(
3624 r'^(\d+)', re.sub(r'\s', '', view_count_text),
3625 'view count', default=None))
3626 uploader = try_get(video, lambda x: x['ownerText']['runs'][0]['text'], compat_str)
3627 total += 1
3628 yield {
3629 '_type': 'url_transparent',
3630 'ie_key': YoutubeIE.ie_key(),
3631 'id': video_id,
3632 'url': video_id,
3633 'title': title,
3634 'description': description,
3635 'duration': duration,
3636 'view_count': view_count,
3637 'uploader': uploader,
3638 }
3639 if total == n:
3640 return
3641 if not continuation_token:
3642 break
3643 data['continuation'] = continuation_token
3644
3645 def _get_n_results(self, query, n):
3646 """Get a specified number of results for a query"""
3647 return self.playlist_result(self._entries(query, n), query)
3648
3649
3650class YoutubeSearchDateIE(YoutubeSearchIE):
3651 IE_NAME = YoutubeSearchIE.IE_NAME + ':date'
3652 _SEARCH_KEY = 'ytsearchdate'
3653 IE_DESC = 'YouTube.com searches, newest videos first, "ytsearchdate" keyword'
3654 _SEARCH_PARAMS = 'CAI%3D'
3655
3656
3657class YoutubeSearchURLIE(YoutubeSearchIE):
3658 IE_DESC = 'YouTube.com searches, "ytsearch" keyword'
3659 IE_NAME = YoutubeSearchIE.IE_NAME + '_url'
3660 _VALID_URL = r'https?://(?:www\.)?youtube\.com/results\?(.*?&)?(?:search_query|q)=(?:[^&]+)(?:[&]|$)'
3661 # _MAX_RESULTS = 100
3662 _TESTS = [{
3663 'url': 'https://www.youtube.com/results?baz=bar&search_query=youtube-dl+test+video&filters=video&lclk=video',
3664 'playlist_mincount': 5,
3665 'info_dict': {
3666 'title': 'youtube-dl test video',
3667 }
3668 }, {
3669 'url': 'https://www.youtube.com/results?q=test&sp=EgQIBBgB',
3670 'only_matching': True,
3671 }]
3672
3673 @classmethod
3674 def _make_valid_url(cls):
3675 return cls._VALID_URL
3676
3677 def _real_extract(self, url):
3678 qs = compat_parse_qs(compat_urllib_parse_urlparse(url).query)
3679 query = (qs.get('search_query') or qs.get('q'))[0]
3680 self._SEARCH_PARAMS = qs.get('sp', ('',))[0]
3681 return self._get_n_results(query, self._MAX_RESULTS)
3682
3683
3684class YoutubeFeedsInfoExtractor(YoutubeTabIE):
3685 """
3686 Base class for feed extractors
3687 Subclasses must define the _FEED_NAME property.
3688 """
3689 _LOGIN_REQUIRED = True
3690 # _MAX_PAGES = 5
3691 _TESTS = []
3692
3693 @property
3694 def IE_NAME(self):
3695 return 'youtube:%s' % self._FEED_NAME
3696
3697 def _real_initialize(self):
3698 self._login()
3699
3700 def _real_extract(self, url):
3701 return self.url_result(
3702 'https://www.youtube.com/feed/%s' % self._FEED_NAME,
3703 ie=YoutubeTabIE.ie_key())
3704
3705
3706class YoutubeWatchLaterIE(InfoExtractor):
3707 IE_NAME = 'youtube:watchlater'
3708 IE_DESC = 'Youtube watch later list, ":ytwatchlater" for short (requires authentication)'
3709 _VALID_URL = r':ytwatchlater'
3710 _TESTS = [{
3711 'url': ':ytwatchlater',
3712 'only_matching': True,
3713 }]
3714
3715 def _real_extract(self, url):
3716 return self.url_result(
3717 'https://www.youtube.com/playlist?list=WL', ie=YoutubeTabIE.ie_key())
3718
3719
3720class YoutubeRecommendedIE(YoutubeFeedsInfoExtractor):
3721 IE_DESC = 'YouTube.com recommended videos, ":ytrec" for short (requires authentication)'
3722 _VALID_URL = r'https?://(?:www\.)?youtube\.com/?(?:[?#]|$)|:ytrec(?:ommended)?'
3723 _FEED_NAME = 'recommended'
3724 _TESTS = [{
3725 'url': ':ytrec',
3726 'only_matching': True,
3727 }, {
3728 'url': ':ytrecommended',
3729 'only_matching': True,
3730 }, {
3731 'url': 'https://youtube.com',
3732 'only_matching': True,
3733 }]
3734
3735
3736class YoutubeSubscriptionsIE(YoutubeFeedsInfoExtractor):
3737 IE_DESC = 'YouTube.com subscriptions feed, ":ytsubs" for short (requires authentication)'
3738 _VALID_URL = r':ytsub(?:scription)?s?'
3739 _FEED_NAME = 'subscriptions'
3740 _TESTS = [{
3741 'url': ':ytsubs',
3742 'only_matching': True,
3743 }, {
3744 'url': ':ytsubscriptions',
3745 'only_matching': True,
3746 }]
3747
3748
3749class YoutubeHistoryIE(YoutubeFeedsInfoExtractor):
3750 IE_DESC = 'Youtube watch history, ":ythistory" for short (requires authentication)'
3751 _VALID_URL = r':ythistory'
3752 _FEED_NAME = 'history'
3753 _TESTS = [{
3754 'url': ':ythistory',
3755 'only_matching': True,
3756 }]
3757
3758
3759class YoutubeTruncatedURLIE(InfoExtractor):
3760 IE_NAME = 'youtube:truncated_url'
3761 IE_DESC = False # Do not list
3762 _VALID_URL = r'''(?x)
3763 (?:https?://)?
3764 (?:\w+\.)?[yY][oO][uU][tT][uU][bB][eE](?:-nocookie)?\.com/
3765 (?:watch\?(?:
3766 feature=[a-z_]+|
3767 annotation_id=annotation_[^&]+|
3768 x-yt-cl=[0-9]+|
3769 hl=[^&]*|
3770 t=[0-9]+
3771 )?
3772 |
3773 attribution_link\?a=[^&]+
3774 )
3775 $
3776 '''
3777
3778 _TESTS = [{
3779 'url': 'https://www.youtube.com/watch?annotation_id=annotation_3951667041',
3780 'only_matching': True,
3781 }, {
3782 'url': 'https://www.youtube.com/watch?',
3783 'only_matching': True,
3784 }, {
3785 'url': 'https://www.youtube.com/watch?x-yt-cl=84503534',
3786 'only_matching': True,
3787 }, {
3788 'url': 'https://www.youtube.com/watch?feature=foo',
3789 'only_matching': True,
3790 }, {
3791 'url': 'https://www.youtube.com/watch?hl=en-GB',
3792 'only_matching': True,
3793 }, {
3794 'url': 'https://www.youtube.com/watch?t=2372',
3795 'only_matching': True,
3796 }]
3797
3798 def _real_extract(self, url):
3799 raise ExtractorError(
3800 'Did you forget to quote the URL? Remember that & is a meta '
3801 'character in most shells, so you want to put the URL in quotes, '
3802 'like youtube-dl '
3803 '"https://www.youtube.com/watch?feature=foo&v=BaW_jenozKc" '
3804 ' or simply youtube-dl BaW_jenozKc .',
3805 expected=True)
3806
3807
3808class YoutubeTruncatedIDIE(InfoExtractor):
3809 IE_NAME = 'youtube:truncated_id'
3810 IE_DESC = False # Do not list
3811 _VALID_URL = r'https?://(?:www\.)?youtube\.com/watch\?v=(?P<id>[0-9A-Za-z_-]{1,10})$'
3812
3813 _TESTS = [{
3814 'url': 'https://www.youtube.com/watch?v=N_708QY7Ob',
3815 'only_matching': True,
3816 }]
3817
3818 def _real_extract(self, url):
3819 video_id = self._match_id(url)
3820 raise ExtractorError(
3821 'Incomplete YouTube ID %s. URL %s looks truncated.' % (video_id, url),
3822 expected=True)
3823
3824
3825# Do Youtube show urls even exist anymore? I couldn't find any
3826r'''
3827class YoutubeShowIE(YoutubeTabIE):
3828 IE_DESC = 'YouTube.com (multi-season) shows'
3829 _VALID_URL = r'https?://(?:www\.)?youtube\.com/show/(?P<id>[^?#]*)'
3830 IE_NAME = 'youtube:show'
3831 _TESTS = [{
3832 'url': 'https://www.youtube.com/show/airdisasters',
3833 'playlist_mincount': 5,
3834 'info_dict': {
3835 'id': 'airdisasters',
3836 'title': 'Air Disasters',
3837 }
3838 }]
3839
3840 def _real_extract(self, url):
3841 playlist_id = self._match_id(url)
3842 return super(YoutubeShowIE, self)._real_extract(
3843 'https://www.youtube.com/show/%s/playlists' % playlist_id)
3844'''