Compare commits
21 Commits
2017.02.27
...
2017.03.02
Author | SHA1 | Date | |
---|---|---|---|
250eea6821 | |||
28d15b73f8 | |||
11bb6ad1a5 | |||
c9612c0487 | |||
af5049f128 | |||
158af5242e | |||
40df485f55 | |||
4b8a984c67 | |||
83e8fce628 | |||
aa9cc2ecbf | |||
1dc24093f8 | |||
11bae9cdde | |||
43b38424a9 | |||
948519b35d | |||
87dadd456a | |||
7c4aa6fd6f | |||
9bd05b5a18 | |||
0a5445ddbe | |||
f48409c7ac | |||
c9619f0a17 | |||
f4c68ba372 |
6
.github/ISSUE_TEMPLATE.md
vendored
6
.github/ISSUE_TEMPLATE.md
vendored
@ -6,8 +6,8 @@
|
|||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
### Make sure you are using the *latest* version: run `youtube-dl --version` and ensure your version is *2017.02.27*. If it's not read [this FAQ entry](https://github.com/rg3/youtube-dl/blob/master/README.md#how-do-i-update-youtube-dl) and update. Issues with outdated version will be rejected.
|
### Make sure you are using the *latest* version: run `youtube-dl --version` and ensure your version is *2017.03.02*. If it's not read [this FAQ entry](https://github.com/rg3/youtube-dl/blob/master/README.md#how-do-i-update-youtube-dl) and update. Issues with outdated version will be rejected.
|
||||||
- [ ] I've **verified** and **I assure** that I'm running youtube-dl **2017.02.27**
|
- [ ] I've **verified** and **I assure** that I'm running youtube-dl **2017.03.02**
|
||||||
|
|
||||||
### Before submitting an *issue* make sure you have:
|
### Before submitting an *issue* make sure you have:
|
||||||
- [ ] At least skimmed through [README](https://github.com/rg3/youtube-dl/blob/master/README.md) and **most notably** [FAQ](https://github.com/rg3/youtube-dl#faq) and [BUGS](https://github.com/rg3/youtube-dl#bugs) sections
|
- [ ] At least skimmed through [README](https://github.com/rg3/youtube-dl/blob/master/README.md) and **most notably** [FAQ](https://github.com/rg3/youtube-dl#faq) and [BUGS](https://github.com/rg3/youtube-dl#bugs) sections
|
||||||
@ -35,7 +35,7 @@ $ youtube-dl -v <your command line>
|
|||||||
[debug] User config: []
|
[debug] User config: []
|
||||||
[debug] Command-line args: [u'-v', u'http://www.youtube.com/watch?v=BaW_jenozKcj']
|
[debug] Command-line args: [u'-v', u'http://www.youtube.com/watch?v=BaW_jenozKcj']
|
||||||
[debug] Encodings: locale cp1251, fs mbcs, out cp866, pref cp1251
|
[debug] Encodings: locale cp1251, fs mbcs, out cp866, pref cp1251
|
||||||
[debug] youtube-dl version 2017.02.27
|
[debug] youtube-dl version 2017.03.02
|
||||||
[debug] Python version 2.7.11 - Windows-2003Server-5.2.3790-SP2
|
[debug] Python version 2.7.11 - Windows-2003Server-5.2.3790-SP2
|
||||||
[debug] exe versions: ffmpeg N-75573-g1d0487f, ffprobe N-75573-g1d0487f, rtmpdump 2.4
|
[debug] exe versions: ffmpeg N-75573-g1d0487f, ffprobe N-75573-g1d0487f, rtmpdump 2.4
|
||||||
[debug] Proxy map: {}
|
[debug] Proxy map: {}
|
||||||
|
28
ChangeLog
28
ChangeLog
@ -1,3 +1,31 @@
|
|||||||
|
version 2017.03.02
|
||||||
|
|
||||||
|
Core
|
||||||
|
+ [adobepass] Add support for Charter Spectrum (#11465)
|
||||||
|
* [YoutubeDL] Don't sanitize identifiers in output template (#12317)
|
||||||
|
|
||||||
|
Extractors
|
||||||
|
* [facebook] Fix extraction (#12323, #12330)
|
||||||
|
* [youtube] Mark errors about rental videos as expected (#12324)
|
||||||
|
+ [npo] Add support for audio
|
||||||
|
* [npo] Adapt to app.php API (#12311, #12320)
|
||||||
|
|
||||||
|
|
||||||
|
version 2017.02.28
|
||||||
|
|
||||||
|
Core
|
||||||
|
+ [utils] Add bytes_to_long and long_to_bytes
|
||||||
|
+ [utils] Add pkcs1pad
|
||||||
|
+ [aes] Add aes_cbc_encrypt
|
||||||
|
|
||||||
|
Extractors
|
||||||
|
+ [azmedien:showplaylist] Add support for show playlists (#12160)
|
||||||
|
+ [youtube:playlist] Recognize another playlist pattern (#11928, #12286)
|
||||||
|
+ [daisuki] Add support for daisuki.net (#2486, #3186, #4738, #6175, #7776,
|
||||||
|
#10060)
|
||||||
|
* [douyu] Fix extraction (#12301)
|
||||||
|
|
||||||
|
|
||||||
version 2017.02.27
|
version 2017.02.27
|
||||||
|
|
||||||
Core
|
Core
|
||||||
|
@ -78,6 +78,7 @@
|
|||||||
- **awaan:video**
|
- **awaan:video**
|
||||||
- **AZMedien**: AZ Medien videos
|
- **AZMedien**: AZ Medien videos
|
||||||
- **AZMedienPlaylist**: AZ Medien playlists
|
- **AZMedienPlaylist**: AZ Medien playlists
|
||||||
|
- **AZMedienShowPlaylist**: AZ Medien show playlists
|
||||||
- **Azubu**
|
- **Azubu**
|
||||||
- **AzubuLive**
|
- **AzubuLive**
|
||||||
- **BaiduVideo**: 百度视频
|
- **BaiduVideo**: 百度视频
|
||||||
@ -191,6 +192,8 @@
|
|||||||
- **dailymotion:playlist**
|
- **dailymotion:playlist**
|
||||||
- **dailymotion:user**
|
- **dailymotion:user**
|
||||||
- **DailymotionCloud**
|
- **DailymotionCloud**
|
||||||
|
- **Daisuki**
|
||||||
|
- **DaisukiPlaylist**
|
||||||
- **daum.net**
|
- **daum.net**
|
||||||
- **daum.net:clip**
|
- **daum.net:clip**
|
||||||
- **daum.net:playlist**
|
- **daum.net:playlist**
|
||||||
|
@ -8,7 +8,7 @@ import sys
|
|||||||
import unittest
|
import unittest
|
||||||
sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
|
sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
|
||||||
|
|
||||||
from youtube_dl.aes import aes_decrypt, aes_encrypt, aes_cbc_decrypt, aes_decrypt_text
|
from youtube_dl.aes import aes_decrypt, aes_encrypt, aes_cbc_decrypt, aes_cbc_encrypt, aes_decrypt_text
|
||||||
from youtube_dl.utils import bytes_to_intlist, intlist_to_bytes
|
from youtube_dl.utils import bytes_to_intlist, intlist_to_bytes
|
||||||
import base64
|
import base64
|
||||||
|
|
||||||
@ -34,6 +34,13 @@ class TestAES(unittest.TestCase):
|
|||||||
decrypted = intlist_to_bytes(aes_cbc_decrypt(data, self.key, self.iv))
|
decrypted = intlist_to_bytes(aes_cbc_decrypt(data, self.key, self.iv))
|
||||||
self.assertEqual(decrypted.rstrip(b'\x08'), self.secret_msg)
|
self.assertEqual(decrypted.rstrip(b'\x08'), self.secret_msg)
|
||||||
|
|
||||||
|
def test_cbc_encrypt(self):
|
||||||
|
data = bytes_to_intlist(self.secret_msg)
|
||||||
|
encrypted = intlist_to_bytes(aes_cbc_encrypt(data, self.key, self.iv))
|
||||||
|
self.assertEqual(
|
||||||
|
encrypted,
|
||||||
|
b"\x97\x92+\xe5\x0b\xc3\x18\x91ky9m&\xb3\xb5@\xe6'\xc2\x96.\xc8u\x88\xab9-[\x9e|\xf1\xcd")
|
||||||
|
|
||||||
def test_decrypt_text(self):
|
def test_decrypt_text(self):
|
||||||
password = intlist_to_bytes(self.key).decode('utf-8')
|
password = intlist_to_bytes(self.key).decode('utf-8')
|
||||||
encrypted = base64.b64encode(
|
encrypted = base64.b64encode(
|
||||||
|
@ -52,6 +52,7 @@ from youtube_dl.utils import (
|
|||||||
parse_filesize,
|
parse_filesize,
|
||||||
parse_count,
|
parse_count,
|
||||||
parse_iso8601,
|
parse_iso8601,
|
||||||
|
pkcs1pad,
|
||||||
read_batch_urls,
|
read_batch_urls,
|
||||||
sanitize_filename,
|
sanitize_filename,
|
||||||
sanitize_path,
|
sanitize_path,
|
||||||
@ -1104,6 +1105,14 @@ The first line
|
|||||||
ohdave_rsa_encrypt(b'aa111222', e, N),
|
ohdave_rsa_encrypt(b'aa111222', e, N),
|
||||||
'726664bd9a23fd0c70f9f1b84aab5e3905ce1e45a584e9cbcf9bcc7510338fc1986d6c599ff990d923aa43c51c0d9013cd572e13bc58f4ae48f2ed8c0b0ba881')
|
'726664bd9a23fd0c70f9f1b84aab5e3905ce1e45a584e9cbcf9bcc7510338fc1986d6c599ff990d923aa43c51c0d9013cd572e13bc58f4ae48f2ed8c0b0ba881')
|
||||||
|
|
||||||
|
def test_pkcs1pad(self):
|
||||||
|
data = [1, 2, 3]
|
||||||
|
padded_data = pkcs1pad(data, 32)
|
||||||
|
self.assertEqual(padded_data[:2], [0, 2])
|
||||||
|
self.assertEqual(padded_data[28:], [0, 1, 2, 3])
|
||||||
|
|
||||||
|
self.assertRaises(ValueError, pkcs1pad, data, 8)
|
||||||
|
|
||||||
def test_encode_base_n(self):
|
def test_encode_base_n(self):
|
||||||
self.assertEqual(encode_base_n(0, 30), '0')
|
self.assertEqual(encode_base_n(0, 30), '0')
|
||||||
self.assertEqual(encode_base_n(80, 30), '2k')
|
self.assertEqual(encode_base_n(80, 30), '2k')
|
||||||
|
@ -616,7 +616,7 @@ class YoutubeDL(object):
|
|||||||
sanitize = lambda k, v: sanitize_filename(
|
sanitize = lambda k, v: sanitize_filename(
|
||||||
compat_str(v),
|
compat_str(v),
|
||||||
restricted=self.params.get('restrictfilenames'),
|
restricted=self.params.get('restrictfilenames'),
|
||||||
is_id=(k == 'id'))
|
is_id=(k == 'id' or k.endswith('_id')))
|
||||||
template_dict = dict((k, v if isinstance(v, compat_numeric_types) else sanitize(k, v))
|
template_dict = dict((k, v if isinstance(v, compat_numeric_types) else sanitize(k, v))
|
||||||
for k, v in template_dict.items()
|
for k, v in template_dict.items()
|
||||||
if v is not None and not isinstance(v, (list, tuple, dict)))
|
if v is not None and not isinstance(v, (list, tuple, dict)))
|
||||||
|
@ -60,6 +60,34 @@ def aes_cbc_decrypt(data, key, iv):
|
|||||||
return decrypted_data
|
return decrypted_data
|
||||||
|
|
||||||
|
|
||||||
|
def aes_cbc_encrypt(data, key, iv):
|
||||||
|
"""
|
||||||
|
Encrypt with aes in CBC mode. Using PKCS#7 padding
|
||||||
|
|
||||||
|
@param {int[]} data cleartext
|
||||||
|
@param {int[]} key 16/24/32-Byte cipher key
|
||||||
|
@param {int[]} iv 16-Byte IV
|
||||||
|
@returns {int[]} encrypted data
|
||||||
|
"""
|
||||||
|
expanded_key = key_expansion(key)
|
||||||
|
block_count = int(ceil(float(len(data)) / BLOCK_SIZE_BYTES))
|
||||||
|
|
||||||
|
encrypted_data = []
|
||||||
|
previous_cipher_block = iv
|
||||||
|
for i in range(block_count):
|
||||||
|
block = data[i * BLOCK_SIZE_BYTES: (i + 1) * BLOCK_SIZE_BYTES]
|
||||||
|
remaining_length = BLOCK_SIZE_BYTES - len(block)
|
||||||
|
block += [remaining_length] * remaining_length
|
||||||
|
mixed_block = xor(block, previous_cipher_block)
|
||||||
|
|
||||||
|
encrypted_block = aes_encrypt(mixed_block, expanded_key)
|
||||||
|
encrypted_data += encrypted_block
|
||||||
|
|
||||||
|
previous_cipher_block = encrypted_block
|
||||||
|
|
||||||
|
return encrypted_data
|
||||||
|
|
||||||
|
|
||||||
def key_expansion(data):
|
def key_expansion(data):
|
||||||
"""
|
"""
|
||||||
Generate key schedule
|
Generate key schedule
|
||||||
|
@ -36,6 +36,11 @@ MSO_INFO = {
|
|||||||
'username_field': 'Ecom_User_ID',
|
'username_field': 'Ecom_User_ID',
|
||||||
'password_field': 'Ecom_Password',
|
'password_field': 'Ecom_Password',
|
||||||
},
|
},
|
||||||
|
'Charter_Direct': {
|
||||||
|
'name': 'Charter Spectrum',
|
||||||
|
'username_field': 'IDToken1',
|
||||||
|
'password_field': 'IDToken2',
|
||||||
|
},
|
||||||
'thr030': {
|
'thr030': {
|
||||||
'name': '3 Rivers Communications'
|
'name': '3 Rivers Communications'
|
||||||
},
|
},
|
||||||
|
@ -1,3 +1,4 @@
|
|||||||
|
# coding: utf-8
|
||||||
from __future__ import unicode_literals
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
import re
|
import re
|
||||||
@ -5,6 +6,7 @@ import re
|
|||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from .kaltura import KalturaIE
|
from .kaltura import KalturaIE
|
||||||
from ..utils import (
|
from ..utils import (
|
||||||
|
get_element_by_class,
|
||||||
get_element_by_id,
|
get_element_by_id,
|
||||||
strip_or_none,
|
strip_or_none,
|
||||||
urljoin,
|
urljoin,
|
||||||
@ -170,3 +172,42 @@ class AZMedienPlaylistIE(AZMedienBaseIE):
|
|||||||
'video-title', webpage)), group='title')
|
'video-title', webpage)), group='title')
|
||||||
|
|
||||||
return self.playlist_result(entries, show_id, title)
|
return self.playlist_result(entries, show_id, title)
|
||||||
|
|
||||||
|
|
||||||
|
class AZMedienShowPlaylistIE(AZMedienBaseIE):
|
||||||
|
IE_DESC = 'AZ Medien show playlists'
|
||||||
|
_VALID_URL = r'''(?x)
|
||||||
|
https?://
|
||||||
|
(?:www\.)?
|
||||||
|
(?:
|
||||||
|
telezueri\.ch|
|
||||||
|
telebaern\.tv|
|
||||||
|
telem1\.ch
|
||||||
|
)/
|
||||||
|
(?:
|
||||||
|
all-episodes|
|
||||||
|
alle-episoden
|
||||||
|
)/
|
||||||
|
(?P<id>[^/?#&]+)
|
||||||
|
'''
|
||||||
|
|
||||||
|
_TEST = {
|
||||||
|
'url': 'http://www.telezueri.ch/all-episodes/astrotalk',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'astrotalk',
|
||||||
|
'title': 'TeleZüri: AstroTalk - alle episoden',
|
||||||
|
'description': 'md5:4c0f7e7d741d906004266e295ceb4a26',
|
||||||
|
},
|
||||||
|
'playlist_mincount': 13,
|
||||||
|
}
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
playlist_id = self._match_id(url)
|
||||||
|
webpage = self._download_webpage(url, playlist_id)
|
||||||
|
episodes = get_element_by_class('search-mobile-box', webpage)
|
||||||
|
entries = [self.url_result(
|
||||||
|
urljoin(url, m.group('url'))) for m in re.finditer(
|
||||||
|
r'<a[^>]+href=(["\'])(?P<url>(?:(?!\1).)+)\1', episodes)]
|
||||||
|
title = self._og_search_title(webpage, fatal=False)
|
||||||
|
description = self._og_search_description(webpage)
|
||||||
|
return self.playlist_result(entries, playlist_id, title, description)
|
||||||
|
159
youtube_dl/extractor/daisuki.py
Normal file
159
youtube_dl/extractor/daisuki.py
Normal file
@ -0,0 +1,159 @@
|
|||||||
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
|
import base64
|
||||||
|
import json
|
||||||
|
import random
|
||||||
|
import re
|
||||||
|
|
||||||
|
from .common import InfoExtractor
|
||||||
|
from ..aes import (
|
||||||
|
aes_cbc_decrypt,
|
||||||
|
aes_cbc_encrypt,
|
||||||
|
)
|
||||||
|
from ..utils import (
|
||||||
|
bytes_to_intlist,
|
||||||
|
bytes_to_long,
|
||||||
|
clean_html,
|
||||||
|
ExtractorError,
|
||||||
|
intlist_to_bytes,
|
||||||
|
get_element_by_id,
|
||||||
|
js_to_json,
|
||||||
|
int_or_none,
|
||||||
|
long_to_bytes,
|
||||||
|
pkcs1pad,
|
||||||
|
remove_end,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class DaisukiIE(InfoExtractor):
|
||||||
|
_VALID_URL = r'https?://(?:www\.)?daisuki\.net/[^/]+/[^/]+/[^/]+/watch\.[^.]+\.(?P<id>\d+)\.html'
|
||||||
|
|
||||||
|
_TEST = {
|
||||||
|
'url': 'http://www.daisuki.net/tw/en/anime/watch.TheIdolMasterCG.11213.html',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '11213',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': '#01 Who is in the pumpkin carriage? - THE IDOLM@STER CINDERELLA GIRLS',
|
||||||
|
'subtitles': {
|
||||||
|
'mul': [{
|
||||||
|
'ext': 'ttml',
|
||||||
|
}],
|
||||||
|
},
|
||||||
|
'creator': 'BANDAI NAMCO Entertainment',
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
'skip_download': True, # AES-encrypted HLS stream
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
# The public key in PEM format can be found in clientlibs_anime_watch.min.js
|
||||||
|
_RSA_KEY = (0xc5524c25e8e14b366b3754940beeb6f96cb7e2feef0b932c7659a0c5c3bf173d602464c2df73d693b513ae06ff1be8f367529ab30bf969c5640522181f2a0c51ea546ae120d3d8d908595e4eff765b389cde080a1ef7f1bbfb07411cc568db73b7f521cedf270cbfbe0ddbc29b1ac9d0f2d8f4359098caffee6d07915020077d, 65537)
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
video_id = self._match_id(url)
|
||||||
|
|
||||||
|
webpage = self._download_webpage(url, video_id)
|
||||||
|
|
||||||
|
flashvars = self._parse_json(self._search_regex(
|
||||||
|
r'(?s)var\s+flashvars\s*=\s*({.+?});', webpage, 'flashvars'),
|
||||||
|
video_id, transform_source=js_to_json)
|
||||||
|
|
||||||
|
iv = [0] * 16
|
||||||
|
|
||||||
|
data = {}
|
||||||
|
for key in ('device_cd', 'mv_id', 'ss1_prm', 'ss2_prm', 'ss3_prm', 'ss_id'):
|
||||||
|
data[key] = flashvars.get(key, '')
|
||||||
|
|
||||||
|
encrypted_rtn = None
|
||||||
|
|
||||||
|
# Some AES keys are rejected. Try it with different AES keys
|
||||||
|
for idx in range(5):
|
||||||
|
aes_key = [random.randint(0, 254) for _ in range(32)]
|
||||||
|
padded_aeskey = intlist_to_bytes(pkcs1pad(aes_key, 128))
|
||||||
|
|
||||||
|
n, e = self._RSA_KEY
|
||||||
|
encrypted_aeskey = long_to_bytes(pow(bytes_to_long(padded_aeskey), e, n))
|
||||||
|
init_data = self._download_json('http://www.daisuki.net/bin/bgn/init', video_id, query={
|
||||||
|
's': flashvars.get('s', ''),
|
||||||
|
'c': flashvars.get('ss3_prm', ''),
|
||||||
|
'e': url,
|
||||||
|
'd': base64.b64encode(intlist_to_bytes(aes_cbc_encrypt(
|
||||||
|
bytes_to_intlist(json.dumps(data)),
|
||||||
|
aes_key, iv))).decode('ascii'),
|
||||||
|
'a': base64.b64encode(encrypted_aeskey).decode('ascii'),
|
||||||
|
}, note='Downloading JSON metadata' + (' (try #%d)' % (idx + 1) if idx > 0 else ''))
|
||||||
|
|
||||||
|
if 'rtn' in init_data:
|
||||||
|
encrypted_rtn = init_data['rtn']
|
||||||
|
break
|
||||||
|
|
||||||
|
self._sleep(5, video_id)
|
||||||
|
|
||||||
|
if encrypted_rtn is None:
|
||||||
|
raise ExtractorError('Failed to fetch init data')
|
||||||
|
|
||||||
|
rtn = self._parse_json(
|
||||||
|
intlist_to_bytes(aes_cbc_decrypt(bytes_to_intlist(
|
||||||
|
base64.b64decode(encrypted_rtn)),
|
||||||
|
aes_key, iv)).decode('utf-8').rstrip('\0'),
|
||||||
|
video_id)
|
||||||
|
|
||||||
|
formats = self._extract_m3u8_formats(
|
||||||
|
rtn['play_url'], video_id, ext='mp4', entry_protocol='m3u8_native')
|
||||||
|
|
||||||
|
title = remove_end(self._og_search_title(webpage), ' - DAISUKI')
|
||||||
|
|
||||||
|
creator = self._html_search_regex(
|
||||||
|
r'Creator\s*:\s*([^<]+)', webpage, 'creator', fatal=False)
|
||||||
|
|
||||||
|
subtitles = {}
|
||||||
|
caption_url = rtn.get('caption_url')
|
||||||
|
if caption_url:
|
||||||
|
# mul: multiple languages
|
||||||
|
subtitles['mul'] = [{
|
||||||
|
'url': caption_url,
|
||||||
|
'ext': 'ttml',
|
||||||
|
}]
|
||||||
|
|
||||||
|
return {
|
||||||
|
'id': video_id,
|
||||||
|
'title': title,
|
||||||
|
'formats': formats,
|
||||||
|
'subtitles': subtitles,
|
||||||
|
'creator': creator,
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class DaisukiPlaylistIE(InfoExtractor):
|
||||||
|
_VALID_URL = r'https?://(?:www\.)daisuki\.net/[^/]+/[^/]+/[^/]+/detail\.(?P<id>[a-zA-Z0-9]+)\.html'
|
||||||
|
|
||||||
|
_TEST = {
|
||||||
|
'url': 'http://www.daisuki.net/tw/en/anime/detail.TheIdolMasterCG.html',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'TheIdolMasterCG',
|
||||||
|
'title': 'THE IDOLM@STER CINDERELLA GIRLS',
|
||||||
|
'description': 'md5:0f2c028a9339f7a2c7fbf839edc5c5d8',
|
||||||
|
},
|
||||||
|
'playlist_count': 26,
|
||||||
|
}
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
playlist_id = self._match_id(url)
|
||||||
|
|
||||||
|
webpage = self._download_webpage(url, playlist_id)
|
||||||
|
|
||||||
|
episode_pattern = r'''(?sx)
|
||||||
|
<img[^>]+delay="[^"]+/(\d+)/movie\.jpg".+?
|
||||||
|
<p[^>]+class=".*?\bepisodeNumber\b.*?">(?:<a[^>]+>)?([^<]+)'''
|
||||||
|
entries = [{
|
||||||
|
'_type': 'url_transparent',
|
||||||
|
'url': url.replace('detail', 'watch').replace('.html', '.' + movie_id + '.html'),
|
||||||
|
'episode_id': episode_id,
|
||||||
|
'episode_number': int_or_none(episode_id),
|
||||||
|
} for movie_id, episode_id in re.findall(episode_pattern, webpage)]
|
||||||
|
|
||||||
|
playlist_title = remove_end(
|
||||||
|
self._og_search_title(webpage, fatal=False), ' - Anime - DAISUKI')
|
||||||
|
playlist_description = clean_html(get_element_by_id('synopsisTxt', webpage))
|
||||||
|
|
||||||
|
return self.playlist_result(entries, playlist_id, playlist_title, playlist_description)
|
@ -1,15 +1,7 @@
|
|||||||
# coding: utf-8
|
# coding: utf-8
|
||||||
from __future__ import unicode_literals
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
import hashlib
|
|
||||||
import time
|
|
||||||
import uuid
|
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from ..compat import (
|
|
||||||
compat_str,
|
|
||||||
compat_urllib_parse_urlencode,
|
|
||||||
)
|
|
||||||
from ..utils import (
|
from ..utils import (
|
||||||
ExtractorError,
|
ExtractorError,
|
||||||
unescapeHTML,
|
unescapeHTML,
|
||||||
@ -24,8 +16,8 @@ class DouyuTVIE(InfoExtractor):
|
|||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': '17732',
|
'id': '17732',
|
||||||
'display_id': 'iseven',
|
'display_id': 'iseven',
|
||||||
'ext': 'flv',
|
'ext': 'mp4',
|
||||||
'title': 're:^清晨醒脑!T-ara根本停不下来! [0-9]{4}-[0-9]{2}-[0-9]{2} [0-9]{2}:[0-9]{2}$',
|
'title': 're:^清晨醒脑!T-ARA根本停不下来! [0-9]{4}-[0-9]{2}-[0-9]{2} [0-9]{2}:[0-9]{2}$',
|
||||||
'description': r're:.*m7show@163\.com.*',
|
'description': r're:.*m7show@163\.com.*',
|
||||||
'thumbnail': r're:^https?://.*\.jpg$',
|
'thumbnail': r're:^https?://.*\.jpg$',
|
||||||
'uploader': '7师傅',
|
'uploader': '7师傅',
|
||||||
@ -39,7 +31,7 @@ class DouyuTVIE(InfoExtractor):
|
|||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': '85982',
|
'id': '85982',
|
||||||
'display_id': '85982',
|
'display_id': '85982',
|
||||||
'ext': 'flv',
|
'ext': 'mp4',
|
||||||
'title': 're:^小漠从零单排记!——CSOL2躲猫猫 [0-9]{4}-[0-9]{2}-[0-9]{2} [0-9]{2}:[0-9]{2}$',
|
'title': 're:^小漠从零单排记!——CSOL2躲猫猫 [0-9]{4}-[0-9]{2}-[0-9]{2} [0-9]{2}:[0-9]{2}$',
|
||||||
'description': 'md5:746a2f7a253966a06755a912f0acc0d2',
|
'description': 'md5:746a2f7a253966a06755a912f0acc0d2',
|
||||||
'thumbnail': r're:^https?://.*\.jpg$',
|
'thumbnail': r're:^https?://.*\.jpg$',
|
||||||
@ -55,8 +47,8 @@ class DouyuTVIE(InfoExtractor):
|
|||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': '17732',
|
'id': '17732',
|
||||||
'display_id': '17732',
|
'display_id': '17732',
|
||||||
'ext': 'flv',
|
'ext': 'mp4',
|
||||||
'title': 're:^清晨醒脑!T-ara根本停不下来! [0-9]{4}-[0-9]{2}-[0-9]{2} [0-9]{2}:[0-9]{2}$',
|
'title': 're:^清晨醒脑!T-ARA根本停不下来! [0-9]{4}-[0-9]{2}-[0-9]{2} [0-9]{2}:[0-9]{2}$',
|
||||||
'description': r're:.*m7show@163\.com.*',
|
'description': r're:.*m7show@163\.com.*',
|
||||||
'thumbnail': r're:^https?://.*\.jpg$',
|
'thumbnail': r're:^https?://.*\.jpg$',
|
||||||
'uploader': '7师傅',
|
'uploader': '7师傅',
|
||||||
@ -96,45 +88,18 @@ class DouyuTVIE(InfoExtractor):
|
|||||||
if room.get('show_status') == '2':
|
if room.get('show_status') == '2':
|
||||||
raise ExtractorError('Live stream is offline', expected=True)
|
raise ExtractorError('Live stream is offline', expected=True)
|
||||||
|
|
||||||
tt = compat_str(int(time.time() / 60))
|
formats = self._extract_m3u8_formats(
|
||||||
did = uuid.uuid4().hex.upper()
|
room['hls_url'], video_id, ext='mp4')
|
||||||
|
|
||||||
sign_content = ''.join((room_id, did, self._API_KEY, tt))
|
|
||||||
sign = hashlib.md5((sign_content).encode('utf-8')).hexdigest()
|
|
||||||
|
|
||||||
flv_data = compat_urllib_parse_urlencode({
|
|
||||||
'cdn': 'ws',
|
|
||||||
'rate': '0',
|
|
||||||
'tt': tt,
|
|
||||||
'did': did,
|
|
||||||
'sign': sign,
|
|
||||||
})
|
|
||||||
|
|
||||||
video_info = self._download_json(
|
|
||||||
'http://www.douyu.com/lapi/live/getPlay/%s' % room_id, video_id,
|
|
||||||
data=flv_data, note='Downloading video info',
|
|
||||||
headers={'Content-Type': 'application/x-www-form-urlencoded'})
|
|
||||||
|
|
||||||
error_code = video_info.get('error', 0)
|
|
||||||
if error_code is not 0:
|
|
||||||
raise ExtractorError(
|
|
||||||
'%s reported error %i' % (self.IE_NAME, error_code),
|
|
||||||
expected=True)
|
|
||||||
|
|
||||||
base_url = video_info['data']['rtmp_url']
|
|
||||||
live_path = video_info['data']['rtmp_live']
|
|
||||||
|
|
||||||
video_url = '%s/%s' % (base_url, live_path)
|
|
||||||
|
|
||||||
title = self._live_title(unescapeHTML(room['room_name']))
|
title = self._live_title(unescapeHTML(room['room_name']))
|
||||||
description = room.get('notice')
|
description = room.get('show_details')
|
||||||
thumbnail = room.get('room_src')
|
thumbnail = room.get('room_src')
|
||||||
uploader = room.get('nickname')
|
uploader = room.get('nickname')
|
||||||
|
|
||||||
return {
|
return {
|
||||||
'id': room_id,
|
'id': room_id,
|
||||||
'display_id': video_id,
|
'display_id': video_id,
|
||||||
'url': video_url,
|
'formats': formats,
|
||||||
'title': title,
|
'title': title,
|
||||||
'description': description,
|
'description': description,
|
||||||
'thumbnail': thumbnail,
|
'thumbnail': thumbnail,
|
||||||
|
@ -83,6 +83,7 @@ from .awaan import (
|
|||||||
from .azmedien import (
|
from .azmedien import (
|
||||||
AZMedienIE,
|
AZMedienIE,
|
||||||
AZMedienPlaylistIE,
|
AZMedienPlaylistIE,
|
||||||
|
AZMedienShowPlaylistIE,
|
||||||
)
|
)
|
||||||
from .azubu import AzubuIE, AzubuLiveIE
|
from .azubu import AzubuIE, AzubuLiveIE
|
||||||
from .baidu import BaiduVideoIE
|
from .baidu import BaiduVideoIE
|
||||||
@ -227,6 +228,10 @@ from .dailymotion import (
|
|||||||
DailymotionUserIE,
|
DailymotionUserIE,
|
||||||
DailymotionCloudIE,
|
DailymotionCloudIE,
|
||||||
)
|
)
|
||||||
|
from .daisuki import (
|
||||||
|
DaisukiIE,
|
||||||
|
DaisukiPlaylistIE,
|
||||||
|
)
|
||||||
from .daum import (
|
from .daum import (
|
||||||
DaumIE,
|
DaumIE,
|
||||||
DaumClipIE,
|
DaumClipIE,
|
||||||
|
@ -303,7 +303,7 @@ class FacebookIE(InfoExtractor):
|
|||||||
if not video_data:
|
if not video_data:
|
||||||
server_js_data = self._parse_json(
|
server_js_data = self._parse_json(
|
||||||
self._search_regex(
|
self._search_regex(
|
||||||
r'bigPipe\.onPageletArrive\(({.+?})\)\s*;\s*}\s*\)\s*,\s*["\']onPageletArrive\s+(?:stream_pagelet|pagelet_group_mall)',
|
r'bigPipe\.onPageletArrive\(({.+?})\)\s*;\s*}\s*\)\s*,\s*["\']onPageletArrive\s+(?:stream_pagelet|pagelet_group_mall|permalink_video_pagelet)',
|
||||||
webpage, 'js data', default='{}'),
|
webpage, 'js data', default='{}'),
|
||||||
video_id, transform_source=js_to_json, fatal=False)
|
video_id, transform_source=js_to_json, fatal=False)
|
||||||
if server_js_data:
|
if server_js_data:
|
||||||
|
@ -3,41 +3,27 @@ from __future__ import unicode_literals
|
|||||||
import re
|
import re
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from ..compat import compat_HTTPError
|
from ..compat import (
|
||||||
|
compat_HTTPError,
|
||||||
|
compat_str,
|
||||||
|
)
|
||||||
from ..utils import (
|
from ..utils import (
|
||||||
|
determine_ext,
|
||||||
|
ExtractorError,
|
||||||
fix_xml_ampersands,
|
fix_xml_ampersands,
|
||||||
orderedSet,
|
orderedSet,
|
||||||
parse_duration,
|
parse_duration,
|
||||||
qualities,
|
qualities,
|
||||||
strip_jsonp,
|
strip_jsonp,
|
||||||
unified_strdate,
|
unified_strdate,
|
||||||
ExtractorError,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
class NPOBaseIE(InfoExtractor):
|
class NPOBaseIE(InfoExtractor):
|
||||||
def _get_token(self, video_id):
|
def _get_token(self, video_id):
|
||||||
token_page = self._download_webpage(
|
return self._download_json(
|
||||||
'http://ida.omroep.nl/npoplayer/i.js',
|
'http://ida.omroep.nl/app.php/auth', video_id,
|
||||||
video_id, note='Downloading token')
|
note='Downloading token')['token']
|
||||||
token = self._search_regex(
|
|
||||||
r'npoplayer\.token = "(.+?)"', token_page, 'token')
|
|
||||||
# Decryption algorithm extracted from http://npoplayer.omroep.nl/csjs/npoplayer-min.js
|
|
||||||
token_l = list(token)
|
|
||||||
first = second = None
|
|
||||||
for i in range(5, len(token_l) - 4):
|
|
||||||
if token_l[i].isdigit():
|
|
||||||
if first is None:
|
|
||||||
first = i
|
|
||||||
elif second is None:
|
|
||||||
second = i
|
|
||||||
if first is None or second is None:
|
|
||||||
first = 12
|
|
||||||
second = 13
|
|
||||||
|
|
||||||
token_l[first], token_l[second] = token_l[second], token_l[first]
|
|
||||||
|
|
||||||
return ''.join(token_l)
|
|
||||||
|
|
||||||
|
|
||||||
class NPOIE(NPOBaseIE):
|
class NPOIE(NPOBaseIE):
|
||||||
@ -58,103 +44,113 @@ class NPOIE(NPOBaseIE):
|
|||||||
(?P<id>[^/?#]+)
|
(?P<id>[^/?#]+)
|
||||||
'''
|
'''
|
||||||
|
|
||||||
_TESTS = [
|
_TESTS = [{
|
||||||
{
|
'url': 'http://www.npo.nl/nieuwsuur/22-06-2014/VPWON_1220719',
|
||||||
'url': 'http://www.npo.nl/nieuwsuur/22-06-2014/VPWON_1220719',
|
'md5': '4b3f9c429157ec4775f2c9cb7b911016',
|
||||||
'md5': '4b3f9c429157ec4775f2c9cb7b911016',
|
'info_dict': {
|
||||||
'info_dict': {
|
'id': 'VPWON_1220719',
|
||||||
'id': 'VPWON_1220719',
|
'ext': 'm4v',
|
||||||
'ext': 'm4v',
|
'title': 'Nieuwsuur',
|
||||||
'title': 'Nieuwsuur',
|
'description': 'Dagelijks tussen tien en elf: nieuws, sport en achtergronden.',
|
||||||
'description': 'Dagelijks tussen tien en elf: nieuws, sport en achtergronden.',
|
'upload_date': '20140622',
|
||||||
'upload_date': '20140622',
|
|
||||||
},
|
|
||||||
},
|
},
|
||||||
{
|
}, {
|
||||||
'url': 'http://www.npo.nl/de-mega-mike-mega-thomas-show/27-02-2009/VARA_101191800',
|
'url': 'http://www.npo.nl/de-mega-mike-mega-thomas-show/27-02-2009/VARA_101191800',
|
||||||
'md5': 'da50a5787dbfc1603c4ad80f31c5120b',
|
'md5': 'da50a5787dbfc1603c4ad80f31c5120b',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'VARA_101191800',
|
'id': 'VARA_101191800',
|
||||||
'ext': 'm4v',
|
'ext': 'm4v',
|
||||||
'title': 'De Mega Mike & Mega Thomas show: The best of.',
|
'title': 'De Mega Mike & Mega Thomas show: The best of.',
|
||||||
'description': 'md5:3b74c97fc9d6901d5a665aac0e5400f4',
|
'description': 'md5:3b74c97fc9d6901d5a665aac0e5400f4',
|
||||||
'upload_date': '20090227',
|
'upload_date': '20090227',
|
||||||
'duration': 2400,
|
'duration': 2400,
|
||||||
},
|
|
||||||
},
|
},
|
||||||
{
|
}, {
|
||||||
'url': 'http://www.npo.nl/tegenlicht/25-02-2013/VPWON_1169289',
|
'url': 'http://www.npo.nl/tegenlicht/25-02-2013/VPWON_1169289',
|
||||||
'md5': 'f8065e4e5a7824068ed3c7e783178f2c',
|
'md5': 'f8065e4e5a7824068ed3c7e783178f2c',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'VPWON_1169289',
|
'id': 'VPWON_1169289',
|
||||||
'ext': 'm4v',
|
'ext': 'm4v',
|
||||||
'title': 'Tegenlicht: De toekomst komt uit Afrika',
|
'title': 'Tegenlicht: Zwart geld. De toekomst komt uit Afrika',
|
||||||
'description': 'md5:52cf4eefbc96fffcbdc06d024147abea',
|
'description': 'md5:52cf4eefbc96fffcbdc06d024147abea',
|
||||||
'upload_date': '20130225',
|
'upload_date': '20130225',
|
||||||
'duration': 3000,
|
'duration': 3000,
|
||||||
},
|
|
||||||
},
|
},
|
||||||
{
|
}, {
|
||||||
'url': 'http://www.npo.nl/de-nieuwe-mens-deel-1/21-07-2010/WO_VPRO_043706',
|
'url': 'http://www.npo.nl/de-nieuwe-mens-deel-1/21-07-2010/WO_VPRO_043706',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'WO_VPRO_043706',
|
'id': 'WO_VPRO_043706',
|
||||||
'ext': 'wmv',
|
'ext': 'm4v',
|
||||||
'title': 'De nieuwe mens - Deel 1',
|
'title': 'De nieuwe mens - Deel 1',
|
||||||
'description': 'md5:518ae51ba1293ffb80d8d8ce90b74e4b',
|
'description': 'md5:518ae51ba1293ffb80d8d8ce90b74e4b',
|
||||||
'duration': 4680,
|
'duration': 4680,
|
||||||
},
|
|
||||||
'params': {
|
|
||||||
# mplayer mms download
|
|
||||||
'skip_download': True,
|
|
||||||
}
|
|
||||||
},
|
},
|
||||||
# non asf in streams
|
'params': {
|
||||||
{
|
'skip_download': True,
|
||||||
'url': 'http://www.npo.nl/hoe-gaat-europa-verder-na-parijs/10-01-2015/WO_NOS_762771',
|
|
||||||
'md5': 'b3da13de374cbe2d5332a7e910bef97f',
|
|
||||||
'info_dict': {
|
|
||||||
'id': 'WO_NOS_762771',
|
|
||||||
'ext': 'mp4',
|
|
||||||
'title': 'Hoe gaat Europa verder na Parijs?',
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
'url': 'http://www.ntr.nl/Aap-Poot-Pies/27/detail/Aap-poot-pies/VPWON_1233944#content',
|
|
||||||
'md5': '01c6a2841675995da1f0cf776f03a9c3',
|
|
||||||
'info_dict': {
|
|
||||||
'id': 'VPWON_1233944',
|
|
||||||
'ext': 'm4v',
|
|
||||||
'title': 'Aap, poot, pies',
|
|
||||||
'description': 'md5:c9c8005d1869ae65b858e82c01a91fde',
|
|
||||||
'upload_date': '20150508',
|
|
||||||
'duration': 599,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
'url': 'http://www.omroepwnl.nl/video/fragment/vandaag-de-dag-verkiezingen__POMS_WNL_853698',
|
|
||||||
'md5': 'd30cd8417b8b9bca1fdff27428860d08',
|
|
||||||
'info_dict': {
|
|
||||||
'id': 'POW_00996502',
|
|
||||||
'ext': 'm4v',
|
|
||||||
'title': '''"Dit is wel een 'landslide'..."''',
|
|
||||||
'description': 'md5:f8d66d537dfb641380226e31ca57b8e8',
|
|
||||||
'upload_date': '20150508',
|
|
||||||
'duration': 462,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
'url': 'http://www.zapp.nl/de-bzt-show/gemist/KN_1687547',
|
|
||||||
'only_matching': True,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
'url': 'http://www.zapp.nl/de-bzt-show/filmpjes/POMS_KN_7315118',
|
|
||||||
'only_matching': True,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
'url': 'http://www.zapp.nl/beste-vrienden-quiz/extra-video-s/WO_NTR_1067990',
|
|
||||||
'only_matching': True,
|
|
||||||
}
|
}
|
||||||
]
|
}, {
|
||||||
|
# non asf in streams
|
||||||
|
'url': 'http://www.npo.nl/hoe-gaat-europa-verder-na-parijs/10-01-2015/WO_NOS_762771',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'WO_NOS_762771',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': 'Hoe gaat Europa verder na Parijs?',
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
'skip_download': True,
|
||||||
|
}
|
||||||
|
}, {
|
||||||
|
'url': 'http://www.ntr.nl/Aap-Poot-Pies/27/detail/Aap-poot-pies/VPWON_1233944#content',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'VPWON_1233944',
|
||||||
|
'ext': 'm4v',
|
||||||
|
'title': 'Aap, poot, pies',
|
||||||
|
'description': 'md5:c9c8005d1869ae65b858e82c01a91fde',
|
||||||
|
'upload_date': '20150508',
|
||||||
|
'duration': 599,
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
'skip_download': True,
|
||||||
|
}
|
||||||
|
}, {
|
||||||
|
'url': 'http://www.omroepwnl.nl/video/fragment/vandaag-de-dag-verkiezingen__POMS_WNL_853698',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'POW_00996502',
|
||||||
|
'ext': 'm4v',
|
||||||
|
'title': '''"Dit is wel een 'landslide'..."''',
|
||||||
|
'description': 'md5:f8d66d537dfb641380226e31ca57b8e8',
|
||||||
|
'upload_date': '20150508',
|
||||||
|
'duration': 462,
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
'skip_download': True,
|
||||||
|
}
|
||||||
|
}, {
|
||||||
|
# audio
|
||||||
|
'url': 'http://www.npo.nl/jouw-stad-rotterdam/29-01-2017/RBX_FUNX_6683215/RBX_FUNX_7601437',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'RBX_FUNX_6683215',
|
||||||
|
'ext': 'mp3',
|
||||||
|
'title': 'Jouw Stad Rotterdam',
|
||||||
|
'description': 'md5:db251505244f097717ec59fabc372d9f',
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
'skip_download': True,
|
||||||
|
}
|
||||||
|
}, {
|
||||||
|
'url': 'http://www.zapp.nl/de-bzt-show/gemist/KN_1687547',
|
||||||
|
'only_matching': True,
|
||||||
|
}, {
|
||||||
|
'url': 'http://www.zapp.nl/de-bzt-show/filmpjes/POMS_KN_7315118',
|
||||||
|
'only_matching': True,
|
||||||
|
}, {
|
||||||
|
'url': 'http://www.zapp.nl/beste-vrienden-quiz/extra-video-s/WO_NTR_1067990',
|
||||||
|
'only_matching': True,
|
||||||
|
}, {
|
||||||
|
# live stream
|
||||||
|
'url': 'npo:LI_NL1_4188102',
|
||||||
|
'only_matching': True,
|
||||||
|
}]
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
video_id = self._match_id(url)
|
video_id = self._match_id(url)
|
||||||
@ -183,70 +179,115 @@ class NPOIE(NPOBaseIE):
|
|||||||
token = self._get_token(video_id)
|
token = self._get_token(video_id)
|
||||||
|
|
||||||
formats = []
|
formats = []
|
||||||
|
urls = set()
|
||||||
|
|
||||||
pubopties = metadata.get('pubopties')
|
quality = qualities(['adaptive', 'wmv_sb', 'h264_sb', 'wmv_bb', 'h264_bb', 'wvc1_std', 'h264_std'])
|
||||||
if pubopties:
|
items = self._download_json(
|
||||||
quality = qualities(['adaptive', 'wmv_sb', 'h264_sb', 'wmv_bb', 'h264_bb', 'wvc1_std', 'h264_std'])
|
'http://ida.omroep.nl/app.php/%s' % video_id, video_id,
|
||||||
for format_id in pubopties:
|
'Downloading formats JSON', query={
|
||||||
format_info = self._download_json(
|
'adaptive': 'yes',
|
||||||
'http://ida.omroep.nl/odi/?prid=%s&puboptions=%s&adaptive=yes&token=%s'
|
'token': token,
|
||||||
% (video_id, format_id, token),
|
})['items'][0]
|
||||||
video_id, 'Downloading %s JSON' % format_id)
|
for num, item in enumerate(items):
|
||||||
if format_info.get('error_code', 0) or format_info.get('errorcode', 0):
|
item_url = item.get('url')
|
||||||
|
if not item_url or item_url in urls:
|
||||||
|
continue
|
||||||
|
urls.add(item_url)
|
||||||
|
format_id = self._search_regex(
|
||||||
|
r'video/ida/([^/]+)', item_url, 'format id',
|
||||||
|
default=None)
|
||||||
|
|
||||||
|
def add_format_url(format_url):
|
||||||
|
formats.append({
|
||||||
|
'url': format_url,
|
||||||
|
'format_id': format_id,
|
||||||
|
'quality': quality(format_id),
|
||||||
|
})
|
||||||
|
|
||||||
|
# Example: http://www.npo.nl/de-nieuwe-mens-deel-1/21-07-2010/WO_VPRO_043706
|
||||||
|
if item.get('contentType') in ('url', 'audio'):
|
||||||
|
add_format_url(item_url)
|
||||||
|
continue
|
||||||
|
|
||||||
|
try:
|
||||||
|
stream_info = self._download_json(
|
||||||
|
item_url + '&type=json', video_id,
|
||||||
|
'Downloading %s stream JSON'
|
||||||
|
% item.get('label') or item.get('format') or format_id or num)
|
||||||
|
except ExtractorError as ee:
|
||||||
|
if isinstance(ee.cause, compat_HTTPError) and ee.cause.code == 404:
|
||||||
|
error = (self._parse_json(
|
||||||
|
ee.cause.read().decode(), video_id,
|
||||||
|
fatal=False) or {}).get('errorstring')
|
||||||
|
if error:
|
||||||
|
raise ExtractorError(error, expected=True)
|
||||||
|
raise
|
||||||
|
# Stream URL instead of JSON, example: npo:LI_NL1_4188102
|
||||||
|
if isinstance(stream_info, compat_str):
|
||||||
|
if not stream_info.startswith('http'):
|
||||||
continue
|
continue
|
||||||
streams = format_info.get('streams')
|
video_url = stream_info
|
||||||
if streams:
|
# JSON
|
||||||
try:
|
else:
|
||||||
video_info = self._download_json(
|
video_url = stream_info.get('url')
|
||||||
streams[0] + '&type=json',
|
if not video_url or video_url in urls:
|
||||||
video_id, 'Downloading %s stream JSON' % format_id)
|
continue
|
||||||
except ExtractorError as ee:
|
urls.add(item_url)
|
||||||
if isinstance(ee.cause, compat_HTTPError) and ee.cause.code == 404:
|
if determine_ext(video_url) == 'm3u8':
|
||||||
error = (self._parse_json(ee.cause.read().decode(), video_id, fatal=False) or {}).get('errorstring')
|
formats.extend(self._extract_m3u8_formats(
|
||||||
if error:
|
video_url, video_id, ext='mp4',
|
||||||
raise ExtractorError(error, expected=True)
|
entry_protocol='m3u8_native', m3u8_id='hls', fatal=False))
|
||||||
raise
|
else:
|
||||||
else:
|
add_format_url(video_url)
|
||||||
video_info = format_info
|
|
||||||
video_url = video_info.get('url')
|
is_live = metadata.get('medium') == 'live'
|
||||||
if not video_url:
|
|
||||||
|
if not is_live:
|
||||||
|
for num, stream in enumerate(metadata.get('streams', [])):
|
||||||
|
stream_url = stream.get('url')
|
||||||
|
if not stream_url or stream_url in urls:
|
||||||
continue
|
continue
|
||||||
if format_id == 'adaptive':
|
urls.add(stream_url)
|
||||||
formats.extend(self._extract_m3u8_formats(video_url, video_id, 'mp4'))
|
# smooth streaming is not supported
|
||||||
else:
|
stream_type = stream.get('type', '').lower()
|
||||||
|
if stream_type in ['ss', 'ms']:
|
||||||
|
continue
|
||||||
|
if stream_type == 'hds':
|
||||||
|
f4m_formats = self._extract_f4m_formats(
|
||||||
|
stream_url, video_id, fatal=False)
|
||||||
|
# f4m downloader downloads only piece of live stream
|
||||||
|
for f4m_format in f4m_formats:
|
||||||
|
f4m_format['preference'] = -1
|
||||||
|
formats.extend(f4m_formats)
|
||||||
|
elif stream_type == 'hls':
|
||||||
|
formats.extend(self._extract_m3u8_formats(
|
||||||
|
stream_url, video_id, ext='mp4', fatal=False))
|
||||||
|
# Example: http://www.npo.nl/de-nieuwe-mens-deel-1/21-07-2010/WO_VPRO_043706
|
||||||
|
elif '.asf' in stream_url:
|
||||||
|
asx = self._download_xml(
|
||||||
|
stream_url, video_id,
|
||||||
|
'Downloading stream %d ASX playlist' % num,
|
||||||
|
transform_source=fix_xml_ampersands, fatal=False)
|
||||||
|
if not asx:
|
||||||
|
continue
|
||||||
|
ref = asx.find('./ENTRY/Ref')
|
||||||
|
if ref is None:
|
||||||
|
continue
|
||||||
|
video_url = ref.get('href')
|
||||||
|
if not video_url or video_url in urls:
|
||||||
|
continue
|
||||||
|
urls.add(video_url)
|
||||||
formats.append({
|
formats.append({
|
||||||
'url': video_url,
|
'url': video_url,
|
||||||
'format_id': format_id,
|
'ext': stream.get('formaat', 'asf'),
|
||||||
'quality': quality(format_id),
|
'quality': stream.get('kwaliteit'),
|
||||||
|
'preference': -10,
|
||||||
})
|
})
|
||||||
|
else:
|
||||||
streams = metadata.get('streams')
|
|
||||||
if streams:
|
|
||||||
for i, stream in enumerate(streams):
|
|
||||||
stream_url = stream.get('url')
|
|
||||||
if not stream_url:
|
|
||||||
continue
|
|
||||||
if '.asf' not in stream_url:
|
|
||||||
formats.append({
|
formats.append({
|
||||||
'url': stream_url,
|
'url': stream_url,
|
||||||
'quality': stream.get('kwaliteit'),
|
'quality': stream.get('kwaliteit'),
|
||||||
})
|
})
|
||||||
continue
|
|
||||||
asx = self._download_xml(
|
|
||||||
stream_url, video_id,
|
|
||||||
'Downloading stream %d ASX playlist' % i,
|
|
||||||
transform_source=fix_xml_ampersands)
|
|
||||||
ref = asx.find('./ENTRY/Ref')
|
|
||||||
if ref is None:
|
|
||||||
continue
|
|
||||||
video_url = ref.get('href')
|
|
||||||
if not video_url:
|
|
||||||
continue
|
|
||||||
formats.append({
|
|
||||||
'url': video_url,
|
|
||||||
'ext': stream.get('formaat', 'asf'),
|
|
||||||
'quality': stream.get('kwaliteit'),
|
|
||||||
})
|
|
||||||
|
|
||||||
self._sort_formats(formats)
|
self._sort_formats(formats)
|
||||||
|
|
||||||
@ -259,28 +300,28 @@ class NPOIE(NPOBaseIE):
|
|||||||
|
|
||||||
return {
|
return {
|
||||||
'id': video_id,
|
'id': video_id,
|
||||||
'title': title,
|
'title': self._live_title(title) if is_live else title,
|
||||||
'description': metadata.get('info'),
|
'description': metadata.get('info'),
|
||||||
'thumbnail': metadata.get('images', [{'url': None}])[-1]['url'],
|
'thumbnail': metadata.get('images', [{'url': None}])[-1]['url'],
|
||||||
'upload_date': unified_strdate(metadata.get('gidsdatum')),
|
'upload_date': unified_strdate(metadata.get('gidsdatum')),
|
||||||
'duration': parse_duration(metadata.get('tijdsduur')),
|
'duration': parse_duration(metadata.get('tijdsduur')),
|
||||||
'formats': formats,
|
'formats': formats,
|
||||||
'subtitles': subtitles,
|
'subtitles': subtitles,
|
||||||
|
'is_live': is_live,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
class NPOLiveIE(NPOBaseIE):
|
class NPOLiveIE(NPOBaseIE):
|
||||||
IE_NAME = 'npo.nl:live'
|
IE_NAME = 'npo.nl:live'
|
||||||
_VALID_URL = r'https?://(?:www\.)?npo\.nl/live/(?P<id>.+)'
|
_VALID_URL = r'https?://(?:www\.)?npo\.nl/live/(?P<id>[^/?#&]+)'
|
||||||
|
|
||||||
_TEST = {
|
_TEST = {
|
||||||
'url': 'http://www.npo.nl/live/npo-1',
|
'url': 'http://www.npo.nl/live/npo-1',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'LI_NEDERLAND1_136692',
|
'id': 'LI_NL1_4188102',
|
||||||
'display_id': 'npo-1',
|
'display_id': 'npo-1',
|
||||||
'ext': 'mp4',
|
'ext': 'mp4',
|
||||||
'title': 're:^Nederland 1 [0-9]{4}-[0-9]{2}-[0-9]{2} [0-9]{2}:[0-9]{2}$',
|
'title': 're:^NPO 1 [0-9]{4}-[0-9]{2}-[0-9]{2} [0-9]{2}:[0-9]{2}$',
|
||||||
'description': 'Livestream',
|
|
||||||
'is_live': True,
|
'is_live': True,
|
||||||
},
|
},
|
||||||
'params': {
|
'params': {
|
||||||
@ -296,58 +337,12 @@ class NPOLiveIE(NPOBaseIE):
|
|||||||
live_id = self._search_regex(
|
live_id = self._search_regex(
|
||||||
r'data-prid="([^"]+)"', webpage, 'live id')
|
r'data-prid="([^"]+)"', webpage, 'live id')
|
||||||
|
|
||||||
metadata = self._download_json(
|
|
||||||
'http://e.omroep.nl/metadata/%s' % live_id,
|
|
||||||
display_id, transform_source=strip_jsonp)
|
|
||||||
|
|
||||||
token = self._get_token(display_id)
|
|
||||||
|
|
||||||
formats = []
|
|
||||||
|
|
||||||
streams = metadata.get('streams')
|
|
||||||
if streams:
|
|
||||||
for stream in streams:
|
|
||||||
stream_type = stream.get('type').lower()
|
|
||||||
# smooth streaming is not supported
|
|
||||||
if stream_type in ['ss', 'ms']:
|
|
||||||
continue
|
|
||||||
stream_info = self._download_json(
|
|
||||||
'http://ida.omroep.nl/aapi/?stream=%s&token=%s&type=jsonp'
|
|
||||||
% (stream.get('url'), token),
|
|
||||||
display_id, 'Downloading %s JSON' % stream_type)
|
|
||||||
if stream_info.get('error_code', 0) or stream_info.get('errorcode', 0):
|
|
||||||
continue
|
|
||||||
stream_url = self._download_json(
|
|
||||||
stream_info['stream'], display_id,
|
|
||||||
'Downloading %s URL' % stream_type,
|
|
||||||
'Unable to download %s URL' % stream_type,
|
|
||||||
transform_source=strip_jsonp, fatal=False)
|
|
||||||
if not stream_url:
|
|
||||||
continue
|
|
||||||
if stream_type == 'hds':
|
|
||||||
f4m_formats = self._extract_f4m_formats(stream_url, display_id)
|
|
||||||
# f4m downloader downloads only piece of live stream
|
|
||||||
for f4m_format in f4m_formats:
|
|
||||||
f4m_format['preference'] = -1
|
|
||||||
formats.extend(f4m_formats)
|
|
||||||
elif stream_type == 'hls':
|
|
||||||
formats.extend(self._extract_m3u8_formats(stream_url, display_id, 'mp4'))
|
|
||||||
else:
|
|
||||||
formats.append({
|
|
||||||
'url': stream_url,
|
|
||||||
'preference': -10,
|
|
||||||
})
|
|
||||||
|
|
||||||
self._sort_formats(formats)
|
|
||||||
|
|
||||||
return {
|
return {
|
||||||
|
'_type': 'url_transparent',
|
||||||
|
'url': 'npo:%s' % live_id,
|
||||||
|
'ie_key': NPOIE.ie_key(),
|
||||||
'id': live_id,
|
'id': live_id,
|
||||||
'display_id': display_id,
|
'display_id': display_id,
|
||||||
'title': self._live_title(metadata['titel']),
|
|
||||||
'description': metadata['info'],
|
|
||||||
'thumbnail': metadata.get('images', [{'url': None}])[-1]['url'],
|
|
||||||
'formats': formats,
|
|
||||||
'is_live': True,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
@ -1454,7 +1454,7 @@ class YoutubeIE(YoutubeBaseInfoExtractor):
|
|||||||
|
|
||||||
# Check for "rental" videos
|
# Check for "rental" videos
|
||||||
if 'ypc_video_rental_bar_text' in video_info and 'author' not in video_info:
|
if 'ypc_video_rental_bar_text' in video_info and 'author' not in video_info:
|
||||||
raise ExtractorError('"rental" videos not supported')
|
raise ExtractorError('"rental" videos not supported. See https://github.com/rg3/youtube-dl/issues/359 for more information.', expected=True)
|
||||||
|
|
||||||
# Start extracting information
|
# Start extracting information
|
||||||
self.report_information_extraction(video_id)
|
self.report_information_extraction(video_id)
|
||||||
@ -1851,7 +1851,7 @@ class YoutubePlaylistIE(YoutubePlaylistBaseInfoExtractor):
|
|||||||
(?:
|
(?:
|
||||||
youtube\.com/
|
youtube\.com/
|
||||||
(?:
|
(?:
|
||||||
(?:course|view_play_list|my_playlists|artist|playlist|watch|embed/videoseries)
|
(?:course|view_play_list|my_playlists|artist|playlist|watch|embed/(?:videoseries|[0-9A-Za-z_-]{11}))
|
||||||
\? (?:.*?[&;])*? (?:p|a|list)=
|
\? (?:.*?[&;])*? (?:p|a|list)=
|
||||||
| p/
|
| p/
|
||||||
)|
|
)|
|
||||||
@ -1924,6 +1924,13 @@ class YoutubePlaylistIE(YoutubePlaylistBaseInfoExtractor):
|
|||||||
'title': 'JODA15',
|
'title': 'JODA15',
|
||||||
'id': 'PL6IaIsEjSbf96XFRuNccS_RuEXwNdsoEu',
|
'id': 'PL6IaIsEjSbf96XFRuNccS_RuEXwNdsoEu',
|
||||||
}
|
}
|
||||||
|
}, {
|
||||||
|
'url': 'http://www.youtube.com/embed/_xDOZElKyNU?list=PLsyOSbh5bs16vubvKePAQ1x3PhKavfBIl',
|
||||||
|
'playlist_mincount': 485,
|
||||||
|
'info_dict': {
|
||||||
|
'title': '2017 華語最新單曲 (2/24更新)',
|
||||||
|
'id': 'PLsyOSbh5bs16vubvKePAQ1x3PhKavfBIl',
|
||||||
|
}
|
||||||
}, {
|
}, {
|
||||||
'note': 'Embedded SWF player',
|
'note': 'Embedded SWF player',
|
||||||
'url': 'https://www.youtube.com/p/YN5VISEtHet5D4NEvfTd0zcgFk84NqFZ?hl=en_US&fs=1&rel=0',
|
'url': 'https://www.youtube.com/p/YN5VISEtHet5D4NEvfTd0zcgFk84NqFZ?hl=en_US&fs=1&rel=0',
|
||||||
@ -2072,7 +2079,7 @@ class YoutubePlaylistIE(YoutubePlaylistBaseInfoExtractor):
|
|||||||
# Check if it's a video-specific URL
|
# Check if it's a video-specific URL
|
||||||
query_dict = compat_urlparse.parse_qs(compat_urlparse.urlparse(url).query)
|
query_dict = compat_urlparse.parse_qs(compat_urlparse.urlparse(url).query)
|
||||||
video_id = query_dict.get('v', [None])[0] or self._search_regex(
|
video_id = query_dict.get('v', [None])[0] or self._search_regex(
|
||||||
r'(?:^|//)youtu\.be/([0-9A-Za-z_-]{11})', url,
|
r'(?:(?:^|//)youtu\.be/|youtube\.com/embed/(?!videoseries))([0-9A-Za-z_-]{11})', url,
|
||||||
'video id', default=None)
|
'video id', default=None)
|
||||||
if video_id:
|
if video_id:
|
||||||
if self._downloader.params.get('noplaylist'):
|
if self._downloader.params.get('noplaylist'):
|
||||||
|
@ -473,7 +473,8 @@ def timeconvert(timestr):
|
|||||||
def sanitize_filename(s, restricted=False, is_id=False):
|
def sanitize_filename(s, restricted=False, is_id=False):
|
||||||
"""Sanitizes a string so it could be used as part of a filename.
|
"""Sanitizes a string so it could be used as part of a filename.
|
||||||
If restricted is set, use a stricter subset of allowed characters.
|
If restricted is set, use a stricter subset of allowed characters.
|
||||||
Set is_id if this is not an arbitrary string, but an ID that should be kept if possible
|
Set is_id if this is not an arbitrary string, but an ID that should be kept
|
||||||
|
if possible.
|
||||||
"""
|
"""
|
||||||
def replace_insane(char):
|
def replace_insane(char):
|
||||||
if restricted and char in ACCENT_CHARS:
|
if restricted and char in ACCENT_CHARS:
|
||||||
@ -3319,6 +3320,57 @@ class PerRequestProxyHandler(compat_urllib_request.ProxyHandler):
|
|||||||
self, req, proxy, type)
|
self, req, proxy, type)
|
||||||
|
|
||||||
|
|
||||||
|
# Both long_to_bytes and bytes_to_long are adapted from PyCrypto, which is
|
||||||
|
# released into Public Domain
|
||||||
|
# https://github.com/dlitz/pycrypto/blob/master/lib/Crypto/Util/number.py#L387
|
||||||
|
|
||||||
|
def long_to_bytes(n, blocksize=0):
|
||||||
|
"""long_to_bytes(n:long, blocksize:int) : string
|
||||||
|
Convert a long integer to a byte string.
|
||||||
|
|
||||||
|
If optional blocksize is given and greater than zero, pad the front of the
|
||||||
|
byte string with binary zeros so that the length is a multiple of
|
||||||
|
blocksize.
|
||||||
|
"""
|
||||||
|
# after much testing, this algorithm was deemed to be the fastest
|
||||||
|
s = b''
|
||||||
|
n = int(n)
|
||||||
|
while n > 0:
|
||||||
|
s = compat_struct_pack('>I', n & 0xffffffff) + s
|
||||||
|
n = n >> 32
|
||||||
|
# strip off leading zeros
|
||||||
|
for i in range(len(s)):
|
||||||
|
if s[i] != b'\000'[0]:
|
||||||
|
break
|
||||||
|
else:
|
||||||
|
# only happens when n == 0
|
||||||
|
s = b'\000'
|
||||||
|
i = 0
|
||||||
|
s = s[i:]
|
||||||
|
# add back some pad bytes. this could be done more efficiently w.r.t. the
|
||||||
|
# de-padding being done above, but sigh...
|
||||||
|
if blocksize > 0 and len(s) % blocksize:
|
||||||
|
s = (blocksize - len(s) % blocksize) * b'\000' + s
|
||||||
|
return s
|
||||||
|
|
||||||
|
|
||||||
|
def bytes_to_long(s):
|
||||||
|
"""bytes_to_long(string) : long
|
||||||
|
Convert a byte string to a long integer.
|
||||||
|
|
||||||
|
This is (essentially) the inverse of long_to_bytes().
|
||||||
|
"""
|
||||||
|
acc = 0
|
||||||
|
length = len(s)
|
||||||
|
if length % 4:
|
||||||
|
extra = (4 - length % 4)
|
||||||
|
s = b'\000' * extra + s
|
||||||
|
length = length + extra
|
||||||
|
for i in range(0, length, 4):
|
||||||
|
acc = (acc << 32) + compat_struct_unpack('>I', s[i:i + 4])[0]
|
||||||
|
return acc
|
||||||
|
|
||||||
|
|
||||||
def ohdave_rsa_encrypt(data, exponent, modulus):
|
def ohdave_rsa_encrypt(data, exponent, modulus):
|
||||||
'''
|
'''
|
||||||
Implement OHDave's RSA algorithm. See http://www.ohdave.com/rsa/
|
Implement OHDave's RSA algorithm. See http://www.ohdave.com/rsa/
|
||||||
@ -3336,6 +3388,21 @@ def ohdave_rsa_encrypt(data, exponent, modulus):
|
|||||||
return '%x' % encrypted
|
return '%x' % encrypted
|
||||||
|
|
||||||
|
|
||||||
|
def pkcs1pad(data, length):
|
||||||
|
"""
|
||||||
|
Padding input data with PKCS#1 scheme
|
||||||
|
|
||||||
|
@param {int[]} data input data
|
||||||
|
@param {int} length target length
|
||||||
|
@returns {int[]} padded data
|
||||||
|
"""
|
||||||
|
if len(data) > length - 11:
|
||||||
|
raise ValueError('Input data too long for PKCS#1 padding')
|
||||||
|
|
||||||
|
pseudo_random = [random.randint(0, 254) for _ in range(length - len(data) - 3)]
|
||||||
|
return [0, 2] + pseudo_random + [0] + data
|
||||||
|
|
||||||
|
|
||||||
def encode_base_n(num, n, table=None):
|
def encode_base_n(num, n, table=None):
|
||||||
FULL_TABLE = '0123456789abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ'
|
FULL_TABLE = '0123456789abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ'
|
||||||
if not table:
|
if not table:
|
||||||
|
@ -1,3 +1,3 @@
|
|||||||
from __future__ import unicode_literals
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
__version__ = '2017.02.27'
|
__version__ = '2017.03.02'
|
||||||
|
Reference in New Issue
Block a user