Compare commits
54 Commits
2017.02.24
...
2017.03.02
Author | SHA1 | Date | |
---|---|---|---|
250eea6821 | |||
28d15b73f8 | |||
11bb6ad1a5 | |||
c9612c0487 | |||
af5049f128 | |||
158af5242e | |||
40df485f55 | |||
4b8a984c67 | |||
83e8fce628 | |||
aa9cc2ecbf | |||
1dc24093f8 | |||
11bae9cdde | |||
43b38424a9 | |||
948519b35d | |||
87dadd456a | |||
7c4aa6fd6f | |||
9bd05b5a18 | |||
0a5445ddbe | |||
f48409c7ac | |||
c9619f0a17 | |||
f4c68ba372 | |||
ef48a1175d | |||
c6184bcf7b | |||
18abb74376 | |||
dbc01fdb6f | |||
f264c62334 | |||
0dc5a86a32 | |||
0e879f432a | |||
892b47ab6c | |||
fdeea72611 | |||
7fd4655256 | |||
fd5c4aab59 | |||
8878789f11 | |||
a5cf17989b | |||
b3aec47665 | |||
9d0c08a02c | |||
e498758b9c | |||
5fc8d89361 | |||
d374d943f3 | |||
103f8c8d36 | |||
922ab7840b | |||
831217291a | |||
db182c63fb | |||
eeb0a95684 | |||
231bcd0b6b | |||
204efc8509 | |||
5d3a51e1b9 | |||
ad3033037c | |||
f3bc281239 | |||
441d7a32e5 | |||
51ed496307 | |||
68f17a9c2d | |||
39e7277ed1 | |||
42dcdbe11c |
6
.github/ISSUE_TEMPLATE.md
vendored
6
.github/ISSUE_TEMPLATE.md
vendored
@ -6,8 +6,8 @@
|
|||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
### Make sure you are using the *latest* version: run `youtube-dl --version` and ensure your version is *2017.02.24*. If it's not read [this FAQ entry](https://github.com/rg3/youtube-dl/blob/master/README.md#how-do-i-update-youtube-dl) and update. Issues with outdated version will be rejected.
|
### Make sure you are using the *latest* version: run `youtube-dl --version` and ensure your version is *2017.03.02*. If it's not read [this FAQ entry](https://github.com/rg3/youtube-dl/blob/master/README.md#how-do-i-update-youtube-dl) and update. Issues with outdated version will be rejected.
|
||||||
- [ ] I've **verified** and **I assure** that I'm running youtube-dl **2017.02.24**
|
- [ ] I've **verified** and **I assure** that I'm running youtube-dl **2017.03.02**
|
||||||
|
|
||||||
### Before submitting an *issue* make sure you have:
|
### Before submitting an *issue* make sure you have:
|
||||||
- [ ] At least skimmed through [README](https://github.com/rg3/youtube-dl/blob/master/README.md) and **most notably** [FAQ](https://github.com/rg3/youtube-dl#faq) and [BUGS](https://github.com/rg3/youtube-dl#bugs) sections
|
- [ ] At least skimmed through [README](https://github.com/rg3/youtube-dl/blob/master/README.md) and **most notably** [FAQ](https://github.com/rg3/youtube-dl#faq) and [BUGS](https://github.com/rg3/youtube-dl#bugs) sections
|
||||||
@ -35,7 +35,7 @@ $ youtube-dl -v <your command line>
|
|||||||
[debug] User config: []
|
[debug] User config: []
|
||||||
[debug] Command-line args: [u'-v', u'http://www.youtube.com/watch?v=BaW_jenozKcj']
|
[debug] Command-line args: [u'-v', u'http://www.youtube.com/watch?v=BaW_jenozKcj']
|
||||||
[debug] Encodings: locale cp1251, fs mbcs, out cp866, pref cp1251
|
[debug] Encodings: locale cp1251, fs mbcs, out cp866, pref cp1251
|
||||||
[debug] youtube-dl version 2017.02.24
|
[debug] youtube-dl version 2017.03.02
|
||||||
[debug] Python version 2.7.11 - Windows-2003Server-5.2.3790-SP2
|
[debug] Python version 2.7.11 - Windows-2003Server-5.2.3790-SP2
|
||||||
[debug] exe versions: ffmpeg N-75573-g1d0487f, ffprobe N-75573-g1d0487f, rtmpdump 2.4
|
[debug] exe versions: ffmpeg N-75573-g1d0487f, ffprobe N-75573-g1d0487f, rtmpdump 2.4
|
||||||
[debug] Proxy map: {}
|
[debug] Proxy map: {}
|
||||||
|
66
ChangeLog
66
ChangeLog
@ -1,3 +1,69 @@
|
|||||||
|
version 2017.03.02
|
||||||
|
|
||||||
|
Core
|
||||||
|
+ [adobepass] Add support for Charter Spectrum (#11465)
|
||||||
|
* [YoutubeDL] Don't sanitize identifiers in output template (#12317)
|
||||||
|
|
||||||
|
Extractors
|
||||||
|
* [facebook] Fix extraction (#12323, #12330)
|
||||||
|
* [youtube] Mark errors about rental videos as expected (#12324)
|
||||||
|
+ [npo] Add support for audio
|
||||||
|
* [npo] Adapt to app.php API (#12311, #12320)
|
||||||
|
|
||||||
|
|
||||||
|
version 2017.02.28
|
||||||
|
|
||||||
|
Core
|
||||||
|
+ [utils] Add bytes_to_long and long_to_bytes
|
||||||
|
+ [utils] Add pkcs1pad
|
||||||
|
+ [aes] Add aes_cbc_encrypt
|
||||||
|
|
||||||
|
Extractors
|
||||||
|
+ [azmedien:showplaylist] Add support for show playlists (#12160)
|
||||||
|
+ [youtube:playlist] Recognize another playlist pattern (#11928, #12286)
|
||||||
|
+ [daisuki] Add support for daisuki.net (#2486, #3186, #4738, #6175, #7776,
|
||||||
|
#10060)
|
||||||
|
* [douyu] Fix extraction (#12301)
|
||||||
|
|
||||||
|
|
||||||
|
version 2017.02.27
|
||||||
|
|
||||||
|
Core
|
||||||
|
* [downloader/common] Limit displaying 2 digits after decimal point in sleep
|
||||||
|
interval message (#12183)
|
||||||
|
+ [extractor/common] Add preference to _parse_html5_media_entries
|
||||||
|
|
||||||
|
Extractors
|
||||||
|
+ [npo] Add support for zapp.nl
|
||||||
|
+ [npo] Add support for hetklokhuis.nl (#12293)
|
||||||
|
- [scivee] Remove extractor (#9315)
|
||||||
|
+ [cda] Decode download URL (#12255)
|
||||||
|
+ [crunchyroll] Improve uploader extraction (#12267)
|
||||||
|
+ [youtube] Raise GeoRestrictedError
|
||||||
|
+ [dailymotion] Raise GeoRestrictedError
|
||||||
|
+ [mdr] Recognize more URL patterns (#12169)
|
||||||
|
+ [tvigle] Raise GeoRestrictedError
|
||||||
|
* [vevo] Fix extraction for videos with the new streams/streamsV3 format
|
||||||
|
(#11719)
|
||||||
|
+ [freshlive] Add support for freshlive.tv (#12175)
|
||||||
|
+ [xhamster] Capture and output videoClosed error (#12263)
|
||||||
|
+ [etonline] Add support for etonline.com (#12236)
|
||||||
|
+ [njpwworld] Add support for njpwworld.com (#11561)
|
||||||
|
* [amcnetworks] Relax URL regular expression (#12127)
|
||||||
|
|
||||||
|
|
||||||
|
version 2017.02.24.1
|
||||||
|
|
||||||
|
Extractors
|
||||||
|
* [noco] Modernize
|
||||||
|
* [noco] Switch login URL to https (#12246)
|
||||||
|
+ [thescene] Extract more metadata
|
||||||
|
* [thescene] Fix extraction (#12235)
|
||||||
|
+ [tubitv] Use geo bypass mechanism
|
||||||
|
* [openload] Fix extraction (#10408)
|
||||||
|
+ [ivi] Raise GeoRestrictedError
|
||||||
|
|
||||||
|
|
||||||
version 2017.02.24
|
version 2017.02.24
|
||||||
|
|
||||||
Core
|
Core
|
||||||
|
@ -78,6 +78,7 @@
|
|||||||
- **awaan:video**
|
- **awaan:video**
|
||||||
- **AZMedien**: AZ Medien videos
|
- **AZMedien**: AZ Medien videos
|
||||||
- **AZMedienPlaylist**: AZ Medien playlists
|
- **AZMedienPlaylist**: AZ Medien playlists
|
||||||
|
- **AZMedienShowPlaylist**: AZ Medien show playlists
|
||||||
- **Azubu**
|
- **Azubu**
|
||||||
- **AzubuLive**
|
- **AzubuLive**
|
||||||
- **BaiduVideo**: 百度视频
|
- **BaiduVideo**: 百度视频
|
||||||
@ -191,6 +192,8 @@
|
|||||||
- **dailymotion:playlist**
|
- **dailymotion:playlist**
|
||||||
- **dailymotion:user**
|
- **dailymotion:user**
|
||||||
- **DailymotionCloud**
|
- **DailymotionCloud**
|
||||||
|
- **Daisuki**
|
||||||
|
- **DaisukiPlaylist**
|
||||||
- **daum.net**
|
- **daum.net**
|
||||||
- **daum.net:clip**
|
- **daum.net:clip**
|
||||||
- **daum.net:playlist**
|
- **daum.net:playlist**
|
||||||
@ -239,6 +242,7 @@
|
|||||||
- **ESPN**
|
- **ESPN**
|
||||||
- **ESPNArticle**
|
- **ESPNArticle**
|
||||||
- **EsriVideo**
|
- **EsriVideo**
|
||||||
|
- **ETOnline**
|
||||||
- **Europa**
|
- **Europa**
|
||||||
- **EveryonesMixtape**
|
- **EveryonesMixtape**
|
||||||
- **ExpoTV**
|
- **ExpoTV**
|
||||||
@ -274,6 +278,7 @@
|
|||||||
- **francetvinfo.fr**
|
- **francetvinfo.fr**
|
||||||
- **Freesound**
|
- **Freesound**
|
||||||
- **freespeech.org**
|
- **freespeech.org**
|
||||||
|
- **FreshLive**
|
||||||
- **Funimation**
|
- **Funimation**
|
||||||
- **FunnyOrDie**
|
- **FunnyOrDie**
|
||||||
- **Fusion**
|
- **Fusion**
|
||||||
@ -310,6 +315,7 @@
|
|||||||
- **HellPorno**
|
- **HellPorno**
|
||||||
- **Helsinki**: helsinki.fi
|
- **Helsinki**: helsinki.fi
|
||||||
- **HentaiStigma**
|
- **HentaiStigma**
|
||||||
|
- **hetklokhuis**
|
||||||
- **hgtv.com:show**
|
- **hgtv.com:show**
|
||||||
- **HistoricFilms**
|
- **HistoricFilms**
|
||||||
- **history:topic**: History.com Topic
|
- **history:topic**: History.com Topic
|
||||||
@ -511,6 +517,7 @@
|
|||||||
- **Nintendo**
|
- **Nintendo**
|
||||||
- **njoy**: N-JOY
|
- **njoy**: N-JOY
|
||||||
- **njoy:embed**
|
- **njoy:embed**
|
||||||
|
- **NJPWWorld**: 新日本プロレスワールド
|
||||||
- **NobelPrize**
|
- **NobelPrize**
|
||||||
- **Noco**
|
- **Noco**
|
||||||
- **Normalboots**
|
- **Normalboots**
|
||||||
@ -666,7 +673,6 @@
|
|||||||
- **savefrom.net**
|
- **savefrom.net**
|
||||||
- **SBS**: sbs.com.au
|
- **SBS**: sbs.com.au
|
||||||
- **schooltv**
|
- **schooltv**
|
||||||
- **SciVee**
|
|
||||||
- **screen.yahoo:search**: Yahoo screen search
|
- **screen.yahoo:search**: Yahoo screen search
|
||||||
- **Screencast**
|
- **Screencast**
|
||||||
- **ScreencastOMatic**
|
- **ScreencastOMatic**
|
||||||
|
@ -8,7 +8,7 @@ import sys
|
|||||||
import unittest
|
import unittest
|
||||||
sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
|
sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
|
||||||
|
|
||||||
from youtube_dl.aes import aes_decrypt, aes_encrypt, aes_cbc_decrypt, aes_decrypt_text
|
from youtube_dl.aes import aes_decrypt, aes_encrypt, aes_cbc_decrypt, aes_cbc_encrypt, aes_decrypt_text
|
||||||
from youtube_dl.utils import bytes_to_intlist, intlist_to_bytes
|
from youtube_dl.utils import bytes_to_intlist, intlist_to_bytes
|
||||||
import base64
|
import base64
|
||||||
|
|
||||||
@ -34,6 +34,13 @@ class TestAES(unittest.TestCase):
|
|||||||
decrypted = intlist_to_bytes(aes_cbc_decrypt(data, self.key, self.iv))
|
decrypted = intlist_to_bytes(aes_cbc_decrypt(data, self.key, self.iv))
|
||||||
self.assertEqual(decrypted.rstrip(b'\x08'), self.secret_msg)
|
self.assertEqual(decrypted.rstrip(b'\x08'), self.secret_msg)
|
||||||
|
|
||||||
|
def test_cbc_encrypt(self):
|
||||||
|
data = bytes_to_intlist(self.secret_msg)
|
||||||
|
encrypted = intlist_to_bytes(aes_cbc_encrypt(data, self.key, self.iv))
|
||||||
|
self.assertEqual(
|
||||||
|
encrypted,
|
||||||
|
b"\x97\x92+\xe5\x0b\xc3\x18\x91ky9m&\xb3\xb5@\xe6'\xc2\x96.\xc8u\x88\xab9-[\x9e|\xf1\xcd")
|
||||||
|
|
||||||
def test_decrypt_text(self):
|
def test_decrypt_text(self):
|
||||||
password = intlist_to_bytes(self.key).decode('utf-8')
|
password = intlist_to_bytes(self.key).decode('utf-8')
|
||||||
encrypted = base64.b64encode(
|
encrypted = base64.b64encode(
|
||||||
|
@ -52,6 +52,7 @@ from youtube_dl.utils import (
|
|||||||
parse_filesize,
|
parse_filesize,
|
||||||
parse_count,
|
parse_count,
|
||||||
parse_iso8601,
|
parse_iso8601,
|
||||||
|
pkcs1pad,
|
||||||
read_batch_urls,
|
read_batch_urls,
|
||||||
sanitize_filename,
|
sanitize_filename,
|
||||||
sanitize_path,
|
sanitize_path,
|
||||||
@ -1104,6 +1105,14 @@ The first line
|
|||||||
ohdave_rsa_encrypt(b'aa111222', e, N),
|
ohdave_rsa_encrypt(b'aa111222', e, N),
|
||||||
'726664bd9a23fd0c70f9f1b84aab5e3905ce1e45a584e9cbcf9bcc7510338fc1986d6c599ff990d923aa43c51c0d9013cd572e13bc58f4ae48f2ed8c0b0ba881')
|
'726664bd9a23fd0c70f9f1b84aab5e3905ce1e45a584e9cbcf9bcc7510338fc1986d6c599ff990d923aa43c51c0d9013cd572e13bc58f4ae48f2ed8c0b0ba881')
|
||||||
|
|
||||||
|
def test_pkcs1pad(self):
|
||||||
|
data = [1, 2, 3]
|
||||||
|
padded_data = pkcs1pad(data, 32)
|
||||||
|
self.assertEqual(padded_data[:2], [0, 2])
|
||||||
|
self.assertEqual(padded_data[28:], [0, 1, 2, 3])
|
||||||
|
|
||||||
|
self.assertRaises(ValueError, pkcs1pad, data, 8)
|
||||||
|
|
||||||
def test_encode_base_n(self):
|
def test_encode_base_n(self):
|
||||||
self.assertEqual(encode_base_n(0, 30), '0')
|
self.assertEqual(encode_base_n(0, 30), '0')
|
||||||
self.assertEqual(encode_base_n(80, 30), '2k')
|
self.assertEqual(encode_base_n(80, 30), '2k')
|
||||||
|
@ -616,7 +616,7 @@ class YoutubeDL(object):
|
|||||||
sanitize = lambda k, v: sanitize_filename(
|
sanitize = lambda k, v: sanitize_filename(
|
||||||
compat_str(v),
|
compat_str(v),
|
||||||
restricted=self.params.get('restrictfilenames'),
|
restricted=self.params.get('restrictfilenames'),
|
||||||
is_id=(k == 'id'))
|
is_id=(k == 'id' or k.endswith('_id')))
|
||||||
template_dict = dict((k, v if isinstance(v, compat_numeric_types) else sanitize(k, v))
|
template_dict = dict((k, v if isinstance(v, compat_numeric_types) else sanitize(k, v))
|
||||||
for k, v in template_dict.items()
|
for k, v in template_dict.items()
|
||||||
if v is not None and not isinstance(v, (list, tuple, dict)))
|
if v is not None and not isinstance(v, (list, tuple, dict)))
|
||||||
|
@ -60,6 +60,34 @@ def aes_cbc_decrypt(data, key, iv):
|
|||||||
return decrypted_data
|
return decrypted_data
|
||||||
|
|
||||||
|
|
||||||
|
def aes_cbc_encrypt(data, key, iv):
|
||||||
|
"""
|
||||||
|
Encrypt with aes in CBC mode. Using PKCS#7 padding
|
||||||
|
|
||||||
|
@param {int[]} data cleartext
|
||||||
|
@param {int[]} key 16/24/32-Byte cipher key
|
||||||
|
@param {int[]} iv 16-Byte IV
|
||||||
|
@returns {int[]} encrypted data
|
||||||
|
"""
|
||||||
|
expanded_key = key_expansion(key)
|
||||||
|
block_count = int(ceil(float(len(data)) / BLOCK_SIZE_BYTES))
|
||||||
|
|
||||||
|
encrypted_data = []
|
||||||
|
previous_cipher_block = iv
|
||||||
|
for i in range(block_count):
|
||||||
|
block = data[i * BLOCK_SIZE_BYTES: (i + 1) * BLOCK_SIZE_BYTES]
|
||||||
|
remaining_length = BLOCK_SIZE_BYTES - len(block)
|
||||||
|
block += [remaining_length] * remaining_length
|
||||||
|
mixed_block = xor(block, previous_cipher_block)
|
||||||
|
|
||||||
|
encrypted_block = aes_encrypt(mixed_block, expanded_key)
|
||||||
|
encrypted_data += encrypted_block
|
||||||
|
|
||||||
|
previous_cipher_block = encrypted_block
|
||||||
|
|
||||||
|
return encrypted_data
|
||||||
|
|
||||||
|
|
||||||
def key_expansion(data):
|
def key_expansion(data):
|
||||||
"""
|
"""
|
||||||
Generate key schedule
|
Generate key schedule
|
||||||
|
@ -2760,8 +2760,10 @@ else:
|
|||||||
compat_kwargs = lambda kwargs: kwargs
|
compat_kwargs = lambda kwargs: kwargs
|
||||||
|
|
||||||
|
|
||||||
compat_numeric_types = ((int, float, long, complex) if sys.version_info[0] < 3
|
try:
|
||||||
else (int, float, complex))
|
compat_numeric_types = (int, float, long, complex)
|
||||||
|
except NameError: # Python 3
|
||||||
|
compat_numeric_types = (int, float, complex)
|
||||||
|
|
||||||
|
|
||||||
if sys.version_info < (2, 7):
|
if sys.version_info < (2, 7):
|
||||||
|
@ -347,7 +347,10 @@ class FileDownloader(object):
|
|||||||
if min_sleep_interval:
|
if min_sleep_interval:
|
||||||
max_sleep_interval = self.params.get('max_sleep_interval', min_sleep_interval)
|
max_sleep_interval = self.params.get('max_sleep_interval', min_sleep_interval)
|
||||||
sleep_interval = random.uniform(min_sleep_interval, max_sleep_interval)
|
sleep_interval = random.uniform(min_sleep_interval, max_sleep_interval)
|
||||||
self.to_screen('[download] Sleeping %s seconds...' % sleep_interval)
|
self.to_screen(
|
||||||
|
'[download] Sleeping %s seconds...' % (
|
||||||
|
int(sleep_interval) if sleep_interval.is_integer()
|
||||||
|
else '%.2f' % sleep_interval))
|
||||||
time.sleep(sleep_interval)
|
time.sleep(sleep_interval)
|
||||||
|
|
||||||
return self.real_download(filename, info_dict)
|
return self.real_download(filename, info_dict)
|
||||||
|
@ -36,6 +36,11 @@ MSO_INFO = {
|
|||||||
'username_field': 'Ecom_User_ID',
|
'username_field': 'Ecom_User_ID',
|
||||||
'password_field': 'Ecom_Password',
|
'password_field': 'Ecom_Password',
|
||||||
},
|
},
|
||||||
|
'Charter_Direct': {
|
||||||
|
'name': 'Charter Spectrum',
|
||||||
|
'username_field': 'IDToken1',
|
||||||
|
'password_field': 'IDToken2',
|
||||||
|
},
|
||||||
'thr030': {
|
'thr030': {
|
||||||
'name': '3 Rivers Communications'
|
'name': '3 Rivers Communications'
|
||||||
},
|
},
|
||||||
|
@ -10,7 +10,7 @@ from ..utils import (
|
|||||||
|
|
||||||
|
|
||||||
class AMCNetworksIE(ThePlatformIE):
|
class AMCNetworksIE(ThePlatformIE):
|
||||||
_VALID_URL = r'https?://(?:www\.)?(?:amc|bbcamerica|ifc|wetv)\.com/(?:movies/|shows/[^/]+/(?:full-episodes/)?[^/]+/episode-\d+(?:-(?:[^/]+/)?|/))(?P<id>[^/?#]+)'
|
_VALID_URL = r'https?://(?:www\.)?(?:amc|bbcamerica|ifc|wetv)\.com/(?:movies|shows(?:/[^/]+)+)/(?P<id>[^/?#]+)'
|
||||||
_TESTS = [{
|
_TESTS = [{
|
||||||
'url': 'http://www.ifc.com/shows/maron/season-04/episode-01/step-1',
|
'url': 'http://www.ifc.com/shows/maron/season-04/episode-01/step-1',
|
||||||
'md5': '',
|
'md5': '',
|
||||||
@ -44,6 +44,12 @@ class AMCNetworksIE(ThePlatformIE):
|
|||||||
}, {
|
}, {
|
||||||
'url': 'http://www.bbcamerica.com/shows/doctor-who/full-episodes/the-power-of-the-daleks/episode-01-episode-1-color-version',
|
'url': 'http://www.bbcamerica.com/shows/doctor-who/full-episodes/the-power-of-the-daleks/episode-01-episode-1-color-version',
|
||||||
'only_matching': True,
|
'only_matching': True,
|
||||||
|
}, {
|
||||||
|
'url': 'http://www.wetv.com/shows/mama-june-from-not-to-hot/full-episode/season-01/thin-tervention',
|
||||||
|
'only_matching': True,
|
||||||
|
}, {
|
||||||
|
'url': 'http://www.wetv.com/shows/la-hair/videos/season-05/episode-09-episode-9-2/episode-9-sneak-peek-3',
|
||||||
|
'only_matching': True,
|
||||||
}]
|
}]
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
|
@ -1,3 +1,4 @@
|
|||||||
|
# coding: utf-8
|
||||||
from __future__ import unicode_literals
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
import re
|
import re
|
||||||
@ -5,6 +6,7 @@ import re
|
|||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from .kaltura import KalturaIE
|
from .kaltura import KalturaIE
|
||||||
from ..utils import (
|
from ..utils import (
|
||||||
|
get_element_by_class,
|
||||||
get_element_by_id,
|
get_element_by_id,
|
||||||
strip_or_none,
|
strip_or_none,
|
||||||
urljoin,
|
urljoin,
|
||||||
@ -170,3 +172,42 @@ class AZMedienPlaylistIE(AZMedienBaseIE):
|
|||||||
'video-title', webpage)), group='title')
|
'video-title', webpage)), group='title')
|
||||||
|
|
||||||
return self.playlist_result(entries, show_id, title)
|
return self.playlist_result(entries, show_id, title)
|
||||||
|
|
||||||
|
|
||||||
|
class AZMedienShowPlaylistIE(AZMedienBaseIE):
|
||||||
|
IE_DESC = 'AZ Medien show playlists'
|
||||||
|
_VALID_URL = r'''(?x)
|
||||||
|
https?://
|
||||||
|
(?:www\.)?
|
||||||
|
(?:
|
||||||
|
telezueri\.ch|
|
||||||
|
telebaern\.tv|
|
||||||
|
telem1\.ch
|
||||||
|
)/
|
||||||
|
(?:
|
||||||
|
all-episodes|
|
||||||
|
alle-episoden
|
||||||
|
)/
|
||||||
|
(?P<id>[^/?#&]+)
|
||||||
|
'''
|
||||||
|
|
||||||
|
_TEST = {
|
||||||
|
'url': 'http://www.telezueri.ch/all-episodes/astrotalk',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'astrotalk',
|
||||||
|
'title': 'TeleZüri: AstroTalk - alle episoden',
|
||||||
|
'description': 'md5:4c0f7e7d741d906004266e295ceb4a26',
|
||||||
|
},
|
||||||
|
'playlist_mincount': 13,
|
||||||
|
}
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
playlist_id = self._match_id(url)
|
||||||
|
webpage = self._download_webpage(url, playlist_id)
|
||||||
|
episodes = get_element_by_class('search-mobile-box', webpage)
|
||||||
|
entries = [self.url_result(
|
||||||
|
urljoin(url, m.group('url'))) for m in re.finditer(
|
||||||
|
r'<a[^>]+href=(["\'])(?P<url>(?:(?!\1).)+)\1', episodes)]
|
||||||
|
title = self._og_search_title(webpage, fatal=False)
|
||||||
|
description = self._og_search_description(webpage)
|
||||||
|
return self.playlist_result(entries, playlist_id, title, description)
|
||||||
|
@ -1,6 +1,7 @@
|
|||||||
# coding: utf-8
|
# coding: utf-8
|
||||||
from __future__ import unicode_literals
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
|
import codecs
|
||||||
import re
|
import re
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
@ -96,6 +97,10 @@ class CDAIE(InfoExtractor):
|
|||||||
if not video or 'file' not in video:
|
if not video or 'file' not in video:
|
||||||
self.report_warning('Unable to extract %s version information' % version)
|
self.report_warning('Unable to extract %s version information' % version)
|
||||||
return
|
return
|
||||||
|
if video['file'].startswith('uggc'):
|
||||||
|
video['file'] = codecs.decode(video['file'], 'rot_13')
|
||||||
|
if video['file'].endswith('adc.mp4'):
|
||||||
|
video['file'] = video['file'].replace('adc.mp4', '.mp4')
|
||||||
f = {
|
f = {
|
||||||
'url': video['file'],
|
'url': video['file'],
|
||||||
}
|
}
|
||||||
|
@ -2010,7 +2010,7 @@ class InfoExtractor(object):
|
|||||||
})
|
})
|
||||||
return formats
|
return formats
|
||||||
|
|
||||||
def _parse_html5_media_entries(self, base_url, webpage, video_id, m3u8_id=None, m3u8_entry_protocol='m3u8', mpd_id=None):
|
def _parse_html5_media_entries(self, base_url, webpage, video_id, m3u8_id=None, m3u8_entry_protocol='m3u8', mpd_id=None, preference=None):
|
||||||
def absolute_url(video_url):
|
def absolute_url(video_url):
|
||||||
return compat_urlparse.urljoin(base_url, video_url)
|
return compat_urlparse.urljoin(base_url, video_url)
|
||||||
|
|
||||||
@ -2032,7 +2032,8 @@ class InfoExtractor(object):
|
|||||||
is_plain_url = False
|
is_plain_url = False
|
||||||
formats = self._extract_m3u8_formats(
|
formats = self._extract_m3u8_formats(
|
||||||
full_url, video_id, ext='mp4',
|
full_url, video_id, ext='mp4',
|
||||||
entry_protocol=m3u8_entry_protocol, m3u8_id=m3u8_id)
|
entry_protocol=m3u8_entry_protocol, m3u8_id=m3u8_id,
|
||||||
|
preference=preference)
|
||||||
elif ext == 'mpd':
|
elif ext == 'mpd':
|
||||||
is_plain_url = False
|
is_plain_url = False
|
||||||
formats = self._extract_mpd_formats(
|
formats = self._extract_mpd_formats(
|
||||||
|
@ -207,6 +207,21 @@ class CrunchyrollIE(CrunchyrollBaseIE):
|
|||||||
# Just test metadata extraction
|
# Just test metadata extraction
|
||||||
'skip_download': True,
|
'skip_download': True,
|
||||||
},
|
},
|
||||||
|
}, {
|
||||||
|
# make sure we can extract an uploader name that's not a link
|
||||||
|
'url': 'http://www.crunchyroll.com/hakuoki-reimeiroku/episode-1-dawn-of-the-divine-warriors-606899',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '606899',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': 'Hakuoki Reimeiroku Episode 1 – Dawn of the Divine Warriors',
|
||||||
|
'description': 'Ryunosuke was left to die, but Serizawa-san asked him a simple question "Do you want to live?"',
|
||||||
|
'uploader': 'Geneon Entertainment',
|
||||||
|
'upload_date': '20120717',
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
# just test metadata extraction
|
||||||
|
'skip_download': True,
|
||||||
|
},
|
||||||
}]
|
}]
|
||||||
|
|
||||||
_FORMAT_IDS = {
|
_FORMAT_IDS = {
|
||||||
@ -388,8 +403,9 @@ Format: Layer, Start, End, Style, Name, MarginL, MarginR, MarginV, Effect, Text
|
|||||||
if video_upload_date:
|
if video_upload_date:
|
||||||
video_upload_date = unified_strdate(video_upload_date)
|
video_upload_date = unified_strdate(video_upload_date)
|
||||||
video_uploader = self._html_search_regex(
|
video_uploader = self._html_search_regex(
|
||||||
r'<a[^>]+href="/publisher/[^"]+"[^>]*>([^<]+)</a>', webpage,
|
# try looking for both an uploader that's a link and one that's not
|
||||||
'video_uploader', fatal=False)
|
[r'<a[^>]+href="/publisher/[^"]+"[^>]*>([^<]+)</a>', r'<div>\s*Publisher:\s*<span>\s*(.+?)\s*</span>\s*</div>'],
|
||||||
|
webpage, 'video_uploader', fatal=False)
|
||||||
|
|
||||||
available_fmts = []
|
available_fmts = []
|
||||||
for a, fmt in re.findall(r'(<a[^>]+token=["\']showmedia\.([0-9]{3,4})p["\'][^>]+>)', webpage):
|
for a, fmt in re.findall(r'(<a[^>]+token=["\']showmedia\.([0-9]{3,4})p["\'][^>]+>)', webpage):
|
||||||
|
@ -282,9 +282,14 @@ class DailymotionIE(DailymotionBaseInfoExtractor):
|
|||||||
}
|
}
|
||||||
|
|
||||||
def _check_error(self, info):
|
def _check_error(self, info):
|
||||||
|
error = info.get('error')
|
||||||
if info.get('error') is not None:
|
if info.get('error') is not None:
|
||||||
|
title = error['title']
|
||||||
|
# See https://developer.dailymotion.com/api#access-error
|
||||||
|
if error.get('code') == 'DM007':
|
||||||
|
self.raise_geo_restricted(msg=title)
|
||||||
raise ExtractorError(
|
raise ExtractorError(
|
||||||
'%s said: %s' % (self.IE_NAME, info['error']['title']), expected=True)
|
'%s said: %s' % (self.IE_NAME, title), expected=True)
|
||||||
|
|
||||||
def _get_subtitles(self, video_id, webpage):
|
def _get_subtitles(self, video_id, webpage):
|
||||||
try:
|
try:
|
||||||
|
159
youtube_dl/extractor/daisuki.py
Normal file
159
youtube_dl/extractor/daisuki.py
Normal file
@ -0,0 +1,159 @@
|
|||||||
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
|
import base64
|
||||||
|
import json
|
||||||
|
import random
|
||||||
|
import re
|
||||||
|
|
||||||
|
from .common import InfoExtractor
|
||||||
|
from ..aes import (
|
||||||
|
aes_cbc_decrypt,
|
||||||
|
aes_cbc_encrypt,
|
||||||
|
)
|
||||||
|
from ..utils import (
|
||||||
|
bytes_to_intlist,
|
||||||
|
bytes_to_long,
|
||||||
|
clean_html,
|
||||||
|
ExtractorError,
|
||||||
|
intlist_to_bytes,
|
||||||
|
get_element_by_id,
|
||||||
|
js_to_json,
|
||||||
|
int_or_none,
|
||||||
|
long_to_bytes,
|
||||||
|
pkcs1pad,
|
||||||
|
remove_end,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class DaisukiIE(InfoExtractor):
|
||||||
|
_VALID_URL = r'https?://(?:www\.)?daisuki\.net/[^/]+/[^/]+/[^/]+/watch\.[^.]+\.(?P<id>\d+)\.html'
|
||||||
|
|
||||||
|
_TEST = {
|
||||||
|
'url': 'http://www.daisuki.net/tw/en/anime/watch.TheIdolMasterCG.11213.html',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '11213',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': '#01 Who is in the pumpkin carriage? - THE IDOLM@STER CINDERELLA GIRLS',
|
||||||
|
'subtitles': {
|
||||||
|
'mul': [{
|
||||||
|
'ext': 'ttml',
|
||||||
|
}],
|
||||||
|
},
|
||||||
|
'creator': 'BANDAI NAMCO Entertainment',
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
'skip_download': True, # AES-encrypted HLS stream
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
# The public key in PEM format can be found in clientlibs_anime_watch.min.js
|
||||||
|
_RSA_KEY = (0xc5524c25e8e14b366b3754940beeb6f96cb7e2feef0b932c7659a0c5c3bf173d602464c2df73d693b513ae06ff1be8f367529ab30bf969c5640522181f2a0c51ea546ae120d3d8d908595e4eff765b389cde080a1ef7f1bbfb07411cc568db73b7f521cedf270cbfbe0ddbc29b1ac9d0f2d8f4359098caffee6d07915020077d, 65537)
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
video_id = self._match_id(url)
|
||||||
|
|
||||||
|
webpage = self._download_webpage(url, video_id)
|
||||||
|
|
||||||
|
flashvars = self._parse_json(self._search_regex(
|
||||||
|
r'(?s)var\s+flashvars\s*=\s*({.+?});', webpage, 'flashvars'),
|
||||||
|
video_id, transform_source=js_to_json)
|
||||||
|
|
||||||
|
iv = [0] * 16
|
||||||
|
|
||||||
|
data = {}
|
||||||
|
for key in ('device_cd', 'mv_id', 'ss1_prm', 'ss2_prm', 'ss3_prm', 'ss_id'):
|
||||||
|
data[key] = flashvars.get(key, '')
|
||||||
|
|
||||||
|
encrypted_rtn = None
|
||||||
|
|
||||||
|
# Some AES keys are rejected. Try it with different AES keys
|
||||||
|
for idx in range(5):
|
||||||
|
aes_key = [random.randint(0, 254) for _ in range(32)]
|
||||||
|
padded_aeskey = intlist_to_bytes(pkcs1pad(aes_key, 128))
|
||||||
|
|
||||||
|
n, e = self._RSA_KEY
|
||||||
|
encrypted_aeskey = long_to_bytes(pow(bytes_to_long(padded_aeskey), e, n))
|
||||||
|
init_data = self._download_json('http://www.daisuki.net/bin/bgn/init', video_id, query={
|
||||||
|
's': flashvars.get('s', ''),
|
||||||
|
'c': flashvars.get('ss3_prm', ''),
|
||||||
|
'e': url,
|
||||||
|
'd': base64.b64encode(intlist_to_bytes(aes_cbc_encrypt(
|
||||||
|
bytes_to_intlist(json.dumps(data)),
|
||||||
|
aes_key, iv))).decode('ascii'),
|
||||||
|
'a': base64.b64encode(encrypted_aeskey).decode('ascii'),
|
||||||
|
}, note='Downloading JSON metadata' + (' (try #%d)' % (idx + 1) if idx > 0 else ''))
|
||||||
|
|
||||||
|
if 'rtn' in init_data:
|
||||||
|
encrypted_rtn = init_data['rtn']
|
||||||
|
break
|
||||||
|
|
||||||
|
self._sleep(5, video_id)
|
||||||
|
|
||||||
|
if encrypted_rtn is None:
|
||||||
|
raise ExtractorError('Failed to fetch init data')
|
||||||
|
|
||||||
|
rtn = self._parse_json(
|
||||||
|
intlist_to_bytes(aes_cbc_decrypt(bytes_to_intlist(
|
||||||
|
base64.b64decode(encrypted_rtn)),
|
||||||
|
aes_key, iv)).decode('utf-8').rstrip('\0'),
|
||||||
|
video_id)
|
||||||
|
|
||||||
|
formats = self._extract_m3u8_formats(
|
||||||
|
rtn['play_url'], video_id, ext='mp4', entry_protocol='m3u8_native')
|
||||||
|
|
||||||
|
title = remove_end(self._og_search_title(webpage), ' - DAISUKI')
|
||||||
|
|
||||||
|
creator = self._html_search_regex(
|
||||||
|
r'Creator\s*:\s*([^<]+)', webpage, 'creator', fatal=False)
|
||||||
|
|
||||||
|
subtitles = {}
|
||||||
|
caption_url = rtn.get('caption_url')
|
||||||
|
if caption_url:
|
||||||
|
# mul: multiple languages
|
||||||
|
subtitles['mul'] = [{
|
||||||
|
'url': caption_url,
|
||||||
|
'ext': 'ttml',
|
||||||
|
}]
|
||||||
|
|
||||||
|
return {
|
||||||
|
'id': video_id,
|
||||||
|
'title': title,
|
||||||
|
'formats': formats,
|
||||||
|
'subtitles': subtitles,
|
||||||
|
'creator': creator,
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class DaisukiPlaylistIE(InfoExtractor):
|
||||||
|
_VALID_URL = r'https?://(?:www\.)daisuki\.net/[^/]+/[^/]+/[^/]+/detail\.(?P<id>[a-zA-Z0-9]+)\.html'
|
||||||
|
|
||||||
|
_TEST = {
|
||||||
|
'url': 'http://www.daisuki.net/tw/en/anime/detail.TheIdolMasterCG.html',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'TheIdolMasterCG',
|
||||||
|
'title': 'THE IDOLM@STER CINDERELLA GIRLS',
|
||||||
|
'description': 'md5:0f2c028a9339f7a2c7fbf839edc5c5d8',
|
||||||
|
},
|
||||||
|
'playlist_count': 26,
|
||||||
|
}
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
playlist_id = self._match_id(url)
|
||||||
|
|
||||||
|
webpage = self._download_webpage(url, playlist_id)
|
||||||
|
|
||||||
|
episode_pattern = r'''(?sx)
|
||||||
|
<img[^>]+delay="[^"]+/(\d+)/movie\.jpg".+?
|
||||||
|
<p[^>]+class=".*?\bepisodeNumber\b.*?">(?:<a[^>]+>)?([^<]+)'''
|
||||||
|
entries = [{
|
||||||
|
'_type': 'url_transparent',
|
||||||
|
'url': url.replace('detail', 'watch').replace('.html', '.' + movie_id + '.html'),
|
||||||
|
'episode_id': episode_id,
|
||||||
|
'episode_number': int_or_none(episode_id),
|
||||||
|
} for movie_id, episode_id in re.findall(episode_pattern, webpage)]
|
||||||
|
|
||||||
|
playlist_title = remove_end(
|
||||||
|
self._og_search_title(webpage, fatal=False), ' - Anime - DAISUKI')
|
||||||
|
playlist_description = clean_html(get_element_by_id('synopsisTxt', webpage))
|
||||||
|
|
||||||
|
return self.playlist_result(entries, playlist_id, playlist_title, playlist_description)
|
@ -1,15 +1,7 @@
|
|||||||
# coding: utf-8
|
# coding: utf-8
|
||||||
from __future__ import unicode_literals
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
import hashlib
|
|
||||||
import time
|
|
||||||
import uuid
|
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from ..compat import (
|
|
||||||
compat_str,
|
|
||||||
compat_urllib_parse_urlencode,
|
|
||||||
)
|
|
||||||
from ..utils import (
|
from ..utils import (
|
||||||
ExtractorError,
|
ExtractorError,
|
||||||
unescapeHTML,
|
unescapeHTML,
|
||||||
@ -24,8 +16,8 @@ class DouyuTVIE(InfoExtractor):
|
|||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': '17732',
|
'id': '17732',
|
||||||
'display_id': 'iseven',
|
'display_id': 'iseven',
|
||||||
'ext': 'flv',
|
'ext': 'mp4',
|
||||||
'title': 're:^清晨醒脑!T-ara根本停不下来! [0-9]{4}-[0-9]{2}-[0-9]{2} [0-9]{2}:[0-9]{2}$',
|
'title': 're:^清晨醒脑!T-ARA根本停不下来! [0-9]{4}-[0-9]{2}-[0-9]{2} [0-9]{2}:[0-9]{2}$',
|
||||||
'description': r're:.*m7show@163\.com.*',
|
'description': r're:.*m7show@163\.com.*',
|
||||||
'thumbnail': r're:^https?://.*\.jpg$',
|
'thumbnail': r're:^https?://.*\.jpg$',
|
||||||
'uploader': '7师傅',
|
'uploader': '7师傅',
|
||||||
@ -39,7 +31,7 @@ class DouyuTVIE(InfoExtractor):
|
|||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': '85982',
|
'id': '85982',
|
||||||
'display_id': '85982',
|
'display_id': '85982',
|
||||||
'ext': 'flv',
|
'ext': 'mp4',
|
||||||
'title': 're:^小漠从零单排记!——CSOL2躲猫猫 [0-9]{4}-[0-9]{2}-[0-9]{2} [0-9]{2}:[0-9]{2}$',
|
'title': 're:^小漠从零单排记!——CSOL2躲猫猫 [0-9]{4}-[0-9]{2}-[0-9]{2} [0-9]{2}:[0-9]{2}$',
|
||||||
'description': 'md5:746a2f7a253966a06755a912f0acc0d2',
|
'description': 'md5:746a2f7a253966a06755a912f0acc0d2',
|
||||||
'thumbnail': r're:^https?://.*\.jpg$',
|
'thumbnail': r're:^https?://.*\.jpg$',
|
||||||
@ -55,8 +47,8 @@ class DouyuTVIE(InfoExtractor):
|
|||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': '17732',
|
'id': '17732',
|
||||||
'display_id': '17732',
|
'display_id': '17732',
|
||||||
'ext': 'flv',
|
'ext': 'mp4',
|
||||||
'title': 're:^清晨醒脑!T-ara根本停不下来! [0-9]{4}-[0-9]{2}-[0-9]{2} [0-9]{2}:[0-9]{2}$',
|
'title': 're:^清晨醒脑!T-ARA根本停不下来! [0-9]{4}-[0-9]{2}-[0-9]{2} [0-9]{2}:[0-9]{2}$',
|
||||||
'description': r're:.*m7show@163\.com.*',
|
'description': r're:.*m7show@163\.com.*',
|
||||||
'thumbnail': r're:^https?://.*\.jpg$',
|
'thumbnail': r're:^https?://.*\.jpg$',
|
||||||
'uploader': '7师傅',
|
'uploader': '7师傅',
|
||||||
@ -96,45 +88,18 @@ class DouyuTVIE(InfoExtractor):
|
|||||||
if room.get('show_status') == '2':
|
if room.get('show_status') == '2':
|
||||||
raise ExtractorError('Live stream is offline', expected=True)
|
raise ExtractorError('Live stream is offline', expected=True)
|
||||||
|
|
||||||
tt = compat_str(int(time.time() / 60))
|
formats = self._extract_m3u8_formats(
|
||||||
did = uuid.uuid4().hex.upper()
|
room['hls_url'], video_id, ext='mp4')
|
||||||
|
|
||||||
sign_content = ''.join((room_id, did, self._API_KEY, tt))
|
|
||||||
sign = hashlib.md5((sign_content).encode('utf-8')).hexdigest()
|
|
||||||
|
|
||||||
flv_data = compat_urllib_parse_urlencode({
|
|
||||||
'cdn': 'ws',
|
|
||||||
'rate': '0',
|
|
||||||
'tt': tt,
|
|
||||||
'did': did,
|
|
||||||
'sign': sign,
|
|
||||||
})
|
|
||||||
|
|
||||||
video_info = self._download_json(
|
|
||||||
'http://www.douyu.com/lapi/live/getPlay/%s' % room_id, video_id,
|
|
||||||
data=flv_data, note='Downloading video info',
|
|
||||||
headers={'Content-Type': 'application/x-www-form-urlencoded'})
|
|
||||||
|
|
||||||
error_code = video_info.get('error', 0)
|
|
||||||
if error_code is not 0:
|
|
||||||
raise ExtractorError(
|
|
||||||
'%s reported error %i' % (self.IE_NAME, error_code),
|
|
||||||
expected=True)
|
|
||||||
|
|
||||||
base_url = video_info['data']['rtmp_url']
|
|
||||||
live_path = video_info['data']['rtmp_live']
|
|
||||||
|
|
||||||
video_url = '%s/%s' % (base_url, live_path)
|
|
||||||
|
|
||||||
title = self._live_title(unescapeHTML(room['room_name']))
|
title = self._live_title(unescapeHTML(room['room_name']))
|
||||||
description = room.get('notice')
|
description = room.get('show_details')
|
||||||
thumbnail = room.get('room_src')
|
thumbnail = room.get('room_src')
|
||||||
uploader = room.get('nickname')
|
uploader = room.get('nickname')
|
||||||
|
|
||||||
return {
|
return {
|
||||||
'id': room_id,
|
'id': room_id,
|
||||||
'display_id': video_id,
|
'display_id': video_id,
|
||||||
'url': video_url,
|
'formats': formats,
|
||||||
'title': title,
|
'title': title,
|
||||||
'description': description,
|
'description': description,
|
||||||
'thumbnail': thumbnail,
|
'thumbnail': thumbnail,
|
||||||
|
39
youtube_dl/extractor/etonline.py
Normal file
39
youtube_dl/extractor/etonline.py
Normal file
@ -0,0 +1,39 @@
|
|||||||
|
# coding: utf-8
|
||||||
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
|
import re
|
||||||
|
|
||||||
|
from .common import InfoExtractor
|
||||||
|
|
||||||
|
|
||||||
|
class ETOnlineIE(InfoExtractor):
|
||||||
|
_VALID_URL = r'https?://(?:www\.)?etonline\.com/(?:[^/]+/)*(?P<id>[^/?#&]+)'
|
||||||
|
_TESTS = [{
|
||||||
|
'url': 'http://www.etonline.com/tv/211130_dove_cameron_liv_and_maddie_emotional_episode_series_finale/',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '211130_dove_cameron_liv_and_maddie_emotional_episode_series_finale',
|
||||||
|
'title': 'md5:a21ec7d3872ed98335cbd2a046f34ee6',
|
||||||
|
'description': 'md5:8b94484063f463cca709617c79618ccd',
|
||||||
|
},
|
||||||
|
'playlist_count': 2,
|
||||||
|
}, {
|
||||||
|
'url': 'http://www.etonline.com/media/video/here_are_the_stars_who_love_bringing_their_moms_as_dates_to_the_oscars-211359/',
|
||||||
|
'only_matching': True,
|
||||||
|
}]
|
||||||
|
BRIGHTCOVE_URL_TEMPLATE = 'http://players.brightcove.net/1242911076001/default_default/index.html?videoId=ref:%s'
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
playlist_id = self._match_id(url)
|
||||||
|
|
||||||
|
webpage = self._download_webpage(url, playlist_id)
|
||||||
|
|
||||||
|
entries = [
|
||||||
|
self.url_result(
|
||||||
|
self.BRIGHTCOVE_URL_TEMPLATE % video_id, 'BrightcoveNew', video_id)
|
||||||
|
for video_id in re.findall(
|
||||||
|
r'site\.brightcove\s*\([^,]+,\s*["\'](title_\d+)', webpage)]
|
||||||
|
|
||||||
|
return self.playlist_result(
|
||||||
|
entries, playlist_id,
|
||||||
|
self._og_search_title(webpage, fatal=False),
|
||||||
|
self._og_search_description(webpage))
|
@ -83,6 +83,7 @@ from .awaan import (
|
|||||||
from .azmedien import (
|
from .azmedien import (
|
||||||
AZMedienIE,
|
AZMedienIE,
|
||||||
AZMedienPlaylistIE,
|
AZMedienPlaylistIE,
|
||||||
|
AZMedienShowPlaylistIE,
|
||||||
)
|
)
|
||||||
from .azubu import AzubuIE, AzubuLiveIE
|
from .azubu import AzubuIE, AzubuLiveIE
|
||||||
from .baidu import BaiduVideoIE
|
from .baidu import BaiduVideoIE
|
||||||
@ -227,6 +228,10 @@ from .dailymotion import (
|
|||||||
DailymotionUserIE,
|
DailymotionUserIE,
|
||||||
DailymotionCloudIE,
|
DailymotionCloudIE,
|
||||||
)
|
)
|
||||||
|
from .daisuki import (
|
||||||
|
DaisukiIE,
|
||||||
|
DaisukiPlaylistIE,
|
||||||
|
)
|
||||||
from .daum import (
|
from .daum import (
|
||||||
DaumIE,
|
DaumIE,
|
||||||
DaumClipIE,
|
DaumClipIE,
|
||||||
@ -288,6 +293,7 @@ from .espn import (
|
|||||||
ESPNArticleIE,
|
ESPNArticleIE,
|
||||||
)
|
)
|
||||||
from .esri import EsriVideoIE
|
from .esri import EsriVideoIE
|
||||||
|
from .etonline import ETOnlineIE
|
||||||
from .europa import EuropaIE
|
from .europa import EuropaIE
|
||||||
from .everyonesmixtape import EveryonesMixtapeIE
|
from .everyonesmixtape import EveryonesMixtapeIE
|
||||||
from .expotv import ExpoTVIE
|
from .expotv import ExpoTVIE
|
||||||
@ -338,6 +344,7 @@ from .francetv import (
|
|||||||
)
|
)
|
||||||
from .freesound import FreesoundIE
|
from .freesound import FreesoundIE
|
||||||
from .freespeech import FreespeechIE
|
from .freespeech import FreespeechIE
|
||||||
|
from .freshlive import FreshLiveIE
|
||||||
from .funimation import FunimationIE
|
from .funimation import FunimationIE
|
||||||
from .funnyordie import FunnyOrDieIE
|
from .funnyordie import FunnyOrDieIE
|
||||||
from .fusion import FusionIE
|
from .fusion import FusionIE
|
||||||
@ -637,6 +644,7 @@ from .ninecninemedia import (
|
|||||||
from .ninegag import NineGagIE
|
from .ninegag import NineGagIE
|
||||||
from .ninenow import NineNowIE
|
from .ninenow import NineNowIE
|
||||||
from .nintendo import NintendoIE
|
from .nintendo import NintendoIE
|
||||||
|
from .njpwworld import NJPWWorldIE
|
||||||
from .nobelprize import NobelPrizeIE
|
from .nobelprize import NobelPrizeIE
|
||||||
from .noco import NocoIE
|
from .noco import NocoIE
|
||||||
from .normalboots import NormalbootsIE
|
from .normalboots import NormalbootsIE
|
||||||
@ -666,6 +674,7 @@ from .npo import (
|
|||||||
NPORadioIE,
|
NPORadioIE,
|
||||||
NPORadioFragmentIE,
|
NPORadioFragmentIE,
|
||||||
SchoolTVIE,
|
SchoolTVIE,
|
||||||
|
HetKlokhuisIE,
|
||||||
VPROIE,
|
VPROIE,
|
||||||
WNLIE,
|
WNLIE,
|
||||||
)
|
)
|
||||||
@ -835,7 +844,6 @@ from .safari import (
|
|||||||
from .sapo import SapoIE
|
from .sapo import SapoIE
|
||||||
from .savefrom import SaveFromIE
|
from .savefrom import SaveFromIE
|
||||||
from .sbs import SBSIE
|
from .sbs import SBSIE
|
||||||
from .scivee import SciVeeIE
|
|
||||||
from .screencast import ScreencastIE
|
from .screencast import ScreencastIE
|
||||||
from .screencastomatic import ScreencastOMaticIE
|
from .screencastomatic import ScreencastOMaticIE
|
||||||
from .scrippsnetworks import ScrippsNetworksWatchIE
|
from .scrippsnetworks import ScrippsNetworksWatchIE
|
||||||
|
@ -303,7 +303,7 @@ class FacebookIE(InfoExtractor):
|
|||||||
if not video_data:
|
if not video_data:
|
||||||
server_js_data = self._parse_json(
|
server_js_data = self._parse_json(
|
||||||
self._search_regex(
|
self._search_regex(
|
||||||
r'bigPipe\.onPageletArrive\(({.+?})\)\s*;\s*}\s*\)\s*,\s*["\']onPageletArrive\s+(?:stream_pagelet|pagelet_group_mall)',
|
r'bigPipe\.onPageletArrive\(({.+?})\)\s*;\s*}\s*\)\s*,\s*["\']onPageletArrive\s+(?:stream_pagelet|pagelet_group_mall|permalink_video_pagelet)',
|
||||||
webpage, 'js data', default='{}'),
|
webpage, 'js data', default='{}'),
|
||||||
video_id, transform_source=js_to_json, fatal=False)
|
video_id, transform_source=js_to_json, fatal=False)
|
||||||
if server_js_data:
|
if server_js_data:
|
||||||
|
84
youtube_dl/extractor/freshlive.py
Normal file
84
youtube_dl/extractor/freshlive.py
Normal file
@ -0,0 +1,84 @@
|
|||||||
|
# coding: utf-8
|
||||||
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
|
from .common import InfoExtractor
|
||||||
|
from ..compat import compat_str
|
||||||
|
from ..utils import (
|
||||||
|
ExtractorError,
|
||||||
|
int_or_none,
|
||||||
|
try_get,
|
||||||
|
unified_timestamp,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class FreshLiveIE(InfoExtractor):
|
||||||
|
_VALID_URL = r'https?://freshlive\.tv/[^/]+/(?P<id>\d+)'
|
||||||
|
_TEST = {
|
||||||
|
'url': 'https://freshlive.tv/satotv/74712',
|
||||||
|
'md5': '9f0cf5516979c4454ce982df3d97f352',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '74712',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': 'テスト',
|
||||||
|
'description': 'テスト',
|
||||||
|
'thumbnail': r're:^https?://.*\.jpg$',
|
||||||
|
'duration': 1511,
|
||||||
|
'timestamp': 1483619655,
|
||||||
|
'upload_date': '20170105',
|
||||||
|
'uploader': 'サトTV',
|
||||||
|
'uploader_id': 'satotv',
|
||||||
|
'view_count': int,
|
||||||
|
'comment_count': int,
|
||||||
|
'is_live': False,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
video_id = self._match_id(url)
|
||||||
|
|
||||||
|
webpage = self._download_webpage(url, video_id)
|
||||||
|
|
||||||
|
options = self._parse_json(
|
||||||
|
self._search_regex(
|
||||||
|
r'window\.__CONTEXT__\s*=\s*({.+?});\s*</script>',
|
||||||
|
webpage, 'initial context'),
|
||||||
|
video_id)
|
||||||
|
|
||||||
|
info = options['context']['dispatcher']['stores']['ProgramStore']['programs'][video_id]
|
||||||
|
|
||||||
|
title = info['title']
|
||||||
|
|
||||||
|
if info.get('status') == 'upcoming':
|
||||||
|
raise ExtractorError('Stream %s is upcoming' % video_id, expected=True)
|
||||||
|
|
||||||
|
stream_url = info.get('liveStreamUrl') or info['archiveStreamUrl']
|
||||||
|
|
||||||
|
is_live = info.get('liveStreamUrl') is not None
|
||||||
|
|
||||||
|
formats = self._extract_m3u8_formats(
|
||||||
|
stream_url, video_id, ext='mp4',
|
||||||
|
entry_protocol='m3u8' if is_live else 'm3u8_native',
|
||||||
|
m3u8_id='hls')
|
||||||
|
|
||||||
|
if is_live:
|
||||||
|
title = self._live_title(title)
|
||||||
|
|
||||||
|
return {
|
||||||
|
'id': video_id,
|
||||||
|
'formats': formats,
|
||||||
|
'title': title,
|
||||||
|
'description': info.get('description'),
|
||||||
|
'thumbnail': info.get('thumbnailUrl'),
|
||||||
|
'duration': int_or_none(info.get('airTime')),
|
||||||
|
'timestamp': unified_timestamp(info.get('createdAt')),
|
||||||
|
'uploader': try_get(
|
||||||
|
info, lambda x: x['channel']['title'], compat_str),
|
||||||
|
'uploader_id': try_get(
|
||||||
|
info, lambda x: x['channel']['code'], compat_str),
|
||||||
|
'uploader_url': try_get(
|
||||||
|
info, lambda x: x['channel']['permalink'], compat_str),
|
||||||
|
'view_count': int_or_none(info.get('viewCount')),
|
||||||
|
'comment_count': int_or_none(info.get('commentCount')),
|
||||||
|
'tags': info.get('tags', []),
|
||||||
|
'is_live': is_live,
|
||||||
|
}
|
@ -16,6 +16,8 @@ class IviIE(InfoExtractor):
|
|||||||
IE_DESC = 'ivi.ru'
|
IE_DESC = 'ivi.ru'
|
||||||
IE_NAME = 'ivi'
|
IE_NAME = 'ivi'
|
||||||
_VALID_URL = r'https?://(?:www\.)?ivi\.ru/(?:watch/(?:[^/]+/)?|video/player\?.*?videoId=)(?P<id>\d+)'
|
_VALID_URL = r'https?://(?:www\.)?ivi\.ru/(?:watch/(?:[^/]+/)?|video/player\?.*?videoId=)(?P<id>\d+)'
|
||||||
|
_GEO_BYPASS = False
|
||||||
|
_GEO_COUNTRIES = ['RU']
|
||||||
|
|
||||||
_TESTS = [
|
_TESTS = [
|
||||||
# Single movie
|
# Single movie
|
||||||
@ -91,7 +93,11 @@ class IviIE(InfoExtractor):
|
|||||||
|
|
||||||
if 'error' in video_json:
|
if 'error' in video_json:
|
||||||
error = video_json['error']
|
error = video_json['error']
|
||||||
if error['origin'] == 'NoRedisValidData':
|
origin = error['origin']
|
||||||
|
if origin == 'NotAllowedForLocation':
|
||||||
|
self.raise_geo_restricted(
|
||||||
|
msg=error['message'], countries=self._GEO_COUNTRIES)
|
||||||
|
elif origin == 'NoRedisValidData':
|
||||||
raise ExtractorError('Video %s does not exist' % video_id, expected=True)
|
raise ExtractorError('Video %s does not exist' % video_id, expected=True)
|
||||||
raise ExtractorError(
|
raise ExtractorError(
|
||||||
'Unable to download video %s: %s' % (video_id, error['message']),
|
'Unable to download video %s: %s' % (video_id, error['message']),
|
||||||
|
@ -14,7 +14,7 @@ from ..utils import (
|
|||||||
|
|
||||||
class MDRIE(InfoExtractor):
|
class MDRIE(InfoExtractor):
|
||||||
IE_DESC = 'MDR.DE and KiKA'
|
IE_DESC = 'MDR.DE and KiKA'
|
||||||
_VALID_URL = r'https?://(?:www\.)?(?:mdr|kika)\.de/(?:.*)/[a-z]+-?(?P<id>\d+)(?:_.+?)?\.html'
|
_VALID_URL = r'https?://(?:www\.)?(?:mdr|kika)\.de/(?:.*)/[a-z-]+-?(?P<id>\d+)(?:_.+?)?\.html'
|
||||||
|
|
||||||
_TESTS = [{
|
_TESTS = [{
|
||||||
# MDR regularly deletes its videos
|
# MDR regularly deletes its videos
|
||||||
@ -31,6 +31,7 @@ class MDRIE(InfoExtractor):
|
|||||||
'duration': 250,
|
'duration': 250,
|
||||||
'uploader': 'MITTELDEUTSCHER RUNDFUNK',
|
'uploader': 'MITTELDEUTSCHER RUNDFUNK',
|
||||||
},
|
},
|
||||||
|
'skip': '404 not found',
|
||||||
}, {
|
}, {
|
||||||
'url': 'http://www.kika.de/baumhaus/videos/video19636.html',
|
'url': 'http://www.kika.de/baumhaus/videos/video19636.html',
|
||||||
'md5': '4930515e36b06c111213e80d1e4aad0e',
|
'md5': '4930515e36b06c111213e80d1e4aad0e',
|
||||||
@ -41,6 +42,7 @@ class MDRIE(InfoExtractor):
|
|||||||
'duration': 134,
|
'duration': 134,
|
||||||
'uploader': 'KIKA',
|
'uploader': 'KIKA',
|
||||||
},
|
},
|
||||||
|
'skip': '404 not found',
|
||||||
}, {
|
}, {
|
||||||
'url': 'http://www.kika.de/sendungen/einzelsendungen/weihnachtsprogramm/videos/video8182.html',
|
'url': 'http://www.kika.de/sendungen/einzelsendungen/weihnachtsprogramm/videos/video8182.html',
|
||||||
'md5': '5fe9c4dd7d71e3b238f04b8fdd588357',
|
'md5': '5fe9c4dd7d71e3b238f04b8fdd588357',
|
||||||
@ -49,11 +51,21 @@ class MDRIE(InfoExtractor):
|
|||||||
'ext': 'mp4',
|
'ext': 'mp4',
|
||||||
'title': 'Beutolomäus und der geheime Weihnachtswunsch',
|
'title': 'Beutolomäus und der geheime Weihnachtswunsch',
|
||||||
'description': 'md5:b69d32d7b2c55cbe86945ab309d39bbd',
|
'description': 'md5:b69d32d7b2c55cbe86945ab309d39bbd',
|
||||||
'timestamp': 1450950000,
|
'timestamp': 1482541200,
|
||||||
'upload_date': '20151224',
|
'upload_date': '20161224',
|
||||||
'duration': 4628,
|
'duration': 4628,
|
||||||
'uploader': 'KIKA',
|
'uploader': 'KIKA',
|
||||||
},
|
},
|
||||||
|
}, {
|
||||||
|
# audio with alternative playerURL pattern
|
||||||
|
'url': 'http://www.mdr.de/kultur/videos-und-audios/audio-radio/operation-mindfuck-robert-wilson100.html',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '100',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': 'Feature: Operation Mindfuck - Robert Anton Wilson',
|
||||||
|
'duration': 3239,
|
||||||
|
'uploader': 'MITTELDEUTSCHER RUNDFUNK',
|
||||||
|
},
|
||||||
}, {
|
}, {
|
||||||
'url': 'http://www.kika.de/baumhaus/sendungen/video19636_zc-fea7f8a0_zs-4bf89c60.html',
|
'url': 'http://www.kika.de/baumhaus/sendungen/video19636_zc-fea7f8a0_zs-4bf89c60.html',
|
||||||
'only_matching': True,
|
'only_matching': True,
|
||||||
@ -71,7 +83,7 @@ class MDRIE(InfoExtractor):
|
|||||||
webpage = self._download_webpage(url, video_id)
|
webpage = self._download_webpage(url, video_id)
|
||||||
|
|
||||||
data_url = self._search_regex(
|
data_url = self._search_regex(
|
||||||
r'(?:dataURL|playerXml(?:["\'])?)\s*:\s*(["\'])(?P<url>.+/(?:video|audio)-?[0-9]+-avCustom\.xml)\1',
|
r'(?:dataURL|playerXml(?:["\'])?)\s*:\s*(["\'])(?P<url>.+?-avCustom\.xml)\1',
|
||||||
webpage, 'data url', group='url').replace(r'\/', '/')
|
webpage, 'data url', group='url').replace(r'\/', '/')
|
||||||
|
|
||||||
doc = self._download_xml(
|
doc = self._download_xml(
|
||||||
|
83
youtube_dl/extractor/njpwworld.py
Normal file
83
youtube_dl/extractor/njpwworld.py
Normal file
@ -0,0 +1,83 @@
|
|||||||
|
# coding: utf-8
|
||||||
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
|
import re
|
||||||
|
|
||||||
|
from .common import InfoExtractor
|
||||||
|
from ..compat import compat_urlparse
|
||||||
|
from ..utils import (
|
||||||
|
get_element_by_class,
|
||||||
|
urlencode_postdata,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class NJPWWorldIE(InfoExtractor):
|
||||||
|
_VALID_URL = r'https?://njpwworld\.com/p/(?P<id>[a-z0-9_]+)'
|
||||||
|
IE_DESC = '新日本プロレスワールド'
|
||||||
|
_NETRC_MACHINE = 'njpwworld'
|
||||||
|
|
||||||
|
_TEST = {
|
||||||
|
'url': 'http://njpwworld.com/p/s_series_00155_1_9/',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 's_series_00155_1_9',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': '第9試合 ランディ・サベージ vs リック・スタイナー',
|
||||||
|
'tags': list,
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
'skip_download': True, # AES-encrypted m3u8
|
||||||
|
},
|
||||||
|
'skip': 'Requires login',
|
||||||
|
}
|
||||||
|
|
||||||
|
def _real_initialize(self):
|
||||||
|
self._login()
|
||||||
|
|
||||||
|
def _login(self):
|
||||||
|
username, password = self._get_login_info()
|
||||||
|
# No authentication to be performed
|
||||||
|
if not username:
|
||||||
|
return True
|
||||||
|
|
||||||
|
webpage, urlh = self._download_webpage_handle(
|
||||||
|
'https://njpwworld.com/auth/login', None,
|
||||||
|
note='Logging in', errnote='Unable to login',
|
||||||
|
data=urlencode_postdata({'login_id': username, 'pw': password}))
|
||||||
|
# /auth/login will return 302 for successful logins
|
||||||
|
if urlh.geturl() == 'https://njpwworld.com/auth/login':
|
||||||
|
self.report_warning('unable to login')
|
||||||
|
return False
|
||||||
|
|
||||||
|
return True
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
video_id = self._match_id(url)
|
||||||
|
|
||||||
|
webpage = self._download_webpage(url, video_id)
|
||||||
|
|
||||||
|
formats = []
|
||||||
|
for player_url, kind in re.findall(r'<a[^>]+href="(/player[^"]+)".+?<img[^>]+src="[^"]+qf_btn_([^".]+)', webpage):
|
||||||
|
player_url = compat_urlparse.urljoin(url, player_url)
|
||||||
|
|
||||||
|
player_page = self._download_webpage(
|
||||||
|
player_url, video_id, note='Downloading player page')
|
||||||
|
|
||||||
|
entries = self._parse_html5_media_entries(
|
||||||
|
player_url, player_page, video_id, m3u8_id='hls-%s' % kind,
|
||||||
|
m3u8_entry_protocol='m3u8_native',
|
||||||
|
preference=2 if 'hq' in kind else 1)
|
||||||
|
formats.extend(entries[0]['formats'])
|
||||||
|
|
||||||
|
self._sort_formats(formats)
|
||||||
|
|
||||||
|
post_content = get_element_by_class('post-content', webpage)
|
||||||
|
tags = re.findall(
|
||||||
|
r'<li[^>]+class="tag-[^"]+"><a[^>]*>([^<]+)</a></li>', post_content
|
||||||
|
) if post_content else None
|
||||||
|
|
||||||
|
return {
|
||||||
|
'id': video_id,
|
||||||
|
'title': self._og_search_title(webpage),
|
||||||
|
'formats': formats,
|
||||||
|
'tags': tags,
|
||||||
|
}
|
@ -23,7 +23,7 @@ from ..utils import (
|
|||||||
|
|
||||||
class NocoIE(InfoExtractor):
|
class NocoIE(InfoExtractor):
|
||||||
_VALID_URL = r'https?://(?:(?:www\.)?noco\.tv/emission/|player\.noco\.tv/\?idvideo=)(?P<id>\d+)'
|
_VALID_URL = r'https?://(?:(?:www\.)?noco\.tv/emission/|player\.noco\.tv/\?idvideo=)(?P<id>\d+)'
|
||||||
_LOGIN_URL = 'http://noco.tv/do.php'
|
_LOGIN_URL = 'https://noco.tv/do.php'
|
||||||
_API_URL_TEMPLATE = 'https://api.noco.tv/1.1/%s?ts=%s&tk=%s'
|
_API_URL_TEMPLATE = 'https://api.noco.tv/1.1/%s?ts=%s&tk=%s'
|
||||||
_SUB_LANG_TEMPLATE = '&sub_lang=%s'
|
_SUB_LANG_TEMPLATE = '&sub_lang=%s'
|
||||||
_NETRC_MACHINE = 'noco'
|
_NETRC_MACHINE = 'noco'
|
||||||
@ -69,16 +69,17 @@ class NocoIE(InfoExtractor):
|
|||||||
if username is None:
|
if username is None:
|
||||||
return
|
return
|
||||||
|
|
||||||
login_form = {
|
login = self._download_json(
|
||||||
'a': 'login',
|
self._LOGIN_URL, None, 'Logging in as %s' % username,
|
||||||
'cookie': '1',
|
data=urlencode_postdata({
|
||||||
'username': username,
|
'a': 'login',
|
||||||
'password': password,
|
'cookie': '1',
|
||||||
}
|
'username': username,
|
||||||
request = sanitized_Request(self._LOGIN_URL, urlencode_postdata(login_form))
|
'password': password,
|
||||||
request.add_header('Content-Type', 'application/x-www-form-urlencoded; charset=UTF-8')
|
}),
|
||||||
|
headers={
|
||||||
login = self._download_json(request, None, 'Logging in as %s' % username)
|
'Content-Type': 'application/x-www-form-urlencoded; charset=UTF-8',
|
||||||
|
})
|
||||||
|
|
||||||
if 'erreur' in login:
|
if 'erreur' in login:
|
||||||
raise ExtractorError('Unable to login: %s' % clean_html(login['erreur']), expected=True)
|
raise ExtractorError('Unable to login: %s' % clean_html(login['erreur']), expected=True)
|
||||||
|
@ -3,41 +3,27 @@ from __future__ import unicode_literals
|
|||||||
import re
|
import re
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from ..compat import compat_HTTPError
|
from ..compat import (
|
||||||
|
compat_HTTPError,
|
||||||
|
compat_str,
|
||||||
|
)
|
||||||
from ..utils import (
|
from ..utils import (
|
||||||
|
determine_ext,
|
||||||
|
ExtractorError,
|
||||||
fix_xml_ampersands,
|
fix_xml_ampersands,
|
||||||
orderedSet,
|
orderedSet,
|
||||||
parse_duration,
|
parse_duration,
|
||||||
qualities,
|
qualities,
|
||||||
strip_jsonp,
|
strip_jsonp,
|
||||||
unified_strdate,
|
unified_strdate,
|
||||||
ExtractorError,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
class NPOBaseIE(InfoExtractor):
|
class NPOBaseIE(InfoExtractor):
|
||||||
def _get_token(self, video_id):
|
def _get_token(self, video_id):
|
||||||
token_page = self._download_webpage(
|
return self._download_json(
|
||||||
'http://ida.omroep.nl/npoplayer/i.js',
|
'http://ida.omroep.nl/app.php/auth', video_id,
|
||||||
video_id, note='Downloading token')
|
note='Downloading token')['token']
|
||||||
token = self._search_regex(
|
|
||||||
r'npoplayer\.token = "(.+?)"', token_page, 'token')
|
|
||||||
# Decryption algorithm extracted from http://npoplayer.omroep.nl/csjs/npoplayer-min.js
|
|
||||||
token_l = list(token)
|
|
||||||
first = second = None
|
|
||||||
for i in range(5, len(token_l) - 4):
|
|
||||||
if token_l[i].isdigit():
|
|
||||||
if first is None:
|
|
||||||
first = i
|
|
||||||
elif second is None:
|
|
||||||
second = i
|
|
||||||
if first is None or second is None:
|
|
||||||
first = 12
|
|
||||||
second = 13
|
|
||||||
|
|
||||||
token_l[first], token_l[second] = token_l[second], token_l[first]
|
|
||||||
|
|
||||||
return ''.join(token_l)
|
|
||||||
|
|
||||||
|
|
||||||
class NPOIE(NPOBaseIE):
|
class NPOIE(NPOBaseIE):
|
||||||
@ -51,97 +37,120 @@ class NPOIE(NPOBaseIE):
|
|||||||
(?:
|
(?:
|
||||||
npo\.nl/(?!live|radio)(?:[^/]+/){2}|
|
npo\.nl/(?!live|radio)(?:[^/]+/){2}|
|
||||||
ntr\.nl/(?:[^/]+/){2,}|
|
ntr\.nl/(?:[^/]+/){2,}|
|
||||||
omroepwnl\.nl/video/fragment/[^/]+__
|
omroepwnl\.nl/video/fragment/[^/]+__|
|
||||||
|
zapp\.nl/[^/]+/[^/]+/
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
(?P<id>[^/?#]+)
|
(?P<id>[^/?#]+)
|
||||||
'''
|
'''
|
||||||
|
|
||||||
_TESTS = [
|
_TESTS = [{
|
||||||
{
|
'url': 'http://www.npo.nl/nieuwsuur/22-06-2014/VPWON_1220719',
|
||||||
'url': 'http://www.npo.nl/nieuwsuur/22-06-2014/VPWON_1220719',
|
'md5': '4b3f9c429157ec4775f2c9cb7b911016',
|
||||||
'md5': '4b3f9c429157ec4775f2c9cb7b911016',
|
'info_dict': {
|
||||||
'info_dict': {
|
'id': 'VPWON_1220719',
|
||||||
'id': 'VPWON_1220719',
|
'ext': 'm4v',
|
||||||
'ext': 'm4v',
|
'title': 'Nieuwsuur',
|
||||||
'title': 'Nieuwsuur',
|
'description': 'Dagelijks tussen tien en elf: nieuws, sport en achtergronden.',
|
||||||
'description': 'Dagelijks tussen tien en elf: nieuws, sport en achtergronden.',
|
'upload_date': '20140622',
|
||||||
'upload_date': '20140622',
|
|
||||||
},
|
|
||||||
},
|
},
|
||||||
{
|
}, {
|
||||||
'url': 'http://www.npo.nl/de-mega-mike-mega-thomas-show/27-02-2009/VARA_101191800',
|
'url': 'http://www.npo.nl/de-mega-mike-mega-thomas-show/27-02-2009/VARA_101191800',
|
||||||
'md5': 'da50a5787dbfc1603c4ad80f31c5120b',
|
'md5': 'da50a5787dbfc1603c4ad80f31c5120b',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'VARA_101191800',
|
'id': 'VARA_101191800',
|
||||||
'ext': 'm4v',
|
'ext': 'm4v',
|
||||||
'title': 'De Mega Mike & Mega Thomas show: The best of.',
|
'title': 'De Mega Mike & Mega Thomas show: The best of.',
|
||||||
'description': 'md5:3b74c97fc9d6901d5a665aac0e5400f4',
|
'description': 'md5:3b74c97fc9d6901d5a665aac0e5400f4',
|
||||||
'upload_date': '20090227',
|
'upload_date': '20090227',
|
||||||
'duration': 2400,
|
'duration': 2400,
|
||||||
},
|
|
||||||
},
|
},
|
||||||
{
|
}, {
|
||||||
'url': 'http://www.npo.nl/tegenlicht/25-02-2013/VPWON_1169289',
|
'url': 'http://www.npo.nl/tegenlicht/25-02-2013/VPWON_1169289',
|
||||||
'md5': 'f8065e4e5a7824068ed3c7e783178f2c',
|
'md5': 'f8065e4e5a7824068ed3c7e783178f2c',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'VPWON_1169289',
|
'id': 'VPWON_1169289',
|
||||||
'ext': 'm4v',
|
'ext': 'm4v',
|
||||||
'title': 'Tegenlicht: De toekomst komt uit Afrika',
|
'title': 'Tegenlicht: Zwart geld. De toekomst komt uit Afrika',
|
||||||
'description': 'md5:52cf4eefbc96fffcbdc06d024147abea',
|
'description': 'md5:52cf4eefbc96fffcbdc06d024147abea',
|
||||||
'upload_date': '20130225',
|
'upload_date': '20130225',
|
||||||
'duration': 3000,
|
'duration': 3000,
|
||||||
},
|
|
||||||
},
|
},
|
||||||
{
|
}, {
|
||||||
'url': 'http://www.npo.nl/de-nieuwe-mens-deel-1/21-07-2010/WO_VPRO_043706',
|
'url': 'http://www.npo.nl/de-nieuwe-mens-deel-1/21-07-2010/WO_VPRO_043706',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'WO_VPRO_043706',
|
'id': 'WO_VPRO_043706',
|
||||||
'ext': 'wmv',
|
'ext': 'm4v',
|
||||||
'title': 'De nieuwe mens - Deel 1',
|
'title': 'De nieuwe mens - Deel 1',
|
||||||
'description': 'md5:518ae51ba1293ffb80d8d8ce90b74e4b',
|
'description': 'md5:518ae51ba1293ffb80d8d8ce90b74e4b',
|
||||||
'duration': 4680,
|
'duration': 4680,
|
||||||
},
|
|
||||||
'params': {
|
|
||||||
# mplayer mms download
|
|
||||||
'skip_download': True,
|
|
||||||
}
|
|
||||||
},
|
},
|
||||||
# non asf in streams
|
'params': {
|
||||||
{
|
'skip_download': True,
|
||||||
'url': 'http://www.npo.nl/hoe-gaat-europa-verder-na-parijs/10-01-2015/WO_NOS_762771',
|
|
||||||
'md5': 'b3da13de374cbe2d5332a7e910bef97f',
|
|
||||||
'info_dict': {
|
|
||||||
'id': 'WO_NOS_762771',
|
|
||||||
'ext': 'mp4',
|
|
||||||
'title': 'Hoe gaat Europa verder na Parijs?',
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
'url': 'http://www.ntr.nl/Aap-Poot-Pies/27/detail/Aap-poot-pies/VPWON_1233944#content',
|
|
||||||
'md5': '01c6a2841675995da1f0cf776f03a9c3',
|
|
||||||
'info_dict': {
|
|
||||||
'id': 'VPWON_1233944',
|
|
||||||
'ext': 'm4v',
|
|
||||||
'title': 'Aap, poot, pies',
|
|
||||||
'description': 'md5:c9c8005d1869ae65b858e82c01a91fde',
|
|
||||||
'upload_date': '20150508',
|
|
||||||
'duration': 599,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
'url': 'http://www.omroepwnl.nl/video/fragment/vandaag-de-dag-verkiezingen__POMS_WNL_853698',
|
|
||||||
'md5': 'd30cd8417b8b9bca1fdff27428860d08',
|
|
||||||
'info_dict': {
|
|
||||||
'id': 'POW_00996502',
|
|
||||||
'ext': 'm4v',
|
|
||||||
'title': '''"Dit is wel een 'landslide'..."''',
|
|
||||||
'description': 'md5:f8d66d537dfb641380226e31ca57b8e8',
|
|
||||||
'upload_date': '20150508',
|
|
||||||
'duration': 462,
|
|
||||||
},
|
|
||||||
}
|
}
|
||||||
]
|
}, {
|
||||||
|
# non asf in streams
|
||||||
|
'url': 'http://www.npo.nl/hoe-gaat-europa-verder-na-parijs/10-01-2015/WO_NOS_762771',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'WO_NOS_762771',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': 'Hoe gaat Europa verder na Parijs?',
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
'skip_download': True,
|
||||||
|
}
|
||||||
|
}, {
|
||||||
|
'url': 'http://www.ntr.nl/Aap-Poot-Pies/27/detail/Aap-poot-pies/VPWON_1233944#content',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'VPWON_1233944',
|
||||||
|
'ext': 'm4v',
|
||||||
|
'title': 'Aap, poot, pies',
|
||||||
|
'description': 'md5:c9c8005d1869ae65b858e82c01a91fde',
|
||||||
|
'upload_date': '20150508',
|
||||||
|
'duration': 599,
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
'skip_download': True,
|
||||||
|
}
|
||||||
|
}, {
|
||||||
|
'url': 'http://www.omroepwnl.nl/video/fragment/vandaag-de-dag-verkiezingen__POMS_WNL_853698',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'POW_00996502',
|
||||||
|
'ext': 'm4v',
|
||||||
|
'title': '''"Dit is wel een 'landslide'..."''',
|
||||||
|
'description': 'md5:f8d66d537dfb641380226e31ca57b8e8',
|
||||||
|
'upload_date': '20150508',
|
||||||
|
'duration': 462,
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
'skip_download': True,
|
||||||
|
}
|
||||||
|
}, {
|
||||||
|
# audio
|
||||||
|
'url': 'http://www.npo.nl/jouw-stad-rotterdam/29-01-2017/RBX_FUNX_6683215/RBX_FUNX_7601437',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'RBX_FUNX_6683215',
|
||||||
|
'ext': 'mp3',
|
||||||
|
'title': 'Jouw Stad Rotterdam',
|
||||||
|
'description': 'md5:db251505244f097717ec59fabc372d9f',
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
'skip_download': True,
|
||||||
|
}
|
||||||
|
}, {
|
||||||
|
'url': 'http://www.zapp.nl/de-bzt-show/gemist/KN_1687547',
|
||||||
|
'only_matching': True,
|
||||||
|
}, {
|
||||||
|
'url': 'http://www.zapp.nl/de-bzt-show/filmpjes/POMS_KN_7315118',
|
||||||
|
'only_matching': True,
|
||||||
|
}, {
|
||||||
|
'url': 'http://www.zapp.nl/beste-vrienden-quiz/extra-video-s/WO_NTR_1067990',
|
||||||
|
'only_matching': True,
|
||||||
|
}, {
|
||||||
|
# live stream
|
||||||
|
'url': 'npo:LI_NL1_4188102',
|
||||||
|
'only_matching': True,
|
||||||
|
}]
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
video_id = self._match_id(url)
|
video_id = self._match_id(url)
|
||||||
@ -170,70 +179,115 @@ class NPOIE(NPOBaseIE):
|
|||||||
token = self._get_token(video_id)
|
token = self._get_token(video_id)
|
||||||
|
|
||||||
formats = []
|
formats = []
|
||||||
|
urls = set()
|
||||||
|
|
||||||
pubopties = metadata.get('pubopties')
|
quality = qualities(['adaptive', 'wmv_sb', 'h264_sb', 'wmv_bb', 'h264_bb', 'wvc1_std', 'h264_std'])
|
||||||
if pubopties:
|
items = self._download_json(
|
||||||
quality = qualities(['adaptive', 'wmv_sb', 'h264_sb', 'wmv_bb', 'h264_bb', 'wvc1_std', 'h264_std'])
|
'http://ida.omroep.nl/app.php/%s' % video_id, video_id,
|
||||||
for format_id in pubopties:
|
'Downloading formats JSON', query={
|
||||||
format_info = self._download_json(
|
'adaptive': 'yes',
|
||||||
'http://ida.omroep.nl/odi/?prid=%s&puboptions=%s&adaptive=yes&token=%s'
|
'token': token,
|
||||||
% (video_id, format_id, token),
|
})['items'][0]
|
||||||
video_id, 'Downloading %s JSON' % format_id)
|
for num, item in enumerate(items):
|
||||||
if format_info.get('error_code', 0) or format_info.get('errorcode', 0):
|
item_url = item.get('url')
|
||||||
|
if not item_url or item_url in urls:
|
||||||
|
continue
|
||||||
|
urls.add(item_url)
|
||||||
|
format_id = self._search_regex(
|
||||||
|
r'video/ida/([^/]+)', item_url, 'format id',
|
||||||
|
default=None)
|
||||||
|
|
||||||
|
def add_format_url(format_url):
|
||||||
|
formats.append({
|
||||||
|
'url': format_url,
|
||||||
|
'format_id': format_id,
|
||||||
|
'quality': quality(format_id),
|
||||||
|
})
|
||||||
|
|
||||||
|
# Example: http://www.npo.nl/de-nieuwe-mens-deel-1/21-07-2010/WO_VPRO_043706
|
||||||
|
if item.get('contentType') in ('url', 'audio'):
|
||||||
|
add_format_url(item_url)
|
||||||
|
continue
|
||||||
|
|
||||||
|
try:
|
||||||
|
stream_info = self._download_json(
|
||||||
|
item_url + '&type=json', video_id,
|
||||||
|
'Downloading %s stream JSON'
|
||||||
|
% item.get('label') or item.get('format') or format_id or num)
|
||||||
|
except ExtractorError as ee:
|
||||||
|
if isinstance(ee.cause, compat_HTTPError) and ee.cause.code == 404:
|
||||||
|
error = (self._parse_json(
|
||||||
|
ee.cause.read().decode(), video_id,
|
||||||
|
fatal=False) or {}).get('errorstring')
|
||||||
|
if error:
|
||||||
|
raise ExtractorError(error, expected=True)
|
||||||
|
raise
|
||||||
|
# Stream URL instead of JSON, example: npo:LI_NL1_4188102
|
||||||
|
if isinstance(stream_info, compat_str):
|
||||||
|
if not stream_info.startswith('http'):
|
||||||
continue
|
continue
|
||||||
streams = format_info.get('streams')
|
video_url = stream_info
|
||||||
if streams:
|
# JSON
|
||||||
try:
|
else:
|
||||||
video_info = self._download_json(
|
video_url = stream_info.get('url')
|
||||||
streams[0] + '&type=json',
|
if not video_url or video_url in urls:
|
||||||
video_id, 'Downloading %s stream JSON' % format_id)
|
continue
|
||||||
except ExtractorError as ee:
|
urls.add(item_url)
|
||||||
if isinstance(ee.cause, compat_HTTPError) and ee.cause.code == 404:
|
if determine_ext(video_url) == 'm3u8':
|
||||||
error = (self._parse_json(ee.cause.read().decode(), video_id, fatal=False) or {}).get('errorstring')
|
formats.extend(self._extract_m3u8_formats(
|
||||||
if error:
|
video_url, video_id, ext='mp4',
|
||||||
raise ExtractorError(error, expected=True)
|
entry_protocol='m3u8_native', m3u8_id='hls', fatal=False))
|
||||||
raise
|
else:
|
||||||
else:
|
add_format_url(video_url)
|
||||||
video_info = format_info
|
|
||||||
video_url = video_info.get('url')
|
is_live = metadata.get('medium') == 'live'
|
||||||
if not video_url:
|
|
||||||
|
if not is_live:
|
||||||
|
for num, stream in enumerate(metadata.get('streams', [])):
|
||||||
|
stream_url = stream.get('url')
|
||||||
|
if not stream_url or stream_url in urls:
|
||||||
continue
|
continue
|
||||||
if format_id == 'adaptive':
|
urls.add(stream_url)
|
||||||
formats.extend(self._extract_m3u8_formats(video_url, video_id, 'mp4'))
|
# smooth streaming is not supported
|
||||||
else:
|
stream_type = stream.get('type', '').lower()
|
||||||
|
if stream_type in ['ss', 'ms']:
|
||||||
|
continue
|
||||||
|
if stream_type == 'hds':
|
||||||
|
f4m_formats = self._extract_f4m_formats(
|
||||||
|
stream_url, video_id, fatal=False)
|
||||||
|
# f4m downloader downloads only piece of live stream
|
||||||
|
for f4m_format in f4m_formats:
|
||||||
|
f4m_format['preference'] = -1
|
||||||
|
formats.extend(f4m_formats)
|
||||||
|
elif stream_type == 'hls':
|
||||||
|
formats.extend(self._extract_m3u8_formats(
|
||||||
|
stream_url, video_id, ext='mp4', fatal=False))
|
||||||
|
# Example: http://www.npo.nl/de-nieuwe-mens-deel-1/21-07-2010/WO_VPRO_043706
|
||||||
|
elif '.asf' in stream_url:
|
||||||
|
asx = self._download_xml(
|
||||||
|
stream_url, video_id,
|
||||||
|
'Downloading stream %d ASX playlist' % num,
|
||||||
|
transform_source=fix_xml_ampersands, fatal=False)
|
||||||
|
if not asx:
|
||||||
|
continue
|
||||||
|
ref = asx.find('./ENTRY/Ref')
|
||||||
|
if ref is None:
|
||||||
|
continue
|
||||||
|
video_url = ref.get('href')
|
||||||
|
if not video_url or video_url in urls:
|
||||||
|
continue
|
||||||
|
urls.add(video_url)
|
||||||
formats.append({
|
formats.append({
|
||||||
'url': video_url,
|
'url': video_url,
|
||||||
'format_id': format_id,
|
'ext': stream.get('formaat', 'asf'),
|
||||||
'quality': quality(format_id),
|
'quality': stream.get('kwaliteit'),
|
||||||
|
'preference': -10,
|
||||||
})
|
})
|
||||||
|
else:
|
||||||
streams = metadata.get('streams')
|
|
||||||
if streams:
|
|
||||||
for i, stream in enumerate(streams):
|
|
||||||
stream_url = stream.get('url')
|
|
||||||
if not stream_url:
|
|
||||||
continue
|
|
||||||
if '.asf' not in stream_url:
|
|
||||||
formats.append({
|
formats.append({
|
||||||
'url': stream_url,
|
'url': stream_url,
|
||||||
'quality': stream.get('kwaliteit'),
|
'quality': stream.get('kwaliteit'),
|
||||||
})
|
})
|
||||||
continue
|
|
||||||
asx = self._download_xml(
|
|
||||||
stream_url, video_id,
|
|
||||||
'Downloading stream %d ASX playlist' % i,
|
|
||||||
transform_source=fix_xml_ampersands)
|
|
||||||
ref = asx.find('./ENTRY/Ref')
|
|
||||||
if ref is None:
|
|
||||||
continue
|
|
||||||
video_url = ref.get('href')
|
|
||||||
if not video_url:
|
|
||||||
continue
|
|
||||||
formats.append({
|
|
||||||
'url': video_url,
|
|
||||||
'ext': stream.get('formaat', 'asf'),
|
|
||||||
'quality': stream.get('kwaliteit'),
|
|
||||||
})
|
|
||||||
|
|
||||||
self._sort_formats(formats)
|
self._sort_formats(formats)
|
||||||
|
|
||||||
@ -246,28 +300,28 @@ class NPOIE(NPOBaseIE):
|
|||||||
|
|
||||||
return {
|
return {
|
||||||
'id': video_id,
|
'id': video_id,
|
||||||
'title': title,
|
'title': self._live_title(title) if is_live else title,
|
||||||
'description': metadata.get('info'),
|
'description': metadata.get('info'),
|
||||||
'thumbnail': metadata.get('images', [{'url': None}])[-1]['url'],
|
'thumbnail': metadata.get('images', [{'url': None}])[-1]['url'],
|
||||||
'upload_date': unified_strdate(metadata.get('gidsdatum')),
|
'upload_date': unified_strdate(metadata.get('gidsdatum')),
|
||||||
'duration': parse_duration(metadata.get('tijdsduur')),
|
'duration': parse_duration(metadata.get('tijdsduur')),
|
||||||
'formats': formats,
|
'formats': formats,
|
||||||
'subtitles': subtitles,
|
'subtitles': subtitles,
|
||||||
|
'is_live': is_live,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
class NPOLiveIE(NPOBaseIE):
|
class NPOLiveIE(NPOBaseIE):
|
||||||
IE_NAME = 'npo.nl:live'
|
IE_NAME = 'npo.nl:live'
|
||||||
_VALID_URL = r'https?://(?:www\.)?npo\.nl/live/(?P<id>.+)'
|
_VALID_URL = r'https?://(?:www\.)?npo\.nl/live/(?P<id>[^/?#&]+)'
|
||||||
|
|
||||||
_TEST = {
|
_TEST = {
|
||||||
'url': 'http://www.npo.nl/live/npo-1',
|
'url': 'http://www.npo.nl/live/npo-1',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'LI_NEDERLAND1_136692',
|
'id': 'LI_NL1_4188102',
|
||||||
'display_id': 'npo-1',
|
'display_id': 'npo-1',
|
||||||
'ext': 'mp4',
|
'ext': 'mp4',
|
||||||
'title': 're:^Nederland 1 [0-9]{4}-[0-9]{2}-[0-9]{2} [0-9]{2}:[0-9]{2}$',
|
'title': 're:^NPO 1 [0-9]{4}-[0-9]{2}-[0-9]{2} [0-9]{2}:[0-9]{2}$',
|
||||||
'description': 'Livestream',
|
|
||||||
'is_live': True,
|
'is_live': True,
|
||||||
},
|
},
|
||||||
'params': {
|
'params': {
|
||||||
@ -283,58 +337,12 @@ class NPOLiveIE(NPOBaseIE):
|
|||||||
live_id = self._search_regex(
|
live_id = self._search_regex(
|
||||||
r'data-prid="([^"]+)"', webpage, 'live id')
|
r'data-prid="([^"]+)"', webpage, 'live id')
|
||||||
|
|
||||||
metadata = self._download_json(
|
|
||||||
'http://e.omroep.nl/metadata/%s' % live_id,
|
|
||||||
display_id, transform_source=strip_jsonp)
|
|
||||||
|
|
||||||
token = self._get_token(display_id)
|
|
||||||
|
|
||||||
formats = []
|
|
||||||
|
|
||||||
streams = metadata.get('streams')
|
|
||||||
if streams:
|
|
||||||
for stream in streams:
|
|
||||||
stream_type = stream.get('type').lower()
|
|
||||||
# smooth streaming is not supported
|
|
||||||
if stream_type in ['ss', 'ms']:
|
|
||||||
continue
|
|
||||||
stream_info = self._download_json(
|
|
||||||
'http://ida.omroep.nl/aapi/?stream=%s&token=%s&type=jsonp'
|
|
||||||
% (stream.get('url'), token),
|
|
||||||
display_id, 'Downloading %s JSON' % stream_type)
|
|
||||||
if stream_info.get('error_code', 0) or stream_info.get('errorcode', 0):
|
|
||||||
continue
|
|
||||||
stream_url = self._download_json(
|
|
||||||
stream_info['stream'], display_id,
|
|
||||||
'Downloading %s URL' % stream_type,
|
|
||||||
'Unable to download %s URL' % stream_type,
|
|
||||||
transform_source=strip_jsonp, fatal=False)
|
|
||||||
if not stream_url:
|
|
||||||
continue
|
|
||||||
if stream_type == 'hds':
|
|
||||||
f4m_formats = self._extract_f4m_formats(stream_url, display_id)
|
|
||||||
# f4m downloader downloads only piece of live stream
|
|
||||||
for f4m_format in f4m_formats:
|
|
||||||
f4m_format['preference'] = -1
|
|
||||||
formats.extend(f4m_formats)
|
|
||||||
elif stream_type == 'hls':
|
|
||||||
formats.extend(self._extract_m3u8_formats(stream_url, display_id, 'mp4'))
|
|
||||||
else:
|
|
||||||
formats.append({
|
|
||||||
'url': stream_url,
|
|
||||||
'preference': -10,
|
|
||||||
})
|
|
||||||
|
|
||||||
self._sort_formats(formats)
|
|
||||||
|
|
||||||
return {
|
return {
|
||||||
|
'_type': 'url_transparent',
|
||||||
|
'url': 'npo:%s' % live_id,
|
||||||
|
'ie_key': NPOIE.ie_key(),
|
||||||
'id': live_id,
|
'id': live_id,
|
||||||
'display_id': display_id,
|
'display_id': display_id,
|
||||||
'title': self._live_title(metadata['titel']),
|
|
||||||
'description': metadata['info'],
|
|
||||||
'thumbnail': metadata.get('images', [{'url': None}])[-1]['url'],
|
|
||||||
'formats': formats,
|
|
||||||
'is_live': True,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
@ -416,7 +424,21 @@ class NPORadioFragmentIE(InfoExtractor):
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
class SchoolTVIE(InfoExtractor):
|
class NPODataMidEmbedIE(InfoExtractor):
|
||||||
|
def _real_extract(self, url):
|
||||||
|
display_id = self._match_id(url)
|
||||||
|
webpage = self._download_webpage(url, display_id)
|
||||||
|
video_id = self._search_regex(
|
||||||
|
r'data-mid=(["\'])(?P<id>(?:(?!\1).)+)\1', webpage, 'video_id', group='id')
|
||||||
|
return {
|
||||||
|
'_type': 'url_transparent',
|
||||||
|
'ie_key': 'NPO',
|
||||||
|
'url': 'npo:%s' % video_id,
|
||||||
|
'display_id': display_id
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class SchoolTVIE(NPODataMidEmbedIE):
|
||||||
IE_NAME = 'schooltv'
|
IE_NAME = 'schooltv'
|
||||||
_VALID_URL = r'https?://(?:www\.)?schooltv\.nl/video/(?P<id>[^/?#&]+)'
|
_VALID_URL = r'https?://(?:www\.)?schooltv\.nl/video/(?P<id>[^/?#&]+)'
|
||||||
|
|
||||||
@ -435,17 +457,25 @@ class SchoolTVIE(InfoExtractor):
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
def _real_extract(self, url):
|
|
||||||
display_id = self._match_id(url)
|
class HetKlokhuisIE(NPODataMidEmbedIE):
|
||||||
webpage = self._download_webpage(url, display_id)
|
IE_NAME = 'hetklokhuis'
|
||||||
video_id = self._search_regex(
|
_VALID_URL = r'https?://(?:www\.)?hetklokhuis.nl/[^/]+/\d+/(?P<id>[^/?#&]+)'
|
||||||
r'data-mid=(["\'])(?P<id>(?:(?!\1).)+)\1', webpage, 'video_id', group='id')
|
|
||||||
return {
|
_TEST = {
|
||||||
'_type': 'url_transparent',
|
'url': 'http://hetklokhuis.nl/tv-uitzending/3471/Zwaartekrachtsgolven',
|
||||||
'ie_key': 'NPO',
|
'info_dict': {
|
||||||
'url': 'npo:%s' % video_id,
|
'id': 'VPWON_1260528',
|
||||||
'display_id': display_id
|
'display_id': 'Zwaartekrachtsgolven',
|
||||||
|
'ext': 'm4v',
|
||||||
|
'title': 'Het Klokhuis: Zwaartekrachtsgolven',
|
||||||
|
'description': 'md5:c94f31fb930d76c2efa4a4a71651dd48',
|
||||||
|
'upload_date': '20170223',
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
'skip_download': True
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
class NPOPlaylistBaseIE(NPOIE):
|
class NPOPlaylistBaseIE(NPOIE):
|
||||||
|
@ -72,16 +72,21 @@ class OpenloadIE(InfoExtractor):
|
|||||||
raise ExtractorError('File not found', expected=True)
|
raise ExtractorError('File not found', expected=True)
|
||||||
|
|
||||||
ol_id = self._search_regex(
|
ol_id = self._search_regex(
|
||||||
'<span[^>]+id="[^"]+"[^>]*>([0-9]+)</span>',
|
'<span[^>]+id="[^"]+"[^>]*>([0-9A-Za-z]+)</span>',
|
||||||
webpage, 'openload ID')
|
webpage, 'openload ID')
|
||||||
|
|
||||||
first_two_chars = int(float(ol_id[0:][:2]))
|
first_char = int(ol_id[0])
|
||||||
urlcode = []
|
urlcode = []
|
||||||
num = 2
|
num = 1
|
||||||
|
|
||||||
while num < len(ol_id):
|
while num < len(ol_id):
|
||||||
key = int(float(ol_id[num + 3:][:2]))
|
i = ord(ol_id[num])
|
||||||
urlcode.append((key, compat_chr(int(float(ol_id[num:][:3])) - first_two_chars)))
|
key = 0
|
||||||
|
if i <= 90:
|
||||||
|
key = i - 65
|
||||||
|
elif i >= 97:
|
||||||
|
key = 25 + i - 97
|
||||||
|
urlcode.append((key, compat_chr(int(ol_id[num + 2:num + 5]) // int(ol_id[num + 1]) - first_char)))
|
||||||
num += 5
|
num += 5
|
||||||
|
|
||||||
video_url = 'https://openload.co/stream/' + ''.join(
|
video_url = 'https://openload.co/stream/' + ''.join(
|
||||||
|
@ -1,57 +0,0 @@
|
|||||||
from __future__ import unicode_literals
|
|
||||||
|
|
||||||
import re
|
|
||||||
|
|
||||||
from .common import InfoExtractor
|
|
||||||
from ..utils import int_or_none
|
|
||||||
|
|
||||||
|
|
||||||
class SciVeeIE(InfoExtractor):
|
|
||||||
_VALID_URL = r'https?://(?:www\.)?scivee\.tv/node/(?P<id>\d+)'
|
|
||||||
|
|
||||||
_TEST = {
|
|
||||||
'url': 'http://www.scivee.tv/node/62352',
|
|
||||||
'md5': 'b16699b74c9e6a120f6772a44960304f',
|
|
||||||
'info_dict': {
|
|
||||||
'id': '62352',
|
|
||||||
'ext': 'mp4',
|
|
||||||
'title': 'Adam Arkin at the 2014 DOE JGI Genomics of Energy & Environment Meeting',
|
|
||||||
'description': 'md5:81f1710638e11a481358fab1b11059d7',
|
|
||||||
},
|
|
||||||
'skip': 'Not accessible from Travis CI server',
|
|
||||||
}
|
|
||||||
|
|
||||||
def _real_extract(self, url):
|
|
||||||
mobj = re.match(self._VALID_URL, url)
|
|
||||||
video_id = mobj.group('id')
|
|
||||||
|
|
||||||
# annotations XML is malformed
|
|
||||||
annotations = self._download_webpage(
|
|
||||||
'http://www.scivee.tv/assets/annotations/%s' % video_id, video_id, 'Downloading annotations')
|
|
||||||
|
|
||||||
title = self._html_search_regex(r'<title>([^<]+)</title>', annotations, 'title')
|
|
||||||
description = self._html_search_regex(r'<abstract>([^<]+)</abstract>', annotations, 'abstract', fatal=False)
|
|
||||||
filesize = int_or_none(self._html_search_regex(
|
|
||||||
r'<filesize>([^<]+)</filesize>', annotations, 'filesize', fatal=False))
|
|
||||||
|
|
||||||
formats = [
|
|
||||||
{
|
|
||||||
'url': 'http://www.scivee.tv/assets/audio/%s' % video_id,
|
|
||||||
'ext': 'mp3',
|
|
||||||
'format_id': 'audio',
|
|
||||||
},
|
|
||||||
{
|
|
||||||
'url': 'http://www.scivee.tv/assets/video/%s' % video_id,
|
|
||||||
'ext': 'mp4',
|
|
||||||
'format_id': 'video',
|
|
||||||
'filesize': filesize,
|
|
||||||
},
|
|
||||||
]
|
|
||||||
|
|
||||||
return {
|
|
||||||
'id': video_id,
|
|
||||||
'title': title,
|
|
||||||
'description': description,
|
|
||||||
'thumbnail': 'http://www.scivee.tv/assets/videothumb/%s' % video_id,
|
|
||||||
'formats': formats,
|
|
||||||
}
|
|
@ -3,7 +3,10 @@ from __future__ import unicode_literals
|
|||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
|
|
||||||
from ..compat import compat_urlparse
|
from ..compat import compat_urlparse
|
||||||
from ..utils import qualities
|
from ..utils import (
|
||||||
|
int_or_none,
|
||||||
|
qualities,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
class TheSceneIE(InfoExtractor):
|
class TheSceneIE(InfoExtractor):
|
||||||
@ -16,6 +19,11 @@ class TheSceneIE(InfoExtractor):
|
|||||||
'ext': 'mp4',
|
'ext': 'mp4',
|
||||||
'title': 'Narciso Rodriguez: Spring 2013 Ready-to-Wear',
|
'title': 'Narciso Rodriguez: Spring 2013 Ready-to-Wear',
|
||||||
'display_id': 'narciso-rodriguez-spring-2013-ready-to-wear',
|
'display_id': 'narciso-rodriguez-spring-2013-ready-to-wear',
|
||||||
|
'duration': 127,
|
||||||
|
'series': 'Style.com Fashion Shows',
|
||||||
|
'season': 'Ready To Wear Spring 2013',
|
||||||
|
'tags': list,
|
||||||
|
'categories': list,
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -32,21 +40,29 @@ class TheSceneIE(InfoExtractor):
|
|||||||
player = self._download_webpage(player_url, display_id)
|
player = self._download_webpage(player_url, display_id)
|
||||||
info = self._parse_json(
|
info = self._parse_json(
|
||||||
self._search_regex(
|
self._search_regex(
|
||||||
r'(?m)var\s+video\s+=\s+({.+?});$', player, 'info json'),
|
r'(?m)video\s*:\s*({.+?}),$', player, 'info json'),
|
||||||
display_id)
|
display_id)
|
||||||
|
|
||||||
|
video_id = info['id']
|
||||||
|
title = info['title']
|
||||||
|
|
||||||
qualities_order = qualities(('low', 'high'))
|
qualities_order = qualities(('low', 'high'))
|
||||||
formats = [{
|
formats = [{
|
||||||
'format_id': '{0}-{1}'.format(f['type'].split('/')[0], f['quality']),
|
'format_id': '{0}-{1}'.format(f['type'].split('/')[0], f['quality']),
|
||||||
'url': f['src'],
|
'url': f['src'],
|
||||||
'quality': qualities_order(f['quality']),
|
'quality': qualities_order(f['quality']),
|
||||||
} for f in info['sources'][0]]
|
} for f in info['sources']]
|
||||||
self._sort_formats(formats)
|
self._sort_formats(formats)
|
||||||
|
|
||||||
return {
|
return {
|
||||||
'id': info['id'],
|
'id': video_id,
|
||||||
'display_id': display_id,
|
'display_id': display_id,
|
||||||
'title': info['title'],
|
'title': title,
|
||||||
'formats': formats,
|
'formats': formats,
|
||||||
'thumbnail': info.get('poster_frame'),
|
'thumbnail': info.get('poster_frame'),
|
||||||
|
'duration': int_or_none(info.get('duration')),
|
||||||
|
'series': info.get('series_title'),
|
||||||
|
'season': info.get('season_title'),
|
||||||
|
'tags': info.get('tags'),
|
||||||
|
'categories': info.get('categories'),
|
||||||
}
|
}
|
||||||
|
@ -16,6 +16,7 @@ class TubiTvIE(InfoExtractor):
|
|||||||
_VALID_URL = r'https?://(?:www\.)?tubitv\.com/video/(?P<id>[0-9]+)'
|
_VALID_URL = r'https?://(?:www\.)?tubitv\.com/video/(?P<id>[0-9]+)'
|
||||||
_LOGIN_URL = 'http://tubitv.com/login'
|
_LOGIN_URL = 'http://tubitv.com/login'
|
||||||
_NETRC_MACHINE = 'tubitv'
|
_NETRC_MACHINE = 'tubitv'
|
||||||
|
_GEO_COUNTRIES = ['US']
|
||||||
_TEST = {
|
_TEST = {
|
||||||
'url': 'http://tubitv.com/video/283829/the_comedian_at_the_friday',
|
'url': 'http://tubitv.com/video/283829/the_comedian_at_the_friday',
|
||||||
'md5': '43ac06be9326f41912dc64ccf7a80320',
|
'md5': '43ac06be9326f41912dc64ccf7a80320',
|
||||||
|
@ -17,6 +17,9 @@ class TvigleIE(InfoExtractor):
|
|||||||
IE_DESC = 'Интернет-телевидение Tvigle.ru'
|
IE_DESC = 'Интернет-телевидение Tvigle.ru'
|
||||||
_VALID_URL = r'https?://(?:www\.)?(?:tvigle\.ru/(?:[^/]+/)+(?P<display_id>[^/]+)/$|cloud\.tvigle\.ru/video/(?P<id>\d+))'
|
_VALID_URL = r'https?://(?:www\.)?(?:tvigle\.ru/(?:[^/]+/)+(?P<display_id>[^/]+)/$|cloud\.tvigle\.ru/video/(?P<id>\d+))'
|
||||||
|
|
||||||
|
_GEO_BYPASS = False
|
||||||
|
_GEO_COUNTRIES = ['RU']
|
||||||
|
|
||||||
_TESTS = [
|
_TESTS = [
|
||||||
{
|
{
|
||||||
'url': 'http://www.tvigle.ru/video/sokrat/',
|
'url': 'http://www.tvigle.ru/video/sokrat/',
|
||||||
@ -72,8 +75,13 @@ class TvigleIE(InfoExtractor):
|
|||||||
|
|
||||||
error_message = item.get('errorMessage')
|
error_message = item.get('errorMessage')
|
||||||
if not videos and error_message:
|
if not videos and error_message:
|
||||||
raise ExtractorError(
|
if item.get('isGeoBlocked') is True:
|
||||||
'%s returned error: %s' % (self.IE_NAME, error_message), expected=True)
|
self.raise_geo_restricted(
|
||||||
|
msg=error_message, countries=self._GEO_COUNTRIES)
|
||||||
|
else:
|
||||||
|
raise ExtractorError(
|
||||||
|
'%s returned error: %s' % (self.IE_NAME, error_message),
|
||||||
|
expected=True)
|
||||||
|
|
||||||
title = item['title']
|
title = item['title']
|
||||||
description = item.get('description')
|
description = item.get('description')
|
||||||
|
@ -17,12 +17,12 @@ from ..utils import (
|
|||||||
|
|
||||||
|
|
||||||
class VevoBaseIE(InfoExtractor):
|
class VevoBaseIE(InfoExtractor):
|
||||||
def _extract_json(self, webpage, video_id, item):
|
def _extract_json(self, webpage, video_id):
|
||||||
return self._parse_json(
|
return self._parse_json(
|
||||||
self._search_regex(
|
self._search_regex(
|
||||||
r'window\.__INITIAL_STORE__\s*=\s*({.+?});\s*</script>',
|
r'window\.__INITIAL_STORE__\s*=\s*({.+?});\s*</script>',
|
||||||
webpage, 'initial store'),
|
webpage, 'initial store'),
|
||||||
video_id)['default'][item]
|
video_id)
|
||||||
|
|
||||||
|
|
||||||
class VevoIE(VevoBaseIE):
|
class VevoIE(VevoBaseIE):
|
||||||
@ -139,6 +139,11 @@ class VevoIE(VevoBaseIE):
|
|||||||
# no genres available
|
# no genres available
|
||||||
'url': 'http://www.vevo.com/watch/INS171400764',
|
'url': 'http://www.vevo.com/watch/INS171400764',
|
||||||
'only_matching': True,
|
'only_matching': True,
|
||||||
|
}, {
|
||||||
|
# Another case available only via the webpage; using streams/streamsV3 formats
|
||||||
|
# Geo-restricted to Netherlands/Germany
|
||||||
|
'url': 'http://www.vevo.com/watch/boostee/pop-corn-clip-officiel/FR1A91600909',
|
||||||
|
'only_matching': True,
|
||||||
}]
|
}]
|
||||||
_VERSIONS = {
|
_VERSIONS = {
|
||||||
0: 'youtube', # only in AuthenticateVideo videoVersions
|
0: 'youtube', # only in AuthenticateVideo videoVersions
|
||||||
@ -193,7 +198,14 @@ class VevoIE(VevoBaseIE):
|
|||||||
# https://github.com/rg3/youtube-dl/issues/9366)
|
# https://github.com/rg3/youtube-dl/issues/9366)
|
||||||
if not video_versions:
|
if not video_versions:
|
||||||
webpage = self._download_webpage(url, video_id)
|
webpage = self._download_webpage(url, video_id)
|
||||||
video_versions = self._extract_json(webpage, video_id, 'streams')[video_id][0]
|
json_data = self._extract_json(webpage, video_id)
|
||||||
|
if 'streams' in json_data.get('default', {}):
|
||||||
|
video_versions = json_data['default']['streams'][video_id][0]
|
||||||
|
else:
|
||||||
|
video_versions = [
|
||||||
|
value
|
||||||
|
for key, value in json_data['apollo']['data'].items()
|
||||||
|
if key.startswith('%s.streams' % video_id)]
|
||||||
|
|
||||||
uploader = None
|
uploader = None
|
||||||
artist = None
|
artist = None
|
||||||
@ -207,7 +219,7 @@ class VevoIE(VevoBaseIE):
|
|||||||
|
|
||||||
formats = []
|
formats = []
|
||||||
for video_version in video_versions:
|
for video_version in video_versions:
|
||||||
version = self._VERSIONS.get(video_version['version'])
|
version = self._VERSIONS.get(video_version.get('version'), 'generic')
|
||||||
version_url = video_version.get('url')
|
version_url = video_version.get('url')
|
||||||
if not version_url:
|
if not version_url:
|
||||||
continue
|
continue
|
||||||
@ -339,7 +351,7 @@ class VevoPlaylistIE(VevoBaseIE):
|
|||||||
if video_id:
|
if video_id:
|
||||||
return self.url_result('vevo:%s' % video_id, VevoIE.ie_key())
|
return self.url_result('vevo:%s' % video_id, VevoIE.ie_key())
|
||||||
|
|
||||||
playlists = self._extract_json(webpage, playlist_id, '%ss' % playlist_kind)
|
playlists = self._extract_json(webpage, playlist_id)['default']['%ss' % playlist_kind]
|
||||||
|
|
||||||
playlist = (list(playlists.values())[0]
|
playlist = (list(playlists.values())[0]
|
||||||
if playlist_kind == 'playlist' else playlists[playlist_id])
|
if playlist_kind == 'playlist' else playlists[playlist_id])
|
||||||
|
@ -5,6 +5,7 @@ import re
|
|||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from ..utils import (
|
from ..utils import (
|
||||||
dict_get,
|
dict_get,
|
||||||
|
ExtractorError,
|
||||||
int_or_none,
|
int_or_none,
|
||||||
parse_duration,
|
parse_duration,
|
||||||
unified_strdate,
|
unified_strdate,
|
||||||
@ -57,6 +58,10 @@ class XHamsterIE(InfoExtractor):
|
|||||||
}, {
|
}, {
|
||||||
'url': 'https://xhamster.com/movies/2272726/amber_slayed_by_the_knight.html',
|
'url': 'https://xhamster.com/movies/2272726/amber_slayed_by_the_knight.html',
|
||||||
'only_matching': True,
|
'only_matching': True,
|
||||||
|
}, {
|
||||||
|
# This video is visible for marcoalfa123456's friends only
|
||||||
|
'url': 'https://it.xhamster.com/movies/7263980/la_mia_vicina.html',
|
||||||
|
'only_matching': True,
|
||||||
}]
|
}]
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
@ -78,6 +83,12 @@ class XHamsterIE(InfoExtractor):
|
|||||||
mrss_url = '%s://xhamster.com/movies/%s/%s.html' % (proto, video_id, seo)
|
mrss_url = '%s://xhamster.com/movies/%s/%s.html' % (proto, video_id, seo)
|
||||||
webpage = self._download_webpage(mrss_url, video_id)
|
webpage = self._download_webpage(mrss_url, video_id)
|
||||||
|
|
||||||
|
error = self._html_search_regex(
|
||||||
|
r'<div[^>]+id=["\']videoClosed["\'][^>]*>(.+?)</div>',
|
||||||
|
webpage, 'error', default=None)
|
||||||
|
if error:
|
||||||
|
raise ExtractorError(error, expected=True)
|
||||||
|
|
||||||
title = self._html_search_regex(
|
title = self._html_search_regex(
|
||||||
[r'<h1[^>]*>([^<]+)</h1>',
|
[r'<h1[^>]*>([^<]+)</h1>',
|
||||||
r'<meta[^>]+itemprop=".*?caption.*?"[^>]+content="(.+?)"',
|
r'<meta[^>]+itemprop=".*?caption.*?"[^>]+content="(.+?)"',
|
||||||
|
@ -47,7 +47,6 @@ from ..utils import (
|
|||||||
unsmuggle_url,
|
unsmuggle_url,
|
||||||
uppercase_escape,
|
uppercase_escape,
|
||||||
urlencode_postdata,
|
urlencode_postdata,
|
||||||
ISO3166Utils,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
@ -371,6 +370,8 @@ class YoutubeIE(YoutubeBaseInfoExtractor):
|
|||||||
}
|
}
|
||||||
_SUBTITLE_FORMATS = ('ttml', 'vtt')
|
_SUBTITLE_FORMATS = ('ttml', 'vtt')
|
||||||
|
|
||||||
|
_GEO_BYPASS = False
|
||||||
|
|
||||||
IE_NAME = 'youtube'
|
IE_NAME = 'youtube'
|
||||||
_TESTS = [
|
_TESTS = [
|
||||||
{
|
{
|
||||||
@ -917,7 +918,12 @@ class YoutubeIE(YoutubeBaseInfoExtractor):
|
|||||||
# itag 212
|
# itag 212
|
||||||
'url': '1t24XAntNCY',
|
'url': '1t24XAntNCY',
|
||||||
'only_matching': True,
|
'only_matching': True,
|
||||||
}
|
},
|
||||||
|
{
|
||||||
|
# geo restricted to JP
|
||||||
|
'url': 'sJL6WA-aGkQ',
|
||||||
|
'only_matching': True,
|
||||||
|
},
|
||||||
]
|
]
|
||||||
|
|
||||||
def __init__(self, *args, **kwargs):
|
def __init__(self, *args, **kwargs):
|
||||||
@ -1376,11 +1382,11 @@ class YoutubeIE(YoutubeBaseInfoExtractor):
|
|||||||
if 'token' not in video_info:
|
if 'token' not in video_info:
|
||||||
if 'reason' in video_info:
|
if 'reason' in video_info:
|
||||||
if 'The uploader has not made this video available in your country.' in video_info['reason']:
|
if 'The uploader has not made this video available in your country.' in video_info['reason']:
|
||||||
regions_allowed = self._html_search_meta('regionsAllowed', video_webpage, default=None)
|
regions_allowed = self._html_search_meta(
|
||||||
if regions_allowed:
|
'regionsAllowed', video_webpage, default=None)
|
||||||
raise ExtractorError('YouTube said: This video is available in %s only' % (
|
countries = regions_allowed.split(',') if regions_allowed else None
|
||||||
', '.join(map(ISO3166Utils.short2full, regions_allowed.split(',')))),
|
self.raise_geo_restricted(
|
||||||
expected=True)
|
msg=video_info['reason'][0], countries=countries)
|
||||||
raise ExtractorError(
|
raise ExtractorError(
|
||||||
'YouTube said: %s' % video_info['reason'][0],
|
'YouTube said: %s' % video_info['reason'][0],
|
||||||
expected=True, video_id=video_id)
|
expected=True, video_id=video_id)
|
||||||
@ -1448,7 +1454,7 @@ class YoutubeIE(YoutubeBaseInfoExtractor):
|
|||||||
|
|
||||||
# Check for "rental" videos
|
# Check for "rental" videos
|
||||||
if 'ypc_video_rental_bar_text' in video_info and 'author' not in video_info:
|
if 'ypc_video_rental_bar_text' in video_info and 'author' not in video_info:
|
||||||
raise ExtractorError('"rental" videos not supported')
|
raise ExtractorError('"rental" videos not supported. See https://github.com/rg3/youtube-dl/issues/359 for more information.', expected=True)
|
||||||
|
|
||||||
# Start extracting information
|
# Start extracting information
|
||||||
self.report_information_extraction(video_id)
|
self.report_information_extraction(video_id)
|
||||||
@ -1845,7 +1851,7 @@ class YoutubePlaylistIE(YoutubePlaylistBaseInfoExtractor):
|
|||||||
(?:
|
(?:
|
||||||
youtube\.com/
|
youtube\.com/
|
||||||
(?:
|
(?:
|
||||||
(?:course|view_play_list|my_playlists|artist|playlist|watch|embed/videoseries)
|
(?:course|view_play_list|my_playlists|artist|playlist|watch|embed/(?:videoseries|[0-9A-Za-z_-]{11}))
|
||||||
\? (?:.*?[&;])*? (?:p|a|list)=
|
\? (?:.*?[&;])*? (?:p|a|list)=
|
||||||
| p/
|
| p/
|
||||||
)|
|
)|
|
||||||
@ -1918,6 +1924,13 @@ class YoutubePlaylistIE(YoutubePlaylistBaseInfoExtractor):
|
|||||||
'title': 'JODA15',
|
'title': 'JODA15',
|
||||||
'id': 'PL6IaIsEjSbf96XFRuNccS_RuEXwNdsoEu',
|
'id': 'PL6IaIsEjSbf96XFRuNccS_RuEXwNdsoEu',
|
||||||
}
|
}
|
||||||
|
}, {
|
||||||
|
'url': 'http://www.youtube.com/embed/_xDOZElKyNU?list=PLsyOSbh5bs16vubvKePAQ1x3PhKavfBIl',
|
||||||
|
'playlist_mincount': 485,
|
||||||
|
'info_dict': {
|
||||||
|
'title': '2017 華語最新單曲 (2/24更新)',
|
||||||
|
'id': 'PLsyOSbh5bs16vubvKePAQ1x3PhKavfBIl',
|
||||||
|
}
|
||||||
}, {
|
}, {
|
||||||
'note': 'Embedded SWF player',
|
'note': 'Embedded SWF player',
|
||||||
'url': 'https://www.youtube.com/p/YN5VISEtHet5D4NEvfTd0zcgFk84NqFZ?hl=en_US&fs=1&rel=0',
|
'url': 'https://www.youtube.com/p/YN5VISEtHet5D4NEvfTd0zcgFk84NqFZ?hl=en_US&fs=1&rel=0',
|
||||||
@ -2066,7 +2079,7 @@ class YoutubePlaylistIE(YoutubePlaylistBaseInfoExtractor):
|
|||||||
# Check if it's a video-specific URL
|
# Check if it's a video-specific URL
|
||||||
query_dict = compat_urlparse.parse_qs(compat_urlparse.urlparse(url).query)
|
query_dict = compat_urlparse.parse_qs(compat_urlparse.urlparse(url).query)
|
||||||
video_id = query_dict.get('v', [None])[0] or self._search_regex(
|
video_id = query_dict.get('v', [None])[0] or self._search_regex(
|
||||||
r'(?:^|//)youtu\.be/([0-9A-Za-z_-]{11})', url,
|
r'(?:(?:^|//)youtu\.be/|youtube\.com/embed/(?!videoseries))([0-9A-Za-z_-]{11})', url,
|
||||||
'video id', default=None)
|
'video id', default=None)
|
||||||
if video_id:
|
if video_id:
|
||||||
if self._downloader.params.get('noplaylist'):
|
if self._downloader.params.get('noplaylist'):
|
||||||
@ -2226,7 +2239,7 @@ class YoutubeUserIE(YoutubeChannelIE):
|
|||||||
'url': 'https://www.youtube.com/gametrailers',
|
'url': 'https://www.youtube.com/gametrailers',
|
||||||
'only_matching': True,
|
'only_matching': True,
|
||||||
}, {
|
}, {
|
||||||
# This channel is not available.
|
# This channel is not available, geo restricted to JP
|
||||||
'url': 'https://www.youtube.com/user/kananishinoSMEJ/videos',
|
'url': 'https://www.youtube.com/user/kananishinoSMEJ/videos',
|
||||||
'only_matching': True,
|
'only_matching': True,
|
||||||
}]
|
}]
|
||||||
|
@ -473,7 +473,8 @@ def timeconvert(timestr):
|
|||||||
def sanitize_filename(s, restricted=False, is_id=False):
|
def sanitize_filename(s, restricted=False, is_id=False):
|
||||||
"""Sanitizes a string so it could be used as part of a filename.
|
"""Sanitizes a string so it could be used as part of a filename.
|
||||||
If restricted is set, use a stricter subset of allowed characters.
|
If restricted is set, use a stricter subset of allowed characters.
|
||||||
Set is_id if this is not an arbitrary string, but an ID that should be kept if possible
|
Set is_id if this is not an arbitrary string, but an ID that should be kept
|
||||||
|
if possible.
|
||||||
"""
|
"""
|
||||||
def replace_insane(char):
|
def replace_insane(char):
|
||||||
if restricted and char in ACCENT_CHARS:
|
if restricted and char in ACCENT_CHARS:
|
||||||
@ -3319,6 +3320,57 @@ class PerRequestProxyHandler(compat_urllib_request.ProxyHandler):
|
|||||||
self, req, proxy, type)
|
self, req, proxy, type)
|
||||||
|
|
||||||
|
|
||||||
|
# Both long_to_bytes and bytes_to_long are adapted from PyCrypto, which is
|
||||||
|
# released into Public Domain
|
||||||
|
# https://github.com/dlitz/pycrypto/blob/master/lib/Crypto/Util/number.py#L387
|
||||||
|
|
||||||
|
def long_to_bytes(n, blocksize=0):
|
||||||
|
"""long_to_bytes(n:long, blocksize:int) : string
|
||||||
|
Convert a long integer to a byte string.
|
||||||
|
|
||||||
|
If optional blocksize is given and greater than zero, pad the front of the
|
||||||
|
byte string with binary zeros so that the length is a multiple of
|
||||||
|
blocksize.
|
||||||
|
"""
|
||||||
|
# after much testing, this algorithm was deemed to be the fastest
|
||||||
|
s = b''
|
||||||
|
n = int(n)
|
||||||
|
while n > 0:
|
||||||
|
s = compat_struct_pack('>I', n & 0xffffffff) + s
|
||||||
|
n = n >> 32
|
||||||
|
# strip off leading zeros
|
||||||
|
for i in range(len(s)):
|
||||||
|
if s[i] != b'\000'[0]:
|
||||||
|
break
|
||||||
|
else:
|
||||||
|
# only happens when n == 0
|
||||||
|
s = b'\000'
|
||||||
|
i = 0
|
||||||
|
s = s[i:]
|
||||||
|
# add back some pad bytes. this could be done more efficiently w.r.t. the
|
||||||
|
# de-padding being done above, but sigh...
|
||||||
|
if blocksize > 0 and len(s) % blocksize:
|
||||||
|
s = (blocksize - len(s) % blocksize) * b'\000' + s
|
||||||
|
return s
|
||||||
|
|
||||||
|
|
||||||
|
def bytes_to_long(s):
|
||||||
|
"""bytes_to_long(string) : long
|
||||||
|
Convert a byte string to a long integer.
|
||||||
|
|
||||||
|
This is (essentially) the inverse of long_to_bytes().
|
||||||
|
"""
|
||||||
|
acc = 0
|
||||||
|
length = len(s)
|
||||||
|
if length % 4:
|
||||||
|
extra = (4 - length % 4)
|
||||||
|
s = b'\000' * extra + s
|
||||||
|
length = length + extra
|
||||||
|
for i in range(0, length, 4):
|
||||||
|
acc = (acc << 32) + compat_struct_unpack('>I', s[i:i + 4])[0]
|
||||||
|
return acc
|
||||||
|
|
||||||
|
|
||||||
def ohdave_rsa_encrypt(data, exponent, modulus):
|
def ohdave_rsa_encrypt(data, exponent, modulus):
|
||||||
'''
|
'''
|
||||||
Implement OHDave's RSA algorithm. See http://www.ohdave.com/rsa/
|
Implement OHDave's RSA algorithm. See http://www.ohdave.com/rsa/
|
||||||
@ -3336,6 +3388,21 @@ def ohdave_rsa_encrypt(data, exponent, modulus):
|
|||||||
return '%x' % encrypted
|
return '%x' % encrypted
|
||||||
|
|
||||||
|
|
||||||
|
def pkcs1pad(data, length):
|
||||||
|
"""
|
||||||
|
Padding input data with PKCS#1 scheme
|
||||||
|
|
||||||
|
@param {int[]} data input data
|
||||||
|
@param {int} length target length
|
||||||
|
@returns {int[]} padded data
|
||||||
|
"""
|
||||||
|
if len(data) > length - 11:
|
||||||
|
raise ValueError('Input data too long for PKCS#1 padding')
|
||||||
|
|
||||||
|
pseudo_random = [random.randint(0, 254) for _ in range(length - len(data) - 3)]
|
||||||
|
return [0, 2] + pseudo_random + [0] + data
|
||||||
|
|
||||||
|
|
||||||
def encode_base_n(num, n, table=None):
|
def encode_base_n(num, n, table=None):
|
||||||
FULL_TABLE = '0123456789abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ'
|
FULL_TABLE = '0123456789abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ'
|
||||||
if not table:
|
if not table:
|
||||||
|
@ -1,3 +1,3 @@
|
|||||||
from __future__ import unicode_literals
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
__version__ = '2017.02.24'
|
__version__ = '2017.03.02'
|
||||||
|
Reference in New Issue
Block a user