mirror of
https://github.com/yt-dlp/yt-dlp.git
synced 2026-01-19 05:11:29 +00:00
Compare commits
58 Commits
2023.10.13
...
2023.11.16
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
fe6c82ccff | ||
|
|
24f827875c | ||
|
|
15cb3528cb | ||
|
|
2325d03aa7 | ||
|
|
e569c2d1f4 | ||
|
|
a489f07150 | ||
|
|
5efe68b73c | ||
|
|
b530118e7f | ||
|
|
dcfad52812 | ||
|
|
0783fd558e | ||
|
|
0f634dba3a | ||
|
|
21dc069bea | ||
|
|
5d3a3cd493 | ||
|
|
a9d3f4b20a | ||
|
|
b012271d01 | ||
|
|
f04b5bedad | ||
|
|
d4f14a72dc | ||
|
|
87264d4fda | ||
|
|
a00af29853 | ||
|
|
0b6ad22e6a | ||
|
|
5438593a35 | ||
|
|
9970d74c83 | ||
|
|
20314dd46f | ||
|
|
1d03633c5a | ||
|
|
8afd9468b0 | ||
|
|
ef12dbdcd3 | ||
|
|
46acc418a5 | ||
|
|
6ba3085616 | ||
|
|
f6e97090d2 | ||
|
|
2863fcf2b6 | ||
|
|
c76c96677f | ||
|
|
15b252dfd2 | ||
|
|
312a2d1e8b | ||
|
|
54579be436 | ||
|
|
05adfd883a | ||
|
|
3ff494f6f4 | ||
|
|
9b5bedf13a | ||
|
|
cb480e390d | ||
|
|
25a4bd345a | ||
|
|
3906de0755 | ||
|
|
7d337ca977 | ||
|
|
10025b715e | ||
|
|
595ea4a99b | ||
|
|
2622c804d1 | ||
|
|
fd8fcf8f4f | ||
|
|
21b25281c5 | ||
|
|
4a601c9eff | ||
|
|
464327acdb | ||
|
|
ef79d20dc9 | ||
|
|
39abae2354 | ||
|
|
4ce2f29a50 | ||
|
|
177f0d963e | ||
|
|
8e02a4dcc8 | ||
|
|
7b8b1cf5eb | ||
|
|
a40e0b37df | ||
|
|
4e38e2ae9d | ||
|
|
8a8b54523a | ||
|
|
700444c23d |
17
.github/ISSUE_TEMPLATE/1_broken_site.yml
vendored
17
.github/ISSUE_TEMPLATE/1_broken_site.yml
vendored
@@ -18,7 +18,7 @@ body:
|
|||||||
options:
|
options:
|
||||||
- label: I'm reporting that yt-dlp is broken on a **supported** site
|
- label: I'm reporting that yt-dlp is broken on a **supported** site
|
||||||
required: true
|
required: true
|
||||||
- label: I've verified that I'm running yt-dlp version **2023.10.13** ([update instructions](https://github.com/yt-dlp/yt-dlp#update)) or later (specify commit)
|
- label: I've verified that I have **updated yt-dlp to nightly or master** ([update instructions](https://github.com/yt-dlp/yt-dlp#update-channels))
|
||||||
required: true
|
required: true
|
||||||
- label: I've checked that all provided URLs are playable in a browser with the same IP and same login details
|
- label: I've checked that all provided URLs are playable in a browser with the same IP and same login details
|
||||||
required: true
|
required: true
|
||||||
@@ -61,19 +61,18 @@ body:
|
|||||||
description: |
|
description: |
|
||||||
It should start like this:
|
It should start like this:
|
||||||
placeholder: |
|
placeholder: |
|
||||||
[debug] Command-line config: ['-vU', 'test:youtube']
|
[debug] Command-line config: ['-vU', 'https://www.youtube.com/watch?v=BaW_jenozKc']
|
||||||
[debug] Portable config "yt-dlp.conf": ['-i']
|
|
||||||
[debug] Encodings: locale cp65001, fs utf-8, pref cp65001, out utf-8, error utf-8, screen utf-8
|
[debug] Encodings: locale cp65001, fs utf-8, pref cp65001, out utf-8, error utf-8, screen utf-8
|
||||||
[debug] yt-dlp version 2023.10.13 [9d339c4] (win32_exe)
|
[debug] yt-dlp version nightly@... from yt-dlp/yt-dlp [b634ba742] (win_exe)
|
||||||
[debug] Python 3.8.10 (CPython 64bit) - Windows-10-10.0.22000-SP0
|
[debug] Python 3.8.10 (CPython 64bit) - Windows-10-10.0.22000-SP0
|
||||||
[debug] Checking exe version: ffmpeg -bsfs
|
|
||||||
[debug] Checking exe version: ffprobe -bsfs
|
|
||||||
[debug] exe versions: ffmpeg N-106550-g072101bd52-20220410 (fdk,setts), ffprobe N-106624-g391ce570c8-20220415, phantomjs 2.1.1
|
[debug] exe versions: ffmpeg N-106550-g072101bd52-20220410 (fdk,setts), ffprobe N-106624-g391ce570c8-20220415, phantomjs 2.1.1
|
||||||
[debug] Optional libraries: Cryptodome-3.15.0, brotli-1.0.9, certifi-2022.06.15, mutagen-1.45.1, sqlite3-2.6.0, websockets-10.3
|
[debug] Optional libraries: Cryptodome-3.15.0, brotli-1.0.9, certifi-2022.06.15, mutagen-1.45.1, sqlite3-2.6.0, websockets-10.3
|
||||||
[debug] Proxy map: {}
|
[debug] Proxy map: {}
|
||||||
[debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp/releases/latest
|
[debug] Request Handlers: urllib, requests
|
||||||
Latest version: 2023.10.13, Current version: 2023.10.13
|
[debug] Loaded 1893 extractors
|
||||||
yt-dlp is up to date (2023.10.13)
|
[debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp-nightly-builds/releases/latest
|
||||||
|
yt-dlp is up to date (nightly@... from yt-dlp/yt-dlp-nightly-builds)
|
||||||
|
[youtube] Extracting URL: https://www.youtube.com/watch?v=BaW_jenozKc
|
||||||
<more lines>
|
<more lines>
|
||||||
render: shell
|
render: shell
|
||||||
validations:
|
validations:
|
||||||
|
|||||||
@@ -18,7 +18,7 @@ body:
|
|||||||
options:
|
options:
|
||||||
- label: I'm reporting a new site support request
|
- label: I'm reporting a new site support request
|
||||||
required: true
|
required: true
|
||||||
- label: I've verified that I'm running yt-dlp version **2023.10.13** ([update instructions](https://github.com/yt-dlp/yt-dlp#update)) or later (specify commit)
|
- label: I've verified that I have **updated yt-dlp to nightly or master** ([update instructions](https://github.com/yt-dlp/yt-dlp#update-channels))
|
||||||
required: true
|
required: true
|
||||||
- label: I've checked that all provided URLs are playable in a browser with the same IP and same login details
|
- label: I've checked that all provided URLs are playable in a browser with the same IP and same login details
|
||||||
required: true
|
required: true
|
||||||
@@ -73,19 +73,18 @@ body:
|
|||||||
description: |
|
description: |
|
||||||
It should start like this:
|
It should start like this:
|
||||||
placeholder: |
|
placeholder: |
|
||||||
[debug] Command-line config: ['-vU', 'test:youtube']
|
[debug] Command-line config: ['-vU', 'https://www.youtube.com/watch?v=BaW_jenozKc']
|
||||||
[debug] Portable config "yt-dlp.conf": ['-i']
|
|
||||||
[debug] Encodings: locale cp65001, fs utf-8, pref cp65001, out utf-8, error utf-8, screen utf-8
|
[debug] Encodings: locale cp65001, fs utf-8, pref cp65001, out utf-8, error utf-8, screen utf-8
|
||||||
[debug] yt-dlp version 2023.10.13 [9d339c4] (win32_exe)
|
[debug] yt-dlp version nightly@... from yt-dlp/yt-dlp [b634ba742] (win_exe)
|
||||||
[debug] Python 3.8.10 (CPython 64bit) - Windows-10-10.0.22000-SP0
|
[debug] Python 3.8.10 (CPython 64bit) - Windows-10-10.0.22000-SP0
|
||||||
[debug] Checking exe version: ffmpeg -bsfs
|
|
||||||
[debug] Checking exe version: ffprobe -bsfs
|
|
||||||
[debug] exe versions: ffmpeg N-106550-g072101bd52-20220410 (fdk,setts), ffprobe N-106624-g391ce570c8-20220415, phantomjs 2.1.1
|
[debug] exe versions: ffmpeg N-106550-g072101bd52-20220410 (fdk,setts), ffprobe N-106624-g391ce570c8-20220415, phantomjs 2.1.1
|
||||||
[debug] Optional libraries: Cryptodome-3.15.0, brotli-1.0.9, certifi-2022.06.15, mutagen-1.45.1, sqlite3-2.6.0, websockets-10.3
|
[debug] Optional libraries: Cryptodome-3.15.0, brotli-1.0.9, certifi-2022.06.15, mutagen-1.45.1, sqlite3-2.6.0, websockets-10.3
|
||||||
[debug] Proxy map: {}
|
[debug] Proxy map: {}
|
||||||
[debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp/releases/latest
|
[debug] Request Handlers: urllib, requests
|
||||||
Latest version: 2023.10.13, Current version: 2023.10.13
|
[debug] Loaded 1893 extractors
|
||||||
yt-dlp is up to date (2023.10.13)
|
[debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp-nightly-builds/releases/latest
|
||||||
|
yt-dlp is up to date (nightly@... from yt-dlp/yt-dlp-nightly-builds)
|
||||||
|
[youtube] Extracting URL: https://www.youtube.com/watch?v=BaW_jenozKc
|
||||||
<more lines>
|
<more lines>
|
||||||
render: shell
|
render: shell
|
||||||
validations:
|
validations:
|
||||||
|
|||||||
@@ -18,7 +18,7 @@ body:
|
|||||||
options:
|
options:
|
||||||
- label: I'm requesting a site-specific feature
|
- label: I'm requesting a site-specific feature
|
||||||
required: true
|
required: true
|
||||||
- label: I've verified that I'm running yt-dlp version **2023.10.13** ([update instructions](https://github.com/yt-dlp/yt-dlp#update)) or later (specify commit)
|
- label: I've verified that I have **updated yt-dlp to nightly or master** ([update instructions](https://github.com/yt-dlp/yt-dlp#update-channels))
|
||||||
required: true
|
required: true
|
||||||
- label: I've checked that all provided URLs are playable in a browser with the same IP and same login details
|
- label: I've checked that all provided URLs are playable in a browser with the same IP and same login details
|
||||||
required: true
|
required: true
|
||||||
@@ -69,19 +69,18 @@ body:
|
|||||||
description: |
|
description: |
|
||||||
It should start like this:
|
It should start like this:
|
||||||
placeholder: |
|
placeholder: |
|
||||||
[debug] Command-line config: ['-vU', 'test:youtube']
|
[debug] Command-line config: ['-vU', 'https://www.youtube.com/watch?v=BaW_jenozKc']
|
||||||
[debug] Portable config "yt-dlp.conf": ['-i']
|
|
||||||
[debug] Encodings: locale cp65001, fs utf-8, pref cp65001, out utf-8, error utf-8, screen utf-8
|
[debug] Encodings: locale cp65001, fs utf-8, pref cp65001, out utf-8, error utf-8, screen utf-8
|
||||||
[debug] yt-dlp version 2023.10.13 [9d339c4] (win32_exe)
|
[debug] yt-dlp version nightly@... from yt-dlp/yt-dlp [b634ba742] (win_exe)
|
||||||
[debug] Python 3.8.10 (CPython 64bit) - Windows-10-10.0.22000-SP0
|
[debug] Python 3.8.10 (CPython 64bit) - Windows-10-10.0.22000-SP0
|
||||||
[debug] Checking exe version: ffmpeg -bsfs
|
|
||||||
[debug] Checking exe version: ffprobe -bsfs
|
|
||||||
[debug] exe versions: ffmpeg N-106550-g072101bd52-20220410 (fdk,setts), ffprobe N-106624-g391ce570c8-20220415, phantomjs 2.1.1
|
[debug] exe versions: ffmpeg N-106550-g072101bd52-20220410 (fdk,setts), ffprobe N-106624-g391ce570c8-20220415, phantomjs 2.1.1
|
||||||
[debug] Optional libraries: Cryptodome-3.15.0, brotli-1.0.9, certifi-2022.06.15, mutagen-1.45.1, sqlite3-2.6.0, websockets-10.3
|
[debug] Optional libraries: Cryptodome-3.15.0, brotli-1.0.9, certifi-2022.06.15, mutagen-1.45.1, sqlite3-2.6.0, websockets-10.3
|
||||||
[debug] Proxy map: {}
|
[debug] Proxy map: {}
|
||||||
[debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp/releases/latest
|
[debug] Request Handlers: urllib, requests
|
||||||
Latest version: 2023.10.13, Current version: 2023.10.13
|
[debug] Loaded 1893 extractors
|
||||||
yt-dlp is up to date (2023.10.13)
|
[debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp-nightly-builds/releases/latest
|
||||||
|
yt-dlp is up to date (nightly@... from yt-dlp/yt-dlp-nightly-builds)
|
||||||
|
[youtube] Extracting URL: https://www.youtube.com/watch?v=BaW_jenozKc
|
||||||
<more lines>
|
<more lines>
|
||||||
render: shell
|
render: shell
|
||||||
validations:
|
validations:
|
||||||
|
|||||||
17
.github/ISSUE_TEMPLATE/4_bug_report.yml
vendored
17
.github/ISSUE_TEMPLATE/4_bug_report.yml
vendored
@@ -18,7 +18,7 @@ body:
|
|||||||
options:
|
options:
|
||||||
- label: I'm reporting a bug unrelated to a specific site
|
- label: I'm reporting a bug unrelated to a specific site
|
||||||
required: true
|
required: true
|
||||||
- label: I've verified that I'm running yt-dlp version **2023.10.13** ([update instructions](https://github.com/yt-dlp/yt-dlp#update)) or later (specify commit)
|
- label: I've verified that I have **updated yt-dlp to nightly or master** ([update instructions](https://github.com/yt-dlp/yt-dlp#update-channels))
|
||||||
required: true
|
required: true
|
||||||
- label: I've checked that all provided URLs are playable in a browser with the same IP and same login details
|
- label: I've checked that all provided URLs are playable in a browser with the same IP and same login details
|
||||||
required: true
|
required: true
|
||||||
@@ -54,19 +54,18 @@ body:
|
|||||||
description: |
|
description: |
|
||||||
It should start like this:
|
It should start like this:
|
||||||
placeholder: |
|
placeholder: |
|
||||||
[debug] Command-line config: ['-vU', 'test:youtube']
|
[debug] Command-line config: ['-vU', 'https://www.youtube.com/watch?v=BaW_jenozKc']
|
||||||
[debug] Portable config "yt-dlp.conf": ['-i']
|
|
||||||
[debug] Encodings: locale cp65001, fs utf-8, pref cp65001, out utf-8, error utf-8, screen utf-8
|
[debug] Encodings: locale cp65001, fs utf-8, pref cp65001, out utf-8, error utf-8, screen utf-8
|
||||||
[debug] yt-dlp version 2023.10.13 [9d339c4] (win32_exe)
|
[debug] yt-dlp version nightly@... from yt-dlp/yt-dlp [b634ba742] (win_exe)
|
||||||
[debug] Python 3.8.10 (CPython 64bit) - Windows-10-10.0.22000-SP0
|
[debug] Python 3.8.10 (CPython 64bit) - Windows-10-10.0.22000-SP0
|
||||||
[debug] Checking exe version: ffmpeg -bsfs
|
|
||||||
[debug] Checking exe version: ffprobe -bsfs
|
|
||||||
[debug] exe versions: ffmpeg N-106550-g072101bd52-20220410 (fdk,setts), ffprobe N-106624-g391ce570c8-20220415, phantomjs 2.1.1
|
[debug] exe versions: ffmpeg N-106550-g072101bd52-20220410 (fdk,setts), ffprobe N-106624-g391ce570c8-20220415, phantomjs 2.1.1
|
||||||
[debug] Optional libraries: Cryptodome-3.15.0, brotli-1.0.9, certifi-2022.06.15, mutagen-1.45.1, sqlite3-2.6.0, websockets-10.3
|
[debug] Optional libraries: Cryptodome-3.15.0, brotli-1.0.9, certifi-2022.06.15, mutagen-1.45.1, sqlite3-2.6.0, websockets-10.3
|
||||||
[debug] Proxy map: {}
|
[debug] Proxy map: {}
|
||||||
[debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp/releases/latest
|
[debug] Request Handlers: urllib, requests
|
||||||
Latest version: 2023.10.13, Current version: 2023.10.13
|
[debug] Loaded 1893 extractors
|
||||||
yt-dlp is up to date (2023.10.13)
|
[debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp-nightly-builds/releases/latest
|
||||||
|
yt-dlp is up to date (nightly@... from yt-dlp/yt-dlp-nightly-builds)
|
||||||
|
[youtube] Extracting URL: https://www.youtube.com/watch?v=BaW_jenozKc
|
||||||
<more lines>
|
<more lines>
|
||||||
render: shell
|
render: shell
|
||||||
validations:
|
validations:
|
||||||
|
|||||||
17
.github/ISSUE_TEMPLATE/5_feature_request.yml
vendored
17
.github/ISSUE_TEMPLATE/5_feature_request.yml
vendored
@@ -20,7 +20,7 @@ body:
|
|||||||
required: true
|
required: true
|
||||||
- label: I've looked through the [README](https://github.com/yt-dlp/yt-dlp#readme)
|
- label: I've looked through the [README](https://github.com/yt-dlp/yt-dlp#readme)
|
||||||
required: true
|
required: true
|
||||||
- label: I've verified that I'm running yt-dlp version **2023.10.13** ([update instructions](https://github.com/yt-dlp/yt-dlp#update)) or later (specify commit)
|
- label: I've verified that I have **updated yt-dlp to nightly or master** ([update instructions](https://github.com/yt-dlp/yt-dlp#update-channels))
|
||||||
required: true
|
required: true
|
||||||
- label: I've searched [known issues](https://github.com/yt-dlp/yt-dlp/issues/3766) and the [bugtracker](https://github.com/yt-dlp/yt-dlp/issues?q=) for similar issues **including closed ones**. DO NOT post duplicates
|
- label: I've searched [known issues](https://github.com/yt-dlp/yt-dlp/issues/3766) and the [bugtracker](https://github.com/yt-dlp/yt-dlp/issues?q=) for similar issues **including closed ones**. DO NOT post duplicates
|
||||||
required: true
|
required: true
|
||||||
@@ -50,18 +50,17 @@ body:
|
|||||||
description: |
|
description: |
|
||||||
It should start like this:
|
It should start like this:
|
||||||
placeholder: |
|
placeholder: |
|
||||||
[debug] Command-line config: ['-vU', 'test:youtube']
|
[debug] Command-line config: ['-vU', 'https://www.youtube.com/watch?v=BaW_jenozKc']
|
||||||
[debug] Portable config "yt-dlp.conf": ['-i']
|
|
||||||
[debug] Encodings: locale cp65001, fs utf-8, pref cp65001, out utf-8, error utf-8, screen utf-8
|
[debug] Encodings: locale cp65001, fs utf-8, pref cp65001, out utf-8, error utf-8, screen utf-8
|
||||||
[debug] yt-dlp version 2023.10.13 [9d339c4] (win32_exe)
|
[debug] yt-dlp version nightly@... from yt-dlp/yt-dlp [b634ba742] (win_exe)
|
||||||
[debug] Python 3.8.10 (CPython 64bit) - Windows-10-10.0.22000-SP0
|
[debug] Python 3.8.10 (CPython 64bit) - Windows-10-10.0.22000-SP0
|
||||||
[debug] Checking exe version: ffmpeg -bsfs
|
|
||||||
[debug] Checking exe version: ffprobe -bsfs
|
|
||||||
[debug] exe versions: ffmpeg N-106550-g072101bd52-20220410 (fdk,setts), ffprobe N-106624-g391ce570c8-20220415, phantomjs 2.1.1
|
[debug] exe versions: ffmpeg N-106550-g072101bd52-20220410 (fdk,setts), ffprobe N-106624-g391ce570c8-20220415, phantomjs 2.1.1
|
||||||
[debug] Optional libraries: Cryptodome-3.15.0, brotli-1.0.9, certifi-2022.06.15, mutagen-1.45.1, sqlite3-2.6.0, websockets-10.3
|
[debug] Optional libraries: Cryptodome-3.15.0, brotli-1.0.9, certifi-2022.06.15, mutagen-1.45.1, sqlite3-2.6.0, websockets-10.3
|
||||||
[debug] Proxy map: {}
|
[debug] Proxy map: {}
|
||||||
[debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp/releases/latest
|
[debug] Request Handlers: urllib, requests
|
||||||
Latest version: 2023.10.13, Current version: 2023.10.13
|
[debug] Loaded 1893 extractors
|
||||||
yt-dlp is up to date (2023.10.13)
|
[debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp-nightly-builds/releases/latest
|
||||||
|
yt-dlp is up to date (nightly@... from yt-dlp/yt-dlp-nightly-builds)
|
||||||
|
[youtube] Extracting URL: https://www.youtube.com/watch?v=BaW_jenozKc
|
||||||
<more lines>
|
<more lines>
|
||||||
render: shell
|
render: shell
|
||||||
|
|||||||
17
.github/ISSUE_TEMPLATE/6_question.yml
vendored
17
.github/ISSUE_TEMPLATE/6_question.yml
vendored
@@ -26,7 +26,7 @@ body:
|
|||||||
required: true
|
required: true
|
||||||
- label: I've looked through the [README](https://github.com/yt-dlp/yt-dlp#readme)
|
- label: I've looked through the [README](https://github.com/yt-dlp/yt-dlp#readme)
|
||||||
required: true
|
required: true
|
||||||
- label: I've verified that I'm running yt-dlp version **2023.10.13** ([update instructions](https://github.com/yt-dlp/yt-dlp#update)) or later (specify commit)
|
- label: I've verified that I have **updated yt-dlp to nightly or master** ([update instructions](https://github.com/yt-dlp/yt-dlp#update-channels))
|
||||||
required: true
|
required: true
|
||||||
- label: I've searched [known issues](https://github.com/yt-dlp/yt-dlp/issues/3766) and the [bugtracker](https://github.com/yt-dlp/yt-dlp/issues?q=) for similar questions **including closed ones**. DO NOT post duplicates
|
- label: I've searched [known issues](https://github.com/yt-dlp/yt-dlp/issues/3766) and the [bugtracker](https://github.com/yt-dlp/yt-dlp/issues?q=) for similar questions **including closed ones**. DO NOT post duplicates
|
||||||
required: true
|
required: true
|
||||||
@@ -56,18 +56,17 @@ body:
|
|||||||
description: |
|
description: |
|
||||||
It should start like this:
|
It should start like this:
|
||||||
placeholder: |
|
placeholder: |
|
||||||
[debug] Command-line config: ['-vU', 'test:youtube']
|
[debug] Command-line config: ['-vU', 'https://www.youtube.com/watch?v=BaW_jenozKc']
|
||||||
[debug] Portable config "yt-dlp.conf": ['-i']
|
|
||||||
[debug] Encodings: locale cp65001, fs utf-8, pref cp65001, out utf-8, error utf-8, screen utf-8
|
[debug] Encodings: locale cp65001, fs utf-8, pref cp65001, out utf-8, error utf-8, screen utf-8
|
||||||
[debug] yt-dlp version 2023.10.13 [9d339c4] (win32_exe)
|
[debug] yt-dlp version nightly@... from yt-dlp/yt-dlp [b634ba742] (win_exe)
|
||||||
[debug] Python 3.8.10 (CPython 64bit) - Windows-10-10.0.22000-SP0
|
[debug] Python 3.8.10 (CPython 64bit) - Windows-10-10.0.22000-SP0
|
||||||
[debug] Checking exe version: ffmpeg -bsfs
|
|
||||||
[debug] Checking exe version: ffprobe -bsfs
|
|
||||||
[debug] exe versions: ffmpeg N-106550-g072101bd52-20220410 (fdk,setts), ffprobe N-106624-g391ce570c8-20220415, phantomjs 2.1.1
|
[debug] exe versions: ffmpeg N-106550-g072101bd52-20220410 (fdk,setts), ffprobe N-106624-g391ce570c8-20220415, phantomjs 2.1.1
|
||||||
[debug] Optional libraries: Cryptodome-3.15.0, brotli-1.0.9, certifi-2022.06.15, mutagen-1.45.1, sqlite3-2.6.0, websockets-10.3
|
[debug] Optional libraries: Cryptodome-3.15.0, brotli-1.0.9, certifi-2022.06.15, mutagen-1.45.1, sqlite3-2.6.0, websockets-10.3
|
||||||
[debug] Proxy map: {}
|
[debug] Proxy map: {}
|
||||||
[debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp/releases/latest
|
[debug] Request Handlers: urllib, requests
|
||||||
Latest version: 2023.10.13, Current version: 2023.10.13
|
[debug] Loaded 1893 extractors
|
||||||
yt-dlp is up to date (2023.10.13)
|
[debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp-nightly-builds/releases/latest
|
||||||
|
yt-dlp is up to date (nightly@... from yt-dlp/yt-dlp-nightly-builds)
|
||||||
|
[youtube] Extracting URL: https://www.youtube.com/watch?v=BaW_jenozKc
|
||||||
<more lines>
|
<more lines>
|
||||||
render: shell
|
render: shell
|
||||||
|
|||||||
@@ -12,7 +12,7 @@ body:
|
|||||||
options:
|
options:
|
||||||
- label: I'm reporting that yt-dlp is broken on a **supported** site
|
- label: I'm reporting that yt-dlp is broken on a **supported** site
|
||||||
required: true
|
required: true
|
||||||
- label: I've verified that I'm running yt-dlp version **%(version)s** ([update instructions](https://github.com/yt-dlp/yt-dlp#update)) or later (specify commit)
|
- label: I've verified that I have **updated yt-dlp to nightly or master** ([update instructions](https://github.com/yt-dlp/yt-dlp#update-channels))
|
||||||
required: true
|
required: true
|
||||||
- label: I've checked that all provided URLs are playable in a browser with the same IP and same login details
|
- label: I've checked that all provided URLs are playable in a browser with the same IP and same login details
|
||||||
required: true
|
required: true
|
||||||
|
|||||||
@@ -12,7 +12,7 @@ body:
|
|||||||
options:
|
options:
|
||||||
- label: I'm reporting a new site support request
|
- label: I'm reporting a new site support request
|
||||||
required: true
|
required: true
|
||||||
- label: I've verified that I'm running yt-dlp version **%(version)s** ([update instructions](https://github.com/yt-dlp/yt-dlp#update)) or later (specify commit)
|
- label: I've verified that I have **updated yt-dlp to nightly or master** ([update instructions](https://github.com/yt-dlp/yt-dlp#update-channels))
|
||||||
required: true
|
required: true
|
||||||
- label: I've checked that all provided URLs are playable in a browser with the same IP and same login details
|
- label: I've checked that all provided URLs are playable in a browser with the same IP and same login details
|
||||||
required: true
|
required: true
|
||||||
|
|||||||
@@ -12,7 +12,7 @@ body:
|
|||||||
options:
|
options:
|
||||||
- label: I'm requesting a site-specific feature
|
- label: I'm requesting a site-specific feature
|
||||||
required: true
|
required: true
|
||||||
- label: I've verified that I'm running yt-dlp version **%(version)s** ([update instructions](https://github.com/yt-dlp/yt-dlp#update)) or later (specify commit)
|
- label: I've verified that I have **updated yt-dlp to nightly or master** ([update instructions](https://github.com/yt-dlp/yt-dlp#update-channels))
|
||||||
required: true
|
required: true
|
||||||
- label: I've checked that all provided URLs are playable in a browser with the same IP and same login details
|
- label: I've checked that all provided URLs are playable in a browser with the same IP and same login details
|
||||||
required: true
|
required: true
|
||||||
|
|||||||
2
.github/ISSUE_TEMPLATE_tmpl/4_bug_report.yml
vendored
2
.github/ISSUE_TEMPLATE_tmpl/4_bug_report.yml
vendored
@@ -12,7 +12,7 @@ body:
|
|||||||
options:
|
options:
|
||||||
- label: I'm reporting a bug unrelated to a specific site
|
- label: I'm reporting a bug unrelated to a specific site
|
||||||
required: true
|
required: true
|
||||||
- label: I've verified that I'm running yt-dlp version **%(version)s** ([update instructions](https://github.com/yt-dlp/yt-dlp#update)) or later (specify commit)
|
- label: I've verified that I have **updated yt-dlp to nightly or master** ([update instructions](https://github.com/yt-dlp/yt-dlp#update-channels))
|
||||||
required: true
|
required: true
|
||||||
- label: I've checked that all provided URLs are playable in a browser with the same IP and same login details
|
- label: I've checked that all provided URLs are playable in a browser with the same IP and same login details
|
||||||
required: true
|
required: true
|
||||||
|
|||||||
@@ -14,7 +14,7 @@ body:
|
|||||||
required: true
|
required: true
|
||||||
- label: I've looked through the [README](https://github.com/yt-dlp/yt-dlp#readme)
|
- label: I've looked through the [README](https://github.com/yt-dlp/yt-dlp#readme)
|
||||||
required: true
|
required: true
|
||||||
- label: I've verified that I'm running yt-dlp version **%(version)s** ([update instructions](https://github.com/yt-dlp/yt-dlp#update)) or later (specify commit)
|
- label: I've verified that I have **updated yt-dlp to nightly or master** ([update instructions](https://github.com/yt-dlp/yt-dlp#update-channels))
|
||||||
required: true
|
required: true
|
||||||
- label: I've searched [known issues](https://github.com/yt-dlp/yt-dlp/issues/3766) and the [bugtracker](https://github.com/yt-dlp/yt-dlp/issues?q=) for similar issues **including closed ones**. DO NOT post duplicates
|
- label: I've searched [known issues](https://github.com/yt-dlp/yt-dlp/issues/3766) and the [bugtracker](https://github.com/yt-dlp/yt-dlp/issues?q=) for similar issues **including closed ones**. DO NOT post duplicates
|
||||||
required: true
|
required: true
|
||||||
|
|||||||
2
.github/ISSUE_TEMPLATE_tmpl/6_question.yml
vendored
2
.github/ISSUE_TEMPLATE_tmpl/6_question.yml
vendored
@@ -20,7 +20,7 @@ body:
|
|||||||
required: true
|
required: true
|
||||||
- label: I've looked through the [README](https://github.com/yt-dlp/yt-dlp#readme)
|
- label: I've looked through the [README](https://github.com/yt-dlp/yt-dlp#readme)
|
||||||
required: true
|
required: true
|
||||||
- label: I've verified that I'm running yt-dlp version **%(version)s** ([update instructions](https://github.com/yt-dlp/yt-dlp#update)) or later (specify commit)
|
- label: I've verified that I have **updated yt-dlp to nightly or master** ([update instructions](https://github.com/yt-dlp/yt-dlp#update-channels))
|
||||||
required: true
|
required: true
|
||||||
- label: I've searched [known issues](https://github.com/yt-dlp/yt-dlp/issues/3766) and the [bugtracker](https://github.com/yt-dlp/yt-dlp/issues?q=) for similar questions **including closed ones**. DO NOT post duplicates
|
- label: I've searched [known issues](https://github.com/yt-dlp/yt-dlp/issues/3766) and the [bugtracker](https://github.com/yt-dlp/yt-dlp/issues?q=) for similar questions **including closed ones**. DO NOT post duplicates
|
||||||
required: true
|
required: true
|
||||||
|
|||||||
6
.github/PULL_REQUEST_TEMPLATE.md
vendored
6
.github/PULL_REQUEST_TEMPLATE.md
vendored
@@ -40,10 +40,4 @@ Fixes #
|
|||||||
- [ ] Core bug fix/improvement
|
- [ ] Core bug fix/improvement
|
||||||
- [ ] New feature (It is strongly [recommended to open an issue first](https://github.com/yt-dlp/yt-dlp/blob/master/CONTRIBUTING.md#adding-new-feature-or-making-overarching-changes))
|
- [ ] New feature (It is strongly [recommended to open an issue first](https://github.com/yt-dlp/yt-dlp/blob/master/CONTRIBUTING.md#adding-new-feature-or-making-overarching-changes))
|
||||||
|
|
||||||
|
|
||||||
<!-- Do NOT edit/remove anything below this! -->
|
|
||||||
</details><details><summary>Copilot Summary</summary>
|
|
||||||
|
|
||||||
copilot:all
|
|
||||||
|
|
||||||
</details>
|
</details>
|
||||||
|
|||||||
73
.github/workflows/build.yml
vendored
73
.github/workflows/build.yml
vendored
@@ -30,6 +30,10 @@ on:
|
|||||||
meta_files:
|
meta_files:
|
||||||
default: true
|
default: true
|
||||||
type: boolean
|
type: boolean
|
||||||
|
origin:
|
||||||
|
required: false
|
||||||
|
default: ''
|
||||||
|
type: string
|
||||||
secrets:
|
secrets:
|
||||||
GPG_SIGNING_KEY:
|
GPG_SIGNING_KEY:
|
||||||
required: false
|
required: false
|
||||||
@@ -37,11 +41,13 @@ on:
|
|||||||
workflow_dispatch:
|
workflow_dispatch:
|
||||||
inputs:
|
inputs:
|
||||||
version:
|
version:
|
||||||
description: Version tag (YYYY.MM.DD[.REV])
|
description: |
|
||||||
|
VERSION: yyyy.mm.dd[.rev] or rev
|
||||||
required: true
|
required: true
|
||||||
type: string
|
type: string
|
||||||
channel:
|
channel:
|
||||||
description: Update channel (stable/nightly/...)
|
description: |
|
||||||
|
SOURCE of this build's updates: stable/nightly/master/<repo>
|
||||||
required: true
|
required: true
|
||||||
default: stable
|
default: stable
|
||||||
type: string
|
type: string
|
||||||
@@ -73,16 +79,34 @@ on:
|
|||||||
description: SHA2-256SUMS, SHA2-512SUMS, _update_spec
|
description: SHA2-256SUMS, SHA2-512SUMS, _update_spec
|
||||||
default: true
|
default: true
|
||||||
type: boolean
|
type: boolean
|
||||||
|
origin:
|
||||||
|
description: .
|
||||||
|
required: false
|
||||||
|
default: ''
|
||||||
|
type: choice
|
||||||
|
options:
|
||||||
|
- ''
|
||||||
|
|
||||||
permissions:
|
permissions:
|
||||||
contents: read
|
contents: read
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
|
process:
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
outputs:
|
||||||
|
origin: ${{ steps.process_origin.outputs.origin }}
|
||||||
|
steps:
|
||||||
|
- name: Process origin
|
||||||
|
id: process_origin
|
||||||
|
run: |
|
||||||
|
echo "origin=${{ inputs.origin || github.repository }}" >> "$GITHUB_OUTPUT"
|
||||||
|
|
||||||
unix:
|
unix:
|
||||||
|
needs: process
|
||||||
if: inputs.unix
|
if: inputs.unix
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v4
|
||||||
- uses: actions/setup-python@v4
|
- uses: actions/setup-python@v4
|
||||||
with:
|
with:
|
||||||
python-version: "3.10"
|
python-version: "3.10"
|
||||||
@@ -96,22 +120,21 @@ jobs:
|
|||||||
auto-activate-base: false
|
auto-activate-base: false
|
||||||
- name: Install Requirements
|
- name: Install Requirements
|
||||||
run: |
|
run: |
|
||||||
sudo apt-get -y install zip pandoc man sed
|
sudo apt -y install zip pandoc man sed
|
||||||
python -m pip install -U pip setuptools wheel
|
|
||||||
python -m pip install -U Pyinstaller -r requirements.txt
|
|
||||||
reqs=$(mktemp)
|
reqs=$(mktemp)
|
||||||
cat > $reqs << EOF
|
cat > "$reqs" << EOF
|
||||||
python=3.10.*
|
python=3.10.*
|
||||||
pyinstaller
|
pyinstaller
|
||||||
cffi
|
cffi
|
||||||
brotli-python
|
brotli-python
|
||||||
|
secretstorage
|
||||||
EOF
|
EOF
|
||||||
sed '/^brotli.*/d' requirements.txt >> $reqs
|
sed -E '/^(brotli|secretstorage).*/d' requirements.txt >> "$reqs"
|
||||||
mamba create -n build --file $reqs
|
mamba create -n build --file "$reqs"
|
||||||
|
|
||||||
- name: Prepare
|
- name: Prepare
|
||||||
run: |
|
run: |
|
||||||
python devscripts/update-version.py -c ${{ inputs.channel }} ${{ inputs.version }}
|
python devscripts/update-version.py -c "${{ inputs.channel }}" -r "${{ needs.process.outputs.origin }}" "${{ inputs.version }}"
|
||||||
python devscripts/make_lazy_extractors.py
|
python devscripts/make_lazy_extractors.py
|
||||||
- name: Build Unix platform-independent binary
|
- name: Build Unix platform-independent binary
|
||||||
run: |
|
run: |
|
||||||
@@ -150,6 +173,7 @@ jobs:
|
|||||||
yt-dlp_linux.zip
|
yt-dlp_linux.zip
|
||||||
|
|
||||||
linux_arm:
|
linux_arm:
|
||||||
|
needs: process
|
||||||
if: inputs.linux_arm
|
if: inputs.linux_arm
|
||||||
permissions:
|
permissions:
|
||||||
contents: read
|
contents: read
|
||||||
@@ -162,7 +186,7 @@ jobs:
|
|||||||
- aarch64
|
- aarch64
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v4
|
||||||
with:
|
with:
|
||||||
path: ./repo
|
path: ./repo
|
||||||
- name: Virtualized Install, Prepare & Build
|
- name: Virtualized Install, Prepare & Build
|
||||||
@@ -180,12 +204,12 @@ jobs:
|
|||||||
apt -y install zlib1g-dev python3.8 python3.8-dev python3.8-distutils python3-pip
|
apt -y install zlib1g-dev python3.8 python3.8-dev python3.8-distutils python3-pip
|
||||||
python3.8 -m pip install -U pip setuptools wheel
|
python3.8 -m pip install -U pip setuptools wheel
|
||||||
# Cannot access requirements.txt from the repo directory at this stage
|
# Cannot access requirements.txt from the repo directory at this stage
|
||||||
python3.8 -m pip install -U Pyinstaller mutagen pycryptodomex websockets brotli certifi
|
python3.8 -m pip install -U Pyinstaller mutagen pycryptodomex websockets brotli certifi secretstorage
|
||||||
|
|
||||||
run: |
|
run: |
|
||||||
cd repo
|
cd repo
|
||||||
python3.8 -m pip install -U Pyinstaller -r requirements.txt # Cached version may be out of date
|
python3.8 -m pip install -U Pyinstaller secretstorage -r requirements.txt # Cached version may be out of date
|
||||||
python3.8 devscripts/update-version.py -c ${{ inputs.channel }} ${{ inputs.version }}
|
python3.8 devscripts/update-version.py -c "${{ inputs.channel }}" -r "${{ needs.process.outputs.origin }}" "${{ inputs.version }}"
|
||||||
python3.8 devscripts/make_lazy_extractors.py
|
python3.8 devscripts/make_lazy_extractors.py
|
||||||
python3.8 pyinst.py
|
python3.8 pyinst.py
|
||||||
|
|
||||||
@@ -206,11 +230,12 @@ jobs:
|
|||||||
repo/dist/yt-dlp_linux_${{ (matrix.architecture == 'armv7' && 'armv7l') || matrix.architecture }}
|
repo/dist/yt-dlp_linux_${{ (matrix.architecture == 'armv7' && 'armv7l') || matrix.architecture }}
|
||||||
|
|
||||||
macos:
|
macos:
|
||||||
|
needs: process
|
||||||
if: inputs.macos
|
if: inputs.macos
|
||||||
runs-on: macos-11
|
runs-on: macos-11
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v4
|
||||||
# NB: Building universal2 does not work with python from actions/setup-python
|
# NB: Building universal2 does not work with python from actions/setup-python
|
||||||
- name: Install Requirements
|
- name: Install Requirements
|
||||||
run: |
|
run: |
|
||||||
@@ -221,7 +246,7 @@ jobs:
|
|||||||
|
|
||||||
- name: Prepare
|
- name: Prepare
|
||||||
run: |
|
run: |
|
||||||
python3 devscripts/update-version.py -c ${{ inputs.channel }} ${{ inputs.version }}
|
python3 devscripts/update-version.py -c "${{ inputs.channel }}" -r "${{ needs.process.outputs.origin }}" "${{ inputs.version }}"
|
||||||
python3 devscripts/make_lazy_extractors.py
|
python3 devscripts/make_lazy_extractors.py
|
||||||
- name: Build
|
- name: Build
|
||||||
run: |
|
run: |
|
||||||
@@ -247,11 +272,12 @@ jobs:
|
|||||||
dist/yt-dlp_macos.zip
|
dist/yt-dlp_macos.zip
|
||||||
|
|
||||||
macos_legacy:
|
macos_legacy:
|
||||||
|
needs: process
|
||||||
if: inputs.macos_legacy
|
if: inputs.macos_legacy
|
||||||
runs-on: macos-latest
|
runs-on: macos-latest
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v4
|
||||||
- name: Install Python
|
- name: Install Python
|
||||||
# We need the official Python, because the GA ones only support newer macOS versions
|
# We need the official Python, because the GA ones only support newer macOS versions
|
||||||
env:
|
env:
|
||||||
@@ -272,7 +298,7 @@ jobs:
|
|||||||
|
|
||||||
- name: Prepare
|
- name: Prepare
|
||||||
run: |
|
run: |
|
||||||
python3 devscripts/update-version.py -c ${{ inputs.channel }} ${{ inputs.version }}
|
python3 devscripts/update-version.py -c "${{ inputs.channel }}" -r "${{ needs.process.outputs.origin }}" "${{ inputs.version }}"
|
||||||
python3 devscripts/make_lazy_extractors.py
|
python3 devscripts/make_lazy_extractors.py
|
||||||
- name: Build
|
- name: Build
|
||||||
run: |
|
run: |
|
||||||
@@ -296,11 +322,12 @@ jobs:
|
|||||||
dist/yt-dlp_macos_legacy
|
dist/yt-dlp_macos_legacy
|
||||||
|
|
||||||
windows:
|
windows:
|
||||||
|
needs: process
|
||||||
if: inputs.windows
|
if: inputs.windows
|
||||||
runs-on: windows-latest
|
runs-on: windows-latest
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v4
|
||||||
- uses: actions/setup-python@v4
|
- uses: actions/setup-python@v4
|
||||||
with: # 3.8 is used for Win7 support
|
with: # 3.8 is used for Win7 support
|
||||||
python-version: "3.8"
|
python-version: "3.8"
|
||||||
@@ -311,7 +338,7 @@ jobs:
|
|||||||
|
|
||||||
- name: Prepare
|
- name: Prepare
|
||||||
run: |
|
run: |
|
||||||
python devscripts/update-version.py -c ${{ inputs.channel }} ${{ inputs.version }}
|
python devscripts/update-version.py -c "${{ inputs.channel }}" -r "${{ needs.process.outputs.origin }}" "${{ inputs.version }}"
|
||||||
python devscripts/make_lazy_extractors.py
|
python devscripts/make_lazy_extractors.py
|
||||||
- name: Build
|
- name: Build
|
||||||
run: |
|
run: |
|
||||||
@@ -343,11 +370,12 @@ jobs:
|
|||||||
dist/yt-dlp_win.zip
|
dist/yt-dlp_win.zip
|
||||||
|
|
||||||
windows32:
|
windows32:
|
||||||
|
needs: process
|
||||||
if: inputs.windows32
|
if: inputs.windows32
|
||||||
runs-on: windows-latest
|
runs-on: windows-latest
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v4
|
||||||
- uses: actions/setup-python@v4
|
- uses: actions/setup-python@v4
|
||||||
with: # 3.7 is used for Vista support. See https://github.com/yt-dlp/yt-dlp/issues/390
|
with: # 3.7 is used for Vista support. See https://github.com/yt-dlp/yt-dlp/issues/390
|
||||||
python-version: "3.7"
|
python-version: "3.7"
|
||||||
@@ -359,7 +387,7 @@ jobs:
|
|||||||
|
|
||||||
- name: Prepare
|
- name: Prepare
|
||||||
run: |
|
run: |
|
||||||
python devscripts/update-version.py -c ${{ inputs.channel }} ${{ inputs.version }}
|
python devscripts/update-version.py -c "${{ inputs.channel }}" -r "${{ needs.process.outputs.origin }}" "${{ inputs.version }}"
|
||||||
python devscripts/make_lazy_extractors.py
|
python devscripts/make_lazy_extractors.py
|
||||||
- name: Build
|
- name: Build
|
||||||
run: |
|
run: |
|
||||||
@@ -387,6 +415,7 @@ jobs:
|
|||||||
meta_files:
|
meta_files:
|
||||||
if: inputs.meta_files && always() && !cancelled()
|
if: inputs.meta_files && always() && !cancelled()
|
||||||
needs:
|
needs:
|
||||||
|
- process
|
||||||
- unix
|
- unix
|
||||||
- linux_arm
|
- linux_arm
|
||||||
- macos
|
- macos
|
||||||
|
|||||||
2
.github/workflows/codeql.yml
vendored
2
.github/workflows/codeql.yml
vendored
@@ -29,7 +29,7 @@ jobs:
|
|||||||
|
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout repository
|
- name: Checkout repository
|
||||||
uses: actions/checkout@v3
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
# Initializes the CodeQL tools for scanning.
|
# Initializes the CodeQL tools for scanning.
|
||||||
- name: Initialize CodeQL
|
- name: Initialize CodeQL
|
||||||
|
|||||||
6
.github/workflows/core.yml
vendored
6
.github/workflows/core.yml
vendored
@@ -27,13 +27,13 @@ jobs:
|
|||||||
python-version: pypy-3.9
|
python-version: pypy-3.9
|
||||||
run-tests-ext: bat
|
run-tests-ext: bat
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v4
|
||||||
- name: Set up Python ${{ matrix.python-version }}
|
- name: Set up Python ${{ matrix.python-version }}
|
||||||
uses: actions/setup-python@v4
|
uses: actions/setup-python@v4
|
||||||
with:
|
with:
|
||||||
python-version: ${{ matrix.python-version }}
|
python-version: ${{ matrix.python-version }}
|
||||||
- name: Install pytest
|
- name: Install dependencies
|
||||||
run: pip install pytest
|
run: pip install pytest -r requirements.txt
|
||||||
- name: Run tests
|
- name: Run tests
|
||||||
continue-on-error: False
|
continue-on-error: False
|
||||||
run: |
|
run: |
|
||||||
|
|||||||
4
.github/workflows/download.yml
vendored
4
.github/workflows/download.yml
vendored
@@ -9,7 +9,7 @@ jobs:
|
|||||||
if: "contains(github.event.head_commit.message, 'ci run dl')"
|
if: "contains(github.event.head_commit.message, 'ci run dl')"
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v4
|
||||||
- name: Set up Python
|
- name: Set up Python
|
||||||
uses: actions/setup-python@v4
|
uses: actions/setup-python@v4
|
||||||
with:
|
with:
|
||||||
@@ -39,7 +39,7 @@ jobs:
|
|||||||
python-version: pypy-3.9
|
python-version: pypy-3.9
|
||||||
run-tests-ext: bat
|
run-tests-ext: bat
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v4
|
||||||
- name: Set up Python ${{ matrix.python-version }}
|
- name: Set up Python ${{ matrix.python-version }}
|
||||||
uses: actions/setup-python@v4
|
uses: actions/setup-python@v4
|
||||||
with:
|
with:
|
||||||
|
|||||||
97
.github/workflows/publish.yml
vendored
97
.github/workflows/publish.yml
vendored
@@ -1,97 +0,0 @@
|
|||||||
name: Publish
|
|
||||||
on:
|
|
||||||
workflow_call:
|
|
||||||
inputs:
|
|
||||||
channel:
|
|
||||||
default: stable
|
|
||||||
required: true
|
|
||||||
type: string
|
|
||||||
version:
|
|
||||||
required: true
|
|
||||||
type: string
|
|
||||||
target_commitish:
|
|
||||||
required: true
|
|
||||||
type: string
|
|
||||||
prerelease:
|
|
||||||
default: false
|
|
||||||
required: true
|
|
||||||
type: boolean
|
|
||||||
secrets:
|
|
||||||
ARCHIVE_REPO_TOKEN:
|
|
||||||
required: false
|
|
||||||
|
|
||||||
permissions:
|
|
||||||
contents: write
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
publish:
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@v3
|
|
||||||
with:
|
|
||||||
fetch-depth: 0
|
|
||||||
- uses: actions/download-artifact@v3
|
|
||||||
- uses: actions/setup-python@v4
|
|
||||||
with:
|
|
||||||
python-version: "3.10"
|
|
||||||
|
|
||||||
- name: Generate release notes
|
|
||||||
run: |
|
|
||||||
printf '%s' \
|
|
||||||
'[]' \
|
|
||||||
'(https://github.com/yt-dlp/yt-dlp#installation "Installation instructions") ' \
|
|
||||||
'[]' \
|
|
||||||
'(https://github.com/yt-dlp/yt-dlp/tree/2023.03.04#readme "Documentation") ' \
|
|
||||||
'[]' \
|
|
||||||
'(https://github.com/yt-dlp/yt-dlp/blob/master/Collaborators.md#collaborators "Donate") ' \
|
|
||||||
'[]' \
|
|
||||||
'(https://discord.gg/H5MNcFW63r "Discord") ' \
|
|
||||||
${{ inputs.channel != 'nightly' && '"[]" \
|
|
||||||
"(https://github.com/yt-dlp/yt-dlp-nightly-builds/releases/latest \"Nightly builds\")"' || '' }} \
|
|
||||||
> ./RELEASE_NOTES
|
|
||||||
printf '\n\n' >> ./RELEASE_NOTES
|
|
||||||
cat >> ./RELEASE_NOTES << EOF
|
|
||||||
#### A description of the various files are in the [README](https://github.com/yt-dlp/yt-dlp#release-files)
|
|
||||||
---
|
|
||||||
$(python ./devscripts/make_changelog.py -vv --collapsible)
|
|
||||||
EOF
|
|
||||||
printf '%s\n\n' '**This is an automated nightly pre-release build**' >> ./NIGHTLY_NOTES
|
|
||||||
cat ./RELEASE_NOTES >> ./NIGHTLY_NOTES
|
|
||||||
printf '%s\n\n' 'Generated from: https://github.com/${{ github.repository }}/commit/${{ inputs.target_commitish }}' >> ./ARCHIVE_NOTES
|
|
||||||
cat ./RELEASE_NOTES >> ./ARCHIVE_NOTES
|
|
||||||
|
|
||||||
- name: Archive nightly release
|
|
||||||
env:
|
|
||||||
GH_TOKEN: ${{ secrets.ARCHIVE_REPO_TOKEN }}
|
|
||||||
GH_REPO: ${{ vars.ARCHIVE_REPO }}
|
|
||||||
if: |
|
|
||||||
inputs.channel == 'nightly' && env.GH_TOKEN != '' && env.GH_REPO != ''
|
|
||||||
run: |
|
|
||||||
gh release create \
|
|
||||||
--notes-file ARCHIVE_NOTES \
|
|
||||||
--title "yt-dlp nightly ${{ inputs.version }}" \
|
|
||||||
${{ inputs.version }} \
|
|
||||||
artifact/*
|
|
||||||
|
|
||||||
- name: Prune old nightly release
|
|
||||||
if: inputs.channel == 'nightly' && !vars.ARCHIVE_REPO
|
|
||||||
env:
|
|
||||||
GH_TOKEN: ${{ github.token }}
|
|
||||||
run: |
|
|
||||||
gh release delete --yes --cleanup-tag "nightly" || true
|
|
||||||
git tag --delete "nightly" || true
|
|
||||||
sleep 5 # Enough time to cover deletion race condition
|
|
||||||
|
|
||||||
- name: Publish release${{ inputs.channel == 'nightly' && ' (nightly)' || '' }}
|
|
||||||
env:
|
|
||||||
GH_TOKEN: ${{ github.token }}
|
|
||||||
if: (inputs.channel == 'nightly' && !vars.ARCHIVE_REPO) || inputs.channel != 'nightly'
|
|
||||||
run: |
|
|
||||||
gh release create \
|
|
||||||
--notes-file ${{ inputs.channel == 'nightly' && 'NIGHTLY_NOTES' || 'RELEASE_NOTES' }} \
|
|
||||||
--target ${{ inputs.target_commitish }} \
|
|
||||||
--title "yt-dlp ${{ inputs.channel == 'nightly' && 'nightly ' || '' }}${{ inputs.version }}" \
|
|
||||||
${{ inputs.prerelease && '--prerelease' || '' }} \
|
|
||||||
${{ inputs.channel == 'nightly' && '"nightly"' || inputs.version }} \
|
|
||||||
artifact/*
|
|
||||||
4
.github/workflows/quick-test.yml
vendored
4
.github/workflows/quick-test.yml
vendored
@@ -9,7 +9,7 @@ jobs:
|
|||||||
if: "!contains(github.event.head_commit.message, 'ci skip all')"
|
if: "!contains(github.event.head_commit.message, 'ci skip all')"
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v4
|
||||||
- name: Set up Python 3.11
|
- name: Set up Python 3.11
|
||||||
uses: actions/setup-python@v4
|
uses: actions/setup-python@v4
|
||||||
with:
|
with:
|
||||||
@@ -25,7 +25,7 @@ jobs:
|
|||||||
if: "!contains(github.event.head_commit.message, 'ci skip all')"
|
if: "!contains(github.event.head_commit.message, 'ci skip all')"
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v4
|
||||||
- uses: actions/setup-python@v4
|
- uses: actions/setup-python@v4
|
||||||
- name: Install flake8
|
- name: Install flake8
|
||||||
run: pip install flake8
|
run: pip install flake8
|
||||||
|
|||||||
28
.github/workflows/release-master.yml
vendored
Normal file
28
.github/workflows/release-master.yml
vendored
Normal file
@@ -0,0 +1,28 @@
|
|||||||
|
name: Release (master)
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
branches:
|
||||||
|
- master
|
||||||
|
paths:
|
||||||
|
- "yt_dlp/**.py"
|
||||||
|
- "!yt_dlp/version.py"
|
||||||
|
- "setup.py"
|
||||||
|
- "pyinst.py"
|
||||||
|
concurrency:
|
||||||
|
group: release-master
|
||||||
|
cancel-in-progress: true
|
||||||
|
permissions:
|
||||||
|
contents: read
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
release:
|
||||||
|
if: vars.BUILD_MASTER != ''
|
||||||
|
uses: ./.github/workflows/release.yml
|
||||||
|
with:
|
||||||
|
prerelease: true
|
||||||
|
source: master
|
||||||
|
permissions:
|
||||||
|
contents: write
|
||||||
|
packages: write
|
||||||
|
id-token: write # mandatory for trusted publishing
|
||||||
|
secrets: inherit
|
||||||
57
.github/workflows/release-nightly.yml
vendored
57
.github/workflows/release-nightly.yml
vendored
@@ -1,52 +1,35 @@
|
|||||||
name: Release (nightly)
|
name: Release (nightly)
|
||||||
on:
|
on:
|
||||||
push:
|
schedule:
|
||||||
branches:
|
- cron: '23 23 * * *'
|
||||||
- master
|
|
||||||
paths:
|
|
||||||
- "yt_dlp/**.py"
|
|
||||||
- "!yt_dlp/version.py"
|
|
||||||
concurrency:
|
|
||||||
group: release-nightly
|
|
||||||
cancel-in-progress: true
|
|
||||||
permissions:
|
permissions:
|
||||||
contents: read
|
contents: read
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
prepare:
|
check_nightly:
|
||||||
if: vars.BUILD_NIGHTLY != ''
|
if: vars.BUILD_NIGHTLY != ''
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
outputs:
|
outputs:
|
||||||
version: ${{ steps.get_version.outputs.version }}
|
commit: ${{ steps.check_for_new_commits.outputs.commit }}
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v4
|
||||||
- name: Get version
|
with:
|
||||||
id: get_version
|
fetch-depth: 0
|
||||||
|
- name: Check for new commits
|
||||||
|
id: check_for_new_commits
|
||||||
run: |
|
run: |
|
||||||
python devscripts/update-version.py "$(date -u +"%H%M%S")" | grep -Po "version=\d+(\.\d+){3}" >> "$GITHUB_OUTPUT"
|
relevant_files=("yt_dlp/*.py" ':!yt_dlp/version.py' "setup.py" "pyinst.py")
|
||||||
|
echo "commit=$(git log --format=%H -1 --since="24 hours ago" -- "${relevant_files[@]}")" | tee "$GITHUB_OUTPUT"
|
||||||
|
|
||||||
build:
|
release:
|
||||||
needs: prepare
|
needs: [check_nightly]
|
||||||
uses: ./.github/workflows/build.yml
|
if: ${{ needs.check_nightly.outputs.commit }}
|
||||||
|
uses: ./.github/workflows/release.yml
|
||||||
with:
|
with:
|
||||||
version: ${{ needs.prepare.outputs.version }}
|
prerelease: true
|
||||||
channel: nightly
|
source: nightly
|
||||||
permissions:
|
|
||||||
contents: read
|
|
||||||
packages: write # For package cache
|
|
||||||
secrets:
|
|
||||||
GPG_SIGNING_KEY: ${{ secrets.GPG_SIGNING_KEY }}
|
|
||||||
|
|
||||||
publish:
|
|
||||||
needs: [prepare, build]
|
|
||||||
uses: ./.github/workflows/publish.yml
|
|
||||||
secrets:
|
|
||||||
ARCHIVE_REPO_TOKEN: ${{ secrets.ARCHIVE_REPO_TOKEN }}
|
|
||||||
permissions:
|
permissions:
|
||||||
contents: write
|
contents: write
|
||||||
with:
|
packages: write
|
||||||
channel: nightly
|
id-token: write # mandatory for trusted publishing
|
||||||
prerelease: true
|
secrets: inherit
|
||||||
version: ${{ needs.prepare.outputs.version }}
|
|
||||||
target_commitish: ${{ github.sha }}
|
|
||||||
|
|||||||
360
.github/workflows/release.yml
vendored
360
.github/workflows/release.yml
vendored
@@ -1,14 +1,45 @@
|
|||||||
name: Release
|
name: Release
|
||||||
on:
|
on:
|
||||||
workflow_dispatch:
|
workflow_call:
|
||||||
inputs:
|
inputs:
|
||||||
version:
|
prerelease:
|
||||||
description: Version tag (YYYY.MM.DD[.REV])
|
required: false
|
||||||
|
default: true
|
||||||
|
type: boolean
|
||||||
|
source:
|
||||||
required: false
|
required: false
|
||||||
default: ''
|
default: ''
|
||||||
type: string
|
type: string
|
||||||
channel:
|
target:
|
||||||
description: Update channel (stable/nightly/...)
|
required: false
|
||||||
|
default: ''
|
||||||
|
type: string
|
||||||
|
version:
|
||||||
|
required: false
|
||||||
|
default: ''
|
||||||
|
type: string
|
||||||
|
workflow_dispatch:
|
||||||
|
inputs:
|
||||||
|
source:
|
||||||
|
description: |
|
||||||
|
SOURCE of this release's updates:
|
||||||
|
channel, repo, tag, or channel/repo@tag
|
||||||
|
(default: <current_repo>)
|
||||||
|
required: false
|
||||||
|
default: ''
|
||||||
|
type: string
|
||||||
|
target:
|
||||||
|
description: |
|
||||||
|
TARGET to publish this release to:
|
||||||
|
channel, tag, or channel@tag
|
||||||
|
(default: <source> if writable else <current_repo>[@source_tag])
|
||||||
|
required: false
|
||||||
|
default: ''
|
||||||
|
type: string
|
||||||
|
version:
|
||||||
|
description: |
|
||||||
|
VERSION: yyyy.mm.dd[.rev] or rev
|
||||||
|
(default: auto-generated)
|
||||||
required: false
|
required: false
|
||||||
default: ''
|
default: ''
|
||||||
type: string
|
type: string
|
||||||
@@ -26,12 +57,18 @@ jobs:
|
|||||||
contents: write
|
contents: write
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
outputs:
|
outputs:
|
||||||
channel: ${{ steps.set_channel.outputs.channel }}
|
channel: ${{ steps.setup_variables.outputs.channel }}
|
||||||
version: ${{ steps.update_version.outputs.version }}
|
version: ${{ steps.setup_variables.outputs.version }}
|
||||||
|
target_repo: ${{ steps.setup_variables.outputs.target_repo }}
|
||||||
|
target_repo_token: ${{ steps.setup_variables.outputs.target_repo_token }}
|
||||||
|
target_tag: ${{ steps.setup_variables.outputs.target_tag }}
|
||||||
|
pypi_project: ${{ steps.setup_variables.outputs.pypi_project }}
|
||||||
|
pypi_suffix: ${{ steps.setup_variables.outputs.pypi_suffix }}
|
||||||
|
pypi_token: ${{ steps.setup_variables.outputs.pypi_token }}
|
||||||
head_sha: ${{ steps.get_target.outputs.head_sha }}
|
head_sha: ${{ steps.get_target.outputs.head_sha }}
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v4
|
||||||
with:
|
with:
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
|
|
||||||
@@ -39,25 +76,133 @@ jobs:
|
|||||||
with:
|
with:
|
||||||
python-version: "3.10"
|
python-version: "3.10"
|
||||||
|
|
||||||
- name: Set channel
|
- name: Process inputs
|
||||||
id: set_channel
|
id: process_inputs
|
||||||
run: |
|
run: |
|
||||||
CHANNEL="${{ github.repository == 'yt-dlp/yt-dlp' && 'stable' || github.repository }}"
|
cat << EOF
|
||||||
echo "channel=${{ inputs.channel || '$CHANNEL' }}" > "$GITHUB_OUTPUT"
|
::group::Inputs
|
||||||
|
prerelease=${{ inputs.prerelease }}
|
||||||
|
source=${{ inputs.source }}
|
||||||
|
target=${{ inputs.target }}
|
||||||
|
version=${{ inputs.version }}
|
||||||
|
::endgroup::
|
||||||
|
EOF
|
||||||
|
IFS='@' read -r source_repo source_tag <<<"${{ inputs.source }}"
|
||||||
|
IFS='@' read -r target_repo target_tag <<<"${{ inputs.target }}"
|
||||||
|
cat << EOF >> "$GITHUB_OUTPUT"
|
||||||
|
source_repo=${source_repo}
|
||||||
|
source_tag=${source_tag}
|
||||||
|
target_repo=${target_repo}
|
||||||
|
target_tag=${target_tag}
|
||||||
|
EOF
|
||||||
|
|
||||||
- name: Update version
|
- name: Setup variables
|
||||||
id: update_version
|
id: setup_variables
|
||||||
|
env:
|
||||||
|
source_repo: ${{ steps.process_inputs.outputs.source_repo }}
|
||||||
|
source_tag: ${{ steps.process_inputs.outputs.source_tag }}
|
||||||
|
target_repo: ${{ steps.process_inputs.outputs.target_repo }}
|
||||||
|
target_tag: ${{ steps.process_inputs.outputs.target_tag }}
|
||||||
run: |
|
run: |
|
||||||
REVISION="${{ vars.PUSH_VERSION_COMMIT == '' && '$(date -u +"%H%M%S")' || '' }}"
|
# unholy bash monstrosity (sincere apologies)
|
||||||
REVISION="${{ inputs.prerelease && '$(date -u +"%H%M%S")' || '$REVISION' }}"
|
fallback_token () {
|
||||||
python devscripts/update-version.py ${{ inputs.version || '$REVISION' }} | \
|
if ${{ !secrets.ARCHIVE_REPO_TOKEN }}; then
|
||||||
grep -Po "version=\d+\.\d+\.\d+(\.\d+)?" >> "$GITHUB_OUTPUT"
|
echo "::error::Repository access secret ${target_repo_token^^} not found"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
target_repo_token=ARCHIVE_REPO_TOKEN
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
|
||||||
|
source_is_channel=0
|
||||||
|
[[ "${source_repo}" == 'stable' ]] && source_repo='yt-dlp/yt-dlp'
|
||||||
|
if [[ -z "${source_repo}" ]]; then
|
||||||
|
source_repo='${{ github.repository }}'
|
||||||
|
elif [[ '${{ vars[format('{0}_archive_repo', env.source_repo)] }}' ]]; then
|
||||||
|
source_is_channel=1
|
||||||
|
source_channel='${{ vars[format('{0}_archive_repo', env.source_repo)] }}'
|
||||||
|
elif [[ -z "${source_tag}" && "${source_repo}" != */* ]]; then
|
||||||
|
source_tag="${source_repo}"
|
||||||
|
source_repo='${{ github.repository }}'
|
||||||
|
fi
|
||||||
|
resolved_source="${source_repo}"
|
||||||
|
if [[ "${source_tag}" ]]; then
|
||||||
|
resolved_source="${resolved_source}@${source_tag}"
|
||||||
|
elif [[ "${source_repo}" == 'yt-dlp/yt-dlp' ]]; then
|
||||||
|
resolved_source='stable'
|
||||||
|
fi
|
||||||
|
|
||||||
|
revision="${{ (inputs.prerelease || !vars.PUSH_VERSION_COMMIT) && '$(date -u +"%H%M%S")' || '' }}"
|
||||||
|
version="$(
|
||||||
|
python devscripts/update-version.py \
|
||||||
|
-c "${resolved_source}" -r "${{ github.repository }}" ${{ inputs.version || '$revision' }} | \
|
||||||
|
grep -Po "version=\K\d+\.\d+\.\d+(\.\d+)?")"
|
||||||
|
|
||||||
|
if [[ "${target_repo}" ]]; then
|
||||||
|
if [[ -z "${target_tag}" ]]; then
|
||||||
|
if [[ '${{ vars[format('{0}_archive_repo', env.target_repo)] }}' ]]; then
|
||||||
|
target_tag="${source_tag:-${version}}"
|
||||||
|
else
|
||||||
|
target_tag="${target_repo}"
|
||||||
|
target_repo='${{ github.repository }}'
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
if [[ "${target_repo}" != '${{ github.repository}}' ]]; then
|
||||||
|
target_repo='${{ vars[format('{0}_archive_repo', env.target_repo)] }}'
|
||||||
|
target_repo_token='${{ env.target_repo }}_archive_repo_token'
|
||||||
|
${{ !!secrets[format('{0}_archive_repo_token', env.target_repo)] }} || fallback_token
|
||||||
|
pypi_project='${{ vars[format('{0}_pypi_project', env.target_repo)] }}'
|
||||||
|
pypi_suffix='${{ vars[format('{0}_pypi_suffix', env.target_repo)] }}'
|
||||||
|
${{ !secrets[format('{0}_pypi_token', env.target_repo)] }} || pypi_token='${{ env.target_repo }}_pypi_token'
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
target_tag="${source_tag:-${version}}"
|
||||||
|
if ((source_is_channel)); then
|
||||||
|
target_repo="${source_channel}"
|
||||||
|
target_repo_token='${{ env.source_repo }}_archive_repo_token'
|
||||||
|
${{ !!secrets[format('{0}_archive_repo_token', env.source_repo)] }} || fallback_token
|
||||||
|
pypi_project='${{ vars[format('{0}_pypi_project', env.source_repo)] }}'
|
||||||
|
pypi_suffix='${{ vars[format('{0}_pypi_suffix', env.source_repo)] }}'
|
||||||
|
${{ !secrets[format('{0}_pypi_token', env.source_repo)] }} || pypi_token='${{ env.source_repo }}_pypi_token'
|
||||||
|
else
|
||||||
|
target_repo='${{ github.repository }}'
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [[ "${target_repo}" == '${{ github.repository }}' ]] && ${{ !inputs.prerelease }}; then
|
||||||
|
pypi_project='${{ vars.PYPI_PROJECT }}'
|
||||||
|
fi
|
||||||
|
if [[ -z "${pypi_token}" && "${pypi_project}" ]]; then
|
||||||
|
if ${{ !secrets.PYPI_TOKEN }}; then
|
||||||
|
pypi_token=OIDC
|
||||||
|
else
|
||||||
|
pypi_token=PYPI_TOKEN
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
|
echo "::group::Output variables"
|
||||||
|
cat << EOF | tee -a "$GITHUB_OUTPUT"
|
||||||
|
channel=${resolved_source}
|
||||||
|
version=${version}
|
||||||
|
target_repo=${target_repo}
|
||||||
|
target_repo_token=${target_repo_token}
|
||||||
|
target_tag=${target_tag}
|
||||||
|
pypi_project=${pypi_project}
|
||||||
|
pypi_suffix=${pypi_suffix}
|
||||||
|
pypi_token=${pypi_token}
|
||||||
|
EOF
|
||||||
|
echo "::endgroup::"
|
||||||
|
|
||||||
- name: Update documentation
|
- name: Update documentation
|
||||||
|
env:
|
||||||
|
version: ${{ steps.setup_variables.outputs.version }}
|
||||||
|
target_repo: ${{ steps.setup_variables.outputs.target_repo }}
|
||||||
|
if: |
|
||||||
|
!inputs.prerelease && env.target_repo == github.repository
|
||||||
run: |
|
run: |
|
||||||
make doc
|
make doc
|
||||||
sed '/### /Q' Changelog.md >> ./CHANGELOG
|
sed '/### /Q' Changelog.md >> ./CHANGELOG
|
||||||
echo '### ${{ steps.update_version.outputs.version }}' >> ./CHANGELOG
|
echo '### ${{ env.version }}' >> ./CHANGELOG
|
||||||
python ./devscripts/make_changelog.py -vv -c >> ./CHANGELOG
|
python ./devscripts/make_changelog.py -vv -c >> ./CHANGELOG
|
||||||
echo >> ./CHANGELOG
|
echo >> ./CHANGELOG
|
||||||
grep -Poz '(?s)### \d+\.\d+\.\d+.+' 'Changelog.md' | head -n -1 >> ./CHANGELOG
|
grep -Poz '(?s)### \d+\.\d+\.\d+.+' 'Changelog.md' | head -n -1 >> ./CHANGELOG
|
||||||
@@ -65,12 +210,16 @@ jobs:
|
|||||||
|
|
||||||
- name: Push to release
|
- name: Push to release
|
||||||
id: push_release
|
id: push_release
|
||||||
if: ${{ !inputs.prerelease }}
|
env:
|
||||||
|
version: ${{ steps.setup_variables.outputs.version }}
|
||||||
|
target_repo: ${{ steps.setup_variables.outputs.target_repo }}
|
||||||
|
if: |
|
||||||
|
!inputs.prerelease && env.target_repo == github.repository
|
||||||
run: |
|
run: |
|
||||||
git config --global user.name github-actions
|
git config --global user.name "github-actions[bot]"
|
||||||
git config --global user.email github-actions@example.com
|
git config --global user.email "41898282+github-actions[bot]@users.noreply.github.com"
|
||||||
git add -u
|
git add -u
|
||||||
git commit -m "Release ${{ steps.update_version.outputs.version }}" \
|
git commit -m "Release ${{ env.version }}" \
|
||||||
-m "Created by: ${{ github.event.sender.login }}" -m ":ci skip all :ci run dl"
|
-m "Created by: ${{ github.event.sender.login }}" -m ":ci skip all :ci run dl"
|
||||||
git push origin --force ${{ github.event.ref }}:release
|
git push origin --force ${{ github.event.ref }}:release
|
||||||
|
|
||||||
@@ -80,7 +229,10 @@ jobs:
|
|||||||
echo "head_sha=$(git rev-parse HEAD)" >> "$GITHUB_OUTPUT"
|
echo "head_sha=$(git rev-parse HEAD)" >> "$GITHUB_OUTPUT"
|
||||||
|
|
||||||
- name: Update master
|
- name: Update master
|
||||||
if: vars.PUSH_VERSION_COMMIT != '' && !inputs.prerelease
|
env:
|
||||||
|
target_repo: ${{ steps.setup_variables.outputs.target_repo }}
|
||||||
|
if: |
|
||||||
|
vars.PUSH_VERSION_COMMIT != '' && !inputs.prerelease && env.target_repo == github.repository
|
||||||
run: git push origin ${{ github.event.ref }}
|
run: git push origin ${{ github.event.ref }}
|
||||||
|
|
||||||
build:
|
build:
|
||||||
@@ -89,75 +241,159 @@ jobs:
|
|||||||
with:
|
with:
|
||||||
version: ${{ needs.prepare.outputs.version }}
|
version: ${{ needs.prepare.outputs.version }}
|
||||||
channel: ${{ needs.prepare.outputs.channel }}
|
channel: ${{ needs.prepare.outputs.channel }}
|
||||||
|
origin: ${{ needs.prepare.outputs.target_repo }}
|
||||||
permissions:
|
permissions:
|
||||||
contents: read
|
contents: read
|
||||||
packages: write # For package cache
|
packages: write # For package cache
|
||||||
secrets:
|
secrets:
|
||||||
GPG_SIGNING_KEY: ${{ secrets.GPG_SIGNING_KEY }}
|
GPG_SIGNING_KEY: ${{ secrets.GPG_SIGNING_KEY }}
|
||||||
|
|
||||||
publish_pypi_homebrew:
|
publish_pypi:
|
||||||
needs: [prepare, build]
|
needs: [prepare, build]
|
||||||
|
if: ${{ needs.prepare.outputs.pypi_project }}
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
|
permissions:
|
||||||
|
id-token: write # mandatory for trusted publishing
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v4
|
||||||
- uses: actions/setup-python@v4
|
- uses: actions/setup-python@v4
|
||||||
with:
|
with:
|
||||||
python-version: "3.10"
|
python-version: "3.10"
|
||||||
|
|
||||||
- name: Install Requirements
|
- name: Install Requirements
|
||||||
run: |
|
run: |
|
||||||
sudo apt-get -y install pandoc man
|
sudo apt -y install pandoc man
|
||||||
python -m pip install -U pip setuptools wheel twine
|
python -m pip install -U pip setuptools wheel twine
|
||||||
python -m pip install -U -r requirements.txt
|
python -m pip install -U -r requirements.txt
|
||||||
|
|
||||||
- name: Prepare
|
- name: Prepare
|
||||||
run: |
|
|
||||||
python devscripts/update-version.py ${{ needs.prepare.outputs.version }}
|
|
||||||
python devscripts/make_lazy_extractors.py
|
|
||||||
|
|
||||||
- name: Build and publish on PyPI
|
|
||||||
env:
|
env:
|
||||||
TWINE_USERNAME: __token__
|
version: ${{ needs.prepare.outputs.version }}
|
||||||
TWINE_PASSWORD: ${{ secrets.PYPI_TOKEN }}
|
suffix: ${{ needs.prepare.outputs.pypi_suffix }}
|
||||||
if: env.TWINE_PASSWORD != '' && !inputs.prerelease
|
channel: ${{ needs.prepare.outputs.channel }}
|
||||||
|
target_repo: ${{ needs.prepare.outputs.target_repo }}
|
||||||
|
pypi_project: ${{ needs.prepare.outputs.pypi_project }}
|
||||||
|
run: |
|
||||||
|
python devscripts/update-version.py -c "${{ env.channel }}" -r "${{ env.target_repo }}" -s "${{ env.suffix }}" "${{ env.version }}"
|
||||||
|
python devscripts/make_lazy_extractors.py
|
||||||
|
sed -i -E "s/(name=')[^']+(', # package name)/\1${{ env.pypi_project }}\2/" setup.py
|
||||||
|
|
||||||
|
- name: Build
|
||||||
run: |
|
run: |
|
||||||
rm -rf dist/*
|
rm -rf dist/*
|
||||||
make pypi-files
|
make pypi-files
|
||||||
python devscripts/set-variant.py pip -M "You installed yt-dlp with pip or using the wheel from PyPi; Use that to update"
|
python devscripts/set-variant.py pip -M "You installed yt-dlp with pip or using the wheel from PyPi; Use that to update"
|
||||||
python setup.py sdist bdist_wheel
|
python setup.py sdist bdist_wheel
|
||||||
|
|
||||||
|
- name: Publish to PyPI via token
|
||||||
|
env:
|
||||||
|
TWINE_USERNAME: __token__
|
||||||
|
TWINE_PASSWORD: ${{ secrets[needs.prepare.outputs.pypi_token] }}
|
||||||
|
if: |
|
||||||
|
needs.prepare.outputs.pypi_token != 'OIDC' && env.TWINE_PASSWORD
|
||||||
|
run: |
|
||||||
twine upload dist/*
|
twine upload dist/*
|
||||||
|
|
||||||
- name: Checkout Homebrew repository
|
- name: Publish to PyPI via trusted publishing
|
||||||
env:
|
if: |
|
||||||
BREW_TOKEN: ${{ secrets.BREW_TOKEN }}
|
needs.prepare.outputs.pypi_token == 'OIDC'
|
||||||
PYPI_TOKEN: ${{ secrets.PYPI_TOKEN }}
|
uses: pypa/gh-action-pypi-publish@release/v1
|
||||||
if: env.BREW_TOKEN != '' && env.PYPI_TOKEN != '' && !inputs.prerelease
|
|
||||||
uses: actions/checkout@v3
|
|
||||||
with:
|
with:
|
||||||
repository: yt-dlp/homebrew-taps
|
verbose: true
|
||||||
path: taps
|
|
||||||
ssh-key: ${{ secrets.BREW_TOKEN }}
|
|
||||||
|
|
||||||
- name: Update Homebrew Formulae
|
|
||||||
env:
|
|
||||||
BREW_TOKEN: ${{ secrets.BREW_TOKEN }}
|
|
||||||
PYPI_TOKEN: ${{ secrets.PYPI_TOKEN }}
|
|
||||||
if: env.BREW_TOKEN != '' && env.PYPI_TOKEN != '' && !inputs.prerelease
|
|
||||||
run: |
|
|
||||||
python devscripts/update-formulae.py taps/Formula/yt-dlp.rb "${{ needs.prepare.outputs.version }}"
|
|
||||||
git -C taps/ config user.name github-actions
|
|
||||||
git -C taps/ config user.email github-actions@example.com
|
|
||||||
git -C taps/ commit -am 'yt-dlp: ${{ needs.prepare.outputs.version }}'
|
|
||||||
git -C taps/ push
|
|
||||||
|
|
||||||
publish:
|
publish:
|
||||||
needs: [prepare, build]
|
needs: [prepare, build]
|
||||||
uses: ./.github/workflows/publish.yml
|
|
||||||
permissions:
|
permissions:
|
||||||
contents: write
|
contents: write
|
||||||
with:
|
runs-on: ubuntu-latest
|
||||||
channel: ${{ needs.prepare.outputs.channel }}
|
|
||||||
prerelease: ${{ inputs.prerelease }}
|
steps:
|
||||||
version: ${{ needs.prepare.outputs.version }}
|
- uses: actions/checkout@v4
|
||||||
target_commitish: ${{ needs.prepare.outputs.head_sha }}
|
with:
|
||||||
|
fetch-depth: 0
|
||||||
|
- uses: actions/download-artifact@v3
|
||||||
|
- uses: actions/setup-python@v4
|
||||||
|
with:
|
||||||
|
python-version: "3.10"
|
||||||
|
|
||||||
|
- name: Generate release notes
|
||||||
|
env:
|
||||||
|
head_sha: ${{ needs.prepare.outputs.head_sha }}
|
||||||
|
target_repo: ${{ needs.prepare.outputs.target_repo }}
|
||||||
|
target_tag: ${{ needs.prepare.outputs.target_tag }}
|
||||||
|
run: |
|
||||||
|
printf '%s' \
|
||||||
|
'[]' \
|
||||||
|
'(https://github.com/${{ github.repository }}#installation "Installation instructions") ' \
|
||||||
|
'[]' \
|
||||||
|
'(https://github.com/${{ github.repository }}' \
|
||||||
|
'${{ env.target_repo == github.repository && format('/tree/{0}', env.target_tag) || '' }}#readme "Documentation") ' \
|
||||||
|
'[]' \
|
||||||
|
'(https://github.com/yt-dlp/yt-dlp/blob/master/Collaborators.md#collaborators "Donate") ' \
|
||||||
|
'[]' \
|
||||||
|
'(https://discord.gg/H5MNcFW63r "Discord") ' \
|
||||||
|
${{ env.target_repo == 'yt-dlp/yt-dlp' && '\
|
||||||
|
"[]" \
|
||||||
|
"(https://github.com/yt-dlp/yt-dlp-nightly-builds/releases/latest \"Nightly builds\") " \
|
||||||
|
"[]" \
|
||||||
|
"(https://github.com/yt-dlp/yt-dlp-master-builds/releases/latest \"Master builds\")"' || '' }} > ./RELEASE_NOTES
|
||||||
|
printf '\n\n' >> ./RELEASE_NOTES
|
||||||
|
cat >> ./RELEASE_NOTES << EOF
|
||||||
|
#### A description of the various files are in the [README](https://github.com/${{ github.repository }}#release-files)
|
||||||
|
---
|
||||||
|
$(python ./devscripts/make_changelog.py -vv --collapsible)
|
||||||
|
EOF
|
||||||
|
printf '%s\n\n' '**This is a pre-release build**' >> ./PRERELEASE_NOTES
|
||||||
|
cat ./RELEASE_NOTES >> ./PRERELEASE_NOTES
|
||||||
|
printf '%s\n\n' 'Generated from: https://github.com/${{ github.repository }}/commit/${{ env.head_sha }}' >> ./ARCHIVE_NOTES
|
||||||
|
cat ./RELEASE_NOTES >> ./ARCHIVE_NOTES
|
||||||
|
|
||||||
|
- name: Publish to archive repo
|
||||||
|
env:
|
||||||
|
GH_TOKEN: ${{ secrets[needs.prepare.outputs.target_repo_token] }}
|
||||||
|
GH_REPO: ${{ needs.prepare.outputs.target_repo }}
|
||||||
|
version: ${{ needs.prepare.outputs.version }}
|
||||||
|
channel: ${{ needs.prepare.outputs.channel }}
|
||||||
|
if: |
|
||||||
|
inputs.prerelease && env.GH_TOKEN != '' && env.GH_REPO != '' && env.GH_REPO != github.repository
|
||||||
|
run: |
|
||||||
|
title="${{ startswith(env.GH_REPO, 'yt-dlp/') && 'yt-dlp ' || '' }}${{ env.channel }}"
|
||||||
|
gh release create \
|
||||||
|
--notes-file ARCHIVE_NOTES \
|
||||||
|
--title "${title} ${{ env.version }}" \
|
||||||
|
${{ env.version }} \
|
||||||
|
artifact/*
|
||||||
|
|
||||||
|
- name: Prune old release
|
||||||
|
env:
|
||||||
|
GH_TOKEN: ${{ github.token }}
|
||||||
|
version: ${{ needs.prepare.outputs.version }}
|
||||||
|
target_repo: ${{ needs.prepare.outputs.target_repo }}
|
||||||
|
target_tag: ${{ needs.prepare.outputs.target_tag }}
|
||||||
|
if: |
|
||||||
|
env.target_repo == github.repository && env.target_tag != env.version
|
||||||
|
run: |
|
||||||
|
gh release delete --yes --cleanup-tag "${{ env.target_tag }}" || true
|
||||||
|
git tag --delete "${{ env.target_tag }}" || true
|
||||||
|
sleep 5 # Enough time to cover deletion race condition
|
||||||
|
|
||||||
|
- name: Publish release
|
||||||
|
env:
|
||||||
|
GH_TOKEN: ${{ github.token }}
|
||||||
|
version: ${{ needs.prepare.outputs.version }}
|
||||||
|
target_repo: ${{ needs.prepare.outputs.target_repo }}
|
||||||
|
target_tag: ${{ needs.prepare.outputs.target_tag }}
|
||||||
|
head_sha: ${{ needs.prepare.outputs.head_sha }}
|
||||||
|
if: |
|
||||||
|
env.target_repo == github.repository
|
||||||
|
run: |
|
||||||
|
title="${{ github.repository == 'yt-dlp/yt-dlp' && 'yt-dlp ' || '' }}"
|
||||||
|
title+="${{ env.target_tag != env.version && format('{0} ', env.target_tag) || '' }}"
|
||||||
|
gh release create \
|
||||||
|
--notes-file ${{ inputs.prerelease && 'PRERELEASE_NOTES' || 'RELEASE_NOTES' }} \
|
||||||
|
--target ${{ env.head_sha }} \
|
||||||
|
--title "${title}${{ env.version }}" \
|
||||||
|
${{ inputs.prerelease && '--prerelease' || '' }} \
|
||||||
|
${{ env.target_tag }} \
|
||||||
|
artifact/*
|
||||||
|
|||||||
15
CONTRIBUTORS
15
CONTRIBUTORS
@@ -513,3 +513,18 @@ awalgarg
|
|||||||
midnightveil
|
midnightveil
|
||||||
naginatana
|
naginatana
|
||||||
Riteo
|
Riteo
|
||||||
|
1100101
|
||||||
|
aniolpages
|
||||||
|
bartbroere
|
||||||
|
CrendKing
|
||||||
|
Esokrates
|
||||||
|
HitomaruKonpaku
|
||||||
|
LoserFox
|
||||||
|
peci1
|
||||||
|
saintliao
|
||||||
|
shubhexists
|
||||||
|
SirElderling
|
||||||
|
almx
|
||||||
|
elivinsky
|
||||||
|
starius
|
||||||
|
TravisDupes
|
||||||
|
|||||||
85
Changelog.md
85
Changelog.md
@@ -4,6 +4,91 @@
|
|||||||
# To create a release, dispatch the https://github.com/yt-dlp/yt-dlp/actions/workflows/release.yml workflow on master
|
# To create a release, dispatch the https://github.com/yt-dlp/yt-dlp/actions/workflows/release.yml workflow on master
|
||||||
-->
|
-->
|
||||||
|
|
||||||
|
### 2023.11.16
|
||||||
|
|
||||||
|
#### Extractor changes
|
||||||
|
- **abc.net.au**: iview, showseries: [Fix extraction](https://github.com/yt-dlp/yt-dlp/commit/15cb3528cbda7b6198f49a6b5953c226d701696b) ([#8586](https://github.com/yt-dlp/yt-dlp/issues/8586)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- **beatbump**: [Update `_VALID_URL`](https://github.com/yt-dlp/yt-dlp/commit/21dc069bea2d4d99345dd969e098f4535c751d45) ([#8576](https://github.com/yt-dlp/yt-dlp/issues/8576)) by [seproDev](https://github.com/seproDev)
|
||||||
|
- **dailymotion**: [Improve `_VALID_URL`](https://github.com/yt-dlp/yt-dlp/commit/a489f071508ec5caf5f32052d142afe86c28df7a) ([#7692](https://github.com/yt-dlp/yt-dlp/issues/7692)) by [TravisDupes](https://github.com/TravisDupes)
|
||||||
|
- **drtv**: [Fix extractor](https://github.com/yt-dlp/yt-dlp/commit/0783fd558ed0d3a8bc754beb75a406256f8b97b2) ([#8484](https://github.com/yt-dlp/yt-dlp/issues/8484)) by [almx](https://github.com/almx), [seproDev](https://github.com/seproDev)
|
||||||
|
- **eltrecetv**: [Add extractor](https://github.com/yt-dlp/yt-dlp/commit/dcfad52812aa8ce007cefbfbe63f58b49f6b1046) ([#8216](https://github.com/yt-dlp/yt-dlp/issues/8216)) by [elivinsky](https://github.com/elivinsky)
|
||||||
|
- **jiosaavn**: [Add extractors](https://github.com/yt-dlp/yt-dlp/commit/b530118e7f48232cacf8050d79a6b20bdfcf5468) ([#8307](https://github.com/yt-dlp/yt-dlp/issues/8307)) by [awalgarg](https://github.com/awalgarg)
|
||||||
|
- **njpwworld**: [Remove](https://github.com/yt-dlp/yt-dlp/commit/e569c2d1f4b665795a2b64f0aaf7f76930664233) ([#8570](https://github.com/yt-dlp/yt-dlp/issues/8570)) by [aarubui](https://github.com/aarubui)
|
||||||
|
- **tv5mondeplus**: [Extract subtitles](https://github.com/yt-dlp/yt-dlp/commit/0f634dba3afdc429ece8839b02f6d56c27b7973a) ([#4209](https://github.com/yt-dlp/yt-dlp/issues/4209)) by [FrankZ85](https://github.com/FrankZ85)
|
||||||
|
- **twitcasting**: [Fix livestream detection](https://github.com/yt-dlp/yt-dlp/commit/2325d03aa7bb80f56ba52cd6992258e44727b424) ([#8574](https://github.com/yt-dlp/yt-dlp/issues/8574)) by [JC-Chung](https://github.com/JC-Chung)
|
||||||
|
- **zenyandex**: [Fix extraction](https://github.com/yt-dlp/yt-dlp/commit/5efe68b73cbf6e907c2e6a3aa338664385084184) ([#8454](https://github.com/yt-dlp/yt-dlp/issues/8454)) by [starius](https://github.com/starius)
|
||||||
|
|
||||||
|
#### Misc. changes
|
||||||
|
- **build**: [Make `secretstorage` an optional dependency](https://github.com/yt-dlp/yt-dlp/commit/24f827875c6ba513f12ed09a3aef2bbed223760d) ([#8585](https://github.com/yt-dlp/yt-dlp/issues/8585)) by [bashonly](https://github.com/bashonly)
|
||||||
|
|
||||||
|
### 2023.11.14
|
||||||
|
|
||||||
|
#### Important changes
|
||||||
|
- **The release channels have been adjusted!**
|
||||||
|
* [`master`](https://github.com/yt-dlp/yt-dlp-master-builds) builds are made after each push, containing the latest fixes (but also possibly bugs). This was previously the `nightly` channel.
|
||||||
|
* [`nightly`](https://github.com/yt-dlp/yt-dlp-nightly-builds) builds are now made once a day, if there were any changes.
|
||||||
|
- Security: [[CVE-2023-46121](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2023-46121)] Patch [Generic Extractor MITM Vulnerability via Arbitrary Proxy Injection](https://github.com/yt-dlp/yt-dlp/security/advisories/GHSA-3ch3-jhc6-5r8x)
|
||||||
|
- Disallow smuggling of arbitrary `http_headers`; extractors now only use specific headers
|
||||||
|
|
||||||
|
#### Core changes
|
||||||
|
- [Add `--compat-option manifest-filesize-approx`](https://github.com/yt-dlp/yt-dlp/commit/10025b715ea01489557eb2c5a3cc04d361fcdb52) ([#8356](https://github.com/yt-dlp/yt-dlp/issues/8356)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- [Fix format sorting with `--load-info-json`](https://github.com/yt-dlp/yt-dlp/commit/595ea4a99b726b8fe9463e7853b7053978d0544e) ([#8521](https://github.com/yt-dlp/yt-dlp/issues/8521)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- [Include build origin in verbose output](https://github.com/yt-dlp/yt-dlp/commit/20314dd46f25e0e0a7e985a7804049aefa8b909f) by [bashonly](https://github.com/bashonly), [Grub4K](https://github.com/Grub4K)
|
||||||
|
- [Only ensure playlist thumbnail dir if writing thumbs](https://github.com/yt-dlp/yt-dlp/commit/a40e0b37dfc8c26916b0e01aa3f29f3bc42250b6) ([#8373](https://github.com/yt-dlp/yt-dlp/issues/8373)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- **update**: [Overhaul self-updater](https://github.com/yt-dlp/yt-dlp/commit/0b6ad22e6a432006a75df968f0283e6c6b3cfae6) by [bashonly](https://github.com/bashonly), [Grub4K](https://github.com/Grub4K)
|
||||||
|
|
||||||
|
#### Extractor changes
|
||||||
|
- [Do not smuggle `http_headers`](https://github.com/yt-dlp/yt-dlp/commit/f04b5bedad7b281bee9814686bba1762bae092eb) by [coletdjnz](https://github.com/coletdjnz)
|
||||||
|
- [Do not test truth value of `xml.etree.ElementTree.Element`](https://github.com/yt-dlp/yt-dlp/commit/d4f14a72dc1dd79396e0e80980268aee902b61e4) ([#8582](https://github.com/yt-dlp/yt-dlp/issues/8582)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- **brilliantpala**: [Fix cookies support](https://github.com/yt-dlp/yt-dlp/commit/9b5bedf13a3323074daceb0ec6ebb3cc6e0b9684) ([#8352](https://github.com/yt-dlp/yt-dlp/issues/8352)) by [pzhlkj6612](https://github.com/pzhlkj6612)
|
||||||
|
- **generic**: [Improve direct video link ext detection](https://github.com/yt-dlp/yt-dlp/commit/4ce2f29a50fcfb9920e6f2ffe42192945a2bad7e) ([#8340](https://github.com/yt-dlp/yt-dlp/issues/8340)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- **laxarxames**: [Add extractor](https://github.com/yt-dlp/yt-dlp/commit/312a2d1e8bc247264f9d85c5ec764e33aa0133b5) ([#8412](https://github.com/yt-dlp/yt-dlp/issues/8412)) by [aniolpages](https://github.com/aniolpages)
|
||||||
|
- **n-tv.de**: [Fix extractor](https://github.com/yt-dlp/yt-dlp/commit/8afd9468b0c822843bc480d366d1c86698daabfb) ([#8414](https://github.com/yt-dlp/yt-dlp/issues/8414)) by [1100101](https://github.com/1100101)
|
||||||
|
- **neteasemusic**: [Improve metadata extraction](https://github.com/yt-dlp/yt-dlp/commit/46acc418a53470b7f32581b3309c3cb87aa8488d) ([#8531](https://github.com/yt-dlp/yt-dlp/issues/8531)) by [LoserFox](https://github.com/LoserFox)
|
||||||
|
- **nhk**: [Improve metadata extraction](https://github.com/yt-dlp/yt-dlp/commit/54579be4364e148277c32e20a5c3efc2c3f52f5b) ([#8388](https://github.com/yt-dlp/yt-dlp/issues/8388)) by [garret1317](https://github.com/garret1317)
|
||||||
|
- **novaembed**: [Improve `_VALID_URL`](https://github.com/yt-dlp/yt-dlp/commit/3ff494f6f41c27549420fa88be27555bd449ffdc) ([#8368](https://github.com/yt-dlp/yt-dlp/issues/8368)) by [peci1](https://github.com/peci1)
|
||||||
|
- **npo**: [Send `POST` request to streams API endpoint](https://github.com/yt-dlp/yt-dlp/commit/8e02a4dcc800f9444e9d461edc41edd7b662f435) ([#8413](https://github.com/yt-dlp/yt-dlp/issues/8413)) by [bartbroere](https://github.com/bartbroere)
|
||||||
|
- **ondemandkorea**: [Overhaul extractor](https://github.com/yt-dlp/yt-dlp/commit/05adfd883a4f2ecae0267e670a62a2e45c351aeb) ([#8386](https://github.com/yt-dlp/yt-dlp/issues/8386)) by [seproDev](https://github.com/seproDev)
|
||||||
|
- **orf**: podcast: [Add extractor](https://github.com/yt-dlp/yt-dlp/commit/6ba3085616652cbf05d1858efc321fdbfc4c6119) ([#8486](https://github.com/yt-dlp/yt-dlp/issues/8486)) by [Esokrates](https://github.com/Esokrates)
|
||||||
|
- **polskieradio**: audition: [Fix playlist extraction](https://github.com/yt-dlp/yt-dlp/commit/464327acdb353ceb91d2115163a5a9621b22fe0d) ([#8459](https://github.com/yt-dlp/yt-dlp/issues/8459)) by [shubhexists](https://github.com/shubhexists)
|
||||||
|
- **qdance**: [Update `_VALID_URL`](https://github.com/yt-dlp/yt-dlp/commit/177f0d963e4b9db749805c482e6f288354c8be84) ([#8426](https://github.com/yt-dlp/yt-dlp/issues/8426)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- **radiocomercial**: [Add extractors](https://github.com/yt-dlp/yt-dlp/commit/ef12dbdcd3e7264bd3d744c1e3107597bd23ad35) ([#8508](https://github.com/yt-dlp/yt-dlp/issues/8508)) by [SirElderling](https://github.com/SirElderling)
|
||||||
|
- **sbs.co.kr**: [Add extractors](https://github.com/yt-dlp/yt-dlp/commit/25a4bd345a0dcfece6fef752d4537eb403da94d9) ([#8326](https://github.com/yt-dlp/yt-dlp/issues/8326)) by [seproDev](https://github.com/seproDev)
|
||||||
|
- **theatercomplextown**: [Add extractors](https://github.com/yt-dlp/yt-dlp/commit/2863fcf2b6876d0c7965ff7d6d9242eea653dc6b) ([#8560](https://github.com/yt-dlp/yt-dlp/issues/8560)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- **thisav**: [Remove](https://github.com/yt-dlp/yt-dlp/commit/cb480e390d85fb3a598c1b6d5eef3438ce729fc9) ([#8346](https://github.com/yt-dlp/yt-dlp/issues/8346)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- **thisoldhouse**: [Add login support](https://github.com/yt-dlp/yt-dlp/commit/c76c96677ff6a056f5844a568ef05ee22c46d6f4) ([#8561](https://github.com/yt-dlp/yt-dlp/issues/8561)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- **twitcasting**: [Fix livestream extraction](https://github.com/yt-dlp/yt-dlp/commit/7b8b1cf5eb8bf44ce70bc24e1f56f0dba2737e98) ([#8427](https://github.com/yt-dlp/yt-dlp/issues/8427)) by [JC-Chung](https://github.com/JC-Chung), [saintliao](https://github.com/saintliao)
|
||||||
|
- **twitter**
|
||||||
|
- broadcast
|
||||||
|
- [Improve metadata extraction](https://github.com/yt-dlp/yt-dlp/commit/7d337ca977d73a0a6c07ab481ed8faa8f6ff8726) ([#8383](https://github.com/yt-dlp/yt-dlp/issues/8383)) by [HitomaruKonpaku](https://github.com/HitomaruKonpaku)
|
||||||
|
- [Support `--wait-for-video`](https://github.com/yt-dlp/yt-dlp/commit/f6e97090d2ed9e05441ab0f4bec3559b816d7a00) ([#8475](https://github.com/yt-dlp/yt-dlp/issues/8475)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- **weibo**: [Fix extraction](https://github.com/yt-dlp/yt-dlp/commit/15b252dfd2c6807fe57afc5a95e59abadb32ccd2) ([#8463](https://github.com/yt-dlp/yt-dlp/issues/8463)) by [c-basalt](https://github.com/c-basalt)
|
||||||
|
- **weverse**: [Fix login error handling](https://github.com/yt-dlp/yt-dlp/commit/4a601c9eff9fb42e24a4c8da3fa03628e035b35b) ([#8458](https://github.com/yt-dlp/yt-dlp/issues/8458)) by [seproDev](https://github.com/seproDev)
|
||||||
|
- **youtube**: [Check newly uploaded iOS HLS formats](https://github.com/yt-dlp/yt-dlp/commit/ef79d20dc9d27ac002a7196f073b37f2f2721aed) ([#8336](https://github.com/yt-dlp/yt-dlp/issues/8336)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- **zoom**: [Extract combined view formats](https://github.com/yt-dlp/yt-dlp/commit/3906de07551fedb00b789345bf24cc27d6ddf128) ([#7847](https://github.com/yt-dlp/yt-dlp/issues/7847)) by [Mipsters](https://github.com/Mipsters)
|
||||||
|
|
||||||
|
#### Downloader changes
|
||||||
|
- **aria2c**: [Remove duplicate `--file-allocation=none`](https://github.com/yt-dlp/yt-dlp/commit/21b25281c51523620706b11bfc1c4a889858e1f2) ([#8332](https://github.com/yt-dlp/yt-dlp/issues/8332)) by [CrendKing](https://github.com/CrendKing)
|
||||||
|
- **dash**: [Force native downloader for `--live-from-start`](https://github.com/yt-dlp/yt-dlp/commit/2622c804d1a5accc3045db398e0fc52074f4bdb3) ([#8339](https://github.com/yt-dlp/yt-dlp/issues/8339)) by [bashonly](https://github.com/bashonly)
|
||||||
|
|
||||||
|
#### Networking changes
|
||||||
|
- **Request Handler**: requests: [Add handler for `requests` HTTP library (#3668)](https://github.com/yt-dlp/yt-dlp/commit/8a8b54523addf46dfd50ef599761a81bc22362e6) by [bashonly](https://github.com/bashonly), [coletdjnz](https://github.com/coletdjnz), [Grub4K](https://github.com/Grub4K) (With fixes in [4e38e2a](https://github.com/yt-dlp/yt-dlp/commit/4e38e2ae9d7380015349e6aee59c78bb3938befd))
|
||||||
|
|
||||||
|
Adds support for HTTPS proxies and persistent connections (keep-alive)
|
||||||
|
|
||||||
|
#### Misc. changes
|
||||||
|
- **build**
|
||||||
|
- [Include secretstorage in Linux builds](https://github.com/yt-dlp/yt-dlp/commit/9970d74c8383432c6c8779aa47d3253dcf412b14) by [bashonly](https://github.com/bashonly)
|
||||||
|
- [Overhaul and unify release workflow](https://github.com/yt-dlp/yt-dlp/commit/1d03633c5a1621b9f3a756f0a4f9dc61fab3aeaa) by [bashonly](https://github.com/bashonly), [Grub4K](https://github.com/Grub4K)
|
||||||
|
- **ci**
|
||||||
|
- [Bump `actions/checkout` to v4](https://github.com/yt-dlp/yt-dlp/commit/5438593a35b7b042fc48fe29cad0b9039f07c9bb) by [bashonly](https://github.com/bashonly)
|
||||||
|
- [Run core tests with dependencies](https://github.com/yt-dlp/yt-dlp/commit/700444c23ddb65f618c2abd942acdc0c58c650b1) by [bashonly](https://github.com/bashonly), [coletdjnz](https://github.com/coletdjnz)
|
||||||
|
- **cleanup**
|
||||||
|
- [Fix changelog typo](https://github.com/yt-dlp/yt-dlp/commit/a9d3f4b20a3533d2a40104c85bc2cc6c2564c800) by [bashonly](https://github.com/bashonly)
|
||||||
|
- [Update documentation for master and nightly channels](https://github.com/yt-dlp/yt-dlp/commit/a00af29853b8c7350ce086f4cab8c2c9cf2fcf1d) by [bashonly](https://github.com/bashonly), [Grub4K](https://github.com/Grub4K)
|
||||||
|
- Miscellaneous: [b012271](https://github.com/yt-dlp/yt-dlp/commit/b012271d01b59759e4eefeab0308698cd9e7224c) by [bashonly](https://github.com/bashonly), [coletdjnz](https://github.com/coletdjnz), [dirkf](https://github.com/dirkf), [gamer191](https://github.com/gamer191), [Grub4K](https://github.com/Grub4K), [seproDev](https://github.com/seproDev)
|
||||||
|
- **test**: update: [Implement simple updater unit tests](https://github.com/yt-dlp/yt-dlp/commit/87264d4fdadcddd91289b968dd0e4bf58d449267) by [bashonly](https://github.com/bashonly)
|
||||||
|
|
||||||
### 2023.10.13
|
### 2023.10.13
|
||||||
|
|
||||||
#### Core changes
|
#### Core changes
|
||||||
|
|||||||
39
README.md
39
README.md
@@ -121,7 +121,7 @@ yt-dlp is a [youtube-dl](https://github.com/ytdl-org/youtube-dl) fork based on t
|
|||||||
|
|
||||||
* **Self updater**: The releases can be updated using `yt-dlp -U`, and downgraded using `--update-to` if required
|
* **Self updater**: The releases can be updated using `yt-dlp -U`, and downgraded using `--update-to` if required
|
||||||
|
|
||||||
* **Nightly builds**: [Automated nightly builds](#update-channels) can be used with `--update-to nightly`
|
* **Automated builds**: [Nightly/master builds](#update-channels) can be used with `--update-to nightly` and `--update-to master`
|
||||||
|
|
||||||
See [changelog](Changelog.md) or [commits](https://github.com/yt-dlp/yt-dlp/commits) for the full list of changes
|
See [changelog](Changelog.md) or [commits](https://github.com/yt-dlp/yt-dlp/commits) for the full list of changes
|
||||||
|
|
||||||
@@ -157,14 +157,16 @@ Some of yt-dlp's default options are different from that of youtube-dl and youtu
|
|||||||
* yt-dlp's sanitization of invalid characters in filenames is different/smarter than in youtube-dl. You can use `--compat-options filename-sanitization` to revert to youtube-dl's behavior
|
* yt-dlp's sanitization of invalid characters in filenames is different/smarter than in youtube-dl. You can use `--compat-options filename-sanitization` to revert to youtube-dl's behavior
|
||||||
* yt-dlp tries to parse the external downloader outputs into the standard progress output if possible (Currently implemented: [~~aria2c~~](https://github.com/yt-dlp/yt-dlp/issues/5931)). You can use `--compat-options no-external-downloader-progress` to get the downloader output as-is
|
* yt-dlp tries to parse the external downloader outputs into the standard progress output if possible (Currently implemented: [~~aria2c~~](https://github.com/yt-dlp/yt-dlp/issues/5931)). You can use `--compat-options no-external-downloader-progress` to get the downloader output as-is
|
||||||
* yt-dlp versions between 2021.09.01 and 2023.01.02 applies `--match-filter` to nested playlists. This was an unintentional side-effect of [8f18ac](https://github.com/yt-dlp/yt-dlp/commit/8f18aca8717bb0dd49054555af8d386e5eda3a88) and is fixed in [d7b460](https://github.com/yt-dlp/yt-dlp/commit/d7b460d0e5fc710950582baed2e3fc616ed98a80). Use `--compat-options playlist-match-filter` to revert this
|
* yt-dlp versions between 2021.09.01 and 2023.01.02 applies `--match-filter` to nested playlists. This was an unintentional side-effect of [8f18ac](https://github.com/yt-dlp/yt-dlp/commit/8f18aca8717bb0dd49054555af8d386e5eda3a88) and is fixed in [d7b460](https://github.com/yt-dlp/yt-dlp/commit/d7b460d0e5fc710950582baed2e3fc616ed98a80). Use `--compat-options playlist-match-filter` to revert this
|
||||||
|
* yt-dlp versions between 2021.11.10 and 2023.06.21 estimated `filesize_approx` values for fragmented/manifest formats. This was added for convenience in [f2fe69](https://github.com/yt-dlp/yt-dlp/commit/f2fe69c7b0d208bdb1f6292b4ae92bc1e1a7444a), but was reverted in [0dff8e](https://github.com/yt-dlp/yt-dlp/commit/0dff8e4d1e6e9fb938f4256ea9af7d81f42fd54f) due to the potentially extreme inaccuracy of the estimated values. Use `--compat-options manifest-filesize-approx` to keep extracting the estimated values
|
||||||
|
* yt-dlp uses modern http client backends such as `requests`. Use `--compat-options prefer-legacy-http-handler` to prefer the legacy http handler (`urllib`) to be used for standard http requests.
|
||||||
|
|
||||||
For ease of use, a few more compat options are available:
|
For ease of use, a few more compat options are available:
|
||||||
|
|
||||||
* `--compat-options all`: Use all compat options (Do NOT use)
|
* `--compat-options all`: Use all compat options (Do NOT use)
|
||||||
* `--compat-options youtube-dl`: Same as `--compat-options all,-multistreams,-playlist-match-filter`
|
* `--compat-options youtube-dl`: Same as `--compat-options all,-multistreams,-playlist-match-filter,-manifest-filesize-approx`
|
||||||
* `--compat-options youtube-dlc`: Same as `--compat-options all,-no-live-chat,-no-youtube-channel-redirect,-playlist-match-filter`
|
* `--compat-options youtube-dlc`: Same as `--compat-options all,-no-live-chat,-no-youtube-channel-redirect,-playlist-match-filter,-manifest-filesize-approx`
|
||||||
* `--compat-options 2021`: Same as `--compat-options 2022,no-certifi,filename-sanitization,no-youtube-prefer-utc-upload-date`
|
* `--compat-options 2021`: Same as `--compat-options 2022,no-certifi,filename-sanitization,no-youtube-prefer-utc-upload-date`
|
||||||
* `--compat-options 2022`: Same as `--compat-options playlist-match-filter,no-external-downloader-progress`. Use this to enable all future compat options
|
* `--compat-options 2022`: Same as `--compat-options playlist-match-filter,no-external-downloader-progress,prefer-legacy-http-handler,manifest-filesize-approx`. Use this to enable all future compat options
|
||||||
|
|
||||||
|
|
||||||
# INSTALLATION
|
# INSTALLATION
|
||||||
@@ -191,9 +193,11 @@ For other third-party package managers, see [the wiki](https://github.com/yt-dlp
|
|||||||
|
|
||||||
<a id="update-channels"/>
|
<a id="update-channels"/>
|
||||||
|
|
||||||
There are currently two release channels for binaries, `stable` and `nightly`.
|
There are currently three release channels for binaries: `stable`, `nightly` and `master`.
|
||||||
`stable` is the default channel, and many of its changes have been tested by users of the nightly channel.
|
|
||||||
The `nightly` channel has releases built after each push to the master branch, and will have the most recent fixes and additions, but also have more risk of regressions. They are available in [their own repo](https://github.com/yt-dlp/yt-dlp-nightly-builds/releases).
|
* `stable` is the default channel, and many of its changes have been tested by users of the `nightly` and `master` channels.
|
||||||
|
* The `nightly` channel has releases scheduled to build every day around midnight UTC, for a snapshot of the project's new patches and changes. This is the **recommended channel for regular users** of yt-dlp. The `nightly` releases are available from [yt-dlp/yt-dlp-nightly-builds](https://github.com/yt-dlp/yt-dlp-nightly-builds/releases) or as development releases of the `yt-dlp` PyPI package (which can be installed with pip's `--pre` flag).
|
||||||
|
* The `master` channel features releases that are built after each push to the master branch, and these will have the very latest fixes and additions, but may also be more prone to regressions. They are available from [yt-dlp/yt-dlp-master-builds](https://github.com/yt-dlp/yt-dlp-master-builds/releases).
|
||||||
|
|
||||||
When using `--update`/`-U`, a release binary will only update to its current channel.
|
When using `--update`/`-U`, a release binary will only update to its current channel.
|
||||||
`--update-to CHANNEL` can be used to switch to a different channel when a newer version is available. `--update-to [CHANNEL@]TAG` can also be used to upgrade or downgrade to specific tags from a channel.
|
`--update-to CHANNEL` can be used to switch to a different channel when a newer version is available. `--update-to [CHANNEL@]TAG` can also be used to upgrade or downgrade to specific tags from a channel.
|
||||||
@@ -201,10 +205,19 @@ When using `--update`/`-U`, a release binary will only update to its current cha
|
|||||||
You may also use `--update-to <repository>` (`<owner>/<repository>`) to update to a channel on a completely different repository. Be careful with what repository you are updating to though, there is no verification done for binaries from different repositories.
|
You may also use `--update-to <repository>` (`<owner>/<repository>`) to update to a channel on a completely different repository. Be careful with what repository you are updating to though, there is no verification done for binaries from different repositories.
|
||||||
|
|
||||||
Example usage:
|
Example usage:
|
||||||
* `yt-dlp --update-to nightly` change to `nightly` channel and update to its latest release
|
* `yt-dlp --update-to master` switch to the `master` channel and update to its latest release
|
||||||
* `yt-dlp --update-to stable@2023.02.17` upgrade/downgrade to release to `stable` channel tag `2023.02.17`
|
* `yt-dlp --update-to stable@2023.07.06` upgrade/downgrade to release to `stable` channel tag `2023.07.06`
|
||||||
* `yt-dlp --update-to 2023.01.06` upgrade/downgrade to tag `2023.01.06` if it exists on the current channel
|
* `yt-dlp --update-to 2023.10.07` upgrade/downgrade to tag `2023.10.07` if it exists on the current channel
|
||||||
* `yt-dlp --update-to example/yt-dlp@2023.03.01` upgrade/downgrade to the release from the `example/yt-dlp` repository, tag `2023.03.01`
|
* `yt-dlp --update-to example/yt-dlp@2023.09.24` upgrade/downgrade to the release from the `example/yt-dlp` repository, tag `2023.09.24`
|
||||||
|
|
||||||
|
**Important**: Any user experiencing an issue with the `stable` release should install or update to the `nightly` release before submitting a bug report:
|
||||||
|
```
|
||||||
|
# To update to nightly from stable executable/binary:
|
||||||
|
yt-dlp --update-to nightly
|
||||||
|
|
||||||
|
# To install nightly with pip:
|
||||||
|
python -m pip install -U --pre yt-dlp
|
||||||
|
```
|
||||||
|
|
||||||
<!-- MANPAGE: BEGIN EXCLUDED SECTION -->
|
<!-- MANPAGE: BEGIN EXCLUDED SECTION -->
|
||||||
## RELEASE FILES
|
## RELEASE FILES
|
||||||
@@ -274,6 +287,7 @@ While all the other dependencies are optional, `ffmpeg` and `ffprobe` are highly
|
|||||||
* [**certifi**](https://github.com/certifi/python-certifi)\* - Provides Mozilla's root certificate bundle. Licensed under [MPLv2](https://github.com/certifi/python-certifi/blob/master/LICENSE)
|
* [**certifi**](https://github.com/certifi/python-certifi)\* - Provides Mozilla's root certificate bundle. Licensed under [MPLv2](https://github.com/certifi/python-certifi/blob/master/LICENSE)
|
||||||
* [**brotli**](https://github.com/google/brotli)\* or [**brotlicffi**](https://github.com/python-hyper/brotlicffi) - [Brotli](https://en.wikipedia.org/wiki/Brotli) content encoding support. Both licensed under MIT <sup>[1](https://github.com/google/brotli/blob/master/LICENSE) [2](https://github.com/python-hyper/brotlicffi/blob/master/LICENSE) </sup>
|
* [**brotli**](https://github.com/google/brotli)\* or [**brotlicffi**](https://github.com/python-hyper/brotlicffi) - [Brotli](https://en.wikipedia.org/wiki/Brotli) content encoding support. Both licensed under MIT <sup>[1](https://github.com/google/brotli/blob/master/LICENSE) [2](https://github.com/python-hyper/brotlicffi/blob/master/LICENSE) </sup>
|
||||||
* [**websockets**](https://github.com/aaugustin/websockets)\* - For downloading over websocket. Licensed under [BSD-3-Clause](https://github.com/aaugustin/websockets/blob/main/LICENSE)
|
* [**websockets**](https://github.com/aaugustin/websockets)\* - For downloading over websocket. Licensed under [BSD-3-Clause](https://github.com/aaugustin/websockets/blob/main/LICENSE)
|
||||||
|
* [**requests**](https://github.com/psf/requests)\* - HTTP library. For HTTPS proxy and persistent connections support. Licensed under [Apache-2.0](https://github.com/psf/requests/blob/main/LICENSE)
|
||||||
|
|
||||||
### Metadata
|
### Metadata
|
||||||
|
|
||||||
@@ -366,7 +380,8 @@ If you fork the project on GitHub, you can run your fork's [build workflow](.git
|
|||||||
CHANNEL can be a repository as well. CHANNEL
|
CHANNEL can be a repository as well. CHANNEL
|
||||||
and TAG default to "stable" and "latest"
|
and TAG default to "stable" and "latest"
|
||||||
respectively if omitted; See "UPDATE" for
|
respectively if omitted; See "UPDATE" for
|
||||||
details. Supported channels: stable, nightly
|
details. Supported channels: stable,
|
||||||
|
nightly, master
|
||||||
-i, --ignore-errors Ignore download and postprocessing errors.
|
-i, --ignore-errors Ignore download and postprocessing errors.
|
||||||
The download will be considered successful
|
The download will be considered successful
|
||||||
even if the postprocessing fails
|
even if the postprocessing fails
|
||||||
|
|||||||
@@ -98,5 +98,21 @@
|
|||||||
"action": "add",
|
"action": "add",
|
||||||
"when": "61bdf15fc7400601c3da1aa7a43917310a5bf391",
|
"when": "61bdf15fc7400601c3da1aa7a43917310a5bf391",
|
||||||
"short": "[priority] Security: [[CVE-2023-40581](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2023-40581)] [Prevent RCE when using `--exec` with `%q` on Windows](https://github.com/yt-dlp/yt-dlp/security/advisories/GHSA-42h4-v29r-42qg)\n - The shell escape function is now using `\"\"` instead of `\\\"`.\n - `utils.Popen` has been patched to properly quote commands."
|
"short": "[priority] Security: [[CVE-2023-40581](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2023-40581)] [Prevent RCE when using `--exec` with `%q` on Windows](https://github.com/yt-dlp/yt-dlp/security/advisories/GHSA-42h4-v29r-42qg)\n - The shell escape function is now using `\"\"` instead of `\\\"`.\n - `utils.Popen` has been patched to properly quote commands."
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"action": "change",
|
||||||
|
"when": "8a8b54523addf46dfd50ef599761a81bc22362e6",
|
||||||
|
"short": "[rh:requests] Add handler for `requests` HTTP library (#3668)\n\n\tAdds support for HTTPS proxies and persistent connections (keep-alive)",
|
||||||
|
"authors": ["bashonly", "coletdjnz", "Grub4K"]
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"action": "add",
|
||||||
|
"when": "1d03633c5a1621b9f3a756f0a4f9dc61fab3aeaa",
|
||||||
|
"short": "[priority] **The release channels have been adjusted!**\n\t* [`master`](https://github.com/yt-dlp/yt-dlp-master-builds) builds are made after each push, containing the latest fixes (but also possibly bugs). This was previously the `nightly` channel.\n\t* [`nightly`](https://github.com/yt-dlp/yt-dlp-nightly-builds) builds are now made once a day, if there were any changes."
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"action": "add",
|
||||||
|
"when": "f04b5bedad7b281bee9814686bba1762bae092eb",
|
||||||
|
"short": "[priority] Security: [[CVE-2023-46121](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2023-46121)] Patch [Generic Extractor MITM Vulnerability via Arbitrary Proxy Injection](https://github.com/yt-dlp/yt-dlp/security/advisories/GHSA-3ch3-jhc6-5r8x)\n\t- Disallow smuggling of arbitrary `http_headers`; extractors now only use specific headers"
|
||||||
}
|
}
|
||||||
]
|
]
|
||||||
|
|||||||
@@ -56,6 +56,7 @@ class CommitGroup(enum.Enum):
|
|||||||
},
|
},
|
||||||
cls.MISC: {
|
cls.MISC: {
|
||||||
'build',
|
'build',
|
||||||
|
'ci',
|
||||||
'cleanup',
|
'cleanup',
|
||||||
'devscripts',
|
'devscripts',
|
||||||
'docs',
|
'docs',
|
||||||
|
|||||||
@@ -12,7 +12,6 @@ import re
|
|||||||
from devscripts.utils import (
|
from devscripts.utils import (
|
||||||
get_filename_args,
|
get_filename_args,
|
||||||
read_file,
|
read_file,
|
||||||
read_version,
|
|
||||||
write_file,
|
write_file,
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -35,19 +34,18 @@ VERBOSE_TMPL = '''
|
|||||||
description: |
|
description: |
|
||||||
It should start like this:
|
It should start like this:
|
||||||
placeholder: |
|
placeholder: |
|
||||||
[debug] Command-line config: ['-vU', 'test:youtube']
|
[debug] Command-line config: ['-vU', 'https://www.youtube.com/watch?v=BaW_jenozKc']
|
||||||
[debug] Portable config "yt-dlp.conf": ['-i']
|
|
||||||
[debug] Encodings: locale cp65001, fs utf-8, pref cp65001, out utf-8, error utf-8, screen utf-8
|
[debug] Encodings: locale cp65001, fs utf-8, pref cp65001, out utf-8, error utf-8, screen utf-8
|
||||||
[debug] yt-dlp version %(version)s [9d339c4] (win32_exe)
|
[debug] yt-dlp version nightly@... from yt-dlp/yt-dlp [b634ba742] (win_exe)
|
||||||
[debug] Python 3.8.10 (CPython 64bit) - Windows-10-10.0.22000-SP0
|
[debug] Python 3.8.10 (CPython 64bit) - Windows-10-10.0.22000-SP0
|
||||||
[debug] Checking exe version: ffmpeg -bsfs
|
|
||||||
[debug] Checking exe version: ffprobe -bsfs
|
|
||||||
[debug] exe versions: ffmpeg N-106550-g072101bd52-20220410 (fdk,setts), ffprobe N-106624-g391ce570c8-20220415, phantomjs 2.1.1
|
[debug] exe versions: ffmpeg N-106550-g072101bd52-20220410 (fdk,setts), ffprobe N-106624-g391ce570c8-20220415, phantomjs 2.1.1
|
||||||
[debug] Optional libraries: Cryptodome-3.15.0, brotli-1.0.9, certifi-2022.06.15, mutagen-1.45.1, sqlite3-2.6.0, websockets-10.3
|
[debug] Optional libraries: Cryptodome-3.15.0, brotli-1.0.9, certifi-2022.06.15, mutagen-1.45.1, sqlite3-2.6.0, websockets-10.3
|
||||||
[debug] Proxy map: {}
|
[debug] Proxy map: {}
|
||||||
[debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp/releases/latest
|
[debug] Request Handlers: urllib, requests
|
||||||
Latest version: %(version)s, Current version: %(version)s
|
[debug] Loaded 1893 extractors
|
||||||
yt-dlp is up to date (%(version)s)
|
[debug] Fetching release info: https://api.github.com/repos/yt-dlp/yt-dlp-nightly-builds/releases/latest
|
||||||
|
yt-dlp is up to date (nightly@... from yt-dlp/yt-dlp-nightly-builds)
|
||||||
|
[youtube] Extracting URL: https://www.youtube.com/watch?v=BaW_jenozKc
|
||||||
<more lines>
|
<more lines>
|
||||||
render: shell
|
render: shell
|
||||||
validations:
|
validations:
|
||||||
@@ -66,7 +64,7 @@ NO_SKIP = '''
|
|||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
fields = {'version': read_version(), 'no_skip': NO_SKIP}
|
fields = {'no_skip': NO_SKIP}
|
||||||
fields['verbose'] = VERBOSE_TMPL % fields
|
fields['verbose'] = VERBOSE_TMPL % fields
|
||||||
fields['verbose_optional'] = re.sub(r'(\n\s+validations:)?\n\s+required: true', '', fields['verbose'])
|
fields['verbose_optional'] = re.sub(r'(\n\s+validations:)?\n\s+required: true', '', fields['verbose'])
|
||||||
|
|
||||||
|
|||||||
@@ -1,39 +0,0 @@
|
|||||||
#!/usr/bin/env python3
|
|
||||||
|
|
||||||
"""
|
|
||||||
Usage: python3 ./devscripts/update-formulae.py <path-to-formulae-rb> <version>
|
|
||||||
version can be either 0-aligned (yt-dlp version) or normalized (PyPi version)
|
|
||||||
"""
|
|
||||||
|
|
||||||
# Allow direct execution
|
|
||||||
import os
|
|
||||||
import sys
|
|
||||||
|
|
||||||
sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
|
|
||||||
|
|
||||||
|
|
||||||
import json
|
|
||||||
import re
|
|
||||||
import urllib.request
|
|
||||||
|
|
||||||
from devscripts.utils import read_file, write_file
|
|
||||||
|
|
||||||
filename, version = sys.argv[1:]
|
|
||||||
|
|
||||||
normalized_version = '.'.join(str(int(x)) for x in version.split('.'))
|
|
||||||
|
|
||||||
pypi_release = json.loads(urllib.request.urlopen(
|
|
||||||
'https://pypi.org/pypi/yt-dlp/%s/json' % normalized_version
|
|
||||||
).read().decode())
|
|
||||||
|
|
||||||
tarball_file = next(x for x in pypi_release['urls'] if x['filename'].endswith('.tar.gz'))
|
|
||||||
|
|
||||||
sha256sum = tarball_file['digests']['sha256']
|
|
||||||
url = tarball_file['url']
|
|
||||||
|
|
||||||
formulae_text = read_file(filename)
|
|
||||||
|
|
||||||
formulae_text = re.sub(r'sha256 "[0-9a-f]*?"', 'sha256 "%s"' % sha256sum, formulae_text, count=1)
|
|
||||||
formulae_text = re.sub(r'url "[^"]*?"', 'url "%s"' % url, formulae_text, count=1)
|
|
||||||
|
|
||||||
write_file(filename, formulae_text)
|
|
||||||
@@ -20,7 +20,7 @@ def get_new_version(version, revision):
|
|||||||
version = datetime.now(timezone.utc).strftime('%Y.%m.%d')
|
version = datetime.now(timezone.utc).strftime('%Y.%m.%d')
|
||||||
|
|
||||||
if revision:
|
if revision:
|
||||||
assert revision.isdigit(), 'Revision must be a number'
|
assert revision.isdecimal(), 'Revision must be a number'
|
||||||
else:
|
else:
|
||||||
old_version = read_version().split('.')
|
old_version = read_version().split('.')
|
||||||
if version.split('.') == old_version[:3]:
|
if version.split('.') == old_version[:3]:
|
||||||
@@ -46,6 +46,10 @@ VARIANT = None
|
|||||||
UPDATE_HINT = None
|
UPDATE_HINT = None
|
||||||
|
|
||||||
CHANNEL = {channel!r}
|
CHANNEL = {channel!r}
|
||||||
|
|
||||||
|
ORIGIN = {origin!r}
|
||||||
|
|
||||||
|
_pkg_version = {package_version!r}
|
||||||
'''
|
'''
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
@@ -53,6 +57,12 @@ if __name__ == '__main__':
|
|||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
'-c', '--channel', default='stable',
|
'-c', '--channel', default='stable',
|
||||||
help='Select update channel (default: %(default)s)')
|
help='Select update channel (default: %(default)s)')
|
||||||
|
parser.add_argument(
|
||||||
|
'-r', '--origin', default='local',
|
||||||
|
help='Select origin/repository (default: %(default)s)')
|
||||||
|
parser.add_argument(
|
||||||
|
'-s', '--suffix', default='',
|
||||||
|
help='Add an alphanumeric suffix to the package version, e.g. "dev"')
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
'-o', '--output', default='yt_dlp/version.py',
|
'-o', '--output', default='yt_dlp/version.py',
|
||||||
help='The output file to write to (default: %(default)s)')
|
help='The output file to write to (default: %(default)s)')
|
||||||
@@ -66,6 +76,7 @@ if __name__ == '__main__':
|
|||||||
args.version if args.version and '.' in args.version
|
args.version if args.version and '.' in args.version
|
||||||
else get_new_version(None, args.version))
|
else get_new_version(None, args.version))
|
||||||
write_file(args.output, VERSION_TEMPLATE.format(
|
write_file(args.output, VERSION_TEMPLATE.format(
|
||||||
version=version, git_head=git_head, channel=args.channel))
|
version=version, git_head=git_head, channel=args.channel, origin=args.origin,
|
||||||
|
package_version=f'{version}{args.suffix}'))
|
||||||
|
|
||||||
print(f'version={version} ({args.channel}), head={git_head}')
|
print(f'version={version} ({args.channel}), head={git_head}')
|
||||||
|
|||||||
@@ -13,10 +13,11 @@ def write_file(fname, content, mode='w'):
|
|||||||
return f.write(content)
|
return f.write(content)
|
||||||
|
|
||||||
|
|
||||||
def read_version(fname='yt_dlp/version.py'):
|
def read_version(fname='yt_dlp/version.py', varname='__version__'):
|
||||||
"""Get the version without importing the package"""
|
"""Get the version without importing the package"""
|
||||||
exec(compile(read_file(fname), fname, 'exec'))
|
items = {}
|
||||||
return locals()['__version__']
|
exec(compile(read_file(fname), fname, 'exec'), items)
|
||||||
|
return items[varname]
|
||||||
|
|
||||||
|
|
||||||
def get_filename_args(has_infile=False, default_outfile=None):
|
def get_filename_args(has_infile=False, default_outfile=None):
|
||||||
|
|||||||
@@ -1,6 +1,8 @@
|
|||||||
mutagen
|
mutagen
|
||||||
pycryptodomex
|
pycryptodomex
|
||||||
websockets
|
websockets
|
||||||
brotli; platform_python_implementation=='CPython'
|
brotli; implementation_name=='cpython'
|
||||||
brotlicffi; platform_python_implementation!='CPython'
|
brotlicffi; implementation_name!='cpython'
|
||||||
certifi
|
certifi
|
||||||
|
requests>=2.31.0,<3
|
||||||
|
urllib3>=1.26.17,<3
|
||||||
|
|||||||
13
setup.py
13
setup.py
@@ -18,7 +18,7 @@ except ImportError:
|
|||||||
|
|
||||||
from devscripts.utils import read_file, read_version
|
from devscripts.utils import read_file, read_version
|
||||||
|
|
||||||
VERSION = read_version()
|
VERSION = read_version(varname='_pkg_version')
|
||||||
|
|
||||||
DESCRIPTION = 'A youtube-dl fork with additional features and patches'
|
DESCRIPTION = 'A youtube-dl fork with additional features and patches'
|
||||||
|
|
||||||
@@ -62,7 +62,14 @@ def py2exe_params():
|
|||||||
'compressed': 1,
|
'compressed': 1,
|
||||||
'optimize': 2,
|
'optimize': 2,
|
||||||
'dist_dir': './dist',
|
'dist_dir': './dist',
|
||||||
'excludes': ['Crypto', 'Cryptodome'], # py2exe cannot import Crypto
|
'excludes': [
|
||||||
|
# py2exe cannot import Crypto
|
||||||
|
'Crypto',
|
||||||
|
'Cryptodome',
|
||||||
|
# py2exe appears to confuse this with our socks library.
|
||||||
|
# We don't use pysocks and urllib3.contrib.socks would fail to import if tried.
|
||||||
|
'urllib3.contrib.socks'
|
||||||
|
],
|
||||||
'dll_excludes': ['w9xpopen.exe', 'crypt32.dll'],
|
'dll_excludes': ['w9xpopen.exe', 'crypt32.dll'],
|
||||||
# Modules that are only imported dynamically must be added here
|
# Modules that are only imported dynamically must be added here
|
||||||
'includes': ['yt_dlp.compat._legacy', 'yt_dlp.compat._deprecated',
|
'includes': ['yt_dlp.compat._legacy', 'yt_dlp.compat._deprecated',
|
||||||
@@ -135,7 +142,7 @@ def main():
|
|||||||
params = build_params()
|
params = build_params()
|
||||||
|
|
||||||
setup(
|
setup(
|
||||||
name='yt-dlp',
|
name='yt-dlp', # package name (do not change/remove comment)
|
||||||
version=VERSION,
|
version=VERSION,
|
||||||
maintainer='pukkandan',
|
maintainer='pukkandan',
|
||||||
maintainer_email='pukkandan.ytdlp@gmail.com',
|
maintainer_email='pukkandan.ytdlp@gmail.com',
|
||||||
|
|||||||
@@ -414,6 +414,7 @@
|
|||||||
- **EllenTubeVideo**
|
- **EllenTubeVideo**
|
||||||
- **Elonet**
|
- **Elonet**
|
||||||
- **ElPais**: El País
|
- **ElPais**: El País
|
||||||
|
- **ElTreceTV**: El Trece TV (Argentina)
|
||||||
- **Embedly**
|
- **Embedly**
|
||||||
- **EMPFlix**
|
- **EMPFlix**
|
||||||
- **Engadget**
|
- **Engadget**
|
||||||
@@ -654,6 +655,8 @@
|
|||||||
- **Jamendo**
|
- **Jamendo**
|
||||||
- **JamendoAlbum**
|
- **JamendoAlbum**
|
||||||
- **JeuxVideo**
|
- **JeuxVideo**
|
||||||
|
- **JioSaavnAlbum**
|
||||||
|
- **JioSaavnSong**
|
||||||
- **Joj**
|
- **Joj**
|
||||||
- **Jove**
|
- **Jove**
|
||||||
- **JStream**
|
- **JStream**
|
||||||
@@ -700,6 +703,7 @@
|
|||||||
- **LastFM**
|
- **LastFM**
|
||||||
- **LastFMPlaylist**
|
- **LastFMPlaylist**
|
||||||
- **LastFMUser**
|
- **LastFMUser**
|
||||||
|
- **LaXarxaMes**: [*laxarxames*](## "netrc machine")
|
||||||
- **lbry**
|
- **lbry**
|
||||||
- **lbry:channel**
|
- **lbry:channel**
|
||||||
- **lbry:playlist**
|
- **lbry:playlist**
|
||||||
@@ -975,7 +979,6 @@
|
|||||||
- **Nitter**
|
- **Nitter**
|
||||||
- **njoy**: N-JOY
|
- **njoy**: N-JOY
|
||||||
- **njoy:embed**
|
- **njoy:embed**
|
||||||
- **NJPWWorld**: [*njpwworld*](## "netrc machine") 新日本プロレスワールド
|
|
||||||
- **NobelPrize**
|
- **NobelPrize**
|
||||||
- **NoicePodcast**
|
- **NoicePodcast**
|
||||||
- **NonkTube**
|
- **NonkTube**
|
||||||
@@ -1026,6 +1029,7 @@
|
|||||||
- **on24**: ON24
|
- **on24**: ON24
|
||||||
- **OnDemandChinaEpisode**
|
- **OnDemandChinaEpisode**
|
||||||
- **OnDemandKorea**
|
- **OnDemandKorea**
|
||||||
|
- **OnDemandKoreaProgram**
|
||||||
- **OneFootball**
|
- **OneFootball**
|
||||||
- **OnePlacePodcast**
|
- **OnePlacePodcast**
|
||||||
- **onet.pl**
|
- **onet.pl**
|
||||||
@@ -1043,6 +1047,7 @@
|
|||||||
- **OraTV**
|
- **OraTV**
|
||||||
- **orf:fm4:story**: fm4.orf.at stories
|
- **orf:fm4:story**: fm4.orf.at stories
|
||||||
- **orf:iptv**: iptv.ORF.at
|
- **orf:iptv**: iptv.ORF.at
|
||||||
|
- **orf:podcast**
|
||||||
- **orf:radio**
|
- **orf:radio**
|
||||||
- **orf:tvthek**: ORF TVthek
|
- **orf:tvthek**: ORF TVthek
|
||||||
- **OsnatelTV**: [*osnateltv*](## "netrc machine")
|
- **OsnatelTV**: [*osnateltv*](## "netrc machine")
|
||||||
@@ -1180,6 +1185,8 @@
|
|||||||
- **radiobremen**
|
- **radiobremen**
|
||||||
- **radiocanada**
|
- **radiocanada**
|
||||||
- **radiocanada:audiovideo**
|
- **radiocanada:audiovideo**
|
||||||
|
- **RadioComercial**
|
||||||
|
- **RadioComercialPlaylist**
|
||||||
- **radiofrance**
|
- **radiofrance**
|
||||||
- **RadioFranceLive**
|
- **RadioFranceLive**
|
||||||
- **RadioFrancePodcast**
|
- **RadioFrancePodcast**
|
||||||
@@ -1306,6 +1313,9 @@
|
|||||||
- **Sapo**: SAPO Vídeos
|
- **Sapo**: SAPO Vídeos
|
||||||
- **savefrom.net**
|
- **savefrom.net**
|
||||||
- **SBS**: sbs.com.au
|
- **SBS**: sbs.com.au
|
||||||
|
- **sbs.co.kr**
|
||||||
|
- **sbs.co.kr:allvod_program**
|
||||||
|
- **sbs.co.kr:programs_vod**
|
||||||
- **schooltv**
|
- **schooltv**
|
||||||
- **ScienceChannel**
|
- **ScienceChannel**
|
||||||
- **screen.yahoo:search**: Yahoo screen search; "yvsearch:" prefix
|
- **screen.yahoo:search**: Yahoo screen search; "yvsearch:" prefix
|
||||||
@@ -1474,6 +1484,8 @@
|
|||||||
- **TenPlaySeason**
|
- **TenPlaySeason**
|
||||||
- **TF1**
|
- **TF1**
|
||||||
- **TFO**
|
- **TFO**
|
||||||
|
- **theatercomplextown:ppv**: [*theatercomplextown*](## "netrc machine")
|
||||||
|
- **theatercomplextown:vod**: [*theatercomplextown*](## "netrc machine")
|
||||||
- **TheHoleTv**
|
- **TheHoleTv**
|
||||||
- **TheIntercept**
|
- **TheIntercept**
|
||||||
- **ThePlatform**
|
- **ThePlatform**
|
||||||
@@ -1482,8 +1494,7 @@
|
|||||||
- **TheSun**
|
- **TheSun**
|
||||||
- **TheWeatherChannel**
|
- **TheWeatherChannel**
|
||||||
- **ThisAmericanLife**
|
- **ThisAmericanLife**
|
||||||
- **ThisAV**
|
- **ThisOldHouse**: [*thisoldhouse*](## "netrc machine")
|
||||||
- **ThisOldHouse**
|
|
||||||
- **ThisVid**
|
- **ThisVid**
|
||||||
- **ThisVidMember**
|
- **ThisVidMember**
|
||||||
- **ThisVidPlaylist**
|
- **ThisVidPlaylist**
|
||||||
|
|||||||
@@ -28,7 +28,7 @@ from http.cookiejar import CookieJar
|
|||||||
|
|
||||||
from test.helper import FakeYDL, http_server_port
|
from test.helper import FakeYDL, http_server_port
|
||||||
from yt_dlp.cookies import YoutubeDLCookieJar
|
from yt_dlp.cookies import YoutubeDLCookieJar
|
||||||
from yt_dlp.dependencies import brotli
|
from yt_dlp.dependencies import brotli, requests, urllib3
|
||||||
from yt_dlp.networking import (
|
from yt_dlp.networking import (
|
||||||
HEADRequest,
|
HEADRequest,
|
||||||
PUTRequest,
|
PUTRequest,
|
||||||
@@ -43,6 +43,7 @@ from yt_dlp.networking.exceptions import (
|
|||||||
HTTPError,
|
HTTPError,
|
||||||
IncompleteRead,
|
IncompleteRead,
|
||||||
NoSupportingHandlers,
|
NoSupportingHandlers,
|
||||||
|
ProxyError,
|
||||||
RequestError,
|
RequestError,
|
||||||
SSLError,
|
SSLError,
|
||||||
TransportError,
|
TransportError,
|
||||||
@@ -305,7 +306,7 @@ class TestRequestHandlerBase:
|
|||||||
|
|
||||||
|
|
||||||
class TestHTTPRequestHandler(TestRequestHandlerBase):
|
class TestHTTPRequestHandler(TestRequestHandlerBase):
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_verify_cert(self, handler):
|
def test_verify_cert(self, handler):
|
||||||
with handler() as rh:
|
with handler() as rh:
|
||||||
with pytest.raises(CertificateVerifyError):
|
with pytest.raises(CertificateVerifyError):
|
||||||
@@ -316,7 +317,7 @@ class TestHTTPRequestHandler(TestRequestHandlerBase):
|
|||||||
assert r.status == 200
|
assert r.status == 200
|
||||||
r.close()
|
r.close()
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_ssl_error(self, handler):
|
def test_ssl_error(self, handler):
|
||||||
# HTTPS server with too old TLS version
|
# HTTPS server with too old TLS version
|
||||||
# XXX: is there a better way to test this than to create a new server?
|
# XXX: is there a better way to test this than to create a new server?
|
||||||
@@ -334,7 +335,7 @@ class TestHTTPRequestHandler(TestRequestHandlerBase):
|
|||||||
validate_and_send(rh, Request(f'https://127.0.0.1:{https_port}/headers'))
|
validate_and_send(rh, Request(f'https://127.0.0.1:{https_port}/headers'))
|
||||||
assert not issubclass(exc_info.type, CertificateVerifyError)
|
assert not issubclass(exc_info.type, CertificateVerifyError)
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_percent_encode(self, handler):
|
def test_percent_encode(self, handler):
|
||||||
with handler() as rh:
|
with handler() as rh:
|
||||||
# Unicode characters should be encoded with uppercase percent-encoding
|
# Unicode characters should be encoded with uppercase percent-encoding
|
||||||
@@ -346,7 +347,7 @@ class TestHTTPRequestHandler(TestRequestHandlerBase):
|
|||||||
assert res.status == 200
|
assert res.status == 200
|
||||||
res.close()
|
res.close()
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_remove_dot_segments(self, handler):
|
def test_remove_dot_segments(self, handler):
|
||||||
with handler() as rh:
|
with handler() as rh:
|
||||||
# This isn't a comprehensive test,
|
# This isn't a comprehensive test,
|
||||||
@@ -361,14 +362,14 @@ class TestHTTPRequestHandler(TestRequestHandlerBase):
|
|||||||
assert res.url == f'http://127.0.0.1:{self.http_port}/headers'
|
assert res.url == f'http://127.0.0.1:{self.http_port}/headers'
|
||||||
res.close()
|
res.close()
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_unicode_path_redirection(self, handler):
|
def test_unicode_path_redirection(self, handler):
|
||||||
with handler() as rh:
|
with handler() as rh:
|
||||||
r = validate_and_send(rh, Request(f'http://127.0.0.1:{self.http_port}/302-non-ascii-redirect'))
|
r = validate_and_send(rh, Request(f'http://127.0.0.1:{self.http_port}/302-non-ascii-redirect'))
|
||||||
assert r.url == f'http://127.0.0.1:{self.http_port}/%E4%B8%AD%E6%96%87.html'
|
assert r.url == f'http://127.0.0.1:{self.http_port}/%E4%B8%AD%E6%96%87.html'
|
||||||
r.close()
|
r.close()
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_raise_http_error(self, handler):
|
def test_raise_http_error(self, handler):
|
||||||
with handler() as rh:
|
with handler() as rh:
|
||||||
for bad_status in (400, 500, 599, 302):
|
for bad_status in (400, 500, 599, 302):
|
||||||
@@ -378,7 +379,7 @@ class TestHTTPRequestHandler(TestRequestHandlerBase):
|
|||||||
# Should not raise an error
|
# Should not raise an error
|
||||||
validate_and_send(rh, Request('http://127.0.0.1:%d/gen_200' % self.http_port)).close()
|
validate_and_send(rh, Request('http://127.0.0.1:%d/gen_200' % self.http_port)).close()
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_response_url(self, handler):
|
def test_response_url(self, handler):
|
||||||
with handler() as rh:
|
with handler() as rh:
|
||||||
# Response url should be that of the last url in redirect chain
|
# Response url should be that of the last url in redirect chain
|
||||||
@@ -389,7 +390,7 @@ class TestHTTPRequestHandler(TestRequestHandlerBase):
|
|||||||
assert res2.url == f'http://127.0.0.1:{self.http_port}/gen_200'
|
assert res2.url == f'http://127.0.0.1:{self.http_port}/gen_200'
|
||||||
res2.close()
|
res2.close()
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_redirect(self, handler):
|
def test_redirect(self, handler):
|
||||||
with handler() as rh:
|
with handler() as rh:
|
||||||
def do_req(redirect_status, method, assert_no_content=False):
|
def do_req(redirect_status, method, assert_no_content=False):
|
||||||
@@ -444,7 +445,7 @@ class TestHTTPRequestHandler(TestRequestHandlerBase):
|
|||||||
with pytest.raises(HTTPError):
|
with pytest.raises(HTTPError):
|
||||||
do_req(code, 'GET')
|
do_req(code, 'GET')
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_request_cookie_header(self, handler):
|
def test_request_cookie_header(self, handler):
|
||||||
# We should accept a Cookie header being passed as in normal headers and handle it appropriately.
|
# We should accept a Cookie header being passed as in normal headers and handle it appropriately.
|
||||||
with handler() as rh:
|
with handler() as rh:
|
||||||
@@ -476,19 +477,19 @@ class TestHTTPRequestHandler(TestRequestHandlerBase):
|
|||||||
assert b'Cookie: test=ytdlp' not in data
|
assert b'Cookie: test=ytdlp' not in data
|
||||||
assert b'Cookie: test=test' in data
|
assert b'Cookie: test=test' in data
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_redirect_loop(self, handler):
|
def test_redirect_loop(self, handler):
|
||||||
with handler() as rh:
|
with handler() as rh:
|
||||||
with pytest.raises(HTTPError, match='redirect loop'):
|
with pytest.raises(HTTPError, match='redirect loop'):
|
||||||
validate_and_send(rh, Request(f'http://127.0.0.1:{self.http_port}/redirect_loop'))
|
validate_and_send(rh, Request(f'http://127.0.0.1:{self.http_port}/redirect_loop'))
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_incompleteread(self, handler):
|
def test_incompleteread(self, handler):
|
||||||
with handler(timeout=2) as rh:
|
with handler(timeout=2) as rh:
|
||||||
with pytest.raises(IncompleteRead):
|
with pytest.raises(IncompleteRead):
|
||||||
validate_and_send(rh, Request('http://127.0.0.1:%d/incompleteread' % self.http_port)).read()
|
validate_and_send(rh, Request('http://127.0.0.1:%d/incompleteread' % self.http_port)).read()
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_cookies(self, handler):
|
def test_cookies(self, handler):
|
||||||
cookiejar = YoutubeDLCookieJar()
|
cookiejar = YoutubeDLCookieJar()
|
||||||
cookiejar.set_cookie(http.cookiejar.Cookie(
|
cookiejar.set_cookie(http.cookiejar.Cookie(
|
||||||
@@ -505,7 +506,7 @@ class TestHTTPRequestHandler(TestRequestHandlerBase):
|
|||||||
rh, Request(f'http://127.0.0.1:{self.http_port}/headers', extensions={'cookiejar': cookiejar})).read()
|
rh, Request(f'http://127.0.0.1:{self.http_port}/headers', extensions={'cookiejar': cookiejar})).read()
|
||||||
assert b'Cookie: test=ytdlp' in data
|
assert b'Cookie: test=ytdlp' in data
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_headers(self, handler):
|
def test_headers(self, handler):
|
||||||
|
|
||||||
with handler(headers=HTTPHeaderDict({'test1': 'test', 'test2': 'test2'})) as rh:
|
with handler(headers=HTTPHeaderDict({'test1': 'test', 'test2': 'test2'})) as rh:
|
||||||
@@ -521,7 +522,7 @@ class TestHTTPRequestHandler(TestRequestHandlerBase):
|
|||||||
assert b'Test2: test2' not in data
|
assert b'Test2: test2' not in data
|
||||||
assert b'Test3: test3' in data
|
assert b'Test3: test3' in data
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_timeout(self, handler):
|
def test_timeout(self, handler):
|
||||||
with handler() as rh:
|
with handler() as rh:
|
||||||
# Default timeout is 20 seconds, so this should go through
|
# Default timeout is 20 seconds, so this should go through
|
||||||
@@ -537,7 +538,7 @@ class TestHTTPRequestHandler(TestRequestHandlerBase):
|
|||||||
validate_and_send(
|
validate_and_send(
|
||||||
rh, Request(f'http://127.0.0.1:{self.http_port}/timeout_1', extensions={'timeout': 4}))
|
rh, Request(f'http://127.0.0.1:{self.http_port}/timeout_1', extensions={'timeout': 4}))
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_source_address(self, handler):
|
def test_source_address(self, handler):
|
||||||
source_address = f'127.0.0.{random.randint(5, 255)}'
|
source_address = f'127.0.0.{random.randint(5, 255)}'
|
||||||
with handler(source_address=source_address) as rh:
|
with handler(source_address=source_address) as rh:
|
||||||
@@ -545,13 +546,13 @@ class TestHTTPRequestHandler(TestRequestHandlerBase):
|
|||||||
rh, Request(f'http://127.0.0.1:{self.http_port}/source_address')).read().decode()
|
rh, Request(f'http://127.0.0.1:{self.http_port}/source_address')).read().decode()
|
||||||
assert source_address == data
|
assert source_address == data
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_gzip_trailing_garbage(self, handler):
|
def test_gzip_trailing_garbage(self, handler):
|
||||||
with handler() as rh:
|
with handler() as rh:
|
||||||
data = validate_and_send(rh, Request(f'http://localhost:{self.http_port}/trailing_garbage')).read().decode()
|
data = validate_and_send(rh, Request(f'http://localhost:{self.http_port}/trailing_garbage')).read().decode()
|
||||||
assert data == '<html><video src="/vid.mp4" /></html>'
|
assert data == '<html><video src="/vid.mp4" /></html>'
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
@pytest.mark.skipif(not brotli, reason='brotli support is not installed')
|
@pytest.mark.skipif(not brotli, reason='brotli support is not installed')
|
||||||
def test_brotli(self, handler):
|
def test_brotli(self, handler):
|
||||||
with handler() as rh:
|
with handler() as rh:
|
||||||
@@ -562,7 +563,7 @@ class TestHTTPRequestHandler(TestRequestHandlerBase):
|
|||||||
assert res.headers.get('Content-Encoding') == 'br'
|
assert res.headers.get('Content-Encoding') == 'br'
|
||||||
assert res.read() == b'<html><video src="/vid.mp4" /></html>'
|
assert res.read() == b'<html><video src="/vid.mp4" /></html>'
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_deflate(self, handler):
|
def test_deflate(self, handler):
|
||||||
with handler() as rh:
|
with handler() as rh:
|
||||||
res = validate_and_send(
|
res = validate_and_send(
|
||||||
@@ -572,7 +573,7 @@ class TestHTTPRequestHandler(TestRequestHandlerBase):
|
|||||||
assert res.headers.get('Content-Encoding') == 'deflate'
|
assert res.headers.get('Content-Encoding') == 'deflate'
|
||||||
assert res.read() == b'<html><video src="/vid.mp4" /></html>'
|
assert res.read() == b'<html><video src="/vid.mp4" /></html>'
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_gzip(self, handler):
|
def test_gzip(self, handler):
|
||||||
with handler() as rh:
|
with handler() as rh:
|
||||||
res = validate_and_send(
|
res = validate_and_send(
|
||||||
@@ -582,7 +583,7 @@ class TestHTTPRequestHandler(TestRequestHandlerBase):
|
|||||||
assert res.headers.get('Content-Encoding') == 'gzip'
|
assert res.headers.get('Content-Encoding') == 'gzip'
|
||||||
assert res.read() == b'<html><video src="/vid.mp4" /></html>'
|
assert res.read() == b'<html><video src="/vid.mp4" /></html>'
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_multiple_encodings(self, handler):
|
def test_multiple_encodings(self, handler):
|
||||||
with handler() as rh:
|
with handler() as rh:
|
||||||
for pair in ('gzip,deflate', 'deflate, gzip', 'gzip, gzip', 'deflate, deflate'):
|
for pair in ('gzip,deflate', 'deflate, gzip', 'gzip, gzip', 'deflate, deflate'):
|
||||||
@@ -593,7 +594,7 @@ class TestHTTPRequestHandler(TestRequestHandlerBase):
|
|||||||
assert res.headers.get('Content-Encoding') == pair
|
assert res.headers.get('Content-Encoding') == pair
|
||||||
assert res.read() == b'<html><video src="/vid.mp4" /></html>'
|
assert res.read() == b'<html><video src="/vid.mp4" /></html>'
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_unsupported_encoding(self, handler):
|
def test_unsupported_encoding(self, handler):
|
||||||
with handler() as rh:
|
with handler() as rh:
|
||||||
res = validate_and_send(
|
res = validate_and_send(
|
||||||
@@ -603,7 +604,7 @@ class TestHTTPRequestHandler(TestRequestHandlerBase):
|
|||||||
assert res.headers.get('Content-Encoding') == 'unsupported'
|
assert res.headers.get('Content-Encoding') == 'unsupported'
|
||||||
assert res.read() == b'raw'
|
assert res.read() == b'raw'
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_read(self, handler):
|
def test_read(self, handler):
|
||||||
with handler() as rh:
|
with handler() as rh:
|
||||||
res = validate_and_send(
|
res = validate_and_send(
|
||||||
@@ -633,7 +634,7 @@ class TestHTTPProxy(TestRequestHandlerBase):
|
|||||||
cls.geo_proxy_thread.daemon = True
|
cls.geo_proxy_thread.daemon = True
|
||||||
cls.geo_proxy_thread.start()
|
cls.geo_proxy_thread.start()
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_http_proxy(self, handler):
|
def test_http_proxy(self, handler):
|
||||||
http_proxy = f'http://127.0.0.1:{self.proxy_port}'
|
http_proxy = f'http://127.0.0.1:{self.proxy_port}'
|
||||||
geo_proxy = f'http://127.0.0.1:{self.geo_port}'
|
geo_proxy = f'http://127.0.0.1:{self.geo_port}'
|
||||||
@@ -659,7 +660,7 @@ class TestHTTPProxy(TestRequestHandlerBase):
|
|||||||
assert res != f'normal: {real_url}'
|
assert res != f'normal: {real_url}'
|
||||||
assert 'Accept' in res
|
assert 'Accept' in res
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_noproxy(self, handler):
|
def test_noproxy(self, handler):
|
||||||
with handler(proxies={'proxy': f'http://127.0.0.1:{self.proxy_port}'}) as rh:
|
with handler(proxies={'proxy': f'http://127.0.0.1:{self.proxy_port}'}) as rh:
|
||||||
# NO_PROXY
|
# NO_PROXY
|
||||||
@@ -669,7 +670,7 @@ class TestHTTPProxy(TestRequestHandlerBase):
|
|||||||
'utf-8')
|
'utf-8')
|
||||||
assert 'Accept' in nop_response
|
assert 'Accept' in nop_response
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_allproxy(self, handler):
|
def test_allproxy(self, handler):
|
||||||
url = 'http://foo.com/bar'
|
url = 'http://foo.com/bar'
|
||||||
with handler() as rh:
|
with handler() as rh:
|
||||||
@@ -677,7 +678,7 @@ class TestHTTPProxy(TestRequestHandlerBase):
|
|||||||
'utf-8')
|
'utf-8')
|
||||||
assert response == f'normal: {url}'
|
assert response == f'normal: {url}'
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_http_proxy_with_idn(self, handler):
|
def test_http_proxy_with_idn(self, handler):
|
||||||
with handler(proxies={
|
with handler(proxies={
|
||||||
'http': f'http://127.0.0.1:{self.proxy_port}',
|
'http': f'http://127.0.0.1:{self.proxy_port}',
|
||||||
@@ -715,27 +716,27 @@ class TestClientCertificate:
|
|||||||
) as rh:
|
) as rh:
|
||||||
validate_and_send(rh, Request(f'https://127.0.0.1:{self.port}/video.html')).read().decode()
|
validate_and_send(rh, Request(f'https://127.0.0.1:{self.port}/video.html')).read().decode()
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_certificate_combined_nopass(self, handler):
|
def test_certificate_combined_nopass(self, handler):
|
||||||
self._run_test(handler, client_cert={
|
self._run_test(handler, client_cert={
|
||||||
'client_certificate': os.path.join(self.certdir, 'clientwithkey.crt'),
|
'client_certificate': os.path.join(self.certdir, 'clientwithkey.crt'),
|
||||||
})
|
})
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_certificate_nocombined_nopass(self, handler):
|
def test_certificate_nocombined_nopass(self, handler):
|
||||||
self._run_test(handler, client_cert={
|
self._run_test(handler, client_cert={
|
||||||
'client_certificate': os.path.join(self.certdir, 'client.crt'),
|
'client_certificate': os.path.join(self.certdir, 'client.crt'),
|
||||||
'client_certificate_key': os.path.join(self.certdir, 'client.key'),
|
'client_certificate_key': os.path.join(self.certdir, 'client.key'),
|
||||||
})
|
})
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_certificate_combined_pass(self, handler):
|
def test_certificate_combined_pass(self, handler):
|
||||||
self._run_test(handler, client_cert={
|
self._run_test(handler, client_cert={
|
||||||
'client_certificate': os.path.join(self.certdir, 'clientwithencryptedkey.crt'),
|
'client_certificate': os.path.join(self.certdir, 'clientwithencryptedkey.crt'),
|
||||||
'client_certificate_password': 'foobar',
|
'client_certificate_password': 'foobar',
|
||||||
})
|
})
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_certificate_nocombined_pass(self, handler):
|
def test_certificate_nocombined_pass(self, handler):
|
||||||
self._run_test(handler, client_cert={
|
self._run_test(handler, client_cert={
|
||||||
'client_certificate': os.path.join(self.certdir, 'client.crt'),
|
'client_certificate': os.path.join(self.certdir, 'client.crt'),
|
||||||
@@ -819,6 +820,75 @@ class TestUrllibRequestHandler(TestRequestHandlerBase):
|
|||||||
assert not isinstance(exc_info.value, TransportError)
|
assert not isinstance(exc_info.value, TransportError)
|
||||||
|
|
||||||
|
|
||||||
|
class TestRequestsRequestHandler(TestRequestHandlerBase):
|
||||||
|
@pytest.mark.parametrize('raised,expected', [
|
||||||
|
(lambda: requests.exceptions.ConnectTimeout(), TransportError),
|
||||||
|
(lambda: requests.exceptions.ReadTimeout(), TransportError),
|
||||||
|
(lambda: requests.exceptions.Timeout(), TransportError),
|
||||||
|
(lambda: requests.exceptions.ConnectionError(), TransportError),
|
||||||
|
(lambda: requests.exceptions.ProxyError(), ProxyError),
|
||||||
|
(lambda: requests.exceptions.SSLError('12[CERTIFICATE_VERIFY_FAILED]34'), CertificateVerifyError),
|
||||||
|
(lambda: requests.exceptions.SSLError(), SSLError),
|
||||||
|
(lambda: requests.exceptions.InvalidURL(), RequestError),
|
||||||
|
(lambda: requests.exceptions.InvalidHeader(), RequestError),
|
||||||
|
# catch-all: https://github.com/psf/requests/blob/main/src/requests/adapters.py#L535
|
||||||
|
(lambda: urllib3.exceptions.HTTPError(), TransportError),
|
||||||
|
(lambda: requests.exceptions.RequestException(), RequestError)
|
||||||
|
# (lambda: requests.exceptions.TooManyRedirects(), HTTPError) - Needs a response object
|
||||||
|
])
|
||||||
|
@pytest.mark.parametrize('handler', ['Requests'], indirect=True)
|
||||||
|
def test_request_error_mapping(self, handler, monkeypatch, raised, expected):
|
||||||
|
with handler() as rh:
|
||||||
|
def mock_get_instance(*args, **kwargs):
|
||||||
|
class MockSession:
|
||||||
|
def request(self, *args, **kwargs):
|
||||||
|
raise raised()
|
||||||
|
return MockSession()
|
||||||
|
|
||||||
|
monkeypatch.setattr(rh, '_get_instance', mock_get_instance)
|
||||||
|
|
||||||
|
with pytest.raises(expected) as exc_info:
|
||||||
|
rh.send(Request('http://fake'))
|
||||||
|
|
||||||
|
assert exc_info.type is expected
|
||||||
|
|
||||||
|
@pytest.mark.parametrize('raised,expected,match', [
|
||||||
|
(lambda: urllib3.exceptions.SSLError(), SSLError, None),
|
||||||
|
(lambda: urllib3.exceptions.TimeoutError(), TransportError, None),
|
||||||
|
(lambda: urllib3.exceptions.ReadTimeoutError(None, None, None), TransportError, None),
|
||||||
|
(lambda: urllib3.exceptions.ProtocolError(), TransportError, None),
|
||||||
|
(lambda: urllib3.exceptions.DecodeError(), TransportError, None),
|
||||||
|
(lambda: urllib3.exceptions.HTTPError(), TransportError, None), # catch-all
|
||||||
|
(
|
||||||
|
lambda: urllib3.exceptions.ProtocolError('error', http.client.IncompleteRead(partial=b'abc', expected=4)),
|
||||||
|
IncompleteRead,
|
||||||
|
'3 bytes read, 4 more expected'
|
||||||
|
),
|
||||||
|
(
|
||||||
|
lambda: urllib3.exceptions.ProtocolError('error', urllib3.exceptions.IncompleteRead(partial=3, expected=5)),
|
||||||
|
IncompleteRead,
|
||||||
|
'3 bytes read, 5 more expected'
|
||||||
|
),
|
||||||
|
])
|
||||||
|
@pytest.mark.parametrize('handler', ['Requests'], indirect=True)
|
||||||
|
def test_response_error_mapping(self, handler, monkeypatch, raised, expected, match):
|
||||||
|
from urllib3.response import HTTPResponse as Urllib3Response
|
||||||
|
from requests.models import Response as RequestsResponse
|
||||||
|
from yt_dlp.networking._requests import RequestsResponseAdapter
|
||||||
|
requests_res = RequestsResponse()
|
||||||
|
requests_res.raw = Urllib3Response(body=b'', status=200)
|
||||||
|
res = RequestsResponseAdapter(requests_res)
|
||||||
|
|
||||||
|
def mock_read(*args, **kwargs):
|
||||||
|
raise raised()
|
||||||
|
monkeypatch.setattr(res.fp, 'read', mock_read)
|
||||||
|
|
||||||
|
with pytest.raises(expected, match=match) as exc_info:
|
||||||
|
res.read()
|
||||||
|
|
||||||
|
assert exc_info.type is expected
|
||||||
|
|
||||||
|
|
||||||
def run_validation(handler, error, req, **handler_kwargs):
|
def run_validation(handler, error, req, **handler_kwargs):
|
||||||
with handler(**handler_kwargs) as rh:
|
with handler(**handler_kwargs) as rh:
|
||||||
if error:
|
if error:
|
||||||
@@ -855,6 +925,10 @@ class TestRequestHandlerValidation:
|
|||||||
('file', UnsupportedRequest, {}),
|
('file', UnsupportedRequest, {}),
|
||||||
('file', False, {'enable_file_urls': True}),
|
('file', False, {'enable_file_urls': True}),
|
||||||
]),
|
]),
|
||||||
|
('Requests', [
|
||||||
|
('http', False, {}),
|
||||||
|
('https', False, {}),
|
||||||
|
]),
|
||||||
(NoCheckRH, [('http', False, {})]),
|
(NoCheckRH, [('http', False, {})]),
|
||||||
(ValidationRH, [('http', UnsupportedRequest, {})])
|
(ValidationRH, [('http', UnsupportedRequest, {})])
|
||||||
]
|
]
|
||||||
@@ -870,6 +944,14 @@ class TestRequestHandlerValidation:
|
|||||||
('socks5h', False),
|
('socks5h', False),
|
||||||
('socks', UnsupportedRequest),
|
('socks', UnsupportedRequest),
|
||||||
]),
|
]),
|
||||||
|
('Requests', [
|
||||||
|
('http', False),
|
||||||
|
('https', False),
|
||||||
|
('socks4', False),
|
||||||
|
('socks4a', False),
|
||||||
|
('socks5', False),
|
||||||
|
('socks5h', False),
|
||||||
|
]),
|
||||||
(NoCheckRH, [('http', False)]),
|
(NoCheckRH, [('http', False)]),
|
||||||
(HTTPSupportedRH, [('http', UnsupportedRequest)]),
|
(HTTPSupportedRH, [('http', UnsupportedRequest)]),
|
||||||
]
|
]
|
||||||
@@ -880,6 +962,10 @@ class TestRequestHandlerValidation:
|
|||||||
('all', False),
|
('all', False),
|
||||||
('unrelated', False),
|
('unrelated', False),
|
||||||
]),
|
]),
|
||||||
|
('Requests', [
|
||||||
|
('all', False),
|
||||||
|
('unrelated', False),
|
||||||
|
]),
|
||||||
(NoCheckRH, [('all', False)]),
|
(NoCheckRH, [('all', False)]),
|
||||||
(HTTPSupportedRH, [('all', UnsupportedRequest)]),
|
(HTTPSupportedRH, [('all', UnsupportedRequest)]),
|
||||||
(HTTPSupportedRH, [('no', UnsupportedRequest)]),
|
(HTTPSupportedRH, [('no', UnsupportedRequest)]),
|
||||||
@@ -894,6 +980,13 @@ class TestRequestHandlerValidation:
|
|||||||
({'timeout': 'notatimeout'}, AssertionError),
|
({'timeout': 'notatimeout'}, AssertionError),
|
||||||
({'unsupported': 'value'}, UnsupportedRequest),
|
({'unsupported': 'value'}, UnsupportedRequest),
|
||||||
]),
|
]),
|
||||||
|
('Requests', [
|
||||||
|
({'cookiejar': 'notacookiejar'}, AssertionError),
|
||||||
|
({'cookiejar': YoutubeDLCookieJar()}, False),
|
||||||
|
({'timeout': 1}, False),
|
||||||
|
({'timeout': 'notatimeout'}, AssertionError),
|
||||||
|
({'unsupported': 'value'}, UnsupportedRequest),
|
||||||
|
]),
|
||||||
(NoCheckRH, [
|
(NoCheckRH, [
|
||||||
({'cookiejar': 'notacookiejar'}, False),
|
({'cookiejar': 'notacookiejar'}, False),
|
||||||
({'somerandom': 'test'}, False), # but any extension is allowed through
|
({'somerandom': 'test'}, False), # but any extension is allowed through
|
||||||
@@ -909,7 +1002,7 @@ class TestRequestHandlerValidation:
|
|||||||
def test_url_scheme(self, handler, scheme, fail, handler_kwargs):
|
def test_url_scheme(self, handler, scheme, fail, handler_kwargs):
|
||||||
run_validation(handler, fail, Request(f'{scheme}://'), **(handler_kwargs or {}))
|
run_validation(handler, fail, Request(f'{scheme}://'), **(handler_kwargs or {}))
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler,fail', [('Urllib', False)], indirect=['handler'])
|
@pytest.mark.parametrize('handler,fail', [('Urllib', False), ('Requests', False)], indirect=['handler'])
|
||||||
def test_no_proxy(self, handler, fail):
|
def test_no_proxy(self, handler, fail):
|
||||||
run_validation(handler, fail, Request('http://', proxies={'no': '127.0.0.1,github.com'}))
|
run_validation(handler, fail, Request('http://', proxies={'no': '127.0.0.1,github.com'}))
|
||||||
run_validation(handler, fail, Request('http://'), proxies={'no': '127.0.0.1,github.com'})
|
run_validation(handler, fail, Request('http://'), proxies={'no': '127.0.0.1,github.com'})
|
||||||
@@ -932,13 +1025,13 @@ class TestRequestHandlerValidation:
|
|||||||
run_validation(handler, fail, Request('http://', proxies={'http': f'{scheme}://example.com'}))
|
run_validation(handler, fail, Request('http://', proxies={'http': f'{scheme}://example.com'}))
|
||||||
run_validation(handler, fail, Request('http://'), proxies={'http': f'{scheme}://example.com'})
|
run_validation(handler, fail, Request('http://'), proxies={'http': f'{scheme}://example.com'})
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler', ['Urllib', HTTPSupportedRH], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', HTTPSupportedRH, 'Requests'], indirect=True)
|
||||||
def test_empty_proxy(self, handler):
|
def test_empty_proxy(self, handler):
|
||||||
run_validation(handler, False, Request('http://', proxies={'http': None}))
|
run_validation(handler, False, Request('http://', proxies={'http': None}))
|
||||||
run_validation(handler, False, Request('http://'), proxies={'http': None})
|
run_validation(handler, False, Request('http://'), proxies={'http': None})
|
||||||
|
|
||||||
@pytest.mark.parametrize('proxy_url', ['//example.com', 'example.com', '127.0.0.1', '/a/b/c'])
|
@pytest.mark.parametrize('proxy_url', ['//example.com', 'example.com', '127.0.0.1', '/a/b/c'])
|
||||||
@pytest.mark.parametrize('handler', ['Urllib'], indirect=True)
|
@pytest.mark.parametrize('handler', ['Urllib', 'Requests'], indirect=True)
|
||||||
def test_invalid_proxy_url(self, handler, proxy_url):
|
def test_invalid_proxy_url(self, handler, proxy_url):
|
||||||
run_validation(handler, UnsupportedRequest, Request('http://', proxies={'http': proxy_url}))
|
run_validation(handler, UnsupportedRequest, Request('http://', proxies={'http': proxy_url}))
|
||||||
|
|
||||||
@@ -1200,6 +1293,10 @@ class TestYoutubeDLNetworking:
|
|||||||
assert 'Youtubedl-no-compression' not in rh.headers
|
assert 'Youtubedl-no-compression' not in rh.headers
|
||||||
assert rh.headers.get('Accept-Encoding') == 'identity'
|
assert rh.headers.get('Accept-Encoding') == 'identity'
|
||||||
|
|
||||||
|
with FakeYDL({'http_headers': {'Ytdl-socks-proxy': 'socks://localhost:1080'}}) as ydl:
|
||||||
|
rh = self.build_handler(ydl)
|
||||||
|
assert 'Ytdl-socks-proxy' not in rh.headers
|
||||||
|
|
||||||
def test_build_handler_params(self):
|
def test_build_handler_params(self):
|
||||||
with FakeYDL({
|
with FakeYDL({
|
||||||
'http_headers': {'test': 'testtest'},
|
'http_headers': {'test': 'testtest'},
|
||||||
@@ -1242,6 +1339,13 @@ class TestYoutubeDLNetworking:
|
|||||||
rh = self.build_handler(ydl, UrllibRH)
|
rh = self.build_handler(ydl, UrllibRH)
|
||||||
assert rh.enable_file_urls is True
|
assert rh.enable_file_urls is True
|
||||||
|
|
||||||
|
def test_compat_opt_prefer_urllib(self):
|
||||||
|
# This assumes urllib only has a preference when this compat opt is given
|
||||||
|
with FakeYDL({'compat_opts': ['prefer-legacy-http-handler']}) as ydl:
|
||||||
|
director = ydl.build_request_director([UrllibRH])
|
||||||
|
assert len(director.preferences) == 1
|
||||||
|
assert director.preferences.pop()(UrllibRH, None)
|
||||||
|
|
||||||
|
|
||||||
class TestRequest:
|
class TestRequest:
|
||||||
|
|
||||||
|
|||||||
@@ -263,7 +263,7 @@ def ctx(request):
|
|||||||
|
|
||||||
|
|
||||||
class TestSocks4Proxy:
|
class TestSocks4Proxy:
|
||||||
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http')], indirect=True)
|
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http'), ('Requests', 'http')], indirect=True)
|
||||||
def test_socks4_no_auth(self, handler, ctx):
|
def test_socks4_no_auth(self, handler, ctx):
|
||||||
with handler() as rh:
|
with handler() as rh:
|
||||||
with ctx.socks_server(Socks4ProxyHandler) as server_address:
|
with ctx.socks_server(Socks4ProxyHandler) as server_address:
|
||||||
@@ -271,7 +271,7 @@ class TestSocks4Proxy:
|
|||||||
rh, proxies={'all': f'socks4://{server_address}'})
|
rh, proxies={'all': f'socks4://{server_address}'})
|
||||||
assert response['version'] == 4
|
assert response['version'] == 4
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http')], indirect=True)
|
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http'), ('Requests', 'http')], indirect=True)
|
||||||
def test_socks4_auth(self, handler, ctx):
|
def test_socks4_auth(self, handler, ctx):
|
||||||
with handler() as rh:
|
with handler() as rh:
|
||||||
with ctx.socks_server(Socks4ProxyHandler, user_id='user') as server_address:
|
with ctx.socks_server(Socks4ProxyHandler, user_id='user') as server_address:
|
||||||
@@ -281,7 +281,7 @@ class TestSocks4Proxy:
|
|||||||
rh, proxies={'all': f'socks4://user:@{server_address}'})
|
rh, proxies={'all': f'socks4://user:@{server_address}'})
|
||||||
assert response['version'] == 4
|
assert response['version'] == 4
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http')], indirect=True)
|
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http'), ('Requests', 'http')], indirect=True)
|
||||||
def test_socks4a_ipv4_target(self, handler, ctx):
|
def test_socks4a_ipv4_target(self, handler, ctx):
|
||||||
with ctx.socks_server(Socks4ProxyHandler) as server_address:
|
with ctx.socks_server(Socks4ProxyHandler) as server_address:
|
||||||
with handler(proxies={'all': f'socks4a://{server_address}'}) as rh:
|
with handler(proxies={'all': f'socks4a://{server_address}'}) as rh:
|
||||||
@@ -289,7 +289,7 @@ class TestSocks4Proxy:
|
|||||||
assert response['version'] == 4
|
assert response['version'] == 4
|
||||||
assert (response['ipv4_address'] == '127.0.0.1') != (response['domain_address'] == '127.0.0.1')
|
assert (response['ipv4_address'] == '127.0.0.1') != (response['domain_address'] == '127.0.0.1')
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http')], indirect=True)
|
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http'), ('Requests', 'http')], indirect=True)
|
||||||
def test_socks4a_domain_target(self, handler, ctx):
|
def test_socks4a_domain_target(self, handler, ctx):
|
||||||
with ctx.socks_server(Socks4ProxyHandler) as server_address:
|
with ctx.socks_server(Socks4ProxyHandler) as server_address:
|
||||||
with handler(proxies={'all': f'socks4a://{server_address}'}) as rh:
|
with handler(proxies={'all': f'socks4a://{server_address}'}) as rh:
|
||||||
@@ -298,7 +298,7 @@ class TestSocks4Proxy:
|
|||||||
assert response['ipv4_address'] is None
|
assert response['ipv4_address'] is None
|
||||||
assert response['domain_address'] == 'localhost'
|
assert response['domain_address'] == 'localhost'
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http')], indirect=True)
|
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http'), ('Requests', 'http')], indirect=True)
|
||||||
def test_ipv4_client_source_address(self, handler, ctx):
|
def test_ipv4_client_source_address(self, handler, ctx):
|
||||||
with ctx.socks_server(Socks4ProxyHandler) as server_address:
|
with ctx.socks_server(Socks4ProxyHandler) as server_address:
|
||||||
source_address = f'127.0.0.{random.randint(5, 255)}'
|
source_address = f'127.0.0.{random.randint(5, 255)}'
|
||||||
@@ -308,7 +308,7 @@ class TestSocks4Proxy:
|
|||||||
assert response['client_address'][0] == source_address
|
assert response['client_address'][0] == source_address
|
||||||
assert response['version'] == 4
|
assert response['version'] == 4
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http')], indirect=True)
|
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http'), ('Requests', 'http')], indirect=True)
|
||||||
@pytest.mark.parametrize('reply_code', [
|
@pytest.mark.parametrize('reply_code', [
|
||||||
Socks4CD.REQUEST_REJECTED_OR_FAILED,
|
Socks4CD.REQUEST_REJECTED_OR_FAILED,
|
||||||
Socks4CD.REQUEST_REJECTED_CANNOT_CONNECT_TO_IDENTD,
|
Socks4CD.REQUEST_REJECTED_CANNOT_CONNECT_TO_IDENTD,
|
||||||
@@ -320,7 +320,7 @@ class TestSocks4Proxy:
|
|||||||
with pytest.raises(ProxyError):
|
with pytest.raises(ProxyError):
|
||||||
ctx.socks_info_request(rh)
|
ctx.socks_info_request(rh)
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http')], indirect=True)
|
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http'), ('Requests', 'http')], indirect=True)
|
||||||
def test_ipv6_socks4_proxy(self, handler, ctx):
|
def test_ipv6_socks4_proxy(self, handler, ctx):
|
||||||
with ctx.socks_server(Socks4ProxyHandler, bind_ip='::1') as server_address:
|
with ctx.socks_server(Socks4ProxyHandler, bind_ip='::1') as server_address:
|
||||||
with handler(proxies={'all': f'socks4://{server_address}'}) as rh:
|
with handler(proxies={'all': f'socks4://{server_address}'}) as rh:
|
||||||
@@ -329,7 +329,7 @@ class TestSocks4Proxy:
|
|||||||
assert response['ipv4_address'] == '127.0.0.1'
|
assert response['ipv4_address'] == '127.0.0.1'
|
||||||
assert response['version'] == 4
|
assert response['version'] == 4
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http')], indirect=True)
|
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http'), ('Requests', 'http')], indirect=True)
|
||||||
def test_timeout(self, handler, ctx):
|
def test_timeout(self, handler, ctx):
|
||||||
with ctx.socks_server(Socks4ProxyHandler, sleep=2) as server_address:
|
with ctx.socks_server(Socks4ProxyHandler, sleep=2) as server_address:
|
||||||
with handler(proxies={'all': f'socks4://{server_address}'}, timeout=0.5) as rh:
|
with handler(proxies={'all': f'socks4://{server_address}'}, timeout=0.5) as rh:
|
||||||
@@ -339,7 +339,7 @@ class TestSocks4Proxy:
|
|||||||
|
|
||||||
class TestSocks5Proxy:
|
class TestSocks5Proxy:
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http')], indirect=True)
|
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http'), ('Requests', 'http')], indirect=True)
|
||||||
def test_socks5_no_auth(self, handler, ctx):
|
def test_socks5_no_auth(self, handler, ctx):
|
||||||
with ctx.socks_server(Socks5ProxyHandler) as server_address:
|
with ctx.socks_server(Socks5ProxyHandler) as server_address:
|
||||||
with handler(proxies={'all': f'socks5://{server_address}'}) as rh:
|
with handler(proxies={'all': f'socks5://{server_address}'}) as rh:
|
||||||
@@ -347,7 +347,7 @@ class TestSocks5Proxy:
|
|||||||
assert response['auth_methods'] == [0x0]
|
assert response['auth_methods'] == [0x0]
|
||||||
assert response['version'] == 5
|
assert response['version'] == 5
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http')], indirect=True)
|
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http'), ('Requests', 'http')], indirect=True)
|
||||||
def test_socks5_user_pass(self, handler, ctx):
|
def test_socks5_user_pass(self, handler, ctx):
|
||||||
with ctx.socks_server(Socks5ProxyHandler, auth=('test', 'testpass')) as server_address:
|
with ctx.socks_server(Socks5ProxyHandler, auth=('test', 'testpass')) as server_address:
|
||||||
with handler() as rh:
|
with handler() as rh:
|
||||||
@@ -360,7 +360,7 @@ class TestSocks5Proxy:
|
|||||||
assert response['auth_methods'] == [Socks5Auth.AUTH_NONE, Socks5Auth.AUTH_USER_PASS]
|
assert response['auth_methods'] == [Socks5Auth.AUTH_NONE, Socks5Auth.AUTH_USER_PASS]
|
||||||
assert response['version'] == 5
|
assert response['version'] == 5
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http')], indirect=True)
|
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http'), ('Requests', 'http')], indirect=True)
|
||||||
def test_socks5_ipv4_target(self, handler, ctx):
|
def test_socks5_ipv4_target(self, handler, ctx):
|
||||||
with ctx.socks_server(Socks5ProxyHandler) as server_address:
|
with ctx.socks_server(Socks5ProxyHandler) as server_address:
|
||||||
with handler(proxies={'all': f'socks5://{server_address}'}) as rh:
|
with handler(proxies={'all': f'socks5://{server_address}'}) as rh:
|
||||||
@@ -368,7 +368,7 @@ class TestSocks5Proxy:
|
|||||||
assert response['ipv4_address'] == '127.0.0.1'
|
assert response['ipv4_address'] == '127.0.0.1'
|
||||||
assert response['version'] == 5
|
assert response['version'] == 5
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http')], indirect=True)
|
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http'), ('Requests', 'http')], indirect=True)
|
||||||
def test_socks5_domain_target(self, handler, ctx):
|
def test_socks5_domain_target(self, handler, ctx):
|
||||||
with ctx.socks_server(Socks5ProxyHandler) as server_address:
|
with ctx.socks_server(Socks5ProxyHandler) as server_address:
|
||||||
with handler(proxies={'all': f'socks5://{server_address}'}) as rh:
|
with handler(proxies={'all': f'socks5://{server_address}'}) as rh:
|
||||||
@@ -376,7 +376,7 @@ class TestSocks5Proxy:
|
|||||||
assert (response['ipv4_address'] == '127.0.0.1') != (response['ipv6_address'] == '::1')
|
assert (response['ipv4_address'] == '127.0.0.1') != (response['ipv6_address'] == '::1')
|
||||||
assert response['version'] == 5
|
assert response['version'] == 5
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http')], indirect=True)
|
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http'), ('Requests', 'http')], indirect=True)
|
||||||
def test_socks5h_domain_target(self, handler, ctx):
|
def test_socks5h_domain_target(self, handler, ctx):
|
||||||
with ctx.socks_server(Socks5ProxyHandler) as server_address:
|
with ctx.socks_server(Socks5ProxyHandler) as server_address:
|
||||||
with handler(proxies={'all': f'socks5h://{server_address}'}) as rh:
|
with handler(proxies={'all': f'socks5h://{server_address}'}) as rh:
|
||||||
@@ -385,7 +385,7 @@ class TestSocks5Proxy:
|
|||||||
assert response['domain_address'] == 'localhost'
|
assert response['domain_address'] == 'localhost'
|
||||||
assert response['version'] == 5
|
assert response['version'] == 5
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http')], indirect=True)
|
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http'), ('Requests', 'http')], indirect=True)
|
||||||
def test_socks5h_ip_target(self, handler, ctx):
|
def test_socks5h_ip_target(self, handler, ctx):
|
||||||
with ctx.socks_server(Socks5ProxyHandler) as server_address:
|
with ctx.socks_server(Socks5ProxyHandler) as server_address:
|
||||||
with handler(proxies={'all': f'socks5h://{server_address}'}) as rh:
|
with handler(proxies={'all': f'socks5h://{server_address}'}) as rh:
|
||||||
@@ -394,7 +394,7 @@ class TestSocks5Proxy:
|
|||||||
assert response['domain_address'] is None
|
assert response['domain_address'] is None
|
||||||
assert response['version'] == 5
|
assert response['version'] == 5
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http')], indirect=True)
|
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http'), ('Requests', 'http')], indirect=True)
|
||||||
def test_socks5_ipv6_destination(self, handler, ctx):
|
def test_socks5_ipv6_destination(self, handler, ctx):
|
||||||
with ctx.socks_server(Socks5ProxyHandler) as server_address:
|
with ctx.socks_server(Socks5ProxyHandler) as server_address:
|
||||||
with handler(proxies={'all': f'socks5://{server_address}'}) as rh:
|
with handler(proxies={'all': f'socks5://{server_address}'}) as rh:
|
||||||
@@ -402,7 +402,7 @@ class TestSocks5Proxy:
|
|||||||
assert response['ipv6_address'] == '::1'
|
assert response['ipv6_address'] == '::1'
|
||||||
assert response['version'] == 5
|
assert response['version'] == 5
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http')], indirect=True)
|
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http'), ('Requests', 'http')], indirect=True)
|
||||||
def test_ipv6_socks5_proxy(self, handler, ctx):
|
def test_ipv6_socks5_proxy(self, handler, ctx):
|
||||||
with ctx.socks_server(Socks5ProxyHandler, bind_ip='::1') as server_address:
|
with ctx.socks_server(Socks5ProxyHandler, bind_ip='::1') as server_address:
|
||||||
with handler(proxies={'all': f'socks5://{server_address}'}) as rh:
|
with handler(proxies={'all': f'socks5://{server_address}'}) as rh:
|
||||||
@@ -413,7 +413,7 @@ class TestSocks5Proxy:
|
|||||||
|
|
||||||
# XXX: is there any feasible way of testing IPv6 source addresses?
|
# XXX: is there any feasible way of testing IPv6 source addresses?
|
||||||
# Same would go for non-proxy source_address test...
|
# Same would go for non-proxy source_address test...
|
||||||
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http')], indirect=True)
|
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http'), ('Requests', 'http')], indirect=True)
|
||||||
def test_ipv4_client_source_address(self, handler, ctx):
|
def test_ipv4_client_source_address(self, handler, ctx):
|
||||||
with ctx.socks_server(Socks5ProxyHandler) as server_address:
|
with ctx.socks_server(Socks5ProxyHandler) as server_address:
|
||||||
source_address = f'127.0.0.{random.randint(5, 255)}'
|
source_address = f'127.0.0.{random.randint(5, 255)}'
|
||||||
@@ -422,7 +422,7 @@ class TestSocks5Proxy:
|
|||||||
assert response['client_address'][0] == source_address
|
assert response['client_address'][0] == source_address
|
||||||
assert response['version'] == 5
|
assert response['version'] == 5
|
||||||
|
|
||||||
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http')], indirect=True)
|
@pytest.mark.parametrize('handler,ctx', [('Urllib', 'http'), ('Requests', 'http')], indirect=True)
|
||||||
@pytest.mark.parametrize('reply_code', [
|
@pytest.mark.parametrize('reply_code', [
|
||||||
Socks5Reply.GENERAL_FAILURE,
|
Socks5Reply.GENERAL_FAILURE,
|
||||||
Socks5Reply.CONNECTION_NOT_ALLOWED,
|
Socks5Reply.CONNECTION_NOT_ALLOWED,
|
||||||
|
|||||||
199
test/test_update.py
Normal file
199
test/test_update.py
Normal file
@@ -0,0 +1,199 @@
|
|||||||
|
#!/usr/bin/env python3
|
||||||
|
|
||||||
|
# Allow direct execution
|
||||||
|
import os
|
||||||
|
import sys
|
||||||
|
import unittest
|
||||||
|
|
||||||
|
sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
|
||||||
|
|
||||||
|
|
||||||
|
from test.helper import FakeYDL, report_warning
|
||||||
|
from yt_dlp.update import Updater, UpdateInfo
|
||||||
|
|
||||||
|
TEST_API_DATA = {
|
||||||
|
'yt-dlp/yt-dlp/latest': {
|
||||||
|
'tag_name': '2023.12.31',
|
||||||
|
'target_commitish': 'bbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbb',
|
||||||
|
'name': 'yt-dlp 2023.12.31',
|
||||||
|
'body': 'BODY',
|
||||||
|
},
|
||||||
|
'yt-dlp/yt-dlp-nightly-builds/latest': {
|
||||||
|
'tag_name': '2023.12.31.123456',
|
||||||
|
'target_commitish': 'master',
|
||||||
|
'name': 'yt-dlp nightly 2023.12.31.123456',
|
||||||
|
'body': 'Generated from: https://github.com/yt-dlp/yt-dlp/commit/cccccccccccccccccccccccccccccccccccccccc',
|
||||||
|
},
|
||||||
|
'yt-dlp/yt-dlp-master-builds/latest': {
|
||||||
|
'tag_name': '2023.12.31.987654',
|
||||||
|
'target_commitish': 'master',
|
||||||
|
'name': 'yt-dlp master 2023.12.31.987654',
|
||||||
|
'body': 'Generated from: https://github.com/yt-dlp/yt-dlp/commit/dddddddddddddddddddddddddddddddddddddddd',
|
||||||
|
},
|
||||||
|
'yt-dlp/yt-dlp/tags/testing': {
|
||||||
|
'tag_name': 'testing',
|
||||||
|
'target_commitish': '9999999999999999999999999999999999999999',
|
||||||
|
'name': 'testing',
|
||||||
|
'body': 'BODY',
|
||||||
|
},
|
||||||
|
'fork/yt-dlp/latest': {
|
||||||
|
'tag_name': '2050.12.31',
|
||||||
|
'target_commitish': 'eeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeee',
|
||||||
|
'name': '2050.12.31',
|
||||||
|
'body': 'BODY',
|
||||||
|
},
|
||||||
|
'fork/yt-dlp/tags/pr0000': {
|
||||||
|
'tag_name': 'pr0000',
|
||||||
|
'target_commitish': 'ffffffffffffffffffffffffffffffffffffffff',
|
||||||
|
'name': 'pr1234 2023.11.11.000000',
|
||||||
|
'body': 'BODY',
|
||||||
|
},
|
||||||
|
'fork/yt-dlp/tags/pr1234': {
|
||||||
|
'tag_name': 'pr1234',
|
||||||
|
'target_commitish': '0000000000000000000000000000000000000000',
|
||||||
|
'name': 'pr1234 2023.12.31.555555',
|
||||||
|
'body': 'BODY',
|
||||||
|
},
|
||||||
|
'fork/yt-dlp/tags/pr9999': {
|
||||||
|
'tag_name': 'pr9999',
|
||||||
|
'target_commitish': '1111111111111111111111111111111111111111',
|
||||||
|
'name': 'pr9999',
|
||||||
|
'body': 'BODY',
|
||||||
|
},
|
||||||
|
'fork/yt-dlp-satellite/tags/pr987': {
|
||||||
|
'tag_name': 'pr987',
|
||||||
|
'target_commitish': 'master',
|
||||||
|
'name': 'pr987',
|
||||||
|
'body': 'Generated from: https://github.com/yt-dlp/yt-dlp/commit/2222222222222222222222222222222222222222',
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
TEST_LOCKFILE_V1 = '''# This file is used for regulating self-update
|
||||||
|
lock 2022.08.18.36 .+ Python 3.6
|
||||||
|
lock 2023.11.13 .+ Python 3.7
|
||||||
|
'''
|
||||||
|
|
||||||
|
TEST_LOCKFILE_V2 = '''# This file is used for regulating self-update
|
||||||
|
lockV2 yt-dlp/yt-dlp 2022.08.18.36 .+ Python 3.6
|
||||||
|
lockV2 yt-dlp/yt-dlp 2023.11.13 .+ Python 3.7
|
||||||
|
'''
|
||||||
|
|
||||||
|
TEST_LOCKFILE_V1_V2 = '''# This file is used for regulating self-update
|
||||||
|
lock 2022.08.18.36 .+ Python 3.6
|
||||||
|
lock 2023.11.13 .+ Python 3.7
|
||||||
|
lockV2 yt-dlp/yt-dlp 2022.08.18.36 .+ Python 3.6
|
||||||
|
lockV2 yt-dlp/yt-dlp 2023.11.13 .+ Python 3.7
|
||||||
|
lockV2 fork/yt-dlp pr0000 .+ Python 3.6
|
||||||
|
lockV2 fork/yt-dlp pr1234 .+ Python 3.7
|
||||||
|
lockV2 fork/yt-dlp pr9999 .+ Python 3.11
|
||||||
|
'''
|
||||||
|
|
||||||
|
|
||||||
|
class FakeUpdater(Updater):
|
||||||
|
current_version = '2022.01.01'
|
||||||
|
current_commit = 'aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa'
|
||||||
|
|
||||||
|
_channel = 'stable'
|
||||||
|
_origin = 'yt-dlp/yt-dlp'
|
||||||
|
|
||||||
|
def _download_update_spec(self, *args, **kwargs):
|
||||||
|
return TEST_LOCKFILE_V1_V2
|
||||||
|
|
||||||
|
def _call_api(self, tag):
|
||||||
|
tag = f'tags/{tag}' if tag != 'latest' else tag
|
||||||
|
return TEST_API_DATA[f'{self.requested_repo}/{tag}']
|
||||||
|
|
||||||
|
def _report_error(self, msg, *args, **kwargs):
|
||||||
|
report_warning(msg)
|
||||||
|
|
||||||
|
|
||||||
|
class TestUpdate(unittest.TestCase):
|
||||||
|
maxDiff = None
|
||||||
|
|
||||||
|
def test_update_spec(self):
|
||||||
|
ydl = FakeYDL()
|
||||||
|
updater = FakeUpdater(ydl, 'stable@latest')
|
||||||
|
|
||||||
|
def test(lockfile, identifier, input_tag, expect_tag, exact=False, repo='yt-dlp/yt-dlp'):
|
||||||
|
updater._identifier = identifier
|
||||||
|
updater._exact = exact
|
||||||
|
updater.requested_repo = repo
|
||||||
|
result = updater._process_update_spec(lockfile, input_tag)
|
||||||
|
self.assertEqual(
|
||||||
|
result, expect_tag,
|
||||||
|
f'{identifier!r} requesting {repo}@{input_tag} (exact={exact}) '
|
||||||
|
f'returned {result!r} instead of {expect_tag!r}')
|
||||||
|
|
||||||
|
test(TEST_LOCKFILE_V1, 'zip Python 3.11.0', '2023.11.13', '2023.11.13')
|
||||||
|
test(TEST_LOCKFILE_V1, 'zip stable Python 3.11.0', '2023.11.13', '2023.11.13', exact=True)
|
||||||
|
test(TEST_LOCKFILE_V1, 'zip Python 3.6.0', '2023.11.13', '2022.08.18.36')
|
||||||
|
test(TEST_LOCKFILE_V1, 'zip stable Python 3.6.0', '2023.11.13', None, exact=True)
|
||||||
|
test(TEST_LOCKFILE_V1, 'zip Python 3.7.0', '2023.11.13', '2023.11.13')
|
||||||
|
test(TEST_LOCKFILE_V1, 'zip stable Python 3.7.1', '2023.11.13', '2023.11.13')
|
||||||
|
test(TEST_LOCKFILE_V1, 'zip Python 3.7.1', '2023.12.31', '2023.11.13')
|
||||||
|
test(TEST_LOCKFILE_V1, 'zip stable Python 3.7.1', '2023.12.31', '2023.11.13')
|
||||||
|
|
||||||
|
test(TEST_LOCKFILE_V2, 'zip Python 3.11.1', '2023.11.13', '2023.11.13')
|
||||||
|
test(TEST_LOCKFILE_V2, 'zip stable Python 3.11.1', '2023.12.31', '2023.12.31')
|
||||||
|
test(TEST_LOCKFILE_V2, 'zip Python 3.6.1', '2023.11.13', '2022.08.18.36')
|
||||||
|
test(TEST_LOCKFILE_V2, 'zip stable Python 3.7.2', '2023.11.13', '2023.11.13')
|
||||||
|
test(TEST_LOCKFILE_V2, 'zip Python 3.7.2', '2023.12.31', '2023.11.13')
|
||||||
|
|
||||||
|
test(TEST_LOCKFILE_V1_V2, 'zip Python 3.11.2', '2023.11.13', '2023.11.13')
|
||||||
|
test(TEST_LOCKFILE_V1_V2, 'zip stable Python 3.11.2', '2023.12.31', '2023.12.31')
|
||||||
|
test(TEST_LOCKFILE_V1_V2, 'zip Python 3.6.2', '2023.11.13', '2022.08.18.36')
|
||||||
|
test(TEST_LOCKFILE_V1_V2, 'zip stable Python 3.7.3', '2023.11.13', '2023.11.13')
|
||||||
|
test(TEST_LOCKFILE_V1_V2, 'zip Python 3.7.3', '2023.12.31', '2023.11.13')
|
||||||
|
test(TEST_LOCKFILE_V1_V2, 'zip Python 3.6.3', 'pr0000', None, repo='fork/yt-dlp')
|
||||||
|
test(TEST_LOCKFILE_V1_V2, 'zip stable Python 3.7.4', 'pr0000', 'pr0000', repo='fork/yt-dlp')
|
||||||
|
test(TEST_LOCKFILE_V1_V2, 'zip Python 3.6.4', 'pr0000', None, repo='fork/yt-dlp')
|
||||||
|
test(TEST_LOCKFILE_V1_V2, 'zip Python 3.7.4', 'pr1234', None, repo='fork/yt-dlp')
|
||||||
|
test(TEST_LOCKFILE_V1_V2, 'zip stable Python 3.8.1', 'pr1234', 'pr1234', repo='fork/yt-dlp')
|
||||||
|
test(TEST_LOCKFILE_V1_V2, 'zip Python 3.7.5', 'pr1234', None, repo='fork/yt-dlp')
|
||||||
|
test(TEST_LOCKFILE_V1_V2, 'zip Python 3.11.3', 'pr9999', None, repo='fork/yt-dlp')
|
||||||
|
test(TEST_LOCKFILE_V1_V2, 'zip stable Python 3.12.0', 'pr9999', 'pr9999', repo='fork/yt-dlp')
|
||||||
|
test(TEST_LOCKFILE_V1_V2, 'zip Python 3.11.4', 'pr9999', None, repo='fork/yt-dlp')
|
||||||
|
|
||||||
|
def test_query_update(self):
|
||||||
|
ydl = FakeYDL()
|
||||||
|
|
||||||
|
def test(target, expected, current_version=None, current_commit=None, identifier=None):
|
||||||
|
updater = FakeUpdater(ydl, target)
|
||||||
|
if current_version:
|
||||||
|
updater.current_version = current_version
|
||||||
|
if current_commit:
|
||||||
|
updater.current_commit = current_commit
|
||||||
|
updater._identifier = identifier or 'zip'
|
||||||
|
update_info = updater.query_update(_output=True)
|
||||||
|
self.assertDictEqual(
|
||||||
|
update_info.__dict__ if update_info else {}, expected.__dict__ if expected else {})
|
||||||
|
|
||||||
|
test('yt-dlp/yt-dlp@latest', UpdateInfo(
|
||||||
|
'2023.12.31', version='2023.12.31', requested_version='2023.12.31', commit='b' * 40))
|
||||||
|
test('yt-dlp/yt-dlp-nightly-builds@latest', UpdateInfo(
|
||||||
|
'2023.12.31.123456', version='2023.12.31.123456', requested_version='2023.12.31.123456', commit='c' * 40))
|
||||||
|
test('yt-dlp/yt-dlp-master-builds@latest', UpdateInfo(
|
||||||
|
'2023.12.31.987654', version='2023.12.31.987654', requested_version='2023.12.31.987654', commit='d' * 40))
|
||||||
|
test('fork/yt-dlp@latest', UpdateInfo(
|
||||||
|
'2050.12.31', version='2050.12.31', requested_version='2050.12.31', commit='e' * 40))
|
||||||
|
test('fork/yt-dlp@pr0000', UpdateInfo(
|
||||||
|
'pr0000', version='2023.11.11.000000', requested_version='2023.11.11.000000', commit='f' * 40))
|
||||||
|
test('fork/yt-dlp@pr1234', UpdateInfo(
|
||||||
|
'pr1234', version='2023.12.31.555555', requested_version='2023.12.31.555555', commit='0' * 40))
|
||||||
|
test('fork/yt-dlp@pr9999', UpdateInfo(
|
||||||
|
'pr9999', version=None, requested_version=None, commit='1' * 40))
|
||||||
|
test('fork/yt-dlp-satellite@pr987', UpdateInfo(
|
||||||
|
'pr987', version=None, requested_version=None, commit='2' * 40))
|
||||||
|
test('yt-dlp/yt-dlp', None, current_version='2024.01.01')
|
||||||
|
test('stable', UpdateInfo(
|
||||||
|
'2023.12.31', version='2023.12.31', requested_version='2023.12.31', commit='b' * 40))
|
||||||
|
test('nightly', UpdateInfo(
|
||||||
|
'2023.12.31.123456', version='2023.12.31.123456', requested_version='2023.12.31.123456', commit='c' * 40))
|
||||||
|
test('master', UpdateInfo(
|
||||||
|
'2023.12.31.987654', version='2023.12.31.987654', requested_version='2023.12.31.987654', commit='d' * 40))
|
||||||
|
test('testing', None, current_commit='9' * 40)
|
||||||
|
test('testing', UpdateInfo('testing', commit='9' * 40))
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == '__main__':
|
||||||
|
unittest.main()
|
||||||
@@ -1,30 +0,0 @@
|
|||||||
#!/usr/bin/env python3
|
|
||||||
|
|
||||||
# Allow direct execution
|
|
||||||
import os
|
|
||||||
import sys
|
|
||||||
import unittest
|
|
||||||
|
|
||||||
sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
|
|
||||||
|
|
||||||
|
|
||||||
import json
|
|
||||||
|
|
||||||
from yt_dlp.update import rsa_verify
|
|
||||||
|
|
||||||
|
|
||||||
class TestUpdate(unittest.TestCase):
|
|
||||||
def test_rsa_verify(self):
|
|
||||||
UPDATES_RSA_KEY = (0x9d60ee4d8f805312fdb15a62f87b95bd66177b91df176765d13514a0f1754bcd2057295c5b6f1d35daa6742c3ffc9a82d3e118861c207995a8031e151d863c9927e304576bc80692bc8e094896fcf11b66f3e29e04e3a71e9a11558558acea1840aec37fc396fb6b65dc81a1c4144e03bd1c011de62e3f1357b327d08426fe93, 65537)
|
|
||||||
with open(os.path.join(os.path.dirname(os.path.abspath(__file__)), 'versions.json'), 'rb') as f:
|
|
||||||
versions_info = f.read().decode()
|
|
||||||
versions_info = json.loads(versions_info)
|
|
||||||
signature = versions_info['signature']
|
|
||||||
del versions_info['signature']
|
|
||||||
self.assertTrue(rsa_verify(
|
|
||||||
json.dumps(versions_info, sort_keys=True).encode(),
|
|
||||||
signature, UPDATES_RSA_KEY))
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
|
||||||
unittest.main()
|
|
||||||
@@ -1,34 +0,0 @@
|
|||||||
{
|
|
||||||
"latest": "2013.01.06",
|
|
||||||
"signature": "72158cdba391628569ffdbea259afbcf279bbe3d8aeb7492690735dc1cfa6afa754f55c61196f3871d429599ab22f2667f1fec98865527b32632e7f4b3675a7ef0f0fbe084d359256ae4bba68f0d33854e531a70754712f244be71d4b92e664302aa99653ee4df19800d955b6c4149cd2b3f24288d6e4b40b16126e01f4c8ce6",
|
|
||||||
"versions": {
|
|
||||||
"2013.01.02": {
|
|
||||||
"bin": [
|
|
||||||
"http://youtube-dl.org/downloads/2013.01.02/youtube-dl",
|
|
||||||
"f5b502f8aaa77675c4884938b1e4871ebca2611813a0c0e74f60c0fbd6dcca6b"
|
|
||||||
],
|
|
||||||
"exe": [
|
|
||||||
"http://youtube-dl.org/downloads/2013.01.02/youtube-dl.exe",
|
|
||||||
"75fa89d2ce297d102ff27675aa9d92545bbc91013f52ec52868c069f4f9f0422"
|
|
||||||
],
|
|
||||||
"tar": [
|
|
||||||
"http://youtube-dl.org/downloads/2013.01.02/youtube-dl-2013.01.02.tar.gz",
|
|
||||||
"6a66d022ac8e1c13da284036288a133ec8dba003b7bd3a5179d0c0daca8c8196"
|
|
||||||
]
|
|
||||||
},
|
|
||||||
"2013.01.06": {
|
|
||||||
"bin": [
|
|
||||||
"http://youtube-dl.org/downloads/2013.01.06/youtube-dl",
|
|
||||||
"64b6ed8865735c6302e836d4d832577321b4519aa02640dc508580c1ee824049"
|
|
||||||
],
|
|
||||||
"exe": [
|
|
||||||
"http://youtube-dl.org/downloads/2013.01.06/youtube-dl.exe",
|
|
||||||
"58609baf91e4389d36e3ba586e21dab882daaaee537e4448b1265392ae86ff84"
|
|
||||||
],
|
|
||||||
"tar": [
|
|
||||||
"http://youtube-dl.org/downloads/2013.01.06/youtube-dl-2013.01.06.tar.gz",
|
|
||||||
"fe77ab20a95d980ed17a659aa67e371fdd4d656d19c4c7950e7b720b0c2f1a86"
|
|
||||||
]
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -60,7 +60,7 @@ from .postprocessor import (
|
|||||||
get_postprocessor,
|
get_postprocessor,
|
||||||
)
|
)
|
||||||
from .postprocessor.ffmpeg import resolve_mapping as resolve_recode_mapping
|
from .postprocessor.ffmpeg import resolve_mapping as resolve_recode_mapping
|
||||||
from .update import REPOSITORY, _get_system_deprecation, current_git_head, detect_variant
|
from .update import REPOSITORY, _get_system_deprecation, _make_label, current_git_head, detect_variant
|
||||||
from .utils import (
|
from .utils import (
|
||||||
DEFAULT_OUTTMPL,
|
DEFAULT_OUTTMPL,
|
||||||
IDENTITY,
|
IDENTITY,
|
||||||
@@ -158,7 +158,7 @@ from .utils.networking import (
|
|||||||
clean_proxies,
|
clean_proxies,
|
||||||
std_headers,
|
std_headers,
|
||||||
)
|
)
|
||||||
from .version import CHANNEL, RELEASE_GIT_HEAD, VARIANT, __version__
|
from .version import CHANNEL, ORIGIN, RELEASE_GIT_HEAD, VARIANT, __version__
|
||||||
|
|
||||||
if compat_os_name == 'nt':
|
if compat_os_name == 'nt':
|
||||||
import ctypes
|
import ctypes
|
||||||
@@ -2338,7 +2338,7 @@ class YoutubeDL:
|
|||||||
return
|
return
|
||||||
|
|
||||||
for f in formats:
|
for f in formats:
|
||||||
if f.get('has_drm'):
|
if f.get('has_drm') or f.get('__needs_testing'):
|
||||||
yield from self._check_formats([f])
|
yield from self._check_formats([f])
|
||||||
else:
|
else:
|
||||||
yield f
|
yield f
|
||||||
@@ -2764,7 +2764,8 @@ class YoutubeDL:
|
|||||||
format['dynamic_range'] = 'SDR'
|
format['dynamic_range'] = 'SDR'
|
||||||
if format.get('aspect_ratio') is None:
|
if format.get('aspect_ratio') is None:
|
||||||
format['aspect_ratio'] = try_call(lambda: round(format['width'] / format['height'], 2))
|
format['aspect_ratio'] = try_call(lambda: round(format['width'] / format['height'], 2))
|
||||||
if (not format.get('manifest_url') # For fragmented formats, "tbr" is often max bitrate and not average
|
# For fragmented formats, "tbr" is often max bitrate and not average
|
||||||
|
if (('manifest-filesize-approx' in self.params['compat_opts'] or not format.get('manifest_url'))
|
||||||
and info_dict.get('duration') and format.get('tbr')
|
and info_dict.get('duration') and format.get('tbr')
|
||||||
and not format.get('filesize') and not format.get('filesize_approx')):
|
and not format.get('filesize') and not format.get('filesize_approx')):
|
||||||
format['filesize_approx'] = int(info_dict['duration'] * format['tbr'] * (1024 / 8))
|
format['filesize_approx'] = int(info_dict['duration'] * format['tbr'] * (1024 / 8))
|
||||||
@@ -3543,14 +3544,14 @@ class YoutubeDL:
|
|||||||
'version': __version__,
|
'version': __version__,
|
||||||
'current_git_head': current_git_head(),
|
'current_git_head': current_git_head(),
|
||||||
'release_git_head': RELEASE_GIT_HEAD,
|
'release_git_head': RELEASE_GIT_HEAD,
|
||||||
'repository': REPOSITORY,
|
'repository': ORIGIN,
|
||||||
})
|
})
|
||||||
|
|
||||||
if remove_private_keys:
|
if remove_private_keys:
|
||||||
reject = lambda k, v: v is None or k.startswith('__') or k in {
|
reject = lambda k, v: v is None or k.startswith('__') or k in {
|
||||||
'requested_downloads', 'requested_formats', 'requested_subtitles', 'requested_entries',
|
'requested_downloads', 'requested_formats', 'requested_subtitles', 'requested_entries',
|
||||||
'entries', 'filepath', '_filename', 'filename', 'infojson_filename', 'original_url',
|
'entries', 'filepath', '_filename', 'filename', 'infojson_filename', 'original_url',
|
||||||
'playlist_autonumber', '_format_sort_fields',
|
'playlist_autonumber',
|
||||||
}
|
}
|
||||||
else:
|
else:
|
||||||
reject = lambda k, v: False
|
reject = lambda k, v: False
|
||||||
@@ -3926,8 +3927,8 @@ class YoutubeDL:
|
|||||||
source += '*'
|
source += '*'
|
||||||
klass = type(self)
|
klass = type(self)
|
||||||
write_debug(join_nonempty(
|
write_debug(join_nonempty(
|
||||||
f'{"yt-dlp" if REPOSITORY == "yt-dlp/yt-dlp" else REPOSITORY} version',
|
f'{REPOSITORY.rpartition("/")[2]} version',
|
||||||
f'{CHANNEL}@{__version__}',
|
_make_label(ORIGIN, CHANNEL.partition('@')[2] or __version__, __version__),
|
||||||
f'[{RELEASE_GIT_HEAD[:9]}]' if RELEASE_GIT_HEAD else '',
|
f'[{RELEASE_GIT_HEAD[:9]}]' if RELEASE_GIT_HEAD else '',
|
||||||
'' if source == 'unknown' else f'({source})',
|
'' if source == 'unknown' else f'({source})',
|
||||||
'' if _IN_CLI else 'API' if klass == YoutubeDL else f'API:{self.__module__}.{klass.__qualname__}',
|
'' if _IN_CLI else 'API' if klass == YoutubeDL else f'API:{self.__module__}.{klass.__qualname__}',
|
||||||
@@ -3968,7 +3969,7 @@ class YoutubeDL:
|
|||||||
})) or 'none'))
|
})) or 'none'))
|
||||||
|
|
||||||
write_debug(f'Proxy map: {self.proxies}')
|
write_debug(f'Proxy map: {self.proxies}')
|
||||||
# write_debug(f'Request Handlers: {", ".join(rh.RH_NAME for rh in self._request_director.handlers.values())}')
|
write_debug(f'Request Handlers: {", ".join(rh.RH_NAME for rh in self._request_director.handlers.values())}')
|
||||||
for plugin_type, plugins in {'Extractor': plugin_ies, 'Post-Processor': plugin_pps}.items():
|
for plugin_type, plugins in {'Extractor': plugin_ies, 'Post-Processor': plugin_pps}.items():
|
||||||
display_list = ['%s%s' % (
|
display_list = ['%s%s' % (
|
||||||
klass.__name__, '' if klass.__name__ == name else f' as {name}')
|
klass.__name__, '' if klass.__name__ == name else f' as {name}')
|
||||||
@@ -4057,6 +4058,9 @@ class YoutubeDL:
|
|||||||
raise RequestError(
|
raise RequestError(
|
||||||
'file:// URLs are disabled by default in yt-dlp for security reasons. '
|
'file:// URLs are disabled by default in yt-dlp for security reasons. '
|
||||||
'Use --enable-file-urls to enable at your own risk.', cause=ue) from ue
|
'Use --enable-file-urls to enable at your own risk.', cause=ue) from ue
|
||||||
|
if 'unsupported proxy type: "https"' in ue.msg.lower():
|
||||||
|
raise RequestError(
|
||||||
|
'To use an HTTPS proxy for this request, one of the following dependencies needs to be installed: requests')
|
||||||
raise
|
raise
|
||||||
except SSLError as e:
|
except SSLError as e:
|
||||||
if 'UNSAFE_LEGACY_RENEGOTIATION_DISABLED' in str(e):
|
if 'UNSAFE_LEGACY_RENEGOTIATION_DISABLED' in str(e):
|
||||||
@@ -4099,6 +4103,8 @@ class YoutubeDL:
|
|||||||
}),
|
}),
|
||||||
))
|
))
|
||||||
director.preferences.update(preferences or [])
|
director.preferences.update(preferences or [])
|
||||||
|
if 'prefer-legacy-http-handler' in self.params['compat_opts']:
|
||||||
|
director.preferences.add(lambda rh, _: 500 if rh.RH_KEY == 'Urllib' else 0)
|
||||||
return director
|
return director
|
||||||
|
|
||||||
def encode(self, s):
|
def encode(self, s):
|
||||||
@@ -4237,7 +4243,7 @@ class YoutubeDL:
|
|||||||
self.write_debug(f'Skipping writing {label} thumbnail')
|
self.write_debug(f'Skipping writing {label} thumbnail')
|
||||||
return ret
|
return ret
|
||||||
|
|
||||||
if not self._ensure_dir_exists(filename):
|
if thumbnails and not self._ensure_dir_exists(filename):
|
||||||
return None
|
return None
|
||||||
|
|
||||||
for idx, t in list(enumerate(thumbnails))[::-1]:
|
for idx, t in list(enumerate(thumbnails))[::-1]:
|
||||||
|
|||||||
@@ -21,9 +21,11 @@ def get_hidden_imports():
|
|||||||
yield from ('yt_dlp.compat._legacy', 'yt_dlp.compat._deprecated')
|
yield from ('yt_dlp.compat._legacy', 'yt_dlp.compat._deprecated')
|
||||||
yield from ('yt_dlp.utils._legacy', 'yt_dlp.utils._deprecated')
|
yield from ('yt_dlp.utils._legacy', 'yt_dlp.utils._deprecated')
|
||||||
yield pycryptodome_module()
|
yield pycryptodome_module()
|
||||||
yield from collect_submodules('websockets')
|
# Only `websockets` is required, others are collected just in case
|
||||||
|
for module in ('websockets', 'requests', 'urllib3'):
|
||||||
|
yield from collect_submodules(module)
|
||||||
# These are auto-detected, but explicitly add them just in case
|
# These are auto-detected, but explicitly add them just in case
|
||||||
yield from ('mutagen', 'brotli', 'certifi')
|
yield from ('mutagen', 'brotli', 'certifi', 'secretstorage')
|
||||||
|
|
||||||
|
|
||||||
hiddenimports = list(get_hidden_imports())
|
hiddenimports = list(get_hidden_imports())
|
||||||
|
|||||||
@@ -58,6 +58,15 @@ except (ImportError, SyntaxError):
|
|||||||
# See https://github.com/yt-dlp/yt-dlp/issues/2633
|
# See https://github.com/yt-dlp/yt-dlp/issues/2633
|
||||||
websockets = None
|
websockets = None
|
||||||
|
|
||||||
|
try:
|
||||||
|
import urllib3
|
||||||
|
except ImportError:
|
||||||
|
urllib3 = None
|
||||||
|
|
||||||
|
try:
|
||||||
|
import requests
|
||||||
|
except ImportError:
|
||||||
|
requests = None
|
||||||
|
|
||||||
try:
|
try:
|
||||||
import xattr # xattr or pyxattr
|
import xattr # xattr or pyxattr
|
||||||
|
|||||||
@@ -15,12 +15,15 @@ class DashSegmentsFD(FragmentFD):
|
|||||||
FD_NAME = 'dashsegments'
|
FD_NAME = 'dashsegments'
|
||||||
|
|
||||||
def real_download(self, filename, info_dict):
|
def real_download(self, filename, info_dict):
|
||||||
if info_dict.get('is_live') and set(info_dict['protocol'].split('+')) != {'http_dash_segments_generator'}:
|
if 'http_dash_segments_generator' in info_dict['protocol'].split('+'):
|
||||||
self.report_error('Live DASH videos are not supported')
|
real_downloader = None # No external FD can support --live-from-start
|
||||||
|
else:
|
||||||
|
if info_dict.get('is_live'):
|
||||||
|
self.report_error('Live DASH videos are not supported')
|
||||||
|
real_downloader = get_suitable_downloader(
|
||||||
|
info_dict, self.params, None, protocol='dash_frag_urls', to_stdout=(filename == '-'))
|
||||||
|
|
||||||
real_start = time.time()
|
real_start = time.time()
|
||||||
real_downloader = get_suitable_downloader(
|
|
||||||
info_dict, self.params, None, protocol='dash_frag_urls', to_stdout=(filename == '-'))
|
|
||||||
|
|
||||||
requested_formats = [{**info_dict, **fmt} for fmt in info_dict.get('requested_formats', [])]
|
requested_formats = [{**info_dict, **fmt} for fmt in info_dict.get('requested_formats', [])]
|
||||||
args = []
|
args = []
|
||||||
|
|||||||
@@ -335,7 +335,7 @@ class Aria2cFD(ExternalFD):
|
|||||||
cmd += ['--auto-file-renaming=false']
|
cmd += ['--auto-file-renaming=false']
|
||||||
|
|
||||||
if 'fragments' in info_dict:
|
if 'fragments' in info_dict:
|
||||||
cmd += ['--file-allocation=none', '--uri-selector=inorder']
|
cmd += ['--uri-selector=inorder']
|
||||||
url_list_file = '%s.frag.urls' % tmpfilename
|
url_list_file = '%s.frag.urls' % tmpfilename
|
||||||
url_list = []
|
url_list = []
|
||||||
for frag_index, fragment in enumerate(info_dict['fragments']):
|
for frag_index, fragment in enumerate(info_dict['fragments']):
|
||||||
|
|||||||
@@ -565,6 +565,7 @@ from .ellentube import (
|
|||||||
)
|
)
|
||||||
from .elonet import ElonetIE
|
from .elonet import ElonetIE
|
||||||
from .elpais import ElPaisIE
|
from .elpais import ElPaisIE
|
||||||
|
from .eltrecetv import ElTreceTVIE
|
||||||
from .embedly import EmbedlyIE
|
from .embedly import EmbedlyIE
|
||||||
from .engadget import EngadgetIE
|
from .engadget import EngadgetIE
|
||||||
from .epicon import (
|
from .epicon import (
|
||||||
@@ -893,6 +894,10 @@ from .japandiet import (
|
|||||||
SangiinIE,
|
SangiinIE,
|
||||||
)
|
)
|
||||||
from .jeuxvideo import JeuxVideoIE
|
from .jeuxvideo import JeuxVideoIE
|
||||||
|
from .jiosaavn import (
|
||||||
|
JioSaavnSongIE,
|
||||||
|
JioSaavnAlbumIE,
|
||||||
|
)
|
||||||
from .jove import JoveIE
|
from .jove import JoveIE
|
||||||
from .joj import JojIE
|
from .joj import JojIE
|
||||||
from .jstream import JStreamIE
|
from .jstream import JStreamIE
|
||||||
@@ -953,6 +958,7 @@ from .lastfm import (
|
|||||||
LastFMPlaylistIE,
|
LastFMPlaylistIE,
|
||||||
LastFMUserIE,
|
LastFMUserIE,
|
||||||
)
|
)
|
||||||
|
from .laxarxames import LaXarxaMesIE
|
||||||
from .lbry import (
|
from .lbry import (
|
||||||
LBRYIE,
|
LBRYIE,
|
||||||
LBRYChannelIE,
|
LBRYChannelIE,
|
||||||
@@ -1319,7 +1325,6 @@ from .ninegag import NineGagIE
|
|||||||
from .ninenow import NineNowIE
|
from .ninenow import NineNowIE
|
||||||
from .nintendo import NintendoIE
|
from .nintendo import NintendoIE
|
||||||
from .nitter import NitterIE
|
from .nitter import NitterIE
|
||||||
from .njpwworld import NJPWWorldIE
|
|
||||||
from .nobelprize import NobelPrizeIE
|
from .nobelprize import NobelPrizeIE
|
||||||
from .noice import NoicePodcastIE
|
from .noice import NoicePodcastIE
|
||||||
from .nonktube import NonkTubeIE
|
from .nonktube import NonkTubeIE
|
||||||
@@ -1387,7 +1392,10 @@ from .oftv import (
|
|||||||
from .oktoberfesttv import OktoberfestTVIE
|
from .oktoberfesttv import OktoberfestTVIE
|
||||||
from .olympics import OlympicsReplayIE
|
from .olympics import OlympicsReplayIE
|
||||||
from .on24 import On24IE
|
from .on24 import On24IE
|
||||||
from .ondemandkorea import OnDemandKoreaIE
|
from .ondemandkorea import (
|
||||||
|
OnDemandKoreaIE,
|
||||||
|
OnDemandKoreaProgramIE,
|
||||||
|
)
|
||||||
from .onefootball import OneFootballIE
|
from .onefootball import OneFootballIE
|
||||||
from .onenewsnz import OneNewsNZIE
|
from .onenewsnz import OneNewsNZIE
|
||||||
from .oneplace import OnePlacePodcastIE
|
from .oneplace import OnePlacePodcastIE
|
||||||
@@ -1416,6 +1424,7 @@ from .orf import (
|
|||||||
ORFTVthekIE,
|
ORFTVthekIE,
|
||||||
ORFFM4StoryIE,
|
ORFFM4StoryIE,
|
||||||
ORFRadioIE,
|
ORFRadioIE,
|
||||||
|
ORFPodcastIE,
|
||||||
ORFIPTVIE,
|
ORFIPTVIE,
|
||||||
)
|
)
|
||||||
from .outsidetv import OutsideTVIE
|
from .outsidetv import OutsideTVIE
|
||||||
@@ -1578,6 +1587,10 @@ from .radiocanada import (
|
|||||||
RadioCanadaIE,
|
RadioCanadaIE,
|
||||||
RadioCanadaAudioVideoIE,
|
RadioCanadaAudioVideoIE,
|
||||||
)
|
)
|
||||||
|
from .radiocomercial import (
|
||||||
|
RadioComercialIE,
|
||||||
|
RadioComercialPlaylistIE,
|
||||||
|
)
|
||||||
from .radiode import RadioDeIE
|
from .radiode import RadioDeIE
|
||||||
from .radiojavan import RadioJavanIE
|
from .radiojavan import RadioJavanIE
|
||||||
from .radiobremen import RadioBremenIE
|
from .radiobremen import RadioBremenIE
|
||||||
@@ -1758,6 +1771,11 @@ from .samplefocus import SampleFocusIE
|
|||||||
from .sapo import SapoIE
|
from .sapo import SapoIE
|
||||||
from .savefrom import SaveFromIE
|
from .savefrom import SaveFromIE
|
||||||
from .sbs import SBSIE
|
from .sbs import SBSIE
|
||||||
|
from .sbscokr import (
|
||||||
|
SBSCoKrIE,
|
||||||
|
SBSCoKrAllvodProgramIE,
|
||||||
|
SBSCoKrProgramsVodIE,
|
||||||
|
)
|
||||||
from .screen9 import Screen9IE
|
from .screen9 import Screen9IE
|
||||||
from .screencast import ScreencastIE
|
from .screencast import ScreencastIE
|
||||||
from .screencastify import ScreencastifyIE
|
from .screencastify import ScreencastifyIE
|
||||||
@@ -1902,6 +1920,8 @@ from .srmediathek import SRMediathekIE
|
|||||||
from .stacommu import (
|
from .stacommu import (
|
||||||
StacommuLiveIE,
|
StacommuLiveIE,
|
||||||
StacommuVODIE,
|
StacommuVODIE,
|
||||||
|
TheaterComplexTownVODIE,
|
||||||
|
TheaterComplexTownPPVIE,
|
||||||
)
|
)
|
||||||
from .stanfordoc import StanfordOpenClassroomIE
|
from .stanfordoc import StanfordOpenClassroomIE
|
||||||
from .startv import StarTVIE
|
from .startv import StarTVIE
|
||||||
@@ -2014,7 +2034,6 @@ from .thestar import TheStarIE
|
|||||||
from .thesun import TheSunIE
|
from .thesun import TheSunIE
|
||||||
from .theweatherchannel import TheWeatherChannelIE
|
from .theweatherchannel import TheWeatherChannelIE
|
||||||
from .thisamericanlife import ThisAmericanLifeIE
|
from .thisamericanlife import ThisAmericanLifeIE
|
||||||
from .thisav import ThisAVIE
|
|
||||||
from .thisoldhouse import ThisOldHouseIE
|
from .thisoldhouse import ThisOldHouseIE
|
||||||
from .thisvid import (
|
from .thisvid import (
|
||||||
ThisVidIE,
|
ThisVidIE,
|
||||||
|
|||||||
@@ -16,6 +16,7 @@ from ..utils import (
|
|||||||
try_get,
|
try_get,
|
||||||
unescapeHTML,
|
unescapeHTML,
|
||||||
update_url_query,
|
update_url_query,
|
||||||
|
url_or_none,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
@@ -379,6 +380,18 @@ class ABCIViewShowSeriesIE(InfoExtractor):
|
|||||||
'noplaylist': True,
|
'noplaylist': True,
|
||||||
'skip_download': 'm3u8',
|
'skip_download': 'm3u8',
|
||||||
},
|
},
|
||||||
|
}, {
|
||||||
|
# 'videoEpisodes' is a dict with `items` key
|
||||||
|
'url': 'https://iview.abc.net.au/show/7-30-mark-humphries-satire',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '178458-0',
|
||||||
|
'title': 'Episodes',
|
||||||
|
'description': 'Satirist Mark Humphries brings his unique perspective on current political events for 7.30.',
|
||||||
|
'series': '7.30 Mark Humphries Satire',
|
||||||
|
'season': 'Episodes',
|
||||||
|
'thumbnail': r're:^https?://cdn\.iview\.abc\.net\.au/thumbs/.*\.jpg$'
|
||||||
|
},
|
||||||
|
'playlist_count': 15,
|
||||||
}]
|
}]
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
@@ -398,12 +411,14 @@ class ABCIViewShowSeriesIE(InfoExtractor):
|
|||||||
series = video_data['selectedSeries']
|
series = video_data['selectedSeries']
|
||||||
return {
|
return {
|
||||||
'_type': 'playlist',
|
'_type': 'playlist',
|
||||||
'entries': [self.url_result(episode['shareUrl'])
|
'entries': [self.url_result(episode_url, ABCIViewIE)
|
||||||
for episode in series['_embedded']['videoEpisodes']],
|
for episode_url in traverse_obj(series, (
|
||||||
|
'_embedded', 'videoEpisodes', (None, 'items'), ..., 'shareUrl', {url_or_none}))],
|
||||||
'id': series.get('id'),
|
'id': series.get('id'),
|
||||||
'title': dict_get(series, ('title', 'displaySubtitle')),
|
'title': dict_get(series, ('title', 'displaySubtitle')),
|
||||||
'description': series.get('description'),
|
'description': series.get('description'),
|
||||||
'series': dict_get(series, ('showTitle', 'displayTitle')),
|
'series': dict_get(series, ('showTitle', 'displayTitle')),
|
||||||
'season': dict_get(series, ('title', 'displaySubtitle')),
|
'season': dict_get(series, ('title', 'displaySubtitle')),
|
||||||
'thumbnail': series.get('thumbnail'),
|
'thumbnail': traverse_obj(
|
||||||
|
series, 'thumbnail', ('images', lambda _, v: v['name'] == 'seriesThumbnail', 'url'), get_all=False),
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -3,14 +3,13 @@ from .youtube import YoutubeIE, YoutubeTabIE
|
|||||||
|
|
||||||
|
|
||||||
class BeatBumpVideoIE(InfoExtractor):
|
class BeatBumpVideoIE(InfoExtractor):
|
||||||
_VALID_URL = r'https://beatbump\.ml/listen\?id=(?P<id>[\w-]+)'
|
_VALID_URL = r'https://beatbump\.(?:ml|io)/listen\?id=(?P<id>[\w-]+)'
|
||||||
_TESTS = [{
|
_TESTS = [{
|
||||||
'url': 'https://beatbump.ml/listen?id=MgNrAu2pzNs',
|
'url': 'https://beatbump.ml/listen?id=MgNrAu2pzNs',
|
||||||
'md5': '5ff3fff41d3935b9810a9731e485fe66',
|
'md5': '5ff3fff41d3935b9810a9731e485fe66',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'MgNrAu2pzNs',
|
'id': 'MgNrAu2pzNs',
|
||||||
'ext': 'mp4',
|
'ext': 'mp4',
|
||||||
'uploader_url': 'http://www.youtube.com/channel/UC-pWHpBjdGG69N9mM2auIAA',
|
|
||||||
'artist': 'Stephen',
|
'artist': 'Stephen',
|
||||||
'thumbnail': 'https://i.ytimg.com/vi_webp/MgNrAu2pzNs/maxresdefault.webp',
|
'thumbnail': 'https://i.ytimg.com/vi_webp/MgNrAu2pzNs/maxresdefault.webp',
|
||||||
'channel_url': 'https://www.youtube.com/channel/UC-pWHpBjdGG69N9mM2auIAA',
|
'channel_url': 'https://www.youtube.com/channel/UC-pWHpBjdGG69N9mM2auIAA',
|
||||||
@@ -22,10 +21,9 @@ class BeatBumpVideoIE(InfoExtractor):
|
|||||||
'alt_title': 'Voyeur Girl',
|
'alt_title': 'Voyeur Girl',
|
||||||
'view_count': int,
|
'view_count': int,
|
||||||
'track': 'Voyeur Girl',
|
'track': 'Voyeur Girl',
|
||||||
'uploader': 'Stephen - Topic',
|
'uploader': 'Stephen',
|
||||||
'title': 'Voyeur Girl',
|
'title': 'Voyeur Girl',
|
||||||
'channel_follower_count': int,
|
'channel_follower_count': int,
|
||||||
'uploader_id': 'UC-pWHpBjdGG69N9mM2auIAA',
|
|
||||||
'age_limit': 0,
|
'age_limit': 0,
|
||||||
'availability': 'public',
|
'availability': 'public',
|
||||||
'live_status': 'not_live',
|
'live_status': 'not_live',
|
||||||
@@ -36,7 +34,12 @@ class BeatBumpVideoIE(InfoExtractor):
|
|||||||
'tags': 'count:11',
|
'tags': 'count:11',
|
||||||
'creator': 'Stephen',
|
'creator': 'Stephen',
|
||||||
'channel_id': 'UC-pWHpBjdGG69N9mM2auIAA',
|
'channel_id': 'UC-pWHpBjdGG69N9mM2auIAA',
|
||||||
}
|
'channel_is_verified': True,
|
||||||
|
'heatmap': 'count:100',
|
||||||
|
},
|
||||||
|
}, {
|
||||||
|
'url': 'https://beatbump.io/listen?id=LDGZAprNGWo',
|
||||||
|
'only_matching': True,
|
||||||
}]
|
}]
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
@@ -45,7 +48,7 @@ class BeatBumpVideoIE(InfoExtractor):
|
|||||||
|
|
||||||
|
|
||||||
class BeatBumpPlaylistIE(InfoExtractor):
|
class BeatBumpPlaylistIE(InfoExtractor):
|
||||||
_VALID_URL = r'https://beatbump\.ml/(?:release\?id=|artist/|playlist/)(?P<id>[\w-]+)'
|
_VALID_URL = r'https://beatbump\.(?:ml|io)/(?:release\?id=|artist/|playlist/)(?P<id>[\w-]+)'
|
||||||
_TESTS = [{
|
_TESTS = [{
|
||||||
'url': 'https://beatbump.ml/release?id=MPREb_gTAcphH99wE',
|
'url': 'https://beatbump.ml/release?id=MPREb_gTAcphH99wE',
|
||||||
'playlist_count': 50,
|
'playlist_count': 50,
|
||||||
@@ -56,25 +59,28 @@ class BeatBumpPlaylistIE(InfoExtractor):
|
|||||||
'title': 'Album - Royalty Free Music Library V2 (50 Songs)',
|
'title': 'Album - Royalty Free Music Library V2 (50 Songs)',
|
||||||
'description': '',
|
'description': '',
|
||||||
'tags': [],
|
'tags': [],
|
||||||
'modified_date': '20221223',
|
'modified_date': '20231110',
|
||||||
}
|
},
|
||||||
|
'expected_warnings': ['YouTube Music is not directly supported'],
|
||||||
}, {
|
}, {
|
||||||
'url': 'https://beatbump.ml/artist/UC_aEa8K-EOJ3D6gOs7HcyNg',
|
'url': 'https://beatbump.ml/artist/UC_aEa8K-EOJ3D6gOs7HcyNg',
|
||||||
'playlist_mincount': 1,
|
'playlist_mincount': 1,
|
||||||
'params': {'flatplaylist': True},
|
'params': {'flatplaylist': True},
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'UC_aEa8K-EOJ3D6gOs7HcyNg',
|
'id': 'UC_aEa8K-EOJ3D6gOs7HcyNg',
|
||||||
'uploader_url': 'https://www.youtube.com/channel/UC_aEa8K-EOJ3D6gOs7HcyNg',
|
'uploader_url': 'https://www.youtube.com/@NoCopyrightSounds',
|
||||||
'channel_url': 'https://www.youtube.com/channel/UC_aEa8K-EOJ3D6gOs7HcyNg',
|
'channel_url': 'https://www.youtube.com/channel/UC_aEa8K-EOJ3D6gOs7HcyNg',
|
||||||
'uploader_id': 'UC_aEa8K-EOJ3D6gOs7HcyNg',
|
'uploader_id': '@NoCopyrightSounds',
|
||||||
'channel_follower_count': int,
|
'channel_follower_count': int,
|
||||||
'title': 'NoCopyrightSounds - Videos',
|
'title': 'NoCopyrightSounds',
|
||||||
'uploader': 'NoCopyrightSounds',
|
'uploader': 'NoCopyrightSounds',
|
||||||
'description': 'md5:cd4fd53d81d363d05eee6c1b478b491a',
|
'description': 'md5:cd4fd53d81d363d05eee6c1b478b491a',
|
||||||
'channel': 'NoCopyrightSounds',
|
'channel': 'NoCopyrightSounds',
|
||||||
'tags': 'count:12',
|
'tags': 'count:65',
|
||||||
'channel_id': 'UC_aEa8K-EOJ3D6gOs7HcyNg',
|
'channel_id': 'UC_aEa8K-EOJ3D6gOs7HcyNg',
|
||||||
|
'channel_is_verified': True,
|
||||||
},
|
},
|
||||||
|
'expected_warnings': ['YouTube Music is not directly supported'],
|
||||||
}, {
|
}, {
|
||||||
'url': 'https://beatbump.ml/playlist/VLPLRBp0Fe2GpgmgoscNFLxNyBVSFVdYmFkq',
|
'url': 'https://beatbump.ml/playlist/VLPLRBp0Fe2GpgmgoscNFLxNyBVSFVdYmFkq',
|
||||||
'playlist_mincount': 1,
|
'playlist_mincount': 1,
|
||||||
@@ -84,16 +90,20 @@ class BeatBumpPlaylistIE(InfoExtractor):
|
|||||||
'uploader_url': 'https://www.youtube.com/@NoCopyrightSounds',
|
'uploader_url': 'https://www.youtube.com/@NoCopyrightSounds',
|
||||||
'description': 'Providing you with copyright free / safe music for gaming, live streaming, studying and more!',
|
'description': 'Providing you with copyright free / safe music for gaming, live streaming, studying and more!',
|
||||||
'view_count': int,
|
'view_count': int,
|
||||||
'channel_url': 'https://www.youtube.com/@NoCopyrightSounds',
|
'channel_url': 'https://www.youtube.com/channel/UC_aEa8K-EOJ3D6gOs7HcyNg',
|
||||||
'uploader_id': 'UC_aEa8K-EOJ3D6gOs7HcyNg',
|
'uploader_id': '@NoCopyrightSounds',
|
||||||
'title': 'NCS : All Releases 💿',
|
'title': 'NCS : All Releases 💿',
|
||||||
'uploader': 'NoCopyrightSounds',
|
'uploader': 'NoCopyrightSounds',
|
||||||
'availability': 'public',
|
'availability': 'public',
|
||||||
'channel': 'NoCopyrightSounds',
|
'channel': 'NoCopyrightSounds',
|
||||||
'tags': [],
|
'tags': [],
|
||||||
'modified_date': '20221225',
|
'modified_date': '20231112',
|
||||||
'channel_id': 'UC_aEa8K-EOJ3D6gOs7HcyNg',
|
'channel_id': 'UC_aEa8K-EOJ3D6gOs7HcyNg',
|
||||||
}
|
},
|
||||||
|
'expected_warnings': ['YouTube Music is not directly supported'],
|
||||||
|
}, {
|
||||||
|
'url': 'https://beatbump.io/playlist/VLPLFCHGavqRG-q_2ZhmgU2XB2--ZY6irT1c',
|
||||||
|
'only_matching': True,
|
||||||
}]
|
}]
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
|
|||||||
@@ -21,10 +21,10 @@ class BrilliantpalaBaseIE(InfoExtractor):
|
|||||||
|
|
||||||
def _get_logged_in_username(self, url, video_id):
|
def _get_logged_in_username(self, url, video_id):
|
||||||
webpage, urlh = self._download_webpage_handle(url, video_id)
|
webpage, urlh = self._download_webpage_handle(url, video_id)
|
||||||
if self._LOGIN_API == urlh.url:
|
if urlh.url.startswith(self._LOGIN_API):
|
||||||
self.raise_login_required()
|
self.raise_login_required()
|
||||||
return self._html_search_regex(
|
return self._html_search_regex(
|
||||||
r'"username"\s*:\s*"(?P<username>[^"]+)"', webpage, 'stream page info', 'username')
|
r'"username"\s*:\s*"(?P<username>[^"]+)"', webpage, 'logged-in username')
|
||||||
|
|
||||||
def _perform_login(self, username, password):
|
def _perform_login(self, username, password):
|
||||||
login_form = self._hidden_inputs(self._download_webpage(
|
login_form = self._hidden_inputs(self._download_webpage(
|
||||||
|
|||||||
@@ -1,8 +1,9 @@
|
|||||||
import re
|
|
||||||
import json
|
|
||||||
import base64
|
import base64
|
||||||
|
import json
|
||||||
|
import re
|
||||||
import time
|
import time
|
||||||
import urllib.parse
|
import urllib.parse
|
||||||
|
import xml.etree.ElementTree
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from ..compat import (
|
from ..compat import (
|
||||||
@@ -387,7 +388,7 @@ class CBCGemIE(InfoExtractor):
|
|||||||
url = re.sub(r'(Manifest\(.*?),format=[\w-]+(.*?\))', r'\1\2', base_url)
|
url = re.sub(r'(Manifest\(.*?),format=[\w-]+(.*?\))', r'\1\2', base_url)
|
||||||
|
|
||||||
secret_xml = self._download_xml(url, video_id, note='Downloading secret XML', fatal=False)
|
secret_xml = self._download_xml(url, video_id, note='Downloading secret XML', fatal=False)
|
||||||
if not secret_xml:
|
if not isinstance(secret_xml, xml.etree.ElementTree.Element):
|
||||||
return
|
return
|
||||||
|
|
||||||
for child in secret_xml:
|
for child in secret_xml:
|
||||||
|
|||||||
@@ -2225,7 +2225,9 @@ class InfoExtractor:
|
|||||||
mpd_url, video_id,
|
mpd_url, video_id,
|
||||||
note='Downloading MPD VOD manifest' if note is None else note,
|
note='Downloading MPD VOD manifest' if note is None else note,
|
||||||
errnote='Failed to download VOD manifest' if errnote is None else errnote,
|
errnote='Failed to download VOD manifest' if errnote is None else errnote,
|
||||||
fatal=False, data=data, headers=headers, query=query) or {}
|
fatal=False, data=data, headers=headers, query=query)
|
||||||
|
if not isinstance(mpd_doc, xml.etree.ElementTree.Element):
|
||||||
|
return None
|
||||||
return int_or_none(parse_duration(mpd_doc.get('mediaPresentationDuration')))
|
return int_or_none(parse_duration(mpd_doc.get('mediaPresentationDuration')))
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
|
|||||||
@@ -105,7 +105,7 @@ class CybraryIE(CybraryBaseIE):
|
|||||||
'chapter': module.get('title'),
|
'chapter': module.get('title'),
|
||||||
'chapter_id': str_or_none(module.get('id')),
|
'chapter_id': str_or_none(module.get('id')),
|
||||||
'title': activity.get('title'),
|
'title': activity.get('title'),
|
||||||
'url': smuggle_url(f'https://player.vimeo.com/video/{vimeo_id}', {'http_headers': {'Referer': 'https://api.cybrary.it'}})
|
'url': smuggle_url(f'https://player.vimeo.com/video/{vimeo_id}', {'referer': 'https://api.cybrary.it'})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -93,7 +93,7 @@ class DailymotionIE(DailymotionBaseInfoExtractor):
|
|||||||
_VALID_URL = r'''(?ix)
|
_VALID_URL = r'''(?ix)
|
||||||
https?://
|
https?://
|
||||||
(?:
|
(?:
|
||||||
(?:(?:www|touch|geo)\.)?dailymotion\.[a-z]{2,3}/(?:(?:(?:(?:embed|swf|\#)/)|player\.html\?)?video|swf)|
|
(?:(?:www|touch|geo)\.)?dailymotion\.[a-z]{2,3}/(?:(?:(?:(?:embed|swf|\#)/)|player(?:/\w+)?\.html\?)?video|swf)|
|
||||||
(?:www\.)?lequipe\.fr/video
|
(?:www\.)?lequipe\.fr/video
|
||||||
)
|
)
|
||||||
[/=](?P<id>[^/?_&]+)(?:.+?\bplaylist=(?P<playlist_id>x[0-9a-z]+))?
|
[/=](?P<id>[^/?_&]+)(?:.+?\bplaylist=(?P<playlist_id>x[0-9a-z]+))?
|
||||||
@@ -107,13 +107,17 @@ class DailymotionIE(DailymotionBaseInfoExtractor):
|
|||||||
'id': 'x5kesuj',
|
'id': 'x5kesuj',
|
||||||
'ext': 'mp4',
|
'ext': 'mp4',
|
||||||
'title': 'Office Christmas Party Review – Jason Bateman, Olivia Munn, T.J. Miller',
|
'title': 'Office Christmas Party Review – Jason Bateman, Olivia Munn, T.J. Miller',
|
||||||
'description': 'Office Christmas Party Review - Jason Bateman, Olivia Munn, T.J. Miller',
|
'description': 'Office Christmas Party Review - Jason Bateman, Olivia Munn, T.J. Miller',
|
||||||
'duration': 187,
|
'duration': 187,
|
||||||
'timestamp': 1493651285,
|
'timestamp': 1493651285,
|
||||||
'upload_date': '20170501',
|
'upload_date': '20170501',
|
||||||
'uploader': 'Deadline',
|
'uploader': 'Deadline',
|
||||||
'uploader_id': 'x1xm8ri',
|
'uploader_id': 'x1xm8ri',
|
||||||
'age_limit': 0,
|
'age_limit': 0,
|
||||||
|
'view_count': int,
|
||||||
|
'like_count': int,
|
||||||
|
'tags': ['hollywood', 'celeb', 'celebrity', 'movies', 'red carpet'],
|
||||||
|
'thumbnail': r're:https://(?:s[12]\.)dmcdn\.net/v/K456B1aXqIx58LKWQ/x1080',
|
||||||
},
|
},
|
||||||
}, {
|
}, {
|
||||||
'url': 'https://geo.dailymotion.com/player.html?video=x89eyek&mute=true',
|
'url': 'https://geo.dailymotion.com/player.html?video=x89eyek&mute=true',
|
||||||
@@ -132,7 +136,7 @@ class DailymotionIE(DailymotionBaseInfoExtractor):
|
|||||||
'view_count': int,
|
'view_count': int,
|
||||||
'like_count': int,
|
'like_count': int,
|
||||||
'tags': ['en_quete_d_esprit'],
|
'tags': ['en_quete_d_esprit'],
|
||||||
'thumbnail': 'https://s2.dmcdn.net/v/Tncwi1YGKdvFbDuDY/x1080',
|
'thumbnail': r're:https://(?:s[12]\.)dmcdn\.net/v/Tncwi1YNg_RUl7ueu/x1080',
|
||||||
}
|
}
|
||||||
}, {
|
}, {
|
||||||
'url': 'https://www.dailymotion.com/video/x2iuewm_steam-machine-models-pricing-listed-on-steam-store-ign-news_videogames',
|
'url': 'https://www.dailymotion.com/video/x2iuewm_steam-machine-models-pricing-listed-on-steam-store-ign-news_videogames',
|
||||||
@@ -201,6 +205,12 @@ class DailymotionIE(DailymotionBaseInfoExtractor):
|
|||||||
}, {
|
}, {
|
||||||
'url': 'https://www.dailymotion.com/video/x3z49k?playlist=xv4bw',
|
'url': 'https://www.dailymotion.com/video/x3z49k?playlist=xv4bw',
|
||||||
'only_matching': True,
|
'only_matching': True,
|
||||||
|
}, {
|
||||||
|
'url': 'https://geo.dailymotion.com/player/x86gw.html?video=k46oCapRs4iikoz9DWy',
|
||||||
|
'only_matching': True,
|
||||||
|
}, {
|
||||||
|
'url': 'https://geo.dailymotion.com/player/xakln.html?video=x8mjju4&customConfig%5BcustomParams%5D=%2Ffr-fr%2Ftennis%2Fwimbledon-mens-singles%2Farticles-video',
|
||||||
|
'only_matching': True,
|
||||||
}]
|
}]
|
||||||
_GEO_BYPASS = False
|
_GEO_BYPASS = False
|
||||||
_COMMON_MEDIA_FIELDS = '''description
|
_COMMON_MEDIA_FIELDS = '''description
|
||||||
|
|||||||
@@ -1,21 +1,17 @@
|
|||||||
import binascii
|
import json
|
||||||
import hashlib
|
import uuid
|
||||||
import re
|
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from ..aes import aes_cbc_decrypt_bytes, unpad_pkcs7
|
|
||||||
from ..compat import compat_urllib_parse_unquote
|
|
||||||
from ..utils import (
|
from ..utils import (
|
||||||
ExtractorError,
|
ExtractorError,
|
||||||
float_or_none,
|
|
||||||
int_or_none,
|
int_or_none,
|
||||||
mimetype2ext,
|
mimetype2ext,
|
||||||
str_or_none,
|
parse_iso8601,
|
||||||
traverse_obj,
|
try_call,
|
||||||
unified_timestamp,
|
|
||||||
update_url_query,
|
update_url_query,
|
||||||
url_or_none,
|
url_or_none,
|
||||||
)
|
)
|
||||||
|
from ..utils.traversal import traverse_obj
|
||||||
|
|
||||||
SERIES_API = 'https://production-cdn.dr-massive.com/api/page?device=web_browser&item_detail_expand=all&lang=da&max_list_prefetch=3&path=%s'
|
SERIES_API = 'https://production-cdn.dr-massive.com/api/page?device=web_browser&item_detail_expand=all&lang=da&max_list_prefetch=3&path=%s'
|
||||||
|
|
||||||
@@ -24,7 +20,7 @@ class DRTVIE(InfoExtractor):
|
|||||||
_VALID_URL = r'''(?x)
|
_VALID_URL = r'''(?x)
|
||||||
https?://
|
https?://
|
||||||
(?:
|
(?:
|
||||||
(?:www\.)?dr\.dk/(?:tv/se|nyheder|(?P<radio>radio|lyd)(?:/ondemand)?)/(?:[^/]+/)*|
|
(?:www\.)?dr\.dk/tv/se(?:/ondemand)?/(?:[^/?#]+/)*|
|
||||||
(?:www\.)?(?:dr\.dk|dr-massive\.com)/drtv/(?:se|episode|program)/
|
(?:www\.)?(?:dr\.dk|dr-massive\.com)/drtv/(?:se|episode|program)/
|
||||||
)
|
)
|
||||||
(?P<id>[\da-z_-]+)
|
(?P<id>[\da-z_-]+)
|
||||||
@@ -53,22 +49,6 @@ class DRTVIE(InfoExtractor):
|
|||||||
},
|
},
|
||||||
'expected_warnings': ['Unable to download f4m manifest'],
|
'expected_warnings': ['Unable to download f4m manifest'],
|
||||||
'skip': 'this video has been removed',
|
'skip': 'this video has been removed',
|
||||||
}, {
|
|
||||||
# embed
|
|
||||||
'url': 'https://www.dr.dk/nyheder/indland/live-christianias-rydning-af-pusher-street-er-i-gang',
|
|
||||||
'info_dict': {
|
|
||||||
'id': 'urn:dr:mu:programcard:57c926176187a50a9c6e83c6',
|
|
||||||
'ext': 'mp4',
|
|
||||||
'title': 'christiania pusher street ryddes drdkrjpo',
|
|
||||||
'description': 'md5:2a71898b15057e9b97334f61d04e6eb5',
|
|
||||||
'timestamp': 1472800279,
|
|
||||||
'upload_date': '20160902',
|
|
||||||
'duration': 131.4,
|
|
||||||
},
|
|
||||||
'params': {
|
|
||||||
'skip_download': True,
|
|
||||||
},
|
|
||||||
'expected_warnings': ['Unable to download f4m manifest'],
|
|
||||||
}, {
|
}, {
|
||||||
# with SignLanguage formats
|
# with SignLanguage formats
|
||||||
'url': 'https://www.dr.dk/tv/se/historien-om-danmark/-/historien-om-danmark-stenalder',
|
'url': 'https://www.dr.dk/tv/se/historien-om-danmark/-/historien-om-danmark-stenalder',
|
||||||
@@ -87,33 +67,54 @@ class DRTVIE(InfoExtractor):
|
|||||||
'season': 'Historien om Danmark',
|
'season': 'Historien om Danmark',
|
||||||
'series': 'Historien om Danmark',
|
'series': 'Historien om Danmark',
|
||||||
},
|
},
|
||||||
'params': {
|
'skip': 'this video has been removed',
|
||||||
'skip_download': True,
|
|
||||||
},
|
|
||||||
}, {
|
}, {
|
||||||
'url': 'https://www.dr.dk/lyd/p4kbh/regionale-nyheder-kh4/p4-nyheder-2019-06-26-17-30-9',
|
'url': 'https://www.dr.dk/drtv/se/frank-and-kastaniegaarden_71769',
|
||||||
'only_matching': True,
|
|
||||||
}, {
|
|
||||||
'url': 'https://www.dr.dk/drtv/se/bonderoeven_71769',
|
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': '00951930010',
|
'id': '00951930010',
|
||||||
'ext': 'mp4',
|
'ext': 'mp4',
|
||||||
'title': 'Bonderøven 2019 (1:8)',
|
'title': 'Frank & Kastaniegaarden',
|
||||||
'description': 'md5:b6dcfe9b6f0bea6703e9a0092739a5bd',
|
'description': 'md5:974e1780934cf3275ef10280204bccb0',
|
||||||
'timestamp': 1654856100,
|
'release_timestamp': 1546545600,
|
||||||
'upload_date': '20220610',
|
'release_date': '20190103',
|
||||||
'duration': 2576.6,
|
'duration': 2576,
|
||||||
'season': 'Bonderøven 2019',
|
'season': 'Frank & Kastaniegaarden',
|
||||||
'season_id': 'urn:dr:mu:bundle:5c201667a11fa01ca4528ce5',
|
'season_id': '67125',
|
||||||
'release_year': 2019,
|
'release_year': 2019,
|
||||||
'season_number': 2019,
|
'season_number': 2019,
|
||||||
'series': 'Frank & Kastaniegaarden',
|
'series': 'Frank & Kastaniegaarden',
|
||||||
'episode_number': 1,
|
'episode_number': 1,
|
||||||
'episode': 'Episode 1',
|
'episode': 'Frank & Kastaniegaarden',
|
||||||
|
'thumbnail': r're:https?://.+',
|
||||||
},
|
},
|
||||||
'params': {
|
'params': {
|
||||||
'skip_download': True,
|
'skip_download': True,
|
||||||
},
|
},
|
||||||
|
}, {
|
||||||
|
# Foreign and Regular subtitle track
|
||||||
|
'url': 'https://www.dr.dk/drtv/se/spise-med-price_-pasta-selv_397445',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '00212301010',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'episode_number': 1,
|
||||||
|
'title': 'Spise med Price: Pasta Selv',
|
||||||
|
'alt_title': '1. Pasta Selv',
|
||||||
|
'release_date': '20230807',
|
||||||
|
'description': 'md5:2da9060524fed707810d71080b3d0cd8',
|
||||||
|
'duration': 1750,
|
||||||
|
'season': 'Spise med Price',
|
||||||
|
'release_timestamp': 1691438400,
|
||||||
|
'season_id': '397440',
|
||||||
|
'episode': 'Spise med Price: Pasta Selv',
|
||||||
|
'thumbnail': r're:https?://.+',
|
||||||
|
'season_number': 15,
|
||||||
|
'series': 'Spise med Price',
|
||||||
|
'release_year': 2022,
|
||||||
|
'subtitles': 'mincount:2',
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
'skip_download': 'm3u8',
|
||||||
|
},
|
||||||
}, {
|
}, {
|
||||||
'url': 'https://www.dr.dk/drtv/episode/bonderoeven_71769',
|
'url': 'https://www.dr.dk/drtv/episode/bonderoeven_71769',
|
||||||
'only_matching': True,
|
'only_matching': True,
|
||||||
@@ -123,226 +124,127 @@ class DRTVIE(InfoExtractor):
|
|||||||
}, {
|
}, {
|
||||||
'url': 'https://www.dr.dk/drtv/program/jagten_220924',
|
'url': 'https://www.dr.dk/drtv/program/jagten_220924',
|
||||||
'only_matching': True,
|
'only_matching': True,
|
||||||
}, {
|
|
||||||
'url': 'https://www.dr.dk/lyd/p4aarhus/regionale-nyheder-ar4/regionale-nyheder-2022-05-05-12-30-3',
|
|
||||||
'info_dict': {
|
|
||||||
'id': 'urn:dr:mu:programcard:6265cb2571401424d0360113',
|
|
||||||
'title': "Regionale nyheder",
|
|
||||||
'ext': 'mp4',
|
|
||||||
'duration': 120.043,
|
|
||||||
'series': 'P4 Østjylland regionale nyheder',
|
|
||||||
'timestamp': 1651746600,
|
|
||||||
'season': 'Regionale nyheder',
|
|
||||||
'release_year': 0,
|
|
||||||
'season_id': 'urn:dr:mu:bundle:61c26889539f0201586b73c5',
|
|
||||||
'description': '',
|
|
||||||
'upload_date': '20220505',
|
|
||||||
},
|
|
||||||
'params': {
|
|
||||||
'skip_download': True,
|
|
||||||
},
|
|
||||||
'skip': 'this video has been removed',
|
|
||||||
}, {
|
|
||||||
'url': 'https://www.dr.dk/lyd/p4kbh/regionale-nyheder-kh4/regionale-nyheder-2023-03-14-10-30-9',
|
|
||||||
'info_dict': {
|
|
||||||
'ext': 'mp4',
|
|
||||||
'id': '14802310112',
|
|
||||||
'timestamp': 1678786200,
|
|
||||||
'duration': 120.043,
|
|
||||||
'season_id': 'urn:dr:mu:bundle:63a4f7c87140143504b6710f',
|
|
||||||
'series': 'P4 København regionale nyheder',
|
|
||||||
'upload_date': '20230314',
|
|
||||||
'release_year': 0,
|
|
||||||
'description': 'Hør seneste regionale nyheder fra P4 København.',
|
|
||||||
'season': 'Regionale nyheder',
|
|
||||||
'title': 'Regionale nyheder',
|
|
||||||
},
|
|
||||||
}]
|
}]
|
||||||
|
|
||||||
|
SUBTITLE_LANGS = {
|
||||||
|
'DanishLanguageSubtitles': 'da',
|
||||||
|
'ForeignLanguageSubtitles': 'da_foreign',
|
||||||
|
'CombinedLanguageSubtitles': 'da_combined',
|
||||||
|
}
|
||||||
|
|
||||||
|
_TOKEN = None
|
||||||
|
|
||||||
|
def _real_initialize(self):
|
||||||
|
if self._TOKEN:
|
||||||
|
return
|
||||||
|
|
||||||
|
token_response = self._download_json(
|
||||||
|
'https://production.dr-massive.com/api/authorization/anonymous-sso', None,
|
||||||
|
note='Downloading anonymous token', headers={
|
||||||
|
'content-type': 'application/json',
|
||||||
|
}, query={
|
||||||
|
'device': 'web_browser',
|
||||||
|
'ff': 'idp,ldp,rpt',
|
||||||
|
'lang': 'da',
|
||||||
|
'supportFallbackToken': 'true',
|
||||||
|
}, data=json.dumps({
|
||||||
|
'deviceId': str(uuid.uuid4()),
|
||||||
|
'scopes': ['Catalog'],
|
||||||
|
'optout': True,
|
||||||
|
}).encode())
|
||||||
|
|
||||||
|
self._TOKEN = traverse_obj(
|
||||||
|
token_response, (lambda _, x: x['type'] == 'UserAccount', 'value', {str}), get_all=False)
|
||||||
|
if not self._TOKEN:
|
||||||
|
raise ExtractorError('Unable to get anonymous token')
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
raw_video_id, is_radio_url = self._match_valid_url(url).group('id', 'radio')
|
url_slug = self._match_id(url)
|
||||||
|
webpage = self._download_webpage(url, url_slug)
|
||||||
|
|
||||||
webpage = self._download_webpage(url, raw_video_id)
|
json_data = self._search_json(
|
||||||
|
r'window\.__data\s*=', webpage, 'data', url_slug, fatal=False) or {}
|
||||||
if '>Programmet er ikke længere tilgængeligt' in webpage:
|
item = traverse_obj(
|
||||||
raise ExtractorError(
|
json_data, ('cache', 'page', ..., (None, ('entries', 0)), 'item', {dict}), get_all=False)
|
||||||
'Video %s is not available' % raw_video_id, expected=True)
|
if item:
|
||||||
|
item_id = item.get('id')
|
||||||
video_id = self._search_regex(
|
|
||||||
(r'data-(?:material-identifier|episode-slug)="([^"]+)"',
|
|
||||||
r'data-resource="[^>"]+mu/programcard/expanded/([^"]+)"'),
|
|
||||||
webpage, 'video id', default=None)
|
|
||||||
|
|
||||||
if not video_id:
|
|
||||||
video_id = self._search_regex(
|
|
||||||
r'(urn(?:%3A|:)dr(?:%3A|:)mu(?:%3A|:)programcard(?:%3A|:)[\da-f]+)',
|
|
||||||
webpage, 'urn', default=None)
|
|
||||||
if video_id:
|
|
||||||
video_id = compat_urllib_parse_unquote(video_id)
|
|
||||||
|
|
||||||
_PROGRAMCARD_BASE = 'https://www.dr.dk/mu-online/api/1.4/programcard'
|
|
||||||
query = {'expanded': 'true'}
|
|
||||||
|
|
||||||
if video_id:
|
|
||||||
programcard_url = '%s/%s' % (_PROGRAMCARD_BASE, video_id)
|
|
||||||
else:
|
else:
|
||||||
programcard_url = _PROGRAMCARD_BASE
|
item_id = url_slug.rsplit('_', 1)[-1]
|
||||||
if is_radio_url:
|
item = self._download_json(
|
||||||
video_id = self._search_nextjs_data(
|
f'https://production-cdn.dr-massive.com/api/items/{item_id}', item_id,
|
||||||
webpage, raw_video_id)['props']['pageProps']['episode']['productionNumber']
|
note='Attempting to download backup item data', query={
|
||||||
else:
|
'device': 'web_browser',
|
||||||
json_data = self._search_json(
|
'expand': 'all',
|
||||||
r'window\.__data\s*=', webpage, 'data', raw_video_id)
|
'ff': 'idp,ldp,rpt',
|
||||||
video_id = traverse_obj(json_data, (
|
'geoLocation': 'dk',
|
||||||
'cache', 'page', ..., (None, ('entries', 0)), 'item', 'customId',
|
'isDeviceAbroad': 'false',
|
||||||
{lambda x: x.split(':')[-1]}), get_all=False)
|
'lang': 'da',
|
||||||
if not video_id:
|
'segments': 'drtv,optedout',
|
||||||
raise ExtractorError('Unable to extract video id')
|
'sub': 'Anonymous',
|
||||||
query['productionnumber'] = video_id
|
})
|
||||||
|
|
||||||
data = self._download_json(
|
video_id = try_call(lambda: item['customId'].rsplit(':', 1)[-1]) or item_id
|
||||||
programcard_url, video_id, 'Downloading video JSON', query=query)
|
stream_data = self._download_json(
|
||||||
|
f'https://production.dr-massive.com/api/account/items/{item_id}/videos', video_id,
|
||||||
supplementary_data = {}
|
note='Downloading stream data', query={
|
||||||
if re.search(r'_\d+$', raw_video_id):
|
'delivery': 'stream',
|
||||||
supplementary_data = self._download_json(
|
'device': 'web_browser',
|
||||||
SERIES_API % f'/episode/{raw_video_id}', raw_video_id, fatal=False) or {}
|
'ff': 'idp,ldp,rpt',
|
||||||
|
'lang': 'da',
|
||||||
title = str_or_none(data.get('Title')) or re.sub(
|
'resolution': 'HD-1080',
|
||||||
r'\s*\|\s*(?:TV\s*\|\s*DR|DRTV)$', '',
|
'sub': 'Anonymous',
|
||||||
self._og_search_title(webpage))
|
}, headers={'authorization': f'Bearer {self._TOKEN}'})
|
||||||
description = self._og_search_description(
|
|
||||||
webpage, default=None) or data.get('Description')
|
|
||||||
|
|
||||||
timestamp = unified_timestamp(
|
|
||||||
data.get('PrimaryBroadcastStartTime') or data.get('SortDateTime'))
|
|
||||||
|
|
||||||
thumbnail = None
|
|
||||||
duration = None
|
|
||||||
|
|
||||||
restricted_to_denmark = False
|
|
||||||
|
|
||||||
formats = []
|
formats = []
|
||||||
subtitles = {}
|
subtitles = {}
|
||||||
|
for stream in traverse_obj(stream_data, (lambda _, x: x['url'])):
|
||||||
|
format_id = stream.get('format', 'na')
|
||||||
|
access_service = stream.get('accessService')
|
||||||
|
preference = None
|
||||||
|
subtitle_suffix = ''
|
||||||
|
if access_service in ('SpokenSubtitles', 'SignLanguage', 'VisuallyInterpreted'):
|
||||||
|
preference = -1
|
||||||
|
format_id += f'-{access_service}'
|
||||||
|
subtitle_suffix = f'-{access_service}'
|
||||||
|
elif access_service == 'StandardVideo':
|
||||||
|
preference = 1
|
||||||
|
fmts, subs = self._extract_m3u8_formats_and_subtitles(
|
||||||
|
stream.get('url'), video_id, preference=preference, m3u8_id=format_id, fatal=False)
|
||||||
|
formats.extend(fmts)
|
||||||
|
|
||||||
assets = []
|
api_subtitles = traverse_obj(stream, ('subtitles', lambda _, v: url_or_none(v['link']), {dict}))
|
||||||
primary_asset = data.get('PrimaryAsset')
|
if not api_subtitles:
|
||||||
if isinstance(primary_asset, dict):
|
self._merge_subtitles(subs, target=subtitles)
|
||||||
assets.append(primary_asset)
|
|
||||||
secondary_assets = data.get('SecondaryAssets')
|
|
||||||
if isinstance(secondary_assets, list):
|
|
||||||
for secondary_asset in secondary_assets:
|
|
||||||
if isinstance(secondary_asset, dict):
|
|
||||||
assets.append(secondary_asset)
|
|
||||||
|
|
||||||
def hex_to_bytes(hex):
|
for sub_track in api_subtitles:
|
||||||
return binascii.a2b_hex(hex.encode('ascii'))
|
lang = sub_track.get('language') or 'da'
|
||||||
|
subtitles.setdefault(self.SUBTITLE_LANGS.get(lang, lang) + subtitle_suffix, []).append({
|
||||||
|
'url': sub_track['link'],
|
||||||
|
'ext': mimetype2ext(sub_track.get('format')) or 'vtt'
|
||||||
|
})
|
||||||
|
|
||||||
def decrypt_uri(e):
|
if not formats and traverse_obj(item, ('season', 'customFields', 'IsGeoRestricted')):
|
||||||
n = int(e[2:10], 16)
|
self.raise_geo_restricted(countries=self._GEO_COUNTRIES)
|
||||||
a = e[10 + n:]
|
|
||||||
data = hex_to_bytes(e[10:10 + n])
|
|
||||||
key = hashlib.sha256(('%s:sRBzYNXBzkKgnjj8pGtkACch' % a).encode('utf-8')).digest()
|
|
||||||
iv = hex_to_bytes(a)
|
|
||||||
decrypted = unpad_pkcs7(aes_cbc_decrypt_bytes(data, key, iv))
|
|
||||||
return decrypted.decode('utf-8').split('?')[0]
|
|
||||||
|
|
||||||
for asset in assets:
|
|
||||||
kind = asset.get('Kind')
|
|
||||||
if kind == 'Image':
|
|
||||||
thumbnail = url_or_none(asset.get('Uri'))
|
|
||||||
elif kind in ('VideoResource', 'AudioResource'):
|
|
||||||
duration = float_or_none(asset.get('DurationInMilliseconds'), 1000)
|
|
||||||
restricted_to_denmark = asset.get('RestrictedToDenmark')
|
|
||||||
asset_target = asset.get('Target')
|
|
||||||
for link in asset.get('Links', []):
|
|
||||||
uri = link.get('Uri')
|
|
||||||
if not uri:
|
|
||||||
encrypted_uri = link.get('EncryptedUri')
|
|
||||||
if not encrypted_uri:
|
|
||||||
continue
|
|
||||||
try:
|
|
||||||
uri = decrypt_uri(encrypted_uri)
|
|
||||||
except Exception:
|
|
||||||
self.report_warning(
|
|
||||||
'Unable to decrypt EncryptedUri', video_id)
|
|
||||||
continue
|
|
||||||
uri = url_or_none(uri)
|
|
||||||
if not uri:
|
|
||||||
continue
|
|
||||||
target = link.get('Target')
|
|
||||||
format_id = target or ''
|
|
||||||
if asset_target in ('SpokenSubtitles', 'SignLanguage', 'VisuallyInterpreted'):
|
|
||||||
preference = -1
|
|
||||||
format_id += '-%s' % asset_target
|
|
||||||
elif asset_target == 'Default':
|
|
||||||
preference = 1
|
|
||||||
else:
|
|
||||||
preference = None
|
|
||||||
if target == 'HDS':
|
|
||||||
f4m_formats = self._extract_f4m_formats(
|
|
||||||
uri + '?hdcore=3.3.0&plugin=aasp-3.3.0.99.43',
|
|
||||||
video_id, preference, f4m_id=format_id, fatal=False)
|
|
||||||
if kind == 'AudioResource':
|
|
||||||
for f in f4m_formats:
|
|
||||||
f['vcodec'] = 'none'
|
|
||||||
formats.extend(f4m_formats)
|
|
||||||
elif target == 'HLS':
|
|
||||||
fmts, subs = self._extract_m3u8_formats_and_subtitles(
|
|
||||||
uri, video_id, 'mp4', entry_protocol='m3u8_native',
|
|
||||||
quality=preference, m3u8_id=format_id, fatal=False)
|
|
||||||
formats.extend(fmts)
|
|
||||||
self._merge_subtitles(subs, target=subtitles)
|
|
||||||
else:
|
|
||||||
bitrate = link.get('Bitrate')
|
|
||||||
if bitrate:
|
|
||||||
format_id += '-%s' % bitrate
|
|
||||||
formats.append({
|
|
||||||
'url': uri,
|
|
||||||
'format_id': format_id,
|
|
||||||
'tbr': int_or_none(bitrate),
|
|
||||||
'ext': link.get('FileFormat'),
|
|
||||||
'vcodec': 'none' if kind == 'AudioResource' else None,
|
|
||||||
'quality': preference,
|
|
||||||
})
|
|
||||||
subtitles_list = asset.get('SubtitlesList') or asset.get('Subtitleslist')
|
|
||||||
if isinstance(subtitles_list, list):
|
|
||||||
LANGS = {
|
|
||||||
'Danish': 'da',
|
|
||||||
}
|
|
||||||
for subs in subtitles_list:
|
|
||||||
if not isinstance(subs, dict):
|
|
||||||
continue
|
|
||||||
sub_uri = url_or_none(subs.get('Uri'))
|
|
||||||
if not sub_uri:
|
|
||||||
continue
|
|
||||||
lang = subs.get('Language') or 'da'
|
|
||||||
subtitles.setdefault(LANGS.get(lang, lang), []).append({
|
|
||||||
'url': sub_uri,
|
|
||||||
'ext': mimetype2ext(subs.get('MimeType')) or 'vtt'
|
|
||||||
})
|
|
||||||
|
|
||||||
if not formats and restricted_to_denmark:
|
|
||||||
self.raise_geo_restricted(
|
|
||||||
'Unfortunately, DR is not allowed to show this program outside Denmark.',
|
|
||||||
countries=self._GEO_COUNTRIES)
|
|
||||||
|
|
||||||
return {
|
return {
|
||||||
'id': video_id,
|
'id': video_id,
|
||||||
'title': title,
|
|
||||||
'description': description,
|
|
||||||
'thumbnail': thumbnail,
|
|
||||||
'timestamp': timestamp,
|
|
||||||
'duration': duration,
|
|
||||||
'formats': formats,
|
'formats': formats,
|
||||||
'subtitles': subtitles,
|
'subtitles': subtitles,
|
||||||
'series': str_or_none(data.get('SeriesTitle')),
|
**traverse_obj(item, {
|
||||||
'season': str_or_none(data.get('SeasonTitle')),
|
'title': 'title',
|
||||||
'season_number': int_or_none(data.get('SeasonNumber')),
|
'alt_title': 'contextualTitle',
|
||||||
'season_id': str_or_none(data.get('SeasonUrn')),
|
'description': 'description',
|
||||||
'episode': traverse_obj(supplementary_data, ('entries', 0, 'item', 'contextualTitle')) or str_or_none(data.get('EpisodeTitle')),
|
'thumbnail': ('images', 'wallpaper'),
|
||||||
'episode_number': traverse_obj(supplementary_data, ('entries', 0, 'item', 'episodeNumber')) or int_or_none(data.get('EpisodeNumber')),
|
'release_timestamp': ('customFields', 'BroadcastTimeDK', {parse_iso8601}),
|
||||||
'release_year': int_or_none(data.get('ProductionYear')),
|
'duration': ('duration', {int_or_none}),
|
||||||
|
'series': ('season', 'show', 'title'),
|
||||||
|
'season': ('season', 'title'),
|
||||||
|
'season_number': ('season', 'seasonNumber', {int_or_none}),
|
||||||
|
'season_id': 'seasonId',
|
||||||
|
'episode': 'episodeName',
|
||||||
|
'episode_number': ('episodeNumber', {int_or_none}),
|
||||||
|
'release_year': ('releaseYear', {int_or_none}),
|
||||||
|
}),
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
@@ -412,6 +314,8 @@ class DRTVSeasonIE(InfoExtractor):
|
|||||||
'display_id': 'frank-and-kastaniegaarden',
|
'display_id': 'frank-and-kastaniegaarden',
|
||||||
'title': 'Frank & Kastaniegaarden',
|
'title': 'Frank & Kastaniegaarden',
|
||||||
'series': 'Frank & Kastaniegaarden',
|
'series': 'Frank & Kastaniegaarden',
|
||||||
|
'season_number': 2008,
|
||||||
|
'alt_title': 'Season 2008',
|
||||||
},
|
},
|
||||||
'playlist_mincount': 8
|
'playlist_mincount': 8
|
||||||
}, {
|
}, {
|
||||||
@@ -421,6 +325,8 @@ class DRTVSeasonIE(InfoExtractor):
|
|||||||
'display_id': 'frank-and-kastaniegaarden',
|
'display_id': 'frank-and-kastaniegaarden',
|
||||||
'title': 'Frank & Kastaniegaarden',
|
'title': 'Frank & Kastaniegaarden',
|
||||||
'series': 'Frank & Kastaniegaarden',
|
'series': 'Frank & Kastaniegaarden',
|
||||||
|
'season_number': 2009,
|
||||||
|
'alt_title': 'Season 2009',
|
||||||
},
|
},
|
||||||
'playlist_mincount': 19
|
'playlist_mincount': 19
|
||||||
}]
|
}]
|
||||||
@@ -434,6 +340,7 @@ class DRTVSeasonIE(InfoExtractor):
|
|||||||
'url': f'https://www.dr.dk/drtv{episode["path"]}',
|
'url': f'https://www.dr.dk/drtv{episode["path"]}',
|
||||||
'ie_key': DRTVIE.ie_key(),
|
'ie_key': DRTVIE.ie_key(),
|
||||||
'title': episode.get('title'),
|
'title': episode.get('title'),
|
||||||
|
'alt_title': episode.get('contextualTitle'),
|
||||||
'episode': episode.get('episodeName'),
|
'episode': episode.get('episodeName'),
|
||||||
'description': episode.get('shortDescription'),
|
'description': episode.get('shortDescription'),
|
||||||
'series': traverse_obj(data, ('entries', 0, 'item', 'title')),
|
'series': traverse_obj(data, ('entries', 0, 'item', 'title')),
|
||||||
@@ -446,6 +353,7 @@ class DRTVSeasonIE(InfoExtractor):
|
|||||||
'id': season_id,
|
'id': season_id,
|
||||||
'display_id': display_id,
|
'display_id': display_id,
|
||||||
'title': traverse_obj(data, ('entries', 0, 'item', 'title')),
|
'title': traverse_obj(data, ('entries', 0, 'item', 'title')),
|
||||||
|
'alt_title': traverse_obj(data, ('entries', 0, 'item', 'contextualTitle')),
|
||||||
'series': traverse_obj(data, ('entries', 0, 'item', 'title')),
|
'series': traverse_obj(data, ('entries', 0, 'item', 'title')),
|
||||||
'entries': entries,
|
'entries': entries,
|
||||||
'season_number': traverse_obj(data, ('entries', 0, 'item', 'seasonNumber'))
|
'season_number': traverse_obj(data, ('entries', 0, 'item', 'seasonNumber'))
|
||||||
@@ -463,6 +371,7 @@ class DRTVSeriesIE(InfoExtractor):
|
|||||||
'display_id': 'frank-and-kastaniegaarden',
|
'display_id': 'frank-and-kastaniegaarden',
|
||||||
'title': 'Frank & Kastaniegaarden',
|
'title': 'Frank & Kastaniegaarden',
|
||||||
'series': 'Frank & Kastaniegaarden',
|
'series': 'Frank & Kastaniegaarden',
|
||||||
|
'alt_title': '',
|
||||||
},
|
},
|
||||||
'playlist_mincount': 15
|
'playlist_mincount': 15
|
||||||
}]
|
}]
|
||||||
@@ -476,6 +385,7 @@ class DRTVSeriesIE(InfoExtractor):
|
|||||||
'url': f'https://www.dr.dk/drtv{season.get("path")}',
|
'url': f'https://www.dr.dk/drtv{season.get("path")}',
|
||||||
'ie_key': DRTVSeasonIE.ie_key(),
|
'ie_key': DRTVSeasonIE.ie_key(),
|
||||||
'title': season.get('title'),
|
'title': season.get('title'),
|
||||||
|
'alt_title': season.get('contextualTitle'),
|
||||||
'series': traverse_obj(data, ('entries', 0, 'item', 'title')),
|
'series': traverse_obj(data, ('entries', 0, 'item', 'title')),
|
||||||
'season_number': traverse_obj(data, ('entries', 0, 'item', 'seasonNumber'))
|
'season_number': traverse_obj(data, ('entries', 0, 'item', 'seasonNumber'))
|
||||||
} for season in traverse_obj(data, ('entries', 0, 'item', 'show', 'seasons', 'items'))]
|
} for season in traverse_obj(data, ('entries', 0, 'item', 'show', 'seasons', 'items'))]
|
||||||
@@ -485,6 +395,7 @@ class DRTVSeriesIE(InfoExtractor):
|
|||||||
'id': series_id,
|
'id': series_id,
|
||||||
'display_id': display_id,
|
'display_id': display_id,
|
||||||
'title': traverse_obj(data, ('entries', 0, 'item', 'title')),
|
'title': traverse_obj(data, ('entries', 0, 'item', 'title')),
|
||||||
|
'alt_title': traverse_obj(data, ('entries', 0, 'item', 'contextualTitle')),
|
||||||
'series': traverse_obj(data, ('entries', 0, 'item', 'title')),
|
'series': traverse_obj(data, ('entries', 0, 'item', 'title')),
|
||||||
'entries': entries
|
'entries': entries
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -138,7 +138,7 @@ class DubokuIE(InfoExtractor):
|
|||||||
# of the video.
|
# of the video.
|
||||||
return {
|
return {
|
||||||
'_type': 'url_transparent',
|
'_type': 'url_transparent',
|
||||||
'url': smuggle_url(data_url, {'http_headers': headers}),
|
'url': smuggle_url(data_url, {'referer': webpage_url}),
|
||||||
'id': video_id,
|
'id': video_id,
|
||||||
'title': title,
|
'title': title,
|
||||||
'series': series_title,
|
'series': series_title,
|
||||||
|
|||||||
62
yt_dlp/extractor/eltrecetv.py
Normal file
62
yt_dlp/extractor/eltrecetv.py
Normal file
@@ -0,0 +1,62 @@
|
|||||||
|
from .common import InfoExtractor
|
||||||
|
|
||||||
|
|
||||||
|
class ElTreceTVIE(InfoExtractor):
|
||||||
|
IE_DESC = 'El Trece TV (Argentina)'
|
||||||
|
_VALID_URL = r'https?://(?:www\.)?eltrecetv\.com\.ar/[\w-]+/capitulos/temporada-\d+/(?P<id>[\w-]+)'
|
||||||
|
_TESTS = [
|
||||||
|
{
|
||||||
|
'url': 'https://www.eltrecetv.com.ar/ahora-caigo/capitulos/temporada-2023/programa-del-061023/',
|
||||||
|
'md5': '71a66673dc63f9a5939d97bfe4b311ba',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'AHCA05102023145553329621094',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': 'AHORA CAIGO - Programa 06/10/23',
|
||||||
|
'thumbnail': 'https://thumbs.vodgc.net/AHCA05102023145553329621094.JPG?649339',
|
||||||
|
}
|
||||||
|
},
|
||||||
|
{
|
||||||
|
'url': 'https://www.eltrecetv.com.ar/poco-correctos/capitulos/temporada-2023/programa-del-250923-invitada-dalia-gutmann/',
|
||||||
|
'only_matching': True,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
'url': 'https://www.eltrecetv.com.ar/argentina-tierra-de-amor-y-venganza/capitulos/temporada-2023/atav-2-capitulo-121-del-250923/',
|
||||||
|
'only_matching': True,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
'url': 'https://www.eltrecetv.com.ar/ahora-caigo/capitulos/temporada-2023/programa-del-250923/',
|
||||||
|
'only_matching': True,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
'url': 'https://www.eltrecetv.com.ar/pasaplatos/capitulos/temporada-2023/pasaplatos-el-restaurante-del-250923/',
|
||||||
|
'only_matching': True,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
'url': 'https://www.eltrecetv.com.ar/el-galpon/capitulos/temporada-2023/programa-del-160923-invitado-raul-lavie/',
|
||||||
|
'only_matching': True,
|
||||||
|
}
|
||||||
|
]
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
slug = self._match_id(url)
|
||||||
|
webpage = self._download_webpage(url, slug)
|
||||||
|
config = self._search_json(
|
||||||
|
r'Fusion.globalContent\s*=', webpage, 'content', slug)['promo_items']['basic']['embed']['config']
|
||||||
|
video_url = config['m3u8']
|
||||||
|
video_id = self._search_regex(r'/(\w+)\.m3u8', video_url, 'video id', default=slug)
|
||||||
|
|
||||||
|
formats, subtitles = self._extract_m3u8_formats_and_subtitles(video_url, video_id, 'mp4', m3u8_id='hls')
|
||||||
|
formats.extend([{
|
||||||
|
'url': f['url'][:-23],
|
||||||
|
'format_id': f['format_id'].replace('hls', 'http'),
|
||||||
|
'width': f.get('width'),
|
||||||
|
'height': f.get('height'),
|
||||||
|
} for f in formats if f['url'].endswith('/tracks-v1a1/index.m3u8') and f.get('height') != 1080])
|
||||||
|
|
||||||
|
return {
|
||||||
|
'id': video_id,
|
||||||
|
'title': config.get('title'),
|
||||||
|
'thumbnail': config.get('thumbnail'),
|
||||||
|
'formats': formats,
|
||||||
|
'subtitles': subtitles,
|
||||||
|
}
|
||||||
@@ -106,4 +106,4 @@ class EmbedlyIE(InfoExtractor):
|
|||||||
return self.url_result(src, YoutubeTabIE)
|
return self.url_result(src, YoutubeTabIE)
|
||||||
return self.url_result(smuggle_url(
|
return self.url_result(smuggle_url(
|
||||||
urllib.parse.unquote(traverse_obj(qs, ('src', 0), ('url', 0))),
|
urllib.parse.unquote(traverse_obj(qs, ('src', 0), ('url', 0))),
|
||||||
{'http_headers': {'Referer': url}}))
|
{'referer': url}))
|
||||||
|
|||||||
@@ -17,6 +17,7 @@ from ..utils import (
|
|||||||
determine_protocol,
|
determine_protocol,
|
||||||
dict_get,
|
dict_get,
|
||||||
extract_basic_auth,
|
extract_basic_auth,
|
||||||
|
filter_dict,
|
||||||
format_field,
|
format_field,
|
||||||
int_or_none,
|
int_or_none,
|
||||||
is_html,
|
is_html,
|
||||||
@@ -34,6 +35,7 @@ from ..utils import (
|
|||||||
unified_timestamp,
|
unified_timestamp,
|
||||||
unsmuggle_url,
|
unsmuggle_url,
|
||||||
update_url_query,
|
update_url_query,
|
||||||
|
urlhandle_detect_ext,
|
||||||
url_or_none,
|
url_or_none,
|
||||||
urljoin,
|
urljoin,
|
||||||
variadic,
|
variadic,
|
||||||
@@ -2434,10 +2436,10 @@ class GenericIE(InfoExtractor):
|
|||||||
# to accept raw bytes and being able to download only a chunk.
|
# to accept raw bytes and being able to download only a chunk.
|
||||||
# It may probably better to solve this by checking Content-Type for application/octet-stream
|
# It may probably better to solve this by checking Content-Type for application/octet-stream
|
||||||
# after a HEAD request, but not sure if we can rely on this.
|
# after a HEAD request, but not sure if we can rely on this.
|
||||||
full_response = self._request_webpage(url, video_id, headers={
|
full_response = self._request_webpage(url, video_id, headers=filter_dict({
|
||||||
'Accept-Encoding': 'identity',
|
'Accept-Encoding': 'identity',
|
||||||
**smuggled_data.get('http_headers', {})
|
'Referer': smuggled_data.get('referer'),
|
||||||
})
|
}))
|
||||||
new_url = full_response.url
|
new_url = full_response.url
|
||||||
url = urllib.parse.urlparse(url)._replace(scheme=urllib.parse.urlparse(new_url).scheme).geturl()
|
url = urllib.parse.urlparse(url)._replace(scheme=urllib.parse.urlparse(new_url).scheme).geturl()
|
||||||
if new_url != extract_basic_auth(url)[0]:
|
if new_url != extract_basic_auth(url)[0]:
|
||||||
@@ -2457,9 +2459,9 @@ class GenericIE(InfoExtractor):
|
|||||||
m = re.match(r'^(?P<type>audio|video|application(?=/(?:ogg$|(?:vnd\.apple\.|x-)?mpegurl)))/(?P<format_id>[^;\s]+)', content_type)
|
m = re.match(r'^(?P<type>audio|video|application(?=/(?:ogg$|(?:vnd\.apple\.|x-)?mpegurl)))/(?P<format_id>[^;\s]+)', content_type)
|
||||||
if m:
|
if m:
|
||||||
self.report_detected('direct video link')
|
self.report_detected('direct video link')
|
||||||
headers = smuggled_data.get('http_headers', {})
|
headers = filter_dict({'Referer': smuggled_data.get('referer')})
|
||||||
format_id = str(m.group('format_id'))
|
format_id = str(m.group('format_id'))
|
||||||
ext = determine_ext(url)
|
ext = determine_ext(url, default_ext=None) or urlhandle_detect_ext(full_response)
|
||||||
subtitles = {}
|
subtitles = {}
|
||||||
if format_id.endswith('mpegurl') or ext == 'm3u8':
|
if format_id.endswith('mpegurl') or ext == 'm3u8':
|
||||||
formats, subtitles = self._extract_m3u8_formats_and_subtitles(url, video_id, 'mp4', headers=headers)
|
formats, subtitles = self._extract_m3u8_formats_and_subtitles(url, video_id, 'mp4', headers=headers)
|
||||||
@@ -2471,6 +2473,7 @@ class GenericIE(InfoExtractor):
|
|||||||
formats = [{
|
formats = [{
|
||||||
'format_id': format_id,
|
'format_id': format_id,
|
||||||
'url': url,
|
'url': url,
|
||||||
|
'ext': ext,
|
||||||
'vcodec': 'none' if m.group('type') == 'audio' else None
|
'vcodec': 'none' if m.group('type') == 'audio' else None
|
||||||
}]
|
}]
|
||||||
info_dict['direct'] = True
|
info_dict['direct'] = True
|
||||||
@@ -2708,7 +2711,7 @@ class GenericIE(InfoExtractor):
|
|||||||
'url': smuggle_url(json_ld['url'], {
|
'url': smuggle_url(json_ld['url'], {
|
||||||
'force_videoid': video_id,
|
'force_videoid': video_id,
|
||||||
'to_generic': True,
|
'to_generic': True,
|
||||||
'http_headers': {'Referer': url},
|
'referer': url,
|
||||||
}),
|
}),
|
||||||
}, json_ld)]
|
}, json_ld)]
|
||||||
|
|
||||||
|
|||||||
79
yt_dlp/extractor/jiosaavn.py
Normal file
79
yt_dlp/extractor/jiosaavn.py
Normal file
@@ -0,0 +1,79 @@
|
|||||||
|
from .common import InfoExtractor
|
||||||
|
from ..utils import (
|
||||||
|
js_to_json,
|
||||||
|
url_or_none,
|
||||||
|
urlencode_postdata,
|
||||||
|
urljoin,
|
||||||
|
)
|
||||||
|
from ..utils.traversal import traverse_obj
|
||||||
|
|
||||||
|
|
||||||
|
class JioSaavnBaseIE(InfoExtractor):
|
||||||
|
def _extract_initial_data(self, url, audio_id):
|
||||||
|
webpage = self._download_webpage(url, audio_id)
|
||||||
|
return self._search_json(
|
||||||
|
r'window\.__INITIAL_DATA__\s*=', webpage,
|
||||||
|
'init json', audio_id, transform_source=js_to_json)
|
||||||
|
|
||||||
|
|
||||||
|
class JioSaavnSongIE(JioSaavnBaseIE):
|
||||||
|
_VALID_URL = r'https?://(?:www\.)?(?:jiosaavn\.com/song/[^/?#]+/|saavn\.com/s/song/(?:[^/?#]+/){3})(?P<id>[^/?#]+)'
|
||||||
|
_TESTS = [{
|
||||||
|
'url': 'https://www.jiosaavn.com/song/leja-re/OQsEfQFVUXk',
|
||||||
|
'md5': '7b1f70de088ede3a152ea34aece4df42',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'OQsEfQFVUXk',
|
||||||
|
'ext': 'mp3',
|
||||||
|
'title': 'Leja Re',
|
||||||
|
'album': 'Leja Re',
|
||||||
|
'thumbnail': 'https://c.saavncdn.com/258/Leja-Re-Hindi-2018-20181124024539-500x500.jpg',
|
||||||
|
},
|
||||||
|
}, {
|
||||||
|
'url': 'https://www.saavn.com/s/song/hindi/Saathiya/O-Humdum-Suniyo-Re/KAMiazoCblU',
|
||||||
|
'only_matching': True,
|
||||||
|
}]
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
audio_id = self._match_id(url)
|
||||||
|
song_data = self._extract_initial_data(url, audio_id)['song']['song']
|
||||||
|
media_data = self._download_json(
|
||||||
|
'https://www.jiosaavn.com/api.php', audio_id, data=urlencode_postdata({
|
||||||
|
'__call': 'song.generateAuthToken',
|
||||||
|
'_format': 'json',
|
||||||
|
'bitrate': '128',
|
||||||
|
'url': song_data['encrypted_media_url'],
|
||||||
|
}))
|
||||||
|
|
||||||
|
return {
|
||||||
|
'id': audio_id,
|
||||||
|
'url': media_data['auth_url'],
|
||||||
|
'ext': media_data.get('type'),
|
||||||
|
'vcodec': 'none',
|
||||||
|
**traverse_obj(song_data, {
|
||||||
|
'title': ('title', 'text'),
|
||||||
|
'album': ('album', 'text'),
|
||||||
|
'thumbnail': ('image', 0, {url_or_none}),
|
||||||
|
}),
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class JioSaavnAlbumIE(JioSaavnBaseIE):
|
||||||
|
_VALID_URL = r'https?://(?:www\.)?(?:jio)?saavn\.com/album/[^/?#]+/(?P<id>[^/?#]+)'
|
||||||
|
_TESTS = [{
|
||||||
|
'url': 'https://www.jiosaavn.com/album/96/buIOjYZDrNA_',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'buIOjYZDrNA_',
|
||||||
|
'title': '96',
|
||||||
|
},
|
||||||
|
'playlist_count': 10,
|
||||||
|
}]
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
album_id = self._match_id(url)
|
||||||
|
album_view = self._extract_initial_data(url, album_id)['albumView']
|
||||||
|
|
||||||
|
return self.playlist_from_matches(
|
||||||
|
traverse_obj(album_view, (
|
||||||
|
'modules', lambda _, x: x['key'] == 'list', 'data', ..., 'title', 'action', {str})),
|
||||||
|
album_id, traverse_obj(album_view, ('album', 'title', 'text', {str})), ie=JioSaavnSongIE,
|
||||||
|
getter=lambda x: urljoin('https://www.jiosaavn.com/', x))
|
||||||
@@ -208,9 +208,9 @@ class LA7PodcastIE(LA7PodcastEpisodeIE): # XXX: Do not subclass from concrete I
|
|||||||
'url': 'https://www.la7.it/propagandalive/podcast',
|
'url': 'https://www.la7.it/propagandalive/podcast',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'propagandalive',
|
'id': 'propagandalive',
|
||||||
'title': "Propaganda Live",
|
'title': 'Propaganda Live',
|
||||||
},
|
},
|
||||||
'playlist_count_min': 10,
|
'playlist_mincount': 10,
|
||||||
}]
|
}]
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
|
|||||||
73
yt_dlp/extractor/laxarxames.py
Normal file
73
yt_dlp/extractor/laxarxames.py
Normal file
@@ -0,0 +1,73 @@
|
|||||||
|
import json
|
||||||
|
|
||||||
|
from .brightcove import BrightcoveNewIE
|
||||||
|
from .common import InfoExtractor
|
||||||
|
from ..utils import ExtractorError
|
||||||
|
from ..utils.traversal import traverse_obj
|
||||||
|
|
||||||
|
|
||||||
|
class LaXarxaMesIE(InfoExtractor):
|
||||||
|
_VALID_URL = r'https?://(?:www\.)?laxarxames\.cat/(?:[^/?#]+/)*?(player|movie-details)/(?P<id>\d+)'
|
||||||
|
_NETRC_MACHINE = 'laxarxames'
|
||||||
|
_TOKEN = None
|
||||||
|
_TESTS = [{
|
||||||
|
'url': 'https://www.laxarxames.cat/player/3459421',
|
||||||
|
'md5': '0966f46c34275934c19af78f3df6e2bc',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '6339612436112',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': 'Resum | UA Horta — UD Viladecans',
|
||||||
|
'timestamp': 1697905186,
|
||||||
|
'thumbnail': r're:https?://.*\.jpg',
|
||||||
|
'description': '',
|
||||||
|
'upload_date': '20231021',
|
||||||
|
'duration': 129.44,
|
||||||
|
'tags': ['ott', 'esports', '23-24', ' futbol', ' futbol-partits', 'elit', 'resum'],
|
||||||
|
'uploader_id': '5779379807001',
|
||||||
|
},
|
||||||
|
'skip': 'Requires login',
|
||||||
|
}]
|
||||||
|
|
||||||
|
def _perform_login(self, username, password):
|
||||||
|
if self._TOKEN:
|
||||||
|
return
|
||||||
|
|
||||||
|
login = self._download_json(
|
||||||
|
'https://api.laxarxames.cat/Authorization/SignIn', None, note='Logging in', headers={
|
||||||
|
'X-Tenantorigin': 'https://laxarxames.cat',
|
||||||
|
'Content-Type': 'application/json',
|
||||||
|
}, data=json.dumps({
|
||||||
|
'Username': username,
|
||||||
|
'Password': password,
|
||||||
|
'Device': {
|
||||||
|
'PlatformCode': 'WEB',
|
||||||
|
'Name': 'Mac OS ()',
|
||||||
|
},
|
||||||
|
}).encode(), expected_status=401)
|
||||||
|
|
||||||
|
self._TOKEN = traverse_obj(login, ('AuthorizationToken', 'Token', {str}))
|
||||||
|
if not self._TOKEN:
|
||||||
|
raise ExtractorError('Login failed', expected=True)
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
video_id = self._match_id(url)
|
||||||
|
if not self._TOKEN:
|
||||||
|
self.raise_login_required()
|
||||||
|
|
||||||
|
media_play_info = self._download_json(
|
||||||
|
'https://api.laxarxames.cat/Media/GetMediaPlayInfo', video_id,
|
||||||
|
data=json.dumps({
|
||||||
|
'MediaId': int(video_id),
|
||||||
|
'StreamType': 'MAIN'
|
||||||
|
}).encode(), headers={
|
||||||
|
'Authorization': f'Bearer {self._TOKEN}',
|
||||||
|
'X-Tenantorigin': 'https://laxarxames.cat',
|
||||||
|
'Content-Type': 'application/json',
|
||||||
|
})
|
||||||
|
|
||||||
|
if not traverse_obj(media_play_info, ('ContentUrl', {str})):
|
||||||
|
self.raise_no_formats('No video found', expected=True)
|
||||||
|
|
||||||
|
return self.url_result(
|
||||||
|
f'https://players.brightcove.net/5779379807001/default_default/index.html?videoId={media_play_info["ContentUrl"]}',
|
||||||
|
BrightcoveNewIE, video_id, media_play_info.get('Title'))
|
||||||
@@ -1,4 +1,5 @@
|
|||||||
import re
|
import re
|
||||||
|
import xml.etree.ElementTree
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from ..compat import compat_str
|
from ..compat import compat_str
|
||||||
@@ -137,7 +138,7 @@ class MTVServicesInfoExtractor(InfoExtractor):
|
|||||||
mediagen_doc = self._download_xml(
|
mediagen_doc = self._download_xml(
|
||||||
mediagen_url, video_id, 'Downloading video urls', fatal=False)
|
mediagen_url, video_id, 'Downloading video urls', fatal=False)
|
||||||
|
|
||||||
if mediagen_doc is False:
|
if not isinstance(mediagen_doc, xml.etree.ElementTree.Element):
|
||||||
return None
|
return None
|
||||||
|
|
||||||
item = mediagen_doc.find('./video/item')
|
item = mediagen_doc.find('./video/item')
|
||||||
|
|||||||
@@ -1,6 +1,7 @@
|
|||||||
import base64
|
import base64
|
||||||
import json
|
import json
|
||||||
import re
|
import re
|
||||||
|
import xml.etree.ElementTree
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from .theplatform import ThePlatformIE, default_ns
|
from .theplatform import ThePlatformIE, default_ns
|
||||||
@@ -803,8 +804,10 @@ class NBCStationsIE(InfoExtractor):
|
|||||||
smil = self._download_xml(
|
smil = self._download_xml(
|
||||||
f'https://link.theplatform.com/s/{pdk_acct}/{player_id}', video_id,
|
f'https://link.theplatform.com/s/{pdk_acct}/{player_id}', video_id,
|
||||||
note='Downloading SMIL data', query=query, fatal=is_live)
|
note='Downloading SMIL data', query=query, fatal=is_live)
|
||||||
subtitles = self._parse_smil_subtitles(smil, default_ns) if smil else {}
|
if not isinstance(smil, xml.etree.ElementTree.Element):
|
||||||
for video in smil.findall(self._xpath_ns('.//video', default_ns)) if smil else []:
|
smil = None
|
||||||
|
subtitles = self._parse_smil_subtitles(smil, default_ns) if smil is not None else {}
|
||||||
|
for video in smil.findall(self._xpath_ns('.//video', default_ns)) if smil is not None else []:
|
||||||
info['duration'] = float_or_none(remove_end(video.get('dur'), 'ms'), 1000)
|
info['duration'] = float_or_none(remove_end(video.get('dur'), 'ms'), 1000)
|
||||||
video_src_url = video.get('src')
|
video_src_url = video.get('src')
|
||||||
ext = mimetype2ext(video.get('type'), default=determine_ext(video_src_url))
|
ext = mimetype2ext(video.get('type'), default=determine_ext(video_src_url))
|
||||||
|
|||||||
@@ -142,6 +142,9 @@ class NetEaseMusicIE(NetEaseMusicBaseIE):
|
|||||||
'subtitles': {'lyrics': [{'ext': 'lrc'}]},
|
'subtitles': {'lyrics': [{'ext': 'lrc'}]},
|
||||||
"duration": 256,
|
"duration": 256,
|
||||||
'thumbnail': r're:^http.*\.jpg',
|
'thumbnail': r're:^http.*\.jpg',
|
||||||
|
'album': '偶像练习生 表演曲目合集',
|
||||||
|
'average_rating': int,
|
||||||
|
'album_artist': '偶像练习生',
|
||||||
},
|
},
|
||||||
}, {
|
}, {
|
||||||
'note': 'No lyrics.',
|
'note': 'No lyrics.',
|
||||||
@@ -155,6 +158,9 @@ class NetEaseMusicIE(NetEaseMusicBaseIE):
|
|||||||
'timestamp': 1202745600,
|
'timestamp': 1202745600,
|
||||||
'duration': 263,
|
'duration': 263,
|
||||||
'thumbnail': r're:^http.*\.jpg',
|
'thumbnail': r're:^http.*\.jpg',
|
||||||
|
'album': 'Piano Solos Vol. 2',
|
||||||
|
'album_artist': 'Dustin O\'Halloran',
|
||||||
|
'average_rating': int,
|
||||||
},
|
},
|
||||||
}, {
|
}, {
|
||||||
'url': 'https://y.music.163.com/m/song?app_version=8.8.45&id=95670&uct2=sKnvS4+0YStsWkqsPhFijw%3D%3D&dlt=0846',
|
'url': 'https://y.music.163.com/m/song?app_version=8.8.45&id=95670&uct2=sKnvS4+0YStsWkqsPhFijw%3D%3D&dlt=0846',
|
||||||
@@ -171,6 +177,9 @@ class NetEaseMusicIE(NetEaseMusicBaseIE):
|
|||||||
'duration': 268,
|
'duration': 268,
|
||||||
'alt_title': '伴唱:现代人乐队 合唱:总政歌舞团',
|
'alt_title': '伴唱:现代人乐队 合唱:总政歌舞团',
|
||||||
'thumbnail': r're:^http.*\.jpg',
|
'thumbnail': r're:^http.*\.jpg',
|
||||||
|
'average_rating': int,
|
||||||
|
'album': '红色摇滚',
|
||||||
|
'album_artist': '侯牧人',
|
||||||
},
|
},
|
||||||
}, {
|
}, {
|
||||||
'url': 'http://music.163.com/#/song?id=32102397',
|
'url': 'http://music.163.com/#/song?id=32102397',
|
||||||
@@ -186,6 +195,9 @@ class NetEaseMusicIE(NetEaseMusicBaseIE):
|
|||||||
'subtitles': {'lyrics': [{'ext': 'lrc'}]},
|
'subtitles': {'lyrics': [{'ext': 'lrc'}]},
|
||||||
'duration': 199,
|
'duration': 199,
|
||||||
'thumbnail': r're:^http.*\.jpg',
|
'thumbnail': r're:^http.*\.jpg',
|
||||||
|
'album': 'Bad Blood',
|
||||||
|
'average_rating': int,
|
||||||
|
'album_artist': 'Taylor Swift',
|
||||||
},
|
},
|
||||||
'skip': 'Blocked outside Mainland China',
|
'skip': 'Blocked outside Mainland China',
|
||||||
}, {
|
}, {
|
||||||
@@ -203,6 +215,9 @@ class NetEaseMusicIE(NetEaseMusicBaseIE):
|
|||||||
'duration': 229,
|
'duration': 229,
|
||||||
'alt_title': '说出愿望吧(Genie)',
|
'alt_title': '说出愿望吧(Genie)',
|
||||||
'thumbnail': r're:^http.*\.jpg',
|
'thumbnail': r're:^http.*\.jpg',
|
||||||
|
'average_rating': int,
|
||||||
|
'album': 'Oh!',
|
||||||
|
'album_artist': '少女时代',
|
||||||
},
|
},
|
||||||
'skip': 'Blocked outside Mainland China',
|
'skip': 'Blocked outside Mainland China',
|
||||||
}]
|
}]
|
||||||
@@ -253,12 +268,15 @@ class NetEaseMusicIE(NetEaseMusicBaseIE):
|
|||||||
'formats': formats,
|
'formats': formats,
|
||||||
'alt_title': '/'.join(traverse_obj(info, (('transNames', 'alias'), ...))) or None,
|
'alt_title': '/'.join(traverse_obj(info, (('transNames', 'alias'), ...))) or None,
|
||||||
'creator': ' / '.join(traverse_obj(info, ('artists', ..., 'name'))) or None,
|
'creator': ' / '.join(traverse_obj(info, ('artists', ..., 'name'))) or None,
|
||||||
|
'album_artist': ' / '.join(traverse_obj(info, ('album', 'artists', ..., 'name'))) or None,
|
||||||
**lyric_data,
|
**lyric_data,
|
||||||
**traverse_obj(info, {
|
**traverse_obj(info, {
|
||||||
'title': ('name', {str}),
|
'title': ('name', {str}),
|
||||||
'timestamp': ('album', 'publishTime', {self.kilo_or_none}),
|
'timestamp': ('album', 'publishTime', {self.kilo_or_none}),
|
||||||
'thumbnail': ('album', 'picUrl', {url_or_none}),
|
'thumbnail': ('album', 'picUrl', {url_or_none}),
|
||||||
'duration': ('duration', {self.kilo_or_none}),
|
'duration': ('duration', {self.kilo_or_none}),
|
||||||
|
'album': ('album', 'name', {str}),
|
||||||
|
'average_rating': ('score', {int_or_none}),
|
||||||
}),
|
}),
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -3,6 +3,8 @@ import re
|
|||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from ..utils import (
|
from ..utils import (
|
||||||
ExtractorError,
|
ExtractorError,
|
||||||
|
clean_html,
|
||||||
|
get_element_by_class,
|
||||||
int_or_none,
|
int_or_none,
|
||||||
join_nonempty,
|
join_nonempty,
|
||||||
parse_duration,
|
parse_duration,
|
||||||
@@ -45,25 +47,36 @@ class NhkBaseIE(InfoExtractor):
|
|||||||
self.cache.store('nhk', 'api_info', api_info)
|
self.cache.store('nhk', 'api_info', api_info)
|
||||||
return api_info
|
return api_info
|
||||||
|
|
||||||
def _extract_formats_and_subtitles(self, vod_id):
|
def _extract_stream_info(self, vod_id):
|
||||||
for refresh in (False, True):
|
for refresh in (False, True):
|
||||||
api_info = self._get_api_info(refresh)
|
api_info = self._get_api_info(refresh)
|
||||||
if not api_info:
|
if not api_info:
|
||||||
continue
|
continue
|
||||||
|
|
||||||
api_url = api_info.pop('url')
|
api_url = api_info.pop('url')
|
||||||
stream_url = traverse_obj(
|
meta = traverse_obj(
|
||||||
self._download_json(
|
self._download_json(
|
||||||
api_url, vod_id, 'Downloading stream url info', fatal=False, query={
|
api_url, vod_id, 'Downloading stream url info', fatal=False, query={
|
||||||
**api_info,
|
**api_info,
|
||||||
'type': 'json',
|
'type': 'json',
|
||||||
'optional_id': vod_id,
|
'optional_id': vod_id,
|
||||||
'active_flg': 1,
|
'active_flg': 1,
|
||||||
}),
|
}), ('meta', 0))
|
||||||
('meta', 0, 'movie_url', ('mb_auto', 'auto_sp', 'auto_pc'), {url_or_none}), get_all=False)
|
stream_url = traverse_obj(
|
||||||
if stream_url:
|
meta, ('movie_url', ('mb_auto', 'auto_sp', 'auto_pc'), {url_or_none}), get_all=False)
|
||||||
return self._extract_m3u8_formats_and_subtitles(stream_url, vod_id)
|
|
||||||
|
|
||||||
|
if stream_url:
|
||||||
|
formats, subtitles = self._extract_m3u8_formats_and_subtitles(stream_url, vod_id)
|
||||||
|
return {
|
||||||
|
**traverse_obj(meta, {
|
||||||
|
'duration': ('duration', {int_or_none}),
|
||||||
|
'timestamp': ('publication_date', {unified_timestamp}),
|
||||||
|
'release_timestamp': ('insert_date', {unified_timestamp}),
|
||||||
|
'modified_timestamp': ('update_date', {unified_timestamp}),
|
||||||
|
}),
|
||||||
|
'formats': formats,
|
||||||
|
'subtitles': subtitles,
|
||||||
|
}
|
||||||
raise ExtractorError('Unable to extract stream url')
|
raise ExtractorError('Unable to extract stream url')
|
||||||
|
|
||||||
def _extract_episode_info(self, url, episode=None):
|
def _extract_episode_info(self, url, episode=None):
|
||||||
@@ -77,11 +90,11 @@ class NhkBaseIE(InfoExtractor):
|
|||||||
if fetch_episode:
|
if fetch_episode:
|
||||||
episode = self._call_api(
|
episode = self._call_api(
|
||||||
episode_id, lang, is_video, True, episode_id[:4] == '9999')[0]
|
episode_id, lang, is_video, True, episode_id[:4] == '9999')[0]
|
||||||
title = episode.get('sub_title_clean') or episode['sub_title']
|
|
||||||
|
|
||||||
def get_clean_field(key):
|
def get_clean_field(key):
|
||||||
return episode.get(key + '_clean') or episode.get(key)
|
return clean_html(episode.get(key + '_clean') or episode.get(key))
|
||||||
|
|
||||||
|
title = get_clean_field('sub_title')
|
||||||
series = get_clean_field('title')
|
series = get_clean_field('title')
|
||||||
|
|
||||||
thumbnails = []
|
thumbnails = []
|
||||||
@@ -96,22 +109,30 @@ class NhkBaseIE(InfoExtractor):
|
|||||||
'url': 'https://www3.nhk.or.jp' + img_path,
|
'url': 'https://www3.nhk.or.jp' + img_path,
|
||||||
})
|
})
|
||||||
|
|
||||||
|
episode_name = title
|
||||||
|
if series and title:
|
||||||
|
title = f'{series} - {title}'
|
||||||
|
elif series and not title:
|
||||||
|
title = series
|
||||||
|
series = None
|
||||||
|
episode_name = None
|
||||||
|
else: # title, no series
|
||||||
|
episode_name = None
|
||||||
|
|
||||||
info = {
|
info = {
|
||||||
'id': episode_id + '-' + lang,
|
'id': episode_id + '-' + lang,
|
||||||
'title': '%s - %s' % (series, title) if series and title else title,
|
'title': title,
|
||||||
'description': get_clean_field('description'),
|
'description': get_clean_field('description'),
|
||||||
'thumbnails': thumbnails,
|
'thumbnails': thumbnails,
|
||||||
'series': series,
|
'series': series,
|
||||||
'episode': title,
|
'episode': episode_name,
|
||||||
}
|
}
|
||||||
|
|
||||||
if is_video:
|
if is_video:
|
||||||
vod_id = episode['vod_id']
|
vod_id = episode['vod_id']
|
||||||
formats, subs = self._extract_formats_and_subtitles(vod_id)
|
|
||||||
|
|
||||||
info.update({
|
info.update({
|
||||||
|
**self._extract_stream_info(vod_id),
|
||||||
'id': vod_id,
|
'id': vod_id,
|
||||||
'formats': formats,
|
|
||||||
'subtitles': subs,
|
|
||||||
})
|
})
|
||||||
|
|
||||||
else:
|
else:
|
||||||
@@ -148,6 +169,14 @@ class NhkVodIE(NhkBaseIE):
|
|||||||
'thumbnail': 'md5:51bcef4a21936e7fea1ff4e06353f463',
|
'thumbnail': 'md5:51bcef4a21936e7fea1ff4e06353f463',
|
||||||
'episode': 'The Tohoku Shinkansen: Full Speed Ahead',
|
'episode': 'The Tohoku Shinkansen: Full Speed Ahead',
|
||||||
'series': 'Japan Railway Journal',
|
'series': 'Japan Railway Journal',
|
||||||
|
'modified_timestamp': 1694243656,
|
||||||
|
'timestamp': 1681428600,
|
||||||
|
'release_timestamp': 1693883728,
|
||||||
|
'duration': 1679,
|
||||||
|
'upload_date': '20230413',
|
||||||
|
'modified_date': '20230909',
|
||||||
|
'release_date': '20230905',
|
||||||
|
|
||||||
},
|
},
|
||||||
}, {
|
}, {
|
||||||
# video clip
|
# video clip
|
||||||
@@ -161,6 +190,13 @@ class NhkVodIE(NhkBaseIE):
|
|||||||
'thumbnail': 'md5:d6a4d9b6e9be90aaadda0bcce89631ed',
|
'thumbnail': 'md5:d6a4d9b6e9be90aaadda0bcce89631ed',
|
||||||
'series': 'Dining with the Chef',
|
'series': 'Dining with the Chef',
|
||||||
'episode': 'Chef Saito\'s Family recipe: MENCHI-KATSU',
|
'episode': 'Chef Saito\'s Family recipe: MENCHI-KATSU',
|
||||||
|
'duration': 148,
|
||||||
|
'upload_date': '20190816',
|
||||||
|
'release_date': '20230902',
|
||||||
|
'release_timestamp': 1693619292,
|
||||||
|
'modified_timestamp': 1694168033,
|
||||||
|
'modified_date': '20230908',
|
||||||
|
'timestamp': 1565997540,
|
||||||
},
|
},
|
||||||
}, {
|
}, {
|
||||||
# radio
|
# radio
|
||||||
@@ -170,7 +206,7 @@ class NhkVodIE(NhkBaseIE):
|
|||||||
'ext': 'm4a',
|
'ext': 'm4a',
|
||||||
'title': 'Living in Japan - Tips for Travelers to Japan / Ramen Vending Machines',
|
'title': 'Living in Japan - Tips for Travelers to Japan / Ramen Vending Machines',
|
||||||
'series': 'Living in Japan',
|
'series': 'Living in Japan',
|
||||||
'description': 'md5:850611969932874b4a3309e0cae06c2f',
|
'description': 'md5:0a0e2077d8f07a03071e990a6f51bfab',
|
||||||
'thumbnail': 'md5:960622fb6e06054a4a1a0c97ea752545',
|
'thumbnail': 'md5:960622fb6e06054a4a1a0c97ea752545',
|
||||||
'episode': 'Tips for Travelers to Japan / Ramen Vending Machines'
|
'episode': 'Tips for Travelers to Japan / Ramen Vending Machines'
|
||||||
},
|
},
|
||||||
@@ -212,6 +248,23 @@ class NhkVodIE(NhkBaseIE):
|
|||||||
'description': 'md5:9c1d6cbeadb827b955b20e99ab920ff0',
|
'description': 'md5:9c1d6cbeadb827b955b20e99ab920ff0',
|
||||||
},
|
},
|
||||||
'skip': 'expires 2023-10-15',
|
'skip': 'expires 2023-10-15',
|
||||||
|
}, {
|
||||||
|
# a one-off (single-episode series). title from the api is just '<p></p>'
|
||||||
|
'url': 'https://www3.nhk.or.jp/nhkworld/en/ondemand/video/3004952/',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'nw_vod_v_en_3004_952_20230723091000_01_1690074552',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': 'Barakan Discovers AMAMI OSHIMA: Isson\'s Treasure Island',
|
||||||
|
'description': 'md5:5db620c46a0698451cc59add8816b797',
|
||||||
|
'thumbnail': 'md5:67d9ff28009ba379bfa85ad1aaa0e2bd',
|
||||||
|
'release_date': '20230905',
|
||||||
|
'timestamp': 1690103400,
|
||||||
|
'duration': 2939,
|
||||||
|
'release_timestamp': 1693898699,
|
||||||
|
'modified_timestamp': 1698057495,
|
||||||
|
'modified_date': '20231023',
|
||||||
|
'upload_date': '20230723',
|
||||||
|
},
|
||||||
}]
|
}]
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
@@ -226,13 +279,15 @@ class NhkVodProgramIE(NhkBaseIE):
|
|||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'sumo',
|
'id': 'sumo',
|
||||||
'title': 'GRAND SUMO Highlights',
|
'title': 'GRAND SUMO Highlights',
|
||||||
|
'description': 'md5:fc20d02dc6ce85e4b72e0273aa52fdbf',
|
||||||
},
|
},
|
||||||
'playlist_mincount': 12,
|
'playlist_mincount': 0,
|
||||||
}, {
|
}, {
|
||||||
'url': 'https://www3.nhk.or.jp/nhkworld/en/ondemand/program/video/japanrailway',
|
'url': 'https://www3.nhk.or.jp/nhkworld/en/ondemand/program/video/japanrailway',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'japanrailway',
|
'id': 'japanrailway',
|
||||||
'title': 'Japan Railway Journal',
|
'title': 'Japan Railway Journal',
|
||||||
|
'description': 'md5:ea39d93af7d05835baadf10d1aae0e3f',
|
||||||
},
|
},
|
||||||
'playlist_mincount': 12,
|
'playlist_mincount': 12,
|
||||||
}, {
|
}, {
|
||||||
@@ -241,6 +296,7 @@ class NhkVodProgramIE(NhkBaseIE):
|
|||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'japanrailway',
|
'id': 'japanrailway',
|
||||||
'title': 'Japan Railway Journal',
|
'title': 'Japan Railway Journal',
|
||||||
|
'description': 'md5:ea39d93af7d05835baadf10d1aae0e3f',
|
||||||
},
|
},
|
||||||
'playlist_mincount': 5,
|
'playlist_mincount': 5,
|
||||||
}, {
|
}, {
|
||||||
@@ -265,11 +321,11 @@ class NhkVodProgramIE(NhkBaseIE):
|
|||||||
entries.append(self._extract_episode_info(
|
entries.append(self._extract_episode_info(
|
||||||
urljoin(url, episode_path), episode))
|
urljoin(url, episode_path), episode))
|
||||||
|
|
||||||
program_title = None
|
html = self._download_webpage(url, program_id)
|
||||||
if entries:
|
program_title = clean_html(get_element_by_class('p-programDetail__title', html))
|
||||||
program_title = entries[0].get('series')
|
program_description = clean_html(get_element_by_class('p-programDetail__text', html))
|
||||||
|
|
||||||
return self.playlist_result(entries, program_id, program_title)
|
return self.playlist_result(entries, program_id, program_title, program_description)
|
||||||
|
|
||||||
|
|
||||||
class NhkForSchoolBangumiIE(InfoExtractor):
|
class NhkForSchoolBangumiIE(InfoExtractor):
|
||||||
@@ -421,6 +477,7 @@ class NhkRadiruIE(InfoExtractor):
|
|||||||
'skip': 'Episode expired on 2023-04-16',
|
'skip': 'Episode expired on 2023-04-16',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'channel': 'NHK-FM',
|
'channel': 'NHK-FM',
|
||||||
|
'uploader': 'NHK-FM',
|
||||||
'description': 'md5:94b08bdeadde81a97df4ec882acce3e9',
|
'description': 'md5:94b08bdeadde81a97df4ec882acce3e9',
|
||||||
'ext': 'm4a',
|
'ext': 'm4a',
|
||||||
'id': '0449_01_3853544',
|
'id': '0449_01_3853544',
|
||||||
@@ -441,6 +498,7 @@ class NhkRadiruIE(InfoExtractor):
|
|||||||
'title': 'ベストオブクラシック',
|
'title': 'ベストオブクラシック',
|
||||||
'description': '世界中の上質な演奏会をじっくり堪能する本格派クラシック番組。',
|
'description': '世界中の上質な演奏会をじっくり堪能する本格派クラシック番組。',
|
||||||
'channel': 'NHK-FM',
|
'channel': 'NHK-FM',
|
||||||
|
'uploader': 'NHK-FM',
|
||||||
'thumbnail': 'https://www.nhk.or.jp/prog/img/458/g458.jpg',
|
'thumbnail': 'https://www.nhk.or.jp/prog/img/458/g458.jpg',
|
||||||
},
|
},
|
||||||
'playlist_mincount': 3,
|
'playlist_mincount': 3,
|
||||||
@@ -454,6 +512,7 @@ class NhkRadiruIE(InfoExtractor):
|
|||||||
'title': '有島武郎「一房のぶどう」',
|
'title': '有島武郎「一房のぶどう」',
|
||||||
'description': '朗読:川野一宇(ラジオ深夜便アンカー)\r\n\r\n(2016年12月8日放送「ラジオ深夜便『アンカー朗読シリーズ』」より)',
|
'description': '朗読:川野一宇(ラジオ深夜便アンカー)\r\n\r\n(2016年12月8日放送「ラジオ深夜便『アンカー朗読シリーズ』」より)',
|
||||||
'channel': 'NHKラジオ第1、NHK-FM',
|
'channel': 'NHKラジオ第1、NHK-FM',
|
||||||
|
'uploader': 'NHKラジオ第1、NHK-FM',
|
||||||
'timestamp': 1635757200,
|
'timestamp': 1635757200,
|
||||||
'thumbnail': 'https://www.nhk.or.jp/radioondemand/json/F300/img/corner/box_109_thumbnail.jpg',
|
'thumbnail': 'https://www.nhk.or.jp/radioondemand/json/F300/img/corner/box_109_thumbnail.jpg',
|
||||||
'release_date': '20161207',
|
'release_date': '20161207',
|
||||||
@@ -469,6 +528,7 @@ class NhkRadiruIE(InfoExtractor):
|
|||||||
'id': 'F261_01_3855109',
|
'id': 'F261_01_3855109',
|
||||||
'ext': 'm4a',
|
'ext': 'm4a',
|
||||||
'channel': 'NHKラジオ第1',
|
'channel': 'NHKラジオ第1',
|
||||||
|
'uploader': 'NHKラジオ第1',
|
||||||
'timestamp': 1681635900,
|
'timestamp': 1681635900,
|
||||||
'release_date': '20230416',
|
'release_date': '20230416',
|
||||||
'series': 'NHKラジオニュース',
|
'series': 'NHKラジオニュース',
|
||||||
@@ -513,6 +573,7 @@ class NhkRadiruIE(InfoExtractor):
|
|||||||
series_meta = traverse_obj(meta, {
|
series_meta = traverse_obj(meta, {
|
||||||
'title': 'program_name',
|
'title': 'program_name',
|
||||||
'channel': 'media_name',
|
'channel': 'media_name',
|
||||||
|
'uploader': 'media_name',
|
||||||
'thumbnail': (('thumbnail_c', 'thumbnail_p'), {url_or_none}),
|
'thumbnail': (('thumbnail_c', 'thumbnail_p'), {url_or_none}),
|
||||||
}, get_all=False)
|
}, get_all=False)
|
||||||
|
|
||||||
@@ -541,6 +602,7 @@ class NhkRadioNewsPageIE(InfoExtractor):
|
|||||||
'thumbnail': 'https://www.nhk.or.jp/radioondemand/json/F261/img/RADIONEWS_640.jpg',
|
'thumbnail': 'https://www.nhk.or.jp/radioondemand/json/F261/img/RADIONEWS_640.jpg',
|
||||||
'description': 'md5:bf2c5b397e44bc7eb26de98d8f15d79d',
|
'description': 'md5:bf2c5b397e44bc7eb26de98d8f15d79d',
|
||||||
'channel': 'NHKラジオ第1',
|
'channel': 'NHKラジオ第1',
|
||||||
|
'uploader': 'NHKラジオ第1',
|
||||||
'title': 'NHKラジオニュース',
|
'title': 'NHKラジオニュース',
|
||||||
}
|
}
|
||||||
}]
|
}]
|
||||||
|
|||||||
@@ -1,82 +0,0 @@
|
|||||||
import re
|
|
||||||
|
|
||||||
from .common import InfoExtractor
|
|
||||||
from ..compat import compat_urlparse
|
|
||||||
from ..utils import (
|
|
||||||
get_element_by_class,
|
|
||||||
urlencode_postdata,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class NJPWWorldIE(InfoExtractor):
|
|
||||||
_VALID_URL = r'https?://(front\.)?njpwworld\.com/p/(?P<id>[a-z0-9_]+)'
|
|
||||||
IE_DESC = '新日本プロレスワールド'
|
|
||||||
_NETRC_MACHINE = 'njpwworld'
|
|
||||||
|
|
||||||
_TESTS = [{
|
|
||||||
'url': 'http://njpwworld.com/p/s_series_00155_1_9/',
|
|
||||||
'info_dict': {
|
|
||||||
'id': 's_series_00155_1_9',
|
|
||||||
'ext': 'mp4',
|
|
||||||
'title': '闘強導夢2000 2000年1月4日 東京ドーム 第9試合 ランディ・サベージ VS リック・スタイナー',
|
|
||||||
'tags': list,
|
|
||||||
},
|
|
||||||
'params': {
|
|
||||||
'skip_download': True, # AES-encrypted m3u8
|
|
||||||
},
|
|
||||||
'skip': 'Requires login',
|
|
||||||
}, {
|
|
||||||
'url': 'https://front.njpwworld.com/p/s_series_00563_16_bs',
|
|
||||||
'info_dict': {
|
|
||||||
'id': 's_series_00563_16_bs',
|
|
||||||
'ext': 'mp4',
|
|
||||||
'title': 'WORLD TAG LEAGUE 2020 & BEST OF THE SUPER Jr.27 2020年12月6日 福岡・福岡国際センター バックステージコメント(字幕あり)',
|
|
||||||
'tags': ["福岡・福岡国際センター", "バックステージコメント", "2020", "20年代"],
|
|
||||||
},
|
|
||||||
'params': {
|
|
||||||
'skip_download': True,
|
|
||||||
},
|
|
||||||
}]
|
|
||||||
|
|
||||||
_LOGIN_URL = 'https://front.njpwworld.com/auth/login'
|
|
||||||
|
|
||||||
def _perform_login(self, username, password):
|
|
||||||
# Setup session (will set necessary cookies)
|
|
||||||
self._request_webpage(
|
|
||||||
'https://njpwworld.com/', None, note='Setting up session')
|
|
||||||
|
|
||||||
webpage, urlh = self._download_webpage_handle(
|
|
||||||
self._LOGIN_URL, None,
|
|
||||||
note='Logging in', errnote='Unable to login',
|
|
||||||
data=urlencode_postdata({'login_id': username, 'pw': password}),
|
|
||||||
headers={'Referer': 'https://front.njpwworld.com/auth'})
|
|
||||||
# /auth/login will return 302 for successful logins
|
|
||||||
if urlh.url == self._LOGIN_URL:
|
|
||||||
self.report_warning('unable to login')
|
|
||||||
return False
|
|
||||||
|
|
||||||
return True
|
|
||||||
|
|
||||||
def _real_extract(self, url):
|
|
||||||
video_id = self._match_id(url)
|
|
||||||
|
|
||||||
webpage = self._download_webpage(url, video_id)
|
|
||||||
|
|
||||||
formats = []
|
|
||||||
for kind, vid in re.findall(r'if\s+\(\s*imageQualityType\s*==\s*\'([^\']+)\'\s*\)\s*{\s*video_id\s*=\s*"(\d+)"', webpage):
|
|
||||||
player_path = '/intent?id=%s&type=url' % vid
|
|
||||||
player_url = compat_urlparse.urljoin(url, player_path)
|
|
||||||
formats += self._extract_m3u8_formats(
|
|
||||||
player_url, video_id, 'mp4', 'm3u8_native', m3u8_id=kind, fatal=False, quality=int(kind == 'high'))
|
|
||||||
|
|
||||||
tag_block = get_element_by_class('tag-block', webpage)
|
|
||||||
tags = re.findall(
|
|
||||||
r'<a[^>]+class="tag-[^"]+"[^>]*>([^<]+)</a>', tag_block
|
|
||||||
) if tag_block else None
|
|
||||||
|
|
||||||
return {
|
|
||||||
'id': video_id,
|
|
||||||
'title': get_element_by_class('article-title', webpage) or self._og_search_title(webpage),
|
|
||||||
'formats': formats,
|
|
||||||
'tags': tags,
|
|
||||||
}
|
|
||||||
@@ -13,7 +13,7 @@ from ..utils import (
|
|||||||
|
|
||||||
|
|
||||||
class NovaEmbedIE(InfoExtractor):
|
class NovaEmbedIE(InfoExtractor):
|
||||||
_VALID_URL = r'https?://media\.cms\.nova\.cz/embed/(?P<id>[^/?#&]+)'
|
_VALID_URL = r'https?://media(?:tn)?\.cms\.nova\.cz/embed/(?P<id>[^/?#&]+)'
|
||||||
_TESTS = [{
|
_TESTS = [{
|
||||||
'url': 'https://media.cms.nova.cz/embed/8o0n0r?autoplay=1',
|
'url': 'https://media.cms.nova.cz/embed/8o0n0r?autoplay=1',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
@@ -37,6 +37,16 @@ class NovaEmbedIE(InfoExtractor):
|
|||||||
'duration': 114,
|
'duration': 114,
|
||||||
},
|
},
|
||||||
'params': {'skip_download': 'm3u8'},
|
'params': {'skip_download': 'm3u8'},
|
||||||
|
}, {
|
||||||
|
'url': 'https://mediatn.cms.nova.cz/embed/EU5ELEsmOHt?autoplay=1',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'EU5ELEsmOHt',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': 'Haptické křeslo, bionická ruka nebo roboti. Reportérka se podívala na Týden inovací',
|
||||||
|
'thumbnail': r're:^https?://.*\.jpg',
|
||||||
|
'duration': 1780,
|
||||||
|
},
|
||||||
|
'params': {'skip_download': 'm3u8'},
|
||||||
}]
|
}]
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
|
|||||||
@@ -245,7 +245,7 @@ class NPOIE(InfoExtractor):
|
|||||||
'quality': 'npoplus',
|
'quality': 'npoplus',
|
||||||
'tokenId': player_token,
|
'tokenId': player_token,
|
||||||
'streamType': 'broadcast',
|
'streamType': 'broadcast',
|
||||||
})
|
}, data=b'') # endpoint requires POST
|
||||||
if not streams:
|
if not streams:
|
||||||
continue
|
continue
|
||||||
stream = streams.get('stream')
|
stream = streams.get('stream')
|
||||||
|
|||||||
@@ -1,21 +1,21 @@
|
|||||||
import re
|
import re
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from ..compat import compat_urlparse
|
|
||||||
from ..utils import (
|
from ..utils import (
|
||||||
int_or_none,
|
int_or_none,
|
||||||
js_to_json,
|
js_to_json,
|
||||||
parse_duration,
|
url_or_none,
|
||||||
)
|
)
|
||||||
|
from ..utils.traversal import traverse_obj
|
||||||
|
|
||||||
|
|
||||||
class NTVDeIE(InfoExtractor):
|
class NTVDeIE(InfoExtractor):
|
||||||
IE_NAME = 'n-tv.de'
|
IE_NAME = 'n-tv.de'
|
||||||
_VALID_URL = r'https?://(?:www\.)?n-tv\.de/mediathek/videos/[^/?#]+/[^/?#]+-article(?P<id>.+)\.html'
|
_VALID_URL = r'https?://(?:www\.)?n-tv\.de/mediathek/(?:videos|magazine)/[^/?#]+/[^/?#]+-article(?P<id>[^/?#]+)\.html'
|
||||||
|
|
||||||
_TESTS = [{
|
_TESTS = [{
|
||||||
'url': 'http://www.n-tv.de/mediathek/videos/panorama/Schnee-und-Glaette-fuehren-zu-zahlreichen-Unfaellen-und-Staus-article14438086.html',
|
'url': 'http://www.n-tv.de/mediathek/videos/panorama/Schnee-und-Glaette-fuehren-zu-zahlreichen-Unfaellen-und-Staus-article14438086.html',
|
||||||
'md5': '6ef2514d4b1e8e03ca24b49e2f167153',
|
'md5': '6bcf2a6638cb83f45d5561659a1cb498',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': '14438086',
|
'id': '14438086',
|
||||||
'ext': 'mp4',
|
'ext': 'mp4',
|
||||||
@@ -23,51 +23,61 @@ class NTVDeIE(InfoExtractor):
|
|||||||
'title': 'Schnee und Glätte führen zu zahlreichen Unfällen und Staus',
|
'title': 'Schnee und Glätte führen zu zahlreichen Unfällen und Staus',
|
||||||
'alt_title': 'Winterchaos auf deutschen Straßen',
|
'alt_title': 'Winterchaos auf deutschen Straßen',
|
||||||
'description': 'Schnee und Glätte sorgen deutschlandweit für einen chaotischen Start in die Woche: Auf den Straßen kommt es zu kilometerlangen Staus und Dutzenden Glätteunfällen. In Düsseldorf und München wirbelt der Schnee zudem den Flugplan durcheinander. Dutzende Flüge landen zu spät, einige fallen ganz aus.',
|
'description': 'Schnee und Glätte sorgen deutschlandweit für einen chaotischen Start in die Woche: Auf den Straßen kommt es zu kilometerlangen Staus und Dutzenden Glätteunfällen. In Düsseldorf und München wirbelt der Schnee zudem den Flugplan durcheinander. Dutzende Flüge landen zu spät, einige fallen ganz aus.',
|
||||||
'duration': 4020,
|
'duration': 67,
|
||||||
'timestamp': 1422892797,
|
'timestamp': 1422892797,
|
||||||
'upload_date': '20150202',
|
'upload_date': '20150202',
|
||||||
},
|
},
|
||||||
|
}, {
|
||||||
|
'url': 'https://www.n-tv.de/mediathek/magazine/auslandsreport/Juedische-Siedler-wollten-Rache-die-wollten-nur-toeten-article24523089.html',
|
||||||
|
'md5': 'c5c6014c014ccc3359470e1d34472bfd',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '24523089',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'thumbnail': r're:^https?://.*\.jpg$',
|
||||||
|
'title': 'Jüdische Siedler "wollten Rache, die wollten nur töten"',
|
||||||
|
'alt_title': 'Israelische Gewalt fern von Gaza',
|
||||||
|
'description': 'Vier Tage nach dem Massaker der Hamas greifen jüdische Siedler das Haus einer palästinensischen Familie im Westjordanland an. Die Überlebenden berichten, sie waren unbewaffnet, die Angreifer seien nur auf "Rache und Töten" aus gewesen. Als die Toten beerdigt werden sollen, eröffnen die Siedler erneut das Feuer.',
|
||||||
|
'duration': 326,
|
||||||
|
'timestamp': 1699688294,
|
||||||
|
'upload_date': '20231111',
|
||||||
|
},
|
||||||
}]
|
}]
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
video_id = self._match_id(url)
|
video_id = self._match_id(url)
|
||||||
webpage = self._download_webpage(url, video_id)
|
webpage = self._download_webpage(url, video_id)
|
||||||
|
|
||||||
info = self._parse_json(self._search_regex(
|
info = self._search_json(
|
||||||
r'(?s)ntv\.pageInfo\.article\s*=\s*(\{.*?\});', webpage, 'info'),
|
r'article:', webpage, 'info', video_id, transform_source=js_to_json)
|
||||||
video_id, transform_source=js_to_json)
|
|
||||||
timestamp = int_or_none(info.get('publishedDateAsUnixTimeStamp'))
|
vdata = self._search_json(
|
||||||
vdata = self._parse_json(self._search_regex(
|
r'\$\(\s*"#playerwrapper"\s*\)\s*\.data\(\s*"player",',
|
||||||
r'(?s)\$\(\s*"\#player"\s*\)\s*\.data\(\s*"player",\s*(\{.*?\})\);',
|
webpage, 'player data', video_id,
|
||||||
webpage, 'player data'), video_id,
|
transform_source=lambda s: js_to_json(re.sub(r'ivw:[^},]+', '', s)))['setup']['source']
|
||||||
transform_source=lambda s: js_to_json(re.sub(r'advertising:\s*{[^}]+},', '', s)))
|
|
||||||
duration = parse_duration(vdata.get('duration'))
|
|
||||||
|
|
||||||
formats = []
|
formats = []
|
||||||
if vdata.get('video'):
|
if vdata.get('progressive'):
|
||||||
formats.append({
|
formats.append({
|
||||||
'format_id': 'flash',
|
'format_id': 'http',
|
||||||
'url': 'rtmp://fms.n-tv.de/%s' % vdata['video'],
|
'url': vdata['progressive'],
|
||||||
})
|
})
|
||||||
if vdata.get('videoMp4'):
|
if vdata.get('hls'):
|
||||||
formats.append({
|
|
||||||
'format_id': 'mobile',
|
|
||||||
'url': compat_urlparse.urljoin('http://video.n-tv.de', vdata['videoMp4']),
|
|
||||||
'tbr': 400, # estimation
|
|
||||||
})
|
|
||||||
if vdata.get('videoM3u8'):
|
|
||||||
m3u8_url = compat_urlparse.urljoin('http://video.n-tv.de', vdata['videoM3u8'])
|
|
||||||
formats.extend(self._extract_m3u8_formats(
|
formats.extend(self._extract_m3u8_formats(
|
||||||
m3u8_url, video_id, ext='mp4', entry_protocol='m3u8_native',
|
vdata['hls'], video_id, 'mp4', m3u8_id='hls', fatal=False))
|
||||||
quality=1, m3u8_id='hls', fatal=False))
|
if vdata.get('dash'):
|
||||||
|
formats.extend(self._extract_mpd_formats(vdata['dash'], video_id, fatal=False, mpd_id='dash'))
|
||||||
|
|
||||||
return {
|
return {
|
||||||
'id': video_id,
|
'id': video_id,
|
||||||
'title': info['headline'],
|
**traverse_obj(info, {
|
||||||
'description': info.get('intro'),
|
'title': 'headline',
|
||||||
'alt_title': info.get('kicker'),
|
'description': 'intro',
|
||||||
'timestamp': timestamp,
|
'alt_title': 'kicker',
|
||||||
'thumbnail': vdata.get('html5VideoPoster'),
|
'timestamp': ('publishedDateAsUnixTimeStamp', {int_or_none}),
|
||||||
'duration': duration,
|
}),
|
||||||
|
**traverse_obj(vdata, {
|
||||||
|
'thumbnail': ('poster', {url_or_none}),
|
||||||
|
'duration': ('length', {int_or_none}),
|
||||||
|
}),
|
||||||
'formats': formats,
|
'formats': formats,
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,87 +1,167 @@
|
|||||||
|
import functools
|
||||||
import re
|
import re
|
||||||
|
import uuid
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
|
from ..networking import HEADRequest
|
||||||
from ..utils import (
|
from ..utils import (
|
||||||
ExtractorError,
|
ExtractorError,
|
||||||
js_to_json,
|
OnDemandPagedList,
|
||||||
|
float_or_none,
|
||||||
|
int_or_none,
|
||||||
|
join_nonempty,
|
||||||
|
parse_age_limit,
|
||||||
|
parse_qs,
|
||||||
|
unified_strdate,
|
||||||
|
url_or_none,
|
||||||
)
|
)
|
||||||
|
from ..utils.traversal import traverse_obj
|
||||||
|
|
||||||
|
|
||||||
class OnDemandKoreaIE(InfoExtractor):
|
class OnDemandKoreaIE(InfoExtractor):
|
||||||
_VALID_URL = r'https?://(?:www\.)?ondemandkorea\.com/(?P<id>[^/]+)\.html'
|
_VALID_URL = r'https?://(?:www\.)?ondemandkorea\.com/(?:en/)?player/vod/[a-z0-9-]+\?(?:[^#]+&)?contentId=(?P<id>\d+)'
|
||||||
_GEO_COUNTRIES = ['US', 'CA']
|
_GEO_COUNTRIES = ['US', 'CA']
|
||||||
|
|
||||||
_TESTS = [{
|
_TESTS = [{
|
||||||
'url': 'https://www.ondemandkorea.com/ask-us-anything-e351.html',
|
'url': 'https://www.ondemandkorea.com/player/vod/ask-us-anything?contentId=686471',
|
||||||
|
'md5': 'e2ff77255d989e3135bde0c5889fbce8',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'ask-us-anything-e351',
|
'id': '686471',
|
||||||
'ext': 'mp4',
|
'ext': 'mp4',
|
||||||
'title': 'Ask Us Anything : Jung Sung-ho, Park Seul-gi, Kim Bo-min, Yang Seung-won - 09/24/2022',
|
'title': 'Ask Us Anything: Jung Sung-ho, Park Seul-gi, Kim Bo-min, Yang Seung-won',
|
||||||
'description': 'A talk show/game show with a school theme where celebrity guests appear as “transfer students.”',
|
'thumbnail': r're:^https?://.*\.(jpg|jpeg|png)',
|
||||||
'thumbnail': r're:^https?://.*\.jpg$',
|
'duration': 5486.955,
|
||||||
|
'release_date': '20220924',
|
||||||
|
'series': 'Ask Us Anything',
|
||||||
|
'series_id': 11790,
|
||||||
|
'episode_number': 351,
|
||||||
|
'episode': 'Jung Sung-ho, Park Seul-gi, Kim Bo-min, Yang Seung-won',
|
||||||
},
|
},
|
||||||
'params': {
|
|
||||||
'skip_download': 'm3u8 download'
|
|
||||||
}
|
|
||||||
}, {
|
}, {
|
||||||
'url': 'https://www.ondemandkorea.com/work-later-drink-now-e1.html',
|
'url': 'https://www.ondemandkorea.com/player/vod/breakup-probation-a-week?contentId=1595796',
|
||||||
|
'md5': '57266c720006962be7ff415b24775caa',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'work-later-drink-now-e1',
|
'id': '1595796',
|
||||||
'ext': 'mp4',
|
'ext': 'mp4',
|
||||||
'title': 'Work Later, Drink Now : E01',
|
'title': 'Breakup Probation, A Week: E08',
|
||||||
'description': 'Work Later, Drink First follows three women who find solace in a glass of liquor at the end of the day. So-hee, who gets comfort from a cup of soju af',
|
'thumbnail': r're:^https?://.*\.(jpg|jpeg|png)',
|
||||||
'thumbnail': r're:^https?://.*\.png$',
|
'duration': 1586.0,
|
||||||
'subtitles': {
|
'release_date': '20231001',
|
||||||
'English': 'mincount:1',
|
'series': 'Breakup Probation, A Week',
|
||||||
},
|
'series_id': 22912,
|
||||||
|
'episode_number': 8,
|
||||||
|
'episode': 'E08',
|
||||||
},
|
},
|
||||||
'params': {
|
}, {
|
||||||
'skip_download': 'm3u8 download'
|
'url': 'https://www.ondemandkorea.com/player/vod/the-outlaws?contentId=369531',
|
||||||
}
|
'md5': 'fa5523b87aa1f6d74fc622a97f2b47cd',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '369531',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'release_date': '20220519',
|
||||||
|
'duration': 7267.0,
|
||||||
|
'title': 'The Outlaws: Main Movie',
|
||||||
|
'thumbnail': r're:^https?://.*\.(jpg|jpeg|png)',
|
||||||
|
'age_limit': 18,
|
||||||
|
},
|
||||||
|
}, {
|
||||||
|
'url': 'https://www.ondemandkorea.com/en/player/vod/capture-the-moment-how-is-that-possible?contentId=1605006',
|
||||||
|
'only_matching': True,
|
||||||
}]
|
}]
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
video_id = self._match_id(url)
|
video_id = self._match_id(url)
|
||||||
webpage = self._download_webpage(url, video_id, fatal=False)
|
|
||||||
|
|
||||||
if not webpage:
|
data = self._download_json(
|
||||||
# Page sometimes returns captcha page with HTTP 403
|
f'https://odkmedia.io/odx/api/v3/playback/{video_id}/', video_id, fatal=False,
|
||||||
raise ExtractorError(
|
headers={'service-name': 'odk'}, query={'did': str(uuid.uuid4())}, expected_status=(403, 404))
|
||||||
'Unable to access page. You may have been blocked.',
|
if not traverse_obj(data, ('result', {dict})):
|
||||||
expected=True)
|
msg = traverse_obj(data, ('messages', '__default'), 'title', expected_type=str)
|
||||||
|
raise ExtractorError(msg or 'Got empty response from playback API', expected=True)
|
||||||
|
|
||||||
if 'msg_block_01.png' in webpage:
|
data = data['result']
|
||||||
self.raise_geo_restricted(
|
|
||||||
msg='This content is not available in your region',
|
|
||||||
countries=self._GEO_COUNTRIES)
|
|
||||||
|
|
||||||
if 'This video is only available to ODK PLUS members.' in webpage:
|
def try_geo_bypass(url):
|
||||||
raise ExtractorError(
|
return traverse_obj(url, ({parse_qs}, 'stream_url', 0, {url_or_none})) or url
|
||||||
'This video is only available to ODK PLUS members.',
|
|
||||||
expected=True)
|
|
||||||
|
|
||||||
if 'ODK PREMIUM Members Only' in webpage:
|
def try_upgrade_quality(url):
|
||||||
raise ExtractorError(
|
mod_url = re.sub(r'_720(p?)\.m3u8', r'_1080\1.m3u8', url)
|
||||||
'This video is only available to ODK PREMIUM members.',
|
return mod_url if mod_url != url and self._request_webpage(
|
||||||
expected=True)
|
HEADRequest(mod_url), video_id, note='Checking for higher quality format',
|
||||||
|
errnote='No higher quality format found', fatal=False) else url
|
||||||
|
|
||||||
title = self._search_regex(
|
formats = []
|
||||||
r'class=["\']episode_title["\'][^>]*>([^<]+)',
|
for m3u8_url in traverse_obj(data, (('sources', 'manifest'), ..., 'url', {url_or_none}, {try_geo_bypass})):
|
||||||
webpage, 'episode_title', fatal=False) or self._og_search_title(webpage)
|
formats.extend(self._extract_m3u8_formats(try_upgrade_quality(m3u8_url), video_id, fatal=False))
|
||||||
|
|
||||||
jw_config = self._parse_json(
|
subtitles = {}
|
||||||
self._search_regex((
|
for track in traverse_obj(data, ('text_tracks', lambda _, v: url_or_none(v['url']))):
|
||||||
r'(?P<options>{\s*[\'"]tracks[\'"].*?})[)\];]+$',
|
subtitles.setdefault(track.get('language', 'und'), []).append({
|
||||||
r'playlist\s*=\s*\[(?P<options>.+)];?$',
|
'url': track['url'],
|
||||||
r'odkPlayer\.init.*?(?P<options>{[^;]+}).*?;',
|
'ext': track.get('codec'),
|
||||||
), webpage, 'jw config', flags=re.MULTILINE | re.DOTALL, group='options'),
|
'name': track.get('label'),
|
||||||
video_id, transform_source=js_to_json)
|
})
|
||||||
info = self._parse_jwplayer_data(
|
|
||||||
jw_config, video_id, require_title=False, m3u8_id='hls',
|
|
||||||
base_url=url)
|
|
||||||
|
|
||||||
info.update({
|
def if_series(key=None):
|
||||||
'title': title,
|
return lambda obj: obj[key] if key and obj['kind'] == 'series' else None
|
||||||
'description': self._og_search_description(webpage),
|
|
||||||
'thumbnail': self._og_search_thumbnail(webpage)
|
return {
|
||||||
})
|
'id': video_id,
|
||||||
return info
|
'title': join_nonempty(
|
||||||
|
('episode', 'program', 'title'),
|
||||||
|
('episode', 'title'), from_dict=data, delim=': '),
|
||||||
|
**traverse_obj(data, {
|
||||||
|
'thumbnail': ('episode', 'images', 'thumbnail', {url_or_none}),
|
||||||
|
'release_date': ('episode', 'release_date', {lambda x: x.replace('-', '')}, {unified_strdate}),
|
||||||
|
'duration': ('duration', {functools.partial(float_or_none, scale=1000)}),
|
||||||
|
'age_limit': ('age_rating', 'name', {lambda x: x.replace('R', '')}, {parse_age_limit}),
|
||||||
|
'series': ('episode', {if_series(key='program')}, 'title'),
|
||||||
|
'series_id': ('episode', {if_series(key='program')}, 'id'),
|
||||||
|
'episode': ('episode', {if_series(key='title')}),
|
||||||
|
'episode_number': ('episode', {if_series(key='number')}, {int_or_none}),
|
||||||
|
}, get_all=False),
|
||||||
|
'formats': formats,
|
||||||
|
'subtitles': subtitles,
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class OnDemandKoreaProgramIE(InfoExtractor):
|
||||||
|
_VALID_URL = r'https?://(?:www\.)?ondemandkorea\.com/(?:en/)?player/vod/(?P<id>[a-z0-9-]+)(?:$|#)'
|
||||||
|
_GEO_COUNTRIES = ['US', 'CA']
|
||||||
|
|
||||||
|
_TESTS = [{
|
||||||
|
'url': 'https://www.ondemandkorea.com/player/vod/uskn-news',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'uskn-news',
|
||||||
|
},
|
||||||
|
'playlist_mincount': 755,
|
||||||
|
}, {
|
||||||
|
'url': 'https://www.ondemandkorea.com/en/player/vod/the-land',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'the-land',
|
||||||
|
},
|
||||||
|
'playlist_count': 52,
|
||||||
|
}]
|
||||||
|
|
||||||
|
_PAGE_SIZE = 100
|
||||||
|
|
||||||
|
def _fetch_page(self, display_id, page):
|
||||||
|
page += 1
|
||||||
|
page_data = self._download_json(
|
||||||
|
f'https://odkmedia.io/odx/api/v3/program/{display_id}/episodes/', display_id,
|
||||||
|
headers={'service-name': 'odk'}, query={
|
||||||
|
'page': page,
|
||||||
|
'page_size': self._PAGE_SIZE,
|
||||||
|
}, note=f'Downloading page {page}', expected_status=404)
|
||||||
|
for episode in traverse_obj(page_data, ('result', 'results', ...)):
|
||||||
|
yield self.url_result(
|
||||||
|
f'https://www.ondemandkorea.com/player/vod/{display_id}?contentId={episode["id"]}',
|
||||||
|
ie=OnDemandKoreaIE, video_title=episode.get('title'))
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
display_id = self._match_id(url)
|
||||||
|
|
||||||
|
entries = OnDemandPagedList(functools.partial(
|
||||||
|
self._fetch_page, display_id), self._PAGE_SIZE)
|
||||||
|
|
||||||
|
return self.playlist_result(entries, display_id)
|
||||||
|
|||||||
@@ -4,15 +4,16 @@ import re
|
|||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from ..networking import HEADRequest
|
from ..networking import HEADRequest
|
||||||
from ..utils import (
|
from ..utils import (
|
||||||
|
InAdvancePagedList,
|
||||||
clean_html,
|
clean_html,
|
||||||
determine_ext,
|
determine_ext,
|
||||||
float_or_none,
|
float_or_none,
|
||||||
InAdvancePagedList,
|
|
||||||
int_or_none,
|
int_or_none,
|
||||||
join_nonempty,
|
join_nonempty,
|
||||||
|
make_archive_id,
|
||||||
|
mimetype2ext,
|
||||||
orderedSet,
|
orderedSet,
|
||||||
remove_end,
|
remove_end,
|
||||||
make_archive_id,
|
|
||||||
smuggle_url,
|
smuggle_url,
|
||||||
strip_jsonp,
|
strip_jsonp,
|
||||||
try_call,
|
try_call,
|
||||||
@@ -21,6 +22,7 @@ from ..utils import (
|
|||||||
unsmuggle_url,
|
unsmuggle_url,
|
||||||
url_or_none,
|
url_or_none,
|
||||||
)
|
)
|
||||||
|
from ..utils.traversal import traverse_obj
|
||||||
|
|
||||||
|
|
||||||
class ORFTVthekIE(InfoExtractor):
|
class ORFTVthekIE(InfoExtractor):
|
||||||
@@ -334,6 +336,45 @@ class ORFRadioIE(InfoExtractor):
|
|||||||
self._entries(data, station or station2), show_id, data.get('title'), clean_html(data.get('subtitle')))
|
self._entries(data, station or station2), show_id, data.get('title'), clean_html(data.get('subtitle')))
|
||||||
|
|
||||||
|
|
||||||
|
class ORFPodcastIE(InfoExtractor):
|
||||||
|
IE_NAME = 'orf:podcast'
|
||||||
|
_STATION_RE = '|'.join(map(re.escape, (
|
||||||
|
'bgl', 'fm4', 'ktn', 'noe', 'oe1', 'oe3',
|
||||||
|
'ooe', 'sbg', 'stm', 'tir', 'tv', 'vbg', 'wie')))
|
||||||
|
_VALID_URL = rf'https?://sound\.orf\.at/podcast/(?P<station>{_STATION_RE})/(?P<show>[\w-]+)/(?P<id>[\w-]+)'
|
||||||
|
_TESTS = [{
|
||||||
|
'url': 'https://sound.orf.at/podcast/oe3/fruehstueck-bei-mir/nicolas-stockhammer-15102023',
|
||||||
|
'md5': '526a5700e03d271a1505386a8721ab9b',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'nicolas-stockhammer-15102023',
|
||||||
|
'ext': 'mp3',
|
||||||
|
'title': 'Nicolas Stockhammer (15.10.2023)',
|
||||||
|
'duration': 3396.0,
|
||||||
|
'series': 'Frühstück bei mir',
|
||||||
|
},
|
||||||
|
'skip': 'ORF podcasts are only available for a limited time'
|
||||||
|
}]
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
station, show, show_id = self._match_valid_url(url).group('station', 'show', 'id')
|
||||||
|
data = self._download_json(
|
||||||
|
f'https://audioapi.orf.at/radiothek/api/2.0/podcast/{station}/{show}/{show_id}', show_id)
|
||||||
|
|
||||||
|
return {
|
||||||
|
'id': show_id,
|
||||||
|
'ext': 'mp3',
|
||||||
|
'vcodec': 'none',
|
||||||
|
**traverse_obj(data, ('payload', {
|
||||||
|
'url': ('enclosures', 0, 'url'),
|
||||||
|
'ext': ('enclosures', 0, 'type', {mimetype2ext}),
|
||||||
|
'title': 'title',
|
||||||
|
'description': ('description', {clean_html}),
|
||||||
|
'duration': ('duration', {functools.partial(float_or_none, scale=1000)}),
|
||||||
|
'series': ('podcast', 'title'),
|
||||||
|
})),
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
class ORFIPTVIE(InfoExtractor):
|
class ORFIPTVIE(InfoExtractor):
|
||||||
IE_NAME = 'orf:iptv'
|
IE_NAME = 'orf:iptv'
|
||||||
IE_DESC = 'iptv.ORF.at'
|
IE_DESC = 'iptv.ORF.at'
|
||||||
|
|||||||
@@ -4,6 +4,7 @@ from ..utils import (
|
|||||||
parse_iso8601,
|
parse_iso8601,
|
||||||
unescapeHTML,
|
unescapeHTML,
|
||||||
)
|
)
|
||||||
|
from ..utils.traversal import traverse_obj
|
||||||
|
|
||||||
|
|
||||||
class PeriscopeBaseIE(InfoExtractor):
|
class PeriscopeBaseIE(InfoExtractor):
|
||||||
@@ -20,22 +21,25 @@ class PeriscopeBaseIE(InfoExtractor):
|
|||||||
title = broadcast.get('status') or 'Periscope Broadcast'
|
title = broadcast.get('status') or 'Periscope Broadcast'
|
||||||
uploader = broadcast.get('user_display_name') or broadcast.get('username')
|
uploader = broadcast.get('user_display_name') or broadcast.get('username')
|
||||||
title = '%s - %s' % (uploader, title) if uploader else title
|
title = '%s - %s' % (uploader, title) if uploader else title
|
||||||
is_live = broadcast.get('state').lower() == 'running'
|
|
||||||
|
|
||||||
thumbnails = [{
|
thumbnails = [{
|
||||||
'url': broadcast[image],
|
'url': broadcast[image],
|
||||||
} for image in ('image_url', 'image_url_small') if broadcast.get(image)]
|
} for image in ('image_url', 'image_url_medium', 'image_url_small') if broadcast.get(image)]
|
||||||
|
|
||||||
return {
|
return {
|
||||||
'id': broadcast.get('id') or video_id,
|
'id': broadcast.get('id') or video_id,
|
||||||
'title': title,
|
'title': title,
|
||||||
'timestamp': parse_iso8601(broadcast.get('created_at')),
|
'timestamp': parse_iso8601(broadcast.get('created_at')) or int_or_none(
|
||||||
|
broadcast.get('created_at_ms'), scale=1000),
|
||||||
|
'release_timestamp': int_or_none(broadcast.get('scheduled_start_ms'), scale=1000),
|
||||||
'uploader': uploader,
|
'uploader': uploader,
|
||||||
'uploader_id': broadcast.get('user_id') or broadcast.get('username'),
|
'uploader_id': broadcast.get('user_id') or broadcast.get('username'),
|
||||||
'thumbnails': thumbnails,
|
'thumbnails': thumbnails,
|
||||||
'view_count': int_or_none(broadcast.get('total_watched')),
|
'view_count': int_or_none(broadcast.get('total_watched')),
|
||||||
'tags': broadcast.get('tags'),
|
'tags': broadcast.get('tags'),
|
||||||
'is_live': is_live,
|
'live_status': {
|
||||||
|
'running': 'is_live',
|
||||||
|
'not_started': 'is_upcoming',
|
||||||
|
}.get(traverse_obj(broadcast, ('state', {str.lower}))) or 'was_live'
|
||||||
}
|
}
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
|
|||||||
@@ -262,14 +262,14 @@ class PolskieRadioAuditionIE(InfoExtractor):
|
|||||||
query=query, headers={'x-api-key': '9bf6c5a2-a7d0-4980-9ed7-a3f7291f2a81'})
|
query=query, headers={'x-api-key': '9bf6c5a2-a7d0-4980-9ed7-a3f7291f2a81'})
|
||||||
|
|
||||||
def _entries(self, playlist_id, has_episodes, has_articles):
|
def _entries(self, playlist_id, has_episodes, has_articles):
|
||||||
for i in itertools.count(1) if has_episodes else []:
|
for i in itertools.count(0) if has_episodes else []:
|
||||||
page = self._call_lp3(
|
page = self._call_lp3(
|
||||||
'AudioArticle/GetListByCategoryId', {
|
'AudioArticle/GetListByCategoryId', {
|
||||||
'categoryId': playlist_id,
|
'categoryId': playlist_id,
|
||||||
'PageSize': 10,
|
'PageSize': 10,
|
||||||
'skip': i,
|
'skip': i,
|
||||||
'format': 400,
|
'format': 400,
|
||||||
}, playlist_id, f'Downloading episode list page {i}')
|
}, playlist_id, f'Downloading episode list page {i + 1}')
|
||||||
if not traverse_obj(page, 'data'):
|
if not traverse_obj(page, 'data'):
|
||||||
break
|
break
|
||||||
for episode in page['data']:
|
for episode in page['data']:
|
||||||
@@ -281,14 +281,14 @@ class PolskieRadioAuditionIE(InfoExtractor):
|
|||||||
'timestamp': parse_iso8601(episode.get('datePublic')),
|
'timestamp': parse_iso8601(episode.get('datePublic')),
|
||||||
}
|
}
|
||||||
|
|
||||||
for i in itertools.count(1) if has_articles else []:
|
for i in itertools.count(0) if has_articles else []:
|
||||||
page = self._call_lp3(
|
page = self._call_lp3(
|
||||||
'Article/GetListByCategoryId', {
|
'Article/GetListByCategoryId', {
|
||||||
'categoryId': playlist_id,
|
'categoryId': playlist_id,
|
||||||
'PageSize': 9,
|
'PageSize': 9,
|
||||||
'skip': i,
|
'skip': i,
|
||||||
'format': 400,
|
'format': 400,
|
||||||
}, playlist_id, f'Downloading article list page {i}')
|
}, playlist_id, f'Downloading article list page {i + 1}')
|
||||||
if not traverse_obj(page, 'data'):
|
if not traverse_obj(page, 'data'):
|
||||||
break
|
break
|
||||||
for article in page['data']:
|
for article in page['data']:
|
||||||
|
|||||||
@@ -15,7 +15,7 @@ from ..utils import (
|
|||||||
|
|
||||||
class QDanceIE(InfoExtractor):
|
class QDanceIE(InfoExtractor):
|
||||||
_NETRC_MACHINE = 'qdance'
|
_NETRC_MACHINE = 'qdance'
|
||||||
_VALID_URL = r'https?://(?:www\.)?q-dance\.com/network/(?:library|live)/(?P<id>\d+)'
|
_VALID_URL = r'https?://(?:www\.)?q-dance\.com/network/(?:library|live)/(?P<id>[\w-]+)'
|
||||||
_TESTS = [{
|
_TESTS = [{
|
||||||
'note': 'vod',
|
'note': 'vod',
|
||||||
'url': 'https://www.q-dance.com/network/library/146542138',
|
'url': 'https://www.q-dance.com/network/library/146542138',
|
||||||
@@ -53,6 +53,27 @@ class QDanceIE(InfoExtractor):
|
|||||||
'channel_id': 'qdancenetwork.video_149170353',
|
'channel_id': 'qdancenetwork.video_149170353',
|
||||||
},
|
},
|
||||||
'skip': 'Completed livestream',
|
'skip': 'Completed livestream',
|
||||||
|
}, {
|
||||||
|
'note': 'vod with alphanumeric id',
|
||||||
|
'url': 'https://www.q-dance.com/network/library/WhDleSIWSfeT3Q9ObBKBeA',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'WhDleSIWSfeT3Q9ObBKBeA',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': 'Aftershock I Defqon.1 Weekend Festival 2023 I Sunday I BLUE',
|
||||||
|
'display_id': 'naam-i-defqon-1-weekend-festival-2023-i-dag-i-podium',
|
||||||
|
'description': 'Relive Defqon.1 Path of the Warrior with Aftershock at the BLUE 🔥',
|
||||||
|
'series': 'Defqon.1',
|
||||||
|
'series_id': '31840378',
|
||||||
|
'season': 'Defqon.1 Weekend Festival 2023',
|
||||||
|
'season_id': '141735599',
|
||||||
|
'duration': 3507,
|
||||||
|
'availability': 'premium_only',
|
||||||
|
'thumbnail': 'https://images.q-dance.network/1698158361-230625-135716-defqon-1-aftershock.jpg',
|
||||||
|
},
|
||||||
|
'params': {'skip_download': 'm3u8'},
|
||||||
|
}, {
|
||||||
|
'url': 'https://www.q-dance.com/network/library/-uRFKXwmRZGVnve7av9uqA',
|
||||||
|
'only_matching': True,
|
||||||
}]
|
}]
|
||||||
|
|
||||||
_access_token = None
|
_access_token = None
|
||||||
|
|||||||
150
yt_dlp/extractor/radiocomercial.py
Normal file
150
yt_dlp/extractor/radiocomercial.py
Normal file
@@ -0,0 +1,150 @@
|
|||||||
|
import itertools
|
||||||
|
|
||||||
|
from .common import InfoExtractor
|
||||||
|
from ..networking.exceptions import HTTPError
|
||||||
|
from ..utils import (
|
||||||
|
ExtractorError,
|
||||||
|
extract_attributes,
|
||||||
|
get_element_by_class,
|
||||||
|
get_element_html_by_class,
|
||||||
|
get_element_text_and_html_by_tag,
|
||||||
|
get_elements_html_by_class,
|
||||||
|
int_or_none,
|
||||||
|
join_nonempty,
|
||||||
|
try_call,
|
||||||
|
unified_strdate,
|
||||||
|
update_url,
|
||||||
|
urljoin
|
||||||
|
)
|
||||||
|
from ..utils.traversal import traverse_obj
|
||||||
|
|
||||||
|
|
||||||
|
class RadioComercialIE(InfoExtractor):
|
||||||
|
_VALID_URL = r'https?://(?:www\.)?radiocomercial\.pt/podcasts/[^/?#]+/t?(?P<season>\d+)/(?P<id>[\w-]+)'
|
||||||
|
_TESTS = [{
|
||||||
|
'url': 'https://radiocomercial.pt/podcasts/o-homem-que-mordeu-o-cao/t6/taylor-swift-entranhando-se-que-nem-uma-espada-no-ventre-dos-fas#page-content-wrapper',
|
||||||
|
'md5': '5f4fe8e485b29d2e8fd495605bc2c7e4',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'taylor-swift-entranhando-se-que-nem-uma-espada-no-ventre-dos-fas',
|
||||||
|
'ext': 'mp3',
|
||||||
|
'title': 'Taylor Swift entranhando-se que nem uma espada no ventre dos fãs.',
|
||||||
|
'release_date': '20231025',
|
||||||
|
'thumbnail': r're:https://radiocomercial.pt/upload/[^.]+.jpg',
|
||||||
|
'season': 6
|
||||||
|
}
|
||||||
|
}, {
|
||||||
|
'url': 'https://radiocomercial.pt/podcasts/convenca-me-num-minuto/t3/convenca-me-num-minuto-que-os-lobisomens-existem',
|
||||||
|
'md5': '47e96c273aef96a8eb160cd6cf46d782',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'convenca-me-num-minuto-que-os-lobisomens-existem',
|
||||||
|
'ext': 'mp3',
|
||||||
|
'title': 'Convença-me num minuto que os lobisomens existem',
|
||||||
|
'release_date': '20231026',
|
||||||
|
'thumbnail': r're:https://radiocomercial.pt/upload/[^.]+.jpg',
|
||||||
|
'season': 3
|
||||||
|
}
|
||||||
|
}, {
|
||||||
|
'url': 'https://radiocomercial.pt/podcasts/inacreditavel-by-ines-castel-branco/t2/o-desastre-de-aviao',
|
||||||
|
'md5': '69be64255420fec23b7259955d771e54',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'o-desastre-de-aviao',
|
||||||
|
'ext': 'mp3',
|
||||||
|
'title': 'O desastre de avião',
|
||||||
|
'description': 'md5:8a82beeb372641614772baab7246245f',
|
||||||
|
'release_date': '20231101',
|
||||||
|
'thumbnail': r're:https://radiocomercial.pt/upload/[^.]+.jpg',
|
||||||
|
'season': 2
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
# inconsistant md5
|
||||||
|
'skip_download': True,
|
||||||
|
},
|
||||||
|
}, {
|
||||||
|
'url': 'https://radiocomercial.pt/podcasts/tnt-todos-no-top/2023/t-n-t-29-de-outubro',
|
||||||
|
'md5': '91d32d4d4b1407272068b102730fc9fa',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 't-n-t-29-de-outubro',
|
||||||
|
'ext': 'mp3',
|
||||||
|
'title': 'T.N.T 29 de outubro',
|
||||||
|
'release_date': '20231029',
|
||||||
|
'thumbnail': r're:https://radiocomercial.pt/upload/[^.]+.jpg',
|
||||||
|
'season': 2023
|
||||||
|
}
|
||||||
|
}]
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
video_id, season = self._match_valid_url(url).group('id', 'season')
|
||||||
|
webpage = self._download_webpage(url, video_id)
|
||||||
|
return {
|
||||||
|
'id': video_id,
|
||||||
|
'title': self._html_extract_title(webpage),
|
||||||
|
'description': self._og_search_description(webpage, default=None),
|
||||||
|
'release_date': unified_strdate(get_element_by_class(
|
||||||
|
'date', get_element_html_by_class('descriptions', webpage) or '')),
|
||||||
|
'thumbnail': self._og_search_thumbnail(webpage),
|
||||||
|
'season': int_or_none(season),
|
||||||
|
'url': extract_attributes(get_element_html_by_class('audiofile', webpage) or '').get('href'),
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class RadioComercialPlaylistIE(InfoExtractor):
|
||||||
|
_VALID_URL = r'https?://(?:www\.)?radiocomercial\.pt/podcasts/(?P<id>[\w-]+)(?:/t?(?P<season>\d+))?/?(?:$|[?#])'
|
||||||
|
_TESTS = [{
|
||||||
|
'url': 'https://radiocomercial.pt/podcasts/convenca-me-num-minuto/t3',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'convenca-me-num-minuto_t3',
|
||||||
|
'title': 'Convença-me num Minuto - Temporada 3',
|
||||||
|
},
|
||||||
|
'playlist_mincount': 32
|
||||||
|
}, {
|
||||||
|
'url': 'https://radiocomercial.pt/podcasts/o-homem-que-mordeu-o-cao',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'o-homem-que-mordeu-o-cao',
|
||||||
|
'title': 'O Homem Que Mordeu o Cão',
|
||||||
|
},
|
||||||
|
'playlist_mincount': 19
|
||||||
|
}, {
|
||||||
|
'url': 'https://radiocomercial.pt/podcasts/as-minhas-coisas-favoritas',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'as-minhas-coisas-favoritas',
|
||||||
|
'title': 'As Minhas Coisas Favoritas',
|
||||||
|
},
|
||||||
|
'playlist_mincount': 131
|
||||||
|
}, {
|
||||||
|
'url': 'https://radiocomercial.pt/podcasts/tnt-todos-no-top/t2023',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'tnt-todos-no-top_t2023',
|
||||||
|
'title': 'TNT - Todos No Top - Temporada 2023',
|
||||||
|
},
|
||||||
|
'playlist_mincount': 39
|
||||||
|
}]
|
||||||
|
|
||||||
|
def _entries(self, url, playlist_id):
|
||||||
|
for page in itertools.count(1):
|
||||||
|
try:
|
||||||
|
webpage = self._download_webpage(
|
||||||
|
f'{url}/{page}', playlist_id, f'Downloading page {page}')
|
||||||
|
except ExtractorError as e:
|
||||||
|
if isinstance(e.cause, HTTPError) and e.cause.status == 404:
|
||||||
|
break
|
||||||
|
raise
|
||||||
|
|
||||||
|
episodes = get_elements_html_by_class('tm-ouvir-podcast', webpage)
|
||||||
|
if not episodes:
|
||||||
|
break
|
||||||
|
for url_path in traverse_obj(episodes, (..., {extract_attributes}, 'href')):
|
||||||
|
episode_url = urljoin(url, url_path)
|
||||||
|
if RadioComercialIE.suitable(episode_url):
|
||||||
|
yield episode_url
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
podcast, season = self._match_valid_url(url).group('id', 'season')
|
||||||
|
playlist_id = join_nonempty(podcast, season, delim='_t')
|
||||||
|
url = update_url(url, query=None, fragment=None)
|
||||||
|
webpage = self._download_webpage(url, playlist_id)
|
||||||
|
|
||||||
|
name = try_call(lambda: get_element_text_and_html_by_tag('h1', webpage)[0])
|
||||||
|
title = name if name == season else join_nonempty(name, season, delim=' - Temporada ')
|
||||||
|
|
||||||
|
return self.playlist_from_matches(
|
||||||
|
self._entries(url, playlist_id), playlist_id, title, ie=RadioComercialIE)
|
||||||
@@ -39,7 +39,7 @@ class RedTubeIE(InfoExtractor):
|
|||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
video_id = self._match_id(url)
|
video_id = self._match_id(url)
|
||||||
webpage = self._download_webpage(
|
webpage = self._download_webpage(
|
||||||
'http://www.redtube.com/%s' % video_id, video_id)
|
f'https://www.redtube.com/{video_id}', video_id)
|
||||||
|
|
||||||
ERRORS = (
|
ERRORS = (
|
||||||
(('video-deleted-info', '>This video has been removed'), 'has been removed'),
|
(('video-deleted-info', '>This video has been removed'), 'has been removed'),
|
||||||
|
|||||||
200
yt_dlp/extractor/sbscokr.py
Normal file
200
yt_dlp/extractor/sbscokr.py
Normal file
@@ -0,0 +1,200 @@
|
|||||||
|
from .common import InfoExtractor
|
||||||
|
from ..utils import (
|
||||||
|
clean_html,
|
||||||
|
int_or_none,
|
||||||
|
parse_iso8601,
|
||||||
|
parse_resolution,
|
||||||
|
url_or_none,
|
||||||
|
)
|
||||||
|
from ..utils.traversal import traverse_obj
|
||||||
|
|
||||||
|
|
||||||
|
class SBSCoKrIE(InfoExtractor):
|
||||||
|
IE_NAME = 'sbs.co.kr'
|
||||||
|
_VALID_URL = [r'https?://allvod\.sbs\.co\.kr/allvod/vod(?:Package)?EndPage\.do\?(?:[^#]+&)?mdaId=(?P<id>\d+)',
|
||||||
|
r'https?://programs\.sbs\.co\.kr/(?:enter|drama|culture|sports|plus|mtv|kth)/[a-z0-9]+/(?:vod|clip|movie)/\d+/(?P<id>(?:OC)?\d+)']
|
||||||
|
|
||||||
|
_TESTS = [{
|
||||||
|
'url': 'https://programs.sbs.co.kr/enter/dongsang2/clip/52007/OC467706746?div=main_pop_clip',
|
||||||
|
'md5': 'c3f6d45e1fb5682039d94cda23c36f19',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'OC467706746',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': '‘아슬아슬’ 박군♥한영의 새 집 인테리어 대첩♨',
|
||||||
|
'description': 'md5:6a71eb1979ee4a94ea380310068ccab4',
|
||||||
|
'thumbnail': 'https://img2.sbs.co.kr/ops_clip_img/2023/10/10/34c4c0f9-a9a5-4ff6-a92e-9bb4b5f6fa65915w1280.jpg',
|
||||||
|
'release_timestamp': 1696889400,
|
||||||
|
'release_date': '20231009',
|
||||||
|
'view_count': int,
|
||||||
|
'like_count': int,
|
||||||
|
'duration': 238,
|
||||||
|
'age_limit': 15,
|
||||||
|
'series': '동상이몽2_너는 내 운명',
|
||||||
|
'episode': '레이디제인, ‘혼전임신설’ ‘3개월’ 앞당긴 결혼식 비하인드 스토리 최초 공개!',
|
||||||
|
'episode_number': 311,
|
||||||
|
},
|
||||||
|
}, {
|
||||||
|
'url': 'https://allvod.sbs.co.kr/allvod/vodPackageEndPage.do?mdaId=22000489324&combiId=PA000000284&packageType=A&isFreeYN=',
|
||||||
|
'md5': 'bf46b2e89fda7ae7de01f5743cef7236',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '22000489324',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': '[다시보기] 트롤리 15회',
|
||||||
|
'description': 'md5:0e55d74bef1ac55c61ae90c73ac485f4',
|
||||||
|
'thumbnail': 'https://img2.sbs.co.kr/img/sbs_cms/WE/2023/02/14/arC1676333794938-1280-720.jpg',
|
||||||
|
'release_timestamp': 1676325600,
|
||||||
|
'release_date': '20230213',
|
||||||
|
'view_count': int,
|
||||||
|
'like_count': int,
|
||||||
|
'duration': 5931,
|
||||||
|
'age_limit': 15,
|
||||||
|
'series': '트롤리',
|
||||||
|
'episode': '이거 다 거짓말이야',
|
||||||
|
'episode_number': 15,
|
||||||
|
},
|
||||||
|
}, {
|
||||||
|
'url': 'https://programs.sbs.co.kr/enter/fourman/vod/69625/22000508948',
|
||||||
|
'md5': '41e8ae4cc6c8424f4e4d76661a4becbf',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '22000508948',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': '[다시보기] 신발 벗고 돌싱포맨 104회',
|
||||||
|
'description': 'md5:c6a247383c4dd661e4b956bf4d3b586e',
|
||||||
|
'thumbnail': 'https://img2.sbs.co.kr/img/sbs_cms/WE/2023/08/30/2vb1693355446261-1280-720.jpg',
|
||||||
|
'release_timestamp': 1693342800,
|
||||||
|
'release_date': '20230829',
|
||||||
|
'view_count': int,
|
||||||
|
'like_count': int,
|
||||||
|
'duration': 7036,
|
||||||
|
'age_limit': 15,
|
||||||
|
'series': '신발 벗고 돌싱포맨',
|
||||||
|
'episode': '돌싱포맨 저격수들 등장!',
|
||||||
|
'episode_number': 104,
|
||||||
|
},
|
||||||
|
}]
|
||||||
|
|
||||||
|
def _call_api(self, video_id, rscuse=''):
|
||||||
|
return self._download_json(
|
||||||
|
f'https://api.play.sbs.co.kr/1.0/sbs_vodall/{video_id}', video_id,
|
||||||
|
note=f'Downloading m3u8 information {rscuse}',
|
||||||
|
query={
|
||||||
|
'platform': 'pcweb',
|
||||||
|
'protocol': 'download',
|
||||||
|
'absolute_show': 'Y',
|
||||||
|
'service': 'program',
|
||||||
|
'ssl': 'Y',
|
||||||
|
'rscuse': rscuse,
|
||||||
|
})
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
video_id = self._match_id(url)
|
||||||
|
|
||||||
|
details = self._call_api(video_id)
|
||||||
|
source = traverse_obj(details, ('vod', 'source', 'mediasource', {dict})) or {}
|
||||||
|
|
||||||
|
formats = []
|
||||||
|
for stream in traverse_obj(details, (
|
||||||
|
'vod', 'source', 'mediasourcelist', lambda _, v: v['mediaurl'] or v['mediarscuse']
|
||||||
|
), default=[source]):
|
||||||
|
if not stream.get('mediaurl'):
|
||||||
|
new_source = traverse_obj(
|
||||||
|
self._call_api(video_id, rscuse=stream['mediarscuse']),
|
||||||
|
('vod', 'source', 'mediasource', {dict})) or {}
|
||||||
|
if new_source.get('mediarscuse') == source.get('mediarscuse') or not new_source.get('mediaurl'):
|
||||||
|
continue
|
||||||
|
stream = new_source
|
||||||
|
formats.append({
|
||||||
|
'url': stream['mediaurl'],
|
||||||
|
'format_id': stream.get('mediarscuse'),
|
||||||
|
'format_note': stream.get('medianame'),
|
||||||
|
**parse_resolution(stream.get('quality')),
|
||||||
|
'preference': int_or_none(stream.get('mediarscuse'))
|
||||||
|
})
|
||||||
|
|
||||||
|
caption_url = traverse_obj(details, ('vod', 'source', 'subtitle', {url_or_none}))
|
||||||
|
|
||||||
|
return {
|
||||||
|
'id': video_id,
|
||||||
|
**traverse_obj(details, ('vod', {
|
||||||
|
'title': ('info', 'title'),
|
||||||
|
'duration': ('info', 'duration', {int_or_none}),
|
||||||
|
'view_count': ('info', 'viewcount', {int_or_none}),
|
||||||
|
'like_count': ('info', 'likecount', {int_or_none}),
|
||||||
|
'description': ('info', 'synopsis', {clean_html}),
|
||||||
|
'episode': ('info', 'content', ('contenttitle', 'title')),
|
||||||
|
'episode_number': ('info', 'content', 'number', {int_or_none}),
|
||||||
|
'series': ('info', 'program', 'programtitle'),
|
||||||
|
'age_limit': ('info', 'targetage', {int_or_none}),
|
||||||
|
'release_timestamp': ('info', 'broaddate', {parse_iso8601}),
|
||||||
|
'thumbnail': ('source', 'thumbnail', 'origin', {url_or_none}),
|
||||||
|
}), get_all=False),
|
||||||
|
'formats': formats,
|
||||||
|
'subtitles': {'ko': [{'url': caption_url}]} if caption_url else None,
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class SBSCoKrAllvodProgramIE(InfoExtractor):
|
||||||
|
IE_NAME = 'sbs.co.kr:allvod_program'
|
||||||
|
_VALID_URL = r'https?://allvod\.sbs\.co\.kr/allvod/vod(?:Free)?ProgramDetail\.do\?(?:[^#]+&)?pgmId=(?P<id>P?\d+)'
|
||||||
|
|
||||||
|
_TESTS = [{
|
||||||
|
'url': 'https://allvod.sbs.co.kr/allvod/vodFreeProgramDetail.do?type=legend&pgmId=22000010159&listOrder=vodCntAsc',
|
||||||
|
'info_dict': {
|
||||||
|
'_type': 'playlist',
|
||||||
|
'id': '22000010159',
|
||||||
|
},
|
||||||
|
'playlist_count': 18,
|
||||||
|
}, {
|
||||||
|
'url': 'https://allvod.sbs.co.kr/allvod/vodProgramDetail.do?pgmId=P460810577',
|
||||||
|
'info_dict': {
|
||||||
|
'_type': 'playlist',
|
||||||
|
'id': 'P460810577',
|
||||||
|
},
|
||||||
|
'playlist_count': 13,
|
||||||
|
}]
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
program_id = self._match_id(url)
|
||||||
|
|
||||||
|
details = self._download_json(
|
||||||
|
'https://allvod.sbs.co.kr/allvod/vodProgramDetail/vodProgramDetailAjax.do',
|
||||||
|
program_id, note='Downloading program details',
|
||||||
|
query={
|
||||||
|
'pgmId': program_id,
|
||||||
|
'currentCount': '10000',
|
||||||
|
})
|
||||||
|
|
||||||
|
return self.playlist_result(
|
||||||
|
[self.url_result(f'https://allvod.sbs.co.kr/allvod/vodEndPage.do?mdaId={video_id}', SBSCoKrIE)
|
||||||
|
for video_id in traverse_obj(details, ('list', ..., 'mdaId'))], program_id)
|
||||||
|
|
||||||
|
|
||||||
|
class SBSCoKrProgramsVodIE(InfoExtractor):
|
||||||
|
IE_NAME = 'sbs.co.kr:programs_vod'
|
||||||
|
_VALID_URL = r'https?://programs\.sbs\.co\.kr/(?:enter|drama|culture|sports|plus|mtv)/(?P<id>[a-z0-9]+)/vods'
|
||||||
|
|
||||||
|
_TESTS = [{
|
||||||
|
'url': 'https://programs.sbs.co.kr/culture/morningwide/vods/65007',
|
||||||
|
'info_dict': {
|
||||||
|
'_type': 'playlist',
|
||||||
|
'id': '00000210215',
|
||||||
|
},
|
||||||
|
'playlist_mincount': 9782,
|
||||||
|
}, {
|
||||||
|
'url': 'https://programs.sbs.co.kr/enter/dongsang2/vods/52006',
|
||||||
|
'info_dict': {
|
||||||
|
'_type': 'playlist',
|
||||||
|
'id': '22000010476',
|
||||||
|
},
|
||||||
|
'playlist_mincount': 312,
|
||||||
|
}]
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
program_slug = self._match_id(url)
|
||||||
|
|
||||||
|
program_id = self._download_json(
|
||||||
|
f'https://static.apis.sbs.co.kr/program-api/1.0/menu/{program_slug}', program_slug,
|
||||||
|
note='Downloading program menu data')['program']['programid']
|
||||||
|
|
||||||
|
return self.url_result(
|
||||||
|
f'https://allvod.sbs.co.kr/allvod/vodProgramDetail.do?pgmId={program_id}', SBSCoKrAllvodProgramIE)
|
||||||
@@ -1,5 +1,6 @@
|
|||||||
import re
|
import re
|
||||||
import urllib.parse
|
import urllib.parse
|
||||||
|
import xml.etree.ElementTree
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from ..utils import (
|
from ..utils import (
|
||||||
@@ -469,11 +470,12 @@ class SlidesLiveIE(InfoExtractor):
|
|||||||
slides = self._download_xml(
|
slides = self._download_xml(
|
||||||
player_info['slides_xml_url'], video_id, fatal=False,
|
player_info['slides_xml_url'], video_id, fatal=False,
|
||||||
note='Downloading slides XML', errnote='Failed to download slides info')
|
note='Downloading slides XML', errnote='Failed to download slides info')
|
||||||
slide_url_template = 'https://cdn.slideslive.com/data/presentations/%s/slides/big/%s%s'
|
if isinstance(slides, xml.etree.ElementTree.Element):
|
||||||
for slide_id, slide in enumerate(slides.findall('./slide') if slides else [], 1):
|
slide_url_template = 'https://cdn.slideslive.com/data/presentations/%s/slides/big/%s%s'
|
||||||
slides_info.append((
|
for slide_id, slide in enumerate(slides.findall('./slide')):
|
||||||
slide_id, xpath_text(slide, './slideName', 'name'), '.jpg',
|
slides_info.append((
|
||||||
int_or_none(xpath_text(slide, './timeSec', 'time'))))
|
slide_id, xpath_text(slide, './slideName', 'name'), '.jpg',
|
||||||
|
int_or_none(xpath_text(slide, './timeSec', 'time'))))
|
||||||
|
|
||||||
chapters, thumbnails = [], []
|
chapters, thumbnails = [], []
|
||||||
if url_or_none(player_info.get('thumbnail')):
|
if url_or_none(player_info.get('thumbnail')):
|
||||||
@@ -528,7 +530,7 @@ class SlidesLiveIE(InfoExtractor):
|
|||||||
if service_name == 'vimeo':
|
if service_name == 'vimeo':
|
||||||
info['url'] = smuggle_url(
|
info['url'] = smuggle_url(
|
||||||
f'https://player.vimeo.com/video/{service_id}',
|
f'https://player.vimeo.com/video/{service_id}',
|
||||||
{'http_headers': {'Referer': url}})
|
{'referer': url})
|
||||||
|
|
||||||
video_slides = traverse_obj(slides, ('slides', ..., 'video', 'id'))
|
video_slides = traverse_obj(slides, ('slides', ..., 'video', 'id'))
|
||||||
if not video_slides:
|
if not video_slides:
|
||||||
|
|||||||
@@ -38,9 +38,48 @@ class StacommuBaseIE(WrestleUniverseBaseIE):
|
|||||||
return None
|
return None
|
||||||
return traverse_obj(encryption_data, {'key': ('key', {decrypt}), 'iv': ('iv', {decrypt})})
|
return traverse_obj(encryption_data, {'key': ('key', {decrypt}), 'iv': ('iv', {decrypt})})
|
||||||
|
|
||||||
|
def _extract_vod(self, url):
|
||||||
|
video_id = self._match_id(url)
|
||||||
|
video_info = self._download_metadata(
|
||||||
|
url, video_id, 'ja', ('dehydratedState', 'queries', 0, 'state', 'data'))
|
||||||
|
hls_info, decrypt = self._call_encrypted_api(
|
||||||
|
video_id, ':watch', 'stream information', data={'method': 1})
|
||||||
|
|
||||||
|
return {
|
||||||
|
'id': video_id,
|
||||||
|
'formats': self._get_formats(hls_info, ('protocolHls', 'url', {url_or_none}), video_id),
|
||||||
|
'hls_aes': self._extract_hls_key(hls_info, 'protocolHls', decrypt),
|
||||||
|
**traverse_obj(video_info, {
|
||||||
|
'title': ('displayName', {str}),
|
||||||
|
'description': ('description', {str}),
|
||||||
|
'timestamp': ('watchStartTime', {int_or_none}),
|
||||||
|
'thumbnail': ('keyVisualUrl', {url_or_none}),
|
||||||
|
'cast': ('casts', ..., 'displayName', {str}),
|
||||||
|
'duration': ('duration', {int}),
|
||||||
|
}),
|
||||||
|
}
|
||||||
|
|
||||||
|
def _extract_ppv(self, url):
|
||||||
|
video_id = self._match_id(url)
|
||||||
|
video_info = self._call_api(video_id, msg='video information', query={'al': 'ja'}, auth=False)
|
||||||
|
hls_info, decrypt = self._call_encrypted_api(
|
||||||
|
video_id, ':watchArchive', 'stream information', data={'method': 1})
|
||||||
|
|
||||||
|
return {
|
||||||
|
'id': video_id,
|
||||||
|
'formats': self._get_formats(hls_info, ('hls', 'urls', ..., {url_or_none}), video_id),
|
||||||
|
'hls_aes': self._extract_hls_key(hls_info, 'hls', decrypt),
|
||||||
|
**traverse_obj(video_info, {
|
||||||
|
'title': ('displayName', {str}),
|
||||||
|
'timestamp': ('startTime', {int_or_none}),
|
||||||
|
'thumbnail': ('keyVisualUrl', {url_or_none}),
|
||||||
|
'duration': ('duration', {int_or_none}),
|
||||||
|
}),
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
class StacommuVODIE(StacommuBaseIE):
|
class StacommuVODIE(StacommuBaseIE):
|
||||||
_VALID_URL = r'https?://www\.stacommu\.jp/videos/episodes/(?P<id>[\da-zA-Z]+)'
|
_VALID_URL = r'https?://www\.stacommu\.jp/(?:en/)?videos/episodes/(?P<id>[\da-zA-Z]+)'
|
||||||
_TESTS = [{
|
_TESTS = [{
|
||||||
# not encrypted
|
# not encrypted
|
||||||
'url': 'https://www.stacommu.jp/videos/episodes/aXcVKjHyAENEjard61soZZ',
|
'url': 'https://www.stacommu.jp/videos/episodes/aXcVKjHyAENEjard61soZZ',
|
||||||
@@ -79,34 +118,19 @@ class StacommuVODIE(StacommuBaseIE):
|
|||||||
'params': {
|
'params': {
|
||||||
'skip_download': 'm3u8',
|
'skip_download': 'm3u8',
|
||||||
},
|
},
|
||||||
|
}, {
|
||||||
|
'url': 'https://www.stacommu.jp/en/videos/episodes/aXcVKjHyAENEjard61soZZ',
|
||||||
|
'only_matching': True,
|
||||||
}]
|
}]
|
||||||
|
|
||||||
_API_PATH = 'videoEpisodes'
|
_API_PATH = 'videoEpisodes'
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
video_id = self._match_id(url)
|
return self._extract_vod(url)
|
||||||
video_info = self._download_metadata(
|
|
||||||
url, video_id, 'ja', ('dehydratedState', 'queries', 0, 'state', 'data'))
|
|
||||||
hls_info, decrypt = self._call_encrypted_api(
|
|
||||||
video_id, ':watch', 'stream information', data={'method': 1})
|
|
||||||
|
|
||||||
return {
|
|
||||||
'id': video_id,
|
|
||||||
'formats': self._get_formats(hls_info, ('protocolHls', 'url', {url_or_none}), video_id),
|
|
||||||
'hls_aes': self._extract_hls_key(hls_info, 'protocolHls', decrypt),
|
|
||||||
**traverse_obj(video_info, {
|
|
||||||
'title': ('displayName', {str}),
|
|
||||||
'description': ('description', {str}),
|
|
||||||
'timestamp': ('watchStartTime', {int_or_none}),
|
|
||||||
'thumbnail': ('keyVisualUrl', {url_or_none}),
|
|
||||||
'cast': ('casts', ..., 'displayName', {str}),
|
|
||||||
'duration': ('duration', {int}),
|
|
||||||
}),
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
class StacommuLiveIE(StacommuBaseIE):
|
class StacommuLiveIE(StacommuBaseIE):
|
||||||
_VALID_URL = r'https?://www\.stacommu\.jp/live/(?P<id>[\da-zA-Z]+)'
|
_VALID_URL = r'https?://www\.stacommu\.jp/(?:en/)?live/(?P<id>[\da-zA-Z]+)'
|
||||||
_TESTS = [{
|
_TESTS = [{
|
||||||
'url': 'https://www.stacommu.jp/live/d2FJ3zLnndegZJCAEzGM3m',
|
'url': 'https://www.stacommu.jp/live/d2FJ3zLnndegZJCAEzGM3m',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
@@ -125,24 +149,83 @@ class StacommuLiveIE(StacommuBaseIE):
|
|||||||
'params': {
|
'params': {
|
||||||
'skip_download': 'm3u8',
|
'skip_download': 'm3u8',
|
||||||
},
|
},
|
||||||
|
}, {
|
||||||
|
'url': 'https://www.stacommu.jp/en/live/d2FJ3zLnndegZJCAEzGM3m',
|
||||||
|
'only_matching': True,
|
||||||
}]
|
}]
|
||||||
|
|
||||||
_API_PATH = 'events'
|
_API_PATH = 'events'
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
video_id = self._match_id(url)
|
return self._extract_ppv(url)
|
||||||
video_info = self._call_api(video_id, msg='video information', query={'al': 'ja'}, auth=False)
|
|
||||||
hls_info, decrypt = self._call_encrypted_api(
|
|
||||||
video_id, ':watchArchive', 'stream information', data={'method': 1})
|
|
||||||
|
|
||||||
return {
|
|
||||||
'id': video_id,
|
class TheaterComplexTownBaseIE(StacommuBaseIE):
|
||||||
'formats': self._get_formats(hls_info, ('hls', 'urls', ..., {url_or_none}), video_id),
|
_NETRC_MACHINE = 'theatercomplextown'
|
||||||
'hls_aes': self._extract_hls_key(hls_info, 'hls', decrypt),
|
_API_HOST = 'api.theater-complex.town'
|
||||||
**traverse_obj(video_info, {
|
_LOGIN_QUERY = {'key': 'AIzaSyAgNCqToaIz4a062EeIrkhI_xetVfAOrfc'}
|
||||||
'title': ('displayName', {str}),
|
_LOGIN_HEADERS = {
|
||||||
'timestamp': ('startTime', {int_or_none}),
|
'Accept': '*/*',
|
||||||
'thumbnail': ('keyVisualUrl', {url_or_none}),
|
'Content-Type': 'application/json',
|
||||||
'duration': ('duration', {int_or_none}),
|
'X-Client-Version': 'Chrome/JsCore/9.23.0/FirebaseCore-web',
|
||||||
}),
|
'Referer': 'https://www.theater-complex.town/',
|
||||||
}
|
'Origin': 'https://www.theater-complex.town',
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class TheaterComplexTownVODIE(TheaterComplexTownBaseIE):
|
||||||
|
_VALID_URL = r'https?://(?:www\.)?theater-complex\.town/(?:en/)?videos/episodes/(?P<id>\w+)'
|
||||||
|
IE_NAME = 'theatercomplextown:vod'
|
||||||
|
_TESTS = [{
|
||||||
|
'url': 'https://www.theater-complex.town/videos/episodes/hoxqidYNoAn7bP92DN6p78',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'hoxqidYNoAn7bP92DN6p78',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': '演劇ドラフトグランプリ2023 劇団『恋のぼり』〜劇団名決定秘話ラジオ',
|
||||||
|
'description': 'md5:a7e2e9cf570379ea67fb630f345ff65d',
|
||||||
|
'cast': ['玉城 裕規', '石川 凌雅'],
|
||||||
|
'thumbnail': 'https://image.theater-complex.town/5URnXX6KCeDysuFrPkP38o/5URnXX6KCeDysuFrPkP38o',
|
||||||
|
'upload_date': '20231103',
|
||||||
|
'timestamp': 1699016400,
|
||||||
|
'duration': 868,
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
'skip_download': 'm3u8',
|
||||||
|
},
|
||||||
|
}, {
|
||||||
|
'url': 'https://www.theater-complex.town/en/videos/episodes/6QT7XYwM9dJz5Gf9VB6K5y',
|
||||||
|
'only_matching': True,
|
||||||
|
}]
|
||||||
|
|
||||||
|
_API_PATH = 'videoEpisodes'
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
return self._extract_vod(url)
|
||||||
|
|
||||||
|
|
||||||
|
class TheaterComplexTownPPVIE(TheaterComplexTownBaseIE):
|
||||||
|
_VALID_URL = r'https?://(?:www\.)?theater-complex\.town/(?:en/)?ppv/(?P<id>\w+)'
|
||||||
|
IE_NAME = 'theatercomplextown:ppv'
|
||||||
|
_TESTS = [{
|
||||||
|
'url': 'https://www.theater-complex.town/ppv/wytW3X7khrjJBUpKuV3jen',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'wytW3X7khrjJBUpKuV3jen',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': 'BREAK FREE STARS 11月5日(日)12:30千秋楽公演',
|
||||||
|
'thumbnail': 'https://image.theater-complex.town/5GWEB31JcTUfjtgdeV5t6o/5GWEB31JcTUfjtgdeV5t6o',
|
||||||
|
'upload_date': '20231105',
|
||||||
|
'timestamp': 1699155000,
|
||||||
|
'duration': 8378,
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
'skip_download': 'm3u8',
|
||||||
|
},
|
||||||
|
}, {
|
||||||
|
'url': 'https://www.theater-complex.town/en/ppv/wytW3X7khrjJBUpKuV3jen',
|
||||||
|
'only_matching': True,
|
||||||
|
}]
|
||||||
|
|
||||||
|
_API_PATH = 'events'
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
return self._extract_ppv(url)
|
||||||
|
|||||||
@@ -32,9 +32,7 @@ class StoryFireBaseIE(InfoExtractor):
|
|||||||
'description': video.get('description'),
|
'description': video.get('description'),
|
||||||
'url': smuggle_url(
|
'url': smuggle_url(
|
||||||
'https://player.vimeo.com/video/' + vimeo_id, {
|
'https://player.vimeo.com/video/' + vimeo_id, {
|
||||||
'http_headers': {
|
'referer': 'https://storyfire.com/',
|
||||||
'Referer': 'https://storyfire.com/',
|
|
||||||
}
|
|
||||||
}),
|
}),
|
||||||
'thumbnail': video.get('storyImage'),
|
'thumbnail': video.get('storyImage'),
|
||||||
'view_count': int_or_none(video.get('views')),
|
'view_count': int_or_none(video.get('views')),
|
||||||
|
|||||||
@@ -1,66 +0,0 @@
|
|||||||
from .common import InfoExtractor
|
|
||||||
from ..utils import remove_end
|
|
||||||
|
|
||||||
|
|
||||||
class ThisAVIE(InfoExtractor):
|
|
||||||
_VALID_URL = r'https?://(?:www\.)?thisav\.com/video/(?P<id>[0-9]+)/.*'
|
|
||||||
_TESTS = [{
|
|
||||||
# jwplayer
|
|
||||||
'url': 'http://www.thisav.com/video/47734/%98%26sup1%3B%83%9E%83%82---just-fit.html',
|
|
||||||
'md5': '0480f1ef3932d901f0e0e719f188f19b',
|
|
||||||
'info_dict': {
|
|
||||||
'id': '47734',
|
|
||||||
'ext': 'flv',
|
|
||||||
'title': '高樹マリア - Just fit',
|
|
||||||
'uploader': 'dj7970',
|
|
||||||
'uploader_id': 'dj7970'
|
|
||||||
}
|
|
||||||
}, {
|
|
||||||
# html5 media
|
|
||||||
'url': 'http://www.thisav.com/video/242352/nerdy-18yo-big-ass-tattoos-and-glasses.html',
|
|
||||||
'md5': 'ba90c076bd0f80203679e5b60bf523ee',
|
|
||||||
'info_dict': {
|
|
||||||
'id': '242352',
|
|
||||||
'ext': 'mp4',
|
|
||||||
'title': 'Nerdy 18yo Big Ass Tattoos and Glasses',
|
|
||||||
'uploader': 'cybersluts',
|
|
||||||
'uploader_id': 'cybersluts',
|
|
||||||
},
|
|
||||||
}]
|
|
||||||
|
|
||||||
def _real_extract(self, url):
|
|
||||||
mobj = self._match_valid_url(url)
|
|
||||||
|
|
||||||
video_id = mobj.group('id')
|
|
||||||
webpage = self._download_webpage(url, video_id)
|
|
||||||
title = remove_end(self._html_extract_title(webpage), ' - 視頻 - ThisAV.com-世界第一中文成人娛樂網站')
|
|
||||||
video_url = self._html_search_regex(
|
|
||||||
r"addVariable\('file','([^']+)'\);", webpage, 'video url', default=None)
|
|
||||||
if video_url:
|
|
||||||
info_dict = {
|
|
||||||
'formats': [{
|
|
||||||
'url': video_url,
|
|
||||||
}],
|
|
||||||
}
|
|
||||||
else:
|
|
||||||
entries = self._parse_html5_media_entries(url, webpage, video_id)
|
|
||||||
if entries:
|
|
||||||
info_dict = entries[0]
|
|
||||||
else:
|
|
||||||
info_dict = self._extract_jwplayer_data(
|
|
||||||
webpage, video_id, require_title=False)
|
|
||||||
uploader = self._html_search_regex(
|
|
||||||
r': <a href="http://www\.thisav\.com/user/[0-9]+/(?:[^"]+)">([^<]+)</a>',
|
|
||||||
webpage, 'uploader name', fatal=False)
|
|
||||||
uploader_id = self._html_search_regex(
|
|
||||||
r': <a href="http://www\.thisav\.com/user/[0-9]+/([^"]+)">(?:[^<]+)</a>',
|
|
||||||
webpage, 'uploader id', fatal=False)
|
|
||||||
|
|
||||||
info_dict.update({
|
|
||||||
'id': video_id,
|
|
||||||
'uploader': uploader,
|
|
||||||
'uploader_id': uploader_id,
|
|
||||||
'title': title,
|
|
||||||
})
|
|
||||||
|
|
||||||
return info_dict
|
|
||||||
@@ -1,11 +1,23 @@
|
|||||||
|
import json
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
|
from .zype import ZypeIE
|
||||||
from ..networking import HEADRequest
|
from ..networking import HEADRequest
|
||||||
|
from ..networking.exceptions import HTTPError
|
||||||
|
from ..utils import (
|
||||||
|
ExtractorError,
|
||||||
|
filter_dict,
|
||||||
|
parse_qs,
|
||||||
|
try_call,
|
||||||
|
urlencode_postdata,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
class ThisOldHouseIE(InfoExtractor):
|
class ThisOldHouseIE(InfoExtractor):
|
||||||
_VALID_URL = r'https?://(?:www\.)?thisoldhouse\.com/(?:watch|how-to|tv-episode|(?:[^/]+/)?\d+)/(?P<id>[^/?#]+)'
|
_NETRC_MACHINE = 'thisoldhouse'
|
||||||
|
_VALID_URL = r'https?://(?:www\.)?thisoldhouse\.com/(?:watch|how-to|tv-episode|(?:[^/?#]+/)?\d+)/(?P<id>[^/?#]+)'
|
||||||
_TESTS = [{
|
_TESTS = [{
|
||||||
'url': 'https://www.thisoldhouse.com/how-to/how-to-build-storage-bench',
|
'url': 'https://www.thisoldhouse.com/furniture/21017078/how-to-build-a-storage-bench',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': '5dcdddf673c3f956ef5db202',
|
'id': '5dcdddf673c3f956ef5db202',
|
||||||
'ext': 'mp4',
|
'ext': 'mp4',
|
||||||
@@ -23,13 +35,16 @@ class ThisOldHouseIE(InfoExtractor):
|
|||||||
'skip_download': True,
|
'skip_download': True,
|
||||||
},
|
},
|
||||||
}, {
|
}, {
|
||||||
|
# Page no longer has video
|
||||||
'url': 'https://www.thisoldhouse.com/watch/arlington-arts-crafts-arts-and-crafts-class-begins',
|
'url': 'https://www.thisoldhouse.com/watch/arlington-arts-crafts-arts-and-crafts-class-begins',
|
||||||
'only_matching': True,
|
'only_matching': True,
|
||||||
}, {
|
}, {
|
||||||
|
# 404 Not Found
|
||||||
'url': 'https://www.thisoldhouse.com/tv-episode/ask-toh-shelf-rough-electric',
|
'url': 'https://www.thisoldhouse.com/tv-episode/ask-toh-shelf-rough-electric',
|
||||||
'only_matching': True,
|
'only_matching': True,
|
||||||
}, {
|
}, {
|
||||||
'url': 'https://www.thisoldhouse.com/furniture/21017078/how-to-build-a-storage-bench',
|
# 404 Not Found
|
||||||
|
'url': 'https://www.thisoldhouse.com/how-to/how-to-build-storage-bench',
|
||||||
'only_matching': True,
|
'only_matching': True,
|
||||||
}, {
|
}, {
|
||||||
'url': 'https://www.thisoldhouse.com/21113884/s41-e13-paradise-lost',
|
'url': 'https://www.thisoldhouse.com/21113884/s41-e13-paradise-lost',
|
||||||
@@ -39,17 +54,51 @@ class ThisOldHouseIE(InfoExtractor):
|
|||||||
'url': 'https://www.thisoldhouse.com/21083431/seaside-transformation-the-westerly-project',
|
'url': 'https://www.thisoldhouse.com/21083431/seaside-transformation-the-westerly-project',
|
||||||
'only_matching': True,
|
'only_matching': True,
|
||||||
}]
|
}]
|
||||||
_ZYPE_TMPL = 'https://player.zype.com/embed/%s.html?api_key=hsOk_yMSPYNrT22e9pu8hihLXjaZf0JW5jsOWv4ZqyHJFvkJn6rtToHl09tbbsbe'
|
|
||||||
|
_LOGIN_URL = 'https://login.thisoldhouse.com/usernamepassword/login'
|
||||||
|
|
||||||
|
def _perform_login(self, username, password):
|
||||||
|
self._request_webpage(
|
||||||
|
HEADRequest('https://www.thisoldhouse.com/insider'), None, 'Requesting session cookies')
|
||||||
|
urlh = self._request_webpage(
|
||||||
|
'https://www.thisoldhouse.com/wp-login.php', None, 'Requesting login info',
|
||||||
|
errnote='Unable to login', query={'redirect_to': 'https://www.thisoldhouse.com/insider'})
|
||||||
|
|
||||||
|
try:
|
||||||
|
auth_form = self._download_webpage(
|
||||||
|
self._LOGIN_URL, None, 'Submitting credentials', headers={
|
||||||
|
'Content-Type': 'application/json',
|
||||||
|
'Referer': urlh.url,
|
||||||
|
}, data=json.dumps(filter_dict({
|
||||||
|
**{('client_id' if k == 'client' else k): v[0] for k, v in parse_qs(urlh.url).items()},
|
||||||
|
'tenant': 'thisoldhouse',
|
||||||
|
'username': username,
|
||||||
|
'password': password,
|
||||||
|
'popup_options': {},
|
||||||
|
'sso': True,
|
||||||
|
'_csrf': try_call(lambda: self._get_cookies(self._LOGIN_URL)['_csrf'].value),
|
||||||
|
'_intstate': 'deprecated',
|
||||||
|
}), separators=(',', ':')).encode())
|
||||||
|
except ExtractorError as e:
|
||||||
|
if isinstance(e.cause, HTTPError) and e.cause.status == 401:
|
||||||
|
raise ExtractorError('Invalid username or password', expected=True)
|
||||||
|
raise
|
||||||
|
|
||||||
|
self._request_webpage(
|
||||||
|
'https://login.thisoldhouse.com/login/callback', None, 'Completing login',
|
||||||
|
data=urlencode_postdata(self._hidden_inputs(auth_form)))
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
display_id = self._match_id(url)
|
display_id = self._match_id(url)
|
||||||
webpage = self._download_webpage(url, display_id)
|
webpage = self._download_webpage(url, display_id)
|
||||||
if 'To Unlock This content' in webpage:
|
if 'To Unlock This content' in webpage:
|
||||||
self.raise_login_required(method='cookies')
|
self.raise_login_required(
|
||||||
video_url = self._search_regex(
|
'This video is only available for subscribers. '
|
||||||
|
'Note that --cookies-from-browser may not work due to this site using session cookies')
|
||||||
|
|
||||||
|
video_url, video_id = self._search_regex(
|
||||||
r'<iframe[^>]+src=[\'"]((?:https?:)?//(?:www\.)?thisoldhouse\.(?:chorus\.build|com)/videos/zype/([0-9a-f]{24})[^\'"]*)[\'"]',
|
r'<iframe[^>]+src=[\'"]((?:https?:)?//(?:www\.)?thisoldhouse\.(?:chorus\.build|com)/videos/zype/([0-9a-f]{24})[^\'"]*)[\'"]',
|
||||||
webpage, 'video url')
|
webpage, 'video url', group=(1, 2))
|
||||||
if 'subscription_required=true' in video_url or 'c-entry-group-labels__image' in webpage:
|
video_url = self._request_webpage(HEADRequest(video_url), video_id, 'Resolving Zype URL').url
|
||||||
return self.url_result(self._request_webpage(HEADRequest(video_url), display_id).url, 'Zype', display_id)
|
|
||||||
video_id = self._search_regex(r'(?:https?:)?//(?:www\.)?thisoldhouse\.(?:chorus\.build|com)/videos/zype/([0-9a-f]{24})', video_url, 'video id')
|
return self.url_result(video_url, ZypeIE, video_id)
|
||||||
return self.url_result(self._ZYPE_TMPL % video_id, 'Zype', video_id)
|
|
||||||
|
|||||||
@@ -84,6 +84,13 @@ class TV5MondePlusIE(InfoExtractor):
|
|||||||
}]
|
}]
|
||||||
_GEO_BYPASS = False
|
_GEO_BYPASS = False
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def _extract_subtitles(data_captions):
|
||||||
|
subtitles = {}
|
||||||
|
for f in traverse_obj(data_captions, ('files', lambda _, v: url_or_none(v['file']))):
|
||||||
|
subtitles.setdefault(f.get('label') or 'fra', []).append({'url': f['file']})
|
||||||
|
return subtitles
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
display_id = self._match_id(url)
|
display_id = self._match_id(url)
|
||||||
webpage = self._download_webpage(url, display_id)
|
webpage = self._download_webpage(url, display_id)
|
||||||
@@ -176,6 +183,8 @@ class TV5MondePlusIE(InfoExtractor):
|
|||||||
'duration': duration,
|
'duration': duration,
|
||||||
'upload_date': upload_date,
|
'upload_date': upload_date,
|
||||||
'formats': formats,
|
'formats': formats,
|
||||||
|
'subtitles': self._extract_subtitles(self._parse_json(
|
||||||
|
traverse_obj(vpl_data, ('data-captions', {str}), default='{}'), display_id, fatal=False)),
|
||||||
'series': series,
|
'series': series,
|
||||||
'episode': episode,
|
'episode': episode,
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -11,6 +11,7 @@ from ..utils import (
|
|||||||
float_or_none,
|
float_or_none,
|
||||||
get_element_by_class,
|
get_element_by_class,
|
||||||
get_element_by_id,
|
get_element_by_id,
|
||||||
|
int_or_none,
|
||||||
parse_duration,
|
parse_duration,
|
||||||
qualities,
|
qualities,
|
||||||
str_to_int,
|
str_to_int,
|
||||||
@@ -142,7 +143,7 @@ class TwitCastingIE(InfoExtractor):
|
|||||||
'https://twitcasting.tv/streamserver.php?target=%s&mode=client' % uploader_id, video_id,
|
'https://twitcasting.tv/streamserver.php?target=%s&mode=client' % uploader_id, video_id,
|
||||||
'Downloading live info', fatal=False)
|
'Downloading live info', fatal=False)
|
||||||
|
|
||||||
is_live = 'data-status="online"' in webpage
|
is_live = any(f'data-{x}' in webpage for x in ['is-onlive="true"', 'live-type="live"', 'status="online"'])
|
||||||
if not traverse_obj(stream_server_data, 'llfmp4') and is_live:
|
if not traverse_obj(stream_server_data, 'llfmp4') and is_live:
|
||||||
self.raise_login_required(method='cookies')
|
self.raise_login_required(method='cookies')
|
||||||
|
|
||||||
@@ -241,6 +242,8 @@ class TwitCastingLiveIE(InfoExtractor):
|
|||||||
'expected_exception': 'UserNotLive',
|
'expected_exception': 'UserNotLive',
|
||||||
}]
|
}]
|
||||||
|
|
||||||
|
_PROTECTED_LIVE_RE = r'(?s)(<span\s*class="tw-movie-thumbnail2-badge"\s*data-status="live">\s*LIVE)'
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
uploader_id = self._match_id(url)
|
uploader_id = self._match_id(url)
|
||||||
self.to_screen(
|
self.to_screen(
|
||||||
@@ -248,24 +251,27 @@ class TwitCastingLiveIE(InfoExtractor):
|
|||||||
'Pass "https://twitcasting.tv/{0}/show" to download the history'.format(uploader_id))
|
'Pass "https://twitcasting.tv/{0}/show" to download the history'.format(uploader_id))
|
||||||
|
|
||||||
webpage = self._download_webpage(url, uploader_id)
|
webpage = self._download_webpage(url, uploader_id)
|
||||||
current_live = self._search_regex(
|
is_live = self._search_regex( # first pattern is for public live
|
||||||
(r'data-type="movie" data-id="(\d+)">',
|
(r'(data-is-onlive="true")', self._PROTECTED_LIVE_RE), webpage, 'is live?', default=None)
|
||||||
r'tw-sound-flag-open-link" data-id="(\d+)" style=',),
|
current_live = int_or_none(self._search_regex(
|
||||||
webpage, 'current live ID', default=None)
|
(r'data-type="movie" data-id="(\d+)">', # not available?
|
||||||
if not current_live:
|
r'tw-sound-flag-open-link" data-id="(\d+)" style=', # not available?
|
||||||
|
r'data-movie-id="(\d+)"'), # if not currently live, value may be 0
|
||||||
|
webpage, 'current live ID', default=None))
|
||||||
|
if is_live and not current_live:
|
||||||
# fetch unfiltered /show to find running livestreams; we can't get ID of the password-protected livestream above
|
# fetch unfiltered /show to find running livestreams; we can't get ID of the password-protected livestream above
|
||||||
webpage = self._download_webpage(
|
webpage = self._download_webpage(
|
||||||
f'https://twitcasting.tv/{uploader_id}/show/', uploader_id,
|
f'https://twitcasting.tv/{uploader_id}/show/', uploader_id,
|
||||||
note='Downloading live history')
|
note='Downloading live history')
|
||||||
is_live = self._search_regex(r'(?s)(<span\s*class="tw-movie-thumbnail-badge"\s*data-status="live">\s*LIVE)', webpage, 'is live?', default=None)
|
is_live = self._search_regex(self._PROTECTED_LIVE_RE, webpage, 'is live?', default=None)
|
||||||
if is_live:
|
if is_live:
|
||||||
# get the first live; running live is always at the first
|
# get the first live; running live is always at the first
|
||||||
current_live = self._search_regex(
|
current_live = self._search_regex(
|
||||||
r'(?s)<a\s+class="tw-movie-thumbnail"\s*href="/[^/]+/movie/(?P<video_id>\d+)"\s*>.+?</a>',
|
r'(?s)<a\s+class="tw-movie-thumbnail2"\s*href="/[^/]+/movie/(?P<video_id>\d+)"\s*>.+?</a>',
|
||||||
webpage, 'current live ID 2', default=None, group='video_id')
|
webpage, 'current live ID 2', default=None, group='video_id')
|
||||||
if not current_live:
|
if not current_live:
|
||||||
raise UserNotLive(video_id=uploader_id)
|
raise UserNotLive(video_id=uploader_id)
|
||||||
return self.url_result('https://twitcasting.tv/%s/movie/%s' % (uploader_id, current_live))
|
return self.url_result(f'https://twitcasting.tv/{uploader_id}/movie/{current_live}', TwitCastingIE)
|
||||||
|
|
||||||
|
|
||||||
class TwitCastingUserIE(InfoExtractor):
|
class TwitCastingUserIE(InfoExtractor):
|
||||||
|
|||||||
@@ -1563,7 +1563,7 @@ class TwitterBroadcastIE(TwitterBaseIE, PeriscopeBaseIE):
|
|||||||
IE_NAME = 'twitter:broadcast'
|
IE_NAME = 'twitter:broadcast'
|
||||||
_VALID_URL = TwitterBaseIE._BASE_REGEX + r'i/broadcasts/(?P<id>[0-9a-zA-Z]{13})'
|
_VALID_URL = TwitterBaseIE._BASE_REGEX + r'i/broadcasts/(?P<id>[0-9a-zA-Z]{13})'
|
||||||
|
|
||||||
_TEST = {
|
_TESTS = [{
|
||||||
# untitled Periscope video
|
# untitled Periscope video
|
||||||
'url': 'https://twitter.com/i/broadcasts/1yNGaQLWpejGj',
|
'url': 'https://twitter.com/i/broadcasts/1yNGaQLWpejGj',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
@@ -1571,11 +1571,42 @@ class TwitterBroadcastIE(TwitterBaseIE, PeriscopeBaseIE):
|
|||||||
'ext': 'mp4',
|
'ext': 'mp4',
|
||||||
'title': 'Andrea May Sahouri - Periscope Broadcast',
|
'title': 'Andrea May Sahouri - Periscope Broadcast',
|
||||||
'uploader': 'Andrea May Sahouri',
|
'uploader': 'Andrea May Sahouri',
|
||||||
'uploader_id': '1PXEdBZWpGwKe',
|
'uploader_id': 'andreamsahouri',
|
||||||
|
'uploader_url': 'https://twitter.com/andreamsahouri',
|
||||||
|
'timestamp': 1590973638,
|
||||||
|
'upload_date': '20200601',
|
||||||
'thumbnail': r're:^https?://[^?#]+\.jpg\?token=',
|
'thumbnail': r're:^https?://[^?#]+\.jpg\?token=',
|
||||||
'view_count': int,
|
'view_count': int,
|
||||||
},
|
},
|
||||||
}
|
}, {
|
||||||
|
'url': 'https://twitter.com/i/broadcasts/1ZkKzeyrPbaxv',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '1ZkKzeyrPbaxv',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': 'Starship | SN10 | High-Altitude Flight Test',
|
||||||
|
'uploader': 'SpaceX',
|
||||||
|
'uploader_id': 'SpaceX',
|
||||||
|
'uploader_url': 'https://twitter.com/SpaceX',
|
||||||
|
'timestamp': 1614812942,
|
||||||
|
'upload_date': '20210303',
|
||||||
|
'thumbnail': r're:^https?://[^?#]+\.jpg\?token=',
|
||||||
|
'view_count': int,
|
||||||
|
},
|
||||||
|
}, {
|
||||||
|
'url': 'https://twitter.com/i/broadcasts/1OyKAVQrgzwGb',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '1OyKAVQrgzwGb',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': 'Starship Flight Test',
|
||||||
|
'uploader': 'SpaceX',
|
||||||
|
'uploader_id': 'SpaceX',
|
||||||
|
'uploader_url': 'https://twitter.com/SpaceX',
|
||||||
|
'timestamp': 1681993964,
|
||||||
|
'upload_date': '20230420',
|
||||||
|
'thumbnail': r're:^https?://[^?#]+\.jpg\?token=',
|
||||||
|
'view_count': int,
|
||||||
|
},
|
||||||
|
}]
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
broadcast_id = self._match_id(url)
|
broadcast_id = self._match_id(url)
|
||||||
@@ -1585,6 +1616,12 @@ class TwitterBroadcastIE(TwitterBaseIE, PeriscopeBaseIE):
|
|||||||
if not broadcast:
|
if not broadcast:
|
||||||
raise ExtractorError('Broadcast no longer exists', expected=True)
|
raise ExtractorError('Broadcast no longer exists', expected=True)
|
||||||
info = self._parse_broadcast_data(broadcast, broadcast_id)
|
info = self._parse_broadcast_data(broadcast, broadcast_id)
|
||||||
|
info['title'] = broadcast.get('status') or info.get('title')
|
||||||
|
info['uploader_id'] = broadcast.get('twitter_username') or info.get('uploader_id')
|
||||||
|
info['uploader_url'] = format_field(broadcast, 'twitter_username', 'https://twitter.com/%s', default=None)
|
||||||
|
if info['live_status'] == 'is_upcoming':
|
||||||
|
return info
|
||||||
|
|
||||||
media_key = broadcast['media_key']
|
media_key = broadcast['media_key']
|
||||||
source = self._call_api(
|
source = self._call_api(
|
||||||
f'live_video_stream/status/{media_key}', media_key)['source']
|
f'live_video_stream/status/{media_key}', media_key)['source']
|
||||||
|
|||||||
@@ -48,6 +48,7 @@ class KnownDRMIE(UnsupportedInfoExtractor):
|
|||||||
r'joyn\.de',
|
r'joyn\.de',
|
||||||
r'amazon\.(?:\w{2}\.)?\w+/gp/video',
|
r'amazon\.(?:\w{2}\.)?\w+/gp/video',
|
||||||
r'music\.amazon\.(?:\w{2}\.)?\w+',
|
r'music\.amazon\.(?:\w{2}\.)?\w+',
|
||||||
|
r'(?:watch|front)\.njpwworld\.com',
|
||||||
)
|
)
|
||||||
|
|
||||||
_TESTS = [{
|
_TESTS = [{
|
||||||
@@ -141,6 +142,13 @@ class KnownDRMIE(UnsupportedInfoExtractor):
|
|||||||
# https://github.com/yt-dlp/yt-dlp/issues/5767
|
# https://github.com/yt-dlp/yt-dlp/issues/5767
|
||||||
'url': 'https://www.hulu.com/movie/anthem-6b25fac9-da2b-45a3-8e09-e4156b0471cc',
|
'url': 'https://www.hulu.com/movie/anthem-6b25fac9-da2b-45a3-8e09-e4156b0471cc',
|
||||||
'only_matching': True,
|
'only_matching': True,
|
||||||
|
}, {
|
||||||
|
# https://github.com/yt-dlp/yt-dlp/pull/8570
|
||||||
|
'url': 'https://watch.njpwworld.com/player/36447/series?assetType=series',
|
||||||
|
'only_matching': True,
|
||||||
|
}, {
|
||||||
|
'url': 'https://front.njpwworld.com/p/s_series_00563_16_bs',
|
||||||
|
'only_matching': True,
|
||||||
}]
|
}]
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
@@ -164,11 +172,15 @@ class KnownPiracyIE(UnsupportedInfoExtractor):
|
|||||||
r'viewsb\.com',
|
r'viewsb\.com',
|
||||||
r'filemoon\.sx',
|
r'filemoon\.sx',
|
||||||
r'hentai\.animestigma\.com',
|
r'hentai\.animestigma\.com',
|
||||||
|
r'thisav\.com',
|
||||||
)
|
)
|
||||||
|
|
||||||
_TESTS = [{
|
_TESTS = [{
|
||||||
'url': 'http://dood.to/e/5s1wmbdacezb',
|
'url': 'http://dood.to/e/5s1wmbdacezb',
|
||||||
'only_matching': True,
|
'only_matching': True,
|
||||||
|
}, {
|
||||||
|
'url': 'https://thisav.com/en/terms',
|
||||||
|
'only_matching': True,
|
||||||
}]
|
}]
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
|
|||||||
@@ -11,6 +11,7 @@ from ..utils import (
|
|||||||
ExtractorError,
|
ExtractorError,
|
||||||
InAdvancePagedList,
|
InAdvancePagedList,
|
||||||
int_or_none,
|
int_or_none,
|
||||||
|
remove_start,
|
||||||
traverse_obj,
|
traverse_obj,
|
||||||
update_url_query,
|
update_url_query,
|
||||||
url_or_none,
|
url_or_none,
|
||||||
@@ -39,11 +40,11 @@ class VideoKenBaseIE(InfoExtractor):
|
|||||||
if not video_url and not video_id:
|
if not video_url and not video_id:
|
||||||
return
|
return
|
||||||
elif not video_url or 'embed/sign-in' in video_url:
|
elif not video_url or 'embed/sign-in' in video_url:
|
||||||
video_url = f'https://slideslive.com/embed/{video_id.lstrip("slideslive-")}'
|
video_url = f'https://slideslive.com/embed/{remove_start(video_id, "slideslive-")}'
|
||||||
if url_or_none(referer):
|
if url_or_none(referer):
|
||||||
return update_url_query(video_url, {
|
return update_url_query(video_url, {
|
||||||
'embed_parent_url': referer,
|
'embed_parent_url': referer,
|
||||||
'embed_container_origin': f'https://{urllib.parse.urlparse(referer).netloc}',
|
'embed_container_origin': f'https://{urllib.parse.urlparse(referer).hostname}',
|
||||||
})
|
})
|
||||||
return video_url
|
return video_url
|
||||||
|
|
||||||
@@ -57,12 +58,12 @@ class VideoKenBaseIE(InfoExtractor):
|
|||||||
video_url = video_id
|
video_url = video_id
|
||||||
ie_key = 'Youtube'
|
ie_key = 'Youtube'
|
||||||
else:
|
else:
|
||||||
video_url = traverse_obj(video, 'embed_url', 'embeddableurl')
|
video_url = traverse_obj(video, 'embed_url', 'embeddableurl', expected_type=url_or_none)
|
||||||
if urllib.parse.urlparse(video_url).netloc == 'slideslive.com':
|
if not video_url:
|
||||||
|
continue
|
||||||
|
elif urllib.parse.urlparse(video_url).hostname == 'slideslive.com':
|
||||||
ie_key = SlidesLiveIE
|
ie_key = SlidesLiveIE
|
||||||
video_url = self._create_slideslive_url(video_url, video_id, url)
|
video_url = self._create_slideslive_url(video_url, video_id, url)
|
||||||
if not video_url:
|
|
||||||
continue
|
|
||||||
yield self.url_result(video_url, ie_key, video_id)
|
yield self.url_result(video_url, ie_key, video_id)
|
||||||
|
|
||||||
|
|
||||||
@@ -178,7 +179,7 @@ class VideoKenIE(VideoKenBaseIE):
|
|||||||
return self.url_result(
|
return self.url_result(
|
||||||
self._create_slideslive_url(None, video_id, url), SlidesLiveIE, video_id)
|
self._create_slideslive_url(None, video_id, url), SlidesLiveIE, video_id)
|
||||||
elif re.match(r'^[\w-]{11}$', video_id):
|
elif re.match(r'^[\w-]{11}$', video_id):
|
||||||
self.url_result(video_id, 'Youtube', video_id)
|
return self.url_result(video_id, 'Youtube', video_id)
|
||||||
else:
|
else:
|
||||||
raise ExtractorError('Unable to extract without VideoKen API response')
|
raise ExtractorError('Unable to extract without VideoKen API response')
|
||||||
|
|
||||||
|
|||||||
@@ -37,14 +37,14 @@ class VimeoBaseInfoExtractor(InfoExtractor):
|
|||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def _smuggle_referrer(url, referrer_url):
|
def _smuggle_referrer(url, referrer_url):
|
||||||
return smuggle_url(url, {'http_headers': {'Referer': referrer_url}})
|
return smuggle_url(url, {'referer': referrer_url})
|
||||||
|
|
||||||
def _unsmuggle_headers(self, url):
|
def _unsmuggle_headers(self, url):
|
||||||
"""@returns (url, smuggled_data, headers)"""
|
"""@returns (url, smuggled_data, headers)"""
|
||||||
url, data = unsmuggle_url(url, {})
|
url, data = unsmuggle_url(url, {})
|
||||||
headers = self.get_param('http_headers').copy()
|
headers = self.get_param('http_headers').copy()
|
||||||
if 'http_headers' in data:
|
if 'referer' in data:
|
||||||
headers.update(data['http_headers'])
|
headers['Referer'] = data['referer']
|
||||||
return url, data, headers
|
return url, data, headers
|
||||||
|
|
||||||
def _perform_login(self, username, password):
|
def _perform_login(self, username, password):
|
||||||
|
|||||||
@@ -1,3 +1,4 @@
|
|||||||
|
import json
|
||||||
import random
|
import random
|
||||||
import itertools
|
import itertools
|
||||||
import urllib.parse
|
import urllib.parse
|
||||||
@@ -18,24 +19,33 @@ from ..utils import (
|
|||||||
|
|
||||||
|
|
||||||
class WeiboBaseIE(InfoExtractor):
|
class WeiboBaseIE(InfoExtractor):
|
||||||
def _update_visitor_cookies(self, video_id):
|
def _update_visitor_cookies(self, visitor_url, video_id):
|
||||||
|
headers = {'Referer': visitor_url}
|
||||||
|
chrome_ver = self._search_regex(
|
||||||
|
r'Chrome/(\d+)', self.get_param('http_headers')['User-Agent'], 'user agent version', default='90')
|
||||||
visitor_data = self._download_json(
|
visitor_data = self._download_json(
|
||||||
'https://passport.weibo.com/visitor/genvisitor', video_id,
|
'https://passport.weibo.com/visitor/genvisitor', video_id,
|
||||||
note='Generating first-visit guest request',
|
note='Generating first-visit guest request',
|
||||||
transform_source=strip_jsonp,
|
headers=headers, transform_source=strip_jsonp,
|
||||||
data=urlencode_postdata({
|
data=urlencode_postdata({
|
||||||
'cb': 'gen_callback',
|
'cb': 'gen_callback',
|
||||||
'fp': '{"os":"2","browser":"Gecko57,0,0,0","fonts":"undefined","screenInfo":"1440*900*24","plugins":""}',
|
'fp': json.dumps({
|
||||||
}))
|
'os': '1',
|
||||||
|
'browser': f'Chrome{chrome_ver},0,0,0',
|
||||||
|
'fonts': 'undefined',
|
||||||
|
'screenInfo': '1920*1080*24',
|
||||||
|
'plugins': ''
|
||||||
|
}, separators=(',', ':'))}))['data']
|
||||||
|
|
||||||
self._download_webpage(
|
self._download_webpage(
|
||||||
'https://passport.weibo.com/visitor/visitor', video_id,
|
'https://passport.weibo.com/visitor/visitor', video_id,
|
||||||
note='Running first-visit callback to get guest cookies',
|
note='Running first-visit callback to get guest cookies',
|
||||||
query={
|
headers=headers, query={
|
||||||
'a': 'incarnate',
|
'a': 'incarnate',
|
||||||
't': visitor_data['data']['tid'],
|
't': visitor_data['tid'],
|
||||||
'w': 2,
|
'w': 3 if visitor_data.get('new_tid') else 2,
|
||||||
'c': '%03d' % visitor_data['data']['confidence'],
|
'c': f'{visitor_data.get("confidence", 100):03d}',
|
||||||
|
'gc': '',
|
||||||
'cb': 'cross_domain',
|
'cb': 'cross_domain',
|
||||||
'from': 'weibo',
|
'from': 'weibo',
|
||||||
'_rand': random.random(),
|
'_rand': random.random(),
|
||||||
@@ -44,7 +54,7 @@ class WeiboBaseIE(InfoExtractor):
|
|||||||
def _weibo_download_json(self, url, video_id, *args, fatal=True, note='Downloading JSON metadata', **kwargs):
|
def _weibo_download_json(self, url, video_id, *args, fatal=True, note='Downloading JSON metadata', **kwargs):
|
||||||
webpage, urlh = self._download_webpage_handle(url, video_id, *args, fatal=fatal, note=note, **kwargs)
|
webpage, urlh = self._download_webpage_handle(url, video_id, *args, fatal=fatal, note=note, **kwargs)
|
||||||
if urllib.parse.urlparse(urlh.url).netloc == 'passport.weibo.com':
|
if urllib.parse.urlparse(urlh.url).netloc == 'passport.weibo.com':
|
||||||
self._update_visitor_cookies(video_id)
|
self._update_visitor_cookies(urlh.url, video_id)
|
||||||
webpage = self._download_webpage(url, video_id, *args, fatal=fatal, note=note, **kwargs)
|
webpage = self._download_webpage(url, video_id, *args, fatal=fatal, note=note, **kwargs)
|
||||||
return self._parse_json(webpage, video_id, fatal=fatal)
|
return self._parse_json(webpage, video_id, fatal=fatal)
|
||||||
|
|
||||||
|
|||||||
@@ -45,10 +45,10 @@ class WeverseBaseIE(InfoExtractor):
|
|||||||
'x-acc-trace-id': str(uuid.uuid4()),
|
'x-acc-trace-id': str(uuid.uuid4()),
|
||||||
'x-clog-user-device-id': str(uuid.uuid4()),
|
'x-clog-user-device-id': str(uuid.uuid4()),
|
||||||
}
|
}
|
||||||
check_username = self._download_json(
|
valid_username = traverse_obj(self._download_json(
|
||||||
f'{self._ACCOUNT_API_BASE}/signup/email/status', None,
|
f'{self._ACCOUNT_API_BASE}/signup/email/status', None, note='Checking username',
|
||||||
note='Checking username', query={'email': username}, headers=headers)
|
query={'email': username}, headers=headers, expected_status=(400, 404)), 'hasPassword')
|
||||||
if not check_username.get('hasPassword'):
|
if not valid_username:
|
||||||
raise ExtractorError('Invalid username provided', expected=True)
|
raise ExtractorError('Invalid username provided', expected=True)
|
||||||
|
|
||||||
headers['content-type'] = 'application/json'
|
headers['content-type'] = 'application/json'
|
||||||
|
|||||||
@@ -194,7 +194,7 @@ class ZenYandexIE(InfoExtractor):
|
|||||||
'id': '60c7c443da18892ebfe85ed7',
|
'id': '60c7c443da18892ebfe85ed7',
|
||||||
'ext': 'mp4',
|
'ext': 'mp4',
|
||||||
'title': 'ВОТ ЭТО Focus. Деды Морозы на гидроциклах',
|
'title': 'ВОТ ЭТО Focus. Деды Морозы на гидроциклах',
|
||||||
'description': 'md5:f3db3d995763b9bbb7b56d4ccdedea89',
|
'description': 'md5:8684912f6086f298f8078d4af0e8a600',
|
||||||
'thumbnail': 're:^https://avatars.dzeninfra.ru/',
|
'thumbnail': 're:^https://avatars.dzeninfra.ru/',
|
||||||
'uploader': 'AcademeG DailyStream'
|
'uploader': 'AcademeG DailyStream'
|
||||||
},
|
},
|
||||||
@@ -209,7 +209,7 @@ class ZenYandexIE(InfoExtractor):
|
|||||||
'id': '60c7c443da18892ebfe85ed7',
|
'id': '60c7c443da18892ebfe85ed7',
|
||||||
'ext': 'mp4',
|
'ext': 'mp4',
|
||||||
'title': 'ВОТ ЭТО Focus. Деды Морозы на гидроциклах',
|
'title': 'ВОТ ЭТО Focus. Деды Морозы на гидроциклах',
|
||||||
'description': 'md5:f3db3d995763b9bbb7b56d4ccdedea89',
|
'description': 'md5:8684912f6086f298f8078d4af0e8a600',
|
||||||
'thumbnail': r're:^https://avatars\.dzeninfra\.ru/',
|
'thumbnail': r're:^https://avatars\.dzeninfra\.ru/',
|
||||||
'uploader': 'AcademeG DailyStream',
|
'uploader': 'AcademeG DailyStream',
|
||||||
'upload_date': '20191111',
|
'upload_date': '20191111',
|
||||||
@@ -258,7 +258,7 @@ class ZenYandexIE(InfoExtractor):
|
|||||||
video_id = self._match_id(redirect)
|
video_id = self._match_id(redirect)
|
||||||
webpage = self._download_webpage(redirect, video_id, note='Redirecting')
|
webpage = self._download_webpage(redirect, video_id, note='Redirecting')
|
||||||
data_json = self._search_json(
|
data_json = self._search_json(
|
||||||
r'data\s*=', webpage, 'metadata', video_id, contains_pattern=r'{["\']_*serverState_*video.+}')
|
r'("data"\s*:|data\s*=)', webpage, 'metadata', video_id, contains_pattern=r'{["\']_*serverState_*video.+}')
|
||||||
serverstate = self._search_regex(r'(_+serverState_+video-site_[^_]+_+)',
|
serverstate = self._search_regex(r'(_+serverState_+video-site_[^_]+_+)',
|
||||||
webpage, 'server state').replace('State', 'Settings')
|
webpage, 'server state').replace('State', 'Settings')
|
||||||
uploader = self._search_regex(r'(<a\s*class=["\']card-channel-link[^"\']+["\'][^>]+>)',
|
uploader = self._search_regex(r'(<a\s*class=["\']card-channel-link[^"\']+["\'][^>]+>)',
|
||||||
@@ -266,22 +266,25 @@ class ZenYandexIE(InfoExtractor):
|
|||||||
uploader_name = extract_attributes(uploader).get('aria-label')
|
uploader_name = extract_attributes(uploader).get('aria-label')
|
||||||
video_json = try_get(data_json, lambda x: x[serverstate]['exportData']['video'], dict)
|
video_json = try_get(data_json, lambda x: x[serverstate]['exportData']['video'], dict)
|
||||||
stream_urls = try_get(video_json, lambda x: x['video']['streams'])
|
stream_urls = try_get(video_json, lambda x: x['video']['streams'])
|
||||||
formats = []
|
formats, subtitles = [], {}
|
||||||
for s_url in stream_urls:
|
for s_url in stream_urls:
|
||||||
ext = determine_ext(s_url)
|
ext = determine_ext(s_url)
|
||||||
if ext == 'mpd':
|
if ext == 'mpd':
|
||||||
formats.extend(self._extract_mpd_formats(s_url, video_id, mpd_id='dash'))
|
fmts, subs = self._extract_mpd_formats_and_subtitles(s_url, video_id, mpd_id='dash')
|
||||||
elif ext == 'm3u8':
|
elif ext == 'm3u8':
|
||||||
formats.extend(self._extract_m3u8_formats(s_url, video_id, 'mp4'))
|
fmts, subs = self._extract_m3u8_formats_and_subtitles(s_url, video_id, 'mp4')
|
||||||
|
formats.extend(fmts)
|
||||||
|
subtitles = self._merge_subtitles(subtitles, subs)
|
||||||
return {
|
return {
|
||||||
'id': video_id,
|
'id': video_id,
|
||||||
'title': video_json.get('title') or self._og_search_title(webpage),
|
'title': video_json.get('title') or self._og_search_title(webpage),
|
||||||
'formats': formats,
|
'formats': formats,
|
||||||
|
'subtitles': subtitles,
|
||||||
'duration': int_or_none(video_json.get('duration')),
|
'duration': int_or_none(video_json.get('duration')),
|
||||||
'view_count': int_or_none(video_json.get('views')),
|
'view_count': int_or_none(video_json.get('views')),
|
||||||
'timestamp': int_or_none(video_json.get('publicationDate')),
|
'timestamp': int_or_none(video_json.get('publicationDate')),
|
||||||
'uploader': uploader_name or data_json.get('authorName') or try_get(data_json, lambda x: x['publisher']['name']),
|
'uploader': uploader_name or data_json.get('authorName') or try_get(data_json, lambda x: x['publisher']['name']),
|
||||||
'description': self._og_search_description(webpage) or try_get(data_json, lambda x: x['og']['description']),
|
'description': video_json.get('description') or self._og_search_description(webpage),
|
||||||
'thumbnail': self._og_search_thumbnail(webpage) or try_get(data_json, lambda x: x['og']['imageUrl']),
|
'thumbnail': self._og_search_thumbnail(webpage) or try_get(data_json, lambda x: x['og']['imageUrl']),
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -296,6 +299,7 @@ class ZenYandexChannelIE(InfoExtractor):
|
|||||||
'description': 'md5:a9e5b3c247b7fe29fd21371a428bcf56',
|
'description': 'md5:a9e5b3c247b7fe29fd21371a428bcf56',
|
||||||
},
|
},
|
||||||
'playlist_mincount': 169,
|
'playlist_mincount': 169,
|
||||||
|
'skip': 'The page does not exist',
|
||||||
}, {
|
}, {
|
||||||
'url': 'https://dzen.ru/tok_media',
|
'url': 'https://dzen.ru/tok_media',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
@@ -304,6 +308,7 @@ class ZenYandexChannelIE(InfoExtractor):
|
|||||||
'description': 'md5:a9e5b3c247b7fe29fd21371a428bcf56',
|
'description': 'md5:a9e5b3c247b7fe29fd21371a428bcf56',
|
||||||
},
|
},
|
||||||
'playlist_mincount': 169,
|
'playlist_mincount': 169,
|
||||||
|
'skip': 'The page does not exist',
|
||||||
}, {
|
}, {
|
||||||
'url': 'https://zen.yandex.ru/id/606fd806cc13cb3c58c05cf5',
|
'url': 'https://zen.yandex.ru/id/606fd806cc13cb3c58c05cf5',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
@@ -318,21 +323,21 @@ class ZenYandexChannelIE(InfoExtractor):
|
|||||||
'url': 'https://zen.yandex.ru/jony_me',
|
'url': 'https://zen.yandex.ru/jony_me',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'jony_me',
|
'id': 'jony_me',
|
||||||
'description': 'md5:a2c62b4ef5cf3e3efb13d25f61f739e1',
|
'description': 'md5:ce0a5cad2752ab58701b5497835b2cc5',
|
||||||
'title': 'JONY ',
|
'title': 'JONY ',
|
||||||
},
|
},
|
||||||
'playlist_count': 20,
|
'playlist_count': 18,
|
||||||
}, {
|
}, {
|
||||||
# Test that the playlist extractor finishes extracting when the
|
# Test that the playlist extractor finishes extracting when the
|
||||||
# channel has more than one page of entries
|
# channel has more than one page of entries
|
||||||
'url': 'https://zen.yandex.ru/tatyanareva',
|
'url': 'https://zen.yandex.ru/tatyanareva',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'tatyanareva',
|
'id': 'tatyanareva',
|
||||||
'description': 'md5:296b588d60841c3756c9105f237b70c6',
|
'description': 'md5:40a1e51f174369ec3ba9d657734ac31f',
|
||||||
'title': 'Татьяна Рева',
|
'title': 'Татьяна Рева',
|
||||||
'entries': 'maxcount:200',
|
'entries': 'maxcount:200',
|
||||||
},
|
},
|
||||||
'playlist_count': 46,
|
'playlist_mincount': 46,
|
||||||
}, {
|
}, {
|
||||||
'url': 'https://dzen.ru/id/606fd806cc13cb3c58c05cf5',
|
'url': 'https://dzen.ru/id/606fd806cc13cb3c58c05cf5',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
@@ -375,7 +380,7 @@ class ZenYandexChannelIE(InfoExtractor):
|
|||||||
item_id = self._match_id(redirect)
|
item_id = self._match_id(redirect)
|
||||||
webpage = self._download_webpage(redirect, item_id, note='Redirecting')
|
webpage = self._download_webpage(redirect, item_id, note='Redirecting')
|
||||||
data = self._search_json(
|
data = self._search_json(
|
||||||
r'var\s+data\s*=', webpage, 'channel data', item_id, contains_pattern=r'{\"__serverState__.+}')
|
r'("data"\s*:|data\s*=)', webpage, 'channel data', item_id, contains_pattern=r'{\"__serverState__.+}')
|
||||||
server_state_json = traverse_obj(data, lambda k, _: k.startswith('__serverState__'), get_all=False)
|
server_state_json = traverse_obj(data, lambda k, _: k.startswith('__serverState__'), get_all=False)
|
||||||
server_settings_json = traverse_obj(data, lambda k, _: k.startswith('__serverSettings__'), get_all=False)
|
server_settings_json = traverse_obj(data, lambda k, _: k.startswith('__serverSettings__'), get_all=False)
|
||||||
|
|
||||||
|
|||||||
@@ -4560,6 +4560,14 @@ class YoutubeIE(YoutubeBaseInfoExtractor):
|
|||||||
self._parse_time_text(self._get_text(vpir, 'dateText'))) or upload_date
|
self._parse_time_text(self._get_text(vpir, 'dateText'))) or upload_date
|
||||||
info['upload_date'] = upload_date
|
info['upload_date'] = upload_date
|
||||||
|
|
||||||
|
if upload_date and live_status not in ('is_live', 'post_live', 'is_upcoming'):
|
||||||
|
# Newly uploaded videos' HLS formats are potentially problematic and need to be checked
|
||||||
|
upload_datetime = datetime_from_str(upload_date).replace(tzinfo=datetime.timezone.utc)
|
||||||
|
if upload_datetime >= datetime_from_str('today-1day'):
|
||||||
|
for fmt in info['formats']:
|
||||||
|
if fmt.get('protocol') == 'm3u8_native':
|
||||||
|
fmt['__needs_testing'] = True
|
||||||
|
|
||||||
for s_k, d_k in [('artist', 'creator'), ('track', 'alt_title')]:
|
for s_k, d_k in [('artist', 'creator'), ('track', 'alt_title')]:
|
||||||
v = info.get(s_k)
|
v = info.get(s_k)
|
||||||
if v:
|
if v:
|
||||||
@@ -6679,7 +6687,7 @@ class YoutubePlaylistIE(InfoExtractor):
|
|||||||
'uploader_url': 'https://www.youtube.com/@milan5503',
|
'uploader_url': 'https://www.youtube.com/@milan5503',
|
||||||
'availability': 'public',
|
'availability': 'public',
|
||||||
},
|
},
|
||||||
'expected_warnings': [r'[Uu]navailable videos? (is|are|will be) hidden'],
|
'expected_warnings': [r'[Uu]navailable videos? (is|are|will be) hidden', 'Retrying', 'Giving up'],
|
||||||
}, {
|
}, {
|
||||||
'url': 'http://www.youtube.com/embed/_xDOZElKyNU?list=PLsyOSbh5bs16vubvKePAQ1x3PhKavfBIl',
|
'url': 'http://www.youtube.com/embed/_xDOZElKyNU?list=PLsyOSbh5bs16vubvKePAQ1x3PhKavfBIl',
|
||||||
'playlist_mincount': 455,
|
'playlist_mincount': 455,
|
||||||
|
|||||||
@@ -2,10 +2,12 @@ from .common import InfoExtractor
|
|||||||
from ..utils import (
|
from ..utils import (
|
||||||
ExtractorError,
|
ExtractorError,
|
||||||
int_or_none,
|
int_or_none,
|
||||||
str_or_none,
|
|
||||||
js_to_json,
|
js_to_json,
|
||||||
parse_filesize,
|
parse_filesize,
|
||||||
|
parse_resolution,
|
||||||
|
str_or_none,
|
||||||
traverse_obj,
|
traverse_obj,
|
||||||
|
url_basename,
|
||||||
urlencode_postdata,
|
urlencode_postdata,
|
||||||
urljoin,
|
urljoin,
|
||||||
)
|
)
|
||||||
@@ -41,6 +43,18 @@ class ZoomIE(InfoExtractor):
|
|||||||
'ext': 'mp4',
|
'ext': 'mp4',
|
||||||
'title': 'Timea Andrea Lelik\'s Personal Meeting Room',
|
'title': 'Timea Andrea Lelik\'s Personal Meeting Room',
|
||||||
},
|
},
|
||||||
|
'skip': 'This recording has expired',
|
||||||
|
}, {
|
||||||
|
# view_with_share URL
|
||||||
|
'url': 'https://cityofdetroit.zoom.us/rec/share/VjE-5kW3xmgbEYqR5KzRgZ1OFZvtMtiXk5HyRJo5kK4m5PYE6RF4rF_oiiO_9qaM.UTAg1MI7JSnF3ZjX',
|
||||||
|
'md5': 'bdc7867a5934c151957fb81321b3c024',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'VjE-5kW3xmgbEYqR5KzRgZ1OFZvtMtiXk5HyRJo5kK4m5PYE6RF4rF_oiiO_9qaM.UTAg1MI7JSnF3ZjX',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': 'February 2022 Detroit Revenue Estimating Conference',
|
||||||
|
'duration': 7299,
|
||||||
|
'formats': 'mincount:3',
|
||||||
|
},
|
||||||
}]
|
}]
|
||||||
|
|
||||||
def _get_page_data(self, webpage, video_id):
|
def _get_page_data(self, webpage, video_id):
|
||||||
@@ -72,6 +86,7 @@ class ZoomIE(InfoExtractor):
|
|||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
base_url, url_type, video_id = self._match_valid_url(url).group('base_url', 'type', 'id')
|
base_url, url_type, video_id = self._match_valid_url(url).group('base_url', 'type', 'id')
|
||||||
|
query = {}
|
||||||
|
|
||||||
if url_type == 'share':
|
if url_type == 'share':
|
||||||
webpage = self._get_real_webpage(url, base_url, video_id, 'share')
|
webpage = self._get_real_webpage(url, base_url, video_id, 'share')
|
||||||
@@ -80,6 +95,7 @@ class ZoomIE(InfoExtractor):
|
|||||||
f'{base_url}nws/recording/1.0/play/share-info/{meeting_id}',
|
f'{base_url}nws/recording/1.0/play/share-info/{meeting_id}',
|
||||||
video_id, note='Downloading share info JSON')['result']['redirectUrl']
|
video_id, note='Downloading share info JSON')['result']['redirectUrl']
|
||||||
url = urljoin(base_url, redirect_path)
|
url = urljoin(base_url, redirect_path)
|
||||||
|
query['continueMode'] = 'true'
|
||||||
|
|
||||||
webpage = self._get_real_webpage(url, base_url, video_id, 'play')
|
webpage = self._get_real_webpage(url, base_url, video_id, 'play')
|
||||||
file_id = self._get_page_data(webpage, video_id)['fileId']
|
file_id = self._get_page_data(webpage, video_id)['fileId']
|
||||||
@@ -88,7 +104,7 @@ class ZoomIE(InfoExtractor):
|
|||||||
raise ExtractorError('Unable to extract file ID')
|
raise ExtractorError('Unable to extract file ID')
|
||||||
|
|
||||||
data = self._download_json(
|
data = self._download_json(
|
||||||
f'{base_url}nws/recording/1.0/play/info/{file_id}', video_id,
|
f'{base_url}nws/recording/1.0/play/info/{file_id}', video_id, query=query,
|
||||||
note='Downloading play info JSON')['result']
|
note='Downloading play info JSON')['result']
|
||||||
|
|
||||||
subtitles = {}
|
subtitles = {}
|
||||||
@@ -104,10 +120,10 @@ class ZoomIE(InfoExtractor):
|
|||||||
if data.get('viewMp4Url'):
|
if data.get('viewMp4Url'):
|
||||||
formats.append({
|
formats.append({
|
||||||
'format_note': 'Camera stream',
|
'format_note': 'Camera stream',
|
||||||
'url': str_or_none(data.get('viewMp4Url')),
|
'url': data['viewMp4Url'],
|
||||||
'width': int_or_none(traverse_obj(data, ('viewResolvtions', 0))),
|
'width': int_or_none(traverse_obj(data, ('viewResolvtions', 0))),
|
||||||
'height': int_or_none(traverse_obj(data, ('viewResolvtions', 1))),
|
'height': int_or_none(traverse_obj(data, ('viewResolvtions', 1))),
|
||||||
'format_id': str_or_none(traverse_obj(data, ('recording', 'id'))),
|
'format_id': 'view',
|
||||||
'ext': 'mp4',
|
'ext': 'mp4',
|
||||||
'filesize_approx': parse_filesize(str_or_none(traverse_obj(data, ('recording', 'fileSizeInMB')))),
|
'filesize_approx': parse_filesize(str_or_none(traverse_obj(data, ('recording', 'fileSizeInMB')))),
|
||||||
'preference': 0
|
'preference': 0
|
||||||
@@ -116,14 +132,26 @@ class ZoomIE(InfoExtractor):
|
|||||||
if data.get('shareMp4Url'):
|
if data.get('shareMp4Url'):
|
||||||
formats.append({
|
formats.append({
|
||||||
'format_note': 'Screen share stream',
|
'format_note': 'Screen share stream',
|
||||||
'url': str_or_none(data.get('shareMp4Url')),
|
'url': data['shareMp4Url'],
|
||||||
'width': int_or_none(traverse_obj(data, ('shareResolvtions', 0))),
|
'width': int_or_none(traverse_obj(data, ('shareResolvtions', 0))),
|
||||||
'height': int_or_none(traverse_obj(data, ('shareResolvtions', 1))),
|
'height': int_or_none(traverse_obj(data, ('shareResolvtions', 1))),
|
||||||
'format_id': str_or_none(traverse_obj(data, ('shareVideo', 'id'))),
|
'format_id': 'share',
|
||||||
'ext': 'mp4',
|
'ext': 'mp4',
|
||||||
'preference': -1
|
'preference': -1
|
||||||
})
|
})
|
||||||
|
|
||||||
|
view_with_share_url = data.get('viewMp4WithshareUrl')
|
||||||
|
if view_with_share_url:
|
||||||
|
formats.append({
|
||||||
|
**parse_resolution(self._search_regex(
|
||||||
|
r'_(\d+x\d+)\.mp4', url_basename(view_with_share_url), 'resolution', default=None)),
|
||||||
|
'format_note': 'Screen share with camera',
|
||||||
|
'url': view_with_share_url,
|
||||||
|
'format_id': 'view_with_share',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'preference': 1
|
||||||
|
})
|
||||||
|
|
||||||
return {
|
return {
|
||||||
'id': video_id,
|
'id': video_id,
|
||||||
'title': str_or_none(traverse_obj(data, ('meet', 'topic'))),
|
'title': str_or_none(traverse_obj(data, ('meet', 'topic'))),
|
||||||
|
|||||||
@@ -1,4 +1,6 @@
|
|||||||
# flake8: noqa: F401
|
# flake8: noqa: F401
|
||||||
|
import warnings
|
||||||
|
|
||||||
from .common import (
|
from .common import (
|
||||||
HEADRequest,
|
HEADRequest,
|
||||||
PUTRequest,
|
PUTRequest,
|
||||||
@@ -11,3 +13,11 @@ from .common import (
|
|||||||
# isort: split
|
# isort: split
|
||||||
# TODO: all request handlers should be safely imported
|
# TODO: all request handlers should be safely imported
|
||||||
from . import _urllib
|
from . import _urllib
|
||||||
|
from ..utils import bug_reports_message
|
||||||
|
|
||||||
|
try:
|
||||||
|
from . import _requests
|
||||||
|
except ImportError:
|
||||||
|
pass
|
||||||
|
except Exception as e:
|
||||||
|
warnings.warn(f'Failed to import "requests" request handler: {e}' + bug_reports_message())
|
||||||
|
|||||||
@@ -11,7 +11,7 @@ import urllib.request
|
|||||||
|
|
||||||
from .exceptions import RequestError, UnsupportedRequest
|
from .exceptions import RequestError, UnsupportedRequest
|
||||||
from ..dependencies import certifi
|
from ..dependencies import certifi
|
||||||
from ..socks import ProxyType
|
from ..socks import ProxyType, sockssocket
|
||||||
from ..utils import format_field, traverse_obj
|
from ..utils import format_field, traverse_obj
|
||||||
|
|
||||||
if typing.TYPE_CHECKING:
|
if typing.TYPE_CHECKING:
|
||||||
@@ -224,6 +224,24 @@ def _socket_connect(ip_addr, timeout, source_address):
|
|||||||
raise
|
raise
|
||||||
|
|
||||||
|
|
||||||
|
def create_socks_proxy_socket(dest_addr, proxy_args, proxy_ip_addr, timeout, source_address):
|
||||||
|
af, socktype, proto, canonname, sa = proxy_ip_addr
|
||||||
|
sock = sockssocket(af, socktype, proto)
|
||||||
|
try:
|
||||||
|
connect_proxy_args = proxy_args.copy()
|
||||||
|
connect_proxy_args.update({'addr': sa[0], 'port': sa[1]})
|
||||||
|
sock.setproxy(**connect_proxy_args)
|
||||||
|
if timeout is not socket._GLOBAL_DEFAULT_TIMEOUT: # noqa: E721
|
||||||
|
sock.settimeout(timeout)
|
||||||
|
if source_address:
|
||||||
|
sock.bind(source_address)
|
||||||
|
sock.connect(dest_addr)
|
||||||
|
return sock
|
||||||
|
except socket.error:
|
||||||
|
sock.close()
|
||||||
|
raise
|
||||||
|
|
||||||
|
|
||||||
def create_connection(
|
def create_connection(
|
||||||
address,
|
address,
|
||||||
timeout=socket._GLOBAL_DEFAULT_TIMEOUT,
|
timeout=socket._GLOBAL_DEFAULT_TIMEOUT,
|
||||||
|
|||||||
398
yt_dlp/networking/_requests.py
Normal file
398
yt_dlp/networking/_requests.py
Normal file
@@ -0,0 +1,398 @@
|
|||||||
|
import contextlib
|
||||||
|
import functools
|
||||||
|
import http.client
|
||||||
|
import logging
|
||||||
|
import re
|
||||||
|
import socket
|
||||||
|
import warnings
|
||||||
|
|
||||||
|
from ..dependencies import brotli, requests, urllib3
|
||||||
|
from ..utils import bug_reports_message, int_or_none, variadic
|
||||||
|
|
||||||
|
if requests is None:
|
||||||
|
raise ImportError('requests module is not installed')
|
||||||
|
|
||||||
|
if urllib3 is None:
|
||||||
|
raise ImportError('urllib3 module is not installed')
|
||||||
|
|
||||||
|
urllib3_version = tuple(int_or_none(x, default=0) for x in urllib3.__version__.split('.'))
|
||||||
|
|
||||||
|
if urllib3_version < (1, 26, 17):
|
||||||
|
raise ImportError('Only urllib3 >= 1.26.17 is supported')
|
||||||
|
|
||||||
|
if requests.__build__ < 0x023100:
|
||||||
|
raise ImportError('Only requests >= 2.31.0 is supported')
|
||||||
|
|
||||||
|
import requests.adapters
|
||||||
|
import requests.utils
|
||||||
|
import urllib3.connection
|
||||||
|
import urllib3.exceptions
|
||||||
|
|
||||||
|
from ._helper import (
|
||||||
|
InstanceStoreMixin,
|
||||||
|
add_accept_encoding_header,
|
||||||
|
create_connection,
|
||||||
|
create_socks_proxy_socket,
|
||||||
|
get_redirect_method,
|
||||||
|
make_socks_proxy_opts,
|
||||||
|
select_proxy,
|
||||||
|
)
|
||||||
|
from .common import (
|
||||||
|
Features,
|
||||||
|
RequestHandler,
|
||||||
|
Response,
|
||||||
|
register_preference,
|
||||||
|
register_rh,
|
||||||
|
)
|
||||||
|
from .exceptions import (
|
||||||
|
CertificateVerifyError,
|
||||||
|
HTTPError,
|
||||||
|
IncompleteRead,
|
||||||
|
ProxyError,
|
||||||
|
RequestError,
|
||||||
|
SSLError,
|
||||||
|
TransportError,
|
||||||
|
)
|
||||||
|
from ..socks import ProxyError as SocksProxyError
|
||||||
|
|
||||||
|
SUPPORTED_ENCODINGS = [
|
||||||
|
'gzip', 'deflate'
|
||||||
|
]
|
||||||
|
|
||||||
|
if brotli is not None:
|
||||||
|
SUPPORTED_ENCODINGS.append('br')
|
||||||
|
|
||||||
|
"""
|
||||||
|
Override urllib3's behavior to not convert lower-case percent-encoded characters
|
||||||
|
to upper-case during url normalization process.
|
||||||
|
|
||||||
|
RFC3986 defines that the lower or upper case percent-encoded hexidecimal characters are equivalent
|
||||||
|
and normalizers should convert them to uppercase for consistency [1].
|
||||||
|
|
||||||
|
However, some sites may have an incorrect implementation where they provide
|
||||||
|
a percent-encoded url that is then compared case-sensitively.[2]
|
||||||
|
|
||||||
|
While this is a very rare case, since urllib does not do this normalization step, it
|
||||||
|
is best to avoid it in requests too for compatability reasons.
|
||||||
|
|
||||||
|
1: https://tools.ietf.org/html/rfc3986#section-2.1
|
||||||
|
2: https://github.com/streamlink/streamlink/pull/4003
|
||||||
|
"""
|
||||||
|
|
||||||
|
|
||||||
|
class Urllib3PercentREOverride:
|
||||||
|
def __init__(self, r: re.Pattern):
|
||||||
|
self.re = r
|
||||||
|
|
||||||
|
# pass through all other attribute calls to the original re
|
||||||
|
def __getattr__(self, item):
|
||||||
|
return self.re.__getattribute__(item)
|
||||||
|
|
||||||
|
def subn(self, repl, string, *args, **kwargs):
|
||||||
|
return string, self.re.subn(repl, string, *args, **kwargs)[1]
|
||||||
|
|
||||||
|
|
||||||
|
# urllib3 >= 1.25.8 uses subn:
|
||||||
|
# https://github.com/urllib3/urllib3/commit/a2697e7c6b275f05879b60f593c5854a816489f0
|
||||||
|
import urllib3.util.url # noqa: E305
|
||||||
|
|
||||||
|
if hasattr(urllib3.util.url, 'PERCENT_RE'):
|
||||||
|
urllib3.util.url.PERCENT_RE = Urllib3PercentREOverride(urllib3.util.url.PERCENT_RE)
|
||||||
|
elif hasattr(urllib3.util.url, '_PERCENT_RE'): # urllib3 >= 2.0.0
|
||||||
|
urllib3.util.url._PERCENT_RE = Urllib3PercentREOverride(urllib3.util.url._PERCENT_RE)
|
||||||
|
else:
|
||||||
|
warnings.warn('Failed to patch PERCENT_RE in urllib3 (does the attribute exist?)' + bug_reports_message())
|
||||||
|
|
||||||
|
"""
|
||||||
|
Workaround for issue in urllib.util.ssl_.py: ssl_wrap_context does not pass
|
||||||
|
server_hostname to SSLContext.wrap_socket if server_hostname is an IP,
|
||||||
|
however this is an issue because we set check_hostname to True in our SSLContext.
|
||||||
|
|
||||||
|
Monkey-patching IS_SECURETRANSPORT forces ssl_wrap_context to pass server_hostname regardless.
|
||||||
|
|
||||||
|
This has been fixed in urllib3 2.0+.
|
||||||
|
See: https://github.com/urllib3/urllib3/issues/517
|
||||||
|
"""
|
||||||
|
|
||||||
|
if urllib3_version < (2, 0, 0):
|
||||||
|
with contextlib.suppress():
|
||||||
|
urllib3.util.IS_SECURETRANSPORT = urllib3.util.ssl_.IS_SECURETRANSPORT = True
|
||||||
|
|
||||||
|
|
||||||
|
# Requests will not automatically handle no_proxy by default
|
||||||
|
# due to buggy no_proxy handling with proxy dict [1].
|
||||||
|
# 1. https://github.com/psf/requests/issues/5000
|
||||||
|
requests.adapters.select_proxy = select_proxy
|
||||||
|
|
||||||
|
|
||||||
|
class RequestsResponseAdapter(Response):
|
||||||
|
def __init__(self, res: requests.models.Response):
|
||||||
|
super().__init__(
|
||||||
|
fp=res.raw, headers=res.headers, url=res.url,
|
||||||
|
status=res.status_code, reason=res.reason)
|
||||||
|
|
||||||
|
self._requests_response = res
|
||||||
|
|
||||||
|
def read(self, amt: int = None):
|
||||||
|
try:
|
||||||
|
# Interact with urllib3 response directly.
|
||||||
|
return self.fp.read(amt, decode_content=True)
|
||||||
|
|
||||||
|
# See urllib3.response.HTTPResponse.read() for exceptions raised on read
|
||||||
|
except urllib3.exceptions.SSLError as e:
|
||||||
|
raise SSLError(cause=e) from e
|
||||||
|
|
||||||
|
except urllib3.exceptions.ProtocolError as e:
|
||||||
|
# IncompleteRead is always contained within ProtocolError
|
||||||
|
# See urllib3.response.HTTPResponse._error_catcher()
|
||||||
|
ir_err = next(
|
||||||
|
(err for err in (e.__context__, e.__cause__, *variadic(e.args))
|
||||||
|
if isinstance(err, http.client.IncompleteRead)), None)
|
||||||
|
if ir_err is not None:
|
||||||
|
# `urllib3.exceptions.IncompleteRead` is subclass of `http.client.IncompleteRead`
|
||||||
|
# but uses an `int` for its `partial` property.
|
||||||
|
partial = ir_err.partial if isinstance(ir_err.partial, int) else len(ir_err.partial)
|
||||||
|
raise IncompleteRead(partial=partial, expected=ir_err.expected) from e
|
||||||
|
raise TransportError(cause=e) from e
|
||||||
|
|
||||||
|
except urllib3.exceptions.HTTPError as e:
|
||||||
|
# catch-all for any other urllib3 response exceptions
|
||||||
|
raise TransportError(cause=e) from e
|
||||||
|
|
||||||
|
|
||||||
|
class RequestsHTTPAdapter(requests.adapters.HTTPAdapter):
|
||||||
|
def __init__(self, ssl_context=None, proxy_ssl_context=None, source_address=None, **kwargs):
|
||||||
|
self._pm_args = {}
|
||||||
|
if ssl_context:
|
||||||
|
self._pm_args['ssl_context'] = ssl_context
|
||||||
|
if source_address:
|
||||||
|
self._pm_args['source_address'] = (source_address, 0)
|
||||||
|
self._proxy_ssl_context = proxy_ssl_context or ssl_context
|
||||||
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
|
def init_poolmanager(self, *args, **kwargs):
|
||||||
|
return super().init_poolmanager(*args, **kwargs, **self._pm_args)
|
||||||
|
|
||||||
|
def proxy_manager_for(self, proxy, **proxy_kwargs):
|
||||||
|
extra_kwargs = {}
|
||||||
|
if not proxy.lower().startswith('socks') and self._proxy_ssl_context:
|
||||||
|
extra_kwargs['proxy_ssl_context'] = self._proxy_ssl_context
|
||||||
|
return super().proxy_manager_for(proxy, **proxy_kwargs, **self._pm_args, **extra_kwargs)
|
||||||
|
|
||||||
|
def cert_verify(*args, **kwargs):
|
||||||
|
# lean on SSLContext for cert verification
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
class RequestsSession(requests.sessions.Session):
|
||||||
|
"""
|
||||||
|
Ensure unified redirect method handling with our urllib redirect handler.
|
||||||
|
"""
|
||||||
|
def rebuild_method(self, prepared_request, response):
|
||||||
|
new_method = get_redirect_method(prepared_request.method, response.status_code)
|
||||||
|
|
||||||
|
# HACK: requests removes headers/body on redirect unless code was a 307/308.
|
||||||
|
if new_method == prepared_request.method:
|
||||||
|
response._real_status_code = response.status_code
|
||||||
|
response.status_code = 308
|
||||||
|
|
||||||
|
prepared_request.method = new_method
|
||||||
|
|
||||||
|
def rebuild_auth(self, prepared_request, response):
|
||||||
|
# HACK: undo status code change from rebuild_method, if applicable.
|
||||||
|
# rebuild_auth runs after requests would remove headers/body based on status code
|
||||||
|
if hasattr(response, '_real_status_code'):
|
||||||
|
response.status_code = response._real_status_code
|
||||||
|
del response._real_status_code
|
||||||
|
return super().rebuild_auth(prepared_request, response)
|
||||||
|
|
||||||
|
|
||||||
|
class Urllib3LoggingFilter(logging.Filter):
|
||||||
|
|
||||||
|
def filter(self, record):
|
||||||
|
# Ignore HTTP request messages since HTTPConnection prints those
|
||||||
|
if record.msg == '%s://%s:%s "%s %s %s" %s %s':
|
||||||
|
return False
|
||||||
|
return True
|
||||||
|
|
||||||
|
|
||||||
|
class Urllib3LoggingHandler(logging.Handler):
|
||||||
|
"""Redirect urllib3 logs to our logger"""
|
||||||
|
def __init__(self, logger, *args, **kwargs):
|
||||||
|
super().__init__(*args, **kwargs)
|
||||||
|
self._logger = logger
|
||||||
|
|
||||||
|
def emit(self, record):
|
||||||
|
try:
|
||||||
|
msg = self.format(record)
|
||||||
|
if record.levelno >= logging.ERROR:
|
||||||
|
self._logger.error(msg)
|
||||||
|
else:
|
||||||
|
self._logger.stdout(msg)
|
||||||
|
|
||||||
|
except Exception:
|
||||||
|
self.handleError(record)
|
||||||
|
|
||||||
|
|
||||||
|
@register_rh
|
||||||
|
class RequestsRH(RequestHandler, InstanceStoreMixin):
|
||||||
|
|
||||||
|
"""Requests RequestHandler
|
||||||
|
https://github.com/psf/requests
|
||||||
|
"""
|
||||||
|
_SUPPORTED_URL_SCHEMES = ('http', 'https')
|
||||||
|
_SUPPORTED_ENCODINGS = tuple(SUPPORTED_ENCODINGS)
|
||||||
|
_SUPPORTED_PROXY_SCHEMES = ('http', 'https', 'socks4', 'socks4a', 'socks5', 'socks5h')
|
||||||
|
_SUPPORTED_FEATURES = (Features.NO_PROXY, Features.ALL_PROXY)
|
||||||
|
RH_NAME = 'requests'
|
||||||
|
|
||||||
|
def __init__(self, *args, **kwargs):
|
||||||
|
super().__init__(*args, **kwargs)
|
||||||
|
|
||||||
|
# Forward urllib3 debug messages to our logger
|
||||||
|
logger = logging.getLogger('urllib3')
|
||||||
|
handler = Urllib3LoggingHandler(logger=self._logger)
|
||||||
|
handler.setFormatter(logging.Formatter('requests: %(message)s'))
|
||||||
|
handler.addFilter(Urllib3LoggingFilter())
|
||||||
|
logger.addHandler(handler)
|
||||||
|
# TODO: Use a logger filter to suppress pool reuse warning instead
|
||||||
|
logger.setLevel(logging.ERROR)
|
||||||
|
|
||||||
|
if self.verbose:
|
||||||
|
# Setting this globally is not ideal, but is easier than hacking with urllib3.
|
||||||
|
# It could technically be problematic for scripts embedding yt-dlp.
|
||||||
|
# However, it is unlikely debug traffic is used in that context in a way this will cause problems.
|
||||||
|
urllib3.connection.HTTPConnection.debuglevel = 1
|
||||||
|
logger.setLevel(logging.DEBUG)
|
||||||
|
# this is expected if we are using --no-check-certificate
|
||||||
|
urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning)
|
||||||
|
|
||||||
|
def close(self):
|
||||||
|
self._clear_instances()
|
||||||
|
|
||||||
|
def _check_extensions(self, extensions):
|
||||||
|
super()._check_extensions(extensions)
|
||||||
|
extensions.pop('cookiejar', None)
|
||||||
|
extensions.pop('timeout', None)
|
||||||
|
|
||||||
|
def _create_instance(self, cookiejar):
|
||||||
|
session = RequestsSession()
|
||||||
|
http_adapter = RequestsHTTPAdapter(
|
||||||
|
ssl_context=self._make_sslcontext(),
|
||||||
|
source_address=self.source_address,
|
||||||
|
max_retries=urllib3.util.retry.Retry(False),
|
||||||
|
)
|
||||||
|
session.adapters.clear()
|
||||||
|
session.headers = requests.models.CaseInsensitiveDict({'Connection': 'keep-alive'})
|
||||||
|
session.mount('https://', http_adapter)
|
||||||
|
session.mount('http://', http_adapter)
|
||||||
|
session.cookies = cookiejar
|
||||||
|
session.trust_env = False # no need, we already load proxies from env
|
||||||
|
return session
|
||||||
|
|
||||||
|
def _send(self, request):
|
||||||
|
|
||||||
|
headers = self._merge_headers(request.headers)
|
||||||
|
add_accept_encoding_header(headers, SUPPORTED_ENCODINGS)
|
||||||
|
|
||||||
|
max_redirects_exceeded = False
|
||||||
|
|
||||||
|
session = self._get_instance(
|
||||||
|
cookiejar=request.extensions.get('cookiejar') or self.cookiejar)
|
||||||
|
|
||||||
|
try:
|
||||||
|
requests_res = session.request(
|
||||||
|
method=request.method,
|
||||||
|
url=request.url,
|
||||||
|
data=request.data,
|
||||||
|
headers=headers,
|
||||||
|
timeout=float(request.extensions.get('timeout') or self.timeout),
|
||||||
|
proxies=request.proxies or self.proxies,
|
||||||
|
allow_redirects=True,
|
||||||
|
stream=True
|
||||||
|
)
|
||||||
|
|
||||||
|
except requests.exceptions.TooManyRedirects as e:
|
||||||
|
max_redirects_exceeded = True
|
||||||
|
requests_res = e.response
|
||||||
|
|
||||||
|
except requests.exceptions.SSLError as e:
|
||||||
|
if 'CERTIFICATE_VERIFY_FAILED' in str(e):
|
||||||
|
raise CertificateVerifyError(cause=e) from e
|
||||||
|
raise SSLError(cause=e) from e
|
||||||
|
|
||||||
|
except requests.exceptions.ProxyError as e:
|
||||||
|
raise ProxyError(cause=e) from e
|
||||||
|
|
||||||
|
except (requests.exceptions.ConnectionError, requests.exceptions.Timeout) as e:
|
||||||
|
raise TransportError(cause=e) from e
|
||||||
|
|
||||||
|
except urllib3.exceptions.HTTPError as e:
|
||||||
|
# Catch any urllib3 exceptions that may leak through
|
||||||
|
raise TransportError(cause=e) from e
|
||||||
|
|
||||||
|
except requests.exceptions.RequestException as e:
|
||||||
|
# Miscellaneous Requests exceptions. May not necessary be network related e.g. InvalidURL
|
||||||
|
raise RequestError(cause=e) from e
|
||||||
|
|
||||||
|
res = RequestsResponseAdapter(requests_res)
|
||||||
|
|
||||||
|
if not 200 <= res.status < 300:
|
||||||
|
raise HTTPError(res, redirect_loop=max_redirects_exceeded)
|
||||||
|
|
||||||
|
return res
|
||||||
|
|
||||||
|
|
||||||
|
@register_preference(RequestsRH)
|
||||||
|
def requests_preference(rh, request):
|
||||||
|
return 100
|
||||||
|
|
||||||
|
|
||||||
|
# Use our socks proxy implementation with requests to avoid an extra dependency.
|
||||||
|
class SocksHTTPConnection(urllib3.connection.HTTPConnection):
|
||||||
|
def __init__(self, _socks_options, *args, **kwargs): # must use _socks_options to pass PoolKey checks
|
||||||
|
self._proxy_args = _socks_options
|
||||||
|
super().__init__(*args, **kwargs)
|
||||||
|
|
||||||
|
def _new_conn(self):
|
||||||
|
try:
|
||||||
|
return create_connection(
|
||||||
|
address=(self._proxy_args['addr'], self._proxy_args['port']),
|
||||||
|
timeout=self.timeout,
|
||||||
|
source_address=self.source_address,
|
||||||
|
_create_socket_func=functools.partial(
|
||||||
|
create_socks_proxy_socket, (self.host, self.port), self._proxy_args))
|
||||||
|
except (socket.timeout, TimeoutError) as e:
|
||||||
|
raise urllib3.exceptions.ConnectTimeoutError(
|
||||||
|
self, f'Connection to {self.host} timed out. (connect timeout={self.timeout})') from e
|
||||||
|
except SocksProxyError as e:
|
||||||
|
raise urllib3.exceptions.ProxyError(str(e), e) from e
|
||||||
|
except (OSError, socket.error) as e:
|
||||||
|
raise urllib3.exceptions.NewConnectionError(
|
||||||
|
self, f'Failed to establish a new connection: {e}') from e
|
||||||
|
|
||||||
|
|
||||||
|
class SocksHTTPSConnection(SocksHTTPConnection, urllib3.connection.HTTPSConnection):
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
class SocksHTTPConnectionPool(urllib3.HTTPConnectionPool):
|
||||||
|
ConnectionCls = SocksHTTPConnection
|
||||||
|
|
||||||
|
|
||||||
|
class SocksHTTPSConnectionPool(urllib3.HTTPSConnectionPool):
|
||||||
|
ConnectionCls = SocksHTTPSConnection
|
||||||
|
|
||||||
|
|
||||||
|
class SocksProxyManager(urllib3.PoolManager):
|
||||||
|
|
||||||
|
def __init__(self, socks_proxy, username=None, password=None, num_pools=10, headers=None, **connection_pool_kw):
|
||||||
|
connection_pool_kw['_socks_options'] = make_socks_proxy_opts(socks_proxy)
|
||||||
|
super().__init__(num_pools, headers, **connection_pool_kw)
|
||||||
|
self.pool_classes_by_scheme = {
|
||||||
|
'http': SocksHTTPConnectionPool,
|
||||||
|
'https': SocksHTTPSConnectionPool
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
requests.adapters.SOCKSProxyManager = SocksProxyManager
|
||||||
@@ -3,7 +3,6 @@ from __future__ import annotations
|
|||||||
import functools
|
import functools
|
||||||
import http.client
|
import http.client
|
||||||
import io
|
import io
|
||||||
import socket
|
|
||||||
import ssl
|
import ssl
|
||||||
import urllib.error
|
import urllib.error
|
||||||
import urllib.parse
|
import urllib.parse
|
||||||
@@ -24,6 +23,7 @@ from ._helper import (
|
|||||||
InstanceStoreMixin,
|
InstanceStoreMixin,
|
||||||
add_accept_encoding_header,
|
add_accept_encoding_header,
|
||||||
create_connection,
|
create_connection,
|
||||||
|
create_socks_proxy_socket,
|
||||||
get_redirect_method,
|
get_redirect_method,
|
||||||
make_socks_proxy_opts,
|
make_socks_proxy_opts,
|
||||||
select_proxy,
|
select_proxy,
|
||||||
@@ -40,7 +40,6 @@ from .exceptions import (
|
|||||||
)
|
)
|
||||||
from ..dependencies import brotli
|
from ..dependencies import brotli
|
||||||
from ..socks import ProxyError as SocksProxyError
|
from ..socks import ProxyError as SocksProxyError
|
||||||
from ..socks import sockssocket
|
|
||||||
from ..utils import update_url_query
|
from ..utils import update_url_query
|
||||||
from ..utils.networking import normalize_url
|
from ..utils.networking import normalize_url
|
||||||
|
|
||||||
@@ -190,25 +189,12 @@ def make_socks_conn_class(base_class, socks_proxy):
|
|||||||
_create_connection = create_connection
|
_create_connection = create_connection
|
||||||
|
|
||||||
def connect(self):
|
def connect(self):
|
||||||
def sock_socket_connect(ip_addr, timeout, source_address):
|
|
||||||
af, socktype, proto, canonname, sa = ip_addr
|
|
||||||
sock = sockssocket(af, socktype, proto)
|
|
||||||
try:
|
|
||||||
connect_proxy_args = proxy_args.copy()
|
|
||||||
connect_proxy_args.update({'addr': sa[0], 'port': sa[1]})
|
|
||||||
sock.setproxy(**connect_proxy_args)
|
|
||||||
if timeout is not socket._GLOBAL_DEFAULT_TIMEOUT: # noqa: E721
|
|
||||||
sock.settimeout(timeout)
|
|
||||||
if source_address:
|
|
||||||
sock.bind(source_address)
|
|
||||||
sock.connect((self.host, self.port))
|
|
||||||
return sock
|
|
||||||
except socket.error:
|
|
||||||
sock.close()
|
|
||||||
raise
|
|
||||||
self.sock = create_connection(
|
self.sock = create_connection(
|
||||||
(proxy_args['addr'], proxy_args['port']), timeout=self.timeout,
|
(proxy_args['addr'], proxy_args['port']),
|
||||||
source_address=self.source_address, _create_socket_func=sock_socket_connect)
|
timeout=self.timeout,
|
||||||
|
source_address=self.source_address,
|
||||||
|
_create_socket_func=functools.partial(
|
||||||
|
create_socks_proxy_socket, (self.host, self.port), proxy_args))
|
||||||
if isinstance(self, http.client.HTTPSConnection):
|
if isinstance(self, http.client.HTTPSConnection):
|
||||||
self.sock = self._context.wrap_socket(self.sock, server_hostname=self.host)
|
self.sock = self._context.wrap_socket(self.sock, server_hostname=self.host)
|
||||||
|
|
||||||
|
|||||||
@@ -75,7 +75,7 @@ class HTTPError(RequestError):
|
|||||||
|
|
||||||
|
|
||||||
class IncompleteRead(TransportError):
|
class IncompleteRead(TransportError):
|
||||||
def __init__(self, partial: int, expected: int = None, **kwargs):
|
def __init__(self, partial: int, expected: int | None = None, **kwargs):
|
||||||
self.partial = partial
|
self.partial = partial
|
||||||
self.expected = expected
|
self.expected = expected
|
||||||
msg = f'{partial} bytes read'
|
msg = f'{partial} bytes read'
|
||||||
|
|||||||
@@ -471,11 +471,12 @@ def create_parser():
|
|||||||
'no-attach-info-json', 'embed-thumbnail-atomicparsley', 'no-external-downloader-progress',
|
'no-attach-info-json', 'embed-thumbnail-atomicparsley', 'no-external-downloader-progress',
|
||||||
'embed-metadata', 'seperate-video-versions', 'no-clean-infojson', 'no-keep-subs', 'no-certifi',
|
'embed-metadata', 'seperate-video-versions', 'no-clean-infojson', 'no-keep-subs', 'no-certifi',
|
||||||
'no-youtube-channel-redirect', 'no-youtube-unavailable-videos', 'no-youtube-prefer-utc-upload-date',
|
'no-youtube-channel-redirect', 'no-youtube-unavailable-videos', 'no-youtube-prefer-utc-upload-date',
|
||||||
|
'prefer-legacy-http-handler', 'manifest-filesize-approx'
|
||||||
}, 'aliases': {
|
}, 'aliases': {
|
||||||
'youtube-dl': ['all', '-multistreams', '-playlist-match-filter'],
|
'youtube-dl': ['all', '-multistreams', '-playlist-match-filter', '-manifest-filesize-approx'],
|
||||||
'youtube-dlc': ['all', '-no-youtube-channel-redirect', '-no-live-chat', '-playlist-match-filter'],
|
'youtube-dlc': ['all', '-no-youtube-channel-redirect', '-no-live-chat', '-playlist-match-filter', '-manifest-filesize-approx'],
|
||||||
'2021': ['2022', 'no-certifi', 'filename-sanitization', 'no-youtube-prefer-utc-upload-date'],
|
'2021': ['2022', 'no-certifi', 'filename-sanitization', 'no-youtube-prefer-utc-upload-date'],
|
||||||
'2022': ['no-external-downloader-progress', 'playlist-match-filter'],
|
'2022': ['no-external-downloader-progress', 'playlist-match-filter', 'prefer-legacy-http-handler', 'manifest-filesize-approx'],
|
||||||
}
|
}
|
||||||
}, help=(
|
}, help=(
|
||||||
'Options that can help keep compatibility with youtube-dl or youtube-dlc '
|
'Options that can help keep compatibility with youtube-dl or youtube-dlc '
|
||||||
|
|||||||
526
yt_dlp/update.py
526
yt_dlp/update.py
@@ -1,3 +1,5 @@
|
|||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
import atexit
|
import atexit
|
||||||
import contextlib
|
import contextlib
|
||||||
import hashlib
|
import hashlib
|
||||||
@@ -7,6 +9,7 @@ import platform
|
|||||||
import re
|
import re
|
||||||
import subprocess
|
import subprocess
|
||||||
import sys
|
import sys
|
||||||
|
from dataclasses import dataclass
|
||||||
from zipimport import zipimporter
|
from zipimport import zipimporter
|
||||||
|
|
||||||
from .compat import functools # isort: split
|
from .compat import functools # isort: split
|
||||||
@@ -14,24 +17,35 @@ from .compat import compat_realpath, compat_shlex_quote
|
|||||||
from .networking import Request
|
from .networking import Request
|
||||||
from .networking.exceptions import HTTPError, network_exceptions
|
from .networking.exceptions import HTTPError, network_exceptions
|
||||||
from .utils import (
|
from .utils import (
|
||||||
|
NO_DEFAULT,
|
||||||
Popen,
|
Popen,
|
||||||
cached_method,
|
|
||||||
deprecation_warning,
|
deprecation_warning,
|
||||||
|
format_field,
|
||||||
remove_end,
|
remove_end,
|
||||||
remove_start,
|
|
||||||
shell_quote,
|
shell_quote,
|
||||||
system_identifier,
|
system_identifier,
|
||||||
version_tuple,
|
version_tuple,
|
||||||
)
|
)
|
||||||
from .version import CHANNEL, UPDATE_HINT, VARIANT, __version__
|
from .version import (
|
||||||
|
CHANNEL,
|
||||||
|
ORIGIN,
|
||||||
|
RELEASE_GIT_HEAD,
|
||||||
|
UPDATE_HINT,
|
||||||
|
VARIANT,
|
||||||
|
__version__,
|
||||||
|
)
|
||||||
|
|
||||||
UPDATE_SOURCES = {
|
UPDATE_SOURCES = {
|
||||||
'stable': 'yt-dlp/yt-dlp',
|
'stable': 'yt-dlp/yt-dlp',
|
||||||
'nightly': 'yt-dlp/yt-dlp-nightly-builds',
|
'nightly': 'yt-dlp/yt-dlp-nightly-builds',
|
||||||
|
'master': 'yt-dlp/yt-dlp-master-builds',
|
||||||
}
|
}
|
||||||
REPOSITORY = UPDATE_SOURCES['stable']
|
REPOSITORY = UPDATE_SOURCES['stable']
|
||||||
|
_INVERSE_UPDATE_SOURCES = {value: key for key, value in UPDATE_SOURCES.items()}
|
||||||
|
|
||||||
_VERSION_RE = re.compile(r'(\d+\.)*\d+')
|
_VERSION_RE = re.compile(r'(\d+\.)*\d+')
|
||||||
|
_HASH_PATTERN = r'[\da-f]{40}'
|
||||||
|
_COMMIT_RE = re.compile(rf'Generated from: https://(?:[^/?#]+/){{3}}commit/(?P<hash>{_HASH_PATTERN})')
|
||||||
|
|
||||||
API_BASE_URL = 'https://api.github.com/repos'
|
API_BASE_URL = 'https://api.github.com/repos'
|
||||||
|
|
||||||
@@ -112,6 +126,10 @@ def is_non_updateable():
|
|||||||
detect_variant(), _NON_UPDATEABLE_REASONS['unknown' if VARIANT else 'other'])
|
detect_variant(), _NON_UPDATEABLE_REASONS['unknown' if VARIANT else 'other'])
|
||||||
|
|
||||||
|
|
||||||
|
def _get_binary_name():
|
||||||
|
return format_field(_FILE_SUFFIXES, detect_variant(), template='yt-dlp%s', ignore=None, default=None)
|
||||||
|
|
||||||
|
|
||||||
def _get_system_deprecation():
|
def _get_system_deprecation():
|
||||||
MIN_SUPPORTED, MIN_RECOMMENDED = (3, 7), (3, 8)
|
MIN_SUPPORTED, MIN_RECOMMENDED = (3, 7), (3, 8)
|
||||||
|
|
||||||
@@ -146,73 +164,117 @@ def _sha256_file(path):
|
|||||||
return h.hexdigest()
|
return h.hexdigest()
|
||||||
|
|
||||||
|
|
||||||
|
def _make_label(origin, tag, version=None):
|
||||||
|
if '/' in origin:
|
||||||
|
channel = _INVERSE_UPDATE_SOURCES.get(origin, origin)
|
||||||
|
else:
|
||||||
|
channel = origin
|
||||||
|
label = f'{channel}@{tag}'
|
||||||
|
if version and version != tag:
|
||||||
|
label += f' build {version}'
|
||||||
|
if channel != origin:
|
||||||
|
label += f' from {origin}'
|
||||||
|
return label
|
||||||
|
|
||||||
|
|
||||||
|
@dataclass
|
||||||
|
class UpdateInfo:
|
||||||
|
"""
|
||||||
|
Update target information
|
||||||
|
|
||||||
|
Can be created by `query_update()` or manually.
|
||||||
|
|
||||||
|
Attributes:
|
||||||
|
tag The release tag that will be updated to. If from query_update,
|
||||||
|
the value is after API resolution and update spec processing.
|
||||||
|
The only property that is required.
|
||||||
|
version The actual numeric version (if available) of the binary to be updated to,
|
||||||
|
after API resolution and update spec processing. (default: None)
|
||||||
|
requested_version Numeric version of the binary being requested (if available),
|
||||||
|
after API resolution only. (default: None)
|
||||||
|
commit Commit hash (if available) of the binary to be updated to,
|
||||||
|
after API resolution and update spec processing. (default: None)
|
||||||
|
This value will only match the RELEASE_GIT_HEAD of prerelease builds.
|
||||||
|
binary_name Filename of the binary to be updated to. (default: current binary name)
|
||||||
|
checksum Expected checksum (if available) of the binary to be
|
||||||
|
updated to. (default: None)
|
||||||
|
"""
|
||||||
|
tag: str
|
||||||
|
version: str | None = None
|
||||||
|
requested_version: str | None = None
|
||||||
|
commit: str | None = None
|
||||||
|
|
||||||
|
binary_name: str | None = _get_binary_name()
|
||||||
|
checksum: str | None = None
|
||||||
|
|
||||||
|
_has_update = True
|
||||||
|
|
||||||
|
|
||||||
class Updater:
|
class Updater:
|
||||||
_exact = True
|
# XXX: use class variables to simplify testing
|
||||||
|
_channel = CHANNEL
|
||||||
|
_origin = ORIGIN
|
||||||
|
|
||||||
def __init__(self, ydl, target=None):
|
def __init__(self, ydl, target: str | None = None):
|
||||||
self.ydl = ydl
|
self.ydl = ydl
|
||||||
|
# For backwards compat, target needs to be treated as if it could be None
|
||||||
|
self.requested_channel, sep, self.requested_tag = (target or self._channel).rpartition('@')
|
||||||
|
# Check if requested_tag is actually the requested repo/channel
|
||||||
|
if not sep and ('/' in self.requested_tag or self.requested_tag in UPDATE_SOURCES):
|
||||||
|
self.requested_channel = self.requested_tag
|
||||||
|
self.requested_tag: str = None # type: ignore (we set it later)
|
||||||
|
elif not self.requested_channel:
|
||||||
|
# User did not specify a channel, so we are requesting the default channel
|
||||||
|
self.requested_channel = self._channel.partition('@')[0]
|
||||||
|
|
||||||
self.target_channel, sep, self.target_tag = (target or CHANNEL).rpartition('@')
|
# --update should not be treated as an exact tag request even if CHANNEL has a @tag
|
||||||
# stable => stable@latest
|
self._exact = bool(target) and target != self._channel
|
||||||
if not sep and ('/' in self.target_tag or self.target_tag in UPDATE_SOURCES):
|
if not self.requested_tag:
|
||||||
self.target_channel = self.target_tag
|
# User did not specify a tag, so we request 'latest' and track that no exact tag was passed
|
||||||
self.target_tag = None
|
self.requested_tag = 'latest'
|
||||||
elif not self.target_channel:
|
|
||||||
self.target_channel = CHANNEL.partition('@')[0]
|
|
||||||
|
|
||||||
if not self.target_tag:
|
|
||||||
self.target_tag = 'latest'
|
|
||||||
self._exact = False
|
self._exact = False
|
||||||
elif self.target_tag != 'latest':
|
|
||||||
self.target_tag = f'tags/{self.target_tag}'
|
|
||||||
|
|
||||||
if '/' in self.target_channel:
|
if '/' in self.requested_channel:
|
||||||
self._target_repo = self.target_channel
|
# requested_channel is actually a repository
|
||||||
if self.target_channel not in (CHANNEL, *UPDATE_SOURCES.values()):
|
self.requested_repo = self.requested_channel
|
||||||
|
if not self.requested_repo.startswith('yt-dlp/') and self.requested_repo != self._origin:
|
||||||
self.ydl.report_warning(
|
self.ydl.report_warning(
|
||||||
f'You are switching to an {self.ydl._format_err("unofficial", "red")} executable '
|
f'You are switching to an {self.ydl._format_err("unofficial", "red")} executable '
|
||||||
f'from {self.ydl._format_err(self._target_repo, self.ydl.Styles.EMPHASIS)}. '
|
f'from {self.ydl._format_err(self.requested_repo, self.ydl.Styles.EMPHASIS)}. '
|
||||||
f'Run {self.ydl._format_err("at your own risk", "light red")}')
|
f'Run {self.ydl._format_err("at your own risk", "light red")}')
|
||||||
self._block_restart('Automatically restarting into custom builds is disabled for security reasons')
|
self._block_restart('Automatically restarting into custom builds is disabled for security reasons')
|
||||||
else:
|
else:
|
||||||
self._target_repo = UPDATE_SOURCES.get(self.target_channel)
|
# Check if requested_channel resolves to a known repository or else raise
|
||||||
if not self._target_repo:
|
self.requested_repo = UPDATE_SOURCES.get(self.requested_channel)
|
||||||
|
if not self.requested_repo:
|
||||||
self._report_error(
|
self._report_error(
|
||||||
f'Invalid update channel {self.target_channel!r} requested. '
|
f'Invalid update channel {self.requested_channel!r} requested. '
|
||||||
f'Valid channels are {", ".join(UPDATE_SOURCES)}', True)
|
f'Valid channels are {", ".join(UPDATE_SOURCES)}', True)
|
||||||
|
|
||||||
def _version_compare(self, a, b, channel=CHANNEL):
|
self._identifier = f'{detect_variant()} {system_identifier()}'
|
||||||
if self._exact and channel != self.target_channel:
|
|
||||||
return False
|
|
||||||
|
|
||||||
if _VERSION_RE.fullmatch(f'{a}.{b}'):
|
@property
|
||||||
a, b = version_tuple(a), version_tuple(b)
|
def current_version(self):
|
||||||
return a == b if self._exact else a >= b
|
"""Current version"""
|
||||||
return a == b
|
return __version__
|
||||||
|
|
||||||
@functools.cached_property
|
@property
|
||||||
def _tag(self):
|
def current_commit(self):
|
||||||
if self._version_compare(self.current_version, self.latest_version):
|
"""Current commit hash"""
|
||||||
return self.target_tag
|
return RELEASE_GIT_HEAD
|
||||||
|
|
||||||
identifier = f'{detect_variant()} {self.target_channel} {system_identifier()}'
|
def _download_asset(self, name, tag=None):
|
||||||
for line in self._download('_update_spec', 'latest').decode().splitlines():
|
if not tag:
|
||||||
if not line.startswith('lock '):
|
tag = self.requested_tag
|
||||||
continue
|
|
||||||
_, tag, pattern = line.split(' ', 2)
|
|
||||||
if re.match(pattern, identifier):
|
|
||||||
if not self._exact:
|
|
||||||
return f'tags/{tag}'
|
|
||||||
elif self.target_tag == 'latest' or not self._version_compare(
|
|
||||||
tag, self.target_tag[5:], channel=self.target_channel):
|
|
||||||
self._report_error(
|
|
||||||
f'yt-dlp cannot be updated above {tag} since you are on an older Python version', True)
|
|
||||||
return f'tags/{self.current_version}'
|
|
||||||
return self.target_tag
|
|
||||||
|
|
||||||
@cached_method
|
path = 'latest/download' if tag == 'latest' else f'download/{tag}'
|
||||||
def _get_version_info(self, tag):
|
url = f'https://github.com/{self.requested_repo}/releases/{path}/{name}'
|
||||||
url = f'{API_BASE_URL}/{self._target_repo}/releases/{tag}'
|
self.ydl.write_debug(f'Downloading {name} from {url}')
|
||||||
|
return self.ydl.urlopen(url).read()
|
||||||
|
|
||||||
|
def _call_api(self, tag):
|
||||||
|
tag = f'tags/{tag}' if tag != 'latest' else tag
|
||||||
|
url = f'{API_BASE_URL}/{self.requested_repo}/releases/{tag}'
|
||||||
self.ydl.write_debug(f'Fetching release info: {url}')
|
self.ydl.write_debug(f'Fetching release info: {url}')
|
||||||
return json.loads(self.ydl.urlopen(Request(url, headers={
|
return json.loads(self.ydl.urlopen(Request(url, headers={
|
||||||
'Accept': 'application/vnd.github+json',
|
'Accept': 'application/vnd.github+json',
|
||||||
@@ -220,105 +282,175 @@ class Updater:
|
|||||||
'X-GitHub-Api-Version': '2022-11-28',
|
'X-GitHub-Api-Version': '2022-11-28',
|
||||||
})).read().decode())
|
})).read().decode())
|
||||||
|
|
||||||
@property
|
def _get_version_info(self, tag: str) -> tuple[str | None, str | None]:
|
||||||
def current_version(self):
|
if _VERSION_RE.fullmatch(tag):
|
||||||
"""Current version"""
|
return tag, None
|
||||||
return __version__
|
|
||||||
|
|
||||||
@staticmethod
|
api_info = self._call_api(tag)
|
||||||
def _label(channel, tag):
|
|
||||||
"""Label for a given channel and tag"""
|
|
||||||
return f'{channel}@{remove_start(tag, "tags/")}'
|
|
||||||
|
|
||||||
def _get_actual_tag(self, tag):
|
if tag == 'latest':
|
||||||
if tag.startswith('tags/'):
|
requested_version = api_info['tag_name']
|
||||||
return tag[5:]
|
else:
|
||||||
return self._get_version_info(tag)['tag_name']
|
match = re.search(rf'\s+(?P<version>{_VERSION_RE.pattern})$', api_info.get('name', ''))
|
||||||
|
requested_version = match.group('version') if match else None
|
||||||
|
|
||||||
@property
|
if re.fullmatch(_HASH_PATTERN, api_info.get('target_commitish', '')):
|
||||||
def new_version(self):
|
target_commitish = api_info['target_commitish']
|
||||||
"""Version of the latest release we can update to"""
|
else:
|
||||||
return self._get_actual_tag(self._tag)
|
match = _COMMIT_RE.match(api_info.get('body', ''))
|
||||||
|
target_commitish = match.group('hash') if match else None
|
||||||
|
|
||||||
@property
|
if not (requested_version or target_commitish):
|
||||||
def latest_version(self):
|
self._report_error('One of either version or commit hash must be available on the release', expected=True)
|
||||||
"""Version of the target release"""
|
|
||||||
return self._get_actual_tag(self.target_tag)
|
|
||||||
|
|
||||||
@property
|
return requested_version, target_commitish
|
||||||
def has_update(self):
|
|
||||||
"""Whether there is an update available"""
|
|
||||||
return not self._version_compare(self.current_version, self.new_version)
|
|
||||||
|
|
||||||
@functools.cached_property
|
def _download_update_spec(self, source_tags):
|
||||||
def filename(self):
|
for tag in source_tags:
|
||||||
"""Filename of the executable"""
|
try:
|
||||||
return compat_realpath(_get_variant_and_executable_path()[1])
|
return self._download_asset('_update_spec', tag=tag).decode()
|
||||||
|
except network_exceptions as error:
|
||||||
|
if isinstance(error, HTTPError) and error.status == 404:
|
||||||
|
continue
|
||||||
|
self._report_network_error(f'fetch update spec: {error}')
|
||||||
|
|
||||||
def _download(self, name, tag):
|
|
||||||
slug = 'latest/download' if tag == 'latest' else f'download/{tag[5:]}'
|
|
||||||
url = f'https://github.com/{self._target_repo}/releases/{slug}/{name}'
|
|
||||||
self.ydl.write_debug(f'Downloading {name} from {url}')
|
|
||||||
return self.ydl.urlopen(url).read()
|
|
||||||
|
|
||||||
@functools.cached_property
|
|
||||||
def release_name(self):
|
|
||||||
"""The release filename"""
|
|
||||||
return f'yt-dlp{_FILE_SUFFIXES[detect_variant()]}'
|
|
||||||
|
|
||||||
@functools.cached_property
|
|
||||||
def release_hash(self):
|
|
||||||
"""Hash of the latest release"""
|
|
||||||
hash_data = dict(ln.split()[::-1] for ln in self._download('SHA2-256SUMS', self._tag).decode().splitlines())
|
|
||||||
return hash_data[self.release_name]
|
|
||||||
|
|
||||||
def _report_error(self, msg, expected=False):
|
|
||||||
self.ydl.report_error(msg, tb=False if expected else None)
|
|
||||||
self.ydl._download_retcode = 100
|
|
||||||
|
|
||||||
def _report_permission_error(self, file):
|
|
||||||
self._report_error(f'Unable to write to {file}; Try running as administrator', True)
|
|
||||||
|
|
||||||
def _report_network_error(self, action, delim=';'):
|
|
||||||
self._report_error(
|
self._report_error(
|
||||||
f'Unable to {action}{delim} visit '
|
f'The requested tag {self.requested_tag} does not exist for {self.requested_repo}', True)
|
||||||
f'https://github.com/{self._target_repo}/releases/{self.target_tag.replace("tags/", "tag/")}', True)
|
return None
|
||||||
|
|
||||||
|
def _process_update_spec(self, lockfile: str, resolved_tag: str):
|
||||||
|
lines = lockfile.splitlines()
|
||||||
|
is_version2 = any(line.startswith('lockV2 ') for line in lines)
|
||||||
|
|
||||||
|
for line in lines:
|
||||||
|
if is_version2:
|
||||||
|
if not line.startswith(f'lockV2 {self.requested_repo} '):
|
||||||
|
continue
|
||||||
|
_, _, tag, pattern = line.split(' ', 3)
|
||||||
|
else:
|
||||||
|
if not line.startswith('lock '):
|
||||||
|
continue
|
||||||
|
_, tag, pattern = line.split(' ', 2)
|
||||||
|
|
||||||
|
if re.match(pattern, self._identifier):
|
||||||
|
if _VERSION_RE.fullmatch(tag):
|
||||||
|
if not self._exact:
|
||||||
|
return tag
|
||||||
|
elif self._version_compare(tag, resolved_tag):
|
||||||
|
return resolved_tag
|
||||||
|
elif tag != resolved_tag:
|
||||||
|
continue
|
||||||
|
|
||||||
|
self._report_error(
|
||||||
|
f'yt-dlp cannot be updated to {resolved_tag} since you are on an older Python version', True)
|
||||||
|
return None
|
||||||
|
|
||||||
|
return resolved_tag
|
||||||
|
|
||||||
|
def _version_compare(self, a: str, b: str):
|
||||||
|
"""
|
||||||
|
Compare two version strings
|
||||||
|
|
||||||
|
This function SHOULD NOT be called if self._exact == True
|
||||||
|
"""
|
||||||
|
if _VERSION_RE.fullmatch(f'{a}.{b}'):
|
||||||
|
return version_tuple(a) >= version_tuple(b)
|
||||||
|
return a == b
|
||||||
|
|
||||||
|
def query_update(self, *, _output=False) -> UpdateInfo | None:
|
||||||
|
"""Fetches and returns info about the available update"""
|
||||||
|
if not self.requested_repo:
|
||||||
|
self._report_error('No target repository could be determined from input')
|
||||||
|
return None
|
||||||
|
|
||||||
def check_update(self):
|
|
||||||
"""Report whether there is an update available"""
|
|
||||||
if not self._target_repo:
|
|
||||||
return False
|
|
||||||
try:
|
try:
|
||||||
self.ydl.to_screen((
|
requested_version, target_commitish = self._get_version_info(self.requested_tag)
|
||||||
f'Available version: {self._label(self.target_channel, self.latest_version)}, ' if self.target_tag == 'latest' else ''
|
|
||||||
) + f'Current version: {self._label(CHANNEL, self.current_version)}')
|
|
||||||
except network_exceptions as e:
|
except network_exceptions as e:
|
||||||
return self._report_network_error(f'obtain version info ({e})', delim='; Please try again later or')
|
self._report_network_error(f'obtain version info ({e})', delim='; Please try again later or')
|
||||||
|
return None
|
||||||
|
|
||||||
|
if self._exact and self._origin != self.requested_repo:
|
||||||
|
has_update = True
|
||||||
|
elif requested_version:
|
||||||
|
if self._exact:
|
||||||
|
has_update = self.current_version != requested_version
|
||||||
|
else:
|
||||||
|
has_update = not self._version_compare(self.current_version, requested_version)
|
||||||
|
elif target_commitish:
|
||||||
|
has_update = target_commitish != self.current_commit
|
||||||
|
else:
|
||||||
|
has_update = False
|
||||||
|
|
||||||
|
resolved_tag = requested_version if self.requested_tag == 'latest' else self.requested_tag
|
||||||
|
current_label = _make_label(self._origin, self._channel.partition("@")[2] or self.current_version, self.current_version)
|
||||||
|
requested_label = _make_label(self.requested_repo, resolved_tag, requested_version)
|
||||||
|
latest_or_requested = f'{"Latest" if self.requested_tag == "latest" else "Requested"} version: {requested_label}'
|
||||||
|
if not has_update:
|
||||||
|
if _output:
|
||||||
|
self.ydl.to_screen(f'{latest_or_requested}\nyt-dlp is up to date ({current_label})')
|
||||||
|
return None
|
||||||
|
|
||||||
|
update_spec = self._download_update_spec(('latest', None) if requested_version else (None,))
|
||||||
|
if not update_spec:
|
||||||
|
return None
|
||||||
|
# `result_` prefixed vars == post-_process_update_spec() values
|
||||||
|
result_tag = self._process_update_spec(update_spec, resolved_tag)
|
||||||
|
if not result_tag or result_tag == self.current_version:
|
||||||
|
return None
|
||||||
|
elif result_tag == resolved_tag:
|
||||||
|
result_version = requested_version
|
||||||
|
elif _VERSION_RE.fullmatch(result_tag):
|
||||||
|
result_version = result_tag
|
||||||
|
else: # actual version being updated to is unknown
|
||||||
|
result_version = None
|
||||||
|
|
||||||
|
checksum = None
|
||||||
|
# Non-updateable variants can get update_info but need to skip checksum
|
||||||
if not is_non_updateable():
|
if not is_non_updateable():
|
||||||
self.ydl.to_screen(f'Current Build Hash: {_sha256_file(self.filename)}')
|
try:
|
||||||
|
hashes = self._download_asset('SHA2-256SUMS', result_tag)
|
||||||
|
except network_exceptions as error:
|
||||||
|
if not isinstance(error, HTTPError) or error.status != 404:
|
||||||
|
self._report_network_error(f'fetch checksums: {error}')
|
||||||
|
return None
|
||||||
|
self.ydl.report_warning('No hash information found for the release, skipping verification')
|
||||||
|
else:
|
||||||
|
for ln in hashes.decode().splitlines():
|
||||||
|
if ln.endswith(_get_binary_name()):
|
||||||
|
checksum = ln.split()[0]
|
||||||
|
break
|
||||||
|
if not checksum:
|
||||||
|
self.ydl.report_warning('The hash could not be found in the checksum file, skipping verification')
|
||||||
|
|
||||||
if self.has_update:
|
if _output:
|
||||||
return True
|
update_label = _make_label(self.requested_repo, result_tag, result_version)
|
||||||
|
self.ydl.to_screen(
|
||||||
|
f'Current version: {current_label}\n{latest_or_requested}'
|
||||||
|
+ (f'\nUpgradable to: {update_label}' if update_label != requested_label else ''))
|
||||||
|
|
||||||
if self.target_tag == self._tag:
|
return UpdateInfo(
|
||||||
self.ydl.to_screen(f'yt-dlp is up to date ({self._label(CHANNEL, self.current_version)})')
|
tag=result_tag,
|
||||||
elif not self._exact:
|
version=result_version,
|
||||||
self.ydl.report_warning('yt-dlp cannot be updated any further since you are on an older Python version')
|
requested_version=requested_version,
|
||||||
return False
|
commit=target_commitish if result_tag == resolved_tag else None,
|
||||||
|
checksum=checksum)
|
||||||
|
|
||||||
def update(self):
|
def update(self, update_info=NO_DEFAULT):
|
||||||
"""Update yt-dlp executable to the latest version"""
|
"""Update yt-dlp executable to the latest version"""
|
||||||
if not self.check_update():
|
if update_info is NO_DEFAULT:
|
||||||
return
|
update_info = self.query_update(_output=True)
|
||||||
|
if not update_info:
|
||||||
|
return False
|
||||||
|
|
||||||
err = is_non_updateable()
|
err = is_non_updateable()
|
||||||
if err:
|
if err:
|
||||||
return self._report_error(err, True)
|
self._report_error(err, True)
|
||||||
self.ydl.to_screen(f'Updating to {self._label(self.target_channel, self.new_version)} ...')
|
return False
|
||||||
if (_VERSION_RE.fullmatch(self.target_tag[5:])
|
|
||||||
and version_tuple(self.target_tag[5:]) < (2023, 3, 2)):
|
self.ydl.to_screen(f'Current Build Hash: {_sha256_file(self.filename)}')
|
||||||
self.ydl.report_warning('You are downgrading to a version without --update-to')
|
|
||||||
self._block_restart('Cannot automatically restart to a version without --update-to')
|
update_label = _make_label(self.requested_repo, update_info.tag, update_info.version)
|
||||||
|
self.ydl.to_screen(f'Updating to {update_label} ...')
|
||||||
|
|
||||||
directory = os.path.dirname(self.filename)
|
directory = os.path.dirname(self.filename)
|
||||||
if not os.access(self.filename, os.W_OK):
|
if not os.access(self.filename, os.W_OK):
|
||||||
@@ -337,20 +469,17 @@ class Updater:
|
|||||||
return self._report_error('Unable to remove the old version')
|
return self._report_error('Unable to remove the old version')
|
||||||
|
|
||||||
try:
|
try:
|
||||||
newcontent = self._download(self.release_name, self._tag)
|
newcontent = self._download_asset(update_info.binary_name, update_info.tag)
|
||||||
except network_exceptions as e:
|
except network_exceptions as e:
|
||||||
if isinstance(e, HTTPError) and e.status == 404:
|
if isinstance(e, HTTPError) and e.status == 404:
|
||||||
return self._report_error(
|
return self._report_error(
|
||||||
f'The requested tag {self._label(self.target_channel, self.target_tag)} does not exist', True)
|
f'The requested tag {self.requested_repo}@{update_info.tag} does not exist', True)
|
||||||
return self._report_network_error(f'fetch updates: {e}')
|
return self._report_network_error(f'fetch updates: {e}', tag=update_info.tag)
|
||||||
|
|
||||||
try:
|
if not update_info.checksum:
|
||||||
expected_hash = self.release_hash
|
self._block_restart('Automatically restarting into unverified builds is disabled for security reasons')
|
||||||
except Exception:
|
elif hashlib.sha256(newcontent).hexdigest() != update_info.checksum:
|
||||||
self.ydl.report_warning('no hash information found for the release')
|
return self._report_network_error('verify the new executable', tag=update_info.tag)
|
||||||
else:
|
|
||||||
if hashlib.sha256(newcontent).hexdigest() != expected_hash:
|
|
||||||
return self._report_network_error('verify the new executable')
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
with open(new_filename, 'wb') as outf:
|
with open(new_filename, 'wb') as outf:
|
||||||
@@ -387,9 +516,14 @@ class Updater:
|
|||||||
return self._report_error(
|
return self._report_error(
|
||||||
f'Unable to set permissions. Run: sudo chmod a+rx {compat_shlex_quote(self.filename)}')
|
f'Unable to set permissions. Run: sudo chmod a+rx {compat_shlex_quote(self.filename)}')
|
||||||
|
|
||||||
self.ydl.to_screen(f'Updated yt-dlp to {self._label(self.target_channel, self.new_version)}')
|
self.ydl.to_screen(f'Updated yt-dlp to {update_label}')
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
@functools.cached_property
|
||||||
|
def filename(self):
|
||||||
|
"""Filename of the executable"""
|
||||||
|
return compat_realpath(_get_variant_and_executable_path()[1])
|
||||||
|
|
||||||
@functools.cached_property
|
@functools.cached_property
|
||||||
def cmd(self):
|
def cmd(self):
|
||||||
"""The command-line to run the executable, if known"""
|
"""The command-line to run the executable, if known"""
|
||||||
@@ -412,6 +546,71 @@ class Updater:
|
|||||||
return self.ydl._download_retcode
|
return self.ydl._download_retcode
|
||||||
self.restart = wrapper
|
self.restart = wrapper
|
||||||
|
|
||||||
|
def _report_error(self, msg, expected=False):
|
||||||
|
self.ydl.report_error(msg, tb=False if expected else None)
|
||||||
|
self.ydl._download_retcode = 100
|
||||||
|
|
||||||
|
def _report_permission_error(self, file):
|
||||||
|
self._report_error(f'Unable to write to {file}; try running as administrator', True)
|
||||||
|
|
||||||
|
def _report_network_error(self, action, delim=';', tag=None):
|
||||||
|
if not tag:
|
||||||
|
tag = self.requested_tag
|
||||||
|
self._report_error(
|
||||||
|
f'Unable to {action}{delim} visit https://github.com/{self.requested_repo}/releases/'
|
||||||
|
+ tag if tag == "latest" else f"tag/{tag}", True)
|
||||||
|
|
||||||
|
# XXX: Everything below this line in this class is deprecated / for compat only
|
||||||
|
@property
|
||||||
|
def _target_tag(self):
|
||||||
|
"""Deprecated; requested tag with 'tags/' prepended when necessary for API calls"""
|
||||||
|
return f'tags/{self.requested_tag}' if self.requested_tag != 'latest' else self.requested_tag
|
||||||
|
|
||||||
|
def _check_update(self):
|
||||||
|
"""Deprecated; report whether there is an update available"""
|
||||||
|
return bool(self.query_update(_output=True))
|
||||||
|
|
||||||
|
def __getattr__(self, attribute: str):
|
||||||
|
"""Compat getter function for deprecated attributes"""
|
||||||
|
deprecated_props_map = {
|
||||||
|
'check_update': '_check_update',
|
||||||
|
'target_tag': '_target_tag',
|
||||||
|
'target_channel': 'requested_channel',
|
||||||
|
}
|
||||||
|
update_info_props_map = {
|
||||||
|
'has_update': '_has_update',
|
||||||
|
'new_version': 'version',
|
||||||
|
'latest_version': 'requested_version',
|
||||||
|
'release_name': 'binary_name',
|
||||||
|
'release_hash': 'checksum',
|
||||||
|
}
|
||||||
|
|
||||||
|
if attribute not in deprecated_props_map and attribute not in update_info_props_map:
|
||||||
|
raise AttributeError(f'{type(self).__name__!r} object has no attribute {attribute!r}')
|
||||||
|
|
||||||
|
msg = f'{type(self).__name__}.{attribute} is deprecated and will be removed in a future version'
|
||||||
|
if attribute in deprecated_props_map:
|
||||||
|
source_name = deprecated_props_map[attribute]
|
||||||
|
if not source_name.startswith('_'):
|
||||||
|
msg += f'. Please use {source_name!r} instead'
|
||||||
|
source = self
|
||||||
|
mapping = deprecated_props_map
|
||||||
|
|
||||||
|
else: # attribute in update_info_props_map
|
||||||
|
msg += '. Please call query_update() instead'
|
||||||
|
source = self.query_update()
|
||||||
|
if source is None:
|
||||||
|
source = UpdateInfo('', None, None, None)
|
||||||
|
source._has_update = False
|
||||||
|
mapping = update_info_props_map
|
||||||
|
|
||||||
|
deprecation_warning(msg)
|
||||||
|
for target_name, source_name in mapping.items():
|
||||||
|
value = getattr(source, source_name)
|
||||||
|
setattr(self, target_name, value)
|
||||||
|
|
||||||
|
return getattr(self, attribute)
|
||||||
|
|
||||||
|
|
||||||
def run_update(ydl):
|
def run_update(ydl):
|
||||||
"""Update the program file with the latest version from the repository
|
"""Update the program file with the latest version from the repository
|
||||||
@@ -420,45 +619,4 @@ def run_update(ydl):
|
|||||||
return Updater(ydl).update()
|
return Updater(ydl).update()
|
||||||
|
|
||||||
|
|
||||||
# Deprecated
|
|
||||||
def update_self(to_screen, verbose, opener):
|
|
||||||
import traceback
|
|
||||||
|
|
||||||
deprecation_warning(f'"{__name__}.update_self" is deprecated and may be removed '
|
|
||||||
f'in a future version. Use "{__name__}.run_update(ydl)" instead')
|
|
||||||
|
|
||||||
printfn = to_screen
|
|
||||||
|
|
||||||
class FakeYDL():
|
|
||||||
to_screen = printfn
|
|
||||||
|
|
||||||
def report_warning(self, msg, *args, **kwargs):
|
|
||||||
return printfn(f'WARNING: {msg}', *args, **kwargs)
|
|
||||||
|
|
||||||
def report_error(self, msg, tb=None):
|
|
||||||
printfn(f'ERROR: {msg}')
|
|
||||||
if not verbose:
|
|
||||||
return
|
|
||||||
if tb is None:
|
|
||||||
# Copied from YoutubeDL.trouble
|
|
||||||
if sys.exc_info()[0]:
|
|
||||||
tb = ''
|
|
||||||
if hasattr(sys.exc_info()[1], 'exc_info') and sys.exc_info()[1].exc_info[0]:
|
|
||||||
tb += ''.join(traceback.format_exception(*sys.exc_info()[1].exc_info))
|
|
||||||
tb += traceback.format_exc()
|
|
||||||
else:
|
|
||||||
tb_data = traceback.format_list(traceback.extract_stack())
|
|
||||||
tb = ''.join(tb_data)
|
|
||||||
if tb:
|
|
||||||
printfn(tb)
|
|
||||||
|
|
||||||
def write_debug(self, msg, *args, **kwargs):
|
|
||||||
printfn(f'[debug] {msg}', *args, **kwargs)
|
|
||||||
|
|
||||||
def urlopen(self, url):
|
|
||||||
return opener.open(url)
|
|
||||||
|
|
||||||
return run_update(FakeYDL())
|
|
||||||
|
|
||||||
|
|
||||||
__all__ = ['Updater']
|
__all__ = ['Updater']
|
||||||
|
|||||||
@@ -123,6 +123,7 @@ def clean_headers(headers: HTTPHeaderDict):
|
|||||||
if 'Youtubedl-No-Compression' in headers: # compat
|
if 'Youtubedl-No-Compression' in headers: # compat
|
||||||
del headers['Youtubedl-No-Compression']
|
del headers['Youtubedl-No-Compression']
|
||||||
headers['Accept-Encoding'] = 'identity'
|
headers['Accept-Encoding'] = 'identity'
|
||||||
|
headers.pop('Ytdl-socks-proxy', None)
|
||||||
|
|
||||||
|
|
||||||
def remove_dot_segments(path):
|
def remove_dot_segments(path):
|
||||||
|
|||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user