mirror of
				https://github.com/yt-dlp/yt-dlp.git
				synced 2025-10-31 06:35:12 +00:00 
			
		
		
		
	Fix all PEP8 issues except E501
This commit is contained in:
		| @@ -22,7 +22,7 @@ def main(): | |||||||
|             continue |             continue | ||||||
|         elif ie_desc is not None: |         elif ie_desc is not None: | ||||||
|             ie_html += ': {}'.format(ie.IE_DESC) |             ie_html += ': {}'.format(ie.IE_DESC) | ||||||
|         if ie.working() == False: |         if not ie.working(): | ||||||
|             ie_html += ' (Currently broken)' |             ie_html += ' (Currently broken)' | ||||||
|         ie_htmls.append('<li>{}</li>'.format(ie_html)) |         ie_htmls.append('<li>{}</li>'.format(ie_html)) | ||||||
|  |  | ||||||
|   | |||||||
| @@ -189,7 +189,7 @@ def _real_main(argv=None): | |||||||
|  |  | ||||||
|     # --all-sub automatically sets --write-sub if --write-auto-sub is not given |     # --all-sub automatically sets --write-sub if --write-auto-sub is not given | ||||||
|     # this was the old behaviour if only --all-sub was given. |     # this was the old behaviour if only --all-sub was given. | ||||||
|     if opts.allsubtitles and (opts.writeautomaticsub == False): |     if opts.allsubtitles and not opts.writeautomaticsub: | ||||||
|         opts.writesubtitles = True |         opts.writesubtitles = True | ||||||
|  |  | ||||||
|     if sys.version_info < (3,): |     if sys.version_info < (3,): | ||||||
|   | |||||||
| @@ -87,7 +87,7 @@ def key_expansion(data): | |||||||
|             temp = sub_bytes(temp) |             temp = sub_bytes(temp) | ||||||
|             data += xor(temp, data[-key_size_bytes: 4 - key_size_bytes]) |             data += xor(temp, data[-key_size_bytes: 4 - key_size_bytes]) | ||||||
|  |  | ||||||
|         for _ in range(3 if key_size_bytes == 32  else 2 if key_size_bytes == 24 else 0): |         for _ in range(3 if key_size_bytes == 32 else 2 if key_size_bytes == 24 else 0): | ||||||
|             temp = data[-4:] |             temp = data[-4:] | ||||||
|             data += xor(temp, data[-key_size_bytes: 4 - key_size_bytes]) |             data += xor(temp, data[-key_size_bytes: 4 - key_size_bytes]) | ||||||
|     data = data[:expanded_key_size_bytes] |     data = data[:expanded_key_size_bytes] | ||||||
|   | |||||||
| @@ -73,10 +73,11 @@ class BambuserChannelIE(InfoExtractor): | |||||||
|         urls = [] |         urls = [] | ||||||
|         last_id = '' |         last_id = '' | ||||||
|         for i in itertools.count(1): |         for i in itertools.count(1): | ||||||
|             req_url = ('http://bambuser.com/xhr-api/index.php?username={user}' |             req_url = ( | ||||||
|                        '&sort=created&access_mode=0%2C1%2C2&limit={count}' |                 'http://bambuser.com/xhr-api/index.php?username={user}' | ||||||
|                        '&method=broadcast&format=json&vid_older_than={last}' |                 '&sort=created&access_mode=0%2C1%2C2&limit={count}' | ||||||
|                 ).format(user=user, count=self._STEP, last=last_id) |                 '&method=broadcast&format=json&vid_older_than={last}' | ||||||
|  |             ).format(user=user, count=self._STEP, last=last_id) | ||||||
|             req = compat_urllib_request.Request(req_url) |             req = compat_urllib_request.Request(req_url) | ||||||
|             # Without setting this header, we wouldn't get any result |             # Without setting this header, we wouldn't get any result | ||||||
|             req.add_header('Referer', 'http://bambuser.com/channel/%s' % user) |             req.add_header('Referer', 'http://bambuser.com/channel/%s' % user) | ||||||
|   | |||||||
| @@ -188,16 +188,17 @@ class Channel9IE(InfoExtractor): | |||||||
|         view_count = self._extract_view_count(html) |         view_count = self._extract_view_count(html) | ||||||
|         comment_count = self._extract_comment_count(html) |         comment_count = self._extract_comment_count(html) | ||||||
|  |  | ||||||
|         common = {'_type': 'video', |         common = { | ||||||
|                   'id': content_path, |             '_type': 'video', | ||||||
|                   'description': description, |             'id': content_path, | ||||||
|                   'thumbnail': thumbnail, |             'description': description, | ||||||
|                   'duration': duration, |             'thumbnail': thumbnail, | ||||||
|                   'avg_rating': avg_rating, |             'duration': duration, | ||||||
|                   'rating_count': rating_count, |             'avg_rating': avg_rating, | ||||||
|                   'view_count': view_count, |             'rating_count': rating_count, | ||||||
|                   'comment_count': comment_count, |             'view_count': view_count, | ||||||
|                 } |             'comment_count': comment_count, | ||||||
|  |         } | ||||||
|  |  | ||||||
|         result = [] |         result = [] | ||||||
|  |  | ||||||
|   | |||||||
| @@ -248,8 +248,9 @@ Format: Layer, Start, End, Style, Name, MarginL, MarginR, MarginV, Effect, Text | |||||||
|         subtitles = {} |         subtitles = {} | ||||||
|         sub_format = self._downloader.params.get('subtitlesformat', 'srt') |         sub_format = self._downloader.params.get('subtitlesformat', 'srt') | ||||||
|         for sub_id, sub_name in re.findall(r'\?ssid=([0-9]+)" title="([^"]+)', webpage): |         for sub_id, sub_name in re.findall(r'\?ssid=([0-9]+)" title="([^"]+)', webpage): | ||||||
|             sub_page = self._download_webpage('http://www.crunchyroll.com/xml/?req=RpcApiSubtitle_GetXml&subtitle_script_id=' + sub_id,\ |             sub_page = self._download_webpage( | ||||||
|                                               video_id, note='Downloading subtitles for ' + sub_name) |                 'http://www.crunchyroll.com/xml/?req=RpcApiSubtitle_GetXml&subtitle_script_id=' + sub_id, | ||||||
|  |                 video_id, note='Downloading subtitles for ' + sub_name) | ||||||
|             id = self._search_regex(r'id=\'([0-9]+)', sub_page, 'subtitle_id', fatal=False) |             id = self._search_regex(r'id=\'([0-9]+)', sub_page, 'subtitle_id', fatal=False) | ||||||
|             iv = self._search_regex(r'<iv>([^<]+)', sub_page, 'subtitle_iv', fatal=False) |             iv = self._search_regex(r'<iv>([^<]+)', sub_page, 'subtitle_iv', fatal=False) | ||||||
|             data = self._search_regex(r'<data>([^<]+)', sub_page, 'subtitle_data', fatal=False) |             data = self._search_regex(r'<data>([^<]+)', sub_page, 'subtitle_data', fatal=False) | ||||||
|   | |||||||
| @@ -77,7 +77,7 @@ class FacebookIE(InfoExtractor): | |||||||
|             'legacy_return': '1', |             'legacy_return': '1', | ||||||
|             'timezone': '-60', |             'timezone': '-60', | ||||||
|             'trynum': '1', |             'trynum': '1', | ||||||
|             } |         } | ||||||
|         request = compat_urllib_request.Request(self._LOGIN_URL, urlencode_postdata(login_form)) |         request = compat_urllib_request.Request(self._LOGIN_URL, urlencode_postdata(login_form)) | ||||||
|         request.add_header('Content-Type', 'application/x-www-form-urlencoded') |         request.add_header('Content-Type', 'application/x-www-form-urlencoded') | ||||||
|         try: |         try: | ||||||
|   | |||||||
| @@ -65,9 +65,9 @@ class FourTubeIE(InfoExtractor): | |||||||
|  |  | ||||||
|         token_url = "http://tkn.4tube.com/{0}/desktop/{1}".format(media_id, "+".join(sources)) |         token_url = "http://tkn.4tube.com/{0}/desktop/{1}".format(media_id, "+".join(sources)) | ||||||
|         headers = { |         headers = { | ||||||
|                 b'Content-Type': b'application/x-www-form-urlencoded', |             b'Content-Type': b'application/x-www-form-urlencoded', | ||||||
|                 b'Origin': b'http://www.4tube.com', |             b'Origin': b'http://www.4tube.com', | ||||||
|                 } |         } | ||||||
|         token_req = compat_urllib_request.Request(token_url, b'{}', headers) |         token_req = compat_urllib_request.Request(token_url, b'{}', headers) | ||||||
|         tokens = self._download_json(token_req, video_id) |         tokens = self._download_json(token_req, video_id) | ||||||
|  |  | ||||||
| @@ -76,7 +76,7 @@ class FourTubeIE(InfoExtractor): | |||||||
|             'format_id': format + 'p', |             'format_id': format + 'p', | ||||||
|             'resolution': format + 'p', |             'resolution': format + 'p', | ||||||
|             'quality': int(format), |             'quality': int(format), | ||||||
|             } for format in sources] |         } for format in sources] | ||||||
|  |  | ||||||
|         self._sort_formats(formats) |         self._sort_formats(formats) | ||||||
|  |  | ||||||
|   | |||||||
| @@ -537,9 +537,9 @@ class GenericIE(InfoExtractor): | |||||||
|  |  | ||||||
|             if default_search in ('error', 'fixup_error'): |             if default_search in ('error', 'fixup_error'): | ||||||
|                 raise ExtractorError( |                 raise ExtractorError( | ||||||
|                     ('%r is not a valid URL. ' |                     '%r is not a valid URL. ' | ||||||
|                      'Set --default-search "ytsearch" (or run  youtube-dl "ytsearch:%s" ) to search YouTube' |                     'Set --default-search "ytsearch" (or run  youtube-dl "ytsearch:%s" ) to search YouTube' | ||||||
|                     ) % (url, url), expected=True) |                     % (url, url), expected=True) | ||||||
|             else: |             else: | ||||||
|                 if ':' not in default_search: |                 if ':' not in default_search: | ||||||
|                     default_search += ':' |                     default_search += ':' | ||||||
|   | |||||||
| @@ -63,8 +63,10 @@ class IGNIE(InfoExtractor): | |||||||
|                 'id': '078fdd005f6d3c02f63d795faa1b984f', |                 'id': '078fdd005f6d3c02f63d795faa1b984f', | ||||||
|                 'ext': 'mp4', |                 'ext': 'mp4', | ||||||
|                 'title': 'Rewind Theater - Wild Trailer Gamescom 2014', |                 'title': 'Rewind Theater - Wild Trailer Gamescom 2014', | ||||||
|                 'description': 'Giant skeletons, bloody hunts, and captivating' |                 'description': ( | ||||||
|                     ' natural beauty take our breath away.', |                     'Giant skeletons, bloody hunts, and captivating' | ||||||
|  |                     ' natural beauty take our breath away.' | ||||||
|  |                 ), | ||||||
|             }, |             }, | ||||||
|         }, |         }, | ||||||
|     ] |     ] | ||||||
|   | |||||||
| @@ -43,7 +43,7 @@ class IviIE(InfoExtractor): | |||||||
|                 'thumbnail': 'http://thumbs.ivi.ru/f15.vcp.digitalaccess.ru/contents/8/4/0068dc0677041f3336b7c2baad8fc0.jpg', |                 'thumbnail': 'http://thumbs.ivi.ru/f15.vcp.digitalaccess.ru/contents/8/4/0068dc0677041f3336b7c2baad8fc0.jpg', | ||||||
|             }, |             }, | ||||||
|             'skip': 'Only works from Russia', |             'skip': 'Only works from Russia', | ||||||
|          } |         } | ||||||
|     ] |     ] | ||||||
|  |  | ||||||
|     # Sorted by quality |     # Sorted by quality | ||||||
|   | |||||||
| @@ -57,9 +57,9 @@ class MyVideoIE(InfoExtractor): | |||||||
|         video_id = mobj.group('id') |         video_id = mobj.group('id') | ||||||
|  |  | ||||||
|         GK = ( |         GK = ( | ||||||
|           b'WXpnME1EZGhNRGhpTTJNM01XVmhOREU0WldNNVpHTTJOakpt' |             b'WXpnME1EZGhNRGhpTTJNM01XVmhOREU0WldNNVpHTTJOakpt' | ||||||
|           b'TW1FMU5tVTBNR05pWkRaa05XRXhNVFJoWVRVd1ptSXhaVEV3' |             b'TW1FMU5tVTBNR05pWkRaa05XRXhNVFJoWVRVd1ptSXhaVEV3' | ||||||
|           b'TnpsbA0KTVRkbU1tSTRNdz09' |             b'TnpsbA0KTVRkbU1tSTRNdz09' | ||||||
|         ) |         ) | ||||||
|  |  | ||||||
|         # Get video webpage |         # Get video webpage | ||||||
|   | |||||||
| @@ -371,7 +371,7 @@ class SoundcloudPlaylistIE(SoundcloudIE): | |||||||
|  |  | ||||||
|         entries = [ |         entries = [ | ||||||
|             self._extract_info_dict(t, quiet=True, secret_token=token) |             self._extract_info_dict(t, quiet=True, secret_token=token) | ||||||
|                 for t in data['tracks']] |             for t in data['tracks']] | ||||||
|  |  | ||||||
|         return { |         return { | ||||||
|             '_type': 'playlist', |             '_type': 'playlist', | ||||||
|   | |||||||
| @@ -36,9 +36,10 @@ class TlcDeIE(InfoExtractor): | |||||||
|             'ext': 'mp4', |             'ext': 'mp4', | ||||||
|             'title': 'Breaking Amish: Die Welt da draußen', |             'title': 'Breaking Amish: Die Welt da draußen', | ||||||
|             'uploader': 'Discovery Networks - Germany', |             'uploader': 'Discovery Networks - Germany', | ||||||
|             'description': 'Vier Amische und eine Mennonitin wagen in New York' |             'description': ( | ||||||
|  |                 'Vier Amische und eine Mennonitin wagen in New York' | ||||||
|                 '  den Sprung in ein komplett anderes Leben. Begleitet sie auf' |                 '  den Sprung in ein komplett anderes Leben. Begleitet sie auf' | ||||||
|                 ' ihrem spannenden Weg.', |                 ' ihrem spannenden Weg.'), | ||||||
|         }, |         }, | ||||||
|     } |     } | ||||||
|  |  | ||||||
|   | |||||||
| @@ -36,7 +36,7 @@ class TudouIE(InfoExtractor): | |||||||
|         'skip': 'Only works from China' |         'skip': 'Only works from China' | ||||||
|     }] |     }] | ||||||
|  |  | ||||||
|     def _url_for_id(self, id, quality = None): |     def _url_for_id(self, id, quality=None): | ||||||
|         info_url = "http://v2.tudou.com/f?id=" + str(id) |         info_url = "http://v2.tudou.com/f?id=" + str(id) | ||||||
|         if quality: |         if quality: | ||||||
|             info_url += '&hd' + quality |             info_url += '&hd' + quality | ||||||
|   | |||||||
| @@ -262,7 +262,8 @@ def parseOpts(overrideArguments=None): | |||||||
|     video_format.add_option( |     video_format.add_option( | ||||||
|         '-f', '--format', |         '-f', '--format', | ||||||
|         action='store', dest='format', metavar='FORMAT', default=None, |         action='store', dest='format', metavar='FORMAT', default=None, | ||||||
|         help='video format code, specify the order of preference using' |         help=( | ||||||
|  |             'video format code, specify the order of preference using' | ||||||
|             ' slashes: -f 22/17/18 .  -f mp4 , -f m4a and  -f flv  are also' |             ' slashes: -f 22/17/18 .  -f mp4 , -f m4a and  -f flv  are also' | ||||||
|             ' supported. You can also use the special names "best",' |             ' supported. You can also use the special names "best",' | ||||||
|             ' "bestvideo", "bestaudio", "worst", "worstvideo" and' |             ' "bestvideo", "bestaudio", "worst", "worstvideo" and' | ||||||
| @@ -271,7 +272,7 @@ def parseOpts(overrideArguments=None): | |||||||
|             ' -f  136/137/mp4/bestvideo,140/m4a/bestaudio.' |             ' -f  136/137/mp4/bestvideo,140/m4a/bestaudio.' | ||||||
|             ' You can merge the video and audio of two formats into a single' |             ' You can merge the video and audio of two formats into a single' | ||||||
|             ' file using -f <video-format>+<audio-format> (requires ffmpeg or' |             ' file using -f <video-format>+<audio-format> (requires ffmpeg or' | ||||||
|             ' avconv), for example -f bestvideo+bestaudio.') |             ' avconv), for example -f bestvideo+bestaudio.')) | ||||||
|     video_format.add_option( |     video_format.add_option( | ||||||
|         '--all-formats', |         '--all-formats', | ||||||
|         action='store_const', dest='format', const='all', |         action='store_const', dest='format', const='all', | ||||||
|   | |||||||
| @@ -240,9 +240,9 @@ def sanitize_open(filename, open_mode): | |||||||
|  |  | ||||||
|         # In case of error, try to remove win32 forbidden chars |         # In case of error, try to remove win32 forbidden chars | ||||||
|         alt_filename = os.path.join( |         alt_filename = os.path.join( | ||||||
|                         re.sub('[/<>:"\\|\\\\?\\*]', '#', path_part) |             re.sub('[/<>:"\\|\\\\?\\*]', '#', path_part) | ||||||
|                         for path_part in os.path.split(filename) |             for path_part in os.path.split(filename) | ||||||
|                        ) |         ) | ||||||
|         if alt_filename == filename: |         if alt_filename == filename: | ||||||
|             raise |             raise | ||||||
|         else: |         else: | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user
	 Philipp Hagemeister
					Philipp Hagemeister