|
|
|
@ -1,32 +1,32 @@
|
|
|
|
|
# -*- coding: utf-8 -*-
|
|
|
|
|
from .common import InfoExtractor
|
|
|
|
|
from ..utils import (
|
|
|
|
|
ExtractorError, # Import ExtractorError for raising specific errors
|
|
|
|
|
int_or_none,
|
|
|
|
|
orderedSet,
|
|
|
|
|
parse_duration,
|
|
|
|
|
parse_iso8601,
|
|
|
|
|
parse_qs,
|
|
|
|
|
qualities,
|
|
|
|
|
traverse_obj,
|
|
|
|
|
traverse_obj, # Useful for safely navigating nested dictionaries
|
|
|
|
|
unified_strdate,
|
|
|
|
|
xpath_text,
|
|
|
|
|
)
|
|
|
|
|
import re # Import re for findall
|
|
|
|
|
|
|
|
|
|
# --- EuropaIE (Older extractor - unchanged) ---
|
|
|
|
|
# This extractor handles older ec.europa.eu/avservices URLs and is likely defunct.
|
|
|
|
|
class EuropaIE(InfoExtractor):
|
|
|
|
|
_WORKING = False
|
|
|
|
|
_WORKING = False # Marked as not working
|
|
|
|
|
_VALID_URL = r'https?://ec\.europa\.eu/avservices/(?:video/player|audio/audioDetails)\.cfm\?.*?\bref=(?P<id>[A-Za-z0-9-]+)'
|
|
|
|
|
_TESTS = [{
|
|
|
|
|
'url': 'http://ec.europa.eu/avservices/video/player.cfm?ref=I107758',
|
|
|
|
|
'md5': '574f080699ddd1e19a675b0ddf010371',
|
|
|
|
|
'info_dict': {
|
|
|
|
|
'id': 'I107758',
|
|
|
|
|
'ext': 'mp4',
|
|
|
|
|
'title': 'TRADE - Wikileaks on TTIP',
|
|
|
|
|
'id': 'I107758', 'ext': 'mp4', 'title': 'TRADE - Wikileaks on TTIP',
|
|
|
|
|
'description': 'NEW LIVE EC Midday press briefing of 11/08/2015',
|
|
|
|
|
'thumbnail': r're:^https?://.*\.jpg$',
|
|
|
|
|
'upload_date': '20150811',
|
|
|
|
|
'duration': 34,
|
|
|
|
|
'view_count': int,
|
|
|
|
|
'formats': 'mincount:3',
|
|
|
|
|
'thumbnail': r're:^https?://.*\.jpg$', 'upload_date': '20150811',
|
|
|
|
|
'duration': 34, 'view_count': int, 'formats': 'mincount:3',
|
|
|
|
|
},
|
|
|
|
|
}, {
|
|
|
|
|
'url': 'http://ec.europa.eu/avservices/video/player.cfm?sitelang=en&ref=I107786',
|
|
|
|
@ -37,189 +37,150 @@ class EuropaIE(InfoExtractor):
|
|
|
|
|
}]
|
|
|
|
|
|
|
|
|
|
def _real_extract(self, url):
|
|
|
|
|
# (Implementation remains the same as previous versions)
|
|
|
|
|
video_id = self._match_id(url)
|
|
|
|
|
|
|
|
|
|
playlist = self._download_xml(
|
|
|
|
|
f'http://ec.europa.eu/avservices/video/player/playlist.cfm?ID={video_id}', video_id)
|
|
|
|
|
|
|
|
|
|
def get_item(type_, preference):
|
|
|
|
|
items = {}
|
|
|
|
|
for item in playlist.findall(f'./info/{type_}/item'):
|
|
|
|
|
lang, label = (
|
|
|
|
|
xpath_text(item, 'lg', default=None),
|
|
|
|
|
xpath_text(item, 'label', default=None)
|
|
|
|
|
)
|
|
|
|
|
if lang and label:
|
|
|
|
|
items[lang] = label.strip()
|
|
|
|
|
lang, label = (xpath_text(item, 'lg', default=None), xpath_text(item, 'label', default=None))
|
|
|
|
|
if lang and label: items[lang] = label.strip()
|
|
|
|
|
for p in preference:
|
|
|
|
|
if items.get(p):
|
|
|
|
|
return items[p]
|
|
|
|
|
|
|
|
|
|
if items.get(p): return items[p]
|
|
|
|
|
query = parse_qs(url)
|
|
|
|
|
preferred_lang = query.get('sitelang', ('en', ))[0]
|
|
|
|
|
preferred_langs = orderedSet((preferred_lang, 'en', 'int'))
|
|
|
|
|
|
|
|
|
|
title = get_item('title', preferred_langs) or video_id
|
|
|
|
|
description = get_item('description', preferred_langs)
|
|
|
|
|
thumbnail = xpath_text(playlist, './info/thumburl', 'thumbnail')
|
|
|
|
|
upload_date = unified_strdate(xpath_text(playlist, './info/date', 'upload date'))
|
|
|
|
|
duration = parse_duration(xpath_text(playlist, './info/duration', 'duration'))
|
|
|
|
|
view_count = int_or_none(xpath_text(playlist, './info/views', 'views'))
|
|
|
|
|
|
|
|
|
|
language_preference = qualities(preferred_langs[::-1])
|
|
|
|
|
|
|
|
|
|
formats = []
|
|
|
|
|
for file_ in playlist.findall('./files/file'):
|
|
|
|
|
video_url = xpath_text(file_, './url')
|
|
|
|
|
if not video_url:
|
|
|
|
|
continue
|
|
|
|
|
if not video_url: continue
|
|
|
|
|
lang = xpath_text(file_, './lg')
|
|
|
|
|
formats.append({
|
|
|
|
|
'url': video_url,
|
|
|
|
|
'format_id': lang,
|
|
|
|
|
'format_note': xpath_text(file_, './lglabel'),
|
|
|
|
|
'language_preference': language_preference(lang),
|
|
|
|
|
})
|
|
|
|
|
|
|
|
|
|
return {
|
|
|
|
|
'id': video_id,
|
|
|
|
|
'title': title,
|
|
|
|
|
'description': description,
|
|
|
|
|
'thumbnail': thumbnail,
|
|
|
|
|
'upload_date': upload_date,
|
|
|
|
|
'duration': duration,
|
|
|
|
|
'view_count': view_count,
|
|
|
|
|
'formats': formats,
|
|
|
|
|
}
|
|
|
|
|
formats.append({'url': video_url, 'format_id': lang, 'format_note': xpath_text(file_, './lglabel'), 'language_preference': language_preference(lang)})
|
|
|
|
|
return {'id': video_id, 'title': title, 'description': description, 'thumbnail': thumbnail, 'upload_date': upload_date, 'duration': duration, 'view_count': view_count, 'formats': formats}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
# --- EuroParlWebstreamIE (Modified extractor to handle potential site changes) ---
|
|
|
|
|
class EuroParlWebstreamIE(InfoExtractor):
|
|
|
|
|
_VALID_URL = r'''(?x)
|
|
|
|
|
https?://multimedia\.europarl\.europa\.eu/
|
|
|
|
|
(?:\w+/)?webstreaming/(?:[\w-]+_)?(?P<id>[\w-]+)
|
|
|
|
|
(?:\w+/)?webstreaming/(?:[\w-]+_)?(?P<id>[\w-]+) # Matches /en/webstreaming/event_id format
|
|
|
|
|
'''
|
|
|
|
|
_TESTS = [{
|
|
|
|
|
# Existing VOD test
|
|
|
|
|
'url': 'https://multimedia.europarl.europa.eu/pl/webstreaming/plenary-session_20220914-0900-PLENARY',
|
|
|
|
|
'info_dict': {
|
|
|
|
|
'id': '62388b15-d85b-4add-99aa-ba12ccf64f0d',
|
|
|
|
|
'display_id': '20220914-0900-PLENARY',
|
|
|
|
|
'ext': 'mp4',
|
|
|
|
|
'title': 'Plenary session',
|
|
|
|
|
'release_timestamp': 1663139069,
|
|
|
|
|
'release_date': '20220914',
|
|
|
|
|
},
|
|
|
|
|
'params': {
|
|
|
|
|
'skip_download': True,
|
|
|
|
|
'id': '62388b15-d85b-4add-99aa-ba12ccf64f0d', 'display_id': '20220914-0900-PLENARY',
|
|
|
|
|
'ext': 'mp4', 'title': 'Plenary session', 'release_timestamp': 1663139069, 'release_date': '20220914',
|
|
|
|
|
},
|
|
|
|
|
'params': {'skip_download': True},
|
|
|
|
|
}, {
|
|
|
|
|
# example of old live webstream
|
|
|
|
|
'url': 'https://multimedia.europarl.europa.eu/en/webstreaming/euroscola_20221115-1000-SPECIAL-EUROSCOLA',
|
|
|
|
|
# Test case that previously failed with regex method
|
|
|
|
|
'url': 'https://multimedia.europarl.europa.eu/en/webstreaming/euroscola_20250328-1000-SPECIAL-EUROSCOLA',
|
|
|
|
|
'info_dict': {
|
|
|
|
|
'id': str, # ID might be a string UUID or similar
|
|
|
|
|
'display_id': '20250328-1000-SPECIAL-EUROSCOLA',
|
|
|
|
|
'ext': 'mp4',
|
|
|
|
|
'id': '510eda7f-ba72-161b-7ee7-0e836cd2e715',
|
|
|
|
|
'release_timestamp': 1668502800,
|
|
|
|
|
'title': 'Euroscola 2022-11-15 19:21',
|
|
|
|
|
'release_date': '20221115',
|
|
|
|
|
'live_status': 'is_live',
|
|
|
|
|
'title': r're:Euroscola', # Expect title containing Euroscola
|
|
|
|
|
'release_timestamp': int, # Expecting a Unix timestamp
|
|
|
|
|
'release_date': '20250328',
|
|
|
|
|
'is_live': bool, # Could be True (if near event time) or False
|
|
|
|
|
},
|
|
|
|
|
'skip': 'not live anymore',
|
|
|
|
|
'params': {'skip_download': True},
|
|
|
|
|
# Note: This test might fail after 2025-03-28 if the URL becomes invalid or content changes significantly
|
|
|
|
|
}]
|
|
|
|
|
|
|
|
|
|
def _real_extract(self, url):
|
|
|
|
|
display_id = self._match_id(url)
|
|
|
|
|
webpage = self._download_webpage(url, display_id)
|
|
|
|
|
display_id = self._match_id(url) # Get ID from URL
|
|
|
|
|
webpage = self._download_webpage(url, display_id) # Get page HTML
|
|
|
|
|
|
|
|
|
|
# Try to parse Next.js data for metadata
|
|
|
|
|
nextjs = self._search_nextjs_data(webpage, display_id, default={})
|
|
|
|
|
page_props = traverse_obj(nextjs, ('props', 'pageProps'), default={})
|
|
|
|
|
media_info = page_props.get('mediaItem') or {} # Look for start/end times here for archives?
|
|
|
|
|
# --- Extract Metadata (prioritize Next.js data) ---
|
|
|
|
|
nextjs_data = self._search_nextjs_data(webpage, display_id, default={})
|
|
|
|
|
# Use traverse_obj for safer nested dictionary access
|
|
|
|
|
media_info = traverse_obj(nextjs_data, ('props', 'pageProps', 'mediaItem')) or {}
|
|
|
|
|
|
|
|
|
|
# Extract basic info, falling back to display_id if metadata is sparse
|
|
|
|
|
internal_id = media_info.get('id') or display_id
|
|
|
|
|
title = media_info.get('title') or media_info.get('name') or display_id
|
|
|
|
|
release_timestamp = None
|
|
|
|
|
# Existing logic uses startDateTime, might need adjustment for archive start/end
|
|
|
|
|
if 'startDateTime' in media_info:
|
|
|
|
|
release_timestamp = parse_iso8601(media_info['startDateTime'])
|
|
|
|
|
|
|
|
|
|
# Determine if it's Live or VOD/Archive (might need refinement)
|
|
|
|
|
# mediaSubType might be 'Live' or 'VOD' or something else
|
|
|
|
|
release_timestamp = traverse_obj(media_info, ('startDateTime', {parse_iso8601}))
|
|
|
|
|
# Determine live status based on metadata hint, if available
|
|
|
|
|
is_live = media_info.get('mediaSubType') == 'Live'
|
|
|
|
|
|
|
|
|
|
# Search for any .m3u8 link first
|
|
|
|
|
m3u8_links = self._search_regex(
|
|
|
|
|
r'(https?://[^"]+live\.media\.eup\.glcloud\.eu/hls/live/\d+/[^"]+\.m3u8[^"]*)',
|
|
|
|
|
webpage, 'm3u8 URL', default=None, group=1, fatal=False
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
# --- Potential modification area START ---
|
|
|
|
|
# If it's NOT live, and we have start/end times, and m3u8_links points to a live URL,
|
|
|
|
|
# try constructing the index-archive.m3u8 URL here.
|
|
|
|
|
# Example (conceptual - requires actual start/end times and base URL logic):
|
|
|
|
|
# if not is_live and media_info.get('startTime') and media_info.get('endTime'):
|
|
|
|
|
# start_time = media_info['startTime'] # Assuming these keys exist and hold timestamps
|
|
|
|
|
# end_time = media_info['endTime']
|
|
|
|
|
# # Assuming m3u8_links contains a base URL that needs modification
|
|
|
|
|
# base_url = m3u8_links.split('/')[0:-1] # Highly simplified base URL extraction
|
|
|
|
|
# archive_url = '/'.join(base_url) + f'/index-archive.m3u8?startTime={start_time}&endTime={end_time}'
|
|
|
|
|
# m3u8_links = archive_url # Replace the found link with the constructed one
|
|
|
|
|
# --- Potential modification area END ---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
if not m3u8_links:
|
|
|
|
|
self.report_warning('Could not find any .m3u8 link in the page. The site structure may have changed.')
|
|
|
|
|
# Return basic info if no HLS manifest found
|
|
|
|
|
return {
|
|
|
|
|
'id': media_info.get('id') or display_id,
|
|
|
|
|
'display_id': display_id,
|
|
|
|
|
'title': title,
|
|
|
|
|
'release_timestamp': release_timestamp,
|
|
|
|
|
'formats': [],
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
# Process all found .m3u8 links (handles case where multiple are found or the first one is a master playlist)
|
|
|
|
|
# The regex used here is identical to the one above, ensures we capture all instances
|
|
|
|
|
import re
|
|
|
|
|
all_links_text = self._html_search_regex(
|
|
|
|
|
r'(https?://[^"]+live\.media\.eup\.glcloud\.eu/hls/live/\d+/[^"]+\.m3u8[^"]*)',
|
|
|
|
|
webpage, 'all m3u8 URLs', default='', fatal=False, group=0 # Find all occurrences
|
|
|
|
|
)
|
|
|
|
|
candidates = re.findall(r'(https?://[^"]+live\.media\.eup\.glcloud\.eu/hls/live/\d+/[^"]+\.m3u8[^"]*)', all_links_text)
|
|
|
|
|
|
|
|
|
|
# If the specific constructed URL was made above, ensure it's prioritized or the only candidate
|
|
|
|
|
# (Refined logic needed here based on the modification above)
|
|
|
|
|
if not candidates and m3u8_links: # Fallback if findall failed but initial search worked
|
|
|
|
|
candidates = [m3u8_links]
|
|
|
|
|
elif m3u8_links not in candidates and m3u8_links: # Ensure the primary (possibly constructed) link is included
|
|
|
|
|
candidates.insert(0, m3u8_links)
|
|
|
|
|
|
|
|
|
|
candidates = list(dict.fromkeys(candidates)) # Make unique, preserving order
|
|
|
|
|
|
|
|
|
|
if not candidates: # Final check if still no candidates
|
|
|
|
|
self.report_warning('Could not extract any valid .m3u8 URLs.')
|
|
|
|
|
return {
|
|
|
|
|
'id': media_info.get('id') or display_id,
|
|
|
|
|
'display_id': display_id,
|
|
|
|
|
'title': title,
|
|
|
|
|
'release_timestamp': release_timestamp,
|
|
|
|
|
'formats': [],
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
formats, subtitles = [], {}
|
|
|
|
|
for link in candidates:
|
|
|
|
|
# Pass the identified m3u8 URL (could be live, index-archive, or norsk-archive)
|
|
|
|
|
# The 'live' flag might need adjustment based on mediaSubType
|
|
|
|
|
fmts, subs = self._extract_m3u8_formats_and_subtitles(
|
|
|
|
|
link, display_id, ext='mp4', live=is_live, fatal=False) # Pass is_live status
|
|
|
|
|
formats.extend(fmts)
|
|
|
|
|
self._merge_subtitles(subs, target=subtitles)
|
|
|
|
|
|
|
|
|
|
hls_url = None # Variable to store the found HLS URL
|
|
|
|
|
|
|
|
|
|
# --- Attempt 1: Find direct HLS URL in media_info ---
|
|
|
|
|
# Check common dictionary keys where the full HLS URL might be stored.
|
|
|
|
|
# Add more potential keys here if observed in website data.
|
|
|
|
|
possible_keys = ('hlsUrl', 'streamUrl', 'manifestUrl', 'url', 'playerUrl', 'videoUrl')
|
|
|
|
|
hls_url = traverse_obj(media_info, possible_keys)
|
|
|
|
|
if hls_url and 'm3u8' in hls_url: # Basic check if it looks like an HLS URL
|
|
|
|
|
self.to_screen(f'Found direct HLS URL in metadata: {hls_url}')
|
|
|
|
|
else:
|
|
|
|
|
hls_url = None # Reset if found value wasn't an HLS URL
|
|
|
|
|
|
|
|
|
|
# --- Attempt 2: Construct HLS URL from IDs in media_info ---
|
|
|
|
|
if not hls_url:
|
|
|
|
|
self.to_screen('Attempting to construct HLS URL from metadata IDs...')
|
|
|
|
|
# Try to extract relevant IDs. Keys like 'eventId', 'channelId' are common,
|
|
|
|
|
# but might differ. Use traverse_obj to safely get values.
|
|
|
|
|
# 'id' from media_info is often the event ID.
|
|
|
|
|
event_id = traverse_obj(media_info, ('id', 'eventId', 'event_id'))
|
|
|
|
|
# Channel ID might be numeric or a string name.
|
|
|
|
|
channel_id = traverse_obj(media_info, ('channelId', 'channel_id', 'channelName', 'channel'))
|
|
|
|
|
|
|
|
|
|
if event_id and channel_id:
|
|
|
|
|
# Construct the URL using the assumed live/default pattern.
|
|
|
|
|
# For archive/VOD, '/index-archive.m3u8?startTime=...&endTime=...' might be needed.
|
|
|
|
|
# This assumes the event is live or uses the default endpoint.
|
|
|
|
|
constructed_url = f'https://live.media.eup.glcloud.eu/hls/live/{event_id}/{channel_id}/index.m3u8'
|
|
|
|
|
hls_url = constructed_url
|
|
|
|
|
self.to_screen(f'Constructed potential HLS URL: {hls_url}')
|
|
|
|
|
else:
|
|
|
|
|
self.to_screen('Could not find sufficient event/channel IDs in metadata to construct URL.')
|
|
|
|
|
|
|
|
|
|
# --- Attempt 3: Fallback to regex search on raw webpage (Original Method) ---
|
|
|
|
|
if not hls_url:
|
|
|
|
|
self.to_screen('Could not find or construct HLS URL from metadata, trying webpage regex search...')
|
|
|
|
|
m3u8_url_pattern = r'(https?://[^"]*live\.media\.eup\.glcloud\.eu/hls/live/\d+/[^"]+\.m3u8[^"]*)'
|
|
|
|
|
hls_url = self._search_regex(
|
|
|
|
|
m3u8_url_pattern, webpage, 'm3u8 URL (regex fallback)', default=None, fatal=False)
|
|
|
|
|
if hls_url:
|
|
|
|
|
self.to_screen(f'Found HLS URL via regex fallback: {hls_url}')
|
|
|
|
|
else:
|
|
|
|
|
# This is where the original "Could not find any .m3u8 link" warning occurred.
|
|
|
|
|
self.report_warning('Could not find any .m3u8 link via metadata or webpage regex.')
|
|
|
|
|
|
|
|
|
|
# --- Process HLS Playlist ---
|
|
|
|
|
if not hls_url:
|
|
|
|
|
# If no URL was found after all attempts, raise an error.
|
|
|
|
|
raise ExtractorError(
|
|
|
|
|
'No HLS URL (.m3u8) could be found or constructed. The website structure might have changed.',
|
|
|
|
|
expected=True) # expected=True prevents stack trace for common errors
|
|
|
|
|
|
|
|
|
|
# Pass the found HLS URL to the HLS processing function.
|
|
|
|
|
# The _extract_m3u8_formats function usually detects live/VOD automatically.
|
|
|
|
|
# The 'live=is_live' hint can sometimes help but isn't strictly necessary.
|
|
|
|
|
formats, subtitles = self._extract_m3u8_formats_and_subtitles(
|
|
|
|
|
hls_url, display_id, ext='mp4', live=is_live, fatal=False)
|
|
|
|
|
|
|
|
|
|
# Check if HLS processing returned any formats
|
|
|
|
|
if not formats:
|
|
|
|
|
raise ExtractorError(f'HLS manifest found at {hls_url} but yielded no video formats.', expected=True)
|
|
|
|
|
|
|
|
|
|
# --- Return Extracted Information ---
|
|
|
|
|
return {
|
|
|
|
|
'id': media_info.get('id') or display_id,
|
|
|
|
|
'id': internal_id,
|
|
|
|
|
'display_id': display_id,
|
|
|
|
|
'title': title,
|
|
|
|
|
'formats': formats,
|
|
|
|
|
'subtitles': subtitles,
|
|
|
|
|
'release_timestamp': release_timestamp,
|
|
|
|
|
# Report 'is_live' based on detected mediaSubType
|
|
|
|
|
'is_live': is_live or None # Report None if not explicitly Live
|
|
|
|
|
'is_live': is_live or None, # Use None if not explicitly marked Live
|
|
|
|
|
}
|
|
|
|
|