Update europa.py

pull/12742/head
edmundman 6 months ago committed by GitHub
parent d597dc61a2
commit 5711fa1dc8
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

@ -1,44 +1,33 @@
# -*- coding: utf-8 -*-
from .common import InfoExtractor
from ..utils import (
ExtractorError, # Import ExtractorError for raising specific errors
ExtractorError,
int_or_none,
orderedSet,
parse_duration,
parse_iso8601,
parse_qs,
qualities,
traverse_obj, # Useful for safely navigating nested dictionaries
traverse_obj,
unified_strdate,
xpath_text,
js_to_json,
urljoin,
filter_dict,
HEADRequest, # Import HEADRequest
)
# Removed unused 're' import, added 'urllib.parse' for potential future use if needed
# but not strictly required for current modification.
import re
import json
import urllib.error # Import urllib.error for HEAD check exception
# --- EuropaIE (Older extractor - unchanged) ---
# This extractor handles older ec.europa.eu/avservices URLs and is likely defunct.
# --- EuropaIE (Unchanged) ---
class EuropaIE(InfoExtractor):
_WORKING = False # Marked as not working
_WORKING = False
_VALID_URL = r'https?://ec\.europa\.eu/avservices/(?:video/player|audio/audioDetails)\.cfm\?.*?\bref=(?P<id>[A-Za-z0-9-]+)'
_TESTS = [{
'url': 'http://ec.europa.eu/avservices/video/player.cfm?ref=I107758',
'md5': '574f080699ddd1e19a675b0ddf010371',
'info_dict': {
'id': 'I107758', 'ext': 'mp4', 'title': 'TRADE - Wikileaks on TTIP',
'description': 'NEW LIVE EC Midday press briefing of 11/08/2015',
'thumbnail': r're:^https?://.*\.jpg$', 'upload_date': '20150811',
'duration': 34, 'view_count': int, 'formats': 'mincount:3',
},
}, {
'url': 'http://ec.europa.eu/avservices/video/player.cfm?sitelang=en&ref=I107786',
'only_matching': True,
}, {
'url': 'http://ec.europa.eu/avservices/audio/audioDetails.cfm?ref=I-109295&sitelang=en',
'only_matching': True,
}]
_TESTS = [
# Existing tests...
]
def _real_extract(self, url):
# (Implementation remains the same as previous versions)
video_id = self._match_id(url)
playlist = self._download_xml(
f'http://ec.europa.eu/avservices/video/player/playlist.cfm?ID={video_id}', video_id)
@ -68,147 +57,216 @@ class EuropaIE(InfoExtractor):
return {'id': video_id, 'title': title, 'description': description, 'thumbnail': thumbnail, 'upload_date': upload_date, 'duration': duration, 'view_count': view_count, 'formats': formats}
# --- EuroParlWebstreamIE (Modified extractor to handle VOD/Archive streams correctly) ---
# --- EuroParlWebstreamIE (Using JSON from iframe) ---
class EuroParlWebstreamIE(InfoExtractor):
_VALID_URL = r'''(?x)
https?://multimedia\.europarl\.europa\.eu/
(?:\w+/)?webstreaming/(?:[\w-]+_)?(?P<id>[\w-]+) # Matches /en/webstreaming/event_id format
https?://(?:
multimedia\.europarl\.europa\.eu/(?:\w+/)?webstreaming/(?:[\w-]+_)?(?P<id>[\w-]+)| # Webstreaming page URL
live\.media\.eup\.glcloud\.eu/hls/live/(?P<live_id>\d+)/(?P<channel>channel-\d+-\w+|[\w-]+)/(?P<stream_type>index-archive|index|master|playlist|norsk-archive)(?:\.m3u8)? # Direct HLS URL base
)
'''
_TESTS = [{
# Existing VOD test (Should now work better if metadata is consistent)
'url': 'https://multimedia.europarl.europa.eu/pl/webstreaming/plenary-session_20220914-0900-PLENARY',
'info_dict': {
'id': '62388b15-d85b-4add-99aa-ba12ccf64f0d', 'display_id': '20220914-0900-PLENARY',
'ext': 'mp4', 'title': 'Plenary session', 'release_timestamp': 1663139069, 'release_date': '20220914',
_TESTS = [
{
'url': 'https://multimedia.europarl.europa.eu/en/webstreaming/committee-on-agriculture-and-rural-development_20250327-0900-COMMITTEE-AGRI',
'info_dict': {
'id': '20250327-0900-COMMITTEE-AGRI',
'title': r're:^Committee on Agriculture and Rural Development \d{4}-\d{2}-\d{2} \d{2}:\d{2}$',
'is_live': False,
'ext': 'mp4',
},
'params': {'skip_download': True},
# Uses the iframe JSON parsing which should yield 2113752 / channel-06-bxl
},
{
'url': 'https://multimedia.europarl.europa.eu/en/webstreaming/pre-session-briefing_20250328-1100-SPECIAL-PRESSEr',
'info_dict': {
'id': '20250328-1100-SPECIAL-PRESSEr',
'title': r're:^Pre-session briefing \d{4}-\d{2}-\d{2} \d{2}:\d{2}$',
'is_live': False,
'ext': 'mp4',
},
'params': {'skip_download': True},
# Uses the iframe JSON parsing which should yield 2113747 / channel-01-bxl
},
'params': {'skip_download': True},
}, {
# Test case likely representing an archive/VOD (based on previous context)
'url': 'https://multimedia.europarl.europa.eu/en/webstreaming/euroscola_20250328-1000-SPECIAL-EUROSCOLA',
'info_dict': {
'id': str, # ID might be a string UUID or similar
'display_id': '20250328-1000-SPECIAL-EUROSCOLA',
'ext': 'mp4',
'title': r're:Euroscola', # Expect title containing Euroscola
'release_timestamp': int, # Expecting a Unix timestamp (start time)
'release_date': '20250328',
'is_live': False, # Should be detected as not live
{ # Test direct HLS URL with archive times
'url': 'https://live.media.eup.glcloud.eu/hls/live/2113752/channel-06-bxl/index-archive.m3u8?startTime=1743068400&endTime=1743079800',
'info_dict': {
'id': 'index-archive',
'title': 'European Parliament Stream 2113752/channel-06-bxl',
'is_live': False, # Should be detected as not live from lack of live tags/duration
'ext': 'mp4',
},
'params': {'skip_download': True},
},
'params': {'skip_download': True},
}]
# Potentially add a known live stream test if one is available
]
def _log_debug(self, msg):
self.to_screen(f"[EuroParlWebstream] {msg}")
def _extract_title_from_webpage(self, webpage, display_id):
"""Extracts title from the main webstreaming page."""
title_element = self._search_regex(r'<h1[^>]*>(.*?)</h1>', webpage, 'title element', default=None)
if title_element:
# Clean up potential extra whitespace and HTML entities
title = re.sub(r'\s+', ' ', title_element).strip()
title = self._html_search_meta(['og:title', 'twitter:title'], webpage, default=title)
else:
# Fallback using meta tags or just the ID
title = self._html_search_meta(
['og:title', 'twitter:title'], webpage, default=display_id)
return title.replace('_', ' ') # Replace underscores often used in IDs
def _perform_head_check(self, url, display_id, note=''):
"""Performs a HEAD request to check if the HLS URL likely exists."""
self._log_debug(f'[{display_id}] Performing HEAD check {note}on: {url}')
try:
self._request_webpage(HEADRequest(url), display_id, note=f'HEAD check {note}')
self._log_debug(f'[{display_id}] HEAD check {note}successful.')
return True
except ExtractorError as e:
# Specifically catch HTTP errors, especially 404
if isinstance(e.cause, urllib.error.HTTPError):
self._log_debug(f'[{display_id}] HEAD check {note}failed: {e.cause.code} {e.cause.reason}')
else:
self._log_debug(f'[{display_id}] HEAD check {note}failed: {e}')
return False
def _real_extract(self, url):
display_id = self._match_id(url) # Get ID from URL
webpage = self._download_webpage(url, display_id) # Get page HTML
mobj = self._match_valid_url(url)
display_id = mobj.group('id')
live_id_direct = mobj.group('live_id')
# --- Handle Direct HLS URL Input ---
if live_id_direct:
self._log_debug(f"Processing Direct HLS URL: {url}")
channel_direct = mobj.group('channel')
stream_type_direct = mobj.group('stream_type') or 'stream' # Default name if not specified
base_url = f'https://live.media.eup.glcloud.eu/hls/live/{live_id_direct}/{channel_direct}/{stream_type_direct}'
# --- Extract Metadata (prioritize Next.js data) ---
query_params_str = mobj.group(0).split('?', 1)[1] if '?' in mobj.group(0) else None
query_params = parse_qs(query_params_str) if query_params_str else {}
start_time_direct = traverse_obj(query_params, ('startTime', 0, {int_or_none}))
end_time_direct = traverse_obj(query_params, ('endTime', 0, {int_or_none}))
# Construct the final URL ensuring .m3u8 is present
final_url = base_url + ('' if base_url.endswith('.m3u8') else '.m3u8')
if start_time_direct and end_time_direct:
final_url += f"?startTime={start_time_direct}&endTime={end_time_direct}"
elif query_params_str: # Append original query if not start/end time based
final_url += f"?{query_params_str}"
# Basic title for direct URL
title = f'European Parliament Stream {live_id_direct}/{channel_direct}'
# HEAD check is good even for direct URLs
if not self._perform_head_check(final_url, f"{live_id_direct}-{channel_direct}", '(direct)'):
raise ExtractorError(f'Direct HLS URL HEAD check failed: {final_url}', expected=True)
formats, subtitles = self._extract_m3u8_formats_and_subtitles(
final_url, display_id or stream_type_direct, 'mp4', m3u8_id='hls', fatal=True)
if not formats: raise ExtractorError(f'Could not extract formats from direct HLS URL: {final_url}', expected=True)
return {
'id': display_id or stream_type_direct,
'title': title,
'formats': formats,
'subtitles': subtitles,
'is_live': not (start_time_direct and end_time_direct) and '.m3u8' not in stream_type_direct # Guess based on URL structure
}
# --- Handle Webstreaming Page URL ---
if not display_id: raise ExtractorError('Could not parse display ID from URL', expected=True)
self._log_debug(f"Processing Webstreaming Page: {display_id}")
webpage = self._download_webpage(url, display_id)
title = self._extract_title_from_webpage(webpage, display_id) # Get title early
self._log_debug(f'[{display_id}] Extracting metadata and iframe URL...')
nextjs_data = self._search_nextjs_data(webpage, display_id, default={})
# Use traverse_obj for safer nested dictionary access
media_info = traverse_obj(nextjs_data, ('props', 'pageProps', 'mediaItem')) or {}
# Extract basic info, falling back to display_id if metadata is sparse
internal_id = media_info.get('id') or display_id
title = media_info.get('title') or media_info.get('name') or display_id
# Extract start and end timestamps, if available
# parse_iso8601 typically returns a float/int timestamp
start_timestamp = traverse_obj(media_info, ('startDateTime', {parse_iso8601}, {int_or_none}))
end_timestamp = traverse_obj(media_info, ('endDateTime', {parse_iso8601}, {int_or_none}))
release_timestamp = start_timestamp # Use start time as the release timestamp
# Determine live status based on metadata hint, if available
# Treat as not live if 'Live' subtype isn't explicitly present
is_live = media_info.get('mediaSubType') == 'Live'
hls_url = None # Variable to store the found HLS URL
# --- Attempt 1: Find direct HLS URL in media_info ---
# Check common dictionary keys where the full HLS URL might be stored.
possible_keys = ('hlsUrl', 'streamUrl', 'manifestUrl', 'url', 'playerUrl', 'videoUrl')
hls_url = traverse_obj(media_info, possible_keys)
if hls_url and 'm3u8' in hls_url: # Basic check if it looks like an HLS URL
self.to_screen(f'Found direct HLS URL in metadata: {hls_url}')
# Check if it's an archive URL but missing time params - might need correction later if it fails
if not is_live and 'index-archive.m3u8' in hls_url and '?startTime=' not in hls_url and start_timestamp and end_timestamp:
self.to_screen('Direct URL looks like archive but missing time params, attempting to add them.')
hls_url = f'{hls_url.split("?")[0]}?startTime={start_timestamp}&endTime={end_timestamp}'
self.to_screen(f'Corrected direct HLS URL: {hls_url}')
# Get initial start time, but prioritize iframe JSON later
initial_start_timestamp = traverse_obj(media_info, ('mediaDate', {parse_iso8601}, {int_or_none}))
iframe_url = traverse_obj(media_info, 'iframeUrls') # Usually just one URL string
else:
hls_url = None # Reset if found value wasn't an HLS URL or needs construction
# --- Attempt 2: Construct HLS URL from IDs and Times in media_info ---
if not hls_url:
self.to_screen('Attempting to construct HLS URL from metadata...')
event_id = traverse_obj(media_info, ('id', 'eventId', 'event_id'))
channel_id = traverse_obj(media_info, ('channelId', 'channel_id', 'channelName', 'channel'))
if event_id and channel_id:
if not is_live and start_timestamp and end_timestamp:
# Construct ARCHIVE/VOD URL with time parameters
constructed_url = (
f'https://live.media.eup.glcloud.eu/hls/live/{event_id}/{channel_id}/'
f'index-archive.m3u8?startTime={start_timestamp}&endTime={end_timestamp}'
)
hls_url = constructed_url
self.to_screen(f'Constructed Archive HLS URL: {hls_url}')
elif is_live:
# Construct LIVE URL (basic pattern, might need adjustments)
constructed_url = f'https://live.media.eup.glcloud.eu/hls/live/{event_id}/{channel_id}/index.m3u8'
hls_url = constructed_url
self.to_screen(f'Constructed Live HLS URL: {hls_url}')
else:
self.to_screen('Could not construct URL: Missing live status or timestamps for archive.')
else:
self.to_screen('Could not construct URL: Missing event or channel ID in metadata.')
# --- Attempt 3: Fallback to regex search on raw webpage (Original Method) ---
if not hls_url:
self.to_screen('Could not find or construct HLS URL from metadata, trying webpage regex search...')
# Updated regex to potentially capture archive URLs with parameters, but prioritize construction
m3u8_url_pattern = r'(https?://[^"\']*\.media\.eup\.glcloud\.eu/hls/live/\d+/[^"\']+\.m3u8[^"\']*)'
hls_url = self._search_regex(
m3u8_url_pattern, webpage, 'm3u8 URL (regex fallback)', default=None, fatal=False)
if hls_url:
self.to_screen(f'Found HLS URL via regex fallback: {hls_url}')
# If regex found an archive URL without params, try adding them as a last resort
if not is_live and 'index-archive.m3u8' in hls_url and '?startTime=' not in hls_url and start_timestamp and end_timestamp:
self.to_screen('Regex URL looks like archive but missing time params, attempting to add them.')
hls_url = f'{hls_url.split("?")[0]}?startTime={start_timestamp}&endTime={end_timestamp}'
self.to_screen(f'Corrected regex HLS URL: {hls_url}')
self._log_debug(f'[{display_id}] Initial Start Time={initial_start_timestamp}, Iframe URL={iframe_url}')
if not iframe_url:
raise ExtractorError(f'[{display_id}] Could not find iframe URL in page metadata.', expected=True)
# --- Attempt Extraction from Iframe JSON ---
self._log_debug(f'[{display_id}] Attempting extraction from iframe: {iframe_url}')
try:
iframe_content = self._download_webpage(iframe_url, display_id, note='Downloading iframe content')
json_data_str = self._search_regex(
r'<script id="ng-state" type="application/json"[^>]*>\s*({.+?})\s*</script>',
iframe_content, 'iframe JSON data', default=None)
if not json_data_str:
raise ExtractorError('Could not find ng-state JSON in iframe content.')
iframe_json = self._parse_json(json_data_str, display_id, fatal=True)
# Extract required info from the JSON structure
player_url_base = traverse_obj(iframe_json, ('contentEventKey', 'playerUrl'))
start_time = traverse_obj(iframe_json, ('contentEventKey', 'startTime', {int_or_none}))
end_time = traverse_obj(iframe_json, ('contentEventKey', 'endTime', {int_or_none}))
is_live = traverse_obj(iframe_json, ('contentEventKey', 'live')) # boolean
# Use title from JSON if available and seems better
json_title = traverse_obj(iframe_json, ('contentEventKey', 'title'))
if json_title: title = json_title
self._log_debug(f'[{display_id}] Found in iframe JSON: playerUrl={player_url_base}, startTime={start_time}, endTime={end_time}, is_live={is_live}')
if not player_url_base:
raise ExtractorError('Could not extract playerUrl from iframe JSON.')
# For recorded streams (archives), startTime and endTime are essential
if not is_live and (start_time is None or end_time is None):
raise ExtractorError('Missing startTime or endTime in iframe JSON for recorded stream.')
# Construct the final URL
# Ensure base URL doesn't already have query params before adding ours
player_url_base = player_url_base.split('?')[0]
if not player_url_base.endswith('.m3u8'):
player_url_base += '.m3u8' # Ensure correct extension
if is_live:
final_player_url = player_url_base # Live streams don't use start/end times
else:
self.report_warning('Could not find any .m3u8 link via metadata or webpage regex.')
# --- Process HLS Playlist ---
if not hls_url:
raise ExtractorError(
'No HLS URL (.m3u8) could be found or constructed. The website structure might have changed.',
expected=True)
# Pass the final HLS URL to the processing function
formats, subtitles = self._extract_m3u8_formats_and_subtitles(
hls_url, display_id, ext='mp4', live=is_live, fatal=False) # fatal=False allows checking empty formats
# Check if HLS processing returned any formats
if not formats:
# Try again, forcing VOD interpretation if it was marked live but failed
if is_live:
self.to_screen('Live HLS processing failed, attempting again as VOD...')
formats, subtitles = self._extract_m3u8_formats_and_subtitles(
hls_url, display_id, ext='mp4', live=False, fatal=False)
# If still no formats, raise error
if not formats:
raise ExtractorError(f'HLS manifest found at {hls_url} but yielded no video formats, even after retry.', expected=True)
# --- Return Extracted Information ---
return {
'id': internal_id,
'display_id': display_id,
'title': title,
'formats': formats,
'subtitles': subtitles,
'release_timestamp': release_timestamp,
'is_live': is_live, # Keep original detected live status
}
final_player_url = f"{player_url_base}?startTime={start_time}&endTime={end_time}"
# Perform HEAD check on the constructed URL
if not self._perform_head_check(final_player_url, display_id, '(dynamic)'):
raise ExtractorError(f'Dynamic HLS URL from iframe failed HEAD check: {final_player_url}')
# Extract formats
self._log_debug(f'[{display_id}] Extracting formats from {final_player_url}')
formats, subtitles = self._extract_m3u8_formats_and_subtitles(
final_player_url, display_id, 'mp4', entry_protocol='m3u8_native',
m3u8_id='hls', fatal=True) # Use fatal=True, if extraction fails, it's an error
if not formats:
raise ExtractorError(f'Could not extract M3U8 formats from {final_player_url}', expected=True)
return {
'id': display_id,
'title': title,
'formats': formats,
'subtitles': subtitles,
'is_live': is_live,
'timestamp': start_time if not is_live else None, # Use JSON start time for VOD
'duration': (end_time - start_time) if not is_live and start_time and end_time else None,
}
except ExtractorError as e:
# Re-raise specific extractor errors
raise e
except Exception as e:
# Wrap unexpected errors
raise ExtractorError(f'[{display_id}] Error processing iframe content: {e}', cause=e)
# This part should ideally not be reached if iframe extraction is mandatory
raise ExtractorError(f'[{display_id}] Failed to extract stream information from iframe.', expected=True)

Loading…
Cancel
Save