|
|
|
|
|
|
|
"""A place for common utilities between wubloader components"""
|
|
|
|
|
|
|
|
|
|
|
|
import base64
|
|
|
|
import datetime
|
|
|
|
import errno
|
|
|
|
import itertools
|
|
|
|
import logging
|
|
|
|
import os
|
|
|
|
import sys
|
|
|
|
from collections import namedtuple
|
|
|
|
|
|
|
|
import dateutil.parser
|
|
|
|
|
|
|
|
|
|
|
|
def dt_to_bustime(start, dt):
|
|
|
|
"""Convert a datetime to bus time. Bus time is seconds since the given start point."""
|
|
|
|
return (dt - start).total_seconds()
|
|
|
|
|
|
|
|
|
|
|
|
def bustime_to_dt(start, bustime):
|
|
|
|
"""Convert from bus time to a datetime"""
|
|
|
|
return start + datetime.timedelta(seconds=bustime)
|
|
|
|
|
|
|
|
|
|
|
|
def format_bustime(bustime, round="millisecond"):
|
|
|
|
"""Convert bustime to a human-readable string (-)HH:MM:SS.fff, with the
|
|
|
|
ending cut off depending on the value of round:
|
|
|
|
"millisecond": (default) Round to the nearest millisecond.
|
|
|
|
"second": Round down to the current second.
|
|
|
|
"minute": Round down to the current minute.
|
|
|
|
Examples:
|
|
|
|
00:00:00.000
|
|
|
|
01:23:00
|
|
|
|
110:50
|
|
|
|
159:59:59.999
|
|
|
|
-10:30:01.100
|
|
|
|
Negative times are formatted as time-until-start, preceeded by a minus
|
|
|
|
sign.
|
|
|
|
eg. "-1:20:00" indicates the run begins in 80 minutes.
|
|
|
|
"""
|
|
|
|
sign = ''
|
|
|
|
if bustime < 0:
|
|
|
|
sign = '-'
|
|
|
|
bustime = -bustime
|
|
|
|
total_mins, secs = divmod(bustime, 60)
|
|
|
|
hours, mins = divmod(total_mins, 60)
|
|
|
|
parts = [
|
|
|
|
"{:02d}".format(int(hours)),
|
|
|
|
"{:02d}".format(int(mins)),
|
|
|
|
]
|
|
|
|
if round == "minute":
|
|
|
|
pass
|
|
|
|
elif round == "second":
|
|
|
|
parts.append("{:02d}".format(int(secs)))
|
|
|
|
elif round == "millisecond":
|
|
|
|
parts.append("{:06.3f}".format(secs))
|
|
|
|
else:
|
|
|
|
raise ValueError("Bad rounding value: {!r}".format(round))
|
|
|
|
return sign + ":".join(parts)
|
|
|
|
|
|
|
|
|
|
|
|
def unpadded_b64_decode(s):
|
|
|
|
"""Decode base64-encoded string that has had its padding removed"""
|
|
|
|
# right-pad with '=' to multiple of 4
|
|
|
|
s = s + '=' * (- len(s) % 4)
|
|
|
|
return base64.b64decode(s, "-_")
|
|
|
|
|
|
|
|
|
|
|
|
class SegmentInfo(
|
|
|
|
namedtuple('SegmentInfoBase', [
|
|
|
|
'path', 'stream', 'variant', 'start', 'duration', 'is_partial', 'hash'
|
|
|
|
])
|
|
|
|
):
|
|
|
|
"""Info parsed from a segment path, including original path."""
|
|
|
|
@property
|
|
|
|
def end(self):
|
|
|
|
return self.start + self.duration
|
|
|
|
|
|
|
|
|
|
|
|
def parse_segment_path(path):
|
|
|
|
"""Parse segment path, returning a SegmentInfo. If path is only the trailing part,
|
|
|
|
eg. just a filename, it will leave unknown fields as None."""
|
|
|
|
parts = path.split('/')
|
|
|
|
# left-pad parts with None up to 4 parts
|
|
|
|
parts = [None] * (4 - len(parts)) + parts
|
|
|
|
# pull info out of path parts
|
|
|
|
stream, variant, hour, filename = parts[-4:]
|
|
|
|
# split filename, which should be TIME-DURATION-TYPE-HASH.ts
|
|
|
|
try:
|
|
|
|
if not filename.endswith('.ts'):
|
|
|
|
raise ValueError("Does not end in .ts")
|
|
|
|
filename = filename[:-len('.ts')] # chop off .ts
|
|
|
|
parts = filename.split('-', 3)
|
|
|
|
if len(parts) != 4:
|
|
|
|
raise ValueError("Not enough dashes in filename")
|
|
|
|
time, duration, type, hash = parts
|
|
|
|
if type not in ('full', 'partial'):
|
|
|
|
raise ValueError("Unknown type {!r}".format(type))
|
|
|
|
return SegmentInfo(
|
|
|
|
path = path,
|
|
|
|
stream = stream,
|
|
|
|
variant = variant,
|
|
|
|
start = dateutil.parser.parse("{}:{}".format(hour, time)),
|
|
|
|
duration = datetime.timedelta(seconds=float(duration)),
|
|
|
|
is_partial = type == "partial",
|
|
|
|
hash = unpadded_b64_decode(hash),
|
|
|
|
)
|
|
|
|
except ValueError as e:
|
|
|
|
# wrap error but preserve original traceback
|
|
|
|
_, _, tb = sys.exc_info()
|
|
|
|
raise ValueError, ValueError("Bad path {!r}: {}".format(path, e)), tb
|
|
|
|
|
|
|
|
|
|
|
|
def get_best_segments(hours_path, start, end):
|
|
|
|
"""Return a list of the best sequence of non-overlapping segments
|
|
|
|
we have for a given time range. Hours path should be the directory containing hour directories.
|
|
|
|
The first segment may start before the time range, and the last may end after it.
|
|
|
|
The returned list contains items that are either:
|
|
|
|
SegmentInfo: a segment
|
|
|
|
None: represents a discontinuity between the previous segment and the next one.
|
|
|
|
ie. as long as two segments appear next to each other, we guarentee there is no gap between
|
|
|
|
them, the second one starts right as the first one finishes.
|
|
|
|
Similarly, unless the first item is None, the first segment starts <= the start of the time
|
|
|
|
range, and unless the last item is None, the last segment ends >= the end of the time range.
|
|
|
|
Example:
|
|
|
|
Suppose you ask for a time range from 10 to 60. We have 10-second segments covering
|
|
|
|
the following times:
|
|
|
|
5 to 15
|
|
|
|
15 to 25
|
|
|
|
30 to 40
|
|
|
|
40 to 50
|
|
|
|
Then the output would look like:
|
|
|
|
segment from 5 to 15
|
|
|
|
segment from 15 to 25
|
|
|
|
None, as the previous segment ends 5sec before the next one begins
|
|
|
|
segment from 30 to 40
|
|
|
|
segment from 40 to 50
|
|
|
|
None, as the previous segment ends 10sec before the requested end time of 60.
|
|
|
|
Note that any is_partial=True segment will be followed by a None, since we can't guarentee
|
|
|
|
it joins on to the next segment fully intact.
|
|
|
|
"""
|
|
|
|
# Note: The exact equality checks in this function are not vulnerable to floating point error,
|
|
|
|
# but only because all input dates and durations are only precise to the millisecond, and
|
|
|
|
# python's datetime types represent these as integer microseconds internally. So the parsing
|
|
|
|
# to these types is exact, and all operations on them are exact, so all operations are exact.
|
|
|
|
|
|
|
|
result = []
|
|
|
|
|
|
|
|
for hour in hour_paths_for_range(hours_path, start, end):
|
|
|
|
# best_segments_by_start will give us the best available segment for each unique start time
|
|
|
|
for segment in best_segments_by_start(hour):
|
|
|
|
|
|
|
|
# special case: first segment
|
|
|
|
if not result:
|
|
|
|
# first segment is allowed to be before start as long as it includes it
|
|
|
|
if segment.start <= start < segment.end:
|
|
|
|
# segment covers start
|
|
|
|
result.append(segment)
|
|
|
|
elif start < segment.start < end:
|
|
|
|
# segment is after start (but before end), so there was no segment that covers start
|
|
|
|
# so we begin with a None
|
|
|
|
result.append(None)
|
|
|
|
result.append(segment)
|
|
|
|
else:
|
|
|
|
# segment is before start, and doesn't cover start, or starts after end.
|
|
|
|
# ignore and go to next.
|
|
|
|
continue
|
|
|
|
else:
|
|
|
|
# normal case: check against previous segment end time
|
|
|
|
prev_end = result[-1].end
|
|
|
|
if segment.start < prev_end:
|
|
|
|
# Overlap! This shouldn't happen, though it might be possible due to weirdness
|
|
|
|
# if the stream drops then starts again quickly. We simply ignore the overlapping
|
|
|
|
# segment and let the algorithm continue.
|
|
|
|
logging.warning("Overlapping segments: {} overlaps end of {}".format(segment, result[-1]))
|
|
|
|
continue
|
|
|
|
if result[-1].is_partial or prev_end < segment.start:
|
|
|
|
# there's a gap between prev end and this start, so add a None
|
|
|
|
result.append(None)
|
|
|
|
result.append(segment)
|
|
|
|
|
|
|
|
# check if we've reached the end
|
|
|
|
if end <= segment.end:
|
|
|
|
break
|
|
|
|
|
|
|
|
# this is a weird little construct that says "if we broke from the inner loop,
|
|
|
|
# then also break from the outer one. otherwise continue."
|
|
|
|
else:
|
|
|
|
continue
|
|
|
|
break
|
|
|
|
|
|
|
|
# check if we need a trailing None because last segment is partial or doesn't reach end
|
|
|
|
if result and (result[-1].is_partial or result[-1].end < end):
|
|
|
|
result.append(None)
|
|
|
|
|
|
|
|
return result
|
|
|
|
|
|
|
|
|
|
|
|
def hour_paths_for_range(hours_path, start, end):
|
|
|
|
"""Generate a list of hour paths to check when looking for segments between start and end."""
|
|
|
|
# truncate start and end to the hour
|
|
|
|
def truncate(dt):
|
|
|
|
return dt.replace(microsecond=0, second=0, minute=0)
|
|
|
|
current = truncate(start)
|
|
|
|
end = truncate(end)
|
|
|
|
# Begin in the hour prior to start, as there may be a segment that starts in that hour
|
|
|
|
# but contains the start time, eg. if the start time is 01:00:01 and there's a segment
|
|
|
|
# at 00:59:59 which goes for 3 seconds.
|
|
|
|
# Checking the entire hour when in most cases it won't be needed is wasteful, but it's also
|
|
|
|
# pretty quick and the complexity of only checking this case when needed just isn't worth it.
|
|
|
|
current -= datetime.timedelta(hours=1)
|
|
|
|
while current <= end:
|
|
|
|
yield os.path.join(hours_path, current.strftime("%Y-%m-%dT%H"))
|
|
|
|
current += datetime.timedelta(hours=1)
|
|
|
|
|
|
|
|
|
|
|
|
def best_segments_by_start(hour):
|
|
|
|
"""Within a given hour path, yield the "best" segment per unique segment start time.
|
|
|
|
Best is defined as non-partial, or failing that the longest partial.
|
|
|
|
Note this means this function may perform os.stat()s in order to find the longest partial.
|
|
|
|
"""
|
|
|
|
try:
|
|
|
|
segment_paths = os.listdir(hour)
|
|
|
|
except OSError as e:
|
|
|
|
if e.errno != errno.ENOENT:
|
|
|
|
raise
|
|
|
|
# path does not exist, treat it as having no files
|
|
|
|
return
|
|
|
|
segment_paths.sort()
|
|
|
|
# note we only parse them as we need them, which is unlikely to save us much time overall
|
|
|
|
# but is easy enough to do, so we might as well.
|
|
|
|
parsed = (parse_segment_path(os.path.join(hour, name)) for name in segment_paths)
|
|
|
|
for start_time, segments in itertools.groupby(parsed, key=lambda segment: segment.start):
|
|
|
|
segments = list(segments)
|
|
|
|
full_segments = [segment for segment in segments if not segment.is_partial]
|
|
|
|
if full_segments:
|
|
|
|
if len(full_segments) != 1:
|
|
|
|
logging.warning("Multiple versions of full segment at start_time {}: {}".format(
|
|
|
|
start_time, ", ".join(map(str, segments))
|
|
|
|
))
|
|
|
|
# we have to pick one, so might as well make it consistent by sorting by path
|
|
|
|
full_segments.sort(key=lambda segment: segment.path)
|
|
|
|
yield full_segments[0]
|
|
|
|
continue
|
|
|
|
# no full segments, fall back to measuring partials.
|
|
|
|
yield max(segments, key=lambda segment: os.stat(segment.path).st_size)
|