[FFmpeg-devel] [RFC PATCH] avformat/dashdec: Improve implementation
Steven Liu
lingjiujianke at gmail.com
Sat Mar 27 04:29:56 EET 2021
zsugabubus <zsugabubus at national.shitposting.agency> 于2021年3月23日周二 上午6:25写道:
>
> Compared to previous implementation, this rework tries to:
>
> - Improve code quality,
> - Provide better error handling (also: reduce numerous (potential)
> memory leaks),
> - Broader coverage of the standard:
> * multiple periods,
> * multiple base URLs (fallback on CDN failure),
> * handle ContentProtection/Role/etc... nodes,
> * manifest chaining,
> * drop ridiculous limits of @id's.
> - Fast startup, thanks to reading/probing streams only when needed.
> - Open multiple connections (HLS can open one more connection in
> advance; DASH can do up to UINT_MAX, configurable via option);
> similarly to HLS, currently it only opens them so only sequential
> reading is possible, however this behavior can be changed in the
> future,
> - Ability to seek live streams.
> - Bonus: Tests show that there are some kind of live streams that old
> implementation was unable to cope with, now it is possible to play
> them.
>
> I mark this patch as RFC since I think:
> - Seeking needs some improvement (I currently lack the necessary
> knowledge to finish it (...I hope a little community help)),
> - Stuff around codec reinitialization (mentioned late opening +
> multiperiod) may require some fine tuning, dunno.
>
> Signed-off-by: zsugabubus <zsugabubus at national.shitposting.agency>
> ---
> libavformat/dashdec.c | 4871 +++++++++++++++++++++++-----------------
> libavformat/internal.h | 14 +
> libavformat/utils.c | 47 +
> 3 files changed, 2868 insertions(+), 2064 deletions(-)
>
> diff --git a/libavformat/dashdec.c b/libavformat/dashdec.c
> index 6f3f28d..0ee7dd8 100644
> --- a/libavformat/dashdec.c
> +++ b/libavformat/dashdec.c
> @@ -1,7 +1,8 @@
> /*
> - * Dynamic Adaptive Streaming over HTTP demux
> + * Dynamic Adaptive Streaming over HTTP demuxer
> * Copyright (c) 2017 samsamsam at o2.pl based on HLS demux
> * Copyright (c) 2017 Steven Liu
> + * Copyright (c) 2021 reworked by zsugabubus
> *
> * This file is part of FFmpeg.
> *
> @@ -19,182 +20,401 @@
> * License along with FFmpeg; if not, write to the Free Software
> * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
> */
> +
> +/**
> + * @file
> + * Dynamic Adaptive Streaming over HTTP demuxer
> + * @author samsamsam at o2.pl
> + * @author Steven Liu
> + * @author reworked by zsugabubus
> + * @see DASH-IF spec: https://dashif-documents.azurewebsites.net/DASH-IF-IOP/master/DASH-IF-IOP.html
> + * @see another spec: https://dashif-documents.azurewebsites.net/Guidelines-TimingModel/master/Guidelines-TimingModel.html
> + * @see test vectors: https://testassets.dashif.org/
> + * @see MPD scheme: https://standards.iso.org/ittf/PubliclyAvailableStandards/MPEG-DASH_schema_files/DASH-MPD.xsd
> + */
> +
> #include <libxml/parser.h>
> +#include "libavutil/encryption_info.h"
> #include "libavutil/intreadwrite.h"
> +#include "libavutil/lfg.h"
> +#include "libavutil/random_seed.h"
> +#include "libavutil/avassert.h"
> #include "libavutil/opt.h"
> #include "libavutil/time.h"
> #include "libavutil/parseutils.h"
> #include "internal.h"
> #include "avio_internal.h"
> -#include "dash.h"
> +#include "libavutil/base64.h"
>
> -#define INITIAL_BUFFER_SIZE 32768
> -#define MAX_BPRINT_READ_SIZE (UINT_MAX - 1)
> -#define DEFAULT_MANIFEST_SIZE 8 * 1024
> +#if CONFIG_HTTP_PROTOCOL
> +# include "http.h"
> +#endif
>
> -struct fragment {
> - int64_t url_offset;
> - int64_t size;
> - char *url;
> +#define MICROSEC_PER_SEC 1000000L
> +#define MILLISEC_PER_SEC (MICROSEC_PER_SEC / 1000L)
> +
> +static const char CENC_NAMESPACE[] = "urn:mpeg:cenc:2013";
> +
> +enum {
> + /* Most init section <1K, but some DRM systems can use 600 or so bytes for
> + * PSSH data. */
> + INITBUF_MAX = 4 * 1024
> };
>
> -/*
> - * reference to : ISO_IEC_23009-1-DASH-2012
> - * Section: 5.3.9.6.2
> - * Table: Table 17 — Semantics of SegmentTimeline element
> - * */
> -struct timeline {
> - /* starttime: Element or Attribute Name
> - * specifies the MPD start time, in @timescale units,
> - * the first Segment in the series starts relative to the beginning of the Period.
> - * The value of this attribute must be equal to or greater than the sum of the previous S
> - * element earliest presentation time and the sum of the contiguous Segment durations.
> - * If the value of the attribute is greater than what is expressed by the previous S element,
> - * it expresses discontinuities in the timeline.
> - * If not present then the value shall be assumed to be zero for the first S element
> - * and for the subsequent S elements, the value shall be assumed to be the sum of
> - * the previous S element's earliest presentation time and contiguous duration
> - * (i.e. previous S at starttime + @duration * (@repeat + 1)).
> - * */
> - int64_t starttime;
> - /* repeat: Element or Attribute Name
> - * specifies the repeat count of the number of following contiguous Segments with
> - * the same duration expressed by the value of @duration. This value is zero-based
> - * (e.g. a value of three means four Segments in the contiguous series).
> - * */
> - int64_t repeat;
> - /* duration: Element or Attribute Name
> - * specifies the Segment duration, in units of the value of the @timescale.
> - * */
> +#define xml_for_each_attr /* { ... } */ \
> + for (xmlAttrPtr attr = node->properties; \
> + attr; \
> + attr = attr->next) \
> + for (xmlChar *value = xmlNodeGetContent(attr->children); \
> + value; \
> + xml_freep(&value))
> +
> +#define xml_for_each_child /* { ... } */ \
> + for (xmlNodePtr child = xmlFirstElementChild(node); \
> + child; \
> + child = xmlNextElementSibling(child))
> +
> +#define ARRAY_SIZE(x) (sizeof(x) / sizeof(*x))
> +
> +#define DASH_RANGE_INITALIZER (DASHRange){ .start = 0, .end = INT64_MAX }
> +
> +typedef struct {
> + int64_t start;
> + int64_t end;
> +} DASHRange;
> +
> +typedef struct {
> + xmlChar *url;
> + DASHRange range;
> +} DASHSegment;
> +
> +typedef struct {
> + xmlChar *id;
> + int64_t start_ts;
> + int64_t end_ts;
> + AVDictionary *metadata;
> +} DASHPeriod;
> +
> +typedef struct {
> + int64_t start_ts; /* Relative to timeline. */
> + uint64_t number; /**< $Number$ */
> + int64_t repeat; /**< number..=(number + repeat) */
> int64_t duration;
> +} DASHSegmentTemplate;
> +
> +typedef struct {
> + unsigned refcount;
> + unsigned nb;
> + char *elems[];
> +} DASHURLList;
> +
> +typedef struct {
> + AVIOContext *pb; /**< IO context for reading segment. */
> + uint64_t segment_size; /**< Size of segment to read. */
> +} DASHSegmentContext;
> +
> +typedef struct {
> + xmlChar *id;
> +
> + unsigned refcount;
> + AVDictionary *metadata;
> + AVEncryptionInitInfo *init_info; /**< Chain of initialization infos
> + extracted from the manifest. */
> +} DASHAdaptationSet;
> +
> +enum DASHTimelineType {
> + TIMELINE_TEMPLATES, /**< DASHTimeline.u.templates */
> + TIMELINE_SEGMENTS, /**< DASHTimeline.u.segments */
> };
>
> -/*
> - * Each playlist has its own demuxer. If it is currently active,
> - * it has an opened AVIOContext too, and potentially an AVPacket
> - * containing the next packet from this stream.
> - */
> -struct representation {
> - char *url_template;
> - AVIOContext pb;
> - AVIOContext *input;
> - AVFormatContext *parent;
> - AVFormatContext *ctx;
> - int stream_index;
> -
> - char *id;
> - char *lang;
> - int bandwidth;
> - AVRational framerate;
> - AVStream *assoc_stream; /* demuxer stream associated with this representation */
> -
> - int n_fragments;
> - struct fragment **fragments; /* VOD list of fragment for profile */
> -
> - int n_timelines;
> - struct timeline **timelines;
> -
> - int64_t first_seq_no;
> - int64_t last_seq_no;
> - int64_t start_number; /* used in case when we have dynamic list of segment to know which segments are new one*/
> -
> - int64_t fragment_duration;
> - int64_t fragment_timescale;
> -
> - int64_t presentation_timeoffset;
> -
> - int64_t cur_seq_no;
> - int64_t cur_seg_offset;
> - int64_t cur_seg_size;
> - struct fragment *cur_seg;
> -
> - /* Currently active Media Initialization Section */
> - struct fragment *init_section;
> - uint8_t *init_sec_buf;
> - uint32_t init_sec_buf_size;
> - uint32_t init_sec_data_len;
> - uint32_t init_sec_buf_read_offset;
> - int64_t cur_timestamp;
> - int is_restart_needed;
> +typedef struct {
> + unsigned refcount;
> + enum DASHTimelineType type;
> + int64_t duration;
> + int64_t timescale;
> + DASHSegment init;
> + union {
> + struct {
> + unsigned nb;
> + DASHSegment elems[];
> + } segments;
> + struct {
> + unsigned nb;
> + DASHSegment master;
> + DASHSegmentTemplate elems[];
> + } templates;
> + } u;
> +} DASHTimeline;
> +
> +#define DASH_PARAMETERS_INITIALIZER { \
> + .sample_aspect_ratio = { 0, 1 }, \
> + .field_order = AV_FIELD_PROGRESSIVE, \
> +}
> +
> +typedef struct {
> + enum AVMediaType codec_type;
> + enum AVCodecID codec_id;
> + int profile;
> + int level;
> + int disposition;
> +
> + /* Audio only. */
> + int sample_rate;
> + int channels;
> + uint64_t channel_layout;
> + enum AVMatrixEncoding matrix_encoding;
> +
> + /* Video only. */
> + int width;
> + int height;
> + AVRational frame_rate;
> + AVRational sample_aspect_ratio;
> + enum AVFieldOrder field_order;
> +} DASHParameters;
> +
> +typedef struct {
> + DASHPeriod *period;
> + DASHAdaptationSet *as;
> + DASHTimeline *timeline;
> + DASHURLList *base;
> + uint32_t bandwidth;
> +
> + char *initbuf;
> + int initbuf_size; /* Limited by INITBUF_MAX. */
> +
> + DASHParameters par;
> +} DASHRepresentationPeriod;
> +
> +typedef struct {
> + xmlChar *id;
> +
> + AVFormatContext *oc; /* := DASHContext.ctx */
> + AVStream **ostreams; /**< Corresponding output streams. */
> + AVFormatContext *ic;
> + unsigned nb_streams; /**< Should be := ic->nb_streams. If less, subdemuxer
> + created new streams. */
> +
> + AVIOContext ic_pb; /**< Custom IO context for ic subdemuxer. */
> +
> + int save_init; /**< Whether to store what we have read from the current segment. */
> + DASHRepresentationPeriod **periods;
> + unsigned nb_periods;
> + DASHRepresentationPeriod *cur_period;
> +
> + int64_t last_pts;
> + int64_t read_ts;
> +
> + unsigned cur_segment;
> + DASHSegmentContext segments[];
> +} DASHRepresentation;
> +
> +enum DASHUTCTiming {
> + DASH_UTC_TIMING_LOCAL,
> + DASH_UTC_TIMING_HTTP_XSDATE,
> + DASH_UTC_TIMING_HTTP_ISO,
> + DASH_UTC_TIMING_HTTP_NTP,
> + DASH_UTC_TIMING_NTP,
> + DASH_UTC_TIMING_HTTP_HEAD,
> + DASH_UTC_TIMING_DIRECT,
> };
>
> -typedef struct DASHContext {
> +typedef struct {
> const AVClass *class;
> - char *base_url;
> -
> - int n_videos;
> - struct representation **videos;
> - int n_audios;
> - struct representation **audios;
> - int n_subtitles;
> - struct representation **subtitles;
> -
> - /* MediaPresentationDescription Attribute */
> - uint64_t media_presentation_duration;
> - uint64_t suggested_presentation_delay;
> - uint64_t availability_start_time;
> - uint64_t availability_end_time;
> - uint64_t publish_time;
> - uint64_t minimum_update_period;
> - uint64_t time_shift_buffer_depth;
> - uint64_t min_buffer_time;
> -
> - /* Period Attribute */
> - uint64_t period_duration;
> - uint64_t period_start;
> -
> - /* AdaptationSet Attribute */
> - char *adaptionset_lang;
> -
> - int is_live;
> - AVIOInterruptCB *interrupt_callback;
> - char *allowed_extensions;
> - AVDictionary *avio_opts;
> - int max_url_size;
> -
> - /* Flags for init section*/
> - int is_init_section_common_video;
> - int is_init_section_common_audio;
> - int is_init_section_common_subtitle;
>
> + xmlChar *id;
> +
> + char *location;
> + xmlChar *chain_next_location;
> + xmlChar *fallback_location;
> +
> + unsigned nb_connections;
> +
> + DASHPeriod **periods;
> + unsigned nb_periods;
> +
> + DASHAdaptationSet **ass;
> + unsigned nb_ass;
> +
> + DASHRepresentation **reps;
> + unsigned nb_reps;
> +
> + int is_live; /**< Segments may get available over time. */
> + int64_t start_ts;
> +
> + int64_t media_presentation_duration;
> + int64_t suggested_presentation_delay;
> + int64_t availability_start_time;
> + int64_t availability_end_time;
> + int64_t publish_time;
> + int64_t min_update_period;
> + int64_t time_shift_buffer_depth;
> + int64_t min_buffer_time;
> + int64_t max_segment_duration;
> +
> + int64_t next_update;
> +
> + int64_t time_diff;
> + int utc_timing; /* DASH_UTC_TIMING_*; < 0 if updated */
> +
> + AVLFG rnd;
> +
> + AVDictionary *protocol_opts;
> } DASHContext;
>
> -static int ishttp(char *url)
> +typedef struct {
> + xmlChar *scheme_id;
> + xmlChar *value;
> + xmlAttrPtr scheme_id_attr;
> +} DASHScheme;
> +
> +#define DASH_CONTENTPROTECTION_INITIALIZER { .has_key_id = 0 }
> +
> +typedef struct {
> + int has_key_id;
> + uint8_t default_kid[16];
> +} DASHContentProtection;
> +
> +#pragma GCC diagnostic push
> +#pragma GCC diagnostic ignored "-Wstrict-prototypes"
> +static av_cold void av_freep_arrayp(void *ptr, void (*free)())
> +{
> + unsigned *nmemb = (unsigned *)((void **)ptr + 1);
> +
> + for (unsigned i = 0; i < *nmemb; ++i)
> + ((void(*)(void *))free)((*(void ***)ptr)[i]);
> +
> + *nmemb = 0;
> + av_freep(ptr);
> +}
> +#pragma GCC diagnostic pop
> +
> +static void xml_free(xmlChar *p)
> +{
> + if (p)
> + xmlFree(p);
> +}
> +
> +static void xml_freep(xmlChar **pp)
> +{
> + if (*pp) {
> + xmlFree(*pp);
> + *pp = NULL;
> + }
> +}
> +
> +static av_cold DASHURLList *dash_ref_urllist(DASHURLList *urls)
> +{
> + ++urls->refcount;
> + return urls;
> +}
> +
> +static av_cold void dash_unref_urllist(DASHURLList *urls)
> +{
> + if (!urls || --urls->refcount)
> + return;
> +
> + for (unsigned i = 0; i < urls->nb; ++i)
> + av_free(urls->elems[i]);
> + av_free(urls);
> +}
> +
> +static int dash_urllist_has(const DASHURLList *urls, const char *url)
> {
> - const char *proto_name = avio_find_protocol_name(url);
> - return proto_name && av_strstart(proto_name, "http", NULL);
> + for (unsigned i = 0; i < urls->nb; ++i)
> + if (!strcmp(urls->elems[i], url))
> + return 1;
> +
> + return 0;
> +}
> +
> +static av_cold void dash_free_period(DASHPeriod *period)
> +{
> + if (!period)
> + return;
> +
> + xml_free(period->id);
> + av_dict_free(&period->metadata);
> + av_free(period);
> +}
> +
> +static av_cold void dash_free_adaptationset(DASHAdaptationSet *as)
> +{
> + if (!as)
> + return;
> +
> + xml_free(as->id);
> + av_encryption_init_info_free(as->init_info);
> + av_dict_free(&as->metadata);
> + av_free(as);
> +}
> +
> +static av_cold DASHTimeline *dash_ref_timeline(DASHTimeline *timeline)
> +{
> + ++timeline->refcount;
> + return timeline;
> +}
> +
> +static void dash_free_segment(DASHSegment *g)
> +{
> + xml_free(g->url);
> +}
> +
> +static void dash_unref_timeline(DASHTimeline *timeline)
> +{
> + if (!timeline || --timeline->refcount)
> + return;
> +
> + dash_free_segment(&timeline->init);
> +
> + if (TIMELINE_SEGMENTS == timeline->type) {
> + for (unsigned i = 0; i < timeline->u.segments.nb; ++i)
> + dash_free_segment(&timeline->u.segments.elems[i]);
> + } else if (TIMELINE_TEMPLATES == timeline->type) {
> + dash_free_segment(&timeline->u.templates.master);
> + } else {
> + abort();
> + }
> +
> + av_free(timeline);
> }
>
> -static int aligned(int val)
> +/**
> + * @return |wall_time| / MICROSEC_PER_SEC * AV_TIME_BASE
> + */
> +static int64_t av_wall2timebase(int64_t wall_time)
> {
> - return ((val + 0x3F) >> 6) << 6;
> + if (MICROSEC_PER_SEC < AV_TIME_BASE)
> + return wall_time * (AV_TIME_BASE / MICROSEC_PER_SEC);
> + else
> + return wall_time / (MICROSEC_PER_SEC / AV_TIME_BASE);
> }
>
> -static uint64_t get_current_time_in_sec(void)
> +/**
> + * @return wall clock time in AV_TIME_BASE
> + */
> +static int64_t dash_gettime(AVFormatContext *s)
> {
> - return av_gettime() / 1000000;
> + DASHContext *c = s->priv_data;
> + return av_wall2timebase(av_gettime()) + c->time_diff;
> }
>
> -static uint64_t get_utc_date_time_insec(AVFormatContext *s, const char *datetime)
> +/**
> + * Parse ISO-8601 date.
> + * @return time in AV_TIME_BASE
> + */
> +static uint64_t dash_parse_date(const char *s, AVFormatContext *log_ctx)
> {
> struct tm timeinfo;
> - int year = 0;
> - int month = 0;
> - int day = 0;
> - int hour = 0;
> - int minute = 0;
> - int ret = 0;
> + unsigned year = 0, month = 0, day = 0;
> + unsigned hour = 0, minute = 0;
> float second = 0.0;
>
> - /* ISO-8601 date parser */
> - if (!datetime)
> - return 0;
> + if (6 != av_sscanf(s, "%u-%u-%uT%u:%u:%fZ",
> + &year, &month, &day,
> + &hour, &minute, &second))
> + av_log(log_ctx, AV_LOG_WARNING, "Malformed ISO-8601 date: '%s'\n", s);
> + /* Fall-through. */
>
> - ret = sscanf(datetime, "%d-%d-%dT%d:%d:%fZ", &year, &month, &day, &hour, &minute, &second);
> - /* year, month, day, hour, minute, second 6 arguments */
> - if (ret != 6) {
> - av_log(s, AV_LOG_WARNING, "get_utc_date_time_insec get a wrong time format\n");
> - }
> timeinfo.tm_year = year - 1900;
> timeinfo.tm_mon = month - 1;
> timeinfo.tm_mday = day;
> @@ -205,2189 +425,2711 @@ static uint64_t get_utc_date_time_insec(AVFormatContext *s, const char *datetime
> return av_timegm(&timeinfo);
> }
>
> -static uint32_t get_duration_insec(AVFormatContext *s, const char *duration)
> -{
> - /* ISO-8601 duration parser */
> - uint32_t days = 0;
> - uint32_t hours = 0;
> - uint32_t mins = 0;
> - uint32_t secs = 0;
> - int size = 0;
> - float value = 0;
> - char type = '\0';
> - const char *ptr = duration;
> -
> - while (*ptr) {
> - if (*ptr == 'P' || *ptr == 'T') {
> - ptr++;
> +/**
> + * Parse ISO-8601 duration.
> + * @return duration in AV_TIME_BASE
> + */
> +static int64_t dash_parse_duration(const char *s, AVFormatContext *log_ctx)
> +{
> + enum {
> + STATE_GROUND,
> + STATE_DATE,
> + STATE_TIME,
> + } state = STATE_GROUND;
> + int64_t ret = 0;
> +
> + for (const char *p = s; *p; ++p) {
> + int len;
> + int64_t unit = 0;
> + /* We require double precision since it is absolutely valid to pass a
> + * huge integer solely in seconds. */
> + double base;
> +
> + switch (*p) {
> + case 'P':
> + if (state != STATE_GROUND)
> + goto invalid;
> + state = STATE_DATE;
> + continue;
> + case 'T':
> + if (state != STATE_DATE)
> + goto invalid;
> + state = STATE_TIME;
> continue;
> }
>
> - if (sscanf(ptr, "%f%c%n", &value, &type, &size) != 2) {
> - av_log(s, AV_LOG_WARNING, "get_duration_insec get a wrong time format\n");
> - return 0; /* parser error */
> + av_sscanf(p, "%lf%n", &base, &len);
> + p += len;
> +
> + switch (state) {
> + case STATE_DATE:
> + switch (*p) {
> + case 'Y':
> + unit = (int64_t)AV_TIME_BASE * 60 * 60 * 24 * 365;
> + break;
> + case 'M':
> + unit = (int64_t)AV_TIME_BASE * 60 * 60 * 24 * 30;
> + break;
> + case 'D':
> + unit = (int64_t)AV_TIME_BASE * 60 * 60 * 24;
> + break;
> + }
> + break;
> + case STATE_TIME:
> + switch (*p) {
> + case 'H':
> + unit = (int64_t)AV_TIME_BASE * 60 * 60;
> + break;
> + case 'M':
> + unit = (int64_t)AV_TIME_BASE * 60;
> + break;
> + case 'S':
> + unit = (int64_t)AV_TIME_BASE;
> + break;
> + }
> + break;
> + }
> +
> + if (!unit) {
> + invalid:
> + av_log(log_ctx, AV_LOG_ERROR, "Malformed ISO-8601 duration: '%s'\n", s);
> + return INT64_MIN;
> }
> +
> + ret += unit * base;
> + }
> +
> + return ret;
> +}
> +
> +static av_cold void dash_subdemuxer_flush(DASHRepresentation *rep)
> +{
> + DASHContext *c = rep->oc->priv_data;
> +
> + avio_flush(&rep->ic_pb);
> + for (unsigned i = 0; i < c->nb_connections; ++i)
> + ff_format_io_close(rep->oc, &rep->segments[i].pb);
> +
> +
> + /* for (unsigned i = 0; i < rep->nb_streams; ++i)
> + avcodec_flush_buffers(rep->ostreams[i]->internal->avctx); */
> +
> + if (rep->ic) {
> + /* for (unsigned i = 0; i < rep->ic->nb_streams; ++i)
> + avcodec_flush_buffers(rep->ic->streams[i]->internal->avctx); */
> + avformat_flush(rep->ic);
> + }
> +
> + /* If receiving init section just has been aborted, we have to forget all
> + * received bytes otherwise at subdemuxer restart on the first read it will
> + * be falsely detected that we have received the complete section because
> + * stored init section size is less than maximum storable. */
> + if (rep->save_init) {
> + rep->save_init = 0;
> + if (rep->cur_period) {
> + rep->cur_period->initbuf_size = 0;
> + rep->cur_period = NULL;
> + }
> + }
> +}
> +
> +static av_cold void dash_subdemuxer_close(DASHRepresentation *rep)
> +{
> + av_log(rep->oc, AV_LOG_DEBUG, "Closing representation '%s'\n", rep->id);
> +
> + /* Only need to deallocate the used array. */
> + av_freep(&rep->ic_pb.buffer);
> +
> + dash_subdemuxer_flush(rep);
> +
> + if (rep->ic) {
> + /* IO context is part of our context so avoid freeing it. */
> + rep->ic->pb = NULL;
> + rep->cur_period = NULL;
> +
> + if (rep->ic->iformat) {
> + avformat_close_input(&rep->ic);
> + } else {
> + avformat_free_context(rep->ic);
> + rep->ic = NULL;
> + }
> + }
> +}
> +
> +static av_cold void dash_free_representationperiod(DASHRepresentationPeriod *period)
> +{
> + dash_unref_timeline(period->timeline);
> + dash_unref_urllist(period->base);
> + av_free(period->initbuf);
> + av_free(period);
> +}
> +
> +static av_cold void dash_free_representation(DASHRepresentation *rep)
> +{
> + dash_subdemuxer_close(rep);
> + av_free(rep->ostreams);
> + av_freep_arrayp(&rep->periods, dash_free_representationperiod);
> + xml_free(rep->id);
> + av_free(rep);
> +}
> +
> +static av_cold xmlNodePtr xml_get_nsnode(xmlNodePtr node, const char *nodename, const char *namespace)
> +{
> + xml_for_each_child
> + if (!strcmp(child->name, nodename) &&
> + (child->ns && !strcmp(child->ns->href, namespace)))
> + return child;
> + return NULL;
> +}
> +
> +static av_cold int dash_parse_contentprotection_pssh(AVFormatContext *s,
> + struct AVEncryptionInitInfo *info,
> + xmlNodePtr node)
> +{
> + int ret = 0;
> + xmlNodePtr pssh_node;
> + int size;
> + char *pssh = NULL;
> + int decoded_size;
> + uint8_t *decoded_pssh;
> +
> + av_free(info->data);
> + info->data = NULL;
> + info->data_size = 0;
> +
> + if (!(pssh_node = xml_get_nsnode(node, "pssh", CENC_NAMESPACE)))
> + goto out;
> +
> + if (!(pssh = xmlNodeGetContent(pssh_node)))
> + goto invalid_content;
> +
> + size = strlen(pssh);
> + decoded_size = AV_BASE64_DECODE_SIZE(size);
> +
> + if (!(decoded_pssh = av_malloc(decoded_size))) {
> + ret = AVERROR(ENOMEM);
> + goto out;
> + }
> +
> + if ((decoded_size = av_base64_decode(decoded_pssh, pssh, decoded_size)) < 0) {
> + av_free(decoded_pssh);
> +
> + invalid_content:
> + av_log(s, AV_LOG_ERROR, "Invalid %s/{%s}:pssh content\n",
> + node->name, CENC_NAMESPACE);
> + ret = AVERROR_INVALIDDATA;
> + goto out;
> + }
> +
> + info->data = decoded_pssh;
> + info->data_size = decoded_size;
> +
> +out:
> + xml_free(pssh);
> + return ret;
> +}
> +
> +static av_cold void dash_free_scheme(DASHScheme *scheme)
> +{
> + xml_free(scheme->scheme_id);
> + xml_free(scheme->value);
> +}
> +
> +static av_cold void dash_log_invalid_attr_value(AVFormatContext *log_ctx, xmlAttrPtr attr, const char *value)
> +{
> + av_log(log_ctx, AV_LOG_ERROR, "Invalid %s/%s@%s%s%s%s='%s'\n",
> + attr->parent->parent->name,
> + attr->parent->ns ? "{" : "",
> + attr->parent->ns ? (const char *)attr->parent->ns->href : "",
> + attr->parent->ns ? "}" : "",
> + attr->parent->name,
> + attr->name,
> + value);
> +}
> +
> +static av_cold void dash_log_missing_attr(AVFormatContext *log_ctx, xmlNodePtr node, const char *attr)
> +{
> + av_log(log_ctx, AV_LOG_ERROR, "Missing %s/%s@%s\n",
> + node->parent->name,
> + node->name,
> + attr);
> +}
> +
> +static av_cold void dash_log_unknown_child(AVFormatContext *log_ctx, xmlNodePtr child)
> +{
> + av_log(log_ctx, AV_LOG_WARNING, "Unknown %s/%s\n",
> + child->parent->name,
> + child->name);
> +}
> +
> +static av_cold void dash_log_unknown_attr(AVFormatContext *log_ctx, xmlAttrPtr attr, const xmlChar *value)
> +{
> + av_log(log_ctx, AV_LOG_WARNING, "Unknown %s%s%s@%s%s%s%s='%s'\n",
> + attr->parent->parent->name ? (const char *)attr->parent->parent->name : "",
> + attr->parent->parent->name ? "/" : "",
> + attr->parent->name,
> + attr->ns ? "{" : "",
> + attr->ns ? (const char *)attr->ns->href : "",
> + attr->ns ? "}:" : "",
> + attr->name,
> + value);
> +}
> +
> +static av_cold void dash_log_unknown_scheme(AVFormatContext *log_ctx, const DASHScheme *scheme)
> +{
> + av_log(log_ctx, AV_LOG_WARNING,
> + "Unknown %s/%s@%s='%s' (@value=%c%s%c)\n",
> + scheme->scheme_id_attr->parent->parent->name,
> + scheme->scheme_id_attr->parent->name,
> + scheme->scheme_id_attr->name,
> + scheme->scheme_id,
> + scheme->value ? '\'' : '(',
> + scheme->value ? (const char *)scheme->value : "not present",
> + scheme->value ? '\'' : ')');
> +}
> +
> +static av_cold int dash_parse_scheme(AVFormatContext *s, DASHScheme *scheme, int value_required, xmlNodePtr node)
> +{
> + scheme->value = NULL;
> +
> + if (!(scheme->scheme_id_attr = xmlHasNsProp(node, "schemeIdUri", NULL))) {
> + dash_log_missing_attr(s, node, "schemeIdUri");
> + return AVERROR_INVALIDDATA;
> + }
> +
> + if (!(scheme->scheme_id = xmlNodeGetContent(scheme->scheme_id_attr->children)))
> + return AVERROR(ENOMEM);
> +
> + if (!(scheme->value = xmlGetNoNsProp(node, "value")) && value_required) {
> + dash_log_missing_attr(s, node, "value");
> + return AVERROR_INVALIDDATA;
> + }
> +
> + return 0;
> +}
> +
> +static av_cold int dash_parse_contentprotection(AVFormatContext *s,
> + DASHAdaptationSet *as,
> + DASHContentProtection *cp,
> + xmlNodePtr node)
> +{
> + int ret = 0;
> + AVEncryptionInitInfo *init_info = NULL;
> + const char *uuid;
> + DASHScheme scheme;
> +
> + if ((ret = dash_parse_scheme(s, &scheme, 0, node)) < 0)
> + goto out;
> +
> + /* Parse Common Encryption element. */
> + if (!cp->has_key_id &&
> + !strcmp(scheme.scheme_id, "urn:mpeg:dash:mp4protection:2011") &&
> + (scheme.value && !strcmp(scheme.value, "cenc")))
> + {
> + char *key_id;
> +
> + if ((key_id = xmlGetNsProp(node, "default_KID", CENC_NAMESPACE))) {
> + if (ff_uuid_to_data(cp->default_kid, key_id) < 0)
> + av_log(s, AV_LOG_ERROR, "Malformed UUID in ContentProtection at cenc:default_KID='%s'\n",
> + scheme.scheme_id);
> + else
> + cp->has_key_id = 1;
> +
> + xmlFree(key_id);
> + } else {
> + dash_log_missing_attr(s, node, "default_KID");
> + }
> +
> + goto out;
> + } else if (cp->has_key_id && av_strstart(scheme.scheme_id, "urn:uuid:", &uuid)) {
> + /* Parse an UUID schema. */
> + init_info = av_encryption_init_info_alloc(/* system_id_size */ 16, 1,
> + /* key_id_size */ 16, /* data_size */ 0,
> + /* extra_data_size */ 0);
Hi zsugabubus,
This line will break the compling work.
CC libavformat/dashdec.o
src/libavformat/dashdec.c:748:73: error: too many arguments to
function call, expected 4, have 5
/* extra_data_size */ 0);
^
src/libavutil/encryption_info.h:176:1: note:
'av_encryption_init_info_alloc' declared here
AVEncryptionInitInfo *av_encryption_init_info_alloc(
^
src/libavformat/dashdec.c:1777:86: warning: incompatible pointer types
passing 'int *' to parameter of type 'size_t *'
(aka 'unsigned long *') [-Wincompatible-pointer-types]
side_data =
av_encryption_init_info_add_side_data(period->as->init_info,
&side_data_size);
^~~~~~~~~~~~~~~
src/libavutil/encryption_info.h:203:47: note: passing argument to
parameter 'side_data_size' here
const AVEncryptionInitInfo *info, size_t *side_data_size);
^
1 warning and 1 error generated.
make: *** [libavformat/dashdec.o] Error 1
> + if (!init_info) {
> + ret = AVERROR(ENOMEM);
> + goto out;
> + }
> +
> + if (ff_uuid_to_data(init_info->system_id, uuid) < 0) {
> + av_log(s, AV_LOG_ERROR, "Malformed UUID in ContentProtection at schemeIdUri='%s'\n",
> + scheme.scheme_id);
> + goto out;
> + }
> +
> + if (0 <= ret)
> + ret = dash_parse_contentprotection_pssh(s, init_info, node);
> + if (ret < 0)
> + goto out;
> +
> + memcpy(init_info->key_ids[0], cp->default_kid, init_info->key_id_size);
> +
> + init_info->next = as->init_info;
> + as->init_info = init_info;
> + init_info = NULL;
> + } else {
> + dash_log_unknown_scheme(s, &scheme);
> + goto out;
> + }
> +
> +out:
> + if (init_info)
> + av_encryption_init_info_free(init_info);
> + dash_free_scheme(&scheme);
> +
> + return ret;
> +}
> +
> +static av_cold int dash_set_location(AVFormatContext *s, char const *url)
> +{
> + DASHContext *c = s->priv_data;
> + URLComponents uc;
> + char const *protocol;
> +
> + if (ff_url_decompose(&uc, url, NULL) < 0)
> + return AVERROR_INVALIDDATA;
> +
> + av_free(c->location);
> + if (!(c->location = av_strdup(url)))
> + return AVERROR(ENOMEM);
> +
> + protocol = avio_find_protocol_name(c->location);
> + if (strstr(protocol, "http")) {
> + char *p = strstr(uc.query, "t=");
> + if (p) {
> + int64_t seconds;
> +
> + p += 2;
> + if (1 == av_sscanf(p, "posix:%"SCNd64, &seconds))
> + c->start_ts = AV_TIME_BASE * seconds;
> + else if (!strncmp(p, "now", 4))
> + c->start_ts = dash_gettime(s);
> + else
> + c->start_ts = dash_parse_date(p, s);
> + }
> + }
> +
> + return 0;
> +}
> +
> +static av_cold int dash_parse_location(AVFormatContext *s,
> + xmlNodePtr node)
> +{
> + int ret = 0;
> + char *value;
> + DASHContext *c = s->priv_data;
> +
> + if (!c->is_live) {
> + av_log(s, AV_LOG_INFO, "MPD at type=\"static\"/Location ignored\n");
> + return 0;
> + }
> +
> + if (!(value = xmlNodeGetContent(node)))
> + return AVERROR(ENOMEM);
> +
> + ret = dash_set_location(s, value);
> +
> + xmlFree(value);
> + return ret;
> +}
> +
> +/**
> + * Like ff_make_absolute_urla() but allocates a big enough memory.
> + * *out_url shall always be av_free()d regard of error.
> + * */
> +static int ff_make_absolute_urla(char **out_url, const char *base, const char *rel)
> +{
> + int out_url_max_size = strlen(base) + strlen(rel) + 1/* NUL */;
> + if (!(*out_url = av_malloc(out_url_max_size)))
> + return AVERROR(ENOMEM);
> +
> + return ff_make_absolute_url(*out_url, out_url_max_size, base, rel);
> +}
> +
> +static av_cold DASHURLList *dash_new_urllist(unsigned nb_urls)
> +{
> + DASHURLList *urls;
> +
> + if (!(urls = av_mallocz(offsetof(DASHURLList, elems[nb_urls]))))
> + return NULL;
> +
> + dash_ref_urllist(urls);
> + urls->nb = nb_urls;
> +
> + return urls;
> +}
> +
> +static av_cold int dash_parse_baseurl(AVFormatContext *s,
> + DASHURLList **urls,
> + xmlNodePtr node)
> +{
> + int ret = 0;
> + DASHContext *c = s->priv_data;
> + char *url_chunk = NULL;
> + char *resolved_url = NULL;
> + DASHURLList *new_urls = NULL;
> + int is_absolute;
> + char *base_url;
> +
> + if (!(url_chunk = xmlNodeGetContent(node)))
> + return AVERROR(ENOMEM);
> +
> + base_url = (*urls)->elems[0];
> + if ((ret = ff_make_absolute_urla(&resolved_url, base_url, url_chunk)) < 0)
> + goto out;
> +
> + if (dash_urllist_has(*urls, resolved_url))
> + goto out;
> +
> + is_absolute = !strcmp(resolved_url, url_chunk);
> +
> + if (1 == (*urls)->refcount) {
> + /* If we own the instance, it means that this node is an alternative
> + * BaseURL. */
> + if (is_absolute) {
> + void *p;
> +
> + if ((p = av_realloc(*urls, offsetof(DASHURLList, elems[(*urls)->nb + 1])))) {
> + /* Put the new URL at a random place. Absolute addresses most
> + * commonly mean different servers/CDNs, so we can help a
> + * little bit on load-balancing. */
> + unsigned to_index;
> +
> + *urls = p;
> + to_index = av_lfg_get(&c->rnd) % ((*urls)->nb + 1);
> + (*urls)->elems[(*urls)->nb++] = (*urls)->elems[to_index];
> + (*urls)->elems[to_index] = resolved_url, resolved_url = NULL;
> + }
> + } else {
> + /* We do not want to explode URL list so we ignore relative
> + * alternative URLs. Also, using different paths on the same
> + * server does not really makes sense. */
> + }
> + } else {
> + /* Either add a single absolute URL to the list or in case of a
> + * relative BaseURL combine it with every parent URL. */
> + if (!(new_urls = dash_new_urllist(is_absolute ? 1 : (*urls)->nb))) {
> + ret = AVERROR(ENOMEM);
> + goto out;
> + }
> +
> + /* We have already done the first one at the top. */
> + new_urls->elems[0] = resolved_url, resolved_url = NULL;
> + if (!is_absolute) {
> + for (unsigned i = 1; i < (*urls)->nb; ++i)
> + if (!(ret = ff_make_absolute_urla(&new_urls->elems[i], (*urls)->elems[i], url_chunk)))
> + goto out;
> + }
> +
> + /* Replace URL list of current level with the modified one. */
> + dash_unref_urllist(*urls);
> + *urls = new_urls;
> + new_urls = NULL;
> + }
> +
> +out:
> + xmlFree(url_chunk);
> + av_free(resolved_url);
> + dash_unref_urllist(new_urls);
> + return ret;
> +}
> +
> +static av_cold int dash_parse_mimetype(DASHParameters *par, const char *value)
> +{
> + if (!strncmp(value, "video/", 6))
> + par->codec_type = AVMEDIA_TYPE_VIDEO;
> + else if (!strncmp(value, "audio/", 6))
> + par->codec_type = AVMEDIA_TYPE_AUDIO;
> + else if (!strncmp(value, "text/", 5))
> + par->codec_type = AVMEDIA_TYPE_SUBTITLE;
> + else
> + return 1;
> +
> + return 0;
> +}
> +
> +static av_cold int dash_parse_contenttype(AVFormatContext *s,
> + DASHParameters *par,
> + const xmlChar *value)
> +{
> + if (!strcmp(value, "video"))
> + par->codec_type = AVMEDIA_TYPE_VIDEO;
> + else if (!strcmp(value, "audio"))
> + par->codec_type = AVMEDIA_TYPE_AUDIO;
> + else {
> + par->codec_type = AVMEDIA_TYPE_UNKNOWN;
> + return 1;
> + }
> +
> + return 0;
> +}
> +
> +static av_cold int dash_parse_codecs(AVFormatContext *s,
> + DASHParameters *par,
> + const xmlChar *value)
> +{
> + int n;
> + char type;
> + int flags;
> +
> + /* https://developer.mozilla.org/en-US/docs/Web/Media/Formats/codecs_parameter */
> + /* https://tools.ietf.org/html/rfc6381 */
> + if (3 == av_sscanf(value, "mp4%c.%d.%d%n", &type, &par->level, &par->profile, &n) &&
> + !value[n])
> + {
> switch (type) {
> - case 'D':
> - days = (uint32_t)value;
> - break;
> - case 'H':
> - hours = (uint32_t)value;
> - break;
> - case 'M':
> - mins = (uint32_t)value;
> - break;
> - case 'S':
> - secs = (uint32_t)value;
> - break;
> - default:
> - // handle invalid type
> - break;
> + case 'a': par->codec_type = AVMEDIA_TYPE_AUDIO; break;
> + case 'v': par->codec_type = AVMEDIA_TYPE_VIDEO; break;
> + default: return 1;
> }
> - ptr += size;
> +
> + } else if (3 == av_sscanf(value, "avc1"/* avcoti */".%02x%02x%02x%n",
> + &par->profile, &flags, &par->level, &n) &&
> + !value[n])
> + {
> + par->codec_type = AVMEDIA_TYPE_VIDEO;
> + par->codec_id = AV_CODEC_ID_H264;
> + par->profile |= (unsigned)flags << 7;
> + } else if (av_sscanf(value, "hev1.%n", &n), n == 5)
> + {
> + par->codec_id = AV_CODEC_ID_HEVC;
> + } else {
> + par->codec_id = AV_CODEC_ID_NONE;
> + return 1;
> }
> - return ((days * 24 + hours) * 60 + mins) * 60 + secs;
> +
> + return 0;
> }
>
> -static int64_t get_segment_start_time_based_on_timeline(struct representation *pls, int64_t cur_seq_no)
> +static av_cold int dash_parse_scantype(AVFormatContext *s,
> + DASHParameters *par,
> + const xmlChar *value)
> {
> - int64_t start_time = 0;
> - int64_t i = 0;
> - int64_t j = 0;
> - int64_t num = 0;
> -
> - if (pls->n_timelines) {
> - for (i = 0; i < pls->n_timelines; i++) {
> - if (pls->timelines[i]->starttime > 0) {
> - start_time = pls->timelines[i]->starttime;
> - }
> - if (num == cur_seq_no)
> - goto finish;
> + if (!strcmp(value, "progressive")) {
> + par->field_order = AV_FIELD_PROGRESSIVE;
> + } else {
> + par->field_order = AV_FIELD_UNKNOWN;
> + return 1;
> + }
>
> - start_time += pls->timelines[i]->duration;
> + return 0;
> +}
>
> - if (pls->timelines[i]->repeat == -1) {
> - start_time = pls->timelines[i]->duration * cur_seq_no;
> - goto finish;
> - }
> +static av_cold int dash_parse_parameters(AVFormatContext *s,
> + DASHParameters *par,
> + xmlAttrPtr attr,
> + const xmlChar *value)
> +{
> + if (!strcmp(attr->name, "height"))
> + (void)av_sscanf(value, "%d", &par->height);
> + else if (!strcmp(attr->name, "width"))
> + (void)av_sscanf(value, "%d", &par->width);
> + else if (!strcmp(attr->name, "mimeType"))
> + return dash_parse_mimetype(par, value);
> + else if (!strcmp(attr->name, "audioSamplingRate"))
> + (void)av_sscanf(value, "%d", &par->sample_rate);
> + else if (!strcmp(attr->name, "par"))
> + /* Ignore. */;
> + else if (!strcmp(attr->name, "sar"))
> + (void)av_parse_ratio(&par->sample_aspect_ratio, value, INT_MAX, AV_LOG_ERROR, s);
> + else if (!strcmp(attr->name, "frameRate"))
> + (void)av_sscanf(value, "%d", &par->frame_rate);
> + else if (!strcmp(attr->name, "codecs"))
> + return dash_parse_codecs(s, par, value);
> + else if (!strcmp(attr->name, "scanType"))
> + return dash_parse_scantype(s, par, value);
> + else if (!strcmp(attr->name, "contentType"))
> + return dash_parse_contenttype(s, par, value);
> + else
> + return 1;
> +
> + return 0;
> +}
>
> - for (j = 0; j < pls->timelines[i]->repeat; j++) {
> - num++;
> - if (num == cur_seq_no)
> - goto finish;
> - start_time += pls->timelines[i]->duration;
> - }
> - num++;
> - }
> +static av_cold int dash_parse_audiochannelconfiguration(AVFormatContext *s,
> + DASHParameters *par,
> + xmlNodePtr node)
> +{
> + int ret = 0;
> + DASHScheme scheme;
> +
> + if ((ret = dash_parse_scheme(s, &scheme, 1, node)) < 0)
> + goto out;
> +
> + par->channels = 0;
> + par->channel_layout = 0;
> + par->matrix_encoding = AV_MATRIX_ENCODING_NONE;
> +
> + /* https://testassets.dashif.org/#testvector/details/586fb3879ae9045678eacd10 */
> + if (!strcmp(scheme.scheme_id, "urn:dolby:dash:audio_channel_configuration:2011") ||
> + !strcmp(scheme.scheme_id, "tag:dolby.com,2014:dash:audio_channel_configuration:2011"))
> + {
> + par->matrix_encoding = AV_MATRIX_ENCODING_DOLBY;
> + (void)av_get_extended_channel_layout(scheme.value, &par->channel_layout, &par->channels);
> + } else if (!strcmp(scheme.scheme_id, "urn:mpeg:dash:23003:3:audio_channel_configuration:2011")) {
> + (void)av_sscanf(scheme.value, "%d", &par->channels);
> + par->channel_layout = av_get_default_channel_layout(par->channels);
> + } else {
> + (void)av_get_extended_channel_layout(scheme.value, &par->channel_layout, &par->channels);
> + dash_log_unknown_scheme(s, &scheme);
> }
> -finish:
> - return start_time;
> +
> +out:
> + dash_free_scheme(&scheme);
> + return ret;
> }
>
> -static int64_t calc_next_seg_no_from_timelines(struct representation *pls, int64_t cur_time)
> +static av_cold int dash_parse_assetidentifier(AVFormatContext *s,
> + DASHPeriod *period,
> + xmlNodePtr node)
> {
> - int64_t i = 0;
> - int64_t j = 0;
> - int64_t num = 0;
> - int64_t start_time = 0;
> + int ret = 0;
> + DASHScheme scheme;
>
> - for (i = 0; i < pls->n_timelines; i++) {
> - if (pls->timelines[i]->starttime > 0) {
> - start_time = pls->timelines[i]->starttime;
> - }
> - if (start_time > cur_time)
> - goto finish;
> -
> - start_time += pls->timelines[i]->duration;
> - for (j = 0; j < pls->timelines[i]->repeat; j++) {
> - num++;
> - if (start_time > cur_time)
> - goto finish;
> - start_time += pls->timelines[i]->duration;
> - }
> - num++;
> - }
> + if ((ret = dash_parse_scheme(s, &scheme, 1, node)) < 0)
> + goto out;
>
> - return -1;
> + if (!strcmp(scheme.scheme_id, "urn:org:dashif:asset-id:2013")) {
> + av_dict_set(&period->metadata, "asset_identifier", scheme.value, 0);
> + } else {
> + dash_log_unknown_scheme(s, &scheme);
> + }
>
> -finish:
> - return num;
> +out:
> + dash_free_scheme(&scheme);
> + return ret;
> }
>
> -static void free_fragment(struct fragment **seg)
> +static av_cold int dash_parse_viewpoint(AVFormatContext *s,
> + DASHAdaptationSet *as,
> + xmlNodePtr node)
> {
> - if (!(*seg)) {
> - return;
> + int ret = 0;
> + DASHScheme scheme;
> +
> + if ((ret = dash_parse_scheme(s, &scheme, 1, node)) < 0)
> + goto out;
> +
> + /* https://testassets.dashif.org/#testvector/details/5cde78e4a5eeda55aa663101 */
> + if (!strcmp(scheme.scheme_id, "urn:mpeg:dash:viewpoint:2011")) {
> + av_dict_set(&as->metadata, "viewpoint", scheme.value, 0);
> + } else {
> + dash_log_unknown_scheme(s, &scheme);
> }
> - av_freep(&(*seg)->url);
> - av_freep(seg);
> +
> +out:
> + dash_free_scheme(&scheme);
> + return ret;
> }
>
> -static void free_fragment_list(struct representation *pls)
> +static av_cold void dash_sync_time(AVFormatContext *s, const xmlChar *value)
> {
> - int i;
> + DASHContext *c = s->priv_data;
> + int ret = 0;
> +
> + switch (c->utc_timing) {
> + case DASH_UTC_TIMING_LOCAL:
> + default:
> + ret = 1;
> + break;
> + case DASH_UTC_TIMING_HTTP_XSDATE:
> + case DASH_UTC_TIMING_HTTP_ISO:
> + case DASH_UTC_TIMING_HTTP_NTP:
> + case DASH_UTC_TIMING_NTP:
> + case DASH_UTC_TIMING_HTTP_HEAD:
> + case DASH_UTC_TIMING_DIRECT:
> + ret = AVERROR_PATCHWELCOME;
> + break;
> + }
>
> - for (i = 0; i < pls->n_fragments; i++) {
> - free_fragment(&pls->fragments[i]);
> + if (!ret) {
> + av_log(s, AV_LOG_DEBUG, "Time synchronized: %lf s\n",
> + (double)c->time_diff / AV_TIME_BASE);
> + c->utc_timing = -c->utc_timing;
> + } else if (ret < 0) {
> + av_log(s, AV_LOG_ERROR, "Failed to synchronized time: %s\n",
> + av_err2str(ret));
> }
> - av_freep(&pls->fragments);
> - pls->n_fragments = 0;
> }
>
> -static void free_timelines_list(struct representation *pls)
> +static av_cold int dash_parse_utctiming(AVFormatContext *s,
> + xmlNodePtr node)
> {
> - int i;
> + int ret = 0;
> + DASHContext *c = s->priv_data;
> + DASHScheme scheme;
> + int utc_timing = c->utc_timing;
> +
> + if ((ret = dash_parse_scheme(s, &scheme, 0, node)) < 0)
> + goto out;
> +
> + c->utc_timing = DASH_UTC_TIMING_LOCAL;
> +
> + if (0);
> +#define PARSE(name, scheme_name) \
> + else if (!strcmp(scheme.scheme_id, "urn:mpeg:dash:utc:"scheme_name":2014")) \
> + c->utc_timing = DASH_UTC_TIMING_##name
> + PARSE(HTTP_XSDATE, "http-xsdate");
> + PARSE(HTTP_ISO, "http-iso");
> + PARSE(HTTP_NTP, "http-ntp");
> + PARSE(NTP, "ntp");
> + PARSE(HTTP_HEAD, "http-head");
> + PARSE(DIRECT, "direct");
> +#undef PARSE
> + else
> + dash_log_unknown_scheme(s, &scheme);
> +
> + if (c->utc_timing == -utc_timing)
> + c->utc_timing = utc_timing;
> +
> + dash_sync_time(s, scheme.value);
> +
> +out:
> + dash_free_scheme(&scheme);
> + return ret;
> +}
>
> - for (i = 0; i < pls->n_timelines; i++) {
> - av_freep(&pls->timelines[i]);
> +static av_cold int dash_parse_role(AVFormatContext *s,
> + DASHParameters *par,
> + xmlNodePtr node)
> +{
> + int ret = 0;
> + DASHScheme scheme;
> +
> + if ((ret = dash_parse_scheme(s, &scheme, 1, node)) < 0)
> + goto out;
> +
> + /* https://testassets.dashif.org/#feature/details/588a48c27459f8cb201b881b */
> + if (!strcmp(scheme.scheme_id, "urn:mpeg:dash:role:2011")) {
> + if (!strcmp(scheme.value, "main"))
> + par->disposition |= AV_DISPOSITION_DEFAULT;
> + else if (!strcmp(scheme.value, "alternate"))
> + par->disposition &= ~AV_DISPOSITION_DEFAULT;
> + else if (!strcmp(scheme.value, "original"))
> + par->disposition |= AV_DISPOSITION_ORIGINAL;
> + else if (!strcmp(scheme.value, "dub"))
> + par->disposition |= AV_DISPOSITION_DUB;
> + else if (!strcmp(scheme.value, "subtitle"))
> + par->codec_type = AVMEDIA_TYPE_SUBTITLE;
> + } else {
> + dash_log_unknown_scheme(s, &scheme);
> }
> - av_freep(&pls->timelines);
> - pls->n_timelines = 0;
> +
> +out:
> + dash_free_scheme(&scheme);
> + return ret;
> }
>
> -static void free_representation(struct representation *pls)
> +static av_cold int dash_parse_property(AVFormatContext *s,
> + DASHAdaptationSet *as,
> + DASHRepresentationPeriod *rep,
> + xmlNodePtr node)
> {
> - free_fragment_list(pls);
> - free_timelines_list(pls);
> - free_fragment(&pls->cur_seg);
> - free_fragment(&pls->init_section);
> - av_freep(&pls->init_sec_buf);
> - av_freep(&pls->pb.buffer);
> - ff_format_io_close(pls->parent, &pls->input);
> - if (pls->ctx) {
> - pls->ctx->pb = NULL;
> - avformat_close_input(&pls->ctx);
> + int ret = 0;
> + DASHContext *c = s->priv_data;
> + DASHScheme scheme;
> +
> + if ((ret = dash_parse_scheme(s, &scheme, 1, node)) < 0)
> + goto out;
> +
> + /* https://testassets.dashif.org/#feature/details/588a48c27459f8cb201b881b */
> + if ((!as && !rep) && !strcmp(scheme.scheme_id, "urn:mpeg:dash:chaining:2016")) {
> + xml_free(c->chain_next_location);
> + c->chain_next_location = scheme.value;
> + scheme.value = NULL;
> + } else if ((!as && !rep) && !strcmp(scheme.scheme_id, "urn:mpeg:dash:fallback:2016")) {
> + xml_free(c->fallback_location);
> + c->fallback_location = scheme.value;
> + scheme.value = NULL;
> + } else {
> + dash_log_unknown_scheme(s, &scheme);
> }
>
> - av_freep(&pls->url_template);
> - av_freep(&pls->lang);
> - av_freep(&pls->id);
> - av_freep(&pls);
> +out:
> + dash_free_scheme(&scheme);
> + return ret;
> }
>
> -static void free_video_list(DASHContext *c)
> +static av_cold int dash_check_existing_timeline(AVFormatContext *s, DASHTimeline **out_timeline, xmlNodePtr timeline_node)
> {
> - int i;
> - for (i = 0; i < c->n_videos; i++) {
> - struct representation *pls = c->videos[i];
> - free_representation(pls);
> - }
> - av_freep(&c->videos);
> - c->n_videos = 0;
> + if (!*out_timeline)
> + return 0;
> +
> + av_log(s, AV_LOG_ERROR, "Multiple timelines specified\n");
> + dash_log_unknown_child(s, timeline_node);
> +
> + return AVERROR_INVALIDDATA;
> }
>
> -static void free_audio_list(DASHContext *c)
> +static int dash_parse_segmentlist(AVFormatContext *s,
> + DASHPeriod *period,
> + DASHTimeline **out_timeline,
> + xmlNodePtr node)
> {
> - int i;
> - for (i = 0; i < c->n_audios; i++) {
> - struct representation *pls = c->audios[i];
> - free_representation(pls);
> + int nb_segments = 0;
> + int64_t duration = 0;
> + int64_t timescale = 1;
> + DASHTimeline *timeline = NULL;
> + DASHSegment *g;
> + int ret = 0;
> +
> + if ((ret = dash_check_existing_timeline(s, out_timeline, node) < 0))
> + return ret;
> +
> + xml_for_each_attr {
> + if (!strcmp(attr->name, "duration"))
> + (void)av_sscanf(value, "%"SCNd64, &duration);
> + else if (!strcmp(attr->name, "timescale"))
> + (void)av_sscanf(value, "%"SCNd64, ×cale);
> + else
> + dash_log_unknown_attr(s, attr, value);
> + }
> +
> + xml_for_each_child {
> + if (!strcmp(child->name, "SegmentURL"))
> + ++nb_segments;
> + else
> + dash_log_unknown_child(s, child);
> + }
> +
> + if (!(timeline = av_mallocz(offsetof(DASHTimeline, u.segments.elems[nb_segments])))) {
> + ret = AVERROR(ENOMEM);
> + goto out;
> + }
> + dash_ref_timeline(timeline);
> + timeline->type = TIMELINE_SEGMENTS;
> + timeline->u.segments.nb = nb_segments;
> + g = &timeline->u.segments.elems[0];
> +
> + xml_for_each_child {
> + xmlNodePtr node = child;
> +
> + if (strcmp(node->name, "SegmentURL"))
> + continue;
> +
> + xml_for_each_attr {
> + if (!strcmp(attr->name, "media")) {
> + g->url = value;
> + value = NULL;
> + } else
> + dash_log_unknown_attr(s, attr, value);
> + }
> + if (!g->url) {
> + dash_log_missing_attr(s, node, "media");
> + ret = AVERROR_INVALIDDATA;
> + goto out;
> + }
> +
> + ++g;
> }
> - av_freep(&c->audios);
> - c->n_audios = 0;
> +
> + *out_timeline = dash_ref_timeline(timeline);
> +
> +out:
> + dash_unref_timeline(timeline);
> + return ret;
> }
>
> -static void free_subtitle_list(DASHContext *c)
> +static av_cold int dash_parse_segmenttimeline(AVFormatContext *s,
> + DASHPeriod *period,
> + uint64_t start_number,
> + DASHTimeline **out_timeline,
> + xmlNodePtr node)
> {
> - int i;
> - for (i = 0; i < c->n_subtitles; i++) {
> - struct representation *pls = c->subtitles[i];
> - free_representation(pls);
> + int ret = 0;
> + unsigned nb_selems = 0;
> + DASHSegmentTemplate *g;
> + DASHTimeline *timeline;
> + int64_t start_ts = 0;
> +
> + if ((ret = dash_check_existing_timeline(s, out_timeline, node) < 0))
> + return ret;
> +
> + xml_for_each_attr {
> + dash_log_unknown_attr(s, attr, value);
> + }
> +
> + xml_for_each_child {
> + if (!strcmp(child->name, "S"))
> + ++nb_selems;
> + else
> + dash_log_unknown_child(s, child);
> + }
> +
> + if (!(timeline = av_malloc(offsetof(DASHTimeline, u.templates.elems[nb_selems])))) {
> + ret = AVERROR(ENOMEM);
> + goto out;
> + }
> + memset(timeline, 0, offsetof(DASHTimeline, u.templates.elems[0]));
> + dash_ref_timeline(timeline);
> + timeline->type = TIMELINE_TEMPLATES;
> + timeline->u.templates.nb = nb_selems;
> + g = &timeline->u.templates.elems[0];
> +
> + xml_for_each_child {
> + xmlNodePtr node = child;
> +
> + if (strcmp(node->name, "S"))
> + continue;
> +
> + *g = (DASHSegmentTemplate){
> + .start_ts = start_ts,
> + .number = start_number
> + };
> +
> + xml_for_each_attr {
> + int64_t num = 0;
> +
> + (void)av_sscanf(value, "%"SCNd64, &num);
> +
> + if (!strcmp(attr->name, "t")) {
> + /* Muste be increasing. */
> + if (num < start_ts) {
> + dash_log_invalid_attr_value(s, attr, value);
> + ret = AVERROR_INVALIDDATA;
> + goto out;
> + }
> +
> + g->start_ts = num;
> + } else if (!strcmp(attr->name, "n"))
> + g->number = num;
> + else if (!strcmp(attr->name, "r"))
> + g->repeat = num;
> + else if (!strcmp(attr->name, "d")) {
> + g->duration = num;
> +
> + if (g->duration <= 0) {
> + dash_log_invalid_attr_value(s, attr, value);
> + ret = AVERROR_INVALIDDATA;
> + goto out;
> + }
> + } else {
> + dash_log_unknown_attr(s, attr, value);
> + }
> + }
> +
> + start_number = g->number + (g->repeat + 1);
> + start_ts = g->start_ts + g->duration * (g->repeat + 1);
> + ++g;
> }
> - av_freep(&c->subtitles);
> - c->n_subtitles = 0;
> +
> + *out_timeline = dash_ref_timeline(timeline);
> +
> +out:
> + dash_unref_timeline(timeline);
> + return ret;
> }
>
> -static int open_url(AVFormatContext *s, AVIOContext **pb, const char *url,
> - AVDictionary **opts, AVDictionary *opts2, int *is_http)
> +/* One init URL and list of template arguments. */
> +static av_cold int dash_parse_segmenttemplate(AVFormatContext *s,
> + DASHPeriod *period,
> + DASHTimeline **out_timeline,
> + xmlNodePtr node)
> {
> - DASHContext *c = s->priv_data;
> - AVDictionary *tmp = NULL;
> - const char *proto_name = NULL;
> - int ret;
> -
> - if (av_strstart(url, "crypto", NULL)) {
> - if (url[6] == '+' || url[6] == ':')
> - proto_name = avio_find_protocol_name(url + 7);
> + int ret = 0;
> + DASHTimeline *timeline = NULL;
> + int64_t duration = -1;
> + uint64_t start_number = 1;
> + int64_t presentation_time_offset = 0;
> + int64_t timescale = 1;
> + xmlChar *init_url = NULL;
> + xmlChar *media_url = NULL;
> +
> + if ((ret = dash_check_existing_timeline(s, out_timeline, node) < 0))
> + return ret;
> +
> + xml_for_each_attr {
> + if (!strcmp(attr->name, "startNumber"))
> + (void)av_sscanf(value, "%"SCNu64, &start_number);
> + else if (!strcmp(attr->name, "duration")) {
> + (void)av_sscanf(value, "%"SCNd64, &duration);
> + if (duration < 0) {
> + dash_log_invalid_attr_value(s, attr, value);
> + ret = AVERROR_INVALIDDATA;
> + }
> + } else if (!strcmp(attr->name, "presentationTimeOffset"))
> + (void)av_sscanf(value, "%"SCNu64, &presentation_time_offset);
> + else if (!strcmp(attr->name, "timescale")) {
> + (void)av_sscanf(value, "%"SCNd64, ×cale);
> + if (timescale <= 0) {
> + dash_log_invalid_attr_value(s, attr, value);
> + ret = AVERROR_INVALIDDATA;
> + }
> + } else if (!strcmp(attr->name, "initialization")) {
> + init_url = value;
> + value = NULL;
> + } else if (!strcmp(attr->name, "media")) {
> + media_url = value;
> + value = NULL;
> + } else
> + dash_log_unknown_attr(s, attr, value);
> }
>
> - if (!proto_name)
> - proto_name = avio_find_protocol_name(url);
> + /* value has to be free()'d */
> + if (ret < 0)
> + goto out;
>
> - if (!proto_name)
> - return AVERROR_INVALIDDATA;
> + if (!init_url || !media_url) {
> + ret = AVERROR_INVALIDDATA;
> + dash_log_missing_attr(s, node, !init_url ? "initialization" : "media");
> + goto out;
> + }
>
> - // only http(s) & file are allowed
> - if (av_strstart(proto_name, "file", NULL)) {
> - if (strcmp(c->allowed_extensions, "ALL") && !av_match_ext(url, c->allowed_extensions)) {
> - av_log(s, AV_LOG_ERROR,
> - "Filename extension of \'%s\' is not a common multimedia extension, blocked for security reasons.\n"
> - "If you wish to override this adjust allowed_extensions, you can set it to \'ALL\' to allow all\n",
> - url);
> - return AVERROR_INVALIDDATA;
> + if (0 <= duration) {
> + DASHSegmentTemplate *g;
> +
> + if (!(timeline = av_mallocz(offsetof(DASHTimeline, u.templates.elems[1])))) {
> + ret = AVERROR(ENOMEM);
> + goto out;
> + }
> + dash_ref_timeline(timeline);
> + timeline->type = TIMELINE_TEMPLATES;
> + timeline->u.templates.nb = 1;
> + g = &timeline->u.templates.elems[0];
> + g->start_ts = 0;
> + g->number = start_number;
> + g->repeat = INT64_MAX;
> + /* We round down in order to fetch segments more instead before than after. */
> + g->duration = duration; /* av_rescale_rnd(duration, AV_TIME_BASE, timescale, AV_ROUND_DOWN); */
> + } else {
> + xml_for_each_child {
> + if (!strcmp(child->name, "SegmentTimeline"))
> + ret = dash_parse_segmenttimeline(s, period, start_number, &timeline, child);
> + else
> + dash_log_unknown_child(s, child);
> +
> + if (ret < 0)
> + goto out;
> }
> - } else if (av_strstart(proto_name, "http", NULL)) {
> - ;
> - } else
> - return AVERROR_INVALIDDATA;
> -
> - if (!strncmp(proto_name, url, strlen(proto_name)) && url[strlen(proto_name)] == ':')
> - ;
> - else if (av_strstart(url, "crypto", NULL) && !strncmp(proto_name, url + 7, strlen(proto_name)) && url[7 + strlen(proto_name)] == ':')
> - ;
> - else if (strcmp(proto_name, "file") || !strncmp(url, "file,", 5))
> - return AVERROR_INVALIDDATA;
> -
> - av_freep(pb);
> - av_dict_copy(&tmp, *opts, 0);
> - av_dict_copy(&tmp, opts2, 0);
> - ret = avio_open2(pb, url, AVIO_FLAG_READ, c->interrupt_callback, &tmp);
> - if (ret >= 0) {
> - // update cookies on http response with setcookies.
> - char *new_cookies = NULL;
> -
> - if (!(s->flags & AVFMT_FLAG_CUSTOM_IO))
> - av_opt_get(*pb, "cookies", AV_OPT_SEARCH_CHILDREN, (uint8_t**)&new_cookies);
> -
> - if (new_cookies) {
> - av_dict_set(opts, "cookies", new_cookies, AV_DICT_DONT_STRDUP_VAL);
> +
> + /* Either a SegmentTemplate at duration or a SegmentTimeline must be
> + * given. */
> + if (!timeline) {
> + av_log(s, AV_LOG_ERROR, "Missing %s/%s",
> + node->name, "SegmentTimeline");
> + ret = AVERROR_INVALIDDATA;
> + goto out;
> }
>
> }
>
> - av_dict_free(&tmp);
> + timeline->duration = INT64_MAX;
> + timeline->timescale = timescale;
> + timeline->init.range = DASH_RANGE_INITALIZER;
> + timeline->init.url = init_url;
> + init_url = NULL;
> + timeline->u.templates.master.url = media_url;
> + media_url = NULL;
>
> - if (is_http)
> - *is_http = av_strstart(proto_name, "http", NULL);
> + timeline->u.templates.master.range = DASH_RANGE_INITALIZER;
>
> + *out_timeline = dash_ref_timeline(timeline);
> +
> +out:
> + xml_free(init_url);
> + xml_free(media_url);
> + dash_unref_timeline(timeline);
> return ret;
> }
>
> -static char *get_content_url(xmlNodePtr *baseurl_nodes,
> - int n_baseurl_nodes,
> - int max_url_size,
> - char *rep_id_val,
> - char *rep_bandwidth_val,
> - char *val)
> +static DASHRange dash_parse_range(const char *s, AVFormatContext *log_ctx)
> {
> - int i;
> - char *text;
> - char *url = NULL;
> - char *tmp_str = av_mallocz(max_url_size);
> + DASHRange range = {
> + .start = 0,
> + .end = INT64_MAX
> + };
> + (void)av_sscanf(s, "%"SCNd64"-%"SCNd64, &range.start, &range.end);
> + return range;
> +}
> +
> +static int dash_parse_segmentbase(AVFormatContext *s,
> + DASHPeriod *period,
> + DASHTimeline **out_timeline,
> + xmlNodePtr node)
> +{
> + int ret = 0;
> + DASHSegment *g;
> + DASHTimeline *timeline;
> +
> + if ((ret = dash_check_existing_timeline(s, out_timeline, node) < 0))
> + return ret;
> +
> + if (!(timeline = av_mallocz(offsetof(DASHTimeline, u.segments.elems[1])))) {
> + ret = AVERROR(ENOMEM);
> + goto out;
> + }
> + dash_ref_timeline(timeline);
> + timeline->type = TIMELINE_SEGMENTS;
> + timeline->duration = INT64_MAX;
> + timeline->u.segments.nb = 1;
> + g = &timeline->u.segments.elems[0];
> +
> + abort();
> + xml_for_each_child {
> + if (!strcmp(child->name, "Initalization")) {
> + xmlNodePtr node = child;
> + xml_for_each_attr {
> + if (!strcmp(attr->name, "range"))
> + timeline->init.range = dash_parse_range(value, s);
> + else
> + dash_log_unknown_attr(s, attr, value);
> + }
> + } else
> + dash_log_unknown_child(s, child);
> + }
> +
> + (void)g;
> +
> + *out_timeline = dash_ref_timeline(timeline);
> +
> +out:
> + dash_unref_timeline(timeline);
> + return 0;
> +}
> +
> +/**
> + * Substitute template arguments in |template| if not NULL and make an URL by
> + * joining it to the absolute |base| part.
> + * @param base Absolute base path.
> + * @param template Absolute or relative path, potentially containing $ template
> + * arguments. May be NULL.
> + * @return The allocated URL that shall be av_free()d by the caller.
> + */
> +static char *dash_make_url(const char *base, const char *template,
> + /* Template arguments. */
> + const char *$RepresentationID$,
> + uint64_t $Number$,
> + int64_t $Time$,
> + uint32_t $Bandwidth$)
> +{
> + enum { MAX_DIGITS = 20 };
> +
> + size_t base_size;
> + size_t $RepresentationID$_size = strlen($RepresentationID$);
> + size_t max_url_size;
> + char *url, *u;
> + URLComponents uc;
> +
> + ff_url_decompose(&uc, template, NULL);
> + base_size = URL_COMPONENT_HAVE(uc, scheme) ? 0 : strlen(base);
> +
> + max_url_size = base_size +
> + (template ? strlen(template) : 0) +
> + $RepresentationID$_size +
> + (MAX_DIGITS * 3) + 1 /* NUL */;
>
> - if (!tmp_str)
> + if (!(url = av_malloc(max_url_size)))
> return NULL;
>
> - for (i = 0; i < n_baseurl_nodes; ++i) {
> - if (baseurl_nodes[i] &&
> - baseurl_nodes[i]->children &&
> - baseurl_nodes[i]->children->type == XML_TEXT_NODE) {
> - text = xmlNodeGetContent(baseurl_nodes[i]->children);
> - if (text) {
> - memset(tmp_str, 0, max_url_size);
> - ff_make_absolute_url(tmp_str, max_url_size, "", text);
> - xmlFree(text);
> + memcpy(url, base, base_size);
> + u = url + base_size;
> +
> + while (template && *template) {
> + char *t;
> +
> + if ((t = strchr(template, '$'))) {
> + size_t len;
> + unsigned digits;
> +
> + /* Append everything before $. */
> + len = t - template;
> + memcpy(u, template, len);
> + u += len;
> + template = t + 1;
> +
> + /* Get length of template name. */
> + len = strcspn(template, "$%");
> +
> + /* Parse formatting. It's easy because we have only one
> + * possibility. */
> + digits = 0;
> + (void)av_sscanf(template + len, "%%0%ud$", &digits);
> + if (MAX_DIGITS < digits)
> + digits = MAX_DIGITS;
> +
> +#define IS(name) (len == sizeof(name) - 1 && !memcmp(template, name, sizeof(name) - 1))
> +
> + /* Substitute template argument. */
> + if (IS("RepresentationID")) {
> + memcpy(u, $RepresentationID$, $RepresentationID$_size);
> + u += $RepresentationID$_size;
> + } else if (IS("Time")) {
> + u += sprintf(u, "%0*"PRId64, digits, $Time$);
> + } else if (IS("Bandwidth")) {
> + u += sprintf(u, "%0*"PRIu32, digits, $Bandwidth$);
> + } else if (IS("Number")) {
> + u += sprintf(u, "%0*"PRIu64, digits, $Number$);
> + } else if (IS("")) {
> + *u++ = '$';
> }
> - }
> - }
>
> - if (val)
> - ff_make_absolute_url(tmp_str, max_url_size, tmp_str, val);
> +#undef IS
>
> - if (rep_id_val) {
> - url = av_strireplace(tmp_str, "$RepresentationID$", rep_id_val);
> - if (!url) {
> - goto end;
> + /* Go over $. */
> + if ((template = strchr(template + len, '$')))
> + ++template;
> + } else {
> + /* Copy remaining. */
> + strcpy(u, template);
> + break;
> }
> - av_strlcpy(tmp_str, url, max_url_size);
> }
> - if (rep_bandwidth_val && tmp_str[0] != '\0') {
> - // free any previously assigned url before reassigning
> - av_free(url);
> - url = av_strireplace(tmp_str, "$Bandwidth$", rep_bandwidth_val);
> - if (!url) {
> - goto end;
> - }
> - }
> -end:
> - av_free(tmp_str);
> +
> return url;
> }
>
> -static char *get_val_from_nodes_tab(xmlNodePtr *nodes, const int n_nodes, const char *attrname)
> +static int64_t dash_subdemuxer_seek(void *opaque, int64_t offset, int whence)
> {
> - int i;
> - char *val;
> -
> - for (i = 0; i < n_nodes; ++i) {
> - if (nodes[i]) {
> - val = xmlGetProp(nodes[i], attrname);
> - if (val)
> - return val;
> - }
> - }
> -
> - return NULL;
> + DASHRepresentation *rep = opaque;
> + return avio_seek(rep->segments[0].pb, offset, whence);
> }
>
> -static xmlNodePtr find_child_node_by_name(xmlNodePtr rootnode, const char *nodename)
> +static av_cold int dash_copy_stream_props(AVStream *slave, AVStream *master)
> {
> - xmlNodePtr node = rootnode;
> - if (!node) {
> - return NULL;
> - }
> + int ret;
>
> - node = xmlFirstElementChild(node);
> - while (node) {
> - if (!av_strcasecmp(node->name, nodename)) {
> - return node;
> - }
> - node = xmlNextElementSibling(node);
> - }
> - return NULL;
> + master->event_flags &= ~AVSTREAM_EVENT_FLAG_METADATA_UPDATED;
> + slave->event_flags |= AVSTREAM_EVENT_FLAG_METADATA_UPDATED;
> +
> + slave->internal->need_context_update = 1;
> +
> + if ((ret = ff_stream_encode_params_copy(slave, master)) < 0)
> + return ret;
> +
> + /* Only for probed context (oc->iformat != NULL). */
> + if (master->time_base.den)
> + avpriv_set_pts_info(slave, master->pts_wrap_bits,
> + master->time_base.num, master->time_base.den);
> +
> + return 0;
> }
>
> -static enum AVMediaType get_content_type(xmlNodePtr node)
> +static av_cold DASHRepresentationPeriod *dash_find_representation_period_at(
> + DASHRepresentation *rep, int64_t timestamp)
> {
> - enum AVMediaType type = AVMEDIA_TYPE_UNKNOWN;
> - int i = 0;
> - const char *attr;
> - char *val = NULL;
> -
> - if (node) {
> - for (i = 0; i < 2; i++) {
> - attr = i ? "mimeType" : "contentType";
> - val = xmlGetProp(node, attr);
> - if (val) {
> - if (av_stristr(val, "video")) {
> - type = AVMEDIA_TYPE_VIDEO;
> - } else if (av_stristr(val, "audio")) {
> - type = AVMEDIA_TYPE_AUDIO;
> - } else if (av_stristr(val, "text")) {
> - type = AVMEDIA_TYPE_SUBTITLE;
> - }
> - xmlFree(val);
> - }
> - }
> + for (unsigned i = 0; i < rep->nb_periods; ++i) {
> + DASHRepresentationPeriod *period = rep->periods[i];
> + if (period->period->start_ts <= timestamp &&
> + timestamp < period->period->end_ts)
> + return period;
> }
> - return type;
> +
> + return NULL;
> }
>
> -static struct fragment * get_Fragment(char *range)
> +/**
> + * Update everything to make outside context of representation.
> + *
> + * Must be called after:
> + * - rep->ic->streams changed,
> + * - rep->cur_period changed.
> + */
> +static av_cold int dash_subdemuxer_update(DASHRepresentation *rep)
> {
> - struct fragment * seg = av_mallocz(sizeof(struct fragment));
> -
> - if (!seg)
> - return NULL;
> + int ret;
> + void *p;
> + AVFormatContext *oc = rep->oc;
> + AVFormatContext *ic = rep->ic;
>
> - seg->size = -1;
> - if (range) {
> - char *str_end_offset;
> - char *str_offset = av_strtok(range, "-", &str_end_offset);
> - seg->url_offset = strtoll(str_offset, NULL, 10);
> - seg->size = strtoll(str_end_offset, NULL, 10) - seg->url_offset + 1;
> + if (rep->nb_streams < ic->nb_streams) {
> + if (!(p = av_realloc(rep->ostreams, ic->nb_streams * sizeof(*rep->ostreams))))
> + return AVERROR(ENOMEM);
> + rep->ostreams = p;
> }
>
> - return seg;
> -}
> + rep->ic->event_flags &= ~AVFMT_EVENT_FLAG_METADATA_UPDATED;
> + rep->oc->event_flags |= AVFMT_EVENT_FLAG_METADATA_UPDATED;
>
> -static int parse_manifest_segmenturlnode(AVFormatContext *s, struct representation *rep,
> - xmlNodePtr fragmenturl_node,
> - xmlNodePtr *baseurl_nodes,
> - char *rep_id_val,
> - char *rep_bandwidth_val)
> -{
> - DASHContext *c = s->priv_data;
> - char *initialization_val = NULL;
> - char *media_val = NULL;
> - char *range_val = NULL;
> - int max_url_size = c ? c->max_url_size: MAX_URL_SIZE;
> - int err;
> -
> - if (!av_strcasecmp(fragmenturl_node->name, "Initialization")) {
> - initialization_val = xmlGetProp(fragmenturl_node, "sourceURL");
> - range_val = xmlGetProp(fragmenturl_node, "range");
> - if (initialization_val || range_val) {
> - free_fragment(&rep->init_section);
> - rep->init_section = get_Fragment(range_val);
> - xmlFree(range_val);
> - if (!rep->init_section) {
> - xmlFree(initialization_val);
> - return AVERROR(ENOMEM);
> - }
> - rep->init_section->url = get_content_url(baseurl_nodes, 4,
> - max_url_size,
> - rep_id_val,
> - rep_bandwidth_val,
> - initialization_val);
> - xmlFree(initialization_val);
> - if (!rep->init_section->url) {
> - av_freep(&rep->init_section);
> + for (unsigned stream_index = 0;
> + stream_index < ic->nb_streams;
> + stream_index++)
> + {
> + AVStream *ist, *ost;
> + DASHRepresentationPeriod *period;
> + AVDictionary *metadata = NULL;
> +
> + ist = ic->streams[stream_index];
> +
> + if (stream_index < rep->nb_streams) {
> + ost = rep->ostreams[stream_index];
> + } else {
> + ost = avformat_new_stream(oc, NULL);
> + if (!ost)
> return AVERROR(ENOMEM);
> - }
> +
> + rep->ostreams[stream_index] = ost;
> }
> - } else if (!av_strcasecmp(fragmenturl_node->name, "SegmentURL")) {
> - media_val = xmlGetProp(fragmenturl_node, "media");
> - range_val = xmlGetProp(fragmenturl_node, "mediaRange");
> - if (media_val || range_val) {
> - struct fragment *seg = get_Fragment(range_val);
> - xmlFree(range_val);
> - if (!seg) {
> - xmlFree(media_val);
> - return AVERROR(ENOMEM);
> - }
> - seg->url = get_content_url(baseurl_nodes, 4,
> - max_url_size,
> - rep_id_val,
> - rep_bandwidth_val,
> - media_val);
> - xmlFree(media_val);
> - if (!seg->url) {
> - av_free(seg);
> +
> + av_log(oc, AV_LOG_VERBOSE,
> + "Match '%s', stream #%u -> DASH stream #%u\n",
> + rep->id, stream_index, ost->index);
> +
> + /* Period specific metadata. */
> + period = rep->cur_period;
> + /* For inactive representations compute where we would be. */
> + if (!period)
> + period = dash_find_representation_period_at(rep, rep->read_ts);
> + if (period) {
> + const DASHParameters *par = &period->par;
> + uint8_t *side_data;
> + int side_data_size;
> +
> + side_data = av_encryption_init_info_add_side_data(period->as->init_info, &side_data_size);
> + if (!side_data)
> return AVERROR(ENOMEM);
> +
> + ret = av_stream_add_side_data(ist, AV_PKT_DATA_ENCRYPTION_INIT_INFO,
> + side_data, side_data_size);
> + if (ret < 0) {
> + av_free(side_data);
> + return ret;
> }
> - err = av_dynarray_add_nofree(&rep->fragments, &rep->n_fragments, seg);
> - if (err < 0) {
> - free_fragment(&seg);
> - return err;
> +
> + av_dict_set_int(&metadata, "variant_bitrate", period->bandwidth, AV_DICT_MULTIKEY);
> +
> + /* AdaptationSet specific metadata. */
> + av_dict_copy(&metadata, period->as->metadata, AV_DICT_MULTIKEY);
> +
> + /* Most parameters are only relevant for elementary streams. */
> + if (rep->ic->nb_streams == 1) {
> + AVCodecParameters *codecpar = ist->codecpar;
> +
> + /* Set unknown parameters for manifest. */
> +
> + if (codecpar->codec_type == AVMEDIA_TYPE_UNKNOWN)
> + codecpar->codec_type = par->codec_type;
> +
> + if (codecpar->codec_id == AV_CODEC_ID_NONE)
> + codecpar->codec_id = par->codec_id;
> +
> + if (!codecpar->sample_rate)
> + codecpar->sample_rate = par->sample_rate;
> +
> + if (!codecpar->channels && !codecpar->channel_layout)
> + codecpar->channels = par->channels,
> + codecpar->channel_layout = par->channel_layout;
> +
> + if (!codecpar->width && !codecpar->height)
> + codecpar->width = par->width,
> + codecpar->height = par->height;
> +
> + if (!ist->avg_frame_rate.num)
> + ist->avg_frame_rate = par->frame_rate;
> + if (!ist->r_frame_rate.num)
> + ist->r_frame_rate = par->frame_rate;
> +
> + if (!codecpar->sample_aspect_ratio.num)
> + codecpar->sample_aspect_ratio = par->sample_aspect_ratio;
> +
> + if (codecpar->field_order == AV_FIELD_UNKNOWN)
> + codecpar->field_order = par->field_order;
> }
> +
> + ist->disposition = par->disposition;
> + }
> +
> + /* Representation specific metadata. */
> + av_dict_set(&metadata, "id", rep->id, AV_DICT_MULTIKEY);
> +
> + /* RepresentationPeriod (stream) specific metadata. */
> + if ((ret = dash_copy_stream_props(ost, ist)) < 0) {
> + av_dict_free(&metadata);
> + return ret;
> }
> +
> + av_dict_copy(&ost->metadata, metadata, AV_DICT_MULTIKEY);
> + av_dict_free(&metadata);
> + }
> +
> + for (unsigned stream_index = ic->nb_streams;
> + stream_index < rep->nb_streams;
> + stream_index++)
> + {
> + AVStream *ist, *ost;
> +
> + /* Dummy format with no streams. */
> + if (!ic->streams)
> + break;
> +
> + ist = ic->streams[stream_index];
> + ost = rep->ostreams[stream_index];
> +
> + /* Reset codec parameters. */
> + avcodec_parameters_free(&ist->codecpar);
> + ist->codecpar = avcodec_parameters_alloc();
> + if (!ist->codecpar)
> + return AVERROR(ENOMEM);
> +
> + if ((ret = dash_copy_stream_props(ost, ist)) < 0)
> + return ret;
> }
>
> + rep->nb_streams = FFMAX(rep->nb_streams, ic->nb_streams);
> +
> return 0;
> }
>
> -static int parse_manifest_segmenttimeline(AVFormatContext *s, struct representation *rep,
> - xmlNodePtr fragment_timeline_node)
> +static void dash_rotate_urllist(AVFormatContext *s, DASHURLList *urls)
> {
> - xmlAttrPtr attr = NULL;
> - char *val = NULL;
> - int err;
> + DASHContext *c = s->priv_data;
>
> - if (!av_strcasecmp(fragment_timeline_node->name, "S")) {
> - struct timeline *tml = av_mallocz(sizeof(struct timeline));
> - if (!tml) {
> - return AVERROR(ENOMEM);
> - }
> - attr = fragment_timeline_node->properties;
> - while (attr) {
> - val = xmlGetProp(fragment_timeline_node, attr->name);
> + if (1 < urls->nb) {
> + unsigned const i = 1 + av_lfg_get(&c->rnd) % (urls->nb - 1);
> + char *tmp = urls->elems[i];
> + urls->elems[i] = urls->elems[0];
> + urls->elems[0] = tmp;
> + }
> +}
>
> - if (!val) {
> - av_log(s, AV_LOG_WARNING, "parse_manifest_segmenttimeline attr->name = %s val is NULL\n", attr->name);
> +static int dash_subdemuxer_read(void *opaque, uint8_t *buf, int buf_size)
> +{
> + int ret;
> + DASHRepresentation *rep = opaque;
> + AVFormatContext *s = rep->oc;
> + DASHContext *c = s->priv_data;
> + int size;
> + unsigned seg = rep->cur_segment;
> +
> +open_segment:;
> + do {
> + DASHRepresentationPeriod *period;
> + char *url;
> + const char *base;
> + const char *template;
> + uint64_t $Number$;
> + int64_t $Time$;
> + DASHRange range;
> + AVDictionary *opts;
> +#if CONFIG_HTTP_PROTOCOL
> + URLContext *uc;
> +#endif
> +
> + if (rep->segments[seg].pb &&
> + !rep->segments[seg].pb->eof_reached)
> + continue;
> +
> + for (unsigned i = 0; i < rep->nb_periods; ++i) {
> + int64_t period_ts;
> + DASHTimeline *timeline;
> +
> + period = rep->periods[i];
> + timeline = period->timeline;
> +
> + /* Check if we are inside Period boundaries. */
> + if (!(period->period->start_ts <= rep->read_ts &&
> + rep->read_ts < period->period->end_ts))
> continue;
> - }
>
> - if (!av_strcasecmp(attr->name, "t")) {
> - tml->starttime = (int64_t)strtoll(val, NULL, 10);
> - } else if (!av_strcasecmp(attr->name, "r")) {
> - tml->repeat =(int64_t) strtoll(val, NULL, 10);
> - } else if (!av_strcasecmp(attr->name, "d")) {
> - tml->duration = (int64_t)strtoll(val, NULL, 10);
> - }
> - attr = attr->next;
> - xmlFree(val);
> - }
> - err = av_dynarray_add_nofree(&rep->timelines, &rep->n_timelines, tml);
> - if (err < 0) {
> - av_free(tml);
> - return err;
> - }
> - }
> + /* period_ts := read_ts relative to Period start. */
> + period_ts = av_rescale_q_rnd(rep->read_ts - period->period->start_ts,
> + AV_TIME_BASE_Q,
> + (AVRational){ 1, timeline->timescale },
> + AV_ROUND_UP);
>
> - return 0;
> -}
> + if (timeline->type == TIMELINE_TEMPLATES) {
> + for (unsigned j = 0; j < timeline->u.templates.nb; ++j) {
> + DASHSegmentTemplate *g = &timeline->u.templates.elems[j];
> + int64_t end_time = g->repeat < 0 || INT64_MAX / (g->repeat + 1) < g->duration
> + ? INT64_MAX
> + : g->start_ts + g->duration * (g->repeat + 1);
> + end_time = FFMIN(end_time, timeline->duration);
>
> -static int resolve_content_path(AVFormatContext *s, const char *url, int *max_url_size, xmlNodePtr *baseurl_nodes, int n_baseurl_nodes)
> -{
> - char *tmp_str = NULL;
> - char *path = NULL;
> - char *mpdName = NULL;
> - xmlNodePtr node = NULL;
> - char *baseurl = NULL;
> - char *root_url = NULL;
> - char *text = NULL;
> - char *tmp = NULL;
> - int isRootHttp = 0;
> - char token ='/';
> - int start = 0;
> - int rootId = 0;
> - int updated = 0;
> - int size = 0;
> - int i;
> - int tmp_max_url_size = strlen(url);
> + if (period_ts < g->start_ts) {
> + /* Gap detected: Missing segments in timeline. */
> + }
>
> - for (i = n_baseurl_nodes-1; i >= 0 ; i--) {
> - text = xmlNodeGetContent(baseurl_nodes[i]);
> - if (!text)
> - continue;
> - tmp_max_url_size += strlen(text);
> - if (ishttp(text)) {
> - xmlFree(text);
> - break;
> - }
> - xmlFree(text);
> - }
> + if (period_ts < end_time) {
> +#if 0
> + av_log(rep->oc, AV_LOG_TRACE, "S@[n=%"PRId64" d*r=%"PRId64"*%"PRId64"]: %"PRId64" <= %"PRId64" <= %"PRId64"?\n",
> + g->number, g->duration, g->repeat, g->start_ts, period_ts, end_time);
> +#endif
>
> - tmp_max_url_size = aligned(tmp_max_url_size);
> - text = av_mallocz(tmp_max_url_size);
> - if (!text) {
> - updated = AVERROR(ENOMEM);
> - goto end;
> - }
> - av_strlcpy(text, url, strlen(url)+1);
> - tmp = text;
> - while (mpdName = av_strtok(tmp, "/", &tmp)) {
> - size = strlen(mpdName);
> - }
> - av_free(text);
> + /* If period changed push init section first. */
> + if (rep->cur_period != period) {
> + if (seg != rep->cur_segment)
> + goto read_segment;
>
> - path = av_mallocz(tmp_max_url_size);
> - tmp_str = av_mallocz(tmp_max_url_size);
> - if (!tmp_str || !path) {
> - updated = AVERROR(ENOMEM);
> - goto end;
> - }
> + if (!rep->save_init) {
> + rep->save_init = 1;
>
> - av_strlcpy (path, url, strlen(url) - size + 1);
> - for (rootId = n_baseurl_nodes - 1; rootId > 0; rootId --) {
> - if (!(node = baseurl_nodes[rootId])) {
> - continue;
> - }
> - text = xmlNodeGetContent(node);
> - if (ishttp(text)) {
> - xmlFree(text);
> - break;
> - }
> - xmlFree(text);
> - }
> + /* Send out stored part. */
> + if (0 < period->initbuf_size) {
> + memcpy(buf, period->initbuf, period->initbuf_size);
> + return period->initbuf_size;
> + }
> + }
>
> - node = baseurl_nodes[rootId];
> - baseurl = xmlNodeGetContent(node);
> - root_url = (av_strcasecmp(baseurl, "")) ? baseurl : path;
> - if (node) {
> - xmlNodeSetContent(node, root_url);
> - updated = 1;
> - }
> + rep->cur_period = period;
> + ret = dash_subdemuxer_update(rep);
> + if (ret < 0)
> + return ret;
>
> - size = strlen(root_url);
> - isRootHttp = ishttp(root_url);
> + range = timeline->init.range;
> + range.start += period->initbuf_size;
>
> - if (size > 0 && root_url[size - 1] != token) {
> - av_strlcat(root_url, "/", size + 2);
> - size += 2;
> - }
> + /* Test whether full segment is stored and thus
> + * previously has been sent out. */
> + if (0 < period->initbuf_size &&
> + (period->initbuf_size < INITBUF_MAX ||
> + range.end <= range.start))
> + {
> + rep->save_init = 0;
> + continue;
> + }
>
> - for (i = 0; i < n_baseurl_nodes; ++i) {
> - if (i == rootId) {
> - continue;
> - }
> - text = xmlNodeGetContent(baseurl_nodes[i]);
> - if (text && !av_strstart(text, "/", NULL)) {
> - memset(tmp_str, 0, strlen(tmp_str));
> - if (!ishttp(text) && isRootHttp) {
> - av_strlcpy(tmp_str, root_url, size + 1);
> - }
> - start = (text[0] == token);
> - if (start && av_stristr(tmp_str, text)) {
> - char *p = tmp_str;
> - if (!av_strncasecmp(tmp_str, "http://", 7)) {
> - p += 7;
> - } else if (!av_strncasecmp(tmp_str, "https://", 8)) {
> - p += 8;
> + $Time$ = 0, $Number$ = 0; /* They should not be used in template URL. */
> + template = timeline->init.url;
> + } else {
> + uint64_t r = (period_ts - g->start_ts) / g->duration;
> + $Time$ = g->start_ts + g->duration * r;
> + $Number$ = g->number + r;
> + template = timeline->u.templates.master.url;
> + range = timeline->u.templates.master.range;
> +
> + rep->read_ts = av_rescale_q_rnd($Time$ + g->duration,
> + (AVRational){ 1, timeline->timescale },
> + AV_TIME_BASE_Q,
> + AV_ROUND_UP);
> +
> + /* Wait until segment becomes available. */
> + if (c->is_live) {
> + int64_t time_shift = rep->read_ts - c->availability_start_time;
> + if (time_shift < 0) {
> + /* Allow suspending thread only when it is really neccessary, i.e. we have no other segments to serve. */
> + if (seg == rep->cur_segment)
> + av_usleep(-time_shift);
> + else
> + goto read_segment;
> + }
> + }
> + }
> + base = period->base->elems[0];
> +
> + goto found;
> + }
> }
> - p = strchr(p, '/');
> - memset(p + 1, 0, strlen(p));
> - }
> - av_strlcat(tmp_str, text + start, tmp_max_url_size);
> - xmlNodeSetContent(baseurl_nodes[i], tmp_str);
> - updated = 1;
> - xmlFree(text);
> - }
> - }
> -
> -end:
> - if (tmp_max_url_size > *max_url_size) {
> - *max_url_size = tmp_max_url_size;
> - }
> - av_free(path);
> - av_free(tmp_str);
> - xmlFree(baseurl);
> - return updated;
> -
> -}
> -
> -static int parse_manifest_representation(AVFormatContext *s, const char *url,
> - xmlNodePtr node,
> - xmlNodePtr adaptionset_node,
> - xmlNodePtr mpd_baseurl_node,
> - xmlNodePtr period_baseurl_node,
> - xmlNodePtr period_segmenttemplate_node,
> - xmlNodePtr period_segmentlist_node,
> - xmlNodePtr fragment_template_node,
> - xmlNodePtr content_component_node,
> - xmlNodePtr adaptionset_baseurl_node,
> - xmlNodePtr adaptionset_segmentlist_node,
> - xmlNodePtr adaptionset_supplementalproperty_node)
> -{
> - int32_t ret = 0;
> - DASHContext *c = s->priv_data;
> - struct representation *rep = NULL;
> - struct fragment *seg = NULL;
> - xmlNodePtr representation_segmenttemplate_node = NULL;
> - xmlNodePtr representation_baseurl_node = NULL;
> - xmlNodePtr representation_segmentlist_node = NULL;
> - xmlNodePtr segmentlists_tab[3];
> - xmlNodePtr fragment_timeline_node = NULL;
> - xmlNodePtr fragment_templates_tab[5];
> - char *val = NULL;
> - xmlNodePtr baseurl_nodes[4];
> - xmlNodePtr representation_node = node;
> - char *rep_bandwidth_val;
> - enum AVMediaType type = AVMEDIA_TYPE_UNKNOWN;
> -
> - // try get information from representation
> - if (type == AVMEDIA_TYPE_UNKNOWN)
> - type = get_content_type(representation_node);
> - // try get information from contentComponen
> - if (type == AVMEDIA_TYPE_UNKNOWN)
> - type = get_content_type(content_component_node);
> - // try get information from adaption set
> - if (type == AVMEDIA_TYPE_UNKNOWN)
> - type = get_content_type(adaptionset_node);
> - if (type != AVMEDIA_TYPE_VIDEO && type != AVMEDIA_TYPE_AUDIO &&
> - type != AVMEDIA_TYPE_SUBTITLE) {
> - av_log(s, AV_LOG_VERBOSE, "Parsing '%s' - skipp not supported representation type\n", url);
> - return 0;
> - }
> -
> - // convert selected representation to our internal struct
> - rep = av_mallocz(sizeof(struct representation));
> - if (!rep)
> - return AVERROR(ENOMEM);
> - if (c->adaptionset_lang) {
> - rep->lang = av_strdup(c->adaptionset_lang);
> - if (!rep->lang) {
> - av_log(s, AV_LOG_ERROR, "alloc language memory failure\n");
> - av_freep(&rep);
> - return AVERROR(ENOMEM);
> - }
> - }
> - rep->parent = s;
> - representation_segmenttemplate_node = find_child_node_by_name(representation_node, "SegmentTemplate");
> - representation_baseurl_node = find_child_node_by_name(representation_node, "BaseURL");
> - representation_segmentlist_node = find_child_node_by_name(representation_node, "SegmentList");
> - rep_bandwidth_val = xmlGetProp(representation_node, "bandwidth");
> - val = xmlGetProp(representation_node, "id");
> - if (val) {
> - rep->id = av_strdup(val);
> - xmlFree(val);
> - if (!rep->id)
> - goto enomem;
> - }
> -
> - baseurl_nodes[0] = mpd_baseurl_node;
> - baseurl_nodes[1] = period_baseurl_node;
> - baseurl_nodes[2] = adaptionset_baseurl_node;
> - baseurl_nodes[3] = representation_baseurl_node;
> -
> - ret = resolve_content_path(s, url, &c->max_url_size, baseurl_nodes, 4);
> - c->max_url_size = aligned(c->max_url_size
> - + (rep->id ? strlen(rep->id) : 0)
> - + (rep_bandwidth_val ? strlen(rep_bandwidth_val) : 0));
> - if (ret == AVERROR(ENOMEM) || ret == 0)
> - goto free;
> - if (representation_segmenttemplate_node || fragment_template_node || period_segmenttemplate_node) {
> - fragment_timeline_node = NULL;
> - fragment_templates_tab[0] = representation_segmenttemplate_node;
> - fragment_templates_tab[1] = adaptionset_segmentlist_node;
> - fragment_templates_tab[2] = fragment_template_node;
> - fragment_templates_tab[3] = period_segmenttemplate_node;
> - fragment_templates_tab[4] = period_segmentlist_node;
> -
> - val = get_val_from_nodes_tab(fragment_templates_tab, 4, "initialization");
> - if (val) {
> - rep->init_section = av_mallocz(sizeof(struct fragment));
> - if (!rep->init_section) {
> - xmlFree(val);
> - goto enomem;
> - }
> - c->max_url_size = aligned(c->max_url_size + strlen(val));
> - rep->init_section->url = get_content_url(baseurl_nodes, 4,
> - c->max_url_size, rep->id,
> - rep_bandwidth_val, val);
> - xmlFree(val);
> - if (!rep->init_section->url)
> - goto enomem;
> - rep->init_section->size = -1;
> - }
> - val = get_val_from_nodes_tab(fragment_templates_tab, 4, "media");
> - if (val) {
> - c->max_url_size = aligned(c->max_url_size + strlen(val));
> - rep->url_template = get_content_url(baseurl_nodes, 4,
> - c->max_url_size, rep->id,
> - rep_bandwidth_val, val);
> - xmlFree(val);
> - }
> - val = get_val_from_nodes_tab(fragment_templates_tab, 4, "presentationTimeOffset");
> - if (val) {
> - rep->presentation_timeoffset = (int64_t) strtoll(val, NULL, 10);
> - av_log(s, AV_LOG_TRACE, "rep->presentation_timeoffset = [%"PRId64"]\n", rep->presentation_timeoffset);
> - xmlFree(val);
> - }
> - val = get_val_from_nodes_tab(fragment_templates_tab, 4, "duration");
> - if (val) {
> - rep->fragment_duration = (int64_t) strtoll(val, NULL, 10);
> - av_log(s, AV_LOG_TRACE, "rep->fragment_duration = [%"PRId64"]\n", rep->fragment_duration);
> - xmlFree(val);
> - }
> - val = get_val_from_nodes_tab(fragment_templates_tab, 4, "timescale");
> - if (val) {
> - rep->fragment_timescale = (int64_t) strtoll(val, NULL, 10);
> - av_log(s, AV_LOG_TRACE, "rep->fragment_timescale = [%"PRId64"]\n", rep->fragment_timescale);
> - xmlFree(val);
> - }
> - val = get_val_from_nodes_tab(fragment_templates_tab, 4, "startNumber");
> - if (val) {
> - rep->start_number = rep->first_seq_no = (int64_t) strtoll(val, NULL, 10);
> - av_log(s, AV_LOG_TRACE, "rep->first_seq_no = [%"PRId64"]\n", rep->first_seq_no);
> - xmlFree(val);
> - }
> - if (adaptionset_supplementalproperty_node) {
> - if (!av_strcasecmp(xmlGetProp(adaptionset_supplementalproperty_node,"schemeIdUri"), "http://dashif.org/guidelines/last-segment-number")) {
> - val = xmlGetProp(adaptionset_supplementalproperty_node,"value");
> - if (!val) {
> - av_log(s, AV_LOG_ERROR, "Missing value attribute in adaptionset_supplementalproperty_node\n");
> + } else if (timeline->type == TIMELINE_SEGMENTS) {
> + DASHSegment *g;
> +
> + if (rep->cur_period != period) {
> + if (seg != rep->cur_segment)
> + goto read_segment;
> +
> + rep->cur_period = period;
> + ret = dash_subdemuxer_update(rep);
> + if (ret < 0)
> + return ret;
> +
> + g = &timeline->init;
> + } else if (0 < timeline->u.segments.nb) {
> + int64_t segment_ts = timeline->duration / timeline->u.segments.nb;
> + g = &timeline->u.segments.elems[period_ts / segment_ts];
> } else {
> - rep->last_seq_no =(int64_t) strtoll(val, NULL, 10) - 1;
> - xmlFree(val);
> + /* One segment timeline. */
> + continue;
> }
> - }
> - }
>
> - fragment_timeline_node = find_child_node_by_name(representation_segmenttemplate_node, "SegmentTimeline");
> + base = period->base->elems[0];
> + template = g->url; /* HACK: URL is not template. */
> + range = g->range;
>
> - if (!fragment_timeline_node)
> - fragment_timeline_node = find_child_node_by_name(fragment_template_node, "SegmentTimeline");
> - if (!fragment_timeline_node)
> - fragment_timeline_node = find_child_node_by_name(adaptionset_segmentlist_node, "SegmentTimeline");
> - if (!fragment_timeline_node)
> - fragment_timeline_node = find_child_node_by_name(period_segmentlist_node, "SegmentTimeline");
> - if (fragment_timeline_node) {
> - fragment_timeline_node = xmlFirstElementChild(fragment_timeline_node);
> - while (fragment_timeline_node) {
> - ret = parse_manifest_segmenttimeline(s, rep, fragment_timeline_node);
> - if (ret < 0)
> - goto free;
> - fragment_timeline_node = xmlNextElementSibling(fragment_timeline_node);
> + goto found;
> + } else {
> + abort();
> }
> - }
> - } else if (representation_baseurl_node && !representation_segmentlist_node) {
> - seg = av_mallocz(sizeof(struct fragment));
> - if (!seg)
> - goto enomem;
> - ret = av_dynarray_add_nofree(&rep->fragments, &rep->n_fragments, seg);
> - if (ret < 0) {
> - av_free(seg);
> - goto free;
> - }
> - seg->url = get_content_url(baseurl_nodes, 4, c->max_url_size,
> - rep->id, rep_bandwidth_val, NULL);
> - if (!seg->url)
> - goto enomem;
> - seg->size = -1;
> - } else if (representation_segmentlist_node) {
> - // TODO: https://www.brendanlong.com/the-structure-of-an-mpeg-dash-mpd.html
> - // http://www-itec.uni-klu.ac.at/dash/ddash/mpdGenerator.php?fragmentlength=15&type=full
> - xmlNodePtr fragmenturl_node = NULL;
> - segmentlists_tab[0] = representation_segmentlist_node;
> - segmentlists_tab[1] = adaptionset_segmentlist_node;
> - segmentlists_tab[2] = period_segmentlist_node;
>
> - val = get_val_from_nodes_tab(segmentlists_tab, 3, "duration");
> - if (val) {
> - rep->fragment_duration = (int64_t) strtoll(val, NULL, 10);
> - av_log(s, AV_LOG_TRACE, "rep->fragment_duration = [%"PRId64"]\n", rep->fragment_duration);
> - xmlFree(val);
> - }
> - val = get_val_from_nodes_tab(segmentlists_tab, 3, "timescale");
> - if (val) {
> - rep->fragment_timescale = (int64_t) strtoll(val, NULL, 10);
> - av_log(s, AV_LOG_TRACE, "rep->fragment_timescale = [%"PRId64"]\n", rep->fragment_timescale);
> - xmlFree(val);
> - }
> - val = get_val_from_nodes_tab(segmentlists_tab, 3, "startNumber");
> - if (val) {
> - rep->start_number = rep->first_seq_no = (int64_t) strtoll(val, NULL, 10);
> - av_log(s, AV_LOG_TRACE, "rep->first_seq_no = [%"PRId64"]\n", rep->first_seq_no);
> - xmlFree(val);
> - }
> + /* Gap detected: No more segments till end of the period. Jump to
> + * the end of the period. */
> + rep->read_ts = period->period->end_ts;
>
> - fragmenturl_node = xmlFirstElementChild(representation_segmentlist_node);
> - while (fragmenturl_node) {
> - ret = parse_manifest_segmenturlnode(s, rep, fragmenturl_node,
> - baseurl_nodes, rep->id,
> - rep_bandwidth_val);
> - if (ret < 0)
> - goto free;
> - fragmenturl_node = xmlNextElementSibling(fragmenturl_node);
> + /* Periods may be out-of-order so start searching next one from the beginning. */
> + i = 0;
> }
>
> - fragment_timeline_node = find_child_node_by_name(adaptionset_segmentlist_node, "SegmentTimeline");
> - if (!fragment_timeline_node)
> - fragment_timeline_node = find_child_node_by_name(period_segmentlist_node, "SegmentTimeline");
> - if (fragment_timeline_node) {
> - fragment_timeline_node = xmlFirstElementChild(fragment_timeline_node);
> - while (fragment_timeline_node) {
> - ret = parse_manifest_segmenttimeline(s, rep, fragment_timeline_node);
> - if (ret < 0)
> - goto free;
> - fragment_timeline_node = xmlNextElementSibling(fragment_timeline_node);
> - }
> - }
> - } else {
> - av_log(s, AV_LOG_ERROR, "Unknown format of Representation node id '%s' \n",
> - rep->id ? rep->id : "");
> - goto free;
> - }
> + if (seg == rep->cur_segment)
> + return AVERROR_EOF;
> + else
> + goto read_segment;
>
> - if (rep->fragment_duration > 0 && !rep->fragment_timescale)
> - rep->fragment_timescale = 1;
> - rep->bandwidth = rep_bandwidth_val ? atoi(rep_bandwidth_val) : 0;
> - rep->framerate = av_make_q(0, 0);
> - if (type == AVMEDIA_TYPE_VIDEO) {
> - char *rep_framerate_val = xmlGetProp(representation_node, "frameRate");
> - if (rep_framerate_val) {
> - ret = av_parse_video_rate(&rep->framerate, rep_framerate_val);
> - if (ret < 0)
> - av_log(s, AV_LOG_VERBOSE, "Ignoring invalid frame rate '%s'\n", rep_framerate_val);
> - xmlFree(rep_framerate_val);
> + found:
> + if (template) {
> + url = dash_make_url(base, template, rep->id, $Number$, $Time$, period->bandwidth);
> + if (!url)
> + return AVERROR(ENOMEM);
> + } else {
> + url = (char *)base;
> }
> - }
> -
> - switch (type) {
> - case AVMEDIA_TYPE_VIDEO:
> - ret = av_dynarray_add_nofree(&c->videos, &c->n_videos, rep);
> - break;
> - case AVMEDIA_TYPE_AUDIO:
> - ret = av_dynarray_add_nofree(&c->audios, &c->n_audios, rep);
> - break;
> - case AVMEDIA_TYPE_SUBTITLE:
> - ret = av_dynarray_add_nofree(&c->subtitles, &c->n_subtitles, rep);
> - break;
> - }
> - if (ret < 0)
> - goto free;
> -
> -end:
> - if (rep_bandwidth_val)
> - xmlFree(rep_bandwidth_val);
>
> - return ret;
> -enomem:
> - ret = AVERROR(ENOMEM);
> -free:
> - free_representation(rep);
> - goto end;
> -}
> + opts = NULL;
>
> -static int parse_manifest_adaptationset_attr(AVFormatContext *s, xmlNodePtr adaptionset_node)
> -{
> - DASHContext *c = s->priv_data;
> + if (0 < range.start)
> + av_dict_set_int(&opts, "offset", range.start, 0);
>
> - if (!adaptionset_node) {
> - av_log(s, AV_LOG_WARNING, "Cannot get AdaptionSet\n");
> - return AVERROR(EINVAL);
> - }
> - c->adaptionset_lang = xmlGetProp(adaptionset_node, "lang");
> + if (range.end < INT64_MAX)
> + av_dict_set_int(&opts, "end_offset", range.end, 0);
>
> - return 0;
> -}
> + av_dict_set(&opts, "multiple_requests", "1", 0);
> + av_dict_copy(&opts, c->protocol_opts, 0);
>
> -static int parse_manifest_adaptationset(AVFormatContext *s, const char *url,
> - xmlNodePtr adaptionset_node,
> - xmlNodePtr mpd_baseurl_node,
> - xmlNodePtr period_baseurl_node,
> - xmlNodePtr period_segmenttemplate_node,
> - xmlNodePtr period_segmentlist_node)
> -{
> - int ret = 0;
> - DASHContext *c = s->priv_data;
> - xmlNodePtr fragment_template_node = NULL;
> - xmlNodePtr content_component_node = NULL;
> - xmlNodePtr adaptionset_baseurl_node = NULL;
> - xmlNodePtr adaptionset_segmentlist_node = NULL;
> - xmlNodePtr adaptionset_supplementalproperty_node = NULL;
> - xmlNodePtr node = NULL;
> -
> - ret = parse_manifest_adaptationset_attr(s, adaptionset_node);
> - if (ret < 0)
> - return ret;
> -
> - node = xmlFirstElementChild(adaptionset_node);
> - while (node) {
> - if (!av_strcasecmp(node->name, "SegmentTemplate")) {
> - fragment_template_node = node;
> - } else if (!av_strcasecmp(node->name, "ContentComponent")) {
> - content_component_node = node;
> - } else if (!av_strcasecmp(node->name, "BaseURL")) {
> - adaptionset_baseurl_node = node;
> - } else if (!av_strcasecmp(node->name, "SegmentList")) {
> - adaptionset_segmentlist_node = node;
> - } else if (!av_strcasecmp(node->name, "SupplementalProperty")) {
> - adaptionset_supplementalproperty_node = node;
> - } else if (!av_strcasecmp(node->name, "Representation")) {
> - ret = parse_manifest_representation(s, url, node,
> - adaptionset_node,
> - mpd_baseurl_node,
> - period_baseurl_node,
> - period_segmenttemplate_node,
> - period_segmentlist_node,
> - fragment_template_node,
> - content_component_node,
> - adaptionset_baseurl_node,
> - adaptionset_segmentlist_node,
> - adaptionset_supplementalproperty_node);
> - if (ret < 0)
> - goto err;
> +#if CONFIG_HTTP_PROTOCOL
> + if (!(uc = ffio_geturlcontext(rep->segments[seg].pb)) ||
> + (rep->segments[seg].pb->eof_reached = 0,
> + rep->segments[seg].pb->pos = 0,
> + (ret = ff_http_do_new_request2(uc, url, &opts)) < 0))
> +#endif
> + {
> + ff_format_io_close(s, &rep->segments[seg].pb);
> + ret = s->io_open(s, &rep->segments[seg].pb, url, AVIO_FLAG_READ, &opts);
> }
> - node = xmlNextElementSibling(node);
> - }
> -
> -err:
> - xmlFree(c->adaptionset_lang);
> - c->adaptionset_lang = NULL;
> - return ret;
> -}
> -
> -static int parse_programinformation(AVFormatContext *s, xmlNodePtr node)
> -{
> - xmlChar *val = NULL;
> -
> - node = xmlFirstElementChild(node);
> - while (node) {
> - if (!av_strcasecmp(node->name, "Title")) {
> - val = xmlNodeGetContent(node);
> - if (val) {
> - av_dict_set(&s->metadata, "Title", val, 0);
> - }
> - } else if (!av_strcasecmp(node->name, "Source")) {
> - val = xmlNodeGetContent(node);
> - if (val) {
> - av_dict_set(&s->metadata, "Source", val, 0);
> - }
> - } else if (!av_strcasecmp(node->name, "Copyright")) {
> - val = xmlNodeGetContent(node);
> - if (val) {
> - av_dict_set(&s->metadata, "Copyright", val, 0);
> - }
> - }
> - node = xmlNextElementSibling(node);
> - xmlFree(val);
> - val = NULL;
> - }
> - return 0;
> -}
> -
> -static int parse_manifest(AVFormatContext *s, const char *url, AVIOContext *in)
> -{
> - DASHContext *c = s->priv_data;
> - int ret = 0;
> - int close_in = 0;
> - int64_t filesize = 0;
> - AVBPrint buf;
> - AVDictionary *opts = NULL;
> - xmlDoc *doc = NULL;
> - xmlNodePtr root_element = NULL;
> - xmlNodePtr node = NULL;
> - xmlNodePtr period_node = NULL;
> - xmlNodePtr tmp_node = NULL;
> - xmlNodePtr mpd_baseurl_node = NULL;
> - xmlNodePtr period_baseurl_node = NULL;
> - xmlNodePtr period_segmenttemplate_node = NULL;
> - xmlNodePtr period_segmentlist_node = NULL;
> - xmlNodePtr adaptionset_node = NULL;
> - xmlAttrPtr attr = NULL;
> - char *val = NULL;
> - uint32_t period_duration_sec = 0;
> - uint32_t period_start_sec = 0;
> -
> - if (!in) {
> - close_in = 1;
>
> - av_dict_copy(&opts, c->avio_opts, 0);
> - ret = avio_open2(&in, url, AVIO_FLAG_READ, c->interrupt_callback, &opts);
> + if (template)
> + av_free(url);
> av_dict_free(&opts);
> - if (ret < 0)
> - return ret;
> - }
>
> - if (av_opt_get(in, "location", AV_OPT_SEARCH_CHILDREN, (uint8_t**)&c->base_url) < 0)
> - c->base_url = av_strdup(url);
> -
> - filesize = avio_size(in);
> - filesize = filesize > 0 ? filesize : DEFAULT_MANIFEST_SIZE;
> -
> - if (filesize > MAX_BPRINT_READ_SIZE) {
> - av_log(s, AV_LOG_ERROR, "Manifest too large: %"PRId64"\n", filesize);
> - return AVERROR_INVALIDDATA;
> - }
> -
> - av_bprint_init(&buf, filesize + 1, AV_BPRINT_SIZE_UNLIMITED);
> -
> - if ((ret = avio_read_to_bprint(in, &buf, MAX_BPRINT_READ_SIZE)) < 0 ||
> - !avio_feof(in) ||
> - (filesize = buf.len) == 0) {
> - av_log(s, AV_LOG_ERROR, "Unable to read to manifest '%s'\n", url);
> - if (ret == 0)
> - ret = AVERROR_INVALIDDATA;
> - } else {
> - LIBXML_TEST_VERSION
> -
> - doc = xmlReadMemory(buf.str, filesize, c->base_url, NULL, 0);
> - root_element = xmlDocGetRootElement(doc);
> - node = root_element;
> -
> - if (!node) {
> - ret = AVERROR_INVALIDDATA;
> - av_log(s, AV_LOG_ERROR, "Unable to parse '%s' - missing root node\n", url);
> - goto cleanup;
> - }
> -
> - if (node->type != XML_ELEMENT_NODE ||
> - av_strcasecmp(node->name, "MPD")) {
> - ret = AVERROR_INVALIDDATA;
> - av_log(s, AV_LOG_ERROR, "Unable to parse '%s' - wrong root node name[%s] type[%d]\n", url, node->name, (int)node->type);
> - goto cleanup;
> - }
> -
> - val = xmlGetProp(node, "type");
> - if (!val) {
> - av_log(s, AV_LOG_ERROR, "Unable to parse '%s' - missing type attrib\n", url);
> - ret = AVERROR_INVALIDDATA;
> - goto cleanup;
> - }
> - if (!av_strcasecmp(val, "dynamic"))
> - c->is_live = 1;
> - xmlFree(val);
> -
> - attr = node->properties;
> - while (attr) {
> - val = xmlGetProp(node, attr->name);
> -
> - if (!av_strcasecmp(attr->name, "availabilityStartTime")) {
> - c->availability_start_time = get_utc_date_time_insec(s, val);
> - av_log(s, AV_LOG_TRACE, "c->availability_start_time = [%"PRId64"]\n", c->availability_start_time);
> - } else if (!av_strcasecmp(attr->name, "availabilityEndTime")) {
> - c->availability_end_time = get_utc_date_time_insec(s, val);
> - av_log(s, AV_LOG_TRACE, "c->availability_end_time = [%"PRId64"]\n", c->availability_end_time);
> - } else if (!av_strcasecmp(attr->name, "publishTime")) {
> - c->publish_time = get_utc_date_time_insec(s, val);
> - av_log(s, AV_LOG_TRACE, "c->publish_time = [%"PRId64"]\n", c->publish_time);
> - } else if (!av_strcasecmp(attr->name, "minimumUpdatePeriod")) {
> - c->minimum_update_period = get_duration_insec(s, val);
> - av_log(s, AV_LOG_TRACE, "c->minimum_update_period = [%"PRId64"]\n", c->minimum_update_period);
> - } else if (!av_strcasecmp(attr->name, "timeShiftBufferDepth")) {
> - c->time_shift_buffer_depth = get_duration_insec(s, val);
> - av_log(s, AV_LOG_TRACE, "c->time_shift_buffer_depth = [%"PRId64"]\n", c->time_shift_buffer_depth);
> - } else if (!av_strcasecmp(attr->name, "minBufferTime")) {
> - c->min_buffer_time = get_duration_insec(s, val);
> - av_log(s, AV_LOG_TRACE, "c->min_buffer_time = [%"PRId64"]\n", c->min_buffer_time);
> - } else if (!av_strcasecmp(attr->name, "suggestedPresentationDelay")) {
> - c->suggested_presentation_delay = get_duration_insec(s, val);
> - av_log(s, AV_LOG_TRACE, "c->suggested_presentation_delay = [%"PRId64"]\n", c->suggested_presentation_delay);
> - } else if (!av_strcasecmp(attr->name, "mediaPresentationDuration")) {
> - c->media_presentation_duration = get_duration_insec(s, val);
> - av_log(s, AV_LOG_TRACE, "c->media_presentation_duration = [%"PRId64"]\n", c->media_presentation_duration);
> - }
> - attr = attr->next;
> - xmlFree(val);
> - }
> -
> - tmp_node = find_child_node_by_name(node, "BaseURL");
> - if (tmp_node) {
> - mpd_baseurl_node = xmlCopyNode(tmp_node,1);
> - } else {
> - mpd_baseurl_node = xmlNewNode(NULL, "BaseURL");
> - }
> -
> - // at now we can handle only one period, with the longest duration
> - node = xmlFirstElementChild(node);
> - while (node) {
> - if (!av_strcasecmp(node->name, "Period")) {
> - period_duration_sec = 0;
> - period_start_sec = 0;
> - attr = node->properties;
> - while (attr) {
> - val = xmlGetProp(node, attr->name);
> - if (!av_strcasecmp(attr->name, "duration")) {
> - period_duration_sec = get_duration_insec(s, val);
> - } else if (!av_strcasecmp(attr->name, "start")) {
> - period_start_sec = get_duration_insec(s, val);
> - }
> - attr = attr->next;
> - xmlFree(val);
> - }
> - if ((period_duration_sec) >= (c->period_duration)) {
> - period_node = node;
> - c->period_duration = period_duration_sec;
> - c->period_start = period_start_sec;
> - if (c->period_start > 0)
> - c->media_presentation_duration = c->period_duration;
> - }
> - } else if (!av_strcasecmp(node->name, "ProgramInformation")) {
> - parse_programinformation(s, node);
> - }
> - node = xmlNextElementSibling(node);
> - }
> - if (!period_node) {
> - av_log(s, AV_LOG_ERROR, "Unable to parse '%s' - missing Period node\n", url);
> - ret = AVERROR_INVALIDDATA;
> - goto cleanup;
> - }
> -
> - adaptionset_node = xmlFirstElementChild(period_node);
> - while (adaptionset_node) {
> - if (!av_strcasecmp(adaptionset_node->name, "BaseURL")) {
> - period_baseurl_node = adaptionset_node;
> - } else if (!av_strcasecmp(adaptionset_node->name, "SegmentTemplate")) {
> - period_segmenttemplate_node = adaptionset_node;
> - } else if (!av_strcasecmp(adaptionset_node->name, "SegmentList")) {
> - period_segmentlist_node = adaptionset_node;
> - } else if (!av_strcasecmp(adaptionset_node->name, "AdaptationSet")) {
> - parse_manifest_adaptationset(s, url, adaptionset_node, mpd_baseurl_node, period_baseurl_node, period_segmenttemplate_node, period_segmentlist_node);
> - }
> - adaptionset_node = xmlNextElementSibling(adaptionset_node);
> - }
> -cleanup:
> - /*free the document */
> - xmlFreeDoc(doc);
> - xmlCleanupParser();
> - xmlFreeNode(mpd_baseurl_node);
> - }
> -
> - av_bprint_finalize(&buf, NULL);
> - if (close_in) {
> - avio_close(in);
> - }
> - return ret;
> -}
> -
> -static int64_t calc_cur_seg_no(AVFormatContext *s, struct representation *pls)
> -{
> - DASHContext *c = s->priv_data;
> - int64_t num = 0;
> - int64_t start_time_offset = 0;
> -
> - if (c->is_live) {
> - if (pls->n_fragments) {
> - av_log(s, AV_LOG_TRACE, "in n_fragments mode\n");
> - num = pls->first_seq_no;
> - } else if (pls->n_timelines) {
> - av_log(s, AV_LOG_TRACE, "in n_timelines mode\n");
> - start_time_offset = get_segment_start_time_based_on_timeline(pls, 0xFFFFFFFF) - 60 * pls->fragment_timescale; // 60 seconds before end
> - num = calc_next_seg_no_from_timelines(pls, start_time_offset);
> - if (num == -1)
> - num = pls->first_seq_no;
> - else
> - num += pls->first_seq_no;
> - } else if (pls->fragment_duration){
> - av_log(s, AV_LOG_TRACE, "in fragment_duration mode fragment_timescale = %"PRId64", presentation_timeoffset = %"PRId64"\n", pls->fragment_timescale, pls->presentation_timeoffset);
> - if (pls->presentation_timeoffset) {
> - num = pls->first_seq_no + (((get_current_time_in_sec() - c->availability_start_time) * pls->fragment_timescale)-pls->presentation_timeoffset) / pls->fragment_duration - c->min_buffer_time;
> - } else if (c->publish_time > 0 && !c->availability_start_time) {
> - if (c->min_buffer_time) {
> - num = pls->first_seq_no + (((c->publish_time + pls->fragment_duration) - c->suggested_presentation_delay) * pls->fragment_timescale) / pls->fragment_duration - c->min_buffer_time;
> - } else {
> - num = pls->first_seq_no + (((c->publish_time - c->time_shift_buffer_depth + pls->fragment_duration) - c->suggested_presentation_delay) * pls->fragment_timescale) / pls->fragment_duration;
> - }
> - } else {
> - num = pls->first_seq_no + (((get_current_time_in_sec() - c->availability_start_time) - c->suggested_presentation_delay) * pls->fragment_timescale) / pls->fragment_duration;
> - }
> - }
> - } else {
> - num = pls->first_seq_no;
> - }
> - return num;
> -}
> -
> -static int64_t calc_min_seg_no(AVFormatContext *s, struct representation *pls)
> -{
> - DASHContext *c = s->priv_data;
> - int64_t num = 0;
> -
> - if (c->is_live && pls->fragment_duration) {
> - av_log(s, AV_LOG_TRACE, "in live mode\n");
> - num = pls->first_seq_no + (((get_current_time_in_sec() - c->availability_start_time) - c->time_shift_buffer_depth) * pls->fragment_timescale) / pls->fragment_duration;
> - } else {
> - num = pls->first_seq_no;
> - }
> - return num;
> -}
> -
> -static int64_t calc_max_seg_no(struct representation *pls, DASHContext *c)
> -{
> - int64_t num = 0;
> -
> - if (pls->n_fragments) {
> - num = pls->first_seq_no + pls->n_fragments - 1;
> - } else if (pls->n_timelines) {
> - int i = 0;
> - num = pls->first_seq_no + pls->n_timelines - 1;
> - for (i = 0; i < pls->n_timelines; i++) {
> - if (pls->timelines[i]->repeat == -1) {
> - int length_of_each_segment = pls->timelines[i]->duration / pls->fragment_timescale;
> - num = c->period_duration / length_of_each_segment;
> - } else {
> - num += pls->timelines[i]->repeat;
> - }
> - }
> - } else if (c->is_live && pls->fragment_duration) {
> - num = pls->first_seq_no + (((get_current_time_in_sec() - c->availability_start_time)) * pls->fragment_timescale) / pls->fragment_duration;
> - } else if (pls->fragment_duration) {
> - num = pls->first_seq_no + (c->media_presentation_duration * pls->fragment_timescale) / pls->fragment_duration;
> - }
> -
> - return num;
> -}
> -
> -static void move_timelines(struct representation *rep_src, struct representation *rep_dest, DASHContext *c)
> -{
> - if (rep_dest && rep_src ) {
> - free_timelines_list(rep_dest);
> - rep_dest->timelines = rep_src->timelines;
> - rep_dest->n_timelines = rep_src->n_timelines;
> - rep_dest->first_seq_no = rep_src->first_seq_no;
> - rep_dest->last_seq_no = calc_max_seg_no(rep_dest, c);
> - rep_src->timelines = NULL;
> - rep_src->n_timelines = 0;
> - rep_dest->cur_seq_no = rep_src->cur_seq_no;
> - }
> -}
> -
> -static void move_segments(struct representation *rep_src, struct representation *rep_dest, DASHContext *c)
> -{
> - if (rep_dest && rep_src ) {
> - free_fragment_list(rep_dest);
> - if (rep_src->start_number > (rep_dest->start_number + rep_dest->n_fragments))
> - rep_dest->cur_seq_no = 0;
> - else
> - rep_dest->cur_seq_no += rep_src->start_number - rep_dest->start_number;
> - rep_dest->fragments = rep_src->fragments;
> - rep_dest->n_fragments = rep_src->n_fragments;
> - rep_dest->parent = rep_src->parent;
> - rep_dest->last_seq_no = calc_max_seg_no(rep_dest, c);
> - rep_src->fragments = NULL;
> - rep_src->n_fragments = 0;
> - }
> -}
> -
> -
> -static int refresh_manifest(AVFormatContext *s)
> -{
> - int ret = 0, i;
> - DASHContext *c = s->priv_data;
> - // save current context
> - int n_videos = c->n_videos;
> - struct representation **videos = c->videos;
> - int n_audios = c->n_audios;
> - struct representation **audios = c->audios;
> - int n_subtitles = c->n_subtitles;
> - struct representation **subtitles = c->subtitles;
> - char *base_url = c->base_url;
> -
> - c->base_url = NULL;
> - c->n_videos = 0;
> - c->videos = NULL;
> - c->n_audios = 0;
> - c->audios = NULL;
> - c->n_subtitles = 0;
> - c->subtitles = NULL;
> - ret = parse_manifest(s, s->url, NULL);
> - if (ret)
> - goto finish;
> -
> - if (c->n_videos != n_videos) {
> - av_log(c, AV_LOG_ERROR,
> - "new manifest has mismatched no. of video representations, %d -> %d\n",
> - n_videos, c->n_videos);
> - return AVERROR_INVALIDDATA;
> - }
> - if (c->n_audios != n_audios) {
> - av_log(c, AV_LOG_ERROR,
> - "new manifest has mismatched no. of audio representations, %d -> %d\n",
> - n_audios, c->n_audios);
> - return AVERROR_INVALIDDATA;
> - }
> - if (c->n_subtitles != n_subtitles) {
> - av_log(c, AV_LOG_ERROR,
> - "new manifest has mismatched no. of subtitles representations, %d -> %d\n",
> - n_subtitles, c->n_subtitles);
> - return AVERROR_INVALIDDATA;
> - }
> -
> - for (i = 0; i < n_videos; i++) {
> - struct representation *cur_video = videos[i];
> - struct representation *ccur_video = c->videos[i];
> - if (cur_video->timelines) {
> - // calc current time
> - int64_t currentTime = get_segment_start_time_based_on_timeline(cur_video, cur_video->cur_seq_no) / cur_video->fragment_timescale;
> - // update segments
> - ccur_video->cur_seq_no = calc_next_seg_no_from_timelines(ccur_video, currentTime * cur_video->fragment_timescale - 1);
> - if (ccur_video->cur_seq_no >= 0) {
> - move_timelines(ccur_video, cur_video, c);
> - }
> - }
> - if (cur_video->fragments) {
> - move_segments(ccur_video, cur_video, c);
> - }
> - }
> - for (i = 0; i < n_audios; i++) {
> - struct representation *cur_audio = audios[i];
> - struct representation *ccur_audio = c->audios[i];
> - if (cur_audio->timelines) {
> - // calc current time
> - int64_t currentTime = get_segment_start_time_based_on_timeline(cur_audio, cur_audio->cur_seq_no) / cur_audio->fragment_timescale;
> - // update segments
> - ccur_audio->cur_seq_no = calc_next_seg_no_from_timelines(ccur_audio, currentTime * cur_audio->fragment_timescale - 1);
> - if (ccur_audio->cur_seq_no >= 0) {
> - move_timelines(ccur_audio, cur_audio, c);
> - }
> - }
> - if (cur_audio->fragments) {
> - move_segments(ccur_audio, cur_audio, c);
> - }
> - }
> -
> -finish:
> - // restore context
> - if (c->base_url)
> - av_free(base_url);
> - else
> - c->base_url = base_url;
> -
> - if (c->subtitles)
> - free_subtitle_list(c);
> - if (c->audios)
> - free_audio_list(c);
> - if (c->videos)
> - free_video_list(c);
> -
> - c->n_subtitles = n_subtitles;
> - c->subtitles = subtitles;
> - c->n_audios = n_audios;
> - c->audios = audios;
> - c->n_videos = n_videos;
> - c->videos = videos;
> - return ret;
> -}
> -
> -static struct fragment *get_current_fragment(struct representation *pls)
> -{
> - int64_t min_seq_no = 0;
> - int64_t max_seq_no = 0;
> - struct fragment *seg = NULL;
> - struct fragment *seg_ptr = NULL;
> - DASHContext *c = pls->parent->priv_data;
> -
> - while (( !ff_check_interrupt(c->interrupt_callback)&& pls->n_fragments > 0)) {
> - if (pls->cur_seq_no < pls->n_fragments) {
> - seg_ptr = pls->fragments[pls->cur_seq_no];
> - seg = av_mallocz(sizeof(struct fragment));
> - if (!seg) {
> - return NULL;
> - }
> - seg->url = av_strdup(seg_ptr->url);
> - if (!seg->url) {
> - av_free(seg);
> - return NULL;
> - }
> - seg->size = seg_ptr->size;
> - seg->url_offset = seg_ptr->url_offset;
> - return seg;
> - } else if (c->is_live) {
> - refresh_manifest(pls->parent);
> - } else {
> - break;
> - }
> - }
> - if (c->is_live) {
> - min_seq_no = calc_min_seg_no(pls->parent, pls);
> - max_seq_no = calc_max_seg_no(pls, c);
> -
> - if (pls->timelines || pls->fragments) {
> - refresh_manifest(pls->parent);
> - }
> - if (pls->cur_seq_no <= min_seq_no) {
> - av_log(pls->parent, AV_LOG_VERBOSE, "old fragment: cur[%"PRId64"] min[%"PRId64"] max[%"PRId64"]\n", (int64_t)pls->cur_seq_no, min_seq_no, max_seq_no);
> - pls->cur_seq_no = calc_cur_seg_no(pls->parent, pls);
> - } else if (pls->cur_seq_no > max_seq_no) {
> - av_log(pls->parent, AV_LOG_VERBOSE, "new fragment: min[%"PRId64"] max[%"PRId64"]\n", min_seq_no, max_seq_no);
> - }
> - seg = av_mallocz(sizeof(struct fragment));
> - if (!seg) {
> - return NULL;
> - }
> - } else if (pls->cur_seq_no <= pls->last_seq_no) {
> - seg = av_mallocz(sizeof(struct fragment));
> - if (!seg) {
> - return NULL;
> - }
> - }
> - if (seg) {
> - char *tmpfilename;
> - if (!pls->url_template) {
> - av_log(pls->parent, AV_LOG_ERROR, "Cannot get fragment, missing template URL\n");
> - av_free(seg);
> - return NULL;
> - }
> - tmpfilename = av_mallocz(c->max_url_size);
> - if (!tmpfilename) {
> - av_free(seg);
> - return NULL;
> - }
> - ff_dash_fill_tmpl_params(tmpfilename, c->max_url_size, pls->url_template, 0, pls->cur_seq_no, 0, get_segment_start_time_based_on_timeline(pls, pls->cur_seq_no));
> - seg->url = av_strireplace(pls->url_template, pls->url_template, tmpfilename);
> - if (!seg->url) {
> - av_log(pls->parent, AV_LOG_WARNING, "Unable to resolve template url '%s', try to use origin template\n", pls->url_template);
> - seg->url = av_strdup(pls->url_template);
> - if (!seg->url) {
> - av_log(pls->parent, AV_LOG_ERROR, "Cannot resolve template url '%s'\n", pls->url_template);
> - av_free(tmpfilename);
> - av_free(seg);
> - return NULL;
> - }
> - }
> - av_free(tmpfilename);
> - seg->size = -1;
> - }
> -
> - return seg;
> -}
> -
> -static int read_from_url(struct representation *pls, struct fragment *seg,
> - uint8_t *buf, int buf_size)
> -{
> - int ret;
> -
> - /* limit read if the fragment was only a part of a file */
> - if (seg->size >= 0)
> - buf_size = FFMIN(buf_size, pls->cur_seg_size - pls->cur_seg_offset);
> -
> - ret = avio_read(pls->input, buf, buf_size);
> - if (ret > 0)
> - pls->cur_seg_offset += ret;
> -
> - return ret;
> -}
> -
> -static int open_input(DASHContext *c, struct representation *pls, struct fragment *seg)
> -{
> - AVDictionary *opts = NULL;
> - char *url = NULL;
> - int ret = 0;
> -
> - url = av_mallocz(c->max_url_size);
> - if (!url) {
> - ret = AVERROR(ENOMEM);
> - goto cleanup;
> - }
> -
> - if (seg->size >= 0) {
> - /* try to restrict the HTTP request to the part we want
> - * (if this is in fact a HTTP request) */
> - av_dict_set_int(&opts, "offset", seg->url_offset, 0);
> - av_dict_set_int(&opts, "end_offset", seg->url_offset + seg->size, 0);
> - }
> -
> - ff_make_absolute_url(url, c->max_url_size, c->base_url, seg->url);
> - av_log(pls->parent, AV_LOG_VERBOSE, "DASH request for url '%s', offset %"PRId64"\n",
> - url, seg->url_offset);
> - ret = open_url(pls->parent, &pls->input, url, &c->avio_opts, opts, NULL);
> -
> -cleanup:
> - av_free(url);
> - av_dict_free(&opts);
> - pls->cur_seg_offset = 0;
> - pls->cur_seg_size = seg->size;
> - return ret;
> -}
> -
> -static int update_init_section(struct representation *pls)
> -{
> - static const int max_init_section_size = 1024 * 1024;
> - DASHContext *c = pls->parent->priv_data;
> - int64_t sec_size;
> - int64_t urlsize;
> - int ret;
> -
> - if (!pls->init_section || pls->init_sec_buf)
> - return 0;
> -
> - ret = open_input(c, pls, pls->init_section);
> - if (ret < 0) {
> - av_log(pls->parent, AV_LOG_WARNING,
> - "Failed to open an initialization section\n");
> - return ret;
> - }
> -
> - if (pls->init_section->size >= 0)
> - sec_size = pls->init_section->size;
> - else if ((urlsize = avio_size(pls->input)) >= 0)
> - sec_size = urlsize;
> - else
> - sec_size = max_init_section_size;
> -
> - av_log(pls->parent, AV_LOG_DEBUG,
> - "Downloading an initialization section of size %"PRId64"\n",
> - sec_size);
> -
> - sec_size = FFMIN(sec_size, max_init_section_size);
> -
> - av_fast_malloc(&pls->init_sec_buf, &pls->init_sec_buf_size, sec_size);
> -
> - ret = read_from_url(pls, pls->init_section, pls->init_sec_buf,
> - pls->init_sec_buf_size);
> - ff_format_io_close(pls->parent, &pls->input);
> -
> - if (ret < 0)
> - return ret;
> -
> - pls->init_sec_data_len = ret;
> - pls->init_sec_buf_read_offset = 0;
> -
> - return 0;
> -}
> -
> -static int64_t seek_data(void *opaque, int64_t offset, int whence)
> -{
> - struct representation *v = opaque;
> - if (v->n_fragments && !v->init_sec_data_len) {
> - return avio_seek(v->input, offset, whence);
> - }
> -
> - return AVERROR(ENOSYS);
> -}
> -
> -static int read_data(void *opaque, uint8_t *buf, int buf_size)
> -{
> - int ret = 0;
> - struct representation *v = opaque;
> - DASHContext *c = v->parent->priv_data;
> -
> -restart:
> - if (!v->input) {
> - free_fragment(&v->cur_seg);
> - v->cur_seg = get_current_fragment(v);
> - if (!v->cur_seg) {
> - ret = AVERROR_EOF;
> - goto end;
> - }
> -
> - /* load/update Media Initialization Section, if any */
> - ret = update_init_section(v);
> - if (ret)
> - goto end;
> -
> - ret = open_input(c, v, v->cur_seg);
> if (ret < 0) {
> - if (ff_check_interrupt(c->interrupt_callback)) {
> - ret = AVERROR_EXIT;
> - goto end;
> + fail_segment:
> + if (seg == rep->cur_segment) {
> + /* We know that we have not reached the end. */
> + if (rep->read_ts < period->period->end_ts) {
> + if (0) {
> + /* TODO: Implement some failure resistance in case of
> + * specific HTTP response codes. */
> + goto open_segment;
> + } else if (!c->is_live && (ret == AVERROR(ETIMEDOUT) || ret == AVERROR(EIO))) {
> + /* Allow dropping some segments only for live streams. */
> + goto open_segment;
> + }
> + }
> +
> + return ret;
> + } else {
> + goto read_segment;
> }
> - av_log(v->parent, AV_LOG_WARNING, "Failed to open fragment of playlist\n");
> - v->cur_seq_no++;
> - goto restart;
> }
> - }
>
> - if (v->init_sec_buf_read_offset < v->init_sec_data_len) {
> - /* Push init section out first before first actual fragment */
> - int copy_size = FFMIN(v->init_sec_data_len - v->init_sec_buf_read_offset, buf_size);
> - memcpy(buf, v->init_sec_buf, copy_size);
> - v->init_sec_buf_read_offset += copy_size;
> - ret = copy_size;
> - goto end;
> - }
> + if (0 < range.start &&
> + avio_seek(s->pb, range.start, SEEK_SET) != range.start)
> + {
> + ret = AVERROR(EIO);
> + goto fail_segment;
> + }
>
> - /* check the v->cur_seg, if it is null, get current and double check if the new v->cur_seg*/
> - if (!v->cur_seg) {
> - v->cur_seg = get_current_fragment(v);
> - }
> - if (!v->cur_seg) {
> - ret = AVERROR_EOF;
> - goto end;
> - }
> - ret = read_from_url(v, v->cur_seg, buf, buf_size);
> - if (ret > 0)
> - goto end;
> + rep->segments[seg].segment_size = range.end - range.start;
> + } while (rep->cur_segment != (seg = (seg + 1) % c->nb_connections) &&
> + /* Use only a single IO context on startup. */
> + rep->ic->iformat);
> +
> +read_segment:;
> + /* Only the current segment can be read. */
> + size = FFMIN(buf_size, (int)FFMIN(rep->segments[rep->cur_segment].segment_size, (uint64_t)INT_MAX));
> + av_assert1(0 < size);
> + ret = avio_read_partial(rep->segments[rep->cur_segment].pb, buf, size);
> + if (0 < ret) {
> + rep->segments[rep->cur_segment].segment_size -= ret;
> + if (rep->save_init) {
> + uint8_t *p;
> + /* How much bytes can we store? */
> + int append_size = FFMIN(rep->cur_period->initbuf_size + ret, INITBUF_MAX) -
> + rep->cur_period->initbuf_size;
>
> - if (c->is_live || v->cur_seq_no < v->last_seq_no) {
> - if (!v->is_restart_needed)
> - v->cur_seq_no++;
> - v->is_restart_needed = 1;
> + if ((p = av_realloc(rep->cur_period->initbuf, rep->cur_period->initbuf_size + append_size))) {
> + memcpy((rep->cur_period->initbuf = p) + rep->cur_period->initbuf_size, buf, append_size);
> + rep->cur_period->initbuf_size += append_size;
> + } else {
> + /* Do not save anything furthermore otherwise cached init data
> + * would may be restored with a missing intermediate chunk. */
> + rep->save_init = 0;
> + /* Heck. Reset stored stuff. We can only handle correctly
> + * received init buffer. */
> + rep->cur_period->initbuf_size = 0;
> + }
> + }
> + return ret;
> + } else if (ret == AVERROR_EOF) {
> + /* Go to next segment. */
> + rep->cur_segment = (rep->cur_segment + 1) % c->nb_connections;
> + rep->save_init = 0;
> + } else if (ret == AVERROR(ETIMEDOUT)) {
> + dash_rotate_urllist(s, rep->cur_period->base);
> + } else if (ret == 0) {
> + if (s->flags & AVFMT_FLAG_NONBLOCK) {
> + return AVERROR(EAGAIN);
> + } else {
> + av_usleep(150 * MILLISEC_PER_SEC);
> + goto read_segment;
> + }
> + } else {
> + return ret;
> }
>
> -end:
> - return ret;
> + goto open_segment;
> }
>
> -static int save_avio_options(AVFormatContext *s)
> +static av_cold int dash_subdemuxer_block_io_open(AVFormatContext *s, AVIOContext **pb, const char *url,
> + int flags, AVDictionary **opts)
> {
> DASHContext *c = s->priv_data;
> - const char *opts[] = {
> - "headers", "user_agent", "cookies", "http_proxy", "referer", "rw_timeout", "icy", NULL };
> - const char **opt = opts;
> - uint8_t *buf = NULL;
> - int ret = 0;
> -
> - while (*opt) {
> - if (av_opt_get(s->pb, *opt, AV_OPT_SEARCH_CHILDREN, &buf) >= 0) {
> - if (buf[0] != '\0') {
> - ret = av_dict_set(&c->avio_opts, *opt, buf, AV_DICT_DONT_STRDUP_VAL);
> - if (ret < 0)
> - return ret;
> - } else {
> - av_freep(&buf);
> - }
> - }
> - opt++;
> - }
> -
> - return ret;
> -}
> -
> -static int nested_io_open(AVFormatContext *s, AVIOContext **pb, const char *url,
> - int flags, AVDictionary **opts)
> -{
> av_log(s, AV_LOG_ERROR,
> - "A DASH playlist item '%s' referred to an external file '%s'. "
> - "Opening this file was forbidden for security reasons\n",
> - s->url, url);
> + "Opening external resource '%s' from DASH segment '%s' has been blocked for security reasons",
> + url, c->location);
> + av_log(s, AV_LOG_VERBOSE,
> + "To avoid this security guard open MPD manifest using file: protocol");
> return AVERROR(EPERM);
> }
>
> -static void close_demux_for_component(struct representation *pls)
> -{
> - /* note: the internal buffer could have changed */
> - av_freep(&pls->pb.buffer);
> - memset(&pls->pb, 0x00, sizeof(AVIOContext));
> - pls->ctx->pb = NULL;
> - avformat_close_input(&pls->ctx);
> -}
> -
> -static int reopen_demux_for_component(AVFormatContext *s, struct representation *pls)
> +static av_cold int dash_subdemuxer_alloc(DASHRepresentation *rep)
> {
> + int ret = 0;
> + AVFormatContext *s = rep->oc;
> DASHContext *c = s->priv_data;
> - ff_const59 AVInputFormat *in_fmt = NULL;
> - AVDictionary *in_fmt_opts = NULL;
> - uint8_t *avio_ctx_buffer = NULL;
> - int ret = 0, i;
> + uint8_t *buf;
> + int buf_size;
>
> - if (pls->ctx) {
> - close_demux_for_component(pls);
> - }
> + dash_subdemuxer_close(rep);
>
> - if (ff_check_interrupt(&s->interrupt_callback)) {
> - ret = AVERROR_EXIT;
> - goto fail;
> - }
> + /* Allocate space that is enough for the theoretically largest segment. */
> + buf_size = rep->periods[0]->bandwidth * c->max_segment_duration / AV_TIME_BASE;
> + /* When maximum is not enough. */
> + buf_size += buf_size / 16 /* +~5% */;
> + buf_size = FFMIN(FFMAX(buf_size, 4 * 1024), 256 * 1024);
>
> - if (!(pls->ctx = avformat_alloc_context())) {
> + if (!(buf = av_malloc(buf_size + AV_INPUT_BUFFER_PADDING_SIZE))) {
> ret = AVERROR(ENOMEM);
> - goto fail;
> + goto out;
> }
>
> - avio_ctx_buffer = av_malloc(INITIAL_BUFFER_SIZE);
> - if (!avio_ctx_buffer ) {
> - ret = AVERROR(ENOMEM);
> - avformat_free_context(pls->ctx);
> - pls->ctx = NULL;
> - goto fail;
> - }
> - ffio_init_context(&pls->pb, avio_ctx_buffer, INITIAL_BUFFER_SIZE, 0,
> - pls, read_data, NULL, c->is_live ? NULL : seek_data);
> - pls->pb.seekable = 0;
> -
> - if ((ret = ff_copy_whiteblacklists(pls->ctx, s)) < 0)
> - goto fail;
> -
> - pls->ctx->flags = AVFMT_FLAG_CUSTOM_IO;
> - pls->ctx->probesize = s->probesize > 0 ? s->probesize : 1024 * 4;
> - pls->ctx->max_analyze_duration = s->max_analyze_duration > 0 ? s->max_analyze_duration : 4 * AV_TIME_BASE;
> - pls->ctx->interrupt_callback = s->interrupt_callback;
> - ret = av_probe_input_buffer(&pls->pb, &in_fmt, "", NULL, 0, 0);
> - if (ret < 0) {
> - av_log(s, AV_LOG_ERROR, "Error when loading first fragment of playlist\n");
> - avformat_free_context(pls->ctx);
> - pls->ctx = NULL;
> - goto fail;
> + if ((ret = ffio_init_context(&rep->ic_pb, buf, buf_size, 0, rep,
> + dash_subdemuxer_read, NULL, dash_subdemuxer_seek) < 0))
> + {
> + /* Note: We purposely do not free I/O buffer since documentation says
> + * nothing about it. We presume inputs are freed on error. Just like in
> + * case av_dict_set() and avformat_alloc_context()... */
> + goto out;
> }
> + rep->ic_pb.seekable = 0;
>
> - pls->ctx->pb = &pls->pb;
> - pls->ctx->io_open = nested_io_open;
> + /* Padding bytes must be zeroed out. */
> + memset(buf + buf_size, 0, AV_INPUT_BUFFER_PADDING_SIZE);
>
> - // provide additional information from mpd if available
> - ret = avformat_open_input(&pls->ctx, "", in_fmt, &in_fmt_opts); //pls->init_section->url
> - av_dict_free(&in_fmt_opts);
> - if (ret < 0)
> - goto fail;
> - if (pls->n_fragments) {
> -#if FF_API_R_FRAME_RATE
> - if (pls->framerate.den) {
> - for (i = 0; i < pls->ctx->nb_streams; i++)
> - pls->ctx->streams[i]->r_frame_rate = pls->framerate;
> - }
> -#endif
> - ret = avformat_find_stream_info(pls->ctx, NULL);
> - if (ret < 0)
> - goto fail;
> + if (!(rep->ic = avformat_alloc_context())) {
> + ret = AVERROR(ENOMEM);
> + goto out;
> }
>
> -fail:
> + rep->ic->pb = &rep->ic_pb;
> + rep->ic->interrupt_callback = s->interrupt_callback;
> +
> + rep->ic->probesize = 0 < s->probesize
> + ? s->probesize
> + : 4 * 1024;
> + rep->ic->max_analyze_duration = 0 < s->max_analyze_duration
> + ? s->max_analyze_duration
> + : 4 * AV_TIME_BASE;
> +
> + /* We supply everything for the demuxer so normall it should not need to
> + * open anything. Apply this restriction unless we use "file" protocol. */
> + if (strcmp(avio_find_protocol_name(s->url), "file"))
> + rep->ic->io_open = dash_subdemuxer_block_io_open;
> +
> + if ((ret = ff_copy_whiteblacklists(rep->ic, s)) < 0)
> + goto out;
> +
> +out:
> + if (ret < 0)
> + dash_subdemuxer_close(rep);
> return ret;
> }
>
> -static int open_demux_for_component(AVFormatContext *s, struct representation *pls)
> +/**
> + * Open the real context.
> + */
> +static av_cold int dash_subdemuxer_open(DASHRepresentation *rep)
> {
> int ret = 0;
> - int i;
> + AVFormatContext *s = rep->oc;
> + AVInputFormat *iformat = NULL;
>
> - pls->parent = s;
> - pls->cur_seq_no = calc_cur_seg_no(s, pls);
> + av_log(s, AV_LOG_DEBUG, "Opening representation '%s'\n", rep->id);
>
> - if (!pls->last_seq_no) {
> - pls->last_seq_no = calc_max_seg_no(pls, s->priv_data);
> - }
> + if ((ret = dash_subdemuxer_alloc(rep)) < 0)
> + goto out;
>
> - ret = reopen_demux_for_component(s, pls);
> - if (ret < 0) {
> - goto fail;
> - }
> - for (i = 0; i < pls->ctx->nb_streams; i++) {
> - AVStream *st = avformat_new_stream(s, NULL);
> - AVStream *ist = pls->ctx->streams[i];
> - if (!st) {
> - ret = AVERROR(ENOMEM);
> - goto fail;
> - }
> - st->id = i;
> - avcodec_parameters_copy(st->codecpar, ist->codecpar);
> - avpriv_set_pts_info(st, ist->pts_wrap_bits, ist->time_base.num, ist->time_base.den);
> -
> - // copy disposition
> - st->disposition = ist->disposition;
> -
> - // copy side data
> - for (int i = 0; i < ist->nb_side_data; i++) {
> - const AVPacketSideData *sd_src = &ist->side_data[i];
> - uint8_t *dst_data;
> -
> - dst_data = av_stream_new_side_data(st, sd_src->type, sd_src->size);
> - if (!dst_data)
> - return AVERROR(ENOMEM);
> - memcpy(dst_data, sd_src->data, sd_src->size);
> - }
> + if ((ret = av_probe_input_buffer(&rep->ic_pb, &iformat, NULL, NULL, 0, 0)) < 0) {
> + av_log(s, AV_LOG_ERROR, "Probe failed for representation '%s'\n", rep->id);
> + goto out;
> }
>
> - return 0;
> -fail:
> + if ((ret = avformat_open_input(&rep->ic, NULL, iformat, NULL)) < 0)
> + goto out;
> +
> + /* Finding stream information may need some essential information, like
> + * encryption init info so stream metadata must be updated before. */
> + if ((ret = dash_subdemuxer_update(rep)))
> + goto out;
> +
> + if ((ret = avformat_find_stream_info(rep->ic, NULL)) < 0)
> + goto out;
> +
> + /* Though new streams may have been created we do not care about them now.
> + * If this is the case or stream information otherwise changed it will get
> + * updated after reading frames. */
> +
> +out:
> + if (ret < 0)
> + dash_subdemuxer_close(rep);
> return ret;
> }
>
> -static int is_common_init_section_exist(struct representation **pls, int n_pls)
> +static av_cold int dash_subdemuxer_init(DASHRepresentation *rep)
> {
> - struct fragment *first_init_section = pls[0]->init_section;
> - char *url =NULL;
> - int64_t url_offset = -1;
> - int64_t size = -1;
> - int i = 0;
> -
> - if (first_init_section == NULL || n_pls == 0)
> - return 0;
> -
> - url = first_init_section->url;
> - url_offset = first_init_section->url_offset;
> - size = pls[0]->init_section->size;
> - for (i=0;i<n_pls;i++) {
> - if (!pls[i]->init_section)
> - continue;
> -
> - if (av_strcasecmp(pls[i]->init_section->url, url) ||
> - pls[i]->init_section->url_offset != url_offset ||
> - pls[i]->init_section->size != size) {
> - return 0;
> - }
> + int ret;
> + AVFormatContext *s = rep->oc;
> + DASHContext *c = s->priv_data;
> +
> + if (c->is_live)
> + rep->read_ts = av_gettime() - FFMIN(10, c->time_shift_buffer_depth) * MICROSEC_PER_SEC;
> + else
> + rep->read_ts = c->start_ts;
> +
> + /* Create an inital stream for this representation though we do not what is
> + * really inside. We expect it to be an elementary stream however it do not
> + * have to be always true. In such cases the new streams will appear in
> + * dash_subdemuxer_open(). */
> + if (!avformat_new_stream(rep->ic, NULL)) {
> + ret = AVERROR(ENOMEM);
> + goto out;
> }
> - return 1;
> +
> + if ((ret = dash_subdemuxer_update(rep)) < 0)
> + goto out;
> +
> +out:
> + if (ret < 0)
> + dash_subdemuxer_close(rep);
> + return ret;
> }
>
> -static int copy_init_section(struct representation *rep_dest, struct representation *rep_src)
> +static void *av_push_array(void *plist, unsigned elem_size)
> {
> - rep_dest->init_sec_buf = av_mallocz(rep_src->init_sec_buf_size);
> - if (!rep_dest->init_sec_buf) {
> - av_log(rep_dest->ctx, AV_LOG_WARNING, "Cannot alloc memory for init_sec_buf\n");
> - return AVERROR(ENOMEM);
> - }
> - memcpy(rep_dest->init_sec_buf, rep_src->init_sec_buf, rep_src->init_sec_data_len);
> - rep_dest->init_sec_buf_size = rep_src->init_sec_buf_size;
> - rep_dest->init_sec_data_len = rep_src->init_sec_data_len;
> - rep_dest->cur_timestamp = rep_src->cur_timestamp;
> + void *p;
> + unsigned *nmemb = (unsigned *)((void **)plist + 1);
>
> - return 0;
> -}
> + if (!(p = av_realloc(*(void **)plist, (*nmemb + 1) * sizeof(void *))))
> + return NULL;
> + *(void **)plist = p;
>
> -static int dash_close(AVFormatContext *s);
> + return ((*(void ***)plist)[(*nmemb)++] = av_mallocz(elem_size));
> +}
>
> -static void move_metadata(AVStream *st, const char *key, char **value)
> +static av_cold void *dash_find_by_id(AVFormatContext *s, void *list, unsigned nmemb, const char *id)
> {
> - if (*value) {
> - av_dict_set(&st->metadata, key, *value, AV_DICT_DONT_STRDUP_VAL);
> - *value = NULL;
> + if (!id)
> + return NULL;
> +
> + for (unsigned i = 0; i < nmemb; ++i) {
> + void *elem = ((void **)list)[i];
> + xmlChar *elem_id = *(xmlChar **)elem;
> + if (elem_id && !strcmp(elem_id, id))
> + return elem;
> }
> +
> + return NULL;
> }
>
> -static int dash_read_header(AVFormatContext *s)
> +static av_cold int dash_push_id_node(AVFormatContext *s, void *elem, void *plist, unsigned elem_size, xmlNodePtr node)
> {
> - DASHContext *c = s->priv_data;
> - struct representation *rep;
> - AVProgram *program;
> - int ret = 0;
> - int stream_index = 0;
> - int i;
> -
> - c->interrupt_callback = &s->interrupt_callback;
> -
> - if ((ret = save_avio_options(s)) < 0)
> - goto fail;
> -
> - if ((ret = parse_manifest(s, s->url, s->pb)) < 0)
> - goto fail;
> -
> - /* If this isn't a live stream, fill the total duration of the
> - * stream. */
> - if (!c->is_live) {
> - s->duration = (int64_t) c->media_presentation_duration * AV_TIME_BASE;
> + int ret;
> + xmlChar *id = xmlGetNoNsProp(node, "id");
> + unsigned *nmemb = (unsigned *)((void **)plist + 1);
> +
> + if ((*(void **)elem = dash_find_by_id(s, *(void ***)plist, *nmemb, id))) {
> + ret = 1;
> + } else if ((*(void **)elem = av_push_array(plist, elem_size))) {
> + *(xmlChar **)(*(void **)elem) = id;
> + id = NULL;
> + ret = 0;
> } else {
> - av_dict_set(&c->avio_opts, "seekable", "0", 0);
> + ret = AVERROR(ENOMEM);
> }
>
> - if(c->n_videos)
> - c->is_init_section_common_video = is_common_init_section_exist(c->videos, c->n_videos);
> + xml_free(id);
> + return ret;
> +}
>
> - /* Open the demuxer for video and audio components if available */
> - for (i = 0; i < c->n_videos; i++) {
> - rep = c->videos[i];
> - if (i > 0 && c->is_init_section_common_video) {
> - ret = copy_init_section(rep, c->videos[0]);
> - if (ret < 0)
> - goto fail;
> - }
> - ret = open_demux_for_component(s, rep);
> +static int dash_parse_representation(AVFormatContext *s,
> + DASHPeriod *period,
> + DASHAdaptationSet *as,
> + DASHURLList *base_urls,
> + DASHTimeline *inherited_timeline,
> + const DASHParameters *inherited_par,
> + xmlNodePtr node)
> +{
> + int ret = 0;
> + DASHContext *c = s->priv_data;
> + DASHURLList *urls = dash_ref_urllist(base_urls);
> + DASHRepresentation *rep;
> + DASHRepresentationPeriod *rep_period;
> + DASHTimeline *timeline = NULL;
> + int64_t availability_time_offset = 0;
> +
> + if ((ret = dash_push_id_node(s, &rep, &c->reps, offsetof(DASHRepresentation, segments[c->nb_connections]), node)) < 0)
> + goto out;
> + else if (!ret) {
> + rep->oc = s;
> + }
> +
> + if (!(rep_period = av_push_array(&rep->periods, sizeof(DASHRepresentationPeriod)))) {
> + ret = AVERROR(ENOMEM);
> + goto out;
> + }
> +
> + rep_period->period = period;
> + rep_period->par = *inherited_par;
> +
> + xml_for_each_attr {
> + if (!strcmp(attr->name, "availabilityTimeOffset"))
> + (void)av_sscanf(value, "%"SCNd64, &availability_time_offset);
> + else if (!strcmp(attr->name, "bandwidth"))
> + (void)av_sscanf(value, "%"SCNu32, &rep_period->bandwidth);
> + else if (!dash_parse_parameters(s, &rep_period->par, attr, value))
> + /* Already done. */;
> + else if (!strcmp(attr->name, "startWithSAP") && !strcmp(value, "1"))
> + /* Just check. */;
> + else if (strcmp(attr->name, "id"))
> + dash_log_unknown_attr(s, attr, value);
> + }
> +
> + xml_for_each_child {
> + if (!strcmp(child->name, "BaseURL"))
> + ret = dash_parse_baseurl(s, &urls, child);
> + else if (!strcmp(child->name, "SegmentTemplate"))
> + ret = dash_parse_segmenttemplate(s, period, &timeline, child);
> + else if (!strcmp(child->name, "SegmentList"))
> + ret = dash_parse_segmentlist(s, period, &timeline, child);
> + else if (!strcmp(child->name, "SegmentBase"))
> + ret = dash_parse_segmentbase(s, period, &timeline, child);
> + else if (!strcmp(child->name, "AudioChannelConfiguration"))
> + ret = dash_parse_audiochannelconfiguration(s, &rep_period->par, child);
> + else if (!strcmp(child->name, "UTCTiming"))
> + ret = dash_parse_utctiming(s, child);
> + else
> + dash_log_unknown_child(s, child);
> +
> + if (ret < 0)
> + goto out;
> + }
> +
> + rep_period->as = as;
> + if (!timeline && inherited_timeline)
> + timeline = dash_ref_timeline(inherited_timeline);
> + /* Default timeline with one big segment. URL is obtained from BaseURLs. */
> + if (!timeline && (timeline = av_mallocz(sizeof(DASHTimeline)))) {
> + dash_ref_timeline(timeline);
> + timeline->type = TIMELINE_SEGMENTS;
> + timeline->duration = INT64_MAX;
> + timeline->init.range = DASH_RANGE_INITALIZER;
> + }
> + if (!timeline) {
> + ret = AVERROR(ENOMEM);
> + goto out;
> + }
> + rep_period->base = dash_ref_urllist(urls);
> + rep_period->timeline = dash_ref_timeline(timeline);
> +
> + if ((ret = dash_subdemuxer_alloc(rep)) < 0 ||
> + (ret = dash_subdemuxer_init(rep)) < 0)
> + {
> + av_log(s, AV_LOG_DEBUG, "Failed to initalize subdemuxer for representation '%s': %s\n",
> + rep->id, av_err2str(ret));
> + goto out;
> + }
> +
> +out:
> + dash_unref_timeline(timeline);
> + dash_unref_urllist(urls);
> + return ret;
> +}
> +
> +static av_cold int dash_parse_label(AVFormatContext *s,
> + DASHAdaptationSet *as,
> + xmlNodePtr node)
> +{
> + xmlChar *lang = NULL;
> + xmlChar *value;
> +
> + xml_for_each_attr {
> + if (!strcmp(attr->name, "lang")) {
> + lang = value;
> + value = NULL;
> + } else
> + dash_log_unknown_attr(s, attr, value);
> + }
> +
> + if ((value = xmlNodeGetContent(node)) && *value) {
> + char key_lang[100];
> + if (lang)
> + sprintf(key_lang, "label-%s", lang);
> + av_dict_set(&s->metadata, lang ? key_lang : "label", value, 0);
> + }
> +
> + xml_free(value);
> + xml_free(lang);
> + return 0;
> +}
> +
> +static av_cold int dash_parse_adaptationset(AVFormatContext *s,
> + DASHPeriod *period,
> + DASHURLList *base_urls,
> + DASHTimeline *inherited_timeline,
> + const DASHParameters *inherited_par,
> + xmlNodePtr node)
> +{
> + int ret = 0;
> + DASHContext *c = s->priv_data;
> + DASHContentProtection cp = DASH_CONTENTPROTECTION_INITIALIZER;
> + DASHURLList *urls = dash_ref_urllist(base_urls);
> + DASHAdaptationSet *as;
> + DASHTimeline *timeline = NULL;
> + DASHParameters par = *inherited_par;
> +
> + if ((ret = dash_push_id_node(s, &as, &c->ass, sizeof(DASHAdaptationSet), node)) < 0)
> + goto out;
> +
> + xml_for_each_attr {
> + if (!strcmp(attr->name, "lang"))
> + av_dict_set(&as->metadata, "language", value, 0);
> + else if (!strcmp(attr->name, "selectionPriority"))
> + av_dict_set(&as->metadata, "priority", value, 0);
> + else if (!strcmp(attr->name, "group"))
> + av_dict_set(&as->metadata, "group", value, 0);
> + else if (!dash_parse_parameters(s, &par, attr, value))
> + /* Already done. */;
> + else if (!strcmp(attr->name, "segmentAlignment") && !strcmp(value, "true"))
> + /* Just check. */;
> + else if (!strcmp(attr->name, "startWithSAP") && !strcmp(value, "1"))
> + /* Just check. */;
> + else if (strcmp(attr->name, "id") &&
> + strncmp(attr->name, "max", 3) &&
> + strncmp(attr->name, "min", 3))
> + dash_log_unknown_attr(s, attr, value);
> + }
> +
> + xml_for_each_child {
> + if (!strcmp(child->name, "SegmentTemplate"))
> + ret = dash_parse_segmenttemplate(s, period, &timeline, child);
> + else if (!strcmp(child->name, "SegmentList"))
> + ret = dash_parse_segmentlist(s, period, &timeline, child);
> + else if (!strcmp(child->name, "SegmentBase"))
> + ret = dash_parse_segmentbase(s, period, &timeline, child);
> + else if (!strcmp(child->name, "ContentProtection"))
> + ret = dash_parse_contentprotection(s, as, &cp, child);
> + else if (!strcmp(child->name, "BaseURL"))
> + ret = dash_parse_baseurl(s, &urls, child);
> + else if (!strcmp(child->name, "Role"))
> + ret = dash_parse_role(s, &par, child);
> + else if (!strcmp(child->name, "Viewpoint"))
> + ret = dash_parse_viewpoint(s, as, child);
> + else if (!strcmp(child->name, "SupplementalProperty") ||
> + !strcmp(child->name, "EssentialProperty"))
> + ret = dash_parse_property(s, as, NULL, child);
> + else if (!strcmp(child->name, "Representation"))
> + ret = dash_parse_representation(s, period, as, urls, (timeline ? timeline : inherited_timeline), &par, child);
> + else if (!strcmp(child->name, "AudioChannelConfiguration"))
> + ret = dash_parse_audiochannelconfiguration(s, &par, child);
> + else if (!strcmp(child->name, "Label"))
> + ret = dash_parse_label(s, as, child);
> + else
> + dash_log_unknown_child(s, child);
> +
> + if (ret < 0)
> + goto out;
> + }
> +
> +out:
> + dash_unref_timeline(timeline);
> + dash_unref_urllist(urls);
> + return ret;
> +}
> +
> +static av_cold int dash_parse_period(AVFormatContext *s,
> + DASHURLList *base_urls,
> + xmlNodePtr node)
> +{
> + int ret = 0;
> + DASHContext *c = s->priv_data;
> + DASHURLList *urls = dash_ref_urllist(base_urls);
> + DASHTimeline *timeline = NULL;
> + DASHPeriod *period;
> + int64_t duration = INT64_MAX;
> + DASHParameters par = DASH_PARAMETERS_INITIALIZER;
> +
> + if ((ret = dash_push_id_node(s, &period, &c->periods, sizeof(DASHPeriod), node)) < 0)
> + goto out;
> +
> + period->start_ts = 0;
> +
> + xml_for_each_attr {
> + if (!strcmp(attr->name, "start"))
> + period->start_ts = dash_parse_duration(value, s);
> + else if (!strcmp(attr->name, "duration"))
> + duration = dash_parse_duration(value, s);
> + else if (!dash_parse_parameters(s, &par, attr, value))
> + /* Already done. */;
> + else if (strcmp(attr->name, "id"))
> + dash_log_unknown_attr(s, attr, value);
> + }
> +
> + period->start_ts += c->availability_start_time;
> + period->end_ts = period->start_ts < INT64_MAX - duration
> + ? period->start_ts + duration
> + : INT64_MAX;
> + if (c->media_presentation_duration < 0)
> + s->duration = FFMAX(s->duration != AV_NOPTS_VALUE ? s->duration : 0, period->end_ts);
> + period->end_ts = FFMIN(period->end_ts, s->duration != AV_NOPTS_VALUE ? s->duration : 0);
> +
> + /* INT64_MAX means infinity, ergo unknown duration. It has to be stored
> + * differently for format. */
> + if (INT64_MAX == s->duration)
> + s->duration = AV_NOPTS_VALUE;
> +
> + /* Restrict duration of previous periods. */
> + for (unsigned i = 0; i < c->nb_periods; ++i) {
> + DASHPeriod *per = c->periods[i];
> + if (per->start_ts < period->start_ts)
> + per->end_ts = FFMIN(per->end_ts, period->start_ts);
> + }
>
> - if (ret)
> - goto fail;
> - rep->stream_index = stream_index;
> - ++stream_index;
> + if (period->end_ts <= period->start_ts) {
> + av_log(s, AV_LOG_DEBUG, "Ignoring empty Period %"PRId64" >= %"PRId64"\n",
> + period->start_ts, period->end_ts);
> + goto out;
> }
>
> - if(c->n_audios)
> - c->is_init_section_common_audio = is_common_init_section_exist(c->audios, c->n_audios);
> + xml_for_each_child {
> + if (!strcmp(child->name, "BaseURL"))
> + ret = dash_parse_baseurl(s, &urls, child);
> + else if (!strcmp(child->name, "SegmentTemplate"))
> + ret = dash_parse_segmenttemplate(s, period, &timeline, child);
> + else if (!strcmp(child->name, "SegmentList"))
> + ret = dash_parse_segmentlist(s, period, &timeline, child);
> + else if (!strcmp(child->name, "SegmentBase"))
> + ret = dash_parse_segmentbase(s, period, &timeline, child);
> + else if (!strcmp(child->name, "AdaptationSet"))
> + ret = dash_parse_adaptationset(s, period, urls, timeline, &par, child);
> + else if (!strcmp(child->name, "AssetIdentifier"))
> + ret = dash_parse_assetidentifier(s, period, child);
> + else
> + dash_log_unknown_child(s, child);
>
> - for (i = 0; i < c->n_audios; i++) {
> - rep = c->audios[i];
> - if (i > 0 && c->is_init_section_common_audio) {
> - ret = copy_init_section(rep, c->audios[0]);
> - if (ret < 0)
> - goto fail;
> + if (ret < 0)
> + goto out;
> + }
> +
> +out:
> + dash_unref_timeline(timeline);
> + dash_unref_urllist(urls);
> + return ret;
> +}
> +
> +static av_cold int dash_parse_programinformation(AVFormatContext *s, xmlNodePtr node)
> +{
> + xmlChar *lang = NULL;
> +
> + xml_for_each_attr {
> + if (!strcmp(attr->name, "lang")) {
> + lang = value;
> + value = NULL;
> + } else
> + dash_log_unknown_attr(s, attr, value);
> + }
> +
> + xml_for_each_child {
> + const char *key = NULL;
> + xmlChar *value;
> +
> + if (!(value = xmlNodeGetContent(child)))
> + continue;
> +
> + if (!strcmp(child->name, "Title"))
> + key = "title";
> + else if (!strcmp(child->name, "Source"))
> + key = "source";
> + else if (!strcmp(child->name, "Copyright"))
> + key = "copyright";
> + else
> + dash_log_unknown_child(s, child);
> +
> + if (key) {
> + char key_lang[100];
> + if (lang)
> + sprintf(key_lang, "%s-%s", key, lang);
> + av_dict_set(&s->metadata, lang ? key_lang : key, value, 0);
> }
> - ret = open_demux_for_component(s, rep);
>
> - if (ret)
> - goto fail;
> - rep->stream_index = stream_index;
> - ++stream_index;
> + xmlFree(value);
> }
>
> - if (c->n_subtitles)
> - c->is_init_section_common_subtitle = is_common_init_section_exist(c->subtitles, c->n_subtitles);
> + xml_free(lang);
> + return 0;
> +}
> +
> +static av_cold void dash_reset(AVFormatContext *s)
> +{
> + DASHContext *c = s->priv_data;
> +
> + /* Gets updated as we parse Periods. */
> + s->duration = AV_NOPTS_VALUE;
> + c->is_live = 0;
> + c->media_presentation_duration = INT64_MIN;
> +
> + av_freep_arrayp(&c->reps, dash_free_representation);
> + av_freep_arrayp(&c->ass, dash_free_adaptationset);
> + av_freep_arrayp(&c->periods, dash_free_period);
> +
> + av_freep(&c->location);
> +
> + xml_freep(&c->id);
> + xml_freep(&c->chain_next_location);
> + xml_freep(&c->fallback_location);
> +
> + av_dict_free(&c->protocol_opts);
> +}
> +
> +static av_cold int dash_parse_mpd(AVFormatContext *s,
> + DASHURLList *base_urls,
> + xmlNodePtr node)
> +{
> + int ret = 0;
> + DASHContext *c = s->priv_data;
> + DASHURLList *urls = dash_ref_urllist(base_urls);
>
> - for (i = 0; i < c->n_subtitles; i++) {
> - rep = c->subtitles[i];
> - if (i > 0 && c->is_init_section_common_subtitle) {
> - ret = copy_init_section(rep, c->subtitles[0]);
> - if (ret < 0)
> - goto fail;
> + /* Nullify periods.
> + * TODO: Maybe they sould be cleaned up after parsing. */
> + for (unsigned i = 0; i < c->nb_periods; ++i) {
> + DASHPeriod *period = c->periods[i];
> + period->start_ts = period->end_ts = 0;
> + }
> +
> + {
> + xmlChar *id = xmlGetNoNsProp(node, "id");
> + if (!c->id || !id || strcmp(c->id, id)) {
> + dash_reset(s);
> + c->id = id;
> + } else {
> + xmlFree(id);
> }
> - ret = open_demux_for_component(s, rep);
> + }
>
> - if (ret)
> - goto fail;
> - rep->stream_index = stream_index;
> - ++stream_index;
> + xml_for_each_attr {
> + if (!strcmp(attr->name, "type")) {
> + /* https://livesim.dashif.org/livesim/scte35_2/testpic_2s/Manifest.mpd */
> + if (!strcmp(value, "dynamic"))
> + c->is_live = 1;
> + else if (!strcmp(value, "static"))
> + c->is_live = 0;
> + else
> + dash_log_invalid_attr_value(s, attr, value);
> + } else if (!strcmp(attr->name, "availabilityStartTime"))
> + c->availability_start_time = dash_parse_date(value, s);
> + else if (!strcmp(attr->name, "availabilityEndTime"))
> + c->availability_end_time = dash_parse_date(value, s);
> + else if (!strcmp(attr->name, "publishTime")) {
> + c->publish_time = dash_parse_date(value, s);
> + /* Take the easy way. */
> + av_dict_set(&s->metadata, "creation_time", value, 0);
> + } else if (!strcmp(attr->name, "minimumUpdatePeriod"))
> + c->min_update_period = dash_parse_duration(value, s);
> + else if (!strcmp(attr->name, "maxSegmentDuration"))
> + c->max_segment_duration = dash_parse_duration(value, s);
> + else if (!strcmp(attr->name, "minBufferTime"))
> + c->min_buffer_time = dash_parse_duration(value, s);
> + else if (!strcmp(attr->name, "timeShiftBufferDepth"))
> + c->time_shift_buffer_depth = dash_parse_duration(value, s);
> + else if (!strcmp(attr->name, "mediaPresentationDuration"))
> + s->duration = c->media_presentation_duration = dash_parse_duration(value, s);
> + else if (strcmp(attr->name, "id") &&
> + strcmp(attr->name, "profiles") &&
> + strcmp(attr->name, "schemaLocation"))
> + dash_log_unknown_attr(s, attr, value);
> }
>
> - if (!stream_index) {
> - ret = AVERROR_INVALIDDATA;
> - goto fail;
> + xml_for_each_child {
> + if (!strcmp(child->name, "ProgramInformation"))
> + ret = dash_parse_programinformation(s, child);
> + else if (!strcmp(child->name, "Location"))
> + ret = dash_parse_location(s, child);
> + else if (!strcmp(child->name, "BaseURL"))
> + ret = dash_parse_baseurl(s, &urls, child);
> + else if (!strcmp(child->name, "Period"))
> + ret = dash_parse_period(s, urls, child);
> + else if (!strcmp(child->name, "UTCTiming"))
> + ret = dash_parse_utctiming(s, child);
> + else if (!strcmp(child->name, "SupplementalProperty") ||
> + !strcmp(child->name, "EssentialProperty"))
> + ret = dash_parse_property(s, NULL, NULL, child);
> + else
> + dash_log_unknown_child(s, child);
> +
> + if (ret < 0)
> + goto out;
> }
>
> - /* Create a program */
> - program = av_new_program(s, 0);
> - if (!program) {
> +out:
> + dash_unref_urllist(urls);
> + return ret;
> +}
> +
> +/**
> + * Remove filename component from an URL.
> + */
> +static int ff_make_absolute_head_urla(char **out_url, const char *url)
> +{
> + int ret;
> +
> + if (0 <= (ret = ff_make_absolute_urla(out_url, url, "./"))) {
> + char *p = *out_url;
> + size_t len = strlen(p);
> +
> + if (3 <= len && !memcmp(p + len - 3, "/./", 4))
> + p[len - 2] = '\0';
> + }
> +
> + return ret;
> +}
> +
> +static av_cold int dash_parse_root(AVFormatContext *s, xmlNodePtr node)
> +{
> + int ret = 0;
> + DASHURLList *urls = NULL;
> + DASHContext *c = s->priv_data;
> +
> + if (!node ||
> + strcmp(node->name, "MPD") ||
> + !node->ns ||
> + strcmp(node->ns->href, "urn:mpeg:dash:schema:mpd:2011"))
> + {
> + av_log(s, AV_LOG_ERROR, "Not an MPD\n");
> + return AVERROR_INVALIDDATA;
> + }
> +
> + if (!(urls = dash_new_urllist(1))) {
> ret = AVERROR(ENOMEM);
> - goto fail;
> - }
> -
> - for (i = 0; i < c->n_videos; i++) {
> - rep = c->videos[i];
> - av_program_add_stream_index(s, 0, rep->stream_index);
> - rep->assoc_stream = s->streams[rep->stream_index];
> - if (rep->bandwidth > 0)
> - av_dict_set_int(&rep->assoc_stream->metadata, "variant_bitrate", rep->bandwidth, 0);
> - move_metadata(rep->assoc_stream, "id", &rep->id);
> - }
> - for (i = 0; i < c->n_audios; i++) {
> - rep = c->audios[i];
> - av_program_add_stream_index(s, 0, rep->stream_index);
> - rep->assoc_stream = s->streams[rep->stream_index];
> - if (rep->bandwidth > 0)
> - av_dict_set_int(&rep->assoc_stream->metadata, "variant_bitrate", rep->bandwidth, 0);
> - move_metadata(rep->assoc_stream, "id", &rep->id);
> - move_metadata(rep->assoc_stream, "language", &rep->lang);
> - }
> - for (i = 0; i < c->n_subtitles; i++) {
> - rep = c->subtitles[i];
> - av_program_add_stream_index(s, 0, rep->stream_index);
> - rep->assoc_stream = s->streams[rep->stream_index];
> - move_metadata(rep->assoc_stream, "id", &rep->id);
> - move_metadata(rep->assoc_stream, "language", &rep->lang);
> + goto out;
> }
>
> - return 0;
> -fail:
> - dash_close(s);
> + if ((ret = ff_make_absolute_head_urla(&urls->elems[0], c->location)) < 0)
> + goto out;
> +
> + if ((ret = dash_parse_mpd(s, urls, node)) < 0)
> + goto out;
> +
> +out:
> + dash_unref_urllist(urls);
> return ret;
> }
>
> -static void recheck_discard_flags(AVFormatContext *s, struct representation **p, int n)
> +static av_cold void dash_libxml_error_handler(void *opaque, const char *fmt, ...)
> {
> - int i, j;
> -
> - for (i = 0; i < n; i++) {
> - struct representation *pls = p[i];
> - int needed = !pls->assoc_stream || pls->assoc_stream->discard < AVDISCARD_ALL;
> -
> - if (needed && !pls->ctx) {
> - pls->cur_seg_offset = 0;
> - pls->init_sec_buf_read_offset = 0;
> - /* Catch up */
> - for (j = 0; j < n; j++) {
> - pls->cur_seq_no = FFMAX(pls->cur_seq_no, p[j]->cur_seq_no);
> + AVFormatContext *s = opaque;
> + va_list ap;
> +
> + va_start(ap, fmt);
> + av_vlog(s, AV_LOG_ERROR, fmt, ap);
> + va_end(ap);
> +}
> +
> +static av_cold int dash_open_manifest(AVFormatContext *s)
> +{
> + int ret = 0;
> + DASHContext *c = s->priv_data;
> + xmlParserCtxtPtr xml;
> + AVIOContext *pb;
> + AVDictionary *opts = NULL;
> + uint8_t *location;
> +
> + av_dict_set(&opts, "icy", "0", 0);
> + ret = s->io_open(s, &pb, c->location, AVIO_FLAG_READ, &opts);
> + av_dict_free(&opts);
> + if (ret < 0)
> + return ret;
> +
> + xmlSetGenericErrorFunc(s, dash_libxml_error_handler);
> +
> + /* Why not SAX? DASH-IF may standardize patching that obviously requires us
> + * to maintain DOM. */
> + for (xml = NULL;;) {
> + char buf[1 << 14];
> +
> + /* libXML2 needs at least 4 bytes at the beginnig to being able to
> + * initialize its parsing context. */
> + if (0 < (ret = (xml ? avio_read_partial : avio_read)(pb, buf, (xml ? sizeof(buf) : sizeof("BOM"))))) {
> + /* Context is initialized by the first read call. */
> + if (!xml) {
> + if ((xml = xmlCreatePushParserCtxt(NULL, NULL, buf, ret, NULL)))
> + continue;
> + } else {
> + if (!xmlParseChunk(xml, buf, ret, 0 /* Terminate? */))
> + continue;
> +
> + parse_error:
> + av_log(s, AV_LOG_ERROR, "Failed to parse manifest file\n");
> + ret = AVERROR_INVALIDDATA;
> + goto out;
> }
> - reopen_demux_for_component(s, pls);
> - av_log(s, AV_LOG_INFO, "Now receiving stream_index %d\n", pls->stream_index);
> - } else if (!needed && pls->ctx) {
> - close_demux_for_component(pls);
> - ff_format_io_close(pls->parent, &pls->input);
> - av_log(s, AV_LOG_INFO, "No longer receiving stream_index %d\n", pls->stream_index);
> + } else if (ret == AVERROR_EOF) {
> + if (!xmlParseChunk(xml, NULL, 0, 1 /* Terminate? */))
> + break;
> +
> + goto parse_error;
> }
> +
> + av_log(s, AV_LOG_ERROR, "Failed to read manifest file\n");
> + goto out;
> }
> +
> + /* For HTTP 3XX redirects standard states that the redirected location
> + * should be used for updates. We are lazy and just check for "location". */
> + if (0 <= av_opt_get(pb, "location", AV_OPT_SEARCH_CHILDREN, &location)) {
> + (void)dash_set_location(s, location);
> + av_free(location);
> + }
> +
> + ret = dash_parse_root(s, xmlDocGetRootElement(xml->myDoc));
> +
> + if (c->is_live) {
> + int64_t accuracy = c->min_update_period / 20; /* 5% */
> + c->next_update = av_gettime() +
> + (av_lfg_get(&c->rnd) % accuracy + 1) +
> + (c->min_update_period - accuracy);
> + } else {
> + c->next_update = 0;
> + }
> +
> +out:
> + if (xml) {
> + if (xml->myDoc)
> + xmlFreeDoc(xml->myDoc);
> + xmlFreeParserCtxt(xml);
> + }
> +
> + /* Reset handlers so passed references to AVFormatContext will not survive. */
> + xmlSetGenericErrorFunc(NULL, NULL);
> +
> + ff_format_io_close(s, &pb);
> + return ret;
> +}
> +
> +static av_cold int dash_close(AVFormatContext *s)
> +{
> + return dash_reset(s), 0;
> +}
> +
> +static av_cold int dash_save_protocol_options(AVFormatContext *s)
> +{
> + static const char OPTIONS[] =
> + "headers\0" "http_proxy\0" "user_agent\0" "cookies\0" "referer\0"
> + "rw_timeout\0" "icy\0";
> +
> + DASHContext *c = s->priv_data;
> +
> + for (const char *opt = OPTIONS; *opt; opt += strlen(opt) + 1 /* NUL */) {
> + uint8_t *buf;
> +
> + if (0 <= av_opt_get(s->pb, opt, AV_OPT_SEARCH_CHILDREN, &buf)) {
> + int ret = av_dict_set(&c->protocol_opts, opt, buf,
> + AV_DICT_DONT_OVERWRITE | AV_DICT_DONT_STRDUP_VAL);
> + if (ret < 0)
> + return ret;
> + }
> + }
> +
> + return 0;
> +}
> +
> +static av_cold int dash_read_header(AVFormatContext *s)
> +{
> + int ret = 0;
> + DASHContext *c = s->priv_data;
> +
> + if ((ret = dash_set_location(s, s->url)) < 0)
> + goto out;
> +
> + /* Signal that we may create additional streams as time pasts. */
> + s->ctx_flags |= AVFMTCTX_NOHEADER;
> +
> + if ((ret = dash_save_protocol_options(s)) < 0)
> + goto out;
> +
> + av_lfg_init(&c->rnd, av_get_random_seed());
> +
> + if ((ret = dash_open_manifest(s)) < 0)
> + goto out;
> +
> +out:
> + if (ret < 0)
> + dash_close(s);
> + return ret;
> +}
> +
> +static DASHRepresentation *dash_get_read_next_representation(AVFormatContext *s)
> +{
> + DASHContext *c = s->priv_data;
> + DASHRepresentation *best_rep = NULL;
> +
> + for (unsigned i = 0; i < c->nb_reps; ++i) {
> + DASHRepresentation *rep = c->reps[i];
> + int needed = 0;
> +
> + /* Subdemuxer not opened for this representation, so we are not
> + * interested in receiving segments for this. */
> + if (!rep->ic)
> + continue;
> +
> + /* Check if any of the streams is needed. */
> + for (unsigned k = 0; k < rep->nb_streams; ++k) {
> + AVStream *ost = rep->ostreams[k];
> + if ((needed = ost->discard < AVDISCARD_ALL))
> + break;
> + }
> +
> + if (!needed)
> + continue;
> +
> + /* A toy scheduler. */
> + if (!best_rep || rep->last_pts < best_rep->last_pts)
> + best_rep = rep;
> + }
> +
> + return best_rep;
> }
>
> static int dash_read_packet(AVFormatContext *s, AVPacket *pkt)
> {
> + int ret;
> + DASHRepresentation *rep;
> + AVStream *ist, /* --packet--> */ *ost;
> DASHContext *c = s->priv_data;
> - int ret = 0, i;
> - int64_t mints = 0;
> - struct representation *cur = NULL;
> - struct representation *rep = NULL;
>
> - recheck_discard_flags(s, c->videos, c->n_videos);
> - recheck_discard_flags(s, c->audios, c->n_audios);
> - recheck_discard_flags(s, c->subtitles, c->n_subtitles);
> -
> - for (i = 0; i < c->n_videos; i++) {
> - rep = c->videos[i];
> - if (!rep->ctx)
> - continue;
> - if (!cur || rep->cur_timestamp < mints) {
> - cur = rep;
> - mints = rep->cur_timestamp;
> + if (c->next_update && c->next_update <= av_gettime())
> + if ((ret = dash_open_manifest(s)) < 0) {
> + c->next_update = av_gettime() * (60 * MICROSEC_PER_SEC) /* 1 min */;
> + av_log(s, AV_LOG_ERROR, "Failed to update manifest\n");
> }
> - }
> - for (i = 0; i < c->n_audios; i++) {
> - rep = c->audios[i];
> - if (!rep->ctx)
> - continue;
> - if (!cur || rep->cur_timestamp < mints) {
> - cur = rep;
> - mints = rep->cur_timestamp;
> +
> + /* Find a representation where we can read from. */
> + for (;;) {
> + if (!(rep = dash_get_read_next_representation(s))) {
> + /* Load next manifest, if any. */
> + if (c->chain_next_location) {
> + if ((ret = dash_set_location(s, c->chain_next_location)) < 0)
> + return ret;
> + xml_freep(&c->chain_next_location);
> +
> + if ((ret = dash_open_manifest(s)) < 0)
> + return ret;
> +
> + continue;
> + }
> +
> + return AVERROR_EOF;
> }
> - }
>
> - for (i = 0; i < c->n_subtitles; i++) {
> - rep = c->subtitles[i];
> - if (!rep->ctx)
> + if ((!rep->ic->iformat &&
> + (ret = dash_subdemuxer_open(rep)) < 0) ||
> + (ret = av_read_frame(rep->ic, pkt)) < 0)
> + {
> + av_log(s, AVERROR_EOF != ret ? AV_LOG_ERROR : AV_LOG_DEBUG,
> + "Failed to read representation '%s': %s\n",
> + rep->id, av_err2str(ret));
> + dash_subdemuxer_close(rep);
> continue;
> - if (!cur || rep->cur_timestamp < mints) {
> - cur = rep;
> - mints = rep->cur_timestamp;
> }
> - }
>
> - if (!cur) {
> - return AVERROR_INVALIDDATA;
> + break;
> }
> - while (!ff_check_interrupt(c->interrupt_callback) && !ret) {
> - ret = av_read_frame(cur->ctx, pkt);
> - if (ret >= 0) {
> - /* If we got a packet, return it */
> - cur->cur_timestamp = av_rescale(pkt->pts, (int64_t)cur->ctx->streams[0]->time_base.num * 90000, cur->ctx->streams[0]->time_base.den);
> - pkt->stream_index = cur->stream_index;
> - return 0;
> - }
> - if (cur->is_restart_needed) {
> - cur->cur_seg_offset = 0;
> - cur->init_sec_buf_read_offset = 0;
> - ff_format_io_close(cur->parent, &cur->input);
> - ret = reopen_demux_for_component(s, cur);
> - cur->is_restart_needed = 0;
> - }
> - }
> - return AVERROR_EOF;
> -}
>
> -static int dash_close(AVFormatContext *s)
> -{
> - DASHContext *c = s->priv_data;
> - free_audio_list(c);
> - free_video_list(c);
> - free_subtitle_list(c);
> - av_dict_free(&c->avio_opts);
> - av_freep(&c->base_url);
> + ist = rep->ic->streams[pkt->stream_index];
> +
> + if (/* New stream had been created by the underlying subdemuxer. */
> + rep->nb_streams <= pkt->stream_index ||
> + /* Something changed. */
> + (ist->event_flags & AVSTREAM_EVENT_FLAG_METADATA_UPDATED))
> + if ((ret = dash_subdemuxer_update(rep)) < 0)
> + return ret;
> +
> + /* Make packet timestamps comparable to each other. */
> + rep->last_pts = av_rescale_q(pkt->pts + pkt->duration, ist->time_base, AV_TIME_BASE_Q);
> +
> + ost = rep->ostreams[pkt->stream_index];
> + if (ost->codecpar->codec_type != ist->codecpar->codec_type ||
> + ost->codecpar->codec_id != ist->codecpar->codec_id ||
> + ost->codecpar->codec_tag != ist->codecpar->codec_tag)
> + if ((ret = dash_subdemuxer_update(rep)) < 0)
> + return ret;
> +
> + /* Translate stream_index from inner to outer context. */
> + pkt->stream_index = ost->index;
> +
> return 0;
> }
>
> -static int dash_seek(AVFormatContext *s, struct representation *pls, int64_t seek_pos_msec, int flags, int dry_run)
> -{
> - int ret = 0;
> - int i = 0;
> - int j = 0;
> - int64_t duration = 0;
> -
> - av_log(pls->parent, AV_LOG_VERBOSE, "DASH seek pos[%"PRId64"ms] %s\n",
> - seek_pos_msec, dry_run ? " (dry)" : "");
> -
> - // single fragment mode
> - if (pls->n_fragments == 1) {
> - pls->cur_timestamp = 0;
> - pls->cur_seg_offset = 0;
> - if (dry_run)
> - return 0;
> - ff_read_frame_flush(pls->ctx);
> - return av_seek_frame(pls->ctx, -1, seek_pos_msec * 1000, flags);
> - }
> -
> - ff_format_io_close(pls->parent, &pls->input);
> -
> - // find the nearest fragment
> - if (pls->n_timelines > 0 && pls->fragment_timescale > 0) {
> - int64_t num = pls->first_seq_no;
> - av_log(pls->parent, AV_LOG_VERBOSE, "dash_seek with SegmentTimeline start n_timelines[%d] "
> - "last_seq_no[%"PRId64"].\n",
> - (int)pls->n_timelines, (int64_t)pls->last_seq_no);
> - for (i = 0; i < pls->n_timelines; i++) {
> - if (pls->timelines[i]->starttime > 0) {
> - duration = pls->timelines[i]->starttime;
> - }
> - duration += pls->timelines[i]->duration;
> - if (seek_pos_msec < ((duration * 1000) / pls->fragment_timescale)) {
> - goto set_seq_num;
> - }
> - for (j = 0; j < pls->timelines[i]->repeat; j++) {
> - duration += pls->timelines[i]->duration;
> - num++;
> - if (seek_pos_msec < ((duration * 1000) / pls->fragment_timescale)) {
> - goto set_seq_num;
> - }
> - }
> - num++;
> - }
> -
> -set_seq_num:
> - pls->cur_seq_no = num > pls->last_seq_no ? pls->last_seq_no : num;
> - av_log(pls->parent, AV_LOG_VERBOSE, "dash_seek with SegmentTimeline end cur_seq_no[%"PRId64"].\n",
> - (int64_t)pls->cur_seq_no);
> - } else if (pls->fragment_duration > 0) {
> - pls->cur_seq_no = pls->first_seq_no + ((seek_pos_msec * pls->fragment_timescale) / pls->fragment_duration) / 1000;
> - } else {
> - av_log(pls->parent, AV_LOG_ERROR, "dash_seek missing timeline or fragment_duration\n");
> - pls->cur_seq_no = pls->first_seq_no;
> - }
> - pls->cur_timestamp = 0;
> - pls->cur_seg_offset = 0;
> - pls->init_sec_buf_read_offset = 0;
> - ret = dry_run ? 0 : reopen_demux_for_component(s, pls);
> -
> - return ret;
> -}
> -
> static int dash_read_seek(AVFormatContext *s, int stream_index, int64_t timestamp, int flags)
> {
> - int ret = 0, i;
> + int ret = 0;
> DASHContext *c = s->priv_data;
> - int64_t seek_pos_msec = av_rescale_rnd(timestamp, 1000,
> - s->streams[stream_index]->time_base.den,
> - flags & AVSEEK_FLAG_BACKWARD ?
> - AV_ROUND_DOWN : AV_ROUND_UP);
> - if ((flags & AVSEEK_FLAG_BYTE) || c->is_live)
> + AVStream *st;
> + int64_t now_ts;
> +
> + if (flags & AVSEEK_FLAG_BYTE)
> return AVERROR(ENOSYS);
>
> - /* Seek in discarded streams with dry_run=1 to avoid reopening them */
> - for (i = 0; i < c->n_videos; i++) {
> - if (!ret)
> - ret = dash_seek(s, c->videos[i], seek_pos_msec, flags, !c->videos[i]->ctx);
> - }
> - for (i = 0; i < c->n_audios; i++) {
> - if (!ret)
> - ret = dash_seek(s, c->audios[i], seek_pos_msec, flags, !c->audios[i]->ctx);
> - }
> - for (i = 0; i < c->n_subtitles; i++) {
> - if (!ret)
> - ret = dash_seek(s, c->subtitles[i], seek_pos_msec, flags, !c->subtitles[i]->ctx);
> + st = s->streams[stream_index];
> + now_ts = av_rescale_q_rnd(timestamp, st->time_base, AV_TIME_BASE_Q,
> + (flags & AVSEEK_FLAG_BACKWARD)
> + ? AV_ROUND_DOWN : AV_ROUND_UP);
> +
> + for (unsigned i = 0; i < c->nb_reps; ++i) {
> + DASHRepresentation *rep = c->reps[i];
> +
> + if (!rep->ic)
> + continue;
> +
> + rep->last_pts = AV_NOPTS_VALUE;
> +
> + rep->read_ts = c->start_ts + now_ts;
> +
> + if ((ret = av_seek_frame(rep->ic, -1, now_ts, flags)) < 0) {
> + av_log(s, AV_LOG_ERROR, "Failed to seek subdemuxer\n");
> + /* abort(); */
> + }
> +
> + dash_subdemuxer_flush(rep);
> }
>
> return ret;
> }
>
> -static int dash_probe(const AVProbeData *p)
> +static av_cold int dash_probe(const AVProbeData *p)
> {
> - if (!av_stristr(p->buf, "<MPD"))
> - return 0;
> + if (strstr(p->buf, "<?xml") &&
> + strstr(p->buf, "<MPD"))
> + return AVPROBE_SCORE_MAX;
>
> - if (av_stristr(p->buf, "dash:profile:isoff-on-demand:2011") ||
> - av_stristr(p->buf, "dash:profile:isoff-live:2011") ||
> - av_stristr(p->buf, "dash:profile:isoff-live:2012") ||
> - av_stristr(p->buf, "dash:profile:isoff-main:2011") ||
> - av_stristr(p->buf, "3GPP:PSS:profile:DASH1")) {
> + if (p->mime_type && !strncmp(p->mime_type, "application/dash+xml", 20))
> return AVPROBE_SCORE_MAX;
> - }
> - if (av_stristr(p->buf, "dash:profile")) {
> - return AVPROBE_SCORE_MAX;
> - }
>
> return 0;
> }
>
> #define OFFSET(x) offsetof(DASHContext, x)
> -#define FLAGS AV_OPT_FLAG_DECODING_PARAM
> static const AVOption dash_options[] = {
> - {"allowed_extensions", "List of file extensions that dash is allowed to access",
> - OFFSET(allowed_extensions), AV_OPT_TYPE_STRING,
> - {.str = "aac,m4a,m4s,m4v,mov,mp4,webm,ts"},
> - INT_MIN, INT_MAX, FLAGS},
> - {NULL}
> + { "connections", "Number of segment requests on the fly (per representation)",
> + OFFSET(nb_connections), AV_OPT_TYPE_INT,
> + { .i64 = 1 }, .min = 1, .max = UINT_MAX, .flags = AV_OPT_FLAG_DECODING_PARAM },
> + { "protocol_opts", "Specify protocol options for opened segments",
> + OFFSET(protocol_opts), AV_OPT_TYPE_DICT,
> + .flags = AV_OPT_FLAG_DECODING_PARAM },
> + { NULL }
> };
>
> static const AVClass dash_class = {
> @@ -2407,5 +3149,6 @@ AVInputFormat ff_dash_demuxer = {
> .read_packet = dash_read_packet,
> .read_close = dash_close,
> .read_seek = dash_read_seek,
> + .extensions = "mpd",
> .flags = AVFMT_NO_BYTE_SEEK,
> };
> diff --git a/libavformat/internal.h b/libavformat/internal.h
> index 3c6b292..ee547e0 100644
> --- a/libavformat/internal.h
> +++ b/libavformat/internal.h
> @@ -33,6 +33,8 @@
> #define PROBE_BUF_MIN 2048
> #define PROBE_BUF_MAX (1 << 20)
>
> +#define UUID_BUF_SIZE 36
> +
> #ifdef DEBUG
> # define hex_dump_debug(class, buf, size) av_hex_dump_log(class, AV_LOG_DEBUG, buf, size)
> #else
> @@ -379,6 +381,18 @@ do {\
> */
> int ff_mkdir_p(const char *path);
>
> +char *ff_uuid_to_hex(char *buff, const uint8_t *src, int lowercase);
> +
> +/**
> + * Parse an UUID string.
> + *
> + * @param data the parsed data is written to this pointer
> + * @param p the string to parse
> + * @return the number of bytes written that is always 16 or negative number on
> + * error
> + */
> +int ff_uuid_to_data(uint8_t *data, const char *p);
> +
> char *ff_data_to_hex(char *buf, const uint8_t *src, int size, int lowercase);
>
> /**
> diff --git a/libavformat/utils.c b/libavformat/utils.c
> index ee947c1..d6b8418 100644
> --- a/libavformat/utils.c
> +++ b/libavformat/utils.c
> @@ -4878,6 +4878,53 @@ int ff_mkdir_p(const char *path)
> return ret;
> }
>
> +char *ff_uuid_to_hex(char *buff, const uint8_t *src, int lowercase)
> +{
> + ff_data_to_hex(buff + 0, src + 0, 4, lowercase);
> + buff[8] = '-';
> + ff_data_to_hex(buff + 9, src + 4, 2, lowercase);
> + buff[13] = '-';
> + ff_data_to_hex(buff + 14, src + 6, 2, lowercase);
> + buff[18] = '-';
> + ff_data_to_hex(buff + 19, src + 8, 2, lowercase);
> + buff[23] = '-';
> + ff_data_to_hex(buff + 24, src + 10, 6, lowercase);
> +
> + return buff;
> +}
> +
> +int ff_uuid_to_data(uint8_t *data, const char *p)
> +{
> + uint8_t len;
> +
> + for (len = 0; len < 16; ) {
> + uint8_t h, l;
> +
> +#define PARSE_XDIGIT(ch, res) \
> + if ('0' <= ch && ch <= '9') \
> + res = ch - '0'; \
> + else if ('A' <= ch && ch <= 'F') \
> + res = ch - 'A' + 10; \
> + else if ('a' <= ch && ch <= 'f') \
> + res = ch - 'a' + 10; \
> + else \
> + return -1;
> +
> + PARSE_XDIGIT(p[0], h);
> + PARSE_XDIGIT(p[1], l);
> + p += 2;
> +
> +#undef PARSE_XDIGIT
> +
> + data[len++] = (h << 4) | l;
> +
> + if ((4 == len || 6 == len || 8 == len || 10 == len) && *p++ != '-')
> + return -1;
> + }
> +
> + return *p == '\0' ? 16 : -1;
> +}
> +
> char *ff_data_to_hex(char *buff, const uint8_t *src, int s, int lowercase)
> {
> int i;
> --
> 2.30.1
>
> _______________________________________________
> ffmpeg-devel mailing list
> ffmpeg-devel at ffmpeg.org
> https://ffmpeg.org/mailman/listinfo/ffmpeg-devel
>
> To unsubscribe, visit link above, or email
> ffmpeg-devel-request at ffmpeg.org with subject "unsubscribe".
More information about the ffmpeg-devel
mailing list