1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2024-11-26 19:01:44 +02:00
FFmpeg/libavformat/dashenc.c
Karthick Jeyapal 44ee858b6d avformat/dashenc: Remove muxer overhead from Bandwidth field in DASH manifest
Fixes bug id #7386
Muxer overhead calculations was intented for HLS playlist as Apple's mediastreamvalidator tests were failing.
But applying the same fix for DASH manifest proved counterproductive, as Bandwidth can be used for segment name templates.
2018-09-11 17:31:09 +05:30

1617 lines
60 KiB
C

/*
* MPEG-DASH ISO BMFF segmenter
* Copyright (c) 2014 Martin Storsjo
* Copyright (c) 2018 Akamai Technologies, Inc.
*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include "config.h"
#if HAVE_UNISTD_H
#include <unistd.h>
#endif
#include "libavutil/avassert.h"
#include "libavutil/avutil.h"
#include "libavutil/avstring.h"
#include "libavutil/intreadwrite.h"
#include "libavutil/mathematics.h"
#include "libavutil/opt.h"
#include "libavutil/rational.h"
#include "libavutil/time_internal.h"
#include "avc.h"
#include "avformat.h"
#include "avio_internal.h"
#include "hlsplaylist.h"
#if CONFIG_HTTP_PROTOCOL
#include "http.h"
#endif
#include "internal.h"
#include "isom.h"
#include "os_support.h"
#include "url.h"
#include "vpcc.h"
#include "dash.h"
typedef enum {
SEGMENT_TYPE_MP4 = 0,
SEGMENT_TYPE_WEBM,
SEGMENT_TYPE_NB
} SegmentType;
typedef struct Segment {
char file[1024];
int64_t start_pos;
int range_length, index_length;
int64_t time;
int duration;
int n;
} Segment;
typedef struct AdaptationSet {
char id[10];
enum AVMediaType media_type;
AVDictionary *metadata;
AVRational min_frame_rate, max_frame_rate;
int ambiguous_frame_rate;
} AdaptationSet;
typedef struct OutputStream {
AVFormatContext *ctx;
int ctx_inited, as_idx;
AVIOContext *out;
int packets_written;
char initfile[1024];
int64_t init_start_pos, pos;
int init_range_length;
int nb_segments, segments_size, segment_index;
Segment **segments;
int64_t first_pts, start_pts, max_pts;
int64_t last_dts, last_pts;
int bit_rate;
char codec_str[100];
int written_len;
char filename[1024];
char full_path[1024];
char temp_path[1024];
double availability_time_offset;
int total_pkt_size;
int muxer_overhead;
} OutputStream;
typedef struct DASHContext {
const AVClass *class; /* Class for private options. */
char *adaptation_sets;
AdaptationSet *as;
int nb_as;
int window_size;
int extra_window_size;
#if FF_API_DASH_MIN_SEG_DURATION
int min_seg_duration;
#endif
int64_t seg_duration;
int remove_at_exit;
int use_template;
int use_timeline;
int single_file;
OutputStream *streams;
int has_video;
int64_t last_duration;
int64_t total_duration;
char availability_start_time[100];
char dirname[1024];
const char *single_file_name;
const char *init_seg_name;
const char *media_seg_name;
const char *utc_timing_url;
const char *method;
const char *user_agent;
int hls_playlist;
int http_persistent;
int master_playlist_created;
AVIOContext *mpd_out;
AVIOContext *m3u8_out;
int streaming;
int64_t timeout;
int index_correction;
char *format_options_str;
SegmentType segment_type;
const char *format_name;
} DASHContext;
static struct codec_string {
int id;
const char *str;
} codecs[] = {
{ AV_CODEC_ID_VP8, "vp8" },
{ AV_CODEC_ID_VP9, "vp9" },
{ AV_CODEC_ID_VORBIS, "vorbis" },
{ AV_CODEC_ID_OPUS, "opus" },
{ 0, NULL }
};
static struct format_string {
SegmentType segment_type;
const char *str;
} formats[] = {
{ SEGMENT_TYPE_MP4, "mp4" },
{ SEGMENT_TYPE_WEBM, "webm" },
{ 0, NULL }
};
static int dashenc_io_open(AVFormatContext *s, AVIOContext **pb, char *filename,
AVDictionary **options) {
DASHContext *c = s->priv_data;
int http_base_proto = filename ? ff_is_http_proto(filename) : 0;
int err = AVERROR_MUXER_NOT_FOUND;
if (!*pb || !http_base_proto || !c->http_persistent) {
err = s->io_open(s, pb, filename, AVIO_FLAG_WRITE, options);
#if CONFIG_HTTP_PROTOCOL
} else {
URLContext *http_url_context = ffio_geturlcontext(*pb);
av_assert0(http_url_context);
err = ff_http_do_new_request(http_url_context, filename);
#endif
}
return err;
}
static void dashenc_io_close(AVFormatContext *s, AVIOContext **pb, char *filename) {
DASHContext *c = s->priv_data;
int http_base_proto = filename ? ff_is_http_proto(filename) : 0;
if (!http_base_proto || !c->http_persistent) {
ff_format_io_close(s, pb);
#if CONFIG_HTTP_PROTOCOL
} else {
URLContext *http_url_context = ffio_geturlcontext(*pb);
av_assert0(http_url_context);
avio_flush(*pb);
ffurl_shutdown(http_url_context, AVIO_FLAG_WRITE);
#endif
}
}
static const char *get_format_str(SegmentType segment_type) {
int i;
for (i = 0; i < SEGMENT_TYPE_NB; i++)
if (formats[i].segment_type == segment_type)
return formats[i].str;
return NULL;
}
static int check_file_extension(const char *filename, const char *extension) {
char *dot;
if (!filename || !extension)
return -1;
dot = strrchr(filename, '.');
if (dot && !strcmp(dot + 1, extension))
return 0;
return -1;
}
static void set_vp9_codec_str(AVFormatContext *s, AVCodecParameters *par,
AVRational *frame_rate, char *str, int size) {
VPCC vpcc;
int ret = ff_isom_get_vpcc_features(s, par, frame_rate, &vpcc);
if (ret == 0) {
av_strlcatf(str, size, "vp09.%02x.%02x.%02x",
vpcc.profile, vpcc.level, vpcc.bitdepth);
} else {
// Default to just vp9 in case of error while finding out profile or level
av_log(s, AV_LOG_WARNING, "Could not find VP9 profile and/or level\n");
av_strlcpy(str, "vp9", size);
}
return;
}
static void set_codec_str(AVFormatContext *s, AVCodecParameters *par,
AVRational *frame_rate, char *str, int size)
{
const AVCodecTag *tags[2] = { NULL, NULL };
uint32_t tag;
int i;
// common Webm codecs are not part of RFC 6381
for (i = 0; codecs[i].id; i++)
if (codecs[i].id == par->codec_id) {
if (codecs[i].id == AV_CODEC_ID_VP9) {
set_vp9_codec_str(s, par, frame_rate, str, size);
} else {
av_strlcpy(str, codecs[i].str, size);
}
return;
}
// for codecs part of RFC 6381
if (par->codec_type == AVMEDIA_TYPE_VIDEO)
tags[0] = ff_codec_movvideo_tags;
else if (par->codec_type == AVMEDIA_TYPE_AUDIO)
tags[0] = ff_codec_movaudio_tags;
else
return;
tag = av_codec_get_tag(tags, par->codec_id);
if (!tag)
return;
if (size < 5)
return;
AV_WL32(str, tag);
str[4] = '\0';
if (!strcmp(str, "mp4a") || !strcmp(str, "mp4v")) {
uint32_t oti;
tags[0] = ff_mp4_obj_type;
oti = av_codec_get_tag(tags, par->codec_id);
if (oti)
av_strlcatf(str, size, ".%02"PRIx32, oti);
else
return;
if (tag == MKTAG('m', 'p', '4', 'a')) {
if (par->extradata_size >= 2) {
int aot = par->extradata[0] >> 3;
if (aot == 31)
aot = ((AV_RB16(par->extradata) >> 5) & 0x3f) + 32;
av_strlcatf(str, size, ".%d", aot);
}
} else if (tag == MKTAG('m', 'p', '4', 'v')) {
// Unimplemented, should output ProfileLevelIndication as a decimal number
av_log(s, AV_LOG_WARNING, "Incomplete RFC 6381 codec string for mp4v\n");
}
} else if (!strcmp(str, "avc1")) {
uint8_t *tmpbuf = NULL;
uint8_t *extradata = par->extradata;
int extradata_size = par->extradata_size;
if (!extradata_size)
return;
if (extradata[0] != 1) {
AVIOContext *pb;
if (avio_open_dyn_buf(&pb) < 0)
return;
if (ff_isom_write_avcc(pb, extradata, extradata_size) < 0) {
ffio_free_dyn_buf(&pb);
return;
}
extradata_size = avio_close_dyn_buf(pb, &extradata);
tmpbuf = extradata;
}
if (extradata_size >= 4)
av_strlcatf(str, size, ".%02x%02x%02x",
extradata[1], extradata[2], extradata[3]);
av_free(tmpbuf);
}
}
static int flush_dynbuf(OutputStream *os, int *range_length)
{
uint8_t *buffer;
if (!os->ctx->pb) {
return AVERROR(EINVAL);
}
// flush
av_write_frame(os->ctx, NULL);
avio_flush(os->ctx->pb);
// write out to file
*range_length = avio_close_dyn_buf(os->ctx->pb, &buffer);
os->ctx->pb = NULL;
avio_write(os->out, buffer + os->written_len, *range_length - os->written_len);
os->written_len = 0;
av_free(buffer);
// re-open buffer
return avio_open_dyn_buf(&os->ctx->pb);
}
static void set_http_options(AVDictionary **options, DASHContext *c)
{
if (c->method)
av_dict_set(options, "method", c->method, 0);
if (c->user_agent)
av_dict_set(options, "user_agent", c->user_agent, 0);
if (c->http_persistent)
av_dict_set_int(options, "multiple_requests", 1, 0);
if (c->timeout >= 0)
av_dict_set_int(options, "timeout", c->timeout, 0);
}
static void get_hls_playlist_name(char *playlist_name, int string_size,
const char *base_url, int id) {
if (base_url)
snprintf(playlist_name, string_size, "%smedia_%d.m3u8", base_url, id);
else
snprintf(playlist_name, string_size, "media_%d.m3u8", id);
}
static int flush_init_segment(AVFormatContext *s, OutputStream *os)
{
DASHContext *c = s->priv_data;
int ret, range_length;
ret = flush_dynbuf(os, &range_length);
if (ret < 0)
return ret;
os->pos = os->init_range_length = range_length;
if (!c->single_file)
ff_format_io_close(s, &os->out);
return 0;
}
static void dash_free(AVFormatContext *s)
{
DASHContext *c = s->priv_data;
int i, j;
if (c->as) {
for (i = 0; i < c->nb_as; i++)
av_dict_free(&c->as[i].metadata);
av_freep(&c->as);
c->nb_as = 0;
}
if (!c->streams)
return;
for (i = 0; i < s->nb_streams; i++) {
OutputStream *os = &c->streams[i];
if (os->ctx && os->ctx_inited)
av_write_trailer(os->ctx);
if (os->ctx && os->ctx->pb)
ffio_free_dyn_buf(&os->ctx->pb);
ff_format_io_close(s, &os->out);
if (os->ctx)
avformat_free_context(os->ctx);
for (j = 0; j < os->nb_segments; j++)
av_free(os->segments[j]);
av_free(os->segments);
}
av_freep(&c->streams);
ff_format_io_close(s, &c->mpd_out);
ff_format_io_close(s, &c->m3u8_out);
}
static void output_segment_list(OutputStream *os, AVIOContext *out, AVFormatContext *s,
int representation_id, int final)
{
DASHContext *c = s->priv_data;
int i, start_index = 0, start_number = 1;
if (c->window_size) {
start_index = FFMAX(os->nb_segments - c->window_size, 0);
start_number = FFMAX(os->segment_index - c->window_size, 1);
}
if (c->use_template) {
int timescale = c->use_timeline ? os->ctx->streams[0]->time_base.den : AV_TIME_BASE;
avio_printf(out, "\t\t\t\t<SegmentTemplate timescale=\"%d\" ", timescale);
if (!c->use_timeline) {
avio_printf(out, "duration=\"%"PRId64"\" ", c->seg_duration);
if (c->streaming && os->availability_time_offset)
avio_printf(out, "availabilityTimeOffset=\"%.3f\" ",
os->availability_time_offset);
}
avio_printf(out, "initialization=\"%s\" media=\"%s\" startNumber=\"%d\">\n", c->init_seg_name, c->media_seg_name, c->use_timeline ? start_number : 1);
if (c->use_timeline) {
int64_t cur_time = 0;
avio_printf(out, "\t\t\t\t\t<SegmentTimeline>\n");
for (i = start_index; i < os->nb_segments; ) {
Segment *seg = os->segments[i];
int repeat = 0;
avio_printf(out, "\t\t\t\t\t\t<S ");
if (i == start_index || seg->time != cur_time) {
cur_time = seg->time;
avio_printf(out, "t=\"%"PRId64"\" ", seg->time);
}
avio_printf(out, "d=\"%d\" ", seg->duration);
while (i + repeat + 1 < os->nb_segments &&
os->segments[i + repeat + 1]->duration == seg->duration &&
os->segments[i + repeat + 1]->time == os->segments[i + repeat]->time + os->segments[i + repeat]->duration)
repeat++;
if (repeat > 0)
avio_printf(out, "r=\"%d\" ", repeat);
avio_printf(out, "/>\n");
i += 1 + repeat;
cur_time += (1 + repeat) * seg->duration;
}
avio_printf(out, "\t\t\t\t\t</SegmentTimeline>\n");
}
avio_printf(out, "\t\t\t\t</SegmentTemplate>\n");
} else if (c->single_file) {
avio_printf(out, "\t\t\t\t<BaseURL>%s</BaseURL>\n", os->initfile);
avio_printf(out, "\t\t\t\t<SegmentList timescale=\"%d\" duration=\"%"PRId64"\" startNumber=\"%d\">\n", AV_TIME_BASE, c->last_duration, start_number);
avio_printf(out, "\t\t\t\t\t<Initialization range=\"%"PRId64"-%"PRId64"\" />\n", os->init_start_pos, os->init_start_pos + os->init_range_length - 1);
for (i = start_index; i < os->nb_segments; i++) {
Segment *seg = os->segments[i];
avio_printf(out, "\t\t\t\t\t<SegmentURL mediaRange=\"%"PRId64"-%"PRId64"\" ", seg->start_pos, seg->start_pos + seg->range_length - 1);
if (seg->index_length)
avio_printf(out, "indexRange=\"%"PRId64"-%"PRId64"\" ", seg->start_pos, seg->start_pos + seg->index_length - 1);
avio_printf(out, "/>\n");
}
avio_printf(out, "\t\t\t\t</SegmentList>\n");
} else {
avio_printf(out, "\t\t\t\t<SegmentList timescale=\"%d\" duration=\"%"PRId64"\" startNumber=\"%d\">\n", AV_TIME_BASE, c->last_duration, start_number);
avio_printf(out, "\t\t\t\t\t<Initialization sourceURL=\"%s\" />\n", os->initfile);
for (i = start_index; i < os->nb_segments; i++) {
Segment *seg = os->segments[i];
avio_printf(out, "\t\t\t\t\t<SegmentURL media=\"%s\" />\n", seg->file);
}
avio_printf(out, "\t\t\t\t</SegmentList>\n");
}
if (c->hls_playlist && start_index < os->nb_segments)
{
int timescale = os->ctx->streams[0]->time_base.den;
char temp_filename_hls[1024];
char filename_hls[1024];
AVDictionary *http_opts = NULL;
int target_duration = 0;
int ret = 0;
const char *proto = avio_find_protocol_name(c->dirname);
int use_rename = proto && !strcmp(proto, "file");
get_hls_playlist_name(filename_hls, sizeof(filename_hls),
c->dirname, representation_id);
snprintf(temp_filename_hls, sizeof(temp_filename_hls), use_rename ? "%s.tmp" : "%s", filename_hls);
set_http_options(&http_opts, c);
dashenc_io_open(s, &c->m3u8_out, temp_filename_hls, &http_opts);
av_dict_free(&http_opts);
for (i = start_index; i < os->nb_segments; i++) {
Segment *seg = os->segments[i];
double duration = (double) seg->duration / timescale;
if (target_duration <= duration)
target_duration = lrint(duration);
}
ff_hls_write_playlist_header(c->m3u8_out, 6, -1, target_duration,
start_number, PLAYLIST_TYPE_NONE);
ff_hls_write_init_file(c->m3u8_out, os->initfile, c->single_file,
os->init_range_length, os->init_start_pos);
for (i = start_index; i < os->nb_segments; i++) {
Segment *seg = os->segments[i];
ret = ff_hls_write_file_entry(c->m3u8_out, 0, c->single_file,
(double) seg->duration / timescale, 0,
seg->range_length, seg->start_pos, NULL,
c->single_file ? os->initfile : seg->file,
NULL);
if (ret < 0) {
av_log(os->ctx, AV_LOG_WARNING, "ff_hls_write_file_entry get error\n");
}
}
if (final)
ff_hls_write_end_list(c->m3u8_out);
dashenc_io_close(s, &c->m3u8_out, temp_filename_hls);
if (use_rename)
if (avpriv_io_move(temp_filename_hls, filename_hls) < 0) {
av_log(os->ctx, AV_LOG_WARNING, "renaming file %s to %s failed\n\n", temp_filename_hls, filename_hls);
}
}
}
static char *xmlescape(const char *str) {
int outlen = strlen(str)*3/2 + 6;
char *out = av_realloc(NULL, outlen + 1);
int pos = 0;
if (!out)
return NULL;
for (; *str; str++) {
if (pos + 6 > outlen) {
char *tmp;
outlen = 2 * outlen + 6;
tmp = av_realloc(out, outlen + 1);
if (!tmp) {
av_free(out);
return NULL;
}
out = tmp;
}
if (*str == '&') {
memcpy(&out[pos], "&amp;", 5);
pos += 5;
} else if (*str == '<') {
memcpy(&out[pos], "&lt;", 4);
pos += 4;
} else if (*str == '>') {
memcpy(&out[pos], "&gt;", 4);
pos += 4;
} else if (*str == '\'') {
memcpy(&out[pos], "&apos;", 6);
pos += 6;
} else if (*str == '\"') {
memcpy(&out[pos], "&quot;", 6);
pos += 6;
} else {
out[pos++] = *str;
}
}
out[pos] = '\0';
return out;
}
static void write_time(AVIOContext *out, int64_t time)
{
int seconds = time / AV_TIME_BASE;
int fractions = time % AV_TIME_BASE;
int minutes = seconds / 60;
int hours = minutes / 60;
seconds %= 60;
minutes %= 60;
avio_printf(out, "PT");
if (hours)
avio_printf(out, "%dH", hours);
if (hours || minutes)
avio_printf(out, "%dM", minutes);
avio_printf(out, "%d.%dS", seconds, fractions / (AV_TIME_BASE / 10));
}
static void format_date_now(char *buf, int size)
{
time_t t = time(NULL);
struct tm *ptm, tmbuf;
ptm = gmtime_r(&t, &tmbuf);
if (ptm) {
if (!strftime(buf, size, "%Y-%m-%dT%H:%M:%SZ", ptm))
buf[0] = '\0';
}
}
static int write_adaptation_set(AVFormatContext *s, AVIOContext *out, int as_index,
int final)
{
DASHContext *c = s->priv_data;
AdaptationSet *as = &c->as[as_index];
AVDictionaryEntry *lang, *role;
int i;
avio_printf(out, "\t\t<AdaptationSet id=\"%s\" contentType=\"%s\" segmentAlignment=\"true\" bitstreamSwitching=\"true\"",
as->id, as->media_type == AVMEDIA_TYPE_VIDEO ? "video" : "audio");
if (as->media_type == AVMEDIA_TYPE_VIDEO && as->max_frame_rate.num && !as->ambiguous_frame_rate && av_cmp_q(as->min_frame_rate, as->max_frame_rate) < 0)
avio_printf(out, " maxFrameRate=\"%d/%d\"", as->max_frame_rate.num, as->max_frame_rate.den);
lang = av_dict_get(as->metadata, "language", NULL, 0);
if (lang)
avio_printf(out, " lang=\"%s\"", lang->value);
avio_printf(out, ">\n");
role = av_dict_get(as->metadata, "role", NULL, 0);
if (role)
avio_printf(out, "\t\t\t<Role schemeIdUri=\"urn:mpeg:dash:role:2011\" value=\"%s\"/>\n", role->value);
for (i = 0; i < s->nb_streams; i++) {
OutputStream *os = &c->streams[i];
char bandwidth_str[64] = {'\0'};
if (os->as_idx - 1 != as_index)
continue;
if (os->bit_rate > 0)
snprintf(bandwidth_str, sizeof(bandwidth_str), " bandwidth=\"%d\"",
os->bit_rate);
if (as->media_type == AVMEDIA_TYPE_VIDEO) {
AVStream *st = s->streams[i];
avio_printf(out, "\t\t\t<Representation id=\"%d\" mimeType=\"video/%s\" codecs=\"%s\"%s width=\"%d\" height=\"%d\"",
i, c->format_name, os->codec_str, bandwidth_str, s->streams[i]->codecpar->width, s->streams[i]->codecpar->height);
if (st->avg_frame_rate.num)
avio_printf(out, " frameRate=\"%d/%d\"", st->avg_frame_rate.num, st->avg_frame_rate.den);
avio_printf(out, ">\n");
} else {
avio_printf(out, "\t\t\t<Representation id=\"%d\" mimeType=\"audio/%s\" codecs=\"%s\"%s audioSamplingRate=\"%d\">\n",
i, c->format_name, os->codec_str, bandwidth_str, s->streams[i]->codecpar->sample_rate);
avio_printf(out, "\t\t\t\t<AudioChannelConfiguration schemeIdUri=\"urn:mpeg:dash:23003:3:audio_channel_configuration:2011\" value=\"%d\" />\n",
s->streams[i]->codecpar->channels);
}
output_segment_list(os, out, s, i, final);
avio_printf(out, "\t\t\t</Representation>\n");
}
avio_printf(out, "\t\t</AdaptationSet>\n");
return 0;
}
static int add_adaptation_set(AVFormatContext *s, AdaptationSet **as, enum AVMediaType type)
{
DASHContext *c = s->priv_data;
void *mem = av_realloc(c->as, sizeof(*c->as) * (c->nb_as + 1));
if (!mem)
return AVERROR(ENOMEM);
c->as = mem;
++c->nb_as;
*as = &c->as[c->nb_as - 1];
memset(*as, 0, sizeof(**as));
(*as)->media_type = type;
return 0;
}
static int adaptation_set_add_stream(AVFormatContext *s, int as_idx, int i)
{
DASHContext *c = s->priv_data;
AdaptationSet *as = &c->as[as_idx - 1];
OutputStream *os = &c->streams[i];
if (as->media_type != s->streams[i]->codecpar->codec_type) {
av_log(s, AV_LOG_ERROR, "Codec type of stream %d doesn't match AdaptationSet's media type\n", i);
return AVERROR(EINVAL);
} else if (os->as_idx) {
av_log(s, AV_LOG_ERROR, "Stream %d is already assigned to an AdaptationSet\n", i);
return AVERROR(EINVAL);
}
os->as_idx = as_idx;
return 0;
}
static int parse_adaptation_sets(AVFormatContext *s)
{
DASHContext *c = s->priv_data;
const char *p = c->adaptation_sets;
enum { new_set, parse_id, parsing_streams } state;
AdaptationSet *as;
int i, n, ret;
// default: one AdaptationSet for each stream
if (!p) {
for (i = 0; i < s->nb_streams; i++) {
if ((ret = add_adaptation_set(s, &as, s->streams[i]->codecpar->codec_type)) < 0)
return ret;
snprintf(as->id, sizeof(as->id), "%d", i);
c->streams[i].as_idx = c->nb_as;
}
goto end;
}
// syntax id=0,streams=0,1,2 id=1,streams=3,4 and so on
state = new_set;
while (*p) {
if (*p == ' ') {
p++;
continue;
} else if (state == new_set && av_strstart(p, "id=", &p)) {
if ((ret = add_adaptation_set(s, &as, AVMEDIA_TYPE_UNKNOWN)) < 0)
return ret;
n = strcspn(p, ",");
snprintf(as->id, sizeof(as->id), "%.*s", n, p);
p += n;
if (*p)
p++;
state = parse_id;
} else if (state == parse_id && av_strstart(p, "streams=", &p)) {
state = parsing_streams;
} else if (state == parsing_streams) {
AdaptationSet *as = &c->as[c->nb_as - 1];
char idx_str[8], *end_str;
n = strcspn(p, " ,");
snprintf(idx_str, sizeof(idx_str), "%.*s", n, p);
p += n;
// if value is "a" or "v", map all streams of that type
if (as->media_type == AVMEDIA_TYPE_UNKNOWN && (idx_str[0] == 'v' || idx_str[0] == 'a')) {
enum AVMediaType type = (idx_str[0] == 'v') ? AVMEDIA_TYPE_VIDEO : AVMEDIA_TYPE_AUDIO;
av_log(s, AV_LOG_DEBUG, "Map all streams of type %s\n", idx_str);
for (i = 0; i < s->nb_streams; i++) {
if (s->streams[i]->codecpar->codec_type != type)
continue;
as->media_type = s->streams[i]->codecpar->codec_type;
if ((ret = adaptation_set_add_stream(s, c->nb_as, i)) < 0)
return ret;
}
} else { // select single stream
i = strtol(idx_str, &end_str, 10);
if (idx_str == end_str || i < 0 || i >= s->nb_streams) {
av_log(s, AV_LOG_ERROR, "Selected stream \"%s\" not found!\n", idx_str);
return AVERROR(EINVAL);
}
av_log(s, AV_LOG_DEBUG, "Map stream %d\n", i);
if (as->media_type == AVMEDIA_TYPE_UNKNOWN) {
as->media_type = s->streams[i]->codecpar->codec_type;
}
if ((ret = adaptation_set_add_stream(s, c->nb_as, i)) < 0)
return ret;
}
if (*p == ' ')
state = new_set;
if (*p)
p++;
} else {
return AVERROR(EINVAL);
}
}
end:
// check for unassigned streams
for (i = 0; i < s->nb_streams; i++) {
OutputStream *os = &c->streams[i];
if (!os->as_idx) {
av_log(s, AV_LOG_ERROR, "Stream %d is not mapped to an AdaptationSet\n", i);
return AVERROR(EINVAL);
}
}
return 0;
}
static int write_manifest(AVFormatContext *s, int final)
{
DASHContext *c = s->priv_data;
AVIOContext *out;
char temp_filename[1024];
int ret, i;
const char *proto = avio_find_protocol_name(s->url);
int use_rename = proto && !strcmp(proto, "file");
static unsigned int warned_non_file = 0;
AVDictionaryEntry *title = av_dict_get(s->metadata, "title", NULL, 0);
AVDictionary *opts = NULL;
if (!use_rename && !warned_non_file++)
av_log(s, AV_LOG_ERROR, "Cannot use rename on non file protocol, this may lead to races and temporary partial files\n");
snprintf(temp_filename, sizeof(temp_filename), use_rename ? "%s.tmp" : "%s", s->url);
set_http_options(&opts, c);
ret = dashenc_io_open(s, &c->mpd_out, temp_filename, &opts);
if (ret < 0) {
av_log(s, AV_LOG_ERROR, "Unable to open %s for writing\n", temp_filename);
return ret;
}
out = c->mpd_out;
av_dict_free(&opts);
avio_printf(out, "<?xml version=\"1.0\" encoding=\"utf-8\"?>\n");
avio_printf(out, "<MPD xmlns:xsi=\"http://www.w3.org/2001/XMLSchema-instance\"\n"
"\txmlns=\"urn:mpeg:dash:schema:mpd:2011\"\n"
"\txmlns:xlink=\"http://www.w3.org/1999/xlink\"\n"
"\txsi:schemaLocation=\"urn:mpeg:DASH:schema:MPD:2011 http://standards.iso.org/ittf/PubliclyAvailableStandards/MPEG-DASH_schema_files/DASH-MPD.xsd\"\n"
"\tprofiles=\"urn:mpeg:dash:profile:isoff-live:2011\"\n"
"\ttype=\"%s\"\n", final ? "static" : "dynamic");
if (final) {
avio_printf(out, "\tmediaPresentationDuration=\"");
write_time(out, c->total_duration);
avio_printf(out, "\"\n");
} else {
int64_t update_period = c->last_duration / AV_TIME_BASE;
char now_str[100];
if (c->use_template && !c->use_timeline)
update_period = 500;
avio_printf(out, "\tminimumUpdatePeriod=\"PT%"PRId64"S\"\n", update_period);
avio_printf(out, "\tsuggestedPresentationDelay=\"PT%"PRId64"S\"\n", c->last_duration / AV_TIME_BASE);
if (c->availability_start_time[0])
avio_printf(out, "\tavailabilityStartTime=\"%s\"\n", c->availability_start_time);
format_date_now(now_str, sizeof(now_str));
if (now_str[0])
avio_printf(out, "\tpublishTime=\"%s\"\n", now_str);
if (c->window_size && c->use_template) {
avio_printf(out, "\ttimeShiftBufferDepth=\"");
write_time(out, c->last_duration * c->window_size);
avio_printf(out, "\"\n");
}
}
avio_printf(out, "\tminBufferTime=\"");
write_time(out, c->last_duration * 2);
avio_printf(out, "\">\n");
avio_printf(out, "\t<ProgramInformation>\n");
if (title) {
char *escaped = xmlescape(title->value);
avio_printf(out, "\t\t<Title>%s</Title>\n", escaped);
av_free(escaped);
}
avio_printf(out, "\t</ProgramInformation>\n");
if (c->window_size && s->nb_streams > 0 && c->streams[0].nb_segments > 0 && !c->use_template) {
OutputStream *os = &c->streams[0];
int start_index = FFMAX(os->nb_segments - c->window_size, 0);
int64_t start_time = av_rescale_q(os->segments[start_index]->time, s->streams[0]->time_base, AV_TIME_BASE_Q);
avio_printf(out, "\t<Period id=\"0\" start=\"");
write_time(out, start_time);
avio_printf(out, "\">\n");
} else {
avio_printf(out, "\t<Period id=\"0\" start=\"PT0.0S\">\n");
}
for (i = 0; i < c->nb_as; i++) {
if ((ret = write_adaptation_set(s, out, i, final)) < 0)
return ret;
}
avio_printf(out, "\t</Period>\n");
if (c->utc_timing_url)
avio_printf(out, "\t<UTCTiming schemeIdUri=\"urn:mpeg:dash:utc:http-xsdate:2014\" value=\"%s\"/>\n", c->utc_timing_url);
avio_printf(out, "</MPD>\n");
avio_flush(out);
dashenc_io_close(s, &c->mpd_out, temp_filename);
if (use_rename) {
if ((ret = avpriv_io_move(temp_filename, s->url)) < 0)
return ret;
}
if (c->hls_playlist && !c->master_playlist_created) {
char filename_hls[1024];
const char *audio_group = "A1";
char audio_codec_str[128] = "\0";
int is_default = 1;
int max_audio_bitrate = 0;
if (*c->dirname)
snprintf(filename_hls, sizeof(filename_hls), "%smaster.m3u8", c->dirname);
else
snprintf(filename_hls, sizeof(filename_hls), "master.m3u8");
snprintf(temp_filename, sizeof(temp_filename), use_rename ? "%s.tmp" : "%s", filename_hls);
set_http_options(&opts, c);
ret = avio_open2(&out, temp_filename, AVIO_FLAG_WRITE, NULL, &opts);
if (ret < 0) {
av_log(s, AV_LOG_ERROR, "Unable to open %s for writing\n", temp_filename);
return ret;
}
av_dict_free(&opts);
ff_hls_write_playlist_version(out, 7);
for (i = 0; i < s->nb_streams; i++) {
char playlist_file[64];
AVStream *st = s->streams[i];
OutputStream *os = &c->streams[i];
if (st->codecpar->codec_type != AVMEDIA_TYPE_AUDIO)
continue;
get_hls_playlist_name(playlist_file, sizeof(playlist_file), NULL, i);
ff_hls_write_audio_rendition(out, (char *)audio_group,
playlist_file, i, is_default);
max_audio_bitrate = FFMAX(st->codecpar->bit_rate +
os->muxer_overhead, max_audio_bitrate);
if (!av_strnstr(audio_codec_str, os->codec_str, sizeof(audio_codec_str))) {
if (strlen(audio_codec_str))
av_strlcat(audio_codec_str, ",", sizeof(audio_codec_str));
av_strlcat(audio_codec_str, os->codec_str, sizeof(audio_codec_str));
}
is_default = 0;
}
for (i = 0; i < s->nb_streams; i++) {
char playlist_file[64];
char codec_str[128];
AVStream *st = s->streams[i];
OutputStream *os = &c->streams[i];
char *agroup = NULL;
int stream_bitrate = st->codecpar->bit_rate + os->muxer_overhead;
av_strlcpy(codec_str, os->codec_str, sizeof(codec_str));
if ((st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO) && max_audio_bitrate) {
agroup = (char *)audio_group;
stream_bitrate += max_audio_bitrate;
av_strlcat(codec_str, ",", sizeof(codec_str));
av_strlcat(codec_str, audio_codec_str, sizeof(codec_str));
}
get_hls_playlist_name(playlist_file, sizeof(playlist_file), NULL, i);
ff_hls_write_stream_info(st, out, stream_bitrate, playlist_file, agroup,
codec_str, NULL);
}
avio_close(out);
if (use_rename)
if ((ret = avpriv_io_move(temp_filename, filename_hls)) < 0)
return ret;
c->master_playlist_created = 1;
}
return 0;
}
static int dict_copy_entry(AVDictionary **dst, const AVDictionary *src, const char *key)
{
AVDictionaryEntry *entry = av_dict_get(src, key, NULL, 0);
if (entry)
av_dict_set(dst, key, entry->value, AV_DICT_DONT_OVERWRITE);
return 0;
}
static int dash_init(AVFormatContext *s)
{
DASHContext *c = s->priv_data;
int ret = 0, i;
char *ptr;
char basename[1024];
if (c->single_file_name)
c->single_file = 1;
if (c->single_file)
c->use_template = 0;
#if FF_API_DASH_MIN_SEG_DURATION
if (c->min_seg_duration != 5000000) {
av_log(s, AV_LOG_WARNING, "The min_seg_duration option is deprecated and will be removed. Please use the -seg_duration\n");
c->seg_duration = c->min_seg_duration;
}
#endif
av_strlcpy(c->dirname, s->url, sizeof(c->dirname));
ptr = strrchr(c->dirname, '/');
if (ptr) {
av_strlcpy(basename, &ptr[1], sizeof(basename));
ptr[1] = '\0';
} else {
c->dirname[0] = '\0';
av_strlcpy(basename, s->url, sizeof(basename));
}
ptr = strrchr(basename, '.');
if (ptr)
*ptr = '\0';
c->streams = av_mallocz(sizeof(*c->streams) * s->nb_streams);
if (!c->streams)
return AVERROR(ENOMEM);
if ((ret = parse_adaptation_sets(s)) < 0)
return ret;
for (i = 0; i < s->nb_streams; i++) {
OutputStream *os = &c->streams[i];
AdaptationSet *as = &c->as[os->as_idx - 1];
AVFormatContext *ctx;
AVStream *st;
AVDictionary *opts = NULL;
char filename[1024];
os->bit_rate = s->streams[i]->codecpar->bit_rate;
if (!os->bit_rate) {
int level = s->strict_std_compliance >= FF_COMPLIANCE_STRICT ?
AV_LOG_ERROR : AV_LOG_WARNING;
av_log(s, level, "No bit rate set for stream %d\n", i);
if (s->strict_std_compliance >= FF_COMPLIANCE_STRICT)
return AVERROR(EINVAL);
}
// copy AdaptationSet language and role from stream metadata
dict_copy_entry(&as->metadata, s->streams[i]->metadata, "language");
dict_copy_entry(&as->metadata, s->streams[i]->metadata, "role");
ctx = avformat_alloc_context();
if (!ctx)
return AVERROR(ENOMEM);
c->format_name = get_format_str(c->segment_type);
if (!c->format_name)
return AVERROR_MUXER_NOT_FOUND;
if (c->segment_type == SEGMENT_TYPE_WEBM) {
if ((!c->single_file && check_file_extension(c->init_seg_name, c->format_name) != 0) ||
(!c->single_file && check_file_extension(c->media_seg_name, c->format_name) != 0) ||
(c->single_file && check_file_extension(c->single_file_name, c->format_name) != 0)) {
av_log(s, AV_LOG_WARNING,
"One or many segment file names doesn't end with .webm. "
"Override -init_seg_name and/or -media_seg_name and/or "
"-single_file_name to end with the extension .webm\n");
}
}
ctx->oformat = av_guess_format(c->format_name, NULL, NULL);
if (!ctx->oformat)
return AVERROR_MUXER_NOT_FOUND;
os->ctx = ctx;
ctx->interrupt_callback = s->interrupt_callback;
ctx->opaque = s->opaque;
ctx->io_close = s->io_close;
ctx->io_open = s->io_open;
ctx->strict_std_compliance = s->strict_std_compliance;
if (!(st = avformat_new_stream(ctx, NULL)))
return AVERROR(ENOMEM);
avcodec_parameters_copy(st->codecpar, s->streams[i]->codecpar);
st->sample_aspect_ratio = s->streams[i]->sample_aspect_ratio;
st->time_base = s->streams[i]->time_base;
st->avg_frame_rate = s->streams[i]->avg_frame_rate;
ctx->avoid_negative_ts = s->avoid_negative_ts;
ctx->flags = s->flags;
if ((ret = avio_open_dyn_buf(&ctx->pb)) < 0)
return ret;
if (c->single_file) {
if (c->single_file_name)
ff_dash_fill_tmpl_params(os->initfile, sizeof(os->initfile), c->single_file_name, i, 0, os->bit_rate, 0);
else
snprintf(os->initfile, sizeof(os->initfile), "%s-stream%d.m4s", basename, i);
} else {
ff_dash_fill_tmpl_params(os->initfile, sizeof(os->initfile), c->init_seg_name, i, 0, os->bit_rate, 0);
}
snprintf(filename, sizeof(filename), "%s%s", c->dirname, os->initfile);
set_http_options(&opts, c);
ret = s->io_open(s, &os->out, filename, AVIO_FLAG_WRITE, &opts);
if (ret < 0)
return ret;
av_dict_free(&opts);
os->init_start_pos = 0;
if (c->format_options_str) {
ret = av_dict_parse_string(&opts, c->format_options_str, "=", ":", 0);
if (ret < 0)
return ret;
}
if (c->segment_type == SEGMENT_TYPE_MP4) {
if (c->streaming)
av_dict_set(&opts, "movflags", "frag_every_frame+dash+delay_moov", 0);
else
av_dict_set(&opts, "movflags", "frag_custom+dash+delay_moov", 0);
} else {
av_dict_set_int(&opts, "cluster_time_limit", c->seg_duration / 1000, 0);
av_dict_set_int(&opts, "cluster_size_limit", 5 * 1024 * 1024, 0); // set a large cluster size limit
av_dict_set_int(&opts, "dash", 1, 0);
av_dict_set_int(&opts, "dash_track_number", i + 1, 0);
av_dict_set_int(&opts, "live", 1, 0);
}
if ((ret = avformat_init_output(ctx, &opts)) < 0)
return ret;
os->ctx_inited = 1;
avio_flush(ctx->pb);
av_dict_free(&opts);
av_log(s, AV_LOG_VERBOSE, "Representation %d init segment will be written to: %s\n", i, filename);
s->streams[i]->time_base = st->time_base;
// If the muxer wants to shift timestamps, request to have them shifted
// already before being handed to this muxer, so we don't have mismatches
// between the MPD and the actual segments.
s->avoid_negative_ts = ctx->avoid_negative_ts;
if (st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO) {
AVRational avg_frame_rate = s->streams[i]->avg_frame_rate;
if (avg_frame_rate.num > 0) {
if (av_cmp_q(avg_frame_rate, as->min_frame_rate) < 0)
as->min_frame_rate = avg_frame_rate;
if (av_cmp_q(as->max_frame_rate, avg_frame_rate) < 0)
as->max_frame_rate = avg_frame_rate;
} else {
as->ambiguous_frame_rate = 1;
}
c->has_video = 1;
}
set_codec_str(s, st->codecpar, &st->avg_frame_rate, os->codec_str,
sizeof(os->codec_str));
os->first_pts = AV_NOPTS_VALUE;
os->max_pts = AV_NOPTS_VALUE;
os->last_dts = AV_NOPTS_VALUE;
os->segment_index = 1;
}
if (!c->has_video && c->seg_duration <= 0) {
av_log(s, AV_LOG_WARNING, "no video stream and no seg duration set\n");
return AVERROR(EINVAL);
}
return 0;
}
static int dash_write_header(AVFormatContext *s)
{
DASHContext *c = s->priv_data;
int i, ret;
for (i = 0; i < s->nb_streams; i++) {
OutputStream *os = &c->streams[i];
if ((ret = avformat_write_header(os->ctx, NULL)) < 0)
return ret;
// Flush init segment
// Only for WebM segment, since for mp4 delay_moov is set and
// the init segment is thus flushed after the first packets.
if (c->segment_type == SEGMENT_TYPE_WEBM &&
(ret = flush_init_segment(s, os)) < 0)
return ret;
}
return ret;
}
static int add_segment(OutputStream *os, const char *file,
int64_t time, int duration,
int64_t start_pos, int64_t range_length,
int64_t index_length, int next_exp_index)
{
int err;
Segment *seg;
if (os->nb_segments >= os->segments_size) {
os->segments_size = (os->segments_size + 1) * 2;
if ((err = av_reallocp(&os->segments, sizeof(*os->segments) *
os->segments_size)) < 0) {
os->segments_size = 0;
os->nb_segments = 0;
return err;
}
}
seg = av_mallocz(sizeof(*seg));
if (!seg)
return AVERROR(ENOMEM);
av_strlcpy(seg->file, file, sizeof(seg->file));
seg->time = time;
seg->duration = duration;
if (seg->time < 0) { // If pts<0, it is expected to be cut away with an edit list
seg->duration += seg->time;
seg->time = 0;
}
seg->start_pos = start_pos;
seg->range_length = range_length;
seg->index_length = index_length;
os->segments[os->nb_segments++] = seg;
os->segment_index++;
//correcting the segment index if it has fallen behind the expected value
if (os->segment_index < next_exp_index) {
av_log(NULL, AV_LOG_WARNING, "Correcting the segment index after file %s: current=%d corrected=%d\n",
file, os->segment_index, next_exp_index);
os->segment_index = next_exp_index;
}
return 0;
}
static void write_styp(AVIOContext *pb)
{
avio_wb32(pb, 24);
ffio_wfourcc(pb, "styp");
ffio_wfourcc(pb, "msdh");
avio_wb32(pb, 0); /* minor */
ffio_wfourcc(pb, "msdh");
ffio_wfourcc(pb, "msix");
}
static void find_index_range(AVFormatContext *s, const char *full_path,
int64_t pos, int *index_length)
{
uint8_t buf[8];
AVIOContext *pb;
int ret;
ret = s->io_open(s, &pb, full_path, AVIO_FLAG_READ, NULL);
if (ret < 0)
return;
if (avio_seek(pb, pos, SEEK_SET) != pos) {
ff_format_io_close(s, &pb);
return;
}
ret = avio_read(pb, buf, 8);
ff_format_io_close(s, &pb);
if (ret < 8)
return;
if (AV_RL32(&buf[4]) != MKTAG('s', 'i', 'd', 'x'))
return;
*index_length = AV_RB32(&buf[0]);
}
static int update_stream_extradata(AVFormatContext *s, OutputStream *os,
AVCodecParameters *par,
AVRational *frame_rate)
{
uint8_t *extradata;
if (os->ctx->streams[0]->codecpar->extradata_size || !par->extradata_size)
return 0;
extradata = av_malloc(par->extradata_size);
if (!extradata)
return AVERROR(ENOMEM);
memcpy(extradata, par->extradata, par->extradata_size);
os->ctx->streams[0]->codecpar->extradata = extradata;
os->ctx->streams[0]->codecpar->extradata_size = par->extradata_size;
set_codec_str(s, par, frame_rate, os->codec_str, sizeof(os->codec_str));
return 0;
}
static void dashenc_delete_file(AVFormatContext *s, char *filename) {
DASHContext *c = s->priv_data;
int http_base_proto = ff_is_http_proto(filename);
if (http_base_proto) {
AVIOContext *out = NULL;
AVDictionary *http_opts = NULL;
set_http_options(&http_opts, c);
av_dict_set(&http_opts, "method", "DELETE", 0);
if (dashenc_io_open(s, &out, filename, &http_opts) < 0) {
av_log(s, AV_LOG_ERROR, "failed to delete %s\n", filename);
}
av_dict_free(&http_opts);
dashenc_io_close(s, &out, filename);
} else if (unlink(filename) < 0) {
av_log(s, AV_LOG_ERROR, "failed to delete %s: %s\n", filename, strerror(errno));
}
}
static int dash_flush(AVFormatContext *s, int final, int stream)
{
DASHContext *c = s->priv_data;
int i, ret = 0;
const char *proto = avio_find_protocol_name(s->url);
int use_rename = proto && !strcmp(proto, "file");
int cur_flush_segment_index = 0, next_exp_index = -1;
if (stream >= 0) {
cur_flush_segment_index = c->streams[stream].segment_index;
//finding the next segment's expected index, based on the current pts value
if (c->use_template && !c->use_timeline && c->index_correction &&
c->streams[stream].last_pts != AV_NOPTS_VALUE &&
c->streams[stream].first_pts != AV_NOPTS_VALUE) {
int64_t pts_diff = av_rescale_q(c->streams[stream].last_pts -
c->streams[stream].first_pts,
s->streams[stream]->time_base,
AV_TIME_BASE_Q);
next_exp_index = (pts_diff / c->seg_duration) + 1;
}
}
for (i = 0; i < s->nb_streams; i++) {
OutputStream *os = &c->streams[i];
AVStream *st = s->streams[i];
int range_length, index_length = 0;
if (!os->packets_written)
continue;
// Flush the single stream that got a keyframe right now.
// Flush all audio streams as well, in sync with video keyframes,
// but not the other video streams.
if (stream >= 0 && i != stream) {
if (s->streams[i]->codecpar->codec_type != AVMEDIA_TYPE_AUDIO)
continue;
// Make sure we don't flush audio streams multiple times, when
// all video streams are flushed one at a time.
if (c->has_video && os->segment_index > cur_flush_segment_index)
continue;
}
if (!c->single_file) {
if (c->segment_type == SEGMENT_TYPE_MP4 && !os->written_len)
write_styp(os->ctx->pb);
} else {
snprintf(os->full_path, sizeof(os->full_path), "%s%s", c->dirname, os->initfile);
}
ret = flush_dynbuf(os, &range_length);
if (ret < 0)
break;
os->packets_written = 0;
if (c->single_file) {
find_index_range(s, os->full_path, os->pos, &index_length);
} else {
dashenc_io_close(s, &os->out, os->temp_path);
if (use_rename) {
ret = avpriv_io_move(os->temp_path, os->full_path);
if (ret < 0)
break;
}
}
if (!os->muxer_overhead)
os->muxer_overhead = ((int64_t) (range_length - os->total_pkt_size) *
8 * AV_TIME_BASE) /
av_rescale_q(os->max_pts - os->start_pts,
st->time_base, AV_TIME_BASE_Q);
os->total_pkt_size = 0;
if (!os->bit_rate) {
// calculate average bitrate of first segment
int64_t bitrate = (int64_t) range_length * 8 * AV_TIME_BASE / av_rescale_q(os->max_pts - os->start_pts,
st->time_base,
AV_TIME_BASE_Q);
if (bitrate >= 0)
os->bit_rate = bitrate;
}
add_segment(os, os->filename, os->start_pts, os->max_pts - os->start_pts, os->pos, range_length, index_length, next_exp_index);
av_log(s, AV_LOG_VERBOSE, "Representation %d media segment %d written to: %s\n", i, os->segment_index, os->full_path);
os->pos += range_length;
}
if (c->window_size || (final && c->remove_at_exit)) {
for (i = 0; i < s->nb_streams; i++) {
OutputStream *os = &c->streams[i];
int j;
int remove = os->nb_segments - c->window_size - c->extra_window_size;
if (final && c->remove_at_exit)
remove = os->nb_segments;
if (remove > 0) {
for (j = 0; j < remove; j++) {
char filename[1024];
snprintf(filename, sizeof(filename), "%s%s", c->dirname, os->segments[j]->file);
dashenc_delete_file(s, filename);
av_free(os->segments[j]);
}
os->nb_segments -= remove;
memmove(os->segments, os->segments + remove, os->nb_segments * sizeof(*os->segments));
}
}
}
if (ret >= 0)
ret = write_manifest(s, final);
return ret;
}
static int dash_write_packet(AVFormatContext *s, AVPacket *pkt)
{
DASHContext *c = s->priv_data;
AVStream *st = s->streams[pkt->stream_index];
OutputStream *os = &c->streams[pkt->stream_index];
int64_t seg_end_duration, elapsed_duration;
int ret;
ret = update_stream_extradata(s, os, st->codecpar, &st->avg_frame_rate);
if (ret < 0)
return ret;
// Fill in a heuristic guess of the packet duration, if none is available.
// The mp4 muxer will do something similar (for the last packet in a fragment)
// if nothing is set (setting it for the other packets doesn't hurt).
// By setting a nonzero duration here, we can be sure that the mp4 muxer won't
// invoke its heuristic (this doesn't have to be identical to that algorithm),
// so that we know the exact timestamps of fragments.
if (!pkt->duration && os->last_dts != AV_NOPTS_VALUE)
pkt->duration = pkt->dts - os->last_dts;
os->last_dts = pkt->dts;
// If forcing the stream to start at 0, the mp4 muxer will set the start
// timestamps to 0. Do the same here, to avoid mismatches in duration/timestamps.
if (os->first_pts == AV_NOPTS_VALUE &&
s->avoid_negative_ts == AVFMT_AVOID_NEG_TS_MAKE_ZERO) {
pkt->pts -= pkt->dts;
pkt->dts = 0;
}
if (os->first_pts == AV_NOPTS_VALUE)
os->first_pts = pkt->pts;
os->last_pts = pkt->pts;
if (!c->availability_start_time[0])
format_date_now(c->availability_start_time,
sizeof(c->availability_start_time));
if (!os->availability_time_offset && pkt->duration) {
int64_t frame_duration = av_rescale_q(pkt->duration, st->time_base,
AV_TIME_BASE_Q);
os->availability_time_offset = ((double) c->seg_duration -
frame_duration) / AV_TIME_BASE;
}
if (c->use_template && !c->use_timeline) {
elapsed_duration = pkt->pts - os->first_pts;
seg_end_duration = (int64_t) os->segment_index * c->seg_duration;
} else {
elapsed_duration = pkt->pts - os->start_pts;
seg_end_duration = c->seg_duration;
}
if ((!c->has_video || st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO) &&
pkt->flags & AV_PKT_FLAG_KEY && os->packets_written &&
av_compare_ts(elapsed_duration, st->time_base,
seg_end_duration, AV_TIME_BASE_Q) >= 0) {
int64_t prev_duration = c->last_duration;
c->last_duration = av_rescale_q(pkt->pts - os->start_pts,
st->time_base,
AV_TIME_BASE_Q);
c->total_duration = av_rescale_q(pkt->pts - os->first_pts,
st->time_base,
AV_TIME_BASE_Q);
if ((!c->use_timeline || !c->use_template) && prev_duration) {
if (c->last_duration < prev_duration*9/10 ||
c->last_duration > prev_duration*11/10) {
av_log(s, AV_LOG_WARNING,
"Segment durations differ too much, enable use_timeline "
"and use_template, or keep a stricter keyframe interval\n");
}
}
if ((ret = dash_flush(s, 0, pkt->stream_index)) < 0)
return ret;
}
if (!os->packets_written) {
// If we wrote a previous segment, adjust the start time of the segment
// to the end of the previous one (which is the same as the mp4 muxer
// does). This avoids gaps in the timeline.
if (os->max_pts != AV_NOPTS_VALUE)
os->start_pts = os->max_pts;
else
os->start_pts = pkt->pts;
}
if (os->max_pts == AV_NOPTS_VALUE)
os->max_pts = pkt->pts + pkt->duration;
else
os->max_pts = FFMAX(os->max_pts, pkt->pts + pkt->duration);
os->packets_written++;
os->total_pkt_size += pkt->size;
if ((ret = ff_write_chained(os->ctx, 0, pkt, s, 0)) < 0)
return ret;
if (!os->init_range_length)
flush_init_segment(s, os);
//open the output context when the first frame of a segment is ready
if (!c->single_file && os->packets_written == 1) {
AVDictionary *opts = NULL;
const char *proto = avio_find_protocol_name(s->url);
int use_rename = proto && !strcmp(proto, "file");
os->filename[0] = os->full_path[0] = os->temp_path[0] = '\0';
ff_dash_fill_tmpl_params(os->filename, sizeof(os->filename),
c->media_seg_name, pkt->stream_index,
os->segment_index, os->bit_rate, os->start_pts);
snprintf(os->full_path, sizeof(os->full_path), "%s%s", c->dirname,
os->filename);
snprintf(os->temp_path, sizeof(os->temp_path),
use_rename ? "%s.tmp" : "%s", os->full_path);
set_http_options(&opts, c);
ret = dashenc_io_open(s, &os->out, os->temp_path, &opts);
if (ret < 0)
return ret;
av_dict_free(&opts);
}
//write out the data immediately in streaming mode
if (c->streaming && c->segment_type == SEGMENT_TYPE_MP4) {
int len = 0;
uint8_t *buf = NULL;
if (!os->written_len)
write_styp(os->ctx->pb);
avio_flush(os->ctx->pb);
len = avio_get_dyn_buf (os->ctx->pb, &buf);
avio_write(os->out, buf + os->written_len, len - os->written_len);
os->written_len = len;
avio_flush(os->out);
}
return ret;
}
static int dash_write_trailer(AVFormatContext *s)
{
DASHContext *c = s->priv_data;
if (s->nb_streams > 0) {
OutputStream *os = &c->streams[0];
// If no segments have been written so far, try to do a crude
// guess of the segment duration
if (!c->last_duration)
c->last_duration = av_rescale_q(os->max_pts - os->start_pts,
s->streams[0]->time_base,
AV_TIME_BASE_Q);
c->total_duration = av_rescale_q(os->max_pts - os->first_pts,
s->streams[0]->time_base,
AV_TIME_BASE_Q);
}
dash_flush(s, 1, -1);
if (c->remove_at_exit) {
char filename[1024];
int i;
for (i = 0; i < s->nb_streams; i++) {
OutputStream *os = &c->streams[i];
snprintf(filename, sizeof(filename), "%s%s", c->dirname, os->initfile);
dashenc_delete_file(s, filename);
}
dashenc_delete_file(s, s->url);
}
return 0;
}
static int dash_check_bitstream(struct AVFormatContext *s, const AVPacket *avpkt)
{
DASHContext *c = s->priv_data;
OutputStream *os = &c->streams[avpkt->stream_index];
AVFormatContext *oc = os->ctx;
if (oc->oformat->check_bitstream) {
int ret;
AVPacket pkt = *avpkt;
pkt.stream_index = 0;
ret = oc->oformat->check_bitstream(oc, &pkt);
if (ret == 1) {
AVStream *st = s->streams[avpkt->stream_index];
AVStream *ost = oc->streams[0];
st->internal->bsfcs = ost->internal->bsfcs;
st->internal->nb_bsfcs = ost->internal->nb_bsfcs;
ost->internal->bsfcs = NULL;
ost->internal->nb_bsfcs = 0;
}
return ret;
}
return 1;
}
#define OFFSET(x) offsetof(DASHContext, x)
#define E AV_OPT_FLAG_ENCODING_PARAM
static const AVOption options[] = {
{ "adaptation_sets", "Adaptation sets. Syntax: id=0,streams=0,1,2 id=1,streams=3,4 and so on", OFFSET(adaptation_sets), AV_OPT_TYPE_STRING, { 0 }, 0, 0, AV_OPT_FLAG_ENCODING_PARAM },
{ "window_size", "number of segments kept in the manifest", OFFSET(window_size), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, E },
{ "extra_window_size", "number of segments kept outside of the manifest before removing from disk", OFFSET(extra_window_size), AV_OPT_TYPE_INT, { .i64 = 5 }, 0, INT_MAX, E },
#if FF_API_DASH_MIN_SEG_DURATION
{ "min_seg_duration", "minimum segment duration (in microseconds) (will be deprecated)", OFFSET(min_seg_duration), AV_OPT_TYPE_INT, { .i64 = 5000000 }, 0, INT_MAX, E },
#endif
{ "seg_duration", "segment duration (in seconds, fractional value can be set)", OFFSET(seg_duration), AV_OPT_TYPE_DURATION, { .i64 = 5000000 }, 0, INT_MAX, E },
{ "remove_at_exit", "remove all segments when finished", OFFSET(remove_at_exit), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, E },
{ "use_template", "Use SegmentTemplate instead of SegmentList", OFFSET(use_template), AV_OPT_TYPE_BOOL, { .i64 = 1 }, 0, 1, E },
{ "use_timeline", "Use SegmentTimeline in SegmentTemplate", OFFSET(use_timeline), AV_OPT_TYPE_BOOL, { .i64 = 1 }, 0, 1, E },
{ "single_file", "Store all segments in one file, accessed using byte ranges", OFFSET(single_file), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, E },
{ "single_file_name", "DASH-templated name to be used for baseURL. Implies storing all segments in one file, accessed using byte ranges", OFFSET(single_file_name), AV_OPT_TYPE_STRING, { .str = NULL }, 0, 0, E },
{ "init_seg_name", "DASH-templated name to used for the initialization segment", OFFSET(init_seg_name), AV_OPT_TYPE_STRING, {.str = "init-stream$RepresentationID$.m4s"}, 0, 0, E },
{ "media_seg_name", "DASH-templated name to used for the media segments", OFFSET(media_seg_name), AV_OPT_TYPE_STRING, {.str = "chunk-stream$RepresentationID$-$Number%05d$.m4s"}, 0, 0, E },
{ "utc_timing_url", "URL of the page that will return the UTC timestamp in ISO format", OFFSET(utc_timing_url), AV_OPT_TYPE_STRING, { 0 }, 0, 0, E },
{ "method", "set the HTTP method", OFFSET(method), AV_OPT_TYPE_STRING, {.str = NULL}, 0, 0, E },
{ "http_user_agent", "override User-Agent field in HTTP header", OFFSET(user_agent), AV_OPT_TYPE_STRING, {.str = NULL}, 0, 0, E},
{ "http_persistent", "Use persistent HTTP connections", OFFSET(http_persistent), AV_OPT_TYPE_BOOL, {.i64 = 0 }, 0, 1, E },
{ "hls_playlist", "Generate HLS playlist files(master.m3u8, media_%d.m3u8)", OFFSET(hls_playlist), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, E },
{ "streaming", "Enable/Disable streaming mode of output. Each frame will be moof fragment", OFFSET(streaming), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, E },
{ "timeout", "set timeout for socket I/O operations", OFFSET(timeout), AV_OPT_TYPE_DURATION, { .i64 = -1 }, -1, INT_MAX, .flags = E },
{ "index_correction", "Enable/Disable segment index correction logic", OFFSET(index_correction), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, E },
{ "format_options","set list of options for the container format (mp4/webm) used for dash", OFFSET(format_options_str), AV_OPT_TYPE_STRING, {.str = NULL}, 0, 0, E},
{ "dash_segment_type", "set dash segment files type", OFFSET(segment_type), AV_OPT_TYPE_INT, {.i64 = SEGMENT_TYPE_MP4 }, 0, SEGMENT_TYPE_NB - 1, E, "segment_type"},
{ "mp4", "make segment file in ISOBMFF format", 0, AV_OPT_TYPE_CONST, {.i64 = SEGMENT_TYPE_MP4 }, 0, UINT_MAX, E, "segment_type"},
{ "webm", "make segment file in WebM format", 0, AV_OPT_TYPE_CONST, {.i64 = SEGMENT_TYPE_WEBM }, 0, UINT_MAX, E, "segment_type"},
{ NULL },
};
static const AVClass dash_class = {
.class_name = "dash muxer",
.item_name = av_default_item_name,
.option = options,
.version = LIBAVUTIL_VERSION_INT,
};
AVOutputFormat ff_dash_muxer = {
.name = "dash",
.long_name = NULL_IF_CONFIG_SMALL("DASH Muxer"),
.extensions = "mpd",
.priv_data_size = sizeof(DASHContext),
.audio_codec = AV_CODEC_ID_AAC,
.video_codec = AV_CODEC_ID_H264,
.flags = AVFMT_GLOBALHEADER | AVFMT_NOFILE | AVFMT_TS_NEGATIVE,
.init = dash_init,
.write_header = dash_write_header,
.write_packet = dash_write_packet,
.write_trailer = dash_write_trailer,
.deinit = dash_free,
.check_bitstream = dash_check_bitstream,
.priv_class = &dash_class,
};