mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-12-23 12:43:46 +02:00
059fc2d9da
Currently said list contains only the pixel formats that are always supported irrespective of the range and the value of strict_std_compliance. This makes the MJPEG encoder an outlier as all other codecs put all potentially supported pixel formats into said list and error out if the chosen pixel format is unsupported. This commit brings it therefore in line with the other encoders. The behaviour of fftools/ffmpeg_filter.c has been preserved. A more informed decision would be possible if colour range were available at this point, but it isn't. Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@outlook.com>
681 lines
23 KiB
C
681 lines
23 KiB
C
/*
|
|
* generic encoding-related code
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "libavutil/attributes.h"
|
|
#include "libavutil/avassert.h"
|
|
#include "libavutil/frame.h"
|
|
#include "libavutil/imgutils.h"
|
|
#include "libavutil/internal.h"
|
|
#include "libavutil/samplefmt.h"
|
|
|
|
#include "avcodec.h"
|
|
#include "encode.h"
|
|
#include "frame_thread_encoder.h"
|
|
#include "internal.h"
|
|
|
|
int ff_alloc_packet2(AVCodecContext *avctx, AVPacket *avpkt, int64_t size, int64_t min_size)
|
|
{
|
|
if (size < 0 || size > INT_MAX - AV_INPUT_BUFFER_PADDING_SIZE) {
|
|
av_log(avctx, AV_LOG_ERROR, "Invalid minimum required packet size %"PRId64" (max allowed is %d)\n",
|
|
size, INT_MAX - AV_INPUT_BUFFER_PADDING_SIZE);
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
av_assert0(!avpkt->data);
|
|
|
|
if (avctx && 2*min_size < size) { // FIXME The factor needs to be finetuned
|
|
av_fast_padded_malloc(&avctx->internal->byte_buffer, &avctx->internal->byte_buffer_size, size);
|
|
avpkt->data = avctx->internal->byte_buffer;
|
|
avpkt->size = size;
|
|
}
|
|
|
|
if (!avpkt->data) {
|
|
int ret = av_new_packet(avpkt, size);
|
|
if (ret < 0)
|
|
av_log(avctx, AV_LOG_ERROR, "Failed to allocate packet of size %"PRId64"\n", size);
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int avcodec_default_get_encode_buffer(AVCodecContext *avctx, AVPacket *avpkt, int flags)
|
|
{
|
|
int ret;
|
|
|
|
if (avpkt->size < 0 || avpkt->size > INT_MAX - AV_INPUT_BUFFER_PADDING_SIZE)
|
|
return AVERROR(EINVAL);
|
|
|
|
if (avpkt->data || avpkt->buf) {
|
|
av_log(avctx, AV_LOG_ERROR, "avpkt->{data,buf} != NULL in avcodec_default_get_encode_buffer()\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
ret = av_buffer_realloc(&avpkt->buf, avpkt->size + AV_INPUT_BUFFER_PADDING_SIZE);
|
|
if (ret < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "Failed to allocate packet of size %d\n", avpkt->size);
|
|
return ret;
|
|
}
|
|
avpkt->data = avpkt->buf->data;
|
|
memset(avpkt->data + avpkt->size, 0, AV_INPUT_BUFFER_PADDING_SIZE);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int ff_get_encode_buffer(AVCodecContext *avctx, AVPacket *avpkt, int64_t size, int flags)
|
|
{
|
|
int ret;
|
|
|
|
if (size < 0 || size > INT_MAX - AV_INPUT_BUFFER_PADDING_SIZE)
|
|
return AVERROR(EINVAL);
|
|
|
|
av_assert0(!avpkt->data && !avpkt->buf);
|
|
|
|
avpkt->size = size;
|
|
ret = avctx->get_encode_buffer(avctx, avpkt, flags);
|
|
if (ret < 0)
|
|
goto fail;
|
|
|
|
if (!avpkt->data || !avpkt->buf) {
|
|
av_log(avctx, AV_LOG_ERROR, "No buffer returned by get_encode_buffer()\n");
|
|
ret = AVERROR(EINVAL);
|
|
goto fail;
|
|
}
|
|
|
|
ret = 0;
|
|
fail:
|
|
if (ret < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "get_encode_buffer() failed\n");
|
|
av_packet_unref(avpkt);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* Pad last frame with silence.
|
|
*/
|
|
static int pad_last_frame(AVCodecContext *s, AVFrame *frame, const AVFrame *src)
|
|
{
|
|
int ret;
|
|
|
|
frame->format = src->format;
|
|
frame->channel_layout = src->channel_layout;
|
|
frame->channels = src->channels;
|
|
frame->nb_samples = s->frame_size;
|
|
ret = av_frame_get_buffer(frame, 0);
|
|
if (ret < 0)
|
|
goto fail;
|
|
|
|
ret = av_frame_copy_props(frame, src);
|
|
if (ret < 0)
|
|
goto fail;
|
|
|
|
if ((ret = av_samples_copy(frame->extended_data, src->extended_data, 0, 0,
|
|
src->nb_samples, s->channels, s->sample_fmt)) < 0)
|
|
goto fail;
|
|
if ((ret = av_samples_set_silence(frame->extended_data, src->nb_samples,
|
|
frame->nb_samples - src->nb_samples,
|
|
s->channels, s->sample_fmt)) < 0)
|
|
goto fail;
|
|
|
|
return 0;
|
|
|
|
fail:
|
|
av_frame_unref(frame);
|
|
return ret;
|
|
}
|
|
|
|
int avcodec_encode_subtitle(AVCodecContext *avctx, uint8_t *buf, int buf_size,
|
|
const AVSubtitle *sub)
|
|
{
|
|
int ret;
|
|
if (sub->start_display_time) {
|
|
av_log(avctx, AV_LOG_ERROR, "start_display_time must be 0.\n");
|
|
return -1;
|
|
}
|
|
|
|
ret = avctx->codec->encode_sub(avctx, buf, buf_size, sub);
|
|
avctx->frame_number++;
|
|
return ret;
|
|
}
|
|
|
|
int ff_encode_get_frame(AVCodecContext *avctx, AVFrame *frame)
|
|
{
|
|
AVCodecInternal *avci = avctx->internal;
|
|
|
|
if (avci->draining)
|
|
return AVERROR_EOF;
|
|
|
|
if (!avci->buffer_frame->buf[0])
|
|
return AVERROR(EAGAIN);
|
|
|
|
av_frame_move_ref(frame, avci->buffer_frame);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int encode_simple_internal(AVCodecContext *avctx, AVPacket *avpkt)
|
|
{
|
|
AVCodecInternal *avci = avctx->internal;
|
|
EncodeSimpleContext *es = &avci->es;
|
|
AVFrame *frame = es->in_frame;
|
|
int got_packet;
|
|
int ret;
|
|
|
|
if (avci->draining_done)
|
|
return AVERROR_EOF;
|
|
|
|
if (!frame->buf[0] && !avci->draining) {
|
|
av_frame_unref(frame);
|
|
ret = ff_encode_get_frame(avctx, frame);
|
|
if (ret < 0 && ret != AVERROR_EOF)
|
|
return ret;
|
|
}
|
|
|
|
if (!frame->buf[0]) {
|
|
if (!(avctx->codec->capabilities & AV_CODEC_CAP_DELAY ||
|
|
(avci->frame_thread_encoder && avctx->active_thread_type & FF_THREAD_FRAME)))
|
|
return AVERROR_EOF;
|
|
|
|
// Flushing is signaled with a NULL frame
|
|
frame = NULL;
|
|
}
|
|
|
|
got_packet = 0;
|
|
|
|
av_assert0(avctx->codec->encode2);
|
|
|
|
if (CONFIG_FRAME_THREAD_ENCODER &&
|
|
avci->frame_thread_encoder && (avctx->active_thread_type & FF_THREAD_FRAME))
|
|
/* This might modify frame, but it doesn't matter, because
|
|
* the frame properties used below are not used for video
|
|
* (due to the delay inherent in frame threaded encoding, it makes
|
|
* no sense to use the properties of the current frame anyway). */
|
|
ret = ff_thread_video_encode_frame(avctx, avpkt, frame, &got_packet);
|
|
else {
|
|
ret = avctx->codec->encode2(avctx, avpkt, frame, &got_packet);
|
|
if (avctx->codec->type == AVMEDIA_TYPE_VIDEO && !ret && got_packet &&
|
|
!(avctx->codec->capabilities & AV_CODEC_CAP_DELAY))
|
|
avpkt->pts = avpkt->dts = frame->pts;
|
|
}
|
|
|
|
av_assert0(ret <= 0);
|
|
|
|
emms_c();
|
|
|
|
if (!ret && got_packet) {
|
|
if (avpkt->data) {
|
|
ret = av_packet_make_refcounted(avpkt);
|
|
if (ret < 0)
|
|
goto end;
|
|
}
|
|
|
|
if (frame && !(avctx->codec->capabilities & AV_CODEC_CAP_DELAY)) {
|
|
if (avctx->codec->type == AVMEDIA_TYPE_AUDIO) {
|
|
if (avpkt->pts == AV_NOPTS_VALUE)
|
|
avpkt->pts = frame->pts;
|
|
if (!avpkt->duration)
|
|
avpkt->duration = ff_samples_to_time_base(avctx,
|
|
frame->nb_samples);
|
|
}
|
|
}
|
|
if (avctx->codec->type == AVMEDIA_TYPE_AUDIO) {
|
|
/* NOTE: if we add any audio encoders which output non-keyframe packets,
|
|
* this needs to be moved to the encoders, but for now we can do it
|
|
* here to simplify things */
|
|
avpkt->flags |= AV_PKT_FLAG_KEY;
|
|
avpkt->dts = avpkt->pts;
|
|
}
|
|
}
|
|
|
|
if (avci->draining && !got_packet)
|
|
avci->draining_done = 1;
|
|
|
|
end:
|
|
if (ret < 0 || !got_packet)
|
|
av_packet_unref(avpkt);
|
|
|
|
if (frame) {
|
|
if (!ret)
|
|
avctx->frame_number++;
|
|
av_frame_unref(frame);
|
|
}
|
|
|
|
if (got_packet)
|
|
// Encoders must always return ref-counted buffers.
|
|
// Side-data only packets have no data and can be not ref-counted.
|
|
av_assert0(!avpkt->data || avpkt->buf);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int encode_simple_receive_packet(AVCodecContext *avctx, AVPacket *avpkt)
|
|
{
|
|
int ret;
|
|
|
|
while (!avpkt->data && !avpkt->side_data) {
|
|
ret = encode_simple_internal(avctx, avpkt);
|
|
if (ret < 0)
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int encode_receive_packet_internal(AVCodecContext *avctx, AVPacket *avpkt)
|
|
{
|
|
AVCodecInternal *avci = avctx->internal;
|
|
int ret;
|
|
|
|
if (avci->draining_done)
|
|
return AVERROR_EOF;
|
|
|
|
av_assert0(!avpkt->data && !avpkt->side_data);
|
|
|
|
if (avctx->codec->type == AVMEDIA_TYPE_VIDEO) {
|
|
if ((avctx->flags & AV_CODEC_FLAG_PASS1) && avctx->stats_out)
|
|
avctx->stats_out[0] = '\0';
|
|
if (av_image_check_size2(avctx->width, avctx->height, avctx->max_pixels, AV_PIX_FMT_NONE, 0, avctx))
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
if (avctx->codec->receive_packet) {
|
|
ret = avctx->codec->receive_packet(avctx, avpkt);
|
|
if (ret < 0)
|
|
av_packet_unref(avpkt);
|
|
else
|
|
// Encoders must always return ref-counted buffers.
|
|
// Side-data only packets have no data and can be not ref-counted.
|
|
av_assert0(!avpkt->data || avpkt->buf);
|
|
} else
|
|
ret = encode_simple_receive_packet(avctx, avpkt);
|
|
|
|
if (ret == AVERROR_EOF)
|
|
avci->draining_done = 1;
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int encode_send_frame_internal(AVCodecContext *avctx, const AVFrame *src)
|
|
{
|
|
AVCodecInternal *avci = avctx->internal;
|
|
AVFrame *dst = avci->buffer_frame;
|
|
int ret;
|
|
|
|
if (avctx->codec->type == AVMEDIA_TYPE_AUDIO) {
|
|
/* extract audio service type metadata */
|
|
AVFrameSideData *sd = av_frame_get_side_data(src, AV_FRAME_DATA_AUDIO_SERVICE_TYPE);
|
|
if (sd && sd->size >= sizeof(enum AVAudioServiceType))
|
|
avctx->audio_service_type = *(enum AVAudioServiceType*)sd->data;
|
|
|
|
/* check for valid frame size */
|
|
if (avctx->codec->capabilities & AV_CODEC_CAP_SMALL_LAST_FRAME) {
|
|
if (src->nb_samples > avctx->frame_size) {
|
|
av_log(avctx, AV_LOG_ERROR, "more samples than frame size\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
} else if (!(avctx->codec->capabilities & AV_CODEC_CAP_VARIABLE_FRAME_SIZE)) {
|
|
/* if we already got an undersized frame, that must have been the last */
|
|
if (avctx->internal->last_audio_frame) {
|
|
av_log(avctx, AV_LOG_ERROR, "frame_size (%d) was not respected for a non-last frame\n", avctx->frame_size);
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
if (src->nb_samples < avctx->frame_size) {
|
|
ret = pad_last_frame(avctx, dst, src);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
avctx->internal->last_audio_frame = 1;
|
|
} else if (src->nb_samples > avctx->frame_size) {
|
|
av_log(avctx, AV_LOG_ERROR, "nb_samples (%d) != frame_size (%d)\n", src->nb_samples, avctx->frame_size);
|
|
return AVERROR(EINVAL);
|
|
}
|
|
}
|
|
}
|
|
|
|
if (!dst->data[0]) {
|
|
ret = av_frame_ref(dst, src);
|
|
if (ret < 0)
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int attribute_align_arg avcodec_send_frame(AVCodecContext *avctx, const AVFrame *frame)
|
|
{
|
|
AVCodecInternal *avci = avctx->internal;
|
|
int ret;
|
|
|
|
if (!avcodec_is_open(avctx) || !av_codec_is_encoder(avctx->codec))
|
|
return AVERROR(EINVAL);
|
|
|
|
if (avci->draining)
|
|
return AVERROR_EOF;
|
|
|
|
if (avci->buffer_frame->data[0])
|
|
return AVERROR(EAGAIN);
|
|
|
|
if (!frame) {
|
|
avci->draining = 1;
|
|
} else {
|
|
ret = encode_send_frame_internal(avctx, frame);
|
|
if (ret < 0)
|
|
return ret;
|
|
}
|
|
|
|
if (!avci->buffer_pkt->data && !avci->buffer_pkt->side_data) {
|
|
ret = encode_receive_packet_internal(avctx, avci->buffer_pkt);
|
|
if (ret < 0 && ret != AVERROR(EAGAIN) && ret != AVERROR_EOF)
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int attribute_align_arg avcodec_receive_packet(AVCodecContext *avctx, AVPacket *avpkt)
|
|
{
|
|
AVCodecInternal *avci = avctx->internal;
|
|
int ret;
|
|
|
|
av_packet_unref(avpkt);
|
|
|
|
if (!avcodec_is_open(avctx) || !av_codec_is_encoder(avctx->codec))
|
|
return AVERROR(EINVAL);
|
|
|
|
if (avci->buffer_pkt->data || avci->buffer_pkt->side_data) {
|
|
av_packet_move_ref(avpkt, avci->buffer_pkt);
|
|
} else {
|
|
ret = encode_receive_packet_internal(avctx, avpkt);
|
|
if (ret < 0)
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
#if FF_API_OLD_ENCDEC
|
|
static int compat_encode(AVCodecContext *avctx, AVPacket *avpkt,
|
|
int *got_packet, const AVFrame *frame)
|
|
{
|
|
AVCodecInternal *avci = avctx->internal;
|
|
AVPacket user_pkt;
|
|
int ret;
|
|
|
|
*got_packet = 0;
|
|
|
|
if (frame && avctx->codec->type == AVMEDIA_TYPE_VIDEO) {
|
|
if (frame->format == AV_PIX_FMT_NONE)
|
|
av_log(avctx, AV_LOG_WARNING, "AVFrame.format is not set\n");
|
|
if (frame->width == 0 || frame->height == 0)
|
|
av_log(avctx, AV_LOG_WARNING, "AVFrame.width or height is not set\n");
|
|
}
|
|
|
|
if (avctx->codec->capabilities & AV_CODEC_CAP_DR1) {
|
|
av_log(avctx, AV_LOG_WARNING, "The deprecated avcodec_encode_* API does not support "
|
|
"AV_CODEC_CAP_DR1 encoders\n");
|
|
return AVERROR(ENOSYS);
|
|
}
|
|
|
|
ret = avcodec_send_frame(avctx, frame);
|
|
if (ret == AVERROR_EOF)
|
|
ret = 0;
|
|
else if (ret == AVERROR(EAGAIN)) {
|
|
/* we fully drain all the output in each encode call, so this should not
|
|
* ever happen */
|
|
return AVERROR_BUG;
|
|
} else if (ret < 0)
|
|
return ret;
|
|
|
|
av_packet_move_ref(&user_pkt, avpkt);
|
|
while (ret >= 0) {
|
|
ret = avcodec_receive_packet(avctx, avpkt);
|
|
if (ret < 0) {
|
|
if (ret == AVERROR(EAGAIN) || ret == AVERROR_EOF)
|
|
ret = 0;
|
|
goto finish;
|
|
}
|
|
|
|
if (avpkt != avci->compat_encode_packet) {
|
|
if (avpkt->data && user_pkt.data) {
|
|
if (user_pkt.size >= avpkt->size) {
|
|
memcpy(user_pkt.data, avpkt->data, avpkt->size);
|
|
av_buffer_unref(&avpkt->buf);
|
|
avpkt->buf = user_pkt.buf;
|
|
avpkt->data = user_pkt.data;
|
|
FF_DISABLE_DEPRECATION_WARNINGS
|
|
av_init_packet(&user_pkt);
|
|
FF_ENABLE_DEPRECATION_WARNINGS
|
|
} else {
|
|
av_log(avctx, AV_LOG_ERROR, "Provided packet is too small, needs to be %d\n", avpkt->size);
|
|
av_packet_unref(avpkt);
|
|
ret = AVERROR(EINVAL);
|
|
goto finish;
|
|
}
|
|
}
|
|
|
|
*got_packet = 1;
|
|
avpkt = avci->compat_encode_packet;
|
|
} else {
|
|
if (!avci->compat_decode_warned) {
|
|
av_log(avctx, AV_LOG_WARNING, "The deprecated avcodec_encode_* "
|
|
"API cannot return all the packets for this encoder. "
|
|
"Some packets will be dropped. Update your code to the "
|
|
"new encoding API to fix this.\n");
|
|
avci->compat_decode_warned = 1;
|
|
av_packet_unref(avpkt);
|
|
}
|
|
}
|
|
|
|
if (avci->draining)
|
|
break;
|
|
}
|
|
|
|
finish:
|
|
if (ret < 0)
|
|
av_packet_unref(&user_pkt);
|
|
|
|
return ret;
|
|
}
|
|
|
|
int attribute_align_arg avcodec_encode_audio2(AVCodecContext *avctx,
|
|
AVPacket *avpkt,
|
|
const AVFrame *frame,
|
|
int *got_packet_ptr)
|
|
{
|
|
int ret = compat_encode(avctx, avpkt, got_packet_ptr, frame);
|
|
|
|
if (ret < 0)
|
|
av_packet_unref(avpkt);
|
|
|
|
return ret;
|
|
}
|
|
|
|
int attribute_align_arg avcodec_encode_video2(AVCodecContext *avctx,
|
|
AVPacket *avpkt,
|
|
const AVFrame *frame,
|
|
int *got_packet_ptr)
|
|
{
|
|
int ret = compat_encode(avctx, avpkt, got_packet_ptr, frame);
|
|
|
|
if (ret < 0)
|
|
av_packet_unref(avpkt);
|
|
|
|
return ret;
|
|
}
|
|
#endif
|
|
|
|
int ff_encode_preinit(AVCodecContext *avctx)
|
|
{
|
|
int i;
|
|
#if FF_API_CODED_FRAME
|
|
FF_DISABLE_DEPRECATION_WARNINGS
|
|
avctx->coded_frame = av_frame_alloc();
|
|
if (!avctx->coded_frame) {
|
|
return AVERROR(ENOMEM);
|
|
}
|
|
FF_ENABLE_DEPRECATION_WARNINGS
|
|
#endif
|
|
|
|
if (avctx->time_base.num <= 0 || avctx->time_base.den <= 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "The encoder timebase is not set.\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
if (avctx->codec->sample_fmts) {
|
|
for (i = 0; avctx->codec->sample_fmts[i] != AV_SAMPLE_FMT_NONE; i++) {
|
|
if (avctx->sample_fmt == avctx->codec->sample_fmts[i])
|
|
break;
|
|
if (avctx->channels == 1 &&
|
|
av_get_planar_sample_fmt(avctx->sample_fmt) ==
|
|
av_get_planar_sample_fmt(avctx->codec->sample_fmts[i])) {
|
|
avctx->sample_fmt = avctx->codec->sample_fmts[i];
|
|
break;
|
|
}
|
|
}
|
|
if (avctx->codec->sample_fmts[i] == AV_SAMPLE_FMT_NONE) {
|
|
char buf[128];
|
|
snprintf(buf, sizeof(buf), "%d", avctx->sample_fmt);
|
|
av_log(avctx, AV_LOG_ERROR, "Specified sample format %s is invalid or not supported\n",
|
|
(char *)av_x_if_null(av_get_sample_fmt_name(avctx->sample_fmt), buf));
|
|
return AVERROR(EINVAL);
|
|
}
|
|
}
|
|
if (avctx->codec->pix_fmts) {
|
|
for (i = 0; avctx->codec->pix_fmts[i] != AV_PIX_FMT_NONE; i++)
|
|
if (avctx->pix_fmt == avctx->codec->pix_fmts[i])
|
|
break;
|
|
if (avctx->codec->pix_fmts[i] == AV_PIX_FMT_NONE) {
|
|
char buf[128];
|
|
snprintf(buf, sizeof(buf), "%d", avctx->pix_fmt);
|
|
av_log(avctx, AV_LOG_ERROR, "Specified pixel format %s is invalid or not supported\n",
|
|
(char *)av_x_if_null(av_get_pix_fmt_name(avctx->pix_fmt), buf));
|
|
return AVERROR(EINVAL);
|
|
}
|
|
if (avctx->codec->pix_fmts[i] == AV_PIX_FMT_YUVJ420P ||
|
|
avctx->codec->pix_fmts[i] == AV_PIX_FMT_YUVJ411P ||
|
|
avctx->codec->pix_fmts[i] == AV_PIX_FMT_YUVJ422P ||
|
|
avctx->codec->pix_fmts[i] == AV_PIX_FMT_YUVJ440P ||
|
|
avctx->codec->pix_fmts[i] == AV_PIX_FMT_YUVJ444P)
|
|
avctx->color_range = AVCOL_RANGE_JPEG;
|
|
}
|
|
if (avctx->codec->supported_samplerates) {
|
|
for (i = 0; avctx->codec->supported_samplerates[i] != 0; i++)
|
|
if (avctx->sample_rate == avctx->codec->supported_samplerates[i])
|
|
break;
|
|
if (avctx->codec->supported_samplerates[i] == 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "Specified sample rate %d is not supported\n",
|
|
avctx->sample_rate);
|
|
return AVERROR(EINVAL);
|
|
}
|
|
}
|
|
if (avctx->sample_rate < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "Specified sample rate %d is not supported\n",
|
|
avctx->sample_rate);
|
|
return AVERROR(EINVAL);
|
|
}
|
|
if (avctx->codec->channel_layouts) {
|
|
if (!avctx->channel_layout) {
|
|
av_log(avctx, AV_LOG_WARNING, "Channel layout not specified\n");
|
|
} else {
|
|
for (i = 0; avctx->codec->channel_layouts[i] != 0; i++)
|
|
if (avctx->channel_layout == avctx->codec->channel_layouts[i])
|
|
break;
|
|
if (avctx->codec->channel_layouts[i] == 0) {
|
|
char buf[512];
|
|
av_get_channel_layout_string(buf, sizeof(buf), -1, avctx->channel_layout);
|
|
av_log(avctx, AV_LOG_ERROR, "Specified channel layout '%s' is not supported\n", buf);
|
|
return AVERROR(EINVAL);
|
|
}
|
|
}
|
|
}
|
|
if (avctx->channel_layout && avctx->channels) {
|
|
int channels = av_get_channel_layout_nb_channels(avctx->channel_layout);
|
|
if (channels != avctx->channels) {
|
|
char buf[512];
|
|
av_get_channel_layout_string(buf, sizeof(buf), -1, avctx->channel_layout);
|
|
av_log(avctx, AV_LOG_ERROR,
|
|
"Channel layout '%s' with %d channels does not match number of specified channels %d\n",
|
|
buf, channels, avctx->channels);
|
|
return AVERROR(EINVAL);
|
|
}
|
|
} else if (avctx->channel_layout) {
|
|
avctx->channels = av_get_channel_layout_nb_channels(avctx->channel_layout);
|
|
}
|
|
if (avctx->channels < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "Specified number of channels %d is not supported\n",
|
|
avctx->channels);
|
|
return AVERROR(EINVAL);
|
|
}
|
|
if(avctx->codec_type == AVMEDIA_TYPE_VIDEO) {
|
|
const AVPixFmtDescriptor *pixdesc = av_pix_fmt_desc_get(avctx->pix_fmt);
|
|
if ( avctx->bits_per_raw_sample < 0
|
|
|| (avctx->bits_per_raw_sample > 8 && pixdesc->comp[0].depth <= 8)) {
|
|
av_log(avctx, AV_LOG_WARNING, "Specified bit depth %d not possible with the specified pixel formats depth %d\n",
|
|
avctx->bits_per_raw_sample, pixdesc->comp[0].depth);
|
|
avctx->bits_per_raw_sample = pixdesc->comp[0].depth;
|
|
}
|
|
if (avctx->width <= 0 || avctx->height <= 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "dimensions not set\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
}
|
|
if ( (avctx->codec_type == AVMEDIA_TYPE_VIDEO || avctx->codec_type == AVMEDIA_TYPE_AUDIO)
|
|
&& avctx->bit_rate>0 && avctx->bit_rate<1000) {
|
|
av_log(avctx, AV_LOG_WARNING, "Bitrate %"PRId64" is extremely low, maybe you mean %"PRId64"k\n", avctx->bit_rate, avctx->bit_rate);
|
|
}
|
|
|
|
if (!avctx->rc_initial_buffer_occupancy)
|
|
avctx->rc_initial_buffer_occupancy = avctx->rc_buffer_size * 3LL / 4;
|
|
|
|
if (avctx->ticks_per_frame && avctx->time_base.num &&
|
|
avctx->ticks_per_frame > INT_MAX / avctx->time_base.num) {
|
|
av_log(avctx, AV_LOG_ERROR,
|
|
"ticks_per_frame %d too large for the timebase %d/%d.",
|
|
avctx->ticks_per_frame,
|
|
avctx->time_base.num,
|
|
avctx->time_base.den);
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
if (avctx->hw_frames_ctx) {
|
|
AVHWFramesContext *frames_ctx = (AVHWFramesContext*)avctx->hw_frames_ctx->data;
|
|
if (frames_ctx->format != avctx->pix_fmt) {
|
|
av_log(avctx, AV_LOG_ERROR,
|
|
"Mismatching AVCodecContext.pix_fmt and AVHWFramesContext.format\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
if (avctx->sw_pix_fmt != AV_PIX_FMT_NONE &&
|
|
avctx->sw_pix_fmt != frames_ctx->sw_format) {
|
|
av_log(avctx, AV_LOG_ERROR,
|
|
"Mismatching AVCodecContext.sw_pix_fmt (%s) "
|
|
"and AVHWFramesContext.sw_format (%s)\n",
|
|
av_get_pix_fmt_name(avctx->sw_pix_fmt),
|
|
av_get_pix_fmt_name(frames_ctx->sw_format));
|
|
return AVERROR(EINVAL);
|
|
}
|
|
avctx->sw_pix_fmt = frames_ctx->sw_format;
|
|
}
|
|
|
|
return 0;
|
|
}
|