mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-12-23 12:43:46 +02:00
Merge commit 'e3db34291f4401a16f6ac92721617a9f33cd4c31'
* commit 'e3db34291f4401a16f6ac92721617a9f33cd4c31': amrnb: decode directly to the user-provided AVFrame als: decode directly to the user-provided AVFrame alac: decode directly to the user-provided AVFrame adxenc: alloc/free coded_frame instead of keeping it in the ADXContext adx: decode directly to the user-provided AVFrame Conflicts: libavcodec/alsdec.c libavcodec/amrnbdec.c Merged-by: Michael Niedermayer <michaelni@gmx.at>
This commit is contained in:
commit
d13f434dbb
@ -40,7 +40,6 @@ typedef struct ADXChannelState {
|
|||||||
} ADXChannelState;
|
} ADXChannelState;
|
||||||
|
|
||||||
typedef struct ADXContext {
|
typedef struct ADXContext {
|
||||||
AVFrame frame;
|
|
||||||
int channels;
|
int channels;
|
||||||
ADXChannelState prev[2];
|
ADXChannelState prev[2];
|
||||||
int header_parsed;
|
int header_parsed;
|
||||||
|
@ -52,9 +52,6 @@ static av_cold int adx_decode_init(AVCodecContext *avctx)
|
|||||||
|
|
||||||
avctx->sample_fmt = AV_SAMPLE_FMT_S16P;
|
avctx->sample_fmt = AV_SAMPLE_FMT_S16P;
|
||||||
|
|
||||||
avcodec_get_frame_defaults(&c->frame);
|
|
||||||
avctx->coded_frame = &c->frame;
|
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -98,6 +95,7 @@ static int adx_decode(ADXContext *c, int16_t *out, int offset,
|
|||||||
static int adx_decode_frame(AVCodecContext *avctx, void *data,
|
static int adx_decode_frame(AVCodecContext *avctx, void *data,
|
||||||
int *got_frame_ptr, AVPacket *avpkt)
|
int *got_frame_ptr, AVPacket *avpkt)
|
||||||
{
|
{
|
||||||
|
AVFrame *frame = data;
|
||||||
int buf_size = avpkt->size;
|
int buf_size = avpkt->size;
|
||||||
ADXContext *c = avctx->priv_data;
|
ADXContext *c = avctx->priv_data;
|
||||||
int16_t **samples;
|
int16_t **samples;
|
||||||
@ -143,12 +141,12 @@ static int adx_decode_frame(AVCodecContext *avctx, void *data,
|
|||||||
}
|
}
|
||||||
|
|
||||||
/* get output buffer */
|
/* get output buffer */
|
||||||
c->frame.nb_samples = num_blocks * BLOCK_SAMPLES;
|
frame->nb_samples = num_blocks * BLOCK_SAMPLES;
|
||||||
if ((ret = ff_get_buffer(avctx, &c->frame)) < 0) {
|
if ((ret = ff_get_buffer(avctx, frame)) < 0) {
|
||||||
av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n");
|
av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n");
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
samples = (int16_t **)c->frame.extended_data;
|
samples = (int16_t **)frame->extended_data;
|
||||||
samples_offset = 0;
|
samples_offset = 0;
|
||||||
|
|
||||||
while (num_blocks--) {
|
while (num_blocks--) {
|
||||||
@ -165,7 +163,6 @@ static int adx_decode_frame(AVCodecContext *avctx, void *data,
|
|||||||
}
|
}
|
||||||
|
|
||||||
*got_frame_ptr = 1;
|
*got_frame_ptr = 1;
|
||||||
*(AVFrame *)data = c->frame;
|
|
||||||
|
|
||||||
return buf - avpkt->data;
|
return buf - avpkt->data;
|
||||||
}
|
}
|
||||||
|
@ -107,6 +107,14 @@ static int adx_encode_header(AVCodecContext *avctx, uint8_t *buf, int bufsize)
|
|||||||
return HEADER_SIZE;
|
return HEADER_SIZE;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#if FF_API_OLD_ENCODE_AUDIO
|
||||||
|
static av_cold int adx_encode_close(AVCodecContext *avctx)
|
||||||
|
{
|
||||||
|
av_freep(&avctx->coded_frame);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
static av_cold int adx_encode_init(AVCodecContext *avctx)
|
static av_cold int adx_encode_init(AVCodecContext *avctx)
|
||||||
{
|
{
|
||||||
ADXContext *c = avctx->priv_data;
|
ADXContext *c = avctx->priv_data;
|
||||||
@ -118,8 +126,8 @@ static av_cold int adx_encode_init(AVCodecContext *avctx)
|
|||||||
avctx->frame_size = BLOCK_SAMPLES;
|
avctx->frame_size = BLOCK_SAMPLES;
|
||||||
|
|
||||||
#if FF_API_OLD_ENCODE_AUDIO
|
#if FF_API_OLD_ENCODE_AUDIO
|
||||||
avcodec_get_frame_defaults(&c->frame);
|
if (!(avctx->coded_frame = avcodec_alloc_frame()))
|
||||||
avctx->coded_frame = &c->frame;
|
return AVERROR(ENOMEM);
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
/* the cutoff can be adjusted, but this seems to work pretty well */
|
/* the cutoff can be adjusted, but this seems to work pretty well */
|
||||||
@ -167,6 +175,9 @@ AVCodec ff_adpcm_adx_encoder = {
|
|||||||
.id = AV_CODEC_ID_ADPCM_ADX,
|
.id = AV_CODEC_ID_ADPCM_ADX,
|
||||||
.priv_data_size = sizeof(ADXContext),
|
.priv_data_size = sizeof(ADXContext),
|
||||||
.init = adx_encode_init,
|
.init = adx_encode_init,
|
||||||
|
#if FF_API_OLD_ENCODE_AUDIO
|
||||||
|
.close = adx_encode_close,
|
||||||
|
#endif
|
||||||
.encode2 = adx_encode_frame,
|
.encode2 = adx_encode_frame,
|
||||||
.sample_fmts = (const enum AVSampleFormat[]) { AV_SAMPLE_FMT_S16,
|
.sample_fmts = (const enum AVSampleFormat[]) { AV_SAMPLE_FMT_S16,
|
||||||
AV_SAMPLE_FMT_NONE },
|
AV_SAMPLE_FMT_NONE },
|
||||||
|
@ -58,7 +58,6 @@
|
|||||||
|
|
||||||
typedef struct {
|
typedef struct {
|
||||||
AVCodecContext *avctx;
|
AVCodecContext *avctx;
|
||||||
AVFrame frame;
|
|
||||||
GetBitContext gb;
|
GetBitContext gb;
|
||||||
int channels;
|
int channels;
|
||||||
|
|
||||||
@ -254,7 +253,7 @@ static void append_extra_bits(int32_t *buffer[2], int32_t *extra_bits_buffer[2],
|
|||||||
buffer[ch][i] = (buffer[ch][i] << extra_bits) | extra_bits_buffer[ch][i];
|
buffer[ch][i] = (buffer[ch][i] << extra_bits) | extra_bits_buffer[ch][i];
|
||||||
}
|
}
|
||||||
|
|
||||||
static int decode_element(AVCodecContext *avctx, void *data, int ch_index,
|
static int decode_element(AVCodecContext *avctx, AVFrame *frame, int ch_index,
|
||||||
int channels)
|
int channels)
|
||||||
{
|
{
|
||||||
ALACContext *alac = avctx->priv_data;
|
ALACContext *alac = avctx->priv_data;
|
||||||
@ -289,8 +288,8 @@ static int decode_element(AVCodecContext *avctx, void *data, int ch_index,
|
|||||||
}
|
}
|
||||||
if (!alac->nb_samples) {
|
if (!alac->nb_samples) {
|
||||||
/* get output buffer */
|
/* get output buffer */
|
||||||
alac->frame.nb_samples = output_samples;
|
frame->nb_samples = output_samples;
|
||||||
if ((ret = ff_get_buffer(avctx, &alac->frame)) < 0) {
|
if ((ret = ff_get_buffer(avctx, frame)) < 0) {
|
||||||
av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n");
|
av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n");
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
@ -302,7 +301,7 @@ static int decode_element(AVCodecContext *avctx, void *data, int ch_index,
|
|||||||
alac->nb_samples = output_samples;
|
alac->nb_samples = output_samples;
|
||||||
if (alac->direct_output) {
|
if (alac->direct_output) {
|
||||||
for (ch = 0; ch < channels; ch++)
|
for (ch = 0; ch < channels; ch++)
|
||||||
alac->output_samples_buffer[ch] = (int32_t *)alac->frame.extended_data[ch_index + ch];
|
alac->output_samples_buffer[ch] = (int32_t *)frame->extended_data[ch_index + ch];
|
||||||
}
|
}
|
||||||
|
|
||||||
if (is_compressed) {
|
if (is_compressed) {
|
||||||
@ -390,7 +389,7 @@ static int decode_element(AVCodecContext *avctx, void *data, int ch_index,
|
|||||||
switch(alac->sample_size) {
|
switch(alac->sample_size) {
|
||||||
case 16: {
|
case 16: {
|
||||||
for (ch = 0; ch < channels; ch++) {
|
for (ch = 0; ch < channels; ch++) {
|
||||||
int16_t *outbuffer = (int16_t *)alac->frame.extended_data[ch_index + ch];
|
int16_t *outbuffer = (int16_t *)frame->extended_data[ch_index + ch];
|
||||||
for (i = 0; i < alac->nb_samples; i++)
|
for (i = 0; i < alac->nb_samples; i++)
|
||||||
*outbuffer++ = alac->output_samples_buffer[ch][i];
|
*outbuffer++ = alac->output_samples_buffer[ch][i];
|
||||||
}}
|
}}
|
||||||
@ -405,7 +404,7 @@ static int decode_element(AVCodecContext *avctx, void *data, int ch_index,
|
|||||||
}else{
|
}else{
|
||||||
switch(alac->sample_size) {
|
switch(alac->sample_size) {
|
||||||
case 16: {
|
case 16: {
|
||||||
int16_t *outbuffer = ((int16_t *)alac->frame.extended_data[0]) + ch_index;
|
int16_t *outbuffer = ((int16_t *)frame->extended_data[0]) + ch_index;
|
||||||
for (i = 0; i < alac->nb_samples; i++) {
|
for (i = 0; i < alac->nb_samples; i++) {
|
||||||
for (ch = 0; ch < channels; ch++)
|
for (ch = 0; ch < channels; ch++)
|
||||||
*outbuffer++ = alac->output_samples_buffer[ch][i];
|
*outbuffer++ = alac->output_samples_buffer[ch][i];
|
||||||
@ -414,7 +413,7 @@ static int decode_element(AVCodecContext *avctx, void *data, int ch_index,
|
|||||||
}
|
}
|
||||||
break;
|
break;
|
||||||
case 24: {
|
case 24: {
|
||||||
int32_t *outbuffer = ((int32_t *)alac->frame.extended_data[0]) + ch_index;
|
int32_t *outbuffer = ((int32_t *)frame->extended_data[0]) + ch_index;
|
||||||
for (i = 0; i < alac->nb_samples; i++) {
|
for (i = 0; i < alac->nb_samples; i++) {
|
||||||
for (ch = 0; ch < channels; ch++)
|
for (ch = 0; ch < channels; ch++)
|
||||||
*outbuffer++ = alac->output_samples_buffer[ch][i] << 8;
|
*outbuffer++ = alac->output_samples_buffer[ch][i] << 8;
|
||||||
@ -423,7 +422,7 @@ static int decode_element(AVCodecContext *avctx, void *data, int ch_index,
|
|||||||
}
|
}
|
||||||
break;
|
break;
|
||||||
case 32: {
|
case 32: {
|
||||||
int32_t *outbuffer = ((int32_t *)alac->frame.extended_data[0]) + ch_index;
|
int32_t *outbuffer = ((int32_t *)frame->extended_data[0]) + ch_index;
|
||||||
for (i = 0; i < alac->nb_samples; i++) {
|
for (i = 0; i < alac->nb_samples; i++) {
|
||||||
for (ch = 0; ch < channels; ch++)
|
for (ch = 0; ch < channels; ch++)
|
||||||
*outbuffer++ = alac->output_samples_buffer[ch][i];
|
*outbuffer++ = alac->output_samples_buffer[ch][i];
|
||||||
@ -441,6 +440,7 @@ static int alac_decode_frame(AVCodecContext *avctx, void *data,
|
|||||||
int *got_frame_ptr, AVPacket *avpkt)
|
int *got_frame_ptr, AVPacket *avpkt)
|
||||||
{
|
{
|
||||||
ALACContext *alac = avctx->priv_data;
|
ALACContext *alac = avctx->priv_data;
|
||||||
|
AVFrame *frame = data;
|
||||||
enum AlacRawDataBlockType element;
|
enum AlacRawDataBlockType element;
|
||||||
int channels;
|
int channels;
|
||||||
int ch, ret, got_end;
|
int ch, ret, got_end;
|
||||||
@ -469,7 +469,7 @@ static int alac_decode_frame(AVCodecContext *avctx, void *data,
|
|||||||
return AVERROR_INVALIDDATA;
|
return AVERROR_INVALIDDATA;
|
||||||
}
|
}
|
||||||
|
|
||||||
ret = decode_element(avctx, data,
|
ret = decode_element(avctx, frame,
|
||||||
ff_alac_channel_layout_offsets[alac->channels - 1][ch],
|
ff_alac_channel_layout_offsets[alac->channels - 1][ch],
|
||||||
channels);
|
channels);
|
||||||
if (ret < 0 && get_bits_left(&alac->gb))
|
if (ret < 0 && get_bits_left(&alac->gb))
|
||||||
@ -488,7 +488,6 @@ static int alac_decode_frame(AVCodecContext *avctx, void *data,
|
|||||||
}
|
}
|
||||||
|
|
||||||
*got_frame_ptr = 1;
|
*got_frame_ptr = 1;
|
||||||
*(AVFrame *)data = alac->frame;
|
|
||||||
|
|
||||||
return avpkt->size;
|
return avpkt->size;
|
||||||
}
|
}
|
||||||
@ -616,9 +615,6 @@ static av_cold int alac_decode_init(AVCodecContext * avctx)
|
|||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
avcodec_get_frame_defaults(&alac->frame);
|
|
||||||
avctx->coded_frame = &alac->frame;
|
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -192,7 +192,6 @@ typedef struct {
|
|||||||
|
|
||||||
typedef struct {
|
typedef struct {
|
||||||
AVCodecContext *avctx;
|
AVCodecContext *avctx;
|
||||||
AVFrame frame;
|
|
||||||
ALSSpecificConfig sconf;
|
ALSSpecificConfig sconf;
|
||||||
GetBitContext gb;
|
GetBitContext gb;
|
||||||
DSPContext dsp;
|
DSPContext dsp;
|
||||||
@ -1450,6 +1449,7 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *got_frame_ptr,
|
|||||||
AVPacket *avpkt)
|
AVPacket *avpkt)
|
||||||
{
|
{
|
||||||
ALSDecContext *ctx = avctx->priv_data;
|
ALSDecContext *ctx = avctx->priv_data;
|
||||||
|
AVFrame *frame = data;
|
||||||
ALSSpecificConfig *sconf = &ctx->sconf;
|
ALSSpecificConfig *sconf = &ctx->sconf;
|
||||||
const uint8_t *buffer = avpkt->data;
|
const uint8_t *buffer = avpkt->data;
|
||||||
int buffer_size = avpkt->size;
|
int buffer_size = avpkt->size;
|
||||||
@ -1479,8 +1479,8 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *got_frame_ptr,
|
|||||||
ctx->frame_id++;
|
ctx->frame_id++;
|
||||||
|
|
||||||
/* get output buffer */
|
/* get output buffer */
|
||||||
ctx->frame.nb_samples = ctx->cur_frame_length;
|
frame->nb_samples = ctx->cur_frame_length;
|
||||||
if ((ret = ff_get_buffer(avctx, &ctx->frame)) < 0) {
|
if ((ret = ff_get_buffer(avctx, frame)) < 0) {
|
||||||
av_log(avctx, AV_LOG_ERROR, "get_buffer() failed.\n");
|
av_log(avctx, AV_LOG_ERROR, "get_buffer() failed.\n");
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
@ -1488,7 +1488,7 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *got_frame_ptr,
|
|||||||
// transform decoded frame into output format
|
// transform decoded frame into output format
|
||||||
#define INTERLEAVE_OUTPUT(bps) \
|
#define INTERLEAVE_OUTPUT(bps) \
|
||||||
{ \
|
{ \
|
||||||
int##bps##_t *dest = (int##bps##_t*)ctx->frame.data[0]; \
|
int##bps##_t *dest = (int##bps##_t*)frame->data[0]; \
|
||||||
shift = bps - ctx->avctx->bits_per_raw_sample; \
|
shift = bps - ctx->avctx->bits_per_raw_sample; \
|
||||||
if (!ctx->cs_switch) { \
|
if (!ctx->cs_switch) { \
|
||||||
for (sample = 0; sample < ctx->cur_frame_length; sample++) \
|
for (sample = 0; sample < ctx->cur_frame_length; sample++) \
|
||||||
@ -1512,7 +1512,7 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *got_frame_ptr,
|
|||||||
int swap = HAVE_BIGENDIAN != sconf->msb_first;
|
int swap = HAVE_BIGENDIAN != sconf->msb_first;
|
||||||
|
|
||||||
if (ctx->avctx->bits_per_raw_sample == 24) {
|
if (ctx->avctx->bits_per_raw_sample == 24) {
|
||||||
int32_t *src = (int32_t *)ctx->frame.data[0];
|
int32_t *src = (int32_t *)frame->data[0];
|
||||||
|
|
||||||
for (sample = 0;
|
for (sample = 0;
|
||||||
sample < ctx->cur_frame_length * avctx->channels;
|
sample < ctx->cur_frame_length * avctx->channels;
|
||||||
@ -1533,7 +1533,7 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *got_frame_ptr,
|
|||||||
|
|
||||||
if (swap) {
|
if (swap) {
|
||||||
if (ctx->avctx->bits_per_raw_sample <= 16) {
|
if (ctx->avctx->bits_per_raw_sample <= 16) {
|
||||||
int16_t *src = (int16_t*) ctx->frame.data[0];
|
int16_t *src = (int16_t*) frame->data[0];
|
||||||
int16_t *dest = (int16_t*) ctx->crc_buffer;
|
int16_t *dest = (int16_t*) ctx->crc_buffer;
|
||||||
for (sample = 0;
|
for (sample = 0;
|
||||||
sample < ctx->cur_frame_length * avctx->channels;
|
sample < ctx->cur_frame_length * avctx->channels;
|
||||||
@ -1541,12 +1541,12 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *got_frame_ptr,
|
|||||||
*dest++ = av_bswap16(src[sample]);
|
*dest++ = av_bswap16(src[sample]);
|
||||||
} else {
|
} else {
|
||||||
ctx->dsp.bswap_buf((uint32_t*)ctx->crc_buffer,
|
ctx->dsp.bswap_buf((uint32_t*)ctx->crc_buffer,
|
||||||
(uint32_t *)ctx->frame.data[0],
|
(uint32_t *)frame->data[0],
|
||||||
ctx->cur_frame_length * avctx->channels);
|
ctx->cur_frame_length * avctx->channels);
|
||||||
}
|
}
|
||||||
crc_source = ctx->crc_buffer;
|
crc_source = ctx->crc_buffer;
|
||||||
} else {
|
} else {
|
||||||
crc_source = ctx->frame.data[0];
|
crc_source = frame->data[0];
|
||||||
}
|
}
|
||||||
|
|
||||||
ctx->crc = av_crc(ctx->crc_table, ctx->crc, crc_source,
|
ctx->crc = av_crc(ctx->crc_table, ctx->crc, crc_source,
|
||||||
@ -1563,8 +1563,6 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *got_frame_ptr,
|
|||||||
}
|
}
|
||||||
|
|
||||||
*got_frame_ptr = 1;
|
*got_frame_ptr = 1;
|
||||||
*(AVFrame *)data = ctx->frame;
|
|
||||||
|
|
||||||
|
|
||||||
bytes_read = invalid_frame ? buffer_size :
|
bytes_read = invalid_frame ? buffer_size :
|
||||||
(get_bits_count(&ctx->gb) + 7) >> 3;
|
(get_bits_count(&ctx->gb) + 7) >> 3;
|
||||||
@ -1761,9 +1759,6 @@ static av_cold int decode_init(AVCodecContext *avctx)
|
|||||||
|
|
||||||
ff_dsputil_init(&ctx->dsp, avctx);
|
ff_dsputil_init(&ctx->dsp, avctx);
|
||||||
|
|
||||||
avcodec_get_frame_defaults(&ctx->frame);
|
|
||||||
avctx->coded_frame = &ctx->frame;
|
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -98,7 +98,6 @@
|
|||||||
#define AMR_AGC_ALPHA 0.9
|
#define AMR_AGC_ALPHA 0.9
|
||||||
|
|
||||||
typedef struct AMRContext {
|
typedef struct AMRContext {
|
||||||
AVFrame avframe; ///< AVFrame for decoded samples
|
|
||||||
AMRNBFrame frame; ///< decoded AMR parameters (lsf coefficients, codebook indexes, etc)
|
AMRNBFrame frame; ///< decoded AMR parameters (lsf coefficients, codebook indexes, etc)
|
||||||
uint8_t bad_frame_indicator; ///< bad frame ? 1 : 0
|
uint8_t bad_frame_indicator; ///< bad frame ? 1 : 0
|
||||||
enum Mode cur_frame_mode;
|
enum Mode cur_frame_mode;
|
||||||
@ -185,9 +184,6 @@ static av_cold int amrnb_decode_init(AVCodecContext *avctx)
|
|||||||
for (i = 0; i < 4; i++)
|
for (i = 0; i < 4; i++)
|
||||||
p->prediction_error[i] = MIN_ENERGY;
|
p->prediction_error[i] = MIN_ENERGY;
|
||||||
|
|
||||||
avcodec_get_frame_defaults(&p->avframe);
|
|
||||||
avctx->coded_frame = &p->avframe;
|
|
||||||
|
|
||||||
ff_acelp_filter_init(&p->acelpf_ctx);
|
ff_acelp_filter_init(&p->acelpf_ctx);
|
||||||
ff_acelp_vectors_init(&p->acelpv_ctx);
|
ff_acelp_vectors_init(&p->acelpv_ctx);
|
||||||
ff_celp_filter_init(&p->celpf_ctx);
|
ff_celp_filter_init(&p->celpf_ctx);
|
||||||
@ -954,6 +950,7 @@ static int amrnb_decode_frame(AVCodecContext *avctx, void *data,
|
|||||||
{
|
{
|
||||||
|
|
||||||
AMRContext *p = avctx->priv_data; // pointer to private data
|
AMRContext *p = avctx->priv_data; // pointer to private data
|
||||||
|
AVFrame *frame = data;
|
||||||
const uint8_t *buf = avpkt->data;
|
const uint8_t *buf = avpkt->data;
|
||||||
int buf_size = avpkt->size;
|
int buf_size = avpkt->size;
|
||||||
float *buf_out; // pointer to the output data buffer
|
float *buf_out; // pointer to the output data buffer
|
||||||
@ -965,12 +962,12 @@ static int amrnb_decode_frame(AVCodecContext *avctx, void *data,
|
|||||||
const float *synth_fixed_vector; // pointer to the fixed vector that synthesis should use
|
const float *synth_fixed_vector; // pointer to the fixed vector that synthesis should use
|
||||||
|
|
||||||
/* get output buffer */
|
/* get output buffer */
|
||||||
p->avframe.nb_samples = AMR_BLOCK_SIZE;
|
frame->nb_samples = AMR_BLOCK_SIZE;
|
||||||
if ((ret = ff_get_buffer(avctx, &p->avframe)) < 0) {
|
if ((ret = ff_get_buffer(avctx, frame)) < 0) {
|
||||||
av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n");
|
av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n");
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
buf_out = (float *)p->avframe.data[0];
|
buf_out = (float *)frame->data[0];
|
||||||
|
|
||||||
p->cur_frame_mode = unpack_bitstream(p, buf, buf_size);
|
p->cur_frame_mode = unpack_bitstream(p, buf, buf_size);
|
||||||
if (p->cur_frame_mode == NO_DATA) {
|
if (p->cur_frame_mode == NO_DATA) {
|
||||||
@ -1079,7 +1076,6 @@ static int amrnb_decode_frame(AVCodecContext *avctx, void *data,
|
|||||||
0.84, 0.16, LP_FILTER_ORDER);
|
0.84, 0.16, LP_FILTER_ORDER);
|
||||||
|
|
||||||
*got_frame_ptr = 1;
|
*got_frame_ptr = 1;
|
||||||
*(AVFrame *)data = p->avframe;
|
|
||||||
|
|
||||||
/* return the amount of bytes consumed if everything was OK */
|
/* return the amount of bytes consumed if everything was OK */
|
||||||
return frame_sizes_nb[p->cur_frame_mode] + 1; // +7 for rounding and +8 for TOC
|
return frame_sizes_nb[p->cur_frame_mode] + 1; // +7 for rounding and +8 for TOC
|
||||||
|
Loading…
Reference in New Issue
Block a user