mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-12-23 12:43:46 +02:00
20f9727018
Up until now, codec.h contains both public and private parts of AVCodec. This exposes the internals of AVCodec to users and leads them into the temptation of actually using them and forces us to forward-declare structures and types that users can't use at all. This commit changes this by adding a new structure FFCodec to codec_internal.h that extends AVCodec, i.e. contains the public AVCodec as first member; the private fields of AVCodec are moved to this structure, leaving codec.h clean. Reviewed-by: Anton Khirnov <anton@khirnov.net> Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@outlook.com>
405 lines
13 KiB
C
405 lines
13 KiB
C
/*
|
|
* PNM image format
|
|
* Copyright (c) 2002, 2003 Fabrice Bellard
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "config_components.h"
|
|
|
|
#include "avcodec.h"
|
|
#include "codec_internal.h"
|
|
#include "internal.h"
|
|
#include "put_bits.h"
|
|
#include "pnm.h"
|
|
|
|
static void samplecpy(uint8_t *dst, const uint8_t *src, int n, int maxval)
|
|
{
|
|
if (maxval <= 255) {
|
|
memcpy(dst, src, n);
|
|
} else {
|
|
int i;
|
|
for (i=0; i<n/2; i++) {
|
|
((uint16_t *)dst)[i] = AV_RB16(src+2*i);
|
|
}
|
|
}
|
|
}
|
|
|
|
static int pnm_decode_frame(AVCodecContext *avctx, void *data,
|
|
int *got_frame, AVPacket *avpkt)
|
|
{
|
|
const uint8_t *buf = avpkt->data;
|
|
int buf_size = avpkt->size;
|
|
PNMContext * const s = avctx->priv_data;
|
|
AVFrame * const p = data;
|
|
int i, j, k, n, linesize, h, upgrade = 0, is_mono = 0;
|
|
unsigned char *ptr;
|
|
int components, sample_len, ret;
|
|
float scale;
|
|
|
|
s->bytestream_start =
|
|
s->bytestream = (uint8_t *)buf;
|
|
s->bytestream_end = (uint8_t *)buf + buf_size;
|
|
|
|
if ((ret = ff_pnm_decode_header(avctx, s)) < 0)
|
|
return ret;
|
|
|
|
if ((ret = ff_get_buffer(avctx, p, 0)) < 0)
|
|
return ret;
|
|
p->pict_type = AV_PICTURE_TYPE_I;
|
|
p->key_frame = 1;
|
|
avctx->bits_per_raw_sample = av_log2(s->maxval) + 1;
|
|
|
|
switch (avctx->pix_fmt) {
|
|
default:
|
|
return AVERROR(EINVAL);
|
|
case AV_PIX_FMT_RGBA64:
|
|
n = avctx->width * 8;
|
|
components=4;
|
|
sample_len=16;
|
|
if (s->maxval < 65535)
|
|
upgrade = 2;
|
|
goto do_read;
|
|
case AV_PIX_FMT_RGB48:
|
|
n = avctx->width * 6;
|
|
components=3;
|
|
sample_len=16;
|
|
if (s->maxval < 65535)
|
|
upgrade = 2;
|
|
goto do_read;
|
|
case AV_PIX_FMT_RGBA:
|
|
n = avctx->width * 4;
|
|
components=4;
|
|
sample_len=8;
|
|
goto do_read;
|
|
case AV_PIX_FMT_RGB24:
|
|
n = avctx->width * 3;
|
|
components=3;
|
|
sample_len=8;
|
|
if (s->maxval < 255)
|
|
upgrade = 1;
|
|
goto do_read;
|
|
case AV_PIX_FMT_GRAY8:
|
|
n = avctx->width;
|
|
components=1;
|
|
sample_len=8;
|
|
if (s->maxval < 255)
|
|
upgrade = 1;
|
|
goto do_read;
|
|
case AV_PIX_FMT_GRAY8A:
|
|
n = avctx->width * 2;
|
|
components=2;
|
|
sample_len=8;
|
|
goto do_read;
|
|
case AV_PIX_FMT_GRAY16:
|
|
n = avctx->width * 2;
|
|
components=1;
|
|
sample_len=16;
|
|
if (s->maxval < 65535)
|
|
upgrade = 2;
|
|
goto do_read;
|
|
case AV_PIX_FMT_YA16:
|
|
n = avctx->width * 4;
|
|
components=2;
|
|
sample_len=16;
|
|
if (s->maxval < 65535)
|
|
upgrade = 2;
|
|
goto do_read;
|
|
case AV_PIX_FMT_MONOWHITE:
|
|
case AV_PIX_FMT_MONOBLACK:
|
|
n = (avctx->width + 7) >> 3;
|
|
components=1;
|
|
sample_len=1;
|
|
is_mono = 1;
|
|
do_read:
|
|
ptr = p->data[0];
|
|
linesize = p->linesize[0];
|
|
if (n * avctx->height > s->bytestream_end - s->bytestream)
|
|
return AVERROR_INVALIDDATA;
|
|
if(s->type < 4 || (is_mono && s->type==7)){
|
|
for (i=0; i<avctx->height; i++) {
|
|
PutBitContext pb;
|
|
init_put_bits(&pb, ptr, linesize);
|
|
for(j=0; j<avctx->width * components; j++){
|
|
unsigned int c=0;
|
|
unsigned v=0;
|
|
if(s->type < 4)
|
|
while(s->bytestream < s->bytestream_end && (*s->bytestream < '0' || *s->bytestream > '9' ))
|
|
s->bytestream++;
|
|
if(s->bytestream >= s->bytestream_end)
|
|
return AVERROR_INVALIDDATA;
|
|
if (is_mono) {
|
|
/* read a single digit */
|
|
v = (*s->bytestream++)&1;
|
|
} else {
|
|
/* read a sequence of digits */
|
|
for (k = 0; k < 6 && c <= 9; k += 1) {
|
|
v = 10*v + c;
|
|
c = (*s->bytestream++) - '0';
|
|
}
|
|
if (v > s->maxval) {
|
|
av_log(avctx, AV_LOG_ERROR, "value %d larger than maxval %d\n", v, s->maxval);
|
|
return AVERROR_INVALIDDATA;
|
|
}
|
|
}
|
|
if (sample_len == 16) {
|
|
((uint16_t*)ptr)[j] = (((1<<sample_len)-1)*v + (s->maxval>>1))/s->maxval;
|
|
} else
|
|
put_bits(&pb, sample_len, (((1<<sample_len)-1)*v + (s->maxval>>1))/s->maxval);
|
|
}
|
|
if (sample_len != 16)
|
|
flush_put_bits(&pb);
|
|
ptr+= linesize;
|
|
}
|
|
}else{
|
|
for (i = 0; i < avctx->height; i++) {
|
|
if (!upgrade)
|
|
samplecpy(ptr, s->bytestream, n, s->maxval);
|
|
else if (upgrade == 1) {
|
|
unsigned int j, f = (255 * 128 + s->maxval / 2) / s->maxval;
|
|
for (j = 0; j < n; j++)
|
|
ptr[j] = (s->bytestream[j] * f + 64) >> 7;
|
|
} else if (upgrade == 2) {
|
|
unsigned int j, v, f = (65535 * 32768 + s->maxval / 2) / s->maxval;
|
|
for (j = 0; j < n / 2; j++) {
|
|
v = AV_RB16(s->bytestream + 2*j);
|
|
((uint16_t *)ptr)[j] = (v * f + 16384) >> 15;
|
|
}
|
|
}
|
|
s->bytestream += n;
|
|
ptr += linesize;
|
|
}
|
|
}
|
|
break;
|
|
case AV_PIX_FMT_YUV420P:
|
|
case AV_PIX_FMT_YUV420P9:
|
|
case AV_PIX_FMT_YUV420P10:
|
|
{
|
|
unsigned char *ptr1, *ptr2;
|
|
|
|
n = avctx->width;
|
|
ptr = p->data[0];
|
|
linesize = p->linesize[0];
|
|
if (s->maxval >= 256)
|
|
n *= 2;
|
|
if (n * avctx->height * 3 / 2 > s->bytestream_end - s->bytestream)
|
|
return AVERROR_INVALIDDATA;
|
|
for (i = 0; i < avctx->height; i++) {
|
|
samplecpy(ptr, s->bytestream, n, s->maxval);
|
|
s->bytestream += n;
|
|
ptr += linesize;
|
|
}
|
|
ptr1 = p->data[1];
|
|
ptr2 = p->data[2];
|
|
n >>= 1;
|
|
h = avctx->height >> 1;
|
|
for (i = 0; i < h; i++) {
|
|
samplecpy(ptr1, s->bytestream, n, s->maxval);
|
|
s->bytestream += n;
|
|
samplecpy(ptr2, s->bytestream, n, s->maxval);
|
|
s->bytestream += n;
|
|
ptr1 += p->linesize[1];
|
|
ptr2 += p->linesize[2];
|
|
}
|
|
}
|
|
break;
|
|
case AV_PIX_FMT_YUV420P16:
|
|
{
|
|
uint16_t *ptr1, *ptr2;
|
|
const int f = (65535 * 32768 + s->maxval / 2) / s->maxval;
|
|
unsigned int j, v;
|
|
|
|
n = avctx->width * 2;
|
|
ptr = p->data[0];
|
|
linesize = p->linesize[0];
|
|
if (n * avctx->height * 3 / 2 > s->bytestream_end - s->bytestream)
|
|
return AVERROR_INVALIDDATA;
|
|
for (i = 0; i < avctx->height; i++) {
|
|
for (j = 0; j < n / 2; j++) {
|
|
v = AV_RB16(s->bytestream + 2*j);
|
|
((uint16_t *)ptr)[j] = (v * f + 16384) >> 15;
|
|
}
|
|
s->bytestream += n;
|
|
ptr += linesize;
|
|
}
|
|
ptr1 = (uint16_t*)p->data[1];
|
|
ptr2 = (uint16_t*)p->data[2];
|
|
n >>= 1;
|
|
h = avctx->height >> 1;
|
|
for (i = 0; i < h; i++) {
|
|
for (j = 0; j < n / 2; j++) {
|
|
v = AV_RB16(s->bytestream + 2*j);
|
|
ptr1[j] = (v * f + 16384) >> 15;
|
|
}
|
|
s->bytestream += n;
|
|
|
|
for (j = 0; j < n / 2; j++) {
|
|
v = AV_RB16(s->bytestream + 2*j);
|
|
ptr2[j] = (v * f + 16384) >> 15;
|
|
}
|
|
s->bytestream += n;
|
|
|
|
ptr1 += p->linesize[1] / 2;
|
|
ptr2 += p->linesize[2] / 2;
|
|
}
|
|
}
|
|
break;
|
|
case AV_PIX_FMT_GBRPF32:
|
|
if (avctx->width * avctx->height * 12 > s->bytestream_end - s->bytestream)
|
|
return AVERROR_INVALIDDATA;
|
|
scale = 1.f / s->scale;
|
|
if (s->endian) {
|
|
float *r, *g, *b;
|
|
|
|
r = (float *)p->data[2];
|
|
g = (float *)p->data[0];
|
|
b = (float *)p->data[1];
|
|
for (int i = 0; i < avctx->height; i++) {
|
|
for (int j = 0; j < avctx->width; j++) {
|
|
r[j] = av_int2float(AV_RL32(s->bytestream+0)) * scale;
|
|
g[j] = av_int2float(AV_RL32(s->bytestream+4)) * scale;
|
|
b[j] = av_int2float(AV_RL32(s->bytestream+8)) * scale;
|
|
s->bytestream += 12;
|
|
}
|
|
|
|
r += p->linesize[2] / 4;
|
|
g += p->linesize[0] / 4;
|
|
b += p->linesize[1] / 4;
|
|
}
|
|
} else {
|
|
float *r, *g, *b;
|
|
|
|
r = (float *)p->data[2];
|
|
g = (float *)p->data[0];
|
|
b = (float *)p->data[1];
|
|
for (int i = 0; i < avctx->height; i++) {
|
|
for (int j = 0; j < avctx->width; j++) {
|
|
r[j] = av_int2float(AV_RB32(s->bytestream+0)) * scale;
|
|
g[j] = av_int2float(AV_RB32(s->bytestream+4)) * scale;
|
|
b[j] = av_int2float(AV_RB32(s->bytestream+8)) * scale;
|
|
s->bytestream += 12;
|
|
}
|
|
|
|
r += p->linesize[2] / 4;
|
|
g += p->linesize[0] / 4;
|
|
b += p->linesize[1] / 4;
|
|
}
|
|
}
|
|
break;
|
|
case AV_PIX_FMT_GRAYF32:
|
|
if (avctx->width * avctx->height * 4 > s->bytestream_end - s->bytestream)
|
|
return AVERROR_INVALIDDATA;
|
|
scale = 1.f / s->scale;
|
|
if (s->endian) {
|
|
float *g = (float *)p->data[0];
|
|
for (int i = 0; i < avctx->height; i++) {
|
|
for (int j = 0; j < avctx->width; j++) {
|
|
g[j] = av_int2float(AV_RL32(s->bytestream)) * scale;
|
|
s->bytestream += 4;
|
|
}
|
|
g += p->linesize[0] / 4;
|
|
}
|
|
} else {
|
|
float *g = (float *)p->data[0];
|
|
for (int i = 0; i < avctx->height; i++) {
|
|
for (int j = 0; j < avctx->width; j++) {
|
|
g[j] = av_int2float(AV_RB32(s->bytestream)) * scale;
|
|
s->bytestream += 4;
|
|
}
|
|
g += p->linesize[0] / 4;
|
|
}
|
|
}
|
|
break;
|
|
}
|
|
*got_frame = 1;
|
|
|
|
return s->bytestream - s->bytestream_start;
|
|
}
|
|
|
|
|
|
#if CONFIG_PGM_DECODER
|
|
const FFCodec ff_pgm_decoder = {
|
|
.p.name = "pgm",
|
|
.p.long_name = NULL_IF_CONFIG_SMALL("PGM (Portable GrayMap) image"),
|
|
.p.type = AVMEDIA_TYPE_VIDEO,
|
|
.p.id = AV_CODEC_ID_PGM,
|
|
.p.capabilities = AV_CODEC_CAP_DR1,
|
|
.priv_data_size = sizeof(PNMContext),
|
|
.decode = pnm_decode_frame,
|
|
};
|
|
#endif
|
|
|
|
#if CONFIG_PGMYUV_DECODER
|
|
const FFCodec ff_pgmyuv_decoder = {
|
|
.p.name = "pgmyuv",
|
|
.p.long_name = NULL_IF_CONFIG_SMALL("PGMYUV (Portable GrayMap YUV) image"),
|
|
.p.type = AVMEDIA_TYPE_VIDEO,
|
|
.p.id = AV_CODEC_ID_PGMYUV,
|
|
.p.capabilities = AV_CODEC_CAP_DR1,
|
|
.priv_data_size = sizeof(PNMContext),
|
|
.decode = pnm_decode_frame,
|
|
};
|
|
#endif
|
|
|
|
#if CONFIG_PPM_DECODER
|
|
const FFCodec ff_ppm_decoder = {
|
|
.p.name = "ppm",
|
|
.p.long_name = NULL_IF_CONFIG_SMALL("PPM (Portable PixelMap) image"),
|
|
.p.type = AVMEDIA_TYPE_VIDEO,
|
|
.p.id = AV_CODEC_ID_PPM,
|
|
.p.capabilities = AV_CODEC_CAP_DR1,
|
|
.priv_data_size = sizeof(PNMContext),
|
|
.decode = pnm_decode_frame,
|
|
};
|
|
#endif
|
|
|
|
#if CONFIG_PBM_DECODER
|
|
const FFCodec ff_pbm_decoder = {
|
|
.p.name = "pbm",
|
|
.p.long_name = NULL_IF_CONFIG_SMALL("PBM (Portable BitMap) image"),
|
|
.p.type = AVMEDIA_TYPE_VIDEO,
|
|
.p.id = AV_CODEC_ID_PBM,
|
|
.p.capabilities = AV_CODEC_CAP_DR1,
|
|
.priv_data_size = sizeof(PNMContext),
|
|
.decode = pnm_decode_frame,
|
|
};
|
|
#endif
|
|
|
|
#if CONFIG_PAM_DECODER
|
|
const FFCodec ff_pam_decoder = {
|
|
.p.name = "pam",
|
|
.p.long_name = NULL_IF_CONFIG_SMALL("PAM (Portable AnyMap) image"),
|
|
.p.type = AVMEDIA_TYPE_VIDEO,
|
|
.p.id = AV_CODEC_ID_PAM,
|
|
.p.capabilities = AV_CODEC_CAP_DR1,
|
|
.priv_data_size = sizeof(PNMContext),
|
|
.decode = pnm_decode_frame,
|
|
};
|
|
#endif
|
|
|
|
#if CONFIG_PFM_DECODER
|
|
const FFCodec ff_pfm_decoder = {
|
|
.p.name = "pfm",
|
|
.p.long_name = NULL_IF_CONFIG_SMALL("PFM (Portable FloatMap) image"),
|
|
.p.type = AVMEDIA_TYPE_VIDEO,
|
|
.p.id = AV_CODEC_ID_PFM,
|
|
.p.capabilities = AV_CODEC_CAP_DR1,
|
|
.priv_data_size = sizeof(PNMContext),
|
|
.decode = pnm_decode_frame,
|
|
};
|
|
#endif
|