1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2024-11-26 19:01:44 +02:00
FFmpeg/libavcodec/tiff.c
Niklas Haas 61ffa23c2e avcodec/codec_internal: add cap for ICC profile support
Codecs that can read/write ICC profiles deserve a special capability so
the common logic in encode.c/decode.c can decide whether or not there
needs to be any special handling for ICC profiles. The motivation here
is to be able to use it to decide whether or not an ICC profile needs to
be generated in the encode path, but it might as well get added to
decoders as well for purely informative reasons.

It's not entirely clear to me whether the "thp" and "smvjpeg" variants
of "mjpeg" should have this capability set or not, given that the code
technically supports it but I somehow doubt these files may contain
them. In either case, this cap is purely informative for decoders so it
doesn't matter too much either way.

It's also not entirely clear whether the "amv" encoder should signal ICC
profile support, but again erring on the side of caution, we probably
*shouldn't* be generating (and encoding!) ICC profiles for this type of
media file.

Signed-off-by: Niklas Haas <git@haasn.dev>
2022-07-30 11:42:06 +02:00

2199 lines
75 KiB
C

/*
* Copyright (c) 2006 Konstantin Shishkov
*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
/**
* @file
* TIFF image decoder
* @author Konstantin Shishkov
*/
#include "config.h"
#if CONFIG_ZLIB
#include <zlib.h>
#endif
#if CONFIG_LZMA
#define LZMA_API_STATIC
#include <lzma.h>
#endif
#include "libavutil/attributes.h"
#include "libavutil/avstring.h"
#include "libavutil/error.h"
#include "libavutil/intreadwrite.h"
#include "libavutil/imgutils.h"
#include "libavutil/opt.h"
#include "libavutil/reverse.h"
#include "avcodec.h"
#include "bytestream.h"
#include "codec_internal.h"
#include "faxcompr.h"
#include "internal.h"
#include "lzw.h"
#include "mathops.h"
#include "tiff.h"
#include "tiff_data.h"
#include "mjpegdec.h"
#include "thread.h"
#include "get_bits.h"
typedef struct TiffContext {
AVClass *class;
AVCodecContext *avctx;
GetByteContext gb;
/* JPEG decoding for DNG */
AVCodecContext *avctx_mjpeg; // wrapper context for MJPEG
AVPacket *jpkt; // encoded JPEG tile
AVFrame *jpgframe; // decoded JPEG tile
int get_subimage;
uint16_t get_page;
int get_thumbnail;
enum TiffType tiff_type;
int width, height;
unsigned int bpp, bppcount;
uint32_t palette[256];
int palette_is_set;
int le;
enum TiffCompr compr;
enum TiffPhotometric photometric;
int planar;
int subsampling[2];
int fax_opts;
int predictor;
int fill_order;
uint32_t res[4];
int is_thumbnail;
unsigned last_tag;
int is_bayer;
uint8_t pattern[4];
unsigned black_level;
unsigned white_level;
uint16_t dng_lut[65536];
uint32_t sub_ifd;
uint16_t cur_page;
int strips, rps, sstype;
int sot;
int stripsizesoff, stripsize, stripoff, strippos;
LZWState *lzw;
/* Tile support */
int is_tiled;
int tile_byte_counts_offset, tile_offsets_offset;
int tile_width, tile_length;
int tile_count;
int is_jpeg;
uint8_t *deinvert_buf;
int deinvert_buf_size;
uint8_t *yuv_line;
unsigned int yuv_line_size;
int geotag_count;
TiffGeoTag *geotags;
} TiffContext;
static void tiff_set_type(TiffContext *s, enum TiffType tiff_type) {
if (s->tiff_type < tiff_type) // Prioritize higher-valued entries
s->tiff_type = tiff_type;
}
static void free_geotags(TiffContext *const s)
{
int i;
for (i = 0; i < s->geotag_count; i++) {
if (s->geotags[i].val)
av_freep(&s->geotags[i].val);
}
av_freep(&s->geotags);
s->geotag_count = 0;
}
#define RET_GEOKEY(TYPE, array, element)\
if (key >= TIFF_##TYPE##_KEY_ID_OFFSET &&\
key - TIFF_##TYPE##_KEY_ID_OFFSET < FF_ARRAY_ELEMS(tiff_##array##_name_type_map))\
return tiff_##array##_name_type_map[key - TIFF_##TYPE##_KEY_ID_OFFSET].element;
static const char *get_geokey_name(int key)
{
RET_GEOKEY(VERT, vert, name);
RET_GEOKEY(PROJ, proj, name);
RET_GEOKEY(GEOG, geog, name);
RET_GEOKEY(CONF, conf, name);
return NULL;
}
static int get_geokey_type(int key)
{
RET_GEOKEY(VERT, vert, type);
RET_GEOKEY(PROJ, proj, type);
RET_GEOKEY(GEOG, geog, type);
RET_GEOKEY(CONF, conf, type);
return AVERROR_INVALIDDATA;
}
static int cmp_id_key(const void *id, const void *k)
{
return *(const int*)id - ((const TiffGeoTagKeyName*)k)->key;
}
static const char *search_keyval(const TiffGeoTagKeyName *keys, int n, int id)
{
TiffGeoTagKeyName *r = bsearch(&id, keys, n, sizeof(keys[0]), cmp_id_key);
if(r)
return r->name;
return NULL;
}
static char *get_geokey_val(int key, int val)
{
char *ap;
if (val == TIFF_GEO_KEY_UNDEFINED)
return av_strdup("undefined");
if (val == TIFF_GEO_KEY_USER_DEFINED)
return av_strdup("User-Defined");
#define RET_GEOKEY_VAL(TYPE, array)\
if (val >= TIFF_##TYPE##_OFFSET &&\
val - TIFF_##TYPE##_OFFSET < FF_ARRAY_ELEMS(tiff_##array##_codes))\
return av_strdup(tiff_##array##_codes[val - TIFF_##TYPE##_OFFSET]);
switch (key) {
case TIFF_GT_MODEL_TYPE_GEOKEY:
RET_GEOKEY_VAL(GT_MODEL_TYPE, gt_model_type);
break;
case TIFF_GT_RASTER_TYPE_GEOKEY:
RET_GEOKEY_VAL(GT_RASTER_TYPE, gt_raster_type);
break;
case TIFF_GEOG_LINEAR_UNITS_GEOKEY:
case TIFF_PROJ_LINEAR_UNITS_GEOKEY:
case TIFF_VERTICAL_UNITS_GEOKEY:
RET_GEOKEY_VAL(LINEAR_UNIT, linear_unit);
break;
case TIFF_GEOG_ANGULAR_UNITS_GEOKEY:
case TIFF_GEOG_AZIMUTH_UNITS_GEOKEY:
RET_GEOKEY_VAL(ANGULAR_UNIT, angular_unit);
break;
case TIFF_GEOGRAPHIC_TYPE_GEOKEY:
RET_GEOKEY_VAL(GCS_TYPE, gcs_type);
RET_GEOKEY_VAL(GCSE_TYPE, gcse_type);
break;
case TIFF_GEOG_GEODETIC_DATUM_GEOKEY:
RET_GEOKEY_VAL(GEODETIC_DATUM, geodetic_datum);
RET_GEOKEY_VAL(GEODETIC_DATUM_E, geodetic_datum_e);
break;
case TIFF_GEOG_ELLIPSOID_GEOKEY:
RET_GEOKEY_VAL(ELLIPSOID, ellipsoid);
break;
case TIFF_GEOG_PRIME_MERIDIAN_GEOKEY:
RET_GEOKEY_VAL(PRIME_MERIDIAN, prime_meridian);
break;
case TIFF_PROJECTED_CS_TYPE_GEOKEY:
ap = av_strdup(search_keyval(tiff_proj_cs_type_codes, FF_ARRAY_ELEMS(tiff_proj_cs_type_codes), val));
if(ap) return ap;
break;
case TIFF_PROJECTION_GEOKEY:
ap = av_strdup(search_keyval(tiff_projection_codes, FF_ARRAY_ELEMS(tiff_projection_codes), val));
if(ap) return ap;
break;
case TIFF_PROJ_COORD_TRANS_GEOKEY:
RET_GEOKEY_VAL(COORD_TRANS, coord_trans);
break;
case TIFF_VERTICAL_CS_TYPE_GEOKEY:
RET_GEOKEY_VAL(VERT_CS, vert_cs);
RET_GEOKEY_VAL(ORTHO_VERT_CS, ortho_vert_cs);
break;
}
ap = av_malloc(14);
if (ap)
snprintf(ap, 14, "Unknown-%d", val);
return ap;
}
static char *doubles2str(double *dp, int count, const char *sep)
{
int i;
char *ap, *ap0;
uint64_t component_len;
if (!sep) sep = ", ";
component_len = 24LL + strlen(sep);
if (count >= (INT_MAX - 1)/component_len)
return NULL;
ap = av_malloc(component_len * count + 1);
if (!ap)
return NULL;
ap0 = ap;
ap[0] = '\0';
for (i = 0; i < count; i++) {
unsigned l = snprintf(ap, component_len, "%.15g%s", dp[i], sep);
if(l >= component_len) {
av_free(ap0);
return NULL;
}
ap += l;
}
ap0[strlen(ap0) - strlen(sep)] = '\0';
return ap0;
}
static int add_metadata(int count, int type,
const char *name, const char *sep, TiffContext *s, AVFrame *frame)
{
switch(type) {
case TIFF_DOUBLE: return ff_tadd_doubles_metadata(count, name, sep, &s->gb, s->le, &frame->metadata);
case TIFF_SHORT : return ff_tadd_shorts_metadata(count, name, sep, &s->gb, s->le, 0, &frame->metadata);
case TIFF_STRING: return ff_tadd_string_metadata(count, name, &s->gb, s->le, &frame->metadata);
default : return AVERROR_INVALIDDATA;
};
}
/**
* Map stored raw sensor values into linear reference values (see: DNG Specification - Chapter 5)
*/
static uint16_t av_always_inline dng_process_color16(uint16_t value,
const uint16_t *lut,
uint16_t black_level,
float scale_factor)
{
float value_norm;
// Lookup table lookup
if (lut)
value = lut[value];
// Black level subtraction
value = av_clip_uint16_c((unsigned)value - black_level);
// Color scaling
value_norm = (float)value * scale_factor;
value = av_clip_uint16_c(value_norm * 65535);
return value;
}
static uint16_t av_always_inline dng_process_color8(uint16_t value,
const uint16_t *lut,
uint16_t black_level,
float scale_factor)
{
return dng_process_color16(value, lut, black_level, scale_factor) >> 8;
}
static void av_always_inline dng_blit(TiffContext *s, uint8_t *dst, int dst_stride,
const uint8_t *src, int src_stride, int width, int height,
int is_single_comp, int is_u16)
{
int line, col;
float scale_factor;
scale_factor = 1.0f / (s->white_level - s->black_level);
if (is_single_comp) {
if (!is_u16)
return; /* <= 8bpp unsupported */
/* Image is double the width and half the height we need, each row comprises 2 rows of the output
(split vertically in the middle). */
for (line = 0; line < height / 2; line++) {
uint16_t *dst_u16 = (uint16_t *)dst;
uint16_t *src_u16 = (uint16_t *)src;
/* Blit first half of input row row to initial row of output */
for (col = 0; col < width; col++)
*dst_u16++ = dng_process_color16(*src_u16++, s->dng_lut, s->black_level, scale_factor);
/* Advance the destination pointer by a row (source pointer remains in the same place) */
dst += dst_stride * sizeof(uint16_t);
dst_u16 = (uint16_t *)dst;
/* Blit second half of input row row to next row of output */
for (col = 0; col < width; col++)
*dst_u16++ = dng_process_color16(*src_u16++, s->dng_lut, s->black_level, scale_factor);
dst += dst_stride * sizeof(uint16_t);
src += src_stride * sizeof(uint16_t);
}
} else {
/* Input and output image are the same size and the MJpeg decoder has done per-component
deinterleaving, so blitting here is straightforward. */
if (is_u16) {
for (line = 0; line < height; line++) {
uint16_t *dst_u16 = (uint16_t *)dst;
uint16_t *src_u16 = (uint16_t *)src;
for (col = 0; col < width; col++)
*dst_u16++ = dng_process_color16(*src_u16++, s->dng_lut, s->black_level, scale_factor);
dst += dst_stride * sizeof(uint16_t);
src += src_stride * sizeof(uint16_t);
}
} else {
for (line = 0; line < height; line++) {
uint8_t *dst_u8 = dst;
const uint8_t *src_u8 = src;
for (col = 0; col < width; col++)
*dst_u8++ = dng_process_color8(*src_u8++, s->dng_lut, s->black_level, scale_factor);
dst += dst_stride;
src += src_stride;
}
}
}
}
static void av_always_inline horizontal_fill(TiffContext *s,
unsigned int bpp, uint8_t* dst,
int usePtr, const uint8_t *src,
uint8_t c, int width, int offset)
{
switch (bpp) {
case 1:
while (--width >= 0) {
dst[(width+offset)*8+7] = (usePtr ? src[width] : c) & 0x1;
dst[(width+offset)*8+6] = (usePtr ? src[width] : c) >> 1 & 0x1;
dst[(width+offset)*8+5] = (usePtr ? src[width] : c) >> 2 & 0x1;
dst[(width+offset)*8+4] = (usePtr ? src[width] : c) >> 3 & 0x1;
dst[(width+offset)*8+3] = (usePtr ? src[width] : c) >> 4 & 0x1;
dst[(width+offset)*8+2] = (usePtr ? src[width] : c) >> 5 & 0x1;
dst[(width+offset)*8+1] = (usePtr ? src[width] : c) >> 6 & 0x1;
dst[(width+offset)*8+0] = (usePtr ? src[width] : c) >> 7;
}
break;
case 2:
while (--width >= 0) {
dst[(width+offset)*4+3] = (usePtr ? src[width] : c) & 0x3;
dst[(width+offset)*4+2] = (usePtr ? src[width] : c) >> 2 & 0x3;
dst[(width+offset)*4+1] = (usePtr ? src[width] : c) >> 4 & 0x3;
dst[(width+offset)*4+0] = (usePtr ? src[width] : c) >> 6;
}
break;
case 4:
while (--width >= 0) {
dst[(width+offset)*2+1] = (usePtr ? src[width] : c) & 0xF;
dst[(width+offset)*2+0] = (usePtr ? src[width] : c) >> 4;
}
break;
case 10:
case 12:
case 14: {
uint16_t *dst16 = (uint16_t *)dst;
int is_dng = (s->tiff_type == TIFF_TYPE_DNG || s->tiff_type == TIFF_TYPE_CINEMADNG);
uint8_t shift = is_dng ? 0 : 16 - bpp;
GetBitContext gb;
init_get_bits8(&gb, src, width);
for (int i = 0; i < s->width; i++) {
dst16[i] = get_bits(&gb, bpp) << shift;
}
}
break;
default:
if (usePtr) {
memcpy(dst + offset, src, width);
} else {
memset(dst + offset, c, width);
}
}
}
static int deinvert_buffer(TiffContext *s, const uint8_t *src, int size)
{
int i;
av_fast_padded_malloc(&s->deinvert_buf, &s->deinvert_buf_size, size);
if (!s->deinvert_buf)
return AVERROR(ENOMEM);
for (i = 0; i < size; i++)
s->deinvert_buf[i] = ff_reverse[src[i]];
return 0;
}
static void unpack_gray(TiffContext *s, AVFrame *p,
const uint8_t *src, int lnum, int width, int bpp)
{
GetBitContext gb;
uint16_t *dst = (uint16_t *)(p->data[0] + lnum * p->linesize[0]);
init_get_bits8(&gb, src, width);
for (int i = 0; i < s->width; i++) {
dst[i] = get_bits(&gb, bpp);
}
}
static void unpack_yuv(TiffContext *s, AVFrame *p,
const uint8_t *src, int lnum)
{
int i, j, k;
int w = (s->width - 1) / s->subsampling[0] + 1;
uint8_t *pu = &p->data[1][lnum / s->subsampling[1] * p->linesize[1]];
uint8_t *pv = &p->data[2][lnum / s->subsampling[1] * p->linesize[2]];
if (s->width % s->subsampling[0] || s->height % s->subsampling[1]) {
for (i = 0; i < w; i++) {
for (j = 0; j < s->subsampling[1]; j++)
for (k = 0; k < s->subsampling[0]; k++)
p->data[0][FFMIN(lnum + j, s->height-1) * p->linesize[0] +
FFMIN(i * s->subsampling[0] + k, s->width-1)] = *src++;
*pu++ = *src++;
*pv++ = *src++;
}
}else{
for (i = 0; i < w; i++) {
for (j = 0; j < s->subsampling[1]; j++)
for (k = 0; k < s->subsampling[0]; k++)
p->data[0][(lnum + j) * p->linesize[0] +
i * s->subsampling[0] + k] = *src++;
*pu++ = *src++;
*pv++ = *src++;
}
}
}
#if CONFIG_ZLIB
static int tiff_uncompress(uint8_t *dst, unsigned long *len, const uint8_t *src,
int size)
{
z_stream zstream = { 0 };
int zret;
zstream.next_in = src;
zstream.avail_in = size;
zstream.next_out = dst;
zstream.avail_out = *len;
zret = inflateInit(&zstream);
if (zret != Z_OK) {
av_log(NULL, AV_LOG_ERROR, "Inflate init error: %d\n", zret);
return zret;
}
zret = inflate(&zstream, Z_SYNC_FLUSH);
inflateEnd(&zstream);
*len = zstream.total_out;
return zret == Z_STREAM_END ? Z_OK : zret;
}
static int tiff_unpack_zlib(TiffContext *s, AVFrame *p, uint8_t *dst, int stride,
const uint8_t *src, int size, int width, int lines,
int strip_start, int is_yuv)
{
uint8_t *zbuf;
unsigned long outlen;
int ret, line;
outlen = width * lines;
zbuf = av_malloc(outlen);
if (!zbuf)
return AVERROR(ENOMEM);
if (s->fill_order) {
if ((ret = deinvert_buffer(s, src, size)) < 0) {
av_free(zbuf);
return ret;
}
src = s->deinvert_buf;
}
ret = tiff_uncompress(zbuf, &outlen, src, size);
if (ret != Z_OK) {
av_log(s->avctx, AV_LOG_ERROR,
"Uncompressing failed (%lu of %lu) with error %d\n", outlen,
(unsigned long)width * lines, ret);
av_free(zbuf);
return AVERROR_UNKNOWN;
}
src = zbuf;
for (line = 0; line < lines; line++) {
if (s->bpp < 8 && s->avctx->pix_fmt == AV_PIX_FMT_PAL8) {
horizontal_fill(s, s->bpp, dst, 1, src, 0, width, 0);
} else {
memcpy(dst, src, width);
}
if (is_yuv) {
unpack_yuv(s, p, dst, strip_start + line);
line += s->subsampling[1] - 1;
}
dst += stride;
src += width;
}
av_free(zbuf);
return 0;
}
#endif
#if CONFIG_LZMA
static int tiff_uncompress_lzma(uint8_t *dst, uint64_t *len, const uint8_t *src,
int size)
{
lzma_stream stream = LZMA_STREAM_INIT;
lzma_ret ret;
stream.next_in = (uint8_t *)src;
stream.avail_in = size;
stream.next_out = dst;
stream.avail_out = *len;
ret = lzma_stream_decoder(&stream, UINT64_MAX, 0);
if (ret != LZMA_OK) {
av_log(NULL, AV_LOG_ERROR, "LZMA init error: %d\n", ret);
return ret;
}
ret = lzma_code(&stream, LZMA_RUN);
lzma_end(&stream);
*len = stream.total_out;
return ret == LZMA_STREAM_END ? LZMA_OK : ret;
}
static int tiff_unpack_lzma(TiffContext *s, AVFrame *p, uint8_t *dst, int stride,
const uint8_t *src, int size, int width, int lines,
int strip_start, int is_yuv)
{
uint64_t outlen = width * (uint64_t)lines;
int ret, line;
uint8_t *buf = av_malloc(outlen);
if (!buf)
return AVERROR(ENOMEM);
if (s->fill_order) {
if ((ret = deinvert_buffer(s, src, size)) < 0) {
av_free(buf);
return ret;
}
src = s->deinvert_buf;
}
ret = tiff_uncompress_lzma(buf, &outlen, src, size);
if (ret != LZMA_OK) {
av_log(s->avctx, AV_LOG_ERROR,
"Uncompressing failed (%"PRIu64" of %"PRIu64") with error %d\n", outlen,
(uint64_t)width * lines, ret);
av_free(buf);
return AVERROR_UNKNOWN;
}
src = buf;
for (line = 0; line < lines; line++) {
if (s->bpp < 8 && s->avctx->pix_fmt == AV_PIX_FMT_PAL8) {
horizontal_fill(s, s->bpp, dst, 1, src, 0, width, 0);
} else {
memcpy(dst, src, width);
}
if (is_yuv) {
unpack_yuv(s, p, dst, strip_start + line);
line += s->subsampling[1] - 1;
}
dst += stride;
src += width;
}
av_free(buf);
return 0;
}
#endif
static int tiff_unpack_fax(TiffContext *s, uint8_t *dst, int stride,
const uint8_t *src, int size, int width, int lines)
{
int line;
int ret;
if (s->fill_order) {
if ((ret = deinvert_buffer(s, src, size)) < 0)
return ret;
src = s->deinvert_buf;
}
ret = ff_ccitt_unpack(s->avctx, src, size, dst, lines, stride,
s->compr, s->fax_opts);
if (s->bpp < 8 && s->avctx->pix_fmt == AV_PIX_FMT_PAL8)
for (line = 0; line < lines; line++) {
horizontal_fill(s, s->bpp, dst, 1, dst, 0, width, 0);
dst += stride;
}
return ret;
}
static int dng_decode_jpeg(AVCodecContext *avctx, AVFrame *frame,
int tile_byte_count, int dst_x, int dst_y, int w, int h)
{
TiffContext *s = avctx->priv_data;
uint8_t *dst_data, *src_data;
uint32_t dst_offset; /* offset from dst buffer in pixels */
int is_single_comp, is_u16, pixel_size;
int ret;
if (tile_byte_count < 0 || tile_byte_count > bytestream2_get_bytes_left(&s->gb))
return AVERROR_INVALIDDATA;
/* Prepare a packet and send to the MJPEG decoder */
av_packet_unref(s->jpkt);
s->jpkt->data = (uint8_t*)s->gb.buffer;
s->jpkt->size = tile_byte_count;
if (s->is_bayer) {
MJpegDecodeContext *mjpegdecctx = s->avctx_mjpeg->priv_data;
/* We have to set this information here, there is no way to know if a given JPEG is a DNG-embedded
image or not from its own data (and we need that information when decoding it). */
mjpegdecctx->bayer = 1;
}
ret = avcodec_send_packet(s->avctx_mjpeg, s->jpkt);
if (ret < 0) {
av_log(avctx, AV_LOG_ERROR, "Error submitting a packet for decoding\n");
return ret;
}
ret = avcodec_receive_frame(s->avctx_mjpeg, s->jpgframe);
if (ret < 0) {
av_log(avctx, AV_LOG_ERROR, "JPEG decoding error: %s.\n", av_err2str(ret));
/* Normally skip, error if explode */
if (avctx->err_recognition & AV_EF_EXPLODE)
return AVERROR_INVALIDDATA;
else
return 0;
}
is_u16 = (s->bpp > 8);
/* Copy the outputted tile's pixels from 'jpgframe' to 'frame' (final buffer) */
if (s->jpgframe->width != s->avctx_mjpeg->width ||
s->jpgframe->height != s->avctx_mjpeg->height ||
s->jpgframe->format != s->avctx_mjpeg->pix_fmt)
return AVERROR_INVALIDDATA;
/* See dng_blit for explanation */
if (s->avctx_mjpeg->width == w * 2 &&
s->avctx_mjpeg->height == h / 2 &&
s->avctx_mjpeg->pix_fmt == AV_PIX_FMT_GRAY16LE) {
is_single_comp = 1;
} else if (s->avctx_mjpeg->width >= w &&
s->avctx_mjpeg->height >= h &&
s->avctx_mjpeg->pix_fmt == (is_u16 ? AV_PIX_FMT_GRAY16 : AV_PIX_FMT_GRAY8)
) {
is_single_comp = 0;
} else
return AVERROR_INVALIDDATA;
pixel_size = (is_u16 ? sizeof(uint16_t) : sizeof(uint8_t));
if (is_single_comp && !is_u16) {
av_log(s->avctx, AV_LOG_ERROR, "DNGs with bpp <= 8 and 1 component are unsupported\n");
av_frame_unref(s->jpgframe);
return AVERROR_PATCHWELCOME;
}
dst_offset = dst_x + frame->linesize[0] * dst_y / pixel_size;
dst_data = frame->data[0] + dst_offset * pixel_size;
src_data = s->jpgframe->data[0];
dng_blit(s,
dst_data,
frame->linesize[0] / pixel_size,
src_data,
s->jpgframe->linesize[0] / pixel_size,
w,
h,
is_single_comp,
is_u16);
av_frame_unref(s->jpgframe);
return 0;
}
static int tiff_unpack_strip(TiffContext *s, AVFrame *p, uint8_t *dst, int stride,
const uint8_t *src, int size, int strip_start, int lines)
{
PutByteContext pb;
int c, line, pixels, code, ret;
const uint8_t *ssrc = src;
int width = ((s->width * s->bpp) + 7) >> 3;
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(p->format);
int is_yuv = !(desc->flags & AV_PIX_FMT_FLAG_RGB) &&
(desc->flags & AV_PIX_FMT_FLAG_PLANAR) &&
desc->nb_components >= 3;
int is_dng;
if (s->planar)
width /= s->bppcount;
if (size <= 0)
return AVERROR_INVALIDDATA;
if (is_yuv) {
int bytes_per_row = (((s->width - 1) / s->subsampling[0] + 1) * s->bpp *
s->subsampling[0] * s->subsampling[1] + 7) >> 3;
av_fast_padded_malloc(&s->yuv_line, &s->yuv_line_size, bytes_per_row);
if (s->yuv_line == NULL) {
av_log(s->avctx, AV_LOG_ERROR, "Not enough memory\n");
return AVERROR(ENOMEM);
}
dst = s->yuv_line;
stride = 0;
width = (s->width - 1) / s->subsampling[0] + 1;
width = width * s->subsampling[0] * s->subsampling[1] + 2*width;
av_assert0(width <= bytes_per_row);
av_assert0(s->bpp == 24);
}
if (s->is_bayer) {
av_assert0(width == (s->bpp * s->width + 7) >> 3);
}
av_assert0(!(s->is_bayer && is_yuv));
if (p->format == AV_PIX_FMT_GRAY12) {
av_fast_padded_malloc(&s->yuv_line, &s->yuv_line_size, width);
if (s->yuv_line == NULL) {
av_log(s->avctx, AV_LOG_ERROR, "Not enough memory\n");
return AVERROR(ENOMEM);
}
dst = s->yuv_line;
stride = 0;
}
if (s->compr == TIFF_DEFLATE || s->compr == TIFF_ADOBE_DEFLATE) {
#if CONFIG_ZLIB
return tiff_unpack_zlib(s, p, dst, stride, src, size, width, lines,
strip_start, is_yuv);
#else
av_log(s->avctx, AV_LOG_ERROR,
"zlib support not enabled, "
"deflate compression not supported\n");
return AVERROR(ENOSYS);
#endif
}
if (s->compr == TIFF_LZMA) {
#if CONFIG_LZMA
return tiff_unpack_lzma(s, p, dst, stride, src, size, width, lines,
strip_start, is_yuv);
#else
av_log(s->avctx, AV_LOG_ERROR,
"LZMA support not enabled\n");
return AVERROR(ENOSYS);
#endif
}
if (s->compr == TIFF_LZW) {
if (s->fill_order) {
if ((ret = deinvert_buffer(s, src, size)) < 0)
return ret;
ssrc = src = s->deinvert_buf;
}
if (size > 1 && !src[0] && (src[1]&1)) {
av_log(s->avctx, AV_LOG_ERROR, "Old style LZW is unsupported\n");
}
if ((ret = ff_lzw_decode_init(s->lzw, 8, src, size, FF_LZW_TIFF)) < 0) {
av_log(s->avctx, AV_LOG_ERROR, "Error initializing LZW decoder\n");
return ret;
}
for (line = 0; line < lines; line++) {
pixels = ff_lzw_decode(s->lzw, dst, width);
if (pixels < width) {
av_log(s->avctx, AV_LOG_ERROR, "Decoded only %i bytes of %i\n",
pixels, width);
return AVERROR_INVALIDDATA;
}
if (s->bpp < 8 && s->avctx->pix_fmt == AV_PIX_FMT_PAL8)
horizontal_fill(s, s->bpp, dst, 1, dst, 0, width, 0);
if (is_yuv) {
unpack_yuv(s, p, dst, strip_start + line);
line += s->subsampling[1] - 1;
} else if (p->format == AV_PIX_FMT_GRAY12) {
unpack_gray(s, p, dst, strip_start + line, width, s->bpp);
}
dst += stride;
}
return 0;
}
if (s->compr == TIFF_CCITT_RLE ||
s->compr == TIFF_G3 ||
s->compr == TIFF_G4) {
if (is_yuv || p->format == AV_PIX_FMT_GRAY12)
return AVERROR_INVALIDDATA;
return tiff_unpack_fax(s, dst, stride, src, size, width, lines);
}
bytestream2_init(&s->gb, src, size);
bytestream2_init_writer(&pb, dst, is_yuv ? s->yuv_line_size : (stride * lines));
is_dng = (s->tiff_type == TIFF_TYPE_DNG || s->tiff_type == TIFF_TYPE_CINEMADNG);
/* Decode JPEG-encoded DNGs with strips */
if (s->compr == TIFF_NEWJPEG && is_dng) {
if (s->strips > 1) {
av_log(s->avctx, AV_LOG_ERROR, "More than one DNG JPEG strips unsupported\n");
return AVERROR_PATCHWELCOME;
}
if (!s->is_bayer)
return AVERROR_PATCHWELCOME;
if ((ret = dng_decode_jpeg(s->avctx, p, s->stripsize, 0, 0, s->width, s->height)) < 0)
return ret;
return 0;
}
if (is_dng && stride == 0)
return AVERROR_INVALIDDATA;
for (line = 0; line < lines; line++) {
if (src - ssrc > size) {
av_log(s->avctx, AV_LOG_ERROR, "Source data overread\n");
return AVERROR_INVALIDDATA;
}
if (bytestream2_get_bytes_left(&s->gb) == 0 || bytestream2_get_eof(&pb))
break;
bytestream2_seek_p(&pb, stride * line, SEEK_SET);
switch (s->compr) {
case TIFF_RAW:
if (ssrc + size - src < width)
return AVERROR_INVALIDDATA;
if (!s->fill_order) {
horizontal_fill(s, s->bpp * (s->avctx->pix_fmt == AV_PIX_FMT_PAL8 || s->is_bayer),
dst, 1, src, 0, width, 0);
} else {
int i;
for (i = 0; i < width; i++)
dst[i] = ff_reverse[src[i]];
}
/* Color processing for DNG images with uncompressed strips (non-tiled) */
if (is_dng) {
int is_u16, pixel_size_bytes, pixel_size_bits, elements;
is_u16 = (s->bpp / s->bppcount > 8);
pixel_size_bits = (is_u16 ? 16 : 8);
pixel_size_bytes = (is_u16 ? sizeof(uint16_t) : sizeof(uint8_t));
elements = width / pixel_size_bytes * pixel_size_bits / s->bpp * s->bppcount; // need to account for [1, 16] bpp
av_assert0 (elements * pixel_size_bytes <= FFABS(stride));
dng_blit(s,
dst,
0, // no stride, only 1 line
dst,
0, // no stride, only 1 line
elements,
1,
0, // single-component variation is only preset in JPEG-encoded DNGs
is_u16);
}
src += width;
break;
case TIFF_PACKBITS:
for (pixels = 0; pixels < width;) {
if (ssrc + size - src < 2) {
av_log(s->avctx, AV_LOG_ERROR, "Read went out of bounds\n");
return AVERROR_INVALIDDATA;
}
code = s->fill_order ? (int8_t) ff_reverse[*src++]: (int8_t) *src++;
if (code >= 0) {
code++;
if (pixels + code > width ||
ssrc + size - src < code) {
av_log(s->avctx, AV_LOG_ERROR,
"Copy went out of bounds\n");
return AVERROR_INVALIDDATA;
}
horizontal_fill(s, s->bpp * (s->avctx->pix_fmt == AV_PIX_FMT_PAL8),
dst, 1, src, 0, code, pixels);
src += code;
pixels += code;
} else if (code != -128) { // -127..-1
code = (-code) + 1;
if (pixels + code > width) {
av_log(s->avctx, AV_LOG_ERROR,
"Run went out of bounds\n");
return AVERROR_INVALIDDATA;
}
c = *src++;
horizontal_fill(s, s->bpp * (s->avctx->pix_fmt == AV_PIX_FMT_PAL8),
dst, 0, NULL, c, code, pixels);
pixels += code;
}
}
if (s->fill_order) {
int i;
for (i = 0; i < width; i++)
dst[i] = ff_reverse[dst[i]];
}
break;
}
if (is_yuv) {
unpack_yuv(s, p, dst, strip_start + line);
line += s->subsampling[1] - 1;
} else if (p->format == AV_PIX_FMT_GRAY12) {
unpack_gray(s, p, dst, strip_start + line, width, s->bpp);
}
dst += stride;
}
return 0;
}
static int dng_decode_tiles(AVCodecContext *avctx, AVFrame *frame,
const AVPacket *avpkt)
{
TiffContext *s = avctx->priv_data;
int tile_idx;
int tile_offset_offset, tile_offset;
int tile_byte_count_offset, tile_byte_count;
int tile_count_x, tile_count_y;
int tile_width, tile_length;
int has_width_leftover, has_height_leftover;
int tile_x = 0, tile_y = 0;
int pos_x = 0, pos_y = 0;
int ret;
if (s->tile_width <= 0 || s->tile_length <= 0)
return AVERROR_INVALIDDATA;
has_width_leftover = (s->width % s->tile_width != 0);
has_height_leftover = (s->height % s->tile_length != 0);
/* Calculate tile counts (round up) */
tile_count_x = (s->width + s->tile_width - 1) / s->tile_width;
tile_count_y = (s->height + s->tile_length - 1) / s->tile_length;
/* Iterate over the number of tiles */
for (tile_idx = 0; tile_idx < s->tile_count; tile_idx++) {
tile_x = tile_idx % tile_count_x;
tile_y = tile_idx / tile_count_x;
if (has_width_leftover && tile_x == tile_count_x - 1) // If on the right-most tile
tile_width = s->width % s->tile_width;
else
tile_width = s->tile_width;
if (has_height_leftover && tile_y == tile_count_y - 1) // If on the bottom-most tile
tile_length = s->height % s->tile_length;
else
tile_length = s->tile_length;
/* Read tile offset */
tile_offset_offset = s->tile_offsets_offset + tile_idx * sizeof(int);
bytestream2_seek(&s->gb, tile_offset_offset, SEEK_SET);
tile_offset = ff_tget_long(&s->gb, s->le);
/* Read tile byte size */
tile_byte_count_offset = s->tile_byte_counts_offset + tile_idx * sizeof(int);
bytestream2_seek(&s->gb, tile_byte_count_offset, SEEK_SET);
tile_byte_count = ff_tget_long(&s->gb, s->le);
/* Seek to tile data */
bytestream2_seek(&s->gb, tile_offset, SEEK_SET);
/* Decode JPEG tile and copy it in the reference frame */
ret = dng_decode_jpeg(avctx, frame, tile_byte_count, pos_x, pos_y, tile_width, tile_length);
if (ret < 0)
return ret;
/* Advance current positions */
pos_x += tile_width;
if (tile_x == tile_count_x - 1) { // If on the right edge
pos_x = 0;
pos_y += tile_length;
}
}
/* Frame is ready to be output */
frame->pict_type = AV_PICTURE_TYPE_I;
frame->key_frame = 1;
return avpkt->size;
}
static int init_image(TiffContext *s, AVFrame *frame)
{
int ret;
int create_gray_palette = 0;
// make sure there is no aliasing in the following switch
if (s->bpp >= 100 || s->bppcount >= 10) {
av_log(s->avctx, AV_LOG_ERROR,
"Unsupported image parameters: bpp=%d, bppcount=%d\n",
s->bpp, s->bppcount);
return AVERROR_INVALIDDATA;
}
switch (s->planar * 1000 + s->bpp * 10 + s->bppcount + s->is_bayer * 10000) {
case 11:
if (!s->palette_is_set) {
s->avctx->pix_fmt = AV_PIX_FMT_MONOBLACK;
break;
}
case 21:
case 41:
s->avctx->pix_fmt = AV_PIX_FMT_PAL8;
if (!s->palette_is_set) {
create_gray_palette = 1;
}
break;
case 81:
s->avctx->pix_fmt = s->palette_is_set ? AV_PIX_FMT_PAL8 : AV_PIX_FMT_GRAY8;
break;
case 121:
s->avctx->pix_fmt = AV_PIX_FMT_GRAY12;
break;
case 10081:
switch (AV_RL32(s->pattern)) {
case 0x02010100:
s->avctx->pix_fmt = AV_PIX_FMT_BAYER_RGGB8;
break;
case 0x00010102:
s->avctx->pix_fmt = AV_PIX_FMT_BAYER_BGGR8;
break;
case 0x01000201:
s->avctx->pix_fmt = AV_PIX_FMT_BAYER_GBRG8;
break;
case 0x01020001:
s->avctx->pix_fmt = AV_PIX_FMT_BAYER_GRBG8;
break;
default:
av_log(s->avctx, AV_LOG_ERROR, "Unsupported Bayer pattern: 0x%X\n",
AV_RL32(s->pattern));
return AVERROR_PATCHWELCOME;
}
break;
case 10101:
case 10121:
case 10141:
case 10161:
switch (AV_RL32(s->pattern)) {
case 0x02010100:
s->avctx->pix_fmt = AV_PIX_FMT_BAYER_RGGB16;
break;
case 0x00010102:
s->avctx->pix_fmt = AV_PIX_FMT_BAYER_BGGR16;
break;
case 0x01000201:
s->avctx->pix_fmt = AV_PIX_FMT_BAYER_GBRG16;
break;
case 0x01020001:
s->avctx->pix_fmt = AV_PIX_FMT_BAYER_GRBG16;
break;
default:
av_log(s->avctx, AV_LOG_ERROR, "Unsupported Bayer pattern: 0x%X\n",
AV_RL32(s->pattern));
return AVERROR_PATCHWELCOME;
}
break;
case 243:
if (s->photometric == TIFF_PHOTOMETRIC_YCBCR) {
if (s->subsampling[0] == 1 && s->subsampling[1] == 1) {
s->avctx->pix_fmt = AV_PIX_FMT_YUV444P;
} else if (s->subsampling[0] == 2 && s->subsampling[1] == 1) {
s->avctx->pix_fmt = AV_PIX_FMT_YUV422P;
} else if (s->subsampling[0] == 4 && s->subsampling[1] == 1) {
s->avctx->pix_fmt = AV_PIX_FMT_YUV411P;
} else if (s->subsampling[0] == 1 && s->subsampling[1] == 2) {
s->avctx->pix_fmt = AV_PIX_FMT_YUV440P;
} else if (s->subsampling[0] == 2 && s->subsampling[1] == 2) {
s->avctx->pix_fmt = AV_PIX_FMT_YUV420P;
} else if (s->subsampling[0] == 4 && s->subsampling[1] == 4) {
s->avctx->pix_fmt = AV_PIX_FMT_YUV410P;
} else {
av_log(s->avctx, AV_LOG_ERROR, "Unsupported YCbCr subsampling\n");
return AVERROR_PATCHWELCOME;
}
} else
s->avctx->pix_fmt = AV_PIX_FMT_RGB24;
break;
case 161:
s->avctx->pix_fmt = s->le ? AV_PIX_FMT_GRAY16LE : AV_PIX_FMT_GRAY16BE;
break;
case 162:
s->avctx->pix_fmt = AV_PIX_FMT_YA8;
break;
case 322:
s->avctx->pix_fmt = s->le ? AV_PIX_FMT_YA16LE : AV_PIX_FMT_YA16BE;
break;
case 324:
s->avctx->pix_fmt = s->photometric == TIFF_PHOTOMETRIC_SEPARATED ? AV_PIX_FMT_RGB0 : AV_PIX_FMT_RGBA;
break;
case 405:
if (s->photometric == TIFF_PHOTOMETRIC_SEPARATED)
s->avctx->pix_fmt = AV_PIX_FMT_RGBA;
else {
av_log(s->avctx, AV_LOG_ERROR,
"bpp=40 without PHOTOMETRIC_SEPARATED is unsupported\n");
return AVERROR_PATCHWELCOME;
}
break;
case 483:
s->avctx->pix_fmt = s->le ? AV_PIX_FMT_RGB48LE : AV_PIX_FMT_RGB48BE;
break;
case 644:
s->avctx->pix_fmt = s->le ? AV_PIX_FMT_RGBA64LE : AV_PIX_FMT_RGBA64BE;
break;
case 1243:
s->avctx->pix_fmt = AV_PIX_FMT_GBRP;
break;
case 1324:
s->avctx->pix_fmt = AV_PIX_FMT_GBRAP;
break;
case 1483:
s->avctx->pix_fmt = s->le ? AV_PIX_FMT_GBRP16LE : AV_PIX_FMT_GBRP16BE;
break;
case 1644:
s->avctx->pix_fmt = s->le ? AV_PIX_FMT_GBRAP16LE : AV_PIX_FMT_GBRAP16BE;
break;
default:
av_log(s->avctx, AV_LOG_ERROR,
"This format is not supported (bpp=%d, bppcount=%d)\n",
s->bpp, s->bppcount);
return AVERROR_INVALIDDATA;
}
if (s->photometric == TIFF_PHOTOMETRIC_YCBCR) {
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(s->avctx->pix_fmt);
if((desc->flags & AV_PIX_FMT_FLAG_RGB) ||
!(desc->flags & AV_PIX_FMT_FLAG_PLANAR) ||
desc->nb_components < 3) {
av_log(s->avctx, AV_LOG_ERROR, "Unsupported YCbCr variant\n");
return AVERROR_INVALIDDATA;
}
}
if (s->width != s->avctx->width || s->height != s->avctx->height) {
ret = ff_set_dimensions(s->avctx, s->width, s->height);
if (ret < 0)
return ret;
}
if ((ret = ff_thread_get_buffer(s->avctx, frame, 0)) < 0)
return ret;
if (s->avctx->pix_fmt == AV_PIX_FMT_PAL8) {
if (!create_gray_palette)
memcpy(frame->data[1], s->palette, sizeof(s->palette));
else {
/* make default grayscale pal */
int i;
uint32_t *pal = (uint32_t *)frame->data[1];
for (i = 0; i < 1<<s->bpp; i++)
pal[i] = 0xFFU << 24 | i * 255 / ((1<<s->bpp) - 1) * 0x010101;
}
}
return 0;
}
static void set_sar(TiffContext *s, unsigned tag, unsigned num, unsigned den)
{
int offset = tag == TIFF_YRES ? 2 : 0;
s->res[offset++] = num;
s->res[offset] = den;
if (s->res[0] && s->res[1] && s->res[2] && s->res[3]) {
uint64_t num = s->res[2] * (uint64_t)s->res[1];
uint64_t den = s->res[0] * (uint64_t)s->res[3];
if (num > INT64_MAX || den > INT64_MAX) {
num = num >> 1;
den = den >> 1;
}
av_reduce(&s->avctx->sample_aspect_ratio.num, &s->avctx->sample_aspect_ratio.den,
num, den, INT32_MAX);
if (!s->avctx->sample_aspect_ratio.den)
s->avctx->sample_aspect_ratio = (AVRational) {0, 1};
}
}
static int tiff_decode_tag(TiffContext *s, AVFrame *frame)
{
AVFrameSideData *sd;
GetByteContext gb_temp;
unsigned tag, type, count, off, value = 0, value2 = 1; // value2 is a denominator so init. to 1
int i, start;
int pos;
int ret;
double *dp;
ret = ff_tread_tag(&s->gb, s->le, &tag, &type, &count, &start);
if (ret < 0) {
goto end;
}
if (tag <= s->last_tag)
return AVERROR_INVALIDDATA;
// We ignore TIFF_STRIP_SIZE as it is sometimes in the logic but wrong order around TIFF_STRIP_OFFS
if (tag != TIFF_STRIP_SIZE)
s->last_tag = tag;
off = bytestream2_tell(&s->gb);
if (count == 1) {
switch (type) {
case TIFF_BYTE:
case TIFF_SHORT:
case TIFF_LONG:
value = ff_tget(&s->gb, type, s->le);
break;
case TIFF_RATIONAL:
value = ff_tget(&s->gb, TIFF_LONG, s->le);
value2 = ff_tget(&s->gb, TIFF_LONG, s->le);
if (!value2) {
av_log(s->avctx, AV_LOG_WARNING, "Invalid denominator in rational\n");
value2 = 1;
}
break;
case TIFF_STRING:
if (count <= 4) {
break;
}
default:
value = UINT_MAX;
}
}
switch (tag) {
case TIFF_SUBFILE:
s->is_thumbnail = (value != 0);
break;
case TIFF_WIDTH:
s->width = value;
break;
case TIFF_HEIGHT:
s->height = value;
break;
case TIFF_BPP:
if (count > 5 || count <= 0) {
av_log(s->avctx, AV_LOG_ERROR,
"This format is not supported (bpp=%d, %d components)\n",
value, count);
return AVERROR_INVALIDDATA;
}
s->bppcount = count;
if (count == 1)
s->bpp = value;
else {
switch (type) {
case TIFF_BYTE:
case TIFF_SHORT:
case TIFF_LONG:
s->bpp = 0;
if (bytestream2_get_bytes_left(&s->gb) < type_sizes[type] * count)
return AVERROR_INVALIDDATA;
for (i = 0; i < count; i++)
s->bpp += ff_tget(&s->gb, type, s->le);
break;
default:
s->bpp = -1;
}
}
break;
case TIFF_SAMPLES_PER_PIXEL:
if (count != 1) {
av_log(s->avctx, AV_LOG_ERROR,
"Samples per pixel requires a single value, many provided\n");
return AVERROR_INVALIDDATA;
}
if (value > 5 || value <= 0) {
av_log(s->avctx, AV_LOG_ERROR,
"Invalid samples per pixel %d\n", value);
return AVERROR_INVALIDDATA;
}
if (s->bppcount == 1)
s->bpp *= value;
s->bppcount = value;
break;
case TIFF_COMPR:
s->compr = value;
av_log(s->avctx, AV_LOG_DEBUG, "compression: %d\n", s->compr);
s->predictor = 0;
switch (s->compr) {
case TIFF_RAW:
case TIFF_PACKBITS:
case TIFF_LZW:
case TIFF_CCITT_RLE:
break;
case TIFF_G3:
case TIFF_G4:
s->fax_opts = 0;
break;
case TIFF_DEFLATE:
case TIFF_ADOBE_DEFLATE:
#if CONFIG_ZLIB
break;
#else
av_log(s->avctx, AV_LOG_ERROR, "Deflate: ZLib not compiled in\n");
return AVERROR(ENOSYS);
#endif
case TIFF_JPEG:
case TIFF_NEWJPEG:
s->is_jpeg = 1;
break;
case TIFF_LZMA:
#if CONFIG_LZMA
break;
#else
av_log(s->avctx, AV_LOG_ERROR, "LZMA not compiled in\n");
return AVERROR(ENOSYS);
#endif
default:
av_log(s->avctx, AV_LOG_ERROR, "Unknown compression method %i\n",
s->compr);
return AVERROR_INVALIDDATA;
}
break;
case TIFF_ROWSPERSTRIP:
if (!value || (type == TIFF_LONG && value == UINT_MAX))
value = s->height;
s->rps = FFMIN(value, s->height);
break;
case TIFF_STRIP_OFFS:
if (count == 1) {
if (value > INT_MAX) {
av_log(s->avctx, AV_LOG_ERROR,
"strippos %u too large\n", value);
return AVERROR_INVALIDDATA;
}
s->strippos = 0;
s->stripoff = value;
} else
s->strippos = off;
s->strips = count;
if (s->strips == 1)
s->rps = s->height;
s->sot = type;
break;
case TIFF_STRIP_SIZE:
if (count == 1) {
if (value > INT_MAX) {
av_log(s->avctx, AV_LOG_ERROR,
"stripsize %u too large\n", value);
return AVERROR_INVALIDDATA;
}
s->stripsizesoff = 0;
s->stripsize = value;
s->strips = 1;
} else {
s->stripsizesoff = off;
}
s->strips = count;
s->sstype = type;
break;
case TIFF_XRES:
case TIFF_YRES:
set_sar(s, tag, value, value2);
break;
case TIFF_TILE_OFFSETS:
s->tile_offsets_offset = off;
s->tile_count = count;
s->is_tiled = 1;
break;
case TIFF_TILE_BYTE_COUNTS:
s->tile_byte_counts_offset = off;
break;
case TIFF_TILE_LENGTH:
s->tile_length = value;
break;
case TIFF_TILE_WIDTH:
s->tile_width = value;
break;
case TIFF_PREDICTOR:
s->predictor = value;
break;
case TIFF_SUB_IFDS:
if (count == 1)
s->sub_ifd = value;
else if (count > 1)
s->sub_ifd = ff_tget(&s->gb, TIFF_LONG, s->le); /** Only get the first SubIFD */
break;
case DNG_LINEARIZATION_TABLE:
if (count > FF_ARRAY_ELEMS(s->dng_lut))
return AVERROR_INVALIDDATA;
for (int i = 0; i < count; i++)
s->dng_lut[i] = ff_tget(&s->gb, type, s->le);
break;
case DNG_BLACK_LEVEL:
if (count > 1) { /* Use the first value in the pattern (assume they're all the same) */
if (type == TIFF_RATIONAL) {
value = ff_tget(&s->gb, TIFF_LONG, s->le);
value2 = ff_tget(&s->gb, TIFF_LONG, s->le);
if (!value2) {
av_log(s->avctx, AV_LOG_WARNING, "Invalid black level denominator\n");
value2 = 1;
}
s->black_level = value / value2;
} else
s->black_level = ff_tget(&s->gb, type, s->le);
av_log(s->avctx, AV_LOG_WARNING, "Assuming black level pattern values are identical\n");
} else {
s->black_level = value / value2;
}
break;
case DNG_WHITE_LEVEL:
s->white_level = value;
break;
case TIFF_CFA_PATTERN_DIM:
if (count != 2 || (ff_tget(&s->gb, type, s->le) != 2 &&
ff_tget(&s->gb, type, s->le) != 2)) {
av_log(s->avctx, AV_LOG_ERROR, "CFA Pattern dimensions are not 2x2\n");
return AVERROR_INVALIDDATA;
}
break;
case TIFF_CFA_PATTERN:
s->is_bayer = 1;
s->pattern[0] = ff_tget(&s->gb, type, s->le);
s->pattern[1] = ff_tget(&s->gb, type, s->le);
s->pattern[2] = ff_tget(&s->gb, type, s->le);
s->pattern[3] = ff_tget(&s->gb, type, s->le);
break;
case TIFF_PHOTOMETRIC:
switch (value) {
case TIFF_PHOTOMETRIC_WHITE_IS_ZERO:
case TIFF_PHOTOMETRIC_BLACK_IS_ZERO:
case TIFF_PHOTOMETRIC_RGB:
case TIFF_PHOTOMETRIC_PALETTE:
case TIFF_PHOTOMETRIC_SEPARATED:
case TIFF_PHOTOMETRIC_YCBCR:
case TIFF_PHOTOMETRIC_CFA:
case TIFF_PHOTOMETRIC_LINEAR_RAW: // Used by DNG images
s->photometric = value;
break;
case TIFF_PHOTOMETRIC_ALPHA_MASK:
case TIFF_PHOTOMETRIC_CIE_LAB:
case TIFF_PHOTOMETRIC_ICC_LAB:
case TIFF_PHOTOMETRIC_ITU_LAB:
case TIFF_PHOTOMETRIC_LOG_L:
case TIFF_PHOTOMETRIC_LOG_LUV:
avpriv_report_missing_feature(s->avctx,
"PhotometricInterpretation 0x%04X",
value);
return AVERROR_PATCHWELCOME;
default:
av_log(s->avctx, AV_LOG_ERROR, "PhotometricInterpretation %u is "
"unknown\n", value);
return AVERROR_INVALIDDATA;
}
break;
case TIFF_FILL_ORDER:
if (value < 1 || value > 2) {
av_log(s->avctx, AV_LOG_ERROR,
"Unknown FillOrder value %d, trying default one\n", value);
value = 1;
}
s->fill_order = value - 1;
break;
case TIFF_PAL: {
GetByteContext pal_gb[3];
off = type_sizes[type];
if (count / 3 > 256 ||
bytestream2_get_bytes_left(&s->gb) < count / 3 * off * 3)
return AVERROR_INVALIDDATA;
pal_gb[0] = pal_gb[1] = pal_gb[2] = s->gb;
bytestream2_skip(&pal_gb[1], count / 3 * off);
bytestream2_skip(&pal_gb[2], count / 3 * off * 2);
off = (type_sizes[type] - 1) << 3;
if (off > 31U) {
av_log(s->avctx, AV_LOG_ERROR, "palette shift %d is out of range\n", off);
return AVERROR_INVALIDDATA;
}
for (i = 0; i < count / 3; i++) {
uint32_t p = 0xFF000000;
p |= (ff_tget(&pal_gb[0], type, s->le) >> off) << 16;
p |= (ff_tget(&pal_gb[1], type, s->le) >> off) << 8;
p |= ff_tget(&pal_gb[2], type, s->le) >> off;
s->palette[i] = p;
}
s->palette_is_set = 1;
break;
}
case TIFF_PLANAR:
s->planar = value == 2;
break;
case TIFF_YCBCR_SUBSAMPLING:
if (count != 2) {
av_log(s->avctx, AV_LOG_ERROR, "subsample count invalid\n");
return AVERROR_INVALIDDATA;
}
for (i = 0; i < count; i++) {
s->subsampling[i] = ff_tget(&s->gb, type, s->le);
if (s->subsampling[i] <= 0) {
av_log(s->avctx, AV_LOG_ERROR, "subsampling %d is invalid\n", s->subsampling[i]);
s->subsampling[i] = 1;
return AVERROR_INVALIDDATA;
}
}
break;
case TIFF_T4OPTIONS:
if (s->compr == TIFF_G3)
s->fax_opts = value;
break;
case TIFF_T6OPTIONS:
if (s->compr == TIFF_G4)
s->fax_opts = value;
break;
#define ADD_METADATA(count, name, sep)\
if ((ret = add_metadata(count, type, name, sep, s, frame)) < 0) {\
av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");\
goto end;\
}
case TIFF_MODEL_PIXEL_SCALE:
ADD_METADATA(count, "ModelPixelScaleTag", NULL);
break;
case TIFF_MODEL_TRANSFORMATION:
ADD_METADATA(count, "ModelTransformationTag", NULL);
break;
case TIFF_MODEL_TIEPOINT:
ADD_METADATA(count, "ModelTiepointTag", NULL);
break;
case TIFF_GEO_KEY_DIRECTORY:
if (s->geotag_count) {
avpriv_request_sample(s->avctx, "Multiple geo key directories");
return AVERROR_INVALIDDATA;
}
ADD_METADATA(1, "GeoTIFF_Version", NULL);
ADD_METADATA(2, "GeoTIFF_Key_Revision", ".");
s->geotag_count = ff_tget_short(&s->gb, s->le);
if (s->geotag_count > count / 4 - 1) {
s->geotag_count = count / 4 - 1;
av_log(s->avctx, AV_LOG_WARNING, "GeoTIFF key directory buffer shorter than specified\n");
}
if ( bytestream2_get_bytes_left(&s->gb) < s->geotag_count * sizeof(int16_t) * 4
|| s->geotag_count == 0) {
s->geotag_count = 0;
return -1;
}
s->geotags = av_calloc(s->geotag_count, sizeof(*s->geotags));
if (!s->geotags) {
av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");
s->geotag_count = 0;
goto end;
}
for (i = 0; i < s->geotag_count; i++) {
s->geotags[i].key = ff_tget_short(&s->gb, s->le);
s->geotags[i].type = ff_tget_short(&s->gb, s->le);
s->geotags[i].count = ff_tget_short(&s->gb, s->le);
if (!s->geotags[i].type)
s->geotags[i].val = get_geokey_val(s->geotags[i].key, ff_tget_short(&s->gb, s->le));
else
s->geotags[i].offset = ff_tget_short(&s->gb, s->le);
}
break;
case TIFF_GEO_DOUBLE_PARAMS:
if (count >= INT_MAX / sizeof(int64_t))
return AVERROR_INVALIDDATA;
if (bytestream2_get_bytes_left(&s->gb) < count * sizeof(int64_t))
return AVERROR_INVALIDDATA;
dp = av_malloc_array(count, sizeof(double));
if (!dp) {
av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");
goto end;
}
for (i = 0; i < count; i++)
dp[i] = ff_tget_double(&s->gb, s->le);
for (i = 0; i < s->geotag_count; i++) {
if (s->geotags[i].type == TIFF_GEO_DOUBLE_PARAMS) {
if (s->geotags[i].count == 0
|| s->geotags[i].offset + s->geotags[i].count > count) {
av_log(s->avctx, AV_LOG_WARNING, "Invalid GeoTIFF key %d\n", s->geotags[i].key);
} else if (s->geotags[i].val) {
av_log(s->avctx, AV_LOG_WARNING, "Duplicate GeoTIFF key %d\n", s->geotags[i].key);
} else {
char *ap = doubles2str(&dp[s->geotags[i].offset], s->geotags[i].count, ", ");
if (!ap) {
av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");
av_freep(&dp);
return AVERROR(ENOMEM);
}
s->geotags[i].val = ap;
}
}
}
av_freep(&dp);
break;
case TIFF_GEO_ASCII_PARAMS:
pos = bytestream2_tell(&s->gb);
for (i = 0; i < s->geotag_count; i++) {
if (s->geotags[i].type == TIFF_GEO_ASCII_PARAMS) {
if (s->geotags[i].count == 0
|| s->geotags[i].offset + s->geotags[i].count > count) {
av_log(s->avctx, AV_LOG_WARNING, "Invalid GeoTIFF key %d\n", s->geotags[i].key);
} else {
char *ap;
bytestream2_seek(&s->gb, pos + s->geotags[i].offset, SEEK_SET);
if (bytestream2_get_bytes_left(&s->gb) < s->geotags[i].count)
return AVERROR_INVALIDDATA;
if (s->geotags[i].val)
return AVERROR_INVALIDDATA;
ap = av_malloc(s->geotags[i].count);
if (!ap) {
av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");
return AVERROR(ENOMEM);
}
bytestream2_get_bufferu(&s->gb, ap, s->geotags[i].count);
ap[s->geotags[i].count - 1] = '\0'; //replace the "|" delimiter with a 0 byte
s->geotags[i].val = ap;
}
}
}
break;
case TIFF_ICC_PROFILE:
gb_temp = s->gb;
bytestream2_seek(&gb_temp, SEEK_SET, off);
if (bytestream2_get_bytes_left(&gb_temp) < count)
return AVERROR_INVALIDDATA;
sd = av_frame_new_side_data(frame, AV_FRAME_DATA_ICC_PROFILE, count);
if (!sd)
return AVERROR(ENOMEM);
bytestream2_get_bufferu(&gb_temp, sd->data, count);
break;
case TIFF_ARTIST:
ADD_METADATA(count, "artist", NULL);
break;
case TIFF_COPYRIGHT:
ADD_METADATA(count, "copyright", NULL);
break;
case TIFF_DATE:
ADD_METADATA(count, "date", NULL);
break;
case TIFF_DOCUMENT_NAME:
ADD_METADATA(count, "document_name", NULL);
break;
case TIFF_HOST_COMPUTER:
ADD_METADATA(count, "computer", NULL);
break;
case TIFF_IMAGE_DESCRIPTION:
ADD_METADATA(count, "description", NULL);
break;
case TIFF_MAKE:
ADD_METADATA(count, "make", NULL);
break;
case TIFF_MODEL:
ADD_METADATA(count, "model", NULL);
break;
case TIFF_PAGE_NAME:
ADD_METADATA(count, "page_name", NULL);
break;
case TIFF_PAGE_NUMBER:
ADD_METADATA(count, "page_number", " / ");
// need to seek back to re-read the page number
bytestream2_seek(&s->gb, -count * sizeof(uint16_t), SEEK_CUR);
// read the page number
s->cur_page = ff_tget(&s->gb, TIFF_SHORT, s->le);
// get back to where we were before the previous seek
bytestream2_seek(&s->gb, count * sizeof(uint16_t) - sizeof(uint16_t), SEEK_CUR);
break;
case TIFF_SOFTWARE_NAME:
ADD_METADATA(count, "software", NULL);
break;
case DNG_VERSION:
if (count == 4) {
unsigned int ver[4];
ver[0] = ff_tget(&s->gb, type, s->le);
ver[1] = ff_tget(&s->gb, type, s->le);
ver[2] = ff_tget(&s->gb, type, s->le);
ver[3] = ff_tget(&s->gb, type, s->le);
av_log(s->avctx, AV_LOG_DEBUG, "DNG file, version %u.%u.%u.%u\n",
ver[0], ver[1], ver[2], ver[3]);
tiff_set_type(s, TIFF_TYPE_DNG);
}
break;
case CINEMADNG_TIME_CODES:
case CINEMADNG_FRAME_RATE:
case CINEMADNG_T_STOP:
case CINEMADNG_REEL_NAME:
case CINEMADNG_CAMERA_LABEL:
tiff_set_type(s, TIFF_TYPE_CINEMADNG);
break;
default:
if (s->avctx->err_recognition & AV_EF_EXPLODE) {
av_log(s->avctx, AV_LOG_ERROR,
"Unknown or unsupported tag %d/0x%0X\n",
tag, tag);
return AVERROR_INVALIDDATA;
}
}
end:
if (s->bpp > 64U) {
av_log(s->avctx, AV_LOG_ERROR,
"This format is not supported (bpp=%d, %d components)\n",
s->bpp, count);
s->bpp = 0;
return AVERROR_INVALIDDATA;
}
bytestream2_seek(&s->gb, start, SEEK_SET);
return 0;
}
static int decode_frame(AVCodecContext *avctx, AVFrame *p,
int *got_frame, AVPacket *avpkt)
{
TiffContext *const s = avctx->priv_data;
unsigned off, last_off;
int le, ret, plane, planes;
int i, j, entries, stride;
unsigned soff, ssize;
uint8_t *dst;
GetByteContext stripsizes;
GetByteContext stripdata;
int retry_for_subifd, retry_for_page;
int is_dng;
int has_tile_bits, has_strip_bits;
bytestream2_init(&s->gb, avpkt->data, avpkt->size);
// parse image header
if ((ret = ff_tdecode_header(&s->gb, &le, &off))) {
av_log(avctx, AV_LOG_ERROR, "Invalid TIFF header\n");
return ret;
} else if (off >= UINT_MAX - 14 || avpkt->size < off + 14) {
av_log(avctx, AV_LOG_ERROR, "IFD offset is greater than image size\n");
return AVERROR_INVALIDDATA;
}
s->le = le;
// TIFF_BPP is not a required tag and defaults to 1
s->tiff_type = TIFF_TYPE_TIFF;
again:
s->is_thumbnail = 0;
s->bppcount = s->bpp = 1;
s->photometric = TIFF_PHOTOMETRIC_NONE;
s->compr = TIFF_RAW;
s->fill_order = 0;
s->white_level = 0;
s->is_bayer = 0;
s->is_tiled = 0;
s->is_jpeg = 0;
s->cur_page = 0;
s->last_tag = 0;
for (i = 0; i < 65536; i++)
s->dng_lut[i] = i;
free_geotags(s);
// Reset these offsets so we can tell if they were set this frame
s->stripsizesoff = s->strippos = 0;
/* parse image file directory */
bytestream2_seek(&s->gb, off, SEEK_SET);
entries = ff_tget_short(&s->gb, le);
if (bytestream2_get_bytes_left(&s->gb) < entries * 12)
return AVERROR_INVALIDDATA;
for (i = 0; i < entries; i++) {
if ((ret = tiff_decode_tag(s, p)) < 0)
return ret;
}
if (s->get_thumbnail && !s->is_thumbnail) {
av_log(avctx, AV_LOG_INFO, "No embedded thumbnail present\n");
return AVERROR_EOF;
}
/** whether we should process this IFD's SubIFD */
retry_for_subifd = s->sub_ifd && (s->get_subimage || (!s->get_thumbnail && s->is_thumbnail));
/** whether we should process this multi-page IFD's next page */
retry_for_page = s->get_page && s->cur_page + 1 < s->get_page; // get_page is 1-indexed
last_off = off;
if (retry_for_page) {
// set offset to the next IFD
off = ff_tget_long(&s->gb, le);
} else if (retry_for_subifd) {
// set offset to the SubIFD
off = s->sub_ifd;
}
if (retry_for_subifd || retry_for_page) {
if (!off) {
av_log(avctx, AV_LOG_ERROR, "Requested entry not found\n");
return AVERROR_INVALIDDATA;
}
if (off <= last_off) {
avpriv_request_sample(s->avctx, "non increasing IFD offset");
return AVERROR_INVALIDDATA;
}
if (off >= UINT_MAX - 14 || avpkt->size < off + 14) {
av_log(avctx, AV_LOG_ERROR, "IFD offset is greater than image size\n");
return AVERROR_INVALIDDATA;
}
s->sub_ifd = 0;
goto again;
}
/* At this point we've decided on which (Sub)IFD to process */
is_dng = (s->tiff_type == TIFF_TYPE_DNG || s->tiff_type == TIFF_TYPE_CINEMADNG);
for (i = 0; i<s->geotag_count; i++) {
const char *keyname = get_geokey_name(s->geotags[i].key);
if (!keyname) {
av_log(avctx, AV_LOG_WARNING, "Unknown or unsupported GeoTIFF key %d\n", s->geotags[i].key);
continue;
}
if (get_geokey_type(s->geotags[i].key) != s->geotags[i].type) {
av_log(avctx, AV_LOG_WARNING, "Type of GeoTIFF key %d is wrong\n", s->geotags[i].key);
continue;
}
ret = av_dict_set(&p->metadata, keyname, s->geotags[i].val, 0);
if (ret<0) {
av_log(avctx, AV_LOG_ERROR, "Writing metadata with key '%s' failed\n", keyname);
return ret;
}
}
if (is_dng) {
int bps;
if (s->bpp % s->bppcount)
return AVERROR_INVALIDDATA;
bps = s->bpp / s->bppcount;
if (bps < 8 || bps > 32)
return AVERROR_INVALIDDATA;
if (s->white_level == 0)
s->white_level = (1LL << bps) - 1; /* Default value as per the spec */
if (s->white_level <= s->black_level) {
av_log(avctx, AV_LOG_ERROR, "BlackLevel (%"PRId32") must be less than WhiteLevel (%"PRId32")\n",
s->black_level, s->white_level);
return AVERROR_INVALIDDATA;
}
if (s->planar)
return AVERROR_PATCHWELCOME;
}
if (!s->is_tiled && !s->strippos && !s->stripoff) {
av_log(avctx, AV_LOG_ERROR, "Image data is missing\n");
return AVERROR_INVALIDDATA;
}
has_tile_bits = s->is_tiled || s->tile_byte_counts_offset || s->tile_offsets_offset || s->tile_width || s->tile_length || s->tile_count;
has_strip_bits = s->strippos || s->strips || s->stripoff || s->rps || s->sot || s->sstype || s->stripsize || s->stripsizesoff;
if (has_tile_bits && has_strip_bits) {
int tiled_dng = s->is_tiled && is_dng;
av_log(avctx, tiled_dng ? AV_LOG_WARNING : AV_LOG_ERROR, "Tiled TIFF is not allowed to strip\n");
if (!tiled_dng)
return AVERROR_INVALIDDATA;
}
/* now we have the data and may start decoding */
if ((ret = init_image(s, p)) < 0)
return ret;
if (!s->is_tiled || has_strip_bits) {
if (s->strips == 1 && !s->stripsize) {
av_log(avctx, AV_LOG_WARNING, "Image data size missing\n");
s->stripsize = avpkt->size - s->stripoff;
}
if (s->stripsizesoff) {
if (s->stripsizesoff >= (unsigned)avpkt->size)
return AVERROR_INVALIDDATA;
bytestream2_init(&stripsizes, avpkt->data + s->stripsizesoff,
avpkt->size - s->stripsizesoff);
}
if (s->strippos) {
if (s->strippos >= (unsigned)avpkt->size)
return AVERROR_INVALIDDATA;
bytestream2_init(&stripdata, avpkt->data + s->strippos,
avpkt->size - s->strippos);
}
if (s->rps <= 0 || s->rps % s->subsampling[1]) {
av_log(avctx, AV_LOG_ERROR, "rps %d invalid\n", s->rps);
return AVERROR_INVALIDDATA;
}
}
if (s->photometric == TIFF_PHOTOMETRIC_LINEAR_RAW ||
s->photometric == TIFF_PHOTOMETRIC_CFA) {
p->color_trc = AVCOL_TRC_LINEAR;
} else if (s->photometric == TIFF_PHOTOMETRIC_BLACK_IS_ZERO) {
p->color_trc = AVCOL_TRC_GAMMA22;
}
/* Handle DNG images with JPEG-compressed tiles */
if (is_dng && s->is_tiled) {
if (!s->is_jpeg) {
avpriv_report_missing_feature(avctx, "DNG uncompressed tiled images");
return AVERROR_PATCHWELCOME;
} else if (!s->is_bayer) {
avpriv_report_missing_feature(avctx, "DNG JPG-compressed tiled non-bayer-encoded images");
return AVERROR_PATCHWELCOME;
} else {
if ((ret = dng_decode_tiles(avctx, p, avpkt)) > 0)
*got_frame = 1;
return ret;
}
}
/* Handle TIFF images and DNG images with uncompressed strips (non-tiled) */
planes = s->planar ? s->bppcount : 1;
for (plane = 0; plane < planes; plane++) {
uint8_t *five_planes = NULL;
int remaining = avpkt->size;
int decoded_height;
stride = p->linesize[plane];
dst = p->data[plane];
if (s->photometric == TIFF_PHOTOMETRIC_SEPARATED &&
s->avctx->pix_fmt == AV_PIX_FMT_RGBA) {
stride = stride * 5 / 4;
five_planes =
dst = av_malloc(stride * s->height);
if (!dst)
return AVERROR(ENOMEM);
}
for (i = 0; i < s->height; i += s->rps) {
if (i)
dst += s->rps * stride;
if (s->stripsizesoff)
ssize = ff_tget(&stripsizes, s->sstype, le);
else
ssize = s->stripsize;
if (s->strippos)
soff = ff_tget(&stripdata, s->sot, le);
else
soff = s->stripoff;
if (soff > avpkt->size || ssize > avpkt->size - soff || ssize > remaining) {
av_log(avctx, AV_LOG_ERROR, "Invalid strip size/offset\n");
av_freep(&five_planes);
return AVERROR_INVALIDDATA;
}
remaining -= ssize;
if ((ret = tiff_unpack_strip(s, p, dst, stride, avpkt->data + soff, ssize, i,
FFMIN(s->rps, s->height - i))) < 0) {
if (avctx->err_recognition & AV_EF_EXPLODE) {
av_freep(&five_planes);
return ret;
}
break;
}
}
decoded_height = FFMIN(i, s->height);
if (s->predictor == 2) {
if (s->photometric == TIFF_PHOTOMETRIC_YCBCR) {
av_log(s->avctx, AV_LOG_ERROR, "predictor == 2 with YUV is unsupported");
return AVERROR_PATCHWELCOME;
}
dst = five_planes ? five_planes : p->data[plane];
soff = s->bpp >> 3;
if (s->planar)
soff = FFMAX(soff / s->bppcount, 1);
ssize = s->width * soff;
if (s->avctx->pix_fmt == AV_PIX_FMT_RGB48LE ||
s->avctx->pix_fmt == AV_PIX_FMT_RGBA64LE ||
s->avctx->pix_fmt == AV_PIX_FMT_GRAY16LE ||
s->avctx->pix_fmt == AV_PIX_FMT_YA16LE ||
s->avctx->pix_fmt == AV_PIX_FMT_GBRP16LE ||
s->avctx->pix_fmt == AV_PIX_FMT_GBRAP16LE) {
for (i = 0; i < decoded_height; i++) {
for (j = soff; j < ssize; j += 2)
AV_WL16(dst + j, AV_RL16(dst + j) + AV_RL16(dst + j - soff));
dst += stride;
}
} else if (s->avctx->pix_fmt == AV_PIX_FMT_RGB48BE ||
s->avctx->pix_fmt == AV_PIX_FMT_RGBA64BE ||
s->avctx->pix_fmt == AV_PIX_FMT_GRAY16BE ||
s->avctx->pix_fmt == AV_PIX_FMT_YA16BE ||
s->avctx->pix_fmt == AV_PIX_FMT_GBRP16BE ||
s->avctx->pix_fmt == AV_PIX_FMT_GBRAP16BE) {
for (i = 0; i < decoded_height; i++) {
for (j = soff; j < ssize; j += 2)
AV_WB16(dst + j, AV_RB16(dst + j) + AV_RB16(dst + j - soff));
dst += stride;
}
} else {
for (i = 0; i < decoded_height; i++) {
for (j = soff; j < ssize; j++)
dst[j] += dst[j - soff];
dst += stride;
}
}
}
if (s->photometric == TIFF_PHOTOMETRIC_WHITE_IS_ZERO) {
int c = (s->avctx->pix_fmt == AV_PIX_FMT_PAL8 ? (1<<s->bpp) - 1 : 255);
dst = p->data[plane];
for (i = 0; i < s->height; i++) {
for (j = 0; j < stride; j++)
dst[j] = c - dst[j];
dst += stride;
}
}
if (s->photometric == TIFF_PHOTOMETRIC_SEPARATED &&
(s->avctx->pix_fmt == AV_PIX_FMT_RGB0 || s->avctx->pix_fmt == AV_PIX_FMT_RGBA)) {
int x = s->avctx->pix_fmt == AV_PIX_FMT_RGB0 ? 4 : 5;
uint8_t *src = five_planes ? five_planes : p->data[plane];
dst = p->data[plane];
for (i = 0; i < s->height; i++) {
for (j = 0; j < s->width; j++) {
int k = 255 - src[x * j + 3];
int r = (255 - src[x * j ]) * k;
int g = (255 - src[x * j + 1]) * k;
int b = (255 - src[x * j + 2]) * k;
dst[4 * j ] = r * 257 >> 16;
dst[4 * j + 1] = g * 257 >> 16;
dst[4 * j + 2] = b * 257 >> 16;
dst[4 * j + 3] = s->avctx->pix_fmt == AV_PIX_FMT_RGBA ? src[x * j + 4] : 255;
}
src += stride;
dst += p->linesize[plane];
}
av_freep(&five_planes);
} else if (s->photometric == TIFF_PHOTOMETRIC_SEPARATED &&
s->avctx->pix_fmt == AV_PIX_FMT_RGBA64BE) {
dst = p->data[plane];
for (i = 0; i < s->height; i++) {
for (j = 0; j < s->width; j++) {
uint64_t k = 65535 - AV_RB16(dst + 8 * j + 6);
uint64_t r = (65535 - AV_RB16(dst + 8 * j )) * k;
uint64_t g = (65535 - AV_RB16(dst + 8 * j + 2)) * k;
uint64_t b = (65535 - AV_RB16(dst + 8 * j + 4)) * k;
AV_WB16(dst + 8 * j , r * 65537 >> 32);
AV_WB16(dst + 8 * j + 2, g * 65537 >> 32);
AV_WB16(dst + 8 * j + 4, b * 65537 >> 32);
AV_WB16(dst + 8 * j + 6, 65535);
}
dst += p->linesize[plane];
}
}
}
if (s->planar && s->bppcount > 2) {
FFSWAP(uint8_t*, p->data[0], p->data[2]);
FFSWAP(int, p->linesize[0], p->linesize[2]);
FFSWAP(uint8_t*, p->data[0], p->data[1]);
FFSWAP(int, p->linesize[0], p->linesize[1]);
}
if (s->is_bayer && s->white_level && s->bpp == 16 && !is_dng) {
uint16_t *dst = (uint16_t *)p->data[0];
for (i = 0; i < s->height; i++) {
for (j = 0; j < s->width; j++)
dst[j] = FFMIN((dst[j] / (float)s->white_level) * 65535, 65535);
dst += stride / 2;
}
}
*got_frame = 1;
return avpkt->size;
}
static av_cold int tiff_init(AVCodecContext *avctx)
{
TiffContext *s = avctx->priv_data;
const AVCodec *codec;
int ret;
s->width = 0;
s->height = 0;
s->subsampling[0] =
s->subsampling[1] = 1;
s->avctx = avctx;
ff_lzw_decode_open(&s->lzw);
if (!s->lzw)
return AVERROR(ENOMEM);
ff_ccitt_unpack_init();
/* Allocate JPEG frame */
s->jpgframe = av_frame_alloc();
s->jpkt = av_packet_alloc();
if (!s->jpgframe || !s->jpkt)
return AVERROR(ENOMEM);
/* Prepare everything needed for JPEG decoding */
codec = avcodec_find_decoder(AV_CODEC_ID_MJPEG);
if (!codec)
return AVERROR_BUG;
s->avctx_mjpeg = avcodec_alloc_context3(codec);
if (!s->avctx_mjpeg)
return AVERROR(ENOMEM);
s->avctx_mjpeg->flags = avctx->flags;
s->avctx_mjpeg->flags2 = avctx->flags2;
s->avctx_mjpeg->dct_algo = avctx->dct_algo;
s->avctx_mjpeg->idct_algo = avctx->idct_algo;
s->avctx_mjpeg->max_pixels = avctx->max_pixels;
ret = avcodec_open2(s->avctx_mjpeg, codec, NULL);
if (ret < 0) {
return ret;
}
return 0;
}
static av_cold int tiff_end(AVCodecContext *avctx)
{
TiffContext *const s = avctx->priv_data;
free_geotags(s);
ff_lzw_decode_close(&s->lzw);
av_freep(&s->deinvert_buf);
s->deinvert_buf_size = 0;
av_freep(&s->yuv_line);
s->yuv_line_size = 0;
av_frame_free(&s->jpgframe);
av_packet_free(&s->jpkt);
avcodec_free_context(&s->avctx_mjpeg);
return 0;
}
#define OFFSET(x) offsetof(TiffContext, x)
static const AVOption tiff_options[] = {
{ "subimage", "decode subimage instead if available", OFFSET(get_subimage), AV_OPT_TYPE_BOOL, {.i64=0}, 0, 1, AV_OPT_FLAG_DECODING_PARAM | AV_OPT_FLAG_VIDEO_PARAM },
{ "thumbnail", "decode embedded thumbnail subimage instead if available", OFFSET(get_thumbnail), AV_OPT_TYPE_BOOL, {.i64=0}, 0, 1, AV_OPT_FLAG_DECODING_PARAM | AV_OPT_FLAG_VIDEO_PARAM },
{ "page", "page number of multi-page image to decode (starting from 1)", OFFSET(get_page), AV_OPT_TYPE_INT, {.i64=0}, 0, UINT16_MAX, AV_OPT_FLAG_DECODING_PARAM | AV_OPT_FLAG_VIDEO_PARAM },
{ NULL },
};
static const AVClass tiff_decoder_class = {
.class_name = "TIFF decoder",
.item_name = av_default_item_name,
.option = tiff_options,
.version = LIBAVUTIL_VERSION_INT,
};
const FFCodec ff_tiff_decoder = {
.p.name = "tiff",
.p.long_name = NULL_IF_CONFIG_SMALL("TIFF image"),
.p.type = AVMEDIA_TYPE_VIDEO,
.p.id = AV_CODEC_ID_TIFF,
.priv_data_size = sizeof(TiffContext),
.init = tiff_init,
.close = tiff_end,
FF_CODEC_DECODE_CB(decode_frame),
.p.capabilities = AV_CODEC_CAP_DR1 | AV_CODEC_CAP_FRAME_THREADS,
.caps_internal = FF_CODEC_CAP_INIT_CLEANUP | FF_CODEC_CAP_ICC_PROFILES,
.p.priv_class = &tiff_decoder_class,
};