mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-12-23 12:43:46 +02:00
4243da4ff4
This is possible, because every given FFCodec has to implement exactly one of these. Doing so decreases sizeof(FFCodec) and therefore decreases the size of the binary. Notice that in case of position-independent code the decrease is in .data.rel.ro, so that this translates to decreased memory consumption. Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@outlook.com>
1371 lines
42 KiB
C
1371 lines
42 KiB
C
/*
|
|
* Interplay MVE Video Decoder
|
|
* Copyright (C) 2003 The FFmpeg project
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
/**
|
|
* @file
|
|
* Interplay MVE Video Decoder by Mike Melanson (melanson@pcisys.net)
|
|
* For more information about the Interplay MVE format, visit:
|
|
* http://www.pcisys.net/~melanson/codecs/interplay-mve.txt
|
|
* This code is written in such a way that the identifiers match up
|
|
* with the encoding descriptions in the document.
|
|
*
|
|
* This decoder presently only supports a PAL8 output colorspace.
|
|
*
|
|
* An Interplay video frame consists of 2 parts: The decoding map and
|
|
* the video data. A demuxer must load these 2 parts together in a single
|
|
* buffer before sending it through the stream to this decoder.
|
|
*/
|
|
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
#include <string.h>
|
|
|
|
#include "libavutil/intreadwrite.h"
|
|
|
|
#define BITSTREAM_READER_LE
|
|
#include "avcodec.h"
|
|
#include "bytestream.h"
|
|
#include "codec_internal.h"
|
|
#include "decode.h"
|
|
#include "get_bits.h"
|
|
#include "hpeldsp.h"
|
|
#include "internal.h"
|
|
|
|
#define PALETTE_COUNT 256
|
|
|
|
typedef struct IpvideoContext {
|
|
|
|
AVCodecContext *avctx;
|
|
HpelDSPContext hdsp;
|
|
AVFrame *second_last_frame;
|
|
AVFrame *last_frame;
|
|
|
|
/* For format 0x10 */
|
|
AVFrame *cur_decode_frame;
|
|
AVFrame *prev_decode_frame;
|
|
|
|
const unsigned char *decoding_map;
|
|
int decoding_map_size;
|
|
const unsigned char *skip_map;
|
|
int skip_map_size;
|
|
|
|
int is_16bpp;
|
|
GetByteContext stream_ptr, mv_ptr;
|
|
unsigned char *pixel_ptr;
|
|
int line_inc;
|
|
int stride;
|
|
int upper_motion_limit_offset;
|
|
|
|
uint32_t pal[256];
|
|
} IpvideoContext;
|
|
|
|
static int copy_from(IpvideoContext *s, AVFrame *src, AVFrame *dst, int delta_x, int delta_y)
|
|
{
|
|
int width = dst->width;
|
|
int current_offset = s->pixel_ptr - dst->data[0];
|
|
int x = (current_offset % dst->linesize[0]) / (1 + s->is_16bpp);
|
|
int y = current_offset / dst->linesize[0];
|
|
int dx = delta_x + x - ((delta_x + x >= width) - (delta_x + x < 0)) * width;
|
|
int dy = delta_y + y + (delta_x + x >= width) - (delta_x + x < 0);
|
|
int motion_offset = dy * src->linesize[0] + dx * (1 + s->is_16bpp);
|
|
|
|
if (motion_offset < 0) {
|
|
av_log(s->avctx, AV_LOG_ERROR, "motion offset < 0 (%d)\n", motion_offset);
|
|
return AVERROR_INVALIDDATA;
|
|
} else if (motion_offset > s->upper_motion_limit_offset) {
|
|
av_log(s->avctx, AV_LOG_ERROR, "motion offset above limit (%d >= %d)\n",
|
|
motion_offset, s->upper_motion_limit_offset);
|
|
return AVERROR_INVALIDDATA;
|
|
}
|
|
if (!src->data[0]) {
|
|
av_log(s->avctx, AV_LOG_ERROR, "Invalid decode type, corrupted header?\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
s->hdsp.put_pixels_tab[!s->is_16bpp][0](s->pixel_ptr, src->data[0] + motion_offset,
|
|
dst->linesize[0], 8);
|
|
return 0;
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0x0(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
return copy_from(s, s->last_frame, frame, 0, 0);
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0x1(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
return copy_from(s, s->second_last_frame, frame, 0, 0);
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0x2(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
unsigned char B;
|
|
int x, y;
|
|
|
|
/* copy block from 2 frames ago using a motion vector; need 1 more byte */
|
|
if (!s->is_16bpp) {
|
|
B = bytestream2_get_byte(&s->stream_ptr);
|
|
} else {
|
|
B = bytestream2_get_byte(&s->mv_ptr);
|
|
}
|
|
|
|
if (B < 56) {
|
|
x = 8 + (B % 7);
|
|
y = B / 7;
|
|
} else {
|
|
x = -14 + ((B - 56) % 29);
|
|
y = 8 + ((B - 56) / 29);
|
|
}
|
|
|
|
ff_tlog(s->avctx, "motion byte = %d, (x, y) = (%d, %d)\n", B, x, y);
|
|
return copy_from(s, s->second_last_frame, frame, x, y);
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0x3(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
unsigned char B;
|
|
int x, y;
|
|
|
|
/* copy 8x8 block from current frame from an up/left block */
|
|
|
|
/* need 1 more byte for motion */
|
|
if (!s->is_16bpp) {
|
|
B = bytestream2_get_byte(&s->stream_ptr);
|
|
} else {
|
|
B = bytestream2_get_byte(&s->mv_ptr);
|
|
}
|
|
|
|
if (B < 56) {
|
|
x = -(8 + (B % 7));
|
|
y = -(B / 7);
|
|
} else {
|
|
x = -(-14 + ((B - 56) % 29));
|
|
y = -( 8 + ((B - 56) / 29));
|
|
}
|
|
|
|
ff_tlog(s->avctx, "motion byte = %d, (x, y) = (%d, %d)\n", B, x, y);
|
|
return copy_from(s, frame, frame, x, y);
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0x4(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
int x, y;
|
|
unsigned char B, BL, BH;
|
|
|
|
/* copy a block from the previous frame; need 1 more byte */
|
|
if (!s->is_16bpp) {
|
|
B = bytestream2_get_byte(&s->stream_ptr);
|
|
} else {
|
|
B = bytestream2_get_byte(&s->mv_ptr);
|
|
}
|
|
|
|
BL = B & 0x0F;
|
|
BH = (B >> 4) & 0x0F;
|
|
x = -8 + BL;
|
|
y = -8 + BH;
|
|
|
|
ff_tlog(s->avctx, "motion byte = %d, (x, y) = (%d, %d)\n", B, x, y);
|
|
return copy_from(s, s->last_frame, frame, x, y);
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0x5(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
signed char x, y;
|
|
|
|
/* copy a block from the previous frame using an expanded range;
|
|
* need 2 more bytes */
|
|
x = bytestream2_get_byte(&s->stream_ptr);
|
|
y = bytestream2_get_byte(&s->stream_ptr);
|
|
|
|
ff_tlog(s->avctx, "motion bytes = %d, %d\n", x, y);
|
|
return copy_from(s, s->last_frame, frame, x, y);
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0x6(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
/* mystery opcode? skip multiple blocks? */
|
|
av_log(s->avctx, AV_LOG_ERROR, "Help! Mystery opcode 0x6 seen\n");
|
|
|
|
/* report success */
|
|
return 0;
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0x7(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
int x, y;
|
|
unsigned char P[2];
|
|
unsigned int flags;
|
|
|
|
if (bytestream2_get_bytes_left(&s->stream_ptr) < 4) {
|
|
av_log(s->avctx, AV_LOG_ERROR, "too little data for opcode 0x7\n");
|
|
return AVERROR_INVALIDDATA;
|
|
}
|
|
|
|
/* 2-color encoding */
|
|
P[0] = bytestream2_get_byte(&s->stream_ptr);
|
|
P[1] = bytestream2_get_byte(&s->stream_ptr);
|
|
|
|
if (P[0] <= P[1]) {
|
|
|
|
/* need 8 more bytes from the stream */
|
|
for (y = 0; y < 8; y++) {
|
|
flags = bytestream2_get_byte(&s->stream_ptr) | 0x100;
|
|
for (; flags != 1; flags >>= 1)
|
|
*s->pixel_ptr++ = P[flags & 1];
|
|
s->pixel_ptr += s->line_inc;
|
|
}
|
|
|
|
} else {
|
|
|
|
/* need 2 more bytes from the stream */
|
|
flags = bytestream2_get_le16(&s->stream_ptr);
|
|
for (y = 0; y < 8; y += 2) {
|
|
for (x = 0; x < 8; x += 2, flags >>= 1) {
|
|
s->pixel_ptr[x ] =
|
|
s->pixel_ptr[x + 1 ] =
|
|
s->pixel_ptr[x + s->stride] =
|
|
s->pixel_ptr[x + 1 + s->stride] = P[flags & 1];
|
|
}
|
|
s->pixel_ptr += s->stride * 2;
|
|
}
|
|
}
|
|
|
|
/* report success */
|
|
return 0;
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0x8(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
int x, y;
|
|
unsigned char P[4];
|
|
unsigned int flags = 0;
|
|
|
|
if (bytestream2_get_bytes_left(&s->stream_ptr) < 12) {
|
|
av_log(s->avctx, AV_LOG_ERROR, "too little data for opcode 0x8\n");
|
|
return AVERROR_INVALIDDATA;
|
|
}
|
|
|
|
/* 2-color encoding for each 4x4 quadrant, or 2-color encoding on
|
|
* either top and bottom or left and right halves */
|
|
P[0] = bytestream2_get_byte(&s->stream_ptr);
|
|
P[1] = bytestream2_get_byte(&s->stream_ptr);
|
|
|
|
if (P[0] <= P[1]) {
|
|
for (y = 0; y < 16; y++) {
|
|
// new values for each 4x4 block
|
|
if (!(y & 3)) {
|
|
if (y) {
|
|
P[0] = bytestream2_get_byte(&s->stream_ptr);
|
|
P[1] = bytestream2_get_byte(&s->stream_ptr);
|
|
}
|
|
flags = bytestream2_get_le16(&s->stream_ptr);
|
|
}
|
|
|
|
for (x = 0; x < 4; x++, flags >>= 1)
|
|
*s->pixel_ptr++ = P[flags & 1];
|
|
s->pixel_ptr += s->stride - 4;
|
|
// switch to right half
|
|
if (y == 7) s->pixel_ptr -= 8 * s->stride - 4;
|
|
}
|
|
|
|
} else {
|
|
flags = bytestream2_get_le32(&s->stream_ptr);
|
|
P[2] = bytestream2_get_byte(&s->stream_ptr);
|
|
P[3] = bytestream2_get_byte(&s->stream_ptr);
|
|
|
|
if (P[2] <= P[3]) {
|
|
|
|
/* vertical split; left & right halves are 2-color encoded */
|
|
|
|
for (y = 0; y < 16; y++) {
|
|
for (x = 0; x < 4; x++, flags >>= 1)
|
|
*s->pixel_ptr++ = P[flags & 1];
|
|
s->pixel_ptr += s->stride - 4;
|
|
// switch to right half
|
|
if (y == 7) {
|
|
s->pixel_ptr -= 8 * s->stride - 4;
|
|
P[0] = P[2];
|
|
P[1] = P[3];
|
|
flags = bytestream2_get_le32(&s->stream_ptr);
|
|
}
|
|
}
|
|
|
|
} else {
|
|
|
|
/* horizontal split; top & bottom halves are 2-color encoded */
|
|
|
|
for (y = 0; y < 8; y++) {
|
|
if (y == 4) {
|
|
P[0] = P[2];
|
|
P[1] = P[3];
|
|
flags = bytestream2_get_le32(&s->stream_ptr);
|
|
}
|
|
|
|
for (x = 0; x < 8; x++, flags >>= 1)
|
|
*s->pixel_ptr++ = P[flags & 1];
|
|
s->pixel_ptr += s->line_inc;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* report success */
|
|
return 0;
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0x9(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
int x, y;
|
|
unsigned char P[4];
|
|
|
|
if (bytestream2_get_bytes_left(&s->stream_ptr) < 8) {
|
|
av_log(s->avctx, AV_LOG_ERROR, "too little data for opcode 0x9\n");
|
|
return AVERROR_INVALIDDATA;
|
|
}
|
|
|
|
/* 4-color encoding */
|
|
bytestream2_get_buffer(&s->stream_ptr, P, 4);
|
|
|
|
if (P[0] <= P[1]) {
|
|
if (P[2] <= P[3]) {
|
|
|
|
/* 1 of 4 colors for each pixel, need 16 more bytes */
|
|
for (y = 0; y < 8; y++) {
|
|
/* get the next set of 8 2-bit flags */
|
|
int flags = bytestream2_get_le16(&s->stream_ptr);
|
|
for (x = 0; x < 8; x++, flags >>= 2)
|
|
*s->pixel_ptr++ = P[flags & 0x03];
|
|
s->pixel_ptr += s->line_inc;
|
|
}
|
|
|
|
} else {
|
|
uint32_t flags;
|
|
|
|
/* 1 of 4 colors for each 2x2 block, need 4 more bytes */
|
|
flags = bytestream2_get_le32(&s->stream_ptr);
|
|
|
|
for (y = 0; y < 8; y += 2) {
|
|
for (x = 0; x < 8; x += 2, flags >>= 2) {
|
|
s->pixel_ptr[x ] =
|
|
s->pixel_ptr[x + 1 ] =
|
|
s->pixel_ptr[x + s->stride] =
|
|
s->pixel_ptr[x + 1 + s->stride] = P[flags & 0x03];
|
|
}
|
|
s->pixel_ptr += s->stride * 2;
|
|
}
|
|
|
|
}
|
|
} else {
|
|
uint64_t flags;
|
|
|
|
/* 1 of 4 colors for each 2x1 or 1x2 block, need 8 more bytes */
|
|
flags = bytestream2_get_le64(&s->stream_ptr);
|
|
if (P[2] <= P[3]) {
|
|
for (y = 0; y < 8; y++) {
|
|
for (x = 0; x < 8; x += 2, flags >>= 2) {
|
|
s->pixel_ptr[x ] =
|
|
s->pixel_ptr[x + 1] = P[flags & 0x03];
|
|
}
|
|
s->pixel_ptr += s->stride;
|
|
}
|
|
} else {
|
|
for (y = 0; y < 8; y += 2) {
|
|
for (x = 0; x < 8; x++, flags >>= 2) {
|
|
s->pixel_ptr[x ] =
|
|
s->pixel_ptr[x + s->stride] = P[flags & 0x03];
|
|
}
|
|
s->pixel_ptr += s->stride * 2;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* report success */
|
|
return 0;
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0xA(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
int x, y;
|
|
unsigned char P[8];
|
|
int flags = 0;
|
|
|
|
if (bytestream2_get_bytes_left(&s->stream_ptr) < 16) {
|
|
av_log(s->avctx, AV_LOG_ERROR, "too little data for opcode 0xA\n");
|
|
return AVERROR_INVALIDDATA;
|
|
}
|
|
|
|
bytestream2_get_buffer(&s->stream_ptr, P, 4);
|
|
|
|
/* 4-color encoding for each 4x4 quadrant, or 4-color encoding on
|
|
* either top and bottom or left and right halves */
|
|
if (P[0] <= P[1]) {
|
|
|
|
/* 4-color encoding for each quadrant; need 32 bytes */
|
|
for (y = 0; y < 16; y++) {
|
|
// new values for each 4x4 block
|
|
if (!(y & 3)) {
|
|
if (y) bytestream2_get_buffer(&s->stream_ptr, P, 4);
|
|
flags = bytestream2_get_le32(&s->stream_ptr);
|
|
}
|
|
|
|
for (x = 0; x < 4; x++, flags >>= 2)
|
|
*s->pixel_ptr++ = P[flags & 0x03];
|
|
|
|
s->pixel_ptr += s->stride - 4;
|
|
// switch to right half
|
|
if (y == 7) s->pixel_ptr -= 8 * s->stride - 4;
|
|
}
|
|
|
|
} else {
|
|
// vertical split?
|
|
int vert;
|
|
uint64_t flags = bytestream2_get_le64(&s->stream_ptr);
|
|
|
|
bytestream2_get_buffer(&s->stream_ptr, P + 4, 4);
|
|
vert = P[4] <= P[5];
|
|
|
|
/* 4-color encoding for either left and right or top and bottom
|
|
* halves */
|
|
|
|
for (y = 0; y < 16; y++) {
|
|
for (x = 0; x < 4; x++, flags >>= 2)
|
|
*s->pixel_ptr++ = P[flags & 0x03];
|
|
|
|
if (vert) {
|
|
s->pixel_ptr += s->stride - 4;
|
|
// switch to right half
|
|
if (y == 7) s->pixel_ptr -= 8 * s->stride - 4;
|
|
} else if (y & 1) s->pixel_ptr += s->line_inc;
|
|
|
|
// load values for second half
|
|
if (y == 7) {
|
|
memcpy(P, P + 4, 4);
|
|
flags = bytestream2_get_le64(&s->stream_ptr);
|
|
}
|
|
}
|
|
}
|
|
|
|
/* report success */
|
|
return 0;
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0xB(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
int y;
|
|
|
|
/* 64-color encoding (each pixel in block is a different color) */
|
|
for (y = 0; y < 8; y++) {
|
|
bytestream2_get_buffer(&s->stream_ptr, s->pixel_ptr, 8);
|
|
s->pixel_ptr += s->stride;
|
|
}
|
|
|
|
/* report success */
|
|
return 0;
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0xC(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
int x, y;
|
|
|
|
/* 16-color block encoding: each 2x2 block is a different color */
|
|
for (y = 0; y < 8; y += 2) {
|
|
for (x = 0; x < 8; x += 2) {
|
|
s->pixel_ptr[x ] =
|
|
s->pixel_ptr[x + 1 ] =
|
|
s->pixel_ptr[x + s->stride] =
|
|
s->pixel_ptr[x + 1 + s->stride] = bytestream2_get_byte(&s->stream_ptr);
|
|
}
|
|
s->pixel_ptr += s->stride * 2;
|
|
}
|
|
|
|
/* report success */
|
|
return 0;
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0xD(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
int y;
|
|
unsigned char P[2];
|
|
|
|
if (bytestream2_get_bytes_left(&s->stream_ptr) < 4) {
|
|
av_log(s->avctx, AV_LOG_ERROR, "too little data for opcode 0xD\n");
|
|
return AVERROR_INVALIDDATA;
|
|
}
|
|
|
|
/* 4-color block encoding: each 4x4 block is a different color */
|
|
for (y = 0; y < 8; y++) {
|
|
if (!(y & 3)) {
|
|
P[0] = bytestream2_get_byte(&s->stream_ptr);
|
|
P[1] = bytestream2_get_byte(&s->stream_ptr);
|
|
}
|
|
memset(s->pixel_ptr, P[0], 4);
|
|
memset(s->pixel_ptr + 4, P[1], 4);
|
|
s->pixel_ptr += s->stride;
|
|
}
|
|
|
|
/* report success */
|
|
return 0;
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0xE(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
int y;
|
|
unsigned char pix;
|
|
|
|
/* 1-color encoding: the whole block is 1 solid color */
|
|
pix = bytestream2_get_byte(&s->stream_ptr);
|
|
|
|
for (y = 0; y < 8; y++) {
|
|
memset(s->pixel_ptr, pix, 8);
|
|
s->pixel_ptr += s->stride;
|
|
}
|
|
|
|
/* report success */
|
|
return 0;
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0xF(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
int x, y;
|
|
unsigned char sample[2];
|
|
|
|
/* dithered encoding */
|
|
sample[0] = bytestream2_get_byte(&s->stream_ptr);
|
|
sample[1] = bytestream2_get_byte(&s->stream_ptr);
|
|
|
|
for (y = 0; y < 8; y++) {
|
|
for (x = 0; x < 8; x += 2) {
|
|
*s->pixel_ptr++ = sample[ y & 1 ];
|
|
*s->pixel_ptr++ = sample[!(y & 1)];
|
|
}
|
|
s->pixel_ptr += s->line_inc;
|
|
}
|
|
|
|
/* report success */
|
|
return 0;
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0x6_16(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
signed char x, y;
|
|
|
|
/* copy a block from the second last frame using an expanded range */
|
|
x = bytestream2_get_byte(&s->stream_ptr);
|
|
y = bytestream2_get_byte(&s->stream_ptr);
|
|
|
|
ff_tlog(s->avctx, "motion bytes = %d, %d\n", x, y);
|
|
return copy_from(s, s->second_last_frame, frame, x, y);
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0x7_16(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
int x, y;
|
|
uint16_t P[2];
|
|
unsigned int flags;
|
|
uint16_t *pixel_ptr = (uint16_t*)s->pixel_ptr;
|
|
|
|
/* 2-color encoding */
|
|
P[0] = bytestream2_get_le16(&s->stream_ptr);
|
|
P[1] = bytestream2_get_le16(&s->stream_ptr);
|
|
|
|
if (!(P[0] & 0x8000)) {
|
|
|
|
for (y = 0; y < 8; y++) {
|
|
flags = bytestream2_get_byte(&s->stream_ptr) | 0x100;
|
|
for (; flags != 1; flags >>= 1)
|
|
*pixel_ptr++ = P[flags & 1];
|
|
pixel_ptr += s->line_inc;
|
|
}
|
|
|
|
} else {
|
|
|
|
flags = bytestream2_get_le16(&s->stream_ptr);
|
|
for (y = 0; y < 8; y += 2) {
|
|
for (x = 0; x < 8; x += 2, flags >>= 1) {
|
|
pixel_ptr[x ] =
|
|
pixel_ptr[x + 1 ] =
|
|
pixel_ptr[x + s->stride] =
|
|
pixel_ptr[x + 1 + s->stride] = P[flags & 1];
|
|
}
|
|
pixel_ptr += s->stride * 2;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0x8_16(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
int x, y;
|
|
uint16_t P[4];
|
|
unsigned int flags = 0;
|
|
uint16_t *pixel_ptr = (uint16_t*)s->pixel_ptr;
|
|
|
|
/* 2-color encoding for each 4x4 quadrant, or 2-color encoding on
|
|
* either top and bottom or left and right halves */
|
|
P[0] = bytestream2_get_le16(&s->stream_ptr);
|
|
P[1] = bytestream2_get_le16(&s->stream_ptr);
|
|
|
|
if (!(P[0] & 0x8000)) {
|
|
|
|
for (y = 0; y < 16; y++) {
|
|
// new values for each 4x4 block
|
|
if (!(y & 3)) {
|
|
if (y) {
|
|
P[0] = bytestream2_get_le16(&s->stream_ptr);
|
|
P[1] = bytestream2_get_le16(&s->stream_ptr);
|
|
}
|
|
flags = bytestream2_get_le16(&s->stream_ptr);
|
|
}
|
|
|
|
for (x = 0; x < 4; x++, flags >>= 1)
|
|
*pixel_ptr++ = P[flags & 1];
|
|
pixel_ptr += s->stride - 4;
|
|
// switch to right half
|
|
if (y == 7) pixel_ptr -= 8 * s->stride - 4;
|
|
}
|
|
|
|
} else {
|
|
|
|
flags = bytestream2_get_le32(&s->stream_ptr);
|
|
P[2] = bytestream2_get_le16(&s->stream_ptr);
|
|
P[3] = bytestream2_get_le16(&s->stream_ptr);
|
|
|
|
if (!(P[2] & 0x8000)) {
|
|
|
|
/* vertical split; left & right halves are 2-color encoded */
|
|
|
|
for (y = 0; y < 16; y++) {
|
|
for (x = 0; x < 4; x++, flags >>= 1)
|
|
*pixel_ptr++ = P[flags & 1];
|
|
pixel_ptr += s->stride - 4;
|
|
// switch to right half
|
|
if (y == 7) {
|
|
pixel_ptr -= 8 * s->stride - 4;
|
|
P[0] = P[2];
|
|
P[1] = P[3];
|
|
flags = bytestream2_get_le32(&s->stream_ptr);
|
|
}
|
|
}
|
|
|
|
} else {
|
|
|
|
/* horizontal split; top & bottom halves are 2-color encoded */
|
|
|
|
for (y = 0; y < 8; y++) {
|
|
if (y == 4) {
|
|
P[0] = P[2];
|
|
P[1] = P[3];
|
|
flags = bytestream2_get_le32(&s->stream_ptr);
|
|
}
|
|
|
|
for (x = 0; x < 8; x++, flags >>= 1)
|
|
*pixel_ptr++ = P[flags & 1];
|
|
pixel_ptr += s->line_inc;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* report success */
|
|
return 0;
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0x9_16(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
int x, y;
|
|
uint16_t P[4];
|
|
uint16_t *pixel_ptr = (uint16_t*)s->pixel_ptr;
|
|
|
|
/* 4-color encoding */
|
|
for (x = 0; x < 4; x++)
|
|
P[x] = bytestream2_get_le16(&s->stream_ptr);
|
|
|
|
if (!(P[0] & 0x8000)) {
|
|
if (!(P[2] & 0x8000)) {
|
|
|
|
/* 1 of 4 colors for each pixel */
|
|
for (y = 0; y < 8; y++) {
|
|
/* get the next set of 8 2-bit flags */
|
|
int flags = bytestream2_get_le16(&s->stream_ptr);
|
|
for (x = 0; x < 8; x++, flags >>= 2)
|
|
*pixel_ptr++ = P[flags & 0x03];
|
|
pixel_ptr += s->line_inc;
|
|
}
|
|
|
|
} else {
|
|
uint32_t flags;
|
|
|
|
/* 1 of 4 colors for each 2x2 block */
|
|
flags = bytestream2_get_le32(&s->stream_ptr);
|
|
|
|
for (y = 0; y < 8; y += 2) {
|
|
for (x = 0; x < 8; x += 2, flags >>= 2) {
|
|
pixel_ptr[x ] =
|
|
pixel_ptr[x + 1 ] =
|
|
pixel_ptr[x + s->stride] =
|
|
pixel_ptr[x + 1 + s->stride] = P[flags & 0x03];
|
|
}
|
|
pixel_ptr += s->stride * 2;
|
|
}
|
|
|
|
}
|
|
} else {
|
|
uint64_t flags;
|
|
|
|
/* 1 of 4 colors for each 2x1 or 1x2 block */
|
|
flags = bytestream2_get_le64(&s->stream_ptr);
|
|
if (!(P[2] & 0x8000)) {
|
|
for (y = 0; y < 8; y++) {
|
|
for (x = 0; x < 8; x += 2, flags >>= 2) {
|
|
pixel_ptr[x ] =
|
|
pixel_ptr[x + 1] = P[flags & 0x03];
|
|
}
|
|
pixel_ptr += s->stride;
|
|
}
|
|
} else {
|
|
for (y = 0; y < 8; y += 2) {
|
|
for (x = 0; x < 8; x++, flags >>= 2) {
|
|
pixel_ptr[x ] =
|
|
pixel_ptr[x + s->stride] = P[flags & 0x03];
|
|
}
|
|
pixel_ptr += s->stride * 2;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* report success */
|
|
return 0;
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0xA_16(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
int x, y;
|
|
uint16_t P[8];
|
|
int flags = 0;
|
|
uint16_t *pixel_ptr = (uint16_t*)s->pixel_ptr;
|
|
|
|
for (x = 0; x < 4; x++)
|
|
P[x] = bytestream2_get_le16(&s->stream_ptr);
|
|
|
|
/* 4-color encoding for each 4x4 quadrant, or 4-color encoding on
|
|
* either top and bottom or left and right halves */
|
|
if (!(P[0] & 0x8000)) {
|
|
|
|
/* 4-color encoding for each quadrant */
|
|
for (y = 0; y < 16; y++) {
|
|
// new values for each 4x4 block
|
|
if (!(y & 3)) {
|
|
if (y)
|
|
for (x = 0; x < 4; x++)
|
|
P[x] = bytestream2_get_le16(&s->stream_ptr);
|
|
flags = bytestream2_get_le32(&s->stream_ptr);
|
|
}
|
|
|
|
for (x = 0; x < 4; x++, flags >>= 2)
|
|
*pixel_ptr++ = P[flags & 0x03];
|
|
|
|
pixel_ptr += s->stride - 4;
|
|
// switch to right half
|
|
if (y == 7) pixel_ptr -= 8 * s->stride - 4;
|
|
}
|
|
|
|
} else {
|
|
// vertical split?
|
|
int vert;
|
|
uint64_t flags = bytestream2_get_le64(&s->stream_ptr);
|
|
|
|
for (x = 4; x < 8; x++)
|
|
P[x] = bytestream2_get_le16(&s->stream_ptr);
|
|
vert = !(P[4] & 0x8000);
|
|
|
|
/* 4-color encoding for either left and right or top and bottom
|
|
* halves */
|
|
|
|
for (y = 0; y < 16; y++) {
|
|
for (x = 0; x < 4; x++, flags >>= 2)
|
|
*pixel_ptr++ = P[flags & 0x03];
|
|
|
|
if (vert) {
|
|
pixel_ptr += s->stride - 4;
|
|
// switch to right half
|
|
if (y == 7) pixel_ptr -= 8 * s->stride - 4;
|
|
} else if (y & 1) pixel_ptr += s->line_inc;
|
|
|
|
// load values for second half
|
|
if (y == 7) {
|
|
memcpy(P, P + 4, 8);
|
|
flags = bytestream2_get_le64(&s->stream_ptr);
|
|
}
|
|
}
|
|
}
|
|
|
|
/* report success */
|
|
return 0;
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0xB_16(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
int x, y;
|
|
uint16_t *pixel_ptr = (uint16_t*)s->pixel_ptr;
|
|
|
|
/* 64-color encoding (each pixel in block is a different color) */
|
|
for (y = 0; y < 8; y++) {
|
|
for (x = 0; x < 8; x++)
|
|
pixel_ptr[x] = bytestream2_get_le16(&s->stream_ptr);
|
|
pixel_ptr += s->stride;
|
|
}
|
|
|
|
/* report success */
|
|
return 0;
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0xC_16(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
int x, y;
|
|
uint16_t *pixel_ptr = (uint16_t*)s->pixel_ptr;
|
|
|
|
/* 16-color block encoding: each 2x2 block is a different color */
|
|
for (y = 0; y < 8; y += 2) {
|
|
for (x = 0; x < 8; x += 2) {
|
|
pixel_ptr[x ] =
|
|
pixel_ptr[x + 1 ] =
|
|
pixel_ptr[x + s->stride] =
|
|
pixel_ptr[x + 1 + s->stride] = bytestream2_get_le16(&s->stream_ptr);
|
|
}
|
|
pixel_ptr += s->stride * 2;
|
|
}
|
|
|
|
/* report success */
|
|
return 0;
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0xD_16(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
int x, y;
|
|
uint16_t P[2];
|
|
uint16_t *pixel_ptr = (uint16_t*)s->pixel_ptr;
|
|
|
|
/* 4-color block encoding: each 4x4 block is a different color */
|
|
for (y = 0; y < 8; y++) {
|
|
if (!(y & 3)) {
|
|
P[0] = bytestream2_get_le16(&s->stream_ptr);
|
|
P[1] = bytestream2_get_le16(&s->stream_ptr);
|
|
}
|
|
for (x = 0; x < 8; x++)
|
|
pixel_ptr[x] = P[x >> 2];
|
|
pixel_ptr += s->stride;
|
|
}
|
|
|
|
/* report success */
|
|
return 0;
|
|
}
|
|
|
|
static int ipvideo_decode_block_opcode_0xE_16(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
int x, y;
|
|
uint16_t pix;
|
|
uint16_t *pixel_ptr = (uint16_t*)s->pixel_ptr;
|
|
|
|
/* 1-color encoding: the whole block is 1 solid color */
|
|
pix = bytestream2_get_le16(&s->stream_ptr);
|
|
|
|
for (y = 0; y < 8; y++) {
|
|
for (x = 0; x < 8; x++)
|
|
pixel_ptr[x] = pix;
|
|
pixel_ptr += s->stride;
|
|
}
|
|
|
|
/* report success */
|
|
return 0;
|
|
}
|
|
|
|
static int (* const ipvideo_decode_block[])(IpvideoContext *s, AVFrame *frame) = {
|
|
ipvideo_decode_block_opcode_0x0, ipvideo_decode_block_opcode_0x1,
|
|
ipvideo_decode_block_opcode_0x2, ipvideo_decode_block_opcode_0x3,
|
|
ipvideo_decode_block_opcode_0x4, ipvideo_decode_block_opcode_0x5,
|
|
ipvideo_decode_block_opcode_0x6, ipvideo_decode_block_opcode_0x7,
|
|
ipvideo_decode_block_opcode_0x8, ipvideo_decode_block_opcode_0x9,
|
|
ipvideo_decode_block_opcode_0xA, ipvideo_decode_block_opcode_0xB,
|
|
ipvideo_decode_block_opcode_0xC, ipvideo_decode_block_opcode_0xD,
|
|
ipvideo_decode_block_opcode_0xE, ipvideo_decode_block_opcode_0xF,
|
|
};
|
|
|
|
static int (* const ipvideo_decode_block16[])(IpvideoContext *s, AVFrame *frame) = {
|
|
ipvideo_decode_block_opcode_0x0, ipvideo_decode_block_opcode_0x1,
|
|
ipvideo_decode_block_opcode_0x2, ipvideo_decode_block_opcode_0x3,
|
|
ipvideo_decode_block_opcode_0x4, ipvideo_decode_block_opcode_0x5,
|
|
ipvideo_decode_block_opcode_0x6_16, ipvideo_decode_block_opcode_0x7_16,
|
|
ipvideo_decode_block_opcode_0x8_16, ipvideo_decode_block_opcode_0x9_16,
|
|
ipvideo_decode_block_opcode_0xA_16, ipvideo_decode_block_opcode_0xB_16,
|
|
ipvideo_decode_block_opcode_0xC_16, ipvideo_decode_block_opcode_0xD_16,
|
|
ipvideo_decode_block_opcode_0xE_16, ipvideo_decode_block_opcode_0x1,
|
|
};
|
|
|
|
static void ipvideo_format_06_firstpass(IpvideoContext *s, AVFrame *frame, int16_t opcode)
|
|
{
|
|
int line;
|
|
|
|
if (!opcode) {
|
|
for (line = 0; line < 8; ++line) {
|
|
bytestream2_get_buffer(&s->stream_ptr, s->pixel_ptr, 8);
|
|
s->pixel_ptr += s->stride;
|
|
}
|
|
} else {
|
|
/* Don't try to copy second_last_frame data on the first frames */
|
|
if (s->avctx->frame_number > 2)
|
|
copy_from(s, s->second_last_frame, frame, 0, 0);
|
|
}
|
|
}
|
|
|
|
static void ipvideo_format_06_secondpass(IpvideoContext *s, AVFrame *frame, int16_t opcode)
|
|
{
|
|
int off_x, off_y;
|
|
|
|
if (opcode < 0) {
|
|
off_x = ((uint16_t)opcode - 0xC000) % frame->width;
|
|
off_y = ((uint16_t)opcode - 0xC000) / frame->width;
|
|
copy_from(s, s->last_frame, frame, off_x, off_y);
|
|
} else if (opcode > 0) {
|
|
off_x = ((uint16_t)opcode - 0x4000) % frame->width;
|
|
off_y = ((uint16_t)opcode - 0x4000) / frame->width;
|
|
copy_from(s, frame, frame, off_x, off_y);
|
|
}
|
|
}
|
|
|
|
static void (* const ipvideo_format_06_passes[])(IpvideoContext *s, AVFrame *frame, int16_t op) = {
|
|
ipvideo_format_06_firstpass, ipvideo_format_06_secondpass,
|
|
};
|
|
|
|
static void ipvideo_decode_format_06_opcodes(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
int pass, x, y;
|
|
int16_t opcode;
|
|
GetByteContext decoding_map_ptr;
|
|
|
|
/* this is PAL8, so make the palette available */
|
|
memcpy(frame->data[1], s->pal, AVPALETTE_SIZE);
|
|
s->stride = frame->linesize[0];
|
|
|
|
s->line_inc = s->stride - 8;
|
|
s->upper_motion_limit_offset = (s->avctx->height - 8) * frame->linesize[0]
|
|
+ (s->avctx->width - 8) * (1 + s->is_16bpp);
|
|
|
|
bytestream2_init(&decoding_map_ptr, s->decoding_map, s->decoding_map_size);
|
|
|
|
for (pass = 0; pass < 2; ++pass) {
|
|
bytestream2_seek(&decoding_map_ptr, 0, SEEK_SET);
|
|
for (y = 0; y < s->avctx->height; y += 8) {
|
|
for (x = 0; x < s->avctx->width; x += 8) {
|
|
opcode = bytestream2_get_le16(&decoding_map_ptr);
|
|
|
|
ff_tlog(s->avctx,
|
|
" block @ (%3d, %3d): opcode 0x%X, data ptr offset %d\n",
|
|
x, y, opcode, bytestream2_tell(&s->stream_ptr));
|
|
|
|
s->pixel_ptr = frame->data[0] + x + y * frame->linesize[0];
|
|
ipvideo_format_06_passes[pass](s, frame, opcode);
|
|
}
|
|
}
|
|
}
|
|
|
|
if (bytestream2_get_bytes_left(&s->stream_ptr) > 1) {
|
|
av_log(s->avctx, AV_LOG_DEBUG,
|
|
"decode finished with %d bytes left over\n",
|
|
bytestream2_get_bytes_left(&s->stream_ptr));
|
|
}
|
|
}
|
|
|
|
static void ipvideo_format_10_firstpass(IpvideoContext *s, AVFrame *frame, int16_t opcode)
|
|
{
|
|
int line;
|
|
|
|
if (!opcode) {
|
|
for (line = 0; line < 8; ++line) {
|
|
bytestream2_get_buffer(&s->stream_ptr, s->pixel_ptr, 8);
|
|
s->pixel_ptr += s->stride;
|
|
}
|
|
}
|
|
}
|
|
|
|
static void ipvideo_format_10_secondpass(IpvideoContext *s, AVFrame *frame, int16_t opcode)
|
|
{
|
|
int off_x, off_y;
|
|
|
|
if (opcode < 0) {
|
|
off_x = ((uint16_t)opcode - 0xC000) % s->cur_decode_frame->width;
|
|
off_y = ((uint16_t)opcode - 0xC000) / s->cur_decode_frame->width;
|
|
copy_from(s, s->prev_decode_frame, s->cur_decode_frame, off_x, off_y);
|
|
} else if (opcode > 0) {
|
|
off_x = ((uint16_t)opcode - 0x4000) % s->cur_decode_frame->width;
|
|
off_y = ((uint16_t)opcode - 0x4000) / s->cur_decode_frame->width;
|
|
copy_from(s, s->cur_decode_frame, s->cur_decode_frame, off_x, off_y);
|
|
}
|
|
}
|
|
|
|
static void (* const ipvideo_format_10_passes[])(IpvideoContext *s, AVFrame *frame, int16_t op) = {
|
|
ipvideo_format_10_firstpass, ipvideo_format_10_secondpass,
|
|
};
|
|
|
|
static void ipvideo_decode_format_10_opcodes(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
int pass, x, y, changed_block;
|
|
int16_t opcode, skip;
|
|
GetByteContext decoding_map_ptr;
|
|
GetByteContext skip_map_ptr;
|
|
|
|
bytestream2_skip(&s->stream_ptr, 14); /* data starts 14 bytes in */
|
|
|
|
/* this is PAL8, so make the palette available */
|
|
memcpy(frame->data[1], s->pal, AVPALETTE_SIZE);
|
|
s->stride = frame->linesize[0];
|
|
|
|
s->line_inc = s->stride - 8;
|
|
s->upper_motion_limit_offset = (s->avctx->height - 8) * frame->linesize[0]
|
|
+ (s->avctx->width - 8) * (1 + s->is_16bpp);
|
|
|
|
bytestream2_init(&decoding_map_ptr, s->decoding_map, s->decoding_map_size);
|
|
bytestream2_init(&skip_map_ptr, s->skip_map, s->skip_map_size);
|
|
|
|
for (pass = 0; pass < 2; ++pass) {
|
|
bytestream2_seek(&decoding_map_ptr, 0, SEEK_SET);
|
|
bytestream2_seek(&skip_map_ptr, 0, SEEK_SET);
|
|
skip = bytestream2_get_le16(&skip_map_ptr);
|
|
|
|
for (y = 0; y < s->avctx->height; y += 8) {
|
|
for (x = 0; x < s->avctx->width; x += 8) {
|
|
s->pixel_ptr = s->cur_decode_frame->data[0] + x + y * s->cur_decode_frame->linesize[0];
|
|
|
|
while (skip <= 0) {
|
|
if (skip != -0x8000 && skip) {
|
|
opcode = bytestream2_get_le16(&decoding_map_ptr);
|
|
ipvideo_format_10_passes[pass](s, frame, opcode);
|
|
break;
|
|
}
|
|
if (bytestream2_get_bytes_left(&skip_map_ptr) < 2)
|
|
return;
|
|
skip = bytestream2_get_le16(&skip_map_ptr);
|
|
}
|
|
skip *= 2;
|
|
}
|
|
}
|
|
}
|
|
|
|
bytestream2_seek(&skip_map_ptr, 0, SEEK_SET);
|
|
skip = bytestream2_get_le16(&skip_map_ptr);
|
|
for (y = 0; y < s->avctx->height; y += 8) {
|
|
for (x = 0; x < s->avctx->width; x += 8) {
|
|
changed_block = 0;
|
|
s->pixel_ptr = frame->data[0] + x + y*frame->linesize[0];
|
|
|
|
while (skip <= 0) {
|
|
if (skip != -0x8000 && skip) {
|
|
changed_block = 1;
|
|
break;
|
|
}
|
|
if (bytestream2_get_bytes_left(&skip_map_ptr) < 2)
|
|
return;
|
|
skip = bytestream2_get_le16(&skip_map_ptr);
|
|
}
|
|
|
|
if (changed_block) {
|
|
copy_from(s, s->cur_decode_frame, frame, 0, 0);
|
|
} else {
|
|
/* Don't try to copy last_frame data on the first frame */
|
|
if (s->avctx->frame_number)
|
|
copy_from(s, s->last_frame, frame, 0, 0);
|
|
}
|
|
skip *= 2;
|
|
}
|
|
}
|
|
|
|
FFSWAP(AVFrame*, s->prev_decode_frame, s->cur_decode_frame);
|
|
|
|
if (bytestream2_get_bytes_left(&s->stream_ptr) > 1) {
|
|
av_log(s->avctx, AV_LOG_DEBUG,
|
|
"decode finished with %d bytes left over\n",
|
|
bytestream2_get_bytes_left(&s->stream_ptr));
|
|
}
|
|
}
|
|
|
|
static void ipvideo_decode_format_11_opcodes(IpvideoContext *s, AVFrame *frame)
|
|
{
|
|
int x, y;
|
|
unsigned char opcode;
|
|
int ret;
|
|
GetBitContext gb;
|
|
|
|
bytestream2_skip(&s->stream_ptr, 14); /* data starts 14 bytes in */
|
|
if (!s->is_16bpp) {
|
|
/* this is PAL8, so make the palette available */
|
|
memcpy(frame->data[1], s->pal, AVPALETTE_SIZE);
|
|
|
|
s->stride = frame->linesize[0];
|
|
} else {
|
|
s->stride = frame->linesize[0] >> 1;
|
|
s->mv_ptr = s->stream_ptr;
|
|
bytestream2_skip(&s->mv_ptr, bytestream2_get_le16(&s->stream_ptr));
|
|
}
|
|
s->line_inc = s->stride - 8;
|
|
s->upper_motion_limit_offset = (s->avctx->height - 8) * frame->linesize[0]
|
|
+ (s->avctx->width - 8) * (1 + s->is_16bpp);
|
|
|
|
init_get_bits(&gb, s->decoding_map, s->decoding_map_size * 8);
|
|
for (y = 0; y < s->avctx->height; y += 8) {
|
|
for (x = 0; x < s->avctx->width; x += 8) {
|
|
if (get_bits_left(&gb) < 4)
|
|
return;
|
|
opcode = get_bits(&gb, 4);
|
|
|
|
ff_tlog(s->avctx,
|
|
" block @ (%3d, %3d): encoding 0x%X, data ptr offset %d\n",
|
|
x, y, opcode, bytestream2_tell(&s->stream_ptr));
|
|
|
|
if (!s->is_16bpp) {
|
|
s->pixel_ptr = frame->data[0] + x
|
|
+ y*frame->linesize[0];
|
|
ret = ipvideo_decode_block[opcode](s, frame);
|
|
} else {
|
|
s->pixel_ptr = frame->data[0] + x*2
|
|
+ y*frame->linesize[0];
|
|
ret = ipvideo_decode_block16[opcode](s, frame);
|
|
}
|
|
if (ret != 0) {
|
|
av_log(s->avctx, AV_LOG_ERROR, "decode problem on frame %d, @ block (%d, %d)\n",
|
|
s->avctx->frame_number, x, y);
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
if (bytestream2_get_bytes_left(&s->stream_ptr) > 1) {
|
|
av_log(s->avctx, AV_LOG_DEBUG,
|
|
"decode finished with %d bytes left over\n",
|
|
bytestream2_get_bytes_left(&s->stream_ptr));
|
|
}
|
|
}
|
|
|
|
static av_cold int ipvideo_decode_init(AVCodecContext *avctx)
|
|
{
|
|
IpvideoContext *s = avctx->priv_data;
|
|
|
|
s->avctx = avctx;
|
|
|
|
s->is_16bpp = avctx->bits_per_coded_sample == 16;
|
|
avctx->pix_fmt = s->is_16bpp ? AV_PIX_FMT_RGB555 : AV_PIX_FMT_PAL8;
|
|
|
|
ff_hpeldsp_init(&s->hdsp, avctx->flags);
|
|
|
|
s->last_frame = av_frame_alloc();
|
|
s->second_last_frame = av_frame_alloc();
|
|
s->cur_decode_frame = av_frame_alloc();
|
|
s->prev_decode_frame = av_frame_alloc();
|
|
if (!s->last_frame || !s->second_last_frame ||
|
|
!s->cur_decode_frame || !s->prev_decode_frame) {
|
|
return AVERROR(ENOMEM);
|
|
}
|
|
|
|
s->cur_decode_frame->width = avctx->width;
|
|
s->prev_decode_frame->width = avctx->width;
|
|
s->cur_decode_frame->height = avctx->height;
|
|
s->prev_decode_frame->height = avctx->height;
|
|
s->cur_decode_frame->format = avctx->pix_fmt;
|
|
s->prev_decode_frame->format = avctx->pix_fmt;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ipvideo_decode_frame(AVCodecContext *avctx, AVFrame *frame,
|
|
int *got_frame, AVPacket *avpkt)
|
|
{
|
|
const uint8_t *buf = avpkt->data;
|
|
int buf_size = avpkt->size;
|
|
IpvideoContext *s = avctx->priv_data;
|
|
int ret;
|
|
int send_buffer;
|
|
int frame_format;
|
|
int video_data_size;
|
|
|
|
if (av_packet_get_side_data(avpkt, AV_PKT_DATA_PARAM_CHANGE, NULL)) {
|
|
av_frame_unref(s->last_frame);
|
|
av_frame_unref(s->second_last_frame);
|
|
av_frame_unref(s->cur_decode_frame);
|
|
av_frame_unref(s->prev_decode_frame);
|
|
}
|
|
|
|
if (!s->cur_decode_frame->data[0]) {
|
|
ret = ff_get_buffer(avctx, s->cur_decode_frame, 0);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
ret = ff_get_buffer(avctx, s->prev_decode_frame, 0);
|
|
if (ret < 0) {
|
|
av_frame_unref(s->cur_decode_frame);
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
if (buf_size < 8)
|
|
return AVERROR_INVALIDDATA;
|
|
|
|
frame_format = AV_RL8(buf);
|
|
send_buffer = AV_RL8(buf + 1);
|
|
video_data_size = AV_RL16(buf + 2);
|
|
s->decoding_map_size = AV_RL16(buf + 4);
|
|
s->skip_map_size = AV_RL16(buf + 6);
|
|
|
|
switch (frame_format) {
|
|
case 0x06:
|
|
if (s->decoding_map_size) {
|
|
av_log(avctx, AV_LOG_ERROR, "Decoding map for format 0x06\n");
|
|
return AVERROR_INVALIDDATA;
|
|
}
|
|
|
|
if (s->skip_map_size) {
|
|
av_log(avctx, AV_LOG_ERROR, "Skip map for format 0x06\n");
|
|
return AVERROR_INVALIDDATA;
|
|
}
|
|
|
|
if (s->is_16bpp) {
|
|
av_log(avctx, AV_LOG_ERROR, "Video format 0x06 does not support 16bpp movies\n");
|
|
return AVERROR_INVALIDDATA;
|
|
}
|
|
|
|
/* Decoding map for 0x06 frame format is at the top of pixeldata */
|
|
s->decoding_map_size = ((s->avctx->width / 8) * (s->avctx->height / 8)) * 2;
|
|
s->decoding_map = buf + 8 + 14; /* 14 bits of op data */
|
|
video_data_size -= s->decoding_map_size + 14;
|
|
if (video_data_size <= 0 || s->decoding_map_size == 0)
|
|
return AVERROR_INVALIDDATA;
|
|
|
|
if (buf_size < 8 + s->decoding_map_size + 14 + video_data_size)
|
|
return AVERROR_INVALIDDATA;
|
|
|
|
bytestream2_init(&s->stream_ptr, buf + 8 + s->decoding_map_size + 14, video_data_size);
|
|
|
|
break;
|
|
|
|
case 0x10:
|
|
if (! s->decoding_map_size) {
|
|
av_log(avctx, AV_LOG_ERROR, "Empty decoding map for format 0x10\n");
|
|
return AVERROR_INVALIDDATA;
|
|
}
|
|
|
|
if (! s->skip_map_size) {
|
|
av_log(avctx, AV_LOG_ERROR, "Empty skip map for format 0x10\n");
|
|
return AVERROR_INVALIDDATA;
|
|
}
|
|
|
|
if (s->is_16bpp) {
|
|
av_log(avctx, AV_LOG_ERROR, "Video format 0x10 does not support 16bpp movies\n");
|
|
return AVERROR_INVALIDDATA;
|
|
}
|
|
|
|
if (buf_size < 8 + video_data_size + s->decoding_map_size + s->skip_map_size)
|
|
return AVERROR_INVALIDDATA;
|
|
|
|
bytestream2_init(&s->stream_ptr, buf + 8, video_data_size);
|
|
s->decoding_map = buf + 8 + video_data_size;
|
|
s->skip_map = buf + 8 + video_data_size + s->decoding_map_size;
|
|
|
|
break;
|
|
|
|
case 0x11:
|
|
if (! s->decoding_map_size) {
|
|
av_log(avctx, AV_LOG_ERROR, "Empty decoding map for format 0x11\n");
|
|
return AVERROR_INVALIDDATA;
|
|
}
|
|
|
|
if (s->skip_map_size) {
|
|
av_log(avctx, AV_LOG_ERROR, "Skip map for format 0x11\n");
|
|
return AVERROR_INVALIDDATA;
|
|
}
|
|
|
|
if (buf_size < 8 + video_data_size + s->decoding_map_size)
|
|
return AVERROR_INVALIDDATA;
|
|
|
|
bytestream2_init(&s->stream_ptr, buf + 8, video_data_size);
|
|
s->decoding_map = buf + 8 + video_data_size;
|
|
|
|
break;
|
|
|
|
default:
|
|
av_log(avctx, AV_LOG_ERROR, "Frame type 0x%02X unsupported\n", frame_format);
|
|
}
|
|
|
|
/* ensure we can't overread the packet */
|
|
if (buf_size < 8 + s->decoding_map_size + video_data_size + s->skip_map_size) {
|
|
av_log(avctx, AV_LOG_ERROR, "Invalid IP packet size\n");
|
|
return AVERROR_INVALIDDATA;
|
|
}
|
|
|
|
if ((ret = ff_get_buffer(avctx, frame, AV_GET_BUFFER_FLAG_REF)) < 0)
|
|
return ret;
|
|
|
|
if (!s->is_16bpp) {
|
|
frame->palette_has_changed = ff_copy_palette(s->pal, avpkt, avctx);
|
|
}
|
|
|
|
switch (frame_format) {
|
|
case 0x06:
|
|
ipvideo_decode_format_06_opcodes(s, frame);
|
|
break;
|
|
case 0x10:
|
|
ipvideo_decode_format_10_opcodes(s, frame);
|
|
break;
|
|
case 0x11:
|
|
ipvideo_decode_format_11_opcodes(s, frame);
|
|
break;
|
|
}
|
|
|
|
*got_frame = send_buffer;
|
|
|
|
/* shuffle frames */
|
|
av_frame_unref(s->second_last_frame);
|
|
FFSWAP(AVFrame*, s->second_last_frame, s->last_frame);
|
|
if ((ret = av_frame_ref(s->last_frame, frame)) < 0)
|
|
return ret;
|
|
|
|
/* report that the buffer was completely consumed */
|
|
return buf_size;
|
|
}
|
|
|
|
static av_cold int ipvideo_decode_end(AVCodecContext *avctx)
|
|
{
|
|
IpvideoContext *s = avctx->priv_data;
|
|
|
|
av_frame_free(&s->last_frame);
|
|
av_frame_free(&s->second_last_frame);
|
|
av_frame_free(&s->cur_decode_frame);
|
|
av_frame_free(&s->prev_decode_frame);
|
|
|
|
return 0;
|
|
}
|
|
|
|
const FFCodec ff_interplay_video_decoder = {
|
|
.p.name = "interplayvideo",
|
|
.p.long_name = NULL_IF_CONFIG_SMALL("Interplay MVE video"),
|
|
.p.type = AVMEDIA_TYPE_VIDEO,
|
|
.p.id = AV_CODEC_ID_INTERPLAY_VIDEO,
|
|
.priv_data_size = sizeof(IpvideoContext),
|
|
.init = ipvideo_decode_init,
|
|
.close = ipvideo_decode_end,
|
|
FF_CODEC_DECODE_CB(ipvideo_decode_frame),
|
|
.p.capabilities = AV_CODEC_CAP_DR1 | AV_CODEC_CAP_PARAM_CHANGE,
|
|
.caps_internal = FF_CODEC_CAP_INIT_THREADSAFE | FF_CODEC_CAP_INIT_CLEANUP,
|
|
};
|