mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-11-21 10:55:51 +02:00
b85bf991f6
if always_reset is set Originally committed as revision 10460 to svn://svn.ffmpeg.org/ffmpeg/trunk
515 lines
16 KiB
Plaintext
515 lines
16 KiB
Plaintext
=============================================
|
|
SNOW Video Codec Specification Draft 20070103
|
|
=============================================
|
|
|
|
Intro:
|
|
======
|
|
This Specification describes the snow syntax and semmantics as well as
|
|
how to decode snow.
|
|
The decoding process is precissely described and any compliant decoder
|
|
MUST produce the exactly same output for a spec conformant snow stream.
|
|
For encoding though any process which generates a stream compliant to
|
|
the syntactical and semmantical requirements and which is decodeable by
|
|
the process described in this spec shall be considered a conformant
|
|
snow encoder.
|
|
|
|
Definitions:
|
|
============
|
|
|
|
MUST the specific part must be done to conform to this standard
|
|
SHOULD it is recommended to be done that way, but not strictly required
|
|
|
|
ilog2(x) is the rounded down logarithm of x with basis 2
|
|
ilog2(0) = 0
|
|
|
|
Type definitions:
|
|
=================
|
|
|
|
b 1-bit range coded
|
|
u unsigned scalar value range coded
|
|
s signed scalar value range coded
|
|
|
|
|
|
Bitstream syntax:
|
|
=================
|
|
|
|
frame:
|
|
header
|
|
prediction
|
|
residual
|
|
|
|
header:
|
|
keyframe b MID_STATE
|
|
if(keyframe || always_reset)
|
|
reset_contexts
|
|
if(keyframe){
|
|
version u header_state
|
|
always_reset b header_state
|
|
temporal_decomposition_type u header_state
|
|
temporal_decomposition_count u header_state
|
|
spatial_decomposition_count u header_state
|
|
colorspace_type u header_state
|
|
chroma_h_shift u header_state
|
|
chroma_v_shift u header_state
|
|
spatial_scalability b header_state
|
|
max_ref_frames-1 u header_state
|
|
qlogs
|
|
}
|
|
if(!keyframe){
|
|
update_mc b header_state
|
|
if(update_mc){
|
|
for(plane=0; plane<2; plane++){
|
|
diag_mc b header_state
|
|
htaps/2-1 u header_state
|
|
for(i= p->htaps/2; i; i--)
|
|
|hcoeff[i]| u header_state
|
|
}
|
|
}
|
|
update_qlogs b header_state
|
|
if(update_qlogs){
|
|
spatial_decomposition_count u header_state
|
|
qlogs
|
|
}
|
|
}
|
|
|
|
spatial_decomposition_type s header_state
|
|
qlog s header_state
|
|
mv_scale s header_state
|
|
qbias s header_state
|
|
block_max_depth s header_state
|
|
|
|
qlogs:
|
|
for(plane=0; plane<2; plane++){
|
|
quant_table[plane][0][0] s header_state
|
|
for(level=0; level < spatial_decomposition_count; level++){
|
|
quant_table[plane][level][1]s header_state
|
|
quant_table[plane][level][3]s header_state
|
|
}
|
|
}
|
|
|
|
reset_contexts
|
|
*_state[*]= MID_STATE
|
|
|
|
prediction:
|
|
for(y=0; y<block_count_vertical; y++)
|
|
for(x=0; x<block_count_horizontal; x++)
|
|
block(0)
|
|
|
|
block(level):
|
|
if(keyframe){
|
|
intra=1
|
|
y_diff=cb_diff=cr_diff=0
|
|
}else{
|
|
if(level!=max_block_depth){
|
|
s_context= 2*left->level + 2*top->level + topleft->level + topright->level
|
|
leaf b block_state[4 + s_context]
|
|
}
|
|
if(level==max_block_depth || leaf){
|
|
intra b block_state[1 + left->intra + top->intra]
|
|
if(intra){
|
|
y_diff s block_state[32]
|
|
cb_diff s block_state[64]
|
|
cr_diff s block_state[96]
|
|
}else{
|
|
ref_context= ilog2(2*left->ref) + ilog2(2*top->ref)
|
|
if(ref_frames > 1)
|
|
ref u block_state[128 + 1024 + 32*ref_context]
|
|
mx_context= ilog2(2*abs(left->mx - top->mx))
|
|
my_context= ilog2(2*abs(left->my - top->my))
|
|
mvx_diff s block_state[128 + 32*(mx_context + 16*!!ref)]
|
|
mvy_diff s block_state[128 + 32*(my_context + 16*!!ref)]
|
|
}
|
|
}else{
|
|
block(level+1)
|
|
block(level+1)
|
|
block(level+1)
|
|
block(level+1)
|
|
}
|
|
}
|
|
|
|
|
|
residual:
|
|
FIXME
|
|
|
|
|
|
|
|
Tag description:
|
|
----------------
|
|
|
|
version
|
|
0
|
|
this MUST NOT change within a bitstream
|
|
|
|
always_reset
|
|
if 1 then the range coder contexts will be reset after each frame
|
|
|
|
temporal_decomposition_type
|
|
0
|
|
|
|
temporal_decomposition_count
|
|
0
|
|
|
|
spatial_decomposition_count
|
|
FIXME
|
|
|
|
colorspace_type
|
|
0
|
|
this MUST NOT change within a bitstream
|
|
|
|
chroma_h_shift
|
|
log2(luma.width / chroma.width)
|
|
this MUST NOT change within a bitstream
|
|
|
|
chroma_v_shift
|
|
log2(luma.height / chroma.height)
|
|
this MUST NOT change within a bitstream
|
|
|
|
spatial_scalability
|
|
0
|
|
|
|
max_ref_frames
|
|
maximum number of reference frames
|
|
this MUST NOT change within a bitstream
|
|
|
|
update_mc
|
|
indicates that motion compensation filter parameters are stored in the
|
|
header
|
|
|
|
diag_mc
|
|
flag to enable faster diagonal interpolation
|
|
this SHOULD be 1 unless it turns out to be covered by a valid patent
|
|
|
|
htaps
|
|
number of half pel interpolation filter taps, MUST be even, >0 and <10
|
|
|
|
hcoeff
|
|
half pel interpolation filter coefficients, hcoeff[0] are the 2 middle
|
|
coefficients [1] are the next outer ones and so on, resulting in a filter
|
|
like: ...eff[2], hcoeff[1], hcoeff[0], hcoeff[0], hcoeff[1], hcoeff[2] ...
|
|
the sign of the coefficients is not explicitly stored but alternates
|
|
after each coeff and coeff[0] is positive, so ...,+,-,+,-,+,+,-,+,-,+,...
|
|
hcoeff[0] is not explicitly stored but found by subtracting the sum
|
|
of all stored coefficients with signs from 32
|
|
hcoeff[0]= 32 - hcoeff[1] - hcoeff[2] - ...
|
|
a good choice for hcoeff and htaps is
|
|
htaps= 6
|
|
hcoeff={40,-10,2}
|
|
an alternative which requires more computations at both encoder and
|
|
decoder side and may or may not be better is
|
|
htaps= 8
|
|
hcoeff={42,-14,6,-2}
|
|
|
|
|
|
ref_frames
|
|
minimum of the number of available reference frames and max_ref_frames
|
|
for example the first frame after a key frame always has ref_frames=1
|
|
|
|
spatial_decomposition_type
|
|
wavelet type
|
|
0 is a 9/7 symmetric compact integer wavelet
|
|
1 is a 5/3 symmetric compact integer wavelet
|
|
others are reserved
|
|
stored as delta from last, last is reset to 0 if always_reset || keyframe
|
|
|
|
qlog
|
|
quality (logarthmic quantizer scale)
|
|
stored as delta from last, last is reset to 0 if always_reset || keyframe
|
|
|
|
mv_scale
|
|
stored as delta from last, last is reset to 0 if always_reset || keyframe
|
|
FIXME check that everything works fine if this changes between frames
|
|
|
|
qbias
|
|
dequantization bias
|
|
stored as delta from last, last is reset to 0 if always_reset || keyframe
|
|
|
|
block_max_depth
|
|
maximum depth of the block tree
|
|
stored as delta from last, last is reset to 0 if always_reset || keyframe
|
|
|
|
quant_table
|
|
quantiztation table
|
|
|
|
Range Coder:
|
|
============
|
|
FIXME
|
|
|
|
Neighboring Blocks:
|
|
===================
|
|
left and top are set to the respective blocks unless they are outside of
|
|
the image in which case they are set to the Null block
|
|
|
|
top-left is set to the top left block unless it is outside of the image in
|
|
which case it is set to the left block
|
|
|
|
if this block has no larger parent block or it is at the left side of its
|
|
parent block and the top right block is not outside of the image then the
|
|
top right block is used for top-right else the top-left block is used
|
|
|
|
Null block
|
|
y,cb,cr are 128
|
|
level, ref, mx and my are 0
|
|
|
|
|
|
Motion Vector Prediction:
|
|
=========================
|
|
1. the motion vectors of all the neighboring blocks are scaled to
|
|
compensate for the difference of reference frames
|
|
|
|
scaled_mv= (mv * (256 * (current_reference+1) / (mv.reference+1)) + 128)>>8
|
|
|
|
2. the median of the scaled left, top and top-right vectors is used as
|
|
motion vector prediction
|
|
|
|
3. the used motion vector is the sum of the predictor and
|
|
(mvx_diff, mvy_diff)*mv_scale
|
|
|
|
|
|
Intra DC Predicton:
|
|
======================
|
|
the luma and chroma values of the left block are used as predictors
|
|
|
|
the used luma and chroma is the sum of the predictor and y_diff, cb_diff, cr_diff
|
|
to reverse this in the decoder apply the following:
|
|
block[y][x].dc[0] += block[y][x-1].dc[0];
|
|
block[y][x].dc[1] += block[y][x-1].dc[1];
|
|
block[y][x].dc[2] += block[y][x-1].dc[2];
|
|
block[*][-1].dc[*]= 128;
|
|
|
|
|
|
Motion Compensation:
|
|
====================
|
|
|
|
Halfpel interpolation:
|
|
----------------------
|
|
halfpel interpolation is done by convolution with the halfpel filter stored
|
|
in the header:
|
|
|
|
horizontal halfpel samples are found by
|
|
H1[y][x] = hcoeff[0]*(F[y][x ] + F[y][x+1])
|
|
+ hcoeff[1]*(F[y][x-1] + F[y][x+2])
|
|
+ hcoeff[2]*(F[y][x-2] + F[y][x+3])
|
|
+ ...
|
|
h1[y][x] = (H1[y][x] + 32)>>6;
|
|
|
|
vertical halfpel samples are found by
|
|
H2[y][x] = hcoeff[0]*(F[y ][x] + F[y+1][x])
|
|
+ hcoeff[1]*(F[y-1][x] + F[y+2][x])
|
|
+ ...
|
|
h2[y][x] = (H2[y][x] + 32)>>6;
|
|
|
|
vertical+horizontal halfpel samples are found by
|
|
H3[y][x] = hcoeff[0]*(H2[y][x ] + H2[y][x+1])
|
|
+ hcoeff[1]*(H2[y][x-1] + H2[y][x+2])
|
|
+ ...
|
|
H3[y][x] = hcoeff[0]*(H1[y ][x] + H1[y+1][x])
|
|
+ hcoeff[1]*(H1[y+1][x] + H1[y+2][x])
|
|
+ ...
|
|
h3[y][x] = (H3[y][x] + 2048)>>12;
|
|
|
|
|
|
F H1 F
|
|
| | |
|
|
| | |
|
|
| | |
|
|
F H1 F
|
|
| | |
|
|
| | |
|
|
| | |
|
|
F-------F-------F-> H1<-F-------F-------F
|
|
v v v
|
|
H2 H3 H2
|
|
^ ^ ^
|
|
F-------F-------F-> H1<-F-------F-------F
|
|
| | |
|
|
| | |
|
|
| | |
|
|
F H1 F
|
|
| | |
|
|
| | |
|
|
| | |
|
|
F H1 F
|
|
|
|
|
|
unavailable fullpel samples (outside the picture for example) shall be equal
|
|
to the closest available fullpel sample
|
|
|
|
|
|
Smaller pel interpolation:
|
|
--------------------------
|
|
if diag_mc is set then points which lie on a line between 2 vertically,
|
|
horiziontally or diagonally adjacent halfpel points shall be interpolated
|
|
linearls with rounding to nearest and halfway values rounded up.
|
|
points which lie on 2 diagonals at the same time should only use the one
|
|
diagonal not containing the fullpel point
|
|
|
|
|
|
|
|
F-->O---q---O<--h1->O---q---O<--F
|
|
v \ / v \ / v
|
|
O O O O O O O
|
|
| / | \ |
|
|
q q q q q
|
|
| / | \ |
|
|
O O O O O O O
|
|
^ / \ ^ / \ ^
|
|
h2-->O---q---O<--h3->O---q---O<--h2
|
|
v \ / v \ / v
|
|
O O O O O O O
|
|
| \ | / |
|
|
q q q q q
|
|
| \ | / |
|
|
O O O O O O O
|
|
^ / \ ^ / \ ^
|
|
F-->O---q---O<--h1->O---q---O<--F
|
|
|
|
|
|
|
|
the remaining points shall be bilinearly interpolated from the
|
|
up to 4 surrounding halfpel and fullpel points, again rounding should be to
|
|
nearest and halfway values rounded up
|
|
|
|
compliant snow decoders MUST support 1-1/8 pel luma and 1/2-1/16 pel chroma
|
|
interpolation at least
|
|
|
|
|
|
Overlapped block motion compensation:
|
|
-------------------------------------
|
|
FIXME
|
|
|
|
LL band prediction:
|
|
===================
|
|
Each sample in the LL0 subband is predicted by the median of the left, top and
|
|
left+top-topleft samples, samples outside the subband shall be considered to
|
|
be 0. To reverse this prediction in the decoder apply the following.
|
|
for(y=0; y<height; y++){
|
|
for(x=0; x<width; x++){
|
|
sample[y][x] += median(sample[y-1][x],
|
|
sample[y][x-1],
|
|
sample[y-1][x]+sample[y][x-1]-sample[y-1][x-1]);
|
|
}
|
|
}
|
|
sample[-1][*]=sample[*][-1]= 0;
|
|
width,height here are the width and height of the LL0 subband not of the final
|
|
video
|
|
|
|
|
|
Dequantizaton:
|
|
==============
|
|
FIXME
|
|
|
|
Wavelet Transform:
|
|
==================
|
|
|
|
Snow supports 2 wavelet transforms, the symmetric biorthogonal 5/3 integer
|
|
transform and a integer approximation of the symmetric biorthogonal 9/7
|
|
daubechies wavelet.
|
|
|
|
2D IDWT (inverse discrete wavelet transform)
|
|
--------------------------------------------
|
|
The 2D IDWT applies a 2D filter recursively, each time combining the
|
|
4 lowest frequency subbands into a single subband until only 1 subband
|
|
remains.
|
|
The 2D filter is done by first applying a 1D filter in the vertical direction
|
|
and then applying it in the horizontal one.
|
|
--------------- --------------- --------------- ---------------
|
|
|LL0|HL0| | | | | | | | | | | |
|
|
|---+---| HL1 | | L0|H0 | HL1 | | LL1 | HL1 | | | |
|
|
|LH0|HH0| | | | | | | | | | | |
|
|
|-------+-------|->|-------+-------|->|-------+-------|->| L1 | H1 |->...
|
|
| | | | | | | | | | | |
|
|
| LH1 | HH1 | | LH1 | HH1 | | LH1 | HH1 | | | |
|
|
| | | | | | | | | | | |
|
|
--------------- --------------- --------------- ---------------
|
|
|
|
|
|
1D Filter:
|
|
----------
|
|
1. interleave the samples of the low and high frequency subbands like
|
|
s={L0, H0, L1, H1, L2, H2, L3, H3, ... }
|
|
note, this can end with a L or a H, the number of elements shall be w
|
|
s[-1] shall be considered equivalent to s[1 ]
|
|
s[w ] shall be considered equivalent to s[w-2]
|
|
|
|
2. perform the lifting steps in order as described below
|
|
|
|
5/3 Integer filter:
|
|
1. s[i] -= (s[i-1] + s[i+1] + 2)>>2; for all even i < w
|
|
2. s[i] += (s[i-1] + s[i+1] )>>1; for all odd i < w
|
|
|
|
\ | /|\ | /|\ | /|\ | /|\
|
|
\|/ | \|/ | \|/ | \|/ |
|
|
+ | + | + | + | -1/4
|
|
/|\ | /|\ | /|\ | /|\ |
|
|
/ | \|/ | \|/ | \|/ | \|/
|
|
| + | + | + | + +1/2
|
|
|
|
|
|
snows 9/7 Integer filter:
|
|
1. s[i] -= (3*(s[i-1] + s[i+1]) + 4)>>3; for all even i < w
|
|
2. s[i] -= s[i-1] + s[i+1] ; for all odd i < w
|
|
3. s[i] += ( s[i-1] + s[i+1] + 4*s[i] + 8)>>4; for all even i < w
|
|
4. s[i] += (3*(s[i-1] + s[i+1]) )>>1; for all odd i < w
|
|
|
|
\ | /|\ | /|\ | /|\ | /|\
|
|
\|/ | \|/ | \|/ | \|/ |
|
|
+ | + | + | + | -3/8
|
|
/|\ | /|\ | /|\ | /|\ |
|
|
/ | \|/ | \|/ | \|/ | \|/
|
|
(| + (| + (| + (| + -1
|
|
\ + /|\ + /|\ + /|\ + /|\ +1/4
|
|
\|/ | \|/ | \|/ | \|/ |
|
|
+ | + | + | + | +1/16
|
|
/|\ | /|\ | /|\ | /|\ |
|
|
/ | \|/ | \|/ | \|/ | \|/
|
|
| + | + | + | + +3/2
|
|
|
|
optimization tips:
|
|
following are exactly identical
|
|
(3a)>>1 == a + (a>>1)
|
|
(a + 4b + 8)>>4 == ((a>>2) + b + 2)>>2
|
|
|
|
16bit implementation note:
|
|
The IDWT can be implemented with 16bits, but this requires some care to
|
|
prevent overflows, the following list, lists the minimum number of bits needed
|
|
for some terms
|
|
1. lifting step
|
|
A= s[i-1] + s[i+1] 16bit
|
|
3*A + 4 18bit
|
|
A + (A>>1) + 2 17bit
|
|
|
|
3. lifting step
|
|
s[i-1] + s[i+1] 17bit
|
|
|
|
4. lifiting step
|
|
3*(s[i-1] + s[i+1]) 17bit
|
|
|
|
|
|
TODO:
|
|
=====
|
|
Important:
|
|
finetune initial contexts
|
|
spatial_decomposition_count per frame?
|
|
flip wavelet?
|
|
try to use the wavelet transformed predicted image (motion compensated image) as context for coding the residual coefficients
|
|
try the MV length as context for coding the residual coefficients
|
|
use extradata for stuff which is in the keyframes now?
|
|
the MV median predictor is patented IIRC
|
|
change MC so per picture halfpel interpolation can be done and finish the implementation of it
|
|
compare the 6 tap and 8 tap hpel filters (psnr/bitrate and subjective quality)
|
|
try different range coder state transition tables for different contexts
|
|
|
|
Not Important:
|
|
spatial_scalability b vs u (!= 0 breaks syntax anyway so we can add a u later)
|
|
|
|
|
|
Credits:
|
|
========
|
|
Michael Niedermayer
|
|
Loren Merritt
|
|
|
|
|
|
Copyright:
|
|
==========
|
|
GPL + GFDL + whatever is needed to make this a RFC
|