1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2024-12-02 03:06:28 +02:00
FFmpeg/libavfilter/opencl/tonemap.cl
Ruiling Song 8b8b0e2cd2 lavfi: add opencl tonemap filter
This filter does HDR(HDR10/HLG) to SDR conversion with tone-mapping.

An example command to use this filter with vaapi codecs:
FFMPEG -init_hw_device vaapi=va:/dev/dri/renderD128 -init_hw_device \
opencl=ocl@va -hwaccel vaapi -hwaccel_device va -hwaccel_output_format \
vaapi -i INPUT -filter_hw_device ocl -filter_complex \
'[0:v]hwmap,tonemap_opencl=t=bt2020:tonemap=linear:format=p010[x1]; \
[x1]hwmap=derive_device=vaapi:reverse=1' -c:v hevc_vaapi -profile 2 OUTPUT

Signed-off-by: Ruiling Song <ruiling.song@intel.com>
2018-06-21 01:19:18 +01:00

273 lines
9.4 KiB
Common Lisp

/*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#define REFERENCE_WHITE 100.0f
extern float3 lrgb2yuv(float3);
extern float lrgb2y(float3);
extern float3 yuv2lrgb(float3);
extern float3 lrgb2lrgb(float3);
extern float get_luma_src(float3);
extern float get_luma_dst(float3);
extern float3 ootf(float3 c, float peak);
extern float3 inverse_ootf(float3 c, float peak);
extern float3 get_chroma_sample(float3, float3, float3, float3);
struct detection_result {
float peak;
float average;
};
float hable_f(float in) {
float a = 0.15f, b = 0.50f, c = 0.10f, d = 0.20f, e = 0.02f, f = 0.30f;
return (in * (in * a + b * c) + d * e) / (in * (in * a + b) + d * f) - e / f;
}
float direct(float s, float peak) {
return s;
}
float linear(float s, float peak) {
return s * tone_param / peak;
}
float gamma(float s, float peak) {
float p = s > 0.05f ? s /peak : 0.05f / peak;
float v = powr(p, 1.0f / tone_param);
return s > 0.05f ? v : (s * v /0.05f);
}
float clip(float s, float peak) {
return clamp(s * tone_param, 0.0f, 1.0f);
}
float reinhard(float s, float peak) {
return s / (s + tone_param) * (peak + tone_param) / peak;
}
float hable(float s, float peak) {
return hable_f(s)/hable_f(peak);
}
float mobius(float s, float peak) {
float j = tone_param;
float a, b;
if (s <= j)
return s;
a = -j * j * (peak - 1.0f) / (j * j - 2.0f * j + peak);
b = (j * j - 2.0f * j * peak + peak) / max(peak - 1.0f, 1e-6f);
return (b * b + 2.0f * b * j + j * j) / (b - a) * (s + a) / (s + b);
}
// detect peak/average signal of a frame, the algorithm was ported from:
// libplacebo (https://github.com/haasn/libplacebo)
struct detection_result
detect_peak_avg(global uint *util_buf, __local uint *sum_wg,
float signal, float peak) {
// layout of the util buffer
//
// Name: : Size (units of 4-bytes)
// average buffer : detection_frames + 1
// peak buffer : detection_frames + 1
// workgroup counter : 1
// total of peak : 1
// total of average : 1
// frame index : 1
// frame number : 1
global uint *avg_buf = util_buf;
global uint *peak_buf = avg_buf + DETECTION_FRAMES + 1;
global uint *counter_wg_p = peak_buf + DETECTION_FRAMES + 1;
global uint *max_total_p = counter_wg_p + 1;
global uint *avg_total_p = max_total_p + 1;
global uint *frame_idx_p = avg_total_p + 1;
global uint *scene_frame_num_p = frame_idx_p + 1;
uint frame_idx = *frame_idx_p;
uint scene_frame_num = *scene_frame_num_p;
size_t lidx = get_local_id(0);
size_t lidy = get_local_id(1);
size_t lsizex = get_local_size(0);
size_t lsizey = get_local_size(1);
uint num_wg = get_num_groups(0) * get_num_groups(1);
size_t group_idx = get_group_id(0);
size_t group_idy = get_group_id(1);
struct detection_result r = {peak, sdr_avg};
if (lidx == 0 && lidy == 0)
*sum_wg = 0;
barrier(CLK_LOCAL_MEM_FENCE);
// update workgroup sum
atomic_add(sum_wg, (uint)(signal * REFERENCE_WHITE));
barrier(CLK_LOCAL_MEM_FENCE);
// update frame peak/avg using work-group-average.
if (lidx == 0 && lidy == 0) {
uint avg_wg = *sum_wg / (lsizex * lsizey);
atomic_max(&peak_buf[frame_idx], avg_wg);
atomic_add(&avg_buf[frame_idx], avg_wg);
}
if (scene_frame_num > 0) {
float peak = (float)*max_total_p / (REFERENCE_WHITE * scene_frame_num);
float avg = (float)*avg_total_p / (REFERENCE_WHITE * scene_frame_num);
r.peak = max(1.0f, peak);
r.average = max(0.25f, avg);
}
if (lidx == 0 && lidy == 0 && atomic_add(counter_wg_p, 1) == num_wg - 1) {
*counter_wg_p = 0;
avg_buf[frame_idx] /= num_wg;
if (scene_threshold > 0.0f) {
uint cur_max = peak_buf[frame_idx];
uint cur_avg = avg_buf[frame_idx];
int diff = (int)(scene_frame_num * cur_avg) - (int)*avg_total_p;
if (abs(diff) > scene_frame_num * scene_threshold * REFERENCE_WHITE) {
for (uint i = 0; i < DETECTION_FRAMES + 1; i++)
avg_buf[i] = 0;
for (uint i = 0; i < DETECTION_FRAMES + 1; i++)
peak_buf[i] = 0;
*avg_total_p = *max_total_p = 0;
*scene_frame_num_p = 0;
avg_buf[frame_idx] = cur_avg;
peak_buf[frame_idx] = cur_max;
}
}
uint next = (frame_idx + 1) % (DETECTION_FRAMES + 1);
// add current frame, subtract next frame
*max_total_p += peak_buf[frame_idx] - peak_buf[next];
*avg_total_p += avg_buf[frame_idx] - avg_buf[next];
// reset next frame
peak_buf[next] = avg_buf[next] = 0;
*frame_idx_p = next;
*scene_frame_num_p = min(*scene_frame_num_p + 1,
(uint)DETECTION_FRAMES);
}
return r;
}
float3 map_one_pixel_rgb(float3 rgb, float peak, float average) {
float sig = max(max(rgb.x, max(rgb.y, rgb.z)), 1e-6f);
// Rescale the variables in order to bring it into a representation where
// 1.0 represents the dst_peak. This is because all of the tone mapping
// algorithms are defined in such a way that they map to the range [0.0, 1.0].
if (target_peak > 1.0f) {
sig *= 1.0f / target_peak;
peak *= 1.0f / target_peak;
}
float sig_old = sig;
// Scale the signal to compensate for differences in the average brightness
float slope = min(1.0f, sdr_avg / average);
sig *= slope;
peak *= slope;
// Desaturate the color using a coefficient dependent on the signal level
if (desat_param > 0.0f) {
float luma = get_luma_dst(rgb);
float coeff = max(sig - 0.18f, 1e-6f) / max(sig, 1e-6f);
coeff = native_powr(coeff, 10.0f / desat_param);
rgb = mix(rgb, (float3)luma, (float3)coeff);
sig = mix(sig, luma * slope, coeff);
}
sig = TONE_FUNC(sig, peak);
sig = min(sig, 1.0f);
rgb *= (sig/sig_old);
return rgb;
}
// map from source space YUV to destination space RGB
float3 map_to_dst_space_from_yuv(float3 yuv, float peak) {
float3 c = yuv2lrgb(yuv);
c = ootf(c, peak);
c = lrgb2lrgb(c);
return c;
}
__kernel void tonemap(__write_only image2d_t dst1,
__read_only image2d_t src1,
__write_only image2d_t dst2,
__read_only image2d_t src2,
global uint *util_buf,
float peak
)
{
__local uint sum_wg;
const sampler_t sampler = (CLK_NORMALIZED_COORDS_FALSE |
CLK_ADDRESS_CLAMP_TO_EDGE |
CLK_FILTER_NEAREST);
int xi = get_global_id(0);
int yi = get_global_id(1);
// each work item process four pixels
int x = 2 * xi;
int y = 2 * yi;
float y0 = read_imagef(src1, sampler, (int2)(x, y)).x;
float y1 = read_imagef(src1, sampler, (int2)(x + 1, y)).x;
float y2 = read_imagef(src1, sampler, (int2)(x, y + 1)).x;
float y3 = read_imagef(src1, sampler, (int2)(x + 1, y + 1)).x;
float2 uv = read_imagef(src2, sampler, (int2)(xi, yi)).xy;
float3 c0 = map_to_dst_space_from_yuv((float3)(y0, uv.x, uv.y), peak);
float3 c1 = map_to_dst_space_from_yuv((float3)(y1, uv.x, uv.y), peak);
float3 c2 = map_to_dst_space_from_yuv((float3)(y2, uv.x, uv.y), peak);
float3 c3 = map_to_dst_space_from_yuv((float3)(y3, uv.x, uv.y), peak);
float sig0 = max(c0.x, max(c0.y, c0.z));
float sig1 = max(c1.x, max(c1.y, c1.z));
float sig2 = max(c2.x, max(c2.y, c2.z));
float sig3 = max(c3.x, max(c3.y, c3.z));
float sig = max(sig0, max(sig1, max(sig2, sig3)));
struct detection_result r = detect_peak_avg(util_buf, &sum_wg, sig, peak);
float3 c0_old = c0, c1_old = c1, c2_old = c2;
c0 = map_one_pixel_rgb(c0, r.peak, r.average);
c1 = map_one_pixel_rgb(c1, r.peak, r.average);
c2 = map_one_pixel_rgb(c2, r.peak, r.average);
c3 = map_one_pixel_rgb(c3, r.peak, r.average);
c0 = inverse_ootf(c0, target_peak);
c1 = inverse_ootf(c1, target_peak);
c2 = inverse_ootf(c2, target_peak);
c3 = inverse_ootf(c3, target_peak);
y0 = lrgb2y(c0);
y1 = lrgb2y(c1);
y2 = lrgb2y(c2);
y3 = lrgb2y(c3);
float3 chroma_c = get_chroma_sample(c0, c1, c2, c3);
float3 chroma = lrgb2yuv(chroma_c);
if (xi < get_image_width(dst2) && yi < get_image_height(dst2)) {
write_imagef(dst1, (int2)(x, y), (float4)(y0, 0.0f, 0.0f, 1.0f));
write_imagef(dst1, (int2)(x+1, y), (float4)(y1, 0.0f, 0.0f, 1.0f));
write_imagef(dst1, (int2)(x, y+1), (float4)(y2, 0.0f, 0.0f, 1.0f));
write_imagef(dst1, (int2)(x+1, y+1), (float4)(y3, 0.0f, 0.0f, 1.0f));
write_imagef(dst2, (int2)(xi, yi),
(float4)(chroma.y, chroma.z, 0.0f, 1.0f));
}
}