2022-06-26 22:58:09 +02:00
|
|
|
/*
|
|
|
|
* Copyright (c) 2022 Jonathan Swinney <jswinney@amazon.com>
|
|
|
|
*
|
|
|
|
* This file is part of FFmpeg.
|
|
|
|
*
|
|
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
|
|
* License as published by the Free Software Foundation; either
|
|
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
|
|
*
|
|
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
* Lesser General Public License for more details.
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include "libavutil/aarch64/asm.S"
|
|
|
|
|
|
|
|
function ff_pix_abs16_neon, export=1
|
|
|
|
// x0 unused
|
|
|
|
// x1 uint8_t *pix1
|
|
|
|
// x2 uint8_t *pix2
|
|
|
|
// x3 ptrdiff_t stride
|
|
|
|
// w4 int h
|
|
|
|
cmp w4, #4 // if h < 4, jump to completion section
|
2022-07-12 23:48:48 +02:00
|
|
|
movi v16.8h, #0 // clear result accumulator
|
|
|
|
movi v17.8h, #0 // clear result accumulator
|
2022-06-26 22:58:09 +02:00
|
|
|
b.lt 2f
|
|
|
|
1:
|
|
|
|
ld1 {v0.16b}, [x1], x3 // load pix1
|
|
|
|
ld1 {v4.16b}, [x2], x3 // load pix2
|
|
|
|
ld1 {v1.16b}, [x1], x3 // load pix1
|
|
|
|
ld1 {v5.16b}, [x2], x3 // load pix2
|
2022-07-12 23:48:48 +02:00
|
|
|
uabal v16.8h, v0.8b, v4.8b // absolute difference accumulate
|
|
|
|
uabal2 v17.8h, v0.16b, v4.16b
|
2022-06-26 22:58:09 +02:00
|
|
|
ld1 {v2.16b}, [x1], x3 // load pix1
|
|
|
|
ld1 {v6.16b}, [x2], x3 // load pix2
|
|
|
|
uabal v16.8h, v1.8b, v5.8b // absolute difference accumulate
|
|
|
|
uabal2 v17.8h, v1.16b, v5.16b
|
|
|
|
ld1 {v3.16b}, [x1], x3
|
|
|
|
ld1 {v7.16b}, [x2], x3
|
|
|
|
uabal v16.8h, v2.8b, v6.8b
|
|
|
|
uabal2 v17.8h, v2.16b, v6.16b
|
|
|
|
sub w4, w4, #4 // h -= 4
|
|
|
|
uabal v16.8h, v3.8b, v7.8b
|
|
|
|
uabal2 v17.8h, v3.16b, v7.16b
|
|
|
|
cmp w4, #4 // if h >= 4, loop
|
|
|
|
|
|
|
|
b.ge 1b
|
|
|
|
cbnz w4, 2f // if iterations remain, jump to completion section
|
|
|
|
|
2022-07-12 23:48:48 +02:00
|
|
|
add v16.8h, v16.8h, v17.8h
|
|
|
|
uaddlv s16, v16.8h // add up everything in v16 accumulator
|
|
|
|
fmov w0, s16 // copy result to general purpose register
|
2022-06-26 22:58:09 +02:00
|
|
|
ret
|
|
|
|
|
|
|
|
2:
|
|
|
|
ld1 {v0.16b}, [x1], x3 // load pix1
|
|
|
|
ld1 {v4.16b}, [x2], x3 // load pix2
|
|
|
|
subs w4, w4, #1 // h -= 1
|
2022-07-12 23:48:48 +02:00
|
|
|
uabal v16.8h, v0.8b, v4.8b // absolute difference accumulate
|
|
|
|
uabal2 v17.8h, v0.16b, v4.16b
|
2022-06-26 22:58:09 +02:00
|
|
|
b.ne 2b
|
|
|
|
|
2022-07-12 23:48:48 +02:00
|
|
|
add v16.8h, v16.8h, v17.8h
|
|
|
|
uaddlv s16, v16.8h // add up everything in v16 accumulator
|
|
|
|
fmov w0, s16 // copy result to general purpose register
|
2022-06-26 22:58:09 +02:00
|
|
|
ret
|
|
|
|
endfunc
|
|
|
|
|
2022-08-16 14:20:16 +02:00
|
|
|
function ff_pix_abs8_neon, export=1
|
|
|
|
// x0 unused
|
|
|
|
// x1 uint8_t *pix1
|
|
|
|
// x2 uint8_t *pix2
|
|
|
|
// x3 ptrdiff_t stride
|
|
|
|
// w4 int h
|
|
|
|
|
|
|
|
movi v30.8h, #0
|
|
|
|
cmp w4, #4
|
|
|
|
b.lt 2f
|
|
|
|
|
|
|
|
// make 4 iterations at once
|
|
|
|
1:
|
|
|
|
ld1 {v0.8b}, [x1], x3 // Load pix1 for first iteration
|
|
|
|
ld1 {v1.8b}, [x2], x3 // Load pix2 for first iteration
|
|
|
|
ld1 {v2.8b}, [x1], x3 // Load pix1 for second iteration
|
|
|
|
uabal v30.8h, v0.8b, v1.8b // Absolute difference, first iteration
|
|
|
|
ld1 {v3.8b}, [x2], x3 // Load pix2 for second iteration
|
|
|
|
ld1 {v4.8b}, [x1], x3 // Load pix1 for third iteration
|
|
|
|
uabal v30.8h, v2.8b, v3.8b // Absolute difference, second iteration
|
|
|
|
ld1 {v5.8b}, [x2], x3 // Load pix2 for third iteration
|
|
|
|
sub w4, w4, #4 // h -= 4
|
|
|
|
ld1 {v6.8b}, [x1], x3 // Load pix1 for foruth iteration
|
|
|
|
ld1 {v7.8b}, [x2], x3 // Load pix2 for fourth iteration
|
|
|
|
uabal v30.8h, v4.8b, v5.8b // Absolute difference, third iteration
|
|
|
|
cmp w4, #4
|
|
|
|
uabal v30.8h, v6.8b, v7.8b // Absolute difference, foruth iteration
|
|
|
|
b.ge 1b
|
|
|
|
|
|
|
|
cbz w4, 3f
|
|
|
|
|
|
|
|
// iterate by one
|
|
|
|
2:
|
|
|
|
ld1 {v0.8b}, [x1], x3 // Load pix1
|
|
|
|
ld1 {v1.8b}, [x2], x3 // Load pix2
|
|
|
|
|
|
|
|
subs w4, w4, #1
|
|
|
|
uabal v30.8h, v0.8b, v1.8b
|
|
|
|
b.ne 2b
|
|
|
|
|
|
|
|
3:
|
|
|
|
uaddlv s20, v30.8h // Add up vector
|
|
|
|
fmov w0, s20
|
|
|
|
|
|
|
|
ret
|
|
|
|
endfunc
|
|
|
|
|
2022-06-26 22:58:09 +02:00
|
|
|
function ff_pix_abs16_xy2_neon, export=1
|
|
|
|
// x0 unused
|
|
|
|
// x1 uint8_t *pix1
|
|
|
|
// x2 uint8_t *pix2
|
|
|
|
// x3 ptrdiff_t stride
|
|
|
|
// w4 int h
|
|
|
|
|
|
|
|
add x5, x2, x3 // use x5 to hold uint8_t *pix3
|
2022-07-12 23:48:48 +02:00
|
|
|
movi v21.8h, #0 // initialize the result register
|
|
|
|
movi v22.8h, #0 // initialize the result register
|
2022-06-26 22:58:09 +02:00
|
|
|
|
|
|
|
// Load initial pix2 values for either the unrolled version or completion version.
|
|
|
|
ldur q4, [x2, #1] // load pix2+1
|
|
|
|
ldr q3, [x2] // load pix2
|
|
|
|
uaddl v2.8h, v4.8b, v3.8b // pix2 + pix2+1 0..7
|
|
|
|
uaddl2 v3.8h, v4.16b, v3.16b // pix2 + pix2+1 8..15
|
|
|
|
cmp w4, #4 // if h < 4 jump to the completion version
|
|
|
|
b.lt 2f
|
|
|
|
1:
|
|
|
|
// This is an unrolled implementation. It completes 4 iterations of the C for each branch.
|
|
|
|
// In each iteration, pix2[i+1] == pix3[i]. This means we need only three loads per iteration,
|
|
|
|
// plus two at the beginning to start.
|
|
|
|
ldur q5, [x5, #1] // load pix3+1
|
|
|
|
ld1 {v4.16b}, [x5], x3 // load pix3
|
|
|
|
ld1 {v1.16b}, [x1], x3 // load pix1
|
|
|
|
|
|
|
|
ldur q7, [x5, #1] // load pix3+1
|
|
|
|
ld1 {v6.16b}, [x5], x3 // load pix3
|
|
|
|
ld1 {v16.16b}, [x1], x3 // load pix1
|
|
|
|
|
|
|
|
// These blocks compute the average: avg(pix2[n], pix2[n+1], pix3[n], pix3[n+1])
|
|
|
|
uaddl v30.8h, v4.8b, v5.8b // pix3 + pix3+1 0..7
|
|
|
|
uaddl2 v31.8h, v4.16b, v5.16b // pix3 + pix3+1 8..15
|
2022-07-12 23:05:41 +02:00
|
|
|
|
|
|
|
ldur q19, [x5, #1] // load pix3+1
|
|
|
|
|
2022-06-26 22:58:09 +02:00
|
|
|
add v23.8h, v2.8h, v30.8h // add up 0..7, using pix2 + pix2+1 values from previous iteration
|
|
|
|
add v24.8h, v3.8h, v31.8h // add up 8..15, using pix2 + pix2+1 values from previous iteration
|
2022-07-12 23:05:41 +02:00
|
|
|
|
|
|
|
ld1 {v18.16b}, [x5], x3 // load pix3
|
|
|
|
ld1 {v17.16b}, [x1], x3 // load pix1
|
|
|
|
|
2022-06-26 22:58:09 +02:00
|
|
|
rshrn v23.8b, v23.8h, #2 // shift right 2 0..7 (rounding shift right)
|
|
|
|
rshrn2 v23.16b, v24.8h, #2 // shift right 2 8..15
|
|
|
|
|
|
|
|
uaddl v2.8h, v6.8b, v7.8b // pix3 + pix3+1 0..7
|
|
|
|
uaddl2 v3.8h, v6.16b, v7.16b // pix3 + pix3+1 8..15
|
2022-07-12 23:05:41 +02:00
|
|
|
|
2022-07-12 23:48:48 +02:00
|
|
|
ldur q7, [x5, #1] // load pix3+1
|
2022-07-12 23:05:41 +02:00
|
|
|
|
2022-06-26 22:58:09 +02:00
|
|
|
add v26.8h, v30.8h, v2.8h // add up 0..7, using pix2 + pix2+1 values from pix3 above
|
|
|
|
add v27.8h, v31.8h, v3.8h // add up 8..15, using pix2 + pix2+1 values from pix3 above
|
2022-07-12 23:05:41 +02:00
|
|
|
|
2022-07-12 23:48:48 +02:00
|
|
|
uabal v21.8h, v1.8b, v23.8b // absolute difference 0..7, i=0
|
|
|
|
uabal2 v22.8h, v1.16b, v23.16b // absolute difference 8..15, i=0
|
2022-07-12 23:06:31 +02:00
|
|
|
|
2022-07-12 23:48:48 +02:00
|
|
|
ld1 {v6.16b}, [x5], x3 // load pix3
|
2022-07-12 23:05:41 +02:00
|
|
|
ld1 {v20.16b}, [x1], x3 // load pix1
|
|
|
|
|
2022-06-26 22:58:09 +02:00
|
|
|
rshrn v26.8b, v26.8h, #2 // shift right 2 0..7 (rounding shift right)
|
|
|
|
rshrn2 v26.16b, v27.8h, #2 // shift right 2 8..15
|
|
|
|
|
|
|
|
uaddl v4.8h, v18.8b, v19.8b // pix3 + pix3+1 0..7
|
|
|
|
uaddl2 v5.8h, v18.16b, v19.16b // pix3 + pix3+1 8..15
|
|
|
|
add v28.8h, v2.8h, v4.8h // add up 0..7, using pix2 + pix2+1 values from pix3 above
|
|
|
|
add v29.8h, v3.8h, v5.8h // add up 8..15, using pix2 + pix2+1 values from pix3 above
|
|
|
|
rshrn v28.8b, v28.8h, #2 // shift right 2 0..7 (rounding shift right)
|
|
|
|
rshrn2 v28.16b, v29.8h, #2 // shift right 2 8..15
|
|
|
|
|
2022-07-12 23:48:48 +02:00
|
|
|
uabal v21.8h, v16.8b, v26.8b // absolute difference 0..7, i=1
|
|
|
|
uabal2 v22.8h, v16.16b, v26.16b // absolute difference 8..15, i=1
|
2022-07-12 23:06:31 +02:00
|
|
|
|
2022-07-12 23:48:48 +02:00
|
|
|
uaddl v2.8h, v6.8b, v7.8b // pix3 + pix3+1 0..7
|
|
|
|
uaddl2 v3.8h, v6.16b, v7.16b // pix3 + pix3+1 8..15
|
2022-06-26 22:58:09 +02:00
|
|
|
add v30.8h, v4.8h, v2.8h // add up 0..7, using pix2 + pix2+1 values from pix3 above
|
|
|
|
add v31.8h, v5.8h, v3.8h // add up 8..15, using pix2 + pix2+1 values from pix3 above
|
|
|
|
rshrn v30.8b, v30.8h, #2 // shift right 2 0..7 (rounding shift right)
|
|
|
|
rshrn2 v30.16b, v31.8h, #2 // shift right 2 8..15
|
|
|
|
|
|
|
|
sub w4, w4, #4 // h -= 4
|
|
|
|
|
2022-07-12 23:48:48 +02:00
|
|
|
uabal v21.8h, v17.8b, v28.8b // absolute difference 0..7, i=2
|
|
|
|
uabal2 v22.8h, v17.16b, v28.16b // absolute difference 8..15, i=2
|
2022-06-26 22:58:09 +02:00
|
|
|
|
|
|
|
cmp w4, #4 // loop if h >= 4
|
|
|
|
|
2022-07-12 23:48:48 +02:00
|
|
|
|
|
|
|
uabal v21.8h, v20.8b, v30.8b // absolute difference 0..7, i=3
|
|
|
|
uabal2 v22.8h, v20.16b, v30.16b // absolute difference 8..15, i=3
|
2022-06-26 22:58:09 +02:00
|
|
|
|
|
|
|
b.ge 1b
|
|
|
|
cbnz w4, 2f // if iterations remain jump to completion section
|
|
|
|
|
2022-07-12 23:48:48 +02:00
|
|
|
add v4.8h, v21.8h, v22.8h
|
|
|
|
uaddlv s0, v4.8h // finish adding up accumulated values
|
|
|
|
|
2022-06-26 22:58:09 +02:00
|
|
|
fmov w0, s0 // copy result to general purpose register
|
|
|
|
ret
|
|
|
|
2:
|
|
|
|
// v2 and v3 are set either at the end of this loop or at from the unrolled version
|
|
|
|
// which branches here to complete iterations when h % 4 != 0.
|
|
|
|
ldur q5, [x5, #1] // load pix3+1
|
|
|
|
ld1 {v4.16b}, [x5], x3 // load pix3
|
|
|
|
ld1 {v1.16b}, [x1], x3 // load pix1
|
|
|
|
subs w4, w4, #1 // decrement h
|
|
|
|
|
|
|
|
uaddl v18.8h, v4.8b, v5.8b // pix3 + pix3+1 0..7
|
|
|
|
uaddl2 v19.8h, v4.16b, v5.16b // pix3 + pix3+1 8..15
|
|
|
|
add v16.8h, v2.8h, v18.8h // add up 0..7, using pix2 + pix2+1 values from previous iteration
|
|
|
|
add v17.8h, v3.8h, v19.8h // add up 8..15, using pix2 + pix2+1 values from previous iteration
|
|
|
|
// divide by 4 to compute the average of values summed above
|
2022-07-12 23:48:48 +02:00
|
|
|
rshrn v16.8b, v16.8h, #2 // shift right by 2 0..7 (rounding shift right)
|
|
|
|
rshrn2 v16.16b, v17.8h, #2 // shift right by 2 8..15
|
2022-06-26 22:58:09 +02:00
|
|
|
|
2022-07-12 23:48:48 +02:00
|
|
|
uabal v21.8h, v1.8b, v16.8b // absolute difference 0..7
|
|
|
|
uabal2 v22.8h, v1.16b, v16.16b // absolute difference accumulate 8..15
|
2022-06-26 22:58:09 +02:00
|
|
|
mov v2.16b, v18.16b // pix3 -> pix2
|
|
|
|
mov v3.16b, v19.16b // pix3+1 -> pix2+1
|
|
|
|
|
|
|
|
b.ne 2b // loop if h > 0
|
2022-07-12 23:48:48 +02:00
|
|
|
|
|
|
|
add v4.8h, v21.8h, v22.8h
|
|
|
|
uaddlv s0, v4.8h // finish adding up accumulated values
|
2022-06-26 22:58:09 +02:00
|
|
|
fmov w0, s0 // copy result to general purpose register
|
|
|
|
ret
|
|
|
|
endfunc
|
2022-07-12 11:15:33 +02:00
|
|
|
|
|
|
|
function ff_pix_abs16_x2_neon, export=1
|
|
|
|
// x0 unused
|
|
|
|
// x1 uint8_t *pix1
|
|
|
|
// x2 uint8_t *pix2
|
|
|
|
// x3 ptrdiff_t stride
|
|
|
|
// w4 int h
|
|
|
|
|
|
|
|
cmp w4, #4
|
|
|
|
// initialize buffers
|
2022-07-12 23:48:48 +02:00
|
|
|
movi v16.8h, #0
|
|
|
|
movi v17.8h, #0
|
2022-07-12 11:15:33 +02:00
|
|
|
add x5, x2, #1 // pix2 + 1
|
|
|
|
b.lt 2f
|
|
|
|
|
|
|
|
// make 4 iterations at once
|
|
|
|
1:
|
|
|
|
|
|
|
|
// abs(pix1[0] - avg2(pix2[0], pix2[1]))
|
|
|
|
// avg2(a,b) = (((a) + (b) + 1) >> 1)
|
|
|
|
// abs(x) = (x < 0 ? -x : x)
|
|
|
|
|
|
|
|
ld1 {v1.16b}, [x2], x3
|
|
|
|
ld1 {v2.16b}, [x5], x3
|
|
|
|
urhadd v30.16b, v1.16b, v2.16b
|
|
|
|
ld1 {v0.16b}, [x1], x3
|
2022-07-12 23:48:48 +02:00
|
|
|
uabal v16.8h, v0.8b, v30.8b
|
2022-07-12 11:15:33 +02:00
|
|
|
ld1 {v4.16b}, [x2], x3
|
2022-07-12 23:48:48 +02:00
|
|
|
uabal2 v17.8h, v0.16b, v30.16b
|
2022-07-12 11:15:33 +02:00
|
|
|
ld1 {v5.16b}, [x5], x3
|
|
|
|
urhadd v29.16b, v4.16b, v5.16b
|
|
|
|
ld1 {v3.16b}, [x1], x3
|
|
|
|
uabal v16.8h, v3.8b, v29.8b
|
|
|
|
ld1 {v7.16b}, [x2], x3
|
|
|
|
uabal2 v17.8h, v3.16b, v29.16b
|
|
|
|
ld1 {v22.16b}, [x5], x3
|
|
|
|
urhadd v28.16b, v7.16b, v22.16b
|
|
|
|
ld1 {v6.16b}, [x1], x3
|
|
|
|
uabal v16.8h, v6.8b, v28.8b
|
|
|
|
ld1 {v24.16b}, [x2], x3
|
2022-07-12 23:48:48 +02:00
|
|
|
sub w4, w4, #4
|
2022-07-12 11:15:33 +02:00
|
|
|
uabal2 v17.8h, v6.16b, v28.16b
|
|
|
|
ld1 {v25.16b}, [x5], x3
|
|
|
|
urhadd v27.16b, v24.16b, v25.16b
|
|
|
|
ld1 {v23.16b}, [x1], x3
|
2022-07-12 23:48:48 +02:00
|
|
|
cmp w4, #4
|
2022-07-12 11:15:33 +02:00
|
|
|
uabal v16.8h, v23.8b, v27.8b
|
|
|
|
uabal2 v17.8h, v23.16b, v27.16b
|
|
|
|
|
|
|
|
b.ge 1b
|
|
|
|
cbz w4, 3f
|
|
|
|
|
|
|
|
// iterate by one
|
|
|
|
2:
|
|
|
|
ld1 {v1.16b}, [x2], x3
|
|
|
|
ld1 {v2.16b}, [x5], x3
|
2022-07-12 23:48:48 +02:00
|
|
|
subs w4, w4, #1
|
2022-07-12 11:15:33 +02:00
|
|
|
urhadd v29.16b, v1.16b, v2.16b
|
|
|
|
ld1 {v0.16b}, [x1], x3
|
2022-07-12 23:48:48 +02:00
|
|
|
uabal v16.8h, v0.8b, v29.8b
|
|
|
|
uabal2 v17.8h, v0.16b, v29.16b
|
2022-07-12 11:15:33 +02:00
|
|
|
|
|
|
|
|
|
|
|
b.ne 2b
|
|
|
|
|
|
|
|
3:
|
2022-07-12 23:48:48 +02:00
|
|
|
add v16.8h, v16.8h, v17.8h
|
|
|
|
uaddlv s16, v16.8h
|
|
|
|
fmov w0, s16
|
2022-07-12 11:15:33 +02:00
|
|
|
|
|
|
|
ret
|
|
|
|
endfunc
|
2022-08-16 14:20:12 +02:00
|
|
|
|
2022-08-16 14:20:14 +02:00
|
|
|
function ff_pix_abs16_y2_neon, export=1
|
|
|
|
// x0 unused
|
|
|
|
// x1 uint8_t *pix1
|
|
|
|
// x2 uint8_t *pix2
|
|
|
|
// x3 ptrdiff_t stride
|
|
|
|
// w4 int h
|
|
|
|
|
|
|
|
// initialize buffers
|
|
|
|
movi v29.8h, #0 // clear the accumulator
|
|
|
|
movi v28.8h, #0 // clear the accumulator
|
|
|
|
add x5, x2, x3 // pix2 + stride
|
|
|
|
cmp w4, #4
|
|
|
|
b.lt 2f
|
|
|
|
|
|
|
|
// make 4 iterations at once
|
|
|
|
1:
|
|
|
|
|
|
|
|
// abs(pix1[0], avg2(pix2[0], pix2[0 + stride]))
|
|
|
|
// avg2(a, b) = (((a) + (b) + 1) >> 1)
|
|
|
|
// abs(x) = (x < 0 ? (-x) : (x))
|
|
|
|
|
|
|
|
ld1 {v1.16b}, [x2], x3 // Load pix2 for first iteration
|
|
|
|
ld1 {v2.16b}, [x5], x3 // Load pix3 for first iteration
|
|
|
|
ld1 {v0.16b}, [x1], x3 // Load pix1 for first iteration
|
|
|
|
urhadd v30.16b, v1.16b, v2.16b // Rounding halving add, first iteration
|
|
|
|
ld1 {v4.16b}, [x2], x3 // Load pix2 for second iteration
|
|
|
|
ld1 {v5.16b}, [x5], x3 // Load pix3 for second iteartion
|
|
|
|
uabal v29.8h, v0.8b, v30.8b // Absolute difference of lower half, first iteration
|
|
|
|
uabal2 v28.8h, v0.16b, v30.16b // Absolute difference of upper half, first iteration
|
|
|
|
ld1 {v3.16b}, [x1], x3 // Load pix1 for second iteration
|
|
|
|
urhadd v27.16b, v4.16b, v5.16b // Rounding halving add, second iteration
|
|
|
|
ld1 {v7.16b}, [x2], x3 // Load pix2 for third iteration
|
|
|
|
ld1 {v20.16b}, [x5], x3 // Load pix3 for third iteration
|
|
|
|
uabal v29.8h, v3.8b, v27.8b // Absolute difference of lower half for second iteration
|
|
|
|
uabal2 v28.8h, v3.16b, v27.16b // Absolute difference of upper half for second iteration
|
|
|
|
ld1 {v6.16b}, [x1], x3 // Load pix1 for third iteration
|
|
|
|
urhadd v26.16b, v7.16b, v20.16b // Rounding halving add, third iteration
|
|
|
|
ld1 {v22.16b}, [x2], x3 // Load pix2 for fourth iteration
|
|
|
|
ld1 {v23.16b}, [x5], x3 // Load pix3 for fourth iteration
|
|
|
|
uabal v29.8h, v6.8b, v26.8b // Absolute difference of lower half for third iteration
|
|
|
|
uabal2 v28.8h, v6.16b, v26.16b // Absolute difference of upper half for third iteration
|
|
|
|
ld1 {v21.16b}, [x1], x3 // Load pix1 for fourth iteration
|
|
|
|
sub w4, w4, #4 // h-= 4
|
|
|
|
urhadd v25.16b, v22.16b, v23.16b // Rounding halving add
|
|
|
|
cmp w4, #4
|
|
|
|
uabal v29.8h, v21.8b, v25.8b // Absolute difference of lower half for fourth iteration
|
|
|
|
uabal2 v28.8h, v21.16b, v25.16b // Absolute difference of upper half for fourth iteration
|
|
|
|
|
|
|
|
b.ge 1b
|
|
|
|
cbz w4, 3f
|
|
|
|
|
|
|
|
// iterate by one
|
|
|
|
2:
|
|
|
|
|
|
|
|
ld1 {v1.16b}, [x2], x3 // Load pix2
|
|
|
|
ld1 {v2.16b}, [x5], x3 // Load pix3
|
|
|
|
subs w4, w4, #1
|
|
|
|
ld1 {v0.16b}, [x1], x3 // Load pix1
|
|
|
|
urhadd v30.16b, v1.16b, v2.16b // Rounding halving add
|
|
|
|
uabal v29.8h, v30.8b, v0.8b
|
|
|
|
uabal2 v28.8h, v30.16b, v0.16b
|
|
|
|
|
|
|
|
b.ne 2b
|
|
|
|
|
|
|
|
3:
|
|
|
|
add v29.8h, v29.8h, v28.8h // Add vectors together
|
|
|
|
uaddlv s16, v29.8h // Add up vector values
|
|
|
|
|
2022-08-18 11:14:15 +02:00
|
|
|
fmov w0, s16
|
2022-08-16 14:20:14 +02:00
|
|
|
|
|
|
|
ret
|
|
|
|
endfunc
|
|
|
|
|
2022-08-16 14:20:12 +02:00
|
|
|
function sse16_neon, export=1
|
|
|
|
// x0 - unused
|
|
|
|
// x1 - pix1
|
|
|
|
// x2 - pix2
|
|
|
|
// x3 - stride
|
|
|
|
// w4 - h
|
|
|
|
|
|
|
|
cmp w4, #4
|
|
|
|
movi v17.4s, #0
|
|
|
|
b.lt 2f
|
|
|
|
|
|
|
|
// Make 4 iterations at once
|
|
|
|
1:
|
|
|
|
|
|
|
|
// res = abs(pix1[0] - pix2[0])
|
|
|
|
// res * res
|
|
|
|
|
|
|
|
ld1 {v0.16b}, [x1], x3 // Load pix1 vector for first iteration
|
|
|
|
ld1 {v1.16b}, [x2], x3 // Load pix2 vector for first iteration
|
|
|
|
ld1 {v2.16b}, [x1], x3 // Load pix1 vector for second iteration
|
|
|
|
uabd v30.16b, v0.16b, v1.16b // Absolute difference, first iteration
|
|
|
|
ld1 {v3.16b}, [x2], x3 // Load pix2 vector for second iteration
|
|
|
|
umull v29.8h, v30.8b, v30.8b // Multiply lower half of vectors, first iteration
|
|
|
|
umull2 v28.8h, v30.16b, v30.16b // Multiply upper half of vectors, first iteration
|
|
|
|
uabd v27.16b, v2.16b, v3.16b // Absolute difference, second iteration
|
|
|
|
uadalp v17.4s, v29.8h // Pairwise add, first iteration
|
|
|
|
ld1 {v4.16b}, [x1], x3 // Load pix1 for third iteration
|
|
|
|
umull v26.8h, v27.8b, v27.8b // Mulitply lower half, second iteration
|
|
|
|
umull2 v25.8h, v27.16b, v27.16b // Multiply upper half, second iteration
|
|
|
|
ld1 {v5.16b}, [x2], x3 // Load pix2 for third iteration
|
|
|
|
uadalp v17.4s, v26.8h // Pairwise add and accumulate, second iteration
|
|
|
|
uabd v24.16b, v4.16b, v5.16b // Absolute difference, third iteration
|
|
|
|
ld1 {v6.16b}, [x1], x3 // Load pix1 for fourth iteration
|
|
|
|
uadalp v17.4s, v25.8h // Pairwise add and accumulate, second iteration
|
|
|
|
umull v23.8h, v24.8b, v24.8b // Multiply lower half, third iteration
|
|
|
|
umull2 v22.8h, v24.16b, v24.16b // Multiply upper half, third iteration
|
|
|
|
uadalp v17.4s, v23.8h // Pairwise add and accumulate, third iteration
|
|
|
|
ld1 {v7.16b}, [x2], x3 // Load pix2 for fouth iteration
|
|
|
|
uadalp v17.4s, v22.8h // Pairwise add and accumulate, third iteration
|
|
|
|
uabd v21.16b, v6.16b, v7.16b // Absolute difference, fourth iteration
|
|
|
|
uadalp v17.4s, v28.8h // Pairwise add and accumulate, first iteration
|
|
|
|
umull v20.8h, v21.8b, v21.8b // Multiply lower half, fourth iteration
|
|
|
|
sub w4, w4, #4 // h -= 4
|
|
|
|
umull2 v19.8h, v21.16b, v21.16b // Multiply upper half, fourth iteration
|
|
|
|
uadalp v17.4s, v20.8h // Pairwise add and accumulate, fourth iteration
|
|
|
|
cmp w4, #4
|
|
|
|
uadalp v17.4s, v19.8h // Pairwise add and accumulate, fourth iteration
|
|
|
|
b.ge 1b
|
|
|
|
|
|
|
|
cbz w4, 3f
|
|
|
|
|
|
|
|
// iterate by one
|
|
|
|
2:
|
|
|
|
|
|
|
|
ld1 {v0.16b}, [x1], x3 // Load pix1
|
|
|
|
ld1 {v1.16b}, [x2], x3 // Load pix2
|
|
|
|
|
|
|
|
uabd v30.16b, v0.16b, v1.16b
|
|
|
|
umull v29.8h, v30.8b, v30.8b
|
|
|
|
umull2 v28.8h, v30.16b, v30.16b
|
|
|
|
uadalp v17.4s, v29.8h
|
|
|
|
subs w4, w4, #1
|
|
|
|
uadalp v17.4s, v28.8h
|
|
|
|
|
|
|
|
b.ne 2b
|
|
|
|
|
|
|
|
3:
|
|
|
|
uaddlv d16, v17.4s // add up accumulator vector
|
|
|
|
|
|
|
|
fmov w0, s16
|
|
|
|
|
|
|
|
ret
|
|
|
|
endfunc
|
2022-08-16 14:20:13 +02:00
|
|
|
|
2022-08-16 14:20:15 +02:00
|
|
|
function sse8_neon, export=1
|
|
|
|
// x0 - unused
|
|
|
|
// x1 - pix1
|
|
|
|
// x2 - pix2
|
|
|
|
// x3 - stride
|
|
|
|
// w4 - h
|
|
|
|
|
|
|
|
movi v21.4s, #0
|
|
|
|
movi v20.4s, #0
|
|
|
|
cmp w4, #4
|
|
|
|
b.le 2f
|
|
|
|
|
|
|
|
// make 4 iterations at once
|
|
|
|
1:
|
|
|
|
|
|
|
|
// res = abs(pix1[0] - pix2[0])
|
|
|
|
// res * res
|
|
|
|
|
|
|
|
ld1 {v0.8b}, [x1], x3 // Load pix1 for first iteration
|
|
|
|
ld1 {v1.8b}, [x2], x3 // Load pix2 for second iteration
|
|
|
|
ld1 {v2.8b}, [x1], x3 // Load pix1 for second iteration
|
|
|
|
ld1 {v3.8b}, [x2], x3 // Load pix2 for second iteration
|
|
|
|
uabdl v30.8h, v0.8b, v1.8b // Absolute difference, first iteration
|
|
|
|
ld1 {v4.8b}, [x1], x3 // Load pix1 for third iteration
|
|
|
|
ld1 {v5.8b}, [x2], x3 // Load pix2 for third iteration
|
|
|
|
uabdl v29.8h, v2.8b, v3.8b // Absolute difference, second iteration
|
|
|
|
umlal v21.4s, v30.4h, v30.4h // Multiply lower half, first iteration
|
|
|
|
ld1 {v6.8b}, [x1], x3 // Load pix1 for fourth iteration
|
|
|
|
ld1 {v7.8b}, [x2], x3 // Load pix2 for fourth iteration
|
|
|
|
uabdl v28.8h, v4.8b, v5.8b // Absolute difference, third iteration
|
|
|
|
umlal v21.4s, v29.4h, v29.4h // Multiply lower half, second iteration
|
|
|
|
umlal2 v20.4s, v30.8h, v30.8h // Multiply upper half, first iteration
|
|
|
|
uabdl v27.8h, v6.8b, v7.8b // Absolute difference, fourth iteration
|
|
|
|
umlal v21.4s, v28.4h, v28.4h // Multiply lower half, third iteration
|
|
|
|
umlal2 v20.4s, v29.8h, v29.8h // Multiply upper half, second iteration
|
|
|
|
sub w4, w4, #4 // h -= 4
|
|
|
|
umlal2 v20.4s, v28.8h, v28.8h // Multiply upper half, third iteration
|
|
|
|
umlal v21.4s, v27.4h, v27.4h // Multiply lower half, fourth iteration
|
|
|
|
cmp w4, #4
|
|
|
|
umlal2 v20.4s, v27.8h, v27.8h // Multiply upper half, fourth iteration
|
|
|
|
b.ge 1b
|
|
|
|
|
|
|
|
cbz w4, 3f
|
|
|
|
|
|
|
|
// iterate by one
|
|
|
|
2:
|
|
|
|
ld1 {v0.8b}, [x1], x3 // Load pix1
|
|
|
|
ld1 {v1.8b}, [x2], x3 // Load pix2
|
|
|
|
subs w4, w4, #1
|
|
|
|
uabdl v30.8h, v0.8b, v1.8b
|
|
|
|
umlal v21.4s, v30.4h, v30.4h
|
|
|
|
umlal2 v20.4s, v30.8h, v30.8h
|
|
|
|
|
|
|
|
b.ne 2b
|
|
|
|
|
|
|
|
3:
|
|
|
|
add v21.4s, v21.4s, v20.4s // Add accumulator vectors together
|
|
|
|
uaddlv d17, v21.4s // Add up vector
|
|
|
|
|
|
|
|
fmov w0, s17
|
|
|
|
ret
|
|
|
|
|
|
|
|
endfunc
|
|
|
|
|
2022-08-16 14:20:13 +02:00
|
|
|
function sse4_neon, export=1
|
|
|
|
// x0 - unused
|
|
|
|
// x1 - pix1
|
|
|
|
// x2 - pix2
|
|
|
|
// x3 - stride
|
|
|
|
// w4 - h
|
|
|
|
|
|
|
|
movi v16.4s, #0 // clear the result accumulator
|
|
|
|
cmp w4, #4
|
|
|
|
b.le 2f
|
|
|
|
|
|
|
|
// make 4 iterations at once
|
|
|
|
1:
|
|
|
|
|
|
|
|
// res = abs(pix1[0] - pix2[0])
|
|
|
|
// res * res
|
|
|
|
|
|
|
|
ld1 {v0.s}[0], [x1], x3 // Load pix1, first iteration
|
|
|
|
ld1 {v1.s}[0], [x2], x3 // Load pix2, first iteration
|
|
|
|
ld1 {v2.s}[0], [x1], x3 // Load pix1, second iteration
|
|
|
|
ld1 {v3.s}[0], [x2], x3 // Load pix2, second iteration
|
|
|
|
uabdl v30.8h, v0.8b, v1.8b // Absolute difference, first iteration
|
|
|
|
ld1 {v4.s}[0], [x1], x3 // Load pix1, third iteration
|
|
|
|
ld1 {v5.s}[0], [x2], x3 // Load pix2, third iteration
|
|
|
|
uabdl v29.8h, v2.8b, v3.8b // Absolute difference, second iteration
|
|
|
|
umlal v16.4s, v30.4h, v30.4h // Multiply vectors, first iteration
|
|
|
|
ld1 {v6.s}[0], [x1], x3 // Load pix1, fourth iteration
|
|
|
|
ld1 {v7.s}[0], [x2], x3 // Load pix2, fourth iteration
|
|
|
|
uabdl v28.8h, v4.8b, v5.8b // Absolute difference, third iteration
|
|
|
|
umlal v16.4s, v29.4h, v29.4h // Multiply and accumulate, second iteration
|
|
|
|
sub w4, w4, #4
|
|
|
|
uabdl v27.8h, v6.8b, v7.8b // Absolue difference, fourth iteration
|
|
|
|
umlal v16.4s, v28.4h, v28.4h // Multiply and accumulate, third iteration
|
|
|
|
cmp w4, #4
|
|
|
|
umlal v16.4s, v27.4h, v27.4h // Multiply and accumulate, fourth iteration
|
|
|
|
b.ge 1b
|
|
|
|
|
|
|
|
cbz w4, 3f
|
|
|
|
|
|
|
|
// iterate by one
|
|
|
|
2:
|
|
|
|
ld1 {v0.s}[0], [x1], x3 // Load pix1
|
|
|
|
ld1 {v1.s}[0], [x2], x3 // Load pix2
|
|
|
|
uabdl v30.8h, v0.8b, v1.8b
|
|
|
|
subs w4, w4, #1
|
|
|
|
umlal v16.4s, v30.4h, v30.4h
|
|
|
|
|
|
|
|
b.ne 2b
|
|
|
|
|
|
|
|
3:
|
|
|
|
uaddlv d17, v16.4s // Add vector
|
|
|
|
fmov w0, s17
|
|
|
|
|
|
|
|
ret
|
|
|
|
endfunc
|
2022-09-08 11:25:03 +02:00
|
|
|
|
|
|
|
function vsad16_neon, export=1
|
|
|
|
// x0 unused
|
|
|
|
// x1 uint8_t *pix1
|
|
|
|
// x2 uint8_t *pix2
|
|
|
|
// x3 ptrdiff_t stride
|
|
|
|
// w4 int h
|
|
|
|
|
|
|
|
ld1 {v0.16b}, [x1], x3 // Load pix1[0], first iteration
|
|
|
|
ld1 {v1.16b}, [x2], x3 // Load pix2[0], first iteration
|
|
|
|
|
|
|
|
sub w4, w4, #1 // we need to make h-1 iterations
|
|
|
|
movi v16.8h, #0
|
|
|
|
|
|
|
|
cmp w4, #3 // check if we can make 3 iterations at once
|
|
|
|
usubl v31.8h, v0.8b, v1.8b // Signed difference pix1[0] - pix2[0], first iteration
|
|
|
|
usubl2 v30.8h, v0.16b, v1.16b // Signed difference pix1[0] - pix2[0], first iteration
|
|
|
|
|
|
|
|
b.lt 2f
|
|
|
|
|
|
|
|
1:
|
|
|
|
// abs(pix1[0] - pix2[0] - pix1[0 + stride] + pix2[0 + stride])
|
|
|
|
ld1 {v0.16b}, [x1], x3 // Load pix1[0 + stride], first iteration
|
|
|
|
ld1 {v1.16b}, [x2], x3 // Load pix2[0 + stride], first iteration
|
|
|
|
ld1 {v2.16b}, [x1], x3 // Load pix1[0 + stride], second iteration
|
|
|
|
ld1 {v3.16b}, [x2], x3 // Load pix2[0 + stride], second iteration
|
|
|
|
usubl v29.8h, v0.8b, v1.8b // Signed difference pix1[0 + stride] - pix2[0 + stride], first iteration
|
|
|
|
usubl2 v28.8h, v0.16b, v1.16b // Signed difference pix1[0 + stride] - pix2[0 + stride], first iteration
|
|
|
|
ld1 {v4.16b}, [x1], x3 // Load pix1[0 + stride], third iteration
|
|
|
|
ld1 {v5.16b}, [x2], x3 // Load pix2[0 + stride], third iteration
|
|
|
|
usubl v27.8h, v2.8b, v3.8b // Signed difference pix1[0 + stride] - pix2[0 + stride], second iteration
|
|
|
|
saba v16.8h, v31.8h, v29.8h // Signed absolute difference and accumulate the result. first iteration
|
|
|
|
usubl2 v26.8h, v2.16b, v3.16b // Signed difference pix1[0 + stride] - pix2[0 + stride], second iteration
|
|
|
|
saba v16.8h, v30.8h, v28.8h // Signed absolute difference and accumulate the result. first iteration
|
|
|
|
usubl v25.8h, v4.8b, v5.8b // Signed difference pix1[0 + stride] - pix2[0 + stride], third iteration
|
|
|
|
usubl2 v24.8h, v4.16b, v5.16b // Signed difference pix1[0 + stride] - pix2[0 + stride], third iteration
|
|
|
|
saba v16.8h, v29.8h, v27.8h // Signed absolute difference and accumulate the result. second iteration
|
|
|
|
mov v31.16b, v25.16b
|
|
|
|
saba v16.8h, v28.8h, v26.8h // Signed absolute difference and accumulate the result. second iteration
|
|
|
|
sub w4, w4, #3 // h -= 3
|
|
|
|
mov v30.16b, v24.16b
|
|
|
|
saba v16.8h, v27.8h, v25.8h // Signed absolute difference and accumulate the result. third iteration
|
|
|
|
cmp w4, #3
|
|
|
|
saba v16.8h, v26.8h, v24.8h // Signed absolute difference and accumulate the result. third iteration
|
|
|
|
|
|
|
|
b.ge 1b
|
|
|
|
cbz w4, 3f
|
|
|
|
2:
|
|
|
|
ld1 {v0.16b}, [x1], x3
|
|
|
|
ld1 {v1.16b}, [x2], x3
|
|
|
|
subs w4, w4, #1
|
|
|
|
usubl v29.8h, v0.8b, v1.8b
|
|
|
|
usubl2 v28.8h, v0.16b, v1.16b
|
|
|
|
saba v16.8h, v31.8h, v29.8h
|
|
|
|
mov v31.16b, v29.16b
|
|
|
|
saba v16.8h, v30.8h, v28.8h
|
|
|
|
mov v30.16b, v28.16b
|
|
|
|
|
|
|
|
b.ne 2b
|
|
|
|
3:
|
|
|
|
uaddlv s17, v16.8h
|
|
|
|
fmov w0, s17
|
|
|
|
|
|
|
|
ret
|
|
|
|
endfunc
|
2022-09-08 11:25:04 +02:00
|
|
|
|
|
|
|
function vsse16_neon, export=1
|
|
|
|
// x0 unused
|
|
|
|
// x1 uint8_t *pix1
|
|
|
|
// x2 uint8_t *pix2
|
|
|
|
// x3 ptrdiff_t stride
|
|
|
|
// w4 int h
|
|
|
|
|
|
|
|
ld1 {v0.16b}, [x1], x3 // Load pix1[0], first iteration
|
|
|
|
ld1 {v1.16b}, [x2], x3 // Load pix2[0], first iteration
|
|
|
|
|
|
|
|
sub w4, w4, #1 // we need to make h-1 iterations
|
|
|
|
movi v16.4s, #0
|
|
|
|
movi v17.4s, #0
|
|
|
|
|
|
|
|
cmp w4, #3 // check if we can make 3 iterations at once
|
|
|
|
usubl v31.8h, v0.8b, v1.8b // Signed difference of pix1[0] - pix2[0], first iteration
|
|
|
|
usubl2 v30.8h, v0.16b, v1.16b // Signed difference of pix1[0] - pix2[0], first iteration
|
|
|
|
b.le 2f
|
|
|
|
|
|
|
|
|
|
|
|
1:
|
|
|
|
// x = abs(pix1[0] - pix2[0] - pix1[0 + stride] + pix2[0 + stride])
|
|
|
|
// res = (x) * (x)
|
|
|
|
ld1 {v0.16b}, [x1], x3 // Load pix1[0 + stride], first iteration
|
|
|
|
ld1 {v1.16b}, [x2], x3 // Load pix2[0 + stride], first iteration
|
|
|
|
ld1 {v2.16b}, [x1], x3 // Load pix1[0 + stride], second iteration
|
|
|
|
ld1 {v3.16b}, [x2], x3 // Load pix2[0 + stride], second iteration
|
|
|
|
usubl v29.8h, v0.8b, v1.8b
|
|
|
|
usubl2 v28.8h, v0.16b, v1.16b
|
|
|
|
ld1 {v4.16b}, [x1], x3 // Load pix1[0 + stride], third iteration
|
|
|
|
ld1 {v5.16b}, [x2], x3 // Load pix1[0 + stride], third iteration
|
|
|
|
sabd v31.8h, v31.8h, v29.8h
|
|
|
|
sabd v30.8h, v30.8h, v28.8h
|
|
|
|
usubl v27.8h, v2.8b, v3.8b
|
|
|
|
usubl2 v26.8h, v2.16b, v3.16b
|
|
|
|
usubl v25.8h, v4.8b, v5.8b
|
|
|
|
usubl2 v24.8h, v4.16b, v5.16b
|
|
|
|
sabd v29.8h, v29.8h, v27.8h
|
|
|
|
sabd v27.8h, v27.8h, v25.8h
|
|
|
|
umlal v16.4s, v31.4h, v31.4h
|
|
|
|
umlal2 v17.4s, v31.8h, v31.8h
|
|
|
|
sabd v28.8h, v28.8h, v26.8h
|
|
|
|
sabd v26.8h, v26.8h, v24.8h
|
|
|
|
umlal v16.4s, v30.4h, v30.4h
|
|
|
|
umlal2 v17.4s, v30.8h, v30.8h
|
|
|
|
mov v31.16b, v25.16b
|
|
|
|
umlal v16.4s, v29.4h, v29.4h
|
|
|
|
umlal2 v17.4s, v29.8h, v29.8h
|
|
|
|
mov v30.16b, v24.16b
|
|
|
|
umlal v16.4s, v28.4h, v28.4h
|
|
|
|
umlal2 v17.4s, v28.8h, v28.8h
|
|
|
|
sub w4, w4, #3
|
|
|
|
umlal v16.4s, v27.4h, v27.4h
|
|
|
|
umlal2 v17.4s, v27.8h, v27.8h
|
|
|
|
cmp w4, #3
|
|
|
|
umlal v16.4s, v26.4h, v26.4h
|
|
|
|
umlal2 v17.4s, v26.8h, v26.8h
|
|
|
|
|
|
|
|
b.ge 1b
|
|
|
|
|
|
|
|
cbz w4, 3f
|
|
|
|
|
|
|
|
// iterate by once
|
|
|
|
2:
|
|
|
|
ld1 {v0.16b}, [x1], x3
|
|
|
|
ld1 {v1.16b}, [x2], x3
|
|
|
|
subs w4, w4, #1
|
|
|
|
usubl v29.8h, v0.8b, v1.8b
|
|
|
|
usubl2 v28.8h, v0.16b, v1.16b
|
|
|
|
sabd v31.8h, v31.8h, v29.8h
|
|
|
|
sabd v30.8h, v30.8h, v28.8h
|
|
|
|
umlal v16.4s, v31.4h, v31.4h
|
|
|
|
umlal2 v17.4s, v31.8h, v31.8h
|
|
|
|
mov v31.16b, v29.16b
|
|
|
|
umlal v16.4s, v30.4h, v30.4h
|
|
|
|
umlal2 v17.4s, v30.8h, v30.8h
|
|
|
|
mov v30.16b, v28.16b
|
|
|
|
b.ne 2b
|
|
|
|
|
|
|
|
3:
|
|
|
|
add v16.4s, v16.4s, v17.4s
|
|
|
|
uaddlv d17, v16.4s
|
|
|
|
fmov w0, s17
|
|
|
|
|
|
|
|
ret
|
|
|
|
endfunc
|