[x265] [PATCH v2] AArch64: Optimize and clean up addAvg Neon and SVE2 functions

chen chenm003 at 163.com
Thu Jun 26 05:40:10 UTC 2025


Thank for the revised patch

Some comments,

+            vst1q_u16(dst + w, vreinterpretq_u16_s16(d0_lo));
+            vst1q_u16(dst + w + 8, vreinterpretq_u16_s16(d0_hi));

How about performance on vst1q_u16_x2 or vst2q_u64 ?




+            int16x8_t d0 = vrsraq_n_s16(addon, vaddq_s16(s0, s1), shiftNum);
+            d0 = vminq_s16(d0, vdupq_n_s16((1 << X265_DEPTH) - 1));
+            d0 = vmaxq_s16(d0, vdupq_n_s16(0));

vrsraq_n_s16 looks not single instruction, is it faster than SQRSHRUN ?





At 2025-06-25 17:03:26, "Li Zhang" <li.zhang2 at arm.com> wrote:
>Extend the neon intrinsics implementation to support all block sizes and
>optimize it to use rounding-shift-and-accumulate instead of separate
>widening, add, and shift steps. Also unroll the loops for larger block
>sizes to enable the compiler to emit LDP and STP instructions.
>
>Delete the Neon and SVE2 assembly implementations as they are 1-2x
>slower than Neon intrinsics implementation.
>---
> source/common/aarch64/asm-primitives.cpp |  16 -
> source/common/aarch64/mc-a-sve2.S        | 606 -----------------------
> source/common/aarch64/mc-a.S             | 341 -------------
> source/common/aarch64/mem-neon.h         |  20 +
> source/common/aarch64/pixel-prim.cpp     | 207 +++++---
> 5 files changed, 169 insertions(+), 1021 deletions(-)
>
>diff --git a/source/common/aarch64/asm-primitives.cpp b/source/common/aarch64/asm-primitives.cpp
>index 55aaecfa7..34364ef8e 100644
>--- a/source/common/aarch64/asm-primitives.cpp
>+++ b/source/common/aarch64/asm-primitives.cpp
>@@ -462,14 +462,6 @@ void setupNeonPrimitives(EncoderPrimitives &p)
>     ALL_LUMA_PU(pixelavg_pp[NONALIGNED], pixel_avg_pp, neon);
>     ALL_LUMA_PU(pixelavg_pp[ALIGNED], pixel_avg_pp, neon);
> 
>-    // addAvg
>-    ALL_LUMA_PU(addAvg[NONALIGNED], addAvg, neon);
>-    ALL_LUMA_PU(addAvg[ALIGNED], addAvg, neon);
>-    ALL_CHROMA_420_PU(addAvg[NONALIGNED], addAvg, neon);
>-    ALL_CHROMA_422_PU(addAvg[NONALIGNED], addAvg, neon);
>-    ALL_CHROMA_420_PU(addAvg[ALIGNED], addAvg, neon);
>-    ALL_CHROMA_422_PU(addAvg[ALIGNED], addAvg, neon);
>-
>     // calc_Residual
>     p.cu[BLOCK_4x4].calcresidual[NONALIGNED]   = PFX(getResidual4_neon);
>     p.cu[BLOCK_8x8].calcresidual[NONALIGNED]   = PFX(getResidual8_neon);
>@@ -597,14 +589,6 @@ void setupSve2Primitives(EncoderPrimitives &p)
>     LUMA_PU_MULTIPLE_ARCHS_3(pixelavg_pp[NONALIGNED], pixel_avg_pp, sve2);
>     LUMA_PU_MULTIPLE_ARCHS_3(pixelavg_pp[ALIGNED], pixel_avg_pp, sve2);
> 
>-    // addAvg
>-    LUMA_PU_CAN_USE_SVE2(addAvg[NONALIGNED], addAvg);
>-    LUMA_PU_CAN_USE_SVE2(addAvg[ALIGNED], addAvg);
>-    CHROMA_420_PU_MULTIPLE_ARCHS(addAvg[NONALIGNED], addAvg, sve2);
>-    CHROMA_420_PU_MULTIPLE_ARCHS(addAvg[ALIGNED], addAvg, sve2);
>-    CHROMA_422_PU_CAN_USE_SVE2(addAvg[NONALIGNED], addAvg);
>-    CHROMA_422_PU_CAN_USE_SVE2(addAvg[ALIGNED], addAvg);
>-
>     // calc_Residual
>     p.cu[BLOCK_16x16].calcresidual[NONALIGNED] = PFX(getResidual16_sve2);
>     p.cu[BLOCK_32x32].calcresidual[NONALIGNED] = PFX(getResidual32_sve2);
>diff --git a/source/common/aarch64/mc-a-sve2.S b/source/common/aarch64/mc-a-sve2.S
>index 00fb0048f..fc0a6f3e8 100644
>--- a/source/common/aarch64/mc-a-sve2.S
>+++ b/source/common/aarch64/mc-a-sve2.S
>@@ -298,609 +298,3 @@ pixel_avg_pp_64xN_sve2 16
> pixel_avg_pp_64xN_sve2 32
> pixel_avg_pp_64xN_sve2 48
> pixel_avg_pp_64xN_sve2 64
>-
>-// void addAvg(const int16_t* src0, const int16_t* src1, pixel* dst, intptr_t src0Stride, intptr_t src1Stride, intptr_t dstStride)
>-
>-.macro addAvg_2xN_sve2 h
>-function PFX(addAvg_2x\h\()_sve2)
>-    ptrue           p0.s, vl2
>-    ptrue           p1.h, vl4
>-    ptrue           p2.h, vl2
>-.rept \h / 2
>-    ld1rw           {z0.s}, p0/z, [x0]
>-    ld1rw           {z1.s}, p0/z, [x1]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    ld1rw           {z2.s}, p0/z, [x0]
>-    ld1rw           {z3.s}, p0/z, [x1]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    add             z0.h, p1/m, z0.h, z1.h
>-    add             z2.h, p1/m, z2.h, z3.h
>-    sqrshrnb        z0.b, z0.h, #7
>-    add             z0.b, z0.b, #0x80
>-    sqrshrnb        z2.b, z2.h, #7
>-    add             z2.b, z2.b, #0x80
>-    st1b            {z0.h}, p2, [x2]
>-    add             x2, x2, x5
>-    st1b            {z2.h}, p2, [x2]
>-    add             x2, x2, x5
>-.endr
>-    ret
>-endfunc
>-.endm
>-
>-addAvg_2xN_sve2 4
>-addAvg_2xN_sve2 8
>-addAvg_2xN_sve2 16
>-
>-.macro addAvg_6xN_sve2 h
>-function PFX(addAvg_6x\h\()_sve2)
>-    mov             w12, #\h / 2
>-    ptrue           p0.b, vl16
>-    ptrue           p2.h, vl6
>-.Loop_sve2_addavg_6x\h\():
>-    sub             w12, w12, #1
>-    ld1b            {z0.b}, p0/z, [x0]
>-    ld1b            {z1.b}, p0/z, [x1]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    ld1b            {z2.b}, p0/z, [x0]
>-    ld1b            {z3.b}, p0/z, [x1]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    add             z0.h, p0/m, z0.h, z1.h
>-    add             z2.h, p0/m, z2.h, z3.h
>-    sqrshrnb        z0.b, z0.h, #7
>-    sqrshrnb        z2.b, z2.h, #7
>-    add             z0.b, z0.b, #0x80
>-    add             z2.b, z2.b, #0x80
>-    st1b            {z0.h}, p2, [x2]
>-    add             x2, x2, x5
>-    st1b            {z2.h}, p2, [x2]
>-    add             x2, x2, x5
>-    cbnz            w12, .Loop_sve2_addavg_6x\h
>-    ret
>-endfunc
>-.endm
>-
>-addAvg_6xN_sve2 8
>-addAvg_6xN_sve2 16
>-
>-.macro addAvg_8xN_sve2 h
>-function PFX(addAvg_8x\h\()_sve2)
>-    ptrue           p0.b, vl16
>-.rept \h / 2
>-    ld1b            {z0.b}, p0/z, [x0]
>-    ld1b            {z1.b}, p0/z, [x1]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    ld1b            {z2.b}, p0/z, [x0]
>-    ld1b            {z3.b}, p0/z, [x1]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    add             z0.h, p0/m, z0.h, z1.h
>-    add             z2.h, p0/m, z2.h, z3.h
>-    sqrshrnb        z0.b, z0.h, #7
>-    add             z0.b, z0.b, #0x80
>-    sqrshrnb        z2.b, z2.h, #7
>-    add             z2.b, z2.b, #0x80
>-    st1b            {z0.h}, p0, [x2]
>-    add             x2, x2, x5
>-    st1b            {z2.h}, p0, [x2]
>-    add             x2, x2, x5
>-.endr
>-    ret
>-endfunc
>-.endm
>-
>-.macro addAvg_8xN1_sve2 h
>-function PFX(addAvg_8x\h\()_sve2)
>-    mov             w12, #\h / 2
>-    ptrue           p0.b, vl16
>-.Loop_sve2_addavg_8x\h\():
>-    sub             w12, w12, #1
>-    ld1b            {z0.b}, p0/z, [x0]
>-    ld1b            {z1.b}, p0/z, [x1]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    ld1b            {z2.b}, p0/z, [x0]
>-    ld1b            {z3.b}, p0/z, [x1]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    add             z0.h, p0/m, z0.h, z1.h
>-    add             z2.h, p0/m, z2.h, z3.h
>-    sqrshrnb        z0.b, z0.h, #7
>-    add             z0.b, z0.b, #0x80
>-    sqrshrnb        z2.b, z2.h, #7
>-    add             z2.b, z2.b, #0x80
>-    st1b            {z0.h}, p0, [x2]
>-    add             x2, x2, x5
>-    st1b            {z2.h}, p0, [x2]
>-    add             x2, x2, x5
>-    cbnz            w12, .Loop_sve2_addavg_8x\h
>-    ret
>-endfunc
>-.endm
>-
>-addAvg_8xN_sve2 2
>-addAvg_8xN_sve2 4
>-addAvg_8xN_sve2 6
>-addAvg_8xN_sve2 8
>-addAvg_8xN_sve2 12
>-addAvg_8xN_sve2 16
>-addAvg_8xN1_sve2 32
>-addAvg_8xN1_sve2 64
>-
>-.macro addAvg_12xN_sve2 h
>-function PFX(addAvg_12x\h\()_sve2)
>-    mov             w12, #\h
>-    rdvl            x9, #1
>-    cmp             x9, #16
>-    bgt             .vl_gt_16_addAvg_12x\h
>-    ptrue           p0.b, vl16
>-    ptrue           p1.b, vl8
>-.Loop_sve2_addavg_12x\h\():
>-    sub             w12, w12, #1
>-    ld1b            {z0.b}, p0/z, [x0]
>-    ld1b            {z1.b}, p0/z, [x1]
>-    ld1b            {z2.b}, p1/z, [x0, #1, mul vl]
>-    ld1b            {z3.b}, p1/z, [x1, #1, mul vl]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    add             z0.h, p0/m, z0.h, z1.h
>-    add             z2.h, p1/m, z2.h, z3.h
>-    sqrshrnb        z0.b, z0.h, #7
>-    add             z0.b, z0.b, #0x80
>-    sqrshrnb        z2.b, z2.h, #7
>-    add             z2.b, z2.b, #0x80
>-    st1b            {z0.h}, p0, [x2]
>-    st1b            {z2.h}, p1, [x2, #1, mul vl]
>-    add             x2, x2, x5
>-    cbnz            w12, .Loop_sve2_addavg_12x\h
>-    ret
>-.vl_gt_16_addAvg_12x\h\():
>-    mov             x10, #24
>-    mov             x11, #0
>-    whilelt         p0.b, x11, x10
>-.Loop_sve2_gt_16_addavg_12x\h\():
>-    sub             w12, w12, #1
>-    ld1b            {z0.b}, p0/z, [x0]
>-    ld1b            {z1.b}, p0/z, [x1]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    add             z0.h, p0/m, z0.h, z1.h
>-    sqrshrnb        z0.b, z0.h, #7
>-    add             z0.b, z0.b, #0x80
>-    sqrshrnb        z2.b, z2.h, #7
>-    add             z2.b, z2.b, #0x80
>-    st1b            {z0.h}, p0, [x2]
>-    add             x2, x2, x5
>-    cbnz            w12, .Loop_sve2_gt_16_addavg_12x\h
>-    ret
>-endfunc
>-.endm
>-
>-addAvg_12xN_sve2 16
>-addAvg_12xN_sve2 32
>-
>-.macro addAvg_16xN_sve2 h
>-function PFX(addAvg_16x\h\()_sve2)
>-    mov             w12, #\h
>-    rdvl            x9, #1
>-    cmp             x9, #16
>-    bgt             .vl_gt_16_addAvg_16x\h
>-    ptrue           p0.b, vl16
>-.Loop_eq_16_sve2_addavg_16x\h\():
>-    sub             w12, w12, #1
>-    ld1b            {z0.b}, p0/z, [x0]
>-    ld1b            {z1.b}, p0/z, [x1]
>-    ld1b            {z2.b}, p0/z, [x0, #1, mul vl]
>-    ld1b            {z3.b}, p0/z, [x1, #1, mul vl]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    add             z0.h, p0/m, z0.h, z1.h
>-    add             z2.h, p0/m, z2.h, z3.h
>-    sqrshrnb        z0.b, z0.h, #7
>-    add             z0.b, z0.b, #0x80
>-    sqrshrnb        z2.b, z2.h, #7
>-    add             z2.b, z2.b, #0x80
>-    st1b            {z0.h}, p0, [x2]
>-    st1b            {z2.h}, p0, [x2, #1, mul vl]
>-    add             x2, x2, x5
>-    cbnz            w12, .Loop_eq_16_sve2_addavg_16x\h
>-    ret
>-.vl_gt_16_addAvg_16x\h\():
>-    ptrue           p0.b, vl32
>-.Loop_gt_16_sve2_addavg_16x\h\():
>-    sub             w12, w12, #1
>-    ld1b            {z0.b}, p0/z, [x0]
>-    ld1b            {z1.b}, p0/z, [x1]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    add             z0.h, p0/m, z0.h, z1.h
>-    sqrshrnb        z0.b, z0.h, #7
>-    add             z0.b, z0.b, #0x80
>-    st1b            {z0.h}, p0, [x2]
>-    add             x2, x2, x5
>-    cbnz            w12, .Loop_gt_16_sve2_addavg_16x\h
>-    ret
>-endfunc
>-.endm
>-
>-addAvg_16xN_sve2 4
>-addAvg_16xN_sve2 8
>-addAvg_16xN_sve2 12
>-addAvg_16xN_sve2 16
>-addAvg_16xN_sve2 24
>-addAvg_16xN_sve2 32
>-addAvg_16xN_sve2 64
>-
>-.macro addAvg_24xN_sve2 h
>-function PFX(addAvg_24x\h\()_sve2)
>-    mov             w12, #\h
>-    rdvl            x9, #1
>-    cmp             x9, #16
>-    bgt             .vl_gt_16_addAvg_24x\h
>-    addAvg_start
>-.Loop_eq_16_sve2_addavg_24x\h\():
>-    sub             w12, w12, #1
>-    ld1             {v0.16b-v2.16b}, [x0], x3
>-    ld1             {v3.16b-v5.16b}, [x1], x4
>-    addavg_1        v0, v3
>-    addavg_1        v1, v4
>-    addavg_1        v2, v5
>-    sqxtun          v0.8b, v0.8h
>-    sqxtun          v1.8b, v1.8h
>-    sqxtun          v2.8b, v2.8h
>-    st1             {v0.8b-v2.8b}, [x2], x5
>-    cbnz            w12, .Loop_eq_16_sve2_addavg_24x\h
>-    ret
>-.vl_gt_16_addAvg_24x\h\():
>-    cmp             x9, #48
>-    bgt             .vl_gt_48_addAvg_24x\h
>-    ptrue           p0.b, vl32
>-    ptrue           p1.b, vl16
>-.Loop_gt_16_sve2_addavg_24x\h\():
>-    sub             w12, w12, #1
>-    ld1b            {z0.b}, p0/z, [x0]
>-    ld1b            {z1.b}, p1/z, [x0, #1, mul vl]
>-    ld1b            {z2.b}, p0/z, [x1]
>-    ld1b            {z3.b}, p1/z, [x1, #1, mul vl]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    add             z0.h, p0/m, z0.h, z2.h
>-    add             z1.h, p1/m, z1.h, z3.h
>-    sqrshrnb        z0.b, z0.h, #7
>-    add             z0.b, z0.b, #0x80
>-    sqrshrnb        z1.b, z1.h, #7
>-    add             z1.b, z1.b, #0x80
>-    st1b            {z0.h}, p0, [x2]
>-    st1b            {z1.h}, p1, [x2, #1, mul vl]
>-    add             x2, x2, x5
>-    cbnz            w12, .Loop_gt_16_sve2_addavg_24x\h
>-    ret
>-.vl_gt_48_addAvg_24x\h\():
>-    mov             x10, #48
>-    mov             x11, #0
>-    whilelt         p0.b, x11, x10
>-.Loop_gt_48_sve2_addavg_24x\h\():
>-    sub             w12, w12, #1
>-    ld1b            {z0.b}, p0/z, [x0]
>-    ld1b            {z2.b}, p0/z, [x1]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    add             z0.h, p0/m, z0.h, z2.h
>-    sqrshrnb        z0.b, z0.h, #7
>-    add             z0.b, z0.b, #0x80
>-    st1b            {z0.h}, p0, [x2]
>-    add             x2, x2, x5
>-    cbnz            w12, .Loop_gt_48_sve2_addavg_24x\h
>-    ret
>-endfunc
>-.endm
>-
>-addAvg_24xN_sve2 32
>-addAvg_24xN_sve2 64
>-
>-.macro addAvg_32xN_sve2 h
>-function PFX(addAvg_32x\h\()_sve2)
>-    mov             w12, #\h
>-    rdvl            x9, #1
>-    cmp             x9, #16
>-    bgt             .vl_gt_16_addAvg_32x\h
>-    ptrue           p0.b, vl16
>-.Loop_eq_16_sve2_addavg_32x\h\():
>-    sub             w12, w12, #1
>-    ld1b            {z0.b}, p0/z, [x0]
>-    ld1b            {z1.b}, p0/z, [x0, #1, mul vl]
>-    ld1b            {z2.b}, p0/z, [x0, #2, mul vl]
>-    ld1b            {z3.b}, p0/z, [x0, #3, mul vl]
>-    ld1b            {z4.b}, p0/z, [x1]
>-    ld1b            {z5.b}, p0/z, [x1, #1, mul vl]
>-    ld1b            {z6.b}, p0/z, [x1, #2, mul vl]
>-    ld1b            {z7.b}, p0/z, [x1, #3, mul vl]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    add             z0.h, p0/m, z0.h, z4.h
>-    add             z1.h, p0/m, z1.h, z5.h
>-    add             z2.h, p0/m, z2.h, z6.h
>-    add             z3.h, p0/m, z3.h, z7.h
>-    sqrshrnb        z0.b, z0.h, #7
>-    add             z0.b, z0.b, #0x80
>-    sqrshrnb        z1.b, z1.h, #7
>-    add             z1.b, z1.b, #0x80
>-    sqrshrnb        z2.b, z2.h, #7
>-    add             z2.b, z2.b, #0x80
>-    sqrshrnb        z3.b, z3.h, #7
>-    add             z3.b, z3.b, #0x80
>-    st1b            {z0.h}, p0, [x2]
>-    st1b            {z1.h}, p0, [x2, #1, mul vl]
>-    st1b            {z2.h}, p0, [x2, #2, mul vl]
>-    st1b            {z3.h}, p0, [x2, #3, mul vl]
>-    add             x2, x2, x5
>-    cbnz            w12, .Loop_eq_16_sve2_addavg_32x\h
>-    ret
>-.vl_gt_16_addAvg_32x\h\():
>-    cmp             x9, #48
>-    bgt             .vl_gt_48_addAvg_32x\h
>-    ptrue           p0.b, vl32
>-.Loop_gt_eq_32_sve2_addavg_32x\h\():
>-    sub             w12, w12, #1
>-    ld1b            {z0.b}, p0/z, [x0]
>-    ld1b            {z1.b}, p0/z, [x0, #1, mul vl]
>-    ld1b            {z2.b}, p0/z, [x1]
>-    ld1b            {z3.b}, p0/z, [x1, #1, mul vl]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    add             z0.h, p0/m, z0.h, z2.h
>-    add             z1.h, p0/m, z1.h, z3.h
>-    sqrshrnb        z0.b, z0.h, #7
>-    add             z0.b, z0.b, #0x80
>-    sqrshrnb        z1.b, z1.h, #7
>-    add             z1.b, z1.b, #0x80
>-    st1b            {z0.h}, p0, [x2]
>-    st1b            {z1.h}, p0, [x2, #1, mul vl]
>-    add             x2, x2, x5
>-    cbnz            w12, .Loop_gt_eq_32_sve2_addavg_32x\h
>-    ret
>-.vl_gt_48_addAvg_32x\h\():
>-    ptrue           p0.b, vl64
>-.Loop_eq_64_sve2_addavg_32x\h\():
>-    sub             w12, w12, #1
>-    ld1b            {z0.b}, p0/z, [x0]
>-    ld1b            {z1.b}, p0/z, [x1]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    add             z0.h, p0/m, z0.h, z1.h
>-    sqrshrnb        z0.b, z0.h, #7
>-    add             z0.b, z0.b, #0x80
>-    st1b            {z0.h}, p0, [x2]
>-    add             x2, x2, x5
>-    cbnz            w12, .Loop_eq_64_sve2_addavg_32x\h
>-    ret
>-endfunc
>-.endm
>-
>-addAvg_32xN_sve2 8
>-addAvg_32xN_sve2 16
>-addAvg_32xN_sve2 24
>-addAvg_32xN_sve2 32
>-addAvg_32xN_sve2 48
>-addAvg_32xN_sve2 64
>-
>-function PFX(addAvg_48x64_sve2)
>-    mov             w12, #64
>-    rdvl            x9, #1
>-    cmp             x9, #16
>-    bgt             .vl_gt_16_addAvg_48x64
>-    addAvg_start
>-    sub             x3, x3, #64
>-    sub             x4, x4, #64
>-.Loop_eq_16_sve2_addavg_48x64:
>-    sub             w12, w12, #1
>-    ld1             {v0.8h-v3.8h}, [x0], #64
>-    ld1             {v4.8h-v7.8h}, [x1], #64
>-    ld1             {v20.8h-v21.8h}, [x0], x3
>-    ld1             {v22.8h-v23.8h}, [x1], x4
>-    addavg_1        v0, v4
>-    addavg_1        v1, v5
>-    addavg_1        v2, v6
>-    addavg_1        v3, v7
>-    addavg_1        v20, v22
>-    addavg_1        v21, v23
>-    sqxtun          v0.8b, v0.8h
>-    sqxtun2         v0.16b, v1.8h
>-    sqxtun          v1.8b, v2.8h
>-    sqxtun2         v1.16b, v3.8h
>-    sqxtun          v2.8b, v20.8h
>-    sqxtun2         v2.16b, v21.8h
>-    st1             {v0.16b-v2.16b}, [x2], x5
>-    cbnz            w12, .Loop_eq_16_sve2_addavg_48x64
>-    ret
>-.vl_gt_16_addAvg_48x64:
>-    cmp             x9, #48
>-    bgt             .vl_gt_48_addAvg_48x64
>-    ptrue           p0.b, vl32
>-.Loop_gt_eq_32_sve2_addavg_48x64:
>-    sub             w12, w12, #1
>-    ld1b            {z0.b}, p0/z, [x0]
>-    ld1b            {z1.b}, p0/z, [x0, #1, mul vl]
>-    ld1b            {z2.b}, p0/z, [x0, #2, mul vl]
>-    ld1b            {z4.b}, p0/z, [x1]
>-    ld1b            {z5.b}, p0/z, [x1, #1, mul vl]
>-    ld1b            {z6.b}, p0/z, [x1, #2, mul vl]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    add             z0.h, p0/m, z0.h, z4.h
>-    add             z1.h, p0/m, z1.h, z5.h
>-    add             z2.h, p0/m, z2.h, z6.h
>-    sqrshrnb        z0.b, z0.h, #7
>-    add             z0.b, z0.b, #0x80
>-    sqrshrnb        z1.b, z1.h, #7
>-    add             z1.b, z1.b, #0x80
>-    sqrshrnb        z2.b, z2.h, #7
>-    add             z2.b, z2.b, #0x80
>-    st1b            {z0.h}, p0, [x2]
>-    st1b            {z1.h}, p0, [x2, #1, mul vl]
>-    st1b            {z2.h}, p0, [x2, #2, mul vl]
>-    add             x2, x2, x5
>-    cbnz            w12, .Loop_gt_eq_32_sve2_addavg_48x64
>-    ret
>-.vl_gt_48_addAvg_48x64:
>-    cmp             x9, #112
>-    bgt             .vl_gt_112_addAvg_48x64
>-    ptrue           p0.b, vl64
>-    ptrue           p1.b, vl32
>-.Loop_gt_48_sve2_addavg_48x64:
>-    sub             w12, w12, #1
>-    ld1b            {z0.b}, p0/z, [x0]
>-    ld1b            {z1.b}, p1/z, [x0, #1, mul vl]
>-    ld1b            {z4.b}, p0/z, [x1]
>-    ld1b            {z5.b}, p1/z, [x1, #1, mul vl]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    add             z0.h, p0/m, z0.h, z4.h
>-    add             z1.h, p1/m, z1.h, z5.h
>-    sqrshrnb        z0.b, z0.h, #7
>-    add             z0.b, z0.b, #0x80
>-    sqrshrnb        z1.b, z1.h, #7
>-    add             z1.b, z1.b, #0x80
>-    st1b            {z0.h}, p0, [x2]
>-    st1b            {z1.h}, p1, [x2, #1, mul vl]
>-    add             x2, x2, x5
>-    cbnz            w12, .Loop_gt_48_sve2_addavg_48x64
>-    ret
>-.vl_gt_112_addAvg_48x64:
>-    mov             x10, #96
>-    mov             x11, #0
>-    whilelt         p0.b, x11, x10
>-.Loop_gt_112_sve2_addavg_48x64:
>-    sub             w12, w12, #1
>-    ld1b            {z0.b}, p0/z, [x0]
>-    ld1b            {z4.b}, p0/z, [x1]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    add             z0.h, p0/m, z0.h, z4.h
>-    sqrshrnb        z0.b, z0.h, #7
>-    add             z0.b, z0.b, #0x80
>-    st1b            {z0.h}, p0, [x2]
>-    add             x2, x2, x5
>-    cbnz            w12, .Loop_gt_112_sve2_addavg_48x64
>-    ret
>-endfunc
>-
>-.macro addAvg_64xN_sve2 h
>-function PFX(addAvg_64x\h\()_sve2)
>-    mov             w12, #\h
>-    rdvl            x9, #1
>-    cmp             x9, #16
>-    bgt             .vl_gt_16_addAvg_64x\h
>-    addAvg_start
>-    sub             x3, x3, #64
>-    sub             x4, x4, #64
>-.Loop_eq_16_sve2_addavg_64x\h\():
>-    sub             w12, w12, #1
>-    ld1             {v0.8h-v3.8h}, [x0], #64
>-    ld1             {v4.8h-v7.8h}, [x1], #64
>-    ld1             {v20.8h-v23.8h}, [x0], x3
>-    ld1             {v24.8h-v27.8h}, [x1], x4
>-    addavg_1        v0, v4
>-    addavg_1        v1, v5
>-    addavg_1        v2, v6
>-    addavg_1        v3, v7
>-    addavg_1        v20, v24
>-    addavg_1        v21, v25
>-    addavg_1        v22, v26
>-    addavg_1        v23, v27
>-    sqxtun          v0.8b, v0.8h
>-    sqxtun2         v0.16b, v1.8h
>-    sqxtun          v1.8b, v2.8h
>-    sqxtun2         v1.16b, v3.8h
>-    sqxtun          v2.8b, v20.8h
>-    sqxtun2         v2.16b, v21.8h
>-    sqxtun          v3.8b, v22.8h
>-    sqxtun2         v3.16b, v23.8h
>-    st1             {v0.16b-v3.16b}, [x2], x5
>-    cbnz            w12, .Loop_eq_16_sve2_addavg_64x\h
>-    ret
>-.vl_gt_16_addAvg_64x\h\():
>-    cmp             x9, #48
>-    bgt             .vl_gt_48_addAvg_64x\h
>-    ptrue           p0.b, vl32
>-.Loop_gt_eq_32_sve2_addavg_64x\h\():
>-    sub             w12, w12, #1
>-    ld1b            {z0.b}, p0/z, [x0]
>-    ld1b            {z1.b}, p0/z, [x0, #1, mul vl]
>-    ld1b            {z2.b}, p0/z, [x0, #2, mul vl]
>-    ld1b            {z3.b}, p0/z, [x0, #3, mul vl]
>-    ld1b            {z4.b}, p0/z, [x1]
>-    ld1b            {z5.b}, p0/z, [x1, #1, mul vl]
>-    ld1b            {z6.b}, p0/z, [x1, #2, mul vl]
>-    ld1b            {z7.b}, p0/z, [x1, #3, mul vl]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    add             z0.h, p0/m, z0.h, z4.h
>-    add             z1.h, p0/m, z1.h, z5.h
>-    add             z2.h, p0/m, z2.h, z6.h
>-    add             z3.h, p0/m, z3.h, z7.h
>-    sqrshrnb        z0.b, z0.h, #7
>-    add             z0.b, z0.b, #0x80
>-    sqrshrnb        z1.b, z1.h, #7
>-    add             z1.b, z1.b, #0x80
>-    sqrshrnb        z2.b, z2.h, #7
>-    add             z2.b, z2.b, #0x80
>-    sqrshrnb        z3.b, z3.h, #7
>-    add             z3.b, z3.b, #0x80
>-    st1b            {z0.h}, p0, [x2]
>-    st1b            {z1.h}, p0, [x2, #1, mul vl]
>-    st1b            {z2.h}, p0, [x2, #2, mul vl]
>-    st1b            {z3.h}, p0, [x2, #3, mul vl]
>-    add             x2, x2, x5
>-    cbnz            w12, .Loop_gt_eq_32_sve2_addavg_64x\h
>-    ret
>-.vl_gt_48_addAvg_64x\h\():
>-    cmp             x9, #112
>-    bgt             .vl_gt_112_addAvg_64x\h
>-    ptrue           p0.b, vl64
>-.Loop_gt_eq_48_sve2_addavg_64x\h\():
>-    sub             w12, w12, #1
>-    ld1b            {z0.b}, p0/z, [x0]
>-    ld1b            {z1.b}, p0/z, [x0, #1, mul vl]
>-    ld1b            {z4.b}, p0/z, [x1]
>-    ld1b            {z5.b}, p0/z, [x1, #1, mul vl]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    add             z0.h, p0/m, z0.h, z4.h
>-    add             z1.h, p0/m, z1.h, z5.h
>-    sqrshrnb        z0.b, z0.h, #7
>-    add             z0.b, z0.b, #0x80
>-    sqrshrnb        z1.b, z1.h, #7
>-    add             z1.b, z1.b, #0x80
>-    st1b            {z0.h}, p0, [x2]
>-    st1b            {z1.h}, p0, [x2, #1, mul vl]
>-    add             x2, x2, x5
>-    cbnz            w12, .Loop_gt_eq_48_sve2_addavg_64x\h
>-    ret
>-.vl_gt_112_addAvg_64x\h\():
>-    ptrue           p0.b, vl128
>-.Loop_gt_eq_128_sve2_addavg_64x\h\():
>-    sub             w12, w12, #1
>-    ld1b            {z0.b}, p0/z, [x0]
>-    ld1b            {z4.b}, p0/z, [x1]
>-    add             x0, x0, x3, lsl #1
>-    add             x1, x1, x4, lsl #1
>-    add             z0.h, p0/m, z0.h, z4.h
>-    sqrshrnb        z0.b, z0.h, #7
>-    add             z0.b, z0.b, #0x80
>-    st1b            {z0.h}, p0, [x2]
>-    add             x2, x2, x5
>-    cbnz            w12, .Loop_gt_eq_128_sve2_addavg_64x\h
>-    ret
>-endfunc
>-.endm
>-
>-addAvg_64xN_sve2 16
>-addAvg_64xN_sve2 32
>-addAvg_64xN_sve2 48
>-addAvg_64xN_sve2 64
>diff --git a/source/common/aarch64/mc-a.S b/source/common/aarch64/mc-a.S
>index 130bf1a4a..876228473 100644
>--- a/source/common/aarch64/mc-a.S
>+++ b/source/common/aarch64/mc-a.S
>@@ -214,344 +214,3 @@ pixel_avg_pp_64xN_neon 16
> pixel_avg_pp_64xN_neon 32
> pixel_avg_pp_64xN_neon 48
> pixel_avg_pp_64xN_neon 64
>-
>-// void addAvg(const int16_t* src0, const int16_t* src1, pixel* dst, intptr_t src0Stride, intptr_t src1Stride, intptr_t dstStride)
>-.macro addAvg_2xN h
>-function PFX(addAvg_2x\h\()_neon)
>-    addAvg_start
>-.rept \h / 2
>-    ldr             w10, [x0]
>-    ldr             w11, [x1]
>-    add             x0, x0, x3
>-    add             x1, x1, x4
>-    ldr             w12, [x0]
>-    ldr             w13, [x1]
>-    add             x0, x0, x3
>-    add             x1, x1, x4
>-    dup             v0.2s, w10
>-    dup             v1.2s, w11
>-    dup             v2.2s, w12
>-    dup             v3.2s, w13
>-    add             v0.4h, v0.4h, v1.4h
>-    add             v2.4h, v2.4h, v3.4h
>-    saddl           v0.4s, v0.4h, v30.4h
>-    saddl           v2.4s, v2.4h, v30.4h
>-    shrn            v0.4h, v0.4s, #7
>-    shrn2           v0.8h, v2.4s, #7
>-    sqxtun          v0.8b, v0.8h
>-    st1             {v0.h}[0], [x2], x5
>-    st1             {v0.h}[2], [x2], x5
>-.endr
>-    ret
>-endfunc
>-.endm
>-
>-addAvg_2xN 4
>-addAvg_2xN 8
>-addAvg_2xN 16
>-
>-.macro addAvg_4xN h
>-function PFX(addAvg_4x\h\()_neon)
>-    addAvg_start
>-.rept \h / 2
>-    ld1             {v0.8b}, [x0], x3
>-    ld1             {v1.8b}, [x1], x4
>-    ld1             {v2.8b}, [x0], x3
>-    ld1             {v3.8b}, [x1], x4
>-    add             v0.4h, v0.4h, v1.4h
>-    add             v2.4h, v2.4h, v3.4h
>-    saddl           v0.4s, v0.4h, v30.4h
>-    saddl           v2.4s, v2.4h, v30.4h
>-    shrn            v0.4h, v0.4s, #7
>-    shrn2           v0.8h, v2.4s, #7
>-    sqxtun          v0.8b, v0.8h
>-    st1             {v0.s}[0], [x2], x5
>-    st1             {v0.s}[1], [x2], x5
>-.endr
>-    ret
>-endfunc
>-.endm
>-
>-addAvg_4xN 2
>-addAvg_4xN 4
>-addAvg_4xN 8
>-addAvg_4xN 16
>-addAvg_4xN 32
>-
>-.macro addAvg_6xN h
>-function PFX(addAvg_6x\h\()_neon)
>-    addAvg_start
>-    mov             w12, #\h / 2
>-    sub             x5, x5, #4
>-.Loop_addavg_6x\h:
>-    sub             w12, w12, #1
>-    ld1             {v0.16b}, [x0], x3
>-    ld1             {v1.16b}, [x1], x4
>-    ld1             {v2.16b}, [x0], x3
>-    ld1             {v3.16b}, [x1], x4
>-    add             v0.8h, v0.8h, v1.8h
>-    add             v2.8h, v2.8h, v3.8h
>-    saddl           v16.4s, v0.4h, v30.4h
>-    saddl2          v17.4s, v0.8h, v30.8h
>-    saddl           v18.4s, v2.4h, v30.4h
>-    saddl2          v19.4s, v2.8h, v30.8h
>-    shrn            v0.4h, v16.4s, #7
>-    shrn2           v0.8h, v17.4s, #7
>-    shrn            v1.4h, v18.4s, #7
>-    shrn2           v1.8h, v19.4s, #7
>-    sqxtun          v0.8b, v0.8h
>-    sqxtun          v1.8b, v1.8h
>-    str             s0, [x2], #4
>-    st1             {v0.h}[2], [x2], x5
>-    str             s1, [x2], #4
>-    st1             {v1.h}[2], [x2], x5
>-    cbnz            w12, .Loop_addavg_6x\h
>-    ret
>-endfunc
>-.endm
>-
>-addAvg_6xN 8
>-addAvg_6xN 16
>-
>-.macro addAvg_8xN h
>-function PFX(addAvg_8x\h\()_neon)
>-    addAvg_start
>-.rept \h / 2
>-    ld1             {v0.16b}, [x0], x3
>-    ld1             {v1.16b}, [x1], x4
>-    ld1             {v2.16b}, [x0], x3
>-    ld1             {v3.16b}, [x1], x4
>-    add             v0.8h, v0.8h, v1.8h
>-    add             v2.8h, v2.8h, v3.8h
>-    saddl           v16.4s, v0.4h, v30.4h
>-    saddl2          v17.4s, v0.8h, v30.8h
>-    saddl           v18.4s, v2.4h, v30.4h
>-    saddl2          v19.4s, v2.8h, v30.8h
>-    shrn            v0.4h, v16.4s, #7
>-    shrn2           v0.8h, v17.4s, #7
>-    shrn            v1.4h, v18.4s, #7
>-    shrn2           v1.8h, v19.4s, #7
>-    sqxtun          v0.8b, v0.8h
>-    sqxtun          v1.8b, v1.8h
>-    st1             {v0.8b}, [x2], x5
>-    st1             {v1.8b}, [x2], x5
>-.endr
>-    ret
>-endfunc
>-.endm
>-
>-.macro addAvg_8xN1 h
>-function PFX(addAvg_8x\h\()_neon)
>-    addAvg_start
>-    mov             w12, #\h / 2
>-.Loop_addavg_8x\h:
>-    sub             w12, w12, #1
>-    ld1             {v0.16b}, [x0], x3
>-    ld1             {v1.16b}, [x1], x4
>-    ld1             {v2.16b}, [x0], x3
>-    ld1             {v3.16b}, [x1], x4
>-    add             v0.8h, v0.8h, v1.8h
>-    add             v2.8h, v2.8h, v3.8h
>-    saddl           v16.4s, v0.4h, v30.4h
>-    saddl2          v17.4s, v0.8h, v30.8h
>-    saddl           v18.4s, v2.4h, v30.4h
>-    saddl2          v19.4s, v2.8h, v30.8h
>-    shrn            v0.4h, v16.4s, #7
>-    shrn2           v0.8h, v17.4s, #7
>-    shrn            v1.4h, v18.4s, #7
>-    shrn2           v1.8h, v19.4s, #7
>-    sqxtun          v0.8b, v0.8h
>-    sqxtun          v1.8b, v1.8h
>-    st1             {v0.8b}, [x2], x5
>-    st1             {v1.8b}, [x2], x5
>-    cbnz            w12, .Loop_addavg_8x\h
>-    ret
>-endfunc
>-.endm
>-
>-addAvg_8xN 2
>-addAvg_8xN 4
>-addAvg_8xN 6
>-addAvg_8xN 8
>-addAvg_8xN 12
>-addAvg_8xN 16
>-addAvg_8xN1 32
>-addAvg_8xN1 64
>-
>-.macro addAvg_12xN h
>-function PFX(addAvg_12x\h\()_neon)
>-    addAvg_start
>-    sub             x3, x3, #16
>-    sub             x4, x4, #16
>-    sub             x5, x5, #8
>-    mov             w12, #\h
>-.Loop_addAvg_12X\h\():
>-    sub             w12, w12, #1
>-    ld1             {v0.16b}, [x0], #16
>-    ld1             {v1.16b}, [x1], #16
>-    ld1             {v2.8b}, [x0], x3
>-    ld1             {v3.8b}, [x1], x4
>-    add             v0.8h, v0.8h, v1.8h
>-    add             v2.4h, v2.4h, v3.4h
>-    saddl           v16.4s, v0.4h, v30.4h
>-    saddl2          v17.4s, v0.8h, v30.8h
>-    saddl           v18.4s, v2.4h, v30.4h
>-    shrn            v0.4h, v16.4s, #7
>-    shrn2           v0.8h, v17.4s, #7
>-    shrn            v1.4h, v18.4s, #7
>-    sqxtun          v0.8b, v0.8h
>-    sqxtun          v1.8b, v1.8h
>-    st1             {v0.8b}, [x2], #8
>-    st1             {v1.s}[0], [x2], x5
>-    cbnz            w12, .Loop_addAvg_12X\h
>-    ret
>-endfunc
>-.endm
>-
>-addAvg_12xN 16
>-addAvg_12xN 32
>-
>-.macro addAvg_16xN h
>-function PFX(addAvg_16x\h\()_neon)
>-    addAvg_start
>-    mov             w12, #\h
>-.Loop_addavg_16x\h:
>-    sub             w12, w12, #1
>-    ld1             {v0.8h-v1.8h}, [x0], x3
>-    ld1             {v2.8h-v3.8h}, [x1], x4
>-    addavg_1        v0, v2
>-    addavg_1        v1, v3
>-    sqxtun          v0.8b, v0.8h
>-    sqxtun2         v0.16b, v1.8h
>-    st1             {v0.16b}, [x2], x5
>-    cbnz            w12, .Loop_addavg_16x\h
>-    ret
>-endfunc
>-.endm
>-
>-addAvg_16xN 4
>-addAvg_16xN 8
>-addAvg_16xN 12
>-addAvg_16xN 16
>-addAvg_16xN 24
>-addAvg_16xN 32
>-addAvg_16xN 64
>-
>-.macro addAvg_24xN h
>-function PFX(addAvg_24x\h\()_neon)
>-    addAvg_start
>-    mov             w12, #\h
>-.Loop_addavg_24x\h\():
>-    sub             w12, w12, #1
>-    ld1             {v0.16b-v2.16b}, [x0], x3
>-    ld1             {v3.16b-v5.16b}, [x1], x4
>-    addavg_1        v0, v3
>-    addavg_1        v1, v4
>-    addavg_1        v2, v5
>-    sqxtun          v0.8b, v0.8h
>-    sqxtun          v1.8b, v1.8h
>-    sqxtun          v2.8b, v2.8h
>-    st1             {v0.8b-v2.8b}, [x2], x5
>-    cbnz            w12, .Loop_addavg_24x\h
>-    ret
>-endfunc
>-.endm
>-
>-addAvg_24xN 32
>-addAvg_24xN 64
>-
>-.macro addAvg_32xN h
>-function PFX(addAvg_32x\h\()_neon)
>-    addAvg_start
>-    mov             w12, #\h
>-.Loop_addavg_32x\h\():
>-    sub             w12, w12, #1
>-    ld1             {v0.8h-v3.8h}, [x0], x3
>-    ld1             {v4.8h-v7.8h}, [x1], x4
>-    addavg_1        v0, v4
>-    addavg_1        v1, v5
>-    addavg_1        v2, v6
>-    addavg_1        v3, v7
>-    sqxtun          v0.8b, v0.8h
>-    sqxtun          v1.8b, v1.8h
>-    sqxtun          v2.8b, v2.8h
>-    sqxtun          v3.8b, v3.8h
>-    st1             {v0.8b-v3.8b}, [x2], x5
>-    cbnz            w12, .Loop_addavg_32x\h
>-    ret
>-endfunc
>-.endm
>-
>-addAvg_32xN 8
>-addAvg_32xN 16
>-addAvg_32xN 24
>-addAvg_32xN 32
>-addAvg_32xN 48
>-addAvg_32xN 64
>-
>-function PFX(addAvg_48x64_neon)
>-    addAvg_start
>-    sub             x3, x3, #64
>-    sub             x4, x4, #64
>-    mov             w12, #64
>-.Loop_addavg_48x64:
>-    sub             w12, w12, #1
>-    ld1             {v0.8h-v3.8h}, [x0], #64
>-    ld1             {v4.8h-v7.8h}, [x1], #64
>-    ld1             {v20.8h-v21.8h}, [x0], x3
>-    ld1             {v22.8h-v23.8h}, [x1], x4
>-    addavg_1        v0, v4
>-    addavg_1        v1, v5
>-    addavg_1        v2, v6
>-    addavg_1        v3, v7
>-    addavg_1        v20, v22
>-    addavg_1        v21, v23
>-    sqxtun          v0.8b, v0.8h
>-    sqxtun2         v0.16b, v1.8h
>-    sqxtun          v1.8b, v2.8h
>-    sqxtun2         v1.16b, v3.8h
>-    sqxtun          v2.8b, v20.8h
>-    sqxtun2         v2.16b, v21.8h
>-    st1             {v0.16b-v2.16b}, [x2], x5
>-    cbnz            w12, .Loop_addavg_48x64
>-    ret
>-endfunc
>-
>-.macro addAvg_64xN h
>-function PFX(addAvg_64x\h\()_neon)
>-    addAvg_start
>-    mov             w12, #\h
>-    sub             x3, x3, #64
>-    sub             x4, x4, #64
>-.Loop_addavg_64x\h\():
>-    sub             w12, w12, #1
>-    ld1             {v0.8h-v3.8h}, [x0], #64
>-    ld1             {v4.8h-v7.8h}, [x1], #64
>-    ld1             {v20.8h-v23.8h}, [x0], x3
>-    ld1             {v24.8h-v27.8h}, [x1], x4
>-    addavg_1        v0, v4
>-    addavg_1        v1, v5
>-    addavg_1        v2, v6
>-    addavg_1        v3, v7
>-    addavg_1        v20, v24
>-    addavg_1        v21, v25
>-    addavg_1        v22, v26
>-    addavg_1        v23, v27
>-    sqxtun          v0.8b, v0.8h
>-    sqxtun2         v0.16b, v1.8h
>-    sqxtun          v1.8b, v2.8h
>-    sqxtun2         v1.16b, v3.8h
>-    sqxtun          v2.8b, v20.8h
>-    sqxtun2         v2.16b, v21.8h
>-    sqxtun          v3.8b, v22.8h
>-    sqxtun2         v3.16b, v23.8h
>-    st1             {v0.16b-v3.16b}, [x2], x5
>-    cbnz            w12, .Loop_addavg_64x\h
>-    ret
>-endfunc
>-.endm
>-
>-addAvg_64xN 16
>-addAvg_64xN 32
>-addAvg_64xN 48
>-addAvg_64xN 64
>diff --git a/source/common/aarch64/mem-neon.h b/source/common/aarch64/mem-neon.h
>index 8bd5fbee9..27d72b70c 100644
>--- a/source/common/aarch64/mem-neon.h
>+++ b/source/common/aarch64/mem-neon.h
>@@ -74,6 +74,26 @@ static void inline store_u8x4x1(uint8_t *d, const uint8x8_t s)
>     vst1_lane_u32((uint32_t *)d, vreinterpret_u32_u8(s), 0);
> }
> 
>+// Store 2 bytes from the low half of a uint8x8_t.
>+static void inline store_u8x2x1(uint8_t *d, const uint8x8_t s)
>+{
>+    vst1_lane_u16((uint16_t *)d, vreinterpret_u16_u8(s), 0);
>+}
>+
>+// Load 2 int16_t into a int16x8_t.
>+static inline int16x8_t load_s16x2x1(const int16_t *p)
>+{
>+    int32x4_t ret = vld1q_lane_s32((const int32_t *)p, vdupq_n_s32(0), 0);
>+
>+    return vreinterpretq_s16_s32(ret);
>+}
>+
>+// Store 2 uint16_t from the low half of a uint16x8_t.
>+static inline void store_u16x2x1(const uint16_t *d, const uint16x8_t s)
>+{
>+    vst1q_lane_u32((uint32_t *)d, vreinterpretq_u32_u16(s), 0);
>+}
>+
> // Store N blocks of 32-bits from (N / 2) D-Registers.
> template<int N>
> static void inline store_u8x4_strided_xN(uint8_t *d, intptr_t stride,
>diff --git a/source/common/aarch64/pixel-prim.cpp b/source/common/aarch64/pixel-prim.cpp
>index d565c7f0b..50df51ca1 100644
>--- a/source/common/aarch64/pixel-prim.cpp
>+++ b/source/common/aarch64/pixel-prim.cpp
>@@ -1196,49 +1196,138 @@ void pixel_add_ps_neon(pixel *dst, intptr_t dstride, const pixel *src0,
>     }
> }
> 
>-template<int bx, int by>
>-void addAvg_neon(const int16_t *src0, const int16_t *src1, pixel *dst, intptr_t src0Stride, intptr_t src1Stride,
>-                 intptr_t dstStride)
>+template<int width, int height>
>+void addAvg_neon(const int16_t *src0, const int16_t *src1, pixel *dst,
>+                 intptr_t src0Stride, intptr_t src1Stride, intptr_t dstStride)
> {
>-
>     const int shiftNum = IF_INTERNAL_PREC + 1 - X265_DEPTH;
>-    const int offset = (1 << (shiftNum - 1)) + 2 * IF_INTERNAL_OFFS;
>+    const int offset = 2 * IF_INTERNAL_OFFS;
> 
>-    const int32x4_t addon = vdupq_n_s32(offset);
>-    for (int y = 0; y < by; y++)
>+#if HIGH_BIT_DEPTH
>+    const int16x8_t addon = vdupq_n_s16(offset >> shiftNum);
>+
>+    for (int h = 0; h < height; h++)
>     {
>-        int x = 0;
>+        int w = 0;
>+        for (; w + 16 <= width; w += 16)
>+        {
>+            int16x8_t s0[2], s1[2];
>+            load_s16x8xn<2>(src0 + w, 8, s0);
>+            load_s16x8xn<2>(src1 + w, 8, s1);
> 
>-        for (; (x + 8) <= bx; x += 8)
>+            int16x8_t d0_lo = vrsraq_n_s16(addon, vaddq_s16(s0[0], s1[0]), shiftNum);
>+            int16x8_t d0_hi = vrsraq_n_s16(addon, vaddq_s16(s0[1], s1[1]), shiftNum);
>+
>+            d0_lo = vminq_s16(d0_lo, vdupq_n_s16((1 << X265_DEPTH) - 1));
>+            d0_lo = vmaxq_s16(d0_lo, vdupq_n_s16(0));
>+            d0_hi = vminq_s16(d0_hi, vdupq_n_s16((1 << X265_DEPTH) - 1));
>+            d0_hi = vmaxq_s16(d0_hi, vdupq_n_s16(0));
>+
>+            vst1q_u16(dst + w, vreinterpretq_u16_s16(d0_lo));
>+            vst1q_u16(dst + w + 8, vreinterpretq_u16_s16(d0_hi));
>+        }
>+        if (width & 8)
>         {
>-            int16x8_t in0 = vld1q_s16(src0 + x);
>-            int16x8_t in1 = vld1q_s16(src1 + x);
>-            int32x4_t t1 = vaddl_s16(vget_low_s16(in0), vget_low_s16(in1));
>-            int32x4_t t2 = vaddl_high_s16(in0, in1);
>-            t1 = vaddq_s32(t1, addon);
>-            t2 = vaddq_s32(t2, addon);
>-            t1 = vshrq_n_s32(t1, shiftNum);
>-            t2 = vshrq_n_s32(t2, shiftNum);
>-            int16x8_t t = vuzp1q_s16(vreinterpretq_s16_s32(t1),
>-                                     vreinterpretq_s16_s32(t2));
>-#if HIGH_BIT_DEPTH
>-            t = vminq_s16(t, vdupq_n_s16((1 << X265_DEPTH) - 1));
>-            t = vmaxq_s16(t, vdupq_n_s16(0));
>-            vst1q_u16(dst + x, vreinterpretq_u16_s16(t));
>-#else
>-            vst1_u8(dst + x, vqmovun_s16(t));
>-#endif
>+            int16x8_t s0 = vld1q_s16(src0 + w);
>+            int16x8_t s1 = vld1q_s16(src1 + w);
>+
>+            int16x8_t d0 = vrsraq_n_s16(addon, vaddq_s16(s0, s1), shiftNum);
>+            d0 = vminq_s16(d0, vdupq_n_s16((1 << X265_DEPTH) - 1));
>+            d0 = vmaxq_s16(d0, vdupq_n_s16(0));
>+
>+            vst1q_u16(dst + w, vreinterpretq_u16_s16(d0));
>+
>+            w += 8;
>+        }
>+        if (width & 4)
>+        {
>+            int16x4_t s0 = vld1_s16(src0 + w);
>+            int16x4_t s1 = vld1_s16(src1 + w);
>+
>+            int16x4_t d0 = vrsra_n_s16(vget_low_s16(addon), vadd_s16(s0, s1), shiftNum);
>+            d0 = vmin_s16(d0, vdup_n_s16((1 << X265_DEPTH) - 1));
>+            d0 = vmax_s16(d0, vdup_n_s16(0));
>+
>+            vst1_u16(dst + w, vreinterpret_u16_s16(d0));
>+
>+            w += 4;
>+        }
>+        if (width & 2)
>+        {
>+            int16x8_t s0 = load_s16x2x1(src0 + w);
>+            int16x8_t s1 = load_s16x2x1(src1 + w);
>+
>+            int16x8_t d0 = vrsraq_n_s16(addon, vaddq_s16(s0, s1), shiftNum);
>+            d0 = vminq_s16(d0, vdupq_n_s16((1 << X265_DEPTH) - 1));
>+            d0 = vmaxq_s16(d0, vdupq_n_s16(0));
>+
>+            store_u16x2x1(dst + w, vreinterpretq_u16_s16(d0));
>+        }
>+
>+        src0 += src0Stride;
>+        src1 += src1Stride;
>+        dst  += dstStride;
>+    }
>+#else // !HIGH_BIT_DEPTH
>+    const uint8x8_t addon = vdup_n_u8(offset >> shiftNum);
>+
>+    for (int h = 0; h < height; h++)
>+    {
>+        int w = 0;
>+        for (; w + 16 <= width; w += 16)
>+        {
>+            int16x8_t s0[2], s1[2];
>+            load_s16x8xn<2>(src0 + w, 8, s0);
>+            load_s16x8xn<2>(src1 + w, 8, s1);
>+
>+            int8x8_t sum01_s8_lo = vqrshrn_n_s16(vaddq_s16(s0[0], s1[0]), shiftNum);
>+            int8x8_t sum01_s8_hi = vqrshrn_n_s16(vaddq_s16(s0[1], s1[1]), shiftNum);
>+            uint8x8_t d0_lo = vadd_u8(vreinterpret_u8_s8(sum01_s8_lo), addon);
>+            uint8x8_t d0_hi = vadd_u8(vreinterpret_u8_s8(sum01_s8_hi), addon);
>+
>+            vst1_u8(dst + w, d0_lo);
>+            vst1_u8(dst + w + 8, d0_hi);
>+        }
>+        if (width & 8)
>+        {
>+            int16x8_t s0 = vld1q_s16(src0 + w);
>+            int16x8_t s1 = vld1q_s16(src1 + w);
>+
>+            int8x8_t sum01_s8 = vqrshrn_n_s16(vaddq_s16(s0, s1), shiftNum);
>+            uint8x8_t d0 = vadd_u8(vreinterpret_u8_s8(sum01_s8), addon);
>+
>+            vst1_u8(dst + w, d0);
>+
>+            w += 8;
>         }
>-        for (; x < bx; x += 2)
>+        if (width & 4)
>         {
>-            dst[x + 0] = x265_clip((src0[x + 0] + src1[x + 0] + offset) >> shiftNum);
>-            dst[x + 1] = x265_clip((src0[x + 1] + src1[x + 1] + offset) >> shiftNum);
>+            int16x8_t s0 = vcombine_s16(vld1_s16(src0 + w), vdup_n_s16(0));
>+            int16x8_t s1 = vcombine_s16(vld1_s16(src1 + w), vdup_n_s16(0));
>+
>+            int8x8_t sum01_s8 = vqrshrn_n_s16(vaddq_s16(s0, s1), shiftNum);
>+            uint8x8_t d0 = vadd_u8(vreinterpret_u8_s8(sum01_s8), addon);
>+
>+            store_u8x4x1(dst + w, d0);
>+
>+            w += 4;
>+        }
>+        if (width & 2)
>+        {
>+            int16x8_t s0 = load_s16x2x1(src0 + w);
>+            int16x8_t s1 = load_s16x2x1(src1 + w);
>+
>+            int8x8_t sum01_s8 = vqrshrn_n_s16(vaddq_s16(s0, s1), shiftNum);
>+            uint8x8_t d0 = vadd_u8(vreinterpret_u8_s8(sum01_s8), addon);
>+
>+            store_u8x2x1(dst + w, d0);
>         }
> 
>         src0 += src0Stride;
>         src1 += src1Stride;
>         dst  += dstStride;
>     }
>+#endif
> }
> 
> void planecopy_cp_neon(const uint8_t *src, intptr_t srcStride, pixel *dst,
>@@ -2249,29 +2338,30 @@ void setupPixelPrimitives_neon(EncoderPrimitives &p)
>     p.chroma[X265_CSP_I420].pu[CHROMA_420_ ## W ## x ## H].copy_pp = blockcopy_pp_neon<W, H>; \
> 
> 
>-    CHROMA_PU_420(4, 4);
>-    CHROMA_PU_420(8, 8);
>-    CHROMA_PU_420(16, 16);
>-    CHROMA_PU_420(32, 32);
>+    CHROMA_PU_420(2, 4);
>+    CHROMA_PU_420(2, 8);
>     CHROMA_PU_420(4, 2);
>-    CHROMA_PU_420(8, 4);
>+    CHROMA_PU_420(4, 4);
>     CHROMA_PU_420(4, 8);
>-    CHROMA_PU_420(8, 6);
>     CHROMA_PU_420(6, 8);
>+    CHROMA_PU_420(4, 16);
>     CHROMA_PU_420(8, 2);
>-    CHROMA_PU_420(2, 8);
>-    CHROMA_PU_420(16, 8);
>-    CHROMA_PU_420(8,  16);
>-    CHROMA_PU_420(16, 12);
>+    CHROMA_PU_420(8, 4);
>+    CHROMA_PU_420(8, 6);
>+    CHROMA_PU_420(8, 8);
>+    CHROMA_PU_420(8, 16);
>+    CHROMA_PU_420(8, 32);
>     CHROMA_PU_420(12, 16);
>     CHROMA_PU_420(16, 4);
>-    CHROMA_PU_420(4,  16);
>-    CHROMA_PU_420(32, 16);
>+    CHROMA_PU_420(16, 8);
>+    CHROMA_PU_420(16, 12);
>+    CHROMA_PU_420(16, 16);
>     CHROMA_PU_420(16, 32);
>-    CHROMA_PU_420(32, 24);
>     CHROMA_PU_420(24, 32);
>     CHROMA_PU_420(32, 8);
>-    CHROMA_PU_420(8,  32);
>+    CHROMA_PU_420(32, 16);
>+    CHROMA_PU_420(32, 24);
>+    CHROMA_PU_420(32, 32);
> 
> 
> 
>@@ -2353,30 +2443,31 @@ void setupPixelPrimitives_neon(EncoderPrimitives &p)
>     p.chroma[X265_CSP_I422].pu[CHROMA_422_ ## W ## x ## H].copy_pp = blockcopy_pp_neon<W, H>; \
> 
> 
>-    CHROMA_PU_422(4, 8);
>-    CHROMA_PU_422(8, 16);
>-    CHROMA_PU_422(16, 32);
>-    CHROMA_PU_422(32, 64);
>-    CHROMA_PU_422(4, 4);
>+    CHROMA_PU_422(2, 4);
>     CHROMA_PU_422(2, 8);
>-    CHROMA_PU_422(8, 8);
>+    CHROMA_PU_422(2, 16);
>+    CHROMA_PU_422(4, 4);
>+    CHROMA_PU_422(4, 8);
>     CHROMA_PU_422(4, 16);
>-    CHROMA_PU_422(8, 12);
>-    CHROMA_PU_422(6, 16);
>+    CHROMA_PU_422(4, 32);
>     CHROMA_PU_422(8, 4);
>-    CHROMA_PU_422(2, 16);
>-    CHROMA_PU_422(16, 16);
>+    CHROMA_PU_422(8, 8);
>+    CHROMA_PU_422(8, 12);
>+    CHROMA_PU_422(8, 16);
>     CHROMA_PU_422(8, 32);
>-    CHROMA_PU_422(16, 24);
>+    CHROMA_PU_422(8, 64);
>+    CHROMA_PU_422(6, 16);
>     CHROMA_PU_422(12, 32);
>     CHROMA_PU_422(16, 8);
>-    CHROMA_PU_422(4,  32);
>-    CHROMA_PU_422(32, 32);
>+    CHROMA_PU_422(16, 16);
>+    CHROMA_PU_422(16, 24);
>+    CHROMA_PU_422(16, 32);
>     CHROMA_PU_422(16, 64);
>-    CHROMA_PU_422(32, 48);
>     CHROMA_PU_422(24, 64);
>     CHROMA_PU_422(32, 16);
>-    CHROMA_PU_422(8,  64);
>+    CHROMA_PU_422(32, 32);
>+    CHROMA_PU_422(32, 48);
>+    CHROMA_PU_422(32, 64);
> 
> 
>     p.chroma[X265_CSP_I422].pu[CHROMA_422_2x4].satd   = NULL;
>-- 
>2.39.5 (Apple Git-154)
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.videolan.org/pipermail/x265-devel/attachments/20250626/ca20c2f1/attachment-0001.htm>


More information about the x265-devel mailing list