[x265] [PATCH] asm: pixelsub_ps routine for all block sizes

murugan at multicorewareinc.com murugan at multicorewareinc.com
Mon Nov 11 22:47:02 CET 2013


# HG changeset patch
# User Murugan Vairavel <murugan at multicorewareinc.com>
# Date 1384206405 -19800
#      Tue Nov 12 03:16:45 2013 +0530
# Node ID 85749f42e3fa7c03ce29903d1457dc0ed873c120
# Parent  9642b5b6500b5553ab3ce70a360aaaadad5d7234
asm: pixelsub_ps routine for all block sizes

diff -r 9642b5b6500b -r 85749f42e3fa source/common/x86/pixel-a.asm
--- a/source/common/x86/pixel-a.asm	Mon Nov 11 17:41:32 2013 +0800
+++ b/source/common/x86/pixel-a.asm	Tue Nov 12 03:16:45 2013 +0530
@@ -5284,3 +5284,1060 @@
     jl .loop
     movifnidn eax, r0d
     RET
+
+;-----------------------------------------------------------------------------
+; void pixel_sub_ps_c_2x4(int16_t *dest, intptr_t destride, pixel *src0, pixel *src1, intptr_t srcstride0, intptr_t srcstride1);
+;-----------------------------------------------------------------------------
+INIT_XMM sse4
+cglobal pixel_sub_ps_2x4, 6, 6, 4, dest, deststride, src0, src1, srcstride0, srcstride1
+
+add         r1,    r1
+
+movd        m0,    [r2]
+pinsrw      m0,    [r2 + r4],    2
+movd        m1,    [r2 + 2 * r4]
+
+movd        m2,    [r3]
+pinsrw      m2,    [r3 + r5],    2
+movd        m3,    [r3 + 2 * r5]
+
+lea         r2,    [r2 + 2 * r4]
+lea         r3,    [r3 + 2 * r5]
+
+pinsrw      m1,    [r2 + r4],    2
+pinsrw      m3,    [r3 + r5],    2
+
+pmovzxbw    m0,    m0
+pmovzxbw    m1,    m1
+pmovzxbw    m2,    m2
+pmovzxbw    m3,    m3
+
+psubw       m0,    m2
+psubw       m1,    m3
+
+movd      [r0],            m0
+pextrd    [r0 + r1],       m0,    2
+movd      [r0 + 2* r1],    m1
+
+lea     r0,              [r0 + 2 * r1]
+
+pextrd    [r0 + r1],       m1,    2
+
+RET
+
+;-----------------------------------------------------------------------------
+; void pixel_sub_ps_c_%1x%2(int16_t *dest, intptr_t destride, pixel *src0, pixel *src1, intptr_t srcstride0, intptr_t srcstride1);
+;-----------------------------------------------------------------------------
+%macro PIXELSUB_PS_W2_H4 2
+INIT_XMM sse4
+cglobal pixel_sub_ps_%1x%2, 6, 7, 4, dest, deststride, src0, src1, srcstride0, srcstride1
+
+add    r1,     r1
+mov    r6d,    %2/4
+
+.loop
+
+    movd        m0,    [r2]
+    pinsrw      m0,    [r2 + r4],    2
+    movd        m1,    [r2 + 2 * r4]
+
+    movd        m2,    [r3]
+    pinsrw      m2,    [r3 + r5],    2
+    movd        m3,    [r3 + 2 * r5]
+
+    lea         r2,    [r2 + 2 * r4]
+    lea         r3,    [r3 + 2 * r5]
+
+    pinsrw      m1,    [r2 + r4],    2
+    pinsrw      m3,    [r3 + r5],    2
+
+    pmovzxbw    m0,    m0
+    pmovzxbw    m1,    m1
+    pmovzxbw    m2,    m2
+    pmovzxbw    m3,    m3
+
+    psubw       m0,    m2
+    psubw       m1,    m3
+
+    movd      [r0],            m0
+    pextrd    [r0 + r1],       m0,    2
+    movd      [r0 + 2* r1],    m1
+
+    lea     r0,              [r0 + 2 * r1]
+
+    pextrd    [r0 + r1],       m1,    2
+
+    lea     r2,              [r2 + 2 * r4]
+    lea     r3,              [r3 + 2 * r5]
+    lea     r0,              [r0 + 2 * r1]
+
+    dec     r6d
+
+jnz    .loop
+
+RET
+%endmacro
+
+PIXELSUB_PS_W2_H4 2, 8
+
+;-----------------------------------------------------------------------------
+; void pixel_sub_sp_c_4x2(int16_t *dest, intptr_t destride, pixel *src0, pixel *src1, intptr_t srcstride0, intptr_t srcstride1);
+;-----------------------------------------------------------------------------
+INIT_XMM sse4
+cglobal pixel_sub_ps_4x2, 6, 6, 4, dest, deststride, src0, src1, srcstride0, srcstride1
+
+add          r1,    r1
+
+movd         m0,    [r2]
+movd         m1,    [r3]
+
+movd         m2,    [r2 + r4]
+movd         m3,    [r3 + r5]
+
+punpckldq    m0,    m2
+punpckldq    m1,    m3
+pmovzxbw     m0,    m0
+pmovzxbw     m1,    m1
+
+psubw        m0,    m1
+
+movlps    [r0],         m0
+movhps    [r0 + r1],    m0
+
+RET
+
+;-----------------------------------------------------------------------------
+; void pixel_sub_ps_c_4x4(int16_t *dest, intptr_t destride, pixel *src0, pixel *src1, intptr_t srcstride0, intptr_t srcstride1);
+;-----------------------------------------------------------------------------
+INIT_XMM sse4
+cglobal pixel_sub_ps_4x4, 6, 6, 8, dest, deststride, src0, src1, srcstride0, srcstride1
+
+add          r1,    r1
+
+movd         m0,    [r2]
+movd         m1,    [r3]
+
+movd         m2,    [r2 + r4]
+movd         m3,    [r3 + r5]
+
+movd         m4,    [r2 + 2 * r4]
+movd         m5,    [r3 + 2 * r5]
+
+lea          r2,    [r2 + 2 * r4]
+lea          r3,    [r3 + 2 * r5]
+
+movd         m6,    [r2 + r4]
+movd         m7,    [r3 + r5]
+
+punpckldq    m0,    m2
+punpckldq    m1,    m3
+punpckldq    m4,    m6
+punpckldq    m5,    m7
+
+pmovzxbw     m0,    m0
+pmovzxbw     m1,    m1
+pmovzxbw     m4,    m4
+pmovzxbw     m5,    m5
+
+psubw        m0,    m1
+psubw        m4,    m5
+
+movlps    [r0],             m0
+movhps    [r0 + r1],        m0
+movlps    [r0 + 2 * r1],    m4
+
+lea       r0,               [r0 + 2 * r1]
+
+movhps    [r0 + r1],        m4
+
+RET
+
+;-----------------------------------------------------------------------------
+; void pixel_sub_ps_c_%1x%2(int16_t *dest, intptr_t destride, pixel *src0, pixel *src1, intptr_t srcstride0, intptr_t srcstride1);
+;-----------------------------------------------------------------------------
+%macro PIXELSUB_PS_W4_H4 2
+INIT_XMM sse4
+cglobal pixel_sub_ps_%1x%2, 6, 7, 8, dest, deststride, src0, src1, srcstride0, srcstride1
+
+add    r1,     r1
+mov    r6d,    %2/4
+
+.loop
+
+    movd         m0,    [r2]
+    movd         m1,    [r3]
+
+    movd         m2,    [r2 + r4]
+    movd         m3,    [r3 + r5]
+
+    movd         m4,    [r2 + 2 * r4]
+    movd         m5,    [r3 + 2 * r5]
+
+    lea          r2,    [r2 + 2 * r4]
+    lea          r3,    [r3 + 2 * r5]
+
+    movd         m6,    [r2 + r4]
+    movd         m7,    [r3 + r5]
+
+    punpckldq    m0,    m2
+    punpckldq    m1,    m3
+    punpckldq    m4,    m6
+    punpckldq    m5,    m7
+
+    pmovzxbw     m0,    m0
+    pmovzxbw     m1,    m1
+    pmovzxbw     m4,    m4
+    pmovzxbw     m5,    m5
+
+    psubw        m0,    m1
+    psubw        m4,    m5
+
+    movlps    [r0],             m0
+    movhps    [r0 + r1],        m0
+    movlps    [r0 + 2 * r1],    m4
+
+    lea       r0,               [r0 + 2 * r1]
+
+    movhps    [r0 + r1],        m4
+
+    lea       r2,               [r2 + 2 * r4]
+    lea       r3,               [r3 + 2 * r5]
+    lea       r0,               [r0 + 2 * r1]
+
+    dec       r6d
+
+jnz    .loop
+
+RET
+%endmacro
+
+PIXELSUB_PS_W4_H4 4, 8
+PIXELSUB_PS_W4_H4 4, 16
+
+;-----------------------------------------------------------------------------
+; void pixel_sub_ps_c_%1x%2(int16_t *dest, intptr_t destride, pixel *src0, pixel *src1, intptr_t srcstride0, intptr_t srcstride1);
+;-----------------------------------------------------------------------------
+%macro PIXELSUB_PS_W6_H4 2
+INIT_XMM sse4
+cglobal pixel_sub_ps_%1x%2, 6, 7, 8, dest, deststride, src0, src1, srcstride0, srcstride1
+
+add    r1,     r1
+mov    r6d,    %2/4
+
+.loop
+
+    movh        m0,    [r2]
+    movh        m1,    [r3]
+
+    movh        m2,    [r2 + r4]
+    movh        m3,    [r3 + r5]
+
+    movh        m4,    [r2 + 2 * r4]
+    movh        m5,    [r3 + 2 * r5]
+
+    lea         r2,    [r2 + 2 * r4]
+    lea         r3,    [r3 + 2 * r5]
+
+    movh        m6,    [r2 + r4]
+    movh        m7,    [r3 + r5]
+
+    pmovzxbw    m0,    m0
+    pmovzxbw    m1,    m1
+    pmovzxbw    m2,    m2
+    pmovzxbw    m3,    m3
+    pmovzxbw    m4,    m4
+    pmovzxbw    m5,    m5
+    pmovzxbw    m6,    m6
+    pmovzxbw    m7,    m7
+
+    psubw       m0,    m1
+    psubw       m2,    m3
+    psubw       m4,    m5
+    psubw       m6,    m7
+
+    movh      [r0],                 m0
+    pextrd    [r0 + 8],             m0,    2
+    movh      [r0 + r1],            m2
+    pextrd    [r0 + r1 + 8],        m2,    2
+    movh      [r0 + 2* r1],         m4
+    pextrd    [r0 + 2 * r1 + 8],    m4,    2
+
+    lea       r0,                   [r0 + 2 * r1]
+
+    movh      [r0 + r1],            m6
+    pextrd    [r0 + r1 + 8],        m6,    2
+
+    lea       r2,                   [r2 + 2 * r4]
+    lea       r3,                   [r3 + 2 * r5]
+    lea       r0,                   [r0 + 2 * r1]
+
+    dec     r6d
+
+jnz    .loop
+
+RET
+%endmacro
+
+PIXELSUB_PS_W6_H4 6, 8
+
+;-----------------------------------------------------------------------------
+; void pixel_sub_ps_c_8x2(int16_t *dest, intptr_t destride, pixel *src0, pixel *src1, intptr_t srcstride0, intptr_t srcstride1);
+;-----------------------------------------------------------------------------
+INIT_XMM sse4
+cglobal pixel_sub_ps_8x2, 6, 6, 4, dest, deststride, src0, src1, srcstride0, srcstride1
+
+add         r1,    r1
+
+movh        m0,    [r2]
+movh        m1,    [r3]
+pmovzxbw    m0,    m0
+pmovzxbw    m1,    m1
+
+movh        m2,    [r2 + r4]
+movh        m3,    [r3 + r5]
+pmovzxbw    m2,    m2
+pmovzxbw    m3,    m3
+
+psubw       m0,    m1
+psubw       m2,    m3
+
+movu    [r0],         m0
+movu    [r0 + r1],    m2
+
+RET
+
+;-----------------------------------------------------------------------------
+; void pixel_sub_ps_c_8x4(int16_t *dest, intptr_t destride, pixel *src0, pixel *src1, intptr_t srcstride0, intptr_t srcstride1);
+;-----------------------------------------------------------------------------
+INIT_XMM sse4
+cglobal pixel_sub_ps_8x4, 6, 6, 8, dest, deststride, src0, src1, srcstride0, srcstride1
+
+add         r1,    r1
+
+movh        m0,    [r2]
+movh        m1,    [r3]
+pmovzxbw    m0,    m0
+pmovzxbw    m1,    m1
+
+movh        m2,    [r2 + r4]
+movh        m3,    [r3 + r5]
+pmovzxbw    m2,    m2
+pmovzxbw    m3,    m3
+
+movh        m4,    [r2 + 2 * r4]
+movh        m5,    [r3 + 2 * r5]
+pmovzxbw    m4,    m4
+pmovzxbw    m5,    m5
+
+psubw       m0,    m1
+psubw       m2,    m3
+psubw       m4,    m5
+
+lea         r2,    [r2 + 2 * r4]
+lea         r3,    [r3 + 2 * r5]
+
+movh        m6,    [r2 + r4]
+movh        m7,    [r3 + r5]
+pmovzxbw    m6,    m6
+pmovzxbw    m7,    m7
+
+psubw       m6,    m7
+
+movu    [r0],             m0
+movu    [r0 + r1],        m2
+movu    [r0 + 2 * r1],    m4
+
+lea     r0,               [r0 + 2 * r1]
+
+movu    [r0 + r1],        m6
+
+RET
+
+;-----------------------------------------------------------------------------
+; void pixel_sub_ps_c_8x6(int16_t *dest, intptr_t destride, pixel *src0, pixel *src1, intptr_t srcstride0, intptr_t srcstride1);
+;-----------------------------------------------------------------------------
+INIT_XMM sse4
+cglobal pixel_sub_ps_8x6, 6, 6, 8, dest, deststride, src0, src1, srcstride0, srcstride1
+
+add         r1,    r1
+
+movh        m0,    [r2]
+movh        m1,    [r3]
+pmovzxbw    m0,    m0
+pmovzxbw    m1,    m1
+
+movh        m2,    [r2 + r4]
+movh        m3,    [r3 + r5]
+pmovzxbw    m2,    m2
+pmovzxbw    m3,    m3
+
+movh        m4,    [r2 + 2 * r4]
+movh        m5,    [r3 + 2 * r5]
+pmovzxbw    m4,    m4
+pmovzxbw    m5,    m5
+
+psubw       m0,    m1
+psubw       m2,    m3
+psubw       m4,    m5
+
+lea         r2,    [r2 + 2 * r4]
+lea         r3,    [r3 + 2 * r5]
+
+movh        m6,    [r2 + r4]
+movh        m7,    [r3 + r5]
+pmovzxbw    m6,    m6
+pmovzxbw    m7,    m7
+
+movh        m1,    [r2 + 2 * r4]
+movh        m3,    [r3 + 2 * r5]
+pmovzxbw    m1,    m1
+pmovzxbw    m3,    m3
+
+psubw       m6,    m7
+psubw       m1,    m3
+
+lea         r2,    [r2 + 2 * r4]
+lea         r3,    [r3 + 2 * r5]
+
+movh        m3,    [r2 + r4]
+movh        m5,    [r3 + r5]
+pmovzxbw    m3,    m3
+pmovzxbw    m5,    m5
+
+psubw       m3,     m5
+
+movu    [r0],             m0
+movu    [r0 + r1],        m2
+movu    [r0 + 2 * r1],    m4
+
+lea     r0,               [r0 + 2 * r1]
+
+movu    [r0 + r1],        m6
+movu    [r0 + 2 * r1],    m1
+
+lea     r0,               [r0 + 2 * r1]
+
+movu    [r0 + r1],        m3
+
+RET
+
+;-----------------------------------------------------------------------------
+; void pixel_sub_ps_c_%1x%2(int16_t *dest, intptr_t destride, pixel *src0, pixel *src1, intptr_t srcstride0, intptr_t srcstride1);
+;-----------------------------------------------------------------------------
+%macro PIXELSUB_PS_W8_H4 2
+INIT_XMM sse4
+cglobal pixel_sub_ps_%1x%2, 6, 7, 8, dest, deststride, src0, src1, srcstride0, srcstride1
+
+add    r1,     r1
+mov    r6d,    %2/4
+
+.loop
+
+    movh        m0,    [r2]
+    movh        m1,    [r3]
+    pmovzxbw    m0,    m0
+    pmovzxbw    m1,    m1
+
+    movh        m2,    [r2 + r4]
+    movh        m3,    [r3 + r5]
+    pmovzxbw    m2,    m2
+    pmovzxbw    m3,    m3
+
+    movh        m4,    [r2 + 2 * r4]
+    movh        m5,    [r3 + 2 * r5]
+    pmovzxbw    m4,    m4
+    pmovzxbw    m5,    m5
+
+    psubw       m0,    m1
+    psubw       m2,    m3
+    psubw       m4,    m5
+
+    lea         r2,    [r2 + 2 * r4]
+    lea         r3,    [r3 + 2 * r5]
+
+    movh        m6,    [r2 + r4]
+    movh        m7,    [r3 + r5]
+    pmovzxbw    m6,    m6
+    pmovzxbw    m7,    m7
+
+    psubw       m6,    m7
+
+    movu    [r0],             m0
+    movu    [r0 + r1],        m2
+    movu    [r0 + 2 * r1],    m4
+
+    lea     r0,               [r0 + 2 * r1]
+
+    movu    [r0 + r1],        m6
+
+    lea     r2,               [r2 + 2 * r4]
+    lea     r3,               [r3 + 2 * r5]
+    lea     r0,               [r0 + 2 * r1]
+
+    dec     r6d
+
+jnz    .loop
+
+RET
+%endmacro
+
+PIXELSUB_PS_W8_H4 8, 8
+PIXELSUB_PS_W8_H4 8, 16
+PIXELSUB_PS_W8_H4 8, 32
+
+;-----------------------------------------------------------------------------
+; void pixel_sub_ps_c_%1x%2(int16_t *dest, intptr_t destride, pixel *src0, pixel *src1, intptr_t srcstride0, intptr_t srcstride1);
+;-----------------------------------------------------------------------------
+%macro PIXELSUB_PS_W12_H4 2
+INIT_XMM sse4
+cglobal pixel_sub_ps_%1x%2, 6, 7, 6, dest, deststride, src0, src1, srcstride0, srcstride1
+
+add    r1,     r1
+mov    r6d,    %2/4
+
+.loop
+
+    movu        m0,    [r2]
+    movu        m1,    [r3]
+    movu        m2,    [r2 + r4]
+    movu        m3,    [r3 + r5]
+
+    mova        m4,    m0
+    mova        m5,    m1
+    punpckhdq   m4,    m2
+    punpckhdq   m5,    m3
+
+    pmovzxbw    m0,    m0
+    pmovzxbw    m1,    m1
+    pmovzxbw    m2,    m2
+    pmovzxbw    m3,    m3
+    pmovzxbw    m4,    m4
+    pmovzxbw    m5,    m5
+
+    psubw       m0,    m1
+    psubw       m2,    m3
+    psubw       m4,    m5
+
+    movu      [r0],              m0
+    movlps    [r0 + 16],         m4
+    movu      [r0 + r1],         m2
+    movhps    [r0 + r1 + 16],    m4
+
+    movu      m0,    [r2 + 2 * r4]
+    movu      m1,    [r3 + 2 * r5]
+
+    lea       r2,    [r2 + 2 * r4]
+    lea       r3,    [r3 + 2 * r5]
+
+    movu      m2,    [r2 + r4]
+    movu      m3,    [r3 + r5]
+
+    mova         m4,    m0
+    mova         m5,    m1
+    punpckhdq    m4,    m2
+    punpckhdq    m5,    m3
+
+    pmovzxbw     m0,    m0
+    pmovzxbw     m1,    m1
+    pmovzxbw     m2,    m2
+    pmovzxbw     m3,    m3
+    pmovzxbw     m4,    m4
+    pmovzxbw     m5,    m5
+
+    psubw        m0,    m1
+    psubw        m2,    m3
+    psubw        m4,    m5
+
+    movu      [r0 + 2 * r1],         m0
+    movlps    [r0 + 2 * r1 + 16],    m4
+
+    lea       r0,                    [r0 + 2 * r1]
+
+    movu      [r0 + r1],             m2
+    movhps    [r0 + r1 + 16],        m4
+
+    lea       r2,                    [r2 + 2 * r4]
+    lea       r3,                    [r3 + 2 * r5]
+    lea       r0,                    [r0 + 2 * r1]
+
+    dec    r6d
+
+jnz    .loop
+
+RET
+%endmacro
+
+PIXELSUB_PS_W12_H4 12, 16
+
+;-----------------------------------------------------------------------------
+; void pixel_sub_ps_c_%1x%2(int16_t *dest, intptr_t destride, pixel *src0, pixel *src1, intptr_t srcstride0, intptr_t srcstride1);
+;-----------------------------------------------------------------------------
+%macro PIXELSUB_PS_W16_H4 2
+INIT_XMM sse4
+cglobal pixel_sub_ps_%1x%2, 6, 7, 7, dest, deststride, src0, src1, srcstride0, srcstride1
+
+add    r1,     r1
+mov    r6d,    %2/4
+pxor   m6,     m6
+
+.loop
+
+    movu         m1,    [r2]
+    pmovzxbw     m0,    m1
+    punpckhbw    m1,    m6
+    movu         m3,    [r3]
+    pmovzxbw     m2,    m3
+    punpckhbw    m3,    m6
+
+    psubw        m0,    m2
+    psubw        m1,    m3
+
+    movu         m5,    [r2 + r4]
+    pmovzxbw     m4,    m5
+    punpckhbw    m5,    m6
+    movu         m3,    [r3 + r5]
+    pmovzxbw     m2,    m3
+    punpckhbw    m3,    m6
+
+    psubw        m4,    m2
+    psubw        m5,    m3
+
+    movu    [r0],              m0
+    movu    [r0 + 16],         m1
+    movu    [r0 + r1],         m4
+    movu    [r0 + r1 + 16],    m5
+
+    movu         m1,    [r2 + 2 * r4]
+    pmovzxbw     m0,    m1
+    punpckhbw    m1,    m6
+    movu         m3,    [r3 + 2 * r5]
+    pmovzxbw     m2,    m3
+    punpckhbw    m3,    m6
+
+    lea          r2,    [r2 + 2 * r4]
+    lea          r3,    [r3 + 2 * r5]
+
+    psubw        m0,    m2
+    psubw        m1,    m3
+
+    movu         m5,    [r2 + r4]
+    pmovzxbw     m4,    m5
+    punpckhbw    m5,    m6
+    movu         m3,    [r3 + r5]
+    pmovzxbw     m2,    m3
+    punpckhbw    m3,    m6
+
+    psubw        m4,    m2
+    psubw        m5,    m3
+
+    movu    [r0 + 2 * r1],         m0
+    movu    [r0 + 2 * r1 + 16],    m1
+
+    lea     r0,                    [r0 + 2 * r1]
+
+    movu    [r0 + r1],             m4
+    movu    [r0 + r1 + 16],        m5
+
+    lea     r2,                    [r2 + 2 * r4]
+    lea     r3,                    [r3 + 2 * r5]
+    lea     r0,                    [r0 + 2 * r1]
+
+    dec    r6d
+
+jnz    .loop
+
+RET
+%endmacro
+
+PIXELSUB_PS_W16_H4 16, 4
+PIXELSUB_PS_W16_H4 16, 8
+PIXELSUB_PS_W16_H4 16, 12
+PIXELSUB_PS_W16_H4 16, 16
+PIXELSUB_PS_W16_H4 16, 32
+PIXELSUB_PS_W16_H4 16, 64
+
+;-----------------------------------------------------------------------------
+; void pixel_sub_ps_c_%1x%2(int16_t *dest, intptr_t destride, pixel *src0, pixel *src1, intptr_t srcstride0, intptr_t srcstride1);
+;-----------------------------------------------------------------------------
+%macro PIXELSUB_PS_W24_H2 2
+INIT_XMM sse4
+cglobal pixel_sub_ps_%1x%2, 6, 7, 7, dest, deststride, src0, src1, srcstride0, srcstride1
+
+add    r1,     r1
+mov    r6d,    %2/2
+pxor   m6,     m6
+
+.loop
+
+    movu         m1,    [r2]
+    pmovzxbw     m0,    m1
+    punpckhbw    m1,    m6
+    movh         m2,    [r2 + 16]
+    pmovzxbw     m2,    m2
+    movu         m4,    [r3]
+    pmovzxbw     m3,    m4
+    punpckhbw    m4,    m6
+    movh         m5,    [r3 + 16]
+    pmovzxbw     m5,    m5
+
+    psubw        m0,    m3
+    psubw        m1,    m4
+    psubw        m2,    m5
+
+    movu    [r0],         m0
+    movu    [r0 + 16],    m1
+    movu    [r0 + 32],    m2
+
+    movu         m1,    [r2 + r4]
+    pmovzxbw     m0,    m1
+    punpckhbw    m1,    m6
+    movh         m2,    [r2 + r4 + 16]
+    pmovzxbw     m2,    m2
+    movu         m4,    [r3 + r5]
+    pmovzxbw     m3,    m4
+    punpckhbw    m4,    m6
+    movh         m5,    [r3 + r5 + 16]
+    pmovzxbw     m5,    m5
+
+    psubw        m0,    m3
+    psubw        m1,    m4
+    psubw        m2,    m5
+
+    movu    [r0 + r1],         m0
+    movu    [r0 + r1 + 16],    m1
+    movu    [r0 + r1 + 32],    m2
+
+    lea    r2,    [r2 + 2 * r4]
+    lea    r3,    [r3 + 2 * r5]
+    lea    r0,    [r0 + 2 * r1]
+
+    dec    r6d
+
+jnz    .loop
+
+RET
+%endmacro
+
+PIXELSUB_PS_W24_H2 24, 32
+
+;-----------------------------------------------------------------------------
+; void pixel_sub_ps_c_%1x%2(int16_t *dest, intptr_t destride, pixel *src0, pixel *src1, intptr_t srcstride0, intptr_t srcstride1);
+;-----------------------------------------------------------------------------
+%macro PIXELSUB_PS_W32_H2 2
+INIT_XMM sse4
+cglobal pixel_sub_ps_%1x%2, 6, 7, 8, dest, deststride, src0, src1, srcstride0, srcstride1
+
+add    r1,     r1
+mov    r6d,    %2/2
+
+.loop
+
+    movh        m0,    [r2]
+    movh        m1,    [r2 + 8]
+    movh        m2,    [r2 + 16]
+    movh        m3,    [r2 + 24]
+    movh        m4,    [r3]
+    movh        m5,    [r3 + 8]
+    movh        m6,    [r3 + 16]
+    movh        m7,    [r3 + 24]
+
+    pmovzxbw    m0,    m0
+    pmovzxbw    m1,    m1
+    pmovzxbw    m2,    m2
+    pmovzxbw    m3,    m3
+    pmovzxbw    m4,    m4
+    pmovzxbw    m5,    m5
+    pmovzxbw    m6,    m6
+    pmovzxbw    m7,    m7
+
+    psubw       m0,    m4
+    psubw       m1,    m5
+    psubw       m2,    m6
+    psubw       m3,    m7
+
+    movu    [r0],         m0
+    movu    [r0 + 16],    m1
+    movu    [r0 + 32],    m2
+    movu    [r0 + 48],    m3
+
+    movh        m0,    [r2 + r4]
+    movh        m1,    [r2 + r4 + 8]
+    movh        m2,    [r2 + r4 + 16]
+    movh        m3,    [r2 + r4 + 24]
+    movh        m4,    [r3 + r5]
+    movh        m5,    [r3 + r5 + 8]
+    movh        m6,    [r3 + r5 + 16]
+    movh        m7,    [r3 + r5 + 24]
+
+    pmovzxbw    m0,    m0
+    pmovzxbw    m1,    m1
+    pmovzxbw    m2,    m2
+    pmovzxbw    m3,    m3
+    pmovzxbw    m4,    m4
+    pmovzxbw    m5,    m5
+    pmovzxbw    m6,    m6
+    pmovzxbw    m7,    m7
+
+    psubw       m0,    m4
+    psubw       m1,    m5
+    psubw       m2,    m6
+    psubw       m3,    m7
+
+    movu    [r0 + r1],         m0
+    movu    [r0 + r1 + 16],    m1
+    movu    [r0 + r1 + 32],    m2
+    movu    [r0 + r1 + 48],    m3
+
+    lea    r2,    [r2 + 2 * r4]
+    lea    r3,    [r3 + 2 * r5]
+    lea    r0,    [r0 + 2 * r1]
+
+    dec    r6d
+
+jnz    .loop
+
+RET
+%endmacro
+
+PIXELSUB_PS_W32_H2 32, 8
+PIXELSUB_PS_W32_H2 32, 16
+PIXELSUB_PS_W32_H2 32, 24
+PIXELSUB_PS_W32_H2 32, 32
+PIXELSUB_PS_W32_H2 32, 64
+
+;-----------------------------------------------------------------------------
+; void pixel_sub_ps_c_%1x%2(int16_t *dest, intptr_t destride, pixel *src0, pixel *src1, intptr_t srcstride0, intptr_t srcstride1);
+;-----------------------------------------------------------------------------
+%macro PIXELSUB_PS_W48_H2 2
+INIT_XMM sse4
+cglobal pixel_sub_ps_%1x%2, 6, 7, 7, dest, deststride, src0, src1, srcstride0, srcstride1
+
+add    r1,     r1
+mov    r6d,    %2/2
+pxor   m6,    m6
+
+.loop
+
+    movu         m1,    [r2]
+    pmovzxbw     m0,    m1
+    punpckhbw    m1,    m6
+    movu         m3,    [r3]
+    pmovzxbw     m2,    m3
+    punpckhbw    m3,    m6
+    movu         m5,    [r2 + 16]
+    pmovzxbw     m4,    m5
+    punpckhbw    m5,    m6
+
+    psubw        m0,    m2
+    psubw        m1,    m3
+
+    movu    [r0],         m0
+    movu    [r0 + 16],    m1
+
+    movu         m3,    [r3 + 16]
+    pmovzxbw     m2,    m3
+    punpckhbw    m3,    m6
+
+    psubw        m4,    m2
+    psubw        m5,    m3
+
+    movu    [r0 + 32],    m4
+    movu    [r0 + 48],    m5
+
+    movu         m1,    [r2 + 32]
+    pmovzxbw     m0,    m1
+    punpckhbw    m1,    m6
+    movu         m3,    [r3 + 32]
+    pmovzxbw     m2,    m3
+    punpckhbw    m3,    m6
+
+    psubw        m0,    m2
+    psubw        m1,    m3
+
+    movu    [r0 + 64],    m0
+    movu    [r0 + 80],    m1
+
+    movu         m1,    [r2 + r4]
+    pmovzxbw     m0,    m1
+    punpckhbw    m1,    m6
+    movu         m3,    [r3 + r5]
+    pmovzxbw     m2,    m3
+    punpckhbw    m3,    m6
+    movu         m5,    [r2 + r5 + 16]
+    pmovzxbw     m4,    m5
+    punpckhbw    m5,    m6
+
+    psubw        m0,    m2
+    psubw        m1,    m3
+
+    movu    [r0 + r1],         m0
+    movu    [r0 + r1 + 16],    m1
+
+    movu         m3,    [r3 + r4 + 16]
+    pmovzxbw     m2,    m3
+    punpckhbw    m3,    m6
+
+    psubw        m4,    m2
+    psubw        m5,    m3
+
+    movu    [r0 + r1 + 32],    m4
+    movu    [r0 + r1 + 48],    m5
+
+    movu         m1,    [r2 + r4 + 32]
+    pmovzxbw     m0,    m1
+    punpckhbw    m1,    m6
+    movu         m3,    [r3 + r5 + 32]
+    pmovzxbw     m2,    m3
+    punpckhbw    m3,    m6
+
+    psubw        m0,    m2
+    psubw        m1,    m3
+
+    movu    [r0 + r1 + 64],    m0
+    movu    [r0 + r1 + 80],    m1
+
+    lea     r2,                [r2 + 2 * r4]
+    lea     r3,                [r3 + 2 * r5]
+    lea     r0,                [r0 + 2 * r1]
+
+    dec    r6d
+
+jnz    .loop
+
+RET
+%endmacro
+
+PIXELSUB_PS_W48_H2 48, 64
+
+;-----------------------------------------------------------------------------
+; void pixel_sub_ps_c_%1x%2(int16_t *dest, intptr_t destride, pixel *src0, pixel *src1, intptr_t srcstride0, intptr_t srcstride1);
+;-----------------------------------------------------------------------------
+%macro PIXELSUB_PS_W64_H2 2
+INIT_XMM sse4
+cglobal pixel_sub_ps_%1x%2, 6, 7, 7, dest, deststride, src0, src1, srcstride0, srcstride1
+
+add    r1,     r1
+mov    r6d,    %2/2
+pxor   m6,    m6
+
+.loop
+
+    movu         m1,    [r2]
+    pmovzxbw     m0,    m1
+    punpckhbw    m1,    m6
+    movu         m3,    [r3]
+    pmovzxbw     m2,    m3
+    punpckhbw    m3,    m6
+    movu         m5,    [r2 + 16]
+    pmovzxbw     m4,    m5
+    punpckhbw    m5,    m6
+
+    psubw        m0,    m2
+    psubw        m1,    m3
+
+    movu    [r0],         m0
+    movu    [r0 + 16],    m1
+
+    movu         m1,    [r3 + 16]
+    pmovzxbw     m0,    m1
+    punpckhbw    m1,    m6
+    movu         m3,    [r2 + 32]
+    pmovzxbw     m2,    m3
+    punpckhbw    m3,    m6
+
+    psubw        m4,    m0
+    psubw        m5,    m1
+
+    movu    [r0 + 32],    m4
+    movu    [r0 + 48],    m5
+
+    movu         m5,    [r3 + 32]
+    pmovzxbw     m4,    m5
+    punpckhbw    m5,    m6
+    movu         m1,    [r2 + 48]
+    pmovzxbw     m0,    m1
+    punpckhbw    m1,    m6
+
+    psubw        m2,    m4
+    psubw        m3,    m5
+
+    movu    [r0 + 64],    m2
+    movu    [r0 + 80],    m3
+
+    movu         m3,    [r3 + 48]
+    pmovzxbw     m2,    m3
+    punpckhbw    m3,    m6
+    movu         m5,    [r2 + r4]
+    pmovzxbw     m4,    m5
+    punpckhbw    m5,    m6
+
+    psubw        m0,    m2
+    psubw        m1,    m3
+
+    movu    [r0 + 96],     m0
+    movu    [r0 + 112],    m1
+
+    movu         m1,    [r3 + r5]
+    pmovzxbw     m0,    m1
+    punpckhbw    m1,    m6
+    movu         m3,    [r2 + r4 + 16]
+    pmovzxbw     m2,    m3
+    punpckhbw    m3,    m6
+
+    psubw        m4,    m0
+    psubw        m5,    m1
+
+    movu    [r0 + r1],         m4
+    movu    [r0 + r1 + 16],    m5
+
+    movu         m5,    [r3 + r5 + 16]
+    pmovzxbw     m4,    m5
+    punpckhbw    m5,    m6
+    movu         m1,    [r2 + r4 + 32]
+    pmovzxbw     m0,    m1
+    punpckhbw    m1,    m6
+
+    psubw        m2,    m4
+    psubw        m3,    m5
+
+    movu    [r0 + r1 + 32],    m2
+    movu    [r0 + r1 + 48],    m3
+
+    movu         m3,    [r3 + r5 + 32]
+    pmovzxbw     m2,    m3
+    punpckhbw    m3,    m6
+    movu         m5,    [r2 + r4 + 48]
+    pmovzxbw     m4,    m5
+    punpckhbw    m5,    m6
+
+    psubw        m0,    m2
+    psubw        m1,    m3
+
+    movu    [r0 + r1 + 64],    m0
+    movu    [r0 + r1 + 80],    m1
+
+    movu         m1,    [r3 + r5 + 48]
+    pmovzxbw     m0,    m1
+    punpckhbw    m1,    m6
+
+    psubw        m4,    m0
+    psubw        m5,    m1
+
+    movu    [r0 + r1 + 96],     m4
+    movu    [r0 + r1 + 112],    m5
+
+    lea     r2,                 [r2 + 2 * r4]
+    lea     r3,                 [r3 + 2 * r5]
+    lea     r0,                 [r0 + 2 * r1]
+
+    dec    r6d
+
+jnz    .loop
+
+RET
+%endmacro
+
+PIXELSUB_PS_W64_H2 64, 16
+PIXELSUB_PS_W64_H2 64, 32
+PIXELSUB_PS_W64_H2 64, 48
+PIXELSUB_PS_W64_H2 64, 64


More information about the x265-devel mailing list