[x265] [PATCH] asm: idct[8x8] sse2 12232 -> 3500 over c code 3550 -> 3500 over intrinsic
chen
chenm003 at 163.com
Fri Dec 19 00:24:09 CET 2014
This code is right, thanks
There just a little mistake at below.
of course, this code difficult to maintenance,
At 2014-12-19 06:57:26,dtyx265 at gmail.com wrote:
># HG changeset patch
># User David T Yuen <dtyx265 at gmail.com>
># Date 1418943389 28800
># Node ID 745b4ad1b0042aec5a1962b0d67053f5ddee0642
># Parent 78ae7996a1ceb60d24cff790cc2fa233d4c31435
>asm: idct[8x8] sse2 12232 -> 3500 over c code 3550 -> 3500 over intrinsic
>
>diff -r 78ae7996a1ce -r 745b4ad1b004 source/common/x86/asm-primitives.cpp
>--- a/source/common/x86/asm-primitives.cpp Wed Dec 17 14:31:50 2014 -0600
>+++ b/source/common/x86/asm-primitives.cpp Thu Dec 18 14:56:29 2014 -0800
>@@ -1387,6 +1387,9 @@
>
> p.dct[DCT_4x4] = x265_dct4_sse2;
> p.idct[IDCT_4x4] = x265_idct4_sse2;
>+#if X86_64
>+ p.idct[IDCT_8x8] = x265_idct8_sse2;
>+#endif
> p.idct[IDST_4x4] = x265_idst4_sse2;
>
> LUMA_SS_FILTERS(_sse2);
>@@ -1593,6 +1596,9 @@
>
> p.dct[DCT_4x4] = x265_dct4_sse2;
> p.idct[IDCT_4x4] = x265_idct4_sse2;
>+#if X86_64
>+ p.idct[IDCT_8x8] = x265_idct8_sse2;
>+#endif
> p.idct[IDST_4x4] = x265_idst4_sse2;
>
> p.planecopy_sp = x265_downShift_16_sse2;
>diff -r 78ae7996a1ce -r 745b4ad1b004 source/common/x86/dct8.asm
>--- a/source/common/x86/dct8.asm Wed Dec 17 14:31:50 2014 -0600
>+++ b/source/common/x86/dct8.asm Thu Dec 18 14:56:29 2014 -0800
>@@ -976,6 +976,382 @@
> ;-------------------------------------------------------
> ; void idct8(const int16_t* src, int16_t* dst, intptr_t dstStride)
> ;-------------------------------------------------------
>+%if ARCH_X86_64
>+INIT_XMM sse2
>+%if BIT_DEPTH == 10
>+ %define IDCT_SHIFT 10
>+ %define IDCT_ADD pd_512
>+%elif BIT_DEPTH == 8
>+ %define IDCT_SHIFT 12
>+ %define IDCT_ADD pd_2048
>+%else
>+ %error Unsupported BIT_DEPTH!
>+%endif
>+
>+cglobal idct8, 3, 6, 16, 0-5*mmsize
>+ mova m9, [r0 + 1 * mmsize]
>+ mova m1, [r0 + 3 * mmsize]
>+ mova m7, m9
>+ punpcklwd m7, m1
>+ punpckhwd m9, m1
>+ mova m14, [tab_idct8_3]
>+ mova m3, m14
>+ pmaddwd m14, m7
>+ pmaddwd m3, m9
>+ mova m0, [r0 + 5 * mmsize]
>+ mova m10, [r0 + 7 * mmsize]
>+ mova m2, m0
>+ punpcklwd m2, m10
>+ punpckhwd m0, m10
>+ mova m15, [tab_idct8_3 + 1 * mmsize]
>+ mova m11, [tab_idct8_3 + 1 * mmsize]
>+ pmaddwd m15, m2
>+ mova m4, [tab_idct8_3 + 2 * mmsize]
>+ pmaddwd m11, m0
>+ mova m1, [tab_idct8_3 + 2 * mmsize]
>+ paddd m15, m14
>+ mova m5, [tab_idct8_3 + 4 * mmsize]
>+ mova m12, [tab_idct8_3 + 4 * mmsize]
>+ paddd m11, m3
>+ mova [rsp + 0 * mmsize], m11
>+ mova [rsp + 1 * mmsize], m15
>+ pmaddwd m4, m7
>+ pmaddwd m1, m9
>+ mova m14, [tab_idct8_3 + 3 * mmsize]
>+ mova m3, [tab_idct8_3 + 3 * mmsize]
>+ pmaddwd m14, m2
>+ pmaddwd m3, m0
>+ paddd m14, m4
>+ paddd m3, m1
>+ mova [rsp + 2 * mmsize], m3
>+ pmaddwd m5, m9
>+ pmaddwd m9, [tab_idct8_3 + 6 * mmsize]
>+ mova m6, [tab_idct8_3 + 5 * mmsize]
>+ pmaddwd m12, m7
>+ pmaddwd m7, [tab_idct8_3 + 6 * mmsize]
>+ mova m4, [tab_idct8_3 + 5 * mmsize]
>+ pmaddwd m6, m2
>+ paddd m6, m12
>+ pmaddwd m2, [tab_idct8_3 + 7 * mmsize]
>+ paddd m7, m2
>+ mova [rsp + 3 * mmsize], m6
>+ pmaddwd m4, m0
>+ pmaddwd m0, [tab_idct8_3 + 7 * mmsize]
>+ paddd m9, m0
>+ paddd m5, m4
>+ mova m6, [r0 + 0 * mmsize]
>+ mova m0, [r0 + 4 * mmsize]
>+ mova m4, m6
>+ punpcklwd m4, m0
>+ punpckhwd m6, m0
>+ mova m12, [r0 + 2 * mmsize]
>+ mova m0, [r0 + 6 * mmsize]
>+ mova m13, m12
>+ mova m8, [tab_dct4]
>+ punpcklwd m13, m0
>+ mova m10, [tab_dct4]
>+ punpckhwd m12, m0
>+ pmaddwd m8, m4
>+ mova m3, m8
>+ pmaddwd m4, [tab_dct4 + 2 * mmsize]
>+ pmaddwd m10, m6
>+ mova m2, [tab_dct4 + 1 * mmsize]
>+ mova m1, m10
>+ pmaddwd m6, [tab_dct4 + 2 * mmsize]
>+ mova m0, [tab_dct4 + 1 * mmsize]
>+ pmaddwd m2, m13
>+ paddd m3, m2
>+ psubd m8, m2
>+ mova m2, m6
>+ pmaddwd m13, [tab_dct4 + 3 * mmsize]
>+ pmaddwd m0, m12
>+ paddd m1, m0
>+ psubd m10, m0
>+ mova m0, m4
>+ pmaddwd m12, [tab_dct4 + 3 * mmsize]
>+ paddd m3, [pd_64]
>+ paddd m1, [pd_64]
>+ paddd m8, [pd_64]
>+ paddd m10, [pd_64]
>+ paddd m0, m13
>+ paddd m2, m12
>+ paddd m0, [pd_64]
>+ paddd m2, [pd_64]
>+ psubd m4, m13
>+ psubd m6, m12
>+ paddd m4, [pd_64]
>+ paddd m6, [pd_64]
>+ mova m12, m8
>+ psubd m8, m7
>+ psrad m8, 7
>+ paddd m15, m3
>+ psubd m3, [rsp + 1 * mmsize]
>+ psrad m15, 7
>+ paddd m12, m7
>+ psrad m12, 7
>+ paddd m11, m1
>+ mova m13, m14
>+ psrad m11, 7
>+ packssdw m15, m11
>+ psubd m1, [rsp + 0 * mmsize]
>+ psrad m1, 7
>+ mova m11, [rsp + 2 * mmsize]
>+ paddd m14, m0
>+ psrad m14, 7
>+ psubd m0, m13
>+ psrad m0, 7
>+ paddd m11, m2
>+ mova m13, [rsp + 3 * mmsize]
>+ psrad m11, 7
>+ packssdw m14, m11
>+ mova m11, m6
>+ psubd m6, m5
>+ paddd m13, m4
>+ psrad m13, 7
>+ psrad m6, 7
>+ paddd m11, m5
>+ psrad m11, 7
>+ packssdw m13, m11
>+ mova m11, m10
>+ psubd m4, [rsp + 3 * mmsize]
>+ psubd m10, m9
>+ psrad m4, 7
>+ psrad m10, 7
>+ packssdw m4, m6
>+ packssdw m8, m10
>+ paddd m11, m9
>+ psrad m11, 7
>+ packssdw m12, m11
>+ psubd m2, [rsp + 2 * mmsize]
>+ mova m5, m15
>+ psrad m2, 7
>+ packssdw m0, m2
>+ mova m2, m14
>+ psrad m3, 7
>+ packssdw m3, m1
>+ mova m6, m13
>+ punpcklwd m5, m8
>+ punpcklwd m2, m4
>+ mova m1, m12
>+ punpcklwd m6, m0
>+ punpcklwd m1, m3
>+ mova m9, m5
>+ punpckhwd m13, m0
>+ mova m0, m2
>+ punpcklwd m9, m6
>+ punpckhwd m5, m6
>+ punpcklwd m0, m1
>+ punpckhwd m2, m1
>+ punpckhwd m15, m8
>+ mova m1, m5
>+ punpckhwd m14, m4
>+ punpckhwd m12, m3
>+ mova m6, m9
>+ punpckhwd m9, m0
>+ punpcklwd m1, m2
>+ mova m4, [tab_idct8_3 + 0 * mmsize]
>+ punpckhwd m5, m2
>+ punpcklwd m6, m0
>+ mova m2, m15
>+ mova m0, m14
>+ mova m7, m9
>+ punpcklwd m2, m13
>+ punpcklwd m0, m12
>+ punpcklwd m7, m5
>+ punpckhwd m14, m12
>+ mova m10, m2
>+ punpckhwd m15, m13
>+ punpckhwd m9, m5
>+ pmaddwd m4, m7
>+ mova m13, m1
>+ punpckhwd m2, m0
>+ punpcklwd m10, m0
>+ mova m0, m15
>+ punpckhwd m15, m14
>+ mova m12, m1
>+ mova m3, [tab_idct8_3 + 0 * mmsize]
>+ punpcklwd m0, m14
>+ pmaddwd m3, m9
>+ mova m11, m2
>+ punpckhwd m2, m15
>+ punpcklwd m11, m15
>+ mova m8, [tab_idct8_3 + 1 * mmsize]
>+ punpcklwd m13, m0
>+ punpckhwd m12, m0
>+ pmaddwd m8, m11
>+ paddd m8, m4
>+ mova [rsp + 4 * mmsize], m8
>+ mova m4, [tab_idct8_3 + 2 * mmsize]
>+ pmaddwd m4, m7
>+ mova m15, [tab_idct8_3 + 2 * mmsize]
>+ mova m5, [tab_idct8_3 + 1 * mmsize]
>+ pmaddwd m15, m9
>+ pmaddwd m5, m2
>+ paddd m5, m3
>+ mova [rsp + 3 * mmsize], m5
>+ mova m14, [tab_idct8_3 + 3 * mmsize]
>+ mova m5, [tab_idct8_3 + 3 * mmsize]
>+ pmaddwd m14, m11
>+ paddd m14, m4
>+ mova [rsp + 2 * mmsize], m14
>+ pmaddwd m5, m2
>+ paddd m5, m15
>+ mova [rsp + 1 * mmsize], m5
>+ mova m15, [tab_idct8_3 + 4 * mmsize]
>+ mova m5, [tab_idct8_3 + 4 * mmsize]
>+ pmaddwd m15, m7
>+ pmaddwd m7, [tab_idct8_3 + 6 * mmsize]
>+ pmaddwd m5, m9
>+ pmaddwd m9, [tab_idct8_3 + 6 * mmsize]
>+ mova m4, [tab_idct8_3 + 5 * mmsize]
>+ pmaddwd m4, m2
>+ paddd m5, m4
>+ mova m4, m6
>+ mova m8, [tab_idct8_3 + 5 * mmsize]
>+ punpckhwd m6, m10
>+ pmaddwd m2, [tab_idct8_3 + 7 * mmsize]
>+ punpcklwd m4, m10
>+ paddd m9, m2
>+ pmaddwd m8, m11
>+ mova m10, [tab_dct4]
>+ paddd m8, m15
>+ pmaddwd m11, [tab_idct8_3 + 7 * mmsize]
>+ paddd m7, m11
>+ mova [rsp + 0 * mmsize], m8
>+ pmaddwd m10, m6
>+ pmaddwd m6, [tab_dct4 + 2 * mmsize]
>+ mova m1, m10
>+ mova m8, [tab_dct4]
>+ mova m3, [tab_dct4 + 1 * mmsize]
>+ pmaddwd m8, m4
>+ pmaddwd m4, [tab_dct4 + 2 * mmsize]
>+ mova m0, m8
>+ mova m2, [tab_dct4 + 1 * mmsize]
>+ pmaddwd m3, m13
>+ psubd m8, m3
>+ paddd m0, m3
>+ mova m3, m6
>+ pmaddwd m13, [tab_dct4 + 3 * mmsize]
>+ pmaddwd m2, m12
>+ paddd m1, m2
>+ psubd m10, m2
>+ mova m2, m4
>+ pmaddwd m12, [tab_dct4 + 3 * mmsize]
>+ paddd m0, [IDCT_ADD]
>+ paddd m1, [IDCT_ADD]
>+ paddd m8, [IDCT_ADD]
>+ paddd m10, [IDCT_ADD]
>+ paddd m2, m13
>+ paddd m3, m12
>+ paddd m2, [IDCT_ADD]
>+ paddd m3, [IDCT_ADD]
>+ psubd m4, m13
>+ psubd m6, m12
>+ paddd m4, [IDCT_ADD]
>+ paddd m6, [IDCT_ADD]
>+ mova m15, [rsp + 4 * mmsize]
>+ mova m12, m8
>+ psubd m8, m7
>+ psrad m8, IDCT_SHIFT
>+ mova m11, [rsp + 3 * mmsize]
>+ paddd m15, m0
>+ psrad m15, IDCT_SHIFT
>+ psubd m0, [rsp + 4 * mmsize]
>+ psrad m0, IDCT_SHIFT
>+ paddd m12, m7
>+ paddd m11, m1
>+ mova m14, [rsp + 2 * mmsize]
>+ psrad m11, IDCT_SHIFT
>+ packssdw m15, m11
>+ psubd m1, [rsp + 3 * mmsize]
>+ psrad m1, IDCT_SHIFT
>+ mova m11, [rsp + 1 * mmsize]
>+ paddd m14, m2
>+ psrad m14, IDCT_SHIFT
>+ packssdw m0, m1
>+ psrad m12, IDCT_SHIFT
>+ psubd m2, [rsp + 2 * mmsize]
>+ paddd m11, m3
>+ mova m13, [rsp + 0 * mmsize]
>+ psrad m11, IDCT_SHIFT
>+ packssdw m14, m11
>+ mova m11, m6
>+ psubd m6, m5
>+ paddd m13, m4
>+ psrad m13, IDCT_SHIFT
>+ mova m1, m15
>+ paddd m11, m5
>+ psrad m11, IDCT_SHIFT
>+ packssdw m13, m11
>+ mova m11, m10
>+ psubd m10, m9
>+ psrad m10, IDCT_SHIFT
>+ packssdw m8, m10
>+ psrad m6, IDCT_SHIFT
>+ psubd m4, [rsp + 0 * mmsize]
>+ paddd m11, m9
>+ psrad m11, IDCT_SHIFT
>+ packssdw m12, m11
>+ punpcklwd m1, m14
>+ mova m5, m13
>+ psrad m4, IDCT_SHIFT
>+ packssdw m4, m6
>+ psubd m3, [rsp + 1 * mmsize]
>+ psrad m2, IDCT_SHIFT
>+ mova m6, m8
>+ psrad m3, IDCT_SHIFT
>+ punpcklwd m5, m12
>+ packssdw m2, m3
>+ punpcklwd m6, m4
>+ punpckhwd m8, m4
>+ mova m4, m1
>+ mova m3, m2
>+ punpckhdq m1, m5
>+ punpckldq m4, m5
>+ punpcklwd m3, m0
>+ punpckhwd m2, m0
>+ mova m0, m6
>+ lea r2, [r2 + r2]
>+ lea r4, [r2 + r2]
>+ lea r3, [r4 + r2]
>+ lea r4, [r4 + r3]
>+ lea r0, [r4 + r2 * 2]
>+ movq [r1], m4
>+ punpckhwd m15, m14
>+ movhps [r1 + r2], m4
>+ punpckhdq m0, m3
>+ movq [r1 + r2 * 2], m1
>+ punpckhwd m13, m12
>+ movhps [r1 + r3], m1
>+ mova m1, m6
>+ punpckldq m1, m3
>+ movq [r1 + 8], m1
>+ movhps [r1 + r2 + 8], m1
>+ movq [r1 + r2 * 2 + 8], m0
>+ movhps [r1 + r3 + 8], m0
>+ mova m0, m15
>+ punpckhdq m15, m13
>+ punpckldq m0, m13
>+ movq [r1 + r2 * 4], m0
>+ movhps [r1 + r4], m0
>+ mova m0, m8
>+ punpckhdq m8, m2
>+ movq [r1 + r3 * 2], m15
>+ punpckldq m0, m2
>+ movhps [r1 + r0], m15
>+ movq [r1 + r2 * 4 + 8], m0
>+ movhps [r1 + r4 + 8], m0
>+ movq [r1 + r3 * 2 + 8], m8
>+ movhps [r1 + r0 + 8], m8
>+ RET
>+
>+%undef IDCT_SHIFT
>+%undef IDCT_ADD
>+%endif
>+
>+;-------------------------------------------------------
>+; void idct8(const int16_t* src, int16_t* dst, intptr_t dstStride)
>+;-------------------------------------------------------
> INIT_XMM ssse3
>
> cglobal patial_butterfly_inverse_internal_pass1
>diff -r 78ae7996a1ce -r 745b4ad1b004 source/common/x86/dct8.h
>--- a/source/common/x86/dct8.h Wed Dec 17 14:31:50 2014 -0600
>+++ b/source/common/x86/dct8.h Thu Dec 18 14:56:29 2014 -0800
>@@ -34,6 +34,9 @@
> void x265_idst4_sse2(const int16_t* src, int16_t* dst, intptr_t dstStride);
> void x265_idct4_sse2(const int16_t* src, int16_t* dst, intptr_t dstStride);
> void x265_idct4_avx2(const int16_t* src, int16_t* dst, intptr_t dstStride);
>+#if X86_64
>+void x265_idct8_sse2(const int16_t* src, int16_t* dst, intptr_t dstStride);
>+#endif
we don't need check enviorment here, the declare may share with any mode.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.videolan.org/pipermail/x265-devel/attachments/20141219/83f37c66/attachment-0001.html>
More information about the x265-devel
mailing list