[x265] [PATCH 1 of 3] asm: residual buffer is alignment to size, so we can use alignment load instruction
chen
chenm003 at 163.com
Sat Nov 16 10:25:17 CET 2013
Here is origin patch
在 2013-11-16 13:42:16,"Steve Borho" <steve at borho.org> 写道:
On Thu, Nov 14, 2013 at 11:18 PM, Min Chen <chenm003 at 163.com> wrote:
# HG changeset patch
# User Min Chen <chenm003 at 163.com>
# Date 1384492636 -28800
# Node ID 9592525e376f4b41d4fde000ae77814a00b06822
# Parent ee42f57411ae746095dd36e36064145ed869d73c
asm: residual buffer is alignment to size, so we can use alignment load instruction
this patch is malformed, can you send it as a patch file?
diff -r ee42f57411ae -r 9592525e376f source/Lib/TLibEncoder/TEncSearch.cpp
--- a/source/Lib/TLibEncoder/TEncSearch.cpp Thu Nov 14 13:38:07 2013 -0600
+++ b/source/Lib/TLibEncoder/TEncSearch.cpp Fri Nov 15 13:17:16 2013 +0800
@@ -502,6 +502,8 @@
}
//===== reconstruction =====
+ assert(((uint32_t)residual & (width - 1)) == 0);
+ assert(width <= 32);
primitives.calcrecon[size](pred, residual, recon, reconQt, reconIPred, stride, MAX_CU_SIZE, reconIPredStride);
//===== update distortion =====
@@ -636,6 +638,8 @@
}
//===== reconstruction =====
+ assert(((uint32_t)residual & (width - 1)) == 0);
+ assert(width <= 32);
primitives.calcrecon[size](pred, residual, recon, reconQt, reconIPred, stride, MAX_CU_SIZE / 2, reconIPredStride);
//===== update distortion =====
diff -r ee42f57411ae -r 9592525e376f source/common/x86/pixel-util.asm
--- a/source/common/x86/pixel-util.asm Thu Nov 14 13:38:07 2013 -0600
+++ b/source/common/x86/pixel-util.asm Fri Nov 15 13:17:16 2013 +0800
@@ -1,475 +1,469 @@
-;*****************************************************************************
-;* Copyright (C) 2013 x265 project
-;*
-;* Authors: Min Chen <chenm003 at 163.com> <min.chen at multicorewareinc.com>
-;*
-;* This program is free software; you can redistribute it and/or modify
-;* it under the terms of the GNU General Public License as published by
-;* the Free Software Foundation; either version 2 of the License, or
-;* (at your option) any later version.
-;*
-;* This program is distributed in the hope that it will be useful,
-;* but WITHOUT ANY WARRANTY; without even the implied warranty of
-;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-;* GNU General Public License for more details.
-;*
-;* You should have received a copy of the GNU General Public License
-;* along with this program; if not, write to the Free Software
-;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02111, USA.
-;*
-;* This program is also available under a commercial proprietary license.
-;* For more information, contact us at licensing at multicorewareinc.com.
-;*****************************************************************************/
-
-%include "x86inc.asm"
-%include "x86util.asm"
-
-SECTION_RODATA 32
-
-SECTION .text
-
-
-;-----------------------------------------------------------------------------
-; void cvt32to16_shr(short *dst, int *src, intptr_t stride, int shift, int size)
-;-----------------------------------------------------------------------------
-INIT_XMM sse2
-cglobal cvt32to16_shr, 5, 7, 1, dst, src, stride
-%define rnd m7
-%define shift m6
-
- ; make shift
- mov r5d, r3m
- movd shift, r5d
-
- ; make round
- dec r5
- xor r6, r6
- bts r6, r5
-
- movd rnd, r6d
- pshufd rnd, rnd, 0
-
- ; register alloc
- ; r0 - dst
- ; r1 - src
- ; r2 - stride * 2 (short*)
- ; r3 - lx
- ; r4 - size
- ; r5 - ly
- ; r6 - diff
- lea r2, [r2 * 2]
-
- mov r4d, r4m
- mov r5, r4
- mov r6, r2
- sub r6, r4
- lea r6, [r6 * 2]
-
- shr r5, 1
-.loop_row:
-
- mov r3, r4
- shr r3, 2
-.loop_col:
- ; row 0
- movu m0, [r1]
- paddd m0, rnd
- psrad m0, shift
- packssdw m0, m0
- movh [r0], m0
-
- ; row 1
- movu m0, [r1 + r4 * 4]
- paddd m0, rnd
- psrad m0, shift
- packssdw m0, m0
- movh [r0 + r2], m0
-
- ; move col pointer
- add r1, 16
- add r0, 8
-
- dec r3
- jg .loop_col
-
- ; update pointer
- lea r1, [r1 + r4 * 4]
- add r0, r6
-
- ; end of loop_row
- dec r5
- jg .loop_row
-
- RET
-
-
-;-----------------------------------------------------------------------------
-; void calcrecon(pixel* pred, int16_t* residual, pixel* recon, int16_t* reconqt, pixel *reconipred, int stride, int strideqt, int strideipred)
-;-----------------------------------------------------------------------------
-INIT_XMM sse2
-cglobal calcRecons4
-%if ARCH_X86_64 == 1
- DECLARE_REG_TMP 0,1,2,3,4,5,6,7,8
- PROLOGUE 6,9,4
-%else
- DECLARE_REG_TMP 0,1,2,3,4,5
- PROLOGUE 6,7,4
- %define t6 r6m
- %define t6d r6d
- %define t7 r7m
- %define t8d r6d
-%endif
-
- mov t6d, r6m
-%if ARCH_X86_64 == 0
- add t6d, t6d
- mov r6m, t6d
-%else
- mov r5d, r5m
- mov r7d, r7m
- add t6d, t6d
-%endif
-
- pxor m0, m0
- mov t8d, 4/2
-.loop:
- movd m1, [t0]
- movd m2, [t0 + t5]
- punpckldq m1, m2
- punpcklbw m1, m0
- movh m2, [t1]
- movh m3, [t1 + t5 * 2]
- punpcklqdq m2, m3
- paddw m1, m2
- packuswb m1, m1
-
- ; store recon[] and recipred[]
- movd [t2], m1
- movd [t4], m1
- add t4, t7
- pshufd m2, m1, 1
- movd [t2 + t5], m2
- movd [t4], m2
- add t4, t7
-
- ; store recqt[]
- punpcklbw m1, m0
- movlps [t3], m1
- add t3, t6
- movhps [t3], m1
- add t3, t6
-
- lea t0, [t0 + t5 * 2]
- lea t1, [t1 + t5 * 4]
- lea t2, [t2 + t5 * 2]
-
- dec t8d
- jnz .loop
- RET
-
-
-INIT_XMM sse2
-cglobal calcRecons8
-%if ARCH_X86_64 == 1
- DECLARE_REG_TMP 0,1,2,3,4,5,6,7,8
- PROLOGUE 6,9,5
-%else
- DECLARE_REG_TMP 0,1,2,3,4,5
- PROLOGUE 6,7,5
- %define t6 r6m
- %define t6d r6d
- %define t7 r7m
- %define t8d r6d
-%endif
-
- mov t6d, r6m
-%if ARCH_X86_64 == 0
- add t6d, t6d
- mov r6m, t6d
-%else
- mov r5d, r5m
- mov r7d, r7m
- add t6d, t6d
-%endif
-
- pxor m0, m0
- mov t8d, 8/2
-.loop:
- movh m1, [t0]
- movh m2, [t0 + t5]
- punpcklbw m1, m0
- punpcklbw m2, m0
- movu m3, [t1]
- movu m4, [t1 + t5 * 2]
- paddw m1, m3
- paddw m2, m4
- packuswb m1, m2
-
- ; store recon[] and recipred[]
- movlps [t2], m1
- movhps [t2 + t5], m1
- movlps [t4], m1
-%if ARCH_X86_64 == 0
- add t4, t7
- movhps [t4], m1
- add t4, t7
-%else
- movhps [t4 + t7], m1
- lea t4, [t4 + t7 * 2]
-%endif
-
- ; store recqt[]
- punpcklbw m2, m1, m0
- punpckhbw m1, m0
- movu [t3], m2
- add t3, t6
- movu [t3], m1
- add t3, t6
-
- lea t0, [t0 + t5 * 2]
- lea t1, [t1 + t5 * 4]
- lea t2, [t2 + t5 * 2]
-
- dec t8d
- jnz .loop
- RET
-
-
-INIT_XMM sse4
-cglobal calcRecons16
-%if ARCH_X86_64 == 1
- DECLARE_REG_TMP 0,1,2,3,4,5,6,7,8
- PROLOGUE 6,9,5
-%else
- DECLARE_REG_TMP 0,1,2,3,4,5
- PROLOGUE 6,7,5
- %define t6 r6m
- %define t6d r6d
- %define t7 r7m
- %define t8d r6d
-%endif
-
- mov t6d, r6m
-%if ARCH_X86_64 == 0
- add t6d, t6d
- mov r6m, t6d
-%else
- mov r5d, r5m
- mov r7d, r7m
- add t6d, t6d
-%endif
-
- pxor m0, m0
- mov t8d, 16
-.loop:
- movu m2, [t0]
- pmovzxbw m1, m2
- punpckhbw m2, m0
- movu m3, [t1]
- movu m4, [t1 + 16]
- paddw m1, m3
- paddw m2, m4
- packuswb m1, m2
-
- ; store recon[] and recipred[]
- movu [t2], m1
- movu [t4], m1
-
- ; store recqt[]
- pmovzxbw m2, m1
- punpckhbw m1, m0
- movu [t3], m2
- movu [t3 + 16], m1
-
- add t3, t6
- add t4, t7
- add t0, t5
- lea t1, [t1 + t5 * 2]
- add t2, t5
-
- dec t8d
- jnz .loop
- RET
-
-
-INIT_XMM sse4
-cglobal calcRecons32
-%if ARCH_X86_64 == 1
- DECLARE_REG_TMP 0,1,2,3,4,5,6,7,8
- PROLOGUE 6,9,7
-%else
- DECLARE_REG_TMP 0,1,2,3,4,5
- PROLOGUE 6,7,7
- %define t6 r6m
- %define t6d r6d
- %define t7 r7m
- %define t8d r6d
-%endif
-
- mov t6d, r6m
-%if ARCH_X86_64 == 0
- add t6d, t6d
- mov r6m, t6d
-%else
- mov r5d, r5m
- mov r7d, r7m
- add t6d, t6d
-%endif
-
- pxor m0, m0
- mov t8d, 32
-.loop:
- movu m2, [t0]
- movu m4, [t0 + 16]
- pmovzxbw m1, m2
- punpckhbw m2, m0
- pmovzxbw m3, m4
- punpckhbw m4, m0
-
- movu m5, [t1 + 0 * 16]
- movu m6, [t1 + 1 * 16]
- paddw m1, m5
- paddw m2, m6
- packuswb m1, m2
-
- movu m5, [t1 + 2 * 16]
- movu m6, [t1 + 3 * 16]
- paddw m3, m5
- paddw m4, m6
- packuswb m3, m4
-
- ; store recon[] and recipred[]
- movu [t2], m1
- movu [t2 + 16], m3
- movu [t4], m1
- movu [t4 + 16], m3
-
- ; store recqt[]
- pmovzxbw m2, m1
- punpckhbw m1, m0
- movu [t3 + 0 * 16], m2
- movu [t3 + 1 * 16], m1
- pmovzxbw m4, m3
- punpckhbw m3, m0
- movu [t3 + 2 * 16], m4
- movu [t3 + 3 * 16], m3
-
- add t3, t6
- add t4, t7
- add t0, t5
- lea t1, [t1 + t5 * 2]
- add t2, t5
-
- dec t8d
- jnz .loop
- RET
-
-
-INIT_XMM sse4
-cglobal calcRecons64
-%if ARCH_X86_64 == 1
- DECLARE_REG_TMP 0,1,2,3,4,5,6,7,8
- PROLOGUE 6,9,7
-%else
- DECLARE_REG_TMP 0,1,2,3,4,5
- PROLOGUE 6,7,7
- %define t6 r6m
- %define t6d r6d
- %define t7 r7m
- %define t8d r6d
-%endif
-
- mov t6d, r6m
-%if ARCH_X86_64 == 0
- add t6d, t6d
- mov r6m, t6d
-%else
- mov r5d, r5m
- mov r7d, r7m
- add t6d, t6d
-%endif
-
- pxor m0, m0
- mov t8d, 64
-.loop:
- ; left 32 pixel
- movu m2, [t0 + 0 * 16]
- movu m4, [t0 + 1 * 16]
- pmovzxbw m1, m2
- punpckhbw m2, m0
- pmovzxbw m3, m4
- punpckhbw m4, m0
-
- movu m5, [t1 + 0 * 16]
- movu m6, [t1 + 1 * 16]
- paddw m1, m5
- paddw m2, m6
- packuswb m1, m2
-
- movu m5, [t1 + 2 * 16]
- movu m6, [t1 + 3 * 16]
- paddw m3, m5
- paddw m4, m6
- packuswb m3, m4
-
- ; store recon[] and recipred[]
- movu [t2 + 0 * 16], m1
- movu [t2 + 1 * 16], m3
- movu [t4 + 0 * 16], m1
- movu [t4 + 1 * 16], m3
-
- ; store recqt[]
- pmovzxbw m2, m1
- punpckhbw m1, m0
- movu [t3 + 0 * 16], m2
- movu [t3 + 1 * 16], m1
- pmovzxbw m4, m3
- punpckhbw m3, m0
- movu [t3 + 2 * 16], m4
- movu [t3 + 3 * 16], m3
-
- ; right 32 pixel
- movu m2, [t0 + 2 * 16]
- movu m4, [t0 + 3 * 16]
- pmovzxbw m1, m2
- punpckhbw m2, m0
- pmovzxbw m3, m4
- punpckhbw m4, m0
-
- movu m5, [t1 + 4 * 16]
- movu m6, [t1 + 5 * 16]
- paddw m1, m5
- paddw m2, m6
- packuswb m1, m2
-
- movu m5, [t1 + 6 * 16]
- movu m6, [t1 + 7 * 16]
- paddw m3, m5
- paddw m4, m6
- packuswb m3, m4
-
- ; store recon[] and recipred[]
- movu [t2 + 2 * 16], m1
- movu [t2 + 3 * 16], m3
- movu [t4 + 2 * 16], m1
- movu [t4 + 3 * 16], m3
-
- ; store recqt[]
- pmovzxbw m2, m1
- punpckhbw m1, m0
- movu [t3 + 4 * 16], m2
- movu [t3 + 5 * 16], m1
- pmovzxbw m4, m3
- punpckhbw m3, m0
- movu [t3 + 6 * 16], m4
- movu [t3 + 7 * 16], m3
-
- add t3, t6
- add t4, t7
- add t0, t5
- lea t1, [t1 + t5 * 2]
- add t2, t5
-
- dec t8d
- jnz .loop
- RET
+;*****************************************************************************
+;* Copyright (C) 2013 x265 project
+;*
+;* Authors: Min Chen <chenm003 at 163.com> <min.chen at multicorewareinc.com>
+;*
+;* This program is free software; you can redistribute it and/or modify
+;* it under the terms of the GNU General Public License as published by
+;* the Free Software Foundation; either version 2 of the License, or
+;* (at your option) any later version.
+;*
+;* This program is distributed in the hope that it will be useful,
+;* but WITHOUT ANY WARRANTY; without even the implied warranty of
+;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+;* GNU General Public License for more details.
+;*
+;* You should have received a copy of the GNU General Public License
+;* along with this program; if not, write to the Free Software
+;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02111, USA.
+;*
+;* This program is also available under a commercial proprietary license.
+;* For more information, contact us at licensing at multicorewareinc.com.
+;*****************************************************************************/
+
+%include "x86inc.asm"
+%include "x86util.asm"
+
+SECTION_RODATA 32
+
+SECTION .text
+
+
+;-----------------------------------------------------------------------------
+; void cvt32to16_shr(short *dst, int *src, intptr_t stride, int shift, int size)
+;-----------------------------------------------------------------------------
+INIT_XMM sse2
+cglobal cvt32to16_shr, 5, 7, 1, dst, src, stride
+%define rnd m7
+%define shift m6
+
+ ; make shift
+ mov r5d, r3m
+ movd shift, r5d
+
+ ; make round
+ dec r5
+ xor r6, r6
+ bts r6, r5
+
+ movd rnd, r6d
+ pshufd rnd, rnd, 0
+
+ ; register alloc
+ ; r0 - dst
+ ; r1 - src
+ ; r2 - stride * 2 (short*)
+ ; r3 - lx
+ ; r4 - size
+ ; r5 - ly
+ ; r6 - diff
+ lea r2, [r2 * 2]
+
+ mov r4d, r4m
+ mov r5, r4
+ mov r6, r2
+ sub r6, r4
+ lea r6, [r6 * 2]
+
+ shr r5, 1
+.loop_row:
+
+ mov r3, r4
+ shr r3, 2
+.loop_col:
+ ; row 0
+ movu m0, [r1]
+ paddd m0, rnd
+ psrad m0, shift
+ packssdw m0, m0
+ movh [r0], m0
+
+ ; row 1
+ movu m0, [r1 + r4 * 4]
+ paddd m0, rnd
+ psrad m0, shift
+ packssdw m0, m0
+ movh [r0 + r2], m0
+
+ ; move col pointer
+ add r1, 16
+ add r0, 8
+
+ dec r3
+ jg .loop_col
+
+ ; update pointer
+ lea r1, [r1 + r4 * 4]
+ add r0, r6
+
+ ; end of loop_row
+ dec r5
+ jg .loop_row
+
+ RET
+
+
+;-----------------------------------------------------------------------------
+; void calcrecon(pixel* pred, int16_t* residual, pixel* recon, int16_t* reconqt, pixel *reconipred, int stride, int strideqt, int strideipred)
+;-----------------------------------------------------------------------------
+INIT_XMM sse2
+cglobal calcRecons4
+%if ARCH_X86_64 == 1
+ DECLARE_REG_TMP 0,1,2,3,4,5,6,7,8
+ PROLOGUE 6,9,4
+%else
+ DECLARE_REG_TMP 0,1,2,3,4,5
+ PROLOGUE 6,7,4
+ %define t6 r6m
+ %define t6d r6d
+ %define t7 r7m
+ %define t8d r6d
+%endif
+
+ mov t6d, r6m
+%if ARCH_X86_64 == 0
+ add t6d, t6d
+ mov r6m, t6d
+%else
+ mov r5d, r5m
+ mov r7d, r7m
+ add t6d, t6d
+%endif
+
+ pxor m0, m0
+ mov t8d, 4/2
+.loop:
+ movd m1, [t0]
+ movd m2, [t0 + t5]
+ punpckldq m1, m2
+ punpcklbw m1, m0
+ movh m2, [t1]
+ movh m3, [t1 + t5 * 2]
+ punpcklqdq m2, m3
+ paddw m1, m2
+ packuswb m1, m1
+
+ ; store recon[] and recipred[]
+ movd [t2], m1
+ movd [t4], m1
+ add t4, t7
+ pshufd m2, m1, 1
+ movd [t2 + t5], m2
+ movd [t4], m2
+ add t4, t7
+
+ ; store recqt[]
+ punpcklbw m1, m0
+ movlps [t3], m1
+ add t3, t6
+ movhps [t3], m1
+ add t3, t6
+
+ lea t0, [t0 + t5 * 2]
+ lea t1, [t1 + t5 * 4]
+ lea t2, [t2 + t5 * 2]
+
+ dec t8d
+ jnz .loop
+ RET
+
+
+INIT_XMM sse2
+cglobal calcRecons8
+%if ARCH_X86_64 == 1
+ DECLARE_REG_TMP 0,1,2,3,4,5,6,7,8
+ PROLOGUE 6,9,5
+%else
+ DECLARE_REG_TMP 0,1,2,3,4,5
+ PROLOGUE 6,7,5
+ %define t6 r6m
+ %define t6d r6d
+ %define t7 r7m
+ %define t8d r6d
+%endif
+
+ mov t6d, r6m
+%if ARCH_X86_64 == 0
+ add t6d, t6d
+ mov r6m, t6d
+%else
+ mov r5d, r5m
+ mov r7d, r7m
+ add t6d, t6d
+%endif
+
+ pxor m0, m0
+ mov t8d, 8/2
+.loop:
+ movh m1, [t0]
+ movh m2, [t0 + t5]
+ punpcklbw m1, m0
+ punpcklbw m2, m0
+ movu m3, [t1]
+ movu m4, [t1 + t5 * 2]
+ paddw m1, m3
+ paddw m2, m4
+ packuswb m1, m2
+
+ ; store recon[] and recipred[]
+ movlps [t2], m1
+ movhps [t2 + t5], m1
+ movlps [t4], m1
+%if ARCH_X86_64 == 0
+ add t4, t7
+ movhps [t4], m1
+ add t4, t7
+%else
+ movhps [t4 + t7], m1
+ lea t4, [t4 + t7 * 2]
+%endif
+
+ ; store recqt[]
+ punpcklbw m2, m1, m0
+ punpckhbw m1, m0
+ movu [t3], m2
+ add t3, t6
+ movu [t3], m1
+ add t3, t6
+
+ lea t0, [t0 + t5 * 2]
+ lea t1, [t1 + t5 * 4]
+ lea t2, [t2 + t5 * 2]
+
+ dec t8d
+ jnz .loop
+ RET
+
+
+INIT_XMM sse4
+cglobal calcRecons16
+%if ARCH_X86_64 == 1
+ DECLARE_REG_TMP 0,1,2,3,4,5,6,7,8
+ PROLOGUE 6,9,3
+%else
+ DECLARE_REG_TMP 0,1,2,3,4,5
+ PROLOGUE 6,7,3
+ %define t6 r6m
+ %define t6d r6d
+ %define t7 r7m
+ %define t8d r6d
+%endif
+
+ mov t6d, r6m
+%if ARCH_X86_64 == 0
+ add t6d, t6d
+ mov r6m, t6d
+%else
+ mov r5d, r5m
+ mov r7d, r7m
+ add t6d, t6d
+%endif
+
+ pxor m0, m0
+ mov t8d, 16
+.loop:
+ movu m2, [t0]
+ pmovzxbw m1, m2
+ punpckhbw m2, m0
+ paddw m1, [t1]
+ paddw m2, [t1 + 16]
+ packuswb m1, m2
+
+ ; store recon[] and recipred[]
+ movu [t2], m1
+ movu [t4], m1
+
+ ; store recqt[]
+ pmovzxbw m2, m1
+ punpckhbw m1, m0
+ movu [t3], m2
+ movu [t3 + 16], m1
+
+ add t3, t6
+ add t4, t7
+ add t0, t5
+ lea t1, [t1 + t5 * 2]
+ add t2, t5
+
+ dec t8d
+ jnz .loop
+ RET
+
+
+INIT_XMM sse4
+cglobal calcRecons32
+%if ARCH_X86_64 == 1
+ DECLARE_REG_TMP 0,1,2,3,4,5,6,7,8
+ PROLOGUE 6,9,5
+%else
+ DECLARE_REG_TMP 0,1,2,3,4,5
+ PROLOGUE 6,7,5
+ %define t6 r6m
+ %define t6d r6d
+ %define t7 r7m
+ %define t8d r6d
+%endif
+
+ mov t6d, r6m
+%if ARCH_X86_64 == 0
+ add t6d, t6d
+ mov r6m, t6d
+%else
+ mov r5d, r5m
+ mov r7d, r7m
+ add t6d, t6d
+%endif
+
+ pxor m0, m0
+ mov t8d, 32
+.loop:
+ movu m2, [t0]
+ movu m4, [t0 + 16]
+ pmovzxbw m1, m2
+ punpckhbw m2, m0
+ pmovzxbw m3, m4
+ punpckhbw m4, m0
+
+ paddw m1, [t1 + 0 * 16]
+ paddw m2, [t1 + 1 * 16]
+ packuswb m1, m2
+
+ paddw m3, [t1 + 2 * 16]
+ paddw m4, [t1 + 3 * 16]
+ packuswb m3, m4
+
+ ; store recon[] and recipred[]
+ movu [t2], m1
+ movu [t2 + 16], m3
+ movu [t4], m1
+ movu [t4 + 16], m3
+
+ ; store recqt[]
+ pmovzxbw m2, m1
+ punpckhbw m1, m0
+ movu [t3 + 0 * 16], m2
+ movu [t3 + 1 * 16], m1
+ pmovzxbw m4, m3
+ punpckhbw m3, m0
+ movu [t3 + 2 * 16], m4
+ movu [t3 + 3 * 16], m3
+
+ add t3, t6
+ add t4, t7
+ add t0, t5
+ lea t1, [t1 + t5 * 2]
+ add t2, t5
+
+ dec t8d
+ jnz .loop
+ RET
+
+
+INIT_XMM sse4
+cglobal calcRecons64
+%if ARCH_X86_64 == 1
+ DECLARE_REG_TMP 0,1,2,3,4,5,6,7,8
+ PROLOGUE 6,9,7
+%else
+ DECLARE_REG_TMP 0,1,2,3,4,5
+ PROLOGUE 6,7,7
+ %define t6 r6m
+ %define t6d r6d
+ %define t7 r7m
+ %define t8d r6d
+%endif
+
+ mov t6d, r6m
+%if ARCH_X86_64 == 0
+ add t6d, t6d
+ mov r6m, t6d
+%else
+ mov r5d, r5m
+ mov r7d, r7m
+ add t6d, t6d
+%endif
+
+ pxor m0, m0
+ mov t8d, 64
+.loop:
+ ; left 32 pixel
+ movu m2, [t0 + 0 * 16]
+ movu m4, [t0 + 1 * 16]
+ pmovzxbw m1, m2
+ punpckhbw m2, m0
+ pmovzxbw m3, m4
+ punpckhbw m4, m0
+
+ movu m5, [t1 + 0 * 16]
+ movu m6, [t1 + 1 * 16]
+ paddw m1, m5
+ paddw m2, m6
+ packuswb m1, m2
+
+ movu m5, [t1 + 2 * 16]
+ movu m6, [t1 + 3 * 16]
+ paddw m3, m5
+ paddw m4, m6
+ packuswb m3, m4
+
+ ; store recon[] and recipred[]
+ movu [t2 + 0 * 16], m1
+ movu [t2 + 1 * 16], m3
+ movu [t4 + 0 * 16], m1
+ movu [t4 + 1 * 16], m3
+
+ ; store recqt[]
+ pmovzxbw m2, m1
+ punpckhbw m1, m0
+ movu [t3 + 0 * 16], m2
+ movu [t3 + 1 * 16], m1
+ pmovzxbw m4, m3
+ punpckhbw m3, m0
+ movu [t3 + 2 * 16], m4
+ movu [t3 + 3 * 16], m3
+
+ ; right 32 pixel
+ movu m2, [t0 + 2 * 16]
+ movu m4, [t0 + 3 * 16]
+ pmovzxbw m1, m2
+ punpckhbw m2, m0
+ pmovzxbw m3, m4
+ punpckhbw m4, m0
+
+ movu m5, [t1 + 4 * 16]
+ movu m6, [t1 + 5 * 16]
+ paddw m1, m5
+ paddw m2, m6
+ packuswb m1, m2
+
+ movu m5, [t1 + 6 * 16]
+ movu m6, [t1 + 7 * 16]
+ paddw m3, m5
+ paddw m4, m6
+ packuswb m3, m4
+
+ ; store recon[] and recipred[]
+ movu [t2 + 2 * 16], m1
+ movu [t2 + 3 * 16], m3
+ movu [t4 + 2 * 16], m1
+ movu [t4 + 3 * 16], m3
+
+ ; store recqt[]
+ pmovzxbw m2, m1
+ punpckhbw m1, m0
+ movu [t3 + 4 * 16], m2
+ movu [t3 + 5 * 16], m1
+ pmovzxbw m4, m3
+ punpckhbw m3, m0
+ movu [t3 + 6 * 16], m4
+ movu [t3 + 7 * 16], m3
+
+ add t3, t6
+ add t4, t7
+ add t0, t5
+ lea t1, [t1 + t5 * 2]
+ add t2, t5
+
+ dec t8d
+ jnz .loop
+ RET
_______________________________________________
x265-devel mailing list
x265-devel at videolan.org
https://mailman.videolan.org/listinfo/x265-devel
--
Steve Borho
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.videolan.org/pipermail/x265-devel/attachments/20131116/f6789aec/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: 0003-asm-assembly-code-for-calcresidual.patch
Type: application/octet-stream
Size: 9500 bytes
Desc: not available
URL: <http://mailman.videolan.org/pipermail/x265-devel/attachments/20131116/f6789aec/attachment-0003.obj>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: 0001-asm-residual-buffer-is-alignment-to-size-so-we-can-u.patch
Type: application/octet-stream
Size: 3083 bytes
Desc: not available
URL: <http://mailman.videolan.org/pipermail/x265-devel/attachments/20131116/f6789aec/attachment-0004.obj>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: 0002-cleanup-remove-unused-64x64-functions.patch
Type: application/octet-stream
Size: 7505 bytes
Desc: not available
URL: <http://mailman.videolan.org/pipermail/x265-devel/attachments/20131116/f6789aec/attachment-0005.obj>
More information about the x265-devel
mailing list