1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798 |
- ;
- ; Copyright (c) 2010 The WebM project authors. All Rights Reserved.
- ;
- ; Use of this source code is governed by a BSD-style license
- ; that can be found in the LICENSE file in the root of the source
- ; tree. An additional intellectual property rights grant can be found
- ; in the file PATENTS. All contributing project authors may
- ; be found in the AUTHORS file in the root of the source tree.
- ;
- %define private_prefix vp9
- %include "third_party/x86inc/x86inc.asm"
- SECTION .text
- ALIGN 16
- ;
- ; int64_t vp9_highbd_block_error_8bit(int32_t *coeff, int32_t *dqcoeff,
- ; intptr_t block_size, int64_t *ssz)
- ;
- INIT_XMM sse2
- cglobal highbd_block_error_8bit, 3, 3, 8, uqc, dqc, size, ssz
- pxor m4, m4 ; sse accumulator
- pxor m6, m6 ; ssz accumulator
- pxor m5, m5 ; dedicated zero register
- lea uqcq, [uqcq+sizeq*4]
- lea dqcq, [dqcq+sizeq*4]
- neg sizeq
- ALIGN 16
- .loop:
- mova m0, [dqcq+sizeq*4]
- packssdw m0, [dqcq+sizeq*4+mmsize]
- mova m2, [uqcq+sizeq*4]
- packssdw m2, [uqcq+sizeq*4+mmsize]
- mova m1, [dqcq+sizeq*4+mmsize*2]
- packssdw m1, [dqcq+sizeq*4+mmsize*3]
- mova m3, [uqcq+sizeq*4+mmsize*2]
- packssdw m3, [uqcq+sizeq*4+mmsize*3]
- add sizeq, mmsize
- ; individual errors are max. 15bit+sign, so squares are 30bit, and
- ; thus the sum of 2 should fit in a 31bit integer (+ unused sign bit)
- psubw m0, m2
- pmaddwd m2, m2
- pmaddwd m0, m0
- psubw m1, m3
- pmaddwd m3, m3
- pmaddwd m1, m1
- ; accumulate in 64bit
- punpckldq m7, m0, m5
- punpckhdq m0, m5
- paddq m4, m7
- punpckldq m7, m2, m5
- punpckhdq m2, m5
- paddq m6, m7
- punpckldq m7, m1, m5
- punpckhdq m1, m5
- paddq m4, m7
- punpckldq m7, m3, m5
- punpckhdq m3, m5
- paddq m6, m7
- paddq m4, m0
- paddq m4, m1
- paddq m6, m2
- paddq m6, m3
- jnz .loop
- ; accumulate horizontally and store in return value
- movhlps m5, m4
- movhlps m7, m6
- paddq m4, m5
- paddq m6, m7
- %if ARCH_X86_64
- movq rax, m4
- movq [sszq], m6
- %else
- mov eax, sszm
- pshufd m5, m4, 0x1
- movq [eax], m6
- movd eax, m4
- movd edx, m5
- %endif
- RET
|