123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548 |
- ;******************************************************************************
- ;* AAC Spectral Band Replication decoding functions
- ;* Copyright (C) 2012 Christophe Gisquet <christophe.gisquet@gmail.com>
- ;*
- ;* This file is part of FFmpeg.
- ;*
- ;* FFmpeg is free software; you can redistribute it and/or
- ;* modify it under the terms of the GNU Lesser General Public
- ;* License as published by the Free Software Foundation; either
- ;* version 2.1 of the License, or (at your option) any later version.
- ;*
- ;* FFmpeg is distributed in the hope that it will be useful,
- ;* but WITHOUT ANY WARRANTY; without even the implied warranty of
- ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- ;* Lesser General Public License for more details.
- ;*
- ;* You should have received a copy of the GNU Lesser General Public
- ;* License along with FFmpeg; if not, write to the Free Software
- ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
- ;******************************************************************************
- %include "libavutil/x86/x86util.asm"
- SECTION_RODATA
- ; mask equivalent for multiply by -1.0 1.0
- ps_mask times 2 dd 1<<31, 0
- ps_mask2 times 2 dd 0, 1<<31
- ps_mask3 dd 0, 0, 0, 1<<31
- ps_noise0 times 2 dd 1.0, 0.0,
- ps_noise2 times 2 dd -1.0, 0.0
- ps_noise13 dd 0.0, 1.0, 0.0, -1.0
- dd 0.0, -1.0, 0.0, 1.0
- dd 0.0, 1.0, 0.0, -1.0
- cextern sbr_noise_table
- cextern ps_neg
- SECTION .text
- INIT_XMM sse
- cglobal sbr_sum_square, 2, 3, 6
- mov r2d, r1d
- xorps m0, m0
- xorps m1, m1
- sar r2, 3
- jz .prepare
- .loop:
- movu m2, [r0 + 0]
- movu m3, [r0 + 16]
- movu m4, [r0 + 32]
- movu m5, [r0 + 48]
- mulps m2, m2
- mulps m3, m3
- mulps m4, m4
- mulps m5, m5
- addps m0, m2
- addps m1, m3
- addps m0, m4
- addps m1, m5
- add r0, 64
- dec r2
- jnz .loop
- .prepare:
- and r1, 7
- sar r1, 1
- jz .end
- ; len is a multiple of 2, thus there are at least 4 elements to process
- .endloop:
- movu m2, [r0]
- add r0, 16
- mulps m2, m2
- dec r1
- addps m0, m2
- jnz .endloop
- .end:
- addps m0, m1
- movhlps m2, m0
- addps m0, m2
- movss m1, m0
- shufps m0, m0, 1
- addss m0, m1
- %if ARCH_X86_64 == 0
- movss r0m, m0
- fld dword r0m
- %endif
- RET
- %define STEP 40*4*2
- cglobal sbr_hf_g_filt, 5, 6, 5
- lea r1, [r1 + 8*r4] ; offset by ixh elements into X_high
- mov r5, r3
- and r3, 0xFC
- lea r2, [r2 + r3*4]
- lea r0, [r0 + r3*8]
- neg r3
- jz .loop1
- .loop4:
- movlps m0, [r2 + 4*r3 + 0]
- movlps m1, [r2 + 4*r3 + 8]
- movlps m2, [r1 + 0*STEP]
- movlps m3, [r1 + 2*STEP]
- movhps m2, [r1 + 1*STEP]
- movhps m3, [r1 + 3*STEP]
- unpcklps m0, m0
- unpcklps m1, m1
- mulps m0, m2
- mulps m1, m3
- movu [r0 + 8*r3 + 0], m0
- movu [r0 + 8*r3 + 16], m1
- add r1, 4*STEP
- add r3, 4
- jnz .loop4
- and r5, 3 ; number of single element loops
- jz .end
- .loop1: ; element 0 and 1 can be computed at the same time
- movss m0, [r2]
- movlps m2, [r1]
- unpcklps m0, m0
- mulps m2, m0
- movlps [r0], m2
- add r0, 8
- add r2, 4
- add r1, STEP
- dec r5
- jnz .loop1
- .end:
- RET
- ; void ff_sbr_hf_gen_sse(float (*X_high)[2], const float (*X_low)[2],
- ; const float alpha0[2], const float alpha1[2],
- ; float bw, int start, int end)
- ;
- cglobal sbr_hf_gen, 4,4,8, X_high, X_low, alpha0, alpha1, BW, S, E
- ; load alpha factors
- %define bw m0
- %if ARCH_X86_64 == 0 || WIN64
- movss bw, BWm
- %endif
- movlps m2, [alpha1q]
- movlps m1, [alpha0q]
- shufps bw, bw, 0
- mulps m2, bw ; (a1[0] a1[1])*bw
- mulps m1, bw ; (a0[0] a0[1])*bw = (a2 a3)
- mulps m2, bw ; (a1[0] a1[1])*bw*bw = (a0 a1)
- mova m3, m1
- mova m4, m2
- ; Set pointers
- %if ARCH_X86_64 == 0 || WIN64
- ; start and end 6th and 7th args on stack
- mov r2d, Sm
- mov r3d, Em
- DEFINE_ARGS X_high, X_low, start, end
- %else
- ; BW does not actually occupy a register, so shift by 1
- DEFINE_ARGS X_high, X_low, alpha0, alpha1, start, end
- movsxd startq, startd
- movsxd endq, endd
- %endif
- sub startq, endq ; neg num of loops
- lea X_highq, [X_highq + endq*2*4]
- lea X_lowq, [X_lowq + endq*2*4 - 2*2*4]
- shl startq, 3 ; offset from num loops
- mova m0, [X_lowq + startq]
- shufps m3, m3, q1111
- shufps m4, m4, q1111
- xorps m3, [ps_mask]
- shufps m1, m1, q0000
- shufps m2, m2, q0000
- xorps m4, [ps_mask]
- .loop2:
- movu m7, [X_lowq + startq + 8] ; BbCc
- mova m6, m0
- mova m5, m7
- shufps m0, m0, q2301 ; aAbB
- shufps m7, m7, q2301 ; bBcC
- mulps m0, m4
- mulps m7, m3
- mulps m6, m2
- mulps m5, m1
- addps m7, m0
- mova m0, [X_lowq + startq + 16] ; CcDd
- addps m7, m0
- addps m6, m5
- addps m7, m6
- mova [X_highq + startq], m7
- add startq, 16
- jnz .loop2
- RET
- cglobal sbr_sum64x5, 1,2,4,z
- lea r1q, [zq+ 256]
- .loop:
- mova m0, [zq+ 0]
- mova m2, [zq+ 16]
- mova m1, [zq+ 256]
- mova m3, [zq+ 272]
- addps m0, [zq+ 512]
- addps m2, [zq+ 528]
- addps m1, [zq+ 768]
- addps m3, [zq+ 784]
- addps m0, [zq+1024]
- addps m2, [zq+1040]
- addps m0, m1
- addps m2, m3
- mova [zq], m0
- mova [zq+16], m2
- add zq, 32
- cmp zq, r1q
- jne .loop
- REP_RET
- INIT_XMM sse
- cglobal sbr_qmf_post_shuffle, 2,3,4,W,z
- lea r2q, [zq + (64-4)*4]
- mova m3, [ps_neg]
- .loop:
- mova m1, [zq]
- xorps m0, m3, [r2q]
- shufps m0, m0, m0, q0123
- unpcklps m2, m0, m1
- unpckhps m0, m0, m1
- mova [Wq + 0], m2
- mova [Wq + 16], m0
- add Wq, 32
- sub r2q, 16
- add zq, 16
- cmp zq, r2q
- jl .loop
- REP_RET
- INIT_XMM sse
- cglobal sbr_neg_odd_64, 1,2,4,z
- lea r1q, [zq+256]
- .loop:
- mova m0, [zq+ 0]
- mova m1, [zq+16]
- mova m2, [zq+32]
- mova m3, [zq+48]
- xorps m0, [ps_mask2]
- xorps m1, [ps_mask2]
- xorps m2, [ps_mask2]
- xorps m3, [ps_mask2]
- mova [zq+ 0], m0
- mova [zq+16], m1
- mova [zq+32], m2
- mova [zq+48], m3
- add zq, 64
- cmp zq, r1q
- jne .loop
- REP_RET
- ; void ff_sbr_qmf_deint_bfly_sse2(float *v, const float *src0, const float *src1)
- %macro SBR_QMF_DEINT_BFLY 0
- cglobal sbr_qmf_deint_bfly, 3,5,8, v,src0,src1,vrev,c
- mov cq, 64*4-2*mmsize
- lea vrevq, [vq + 64*4]
- .loop:
- mova m0, [src0q+cq]
- mova m1, [src1q]
- mova m4, [src0q+cq+mmsize]
- mova m5, [src1q+mmsize]
- %if cpuflag(sse2)
- pshufd m2, m0, q0123
- pshufd m3, m1, q0123
- pshufd m6, m4, q0123
- pshufd m7, m5, q0123
- %else
- shufps m2, m0, m0, q0123
- shufps m3, m1, m1, q0123
- shufps m6, m4, m4, q0123
- shufps m7, m5, m5, q0123
- %endif
- addps m5, m2
- subps m0, m7
- addps m1, m6
- subps m4, m3
- mova [vrevq], m1
- mova [vrevq+mmsize], m5
- mova [vq+cq], m0
- mova [vq+cq+mmsize], m4
- add src1q, 2*mmsize
- add vrevq, 2*mmsize
- sub cq, 2*mmsize
- jge .loop
- REP_RET
- %endmacro
- INIT_XMM sse
- SBR_QMF_DEINT_BFLY
- INIT_XMM sse2
- SBR_QMF_DEINT_BFLY
- INIT_XMM sse2
- cglobal sbr_qmf_pre_shuffle, 1,4,6,z
- %define OFFSET (32*4-2*mmsize)
- mov r3q, OFFSET
- lea r1q, [zq + (32+1)*4]
- lea r2q, [zq + 64*4]
- mova m5, [ps_neg]
- .loop:
- movu m0, [r1q]
- movu m2, [r1q + mmsize]
- movu m1, [zq + r3q + 4 + mmsize]
- movu m3, [zq + r3q + 4]
- pxor m2, m5
- pxor m0, m5
- pshufd m2, m2, q0123
- pshufd m0, m0, q0123
- SBUTTERFLY dq, 2, 3, 4
- SBUTTERFLY dq, 0, 1, 4
- mova [r2q + 2*r3q + 0*mmsize], m2
- mova [r2q + 2*r3q + 1*mmsize], m3
- mova [r2q + 2*r3q + 2*mmsize], m0
- mova [r2q + 2*r3q + 3*mmsize], m1
- add r1q, 2*mmsize
- sub r3q, 2*mmsize
- jge .loop
- movq m2, [zq]
- movq [r2q], m2
- REP_RET
- %ifdef PIC
- %define NREGS 1
- %if UNIX64
- %define NOISE_TABLE r6q ; r5q is m_max
- %else
- %define NOISE_TABLE r5q
- %endif
- %else
- %define NREGS 0
- %define NOISE_TABLE sbr_noise_table
- %endif
- %macro LOAD_NST 1
- %ifdef PIC
- lea NOISE_TABLE, [%1]
- mova m0, [kxq + NOISE_TABLE]
- %else
- mova m0, [kxq + %1]
- %endif
- %endmacro
- INIT_XMM sse2
- ; sbr_hf_apply_noise_0(float (*Y)[2], const float *s_m,
- ; const float *q_filt, int noise,
- ; int kx, int m_max)
- cglobal sbr_hf_apply_noise_0, 5,5+NREGS+UNIX64,8, Y,s_m,q_filt,noise,kx,m_max
- mova m0, [ps_noise0]
- jmp apply_noise_main
- ; sbr_hf_apply_noise_1(float (*Y)[2], const float *s_m,
- ; const float *q_filt, int noise,
- ; int kx, int m_max)
- cglobal sbr_hf_apply_noise_1, 5,5+NREGS+UNIX64,8, Y,s_m,q_filt,noise,kx,m_max
- and kxq, 1
- shl kxq, 4
- LOAD_NST ps_noise13
- jmp apply_noise_main
- ; sbr_hf_apply_noise_2(float (*Y)[2], const float *s_m,
- ; const float *q_filt, int noise,
- ; int kx, int m_max)
- cglobal sbr_hf_apply_noise_2, 5,5+NREGS+UNIX64,8, Y,s_m,q_filt,noise,kx,m_max
- mova m0, [ps_noise2]
- jmp apply_noise_main
- ; sbr_hf_apply_noise_3(float (*Y)[2], const float *s_m,
- ; const float *q_filt, int noise,
- ; int kx, int m_max)
- cglobal sbr_hf_apply_noise_3, 5,5+NREGS+UNIX64,8, Y,s_m,q_filt,noise,kx,m_max
- and kxq, 1
- shl kxq, 4
- LOAD_NST ps_noise13+16
- apply_noise_main:
- %if ARCH_X86_64 == 0 || WIN64
- mov kxd, m_maxm
- DEFINE_ARGS Y, s_m, q_filt, noise, count
- %else
- DEFINE_ARGS Y, s_m, q_filt, noise, kx, count
- %endif
- movsxdifnidn noiseq, noised
- dec noiseq
- shl countd, 2
- %ifdef PIC
- lea NOISE_TABLE, [sbr_noise_table]
- %endif
- lea Yq, [Yq + 2*countq]
- add s_mq, countq
- add q_filtq, countq
- shl noiseq, 3
- pxor m5, m5
- neg countq
- .loop:
- mova m1, [q_filtq + countq]
- movu m3, [noiseq + NOISE_TABLE + 1*mmsize]
- movu m4, [noiseq + NOISE_TABLE + 2*mmsize]
- add noiseq, 2*mmsize
- and noiseq, 0x1ff<<3
- punpckhdq m2, m1, m1
- punpckldq m1, m1
- mulps m1, m3 ; m2 = q_filt[m] * ff_sbr_noise_table[noise]
- mulps m2, m4 ; m2 = q_filt[m] * ff_sbr_noise_table[noise]
- mova m3, [s_mq + countq]
- ; TODO: replace by a vpermd in AVX2
- punpckhdq m4, m3, m3
- punpckldq m3, m3
- pcmpeqd m6, m3, m5 ; m6 == 0
- pcmpeqd m7, m4, m5 ; m7 == 0
- mulps m3, m0 ; s_m[m] * phi_sign
- mulps m4, m0 ; s_m[m] * phi_sign
- pand m1, m6
- pand m2, m7
- movu m6, [Yq + 2*countq]
- movu m7, [Yq + 2*countq + mmsize]
- addps m3, m1
- addps m4, m2
- addps m6, m3
- addps m7, m4
- movu [Yq + 2*countq], m6
- movu [Yq + 2*countq + mmsize], m7
- add countq, mmsize
- jl .loop
- RET
- INIT_XMM sse
- cglobal sbr_qmf_deint_neg, 2,4,4,v,src,vrev,c
- %define COUNT 32*4
- %define OFFSET 32*4
- mov cq, -COUNT
- lea vrevq, [vq + OFFSET + COUNT]
- add vq, OFFSET-mmsize
- add srcq, 2*COUNT
- mova m3, [ps_neg]
- .loop:
- mova m0, [srcq + 2*cq + 0*mmsize]
- mova m1, [srcq + 2*cq + 1*mmsize]
- shufps m2, m0, m1, q2020
- shufps m1, m0, q1313
- xorps m2, m3
- mova [vq], m1
- mova [vrevq + cq], m2
- sub vq, mmsize
- add cq, mmsize
- jl .loop
- REP_RET
- %macro SBR_AUTOCORRELATE 0
- cglobal sbr_autocorrelate, 2,3,8,32, x, phi, cnt
- mov cntq, 37*8
- add xq, cntq
- neg cntq
- %if cpuflag(sse3)
- %define MOVH movsd
- movddup m5, [xq+cntq]
- %else
- %define MOVH movlps
- movlps m5, [xq+cntq]
- movlhps m5, m5
- %endif
- MOVH m7, [xq+cntq+8 ]
- MOVH m1, [xq+cntq+16]
- shufps m7, m7, q0110
- shufps m1, m1, q0110
- mulps m3, m5, m7 ; x[0][0] * x[1][0], x[0][1] * x[1][1], x[0][0] * x[1][1], x[0][1] * x[1][0]
- mulps m4, m5, m5 ; x[0][0] * x[0][0], x[0][1] * x[0][1];
- mulps m5, m1 ; real_sum2 = x[0][0] * x[2][0], x[0][1] * x[2][1]; imag_sum2 = x[0][0] * x[2][1], x[0][1] * x[2][0]
- movaps [rsp ], m3
- movaps [rsp+16], m4
- add cntq, 8
- MOVH m2, [xq+cntq+16]
- movlhps m7, m7
- shufps m2, m2, q0110
- mulps m6, m7, m1 ; real_sum1 = x[1][0] * x[2][0], x[1][1] * x[2][1]; imag_sum1 += x[1][0] * x[2][1], x[1][1] * x[2][0]
- mulps m4, m7, m2
- mulps m7, m7 ; real_sum0 = x[1][0] * x[1][0], x[1][1] * x[1][1];
- addps m5, m4 ; real_sum2 += x[1][0] * x[3][0], x[1][1] * x[3][1]; imag_sum2 += x[1][0] * x[3][1], x[1][1] * x[3][0]
- align 16
- .loop:
- add cntq, 8
- MOVH m0, [xq+cntq+16]
- movlhps m1, m1
- shufps m0, m0, q0110
- mulps m3, m1, m2
- mulps m4, m1, m0
- mulps m1, m1
- addps m6, m3 ; real_sum1 += x[i][0] * x[i + 1][0], x[i][1] * x[i + 1][1]; imag_sum1 += x[i][0] * x[i + 1][1], x[i][1] * x[i + 1][0];
- addps m5, m4 ; real_sum2 += x[i][0] * x[i + 2][0], x[i][1] * x[i + 2][1]; imag_sum2 += x[i][0] * x[i + 2][1], x[i][1] * x[i + 2][0];
- addps m7, m1 ; real_sum0 += x[i][0] * x[i][0], x[i][1] * x[i][1];
- add cntq, 8
- MOVH m1, [xq+cntq+16]
- movlhps m2, m2
- shufps m1, m1, q0110
- mulps m3, m2, m0
- mulps m4, m2, m1
- mulps m2, m2
- addps m6, m3 ; real_sum1 += x[i][0] * x[i + 1][0], x[i][1] * x[i + 1][1]; imag_sum1 += x[i][0] * x[i + 1][1], x[i][1] * x[i + 1][0];
- addps m5, m4 ; real_sum2 += x[i][0] * x[i + 2][0], x[i][1] * x[i + 2][1]; imag_sum2 += x[i][0] * x[i + 2][1], x[i][1] * x[i + 2][0];
- addps m7, m2 ; real_sum0 += x[i][0] * x[i][0], x[i][1] * x[i][1];
- add cntq, 8
- MOVH m2, [xq+cntq+16]
- movlhps m0, m0
- shufps m2, m2, q0110
- mulps m3, m0, m1
- mulps m4, m0, m2
- mulps m0, m0
- addps m6, m3 ; real_sum1 += x[i][0] * x[i + 1][0], x[i][1] * x[i + 1][1]; imag_sum1 += x[i][0] * x[i + 1][1], x[i][1] * x[i + 1][0];
- addps m5, m4 ; real_sum2 += x[i][0] * x[i + 2][0], x[i][1] * x[i + 2][1]; imag_sum2 += x[i][0] * x[i + 2][1], x[i][1] * x[i + 2][0];
- addps m7, m0 ; real_sum0 += x[i][0] * x[i][0], x[i][1] * x[i][1];
- jl .loop
- movlhps m1, m1
- mulps m2, m1
- mulps m1, m1
- addps m2, m6 ; real_sum1 + x[38][0] * x[39][0], x[38][1] * x[39][1]; imag_sum1 + x[38][0] * x[39][1], x[38][1] * x[39][0];
- addps m1, m7 ; real_sum0 + x[38][0] * x[38][0], x[38][1] * x[38][1];
- addps m6, [rsp ] ; real_sum1 + x[ 0][0] * x[ 1][0], x[ 0][1] * x[ 1][1]; imag_sum1 + x[ 0][0] * x[ 1][1], x[ 0][1] * x[ 1][0];
- addps m7, [rsp+16] ; real_sum0 + x[ 0][0] * x[ 0][0], x[ 0][1] * x[ 0][1];
- xorps m2, [ps_mask3]
- xorps m5, [ps_mask3]
- xorps m6, [ps_mask3]
- HADDPS m2, m5, m3
- HADDPS m7, m6, m4
- %if cpuflag(sse3)
- movshdup m0, m1
- %else
- movss m0, m1
- shufps m1, m1, q0001
- %endif
- addss m1, m0
- movaps [phiq ], m2
- movhps [phiq+0x18], m7
- movss [phiq+0x28], m7
- movss [phiq+0x10], m1
- RET
- %endmacro
- INIT_XMM sse
- SBR_AUTOCORRELATE
- INIT_XMM sse3
- SBR_AUTOCORRELATE
|