123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290 |
- ;******************************************************************************
- ;* linear least squares model
- ;*
- ;* Copyright (c) 2013 Loren Merritt
- ;*
- ;* This file is part of FFmpeg.
- ;*
- ;* FFmpeg is free software; you can redistribute it and/or
- ;* modify it under the terms of the GNU Lesser General Public
- ;* License as published by the Free Software Foundation; either
- ;* version 2.1 of the License, or (at your option) any later version.
- ;*
- ;* FFmpeg is distributed in the hope that it will be useful,
- ;* but WITHOUT ANY WARRANTY; without even the implied warranty of
- ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- ;* Lesser General Public License for more details.
- ;*
- ;* You should have received a copy of the GNU Lesser General Public
- ;* License along with FFmpeg; if not, write to the Free Software
- ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
- ;******************************************************************************
- %include "x86util.asm"
- SECTION .text
- %define MAX_VARS 32
- %define MAX_VARS_ALIGN (MAX_VARS+4)
- %define COVAR_STRIDE MAX_VARS_ALIGN*8
- %define COVAR(x,y) [covarq + (x)*8 + (y)*COVAR_STRIDE]
- struc LLSModel
- .covariance: resq MAX_VARS_ALIGN*MAX_VARS_ALIGN
- .coeff: resq MAX_VARS*MAX_VARS
- .variance: resq MAX_VARS
- .indep_count: resd 1
- endstruc
- %macro ADDPD_MEM 2
- %if cpuflag(avx)
- vaddpd %2, %2, %1
- %else
- addpd %2, %1
- %endif
- mova %1, %2
- %endmacro
- INIT_XMM sse2
- %define movdqa movaps
- cglobal update_lls, 2,5,8, ctx, var, i, j, covar2
- %define covarq ctxq
- mov id, [ctxq + LLSModel.indep_count]
- lea varq, [varq + iq*8]
- neg iq
- mov covar2q, covarq
- .loopi:
- ; Compute all 3 pairwise products of a 2x2 block that lies on the diagonal
- mova m1, [varq + iq*8]
- mova m3, [varq + iq*8 + 16]
- pshufd m4, m1, q1010
- pshufd m5, m1, q3232
- pshufd m6, m3, q1010
- pshufd m7, m3, q3232
- mulpd m0, m1, m4
- mulpd m1, m1, m5
- lea covarq, [covar2q + 16]
- ADDPD_MEM COVAR(-2,0), m0
- ADDPD_MEM COVAR(-2,1), m1
- lea jq, [iq + 2]
- cmp jd, -2
- jg .skip4x4
- .loop4x4:
- ; Compute all 16 pairwise products of a 4x4 block
- mulpd m0, m4, m3
- mulpd m1, m5, m3
- mulpd m2, m6, m3
- mulpd m3, m3, m7
- ADDPD_MEM COVAR(0,0), m0
- ADDPD_MEM COVAR(0,1), m1
- ADDPD_MEM COVAR(0,2), m2
- ADDPD_MEM COVAR(0,3), m3
- mova m3, [varq + jq*8 + 16]
- mulpd m0, m4, m3
- mulpd m1, m5, m3
- mulpd m2, m6, m3
- mulpd m3, m3, m7
- ADDPD_MEM COVAR(2,0), m0
- ADDPD_MEM COVAR(2,1), m1
- ADDPD_MEM COVAR(2,2), m2
- ADDPD_MEM COVAR(2,3), m3
- mova m3, [varq + jq*8 + 32]
- add covarq, 32
- add jq, 4
- cmp jd, -2
- jle .loop4x4
- .skip4x4:
- test jd, jd
- jg .skip2x4
- mulpd m4, m3
- mulpd m5, m3
- mulpd m6, m3
- mulpd m7, m3
- ADDPD_MEM COVAR(0,0), m4
- ADDPD_MEM COVAR(0,1), m5
- ADDPD_MEM COVAR(0,2), m6
- ADDPD_MEM COVAR(0,3), m7
- .skip2x4:
- add iq, 4
- add covar2q, 4*COVAR_STRIDE+32
- cmp id, -2
- jle .loopi
- test id, id
- jg .ret
- mov jq, iq
- %define covarq covar2q
- .loop2x1:
- movsd m0, [varq + iq*8]
- movlhps m0, m0
- mulpd m0, [varq + jq*8]
- ADDPD_MEM COVAR(0,0), m0
- inc iq
- add covarq, COVAR_STRIDE
- test id, id
- jle .loop2x1
- .ret:
- REP_RET
- %macro UPDATE_LLS 0
- cglobal update_lls, 3,6,8, ctx, var, count, i, j, count2
- %define covarq ctxq
- mov countd, [ctxq + LLSModel.indep_count]
- lea count2d, [countq-2]
- xor id, id
- .loopi:
- ; Compute all 10 pairwise products of a 4x4 block that lies on the diagonal
- mova ymm1, [varq + iq*8]
- vbroadcastsd ymm4, [varq + iq*8]
- vbroadcastsd ymm5, [varq + iq*8 + 8]
- vbroadcastsd ymm6, [varq + iq*8 + 16]
- vbroadcastsd ymm7, [varq + iq*8 + 24]
- vextractf128 xmm3, ymm1, 1
- %if cpuflag(fma3)
- mova ymm0, COVAR(iq ,0)
- mova xmm2, COVAR(iq+2,2)
- fmaddpd ymm0, ymm1, ymm4, ymm0
- fmaddpd xmm2, xmm3, xmm6, xmm2
- fmaddpd ymm1, ymm5, ymm1, COVAR(iq ,1)
- fmaddpd xmm3, xmm7, xmm3, COVAR(iq+2,3)
- mova COVAR(iq ,0), ymm0
- mova COVAR(iq ,1), ymm1
- mova COVAR(iq+2,2), xmm2
- mova COVAR(iq+2,3), xmm3
- %else
- vmulpd ymm0, ymm1, ymm4
- vmulpd ymm1, ymm1, ymm5
- vmulpd xmm2, xmm3, xmm6
- vmulpd xmm3, xmm3, xmm7
- ADDPD_MEM COVAR(iq ,0), ymm0
- ADDPD_MEM COVAR(iq ,1), ymm1
- ADDPD_MEM COVAR(iq+2,2), xmm2
- ADDPD_MEM COVAR(iq+2,3), xmm3
- %endif ; cpuflag(fma3)
- lea jd, [iq + 4]
- cmp jd, count2d
- jg .skip4x4
- .loop4x4:
- ; Compute all 16 pairwise products of a 4x4 block
- mova ymm3, [varq + jq*8]
- %if cpuflag(fma3)
- mova ymm0, COVAR(jq, 0)
- mova ymm1, COVAR(jq, 1)
- mova ymm2, COVAR(jq, 2)
- fmaddpd ymm0, ymm3, ymm4, ymm0
- fmaddpd ymm1, ymm3, ymm5, ymm1
- fmaddpd ymm2, ymm3, ymm6, ymm2
- fmaddpd ymm3, ymm7, ymm3, COVAR(jq,3)
- mova COVAR(jq, 0), ymm0
- mova COVAR(jq, 1), ymm1
- mova COVAR(jq, 2), ymm2
- mova COVAR(jq, 3), ymm3
- %else
- vmulpd ymm0, ymm3, ymm4
- vmulpd ymm1, ymm3, ymm5
- vmulpd ymm2, ymm3, ymm6
- vmulpd ymm3, ymm3, ymm7
- ADDPD_MEM COVAR(jq,0), ymm0
- ADDPD_MEM COVAR(jq,1), ymm1
- ADDPD_MEM COVAR(jq,2), ymm2
- ADDPD_MEM COVAR(jq,3), ymm3
- %endif ; cpuflag(fma3)
- add jd, 4
- cmp jd, count2d
- jle .loop4x4
- .skip4x4:
- cmp jd, countd
- jg .skip2x4
- mova xmm3, [varq + jq*8]
- %if cpuflag(fma3)
- mova xmm0, COVAR(jq, 0)
- mova xmm1, COVAR(jq, 1)
- mova xmm2, COVAR(jq, 2)
- fmaddpd xmm0, xmm3, xmm4, xmm0
- fmaddpd xmm1, xmm3, xmm5, xmm1
- fmaddpd xmm2, xmm3, xmm6, xmm2
- fmaddpd xmm3, xmm7, xmm3, COVAR(jq,3)
- mova COVAR(jq, 0), xmm0
- mova COVAR(jq, 1), xmm1
- mova COVAR(jq, 2), xmm2
- mova COVAR(jq, 3), xmm3
- %else
- vmulpd xmm0, xmm3, xmm4
- vmulpd xmm1, xmm3, xmm5
- vmulpd xmm2, xmm3, xmm6
- vmulpd xmm3, xmm3, xmm7
- ADDPD_MEM COVAR(jq,0), xmm0
- ADDPD_MEM COVAR(jq,1), xmm1
- ADDPD_MEM COVAR(jq,2), xmm2
- ADDPD_MEM COVAR(jq,3), xmm3
- %endif ; cpuflag(fma3)
- .skip2x4:
- add id, 4
- add covarq, 4*COVAR_STRIDE
- cmp id, count2d
- jle .loopi
- cmp id, countd
- jg .ret
- mov jd, id
- .loop2x1:
- vmovddup xmm0, [varq + iq*8]
- %if cpuflag(fma3)
- mova xmm1, [varq + jq*8]
- fmaddpd xmm0, xmm1, xmm0, COVAR(jq,0)
- mova COVAR(jq,0), xmm0
- %else
- vmulpd xmm0, [varq + jq*8]
- ADDPD_MEM COVAR(jq,0), xmm0
- %endif ; cpuflag(fma3)
- inc id
- add covarq, COVAR_STRIDE
- cmp id, countd
- jle .loop2x1
- .ret:
- REP_RET
- %endmacro ; UPDATE_LLS
- %if HAVE_AVX_EXTERNAL
- INIT_YMM avx
- UPDATE_LLS
- %endif
- %if HAVE_FMA3_EXTERNAL
- INIT_YMM fma3
- UPDATE_LLS
- %endif
- INIT_XMM sse2
- cglobal evaluate_lls, 3,4,2, ctx, var, order, i
- ; This function is often called on the same buffer as update_lls, but with
- ; an offset. They can't both be aligned.
- ; Load halves rather than movu to avoid store-forwarding stalls, since the
- ; input was initialized immediately prior to this function using scalar math.
- %define coefsq ctxq
- mov id, orderd
- imul orderd, MAX_VARS
- lea coefsq, [ctxq + LLSModel.coeff + orderq*8]
- movsd m0, [varq]
- movhpd m0, [varq + 8]
- mulpd m0, [coefsq]
- lea coefsq, [coefsq + iq*8]
- lea varq, [varq + iq*8]
- neg iq
- add iq, 2
- .loop:
- movsd m1, [varq + iq*8]
- movhpd m1, [varq + iq*8 + 8]
- mulpd m1, [coefsq + iq*8]
- addpd m0, m1
- add iq, 2
- jl .loop
- jg .skip1
- movsd m1, [varq + iq*8]
- mulsd m1, [coefsq + iq*8]
- addpd m0, m1
- .skip1:
- movhlps m1, m0
- addsd m0, m1
- %if ARCH_X86_32
- movsd r0m, m0
- fld qword r0m
- %endif
- RET
|