123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336 |
- /* Copyright (C) 2002 Jean-Marc Valin */
- /**
- @file filters_sse.h
- @brief Various analysis/synthesis filters (SSE version)
- */
- /*
- Redistribution and use in source and binary forms, with or without
- modification, are permitted provided that the following conditions
- are met:
- - Redistributions of source code must retain the above copyright
- notice, this list of conditions and the following disclaimer.
- - Redistributions in binary form must reproduce the above copyright
- notice, this list of conditions and the following disclaimer in the
- documentation and/or other materials provided with the distribution.
- - Neither the name of the Xiph.org Foundation nor the names of its
- contributors may be used to endorse or promote products derived from
- this software without specific prior written permission.
- THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
- ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
- LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
- A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR
- CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
- EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
- PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
- PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
- LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
- NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
- SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
- */
- #include <xmmintrin.h>
- void filter_mem16_10(const float *x, const float *_num, const float *_den, float *y, int N, int ord, float *_mem)
- {
- __m128 num[3], den[3], mem[3];
- int i;
- /* Copy numerator, denominator and memory to aligned xmm */
- for (i=0;i<2;i++)
- {
- mem[i] = _mm_loadu_ps(_mem+4*i);
- num[i] = _mm_loadu_ps(_num+4*i);
- den[i] = _mm_loadu_ps(_den+4*i);
- }
- mem[2] = _mm_setr_ps(_mem[8], _mem[9], 0, 0);
- num[2] = _mm_setr_ps(_num[8], _num[9], 0, 0);
- den[2] = _mm_setr_ps(_den[8], _den[9], 0, 0);
- for (i=0;i<N;i++)
- {
- __m128 xx;
- __m128 yy;
- /* Compute next filter result */
- xx = _mm_load_ps1(x+i);
- yy = _mm_add_ss(xx, mem[0]);
- _mm_store_ss(y+i, yy);
- yy = _mm_shuffle_ps(yy, yy, 0);
- /* Update memory */
- mem[0] = _mm_move_ss(mem[0], mem[1]);
- mem[0] = _mm_shuffle_ps(mem[0], mem[0], 0x39);
- mem[0] = _mm_add_ps(mem[0], _mm_mul_ps(xx, num[0]));
- mem[0] = _mm_sub_ps(mem[0], _mm_mul_ps(yy, den[0]));
- mem[1] = _mm_move_ss(mem[1], mem[2]);
- mem[1] = _mm_shuffle_ps(mem[1], mem[1], 0x39);
- mem[1] = _mm_add_ps(mem[1], _mm_mul_ps(xx, num[1]));
- mem[1] = _mm_sub_ps(mem[1], _mm_mul_ps(yy, den[1]));
- mem[2] = _mm_shuffle_ps(mem[2], mem[2], 0xfd);
- mem[2] = _mm_add_ps(mem[2], _mm_mul_ps(xx, num[2]));
- mem[2] = _mm_sub_ps(mem[2], _mm_mul_ps(yy, den[2]));
- }
- /* Put memory back in its place */
- _mm_storeu_ps(_mem, mem[0]);
- _mm_storeu_ps(_mem+4, mem[1]);
- _mm_store_ss(_mem+8, mem[2]);
- mem[2] = _mm_shuffle_ps(mem[2], mem[2], 0x55);
- _mm_store_ss(_mem+9, mem[2]);
- }
- void filter_mem16_8(const float *x, const float *_num, const float *_den, float *y, int N, int ord, float *_mem)
- {
- __m128 num[2], den[2], mem[2];
- int i;
- /* Copy numerator, denominator and memory to aligned xmm */
- for (i=0;i<2;i++)
- {
- mem[i] = _mm_loadu_ps(_mem+4*i);
- num[i] = _mm_loadu_ps(_num+4*i);
- den[i] = _mm_loadu_ps(_den+4*i);
- }
- for (i=0;i<N;i++)
- {
- __m128 xx;
- __m128 yy;
- /* Compute next filter result */
- xx = _mm_load_ps1(x+i);
- yy = _mm_add_ss(xx, mem[0]);
- _mm_store_ss(y+i, yy);
- yy = _mm_shuffle_ps(yy, yy, 0);
- /* Update memory */
- mem[0] = _mm_move_ss(mem[0], mem[1]);
- mem[0] = _mm_shuffle_ps(mem[0], mem[0], 0x39);
- mem[0] = _mm_add_ps(mem[0], _mm_mul_ps(xx, num[0]));
- mem[0] = _mm_sub_ps(mem[0], _mm_mul_ps(yy, den[0]));
- mem[1] = _mm_sub_ss(mem[1], mem[1]);
- mem[1] = _mm_shuffle_ps(mem[1], mem[1], 0x39);
- mem[1] = _mm_add_ps(mem[1], _mm_mul_ps(xx, num[1]));
- mem[1] = _mm_sub_ps(mem[1], _mm_mul_ps(yy, den[1]));
- }
- /* Put memory back in its place */
- _mm_storeu_ps(_mem, mem[0]);
- _mm_storeu_ps(_mem+4, mem[1]);
- }
- #define OVERRIDE_FILTER_MEM16
- void filter_mem16(const float *x, const float *_num, const float *_den, float *y, int N, int ord, float *_mem, char *stack)
- {
- if(ord==10)
- filter_mem16_10(x, _num, _den, y, N, ord, _mem);
- else if (ord==8)
- filter_mem16_8(x, _num, _den, y, N, ord, _mem);
- }
- void iir_mem16_10(const float *x, const float *_den, float *y, int N, int ord, float *_mem)
- {
- __m128 den[3], mem[3];
- int i;
- /* Copy numerator, denominator and memory to aligned xmm */
- for (i=0;i<2;i++)
- {
- mem[i] = _mm_loadu_ps(_mem+4*i);
- den[i] = _mm_loadu_ps(_den+4*i);
- }
- mem[2] = _mm_setr_ps(_mem[8], _mem[9], 0, 0);
- den[2] = _mm_setr_ps(_den[8], _den[9], 0, 0);
- for (i=0;i<N;i++)
- {
- __m128 xx;
- __m128 yy;
- /* Compute next filter result */
- xx = _mm_load_ps1(x+i);
- yy = _mm_add_ss(xx, mem[0]);
- _mm_store_ss(y+i, yy);
- yy = _mm_shuffle_ps(yy, yy, 0);
- /* Update memory */
- mem[0] = _mm_move_ss(mem[0], mem[1]);
- mem[0] = _mm_shuffle_ps(mem[0], mem[0], 0x39);
- mem[0] = _mm_sub_ps(mem[0], _mm_mul_ps(yy, den[0]));
- mem[1] = _mm_move_ss(mem[1], mem[2]);
- mem[1] = _mm_shuffle_ps(mem[1], mem[1], 0x39);
- mem[1] = _mm_sub_ps(mem[1], _mm_mul_ps(yy, den[1]));
- mem[2] = _mm_shuffle_ps(mem[2], mem[2], 0xfd);
- mem[2] = _mm_sub_ps(mem[2], _mm_mul_ps(yy, den[2]));
- }
- /* Put memory back in its place */
- _mm_storeu_ps(_mem, mem[0]);
- _mm_storeu_ps(_mem+4, mem[1]);
- _mm_store_ss(_mem+8, mem[2]);
- mem[2] = _mm_shuffle_ps(mem[2], mem[2], 0x55);
- _mm_store_ss(_mem+9, mem[2]);
- }
- void iir_mem16_8(const float *x, const float *_den, float *y, int N, int ord, float *_mem)
- {
- __m128 den[2], mem[2];
- int i;
- /* Copy numerator, denominator and memory to aligned xmm */
- for (i=0;i<2;i++)
- {
- mem[i] = _mm_loadu_ps(_mem+4*i);
- den[i] = _mm_loadu_ps(_den+4*i);
- }
- for (i=0;i<N;i++)
- {
- __m128 xx;
- __m128 yy;
- /* Compute next filter result */
- xx = _mm_load_ps1(x+i);
- yy = _mm_add_ss(xx, mem[0]);
- _mm_store_ss(y+i, yy);
- yy = _mm_shuffle_ps(yy, yy, 0);
- /* Update memory */
- mem[0] = _mm_move_ss(mem[0], mem[1]);
- mem[0] = _mm_shuffle_ps(mem[0], mem[0], 0x39);
- mem[0] = _mm_sub_ps(mem[0], _mm_mul_ps(yy, den[0]));
- mem[1] = _mm_sub_ss(mem[1], mem[1]);
- mem[1] = _mm_shuffle_ps(mem[1], mem[1], 0x39);
- mem[1] = _mm_sub_ps(mem[1], _mm_mul_ps(yy, den[1]));
- }
- /* Put memory back in its place */
- _mm_storeu_ps(_mem, mem[0]);
- _mm_storeu_ps(_mem+4, mem[1]);
- }
- #define OVERRIDE_IIR_MEM16
- void iir_mem16(const float *x, const float *_den, float *y, int N, int ord, float *_mem, char *stack)
- {
- if(ord==10)
- iir_mem16_10(x, _den, y, N, ord, _mem);
- else if (ord==8)
- iir_mem16_8(x, _den, y, N, ord, _mem);
- }
- void fir_mem16_10(const float *x, const float *_num, float *y, int N, int ord, float *_mem)
- {
- __m128 num[3], mem[3];
- int i;
- /* Copy numerator, denominator and memory to aligned xmm */
- for (i=0;i<2;i++)
- {
- mem[i] = _mm_loadu_ps(_mem+4*i);
- num[i] = _mm_loadu_ps(_num+4*i);
- }
- mem[2] = _mm_setr_ps(_mem[8], _mem[9], 0, 0);
- num[2] = _mm_setr_ps(_num[8], _num[9], 0, 0);
- for (i=0;i<N;i++)
- {
- __m128 xx;
- __m128 yy;
- /* Compute next filter result */
- xx = _mm_load_ps1(x+i);
- yy = _mm_add_ss(xx, mem[0]);
- _mm_store_ss(y+i, yy);
- yy = _mm_shuffle_ps(yy, yy, 0);
- /* Update memory */
- mem[0] = _mm_move_ss(mem[0], mem[1]);
- mem[0] = _mm_shuffle_ps(mem[0], mem[0], 0x39);
- mem[0] = _mm_add_ps(mem[0], _mm_mul_ps(xx, num[0]));
- mem[1] = _mm_move_ss(mem[1], mem[2]);
- mem[1] = _mm_shuffle_ps(mem[1], mem[1], 0x39);
- mem[1] = _mm_add_ps(mem[1], _mm_mul_ps(xx, num[1]));
- mem[2] = _mm_shuffle_ps(mem[2], mem[2], 0xfd);
- mem[2] = _mm_add_ps(mem[2], _mm_mul_ps(xx, num[2]));
- }
- /* Put memory back in its place */
- _mm_storeu_ps(_mem, mem[0]);
- _mm_storeu_ps(_mem+4, mem[1]);
- _mm_store_ss(_mem+8, mem[2]);
- mem[2] = _mm_shuffle_ps(mem[2], mem[2], 0x55);
- _mm_store_ss(_mem+9, mem[2]);
- }
- void fir_mem16_8(const float *x, const float *_num, float *y, int N, int ord, float *_mem)
- {
- __m128 num[2], mem[2];
- int i;
- /* Copy numerator, denominator and memory to aligned xmm */
- for (i=0;i<2;i++)
- {
- mem[i] = _mm_loadu_ps(_mem+4*i);
- num[i] = _mm_loadu_ps(_num+4*i);
- }
- for (i=0;i<N;i++)
- {
- __m128 xx;
- __m128 yy;
- /* Compute next filter result */
- xx = _mm_load_ps1(x+i);
- yy = _mm_add_ss(xx, mem[0]);
- _mm_store_ss(y+i, yy);
- yy = _mm_shuffle_ps(yy, yy, 0);
- /* Update memory */
- mem[0] = _mm_move_ss(mem[0], mem[1]);
- mem[0] = _mm_shuffle_ps(mem[0], mem[0], 0x39);
- mem[0] = _mm_add_ps(mem[0], _mm_mul_ps(xx, num[0]));
- mem[1] = _mm_sub_ss(mem[1], mem[1]);
- mem[1] = _mm_shuffle_ps(mem[1], mem[1], 0x39);
- mem[1] = _mm_add_ps(mem[1], _mm_mul_ps(xx, num[1]));
- }
- /* Put memory back in its place */
- _mm_storeu_ps(_mem, mem[0]);
- _mm_storeu_ps(_mem+4, mem[1]);
- }
- #define OVERRIDE_FIR_MEM16
- void fir_mem16(const float *x, const float *_num, float *y, int N, int ord, float *_mem, char *stack)
- {
- if(ord==10)
- fir_mem16_10(x, _num, y, N, ord, _mem);
- else if (ord==8)
- fir_mem16_8(x, _num, y, N, ord, _mem);
- }
|