fwd_txfm_msa.h 25 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364
  1. /*
  2. * Copyright (c) 2015 The WebM project authors. All Rights Reserved.
  3. *
  4. * Use of this source code is governed by a BSD-style license
  5. * that can be found in the LICENSE file in the root of the source
  6. * tree. An additional intellectual property rights grant can be found
  7. * in the file PATENTS. All contributing project authors may
  8. * be found in the AUTHORS file in the root of the source tree.
  9. */
  10. #ifndef VPX_VPX_DSP_MIPS_FWD_TXFM_MSA_H_
  11. #define VPX_VPX_DSP_MIPS_FWD_TXFM_MSA_H_
  12. #include "vpx_dsp/mips/txfm_macros_msa.h"
  13. #include "vpx_dsp/txfm_common.h"
  14. #define VP9_FDCT4(in0, in1, in2, in3, out0, out1, out2, out3) \
  15. { \
  16. v8i16 cnst0_m, cnst1_m, cnst2_m, cnst3_m; \
  17. v8i16 vec0_m, vec1_m, vec2_m, vec3_m; \
  18. v4i32 vec4_m, vec5_m, vec6_m, vec7_m; \
  19. v8i16 coeff_m = { \
  20. cospi_16_64, -cospi_16_64, cospi_8_64, cospi_24_64, -cospi_8_64, 0, 0, 0 \
  21. }; \
  22. \
  23. BUTTERFLY_4(in0, in1, in2, in3, vec0_m, vec1_m, vec2_m, vec3_m); \
  24. ILVR_H2_SH(vec1_m, vec0_m, vec3_m, vec2_m, vec0_m, vec2_m); \
  25. SPLATI_H2_SH(coeff_m, 0, 1, cnst0_m, cnst1_m); \
  26. cnst1_m = __msa_ilvev_h(cnst1_m, cnst0_m); \
  27. vec5_m = __msa_dotp_s_w(vec0_m, cnst1_m); \
  28. \
  29. SPLATI_H2_SH(coeff_m, 4, 3, cnst2_m, cnst3_m); \
  30. cnst2_m = __msa_ilvev_h(cnst3_m, cnst2_m); \
  31. vec7_m = __msa_dotp_s_w(vec2_m, cnst2_m); \
  32. \
  33. vec4_m = __msa_dotp_s_w(vec0_m, cnst0_m); \
  34. cnst2_m = __msa_splati_h(coeff_m, 2); \
  35. cnst2_m = __msa_ilvev_h(cnst2_m, cnst3_m); \
  36. vec6_m = __msa_dotp_s_w(vec2_m, cnst2_m); \
  37. \
  38. SRARI_W4_SW(vec4_m, vec5_m, vec6_m, vec7_m, DCT_CONST_BITS); \
  39. PCKEV_H4_SH(vec4_m, vec4_m, vec5_m, vec5_m, vec6_m, vec6_m, vec7_m, \
  40. vec7_m, out0, out2, out1, out3); \
  41. }
  42. #define SRLI_AVE_S_4V_H(in0, in1, in2, in3, in4, in5, in6, in7) \
  43. { \
  44. v8i16 vec0_m, vec1_m, vec2_m, vec3_m, vec4_m, vec5_m, vec6_m, vec7_m; \
  45. \
  46. SRLI_H4_SH(in0, in1, in2, in3, vec0_m, vec1_m, vec2_m, vec3_m, 15); \
  47. SRLI_H4_SH(in4, in5, in6, in7, vec4_m, vec5_m, vec6_m, vec7_m, 15); \
  48. AVE_SH4_SH(vec0_m, in0, vec1_m, in1, vec2_m, in2, vec3_m, in3, in0, in1, \
  49. in2, in3); \
  50. AVE_SH4_SH(vec4_m, in4, vec5_m, in5, vec6_m, in6, vec7_m, in7, in4, in5, \
  51. in6, in7); \
  52. }
  53. #define VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, \
  54. out3, out4, out5, out6, out7) \
  55. { \
  56. v8i16 s0_m, s1_m, s2_m, s3_m, s4_m, s5_m, s6_m; \
  57. v8i16 s7_m, x0_m, x1_m, x2_m, x3_m; \
  58. v8i16 coeff_m = { cospi_16_64, -cospi_16_64, cospi_8_64, cospi_24_64, \
  59. cospi_4_64, cospi_28_64, cospi_12_64, cospi_20_64 }; \
  60. \
  61. /* FDCT stage1 */ \
  62. BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, s0_m, s1_m, s2_m, \
  63. s3_m, s4_m, s5_m, s6_m, s7_m); \
  64. BUTTERFLY_4(s0_m, s1_m, s2_m, s3_m, x0_m, x1_m, x2_m, x3_m); \
  65. ILVL_H2_SH(x1_m, x0_m, x3_m, x2_m, s0_m, s2_m); \
  66. ILVR_H2_SH(x1_m, x0_m, x3_m, x2_m, s1_m, s3_m); \
  67. SPLATI_H2_SH(coeff_m, 0, 1, x0_m, x1_m); \
  68. x1_m = __msa_ilvev_h(x1_m, x0_m); \
  69. out4 = DOT_SHIFT_RIGHT_PCK_H(s0_m, s1_m, x1_m); \
  70. \
  71. SPLATI_H2_SH(coeff_m, 2, 3, x2_m, x3_m); \
  72. x2_m = -x2_m; \
  73. x2_m = __msa_ilvev_h(x3_m, x2_m); \
  74. out6 = DOT_SHIFT_RIGHT_PCK_H(s2_m, s3_m, x2_m); \
  75. \
  76. out0 = DOT_SHIFT_RIGHT_PCK_H(s0_m, s1_m, x0_m); \
  77. x2_m = __msa_splati_h(coeff_m, 2); \
  78. x2_m = __msa_ilvev_h(x2_m, x3_m); \
  79. out2 = DOT_SHIFT_RIGHT_PCK_H(s2_m, s3_m, x2_m); \
  80. \
  81. /* stage2 */ \
  82. ILVRL_H2_SH(s5_m, s6_m, s1_m, s0_m); \
  83. \
  84. s6_m = DOT_SHIFT_RIGHT_PCK_H(s0_m, s1_m, x0_m); \
  85. s5_m = DOT_SHIFT_RIGHT_PCK_H(s0_m, s1_m, x1_m); \
  86. \
  87. /* stage3 */ \
  88. BUTTERFLY_4(s4_m, s7_m, s6_m, s5_m, x0_m, x3_m, x2_m, x1_m); \
  89. \
  90. /* stage4 */ \
  91. ILVL_H2_SH(x3_m, x0_m, x2_m, x1_m, s4_m, s6_m); \
  92. ILVR_H2_SH(x3_m, x0_m, x2_m, x1_m, s5_m, s7_m); \
  93. \
  94. SPLATI_H2_SH(coeff_m, 4, 5, x0_m, x1_m); \
  95. x1_m = __msa_ilvev_h(x0_m, x1_m); \
  96. out1 = DOT_SHIFT_RIGHT_PCK_H(s4_m, s5_m, x1_m); \
  97. \
  98. SPLATI_H2_SH(coeff_m, 6, 7, x2_m, x3_m); \
  99. x2_m = __msa_ilvev_h(x3_m, x2_m); \
  100. out5 = DOT_SHIFT_RIGHT_PCK_H(s6_m, s7_m, x2_m); \
  101. \
  102. x1_m = __msa_splati_h(coeff_m, 5); \
  103. x0_m = -x0_m; \
  104. x0_m = __msa_ilvev_h(x1_m, x0_m); \
  105. out7 = DOT_SHIFT_RIGHT_PCK_H(s4_m, s5_m, x0_m); \
  106. \
  107. x2_m = __msa_splati_h(coeff_m, 6); \
  108. x3_m = -x3_m; \
  109. x2_m = __msa_ilvev_h(x2_m, x3_m); \
  110. out3 = DOT_SHIFT_RIGHT_PCK_H(s6_m, s7_m, x2_m); \
  111. }
  112. #define FDCT8x16_EVEN(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, \
  113. out2, out3, out4, out5, out6, out7) \
  114. { \
  115. v8i16 s0_m, s1_m, s2_m, s3_m, s4_m, s5_m, s6_m, s7_m; \
  116. v8i16 x0_m, x1_m, x2_m, x3_m; \
  117. v8i16 coeff_m = { cospi_16_64, -cospi_16_64, cospi_8_64, cospi_24_64, \
  118. cospi_4_64, cospi_28_64, cospi_12_64, cospi_20_64 }; \
  119. \
  120. /* FDCT stage1 */ \
  121. BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, s0_m, s1_m, s2_m, \
  122. s3_m, s4_m, s5_m, s6_m, s7_m); \
  123. BUTTERFLY_4(s0_m, s1_m, s2_m, s3_m, x0_m, x1_m, x2_m, x3_m); \
  124. ILVL_H2_SH(x1_m, x0_m, x3_m, x2_m, s0_m, s2_m); \
  125. ILVR_H2_SH(x1_m, x0_m, x3_m, x2_m, s1_m, s3_m); \
  126. SPLATI_H2_SH(coeff_m, 0, 1, x0_m, x1_m); \
  127. x1_m = __msa_ilvev_h(x1_m, x0_m); \
  128. out4 = DOT_SHIFT_RIGHT_PCK_H(s0_m, s1_m, x1_m); \
  129. \
  130. SPLATI_H2_SH(coeff_m, 2, 3, x2_m, x3_m); \
  131. x2_m = -x2_m; \
  132. x2_m = __msa_ilvev_h(x3_m, x2_m); \
  133. out6 = DOT_SHIFT_RIGHT_PCK_H(s2_m, s3_m, x2_m); \
  134. \
  135. out0 = DOT_SHIFT_RIGHT_PCK_H(s0_m, s1_m, x0_m); \
  136. x2_m = __msa_splati_h(coeff_m, 2); \
  137. x2_m = __msa_ilvev_h(x2_m, x3_m); \
  138. out2 = DOT_SHIFT_RIGHT_PCK_H(s2_m, s3_m, x2_m); \
  139. \
  140. /* stage2 */ \
  141. ILVRL_H2_SH(s5_m, s6_m, s1_m, s0_m); \
  142. \
  143. s6_m = DOT_SHIFT_RIGHT_PCK_H(s0_m, s1_m, x0_m); \
  144. s5_m = DOT_SHIFT_RIGHT_PCK_H(s0_m, s1_m, x1_m); \
  145. \
  146. /* stage3 */ \
  147. BUTTERFLY_4(s4_m, s7_m, s6_m, s5_m, x0_m, x3_m, x2_m, x1_m); \
  148. \
  149. /* stage4 */ \
  150. ILVL_H2_SH(x3_m, x0_m, x2_m, x1_m, s4_m, s6_m); \
  151. ILVR_H2_SH(x3_m, x0_m, x2_m, x1_m, s5_m, s7_m); \
  152. \
  153. SPLATI_H2_SH(coeff_m, 4, 5, x0_m, x1_m); \
  154. x1_m = __msa_ilvev_h(x0_m, x1_m); \
  155. out1 = DOT_SHIFT_RIGHT_PCK_H(s4_m, s5_m, x1_m); \
  156. \
  157. SPLATI_H2_SH(coeff_m, 6, 7, x2_m, x3_m); \
  158. x2_m = __msa_ilvev_h(x3_m, x2_m); \
  159. out5 = DOT_SHIFT_RIGHT_PCK_H(s6_m, s7_m, x2_m); \
  160. \
  161. x1_m = __msa_splati_h(coeff_m, 5); \
  162. x0_m = -x0_m; \
  163. x0_m = __msa_ilvev_h(x1_m, x0_m); \
  164. out7 = DOT_SHIFT_RIGHT_PCK_H(s4_m, s5_m, x0_m); \
  165. \
  166. x2_m = __msa_splati_h(coeff_m, 6); \
  167. x3_m = -x3_m; \
  168. x2_m = __msa_ilvev_h(x2_m, x3_m); \
  169. out3 = DOT_SHIFT_RIGHT_PCK_H(s6_m, s7_m, x2_m); \
  170. }
  171. #define FDCT8x16_ODD(input0, input1, input2, input3, input4, input5, input6, \
  172. input7, out1, out3, out5, out7, out9, out11, out13, \
  173. out15) \
  174. { \
  175. v8i16 stp21_m, stp22_m, stp23_m, stp24_m, stp25_m, stp26_m; \
  176. v8i16 stp30_m, stp31_m, stp32_m, stp33_m, stp34_m, stp35_m; \
  177. v8i16 stp36_m, stp37_m, vec0_m, vec1_m; \
  178. v8i16 vec2_m, vec3_m, vec4_m, vec5_m, vec6_m; \
  179. v8i16 cnst0_m, cnst1_m, cnst4_m, cnst5_m; \
  180. v8i16 coeff_m = { cospi_16_64, -cospi_16_64, cospi_8_64, cospi_24_64, \
  181. -cospi_8_64, -cospi_24_64, cospi_12_64, cospi_20_64 }; \
  182. v8i16 coeff1_m = { cospi_2_64, cospi_30_64, cospi_14_64, cospi_18_64, \
  183. cospi_10_64, cospi_22_64, cospi_6_64, cospi_26_64 }; \
  184. v8i16 coeff2_m = { \
  185. -cospi_2_64, -cospi_10_64, -cospi_18_64, -cospi_26_64, 0, 0, 0, 0 \
  186. }; \
  187. \
  188. /* stp 1 */ \
  189. ILVL_H2_SH(input2, input5, input3, input4, vec2_m, vec4_m); \
  190. ILVR_H2_SH(input2, input5, input3, input4, vec3_m, vec5_m); \
  191. \
  192. cnst4_m = __msa_splati_h(coeff_m, 0); \
  193. stp25_m = DOT_SHIFT_RIGHT_PCK_H(vec2_m, vec3_m, cnst4_m); \
  194. \
  195. cnst5_m = __msa_splati_h(coeff_m, 1); \
  196. cnst5_m = __msa_ilvev_h(cnst5_m, cnst4_m); \
  197. stp22_m = DOT_SHIFT_RIGHT_PCK_H(vec2_m, vec3_m, cnst5_m); \
  198. stp24_m = DOT_SHIFT_RIGHT_PCK_H(vec4_m, vec5_m, cnst4_m); \
  199. stp23_m = DOT_SHIFT_RIGHT_PCK_H(vec4_m, vec5_m, cnst5_m); \
  200. \
  201. /* stp2 */ \
  202. BUTTERFLY_4(input0, input1, stp22_m, stp23_m, stp30_m, stp31_m, stp32_m, \
  203. stp33_m); \
  204. BUTTERFLY_4(input7, input6, stp25_m, stp24_m, stp37_m, stp36_m, stp35_m, \
  205. stp34_m); \
  206. \
  207. ILVL_H2_SH(stp36_m, stp31_m, stp35_m, stp32_m, vec2_m, vec4_m); \
  208. ILVR_H2_SH(stp36_m, stp31_m, stp35_m, stp32_m, vec3_m, vec5_m); \
  209. \
  210. SPLATI_H2_SH(coeff_m, 2, 3, cnst0_m, cnst1_m); \
  211. cnst0_m = __msa_ilvev_h(cnst0_m, cnst1_m); \
  212. stp26_m = DOT_SHIFT_RIGHT_PCK_H(vec2_m, vec3_m, cnst0_m); \
  213. \
  214. cnst0_m = __msa_splati_h(coeff_m, 4); \
  215. cnst1_m = __msa_ilvev_h(cnst1_m, cnst0_m); \
  216. stp21_m = DOT_SHIFT_RIGHT_PCK_H(vec2_m, vec3_m, cnst1_m); \
  217. \
  218. SPLATI_H2_SH(coeff_m, 5, 2, cnst0_m, cnst1_m); \
  219. cnst1_m = __msa_ilvev_h(cnst0_m, cnst1_m); \
  220. stp25_m = DOT_SHIFT_RIGHT_PCK_H(vec4_m, vec5_m, cnst1_m); \
  221. \
  222. cnst0_m = __msa_splati_h(coeff_m, 3); \
  223. cnst1_m = __msa_ilvev_h(cnst1_m, cnst0_m); \
  224. stp22_m = DOT_SHIFT_RIGHT_PCK_H(vec4_m, vec5_m, cnst1_m); \
  225. \
  226. /* stp4 */ \
  227. BUTTERFLY_4(stp30_m, stp37_m, stp26_m, stp21_m, vec6_m, vec2_m, vec4_m, \
  228. vec5_m); \
  229. BUTTERFLY_4(stp33_m, stp34_m, stp25_m, stp22_m, stp21_m, stp23_m, stp24_m, \
  230. stp31_m); \
  231. \
  232. ILVRL_H2_SH(vec2_m, vec6_m, vec1_m, vec0_m); \
  233. SPLATI_H2_SH(coeff1_m, 0, 1, cnst0_m, cnst1_m); \
  234. cnst0_m = __msa_ilvev_h(cnst0_m, cnst1_m); \
  235. \
  236. out1 = DOT_SHIFT_RIGHT_PCK_H(vec0_m, vec1_m, cnst0_m); \
  237. \
  238. cnst0_m = __msa_splati_h(coeff2_m, 0); \
  239. cnst0_m = __msa_ilvev_h(cnst1_m, cnst0_m); \
  240. out15 = DOT_SHIFT_RIGHT_PCK_H(vec0_m, vec1_m, cnst0_m); \
  241. \
  242. ILVRL_H2_SH(vec4_m, vec5_m, vec1_m, vec0_m); \
  243. SPLATI_H2_SH(coeff1_m, 2, 3, cnst0_m, cnst1_m); \
  244. cnst1_m = __msa_ilvev_h(cnst1_m, cnst0_m); \
  245. \
  246. out9 = DOT_SHIFT_RIGHT_PCK_H(vec0_m, vec1_m, cnst1_m); \
  247. \
  248. cnst1_m = __msa_splati_h(coeff2_m, 2); \
  249. cnst0_m = __msa_ilvev_h(cnst0_m, cnst1_m); \
  250. out7 = DOT_SHIFT_RIGHT_PCK_H(vec0_m, vec1_m, cnst0_m); \
  251. \
  252. ILVRL_H2_SH(stp23_m, stp21_m, vec1_m, vec0_m); \
  253. SPLATI_H2_SH(coeff1_m, 4, 5, cnst0_m, cnst1_m); \
  254. cnst0_m = __msa_ilvev_h(cnst0_m, cnst1_m); \
  255. out5 = DOT_SHIFT_RIGHT_PCK_H(vec0_m, vec1_m, cnst0_m); \
  256. \
  257. cnst0_m = __msa_splati_h(coeff2_m, 1); \
  258. cnst0_m = __msa_ilvev_h(cnst1_m, cnst0_m); \
  259. out11 = DOT_SHIFT_RIGHT_PCK_H(vec0_m, vec1_m, cnst0_m); \
  260. \
  261. ILVRL_H2_SH(stp24_m, stp31_m, vec1_m, vec0_m); \
  262. SPLATI_H2_SH(coeff1_m, 6, 7, cnst0_m, cnst1_m); \
  263. cnst1_m = __msa_ilvev_h(cnst1_m, cnst0_m); \
  264. \
  265. out13 = DOT_SHIFT_RIGHT_PCK_H(vec0_m, vec1_m, cnst1_m); \
  266. \
  267. cnst1_m = __msa_splati_h(coeff2_m, 3); \
  268. cnst0_m = __msa_ilvev_h(cnst0_m, cnst1_m); \
  269. out3 = DOT_SHIFT_RIGHT_PCK_H(vec0_m, vec1_m, cnst0_m); \
  270. }
  271. #define FDCT_POSTPROC_2V_NEG_H(vec0, vec1) \
  272. { \
  273. v8i16 tp0_m, tp1_m; \
  274. v8i16 one_m = __msa_ldi_h(1); \
  275. \
  276. tp0_m = __msa_clti_s_h(vec0, 0); \
  277. tp1_m = __msa_clti_s_h(vec1, 0); \
  278. vec0 += 1; \
  279. vec1 += 1; \
  280. tp0_m = one_m & tp0_m; \
  281. tp1_m = one_m & tp1_m; \
  282. vec0 += tp0_m; \
  283. vec1 += tp1_m; \
  284. vec0 >>= 2; \
  285. vec1 >>= 2; \
  286. }
  287. #define FDCT32_POSTPROC_NEG_W(vec) \
  288. { \
  289. v4i32 temp_m; \
  290. v4i32 one_m = __msa_ldi_w(1); \
  291. \
  292. temp_m = __msa_clti_s_w(vec, 0); \
  293. vec += 1; \
  294. temp_m = one_m & temp_m; \
  295. vec += temp_m; \
  296. vec >>= 2; \
  297. }
  298. #define FDCT32_POSTPROC_2V_POS_H(vec0, vec1) \
  299. { \
  300. v8i16 tp0_m, tp1_m; \
  301. v8i16 one = __msa_ldi_h(1); \
  302. \
  303. tp0_m = __msa_clei_s_h(vec0, 0); \
  304. tp1_m = __msa_clei_s_h(vec1, 0); \
  305. tp0_m = (v8i16)__msa_xori_b((v16u8)tp0_m, 255); \
  306. tp1_m = (v8i16)__msa_xori_b((v16u8)tp1_m, 255); \
  307. vec0 += 1; \
  308. vec1 += 1; \
  309. tp0_m = one & tp0_m; \
  310. tp1_m = one & tp1_m; \
  311. vec0 += tp0_m; \
  312. vec1 += tp1_m; \
  313. vec0 >>= 2; \
  314. vec1 >>= 2; \
  315. }
  316. #define DOTP_CONST_PAIR_W(reg0_left, reg1_left, reg0_right, reg1_right, \
  317. const0, const1, out0, out1, out2, out3) \
  318. { \
  319. v4i32 s0_m, s1_m, s2_m, s3_m, s4_m, s5_m, s6_m, s7_m; \
  320. v2i64 tp0_m, tp1_m, tp2_m, tp3_m; \
  321. v4i32 k0_m = __msa_fill_w((int32_t)const0); \
  322. \
  323. s0_m = __msa_fill_w((int32_t)const1); \
  324. k0_m = __msa_ilvev_w(s0_m, k0_m); \
  325. \
  326. ILVRL_W2_SW(-reg1_left, reg0_left, s1_m, s0_m); \
  327. ILVRL_W2_SW(reg0_left, reg1_left, s3_m, s2_m); \
  328. ILVRL_W2_SW(-reg1_right, reg0_right, s5_m, s4_m); \
  329. ILVRL_W2_SW(reg0_right, reg1_right, s7_m, s6_m); \
  330. \
  331. DOTP_SW2_SD(s0_m, s1_m, k0_m, k0_m, tp0_m, tp1_m); \
  332. DOTP_SW2_SD(s4_m, s5_m, k0_m, k0_m, tp2_m, tp3_m); \
  333. tp0_m = __msa_srari_d(tp0_m, DCT_CONST_BITS); \
  334. tp1_m = __msa_srari_d(tp1_m, DCT_CONST_BITS); \
  335. tp2_m = __msa_srari_d(tp2_m, DCT_CONST_BITS); \
  336. tp3_m = __msa_srari_d(tp3_m, DCT_CONST_BITS); \
  337. out0 = __msa_pckev_w((v4i32)tp0_m, (v4i32)tp1_m); \
  338. out1 = __msa_pckev_w((v4i32)tp2_m, (v4i32)tp3_m); \
  339. \
  340. DOTP_SW2_SD(s2_m, s3_m, k0_m, k0_m, tp0_m, tp1_m); \
  341. DOTP_SW2_SD(s6_m, s7_m, k0_m, k0_m, tp2_m, tp3_m); \
  342. tp0_m = __msa_srari_d(tp0_m, DCT_CONST_BITS); \
  343. tp1_m = __msa_srari_d(tp1_m, DCT_CONST_BITS); \
  344. tp2_m = __msa_srari_d(tp2_m, DCT_CONST_BITS); \
  345. tp3_m = __msa_srari_d(tp3_m, DCT_CONST_BITS); \
  346. out2 = __msa_pckev_w((v4i32)tp0_m, (v4i32)tp1_m); \
  347. out3 = __msa_pckev_w((v4i32)tp2_m, (v4i32)tp3_m); \
  348. }
  349. void fdct8x16_1d_column(const int16_t *input, int16_t *tmp_ptr,
  350. int32_t src_stride);
  351. void fdct16x8_1d_row(int16_t *input, int16_t *output);
  352. #endif // VPX_VPX_DSP_MIPS_FWD_TXFM_MSA_H_