/external/libvpx/libvpx/vpx_dsp/mips/ |
D | fwd_txfm_msa.c | 32 v8i16 in8, in9, in10, in11, in12, in13, in14, in15; in fdct8x16_1d_column() local 44 LD_SH16(input, src_stride, in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in fdct8x16_1d_column() 48 SLLI_4V(in8, in9, in10, in11, 2); in fdct8x16_1d_column() 51 ADD4(in4, in11, in5, in10, in6, in9, in7, in8, tmp4, tmp5, tmp6, tmp7); in fdct8x16_1d_column() 56 SUB4(in4, in11, in5, in10, in6, in9, in7, in8, in11, in10, in9, in8); in fdct8x16_1d_column() 74 BUTTERFLY_4(in8, in9, stp22, stp23, stp30, stp31, stp32, stp33); in fdct8x16_1d_column() 86 BUTTERFLY_4(stp30, stp37, stp26, stp21, in8, in15, in14, in9); in fdct8x16_1d_column() 87 ILVRL_H2_SH(in15, in8, vec1, vec0); in fdct8x16_1d_column() 91 in8 = DOT_SHIFT_RIGHT_PCK_H(vec0, vec1, cnst0); in fdct8x16_1d_column() 92 ST_SH(in8, tmp_ptr); in fdct8x16_1d_column() [all …]
|
D | fwd_dct32x32_msa.c | 60 v8i16 in8, in9, in10, in11, in12, in13, in14, in15; in fdct8x32_1d_column_even_store() local 70 LD_SH4(input + 64, 8, in8, in9, in10, in11); in fdct8x32_1d_column_even_store() 71 BUTTERFLY_8(in4, in5, in6, in7, in8, in9, in10, in11, vec4, vec5, vec6, vec7, in fdct8x32_1d_column_even_store() 72 in8, in9, in10, in11); in fdct8x32_1d_column_even_store() 103 ADD4(in8, vec3, in9, vec2, in14, vec5, in15, vec4, in0, vec1, vec6, in2); in fdct8x32_1d_column_even_store() 119 SUB4(in8, vec3, in15, vec4, in3, in2, in0, in1, in3, in0, vec2, vec5); in fdct8x32_1d_column_even_store() 256 v8i16 in8, in9, in10, in11, in12, in13, in14, in15; in fdct8x32_1d_row_load_butterfly() local 260 LD_SH8(temp_buff + 24, 32, in8, in9, in10, in11, in12, in13, in14, in15); in fdct8x32_1d_row_load_butterfly() 263 TRANSPOSE8x8_SH_SH(in8, in9, in10, in11, in12, in13, in14, in15, in8, in9, in fdct8x32_1d_row_load_butterfly() 265 BUTTERFLY_16(in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in10, in11, in fdct8x32_1d_row_load_butterfly() [all …]
|
D | deblock_msa.c | 74 #define TRANSPOSE12x16_B(in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, \ argument 86 ILVR_B2_SH(in9, in8, in11, in10, temp4, temp5); \ 87 ILVR_B2_SH(in9, in8, in11, in10, temp4, temp5); \ 101 ILVL_B4_SH(in9, in8, in11, in10, in13, in12, in15, in14, temp2, temp3, \ 106 in8 = (v16u8)__msa_ilvr_d((v2i64)temp1, (v2i64)temp0); \ 113 #define VPX_TRANSPOSE12x8_UB_UB(in0, in1, in2, in3, in4, in5, in6, in7, in8, \ argument 134 in8 = (v16u8)temp6; \
|
D | macros_msa.h | 1073 #define ILVR_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in10, \ argument 1079 ILVR_B4(RTYPE, in8, in9, in10, in11, in12, in13, in14, in15, out4, out5, \ 1690 #define BUTTERFLY_16(in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in10, \ argument 1702 out7 = in7 + in8; \ 1704 out8 = in7 - in8; \ 1742 #define TRANSPOSE16x8_UB_UB(in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, \ argument 1749 ILVEV_D2_UB(in0, in8, in1, in9, out7, out6); \
|
D | vpx_convolve8_msa.c | 829 v16u8 in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in10, in11, in12; in transpose16x16_to_dst() local 834 LD_UB8(src + 16 * 8, 16, in8, in9, in10, in11, in12, in13, in14, in15); in transpose16x16_to_dst() 836 TRANSPOSE16x8_UB_UB(in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in10, in transpose16x16_to_dst() 844 SLDI_B4_0_UB(in8, in9, in10, in11, in8, in9, in10, in11, 8); in transpose16x16_to_dst() 847 TRANSPOSE16x8_UB_UB(in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in10, in transpose16x16_to_dst()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | aggregate_ops_cpu.h | 107 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8) { 109 in7, in8); 120 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8) { 122 in7, in8); 133 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8, 136 in7, in8, in9); 215 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8) { 217 in7, in8); 228 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8) { 230 in7, in8); [all …]
|
D | aggregate_ops_gpu.cu.cc | 109 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8) { in operator ()() 111 in7, in8); in operator ()() 122 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8) { in operator ()() 124 in7, in8); in operator ()() 135 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8, in operator ()() 138 in7, in8, in9); in operator ()()
|
D | aggregate_ops.h | 160 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8); 170 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8) { in Compute() 171 out.device(d) = in1 + in2 + in3 + in4 + in5 + in6 + in7 + in8; in Compute() 184 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8); 194 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8) { in Compute() 195 out.device(d) += in1 + in2 + in3 + in4 + in5 + in6 + in7 + in8; in Compute() 206 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8, 217 typename TTypes<T>::ConstFlat in7, typename TTypes<T>::ConstFlat in8, in Compute() 219 out.device(d) = in1 + in2 + in3 + in4 + in5 + in6 + in7 + in8 + in9; in Compute()
|
/external/llvm/test/CodeGen/Hexagon/ |
D | circ_ldd_bug.ll | 86 %var8.0.in8.unr = phi i8* [ %4, %unr.cmp ], [ %11, %for.body.unr ] 88 %16 = call i8* @llvm.hexagon.circ.ldd(i8* %var8.0.in8.unr, i8* %3, i32 %or, i32 -8) 100 %var8.0.in8.unr19 = phi i8* [ %4, %unr.cmp24 ], [ %16, %for.body.unr13 ] 102 %21 = call i8* @llvm.hexagon.circ.ldd(i8* %var8.0.in8.unr19, i8* %3, i32 %or, i32 -8) 114 %var8.0.in8.unr28 = phi i8* [ %4, %unr.cmp33 ], [ %21, %for.body.unr17 ] 116 %26 = call i8* @llvm.hexagon.circ.ldd(i8* %var8.0.in8.unr28, i8* %3, i32 %or, i32 -8) 128 %var8.0.in8.unr37 = phi i8* [ %4, %unr.cmp42 ], [ %26, %for.body.unr26 ] 130 %31 = call i8* @llvm.hexagon.circ.ldd(i8* %var8.0.in8.unr37, i8* %3, i32 %or, i32 -8) 142 %var8.0.in8.unr46 = phi i8* [ %4, %unr.cmp51 ], [ %31, %for.body.unr35 ] 144 %36 = call i8* @llvm.hexagon.circ.ldd(i8* %var8.0.in8.unr46, i8* %3, i32 %or, i32 -8) [all …]
|
/external/webp/src/dsp/ |
D | enc_sse41.c | 210 __m128i in8 = _mm_loadu_si128((__m128i*)&in[8]); in DoQuantizeBlock_SSE41() local 218 __m128i coeff8 = _mm_abs_epi16(in8); in DoQuantizeBlock_SSE41() 266 out8 = _mm_sign_epi16(out8, in8); in DoQuantizeBlock_SSE41() 270 in8 = _mm_mullo_epi16(out8, q8); in DoQuantizeBlock_SSE41() 273 _mm_storeu_si128((__m128i*)&in[8], in8); in DoQuantizeBlock_SSE41()
|
D | enc_sse2.c | 1231 __m128i in8 = _mm_loadu_si128((__m128i*)&in[8]); in DoQuantizeBlock_SSE2() local 1239 const __m128i sign8 = _mm_cmpgt_epi16(zero, in8); in DoQuantizeBlock_SSE2() 1243 coeff8 = _mm_xor_si128(in8, sign8); in DoQuantizeBlock_SSE2() 1299 in8 = _mm_mullo_epi16(out8, q8); in DoQuantizeBlock_SSE2() 1302 _mm_storeu_si128((__m128i*)&in[8], in8); in DoQuantizeBlock_SSE2()
|
D | msa_macro.h | 1283 in8, in9, in10, in11, in12, in13, in14, in15, \ argument 1286 ILVEV_W2_SD(in0, in4, in8, in12, tmp2_m, tmp3_m); \ 1305 in8, in9, in10, in11, in12, in13, in14, in15, \ argument 1310 ILVEV_D2_UB(in0, in8, in1, in9, out7, out6); \
|
/external/libvpx/libvpx/vp9/encoder/mips/msa/ |
D | vp9_fdct16x16_msa.c | 368 v8i16 in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in10, in11; in postproc_fdct16x8_1d_row() local 373 LD_SH8(temp, 16, in8, in9, in10, in11, in12, in13, in14, in15); in postproc_fdct16x8_1d_row() 376 TRANSPOSE8x8_SH_SH(in8, in9, in10, in11, in12, in13, in14, in15, in8, in9, in postproc_fdct16x8_1d_row() 382 FDCT_POSTPROC_2V_NEG_H(in8, in9); in postproc_fdct16x8_1d_row() 386 BUTTERFLY_16(in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in10, in11, in postproc_fdct16x8_1d_row() 388 tmp7, in8, in9, in10, in11, in12, in13, in14, in15); in postproc_fdct16x8_1d_row() 390 ST_SH8(in8, in9, in10, in11, in12, in13, in14, in15, temp, 16); in postproc_fdct16x8_1d_row() 394 LD_SH8(temp, 16, in8, in9, in10, in11, in12, in13, in14, in15); in postproc_fdct16x8_1d_row() 395 FDCT8x16_ODD(in8, in9, in10, in11, in12, in13, in14, in15, in0, in1, in2, in3, in postproc_fdct16x8_1d_row()
|
/external/libvpx/libvpx/vpx_dsp/ppc/ |
D | inv_txfm_vsx.c | 333 in6, in7, in8, in9, inA, inB, inC, inD, inE, inF) \ argument 342 in8 = load(8 * (step) + (offset), source); \ 369 #define IDCT16(in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, inA, inB, \ argument 374 out1 = in8; \ 400 STEP8_0(out8, outF, in8, inF, cospi30_v, cospi2_v); \ 414 out8 = vec_add(in8, in9); \ 415 out9 = vec_sub(in8, in9); \ 431 in8 = out8; \ 473 out8 = vec_add(in8, inB); \ 476 outB = vec_sub(in8, inB); \ [all …]
|
/external/libvpx/libvpx/vp8/common/mips/msa/ |
D | vp8_macros_msa.h | 1572 #define TRANSPOSE16x4_UB_UB(in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, \ argument 1578 ILVEV_W2_SD(in0, in4, in8, in12, tmp0_m, tmp1_m); \ 1606 #define TRANSPOSE16x8_UB_UB(in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, \ argument 1613 ILVEV_D2_UB(in0, in8, in1, in9, out7, out6); \
|
/external/deqp/external/vulkancts/data/vulkan/glsl/es310/ |
D | conversions.test | 12626 input bool in8 = [ true | true | true | false | true | false | true | false | false | false ]; 12641 out0 = mat3x4(in0, in1, in2, in3, in4, in5, in6, in7, in8, in9); 12964 input bool in8 = [ true | false | false | true | true | false | true | true | false | false ]; 12979 out0 = mat4x3(in0, in1, in2, in3, in4, in5, in6, in7, in8, in9);
|
/external/deqp/data/gles3/shaders/ |
D | conversions.test | 13686 input bool in8 = [ true | true | true | false | true | false | true | false | false | false ]; 13701 out0 = mat3x4(in0, in1, in2, in3, in4, in5, in6, in7, in8, in9); 14024 input bool in8 = [ true | false | false | true | true | false | true | true | false | false ]; 14039 out0 = mat4x3(in0, in1, in2, in3, in4, in5, in6, in7, in8, in9);
|
/external/boringssl/src/crypto/fipsmodule/aes/asm/ |
D | aesni-x86_64.pl | 1066 movaps %xmm8,0x20(%rsp) # $in8
|