/external/libvpx/libvpx/vpx_dsp/ppc/ |
D | inv_txfm_vsx.c | 155 t1 = vec_sub(in0, in1); \ 161 temp3 = vec_sub(vec_mule(tmp16_0, cospi24_v), vec_mulo(tmp16_0, cospi8_v)); \ 169 out1 = vec_sub(step0, step1); \ 266 temp10 = vec_sub(vec_mule(tmp16_0, cospi0), vec_mulo(tmp16_0, cospi1)); \ 267 temp11 = vec_sub(vec_mule(tmp16_1, cospi0), vec_mulo(tmp16_1, cospi1)); \ 278 tmp16_2 = vec_sub(inpt0, inpt1); \ 307 in5 = vec_sub(step4, step5); \ 308 in6 = vec_sub(step7, step6); \ 314 step2 = vec_sub(in1, in2); \ 315 step3 = vec_sub(in0, in3); \ [all …]
|
D | hadamard_vsx.c | 18 const int16x8_t b1 = vec_sub(v[0], v[1]); in vpx_hadamard_s16_8x8_one_pass() 20 const int16x8_t b3 = vec_sub(v[2], v[3]); in vpx_hadamard_s16_8x8_one_pass() 22 const int16x8_t b5 = vec_sub(v[4], v[5]); in vpx_hadamard_s16_8x8_one_pass() 24 const int16x8_t b7 = vec_sub(v[6], v[7]); in vpx_hadamard_s16_8x8_one_pass() 28 const int16x8_t c2 = vec_sub(b0, b2); in vpx_hadamard_s16_8x8_one_pass() 29 const int16x8_t c3 = vec_sub(b1, b3); in vpx_hadamard_s16_8x8_one_pass() 32 const int16x8_t c6 = vec_sub(b4, b6); in vpx_hadamard_s16_8x8_one_pass() 33 const int16x8_t c7 = vec_sub(b5, b7); in vpx_hadamard_s16_8x8_one_pass() 36 v[1] = vec_sub(c2, c6); in vpx_hadamard_s16_8x8_one_pass() 37 v[2] = vec_sub(c0, c4); in vpx_hadamard_s16_8x8_one_pass() [all …]
|
D | fdct32x32_vsx.c | 31 const int32x4_t diff_e = vec_sub(ac_e, bc_e); in single_butterfly() 32 const int32x4_t diff_o = vec_sub(ac_o, bc_o); in single_butterfly() 66 const int32x4_t diff_o = vec_sub(ac2_o, bc1_o); in double_butterfly() 67 const int32x4_t diff_e = vec_sub(ac2_e, bc1_e); in double_butterfly() 216 return vec_sra(vec_sub(vec_add(a, vec_twos_s16), sign), vec_dct_scale_log2); in sub_round_shift() 236 temp1[i + 16] = vec_sub(in[15 - i], in[i + 16]); in vpx_fdct32_vsx() 243 temp1[i + 8] = vec_sub(temp0[7 - i], temp0[i + 8]); in vpx_fdct32_vsx() 295 temp2[5] = vec_sub(temp1[2], temp1[5]); in vpx_fdct32_vsx() 296 temp2[6] = vec_sub(temp1[1], temp1[6]); in vpx_fdct32_vsx() 308 temp2[20] = vec_sub(temp1[19], temp1[20]); in vpx_fdct32_vsx() [all …]
|
D | subtract_vsx.c | 34 const int16x8_t da = vec_sub(unpack_to_s16_h(s0), unpack_to_s16_h(p0)); in subtract_block4x4() 35 const int16x8_t db = vec_sub(unpack_to_s16_h(s1), unpack_to_s16_h(p1)); in subtract_block4x4() 59 vec_sub(unpack_to_s16_l(s0), unpack_to_s16_l(p0)); in vpx_subtract_block_vsx() 61 vec_sub(unpack_to_s16_h(s0), unpack_to_s16_h(p0)); in vpx_subtract_block_vsx() 63 vec_sub(unpack_to_s16_l(s1), unpack_to_s16_l(p1)); in vpx_subtract_block_vsx() 65 vec_sub(unpack_to_s16_h(s1), unpack_to_s16_h(p1)); in vpx_subtract_block_vsx() 80 const int16x8_t d0l = vec_sub(unpack_to_s16_l(s0), unpack_to_s16_l(p0)); in vpx_subtract_block_vsx() 81 const int16x8_t d0h = vec_sub(unpack_to_s16_h(s0), unpack_to_s16_h(p0)); in vpx_subtract_block_vsx() 93 const int16x8_t d0h = vec_sub(unpack_to_s16_h(s0), unpack_to_s16_h(p0)); in vpx_subtract_block_vsx()
|
D | intrapred_vsx.c | 246 val = vec_sub(vec_add(vec_splat(l, 0), a), tl); in vpx_tm_predictor_4x4_vsx() 252 val = vec_sub(vec_add(vec_splat(l, 1), a), tl); in vpx_tm_predictor_4x4_vsx() 258 val = vec_sub(vec_add(vec_splat(l, 2), a), tl); in vpx_tm_predictor_4x4_vsx() 264 val = vec_sub(vec_add(vec_splat(l, 3), a), tl); in vpx_tm_predictor_4x4_vsx() 276 val = vec_sub(vec_add(vec_splat(l, 0), a), tl); in vpx_tm_predictor_8x8_vsx() 281 val = vec_sub(vec_add(vec_splat(l, 1), a), tl); in vpx_tm_predictor_8x8_vsx() 286 val = vec_sub(vec_add(vec_splat(l, 2), a), tl); in vpx_tm_predictor_8x8_vsx() 291 val = vec_sub(vec_add(vec_splat(l, 3), a), tl); in vpx_tm_predictor_8x8_vsx() 296 val = vec_sub(vec_add(vec_splat(l, 4), a), tl); in vpx_tm_predictor_8x8_vsx() 301 val = vec_sub(vec_add(vec_splat(l, 5), a), tl); in vpx_tm_predictor_8x8_vsx() [all …]
|
D | variance_vsx.c | 27 const int16x8_t d0 = vec_sub(a0, b0); in vpx_get4x4sse_cs_vsx() 28 const int16x8_t d1 = vec_sub(a1, b1); in vpx_get4x4sse_cs_vsx() 120 const int16x8_t d0 = vec_sub(a0, b0); in variance_inner_32() 121 const int16x8_t d1 = vec_sub(a1, b1); in variance_inner_32() 122 const int16x8_t d2 = vec_sub(a2, b2); in variance_inner_32() 123 const int16x8_t d3 = vec_sub(a3, b3); in variance_inner_32() 150 const int16x8_t d = vec_sub(a0, b0); in variance() 161 const int16x8_t d = vec_sub(a0, b0); in variance() 177 const int16x8_t d0 = vec_sub(a0, b0); in variance() 178 const int16x8_t d1 = vec_sub(a1, b1); in variance()
|
D | deblock_vsx.c | 182 const int32x4_t thres_odd = vec_sub(sumsq_odd_scaled, vec_mulo(sum, sum)); in mask_s16() 183 const int32x4_t thres_even = vec_sub(sumsq_even_scaled, vec_mule(sum, sum)); in mask_s16() 235 const int16x8_t x = vec_sub(right_ctx, left_ctx); in vpx_mbpost_proc_across_ip_vsx() 237 vec_sub(vec_mule(right_ctx, right_ctx), vec_mule(left_ctx, left_ctx)); in vpx_mbpost_proc_across_ip_vsx() 239 vec_sub(vec_mulo(right_ctx, right_ctx), vec_mulo(left_ctx, left_ctx)); in vpx_mbpost_proc_across_ip_vsx() 340 sum = vec_add(sum, vec_sub(window[15], window[0])); in vpx_mbpost_proc_down_vsx() 350 sumsq_odd = vec_add(sumsq_odd, vec_sub(d15_odd, d0_odd)); in vpx_mbpost_proc_down_vsx() 351 sumsq_even = vec_add(sumsq_even, vec_sub(d15_even, d0_even)); in vpx_mbpost_proc_down_vsx()
|
D | sad_vsx.c | 159 v_subh = vec_sub(v_h, v_bh); \ 160 v_subl = vec_sub(v_l, v_bl); \
|
D | types_vsx.h | 79 #define vec_absd(a, b) vec_sub(vec_max(a, b), vec_min(a, b))
|
/external/libjpeg-turbo/simd/powerpc/ |
D | jquanti-altivec.c | 81 out0 = vec_sub(out0, pw_centerjsamp); in jsimd_convsamp_altivec() 82 out1 = vec_sub(out1, pw_centerjsamp); in jsimd_convsamp_altivec() 83 out2 = vec_sub(out2, pw_centerjsamp); in jsimd_convsamp_altivec() 84 out3 = vec_sub(out3, pw_centerjsamp); in jsimd_convsamp_altivec() 85 out4 = vec_sub(out4, pw_centerjsamp); in jsimd_convsamp_altivec() 86 out5 = vec_sub(out5, pw_centerjsamp); in jsimd_convsamp_altivec() 87 out6 = vec_sub(out6, pw_centerjsamp); in jsimd_convsamp_altivec() 88 out7 = vec_sub(out7, pw_centerjsamp); in jsimd_convsamp_altivec() 162 row0 = vec_sub(row0, row0s); in jsimd_quantize_altivec() 163 row1 = vec_sub(row1, row1s); in jsimd_quantize_altivec() [all …]
|
D | jfdctfst-altivec.c | 49 tmp13 = vec_sub(tmp0, tmp3); \ 51 tmp12 = vec_sub(tmp1, tmp2); \ 54 out4 = vec_sub(tmp10, tmp11); \ 61 out6 = vec_sub(tmp13, z1); \ 71 z5 = vec_sub(tmp10, tmp12); \ 81 z13 = vec_sub(tmp7, z3); \ 84 out3 = vec_sub(z13, z2); \ 86 out7 = vec_sub(z11, z4); \ 121 tmp7 = vec_sub(col0, col7); in jsimd_fdct_ifast_altivec() 123 tmp6 = vec_sub(col1, col6); in jsimd_fdct_ifast_altivec() [all …]
|
D | jidctfst-altivec.c | 51 tmp11 = vec_sub(in##0, in##4); \ 54 tmp12 = vec_sub(in##2, in##6); \ 57 tmp12 = vec_sub(tmp12, tmp13); \ 60 tmp3 = vec_sub(tmp10, tmp13); \ 62 tmp2 = vec_sub(tmp11, tmp12); \ 67 z10 = vec_sub(in##5, in##3); \ 70 z12s = vec_sub(in##1, in##7); \ 73 tmp11 = vec_sub(z11, z13); \ 93 tmp10 = vec_sub(tmp10, z5); \ 95 tmp12 = vec_sub(tmp12, z10); \ [all …]
|
D | jfdctint-altivec.c | 151 tmp13 = vec_sub(tmp0, tmp3); \ 153 tmp12 = vec_sub(tmp1, tmp2); \ 157 out4 = vec_sub(tmp10, tmp11); \ 167 tmp13 = vec_sub(tmp0, tmp3); \ 169 tmp12 = vec_sub(tmp1, tmp2); \ 174 out4 = vec_sub(tmp10, tmp11); \ 225 tmp7 = vec_sub(col0, col7); in jsimd_fdct_islow_altivec() 227 tmp6 = vec_sub(col1, col6); in jsimd_fdct_islow_altivec() 229 tmp5 = vec_sub(col2, col5); in jsimd_fdct_islow_altivec() 231 tmp4 = vec_sub(col3, col4); in jsimd_fdct_islow_altivec() [all …]
|
D | jidctint-altivec.c | 69 tmp1 = vec_sub(in##0, in##4); \ 80 tmp13l = vec_sub(tmp0l, tmp3l); \ 81 tmp13h = vec_sub(tmp0h, tmp3h); \ 92 tmp12l = vec_sub(tmp1l, tmp2l); \ 93 tmp12h = vec_sub(tmp1h, tmp2h); \ 155 out7l = vec_sub(tmp10l, tmp3l); \ 156 out7h = vec_sub(tmp10h, tmp3h); \ 168 out6l = vec_sub(tmp11l, tmp2l); \ 169 out6h = vec_sub(tmp11h, tmp2h); \ 181 out5l = vec_sub(tmp12l, tmp1l); \ [all …]
|
D | jdcolext-altivec.c | 92 cbl = vec_sub(cbl, pw_cj); in jsimd_ycc_rgb_convert_altivec() 93 cbh = vec_sub(cbh, pw_cj); in jsimd_ycc_rgb_convert_altivec() 98 crl = vec_sub(crl, pw_cj); in jsimd_ycc_rgb_convert_altivec() 99 crh = vec_sub(crh, pw_cj); in jsimd_ycc_rgb_convert_altivec() 150 gl = vec_sub(gl, crl); in jsimd_ycc_rgb_convert_altivec() 151 gh = vec_sub(gh, crh); in jsimd_ycc_rgb_convert_altivec()
|
D | jdmrgext-altivec.c | 94 cbl = vec_sub(cbl, pw_cj); in jsimd_h2v1_merged_upsample_altivec() 95 cbh = vec_sub(cbh, pw_cj); in jsimd_h2v1_merged_upsample_altivec() 100 crl = vec_sub(crl, pw_cj); in jsimd_h2v1_merged_upsample_altivec() 101 crh = vec_sub(crh, pw_cj); in jsimd_h2v1_merged_upsample_altivec() 150 g_yl = vec_sub(g_yl, crl); in jsimd_h2v1_merged_upsample_altivec() 151 g_yh = vec_sub(g_yh, crh); in jsimd_h2v1_merged_upsample_altivec()
|
/external/libaom/libaom/av1/common/ppc/ |
D | cfl_ppc.c | 76 vec_vsx_st(vec_sub(vec_vsx_ld(OFF_0, dst), vec_avg), OFF_0, dst); in subtract_average_vsx() 77 vec_vsx_st(vec_sub(vec_vsx_ld(OFF_0 + CFL_LINE_1, dst), vec_avg), in subtract_average_vsx() 79 vec_vsx_st(vec_sub(vec_vsx_ld(OFF_0 + CFL_LINE_2, dst), vec_avg), in subtract_average_vsx() 81 vec_vsx_st(vec_sub(vec_vsx_ld(OFF_0 + CFL_LINE_3, dst), vec_avg), in subtract_average_vsx() 84 vec_vsx_st(vec_sub(vec_vsx_ld(OFF_1, dst), vec_avg), OFF_1, dst); in subtract_average_vsx() 85 vec_vsx_st(vec_sub(vec_vsx_ld(OFF_1 + CFL_LINE_1, dst), vec_avg), in subtract_average_vsx() 87 vec_vsx_st(vec_sub(vec_vsx_ld(OFF_1 + CFL_LINE_2, dst), vec_avg), in subtract_average_vsx() 89 vec_vsx_st(vec_sub(vec_vsx_ld(OFF_1 + CFL_LINE_3, dst), vec_avg), in subtract_average_vsx() 93 vec_vsx_st(vec_sub(vec_vsx_ld(OFF_2, dst), vec_avg), OFF_2, dst); in subtract_average_vsx() 94 vec_vsx_st(vec_sub(vec_vsx_ld(OFF_2 + CFL_LINE_1, dst), vec_avg), in subtract_average_vsx() [all …]
|
/external/libpng/powerpc/ |
D | filter_vsx_intrinsics.c | 349 avg_vec = vec_sub(avg_vec, vec_and(vec_xor(rp_part_vec,pp_part_vec),vec_splat_u8(1))); in png_read_filter_row_avg4_vsx() 355 avg_vec = vec_sub(avg_vec, vec_and(vec_xor(rp_part_vec,pp_part_vec),vec_splat_u8(1))); in png_read_filter_row_avg4_vsx() 361 avg_vec = vec_sub(avg_vec, vec_and(vec_xor(rp_part_vec,pp_part_vec),vec_splat_u8(1))); in png_read_filter_row_avg4_vsx() 436 avg_vec = vec_sub(avg_vec, vec_and(vec_xor(rp_part_vec,pp_part_vec),vec_splat_u8(1))); in png_read_filter_row_avg3_vsx() 442 avg_vec = vec_sub(avg_vec, vec_and(vec_xor(rp_part_vec,pp_part_vec),vec_splat_u8(1))); in png_read_filter_row_avg3_vsx() 448 avg_vec = vec_sub(avg_vec, vec_and(vec_xor(rp_part_vec,pp_part_vec),vec_splat_u8(1))); in png_read_filter_row_avg3_vsx() 454 avg_vec = vec_sub(avg_vec, vec_and(vec_xor(rp_part_vec,pp_part_vec),vec_splat_u8(1))); in png_read_filter_row_avg3_vsx() 545 pa_vec = (vector signed short) vec_sub(b_vec,c_vec); in png_read_filter_row_paeth4_vsx() 546 pb_vec = (vector signed short) vec_sub(a_vec , c_vec); in png_read_filter_row_paeth4_vsx() 566 pa_vec = (vector signed short) vec_sub(b_vec,c_vec); in png_read_filter_row_paeth4_vsx() [all …]
|
/external/clang/test/CodeGen/ |
D | builtins-ppc-quadword.c | 95 res_vlll = vec_sub(vlll, vlll); in test1() 100 res_vulll = vec_sub(vulll, vulll); in test1()
|
D | builtins-ppc-p8vector.c | 1458 res_vsll = vec_sub(vsll, vsll); in test1() 1462 res_vull = vec_sub(vull, vull); in test1() 1466 res_vd = vec_sub(vda, vda); in test1() 1470 res_vsx = vec_sub(vsx, vsx); in test1() 1474 res_vux = vec_sub(vux, vux); in test1()
|
D | builtins-ppc-altivec.c | 5086 res_vsc = vec_sub(vsc, vsc); in test6() 5090 res_vsc = vec_sub(vbc, vsc); in test6() 5094 res_vsc = vec_sub(vsc, vbc); in test6() 5098 res_vuc = vec_sub(vuc, vuc); in test6() 5102 res_vuc = vec_sub(vbc, vuc); in test6() 5106 res_vuc = vec_sub(vuc, vbc); in test6() 5110 res_vs = vec_sub(vs, vs); in test6() 5114 res_vs = vec_sub(vbs, vs); in test6() 5118 res_vs = vec_sub(vs, vbs); in test6() 5122 res_vus = vec_sub(vus, vus); in test6() [all …]
|
/external/libunwind/tests/ |
D | ppc64-test-altivec.c | 130 vec_nor (v1, vec_add (v2, vec_sub (v3, vec_and (v4, vec_or (v5, v6))))); in vec_stack() 145 vec_sub (v3, in vec_stack()
|
/external/fec/ |
D | viterbi27_av.c | 138 m_metric = vec_sub((vector unsigned char)(31),metric); in update_viterbi27_blk_av() 149 m_metric = vec_sub((vector unsigned char)(31),metric); in update_viterbi27_blk_av()
|
D | viterbi39_av.c | 143 m_metric = vec_sub((vector unsigned short)(765),metric); in update_viterbi39_blk_av() 166 …decisions = vec_sub(decisions,(vector unsigned char)vec_pack(vec_mergeh(decision0,decision1),vec_m… in update_viterbi39_blk_av()
|
D | viterbi615_av.c | 148 m_metric = vec_sub((vector unsigned short)(1530),metric); in update_viterbi615_blk_av() 171 …decisions = vec_sub(decisions,(vector unsigned char)vec_pack(vec_mergeh(decision0,decision1),vec_m… in update_viterbi615_blk_av()
|