Home
last modified time | relevance | path

Searched refs:src0 (Results 1 – 25 of 478) sorted by relevance

12345678910>>...20

/external/mesa3d/prebuilt-intermediates/nir/
Dnir_builder_opcodes.h29 nir_amul(nir_builder *build, nir_ssa_def *src0, nir_ssa_def *src1) in nir_amul() argument
31 return nir_build_alu(build, nir_op_amul, src0, src1, NULL, NULL); in nir_amul()
34 nir_b16all_fequal16(nir_builder *build, nir_ssa_def *src0, nir_ssa_def *src1) in nir_b16all_fequal16() argument
36 return nir_build_alu(build, nir_op_b16all_fequal16, src0, src1, NULL, NULL); in nir_b16all_fequal16()
39 nir_b16all_fequal2(nir_builder *build, nir_ssa_def *src0, nir_ssa_def *src1) in nir_b16all_fequal2() argument
41 return nir_build_alu(build, nir_op_b16all_fequal2, src0, src1, NULL, NULL); in nir_b16all_fequal2()
44 nir_b16all_fequal3(nir_builder *build, nir_ssa_def *src0, nir_ssa_def *src1) in nir_b16all_fequal3() argument
46 return nir_build_alu(build, nir_op_b16all_fequal3, src0, src1, NULL, NULL); in nir_b16all_fequal3()
49 nir_b16all_fequal4(nir_builder *build, nir_ssa_def *src0, nir_ssa_def *src1) in nir_b16all_fequal4() argument
51 return nir_build_alu(build, nir_op_b16all_fequal4, src0, src1, NULL, NULL); in nir_b16all_fequal4()
[all …]
Dnir_constant_expressions.c576 const int1_t src0 = -(int1_t)_src[0][_i].b; in evaluate_amul() local
580 int1_t dst = src0 * src1; in evaluate_amul()
595 const int8_t src0 = in evaluate_amul() local
600 int8_t dst = src0 * src1; in evaluate_amul()
614 const int16_t src0 = in evaluate_amul() local
619 int16_t dst = src0 * src1; in evaluate_amul()
633 const int32_t src0 = in evaluate_amul() local
638 int32_t dst = src0 * src1; in evaluate_amul()
652 const int64_t src0 = in evaluate_amul() local
657 int64_t dst = src0 * src1; in evaluate_amul()
[all …]
/external/rust/crates/libz-sys/src/zlib-ng/
Dcompare256.c10 static inline uint32_t compare256_c_static(const uint8_t *src0, const uint8_t *src1) { in compare256_c_static() argument
14 if (*src0 != *src1) in compare256_c_static()
16 src0 += 1, src1 += 1, len += 1; in compare256_c_static()
17 if (*src0 != *src1) in compare256_c_static()
19 src0 += 1, src1 += 1, len += 1; in compare256_c_static()
20 if (*src0 != *src1) in compare256_c_static()
22 src0 += 1, src1 += 1, len += 1; in compare256_c_static()
23 if (*src0 != *src1) in compare256_c_static()
25 src0 += 1, src1 += 1, len += 1; in compare256_c_static()
26 if (*src0 != *src1) in compare256_c_static()
[all …]
/external/mesa3d/prebuilt-intermediates/bifrost/
Dbi_generated_pack.h35 unsigned src0 = bi_get_src(ins, regs, 0); in pan_pack_fma_rshift_and_i32() local
36 assert((1 << src0) & 0xfb); in pan_pack_fma_rshift_and_i32()
57 …return 0x301000 | (src0 << 0) | (src1 << 3) | (src2 << 6) | (lane2 << 9) | (not1 << 14) | (not_res… in pan_pack_fma_rshift_and_i32()
63 unsigned src0 = bi_get_src(ins, regs, 0); in pan_pack_add_iadd_u32() local
87 return 0xbc600 | (src0 << 0) | (src1 << 3) | (saturate << 8) | (derived_7 << 7); in pan_pack_add_iadd_u32()
99 …return 0xbec00 | (src0 << 0) | (src1 << 3) | (saturate << 8) | (derived_7 << 7) | (derived_9 << 9); in pan_pack_add_iadd_u32()
113 …return 0xbe000 | (src0 << 0) | (src1 << 3) | (saturate << 8) | (derived_7 << 7) | (derived_9 << 9); in pan_pack_add_iadd_u32()
122 unsigned src0 = bi_get_src(ins, regs, 0); in pan_pack_add_ld_var_flat() local
150 …return 0x538c0 | (src0 << 3) | (vecsize << 8) | (function << 0) | (derived_10 << 10) | (derived_19… in pan_pack_add_ld_var_flat()
152 return 0xcf8c0 | (src0 << 3) | (vecsize << 8) | (function << 0); in pan_pack_add_ld_var_flat()
[all …]
/external/ComputeLibrary/src/cpu/operators/
DCpuElementwise.cpp49 void CpuElementwiseArithmetic<op>::configure(const ITensorInfo *src0, const ITensorInfo *src1, ITen… in configure() argument
51 ARM_COMPUTE_LOG_PARAMS(src0, src1, dst); in configure()
53 k->configure(op, src0, src1, dst); in configure()
58 Status CpuElementwiseArithmetic<op>::validate(const ITensorInfo *src0, const ITensorInfo *src1, con… in validate() argument
60 return kernels::CpuArithmeticKernel::validate(op, src0, src1, dst); in validate()
68 void CpuElementwiseDivision::configure(const ITensorInfo *src0, const ITensorInfo *src1, ITensorInf… in configure() argument
70 ARM_COMPUTE_LOG_PARAMS(src0, src1, dst); in configure()
72 k->configure(src0, src1, dst); in configure()
76 Status CpuElementwiseDivision::validate(const ITensorInfo *src0, const ITensorInfo *src1, const ITe… in validate() argument
78 return kernels::CpuDivisionKernel::validate(src0, src1, dst); in validate()
[all …]
/external/libvpx/vpx_dsp/mips/
Dsum_squares_msa.c22 uint64_t src0, src1, src2, src3; in vpx_sum_squares_2d_i16_msa() local
26 LD4(src, src_stride, src0, src1, src2, src3); in vpx_sum_squares_2d_i16_msa()
27 INSERT_D2_SH(src0, src1, diff0); in vpx_sum_squares_2d_i16_msa()
35 v8i16 src0, src1, src2, src3, src4, src5, src6, src7; in vpx_sum_squares_2d_i16_msa() local
37 LD_SH8(src, src_stride, src0, src1, src2, src3, src4, src5, src6, src7); in vpx_sum_squares_2d_i16_msa()
38 DOTP_SH2_SW(src0, src1, src0, src1, mul0, mul1); in vpx_sum_squares_2d_i16_msa()
47 v8i16 src0, src1, src2, src3, src4, src5, src6, src7; in vpx_sum_squares_2d_i16_msa() local
49 LD_SH8(src, src_stride, src0, src1, src2, src3, src4, src5, src6, src7); in vpx_sum_squares_2d_i16_msa()
50 DOTP_SH2_SW(src0, src1, src0, src1, mul0, mul1); in vpx_sum_squares_2d_i16_msa()
54 LD_SH8(src + 8, src_stride, src0, src1, src2, src3, src4, src5, src6, src7); in vpx_sum_squares_2d_i16_msa()
[all …]
Dvariance_msa.c45 uint32_t src0, src1, src2, src3; in sse_diff_4width_msa() local
54 LW4(src_ptr, src_stride, src0, src1, src2, src3); in sse_diff_4width_msa()
59 INSERT_W4_UB(src0, src1, src2, src3, src); in sse_diff_4width_msa()
74 v16u8 src0, src1, src2, src3; in sse_diff_8width_msa() local
80 LD_UB4(src_ptr, src_stride, src0, src1, src2, src3); in sse_diff_8width_msa()
85 PCKEV_D4_UB(src1, src0, src3, src2, ref1, ref0, ref3, ref2, src0, src1, in sse_diff_8width_msa()
87 CALC_MSE_AVG_B(src0, ref0, var, avg); in sse_diff_8width_msa()
141 v16u8 src0, src1, ref0, ref1; in sse_diff_32width_msa() local
146 LD_UB2(src_ptr, 16, src0, src1); in sse_diff_32width_msa()
150 CALC_MSE_AVG_B(src0, ref0, var, avg); in sse_diff_32width_msa()
[all …]
Dvpx_convolve_copy_msa.c19 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; in copy_width8_msa() local
23 LD_UB8(src, src_stride, src0, src1, src2, src3, src4, src5, src6, src7); in copy_width8_msa()
26 out0 = __msa_copy_u_d((v2i64)src0, 0); in copy_width8_msa()
40 LD_UB4(src, src_stride, src0, src1, src2, src3); in copy_width8_msa()
43 out0 = __msa_copy_u_d((v2i64)src0, 0); in copy_width8_msa()
52 LD_UB8(src, src_stride, src0, src1, src2, src3, src4, src5, src6, src7); in copy_width8_msa()
55 out0 = __msa_copy_u_d((v2i64)src0, 0); in copy_width8_msa()
71 LD_UB4(src, src_stride, src0, src1, src2, src3); in copy_width8_msa()
73 out0 = __msa_copy_u_d((v2i64)src0, 0); in copy_width8_msa()
83 LD_UB2(src, src_stride, src0, src1); in copy_width8_msa()
[all …]
Dvpx_convolve8_horiz_msa.c19 v16i8 src0, src1, src2, src3, filt0, filt1, filt2, filt3; in common_hz_8t_4x4_msa() local
33 LD_SB4(src, src_stride, src0, src1, src2, src3); in common_hz_8t_4x4_msa()
34 XORI_B4_128_SB(src0, src1, src2, src3); in common_hz_8t_4x4_msa()
35 HORIZ_8TAP_4WID_4VECS_FILT(src0, src1, src2, src3, mask0, mask1, mask2, mask3, in common_hz_8t_4x4_msa()
47 v16i8 src0, src1, src2, src3; in common_hz_8t_4x8_msa() local
62 LD_SB4(src, src_stride, src0, src1, src2, src3); in common_hz_8t_4x8_msa()
63 XORI_B4_128_SB(src0, src1, src2, src3); in common_hz_8t_4x8_msa()
65 HORIZ_8TAP_4WID_4VECS_FILT(src0, src1, src2, src3, mask0, mask1, mask2, mask3, in common_hz_8t_4x8_msa()
67 LD_SB4(src, src_stride, src0, src1, src2, src3); in common_hz_8t_4x8_msa()
68 XORI_B4_128_SB(src0, src1, src2, src3); in common_hz_8t_4x8_msa()
[all …]
/external/libaom/av1/common/x86/
Dreconinter_avx2.c31 const uint8_t *src0, int src0_stride, in av1_build_compound_diffwtd_mask_avx2() argument
39 const __m128i s0A = xx_loadl_32(src0); in av1_build_compound_diffwtd_mask_avx2()
40 const __m128i s0B = xx_loadl_32(src0 + src0_stride); in av1_build_compound_diffwtd_mask_avx2()
41 const __m128i s0C = xx_loadl_32(src0 + src0_stride * 2); in av1_build_compound_diffwtd_mask_avx2()
42 const __m128i s0D = xx_loadl_32(src0 + src0_stride * 3); in av1_build_compound_diffwtd_mask_avx2()
61 src0 += (src0_stride << 2); in av1_build_compound_diffwtd_mask_avx2()
68 const __m128i s0A = xx_loadl_64(src0); in av1_build_compound_diffwtd_mask_avx2()
69 const __m128i s0B = xx_loadl_64(src0 + src0_stride); in av1_build_compound_diffwtd_mask_avx2()
70 const __m128i s0C = xx_loadl_64(src0 + src0_stride * 2); in av1_build_compound_diffwtd_mask_avx2()
71 const __m128i s0D = xx_loadl_64(src0 + src0_stride * 3); in av1_build_compound_diffwtd_mask_avx2()
[all …]
/external/ComputeLibrary/src/cpu/kernels/
DCpuDirectConv3dKernel.cpp78 Status validate_arguments(const ITensorInfo *src0, const ITensorInfo *src1, const ITensorInfo *src2… in validate_arguments() argument
80 ARM_COMPUTE_RETURN_ERROR_ON_NULLPTR(src0, src1, dst); in validate_arguments()
81 ARM_COMPUTE_RETURN_ERROR_ON(src0->data_layout() != DataLayout::NDHWC); in validate_arguments()
82 ARM_COMPUTE_RETURN_ERROR_ON_MISMATCHING_DATA_LAYOUT(src0, src1, dst); in validate_arguments()
83 ARM_COMPUTE_RETURN_ERROR_ON_CPU_F16_UNSUPPORTED(src0); in validate_arguments()
84 …ARM_COMPUTE_RETURN_ERROR_ON_DATA_TYPE_CHANNEL_NOT_IN(src0, 1, DataType::F16, DataType::F32, DataTy… in validate_arguments()
85 ARM_COMPUTE_RETURN_ERROR_ON_MISMATCHING_DATA_TYPES(src0, src1); in validate_arguments()
88 …const auto *uk = CpuDirectConv3dKernel::get_implementation(DataTypeISASelectorData{ src0->data_typ… in validate_arguments()
92 const DataLayout data_layout = src0->data_layout(); in validate_arguments()
97 ARM_COMPUTE_RETURN_ERROR_ON(src1->dimension(1) != src0->dimension(channel_idx)); in validate_arguments()
[all …]
DCpuElementwiseKernel.cpp301 Status CpuElementwiseKernel<Derived>::validate_arguments_common(const ITensorInfo &src0, const ITen… in validate_arguments_common() argument
303 ARM_COMPUTE_RETURN_ERROR_ON_CPU_F16_UNSUPPORTED(&src0); in validate_arguments_common()
304 ARM_COMPUTE_RETURN_ERROR_ON_MISMATCHING_DATA_TYPES(&src0, &src1); in validate_arguments_common()
306 …const TensorShape out_shape = TensorShape::broadcast_shape(src0.tensor_shape(), src1.tensor_shape(… in validate_arguments_common()
320 void CpuArithmeticKernel::configure_common(const ITensorInfo *src0, const ITensorInfo *src1, ITenso… in configure_common() argument
322 ARM_COMPUTE_ERROR_ON_NULLPTR(src0, src1, dst); in configure_common()
324 …uArithmeticKernel::get_implementation(ElementwiseDataTypeISASelectorData{ src0->data_type(), CPUIn… in configure_common()
332 if(src0->is_dynamic() || src1->is_dynamic()) in configure_common()
337 …auto shape_and_window = compute_output_shape_and_window(src0->tensor_shape(), src1->tensor_shape()… in configure_common()
338 auto_init_if_empty(*dst, shape_and_window.first, 1, src0->data_type()); in configure_common()
[all …]
DCpuSubKernel.cpp108 inline Status validate_arguments(const ITensorInfo &src0, const ITensorInfo &src1, const ITensorInf… in validate_arguments() argument
111 ARM_COMPUTE_RETURN_ERROR_ON_CPU_F16_UNSUPPORTED(&src0); in validate_arguments()
112 …ARM_COMPUTE_RETURN_ERROR_ON_DATA_TYPE_CHANNEL_NOT_IN(&src0, 1, DataType::U8, DataType::QASYMM8, Da… in validate_arguments()
114 ARM_COMPUTE_RETURN_ERROR_ON_MISMATCHING_DATA_TYPES(&src0, &src1); in validate_arguments()
116 const auto can_use_fixedpoint = sub_q8_neon_fixedpoint_possible(&src0, &src1, &dst); in validate_arguments()
117 …CpuSubKernelDataTypeISASelectorData>(CpuSubKernelDataTypeISASelectorData{ src0.data_type(), CPUInf… in validate_arguments()
121 …const TensorShape out_shape = TensorShape::broadcast_shape(src0.tensor_shape(), src1.tensor_shape(… in validate_arguments()
124 …ARM_COMPUTE_RETURN_ERROR_ON_MSG(is_data_type_quantized(src0.data_type()) && (policy == ConvertPoli… in validate_arguments()
130 ARM_COMPUTE_RETURN_ERROR_ON_MISMATCHING_DATA_TYPES(&src0, &dst); in validate_arguments()
138 void CpuSubKernel::configure(const ITensorInfo *src0, const ITensorInfo *src1, ITensorInfo *dst, Co… in configure() argument
[all …]
DCpuAddKernel.cpp179 Status validate_arguments(const ITensorInfo &src0, const ITensorInfo &src1, const ITensorInfo &dst,… in validate_arguments() argument
183 ARM_COMPUTE_RETURN_ERROR_ON_CPU_F16_UNSUPPORTED(&src0); in validate_arguments()
184 …ARM_COMPUTE_RETURN_ERROR_ON_DATA_TYPE_CHANNEL_NOT_IN(&src0, 1, DataType::U8, DataType::QASYMM8, Da… in validate_arguments()
187 ARM_COMPUTE_RETURN_ERROR_ON_MISMATCHING_DATA_TYPES(&src0, &src1); in validate_arguments()
189 …const TensorShape out_shape = TensorShape::broadcast_shape(src0.tensor_shape(), src1.tensor_shape(… in validate_arguments()
192 …UTE_RETURN_ERROR_ON_MSG((src0.tensor_shape().x() != src1.tensor_shape().x()) && ((src0.data_type()… in validate_arguments()
199 ARM_COMPUTE_RETURN_ERROR_ON_MISMATCHING_DATA_TYPES(&src0, &dst); in validate_arguments()
204 const auto can_use_fixedpoint = add_q8_neon_fixedpoint_possible(&src0, &src1, &dst); in validate_arguments()
205 …ntation<CpuAddKernelDataTypeISASelectorData>(CpuAddKernelDataTypeISASelectorData{ src0.data_type(), in validate_arguments()
213 void CpuAddKernel::configure(const ITensorInfo *src0, const ITensorInfo *src1, ITensorInfo *dst, Co… in configure() argument
[all …]
/external/libaom/aom_dsp/x86/
Dblend_a64_vmask_sse4.c31 const uint8_t *src0, uint32_t src0_stride, in blend_a64_vmask_w4_sse4_1() argument
42 const __m128i v_res_w = blend_4(src0, src1, &v_m0_w, &v_m1_w); in blend_a64_vmask_w4_sse4_1()
49 src0 += src0_stride; in blend_a64_vmask_w4_sse4_1()
56 const uint8_t *src0, uint32_t src0_stride, in blend_a64_vmask_w8_sse4_1() argument
67 const __m128i v_res_w = blend_8(src0, src1, &v_m0_w, &v_m1_w); in blend_a64_vmask_w8_sse4_1()
74 src0 += src0_stride; in blend_a64_vmask_w8_sse4_1()
81 const uint8_t *src0, in blend_a64_vmask_w16n_sse4_1() argument
93 const __m128i v_resl_w = blend_8(src0 + c, src1 + c, &v_m0_w, &v_m1_w); in blend_a64_vmask_w16n_sse4_1()
95 blend_8(src0 + c + 8, src1 + c + 8, &v_m0_w, &v_m1_w); in blend_a64_vmask_w16n_sse4_1()
102 src0 += src0_stride; in blend_a64_vmask_w16n_sse4_1()
[all …]
Dblend_a64_mask_sse4.c32 const uint8_t *src0, uint32_t src0_stride, in blend_a64_mask_w4_sse4_1() argument
42 const __m128i v_res_b = blend_4_u8(src0, src1, &v_m0_b, &v_m1_b, &_r); in blend_a64_mask_w4_sse4_1()
46 src0 += src0_stride; in blend_a64_mask_w4_sse4_1()
53 const uint8_t *src0, uint32_t src0_stride, in blend_a64_mask_w8_sse4_1() argument
63 const __m128i v_res_b = blend_8_u8(src0, src1, &v_m0_b, &v_m1_b, &_r); in blend_a64_mask_w8_sse4_1()
67 src0 += src0_stride; in blend_a64_mask_w8_sse4_1()
74 uint8_t *dst, uint32_t dst_stride, const uint8_t *src0, in blend_a64_mask_w16n_sse4_1() argument
87 blend_16_u8(src0 + c, src1 + c, &v_m0_b, &v_m1_b, &_r); in blend_a64_mask_w16n_sse4_1()
92 src0 += src0_stride; in blend_a64_mask_w16n_sse4_1()
103 uint8_t *dst, uint32_t dst_stride, const uint8_t *src0, in blend_a64_mask_sx_w4_sse4_1() argument
[all …]
Dblend_a64_mask_avx2.c29 uint8_t *dst, const CONV_BUF_TYPE *src0, const CONV_BUF_TYPE *src1, in blend_a64_d16_mask_w16_avx2() argument
33 const __m256i s0_0 = yy_loadu_256(src0); in blend_a64_d16_mask_w16_avx2()
50 uint8_t *dst, const CONV_BUF_TYPE *src0, const CONV_BUF_TYPE *src1, in blend_a64_d16_mask_w32_avx2() argument
55 const __m256i s0_0 = yy_loadu_256(src0); in blend_a64_d16_mask_w32_avx2()
56 const __m256i s0_1 = yy_loadu_256(src0 + 16); in blend_a64_d16_mask_w32_avx2()
83 uint8_t *dst, uint32_t dst_stride, const CONV_BUF_TYPE *src0, in lowbd_blend_a64_d16_mask_subw0_subh0_w16_avx2() argument
92 blend_a64_d16_mask_w16_avx2(dst, src0, src1, &m0, round_offset, &v_maxval, in lowbd_blend_a64_d16_mask_subw0_subh0_w16_avx2()
96 src0 += src0_stride; in lowbd_blend_a64_d16_mask_subw0_subh0_w16_avx2()
102 uint8_t *dst, uint32_t dst_stride, const CONV_BUF_TYPE *src0, in lowbd_blend_a64_d16_mask_subw0_subh0_w32_avx2() argument
113 blend_a64_d16_mask_w32_avx2(dst + j, src0 + j, src1 + j, &m0, &m1, in lowbd_blend_a64_d16_mask_subw0_subh0_w32_avx2()
[all …]
/external/webp/src/dsp/
Dlossless_enc_msa.c21 #define TRANSFORM_COLOR_8(src0, src1, dst0, dst1, c0, c1, mask0, mask1) do { \ argument
24 VSHF_B2_SH(src0, src0, src1, src1, mask0, mask0, g0, g1); \
27 t0 = __msa_subv_h((v8i16)src0, t0); \
29 t4 = __msa_srli_w((v4i32)src0, 16); \
34 VSHF_B2_UB(src0, t0, src1, t1, mask1, mask1, dst0, dst1); \
53 v16u8 src0, dst0; in TransformColor_MSA() local
64 LD_UB2(data, 4, src0, src1); in TransformColor_MSA()
65 TRANSFORM_COLOR_8(src0, src1, dst0, dst1, g2br, r2b, mask0, mask1); in TransformColor_MSA()
72 src0 = LD_UB(data); in TransformColor_MSA()
73 TRANSFORM_COLOR_4(src0, dst0, g2br, r2b, mask0, mask1); in TransformColor_MSA()
[all …]
Dlossless_msa.c25 v16u8 src0, src1, src2, src3, dst0, dst1, dst2; \
26 LD_UB4(psrc, 16, src0, src1, src2, src3); \
27 VSHF_B2_UB(src0, src1, src1, src2, m0, m1, dst0, dst1); \
35 v16u8 src0, src1, src2, dst0, dst1, dst2; \
36 LD_UB3(psrc, 16, src0, src1, src2); \
37 VSHF_B2_UB(src0, src1, src1, src2, m0, m1, dst0, dst1); \
46 v16u8 src0, src1, src2 = { 0 }, dst0, dst1; \
47 LD_UB2(psrc, 16, src0, src1); \
48 VSHF_B2_UB(src0, src1, src1, src2, m0, m1, dst0, dst1); \
55 const v16u8 src0 = LD_UB(psrc); \
[all …]
Drescaler_msa.c125 v4u32 src0, src1, src2, src3; in ExportRowExpand_0() local
127 LD_UW4(frow, 4, src0, src1, src2, src3); in ExportRowExpand_0()
128 CALC_MULT_FIX_16(src0, src1, src2, src3, scale, shift, out); in ExportRowExpand_0()
138 v4u32 src0, src1, src2; in ExportRowExpand_0() local
139 LD_UW3(frow, 4, src0, src1, src2); in ExportRowExpand_0()
140 CALC_MULT_FIX_4(src0, scale, shift, val0_m); in ExportRowExpand_0()
149 v4u32 src0, src1; in ExportRowExpand_0() local
150 LD_UW2(frow, 4, src0, src1); in ExportRowExpand_0()
151 CALC_MULT_FIX_4(src0, scale, shift, val0_m); in ExportRowExpand_0()
159 const v4u32 src0 = LD_UW(frow); in ExportRowExpand_0() local
[all …]
/external/mesa3d/src/freedreno/ir3/
Dir3_a4xx.c43 struct ir3_instruction *ldgb, *src0, *src1, *byte_offset, *offset; in emit_intrinsic_load_ssbo() local
51 src0 = ir3_create_collect(ctx, (struct ir3_instruction*[]){ in emit_intrinsic_load_ssbo()
58 src0, 0, src1, 0); in emit_intrinsic_load_ssbo()
74 struct ir3_instruction *stgb, *src0, *src1, *src2, *byte_offset, *offset; in emit_intrinsic_store_ssbo() local
88 src0 = ir3_create_collect(ctx, ir3_get_src(ctx, &intr->src[0]), ncomp); in emit_intrinsic_store_ssbo()
95 stgb = ir3_STGB(b, ssbo, 0, src0, 0, src1, 0, src2, 0); in emit_intrinsic_store_ssbo()
126 struct ir3_instruction *atomic, *ssbo, *src0, *src1, *src2, *byte_offset, in emit_intrinsic_atomic_ssbo() local
141 src0 = ir3_get_src(ctx, &intr->src[2])[0]; in emit_intrinsic_atomic_ssbo()
150 atomic = ir3_ATOMIC_ADD_G(b, ssbo, 0, src0, 0, src1, 0, src2, 0); in emit_intrinsic_atomic_ssbo()
153 atomic = ir3_ATOMIC_MIN_G(b, ssbo, 0, src0, 0, src1, 0, src2, 0); in emit_intrinsic_atomic_ssbo()
[all …]
/external/ComputeLibrary/src/gpu/cl/kernels/
DClGemmLowpMatrixMultiplyNativeKernel.cpp53 Status validate_arguments(const ITensorInfo *src0, const ITensorInfo *src1, const ITensorInfo *dst,… in validate_arguments() argument
56 ARM_COMPUTE_RETURN_ERROR_ON_NULLPTR(src0, src1, dst); in validate_arguments()
57 …ARM_COMPUTE_RETURN_ERROR_ON_DATA_TYPE_CHANNEL_NOT_IN(src0, 1, DataType::QASYMM8, DataType::QASYMM8… in validate_arguments()
58 if(src0->data_type() == DataType::QASYMM8) in validate_arguments()
60 ARM_COMPUTE_RETURN_ERROR_ON_MISMATCHING_DATA_TYPES(src0, src1); in validate_arguments()
66 …ARM_COMPUTE_RETURN_ERROR_ON_MSG(src0->num_dimensions() > 4, "The number of dimensions for the LHS … in validate_arguments()
83 ARM_COMPUTE_RETURN_ERROR_ON(src0->dimension(0) != static_cast<unsigned int>(k)); in validate_arguments()
88 …ARM_COMPUTE_RETURN_ERROR_ON(src0->dimension(1) * src0->dimension(2) != static_cast<unsigned int>(m… in validate_arguments()
92 ARM_COMPUTE_RETURN_ERROR_ON(src0->dimension(1) != static_cast<unsigned int>(m)); in validate_arguments()
97 … dst->clone()->set_tensor_shape(misc::shape_calculator::compute_mm_shape(*src0, *src1, gemm_info)); in validate_arguments()
[all …]
DClGemmMatrixMultiplyReshapedOnlyRhsMMULKernel.cpp58 Status validate_arguments(const ITensorInfo *src0, const ITensorInfo *src1, const ITensorInfo *src2… in validate_arguments() argument
63 ARM_COMPUTE_RETURN_ERROR_ON_NULLPTR(src0, src1, dst); in validate_arguments()
65 ARM_COMPUTE_RETURN_ERROR_ON_DATA_TYPE_CHANNEL_NOT_IN(src0, 1, DataType::F16, DataType::F32); in validate_arguments()
66 ARM_COMPUTE_RETURN_ERROR_ON_MISMATCHING_DATA_TYPES(src0, src1); in validate_arguments()
67 …ARM_COMPUTE_RETURN_ERROR_ON_MSG(src0->num_dimensions() > 4, "The number of dimensions for the LHS … in validate_arguments()
86 ARM_COMPUTE_RETURN_ERROR_ON(src0->dimension(0) != k); in validate_arguments()
91 ARM_COMPUTE_RETURN_ERROR_ON(src0->dimension(1) * src0->dimension(2) != m); in validate_arguments()
95 ARM_COMPUTE_RETURN_ERROR_ON(src0->dimension(1) != m); in validate_arguments()
103 ARM_COMPUTE_RETURN_ERROR_ON(src0->dimension(3) != src1->dimension(2)); in validate_arguments()
107 ARM_COMPUTE_RETURN_ERROR_ON(src0->dimension(2) != src1->dimension(2)); in validate_arguments()
[all …]
/external/libyuv/files/source/
Dscale_lsx.cc43 __m128i src0, src1, dst0; in ScaleARGBRowDown2_LSX() local
46 DUP2_ARG2(__lsx_vld, src_argb, 0, src_argb, 16, src0, src1); in ScaleARGBRowDown2_LSX()
47 dst0 = __lsx_vpickod_w(src1, src0); in ScaleARGBRowDown2_LSX()
61 __m128i src0, src1, tmp0, tmp1, dst0; in ScaleARGBRowDown2Linear_LSX() local
64 DUP2_ARG2(__lsx_vld, src_argb, 0, src_argb, 16, src0, src1); in ScaleARGBRowDown2Linear_LSX()
65 tmp0 = __lsx_vpickev_w(src1, src0); in ScaleARGBRowDown2Linear_LSX()
66 tmp1 = __lsx_vpickod_w(src1, src0); in ScaleARGBRowDown2Linear_LSX()
82 __m128i src0, src1, src2, src3, tmp0, tmp1, tmp2, tmp3, dst0; in ScaleARGBRowDown2Box_LSX() local
87 DUP2_ARG2(__lsx_vld, s, 0, s, 16, src0, src1); in ScaleARGBRowDown2Box_LSX()
89 DUP4_ARG3(__lsx_vshuf_b, src0, src0, shuff, src1, src1, shuff, src2, src2, in ScaleARGBRowDown2Box_LSX()
[all …]
/external/libvpx/vpx_dsp/loongarch/
Dvpx_convolve8_horiz_lsx.c27 __m128i src0, src1, src2, src3; in common_hz_8t_4x4_lsx() local
39 LSX_LD_4(src, src_stride, src0, src1, src2, src3); in common_hz_8t_4x4_lsx()
40 DUP4_ARG2(__lsx_vxori_b, src0, 128, src1, 128, src2, 128, src3, 128, src0, in common_hz_8t_4x4_lsx()
42 HORIZ_8TAP_4WID_4VECS_FILT(src0, src1, src2, src3, mask0, mask1, mask2, mask3, in common_hz_8t_4x4_lsx()
61 __m128i src0, src1, src2, src3; in common_hz_8t_4x8_lsx() local
73 src0 = __lsx_vld(_src, 0); in common_hz_8t_4x8_lsx()
77 DUP4_ARG2(__lsx_vxori_b, src0, 128, src1, 128, src2, 128, src3, 128, src0, in common_hz_8t_4x8_lsx()
79 HORIZ_8TAP_4WID_4VECS_FILT(src0, src1, src2, src3, mask0, mask1, mask2, mask3, in common_hz_8t_4x8_lsx()
81 src0 = __lsx_vld(_src, 0); in common_hz_8t_4x8_lsx()
84 DUP4_ARG2(__lsx_vxori_b, src0, 128, src1, 128, src2, 128, src3, 128, src0, in common_hz_8t_4x8_lsx()
[all …]

12345678910>>...20