/external/XNNPACK/test/ |
D | prelu-operator-tester.h | 43 inline PReLUOperatorTester& x_stride(size_t x_stride) { in x_stride() argument 44 assert(x_stride != 0); in x_stride() 45 this->x_stride_ = x_stride; in x_stride() 49 inline size_t x_stride() const { in x_stride() function 88 …std::vector<float> x((batch_size() - 1) * x_stride() + channels() + XNN_EXTRA_BYTES / sizeof(float… in TestF32() 100 …ef[i * channels() + c] = std::signbit(x[i * x_stride() + c]) ? x[i * x_stride() + c] * w[c] : x[i … in TestF32() 110 channels(), x_stride(), y_stride(), in TestF32()
|
D | pack-microkernel-tester.h | 55 inline PackMicrokernelTester& x_stride(size_t x_stride) { in x_stride() function 56 assert(x_stride != 0); in x_stride() 57 this->x_stride_ = x_stride; in x_stride() 61 inline size_t x_stride() const { in x_stride() function 85 std::vector<uint32_t> x(k() + (m() - 1) * x_stride() + XNN_EXTRA_BYTES / sizeof(uint32_t)); in Test() 96 y_ref[j * mr() + i] = x[std::min(i, m() - 1) * x_stride() + j]; in Test() 103 x.data(), x_stride() * sizeof(uint32_t), in Test()
|
D | prelu-nc.cc | 38 .x_stride(123) in TEST() 60 .x_stride(123) in TEST() 82 .x_stride(123) in TEST() 104 .x_stride(123) in TEST()
|
D | x32-packx.cc | 120 .x_stride(23) in TEST() 227 .x_stride(23) in TEST() 325 .x_stride(23) in TEST() 378 .x_stride(7) in TEST() 429 .x_stride(7) in TEST() 480 .x_stride(7) in TEST()
|
/external/webp/src/dsp/ |
D | rescaler.c | 31 const int x_stride = wrk->num_channels; in WebPRescalerImportRowExpand_C() local 36 for (channel = 0; channel < x_stride; ++channel) { in WebPRescalerImportRowExpand_C() 42 int right = (wrk->src_width > 1) ? src[x_in + x_stride] : left; in WebPRescalerImportRowExpand_C() 43 x_in += x_stride; in WebPRescalerImportRowExpand_C() 46 x_out += x_stride; in WebPRescalerImportRowExpand_C() 51 x_in += x_stride; in WebPRescalerImportRowExpand_C() 52 assert(x_in < wrk->src_width * x_stride); in WebPRescalerImportRowExpand_C() 63 const int x_stride = wrk->num_channels; in WebPRescalerImportRowShrink_C() local 68 for (channel = 0; channel < x_stride; ++channel) { in WebPRescalerImportRowShrink_C() 78 assert(x_in < wrk->src_width * x_stride); in WebPRescalerImportRowShrink_C() [all …]
|
D | rescaler_mips32.c | 26 const int x_stride = wrk->num_channels; in ImportRowShrink_MIPS32() local 31 const int x_stride1 = x_stride << 2; in ImportRowShrink_MIPS32() 36 for (channel = 0; channel < x_stride; ++channel) { in ImportRowShrink_MIPS32() 75 : [x_stride]"r"(x_stride), [fx_scale]"r"(fx_scale), in ImportRowShrink_MIPS32() 86 const int x_stride = wrk->num_channels; in ImportRowExpand_MIPS32() local 91 const int x_stride1 = x_stride << 2; in ImportRowExpand_MIPS32() 96 for (channel = 0; channel < x_stride; ++channel) { in ImportRowExpand_MIPS32() 137 : [x_stride]"r"(x_stride), [x_add]"r"(x_add), [x_sub]"r"(x_sub), in ImportRowExpand_MIPS32()
|
/external/XNNPACK/src/x32-packx/ |
D | x4-scalar.c | 15 size_t x_stride, in xnn_x32_packx_ukernel_4x__scalar() argument 22 const float* x1 = (const float*) ((uintptr_t) x0 + x_stride); in xnn_x32_packx_ukernel_4x__scalar() 26 const float* x2 = (const float*) ((uintptr_t) x1 + x_stride); in xnn_x32_packx_ukernel_4x__scalar() 30 const float* x3 = (const float*) ((uintptr_t) x2 + x_stride); in xnn_x32_packx_ukernel_4x__scalar()
|
D | x4-neon-st4.c | 17 size_t x_stride, in xnn_x32_packx_ukernel_4x__neon_st4() argument 24 const uint32_t* x1 = (const uint32_t*) ((uintptr_t) x0 + x_stride); in xnn_x32_packx_ukernel_4x__neon_st4() 28 const uint32_t* x2 = (const uint32_t*) ((uintptr_t) x1 + x_stride); in xnn_x32_packx_ukernel_4x__neon_st4() 32 const uint32_t* x3 = (const uint32_t*) ((uintptr_t) x2 + x_stride); in xnn_x32_packx_ukernel_4x__neon_st4()
|
D | x4-wasmsimd.c | 17 size_t x_stride, in xnn_x32_packx_ukernel_4x__wasmsimd() argument 24 const float* x1 = (const float*) ((uintptr_t) x0 + x_stride); in xnn_x32_packx_ukernel_4x__wasmsimd() 28 const float* x2 = (const float*) ((uintptr_t) x1 + x_stride); in xnn_x32_packx_ukernel_4x__wasmsimd() 32 const float* x3 = (const float*) ((uintptr_t) x2 + x_stride); in xnn_x32_packx_ukernel_4x__wasmsimd()
|
D | x3-scalar.c | 15 size_t x_stride, in xnn_x32_packx_ukernel_3x__scalar() argument 22 const float* x1 = (const float*) ((uintptr_t) x0 + x_stride); in xnn_x32_packx_ukernel_3x__scalar() 26 const float* x2 = (const float*) ((uintptr_t) x1 + x_stride); in xnn_x32_packx_ukernel_3x__scalar()
|
D | x4-sse.c | 17 size_t x_stride, in xnn_x32_packx_ukernel_4x__sse() argument 24 const float* x1 = (const float*) ((uintptr_t) x0 + x_stride); in xnn_x32_packx_ukernel_4x__sse() 28 const float* x2 = (const float*) ((uintptr_t) x1 + x_stride); in xnn_x32_packx_ukernel_4x__sse() 32 const float* x3 = (const float*) ((uintptr_t) x2 + x_stride); in xnn_x32_packx_ukernel_4x__sse()
|
D | x2-scalar.c | 15 size_t x_stride, in xnn_x32_packx_ukernel_2x__scalar() argument 22 const float* x1 = (const float*) ((uintptr_t) x0 + x_stride); in xnn_x32_packx_ukernel_2x__scalar()
|
/external/XNNPACK/src/xnnpack/ |
D | vmulcaddc.h | 24 size_t x_stride, \ 58 size_t x_stride, \
|
D | compute.h | 786 size_t x_stride; member 810 size_t x_stride; member 825 size_t x_stride; member 842 size_t x_stride; member 862 size_t x_stride; member 883 size_t x_stride; member 900 size_t x_stride; member 941 size_t x_stride; member 958 size_t x_stride; member
|
D | packx.h | 23 size_t x_stride, \
|
D | params.h | 907 size_t x_stride, 914 size_t x_stride, 1646 size_t x_stride, 1656 size_t x_stride, 1666 size_t x_stride, 1676 size_t x_stride, 1685 size_t x_stride, 1694 size_t x_stride,
|
/external/XNNPACK/src/ |
D | operator-run.c | 724 const size_t x_stride = context->x_stride; in xnn_compute_prelu() local 726 const void* x = (const void*) ((uintptr_t) context->x + x_stride * batch_start); in xnn_compute_prelu() 729 context->ukernel(batch_range, context->n, x, x_stride, context->w, y, y_stride); in xnn_compute_prelu() 783 const void* x = (const void*) ((uintptr_t) context->x + index * context->x_stride); in xnn_compute_channel_shuffle_fixed() 793 const void* x = (const void*) ((uintptr_t) context->x + index * context->x_stride); in xnn_compute_channel_shuffle_variable() 803 const void* x = (const void*) ((uintptr_t) context->x + context->x_stride * batch_index); in xnn_compute_lut_strided() 827 const void* x = (const void*) ((uintptr_t) context->x + context->x_stride * batch_index); in xnn_compute_univector_strided() 846 const uint8_t* x = (const uint8_t*) ((uintptr_t) context->x + context->x_stride * batch_index); in xnn_compute_u8_softmax() 861 const float* x = (const float*) ((uintptr_t) context->x + context->x_stride * batch_index); in xnn_compute_f32_three_pass_softmax() 883 const size_t x_stride = context->x_stride; in xnn_compute_vmulcaddc() local [all …]
|
/external/XNNPACK/src/operators/ |
D | sigmoid-nc.c | 192 .x_stride = input_stride * sizeof(uint8_t), in xnn_setup_sigmoid_nc_qu8() 206 .x_stride = input_stride * sizeof(uint8_t), in xnn_setup_sigmoid_nc_qu8()
|
D | leaky-relu-nc.c | 198 .x_stride = input_stride * sizeof(uint8_t), in xnn_setup_leaky_relu_nc_qu8() 212 .x_stride = input_stride * sizeof(uint8_t), in xnn_setup_leaky_relu_nc_qu8()
|
D | softmax-nc.c | 170 .x_stride = softmax_op->input_pixel_stride * sizeof(uint8_t), in xnn_setup_softmax_nc_qu8() 285 .x_stride = softmax_op->input_pixel_stride * sizeof(float), in xnn_setup_softmax_nc_f32()
|
D | prelu-nc.c | 128 .x_stride = prelu_op->input_pixel_stride * sizeof(float), in xnn_setup_prelu_nc_f32()
|
D | channel-shuffle-nc.c | 165 .x_stride = channel_shuffle_op->input_pixel_stride << log2_element_size, in setup_channel_shuffle_nc()
|
D | unary-elementwise-nc.c | 116 .x_stride = input_stride << log2_element_size, in setup_unary_elementwise_nc() 132 .x_stride = input_stride << log2_element_size, in setup_unary_elementwise_nc()
|
/external/tensorflow/tensorflow/cc/framework/ |
D | gradient_checker.cc | 142 const int x_stride = JacobianStride<X_T>::value; in ComputeTheoreticalJacobianTranspose() local 165 SetJacobian<X_T, JAC_T>(&jacobian, r * x_stride, in ComputeTheoreticalJacobianTranspose() 217 const int x_stride = JacobianStride<X_T>::value; in ComputeNumericJacobianTranspose() local 253 SetJacobian<Y_T, JAC_T>(&jacobian, r * x_stride + unit_dimension, in ComputeNumericJacobianTranspose()
|
/external/mesa3d/src/gallium/auxiliary/gallivm/ |
D | lp_bld_sample_aos.c | 455 LLVMValueRef x_stride; in lp_build_sample_image_nearest() local 503 x_stride = lp_build_const_vec(bld->gallivm, in lp_build_sample_image_nearest() 511 width_vec, x_stride, offsets[0], in lp_build_sample_image_nearest() 767 LLVMValueRef x_stride, y_stride, z_stride; in lp_build_sample_image_linear() local 859 x_stride = lp_build_const_vec(bld->gallivm, bld->int_coord_bld.type, in lp_build_sample_image_linear() 868 width_vec, x_stride, offsets[0], in lp_build_sample_image_linear()
|