/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/gpu/cuda_impl/ |
D | cumsum_impl.cu | 29 __global__ void LeftMoveSum(const T *input, T *output, size_t dim0, size_t dim1, size_t dim2, size_… in LeftMoveSum() argument 31 size_t num = dim0 * dim2; in LeftMoveSum() 35 i = write_index / dim2 % dim0; in LeftMoveSum() 36 k = write_index % dim2; in LeftMoveSum() 51 __global__ void RightMoveSum(const T *input, T *output, size_t dim0, size_t dim1, size_t dim2, size… in RightMoveSum() argument 53 size_t num = dim0 * dim2; in RightMoveSum() 57 i = write_index / dim2 % dim0; in RightMoveSum() 58 k = write_index % dim2; in RightMoveSum() 72 …umSumKernelReverse(const T *input, T *output, size_t dim0, size_t dim1, size_t dim2, size_t stride, in CumSumKernelReverse() argument 74 size_t num = dim0 * dim2; in CumSumKernelReverse() [all …]
|
D | cumprod_impl.cu | 29 __global__ void LeftMoveProd(const T *input, T *output, size_t dim0, size_t dim1, size_t dim2, size… in LeftMoveProd() argument 31 size_t num = dim0 * dim2; in LeftMoveProd() 35 i = write_index / dim2 % dim0; in LeftMoveProd() 36 k = write_index % dim2; in LeftMoveProd() 51 __global__ void RightMoveProd(const T *input, T *output, size_t dim0, size_t dim1, size_t dim2, siz… in RightMoveProd() argument 53 size_t num = dim0 * dim2; in RightMoveProd() 57 i = write_index / dim2 % dim0; in RightMoveProd() 58 k = write_index % dim2; in RightMoveProd() 72 …mProdKernelReverse(const T *input, T *output, size_t dim0, size_t dim1, size_t dim2, size_t stride, in CumProdKernelReverse() argument 74 size_t num = dim0 * dim2; in CumProdKernelReverse() [all …]
|
D | cumprod_impl.cuh | 20 void CumProd(const T *input, T *output, T *workspace, size_t dim0, size_t dim1, size_t dim2, size_t…
|
D | cumsum_impl.cuh | 20 void CumSum(const T *input, T *output, T *workspace, size_t dim0, size_t dim1, size_t dim2, size_t …
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/ |
D | cumsum_cpu_kernel.cc | 113 void CumSumCPUKernel::LeftMove(const T *input, T *output, size_t dim0, size_t dim1, size_t dim2, si… in LeftMove() argument 116 size_t k1 = i / dim2 % dim0; in LeftMove() 117 size_t k2 = i % dim2; in LeftMove() 132 void CumSumCPUKernel::RightMove(const T *input, T *output, size_t dim0, size_t dim1, size_t dim2, s… in RightMove() argument 135 size_t k1 = i / dim2 % dim0; in RightMove() 136 size_t k2 = i % dim2; in RightMove() 151 void CumSumCPUKernel::Copy(T *input, T *output, size_t dim0, size_t dim1, size_t dim2, size_t strid… in Copy() argument 154 size_t k1 = i / dim2 % dim0; in Copy() 155 size_t k2 = i % dim2; in Copy() 165 …SumCPUKernel::CumSumKernelReverse(const T *input, T *output, size_t dim0, size_t dim1, size_t dim2, in CumSumKernelReverse() argument [all …]
|
D | cumsum_cpu_kernel.h | 47 …void LeftMove(const T *input, T *output, size_t dim0, size_t dim1, size_t dim2, size_t stride, siz… 51 …void RightMove(const T *input, T *output, size_t dim0, size_t dim1, size_t dim2, size_t stride, si… 55 …void Copy(T *input, T *output, size_t dim0, size_t dim1, size_t dim2, size_t stride, size_t stride… 59 …void CumSumKernelReverse(const T *input, T *output, size_t dim0, size_t dim1, size_t dim2, size_t … 63 …void CumSumKernel(const T *input, T *output, size_t dim0, size_t dim1, size_t dim2, size_t stride,…
|
D | cpu_kernel.cc | 67 …CPUKernelUtils::CalcOffset(const std::vector<size_t> &shape, size_t dim0, size_t dim1, size_t dim2, in CalcOffset() argument 69 …size_t offset = dim0 * shape[1] * shape[2] * shape[3] + dim1 * shape[2] * shape[3] + dim2 * shape[… in CalcOffset()
|
D | cpu_kernel.h | 146 …static size_t CalcOffset(const std::vector<size_t> &shape, size_t dim0, size_t dim1, size_t dim2, …
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/nnacl/fp32/ |
D | strided_slice_fp32.c | 82 int32_t dim0, dim1, dim2, dim3, dim4, dim5, dim6, dim7; in DoStridedSliceIntFp64Bool() local 85 for (dim2 = begins[2]; LoopContinue(strides[2], dim2, ends[2]); dim2 += strides[2]) { in DoStridedSliceIntFp64Bool() 91 … int32_t in_offset = dim0 * dim_offset[0] + dim1 * dim_offset[1] + dim2 * dim_offset[2] + in DoStridedSliceIntFp64Bool() 144 int32_t dim0, dim1, dim2, dim3, dim4, dim5, dim6, dim7; in DoStridedSlice() local 147 for (dim2 = begins[2]; LoopContinue(strides[2], dim2, ends[2]); dim2 += strides[2]) { in DoStridedSlice() 153 … int32_t in_offset = dim0 * dim_offset[0] + dim1 * dim_offset[1] + dim2 * dim_offset[2] + in DoStridedSlice()
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/nnacl/ |
D | common_func.c | 19 int Offset(const int *shape, const int dim0, const int dim1, const int dim2, const int dim3) { in Offset() argument 20 return ((dim0 * shape[1] + dim1) * shape[2] + dim2) * shape[3] + dim3; in Offset() 23 int OffsetComm(const int *shape, const int dim0, const int dim1, const int dim2) { in OffsetComm() argument 24 return ((dim0 * shape[1] + dim1) * shape[2] + dim2) * shape[3]; in OffsetComm()
|
D | common_func.h | 31 int Offset(const int *shape, const int dim0, const int dim1, const int dim2, const int dim3); 32 int OffsetComm(const int *shape, const int dim0, const int dim1, const int dim2);
|
/third_party/mindspore/tests/ut/python/train/summary/ |
D | test_histogram_summary.py | 111 dim2 = 100 113 test_data = _wrap_test_data(Tensor(np.ones([dim1, dim2]))) 122 … assert len(event.summary.value[0].histogram.buckets) == _calc_histogram_bins(dim1 * dim2) 152 dim2 = 100 154 arr = np.ones([dim1, dim2])
|
/third_party/boost/libs/units/test/ |
D | fail_base_dimension.cpp | 28 struct dim2 : boost::units::base_dimension<dim2, 1> {}; struct
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/nnacl/base/ |
D | slice_base.c | 134 for (int32_t dim2 = param->begin_[2]; dim2 < param->end_[2]; ++dim2) { in DoSliceNoParallel() local 135 int in_offset2 = in_offset1 + dim2 * in_stride[2]; in DoSliceNoParallel()
|
/third_party/boost/boost/units/detail/ |
D | dimension_impl.hpp | 197 typedef typename Begin2::item dim2; typedef 199 typedef typename merge_dimensions_func<(mpl::less<dim1,dim2>::value == true), 200 (mpl::less<dim2,dim1>::value == true)>::template apply<
|
/third_party/boost/libs/geometry/test/algorithms/detail/sections/ |
D | range_by_section.cpp | 33 typedef boost::mpl::vector_c<std::size_t, 0> dim2; in test_sectionalize() typedef 36 bg::sectionalize<Reverse, dim2>(geometry, bg::detail::no_rescale_policy(), s); in test_sectionalize()
|
D | sectionalize.cpp | 206 typedef boost::mpl::vector_c<std::size_t, 0, 1> dim2; in test_sectionalize() typedef 209 test_sectionalize<dim2, Reverse>(caseid + "_d2", g, count2, s2, d2, max_count); in test_sectionalize()
|
/third_party/boost/boost/geometry/srs/ |
D | projection.hpp | 79 static const std::size_t dim2 = geometry::dimension<Point2>::value; in copy_higher_dimensions() local 80 static const std::size_t lesser_dim = dim1 < dim2 ? dim1 : dim2; in copy_higher_dimensions()
|
/third_party/skia/third_party/externals/angle2/src/tests/gl_tests/ |
D | GLSLTest.cpp | 10731 for (int dim2 = 0; dim2 < 3; ++dim2) in TEST_P() local 10733 int textureUnit = dim1 * 3 + dim2; in TEST_P() 10735 glBindTexture(GL_TEXTURE_2D, textures[dim1][dim2]); in TEST_P() 10742 uniformName << "smplr[" << dim1 << "][" << dim2 << "]"; in TEST_P() 10840 for (int dim2 = 0; dim2 < 3; ++dim2) in TEST_P() local 10842 int textureUnit = dim1 * 3 + dim2; in TEST_P() 10844 glBindTexture(GL_TEXTURE_2D, textures[dim1][dim2]); in TEST_P() 10851 uniformName << "smplr[" << dim1 << "][" << dim2 << "]"; in TEST_P() 10973 for (int dim2 = 0; dim2 < 3; ++dim2) in TEST_P() local 10977 int textureUnit = (dim1 * 3 + dim2) * 4 + dim3; in TEST_P() [all …]
|
/third_party/skia/third_party/externals/swiftshader/src/OpenGL/compiler/ |
D | OutputASM.h | 320 static int dim2(TIntermNode *m);
|
D | OutputASM.cpp | 3825 int OutputASM::dim2(TIntermNode *m) in dim2() function in glsl::OutputASM
|
/third_party/libxml2/ |
D | xmlregexp.c | 441 xmlRegCalloc2(size_t dim1, size_t dim2, size_t elemSize) { in xmlRegCalloc2() argument 446 if (dim1 > SIZE_MAX / dim2 / elemSize) in xmlRegCalloc2() 448 totalSize = dim1 * dim2 * elemSize; in xmlRegCalloc2()
|