/third_party/mindspore/mindspore/lite/micro/coder/opcoders/nnacl/fp32/ |
D | softmax_fp32_coder.cc | 30 int n_dim = softmax_param_->n_dim_; in Prepare() local 33 softmax_param_->axis_ += n_dim; in Prepare() 42 for (int i = axis + 1; i < n_dim; i++) { in Prepare()
|
D | batchnorm_fp32_coder.cc | 34 int n_dim = static_cast<int>(input_shapes.size()); in Init() local 35 bn_parameter->channel_ = input_shapes.at(n_dim - 1); in Init() 37 for (int i = 0; i < n_dim - 1; i++) { in Init()
|
/third_party/mindspore/mindspore/lite/src/runtime/kernel/arm/fp32/ |
D | batchnorm_fp32.cc | 65 auto n_dim = input_shapes.size(); in FillParam() local 67 CHECK_LESS_RETURN(n_dim - 1, 0); in FillParam() 68 param->channel_ = input_shapes[n_dim - 1]; in FillParam() 70 for (size_t i = 0; i < n_dim - 1; i++) { in FillParam()
|
D | prelu_fp32.cc | 76 auto n_dim = in_shape.size(); in ReSize() local 77 auto channel_num = in_shape.at(n_dim - 1); in ReSize() 79 for (size_t i = 0; i < n_dim - 1; ++i) { in ReSize()
|
D | log_softmax_fp32.cc | 56 auto n_dim = softmax_param_->n_dim_; in ReSize() local 64 for (int i = axis + 1; i < n_dim; i++) { in ReSize()
|
D | softmax_fp32.cc | 51 auto n_dim = softmax_param_->n_dim_; in ReSize() local 59 for (int i = axis + 1; i < n_dim; i++) { in ReSize()
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/nnacl/fp32_grad/ |
D | softmax_grad.c | 24 int n_dim = parameter->n_dim_; in SoftmaxGrad() local 33 for (int i = axis + 1; i < n_dim; i++) { in SoftmaxGrad() 38 for (int i = 0; i < n_dim; i++) dim *= input_shape[i]; in SoftmaxGrad()
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/nnacl/int8/ |
D | softmax_int8.c | 23 int n_dim = parameter->n_dim_; in SoftmaxInt8() local 28 if (n_dim > DIMENSION_5D) { in SoftmaxInt8() 31 for (int i = axis + 1; i < n_dim; i++) { in SoftmaxInt8()
|
/third_party/mindspore/mindspore/lite/src/runtime/kernel/arm/base/ |
D | stack_base.cc | 36 static int GetCopyNum(const std::vector<int> &in_shape, int axis, int n_dim) { in GetCopyNum() argument 39 for (int j = n_dim - 1; j > axis - 1; j--) { in GetCopyNum() 43 for (int i = 0; i < n_dim; ++i) { in GetCopyNum()
|
/third_party/mindspore/mindspore/lite/src/runtime/kernel/arm/int8/ |
D | batchnorm_int8.cc | 173 auto n_dim = input_shapes.size(); in Init() local 174 CHECK_LESS_RETURN(n_dim - 1, 0); in Init() 175 batchnorm_param_->channel_ = input_shapes[n_dim - 1]; in Init() 177 for (size_t i = 0; i < n_dim - 1; i++) { in Init()
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/nnacl/fp32/ |
D | log_softmax_fp32.c | 48 int n_dim = parameter->n_dim_; in LogSoftmax() local 56 for (int i = axis + 1; i < n_dim; i++) { in LogSoftmax()
|
D | softmax_fp32.c | 107 int n_dim = parameter->n_dim_; in Softmax() local 115 for (int i = axis + 1; i < n_dim; i++) { in Softmax()
|
/third_party/mindspore/mindspore/lite/micro/coder/opcoders/nnacl/int8/ |
D | batchnorm_int8_coder.cc | 29 size_t n_dim = input_shapes.size(); in Prepare() local 30 batchnorm_param_->channel_ = input_shapes[n_dim - 1]; in Prepare() 32 for (size_t i = 0; i < n_dim - 1; i++) { in Prepare()
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/nnacl/fp16/ |
D | log_softmax_fp16.c | 51 int n_dim = parameter->n_dim_; in LogSoftmaxFp16() local 59 for (int i = axis + 1; i < n_dim; i++) { in LogSoftmaxFp16()
|
D | softmax_fp16.c | 97 int n_dim = parameter->n_dim_; in SoftmaxFp16() local 105 for (int i = axis + 1; i < n_dim; i++) { in SoftmaxFp16()
|
/third_party/mindspore/mindspore/lite/src/runtime/kernel/arm/fp16/ |
D | log_softmax_fp16.cc | 58 auto n_dim = softmax_param_->n_dim_; in ReSize() local 66 for (int i = axis + 1; i < n_dim; i++) { in ReSize()
|
D | softmax_fp16.cc | 51 auto n_dim = softmax_param_->n_dim_; in ReSize() local 59 for (int i = axis + 1; i < n_dim; i++) { in ReSize()
|
/third_party/mindspore/mindspore/lite/micro/example/mnist_stm32f746/mnist_stm32f746/operator_library/kernels/nnacl/fp32/ |
D | softmax_fp32.c | 107 int n_dim = parameter->n_dim_; in Softmax() local 115 for (int i = axis + 1; i < n_dim; i++) { in Softmax()
|
/third_party/mindspore/mindspore/lite/src/runtime/kernel/arm/fp32_grad/ |
D | sparse_softmax_cross_entropy_with_logits.cc | 130 int n_dim = sm_params_.n_dim_; in Run() local 144 for (int i = axis + 1; i < n_dim; i++) { in Run()
|
/third_party/mindspore/mindspore/nn/loss/ |
D | loss.py | 943 n_dim = self.shape(all_w)[1] 945 true_w = self.slice_op(all_w, [0, 0], [n_true, n_dim]) 946 sampled_w = self.slice_op(all_w, [n_true, 0], [n_sampled, n_dim]) 953 new_true_w_shape = (-1, num_true, n_dim) 959 dots_as_matrix = self.reshape(row_wise_dots, (-1, n_dim))
|
/third_party/mindspore/mindspore/ccsrc/common/ |
D | trans.cc | 1979 auto n_dim = args.host_shape[kN]; in NchwFracZTransWithGroups() local 1986 auto cout_ori = n_dim / group_size; in NchwFracZTransWithGroups()
|