/third_party/gettext/gettext-tools/src/ |
D | xg-arglist-callshape.c | 185 struct callshapes *shapes = XMALLOC (struct callshapes); in insert_keyword_callshape() local 186 shapes->nshapes = 1; in insert_keyword_callshape() 187 shapes->shapes[0] = *shape; in insert_keyword_callshape() 189 (const char *) hash_insert_entry (table, keyword, keyword_len, shapes); in insert_keyword_callshape() 192 shapes->keyword = keyword; in insert_keyword_callshape() 193 shapes->keyword_len = keyword_len; in insert_keyword_callshape() 205 if (old_shapes->shapes[i].argnum1 == shape->argnum1 in insert_keyword_callshape() 206 && old_shapes->shapes[i].argnum2 == shape->argnum2 in insert_keyword_callshape() 207 && old_shapes->shapes[i].argnumc == shape->argnumc in insert_keyword_callshape() 208 && old_shapes->shapes[i].argnum1_glib_context in insert_keyword_callshape() [all …]
|
D | xg-arglist-parser.c | 41 arglist_parser_alloc (message_list_ty *mlp, const struct callshapes *shapes) in arglist_parser_alloc() argument 43 if (shapes == NULL || shapes->nshapes == 0) in arglist_parser_alloc() 62 xtimes (shapes->nshapes - 1, in arglist_parser_alloc() 67 ap->keyword = shapes->keyword; in arglist_parser_alloc() 68 ap->keyword_len = shapes->keyword_len; in arglist_parser_alloc() 70 ap->nalternatives = shapes->nshapes; in arglist_parser_alloc() 71 for (i = 0; i < shapes->nshapes; i++) in arglist_parser_alloc() 73 ap->alternative[i].argnumc = shapes->shapes[i].argnumc; in arglist_parser_alloc() 74 ap->alternative[i].argnum1 = shapes->shapes[i].argnum1; in arglist_parser_alloc() 75 ap->alternative[i].argnum2 = shapes->shapes[i].argnum2; in arglist_parser_alloc() [all …]
|
/third_party/mindspore/tests/ut/python/model/ |
D | test_utils.py | 24 shapes = ((1, 3, 224, 224), (32,)) 25 _construct_tensor_list(types, shapes) 30 shapes = ((1, 3, 224, 224),) 31 _construct_tensor_list(types, shapes) 36 shapes = ((1, 3, 224, 224),) 38 _construct_tensor_list(types, shapes)
|
/third_party/mindspore/tests/st/ops/cpu/ |
D | test_rl_buffer_net.py | 28 def create_tensor(capcity, shapes, types): argument 30 for i in range(len(shapes)): 31 buffer.append(Parameter(Tensor(np.zeros(((capcity,)+shapes[i])), types[i]), \ 37 def __init__(self, batch_size, capcity, shapes, types): argument 39 self.buffer = create_tensor(capcity, shapes, types) 44 self.buffer_append = P.BufferAppend(self._capacity, shapes, types) 45 self.buffer_get = P.BufferGetItem(self._capacity, shapes, types) 47 self._capacity, batch_size, shapes, types) 75 buffer = RLBuffer(batch_size=32, capcity=100, shapes=[(4,), (2,), (1,), (4,)], types=[
|
D | test_rl_buffer_op.py | 29 def __init__(self, capcity, shapes, types): argument 34 self.buffer_append = P.BufferAppend(self._capacity, shapes, types) 42 def __init__(self, capcity, shapes, types): argument 47 self.buffer_get = P.BufferGetItem(self._capacity, shapes, types) 55 def __init__(self, capcity, batch_size, shapes, types): argument 60 self.buffer_sample = P.BufferSample(self._capacity, batch_size, shapes, types) 90 buffer_sample = RLBufferSample(capcity=5, batch_size=3, shapes=[(4,), (2,), (1,), (4,)], types=[ 101 buffer_get = RLBufferGet(capcity=5, shapes=[(4,), (2,), (1,), (4,)], types=[ 119 buffer_append = RLBufferAppend(capcity=5, shapes=[(4,), (2,), (1,), (4,)], types=[
|
/third_party/mindspore/tests/st/ops/gpu/ |
D | test_rl_buffer_net.py | 28 def create_tensor(capcity, shapes, dtypes): argument 30 for i in range(len(shapes)): 31 buffer.append(Tensor(np.zeros(((capcity,)+shapes[i])), dtypes[i])) 36 def __init__(self, batch_size, capcity, shapes, types): argument 38 self.buffer = create_tensor(capcity, shapes, types) 42 self.buffer_append = P.BufferAppend(self._capacity, shapes, types) 43 self.buffer_get = P.BufferGetItem(self._capacity, shapes, types) 45 self._capacity, batch_size, shapes, types) 75 buffer = RLBuffer(batch_size=32, capcity=100, shapes=[(4,), (2,), (1,), (4,)], types=[
|
D | test_rl_buffer_op.py | 29 def __init__(self, capcity, shapes, types): argument 34 self.buffer_append = P.BufferAppend(self._capacity, shapes, types) 42 def __init__(self, capcity, shapes, types): argument 47 self.buffer_get = P.BufferGetItem(self._capacity, shapes, types) 55 def __init__(self, capcity, batch_size, shapes, types): argument 61 self._capacity, batch_size, shapes, types) 91 buffer_sample = RLBufferSample(capcity=5, batch_size=3, shapes=[(4,), (2,), (1,), (4,)], types=[ 102 buffer_get = RLBufferGet(capcity=5, shapes=[(4,), (2,), (1,), (4,)], types=[ 120 buffer_append = RLBufferAppend(capcity=5, shapes=[(4,), (2,), (1,), (4,)], types=[
|
/third_party/skia/third_party/externals/harfbuzz/src/ |
D | gen-arabic-table.py | 158 shapes = {} 190 if items[0] not in shapes: 191 shapes[items[0]] = {} 192 shapes[items[0]][shape] = c 198 keys = shapes.keys () 201 s = [shapes[u][shape] if u in shapes and shape in shapes[u] else 0 217 liga = (shapes[pair[0]]['initial'], shapes[pair[1]]['final']) 219 liga = (shapes[pair[0]]['medial'], shapes[pair[1]]['final'])
|
/third_party/harfbuzz/src/ |
D | gen-arabic-table.py | 158 shapes = {} 190 if items[0] not in shapes: 191 shapes[items[0]] = {} 192 shapes[items[0]][shape] = c 198 keys = shapes.keys () 201 s = [shapes[u][shape] if u in shapes and shape in shapes[u] else 0 217 liga = (shapes[pair[0]]['initial'], shapes[pair[1]]['final']) 219 liga = (shapes[pair[0]]['medial'], shapes[pair[1]]['final'])
|
/third_party/flutter/skia/third_party/externals/harfbuzz/src/ |
D | gen-arabic-table.py | 153 shapes = {} 186 if items[0] not in shapes: 187 shapes[items[0]] = {} 188 shapes[items[0]][shape] = c 194 keys = shapes.keys () 197 s = [shapes[u][shape] if u in shapes and shape in shapes[u] else 0 213 liga = (shapes[pair[0]]['initial'], shapes[pair[1]]['final']) 215 liga = (shapes[pair[0]]['medial'], shapes[pair[1]]['final'])
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/gpu/data/ |
D | dataset_utils.cc | 36 void GetShapeAndType(const CNodePtr &kernel_node, std::vector<std::vector<int>> *shapes, std::vecto… in GetShapeAndType() argument 38 MS_EXCEPTION_IF_NULL(shapes); in GetShapeAndType() 42 (void)std::transform(shapes_me.begin(), shapes_me.end(), std::back_inserter(*shapes), in GetShapeAndType() 55 if (shapes->size() != types->size()) { in GetShapeAndType() 56 MS_LOG(EXCEPTION) << "Invalid shapes: " << *shapes << ", types: " << *types; in GetShapeAndType()
|
D | dataset_init_kernel.cc | 40 std::vector<std::vector<int>> shapes; in Init() local 42 GetShapeAndType(kernel_node, &shapes, &types); in Init() 46 for (size_t i = 0; i < shapes.size(); i++) { in Init() 48 int nums = ElementNums(shapes[i]); in Init()
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/ps/ |
D | sparse_apply_adam_ps_kernel.cc | 29 …NodePtr &cnode, const std::shared_ptr<std::vector<std::shared_ptr<std::vector<size_t>>>> &shapes) { in InitKernel() argument 31 MS_EXCEPTION_IF_NULL(shapes); in InitKernel() 32 const std::vector<std::shared_ptr<std::vector<size_t>>> &shape_vec = *shapes; in InitKernel() 81 void SparseApplyAdamPSKernel::ReInit(const std::vector<std::vector<size_t>> &shapes) { in ReInit() argument 82 if (shapes.empty() || shapes[0].empty()) { in ReInit() 85 const std::vector<size_t> &indices_shape = shapes[0]; in ReInit()
|
D | sparse_apply_lazy_adam_ps_kernel.cc | 29 …NodePtr &cnode, const std::shared_ptr<std::vector<std::shared_ptr<std::vector<size_t>>>> &shapes) { in InitKernel() argument 31 MS_EXCEPTION_IF_NULL(shapes); in InitKernel() 32 const std::vector<std::shared_ptr<std::vector<size_t>>> &shape_vec = *shapes; in InitKernel() 82 void SparseApplyLazyAdamPSKernel::ReInit(const std::vector<std::vector<size_t>> &shapes) { in ReInit() argument 83 if (shapes.empty() || shapes[0].empty()) { in ReInit() 86 const std::vector<size_t> &indices_shape = shapes[0]; in ReInit()
|
D | sparse_apply_ftrl_ps_kernel.cc | 26 …NodePtr &cnode, const std::shared_ptr<std::vector<std::shared_ptr<std::vector<size_t>>>> &shapes) { in InitKernel() argument 28 MS_EXCEPTION_IF_NULL(shapes); in InitKernel() 29 const std::vector<std::shared_ptr<std::vector<size_t>>> &shape_vec = *shapes; in InitKernel() 92 void SparseApplyFtrlPSKernel::ReInit(const std::vector<std::vector<size_t>> &shapes) { in ReInit() argument 93 if (shapes.empty() || shapes[0].empty()) { in ReInit() 96 const std::vector<size_t> &indices_shape = shapes[0]; in ReInit()
|
D | embedding_look_up_ps_kernel.cc | 32 const std::shared_ptr<std::vector<std::shared_ptr<std::vector<size_t>>>> &shapes) { in InitKernel() argument 33 const std::vector<std::shared_ptr<std::vector<size_t>>> &shape_vec = *shapes; in InitKernel() 70 void EmbeddingLookUpPSKernel::ReInit(const std::vector<std::vector<size_t>> &shapes) { in ReInit() argument 71 if (shapes.empty() || shapes[0].empty()) { in ReInit() 74 const auto &indices_shape = shapes[0]; in ReInit()
|
/third_party/mindspore/tests/ut/cpp/dataset/ |
D | c_api_dataset_cifar_test.cc | 139 std::vector<TensorShape> shapes = ToTensorShapeVec(ds->GetOutputShapes()); in TEST_F() local 145 EXPECT_EQ(shapes.size(), 2); in TEST_F() 146 EXPECT_EQ(shapes[0].ToString(), "<32,32,3>"); in TEST_F() 147 EXPECT_EQ(shapes[1].ToString(), "<>"); in TEST_F() 154 EXPECT_EQ(ToTensorShapeVec(ds->GetOutputShapes()), shapes); in TEST_F() 160 EXPECT_EQ(ToTensorShapeVec(ds->GetOutputShapes()), shapes); in TEST_F() 212 std::vector<TensorShape> shapes = ToTensorShapeVec(ds->GetOutputShapes()); in TEST_F() local 219 EXPECT_EQ(shapes.size(), 3); in TEST_F() 220 EXPECT_EQ(shapes[0].ToString(), "<32,32,3>"); in TEST_F() 221 EXPECT_EQ(shapes[1].ToString(), "<>"); in TEST_F() [all …]
|
/third_party/mindspore/mindspore/lite/tools/converter/quantizer/ |
D | quantize_util.cc | 685 void CalQuantAssitInfo(const PrimitivePtr &primitive, const ShapeVector &shapes, int index, bool *c… in CalQuantAssitInfo() argument 688 if (shapes.empty()) { in CalQuantAssitInfo() 692 if (primitive->name() == ops::kNameMatMul && static_cast<int>(shapes.size()) == DIMENSION_2D) { in CalQuantAssitInfo() 699 if (shapes.size() != kLstmWeightShapeSize) { in CalQuantAssitInfo() 700 MS_LOG(WARNING) << "unexpected lstm shape size: " << shapes.size(); in CalQuantAssitInfo() 702 *channel_cnt = shapes[0] * shapes[1]; in CalQuantAssitInfo() 705 if (shapes.size() != kLstmBiasShapeSize) { in CalQuantAssitInfo() 706 MS_LOG(WARNING) << "unexpected lstm shape size: " << shapes.size(); in CalQuantAssitInfo() 708 auto tensor_elem_cnt = shapes[0] * shapes[1]; in CalQuantAssitInfo() 719 void CalQuantAssitInfo(const schema::PrimitiveT &primitive, const std::vector<int> &shapes, int ind… in CalQuantAssitInfo() argument [all …]
|
/third_party/flutter/skia/third_party/externals/sdl/premake/Xcode/Xcode4/tests/testshape/testshape.xcodeproj/ |
D | project.pbxproj | 252 …shapes\" ]; then mkdir -p \"./Build/Debug/shapes\"; fi\ncp ./../../../../../test/shapes/*.bmp \"./…
|
/third_party/mindspore/mindspore/ccsrc/backend/optimizer/ascend/mindir/ |
D | fake_learned_scale_quant_grad_unify_mindir.cc | 50 auto shapes = {AnfAlgo::GetOutputInferShape(lsq_perlayer_grad_node, 0), in CreateOutputsOfLSQPerLayerGradD() local 52 AnfAlgo::SetOutputInferTypeAndShape(types, shapes, lsq_perlayer_grad_d.get()); in CreateOutputsOfLSQPerLayerGradD() 82 auto shapes = {AnfAlgo::GetOutputInferShape(lsq_perlayer_grad_node, 1)}; in CreateOutputsOfLSQPerLayerReduceGrad() local 83 AnfAlgo::SetOutputInferTypeAndShape(types, shapes, lsq_perlayer_reduce_grad.get()); in CreateOutputsOfLSQPerLayerReduceGrad() 107 auto shapes = {AnfAlgo::GetOutputInferShape(lsq_perchannel_grad_node, 0), in CreateOutputsOfLSQPerChannelGradD() local 109 AnfAlgo::SetOutputInferTypeAndShape(types, shapes, lsq_perchannel_grad_d.get()); in CreateOutputsOfLSQPerChannelGradD() 140 auto shapes = {AnfAlgo::GetOutputInferShape(lsq_perchannel_grad_node, 1)}; in CreateOutputsOfLSQPerChannelReduceGrad() local 141 AnfAlgo::SetOutputInferTypeAndShape(types, shapes, lsq_perchannel_reduce_grad.get()); in CreateOutputsOfLSQPerChannelReduceGrad()
|
/third_party/mindspore/mindspore/core/abstract/ |
D | dshape.h | 100 explicit SequeueShape(const BaseShapePtrList &shapes) : p_shapes_(shapes) {} in SequeueShape() argument 125 explicit TupleShape(const BaseShapePtrList &shapes) : SequeueShape(shapes) {} in TupleShape() argument 140 explicit ListShape(const BaseShapePtrList &shapes) : SequeueShape(shapes) {} in ListShape() argument
|
/third_party/mindspore/mindspore/ccsrc/frontend/parallel/ |
D | step_parallel_utils.cc | 84 Shapes shapes; in GetValueListShape() local 100 shapes.push_back(one_shape); in GetValueListShape() 102 return shapes; in GetValueListShape() 107 Shapes shapes; in GetNodeShape() local 142 shapes.push_back(each_shape->shape()); in GetNodeShape() 147 shapes.push_back(shape_ptr->shape()); in GetNodeShape() 149 return shapes; in GetNodeShape()
|
/third_party/flutter/skia/third_party/externals/sdl/premake/Xcode/Xcode3/tests/testshape/testshape.xcodeproj/ |
D | project.pbxproj | 252 …shapes\" ]; then mkdir -p \"./Build/Debug/shapes\"; fi\ncp ./../../../../../test/shapes/*.bmp \"./…
|
/third_party/mindspore/tests/st/data_transfer/ |
D | test_tdt_data_transfer.py | 67 def convert_type(shapes, types): argument 69 for np_shape, np_type in zip(shapes, types): 95 def __init__(self, shapes): argument 97 self.shapes = tuple(shapes[0]) 101 return self.Op_Reshape_network(network_input, self.shapes)
|
/third_party/boost/boost/math/distributions/ |
D | skew_normal.hpp | 341 static const RealType shapes[] = { in mode_fallback() local 389 const RealType* result_ptr = std::lower_bound(shapes, shapes+21, shape); in mode_fallback() 393 const diff_type d = std::distance(shapes, result_ptr); in mode_fallback() 401 … + (guess[d]-guess[d-static_cast<diff_type>(1)])/(shapes[d]-shapes[d-static_cast<diff_type>(1)]) in mode_fallback() 402 * (shape-shapes[d-static_cast<diff_type>(1)]); in mode_fallback() 479 static const RealType shapes[] = { in mode() local 528 const RealType* result_ptr = std::lower_bound(shapes, shapes+21, shape); in mode() 532 const diff_type d = std::distance(shapes, result_ptr); in mode() 545 … + (guess[d]-guess[d-static_cast<diff_type>(1)])/(shapes[d]-shapes[d-static_cast<diff_type>(1)]) in mode() 546 * (shape-shapes[d-static_cast<diff_type>(1)]); in mode()
|