/external/tensorflow/tensorflow/python/framework/ |
D | common_shapes.py | 45 for (dim_x, dim_y) in broadcasted_dims: 46 if dim_x.value is None or dim_y.value is None: 52 if dim_x.value is not None and dim_x.value > 1: 53 return_dims.append(dim_x) 58 elif dim_x.value == 1: 63 return_dims.append(dim_x) 64 elif dim_x.value == dim_y.value: 67 return_dims.append(dim_x.merge_with(dim_y))
|
/external/eigen/unsupported/test/ |
D | cxx11_tensor_reduction_cuda.cu | 61 int dim_x = 33; in test_first_dim_reductions() local 65 Tensor<Type, 3, DataLayout> in(dim_x, dim_y, dim_z); in test_first_dim_reductions() 77 Type* in_data = (Type*)dev.allocate(dim_x*dim_y*dim_z*sizeof(Type)); in test_first_dim_reductions() 79 Eigen::TensorMap<Eigen::Tensor<Type, 3, DataLayout> > gpu_in(in_data, dim_x, dim_y, dim_z); in test_first_dim_reductions() 101 int dim_x = 128; in test_last_dim_reductions() local 105 Tensor<Type, 3, DataLayout> in(dim_x, dim_y, dim_z); in test_last_dim_reductions() 117 Type* in_data = (Type*)dev.allocate(dim_x*dim_y*dim_z*sizeof(Type)); in test_last_dim_reductions() 118 Type* out_data = (Type*)dev.allocate(dim_x*dim_y*sizeof(Type)); in test_last_dim_reductions() 119 Eigen::TensorMap<Eigen::Tensor<Type, 3, DataLayout> > gpu_in(in_data, dim_x, dim_y, dim_z); in test_last_dim_reductions() 120 Eigen::TensorMap<Eigen::Tensor<Type, 2, DataLayout> > gpu_out(out_data, dim_x, dim_y); in test_last_dim_reductions() [all …]
|
D | cxx11_tensor_reduction_sycl.cpp | 57 int dim_x = 145; in test_first_dim_reductions_sycl() local 61 array<int, 3> tensorRange = {{dim_x, dim_y, dim_z}}; in test_first_dim_reductions_sycl() 95 int dim_x = 567; in test_last_dim_reductions_sycl() local 99 array<int, 3> tensorRange = {{dim_x, dim_y, dim_z}}; in test_last_dim_reductions_sycl() 102 array<int, 2> reduced_tensorRange = {{dim_x, dim_y}}; in test_last_dim_reductions_sycl()
|
/external/llvm-project/mlir/test/mlir-cpu-runner/ |
D | memref_reinterpret_cast.mlir | 14 %dim_x = dim %input, %c0 : memref<2x3xf32> 16 scf.parallel (%i, %j) = (%c0, %c0) to (%dim_x, %dim_y) step (%c1, %c1) {
|
D | memref_reshape.mlir | 15 %dim_x = dim %input, %c0 : memref<2x3xf32> 17 scf.parallel (%i, %j) = (%c0, %c0) to (%dim_x, %dim_y) step (%c1, %c1) {
|
/external/tensorflow/tensorflow/core/framework/ |
D | common_shape_fns.cc | 2012 const auto dim_x = i < (rank_out - rank_x) in BroadcastBinaryOpOutputShapeFnHelper() local 2018 if (!c->ValueKnown(dim_x) || !c->ValueKnown(dim_y)) { in BroadcastBinaryOpOutputShapeFnHelper() 2028 if (c->Value(dim_x) > 1) { in BroadcastBinaryOpOutputShapeFnHelper() 2033 dims.push_back(dim_x); in BroadcastBinaryOpOutputShapeFnHelper() 2040 } else if (c->Value(dim_x) == 1) { in BroadcastBinaryOpOutputShapeFnHelper() 2043 dims.push_back(dim_x); in BroadcastBinaryOpOutputShapeFnHelper() 2044 } else if (dim_y.SameHandle(dim_x)) { in BroadcastBinaryOpOutputShapeFnHelper() 2045 dims.push_back(dim_x); in BroadcastBinaryOpOutputShapeFnHelper() 2046 } else if (!c->ValueKnown(dim_x) && !c->ValueKnown(dim_y)) { in BroadcastBinaryOpOutputShapeFnHelper() 2055 } else if (c->Value(dim_x) == 1 || c->Value(dim_y) == 1) { in BroadcastBinaryOpOutputShapeFnHelper() [all …]
|
/external/tensorflow/tensorflow/python/keras/saving/saved_model/ |
D | load.py | 1097 for dim_x, dim_y in zip(x.dims, y.dims): 1098 if (dim_x != dim_y 1099 or tensor_shape.dimension_value(dim_x) is None 1103 dims.append(tensor_shape.dimension_value(dim_x))
|
/external/llvm-project/lldb/source/Plugins/LanguageRuntime/RenderScript/RenderScriptRuntime/ |
D | RenderScriptRuntime.cpp | 2118 uint32_t dim_x = alloc->dimension.get()->dim_1; in JITAllocationSize() local 2126 if (dim_x == 0) in JITAllocationSize() 2127 dim_x = 1; in JITAllocationSize() 2133 alloc->size = dim_x * dim_y * dim_z * *alloc->element.datum_size.get(); in JITAllocationSize() 2144 dim_x = dim_x == 0 ? 0 : dim_x - 1; in JITAllocationSize() 2149 *alloc->address.get(), dim_x, dim_y, dim_z); in JITAllocationSize() 3255 uint32_t dim_x = alloc->dimension.get()->dim_1; in DumpAllocation() local 3256 dim_x = dim_x == 0 ? 1 : dim_x; in DumpAllocation() 3283 for (uint32_t x = 0; x < dim_x; ++x) { in DumpAllocation()
|
/external/tensorflow/tensorflow/lite/toco/graph_transformations/ |
D | propagate_fixed_sizes.cc | 98 int dim_x = i < (rank_out - rank_x) in ComputeBinaryOperatorOutputSize() local 103 if (dim_x == -1 || dim_y == -1) { in ComputeBinaryOperatorOutputSize() 106 } else if (dim_x == 1 || dim_y == 1) { in ComputeBinaryOperatorOutputSize() 108 if (dim_x == 1 && !dim_y_is_one) { in ComputeBinaryOperatorOutputSize() 114 dims_out->push_back(dim_x); in ComputeBinaryOperatorOutputSize() 118 CHECK_EQ(dim_x, dim_y) << "Dimensions must match"; in ComputeBinaryOperatorOutputSize() 119 dims_out->push_back(dim_x); in ComputeBinaryOperatorOutputSize()
|
/external/tensorflow/tensorflow/python/eager/ |
D | function.py | 208 for dim_x, dim_y in zip(x.dims, y.dims): 209 if (dim_x != dim_y 210 or tensor_shape.dimension_value(dim_x) is None 214 dims.append(tensor_shape.dimension_value(dim_x))
|
/external/freetype/src/pshinter/ |
D | pshalgo.c | 2112 PSH_Dimension dim_x = &glyph->globals->dimension[0]; in ps_hints_apply() local 2115 FT_Fixed x_scale = dim_x->scale_mult; in ps_hints_apply()
|