/external/tensorflow/tensorflow/core/distributed_runtime/ |
D | collective_param_resolver_distributed_test.cc | 145 void DefineWorkers(int num_workers, int num_devices, in DefineWorkers() argument 153 DefineWorker(config, name, device_type, num_devices); in DefineWorkers() 158 const string& device_type, int num_devices) { in DefineWorker() argument 160 for (int i = 0; i < num_devices; ++i) { in DefineWorker() 183 void DefineCollectiveParams(int num_workers, int num_devices) { in DefineCollectiveParams() argument 188 for (int di = 0; di < num_devices; ++di) { in DefineCollectiveParams() 193 cp.group.group_size = num_workers * num_devices; in DefineCollectiveParams() 205 void IssueRequests(int num_workers, int num_devices) { in IssueRequests() argument 206 const int device_count = num_workers * num_devices; in IssueRequests() 215 for (int di = 0; di < num_devices; ++di) { in IssueRequests() [all …]
|
D | device_resolver_distributed_test.cc | 144 void DefineWorkers(int num_workers, int num_devices, in DefineWorkers() argument 148 DefineWorker(name, device_type, num_devices); in DefineWorkers() 153 int num_devices) { in DefineWorker() argument 155 for (int i = 0; i < num_devices; ++i) { in DefineWorker()
|
/external/tensorflow/tensorflow/python/distribute/ |
D | all_reduce.py | 317 num_devices = len(input_tensors) 318 if num_devices == 0: 320 if num_devices == 1: 325 num_chunks = num_devices * num_subchunks 326 num_ticks = num_devices - 1 330 for d in range(0, num_devices): 339 for d in range(0, num_devices): 343 seg_index = (rank + num_devices - (2 + tick)) % num_devices 350 for d in range(0, num_devices): 353 seg_index = (rank + num_devices - (2 + tick)) % num_devices [all …]
|
D | cross_device_utils.py | 77 num_devices = len(avail_devices) 80 group_size = num_devices // 2 82 group_0_main_device = i % num_devices 83 group_1_main_device = (group_0_main_device + group_size) % num_devices 184 num_devices = len(devices) 185 if group_size > num_devices: 187 'only %d devices, but group_size=%d' % (num_devices, group_size)) 189 num_devices // group_size + (1 if (num_devices % group_size != 0) else 0)) 192 groups[i % num_groups].append(devices[i % num_devices]) 355 num_devices = len(devices) [all …]
|
D | all_reduce_test.py | 130 num_devices = len(dev_list) 132 for d in range(0, num_devices): 152 num_devices = num_workers * num_gpus 154 for _ in range(num_devices)] 158 x, constant_op.constant(num_devices, dtype=types_pb2.DT_FLOAT))
|
D | shared_variable_creator_test.py | 53 num_devices = 3 55 for i in range(num_devices):
|
/external/tensorflow/tensorflow/core/kernels/ |
D | nccl_ops.cc | 52 int num_devices() const { return num_devices_; } in num_devices() function in tensorflow::__anon49c74ebe0111::NcclAsyncOpBase 113 /*num_local_devices=*/num_devices(), in ComputeAsync() 114 /*num_global_devices=*/num_devices(), in ComputeAsync() 145 /*num_local_devices=*/num_devices(), in ComputeAsync() 146 /*num_global_devices=*/num_devices(), in ComputeAsync() 182 /*num_local_devices=*/num_devices(), in ComputeAsync() 183 /*num_global_devices=*/num_devices(), in ComputeAsync() 216 /*num_local_devices=*/num_devices(), in ComputeAsync() 217 /*num_global_devices=*/num_devices(), in ComputeAsync() 253 /*num_local_devices=*/num_devices(), in ComputeAsync() [all …]
|
/external/tensorflow/tensorflow/compiler/xla/experimental/xla_sharding/ |
D | xla_sharding.py | 92 def split(cls, tensor, split_dimension, num_devices): argument 109 shape[split_dimension] < num_devices): 112 (shape, split_dimension, num_devices)) 115 tile_assignment_dims[split_dimension] = num_devices 121 tile_assignment_devices=range(num_devices))) 184 def split(tensor, split_dimension, num_devices): argument 185 Sharding.split(tensor, split_dimension, num_devices).apply_to_tensor(tensor)
|
/external/mesa3d/src/gallium/state_trackers/clover/api/ |
D | context.cpp | 71 cl_uint num_devices; in clCreateContextFromType() local 77 ret = clGetDeviceIDs(d_platform, type, 0, NULL, &num_devices); in clCreateContextFromType() 80 devs.resize(num_devices); in clCreateContextFromType() 81 ret = clGetDeviceIDs(d_platform, type, num_devices, devs.data(), 0); in clCreateContextFromType() 85 return clCreateContext(d_props, num_devices, devs.data(), pfn_notify, in clCreateContextFromType()
|
D | dispatch.hpp | 56 cl_uint *num_devices); 67 cl_uint num_devices, 204 cl_uint num_devices, 219 cl_uint num_devices, 630 cl_uint *num_devices); 648 cl_uint *num_devices); 666 cl_uint num_devices, 673 cl_uint num_devices, 684 cl_uint num_devices,
|
/external/mesa3d/src/loader/ |
D | loader.c | 163 int i, num_devices, fd; in loader_get_user_preferred_fd() local 182 num_devices = drmGetDevices2(0, devices, MAX_DRM_DEVICES); in loader_get_user_preferred_fd() 183 if (num_devices < 0) in loader_get_user_preferred_fd() 195 for (i = 0; i < num_devices; i++) { in loader_get_user_preferred_fd() 204 for (i = 0; i < num_devices; i++) { in loader_get_user_preferred_fd() 215 drmFreeDevices(devices, num_devices); in loader_get_user_preferred_fd() 220 drmFreeDevices(devices, num_devices); in loader_get_user_preferred_fd()
|
/external/tensorflow/tensorflow/core/common_runtime/ |
D | ring_gatherer_test.cc | 133 void Init(int num_workers, int num_devices, DataType dtype, in Init() argument 145 for (int di = 0; di < num_devices; ++di) { in Init() 152 int dev_idx = (wi * num_devices) + di; in Init() 179 col_params_.group.group_size = num_workers * num_devices; in Init() 188 int subdiv_stride = num_devices / num_subdivs; in Init() 197 for (int di = 0; di < num_devices; ++di) { in Init() 200 for (int di = 0; di < num_devices; ++di) { in Init() 202 int other = (di + (is_odd ? 7 : 3)) % num_devices; in Init() 213 for (int di = 0; di < num_devices; ++di) { in Init() 228 num_devices; in Init() [all …]
|
D | ring_reducer_test.cc | 155 void Init(int num_workers, int num_devices, DataType dtype, in Init() argument 167 for (int di = 0; di < num_devices; ++di) { in Init() 174 int dev_idx = (wi * num_devices) + di; in Init() 201 col_params_.group.group_size = num_workers * num_devices; in Init() 210 int subdiv_stride = num_devices / num_subdivs; in Init() 219 for (int di = 0; di < num_devices; ++di) { in Init() 222 for (int di = 0; di < num_devices; ++di) { in Init() 224 int other = (di + (is_odd ? 7 : 3)) % num_devices; in Init() 235 for (int di = 0; di < num_devices; ++di) { in Init() 250 num_devices; in Init() [all …]
|
/external/eigen/unsupported/Eigen/CXX11/src/Tensor/ |
D | TensorDeviceCuda.h | 60 int num_devices; in initializeDeviceProp() local 61 cudaError_t status = cudaGetDeviceCount(&num_devices); in initializeDeviceProp() 68 m_deviceProperties = new cudaDeviceProp[num_devices]; in initializeDeviceProp() 69 for (int i = 0; i < num_devices; ++i) { in initializeDeviceProp() 119 int num_devices; in initializeDeviceProp() local 120 cudaError_t err = cudaGetDeviceCount(&num_devices); in initializeDeviceProp() 123 assert(device < num_devices); in initializeDeviceProp()
|
/external/tensorflow/tensorflow/core/nccl/ |
D | nccl_rewrite.cc | 36 int num_devices = node->num_inputs(); in ReplaceReduce() local 41 .Attr("num_devices", num_devices) in ReplaceReduce() 121 int num_devices = 0; // Number of distinct devices, incremented below. in ReplaceBroadcast() local 136 recv_index_map[dst_dev] = num_devices; in ReplaceBroadcast() 137 ++num_devices; in ReplaceBroadcast() 142 if (num_devices <= 1) { in ReplaceBroadcast() 164 .Attr("num_devices", num_devices) in ReplaceBroadcast()
|
/external/tensorflow/tensorflow/core/api_def/base_api/ |
D | api_def_NcclReduce.pbtxt | 3 summary: "Reduces `input` from `num_devices` using `reduction` to a single device." 5 Reduces `input` from `num_devices` using `reduction` to a single device. 11 data: the value of the reduction across all `num_devices` devices.
|
D | api_def_NcclAllReduce.pbtxt | 9 then `num_devices` ops will run with shared_name value `c`. Failure to do so 13 data: the value of the reduction across all `num_devices` devices. 15 num_devices: The number of devices participating in this reduction.
|
/external/tensorflow/tensorflow/python/ops/ |
D | nccl_ops.py | 68 num_devices = op.get_attr('num_devices') 75 num_devices=num_devices, 227 num_devices=len(tensors),
|
/external/webrtc/webrtc/tools/agc/ |
D | agc_harness.cc | 158 int num_devices = 0; in PrintDevices() local 161 RTC_CHECK_EQ(0, hardware_->GetNumOfRecordingDevices(num_devices)); in PrintDevices() 163 for (int i = 0; i < num_devices; i++) { in PrintDevices() 167 RTC_CHECK_EQ(0, hardware_->GetNumOfPlayoutDevices(num_devices)); in PrintDevices() 169 for (int i = 0; i < num_devices; i++) { in PrintDevices()
|
/external/webrtc/webrtc/tools/e2e_quality/audio/ |
D | audio_e2e_harness.cc | 70 int num_devices = 0; in RunHarness() local 71 ASSERT_EQ(0, hardware->GetNumOfPlayoutDevices(num_devices)); in RunHarness() 76 for (device_index = 0; device_index < num_devices; device_index++) { in RunHarness()
|
/external/u-boot/test/dm/ |
D | test-fdt.c | 132 int dm_check_devices(struct unit_test_state *uts, int num_devices) in dm_check_devices() argument 142 for (i = 0; i < num_devices; i++) { in dm_check_devices() 169 const int num_devices = 7; in dm_test_fdt() local 182 ut_asserteq(num_devices, list_count_items(&uc->dev_head)); in dm_test_fdt() 185 for (i = 0; i < num_devices; i++) { in dm_test_fdt() 192 ut_assertok(dm_check_devices(uts, num_devices)); in dm_test_fdt()
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
D | hlo_sharding.cc | 271 Status HloSharding::ValidateTuple(const Shape& shape, int64 num_devices) const { in ValidateTuple() 283 ShapeUtil::GetSubshape(shape, index_to_sharding.first), num_devices); in ValidateTuple() 295 Status HloSharding::Validate(const Shape& shape, int64 num_devices) const { in Validate() 296 Status status = IsTuple() ? ValidateTuple(shape, num_devices) in Validate() 297 : ValidateNonTuple(shape, num_devices); in Validate() 307 int64 num_devices) const { in ValidateNonTuple() 324 if (core >= num_devices) { in ValidateNonTuple() 326 "core ", core, " > ", num_devices, " in tile assignment")); in ValidateNonTuple()
|
D | hlo_sharding.h | 94 Status Validate(const Shape& shape, int64 num_devices) const; 252 Status ValidateTuple(const Shape& shape, int64 num_devices) const; 255 Status ValidateNonTuple(const Shape& shape, int64 num_devices) const;
|
/external/webrtc/talk/media/devices/ |
D | macdevicemanager.cc | 121 size_t num_devices = propsize / sizeof(AudioDeviceID); in GetAudioDeviceIDs() local 123 new AudioDeviceID[num_devices]); in GetAudioDeviceIDs() 133 for (size_t i = 0; i < num_devices; ++i) { in GetAudioDeviceIDs()
|
/external/blktrace/iowatcher/ |
D | tracers.c | 52 int start_blktrace(char **devices, int num_devices, char *trace_name, char *dest) in start_blktrace() argument 76 if (num_devices == 1) { in start_blktrace() 86 for (i = 0; i < num_devices; i++) { in start_blktrace()
|