/external/mesa3d/src/gallium/auxiliary/pipebuffer/ |
D | pb_slab.c | 64 struct pb_slab_group *group = &slabs->groups[entry->group_index]; in pb_slab_reclaim() 101 unsigned group_index; in pb_slab_alloc() local 109 group_index = heap * slabs->num_orders + (order - slabs->min_order); in pb_slab_alloc() 110 group = &slabs->groups[group_index]; in pb_slab_alloc() 139 slab = slabs->slab_alloc(slabs->priv, heap, 1 << order, group_index); in pb_slab_alloc()
|
D | pb_slab.h | 64 unsigned group_index; /* index into pb_slabs::groups */ member 94 unsigned group_index);
|
/external/tensorflow/tensorflow/core/kernels/ |
D | random_op_gpu.h | 176 int64 group_index = thread_id; 177 int64 offset = group_index * kGroupSize; 183 gen.Skip(group_index * kGeneratorSkipPerOutputGroup); 197 group_index += total_thread_count;
|
D | random_op.cc | 123 int64 group_index; in Run() local 124 for (group_index = start_group; group_index < limit_group_full; in Run() 125 ++group_index) { in Run() 130 gen.Skip(group_index * kGeneratorSkipPerOutputGroup); in Run() 141 gen.Skip(group_index * kGeneratorSkipPerOutputGroup); in Run() 644 size_t group_index = item_id; in operator ()() local 645 size_t offset = group_index * kGroupSize; in operator ()() 654 gen.Skip(group_index * kGeneratorSkipPerOutputGroup); in operator ()() 668 group_index += total_item_count; in operator ()()
|
D | bias_op_gpu.cu.cc | 164 int32 group_index = blockIdx.x / bias_size; in BiasGradNCHW_SharedAtomics() local 167 for (int32 index = group_index * blockDim.x + threadIdx.x; in BiasGradNCHW_SharedAtomics()
|
/external/libxcam/cl_kernel/ |
D | kernel_wavelet_denoise.cl | 56 int group_index = x * 4 + y * (imageWidth / 4); 212 (*(__global uint4*)(approxOut + group_index + approxOffset)) = approx; 214 (*(__global uint4*)(approxOut + group_index)) = approx; 221 … __global uint4* dest_p = (__global uint4*)(&dest[group_index + outputUVOffset * imageWidth / 4]); 224 __global uint4* dest_p = (__global uint4*)(&dest[group_index]);
|
/external/tensorflow/tensorflow/python/distribute/ |
D | cross_device_utils.py | 450 group_index = 0 456 if is_hierarchical else aux_device_groups[group_index], num_shards)) 457 group_index = (group_index + 1) % len(aux_device_groups)
|
/external/mesa3d/src/gallium/winsys/radeon/drm/ |
D | radeon_drm_bo.h | 86 unsigned group_index);
|
D | radeon_drm_bo.c | 735 unsigned group_index) in radeon_bo_slab_alloc() argument 776 bo->u.slab.entry.group_index = group_index; in radeon_bo_slab_alloc()
|
/external/tensorflow/tensorflow/python/grappler/ |
D | hierarchical_controller.py | 621 group_index = grouping_actions[child_id][topo_op_index] 628 dag_matrix[group_index, output_group_index] += 1.0 670 group_index = grouping_actions[child_id][ 674 group_embedding[group_index, type_index] += 1 675 group_embedding[group_index, :self.hparams.max_num_outputs * self.hparams. 971 group_index = grouping_actions[topo_order_index] 972 op.device = self.devices[actions[group_index]].name
|
/external/mesa3d/src/gallium/winsys/amdgpu/drm/ |
D | amdgpu_bo.h | 131 unsigned group_index);
|
D | amdgpu_bo.c | 522 unsigned group_index) in amdgpu_bo_slab_alloc() argument 564 bo->u.slab.entry.group_index = group_index; in amdgpu_bo_slab_alloc()
|
/external/webrtc/webrtc/libjingle/xmpp/ |
D | rostermoduleimpl.cc | 464 size_t group_index = 0; in GetGroup() local 465 while(group_index < index) { in GetGroup() 467 group_index++; in GetGroup() 472 me->group_index_returned_ = static_cast<int>(group_index); in GetGroup()
|
/external/mesa3d/src/gallium/drivers/r600/sb/ |
D | sb_bc_dump.cpp | 73 sblog.print_w(++group_index, 5); in visit() 443 new_group(), group_index() { in bc_dump()
|
D | sb_pass.h | 95 unsigned new_group, group_index; variable 102 vpass(s), bc_data(bc_ptr), ndw(ndw), id(), new_group(), group_index() {} in bc_dump()
|
/external/mesa3d/src/gallium/drivers/radeonsi/ |
D | si_state.h | 363 unsigned group_index);
|