/external/valgrind/callgrind/ |
D | jumps.c | 87 UInt new_idx; in resize_jcc_table() local 104 new_idx = jcc_hash_idx(curr_jcc->from, curr_jcc->jmp, in resize_jcc_table() 107 curr_jcc->next_hash = new_table[new_idx]; in resize_jcc_table() 108 new_table[new_idx] = curr_jcc; in resize_jcc_table() 139 UInt new_idx; in new_jcc() local 170 new_idx = jcc_hash_idx(from, jmp, to, current_jccs.size); in new_jcc() 171 jcc->next_hash = current_jccs.table[new_idx]; in new_jcc() 172 current_jccs.table[new_idx] = jcc; in new_jcc()
|
D | context.c | 90 UInt new_idx; in resize_cxt_table() local 106 new_idx = (UInt) (curr->hash % new_size); in resize_cxt_table() 108 curr->next = new_table[new_idx]; in resize_cxt_table() 109 new_table[new_idx] = curr; in resize_cxt_table()
|
D | bb.c | 71 UInt new_idx; in resize_bb_table() local 87 new_idx = bb_hash_idx(curr->obj, curr->offset, new_size); in resize_bb_table() 89 curr->next = new_table[new_idx]; in resize_bb_table() 90 new_table[new_idx] = curr; in resize_bb_table()
|
D | bbcc.c | 197 UInt new_idx; in resize_bbcc_hash() local 214 new_idx = bbcc_hash_idx(curr_BBCC->bb, in resize_bbcc_hash() 218 curr_BBCC->next = new_table[new_idx]; in resize_bbcc_hash() 219 new_table[new_idx] = curr_BBCC; in resize_bbcc_hash()
|
/external/vulkan-validation-layers/loader/ |
D | extension_manual.c | 1034 for (uint32_t new_idx = 0; new_idx < total_count; new_idx++) { in setupLoaderTrampPhysDevGroups() local 1037 …if (local_phys_dev_groups[new_idx].physicalDeviceCount == inst->phys_dev_groups_tramp[old_idx]->ph… in setupLoaderTrampPhysDevGroups() 1041 …for (uint32_t new_gpu = 0; new_gpu < local_phys_dev_groups[new_idx].physicalDeviceCount; new_gpu++… in setupLoaderTrampPhysDevGroups() 1042 …if (local_phys_dev_groups[new_idx].physicalDevices[new_gpu] == inst->phys_dev_groups_tramp[old_idx… in setupLoaderTrampPhysDevGroups() 1056 new_phys_dev_groups[new_idx] = inst->phys_dev_groups_tramp[old_idx]; in setupLoaderTrampPhysDevGroups() 1063 if (NULL == new_phys_dev_groups[new_idx]) { in setupLoaderTrampPhysDevGroups() 1064 … new_phys_dev_groups[new_idx] = (VkPhysicalDeviceGroupPropertiesKHX *)loader_instance_heap_alloc( in setupLoaderTrampPhysDevGroups() 1066 if (NULL == new_phys_dev_groups[new_idx]) { in setupLoaderTrampPhysDevGroups() 1070 new_idx); in setupLoaderTrampPhysDevGroups() 1071 total_count = new_idx; in setupLoaderTrampPhysDevGroups() [all …]
|
D | loader.c | 5489 for (uint32_t new_idx = 0; new_idx < total_count; new_idx++) { local 5492 if (local_phys_devs[new_idx] == inst->phys_devs_tramp[old_idx]->phys_dev) { 5493 new_phys_devs[new_idx] = inst->phys_devs_tramp[old_idx]; 5499 if (NULL == new_phys_devs[new_idx]) { 5500 … new_phys_devs[new_idx] = (struct loader_physical_device_tramp *)loader_instance_heap_alloc( 5502 if (NULL == new_phys_devs[new_idx]) { 5506 new_idx); 5507 total_count = new_idx; 5513 loader_set_dispatch((void *)new_phys_devs[new_idx], inst->disp); 5514 new_phys_devs[new_idx]->this_instance = inst; [all …]
|
/external/kernel-headers/original/uapi/linux/ |
D | virtio_ring.h | 164 static inline int vring_need_event(__u16 event_idx, __u16 new_idx, __u16 old) in vring_need_event() argument 171 return (__u16)(new_idx - event_idx - 1) < (__u16)(new_idx - old); in vring_need_event()
|
/external/libxkbcommon/xkbcommon/src/xkbcomp/ |
D | keycodes.c | 88 LedNameInfo *new, xkb_led_index_t new_idx) in AddLedName() argument 99 if (old_idx == new_idx) { in AddLedName() 108 xkb_led_index_t use = (replace ? new_idx + 1 : old_idx + 1); in AddLedName() 109 xkb_led_index_t ignore = (replace ? old_idx + 1 : new_idx + 1); in AddLedName() 121 if (new_idx >= info->num_led_names) in AddLedName() 122 info->num_led_names = new_idx + 1; in AddLedName() 125 old = &info->led_names[new_idx]; in AddLedName() 131 "Using %s, ignoring %s\n", new_idx + 1, in AddLedName()
|
/external/e2fsprogs/lib/et/ |
D | et_h.awk | 170 new_idx = $2 171 curr_low += curr_sign * (new_idx - curr_idx) 172 curr_idx = new_idx
|
D | et_c.awk | 209 new_idx = $2 210 for (i = table_item_count ; i < new_idx; i++) {
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
D | while_loop_simplifier.cc | 407 for (int64 new_idx = 0; new_idx < new_to_old_tuple_idx.size(); ++new_idx) { in TryRemoveDeadWhileParams() local 408 int64 old_idx = new_to_old_tuple_idx[new_idx]; in TryRemoveDeadWhileParams() 409 old_to_new_tuple_idx[old_idx] = new_idx; in TryRemoveDeadWhileParams() 410 VLOG(2) << "Remapping tuple index " << old_idx << " to " << new_idx; in TryRemoveDeadWhileParams()
|
/external/mesa3d/src/gallium/drivers/r300/compiler/ |
D | radeon_compiler_util.c | 687 unsigned int new_idx = 0; in rc_make_conversion_swizzle() local 691 for ( ; new_idx < 4; new_idx++) { in rc_make_conversion_swizzle() 692 if (GET_BIT(new_mask, new_idx)) { in rc_make_conversion_swizzle() 693 SET_SWZ(conversion_swizzle, old_idx, new_idx); in rc_make_conversion_swizzle() 694 new_idx++; in rc_make_conversion_swizzle()
|
/external/adhd/cras/src/server/ |
D | cras_iodev_list.c | 112 uint32_t new_idx; in add_dev_to_list() local 124 new_idx = next_iodev_idx; in add_dev_to_list() 126 if (new_idx < MAX_SPECIAL_DEVICE_IDX) in add_dev_to_list() 127 new_idx = MAX_SPECIAL_DEVICE_IDX; in add_dev_to_list() 128 DL_SEARCH_SCALAR(list->iodevs, tmp, info.idx, new_idx); in add_dev_to_list() 131 new_idx++; in add_dev_to_list() 133 dev->info.idx = new_idx; in add_dev_to_list() 134 next_iodev_idx = new_idx + 1; in add_dev_to_list()
|
/external/libvpx/libvpx/vp8/decoder/ |
D | onyxd_if.c | 45 static void ref_cnt_fb(int *buf, int *idx, int new_idx); 203 static void ref_cnt_fb(int *buf, int *idx, int new_idx) { in ref_cnt_fb() argument 206 *idx = new_idx; in ref_cnt_fb() 208 buf[new_idx]++; in ref_cnt_fb()
|
/external/libvncserver/libvncserver/ |
D | tight.c | 1361 int hash_key, idx, new_idx, count; in PaletteInsert() local 1370 new_idx = idx = pnode->idx; in PaletteInsert() 1372 if (new_idx && palette.entry[new_idx-1].numPixels < count) { in PaletteInsert() 1374 palette.entry[new_idx] = palette.entry[new_idx-1]; in PaletteInsert() 1375 palette.entry[new_idx].listNode->idx = new_idx; in PaletteInsert() 1376 new_idx--; in PaletteInsert() 1378 while (new_idx && palette.entry[new_idx-1].numPixels < count); in PaletteInsert() 1379 palette.entry[new_idx].listNode = pnode; in PaletteInsert() 1380 pnode->idx = new_idx; in PaletteInsert() 1382 palette.entry[new_idx].numPixels = count; in PaletteInsert()
|
/external/swiftshader/third_party/LLVM/include/llvm/Support/ |
D | CFG.h | 173 unsigned new_idx = idx + Right; variable 174 assert(index_is_valid(new_idx) && "Iterator index out of bound"); 175 idx = new_idx;
|
/external/libvpx/libvpx/vp9/common/ |
D | vp9_onyxc_int.h | 289 static INLINE void ref_cnt_fb(RefCntBuffer *bufs, int *idx, int new_idx) { in ref_cnt_fb() argument 295 *idx = new_idx; in ref_cnt_fb() 297 bufs[new_idx].ref_count++; in ref_cnt_fb()
|
/external/tensorflow/tensorflow/contrib/boosted_trees/kernels/ |
D | training_ops.cc | 707 auto new_idx = tree_config->nodes_size(); in PruneTree() local 709 nodes_map[child_idx] = new_idx; in PruneTree() 710 child_idx = new_idx; in PruneTree()
|
/external/llvm/include/llvm/IR/ |
D | InstrTypes.h | 213 unsigned new_idx = idx + Right; variable 214 assert(index_is_valid(new_idx) && "Iterator index out of bound"); 215 idx = new_idx;
|