/kernel/ |
D | cfi.c | 73 unsigned long index; in ptr_to_shadow() local 79 index = page - s->base; in ptr_to_shadow() 81 if (index >= SHADOW_ARR_SLOTS) in ptr_to_shadow() 84 return (int)index; in ptr_to_shadow() 89 int index) in shadow_to_ptr() argument 91 if (unlikely(index < 0 || index >= SHADOW_ARR_SLOTS)) in shadow_to_ptr() 94 return (s->base + index) << PAGE_SHIFT; in shadow_to_ptr() 99 int index) in shadow_to_check_fn() argument 101 if (unlikely(index < 0 || index >= SHADOW_ARR_SLOTS)) in shadow_to_check_fn() 104 if (unlikely(s->shadow[index] == SHADOW_INVALID)) in shadow_to_check_fn() [all …]
|
D | audit_tree.c | 36 unsigned index; /* index; upper bit indicates 'will prune' */ member 203 chunk->owners[i].index = i; in alloc_chunk() 274 int index = p->index & ~(1U<<31); in find_chunk() local 275 p -= index; in find_chunk() 309 new->owners[i].index = old->owners[j].index - j + i; in replace_chunk() 432 chunk->owners[0].index = (1U << 31); in create_chunk() 504 p->index = (chunk->count - 1) | (1U<<31); in tag_chunk() 579 if (tagged && !(p->index & (1U<<31))) in prune_tree_chunks() 621 if (node->index & (1U<<31)) { in trim_marked() 709 node->index |= 1U<<31; in audit_trim_trees() [all …]
|
D | module.c | 661 Elf_Shdr *pcpusec = &info->sechdrs[info->index.pcpu]; in percpu_modalloc() 758 if (info->sechdrs[info->index.pcpu].sh_size != 0) in percpu_modalloc() 1317 unsigned int versindex = info->index.vers; in check_version() 2164 symndx = info->index.sym; in copy_module_elf() 2377 Elf_Shdr *symsec = &info->sechdrs[info->index.sym]; in simplify_symbols() 2432 if (sym[i].st_shndx == info->index.pcpu) in simplify_symbols() 2465 info->index.sym, i, in apply_relocations() 2469 info->index.sym, i, mod); in apply_relocations() 2472 info->index.sym, i, mod); in apply_relocations() 2629 Elf_Shdr *infosec = &info->sechdrs[info->index.info]; in get_next_modinfo() [all …]
|
D | module-internal.h | 28 } index; member
|
/kernel/trace/ |
D | fgraph.c | 63 int index; in ftrace_push_return_trace() local 85 index = ++current->curr_ret_stack; in ftrace_push_return_trace() 87 current->ret_stack[index].ret = ret; in ftrace_push_return_trace() 88 current->ret_stack[index].func = func; in ftrace_push_return_trace() 89 current->ret_stack[index].calltime = calltime; in ftrace_push_return_trace() 91 current->ret_stack[index].fp = frame_pointer; in ftrace_push_return_trace() 94 current->ret_stack[index].retp = retp; in ftrace_push_return_trace() 150 int index; in ftrace_pop_return_trace() local 152 index = current->curr_ret_stack; in ftrace_pop_return_trace() 154 if (unlikely(index < 0 || index >= FTRACE_RETFUNC_DEPTH)) { in ftrace_pop_return_trace() [all …]
|
D | ftrace.c | 397 unsigned long index; member 437 if ((void *)rec >= (void *)&pg->records[pg->index]) { in function_stat_next() 576 pg->index = 0; in ftrace_profile_reset() 739 if (stat->pages->index == PROFILES_PER_PAGE) { in ftrace_profile_alloc() 745 rec = &stat->pages->records[stat->pages->index++]; in ftrace_profile_alloc() 1093 int index; member 1511 for (_____i = 0; _____i < pg->index; _____i++) { \ 1541 if (pg->index == 0 || in lookup_rec() 1543 start >= (pg->records[pg->index - 1].ip + MCOUNT_INSN_SIZE)) in lookup_rec() 1545 rec = bsearch(&key, pg->records, pg->index, in lookup_rec() [all …]
|
D | bpf_trace.c | 809 u64 index = flags & BPF_F_INDEX_MASK; in get_map_perf_counter() local 814 if (index == BPF_F_CURRENT_CPU) in get_map_perf_counter() 815 index = cpu; in get_map_perf_counter() 816 if (unlikely(index >= array->map.max_entries)) in get_map_perf_counter() 819 ee = READ_ONCE(array->ptrs[index]); in get_map_perf_counter() 882 u64 index = flags & BPF_F_INDEX_MASK; in __bpf_perf_event_output() local 886 if (index == BPF_F_CURRENT_CPU) in __bpf_perf_event_output() 887 index = cpu; in __bpf_perf_event_output() 888 if (unlikely(index >= array->map.max_entries)) in __bpf_perf_event_output() 891 ee = READ_ONCE(array->ptrs[index]); in __bpf_perf_event_output()
|
D | trace_probe.c | 150 trace_probe_log.index = 0; in trace_probe_log_init() 158 void trace_probe_log_set_index(int index) in trace_probe_log_set_index() argument 160 trace_probe_log.index = index; in trace_probe_log_set_index() 173 if (i == trace_probe_log.index) in __trace_probe_log_err() 181 if (trace_probe_log.index >= trace_probe_log.argc) { in __trace_probe_log_err()
|
/kernel/dma/ |
D | swiotlb.c | 478 static unsigned int wrap_index(unsigned int index) in wrap_index() argument 480 if (index >= io_tlb_nslabs) in wrap_index() 482 return index; in wrap_index() 499 unsigned int index, wrap, count = 0, i; in find_slots() local 517 index = wrap = wrap_index(ALIGN(io_tlb_index, stride)); in find_slots() 519 if ((slot_addr(tbl_dma_addr, index) & iotlb_align_mask) != in find_slots() 521 index = wrap_index(index + 1); in find_slots() 530 if (!iommu_is_span_boundary(index, nslots, in find_slots() 533 if (io_tlb_list[index] >= nslots) in find_slots() 536 index = wrap_index(index + stride); in find_slots() [all …]
|
/kernel/irq/ |
D | timings.c | 384 int index, i, period_max, count, start, min = INT_MAX; in __irq_timings_next_event() local 422 int index = (start + i) & IRQ_TIMINGS_MASK; in __irq_timings_next_event() local 424 irqs->timings[i] = irqs->circ_timings[index]; in __irq_timings_next_event() 428 index = irq_timings_next_event_index(irqs->timings, count, period_max); in __irq_timings_next_event() 429 if (index < 0) in __irq_timings_next_event() 432 return irqs->last_ts + irqs->ema_time[index]; in __irq_timings_next_event() 449 int index; in __irq_timings_store() local 454 index = irq_timings_interval_index(interval); in __irq_timings_store() 456 if (index > PREDICTION_BUFFER_SIZE - 1) { in __irq_timings_store() 465 irqs->circ_timings[irqs->count & IRQ_TIMINGS_MASK] = index; in __irq_timings_store() [all …]
|
/kernel/cgroup/ |
D | rdma.c | 83 int index, int new_max) in set_resource_limit() argument 86 if (rpool->resources[index].max != S32_MAX) in set_resource_limit() 89 if (rpool->resources[index].max == S32_MAX) in set_resource_limit() 92 rpool->resources[index].max = new_max; in set_resource_limit() 164 enum rdmacg_resource_type index) in uncharge_cg_locked() argument 180 rpool->resources[index].usage--; in uncharge_cg_locked() 186 WARN_ON_ONCE(rpool->resources[index].usage < 0); in uncharge_cg_locked() 208 enum rdmacg_resource_type index) in rdmacg_uncharge_hierarchy() argument 215 uncharge_cg_locked(p, device, index); in rdmacg_uncharge_hierarchy() 229 enum rdmacg_resource_type index) in rdmacg_uncharge() argument [all …]
|
D | cgroup-v1.c | 402 int index = 0, pid = *pos; in cgroup_pidlist_start() local 430 while (index < end) { in cgroup_pidlist_start() 431 int mid = (index + end) / 2; in cgroup_pidlist_start() 433 index = mid; in cgroup_pidlist_start() 436 index = mid + 1; in cgroup_pidlist_start() 442 if (index >= l->length) in cgroup_pidlist_start() 445 iter = l->list + index; in cgroup_pidlist_start()
|
/kernel/sched/ |
D | cpuacct.c | 96 enum cpuacct_stat_index index) in cpuacct_cpuusage_read() argument 106 BUG_ON(index > CPUACCT_STAT_NSTATS); in cpuacct_cpuusage_read() 115 switch (index) { in cpuacct_cpuusage_read() 162 enum cpuacct_stat_index index) in __cpuusage_read() argument 169 totalcpuusage += cpuacct_cpuusage_read(ca, i, index); in __cpuusage_read() 210 enum cpuacct_stat_index index) in __cpuacct_percpu_seq_show() argument 217 percpu = cpuacct_cpuusage_read(ca, i, index); in __cpuacct_percpu_seq_show() 242 int index; in cpuacct_all_seq_show() local 246 for (index = 0; index < CPUACCT_STAT_NSTATS; index++) in cpuacct_all_seq_show() 247 seq_printf(m, " %s", cpuacct_stat_desc[index]); in cpuacct_all_seq_show() [all …]
|
D | cputime.c | 103 static inline void task_group_account_field(struct task_struct *p, int index, in task_group_account_field() argument 112 __this_cpu_add(kernel_cpustat.cpustat[index], tmp); in task_group_account_field() 114 cgroup_account_cputime_field(p, index, tmp); in task_group_account_field() 124 int index; in account_user_time() local 130 index = (task_nice(p) > 0) ? CPUTIME_NICE : CPUTIME_USER; in account_user_time() 133 task_group_account_field(p, index, cputime); in account_user_time() 173 u64 cputime, enum cpu_usage_stat index) in account_system_index_time() argument 180 task_group_account_field(p, index, cputime); in account_system_index_time() 197 int index; in account_system_time() local 205 index = CPUTIME_IRQ; in account_system_time() [all …]
|
/kernel/bpf/ |
D | reuseport_array.c | 55 u32 index = *(u32 *)key; in reuseport_array_lookup_elem() local 57 if (unlikely(index >= array->map.max_entries)) in reuseport_array_lookup_elem() 60 return rcu_dereference(array->ptrs[index]); in reuseport_array_lookup_elem() 67 u32 index = *(u32 *)key; in reuseport_array_delete_elem() local 71 if (index >= map->max_entries) in reuseport_array_delete_elem() 74 if (!rcu_access_pointer(array->ptrs[index])) in reuseport_array_delete_elem() 79 sk = rcu_dereference_protected(array->ptrs[index], in reuseport_array_delete_elem() 84 RCU_INIT_POINTER(array->ptrs[index], NULL); in reuseport_array_delete_elem() 255 u32 index = *(u32 *)key; in bpf_fd_reuseport_array_update_elem() local 263 if (index >= map->max_entries) in bpf_fd_reuseport_array_update_elem() [all …]
|
D | arraymap.c | 178 u32 index = *(u32 *)key; in array_map_lookup_elem() local 180 if (unlikely(index >= array->map.max_entries)) in array_map_lookup_elem() 183 return array->value + array->elem_size * (index & array->index_mask); in array_map_lookup_elem() 224 const int index = BPF_REG_2; in array_map_gen_lookup() local 230 *insn++ = BPF_LDX_MEM(BPF_W, ret, index, 0); in array_map_gen_lookup() 253 u32 index = *(u32 *)key; in percpu_array_map_lookup_elem() local 255 if (unlikely(index >= array->map.max_entries)) in percpu_array_map_lookup_elem() 258 return this_cpu_ptr(array->pptrs[index & array->index_mask]); in percpu_array_map_lookup_elem() 264 u32 index = *(u32 *)key; in bpf_percpu_array_copy() local 269 if (unlikely(index >= array->map.max_entries)) in bpf_percpu_array_copy() [all …]
|
D | queue_stack_maps.c | 147 u32 index; in __stack_map_get() local 162 index = qs->head - 1; in __stack_map_get() 163 if (unlikely(index >= qs->size)) in __stack_map_get() 164 index = qs->size - 1; in __stack_map_get() 166 ptr = &qs->elements[index * qs->map.value_size]; in __stack_map_get() 170 qs->head = index; in __stack_map_get()
|
D | bpf_iter.c | 113 p = seq->op->start(seq, &seq->index); in bpf_seq_read() 139 loff_t pos = seq->index; in bpf_seq_read() 143 p = seq->op->next(seq, p, &seq->index); in bpf_seq_read() 144 if (pos == seq->index) { in bpf_seq_read() 148 seq->index++; in bpf_seq_read()
|
D | cpumap.c | 642 u32 index = key ? *(u32 *)key : U32_MAX; in cpu_map_get_next_key() local 645 if (index >= cmap->map.max_entries) { in cpu_map_get_next_key() 650 if (index == cmap->map.max_entries - 1) in cpu_map_get_next_key() 652 *next = index + 1; in cpu_map_get_next_key()
|
/kernel/power/ |
D | main.c | 332 int index; in last_failed_dev_show() local 335 index = suspend_stats.last_failed_dev + REC_FAILED_NUM - 1; in last_failed_dev_show() 336 index %= REC_FAILED_NUM; in last_failed_dev_show() 337 last_failed_dev = suspend_stats.failed_devs[index]; in last_failed_dev_show() 346 int index; in last_failed_errno_show() local 349 index = suspend_stats.last_failed_errno + REC_FAILED_NUM - 1; in last_failed_errno_show() 350 index %= REC_FAILED_NUM; in last_failed_errno_show() 351 last_failed_errno = suspend_stats.errno[index]; in last_failed_errno_show() 360 int index; in last_failed_step_show() local 364 index = suspend_stats.last_failed_step + REC_FAILED_NUM - 1; in last_failed_step_show() [all …]
|
D | snapshot.c | 465 int index; in add_rtree_block() local 475 index = block_nr >> ((i - 1) * BM_RTREE_LEVEL_SHIFT); in add_rtree_block() 476 index &= BM_RTREE_LEVEL_MASK; in add_rtree_block() 477 dst = (struct rtree_node **)&((*dst)->data[index]); in add_rtree_block() 751 int index; in memory_bm_find_bit() local 753 index = block_nr >> ((i - 1) * BM_RTREE_LEVEL_SHIFT); in memory_bm_find_bit() 754 index &= BM_RTREE_LEVEL_MASK; in memory_bm_find_bit() 755 BUG_ON(node->data[index] == 0); in memory_bm_find_bit() 756 node = (struct rtree_node *)node->data[index]; in memory_bm_find_bit()
|
/kernel/printk/ |
D | printk.c | 2186 if (strcmp(c->name, name) == 0 && c->index == idx) { in __add_preferred_console() 2203 c->index = idx; in __add_preferred_console() 2652 struct tty_driver *console_device(int *index) in console_device() argument 2661 driver = c->device(c, index); in console_device() 2722 newcon->match(newcon, c->name, c->index, c->options) != 0) { in try_enable_new_console() 2727 if (newcon->index >= 0 && in try_enable_new_console() 2728 newcon->index != c->index) in try_enable_new_console() 2730 if (newcon->index < 0) in try_enable_new_console() 2731 newcon->index = c->index; in try_enable_new_console() 2786 bcon->name, bcon->index)) in register_console() [all …]
|
D | console_cmdline.h | 8 int index; /* Minor dev. to use */ member
|
D | braille.c | 44 rtn = braille_register_console(console, c->index, c->options, in _braille_register_console()
|
/kernel/time/ |
D | hrtimer.c | 74 .index = HRTIMER_BASE_MONOTONIC, 79 .index = HRTIMER_BASE_REALTIME, 84 .index = HRTIMER_BASE_BOOTTIME, 89 .index = HRTIMER_BASE_TAI, 94 .index = HRTIMER_BASE_MONOTONIC_SOFT, 99 .index = HRTIMER_BASE_REALTIME_SOFT, 104 .index = HRTIMER_BASE_BOOTTIME_SOFT, 109 .index = HRTIMER_BASE_TAI_SOFT, 230 int basenum = base->index; in switch_hrtimer_base() 984 base->cpu_base->active_bases |= 1 << base->index; in enqueue_hrtimer() [all …]
|