/kernel/ |
D | pid_namespace.c | 37 static struct kmem_cache *create_pid_cachep(unsigned int level) in create_pid_cachep() argument 40 struct kmem_cache **pkc = &pid_cache[level - 1]; in create_pid_cachep() 49 snprintf(name, sizeof(name), "pid_%u", level + 1); in create_pid_cachep() 50 len = sizeof(struct pid) + level * sizeof(struct upid); in create_pid_cachep() 75 unsigned int level = parent_pid_ns->level + 1; in create_pid_namespace() local 84 if (level > MAX_PID_NS_LEVEL) in create_pid_namespace() 97 ns->pid_cachep = create_pid_cachep(level); in create_pid_namespace() 107 ns->level = level; in create_pid_namespace() 410 if (new->level < active->level) in pidns_install() 414 while (ancestor->level > active->level) in pidns_install()
|
D | pid.c | 55 .level = 0, 79 .level = 0, 112 ns = pid->numbers[pid->level].ns; in put_pid() 133 for (i = 0; i <= pid->level; i++) { in free_pid() 177 if (set_tid_size > ns->level + 1) in alloc_pid() 185 pid->level = ns->level; in alloc_pid() 187 for (i = ns->level; i >= 0; i--) { in alloc_pid() 191 tid = set_tid[ns->level - i]; in alloc_pid() 268 upid = pid->numbers + ns->level; in alloc_pid() 287 while (++i <= ns->level) { in alloc_pid() [all …]
|
D | params.c | 132 if (params[i].level < min_level in parse_one() 133 || params[i].level > max_level) in parse_one() 419 s16 level, in param_array() argument 429 kp.level = level; in param_array() 468 arr->elemsize, arr->ops->set, kp->level, in param_array_set()
|
D | user_namespace.c | 90 if (parent_ns->level > 32) in create_user_ns() 130 ns->level = parent_ns->level + 1; in create_user_ns() 1299 for (ns = child; ns->level > ancestor->level; ns = ns->parent) in in_userns()
|
D | Kconfig.preempt | 69 low level and critical code paths (entry code, scheduler, low 70 level interrupt handling) fully preemptible and brings most
|
/kernel/printk/ |
D | index.c | 77 int level = LOGLEVEL_DEFAULT; in pi_show() local 89 if (entry->level) in pi_show() 90 printk_parse_prefix(entry->level, &level, &flags); in pi_show() 92 prefix_len = printk_parse_prefix(entry->fmt, &level, &flags); in pi_show() 101 if (level == LOGLEVEL_DEFAULT) in pi_show() 104 seq_printf(s, "<%d,c>", level); in pi_show() 106 seq_printf(s, "<%d>", level); in pi_show()
|
D | printk.c | 577 (info->facility << 3) | info->level, info->seq, in info_print_ext_header() 650 int devkmsg_emit(int facility, int level, const char *fmt, ...) in devkmsg_emit() argument 656 r = vprintk_emit(facility, level, NULL, fmt, args); in devkmsg_emit() 665 int level = default_message_loglevel; in devkmsg_write() local 711 level = LOG_LEVEL(u); in devkmsg_write() 719 devkmsg_emit(facility, level, "%s", line); in devkmsg_write() 1065 dest_r.info->level = r->info->level; in add_to_rb() 1209 static bool suppress_message_printing(int level) in suppress_message_printing() argument 1211 return (level >= console_loglevel && !ignore_loglevel); in suppress_message_printing() 1237 static void boot_delay_msec(int level) in boot_delay_msec() argument [all …]
|
D | internal.h | 16 int vprintk_store(int facility, int level, 39 u16 printk_parse_prefix(const char *text, int *level,
|
D | printk_ringbuffer.h | 21 u8 level:3; /* syslog level */ member
|
/kernel/power/ |
D | main.c | 243 int level; in pm_test_show() local 245 for (level = TEST_FIRST; level <= TEST_MAX; level++) in pm_test_show() 246 if (pm_tests[level]) { in pm_test_show() 247 if (level == pm_test_level) in pm_test_show() 248 s += sprintf(s, "[%s] ", pm_tests[level]); in pm_test_show() 250 s += sprintf(s, "%s ", pm_tests[level]); in pm_test_show() 264 int level; in pm_test_store() local 274 level = TEST_FIRST; in pm_test_store() 275 for (s = &pm_tests[level]; level <= TEST_MAX; s++, level++) in pm_test_store() 277 pm_test_level = level; in pm_test_store()
|
D | hibernate.c | 128 static int hibernation_test(int level) in hibernation_test() argument 130 if (pm_test_level == level) { in hibernation_test() 137 static int hibernation_test(int level) { return 0; } in hibernation_test() argument
|
D | suspend.c | 325 static int suspend_test(int level) in suspend_test() argument 328 if (pm_test_level == level) { in suspend_test()
|
/kernel/rcu/ |
D | srcutree.c | 87 int level = 0; in init_srcu_struct_nodes() local 97 ssp->level[0] = &ssp->node[0]; in init_srcu_struct_nodes() 99 ssp->level[i] = ssp->level[i - 1] + num_rcu_lvl[i - 1]; in init_srcu_struct_nodes() 121 if (snp == ssp->level[level + 1]) in init_srcu_struct_nodes() 122 level++; in init_srcu_struct_nodes() 123 snp->srcu_parent = ssp->level[level - 1] + in init_srcu_struct_nodes() 124 (snp - ssp->level[level]) / in init_srcu_struct_nodes() 125 levelspread[level - 1]; in init_srcu_struct_nodes() 134 level = rcu_num_lvls - 1; in init_srcu_struct_nodes() 135 snp_first = ssp->level[level]; in init_srcu_struct_nodes() [all …]
|
D | Kconfig | 40 bool "Make expert-level adjustments to RCU configuration" 44 expert-level adjustments to RCU configuration. By default, 50 Say Y if you need to make expert-level adjustments to RCU. 139 int "Tree-based hierarchical RCU leaf-level fanout value" 147 This option controls the leaf-level fanout of hierarchical 151 want the default because the smaller leaf-level fanout keeps 159 leaf-level fanouts work well. That said, setting leaf-level 161 lock contention on the leaf-level rcu_node structures unless
|
D | tree.h | 85 u8 level; /* root is at level 0. */ member 307 struct rcu_node *level[RCU_NUM_LVLS + 1]; member 450 static int rcu_nocb_need_deferred_wakeup(struct rcu_data *rdp, int level);
|
D | tree.c | 87 .level = { &rcu_state.node[0] }, 1265 __func__, rnp->grplo, rnp->grphi, rnp->level, in rcu_implicit_dynticks_qs() 1349 gp_seq_req, rnp->level, in trace_rcu_this_gp() 1861 rnp->level, rnp->grplo, in rcu_gp_init() 2213 mask, rnp->qsmask, rnp->level, in rcu_report_qs_rnp() 4641 rcu_state.level[i] = in rcu_init_one() 4642 rcu_state.level[i - 1] + num_rcu_lvl[i - 1]; in rcu_init_one() 4649 rnp = rcu_state.level[i]; in rcu_init_one() 4673 rnp->parent = rcu_state.level[i - 1] + in rcu_init_one() 4676 rnp->level = i; in rcu_init_one() [all …]
|
D | rcu.h | 314 #define rcu_first_leaf_node() (rcu_state.level[rcu_num_lvls - 1]) 317 #define rcu_is_leaf_node(rnp) ((rnp)->level == rcu_num_lvls - 1)
|
D | tree_exp.h | 310 trace_rcu_exp_funnel_lock(rcu_state.name, rnp->level, in exp_funnel_lock() 319 trace_rcu_exp_funnel_lock(rcu_state.name, rnp->level, in exp_funnel_lock() 657 rnp->level, rnp->grplo, rnp->grphi, in synchronize_rcu_expedited_wait()
|
D | tree_nocb.h | 980 static int rcu_nocb_need_deferred_wakeup(struct rcu_data *rdp, int level) in rcu_nocb_need_deferred_wakeup() argument 982 return READ_ONCE(rdp->nocb_defer_wakeup) >= level; in rcu_nocb_need_deferred_wakeup() 987 struct rcu_data *rdp, int level, in do_nocb_deferred_wakeup_common() argument 994 if (!rcu_nocb_need_deferred_wakeup(rdp_gp, level)) { in do_nocb_deferred_wakeup_common() 1637 static int rcu_nocb_need_deferred_wakeup(struct rcu_data *rdp, int level) in rcu_nocb_need_deferred_wakeup() argument
|
/kernel/bpf/ |
D | cgroup.c | 1432 int __cgroup_bpf_run_filter_setsockopt(struct sock *sk, int *level, in __cgroup_bpf_run_filter_setsockopt() argument 1440 .level = *level, in __cgroup_bpf_run_filter_setsockopt() 1496 *level = ctx.level; in __cgroup_bpf_run_filter_setsockopt() 1532 int __cgroup_bpf_run_filter_getsockopt(struct sock *sk, int level, in __cgroup_bpf_run_filter_getsockopt() argument 1541 .level = level, in __cgroup_bpf_run_filter_getsockopt() 1635 int __cgroup_bpf_run_filter_getsockopt_kern(struct sock *sk, int level, in __cgroup_bpf_run_filter_getsockopt_kern() argument 1642 .level = level, in __cgroup_bpf_run_filter_getsockopt_kern() 2018 case offsetof(struct bpf_sockopt, level): in cg_sockopt_is_valid_access() 2075 case offsetof(struct bpf_sockopt, level): in cg_sockopt_convert_ctx_access() 2077 *insn++ = CG_SOCKOPT_ACCESS_FIELD(BPF_STX_MEM, level); in cg_sockopt_convert_ctx_access() [all …]
|
D | verifier.c | 299 if (log->level == BPF_LOG_KERNEL) { in bpf_verifier_vlog() 1828 if (env->log.level & BPF_LOG_LEVEL2) in add_subprog_and_kfunc() 2183 if (env->log.level & BPF_LOG_LEVEL) { in backtrack_insn() 2581 if (env->log.level & BPF_LOG_LEVEL) in __mark_chain_precision() 2697 if (env->log.level & BPF_LOG_LEVEL) { in __mark_chain_precision() 3420 if (env->log.level & BPF_LOG_LEVEL) in check_mem_region_access() 6000 if (env->log.level & BPF_LOG_LEVEL) in __check_func_call() 6083 if (env->log.level & BPF_LOG_LEVEL) { in __check_func_call() 6270 if (env->log.level & BPF_LOG_LEVEL) { in prepare_func_exit() 9376 if (env->log.level & BPF_LOG_LEVEL) in check_cond_jmp_op() [all …]
|
/kernel/irq/ |
D | Kconfig | 21 # Print level/edge extra information 78 # Support for hierarchical fasteoi+edge and fasteoi+level handlers 150 Allow to specify the low level IRQ handler at run time.
|
/kernel/sched/ |
D | topology.c | 39 static int sched_domain_debug_one(struct sched_domain *sd, int cpu, int level, in sched_domain_debug_one() argument 48 printk(KERN_DEBUG "%*s domain-%d: ", level, "", level); in sched_domain_debug_one() 74 printk(KERN_DEBUG "%*s groups:", level + 1, ""); in sched_domain_debug_one() 137 int level = 0; in sched_domain_debug() local 150 if (sched_domain_debug_one(sd, cpu, level, sched_domains_tmpmask)) in sched_domain_debug() 152 level++; in sched_domain_debug() 1401 if (sd->level > request) { in set_domain_attribute() 2106 sd->level = child->level + 1; in build_sched_domain() 2107 sched_domain_level_max = max(sched_domain_level_max, sd->level); in build_sched_domain()
|
/kernel/debug/kdb/ |
D | kdb_cmds | 7 # Standard debugging information for first level support, just type archkdb
|
/kernel/trace/ |
D | trace_events_filter.c | 1561 int level = open; in calc_stack() local 1572 if (level == open) { in calc_stack() 1576 level--; in calc_stack() 1579 level++; in calc_stack()
|