/kernel/configs/ |
D | debug.config | 29 # CONFIG_UBSAN_ALIGNMENT is not set 30 # CONFIG_UBSAN_DIV_ZERO is not set 31 # CONFIG_UBSAN_TRAP is not set 32 # CONFIG_WARN_ALL_UNSEEDED_RANDOM is not set 45 # CONFIG_DEBUG_PAGEALLOC is not set 46 # CONFIG_DEBUG_KMEMLEAK_DEFAULT_OFF is not set 47 # CONFIG_DEBUG_RODATA_TEST is not set 48 # CONFIG_DEBUG_WX is not set 49 # CONFIG_KFENCE is not set 50 # CONFIG_PAGE_POISONING is not set [all …]
|
D | tiny.config | 1 # CONFIG_CC_OPTIMIZE_FOR_PERFORMANCE is not set 3 # CONFIG_KERNEL_GZIP is not set 4 # CONFIG_KERNEL_BZIP2 is not set 5 # CONFIG_KERNEL_LZMA is not set 7 # CONFIG_KERNEL_LZO is not set 8 # CONFIG_KERNEL_LZ4 is not set
|
/kernel/trace/ |
D | trace_irqsoff.c | 65 static int irqsoff_display_graph(struct trace_array *tr, int set); 68 static inline int irqsoff_display_graph(struct trace_array *tr, int set) in irqsoff_display_graph() argument 160 static int irqsoff_display_graph(struct trace_array *tr, int set) in irqsoff_display_graph() argument 164 if (!(is_graph(tr) ^ set)) in irqsoff_display_graph() 167 stop_irqsoff_tracer(irqsoff_trace, !set); in irqsoff_display_graph() 175 return start_irqsoff_tracer(irqsoff_trace, set); in irqsoff_display_graph() 457 static int register_irqsoff_function(struct trace_array *tr, int graph, int set) in register_irqsoff_function() argument 462 if (function_enabled || (!set && !(tr->trace_flags & TRACE_ITER_FUNCTION))) in register_irqsoff_function() 489 static int irqsoff_function_set(struct trace_array *tr, u32 mask, int set) in irqsoff_function_set() argument 494 if (set) in irqsoff_function_set() [all …]
|
D | trace_sched_wakeup.c | 102 static int wakeup_display_graph(struct trace_array *tr, int set) in wakeup_display_graph() argument 104 if (!(is_graph(tr) ^ set)) in wakeup_display_graph() 107 stop_func_tracer(tr, !set); in wakeup_display_graph() 112 return start_func_tracer(tr, set); in wakeup_display_graph() 232 static int register_wakeup_function(struct trace_array *tr, int graph, int set) in register_wakeup_function() argument 237 if (function_enabled || (!set && !(tr->trace_flags & TRACE_ITER_FUNCTION))) in register_wakeup_function() 264 static int wakeup_function_set(struct trace_array *tr, u32 mask, int set) in wakeup_function_set() argument 269 if (set) in wakeup_function_set() 276 static int register_wakeup_function(struct trace_array *tr, int graph, int set) in register_wakeup_function() argument 281 static int wakeup_function_set(struct trace_array *tr, u32 mask, int set) in wakeup_function_set() argument [all …]
|
D | trace_nop.c | 64 static int nop_set_flag(struct trace_array *tr, u32 old_flags, u32 bit, int set) in nop_set_flag() argument 73 set); in nop_set_flag() 80 set); in nop_set_flag()
|
D | trace_events.c | 1029 const char *sub, const char *event, int set) in __ftrace_set_clr_event_nolock() argument 1059 ret = ftrace_event_enable_disable(file, set); in __ftrace_set_clr_event_nolock() 1076 const char *sub, const char *event, int set) in __ftrace_set_clr_event() argument 1081 ret = __ftrace_set_clr_event_nolock(tr, match, sub, event, set); in __ftrace_set_clr_event() 1087 int ftrace_set_clr_event(struct trace_array *tr, char *buf, int set) in ftrace_set_clr_event() argument 1118 ret = __ftrace_set_clr_event(tr, match, sub, event, set); in ftrace_set_clr_event() 1139 int trace_set_clr_event(const char *system, const char *event, int set) in trace_set_clr_event() argument 1146 return __ftrace_set_clr_event(tr, NULL, system, event, set); in trace_set_clr_event() 1166 int set; in trace_array_set_clr_event() local 1171 set = (enable == true) ? 1 : 0; in trace_array_set_clr_event() [all …]
|
D | trace_functions.c | 396 func_set_flag(struct trace_array *tr, u32 old_flags, u32 bit, int set) in func_set_flag() argument 402 if (!!set == !!(func_flags.val & bit)) in func_set_flag() 409 new_flags = (func_flags.val & ~bit) | (set ? bit : 0); in func_set_flag()
|
D | trace.h | 429 extern int tracing_set_filter_buffering(struct trace_array *tr, bool set); 584 u32 old_flags, u32 bit, int set); 587 u32 mask, int set); 1890 int trace_keep_overwrite(struct tracer *tracer, u32 mask, int set); 1959 extern int ftrace_set_clr_event(struct trace_array *tr, char *buf, int set);
|
D | trace_functions_graph.c | 1324 func_graph_set_flag(struct trace_array *tr, u32 old_flags, u32 bit, int set) in func_graph_set_flag() argument 1327 ftrace_graph_skip_irqs = !set; in func_graph_set_flag() 1330 ftrace_graph_sleep_time_control(set); in func_graph_set_flag() 1333 ftrace_graph_graph_time_control(set); in func_graph_set_flag()
|
/kernel/ |
D | compat.c | 35 static inline void compat_sig_setmask(sigset_t *blocked, compat_sigset_word set) in compat_sig_setmask() argument 37 memcpy(blocked->sig, &set, sizeof(set)); in compat_sig_setmask() 250 get_compat_sigset(sigset_t *set, const compat_sigset_t __user *compat) in get_compat_sigset() argument 257 case 4: set->sig[3] = v.sig[6] | (((long)v.sig[7]) << 32 ); in get_compat_sigset() 259 case 3: set->sig[2] = v.sig[4] | (((long)v.sig[5]) << 32 ); in get_compat_sigset() 261 case 2: set->sig[1] = v.sig[2] | (((long)v.sig[3]) << 32 ); in get_compat_sigset() 263 case 1: set->sig[0] = v.sig[0] | (((long)v.sig[1]) << 32 ); in get_compat_sigset() 266 if (copy_from_user(set, compat, sizeof(compat_sigset_t))) in get_compat_sigset()
|
D | ucount.c | 32 return ¤t_user_ns()->set; in set_lookup() 35 static int set_is_seen(struct ctl_table_set *set) in set_is_seen() argument 37 return ¤t_user_ns()->set == set; in set_is_seen() 44 container_of(head->set, struct user_namespace, set); in set_permissions() 100 setup_sysctl_set(&ns->set, &set_root, set_is_seen); in setup_userns_sysctls() 107 ns->sysctls = __register_sysctl_table(&ns->set, "user", tbl, in setup_userns_sysctls() 112 retire_sysctl_set(&ns->set); in setup_userns_sysctls() 126 retire_sysctl_set(&ns->set); in retire_userns_sysctls()
|
D | params.c | 141 params[i].ops->set); in parse_one() 144 err = params[i].ops->set(val, ¶ms[i]); in parse_one() 229 .set = param_set_##name, \ 301 .set = param_set_charp, 327 .set = param_set_bool, 358 .set = param_set_bool_enable_only, 385 .set = param_set_invbool, 408 .set = param_set_bint, 419 int (*set)(const char *, const struct kernel_param *kp), in param_array() 447 ret = set(val, &kp); in param_array() [all …]
|
D | cred.c | 352 static bool cred_cap_issubset(const struct cred *set, const struct cred *subset) in cred_cap_issubset() argument 354 const struct user_namespace *set_ns = set->user_ns; in cred_cap_issubset() 361 return cap_issubset(subset->cap_permitted, set->cap_permitted); in cred_cap_issubset() 370 uid_eq(subset_ns->owner, set->euid)) in cred_cap_issubset()
|
D | signal.c | 3119 int sigprocmask(int how, sigset_t *set, sigset_t *oldset) in sigprocmask() argument 3130 sigorsets(&newset, &tsk->blocked, set); in sigprocmask() 3133 sigandnsets(&newset, &tsk->blocked, set); in sigprocmask() 3136 newset = *set; in sigprocmask() 3257 static void do_sigpending(sigset_t *set) in do_sigpending() argument 3260 sigorsets(set, ¤t->pending.signal, in do_sigpending() 3265 sigandsets(set, ¤t->blocked, set); in do_sigpending() 3276 sigset_t set; in SYSCALL_DEFINE2() local 3281 do_sigpending(&set); in SYSCALL_DEFINE2() 3283 if (copy_to_user(uset, &set, sigsetsize)) in SYSCALL_DEFINE2() [all …]
|
D | watch_queue.c | 323 goto set; in watch_queue_set_filter() 372 set: in watch_queue_set_filter()
|
/kernel/irq/ |
D | devres.c | 237 unsigned int set; member 244 irq_remove_generic_chip(this->gc, this->msk, this->clr, this->set); in devm_irq_remove_generic_chip() 264 unsigned int clr, unsigned int set) in devm_irq_setup_generic_chip() argument 273 irq_setup_generic_chip(gc, msk, flags, clr, set); in devm_irq_setup_generic_chip() 278 dr->set = set; in devm_irq_setup_generic_chip()
|
D | generic-chip.c | 288 unsigned int clr, unsigned int set, in __irq_alloc_domain_generic_chips() argument 317 dgc->irq_flags_to_set = set; in __irq_alloc_domain_generic_chips() 475 unsigned int set) in irq_setup_generic_chip() argument 505 irq_modify_status(i, clr, set); in irq_setup_generic_chip() 545 unsigned int clr, unsigned int set) in irq_remove_generic_chip() argument 574 irq_modify_status(virq, clr, set); in irq_remove_generic_chip()
|
D | settings.h | 42 irq_settings_clr_and_set(struct irq_desc *desc, u32 clr, u32 set) in irq_settings_clr_and_set() argument 45 desc->status_use_accessors |= (set & _IRQF_MODIFY_MASK); in irq_settings_clr_and_set()
|
/kernel/sched/ |
D | psi.c | 785 unsigned int clear, unsigned int set, u64 now, in psi_group_change() argument 816 } else if (unlikely(set & TSK_ONCPU)) { in psi_group_change() 818 set &= ~TSK_ONCPU; in psi_group_change() 836 groupc->tasks[3], clear, set); in psi_group_change() 841 for (t = 0; set; set &= ~(1 << t), t++) in psi_group_change() 842 if (set & (1 << t)) in psi_group_change() 900 static void psi_flags_change(struct task_struct *task, int clear, int set) in psi_flags_change() argument 902 if (((task->psi_flags & set) || in psi_flags_change() 907 task->psi_flags, clear, set); in psi_flags_change() 912 task->psi_flags |= set; in psi_flags_change() [all …]
|
D | stats.h | 110 void psi_task_change(struct task_struct *task, int clear, int set); 127 int clear = 0, set = TSK_RUNNING; in psi_enqueue() local 133 set |= TSK_MEMSTALL_RUNNING; in psi_enqueue() 137 set |= TSK_MEMSTALL; in psi_enqueue() 143 psi_task_change(p, clear, set); in psi_enqueue()
|
/kernel/debug/kdb/ |
D | kdb_cmds | 4 # registers) are not reliable this early. set and bp commands should 11 set BTAPROMPT 0 12 set LINES 10000
|
/kernel/rcu/ |
D | Kconfig | 50 obscure RCU options to be set up. 174 (hundreds or thousands of CPUs) will instead want to set this 188 please understand that you may also need to set the skew_tick 248 This option offloads callback invocation from the set of CPUs 256 be used to force the kthreads to run on whatever set of CPUs is 288 Say Y here if you want to set RT priority for offloading kthreads. 316 Requires rcu_nocbs=all to be set.
|
/kernel/dma/ |
D | Kconfig | 17 # mapping fast path should select this option and set the dma_ops_bypass 125 # The only thing that is really required is a way to set an uncached bit 168 You can set the size of pernuma CMA by specifying "cma_pernuma=size" 169 or set the node id and its size of CMA by specifying "numa_cma=
|
/kernel/bpf/ |
D | btf.c | 5320 struct btf_id_set set; in btf_parse_struct_metas() member 5341 aof.set.ids[aof.set.cnt++] = id; in btf_parse_struct_metas() 5344 if (!aof.set.cnt) in btf_parse_struct_metas() 5346 sort(&aof.set.ids, aof.set.cnt, sizeof(aof.set.ids[0]), btf_id_cmp_func, NULL); in btf_parse_struct_metas() 5368 if (btf_id_set_contains(&aof.set, member->type)) in btf_parse_struct_metas() 7681 struct btf_id_set8 *add_set = kset->set; in btf_populate_kfunc_set() 7685 struct btf_id_set8 *set; in btf_populate_kfunc_set() local 7723 set = tab->sets[hook]; in btf_populate_kfunc_set() 7727 if (WARN_ON_ONCE(set && !vmlinux_set)) { in btf_populate_kfunc_set() 7748 set_cnt = set ? set->cnt : 0; in btf_populate_kfunc_set() [all …]
|
D | map_iter.c | 224 .set = &bpf_map_iter_kfunc_ids,
|