/kernel/printk/ |
D | printk.c | 811 struct devkmsg_user *user = file->private_data; in devkmsg_write() local 815 if (!user || len > LOG_LINE_MAX) in devkmsg_write() 824 if (!___ratelimit(&user->rs, current->comm)) in devkmsg_write() 871 struct devkmsg_user *user = file->private_data; in devkmsg_read() local 876 if (!user) in devkmsg_read() 879 ret = mutex_lock_interruptible(&user->lock); in devkmsg_read() 884 while (user->seq == log_next_seq) { in devkmsg_read() 893 user->seq != log_next_seq); in devkmsg_read() 899 if (user->seq < log_first_seq) { in devkmsg_read() 901 user->idx = log_first_idx; in devkmsg_read() [all …]
|
/kernel/bpf/ |
D | stackmap.c | 283 u64 *ips, u32 trace_nr, bool user) in stack_map_get_build_id_offset() argument 307 if (!user || !current || !current->mm || irq_work_busy || in stack_map_get_build_id_offset() 357 bool user = flags & BPF_F_USER_STACK; in BPF_CALL_3() local 358 bool kernel = !user; in BPF_CALL_3() 366 trace = get_perf_callchain(regs, init_nr, kernel, user, in BPF_CALL_3() 403 ips, trace_nr, user); in BPF_CALL_3() 452 bool user = flags & BPF_F_USER_STACK; in BPF_CALL_4() local 454 bool kernel = !user; in BPF_CALL_4() 464 elem_size = (user && user_build_id) ? sizeof(struct bpf_stack_build_id) in BPF_CALL_4() 474 trace = get_perf_callchain(regs, init_nr, kernel, user, in BPF_CALL_4() [all …]
|
D | syscall.c | 186 static int bpf_charge_memlock(struct user_struct *user, u32 pages) in bpf_charge_memlock() argument 190 if (atomic_long_add_return(pages, &user->locked_vm) > memlock_limit) { in bpf_charge_memlock() 191 atomic_long_sub(pages, &user->locked_vm); in bpf_charge_memlock() 197 static void bpf_uncharge_memlock(struct user_struct *user, u32 pages) in bpf_uncharge_memlock() argument 199 if (user) in bpf_uncharge_memlock() 200 atomic_long_sub(pages, &user->locked_vm); in bpf_uncharge_memlock() 206 struct user_struct *user; in bpf_map_charge_init() local 212 user = get_current_user(); in bpf_map_charge_init() 213 ret = bpf_charge_memlock(user, pages); in bpf_map_charge_init() 215 free_uid(user); in bpf_map_charge_init() [all …]
|
D | core.c | 232 ret = __bpf_prog_charge(fp_old->aux->user, delta); in bpf_prog_realloc() 238 __bpf_prog_uncharge(fp_old->aux->user, delta); in bpf_prog_realloc()
|
/kernel/ |
D | user.c | 124 struct user_struct *user; in uid_hash_find() local 126 hlist_for_each_entry(user, hashent, uidhash_node) { in uid_hash_find() 127 if (uid_eq(user->uid, uid)) { in uid_hash_find() 128 refcount_inc(&user->__count); in uid_hash_find() 129 return user; in uid_hash_find()
|
D | uid16.c | 23 SYSCALL_DEFINE3(chown16, const char __user *, filename, old_uid_t, user, old_gid_t, group) in SYSCALL_DEFINE3() argument 25 return ksys_chown(filename, low2highuid(user), low2highgid(group)); in SYSCALL_DEFINE3() 28 SYSCALL_DEFINE3(lchown16, const char __user *, filename, old_uid_t, user, old_gid_t, group) in SYSCALL_DEFINE3() argument 30 return ksys_lchown(filename, low2highuid(user), low2highgid(group)); in SYSCALL_DEFINE3() 33 SYSCALL_DEFINE3(fchown16, unsigned int, fd, old_uid_t, user, old_gid_t, group) in SYSCALL_DEFINE3() argument 35 return ksys_fchown(fd, low2highuid(user), low2highgid(group)); in SYSCALL_DEFINE3()
|
D | cred.c | 60 .user = INIT_USER, 121 free_uid(cred->user); in put_cred_rcu() 271 get_uid(new->user); in prepare_creds() 351 atomic_inc(&p->cred->user->processes); in copy_creds() 384 atomic_inc(&new->user->processes); in copy_creds() 485 if (new->user != old->user) in commit_creds() 486 atomic_inc(&new->user->processes); in commit_creds() 489 if (new->user != old->user) in commit_creds() 490 atomic_dec(&old->user->processes); in commit_creds() 703 get_uid(new->user); in prepare_kernel_cred()
|
D | sys.c | 201 struct user_struct *user; in SYSCALL_DEFINE3() local 239 user = cred->user; in SYSCALL_DEFINE3() 243 user = find_user(uid); in SYSCALL_DEFINE3() 244 if (!user) in SYSCALL_DEFINE3() 252 free_uid(user); /* For find_user() */ in SYSCALL_DEFINE3() 271 struct user_struct *user; in SYSCALL_DEFINE2() local 307 user = cred->user; in SYSCALL_DEFINE2() 311 user = find_user(uid); in SYSCALL_DEFINE2() 312 if (!user) in SYSCALL_DEFINE2() 323 free_uid(user); /* for find_user() */ in SYSCALL_DEFINE2() [all …]
|
D | Makefile | 8 sysctl.o sysctl_binary.o capability.o ptrace.o user.o \ 113 obj-$(CONFIG_USER_RETURN_NOTIFIER) += user-return-notifier.o
|
D | Kconfig.hz | 13 a fast response for user interaction and that may experience bus
|
D | signal.c | 415 struct user_struct *user; in __sigqueue_alloc() local 422 user = get_uid(__task_cred(t)->user); in __sigqueue_alloc() 423 atomic_inc(&user->sigpending); in __sigqueue_alloc() 427 atomic_read(&user->sigpending) <= in __sigqueue_alloc() 435 atomic_dec(&user->sigpending); in __sigqueue_alloc() 436 free_uid(user); in __sigqueue_alloc() 440 q->user = user; in __sigqueue_alloc() 450 atomic_dec(&q->user->sigpending); in __sigqueue_free() 451 free_uid(q->user); in __sigqueue_free()
|
D | fork.c | 1885 if (atomic_read(&p->real_cred->user->processes) >= in copy_process() 1887 if (p->real_cred->user != INIT_USER && in copy_process() 2290 atomic_dec(&p->cred->user->processes); in copy_process()
|
/kernel/power/ |
D | Kconfig | 27 Skip the kernel sys_sync() before freezing user processes. 30 user-space before invoking suspend. Say Y if that's your case. 44 called "hibernation" in user interfaces. STD checkpoints the 91 The partition specified here will be different for almost every user. 140 Allow user space to create, activate and deactivate wakeup source 144 int "Maximum number of user space wakeup sources (0 = no limit)" 150 bool "Garbage collector for user space wakeup sources" 180 fields of device objects from user space. If you are not a kernel 255 battery status information, and user-space programs will receive
|
D | Makefile | 13 obj-$(CONFIG_HIBERNATION) += hibernate.o snapshot.o swap.o user.o
|
/kernel/rcu/ |
D | tiny.c | 66 void rcu_sched_clock_irq(int user) in rcu_sched_clock_irq() argument 68 if (user) { in rcu_sched_clock_irq()
|
D | tree.c | 565 static void rcu_eqs_enter(bool user) in rcu_eqs_enter() argument 580 WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !user && !is_idle_task(current)); in rcu_eqs_enter() 730 static void rcu_eqs_exit(bool user) in rcu_eqs_exit() argument 747 WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !user && !is_idle_task(current)); in rcu_eqs_exit() 2230 void rcu_sched_clock_irq(int user) in rcu_sched_clock_irq() argument 2237 if (!rcu_is_cpu_rrupt_from_idle() && !user) { in rcu_sched_clock_irq() 2243 rcu_flavor_sched_clock_irq(user); in rcu_sched_clock_irq()
|
D | tree.h | 416 static void rcu_flavor_sched_clock_irq(int user);
|
D | Kconfig.debug | 91 Say N here if you need ultimate kernel/user switch latencies
|
D | tree_plugin.h | 682 static void rcu_flavor_sched_clock_irq(int user) in rcu_flavor_sched_clock_irq() argument 686 if (user || rcu_is_cpu_rrupt_from_idle()) { in rcu_flavor_sched_clock_irq() 901 static void rcu_flavor_sched_clock_irq(int user) in rcu_flavor_sched_clock_irq() argument 903 if (user || rcu_is_cpu_rrupt_from_idle()) { in rcu_flavor_sched_clock_irq()
|
D | Kconfig | 78 user-mode execution as quiescent states.
|
/kernel/events/ |
D | callchain.c | 178 get_perf_callchain(struct pt_regs *regs, u32 init_nr, bool kernel, bool user, in get_perf_callchain() argument 204 if (user) { in get_perf_callchain()
|
D | core.c | 5715 struct user_struct *user = current_user(); in perf_mmap() local 5842 user_locked = atomic_long_read(&user->locked_vm) + user_extra; in perf_mmap() 5846 } else if (atomic_long_read(&user->locked_vm) >= user_lock_limit) { in perf_mmap() 5901 atomic_long_add(user_extra, &user->locked_vm); in perf_mmap() 6565 bool user = !event->attr.exclude_callchain_user; in perf_callchain() local 6571 if (!kernel && !user) in perf_callchain() 6574 callchain = get_perf_callchain(regs, 0, kernel, user, in perf_callchain()
|
/kernel/trace/ |
D | Kconfig | 288 allowing the user to pick and choose which trace point they 468 This allows the user to add tracing events (similar to tracepoints) 509 This allows the user to add tracing events on top of userspace 514 of perf tools on user space applications. 522 This allows the user to attach BPF programs to kprobe, uprobe, and 750 tracers by executing a preempt or irq disable section with a user 765 instead of their values. This can cause problems for user space tools 766 that use this string to parse the raw data as user space does not know
|
/kernel/time/ |
D | Kconfig | 113 This is implemented at the expense of some overhead in user <-> kernel
|
/kernel/sched/ |
D | core.c | 4753 bool user, bool pi) in __sched_setscheduler() argument 4797 if (user && !capable(CAP_SYS_NICE)) { in __sched_setscheduler() 4845 if (user) { in __sched_setscheduler() 4902 if (user) { in __sched_setscheduler()
|