| /include/linux/ |
| D | vtime.h | 52 vtime_account_kernel(current); in vtime_account_guest_enter() 53 current->flags |= PF_VCPU; in vtime_account_guest_enter() 58 vtime_account_kernel(current); in vtime_account_guest_exit() 59 current->flags &= ~PF_VCPU; in vtime_account_guest_exit() 96 vtime_guest_enter(current); in vtime_account_guest_enter() 98 current->flags |= PF_VCPU; in vtime_account_guest_enter() 104 vtime_guest_exit(current); in vtime_account_guest_exit() 106 current->flags &= ~PF_VCPU; in vtime_account_guest_exit() 116 current->flags |= PF_VCPU; in vtime_account_guest_enter() 121 current->flags &= ~PF_VCPU; in vtime_account_guest_exit()
|
| D | resume_user_mode.h | 49 if (unlikely(task_work_pending(current))) in resume_user_mode_work() 53 if (unlikely(current->cached_requested_key)) { in resume_user_mode_work() 54 key_put(current->cached_requested_key); in resume_user_mode_work() 55 current->cached_requested_key = NULL; in resume_user_mode_work()
|
| D | delayacct.h | 112 if (current->delays) in delayacct_blkio_start() 137 if (current->delays) in delayacct_freepages_start() 146 if (current->delays) in delayacct_freepages_end() 155 if (current->delays) in delayacct_thrashing_start() 164 if (current->delays) in delayacct_thrashing_end() 173 if (current->delays) in delayacct_swapin_start() 182 if (current->delays) in delayacct_swapin_end() 191 if (current->delays) in delayacct_compact_start() 200 if (current->delays) in delayacct_compact_end() 209 if (current->delays) in delayacct_wpcopy_start() [all …]
|
| D | irqflags.h | 53 current->hardirq_threaded = 0; \ 57 current->hardirq_threaded = 1; \ 69 current->irq_config = 1; \ 78 current->irq_config = 0; \ 83 current->irq_config = 1; \ 88 current->irq_config = 0; \ 94 current->irq_config = 1; \ 99 current->irq_config = 0; \ 127 current->softirq_context++; \ 131 current->softirq_context--; \
|
| D | cpuset.h | 81 #define cpuset_current_mems_allowed (current->mems_allowed) 123 return task_spread_page(current); in cpuset_do_page_mem_spread() 144 return read_seqcount_begin(¤t->mems_allowed_seq); in read_mems_allowed_begin() 158 return read_seqcount_retry(¤t->mems_allowed_seq, seq); in read_mems_allowed_retry() 165 task_lock(current); in set_mems_allowed() 167 write_seqcount_begin(¤t->mems_allowed_seq); in set_mems_allowed() 168 current->mems_allowed = nodemask; in set_mems_allowed() 169 write_seqcount_end(¤t->mems_allowed_seq); in set_mems_allowed() 171 task_unlock(current); in set_mems_allowed()
|
| D | rseq.h | 37 if (current->rseq) in rseq_handle_notify_resume() 45 __set_bit(RSEQ_EVENT_SIGNAL_BIT, ¤t->rseq_event_mask); in rseq_signal_deliver() 76 t->rseq = current->rseq; in rseq_fork() 77 t->rseq_len = current->rseq_len; in rseq_fork() 78 t->rseq_sig = current->rseq_sig; in rseq_fork() 79 t->rseq_event_mask = current->rseq_event_mask; in rseq_fork()
|
| D | ptrace.h | 150 if (unlikely(ptrace_event_enabled(current, event))) { in ptrace_event() 154 if ((current->ptrace & (PT_PTRACED|PT_SEIZED)) == PT_PTRACED) in ptrace_event() 155 send_sig(SIGTRAP, current, 0); in ptrace_event() 182 ns = task_active_pid_ns(rcu_dereference(current->parent)); in ptrace_event_pid() 208 if (unlikely(ptrace) && current->ptrace) { in ptrace_init_task() 209 child->ptrace = current->ptrace; in ptrace_init_task() 210 __ptrace_link(child, current->parent, current->ptracer_cred); in ptrace_init_task() 389 #define current_pt_regs() task_pt_regs(current) 409 int ptrace = current->ptrace; in ptrace_report_syscall() 424 send_sig(signr, current, 1); in ptrace_report_syscall() [all …]
|
| D | trace_recursion.h | 51 #define trace_recursion_set(bit) do { (current)->trace_recursion |= (1<<(bit)); } while (0) 52 #define trace_recursion_clear(bit) do { (current)->trace_recursion &= ~(1<<(bit)); } while (0) 53 #define trace_recursion_test(bit) ((current)->trace_recursion & (1<<(bit))) 120 unsigned int val = READ_ONCE(current->trace_recursion); in trace_test_and_set_recursion() 144 current->trace_recursion = val; in trace_test_and_set_recursion()
|
| D | io_uring.h | 18 if (current->io_uring) in io_uring_files_cancel() 23 if (current->io_uring) in io_uring_task_cancel()
|
| D | sched.h | 165 current->task_state_change = _THIS_IP_; \ 171 current->task_state_change = _THIS_IP_; \ 176 current->saved_state_change = current->task_state_change;\ 177 current->task_state_change = _THIS_IP_; \ 182 current->task_state_change = current->saved_state_change;\ 232 WRITE_ONCE(current->__state, (state_value)); \ 238 smp_store_mb(current->__state, (state_value)); \ 251 raw_spin_lock_irqsave(¤t->pi_lock, flags); \ 253 WRITE_ONCE(current->__state, (state_value)); \ 254 raw_spin_unlock_irqrestore(¤t->pi_lock, flags); \ [all …]
|
| D | time_namespace.h | 70 struct timens_offsets *ns_offsets = ¤t->nsproxy->time_ns->offsets; in timens_add_monotonic() 77 struct timens_offsets *ns_offsets = ¤t->nsproxy->time_ns->offsets; in timens_add_boottime() 84 struct timens_offsets *ns_offsets = ¤t->nsproxy->time_ns->offsets; in timens_add_boottime_ns() 91 struct timens_offsets *ns_offsets = ¤t->nsproxy->time_ns->offsets; in timens_sub_boottime() 101 struct time_namespace *ns = current->nsproxy->time_ns; in timens_ktime_to_host()
|
| D | perf_regs.h | 46 regs_user->regs = task_pt_regs(current); in perf_get_regs_user() 47 regs_user->abi = perf_reg_abi(current); in perf_get_regs_user()
|
| D | task_io_accounting_ops.h | 13 current->ioac.read_bytes += bytes; in task_io_account_read() 27 current->ioac.write_bytes += bytes; in task_io_account_write() 41 current->ioac.cancelled_write_bytes += bytes; in task_io_account_cancelled_write()
|
| D | hardirq.h | 38 account_hardirq_enter(current); \ 66 account_hardirq_exit(current); \
|
| D | freezer.h | 54 if (likely(!freezing(current))) in try_to_freeze() 56 if (!(current->flags & PF_NOFREEZE)) in try_to_freeze()
|
| D | user_events.h | 40 if (!t || !current->user_event_mm) in user_events_fork() 43 old_mm = current->user_event_mm; in user_events_fork()
|
| D | rcupdate_trace.h | 50 struct task_struct *t = current; in rcu_read_lock_trace() 72 struct task_struct *t = current; in rcu_read_unlock_trace()
|
| D | ioprio.h | 66 if (p != current) in __get_task_ioprio() 84 return __get_task_ioprio(current); in get_current_ioprio()
|
| D | rcuwait.h | 36 rcu_assign_pointer(w->task, current); in prepare_to_rcuwait() 54 if (signal_pending_state(state, current)) { \
|
| D | oom.h | 62 current->signal->oom_flag_origin = true; in set_current_oom_origin() 67 current->signal->oom_flag_origin = false; in clear_current_oom_origin()
|
| /include/linux/sched/ |
| D | signal.h | 289 struct task_struct *task = current; in kernel_dequeue_signal() 303 spin_lock_irq(¤t->sighand->siglock); in kernel_signal_stop() 304 if (current->jobctl & JOBCTL_STOP_DEQUEUED) { in kernel_signal_stop() 305 current->jobctl |= JOBCTL_STOPPED; in kernel_signal_stop() 308 spin_unlock_irq(¤t->sighand->siglock); in kernel_signal_stop() 380 set_tsk_thread_flag(current, TIF_SIGPENDING); in restart_syscall() 431 (fatal_signal_pending(current) || in fault_signal_pending() 432 (user_mode(regs) && signal_pending(current)))); in fault_signal_pending() 516 current->restore_sigmask = true; in set_restore_sigmask() 524 current->restore_sigmask = false; in clear_restore_sigmask() [all …]
|
| D | mm.h | 252 unsigned int pflags = READ_ONCE(current->flags); in current_gfp_context() 332 unsigned oldflags = ~current->flags & flags; in memalloc_flags_save() 333 current->flags |= flags; in memalloc_flags_save() 339 current->flags &= ~flags; in memalloc_flags_restore() 497 old = current->active_memcg; in set_active_memcg() 498 current->active_memcg = memcg; in set_active_memcg() 534 if (current->mm != mm) in membarrier_mm_sync_core_before_usermode()
|
| /include/linux/sunrpc/ |
| D | types.h | 22 #define signalled() (signal_pending(current))
|
| /include/net/ |
| D | cls_cgroup.h | 46 classid = task_cls_classid(current); in sock_update_classid() 57 u32 classid = __task_get_classid(current); in task_get_classid()
|
| /include/asm-generic/ |
| D | current.h | 9 #define current get_current() macro
|