/mm/ |
D | mmu_context.c | 23 struct task_struct *tsk = current; in use_mm() local 25 task_lock(tsk); in use_mm() 26 active_mm = tsk->active_mm; in use_mm() 29 tsk->active_mm = mm; in use_mm() 31 tsk->mm = mm; in use_mm() 32 switch_mm(active_mm, mm, tsk); in use_mm() 33 task_unlock(tsk); in use_mm() 53 struct task_struct *tsk = current; in unuse_mm() local 55 task_lock(tsk); in unuse_mm() 57 tsk->mm = NULL; in unuse_mm() [all …]
|
D | memory-failure.c | 265 struct task_struct *tsk; member 280 static void add_to_kill(struct task_struct *tsk, struct page *p, in add_to_kill() argument 309 page_to_pfn(p), tsk->comm); in add_to_kill() 312 get_task_struct(tsk); in add_to_kill() 313 tk->tsk = tsk; in add_to_kill() 341 pfn, tk->tsk->comm, tk->tsk->pid); in kill_procs() 342 force_sig(SIGKILL, tk->tsk); in kill_procs() 351 else if (kill_proc(tk->tsk, tk->addr, trapno, in kill_procs() 355 pfn, tk->tsk->comm, tk->tsk->pid); in kill_procs() 357 put_task_struct(tk->tsk); in kill_procs() [all …]
|
D | gup.c | 321 static int faultin_page(struct task_struct *tsk, struct vm_area_struct *vma, in faultin_page() argument 353 if (tsk) { in faultin_page() 355 tsk->maj_flt++; in faultin_page() 357 tsk->min_flt++; in faultin_page() 477 long __get_user_pages(struct task_struct *tsk, struct mm_struct *mm, in __get_user_pages() argument 544 ret = faultin_page(tsk, vma, start, &foll_flags, in __get_user_pages() 617 int fixup_user_fault(struct task_struct *tsk, struct mm_struct *mm, in fixup_user_fault() argument 642 if (tsk) { in fixup_user_fault() 644 tsk->maj_flt++; in fixup_user_fault() 646 tsk->min_flt++; in fixup_user_fault() [all …]
|
D | oom_kill.c | 61 struct task_struct *tsk; in has_intersects_mems_allowed() local 65 for_each_thread(start, tsk) { in has_intersects_mems_allowed() 73 ret = mempolicy_nodemask_intersects(tsk, mask); in has_intersects_mems_allowed() 79 ret = cpuset_mems_allowed_intersects(current, tsk); in has_intersects_mems_allowed() 89 static bool has_intersects_mems_allowed(struct task_struct *tsk, in has_intersects_mems_allowed() argument 417 void mark_oom_victim(struct task_struct *tsk) in mark_oom_victim() argument 421 if (test_and_set_tsk_thread_flag(tsk, TIF_MEMDIE)) in mark_oom_victim() 429 __thaw_task(tsk); in mark_oom_victim()
|
D | nommu.c | 135 long __get_user_pages(struct task_struct *tsk, struct mm_struct *mm, in __get_user_pages() argument 185 long get_user_pages(struct task_struct *tsk, struct mm_struct *mm, in get_user_pages() argument 190 return __get_user_pages(tsk, mm, start, nr_pages, in get_user_pages() 195 long get_user_pages_locked(struct task_struct *tsk, struct mm_struct *mm, in get_user_pages_locked() argument 200 return get_user_pages(tsk, mm, start, nr_pages, gup_flags, in get_user_pages_locked() 205 long __get_user_pages_unlocked(struct task_struct *tsk, struct mm_struct *mm, in __get_user_pages_unlocked() argument 211 ret = __get_user_pages(tsk, mm, start, nr_pages, gup_flags, pages, in __get_user_pages_unlocked() 218 long get_user_pages_unlocked(struct task_struct *tsk, struct mm_struct *mm, in get_user_pages_unlocked() argument 222 return __get_user_pages_unlocked(tsk, mm, start, nr_pages, in get_user_pages_unlocked() 1935 static int __access_remote_vm(struct task_struct *tsk, struct mm_struct *mm, in __access_remote_vm() argument [all …]
|
D | page-writeback.c | 367 struct task_struct *tsk; in domain_dirty_limits() local 401 tsk = current; in domain_dirty_limits() 402 if (tsk->flags & PF_LESS_THROTTLE || rt_task(tsk)) { in domain_dirty_limits() 443 struct task_struct *tsk = current; in zone_dirty_limit() local 452 if (tsk->flags & PF_LESS_THROTTLE || rt_task(tsk)) in zone_dirty_limit()
|
D | mempolicy.c | 431 void mpol_rebind_task(struct task_struct *tsk, const nodemask_t *new, in mpol_rebind_task() argument 434 mpol_rebind_policy(tsk->mempolicy, new, step); in mpol_rebind_task() 1903 bool mempolicy_nodemask_intersects(struct task_struct *tsk, in mempolicy_nodemask_intersects() argument 1911 task_lock(tsk); in mempolicy_nodemask_intersects() 1912 mempolicy = tsk->mempolicy; in mempolicy_nodemask_intersects() 1933 task_unlock(tsk); in mempolicy_nodemask_intersects()
|
D | vmscan.c | 3429 struct task_struct *tsk = current; in kswapd() local 3439 set_cpus_allowed_ptr(tsk, cpumask); in kswapd() 3454 tsk->flags |= PF_MEMALLOC | PF_SWAPWRITE | PF_KSWAPD; in kswapd() 3511 tsk->flags &= ~(PF_MEMALLOC | PF_SWAPWRITE | PF_KSWAPD); in kswapd()
|
D | memory.c | 3715 static int __access_remote_vm(struct task_struct *tsk, struct mm_struct *mm, in __access_remote_vm() argument 3729 ret = get_user_pages(tsk, mm, addr, 1, in __access_remote_vm() 3797 int access_process_vm(struct task_struct *tsk, unsigned long addr, in access_process_vm() argument 3804 mm = get_task_mm(tsk); in access_process_vm() 3811 ret = __access_remote_vm(tsk, mm, addr, buf, len, flags); in access_process_vm()
|