/mm/ |
D | oom_kill.c | 309 static int oom_evaluate_task(struct task_struct *task, void *arg) in oom_evaluate_task() argument 314 if (oom_unkillable_task(task)) in oom_evaluate_task() 318 if (!is_memcg_oom(oc) && !oom_cpuset_eligible(task, oc)) in oom_evaluate_task() 327 if (!is_sysrq_oom(oc) && tsk_is_oom_victim(task)) { in oom_evaluate_task() 328 if (test_bit(MMF_OOM_SKIP, &task->signal->oom_mm->flags)) in oom_evaluate_task() 337 if (oom_task_origin(task)) { in oom_evaluate_task() 342 points = oom_badness(task, oc->totalpages); in oom_evaluate_task() 349 get_task_struct(task); in oom_evaluate_task() 350 oc->chosen = task; in oom_evaluate_task() 383 struct task_struct *task; in dump_task() local [all …]
|
D | process_vm_access.c | 79 struct task_struct *task, in process_vm_rw_single_vec() argument 109 pages = get_user_pages_remote(task, mm, pa, pages, flags, in process_vm_rw_single_vec() 156 struct task_struct *task; in process_vm_rw_core() local 197 task = find_get_task_by_vpid(pid); in process_vm_rw_core() 198 if (!task) { in process_vm_rw_core() 203 mm = mm_access(task, PTRACE_MODE_ATTACH_REALCREDS); in process_vm_rw_core() 218 iter, process_pages, mm, task, vm_write); in process_vm_rw_core() 232 put_task_struct(task); in process_vm_rw_core()
|
D | util.c | 425 struct task_struct *task, bool bypass_rlim) in __account_locked_vm() argument 435 limit = task_rlimit(task, RLIMIT_MEMLOCK) >> PAGE_SHIFT; in __account_locked_vm() 446 pr_debug("%s: [%d] caller %ps %c%lu %lu/%lu%s\n", __func__, task->pid, in __account_locked_vm() 448 locked_vm << PAGE_SHIFT, task_rlimit(task, RLIMIT_MEMLOCK), in __account_locked_vm() 843 int get_cmdline(struct task_struct *task, char *buffer, int buflen) in get_cmdline() argument 847 struct mm_struct *mm = get_task_mm(task); in get_cmdline() 866 res = access_process_vm(task, arg_start, buffer, len, FOLL_FORCE); in get_cmdline() 880 res += access_process_vm(task, env_start, in get_cmdline()
|
D | mempolicy.c | 1465 struct task_struct *task; in kernel_migrate_pages() local 1488 task = pid ? find_task_by_vpid(pid) : current; in kernel_migrate_pages() 1489 if (!task) { in kernel_migrate_pages() 1494 get_task_struct(task); in kernel_migrate_pages() 1502 if (!ptrace_may_access(task, PTRACE_MODE_READ_REALCREDS)) { in kernel_migrate_pages() 1509 task_nodes = cpuset_mems_allowed(task); in kernel_migrate_pages() 1521 err = security_task_movememory(task); in kernel_migrate_pages() 1525 mm = get_task_mm(task); in kernel_migrate_pages() 1526 put_task_struct(task); in kernel_migrate_pages() 1543 put_task_struct(task); in kernel_migrate_pages() [all …]
|
D | migrate.c | 1770 struct task_struct *task; in kernel_move_pages() local 1784 task = pid ? find_task_by_vpid(pid) : current; in kernel_move_pages() 1785 if (!task) { in kernel_move_pages() 1789 get_task_struct(task); in kernel_move_pages() 1795 if (!ptrace_may_access(task, PTRACE_MODE_READ_REALCREDS)) { in kernel_move_pages() 1802 err = security_task_movememory(task); in kernel_move_pages() 1806 task_nodes = cpuset_mems_allowed(task); in kernel_move_pages() 1807 mm = get_task_mm(task); in kernel_move_pages() 1808 put_task_struct(task); in kernel_move_pages() 1823 put_task_struct(task); in kernel_move_pages()
|
D | hugetlb_cgroup.c | 43 struct hugetlb_cgroup *hugetlb_cgroup_from_task(struct task_struct *task) in hugetlb_cgroup_from_task() argument 45 return hugetlb_cgroup_from_css(task_css(task, hugetlb_cgrp_id)); in hugetlb_cgroup_from_task()
|
D | memory.c | 180 struct task_struct *task = current; in add_mm_counter_fast() local 182 if (likely(task->mm == mm)) in add_mm_counter_fast() 183 task->rss_stat.count[member] += val; in add_mm_counter_fast() 192 static void check_sync_rss_stat(struct task_struct *task) in check_sync_rss_stat() argument 194 if (unlikely(task != current)) in check_sync_rss_stat() 196 if (unlikely(task->rss_stat.events++ > TASK_RSS_EVENTS_THRESH)) in check_sync_rss_stat() 197 sync_mm_rss(task->mm); in check_sync_rss_stat() 204 static void check_sync_rss_stat(struct task_struct *task) in check_sync_rss_stat() argument
|
D | vmscan.c | 174 static void set_task_reclaim_state(struct task_struct *task, in set_task_reclaim_state() argument 178 WARN_ON_ONCE(rs && task->reclaim_state); in set_task_reclaim_state() 181 WARN_ON_ONCE(!rs && !task->reclaim_state); in set_task_reclaim_state() 183 task->reclaim_state = rs; in set_task_reclaim_state()
|
D | memcontrol.c | 1211 struct task_struct *task; in mem_cgroup_scan_tasks() local 1214 while (!ret && (task = css_task_iter_next(&it))) in mem_cgroup_scan_tasks() 1215 ret = fn(task, arg); in mem_cgroup_scan_tasks()
|
/mm/kasan/ |
D | common.c | 172 static void __kasan_unpoison_stack(struct task_struct *task, const void *sp) in __kasan_unpoison_stack() argument 174 void *base = task_stack_page(task); in __kasan_unpoison_stack() 181 void kasan_unpoison_task_stack(struct task_struct *task) in kasan_unpoison_task_stack() argument 183 __kasan_unpoison_stack(task, task_stack_page(task) + THREAD_SIZE); in kasan_unpoison_task_stack()
|