Home
last modified time | relevance | path

Searched refs:mm (Results 1 – 19 of 19) sorted by relevance

/fs/proc/
Dtask_nommu.c21 void task_mem(struct seq_file *m, struct mm_struct *mm) in task_mem() argument
28 mmap_read_lock(mm); in task_mem()
29 for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) { in task_mem()
42 if (atomic_read(&mm->mm_count) > 1 || in task_mem()
52 if (atomic_read(&mm->mm_count) > 1) in task_mem()
53 sbytes += kobjsize(mm); in task_mem()
55 bytes += kobjsize(mm); in task_mem()
80 mmap_read_unlock(mm); in task_mem()
83 unsigned long task_vsize(struct mm_struct *mm) in task_vsize() argument
89 mmap_read_lock(mm); in task_vsize()
[all …]
Dtask_mmu.c32 void task_mem(struct seq_file *m, struct mm_struct *mm) in task_mem() argument
37 anon = get_mm_counter(mm, MM_ANONPAGES); in task_mem()
38 file = get_mm_counter(mm, MM_FILEPAGES); in task_mem()
39 shmem = get_mm_counter(mm, MM_SHMEMPAGES); in task_mem()
48 hiwater_vm = total_vm = mm->total_vm; in task_mem()
49 if (hiwater_vm < mm->hiwater_vm) in task_mem()
50 hiwater_vm = mm->hiwater_vm; in task_mem()
52 if (hiwater_rss < mm->hiwater_rss) in task_mem()
53 hiwater_rss = mm->hiwater_rss; in task_mem()
56 text = PAGE_ALIGN(mm->end_code) - (mm->start_code & PAGE_MASK); in task_mem()
[all …]
Darray.c411 static inline void task_core_dumping(struct seq_file *m, struct mm_struct *mm) in task_core_dumping() argument
413 seq_put_decimal_ull(m, "CoreDumping:\t", !!mm->core_state); in task_core_dumping()
417 static inline void task_thp_status(struct seq_file *m, struct mm_struct *mm) in task_thp_status() argument
422 thp_enabled = !test_bit(MMF_DISABLE_THP, &mm->flags); in task_thp_status()
429 struct mm_struct *mm = get_task_mm(task); in proc_pid_status() local
437 if (mm) { in proc_pid_status()
438 task_mem(m, mm); in proc_pid_status()
439 task_core_dumping(m, mm); in proc_pid_status()
440 task_thp_status(m, mm); in proc_pid_status()
441 mmput(mm); in proc_pid_status()
[all …]
Dbase.c221 static ssize_t get_mm_proctitle(struct mm_struct *mm, char __user *buf, in get_mm_proctitle() argument
236 got = access_remote_vm(mm, arg_start, page, PAGE_SIZE, FOLL_ANON); in get_mm_proctitle()
258 static ssize_t get_mm_cmdline(struct mm_struct *mm, char __user *buf, in get_mm_cmdline() argument
266 if (!mm->env_end) in get_mm_cmdline()
269 spin_lock(&mm->arg_lock); in get_mm_cmdline()
270 arg_start = mm->arg_start; in get_mm_cmdline()
271 arg_end = mm->arg_end; in get_mm_cmdline()
272 env_start = mm->env_start; in get_mm_cmdline()
273 env_end = mm->env_end; in get_mm_cmdline()
274 spin_unlock(&mm->arg_lock); in get_mm_cmdline()
[all …]
Dinternal.h291 struct mm_struct *mm; member
Dinode.c460 get_area = current->mm->get_unmapped_area; in pde_get_unmapped_area()
/fs/
Duserfaultfd.c73 struct mm_struct *mm; member
186 mmdrop(ctx->mm); in userfaultfd_ctx_put()
241 struct mm_struct *mm = ctx->mm; in userfaultfd_huge_must_wait() local
245 mmap_assert_locked(mm); in userfaultfd_huge_must_wait()
247 ptep = huge_pte_offset(mm, address, vma_mmu_pagesize(vma)); in userfaultfd_huge_must_wait()
289 struct mm_struct *mm = ctx->mm; in userfaultfd_must_wait() local
297 mmap_assert_locked(mm); in userfaultfd_must_wait()
299 pgd = pgd_offset(mm, address); in userfaultfd_must_wait()
396 struct mm_struct *mm = vmf->vma->vm_mm; in handle_userfault() local
421 mmap_assert_locked(mm); in handle_userfault()
[all …]
Dbinfmt_elf_fdpic.c358 current->mm->start_code = 0; in load_elf_fdpic_binary()
359 current->mm->end_code = 0; in load_elf_fdpic_binary()
360 current->mm->start_stack = 0; in load_elf_fdpic_binary()
361 current->mm->start_data = 0; in load_elf_fdpic_binary()
362 current->mm->end_data = 0; in load_elf_fdpic_binary()
363 current->mm->context.exec_fdpic_loadmap = 0; in load_elf_fdpic_binary()
364 current->mm->context.interp_fdpic_loadmap = 0; in load_elf_fdpic_binary()
369 &current->mm->start_stack, in load_elf_fdpic_binary()
370 &current->mm->start_brk); in load_elf_fdpic_binary()
372 retval = setup_arg_pages(bprm, current->mm->start_stack, in load_elf_fdpic_binary()
[all …]
Dexec.c190 struct mm_struct *mm = current->mm; in acct_arg_size() local
193 if (!mm || !diff) in acct_arg_size()
197 add_mm_counter(mm, MM_ANONPAGES, diff); in acct_arg_size()
222 mmap_read_lock(bprm->mm); in get_arg_page()
223 ret = get_user_pages_remote(bprm->mm, pos, 1, gup_flags, in get_arg_page()
225 mmap_read_unlock(bprm->mm); in get_arg_page()
254 struct mm_struct *mm = bprm->mm; in __bprm_mm_init() local
256 bprm->vma = vma = vm_area_alloc(mm); in __bprm_mm_init()
261 if (mmap_write_lock_killable(mm)) { in __bprm_mm_init()
278 err = insert_vm_struct(mm, vma); in __bprm_mm_init()
[all …]
Dcoredump.c166 exe_file = get_mm_exe_file(current->mm); in cn_print_exe_file()
366 if (t != current && t->mm) { in zap_process()
376 static int zap_threads(struct task_struct *tsk, struct mm_struct *mm, in zap_threads() argument
385 mm->core_state = core_state; in zap_threads()
395 if (atomic_read(&mm->mm_users) == nr + 1) in zap_threads()
435 if (unlikely(!p->mm)) in zap_threads()
437 if (unlikely(p->mm == mm)) { in zap_threads()
455 struct mm_struct *mm = tsk->mm; in coredump_wait() local
462 if (mmap_write_lock_killable(mm)) in coredump_wait()
465 if (!mm->core_state) in coredump_wait()
[all …]
Dbinfmt_flat.c136 sp = (unsigned long __user *)current->mm->start_stack; in create_flat_tables()
144 current->mm->start_stack = (unsigned long)sp & -FLAT_STACK_ALIGN; in create_flat_tables()
145 sp = (unsigned long __user *)current->mm->start_stack; in create_flat_tables()
157 current->mm->arg_start = (unsigned long)p; in create_flat_tables()
168 current->mm->arg_end = (unsigned long)p; in create_flat_tables()
170 current->mm->env_start = (unsigned long) p; in create_flat_tables()
181 current->mm->env_end = (unsigned long)p; in create_flat_tables()
404 ptr = (unsigned long __user *)(current->mm->start_code + r.reloc.offset); in old_reloc()
406 ptr = (unsigned long __user *)(current->mm->start_data + r.reloc.offset); in old_reloc()
416 val += current->mm->start_code; in old_reloc()
[all …]
Dbinfmt_aout.c90 current->mm->arg_start = (unsigned long) p; in create_aout_tables()
99 current->mm->arg_end = current->mm->env_start = (unsigned long) p; in create_aout_tables()
108 current->mm->env_end = (unsigned long) p; in create_aout_tables()
166 current->mm->end_code = ex.a_text + in load_aout_binary()
167 (current->mm->start_code = N_TXTADDR(ex)); in load_aout_binary()
168 current->mm->end_data = ex.a_data + in load_aout_binary()
169 (current->mm->start_data = N_DATADDR(ex)); in load_aout_binary()
170 current->mm->brk = ex.a_bss + in load_aout_binary()
171 (current->mm->start_brk = N_BSSADDR(ex)); in load_aout_binary()
240 retval = set_brk(current->mm->start_brk, current->mm->brk); in load_aout_binary()
[all …]
Dbinfmt_elf.c125 current->mm->start_brk = current->mm->brk = end; in set_brk()
176 struct mm_struct *mm = current->mm; in create_elf_tables() local
240 elf_info = (elf_addr_t *)mm->saved_auxv; in create_elf_tables()
291 memset(elf_info, 0, (char *)mm->saved_auxv + in create_elf_tables()
292 sizeof(mm->saved_auxv) - (char *)elf_info); in create_elf_tables()
297 ei_index = elf_info - (elf_addr_t *)mm->saved_auxv; in create_elf_tables()
316 if (mmap_read_lock_killable(mm)) in create_elf_tables()
318 vma = find_extend_vma(mm, bprm->p); in create_elf_tables()
319 mmap_read_unlock(mm); in create_elf_tables()
328 p = mm->arg_end = mm->arg_start; in create_elf_tables()
[all …]
Daio.c330 struct mm_struct *mm = vma->vm_mm; in aio_ring_mremap() local
334 spin_lock(&mm->ioctx_lock); in aio_ring_mremap()
336 table = rcu_dereference(mm->ioctx_table); in aio_ring_mremap()
355 spin_unlock(&mm->ioctx_lock); in aio_ring_mremap()
467 struct mm_struct *mm = current->mm; in aio_setup_ring() local
526 if (mmap_write_lock_killable(mm)) { in aio_setup_ring()
535 mmap_write_unlock(mm); in aio_setup_ring()
644 static int ioctx_add_table(struct kioctx *ctx, struct mm_struct *mm) in ioctx_add_table() argument
650 spin_lock(&mm->ioctx_lock); in ioctx_add_table()
651 table = rcu_dereference_raw(mm->ioctx_table); in ioctx_add_table()
[all …]
DKconfig242 <file:Documentation/admin-guide/mm/hugetlbpage.rst> for details.
/fs/hugetlbfs/
Dinode.c219 info.low_limit = current->mm->mmap_base; in hugetlb_get_unmapped_area_bottomup()
236 info.high_limit = arch_get_mmap_base(addr, current->mm->mmap_base); in hugetlb_get_unmapped_area_topdown()
250 info.low_limit = current->mm->mmap_base; in hugetlb_get_unmapped_area_topdown()
262 struct mm_struct *mm = current->mm; in hugetlb_get_unmapped_area() local
280 vma = find_vma(mm, addr); in hugetlb_get_unmapped_area()
291 if (mm->get_unmapped_area == arch_get_unmapped_area_topdown) in hugetlb_get_unmapped_area()
654 struct mm_struct *mm = current->mm; in hugetlbfs_fallocate() local
692 vma_init(&pseudo_vma, mm); in hugetlbfs_fallocate()
/fs/ramfs/
Dfile-mmu.c38 return current->mm->get_unmapped_area(file, addr, len, pgoff, flags); in ramfs_mmu_get_unmapped_area()
/fs/notify/inotify/
Dinotify_user.c656 group->memcg = get_mem_cgroup_from_mm(current->mm); in inotify_new_group()
/fs/notify/fanotify/
Dfanotify_user.c1232 group->memcg = get_mem_cgroup_from_mm(current->mm); in SYSCALL_DEFINE2()