• Home
  • Raw
  • Download

Lines Matching refs:mm

27 void task_mem(struct seq_file *m, struct mm_struct *mm)  in task_mem()  argument
32 anon = get_mm_counter(mm, MM_ANONPAGES); in task_mem()
33 file = get_mm_counter(mm, MM_FILEPAGES); in task_mem()
34 shmem = get_mm_counter(mm, MM_SHMEMPAGES); in task_mem()
43 hiwater_vm = total_vm = mm->total_vm; in task_mem()
44 if (hiwater_vm < mm->hiwater_vm) in task_mem()
45 hiwater_vm = mm->hiwater_vm; in task_mem()
47 if (hiwater_rss < mm->hiwater_rss) in task_mem()
48 hiwater_rss = mm->hiwater_rss; in task_mem()
50 text = (PAGE_ALIGN(mm->end_code) - (mm->start_code & PAGE_MASK)) >> 10; in task_mem()
51 lib = (mm->exec_vm << (PAGE_SHIFT-10)) - text; in task_mem()
52 swap = get_mm_counter(mm, MM_SWAPENTS); in task_mem()
53 ptes = PTRS_PER_PTE * sizeof(pte_t) * atomic_long_read(&mm->nr_ptes); in task_mem()
54 pmds = PTRS_PER_PMD * sizeof(pmd_t) * mm_nr_pmds(mm); in task_mem()
74 mm->locked_vm << (PAGE_SHIFT-10), in task_mem()
75 mm->pinned_vm << (PAGE_SHIFT-10), in task_mem()
81 mm->data_vm << (PAGE_SHIFT-10), in task_mem()
82 mm->stack_vm << (PAGE_SHIFT-10), text, lib, in task_mem()
86 hugetlb_report_usage(m, mm); in task_mem()
89 unsigned long task_vsize(struct mm_struct *mm) in task_vsize() argument
91 return PAGE_SIZE * mm->total_vm; in task_vsize()
94 unsigned long task_statm(struct mm_struct *mm, in task_statm() argument
98 *shared = get_mm_counter(mm, MM_FILEPAGES) + in task_statm()
99 get_mm_counter(mm, MM_SHMEMPAGES); in task_statm()
100 *text = (PAGE_ALIGN(mm->end_code) - (mm->start_code & PAGE_MASK)) in task_statm()
102 *data = mm->data_vm + mm->stack_vm; in task_statm()
103 *resident = *shared + get_mm_counter(mm, MM_ANONPAGES); in task_statm()
104 return mm->total_vm; in task_statm()
136 struct mm_struct *mm = vma->vm_mm; in seq_print_vma_name() local
157 pages_pinned = get_user_pages_remote(current, mm, in seq_print_vma_name()
185 struct mm_struct *mm = priv->mm; in vma_stop() local
188 up_read(&mm->mmap_sem); in vma_stop()
189 mmput(mm); in vma_stop()
210 struct mm_struct *mm; in m_start() local
222 mm = priv->mm; in m_start()
223 if (!mm || !mmget_not_zero(mm)) in m_start()
226 down_read(&mm->mmap_sem); in m_start()
228 priv->tail_vma = get_gate_vma(mm); in m_start()
231 vma = find_vma(mm, last_addr - 1); in m_start()
239 if (pos < mm->map_count) { in m_start()
240 for (vma = mm->mmap; pos; pos--) { in m_start()
248 if (pos == mm->map_count && priv->tail_vma) in m_start()
288 priv->mm = proc_mem_open(inode, PTRACE_MODE_READ); in proc_maps_open()
289 if (IS_ERR(priv->mm)) { in proc_maps_open()
290 int err = PTR_ERR(priv->mm); in proc_maps_open()
304 if (priv->mm) in proc_map_release()
305 mmdrop(priv->mm); in proc_map_release()
353 struct mm_struct *mm = vma->vm_mm; in show_map_vma() local
391 if (!mm) { in show_map_vma()
396 if (vma->vm_start <= mm->brk && in show_map_vma()
397 vma->vm_end >= mm->start_brk) { in show_map_vma()
814 .mm = vma->vm_mm, in show_smap()
1165 struct mm_struct *mm; in clear_refs_write() local
1187 mm = get_task_mm(task); in clear_refs_write()
1188 if (mm) { in clear_refs_write()
1195 .mm = mm, in clear_refs_write()
1200 if (down_write_killable(&mm->mmap_sem)) { in clear_refs_write()
1209 reset_mm_hiwater_rss(mm); in clear_refs_write()
1210 up_write(&mm->mmap_sem); in clear_refs_write()
1214 down_read(&mm->mmap_sem); in clear_refs_write()
1215 tlb_gather_mmu(&tlb, mm, 0, -1); in clear_refs_write()
1217 for (vma = mm->mmap; vma; vma = vma->vm_next) { in clear_refs_write()
1220 up_read(&mm->mmap_sem); in clear_refs_write()
1221 if (down_write_killable(&mm->mmap_sem)) { in clear_refs_write()
1230 if (!mmget_still_valid(mm)) { in clear_refs_write()
1240 up_write(&mm->mmap_sem); in clear_refs_write()
1243 for (vma = mm->mmap; vma; vma = vma->vm_next) { in clear_refs_write()
1247 downgrade_write(&mm->mmap_sem); in clear_refs_write()
1250 mmu_notifier_invalidate_range_start(mm, 0, -1); in clear_refs_write()
1252 walk_page_range(0, mm->highest_vm_end, &clear_refs_walk); in clear_refs_write()
1254 mmu_notifier_invalidate_range_end(mm, 0, -1); in clear_refs_write()
1256 up_read(&mm->mmap_sem); in clear_refs_write()
1258 mmput(mm); in clear_refs_write()
1316 struct vm_area_struct *vma = find_vma(walk->mm, addr); in pagemap_pte_hole()
1463 orig_pte = pte = pte_offset_map_lock(walk->mm, pmdp, addr, &ptl); in pagemap_pmd_range()
1555 struct mm_struct *mm = file->private_data; in pagemap_read() local
1564 if (!mm || !mmget_not_zero(mm)) in pagemap_read()
1590 pagemap_walk.mm = mm; in pagemap_read()
1596 end_vaddr = mm->task_size; in pagemap_read()
1599 if (svpfn > mm->task_size >> PAGE_SHIFT) in pagemap_read()
1618 down_read(&mm->mmap_sem); in pagemap_read()
1620 up_read(&mm->mmap_sem); in pagemap_read()
1639 mmput(mm); in pagemap_read()
1646 struct mm_struct *mm; in pagemap_open() local
1648 mm = proc_mem_open(inode, PTRACE_MODE_READ); in pagemap_open()
1649 if (IS_ERR(mm)) in pagemap_open()
1650 return PTR_ERR(mm); in pagemap_open()
1651 file->private_data = mm; in pagemap_open()
1657 struct mm_struct *mm = file->private_data; in pagemap_release() local
1659 if (mm) in pagemap_release()
1660 mmdrop(mm); in pagemap_release()
1791 orig_pte = pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); in gather_pte_stats()
1841 struct mm_struct *mm = vma->vm_mm; in show_numa_map() local
1846 .mm = mm, in show_numa_map()
1852 if (!mm) in show_numa_map()
1871 } else if (vma->vm_start <= mm->brk && vma->vm_end >= mm->start_brk) { in show_numa_map()