• Home
  • Raw
  • Download

Lines Matching refs:vma

5841 	struct perf_event *event = vmf->vma->vm_file->private_data;  in perf_mmap_fault()
5864 vmf->page->mapping = vmf->vma->vm_file->f_mapping; in perf_mmap_fault()
5979 static void perf_mmap_open(struct vm_area_struct *vma) in perf_mmap_open() argument
5981 struct perf_event *event = vma->vm_file->private_data; in perf_mmap_open()
5986 if (vma->vm_pgoff) in perf_mmap_open()
5990 event->pmu->event_mapped(event, vma->vm_mm); in perf_mmap_open()
6003 static void perf_mmap_close(struct vm_area_struct *vma) in perf_mmap_close() argument
6005 struct perf_event *event = vma->vm_file->private_data; in perf_mmap_close()
6013 event->pmu->event_unmapped(event, vma->vm_mm); in perf_mmap_close()
6020 if (rb_has_aux(rb) && vma->vm_pgoff == rb->aux_pgoff && in perf_mmap_close()
6032 atomic64_sub(rb->aux_mmap_locked, &vma->vm_mm->pinned_vm); in perf_mmap_close()
6107 atomic64_sub(mmap_locked, &vma->vm_mm->pinned_vm); in perf_mmap_close()
6121 static int perf_mmap(struct file *file, struct vm_area_struct *vma) in perf_mmap() argument
6141 if (!(vma->vm_flags & VM_SHARED)) in perf_mmap()
6148 vma_size = vma->vm_end - vma->vm_start; in perf_mmap()
6150 if (vma->vm_pgoff == 0) { in perf_mmap()
6178 if (aux_offset != vma->vm_pgoff << PAGE_SHIFT) in perf_mmap()
6182 if (rb_has_aux(rb) && rb->aux_pgoff != vma->vm_pgoff) in perf_mmap()
6273 locked = atomic64_read(&vma->vm_mm->pinned_vm) + extra; in perf_mmap()
6283 if (vma->vm_flags & VM_WRITE) in perf_mmap()
6307 ret = rb_alloc_aux(rb, event, vma->vm_pgoff, nr_pages, in perf_mmap()
6316 atomic64_add(extra, &vma->vm_mm->pinned_vm); in perf_mmap()
6329 vma->vm_flags |= VM_DONTCOPY | VM_DONTEXPAND | VM_DONTDUMP; in perf_mmap()
6330 vma->vm_ops = &perf_mmap_vmops; in perf_mmap()
6333 event->pmu->event_mapped(event, vma->vm_mm); in perf_mmap()
8052 struct vm_area_struct *vma; member
8076 struct vm_area_struct *vma = mmap_event->vma; in perf_event_mmap_match() local
8077 int executable = vma->vm_flags & VM_EXEC; in perf_event_mmap_match()
8139 struct vm_area_struct *vma = mmap_event->vma; in perf_event_mmap_event() local
8140 struct file *file = vma->vm_file; in perf_event_mmap_event()
8149 if (vma->vm_flags & VM_READ) in perf_event_mmap_event()
8151 if (vma->vm_flags & VM_WRITE) in perf_event_mmap_event()
8153 if (vma->vm_flags & VM_EXEC) in perf_event_mmap_event()
8156 if (vma->vm_flags & VM_MAYSHARE) in perf_event_mmap_event()
8161 if (vma->vm_flags & VM_DENYWRITE) in perf_event_mmap_event()
8163 if (vma->vm_flags & VM_MAYEXEC) in perf_event_mmap_event()
8165 if (vma->vm_flags & VM_LOCKED) in perf_event_mmap_event()
8167 if (is_vm_hugetlb_page(vma)) in perf_event_mmap_event()
8189 inode = file_inode(vma->vm_file); in perf_event_mmap_event()
8198 if (vma->vm_ops && vma->vm_ops->name) { in perf_event_mmap_event()
8199 name = (char *) vma->vm_ops->name(vma); in perf_event_mmap_event()
8204 name = (char *)arch_vma_name(vma); in perf_event_mmap_event()
8208 if (vma->vm_start <= vma->vm_mm->start_brk && in perf_event_mmap_event()
8209 vma->vm_end >= vma->vm_mm->brk) { in perf_event_mmap_event()
8213 if (vma->vm_start <= vma->vm_mm->start_stack && in perf_event_mmap_event()
8214 vma->vm_end >= vma->vm_mm->start_stack) { in perf_event_mmap_event()
8245 if (!(vma->vm_flags & VM_EXEC)) in perf_event_mmap_event()
8281 struct vm_area_struct *vma, in perf_addr_filter_vma_adjust() argument
8284 unsigned long vma_size = vma->vm_end - vma->vm_start; in perf_addr_filter_vma_adjust()
8285 unsigned long off = vma->vm_pgoff << PAGE_SHIFT; in perf_addr_filter_vma_adjust()
8286 struct file *file = vma->vm_file; in perf_addr_filter_vma_adjust()
8292 fr->start = vma->vm_start; in perf_addr_filter_vma_adjust()
8295 fr->start = vma->vm_start + filter->offset - off; in perf_addr_filter_vma_adjust()
8296 fr->size = min(vma->vm_end - fr->start, filter->size); in perf_addr_filter_vma_adjust()
8305 struct vm_area_struct *vma = data; in __perf_addr_filters_adjust() local
8313 if (!vma->vm_file) in __perf_addr_filters_adjust()
8318 if (perf_addr_filter_vma_adjust(filter, vma, in __perf_addr_filters_adjust()
8336 static void perf_addr_filters_adjust(struct vm_area_struct *vma) in perf_addr_filters_adjust() argument
8345 if (!(vma->vm_flags & VM_EXEC)) in perf_addr_filters_adjust()
8354 perf_iterate_ctx(ctx, __perf_addr_filters_adjust, vma, true); in perf_addr_filters_adjust()
8359 void perf_event_mmap(struct vm_area_struct *vma) in perf_event_mmap() argument
8367 .vma = vma, in perf_event_mmap()
8378 .start = vma->vm_start, in perf_event_mmap()
8379 .len = vma->vm_end - vma->vm_start, in perf_event_mmap()
8380 .pgoff = (u64)vma->vm_pgoff << PAGE_SHIFT, in perf_event_mmap()
8390 perf_addr_filters_adjust(vma); in perf_event_mmap()
10013 struct vm_area_struct *vma; in perf_addr_filter_apply() local
10015 for (vma = mm->mmap; vma; vma = vma->vm_next) { in perf_addr_filter_apply()
10016 if (!vma->vm_file) in perf_addr_filter_apply()
10019 if (perf_addr_filter_vma_adjust(filter, vma, fr)) in perf_addr_filter_apply()