Lines Matching refs:p
54 struct vfree_deferred *p = container_of(w, struct vfree_deferred, wq); in free_work() local
57 llist_for_each_safe(llnode, t, llist_del_all(&p->list)) in free_work()
1820 struct vm_struct *tmp, **p; in vm_area_add_early() local
1823 for (p = &vmlist; (tmp = *p) != NULL; p = &tmp->next) { in vm_area_add_early()
1830 vm->next = *p; in vm_area_add_early()
1831 *p = vm; in vm_area_add_early()
1913 struct vfree_deferred *p; in vmalloc_init() local
1918 p = &per_cpu(vfree_deferred, i); in vmalloc_init()
1919 init_llist_head(&p->list); in vmalloc_init()
1920 INIT_WORK(&p->wq, free_work); in vmalloc_init()
2276 struct vfree_deferred *p = raw_cpu_ptr(&vfree_deferred); in __vfree_deferred() local
2278 if (llist_add((struct llist_node *)addr, &p->list)) in __vfree_deferred()
2279 schedule_work(&p->wq); in __vfree_deferred()
2747 struct page *p; in aligned_vread() local
2757 p = vmalloc_to_page(addr); in aligned_vread()
2765 if (p) { in aligned_vread()
2770 void *map = kmap_atomic(p); in aligned_vread()
2786 struct page *p; in aligned_vwrite() local
2796 p = vmalloc_to_page(addr); in aligned_vwrite()
2804 if (p) { in aligned_vwrite()
2809 void *map = kmap_atomic(p); in aligned_vwrite()
3066 pte_t ***p = data; in f() local
3068 if (p) { in f()
3069 *(*p) = pte; in f()
3070 (*p)++; in f()
3426 static void *s_next(struct seq_file *m, void *p, loff_t *pos) in s_next() argument
3428 return seq_list_next(p, &vmap_area_list, pos); in s_next()
3431 static void s_stop(struct seq_file *m, void *p) in s_stop() argument
3477 static int s_show(struct seq_file *m, void *p) in s_show() argument
3482 va = list_entry(p, struct vmap_area, list); in s_show()