/mm/ |
D | maccess.c | 10 probe_read_common(void *dst, const void __user *src, size_t size) in probe_read_common() argument 15 ret = __copy_from_user_inatomic(dst, src, size); in probe_read_common() 22 probe_write_common(void __user *dst, const void *src, size_t size) in probe_write_common() argument 27 ret = __copy_to_user_inatomic(dst, src, size); in probe_write_common() 48 long __weak probe_kernel_read(void *dst, const void *src, size_t size) 51 long __probe_kernel_read(void *dst, const void *src, size_t size) in __probe_kernel_read() argument 57 ret = probe_read_common(dst, (__force const void __user *)src, size); in __probe_kernel_read() 74 long __weak probe_user_read(void *dst, const void __user *src, size_t size) 77 long __probe_user_read(void *dst, const void __user *src, size_t size) in __probe_user_read() argument 84 ret = probe_read_common(dst, src, size); in __probe_user_read() [all …]
|
D | zswap.c | 377 u8 *dst; in zswap_dstmem_prepare() local 379 dst = kmalloc_node(PAGE_SIZE * 2, GFP_KERNEL, cpu_to_node(cpu)); in zswap_dstmem_prepare() 380 if (!dst) in zswap_dstmem_prepare() 383 per_cpu(zswap_dstmem, cpu) = dst; in zswap_dstmem_prepare() 389 u8 *dst; in zswap_dstmem_dead() local 391 dst = per_cpu(zswap_dstmem, cpu); in zswap_dstmem_dead() 392 kfree(dst); in zswap_dstmem_dead() 849 u8 *src, *dst; in zswap_writeback_entry() local 890 dst = kmap_atomic(page); in zswap_writeback_entry() 893 dst, &dlen); in zswap_writeback_entry() [all …]
|
D | migrate.c | 548 static void __copy_gigantic_page(struct page *dst, struct page *src, in __copy_gigantic_page() argument 552 struct page *dst_base = dst; in __copy_gigantic_page() 557 copy_highpage(dst, src); in __copy_gigantic_page() 560 dst = mem_map_next(dst, dst_base, i); in __copy_gigantic_page() 565 static void copy_huge_page(struct page *dst, struct page *src) in copy_huge_page() argument 576 __copy_gigantic_page(dst, src, nr_pages); in copy_huge_page() 587 copy_highpage(dst + i, src + i); in copy_huge_page() 2144 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_hole() 2160 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_skip() 2318 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_pmd() [all …]
|
D | list_lru.c | 546 struct list_lru_one *src, *dst; in memcg_drain_list_lru_node() local 556 dst = list_lru_from_memcg_idx(nlru, dst_idx); in memcg_drain_list_lru_node() 558 list_splice_init(&src->list, &dst->list); in memcg_drain_list_lru_node() 559 set = (!dst->nr_items && src->nr_items); in memcg_drain_list_lru_node() 560 dst->nr_items += src->nr_items; in memcg_drain_list_lru_node()
|
D | rmap.c | 262 int anon_vma_clone(struct vm_area_struct *dst, struct vm_area_struct *src) in anon_vma_clone() argument 280 anon_vma_chain_link(dst, avc, anon_vma); in anon_vma_clone() 290 if (!dst->anon_vma && anon_vma != src->anon_vma && in anon_vma_clone() 292 dst->anon_vma = anon_vma; in anon_vma_clone() 294 if (dst->anon_vma) in anon_vma_clone() 295 dst->anon_vma->degree++; in anon_vma_clone() 306 dst->anon_vma = NULL; in anon_vma_clone() 307 unlink_anon_vmas(dst); in anon_vma_clone()
|
D | memory.c | 2167 static inline void cow_user_page(struct page *dst, struct page *src, unsigned long va, struct vm_ar… in cow_user_page() argument 2178 void *kaddr = kmap_atomic(dst); in cow_user_page() 2190 flush_dcache_page(dst); in cow_user_page() 2192 copy_user_highpage(dst, src, va, vma); in cow_user_page() 4579 static void copy_user_gigantic_page(struct page *dst, struct page *src, in copy_user_gigantic_page() argument 4585 struct page *dst_base = dst; in copy_user_gigantic_page() 4590 copy_user_highpage(dst, src, addr + i*PAGE_SIZE, vma); in copy_user_gigantic_page() 4593 dst = mem_map_next(dst, dst_base, i); in copy_user_gigantic_page() 4599 struct page *dst; member 4608 copy_user_highpage(copy_arg->dst + idx, copy_arg->src + idx, in copy_subpage() [all …]
|
D | hugetlb.c | 3419 int copy_hugetlb_page_range(struct mm_struct *dst, struct mm_struct *src, in copy_hugetlb_page_range() argument 3445 dst_pte = huge_pte_alloc(dst, addr, sz); in copy_hugetlb_page_range() 3464 dst_ptl = huge_pte_lock(h, dst, dst_pte); in copy_hugetlb_page_range() 3490 set_huge_swap_pte_at(dst, addr, dst_pte, entry, sz); in copy_hugetlb_page_range() 3506 set_huge_pte_at(dst, addr, dst_pte, entry); in copy_hugetlb_page_range() 3507 hugetlb_count_add(pages_per_huge_page(h), dst); in copy_hugetlb_page_range()
|
D | zsmalloc.c | 1585 static void zs_object_copy(struct size_class *class, unsigned long dst, in zs_object_copy() argument 1598 obj_to_location(dst, &d_page, &d_objidx); in zs_object_copy()
|
D | mempolicy.c | 2201 int vma_dup_policy(struct vm_area_struct *src, struct vm_area_struct *dst) in vma_dup_policy() argument 2207 dst->vm_policy = pol; in vma_dup_policy()
|
D | vmscan.c | 1684 struct lruvec *lruvec, struct list_head *dst, in isolate_lru_pages() argument 1731 list_move(&page->lru, dst); in isolate_lru_pages()
|