/mm/ |
D | maccess.c | 17 #define copy_from_kernel_nofault_loop(dst, src, len, type, err_label) \ argument 19 __get_kernel_nofault(dst, src, type, err_label); \ 20 dst += sizeof(type); \ 25 long copy_from_kernel_nofault(void *dst, const void *src, size_t size) in copy_from_kernel_nofault() argument 31 copy_from_kernel_nofault_loop(dst, src, size, u64, Efault); in copy_from_kernel_nofault() 32 copy_from_kernel_nofault_loop(dst, src, size, u32, Efault); in copy_from_kernel_nofault() 33 copy_from_kernel_nofault_loop(dst, src, size, u16, Efault); in copy_from_kernel_nofault() 34 copy_from_kernel_nofault_loop(dst, src, size, u8, Efault); in copy_from_kernel_nofault() 43 #define copy_to_kernel_nofault_loop(dst, src, len, type, err_label) \ argument 45 __put_kernel_nofault(dst, src, type, err_label); \ [all …]
|
D | zswap.c | 394 u8 *dst; in zswap_dstmem_prepare() local 396 dst = kmalloc_node(PAGE_SIZE * 2, GFP_KERNEL, cpu_to_node(cpu)); in zswap_dstmem_prepare() 397 if (!dst) in zswap_dstmem_prepare() 400 per_cpu(zswap_dstmem, cpu) = dst; in zswap_dstmem_prepare() 406 u8 *dst; in zswap_dstmem_dead() local 408 dst = per_cpu(zswap_dstmem, cpu); in zswap_dstmem_dead() 409 kfree(dst); in zswap_dstmem_dead() 880 u8 *src, *dst; in zswap_writeback_entry() local 921 dst = kmap_atomic(page); in zswap_writeback_entry() 924 dst, &dlen); in zswap_writeback_entry() [all …]
|
D | migrate.c | 541 static void __copy_gigantic_page(struct page *dst, struct page *src, in __copy_gigantic_page() argument 545 struct page *dst_base = dst; in __copy_gigantic_page() 550 copy_highpage(dst, src); in __copy_gigantic_page() 553 dst = mem_map_next(dst, dst_base, i); in __copy_gigantic_page() 558 static void copy_huge_page(struct page *dst, struct page *src) in copy_huge_page() argument 569 __copy_gigantic_page(dst, src, nr_pages); in copy_huge_page() 580 copy_highpage(dst + i, src + i); in copy_huge_page() 2262 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_hole() 2270 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_hole() 2286 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_skip() [all …]
|
D | list_lru.c | 536 struct list_lru_one *src, *dst; in memcg_drain_list_lru_node() local 545 dst = list_lru_from_memcg_idx(nlru, dst_idx); in memcg_drain_list_lru_node() 547 list_splice_init(&src->list, &dst->list); in memcg_drain_list_lru_node() 550 dst->nr_items += src->nr_items; in memcg_drain_list_lru_node()
|
D | rmap.c | 277 int anon_vma_clone(struct vm_area_struct *dst, struct vm_area_struct *src) in anon_vma_clone() argument 295 anon_vma_chain_link(dst, avc, anon_vma); in anon_vma_clone() 304 if (!dst->anon_vma && src->anon_vma && in anon_vma_clone() 307 dst->anon_vma = anon_vma; in anon_vma_clone() 309 if (dst->anon_vma) in anon_vma_clone() 310 dst->anon_vma->num_active_vmas++; in anon_vma_clone() 321 dst->anon_vma = NULL; in anon_vma_clone() 322 unlink_anon_vmas(dst); in anon_vma_clone()
|
D | memory.c | 2901 static inline bool cow_user_page(struct page *dst, struct page *src, in cow_user_page() argument 2913 copy_user_highpage(dst, src, addr, vma); in cow_user_page() 2923 kaddr = kmap_atomic(dst); in cow_user_page() 2991 flush_dcache_page(dst); in cow_user_page() 5851 static void copy_user_gigantic_page(struct page *dst, struct page *src, in copy_user_gigantic_page() argument 5857 struct page *dst_base = dst; in copy_user_gigantic_page() 5862 copy_user_highpage(dst, src, addr + i*PAGE_SIZE, vma); in copy_user_gigantic_page() 5865 dst = mem_map_next(dst, dst_base, i); in copy_user_gigantic_page() 5871 struct page *dst; member 5880 copy_user_highpage(copy_arg->dst + idx, copy_arg->src + idx, in copy_subpage() [all …]
|
D | hugetlb.c | 3815 int copy_hugetlb_page_range(struct mm_struct *dst, struct mm_struct *src, in copy_hugetlb_page_range() argument 3850 dst_pte = huge_pte_alloc(dst, vma, addr, sz); in copy_hugetlb_page_range() 3869 dst_ptl = huge_pte_lock(h, dst, dst_pte); in copy_hugetlb_page_range() 3895 set_huge_swap_pte_at(dst, addr, dst_pte, entry, sz); in copy_hugetlb_page_range() 3911 set_huge_pte_at(dst, addr, dst_pte, entry); in copy_hugetlb_page_range() 3912 hugetlb_count_add(pages_per_huge_page(h), dst); in copy_hugetlb_page_range()
|
D | zsmalloc.c | 1539 static void zs_object_copy(struct size_class *class, unsigned long dst, in zs_object_copy() argument 1552 obj_to_location(dst, &d_page, &d_objidx); in zs_object_copy()
|
D | mempolicy.c | 2290 int vma_dup_policy(struct vm_area_struct *src, struct vm_area_struct *dst) in vma_dup_policy() argument 2296 dst->vm_policy = pol; in vma_dup_policy()
|
D | vmscan.c | 1717 struct lruvec *lruvec, struct list_head *dst, in isolate_lru_pages() argument 1765 list_move(&page->lru, dst); in isolate_lru_pages()
|