Home
last modified time | relevance | path

Searched refs:dst (Results 1 – 10 of 10) sorted by relevance

/mm/
Dmaccess.c10 probe_read_common(void *dst, const void __user *src, size_t size) in probe_read_common() argument
15 ret = __copy_from_user_inatomic(dst, src, size); in probe_read_common()
22 probe_write_common(void __user *dst, const void *src, size_t size) in probe_write_common() argument
27 ret = __copy_to_user_inatomic(dst, src, size); in probe_write_common()
48 long __weak probe_kernel_read(void *dst, const void *src, size_t size)
51 long __probe_kernel_read(void *dst, const void *src, size_t size) in __probe_kernel_read() argument
57 ret = probe_read_common(dst, (__force const void __user *)src, size); in __probe_kernel_read()
74 long __weak probe_user_read(void *dst, const void __user *src, size_t size)
77 long __probe_user_read(void *dst, const void __user *src, size_t size) in __probe_user_read() argument
84 ret = probe_read_common(dst, src, size); in __probe_user_read()
[all …]
Dzswap.c377 u8 *dst; in zswap_dstmem_prepare() local
379 dst = kmalloc_node(PAGE_SIZE * 2, GFP_KERNEL, cpu_to_node(cpu)); in zswap_dstmem_prepare()
380 if (!dst) in zswap_dstmem_prepare()
383 per_cpu(zswap_dstmem, cpu) = dst; in zswap_dstmem_prepare()
389 u8 *dst; in zswap_dstmem_dead() local
391 dst = per_cpu(zswap_dstmem, cpu); in zswap_dstmem_dead()
392 kfree(dst); in zswap_dstmem_dead()
849 u8 *src, *dst; in zswap_writeback_entry() local
890 dst = kmap_atomic(page); in zswap_writeback_entry()
893 dst, &dlen); in zswap_writeback_entry()
[all …]
Dmigrate.c548 static void __copy_gigantic_page(struct page *dst, struct page *src, in __copy_gigantic_page() argument
552 struct page *dst_base = dst; in __copy_gigantic_page()
557 copy_highpage(dst, src); in __copy_gigantic_page()
560 dst = mem_map_next(dst, dst_base, i); in __copy_gigantic_page()
565 static void copy_huge_page(struct page *dst, struct page *src) in copy_huge_page() argument
576 __copy_gigantic_page(dst, src, nr_pages); in copy_huge_page()
587 copy_highpage(dst + i, src + i); in copy_huge_page()
2144 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_hole()
2160 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_skip()
2318 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_pmd()
[all …]
Dlist_lru.c546 struct list_lru_one *src, *dst; in memcg_drain_list_lru_node() local
556 dst = list_lru_from_memcg_idx(nlru, dst_idx); in memcg_drain_list_lru_node()
558 list_splice_init(&src->list, &dst->list); in memcg_drain_list_lru_node()
559 set = (!dst->nr_items && src->nr_items); in memcg_drain_list_lru_node()
560 dst->nr_items += src->nr_items; in memcg_drain_list_lru_node()
Drmap.c262 int anon_vma_clone(struct vm_area_struct *dst, struct vm_area_struct *src) in anon_vma_clone() argument
280 anon_vma_chain_link(dst, avc, anon_vma); in anon_vma_clone()
290 if (!dst->anon_vma && anon_vma != src->anon_vma && in anon_vma_clone()
292 dst->anon_vma = anon_vma; in anon_vma_clone()
294 if (dst->anon_vma) in anon_vma_clone()
295 dst->anon_vma->degree++; in anon_vma_clone()
306 dst->anon_vma = NULL; in anon_vma_clone()
307 unlink_anon_vmas(dst); in anon_vma_clone()
Dmemory.c2167 static inline void cow_user_page(struct page *dst, struct page *src, unsigned long va, struct vm_ar… in cow_user_page() argument
2178 void *kaddr = kmap_atomic(dst); in cow_user_page()
2190 flush_dcache_page(dst); in cow_user_page()
2192 copy_user_highpage(dst, src, va, vma); in cow_user_page()
4579 static void copy_user_gigantic_page(struct page *dst, struct page *src, in copy_user_gigantic_page() argument
4585 struct page *dst_base = dst; in copy_user_gigantic_page()
4590 copy_user_highpage(dst, src, addr + i*PAGE_SIZE, vma); in copy_user_gigantic_page()
4593 dst = mem_map_next(dst, dst_base, i); in copy_user_gigantic_page()
4599 struct page *dst; member
4608 copy_user_highpage(copy_arg->dst + idx, copy_arg->src + idx, in copy_subpage()
[all …]
Dhugetlb.c3419 int copy_hugetlb_page_range(struct mm_struct *dst, struct mm_struct *src, in copy_hugetlb_page_range() argument
3445 dst_pte = huge_pte_alloc(dst, addr, sz); in copy_hugetlb_page_range()
3464 dst_ptl = huge_pte_lock(h, dst, dst_pte); in copy_hugetlb_page_range()
3490 set_huge_swap_pte_at(dst, addr, dst_pte, entry, sz); in copy_hugetlb_page_range()
3506 set_huge_pte_at(dst, addr, dst_pte, entry); in copy_hugetlb_page_range()
3507 hugetlb_count_add(pages_per_huge_page(h), dst); in copy_hugetlb_page_range()
Dzsmalloc.c1585 static void zs_object_copy(struct size_class *class, unsigned long dst, in zs_object_copy() argument
1598 obj_to_location(dst, &d_page, &d_objidx); in zs_object_copy()
Dmempolicy.c2201 int vma_dup_policy(struct vm_area_struct *src, struct vm_area_struct *dst) in vma_dup_policy() argument
2207 dst->vm_policy = pol; in vma_dup_policy()
Dvmscan.c1684 struct lruvec *lruvec, struct list_head *dst, in isolate_lru_pages() argument
1731 list_move(&page->lru, dst); in isolate_lru_pages()