/mm/ |
D | maccess.c | 23 long __weak probe_kernel_read(void *dst, const void *src, size_t size) 26 long __probe_kernel_read(void *dst, const void *src, size_t size) in __probe_kernel_read() argument 33 ret = __copy_from_user_inatomic(dst, in __probe_kernel_read() 51 long __weak probe_kernel_write(void *dst, const void *src, size_t size) 54 long __probe_kernel_write(void *dst, const void *src, size_t size) in __probe_kernel_write() argument 61 ret = __copy_to_user_inatomic((__force void __user *)dst, src, size); in __probe_kernel_write() 86 long strncpy_from_unsafe(char *dst, const void *unsafe_addr, long count) in strncpy_from_unsafe() argument 99 ret = __get_user(*dst++, (const char __user __force *)src++); in strncpy_from_unsafe() 100 } while (dst[-1] && ret == 0 && src - unsafe_addr < count); in strncpy_from_unsafe() 102 dst[-1] = '\0'; in strncpy_from_unsafe()
|
D | zswap.c | 371 u8 *dst; in zswap_dstmem_prepare() local 373 dst = kmalloc_node(PAGE_SIZE * 2, GFP_KERNEL, cpu_to_node(cpu)); in zswap_dstmem_prepare() 374 if (!dst) in zswap_dstmem_prepare() 377 per_cpu(zswap_dstmem, cpu) = dst; in zswap_dstmem_prepare() 383 u8 *dst; in zswap_dstmem_dead() local 385 dst = per_cpu(zswap_dstmem, cpu); in zswap_dstmem_dead() 386 kfree(dst); in zswap_dstmem_dead() 843 u8 *src, *dst; in zswap_writeback_entry() local 885 dst = kmap_atomic(page); in zswap_writeback_entry() 888 dst, &dlen); in zswap_writeback_entry() [all …]
|
D | migrate.c | 613 static void __copy_gigantic_page(struct page *dst, struct page *src, in __copy_gigantic_page() argument 617 struct page *dst_base = dst; in __copy_gigantic_page() 622 copy_highpage(dst, src); in __copy_gigantic_page() 625 dst = mem_map_next(dst, dst_base, i); in __copy_gigantic_page() 630 static void copy_huge_page(struct page *dst, struct page *src) in copy_huge_page() argument 641 __copy_gigantic_page(dst, src, nr_pages); in copy_huge_page() 652 copy_highpage(dst + i, src + i); in copy_huge_page() 2158 unsigned long *dst; member 2175 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_hole() 2191 migrate->dst[migrate->npages] = 0; in migrate_vma_collect_skip() [all …]
|
D | rmap.c | 261 int anon_vma_clone(struct vm_area_struct *dst, struct vm_area_struct *src) in anon_vma_clone() argument 279 anon_vma_chain_link(dst, avc, anon_vma); in anon_vma_clone() 289 if (!dst->anon_vma && anon_vma != src->anon_vma && in anon_vma_clone() 291 dst->anon_vma = anon_vma; in anon_vma_clone() 293 if (dst->anon_vma) in anon_vma_clone() 294 dst->anon_vma->degree++; in anon_vma_clone() 305 dst->anon_vma = NULL; in anon_vma_clone() 306 unlink_anon_vmas(dst); in anon_vma_clone()
|
D | list_lru.c | 480 struct list_lru_one *src, *dst; in memcg_drain_list_lru_node() local 489 dst = list_lru_from_memcg_idx(nlru, dst_idx); in memcg_drain_list_lru_node() 491 list_splice_init(&src->list, &dst->list); in memcg_drain_list_lru_node() 492 dst->nr_items += src->nr_items; in memcg_drain_list_lru_node()
|
D | memory.c | 2327 static inline void cow_user_page(struct page *dst, struct page *src, unsigned long va, struct vm_ar… in cow_user_page() argument 2338 void *kaddr = kmap_atomic(dst); in cow_user_page() 2350 flush_dcache_page(dst); in cow_user_page() 2352 copy_user_highpage(dst, src, va, vma); in cow_user_page() 4656 static void copy_user_gigantic_page(struct page *dst, struct page *src, in copy_user_gigantic_page() argument 4662 struct page *dst_base = dst; in copy_user_gigantic_page() 4667 copy_user_highpage(dst, src, addr + i*PAGE_SIZE, vma); in copy_user_gigantic_page() 4670 dst = mem_map_next(dst, dst_base, i); in copy_user_gigantic_page() 4675 void copy_user_huge_page(struct page *dst, struct page *src, in copy_user_huge_page() argument 4682 copy_user_gigantic_page(dst, src, addr, vma, in copy_user_huge_page() [all …]
|
D | hugetlb.c | 3221 int copy_hugetlb_page_range(struct mm_struct *dst, struct mm_struct *src, in copy_hugetlb_page_range() argument 3246 dst_pte = huge_pte_alloc(dst, addr, sz); in copy_hugetlb_page_range() 3265 dst_ptl = huge_pte_lock(h, dst, dst_pte); in copy_hugetlb_page_range() 3291 set_huge_swap_pte_at(dst, addr, dst_pte, entry, sz); in copy_hugetlb_page_range() 3302 set_huge_pte_at(dst, addr, dst_pte, entry); in copy_hugetlb_page_range() 3303 hugetlb_count_add(pages_per_huge_page(h), dst); in copy_hugetlb_page_range()
|
D | zsmalloc.c | 1632 static void zs_object_copy(struct size_class *class, unsigned long dst, in zs_object_copy() argument 1645 obj_to_location(dst, &d_page, &d_objidx); in zs_object_copy()
|
D | mempolicy.c | 2122 int vma_dup_policy(struct vm_area_struct *src, struct vm_area_struct *dst) in vma_dup_policy() argument 2128 dst->vm_policy = pol; in vma_dup_policy()
|
D | vmscan.c | 1530 struct lruvec *lruvec, struct list_head *dst, in isolate_lru_pages() argument 1571 list_move(&page->lru, dst); in isolate_lru_pages()
|