/mm/ |
D | maccess.c | 10 probe_read_common(void *dst, const void __user *src, size_t size) in probe_read_common() argument 15 ret = __copy_from_user_inatomic(dst, src, size); in probe_read_common() 22 probe_write_common(void __user *dst, const void *src, size_t size) in probe_write_common() argument 27 ret = __copy_to_user_inatomic(dst, src, size); in probe_write_common() 48 long __weak probe_kernel_read(void *dst, const void *src, size_t size) 51 long __probe_kernel_read(void *dst, const void *src, size_t size) in __probe_kernel_read() argument 57 ret = probe_read_common(dst, (__force const void __user *)src, size); in __probe_kernel_read() 74 long __weak probe_user_read(void *dst, const void __user *src, size_t size) 77 long __probe_user_read(void *dst, const void __user *src, size_t size) in __probe_user_read() argument 83 if (access_ok(src, size)) in __probe_user_read() [all …]
|
D | migrate.c | 553 static void __copy_gigantic_page(struct page *dst, struct page *src, in __copy_gigantic_page() argument 558 struct page *src_base = src; in __copy_gigantic_page() 562 copy_highpage(dst, src); in __copy_gigantic_page() 566 src = mem_map_next(src, src_base, i); in __copy_gigantic_page() 570 static void copy_huge_page(struct page *dst, struct page *src) in copy_huge_page() argument 575 if (PageHuge(src)) { in copy_huge_page() 577 struct hstate *h = page_hstate(src); in copy_huge_page() 581 __copy_gigantic_page(dst, src, nr_pages); in copy_huge_page() 586 BUG_ON(!PageTransHuge(src)); in copy_huge_page() 587 nr_pages = hpage_nr_pages(src); in copy_huge_page() [all …]
|
D | zswap.c | 849 u8 *src, *dst; in zswap_writeback_entry() local 889 src = (u8 *)zhdr + sizeof(struct zswap_header); in zswap_writeback_entry() 892 ret = crypto_comp_decompress(tfm, src, entry->length, in zswap_writeback_entry() 997 u8 *src, *dst; in zswap_frontswap_store() local 1040 src = kmap_atomic(page); in zswap_frontswap_store() 1041 if (zswap_is_page_same_filled(src, &value)) { in zswap_frontswap_store() 1042 kunmap_atomic(src); in zswap_frontswap_store() 1049 kunmap_atomic(src); in zswap_frontswap_store() 1062 src = kmap_atomic(page); in zswap_frontswap_store() 1063 ret = crypto_comp_compress(tfm, src, PAGE_SIZE, dst, &dlen); in zswap_frontswap_store() [all …]
|
D | util.c | 123 void *kmemdup(const void *src, size_t len, gfp_t gfp) in kmemdup() argument 129 memcpy(p, src, len); in kmemdup() 168 void *memdup_user(const void __user *src, size_t len) in memdup_user() argument 176 if (copy_from_user(p, src, len)) { in memdup_user() 194 void *vmemdup_user(const void __user *src, size_t len) in vmemdup_user() argument 202 if (copy_from_user(p, src, len)) { in vmemdup_user() 250 void *memdup_user_nul(const void __user *src, size_t len) in memdup_user_nul() argument 263 if (copy_from_user(p, src, len)) { in memdup_user_nul()
|
D | early_ioremap.c | 254 void __init copy_from_early_mem(void *dest, phys_addr_t src, unsigned long size) in copy_from_early_mem() argument 260 slop = offset_in_page(src); in copy_from_early_mem() 264 p = early_memremap(src & PAGE_MASK, clen + slop); in copy_from_early_mem() 268 src += clen; in copy_from_early_mem()
|
D | list_lru.c | 546 struct list_lru_one *src, *dst; in memcg_drain_list_lru_node() local 554 src = list_lru_from_memcg_idx(nlru, src_idx); in memcg_drain_list_lru_node() 557 list_splice_init(&src->list, &dst->list); in memcg_drain_list_lru_node() 559 if (src->nr_items) { in memcg_drain_list_lru_node() 560 dst->nr_items += src->nr_items; in memcg_drain_list_lru_node() 562 src->nr_items = 0; in memcg_drain_list_lru_node()
|
D | memory.c | 2203 static inline bool cow_user_page(struct page *dst, struct page *src, in cow_user_page() argument 2214 debug_dma_assert_idle(src); in cow_user_page() 2216 if (likely(src)) { in cow_user_page() 2217 copy_user_highpage(dst, src, addr, vma); in cow_user_page() 4720 static void copy_user_gigantic_page(struct page *dst, struct page *src, in copy_user_gigantic_page() argument 4727 struct page *src_base = src; in copy_user_gigantic_page() 4731 copy_user_highpage(dst, src, addr + i*PAGE_SIZE, vma); in copy_user_gigantic_page() 4735 src = mem_map_next(src, src_base, i); in copy_user_gigantic_page() 4741 struct page *src; member 4749 copy_user_highpage(copy_arg->dst + idx, copy_arg->src + idx, in copy_subpage() [all …]
|
D | vmscan.c | 1695 struct list_head *src = &lruvec->lists[lru]; in isolate_lru_pages() local 1706 while (scan < nr_to_scan && !list_empty(src)) { in isolate_lru_pages() 1709 page = lru_to_page(src); in isolate_lru_pages() 1710 prefetchw_prev_lru_page(page, src, flags); in isolate_lru_pages() 1743 list_move(&page->lru, src); in isolate_lru_pages() 1761 list_splice(&pages_skipped, src); in isolate_lru_pages()
|
D | rmap.c | 263 int anon_vma_clone(struct vm_area_struct *dst, struct vm_area_struct *src) in anon_vma_clone() argument 268 list_for_each_entry_reverse(pavc, &src->anon_vma_chain, same_vma) { in anon_vma_clone() 290 if (!dst->anon_vma && src->anon_vma && in anon_vma_clone()
|
D | hugetlb.c | 3479 int copy_hugetlb_page_range(struct mm_struct *dst, struct mm_struct *src, in copy_hugetlb_page_range() argument 3494 mmu_notifier_range_init(&range, MMU_NOTIFY_CLEAR, 0, vma, src, in copy_hugetlb_page_range() 3502 src_pte = huge_pte_offset(src, addr, sz); in copy_hugetlb_page_range() 3525 src_ptl = huge_pte_lockptr(h, src, src_pte); in copy_hugetlb_page_range() 3547 set_huge_swap_pte_at(src, addr, src_pte, in copy_hugetlb_page_range() 3560 huge_ptep_set_wrprotect(src, addr, src_pte); in copy_hugetlb_page_range()
|
D | zsmalloc.c | 1540 unsigned long src) in zs_object_copy() argument 1551 obj_to_location(src, &s_page, &s_objidx); in zs_object_copy()
|
D | mempolicy.c | 2197 int vma_dup_policy(struct vm_area_struct *src, struct vm_area_struct *dst) in vma_dup_policy() argument 2199 struct mempolicy *pol = mpol_dup(vma_policy(src)); in vma_dup_policy()
|
/mm/kasan/ |
D | common.c | 111 void *memmove(void *dest, const void *src, size_t len) in memmove() argument 113 check_memory_region((unsigned long)src, len, false, _RET_IP_); in memmove() 116 return __memmove(dest, src, len); in memmove() 120 void *memcpy(void *dest, const void *src, size_t len) in memcpy() argument 122 check_memory_region((unsigned long)src, len, false, _RET_IP_); in memcpy() 125 return __memcpy(dest, src, len); in memcpy()
|