Searched refs:copied (Results 1 – 6 of 6) sorted by relevance
/mm/ |
D | userfaultfd.c | 299 long copied; in __mcopy_atomic_hugetlb() local 319 copied = 0; in __mcopy_atomic_hugetlb() 418 copied += vma_hpagesize; in __mcopy_atomic_hugetlb() 432 BUG_ON(copied < 0); in __mcopy_atomic_hugetlb() 434 BUG_ON(!copied && !err); in __mcopy_atomic_hugetlb() 435 return copied ? copied : err; in __mcopy_atomic_hugetlb() 504 long copied; in __mcopy_atomic() local 520 copied = 0; in __mcopy_atomic() 634 if (copied && mmap_lock_is_contended(dst_mm)) in __mcopy_atomic() 657 copied += PAGE_SIZE; in __mcopy_atomic() [all …]
|
D | process_vm_access.c | 37 size_t copied; in process_vm_rw_pages() local 43 copied = copy_page_from_iter(page, offset, copy, iter); in process_vm_rw_pages() 45 copied = copy_page_to_iter(page, offset, copy, iter); in process_vm_rw_pages() 47 len -= copied; in process_vm_rw_pages() 48 if (copied < copy && iov_iter_count(iter)) in process_vm_rw_pages()
|
D | filemap.c | 2685 size_t copied; in filemap_read() local 2703 copied = copy_page_to_iter(page, offset, bytes, iter); in filemap_read() 2705 already_read += copied; in filemap_read() 2706 iocb->ki_pos += copied; in filemap_read() 2709 if (copied < bytes) { in filemap_read() 3676 loff_t pos, unsigned len, unsigned copied, in pagecache_write_end() argument 3681 return aops->write_end(file, mapping, pos, len, copied, page, fsdata); in pagecache_write_end() 3823 size_t copied; /* Bytes copied from user */ in generic_perform_write() local 3855 copied = copy_page_from_iter_atomic(page, offset, bytes, i); in generic_perform_write() 3858 status = a_ops->write_end(file, mapping, pos, bytes, copied, in generic_perform_write() [all …]
|
D | shmem.c | 2495 loff_t pos, unsigned len, unsigned copied, in shmem_write_end() argument 2500 if (pos + copied > inode->i_size) in shmem_write_end() 2501 i_size_write(inode, pos + copied); in shmem_write_end() 2515 if (copied < PAGE_SIZE) { in shmem_write_end() 2518 from + copied, PAGE_SIZE); in shmem_write_end() 2526 return copied; in shmem_write_end()
|
D | vmalloc.c | 3336 int copied = 0; in aligned_vread() local 3363 copied += length; in aligned_vread() 3366 return copied; in aligned_vread()
|
D | Kconfig | 445 the page is copied into the kernel and a disk access is avoided.
|