/mm/ |
D | iov_iter.c | 7 static size_t copy_to_iter_iovec(void *from, size_t bytes, struct iov_iter *i) in copy_to_iter_iovec() argument 13 if (unlikely(bytes > i->count)) in copy_to_iter_iovec() 14 bytes = i->count; in copy_to_iter_iovec() 16 if (unlikely(!bytes)) in copy_to_iter_iovec() 19 wanted = bytes; in copy_to_iter_iovec() 23 copy = min(bytes, iov->iov_len - skip); in copy_to_iter_iovec() 29 bytes -= copy; in copy_to_iter_iovec() 30 while (unlikely(!left && bytes)) { in copy_to_iter_iovec() 33 copy = min(bytes, iov->iov_len); in copy_to_iter_iovec() 38 bytes -= copy; in copy_to_iter_iovec() [all …]
|
D | filemap_xip.c | 331 size_t bytes; in __xip_file_write() local 345 bytes = PAGE_CACHE_SIZE - offset; in __xip_file_write() 346 if (bytes > count) in __xip_file_write() 347 bytes = count; in __xip_file_write() 365 copied = bytes - in __xip_file_write() 366 __copy_from_user_nocache(xip_mem + offset, buf, bytes); in __xip_file_write() 378 if (unlikely(copied != bytes)) in __xip_file_write()
|
D | process_vm_access.c | 99 size_t bytes; in process_vm_rw_single_vec() local 110 bytes = pages * PAGE_SIZE - start_offset; in process_vm_rw_single_vec() 111 if (bytes > len) in process_vm_rw_single_vec() 112 bytes = len; in process_vm_rw_single_vec() 115 start_offset, bytes, iter, in process_vm_rw_single_vec() 117 len -= bytes; in process_vm_rw_single_vec()
|
D | debug-pagealloc.c | 48 static void check_poison_mem(unsigned char *mem, size_t bytes) in check_poison_mem() argument 54 start = memchr_inv(mem, PAGE_POISON, bytes); in check_poison_mem() 58 for (end = mem + bytes - 1; end > start; end--) { in check_poison_mem()
|
D | bootmem.c | 59 unsigned long bytes = DIV_ROUND_UP(pages, 8); in bootmap_bytes() local 61 return ALIGN(bytes, sizeof(long)); in bootmap_bytes() 70 unsigned long bytes = bootmap_bytes(pages); in bootmem_bootmap_pages() local 72 return PAGE_ALIGN(bytes) >> PAGE_SHIFT; in bootmem_bootmap_pages()
|
D | filemap.c | 2468 unsigned long bytes; /* Bytes to write to page */ in generic_perform_write() local 2473 bytes = min_t(unsigned long, PAGE_CACHE_SIZE - offset, in generic_perform_write() 2487 if (unlikely(iov_iter_fault_in_readable(i, bytes))) { in generic_perform_write() 2497 status = a_ops->write_begin(file, mapping, pos, bytes, flags, in generic_perform_write() 2505 copied = iov_iter_copy_from_user_atomic(page, i, offset, bytes); in generic_perform_write() 2508 status = a_ops->write_end(file, mapping, pos, bytes, copied, in generic_perform_write() 2526 bytes = min_t(unsigned long, PAGE_CACHE_SIZE - offset, in generic_perform_write()
|
D | memory.c | 3549 int bytes, ret, offset; in __access_remote_vm() local 3571 bytes = ret; in __access_remote_vm() 3574 bytes = len; in __access_remote_vm() 3576 if (bytes > PAGE_SIZE-offset) in __access_remote_vm() 3577 bytes = PAGE_SIZE-offset; in __access_remote_vm() 3582 maddr + offset, buf, bytes); in __access_remote_vm() 3586 buf, maddr + offset, bytes); in __access_remote_vm() 3591 len -= bytes; in __access_remote_vm() 3592 buf += bytes; in __access_remote_vm() 3593 addr += bytes; in __access_remote_vm()
|
D | slab_common.c | 643 static inline int size_index_elem(size_t bytes) in size_index_elem() argument 645 return (bytes - 1) / 8; in size_index_elem()
|
D | memcontrol.c | 2341 unsigned long bytes = stock->nr_pages * PAGE_SIZE; in drain_stock() local 2343 res_counter_uncharge(&old->res, bytes); in drain_stock() 2345 res_counter_uncharge(&old->memsw, bytes); in drain_stock() 2614 unsigned long bytes = nr_pages * PAGE_SIZE; in cancel_charge() local 2619 res_counter_uncharge(&memcg->res, bytes); in cancel_charge() 2621 res_counter_uncharge(&memcg->memsw, bytes); in cancel_charge() 2631 unsigned long bytes = nr_pages * PAGE_SIZE; in __mem_cgroup_cancel_local_charge() local 2636 res_counter_uncharge_until(&memcg->res, memcg->res.parent, bytes); in __mem_cgroup_cancel_local_charge() 2639 memcg->memsw.parent, bytes); in __mem_cgroup_cancel_local_charge()
|
D | slub.c | 708 u8 *start, unsigned int value, unsigned int bytes) in check_bytes_and_report() argument 713 fault = memchr_inv(start, value, bytes); in check_bytes_and_report() 717 end = start + bytes; in check_bytes_and_report()
|
D | Kconfig | 220 # PA-RISC 7xxx's spinlock_t would enlarge struct page from 32 to 44 bytes.
|