/mm/kasan/ |
D | quarantine.c | 36 size_t bytes; member 50 q->bytes = 0; in qlist_init() 62 q->bytes += size; in qlist_put() 78 to->bytes += from->bytes; in qlist_move_all() 201 if (unlikely(q->bytes > QUARANTINE_PERCPU_SIZE)) { in kasan_quarantine_put() 205 WRITE_ONCE(quarantine_size, quarantine_size + temp.bytes); in kasan_quarantine_put() 207 if (global_quarantine[quarantine_tail].bytes >= in kasan_quarantine_put() 264 WRITE_ONCE(quarantine_size, quarantine_size - to_free.bytes); in kasan_quarantine_reduce()
|
/mm/ |
D | process_vm_access.c | 98 size_t bytes; in process_vm_rw_single_vec() local 114 bytes = pinned_pages * PAGE_SIZE - start_offset; in process_vm_rw_single_vec() 115 if (bytes > len) in process_vm_rw_single_vec() 116 bytes = len; in process_vm_rw_single_vec() 119 start_offset, bytes, iter, in process_vm_rw_single_vec() 121 len -= bytes; in process_vm_rw_single_vec()
|
D | page_poison.c | 48 static void check_poison_mem(unsigned char *mem, size_t bytes) in check_poison_mem() argument 54 start = memchr_inv(mem, PAGE_POISON, bytes); in check_poison_mem() 58 for (end = mem + bytes - 1; end > start; end--) { in check_poison_mem()
|
D | page_counter.c | 248 u64 bytes; in page_counter_memparse() local 255 bytes = memparse(buf, &end); in page_counter_memparse() 259 *nr_pages = min(bytes / PAGE_SIZE, (u64)PAGE_COUNTER_MAX); in page_counter_memparse()
|
D | util.c | 709 size_t bytes; in __vmalloc_array() local 711 if (unlikely(check_mul_overflow(n, size, &bytes))) in __vmalloc_array() 713 return __vmalloc(bytes, flags); in __vmalloc_array()
|
D | page-writeback.c | 399 unsigned long bytes = vm_dirty_bytes; in domain_dirty_limits() local 419 if (bytes) in domain_dirty_limits() 420 ratio = min(DIV_ROUND_UP(bytes, global_avail), in domain_dirty_limits() 425 bytes = bg_bytes = 0; in domain_dirty_limits() 428 if (bytes) in domain_dirty_limits() 429 thresh = DIV_ROUND_UP(bytes, PAGE_SIZE); in domain_dirty_limits()
|
D | filemap.c | 3482 unsigned long bytes; /* Bytes to write to page */ in generic_perform_write() local 3487 bytes = min_t(unsigned long, PAGE_SIZE - offset, in generic_perform_write() 3501 if (unlikely(iov_iter_fault_in_readable(i, bytes))) { in generic_perform_write() 3511 status = a_ops->write_begin(file, mapping, pos, bytes, flags, in generic_perform_write() 3519 copied = iov_iter_copy_from_user_atomic(page, i, offset, bytes); in generic_perform_write() 3522 status = a_ops->write_end(file, mapping, pos, bytes, copied, in generic_perform_write() 3540 bytes = min_t(unsigned long, PAGE_SIZE - offset, in generic_perform_write()
|
D | memory.c | 5616 int bytes, ret, offset; in __access_remote_vm() local 5638 bytes = ret; in __access_remote_vm() 5641 bytes = len; in __access_remote_vm() 5643 if (bytes > PAGE_SIZE-offset) in __access_remote_vm() 5644 bytes = PAGE_SIZE-offset; in __access_remote_vm() 5649 maddr + offset, buf, bytes); in __access_remote_vm() 5653 buf, maddr + offset, bytes); in __access_remote_vm() 5658 len -= bytes; in __access_remote_vm() 5659 buf += bytes; in __access_remote_vm() 5660 addr += bytes; in __access_remote_vm()
|
D | slab_common.c | 640 static inline unsigned int size_index_elem(unsigned int bytes) in size_index_elem() argument 642 return (bytes - 1) / 8; in size_index_elem()
|
D | swapfile.c | 2847 unsigned int bytes, inuse; in swap_show() local 2854 bytes = si->pages << (PAGE_SHIFT - 10); in swap_show() 2863 bytes, bytes < 10000000 ? "\t" : "", in swap_show()
|
D | slub.c | 814 u8 *start, unsigned int value, unsigned int bytes) in check_bytes_and_report() argument 821 fault = memchr_inv(kasan_reset_tag(start), value, bytes); in check_bytes_and_report() 826 end = start + bytes; in check_bytes_and_report()
|
D | Kconfig | 191 # PA-RISC 7xxx's spinlock_t would enlarge struct page from 32 to 44 bytes.
|