/mm/ |
D | early_ioremap.c | 35 unsigned long size, in early_memremap_pgprot_adjust() 106 __early_ioremap(resource_size_t phys_addr, unsigned long size, pgprot_t prot) in __early_ioremap() 168 void __init early_iounmap(void __iomem *addr, unsigned long size) in early_iounmap() 217 early_ioremap(resource_size_t phys_addr, unsigned long size) in early_ioremap() 224 early_memremap(resource_size_t phys_addr, unsigned long size) in early_memremap() 233 early_memremap_ro(resource_size_t phys_addr, unsigned long size) in early_memremap_ro() 244 early_memremap_prot(resource_size_t phys_addr, unsigned long size, in early_memremap_prot() 254 void __init copy_from_early_mem(void *dest, phys_addr_t src, unsigned long size) in copy_from_early_mem() 276 early_ioremap(resource_size_t phys_addr, unsigned long size) in early_ioremap() 283 early_memremap(resource_size_t phys_addr, unsigned long size) in early_memremap() [all …]
|
D | memblock.c | 148 static inline phys_addr_t memblock_cap_size(phys_addr_t base, phys_addr_t *size) in memblock_cap_size() 163 phys_addr_t base, phys_addr_t size) in memblock_overlaps_region() 191 phys_addr_t size, phys_addr_t align, int nid, in __memblock_find_range_bottom_up() 226 phys_addr_t size, phys_addr_t align, int nid, in __memblock_find_range_top_down() 271 static phys_addr_t __init_memblock memblock_find_in_range_node(phys_addr_t size, in memblock_find_in_range_node() 336 phys_addr_t end, phys_addr_t size, in memblock_find_in_range() 380 phys_addr_t addr, size; in memblock_discard() local 550 phys_addr_t size, in memblock_insert_region() 583 phys_addr_t base, phys_addr_t size, in memblock_add_range() 680 int __init_memblock memblock_add_node(phys_addr_t base, phys_addr_t size, in memblock_add_node() [all …]
|
D | slob.c | 126 #define SLOB_UNITS(size) DIV_ROUND_UP(size, SLOB_UNIT) argument 135 int size; member 146 static void set_slob(slob_t *s, slobidx_t size, slob_t *next) in set_slob() 237 static void *slob_page_alloc(struct page *sp, size_t size, int align, in slob_page_alloc() 301 static void *slob_alloc(size_t size, gfp_t gfp, int align, int node, in slob_alloc() 382 static void slob_free(void *block, int size) in slob_free() 469 __do_kmalloc_node(size_t size, gfp_t gfp, int node, unsigned long caller) in __do_kmalloc_node() 517 void *__kmalloc(size_t size, gfp_t gfp) in __kmalloc() 523 void *__kmalloc_track_caller(size_t size, gfp_t gfp, unsigned long caller) in __kmalloc_track_caller() 529 void *__kmalloc_node_track_caller(size_t size, gfp_t gfp, in __kmalloc_node_track_caller() [all …]
|
D | vmalloc.c | 560 unsigned long size; in augment_tree_propagate_check() local 764 is_within_this_va(struct vmap_area *va, unsigned long size, in is_within_this_va() 788 find_vmap_lowest_match(unsigned long size, in find_vmap_lowest_match() 847 find_vmap_lowest_linear_match(unsigned long size, in find_vmap_lowest_linear_match() 863 find_vmap_lowest_match_check(unsigned long size) in find_vmap_lowest_match_check() 891 unsigned long nva_start_addr, unsigned long size) in classify_va_fit_type() 917 unsigned long nva_start_addr, unsigned long size, in adjust_va_to_fit_type() 1007 __alloc_vmap_area(unsigned long size, unsigned long align, in __alloc_vmap_area() 1049 static struct vmap_area *alloc_vmap_area(unsigned long size, in alloc_vmap_area() 1572 static void *vb_alloc(unsigned long size, gfp_t gfp_mask) in vb_alloc() [all …]
|
D | maccess.c | 10 probe_read_common(void *dst, const void __user *src, size_t size) in probe_read_common() 22 probe_write_common(void __user *dst, const void *src, size_t size) in probe_write_common() 51 long __probe_kernel_read(void *dst, const void *src, size_t size) in __probe_kernel_read() 77 long __probe_user_read(void *dst, const void __user *src, size_t size) in __probe_user_read() 104 long __probe_kernel_write(void *dst, const void *src, size_t size) in __probe_kernel_write() 130 long __probe_user_write(void __user *dst, const void *src, size_t size) in __probe_user_write()
|
D | slab_common.c | 86 static int kmem_cache_sanity_check(const char *name, unsigned int size) in kmem_cache_sanity_check() 98 static inline int kmem_cache_sanity_check(const char *name, unsigned int size) in kmem_cache_sanity_check() 282 unsigned int align, unsigned int size) in calculate_alignment() 332 struct kmem_cache *find_mergeable(unsigned int size, unsigned int align, in find_mergeable() 455 unsigned int size, unsigned int align, in kmem_cache_create_usercopy() 561 kmem_cache_create(const char *name, unsigned int size, unsigned int align, in kmem_cache_create() 1043 unsigned int size, slab_flags_t flags, in create_boot_cache() 1075 unsigned int size, slab_flags_t flags, in create_kmalloc_cache() 1137 struct kmem_cache *kmalloc_slab(size_t size, gfp_t flags) in kmalloc_slab() 1225 kmalloc_cache_name(const char *prefix, unsigned int size) in kmalloc_cache_name() [all …]
|
D | sparse-vmemmap.c | 41 unsigned long size, in __earlyonly_bootmem_alloc() 49 void * __meminit vmemmap_alloc_block(unsigned long size, int node) in vmemmap_alloc_block() 74 void * __meminit vmemmap_alloc_block_buf(unsigned long size, int node) in vmemmap_alloc_block_buf() 105 void * __meminit altmap_alloc_block_buf(unsigned long size, in altmap_alloc_block_buf() 157 static void * __meminit vmemmap_alloc_block_zero(unsigned long size, int node) in vmemmap_alloc_block_zero()
|
D | nommu.c | 143 void *__vmalloc(unsigned long size, gfp_t gfp_mask, pgprot_t prot) in __vmalloc() 153 void *__vmalloc_node_flags(unsigned long size, int node, gfp_t flags) in __vmalloc_node_flags() 158 void *vmalloc_user(unsigned long size) in vmalloc_user() 220 void *vmalloc(unsigned long size) in vmalloc() 238 void *vzalloc(unsigned long size) in vzalloc() 256 void *vmalloc_node(unsigned long size, int node) in vmalloc_node() 274 void *vzalloc_node(unsigned long size, int node) in vzalloc_node() 292 void *vmalloc_exec(unsigned long size) in vmalloc_exec() 304 void *vmalloc_32(unsigned long size) in vmalloc_32() 320 void *vmalloc_32_user(unsigned long size) in vmalloc_32_user() [all …]
|
D | kmemleak.c | 123 size_t size; member 147 size_t size; member 570 static struct kmemleak_object *create_object(unsigned long ptr, size_t size, in create_object() 700 static void delete_object_part(unsigned long ptr, size_t size) in delete_object_part() 785 static void add_scan_area(unsigned long ptr, size_t size, gfp_t gfp) in add_scan_area() 887 void __ref kmemleak_alloc(const void *ptr, size_t size, int min_count, in kmemleak_alloc() 906 void __ref kmemleak_alloc_percpu(const void __percpu *ptr, size_t size, in kmemleak_alloc_percpu() 933 void __ref kmemleak_vmalloc(const struct vm_struct *area, size_t size, gfp_t gfp) in kmemleak_vmalloc() 974 void __ref kmemleak_free_part(const void *ptr, size_t size) in kmemleak_free_part() 1082 void __ref kmemleak_scan_area(const void *ptr, size_t size, gfp_t gfp) in kmemleak_scan_area() [all …]
|
D | sparse.c | 264 unsigned long size, align; in memory_present() local 356 unsigned long size) in sparse_early_usemaps_alloc_pgdat_section() 428 unsigned long size) in sparse_early_usemaps_alloc_pgdat_section() 454 unsigned long size = section_map_size(); in __populate_section_memmap() local 475 static inline void __meminit sparse_buffer_free(unsigned long size) in sparse_buffer_free() 481 static void __init sparse_buffer_init(unsigned long size, int nid) in sparse_buffer_init() 494 unsigned long size = sparsemap_buf_end - sparsemap_buf; in sparse_buffer_fini() local 501 void * __meminit sparse_buffer_alloc(unsigned long size) in sparse_buffer_alloc()
|
D | dmapool.c | 45 size_t size; member 68 unsigned size; in show_pools() local 131 size_t size, size_t align, size_t boundary) in dma_pool_create() 507 size_t size, size_t align, size_t allocation) in dmam_pool_create()
|
D | slab.c | 1442 int size = cachep->object_size; in poison_obj() local 1483 int i, size; in print_objinfo() local 1508 int size, i; in check_poison_obj() local 1654 size_t size, slab_flags_t flags) in calculate_slab_order() 1725 size_t size; in alloc_kmem_cache_cpus() local 1789 __kmem_cache_alias(const char *name, unsigned int size, unsigned int align, in __kmem_cache_alias() 1808 size_t size, slab_flags_t flags) in set_objfreelist_slab_cache() 1839 size_t size, slab_flags_t flags) in set_off_slab_cache() 1873 size_t size, slab_flags_t flags) in set_on_slab_cache() 1916 unsigned int size = cachep->size; in __kmem_cache_create() local [all …]
|
D | memory_hotplug.c | 104 static struct resource *register_memory_resource(u64 start, u64 size) in register_memory_resource() 1002 static int check_hotplug_memory_range(u64 start, u64 size) in check_hotplug_memory_range() 1029 u64 start, size; in add_memory_resource() local 1103 int __ref __add_memory(int nid, u64 start, u64 size) in __add_memory() 1118 int add_memory(int nid, u64 start, u64 size) in add_memory() 1716 resource_size_t size) in __release_memory_resource() 1735 static int __ref try_remove_memory(int nid, u64 start, u64 size) in try_remove_memory() 1780 void __remove_memory(int nid, u64 start, u64 size) in __remove_memory() 1795 int remove_memory(int nid, u64 start, u64 size) in remove_memory()
|
D | mempool.c | 25 static void poison_error(mempool_t *pool, void *element, size_t size, in poison_error() 42 static void __check_element(mempool_t *pool, void *element, size_t size) in __check_element() 74 static void __poison_element(void *element, size_t size) in __poison_element() 530 size_t size = (size_t)pool_data; in mempool_kmalloc() local
|
D | percpu.c | 220 static int __pcpu_size_to_slot(int size) in __pcpu_size_to_slot() 226 static int pcpu_size_to_slot(int size) in pcpu_size_to_slot() 503 static void *pcpu_mem_zalloc(size_t size, gfp_t gfp) in pcpu_mem_zalloc() 1132 unsigned long size, in pcpu_find_zero_area() 1586 static void __percpu *pcpu_alloc(size_t size, size_t align, bool reserved, in pcpu_alloc() 1781 void __percpu *__alloc_percpu_gfp(size_t size, size_t align, gfp_t gfp) in __alloc_percpu_gfp() 1794 void __percpu *__alloc_percpu(size_t size, size_t align) in __alloc_percpu() 1816 void __percpu *__alloc_reserved_percpu(size_t size, size_t align) in __alloc_reserved_percpu() 2949 static void * __init pcpu_dfl_fc_alloc(unsigned int cpu, size_t size, in pcpu_dfl_fc_alloc() 2955 static void __init pcpu_dfl_fc_free(void *ptr, size_t size) in pcpu_dfl_fc_free()
|
D | shmem.c | 90 #define VM_ACCT(size) (PAGE_ALIGN(size) >> PAGE_SHIFT) argument 169 static inline int shmem_acct_size(unsigned long flags, loff_t size) in shmem_acct_size() 175 static inline void shmem_unacct_size(unsigned long flags, loff_t size) in shmem_unacct_size() 3229 const char *name, void *buffer, size_t size, in shmem_xattr_handler_get() 3241 size_t size, int flags) in shmem_xattr_handler_set() 3271 static ssize_t shmem_listxattr(struct dentry *dentry, char *buffer, size_t size) in shmem_listxattr() 3414 unsigned long long size; in shmem_parse_one() local 4086 #define shmem_acct_size(flags, size) 0 argument 4087 #define shmem_unacct_size(flags, size) do {} while (0) argument 4093 static struct file *__shmem_file_setup(struct vfsmount *mnt, const char *name, loff_t size, in __shmem_file_setup() [all …]
|
D | zbud.c | 160 static int zbud_zpool_malloc(void *pool, size_t size, gfp_t gfp, in zbud_zpool_malloc() 230 static int size_to_chunks(size_t size) in size_to_chunks() 354 int zbud_alloc(struct zbud_pool *pool, size_t size, gfp_t gfp, in zbud_alloc()
|
D | zsmalloc.c | 207 int size; member 404 static int zs_zpool_malloc(void *pool, size_t size, gfp_t gfp, in zs_zpool_malloc() 539 static int get_size_class_index(int size) in get_size_class_index() 1139 struct page *pages[2], int off, int size) in __zs_map_object() 1147 struct page *pages[2], int off, int size) in __zs_unmap_object() 1177 struct page *pages[2], int off, int size) in __zs_map_object() 1205 struct page *pages[2], int off, int size) in __zs_unmap_object() 1459 unsigned long zs_malloc(struct zs_pool *pool, size_t size, gfp_t gfp) in zs_malloc() 1592 int s_size, d_size, size; in zs_object_copy() local 2429 int size; in zs_create_pool() local
|
D | slab.h | 22 unsigned int size; /* The aligned/padded/added on size */ member 143 unsigned int size; member 179 __kmem_cache_alias(const char *name, unsigned int size, unsigned int align, in __kmem_cache_alias() 578 size_t size, void **p) in slab_post_alloc_hook()
|
D | slub.c | 324 static inline unsigned int order_objects(unsigned int order, unsigned int size) in order_objects() 330 unsigned int size) in oo_make() 1388 static inline void *kmalloc_large_node_hook(void *ptr, size_t size, gfp_t flags) in kmalloc_large_node_hook() 2781 void *kmem_cache_alloc_trace(struct kmem_cache *s, gfp_t gfpflags, size_t size) in kmem_cache_alloc_trace() 2806 int node, size_t size) in kmem_cache_alloc_node_trace() 3046 int build_detached_freelist(struct kmem_cache *s, size_t size, in build_detached_freelist() 3117 void kmem_cache_free_bulk(struct kmem_cache *s, size_t size, void **p) in kmem_cache_free_bulk() 3135 int kmem_cache_alloc_bulk(struct kmem_cache *s, gfp_t flags, size_t size, in kmem_cache_alloc_bulk() 3246 static inline unsigned int slab_order(unsigned int size, in slab_order() 3271 static inline int calculate_order(unsigned int size) in calculate_order() [all …]
|
/mm/kasan/ |
D | generic.c | 63 unsigned long size) in memory_is_poisoned_2_4_8() 89 size_t size) in bytes_is_nonzero() 131 size_t size) in memory_is_poisoned_n() 149 static __always_inline bool memory_is_poisoned(unsigned long addr, size_t size) in memory_is_poisoned() 170 size_t size, bool write, in check_memory_region_inline() 189 bool check_memory_region(unsigned long addr, size_t size, bool write, in check_memory_region() 217 void __asan_register_globals(struct kasan_global *globals, size_t size) in __asan_register_globals() 226 void __asan_unregister_globals(struct kasan_global *globals, size_t size) in __asan_unregister_globals() 231 #define DEFINE_ASAN_LOAD_STORE(size) \ argument 255 void __asan_loadN(unsigned long addr, size_t size) in __asan_loadN() [all …]
|
D | generic_report.c | 37 void *find_first_bad_addr(void *addr, size_t size) in find_first_bad_addr() 115 #define DEFINE_ASAN_REPORT_LOAD(size) \ argument 122 #define DEFINE_ASAN_REPORT_STORE(size) \ argument 140 void __asan_report_load_n_noabort(unsigned long addr, size_t size) in __asan_report_load_n_noabort() 146 void __asan_report_store_n_noabort(unsigned long addr, size_t size) in __asan_report_store_n_noabort()
|
D | common.c | 90 bool __kasan_check_read(const volatile void *p, unsigned int size) in __kasan_check_read() 96 bool __kasan_check_write(const volatile void *p, unsigned int size) in __kasan_check_write() 132 void kasan_poison_shadow(const void *address, size_t size, u8 value) in kasan_poison_shadow() 149 void kasan_unpoison_shadow(const void *address, size_t size) in kasan_unpoison_shadow() 175 size_t size = sp - base; in __kasan_unpoison_stack() local 207 size_t size = watermark - sp; in kasan_unpoison_stack_above_sp_to() local 255 void kasan_cache_create(struct kmem_cache *cache, unsigned int *size, in kasan_cache_create() 484 size_t size, gfp_t flags, bool keep_tag) in __kasan_kmalloc() 522 size_t size, gfp_t flags) in kasan_kmalloc() 528 void * __must_check kasan_kmalloc_large(const void *ptr, size_t size, in kasan_kmalloc_large() [all …]
|
D | tags.c | 79 bool check_memory_region(unsigned long addr, size_t size, bool write, in check_memory_region() 129 #define DEFINE_HWASAN_LOAD_STORE(size) \ argument 147 void __hwasan_loadN_noabort(unsigned long addr, unsigned long size) in __hwasan_loadN_noabort() 153 void __hwasan_storeN_noabort(unsigned long addr, unsigned long size) in __hwasan_storeN_noabort() 159 void __hwasan_tag_memory(unsigned long addr, u8 tag, unsigned long size) in __hwasan_tag_memory()
|
D | init.c | 84 static __init void *early_alloc(size_t size, int node) in early_alloc() 452 void kasan_remove_zero_shadow(void *start, unsigned long size) in kasan_remove_zero_shadow() 487 int kasan_add_zero_shadow(void *start, unsigned long size) in kasan_add_zero_shadow()
|