/mm/ |
D | memblock.c | 68 static inline phys_addr_t memblock_cap_size(phys_addr_t base, phys_addr_t *size) in memblock_cap_size() 83 phys_addr_t base, phys_addr_t size) in memblock_overlaps_region() 110 phys_addr_t size, phys_addr_t align, int nid, in __memblock_find_range_bottom_up() 144 phys_addr_t size, phys_addr_t align, int nid, in __memblock_find_range_top_down() 188 phys_addr_t __init_memblock memblock_find_in_range_node(phys_addr_t size, in memblock_find_in_range_node() 249 phys_addr_t end, phys_addr_t size, in memblock_find_in_range() 293 phys_addr_t addr, size; in memblock_discard() local 473 phys_addr_t size, in memblock_insert_region() 505 phys_addr_t base, phys_addr_t size, in memblock_add_range() 590 int __init_memblock memblock_add_node(phys_addr_t base, phys_addr_t size, in memblock_add_node() [all …]
|
D | nobootmem.c | 40 static void * __init __alloc_memory_core_early(int nid, u64 size, u64 align, in __alloc_memory_core_early() 84 void __init free_bootmem_late(unsigned long addr, unsigned long size) in free_bootmem_late() 204 unsigned long size) in free_bootmem_node() 218 void __init free_bootmem(unsigned long addr, unsigned long size) in free_bootmem() 223 static void * __init ___alloc_bootmem_nopanic(unsigned long size, in ___alloc_bootmem_nopanic() 261 void * __init __alloc_bootmem_nopanic(unsigned long size, unsigned long align, in __alloc_bootmem_nopanic() 269 static void * __init ___alloc_bootmem(unsigned long size, unsigned long align, in ___alloc_bootmem() 297 void * __init __alloc_bootmem(unsigned long size, unsigned long align, in __alloc_bootmem() 306 unsigned long size, in ___alloc_bootmem_node_nopanic() 332 void * __init __alloc_bootmem_node_nopanic(pg_data_t *pgdat, unsigned long size, in __alloc_bootmem_node_nopanic() [all …]
|
D | early_ioremap.c | 35 unsigned long size, in early_memremap_pgprot_adjust() 106 __early_ioremap(resource_size_t phys_addr, unsigned long size, pgprot_t prot) in __early_ioremap() 168 void __init early_iounmap(void __iomem *addr, unsigned long size) in early_iounmap() 217 early_ioremap(resource_size_t phys_addr, unsigned long size) in early_ioremap() 224 early_memremap(resource_size_t phys_addr, unsigned long size) in early_memremap() 233 early_memremap_ro(resource_size_t phys_addr, unsigned long size) in early_memremap_ro() 244 early_memremap_prot(resource_size_t phys_addr, unsigned long size, in early_memremap_prot() 254 void __init copy_from_early_mem(void *dest, phys_addr_t src, unsigned long size) in copy_from_early_mem() 276 early_ioremap(resource_size_t phys_addr, unsigned long size) in early_ioremap() 283 early_memremap(resource_size_t phys_addr, unsigned long size) in early_memremap() [all …]
|
D | bootmem.c | 155 void __init free_bootmem_late(unsigned long physaddr, unsigned long size) in free_bootmem_late() 399 unsigned long size) in free_bootmem_node() 420 void __init free_bootmem(unsigned long physaddr, unsigned long size) in free_bootmem() 444 unsigned long size, int flags) in reserve_bootmem_node() 464 int __init reserve_bootmem(unsigned long addr, unsigned long size, in reserve_bootmem() 499 unsigned long size, unsigned long align, in alloc_bootmem_bdata() 605 static void * __init alloc_bootmem_core(unsigned long size, in alloc_bootmem_core() 630 static void * __init ___alloc_bootmem_nopanic(unsigned long size, in ___alloc_bootmem_nopanic() 662 void * __init __alloc_bootmem_nopanic(unsigned long size, unsigned long align, in __alloc_bootmem_nopanic() 670 static void * __init ___alloc_bootmem(unsigned long size, unsigned long align, in ___alloc_bootmem() [all …]
|
D | slob.c | 126 #define SLOB_UNITS(size) DIV_ROUND_UP(size, SLOB_UNIT) argument 135 int size; member 146 static void set_slob(slob_t *s, slobidx_t size, slob_t *next) in set_slob() 218 static void *slob_page_alloc(struct page *sp, size_t size, int align) in slob_page_alloc() 269 static void *slob_alloc(size_t size, gfp_t gfp, int align, int node) in slob_alloc() 341 static void slob_free(void *block, int size) in slob_free() 428 __do_kmalloc_node(size_t size, gfp_t gfp, int node, unsigned long caller) in __do_kmalloc_node() 467 void *__kmalloc(size_t size, gfp_t gfp) in __kmalloc() 473 void *__kmalloc_track_caller(size_t size, gfp_t gfp, unsigned long caller) in __kmalloc_track_caller() 479 void *__kmalloc_node_track_caller(size_t size, gfp_t gfp, in __kmalloc_node_track_caller() [all …]
|
D | vmalloc.c | 402 static struct vmap_area *alloc_vmap_area(unsigned long size, in alloc_vmap_area() 982 static void *vb_alloc(unsigned long size, gfp_t gfp_mask) in vb_alloc() 1035 static void vb_free(const void *addr, unsigned long size) in vb_free() 1138 unsigned long size = (unsigned long)count << PAGE_SHIFT; in vm_unmap_ram() local 1179 unsigned long size = (unsigned long)count << PAGE_SHIFT; in vm_map_ram() local 1310 int map_kernel_range_noflush(unsigned long addr, unsigned long size, in map_kernel_range_noflush() 1330 void unmap_kernel_range_noflush(unsigned long addr, unsigned long size) in unmap_kernel_range_noflush() 1344 void unmap_kernel_range(unsigned long addr, unsigned long size) in unmap_kernel_range() 1390 static struct vm_struct *__get_vm_area_node(unsigned long size, in __get_vm_area_node() 1424 struct vm_struct *__get_vm_area(unsigned long size, unsigned long flags, in __get_vm_area() [all …]
|
D | slab_common.c | 76 static int kmem_cache_sanity_check(const char *name, size_t size) in kmem_cache_sanity_check() 107 static inline int kmem_cache_sanity_check(const char *name, size_t size) in kmem_cache_sanity_check() 293 struct kmem_cache *find_mergeable(size_t size, size_t align, in find_mergeable() 345 unsigned long align, unsigned long size) in calculate_alignment() 368 size_t object_size, size_t size, size_t align, in create_cache() 433 kmem_cache_create(const char *name, size_t size, size_t align, in kmem_cache_create() 885 void __init create_boot_cache(struct kmem_cache *s, const char *name, size_t size, in create_boot_cache() 905 struct kmem_cache *__init create_kmalloc_cache(const char *name, size_t size, in create_kmalloc_cache() 970 struct kmem_cache *kmalloc_slab(size_t size, gfp_t flags) in kmalloc_slab() 1102 int size = kmalloc_size(i); in create_kmalloc_caches() local [all …]
|
D | sparse-vmemmap.c | 41 unsigned long size, in __earlyonly_bootmem_alloc() 52 void * __meminit vmemmap_alloc_block(unsigned long size, int node) in vmemmap_alloc_block() 70 static void * __meminit alloc_block_buf(unsigned long size, int node) in alloc_block_buf() 125 static void * __meminit altmap_alloc_block_buf(unsigned long size, in altmap_alloc_block_buf() 150 void * __meminit __vmemmap_alloc_block_buf(unsigned long size, int node, in __vmemmap_alloc_block_buf() 285 unsigned long size = sizeof(struct page) * PAGES_PER_SECTION; in sparse_mem_maps_populate_node() local
|
D | kmemleak.c | 136 size_t size; member 160 size_t size; member 270 size_t size; /* memory block size */ member 548 static struct kmemleak_object *create_object(unsigned long ptr, size_t size, in create_object() 676 static void delete_object_part(unsigned long ptr, size_t size) in delete_object_part() 763 static void add_scan_area(unsigned long ptr, size_t size, gfp_t gfp) in add_scan_area() 853 static void __init log_early(int op_type, const void *ptr, size_t size, in log_early() 943 void __ref kmemleak_alloc(const void *ptr, size_t size, int min_count, in kmemleak_alloc() 964 void __ref kmemleak_alloc_percpu(const void __percpu *ptr, size_t size, in kmemleak_alloc_percpu() 993 void __ref kmemleak_vmalloc(const struct vm_struct *area, size_t size, gfp_t gfp) in kmemleak_vmalloc() [all …]
|
D | nommu.c | 230 void *__vmalloc(unsigned long size, gfp_t gfp_mask, pgprot_t prot) in __vmalloc() 240 void *__vmalloc_node_flags(unsigned long size, int node, gfp_t flags) in __vmalloc_node_flags() 245 void *vmalloc_user(unsigned long size) in vmalloc_user() 307 void *vmalloc(unsigned long size) in vmalloc() 325 void *vzalloc(unsigned long size) in vzalloc() 343 void *vmalloc_node(unsigned long size, int node) in vmalloc_node() 361 void *vzalloc_node(unsigned long size, int node) in vzalloc_node() 383 void *vmalloc_exec(unsigned long size) in vmalloc_exec() 395 void *vmalloc_32(unsigned long size) in vmalloc_32() 411 void *vmalloc_32_user(unsigned long size) in vmalloc_32_user() [all …]
|
D | sparse.c | 212 unsigned long size, align; in memory_present() local 310 unsigned long size) in sparse_early_usemaps_alloc_pgdat_section() 383 unsigned long size) in sparse_early_usemaps_alloc_pgdat_section() 401 int size = usemap_size(); in sparse_early_usemaps_alloc_node() local 423 unsigned long size; in sparse_mem_map_populate() local 442 unsigned long size = sizeof(struct page) * PAGES_PER_SECTION; in sparse_mem_maps_populate_node() local 573 int size; in sparse_init() local
|
D | slab.c | 1494 int size = cachep->object_size; in store_stackinfo() local 1543 int size = cachep->object_size; in poison_obj() local 1584 int i, size; in print_objinfo() local 1609 int size, i; in check_poison_obj() local 1753 size_t size, unsigned long flags) in calculate_slab_order() 1824 size_t size; in alloc_kmem_cache_cpus() local 1888 __kmem_cache_alias(const char *name, size_t size, size_t align, in __kmem_cache_alias() 1907 size_t size, unsigned long flags) in set_objfreelist_slab_cache() 1930 size_t size, unsigned long flags) in set_off_slab_cache() 1964 size_t size, unsigned long flags) in set_on_slab_cache() [all …]
|
D | dmapool.c | 48 size_t size; member 71 unsigned size; in show_pools() local 132 size_t size, size_t align, size_t boundary) in dma_pool_create() 503 size_t size, size_t align, size_t allocation) in dmam_pool_create()
|
D | mempool.c | 25 static void poison_error(mempool_t *pool, void *element, size_t size, in poison_error() 42 static void __check_element(mempool_t *pool, void *element, size_t size) in __check_element() 74 static void __poison_element(void *element, size_t size) in __poison_element() 467 size_t size = (size_t)pool_data; in mempool_kmalloc() local
|
D | maccess.c | 26 long __probe_kernel_read(void *dst, const void *src, size_t size) in __probe_kernel_read() 54 long __probe_kernel_write(void *dst, const void *src, size_t size) in __probe_kernel_write()
|
D | percpu.c | 211 static int __pcpu_size_to_slot(int size) in __pcpu_size_to_slot() 217 static int pcpu_size_to_slot(int size) in pcpu_size_to_slot() 464 static void *pcpu_mem_zalloc(size_t size, gfp_t gfp) in pcpu_mem_zalloc() 1346 static void __percpu *pcpu_alloc(size_t size, size_t align, bool reserved, in pcpu_alloc() 1527 void __percpu *__alloc_percpu_gfp(size_t size, size_t align, gfp_t gfp) in __alloc_percpu_gfp() 1540 void __percpu *__alloc_percpu(size_t size, size_t align) in __alloc_percpu() 1562 void __percpu *__alloc_reserved_percpu(size_t size, size_t align) in __alloc_reserved_percpu() 2671 static void * __init pcpu_dfl_fc_alloc(unsigned int cpu, size_t size, in pcpu_dfl_fc_alloc() 2678 static void __init pcpu_dfl_fc_free(void *ptr, size_t size) in pcpu_dfl_fc_free()
|
D | memory_hotplug.c | 101 static struct resource *register_memory_resource(u64 start, u64 size) in register_memory_resource() 552 resource_size_t start, size; in __remove_pages() local 1054 static int check_hotplug_memory_range(u64 start, u64 size) in check_hotplug_memory_range() 1084 u64 start, size; in add_memory_resource() local 1176 int __ref __add_memory(int nid, u64 start, u64 size) in __add_memory() 1191 int add_memory(int nid, u64 start, u64 size) in add_memory() 1894 void __ref remove_memory(int nid, u64 start, u64 size) in remove_memory()
|
D | slab.h | 22 unsigned int size; /* The aligned/padded/added on size */ member 78 unsigned long size; member 115 __kmem_cache_alias(const char *name, size_t size, size_t align, in __kmem_cache_alias() 432 size_t size, void **p) in slab_post_alloc_hook()
|
D | zbud.c | 159 static int zbud_zpool_malloc(void *pool, size_t size, gfp_t gfp, in zbud_zpool_malloc() 229 static int size_to_chunks(size_t size) in size_to_chunks() 353 int zbud_alloc(struct zbud_pool *pool, size_t size, gfp_t gfp, in zbud_alloc()
|
D | shmem.c | 87 #define VM_ACCT(size) (PAGE_ALIGN(size) >> PAGE_SHIFT) argument 146 static inline int shmem_acct_size(unsigned long flags, loff_t size) in shmem_acct_size() 152 static inline void shmem_unacct_size(unsigned long flags, loff_t size) in shmem_unacct_size() 3400 const char *name, void *buffer, size_t size) in shmem_xattr_handler_get() 3411 size_t size, int flags) in shmem_xattr_handler_set() 3441 static ssize_t shmem_listxattr(struct dentry *dentry, char *buffer, size_t size) in shmem_listxattr() 3570 unsigned long long size; in shmem_parse_options() local 4243 #define shmem_acct_size(flags, size) 0 argument 4244 #define shmem_unacct_size(flags, size) do {} while (0) argument 4254 static struct file *__shmem_file_setup(const char *name, loff_t size, in __shmem_file_setup() [all …]
|
D | slub.c | 317 static inline int order_objects(int order, unsigned long size, int reserved) in order_objects() 323 unsigned long size, int reserved) in oo_make() 1349 static inline void kmalloc_large_node_hook(void *ptr, size_t size, gfp_t flags) in kmalloc_large_node_hook() 2754 void *kmem_cache_alloc_trace(struct kmem_cache *s, gfp_t gfpflags, size_t size) in kmem_cache_alloc_trace() 2779 int node, size_t size) in kmem_cache_alloc_node_trace() 3024 int build_detached_freelist(struct kmem_cache *s, size_t size, in build_detached_freelist() 3095 void kmem_cache_free_bulk(struct kmem_cache *s, size_t size, void **p) in kmem_cache_free_bulk() 3113 int kmem_cache_alloc_bulk(struct kmem_cache *s, gfp_t flags, size_t size, in kmem_cache_alloc_bulk() 3230 static inline int slab_order(int size, int min_objects, in slab_order() 3254 static inline int calculate_order(int size, int reserved) in calculate_order() [all …]
|
D | zsmalloc.c | 203 int size; member 404 static int zs_zpool_malloc(void *pool, size_t size, gfp_t gfp, in zs_zpool_malloc() 549 static int get_size_class_index(int size) in get_size_class_index() 1187 struct page *pages[2], int off, int size) in __zs_map_object() 1195 struct page *pages[2], int off, int size) in __zs_unmap_object() 1225 struct page *pages[2], int off, int size) in __zs_map_object() 1253 struct page *pages[2], int off, int size) in __zs_unmap_object() 1506 unsigned long zs_malloc(struct zs_pool *pool, size_t size, gfp_t gfp) in zs_malloc() 1639 int s_size, d_size, size; in zs_object_copy() local 2471 int size; in zs_create_pool() local
|
D | cma.c | 174 int __init cma_init_reserved_mem(phys_addr_t base, phys_addr_t size, in cma_init_reserved_mem() 243 phys_addr_t size, phys_addr_t limit, in cma_declare_contiguous()
|
/mm/kasan/ |
D | kasan.c | 57 static void kasan_poison_shadow(const void *address, size_t size, u8 value) in kasan_poison_shadow() 67 void kasan_unpoison_shadow(const void *address, size_t size) in kasan_unpoison_shadow() 80 size_t size = sp - base; in __kasan_unpoison_stack() local 112 size_t size = watermark - sp; in kasan_unpoison_stack_above_sp_to() local 138 unsigned long size) in memory_is_poisoned_2_4_8() 164 size_t size) in bytes_is_nonzero() 206 size_t size) in memory_is_poisoned_n() 224 static __always_inline bool memory_is_poisoned(unsigned long addr, size_t size) in memory_is_poisoned() 245 size_t size, bool write, in check_memory_region_inline() 264 size_t size, bool write, in check_memory_region() [all …]
|
D | report.c | 42 static const void *find_first_bad_addr(const void *addr, size_t size) in find_first_bad_addr() 393 void kasan_report(unsigned long addr, size_t size, in kasan_report() 413 #define DEFINE_ASAN_REPORT_LOAD(size) \ argument 420 #define DEFINE_ASAN_REPORT_STORE(size) \ argument 438 void __asan_report_load_n_noabort(unsigned long addr, size_t size) in __asan_report_load_n_noabort() 444 void __asan_report_store_n_noabort(unsigned long addr, size_t size) in __asan_report_store_n_noabort()
|