• Home
  • Raw
  • Download

Lines Matching refs:size

91 static int kmem_cache_sanity_check(const char *name, unsigned int size)  in kmem_cache_sanity_check()  argument
93 if (!name || in_interrupt() || size > KMALLOC_MAX_SIZE) { in kmem_cache_sanity_check()
102 static inline int kmem_cache_sanity_check(const char *name, unsigned int size) in kmem_cache_sanity_check() argument
140 unsigned int align, unsigned int size) in calculate_alignment() argument
153 while (size <= ralign / 2) in calculate_alignment()
186 struct kmem_cache *find_mergeable(unsigned int size, unsigned int align, in find_mergeable() argument
197 size = ALIGN(size, sizeof(void *)); in find_mergeable()
198 align = calculate_alignment(flags, align, size); in find_mergeable()
199 size = ALIGN(size, align); in find_mergeable()
200 flags = kmem_cache_flags(size, flags, name); in find_mergeable()
209 if (size > s->size) in find_mergeable()
218 if ((s->size & ~(align - 1)) != s->size) in find_mergeable()
221 if (s->size - size >= sizeof(void *)) in find_mergeable()
251 s->size = s->object_size = object_size; in create_cache()
303 unsigned int size, unsigned int align, in kmem_cache_create_usercopy() argument
327 err = kmem_cache_sanity_check(name, size); in kmem_cache_create_usercopy()
348 WARN_ON(size < usersize || size - usersize < useroffset)) in kmem_cache_create_usercopy()
352 s = __kmem_cache_alias(name, size, align, flags, ctor); in kmem_cache_create_usercopy()
362 s = create_cache(cache_name, size, in kmem_cache_create_usercopy()
363 calculate_alignment(flags, align, size), in kmem_cache_create_usercopy()
417 kmem_cache_create(const char *name, unsigned int size, unsigned int align, in kmem_cache_create() argument
420 return kmem_cache_create_usercopy(name, size, align, flags, 0, 0, in kmem_cache_create()
557 unsigned int size, slab_flags_t flags, in create_boot_cache() argument
564 s->size = s->object_size = size; in create_boot_cache()
570 if (is_power_of_2(size)) in create_boot_cache()
571 align = max(align, size); in create_boot_cache()
572 s->align = calculate_alignment(flags, align, size); in create_boot_cache()
581 name, size, err); in create_boot_cache()
587 unsigned int size, slab_flags_t flags, in create_kmalloc_cache() argument
595 create_boot_cache(s, name, size, flags, useroffset, usersize); in create_kmalloc_cache()
649 struct kmem_cache *kmalloc_slab(size_t size, gfp_t flags) in kmalloc_slab() argument
654 if (size <= 192) { in kmalloc_slab()
655 if (!size) in kmalloc_slab()
658 index = size_index[size_index_elem(size)]; in kmalloc_slab()
660 if (WARN_ON_ONCE(size > KMALLOC_MAX_CACHE_SIZE)) in kmalloc_slab()
662 index = fls(size - 1); in kmalloc_slab()
678 .size = __size, \
685 .size = __size, \
779 kmalloc_info[idx].size, flags, 0, in new_kmalloc_cache()
780 kmalloc_info[idx].size); in new_kmalloc_cache()
822 kmalloc_info[i].size, in create_kmalloc_caches()
824 kmalloc_info[i].size); in create_kmalloc_caches()
848 void *kmalloc_order(size_t size, gfp_t flags, unsigned int order) in kmalloc_order() argument
863 ret = kasan_kmalloc_large(ret, size, flags); in kmalloc_order()
865 kmemleak_alloc(ret, size, 1, flags); in kmalloc_order()
871 void *kmalloc_order_trace(size_t size, gfp_t flags, unsigned int order) in kmalloc_order_trace() argument
873 void *ret = kmalloc_order(size, flags, order); in kmalloc_order_trace()
874 trace_kmalloc(_RET_IP_, ret, size, PAGE_SIZE << order, flags); in kmalloc_order_trace()
979 s->name, sinfo.active_objs, sinfo.num_objs, s->size, in cache_show()
1028 (sinfo.active_objs * s->size) / 1024, in dump_unreclaimable_slab()
1029 (sinfo.num_objs * s->size) / 1024); in dump_unreclaimable_slab()
1188 size_t size; in ksize() local
1208 size = kfence_ksize(objp) ?: __ksize(objp); in ksize()
1213 kasan_unpoison_range(objp, size); in ksize()
1214 return size; in ksize()