Searched refs:object_size (Results 1 – 8 of 8) sorted by relevance
/mm/kasan/ |
D | common.c | 127 static inline unsigned int optimal_redzone(unsigned int object_size) in optimal_redzone() argument 130 object_size <= 64 - 16 ? 16 : in optimal_redzone() 131 object_size <= 128 - 32 ? 32 : in optimal_redzone() 132 object_size <= 512 - 64 ? 64 : in optimal_redzone() 133 object_size <= 4096 - 128 ? 128 : in optimal_redzone() 134 object_size <= (1 << 14) - 256 ? 256 : in optimal_redzone() 135 object_size <= (1 << 15) - 512 ? 512 : in optimal_redzone() 136 object_size <= (1 << 16) - 1024 ? 1024 : 2048; in optimal_redzone() 192 cache->object_size < sizeof(struct kasan_free_meta)) { in __kasan_cache_create() 206 optimal_size = cache->object_size + optimal_redzone(cache->object_size); in __kasan_cache_create() [all …]
|
D | report.c | 164 object, cache->name, cache->object_size); in describe_object_addr() 172 } else if (access_addr >= object_addr + cache->object_size) { in describe_object_addr() 174 rel_bytes = access_addr - (object_addr + cache->object_size); in describe_object_addr() 182 rel_bytes, rel_type, cache->object_size, (void *)object_addr, in describe_object_addr() 183 (void *)(object_addr + cache->object_size)); in describe_object_addr()
|
/mm/ |
D | slab.h | 21 unsigned int object_size;/* The original size of the object */ member 130 slab_flags_t kmem_cache_flags(unsigned int object_size, 138 static inline slab_flags_t kmem_cache_flags(unsigned int object_size, in kmem_cache_flags() argument 490 return s->object_size; in slab_ksize() 499 return s->object_size; in slab_ksize() 502 return s->object_size; in slab_ksize() 552 memset(p[i], 0, s->object_size); in slab_post_alloc_hook() 553 kmemleak_alloc_recursive(p[i], s->object_size, 1, in slab_post_alloc_hook()
|
D | slub.c | 858 min_t(unsigned int, s->object_size, PAGE_SIZE)); in print_trailer() 860 print_section(KERN_ERR, "Redzone ", p + s->object_size, in print_trailer() 861 s->inuse - s->object_size); in print_trailer() 915 memset(p, POISON_FREE, s->object_size - 1); in init_object() 916 p[s->object_size - 1] = POISON_END; in init_object() 920 memset(p + s->object_size, val, s->inuse - s->object_size); in init_object() 1059 u8 *endobject = object + s->object_size; in check_object() 1067 endobject, val, s->inuse - s->object_size)) in check_object() 1070 if ((s->flags & SLAB_POISON) && s->object_size < s->inuse) { in check_object() 1073 s->inuse - s->object_size); in check_object() [all …]
|
D | slab.c | 1450 int size = cachep->object_size; in poison_obj() 1503 size = cachep->object_size; in print_objinfo() 1523 size = cachep->object_size; in check_poison_obj() 1793 slab_flags_t kmem_cache_flags(unsigned int object_size, in kmem_cache_flags() argument 1813 cachep->object_size = max_t(int, cachep->object_size, size); in __kmem_cache_alias() 1841 if (cachep->num * sizeof(freelist_idx_t) > cachep->object_size) in set_objfreelist_slab_cache() 2022 size >= 256 && cachep->object_size > cache_line_size()) { in __kmem_cache_create() 3443 memset(objp, 0, cachep->object_size); in __cache_free() 3450 __kcsan_check_access(objp, cachep->object_size, in __cache_free() 3507 void *ret = slab_alloc(cachep, flags, cachep->object_size, _RET_IP_); in kmem_cache_alloc() [all …]
|
D | slab_common.c | 94 return s->object_size; in kmem_cache_size() 242 unsigned int object_size, unsigned int align, in create_cache() argument 250 if (WARN_ON(useroffset + usersize > object_size)) in create_cache() 259 s->size = s->object_size = object_size; in create_cache() 658 s->size = s->object_size = size; in create_boot_cache()
|
D | failslab.c | 33 return should_fail(&failslab.attr, s->object_size); in __should_failslab()
|
D | slob.c | 617 trace_kmem_cache_alloc_node(_RET_IP_, b, c->object_size, in slob_alloc_node() 622 trace_kmem_cache_alloc_node(_RET_IP_, b, c->object_size, in slob_alloc_node()
|