• Home
  • Raw
  • Download

Lines Matching refs:size

370 		return (unsigned long long *)(objp + cachep->size -  in dbg_redzone2()
373 return (unsigned long long *) (objp + cachep->size - in dbg_redzone2()
380 return (void **)(objp + cachep->size - BYTES_PER_WORD); in dbg_userword()
443 return page->s_mem + cache->size * idx; in index_to_obj()
464 .size = sizeof(struct kmem_cache),
1531 cachep->name, cachep->size, cachep->gfporder); in slab_out_of_memory()
1662 int size = cachep->object_size; in store_stackinfo() local
1666 if (size < 5 * sizeof(unsigned long)) in store_stackinfo()
1672 size -= 3 * sizeof(unsigned long); in store_stackinfo()
1681 size -= sizeof(unsigned long); in store_stackinfo()
1682 if (size <= sizeof(unsigned long)) in store_stackinfo()
1694 int size = cachep->object_size; in poison_obj() local
1697 memset(addr, val, size); in poison_obj()
1698 *(unsigned char *)(addr + size - 1) = POISON_END; in poison_obj()
1737 int i, size; in print_objinfo() local
1752 size = cachep->object_size; in print_objinfo()
1753 for (i = 0; i < size && lines; i += 16, lines--) { in print_objinfo()
1756 if (i + limit > size) in print_objinfo()
1757 limit = size - i; in print_objinfo()
1765 int size, i; in check_poison_obj() local
1769 size = cachep->object_size; in check_poison_obj()
1771 for (i = 0; i < size; i++) { in check_poison_obj()
1773 if (i == size - 1) in check_poison_obj()
1782 print_tainted(), cachep->name, realobj, size); in check_poison_obj()
1788 if (i + limit > size) in check_poison_obj()
1789 limit = size - i; in check_poison_obj()
1810 realobj, size); in check_poison_obj()
1817 realobj, size); in check_poison_obj()
1834 if (cachep->size % PAGE_SIZE == 0 && in slab_destroy_debugcheck()
1837 cachep->size / PAGE_SIZE, 1); in slab_destroy_debugcheck()
1924 size_t size, size_t align, unsigned long flags) in calculate_slab_order() argument
1934 cache_estimate(gfporder, size, align, flags, &remainder, &num); in calculate_slab_order()
1951 offslab_limit = size; in calculate_slab_order()
1991 size_t size; in alloc_kmem_cache_cpus() local
1994 size = sizeof(void *) * entries + sizeof(struct array_cache); in alloc_kmem_cache_cpus()
1995 cpu_cache = __alloc_percpu(size, sizeof(void *)); in alloc_kmem_cache_cpus()
2055 __kmem_cache_alias(const char *name, size_t size, size_t align, in __kmem_cache_alias() argument
2060 cachep = find_mergeable(size, align, flags, name, ctor); in __kmem_cache_alias()
2068 cachep->object_size = max_t(int, cachep->object_size, size); in __kmem_cache_alias()
2101 size_t size = cachep->size; in __kmem_cache_create() local
2111 if (size < 4096 || fls(size - 1) == fls(size-1 + REDZONE_ALIGN + in __kmem_cache_create()
2126 if (size & (BYTES_PER_WORD - 1)) { in __kmem_cache_create()
2127 size += (BYTES_PER_WORD - 1); in __kmem_cache_create()
2128 size &= ~(BYTES_PER_WORD - 1); in __kmem_cache_create()
2135 size += REDZONE_ALIGN - 1; in __kmem_cache_create()
2136 size &= ~(REDZONE_ALIGN - 1); in __kmem_cache_create()
2165 size += 2 * sizeof(unsigned long long); in __kmem_cache_create()
2173 size += REDZONE_ALIGN; in __kmem_cache_create()
2175 size += BYTES_PER_WORD; in __kmem_cache_create()
2185 if (!slab_early_init && size >= kmalloc_size(INDEX_NODE) && in __kmem_cache_create()
2186 size >= 256 && cachep->object_size > cache_line_size() && in __kmem_cache_create()
2187 ALIGN(size, cachep->align) < PAGE_SIZE) { in __kmem_cache_create()
2188 cachep->obj_offset += PAGE_SIZE - ALIGN(size, cachep->align); in __kmem_cache_create()
2189 size = PAGE_SIZE; in __kmem_cache_create()
2200 if ((size >= (PAGE_SIZE >> 5)) && !slab_early_init && in __kmem_cache_create()
2208 size = ALIGN(size, cachep->align); in __kmem_cache_create()
2213 if (FREELIST_BYTE_INDEX && size < SLAB_OBJ_MIN_SIZE) in __kmem_cache_create()
2214 size = ALIGN(SLAB_OBJ_MIN_SIZE, cachep->align); in __kmem_cache_create()
2216 left_over = calculate_slab_order(cachep, size, cachep->align, flags); in __kmem_cache_create()
2241 if (size % PAGE_SIZE == 0 && flags & SLAB_POISON) in __kmem_cache_create()
2256 cachep->size = size; in __kmem_cache_create()
2257 cachep->reciprocal_buffer_size = reciprocal_value(size); in __kmem_cache_create()
2512 if ((cachep->size % PAGE_SIZE) == 0 && in cache_init_objs()
2515 cachep->size / PAGE_SIZE, 0); in cache_init_objs()
2732 if ((cachep->size % PAGE_SIZE)==0 && OFF_SLAB(cachep)) { in cache_free_debugcheck()
2735 cachep->size / PAGE_SIZE, 0); in cache_free_debugcheck()
2869 if ((cachep->size % PAGE_SIZE) == 0 && OFF_SLAB(cachep)) in cache_alloc_debugcheck_after()
2871 cachep->size / PAGE_SIZE, 1); in cache_alloc_debugcheck_after()
3400 cachep->object_size, cachep->size, flags); in kmem_cache_alloc()
3408 kmem_cache_alloc_trace(struct kmem_cache *cachep, gfp_t flags, size_t size) in kmem_cache_alloc_trace() argument
3415 size, cachep->size, flags); in kmem_cache_alloc_trace()
3438 cachep->object_size, cachep->size, in kmem_cache_alloc_node()
3449 size_t size) in kmem_cache_alloc_node_trace() argument
3456 size, cachep->size, in kmem_cache_alloc_node_trace()
3464 __do_kmalloc_node(size_t size, gfp_t flags, int node, unsigned long caller) in __do_kmalloc_node() argument
3468 cachep = kmalloc_slab(size, flags); in __do_kmalloc_node()
3471 return kmem_cache_alloc_node_trace(cachep, flags, node, size); in __do_kmalloc_node()
3474 void *__kmalloc_node(size_t size, gfp_t flags, int node) in __kmalloc_node() argument
3476 return __do_kmalloc_node(size, flags, node, _RET_IP_); in __kmalloc_node()
3480 void *__kmalloc_node_track_caller(size_t size, gfp_t flags, in __kmalloc_node_track_caller() argument
3483 return __do_kmalloc_node(size, flags, node, caller); in __kmalloc_node_track_caller()
3494 static __always_inline void *__do_kmalloc(size_t size, gfp_t flags, in __do_kmalloc() argument
3500 cachep = kmalloc_slab(size, flags); in __do_kmalloc()
3506 size, cachep->size, flags); in __do_kmalloc()
3511 void *__kmalloc(size_t size, gfp_t flags) in __kmalloc() argument
3513 return __do_kmalloc(size, flags, _RET_IP_); in __kmalloc()
3517 void *__kmalloc_track_caller(size_t size, gfp_t flags, unsigned long caller) in __kmalloc_track_caller() argument
3519 return __do_kmalloc(size, flags, caller); in __kmalloc_track_caller()
3761 if (cachep->size > 131072) in enable_cpucache()
3763 else if (cachep->size > PAGE_SIZE) in enable_cpucache()
3765 else if (cachep->size > 1024) in enable_cpucache()
3767 else if (cachep->size > 256) in enable_cpucache()
3782 if (cachep->size <= PAGE_SIZE && num_possible_cpus() > 1) in enable_cpucache()
4097 for (i = 0, p = page->s_mem; i < c->num; i++, p += c->size) { in handle_slab()
4109 unsigned long offset, size; in show_symbol() local
4112 if (lookup_symbol_attrs(address, &size, &offset, modname, name) == 0) { in show_symbol()
4113 seq_printf(m, "%s+%#lx/%#lx", name, offset, size); in show_symbol()
4167 m->count = m->size; in leaks_show()