Lines Matching refs:objects
519 if (object < base || object >= base + page->objects * s->size || in check_valid_pointer()
615 page, page->objects, page->inuse, page->freelist, page->flags); in print_page_info()
922 if (page->objects > maxobj) { in check_slab()
924 page->objects, maxobj); in check_slab()
927 if (page->inuse > page->objects) { in check_slab()
929 page->inuse, page->objects); in check_slab()
949 while (fp && nr <= page->objects) { in on_freelist()
960 page->inuse = page->objects; in on_freelist()
975 if (page->objects != max_objects) { in on_freelist()
977 page->objects, max_objects); in on_freelist()
978 page->objects = max_objects; in on_freelist()
981 if (page->inuse != page->objects - nr) { in on_freelist()
983 page->inuse, page->objects - nr); in on_freelist()
984 page->inuse = page->objects - nr; in on_freelist()
1043 static inline void inc_slabs_node(struct kmem_cache *s, int node, int objects) in inc_slabs_node() argument
1055 atomic_long_add(objects, &n->total_objects); in inc_slabs_node()
1058 static inline void dec_slabs_node(struct kmem_cache *s, int node, int objects) in dec_slabs_node() argument
1063 atomic_long_sub(objects, &n->total_objects); in dec_slabs_node()
1129 page->inuse = page->objects; in alloc_debug_processing()
1378 int objects) {} in inc_slabs_node() argument
1380 int objects) {} in dec_slabs_node() argument
1578 if (page->objects < 2 || !s->random_seq) in shuffle_freelist()
1584 page_limit = page->objects * s->size; in shuffle_freelist()
1593 for (idx = 1; idx < page->objects; idx++) { in shuffle_freelist()
1654 page->objects = oo_objects(oo); in allocate_slab()
1673 for (idx = 0, p = start; idx < page->objects - 1; idx++) { in allocate_slab()
1682 page->inuse = page->objects; in allocate_slab()
1691 inc_slabs_node(s, page_to_nid(page), page->objects); in allocate_slab()
1720 page->objects) in __free_slab()
1751 dec_slabs_node(s, page_to_nid(page), page->objects); in discard_slab()
1791 int mode, int *objects) in acquire_slab() argument
1807 *objects = new.objects - new.inuse; in acquire_slab()
1809 new.inuse = page->objects; in acquire_slab()
1841 int objects; in get_partial_node() local
1859 t = acquire_slab(s, n, page, object == NULL, &objects); in get_partial_node()
1863 available += objects; in get_partial_node()
2284 pobjects += page->objects - page->inuse; in put_cpu_partial()
2381 return page->objects - page->inuse; in count_free()
2509 new.inuse = page->objects; in get_freelist()
3394 inc_slabs_node(kmem_cache_node, node, page->objects); in early_kmem_cache_node_alloc()
3677 unsigned long *map = bitmap_zalloc(page->objects, GFP_ATOMIC); in list_slab_objects()
3684 for_each_object(p, s, addr, page->objects) { in list_slab_objects()
3998 int free = page->objects - page->inuse; in __kmem_cache_shrink()
4006 if (free == page->objects) { in __kmem_cache_shrink()
4381 return page->objects; in count_total()
4397 bitmap_zero(map, page->objects); in validate_slab()
4400 for_each_object(p, s, addr, page->objects) { in validate_slab()
4406 for_each_object(p, s, addr, page->objects) in validate_slab()
4603 bitmap_zero(map, page->objects); in process_slab()
4606 for_each_object(p, s, addr, page->objects) in process_slab()
4814 x = page->objects; in show_slab_objects()
4999 unsigned int objects; in cpu_partial_store() local
5002 err = kstrtouint(buf, 10, &objects); in cpu_partial_store()
5005 if (objects && !kmem_cache_has_cpu_partial(s)) in cpu_partial_store()
5008 slub_set_cpu_partial(s, objects); in cpu_partial_store()
5044 SLAB_ATTR_RO(objects);
5054 int objects = 0; in slabs_cpu_partial_show() local
5066 objects += page->pobjects; in slabs_cpu_partial_show()
5070 len = sprintf(buf, "%d(%d)", objects, pages); in slabs_cpu_partial_show()