• Home
  • Raw
  • Download

Lines Matching refs:objects

520 	if (object < base || object >= base + page->objects * s->size ||  in check_valid_pointer()
633 page, page->objects, page->inuse, page->freelist, page->flags); in print_page_info()
947 if (page->objects > maxobj) { in check_slab()
949 page->objects, maxobj); in check_slab()
952 if (page->inuse > page->objects) { in check_slab()
954 page->inuse, page->objects); in check_slab()
974 while (fp && nr <= page->objects) { in on_freelist()
985 page->inuse = page->objects; in on_freelist()
1000 if (page->objects != max_objects) { in on_freelist()
1002 page->objects, max_objects); in on_freelist()
1003 page->objects = max_objects; in on_freelist()
1006 if (page->inuse != page->objects - nr) { in on_freelist()
1008 page->inuse, page->objects - nr); in on_freelist()
1009 page->inuse = page->objects - nr; in on_freelist()
1068 static inline void inc_slabs_node(struct kmem_cache *s, int node, int objects) in inc_slabs_node() argument
1080 atomic_long_add(objects, &n->total_objects); in inc_slabs_node()
1083 static inline void dec_slabs_node(struct kmem_cache *s, int node, int objects) in dec_slabs_node() argument
1088 atomic_long_sub(objects, &n->total_objects); in dec_slabs_node()
1154 page->inuse = page->objects; in alloc_debug_processing()
1403 int objects) {} in inc_slabs_node() argument
1405 int objects) {} in dec_slabs_node() argument
1615 if (page->objects < 2 || !s->random_seq) in shuffle_freelist()
1621 page_limit = page->objects * s->size; in shuffle_freelist()
1630 for (idx = 1; idx < page->objects; idx++) { in shuffle_freelist()
1691 page->objects = oo_objects(oo); in allocate_slab()
1710 for (idx = 0, p = start; idx < page->objects - 1; idx++) { in allocate_slab()
1719 page->inuse = page->objects; in allocate_slab()
1728 inc_slabs_node(s, page_to_nid(page), page->objects); in allocate_slab()
1757 page->objects) in __free_slab()
1788 dec_slabs_node(s, page_to_nid(page), page->objects); in discard_slab()
1828 int mode, int *objects) in acquire_slab() argument
1844 *objects = new.objects - new.inuse; in acquire_slab()
1846 new.inuse = page->objects; in acquire_slab()
1878 int objects; in get_partial_node() local
1896 t = acquire_slab(s, n, page, object == NULL, &objects); in get_partial_node()
1900 available += objects; in get_partial_node()
2328 pobjects += page->objects - page->inuse; in put_cpu_partial()
2423 return page->objects - page->inuse; in count_free()
2551 new.inuse = page->objects; in get_freelist()
3458 inc_slabs_node(kmem_cache_node, node, page->objects); in early_kmem_cache_node_alloc()
3748 unsigned long *map = bitmap_zalloc(page->objects, GFP_ATOMIC); in list_slab_objects()
3755 for_each_object(p, s, addr, page->objects) { in list_slab_objects()
4069 int free = page->objects - page->inuse; in __kmem_cache_shrink()
4077 if (free == page->objects) { in __kmem_cache_shrink()
4454 return page->objects; in count_total()
4470 bitmap_zero(map, page->objects); in validate_slab()
4473 for_each_object(p, s, addr, page->objects) { in validate_slab()
4479 for_each_object(p, s, addr, page->objects) in validate_slab()
4676 bitmap_zero(map, page->objects); in process_slab()
4679 for_each_object(p, s, addr, page->objects) in process_slab()
4887 x = page->objects; in show_slab_objects()
5072 unsigned int objects; in cpu_partial_store() local
5075 err = kstrtouint(buf, 10, &objects); in cpu_partial_store()
5078 if (objects && !kmem_cache_has_cpu_partial(s)) in cpu_partial_store()
5081 slub_set_cpu_partial(s, objects); in cpu_partial_store()
5117 SLAB_ATTR_RO(objects);
5127 int objects = 0; in slabs_cpu_partial_show() local
5139 objects += page->pobjects; in slabs_cpu_partial_show()
5143 len = sprintf(buf, "%d(%d)", objects, pages); in slabs_cpu_partial_show()