Lines Matching refs:c
582 int __kmem_cache_create(struct kmem_cache *c, slab_flags_t flags) in __kmem_cache_create() argument
586 c->size += sizeof(struct slob_rcu); in __kmem_cache_create()
588 c->flags = flags; in __kmem_cache_create()
592 static void *slob_alloc_node(struct kmem_cache *c, gfp_t flags, int node) in slob_alloc_node() argument
601 if (c->size < PAGE_SIZE) { in slob_alloc_node()
602 b = slob_alloc(c->size, flags, c->align, node, 0); in slob_alloc_node()
603 trace_kmem_cache_alloc_node(_RET_IP_, b, c->object_size, in slob_alloc_node()
604 SLOB_UNITS(c->size) * SLOB_UNIT, in slob_alloc_node()
607 b = slob_new_pages(flags, get_order(c->size), node); in slob_alloc_node()
608 trace_kmem_cache_alloc_node(_RET_IP_, b, c->object_size, in slob_alloc_node()
609 PAGE_SIZE << get_order(c->size), in slob_alloc_node()
613 if (b && c->ctor) { in slob_alloc_node()
615 c->ctor(b); in slob_alloc_node()
618 kmemleak_alloc_recursive(b, c->size, 1, c->flags, flags); in slob_alloc_node()
658 void kmem_cache_free(struct kmem_cache *c, void *b) in kmem_cache_free() argument
660 kmemleak_free_recursive(b, c->flags); in kmem_cache_free()
661 if (unlikely(c->flags & SLAB_TYPESAFE_BY_RCU)) { in kmem_cache_free()
663 slob_rcu = b + (c->size - sizeof(struct slob_rcu)); in kmem_cache_free()
664 slob_rcu->size = c->size; in kmem_cache_free()
667 __kmem_cache_free(b, c->size); in kmem_cache_free()
687 int __kmem_cache_shutdown(struct kmem_cache *c) in __kmem_cache_shutdown() argument
693 void __kmem_cache_release(struct kmem_cache *c) in __kmem_cache_release() argument