/lib/ |
D | list_debug.c | 22 void __list_add(struct list_head *new, in __list_add() argument 34 WARN(new == prev || new == next, in __list_add() 36 new, prev, next); in __list_add() 37 next->prev = new; in __list_add() 38 new->next = next; in __list_add() 39 new->prev = prev; in __list_add() 40 prev->next = new; in __list_add() 86 void __list_add_rcu(struct list_head *new, in __list_add_rcu() argument 95 new->next = next; in __list_add_rcu() 96 new->prev = prev; in __list_add_rcu() [all …]
|
D | lockref.c | 23 struct lockref new = old, prev = old; \ 27 new.lock_count); \ 51 new.count++; in lockref_get() 72 new.count++; in lockref_get_not_zero() 99 new.count++; in lockref_get_or_lock() 125 new.count--; in lockref_put_return() 129 return new.count; in lockref_put_return() 143 new.count--; in lockref_put_or_lock() 180 new.count++; in lockref_get_not_dead()
|
D | idr.c | 97 struct idr_layer *new; in idr_layer_alloc() local 110 new = kmem_cache_zalloc(idr_layer_cache, gfp_mask | __GFP_NOWARN); in idr_layer_alloc() 111 if (new) in idr_layer_alloc() 112 return new; in idr_layer_alloc() 120 new = __this_cpu_read(idr_preload_head); in idr_layer_alloc() 121 if (new) { in idr_layer_alloc() 122 __this_cpu_write(idr_preload_head, new->ary[0]); in idr_layer_alloc() 124 new->ary[0] = NULL; in idr_layer_alloc() 127 if (new) in idr_layer_alloc() 128 return new; in idr_layer_alloc() [all …]
|
D | rbtree_test.c | 26 struct rb_node **new = &root->rb_node, *parent = NULL; in insert() local 29 while (*new) { in insert() 30 parent = *new; in insert() 32 new = &parent->rb_left; in insert() 34 new = &parent->rb_right; in insert() 37 rb_link_node(&node->rb, parent, new); in insert() 69 struct rb_node **new = &root->rb_node, *rb_parent = NULL; in RB_DECLARE_CALLBACKS() local 74 while (*new) { in RB_DECLARE_CALLBACKS() 75 rb_parent = *new; in RB_DECLARE_CALLBACKS() 80 new = &parent->rb.rb_left; in RB_DECLARE_CALLBACKS() [all …]
|
D | rbtree.c | 63 __rb_rotate_set_parents(struct rb_node *old, struct rb_node *new, in __rb_rotate_set_parents() argument 67 new->__rb_parent_color = old->__rb_parent_color; in __rb_rotate_set_parents() 68 rb_set_parent_color(old, new, color); in __rb_rotate_set_parents() 69 __rb_change_child(old, new, parent, root); in __rb_rotate_set_parents() 74 void (*augment_rotate)(struct rb_node *old, struct rb_node *new)) in __rb_insert() argument 203 void (*augment_rotate)(struct rb_node *old, struct rb_node *new)) in ____rb_erase_color() argument 365 void (*augment_rotate)(struct rb_node *old, struct rb_node *new)) in __rb_erase_color() argument 379 static inline void dummy_copy(struct rb_node *old, struct rb_node *new) {} in dummy_copy() argument 380 static inline void dummy_rotate(struct rb_node *old, struct rb_node *new) {} in dummy_rotate() argument 409 void (*augment_rotate)(struct rb_node *old, struct rb_node *new)) in __rb_insert_augmented() argument [all …]
|
D | debugobjects.c | 88 struct debug_obj *new; in fill_pool() local 99 new = kmem_cache_zalloc(obj_cache, gfp); in fill_pool() 100 if (!new) in fill_pool() 104 hlist_add_head(&new->node, &obj_pool); in fill_pool() 1025 struct debug_obj *obj, *new; in debug_objects_replace_static_objects() local 1054 new = hlist_entry(obj_pool.first, typeof(*obj), node); in debug_objects_replace_static_objects() 1055 hlist_del(&new->node); in debug_objects_replace_static_objects() 1057 *new = *obj; in debug_objects_replace_static_objects() 1058 hlist_add_head(&new->node, &db->list); in debug_objects_replace_static_objects()
|
D | atomic64.c | 140 long long atomic64_xchg(atomic64_t *v, long long new) in atomic64_xchg() argument 148 v->counter = new; in atomic64_xchg()
|
D | bitmap.c | 821 const unsigned long *old, const unsigned long *new, in bitmap_remap() argument 830 w = bitmap_weight(new, bits); in bitmap_remap() 837 set_bit(bitmap_ord_to_pos(new, n % w, bits), dst); in bitmap_remap() 869 const unsigned long *new, int bits) in bitmap_bitremap() argument 871 int w = bitmap_weight(new, bits); in bitmap_bitremap() 876 return bitmap_ord_to_pos(new, n % w, bits); in bitmap_bitremap()
|
D | btree.c | 471 unsigned long *new; in btree_insert_level() local 473 new = btree_node_alloc(head, gfp); in btree_insert_level() 474 if (!new) in btree_insert_level() 478 new, level + 1, gfp); in btree_insert_level() 480 mempool_free(new, head->mempool); in btree_insert_level() 484 setkey(geo, new, i, bkey(geo, node, i)); in btree_insert_level() 485 setval(geo, new, i, bval(geo, node, i)); in btree_insert_level()
|
D | lru_cache.c | 78 unsigned long old, new, val; in lc_try_lock() 81 new = old | LC_LOCKED; in lc_try_lock() 82 val = cmpxchg(&lc->flags, old, new); in lc_try_lock()
|
D | string.c | 953 char *strreplace(char *s, char old, char new) in strreplace() argument 957 *s = new; in strreplace()
|
D | Kconfig.kmemcheck | 61 be scheduled. If the queue is full, new error reports will be
|
D | Kconfig.kgdb | 37 intended to for the development of new kgdb stubs
|