/lib/ |
D | xarray.c | 181 void *entry; in xas_start() local 188 entry = xa_head(xas->xa); in xas_start() 189 if (!xa_is_node(entry)) { in xas_start() 193 if ((xas->xa_index >> xa_to_node(entry)->shift) > XA_CHUNK_MASK) in xas_start() 198 return entry; in xas_start() 204 void *entry = xa_entry(xas->xa, node, offset); in xas_descend() local 207 if (xa_is_sibling(entry)) { in xas_descend() 208 offset = xa_to_sibling(entry); in xas_descend() 209 entry = xa_entry(xas->xa, node, offset); in xas_descend() 213 return entry; in xas_descend() [all …]
|
D | list_debug.c | 38 bool __list_del_entry_valid(struct list_head *entry) in __list_del_entry_valid() argument 42 prev = entry->prev; in __list_del_entry_valid() 43 next = entry->next; in __list_del_entry_valid() 47 entry, LIST_POISON1) || in __list_del_entry_valid() 50 entry, LIST_POISON2) || in __list_del_entry_valid() 51 CHECK_DATA_CORRUPTION(prev->next != entry, in __list_del_entry_valid() 53 entry, prev->next) || in __list_del_entry_valid() 54 CHECK_DATA_CORRUPTION(next->prev != entry, in __list_del_entry_valid() 56 entry, next->prev)) in __list_del_entry_valid()
|
D | logic_pio.c | 238 struct logic_pio_hwaddr *entry = find_io_range(addr); \ 240 if (entry && entry->ops) \ 241 ret = entry->ops->in(entry->hostdata, \ 254 struct logic_pio_hwaddr *entry = find_io_range(addr); \ 256 if (entry && entry->ops) \ 257 entry->ops->out(entry->hostdata, \ 270 struct logic_pio_hwaddr *entry = find_io_range(addr); \ 272 if (entry && entry->ops) \ 273 entry->ops->ins(entry->hostdata, \ 287 struct logic_pio_hwaddr *entry = find_io_range(addr); \ [all …]
|
D | llist.c | 55 struct llist_node *entry, *old_entry, *next; in llist_del_first() local 57 entry = smp_load_acquire(&head->first); in llist_del_first() 59 if (entry == NULL) in llist_del_first() 61 old_entry = entry; in llist_del_first() 62 next = READ_ONCE(entry->next); in llist_del_first() 63 entry = cmpxchg(&head->first, old_entry, next); in llist_del_first() 64 if (entry == old_entry) in llist_del_first() 68 return entry; in llist_del_first()
|
D | idr.c | 231 void *entry = NULL; in idr_get_next_ul() local 237 entry = rcu_dereference_raw(*slot); in idr_get_next_ul() 238 if (!entry) in idr_get_next_ul() 240 if (!xa_is_internal(entry)) in idr_get_next_ul() 242 if (slot != &idr->idr_rt.xa_head && !xa_is_retry(entry)) in idr_get_next_ul() 250 return entry; in idr_get_next_ul() 267 void *entry = idr_get_next_ul(idr, &id); in idr_get_next() local 272 return entry; in idr_get_next() 294 void *entry; in idr_replace() local 298 entry = __radix_tree_lookup(&idr->idr_rt, id, &node, &slot); in idr_replace() [all …]
|
D | test_xarray.c | 69 unsigned order, void *entry, gfp_t gfp) in xa_store_order() argument 76 curr = xas_store(&xas, entry); in xa_store_order() 102 void *entry; in check_xas_retry() local 135 xas_for_each(&xas, entry, ULONG_MAX) { in check_xas_retry() 150 void *entry = xa_load(xa, j); in check_xa_load() local 152 XA_BUG_ON(xa, xa_to_value(entry) != j); in check_xa_load() 154 XA_BUG_ON(xa, entry); in check_xa_load() 161 void *entry = xa_load(xa, j); in check_xa_load() local 163 XA_BUG_ON(xa, xa_to_value(entry) != j); in check_xa_load() 165 XA_BUG_ON(xa, entry); in check_xa_load() [all …]
|
D | error-inject.c | 63 unsigned long entry, offset = 0, size = 0; in populate_error_injection_list() local 67 entry = arch_deref_entry_point((void *)iter->addr); in populate_error_injection_list() 69 if (!kernel_text_address(entry) || in populate_error_injection_list() 70 !kallsyms_lookup_size_offset(entry, &size, &offset)) { in populate_error_injection_list() 72 (void *)entry); in populate_error_injection_list() 79 ent->start_addr = entry; in populate_error_injection_list() 80 ent->end_addr = entry + size; in populate_error_injection_list()
|
D | radix-tree.c | 98 void __rcu **entry = rcu_dereference_raw(parent->slots[offset]); in radix_tree_descend() local 100 *nodep = (void *)entry; in radix_tree_descend() 420 void *entry; in radix_tree_extend() local 429 entry = rcu_dereference_raw(root->xa_head); in radix_tree_extend() 430 if (!entry && (!is_idr(root) || root_tag_get(root, IDR_FREE))) in radix_tree_extend() 454 if (radix_tree_is_internal_node(entry)) { in radix_tree_extend() 455 entry_to_node(entry)->parent = node; in radix_tree_extend() 456 } else if (xa_is_value(entry)) { in radix_tree_extend() 464 node->slots[0] = (void __rcu *)entry; in radix_tree_extend() 465 entry = node_to_entry(node); in radix_tree_extend() [all …]
|
D | nlattr.c | 75 const struct nlattr *entry; in nla_validate_array() local 78 nla_for_each_attr(entry, head, len, rem) { in nla_validate_array() 81 if (nla_len(entry) == 0) in nla_validate_array() 84 if (nla_len(entry) < NLA_HDRLEN) { in nla_validate_array() 85 NL_SET_ERR_MSG_ATTR(extack, entry, in nla_validate_array() 90 ret = __nla_validate(nla_data(entry), nla_len(entry), in nla_validate_array()
|
D | rhashtable.c | 222 struct rhash_head *head, *next, *entry; in rhashtable_rehash_one() local 231 rht_for_each_from(entry, rht_ptr(bkt, old_tbl, old_hash), in rhashtable_rehash_one() 234 next = rht_dereference_bucket(entry->next, old_tbl, old_hash); in rhashtable_rehash_one() 239 pprev = &entry->next; in rhashtable_rehash_one() 245 new_hash = head_hashfn(ht, new_tbl, entry); in rhashtable_rehash_one() 251 RCU_INIT_POINTER(entry->next, head); in rhashtable_rehash_one() 253 rht_assign_unlock(new_tbl, &new_tbl->buckets[new_hash], entry); in rhashtable_rehash_one()
|
D | kobject.c | 196 list_add_tail(&kobj->entry, &kobj->kset->list); in kobj_kset_join() 207 list_del_init(&kobj->entry); in kobj_kset_leave() 217 INIT_LIST_HEAD(&kobj->entry); in kobject_init_internal() 901 list_for_each_entry(k, &kset->list, entry) { in kset_find_obj()
|
D | sbitmap.c | 677 list_del_init(&sbq_wait->wait.entry); in sbitmap_del_wait_queue()
|