| /lib/ |
| D | list_debug.c | 23 struct list_head *next) in __list_add_valid_or_report() argument 27 CHECK_DATA_CORRUPTION(next == NULL, in __list_add_valid_or_report() 29 CHECK_DATA_CORRUPTION(next->prev != prev, in __list_add_valid_or_report() 31 prev, next->prev, next) || in __list_add_valid_or_report() 32 CHECK_DATA_CORRUPTION(prev->next != next, in __list_add_valid_or_report() 34 next, prev->next, prev) || in __list_add_valid_or_report() 35 CHECK_DATA_CORRUPTION(new == prev || new == next, in __list_add_valid_or_report() 37 new, prev, next)) in __list_add_valid_or_report() 47 struct list_head *prev, *next; in __list_del_entry_valid_or_report() local 50 next = entry->next; in __list_del_entry_valid_or_report() [all …]
|
| D | list_sort.c | 25 tail = &a->next; in merge() 26 a = a->next; in merge() 33 tail = &b->next; in merge() 34 b = b->next; in merge() 61 tail->next = a; in merge_final() 64 a = a->next; in merge_final() 68 tail->next = b; in merge_final() 71 b = b->next; in merge_final() 80 tail->next = b; in merge_final() 93 b = b->next; in merge_final() [all …]
|
| D | cpumask.c | 24 unsigned int next; in cpumask_next_wrap() local 27 next = cpumask_next(n, mask); in cpumask_next_wrap() 29 if (wrap && n < start && next >= start) { in cpumask_next_wrap() 32 } else if (next >= nr_cpumask_bits) { in cpumask_next_wrap() 38 return next; in cpumask_next_wrap() 172 unsigned int next, prev; in cpumask_any_and_distribute() local 177 next = find_next_and_bit_wrap(cpumask_bits(src1p), cpumask_bits(src2p), in cpumask_any_and_distribute() 179 if (next < nr_cpu_ids) in cpumask_any_and_distribute() 180 __this_cpu_write(distribute_cpu_mask_prev, next); in cpumask_any_and_distribute() 182 return next; in cpumask_any_and_distribute() [all …]
|
| D | llist.c | 32 new_last->next = first; in llist_add_batch() 55 struct llist_node *entry, *next; in llist_del_first() local 61 next = READ_ONCE(entry->next); in llist_del_first() 62 } while (!try_cmpxchg(&head->first, &entry, next)); in llist_del_first() 82 struct llist_node *entry, *next; in llist_del_first_this() local 89 next = READ_ONCE(entry->next); in llist_del_first_this() 90 } while (!try_cmpxchg(&head->first, &entry, next)); in llist_del_first_this() 109 head = head->next; in llist_reverse_order() 110 tmp->next = new_head; in llist_reverse_order()
|
| D | plist.c | 35 WARN(n->prev != p || p->next != n, in plist_check_prev_next() 39 t, t->next, t->prev, in plist_check_prev_next() 40 p, p->next, p->prev, in plist_check_prev_next() 41 n, n->next, n->prev); in plist_check_prev_next() 46 struct list_head *prev = top, *next = top->next; in plist_check_list() local 48 plist_check_prev_next(top, prev, next); in plist_check_list() 49 while (next != top) { in plist_check_list() 50 WRITE_ONCE(prev, next); in plist_check_list() 51 WRITE_ONCE(next, prev->next); in plist_check_list() 52 plist_check_prev_next(top, prev, next); in plist_check_list() [all …]
|
| D | bootconfig.c | 171 return node->next ? &xbc_nodes[node->next] : NULL; in xbc_node_get_next() 343 struct xbc_node *next; in xbc_node_find_next_leaf() local 354 next = xbc_node_get_subkey(node); in xbc_node_find_next_leaf() 355 if (next) { in xbc_node_find_next_leaf() 356 node = next; in xbc_node_find_next_leaf() 363 while (!node->next) { in xbc_node_find_next_leaf() 419 node->next = 0; in xbc_init_node() 440 while (node->next) in xbc_last_sibling() 463 sib->next = xbc_node_index(node); in __xbc_add_sibling() 467 node->next = last_parent->child; in __xbc_add_sibling() [all …]
|
| D | kobject_uevent.c | 101 const char *next = buf; in action_arg_word_end() local 103 while (next <= buf_end && *next != delim) in action_arg_word_end() 104 if (!isalnum(*next++)) in action_arg_word_end() 107 if (next == buf) in action_arg_word_end() 110 return next; in action_arg_word_end() 117 const char *next, *buf_end, *key; in kobject_action_args() local 140 next = buf + UUID_STRING_LEN; in kobject_action_args() 143 while (next <= buf_end) { in kobject_action_args() 144 if (*next != ' ') in kobject_action_args() 148 key = ++next; in kobject_action_args() [all …]
|
| D | klist.c | 379 struct klist_node *next; in klist_next() local 385 next = to_klist_node(last->n_node.next); in klist_next() 389 next = to_klist_node(i->i_klist->k_list.next); in klist_next() 392 while (next != to_klist_node(&i->i_klist->k_list)) { in klist_next() 393 if (likely(!knode_dead(next))) { in klist_next() 394 kref_get(&next->n_ref); in klist_next() 395 i->i_cur = next; in klist_next() 398 next = to_klist_node(next->n_node.next); in klist_next()
|
| D | timerqueue.c | 75 struct rb_node *next; in timerqueue_iterate_next() local 79 next = rb_next(&node->node); in timerqueue_iterate_next() 80 if (!next) in timerqueue_iterate_next() 82 return container_of(next, struct timerqueue_node, node); in timerqueue_iterate_next()
|
| D | test_list_sort.c | 84 for (cur = head.next; cur->next != &head; cur = cur->next) { in list_sort_test() 88 KUNIT_ASSERT_PTR_EQ_MSG(test, cur->next->prev, cur, in list_sort_test() 91 cmp_result = cmp(test, cur, cur->next); in list_sort_test() 95 el1 = container_of(cur->next, struct debug_el, list); in list_sort_test()
|
| D | rhashtable.c | 235 struct rhash_head *head, *next, *entry; in rhashtable_rehash_one() local 248 next = rht_dereference_bucket(entry->next, old_tbl, old_hash); in rhashtable_rehash_one() 250 if (rht_is_a_nulls(next)) in rhashtable_rehash_one() 253 pprev = &entry->next; in rhashtable_rehash_one() 266 RCU_INIT_POINTER(entry->next, head); in rhashtable_rehash_one() 271 rcu_assign_pointer(*pprev, next); in rhashtable_rehash_one() 274 rht_assign_locked(bkt, next); in rhashtable_rehash_one() 517 pprev = &head->next; in rhashtable_lookup_one() 527 RCU_INIT_POINTER(list->next, plist); in rhashtable_lookup_one() 528 head = rht_dereference_bucket(head->next, tbl, hash); in rhashtable_lookup_one() [all …]
|
| D | list-test.c | 56 KUNIT_EXPECT_PTR_EQ(test, list.next, &b); in list_test_list_add() 58 KUNIT_EXPECT_PTR_EQ(test, b.next, &a); in list_test_list_add() 70 KUNIT_EXPECT_PTR_EQ(test, list.next, &a); in list_test_list_add_tail() 72 KUNIT_EXPECT_PTR_EQ(test, a.next, &b); in list_test_list_add_tail() 87 KUNIT_EXPECT_PTR_EQ(test, list.next, &b); in list_test_list_del() 103 KUNIT_EXPECT_PTR_EQ(test, list.next, &a_new); in list_test_list_replace() 105 KUNIT_EXPECT_PTR_EQ(test, a_new.next, &b); in list_test_list_replace() 121 KUNIT_EXPECT_PTR_EQ(test, list.next, &a_new); in list_test_list_replace_init() 123 KUNIT_EXPECT_PTR_EQ(test, a_new.next, &b); in list_test_list_replace_init() 142 KUNIT_EXPECT_PTR_EQ(test, &b, list.next); in list_test_list_swap() [all …]
|
| D | test_hmm.c | 785 unsigned long next; in dmirror_exclusive() local 798 for (addr = start; addr < end; addr = next) { in dmirror_exclusive() 802 next = min(end, addr + (ARRAY_SIZE(pages) << PAGE_SHIFT)); in dmirror_exclusive() 804 ret = make_device_exclusive_range(mm, addr, next, pages, NULL); in dmirror_exclusive() 810 if (ret == (next - addr) >> PAGE_SHIFT) in dmirror_exclusive() 811 mapped = dmirror_atomic_map(addr, next, pages, dmirror); in dmirror_exclusive() 819 if (addr + (mapped << PAGE_SHIFT) < next) { in dmirror_exclusive() 907 unsigned long next; in dmirror_migrate_to_system() local 921 for (addr = start; addr < end; addr = next) { in dmirror_migrate_to_system() 927 next = min(end, addr + (ARRAY_SIZE(src_pfns) << PAGE_SHIFT)); in dmirror_migrate_to_system() [all …]
|
| D | oid_registry.c | 80 goto next; in look_up_OID() 84 goto next; in look_up_OID() 88 next: in look_up_OID()
|
| D | ts_fsm.c | 137 struct ts_fsm_token *cur = NULL, *next; in fsm_find() local 169 next = &fsm->tokens[tok_idx + 1]; in fsm_find() 171 next = NULL; in fsm_find() 199 if (next == NULL) in fsm_find() 205 while (!match_token(next, data[block_idx])) { in fsm_find() 222 while (!match_token(next, data[block_idx])) { in fsm_find()
|
| D | test_xarray.c | 209 unsigned long next = base + (1UL << order); in check_xa_mark_1() local 216 XA_BUG_ON(xa, xa_store_index(xa, next, GFP_KERNEL)); in check_xa_mark_1() 219 for (i = base; i < next; i++) { in check_xa_mark_1() 244 XA_BUG_ON(xa, xa_get_mark(xa, next, XA_MARK_0)); in check_xa_mark_1() 245 XA_BUG_ON(xa, xa_get_mark(xa, next, XA_MARK_1)); in check_xa_mark_1() 246 XA_BUG_ON(xa, xa_get_mark(xa, next, XA_MARK_2)); in check_xa_mark_1() 248 xa_erase_index(xa, next); in check_xa_mark_1() 1039 u32 next = 0; in check_xa_alloc_3() local 1045 &next, GFP_KERNEL) != 0); in check_xa_alloc_3() 1048 next = 0x3ffd; in check_xa_alloc_3() [all …]
|
| D | xarray.c | 271 struct xa_node *next, *node = xas->xa_alloc; in xas_destroy() local 275 next = rcu_dereference_raw(node->parent); in xas_destroy() 277 xas->xa_alloc = node = next; in xas_destroy() 783 void *first, *next; in xas_store() local 801 next = first; in xas_store() 821 if (xa_is_node(next) && (!node || node->shift)) in xas_store() 822 xas_free_nodes(xas, xa_to_node(next)); in xas_store() 825 count += !next - !entry; in xas_store() 836 next = xa_entry_locked(xas->xa, node, ++offset); in xas_store() 837 if (!xa_is_sibling(next)) { in xas_store() [all …]
|
| D | test_rhashtable.c | 483 struct rhash_head *pos, *next; in print_ht() local 487 next = !rht_is_a_nulls(pos) ? rht_dereference(pos->next, ht) : NULL; in print_ht() 498 list = rht_dereference(list->next, ht); in print_ht() 506 pos = next, in print_ht() 507 next = !rht_is_a_nulls(pos) ? in print_ht() 508 rht_dereference(pos->next, ht) : NULL; in print_ht()
|
| D | maple_tree.c | 891 void *next; in mt_clear_meta() local 898 next = mt_slot_locked(mt, slots, in mt_clear_meta() 900 if (unlikely((mte_to_node(next) && in mt_clear_meta() 901 mte_node_type(next)))) in mt_clear_meta() 964 mte_to_mat(dead_enode)->next = NULL; in mat_add() 970 mte_to_mat(mat->tail)->next = dead_enode; in mat_add() 986 struct maple_enode *next; in mas_mat_destroy() local 991 next = mte_to_mat(mat->head)->next; in mas_mat_destroy() 996 mat->head = next; in mas_mat_destroy() 2744 struct maple_enode *next, *last; in mtree_range_walk() local [all …]
|
| /lib/lzo/ |
| D | lzo1x_decompress_safe.c | 44 size_t t, next; in lzo1x_decompress_safe() local 68 next = t; in lzo1x_decompress_safe() 121 next = t & 3; in lzo1x_decompress_safe() 132 next = t & 3; in lzo1x_decompress_safe() 139 next = t & 3; in lzo1x_decompress_safe() 163 next = get_unaligned_le16(ip); in lzo1x_decompress_safe() 165 m_pos -= next >> 2; in lzo1x_decompress_safe() 166 next &= 3; in lzo1x_decompress_safe() 169 next = get_unaligned_le16(ip); in lzo1x_decompress_safe() 170 if (((next & 0xfffc) == 0xfffc) && in lzo1x_decompress_safe() [all …]
|
| /lib/zlib_inflate/ |
| D | inftrees.c | 40 code *next; /* next available space in table */ in zlib_inflate_table() local 190 next = *table; /* current table to fill in */ in zlib_inflate_table() 224 next[(huff >> drop) + fill] = this; in zlib_inflate_table() 252 next += min; /* here min is 1 << curr */ in zlib_inflate_table() 273 (*table)[low].val = (unsigned short)(next - *table); in zlib_inflate_table() 292 next = *table; in zlib_inflate_table() 297 next[huff >> drop] = this; in zlib_inflate_table()
|
| D | inflate.c | 48 state->lencode = state->distcode = state->next = state->codes; in zlib_inflateReset() 189 next = strm->next_in; \ 200 strm->next_in = next; \ 219 hold += (unsigned long)(*next++) << bits; \ 334 const unsigned char *next; /* next input */ in zlib_inflate() local 456 memcpy(put, next, copy); in zlib_inflate() 458 next += copy; in zlib_inflate() 492 state->next = state->codes; in zlib_inflate() 493 state->lencode = (code const *)(state->next); in zlib_inflate() 495 ret = zlib_inflate_table(CODES, state->lens, 19, &(state->next), in zlib_inflate() [all …]
|
| /lib/crypto/mpi/ |
| D | mpih-mul.c | 374 if (!ctx->next) { in mpihelp_mul_karatsuba_case() 375 ctx->next = kzalloc(sizeof *ctx, GFP_KERNEL); in mpihelp_mul_karatsuba_case() 376 if (!ctx->next) in mpihelp_mul_karatsuba_case() 382 ctx->next) < 0) in mpihelp_mul_karatsuba_case() 401 for (ctx = ctx->next; ctx; ctx = ctx2) { in mpihelp_release_karatsuba_ctx() 402 ctx2 = ctx->next; in mpihelp_release_karatsuba_ctx()
|
| /lib/zlib_deflate/ |
| D | deflate.c | 200 char *next; in zlib_deflateInit2() local 229 next = (char *) mem; in zlib_deflateInit2() 230 next += sizeof(*mem); in zlib_deflateInit2() 236 mem->window_memory = (Byte *) PTR_ALIGN(next, PAGE_SIZE); in zlib_deflateInit2() 238 mem->window_memory = (Byte *) next; in zlib_deflateInit2() 240 next += zlib_deflate_window_memsize(windowBits); in zlib_deflateInit2() 241 mem->prev_memory = (Pos *) next; in zlib_deflateInit2() 242 next += zlib_deflate_prev_memsize(windowBits); in zlib_deflateInit2() 243 mem->head_memory = (Pos *) next; in zlib_deflateInit2() 244 next += zlib_deflate_head_memsize(memLevel); in zlib_deflateInit2() [all …]
|
| /lib/842/ |
| D | 842_compress.c | 482 u64 last, next, pad, total; in sw842_compress() local 517 next = get_unaligned((u64 *)p->in); in sw842_compress() 528 if (next == last) { in sw842_compress() 538 if (next == last) /* reached max repeat bits */ in sw842_compress() 542 if (next == 0) in sw842_compress() 551 last = next; in sw842_compress()
|