/lib/ |
D | list_debug.c | 21 struct list_head *next) in __list_add_valid() argument 23 if (CHECK_DATA_CORRUPTION(next->prev != prev, in __list_add_valid() 25 prev, next->prev, next) || in __list_add_valid() 26 CHECK_DATA_CORRUPTION(prev->next != next, in __list_add_valid() 28 next, prev->next, prev) || in __list_add_valid() 29 CHECK_DATA_CORRUPTION(new == prev || new == next, in __list_add_valid() 31 new, prev, next)) in __list_add_valid() 40 struct list_head *prev, *next; in __list_del_entry_valid() local 43 next = entry->next; in __list_del_entry_valid() 45 if (CHECK_DATA_CORRUPTION(next == LIST_POISON1, in __list_del_entry_valid() [all …]
|
D | ioremap.c | 107 unsigned long next; in ioremap_pmd_range() local 113 next = pmd_addr_end(addr, end); in ioremap_pmd_range() 115 if (ioremap_try_huge_pmd(pmd, addr, next, phys_addr, prot)) in ioremap_pmd_range() 118 if (ioremap_pte_range(pmd, addr, next, phys_addr, prot)) in ioremap_pmd_range() 120 } while (pmd++, phys_addr += (next - addr), addr = next, addr != end); in ioremap_pmd_range() 150 unsigned long next; in ioremap_pud_range() local 156 next = pud_addr_end(addr, end); in ioremap_pud_range() 158 if (ioremap_try_huge_pud(pud, addr, next, phys_addr, prot)) in ioremap_pud_range() 161 if (ioremap_pmd_range(pud, addr, next, phys_addr, prot)) in ioremap_pud_range() 163 } while (pud++, phys_addr += (next - addr), addr = next, addr != end); in ioremap_pud_range() [all …]
|
D | list_sort.c | 28 tail = &a->next; in merge() 29 a = a->next; in merge() 36 tail = &b->next; in merge() 37 b = b->next; in merge() 64 tail->next = a; in merge_final() 67 a = a->next; in merge_final() 71 tail->next = b; in merge_final() 74 b = b->next; in merge_final() 83 tail->next = b; in merge_final() 96 b = b->next; in merge_final() [all …]
|
D | plist.c | 35 WARN(n->prev != p || p->next != n, in plist_check_prev_next() 39 t, t->next, t->prev, in plist_check_prev_next() 40 p, p->next, p->prev, in plist_check_prev_next() 41 n, n->next, n->prev); in plist_check_prev_next() 46 struct list_head *prev = top, *next = top->next; in plist_check_list() local 48 plist_check_prev_next(top, prev, next); in plist_check_list() 49 while (next != top) { in plist_check_list() 50 prev = next; in plist_check_list() 51 next = prev->next; in plist_check_list() 52 plist_check_prev_next(top, prev, next); in plist_check_list() [all …]
|
D | llist.c | 32 new_last->next = first = READ_ONCE(head->first); in llist_add_batch() 55 struct llist_node *entry, *old_entry, *next; in llist_del_first() local 62 next = READ_ONCE(entry->next); in llist_del_first() 63 entry = cmpxchg(&head->first, old_entry, next); in llist_del_first() 85 head = head->next; in llist_reverse_order() 86 tmp->next = new_head; in llist_reverse_order()
|
D | kobject_uevent.c | 102 const char *next = buf; in action_arg_word_end() local 104 while (next <= buf_end && *next != delim) in action_arg_word_end() 105 if (!isalnum(*next++)) in action_arg_word_end() 108 if (next == buf) in action_arg_word_end() 111 return next; in action_arg_word_end() 118 const char *next, *buf_end, *key; in kobject_action_args() local 141 next = buf + UUID_STRING_LEN; in kobject_action_args() 144 while (next <= buf_end) { in kobject_action_args() 145 if (*next != ' ') in kobject_action_args() 149 key = ++next; in kobject_action_args() [all …]
|
D | klist.c | 379 struct klist_node *next; in klist_next() local 385 next = to_klist_node(last->n_node.next); in klist_next() 389 next = to_klist_node(i->i_klist->k_list.next); in klist_next() 392 while (next != to_klist_node(&i->i_klist->k_list)) { in klist_next() 393 if (likely(!knode_dead(next))) { in klist_next() 394 kref_get(&next->n_ref); in klist_next() 395 i->i_cur = next; in klist_next() 398 next = to_klist_node(next->n_node.next); in klist_next()
|
D | test_list_sort.c | 100 for (cur = head.next; cur->next != &head; cur = cur->next) { in list_sort_test() 104 if (cur->next->prev != cur) { in list_sort_test() 109 cmp_result = cmp(NULL, cur, cur->next); in list_sort_test() 116 el1 = container_of(cur->next, struct debug_el, list); in list_sort_test()
|
D | timerqueue.c | 85 struct rb_node *next; in timerqueue_iterate_next() local 89 next = rb_next(&node->node); in timerqueue_iterate_next() 90 if (!next) in timerqueue_iterate_next() 92 return container_of(next, struct timerqueue_node, node); in timerqueue_iterate_next()
|
D | cpumask.c | 79 int next; in cpumask_next_wrap() local 82 next = cpumask_next(n, mask); in cpumask_next_wrap() 84 if (wrap && n < start && next >= start) { in cpumask_next_wrap() 87 } else if (next >= nr_cpumask_bits) { in cpumask_next_wrap() 93 return next; in cpumask_next_wrap()
|
D | rhashtable.c | 222 struct rhash_head *head, *next, *entry; in rhashtable_rehash_one() local 234 next = rht_dereference_bucket(entry->next, old_tbl, old_hash); in rhashtable_rehash_one() 236 if (rht_is_a_nulls(next)) in rhashtable_rehash_one() 239 pprev = &entry->next; in rhashtable_rehash_one() 251 RCU_INIT_POINTER(entry->next, head); in rhashtable_rehash_one() 256 rcu_assign_pointer(*pprev, next); in rhashtable_rehash_one() 259 rht_assign_locked(bkt, next); in rhashtable_rehash_one() 501 pprev = &head->next; in rhashtable_lookup_one() 511 RCU_INIT_POINTER(list->next, plist); in rhashtable_lookup_one() 512 head = rht_dereference_bucket(head->next, tbl, hash); in rhashtable_lookup_one() [all …]
|
D | cmdline.c | 205 char *next; in next_arg() local 243 next = args + i + 1; in next_arg() 245 next = args + i; in next_arg() 248 return skip_spaces(next); in next_arg()
|
D | oid_registry.c | 79 goto next; in look_up_OID() 83 goto next; in look_up_OID() 87 next: in look_up_OID()
|
D | test_xarray.c | 206 unsigned long next = base + (1UL << order); in check_xa_mark_1() local 213 XA_BUG_ON(xa, xa_store_index(xa, next, GFP_KERNEL)); in check_xa_mark_1() 216 for (i = base; i < next; i++) { in check_xa_mark_1() 241 XA_BUG_ON(xa, xa_get_mark(xa, next, XA_MARK_0)); in check_xa_mark_1() 242 XA_BUG_ON(xa, xa_get_mark(xa, next, XA_MARK_1)); in check_xa_mark_1() 243 XA_BUG_ON(xa, xa_get_mark(xa, next, XA_MARK_2)); in check_xa_mark_1() 245 xa_erase_index(xa, next); in check_xa_mark_1() 778 u32 next = 0; in check_xa_alloc_3() local 784 &next, GFP_KERNEL) != 0); in check_xa_alloc_3() 787 next = 0x3ffd; in check_xa_alloc_3() [all …]
|
D | stackdepot.c | 60 struct stack_record *next; /* Link in the hashtable */ member 179 for (found = bucket; found; found = found->next) { in find_stack() 276 new->next = *bucket; in stack_depot_save()
|
D | ts_fsm.c | 137 struct ts_fsm_token *cur = NULL, *next; in fsm_find() local 169 next = &fsm->tokens[tok_idx + 1]; in fsm_find() 171 next = NULL; in fsm_find() 199 if (next == NULL) in fsm_find() 205 while (!match_token(next, data[block_idx])) { in fsm_find() 222 while (!match_token(next, data[block_idx])) { in fsm_find()
|
D | test_rhashtable.c | 497 struct rhash_head *pos, *next; in print_ht() local 501 next = !rht_is_a_nulls(pos) ? rht_dereference(pos->next, ht) : NULL; in print_ht() 512 list = rht_dereference(list->next, ht); in print_ht() 520 pos = next, in print_ht() 521 next = !rht_is_a_nulls(pos) ? in print_ht() 522 rht_dereference(pos->next, ht) : NULL; in print_ht()
|
D | xarray.c | 776 void *first, *next; in xas_store() local 794 next = first; in xas_store() 814 if (xa_is_node(next) && (!node || node->shift)) in xas_store() 815 xas_free_nodes(xas, xa_to_node(next)); in xas_store() 818 count += !next - !entry; in xas_store() 829 next = xa_entry_locked(xas->xa, node, ++offset); in xas_store() 830 if (!xa_is_sibling(next)) { in xas_store() 833 first = next; in xas_store() 1660 struct xa_limit limit, u32 *next, gfp_t gfp) in __xa_alloc_cyclic() argument 1665 limit.min = max(min, *next); in __xa_alloc_cyclic() [all …]
|
D | ts_bm.c | 79 goto next; in bm_find() 85 next: bs = bm->bad_shift[text[shift-i]]; in bm_find()
|
/lib/lzo/ |
D | lzo1x_decompress_safe.c | 44 size_t t, next; in lzo1x_decompress_safe() local 68 next = t; in lzo1x_decompress_safe() 121 next = t & 3; in lzo1x_decompress_safe() 132 next = t & 3; in lzo1x_decompress_safe() 139 next = t & 3; in lzo1x_decompress_safe() 163 next = get_unaligned_le16(ip); in lzo1x_decompress_safe() 165 m_pos -= next >> 2; in lzo1x_decompress_safe() 166 next &= 3; in lzo1x_decompress_safe() 169 next = get_unaligned_le16(ip); in lzo1x_decompress_safe() 170 if (((next & 0xfffc) == 0xfffc) && in lzo1x_decompress_safe() [all …]
|
/lib/zlib_inflate/ |
D | inftrees.c | 40 code *next; /* next available space in table */ in zlib_inflate_table() local 190 next = *table; /* current table to fill in */ in zlib_inflate_table() 224 next[(huff >> drop) + fill] = this; in zlib_inflate_table() 252 next += min; /* here min is 1 << curr */ in zlib_inflate_table() 273 (*table)[low].val = (unsigned short)(next - *table); in zlib_inflate_table() 292 next = *table; in zlib_inflate_table() 297 next[huff >> drop] = this; in zlib_inflate_table()
|
D | inflate.c | 38 state->lencode = state->distcode = state->next = state->codes; in zlib_inflateReset() 170 next = strm->next_in; \ 181 strm->next_in = next; \ 200 hold += (unsigned long)(*next++) << bits; \ 320 const unsigned char *next; /* next input */ in zlib_inflate() local 441 memcpy(put, next, copy); in zlib_inflate() 443 next += copy; in zlib_inflate() 477 state->next = state->codes; in zlib_inflate() 478 state->lencode = (code const *)(state->next); in zlib_inflate() 480 ret = zlib_inflate_table(CODES, state->lens, 19, &(state->next), in zlib_inflate() [all …]
|
/lib/mpi/ |
D | mpih-mul.c | 374 if (!ctx->next) { in mpihelp_mul_karatsuba_case() 375 ctx->next = kzalloc(sizeof *ctx, GFP_KERNEL); in mpihelp_mul_karatsuba_case() 376 if (!ctx->next) in mpihelp_mul_karatsuba_case() 382 ctx->next) < 0) in mpihelp_mul_karatsuba_case() 401 for (ctx = ctx->next; ctx; ctx = ctx2) { in mpihelp_release_karatsuba_ctx() 402 ctx2 = ctx->next; in mpihelp_release_karatsuba_ctx()
|
/lib/zlib_deflate/ |
D | deflate.c | 179 char *next; in zlib_deflateInit2() local 208 next = (char *) mem; in zlib_deflateInit2() 209 next += sizeof(*mem); in zlib_deflateInit2() 210 mem->window_memory = (Byte *) next; in zlib_deflateInit2() 211 next += zlib_deflate_window_memsize(windowBits); in zlib_deflateInit2() 212 mem->prev_memory = (Pos *) next; in zlib_deflateInit2() 213 next += zlib_deflate_prev_memsize(windowBits); in zlib_deflateInit2() 214 mem->head_memory = (Pos *) next; in zlib_deflateInit2() 215 next += zlib_deflate_head_memsize(memLevel); in zlib_deflateInit2() 216 mem->overlay_memory = next; in zlib_deflateInit2()
|
/lib/842/ |
D | 842_compress.c | 482 u64 last, next, pad, total; in sw842_compress() local 517 next = get_unaligned((u64 *)p->in); in sw842_compress() 528 if (next == last) { in sw842_compress() 536 if (next == last) /* reached max repeat bits */ in sw842_compress() 540 if (next == 0) in sw842_compress() 549 last = next; in sw842_compress()
|