Searched refs:entries (Results 1 – 7 of 7) sorted by relevance
/lib/ |
D | stackdepot.c | 69 unsigned long entries[]; /* Variable-sized array of entries. */ member 120 depot_alloc_stack(unsigned long *entries, int size, u32 hash, void **prealloc) in depot_alloc_stack() argument 123 size_t required_size = struct_size(stack, entries, size); in depot_alloc_stack() 154 memcpy(stack->entries, entries, flex_array_size(stack, entries, size)); in depot_alloc_stack() 196 unsigned long entries = 0; in stack_depot_early_init() local 211 entries = 1UL << stack_hash_order; in stack_depot_early_init() 214 entries, in stack_depot_early_init() 238 unsigned long entries; in stack_depot_init() local 242 entries = 1UL << stack_hash_order; in stack_depot_init() 244 entries = nr_free_buffer_pages(); in stack_depot_init() [all …]
|
D | list-test.c | 386 struct list_head entries[3], *cur; in list_test_list_cut_position() local 391 list_add_tail(&entries[0], &list1); in list_test_list_cut_position() 392 list_add_tail(&entries[1], &list1); in list_test_list_cut_position() 393 list_add_tail(&entries[2], &list1); in list_test_list_cut_position() 396 list_cut_position(&list2, &list1, &entries[1]); in list_test_list_cut_position() 400 KUNIT_EXPECT_PTR_EQ(test, cur, &entries[i]); in list_test_list_cut_position() 407 KUNIT_EXPECT_PTR_EQ(test, cur, &entries[i]); in list_test_list_cut_position() 414 struct list_head entries[3], *cur; in list_test_list_cut_before() local 419 list_add_tail(&entries[0], &list1); in list_test_list_cut_before() 420 list_add_tail(&entries[1], &list1); in list_test_list_cut_before() [all …]
|
D | test_rhashtable.c | 73 unsigned int entries; member 138 unsigned int entries) in test_rht_lookup() argument 142 for (i = 0; i < entries; i++) { in test_rht_lookup() 175 static void test_bucket_stats(struct rhashtable *ht, unsigned int entries) in test_bucket_stats() argument 202 total, atomic_read(&ht->nelems), entries, chain_len); in test_bucket_stats() 204 if (total != atomic_read(&ht->nelems) || total != entries) in test_bucket_stats() 209 unsigned int entries) in test_rhashtable() argument 220 pr_info(" Adding %d keys\n", entries); in test_rhashtable() 222 for (i = 0; i < entries; i++) { in test_rhashtable() 237 test_bucket_stats(ht, entries); in test_rhashtable() [all …]
|
D | ref_tracker.c | 71 unsigned long entries[REF_TRACKER_STACK_ENTRIES]; in ref_tracker_alloc() local 91 nr_entries = stack_trace_save(entries, ARRAY_SIZE(entries), 1); in ref_tracker_alloc() 92 tracker->alloc_stack_handle = stack_depot_save(entries, nr_entries, gfp); in ref_tracker_alloc() 104 unsigned long entries[REF_TRACKER_STACK_ENTRIES]; in ref_tracker_free() local 121 nr_entries = stack_trace_save(entries, ARRAY_SIZE(entries), 1); in ref_tracker_free() 122 stack_handle = stack_depot_save(entries, nr_entries, GFP_ATOMIC); in ref_tracker_free()
|
D | fault-inject.c | 70 unsigned long entries[MAX_STACK_TRACE_DEPTH]; in fail_stacktrace() local 77 nr_entries = stack_trace_save(entries, depth, 1); in fail_stacktrace() 79 if (attr->reject_start <= entries[n] && in fail_stacktrace() 80 entries[n] < attr->reject_end) in fail_stacktrace() 82 if (attr->require_start <= entries[n] && in fail_stacktrace() 83 entries[n] < attr->require_end) in fail_stacktrace()
|
D | Kconfig | 485 Support entries which occupy multiple consecutive indices in the
|
D | Kconfig.debug | 1951 bool "Debugfs entries for fault-injection capabilities"
|