/external/linux-kselftest/tools/testing/selftests/bpf/map_tests/ |
D | htab_map_batch_ops.c | 13 static void map_batch_update(int map_fd, __u32 max_entries, int *keys, in map_batch_update() argument 27 for (i = 0; i < max_entries; i++) { in map_batch_update() 36 err = bpf_map_update_batch(map_fd, keys, values, &max_entries, &opts); in map_batch_update() 40 static void map_batch_verify(int *visited, __u32 max_entries, in map_batch_verify() argument 50 memset(visited, 0, max_entries * sizeof(*visited)); in map_batch_verify() 51 for (i = 0; i < max_entries; i++) { in map_batch_verify() 70 for (i = 0; i < max_entries; i++) { in map_batch_verify() 81 const __u32 max_entries = 10; in __test_map_lookup_and_delete_batch() local 82 value pcpu_values[max_entries]; in __test_map_lookup_and_delete_batch() 92 "hash_map", sizeof(int), sizeof(int), max_entries, NULL); in __test_map_lookup_and_delete_batch() [all …]
|
D | lpm_trie_map_batch_ops.c | 21 static void map_batch_update(int map_fd, __u32 max_entries, in map_batch_update() argument 32 for (i = 0; i < max_entries; i++) { in map_batch_update() 39 err = bpf_map_update_batch(map_fd, keys, values, &max_entries, &opts); in map_batch_update() 43 static void map_batch_verify(int *visited, __u32 max_entries, in map_batch_verify() argument 50 memset(visited, 0, max_entries * sizeof(*visited)); in map_batch_verify() 51 for (i = 0; i < max_entries; i++) { in map_batch_verify() 59 for (i = 0; i < max_entries; i++) { in map_batch_verify() 71 const __u32 max_entries = 10; in test_lpm_trie_map_batch_ops() local 81 max_entries, &create_opts); in test_lpm_trie_map_batch_ops() 85 keys = malloc(max_entries * sizeof(struct test_lpm_key)); in test_lpm_trie_map_batch_ops() [all …]
|
D | array_map_batch_ops.c | 14 static void map_batch_update(int map_fd, __u32 max_entries, int *keys, in map_batch_update() argument 24 for (i = 0; i < max_entries; i++) { in map_batch_update() 35 err = bpf_map_update_batch(map_fd, keys, values, &max_entries, &opts); in map_batch_update() 39 static void map_batch_verify(int *visited, __u32 max_entries, int *keys, in map_batch_verify() argument 45 memset(visited, 0, max_entries * sizeof(*visited)); in map_batch_verify() 46 for (i = 0; i < max_entries; i++) { in map_batch_verify() 63 for (i = 0; i < max_entries; i++) { in map_batch_verify() 73 const __u32 max_entries = 10; in __test_map_lookup_and_update_batch() local 83 "array_map", sizeof(int), sizeof(__s64), max_entries, NULL); in __test_map_lookup_and_update_batch() 91 keys = calloc(max_entries, sizeof(*keys)); in __test_map_lookup_and_update_batch() [all …]
|
/external/cronet/net/third_party/quiche/src/quiche/quic/core/qpack/ |
D | qpack_required_insert_count.cc | 14 uint64_t max_entries) { in QpackEncodeRequiredInsertCount() argument 19 return required_insert_count % (2 * max_entries) + 1; in QpackEncodeRequiredInsertCount() 23 uint64_t max_entries, in QpackDecodeRequiredInsertCount() argument 33 QUICHE_DCHECK_LE(max_entries, std::numeric_limits<uint64_t>::max() / 32); in QpackDecodeRequiredInsertCount() 35 if (encoded_required_insert_count > 2 * max_entries) { in QpackDecodeRequiredInsertCount() 43 uint64_t current_wrapped = total_number_of_inserts % (2 * max_entries); in QpackDecodeRequiredInsertCount() 46 if (current_wrapped >= *required_insert_count + max_entries) { in QpackDecodeRequiredInsertCount() 48 *required_insert_count += 2 * max_entries; in QpackDecodeRequiredInsertCount() 49 } else if (current_wrapped + max_entries < *required_insert_count) { in QpackDecodeRequiredInsertCount() 51 current_wrapped += 2 * max_entries; in QpackDecodeRequiredInsertCount()
|
D | qpack_required_insert_count_test.cc | 30 uint64_t max_entries; member 55 const uint64_t max_entries = kTestData[i].max_entries; in TEST() local 61 ASSERT_LT(0u, max_entries) << i; in TEST() 65 ASSERT_LT(total_number_of_inserts, required_insert_count + max_entries) in TEST() 70 ASSERT_LE(required_insert_count, total_number_of_inserts + max_entries) in TEST() 75 QpackEncodeRequiredInsertCount(required_insert_count, max_entries); in TEST() 82 encoded_required_insert_count, max_entries, total_number_of_inserts, in TEST() 95 uint64_t max_entries; member 116 kInvalidTestData[i].max_entries, in TEST()
|
/external/linux-kselftest/tools/testing/selftests/bpf/progs/ |
D | map_ptr_kern.c | 34 __u32 max_entries; member 39 __u32 value_size, __u32 max_entries) in check_bpf_map_fields() argument 44 VERIFY(map->max_entries == max_entries); in check_bpf_map_fields() 56 VERIFY(indirect->max_entries == direct->max_entries); in check_bpf_map_ptr() 63 __u32 key_size, __u32 value_size, __u32 max_entries) in check() argument 67 max_entries)); in check() 101 __uint(max_entries, MAX_ENTRIES); 137 __uint(max_entries, MAX_ENTRIES); 152 for (i = 0; i < array->map.max_entries && i < LOOP_BOUND; ++i) { in check_array() 169 __uint(max_entries, MAX_ENTRIES); [all …]
|
D | test_btf_map_in_map.c | 8 __uint(max_entries, 1); 16 __uint(max_entries, 2); 23 __uint(max_entries, 3); 31 __uint(max_entries, 1); 47 __uint(max_entries, 3); 56 __uint(max_entries, 5); 63 __uint(max_entries, 3); 69 __uint(max_entries, 1); 83 __uint(max_entries, 5); 99 __uint(max_entries, 1); [all …]
|
D | lsm.c | 14 __uint(max_entries, 1); 21 __uint(max_entries, 1); 28 __uint(max_entries, 1); 35 __uint(max_entries, 1); 42 __uint(max_entries, 1); 49 __uint(max_entries, 1); 56 __uint(max_entries, 1); 63 __uint(max_entries, 1); 73 __uint(max_entries, 1);
|
D | sockmap_verdict_prog.c | 7 __uint(max_entries, 20); 14 __uint(max_entries, 20); 21 __uint(max_entries, 20); 28 __uint(max_entries, 20);
|
D | test_stacktrace_build_id.c | 13 __uint(max_entries, 1); 20 __uint(max_entries, 16384); 29 __uint(max_entries, 128); 37 __uint(max_entries, 128);
|
D | test_stacktrace_map.c | 13 __uint(max_entries, 1); 20 __uint(max_entries, 16384); 29 __uint(max_entries, 16384); 36 __uint(max_entries, 16384);
|
D | test_sockmap_kern.h | 33 __uint(max_entries, 20); 40 __uint(max_entries, 20); 47 __uint(max_entries, 20); 54 __uint(max_entries, 1); 61 __uint(max_entries, 1); 68 __uint(max_entries, 6); 75 __uint(max_entries, 1); 82 __uint(max_entries, 3); 89 __uint(max_entries, 20);
|
D | test_pinning.c | 8 __uint(max_entries, 1); 16 __uint(max_entries, 1); 23 __uint(max_entries, 1);
|
D | test_select_reuseport_kern.c | 24 __uint(max_entries, 1); 31 __uint(max_entries, NR_RESULTS); 38 __uint(max_entries, 1); 45 __uint(max_entries, 1); 52 __uint(max_entries, 1);
|
D | test_sockmap_update.c | 8 __uint(max_entries, 1); 15 __uint(max_entries, 1); 22 __uint(max_entries, 1);
|
/external/linux-kselftest/tools/testing/selftests/bpf/prog_tests/ |
D | btf.c | 74 __u32 max_entries; member 142 .max_entries = 4, 197 .max_entries = 4, 222 .max_entries = 4, 263 .max_entries = 4, 308 .max_entries = 1, 330 .max_entries = 1, 352 .max_entries = 1, 374 .max_entries = 1, 399 .max_entries = 1, [all …]
|
D | for_each.c | 12 int i, err, hashmap_fd, max_entries, percpu_map_fd; in test_hash_map() local 23 max_entries = bpf_map__max_entries(skel->maps.hashmap); in test_hash_map() 24 for (i = 0; i < max_entries; i++) { in test_hash_map() 53 ASSERT_EQ(skel->bss->hashmap_elems, max_entries, "hashmap_elems"); in test_hash_map() 72 __u32 key, num_cpus, max_entries, retval; in test_array_map() local 84 max_entries = bpf_map__max_entries(skel->maps.arraymap); in test_array_map() 85 for (i = 0; i < max_entries; i++) { in test_array_map() 89 if (i != max_entries - 1) in test_array_map()
|
D | tailcalls.c | 40 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_1() 56 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_1() 72 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_1() 93 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_1() 94 j = bpf_map__def(prog_array)->max_entries - 1 - i; in test_tailcall_1() 110 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_1() 111 j = bpf_map__def(prog_array)->max_entries - 1 - i; in test_tailcall_1() 128 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_1() 178 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_2() 356 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_4() [all …]
|
/external/bcc/src/cc/includes/ |
D | table_desc.h | 52 max_entries(that.max_entries), in TableDesc() 69 max_entries(0), in TableDesc() 74 size_t leaf_size, size_t max_entries, int flags) in TableDesc() argument 80 max_entries(max_entries), in TableDesc() 97 size_t max_entries; variable
|
/external/bcc/src/cc/ |
D | table_desc.h | 52 max_entries(that.max_entries), in TableDesc() 69 max_entries(0), in TableDesc() 74 size_t leaf_size, size_t max_entries, int flags) in TableDesc() argument 80 max_entries(max_entries), in TableDesc() 97 size_t max_entries; variable
|
/external/cronet/net/disk_cache/blockfile/ |
D | block_files.cc | 76 for (int i = 0; i < header_->max_entries / 32; i++, current++) { in CreateMapBlock() 77 if (current == header_->max_entries / 32) in CreateMapBlock() 174 for (int i = 0; i < header_->max_entries / 32; i++) { in FixAllocationCounters() 228 return header_->max_entries; in Capacity() 232 if (header_->max_entries < 0 || header_->max_entries > kMaxBlocks || in ValidateCounters() 237 if (empty_blocks + header_->num_entries > header_->max_entries) in ValidateCounters() 474 header->max_entries * header->entry_size + kBlockHeaderSize) { in OpenBlockFile() 492 if (kMaxBlocks == header->max_entries) in GrowBlockFile() 497 int new_size = header->max_entries + 1024; in GrowBlockFile() 511 return (header->max_entries >= new_size); in GrowBlockFile() [all …]
|
/external/mesa3d/src/intel/dev/ |
D | gen_device_info.c | 166 .max_entries = { 195 .max_entries = { 232 .max_entries = { 261 .max_entries = { 290 .max_entries = { 324 .max_entries = { 352 .max_entries = { 380 .max_entries = { 427 .max_entries = { 451 .max_entries = { [all …]
|
/external/mesa3d/src/mesa/drivers/dri/i965/ |
D | gen6_urb.c | 67 if (nr_vs_entries > devinfo->urb.max_entries[MESA_SHADER_VERTEX]) in gen6_upload_urb() 68 nr_vs_entries = devinfo->urb.max_entries[MESA_SHADER_VERTEX]; in gen6_upload_urb() 70 if (nr_gs_entries > devinfo->urb.max_entries[MESA_SHADER_GEOMETRY]) in gen6_upload_urb() 71 nr_gs_entries = devinfo->urb.max_entries[MESA_SHADER_GEOMETRY]; in gen6_upload_urb()
|
/external/mesa3d/src/util/ |
D | hash_table.c | 84 uint32_t max_entries, size, rehash; member 87 #define ENTRY(max_entries, size, rehash) \ argument 88 { max_entries, size, rehash, \ 160 ht->max_entries = hash_sizes[ht->size_index].max_entries; in _mesa_hash_table_init() 381 ht->max_entries = hash_sizes[ht->size_index].max_entries; in _mesa_hash_table_rehash() 402 if (ht->entries >= ht->max_entries) { in hash_table_insert() 404 } else if (ht->deleted_entries + ht->entries >= ht->max_entries) { in hash_table_insert() 667 if (size < ht->max_entries) in _mesa_hash_table_reserve() 670 if (hash_sizes[i].max_entries >= size) { in _mesa_hash_table_reserve() 675 return ht->max_entries >= size; in _mesa_hash_table_reserve()
|
/external/virglrenderer/src/mesa/util/ |
D | hash_table.c | 84 uint32_t max_entries, size, rehash; member 87 #define ENTRY(max_entries, size, rehash) \ argument 88 { max_entries, size, rehash, \ 160 ht->max_entries = hash_sizes[ht->size_index].max_entries; in _mesa_hash_table_init() 394 if (ht->size_index == new_size_index && ht->deleted_entries == ht->max_entries) { in _mesa_hash_table_rehash() 416 ht->max_entries = hash_sizes[ht->size_index].max_entries; in _mesa_hash_table_rehash() 437 if (ht->entries >= ht->max_entries) { in hash_table_insert() 439 } else if (ht->deleted_entries + ht->entries >= ht->max_entries) { in hash_table_insert() 729 if (size < ht->max_entries) in _mesa_hash_table_reserve() 732 if (hash_sizes[i].max_entries >= size) { in _mesa_hash_table_reserve() [all …]
|