/external/gwp_asan/gwp_asan/tests/ |
D | slot_reuse.cpp | 59 constexpr unsigned kPoolSize = 2; in TEST_F() local 60 InitNumSlots(kPoolSize); in TEST_F() 61 runNoReuseBeforeNecessary(&GPA, kPoolSize); in TEST_F() 65 constexpr unsigned kPoolSize = 128; in TEST_F() local 66 InitNumSlots(kPoolSize); in TEST_F() 67 runNoReuseBeforeNecessary(&GPA, kPoolSize); in TEST_F() 71 constexpr unsigned kPoolSize = 129; in TEST_F() local 72 InitNumSlots(kPoolSize); in TEST_F() 73 runNoReuseBeforeNecessary(&GPA, kPoolSize); in TEST_F()
|
/external/llvm-project/compiler-rt/lib/gwp_asan/tests/ |
D | slot_reuse.cpp | 59 constexpr unsigned kPoolSize = 2; in TEST_F() local 60 InitNumSlots(kPoolSize); in TEST_F() 61 runNoReuseBeforeNecessary(&GPA, kPoolSize); in TEST_F() 65 constexpr unsigned kPoolSize = 128; in TEST_F() local 66 InitNumSlots(kPoolSize); in TEST_F() 67 runNoReuseBeforeNecessary(&GPA, kPoolSize); in TEST_F() 71 constexpr unsigned kPoolSize = 129; in TEST_F() local 72 InitNumSlots(kPoolSize); in TEST_F() 73 runNoReuseBeforeNecessary(&GPA, kPoolSize); in TEST_F()
|
/external/perfetto/src/trace_processor/containers/ |
D | bit_vector_benchmark.cc | 94 static constexpr uint32_t kPoolSize = 1024 * 1024; in BM_BitVectorSet() local 95 std::vector<bool> bit_pool(kPoolSize); in BM_BitVectorSet() 96 std::vector<uint32_t> row_pool(kPoolSize); in BM_BitVectorSet() 97 for (uint32_t i = 0; i < kPoolSize; ++i) { in BM_BitVectorSet() 105 pool_idx = (pool_idx + 1) % kPoolSize; in BM_BitVectorSet() 120 static constexpr uint32_t kPoolSize = 1024 * 1024; in BM_BitVectorClear() local 121 std::vector<uint32_t> row_pool(kPoolSize); in BM_BitVectorClear() 122 for (uint32_t i = 0; i < kPoolSize; ++i) { in BM_BitVectorClear() 129 pool_idx = (pool_idx + 1) % kPoolSize; in BM_BitVectorClear() 143 static constexpr uint32_t kPoolSize = 1024 * 1024; in BM_BitVectorIndexOfNthSet() local [all …]
|
D | nullable_vector_benchmark.cc | 23 static constexpr uint32_t kPoolSize = 100000; variable 29 std::vector<uint8_t> data_pool(kPoolSize); in BM_NullableVectorAppendNonNull() 33 for (uint32_t i = 0; i < kPoolSize; ++i) { in BM_NullableVectorAppendNonNull() 41 pool_idx = (pool_idx + 1) % kPoolSize; in BM_NullableVectorAppendNonNull() 48 std::vector<uint32_t> idx_pool(kPoolSize); in BM_NullableVectorGetNonNull() 56 for (uint32_t i = 0; i < kPoolSize; ++i) { in BM_NullableVectorGetNonNull() 63 pool_idx = (pool_idx + 1) % kPoolSize; in BM_NullableVectorGetNonNull()
|
D | row_map_benchmark.cc | 26 static constexpr uint32_t kPoolSize = 100000; variable 63 auto pool_vec = CreateIndexVector(kPoolSize, rm.size()); in BenchRowMapGet() 68 pool_idx = (pool_idx + 1) % kPoolSize; in BenchRowMapGet() 74 auto pool_vec = CreateIndexVector(kPoolSize, kSize); in BenchRowMapInsertIntoEmpty() 81 pool_idx = (pool_idx + 1) % kPoolSize; in BenchRowMapInsertIntoEmpty() 99 auto pool_vec = CreateIndexVector(kPoolSize, kSize); in BenchRowMapFilterInto() 111 pool_idx = (pool_idx + 1) % kPoolSize; in BenchRowMapFilterInto()
|
/external/rust/crates/grpcio-sys/grpc/third_party/abseil-cpp/absl/random/internal/ |
D | pool_urbg.cc | 134 static constexpr int kPoolSize = 8; variable 138 ABSL_CACHELINE_ALIGNED static RandenPoolEntry* shared_pools[kPoolSize]; 151 static_assert(kPoolSize >= 1, in GetPoolID() 159 my_pool_id = (sequence++ % kPoolSize); in GetPoolID() 178 my_pool_id = (sequence++ % kPoolSize) + 1; in GetPoolID() 214 uint32_t seed_material[kPoolSize * kSeedSize]; in InitPoolURBG() 219 for (int i = 0; i < kPoolSize; i++) { in InitPoolURBG()
|
/external/angle/third_party/abseil-cpp/absl/random/internal/ |
D | pool_urbg.cc | 134 static constexpr int kPoolSize = 8; variable 138 ABSL_CACHELINE_ALIGNED static RandenPoolEntry* shared_pools[kPoolSize]; 151 static_assert(kPoolSize >= 1, in GetPoolID() 159 my_pool_id = (sequence++ % kPoolSize); in GetPoolID() 178 my_pool_id = (sequence++ % kPoolSize) + 1; in GetPoolID() 213 uint32_t seed_material[kPoolSize * kSeedSize]; in InitPoolURBG() 218 for (int i = 0; i < kPoolSize; i++) { in InitPoolURBG()
|
/external/libtextclassifier/abseil-cpp/absl/random/internal/ |
D | pool_urbg.cc | 134 static constexpr int kPoolSize = 8; variable 138 ABSL_CACHELINE_ALIGNED static RandenPoolEntry* shared_pools[kPoolSize]; 151 static_assert(kPoolSize >= 1, in GetPoolID() 159 my_pool_id = (sequence++ % kPoolSize); in GetPoolID() 178 my_pool_id = (sequence++ % kPoolSize) + 1; in GetPoolID() 214 uint32_t seed_material[kPoolSize * kSeedSize]; in InitPoolURBG() 219 for (int i = 0; i < kPoolSize; i++) { in InitPoolURBG()
|
/external/abseil-cpp/absl/random/internal/ |
D | pool_urbg.cc | 134 static constexpr int kPoolSize = 8; variable 138 ABSL_CACHELINE_ALIGNED static RandenPoolEntry* shared_pools[kPoolSize]; 151 static_assert(kPoolSize >= 1, in GetPoolID() 159 my_pool_id = (sequence++ % kPoolSize); in GetPoolID() 178 my_pool_id = (sequence++ % kPoolSize) + 1; in GetPoolID() 214 uint32_t seed_material[kPoolSize * kSeedSize]; in InitPoolURBG() 219 for (int i = 0; i < kPoolSize; i++) { in InitPoolURBG()
|
/external/webrtc/third_party/abseil-cpp/absl/random/internal/ |
D | pool_urbg.cc | 134 static constexpr int kPoolSize = 8; variable 138 ABSL_CACHELINE_ALIGNED static RandenPoolEntry* shared_pools[kPoolSize]; 151 static_assert(kPoolSize >= 1, in GetPoolID() 159 my_pool_id = (sequence++ % kPoolSize); in GetPoolID() 178 my_pool_id = (sequence++ % kPoolSize) + 1; in GetPoolID() 214 uint32_t seed_material[kPoolSize * kSeedSize]; in InitPoolURBG() 219 for (int i = 0; i < kPoolSize; i++) { in InitPoolURBG()
|
/external/openscreen/third_party/abseil/src/absl/random/internal/ |
D | pool_urbg.cc | 134 static constexpr int kPoolSize = 8; variable 138 ABSL_CACHELINE_ALIGNED static RandenPoolEntry* shared_pools[kPoolSize]; 151 static_assert(kPoolSize >= 1, in GetPoolID() 159 my_pool_id = (sequence++ % kPoolSize); in GetPoolID() 178 my_pool_id = (sequence++ % kPoolSize) + 1; in GetPoolID() 214 uint32_t seed_material[kPoolSize * kSeedSize]; in InitPoolURBG() 219 for (int i = 0; i < kPoolSize; i++) { in InitPoolURBG()
|
/external/vixl/test/ |
D | test-pool-manager.cc | 712 const int kPoolSize = kVeneerSize * kNumBranches + kHeaderSize; in TEST() local 720 pc + kPoolSize + kBranchSize - 1); in TEST() 731 pc + kPoolSize + kBranchSize); in TEST() 741 pc + kPoolSize + kBranchSize, in TEST() 751 VIXL_ASSERT(new_pc == pc + kPoolSize); in TEST() 810 const int kPoolSize = kVeneerSize * kNumBranches + kHeaderSize; in TEST() local 813 ForwardReference<int32_t> temp_ref(pc, kBranchSize, pc, pc + kPoolSize); in TEST() 821 ForwardReference<int32_t> temp_ref(pc, kBranchSize, pc, pc + kPoolSize); in TEST() 835 VIXL_ASSERT(new_pc == pc + kPoolSize); in TEST()
|