/external/pigweed/pw_kvs/ |
D | key_value_store_binary_format_test.cc | 373 EXPECT_EQ(stats.in_use_bytes, (32u * kvs_.redundancy())); in TEST_F() 375 EXPECT_EQ(stats.writable_bytes, 512u * 2 - (32 * kvs_.redundancy())); in TEST_F() 596 EXPECT_EQ(stats.in_use_bytes, (32u * kvs_.redundancy())); in TEST_F() 598 EXPECT_EQ(stats.writable_bytes, 512u * 2 - (32 * kvs_.redundancy())); in TEST_F() 678 EXPECT_EQ(stats.in_use_bytes, (192u * kvs_.redundancy())); in TEST_F() 680 EXPECT_EQ(stats.writable_bytes, 512u * 3 - (192 * kvs_.redundancy())); in TEST_F() 695 EXPECT_EQ(stats.in_use_bytes, (192u * kvs_.redundancy())); in TEST_F() 697 EXPECT_EQ(stats.writable_bytes, 512u * 2 - (192 * (kvs_.redundancy() - 1))); in TEST_F() 703 EXPECT_EQ(stats.in_use_bytes, (192u * kvs_.redundancy())); in TEST_F() 705 EXPECT_EQ(stats.writable_bytes, 512u * 3 - (192 * kvs_.redundancy())); in TEST_F() [all …]
|
D | key_value_store_map_test.cc | 55 size_t redundancy; member 167 label << ((kvs_.redundancy() > 1) ? "Redundant" : ""); in Test_RandomValidInputs() 399 (kParams.sector_count * kParams.redundancy)> 404 KeyValueStoreBuffer<kMaxEntries, kMaxUsableSectors, kParams.redundancy> kvs_; 412 (kParams.sector_count * kParams.redundancy)> 415 (kParams.sector_count * kParams.redundancy)>( 466 .redundancy = 1, 475 .redundancy = 2, 484 .redundancy = 1, 493 .redundancy = 2, [all …]
|
D | key_value_store.cc | 44 size_t redundancy, in KeyValueStore() argument 52 entry_cache_(key_descriptor_list, addresses, redundancy), in KeyValueStore() 233 if (metadata.addresses().size() < redundancy()) { in InitializeMetadata() 237 unsigned(redundancy())); in InitializeMetadata() 284 unsigned(redundancy())); in InitializeMetadata() 341 if (redundancy() > 1) { in CheckForErrors() 343 if (metadata.addresses().size() < redundancy()) { in CheckForErrors() 683 for (size_t i = 1; i < redundancy(); ++i) { in WriteEntry() 721 for (size_t i = 0; i < redundancy(); i++) { in GetAddressesForWrite() 1053 for (size_t i = 1; i < redundancy(); ++i) { in UpdateEntriesToPrimaryFormat() [all …]
|
D | entry_cache.cc | 208 for (size_t i = 0; i < redundancy(); ++i) { in AddAddressIfRoom() 221 while (size < redundancy() && addresses[size] != kNoAddress) { in addresses()
|
D | docs.rst | 106 The flash storage used by KVS is multiplied by `redundancy`_ used. A redundancy 133 KVS supports storing redundant copies of KV entries. For a given redundancy 138 Redundancy increases flash usage proportional to the redundancy level. The RAM 139 usage for KVS internal state has a small increase with redundancy.
|
/external/libopus/src/ |
D | opus_decoder.c | 243 int redundancy=0; in opus_decode_frame() local 426 redundancy = ec_dec_bit_logp(&dec, 12); in opus_decode_frame() 428 redundancy = 1; in opus_decode_frame() 429 if (redundancy) in opus_decode_frame() 444 redundancy = 0; in opus_decode_frame() 453 if (redundancy) in opus_decode_frame() 496 redundant_audio_size = redundancy ? F5*st->channels : ALLOC_NONE; in opus_decode_frame() 500 if (redundancy && celt_to_silk) in opus_decode_frame() 529 if (st->prev_mode == MODE_HYBRID && !(redundancy && celt_to_silk && st->prev_redundancy) ) in opus_decode_frame() 554 if (redundancy && !celt_to_silk) in opus_decode_frame() [all …]
|
D | opus_encoder.c | 1061 int redundancy = 0; in opus_encode_native() local 1414 redundancy = 1; in opus_encode_native() 1424 redundancy=0; in opus_encode_native() 1601 redundancy = 1; in opus_encode_native() 1611 redundancy = 0; in opus_encode_native() 1613 if (redundancy) in opus_encode_native() 1617 redundancy = 0; in opus_encode_native() 1777 if (redundancy && redundancy_bytes >= 2) in opus_encode_native() 1868 redundancy = (redundancy_bytes != 0); in opus_encode_native() 1977 ec_enc_bit_logp(&enc, redundancy, 12); in opus_encode_native() [all …]
|
/external/pigweed/pw_kvs/public/pw_kvs/internal/ |
D | entry_cache.h | 149 size_t redundancy) in EntryCache() argument 152 redundancy_(redundancy) {} in EntryCache() 194 size_t redundancy() const { return redundancy_; } in redundancy() function
|
/external/scapy/scapy/contrib/ |
D | pnio_rtc.uts | 73 raw(PNIORealTime(len=20, dataLen=15, cycleCounter=0x1234, dataStatus='redundancy+validData+no_probl… 83 p == PNIORealTime(len=20, dataLen=15, cycleCounter=0x1234, dataStatus='redundancy+validData+no_prob… 99 …len=20, dataLen=15, cycleCounter=0x1234, dataStatus='redundancy+validData+no_problem', transferSta…
|
/external/lz4/examples/ |
D | streaming_api_basics.md | 23 In other words, LZ4 library find redundancy from single contiguous memory block. 25 So LZ4 library could find more redundancy than Block API.
|
/external/ltp/testcases/open_posix_testsuite/conformance/interfaces/mq_timedsend/ |
D | coverage.txt | 5 I could find to reduce this redundancy.
|
/external/pigweed/pw_kvs/public/pw_kvs/ |
D | key_value_store.h | 323 size_t redundancy() const { return entry_cache_.redundancy(); } in redundancy() function 354 size_t redundancy,
|
/external/llvm/test/Transforms/GVN/ |
D | non-local-offset.ll | 31 ; redundancy.
|
D | invariant-load.ll | 31 ; be a case for PRE becomes a full redundancy.
|
D | volatile.ll | 76 ; Does cross block redundancy elimination work with volatiles?
|
/external/llvm-project/llvm/test/Transforms/GVN/ |
D | non-local-offset.ll | 31 ; redundancy.
|
/external/llvm-project/llvm/test/Transforms/NewGVN/ |
D | non-local-offset.ll | 31 ; redundancy.
|
/external/llvm-project/llvm/test/Transforms/LoopDistribute/ |
D | bounds-expansion-bug.ll | 13 ; redundancy. However, due to logic in SCEVExpander::ReuseOrCreateCast, we
|
/external/llvm/test/Transforms/LoopDistribute/ |
D | bounds-expansion-bug.ll | 13 ; redundancy. However, due to logic in SCEVExpander::ReuseOrCreateCast, we
|
/external/llvm-project/llvm/test/Transforms/GVN/PRE/ |
D | invariant-load.ll | 31 ; be a case for PRE becomes a full redundancy.
|
/external/llvm-project/llvm/test/Transforms/MemCpyOpt/ |
D | nonlocal-memcpy-memcpy.ll | 21 ; A simple partial redundancy. Test that the second memcpy is optimized 63 ; A CFG triangle with a partial redundancy targeting an alloca. Test that the
|
/external/okio/docs/ |
D | code_of_conduct.md | 26 * **Be collaborative**: Collaboration reduces redundancy and improves the quality of our work. We
|
/external/libabigail/ |
D | COMMIT-LOG-GUIDELINES | 6 that redundancy does help in understanding the commit diff better. If
|
/external/llvm/docs/ |
D | Lexicon.rst | 199 Rearranging associative expressions to promote better redundancy elimination
|
/external/mesa3d/docs/relnotes/ |
D | 13.0.4.rst | 209 - glsl: fix opt_minmax redundancy checks against baserange
|