/external/perfetto/src/tracing/core/ |
D | shared_memory_abi_unittest.cc | 88 for (size_t chunk_idx = 0; chunk_idx < num_chunks; chunk_idx++) { in TEST_P() local 89 Chunk& chunk = chunks[chunk_idx]; in TEST_P() 93 abi.GetChunkState(page_idx, chunk_idx)); in TEST_P() 100 uint16_t packets_count = static_cast<uint16_t>(chunk_idx * 10); in TEST_P() 102 uint8_t flags = static_cast<uint8_t>((0xffu - chunk_idx) & kFlagsMask); in TEST_P() 105 chunk = abi.TryAcquireChunkForWriting(page_idx, chunk_idx, &header); in TEST_P() 108 abi.GetChunkState(page_idx, chunk_idx)); in TEST_P() 146 ASSERT_FALSE(abi.TryAcquireChunkForWriting(page_idx, chunk_idx, &header) in TEST_P() 151 for (size_t chunk_idx = 0; chunk_idx < num_chunks; chunk_idx++) { in TEST_P() local 152 Chunk& chunk = chunks[chunk_idx]; in TEST_P() [all …]
|
D | shared_memory_abi.cc | 144 size_t chunk_idx) { in GetChunkUnchecked() argument 146 PERFETTO_DCHECK(chunk_idx < num_chunks); in GetChunkUnchecked() 149 size_t chunk_offset_in_page = sizeof(PageHeader) + chunk_idx * chunk_size; in GetChunkUnchecked() 152 static_cast<uint8_t>(chunk_idx)); in GetChunkUnchecked() 159 size_t chunk_idx, in TryAcquireChunk() argument 170 if (chunk_idx >= num_chunks) in TryAcquireChunk() 179 auto cur_chunk_state = (layout >> (chunk_idx * kChunkShift)) & kChunkMask; in TryAcquireChunk() 184 next_layout &= ~(kChunkMask << (chunk_idx * kChunkShift)); in TryAcquireChunk() 185 next_layout |= (desired_chunk_state << (chunk_idx * kChunkShift)); in TryAcquireChunk() 189 Chunk chunk = GetChunkUnchecked(page_idx, layout, chunk_idx); in TryAcquireChunk() [all …]
|
D | trace_writer_impl_unittest.cc | 84 for (size_t chunk_idx = 0; chunk_idx < num_chunks; chunk_idx++) { in TEST_P() local 85 auto chunk_state = abi->GetChunkState(page_idx, chunk_idx); in TEST_P() 88 auto chunk = abi->TryAcquireChunkForReading(page_idx, chunk_idx); in TEST_P()
|
D | shared_memory_arbiter_impl.cc | 97 for (uint32_t chunk_idx = 0; free_chunks; in GetNewChunk() local 98 chunk_idx++, free_chunks >>= 1) { in GetNewChunk() 103 page_idx_, chunk_idx, &header); in GetNewChunk() 187 uint8_t chunk_idx = chunk.chunk_idx(); in UpdateCommitDataRequest() local 196 ctm->set_chunk(chunk_idx); in UpdateCommitDataRequest()
|
D | startup_trace_writer_unittest.cc | 81 for (size_t chunk_idx = 0; chunk_idx < num_chunks; chunk_idx++) { in VerifyPackets() local 82 auto chunk_state = abi->GetChunkState(page_idx, chunk_idx); in VerifyPackets() 85 auto chunk = abi->TryAcquireChunkForReading(page_idx, chunk_idx); in VerifyPackets()
|
D | tracing_service_impl.cc | 1221 for (uint32_t chunk_idx = 0; used_chunks; chunk_idx++, used_chunks >>= 1) { in ScrapeSharedMemoryBuffers() local 1226 SharedMemoryABI::GetChunkStateFromLayout(layout, chunk_idx); in ScrapeSharedMemoryBuffers() 1232 abi->GetChunkUnchecked(page_idx, layout, chunk_idx); in ScrapeSharedMemoryBuffers()
|
/external/perfetto/include/perfetto/tracing/core/ |
D | shared_memory_abi.h | 363 uint8_t chunk_idx() const { return chunk_idx_; } in chunk_idx() function 419 Chunk(uint8_t* begin, uint16_t size, uint8_t chunk_idx); 500 size_t chunk_idx, in TryAcquireChunkForWriting() argument 502 return TryAcquireChunk(page_idx, chunk_idx, kChunkBeingWritten, header); in TryAcquireChunkForWriting() 507 Chunk TryAcquireChunkForReading(size_t page_idx, size_t chunk_idx) { in TryAcquireChunkForReading() argument 508 return TryAcquireChunk(page_idx, chunk_idx, kChunkBeingRead, nullptr); in TryAcquireChunkForReading() 514 size_t chunk_idx); 526 ChunkState GetChunkState(size_t page_idx, size_t chunk_idx) { in GetChunkState() argument 529 return GetChunkStateFromLayout(layout, chunk_idx); in GetChunkState() 539 size_t chunk_idx) { in GetChunkStateFromLayout() argument [all …]
|
/external/tensorflow/tensorflow/core/common_runtime/ |
D | ring_alg.cc | 309 void RingAlg::InitRingField(RingField* rf, int chunk_idx, int subdiv_idx, in InitRingField() argument 318 DCHECK_EQ(field_idx, (chunk_idx * num_subdivs_) + subdiv_idx); in InitRingField() 319 rf->chunk_idx = chunk_idx; in InitRingField() 336 rf->do_recv = (rf->chunk_idx != rf->rank); in InitRingField() 339 (rf->rank != ((rf->chunk_idx + (group_size_ - 1)) % group_size_)); in InitRingField() 342 (rf->rank == ((rf->chunk_idx + (group_size_ - 1)) % group_size_)); in InitRingField() 360 (rf->rank != ((rf->chunk_idx + (group_size_ - 1)) % group_size_)); in AdvanceToSecondPass() 362 (rf->rank != ((rf->chunk_idx + (group_size_ - 2)) % group_size_)); in AdvanceToSecondPass() 365 (rf->rank == ((rf->chunk_idx + (group_size_ - 2)) % group_size_)); in AdvanceToSecondPass() 370 string rv = strings::StrCat("RingField rank=", rank, " chunk_idx=", chunk_idx, in DebugString()
|
D | ring_reducer.cc | 145 void RingReducer::InitRingField(RingField* rf, int chunk_idx, int subdiv_idx, in InitRingField() argument 147 RingAlg::InitRingField(rf, chunk_idx, subdiv_idx, field_idx); in InitRingField() 164 for (int chunk_idx = 0; chunk_idx < group_size_; ++chunk_idx) { in RunAsyncParts() local 166 int rf_index = (chunk_idx * num_subdivs_) + subdiv_idx; in RunAsyncParts() 167 InitRingField(&rfv_[rf_index], chunk_idx, subdiv_idx, rf_index); in RunAsyncParts()
|
D | ring_gatherer.cc | 128 for (int chunk_idx = 0; chunk_idx < group_size_; ++chunk_idx) { in RunAsyncParts() local 130 int rf_index = (chunk_idx * num_subdivs_) + subdiv_idx; in RunAsyncParts() 131 InitRingField(&rfv_[rf_index], chunk_idx, subdiv_idx, rf_index); in RunAsyncParts()
|
D | ring_alg.h | 68 int16 chunk_idx; // major division index member 85 virtual void InitRingField(RingField* rf, int chunk_idx, int subdiv_idx,
|
D | ring_reducer.h | 45 void InitRingField(RingField* rf, int chunk_idx, int subdiv_idx,
|
/external/compiler-rt/lib/sanitizer_common/ |
D | sanitizer_allocator.h | 389 uptr chunk_idx = GetChunkIdx((uptr)p, size); in GetBlockBegin() local 391 uptr beg = chunk_idx * size; in GetBlockBegin() 410 uptr chunk_idx = GetChunkIdx(reinterpret_cast<uptr>(p), size); in GetMetaData() local 413 (1 + chunk_idx) * kMetadataSize); in GetMetaData() 1079 h->chunk_idx = idx; in Allocate() 1105 uptr idx = h->chunk_idx; in Deallocate() 1109 chunks_[idx]->chunk_idx = idx; in Deallocate() 1126 CHECK_EQ(h->chunk_idx, i); in TotalMemoryUsed() 1183 chunks_[i]->chunk_idx = i; in GetBlockBeginFastLocked() 1251 uptr chunk_idx; member
|
/external/avb/ |
D | avbtool | 956 chunk_idx = bisect.bisect_right(self._chunk_output_offsets, 961 chunk = self._chunks[chunk_idx] 978 chunk_idx += 1 980 if chunk_idx >= len(self._chunks): 1015 chunk_idx = bisect.bisect_right(self._chunk_output_offsets, size) - 1 1016 chunk = self._chunks[chunk_idx] 1020 chunk_idx_for_update = chunk_idx + 1 1047 chunk_idx_for_update = chunk_idx
|