/art/runtime/ |
D | mapping_table.h | 69 table_(table), element_(element), end_(table_->DexToPcSize()), encoded_table_ptr_(nullptr), in DexToPcIterator() 72 if (end_ > 0) { in DexToPcIterator() 90 if (element_ != end_) { // Avoid reading beyond the end of the table. 108 const uint32_t end_; // Equal to table_->DexToPcSize(). variable 146 table_(table), element_(element), end_(table_->PcToDexSize()), encoded_table_ptr_(nullptr), in PcToDexIterator() 149 if (end_ > 0) { in PcToDexIterator() 167 if (element_ != end_) { // Avoid reading beyond the end of the table. 185 const uint32_t end_; // Equal to table_->PcToDexSize(). variable
|
D | oat_file.cc | 52 oat_file->end_ = elf_file->Begin() + hdr->sh_size + hdr->sh_offset; in OpenWithElfFile() 63 oat_file->end_ = &oat_contents[oat_contents.size()]; in OpenMemory() 141 : location_(location), begin_(NULL), end_(NULL), is_executable_(is_executable), in OatFile() 180 end_ = reinterpret_cast<byte*>(dlsym(dlopen_handle_, "oatlastword")); in Dlopen() 181 if (end_ == NULL) { in Dlopen() 187 end_ += sizeof(uint32_t); in Dlopen() 215 end_ = elf_file_->FindDynamicSymbolAddress("oatlastword"); in ElfFileOpen() 216 if (end_ == NULL) { in ElfFileOpen() 221 end_ += sizeof(uint32_t); in ElfFileOpen() 357 CHECK(end_ != NULL); in End() [all …]
|
D | oat_file.h | 325 const byte* end_; variable
|
/art/runtime/jdwp/ |
D | jdwp_request.cc | 30 end_ = bytes + byte_count_; in Request() 47 if (p_ < end_) { in CheckConsumed() 48 CHECK(p_ == end_) << "read too few bytes: " << (end_ - p_); in CheckConsumed() 49 } else if (p_ > end_) { in CheckConsumed() 50 CHECK(p_ == end_) << "read too many bytes: " << (p_ - end_); in CheckConsumed()
|
D | jdwp.h | 445 size_t size() { return end_ - p_; } in size() 466 const uint8_t* end_; variable
|
/art/runtime/gc/collector/ |
D | immune_region.h | 56 end_ = end; in SetEnd() 65 size_ = reinterpret_cast<uintptr_t>(end_) - reinterpret_cast<uintptr_t>(begin_); in UpdateSize() 69 mirror::Object* end_; variable
|
D | immune_region.cc | 49 } else if (space_begin >= end_) { // Space is after the immune region. in AddContinuousSpace() 61 end_ >= reinterpret_cast<mirror::Object*>(space->Limit()); in ContainsSpace()
|
D | mark_sweep.cc | 696 end_(end), in CardScanTask() 703 byte* const end_; member in art::gc::collector::CardScanTask 713 size_t cards_scanned = card_table->Scan(bitmap_, begin_, end_, visitor, minimum_age_); in Run() 715 << reinterpret_cast<void*>(end_) << " = " << cards_scanned; in Run() 827 end_(end) { in RecursiveMarkTask() 833 const uintptr_t end_; member in art::gc::collector::RecursiveMarkTask 842 bitmap_->VisitMarkedRange(begin_, end_, visitor); in Run()
|
/art/runtime/gc/space/ |
D | bump_pointer_space-inl.h | 44 byte* end = end_.LoadRelaxed(); in AllocThreadUnsafe() 49 end_.StoreRelaxed(end + num_bytes); in AllocThreadUnsafe() 65 old_end = end_.LoadRelaxed(); in AllocNonvirtualWithoutAccounting() 71 } while (!end_.CompareExchangeWeakSequentiallyConsistent(old_end, new_end)); in AllocNonvirtualWithoutAccounting()
|
D | large_object_space.cc | 90 total_objects_allocated_(0), begin_(begin), end_(end) { in LargeObjectSpace() 127 if (end_ == nullptr || obj_end > end_) { in Alloc() 128 end_ = obj_end; in Alloc() 315 const uintptr_t free_end_start = reinterpret_cast<uintptr_t>(end_) - free_end_; in Walk() 353 uintptr_t free_end_start = reinterpret_cast<uintptr_t>(end_) - free_end_; in Free() 469 uintptr_t free_end_start = reinterpret_cast<uintptr_t>(end_) - free_end_; in Dump()
|
D | space.h | 255 return end_.LoadRelaxed(); in End() 266 end_.StoreRelaxed(end); in SetEnd() 306 Space(name, gc_retention_policy), begin_(begin), end_(end), limit_(limit) { in ContinuousSpace() 313 Atomic<byte*> end_; variable
|
D | large_object_space.h | 79 return end_; in End() 104 byte* end_; variable
|
/art/runtime/gc/accounting/ |
D | space_bitmap_test.cc | 46 end_(end) {} in BitmapVerify() 50 EXPECT_TRUE(obj <= end_); in operator ()() 56 const mirror::Object* end_; member in art::gc::accounting::BitmapVerify
|
/art/compiler/utils/ |
D | arena_allocator.cc | 205 end_(nullptr), in ArenaAllocator() 221 if (UNLIKELY(ptr_ + rounded_bytes > end_)) { in AllocValgrind() 252 end_ = new_arena->End(); in ObtainNewArenaForAllocation() 271 (arena_head_ == nullptr) ? 0 : (end_ - ptr_) - arena_head_->RemainingSpace(); in GetMemStats()
|
D | arena_allocator.h | 168 if (UNLIKELY(ptr_ + bytes > end_)) { in Alloc() 195 uint8_t* end_; variable
|
/art/compiler/optimizing/ |
D | ssa_liveness_analysis.h | 55 LiveRange(size_t start, size_t end, LiveRange* next) : start_(start), end_(end), next_(next) { in LiveRange() 61 size_t GetEnd() const { return end_; } in GetEnd() 65 return (start_ >= other.start_ && start_ < other.end_) in IntersectsWith() 66 || (other.start_ >= start_ && other.start_ < end_); in IntersectsWith() 70 return end_ <= other.start_; in IsBefore() 74 stream << "[" << start_ << ", " << end_ << ")"; in Dump() 79 size_t end_; variable
|
/art/compiler/driver/ |
D | compiler_driver.cc | 1428 end_(end), in ForAllClosure() 1434 if (UNLIKELY(index >= end_)) { in Run() 1448 const size_t end_; member in art::ParallelCompilationManager::ForAllClosure
|