// Copyright 2019 the V8 project authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #include "src/objects/code.h" #include #include "src/codegen/assembler-inl.h" #include "src/codegen/cpu-features.h" #include "src/codegen/reloc-info.h" #include "src/codegen/safepoint-table.h" #include "src/codegen/source-position.h" #include "src/deoptimizer/deoptimizer.h" #include "src/execution/isolate-utils-inl.h" #include "src/interpreter/bytecode-array-iterator.h" #include "src/interpreter/bytecode-decoder.h" #include "src/interpreter/interpreter.h" #include "src/objects/allocation-site-inl.h" #include "src/objects/code-kind.h" #include "src/objects/fixed-array.h" #include "src/roots/roots-inl.h" #include "src/snapshot/embedded/embedded-data-inl.h" #include "src/utils/ostreams.h" #ifdef ENABLE_DISASSEMBLER #include "src/codegen/code-comments.h" #include "src/diagnostics/disasm.h" #include "src/diagnostics/disassembler.h" #include "src/diagnostics/eh-frame.h" #endif namespace v8 { namespace internal { namespace { // Helper function for getting an EmbeddedData that can handle un-embedded // builtins when short builtin calls are enabled. inline EmbeddedData EmbeddedDataWithMaybeRemappedEmbeddedBuiltins( HeapObject code) { #if defined(V8_COMPRESS_POINTERS_IN_ISOLATE_CAGE) // GetIsolateFromWritableObject(*this) works for both read-only and writable // objects when pointer compression is enabled with a per-Isolate cage. return EmbeddedData::FromBlob(GetIsolateFromWritableObject(code)); #elif defined(V8_COMPRESS_POINTERS_IN_SHARED_CAGE) // When pointer compression is enabled with a shared cage, there is also a // shared CodeRange. When short builtin calls are enabled, there is a single // copy of the re-embedded builtins in the shared CodeRange, so use that if // it's present. if (FLAG_jitless) return EmbeddedData::FromBlob(); CodeRange* code_range = CodeRange::GetProcessWideCodeRange().get(); return (code_range && code_range->embedded_blob_code_copy() != nullptr) ? EmbeddedData::FromBlob(code_range) : EmbeddedData::FromBlob(); #else // Otherwise there is a single copy of the blob across all Isolates, use the // global atomic variables. return EmbeddedData::FromBlob(); #endif } } // namespace Address OffHeapInstructionStart(HeapObject code, Builtin builtin) { // TODO(11527): Here and below: pass Isolate as an argument for getting // the EmbeddedData. EmbeddedData d = EmbeddedDataWithMaybeRemappedEmbeddedBuiltins(code); return d.InstructionStartOfBuiltin(builtin); } Address OffHeapInstructionEnd(HeapObject code, Builtin builtin) { EmbeddedData d = EmbeddedDataWithMaybeRemappedEmbeddedBuiltins(code); return d.InstructionStartOfBuiltin(builtin) + d.InstructionSizeOfBuiltin(builtin); } int OffHeapInstructionSize(HeapObject code, Builtin builtin) { EmbeddedData d = EmbeddedDataWithMaybeRemappedEmbeddedBuiltins(code); return d.InstructionSizeOfBuiltin(builtin); } Address OffHeapMetadataStart(HeapObject code, Builtin builtin) { EmbeddedData d = EmbeddedDataWithMaybeRemappedEmbeddedBuiltins(code); return d.MetadataStartOfBuiltin(builtin); } Address OffHeapMetadataEnd(HeapObject code, Builtin builtin) { EmbeddedData d = EmbeddedDataWithMaybeRemappedEmbeddedBuiltins(code); return d.MetadataStartOfBuiltin(builtin) + d.MetadataSizeOfBuiltin(builtin); } int OffHeapMetadataSize(HeapObject code, Builtin builtin) { EmbeddedData d = EmbeddedDataWithMaybeRemappedEmbeddedBuiltins(code); return d.MetadataSizeOfBuiltin(builtin); } Address OffHeapSafepointTableAddress(HeapObject code, Builtin builtin) { EmbeddedData d = EmbeddedDataWithMaybeRemappedEmbeddedBuiltins(code); return d.SafepointTableStartOf(builtin); } int OffHeapSafepointTableSize(HeapObject code, Builtin builtin) { EmbeddedData d = EmbeddedDataWithMaybeRemappedEmbeddedBuiltins(code); return d.SafepointTableSizeOf(builtin); } Address OffHeapHandlerTableAddress(HeapObject code, Builtin builtin) { EmbeddedData d = EmbeddedDataWithMaybeRemappedEmbeddedBuiltins(code); return d.HandlerTableStartOf(builtin); } int OffHeapHandlerTableSize(HeapObject code, Builtin builtin) { EmbeddedData d = EmbeddedDataWithMaybeRemappedEmbeddedBuiltins(code); return d.HandlerTableSizeOf(builtin); } Address OffHeapConstantPoolAddress(HeapObject code, Builtin builtin) { EmbeddedData d = EmbeddedDataWithMaybeRemappedEmbeddedBuiltins(code); return d.ConstantPoolStartOf(builtin); } int OffHeapConstantPoolSize(HeapObject code, Builtin builtin) { EmbeddedData d = EmbeddedDataWithMaybeRemappedEmbeddedBuiltins(code); return d.ConstantPoolSizeOf(builtin); } Address OffHeapCodeCommentsAddress(HeapObject code, Builtin builtin) { EmbeddedData d = EmbeddedDataWithMaybeRemappedEmbeddedBuiltins(code); return d.CodeCommentsStartOf(builtin); } int OffHeapCodeCommentsSize(HeapObject code, Builtin builtin) { EmbeddedData d = EmbeddedDataWithMaybeRemappedEmbeddedBuiltins(code); return d.CodeCommentsSizeOf(builtin); } Address OffHeapUnwindingInfoAddress(HeapObject code, Builtin builtin) { EmbeddedData d = EmbeddedDataWithMaybeRemappedEmbeddedBuiltins(code); return d.UnwindingInfoStartOf(builtin); } int OffHeapUnwindingInfoSize(HeapObject code, Builtin builtin) { EmbeddedData d = EmbeddedDataWithMaybeRemappedEmbeddedBuiltins(code); return d.UnwindingInfoSizeOf(builtin); } void Code::ClearEmbeddedObjects(Heap* heap) { HeapObject undefined = ReadOnlyRoots(heap).undefined_value(); int mode_mask = RelocInfo::EmbeddedObjectModeMask(); for (RelocIterator it(*this, mode_mask); !it.done(); it.next()) { DCHECK(RelocInfo::IsEmbeddedObjectMode(it.rinfo()->rmode())); it.rinfo()->set_target_object(heap, undefined, SKIP_WRITE_BARRIER); } set_embedded_objects_cleared(true); } void Code::Relocate(intptr_t delta) { for (RelocIterator it(*this, RelocInfo::kApplyMask); !it.done(); it.next()) { it.rinfo()->apply(delta); } FlushICache(); } void Code::FlushICache() const { FlushInstructionCache(raw_instruction_start(), raw_instruction_size()); } void Code::CopyFromNoFlush(ByteArray reloc_info, Heap* heap, const CodeDesc& desc) { // Copy code. STATIC_ASSERT(kOnHeapBodyIsContiguous); CopyBytes(reinterpret_cast(raw_instruction_start()), desc.buffer, static_cast(desc.instr_size)); // TODO(jgruber,v8:11036): Merge with the above. CopyBytes(reinterpret_cast(raw_instruction_start() + desc.instr_size), desc.unwinding_info, static_cast(desc.unwinding_info_size)); // Copy reloc info. CopyRelocInfoToByteArray(reloc_info, desc); // Unbox handles and relocate. RelocateFromDesc(reloc_info, heap, desc); } void Code::RelocateFromDesc(ByteArray reloc_info, Heap* heap, const CodeDesc& desc) { // Unbox handles and relocate. Assembler* origin = desc.origin; const int mode_mask = RelocInfo::PostCodegenRelocationMask(); for (RelocIterator it(*this, reloc_info, mode_mask); !it.done(); it.next()) { RelocInfo::Mode mode = it.rinfo()->rmode(); if (RelocInfo::IsEmbeddedObjectMode(mode)) { Handle p = it.rinfo()->target_object_handle(origin); it.rinfo()->set_target_object(heap, *p, UPDATE_WRITE_BARRIER, SKIP_ICACHE_FLUSH); } else if (RelocInfo::IsCodeTargetMode(mode)) { // Rewrite code handles to direct pointers to the first instruction in the // code object. Handle p = it.rinfo()->target_object_handle(origin); DCHECK(p->IsCodeT(GetPtrComprCageBaseSlow(*p))); Code code = FromCodeT(CodeT::cast(*p)); it.rinfo()->set_target_address(code.raw_instruction_start(), UPDATE_WRITE_BARRIER, SKIP_ICACHE_FLUSH); } else if (RelocInfo::IsRuntimeEntry(mode)) { Address p = it.rinfo()->target_runtime_entry(origin); it.rinfo()->set_target_runtime_entry(p, UPDATE_WRITE_BARRIER, SKIP_ICACHE_FLUSH); } else { intptr_t delta = raw_instruction_start() - reinterpret_cast
(desc.buffer); it.rinfo()->apply(delta); } } } SafepointEntry Code::GetSafepointEntry(Isolate* isolate, Address pc) { SafepointTable table(isolate, pc, *this); return table.FindEntry(pc); } Address Code::OffHeapInstructionStart(Isolate* isolate, Address pc) const { DCHECK(is_off_heap_trampoline()); EmbeddedData d = EmbeddedData::GetEmbeddedDataForPC(isolate, pc); return d.InstructionStartOfBuiltin(builtin_id()); } Address Code::OffHeapInstructionEnd(Isolate* isolate, Address pc) const { DCHECK(is_off_heap_trampoline()); EmbeddedData d = EmbeddedData::GetEmbeddedDataForPC(isolate, pc); return d.InstructionStartOfBuiltin(builtin_id()) + d.InstructionSizeOfBuiltin(builtin_id()); } // TODO(cbruni): Move to BytecodeArray int AbstractCode::SourcePosition(int offset) { CHECK_NE(kind(), CodeKind::BASELINE); Object maybe_table = SourcePositionTableInternal(); if (maybe_table.IsException()) return kNoSourcePosition; ByteArray source_position_table = ByteArray::cast(maybe_table); // Subtract one because the current PC is one instruction after the call site. if (IsCode()) offset--; int position = 0; for (SourcePositionTableIterator iterator( source_position_table, SourcePositionTableIterator::kJavaScriptOnly, SourcePositionTableIterator::kDontSkipFunctionEntry); !iterator.done() && iterator.code_offset() <= offset; iterator.Advance()) { position = iterator.source_position().ScriptOffset(); } return position; } // TODO(cbruni): Move to BytecodeArray int AbstractCode::SourceStatementPosition(int offset) { CHECK_NE(kind(), CodeKind::BASELINE); // First find the closest position. int position = SourcePosition(offset); // Now find the closest statement position before the position. int statement_position = 0; for (SourcePositionTableIterator it(SourcePositionTableInternal()); !it.done(); it.Advance()) { if (it.is_statement()) { int p = it.source_position().ScriptOffset(); if (statement_position < p && p <= position) { statement_position = p; } } } return statement_position; } bool Code::CanDeoptAt(Isolate* isolate, Address pc) { DeoptimizationData deopt_data = DeoptimizationData::cast(deoptimization_data()); Address code_start_address = InstructionStart(isolate, pc); for (int i = 0; i < deopt_data.DeoptCount(); i++) { if (deopt_data.Pc(i).value() == -1) continue; Address address = code_start_address + deopt_data.Pc(i).value(); if (address == pc && deopt_data.GetBytecodeOffset(i) != BytecodeOffset::None()) { return true; } } return false; } bool Code::IsIsolateIndependent(Isolate* isolate) { static constexpr int kModeMask = RelocInfo::AllRealModesMask() & ~RelocInfo::ModeMask(RelocInfo::CONST_POOL) & ~RelocInfo::ModeMask(RelocInfo::OFF_HEAP_TARGET) & ~RelocInfo::ModeMask(RelocInfo::VENEER_POOL); STATIC_ASSERT(kModeMask == (RelocInfo::ModeMask(RelocInfo::CODE_TARGET) | RelocInfo::ModeMask(RelocInfo::RELATIVE_CODE_TARGET) | RelocInfo::ModeMask(RelocInfo::COMPRESSED_EMBEDDED_OBJECT) | RelocInfo::ModeMask(RelocInfo::FULL_EMBEDDED_OBJECT) | RelocInfo::ModeMask(RelocInfo::DATA_EMBEDDED_OBJECT) | RelocInfo::ModeMask(RelocInfo::EXTERNAL_REFERENCE) | RelocInfo::ModeMask(RelocInfo::INTERNAL_REFERENCE) | RelocInfo::ModeMask(RelocInfo::INTERNAL_REFERENCE_ENCODED) | RelocInfo::ModeMask(RelocInfo::RUNTIME_ENTRY) | RelocInfo::ModeMask(RelocInfo::WASM_CALL) | RelocInfo::ModeMask(RelocInfo::WASM_STUB_CALL))); #if defined(V8_TARGET_ARCH_PPC) || defined(V8_TARGET_ARCH_PPC64) || \ defined(V8_TARGET_ARCH_MIPS64) return RelocIterator(*this, kModeMask).done(); #elif defined(V8_TARGET_ARCH_X64) || defined(V8_TARGET_ARCH_ARM64) || \ defined(V8_TARGET_ARCH_ARM) || defined(V8_TARGET_ARCH_MIPS) || \ defined(V8_TARGET_ARCH_S390) || defined(V8_TARGET_ARCH_IA32) || \ defined(V8_TARGET_ARCH_RISCV64) || defined(V8_TARGET_ARCH_LOONG64) for (RelocIterator it(*this, kModeMask); !it.done(); it.next()) { // On these platforms we emit relative builtin-to-builtin // jumps for isolate independent builtins in the snapshot. They are later // rewritten as pc-relative jumps to the off-heap instruction stream and are // thus process-independent. See also: FinalizeEmbeddedCodeTargets. if (RelocInfo::IsCodeTargetMode(it.rinfo()->rmode())) { Address target_address = it.rinfo()->target_address(); if (OffHeapInstructionStream::PcIsOffHeap(isolate, target_address)) continue; Code target = Code::GetCodeFromTargetAddress(target_address); CHECK(target.IsCode()); if (Builtins::IsIsolateIndependentBuiltin(target)) continue; } return false; } return true; #else #error Unsupported architecture. #endif } bool Code::Inlines(SharedFunctionInfo sfi) { // We can only check for inlining for optimized code. DCHECK(is_optimized_code()); DisallowGarbageCollection no_gc; DeoptimizationData const data = DeoptimizationData::cast(deoptimization_data()); if (data.length() == 0) return false; if (data.SharedFunctionInfo() == sfi) return true; DeoptimizationLiteralArray const literals = data.LiteralArray(); int const inlined_count = data.InlinedFunctionCount().value(); for (int i = 0; i < inlined_count; ++i) { if (SharedFunctionInfo::cast(literals.get(i)) == sfi) return true; } return false; } Code::OptimizedCodeIterator::OptimizedCodeIterator(Isolate* isolate) { isolate_ = isolate; Object list = isolate->heap()->native_contexts_list(); next_context_ = list.IsUndefined(isolate_) ? NativeContext() : NativeContext::cast(list); } Code Code::OptimizedCodeIterator::Next() { do { Object next; if (!current_code_.is_null()) { // Get next code in the linked list. next = current_code_.next_code_link(); } else if (!next_context_.is_null()) { // Linked list of code exhausted. Get list of next context. next = next_context_.OptimizedCodeListHead(); Object next_context = next_context_.next_context_link(); next_context_ = next_context.IsUndefined(isolate_) ? NativeContext() : NativeContext::cast(next_context); } else { // Exhausted contexts. return Code(); } current_code_ = next.IsUndefined(isolate_) ? Code() : FromCodeT(CodeT::cast(next)); } while (current_code_.is_null()); DCHECK(CodeKindCanDeoptimize(current_code_.kind())); return current_code_; } Handle DeoptimizationData::New(Isolate* isolate, int deopt_entry_count, AllocationType allocation) { return Handle::cast(isolate->factory()->NewFixedArray( LengthFor(deopt_entry_count), allocation)); } Handle DeoptimizationData::Empty(Isolate* isolate) { return Handle::cast( isolate->factory()->empty_fixed_array()); } SharedFunctionInfo DeoptimizationData::GetInlinedFunction(int index) { if (index == -1) { return SharedFunctionInfo::cast(SharedFunctionInfo()); } else { return SharedFunctionInfo::cast(LiteralArray().get(index)); } } #ifdef ENABLE_DISASSEMBLER const char* Code::GetName(Isolate* isolate) const { if (kind() == CodeKind::BYTECODE_HANDLER) { return isolate->interpreter()->LookupNameOfBytecodeHandler(*this); } else { // There are some handlers and ICs that we can also find names for with // Builtins::Lookup. return isolate->builtins()->Lookup(raw_instruction_start()); } } namespace { void print_pc(std::ostream& os, int pc) { if (pc == -1) { os << "NA"; } else { os << std::hex << pc << std::dec; } } } // anonymous namespace void DeoptimizationData::DeoptimizationDataPrint(std::ostream& os) { if (length() == 0) { os << "Deoptimization Input Data invalidated by lazy deoptimization\n"; return; } int const inlined_function_count = InlinedFunctionCount().value(); os << "Inlined functions (count = " << inlined_function_count << ")\n"; for (int id = 0; id < inlined_function_count; ++id) { Object info = LiteralArray().get(id); os << " " << Brief(SharedFunctionInfo::cast(info)) << "\n"; } os << "\n"; int deopt_count = DeoptCount(); os << "Deoptimization Input Data (deopt points = " << deopt_count << ")\n"; if (0 != deopt_count) { #ifdef DEBUG os << " index bytecode-offset node-id pc"; #else // DEBUG os << " index bytecode-offset pc"; #endif // DEBUG if (FLAG_print_code_verbose) os << " commands"; os << "\n"; } for (int i = 0; i < deopt_count; i++) { os << std::setw(6) << i << " " << std::setw(15) << GetBytecodeOffset(i).ToInt() << " " #ifdef DEBUG << std::setw(7) << NodeId(i).value() << " " #endif // DEBUG << std::setw(4); print_pc(os, Pc(i).value()); os << std::setw(2); if (!FLAG_print_code_verbose) { os << "\n"; continue; } TranslationArrayPrintSingleFrame(os, TranslationByteArray(), TranslationIndex(i).value(), LiteralArray()); } } namespace { inline void DisassembleCodeRange(Isolate* isolate, std::ostream& os, Code code, Address begin, size_t size, Address current_pc) { Address end = begin + size; AllowHandleAllocation allow_handles; DisallowGarbageCollection no_gc; HandleScope handle_scope(isolate); Disassembler::Decode(isolate, os, reinterpret_cast(begin), reinterpret_cast(end), CodeReference(handle(code, isolate)), current_pc); } } // namespace void Code::Disassemble(const char* name, std::ostream& os, Isolate* isolate, Address current_pc) { os << "kind = " << CodeKindToString(kind()) << "\n"; if (name == nullptr) { name = GetName(isolate); } if ((name != nullptr) && (name[0] != '\0')) { os << "name = " << name << "\n"; } if (CodeKindIsOptimizedJSFunction(kind()) && kind() != CodeKind::BASELINE) { os << "stack_slots = " << stack_slots() << "\n"; } os << "compiler = " << (is_turbofanned() ? "turbofan" : is_maglevved() ? "turbofan" : kind() == CodeKind::BASELINE ? "baseline" : "unknown") << "\n"; os << "address = " << reinterpret_cast(ptr()) << "\n\n"; if (is_off_heap_trampoline()) { int trampoline_size = raw_instruction_size(); os << "Trampoline (size = " << trampoline_size << ")\n"; DisassembleCodeRange(isolate, os, *this, raw_instruction_start(), trampoline_size, current_pc); os << "\n"; } { int code_size = InstructionSize(); os << "Instructions (size = " << code_size << ")\n"; DisassembleCodeRange(isolate, os, *this, InstructionStart(), code_size, current_pc); if (int pool_size = constant_pool_size()) { DCHECK_EQ(pool_size & kPointerAlignmentMask, 0); os << "\nConstant Pool (size = " << pool_size << ")\n"; base::Vector buf = base::Vector::New(50); intptr_t* ptr = reinterpret_cast(constant_pool()); for (int i = 0; i < pool_size; i += kSystemPointerSize, ptr++) { SNPrintF(buf, "%4d %08" V8PRIxPTR, i, *ptr); os << static_cast(ptr) << " " << buf.begin() << "\n"; } } } os << "\n"; // TODO(cbruni): add support for baseline code. if (kind() != CodeKind::BASELINE) { { SourcePositionTableIterator it( source_position_table(), SourcePositionTableIterator::kJavaScriptOnly); if (!it.done()) { os << "Source positions:\n pc offset position\n"; for (; !it.done(); it.Advance()) { os << std::setw(10) << std::hex << it.code_offset() << std::dec << std::setw(10) << it.source_position().ScriptOffset() << (it.is_statement() ? " statement" : "") << "\n"; } os << "\n"; } } { SourcePositionTableIterator it( source_position_table(), SourcePositionTableIterator::kExternalOnly); if (!it.done()) { os << "External Source positions:\n pc offset fileid line\n"; for (; !it.done(); it.Advance()) { DCHECK(it.source_position().IsExternal()); os << std::setw(10) << std::hex << it.code_offset() << std::dec << std::setw(10) << it.source_position().ExternalFileId() << std::setw(10) << it.source_position().ExternalLine() << "\n"; } os << "\n"; } } } if (CodeKindCanDeoptimize(kind())) { DeoptimizationData data = DeoptimizationData::cast(this->deoptimization_data()); data.DeoptimizationDataPrint(os); } os << "\n"; if (uses_safepoint_table()) { SafepointTable table(isolate, current_pc, *this); table.Print(os); os << "\n"; } if (has_handler_table()) { HandlerTable table(*this); os << "Handler Table (size = " << table.NumberOfReturnEntries() << ")\n"; if (CodeKindIsOptimizedJSFunction(kind())) { table.HandlerTableReturnPrint(os); } os << "\n"; } os << "RelocInfo (size = " << relocation_size() << ")\n"; for (RelocIterator it(*this); !it.done(); it.next()) { it.rinfo()->Print(isolate, os); } os << "\n"; if (has_unwinding_info()) { os << "UnwindingInfo (size = " << unwinding_info_size() << ")\n"; EhFrameDisassembler eh_frame_disassembler( reinterpret_cast(unwinding_info_start()), reinterpret_cast(unwinding_info_end())); eh_frame_disassembler.DisassembleToStream(os); os << "\n"; } } #endif // ENABLE_DISASSEMBLER void BytecodeArray::Disassemble(std::ostream& os) { DisallowGarbageCollection no_gc; os << "Parameter count " << parameter_count() << "\n"; os << "Register count " << register_count() << "\n"; os << "Frame size " << frame_size() << "\n"; os << "OSR urgency: " << osr_urgency() << "\n"; os << "Bytecode age: " << bytecode_age() << "\n"; Address base_address = GetFirstBytecodeAddress(); SourcePositionTableIterator source_positions(SourcePositionTable()); // Storage for backing the handle passed to the iterator. This handle won't be // updated by the gc, but that's ok because we've disallowed GCs anyway. BytecodeArray handle_storage = *this; Handle handle(reinterpret_cast(&handle_storage)); interpreter::BytecodeArrayIterator iterator(handle); while (!iterator.done()) { if (!source_positions.done() && iterator.current_offset() == source_positions.code_offset()) { os << std::setw(5) << source_positions.source_position().ScriptOffset(); os << (source_positions.is_statement() ? " S> " : " E> "); source_positions.Advance(); } else { os << " "; } Address current_address = base_address + iterator.current_offset(); os << reinterpret_cast(current_address) << " @ " << std::setw(4) << iterator.current_offset() << " : "; interpreter::BytecodeDecoder::Decode( os, reinterpret_cast(current_address)); if (interpreter::Bytecodes::IsJump(iterator.current_bytecode())) { Address jump_target = base_address + iterator.GetJumpTargetOffset(); os << " (" << reinterpret_cast(jump_target) << " @ " << iterator.GetJumpTargetOffset() << ")"; } if (interpreter::Bytecodes::IsSwitch(iterator.current_bytecode())) { os << " {"; bool first_entry = true; for (interpreter::JumpTableTargetOffset entry : iterator.GetJumpTableTargetOffsets()) { if (first_entry) { first_entry = false; } else { os << ","; } os << " " << entry.case_value << ": @" << entry.target_offset; } os << " }"; } os << std::endl; iterator.Advance(); } os << "Constant pool (size = " << constant_pool().length() << ")\n"; #ifdef OBJECT_PRINT if (constant_pool().length() > 0) { constant_pool().Print(os); } #endif os << "Handler Table (size = " << handler_table().length() << ")\n"; #ifdef ENABLE_DISASSEMBLER if (handler_table().length() > 0) { HandlerTable table(*this); table.HandlerTableRangePrint(os); } #endif ByteArray source_position_table = SourcePositionTable(); os << "Source Position Table (size = " << source_position_table.length() << ")\n"; #ifdef OBJECT_PRINT if (source_position_table.length() > 0) { os << Brief(source_position_table) << std::endl; } #endif } void BytecodeArray::CopyBytecodesTo(BytecodeArray to) { BytecodeArray from = *this; DCHECK_EQ(from.length(), to.length()); CopyBytes(reinterpret_cast(to.GetFirstBytecodeAddress()), reinterpret_cast(from.GetFirstBytecodeAddress()), from.length()); } void BytecodeArray::MakeOlder() { // BytecodeArray is aged in concurrent marker. // The word must be completely within the byte code array. Address age_addr = address() + kBytecodeAgeOffset; DCHECK_LE(RoundDown(age_addr, kTaggedSize) + kTaggedSize, address() + Size()); Age age = bytecode_age(); if (age < kLastBytecodeAge) { static_assert(kBytecodeAgeSize == kUInt16Size); base::AsAtomic16::Relaxed_CompareAndSwap( reinterpret_cast(age_addr), age, age + 1); } DCHECK_GE(bytecode_age(), kFirstBytecodeAge); DCHECK_LE(bytecode_age(), kLastBytecodeAge); } bool BytecodeArray::IsOld() const { return bytecode_age() >= kIsOldBytecodeAge; } DependentCode DependentCode::GetDependentCode(Handle object) { if (object->IsMap()) { return Handle::cast(object)->dependent_code(); } else if (object->IsPropertyCell()) { return Handle::cast(object)->dependent_code(); } else if (object->IsAllocationSite()) { return Handle::cast(object)->dependent_code(); } UNREACHABLE(); } void DependentCode::SetDependentCode(Handle object, Handle dep) { if (object->IsMap()) { Handle::cast(object)->set_dependent_code(*dep); } else if (object->IsPropertyCell()) { Handle::cast(object)->set_dependent_code(*dep); } else if (object->IsAllocationSite()) { Handle::cast(object)->set_dependent_code(*dep); } else { UNREACHABLE(); } } namespace { void PrintDependencyGroups(DependentCode::DependencyGroups groups) { while (groups != 0) { auto group = static_cast( 1 << base::bits::CountTrailingZeros(static_cast(groups))); StdoutStream{} << DependentCode::DependencyGroupName(group); groups &= ~group; if (groups != 0) StdoutStream{} << ","; } } } // namespace void DependentCode::InstallDependency(Isolate* isolate, Handle code, Handle object, DependencyGroups groups) { if (V8_UNLIKELY(FLAG_trace_compilation_dependencies)) { StdoutStream{} << "Installing dependency of [" << code->GetHeapObject() << "] on [" << object << "] in groups ["; PrintDependencyGroups(groups); StdoutStream{} << "]\n"; } Handle old_deps(DependentCode::GetDependentCode(object), isolate); Handle new_deps = InsertWeakCode(isolate, old_deps, groups, code); // Update the list head if necessary. if (!new_deps.is_identical_to(old_deps)) { DependentCode::SetDependentCode(object, new_deps); } } Handle DependentCode::InsertWeakCode( Isolate* isolate, Handle entries, DependencyGroups groups, Handle code) { if (entries->length() == entries->capacity()) { // We'd have to grow - try to compact first. entries->IterateAndCompact([](CodeT, DependencyGroups) { return false; }); } MaybeObjectHandle code_slot(HeapObjectReference::Weak(ToCodeT(*code)), isolate); MaybeObjectHandle group_slot(MaybeObject::FromSmi(Smi::FromInt(groups)), isolate); entries = Handle::cast( WeakArrayList::AddToEnd(isolate, entries, code_slot, group_slot)); return entries; } Handle DependentCode::New(Isolate* isolate, DependencyGroups groups, Handle code) { Handle result = Handle::cast( isolate->factory()->NewWeakArrayList(LengthFor(1), AllocationType::kOld)); result->Set(0, HeapObjectReference::Weak(ToCodeT(*code))); result->Set(1, Smi::FromInt(groups)); return result; } void DependentCode::IterateAndCompact(const IterateAndCompactFn& fn) { DisallowGarbageCollection no_gc; int len = length(); if (len == 0) return; // We compact during traversal, thus use a somewhat custom loop construct: // // - Loop back-to-front s.t. trailing cleared entries can simply drop off // the back of the list. // - Any cleared slots are filled from the back of the list. int i = len - kSlotsPerEntry; while (i >= 0) { MaybeObject obj = Get(i + kCodeSlotOffset); if (obj->IsCleared()) { len = FillEntryFromBack(i, len); i -= kSlotsPerEntry; continue; } if (fn(CodeT::cast(obj->GetHeapObjectAssumeWeak()), static_cast( Get(i + kGroupsSlotOffset).ToSmi().value()))) { len = FillEntryFromBack(i, len); } i -= kSlotsPerEntry; } set_length(len); } bool DependentCode::MarkCodeForDeoptimization( DependentCode::DependencyGroups deopt_groups) { DisallowGarbageCollection no_gc; bool marked_something = false; IterateAndCompact([&](CodeT codet, DependencyGroups groups) { if ((groups & deopt_groups) == 0) return false; // TODO(v8:11880): avoid roundtrips between cdc and code. Code code = FromCodeT(codet); if (!code.marked_for_deoptimization()) { code.SetMarkedForDeoptimization("code dependencies"); marked_something = true; } return true; }); return marked_something; } int DependentCode::FillEntryFromBack(int index, int length) { DCHECK_EQ(index % 2, 0); DCHECK_EQ(length % 2, 0); for (int i = length - kSlotsPerEntry; i > index; i -= kSlotsPerEntry) { MaybeObject obj = Get(i + kCodeSlotOffset); if (obj->IsCleared()) continue; Set(index + kCodeSlotOffset, obj); Set(index + kGroupsSlotOffset, Get(i + kGroupsSlotOffset), SKIP_WRITE_BARRIER); return i; } return index; // No non-cleared entry found. } void DependentCode::DeoptimizeDependentCodeGroup( Isolate* isolate, DependentCode::DependencyGroups groups) { DisallowGarbageCollection no_gc_scope; bool marked_something = MarkCodeForDeoptimization(groups); if (marked_something) { DCHECK(AllowCodeDependencyChange::IsAllowed()); Deoptimizer::DeoptimizeMarkedCode(isolate); } } // static DependentCode DependentCode::empty_dependent_code(const ReadOnlyRoots& roots) { return DependentCode::cast(roots.empty_weak_array_list()); } void Code::SetMarkedForDeoptimization(const char* reason) { set_marked_for_deoptimization(true); Deoptimizer::TraceMarkForDeoptimization(*this, reason); } const char* DependentCode::DependencyGroupName(DependencyGroup group) { switch (group) { case kTransitionGroup: return "transition"; case kPrototypeCheckGroup: return "prototype-check"; case kPropertyCellChangedGroup: return "property-cell-changed"; case kFieldConstGroup: return "field-const"; case kFieldTypeGroup: return "field-type"; case kFieldRepresentationGroup: return "field-representation"; case kInitialMapChangedGroup: return "initial-map-changed"; case kAllocationSiteTenuringChangedGroup: return "allocation-site-tenuring-changed"; case kAllocationSiteTransitionChangedGroup: return "allocation-site-transition-changed"; } UNREACHABLE(); } } // namespace internal } // namespace v8