// Copyright 2018 the V8 project authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #include "src/builtins/builtins-lazy-gen.h" #include "src/builtins/builtins-utils-gen.h" #include "src/builtins/builtins.h" #include "src/common/globals.h" #include "src/objects/code-inl.h" #include "src/objects/feedback-vector.h" #include "src/objects/shared-function-info.h" namespace v8 { namespace internal { void LazyBuiltinsAssembler::GenerateTailCallToJSCode( TNode code, TNode function) { auto argc = UncheckedParameter(Descriptor::kActualArgumentsCount); auto context = Parameter(Descriptor::kContext); auto new_target = Parameter(Descriptor::kNewTarget); TailCallJSCode(code, context, function, new_target, argc); } void LazyBuiltinsAssembler::GenerateTailCallToReturnedCode( Runtime::FunctionId function_id, TNode function) { auto context = Parameter(Descriptor::kContext); TNode code = CAST(CallRuntime(function_id, context, function)); GenerateTailCallToJSCode(code, function); } void LazyBuiltinsAssembler::TailCallRuntimeIfStateEquals( TNode state, TieringState expected_state, Runtime::FunctionId function_id, TNode function) { Label no_match(this); GotoIfNot( Word32Equal(state, Uint32Constant(static_cast(expected_state))), &no_match); GenerateTailCallToReturnedCode(function_id, function); BIND(&no_match); } void LazyBuiltinsAssembler::MaybeTailCallOptimizedCodeSlot( TNode function, TNode feedback_vector) { Label fallthrough(this), may_have_optimized_code(this); TNode optimization_state = LoadObjectField(feedback_vector, FeedbackVector::kFlagsOffset); // Fall through if no optimization trigger or optimized code. GotoIfNot( IsSetWord32( optimization_state, FeedbackVector::kHasOptimizedCodeOrTieringStateIsAnyRequestMask), &fallthrough); GotoIfNot(IsSetWord32(optimization_state, FeedbackVector::kTieringStateIsAnyRequestMask), &may_have_optimized_code); // TODO(ishell): introduce Runtime::kHandleTieringState and check // all these state values there. TNode state = DecodeWord32(optimization_state); TailCallRuntimeIfStateEquals(state, TieringState::kRequestTurbofan_Synchronous, Runtime::kCompileTurbofan_Synchronous, function); TailCallRuntimeIfStateEquals(state, TieringState::kRequestTurbofan_Concurrent, Runtime::kCompileTurbofan_Concurrent, function); TailCallRuntimeIfStateEquals(state, TieringState::kRequestMaglev_Synchronous, Runtime::kCompileMaglev_Synchronous, function); TailCallRuntimeIfStateEquals(state, TieringState::kRequestMaglev_Concurrent, Runtime::kCompileMaglev_Concurrent, function); Unreachable(); BIND(&may_have_optimized_code); { Label heal_optimized_code_slot(this); TNode maybe_optimized_code_entry = LoadMaybeWeakObjectField( feedback_vector, FeedbackVector::kMaybeOptimizedCodeOffset); // Optimized code slot is a weak reference to CodeT object. TNode optimized_code = CAST(GetHeapObjectAssumeWeak( maybe_optimized_code_entry, &heal_optimized_code_slot)); // Check if the optimized code is marked for deopt. If it is, call the // runtime to clear it. TNode code_data_container = CodeDataContainerFromCodeT(optimized_code); TNode code_kind_specific_flags = LoadObjectField( code_data_container, CodeDataContainer::kKindSpecificFlagsOffset); GotoIf(IsSetWord32( code_kind_specific_flags), &heal_optimized_code_slot); // Optimized code is good, get it into the closure and link the closure into // the optimized functions list, then tail call the optimized code. StoreObjectField(function, JSFunction::kCodeOffset, optimized_code); Comment("MaybeTailCallOptimizedCodeSlot:: GenerateTailCallToJSCode"); GenerateTailCallToJSCode(optimized_code, function); // Optimized code slot contains deoptimized code, or the code is cleared // and tiering state hasn't yet been updated. Evict the code, update the // state and re-enter the closure's code. BIND(&heal_optimized_code_slot); GenerateTailCallToReturnedCode(Runtime::kHealOptimizedCodeSlot, function); } // Fall-through if the optimized code cell is clear and the tiering state is // kNone. BIND(&fallthrough); } void LazyBuiltinsAssembler::CompileLazy(TNode function) { // First lookup code, maybe we don't need to compile! Label compile_function(this, Label::kDeferred); // Check the code object for the SFI. If SFI's code entry points to // CompileLazy, then we need to lazy compile regardless of the function or // tiering state. TNode shared = CAST(LoadObjectField(function, JSFunction::kSharedFunctionInfoOffset)); TVARIABLE(Uint16T, sfi_data_type); TNode sfi_code = GetSharedFunctionInfoCode(shared, &sfi_data_type, &compile_function); TNode feedback_cell_value = LoadFeedbackCellValue(function); // If feedback cell isn't initialized, compile function GotoIf(IsUndefined(feedback_cell_value), &compile_function); CSA_DCHECK(this, TaggedNotEqual(sfi_code, HeapConstant(BUILTIN_CODE( isolate(), CompileLazy)))); StoreObjectField(function, JSFunction::kCodeOffset, sfi_code); Label maybe_use_sfi_code(this); // If there is no feedback, don't check for optimized code. GotoIf(HasInstanceType(feedback_cell_value, CLOSURE_FEEDBACK_CELL_ARRAY_TYPE), &maybe_use_sfi_code); // If it isn't undefined or fixed array it must be a feedback vector. CSA_DCHECK(this, IsFeedbackVector(feedback_cell_value)); // Is there a tiering state or optimized code in the feedback vector? MaybeTailCallOptimizedCodeSlot(function, CAST(feedback_cell_value)); Goto(&maybe_use_sfi_code); // At this point we have a candidate Code object. It's *not* a cached // optimized Code object (we'd have tail-called it above). A usual case would // be the InterpreterEntryTrampoline to start executing existing bytecode. BIND(&maybe_use_sfi_code); Label tailcall_code(this), baseline(this); TVARIABLE(CodeT, code); // Check if we have baseline code. GotoIf(InstanceTypeEqual(sfi_data_type.value(), CODET_TYPE), &baseline); code = sfi_code; Goto(&tailcall_code); BIND(&baseline); // Ensure we have a feedback vector. code = Select( IsFeedbackVector(feedback_cell_value), [=]() { return sfi_code; }, [=]() { return CAST(CallRuntime(Runtime::kInstallBaselineCode, Parameter(Descriptor::kContext), function)); }); Goto(&tailcall_code); BIND(&tailcall_code); GenerateTailCallToJSCode(code.value(), function); BIND(&compile_function); GenerateTailCallToReturnedCode(Runtime::kCompileLazy, function); } TF_BUILTIN(CompileLazy, LazyBuiltinsAssembler) { auto function = Parameter(Descriptor::kTarget); CompileLazy(function); } TF_BUILTIN(CompileLazyDeoptimizedCode, LazyBuiltinsAssembler) { auto function = Parameter(Descriptor::kTarget); TNode code = HeapConstant(BUILTIN_CODE(isolate(), CompileLazy)); // Set the code slot inside the JSFunction to CompileLazy. StoreObjectField(function, JSFunction::kCodeOffset, code); GenerateTailCallToJSCode(code, function); } } // namespace internal } // namespace v8