• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 // Copyright 2015 the V8 project authors. All rights reserved.
2 // Use of this source code is governed by a BSD-style license that can be
3 // found in the LICENSE file.
4 
5 #include "src/interpreter/interpreter.h"
6 
7 #include <fstream>
8 
9 #include "src/ast/prettyprinter.h"
10 #include "src/code-factory.h"
11 #include "src/compiler.h"
12 #include "src/factory.h"
13 #include "src/interpreter/bytecode-generator.h"
14 #include "src/interpreter/bytecodes.h"
15 #include "src/interpreter/interpreter-assembler.h"
16 #include "src/interpreter/interpreter-intrinsics.h"
17 #include "src/log.h"
18 #include "src/zone.h"
19 
20 namespace v8 {
21 namespace internal {
22 namespace interpreter {
23 
24 using compiler::Node;
25 typedef CodeStubAssembler::Label Label;
26 typedef CodeStubAssembler::Variable Variable;
27 
28 #define __ assembler->
29 
Interpreter(Isolate * isolate)30 Interpreter::Interpreter(Isolate* isolate) : isolate_(isolate) {
31   memset(dispatch_table_, 0, sizeof(dispatch_table_));
32 }
33 
Initialize()34 void Interpreter::Initialize() {
35   if (IsDispatchTableInitialized()) return;
36   Zone zone(isolate_->allocator());
37   HandleScope scope(isolate_);
38 
39   if (FLAG_trace_ignition_dispatches) {
40     static const int kBytecodeCount = static_cast<int>(Bytecode::kLast) + 1;
41     bytecode_dispatch_counters_table_.Reset(
42         new uintptr_t[kBytecodeCount * kBytecodeCount]);
43     memset(bytecode_dispatch_counters_table_.get(), 0,
44            sizeof(uintptr_t) * kBytecodeCount * kBytecodeCount);
45   }
46 
47   // Generate bytecode handlers for all bytecodes and scales.
48   const OperandScale kOperandScales[] = {
49 #define VALUE(Name, _) OperandScale::k##Name,
50       OPERAND_SCALE_LIST(VALUE)
51 #undef VALUE
52   };
53 
54   for (OperandScale operand_scale : kOperandScales) {
55 #define GENERATE_CODE(Name, ...)                                               \
56   {                                                                            \
57     if (Bytecodes::BytecodeHasHandler(Bytecode::k##Name, operand_scale)) {     \
58       InterpreterAssembler assembler(isolate_, &zone, Bytecode::k##Name,       \
59                                      operand_scale);                           \
60       Do##Name(&assembler);                                                    \
61       Handle<Code> code = assembler.GenerateCode();                            \
62       size_t index = GetDispatchTableIndex(Bytecode::k##Name, operand_scale);  \
63       dispatch_table_[index] = code->entry();                                  \
64       TraceCodegen(code);                                                      \
65       PROFILE(                                                                 \
66           isolate_,                                                            \
67           CodeCreateEvent(                                                     \
68               CodeEventListener::BYTECODE_HANDLER_TAG,                         \
69               AbstractCode::cast(*code),                                       \
70               Bytecodes::ToString(Bytecode::k##Name, operand_scale).c_str())); \
71     }                                                                          \
72   }
73     BYTECODE_LIST(GENERATE_CODE)
74 #undef GENERATE_CODE
75   }
76 
77   // Fill unused entries will the illegal bytecode handler.
78   size_t illegal_index =
79       GetDispatchTableIndex(Bytecode::kIllegal, OperandScale::kSingle);
80   for (size_t index = 0; index < arraysize(dispatch_table_); ++index) {
81     if (dispatch_table_[index] == nullptr) {
82       dispatch_table_[index] = dispatch_table_[illegal_index];
83     }
84   }
85 }
86 
GetBytecodeHandler(Bytecode bytecode,OperandScale operand_scale)87 Code* Interpreter::GetBytecodeHandler(Bytecode bytecode,
88                                       OperandScale operand_scale) {
89   DCHECK(IsDispatchTableInitialized());
90   DCHECK(Bytecodes::BytecodeHasHandler(bytecode, operand_scale));
91   size_t index = GetDispatchTableIndex(bytecode, operand_scale);
92   Address code_entry = dispatch_table_[index];
93   return Code::GetCodeFromTargetAddress(code_entry);
94 }
95 
96 // static
GetDispatchTableIndex(Bytecode bytecode,OperandScale operand_scale)97 size_t Interpreter::GetDispatchTableIndex(Bytecode bytecode,
98                                           OperandScale operand_scale) {
99   static const size_t kEntriesPerOperandScale = 1u << kBitsPerByte;
100   size_t index = static_cast<size_t>(bytecode);
101   switch (operand_scale) {
102     case OperandScale::kSingle:
103       return index;
104     case OperandScale::kDouble:
105       return index + kEntriesPerOperandScale;
106     case OperandScale::kQuadruple:
107       return index + 2 * kEntriesPerOperandScale;
108   }
109   UNREACHABLE();
110   return 0;
111 }
112 
IterateDispatchTable(ObjectVisitor * v)113 void Interpreter::IterateDispatchTable(ObjectVisitor* v) {
114   for (int i = 0; i < kDispatchTableSize; i++) {
115     Address code_entry = dispatch_table_[i];
116     Object* code = code_entry == nullptr
117                        ? nullptr
118                        : Code::GetCodeFromTargetAddress(code_entry);
119     Object* old_code = code;
120     v->VisitPointer(&code);
121     if (code != old_code) {
122       dispatch_table_[i] = reinterpret_cast<Code*>(code)->entry();
123     }
124   }
125 }
126 
127 // static
InterruptBudget()128 int Interpreter::InterruptBudget() {
129   // TODO(ignition): Tune code size multiplier.
130   const int kCodeSizeMultiplier = 32;
131   return FLAG_interrupt_budget * kCodeSizeMultiplier;
132 }
133 
MakeBytecode(CompilationInfo * info)134 bool Interpreter::MakeBytecode(CompilationInfo* info) {
135   RuntimeCallTimerScope runtimeTimer(info->isolate(),
136                                      &RuntimeCallStats::CompileIgnition);
137   TimerEventScope<TimerEventCompileIgnition> timer(info->isolate());
138   TRACE_EVENT0("v8", "V8.CompileIgnition");
139 
140   if (FLAG_print_bytecode || FLAG_print_source || FLAG_print_ast) {
141     OFStream os(stdout);
142     base::SmartArrayPointer<char> name = info->GetDebugName();
143     os << "[generating bytecode for function: " << info->GetDebugName().get()
144        << "]" << std::endl
145        << std::flush;
146   }
147 
148 #ifdef DEBUG
149   if (info->parse_info() && FLAG_print_source) {
150     OFStream os(stdout);
151     os << "--- Source from AST ---" << std::endl
152        << PrettyPrinter(info->isolate()).PrintProgram(info->literal())
153        << std::endl
154        << std::flush;
155   }
156 
157   if (info->parse_info() && FLAG_print_ast) {
158     OFStream os(stdout);
159     os << "--- AST ---" << std::endl
160        << AstPrinter(info->isolate()).PrintProgram(info->literal()) << std::endl
161        << std::flush;
162   }
163 #endif  // DEBUG
164 
165   BytecodeGenerator generator(info);
166   Handle<BytecodeArray> bytecodes = generator.MakeBytecode();
167 
168   if (generator.HasStackOverflow()) return false;
169 
170   if (FLAG_print_bytecode) {
171     OFStream os(stdout);
172     bytecodes->Print(os);
173     os << std::flush;
174   }
175 
176   info->SetBytecodeArray(bytecodes);
177   info->SetCode(info->isolate()->builtins()->InterpreterEntryTrampoline());
178   return true;
179 }
180 
IsDispatchTableInitialized()181 bool Interpreter::IsDispatchTableInitialized() {
182   if (FLAG_trace_ignition || FLAG_trace_ignition_codegen ||
183       FLAG_trace_ignition_dispatches) {
184     // Regenerate table to add bytecode tracing operations, print the assembly
185     // code generated by TurboFan or instrument handlers with dispatch counters.
186     return false;
187   }
188   return dispatch_table_[0] != nullptr;
189 }
190 
TraceCodegen(Handle<Code> code)191 void Interpreter::TraceCodegen(Handle<Code> code) {
192 #ifdef ENABLE_DISASSEMBLER
193   if (FLAG_trace_ignition_codegen) {
194     OFStream os(stdout);
195     code->Disassemble(nullptr, os);
196     os << std::flush;
197   }
198 #endif  // ENABLE_DISASSEMBLER
199 }
200 
LookupNameOfBytecodeHandler(Code * code)201 const char* Interpreter::LookupNameOfBytecodeHandler(Code* code) {
202 #ifdef ENABLE_DISASSEMBLER
203 #define RETURN_NAME(Name, ...)                                 \
204   if (dispatch_table_[Bytecodes::ToByte(Bytecode::k##Name)] == \
205       code->entry()) {                                         \
206     return #Name;                                              \
207   }
208   BYTECODE_LIST(RETURN_NAME)
209 #undef RETURN_NAME
210 #endif  // ENABLE_DISASSEMBLER
211   return nullptr;
212 }
213 
GetDispatchCounter(Bytecode from,Bytecode to) const214 uintptr_t Interpreter::GetDispatchCounter(Bytecode from, Bytecode to) const {
215   int from_index = Bytecodes::ToByte(from);
216   int to_index = Bytecodes::ToByte(to);
217   return bytecode_dispatch_counters_table_[from_index * kNumberOfBytecodes +
218                                            to_index];
219 }
220 
GetDispatchCountersObject()221 Local<v8::Object> Interpreter::GetDispatchCountersObject() {
222   v8::Isolate* isolate = reinterpret_cast<v8::Isolate*>(isolate_);
223   Local<v8::Context> context = isolate->GetCurrentContext();
224 
225   Local<v8::Object> counters_map = v8::Object::New(isolate);
226 
227   // Output is a JSON-encoded object of objects.
228   //
229   // The keys on the top level object are source bytecodes,
230   // and corresponding value are objects. Keys on these last are the
231   // destinations of the dispatch and the value associated is a counter for
232   // the correspondent source-destination dispatch chain.
233   //
234   // Only non-zero counters are written to file, but an entry in the top-level
235   // object is always present, even if the value is empty because all counters
236   // for that source are zero.
237 
238   for (int from_index = 0; from_index < kNumberOfBytecodes; ++from_index) {
239     Bytecode from_bytecode = Bytecodes::FromByte(from_index);
240     Local<v8::Object> counters_row = v8::Object::New(isolate);
241 
242     for (int to_index = 0; to_index < kNumberOfBytecodes; ++to_index) {
243       Bytecode to_bytecode = Bytecodes::FromByte(to_index);
244       uintptr_t counter = GetDispatchCounter(from_bytecode, to_bytecode);
245 
246       if (counter > 0) {
247         std::string to_name = Bytecodes::ToString(to_bytecode);
248         Local<v8::String> to_name_object =
249             v8::String::NewFromUtf8(isolate, to_name.c_str(),
250                                     NewStringType::kNormal)
251                 .ToLocalChecked();
252         Local<v8::Number> counter_object = v8::Number::New(isolate, counter);
253         CHECK(counters_row
254                   ->DefineOwnProperty(context, to_name_object, counter_object)
255                   .IsJust());
256       }
257     }
258 
259     std::string from_name = Bytecodes::ToString(from_bytecode);
260     Local<v8::String> from_name_object =
261         v8::String::NewFromUtf8(isolate, from_name.c_str(),
262                                 NewStringType::kNormal)
263             .ToLocalChecked();
264 
265     CHECK(
266         counters_map->DefineOwnProperty(context, from_name_object, counters_row)
267             .IsJust());
268   }
269 
270   return counters_map;
271 }
272 
273 // LdaZero
274 //
275 // Load literal '0' into the accumulator.
DoLdaZero(InterpreterAssembler * assembler)276 void Interpreter::DoLdaZero(InterpreterAssembler* assembler) {
277   Node* zero_value = __ NumberConstant(0.0);
278   __ SetAccumulator(zero_value);
279   __ Dispatch();
280 }
281 
282 // LdaSmi <imm>
283 //
284 // Load an integer literal into the accumulator as a Smi.
DoLdaSmi(InterpreterAssembler * assembler)285 void Interpreter::DoLdaSmi(InterpreterAssembler* assembler) {
286   Node* raw_int = __ BytecodeOperandImm(0);
287   Node* smi_int = __ SmiTag(raw_int);
288   __ SetAccumulator(smi_int);
289   __ Dispatch();
290 }
291 
292 // LdaConstant <idx>
293 //
294 // Load constant literal at |idx| in the constant pool into the accumulator.
DoLdaConstant(InterpreterAssembler * assembler)295 void Interpreter::DoLdaConstant(InterpreterAssembler* assembler) {
296   Node* index = __ BytecodeOperandIdx(0);
297   Node* constant = __ LoadConstantPoolEntry(index);
298   __ SetAccumulator(constant);
299   __ Dispatch();
300 }
301 
302 // LdaUndefined
303 //
304 // Load Undefined into the accumulator.
DoLdaUndefined(InterpreterAssembler * assembler)305 void Interpreter::DoLdaUndefined(InterpreterAssembler* assembler) {
306   Node* undefined_value =
307       __ HeapConstant(isolate_->factory()->undefined_value());
308   __ SetAccumulator(undefined_value);
309   __ Dispatch();
310 }
311 
312 // LdrUndefined <reg>
313 //
314 // Loads undefined into the accumulator and |reg|.
DoLdrUndefined(InterpreterAssembler * assembler)315 void Interpreter::DoLdrUndefined(InterpreterAssembler* assembler) {
316   Node* undefined_value =
317       __ HeapConstant(isolate_->factory()->undefined_value());
318   Node* destination = __ BytecodeOperandReg(0);
319   __ StoreRegister(undefined_value, destination);
320   __ Dispatch();
321 }
322 
323 // LdaNull
324 //
325 // Load Null into the accumulator.
DoLdaNull(InterpreterAssembler * assembler)326 void Interpreter::DoLdaNull(InterpreterAssembler* assembler) {
327   Node* null_value = __ HeapConstant(isolate_->factory()->null_value());
328   __ SetAccumulator(null_value);
329   __ Dispatch();
330 }
331 
332 // LdaTheHole
333 //
334 // Load TheHole into the accumulator.
DoLdaTheHole(InterpreterAssembler * assembler)335 void Interpreter::DoLdaTheHole(InterpreterAssembler* assembler) {
336   Node* the_hole_value = __ HeapConstant(isolate_->factory()->the_hole_value());
337   __ SetAccumulator(the_hole_value);
338   __ Dispatch();
339 }
340 
341 // LdaTrue
342 //
343 // Load True into the accumulator.
DoLdaTrue(InterpreterAssembler * assembler)344 void Interpreter::DoLdaTrue(InterpreterAssembler* assembler) {
345   Node* true_value = __ HeapConstant(isolate_->factory()->true_value());
346   __ SetAccumulator(true_value);
347   __ Dispatch();
348 }
349 
350 // LdaFalse
351 //
352 // Load False into the accumulator.
DoLdaFalse(InterpreterAssembler * assembler)353 void Interpreter::DoLdaFalse(InterpreterAssembler* assembler) {
354   Node* false_value = __ HeapConstant(isolate_->factory()->false_value());
355   __ SetAccumulator(false_value);
356   __ Dispatch();
357 }
358 
359 // Ldar <src>
360 //
361 // Load accumulator with value from register <src>.
DoLdar(InterpreterAssembler * assembler)362 void Interpreter::DoLdar(InterpreterAssembler* assembler) {
363   Node* reg_index = __ BytecodeOperandReg(0);
364   Node* value = __ LoadRegister(reg_index);
365   __ SetAccumulator(value);
366   __ Dispatch();
367 }
368 
369 // Star <dst>
370 //
371 // Store accumulator to register <dst>.
DoStar(InterpreterAssembler * assembler)372 void Interpreter::DoStar(InterpreterAssembler* assembler) {
373   Node* reg_index = __ BytecodeOperandReg(0);
374   Node* accumulator = __ GetAccumulator();
375   __ StoreRegister(accumulator, reg_index);
376   __ Dispatch();
377 }
378 
379 // Mov <src> <dst>
380 //
381 // Stores the value of register <src> to register <dst>.
DoMov(InterpreterAssembler * assembler)382 void Interpreter::DoMov(InterpreterAssembler* assembler) {
383   Node* src_index = __ BytecodeOperandReg(0);
384   Node* src_value = __ LoadRegister(src_index);
385   Node* dst_index = __ BytecodeOperandReg(1);
386   __ StoreRegister(src_value, dst_index);
387   __ Dispatch();
388 }
389 
BuildLoadGlobal(Callable ic,InterpreterAssembler * assembler)390 Node* Interpreter::BuildLoadGlobal(Callable ic,
391                                    InterpreterAssembler* assembler) {
392   // Get the global object.
393   Node* context = __ GetContext();
394 
395   // Load the global via the LoadGlobalIC.
396   Node* code_target = __ HeapConstant(ic.code());
397   Node* raw_slot = __ BytecodeOperandIdx(0);
398   Node* smi_slot = __ SmiTag(raw_slot);
399   Node* type_feedback_vector = __ LoadTypeFeedbackVector();
400   return __ CallStub(ic.descriptor(), code_target, context, smi_slot,
401                      type_feedback_vector);
402 }
403 
404 // LdaGlobal <slot>
405 //
406 // Load the global with name in constant pool entry <name_index> into the
407 // accumulator using FeedBackVector slot <slot> outside of a typeof.
DoLdaGlobal(InterpreterAssembler * assembler)408 void Interpreter::DoLdaGlobal(InterpreterAssembler* assembler) {
409   Callable ic =
410       CodeFactory::LoadGlobalICInOptimizedCode(isolate_, NOT_INSIDE_TYPEOF);
411   Node* result = BuildLoadGlobal(ic, assembler);
412   __ SetAccumulator(result);
413   __ Dispatch();
414 }
415 
416 // LdrGlobal <slot> <reg>
417 //
418 // Load the global with name in constant pool entry <name_index> into
419 // register <reg> using FeedBackVector slot <slot> outside of a typeof.
DoLdrGlobal(InterpreterAssembler * assembler)420 void Interpreter::DoLdrGlobal(InterpreterAssembler* assembler) {
421   Callable ic =
422       CodeFactory::LoadGlobalICInOptimizedCode(isolate_, NOT_INSIDE_TYPEOF);
423   Node* result = BuildLoadGlobal(ic, assembler);
424   Node* destination = __ BytecodeOperandReg(1);
425   __ StoreRegister(result, destination);
426   __ Dispatch();
427 }
428 
429 // LdaGlobalInsideTypeof <slot>
430 //
431 // Load the global with name in constant pool entry <name_index> into the
432 // accumulator using FeedBackVector slot <slot> inside of a typeof.
DoLdaGlobalInsideTypeof(InterpreterAssembler * assembler)433 void Interpreter::DoLdaGlobalInsideTypeof(InterpreterAssembler* assembler) {
434   Callable ic =
435       CodeFactory::LoadGlobalICInOptimizedCode(isolate_, INSIDE_TYPEOF);
436   Node* result = BuildLoadGlobal(ic, assembler);
437   __ SetAccumulator(result);
438   __ Dispatch();
439 }
440 
DoStaGlobal(Callable ic,InterpreterAssembler * assembler)441 void Interpreter::DoStaGlobal(Callable ic, InterpreterAssembler* assembler) {
442   // Get the global object.
443   Node* context = __ GetContext();
444   Node* native_context =
445       __ LoadContextSlot(context, Context::NATIVE_CONTEXT_INDEX);
446   Node* global = __ LoadContextSlot(native_context, Context::EXTENSION_INDEX);
447 
448   // Store the global via the StoreIC.
449   Node* code_target = __ HeapConstant(ic.code());
450   Node* constant_index = __ BytecodeOperandIdx(0);
451   Node* name = __ LoadConstantPoolEntry(constant_index);
452   Node* value = __ GetAccumulator();
453   Node* raw_slot = __ BytecodeOperandIdx(1);
454   Node* smi_slot = __ SmiTag(raw_slot);
455   Node* type_feedback_vector = __ LoadTypeFeedbackVector();
456   __ CallStub(ic.descriptor(), code_target, context, global, name, value,
457               smi_slot, type_feedback_vector);
458   __ Dispatch();
459 }
460 
461 // StaGlobalSloppy <name_index> <slot>
462 //
463 // Store the value in the accumulator into the global with name in constant pool
464 // entry <name_index> using FeedBackVector slot <slot> in sloppy mode.
DoStaGlobalSloppy(InterpreterAssembler * assembler)465 void Interpreter::DoStaGlobalSloppy(InterpreterAssembler* assembler) {
466   Callable ic = CodeFactory::StoreICInOptimizedCode(isolate_, SLOPPY);
467   DoStaGlobal(ic, assembler);
468 }
469 
470 // StaGlobalStrict <name_index> <slot>
471 //
472 // Store the value in the accumulator into the global with name in constant pool
473 // entry <name_index> using FeedBackVector slot <slot> in strict mode.
DoStaGlobalStrict(InterpreterAssembler * assembler)474 void Interpreter::DoStaGlobalStrict(InterpreterAssembler* assembler) {
475   Callable ic = CodeFactory::StoreICInOptimizedCode(isolate_, STRICT);
476   DoStaGlobal(ic, assembler);
477 }
478 
BuildLoadContextSlot(InterpreterAssembler * assembler)479 compiler::Node* Interpreter::BuildLoadContextSlot(
480     InterpreterAssembler* assembler) {
481   Node* reg_index = __ BytecodeOperandReg(0);
482   Node* context = __ LoadRegister(reg_index);
483   Node* slot_index = __ BytecodeOperandIdx(1);
484   return __ LoadContextSlot(context, slot_index);
485 }
486 
487 // LdaContextSlot <context> <slot_index>
488 //
489 // Load the object in |slot_index| of |context| into the accumulator.
DoLdaContextSlot(InterpreterAssembler * assembler)490 void Interpreter::DoLdaContextSlot(InterpreterAssembler* assembler) {
491   Node* result = BuildLoadContextSlot(assembler);
492   __ SetAccumulator(result);
493   __ Dispatch();
494 }
495 
496 // LdrContextSlot <context> <slot_index> <reg>
497 //
498 // Load the object in <slot_index> of <context> into register <reg>.
DoLdrContextSlot(InterpreterAssembler * assembler)499 void Interpreter::DoLdrContextSlot(InterpreterAssembler* assembler) {
500   Node* result = BuildLoadContextSlot(assembler);
501   Node* destination = __ BytecodeOperandReg(2);
502   __ StoreRegister(result, destination);
503   __ Dispatch();
504 }
505 
506 // StaContextSlot <context> <slot_index>
507 //
508 // Stores the object in the accumulator into |slot_index| of |context|.
DoStaContextSlot(InterpreterAssembler * assembler)509 void Interpreter::DoStaContextSlot(InterpreterAssembler* assembler) {
510   Node* value = __ GetAccumulator();
511   Node* reg_index = __ BytecodeOperandReg(0);
512   Node* context = __ LoadRegister(reg_index);
513   Node* slot_index = __ BytecodeOperandIdx(1);
514   __ StoreContextSlot(context, slot_index, value);
515   __ Dispatch();
516 }
517 
DoLdaLookupSlot(Runtime::FunctionId function_id,InterpreterAssembler * assembler)518 void Interpreter::DoLdaLookupSlot(Runtime::FunctionId function_id,
519                                   InterpreterAssembler* assembler) {
520   Node* index = __ BytecodeOperandIdx(0);
521   Node* name = __ LoadConstantPoolEntry(index);
522   Node* context = __ GetContext();
523   Node* result = __ CallRuntime(function_id, context, name);
524   __ SetAccumulator(result);
525   __ Dispatch();
526 }
527 
528 // LdaLookupSlot <name_index>
529 //
530 // Lookup the object with the name in constant pool entry |name_index|
531 // dynamically.
DoLdaLookupSlot(InterpreterAssembler * assembler)532 void Interpreter::DoLdaLookupSlot(InterpreterAssembler* assembler) {
533   DoLdaLookupSlot(Runtime::kLoadLookupSlot, assembler);
534 }
535 
536 // LdaLookupSlotInsideTypeof <name_index>
537 //
538 // Lookup the object with the name in constant pool entry |name_index|
539 // dynamically without causing a NoReferenceError.
DoLdaLookupSlotInsideTypeof(InterpreterAssembler * assembler)540 void Interpreter::DoLdaLookupSlotInsideTypeof(InterpreterAssembler* assembler) {
541   DoLdaLookupSlot(Runtime::kLoadLookupSlotInsideTypeof, assembler);
542 }
543 
DoStaLookupSlot(LanguageMode language_mode,InterpreterAssembler * assembler)544 void Interpreter::DoStaLookupSlot(LanguageMode language_mode,
545                                   InterpreterAssembler* assembler) {
546   Node* value = __ GetAccumulator();
547   Node* index = __ BytecodeOperandIdx(0);
548   Node* name = __ LoadConstantPoolEntry(index);
549   Node* context = __ GetContext();
550   Node* result = __ CallRuntime(is_strict(language_mode)
551                                     ? Runtime::kStoreLookupSlot_Strict
552                                     : Runtime::kStoreLookupSlot_Sloppy,
553                                 context, name, value);
554   __ SetAccumulator(result);
555   __ Dispatch();
556 }
557 
558 // StaLookupSlotSloppy <name_index>
559 //
560 // Store the object in accumulator to the object with the name in constant
561 // pool entry |name_index| in sloppy mode.
DoStaLookupSlotSloppy(InterpreterAssembler * assembler)562 void Interpreter::DoStaLookupSlotSloppy(InterpreterAssembler* assembler) {
563   DoStaLookupSlot(LanguageMode::SLOPPY, assembler);
564 }
565 
566 // StaLookupSlotStrict <name_index>
567 //
568 // Store the object in accumulator to the object with the name in constant
569 // pool entry |name_index| in strict mode.
DoStaLookupSlotStrict(InterpreterAssembler * assembler)570 void Interpreter::DoStaLookupSlotStrict(InterpreterAssembler* assembler) {
571   DoStaLookupSlot(LanguageMode::STRICT, assembler);
572 }
573 
BuildLoadNamedProperty(Callable ic,InterpreterAssembler * assembler)574 Node* Interpreter::BuildLoadNamedProperty(Callable ic,
575                                           InterpreterAssembler* assembler) {
576   Node* code_target = __ HeapConstant(ic.code());
577   Node* register_index = __ BytecodeOperandReg(0);
578   Node* object = __ LoadRegister(register_index);
579   Node* constant_index = __ BytecodeOperandIdx(1);
580   Node* name = __ LoadConstantPoolEntry(constant_index);
581   Node* raw_slot = __ BytecodeOperandIdx(2);
582   Node* smi_slot = __ SmiTag(raw_slot);
583   Node* type_feedback_vector = __ LoadTypeFeedbackVector();
584   Node* context = __ GetContext();
585   return __ CallStub(ic.descriptor(), code_target, context, object, name,
586                      smi_slot, type_feedback_vector);
587 }
588 
589 // LdaNamedProperty <object> <name_index> <slot>
590 //
591 // Calls the LoadIC at FeedBackVector slot <slot> for <object> and the name at
592 // constant pool entry <name_index>.
DoLdaNamedProperty(InterpreterAssembler * assembler)593 void Interpreter::DoLdaNamedProperty(InterpreterAssembler* assembler) {
594   Callable ic = CodeFactory::LoadICInOptimizedCode(isolate_);
595   Node* result = BuildLoadNamedProperty(ic, assembler);
596   __ SetAccumulator(result);
597   __ Dispatch();
598 }
599 
600 // LdrNamedProperty <object> <name_index> <slot> <reg>
601 //
602 // Calls the LoadIC at FeedBackVector slot <slot> for <object> and the name at
603 // constant pool entry <name_index> and puts the result into register <reg>.
DoLdrNamedProperty(InterpreterAssembler * assembler)604 void Interpreter::DoLdrNamedProperty(InterpreterAssembler* assembler) {
605   Callable ic = CodeFactory::LoadICInOptimizedCode(isolate_);
606   Node* result = BuildLoadNamedProperty(ic, assembler);
607   Node* destination = __ BytecodeOperandReg(3);
608   __ StoreRegister(result, destination);
609   __ Dispatch();
610 }
611 
BuildLoadKeyedProperty(Callable ic,InterpreterAssembler * assembler)612 Node* Interpreter::BuildLoadKeyedProperty(Callable ic,
613                                           InterpreterAssembler* assembler) {
614   Node* code_target = __ HeapConstant(ic.code());
615   Node* reg_index = __ BytecodeOperandReg(0);
616   Node* object = __ LoadRegister(reg_index);
617   Node* name = __ GetAccumulator();
618   Node* raw_slot = __ BytecodeOperandIdx(1);
619   Node* smi_slot = __ SmiTag(raw_slot);
620   Node* type_feedback_vector = __ LoadTypeFeedbackVector();
621   Node* context = __ GetContext();
622   return __ CallStub(ic.descriptor(), code_target, context, object, name,
623                      smi_slot, type_feedback_vector);
624 }
625 
626 // KeyedLoadIC <object> <slot>
627 //
628 // Calls the KeyedLoadIC at FeedBackVector slot <slot> for <object> and the key
629 // in the accumulator.
DoLdaKeyedProperty(InterpreterAssembler * assembler)630 void Interpreter::DoLdaKeyedProperty(InterpreterAssembler* assembler) {
631   Callable ic = CodeFactory::KeyedLoadICInOptimizedCode(isolate_);
632   Node* result = BuildLoadKeyedProperty(ic, assembler);
633   __ SetAccumulator(result);
634   __ Dispatch();
635 }
636 
637 // LdrKeyedProperty <object> <slot> <reg>
638 //
639 // Calls the KeyedLoadIC at FeedBackVector slot <slot> for <object> and the key
640 // in the accumulator and puts the result in register <reg>.
DoLdrKeyedProperty(InterpreterAssembler * assembler)641 void Interpreter::DoLdrKeyedProperty(InterpreterAssembler* assembler) {
642   Callable ic = CodeFactory::KeyedLoadICInOptimizedCode(isolate_);
643   Node* result = BuildLoadKeyedProperty(ic, assembler);
644   Node* destination = __ BytecodeOperandReg(2);
645   __ StoreRegister(result, destination);
646   __ Dispatch();
647 }
648 
DoStoreIC(Callable ic,InterpreterAssembler * assembler)649 void Interpreter::DoStoreIC(Callable ic, InterpreterAssembler* assembler) {
650   Node* code_target = __ HeapConstant(ic.code());
651   Node* object_reg_index = __ BytecodeOperandReg(0);
652   Node* object = __ LoadRegister(object_reg_index);
653   Node* constant_index = __ BytecodeOperandIdx(1);
654   Node* name = __ LoadConstantPoolEntry(constant_index);
655   Node* value = __ GetAccumulator();
656   Node* raw_slot = __ BytecodeOperandIdx(2);
657   Node* smi_slot = __ SmiTag(raw_slot);
658   Node* type_feedback_vector = __ LoadTypeFeedbackVector();
659   Node* context = __ GetContext();
660   __ CallStub(ic.descriptor(), code_target, context, object, name, value,
661               smi_slot, type_feedback_vector);
662   __ Dispatch();
663 }
664 
665 // StaNamedPropertySloppy <object> <name_index> <slot>
666 //
667 // Calls the sloppy mode StoreIC at FeedBackVector slot <slot> for <object> and
668 // the name in constant pool entry <name_index> with the value in the
669 // accumulator.
DoStaNamedPropertySloppy(InterpreterAssembler * assembler)670 void Interpreter::DoStaNamedPropertySloppy(InterpreterAssembler* assembler) {
671   Callable ic = CodeFactory::StoreICInOptimizedCode(isolate_, SLOPPY);
672   DoStoreIC(ic, assembler);
673 }
674 
675 // StaNamedPropertyStrict <object> <name_index> <slot>
676 //
677 // Calls the strict mode StoreIC at FeedBackVector slot <slot> for <object> and
678 // the name in constant pool entry <name_index> with the value in the
679 // accumulator.
DoStaNamedPropertyStrict(InterpreterAssembler * assembler)680 void Interpreter::DoStaNamedPropertyStrict(InterpreterAssembler* assembler) {
681   Callable ic = CodeFactory::StoreICInOptimizedCode(isolate_, STRICT);
682   DoStoreIC(ic, assembler);
683 }
684 
DoKeyedStoreIC(Callable ic,InterpreterAssembler * assembler)685 void Interpreter::DoKeyedStoreIC(Callable ic, InterpreterAssembler* assembler) {
686   Node* code_target = __ HeapConstant(ic.code());
687   Node* object_reg_index = __ BytecodeOperandReg(0);
688   Node* object = __ LoadRegister(object_reg_index);
689   Node* name_reg_index = __ BytecodeOperandReg(1);
690   Node* name = __ LoadRegister(name_reg_index);
691   Node* value = __ GetAccumulator();
692   Node* raw_slot = __ BytecodeOperandIdx(2);
693   Node* smi_slot = __ SmiTag(raw_slot);
694   Node* type_feedback_vector = __ LoadTypeFeedbackVector();
695   Node* context = __ GetContext();
696   __ CallStub(ic.descriptor(), code_target, context, object, name, value,
697               smi_slot, type_feedback_vector);
698   __ Dispatch();
699 }
700 
701 // StaKeyedPropertySloppy <object> <key> <slot>
702 //
703 // Calls the sloppy mode KeyStoreIC at FeedBackVector slot <slot> for <object>
704 // and the key <key> with the value in the accumulator.
DoStaKeyedPropertySloppy(InterpreterAssembler * assembler)705 void Interpreter::DoStaKeyedPropertySloppy(InterpreterAssembler* assembler) {
706   Callable ic = CodeFactory::KeyedStoreICInOptimizedCode(isolate_, SLOPPY);
707   DoKeyedStoreIC(ic, assembler);
708 }
709 
710 // StaKeyedPropertyStrict <object> <key> <slot>
711 //
712 // Calls the strict mode KeyStoreIC at FeedBackVector slot <slot> for <object>
713 // and the key <key> with the value in the accumulator.
DoStaKeyedPropertyStrict(InterpreterAssembler * assembler)714 void Interpreter::DoStaKeyedPropertyStrict(InterpreterAssembler* assembler) {
715   Callable ic = CodeFactory::KeyedStoreICInOptimizedCode(isolate_, STRICT);
716   DoKeyedStoreIC(ic, assembler);
717 }
718 
719 // PushContext <context>
720 //
721 // Saves the current context in <context>, and pushes the accumulator as the
722 // new current context.
DoPushContext(InterpreterAssembler * assembler)723 void Interpreter::DoPushContext(InterpreterAssembler* assembler) {
724   Node* reg_index = __ BytecodeOperandReg(0);
725   Node* new_context = __ GetAccumulator();
726   Node* old_context = __ GetContext();
727   __ StoreRegister(old_context, reg_index);
728   __ SetContext(new_context);
729   __ Dispatch();
730 }
731 
732 // PopContext <context>
733 //
734 // Pops the current context and sets <context> as the new context.
DoPopContext(InterpreterAssembler * assembler)735 void Interpreter::DoPopContext(InterpreterAssembler* assembler) {
736   Node* reg_index = __ BytecodeOperandReg(0);
737   Node* context = __ LoadRegister(reg_index);
738   __ SetContext(context);
739   __ Dispatch();
740 }
741 
742 template <class Generator>
DoBinaryOp(InterpreterAssembler * assembler)743 void Interpreter::DoBinaryOp(InterpreterAssembler* assembler) {
744   Node* reg_index = __ BytecodeOperandReg(0);
745   Node* lhs = __ LoadRegister(reg_index);
746   Node* rhs = __ GetAccumulator();
747   Node* context = __ GetContext();
748   Node* result = Generator::Generate(assembler, lhs, rhs, context);
749   __ SetAccumulator(result);
750   __ Dispatch();
751 }
752 
753 // Add <src>
754 //
755 // Add register <src> to accumulator.
DoAdd(InterpreterAssembler * assembler)756 void Interpreter::DoAdd(InterpreterAssembler* assembler) {
757   DoBinaryOp<AddStub>(assembler);
758 }
759 
760 // Sub <src>
761 //
762 // Subtract register <src> from accumulator.
DoSub(InterpreterAssembler * assembler)763 void Interpreter::DoSub(InterpreterAssembler* assembler) {
764   DoBinaryOp<SubtractStub>(assembler);
765 }
766 
767 // Mul <src>
768 //
769 // Multiply accumulator by register <src>.
DoMul(InterpreterAssembler * assembler)770 void Interpreter::DoMul(InterpreterAssembler* assembler) {
771   DoBinaryOp<MultiplyStub>(assembler);
772 }
773 
774 // Div <src>
775 //
776 // Divide register <src> by accumulator.
DoDiv(InterpreterAssembler * assembler)777 void Interpreter::DoDiv(InterpreterAssembler* assembler) {
778   DoBinaryOp<DivideStub>(assembler);
779 }
780 
781 // Mod <src>
782 //
783 // Modulo register <src> by accumulator.
DoMod(InterpreterAssembler * assembler)784 void Interpreter::DoMod(InterpreterAssembler* assembler) {
785   DoBinaryOp<ModulusStub>(assembler);
786 }
787 
788 // BitwiseOr <src>
789 //
790 // BitwiseOr register <src> to accumulator.
DoBitwiseOr(InterpreterAssembler * assembler)791 void Interpreter::DoBitwiseOr(InterpreterAssembler* assembler) {
792   DoBinaryOp<BitwiseOrStub>(assembler);
793 }
794 
795 // BitwiseXor <src>
796 //
797 // BitwiseXor register <src> to accumulator.
DoBitwiseXor(InterpreterAssembler * assembler)798 void Interpreter::DoBitwiseXor(InterpreterAssembler* assembler) {
799   DoBinaryOp<BitwiseXorStub>(assembler);
800 }
801 
802 // BitwiseAnd <src>
803 //
804 // BitwiseAnd register <src> to accumulator.
DoBitwiseAnd(InterpreterAssembler * assembler)805 void Interpreter::DoBitwiseAnd(InterpreterAssembler* assembler) {
806   DoBinaryOp<BitwiseAndStub>(assembler);
807 }
808 
809 // ShiftLeft <src>
810 //
811 // Left shifts register <src> by the count specified in the accumulator.
812 // Register <src> is converted to an int32 and the accumulator to uint32
813 // before the operation. 5 lsb bits from the accumulator are used as count
814 // i.e. <src> << (accumulator & 0x1F).
DoShiftLeft(InterpreterAssembler * assembler)815 void Interpreter::DoShiftLeft(InterpreterAssembler* assembler) {
816   DoBinaryOp<ShiftLeftStub>(assembler);
817 }
818 
819 // ShiftRight <src>
820 //
821 // Right shifts register <src> by the count specified in the accumulator.
822 // Result is sign extended. Register <src> is converted to an int32 and the
823 // accumulator to uint32 before the operation. 5 lsb bits from the accumulator
824 // are used as count i.e. <src> >> (accumulator & 0x1F).
DoShiftRight(InterpreterAssembler * assembler)825 void Interpreter::DoShiftRight(InterpreterAssembler* assembler) {
826   DoBinaryOp<ShiftRightStub>(assembler);
827 }
828 
829 // ShiftRightLogical <src>
830 //
831 // Right Shifts register <src> by the count specified in the accumulator.
832 // Result is zero-filled. The accumulator and register <src> are converted to
833 // uint32 before the operation 5 lsb bits from the accumulator are used as
834 // count i.e. <src> << (accumulator & 0x1F).
DoShiftRightLogical(InterpreterAssembler * assembler)835 void Interpreter::DoShiftRightLogical(InterpreterAssembler* assembler) {
836   DoBinaryOp<ShiftRightLogicalStub>(assembler);
837 }
838 
DoUnaryOp(Callable callable,InterpreterAssembler * assembler)839 void Interpreter::DoUnaryOp(Callable callable,
840                             InterpreterAssembler* assembler) {
841   Node* target = __ HeapConstant(callable.code());
842   Node* accumulator = __ GetAccumulator();
843   Node* context = __ GetContext();
844   Node* result =
845       __ CallStub(callable.descriptor(), target, context, accumulator);
846   __ SetAccumulator(result);
847   __ Dispatch();
848 }
849 
850 template <class Generator>
DoUnaryOp(InterpreterAssembler * assembler)851 void Interpreter::DoUnaryOp(InterpreterAssembler* assembler) {
852   Node* value = __ GetAccumulator();
853   Node* context = __ GetContext();
854   Node* result = Generator::Generate(assembler, value, context);
855   __ SetAccumulator(result);
856   __ Dispatch();
857 }
858 
859 // ToName
860 //
861 // Cast the object referenced by the accumulator to a name.
DoToName(InterpreterAssembler * assembler)862 void Interpreter::DoToName(InterpreterAssembler* assembler) {
863   DoUnaryOp(CodeFactory::ToName(isolate_), assembler);
864 }
865 
866 // ToNumber
867 //
868 // Cast the object referenced by the accumulator to a number.
DoToNumber(InterpreterAssembler * assembler)869 void Interpreter::DoToNumber(InterpreterAssembler* assembler) {
870   DoUnaryOp(CodeFactory::ToNumber(isolate_), assembler);
871 }
872 
873 // ToObject
874 //
875 // Cast the object referenced by the accumulator to a JSObject.
DoToObject(InterpreterAssembler * assembler)876 void Interpreter::DoToObject(InterpreterAssembler* assembler) {
877   DoUnaryOp(CodeFactory::ToObject(isolate_), assembler);
878 }
879 
880 // Inc
881 //
882 // Increments value in the accumulator by one.
DoInc(InterpreterAssembler * assembler)883 void Interpreter::DoInc(InterpreterAssembler* assembler) {
884   DoUnaryOp<IncStub>(assembler);
885 }
886 
887 // Dec
888 //
889 // Decrements value in the accumulator by one.
DoDec(InterpreterAssembler * assembler)890 void Interpreter::DoDec(InterpreterAssembler* assembler) {
891   DoUnaryOp<DecStub>(assembler);
892 }
893 
BuildToBoolean(Node * value,InterpreterAssembler * assembler)894 Node* Interpreter::BuildToBoolean(Node* value,
895                                   InterpreterAssembler* assembler) {
896   Node* context = __ GetContext();
897   return ToBooleanStub::Generate(assembler, value, context);
898 }
899 
BuildLogicalNot(Node * value,InterpreterAssembler * assembler)900 Node* Interpreter::BuildLogicalNot(Node* value,
901                                    InterpreterAssembler* assembler) {
902   Variable result(assembler, MachineRepresentation::kTagged);
903   Label if_true(assembler), if_false(assembler), end(assembler);
904   Node* true_value = __ BooleanConstant(true);
905   Node* false_value = __ BooleanConstant(false);
906   __ BranchIfWordEqual(value, true_value, &if_true, &if_false);
907   __ Bind(&if_true);
908   {
909     result.Bind(false_value);
910     __ Goto(&end);
911   }
912   __ Bind(&if_false);
913   {
914     if (FLAG_debug_code) {
915       __ AbortIfWordNotEqual(value, false_value,
916                              BailoutReason::kExpectedBooleanValue);
917     }
918     result.Bind(true_value);
919     __ Goto(&end);
920   }
921   __ Bind(&end);
922   return result.value();
923 }
924 
925 // LogicalNot
926 //
927 // Perform logical-not on the accumulator, first casting the
928 // accumulator to a boolean value if required.
929 // ToBooleanLogicalNot
DoToBooleanLogicalNot(InterpreterAssembler * assembler)930 void Interpreter::DoToBooleanLogicalNot(InterpreterAssembler* assembler) {
931   Node* value = __ GetAccumulator();
932   Node* to_boolean_value = BuildToBoolean(value, assembler);
933   Node* result = BuildLogicalNot(to_boolean_value, assembler);
934   __ SetAccumulator(result);
935   __ Dispatch();
936 }
937 
938 // LogicalNot
939 //
940 // Perform logical-not on the accumulator, which must already be a boolean
941 // value.
DoLogicalNot(InterpreterAssembler * assembler)942 void Interpreter::DoLogicalNot(InterpreterAssembler* assembler) {
943   Node* value = __ GetAccumulator();
944   Node* result = BuildLogicalNot(value, assembler);
945   __ SetAccumulator(result);
946   __ Dispatch();
947 }
948 
949 // TypeOf
950 //
951 // Load the accumulator with the string representating type of the
952 // object in the accumulator.
DoTypeOf(InterpreterAssembler * assembler)953 void Interpreter::DoTypeOf(InterpreterAssembler* assembler) {
954   DoUnaryOp(CodeFactory::Typeof(isolate_), assembler);
955 }
956 
DoDelete(Runtime::FunctionId function_id,InterpreterAssembler * assembler)957 void Interpreter::DoDelete(Runtime::FunctionId function_id,
958                            InterpreterAssembler* assembler) {
959   Node* reg_index = __ BytecodeOperandReg(0);
960   Node* object = __ LoadRegister(reg_index);
961   Node* key = __ GetAccumulator();
962   Node* context = __ GetContext();
963   Node* result = __ CallRuntime(function_id, context, object, key);
964   __ SetAccumulator(result);
965   __ Dispatch();
966 }
967 
968 // DeletePropertyStrict
969 //
970 // Delete the property specified in the accumulator from the object
971 // referenced by the register operand following strict mode semantics.
DoDeletePropertyStrict(InterpreterAssembler * assembler)972 void Interpreter::DoDeletePropertyStrict(InterpreterAssembler* assembler) {
973   DoDelete(Runtime::kDeleteProperty_Strict, assembler);
974 }
975 
976 // DeletePropertySloppy
977 //
978 // Delete the property specified in the accumulator from the object
979 // referenced by the register operand following sloppy mode semantics.
DoDeletePropertySloppy(InterpreterAssembler * assembler)980 void Interpreter::DoDeletePropertySloppy(InterpreterAssembler* assembler) {
981   DoDelete(Runtime::kDeleteProperty_Sloppy, assembler);
982 }
983 
DoJSCall(InterpreterAssembler * assembler,TailCallMode tail_call_mode)984 void Interpreter::DoJSCall(InterpreterAssembler* assembler,
985                            TailCallMode tail_call_mode) {
986   Node* function_reg = __ BytecodeOperandReg(0);
987   Node* function = __ LoadRegister(function_reg);
988   Node* receiver_reg = __ BytecodeOperandReg(1);
989   Node* receiver_arg = __ RegisterLocation(receiver_reg);
990   Node* receiver_args_count = __ BytecodeOperandCount(2);
991   Node* receiver_count = __ Int32Constant(1);
992   Node* args_count = __ Int32Sub(receiver_args_count, receiver_count);
993   Node* context = __ GetContext();
994   // TODO(rmcilroy): Use the call type feedback slot to call via CallStub.
995   Node* result =
996       __ CallJS(function, context, receiver_arg, args_count, tail_call_mode);
997   __ SetAccumulator(result);
998   __ Dispatch();
999 }
1000 
1001 // Call <callable> <receiver> <arg_count>
1002 //
1003 // Call a JSfunction or Callable in |callable| with the |receiver| and
1004 // |arg_count| arguments in subsequent registers.
DoCall(InterpreterAssembler * assembler)1005 void Interpreter::DoCall(InterpreterAssembler* assembler) {
1006   DoJSCall(assembler, TailCallMode::kDisallow);
1007 }
1008 
1009 // TailCall <callable> <receiver> <arg_count>
1010 //
1011 // Tail call a JSfunction or Callable in |callable| with the |receiver| and
1012 // |arg_count| arguments in subsequent registers.
DoTailCall(InterpreterAssembler * assembler)1013 void Interpreter::DoTailCall(InterpreterAssembler* assembler) {
1014   DoJSCall(assembler, TailCallMode::kAllow);
1015 }
1016 
DoCallRuntimeCommon(InterpreterAssembler * assembler)1017 void Interpreter::DoCallRuntimeCommon(InterpreterAssembler* assembler) {
1018   Node* function_id = __ BytecodeOperandRuntimeId(0);
1019   Node* first_arg_reg = __ BytecodeOperandReg(1);
1020   Node* first_arg = __ RegisterLocation(first_arg_reg);
1021   Node* args_count = __ BytecodeOperandCount(2);
1022   Node* context = __ GetContext();
1023   Node* result = __ CallRuntimeN(function_id, context, first_arg, args_count);
1024   __ SetAccumulator(result);
1025   __ Dispatch();
1026 }
1027 
1028 // CallRuntime <function_id> <first_arg> <arg_count>
1029 //
1030 // Call the runtime function |function_id| with the first argument in
1031 // register |first_arg| and |arg_count| arguments in subsequent
1032 // registers.
DoCallRuntime(InterpreterAssembler * assembler)1033 void Interpreter::DoCallRuntime(InterpreterAssembler* assembler) {
1034   DoCallRuntimeCommon(assembler);
1035 }
1036 
1037 // InvokeIntrinsic <function_id> <first_arg> <arg_count>
1038 //
1039 // Implements the semantic equivalent of calling the runtime function
1040 // |function_id| with the first argument in |first_arg| and |arg_count|
1041 // arguments in subsequent registers.
DoInvokeIntrinsic(InterpreterAssembler * assembler)1042 void Interpreter::DoInvokeIntrinsic(InterpreterAssembler* assembler) {
1043   Node* function_id = __ BytecodeOperandIntrinsicId(0);
1044   Node* first_arg_reg = __ BytecodeOperandReg(1);
1045   Node* arg_count = __ BytecodeOperandCount(2);
1046   Node* context = __ GetContext();
1047   IntrinsicsHelper helper(assembler);
1048   Node* result =
1049       helper.InvokeIntrinsic(function_id, context, first_arg_reg, arg_count);
1050   __ SetAccumulator(result);
1051   __ Dispatch();
1052 }
1053 
DoCallRuntimeForPairCommon(InterpreterAssembler * assembler)1054 void Interpreter::DoCallRuntimeForPairCommon(InterpreterAssembler* assembler) {
1055   // Call the runtime function.
1056   Node* function_id = __ BytecodeOperandRuntimeId(0);
1057   Node* first_arg_reg = __ BytecodeOperandReg(1);
1058   Node* first_arg = __ RegisterLocation(first_arg_reg);
1059   Node* args_count = __ BytecodeOperandCount(2);
1060   Node* context = __ GetContext();
1061   Node* result_pair =
1062       __ CallRuntimeN(function_id, context, first_arg, args_count, 2);
1063 
1064   // Store the results in <first_return> and <first_return + 1>
1065   Node* first_return_reg = __ BytecodeOperandReg(3);
1066   Node* second_return_reg = __ NextRegister(first_return_reg);
1067   Node* result0 = __ Projection(0, result_pair);
1068   Node* result1 = __ Projection(1, result_pair);
1069   __ StoreRegister(result0, first_return_reg);
1070   __ StoreRegister(result1, second_return_reg);
1071   __ Dispatch();
1072 }
1073 
1074 // CallRuntimeForPair <function_id> <first_arg> <arg_count> <first_return>
1075 //
1076 // Call the runtime function |function_id| which returns a pair, with the
1077 // first argument in register |first_arg| and |arg_count| arguments in
1078 // subsequent registers. Returns the result in <first_return> and
1079 // <first_return + 1>
DoCallRuntimeForPair(InterpreterAssembler * assembler)1080 void Interpreter::DoCallRuntimeForPair(InterpreterAssembler* assembler) {
1081   DoCallRuntimeForPairCommon(assembler);
1082 }
1083 
DoCallJSRuntimeCommon(InterpreterAssembler * assembler)1084 void Interpreter::DoCallJSRuntimeCommon(InterpreterAssembler* assembler) {
1085   Node* context_index = __ BytecodeOperandIdx(0);
1086   Node* receiver_reg = __ BytecodeOperandReg(1);
1087   Node* first_arg = __ RegisterLocation(receiver_reg);
1088   Node* receiver_args_count = __ BytecodeOperandCount(2);
1089   Node* receiver_count = __ Int32Constant(1);
1090   Node* args_count = __ Int32Sub(receiver_args_count, receiver_count);
1091 
1092   // Get the function to call from the native context.
1093   Node* context = __ GetContext();
1094   Node* native_context =
1095       __ LoadContextSlot(context, Context::NATIVE_CONTEXT_INDEX);
1096   Node* function = __ LoadContextSlot(native_context, context_index);
1097 
1098   // Call the function.
1099   Node* result = __ CallJS(function, context, first_arg, args_count,
1100                            TailCallMode::kDisallow);
1101   __ SetAccumulator(result);
1102   __ Dispatch();
1103 }
1104 
1105 // CallJSRuntime <context_index> <receiver> <arg_count>
1106 //
1107 // Call the JS runtime function that has the |context_index| with the receiver
1108 // in register |receiver| and |arg_count| arguments in subsequent registers.
DoCallJSRuntime(InterpreterAssembler * assembler)1109 void Interpreter::DoCallJSRuntime(InterpreterAssembler* assembler) {
1110   DoCallJSRuntimeCommon(assembler);
1111 }
1112 
DoCallConstruct(InterpreterAssembler * assembler)1113 void Interpreter::DoCallConstruct(InterpreterAssembler* assembler) {
1114   Callable ic = CodeFactory::InterpreterPushArgsAndConstruct(isolate_);
1115   Node* new_target = __ GetAccumulator();
1116   Node* constructor_reg = __ BytecodeOperandReg(0);
1117   Node* constructor = __ LoadRegister(constructor_reg);
1118   Node* first_arg_reg = __ BytecodeOperandReg(1);
1119   Node* first_arg = __ RegisterLocation(first_arg_reg);
1120   Node* args_count = __ BytecodeOperandCount(2);
1121   Node* context = __ GetContext();
1122   Node* result =
1123       __ CallConstruct(constructor, context, new_target, first_arg, args_count);
1124   __ SetAccumulator(result);
1125   __ Dispatch();
1126 }
1127 
1128 // New <constructor> <first_arg> <arg_count>
1129 //
1130 // Call operator new with |constructor| and the first argument in
1131 // register |first_arg| and |arg_count| arguments in subsequent
1132 // registers. The new.target is in the accumulator.
1133 //
DoNew(InterpreterAssembler * assembler)1134 void Interpreter::DoNew(InterpreterAssembler* assembler) {
1135   DoCallConstruct(assembler);
1136 }
1137 
1138 // TestEqual <src>
1139 //
1140 // Test if the value in the <src> register equals the accumulator.
DoTestEqual(InterpreterAssembler * assembler)1141 void Interpreter::DoTestEqual(InterpreterAssembler* assembler) {
1142   DoBinaryOp<EqualStub>(assembler);
1143 }
1144 
1145 // TestNotEqual <src>
1146 //
1147 // Test if the value in the <src> register is not equal to the accumulator.
DoTestNotEqual(InterpreterAssembler * assembler)1148 void Interpreter::DoTestNotEqual(InterpreterAssembler* assembler) {
1149   DoBinaryOp<NotEqualStub>(assembler);
1150 }
1151 
1152 // TestEqualStrict <src>
1153 //
1154 // Test if the value in the <src> register is strictly equal to the accumulator.
DoTestEqualStrict(InterpreterAssembler * assembler)1155 void Interpreter::DoTestEqualStrict(InterpreterAssembler* assembler) {
1156   DoBinaryOp<StrictEqualStub>(assembler);
1157 }
1158 
1159 // TestLessThan <src>
1160 //
1161 // Test if the value in the <src> register is less than the accumulator.
DoTestLessThan(InterpreterAssembler * assembler)1162 void Interpreter::DoTestLessThan(InterpreterAssembler* assembler) {
1163   DoBinaryOp<LessThanStub>(assembler);
1164 }
1165 
1166 // TestGreaterThan <src>
1167 //
1168 // Test if the value in the <src> register is greater than the accumulator.
DoTestGreaterThan(InterpreterAssembler * assembler)1169 void Interpreter::DoTestGreaterThan(InterpreterAssembler* assembler) {
1170   DoBinaryOp<GreaterThanStub>(assembler);
1171 }
1172 
1173 // TestLessThanOrEqual <src>
1174 //
1175 // Test if the value in the <src> register is less than or equal to the
1176 // accumulator.
DoTestLessThanOrEqual(InterpreterAssembler * assembler)1177 void Interpreter::DoTestLessThanOrEqual(InterpreterAssembler* assembler) {
1178   DoBinaryOp<LessThanOrEqualStub>(assembler);
1179 }
1180 
1181 // TestGreaterThanOrEqual <src>
1182 //
1183 // Test if the value in the <src> register is greater than or equal to the
1184 // accumulator.
DoTestGreaterThanOrEqual(InterpreterAssembler * assembler)1185 void Interpreter::DoTestGreaterThanOrEqual(InterpreterAssembler* assembler) {
1186   DoBinaryOp<GreaterThanOrEqualStub>(assembler);
1187 }
1188 
1189 // TestIn <src>
1190 //
1191 // Test if the object referenced by the register operand is a property of the
1192 // object referenced by the accumulator.
DoTestIn(InterpreterAssembler * assembler)1193 void Interpreter::DoTestIn(InterpreterAssembler* assembler) {
1194   DoBinaryOp<HasPropertyStub>(assembler);
1195 }
1196 
1197 // TestInstanceOf <src>
1198 //
1199 // Test if the object referenced by the <src> register is an an instance of type
1200 // referenced by the accumulator.
DoTestInstanceOf(InterpreterAssembler * assembler)1201 void Interpreter::DoTestInstanceOf(InterpreterAssembler* assembler) {
1202   DoBinaryOp<InstanceOfStub>(assembler);
1203 }
1204 
1205 // Jump <imm>
1206 //
1207 // Jump by number of bytes represented by the immediate operand |imm|.
DoJump(InterpreterAssembler * assembler)1208 void Interpreter::DoJump(InterpreterAssembler* assembler) {
1209   Node* relative_jump = __ BytecodeOperandImm(0);
1210   __ Jump(relative_jump);
1211 }
1212 
1213 // JumpConstant <idx>
1214 //
1215 // Jump by number of bytes in the Smi in the |idx| entry in the constant pool.
DoJumpConstant(InterpreterAssembler * assembler)1216 void Interpreter::DoJumpConstant(InterpreterAssembler* assembler) {
1217   Node* index = __ BytecodeOperandIdx(0);
1218   Node* constant = __ LoadConstantPoolEntry(index);
1219   Node* relative_jump = __ SmiUntag(constant);
1220   __ Jump(relative_jump);
1221 }
1222 
1223 // JumpIfTrue <imm>
1224 //
1225 // Jump by number of bytes represented by an immediate operand if the
1226 // accumulator contains true.
DoJumpIfTrue(InterpreterAssembler * assembler)1227 void Interpreter::DoJumpIfTrue(InterpreterAssembler* assembler) {
1228   Node* accumulator = __ GetAccumulator();
1229   Node* relative_jump = __ BytecodeOperandImm(0);
1230   Node* true_value = __ BooleanConstant(true);
1231   __ JumpIfWordEqual(accumulator, true_value, relative_jump);
1232 }
1233 
1234 // JumpIfTrueConstant <idx>
1235 //
1236 // Jump by number of bytes in the Smi in the |idx| entry in the constant pool
1237 // if the accumulator contains true.
DoJumpIfTrueConstant(InterpreterAssembler * assembler)1238 void Interpreter::DoJumpIfTrueConstant(InterpreterAssembler* assembler) {
1239   Node* accumulator = __ GetAccumulator();
1240   Node* index = __ BytecodeOperandIdx(0);
1241   Node* constant = __ LoadConstantPoolEntry(index);
1242   Node* relative_jump = __ SmiUntag(constant);
1243   Node* true_value = __ BooleanConstant(true);
1244   __ JumpIfWordEqual(accumulator, true_value, relative_jump);
1245 }
1246 
1247 // JumpIfFalse <imm>
1248 //
1249 // Jump by number of bytes represented by an immediate operand if the
1250 // accumulator contains false.
DoJumpIfFalse(InterpreterAssembler * assembler)1251 void Interpreter::DoJumpIfFalse(InterpreterAssembler* assembler) {
1252   Node* accumulator = __ GetAccumulator();
1253   Node* relative_jump = __ BytecodeOperandImm(0);
1254   Node* false_value = __ BooleanConstant(false);
1255   __ JumpIfWordEqual(accumulator, false_value, relative_jump);
1256 }
1257 
1258 // JumpIfFalseConstant <idx>
1259 //
1260 // Jump by number of bytes in the Smi in the |idx| entry in the constant pool
1261 // if the accumulator contains false.
DoJumpIfFalseConstant(InterpreterAssembler * assembler)1262 void Interpreter::DoJumpIfFalseConstant(InterpreterAssembler* assembler) {
1263   Node* accumulator = __ GetAccumulator();
1264   Node* index = __ BytecodeOperandIdx(0);
1265   Node* constant = __ LoadConstantPoolEntry(index);
1266   Node* relative_jump = __ SmiUntag(constant);
1267   Node* false_value = __ BooleanConstant(false);
1268   __ JumpIfWordEqual(accumulator, false_value, relative_jump);
1269 }
1270 
1271 // JumpIfToBooleanTrue <imm>
1272 //
1273 // Jump by number of bytes represented by an immediate operand if the object
1274 // referenced by the accumulator is true when the object is cast to boolean.
DoJumpIfToBooleanTrue(InterpreterAssembler * assembler)1275 void Interpreter::DoJumpIfToBooleanTrue(InterpreterAssembler* assembler) {
1276   Node* accumulator = __ GetAccumulator();
1277   Node* to_boolean_value = BuildToBoolean(accumulator, assembler);
1278   Node* relative_jump = __ BytecodeOperandImm(0);
1279   Node* true_value = __ BooleanConstant(true);
1280   __ JumpIfWordEqual(to_boolean_value, true_value, relative_jump);
1281 }
1282 
1283 // JumpIfToBooleanTrueConstant <idx>
1284 //
1285 // Jump by number of bytes in the Smi in the |idx| entry in the constant pool
1286 // if the object referenced by the accumulator is true when the object is cast
1287 // to boolean.
DoJumpIfToBooleanTrueConstant(InterpreterAssembler * assembler)1288 void Interpreter::DoJumpIfToBooleanTrueConstant(
1289     InterpreterAssembler* assembler) {
1290   Node* accumulator = __ GetAccumulator();
1291   Node* to_boolean_value = BuildToBoolean(accumulator, assembler);
1292   Node* index = __ BytecodeOperandIdx(0);
1293   Node* constant = __ LoadConstantPoolEntry(index);
1294   Node* relative_jump = __ SmiUntag(constant);
1295   Node* true_value = __ BooleanConstant(true);
1296   __ JumpIfWordEqual(to_boolean_value, true_value, relative_jump);
1297 }
1298 
1299 // JumpIfToBooleanFalse <imm>
1300 //
1301 // Jump by number of bytes represented by an immediate operand if the object
1302 // referenced by the accumulator is false when the object is cast to boolean.
DoJumpIfToBooleanFalse(InterpreterAssembler * assembler)1303 void Interpreter::DoJumpIfToBooleanFalse(InterpreterAssembler* assembler) {
1304   Node* accumulator = __ GetAccumulator();
1305   Node* to_boolean_value = BuildToBoolean(accumulator, assembler);
1306   Node* relative_jump = __ BytecodeOperandImm(0);
1307   Node* false_value = __ BooleanConstant(false);
1308   __ JumpIfWordEqual(to_boolean_value, false_value, relative_jump);
1309 }
1310 
1311 // JumpIfToBooleanFalseConstant <idx>
1312 //
1313 // Jump by number of bytes in the Smi in the |idx| entry in the constant pool
1314 // if the object referenced by the accumulator is false when the object is cast
1315 // to boolean.
DoJumpIfToBooleanFalseConstant(InterpreterAssembler * assembler)1316 void Interpreter::DoJumpIfToBooleanFalseConstant(
1317     InterpreterAssembler* assembler) {
1318   Node* accumulator = __ GetAccumulator();
1319   Node* to_boolean_value = BuildToBoolean(accumulator, assembler);
1320   Node* index = __ BytecodeOperandIdx(0);
1321   Node* constant = __ LoadConstantPoolEntry(index);
1322   Node* relative_jump = __ SmiUntag(constant);
1323   Node* false_value = __ BooleanConstant(false);
1324   __ JumpIfWordEqual(to_boolean_value, false_value, relative_jump);
1325 }
1326 
1327 // JumpIfNull <imm>
1328 //
1329 // Jump by number of bytes represented by an immediate operand if the object
1330 // referenced by the accumulator is the null constant.
DoJumpIfNull(InterpreterAssembler * assembler)1331 void Interpreter::DoJumpIfNull(InterpreterAssembler* assembler) {
1332   Node* accumulator = __ GetAccumulator();
1333   Node* null_value = __ HeapConstant(isolate_->factory()->null_value());
1334   Node* relative_jump = __ BytecodeOperandImm(0);
1335   __ JumpIfWordEqual(accumulator, null_value, relative_jump);
1336 }
1337 
1338 // JumpIfNullConstant <idx>
1339 //
1340 // Jump by number of bytes in the Smi in the |idx| entry in the constant pool
1341 // if the object referenced by the accumulator is the null constant.
DoJumpIfNullConstant(InterpreterAssembler * assembler)1342 void Interpreter::DoJumpIfNullConstant(InterpreterAssembler* assembler) {
1343   Node* accumulator = __ GetAccumulator();
1344   Node* null_value = __ HeapConstant(isolate_->factory()->null_value());
1345   Node* index = __ BytecodeOperandIdx(0);
1346   Node* constant = __ LoadConstantPoolEntry(index);
1347   Node* relative_jump = __ SmiUntag(constant);
1348   __ JumpIfWordEqual(accumulator, null_value, relative_jump);
1349 }
1350 
1351 // JumpIfUndefined <imm>
1352 //
1353 // Jump by number of bytes represented by an immediate operand if the object
1354 // referenced by the accumulator is the undefined constant.
DoJumpIfUndefined(InterpreterAssembler * assembler)1355 void Interpreter::DoJumpIfUndefined(InterpreterAssembler* assembler) {
1356   Node* accumulator = __ GetAccumulator();
1357   Node* undefined_value =
1358       __ HeapConstant(isolate_->factory()->undefined_value());
1359   Node* relative_jump = __ BytecodeOperandImm(0);
1360   __ JumpIfWordEqual(accumulator, undefined_value, relative_jump);
1361 }
1362 
1363 // JumpIfUndefinedConstant <idx>
1364 //
1365 // Jump by number of bytes in the Smi in the |idx| entry in the constant pool
1366 // if the object referenced by the accumulator is the undefined constant.
DoJumpIfUndefinedConstant(InterpreterAssembler * assembler)1367 void Interpreter::DoJumpIfUndefinedConstant(InterpreterAssembler* assembler) {
1368   Node* accumulator = __ GetAccumulator();
1369   Node* undefined_value =
1370       __ HeapConstant(isolate_->factory()->undefined_value());
1371   Node* index = __ BytecodeOperandIdx(0);
1372   Node* constant = __ LoadConstantPoolEntry(index);
1373   Node* relative_jump = __ SmiUntag(constant);
1374   __ JumpIfWordEqual(accumulator, undefined_value, relative_jump);
1375 }
1376 
1377 // JumpIfNotHole <imm>
1378 //
1379 // Jump by number of bytes represented by an immediate operand if the object
1380 // referenced by the accumulator is the hole.
DoJumpIfNotHole(InterpreterAssembler * assembler)1381 void Interpreter::DoJumpIfNotHole(InterpreterAssembler* assembler) {
1382   Node* accumulator = __ GetAccumulator();
1383   Node* the_hole_value = __ HeapConstant(isolate_->factory()->the_hole_value());
1384   Node* relative_jump = __ BytecodeOperandImm(0);
1385   __ JumpIfWordNotEqual(accumulator, the_hole_value, relative_jump);
1386 }
1387 
1388 // JumpIfNotHoleConstant <idx>
1389 //
1390 // Jump by number of bytes in the Smi in the |idx| entry in the constant pool
1391 // if the object referenced by the accumulator is the hole constant.
DoJumpIfNotHoleConstant(InterpreterAssembler * assembler)1392 void Interpreter::DoJumpIfNotHoleConstant(InterpreterAssembler* assembler) {
1393   Node* accumulator = __ GetAccumulator();
1394   Node* the_hole_value = __ HeapConstant(isolate_->factory()->the_hole_value());
1395   Node* index = __ BytecodeOperandIdx(0);
1396   Node* constant = __ LoadConstantPoolEntry(index);
1397   Node* relative_jump = __ SmiUntag(constant);
1398   __ JumpIfWordNotEqual(accumulator, the_hole_value, relative_jump);
1399 }
1400 
1401 // CreateRegExpLiteral <pattern_idx> <literal_idx> <flags>
1402 //
1403 // Creates a regular expression literal for literal index <literal_idx> with
1404 // <flags> and the pattern in <pattern_idx>.
DoCreateRegExpLiteral(InterpreterAssembler * assembler)1405 void Interpreter::DoCreateRegExpLiteral(InterpreterAssembler* assembler) {
1406   Callable callable = CodeFactory::FastCloneRegExp(isolate_);
1407   Node* target = __ HeapConstant(callable.code());
1408   Node* index = __ BytecodeOperandIdx(0);
1409   Node* pattern = __ LoadConstantPoolEntry(index);
1410   Node* literal_index_raw = __ BytecodeOperandIdx(1);
1411   Node* literal_index = __ SmiTag(literal_index_raw);
1412   Node* flags_raw = __ BytecodeOperandFlag(2);
1413   Node* flags = __ SmiTag(flags_raw);
1414   Node* closure = __ LoadRegister(Register::function_closure());
1415   Node* context = __ GetContext();
1416   Node* result = __ CallStub(callable.descriptor(), target, context, closure,
1417                              literal_index, pattern, flags);
1418   __ SetAccumulator(result);
1419   __ Dispatch();
1420 }
1421 
1422 // CreateArrayLiteral <element_idx> <literal_idx> <flags>
1423 //
1424 // Creates an array literal for literal index <literal_idx> with flags <flags>
1425 // and constant elements in <element_idx>.
DoCreateArrayLiteral(InterpreterAssembler * assembler)1426 void Interpreter::DoCreateArrayLiteral(InterpreterAssembler* assembler) {
1427   Node* index = __ BytecodeOperandIdx(0);
1428   Node* constant_elements = __ LoadConstantPoolEntry(index);
1429   Node* literal_index_raw = __ BytecodeOperandIdx(1);
1430   Node* literal_index = __ SmiTag(literal_index_raw);
1431   Node* flags_raw = __ BytecodeOperandFlag(2);
1432   Node* flags = __ SmiTag(flags_raw);
1433   Node* closure = __ LoadRegister(Register::function_closure());
1434   Node* context = __ GetContext();
1435   Node* result = __ CallRuntime(Runtime::kCreateArrayLiteral, context, closure,
1436                                 literal_index, constant_elements, flags);
1437   __ SetAccumulator(result);
1438   __ Dispatch();
1439 }
1440 
1441 // CreateObjectLiteral <element_idx> <literal_idx> <flags>
1442 //
1443 // Creates an object literal for literal index <literal_idx> with
1444 // CreateObjectLiteralFlags <flags> and constant elements in <element_idx>.
DoCreateObjectLiteral(InterpreterAssembler * assembler)1445 void Interpreter::DoCreateObjectLiteral(InterpreterAssembler* assembler) {
1446   Node* literal_index_raw = __ BytecodeOperandIdx(1);
1447   Node* literal_index = __ SmiTag(literal_index_raw);
1448   Node* bytecode_flags = __ BytecodeOperandFlag(2);
1449   Node* closure = __ LoadRegister(Register::function_closure());
1450 
1451   // Check if we can do a fast clone or have to call the runtime.
1452   Label if_fast_clone(assembler),
1453       if_not_fast_clone(assembler, Label::kDeferred);
1454   Node* fast_clone_properties_count =
1455       __ BitFieldDecode<CreateObjectLiteralFlags::FastClonePropertiesCountBits>(
1456           bytecode_flags);
1457   __ BranchIf(fast_clone_properties_count, &if_fast_clone, &if_not_fast_clone);
1458 
1459   __ Bind(&if_fast_clone);
1460   {
1461     // If we can do a fast clone do the fast-path in FastCloneShallowObjectStub.
1462     Node* result = FastCloneShallowObjectStub::GenerateFastPath(
1463         assembler, &if_not_fast_clone, closure, literal_index,
1464         fast_clone_properties_count);
1465     __ SetAccumulator(result);
1466     __ Dispatch();
1467   }
1468 
1469   __ Bind(&if_not_fast_clone);
1470   {
1471     // If we can't do a fast clone, call into the runtime.
1472     Node* index = __ BytecodeOperandIdx(0);
1473     Node* constant_elements = __ LoadConstantPoolEntry(index);
1474     Node* context = __ GetContext();
1475 
1476     STATIC_ASSERT(CreateObjectLiteralFlags::FlagsBits::kShift == 0);
1477     Node* flags_raw = __ Word32And(
1478         bytecode_flags,
1479         __ Int32Constant(CreateObjectLiteralFlags::FlagsBits::kMask));
1480     Node* flags = __ SmiTag(flags_raw);
1481 
1482     Node* result =
1483         __ CallRuntime(Runtime::kCreateObjectLiteral, context, closure,
1484                        literal_index, constant_elements, flags);
1485     __ SetAccumulator(result);
1486     __ Dispatch();
1487   }
1488 }
1489 
1490 // CreateClosure <index> <tenured>
1491 //
1492 // Creates a new closure for SharedFunctionInfo at position |index| in the
1493 // constant pool and with the PretenureFlag <tenured>.
DoCreateClosure(InterpreterAssembler * assembler)1494 void Interpreter::DoCreateClosure(InterpreterAssembler* assembler) {
1495   // TODO(rmcilroy): Possibly call FastNewClosureStub when possible instead of
1496   // calling into the runtime.
1497   Node* index = __ BytecodeOperandIdx(0);
1498   Node* shared = __ LoadConstantPoolEntry(index);
1499   Node* tenured_raw = __ BytecodeOperandFlag(1);
1500   Node* tenured = __ SmiTag(tenured_raw);
1501   Node* context = __ GetContext();
1502   Node* result =
1503       __ CallRuntime(Runtime::kInterpreterNewClosure, context, shared, tenured);
1504   __ SetAccumulator(result);
1505   __ Dispatch();
1506 }
1507 
1508 // CreateMappedArguments
1509 //
1510 // Creates a new mapped arguments object.
DoCreateMappedArguments(InterpreterAssembler * assembler)1511 void Interpreter::DoCreateMappedArguments(InterpreterAssembler* assembler) {
1512   Node* closure = __ LoadRegister(Register::function_closure());
1513   Node* context = __ GetContext();
1514 
1515   Label if_duplicate_parameters(assembler, Label::kDeferred);
1516   Label if_not_duplicate_parameters(assembler);
1517 
1518   // Check if function has duplicate parameters.
1519   // TODO(rmcilroy): Remove this check when FastNewSloppyArgumentsStub supports
1520   // duplicate parameters.
1521   Node* shared_info =
1522       __ LoadObjectField(closure, JSFunction::kSharedFunctionInfoOffset);
1523   Node* compiler_hints = __ LoadObjectField(
1524       shared_info, SharedFunctionInfo::kHasDuplicateParametersByteOffset,
1525       MachineType::Uint8());
1526   Node* duplicate_parameters_bit = __ Int32Constant(
1527       1 << SharedFunctionInfo::kHasDuplicateParametersBitWithinByte);
1528   Node* compare = __ Word32And(compiler_hints, duplicate_parameters_bit);
1529   __ BranchIf(compare, &if_duplicate_parameters, &if_not_duplicate_parameters);
1530 
1531   __ Bind(&if_not_duplicate_parameters);
1532   {
1533     // TODO(rmcilroy): Inline FastNewSloppyArguments when it is a TurboFan stub.
1534     Callable callable = CodeFactory::FastNewSloppyArguments(isolate_, true);
1535     Node* target = __ HeapConstant(callable.code());
1536     Node* result = __ CallStub(callable.descriptor(), target, context, closure);
1537     __ SetAccumulator(result);
1538     __ Dispatch();
1539   }
1540 
1541   __ Bind(&if_duplicate_parameters);
1542   {
1543     Node* result =
1544         __ CallRuntime(Runtime::kNewSloppyArguments_Generic, context, closure);
1545     __ SetAccumulator(result);
1546     __ Dispatch();
1547   }
1548 }
1549 
1550 // CreateUnmappedArguments
1551 //
1552 // Creates a new unmapped arguments object.
DoCreateUnmappedArguments(InterpreterAssembler * assembler)1553 void Interpreter::DoCreateUnmappedArguments(InterpreterAssembler* assembler) {
1554   // TODO(rmcilroy): Inline FastNewStrictArguments when it is a TurboFan stub.
1555   Callable callable = CodeFactory::FastNewStrictArguments(isolate_, true);
1556   Node* target = __ HeapConstant(callable.code());
1557   Node* context = __ GetContext();
1558   Node* closure = __ LoadRegister(Register::function_closure());
1559   Node* result = __ CallStub(callable.descriptor(), target, context, closure);
1560   __ SetAccumulator(result);
1561   __ Dispatch();
1562 }
1563 
1564 // CreateRestParameter
1565 //
1566 // Creates a new rest parameter array.
DoCreateRestParameter(InterpreterAssembler * assembler)1567 void Interpreter::DoCreateRestParameter(InterpreterAssembler* assembler) {
1568   // TODO(rmcilroy): Inline FastNewRestArguments when it is a TurboFan stub.
1569   Callable callable = CodeFactory::FastNewRestParameter(isolate_, true);
1570   Node* target = __ HeapConstant(callable.code());
1571   Node* closure = __ LoadRegister(Register::function_closure());
1572   Node* context = __ GetContext();
1573   Node* result = __ CallStub(callable.descriptor(), target, context, closure);
1574   __ SetAccumulator(result);
1575   __ Dispatch();
1576 }
1577 
1578 // StackCheck
1579 //
1580 // Performs a stack guard check.
DoStackCheck(InterpreterAssembler * assembler)1581 void Interpreter::DoStackCheck(InterpreterAssembler* assembler) {
1582   Label ok(assembler), stack_check_interrupt(assembler, Label::kDeferred);
1583 
1584   Node* interrupt = __ StackCheckTriggeredInterrupt();
1585   __ BranchIf(interrupt, &stack_check_interrupt, &ok);
1586 
1587   __ Bind(&ok);
1588   __ Dispatch();
1589 
1590   __ Bind(&stack_check_interrupt);
1591   {
1592     Node* context = __ GetContext();
1593     __ CallRuntime(Runtime::kStackGuard, context);
1594     __ Dispatch();
1595   }
1596 }
1597 
1598 // Throw
1599 //
1600 // Throws the exception in the accumulator.
DoThrow(InterpreterAssembler * assembler)1601 void Interpreter::DoThrow(InterpreterAssembler* assembler) {
1602   Node* exception = __ GetAccumulator();
1603   Node* context = __ GetContext();
1604   __ CallRuntime(Runtime::kThrow, context, exception);
1605   // We shouldn't ever return from a throw.
1606   __ Abort(kUnexpectedReturnFromThrow);
1607 }
1608 
1609 // ReThrow
1610 //
1611 // Re-throws the exception in the accumulator.
DoReThrow(InterpreterAssembler * assembler)1612 void Interpreter::DoReThrow(InterpreterAssembler* assembler) {
1613   Node* exception = __ GetAccumulator();
1614   Node* context = __ GetContext();
1615   __ CallRuntime(Runtime::kReThrow, context, exception);
1616   // We shouldn't ever return from a throw.
1617   __ Abort(kUnexpectedReturnFromThrow);
1618 }
1619 
1620 // Return
1621 //
1622 // Return the value in the accumulator.
DoReturn(InterpreterAssembler * assembler)1623 void Interpreter::DoReturn(InterpreterAssembler* assembler) {
1624   __ UpdateInterruptBudgetOnReturn();
1625   Node* accumulator = __ GetAccumulator();
1626   __ Return(accumulator);
1627 }
1628 
1629 // Debugger
1630 //
1631 // Call runtime to handle debugger statement.
DoDebugger(InterpreterAssembler * assembler)1632 void Interpreter::DoDebugger(InterpreterAssembler* assembler) {
1633   Node* context = __ GetContext();
1634   __ CallRuntime(Runtime::kHandleDebuggerStatement, context);
1635   __ Dispatch();
1636 }
1637 
1638 // DebugBreak
1639 //
1640 // Call runtime to handle a debug break.
1641 #define DEBUG_BREAK(Name, ...)                                                \
1642   void Interpreter::Do##Name(InterpreterAssembler* assembler) {               \
1643     Node* context = __ GetContext();                                          \
1644     Node* accumulator = __ GetAccumulator();                                  \
1645     Node* original_handler =                                                  \
1646         __ CallRuntime(Runtime::kDebugBreakOnBytecode, context, accumulator); \
1647     __ DispatchToBytecodeHandler(original_handler);                           \
1648   }
1649 DEBUG_BREAK_BYTECODE_LIST(DEBUG_BREAK);
1650 #undef DEBUG_BREAK
1651 
1652 // ForInPrepare <cache_info_triple>
1653 //
1654 // Returns state for for..in loop execution based on the object in the
1655 // accumulator. The result is output in registers |cache_info_triple| to
1656 // |cache_info_triple + 2|, with the registers holding cache_type, cache_array,
1657 // and cache_length respectively.
DoForInPrepare(InterpreterAssembler * assembler)1658 void Interpreter::DoForInPrepare(InterpreterAssembler* assembler) {
1659   Node* object = __ GetAccumulator();
1660   Node* context = __ GetContext();
1661   Node* result_triple = __ CallRuntime(Runtime::kForInPrepare, context, object);
1662 
1663   // Set output registers:
1664   //   0 == cache_type, 1 == cache_array, 2 == cache_length
1665   Node* output_register = __ BytecodeOperandReg(0);
1666   for (int i = 0; i < 3; i++) {
1667     Node* cache_info = __ Projection(i, result_triple);
1668     __ StoreRegister(cache_info, output_register);
1669     output_register = __ NextRegister(output_register);
1670   }
1671   __ Dispatch();
1672 }
1673 
1674 // ForInNext <receiver> <index> <cache_info_pair>
1675 //
1676 // Returns the next enumerable property in the the accumulator.
DoForInNext(InterpreterAssembler * assembler)1677 void Interpreter::DoForInNext(InterpreterAssembler* assembler) {
1678   Node* receiver_reg = __ BytecodeOperandReg(0);
1679   Node* receiver = __ LoadRegister(receiver_reg);
1680   Node* index_reg = __ BytecodeOperandReg(1);
1681   Node* index = __ LoadRegister(index_reg);
1682   Node* cache_type_reg = __ BytecodeOperandReg(2);
1683   Node* cache_type = __ LoadRegister(cache_type_reg);
1684   Node* cache_array_reg = __ NextRegister(cache_type_reg);
1685   Node* cache_array = __ LoadRegister(cache_array_reg);
1686 
1687   // Load the next key from the enumeration array.
1688   Node* key = __ LoadFixedArrayElement(cache_array, index, 0,
1689                                        CodeStubAssembler::SMI_PARAMETERS);
1690 
1691   // Check if we can use the for-in fast path potentially using the enum cache.
1692   Label if_fast(assembler), if_slow(assembler, Label::kDeferred);
1693   Node* receiver_map = __ LoadObjectField(receiver, HeapObject::kMapOffset);
1694   Node* condition = __ WordEqual(receiver_map, cache_type);
1695   __ BranchIf(condition, &if_fast, &if_slow);
1696   __ Bind(&if_fast);
1697   {
1698     // Enum cache in use for {receiver}, the {key} is definitely valid.
1699     __ SetAccumulator(key);
1700     __ Dispatch();
1701   }
1702   __ Bind(&if_slow);
1703   {
1704     // Record the fact that we hit the for-in slow path.
1705     Node* vector_index = __ BytecodeOperandIdx(3);
1706     Node* type_feedback_vector = __ LoadTypeFeedbackVector();
1707     Node* megamorphic_sentinel =
1708         __ HeapConstant(TypeFeedbackVector::MegamorphicSentinel(isolate_));
1709     __ StoreFixedArrayElement(type_feedback_vector, vector_index,
1710                               megamorphic_sentinel, SKIP_WRITE_BARRIER);
1711 
1712     // Need to filter the {key} for the {receiver}.
1713     Node* context = __ GetContext();
1714     Node* result =
1715         __ CallRuntime(Runtime::kForInFilter, context, receiver, key);
1716     __ SetAccumulator(result);
1717     __ Dispatch();
1718   }
1719 }
1720 
1721 // ForInDone <index> <cache_length>
1722 //
1723 // Returns true if the end of the enumerable properties has been reached.
DoForInDone(InterpreterAssembler * assembler)1724 void Interpreter::DoForInDone(InterpreterAssembler* assembler) {
1725   Node* index_reg = __ BytecodeOperandReg(0);
1726   Node* index = __ LoadRegister(index_reg);
1727   Node* cache_length_reg = __ BytecodeOperandReg(1);
1728   Node* cache_length = __ LoadRegister(cache_length_reg);
1729 
1730   // Check if {index} is at {cache_length} already.
1731   Label if_true(assembler), if_false(assembler), end(assembler);
1732   __ BranchIfWordEqual(index, cache_length, &if_true, &if_false);
1733   __ Bind(&if_true);
1734   {
1735     __ SetAccumulator(__ BooleanConstant(true));
1736     __ Goto(&end);
1737   }
1738   __ Bind(&if_false);
1739   {
1740     __ SetAccumulator(__ BooleanConstant(false));
1741     __ Goto(&end);
1742   }
1743   __ Bind(&end);
1744   __ Dispatch();
1745 }
1746 
1747 // ForInStep <index>
1748 //
1749 // Increments the loop counter in register |index| and stores the result
1750 // in the accumulator.
DoForInStep(InterpreterAssembler * assembler)1751 void Interpreter::DoForInStep(InterpreterAssembler* assembler) {
1752   Node* index_reg = __ BytecodeOperandReg(0);
1753   Node* index = __ LoadRegister(index_reg);
1754   Node* one = __ SmiConstant(Smi::FromInt(1));
1755   Node* result = __ SmiAdd(index, one);
1756   __ SetAccumulator(result);
1757   __ Dispatch();
1758 }
1759 
1760 // Wide
1761 //
1762 // Prefix bytecode indicating next bytecode has wide (16-bit) operands.
DoWide(InterpreterAssembler * assembler)1763 void Interpreter::DoWide(InterpreterAssembler* assembler) {
1764   __ DispatchWide(OperandScale::kDouble);
1765 }
1766 
1767 // ExtraWide
1768 //
1769 // Prefix bytecode indicating next bytecode has extra-wide (32-bit) operands.
DoExtraWide(InterpreterAssembler * assembler)1770 void Interpreter::DoExtraWide(InterpreterAssembler* assembler) {
1771   __ DispatchWide(OperandScale::kQuadruple);
1772 }
1773 
1774 // Illegal
1775 //
1776 // An invalid bytecode aborting execution if dispatched.
DoIllegal(InterpreterAssembler * assembler)1777 void Interpreter::DoIllegal(InterpreterAssembler* assembler) {
1778   __ Abort(kInvalidBytecode);
1779 }
1780 
1781 // Nop
1782 //
1783 // No operation.
DoNop(InterpreterAssembler * assembler)1784 void Interpreter::DoNop(InterpreterAssembler* assembler) { __ Dispatch(); }
1785 
1786 // SuspendGenerator <generator>
1787 //
1788 // Exports the register file and stores it into the generator.  Also stores the
1789 // current context, the state given in the accumulator, and the current bytecode
1790 // offset (for debugging purposes) into the generator.
DoSuspendGenerator(InterpreterAssembler * assembler)1791 void Interpreter::DoSuspendGenerator(InterpreterAssembler* assembler) {
1792   Node* generator_reg = __ BytecodeOperandReg(0);
1793   Node* generator = __ LoadRegister(generator_reg);
1794 
1795   Label if_stepping(assembler, Label::kDeferred), ok(assembler);
1796   Node* step_action_address = __ ExternalConstant(
1797       ExternalReference::debug_last_step_action_address(isolate_));
1798   Node* step_action = __ Load(MachineType::Int8(), step_action_address);
1799   STATIC_ASSERT(StepIn > StepNext);
1800   STATIC_ASSERT(StepFrame > StepNext);
1801   STATIC_ASSERT(LastStepAction == StepFrame);
1802   Node* step_next = __ Int32Constant(StepNext);
1803   __ BranchIfInt32LessThanOrEqual(step_next, step_action, &if_stepping, &ok);
1804   __ Bind(&ok);
1805 
1806   Node* array =
1807       __ LoadObjectField(generator, JSGeneratorObject::kOperandStackOffset);
1808   Node* context = __ GetContext();
1809   Node* state = __ GetAccumulator();
1810 
1811   __ ExportRegisterFile(array);
1812   __ StoreObjectField(generator, JSGeneratorObject::kContextOffset, context);
1813   __ StoreObjectField(generator, JSGeneratorObject::kContinuationOffset, state);
1814 
1815   Node* offset = __ SmiTag(__ BytecodeOffset());
1816   __ StoreObjectField(generator, JSGeneratorObject::kInputOrDebugPosOffset,
1817                       offset);
1818 
1819   __ Dispatch();
1820 
1821   __ Bind(&if_stepping);
1822   {
1823     Node* context = __ GetContext();
1824     __ CallRuntime(Runtime::kDebugRecordAsyncFunction, context, generator);
1825     __ Goto(&ok);
1826   }
1827 }
1828 
1829 // ResumeGenerator <generator>
1830 //
1831 // Imports the register file stored in the generator. Also loads the
1832 // generator's state and stores it in the accumulator, before overwriting it
1833 // with kGeneratorExecuting.
DoResumeGenerator(InterpreterAssembler * assembler)1834 void Interpreter::DoResumeGenerator(InterpreterAssembler* assembler) {
1835   Node* generator_reg = __ BytecodeOperandReg(0);
1836   Node* generator = __ LoadRegister(generator_reg);
1837 
1838   __ ImportRegisterFile(
1839       __ LoadObjectField(generator, JSGeneratorObject::kOperandStackOffset));
1840 
1841   Node* old_state =
1842       __ LoadObjectField(generator, JSGeneratorObject::kContinuationOffset);
1843   Node* new_state = __ Int32Constant(JSGeneratorObject::kGeneratorExecuting);
1844   __ StoreObjectField(generator, JSGeneratorObject::kContinuationOffset,
1845       __ SmiTag(new_state));
1846   __ SetAccumulator(old_state);
1847 
1848   __ Dispatch();
1849 }
1850 
1851 }  // namespace interpreter
1852 }  // namespace internal
1853 }  // namespace v8
1854