Lines Matching refs:a3
105 __ GetObjectType(a2, a3, a4); in Generate_InternalArrayCode()
135 __ GetObjectType(a2, a3, a4); in Generate_ArrayCode()
157 __ IncrementCounter(counters->string_ctor_calls(), 1, a2, a3); in Generate_StringConstructCode()
180 a3, // Scratch. in Generate_StringConstructCode()
184 __ IncrementCounter(counters->string_ctor_cached_number(), 1, a3, a4); in Generate_StringConstructCode()
196 a3, // Scratch. in Generate_StringConstructCode()
202 Register map = a3; in Generate_StringConstructCode()
214 __ LoadRoot(a3, Heap::kEmptyFixedArrayRootIndex); in Generate_StringConstructCode()
215 __ sd(a3, FieldMemOperand(v0, JSObject::kPropertiesOffset)); in Generate_StringConstructCode()
216 __ sd(a3, FieldMemOperand(v0, JSObject::kElementsOffset)); in Generate_StringConstructCode()
233 __ lbu(a3, FieldMemOperand(a2, Map::kInstanceTypeOffset)); in Generate_StringConstructCode()
235 __ And(a4, a3, Operand(kIsNotStringMask)); in Generate_StringConstructCode()
238 __ IncrementCounter(counters->string_ctor_conversions(), 1, a3, a4); in Generate_StringConstructCode()
244 __ IncrementCounter(counters->string_ctor_conversions(), 1, a3, a4); in Generate_StringConstructCode()
264 __ IncrementCounter(counters->string_ctor_gc_required(), 1, a3, a4); in Generate_StringConstructCode()
347 __ AssertUndefinedOrAllocationSite(a2, a3); in Generate_JSConstructStubHelper()
371 __ GetObjectType(a2, a3, t0); in Generate_JSConstructStubHelper()
379 __ lbu(a3, FieldMemOperand(a2, Map::kInstanceTypeOffset)); in Generate_JSConstructStubHelper()
380 __ Branch(&rt_call, eq, a3, Operand(JS_FUNCTION_TYPE)); in Generate_JSConstructStubHelper()
412 __ lbu(a3, FieldMemOperand(a2, Map::kInstanceSizeOffset)); in Generate_JSConstructStubHelper()
414 __ Daddu(a3, a3, Operand(AllocationMemento::kSize / kPointerSize)); in Generate_JSConstructStubHelper()
417 __ Allocate(a3, t0, t1, t2, &rt_call, SIZE_IN_WORDS); in Generate_JSConstructStubHelper()
464 __ dsll(at, a3, kPointerSizeLog2); in Generate_JSConstructStubHelper()
478 __ Dsubu(a0, a3, Operand(AllocationMemento::kSize / kPointerSize)); in Generate_JSConstructStubHelper()
495 __ dsll(at, a3, kPointerSizeLog2); in Generate_JSConstructStubHelper()
511 __ lbu(a3, FieldMemOperand(a2, Map::kUnusedPropertyFieldsOffset)); in Generate_JSConstructStubHelper()
517 __ Daddu(a3, a3, Operand(t2)); in Generate_JSConstructStubHelper()
520 __ dsubu(a3, a3, t2); in Generate_JSConstructStubHelper()
523 __ Branch(&allocated, eq, a3, Operand(zero_reg)); in Generate_JSConstructStubHelper()
525 a3, Operand(zero_reg)); in Generate_JSConstructStubHelper()
533 __ Daddu(a0, a3, Operand(FixedArray::kHeaderSize / kPointerSize)); in Generate_JSConstructStubHelper()
551 __ dsll32(a0, a3, 0); in Generate_JSConstructStubHelper()
564 __ dsll(a7, a3, kPointerSizeLog2); in Generate_JSConstructStubHelper()
638 __ ld(a3, FieldMemOperand(a2, in Generate_JSConstructStubHelper()
640 __ Daddu(a3, a3, Operand(Smi::FromInt(1))); in Generate_JSConstructStubHelper()
641 __ sd(a3, FieldMemOperand(a2, in Generate_JSConstructStubHelper()
654 __ ld(a3, MemOperand(sp, 3 * kPointerSize)); in Generate_JSConstructStubHelper()
660 __ SmiUntag(a0, a3); in Generate_JSConstructStubHelper()
672 __ SmiUntag(a3); in Generate_JSConstructStubHelper()
675 __ dsll(a4, a3, kPointerSizeLog2); in Generate_JSConstructStubHelper()
680 __ Daddu(a3, a3, Operand(-1)); in Generate_JSConstructStubHelper()
681 __ Branch(&loop, greater_equal, a3, Operand(zero_reg)); in Generate_JSConstructStubHelper()
718 __ GetObjectType(v0, a1, a3); in Generate_JSConstructStubHelper()
719 __ Branch(&exit, greater_equal, a3, Operand(FIRST_SPEC_OBJECT_TYPE)); in Generate_JSConstructStubHelper()
788 __ dsll32(a3, a3, 0); // int32_t -> int64_t. in Generate_JSEntryTrampolineHelper()
789 __ dsrl32(a3, a3, 0); in Generate_JSEntryTrampolineHelper()
790 __ dsll(a4, a3, kPointerSizeLog2); in Generate_JSEntryTrampolineHelper()
815 __ mov(a0, a3); in Generate_JSEntryTrampolineHelper()
1115 __ lbu(a3, FieldMemOperand(a2, SharedFunctionInfo::kStrictModeByteOffset)); in Generate_FunctionCall()
1116 __ And(a7, a3, Operand(1 << SharedFunctionInfo::kStrictModeBitWithinByte)); in Generate_FunctionCall()
1120 __ lbu(a3, FieldMemOperand(a2, SharedFunctionInfo::kNativeByteOffset)); in Generate_FunctionCall()
1121 __ And(a7, a3, Operand(1 << SharedFunctionInfo::kNativeBitWithinByte)); in Generate_FunctionCall()
1134 __ LoadRoot(a3, Heap::kUndefinedValueRootIndex); in Generate_FunctionCall()
1135 __ Branch(&use_global_proxy, eq, a2, Operand(a3)); in Generate_FunctionCall()
1136 __ LoadRoot(a3, Heap::kNullValueRootIndex); in Generate_FunctionCall()
1137 __ Branch(&use_global_proxy, eq, a2, Operand(a3)); in Generate_FunctionCall()
1140 __ GetObjectType(a2, a3, a3); in Generate_FunctionCall()
1141 __ Branch(&shift_arguments, ge, a3, Operand(FIRST_SPEC_OBJECT_TYPE)); in Generate_FunctionCall()
1169 __ daddu(a3, sp, at); in Generate_FunctionCall()
1170 __ sd(a2, MemOperand(a3, -kPointerSize)); in Generate_FunctionCall()
1246 __ ld(a3, FieldMemOperand(a1, JSFunction::kSharedFunctionInfoOffset)); in Generate_FunctionCall()
1250 FieldMemOperand(a3, SharedFunctionInfo::kFormalParameterCountOffset)); in Generate_FunctionCall()
1255 __ ld(a3, FieldMemOperand(a1, JSFunction::kCodeEntryOffset)); in Generate_FunctionCall()
1257 __ InvokeCode(a3, expected, expected, JUMP_FUNCTION, NullCallWrapper()); in Generate_FunctionCall()
1476 __ ld(a3, FieldMemOperand(a1, JSFunction::kCodeEntryOffset)); in Generate_ArgumentsAdaptorTrampoline()
1566 __ Call(a3); in Generate_ArgumentsAdaptorTrampoline()
1580 __ Jump(a3); in Generate_ArgumentsAdaptorTrampoline()