Lines Matching refs:scratch
58 Register scratch, in InNewSpace() argument
63 if (scratch.is(object)) { in InNewSpace()
64 and_(scratch, Immediate(~Page::kPageAlignmentMask)); in InNewSpace()
66 mov(scratch, Immediate(~Page::kPageAlignmentMask)); in InNewSpace()
67 and_(scratch, object); in InNewSpace()
75 test_b(Operand(scratch, MemoryChunk::kFlagsOffset), in InNewSpace()
84 Register scratch, in RememberedSetHelper() argument
90 JumpIfNotInNewSpace(object, scratch, &ok, Label::kNear); in RememberedSetHelper()
97 mov(scratch, Operand::StaticVariable(store_buffer)); in RememberedSetHelper()
99 mov(Operand(scratch, 0), addr); in RememberedSetHelper()
101 add(scratch, Immediate(kPointerSize)); in RememberedSetHelper()
103 mov(Operand::StaticVariable(store_buffer), scratch); in RememberedSetHelper()
106 test(scratch, Immediate(StoreBuffer::kStoreBufferOverflowBit)); in RememberedSetHelper()
563 Register scratch, in IsObjectJSObjectType() argument
566 IsInstanceJSObjectType(map, scratch, fail); in IsObjectJSObjectType()
571 Register scratch, in IsInstanceJSObjectType() argument
573 movzx_b(scratch, FieldOperand(map, Map::kInstanceTypeOffset)); in IsInstanceJSObjectType()
574 sub(scratch, Immediate(FIRST_NONCALLABLE_SPEC_OBJECT_TYPE)); in IsInstanceJSObjectType()
575 cmp(scratch, in IsInstanceJSObjectType()
910 Register scratch, in CheckAccessGlobalProxy() argument
914 ASSERT(!holder_reg.is(scratch)); in CheckAccessGlobalProxy()
917 mov(scratch, Operand(ebp, StandardFrameConstants::kContextOffset)); in CheckAccessGlobalProxy()
921 cmp(scratch, Immediate(0)); in CheckAccessGlobalProxy()
926 mov(scratch, FieldOperand(scratch, offset)); in CheckAccessGlobalProxy()
927 mov(scratch, FieldOperand(scratch, GlobalObject::kGlobalContextOffset)); in CheckAccessGlobalProxy()
931 push(scratch); in CheckAccessGlobalProxy()
933 mov(scratch, FieldOperand(scratch, HeapObject::kMapOffset)); in CheckAccessGlobalProxy()
934 cmp(scratch, isolate()->factory()->global_context_map()); in CheckAccessGlobalProxy()
936 pop(scratch); in CheckAccessGlobalProxy()
940 cmp(scratch, FieldOperand(holder_reg, JSGlobalProxy::kContextOffset)); in CheckAccessGlobalProxy()
968 mov(scratch, FieldOperand(scratch, token_offset)); in CheckAccessGlobalProxy()
969 cmp(scratch, FieldOperand(holder_reg, token_offset)); in CheckAccessGlobalProxy()
981 void MacroAssembler::GetNumberHash(Register r0, Register scratch) { in GetNumberHash() argument
986 mov(scratch, Immediate(Heap::kHashSeedRootIndex)); in GetNumberHash()
987 mov(scratch, in GetNumberHash()
988 Operand::StaticArray(scratch, times_pointer_size, roots_array_start)); in GetNumberHash()
989 SmiUntag(scratch); in GetNumberHash()
990 xor_(r0, scratch); in GetNumberHash()
997 mov(scratch, r0); in GetNumberHash()
999 shl(scratch, 15); in GetNumberHash()
1000 add(r0, scratch); in GetNumberHash()
1002 mov(scratch, r0); in GetNumberHash()
1003 shr(scratch, 12); in GetNumberHash()
1004 xor_(r0, scratch); in GetNumberHash()
1008 mov(scratch, r0); in GetNumberHash()
1009 shr(scratch, 4); in GetNumberHash()
1010 xor_(r0, scratch); in GetNumberHash()
1014 mov(scratch, r0); in GetNumberHash()
1015 shr(scratch, 16); in GetNumberHash()
1016 xor_(r0, scratch); in GetNumberHash()
1097 Register scratch, in LoadAllocationTopHelper() argument
1105 ASSERT(scratch.is(no_reg)); in LoadAllocationTopHelper()
1115 if (scratch.is(no_reg)) { in LoadAllocationTopHelper()
1118 mov(scratch, Immediate(new_space_allocation_top)); in LoadAllocationTopHelper()
1119 mov(result, Operand(scratch, 0)); in LoadAllocationTopHelper()
1125 Register scratch) { in UpdateAllocationTopHelper() argument
1135 if (scratch.is(no_reg)) { in UpdateAllocationTopHelper()
1138 mov(Operand(scratch, 0), result_end); in UpdateAllocationTopHelper()
1146 Register scratch, in AllocateInNewSpace() argument
1156 if (scratch.is_valid()) { in AllocateInNewSpace()
1157 mov(scratch, Immediate(0x7291)); in AllocateInNewSpace()
1166 LoadAllocationTopHelper(result, scratch, flags); in AllocateInNewSpace()
1183 UpdateAllocationTopHelper(top_reg, scratch); in AllocateInNewSpace()
1203 Register scratch, in AllocateInNewSpace() argument
1211 if (scratch.is_valid()) { in AllocateInNewSpace()
1212 mov(scratch, Immediate(0x7291)); in AllocateInNewSpace()
1222 LoadAllocationTopHelper(result, scratch, flags); in AllocateInNewSpace()
1242 UpdateAllocationTopHelper(result_end, scratch); in AllocateInNewSpace()
1249 Register scratch, in AllocateInNewSpace() argument
1257 if (scratch.is_valid()) { in AllocateInNewSpace()
1258 mov(scratch, Immediate(0x7291)); in AllocateInNewSpace()
1268 LoadAllocationTopHelper(result, scratch, flags); in AllocateInNewSpace()
1287 UpdateAllocationTopHelper(result_end, scratch); in AllocateInNewSpace()
1501 Register scratch) { in CopyBytes() argument
1514 mov(scratch, Operand(source, length, times_1, -4)); in CopyBytes()
1515 mov(Operand(destination, length, times_1, -4), scratch); in CopyBytes()
1516 mov(scratch, ecx); in CopyBytes()
1519 and_(scratch, Immediate(0x3)); in CopyBytes()
1520 add(destination, scratch); in CopyBytes()
1528 mov_b(scratch, Operand(source, 0)); in CopyBytes()
1529 mov_b(Operand(destination, 0), scratch); in CopyBytes()
1581 Register scratch, in NegativeZeroTest() argument
1586 mov(scratch, op1); in NegativeZeroTest()
1587 or_(scratch, op2); in NegativeZeroTest()
1595 Register scratch, in TryGetFunctionPrototype() argument
1607 mov(scratch, in TryGetFunctionPrototype()
1609 BooleanBitTest(scratch, SharedFunctionInfo::kCompilerHintsOffset, in TryGetFunctionPrototype()
1616 movzx_b(scratch, FieldOperand(result, Map::kBitFieldOffset)); in TryGetFunctionPrototype()
1617 test(scratch, Immediate(1 << Map::kHasNonInstancePrototype)); in TryGetFunctionPrototype()
1632 CmpObjectType(result, MAP_TYPE, scratch); in TryGetFunctionPrototype()
2157 Register scratch, in LoadTransitionedArrayMapConditional() argument
2160 mov(scratch, Operand(esi, Context::SlotOffset(Context::GLOBAL_INDEX))); in LoadTransitionedArrayMapConditional()
2161 mov(scratch, FieldOperand(scratch, GlobalObject::kGlobalContextOffset)); in LoadTransitionedArrayMapConditional()
2166 cmp(map_in_out, Operand(scratch, Context::SlotOffset(expected_index))); in LoadTransitionedArrayMapConditional()
2172 mov(map_in_out, Operand(scratch, Context::SlotOffset(trans_index))); in LoadTransitionedArrayMapConditional()
2177 Register function_in, Register scratch, Register map_out) { in LoadInitialArrayMap() argument
2186 scratch, in LoadInitialArrayMap()
2277 void MacroAssembler::Ret(int bytes_dropped, Register scratch) { in Ret() argument
2281 pop(scratch); in Ret()
2283 push(scratch); in Ret()
2459 Register scratch, in LoadPowerOf2() argument
2463 mov(scratch, Immediate(power + HeapNumber::kExponentBias)); in LoadPowerOf2()
2464 movd(dst, scratch); in LoadPowerOf2()
2471 Register scratch, in JumpIfInstanceTypeIsNotSequentialAscii() argument
2473 if (!scratch.is(instance_type)) { in JumpIfInstanceTypeIsNotSequentialAscii()
2474 mov(scratch, instance_type); in JumpIfInstanceTypeIsNotSequentialAscii()
2476 and_(scratch, in JumpIfInstanceTypeIsNotSequentialAscii()
2478 cmp(scratch, kStringTag | kSeqStringTag | kAsciiStringTag); in JumpIfInstanceTypeIsNotSequentialAscii()
2514 void MacroAssembler::PrepareCallCFunction(int num_arguments, Register scratch) { in PrepareCallCFunction() argument
2519 mov(scratch, esp); in PrepareCallCFunction()
2523 mov(Operand(esp, num_arguments * kPointerSize), scratch); in PrepareCallCFunction()
2589 Register scratch, in CheckPageFlag() argument
2595 if (scratch.is(object)) { in CheckPageFlag()
2596 and_(scratch, Immediate(~Page::kPageAlignmentMask)); in CheckPageFlag()
2598 mov(scratch, Immediate(~Page::kPageAlignmentMask)); in CheckPageFlag()
2599 and_(scratch, object); in CheckPageFlag()
2602 test_b(Operand(scratch, MemoryChunk::kFlagsOffset), in CheckPageFlag()
2605 test(Operand(scratch, MemoryChunk::kFlagsOffset), Immediate(mask)); in CheckPageFlag()