Searched refs:mem_op (Results 1 – 2 of 2) sorted by relevance
/art/compiler/optimizing/ |
D | common_arm64.h | 209 const vixl::aarch64::MemOperand& mem_op) { in OperandFromMemOperand() argument 210 if (mem_op.IsImmediateOffset()) { in OperandFromMemOperand() 211 return vixl::aarch64::Operand(mem_op.GetOffset()); in OperandFromMemOperand() 213 DCHECK(mem_op.IsRegisterOffset()); in OperandFromMemOperand() 214 if (mem_op.GetExtend() != vixl::aarch64::NO_EXTEND) { in OperandFromMemOperand() 215 return vixl::aarch64::Operand(mem_op.GetRegisterOffset(), in OperandFromMemOperand() 216 mem_op.GetExtend(), in OperandFromMemOperand() 217 mem_op.GetShiftAmount()); in OperandFromMemOperand() 218 } else if (mem_op.GetShift() != vixl::aarch64::NO_SHIFT) { in OperandFromMemOperand() 219 return vixl::aarch64::Operand(mem_op.GetRegisterOffset(), in OperandFromMemOperand() [all …]
|
D | intrinsics_arm64.cc | 759 MemOperand mem_op(base.X(), offset); in GenUnsafeGet() local 761 codegen->LoadAcquire(invoke, trg, mem_op, /* needs_null_check= */ true); in GenUnsafeGet() 763 codegen->Load(type, trg, mem_op); in GenUnsafeGet() 883 MemOperand mem_op(base.X(), offset); in GenUnsafePut() local 899 codegen->StoreRelease(invoke, type, source, mem_op, /* needs_null_check= */ false); in GenUnsafePut() 901 codegen->Store(type, source, mem_op); in GenUnsafePut()
|