• Home
  • Raw
  • Download

Lines Matching refs:Alu64R

599    i->Ain.Alu64R.op  = op;  in AMD64Instr_Alu64R()
600 i->Ain.Alu64R.src = src; in AMD64Instr_Alu64R()
601 i->Ain.Alu64R.dst = dst; in AMD64Instr_Alu64R()
1047 vex_printf("%sq ", showAMD64AluOp(i->Ain.Alu64R.op)); in ppAMD64Instr()
1048 ppAMD64RMI(i->Ain.Alu64R.src); in ppAMD64Instr()
1050 ppHRegAMD64(i->Ain.Alu64R.dst); in ppAMD64Instr()
1404 addRegUsage_AMD64RMI(u, i->Ain.Alu64R.src); in getRegUsage_AMD64Instr()
1405 if (i->Ain.Alu64R.op == Aalu_MOV) { in getRegUsage_AMD64Instr()
1406 addHRegUse(u, HRmWrite, i->Ain.Alu64R.dst); in getRegUsage_AMD64Instr()
1409 if (i->Ain.Alu64R.op == Aalu_CMP) { in getRegUsage_AMD64Instr()
1410 addHRegUse(u, HRmRead, i->Ain.Alu64R.dst); in getRegUsage_AMD64Instr()
1413 addHRegUse(u, HRmModify, i->Ain.Alu64R.dst); in getRegUsage_AMD64Instr()
1728 mapRegs_AMD64RMI(m, i->Ain.Alu64R.src); in mapRegs_AMD64Instr()
1729 mapReg(m, &i->Ain.Alu64R.dst); in mapRegs_AMD64Instr()
1922 if (i->Ain.Alu64R.op != Aalu_MOV) in isMove_AMD64Instr()
1924 if (i->Ain.Alu64R.src->tag != Armi_Reg) in isMove_AMD64Instr()
1926 *src = i->Ain.Alu64R.src->Armi.Reg.reg; in isMove_AMD64Instr()
1927 *dst = i->Ain.Alu64R.dst; in isMove_AMD64Instr()
2006 && (i->Ain.Alu64R.op == Aalu_MOV || i->Ain.Alu64R.op == Aalu_OR in directReload_AMD64()
2007 || i->Ain.Alu64R.op == Aalu_XOR) in directReload_AMD64()
2008 && i->Ain.Alu64R.src->tag == Armi_Reg in directReload_AMD64()
2009 && sameHReg(i->Ain.Alu64R.src->Armi.Reg.reg, vreg)) { in directReload_AMD64()
2010 vassert(! sameHReg(i->Ain.Alu64R.dst, vreg)); in directReload_AMD64()
2012 i->Ain.Alu64R.op, in directReload_AMD64()
2014 i->Ain.Alu64R.dst in directReload_AMD64()
2022 && (i->Ain.Alu64R.op == Aalu_CMP) in directReload_AMD64()
2023 && i->Ain.Alu64R.src->tag == Armi_Imm in directReload_AMD64()
2024 && sameHReg(i->Ain.Alu64R.dst, vreg)) { in directReload_AMD64()
2026 i->Ain.Alu64R.op, in directReload_AMD64()
2027 AMD64RI_Imm( i->Ain.Alu64R.src->Armi.Imm.imm32 ), in directReload_AMD64()
2488 if (i->Ain.Alu64R.op == Aalu_MOV) { in emit_AMD64Instr()
2489 switch (i->Ain.Alu64R.src->tag) { in emit_AMD64Instr()
2491 if (0 == (i->Ain.Alu64R.src->Armi.Imm.imm32 & ~0xFFFFF)) { in emit_AMD64Instr()
2501 if (1 & iregEnc3(i->Ain.Alu64R.dst)) in emit_AMD64Instr()
2503 *p++ = 0xB8 + iregEnc210(i->Ain.Alu64R.dst); in emit_AMD64Instr()
2504 p = emit32(p, i->Ain.Alu64R.src->Armi.Imm.imm32); in emit_AMD64Instr()
2506 *p++ = toUChar(0x48 + (1 & iregEnc3(i->Ain.Alu64R.dst))); in emit_AMD64Instr()
2508 *p++ = toUChar(0xC0 + iregEnc210(i->Ain.Alu64R.dst)); in emit_AMD64Instr()
2509 p = emit32(p, i->Ain.Alu64R.src->Armi.Imm.imm32); in emit_AMD64Instr()
2513 *p++ = rexAMode_R( i->Ain.Alu64R.src->Armi.Reg.reg, in emit_AMD64Instr()
2514 i->Ain.Alu64R.dst ); in emit_AMD64Instr()
2516 p = doAMode_R(p, i->Ain.Alu64R.src->Armi.Reg.reg, in emit_AMD64Instr()
2517 i->Ain.Alu64R.dst); in emit_AMD64Instr()
2520 *p++ = rexAMode_M(i->Ain.Alu64R.dst, in emit_AMD64Instr()
2521 i->Ain.Alu64R.src->Armi.Mem.am); in emit_AMD64Instr()
2523 p = doAMode_M(p, i->Ain.Alu64R.dst, in emit_AMD64Instr()
2524 i->Ain.Alu64R.src->Armi.Mem.am); in emit_AMD64Instr()
2531 if (i->Ain.Alu64R.op == Aalu_MUL) { in emit_AMD64Instr()
2532 switch (i->Ain.Alu64R.src->tag) { in emit_AMD64Instr()
2534 *p++ = rexAMode_R( i->Ain.Alu64R.dst, in emit_AMD64Instr()
2535 i->Ain.Alu64R.src->Armi.Reg.reg); in emit_AMD64Instr()
2538 p = doAMode_R(p, i->Ain.Alu64R.dst, in emit_AMD64Instr()
2539 i->Ain.Alu64R.src->Armi.Reg.reg); in emit_AMD64Instr()
2542 *p++ = rexAMode_M(i->Ain.Alu64R.dst, in emit_AMD64Instr()
2543 i->Ain.Alu64R.src->Armi.Mem.am); in emit_AMD64Instr()
2546 p = doAMode_M(p, i->Ain.Alu64R.dst, in emit_AMD64Instr()
2547 i->Ain.Alu64R.src->Armi.Mem.am); in emit_AMD64Instr()
2550 if (fits8bits(i->Ain.Alu64R.src->Armi.Imm.imm32)) { in emit_AMD64Instr()
2551 *p++ = rexAMode_R(i->Ain.Alu64R.dst, i->Ain.Alu64R.dst); in emit_AMD64Instr()
2553 p = doAMode_R(p, i->Ain.Alu64R.dst, i->Ain.Alu64R.dst); in emit_AMD64Instr()
2554 *p++ = toUChar(0xFF & i->Ain.Alu64R.src->Armi.Imm.imm32); in emit_AMD64Instr()
2556 *p++ = rexAMode_R(i->Ain.Alu64R.dst, i->Ain.Alu64R.dst); in emit_AMD64Instr()
2558 p = doAMode_R(p, i->Ain.Alu64R.dst, i->Ain.Alu64R.dst); in emit_AMD64Instr()
2559 p = emit32(p, i->Ain.Alu64R.src->Armi.Imm.imm32); in emit_AMD64Instr()
2568 switch (i->Ain.Alu64R.op) { in emit_AMD64Instr()
2587 switch (i->Ain.Alu64R.src->tag) { in emit_AMD64Instr()
2589 if (sameHReg(i->Ain.Alu64R.dst, hregAMD64_RAX()) in emit_AMD64Instr()
2590 && !fits8bits(i->Ain.Alu64R.src->Armi.Imm.imm32)) { in emit_AMD64Instr()
2593 p = emit32(p, i->Ain.Alu64R.src->Armi.Imm.imm32); in emit_AMD64Instr()
2595 if (fits8bits(i->Ain.Alu64R.src->Armi.Imm.imm32)) { in emit_AMD64Instr()
2596 *p++ = rexAMode_R_enc_reg( 0, i->Ain.Alu64R.dst ); in emit_AMD64Instr()
2598 p = doAMode_R_enc_reg(p, subopc_imm, i->Ain.Alu64R.dst); in emit_AMD64Instr()
2599 *p++ = toUChar(0xFF & i->Ain.Alu64R.src->Armi.Imm.imm32); in emit_AMD64Instr()
2601 *p++ = rexAMode_R_enc_reg( 0, i->Ain.Alu64R.dst); in emit_AMD64Instr()
2603 p = doAMode_R_enc_reg(p, subopc_imm, i->Ain.Alu64R.dst); in emit_AMD64Instr()
2604 p = emit32(p, i->Ain.Alu64R.src->Armi.Imm.imm32); in emit_AMD64Instr()
2608 *p++ = rexAMode_R( i->Ain.Alu64R.src->Armi.Reg.reg, in emit_AMD64Instr()
2609 i->Ain.Alu64R.dst); in emit_AMD64Instr()
2611 p = doAMode_R(p, i->Ain.Alu64R.src->Armi.Reg.reg, in emit_AMD64Instr()
2612 i->Ain.Alu64R.dst); in emit_AMD64Instr()
2615 *p++ = rexAMode_M( i->Ain.Alu64R.dst, in emit_AMD64Instr()
2616 i->Ain.Alu64R.src->Armi.Mem.am); in emit_AMD64Instr()
2618 p = doAMode_M(p, i->Ain.Alu64R.dst, in emit_AMD64Instr()
2619 i->Ain.Alu64R.src->Armi.Mem.am); in emit_AMD64Instr()