• Home
  • Raw
  • Download

Lines Matching refs:mkU64

281 static IRExpr* mkU64 ( ULong i )  in mkU64()  function
292 case Ity_I64: return mkU64(i); in mkU()
1670 = mkIRExprVec_5( mkU64(cond), in mk_amd64g_calculate_condition()
1799 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(ccOp)) ); in setFlags_DEP1_DEP2()
1825 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(ccOp)) ); in setFlags_DEP1()
1827 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0)) ); in setFlags_DEP1()
1869 mkU64(ccOp), in setFlags_DEP1_DEP2_shift()
1901 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(ccOp)) ); in setFlags_INC_DEC()
1903 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0)) ); in setFlags_INC_DEC()
1915 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(base_op+0) ) ); in setFlags_MUL()
1918 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(base_op+1) ) ); in setFlags_MUL()
1921 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(base_op+2) ) ); in setFlags_MUL()
1924 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(base_op+3) ) ); in setFlags_MUL()
2025 mkU64(1)) ); in helper_ADC()
2047 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(thunkOp) ) ); in helper_ADC()
2084 mkU64(1)) ); in helper_SBB()
2106 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(thunkOp) ) ); in helper_SBB()
2235 stmt( IRStmt_Put( OFFB_RIP, mkU64(d64) ) ); in jmp_lit()
2268 stmt( IRStmt_Put( OFFB_RIP, mkU64(d64_true) ) ); in jcc_01()
2274 stmt( IRStmt_Put( OFFB_RIP, mkU64(d64_false) ) ); in jcc_01()
2301 binop(Iop_Sub64, mkexpr(new_rsp), mkU64(szB)), in make_redzone_AbiHint()
2500 binop(Iop_Add64,getIRegRexB(8,pfx,rm),mkU64(d)))); in disAMode()
2514 binop(Iop_Add64,getIRegRexB(8,pfx,rm),mkU64(d)))); in disAMode()
2539 binop(Iop_Add64, mkU64(guest_RIP_next_assumed), in disAMode()
2540 mkU64(d)))); in disAMode()
2600 mkU64(d)))); in disAMode()
2615 handleAddrOverrides(vbi, pfx, mkU64(d))); in disAMode()
2643 binop(Iop_Add64, getIRegRexB(8,pfx,base_r), mkU64(d)) )); in disAMode()
2663 mkU64(d)))); in disAMode()
2690 binop(Iop_Add64, getIRegRexB(8,pfx,base_r), mkU64(d)) )); in disAMode()
2710 mkU64(d)))); in disAMode()
2768 return disAMode_copy2tmp( mkU64(d) ); in disAVSIBMode()
2800 mkU64(d)) ); in disAVSIBMode()
3578 mkU64(sz) ); in dis_Grp2()
3593 mkU64(-sz) ); in dis_Grp2()
3605 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(AMD64G_CC_OP_COPY) )); in dis_Grp2()
3607 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0) )); in dis_Grp2()
3608 stmt( IRStmt_Put( OFFB_CC_NDEP, mkU64(0) )); in dis_Grp2()
3740 mkU64(ccOp), in dis_Grp2()
3748 mkU64(0), in dis_Grp2()
3870 assign( t2m, binop(Iop_Or64, mkU64(mask), mkexpr(t2)) ); in dis_Grp8_Imm()
3873 assign( t2m, binop(Iop_And64, mkU64(mask), mkexpr(t2)) ); in dis_Grp8_Imm()
3876 assign( t2m, binop(Iop_Xor64, mkU64(mask), mkexpr(t2)) ); in dis_Grp8_Imm()
3901 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(AMD64G_CC_OP_COPY) )); in dis_Grp8_Imm()
3902 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0) )); in dis_Grp8_Imm()
3907 mkU64(1)) in dis_Grp8_Imm()
3911 stmt( IRStmt_Put( OFFB_CC_NDEP, mkU64(0) )); in dis_Grp8_Imm()
4306 assign(t2, binop(Iop_Sub64, getIReg64(R_RSP), mkU64(8))); in dis_Grp5()
4308 storeLE( mkexpr(t2), mkU64(guest_RIP_bbstart+delta+1)); in dis_Grp5()
4333 assign( t2, binop(Iop_Sub64,getIReg64(R_RSP),mkU64(sz)) ); in dis_Grp5()
4399 assign(t2, binop(Iop_Sub64, getIReg64(R_RSP), mkU64(8))); in dis_Grp5()
4401 storeLE( mkexpr(t2), mkU64(guest_RIP_bbstart+delta+len)); in dis_Grp5()
4426 assign( t2, binop(Iop_Sub64,getIReg64(R_RSP),mkU64(sz)) ); in dis_Grp5()
4639 cmp = binop(Iop_CmpEQ64, mkexpr(tc), mkU64(0)); in dis_REP_op()
4648 putIReg64(R_RCX, binop(Iop_Sub64, mkexpr(tc), mkU64(1)) ); in dis_REP_op()
4832 assign(mask[0], mkU64(0x5555555555555555ULL)); in gen_POPCOUNT()
4833 assign(mask[1], mkU64(0x3333333333333333ULL)); in gen_POPCOUNT()
4834 assign(mask[2], mkU64(0x0F0F0F0F0F0F0F0FULL)); in gen_POPCOUNT()
4835 assign(mask[3], mkU64(0x00FF00FF00FF00FFULL)); in gen_POPCOUNT()
4836 assign(mask[4], mkU64(0x0000FFFF0000FFFFULL)); in gen_POPCOUNT()
4837 assign(mask[5], mkU64(0x00000000FFFFFFFFULL)); in gen_POPCOUNT()
4880 binop(Iop_CmpEQ64, mkexpr(src64x), mkU64(0)), in gen_LZCNT()
4881 mkU64(8 * sizeofIRType(ty)), in gen_LZCNT()
4908 binop(Iop_CmpEQ64, mkexpr(src64), mkU64(0)), in gen_TZCNT()
4909 mkU64(8 * sizeofIRType(ty)), in gen_TZCNT()
5148 IRExpr* cleared = binop(Iop_And64, get_C3210(), mkU64(~AMD64G_FC_MASK_C2)); in set_C2()
5356 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(AMD64G_CC_OP_COPY) )); in fp_do_ucomi_ST0_STi()
5357 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0) )); in fp_do_ucomi_ST0_STi()
5363 mkU64(0x45) in fp_do_ucomi_ST0_STi()
5382 mkU64(65536)), in x87ishly_qnarrow_32_to_16()
5885 mkU64(1)) ); in dis_FPU()
6026 mkU64(1)) ); in dis_FPU()
6064 mkU64(1)) ); in dis_FPU()
7262 binop(Iop_CmpLT64U,mkexpr(amt),mkU64(size)), in dis_MMX_shiftG_byE()
7264 mkU64(0) in dis_MMX_shiftG_byE()
7272 binop(Iop_CmpLT64U,mkexpr(amt),mkU64(size)), in dis_MMX_shiftG_byE()
7324 ? mkU64(0) in dis_MMX_shiftE_imm()
8032 assign( t_rsp, binop(Iop_Sub64, getIReg64(R_RSP), mkU64(288)) ); in dis_bt_G_E()
8044 mkU64(sz == 8 ? 63 : sz == 4 ? 31 : 15)) ); in dis_bt_G_E()
8067 binop(Iop_And64, mkexpr(t_bitno1), mkU64(7))) ); in dis_bt_G_E()
8109 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(AMD64G_CC_OP_COPY) )); in dis_bt_G_E()
8110 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0) )); in dis_bt_G_E()
8117 mkU64(1))) in dis_bt_G_E()
8121 stmt( IRStmt_Put( OFFB_CC_NDEP, mkU64(0) )); in dis_bt_G_E()
8131 putIReg64(R_RSP, binop(Iop_Add64, mkexpr(t_rsp), mkU64(288)) ); in dis_bt_G_E()
8187 assign( srcB, binop(Iop_ExpCmpNE64, mkexpr(src64), mkU64(0)) ); in dis_bs_E_G()
8191 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(AMD64G_CC_OP_COPY) )); in dis_bs_E_G()
8192 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0) )); in dis_bs_E_G()
8197 mkU64(0), in dis_bs_E_G()
8199 mkU64(AMD64G_CC_MASK_Z) in dis_bs_E_G()
8204 stmt( IRStmt_Put( OFFB_CC_NDEP, mkU64(0) )); in dis_bs_E_G()
8239 mkU64(63), in dis_bs_E_G()
8305 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(AMD64G_CC_OP_COPY) )); in codegen_SAHF()
8306 stmt( IRStmt_Put( OFFB_CC_NDEP, mkU64(0) )); in codegen_SAHF()
8307 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0) )); in codegen_SAHF()
8310 binop(Iop_And64, mkexpr(oldflags), mkU64(AMD64G_CC_MASK_O)), in codegen_SAHF()
8313 mkU64(mask_SZACP)) in codegen_SAHF()
8333 = binop(Iop_And64, getIReg64(R_RAX), mkU64(~0xFF00ULL)); in codegen_LAHF()
8335 = binop(Iop_Or64, binop(Iop_And64, mkexpr(flags), mkU64(mask_SZACP)), in codegen_LAHF()
8336 mkU64(1<<1)); in codegen_LAHF()
8734 assign(t3, binop(Iop_Add64, mkexpr(t1), mkU64(8+d64))); in dis_ret()
9342 binop(Iop_CmpLT64U, mkexpr(amt), mkU64(size)), in dis_SSE_shiftG_byE()
9352 binop(Iop_CmpLT64U, mkexpr(amt), mkU64(size)), in dis_SSE_shiftG_byE()
9429 mkU64(3) )); in get_sse_roundingmode()
9646 assign(one32x2, mkU64( (1ULL << 32) + 1 )); in dis_PMULHRSW_helper()
9710 assign( zero, mkU64(0) ); in dis_PSIGN_helper()
9751 assign( zero, mkU64(0) ); in math_PABS_MMX()
9864 assign( rHi, mkU64(0) ); in math_PALIGNR_XMM()
9868 assign( rHi, mkU64(0) ); in math_PALIGNR_XMM()
9872 assign( rHi, mkU64(0) ); in math_PALIGNR_XMM()
9873 assign( rLo, mkU64(0) ); in math_PALIGNR_XMM()
9893 binop(Iop_And64,mkexpr(effective_addr),mkU64(mask)), in gen_SEGV_if_not_XX_aligned()
9894 mkU64(0)), in gen_SEGV_if_not_XX_aligned()
10080 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(AMD64G_CC_OP_COPY) )); in dis_COMISD()
10081 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0) )); in dis_COMISD()
10087 mkU64(0x45) in dis_COMISD()
10123 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(AMD64G_CC_OP_COPY) )); in dis_COMISS()
10124 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0) )); in dis_COMISS()
10132 mkU64(0x45) in dis_COMISS()
10248 assign( hi64r, mkU64(0) ); in math_PSRLDQ()
10253 assign( hi64r, mkU64(0) ); in math_PSRLDQ()
10295 assign( lo64r, mkU64(0) ); in math_PSLLDQ()
10300 assign( lo64r, mkU64(0) ); in math_PSLLDQ()
10440 binop(Iop_Add64,mkexpr(addr),mkU64(4))) ); in dis_CVTPS2PD_128()
10478 binop(Iop_Add64,mkexpr(addr),mkU64(4))) ); in dis_CVTPS2PD_256()
10480 binop(Iop_Add64,mkexpr(addr),mkU64(8))) ); in dis_CVTPS2PD_256()
10482 binop(Iop_Add64,mkexpr(addr),mkU64(12))) ); in dis_CVTPS2PD_256()
11579 binop(Iop_CmpNE64, unop(Iop_32Uto64,mkexpr(ew)), mkU64(0)), in dis_LDMXCSR()
11602 d0->guard = binop(Iop_CmpEQ64, binop(Iop_And64, mkexpr(rfbm), mkU64(1)), in gen_XSAVE_SEQUENCE()
11603 mkU64(1)); in gen_XSAVE_SEQUENCE()
11643 assign(rfbm_1, binop(Iop_And64, mkexpr(rfbm), mkU64(2))); in gen_XSAVE_SEQUENCE()
11644 assign(rfbm_1or2, binop(Iop_And64, mkexpr(rfbm), mkU64(6))); in gen_XSAVE_SEQUENCE()
11646 IRExpr* guard_1 = binop(Iop_CmpEQ64, mkexpr(rfbm_1), mkU64(2)); in gen_XSAVE_SEQUENCE()
11647 IRExpr* guard_1or2 = binop(Iop_CmpNE64, mkexpr(rfbm_1or2), mkU64(0)); in gen_XSAVE_SEQUENCE()
11669 d1->mAddr = binop(Iop_Add64, mkexpr(addr), mkU64(24)); in gen_XSAVE_SEQUENCE()
11692 binop(Iop_Add64, mkexpr(addr), mkU64(160 + reg * 16)), in gen_XSAVE_SEQUENCE()
11703 assign(rfbm_2, binop(Iop_And64, mkexpr(rfbm), mkU64(4))); in gen_XSAVE_SEQUENCE()
11705 IRExpr* guard_2 = binop(Iop_CmpEQ64, mkexpr(rfbm_2), mkU64(4)); in gen_XSAVE_SEQUENCE()
11710 binop(Iop_Add64, mkexpr(addr), mkU64(576 + reg * 16)), in gen_XSAVE_SEQUENCE()
11750 mkU64(aSSUMED_XCR0_VALUE))); in dis_XSAVE()
11757 assign(addr_plus_512, binop(Iop_Add64, mkexpr(addr), mkU64(512))); in dis_XSAVE()
11788 assign(rfbm, mkU64(3)); in dis_FXSAVE()
11810 assign(rfbm_0, binop(Iop_And64, mkexpr(rfbm), mkU64(1))); in gen_XRSTOR_SEQUENCE()
11811 assign(xstate_bv_0, binop(Iop_And64, mkexpr(xstate_bv), mkU64(1))); in gen_XRSTOR_SEQUENCE()
11814 gen_FINIT_SEQUENCE( binop(Iop_CmpNE64, mkexpr(rfbm_0), mkU64(0)) ); in gen_XRSTOR_SEQUENCE()
11825 d0->guard = binop(Iop_CmpNE64, mkexpr(restore_0), mkU64(0)); in gen_XRSTOR_SEQUENCE()
11868 assign(rfbm_1, binop(Iop_And64, mkexpr(rfbm), mkU64(2))); in gen_XRSTOR_SEQUENCE()
11869 assign(xstate_bv_1, binop(Iop_And64, mkexpr(xstate_bv), mkU64(2))); in gen_XRSTOR_SEQUENCE()
11871 IRExpr* rfbm_1e = binop(Iop_CmpNE64, mkexpr(rfbm_1), mkU64(0)); in gen_XRSTOR_SEQUENCE()
11872 IRExpr* restore_1e = binop(Iop_CmpNE64, mkexpr(restore_1), mkU64(0)); in gen_XRSTOR_SEQUENCE()
11877 assign(rfbm_1or2, binop(Iop_And64, mkexpr(rfbm), mkU64(6))); in gen_XRSTOR_SEQUENCE()
11878 assign(xstate_bv_1or2, binop(Iop_And64, mkexpr(xstate_bv), mkU64(6))); in gen_XRSTOR_SEQUENCE()
11881 IRExpr* rfbm_1or2e = binop(Iop_CmpNE64, mkexpr(rfbm_1or2), mkU64(0)); in gen_XRSTOR_SEQUENCE()
11882 IRExpr* restore_1or2e = binop(Iop_CmpNE64, mkexpr(restore_1or2), mkU64(0)); in gen_XRSTOR_SEQUENCE()
11885 putGuarded(OFFB_SSEROUND, rfbm_1or2e, mkU64(Irrm_NEAREST)); in gen_XRSTOR_SEQUENCE()
11913 d1->mAddr = binop(Iop_Add64, mkexpr(addr), mkU64(24)); in gen_XRSTOR_SEQUENCE()
11936 IRExpr* ea = binop(Iop_Add64, mkexpr(addr), mkU64(160 + reg * 16)); in gen_XRSTOR_SEQUENCE()
11954 assign(rfbm_2, binop(Iop_And64, mkexpr(rfbm), mkU64(4))); in gen_XRSTOR_SEQUENCE()
11955 assign(xstate_bv_2, binop(Iop_And64, mkexpr(xstate_bv), mkU64(4))); in gen_XRSTOR_SEQUENCE()
11958 IRExpr* rfbm_2e = binop(Iop_CmpNE64, mkexpr(rfbm_2), mkU64(0)); in gen_XRSTOR_SEQUENCE()
11959 IRExpr* restore_2e = binop(Iop_CmpNE64, mkexpr(restore_2), mkU64(0)); in gen_XRSTOR_SEQUENCE()
11966 IRExpr* ea = binop(Iop_Add64, mkexpr(addr), mkU64(576 + reg * 16)); in gen_XRSTOR_SEQUENCE()
12005 mkU64(aSSUMED_XCR0_VALUE))); in dis_XRSTOR()
12009 binop(Iop_Add64, mkexpr(addr), mkU64(512+0)))); in dis_XRSTOR()
12013 binop(Iop_Add64, mkexpr(addr), mkU64(512+8)))); in dis_XRSTOR()
12018 binop(Iop_Add64, mkexpr(addr), mkU64(512+16)))); in dis_XRSTOR()
12031 binop(Iop_And64, mkexpr(xstate_bv), mkU64(~aSSUMED_XCR0_VALUE)), in dis_XRSTOR()
12033 stmt( IRStmt_Exit(binop(Iop_CmpNE64, mkexpr(fault_if_nonzero), mkU64(0)), in dis_XRSTOR()
12071 assign(three, mkU64(3)); in dis_FXRSTOR()
12090 assign(tmp128, binop(Iop_64HLtoV128, mkU64(0), mkexpr(halfshift))); in math_PINSRW_128()
12092 assign(tmp128, binop(Iop_64HLtoV128, mkexpr(halfshift), mkU64(0))); in math_PINSRW_128()
13068 mkU64(4) ))); in dis_ESC_0F__SSE2()
13154 mkU64(8) ))); in dis_ESC_0F__SSE2()
13992 putXMMRegLane64( gregOfRexRM(pfx,modrm), 1, mkU64(0) ); in dis_ESC_0F__SSE2()
14126 mkU64( ~(lineszB-1) ))) ); in dis_ESC_0F__SSE2()
14128 stmt( IRStmt_Put(OFFB_CMLEN, mkU64(lineszB) ) ); in dis_ESC_0F__SSE2()
15145 mkexpr(addr), mkU64(16))) ); in dis_MOVDDUP_256()
15480 assign( sevens, mkU64(0x0707070707070707ULL) ); in math_PSHUFB_XMM()
15825 binop(Iop_And64, mkexpr(sV), mkU64(0x0707070707070707ULL)) in dis_ESC_0F38__SupSSE3()
16363 assign( res, mkU64(0) ); in dis_ESC_0F3A__SupSSE3()
16442 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(AMD64G_CC_OP_COPY) )); in dis_ESC_0F__SSE4()
16443 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0) )); in dis_ESC_0F__SSE4()
16444 stmt( IRStmt_Put( OFFB_CC_NDEP, mkU64(0) )); in dis_ESC_0F__SSE4()
16450 mkU64(0))), in dis_ESC_0F__SSE4()
16499 binop(Iop_CmpEQ64, mkexpr(res64), mkU64(0))), in dis_ESC_0F__SSE4()
16503 binop(Iop_CmpEQ64, mkexpr(src64), mkU64(0))), in dis_ESC_0F__SSE4()
16508 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(AMD64G_CC_OP_COPY) )); in dis_ESC_0F__SSE4()
16509 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0) )); in dis_ESC_0F__SSE4()
16510 stmt( IRStmt_Put( OFFB_CC_NDEP, mkU64(0) )); in dis_ESC_0F__SSE4()
16560 binop(Iop_CmpEQ64, mkexpr(res64), mkU64(0))), in dis_ESC_0F__SSE4()
16564 binop(Iop_CmpEQ64, mkexpr(src64), mkU64(0))), in dis_ESC_0F__SSE4()
16569 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(AMD64G_CC_OP_COPY) )); in dis_ESC_0F__SSE4()
16570 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0) )); in dis_ESC_0F__SSE4()
16571 stmt( IRStmt_Put( OFFB_CC_NDEP, mkU64(0) )); in dis_ESC_0F__SSE4()
16786 assign(t0, mkU64(0x8000000080000000ULL)); in finish_xTESTy()
16801 binop(Iop_Sub64, mkU64(0), mkexpr(and64)), in finish_xTESTy()
16808 binop(Iop_Sub64, mkU64(0), mkexpr(andn64)), in finish_xTESTy()
16817 binop(Iop_And64, mkexpr(z64), mkU64(AMD64G_CC_MASK_Z)), in finish_xTESTy()
16818 binop(Iop_And64, mkexpr(c64), mkU64(AMD64G_CC_MASK_C)))); in finish_xTESTy()
16821 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(AMD64G_CC_OP_COPY) )); in finish_xTESTy()
16822 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0) )); in finish_xTESTy()
16823 stmt( IRStmt_Put( OFFB_CC_NDEP, mkU64(0) )); in finish_xTESTy()
17674 IRExpr* opc4 = mkU64(opc); in dis_AESx()
17675 IRExpr* gstOffDe = mkU64(gstOffD); in dis_AESx()
17676 IRExpr* gstOffLe = mkU64(gstOffL); in dis_AESx()
17677 IRExpr* gstOffRe = mkU64(gstOffR); in dis_AESx()
17765 IRExpr* imme = mkU64(imm & 0xFF); in dis_AESKEYGENASSIST()
17766 IRExpr* gstOffLe = mkU64(gstOffL); in dis_AESKEYGENASSIST()
17767 IRExpr* gstOffRe = mkU64(gstOffR); in dis_AESKEYGENASSIST()
18276 mkU64(0xFFFFFFFF))); in dis_ESC_0F38__SSE4()
18602 stmt(IRStmt_Put(OFFB_CC_OP, mkU64(AMD64G_CC_OP_COPY))); in dis_PCMPISTRI_3A()
18604 stmt(IRStmt_Put(OFFB_CC_DEP2, mkU64(0))); in dis_PCMPISTRI_3A()
18605 stmt(IRStmt_Put(OFFB_CC_NDEP, mkU64(0))); in dis_PCMPISTRI_3A()
18694 IRExpr* opc4_and_imm = mkU64((opc << 8) | (imm & 0xFF)); in dis_PCMPxSTRx()
18695 IRExpr* gstOffLe = mkU64(gstOffL); in dis_PCMPxSTRx()
18696 IRExpr* gstOffRe = mkU64(gstOffR); in dis_PCMPxSTRx()
18697 IRExpr* edxIN = isISTRx ? mkU64(0) : getIRegRDX(8); in dis_PCMPxSTRx()
18698 IRExpr* eaxIN = isISTRx ? mkU64(0) : getIRegRAX(8); in dis_PCMPxSTRx()
18732 mkU64(0xFFFF))); in dis_PCMPxSTRx()
18741 binop(Iop_And64, mkexpr(resT), mkU64(0xFFFF)) in dis_PCMPxSTRx()
18743 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(AMD64G_CC_OP_COPY) )); in dis_PCMPxSTRx()
18744 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0) )); in dis_PCMPxSTRx()
18745 stmt( IRStmt_Put( OFFB_CC_NDEP, mkU64(0) )); in dis_PCMPxSTRx()
18763 assign(tmp128, binop(Iop_64HLtoV128, mkU64(0), mkexpr(halfshift))); in math_PINSRB_128()
18765 assign(tmp128, binop(Iop_64HLtoV128, mkexpr(halfshift), mkU64(0))); in math_PINSRB_128()
18818 assign(withZs, binop(Iop_64HLtoV128, mkU64(0), mkexpr(u64))); in math_PINSRQ_128()
18822 assign( withZs, binop(Iop_64HLtoV128, mkexpr(u64), mkU64(0))); in math_PINSRQ_128()
19014 mkU64( 0x80 | (imm8 & 7) )); in math_MPSADBW_128()
19017 mkU64( 0x00 | (imm8 & 7) )); in math_MPSADBW_128()
19094 args = mkIRExprVec_3(mkexpr(t0), mkexpr(t1), mkU64(0)); in math_PCLMULQDQ()
19097 args = mkIRExprVec_3(mkexpr(t0), mkexpr(t1), mkU64(1)); in math_PCLMULQDQ()
19161 binop(Iop_Add64, mkexpr(addr), mkU64(0) ))); in dis_ESC_0F3A__SSE4()
19163 binop(Iop_Add64, mkexpr(addr), mkU64(4) ))); in dis_ESC_0F3A__SSE4()
19165 binop(Iop_Add64, mkexpr(addr), mkU64(8) ))); in dis_ESC_0F3A__SSE4()
19167 binop(Iop_Add64, mkexpr(addr), mkU64(12) ))); in dis_ESC_0F3A__SSE4()
19223 binop(Iop_Add64, mkexpr(addr), mkU64(0) ))); in dis_ESC_0F3A__SSE4()
19225 binop(Iop_Add64, mkexpr(addr), mkU64(8) ))); in dis_ESC_0F3A__SSE4()
20088 assign(t2, binop(Iop_Sub64, getIReg64(R_RSP), mkU64(sz))); in dis_ESC_NONE()
20110 putIReg64(R_RSP, binop(Iop_Add64, mkexpr(t2), mkU64(sz))); in dis_ESC_NONE()
20167 assign( t1, binop(Iop_Sub64,getIReg64(R_RSP),mkU64(sz)) ); in dis_ESC_NONE()
20439 putIReg64(R_RSP, binop(Iop_Add64, mkexpr(t1), mkU64(sz)) ); in dis_ESC_NONE()
20529 assign( t1, binop(Iop_Sub64,getIReg64(R_RSP),mkU64(sz)) ); in dis_ESC_NONE()
20542 mkU64(1<<10))) in dis_ESC_NONE()
20552 mkU64(1<<21))) in dis_ESC_NONE()
20562 mkU64(1<<18))) in dis_ESC_NONE()
20586 putIReg64(R_RSP, binop(Iop_Add64, mkexpr(t2), mkU64(sz))); in dis_ESC_NONE()
20589 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(AMD64G_CC_OP_COPY) )); in dis_ESC_NONE()
20590 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0) )); in dis_ESC_NONE()
20594 mkU64( AMD64G_CC_MASK_C | AMD64G_CC_MASK_P in dis_ESC_NONE()
20609 mkU64(1))), in dis_ESC_NONE()
20610 mkU64(0xFFFFFFFFFFFFFFFFULL), in dis_ESC_NONE()
20611 mkU64(1))) in dis_ESC_NONE()
20621 mkU64(1))), in dis_ESC_NONE()
20622 mkU64(1), in dis_ESC_NONE()
20623 mkU64(0))) in dis_ESC_NONE()
20633 mkU64(1))), in dis_ESC_NONE()
20634 mkU64(1), in dis_ESC_NONE()
20635 mkU64(0))) in dis_ESC_NONE()
20662 assign( addr, handleAddrOverrides(vbi, pfx, mkU64(d64)) ); in dis_ESC_NONE()
20680 assign( addr, handleAddrOverrides(vbi, pfx, mkU64(d64)) ); in dis_ESC_NONE()
20817 putIRegRexB(8, pfx, opc-0xB8, mkU64(d64)); in dis_ESC_NONE()
20961 assign(t2, binop(Iop_Sub64, getIReg64(R_RSP), mkU64(8))); in dis_ESC_NONE()
20966 putIReg64(R_RSP, binop(Iop_Sub64, mkexpr(t2), mkU64(d64))); in dis_ESC_NONE()
20985 putIReg64(R_RSP, binop(Iop_Add64, mkexpr(t1), mkU64(8)) ); in dis_ESC_NONE()
21129 putIReg64(R_RCX, binop(Iop_Sub64, getIReg64(R_RCX), mkU64(1))); in dis_ESC_NONE()
21137 cond = binop(Iop_CmpNE64, count, mkU64(0)); in dis_ESC_NONE()
21170 mkU64(0)), in dis_ESC_NONE()
21180 mkU64(0)), in dis_ESC_NONE()
21193 assign(t1, mkU64( abyte & 0xFF )); in dis_ESC_NONE()
21200 assign(t1, mkU64( abyte & 0xFF )); in dis_ESC_NONE()
21230 mkIRExprVec_2( mkexpr(t1), mkU64(sz) ) in dis_ESC_NONE()
21242 assign( t1, mkU64( abyte & 0xFF ) ); in dis_ESC_NONE()
21249 assign( t1, mkU64( abyte & 0xFF ) ); in dis_ESC_NONE()
21279 mkU64(sz) ) in dis_ESC_NONE()
21292 assign(t1, binop(Iop_Sub64, getIReg64(R_RSP), mkU64(8))); in dis_ESC_NONE()
21294 storeLE( mkexpr(t1), mkU64(guest_RIP_bbstart+delta)); in dis_ESC_NONE()
21296 assign(t2, mkU64((Addr64)d64)); in dis_ESC_NONE()
21352 mkU64(AMD64G_CC_MASK_C))); in dis_ESC_NONE()
21357 mkU64(~AMD64G_CC_MASK_C))); in dis_ESC_NONE()
21362 mkU64(AMD64G_CC_MASK_C))); in dis_ESC_NONE()
21368 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(AMD64G_CC_OP_COPY) )); in dis_ESC_NONE()
21369 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0) )); in dis_ESC_NONE()
21373 stmt( IRStmt_Put( OFFB_CC_NDEP, mkU64(0) )); in dis_ESC_NONE()
21396 stmt( IRStmt_Put( OFFB_DFLAG, mkU64(1)) ); in dis_ESC_NONE()
21402 stmt( IRStmt_Put( OFFB_DFLAG, mkU64(-1ULL)) ); in dis_ESC_NONE()
21449 assign( m8, mkU64(0xFF00FF00FF00FF00ULL) ); in math_BSWAP()
21461 assign( m16, mkU64(0xFFFF0000FFFF0000ULL) ); in math_BSWAP()
21473 assign( m32, mkU64(0xFFFFFFFF00000000ULL) ); in math_BSWAP()
21568 mkU64(gregLO3ofRM(modrm)) ) in dis_ESC_0F()
21614 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(AMD64G_CC_OP_COPY) )); in dis_ESC_0F()
21615 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0) )); in dis_ESC_0F()
21616 stmt( IRStmt_Put( OFFB_CC_DEP1, mkU64(AMD64G_CC_MASK_Z) )); in dis_ESC_0F()
21619 stmt( IRStmt_Put( OFFB_CC_NDEP, mkU64(0) )); in dis_ESC_0F()
21662 putIReg64( R_RCX, mkU64(guest_RIP_next_assumed) ); in dis_ESC_0F()
21672 stmt( IRStmt_Put( OFFB_RIP, mkU64(guest_RIP_curr_instr) ) ); in dis_ESC_0F()
22171 IRExpr* zero = sz==4 ? mkU32(0) : mkU64(0); in dis_ESC_0F()
22273 mkU64(~AMD64G_CC_MASK_Z)), in dis_ESC_0F()
22276 unop(Iop_1Uto64, mkexpr(success)), mkU64(1)), in dis_ESC_0F()
22279 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(AMD64G_CC_OP_COPY) )); in dis_ESC_0F()
22281 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0) )); in dis_ESC_0F()
22285 stmt( IRStmt_Put( OFFB_CC_NDEP, mkU64(0) )); in dis_ESC_0F()
22759 binop(Iop_CmpLT64U, mkexpr(amt), mkU64(size)), in dis_AVX128_shiftV_byE()
22769 binop(Iop_CmpLT64U, mkexpr(amt), mkU64(size)), in dis_AVX128_shiftV_byE()
22833 binop(Iop_CmpLT64U, mkexpr(amt), mkU64(size)), in dis_AVX256_shiftV_byE()
22843 binop(Iop_CmpLT64U, mkexpr(amt), mkU64(size)), in dis_AVX256_shiftV_byE()
22943 size == 32 ? mkU32(size) : mkU64(size)), in dis_AVX_var_shiftV_byE()
22948 : size == 32 ? mkU32(0) : mkU64(0) in dis_AVX_var_shiftV_byE()
22960 ? mkexpr(res[i]) : mkU64(0) ); in dis_AVX_var_shiftV_byE()
26223 putXMMRegLane64( rG, 1, mkU64(0) ); in dis_ESC_0F__VEX()
27752 IRExpr* one = laneIs32 ? mkU32(1) : mkU64(1); in dis_VMASKMOV()
27762 mkU64(i * (laneIs32 ? 4 : 8))); in dis_VMASKMOV()
27767 data, ea, laneIs32 ? mkU32(0) : mkU64(0), mkexpr(cond) in dis_VMASKMOV()
27860 putYMMRegLane64( rV, i, mkU64(0) ); in dis_VGATHER()
27868 putYMMRegLane64( rV, 1, mkU64(0) ); in dis_VGATHER()
27869 putYMMRegLane64( rG, 1, mkU64(0) ); in dis_VGATHER()
29685 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(size == 8 in dis_ESC_0F38__VEX()
29689 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0)) ); in dis_ESC_0F38__VEX()
29722 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(size == 8 in dis_ESC_0F38__VEX()
29756 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(size == 8 in dis_ESC_0F38__VEX()
29790 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(size == 8 in dis_ESC_0F38__VEX()
29859 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(size == 8 in dis_ESC_0F38__VEX()
30067 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(size == 8 in dis_ESC_0F38__VEX()
30071 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0)) ); in dis_ESC_0F38__VEX()
30418 mkexpr(addr), mkU64(0)))); in dis_ESC_0F3A__VEX()
30420 mkexpr(addr), mkU64(16)))); in dis_ESC_0F3A__VEX()
31535 mkexpr(addr), mkU64(0)))); in dis_ESC_0F3A__VEX()
31537 mkexpr(addr), mkU64(16)))); in dis_ESC_0F3A__VEX()
31796 assign(t2, binop(Iop_Sub64, getIReg64(R_RSP), mkU64(8))); in disInstr_AMD64_WRK()
31798 storeLE( mkexpr(t2), mkU64(guest_RIP_bbstart+delta)); in disInstr_AMD64_WRK()
31814 stmt(IRStmt_Put(OFFB_CMSTART, mkU64(guest_RIP_curr_instr))); in disInstr_AMD64_WRK()
31815 stmt(IRStmt_Put(OFFB_CMLEN, mkU64(19))); in disInstr_AMD64_WRK()
31819 stmt( IRStmt_Put( OFFB_RIP, mkU64(guest_RIP_bbstart + delta) ) ); in disInstr_AMD64_WRK()
32107 stmt( IRStmt_Put( OFFB_RIP, mkU64(guest_RIP_curr_instr) ) ); in disInstr_AMD64_WRK()
32124 stmt( IRStmt_Put( OFFB_RIP, mkU64(guest_RIP_bbstart + delta) ) ); in disInstr_AMD64_WRK()
32128 stmt( IRStmt_Put( OFFB_RIP, mkU64(dres.continueAt) ) ); in disInstr_AMD64_WRK()