/external/valgrind/main/VEX/priv/ |
D | guest_s390_helpers.c | 687 #define mkU64(v) IRExpr_Const(IRConst_U64(v)) macro 834 return unop(Iop_1Uto32, binop(Iop_CmpEQ64, cc_dep1, mkU64(0))); in guest_s390x_spechelper() 837 return unop(Iop_1Uto32, binop(Iop_CmpNE64, cc_dep1, mkU64(0))); in guest_s390x_spechelper() 840 return unop(Iop_1Uto32, binop(Iop_CmpLT64S, cc_dep1, mkU64(0))); in guest_s390x_spechelper() 843 return unop(Iop_1Uto32, binop(Iop_CmpLE64S, cc_dep1, mkU64(0))); in guest_s390x_spechelper() 847 return unop(Iop_1Uto32, binop(Iop_CmpLT64S, mkU64(0), cc_dep1)); in guest_s390x_spechelper() 850 return unop(Iop_1Uto32, binop(Iop_CmpLE64S, mkU64(0), cc_dep1)); in guest_s390x_spechelper() 879 return unop(Iop_1Uto32, binop(Iop_CmpEQ64, cc_dep1, mkU64(0))); in guest_s390x_spechelper() 882 return unop(Iop_1Uto32, binop(Iop_CmpNE64, cc_dep1, mkU64(0))); in guest_s390x_spechelper() 985 mkU64(0))); in guest_s390x_spechelper() [all …]
|
D | guest_s390_toIR.c | 197 mkU64(ULong value) in mkU64() function 260 irsb->next = mkU64(callee_address); in call_function_and_chase() 330 irsb->next = mkU64(target); in always_goto_and_chase() 344 stmt(IRStmt_Put(OFFSET_s390x_IP_AT_SYSCALL, mkU64(guest_IA_curr_instr))); in system_call() 349 irsb->next = mkU64(guest_IA_next_instr); in system_call() 457 op = mkU64(opc); in s390_cc_thunk_put1() 459 dep2 = mkU64(0); in s390_cc_thunk_put1() 460 ndep = mkU64(0); in s390_cc_thunk_put1() 471 op = mkU64(opc); in s390_cc_thunk_put2() 474 ndep = mkU64(0); in s390_cc_thunk_put2() [all …]
|
D | guest_amd64_helpers.c | 879 # define mkU64(_n) IRExpr_Const(IRConst_U64(_n)) in guest_amd64_spechelper() macro 914 mkU64(0))); in guest_amd64_spechelper() 1091 binop(Iop_And64, cc_dep1, mkU64(0xFF)), in guest_amd64_spechelper() 1092 binop(Iop_And64, cc_dep2, mkU64(0xFF)))); in guest_amd64_spechelper() 1107 mkU64(1)); in guest_amd64_spechelper() 1118 mkU64(1)), in guest_amd64_spechelper() 1119 mkU64(1)); in guest_amd64_spechelper() 1127 binop(Iop_CmpEQ64, cc_dep1, mkU64(0))); in guest_amd64_spechelper() 1132 binop(Iop_CmpNE64, cc_dep1, mkU64(0))); in guest_amd64_spechelper() 1145 mkU64(0))); in guest_amd64_spechelper() [all …]
|
D | guest_amd64_toIR.c | 286 static IRExpr* mkU64 ( ULong i ) in mkU64() function 297 case Ity_I64: return mkU64(i); in mkU() 1533 = mkIRExprVec_5( mkU64(cond), in mk_amd64g_calculate_condition() 1661 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(ccOp)) ); in setFlags_DEP1_DEP2() 1687 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(ccOp)) ); in setFlags_DEP1() 1689 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0)) ); in setFlags_DEP1() 1728 mkU64(ccOp))) ); in setFlags_DEP1_DEP2_shift() 1759 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(ccOp)) ); in setFlags_INC_DEC() 1761 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0)) ); in setFlags_INC_DEC() 1773 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(base_op+0) ) ); in setFlags_MUL() [all …]
|
D | guest_ppc_toIR.c | 558 static IRExpr* mkU64 ( ULong i ) in mkU64() function 919 return ty == Ity_I64 ? mkU64(imm64) : mkU32((UInt)imm64); in mkSzImm() 936 mkU64(extend_s_16to64(imm16)) : in mkSzExtendS16() 945 mkU64(extend_s_32to64(imm32)) : in mkSzExtendS32() 1335 assign( mask[0], mkU64( 0x5555555555555555ULL ) ); in gen_POPCOUNT() 1336 assign( mask[1], mkU64( 0x3333333333333333ULL ) ); in gen_POPCOUNT() 1337 assign( mask[2], mkU64( 0x0F0F0F0F0F0F0F0FULL ) ); in gen_POPCOUNT() 1338 assign( mask[3], mkU64( 0x00FF00FF00FF00FFULL ) ); in gen_POPCOUNT() 1339 assign( mask[4], mkU64( 0x0000FFFF0000FFFFULL ) ); in gen_POPCOUNT() 1340 assign( mask[5], mkU64( 0x00000000FFFFFFFFULL ) ); in gen_POPCOUNT() [all …]
|
D | guest_arm_toIR.c | 274 static IRExpr* mkU64 ( ULong i ) in mkU64() function 359 return binop(Iop_64HLtoV128, mkU64(i), mkU64(i)); in mkU128() 2893 assign(old_mask, mkU64(0)); in dis_neon_vtb() 2894 assign(old_res, mkU64(0)); in dis_neon_vtb() 2908 assign(new_arg, binop(Iop_Sub8x8, mkexpr(old_arg), mkU64(imm))); in dis_neon_vtb() 2909 assign(cur_mask, binop(cmp, mkU64(imm), mkexpr(old_arg))); in dis_neon_vtb() 3052 imm_val = binop(Iop_64HLtoV128, mkU64(imm), mkU64(imm)); in dis_neon_data_3same() 3055 imm_val = mkU64(imm); in dis_neon_data_3same() 3232 mkU64(one), in dis_neon_data_3same() 3233 mkU64(one))), in dis_neon_data_3same() [all …]
|
D | guest_x86_toIR.c | 683 static IRExpr* mkU64 ( ULong i ) in mkU64() function 5615 mkU64(0), in dis_MMX_shiftG_byE() 5677 ? mkU64(0) in dis_MMX_shiftE_imm() 7567 assign(one32x2, mkU64( (1ULL << 32) + 1 )); in dis_PMULHRSW_helper() 7631 assign( zero, mkU64(0) ); in dis_PSIGN_helper() 7672 assign( zero, mkU64(0) ); in dis_PABS_helper() 8791 putXMMRegLane64( gregOfRM(modrm), 1, mkU64(0) ); in disInstr_X86_WRK() 10343 putXMMRegLane64( gregOfRM(modrm), 1, mkU64(0) ); in disInstr_X86_WRK() 10351 putXMMRegLane64( gregOfRM(modrm), 1, mkU64(0) ); in disInstr_X86_WRK() 11167 assign( lo64r, mkU64(0) ); in disInstr_X86_WRK() [all …]
|
D | host_s390_isel.c | 112 mkU64(ULong value) in mkU64() function 1403 addr_lo = IRExpr_Binop(Iop_Add64, addr_hi, mkU64(8)); in s390_isel_float128_expr_wrk()
|
/external/valgrind/main/coregrind/ |
D | m_translate.c | 873 static IRExpr* mkU64 ( ULong n ) { in mkU64() function 914 IRExpr*(*mkU)(ULong) = mkU64; in gen_PUSH() 1004 IRExpr*(*mkU)(ULong) = mkU64; in gen_POP() 1089 addStmtToIRSB( bb, IRStmt_Put( offB_LR, mkU64( bogus_RA )) ); in gen_push_and_set_LR_R2() 1090 addStmtToIRSB( bb, IRStmt_Put( offB_GPR2, mkU64( new_R2_value )) ); in gen_push_and_set_LR_R2() 1115 bb->next = IRExpr_Binop(Iop_And64, IRExpr_RdTmp(old_LR), mkU64(~(3ULL))); in gen_pop_R2_LR_then_bLR() 1173 nraddr_szB == 8 ? mkU64(0) : mkU32(0) in mk_preamble__set_NRADDR_to_zero() 1182 VG_WORDSIZE==8 ? mkU64(0) : mkU32(0) in mk_preamble__set_NRADDR_to_zero()
|
/external/valgrind/main/exp-dhat/ |
D | dh_main.c | 769 #define mkU64(_n) IRExpr_Const(IRConst_U64(_n)) macro 791 IRStmt* st2 = assign(t2, binop(Iop_Add64, mkexpr(t1), mkU64(n))); in add_counter_update() 847 : binop(Iop_Sub64, mkexpr(sp), mkU64(rz_szB))) in addMemEvent() 865 : binop(Iop_CmpLT64U, mkU64(THRESH), mkexpr(diff))) in addMemEvent() 1022 #undef mkU64
|
/external/valgrind/main/VEX/ |
D | test_main.c | 697 #define mkU64(_n) IRExpr_Const(IRConst_U64(_n)) macro 957 tmp1 = assignNew(mce, Ity_I1, binop(Iop_CmpNE64, vbits, mkU64(0))); in mkPCastTo() 2050 eBias = tyAddr==Ity_I32 ? mkU32(bias) : mkU64(bias); in expr2vbits_LDle_WRK() 2266 eBias0 = tyAddr==Ity_I32 ? mkU32(bias) : mkU64(bias); in do_shadow_STle() 2273 eBias8 = tyAddr==Ity_I32 ? mkU32(bias+8) : mkU64(bias+8); in do_shadow_STle() 2292 eBias = tyAddr==Ity_I32 ? mkU32(bias) : mkU64(bias); in do_shadow_STle()
|
/external/valgrind/main/memcheck/ |
D | mc_translate.c | 418 #define mkU64(_n) IRExpr_Const(IRConst_U64(_n)) macro 850 top = mkU64(0xFFFFFFFFFFFFFFFFULL); in expensiveCmpEQorNE() 966 threeLeft1 = m64 ? mkU64(3<<1) : mkU32(3<<1); in doCmpORD() 989 sevenLeft1 = m64 ? mkU64(7<<1) : mkU32(7<<1); in doCmpORD() 3624 eBias = tyAddr==Ity_I32 ? mkU32(bias) : mkU64(bias); in expr2vbits_Load_WRK() 3938 eBiasLo64 = tyAddr==Ity_I32 ? mkU32(bias+offLo64) : mkU64(bias+offLo64); in do_shadow_Store() 3946 eBiasHi64 = tyAddr==Ity_I32 ? mkU32(bias+offHi64) : mkU64(bias+offHi64); in do_shadow_Store() 3971 IRAtom* eBias = tyAddr==Ity_I32 ? mkU32(bias) : mkU64(bias); in do_shadow_Store() 4549 elemSzB = 8; zero = mkU64(0); in do_shadow_CAS_double() 5345 : mkU64( (Long)(Int)offset ); in gen_load_b() [all …]
|
/external/valgrind/main/helgrind/ |
D | hg_main.c | 4100 #define mkU64(_n) IRExpr_Const(IRConst_U64(_n)) macro 4241 : binop(Iop_Add64, mkexpr(addr_minus_sp), mkU64(rz_szB))) in instrument_mem_access() 4250 : binop(Iop_CmpLT64U, mkU64(THRESH), mkexpr(diff))) in instrument_mem_access() 4507 #undef mkU64
|