• Home
  • Raw
  • Download

Lines Matching refs:mkU64

563 static IRExpr* mkU64 ( ULong i )  in mkU64()  function
924 return ty == Ity_I64 ? mkU64(imm64) : mkU32((UInt)imm64); in mkSzImm()
941 mkU64(extend_s_16to64(imm16)) : in mkSzExtendS16()
950 mkU64(extend_s_32to64(imm32)) : in mkSzExtendS32()
1395 assign( mask[0], mkU64( 0x5555555555555555ULL ) ); in gen_POPCOUNT()
1396 assign( mask[1], mkU64( 0x3333333333333333ULL ) ); in gen_POPCOUNT()
1397 assign( mask[2], mkU64( 0x0F0F0F0F0F0F0F0FULL ) ); in gen_POPCOUNT()
1398 assign( mask[3], mkU64( 0x00FF00FF00FF00FFULL ) ); in gen_POPCOUNT()
1399 assign( mask[4], mkU64( 0x0000FFFF0000FFFFULL ) ); in gen_POPCOUNT()
1400 assign( mask[5], mkU64( 0x00000000FFFFFFFFULL ) ); in gen_POPCOUNT()
1526 binop(Iop_And64, mkexpr(addr), mkU64(align-1)), in gen_SIGBUS_if_misaligned()
1527 mkU64(0)), in gen_SIGBUS_if_misaligned()
1563 binop(Iop_Sub64, getIReg(1), mkU64(szB)), in make_redzone_AbiHint()
1732 binop(Iop_CmpORD64S, result, mkU64(0))) ); in set_CR0()
2032 = AND3( XOR3(argL,argR,mkU64(-1)), in set_XER_OV_64()
2034 mkU64(INT64_MIN) ); in set_XER_OV_64()
2045 binop(Iop_CmpEQ64, argL, mkU64(INT64_MIN)), in set_XER_OV_64()
2046 binop(Iop_CmpEQ64, argR, mkU64(-1)) in set_XER_OV_64()
2048 binop(Iop_CmpEQ64, argR, mkU64(0) ) in set_XER_OV_64()
2055 = binop(Iop_CmpEQ64, argR, mkU64(0)); in set_XER_OV_64()
2074 = binop(Iop_CmpEQ64, argL, mkU64(INT64_MIN)); in set_XER_OV_64()
2082 = AND3( XOR3(NOT(argL),argR,mkU64(-1)), in set_XER_OV_64()
2084 mkU64(INT64_MIN) ); in set_XER_OV_64()
2098 = mkOR1( binop( Iop_CmpEQ64, argR, mkU64( 0 ) ), in set_XER_OV_64()
2099 mkAND1( binop( Iop_CmpEQ64, res, mkU64( 0 ) ), in set_XER_OV_64()
2100 mkAND1( binop( Iop_CmpNE64, argL, mkU64( 0 ) ), in set_XER_OV_64()
2101 binop( Iop_CmpNE64, argR, mkU64( 0 ) ) ) ) ); in set_XER_OV_64()
2106 xer_ov = mkOR1( binop( Iop_CmpEQ64, argR, mkU64( 0 ) ), in set_XER_OV_64()
2289 binop(Iop_CmpEQ64, oldca, mkU64(1)), in set_XER_CA_64()
2303 binop(Iop_CmpEQ64, oldca, mkU64(1)), in set_XER_CA_64()
2332 binop(Iop_Shl64, mkU64(1), in set_XER_CA_64()
2334 mkU64(1) ) in set_XER_CA_64()
2340 unop(Iop_1Uto8, binop(Iop_CmpLT64U, mkU64(31), argR)), in set_XER_CA_64()
2342 unop(Iop_1Uto32, binop(Iop_CmpNE64, xer_ca, mkU64(0))), in set_XER_CA_64()
2363 binop(Iop_Shl64, mkU64(1), in set_XER_CA_64()
2365 mkU64(1) ) in set_XER_CA_64()
2369 = unop(Iop_1Uto32, binop(Iop_CmpNE64, xer_ca, mkU64(0))); in set_XER_CA_64()
2386 binop(Iop_Shl64, mkU64(1), in set_XER_CA_64()
2388 mkU64(1) ) in set_XER_CA_64()
2394 unop(Iop_1Uto8, binop(Iop_CmpLT64U, mkU64(63), argR)), in set_XER_CA_64()
2396 unop(Iop_1Uto32, binop(Iop_CmpNE64, xer_ca, mkU64(0))), in set_XER_CA_64()
2418 binop(Iop_Shl64, mkU64(1), in set_XER_CA_64()
2420 mkU64(1) ) in set_XER_CA_64()
2424 = unop(Iop_1Uto32, binop(Iop_CmpNE64, xer_ca, mkU64(0))); in set_XER_CA_64()
2566 unop( Iop_64HIto32, mkU64( mask ) ) ); in getGST_masked_upper()
2778 mkU64( NONZERO_FRAC_MASK ) )
2803 mkU64( 0x7ff ) ) ); in fp_exp_part()
2853 assign( sign_less_part, binop( Iop_And64, mkexpr( src ), mkU64( SIGN_MASK ) ) ); in is_Zero()
2915 mkU64( signbit_mask ) ) ), in getNegatedResult()
2932 mkU64( ~signbit_mask ) ), in getNegatedResult()
3699 ? mkU64(0) : mkU32(0); in dis_int_cmp()
3718 ? mkU64(0) : mkU32(0); in dis_int_cmp()
3950 irx = binop(Iop_CmpNE64, mkexpr(rS), mkU64(0)); in dis_int_logic()
3952 mkU64(64), in dis_int_logic()
3963 binop( Iop_64HLtoV128, mkU64(0), mkexpr(rS) ), in dis_int_logic()
3964 binop( Iop_64HLtoV128, mkU64(0), mkexpr(rB) ) in dis_int_logic()
4057 IRExpr * res = binop(Iop_And64, mkU64(0), mkU64(0)); in dis_int_logic()
4067 binop( Iop_And64, mkU64( BPERMD_IDX_MASK ), rS_expr ) ); in dis_int_logic()
4069 binop( Iop_CmpLT64U, mkexpr( idx_tmp ), mkU64( 64 ) ) ); in dis_int_logic()
4087 mkU64( BPERMD_BIT_MASK ), in dis_int_logic()
4296 binop(Iop_And64, mkexpr(rot), mkU64(mask64)), in dis_int_rot()
4297 binop(Iop_And64, getIReg(rA_addr), mkU64(~mask64))) ); in dis_int_rot()
4330 assign( rA, binop(Iop_And64, mkexpr(rot), mkU64(mask64)) ); in dis_int_rot()
4376 assign( rA, binop(Iop_And64, mkexpr(rot), mkU64(mask64)) ); in dis_int_rot()
4407 assign( rA, binop(Iop_And64, r, mkU64(mask64)) ); in dis_int_rot()
4413 assign( rA, binop(Iop_And64, r, mkU64(mask64)) ); in dis_int_rot()
4423 assign( rA, binop(Iop_And64, r, mkU64(mask64)) ); in dis_int_rot()
4446 assign( rA, binop(Iop_And64, r, mkU64(mask64)) ); in dis_int_rot()
4463 assign( rA, binop(Iop_And64, r, mkU64(mask64)) ); in dis_int_rot()
4475 binop(Iop_And64, mkU64(mask64), r), in dis_int_rot()
4476 binop(Iop_And64, mkU64(~mask64), in dis_int_rot()
5317 assign(t_tgt, mode64 ? mkU64(tgt) : mkU32(tgt) ); in dis_branch()
5583 IRExpr* const0 = is32bit ? mkU32(0) : mkU64(0); in do_trap()
5584 IRExpr* const2 = is32bit ? mkU32(2) : mkU64(2); in do_trap()
5585 IRExpr* const4 = is32bit ? mkU32(4) : mkU64(4); in do_trap()
5586 IRExpr* const8 = is32bit ? mkU32(8) : mkU64(8); in do_trap()
5685 uncond = do_trap( TO, getIReg(rA_addr), mkU64( (ULong)simm16 ), cia ); in dis_trapi()
6162 binop(Iop_And64, mkexpr(rB), mkU64(63)))), in dis_int_shift()
6177 assign( sh_amt, binop(Iop_And64, mkU64(0x7F), mkexpr(rB)) ); in dis_int_shift()
6180 binop(Iop_CmpLT64U, mkU64(63), in dis_int_shift()
6188 mkU64(63)) )) in dis_int_shift()
6206 mkU64(sh_imm), in dis_int_shift()
6225 binop(Iop_And64, mkexpr(rB), mkU64(63)))), in dis_int_shift()
6336 ty == Ity_I64 ? mkU64( 4 ) : mkU32( 4 ) ); in dis_int_ldst_rev()
6730 mkU64( ~((ULong)clearszB-1) )) ); in dis_cache_manage()
6733 irx_addr = binop( Iop_Add64, mkexpr(addr), mkU64(i*8) ); in dis_cache_manage()
6734 storeBE( irx_addr, mkU64(0) ); in dis_cache_manage()
8352 assign( EA_lo, binop(Iop_Add64, mkexpr(EA_hi), mkU64(8)) ); in dis_fp_pair()
8498 putGST_masked( PPC_GST_FPSCR, mkU64( 1 <<( 31 - crbD ) ), in dis_fp_scr()
8535 putGST_masked( PPC_GST_FPSCR, mkU64( 0 ), 1ULL << ( 31 - crbD ) ); in dis_fp_scr()
9712 binop( Iop_Sub64, mkU64( 397 ), in dis_dfp_quantize_sig_rrnd()
9713 binop( Iop_And64, mkU64( 0xF ), in dis_dfp_quantize_sig_rrnd()
9714 unop( Iop_Not64, mkU64( TE_value ) ) in dis_dfp_quantize_sig_rrnd()
9720 binop( Iop_Add64, mkU64( 398 ), mkU64( TE_value ) ) ) ); in dis_dfp_quantize_sig_rrnd()
9724 unop( Iop_ReinterpI64asD64, mkU64( 1 ) ) ) ); in dis_dfp_quantize_sig_rrnd()
9796 binop( Iop_Sub64, mkU64( 6175 ), in dis_dfp_quantize_sig_rrndq()
9797 binop( Iop_And64, mkU64( 0xF ), in dis_dfp_quantize_sig_rrndq()
9798 unop( Iop_Not64, mkU64( TE_value ) ) in dis_dfp_quantize_sig_rrndq()
9804 binop( Iop_Add64, mkU64( 6176 ), mkU64( TE_value ) ) in dis_dfp_quantize_sig_rrndq()
9811 unop( Iop_ReinterpI64asD64, mkU64( 1 ) ) ) ) ); in dis_dfp_quantize_sig_rrndq()
10276 mkU64( 0x7FFFFFFFFFFFFFFFULL ) ) ) ); in dis_dfp_class_test()
10283 assign( frAI64_lo, mkU64( 0 ) ); in dis_dfp_class_test()
10294 mkU64( 0x2234000000000001ULL ) ) ); // dfp 1.0 in dis_dfp_class_test()
10296 unop( Iop_ReinterpI64asD64, mkU64( 398 - 383 ) ) ); in dis_dfp_class_test()
10312 mkU64( 0x2238000000000000ULL ) ) ) ); in dis_dfp_class_test()
10336 unop( Iop_ReinterpI64asD64, mkU64( 6176 - 6143 ) ) ); in dis_dfp_class_test()
10340 mkU64( 0x2234000000000001ULL ) ) ) ); // dfp 1.0 in dis_dfp_class_test()
10354 mkU64( 0x7FFFFFFFFFFFFFFFULL ) ) ), in dis_dfp_class_test()
10364 mkU64( 0x0ULL ) ) ) ) ); in dis_dfp_class_test()
10787 mkU64( DFP_LONG_BIAS ) ), in dis_dfp_bcd()
10876 mkU64( DFP_LONG_BIAS ) ), in dis_dfp_bcd()
11267 unop( Iop_ReinterpI64asD64, mkU64( DFP_EXTND_BIAS ) ), in dis_dfp_bcdq()
11701 mkexpr( xB ) ), mkU64( 0 ) ) ); in dis_vx_conv()
11712 mkU64( 0ULL ) ) ); in dis_vx_conv()
11722 mkU64( 0ULL ) ) ); in dis_vx_conv()
11732 mkU64( 0 ) ) ); in dis_vx_conv()
11742 mkU64( 0 ) ) ); in dis_vx_conv()
11825 mkU64( 0ULL ) ) ); in dis_vx_conv()
11837 mkU64( 0ULL ) ) ); in dis_vx_conv()
11846 mkU64( 0ULL ) ) ); in dis_vx_conv()
12738 binop(Iop_Or64, mkexpr(frB_I64), mkU64(SNAN_MASK))), in _get_maxmin_fp_NaN()
12740 binop(Iop_Or64, mkexpr(frA_I64), mkU64(SNAN_MASK))); in _get_maxmin_fp_NaN()
12806 mkU64( isMin ? MINUS_ZERO : 0ULL ) ) ); in get_max_min_fp()
12893 mkU64( SNAN_MASK ), in _do_vsx_fp_roundToInt()
13117 mkU64( SIGN_BIT ) ), in dis_vxv_misc()
13120 mkU64( SIGN_MASK ) ) ), in dis_vxv_misc()
13124 mkU64( SIGN_BIT ) ), in dis_vxv_misc()
13127 mkU64( SIGN_MASK ) ) ) ) ); in dis_vxv_misc()
13399 mkU64( 0 ) ) ); in dis_vxs_arith()
13407 mkU64( 0 ) ) ); in dis_vxs_arith()
13421 mkU64( 0 ) ) ); in dis_vxs_arith()
13436 mkU64( 0 ) ) ); in dis_vxs_arith()
13461 mkU64( 0 ) ) ); in dis_vxs_arith()
13480 … putVSReg( XT, binop( Iop_64HLtoV128, mkexpr( getNegatedResult(msubResult) ), mkU64( 0 ) ) ); in dis_vxs_arith()
13491 mkU64( 0 ) ) ); in dis_vxs_arith()
13499 mkU64( 0 ) ) ); in dis_vxs_arith()
13507 mkU64( 0 ) ) ); in dis_vxs_arith()
13869 putVSReg( XT, binop( Iop_64HLtoV128, get_max_min_fp(frA, frB, isMin), mkU64( 0 ) ) ); in dis_vxs_misc()
13890 mkU64( 0 ) ) ); in dis_vxs_misc()
13919 mkU64( 0 ) ) ); in dis_vxs_misc()
14026 high_addr = binop( addOp, mkexpr( EA ), ty == Ity_I64 ? mkU64( ea_off ) in dis_vx_load()
14050 ty == Ity_I64 ? mkU64( ea_off ) : mkU32( ea_off ) ); in dis_vx_load()
14054 ty == Ity_I64 ? mkU64( ea_off ) : mkU32( ea_off ) ); in dis_vx_load()
14058 ty == Ity_I64 ? mkU64( ea_off ) : mkU32( ea_off ) ); in dis_vx_load()
14113 storeBE( binop( mkSzOp( ty, Iop_Add8 ), mkexpr( EA ), ty == Ity_I64 ? mkU64( 8 ) in dis_vx_store()
14134 ty == Ity_I64 ? mkU64( ea_off ) : mkU32( ea_off ) ); in dis_vx_store()
14138 ty == Ity_I64 ? mkU64( ea_off ) : mkU32( ea_off ) ); in dis_vx_store()
14142 ty == Ity_I64 ? mkU64( ea_off ) : mkU32( ea_off ) ); in dis_vx_store()
14918 putVReg( vD_addr, mkV128from4x64S( mkU64(0), mkexpr(z2), in dis_av_arith()
14919 mkU64(0), mkexpr(z0)) ); in dis_av_arith()
14936 putVReg( vD_addr, mkV128from4x64S( mkU64(0), mkU64(0), in dis_av_arith()
14937 mkU64(0), mkexpr(z0)) ); in dis_av_arith()
17476 rA == 0 ? (mode64 ? mkU64(0) : mkU32(0)) in disInstr_PPC_WRK()