/external/valgrind/main/VEX/priv/ |
D | host_x86_isel.c | 209 static void addInstr ( ISelEnv* env, X86Instr* instr ) in addInstr() function 311 addInstr(env, in add_to_esp() 318 addInstr(env, in sub_from_esp() 352 addInstr(env, X86Instr_Push(X86RMI_Reg(r_vecRetAddr))); in pushArg() 356 addInstr(env, X86Instr_Push(X86RMI_Reg(hregX86_EBP()))); in pushArg() 362 addInstr(env, X86Instr_Push(iselIntExpr_RMI(env, arg))); in pushArg() 368 addInstr(env, X86Instr_Push(X86RMI_Reg(rHi))); in pushArg() 369 addInstr(env, X86Instr_Push(X86RMI_Reg(rLo))); in pushArg() 390 addInstr(env, X86Instr_Call( cc, toUInt(Ptr_to_ULong(cee->addr)), in callHelperAndClearArgs() 529 addInstr(env, mk_iMOVsd_RR( hregX86_ESP(), r_vecRetAddr )); in doHelperCall() [all …]
|
D | host_amd64_isel.c | 185 static void addInstr ( ISelEnv* env, AMD64Instr* instr ) in addInstr() function 332 addInstr(env, in add_to_rsp() 340 addInstr(env, in sub_from_rsp() 352 addInstr( env, AMD64Instr_Push(AMD64RMI_Imm( (UInt)uimm64 )) ); in push_uimm64() 355 addInstr( env, AMD64Instr_Imm64(uimm64, tmp) ); in push_uimm64() 356 addInstr( env, AMD64Instr_Push(AMD64RMI_Reg(tmp)) ); in push_uimm64() 580 addInstr(env, fastinstrs[i]); in doHelperCall() 603 addInstr(env, mk_iMOVsd_RR( hregAMD64_RSP(), r_vecRetAddr )); in doHelperCall() 608 addInstr(env, mk_iMOVsd_RR( hregAMD64_RSP(), r_vecRetAddr )); in doHelperCall() 616 addInstr(env, mk_iMOVsd_RR( hregAMD64_RBP(), tmpregs[i])); in doHelperCall() [all …]
|
D | host_s390_isel.c | 173 addInstr(ISelEnv *env, s390_insn *insn) in addInstr() function 550 addInstr(env, s390_insn_alu(4, S390_ALU_SUB, sp, s390_opnd_imm(16))); in doHelperCall() 551 addInstr(env, s390_insn_move(sizeof(ULong), r_vecRetAddr, sp)); in doHelperCall() 567 addInstr(env, s390_insn_move(sizeof(ULong), tmpregs[argreg], in doHelperCall() 572 addInstr(env, s390_insn_move(sizeof(ULong), tmpregs[argreg], in doHelperCall() 599 addInstr(env, s390_insn_move(size, finalreg, tmpregs[i])); in doHelperCall() 635 addInstr(env, s390_insn_helper_call(cc, (Addr64)target, n_args, in doHelperCall() 680 addInstr(env, s390_insn_load_immediate(4, mode, 4)); in set_bfp_rounding_mode_in_fpc() 681 addInstr(env, s390_insn_alu(4, S390_ALU_SUB, mode, s390_opnd_reg(ir))); in set_bfp_rounding_mode_in_fpc() 682 addInstr(env, s390_insn_alu(4, S390_ALU_AND, mode, s390_opnd_imm(3))); in set_bfp_rounding_mode_in_fpc() [all …]
|
D | host_mips_isel.c | 160 static void addInstr(ISelEnv * env, MIPSInstr * instr) in addInstr() function 197 addInstr(env, MIPSInstr_Alu(Malu_DADD, sp, sp, MIPSRH_Imm(True, in add_to_sp() 200 addInstr(env, MIPSInstr_Alu(Malu_ADD, sp, sp, MIPSRH_Imm(True, in add_to_sp() 209 addInstr(env, MIPSInstr_Alu(Malu_DSUB, sp, sp, in sub_from_sp() 212 addInstr(env, MIPSInstr_Alu(Malu_SUB, sp, sp, in sub_from_sp() 289 addInstr(env, MIPSInstr_Shft(Mshft_SLL, True, tmp, irrm, in set_MIPS_rounding_mode() 291 addInstr(env, MIPSInstr_Alu(Malu_XOR, tmp, irrm, MIPSRH_Reg(tmp))); in set_MIPS_rounding_mode() 292 addInstr(env, MIPSInstr_Alu(Malu_AND, irrm, tmp, MIPSRH_Imm(False, 3))); in set_MIPS_rounding_mode() 294 addInstr(env, MIPSInstr_MfFCSR(fcsr_old)); in set_MIPS_rounding_mode() 299 addInstr(env, MIPSInstr_Store(4, am_addr, fcsr_old, mode64)); in set_MIPS_rounding_mode() [all …]
|
D | host_arm_isel.c | 145 static void addInstr ( ISelEnv* env, ARMInstr* instr ) in addInstr() function 304 addInstr(env, ARMInstr_Imm32(rTmp, DEFAULT_FPSCR)); in set_VFP_rounding_default() 305 addInstr(env, ARMInstr_FPSCR(True/*toFPSCR*/, rTmp)); in set_VFP_rounding_default() 342 addInstr(env, ARMInstr_Shift(ARMsh_SHL, tL, irrm, ARMRI5_I5(1))); in set_VFP_rounding_mode() 343 addInstr(env, ARMInstr_Shift(ARMsh_SHR, tR, irrm, ARMRI5_I5(1))); in set_VFP_rounding_mode() 344 addInstr(env, ARMInstr_Alu(ARMalu_AND, tL, tL, ARMRI84_I84(2,0))); in set_VFP_rounding_mode() 345 addInstr(env, ARMInstr_Alu(ARMalu_AND, tR, tR, ARMRI84_I84(1,0))); in set_VFP_rounding_mode() 346 addInstr(env, ARMInstr_Alu(ARMalu_OR, t3, tL, ARMRI84_R(tR))); in set_VFP_rounding_mode() 347 addInstr(env, ARMInstr_Shift(ARMsh_SHL, t3, t3, ARMRI5_I5(22))); in set_VFP_rounding_mode() 348 addInstr(env, ARMInstr_FPSCR(True/*toFPSCR*/, t3)); in set_VFP_rounding_mode() [all …]
|
D | host_ppc_isel.c | 333 static void addInstr ( ISelEnv* env, PPCInstr* instr ) in addInstr() function 494 addInstr(env, PPCInstr_Alu( Palu_ADD, sp, sp, in add_to_sp() 502 addInstr(env, PPCInstr_Alu( Palu_SUB, sp, sp, in sub_from_sp() 516 addInstr(env, mk_iMOVds_RR(r, StackFramePtr(env->mode64))); in get_sp_aligned16() 518 addInstr(env, PPCInstr_Alu( Palu_ADD, r, r, in get_sp_aligned16() 521 addInstr(env, in get_sp_aligned16() 523 addInstr(env, PPCInstr_Alu(Palu_AND, r,r, PPCRH_Reg(align16))); in get_sp_aligned16() 545 addInstr(env, PPCInstr_Store( 4, am_addr0, r_srcHi, env->mode64 )); in mk_LoadRR32toFPR() 546 addInstr(env, PPCInstr_Store( 4, am_addr1, r_srcLo, env->mode64 )); in mk_LoadRR32toFPR() 549 addInstr(env, PPCInstr_FpLdSt(True/*load*/, 8, fr_dst, am_addr0)); in mk_LoadRR32toFPR() [all …]
|
D | host_arm64_isel.c | 139 static void addInstr ( ISelEnv* env, ARM64Instr* instr ) in addInstr() function 297 addInstr(env, ARM64Instr_Arith(r, hregARM64_X21(), in mk_baseblock_128bit_access_addr() 313 addInstr(env, ARM64Instr_Logic(dst, src, mask, ARM64lo_AND)); in widen_z_32_to_64() 323 addInstr(env, ARM64Instr_Shift(dst, src, n48, ARM64sh_SHL)); in widen_s_16_to_64() 324 addInstr(env, ARM64Instr_Shift(dst, dst, n48, ARM64sh_SAR)); in widen_s_16_to_64() 334 addInstr(env, ARM64Instr_Shift(dst, src, n48, ARM64sh_SHL)); in widen_z_16_to_64() 335 addInstr(env, ARM64Instr_Shift(dst, dst, n48, ARM64sh_SHR)); in widen_z_16_to_64() 345 addInstr(env, ARM64Instr_Shift(dst, src, n32, ARM64sh_SHL)); in widen_s_32_to_64() 346 addInstr(env, ARM64Instr_Shift(dst, dst, n32, ARM64sh_SAR)); in widen_s_32_to_64() 356 addInstr(env, ARM64Instr_Shift(dst, src, n56, ARM64sh_SHL)); in widen_s_8_to_64() [all …]
|