/external/valgrind/main/VEX/priv/ |
D | ir_defs.c | 1584 IRConst* c = LibVEX_Alloc(sizeof(IRConst)); in IRConst_U1() 1593 IRConst* c = LibVEX_Alloc(sizeof(IRConst)); in IRConst_U8() 1600 IRConst* c = LibVEX_Alloc(sizeof(IRConst)); in IRConst_U16() 1607 IRConst* c = LibVEX_Alloc(sizeof(IRConst)); in IRConst_U32() 1614 IRConst* c = LibVEX_Alloc(sizeof(IRConst)); in IRConst_U64() 1621 IRConst* c = LibVEX_Alloc(sizeof(IRConst)); in IRConst_F32() 1628 IRConst* c = LibVEX_Alloc(sizeof(IRConst)); in IRConst_F32i() 1635 IRConst* c = LibVEX_Alloc(sizeof(IRConst)); in IRConst_F64() 1642 IRConst* c = LibVEX_Alloc(sizeof(IRConst)); in IRConst_F64i() 1649 IRConst* c = LibVEX_Alloc(sizeof(IRConst)); in IRConst_V128() [all …]
|
D | host_x86_defs.c | 108 *arr = LibVEX_Alloc(*nregs * sizeof(HReg)); in getAllocableRegs_X86() 162 X86AMode* am = LibVEX_Alloc(sizeof(X86AMode)); in X86AMode_IR() 169 X86AMode* am = LibVEX_Alloc(sizeof(X86AMode)); in X86AMode_IRRS() 244 X86RMI* op = LibVEX_Alloc(sizeof(X86RMI)); in X86RMI_Imm() 250 X86RMI* op = LibVEX_Alloc(sizeof(X86RMI)); in X86RMI_Reg() 256 X86RMI* op = LibVEX_Alloc(sizeof(X86RMI)); in X86RMI_Mem() 315 X86RI* op = LibVEX_Alloc(sizeof(X86RI)); in X86RI_Imm() 321 X86RI* op = LibVEX_Alloc(sizeof(X86RI)); in X86RI_Reg() 371 X86RM* op = LibVEX_Alloc(sizeof(X86RM)); in X86RM_Reg() 377 X86RM* op = LibVEX_Alloc(sizeof(X86RM)); in X86RM_Mem() [all …]
|
D | host_amd64_defs.c | 139 *arr = LibVEX_Alloc(*nregs * sizeof(HReg)); in getAllocableRegs_AMD64() 150 *arr = LibVEX_Alloc(*nregs * sizeof(HReg)); in getAllocableRegs_AMD64() 206 AMD64AMode* am = LibVEX_Alloc(sizeof(AMD64AMode)); in AMD64AMode_IR() 213 AMD64AMode* am = LibVEX_Alloc(sizeof(AMD64AMode)); in AMD64AMode_IRRS() 276 AMD64RMI* op = LibVEX_Alloc(sizeof(AMD64RMI)); in AMD64RMI_Imm() 282 AMD64RMI* op = LibVEX_Alloc(sizeof(AMD64RMI)); in AMD64RMI_Reg() 288 AMD64RMI* op = LibVEX_Alloc(sizeof(AMD64RMI)); in AMD64RMI_Mem() 356 AMD64RI* op = LibVEX_Alloc(sizeof(AMD64RI)); in AMD64RI_Imm() 362 AMD64RI* op = LibVEX_Alloc(sizeof(AMD64RI)); in AMD64RI_Reg() 412 AMD64RM* op = LibVEX_Alloc(sizeof(AMD64RM)); in AMD64RM_Reg() [all …]
|
D | host_arm64_defs.c | 135 *arr = LibVEX_Alloc(*nregs * sizeof(HReg)); in getAllocableRegs_ARM64() 231 ARM64AMode* am = LibVEX_Alloc(sizeof(ARM64AMode)); in ARM64AMode_RI9() 240 ARM64AMode* am = LibVEX_Alloc(sizeof(ARM64AMode)); in ARM64AMode_RI12() 254 ARM64AMode* am = LibVEX_Alloc(sizeof(ARM64AMode)); in ARM64AMode_RR() 466 ARM64RIA* riA = LibVEX_Alloc(sizeof(ARM64RIA)); in ARM64RIA_I12() 475 ARM64RIA* riA = LibVEX_Alloc(sizeof(ARM64RIA)); in ARM64RIA_R() 523 ARM64RIL* riL = LibVEX_Alloc(sizeof(ARM64RIL)); in ARM64RIL_I13() 534 ARM64RIL* riL = LibVEX_Alloc(sizeof(ARM64RIL)); in ARM64RIL_R() 584 ARM64RI6* ri6 = LibVEX_Alloc(sizeof(ARM64RI6)); in ARM64RI6_I6() 591 ARM64RI6* ri6 = LibVEX_Alloc(sizeof(ARM64RI6)); in ARM64RI6_R() [all …]
|
D | host_generic_regs.c | 195 HInstrArray* ha = LibVEX_Alloc(sizeof(HInstrArray)); in newHInstrArray() 198 ha->arr = LibVEX_Alloc(ha->arr_size * sizeof(HInstr*)); in newHInstrArray() 211 HInstr** arr2 = LibVEX_Alloc(ha->arr_size * 2 * sizeof(HInstr*)); in addHInstr()
|
D | host_arm_defs.c | 126 *arr = LibVEX_Alloc(*nregs * sizeof(HReg)); in getAllocableRegs_ARM() 220 ARMAMode1* am = LibVEX_Alloc(sizeof(ARMAMode1)); in ARMAMode1_RI() 228 ARMAMode1* am = LibVEX_Alloc(sizeof(ARMAMode1)); in ARMAMode1_RRS() 288 ARMAMode2* am = LibVEX_Alloc(sizeof(ARMAMode2)); in ARMAMode2_RI() 296 ARMAMode2* am = LibVEX_Alloc(sizeof(ARMAMode2)); in ARMAMode2_RR() 354 ARMAModeV* am = LibVEX_Alloc(sizeof(ARMAModeV)); in mkARMAModeV() 380 ARMAModeN* am = LibVEX_Alloc(sizeof(ARMAModeN)); in mkARMAModeN_RR() 388 ARMAModeN* am = LibVEX_Alloc(sizeof(ARMAModeN)); in mkARMAModeN_R() 438 ARMRI84* ri84 = LibVEX_Alloc(sizeof(ARMRI84)); in ARMRI84_I84() 447 ARMRI84* ri84 = LibVEX_Alloc(sizeof(ARMRI84)); in ARMRI84_R() [all …]
|
D | host_ppc_defs.c | 203 *arr = LibVEX_Alloc(*nregs * sizeof(HReg)); in getAllocableRegs_PPC() 320 PPCAMode* am = LibVEX_Alloc(sizeof(PPCAMode)); in PPCAMode_IR() 328 PPCAMode* am = LibVEX_Alloc(sizeof(PPCAMode)); in PPCAMode_RR() 397 PPCRH* op = LibVEX_Alloc(sizeof(PPCRH)); in PPCRH_Imm() 409 PPCRH* op = LibVEX_Alloc(sizeof(PPCRH)); in PPCRH_Reg() 462 PPCRI* op = LibVEX_Alloc(sizeof(PPCRI)); in PPCRI_Imm() 468 PPCRI* op = LibVEX_Alloc(sizeof(PPCRI)); in PPCRI_Reg() 518 PPCVI5s* op = LibVEX_Alloc(sizeof(PPCVI5s)); in PPCVI5s_Imm() 525 PPCVI5s* op = LibVEX_Alloc(sizeof(PPCVI5s)); in PPCVI5s_Reg() 785 PPCInstr* i = LibVEX_Alloc(sizeof(PPCInstr)); in PPCInstr_LI() [all …]
|
D | host_mips_defs.c | 571 *arr = LibVEX_Alloc(*nregs * sizeof(HReg)); in getAllocableRegs_MIPS() 874 MIPSAMode *am = LibVEX_Alloc(sizeof(MIPSAMode)); in MIPSAMode_IR() 884 MIPSAMode *am = LibVEX_Alloc(sizeof(MIPSAMode)); in MIPSAMode_RR() 1007 MIPSRH *op = LibVEX_Alloc(sizeof(MIPSRH)); in MIPSRH_Imm() 1021 MIPSRH *op = LibVEX_Alloc(sizeof(MIPSRH)); in MIPSRH_Reg() 1183 MIPSInstr *i = LibVEX_Alloc(sizeof(MIPSInstr)); in MIPSInstr_LI() 1192 MIPSInstr *i = LibVEX_Alloc(sizeof(MIPSInstr)); in MIPSInstr_Alu() 1204 MIPSInstr *i = LibVEX_Alloc(sizeof(MIPSInstr)); in MIPSInstr_Shft() 1216 MIPSInstr *i = LibVEX_Alloc(sizeof(MIPSInstr)); in MIPSInstr_Unary() 1227 MIPSInstr *i = LibVEX_Alloc(sizeof(MIPSInstr)); in MIPSInstr_Cmp() [all …]
|
D | host_generic_reg_alloc2.c | 228 arr2 = LibVEX_Alloc(2 * *size * sizeof(RRegLR)); in ensureRRLRspace() 466 rreg_state = LibVEX_Alloc(n_rregs * sizeof(RRegState)); in doRegisterAllocation() 467 vreg_state = LibVEX_Alloc(n_vregs * sizeof(Short)); in doRegisterAllocation() 495 vreg_lrs = LibVEX_Alloc(sizeof(VRegLR) * n_vregs); in doRegisterAllocation() 515 rreg_lrs_la = LibVEX_Alloc(rreg_lrs_size * sizeof(RRegLR)); in doRegisterAllocation() 521 rreg_live_after = LibVEX_Alloc(n_available_real_regs * sizeof(Int)); in doRegisterAllocation() 522 rreg_dead_before = LibVEX_Alloc(n_available_real_regs * sizeof(Int)); in doRegisterAllocation() 748 rreg_lrs_db = LibVEX_Alloc(rreg_lrs_used * sizeof(RRegLR)); in doRegisterAllocation()
|
D | host_s390_defs.c | 130 *arr = LibVEX_Alloc(*nregs * sizeof(HReg)); in s390_hreg_get_allocable() 208 s390_amode *am = LibVEX_Alloc(sizeof(s390_amode)); in s390_amode_b12() 225 s390_amode *am = LibVEX_Alloc(sizeof(s390_amode)); in s390_amode_b20() 242 s390_amode *am = LibVEX_Alloc(sizeof(s390_amode)); in s390_amode_bx12() 261 s390_amode *am = LibVEX_Alloc(sizeof(s390_amode)); in s390_amode_bx20() 5355 s390_insn *insn = LibVEX_Alloc(sizeof(s390_insn)); in s390_insn_load() 5371 s390_insn *insn = LibVEX_Alloc(sizeof(s390_insn)); in s390_insn_store() 5387 s390_insn *insn = LibVEX_Alloc(sizeof(s390_insn)); in s390_insn_move() 5403 s390_insn *insn = LibVEX_Alloc(sizeof(s390_insn)); in s390_insn_memcpy() 5424 s390_insn *insn = LibVEX_Alloc(sizeof(s390_insn)); in s390_insn_cond_move() [all …]
|
D | ir_opt.c | 191 HashHW* h = LibVEX_Alloc(sizeof(HashHW)); in newHHW() 194 h->inuse = LibVEX_Alloc(h->size * sizeof(Bool)); in newHHW() 195 h->key = LibVEX_Alloc(h->size * sizeof(HWord)); in newHHW() 196 h->val = LibVEX_Alloc(h->size * sizeof(HWord)); in newHHW() 236 Bool* inuse2 = LibVEX_Alloc(2 * h->size * sizeof(Bool)); in addToHHW() 237 HWord* key2 = LibVEX_Alloc(2 * h->size * sizeof(HWord)); in addToHHW() 238 HWord* val2 = LibVEX_Alloc(2 * h->size * sizeof(HWord)); in addToHHW() 2698 IRExpr** env = LibVEX_Alloc(n_tmps * sizeof(IRExpr*)); in cprop_BB() 3034 Bool* set = LibVEX_Alloc(n_tmps * sizeof(Bool)); in do_deadcode_BB() 3316 *outs = LibVEX_Alloc(n * sizeof(TmpOrConst)); in irExprVec_to_TmpOrConsts() [all …]
|
D | host_arm64_isel.c | 6999 env = LibVEX_Alloc(sizeof(ISelEnv)); in iselSB_ARM64() 7011 env->vregmap = LibVEX_Alloc(env->n_vregmap * sizeof(HReg)); in iselSB_ARM64() 7012 env->vregmapHI = LibVEX_Alloc(env->n_vregmap * sizeof(HReg)); in iselSB_ARM64()
|
D | host_s390_isel.c | 4098 env = LibVEX_Alloc(sizeof(ISelEnv)); in iselSB_S390() 4120 env->vregmap = LibVEX_Alloc(env->n_vregmap * sizeof(HReg)); in iselSB_S390() 4121 env->vregmapHI = LibVEX_Alloc(env->n_vregmap * sizeof(HReg)); in iselSB_S390()
|
D | host_ppc_isel.c | 5930 env = LibVEX_Alloc(sizeof(ISelEnv)); in iselSB_PPC() 5949 env->vregmapLo = LibVEX_Alloc(env->n_vregmap * sizeof(HReg)); in iselSB_PPC() 5950 env->vregmapMedLo = LibVEX_Alloc(env->n_vregmap * sizeof(HReg)); in iselSB_PPC() 5955 env->vregmapMedHi = LibVEX_Alloc(env->n_vregmap * sizeof(HReg)); in iselSB_PPC() 5956 env->vregmapHi = LibVEX_Alloc(env->n_vregmap * sizeof(HReg)); in iselSB_PPC()
|
D | host_mips_isel.c | 4160 env = LibVEX_Alloc(sizeof(ISelEnv)); in iselSB_MIPS() 4174 env->vregmap = LibVEX_Alloc(env->n_vregmap * sizeof(HReg)); in iselSB_MIPS() 4175 env->vregmapHI = LibVEX_Alloc(env->n_vregmap * sizeof(HReg)); in iselSB_MIPS()
|
D | host_x86_isel.c | 4444 env = LibVEX_Alloc(sizeof(ISelEnv)); in iselSB_X86() 4456 env->vregmap = LibVEX_Alloc(env->n_vregmap * sizeof(HReg)); in iselSB_X86() 4457 env->vregmapHI = LibVEX_Alloc(env->n_vregmap * sizeof(HReg)); in iselSB_X86()
|
D | host_amd64_isel.c | 4881 env = LibVEX_Alloc(sizeof(ISelEnv)); in iselSB_AMD64() 4893 env->vregmap = LibVEX_Alloc(env->n_vregmap * sizeof(HReg)); in iselSB_AMD64() 4894 env->vregmapHI = LibVEX_Alloc(env->n_vregmap * sizeof(HReg)); in iselSB_AMD64()
|
D | host_arm_isel.c | 6341 env = LibVEX_Alloc(sizeof(ISelEnv)); in iselSB_ARM() 6353 env->vregmap = LibVEX_Alloc(env->n_vregmap * sizeof(HReg)); in iselSB_ARM() 6354 env->vregmapHI = LibVEX_Alloc(env->n_vregmap * sizeof(HReg)); in iselSB_ARM()
|
/external/valgrind/main/docs/internals/ |
D | performance.txt | 8 - Julian made LibVEX_Alloc() inlinable. Saved a couple of percent.
|
/external/valgrind/main/VEX/pub/ |
D | libvex.h | 478 static inline void* LibVEX_Alloc ( Int nbytes ) in LibVEX_Alloc() function
|
/external/valgrind/main/VEX/ |
D | test_main.c | 2618 mce.tmpMap = LibVEX_Alloc(mce.n_originalTmps * sizeof(IRTemp)); in mc_instrument()
|
/external/valgrind/main/ |
D | NEWS | 741 283671 Robustize alignment computation in LibVEX_Alloc
|