/arch/powerpc/kvm/ |
D | book3s_pr_papr.c | 38 long flags = kvmppc_get_gpr(vcpu, 4); in kvmppc_h_pr_enter() 39 long pte_index = kvmppc_get_gpr(vcpu, 5); in kvmppc_h_pr_enter() 70 hpte[0] = cpu_to_be64(kvmppc_get_gpr(vcpu, 6)); in kvmppc_h_pr_enter() 71 hpte[1] = cpu_to_be64(kvmppc_get_gpr(vcpu, 7)); in kvmppc_h_pr_enter() 88 unsigned long flags= kvmppc_get_gpr(vcpu, 4); in kvmppc_h_pr_remove() 89 unsigned long pte_index = kvmppc_get_gpr(vcpu, 5); in kvmppc_h_pr_remove() 90 unsigned long avpn = kvmppc_get_gpr(vcpu, 6); in kvmppc_h_pr_remove() 153 unsigned long tsh = kvmppc_get_gpr(vcpu, paramnr+(2*i)); in kvmppc_h_pr_bulk_remove() 154 unsigned long tsl = kvmppc_get_gpr(vcpu, paramnr+(2*i)+1); in kvmppc_h_pr_bulk_remove() 216 unsigned long flags = kvmppc_get_gpr(vcpu, 4); in kvmppc_h_pr_protect() [all …]
|
D | book3s_paired_singles.c | 680 ulong addr = ax_ra ? kvmppc_get_gpr(vcpu, ax_ra) : 0; in kvmppc_emulate_paired_single() 690 ulong addr = kvmppc_get_gpr(vcpu, ax_ra); in kvmppc_emulate_paired_single() 703 ulong addr = ax_ra ? kvmppc_get_gpr(vcpu, ax_ra) : 0; in kvmppc_emulate_paired_single() 713 ulong addr = kvmppc_get_gpr(vcpu, ax_ra); in kvmppc_emulate_paired_single() 733 ulong addr = ax_ra ? kvmppc_get_gpr(vcpu, ax_ra) : 0; in kvmppc_emulate_paired_single() 737 addr += kvmppc_get_gpr(vcpu, ax_rb); in kvmppc_emulate_paired_single() 747 ulong addr = kvmppc_get_gpr(vcpu, ax_ra); in kvmppc_emulate_paired_single() 751 addr += kvmppc_get_gpr(vcpu, ax_rb); in kvmppc_emulate_paired_single() 824 ulong addr = ax_ra ? kvmppc_get_gpr(vcpu, ax_ra) : 0; in kvmppc_emulate_paired_single() 828 addr += kvmppc_get_gpr(vcpu, ax_rb); in kvmppc_emulate_paired_single() [all …]
|
D | book3s_emulate.c | 299 ulong rs_val = kvmppc_get_gpr(vcpu, rs); in kvmppc_core_emulate_op_pr() 310 kvmppc_set_msr(vcpu, kvmppc_get_gpr(vcpu, rs)); in kvmppc_core_emulate_op_pr() 328 srnum = (kvmppc_get_gpr(vcpu, rb) >> 28) & 0xf; in kvmppc_core_emulate_op_pr() 339 kvmppc_get_gpr(vcpu, rs)); in kvmppc_core_emulate_op_pr() 343 (kvmppc_get_gpr(vcpu, rb) >> 28) & 0xf, in kvmppc_core_emulate_op_pr() 344 kvmppc_get_gpr(vcpu, rs)); in kvmppc_core_emulate_op_pr() 350 ulong addr = kvmppc_get_gpr(vcpu, rb); in kvmppc_core_emulate_op_pr() 358 ulong cmd = kvmppc_get_gpr(vcpu, 3); in kvmppc_core_emulate_op_pr() 372 ulong gpr = kvmppc_get_gpr(vcpu, 4 + i); in kvmppc_core_emulate_op_pr() 389 kvmppc_get_gpr(vcpu, rs), in kvmppc_core_emulate_op_pr() [all …]
|
D | book3s_hv.c | 461 r, kvmppc_get_gpr(vcpu, r), in kvmppc_dump_regs() 462 r+16, kvmppc_get_gpr(vcpu, r+16)); in kvmppc_dump_regs() 1036 unsigned long type = kvmppc_get_gpr(vcpu, 6); in kvmppc_nested_h_rpt_invalidate() 1045 pid = kvmppc_get_gpr(vcpu, 4); in kvmppc_nested_h_rpt_invalidate() 1046 pg_sizes = kvmppc_get_gpr(vcpu, 7); in kvmppc_nested_h_rpt_invalidate() 1047 start = kvmppc_get_gpr(vcpu, 8); in kvmppc_nested_h_rpt_invalidate() 1048 end = kvmppc_get_gpr(vcpu, 9); in kvmppc_nested_h_rpt_invalidate() 1094 unsigned long req = kvmppc_get_gpr(vcpu, 3); in kvmppc_pseries_do_hcall() 1106 ret = kvmppc_h_remove(vcpu, kvmppc_get_gpr(vcpu, 4), in kvmppc_pseries_do_hcall() 1107 kvmppc_get_gpr(vcpu, 5), in kvmppc_pseries_do_hcall() [all …]
|
D | trace_hv.h | 368 __entry->req = kvmppc_get_gpr(vcpu, 3); 369 __entry->gpr4 = kvmppc_get_gpr(vcpu, 4); 370 __entry->gpr5 = kvmppc_get_gpr(vcpu, 5); 371 __entry->gpr6 = kvmppc_get_gpr(vcpu, 6); 372 __entry->gpr7 = kvmppc_get_gpr(vcpu, 7); 395 __entry->hcall_rc = kvmppc_get_gpr(vcpu, 3);
|
D | book3s_hv_nested.c | 307 hv_ptr = kvmppc_get_gpr(vcpu, 4); in kvmhv_enter_nested_guest() 308 regs_ptr = kvmppc_get_gpr(vcpu, 5); in kvmhv_enter_nested_guest() 533 unsigned long ptcr = kvmppc_get_gpr(vcpu, 4); in kvmhv_set_partition_table() 561 int l1_lpid = kvmppc_get_gpr(vcpu, 4); in kvmhv_copy_tofrom_guest_nested() 562 int pid = kvmppc_get_gpr(vcpu, 5); in kvmhv_copy_tofrom_guest_nested() 563 gva_t eaddr = kvmppc_get_gpr(vcpu, 6); in kvmhv_copy_tofrom_guest_nested() 564 gpa_t gp_to = (gpa_t) kvmppc_get_gpr(vcpu, 7); in kvmhv_copy_tofrom_guest_nested() 565 gpa_t gp_from = (gpa_t) kvmppc_get_gpr(vcpu, 8); in kvmhv_copy_tofrom_guest_nested() 567 unsigned long n = kvmppc_get_gpr(vcpu, 9); in kvmhv_copy_tofrom_guest_nested() 1243 ret = kvmhv_emulate_priv_tlbie(vcpu, kvmppc_get_gpr(vcpu, 4), in kvmhv_do_nested_tlbie() [all …]
|
D | book3s.c | 119 crit_r1 = kvmppc_get_gpr(vcpu, 1); in kvmppc_critical_section() 579 regs->gpr[i] = kvmppc_get_gpr(vcpu, i); in kvm_arch_vcpu_ioctl_get_regs() 944 unsigned long size = kvmppc_get_gpr(vcpu, 4); in kvmppc_h_logical_ci_load() 945 unsigned long addr = kvmppc_get_gpr(vcpu, 5); in kvmppc_h_logical_ci_load() 986 unsigned long size = kvmppc_get_gpr(vcpu, 4); in kvmppc_h_logical_ci_store() 987 unsigned long addr = kvmppc_get_gpr(vcpu, 5); in kvmppc_h_logical_ci_store() 988 unsigned long val = kvmppc_get_gpr(vcpu, 6); in kvmppc_h_logical_ci_store()
|
D | book3s_hv_tm.c | 120 newmsr = kvmppc_get_gpr(vcpu, rs); in kvmhv_p9_tm_emulation() 196 ra = kvmppc_get_gpr(vcpu, ra) & 0xff; in kvmhv_p9_tm_emulation()
|
D | book3s_hv_tm_builtin.c | 76 newmsr = kvmppc_get_gpr(vcpu, rs); in kvmhv_p9_tm_emulation_early()
|
D | booke_emulate.c | 86 kvmppc_set_msr(vcpu, kvmppc_get_gpr(vcpu, rs)); in kvmppc_booke_emulate_op() 91 | (kvmppc_get_gpr(vcpu, rs) & MSR_EE); in kvmppc_booke_emulate_op()
|
D | book3s_xics.c | 893 rc = kvmppc_h_ipoll(vcpu, kvmppc_get_gpr(vcpu, 4)); in kvmppc_xics_hcall() 907 kvmppc_h_cppr(vcpu, kvmppc_get_gpr(vcpu, 4)); in kvmppc_xics_hcall() 910 rc = kvmppc_h_eoi(vcpu, kvmppc_get_gpr(vcpu, 4)); in kvmppc_xics_hcall() 913 rc = kvmppc_h_ipi(vcpu, kvmppc_get_gpr(vcpu, 4), in kvmppc_xics_hcall() 914 kvmppc_get_gpr(vcpu, 5)); in kvmppc_xics_hcall()
|
D | emulate.c | 78 ulong spr_val = kvmppc_get_gpr(vcpu, rs); in kvmppc_emulate_mtspr()
|
D | booke.c | 105 kvmppc_get_gpr(vcpu, i), in kvmppc_dump_vcpu() 106 kvmppc_get_gpr(vcpu, i+1), in kvmppc_dump_vcpu() 107 kvmppc_get_gpr(vcpu, i+2), in kvmppc_dump_vcpu() 108 kvmppc_get_gpr(vcpu, i+3)); in kvmppc_dump_vcpu() 422 ulong crit_r1 = kvmppc_get_gpr(vcpu, 1); in kvmppc_booke_irqprio_deliver() 1270 (((u32)kvmppc_get_gpr(vcpu, 0)) == KVM_SC_MAGIC_R0)) { in kvmppc_handle_exit() 1472 regs->gpr[i] = kvmppc_get_gpr(vcpu, i); in kvm_arch_vcpu_ioctl_get_regs()
|
D | book3s_rtas.c | 230 args_phys = kvmppc_get_gpr(vcpu, 4) & KVM_PAM; in kvmppc_rtas_hcall()
|
D | book3s_pr.c | 1304 ulong cmd = kvmppc_get_gpr(vcpu, 3); in kvmppc_handle_exit_pr() 1316 ulong gpr = kvmppc_get_gpr(vcpu, 4 + i); in kvmppc_handle_exit_pr() 1323 (((u32)kvmppc_get_gpr(vcpu, 3)) == OSI_SC_MAGIC_R3) && in kvmppc_handle_exit_pr() 1324 (((u32)kvmppc_get_gpr(vcpu, 4)) == OSI_SC_MAGIC_R4)) { in kvmppc_handle_exit_pr() 1331 gprs[i] = kvmppc_get_gpr(vcpu, i); in kvmppc_handle_exit_pr() 1335 (((u32)kvmppc_get_gpr(vcpu, 0)) == KVM_SC_MAGIC_R0)) { in kvmppc_handle_exit_pr()
|
D | powerpc.c | 164 int nr = kvmppc_get_gpr(vcpu, 11); in kvmppc_kvm_pv() 166 unsigned long __maybe_unused param1 = kvmppc_get_gpr(vcpu, 3); in kvmppc_kvm_pv() 167 unsigned long __maybe_unused param2 = kvmppc_get_gpr(vcpu, 4); in kvmppc_kvm_pv() 168 unsigned long __maybe_unused param3 = kvmppc_get_gpr(vcpu, 5); in kvmppc_kvm_pv() 169 unsigned long __maybe_unused param4 = kvmppc_get_gpr(vcpu, 6); in kvmppc_kvm_pv()
|
D | book3s_xive.c | 2792 return xive_vm_h_cppr(vcpu, kvmppc_get_gpr(vcpu, 4)); in kvmppc_xive_xics_hcall() 2794 return xive_vm_h_eoi(vcpu, kvmppc_get_gpr(vcpu, 4)); in kvmppc_xive_xics_hcall() 2796 return xive_vm_h_ipi(vcpu, kvmppc_get_gpr(vcpu, 4), in kvmppc_xive_xics_hcall() 2797 kvmppc_get_gpr(vcpu, 5)); in kvmppc_xive_xics_hcall() 2799 return xive_vm_h_ipoll(vcpu, kvmppc_get_gpr(vcpu, 4)); in kvmppc_xive_xics_hcall()
|
/arch/powerpc/include/asm/ |
D | kvm_booke.h | 31 static inline ulong kvmppc_get_gpr(struct kvm_vcpu *vcpu, int num) in kvmppc_get_gpr() function
|
D | kvm_ppc.h | 1059 ea = kvmppc_get_gpr(vcpu, rb); in kvmppc_get_ea_indexed() 1061 ea += kvmppc_get_gpr(vcpu, ra); in kvmppc_get_ea_indexed()
|
D | kvm_book3s.h | 340 static inline ulong kvmppc_get_gpr(struct kvm_vcpu *vcpu, int num) in kvmppc_get_gpr() function
|