| /kernel/linux/linux-5.10/arch/powerpc/kvm/ |
| D | booke_emulate.c | 26 vcpu->arch.regs.nip = vcpu->arch.shared->srr0; in kvmppc_emul_rfi() 27 kvmppc_set_msr(vcpu, vcpu->arch.shared->srr1); in kvmppc_emul_rfi() 32 vcpu->arch.regs.nip = vcpu->arch.dsrr0; in kvmppc_emul_rfdi() 33 kvmppc_set_msr(vcpu, vcpu->arch.dsrr1); in kvmppc_emul_rfdi() 38 vcpu->arch.regs.nip = vcpu->arch.csrr0; in kvmppc_emul_rfci() 39 kvmppc_set_msr(vcpu, vcpu->arch.csrr1); in kvmppc_emul_rfci() 80 kvmppc_set_gpr(vcpu, rt, vcpu->arch.shared->msr); in kvmppc_booke_emulate_op() 90 vcpu->arch.shared->msr = (vcpu->arch.shared->msr & ~MSR_EE) in kvmppc_booke_emulate_op() 96 vcpu->arch.shared->msr = (vcpu->arch.shared->msr & ~MSR_EE) in kvmppc_booke_emulate_op() 127 vcpu->arch.shared->dar = spr_val; in kvmppc_booke_emulate_mtspr() [all …]
|
| D | booke.c | 68 printk("pc: %08lx msr: %08llx\n", vcpu->arch.regs.nip, in kvmppc_dump_vcpu() 69 vcpu->arch.shared->msr); in kvmppc_dump_vcpu() 70 printk("lr: %08lx ctr: %08lx\n", vcpu->arch.regs.link, in kvmppc_dump_vcpu() 71 vcpu->arch.regs.ctr); in kvmppc_dump_vcpu() 72 printk("srr0: %08llx srr1: %08llx\n", vcpu->arch.shared->srr0, in kvmppc_dump_vcpu() 73 vcpu->arch.shared->srr1); in kvmppc_dump_vcpu() 75 printk("exceptions: %08lx\n", vcpu->arch.pending_exceptions); in kvmppc_dump_vcpu() 93 vcpu->arch.shadow_msr &= ~MSR_SPE; in kvmppc_vcpu_disable_spe() 103 vcpu->arch.shadow_msr |= MSR_SPE; in kvmppc_vcpu_enable_spe() 109 if (vcpu->arch.shared->msr & MSR_SPE) { in kvmppc_vcpu_sync_spe() [all …]
|
| D | timing.c | 27 mutex_lock(&vcpu->arch.exit_timing_lock); in kvmppc_init_timing_stats() 29 vcpu->arch.last_exit_type = 0xDEAD; in kvmppc_init_timing_stats() 31 vcpu->arch.timing_count_type[i] = 0; in kvmppc_init_timing_stats() 32 vcpu->arch.timing_max_duration[i] = 0; in kvmppc_init_timing_stats() 33 vcpu->arch.timing_min_duration[i] = 0xFFFFFFFF; in kvmppc_init_timing_stats() 34 vcpu->arch.timing_sum_duration[i] = 0; in kvmppc_init_timing_stats() 35 vcpu->arch.timing_sum_quad_duration[i] = 0; in kvmppc_init_timing_stats() 37 vcpu->arch.timing_last_exit = 0; in kvmppc_init_timing_stats() 38 vcpu->arch.timing_exit.tv64 = 0; in kvmppc_init_timing_stats() 39 vcpu->arch.timing_last_enter.tv64 = 0; in kvmppc_init_timing_stats() [all …]
|
| D | book3s_hv.c | 14 * This file is derived from arch/powerpc/kvm/book3s.c, 134 return kvm->arch.nested_enable && kvm_is_radix(kvm); in nesting_enabled() 241 cpu = READ_ONCE(vcpu->arch.thread_cpu); in kvmppc_fast_vcpu_kick_hv() 277 * Updates to busy_stolen are protected by arch.tbacct_lock; 307 struct kvmppc_vcore *vc = vcpu->arch.vcore; in kvmppc_core_vcpu_load_hv() 319 spin_lock_irqsave(&vcpu->arch.tbacct_lock, flags); in kvmppc_core_vcpu_load_hv() 320 if (vcpu->arch.state == KVMPPC_VCPU_BUSY_IN_HOST && in kvmppc_core_vcpu_load_hv() 321 vcpu->arch.busy_preempt != TB_NIL) { in kvmppc_core_vcpu_load_hv() 322 vcpu->arch.busy_stolen += mftb() - vcpu->arch.busy_preempt; in kvmppc_core_vcpu_load_hv() 323 vcpu->arch.busy_preempt = TB_NIL; in kvmppc_core_vcpu_load_hv() [all …]
|
| D | book3s_hv_tm.c | 19 u64 msr = vcpu->arch.shregs.msr; in emulate_tx_failure() 21 tfiar = vcpu->arch.regs.nip & ~0x3ull; in emulate_tx_failure() 23 if (MSR_TM_SUSPENDED(vcpu->arch.shregs.msr)) in emulate_tx_failure() 29 vcpu->arch.tfiar = tfiar; in emulate_tx_failure() 31 vcpu->arch.texasr = (vcpu->arch.texasr & 0x3ffffff) | texasr; in emulate_tx_failure() 37 * instruction image is in vcpu->arch.emul_inst. If the guest was in 44 u32 instr = vcpu->arch.emul_inst; in kvmhv_p9_tm_emulation() 45 u64 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation() 63 newmsr = vcpu->arch.shregs.srr1; in kvmhv_p9_tm_emulation() 69 vcpu->arch.shregs.msr = newmsr; in kvmhv_p9_tm_emulation() [all …]
|
| D | emulate_loadstore.c | 86 vcpu->arch.mmio_vsx_copy_nums = 0; in kvmppc_emulate_loadstore() 87 vcpu->arch.mmio_vsx_offset = 0; in kvmppc_emulate_loadstore() 88 vcpu->arch.mmio_copy_type = KVMPPC_VSX_COPY_NONE; in kvmppc_emulate_loadstore() 89 vcpu->arch.mmio_sp64_extend = 0; in kvmppc_emulate_loadstore() 90 vcpu->arch.mmio_sign_extend = 0; in kvmppc_emulate_loadstore() 91 vcpu->arch.mmio_vmx_copy_nums = 0; in kvmppc_emulate_loadstore() 92 vcpu->arch.mmio_vmx_offset = 0; in kvmppc_emulate_loadstore() 93 vcpu->arch.mmio_host_swabbed = 0; in kvmppc_emulate_loadstore() 96 vcpu->arch.regs.msr = vcpu->arch.shared->msr; in kvmppc_emulate_loadstore() 97 if (analyse_instr(&op, &vcpu->arch.regs, ppc_inst(inst)) == 0) { in kvmppc_emulate_loadstore() [all …]
|
| D | book3s_emulate.c | 77 if (vcpu->arch.papr_enabled && (level > PRIV_SUPER)) in spr_allowed() 90 memcpy(&vcpu->arch.gpr_tm[0], &vcpu->arch.regs.gpr[0], in kvmppc_copyto_vcpu_tm() 91 sizeof(vcpu->arch.gpr_tm)); in kvmppc_copyto_vcpu_tm() 92 memcpy(&vcpu->arch.fp_tm, &vcpu->arch.fp, in kvmppc_copyto_vcpu_tm() 94 memcpy(&vcpu->arch.vr_tm, &vcpu->arch.vr, in kvmppc_copyto_vcpu_tm() 96 vcpu->arch.ppr_tm = vcpu->arch.ppr; in kvmppc_copyto_vcpu_tm() 97 vcpu->arch.dscr_tm = vcpu->arch.dscr; in kvmppc_copyto_vcpu_tm() 98 vcpu->arch.amr_tm = vcpu->arch.amr; in kvmppc_copyto_vcpu_tm() 99 vcpu->arch.ctr_tm = vcpu->arch.regs.ctr; in kvmppc_copyto_vcpu_tm() 100 vcpu->arch.tar_tm = vcpu->arch.tar; in kvmppc_copyto_vcpu_tm() [all …]
|
| D | e500_emulate.c | 8 * This file is derived from arch/powerpc/kvm/44x_emulate.c, 53 ulong param = vcpu->arch.regs.gpr[rb]; in kvmppc_e500_emul_msgclr() 59 clear_bit(prio, &vcpu->arch.pending_exceptions); in kvmppc_e500_emul_msgclr() 65 ulong param = vcpu->arch.regs.gpr[rb]; in kvmppc_e500_emul_msgsnd() 75 int cpir = cvcpu->arch.shared->pir; in kvmppc_e500_emul_msgsnd() 77 set_bit(prio, &cvcpu->arch.pending_exceptions); in kvmppc_e500_emul_msgsnd() 94 vcpu->run->debug.arch.address = vcpu->arch.regs.nip; in kvmppc_e500_emul_ehpriv() 95 vcpu->run->debug.arch.status = 0; in kvmppc_e500_emul_ehpriv() 225 vcpu->arch.shared->mas0 = spr_val; in kvmppc_core_emulate_mtspr_e500() 228 vcpu->arch.shared->mas1 = spr_val; in kvmppc_core_emulate_mtspr_e500() [all …]
|
| /kernel/linux/linux-4.19/arch/powerpc/kvm/ |
| D | booke_emulate.c | 37 vcpu->arch.regs.nip = vcpu->arch.shared->srr0; in kvmppc_emul_rfi() 38 kvmppc_set_msr(vcpu, vcpu->arch.shared->srr1); in kvmppc_emul_rfi() 43 vcpu->arch.regs.nip = vcpu->arch.dsrr0; in kvmppc_emul_rfdi() 44 kvmppc_set_msr(vcpu, vcpu->arch.dsrr1); in kvmppc_emul_rfdi() 49 vcpu->arch.regs.nip = vcpu->arch.csrr0; in kvmppc_emul_rfci() 50 kvmppc_set_msr(vcpu, vcpu->arch.csrr1); in kvmppc_emul_rfci() 91 kvmppc_set_gpr(vcpu, rt, vcpu->arch.shared->msr); in kvmppc_booke_emulate_op() 101 vcpu->arch.shared->msr = (vcpu->arch.shared->msr & ~MSR_EE) in kvmppc_booke_emulate_op() 107 vcpu->arch.shared->msr = (vcpu->arch.shared->msr & ~MSR_EE) in kvmppc_booke_emulate_op() 138 vcpu->arch.shared->dar = spr_val; in kvmppc_booke_emulate_mtspr() [all …]
|
| D | booke.c | 80 printk("pc: %08lx msr: %08llx\n", vcpu->arch.regs.nip, in kvmppc_dump_vcpu() 81 vcpu->arch.shared->msr); in kvmppc_dump_vcpu() 82 printk("lr: %08lx ctr: %08lx\n", vcpu->arch.regs.link, in kvmppc_dump_vcpu() 83 vcpu->arch.regs.ctr); in kvmppc_dump_vcpu() 84 printk("srr0: %08llx srr1: %08llx\n", vcpu->arch.shared->srr0, in kvmppc_dump_vcpu() 85 vcpu->arch.shared->srr1); in kvmppc_dump_vcpu() 87 printk("exceptions: %08lx\n", vcpu->arch.pending_exceptions); in kvmppc_dump_vcpu() 105 vcpu->arch.shadow_msr &= ~MSR_SPE; in kvmppc_vcpu_disable_spe() 115 vcpu->arch.shadow_msr |= MSR_SPE; in kvmppc_vcpu_enable_spe() 121 if (vcpu->arch.shared->msr & MSR_SPE) { in kvmppc_vcpu_sync_spe() [all …]
|
| D | timing.c | 38 mutex_lock(&vcpu->arch.exit_timing_lock); in kvmppc_init_timing_stats() 40 vcpu->arch.last_exit_type = 0xDEAD; in kvmppc_init_timing_stats() 42 vcpu->arch.timing_count_type[i] = 0; in kvmppc_init_timing_stats() 43 vcpu->arch.timing_max_duration[i] = 0; in kvmppc_init_timing_stats() 44 vcpu->arch.timing_min_duration[i] = 0xFFFFFFFF; in kvmppc_init_timing_stats() 45 vcpu->arch.timing_sum_duration[i] = 0; in kvmppc_init_timing_stats() 46 vcpu->arch.timing_sum_quad_duration[i] = 0; in kvmppc_init_timing_stats() 48 vcpu->arch.timing_last_exit = 0; in kvmppc_init_timing_stats() 49 vcpu->arch.timing_exit.tv64 = 0; in kvmppc_init_timing_stats() 50 vcpu->arch.timing_last_enter.tv64 = 0; in kvmppc_init_timing_stats() [all …]
|
| D | book3s_hv_tm.c | 22 u64 msr = vcpu->arch.shregs.msr; in emulate_tx_failure() 24 tfiar = vcpu->arch.regs.nip & ~0x3ull; in emulate_tx_failure() 26 if (MSR_TM_SUSPENDED(vcpu->arch.shregs.msr)) in emulate_tx_failure() 32 vcpu->arch.tfiar = tfiar; in emulate_tx_failure() 34 vcpu->arch.texasr = (vcpu->arch.texasr & 0x3ffffff) | texasr; in emulate_tx_failure() 40 * instruction image is in vcpu->arch.emul_inst. If the guest was in 47 u32 instr = vcpu->arch.emul_inst; in kvmhv_p9_tm_emulation() 48 u64 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation() 66 newmsr = vcpu->arch.shregs.srr1; in kvmhv_p9_tm_emulation() 72 vcpu->arch.shregs.msr = newmsr; in kvmhv_p9_tm_emulation() [all …]
|
| D | book3s_hv.c | 13 * This file is derived from arch/powerpc/kvm/book3s.c, 223 cpu = READ_ONCE(vcpu->arch.thread_cpu); in kvmppc_fast_vcpu_kick_hv() 259 * Updates to busy_stolen are protected by arch.tbacct_lock; 289 struct kvmppc_vcore *vc = vcpu->arch.vcore; in kvmppc_core_vcpu_load_hv() 301 spin_lock_irqsave(&vcpu->arch.tbacct_lock, flags); in kvmppc_core_vcpu_load_hv() 302 if (vcpu->arch.state == KVMPPC_VCPU_BUSY_IN_HOST && in kvmppc_core_vcpu_load_hv() 303 vcpu->arch.busy_preempt != TB_NIL) { in kvmppc_core_vcpu_load_hv() 304 vcpu->arch.busy_stolen += mftb() - vcpu->arch.busy_preempt; in kvmppc_core_vcpu_load_hv() 305 vcpu->arch.busy_preempt = TB_NIL; in kvmppc_core_vcpu_load_hv() 307 spin_unlock_irqrestore(&vcpu->arch.tbacct_lock, flags); in kvmppc_core_vcpu_load_hv() [all …]
|
| D | e500_emulate.c | 7 * This file is derived from arch/powerpc/kvm/44x_emulate.c, 56 ulong param = vcpu->arch.regs.gpr[rb]; in kvmppc_e500_emul_msgclr() 62 clear_bit(prio, &vcpu->arch.pending_exceptions); in kvmppc_e500_emul_msgclr() 68 ulong param = vcpu->arch.regs.gpr[rb]; in kvmppc_e500_emul_msgsnd() 78 int cpir = cvcpu->arch.shared->pir; in kvmppc_e500_emul_msgsnd() 80 set_bit(prio, &cvcpu->arch.pending_exceptions); in kvmppc_e500_emul_msgsnd() 97 run->debug.arch.address = vcpu->arch.regs.nip; in kvmppc_e500_emul_ehpriv() 98 run->debug.arch.status = 0; in kvmppc_e500_emul_ehpriv() 229 vcpu->arch.shared->mas0 = spr_val; in kvmppc_core_emulate_mtspr_e500() 232 vcpu->arch.shared->mas1 = spr_val; in kvmppc_core_emulate_mtspr_e500() [all …]
|
| D | book3s_emulate.c | 88 if (vcpu->arch.papr_enabled && (level > PRIV_SUPER)) in spr_allowed() 101 memcpy(&vcpu->arch.gpr_tm[0], &vcpu->arch.regs.gpr[0], in kvmppc_copyto_vcpu_tm() 102 sizeof(vcpu->arch.gpr_tm)); in kvmppc_copyto_vcpu_tm() 103 memcpy(&vcpu->arch.fp_tm, &vcpu->arch.fp, in kvmppc_copyto_vcpu_tm() 105 memcpy(&vcpu->arch.vr_tm, &vcpu->arch.vr, in kvmppc_copyto_vcpu_tm() 107 vcpu->arch.ppr_tm = vcpu->arch.ppr; in kvmppc_copyto_vcpu_tm() 108 vcpu->arch.dscr_tm = vcpu->arch.dscr; in kvmppc_copyto_vcpu_tm() 109 vcpu->arch.amr_tm = vcpu->arch.amr; in kvmppc_copyto_vcpu_tm() 110 vcpu->arch.ctr_tm = vcpu->arch.regs.ctr; in kvmppc_copyto_vcpu_tm() 111 vcpu->arch.tar_tm = vcpu->arch.tar; in kvmppc_copyto_vcpu_tm() [all …]
|
| /kernel/linux/linux-5.10/arch/powerpc/kernel/ |
| D | asm-offsets.c | 455 OFFSET(VCPU_HOST_STACK, kvm_vcpu, arch.host_stack); in main() 456 OFFSET(VCPU_HOST_PID, kvm_vcpu, arch.host_pid); in main() 457 OFFSET(VCPU_GUEST_PID, kvm_vcpu, arch.pid); in main() 458 OFFSET(VCPU_GPRS, kvm_vcpu, arch.regs.gpr); in main() 459 OFFSET(VCPU_VRSAVE, kvm_vcpu, arch.vrsave); in main() 460 OFFSET(VCPU_FPRS, kvm_vcpu, arch.fp.fpr); in main() 462 OFFSET(VCPU_VRS, kvm_vcpu, arch.vr.vr); in main() 464 OFFSET(VCPU_XER, kvm_vcpu, arch.regs.xer); in main() 465 OFFSET(VCPU_CTR, kvm_vcpu, arch.regs.ctr); in main() 466 OFFSET(VCPU_LR, kvm_vcpu, arch.regs.link); in main() [all …]
|
| /kernel/linux/linux-4.19/arch/powerpc/kernel/ |
| D | asm-offsets.c | 426 OFFSET(VCPU_HOST_STACK, kvm_vcpu, arch.host_stack); in main() 427 OFFSET(VCPU_HOST_PID, kvm_vcpu, arch.host_pid); in main() 428 OFFSET(VCPU_GUEST_PID, kvm_vcpu, arch.pid); in main() 429 OFFSET(VCPU_GPRS, kvm_vcpu, arch.regs.gpr); in main() 430 OFFSET(VCPU_VRSAVE, kvm_vcpu, arch.vrsave); in main() 431 OFFSET(VCPU_FPRS, kvm_vcpu, arch.fp.fpr); in main() 433 OFFSET(VCPU_VRS, kvm_vcpu, arch.vr.vr); in main() 435 OFFSET(VCPU_XER, kvm_vcpu, arch.regs.xer); in main() 436 OFFSET(VCPU_CTR, kvm_vcpu, arch.regs.ctr); in main() 437 OFFSET(VCPU_LR, kvm_vcpu, arch.regs.link); in main() [all …]
|
| /kernel/linux/linux-4.19/arch/s390/kvm/ |
| D | guestdbg.c | 62 u64 *cr9 = &vcpu->arch.sie_block->gcr[9]; in enable_all_hw_bp() 63 u64 *cr10 = &vcpu->arch.sie_block->gcr[10]; in enable_all_hw_bp() 64 u64 *cr11 = &vcpu->arch.sie_block->gcr[11]; in enable_all_hw_bp() 67 if (vcpu->arch.guestdbg.nr_hw_bp <= 0 || in enable_all_hw_bp() 68 vcpu->arch.guestdbg.hw_bp_info == NULL) in enable_all_hw_bp() 79 for (i = 0; i < vcpu->arch.guestdbg.nr_hw_bp; i++) { in enable_all_hw_bp() 80 start = vcpu->arch.guestdbg.hw_bp_info[i].addr; in enable_all_hw_bp() 81 len = vcpu->arch.guestdbg.hw_bp_info[i].len; in enable_all_hw_bp() 102 u64 *cr9 = &vcpu->arch.sie_block->gcr[9]; in enable_all_hw_wp() 103 u64 *cr10 = &vcpu->arch.sie_block->gcr[10]; in enable_all_hw_wp() [all …]
|
| /kernel/linux/linux-5.10/arch/s390/kvm/ |
| D | guestdbg.c | 62 u64 *cr9 = &vcpu->arch.sie_block->gcr[9]; in enable_all_hw_bp() 63 u64 *cr10 = &vcpu->arch.sie_block->gcr[10]; in enable_all_hw_bp() 64 u64 *cr11 = &vcpu->arch.sie_block->gcr[11]; in enable_all_hw_bp() 67 if (vcpu->arch.guestdbg.nr_hw_bp <= 0 || in enable_all_hw_bp() 68 vcpu->arch.guestdbg.hw_bp_info == NULL) in enable_all_hw_bp() 79 for (i = 0; i < vcpu->arch.guestdbg.nr_hw_bp; i++) { in enable_all_hw_bp() 80 start = vcpu->arch.guestdbg.hw_bp_info[i].addr; in enable_all_hw_bp() 81 len = vcpu->arch.guestdbg.hw_bp_info[i].len; in enable_all_hw_bp() 102 u64 *cr9 = &vcpu->arch.sie_block->gcr[9]; in enable_all_hw_wp() 103 u64 *cr10 = &vcpu->arch.sie_block->gcr[10]; in enable_all_hw_wp() [all …]
|
| D | kvm-s390.c | 285 kvm_clock_sync_scb(vcpu->arch.sie_block, *delta); in kvm_clock_sync() 287 kvm->arch.epoch = vcpu->arch.sie_block->epoch; in kvm_clock_sync() 288 kvm->arch.epdx = vcpu->arch.sie_block->epdx; in kvm_clock_sync() 290 if (vcpu->arch.cputm_enabled) in kvm_clock_sync() 291 vcpu->arch.cputm_start += *delta; in kvm_clock_sync() 292 if (vcpu->arch.vsie_block) in kvm_clock_sync() 293 kvm_clock_sync_scb(vcpu->arch.vsie_block, in kvm_clock_sync() 597 struct gmap *gmap = kvm->arch.gmap; in kvm_arch_sync_dirty_log() 680 kvm->arch.use_irqchip = 1; in kvm_vm_ioctl_enable_cap() 685 kvm->arch.user_sigp = 1; in kvm_vm_ioctl_enable_cap() [all …]
|
| /kernel/linux/linux-5.10/arch/mips/kvm/ |
| D | emulate.c | 46 struct kvm_vcpu_arch *arch = &vcpu->arch; in kvm_compute_return_epc() local 66 arch->gprs[insn.r_format.rd] = epc + 8; in kvm_compute_return_epc() 69 nextpc = arch->gprs[insn.r_format.rs]; in kvm_compute_return_epc() 85 if ((long)arch->gprs[insn.i_format.rs] < 0) in kvm_compute_return_epc() 94 if ((long)arch->gprs[insn.i_format.rs] >= 0) in kvm_compute_return_epc() 103 arch->gprs[31] = epc + 8; in kvm_compute_return_epc() 104 if ((long)arch->gprs[insn.i_format.rs] < 0) in kvm_compute_return_epc() 113 arch->gprs[31] = epc + 8; in kvm_compute_return_epc() 114 if ((long)arch->gprs[insn.i_format.rs] >= 0) in kvm_compute_return_epc() 142 arch->gprs[31] = instpc + 8; in kvm_compute_return_epc() [all …]
|
| /kernel/linux/linux-5.10/tools/perf/ |
| D | check-headers.sh | 30 arch/x86/include/asm/disabled-features.h 31 arch/x86/include/asm/required-features.h 32 arch/x86/include/asm/cpufeatures.h 33 arch/x86/include/asm/inat_types.h 34 arch/x86/include/asm/emulate_prefix.h 35 arch/x86/include/asm/irq_vectors.h 36 arch/x86/include/asm/msr-index.h 37 arch/x86/include/uapi/asm/prctl.h 38 arch/x86/lib/x86-opcode-map.txt 39 arch/x86/tools/gen-insn-attr-x86.awk [all …]
|
| /kernel/linux/linux-4.19/tools/perf/ |
| D | check-headers.sh | 20 arch/x86/include/asm/disabled-features.h 21 arch/x86/include/asm/required-features.h 22 arch/x86/include/asm/cpufeatures.h 23 arch/arm/include/uapi/asm/perf_regs.h 24 arch/arm64/include/uapi/asm/perf_regs.h 25 arch/powerpc/include/uapi/asm/perf_regs.h 26 arch/s390/include/uapi/asm/perf_regs.h 27 arch/x86/include/uapi/asm/perf_regs.h 28 arch/x86/include/uapi/asm/kvm.h 29 arch/x86/include/uapi/asm/kvm_perf.h [all …]
|
| /kernel/linux/linux-4.19/arch/mips/kvm/ |
| D | emulate.c | 46 struct kvm_vcpu_arch *arch = &vcpu->arch; in kvm_compute_return_epc() local 66 arch->gprs[insn.r_format.rd] = epc + 8; in kvm_compute_return_epc() 69 nextpc = arch->gprs[insn.r_format.rs]; in kvm_compute_return_epc() 85 if ((long)arch->gprs[insn.i_format.rs] < 0) in kvm_compute_return_epc() 94 if ((long)arch->gprs[insn.i_format.rs] >= 0) in kvm_compute_return_epc() 103 arch->gprs[31] = epc + 8; in kvm_compute_return_epc() 104 if ((long)arch->gprs[insn.i_format.rs] < 0) in kvm_compute_return_epc() 113 arch->gprs[31] = epc + 8; in kvm_compute_return_epc() 114 if ((long)arch->gprs[insn.i_format.rs] >= 0) in kvm_compute_return_epc() 142 arch->gprs[31] = instpc + 8; in kvm_compute_return_epc() [all …]
|
| /kernel/linux/linux-4.19/arch/ia64/ |
| D | Makefile | 25 KBUILD_LDFLAGS_MODULE += -T $(srctree)/arch/ia64/module.lds 33 GAS_STATUS = $(shell $(srctree)/arch/ia64/scripts/check-gas "$(CC)" "$(OBJDUMP)") 34 KBUILD_CPPFLAGS += $(shell $(srctree)/arch/ia64/scripts/toolchain-flags "$(CC)" "$(OBJDUMP)" "$(REA… 45 head-y := arch/ia64/kernel/head.o 47 libs-y += arch/ia64/lib/ 48 core-y += arch/ia64/kernel/ arch/ia64/mm/ 49 core-$(CONFIG_IA64_DIG) += arch/ia64/dig/ 50 core-$(CONFIG_IA64_DIG_VTD) += arch/ia64/dig/ 51 core-$(CONFIG_IA64_GENERIC) += arch/ia64/dig/ 52 core-$(CONFIG_IA64_HP_ZX1) += arch/ia64/dig/ [all …]
|