Home
last modified time | relevance | path

Searched refs:cpu_if (Results 1 – 6 of 6) sorted by relevance

/arch/arm64/kvm/hyp/
Dvgic-v3-sr.c199 void __vgic_v3_save_state(struct vgic_v3_cpu_if *cpu_if) in __vgic_v3_save_state() argument
201 u64 used_lrs = cpu_if->used_lrs; in __vgic_v3_save_state()
209 if (!cpu_if->vgic_sre) { in __vgic_v3_save_state()
215 if (used_lrs || cpu_if->its_vpe.its_vm) { in __vgic_v3_save_state()
221 write_gicreg(cpu_if->vgic_hcr & ~ICH_HCR_EN, ICH_HCR_EL2); in __vgic_v3_save_state()
225 cpu_if->vgic_lr[i] &= ~ICH_LR_STATE; in __vgic_v3_save_state()
227 cpu_if->vgic_lr[i] = __gic_v3_get_lr(i); in __vgic_v3_save_state()
234 void __vgic_v3_restore_state(struct vgic_v3_cpu_if *cpu_if) in __vgic_v3_restore_state() argument
236 u64 used_lrs = cpu_if->used_lrs; in __vgic_v3_restore_state()
239 if (used_lrs || cpu_if->its_vpe.its_vm) { in __vgic_v3_restore_state()
[all …]
/arch/arm64/kvm/vgic/
Dvgic-v2.c207 struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; in vgic_v2_set_vmcr() local
229 cpu_if->vgic_vmcr = vmcr; in vgic_v2_set_vmcr()
234 struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; in vgic_v2_get_vmcr() local
237 vmcr = cpu_if->vgic_vmcr; in vgic_v2_get_vmcr()
411 struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; in save_lrs() local
412 u64 used_lrs = cpu_if->used_lrs; in save_lrs()
422 cpu_if->vgic_lr[i] &= ~GICH_LR_STATE; in save_lrs()
424 cpu_if->vgic_lr[i] = readl_relaxed(base + GICH_LR0 + (i * 4)); in save_lrs()
446 struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; in vgic_v2_restore_state() local
448 u64 used_lrs = cpu_if->used_lrs; in vgic_v2_restore_state()
[all …]
Dvgic-v3.c195 struct vgic_v3_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v3; in vgic_v3_set_vmcr() local
220 cpu_if->vgic_vmcr = vmcr; in vgic_v3_set_vmcr()
225 struct vgic_v3_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v3; in vgic_v3_get_vmcr() local
229 vmcr = cpu_if->vgic_vmcr; in vgic_v3_get_vmcr()
713 struct vgic_v3_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v3; in vgic_v3_load() local
716 kvm_call_hyp(__vgic_v3_restore_vmcr_aprs, cpu_if); in vgic_v3_load()
719 __vgic_v3_activate_traps(cpu_if); in vgic_v3_load()
726 struct vgic_v3_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v3; in vgic_v3_put() local
731 kvm_call_hyp(__vgic_v3_save_vmcr_aprs, cpu_if); in vgic_v3_put()
734 __vgic_v3_deactivate_traps(cpu_if); in vgic_v3_put()
Dvgic.h262 struct vgic_cpu *cpu_if = &vcpu->arch.vgic_cpu; in vgic_v3_max_apr_idx() local
269 switch (cpu_if->num_pri_bits) { in vgic_v3_max_apr_idx()
/arch/arm64/include/asm/
Dkvm_hyp.h60 void __vgic_v3_save_state(struct vgic_v3_cpu_if *cpu_if);
61 void __vgic_v3_restore_state(struct vgic_v3_cpu_if *cpu_if);
62 void __vgic_v3_activate_traps(struct vgic_v3_cpu_if *cpu_if);
63 void __vgic_v3_deactivate_traps(struct vgic_v3_cpu_if *cpu_if);
64 void __vgic_v3_save_vmcr_aprs(struct vgic_v3_cpu_if *cpu_if);
65 void __vgic_v3_restore_vmcr_aprs(struct vgic_v3_cpu_if *cpu_if);
/arch/arm64/kvm/hyp/nvhe/
Dhyp-main.c889 struct vgic_v3_cpu_if *shadow_cpu_if, *cpu_if; in handle___vgic_v3_save_vmcr_aprs() local
895 cpu_if = &vcpu->arch.vgic_cpu.vgic_v3; in handle___vgic_v3_save_vmcr_aprs()
897 cpu_if->vgic_vmcr = shadow_cpu_if->vgic_vmcr; in handle___vgic_v3_save_vmcr_aprs()
898 for (i = 0; i < ARRAY_SIZE(cpu_if->vgic_ap0r); i++) { in handle___vgic_v3_save_vmcr_aprs()
899 cpu_if->vgic_ap0r[i] = shadow_cpu_if->vgic_ap0r[i]; in handle___vgic_v3_save_vmcr_aprs()
900 cpu_if->vgic_ap1r[i] = shadow_cpu_if->vgic_ap1r[i]; in handle___vgic_v3_save_vmcr_aprs()
917 struct vgic_v3_cpu_if *shadow_cpu_if, *cpu_if; in handle___vgic_v3_restore_vmcr_aprs() local
921 cpu_if = &vcpu->arch.vgic_cpu.vgic_v3; in handle___vgic_v3_restore_vmcr_aprs()
923 shadow_cpu_if->vgic_vmcr = cpu_if->vgic_vmcr; in handle___vgic_v3_restore_vmcr_aprs()
928 for (i = 0; i < ARRAY_SIZE(cpu_if->vgic_ap0r); i++) { in handle___vgic_v3_restore_vmcr_aprs()
[all …]