Home
last modified time | relevance | path

Searched refs:cpu_if (Results 1 – 5 of 5) sorted by relevance

/arch/arm64/kvm/hyp/
Dvgic-v3-sr.c199 void __vgic_v3_save_state(struct vgic_v3_cpu_if *cpu_if) in __vgic_v3_save_state() argument
201 u64 used_lrs = cpu_if->used_lrs; in __vgic_v3_save_state()
209 if (!cpu_if->vgic_sre) { in __vgic_v3_save_state()
215 if (used_lrs || cpu_if->its_vpe.its_vm) { in __vgic_v3_save_state()
221 write_gicreg(cpu_if->vgic_hcr & ~ICH_HCR_EN, ICH_HCR_EL2); in __vgic_v3_save_state()
225 cpu_if->vgic_lr[i] &= ~ICH_LR_STATE; in __vgic_v3_save_state()
227 cpu_if->vgic_lr[i] = __gic_v3_get_lr(i); in __vgic_v3_save_state()
234 void __vgic_v3_restore_state(struct vgic_v3_cpu_if *cpu_if) in __vgic_v3_restore_state() argument
236 u64 used_lrs = cpu_if->used_lrs; in __vgic_v3_restore_state()
239 if (used_lrs || cpu_if->its_vpe.its_vm) { in __vgic_v3_restore_state()
[all …]
/arch/arm64/kvm/vgic/
Dvgic-v2.c207 struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; in vgic_v2_set_vmcr() local
229 cpu_if->vgic_vmcr = vmcr; in vgic_v2_set_vmcr()
234 struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; in vgic_v2_get_vmcr() local
237 vmcr = cpu_if->vgic_vmcr; in vgic_v2_get_vmcr()
405 struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; in save_lrs() local
406 u64 used_lrs = cpu_if->used_lrs; in save_lrs()
416 cpu_if->vgic_lr[i] &= ~GICH_LR_STATE; in save_lrs()
418 cpu_if->vgic_lr[i] = readl_relaxed(base + GICH_LR0 + (i * 4)); in save_lrs()
440 struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; in vgic_v2_restore_state() local
442 u64 used_lrs = cpu_if->used_lrs; in vgic_v2_restore_state()
[all …]
Dvgic-v3.c196 struct vgic_v3_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v3; in vgic_v3_set_vmcr() local
221 cpu_if->vgic_vmcr = vmcr; in vgic_v3_set_vmcr()
226 struct vgic_v3_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v3; in vgic_v3_get_vmcr() local
230 vmcr = cpu_if->vgic_vmcr; in vgic_v3_get_vmcr()
722 struct vgic_v3_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v3; in vgic_v3_load() local
725 kvm_call_hyp(__vgic_v3_restore_vmcr_aprs, cpu_if); in vgic_v3_load()
728 __vgic_v3_activate_traps(cpu_if); in vgic_v3_load()
735 struct vgic_v3_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v3; in vgic_v3_put() local
738 kvm_call_hyp(__vgic_v3_save_vmcr_aprs, cpu_if); in vgic_v3_put()
742 __vgic_v3_deactivate_traps(cpu_if); in vgic_v3_put()
Dvgic.h275 struct vgic_cpu *cpu_if = &vcpu->arch.vgic_cpu; in vgic_v3_max_apr_idx() local
282 switch (cpu_if->num_pri_bits) { in vgic_v3_max_apr_idx()
/arch/arm64/include/asm/
Dkvm_hyp.h82 void __vgic_v3_save_state(struct vgic_v3_cpu_if *cpu_if);
83 void __vgic_v3_restore_state(struct vgic_v3_cpu_if *cpu_if);
84 void __vgic_v3_activate_traps(struct vgic_v3_cpu_if *cpu_if);
85 void __vgic_v3_deactivate_traps(struct vgic_v3_cpu_if *cpu_if);
86 void __vgic_v3_save_vmcr_aprs(struct vgic_v3_cpu_if *cpu_if);
87 void __vgic_v3_restore_vmcr_aprs(struct vgic_v3_cpu_if *cpu_if);