Lines Matching refs:kvm
28 struct vgic_dist *vgic = &vcpu->kvm->arch.vgic; in vgic_mmio_read_v2_misc()
38 value |= (atomic_read(&vcpu->kvm->online_vcpus) - 1) << 5; in vgic_mmio_read_v2_misc()
56 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_mmio_write_v2_misc()
63 vgic_kick_vcpus(vcpu->kvm); in vgic_mmio_write_v2_misc()
90 vcpu->kvm->arch.vgic.v2_groups_user_writable = true; in vgic_mmio_uaccess_write_v2_misc()
102 if (vcpu->kvm->arch.vgic.v2_groups_user_writable) in vgic_mmio_uaccess_write_v2_group()
112 int nr_vcpus = atomic_read(&source_vcpu->kvm->online_vcpus); in vgic_mmio_write_sgir()
134 kvm_for_each_vcpu(c, vcpu, source_vcpu->kvm) { in vgic_mmio_write_sgir()
140 irq = vgic_get_irq(source_vcpu->kvm, vcpu, intid); in vgic_mmio_write_sgir()
146 vgic_queue_irq_unlock(source_vcpu->kvm, irq, flags); in vgic_mmio_write_sgir()
147 vgic_put_irq(source_vcpu->kvm, irq); in vgic_mmio_write_sgir()
159 struct vgic_irq *irq = vgic_get_irq(vcpu->kvm, vcpu, intid + i); in vgic_mmio_read_target()
163 vgic_put_irq(vcpu->kvm, irq); in vgic_mmio_read_target()
174 u8 cpu_mask = GENMASK(atomic_read(&vcpu->kvm->online_vcpus) - 1, 0); in vgic_mmio_write_target()
183 struct vgic_irq *irq = vgic_get_irq(vcpu->kvm, NULL, intid + i); in vgic_mmio_write_target()
190 irq->target_vcpu = kvm_get_vcpu(vcpu->kvm, target); in vgic_mmio_write_target()
193 vgic_put_irq(vcpu->kvm, irq); in vgic_mmio_write_target()
205 struct vgic_irq *irq = vgic_get_irq(vcpu->kvm, vcpu, intid + i); in vgic_mmio_read_sgipend()
209 vgic_put_irq(vcpu->kvm, irq); in vgic_mmio_read_sgipend()
223 struct vgic_irq *irq = vgic_get_irq(vcpu->kvm, vcpu, intid + i); in vgic_mmio_write_sgipendc()
232 vgic_put_irq(vcpu->kvm, irq); in vgic_mmio_write_sgipendc()
245 struct vgic_irq *irq = vgic_get_irq(vcpu->kvm, vcpu, intid + i); in vgic_mmio_write_sgipends()
253 vgic_queue_irq_unlock(vcpu->kvm, irq, flags); in vgic_mmio_write_sgipends()
257 vgic_put_irq(vcpu->kvm, irq); in vgic_mmio_write_sgipends()