Lines Matching full:vpes
166 struct its_vpe **vpes; member
1237 * If the VM wasn't mapped yet, iterate over the vpes and get in its_map_vm()
1246 struct its_vpe *vpe = vm->vpes[i]; in its_map_vm()
1274 its_send_vmapp(its, vm->vpes[i], false); in its_unmap_vm()
1315 /* Ensure all the VPEs are mapped on this ITS */ in its_vlpi_map()
2596 vpe_proxy.vpes[vpe->vpe_proxy_event] = NULL; in its_vpe_db_proxy_unmap_locked()
2603 * effect... Let's just hope VPEs don't migrate too often. in its_vpe_db_proxy_unmap_locked()
2605 if (vpe_proxy.vpes[vpe_proxy.next_victim]) in its_vpe_db_proxy_unmap_locked()
2629 if (vpe_proxy.vpes[vpe_proxy.next_victim]) in its_vpe_db_proxy_map_locked()
2630 its_vpe_db_proxy_unmap_locked(vpe_proxy.vpes[vpe_proxy.next_victim]); in its_vpe_db_proxy_map_locked()
2633 vpe_proxy.vpes[vpe_proxy.next_victim] = vpe; in its_vpe_db_proxy_map_locked()
2985 vm->vpes[i]->vpe_db_lpi = base + i; in its_vpe_irq_domain_alloc()
2986 err = its_vpe_init(vm->vpes[i]); in its_vpe_irq_domain_alloc()
2990 vm->vpes[i]->vpe_db_lpi); in its_vpe_irq_domain_alloc()
2994 &its_vpe_irq_chip, vm->vpes[i]); in its_vpe_irq_domain_alloc()
3381 vpe_proxy.vpes = kcalloc(entries, sizeof(*vpe_proxy.vpes), in its_init_vpe_domain()
3383 if (!vpe_proxy.vpes) { in its_init_vpe_domain()
3392 kfree(vpe_proxy.vpes); in its_init_vpe_domain()