Home
last modified time | relevance | path

Searched full:viommu (Results 1 – 11 of 11) sorted by relevance

/kernel/linux/linux-5.10/drivers/iommu/
Dvirtio-iommu.c65 struct viommu_dev *viommu; member
66 struct mutex mutex; /* protects viommu pointer */
78 struct viommu_dev *viommu; member
136 static off_t viommu_get_write_desc_offset(struct viommu_dev *viommu, in viommu_get_write_desc_offset() argument
143 return len - viommu->probe_size - tail_size; in viommu_get_write_desc_offset()
154 static int __viommu_sync_req(struct viommu_dev *viommu) in __viommu_sync_req() argument
159 struct virtqueue *vq = viommu->vqs[VIOMMU_REQUEST_VQ]; in __viommu_sync_req()
161 assert_spin_locked(&viommu->request_lock); in __viommu_sync_req()
165 while (!list_empty(&viommu->requests)) { in __viommu_sync_req()
187 static int viommu_sync_req(struct viommu_dev *viommu) in viommu_sync_req() argument
[all …]
/kernel/linux/linux-6.6/drivers/iommu/
Dvirtio-iommu.c64 struct viommu_dev *viommu; member
65 struct mutex mutex; /* protects viommu pointer */
78 struct viommu_dev *viommu; member
136 static off_t viommu_get_write_desc_offset(struct viommu_dev *viommu, in viommu_get_write_desc_offset() argument
143 return len - viommu->probe_size - tail_size; in viommu_get_write_desc_offset()
154 static int __viommu_sync_req(struct viommu_dev *viommu) in __viommu_sync_req() argument
159 struct virtqueue *vq = viommu->vqs[VIOMMU_REQUEST_VQ]; in __viommu_sync_req()
161 assert_spin_locked(&viommu->request_lock); in __viommu_sync_req()
165 while (!list_empty(&viommu->requests)) { in __viommu_sync_req()
187 static int viommu_sync_req(struct viommu_dev *viommu) in viommu_sync_req() argument
[all …]
/kernel/linux/linux-6.6/drivers/acpi/
Dviot.c48 struct viot_iommu *viommu; member
77 static int __init viot_get_pci_iommu_fwnode(struct viot_iommu *viommu, in viot_get_pci_iommu_fwnode() argument
103 viommu->fwnode = dev_fwnode(&pdev->dev); in viot_get_pci_iommu_fwnode()
108 static int __init viot_get_mmio_iommu_fwnode(struct viot_iommu *viommu, in viot_get_mmio_iommu_fwnode() argument
123 viommu->fwnode = &adev->fwnode; in viot_get_mmio_iommu_fwnode()
130 struct viot_iommu *viommu; in viot_get_iommu() local
138 list_for_each_entry(viommu, &viot_iommus, list) in viot_get_iommu()
139 if (viommu->offset == offset) in viot_get_iommu()
140 return viommu; in viot_get_iommu()
145 viommu = kzalloc(sizeof(*viommu), GFP_KERNEL); in viot_get_iommu()
[all …]
/kernel/linux/linux-6.6/Documentation/devicetree/bindings/virtio/
Dmmio.yaml56 iommus = <&viommu 23>;
59 viommu: iommu@3100 {
/kernel/linux/linux-5.10/Documentation/devicetree/bindings/virtio/
Dmmio.txt35 iommus = <&viommu 23>
38 viommu: iommu@3100 {
/kernel/linux/linux-6.6/Documentation/userspace-api/
Diommu.rst14 guest IO virtual address (IOVA), wherein the vIOMMU implementation
84 While launching a guest with vIOMMU, it is strongly advised to check
86 vIOMMU operation, such as cache invalidation failures cannot be nicely
/kernel/linux/linux-5.10/Documentation/userspace-api/
Diommu.rst14 guest IO virtual address (IOVA), wherein the vIOMMU implementation
84 While launching a guest with vIOMMU, it is strongly advised to check
86 vIOMMU operation, such as cache invalidation failures cannot be nicely
/kernel/linux/linux-5.10/include/uapi/linux/
Diommu.h281 * Only guest vIOMMU selectable and effective options are passed down to
/kernel/linux/linux-6.6/drivers/hv/
Dhv_common.c489 * Hyper-V does not offer a vIOMMU in the guest in hv_setup_dma_ops()
/kernel/linux/linux-5.10/drivers/iommu/intel/
Diommu.c3301 * is active. All vIOMMU allocators will eventually be calling the same in register_pasid_allocator()
3315 * is no custom allocator. Mixing SM capable vIOMMU in register_pasid_allocator()
3316 * and non-SM vIOMMU are not supported. in register_pasid_allocator()
5489 * vIOMMU in the guest will only expose first level page tables, therefore
5617 * Always flush device IOTLB if ATS is enabled. vIOMMU in intel_iommu_sva_invalidate()
/kernel/linux/linux-6.6/drivers/iommu/amd/
Diommu.c1422 * When NpCache is on, we infer that we run in a VM and use a vIOMMU. in domain_flush_pages()