Lines Matching full:iommu
14 #include <linux/iommu.h>
99 struct iommu_device iommu; member
101 /* Lock to modify the IOMMU registers */
124 struct sun50i_iommu *iommu; member
137 static u32 iommu_read(struct sun50i_iommu *iommu, u32 offset) in iommu_read() argument
139 return readl(iommu->base + offset); in iommu_read()
142 static void iommu_write(struct sun50i_iommu *iommu, u32 offset, u32 value) in iommu_write() argument
144 writel(value, iommu->base + offset); in iommu_write()
148 * The Allwinner H6 IOMMU uses a 2-level page table.
157 * The IOMMU supports a single DT, pointed by the IOMMU_TTB_REG
224 * The way permissions work is that the IOMMU has 16 "domains" that
235 * In order to make it work with the IOMMU framework, we will be using
238 * have each master setup in the same way, since the IOMMU framework
293 struct sun50i_iommu *iommu = sun50i_domain->iommu; in sun50i_table_flush() local
297 dma_sync_single_for_device(iommu->dev, dma, size, DMA_TO_DEVICE); in sun50i_table_flush()
300 static void sun50i_iommu_zap_iova(struct sun50i_iommu *iommu, in sun50i_iommu_zap_iova() argument
306 iommu_write(iommu, IOMMU_TLB_IVLD_ADDR_REG, iova); in sun50i_iommu_zap_iova()
307 iommu_write(iommu, IOMMU_TLB_IVLD_ADDR_MASK_REG, GENMASK(31, 12)); in sun50i_iommu_zap_iova()
308 iommu_write(iommu, IOMMU_TLB_IVLD_ENABLE_REG, in sun50i_iommu_zap_iova()
311 ret = readl_poll_timeout_atomic(iommu->base + IOMMU_TLB_IVLD_ENABLE_REG, in sun50i_iommu_zap_iova()
314 dev_warn(iommu->dev, "TLB invalidation timed out!\n"); in sun50i_iommu_zap_iova()
317 static void sun50i_iommu_zap_ptw_cache(struct sun50i_iommu *iommu, in sun50i_iommu_zap_ptw_cache() argument
323 iommu_write(iommu, IOMMU_PC_IVLD_ADDR_REG, iova); in sun50i_iommu_zap_ptw_cache()
324 iommu_write(iommu, IOMMU_PC_IVLD_ENABLE_REG, in sun50i_iommu_zap_ptw_cache()
327 ret = readl_poll_timeout_atomic(iommu->base + IOMMU_PC_IVLD_ENABLE_REG, in sun50i_iommu_zap_ptw_cache()
330 dev_warn(iommu->dev, "PTW cache invalidation timed out!\n"); in sun50i_iommu_zap_ptw_cache()
333 static void sun50i_iommu_zap_range(struct sun50i_iommu *iommu, in sun50i_iommu_zap_range() argument
336 assert_spin_locked(&iommu->iommu_lock); in sun50i_iommu_zap_range()
338 iommu_write(iommu, IOMMU_AUTO_GATING_REG, 0); in sun50i_iommu_zap_range()
340 sun50i_iommu_zap_iova(iommu, iova); in sun50i_iommu_zap_range()
341 sun50i_iommu_zap_iova(iommu, iova + SPAGE_SIZE); in sun50i_iommu_zap_range()
343 sun50i_iommu_zap_iova(iommu, iova + size); in sun50i_iommu_zap_range()
344 sun50i_iommu_zap_iova(iommu, iova + size + SPAGE_SIZE); in sun50i_iommu_zap_range()
346 sun50i_iommu_zap_ptw_cache(iommu, iova); in sun50i_iommu_zap_range()
347 sun50i_iommu_zap_ptw_cache(iommu, iova + SZ_1M); in sun50i_iommu_zap_range()
349 sun50i_iommu_zap_ptw_cache(iommu, iova + size); in sun50i_iommu_zap_range()
350 sun50i_iommu_zap_ptw_cache(iommu, iova + size + SZ_1M); in sun50i_iommu_zap_range()
353 iommu_write(iommu, IOMMU_AUTO_GATING_REG, IOMMU_AUTO_GATING_ENABLE); in sun50i_iommu_zap_range()
356 static int sun50i_iommu_flush_all_tlb(struct sun50i_iommu *iommu) in sun50i_iommu_flush_all_tlb() argument
361 assert_spin_locked(&iommu->iommu_lock); in sun50i_iommu_flush_all_tlb()
363 iommu_write(iommu, in sun50i_iommu_flush_all_tlb()
374 ret = readl_poll_timeout_atomic(iommu->base + IOMMU_TLB_FLUSH_REG, in sun50i_iommu_flush_all_tlb()
378 dev_warn(iommu->dev, "TLB Flush timed out!\n"); in sun50i_iommu_flush_all_tlb()
386 struct sun50i_iommu *iommu = sun50i_domain->iommu; in sun50i_iommu_flush_iotlb_all() local
391 * .probe_device, and since we link our (single) domain to our iommu in in sun50i_iommu_flush_iotlb_all()
397 if (!iommu) in sun50i_iommu_flush_iotlb_all()
400 spin_lock_irqsave(&iommu->iommu_lock, flags); in sun50i_iommu_flush_iotlb_all()
401 sun50i_iommu_flush_all_tlb(iommu); in sun50i_iommu_flush_iotlb_all()
402 spin_unlock_irqrestore(&iommu->iommu_lock, flags); in sun50i_iommu_flush_iotlb_all()
409 struct sun50i_iommu *iommu = sun50i_domain->iommu; in sun50i_iommu_iotlb_sync_map() local
412 spin_lock_irqsave(&iommu->iommu_lock, flags); in sun50i_iommu_iotlb_sync_map()
413 sun50i_iommu_zap_range(iommu, iova, size); in sun50i_iommu_iotlb_sync_map()
414 spin_unlock_irqrestore(&iommu->iommu_lock, flags); in sun50i_iommu_iotlb_sync_map()
423 static int sun50i_iommu_enable(struct sun50i_iommu *iommu) in sun50i_iommu_enable() argument
429 if (!iommu->domain) in sun50i_iommu_enable()
432 sun50i_domain = to_sun50i_domain(iommu->domain); in sun50i_iommu_enable()
434 ret = reset_control_deassert(iommu->reset); in sun50i_iommu_enable()
438 ret = clk_prepare_enable(iommu->clk); in sun50i_iommu_enable()
442 spin_lock_irqsave(&iommu->iommu_lock, flags); in sun50i_iommu_enable()
444 iommu_write(iommu, IOMMU_TTB_REG, sun50i_domain->dt_dma); in sun50i_iommu_enable()
445 iommu_write(iommu, IOMMU_TLB_PREFETCH_REG, in sun50i_iommu_enable()
452 iommu_write(iommu, IOMMU_INT_ENABLE_REG, IOMMU_INT_MASK); in sun50i_iommu_enable()
453 iommu_write(iommu, IOMMU_DM_AUT_CTRL_REG(SUN50I_IOMMU_ACI_NONE), in sun50i_iommu_enable()
467 iommu_write(iommu, IOMMU_DM_AUT_CTRL_REG(SUN50I_IOMMU_ACI_RD), in sun50i_iommu_enable()
475 iommu_write(iommu, IOMMU_DM_AUT_CTRL_REG(SUN50I_IOMMU_ACI_WR), in sun50i_iommu_enable()
483 ret = sun50i_iommu_flush_all_tlb(iommu); in sun50i_iommu_enable()
485 spin_unlock_irqrestore(&iommu->iommu_lock, flags); in sun50i_iommu_enable()
489 iommu_write(iommu, IOMMU_AUTO_GATING_REG, IOMMU_AUTO_GATING_ENABLE); in sun50i_iommu_enable()
490 iommu_write(iommu, IOMMU_ENABLE_REG, IOMMU_ENABLE_ENABLE); in sun50i_iommu_enable()
492 spin_unlock_irqrestore(&iommu->iommu_lock, flags); in sun50i_iommu_enable()
497 clk_disable_unprepare(iommu->clk); in sun50i_iommu_enable()
500 reset_control_assert(iommu->reset); in sun50i_iommu_enable()
505 static void sun50i_iommu_disable(struct sun50i_iommu *iommu) in sun50i_iommu_disable() argument
509 spin_lock_irqsave(&iommu->iommu_lock, flags); in sun50i_iommu_disable()
511 iommu_write(iommu, IOMMU_ENABLE_REG, 0); in sun50i_iommu_disable()
512 iommu_write(iommu, IOMMU_TTB_REG, 0); in sun50i_iommu_disable()
514 spin_unlock_irqrestore(&iommu->iommu_lock, flags); in sun50i_iommu_disable()
516 clk_disable_unprepare(iommu->clk); in sun50i_iommu_disable()
517 reset_control_assert(iommu->reset); in sun50i_iommu_disable()
520 static void *sun50i_iommu_alloc_page_table(struct sun50i_iommu *iommu, in sun50i_iommu_alloc_page_table() argument
526 page_table = kmem_cache_zalloc(iommu->pt_pool, gfp); in sun50i_iommu_alloc_page_table()
530 pt_dma = dma_map_single(iommu->dev, page_table, PT_SIZE, DMA_TO_DEVICE); in sun50i_iommu_alloc_page_table()
531 if (dma_mapping_error(iommu->dev, pt_dma)) { in sun50i_iommu_alloc_page_table()
532 dev_err(iommu->dev, "Couldn't map L2 Page Table\n"); in sun50i_iommu_alloc_page_table()
533 kmem_cache_free(iommu->pt_pool, page_table); in sun50i_iommu_alloc_page_table()
543 static void sun50i_iommu_free_page_table(struct sun50i_iommu *iommu, in sun50i_iommu_free_page_table() argument
548 dma_unmap_single(iommu->dev, pt_phys, PT_SIZE, DMA_TO_DEVICE); in sun50i_iommu_free_page_table()
549 kmem_cache_free(iommu->pt_pool, page_table); in sun50i_iommu_free_page_table()
555 struct sun50i_iommu *iommu = sun50i_domain->iommu; in sun50i_dte_get_page_table() local
568 page_table = sun50i_iommu_alloc_page_table(iommu, gfp); in sun50i_dte_get_page_table()
582 sun50i_iommu_free_page_table(iommu, drop_pt); in sun50i_dte_get_page_table()
595 struct sun50i_iommu *iommu = sun50i_domain->iommu; in sun50i_iommu_map() local
610 dev_err(iommu->dev, in sun50i_iommu_map()
711 static int sun50i_iommu_attach_domain(struct sun50i_iommu *iommu, in sun50i_iommu_attach_domain() argument
714 iommu->domain = &sun50i_domain->domain; in sun50i_iommu_attach_domain()
715 sun50i_domain->iommu = iommu; in sun50i_iommu_attach_domain()
717 sun50i_domain->dt_dma = dma_map_single(iommu->dev, sun50i_domain->dt, in sun50i_iommu_attach_domain()
719 if (dma_mapping_error(iommu->dev, sun50i_domain->dt_dma)) { in sun50i_iommu_attach_domain()
720 dev_err(iommu->dev, "Couldn't map L1 Page Table\n"); in sun50i_iommu_attach_domain()
724 return sun50i_iommu_enable(iommu); in sun50i_iommu_attach_domain()
727 static void sun50i_iommu_detach_domain(struct sun50i_iommu *iommu, in sun50i_iommu_detach_domain() argument
748 sun50i_iommu_free_page_table(iommu, page_table); in sun50i_iommu_detach_domain()
752 sun50i_iommu_disable(iommu); in sun50i_iommu_detach_domain()
754 dma_unmap_single(iommu->dev, virt_to_phys(sun50i_domain->dt), in sun50i_iommu_detach_domain()
757 iommu->domain = NULL; in sun50i_iommu_detach_domain()
764 struct sun50i_iommu *iommu = dev_iommu_priv_get(dev); in sun50i_iommu_detach_device() local
766 dev_dbg(dev, "Detaching from IOMMU domain\n"); in sun50i_iommu_detach_device()
768 if (iommu->domain != domain) in sun50i_iommu_detach_device()
772 sun50i_iommu_detach_domain(iommu, sun50i_domain); in sun50i_iommu_detach_device()
779 struct sun50i_iommu *iommu; in sun50i_iommu_attach_device() local
781 iommu = sun50i_iommu_from_dev(dev); in sun50i_iommu_attach_device()
782 if (!iommu) in sun50i_iommu_attach_device()
785 dev_dbg(dev, "Attaching to IOMMU domain\n"); in sun50i_iommu_attach_device()
789 if (iommu->domain == domain) in sun50i_iommu_attach_device()
792 if (iommu->domain) in sun50i_iommu_attach_device()
793 sun50i_iommu_detach_device(iommu->domain, dev); in sun50i_iommu_attach_device()
795 sun50i_iommu_attach_domain(iommu, sun50i_domain); in sun50i_iommu_attach_device()
802 struct sun50i_iommu *iommu; in sun50i_iommu_probe_device() local
804 iommu = sun50i_iommu_from_dev(dev); in sun50i_iommu_probe_device()
805 if (!iommu) in sun50i_iommu_probe_device()
808 return &iommu->iommu; in sun50i_iommu_probe_device()
813 struct sun50i_iommu *iommu = sun50i_iommu_from_dev(dev); in sun50i_iommu_device_group() local
815 return iommu_group_ref_get(iommu->group); in sun50i_iommu_device_group()
847 static void sun50i_iommu_report_fault(struct sun50i_iommu *iommu, in sun50i_iommu_report_fault() argument
851 dev_err(iommu->dev, "Page fault for %pad (master %d, dir %s)\n", in sun50i_iommu_report_fault()
854 if (iommu->domain) in sun50i_iommu_report_fault()
855 report_iommu_fault(iommu->domain, iommu->dev, iova, prot); in sun50i_iommu_report_fault()
857 dev_err(iommu->dev, "Page fault while iommu not attached to any domain?\n"); in sun50i_iommu_report_fault()
859 sun50i_iommu_zap_range(iommu, iova, SPAGE_SIZE); in sun50i_iommu_report_fault()
862 static phys_addr_t sun50i_iommu_handle_pt_irq(struct sun50i_iommu *iommu, in sun50i_iommu_handle_pt_irq() argument
870 assert_spin_locked(&iommu->iommu_lock); in sun50i_iommu_handle_pt_irq()
872 iova = iommu_read(iommu, addr_reg); in sun50i_iommu_handle_pt_irq()
873 blame = iommu_read(iommu, blame_reg); in sun50i_iommu_handle_pt_irq()
881 sun50i_iommu_report_fault(iommu, master, iova, IOMMU_FAULT_READ); in sun50i_iommu_handle_pt_irq()
886 static phys_addr_t sun50i_iommu_handle_perm_irq(struct sun50i_iommu *iommu) in sun50i_iommu_handle_perm_irq() argument
894 assert_spin_locked(&iommu->iommu_lock); in sun50i_iommu_handle_perm_irq()
896 blame = iommu_read(iommu, IOMMU_INT_STA_REG); in sun50i_iommu_handle_perm_irq()
898 iova = iommu_read(iommu, IOMMU_INT_ERR_ADDR_REG(master)); in sun50i_iommu_handle_perm_irq()
899 aci = sun50i_get_pte_aci(iommu_read(iommu, in sun50i_iommu_handle_perm_irq()
936 sun50i_iommu_report_fault(iommu, master, iova, dir); in sun50i_iommu_handle_perm_irq()
944 struct sun50i_iommu *iommu = dev_id; in sun50i_iommu_irq() local
946 spin_lock(&iommu->iommu_lock); in sun50i_iommu_irq()
948 status = iommu_read(iommu, IOMMU_INT_STA_REG); in sun50i_iommu_irq()
950 spin_unlock(&iommu->iommu_lock); in sun50i_iommu_irq()
954 l1_status = iommu_read(iommu, IOMMU_L1PG_INT_REG); in sun50i_iommu_irq()
955 l2_status = iommu_read(iommu, IOMMU_L2PG_INT_REG); in sun50i_iommu_irq()
958 sun50i_iommu_handle_pt_irq(iommu, in sun50i_iommu_irq()
962 sun50i_iommu_handle_pt_irq(iommu, in sun50i_iommu_irq()
966 sun50i_iommu_handle_perm_irq(iommu); in sun50i_iommu_irq()
968 iommu_write(iommu, IOMMU_INT_CLR_REG, status); in sun50i_iommu_irq()
971 iommu_write(iommu, IOMMU_RESET_REG, ~resets); in sun50i_iommu_irq()
972 iommu_write(iommu, IOMMU_RESET_REG, IOMMU_RESET_RELEASE_ALL); in sun50i_iommu_irq()
974 spin_unlock(&iommu->iommu_lock); in sun50i_iommu_irq()
981 struct sun50i_iommu *iommu; in sun50i_iommu_probe() local
984 iommu = devm_kzalloc(&pdev->dev, sizeof(*iommu), GFP_KERNEL); in sun50i_iommu_probe()
985 if (!iommu) in sun50i_iommu_probe()
987 spin_lock_init(&iommu->iommu_lock); in sun50i_iommu_probe()
988 platform_set_drvdata(pdev, iommu); in sun50i_iommu_probe()
989 iommu->dev = &pdev->dev; in sun50i_iommu_probe()
991 iommu->pt_pool = kmem_cache_create(dev_name(&pdev->dev), in sun50i_iommu_probe()
995 if (!iommu->pt_pool) in sun50i_iommu_probe()
998 iommu->group = iommu_group_alloc(); in sun50i_iommu_probe()
999 if (IS_ERR(iommu->group)) { in sun50i_iommu_probe()
1000 ret = PTR_ERR(iommu->group); in sun50i_iommu_probe()
1004 iommu->base = devm_platform_ioremap_resource(pdev, 0); in sun50i_iommu_probe()
1005 if (IS_ERR(iommu->base)) { in sun50i_iommu_probe()
1006 ret = PTR_ERR(iommu->base); in sun50i_iommu_probe()
1016 iommu->clk = devm_clk_get(&pdev->dev, NULL); in sun50i_iommu_probe()
1017 if (IS_ERR(iommu->clk)) { in sun50i_iommu_probe()
1019 ret = PTR_ERR(iommu->clk); in sun50i_iommu_probe()
1023 iommu->reset = devm_reset_control_get(&pdev->dev, NULL); in sun50i_iommu_probe()
1024 if (IS_ERR(iommu->reset)) { in sun50i_iommu_probe()
1026 ret = PTR_ERR(iommu->reset); in sun50i_iommu_probe()
1030 ret = iommu_device_sysfs_add(&iommu->iommu, &pdev->dev, in sun50i_iommu_probe()
1035 ret = iommu_device_register(&iommu->iommu, &sun50i_iommu_ops, &pdev->dev); in sun50i_iommu_probe()
1040 dev_name(&pdev->dev), iommu); in sun50i_iommu_probe()
1047 iommu_device_unregister(&iommu->iommu); in sun50i_iommu_probe()
1050 iommu_device_sysfs_remove(&iommu->iommu); in sun50i_iommu_probe()
1053 iommu_group_put(iommu->group); in sun50i_iommu_probe()
1056 kmem_cache_destroy(iommu->pt_pool); in sun50i_iommu_probe()
1062 { .compatible = "allwinner,sun50i-h6-iommu", },
1069 .name = "sun50i-iommu",
1076 MODULE_DESCRIPTION("Allwinner H6 IOMMU driver");