/arch/sparc/kernel/ |
D | of_device_common.c | 19 if (!op || index >= op->archdata.num_irqs) in irq_of_parse_and_map() 22 return op->archdata.irqs[index]; in irq_of_parse_and_map() 34 memcpy(r, &op->archdata.resource[index], sizeof(*r)); in of_address_to_resource() 47 r = &op->archdata.resource[index]; in of_iomap() 58 struct dev_archdata *bus_sd = &bus->dev.archdata; in of_propagate_archdata() 65 op->dev.archdata.iommu = bus_sd->iommu; in of_propagate_archdata() 66 op->dev.archdata.stc = bus_sd->stc; in of_propagate_archdata() 67 op->dev.archdata.host_controller = bus_sd->host_controller; in of_propagate_archdata() 68 op->dev.archdata.numa_node = bus_sd->numa_node; in of_propagate_archdata()
|
D | of_device_32.c | 271 op->resource = op->archdata.resource; in build_device_resources() 350 sd = &op->dev.archdata; in scan_one_device() 357 op->archdata.num_irqs = len / sizeof(struct linux_prom_irqs); in scan_one_device() 358 for (i = 0; i < op->archdata.num_irqs; i++) in scan_one_device() 359 op->archdata.irqs[i] = in scan_one_device() 366 op->archdata.num_irqs = len / sizeof(unsigned int); in scan_one_device() 367 for (i = 0; i < op->archdata.num_irqs; i++) in scan_one_device() 368 op->archdata.irqs[i] = in scan_one_device() 371 op->archdata.num_irqs = 0; in scan_one_device()
|
D | iommu.c | 214 nid = dev->archdata.numa_node; in dma_4u_alloc_coherent() 222 iommu = dev->archdata.iommu; in dma_4u_alloc_coherent() 255 iommu = dev->archdata.iommu; in dma_4u_free_coherent() 277 iommu = dev->archdata.iommu; in dma_4u_map_page() 278 strbuf = dev->archdata.stc; in dma_4u_map_page() 402 iommu = dev->archdata.iommu; in dma_4u_unmap_page() 403 strbuf = dev->archdata.stc; in dma_4u_unmap_page() 449 iommu = dev->archdata.iommu; in dma_4u_map_sg() 450 strbuf = dev->archdata.stc; in dma_4u_map_sg() 620 iommu = dev->archdata.iommu; in dma_4u_unmap_sg() [all …]
|
D | of_device_64.c | 349 op->resource = op->archdata.resource; in build_device_resources() 647 sd = &op->dev.archdata; in scan_one_device() 654 op->archdata.num_irqs = len / 4; in scan_one_device() 657 if (op->archdata.num_irqs > PROMINTR_MAX) { in scan_one_device() 660 dp->full_name, op->archdata.num_irqs, PROMINTR_MAX); in scan_one_device() 661 op->archdata.num_irqs = PROMINTR_MAX; in scan_one_device() 663 memcpy(op->archdata.irqs, irq, op->archdata.num_irqs * 4); in scan_one_device() 665 op->archdata.num_irqs = 0; in scan_one_device() 669 for (i = 0; i < op->archdata.num_irqs; i++) in scan_one_device() 670 op->archdata.irqs[i] = build_one_device_irq(op, parent, op->archdata.irqs[i]); in scan_one_device()
|
D | sbus.c | 61 struct iommu *iommu = dev->archdata.iommu; in sbus_set_sbus64() 211 struct iommu *iommu = op->dev.archdata.iommu; in sbus_build_irq() 273 struct iommu *iommu = op->dev.archdata.iommu; in sysio_ue_handler() 347 struct iommu *iommu = op->dev.archdata.iommu; in sysio_ce_handler() 426 struct iommu *iommu = op->dev.archdata.iommu; in sysio_sbus_error_handler() 495 struct iommu *iommu = op->dev.archdata.iommu; in sysio_register_error_handlers() 562 op->dev.archdata.iommu = iommu; in sbus_iommu_init() 563 op->dev.archdata.stc = strbuf; in sbus_iommu_init() 564 op->dev.archdata.numa_node = -1; in sbus_iommu_init()
|
D | power.c | 40 unsigned int irq = op->archdata.irqs[0]; in power_probe()
|
D | pci_sun4v.c | 79 struct pci_pbm_info *pbm = p->dev->archdata.host_controller; in iommu_batch_flush() 199 nid = dev->archdata.numa_node; in dma_4v_alloc_coherent() 207 iommu = dev->archdata.iommu; in dma_4v_alloc_coherent() 331 iommu = dev->archdata.iommu; in dma_4v_free_coherent() 332 pbm = dev->archdata.host_controller; in dma_4v_free_coherent() 366 iommu = dev->archdata.iommu; in dma_4v_map_page() 444 iommu = dev->archdata.iommu; in dma_4v_unmap_page() 445 pbm = dev->archdata.host_controller; in dma_4v_unmap_page() 484 iommu = dev->archdata.iommu; in dma_4v_map_sg() 637 iommu = dev->archdata.iommu; in dma_4v_unmap_sg() [all …]
|
D | pci.c | 276 sd = &dev->dev.archdata; in of_create_pci_dev() 279 sd = &op->dev.archdata; in of_create_pci_dev() 351 dev->irq = sd->op->archdata.irqs[0]; in of_create_pci_dev() 736 struct pci_pbm_info *pbm = pdev->dev.archdata.host_controller; in __pci_mmap_make_offset_bus() 909 struct pci_pbm_info *pbm = pdev->dev.archdata.host_controller; in arch_setup_msi_irq() 922 struct pci_pbm_info *pbm = pdev->dev.archdata.host_controller; in arch_teardown_msi_irq() 957 struct iommu *iommu = pdev->dev.archdata.iommu; in pci64_dma_supported() 1011 psd = &pdev->dev.archdata; in pcibios_add_device() 1012 pci_init_dev_archdata(&dev->dev.archdata, psd->iommu, in pcibios_add_device()
|
D | pci_schizo.c | 862 err = request_irq(op->archdata.irqs[1], schizo_ue_intr, 0, in tomatillo_register_error_handlers() 869 err = request_irq(op->archdata.irqs[2], schizo_ce_intr, 0, in tomatillo_register_error_handlers() 877 err = request_irq(op->archdata.irqs[0], schizo_pcierr_intr, 0, in tomatillo_register_error_handlers() 880 err = request_irq(op->archdata.irqs[0], schizo_pcierr_intr, 0, in tomatillo_register_error_handlers() 888 err = request_irq(op->archdata.irqs[3], schizo_safarierr_intr, 0, in tomatillo_register_error_handlers() 957 err = request_irq(op->archdata.irqs[1], schizo_ue_intr, 0, in schizo_register_error_handlers() 964 err = request_irq(op->archdata.irqs[2], schizo_ce_intr, 0, in schizo_register_error_handlers() 972 err = request_irq(op->archdata.irqs[0], schizo_pcierr_intr, 0, in schizo_register_error_handlers() 975 err = request_irq(op->archdata.irqs[0], schizo_pcierr_intr, 0, in schizo_register_error_handlers() 983 err = request_irq(op->archdata.irqs[3], schizo_safarierr_intr, 0, in schizo_register_error_handlers()
|
D | pci_sabre.c | 334 if (op->archdata.num_irqs < 4) in sabre_register_error_handlers() 346 err = request_irq(op->archdata.irqs[1], sabre_ue_intr, 0, "SABRE_UE", pbm); in sabre_register_error_handlers() 356 err = request_irq(op->archdata.irqs[2], sabre_ce_intr, 0, "SABRE_CE", pbm); in sabre_register_error_handlers() 360 err = request_irq(op->archdata.irqs[0], psycho_pcierr_intr, 0, in sabre_register_error_handlers()
|
/arch/arm/common/ |
D | dmabounce.c | 91 struct dmabounce_device_info *device_info = dev->archdata.dmabounce; in dmabounce_show() 214 if (!dev || !dev->archdata.dmabounce) in find_safe_buffer_dev() 220 return find_safe_buffer(dev->archdata.dmabounce, dma_addr); in find_safe_buffer_dev() 225 if (!dev || !dev->archdata.dmabounce) in needs_bounce() 243 return !!dev->archdata.dmabounce->needs_bounce(dev, dma_addr, size); in needs_bounce() 250 struct dmabounce_device_info *device_info = dev->archdata.dmabounce; in map_single() 288 DO_STATS(dev->archdata.dmabounce->bounce_count++); in unmap_single() 305 free_safe_buffer(dev->archdata.dmabounce, buf); in unmap_single() 389 DO_STATS(dev->archdata.dmabounce->bounce_count++); in __dmabounce_sync_for_cpu() 429 DO_STATS(dev->archdata.dmabounce->bounce_count++); in __dmabounce_sync_for_device() [all …]
|
/arch/powerpc/kernel/ |
D | dma-swiotlb.c | 29 u64 end, mask, max_direct_dma_addr = dev->archdata.max_direct_dma_addr; in swiotlb_powerpc_get_required() 72 sd = &pdev->dev.archdata; in pci_dma_dev_setup_swiotlb() 87 sd = &dev->archdata; in ppc_swiotlb_bus_notify()
|
D | pci_dn.c | 81 if (pdev->dev.archdata.pci_data) in pci_get_pdn_by_devfn() 82 return pdev->dev.archdata.pci_data; in pci_get_pdn_by_devfn() 114 if (pdev->dev.archdata.pci_data) in pci_get_pdn() 115 return pdev->dev.archdata.pci_data; in pci_get_pdn() 547 if (pdev->dev.archdata.pci_data) in pci_dev_pdn_setup() 552 pdev->dev.archdata.pci_data = pdn; in pci_dev_pdn_setup()
|
/arch/tile/include/asm/ |
D | dma-mapping.h | 39 return dev->archdata.dma_offset; in get_dma_offset() 44 dev->archdata.dma_offset = off; in set_dma_offset()
|
/arch/sparc/mm/ |
D | io-unit.c | 66 op->dev.archdata.iommu = iounit; in iounit_iommu_init() 145 struct iounit_struct *iounit = dev->archdata.iommu; in iounit_get_scsi_one() 156 struct iounit_struct *iounit = dev->archdata.iommu; in iounit_get_scsi_sgl() 172 struct iounit_struct *iounit = dev->archdata.iommu; in iounit_release_scsi_one() 186 struct iounit_struct *iounit = dev->archdata.iommu; in iounit_release_scsi_sgl() 206 struct iounit_struct *iounit = dev->archdata.iommu; in iounit_map_dma_area()
|
D | iommu.c | 131 op->dev.archdata.iommu = iommu; in sbus_iommu_init() 180 struct iommu_struct *iommu = dev->archdata.iommu; in iommu_get_one() 286 struct iommu_struct *iommu = dev->archdata.iommu; in iommu_release_one() 328 struct iommu_struct *iommu = dev->archdata.iommu; in iommu_map_dma_area() 394 struct iommu_struct *iommu = dev->archdata.iommu; in iommu_unmap_dma_area()
|
/arch/powerpc/include/asm/ |
D | dma-mapping.h | 99 return dev->archdata.dma_offset; in get_dma_offset() 107 dev->archdata.dma_offset = off; in set_dma_offset() 121 struct dev_archdata *sd = &dev->archdata; in dma_capable()
|
/arch/arm/include/asm/ |
D | device.h | 35 #define to_dma_iommu_mapping(dev) ((dev)->archdata.mapping)
|
/arch/arm/mm/ |
D | dma-mapping-nommu.c | 231 dev->archdata.dma_coherent = (cacheid) ? coherent : true; in arch_setup_dma_ops() 236 dev->archdata.dma_coherent = (get_cr() & CR_M) ? coherent : true; in arch_setup_dma_ops() 239 dma_ops = arm_nommu_get_dma_map_ops(dev->archdata.dma_coherent); in arch_setup_dma_ops()
|
/arch/ia64/sn/kernel/sn2/ |
D | timer.c | 57 clocksource_sn2.archdata.fsys_mmio = RTC_COUNTER_ADDR; in sn_timer_init()
|
/arch/mips/include/asm/ |
D | dma-mapping.h | 39 dev->archdata.dma_coherent = coherent; in arch_setup_dma_ops()
|
/arch/mips/kernel/ |
D | csrc-r4k.c | 81 clocksource_mips.archdata.vdso_clock_mode = VDSO_CLOCK_R4K; in init_r4k_clocksource()
|
/arch/mips/include/asm/mach-generic/ |
D | dma-coherence.h | 53 return dev->archdata.dma_coherent; in plat_device_is_coherent()
|
/arch/arm/mach-omap2/ |
D | omap_device.h | 102 return pdev ? pdev->archdata.od : NULL; in to_omap_device()
|
/arch/arm64/include/asm/ |
D | dma-mapping.h | 50 return dev->archdata.dma_coherent; in is_device_dma_coherent()
|