/drivers/s390/cio/ |
D | vfio_ccw_cp.c | 70 static int pfn_array_pin(struct pfn_array *pa, struct device *mdev) in pfn_array_pin() argument 74 if (pa->pa_nr <= 0) { in pfn_array_pin() 75 pa->pa_nr = 0; in pfn_array_pin() 79 pa->pa_iova_pfn[0] = pa->pa_iova >> PAGE_SHIFT; in pfn_array_pin() 80 for (i = 1; i < pa->pa_nr; i++) in pfn_array_pin() 81 pa->pa_iova_pfn[i] = pa->pa_iova_pfn[i - 1] + 1; in pfn_array_pin() 83 ret = vfio_pin_pages(mdev, pa->pa_iova_pfn, pa->pa_nr, in pfn_array_pin() 84 IOMMU_READ | IOMMU_WRITE, pa->pa_pfn); in pfn_array_pin() 86 if (ret > 0 && ret != pa->pa_nr) { in pfn_array_pin() 87 vfio_unpin_pages(mdev, pa->pa_iova_pfn, ret); in pfn_array_pin() [all …]
|
/drivers/gpu/drm/i915/gvt/ |
D | mmio.c | 60 static void failsafe_emulate_mmio_rw(struct intel_vgpu *vgpu, uint64_t pa, in failsafe_emulate_mmio_rw() argument 72 offset = intel_vgpu_gpa_to_mmio_offset(vgpu, pa); in failsafe_emulate_mmio_rw() 97 gp = intel_vgpu_find_guest_page(vgpu, pa >> PAGE_SHIFT); in failsafe_emulate_mmio_rw() 102 intel_gvt_hypervisor_read_gpa(vgpu, pa, in failsafe_emulate_mmio_rw() 105 intel_gvt_hypervisor_write_gpa(vgpu, pa, in failsafe_emulate_mmio_rw() 122 int intel_vgpu_emulate_mmio_read(struct intel_vgpu *vgpu, uint64_t pa, in intel_vgpu_emulate_mmio_read() argument 131 failsafe_emulate_mmio_rw(vgpu, pa, p_data, bytes, true); in intel_vgpu_emulate_mmio_read() 139 gp = intel_vgpu_find_guest_page(vgpu, pa >> PAGE_SHIFT); in intel_vgpu_emulate_mmio_read() 141 ret = intel_gvt_hypervisor_read_gpa(vgpu, pa, in intel_vgpu_emulate_mmio_read() 146 ret, gp->gfn, pa, *(u32 *)p_data, in intel_vgpu_emulate_mmio_read() [all …]
|
/drivers/net/fddi/skfp/ |
D | pmf.c | 39 static int smt_set_para(struct s_smc *smc, struct smt_para *pa, int index, 316 struct smt_para *pa ; in smt_build_pmf_response() local 373 pa = (struct smt_para *) (req + 1) ; in smt_build_pmf_response() 378 if (((u_short)len < pa->p_len + PARA_LEN) || (pa->p_len & 3)) { in smt_build_pmf_response() 383 if (((range = (pa->p_type & 0xf000)) == 0x2000) || in smt_build_pmf_response() 388 index = *((u_char *)pa + PARA_LEN + 3) ;/* index */ in smt_build_pmf_response() 390 if (!set && (pa->p_len != 4)) { in smt_build_pmf_response() 419 if (!set && (pa->p_len != 0)) { in smt_build_pmf_response() 431 pt = smt_get_ptab(pa->p_type) ; in smt_build_pmf_response() 446 else if (pa->p_type != SMT_P_AUTHOR && in smt_build_pmf_response() [all …]
|
/drivers/infiniband/hw/hfi1/ |
D | trace_rx.h | 151 u32 npages, unsigned long va, unsigned long pa, 153 TP_ARGS(ctxt, subctxt, rarr, npages, va, pa, dma), 160 __field(unsigned long, pa) 169 __entry->pa = pa; 177 __entry->pa, 186 unsigned long va, unsigned long pa, dma_addr_t dma), 187 TP_ARGS(ctxt, subctxt, rarr, npages, va, pa, dma)); 192 unsigned long va, unsigned long pa, dma_addr_t dma), 193 TP_ARGS(ctxt, subctxt, rarr, npages, va, pa, dma)); 198 u32 index, u32 type, unsigned long pa, u16 order), [all …]
|
/drivers/dio/ |
D | dio.c | 123 unsigned long pa; in dio_find() local 128 pa = dio_scodetophysaddr(scode); in dio_find() 130 if (!pa) in dio_find() 134 va = (void *)(pa + DIO_VIRADDRBASE); in dio_find() 136 va = ioremap(pa, PAGE_SIZE); in dio_find() 196 unsigned long pa; in dio_init() local 201 pa = dio_scodetophysaddr(scode); in dio_init() 203 if (!pa) in dio_init() 207 va = (void *)(pa + DIO_VIRADDRBASE); in dio_init() 209 va = ioremap(pa, PAGE_SIZE); in dio_init() [all …]
|
/drivers/net/ethernet/amd/xgbe/ |
D | xgbe-desc.c | 286 struct xgbe_page_alloc *pa, int alloc_order, in xgbe_alloc_pages() argument 324 pa->pages = pages; in xgbe_alloc_pages() 325 pa->pages_len = PAGE_SIZE << order; in xgbe_alloc_pages() 326 pa->pages_offset = 0; in xgbe_alloc_pages() 327 pa->pages_dma = pages_dma; in xgbe_alloc_pages() 333 struct xgbe_page_alloc *pa, in xgbe_set_buffer_data() argument 336 get_page(pa->pages); in xgbe_set_buffer_data() 337 bd->pa = *pa; in xgbe_set_buffer_data() 339 bd->dma_base = pa->pages_dma; in xgbe_set_buffer_data() 340 bd->dma_off = pa->pages_offset; in xgbe_set_buffer_data() [all …]
|
/drivers/net/ethernet/synopsys/ |
D | dwc-xlgmac-desc.c | 41 if (desc_data->rx.hdr.pa.pages) in xlgmac_unmap_desc_data() 42 put_page(desc_data->rx.hdr.pa.pages); in xlgmac_unmap_desc_data() 51 if (desc_data->rx.buf.pa.pages) in xlgmac_unmap_desc_data() 52 put_page(desc_data->rx.buf.pa.pages); in xlgmac_unmap_desc_data() 331 struct xlgmac_page_alloc *pa, in xlgmac_alloc_pages() argument 357 pa->pages = pages; in xlgmac_alloc_pages() 358 pa->pages_len = PAGE_SIZE << order; in xlgmac_alloc_pages() 359 pa->pages_offset = 0; in xlgmac_alloc_pages() 360 pa->pages_dma = pages_dma; in xlgmac_alloc_pages() 366 struct xlgmac_page_alloc *pa, in xlgmac_set_buffer_data() argument [all …]
|
/drivers/sfi/ |
D | sfi_core.c | 121 static void sfi_print_table_header(unsigned long long pa, in sfi_print_table_header() argument 125 header->sig, pa, in sfi_print_table_header() 165 static struct sfi_table_header *sfi_map_table(u64 pa) in sfi_map_table() argument 170 if (!TABLE_ON_PAGE(syst_pa, pa, sizeof(struct sfi_table_header))) in sfi_map_table() 171 th = sfi_map_memory(pa, sizeof(struct sfi_table_header)); in sfi_map_table() 173 th = (void *)syst_va + (pa - syst_pa); in sfi_map_table() 181 if (!TABLE_ON_PAGE(syst_pa, pa, sizeof(struct sfi_table_header))) in sfi_map_table() 184 return sfi_map_memory(pa, length); in sfi_map_table() 232 __ref sfi_check_table(u64 pa, struct sfi_table_key *key) in sfi_check_table() argument 237 th = sfi_map_table(pa); in sfi_check_table() [all …]
|
/drivers/isdn/hardware/eicon/ |
D | diva.c | 170 diva_os_xdi_adapter_t *pdiva, *pa; in diva_driver_add_card() local 208 pa = pdiva; in diva_driver_add_card() 210 pa = diva_q_get_next(&pa->link); in diva_driver_add_card() 211 if (pa && !pa->interface.cleanup_adapter_proc) { in diva_driver_add_card() 212 pa->controller = i + 1 + j; in diva_driver_add_card() 213 pa->xdi_adapter.ANum = pa->controller; in diva_driver_add_card() 214 IoAdapters[i + j] = &pa->xdi_adapter; in diva_driver_add_card() 217 pa->controller)) in diva_driver_add_card() 218 create_adapter_proc(pa); /* add adapter to proc file system */ in diva_driver_add_card() 284 diva_os_xdi_adapter_t *pa; in diva_driver_remove_card() local [all …]
|
/drivers/media/cec/ |
D | cec-notifier.c | 37 void (*callback)(struct cec_adapter *adap, u16 pa); 88 void cec_notifier_set_phys_addr(struct cec_notifier *n, u16 pa) in cec_notifier_set_phys_addr() argument 94 n->phys_addr = pa; in cec_notifier_set_phys_addr() 104 u16 pa = CEC_PHYS_ADDR_INVALID; in cec_notifier_set_phys_addr_from_edid() local 110 pa = cec_get_edid_phys_addr((const u8 *)edid, in cec_notifier_set_phys_addr_from_edid() 112 cec_notifier_set_phys_addr(n, pa); in cec_notifier_set_phys_addr_from_edid() 118 void (*callback)(struct cec_adapter *adap, u16 pa)) in cec_notifier_register() argument
|
/drivers/infiniband/hw/usnic/ |
D | usnic_uiom.c | 86 dma_addr_t pa; in usnic_uiom_put_pages() local 91 pa = sg_phys(sg); in usnic_uiom_put_pages() 95 usnic_dbg("pa: %pa\n", &pa); in usnic_uiom_put_pages() 115 dma_addr_t pa; in usnic_uiom_get_pages() local 183 pa = sg_phys(sg); in usnic_uiom_get_pages() 185 cur_base + i*PAGE_SIZE, &pa); in usnic_uiom_get_pages() 263 dma_addr_t pa; in usnic_uiom_map_sorted_intervals() local 277 pa = sg_phys(&chunk->page_list[i]); in usnic_uiom_map_sorted_intervals() 284 pa_start = pa; in usnic_uiom_map_sorted_intervals() 285 pa_end = pa; in usnic_uiom_map_sorted_intervals() [all …]
|
/drivers/tee/ |
D | tee_shm.c | 240 int tee_shm_va2pa(struct tee_shm *shm, void *va, phys_addr_t *pa) in tee_shm_va2pa() argument 249 shm, (unsigned long)va - (unsigned long)shm->kaddr, pa); in tee_shm_va2pa() 260 int tee_shm_pa2va(struct tee_shm *shm, phys_addr_t pa, void **va) in tee_shm_pa2va() argument 263 if (pa < shm->paddr) in tee_shm_pa2va() 265 if (pa >= (shm->paddr + shm->size)) in tee_shm_pa2va() 269 void *v = tee_shm_get_va(shm, pa - shm->paddr); in tee_shm_pa2va() 302 int tee_shm_get_pa(struct tee_shm *shm, size_t offs, phys_addr_t *pa) in tee_shm_get_pa() argument 306 if (pa) in tee_shm_get_pa() 307 *pa = shm->paddr + offs; in tee_shm_get_pa()
|
/drivers/misc/mic/host/ |
D | mic_smpt.c | 32 static inline u8 mic_sys_addr_to_smpt(struct mic_device *mdev, dma_addr_t pa) in mic_sys_addr_to_smpt() argument 34 return (pa - mdev->smpt->info.base) >> mdev->smpt->info.page_shift; in mic_sys_addr_to_smpt() 42 static inline u64 mic_smpt_offset(struct mic_device *mdev, dma_addr_t pa) in mic_smpt_offset() argument 44 return pa & mic_system_page_mask(mdev); in mic_smpt_offset() 47 static inline u64 mic_smpt_align_low(struct mic_device *mdev, dma_addr_t pa) in mic_smpt_align_low() argument 49 return ALIGN(pa - mic_system_page_mask(mdev), in mic_smpt_align_low() 53 static inline u64 mic_smpt_align_high(struct mic_device *mdev, dma_addr_t pa) in mic_smpt_align_high() argument 55 return ALIGN(pa, mdev->smpt->info.page_size); in mic_smpt_align_high() 72 mic_is_system_addr(struct mic_device *mdev, dma_addr_t pa) in mic_is_system_addr() argument 74 return pa >= mdev->smpt->info.base && pa <= mic_max_system_addr(mdev); in mic_is_system_addr()
|
/drivers/net/wireless/ath/wil6210/ |
D | txrx.c | 191 vring->va = dma_alloc_coherent(dev, sz, &vring->pa, GFP_KERNEL); in wil_vring_alloc() 212 vring->va, &vring->pa, vring->ctx); in wil_vring_alloc() 220 dma_addr_t pa = wil_desc_addr(&d->dma.addr); in wil_txdesc_unmap() local 225 dma_unmap_single(dev, pa, dmalen, DMA_TO_DEVICE); in wil_txdesc_unmap() 228 dma_unmap_page(dev, pa, dmalen, DMA_TO_DEVICE); in wil_txdesc_unmap() 247 &vring->pa, vring->ctx); in wil_vring_free() 251 &vring->pa, vring->ctx); in wil_vring_free() 255 dma_addr_t pa; in wil_vring_free() local 284 pa = wil_desc_addr(&d->dma.addr); in wil_vring_free() 286 dma_unmap_single(dev, pa, dmalen, DMA_FROM_DEVICE); in wil_vring_free() [all …]
|
/drivers/net/wireless/broadcom/brcm80211/brcmsmac/ |
D | dma.c | 716 dma_addr_t pa, uint outidx, u32 *flags, u32 bufcount) in dma64_dd_upd() argument 721 if ((di->dataoffsetlow == 0) || !(pa & PCI32ADDR_HIGH)) { in dma64_dd_upd() 722 ddring[outidx].addrlow = cpu_to_le32(pa + di->dataoffsetlow); in dma64_dd_upd() 730 ae = (pa & PCI32ADDR_HIGH) >> PCI32ADDR_HIGH_SHIFT; in dma64_dd_upd() 731 pa &= ~PCI32ADDR_HIGH; in dma64_dd_upd() 734 ddring[outidx].addrlow = cpu_to_le32(pa + di->dataoffsetlow); in dma64_dd_upd() 774 _dma_ddtable_init(struct dma_info *di, uint direction, dma_addr_t pa) in _dma_ddtable_init() argument 778 di->xmtptrbase = pa; in _dma_ddtable_init() 780 di->rcvptrbase = pa; in _dma_ddtable_init() 784 || !(pa & PCI32ADDR_HIGH)) { in _dma_ddtable_init() [all …]
|
/drivers/iommu/ |
D | omap-iommu.c | 134 u32 l, pa; in omap2_iommu_enable() local 139 pa = virt_to_phys(obj->iopgd); in omap2_iommu_enable() 140 if (!IS_ALIGNED(pa, SZ_16K)) in omap2_iommu_enable() 147 iommu_write_reg(obj, pa, MMU_TTB); in omap2_iommu_enable() 315 cr->ram = e->pa | e->endian | e->elsz | e->mixed; in iotlb_alloc_cr() 540 static int iopgd_alloc_section(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) in iopgd_alloc_section() argument 545 if ((da | pa) & ~IOSECTION_MASK) { in iopgd_alloc_section() 547 __func__, da, pa, IOSECTION_SIZE); in iopgd_alloc_section() 551 *iopgd = (pa & IOSECTION_MASK) | prot | IOPGD_SECTION; in iopgd_alloc_section() 556 static int iopgd_alloc_super(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) in iopgd_alloc_super() argument [all …]
|
/drivers/base/ |
D | platform.c | 247 struct platform_object *pa = container_of(dev, struct platform_object, in platform_device_release() local 250 of_device_node_put(&pa->pdev.dev); in platform_device_release() 251 kfree(pa->pdev.dev.platform_data); in platform_device_release() 252 kfree(pa->pdev.mfd_cell); in platform_device_release() 253 kfree(pa->pdev.resource); in platform_device_release() 254 kfree(pa->pdev.driver_override); in platform_device_release() 255 kfree(pa); in platform_device_release() 268 struct platform_object *pa; in platform_device_alloc() local 270 pa = kzalloc(sizeof(*pa) + strlen(name) + 1, GFP_KERNEL); in platform_device_alloc() 271 if (pa) { in platform_device_alloc() [all …]
|
/drivers/infiniband/hw/ocrdma/ |
D | ocrdma_hw.c | 449 ocrdma_build_q_pages(&cmd->pa[0], cmd->num_pages, eq->q.dma, in ocrdma_mbx_create_eq() 545 ocrdma_build_q_pages(&cmd->pa[0], cq->size / OCRDMA_MIN_Q_PAGE_SIZE, in ocrdma_mbx_mq_cq_create() 572 struct ocrdma_pa *pa; in ocrdma_mbx_create_mq() local 593 pa = &cmd->pa[0]; in ocrdma_mbx_create_mq() 595 ocrdma_build_q_pages(pa, num_pages, mq->dma, PAGE_SIZE_4K); in ocrdma_mbx_create_mq() 1303 mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dev->stats_mem.pa & 0xffffffff); in ocrdma_mbx_rdma_stats() 1304 mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dev->stats_mem.pa); in ocrdma_mbx_rdma_stats() 1343 dma.size, &dma.pa, GFP_KERNEL); in ocrdma_mbx_get_ctrl_attribs() 1351 mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dma.pa & 0xffffffff); in ocrdma_mbx_get_ctrl_attribs() 1352 mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dma.pa); in ocrdma_mbx_get_ctrl_attribs() [all …]
|
/drivers/crypto/nx/ |
D | nx-842.h | 101 #define LEN_ON_SIZE(pa, size) ((size) - ((pa) & ((size) - 1))) argument 102 #define LEN_ON_PAGE(pa) LEN_ON_SIZE(pa, PAGE_SIZE) argument
|
/drivers/tty/serial/8250/ |
D | 8250_hp300.c | 126 unsigned long pa = dio_scodetophysaddr(scode); in hp300_setup_serial_console() local 127 if (!pa) in hp300_setup_serial_console() 133 port.mapbase = (pa + UART_OFFSET); in hp300_setup_serial_console() 136 port.irq = DIO_IPL(pa + DIO_VIRADDRBASE); in hp300_setup_serial_console() 139 out_8(pa + DIO_VIRADDRBASE + DCA_IC, DCA_IC_IE); in hp300_setup_serial_console() 141 if (DIO_ID(pa + DIO_VIRADDRBASE) & 0x80) in hp300_setup_serial_console()
|
/drivers/net/ethernet/brocade/bna/ |
D | bfa_ioc.h | 48 u64 pa; /* ! Physical address */ member 58 #define bfa_dma_be_addr_set(dma_addr, pa) \ argument 59 __bfa_dma_be_addr_set(&dma_addr, (u64)pa) 61 __bfa_dma_be_addr_set(union bfi_addr_u *dma_addr, u64 pa) in __bfa_dma_be_addr_set() argument 63 dma_addr->a32.addr_lo = (u32) htonl(pa); in __bfa_dma_be_addr_set() 64 dma_addr->a32.addr_hi = (u32) htonl(upper_32_bits(pa)); in __bfa_dma_be_addr_set() 71 __bfa_alen_set(struct bfi_alen *alen, u32 len, u64 pa) in __bfa_alen_set() argument 74 bfa_dma_be_addr_set(alen->al_addr, pa); in __bfa_alen_set()
|
/drivers/scsi/sym53c8xx_2/ |
D | sym_fw.c | 260 u32 *pa; in sym_fw_setup_bus_addresses() local 269 pa = (u32 *) &np->fwa_bas; in sym_fw_setup_bus_addresses() 271 pa[i] = np->scripta_ba + po[i]; in sym_fw_setup_bus_addresses() 277 pa = (u32 *) &np->fwb_bas; in sym_fw_setup_bus_addresses() 279 pa[i] = np->scriptb_ba + po[i]; in sym_fw_setup_bus_addresses() 285 pa = (u32 *) &np->fwz_bas; in sym_fw_setup_bus_addresses() 287 pa[i] = np->scriptz_ba + po[i]; in sym_fw_setup_bus_addresses()
|
/drivers/perf/ |
D | xgene_pmu.c | 445 XGENE_PMU_EVENT_ATTR(pa-req-buf-alloc-all, 0x01), 446 XGENE_PMU_EVENT_ATTR(pa-req-buf-alloc-rd, 0x02), 447 XGENE_PMU_EVENT_ATTR(pa-req-buf-alloc-wr, 0x03), 448 XGENE_PMU_EVENT_ATTR(pa-all-cp-req, 0x04), 449 XGENE_PMU_EVENT_ATTR(pa-cp-blk-req, 0x05), 450 XGENE_PMU_EVENT_ATTR(pa-cp-ptl-req, 0x06), 451 XGENE_PMU_EVENT_ATTR(pa-cp-rd-req, 0x07), 452 XGENE_PMU_EVENT_ATTR(pa-cp-wr-req, 0x08), 456 XGENE_PMU_EVENT_ATTR(pa-rd-shared-req-issued, 0x10), 457 XGENE_PMU_EVENT_ATTR(pa-rd-exclusive-req-issued, 0x11), [all …]
|
/drivers/media/usb/pulse8-cec/ |
D | pulse8-cec.c | 335 struct cec_log_addrs *log_addrs, u16 *pa) in pulse8_setup() argument 352 *pa = CEC_PHYS_ADDR_INVALID; in pulse8_setup() 432 *pa = (data[0] << 8) | data[1]; in pulse8_setup() 434 cec_phys_addr_exp(*pa)); in pulse8_setup() 455 u16 pa) in pulse8_apply_persistent_config() argument 463 cec_s_phys_addr(pulse8->adap, pa, false); in pulse8_apply_persistent_config() 485 u16 pa = adap->phys_addr; in pulse8_cec_adap_log_addr() local 559 cmd[1] = pa >> 8; in pulse8_cec_adap_log_addr() 560 cmd[2] = pa & 0xff; in pulse8_cec_adap_log_addr() 658 u16 pa = CEC_PHYS_ADDR_INVALID; in pulse8_connect() local [all …]
|
/drivers/net/ethernet/intel/i40evf/ |
D | i40e_adminq.c | 181 cpu_to_le32(upper_32_bits(bi->pa)); in i40e_alloc_arq_bufs() 183 cpu_to_le32(lower_32_bits(bi->pa)); in i40e_alloc_arq_bufs() 270 if (hw->aq.asq.r.asq_bi[i].pa) in i40e_free_asq_bufs() 301 wr32(hw, hw->aq.asq.bal, lower_32_bits(hw->aq.asq.desc_buf.pa)); in i40e_config_asq_regs() 302 wr32(hw, hw->aq.asq.bah, upper_32_bits(hw->aq.asq.desc_buf.pa)); in i40e_config_asq_regs() 306 if (reg != lower_32_bits(hw->aq.asq.desc_buf.pa)) in i40e_config_asq_regs() 330 wr32(hw, hw->aq.arq.bal, lower_32_bits(hw->aq.arq.desc_buf.pa)); in i40e_config_arq_regs() 331 wr32(hw, hw->aq.arq.bah, upper_32_bits(hw->aq.arq.desc_buf.pa)); in i40e_config_arq_regs() 338 if (reg != lower_32_bits(hw->aq.arq.desc_buf.pa)) in i40e_config_arq_regs() 773 cpu_to_le32(upper_32_bits(dma_buff->pa)); in i40evf_asq_send_command() [all …]
|