/drivers/parisc/ |
D | iommu-helpers.h | 21 unsigned long dma_offset = 0, dma_len = 0; in iommu_fill_pdir() local 46 BUG_ON(pdirp && (dma_len != sg_dma_len(dma_sg))); in iommu_fill_pdir() 50 dma_len = sg_dma_len(startsg); in iommu_fill_pdir() 105 unsigned long dma_offset, dma_len; /* start/len of DMA stream */ in iommu_coalesce_chunks() local 115 dma_len = startsg->length; in iommu_coalesce_chunks() 145 if(unlikely(ALIGN(dma_len + dma_offset + startsg->length, in iommu_coalesce_chunks() 149 if (startsg->length + dma_len > max_seg_size) in iommu_coalesce_chunks() 159 dma_len += startsg->length; in iommu_coalesce_chunks() 167 sg_dma_len(contig_sg) = dma_len; in iommu_coalesce_chunks() 168 dma_len = ALIGN(dma_len + dma_offset, IOVP_SIZE); in iommu_coalesce_chunks() [all …]
|
/drivers/net/wireless/brcm80211/brcmsmac/ |
D | brcms_trace_events.h | 162 u16 dma_len), 164 dma_len), 171 __field(u16, dma_len) 179 __entry->dma_len = dma_len; 184 __entry->ampdu_frames, __entry->dma_len)
|
D | ampdu.c | 509 session->dma_len = 0; in brcms_c_ampdu_reset_session() 568 session->dma_len += p->len; in brcms_c_ampdu_add_frame() 644 u16 dma_len = session->dma_len; in brcms_c_ampdu_finalize() local 771 dma_len = min(dma_len, f->ampdu_pld_size); in brcms_c_ampdu_finalize() 772 txh->PreloadSize = cpu_to_le16(dma_len); in brcms_c_ampdu_finalize()
|
D | ampdu.h | 37 u16 dma_len; member
|
/drivers/net/ethernet/sfc/ |
D | tx.c | 342 unsigned int dma_len; in efx_enqueue_skb() local 391 dma_len = efx_max_tx_len(efx, dma_addr); in efx_enqueue_skb() 392 if (likely(dma_len >= len)) in efx_enqueue_skb() 393 dma_len = len; in efx_enqueue_skb() 396 buffer->len = dma_len; in efx_enqueue_skb() 399 len -= dma_len; in efx_enqueue_skb() 400 dma_addr += dma_len; in efx_enqueue_skb() 930 unsigned dma_len; in efx_tx_queue_insert() local 944 dma_len = efx_max_tx_len(efx, dma_addr); in efx_tx_queue_insert() 947 if (dma_len >= len) in efx_tx_queue_insert() [all …]
|
/drivers/usb/host/whci/ |
D | qset.c | 464 size_t dma_len; in qset_add_urb_sg() local 492 dma_len = dma_remaining; in qset_add_urb_sg() 500 if (std->len + dma_len > QTD_MAX_XFER_SIZE) { in qset_add_urb_sg() 501 dma_len = (QTD_MAX_XFER_SIZE / qset->max_packet) in qset_add_urb_sg() 505 std->len += dma_len; in qset_add_urb_sg() 509 ep = dma_addr + dma_len; in qset_add_urb_sg() 529 dma_remaining -= dma_len; in qset_add_urb_sg() 530 remaining -= dma_len; in qset_add_urb_sg()
|
/drivers/net/ethernet/chelsio/cxgb/ |
D | sge.c | 165 DEFINE_DMA_UNMAP_LEN(dma_len); 171 DEFINE_DMA_UNMAP_LEN(dma_len); 513 dma_unmap_len(ce, dma_len), in free_freelQ_buffers() 627 if (likely(dma_unmap_len(ce, dma_len))) { in free_cmdQ_buffers() 629 dma_unmap_len(ce, dma_len), in free_cmdQ_buffers() 830 unsigned int dma_len = q->rx_buffer_size - q->dma_offset; in refill_free_list() local 841 mapping = pci_map_single(pdev, skb->data, dma_len, in refill_free_list() 847 dma_unmap_len_set(ce, dma_len, dma_len); in refill_free_list() 850 e->len_gen = V_CMD_LEN(dma_len) | V_CMD_GEN1(q->genbit); in refill_free_list() 1054 dma_unmap_len(ce, dma_len), in get_packet() [all …]
|
/drivers/crypto/ |
D | talitos.c | 733 int dma_len; member 810 if (edesc->dma_len) in ipsec_esp_unmap() 811 dma_unmap_single(dev, edesc->dma_link_tbl, edesc->dma_len, in ipsec_esp_unmap() 865 if (edesc->dma_len) in ipsec_esp_decrypt_swauth_done() 984 edesc->dma_len, DMA_BIDIRECTIONAL); in ipsec_esp() 1034 edesc->dma_len, in ipsec_esp() 1077 edesc->dma_len, DMA_BIDIRECTIONAL); in ipsec_esp() 1129 int assoc_nents = 0, src_nents, dst_nents, alloc_len, dma_len; in talitos_edesc_alloc() local 1179 dma_len = (src_nents + dst_nents + 2 + assoc_nents) * in talitos_edesc_alloc() 1181 alloc_len += dma_len; in talitos_edesc_alloc() [all …]
|
/drivers/mmc/host/ |
D | bfin_sdh.c | 69 int dma_len; member 158 host->dma_len = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, host->dma_dir); in sdh_setup_data() 169 for_each_sg(data->sg, sg, host->dma_len, i) { in sdh_setup_data() 183 host->dma_len * sizeof(struct dma_desc_array)); in sdh_setup_data() 185 host->sg_cpu[host->dma_len - 1].cfg &= ~(DMAFLOW | NDSIZE); in sdh_setup_data() 186 host->sg_cpu[host->dma_len - 1].cfg |= DI_EN; in sdh_setup_data()
|
D | tifm_sd.c | 267 unsigned int dma_len, dma_blk_cnt, dma_off; in tifm_sd_set_dma_data() local 283 dma_len = sg_dma_len(&r_data->sg[host->sg_pos]) - host->block_pos; in tifm_sd_set_dma_data() 284 if (!dma_len) { in tifm_sd_set_dma_data() 289 dma_len = sg_dma_len(&r_data->sg[host->sg_pos]); in tifm_sd_set_dma_data() 292 if (dma_len < t_size) { in tifm_sd_set_dma_data() 293 dma_blk_cnt = dma_len / r_data->blksz; in tifm_sd_set_dma_data() 304 else if (dma_len) { in tifm_sd_set_dma_data()
|
D | omap_hsmmc.c | 170 unsigned int dma_len; member 196 unsigned int dma_len; member 1348 int dma_len; in omap_hsmmc_pre_dma_transfer() local 1360 dma_len = dma_map_sg(chan->device->dev, data->sg, data->sg_len, in omap_hsmmc_pre_dma_transfer() 1364 dma_len = host->next_data.dma_len; in omap_hsmmc_pre_dma_transfer() 1365 host->next_data.dma_len = 0; in omap_hsmmc_pre_dma_transfer() 1369 if (dma_len == 0) in omap_hsmmc_pre_dma_transfer() 1373 next->dma_len = dma_len; in omap_hsmmc_pre_dma_transfer() 1376 host->dma_len = dma_len; in omap_hsmmc_pre_dma_transfer()
|
D | pxamci.c | 74 unsigned int dma_len; member 212 host->dma_len = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, in pxamci_setup_data() 215 for (i = 0; i < host->dma_len; i++) { in pxamci_setup_data() 233 host->sg_cpu[host->dma_len - 1].ddadr = DDADR_STOP; in pxamci_setup_data()
|
/drivers/scsi/ |
D | esp_scsi.c | 478 static u32 esp_dma_length_limit(struct esp *esp, u32 dma_addr, u32 dma_len) in esp_dma_length_limit() argument 482 if (dma_len > (1U << 24)) in esp_dma_length_limit() 483 dma_len = (1U << 24); in esp_dma_length_limit() 493 if (dma_len > (1U << 16)) in esp_dma_length_limit() 494 dma_len = (1U << 16); in esp_dma_length_limit() 500 end = base + dma_len; in esp_dma_length_limit() 503 dma_len = end - base; in esp_dma_length_limit() 505 return dma_len; in esp_dma_length_limit() 1692 unsigned int dma_len = esp_cur_dma_len(ent, cmd); in esp_process_event() local 1703 dma_len = esp->ops->dma_length_limit(esp, dma_addr, in esp_process_event() [all …]
|
D | mac53c94.c | 367 u32 dma_len; in set_dma_cmds() local 381 dma_len = sg_dma_len(scl); in set_dma_cmds() 382 if (dma_len > 0xffff) in set_dma_cmds() 384 total += dma_len; in set_dma_cmds() 385 st_le16(&dcmds->req_count, dma_len); in set_dma_cmds()
|
D | mac_esp.c | 437 static u32 mac_esp_dma_length_limit(struct esp *esp, u32 dma_addr, u32 dma_len) in mac_esp_dma_length_limit() argument 439 return dma_len > 0xFFFF ? 0xFFFF : dma_len; in mac_esp_dma_length_limit()
|
D | atari_NCR5380.c | 835 hostdata->dma_len = 0; in NCR5380_init() 1148 && !hostdata->dma_len in DEF_SCSI_QCMD() 1214 transfered = hostdata->dma_len - NCR5380_dma_residual(instance); in NCR5380_dma_complete() 1215 hostdata->dma_len = 0; in NCR5380_dma_complete() 1929 hostdata->dma_len = (p & SR_IO) ? in NCR5380_transfer_dma() 1947 hostdata->dma_len = (p & SR_IO) ? in NCR5380_transfer_dma() 2884 hostdata->dma_len = 0; in NCR5380_bus_reset() 2962 hostdata->dma_len = 0; in NCR5380_bus_reset()
|
D | sun3_NCR5380.c | 781 hostdata->dma_len = 0; in NCR5380_init() 1082 && !hostdata->dma_len in DEF_SCSI_QCMD() 1156 transfered = hostdata->dma_len - NCR5380_dma_residual(instance); in NCR5380_dma_complete() 1157 hostdata->dma_len = 0; in NCR5380_dma_complete() 1848 hostdata->dma_len = c; in NCR5380_transfer_dma() 2850 hostdata->dma_len = 0; in NCR5380_bus_reset() 2922 hostdata->dma_len = 0; in NCR5380_bus_reset()
|
/drivers/ata/ |
D | pata_arasan_cf.c | 434 u32 xfer_cnt, sglen, dma_len, xfer_ctr; in sg_xfer() local 474 dma_len = min(xfer_cnt, FIFO_SIZE); in sg_xfer() 475 ret = dma_xfer(acdev, src, dest, dma_len); in sg_xfer() 482 src += dma_len; in sg_xfer() 484 dest += dma_len; in sg_xfer() 486 sglen -= dma_len; in sg_xfer() 487 xfer_cnt -= dma_len; in sg_xfer()
|
/drivers/block/rsxx/ |
D | dma.c | 647 unsigned int dma_len, in rsxx_queue_dma() argument 663 dma->sub_page.cnt = (dma_len >> 9); in rsxx_queue_dma() 694 unsigned int dma_len; in rsxx_dma_queue_bio() local 734 dma_len = min(bv_len, in rsxx_dma_queue_bio() 739 dma_off, dma_len, in rsxx_dma_queue_bio() 747 addr8 += dma_len; in rsxx_dma_queue_bio() 748 bv_off += dma_len; in rsxx_dma_queue_bio() 749 bv_len -= dma_len; in rsxx_dma_queue_bio()
|
/drivers/infiniband/ulp/srp/ |
D | ib_srp.c | 1154 unsigned int dma_len, u32 rkey) in srp_map_desc() argument 1160 desc->len = cpu_to_be32(dma_len); in srp_map_desc() 1162 state->total_len += dma_len; in srp_map_desc() 1181 srp_map_desc(state, 0, state->dma_len, fmr->fmr->rkey); in srp_map_finish_fmr() 1212 wr.wr.fast_reg.length = state->dma_len; in srp_map_finish_fr() 1221 srp_map_desc(state, state->base_dma_addr, state->dma_len, in srp_map_finish_fr() 1236 srp_map_desc(state, state->base_dma_addr, state->dma_len, in srp_finish_mapping() 1245 state->dma_len = 0; in srp_finish_mapping() 1268 unsigned int dma_len = ib_sg_dma_len(ibdev, sg); in srp_map_sg_entry() local 1272 if (!dma_len) in srp_map_sg_entry() [all …]
|
/drivers/infiniband/ulp/srpt/ |
D | ib_srpt.c | 1093 u32 dma_len; in srpt_map_sg_to_ib_sge() local 1128 dma_len = ib_sg_dma_len(dev, &sg[0]); in srpt_map_sg_to_ib_sge() 1150 if (rsize >= dma_len) { in srpt_map_sg_to_ib_sge() 1151 tsize -= dma_len; in srpt_map_sg_to_ib_sge() 1152 rsize -= dma_len; in srpt_map_sg_to_ib_sge() 1153 raddr += dma_len; in srpt_map_sg_to_ib_sge() 1159 dma_len = ib_sg_dma_len( in srpt_map_sg_to_ib_sge() 1165 dma_len -= rsize; in srpt_map_sg_to_ib_sge() 1197 dma_len = ib_sg_dma_len(dev, &sg[0]); in srpt_map_sg_to_ib_sge() 1211 if (rsize >= dma_len) { in srpt_map_sg_to_ib_sge() [all …]
|
/drivers/media/platform/soc_camera/ |
D | pxa_camera.c | 293 int i, offset, dma_len, xfer_len; in calculate_dma_sglen() local 298 dma_len = sg_dma_len(sg); in calculate_dma_sglen() 301 xfer_len = roundup(min(dma_len - offset, size), 8); in calculate_dma_sglen() 339 int dma_len = 0, xfer_len = 0; in pxa_init_dma_channel() local 362 dma_len = sg_dma_len(sg); in pxa_init_dma_channel() 365 xfer_len = roundup(min(dma_len - offset, size), 8); in pxa_init_dma_channel() 399 if (xfer_len >= dma_len) { in pxa_init_dma_channel() 400 *sg_first_ofs = xfer_len - dma_len; in pxa_init_dma_channel()
|
/drivers/net/wireless/p54/ |
D | p54pci.h | 62 __le32 dma_len; member
|
/drivers/infiniband/ulp/iser/ |
D | iser_memory.c | 181 unsigned int dma_len; in iser_sg_to_page_vec() local 193 dma_len = ib_sg_dma_len(ibdev, sg); in iser_sg_to_page_vec() 194 end_addr = start_addr + dma_len; in iser_sg_to_page_vec() 195 total_sz += dma_len; in iser_sg_to_page_vec()
|
/drivers/net/wireless/ath/ath9k/ |
D | init.c | 228 u32 dma_len; in ath_descdma_setup() local 231 dma_len = ndesc_skipped * desc_len; in ath_descdma_setup() 232 dd->dd_desc_len += dma_len; in ath_descdma_setup() 234 ndesc_skipped = ATH_DESC_4KB_BOUND_NUM_SKIPPED(dma_len); in ath_descdma_setup()
|