Home
last modified time | relevance | path

Searched refs:sg_len (Results 1 – 25 of 41) sorted by relevance

12

/drivers/mmc/host/
Dgoldfish.c133 unsigned int sg_len; member
234 dma_unmap_sg(mmc_dev(host->mmc), data->sg, host->sg_len, dma_data_dir); in goldfish_mmc_xfer_done()
238 host->sg_len = 0; in goldfish_mmc_xfer_done()
364 unsigned sg_len; in goldfish_mmc_prepare_data() local
383 sg_len = (data->blocks == 1) ? 1 : data->sg_len; in goldfish_mmc_prepare_data()
390 host->sg_len = dma_map_sg(mmc_dev(host->mmc), data->sg, in goldfish_mmc_prepare_data()
391 sg_len, dma_data_dir); in goldfish_mmc_prepare_data()
396 host->sg_len = dma_map_sg(NULL, data->sg, in goldfish_mmc_prepare_data()
397 sg_len, dma_data_dir); in goldfish_mmc_prepare_data()
Dau1xmmc.c339 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, host->dma.dir); in au1xmmc_data_complete()
384 int sg_len, max, count; in au1xmmc_send_pio() local
399 sg_len = data->sg[host->pio.index].length - host->pio.offset; in au1xmmc_send_pio()
402 max = (sg_len > host->pio.len) ? host->pio.len : sg_len; in au1xmmc_send_pio()
421 if (count == sg_len) { in au1xmmc_send_pio()
439 int max, count, sg_len = 0; in au1xmmc_receive_pio() local
456 sg_len = sg_dma_len(&data->sg[host->pio.index]) - host->pio.offset; in au1xmmc_receive_pio()
459 if (sg_len < max) in au1xmmc_receive_pio()
460 max = sg_len; in au1xmmc_receive_pio()
498 if (sg_len && count == sg_len) { in au1xmmc_receive_pio()
[all …]
Dmmci.h170 unsigned int sg_len; member
183 host->sg_len = data->sg_len; in mmci_init_sg()
192 return --host->sg_len; in mmci_next_sg()
Dtmio_mmc.h119 unsigned int sg_len; member
129 host->sg_len = data->sg_len; in tmio_mmc_init_sg()
138 return --host->sg_len; in tmio_mmc_next_sg()
Dtifm_sd.c105 int sg_len; member
174 if (host->sg_pos == host->sg_len) in tifm_sd_transfer_data()
181 if (host->sg_pos == host->sg_len) { in tifm_sd_transfer_data()
237 if (host->sg_pos == host->sg_len) in tifm_sd_bounce_block()
270 if (host->sg_pos == host->sg_len) in tifm_sd_set_dma_data()
278 if (host->sg_pos == host->sg_len) in tifm_sd_set_dma_data()
286 if (host->sg_pos == host->sg_len) in tifm_sd_set_dma_data()
665 host->sg_len = r_data->sg_len; in tifm_sd_request()
679 host->sg_len = tifm_map_sg(sock, r_data->sg, in tifm_sd_request()
680 r_data->sg_len, in tifm_sd_request()
[all …]
Domap.c145 unsigned int sg_len; member
416 dma_unmap_sg(mmc_dev(host->mmc), data->sg, host->sg_len, in mmc_omap_release_dma()
441 host->sg_len = 0; in mmc_omap_xfer_done()
498 host->sg_len = 0; in mmc_omap_abort_xfer()
673 BUG_ON(host->sg_idx == host->sg_len); in mmc_omap_xfer_data()
996 if (host->sg_idx < host->sg_len) { in mmc_omap_dma_cb()
1082 unsigned sg_len; in mmc_omap_prepare_data() local
1103 sg_len = (data->blocks == 1) ? 1 : data->sg_len; in mmc_omap_prepare_data()
1108 for (i = 0; i < sg_len; i++) { in mmc_omap_prepare_data()
1126 host->sg_len = dma_map_sg(mmc_dev(host->mmc), data->sg, in mmc_omap_prepare_data()
[all …]
Dsdhci.c359 data->sg, data->sg_len, direction); in sdhci_adma_table_pre()
464 data->sg_len, direction); in sdhci_adma_table_pre()
496 data->sg_len, direction); in sdhci_adma_table_post()
515 data->sg_len, direction); in sdhci_adma_table_post()
607 for_each_sg(data->sg, sg, data->sg_len, i) { in sdhci_prepare_data()
642 for_each_sg(data->sg, sg, data->sg_len, i) { in sdhci_prepare_data()
671 data->sg, data->sg_len, in sdhci_prepare_data()
708 data->sg, data->sg_len, SG_MITER_ATOMIC); in sdhci_prepare_data()
753 data->sg_len, (data->flags & MMC_DATA_READ) ? in sdhci_finish_data()
Dmxcmmc.c183 for_each_sg(data->sg, sg, data->sg_len, i) { in mxcmci_setup_data()
193 data->sg_len, host->dma_dir); in mxcmci_setup_data()
201 data->sg_len, host->dma_dir); in mxcmci_setup_data()
429 for_each_sg(data->sg, sg, data->sg_len, i) { in mxcmci_transfer_data()
436 for_each_sg(data->sg, sg, data->sg_len, i) { in mxcmci_transfer_data()
Dimxmmc.c273 data->sg_len, host->dma_dir); in imxmci_setup_data()
275 imx_dma_setup_sg(host->dma, data->sg, data->sg_len, datasz, in imxmci_setup_data()
285 data->sg_len, host->dma_dir); in imxmci_setup_data()
287 imx_dma_setup_sg(host->dma, data->sg, data->sg_len, datasz, in imxmci_setup_data()
Ds3cmci.c206 if (host->pio_sgptr >= host->mrq->data->sg_len) { in get_data_buffer()
208 host->pio_sgptr, host->mrq->data->sg_len); in get_data_buffer()
219 host->pio_sgptr, host->mrq->data->sg_len); in get_data_buffer()
959 dma_len = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, in s3cmci_prepare_dma()
/drivers/mmc/card/
Dqueue.c308 unsigned int sg_len; in mmc_queue_map_sg() local
318 sg_len = blk_rq_map_sg(mq->queue, mq->req, mq->bounce_sg); in mmc_queue_map_sg()
320 mq->bounce_sg_len = sg_len; in mmc_queue_map_sg()
323 for_each_sg(mq->bounce_sg, sg, sg_len, i) in mmc_queue_map_sg()
Dmmc_test.c63 struct mmc_request *mrq, struct scatterlist *sg, unsigned sg_len, in mmc_test_prepare_mrq() argument
91 mrq->data->sg_len = sg_len; in mmc_test_prepare_mrq()
318 struct scatterlist *sg, unsigned sg_len, unsigned dev_addr, in mmc_test_simple_transfer() argument
335 mmc_test_prepare_mrq(test, &mrq, sg, sg_len, dev_addr, in mmc_test_simple_transfer()
385 struct scatterlist *sg, unsigned sg_len, unsigned dev_addr, in mmc_test_transfer() argument
398 sg_copy_from_buffer(sg, sg_len, test->scratch, BUFFER_SIZE); in mmc_test_transfer()
405 ret = mmc_test_simple_transfer(test, sg, sg_len, dev_addr, in mmc_test_transfer()
445 sg_copy_to_buffer(sg, sg_len, test->scratch, BUFFER_SIZE); in mmc_test_transfer()
Dblock.c196 data.sg_len = 1; in mmc_sd_num_wr_blocks()
340 brq.data.sg_len = mmc_queue_map_sg(mq); in mmc_blk_issue_rq()
350 for_each_sg(brq.data.sg, sg, brq.data.sg_len, i) { in mmc_blk_issue_rq()
358 brq.data.sg_len = i; in mmc_blk_issue_rq()
/drivers/ata/
Dsata_fsl.c333 u32 sg_len = sg_dma_len(sg); in sata_fsl_fill_sg() local
336 sg_addr, sg_len); in sata_fsl_fill_sg()
342 if (sg_len & 0x03) in sata_fsl_fill_sg()
344 "s/g len unaligned : 0x%x\n", sg_len); in sata_fsl_fill_sg()
356 ttl_dwords += sg_len; in sata_fsl_fill_sg()
359 cpu_to_le32(DATA_SNOOP_ENABLE | (sg_len & ~0x03)); in sata_fsl_fill_sg()
367 indirect_ext_segment_sz += sg_len; in sata_fsl_fill_sg()
Dlibata-sff.c110 u32 sg_len, len; in ata_fill_sg() local
117 sg_len = sg_dma_len(sg); in ata_fill_sg()
119 while (sg_len) { in ata_fill_sg()
121 len = sg_len; in ata_fill_sg()
122 if ((offset + sg_len) > 0x10000) in ata_fill_sg()
130 sg_len -= len; in ata_fill_sg()
160 u32 sg_len, len, blen; in ata_fill_sg_dumb() local
167 sg_len = sg_dma_len(sg); in ata_fill_sg_dumb()
169 while (sg_len) { in ata_fill_sg_dumb()
171 len = sg_len; in ata_fill_sg_dumb()
[all …]
Dsata_promise.c515 u32 sg_len; in pdc_fill_sg() local
522 sg_len = sg_dma_len(sg); in pdc_fill_sg()
524 while (sg_len) { in pdc_fill_sg()
526 len = sg_len; in pdc_fill_sg()
527 if ((offset + sg_len) > 0x10000) in pdc_fill_sg()
535 sg_len -= len; in pdc_fill_sg()
Dsata_sil.c323 u32 sg_len = sg_dma_len(sg); in sil_fill_sg() local
326 prd->flags_len = cpu_to_le32(sg_len); in sil_fill_sg()
327 VPRINTK("PRD[%u] = (0x%X, 0x%X)\n", si, addr, sg_len); in sil_fill_sg()
/drivers/dma/
Ddw_dmac.c559 unsigned int sg_len, enum dma_data_direction direction, in dwc_prep_slave_sg() argument
576 if (unlikely(!dws || !sg_len)) in dwc_prep_slave_sg()
582 sg_len = dma_map_sg(chan2parent(chan), sgl, sg_len, direction); in dwc_prep_slave_sg()
592 for_each_sg(sgl, sg, sg_len, i) { in dwc_prep_slave_sg()
638 for_each_sg(sgl, sg, sg_len, i) { in dwc_prep_slave_sg()
/drivers/mmc/core/
Dsd_ops.c283 data.sg_len = 1; in mmc_app_send_scr()
335 data.sg_len = 1; in mmc_sd_switch()
Dsdio_ops.c150 data.sg_len = 1; in mmc_io_rw_extended()
/drivers/scsi/
Dgdth_ioctl.h46 ulong32 sg_len; /* length */ member
52 ulong32 sg_len; /* length */ member
Dscsi_lib.c2720 size_t sg_len = 0, len_complete = 0; in scsi_kmap_atomic_sg() local
2727 len_complete = sg_len; /* Complete sg-entries */ in scsi_kmap_atomic_sg()
2728 sg_len += sg->length; in scsi_kmap_atomic_sg()
2729 if (sg_len > *offset) in scsi_kmap_atomic_sg()
2736 __func__, sg_len, *offset, sg_count); in scsi_kmap_atomic_sg()
2749 sg_len = PAGE_SIZE - *offset; in scsi_kmap_atomic_sg()
2750 if (*len > sg_len) in scsi_kmap_atomic_sg()
2751 *len = sg_len; in scsi_kmap_atomic_sg()
Da100u2w.c647 scb->sg_len = cpu_to_le32(0); in orc_device_reset()
862 scb->sg_len = cpu_to_le32(0); in inia100_build_scb()
874 scb->sg_len = cpu_to_le32((u32) (count_sg * 8)); in inia100_build_scb()
881 scb->sg_len = cpu_to_le32(0); in inia100_build_scb()
Da100u2w.h154 u32 sg_len; /*10 SG list # * 8 */ member
Dgdth.c2541 cmdp->u.cache64.sg_lst[i].sg_len = sg_dma_len(sl); in gdth_fill_cache_cmd()
2553 cmdp->u.cache.sg_lst[i].sg_len = sg_dma_len(sl); in gdth_fill_cache_cmd()
2571 cmdp->u.cache64.sg_lst[0].sg_len)); in gdth_fill_cache_cmd()
2580 cmdp->u.cache.sg_lst[0].sg_len)); in gdth_fill_cache_cmd()
2716 cmdp->u.raw64.sg_lst[i].sg_len = sg_dma_len(sl); in gdth_fill_raw_cmd()
2728 cmdp->u.raw.sg_lst[i].sg_len = sg_dma_len(sl); in gdth_fill_raw_cmd()
2744 cmdp->u.raw64.sg_lst[0].sg_len)); in gdth_fill_raw_cmd()
2752 cmdp->u.raw.sg_lst[0].sg_len)); in gdth_fill_raw_cmd()
4200 gen.command.u.cache64.sg_lst[0].sg_len = gen.data_len; in ioc_general()
4201 gen.command.u.cache64.sg_lst[1].sg_len = 0; in ioc_general()
[all …]

12