/drivers/mmc/host/ |
D | goldfish.c | 133 unsigned int sg_len; member 234 dma_unmap_sg(mmc_dev(host->mmc), data->sg, host->sg_len, dma_data_dir); in goldfish_mmc_xfer_done() 238 host->sg_len = 0; in goldfish_mmc_xfer_done() 364 unsigned sg_len; in goldfish_mmc_prepare_data() local 383 sg_len = (data->blocks == 1) ? 1 : data->sg_len; in goldfish_mmc_prepare_data() 390 host->sg_len = dma_map_sg(mmc_dev(host->mmc), data->sg, in goldfish_mmc_prepare_data() 391 sg_len, dma_data_dir); in goldfish_mmc_prepare_data() 396 host->sg_len = dma_map_sg(NULL, data->sg, in goldfish_mmc_prepare_data() 397 sg_len, dma_data_dir); in goldfish_mmc_prepare_data()
|
D | au1xmmc.c | 339 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, host->dma.dir); in au1xmmc_data_complete() 384 int sg_len, max, count; in au1xmmc_send_pio() local 399 sg_len = data->sg[host->pio.index].length - host->pio.offset; in au1xmmc_send_pio() 402 max = (sg_len > host->pio.len) ? host->pio.len : sg_len; in au1xmmc_send_pio() 421 if (count == sg_len) { in au1xmmc_send_pio() 439 int max, count, sg_len = 0; in au1xmmc_receive_pio() local 456 sg_len = sg_dma_len(&data->sg[host->pio.index]) - host->pio.offset; in au1xmmc_receive_pio() 459 if (sg_len < max) in au1xmmc_receive_pio() 460 max = sg_len; in au1xmmc_receive_pio() 498 if (sg_len && count == sg_len) { in au1xmmc_receive_pio() [all …]
|
D | mmci.h | 170 unsigned int sg_len; member 183 host->sg_len = data->sg_len; in mmci_init_sg() 192 return --host->sg_len; in mmci_next_sg()
|
D | tmio_mmc.h | 119 unsigned int sg_len; member 129 host->sg_len = data->sg_len; in tmio_mmc_init_sg() 138 return --host->sg_len; in tmio_mmc_next_sg()
|
D | tifm_sd.c | 105 int sg_len; member 174 if (host->sg_pos == host->sg_len) in tifm_sd_transfer_data() 181 if (host->sg_pos == host->sg_len) { in tifm_sd_transfer_data() 237 if (host->sg_pos == host->sg_len) in tifm_sd_bounce_block() 270 if (host->sg_pos == host->sg_len) in tifm_sd_set_dma_data() 278 if (host->sg_pos == host->sg_len) in tifm_sd_set_dma_data() 286 if (host->sg_pos == host->sg_len) in tifm_sd_set_dma_data() 665 host->sg_len = r_data->sg_len; in tifm_sd_request() 679 host->sg_len = tifm_map_sg(sock, r_data->sg, in tifm_sd_request() 680 r_data->sg_len, in tifm_sd_request() [all …]
|
D | omap.c | 145 unsigned int sg_len; member 416 dma_unmap_sg(mmc_dev(host->mmc), data->sg, host->sg_len, in mmc_omap_release_dma() 441 host->sg_len = 0; in mmc_omap_xfer_done() 498 host->sg_len = 0; in mmc_omap_abort_xfer() 673 BUG_ON(host->sg_idx == host->sg_len); in mmc_omap_xfer_data() 996 if (host->sg_idx < host->sg_len) { in mmc_omap_dma_cb() 1082 unsigned sg_len; in mmc_omap_prepare_data() local 1103 sg_len = (data->blocks == 1) ? 1 : data->sg_len; in mmc_omap_prepare_data() 1108 for (i = 0; i < sg_len; i++) { in mmc_omap_prepare_data() 1126 host->sg_len = dma_map_sg(mmc_dev(host->mmc), data->sg, in mmc_omap_prepare_data() [all …]
|
D | sdhci.c | 359 data->sg, data->sg_len, direction); in sdhci_adma_table_pre() 464 data->sg_len, direction); in sdhci_adma_table_pre() 496 data->sg_len, direction); in sdhci_adma_table_post() 515 data->sg_len, direction); in sdhci_adma_table_post() 607 for_each_sg(data->sg, sg, data->sg_len, i) { in sdhci_prepare_data() 642 for_each_sg(data->sg, sg, data->sg_len, i) { in sdhci_prepare_data() 671 data->sg, data->sg_len, in sdhci_prepare_data() 708 data->sg, data->sg_len, SG_MITER_ATOMIC); in sdhci_prepare_data() 753 data->sg_len, (data->flags & MMC_DATA_READ) ? in sdhci_finish_data()
|
D | mxcmmc.c | 183 for_each_sg(data->sg, sg, data->sg_len, i) { in mxcmci_setup_data() 193 data->sg_len, host->dma_dir); in mxcmci_setup_data() 201 data->sg_len, host->dma_dir); in mxcmci_setup_data() 429 for_each_sg(data->sg, sg, data->sg_len, i) { in mxcmci_transfer_data() 436 for_each_sg(data->sg, sg, data->sg_len, i) { in mxcmci_transfer_data()
|
D | imxmmc.c | 273 data->sg_len, host->dma_dir); in imxmci_setup_data() 275 imx_dma_setup_sg(host->dma, data->sg, data->sg_len, datasz, in imxmci_setup_data() 285 data->sg_len, host->dma_dir); in imxmci_setup_data() 287 imx_dma_setup_sg(host->dma, data->sg, data->sg_len, datasz, in imxmci_setup_data()
|
D | s3cmci.c | 206 if (host->pio_sgptr >= host->mrq->data->sg_len) { in get_data_buffer() 208 host->pio_sgptr, host->mrq->data->sg_len); in get_data_buffer() 219 host->pio_sgptr, host->mrq->data->sg_len); in get_data_buffer() 959 dma_len = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, in s3cmci_prepare_dma()
|
/drivers/mmc/card/ |
D | queue.c | 308 unsigned int sg_len; in mmc_queue_map_sg() local 318 sg_len = blk_rq_map_sg(mq->queue, mq->req, mq->bounce_sg); in mmc_queue_map_sg() 320 mq->bounce_sg_len = sg_len; in mmc_queue_map_sg() 323 for_each_sg(mq->bounce_sg, sg, sg_len, i) in mmc_queue_map_sg()
|
D | mmc_test.c | 63 struct mmc_request *mrq, struct scatterlist *sg, unsigned sg_len, in mmc_test_prepare_mrq() argument 91 mrq->data->sg_len = sg_len; in mmc_test_prepare_mrq() 318 struct scatterlist *sg, unsigned sg_len, unsigned dev_addr, in mmc_test_simple_transfer() argument 335 mmc_test_prepare_mrq(test, &mrq, sg, sg_len, dev_addr, in mmc_test_simple_transfer() 385 struct scatterlist *sg, unsigned sg_len, unsigned dev_addr, in mmc_test_transfer() argument 398 sg_copy_from_buffer(sg, sg_len, test->scratch, BUFFER_SIZE); in mmc_test_transfer() 405 ret = mmc_test_simple_transfer(test, sg, sg_len, dev_addr, in mmc_test_transfer() 445 sg_copy_to_buffer(sg, sg_len, test->scratch, BUFFER_SIZE); in mmc_test_transfer()
|
D | block.c | 196 data.sg_len = 1; in mmc_sd_num_wr_blocks() 340 brq.data.sg_len = mmc_queue_map_sg(mq); in mmc_blk_issue_rq() 350 for_each_sg(brq.data.sg, sg, brq.data.sg_len, i) { in mmc_blk_issue_rq() 358 brq.data.sg_len = i; in mmc_blk_issue_rq()
|
/drivers/ata/ |
D | sata_fsl.c | 333 u32 sg_len = sg_dma_len(sg); in sata_fsl_fill_sg() local 336 sg_addr, sg_len); in sata_fsl_fill_sg() 342 if (sg_len & 0x03) in sata_fsl_fill_sg() 344 "s/g len unaligned : 0x%x\n", sg_len); in sata_fsl_fill_sg() 356 ttl_dwords += sg_len; in sata_fsl_fill_sg() 359 cpu_to_le32(DATA_SNOOP_ENABLE | (sg_len & ~0x03)); in sata_fsl_fill_sg() 367 indirect_ext_segment_sz += sg_len; in sata_fsl_fill_sg()
|
D | libata-sff.c | 110 u32 sg_len, len; in ata_fill_sg() local 117 sg_len = sg_dma_len(sg); in ata_fill_sg() 119 while (sg_len) { in ata_fill_sg() 121 len = sg_len; in ata_fill_sg() 122 if ((offset + sg_len) > 0x10000) in ata_fill_sg() 130 sg_len -= len; in ata_fill_sg() 160 u32 sg_len, len, blen; in ata_fill_sg_dumb() local 167 sg_len = sg_dma_len(sg); in ata_fill_sg_dumb() 169 while (sg_len) { in ata_fill_sg_dumb() 171 len = sg_len; in ata_fill_sg_dumb() [all …]
|
D | sata_promise.c | 515 u32 sg_len; in pdc_fill_sg() local 522 sg_len = sg_dma_len(sg); in pdc_fill_sg() 524 while (sg_len) { in pdc_fill_sg() 526 len = sg_len; in pdc_fill_sg() 527 if ((offset + sg_len) > 0x10000) in pdc_fill_sg() 535 sg_len -= len; in pdc_fill_sg()
|
D | sata_sil.c | 323 u32 sg_len = sg_dma_len(sg); in sil_fill_sg() local 326 prd->flags_len = cpu_to_le32(sg_len); in sil_fill_sg() 327 VPRINTK("PRD[%u] = (0x%X, 0x%X)\n", si, addr, sg_len); in sil_fill_sg()
|
/drivers/dma/ |
D | dw_dmac.c | 559 unsigned int sg_len, enum dma_data_direction direction, in dwc_prep_slave_sg() argument 576 if (unlikely(!dws || !sg_len)) in dwc_prep_slave_sg() 582 sg_len = dma_map_sg(chan2parent(chan), sgl, sg_len, direction); in dwc_prep_slave_sg() 592 for_each_sg(sgl, sg, sg_len, i) { in dwc_prep_slave_sg() 638 for_each_sg(sgl, sg, sg_len, i) { in dwc_prep_slave_sg()
|
/drivers/mmc/core/ |
D | sd_ops.c | 283 data.sg_len = 1; in mmc_app_send_scr() 335 data.sg_len = 1; in mmc_sd_switch()
|
D | sdio_ops.c | 150 data.sg_len = 1; in mmc_io_rw_extended()
|
/drivers/scsi/ |
D | gdth_ioctl.h | 46 ulong32 sg_len; /* length */ member 52 ulong32 sg_len; /* length */ member
|
D | scsi_lib.c | 2720 size_t sg_len = 0, len_complete = 0; in scsi_kmap_atomic_sg() local 2727 len_complete = sg_len; /* Complete sg-entries */ in scsi_kmap_atomic_sg() 2728 sg_len += sg->length; in scsi_kmap_atomic_sg() 2729 if (sg_len > *offset) in scsi_kmap_atomic_sg() 2736 __func__, sg_len, *offset, sg_count); in scsi_kmap_atomic_sg() 2749 sg_len = PAGE_SIZE - *offset; in scsi_kmap_atomic_sg() 2750 if (*len > sg_len) in scsi_kmap_atomic_sg() 2751 *len = sg_len; in scsi_kmap_atomic_sg()
|
D | a100u2w.c | 647 scb->sg_len = cpu_to_le32(0); in orc_device_reset() 862 scb->sg_len = cpu_to_le32(0); in inia100_build_scb() 874 scb->sg_len = cpu_to_le32((u32) (count_sg * 8)); in inia100_build_scb() 881 scb->sg_len = cpu_to_le32(0); in inia100_build_scb()
|
D | a100u2w.h | 154 u32 sg_len; /*10 SG list # * 8 */ member
|
D | gdth.c | 2541 cmdp->u.cache64.sg_lst[i].sg_len = sg_dma_len(sl); in gdth_fill_cache_cmd() 2553 cmdp->u.cache.sg_lst[i].sg_len = sg_dma_len(sl); in gdth_fill_cache_cmd() 2571 cmdp->u.cache64.sg_lst[0].sg_len)); in gdth_fill_cache_cmd() 2580 cmdp->u.cache.sg_lst[0].sg_len)); in gdth_fill_cache_cmd() 2716 cmdp->u.raw64.sg_lst[i].sg_len = sg_dma_len(sl); in gdth_fill_raw_cmd() 2728 cmdp->u.raw.sg_lst[i].sg_len = sg_dma_len(sl); in gdth_fill_raw_cmd() 2744 cmdp->u.raw64.sg_lst[0].sg_len)); in gdth_fill_raw_cmd() 2752 cmdp->u.raw.sg_lst[0].sg_len)); in gdth_fill_raw_cmd() 4200 gen.command.u.cache64.sg_lst[0].sg_len = gen.data_len; in ioc_general() 4201 gen.command.u.cache64.sg_lst[1].sg_len = 0; in ioc_general() [all …]
|