/drivers/crypto/cavium/cpt/ |
D | cptvf_reqmanager.c | 51 struct sglist_component *sg_ptr = NULL; in setup_sgio_components() local 76 sg_ptr = (struct sglist_component *)buffer; in setup_sgio_components() 78 sg_ptr->u.s.len0 = cpu_to_be16(list[i * 4 + 0].size); in setup_sgio_components() 79 sg_ptr->u.s.len1 = cpu_to_be16(list[i * 4 + 1].size); in setup_sgio_components() 80 sg_ptr->u.s.len2 = cpu_to_be16(list[i * 4 + 2].size); in setup_sgio_components() 81 sg_ptr->u.s.len3 = cpu_to_be16(list[i * 4 + 3].size); in setup_sgio_components() 82 sg_ptr->ptr0 = cpu_to_be64(list[i * 4 + 0].dma_addr); in setup_sgio_components() 83 sg_ptr->ptr1 = cpu_to_be64(list[i * 4 + 1].dma_addr); in setup_sgio_components() 84 sg_ptr->ptr2 = cpu_to_be64(list[i * 4 + 2].dma_addr); in setup_sgio_components() 85 sg_ptr->ptr3 = cpu_to_be64(list[i * 4 + 3].dma_addr); in setup_sgio_components() [all …]
|
/drivers/staging/rts5208/ |
D | rtsx_transport.c | 339 struct scatterlist *sg_ptr; in rtsx_transfer_sglist_adma_partial() local 374 sg_ptr = sg; in rtsx_transfer_sglist_adma_partial() 383 sg_ptr = sg_next(sg_ptr); in rtsx_transfer_sglist_adma_partial() 389 addr = sg_dma_address(sg_ptr); in rtsx_transfer_sglist_adma_partial() 390 len = sg_dma_len(sg_ptr); in rtsx_transfer_sglist_adma_partial() 419 sg_ptr = sg_next(sg_ptr); in rtsx_transfer_sglist_adma_partial() 504 struct scatterlist *sg_ptr; in rtsx_transfer_sglist_adma() local 537 sg_ptr = sg; in rtsx_transfer_sglist_adma() 550 dma_addr_t addr = sg_dma_address(sg_ptr); in rtsx_transfer_sglist_adma() 551 unsigned int len = sg_dma_len(sg_ptr); in rtsx_transfer_sglist_adma() [all …]
|
/drivers/mmc/host/ |
D | renesas_sdhi_sys_dmac.c | 153 host->sg_ptr, host->sg_len, in renesas_sdhi_sys_dmac_dma_callback() 157 host->sg_ptr, host->sg_len, in renesas_sdhi_sys_dmac_dma_callback() 172 struct scatterlist *sg = host->sg_ptr, *sg_tmp; in renesas_sdhi_sys_dmac_start_dma_rx() 205 host->sg_ptr = &host->bounce_sg; in renesas_sdhi_sys_dmac_start_dma_rx() 206 sg = host->sg_ptr; in renesas_sdhi_sys_dmac_start_dma_rx() 246 struct scatterlist *sg = host->sg_ptr, *sg_tmp; in renesas_sdhi_sys_dmac_start_dma_tx() 284 host->sg_ptr = &host->bounce_sg; in renesas_sdhi_sys_dmac_start_dma_tx() 285 sg = host->sg_ptr; in renesas_sdhi_sys_dmac_start_dma_tx()
|
D | tmio_mmc_core.c | 120 host->sg_ptr = data->sg; in tmio_mmc_init_sg() 127 host->sg_ptr = sg_next(host->sg_ptr); in tmio_mmc_next_sg() 449 sg_virt = tmio_mmc_kmap_atomic(host->sg_ptr, &flags); in tmio_mmc_pio_irq() 452 count = host->sg_ptr->length - host->sg_off; in tmio_mmc_pio_irq() 464 tmio_mmc_kunmap_atomic(host->sg_ptr, &flags, sg_virt); in tmio_mmc_pio_irq() 466 if (host->sg_off == host->sg_ptr->length) in tmio_mmc_pio_irq() 472 if (host->sg_ptr == &host->bounce_sg) { in tmio_mmc_check_bounce_buffer()
|
D | renesas_sdhi_internal_dmac.c | 139 struct scatterlist *sg = host->sg_ptr; in renesas_sdhi_internal_dmac_start_dma() 211 dma_unmap_sg(&host->pdev->dev, host->sg_ptr, host->sg_len, dir); in renesas_sdhi_internal_dmac_complete_tasklet_fn()
|
D | au1xmmc.c | 397 unsigned char *sg_ptr, val; in au1xmmc_send_pio() local 408 sg_ptr = sg_virt(sg) + host->pio.offset; in au1xmmc_send_pio() 424 val = *sg_ptr++; in au1xmmc_send_pio() 452 unsigned char *sg_ptr = NULL; in au1xmmc_receive_pio() local 465 sg_ptr = sg_virt(sg) + host->pio.offset; in au1xmmc_receive_pio() 503 if (sg_ptr) in au1xmmc_receive_pio() 504 *sg_ptr++ = (unsigned char)(val & 0xFF); in au1xmmc_receive_pio()
|
D | tmio_mmc.h | 143 struct scatterlist *sg_ptr; member
|
/drivers/mmc/core/ |
D | sdio_ops.c | 124 struct scatterlist sg, *sg_ptr; in mmc_io_rw_extended() local 163 for_each_sg(data.sg, sg_ptr, data.sg_len, i) { in mmc_io_rw_extended() 164 sg_set_buf(sg_ptr, buf + i * seg_size, in mmc_io_rw_extended()
|
/drivers/scsi/ |
D | gdth_ioctl.h | 38 u32 sg_ptr; /* address */ member 44 u64 sg_ptr; /* address */ member
|
D | gdth.c | 2528 cmdp->u.cache64.sg_lst[i].sg_ptr = sg_dma_address(sl); in gdth_fill_cache_cmd() 2530 if (cmdp->u.cache64.sg_lst[i].sg_ptr > (u64)0xffffffff) in gdth_fill_cache_cmd() 2543 cmdp->u.cache.sg_lst[i].sg_ptr = sg_dma_address(sl); in gdth_fill_cache_cmd() 2564 cmdp->u.cache64.sg_lst[0].sg_ptr, in gdth_fill_cache_cmd() 2573 cmdp->u.cache.sg_lst[0].sg_ptr, in gdth_fill_cache_cmd() 2703 cmdp->u.raw64.sg_lst[i].sg_ptr = sg_dma_address(sl); in gdth_fill_raw_cmd() 2705 if (cmdp->u.raw64.sg_lst[i].sg_ptr > (u64)0xffffffff) in gdth_fill_raw_cmd() 2718 cmdp->u.raw.sg_lst[i].sg_ptr = sg_dma_address(sl); in gdth_fill_raw_cmd() 2737 cmdp->u.raw64.sg_lst[0].sg_ptr, in gdth_fill_raw_cmd() 2745 cmdp->u.raw.sg_lst[0].sg_ptr, in gdth_fill_raw_cmd() [all …]
|
D | FlashPoint.c | 5140 u32 sg_ptr; in FPT_hostDataXferAbort() local 5199 sg_ptr = pCurrSCCB->Sccb_sgseg + SG_BUF_CNT; in FPT_hostDataXferAbort() 5201 if (sg_ptr > in FPT_hostDataXferAbort() 5205 sg_ptr = (u32)(pCurrSCCB->DataLength / in FPT_hostDataXferAbort() 5213 sg_ptr--; in FPT_hostDataXferAbort() 5215 DataPointer) + (sg_ptr * 2); in FPT_hostDataXferAbort() 5227 pCurrSCCB->Sccb_sgseg = (unsigned short)sg_ptr; in FPT_hostDataXferAbort() 5229 if ((unsigned long)(sg_ptr * SG_ELEMENT_SIZE) == in FPT_hostDataXferAbort()
|
D | advansys.c | 1733 __le32 sg_ptr; /* Pointer to next sg block. */ member 2562 b->sg_cnt, (u32)le32_to_cpu(b->sg_ptr)); in asc_prt_adv_sgblock() 2564 if (b->sg_ptr != 0) in asc_prt_adv_sgblock() 2581 struct adv_sg_block *sg_ptr; in asc_prt_adv_scsi_req_q() local 2612 sg_ptr = &sgblkp->sg_block; in asc_prt_adv_scsi_req_q() 2613 asc_prt_adv_sgblock(sg_blk_cnt, sg_ptr); in asc_prt_adv_scsi_req_q() 2614 if (sg_ptr->sg_ptr == 0) { in asc_prt_adv_scsi_req_q() 7700 prev_sg_block->sg_ptr = cpu_to_le32(sgblk_paddr); in adv_get_sglist() 7716 sg_block->sg_ptr = 0L; /* Last ADV_SG_BLOCK in list. */ in adv_get_sglist()
|
/drivers/vhost/ |
D | scsi.c | 788 struct scatterlist *sg_ptr, *sg_prot_ptr = NULL; in vhost_scsi_submission_work() local 793 sg_ptr = cmd->tvc_sgl; in vhost_scsi_submission_work() 800 sg_ptr = NULL; in vhost_scsi_submission_work() 810 sg_ptr, cmd->tvc_sgl_count, NULL, 0, sg_prot_ptr, in vhost_scsi_submission_work()
|