/drivers/message/i2o/ |
D | memory.c | 33 u16 sg_count = in i2o_sg_tablesize() local 42 sg_count -= 2; in i2o_sg_tablesize() 43 sg_count /= 3; in i2o_sg_tablesize() 45 sg_count /= 2; in i2o_sg_tablesize() 47 if (c->short_req && (sg_count > 8)) in i2o_sg_tablesize() 48 sg_count = 8; in i2o_sg_tablesize() 50 return sg_count; in i2o_sg_tablesize() 128 int sg_count, enum dma_data_direction direction, u32 ** sg_ptr) in i2o_dma_map_sg() argument 144 sg_count = dma_map_sg(&c->pdev->dev, sg, sg_count, direction); in i2o_dma_map_sg() 145 if (!sg_count) in i2o_dma_map_sg() [all …]
|
D | i2o_config.c | 538 u32 sg_count = 0; in i2o_cfg_passthru32() local 614 sg_count = in i2o_cfg_passthru32() 616 if (sg_count > SG_TABLESIZE) { in i2o_cfg_passthru32() 618 c->name, sg_count); in i2o_cfg_passthru32() 623 for (i = 0; i < sg_count; i++) { in i2o_cfg_passthru32() 641 c->name, sg_size, i, sg_count); in i2o_cfg_passthru32() 700 sg_count = in i2o_cfg_passthru32() 705 for (j = 0; j < sg_count; j++) { in i2o_cfg_passthru32() 789 u32 sg_count = 0; in i2o_cfg_passthru() local 857 sg_count = in i2o_cfg_passthru() [all …]
|
/drivers/spi/ |
D | spi-mxs.c | 224 int sg_count; in mxs_spi_txrx_dma() local 253 for (sg_count = 0; sg_count < sgs; sg_count++) { in mxs_spi_txrx_dma() 257 if ((sg_count + 1 == sgs) && *last) in mxs_spi_txrx_dma() 265 dma_xfer[sg_count].pio[0] = ctrl0; in mxs_spi_txrx_dma() 266 dma_xfer[sg_count].pio[3] = min; in mxs_spi_txrx_dma() 280 sg_init_one(&dma_xfer[sg_count].sg, sg_buf, min); in mxs_spi_txrx_dma() 281 ret = dma_map_sg(ssp->dev, &dma_xfer[sg_count].sg, 1, in mxs_spi_txrx_dma() 289 (struct scatterlist *)dma_xfer[sg_count].pio, in mxs_spi_txrx_dma() 292 sg_count ? DMA_PREP_INTERRUPT : 0); in mxs_spi_txrx_dma() 301 &dma_xfer[sg_count].sg, 1, in mxs_spi_txrx_dma() [all …]
|
/drivers/crypto/caam/ |
D | sg_sw_sec4.h | 33 sg_to_sec4_sg(struct scatterlist *sg, int sg_count, in sg_to_sec4_sg() argument 36 while (sg_count) { in sg_to_sec4_sg() 41 sg_count--; in sg_to_sec4_sg() 50 static inline void sg_to_sec4_sg_last(struct scatterlist *sg, int sg_count, in sg_to_sec4_sg_last() argument 54 sec4_sg_ptr = sg_to_sec4_sg(sg, sg_count, sec4_sg_ptr, offset); in sg_to_sec4_sg_last() 77 static inline int sg_count(struct scatterlist *sg_list, int nbytes, in sg_count() function
|
D | caamalg.c | 1150 assoc_nents = sg_count(req->assoc, req->assoclen, &assoc_chained); in aead_edesc_alloc() 1151 src_nents = sg_count(req->src, req->cryptlen, &src_chained); in aead_edesc_alloc() 1154 dst_nents = sg_count(req->dst, req->cryptlen, &dst_chained); in aead_edesc_alloc() 1332 assoc_nents = sg_count(req->assoc, req->assoclen, &assoc_chained); in aead_giv_edesc_alloc() 1333 src_nents = sg_count(req->src, req->cryptlen, &src_chained); in aead_giv_edesc_alloc() 1336 dst_nents = sg_count(req->dst, req->cryptlen, &dst_chained); in aead_giv_edesc_alloc() 1488 src_nents = sg_count(req->src, req->nbytes, &src_chained); in ablkcipher_edesc_alloc() 1491 dst_nents = sg_count(req->dst, req->nbytes, &dst_chained); in ablkcipher_edesc_alloc()
|
/drivers/crypto/ |
D | talitos.c | 915 static int sg_to_link_tbl(struct scatterlist *sg, int sg_count, in sg_to_link_tbl() argument 918 int n_sg = sg_count; in sg_to_link_tbl() 935 sg_count--; in sg_to_link_tbl() 943 return sg_count; in sg_to_link_tbl() 961 int sg_count, ret; in ipsec_esp() local 979 sg_count = sg_to_link_tbl(areq->assoc, edesc->assoc_nents - 1, in ipsec_esp() 983 tbl_ptr += sg_count - 1; in ipsec_esp() 1018 sg_count = talitos_map_sg(dev, areq->src, edesc->src_nents ? : 1, in ipsec_esp() 1023 if (sg_count == 1) { in ipsec_esp() 1031 sg_count = sg_to_link_tbl(areq->src, sg_count, sg_link_tbl_len, in ipsec_esp() [all …]
|
D | picoxcell_crypto.c | 264 static int sg_count(struct scatterlist *sg_list, int nbytes) in sg_count() function 300 nents = sg_count(payload, nbytes); in spacc_sg_to_ddt() 328 unsigned nents = sg_count(areq->src, areq->cryptlen); in spacc_aead_make_ddts() 348 sg_count(areq->assoc, areq->assoclen), DMA_TO_DEVICE); in spacc_aead_make_ddts() 412 unsigned nents = sg_count(areq->src, areq->cryptlen); in spacc_aead_free_ddts() 417 sg_count(areq->dst, areq->cryptlen), in spacc_aead_free_ddts() 423 sg_count(areq->assoc, areq->assoclen), DMA_TO_DEVICE); in spacc_aead_free_ddts() 435 unsigned nents = sg_count(payload, nbytes); in spacc_free_ddt()
|
D | bfin_crc.c | 101 static int sg_count(struct scatterlist *sg_list) in sg_count() function 162 if (sg_count(req->src) > CRC_MAX_DMA_DESC) { in bfin_crypto_crc_init() 382 nsg = ctx->sg_nents = sg_count(ctx->sg); in bfin_crypto_crc_handle_queue()
|
/drivers/scsi/ |
D | storvsc_drv.c | 432 unsigned int sg_count) in destroy_bounce_buffer() argument 437 for (i = 0; i < sg_count; i++) { in destroy_bounce_buffer() 446 static int do_bounce_buffer(struct scatterlist *sgl, unsigned int sg_count) in do_bounce_buffer() argument 451 if (sg_count < 2) in do_bounce_buffer() 455 for (i = 0; i < sg_count; i++) { in do_bounce_buffer() 460 } else if (i == sg_count - 1) { in do_bounce_buffer() 474 unsigned int sg_count, in create_bounce_buffer() argument 1316 unsigned int sg_count = 0; in storvsc_queuecommand() local 1376 sg_count = scsi_sg_count(scmnd); in storvsc_queuecommand() 1399 sg_count = cmd_request->bounce_sgl_count; in storvsc_queuecommand() [all …]
|
D | qlogicpti.c | 895 int sg_count; in load_cmd() local 898 sg_count = dma_map_sg(&qpti->op->dev, sg, in load_cmd() 903 cmd->segment_cnt = sg_count; in load_cmd() 906 n = sg_count; in load_cmd() 913 sg_count -= 4; in load_cmd() 915 while (sg_count > 0) { in load_cmd() 930 n = sg_count; in load_cmd() 937 sg_count -= n; in load_cmd()
|
D | stex.c | 183 __le16 sg_count; member 286 int sg_count; member 425 ccb->sg_count = nseg; in stex_map_sg() 426 dst->sg_count = cpu_to_le16((u16)nseg); in stex_map_sg() 457 ccb->sg_count = nseg; in stex_ss_map_sg() 458 dst->sg_count = cpu_to_le16((u16)nseg); in stex_ss_map_sg() 534 addr += (hba->ccb[tag].sg_count+4)/11; in stex_ss_send_cmd() 693 hba->ccb[tag].sg_count = 0; in stex_queuecommand_lck() 1753 hba->ccb[tag].sg_count = 0; in stex_hba_stop()
|
D | ips.h | 430 uint8_t sg_count; member 639 uint8_t sg_count; member 654 uint16_t sg_count; member 1105 int sg_count; member
|
D | dc395x.c | 236 u8 sg_count; /* No of HW sg entries for this request */ member 990 srb->sg_count = 0; in build_srb() 1019 srb->sg_count = nseg; in build_srb() 1024 srb->sg_count); in build_srb() 1026 scsi_for_each_sg(cmd, sg, srb->sg_count, i) { in build_srb() 1033 sgp += srb->sg_count - 1; in build_srb() 1212 srb->segment_x, srb->sg_count, srb->sg_index, in dump_register_info() 1939 for (; idx < srb->sg_count; psge++, idx++) in sg_verify_length() 1969 for (idx = srb->sg_index; idx < srb->sg_count; idx++) { in sg_update_list() 2284 srb->sg_count, &offset, &len); in data_in_phase0() [all …]
|
D | aha1542.c | 663 int i, sg_count = scsi_sg_count(SCpnt); in aha1542_queuecommand_lck() local 665 SCpnt->host_scribble = kmalloc(sizeof(*cptr)*sg_count, in aha1542_queuecommand_lck() 673 scsi_for_each_sg(SCpnt, sg, sg_count, i) { in aha1542_queuecommand_lck() 677 any2scsi(ccb[mbo].datalen, sg_count * sizeof(struct chain)); in aha1542_queuecommand_lck()
|
D | dpt_i2o.c | 1728 u32 sg_count = 0; in adpt_i2o_passthru() local 1774 sg_count = (size - sg_offset*4) / sizeof(struct sg_simple_element); in adpt_i2o_passthru() 1775 if (sg_count > pHba->sg_tablesize){ in adpt_i2o_passthru() 1776 printk(KERN_DEBUG"%s:IOCTL SG List too large (%u)\n", pHba->name,sg_count); in adpt_i2o_passthru() 1781 for(i = 0; i < sg_count; i++) { in adpt_i2o_passthru() 1794 pHba->name,sg_size,i,sg_count); in adpt_i2o_passthru() 1859 sg_count = (size - sg_offset*4) / sizeof(struct sg_simple_element); in adpt_i2o_passthru() 1863 for (j = 0; j < sg_count; j++) { in adpt_i2o_passthru()
|
D | mvumi.c | 205 void *sgl_p, unsigned char *sg_count) in mvumi_make_sgl() argument 215 *sg_count = pci_map_sg(mhba->pdev, sg, sgnum, in mvumi_make_sgl() 217 if (*sg_count > mhba->max_sge) { in mvumi_make_sgl() 220 *sg_count, mhba->max_sge); in mvumi_make_sgl() 223 for (i = 0; i < *sg_count; i++) { in mvumi_make_sgl() 229 if ((i + 1) == *sg_count) in mvumi_make_sgl() 245 *sg_count = 1; in mvumi_make_sgl()
|
D | scsi_lib.c | 2574 void *scsi_kmap_atomic_sg(struct scatterlist *sgl, int sg_count, in scsi_kmap_atomic_sg() argument 2584 for_each_sg(sgl, sg, sg_count, i) { in scsi_kmap_atomic_sg() 2591 if (unlikely(i == sg_count)) { in scsi_kmap_atomic_sg() 2594 __func__, sg_len, *offset, sg_count); in scsi_kmap_atomic_sg()
|
D | sr.c | 462 int i, size = 0, sg_count = scsi_sg_count(SCpnt); in sr_prep_fn() local 464 scsi_for_each_sg(SCpnt, sg, sg_count, i) in sr_prep_fn()
|
D | 3w-sas.c | 331 int i, sg_count; in twl_scsiop_execute_scsi() local 373 sg_count = twl_map_scsi_sg_data(tw_dev, request_id); in twl_scsiop_execute_scsi() 374 if (sg_count == 0) in twl_scsiop_execute_scsi() 377 scsi_for_each_sg(srb, sg, sg_count, i) { in twl_scsiop_execute_scsi()
|
/drivers/scsi/arm/ |
D | scsi.h | 101 unsigned i, sg_count = scsi_sg_count(SCpnt); in init_SCp() local 103 scsi_for_each_sg(SCpnt, sg, sg_count, i) in init_SCp()
|
/drivers/scsi/fnic/ |
D | fnic_scsi.c | 301 int sg_count) in fnic_queue_wq_copy_desc() argument 315 if (sg_count) { in fnic_queue_wq_copy_desc() 318 for_each_sg(scsi_sglist(sc), sg, sg_count, i) { in fnic_queue_wq_copy_desc() 328 sizeof(io_req->sgl_list[0]) * sg_count, in fnic_queue_wq_copy_desc() 401 int sg_count = 0; in fnic_queuecommand_lck() local 439 sg_count = scsi_dma_map(sc); in fnic_queuecommand_lck() 440 if (sg_count < 0) { in fnic_queuecommand_lck() 443 sg_count, CMD_STATE(sc)); in fnic_queuecommand_lck() 449 io_req->sgl_cnt = sg_count; in fnic_queuecommand_lck() 451 if (sg_count > FNIC_DFLT_SG_DESC_CNT) in fnic_queuecommand_lck() [all …]
|
/drivers/memstick/host/ |
D | r592.c | 278 int len, sg_count; in r592_transfer_fifo_dma() local 296 sg_count = dma_map_sg(&dev->pci_dev->dev, &dev->req->sg, 1, is_write ? in r592_transfer_fifo_dma() 299 if (sg_count != 1 || in r592_transfer_fifo_dma()
|
/drivers/infiniband/hw/qib/ |
D | qib_sdma.c | 549 if (tx->txreq.sg_count > qib_sdma_descq_freecnt(ppd)) { in qib_sdma_verbs_send() 645 ppd->sdma_descq_added += tx->txreq.sg_count; in qib_sdma_verbs_send()
|
/drivers/infiniband/hw/ipath/ |
D | ipath_sdma.c | 693 if (tx->txreq.sg_count > ipath_sdma_descq_freecnt(dd)) { in ipath_sdma_verbs_send() 798 dd->ipath_sdma_descq_added += tx->txreq.sg_count; in ipath_sdma_verbs_send()
|
/drivers/scsi/aic7xxx/ |
D | aic7xxx_osm.c | 485 if ((scb->sg_count + 1) > AHC_NSEG) in ahc_linux_map_seg() 1542 scb->sg_count = 0; in ahc_linux_run_command() 1565 scb->sg_count += consumed; in ahc_linux_run_command() 1586 scb->sg_count = 0; in ahc_linux_run_command()
|