/drivers/infiniband/sw/rdmavt/ |
D | trace_mr.h | 186 TP_PROTO(struct ib_mr *ibmr, int sg_nents, unsigned int *sg_offset), 187 TP_ARGS(ibmr, sg_nents, sg_offset), 195 __field(uint, sg_offset) 204 __entry->sg_offset = sg_offset ? *sg_offset : 0; 214 __entry->sg_offset
|
D | mr.h | 85 int sg_nents, unsigned int *sg_offset);
|
D | mr.c | 633 int sg_nents, unsigned int *sg_offset) in rvt_map_mr_sg() argument 640 ret = ib_sg_to_pages(ibmr, sg, sg_nents, sg_offset, rvt_set_page); in rvt_map_mr_sg() 645 trace_rvt_map_mr_sg(ibmr, sg_nents, sg_offset); in rvt_map_mr_sg()
|
/drivers/infiniband/hw/mlx5/ |
D | mr.c | 2034 unsigned int sg_offset = 0; in mlx5_ib_map_pa_mr_sg_pi() local 2042 sg_offset = *data_sg_offset; in mlx5_ib_map_pa_mr_sg_pi() 2043 mr->data_length = sg_dma_len(data_sg) - sg_offset; in mlx5_ib_map_pa_mr_sg_pi() 2044 mr->data_iova = sg_dma_address(data_sg) + sg_offset; in mlx5_ib_map_pa_mr_sg_pi() 2049 sg_offset = *meta_sg_offset; in mlx5_ib_map_pa_mr_sg_pi() 2051 sg_offset = 0; in mlx5_ib_map_pa_mr_sg_pi() 2052 mr->meta_length = sg_dma_len(meta_sg) - sg_offset; in mlx5_ib_map_pa_mr_sg_pi() 2053 mr->pi_iova = sg_dma_address(meta_sg) + sg_offset; in mlx5_ib_map_pa_mr_sg_pi() 2072 unsigned int sg_offset = sg_offset_p ? *sg_offset_p : 0; in mlx5_ib_sg_to_klms() local 2076 mr->ibmr.iova = sg_dma_address(sg) + sg_offset; in mlx5_ib_sg_to_klms() [all …]
|
/drivers/infiniband/core/ |
D | rw.c | 287 struct scatterlist *sg, u32 sg_cnt, u32 sg_offset, in rdma_rw_ctx_init() argument 308 if (sg_offset < len) in rdma_rw_ctx_init() 312 sg_offset -= len; in rdma_rw_ctx_init() 322 sg_offset, remote_addr, rkey, dir); in rdma_rw_ctx_init() 324 ret = rdma_rw_init_map_wrs(ctx, qp, sg, sg_cnt, sg_offset, in rdma_rw_ctx_init() 327 ret = rdma_rw_init_single_wr(ctx, qp, sg, sg_offset, in rdma_rw_ctx_init()
|
D | verbs.c | 2525 unsigned int *sg_offset, unsigned int page_size) in ib_map_mr_sg() argument 2532 return mr->device->ops.map_mr_sg(mr, sg, sg_nents, sg_offset); in ib_map_mr_sg() 2561 unsigned int sg_offset = sg_offset_p ? *sg_offset_p : 0; in ib_sg_to_pages() local 2566 if (unlikely(sg_nents <= 0 || sg_offset > sg_dma_len(&sgl[0]))) in ib_sg_to_pages() 2569 mr->iova = sg_dma_address(&sgl[0]) + sg_offset; in ib_sg_to_pages() 2573 u64 dma_addr = sg_dma_address(sg) + sg_offset; in ib_sg_to_pages() 2575 unsigned int dma_len = sg_dma_len(sg) - sg_offset; in ib_sg_to_pages() 2600 sg_offset = prev_addr - sg_dma_address(sg); in ib_sg_to_pages() 2603 *sg_offset_p = sg_offset; in ib_sg_to_pages() 2604 return i || sg_offset ? i : ret; in ib_sg_to_pages() [all …]
|
/drivers/usb/storage/ |
D | jumpshot.c | 163 unsigned int sg_offset = 0; in jumpshot_read_data() local 215 &sg, &sg_offset, TO_XFER_BUF); in jumpshot_read_data() 240 unsigned int sg_offset = 0; in jumpshot_write_data() local 271 &sg, &sg_offset, FROM_XFER_BUF); in jumpshot_write_data()
|
D | datafab.c | 145 unsigned int sg_offset = 0; in datafab_read_data() local 203 &sg, &sg_offset, TO_XFER_BUF); in datafab_read_data() 229 unsigned int sg_offset = 0; in datafab_write_data() local 266 &sg, &sg_offset, FROM_XFER_BUF); in datafab_write_data()
|
D | shuttle_usbat.c | 1126 unsigned int sg_offset = 0; in usbat_flash_read_data() local 1181 &sg, &sg_offset, TO_XFER_BUF); in usbat_flash_read_data() 1217 unsigned int sg_offset = 0; in usbat_flash_write_data() local 1257 &sg, &sg_offset, FROM_XFER_BUF); in usbat_flash_write_data() 1297 unsigned int sg_offset = 0; in usbat_hp8200e_handle_read10() local 1386 &sg, &sg_offset, TO_XFER_BUF); in usbat_hp8200e_handle_read10()
|
/drivers/infiniband/hw/vmw_pvrdma/ |
D | pvrdma_mr.c | 313 unsigned int *sg_offset) in pvrdma_map_mr_sg() argument 321 ret = ib_sg_to_pages(ibmr, sg, sg_nents, sg_offset, pvrdma_set_page); in pvrdma_map_mr_sg()
|
D | pvrdma_verbs.h | 411 int sg_nents, unsigned int *sg_offset);
|
/drivers/target/iscsi/cxgbit/ |
D | cxgbit_ddp.c | 178 unsigned int sg_offset = sgl->offset; in cxgbit_ddp_reserve() local 205 sgl->offset = sg_offset; in cxgbit_ddp_reserve()
|
/drivers/infiniband/hw/ocrdma/ |
D | ocrdma_verbs.h | 108 unsigned int *sg_offset);
|
/drivers/infiniband/hw/qedr/ |
D | verbs.h | 85 int sg_nents, unsigned int *sg_offset);
|
/drivers/scsi/ |
D | dpt_i2o.c | 1665 u32 sg_offset = 0; in adpt_i2o_passthru() local 1702 sg_offset = (msg[0]>>4)&0xf; in adpt_i2o_passthru() 1715 if(sg_offset) { in adpt_i2o_passthru() 1717 struct sg_simple_element *sg = (struct sg_simple_element*) (msg+sg_offset); in adpt_i2o_passthru() 1718 sg_count = (size - sg_offset*4) / sizeof(struct sg_simple_element); in adpt_i2o_passthru() 1780 if(sg_offset) { in adpt_i2o_passthru() 1805 sg_count = (size - sg_offset*4) / sizeof(struct sg_simple_element); in adpt_i2o_passthru() 1808 sg = (struct sg_simple_element*)(msg + sg_offset); in adpt_i2o_passthru() 1840 (struct sg_simple_element*) (msg +sg_offset); in adpt_i2o_passthru()
|
D | libiscsi_tcp.c | 94 segment->sg_offset = offset; in iscsi_tcp_segment_init_sg() 143 segment->data = segment->sg_mapped + sg->offset + segment->sg_offset; in iscsi_tcp_segment_map() 210 segment->copied + segment->sg_offset + in iscsi_tcp_segment_done()
|
/drivers/crypto/ccp/ |
D | ccp-ops.c | 177 struct scatterlist *sg, unsigned int sg_offset, in ccp_set_dm_area() argument 185 scatterwalk_map_and_copy(wa->address + wa_offset, sg, sg_offset, len, in ccp_set_dm_area() 191 struct scatterlist *sg, unsigned int sg_offset, in ccp_get_dm_area() argument 196 scatterwalk_map_and_copy(wa->address + wa_offset, sg, sg_offset, len, in ccp_get_dm_area() 203 unsigned int sg_offset, in ccp_reverse_set_dm_area() argument 209 rc = ccp_set_dm_area(wa, wa_offset, sg, sg_offset, len); in ccp_reverse_set_dm_area() 228 unsigned int sg_offset, in ccp_reverse_get_dm_area() argument 243 ccp_get_dm_area(wa, wa_offset, sg, sg_offset, len); in ccp_reverse_get_dm_area()
|
/drivers/infiniband/hw/bnxt_re/ |
D | ib_verbs.h | 201 unsigned int *sg_offset);
|
/drivers/infiniband/hw/mlx4/ |
D | mr.c | 807 unsigned int *sg_offset) in mlx4_ib_map_mr_sg() argument 817 rc = ib_sg_to_pages(ibmr, sg, sg_nents, sg_offset, mlx4_set_page); in mlx4_ib_map_mr_sg()
|
/drivers/nvme/target/ |
D | tcp.c | 285 u32 length, offset, sg_offset; in nvmet_tcp_map_pdu_iovec() local 291 sg_offset = offset % PAGE_SIZE; in nvmet_tcp_map_pdu_iovec() 295 u32 iov_len = min_t(u32, length, sg->length - sg_offset); in nvmet_tcp_map_pdu_iovec() 297 iov->iov_base = kmap(sg_page(sg)) + sg->offset + sg_offset; in nvmet_tcp_map_pdu_iovec()
|
/drivers/infiniband/hw/cxgb4/ |
D | mem.c | 786 unsigned int *sg_offset) in c4iw_map_mr_sg() argument 792 return ib_sg_to_pages(ibmr, sg, sg_nents, sg_offset, c4iw_set_page); in c4iw_map_mr_sg()
|
/drivers/infiniband/sw/rxe/ |
D | rxe_verbs.c | 1027 int sg_nents, unsigned int *sg_offset) in rxe_map_mr_sg() argument 1034 n = ib_sg_to_pages(ibmr, sg, sg_nents, sg_offset, rxe_set_page); in rxe_map_mr_sg()
|
/drivers/scsi/lpfc/ |
D | lpfc_bsg.c | 224 unsigned int sg_offset, dma_offset; in lpfc_bsg_copy_data() local 234 sg_offset = 0; in lpfc_bsg_copy_data() 248 if (sg_offset) { in lpfc_bsg_copy_data() 250 sg_address = miter.addr + sg_offset; in lpfc_bsg_copy_data() 251 transfer_bytes = miter.length - sg_offset; in lpfc_bsg_copy_data() 265 sg_offset += transfer_bytes; in lpfc_bsg_copy_data() 268 if (sg_offset >= miter.length) { in lpfc_bsg_copy_data() 269 sg_offset = 0; in lpfc_bsg_copy_data()
|
/drivers/infiniband/hw/cxgb3/ |
D | iwch_provider.c | 656 int sg_nents, unsigned int *sg_offset) in iwch_map_mr_sg() argument 662 return ib_sg_to_pages(ibmr, sg, sg_nents, sg_offset, iwch_set_page); in iwch_map_mr_sg()
|
/drivers/infiniband/ulp/srp/ |
D | ib_srp.c | 1542 unsigned int sg_offset = sg_offset_p ? *sg_offset_p : 0; in srp_map_finish_fr() local 1544 srp_map_desc(state, sg_dma_address(state->sg) + sg_offset, in srp_map_finish_fr() 1545 sg_dma_len(state->sg) - sg_offset, in srp_map_finish_fr() 1671 unsigned int sg_offset = 0; in srp_map_sg_fr() local 1683 n = srp_map_finish_fr(state, req, ch, count, &sg_offset); in srp_map_sg_fr()
|