Home
last modified time | relevance | path

Searched refs:sqe (Results 1 – 25 of 27) sorted by relevance

12

/drivers/infiniband/sw/siw/
Dsiw_qp.c275 wqe->sqe.flags = 0; in siw_qp_mpa_rts()
276 wqe->sqe.num_sge = 1; in siw_qp_mpa_rts()
277 wqe->sqe.sge[0].length = 0; in siw_qp_mpa_rts()
278 wqe->sqe.sge[0].laddr = 0; in siw_qp_mpa_rts()
279 wqe->sqe.sge[0].lkey = 0; in siw_qp_mpa_rts()
284 wqe->sqe.rkey = 1; in siw_qp_mpa_rts()
285 wqe->sqe.raddr = 0; in siw_qp_mpa_rts()
289 wqe->sqe.opcode = SIW_OP_WRITE; in siw_qp_mpa_rts()
293 wqe->sqe.opcode = SIW_OP_READ; in siw_qp_mpa_rts()
299 siw_read_to_orq(rreq, &wqe->sqe); in siw_qp_mpa_rts()
[all …]
Dsiw_qp_tx.c43 struct siw_sge *sge = &wqe->sqe.sge[0]; in siw_try_1seg()
46 if (bytes > MAX_HDR_INLINE || wqe->sqe.num_sge != 1) in siw_try_1seg()
53 memcpy(paddr, &wqe->sqe.sge[1], bytes); in siw_try_1seg()
137 c_tx->pkt.rreq.sink_stag = htonl(wqe->sqe.sge[0].lkey); in siw_qp_prepare_tx()
139 cpu_to_be64(wqe->sqe.sge[0].laddr); in siw_qp_prepare_tx()
140 c_tx->pkt.rreq.source_stag = htonl(wqe->sqe.rkey); in siw_qp_prepare_tx()
141 c_tx->pkt.rreq.source_to = cpu_to_be64(wqe->sqe.raddr); in siw_qp_prepare_tx()
142 c_tx->pkt.rreq.read_size = htonl(wqe->sqe.sge[0].length); in siw_qp_prepare_tx()
185 c_tx->pkt.send_inv.inval_stag = cpu_to_be32(wqe->sqe.rkey); in siw_qp_prepare_tx()
197 c_tx->pkt.rwrite.sink_stag = htonl(wqe->sqe.rkey); in siw_qp_prepare_tx()
[all …]
Dsiw_verbs.c657 struct siw_sqe *sqe) in siw_copy_inline_sgl() argument
660 void *kbuf = &sqe->sge[1]; in siw_copy_inline_sgl()
663 sqe->sge[0].laddr = (uintptr_t)kbuf; in siw_copy_inline_sgl()
664 sqe->sge[0].lkey = 0; in siw_copy_inline_sgl()
682 sqe->sge[0].length = bytes > 0 ? bytes : 0; in siw_copy_inline_sgl()
683 sqe->num_sge = bytes > 0 ? 1 : 0; in siw_copy_inline_sgl()
692 struct siw_sqe sqe = {}; in siw_sq_flush_wr() local
696 sqe.id = wr->wr_id; in siw_sq_flush_wr()
697 sqe.opcode = wr->opcode; in siw_sq_flush_wr()
698 rv = siw_sqe_complete(qp, &sqe, 0, SIW_WC_WR_FLUSH_ERR); in siw_sq_flush_wr()
[all …]
Dsiw.h204 struct siw_sqe sqe; member
496 #define tx_type(wqe) ((wqe)->sqe.opcode)
498 #define tx_flags(wqe) ((wqe)->sqe.flags)
538 void siw_read_to_orq(struct siw_sqe *rreq, struct siw_sqe *sqe);
539 int siw_sqe_complete(struct siw_qp *qp, struct siw_sqe *sqe, u32 bytes,
641 struct siw_sqe *sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in siw_sq_empty() local
643 return READ_ONCE(sqe->flags) == 0; in siw_sq_empty()
648 struct siw_sqe *sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in sq_get_next() local
650 if (READ_ONCE(sqe->flags) & SIW_WQE_VALID) in sq_get_next()
651 return sqe; in sq_get_next()
Dsiw_qp_rx.c175 srx->ddp_stag = wqe->sqe.sge[0].lkey; in siw_rresp_check_ntoh()
176 srx->ddp_to = wqe->sqe.sge[0].laddr; in siw_rresp_check_ntoh()
690 resp = &tx_work->sqe; in siw_init_rresp()
749 wqe->sqe.id = orqe->id; in siw_orqe_start_rx()
750 wqe->sqe.opcode = orqe->opcode; in siw_orqe_start_rx()
751 wqe->sqe.sge[0].laddr = orqe->sge[0].laddr; in siw_orqe_start_rx()
752 wqe->sqe.sge[0].lkey = orqe->sge[0].lkey; in siw_orqe_start_rx()
753 wqe->sqe.sge[0].length = orqe->sge[0].length; in siw_orqe_start_rx()
754 wqe->sqe.flags = orqe->flags; in siw_orqe_start_rx()
755 wqe->sqe.num_sge = 1; in siw_orqe_start_rx()
[all …]
Dsiw_mem.c271 if (!(wqe->sqe.flags & SIW_WQE_INLINE)) in siw_wqe_put_mem()
272 siw_unref_mem_sgl(wqe->mem, wqe->sqe.num_sge); in siw_wqe_put_mem()
/drivers/crypto/hisilicon/zip/
Dzip_crypto.c81 static void hisi_zip_config_buf_type(struct hisi_zip_sqe *sqe, u8 buf_type) in hisi_zip_config_buf_type() argument
85 val = (sqe->dw9) & ~HZIP_BUF_TYPE_M; in hisi_zip_config_buf_type()
87 sqe->dw9 = val; in hisi_zip_config_buf_type()
90 static void hisi_zip_config_tag(struct hisi_zip_sqe *sqe, u32 tag) in hisi_zip_config_tag() argument
92 sqe->tag = tag; in hisi_zip_config_tag()
95 static void hisi_zip_fill_sqe(struct hisi_zip_sqe *sqe, u8 req_type, in hisi_zip_fill_sqe() argument
99 memset(sqe, 0, sizeof(struct hisi_zip_sqe)); in hisi_zip_fill_sqe()
101 sqe->input_data_length = slen; in hisi_zip_fill_sqe()
102 sqe->dw9 = FIELD_PREP(HZIP_REQ_TYPE_M, req_type); in hisi_zip_fill_sqe()
103 sqe->dest_avail_out = dlen; in hisi_zip_fill_sqe()
[all …]
/drivers/scsi/qedf/
Ddrv_fcoe_fw_funcs.c13 memset(task_params->sqe, 0, sizeof(*(task_params->sqe))); in init_common_sqe()
14 SET_FIELD(task_params->sqe->flags, FCOE_WQE_REQ_TYPE, in init_common_sqe()
16 task_params->sqe->task_id = task_params->itid; in init_common_sqe()
167 task_params->sqe->additional_info_union.burst_length = in init_initiator_midpath_unsolicited_fcoe_task()
169 SET_FIELD(task_params->sqe->flags, in init_initiator_midpath_unsolicited_fcoe_task()
171 SET_FIELD(task_params->sqe->flags, FCOE_WQE_SGL_MODE, in init_initiator_midpath_unsolicited_fcoe_task()
193 task_params->sqe->additional_info_union.seq_rec_updated_offset = in init_initiator_sequence_recovery_fcoe_task()
Dqedf_io.c596 struct fcoe_wqe *sqe) in qedf_init_task() argument
632 io_req->task_params->sqe = sqe; in qedf_init_task()
685 struct e4_fcoe_task_context *task_ctx, struct fcoe_wqe *sqe) in qedf_init_mp_task() argument
711 io_req->task_params->sqe = sqe; in qedf_init_mp_task()
864 struct fcoe_wqe *sqe; in qedf_post_io_req() local
913 sqe = &fcport->sq[sqe_idx]; in qedf_post_io_req()
914 memset(sqe, 0, sizeof(struct fcoe_wqe)); in qedf_post_io_req()
927 qedf_init_task(fcport, lport, io_req, task_ctx, sqe); in qedf_post_io_req()
1841 struct fcoe_wqe *sqe; in qedf_initiate_abts() local
1920 sqe = &fcport->sq[sqe_idx]; in qedf_initiate_abts()
[all …]
Dqedf_els.c23 struct fcoe_wqe *sqe; in qedf_initiate_els() local
120 sqe = &fcport->sq[sqe_idx]; in qedf_initiate_els()
121 memset(sqe, 0, sizeof(struct fcoe_wqe)); in qedf_initiate_els()
125 qedf_init_mp_task(els_req, task, sqe); in qedf_initiate_els()
687 struct fcoe_wqe *sqe; in qedf_initiate_seq_cleanup() local
717 sqe = &fcport->sq[sqe_idx]; in qedf_initiate_seq_cleanup()
718 memset(sqe, 0, sizeof(struct fcoe_wqe)); in qedf_initiate_seq_cleanup()
719 orig_io_req->task_params->sqe = sqe; in qedf_initiate_seq_cleanup()
Ddrv_fcoe_fw_funcs.h16 struct fcoe_wqe *sqe; member
Dqedf.h503 struct e4_fcoe_task_context *task_ctx, struct fcoe_wqe *sqe);
/drivers/scsi/qedi/
Dqedi_fw_api.c98 if (!task_params->sqe) in init_sqe()
101 memset(task_params->sqe, 0, sizeof(*task_params->sqe)); in init_sqe()
102 task_params->sqe->task_id = cpu_to_le16(task_params->itid); in init_sqe()
104 SET_FIELD(task_params->sqe->flags, ISCSI_WQE_WQE_TYPE, in init_sqe()
115 init_dif_context_flags(&task_params->sqe->prot_flags, in init_sqe()
118 SET_FIELD(task_params->sqe->flags, ISCSI_WQE_WQE_TYPE, in init_sqe()
134 SET_FIELD(task_params->sqe->flags, ISCSI_WQE_NUM_SGES, in init_sqe()
136 SET_FIELD(task_params->sqe->contlen_cdbsize, ISCSI_WQE_CONT_LEN, in init_sqe()
141 SET_FIELD(task_params->sqe->contlen_cdbsize, in init_sqe()
147 SET_FIELD(task_params->sqe->flags, ISCSI_WQE_WQE_TYPE, in init_sqe()
[all …]
Dqedi_fw.c1089 task_params.sqe = &ep->sq[sq_idx]; in qedi_send_iscsi_login()
1091 memset(task_params.sqe, 0, sizeof(struct iscsi_wqe)); in qedi_send_iscsi_login()
1163 task_params.sqe = &ep->sq[sq_idx]; in qedi_send_iscsi_logout()
1164 memset(task_params.sqe, 0, sizeof(struct iscsi_wqe)); in qedi_send_iscsi_logout()
1534 task_params.sqe = &ep->sq[sq_idx]; in qedi_send_iscsi_tmf()
1536 memset(task_params.sqe, 0, sizeof(struct iscsi_wqe)); in qedi_send_iscsi_tmf()
1674 task_params.sqe = &ep->sq[sq_idx]; in qedi_send_iscsi_text()
1676 memset(task_params.sqe, 0, sizeof(struct iscsi_wqe)); in qedi_send_iscsi_text()
1791 task_params.sqe = &ep->sq[sq_idx]; in qedi_send_iscsi_nopout()
1793 memset(task_params.sqe, 0, sizeof(struct iscsi_wqe)); in qedi_send_iscsi_nopout()
[all …]
Dqedi_fw_iscsi.h14 struct iscsi_wqe *sqe; member
/drivers/infiniband/hw/bnxt_re/
Dqplib_fp.c1601 struct sq_send_raweth_qp1 *sqe = in bnxt_qplib_post_send() local
1604 sqe->wqe_type = wqe->type; in bnxt_qplib_post_send()
1605 sqe->flags = wqe->flags; in bnxt_qplib_post_send()
1606 sqe->wqe_size = wqe_size16 + in bnxt_qplib_post_send()
1607 ((offsetof(typeof(*sqe), data) + 15) >> 4); in bnxt_qplib_post_send()
1608 sqe->cfa_action = cpu_to_le16(wqe->rawqp1.cfa_action); in bnxt_qplib_post_send()
1609 sqe->lflags = cpu_to_le16(wqe->rawqp1.lflags); in bnxt_qplib_post_send()
1610 sqe->length = cpu_to_le32(data_len); in bnxt_qplib_post_send()
1611 sqe->cfa_meta = cpu_to_le32((wqe->rawqp1.cfa_meta & in bnxt_qplib_post_send()
1621 struct sq_send *sqe = (struct sq_send *)hw_sq_send_hdr; in bnxt_qplib_post_send() local
[all …]
/drivers/infiniband/hw/cxgb4/
Drestrack.c96 struct t4_swsqe *sqe) in fill_swsqe() argument
100 if (rdma_nl_put_driver_u32(msg, "opcode", sqe->opcode)) in fill_swsqe()
102 if (rdma_nl_put_driver_u32(msg, "complete", sqe->complete)) in fill_swsqe()
104 if (sqe->complete && in fill_swsqe()
105 rdma_nl_put_driver_u32(msg, "cqe_status", CQE_STATUS(&sqe->cqe))) in fill_swsqe()
107 if (rdma_nl_put_driver_u32(msg, "signaled", sqe->signaled)) in fill_swsqe()
109 if (rdma_nl_put_driver_u32(msg, "flushed", sqe->flushed)) in fill_swsqe()
/drivers/scsi/bnx2i/
Dbnx2i.h498 struct sqe { struct
634 struct sqe *sq_virt;
638 struct sqe *sq_prod_qe;
639 struct sqe *sq_cons_qe;
640 struct sqe *sq_first_qe;
641 struct sqe *sq_last_qe;
/drivers/nvme/host/
Drdma.c55 struct nvme_rdma_qe sqe; member
284 kfree(req->sqe.data); in nvme_rdma_exit_request()
297 req->sqe.data = kzalloc(sizeof(struct nvme_command), GFP_KERNEL); in nvme_rdma_init_request()
298 if (!req->sqe.data) in nvme_rdma_init_request()
1328 container_of(qe, struct nvme_rdma_request, sqe); in nvme_rdma_send_done()
1417 struct nvme_rdma_qe *sqe = &ctrl->async_event_sqe; in nvme_rdma_submit_async_event() local
1418 struct nvme_command *cmd = sqe->data; in nvme_rdma_submit_async_event()
1422 ib_dma_sync_single_for_cpu(dev, sqe->dma, sizeof(*cmd), DMA_TO_DEVICE); in nvme_rdma_submit_async_event()
1430 sqe->cqe.done = nvme_rdma_async_done; in nvme_rdma_submit_async_event()
1432 ib_dma_sync_single_for_device(dev, sqe->dma, sizeof(*cmd), in nvme_rdma_submit_async_event()
[all …]
Dfc.c1563 struct nvme_command *sqe = &op->cmd_iu.sqe; in nvme_fc_fcpio_done() local
1667 sqe->common.command_id != cqe->command_id)) { in nvme_fc_fcpio_done()
1677 sqe->common.command_id, in nvme_fc_fcpio_done()
1786 struct nvme_command *sqe; in nvme_fc_init_aen_ops() local
1798 sqe = &cmdiu->sqe; in nvme_fc_init_aen_ops()
1810 memset(sqe, 0, sizeof(*sqe)); in nvme_fc_init_aen_ops()
1811 sqe->common.opcode = nvme_admin_async_event; in nvme_fc_init_aen_ops()
1813 sqe->common.command_id = NVME_AQ_BLK_MQ_DEPTH + i; in nvme_fc_init_aen_ops()
2215 struct nvme_command *sqe = &cmdiu->sqe; in nvme_fc_start_fcp_op() local
2253 WARN_ON_ONCE(sqe->common.metadata); in nvme_fc_start_fcp_op()
[all …]
/drivers/crypto/hisilicon/
Dqm.h174 int (*fill_sqe)(void *sqe, void *q_parm, void *d_parm);
183 void *sqe; member
Dqm.c467 qp->req_cb(qp, qp->sqe + qm->sqe_size * cqe->sq_head); in qm_poll_qp()
1094 return qp->sqe + sq_tail * qp->qm->sqe_size; in qm_get_avail_sqe()
1298 QP_INIT_BUF(qp, sqe, qm->sqe_size * QM_Q_DEPTH); in hisi_qm_start_qp()
1304 ver, qp->sqe, (unsigned long)qp->sqe_dma, in hisi_qm_start_qp()
1362 void *sqe = qm_get_avail_sqe(qp); in hisi_qp_send() local
1369 if (!sqe) in hisi_qp_send()
1372 memcpy(sqe, msg, qp->qm->sqe_size); in hisi_qp_send()
/drivers/nvme/target/
Dfc.c1777 struct nvme_common_command *sqe = &fod->cmdiubuf.sqe.common; in nvmet_fc_prep_fcp_rsp() local
1809 nvme_is_fabrics((struct nvme_command *) sqe) || in nvmet_fc_prep_fcp_rsp()
1812 (sqe->flags & (NVME_CMD_FUSE_FIRST | NVME_CMD_FUSE_SECOND)) || in nvmet_fc_prep_fcp_rsp()
2083 struct nvme_common_command *sqe = &fod->cmdiubuf.sqe.common; in __nvmet_fc_fcp_nvme_cmd_done() local
2107 cqe->command_id = sqe->command_id; in __nvmet_fc_fcp_nvme_cmd_done()
2174 if (!nvme_is_write(&cmdiu->sqe)) in nvmet_fc_handle_fcp_rqst()
2178 if (nvme_is_write(&cmdiu->sqe)) in nvmet_fc_handle_fcp_rqst()
2186 fod->req.cmd = &fod->cmdiubuf.sqe; in nvmet_fc_handle_fcp_rqst()
/drivers/scsi/lpfc/
Dlpfc_nvme.c1036 cid = cp->sqe.common.command_id; in lpfc_nvme_io_cmd_wqe_cmpl()
1509 struct nvme_common_command *sqe; in lpfc_nvme_fcp_io_submit() local
1599 sqe = &((struct nvme_fc_cmd_iu *) in lpfc_nvme_fcp_io_submit()
1600 pnvme_fcreq->cmdaddr)->sqe.common; in lpfc_nvme_fcp_io_submit()
1601 if (sqe->opcode == nvme_admin_keep_alive) in lpfc_nvme_fcp_io_submit()
/drivers/scsi/bnx2fc/
Dbnx2fc_hwi.c1377 struct fcoe_sqe *sqe; in bnx2fc_add_2_sq() local
1379 sqe = &tgt->sq[tgt->sq_prod_idx]; in bnx2fc_add_2_sq()
1382 sqe->wqe = xid << FCOE_SQE_TASK_ID_SHIFT; in bnx2fc_add_2_sq()
1383 sqe->wqe |= tgt->sq_curr_toggle_bit << FCOE_SQE_TOGGLE_BIT_SHIFT; in bnx2fc_add_2_sq()

12