| /drivers/media/mc/ |
| D | mc-request.c | 39 static void media_request_clean(struct media_request *req) in media_request_clean() argument 44 WARN_ON(req->state != MEDIA_REQUEST_STATE_CLEANING); in media_request_clean() 45 WARN_ON(req->updating_count); in media_request_clean() 46 WARN_ON(req->access_count); in media_request_clean() 48 list_for_each_entry_safe(obj, obj_safe, &req->objects, list) { in media_request_clean() 53 req->updating_count = 0; in media_request_clean() 54 req->access_count = 0; in media_request_clean() 55 WARN_ON(req->num_incomplete_objects); in media_request_clean() 56 req->num_incomplete_objects = 0; in media_request_clean() 57 wake_up_interruptible_all(&req->poll_wait); in media_request_clean() [all …]
|
| /drivers/s390/scsi/ |
| D | zfcp_fsf.c | 80 static void zfcp_fsf_class_not_supp(struct zfcp_fsf_req *req) in zfcp_fsf_class_not_supp() argument 82 dev_err(&req->adapter->ccw_device->dev, "FCP device not " in zfcp_fsf_class_not_supp() 84 zfcp_erp_adapter_shutdown(req->adapter, 0, "fscns_1"); in zfcp_fsf_class_not_supp() 85 req->status |= ZFCP_STATUS_FSFREQ_ERROR; in zfcp_fsf_class_not_supp() 92 void zfcp_fsf_req_free(struct zfcp_fsf_req *req) in zfcp_fsf_req_free() argument 94 if (likely(req->pool)) { in zfcp_fsf_req_free() 95 if (likely(!zfcp_fsf_req_is_status_read_buffer(req))) in zfcp_fsf_req_free() 96 mempool_free(req->qtcb, req->adapter->pool.qtcb_pool); in zfcp_fsf_req_free() 97 mempool_free(req, req->pool); in zfcp_fsf_req_free() 101 if (likely(!zfcp_fsf_req_is_status_read_buffer(req))) in zfcp_fsf_req_free() [all …]
|
| /drivers/block/drbd/ |
| D | drbd_req.c | 26 struct drbd_request *req; in drbd_req_new() local 28 req = mempool_alloc(&drbd_request_mempool, GFP_NOIO); in drbd_req_new() 29 if (!req) in drbd_req_new() 31 memset(req, 0, sizeof(*req)); in drbd_req_new() 33 req->rq_state = (bio_data_dir(bio_src) == WRITE ? RQ_WRITE : 0) in drbd_req_new() 36 req->device = device; in drbd_req_new() 37 req->master_bio = bio_src; in drbd_req_new() 38 req->epoch = 0; in drbd_req_new() 40 drbd_clear_interval(&req->i); in drbd_req_new() 41 req->i.sector = bio_src->bi_iter.bi_sector; in drbd_req_new() [all …]
|
| /drivers/nvme/target/ |
| D | io-cmd-file.c | 76 static ssize_t nvmet_file_submit_bvec(struct nvmet_req *req, loff_t pos, in nvmet_file_submit_bvec() argument 79 struct kiocb *iocb = &req->f.iocb; in nvmet_file_submit_bvec() 84 if (req->cmd->rw.opcode == nvme_cmd_write) { in nvmet_file_submit_bvec() 85 if (req->cmd->rw.control & cpu_to_le16(NVME_RW_FUA)) in nvmet_file_submit_bvec() 87 call_iter = req->ns->file->f_op->write_iter; in nvmet_file_submit_bvec() 90 call_iter = req->ns->file->f_op->read_iter; in nvmet_file_submit_bvec() 94 iov_iter_bvec(&iter, rw, req->f.bvec, nr_segs, count); in nvmet_file_submit_bvec() 97 iocb->ki_filp = req->ns->file; in nvmet_file_submit_bvec() 105 struct nvmet_req *req = container_of(iocb, struct nvmet_req, f.iocb); in nvmet_file_io_done() local 108 if (req->f.bvec != req->inline_bvec) { in nvmet_file_io_done() [all …]
|
| D | admin-cmd.c | 28 static u32 nvmet_feat_data_len(struct nvmet_req *req, u32 cdw10) in nvmet_feat_data_len() argument 32 return sizeof(req->sq->ctrl->hostid); in nvmet_feat_data_len() 43 static void nvmet_execute_get_log_page_noop(struct nvmet_req *req) in nvmet_execute_get_log_page_noop() argument 45 nvmet_req_complete(req, nvmet_zero_sgl(req, 0, req->transfer_len)); in nvmet_execute_get_log_page_noop() 48 static void nvmet_execute_get_log_page_error(struct nvmet_req *req) in nvmet_execute_get_log_page_error() argument 50 struct nvmet_ctrl *ctrl = req->sq->ctrl; in nvmet_execute_get_log_page_error() 60 if (nvmet_copy_to_sgl(req, offset, &ctrl->slots[slot], in nvmet_execute_get_log_page_error() 71 nvmet_req_complete(req, 0); in nvmet_execute_get_log_page_error() 74 static u16 nvmet_get_smart_log_nsid(struct nvmet_req *req, in nvmet_get_smart_log_nsid() argument 80 status = nvmet_req_find_ns(req); in nvmet_get_smart_log_nsid() [all …]
|
| D | fabrics-cmd-auth.c | 34 static u8 nvmet_auth_negotiate(struct nvmet_req *req, void *d) in nvmet_auth_negotiate() argument 36 struct nvmet_ctrl *ctrl = req->sq->ctrl; in nvmet_auth_negotiate() 41 __func__, ctrl->cntlid, req->sq->qid, in nvmet_auth_negotiate() 45 req->sq->dhchap_tid = le16_to_cpu(data->t_id); in nvmet_auth_negotiate() 70 __func__, ctrl->cntlid, req->sq->qid); in nvmet_auth_negotiate() 74 __func__, ctrl->cntlid, req->sq->qid, in nvmet_auth_negotiate() 98 __func__, ctrl->cntlid, req->sq->qid); in nvmet_auth_negotiate() 102 __func__, ctrl->cntlid, req->sq->qid, in nvmet_auth_negotiate() 107 __func__, ctrl->cntlid, req->sq->qid, in nvmet_auth_negotiate() 112 static u8 nvmet_auth_reply(struct nvmet_req *req, void *d) in nvmet_auth_reply() argument [all …]
|
| D | zns.c | 69 void nvmet_execute_identify_ctrl_zns(struct nvmet_req *req) in nvmet_execute_identify_ctrl_zns() argument 71 u8 zasl = req->sq->ctrl->subsys->zasl; in nvmet_execute_identify_ctrl_zns() 72 struct nvmet_ctrl *ctrl = req->sq->ctrl; in nvmet_execute_identify_ctrl_zns() 87 status = nvmet_copy_to_sgl(req, 0, id, sizeof(*id)); in nvmet_execute_identify_ctrl_zns() 91 nvmet_req_complete(req, status); in nvmet_execute_identify_ctrl_zns() 94 void nvmet_execute_identify_ns_zns(struct nvmet_req *req) in nvmet_execute_identify_ns_zns() argument 101 if (le32_to_cpu(req->cmd->identify.nsid) == NVME_NSID_ALL) { in nvmet_execute_identify_ns_zns() 102 req->error_loc = offsetof(struct nvme_identify, nsid); in nvmet_execute_identify_ns_zns() 113 status = nvmet_req_find_ns(req); in nvmet_execute_identify_ns_zns() 117 if (nvmet_ns_revalidate(req->ns)) { in nvmet_execute_identify_ns_zns() [all …]
|
| D | io-cmd-bdev.c | 126 u16 blk_to_nvme_status(struct nvmet_req *req, blk_status_t blk_sts) in blk_to_nvme_status() argument 141 req->error_loc = offsetof(struct nvme_rw_command, length); in blk_to_nvme_status() 145 req->error_loc = offsetof(struct nvme_rw_command, slba); in blk_to_nvme_status() 149 req->error_loc = offsetof(struct nvme_common_command, opcode); in blk_to_nvme_status() 153 req->error_loc = offsetof(struct nvme_rw_command, nsid); in blk_to_nvme_status() 158 req->error_loc = offsetof(struct nvme_common_command, opcode); in blk_to_nvme_status() 161 switch (req->cmd->common.opcode) { in blk_to_nvme_status() 164 req->error_slba = le64_to_cpu(req->cmd->rw.slba); in blk_to_nvme_status() 167 req->error_slba = in blk_to_nvme_status() 168 le64_to_cpu(req->cmd->write_zeroes.slba); in blk_to_nvme_status() [all …]
|
| D | fabrics-cmd.c | 10 static void nvmet_execute_prop_set(struct nvmet_req *req) in nvmet_execute_prop_set() argument 12 u64 val = le64_to_cpu(req->cmd->prop_set.value); in nvmet_execute_prop_set() 15 if (!nvmet_check_transfer_len(req, 0)) in nvmet_execute_prop_set() 18 if (req->cmd->prop_set.attrib & 1) { in nvmet_execute_prop_set() 19 req->error_loc = in nvmet_execute_prop_set() 25 switch (le32_to_cpu(req->cmd->prop_set.offset)) { in nvmet_execute_prop_set() 27 nvmet_update_cc(req->sq->ctrl, val); in nvmet_execute_prop_set() 30 req->error_loc = in nvmet_execute_prop_set() 35 nvmet_req_complete(req, status); in nvmet_execute_prop_set() 38 static void nvmet_execute_prop_get(struct nvmet_req *req) in nvmet_execute_prop_get() argument [all …]
|
| D | passthru.c | 33 static u16 nvmet_passthru_override_id_descs(struct nvmet_req *req) in nvmet_passthru_override_id_descs() argument 35 struct nvmet_ctrl *ctrl = req->sq->ctrl; in nvmet_passthru_override_id_descs() 49 status = nvmet_copy_from_sgl(req, 0, data, NVME_IDENTIFY_DATA_SIZE); in nvmet_passthru_override_id_descs() 74 status = nvmet_copy_to_sgl(req, 0, data, NVME_IDENTIFY_DATA_SIZE); in nvmet_passthru_override_id_descs() 80 static u16 nvmet_passthru_override_id_ctrl(struct nvmet_req *req) in nvmet_passthru_override_id_ctrl() argument 82 struct nvmet_ctrl *ctrl = req->sq->ctrl; in nvmet_passthru_override_id_ctrl() 93 status = nvmet_copy_from_sgl(req, 0, id, sizeof(*id)); in nvmet_passthru_override_id_ctrl() 143 if (req->port->inline_data_size) in nvmet_passthru_override_id_ctrl() 157 req->port->inline_data_size) / 16); in nvmet_passthru_override_id_ctrl() 168 status = nvmet_copy_to_sgl(req, 0, id, sizeof(struct nvme_id_ctrl)); in nvmet_passthru_override_id_ctrl() [all …]
|
| /drivers/staging/greybus/ |
| D | audio_apbridgea.c | 16 struct audio_apbridgea_set_config_request req; in gb_audio_apbridgea_set_config() local 18 req.hdr.type = AUDIO_APBRIDGEA_TYPE_SET_CONFIG; in gb_audio_apbridgea_set_config() 19 req.hdr.i2s_port = cpu_to_le16(i2s_port); in gb_audio_apbridgea_set_config() 20 req.format = cpu_to_le32(format); in gb_audio_apbridgea_set_config() 21 req.rate = cpu_to_le32(rate); in gb_audio_apbridgea_set_config() 22 req.mclk_freq = cpu_to_le32(mclk_freq); in gb_audio_apbridgea_set_config() 24 return gb_hd_output(connection->hd, &req, sizeof(req), in gb_audio_apbridgea_set_config() 33 struct audio_apbridgea_register_cport_request req; in gb_audio_apbridgea_register_cport() local 36 req.hdr.type = AUDIO_APBRIDGEA_TYPE_REGISTER_CPORT; in gb_audio_apbridgea_register_cport() 37 req.hdr.i2s_port = cpu_to_le16(i2s_port); in gb_audio_apbridgea_register_cport() [all …]
|
| D | audio_gb.c | 50 struct gb_audio_get_control_request req; in gb_audio_gb_get_control() local 54 req.control_id = control_id; in gb_audio_gb_get_control() 55 req.index = index; in gb_audio_gb_get_control() 58 &req, sizeof(req), &resp, sizeof(resp)); in gb_audio_gb_get_control() 72 struct gb_audio_set_control_request req; in gb_audio_gb_set_control() local 74 req.control_id = control_id; in gb_audio_gb_set_control() 75 req.index = index; in gb_audio_gb_set_control() 76 memcpy(&req.value, value, sizeof(req.value)); in gb_audio_gb_set_control() 79 &req, sizeof(req), NULL, 0); in gb_audio_gb_set_control() 86 struct gb_audio_enable_widget_request req; in gb_audio_gb_enable_widget() local [all …]
|
| /drivers/net/ethernet/marvell/octeontx2/af/ |
| D | mcs_rvu_if.c | 21 struct _req_type *req; \ 23 req = (struct _req_type *)otx2_mbox_alloc_msg_rsp( \ 26 if (!req) \ 28 req->hdr.sig = OTX2_MBOX_REQ_SIG; \ 29 req->hdr.id = _id; \ 30 return req; \ 73 struct mcs_set_lmac_mode *req, in rvu_mbox_handler_mcs_set_lmac_mode() argument 78 if (req->mcs_id >= rvu->mcs_blk_cnt) in rvu_mbox_handler_mcs_set_lmac_mode() 81 mcs = mcs_get_pdata(req->mcs_id); in rvu_mbox_handler_mcs_set_lmac_mode() 83 if (BIT_ULL(req->lmac_id) & mcs->hw->lmac_bmap) in rvu_mbox_handler_mcs_set_lmac_mode() [all …]
|
| /drivers/infiniband/hw/hfi1/ |
| D | user_sdma.c | 38 static int user_sdma_send_pkts(struct user_sdma_request *req, u16 maxpkts); 41 static void user_sdma_free_request(struct user_sdma_request *req); 42 static int check_header_template(struct user_sdma_request *req, 45 static int set_txreq_header(struct user_sdma_request *req, 47 static int set_txreq_header_ahg(struct user_sdma_request *req, 290 struct user_sdma_request *req; in hfi1_user_sdma_process_request() local 297 if (iovec[idx].iov_len < sizeof(info) + sizeof(req->hdr)) { in hfi1_user_sdma_process_request() 302 iovec[idx].iov_len, sizeof(info) + sizeof(req->hdr)); in hfi1_user_sdma_process_request() 352 req = pq->reqs + info.comp_idx; in hfi1_user_sdma_process_request() 353 req->data_iovs = req_iovcnt(info.ctrl) - 1; /* subtract header vector */ in hfi1_user_sdma_process_request() [all …]
|
| /drivers/peci/ |
| D | request.c | 88 static u8 peci_request_data_cc(struct peci_request *req) in peci_request_data_cc() argument 90 return req->rx.buf[0]; in peci_request_data_cc() 102 int peci_request_status(struct peci_request *req) in peci_request_status() argument 104 u8 cc = peci_request_data_cc(req); in peci_request_status() 107 dev_dbg(&req->device->dev, "ret: %#02x\n", cc); in peci_request_status() 133 static int peci_request_xfer(struct peci_request *req) in peci_request_xfer() argument 135 struct peci_device *device = req->device; in peci_request_xfer() 140 ret = controller->ops->xfer(controller, device->addr, req); in peci_request_xfer() 146 static int peci_request_xfer_retry(struct peci_request *req) in peci_request_xfer_retry() argument 149 struct peci_device *device = req->device; in peci_request_xfer_retry() [all …]
|
| /drivers/clk/sunxi/ |
| D | clk-sunxi.c | 33 static void sun4i_get_pll1_factors(struct factors_request *req) in sun4i_get_pll1_factors() argument 38 div = req->rate / 6000000; in sun4i_get_pll1_factors() 39 req->rate = 6000000 * div; in sun4i_get_pll1_factors() 42 req->m = 0; in sun4i_get_pll1_factors() 45 if (req->rate >= 768000000 || req->rate == 42000000 || in sun4i_get_pll1_factors() 46 req->rate == 54000000) in sun4i_get_pll1_factors() 47 req->k = 1; in sun4i_get_pll1_factors() 49 req->k = 0; in sun4i_get_pll1_factors() 53 req->p = 3; in sun4i_get_pll1_factors() 57 req->p = 2; in sun4i_get_pll1_factors() [all …]
|
| /drivers/crypto/inside-secure/ |
| D | safexcel_hash.c | 70 static inline u64 safexcel_queued_len(struct safexcel_ahash_req *req) in safexcel_queued_len() argument 72 return req->len - req->processed; in safexcel_queued_len() 109 struct safexcel_ahash_req *req, in safexcel_context_control() argument 123 if (unlikely(req->digest == CONTEXT_CONTROL_DIGEST_XCM)) { in safexcel_context_control() 124 if (req->xcbcmac) in safexcel_context_control() 127 memcpy(ctx->base.ctxr->data, req->state, req->state_sz); in safexcel_context_control() 129 if (!req->finish && req->xcbcmac) in safexcel_context_control() 134 CONTEXT_CONTROL_SIZE(req->state_sz / in safexcel_context_control() 140 CONTEXT_CONTROL_SIZE(req->state_sz / in safexcel_context_control() 143 } else if (!req->processed) { in safexcel_context_control() [all …]
|
| /drivers/macintosh/ |
| D | via-pmu.c | 197 static int pmu_send_request(struct adb_request *req, int sync); 231 int pmu_polled_request(struct adb_request *req); 573 struct adb_request req; in init_pmu() local 579 pmu_request(&req, NULL, 2, PMU_SET_INTR_MASK, pmu_intr_mask); in init_pmu() 581 while (!req.complete) { in init_pmu() 606 pmu_request(&req, NULL, 2, PMU_SYSTEM_READY, 2); in init_pmu() 607 while (!req.complete) in init_pmu() 612 pmu_request(&req, NULL, 1, PMU_GET_VERSION); in init_pmu() 613 pmu_wait_complete(&req); in init_pmu() 614 if (req.reply_len > 0) in init_pmu() [all …]
|
| D | via-macii.c | 91 static int macii_send_request(struct adb_request *req, int sync); 92 static int macii_write(struct adb_request *req); 181 static struct adb_request req; in macii_queue_poll() local 216 adb_request(&req, NULL, ADBREQ_NOSEND, 1, poll_command); in macii_queue_poll() 218 req.sent = 0; in macii_queue_poll() 219 req.complete = 0; in macii_queue_poll() 220 req.reply_len = 0; in macii_queue_poll() 221 req.next = current_req; in macii_queue_poll() 224 current_req = &req; in macii_queue_poll() 226 current_req = &req; in macii_queue_poll() [all …]
|
| /drivers/s390/cio/ |
| D | ccwreq.c | 43 struct ccw_request *req = &cdev->private->req; in ccwreq_next_path() local 45 if (!req->singlepath) { in ccwreq_next_path() 46 req->mask = 0; in ccwreq_next_path() 49 req->retries = req->maxretries; in ccwreq_next_path() 50 req->mask = lpm_adjust(req->mask >> 1, req->lpm); in ccwreq_next_path() 52 return req->mask; in ccwreq_next_path() 60 struct ccw_request *req = &cdev->private->req; in ccwreq_stop() local 62 if (req->done) in ccwreq_stop() 64 req->done = 1; in ccwreq_stop() 67 if (rc && rc != -ENODEV && req->drc) in ccwreq_stop() [all …]
|
| /drivers/infiniband/hw/bnxt_re/ |
| D | qplib_tlv.h | 48 static inline u8 __get_cmdq_base_opcode(struct cmdq_base *req, u32 size) in __get_cmdq_base_opcode() argument 50 if (HAS_TLV_HEADER(req) && size > TLV_BYTES) in __get_cmdq_base_opcode() 51 return ((struct cmdq_base *)GET_TLV_DATA(req))->opcode; in __get_cmdq_base_opcode() 53 return req->opcode; in __get_cmdq_base_opcode() 56 static inline void __set_cmdq_base_opcode(struct cmdq_base *req, in __set_cmdq_base_opcode() argument 59 if (HAS_TLV_HEADER(req) && size > TLV_BYTES) in __set_cmdq_base_opcode() 60 ((struct cmdq_base *)GET_TLV_DATA(req))->opcode = val; in __set_cmdq_base_opcode() 62 req->opcode = val; in __set_cmdq_base_opcode() 65 static inline __le16 __get_cmdq_base_cookie(struct cmdq_base *req, u32 size) in __get_cmdq_base_cookie() argument 67 if (HAS_TLV_HEADER(req) && size > TLV_BYTES) in __get_cmdq_base_cookie() [all …]
|
| /drivers/usb/gadget/udc/ |
| D | renesas_usbf.c | 259 struct usb_request req; member 558 static int usbf_ep0_pio_in(struct usbf_ep *ep0, struct usbf_req *req) in usbf_ep0_pio_in() argument 566 left = req->req.length - req->req.actual; in usbf_ep0_pio_in() 569 if (!req->is_zero_sent) { in usbf_ep0_pio_in() 570 if (req->req.length == 0) { in usbf_ep0_pio_in() 573 req->is_zero_sent = 1; in usbf_ep0_pio_in() 576 if ((req->req.actual % ep0->ep.maxpacket) == 0) { in usbf_ep0_pio_in() 577 if (req->req.zero) { in usbf_ep0_pio_in() 580 req->is_zero_sent = 1; in usbf_ep0_pio_in() 591 buf = req->req.buf; in usbf_ep0_pio_in() [all …]
|
| /drivers/crypto/amcc/ |
| D | crypto4xx_alg.c | 69 static inline int crypto4xx_crypt(struct skcipher_request *req, in crypto4xx_crypt() argument 73 struct crypto_skcipher *cipher = crypto_skcipher_reqtfm(req); in crypto4xx_crypt() 77 if (check_blocksize && !IS_ALIGNED(req->cryptlen, AES_BLOCK_SIZE)) in crypto4xx_crypt() 81 crypto4xx_memcpy_to_le32(iv, req->iv, ivlen); in crypto4xx_crypt() 83 return crypto4xx_build_pd(&req->base, ctx, req->src, req->dst, in crypto4xx_crypt() 84 req->cryptlen, iv, ivlen, decrypt ? ctx->sa_in : ctx->sa_out, in crypto4xx_crypt() 88 int crypto4xx_encrypt_noiv_block(struct skcipher_request *req) in crypto4xx_encrypt_noiv_block() argument 90 return crypto4xx_crypt(req, 0, false, true); in crypto4xx_encrypt_noiv_block() 93 int crypto4xx_encrypt_iv_stream(struct skcipher_request *req) in crypto4xx_encrypt_iv_stream() argument 95 return crypto4xx_crypt(req, AES_IV_SIZE, false, false); in crypto4xx_encrypt_iv_stream() [all …]
|
| /drivers/crypto/marvell/cesa/ |
| D | cipher.c | 42 struct skcipher_request *req) in mv_cesa_skcipher_req_iter_init() argument 44 mv_cesa_req_dma_iter_init(&iter->base, req->cryptlen); in mv_cesa_skcipher_req_iter_init() 45 mv_cesa_sg_dma_iter_init(&iter->src, req->src, DMA_TO_DEVICE); in mv_cesa_skcipher_req_iter_init() 46 mv_cesa_sg_dma_iter_init(&iter->dst, req->dst, DMA_FROM_DEVICE); in mv_cesa_skcipher_req_iter_init() 59 mv_cesa_skcipher_dma_cleanup(struct skcipher_request *req) in mv_cesa_skcipher_dma_cleanup() argument 61 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); in mv_cesa_skcipher_dma_cleanup() 63 if (req->dst != req->src) { in mv_cesa_skcipher_dma_cleanup() 64 dma_unmap_sg(cesa_dev->dev, req->dst, creq->dst_nents, in mv_cesa_skcipher_dma_cleanup() 66 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents, in mv_cesa_skcipher_dma_cleanup() 69 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents, in mv_cesa_skcipher_dma_cleanup() [all …]
|
| /drivers/accel/ivpu/ |
| D | ivpu_jsm_msg.c | 98 struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_REGISTER_DB }; in ivpu_jsm_register_db() local 102 req.payload.register_db.db_idx = db_id; in ivpu_jsm_register_db() 103 req.payload.register_db.jobq_base = jobq_base; in ivpu_jsm_register_db() 104 req.payload.register_db.jobq_size = jobq_size; in ivpu_jsm_register_db() 105 req.payload.register_db.host_ssid = ctx_id; in ivpu_jsm_register_db() 107 ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_REGISTER_DB_DONE, &resp, in ivpu_jsm_register_db() 117 struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_UNREGISTER_DB }; in ivpu_jsm_unregister_db() local 121 req.payload.unregister_db.db_idx = db_id; in ivpu_jsm_unregister_db() 123 ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_UNREGISTER_DB_DONE, &resp, in ivpu_jsm_unregister_db() 133 struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_QUERY_ENGINE_HB }; in ivpu_jsm_get_heartbeat() local [all …]
|