Home
last modified time | relevance | path

Searched refs:req (Results 1 – 25 of 1158) sorted by relevance

12345678910>>...47

/drivers/media/mc/
Dmc-request.c39 static void media_request_clean(struct media_request *req) in media_request_clean() argument
44 WARN_ON(req->state != MEDIA_REQUEST_STATE_CLEANING); in media_request_clean()
45 WARN_ON(req->updating_count); in media_request_clean()
46 WARN_ON(req->access_count); in media_request_clean()
48 list_for_each_entry_safe(obj, obj_safe, &req->objects, list) { in media_request_clean()
53 req->updating_count = 0; in media_request_clean()
54 req->access_count = 0; in media_request_clean()
55 WARN_ON(req->num_incomplete_objects); in media_request_clean()
56 req->num_incomplete_objects = 0; in media_request_clean()
57 wake_up_interruptible_all(&req->poll_wait); in media_request_clean()
[all …]
/drivers/s390/scsi/
Dzfcp_fsf.c80 static void zfcp_fsf_class_not_supp(struct zfcp_fsf_req *req) in zfcp_fsf_class_not_supp() argument
82 dev_err(&req->adapter->ccw_device->dev, "FCP device not " in zfcp_fsf_class_not_supp()
84 zfcp_erp_adapter_shutdown(req->adapter, 0, "fscns_1"); in zfcp_fsf_class_not_supp()
85 req->status |= ZFCP_STATUS_FSFREQ_ERROR; in zfcp_fsf_class_not_supp()
92 void zfcp_fsf_req_free(struct zfcp_fsf_req *req) in zfcp_fsf_req_free() argument
94 if (likely(req->pool)) { in zfcp_fsf_req_free()
95 if (likely(!zfcp_fsf_req_is_status_read_buffer(req))) in zfcp_fsf_req_free()
96 mempool_free(req->qtcb, req->adapter->pool.qtcb_pool); in zfcp_fsf_req_free()
97 mempool_free(req, req->pool); in zfcp_fsf_req_free()
101 if (likely(!zfcp_fsf_req_is_status_read_buffer(req))) in zfcp_fsf_req_free()
[all …]
/drivers/nvme/target/
Dio-cmd-file.c102 static ssize_t nvmet_file_submit_bvec(struct nvmet_req *req, loff_t pos, in nvmet_file_submit_bvec() argument
105 struct kiocb *iocb = &req->f.iocb; in nvmet_file_submit_bvec()
110 if (req->cmd->rw.opcode == nvme_cmd_write) { in nvmet_file_submit_bvec()
111 if (req->cmd->rw.control & cpu_to_le16(NVME_RW_FUA)) in nvmet_file_submit_bvec()
113 call_iter = req->ns->file->f_op->write_iter; in nvmet_file_submit_bvec()
116 call_iter = req->ns->file->f_op->read_iter; in nvmet_file_submit_bvec()
120 iov_iter_bvec(&iter, rw, req->f.bvec, nr_segs, count); in nvmet_file_submit_bvec()
123 iocb->ki_filp = req->ns->file; in nvmet_file_submit_bvec()
124 iocb->ki_flags = ki_flags | iocb_flags(req->ns->file); in nvmet_file_submit_bvec()
131 struct nvmet_req *req = container_of(iocb, struct nvmet_req, f.iocb); in nvmet_file_io_done() local
[all …]
Dadmin-cmd.c28 static u32 nvmet_feat_data_len(struct nvmet_req *req, u32 cdw10) in nvmet_feat_data_len() argument
32 return sizeof(req->sq->ctrl->hostid); in nvmet_feat_data_len()
43 static void nvmet_execute_get_log_page_noop(struct nvmet_req *req) in nvmet_execute_get_log_page_noop() argument
45 nvmet_req_complete(req, nvmet_zero_sgl(req, 0, req->transfer_len)); in nvmet_execute_get_log_page_noop()
48 static void nvmet_execute_get_log_page_error(struct nvmet_req *req) in nvmet_execute_get_log_page_error() argument
50 struct nvmet_ctrl *ctrl = req->sq->ctrl; in nvmet_execute_get_log_page_error()
60 if (nvmet_copy_to_sgl(req, offset, &ctrl->slots[slot], in nvmet_execute_get_log_page_error()
71 nvmet_req_complete(req, 0); in nvmet_execute_get_log_page_error()
74 static u16 nvmet_get_smart_log_nsid(struct nvmet_req *req, in nvmet_get_smart_log_nsid() argument
80 ns = nvmet_find_namespace(req->sq->ctrl, req->cmd->get_log_page.nsid); in nvmet_get_smart_log_nsid()
[all …]
Dio-cmd-bdev.c105 static u16 blk_to_nvme_status(struct nvmet_req *req, blk_status_t blk_sts) in blk_to_nvme_status() argument
120 req->error_loc = offsetof(struct nvme_rw_command, length); in blk_to_nvme_status()
124 req->error_loc = offsetof(struct nvme_rw_command, slba); in blk_to_nvme_status()
127 req->error_loc = offsetof(struct nvme_common_command, opcode); in blk_to_nvme_status()
128 switch (req->cmd->common.opcode) { in blk_to_nvme_status()
139 req->error_loc = offsetof(struct nvme_rw_command, nsid); in blk_to_nvme_status()
144 req->error_loc = offsetof(struct nvme_common_command, opcode); in blk_to_nvme_status()
147 switch (req->cmd->common.opcode) { in blk_to_nvme_status()
150 req->error_slba = le64_to_cpu(req->cmd->rw.slba); in blk_to_nvme_status()
153 req->error_slba = in blk_to_nvme_status()
[all …]
Dfabrics-cmd.c10 static void nvmet_execute_prop_set(struct nvmet_req *req) in nvmet_execute_prop_set() argument
12 u64 val = le64_to_cpu(req->cmd->prop_set.value); in nvmet_execute_prop_set()
15 if (!nvmet_check_transfer_len(req, 0)) in nvmet_execute_prop_set()
18 if (req->cmd->prop_set.attrib & 1) { in nvmet_execute_prop_set()
19 req->error_loc = in nvmet_execute_prop_set()
25 switch (le32_to_cpu(req->cmd->prop_set.offset)) { in nvmet_execute_prop_set()
27 nvmet_update_cc(req->sq->ctrl, val); in nvmet_execute_prop_set()
30 req->error_loc = in nvmet_execute_prop_set()
35 nvmet_req_complete(req, status); in nvmet_execute_prop_set()
38 static void nvmet_execute_prop_get(struct nvmet_req *req) in nvmet_execute_prop_get() argument
[all …]
Dpassthru.c23 static u16 nvmet_passthru_override_id_ctrl(struct nvmet_req *req) in nvmet_passthru_override_id_ctrl() argument
25 struct nvmet_ctrl *ctrl = req->sq->ctrl; in nvmet_passthru_override_id_ctrl()
36 status = nvmet_copy_from_sgl(req, 0, id, sizeof(*id)); in nvmet_passthru_override_id_ctrl()
86 if (req->port->inline_data_size) in nvmet_passthru_override_id_ctrl()
100 req->port->inline_data_size) / 16); in nvmet_passthru_override_id_ctrl()
111 status = nvmet_copy_to_sgl(req, 0, id, sizeof(struct nvme_id_ctrl)); in nvmet_passthru_override_id_ctrl()
118 static u16 nvmet_passthru_override_id_ns(struct nvmet_req *req) in nvmet_passthru_override_id_ns() argument
128 status = nvmet_copy_from_sgl(req, 0, id, sizeof(struct nvme_id_ns)); in nvmet_passthru_override_id_ns()
145 status = nvmet_copy_to_sgl(req, 0, id, sizeof(*id)); in nvmet_passthru_override_id_ns()
154 struct nvmet_req *req = container_of(w, struct nvmet_req, p.work); in nvmet_passthru_execute_cmd_work() local
[all …]
/drivers/block/drbd/
Ddrbd_req.c26 struct drbd_request *req; in drbd_req_new() local
28 req = mempool_alloc(&drbd_request_mempool, GFP_NOIO); in drbd_req_new()
29 if (!req) in drbd_req_new()
31 memset(req, 0, sizeof(*req)); in drbd_req_new()
33 drbd_req_make_private_bio(req, bio_src); in drbd_req_new()
34 req->rq_state = (bio_data_dir(bio_src) == WRITE ? RQ_WRITE : 0) in drbd_req_new()
38 req->device = device; in drbd_req_new()
39 req->master_bio = bio_src; in drbd_req_new()
40 req->epoch = 0; in drbd_req_new()
42 drbd_clear_interval(&req->i); in drbd_req_new()
[all …]
/drivers/staging/greybus/
Daudio_apbridgea.c16 struct audio_apbridgea_set_config_request req; in gb_audio_apbridgea_set_config() local
18 req.hdr.type = AUDIO_APBRIDGEA_TYPE_SET_CONFIG; in gb_audio_apbridgea_set_config()
19 req.hdr.i2s_port = cpu_to_le16(i2s_port); in gb_audio_apbridgea_set_config()
20 req.format = cpu_to_le32(format); in gb_audio_apbridgea_set_config()
21 req.rate = cpu_to_le32(rate); in gb_audio_apbridgea_set_config()
22 req.mclk_freq = cpu_to_le32(mclk_freq); in gb_audio_apbridgea_set_config()
24 return gb_hd_output(connection->hd, &req, sizeof(req), in gb_audio_apbridgea_set_config()
33 struct audio_apbridgea_register_cport_request req; in gb_audio_apbridgea_register_cport() local
36 req.hdr.type = AUDIO_APBRIDGEA_TYPE_REGISTER_CPORT; in gb_audio_apbridgea_register_cport()
37 req.hdr.i2s_port = cpu_to_le16(i2s_port); in gb_audio_apbridgea_register_cport()
[all …]
Daudio_gb.c50 struct gb_audio_get_control_request req; in gb_audio_gb_get_control() local
54 req.control_id = control_id; in gb_audio_gb_get_control()
55 req.index = index; in gb_audio_gb_get_control()
58 &req, sizeof(req), &resp, sizeof(resp)); in gb_audio_gb_get_control()
72 struct gb_audio_set_control_request req; in gb_audio_gb_set_control() local
74 req.control_id = control_id; in gb_audio_gb_set_control()
75 req.index = index; in gb_audio_gb_set_control()
76 memcpy(&req.value, value, sizeof(req.value)); in gb_audio_gb_set_control()
79 &req, sizeof(req), NULL, 0); in gb_audio_gb_set_control()
86 struct gb_audio_enable_widget_request req; in gb_audio_gb_enable_widget() local
[all …]
/drivers/infiniband/hw/hfi1/
Duser_sdma.c79 static int user_sdma_send_pkts(struct user_sdma_request *req, u16 maxpkts);
82 static void user_sdma_free_request(struct user_sdma_request *req);
83 static int check_header_template(struct user_sdma_request *req,
86 static int set_txreq_header(struct user_sdma_request *req,
88 static int set_txreq_header_ahg(struct user_sdma_request *req,
116 static int add_system_pages_to_sdma_packet(struct user_sdma_request *req,
351 struct user_sdma_request *req; in hfi1_user_sdma_process_request() local
358 if (iovec[idx].iov_len < sizeof(info) + sizeof(req->hdr)) { in hfi1_user_sdma_process_request()
363 iovec[idx].iov_len, sizeof(info) + sizeof(req->hdr)); in hfi1_user_sdma_process_request()
413 req = pq->reqs + info.comp_idx; in hfi1_user_sdma_process_request()
[all …]
/drivers/clk/sunxi/
Dclk-sunxi.c33 static void sun4i_get_pll1_factors(struct factors_request *req) in sun4i_get_pll1_factors() argument
38 div = req->rate / 6000000; in sun4i_get_pll1_factors()
39 req->rate = 6000000 * div; in sun4i_get_pll1_factors()
42 req->m = 0; in sun4i_get_pll1_factors()
45 if (req->rate >= 768000000 || req->rate == 42000000 || in sun4i_get_pll1_factors()
46 req->rate == 54000000) in sun4i_get_pll1_factors()
47 req->k = 1; in sun4i_get_pll1_factors()
49 req->k = 0; in sun4i_get_pll1_factors()
53 req->p = 3; in sun4i_get_pll1_factors()
57 req->p = 2; in sun4i_get_pll1_factors()
[all …]
Dclk-sun9i-core.c26 static void sun9i_a80_get_pll4_factors(struct factors_request *req) in sun9i_a80_get_pll4_factors() argument
33 n = DIV_ROUND_UP(req->rate, 6000000); in sun9i_a80_get_pll4_factors()
53 req->rate = ((24000000 * n) >> p) / (m + 1); in sun9i_a80_get_pll4_factors()
54 req->n = n; in sun9i_a80_get_pll4_factors()
55 req->m = m; in sun9i_a80_get_pll4_factors()
56 req->p = p; in sun9i_a80_get_pll4_factors()
99 static void sun9i_a80_get_gt_factors(struct factors_request *req) in sun9i_a80_get_gt_factors() argument
103 if (req->parent_rate < req->rate) in sun9i_a80_get_gt_factors()
104 req->rate = req->parent_rate; in sun9i_a80_get_gt_factors()
106 div = DIV_ROUND_UP(req->parent_rate, req->rate); in sun9i_a80_get_gt_factors()
[all …]
/drivers/crypto/inside-secure/
Dsafexcel_hash.c69 static inline u64 safexcel_queued_len(struct safexcel_ahash_req *req) in safexcel_queued_len() argument
71 return req->len - req->processed; in safexcel_queued_len()
108 struct safexcel_ahash_req *req, in safexcel_context_control() argument
122 if (unlikely(req->digest == CONTEXT_CONTROL_DIGEST_XCM)) { in safexcel_context_control()
123 if (req->xcbcmac) in safexcel_context_control()
126 memcpy(ctx->base.ctxr->data, req->state, req->state_sz); in safexcel_context_control()
128 if (!req->finish && req->xcbcmac) in safexcel_context_control()
133 CONTEXT_CONTROL_SIZE(req->state_sz / in safexcel_context_control()
139 CONTEXT_CONTROL_SIZE(req->state_sz / in safexcel_context_control()
142 } else if (!req->processed) { in safexcel_context_control()
[all …]
/drivers/s390/cio/
Dccwreq.c43 struct ccw_request *req = &cdev->private->req; in ccwreq_next_path() local
45 if (!req->singlepath) { in ccwreq_next_path()
46 req->mask = 0; in ccwreq_next_path()
49 req->retries = req->maxretries; in ccwreq_next_path()
50 req->mask = lpm_adjust(req->mask >> 1, req->lpm); in ccwreq_next_path()
52 return req->mask; in ccwreq_next_path()
60 struct ccw_request *req = &cdev->private->req; in ccwreq_stop() local
62 if (req->done) in ccwreq_stop()
64 req->done = 1; in ccwreq_stop()
67 if (rc && rc != -ENODEV && req->drc) in ccwreq_stop()
[all …]
/drivers/macintosh/
Dvia-pmu.c199 static int pmu_send_request(struct adb_request *req, int sync);
231 int pmu_polled_request(struct adb_request *req);
584 struct adb_request req; in init_pmu() local
590 pmu_request(&req, NULL, 2, PMU_SET_INTR_MASK, pmu_intr_mask); in init_pmu()
592 while (!req.complete) { in init_pmu()
617 pmu_request(&req, NULL, 2, PMU_SYSTEM_READY, 2); in init_pmu()
618 while (!req.complete) in init_pmu()
623 pmu_request(&req, NULL, 1, PMU_GET_VERSION); in init_pmu()
624 pmu_wait_complete(&req); in init_pmu()
625 if (req.reply_len > 0) in init_pmu()
[all …]
Dvia-macii.c93 static int macii_send_request(struct adb_request *req, int sync);
94 static int macii_write(struct adb_request *req);
188 static struct adb_request req; in macii_queue_poll() local
223 adb_request(&req, NULL, ADBREQ_NOSEND, 1, poll_command); in macii_queue_poll()
225 req.sent = 0; in macii_queue_poll()
226 req.complete = 0; in macii_queue_poll()
227 req.reply_len = 0; in macii_queue_poll()
228 req.next = current_req; in macii_queue_poll()
231 current_req = &req; in macii_queue_poll()
233 current_req = &req; in macii_queue_poll()
[all …]
/drivers/crypto/amcc/
Dcrypto4xx_alg.c69 static inline int crypto4xx_crypt(struct skcipher_request *req, in crypto4xx_crypt() argument
73 struct crypto_skcipher *cipher = crypto_skcipher_reqtfm(req); in crypto4xx_crypt()
77 if (check_blocksize && !IS_ALIGNED(req->cryptlen, AES_BLOCK_SIZE)) in crypto4xx_crypt()
81 crypto4xx_memcpy_to_le32(iv, req->iv, ivlen); in crypto4xx_crypt()
83 return crypto4xx_build_pd(&req->base, ctx, req->src, req->dst, in crypto4xx_crypt()
84 req->cryptlen, iv, ivlen, decrypt ? ctx->sa_in : ctx->sa_out, in crypto4xx_crypt()
88 int crypto4xx_encrypt_noiv_block(struct skcipher_request *req) in crypto4xx_encrypt_noiv_block() argument
90 return crypto4xx_crypt(req, 0, false, true); in crypto4xx_encrypt_noiv_block()
93 int crypto4xx_encrypt_iv_stream(struct skcipher_request *req) in crypto4xx_encrypt_iv_stream() argument
95 return crypto4xx_crypt(req, AES_IV_SIZE, false, false); in crypto4xx_encrypt_iv_stream()
[all …]
/drivers/staging/emxx_udc/
Demxx_udc.c165 udc->ep0_req.req.buf = p_buf; in _nbu2ss_create_ep0_packet()
166 udc->ep0_req.req.length = length; in _nbu2ss_create_ep0_packet()
167 udc->ep0_req.req.dma = 0; in _nbu2ss_create_ep0_packet()
168 udc->ep0_req.req.zero = true; in _nbu2ss_create_ep0_packet()
169 udc->ep0_req.req.complete = _nbu2ss_ep0_complete; in _nbu2ss_create_ep0_packet()
170 udc->ep0_req.req.status = -EINPROGRESS; in _nbu2ss_create_ep0_packet()
171 udc->ep0_req.req.context = udc; in _nbu2ss_create_ep0_packet()
172 udc->ep0_req.req.actual = 0; in _nbu2ss_create_ep0_packet()
456 struct nbu2ss_req *req, u8 direct) in _nbu2ss_dma_map_single() argument
458 if (req->req.dma == DMA_ADDR_INVALID) { in _nbu2ss_dma_map_single()
[all …]
/drivers/crypto/marvell/cesa/
Dcipher.c42 struct skcipher_request *req) in mv_cesa_skcipher_req_iter_init() argument
44 mv_cesa_req_dma_iter_init(&iter->base, req->cryptlen); in mv_cesa_skcipher_req_iter_init()
45 mv_cesa_sg_dma_iter_init(&iter->src, req->src, DMA_TO_DEVICE); in mv_cesa_skcipher_req_iter_init()
46 mv_cesa_sg_dma_iter_init(&iter->dst, req->dst, DMA_FROM_DEVICE); in mv_cesa_skcipher_req_iter_init()
59 mv_cesa_skcipher_dma_cleanup(struct skcipher_request *req) in mv_cesa_skcipher_dma_cleanup() argument
61 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); in mv_cesa_skcipher_dma_cleanup()
63 if (req->dst != req->src) { in mv_cesa_skcipher_dma_cleanup()
64 dma_unmap_sg(cesa_dev->dev, req->dst, creq->dst_nents, in mv_cesa_skcipher_dma_cleanup()
66 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents, in mv_cesa_skcipher_dma_cleanup()
69 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents, in mv_cesa_skcipher_dma_cleanup()
[all …]
Dhash.c27 struct ahash_request *req) in mv_cesa_ahash_req_iter_init() argument
29 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); in mv_cesa_ahash_req_iter_init()
30 unsigned int len = req->nbytes + creq->cache_ptr; in mv_cesa_ahash_req_iter_init()
36 mv_cesa_sg_dma_iter_init(&iter->src, req->src, DMA_TO_DEVICE); in mv_cesa_ahash_req_iter_init()
49 mv_cesa_ahash_dma_alloc_cache(struct mv_cesa_ahash_dma_req *req, gfp_t flags) in mv_cesa_ahash_dma_alloc_cache() argument
51 req->cache = dma_pool_alloc(cesa_dev->dma->cache_pool, flags, in mv_cesa_ahash_dma_alloc_cache()
52 &req->cache_dma); in mv_cesa_ahash_dma_alloc_cache()
53 if (!req->cache) in mv_cesa_ahash_dma_alloc_cache()
60 mv_cesa_ahash_dma_free_cache(struct mv_cesa_ahash_dma_req *req) in mv_cesa_ahash_dma_free_cache() argument
62 if (!req->cache) in mv_cesa_ahash_dma_free_cache()
[all …]
/drivers/base/power/
Dqos.c151 static int apply_constraint(struct dev_pm_qos_request *req, in apply_constraint() argument
154 struct dev_pm_qos *qos = req->dev->power.qos; in apply_constraint()
157 switch(req->type) { in apply_constraint()
163 &req->data.pnode, action, value); in apply_constraint()
167 &req->data.pnode, action, value); in apply_constraint()
170 req->dev->power.set_latency_tolerance(req->dev, value); in apply_constraint()
175 ret = freq_qos_apply(&req->data.freq, action, value); in apply_constraint()
178 ret = pm_qos_update_flags(&qos->flags, &req->data.flr, in apply_constraint()
250 struct dev_pm_qos_request *req, *tmp; in dev_pm_qos_constraints_destroy() local
274 plist_for_each_entry_safe(req, tmp, &c->list, data.pnode) { in dev_pm_qos_constraints_destroy()
[all …]
/drivers/usb/gadget/udc/aspeed-vhub/
Dep0.c37 struct usb_request *req = &ep->ep0.req.req; in ast_vhub_reply() local
46 if (WARN_ON(req->status == -EINPROGRESS)) in ast_vhub_reply()
49 req->buf = ptr; in ast_vhub_reply()
50 req->length = len; in ast_vhub_reply()
51 req->complete = NULL; in ast_vhub_reply()
52 req->zero = true; in ast_vhub_reply()
60 if (ep->ep.ops->queue(&ep->ep, req, GFP_ATOMIC)) in ast_vhub_reply()
183 struct ast_vhub_req *req) in ast_vhub_ep0_do_send() argument
191 if (req->req.length == 0) in ast_vhub_ep0_do_send()
192 req->last_desc = 1; in ast_vhub_ep0_do_send()
[all …]
/drivers/net/ethernet/broadcom/bnxt/
Dbnxt_sriov.c28 struct hwrm_fwd_async_event_cmpl_input req = {0}; in bnxt_hwrm_fwd_async_event_cmpl() local
32 bnxt_hwrm_cmd_hdr_init(bp, &req, HWRM_FWD_ASYNC_EVENT_CMPL, -1, -1); in bnxt_hwrm_fwd_async_event_cmpl()
34 req.encap_async_event_target_id = cpu_to_le16(vf->fw_fid); in bnxt_hwrm_fwd_async_event_cmpl()
37 req.encap_async_event_target_id = cpu_to_le16(0xffff); in bnxt_hwrm_fwd_async_event_cmpl()
38 async_cmpl = (struct hwrm_async_event_cmpl *)req.encap_async_event_cmpl; in bnxt_hwrm_fwd_async_event_cmpl()
42 rc = hwrm_send_message(bp, &req, sizeof(req), HWRM_CMD_TIMEOUT); in bnxt_hwrm_fwd_async_event_cmpl()
68 struct hwrm_func_cfg_input req = {0}; in bnxt_set_vf_spoofchk() local
95 bnxt_hwrm_cmd_hdr_init(bp, &req, HWRM_FUNC_CFG, -1, -1); in bnxt_set_vf_spoofchk()
96 req.fid = cpu_to_le16(vf->fw_fid); in bnxt_set_vf_spoofchk()
97 req.flags = cpu_to_le32(func_flags); in bnxt_set_vf_spoofchk()
[all …]
/drivers/tee/optee/
Dsupp.c41 struct optee_supp_req *req; in optee_supp_release() local
47 idr_for_each_entry(&supp->idr, req, id) { in optee_supp_release()
49 req->ret = TEEC_ERROR_COMMUNICATION; in optee_supp_release()
50 complete(&req->c); in optee_supp_release()
54 list_for_each_entry_safe(req, req_tmp, &supp->reqs, link) { in optee_supp_release()
55 list_del(&req->link); in optee_supp_release()
56 req->in_queue = false; in optee_supp_release()
57 req->ret = TEEC_ERROR_COMMUNICATION; in optee_supp_release()
58 complete(&req->c); in optee_supp_release()
82 struct optee_supp_req *req; in optee_supp_thrd_req() local
[all …]

12345678910>>...47