/kernel/linux/linux-5.10/block/ |
D | bsg.c | 68 struct scsi_request *sreq = scsi_req(rq); in bsg_scsi_fill_hdr() local 75 sreq->cmd_len = hdr->request_len; in bsg_scsi_fill_hdr() 76 if (sreq->cmd_len > BLK_MAX_CDB) { in bsg_scsi_fill_hdr() 77 sreq->cmd = kzalloc(sreq->cmd_len, GFP_KERNEL); in bsg_scsi_fill_hdr() 78 if (!sreq->cmd) in bsg_scsi_fill_hdr() 82 if (copy_from_user(sreq->cmd, uptr64(hdr->request), sreq->cmd_len)) in bsg_scsi_fill_hdr() 84 if (blk_verify_command(sreq->cmd, mode)) in bsg_scsi_fill_hdr() 91 struct scsi_request *sreq = scsi_req(rq); in bsg_scsi_complete_rq() local 97 hdr->device_status = sreq->result & 0xff; in bsg_scsi_complete_rq() 98 hdr->transport_status = host_byte(sreq->result); in bsg_scsi_complete_rq() [all …]
|
/kernel/linux/linux-5.10/drivers/crypto/marvell/cesa/ |
D | cipher.c | 86 struct mv_cesa_skcipher_std_req *sreq = &creq->std; in mv_cesa_skcipher_std_step() local 88 size_t len = min_t(size_t, req->cryptlen - sreq->offset, in mv_cesa_skcipher_std_step() 91 mv_cesa_adjust_op(engine, &sreq->op); in mv_cesa_skcipher_std_step() 92 memcpy_toio(engine->sram, &sreq->op, sizeof(sreq->op)); in mv_cesa_skcipher_std_step() 96 len, sreq->offset); in mv_cesa_skcipher_std_step() 98 sreq->size = len; in mv_cesa_skcipher_std_step() 99 mv_cesa_set_crypt_op_len(&sreq->op, len); in mv_cesa_skcipher_std_step() 102 if (!sreq->skip_ctx) { in mv_cesa_skcipher_std_step() 103 memcpy_toio(engine->sram, &sreq->op, sizeof(sreq->op)); in mv_cesa_skcipher_std_step() 104 sreq->skip_ctx = true; in mv_cesa_skcipher_std_step() [all …]
|
D | hash.c | 160 struct mv_cesa_ahash_std_req *sreq = &creq->req.std; in mv_cesa_ahash_std_step() local 172 if (!sreq->offset) { in mv_cesa_ahash_std_step() 183 len = min_t(size_t, req->nbytes + creq->cache_ptr - sreq->offset, in mv_cesa_ahash_std_step() 192 sreq->offset += sg_pcopy_to_buffer(req->src, creq->src_nents, in mv_cesa_ahash_std_step() 197 sreq->offset); in mv_cesa_ahash_std_step() 203 if (creq->last_req && sreq->offset == req->nbytes && in mv_cesa_ahash_std_step() 263 struct mv_cesa_ahash_std_req *sreq = &creq->req.std; in mv_cesa_ahash_std_process() local 265 if (sreq->offset < (req->nbytes - creq->cache_ptr)) in mv_cesa_ahash_std_process() 282 struct mv_cesa_ahash_std_req *sreq = &creq->req.std; in mv_cesa_ahash_std_prepare() local 284 sreq->offset = 0; in mv_cesa_ahash_std_prepare()
|
/kernel/linux/linux-5.10/drivers/crypto/inside-secure/ |
D | safexcel_cipher.c | 507 struct safexcel_cipher_req *sreq, in safexcel_context_control() argument 529 (sreq->direction == SAFEXCEL_ENCRYPT ? in safexcel_context_control() 544 if (sreq->direction == SAFEXCEL_ENCRYPT && in safexcel_context_control() 549 else if (sreq->direction == SAFEXCEL_ENCRYPT) in safexcel_context_control() 559 if (sreq->direction == SAFEXCEL_ENCRYPT) in safexcel_context_control() 612 struct safexcel_cipher_req *sreq, in safexcel_handle_req_result() argument 623 if (unlikely(!sreq->rdescs)) in safexcel_handle_req_result() 626 while (sreq->rdescs--) { in safexcel_handle_req_result() 644 dma_unmap_sg(priv->dev, src, sreq->nr_src, DMA_BIDIRECTIONAL); in safexcel_handle_req_result() 646 dma_unmap_sg(priv->dev, src, sreq->nr_src, DMA_TO_DEVICE); in safexcel_handle_req_result() [all …]
|
D | safexcel_hash.c | 232 struct safexcel_ahash_req *sreq = ahash_request_ctx(areq); in safexcel_handle_req_result() local 249 if (sreq->nents) { in safexcel_handle_req_result() 250 dma_unmap_sg(priv->dev, areq->src, sreq->nents, DMA_TO_DEVICE); in safexcel_handle_req_result() 251 sreq->nents = 0; in safexcel_handle_req_result() 254 if (sreq->result_dma) { in safexcel_handle_req_result() 255 dma_unmap_single(priv->dev, sreq->result_dma, sreq->digest_sz, in safexcel_handle_req_result() 257 sreq->result_dma = 0; in safexcel_handle_req_result() 260 if (sreq->cache_dma) { in safexcel_handle_req_result() 261 dma_unmap_single(priv->dev, sreq->cache_dma, sreq->cache_sz, in safexcel_handle_req_result() 263 sreq->cache_dma = 0; in safexcel_handle_req_result() [all …]
|
/kernel/liteos_a/syscall/ |
D | time_syscall.c | 334 struct timespec sreq; in SysClockNanoSleep() local 337 if (!req || LOS_ArchCopyFromUser(&sreq, req, sizeof(struct timespec))) { in SysClockNanoSleep() 342 ret = clock_nanosleep(clk, flags, &sreq, rem ? &srem : NULL); in SysClockNanoSleep() 494 struct timespec64 sreq; in SysClockNanoSleep64() local 497 if (!req || LOS_ArchCopyFromUser(&sreq, req, sizeof(struct timespec64))) { in SysClockNanoSleep64() 503 rq.tv_sec = (sreq.tv_sec > UINT32_MAX) ? UINT32_MAX : sreq.tv_sec; in SysClockNanoSleep64() 504 rq.tv_nsec = sreq.tv_nsec; in SysClockNanoSleep64()
|
/kernel/linux/linux-5.10/drivers/net/wireless/mediatek/mt76/mt7615/ |
D | mcu.c | 3081 struct cfg80211_scan_request *sreq = &scan_req->req; in mt7615_mcu_hw_scan() local 3083 int ext_channels_num = max_t(int, sreq->n_channels - 32, 0); in mt7615_mcu_hw_scan() 3084 struct ieee80211_channel **scan_list = sreq->channels; in mt7615_mcu_hw_scan() 3106 req->scan_type = sreq->n_ssids ? 1 : 0; in mt7615_mcu_hw_scan() 3107 req->probe_req_num = sreq->n_ssids ? 2 : 0; in mt7615_mcu_hw_scan() 3110 for (i = 0; i < sreq->n_ssids; i++) { in mt7615_mcu_hw_scan() 3111 if (!sreq->ssids[i].ssid_len) in mt7615_mcu_hw_scan() 3114 req->ssids[i].ssid_len = cpu_to_le32(sreq->ssids[i].ssid_len); in mt7615_mcu_hw_scan() 3115 memcpy(req->ssids[i].ssid, sreq->ssids[i].ssid, in mt7615_mcu_hw_scan() 3116 sreq->ssids[i].ssid_len); in mt7615_mcu_hw_scan() [all …]
|
D | mt7615.h | 611 struct cfg80211_sched_scan_request *sreq);
|
/kernel/linux/linux-5.10/drivers/crypto/marvell/octeontx/ |
D | otx_cptvf_algs.c | 146 struct skcipher_request *sreq; in output_iv_copyback() local 152 sreq = container_of(areq, struct skcipher_request, base); in output_iv_copyback() 153 stfm = crypto_skcipher_reqtfm(sreq); in output_iv_copyback() 157 rctx = skcipher_request_ctx(sreq); in output_iv_copyback() 160 start = sreq->cryptlen - ivsize; in output_iv_copyback() 163 scatterwalk_map_and_copy(sreq->iv, sreq->dst, start, in output_iv_copyback() 166 if (sreq->src != sreq->dst) { in output_iv_copyback() 167 scatterwalk_map_and_copy(sreq->iv, sreq->src, in output_iv_copyback() 170 memcpy(sreq->iv, req_info->iv_out, ivsize); in output_iv_copyback()
|
/kernel/linux/linux-5.10/drivers/crypto/hisilicon/sec2/ |
D | sec_crypto.c | 1408 static int sec_skcipher_param_check(struct sec_ctx *ctx, struct sec_req *sreq) in sec_skcipher_param_check() argument 1410 struct skcipher_request *sk_req = sreq->c_req.sk_req; in sec_skcipher_param_check() 1418 sreq->c_req.c_len = sk_req->cryptlen; in sec_skcipher_param_check() 1421 sreq->use_pbuf = true; in sec_skcipher_param_check() 1423 sreq->use_pbuf = false; in sec_skcipher_param_check() 1532 static int sec_aead_param_check(struct sec_ctx *ctx, struct sec_req *sreq) in sec_aead_param_check() argument 1534 struct aead_request *req = sreq->aead_req.aead_req; in sec_aead_param_check() 1548 sreq->use_pbuf = true; in sec_aead_param_check() 1550 sreq->use_pbuf = false; in sec_aead_param_check() 1558 if (sreq->c_req.encrypt) in sec_aead_param_check() [all …]
|
/kernel/linux/linux-5.10/include/linux/netfilter/ |
D | nf_conntrack_pptp.h | 288 struct PptpStartSessionRequest sreq; member
|
/kernel/linux/linux-5.10/arch/s390/kernel/ |
D | perf_cpum_sf.c | 107 struct hws_lsctl_request_block sreq; in sf_disable() local 109 memset(&sreq, 0, sizeof(sreq)); in sf_disable() 110 return lsctl(&sreq); in sf_disable()
|
/kernel/linux/linux-5.10/drivers/crypto/qat/qat_common/ |
D | qat_algs.c | 819 struct skcipher_request *sreq = qat_req->skcipher_req; in qat_skcipher_alg_callback() local 828 memcpy(sreq->iv, qat_req->iv, AES_BLOCK_SIZE); in qat_skcipher_alg_callback() 832 sreq->base.complete(&sreq->base, res); in qat_skcipher_alg_callback()
|
/kernel/linux/linux-5.10/drivers/ide/ |
D | ide-probe.c | 751 scsi_req_init(&req->sreq); in ide_initialize_rq() 752 req->sreq.sense = req->sense; in ide_initialize_rq()
|
/kernel/linux/linux-5.10/include/linux/ |
D | ide.h | 50 struct scsi_request sreq; member
|