Lines Matching refs:IV
74 #define IV AES_BLOCK_SIZE macro
840 chcr_req->sec_cpl.pldlen = htonl(IV + wrparam->bytes); in create_cipher_wr()
842 FILL_SEC_CPL_CIPHERSTOP_HI(0, 0, IV + 1, 0); in create_cipher_wr()
848 0, 0, IV >> 1); in create_cipher_wr()
880 temp = sizeof(struct cpl_rx_phys_dsgl) + dst_size + kctx_len + IV in create_cipher_wr()
1199 memcpy(req->iv, reqctx->init_iv, IV); in chcr_handle_cipher_resp()
1290 reqctx->imm = (transhdr_len + IV + req->cryptlen) <= in process_cipher()
1292 bytes = IV + req->cryptlen; in process_cipher()
1320 memcpy(reqctx->init_iv, reqctx->iv, IV); in process_cipher()
1324 memcpy(reqctx->iv, req->iv, IV); in process_cipher()
1325 memcpy(reqctx->init_iv, req->iv, IV); in process_cipher()
2386 reqctx->scratch_pad = reqctx->iv + IV; in chcr_aead_common_init()
2510 chcr_req->sec_cpl.pldlen = htonl(req->assoclen + IV + req->cryptlen); in create_authenc_wr()
2512 null ? 0 : 1 + IV, in create_authenc_wr()
2513 null ? 0 : IV + req->assoclen, in create_authenc_wr()
2514 req->assoclen + IV + 1, in create_authenc_wr()
2518 null ? 0 : req->assoclen + IV + 1, in create_authenc_wr()
2529 IV >> 1); in create_authenc_wr()
2547 ulptx = (struct ulptx_sgl *)(ivptr + IV); in create_authenc_wr()
2556 memcpy(ivptr, req->iv, IV); in create_authenc_wr()
2561 temp = sizeof(struct cpl_rx_phys_dsgl) + dst_size + IV + in create_authenc_wr()
2599 reqctx->iv_dma = dma_map_single(dev, reqctx->iv, (IV + reqctx->b0_len), in chcr_aead_dma_map()
2604 reqctx->b0_dma = reqctx->iv_dma + IV; in chcr_aead_dma_map()
2632 dma_unmap_single(dev, reqctx->iv_dma, IV, DMA_BIDIRECTIONAL); in chcr_aead_dma_map()
2661 dma_unmap_single(dev, reqctx->iv_dma, (IV + reqctx->b0_len), in chcr_aead_dma_unmap()
2718 dsgl_walk_add_page(&dsgl_walk, IV + reqctx->b0_len, reqctx->iv_dma); in chcr_add_aead_dst_ent()
2733 memcpy(buf, reqctx->iv, IV); in chcr_add_cipher_src_ent()
2734 buf += IV; in chcr_add_cipher_src_ent()
2984 (req->assoclen + IV + 1 + ccm_xtra) : 0; in fill_sec_cpl_for_aead()
2994 htonl(req->assoclen + IV + req->cryptlen + ccm_xtra); in fill_sec_cpl_for_aead()
2997 1 + IV, IV + assoclen + ccm_xtra, in fill_sec_cpl_for_aead()
2998 req->assoclen + IV + 1 + ccm_xtra, 0); in fill_sec_cpl_for_aead()
3007 aeadctx->hmac_ctrl, IV >> 1); in fill_sec_cpl_for_aead()
3107 ulptx = (struct ulptx_sgl *)(ivptr + IV); in create_aead_ccm_wr()
3115 temp = sizeof(struct cpl_rx_phys_dsgl) + dst_size + IV + in create_aead_ccm_wr()
3196 htonl(req->assoclen + IV + req->cryptlen); in create_gcm_wr()
3198 assoclen ? 1 + IV : 0, in create_gcm_wr()
3199 assoclen ? IV + assoclen : 0, in create_gcm_wr()
3200 req->assoclen + IV + 1, 0); in create_gcm_wr()
3202 FILL_SEC_CPL_AUTHINSERT(0, req->assoclen + IV + 1, in create_gcm_wr()
3209 aeadctx->hmac_ctrl, IV >> 1); in create_gcm_wr()
3234 temp = sizeof(struct cpl_rx_phys_dsgl) + dst_size + IV + in create_gcm_wr()