/drivers/md/ |
D | dm-switch.c | 62 struct switch_ctx *sctx; in alloc_switch_ctx() local 64 sctx = kzalloc(sizeof(struct switch_ctx) + nr_paths * sizeof(struct switch_path), in alloc_switch_ctx() 66 if (!sctx) in alloc_switch_ctx() 69 sctx->ti = ti; in alloc_switch_ctx() 70 sctx->region_size = region_size; in alloc_switch_ctx() 72 ti->private = sctx; in alloc_switch_ctx() 74 return sctx; in alloc_switch_ctx() 79 struct switch_ctx *sctx = ti->private; in alloc_region_table() local 83 if (!(sctx->region_size & (sctx->region_size - 1))) in alloc_region_table() 84 sctx->region_size_bits = __ffs(sctx->region_size); in alloc_region_table() [all …]
|
/drivers/crypto/nx/ |
D | nx-sha256.c | 51 struct sha256_state *sctx = shash_desc_ctx(desc); in nx_sha256_init() local 53 memset(sctx, 0, sizeof *sctx); in nx_sha256_init() 55 sctx->state[0] = __cpu_to_be32(SHA256_H0); in nx_sha256_init() 56 sctx->state[1] = __cpu_to_be32(SHA256_H1); in nx_sha256_init() 57 sctx->state[2] = __cpu_to_be32(SHA256_H2); in nx_sha256_init() 58 sctx->state[3] = __cpu_to_be32(SHA256_H3); in nx_sha256_init() 59 sctx->state[4] = __cpu_to_be32(SHA256_H4); in nx_sha256_init() 60 sctx->state[5] = __cpu_to_be32(SHA256_H5); in nx_sha256_init() 61 sctx->state[6] = __cpu_to_be32(SHA256_H6); in nx_sha256_init() 62 sctx->state[7] = __cpu_to_be32(SHA256_H7); in nx_sha256_init() [all …]
|
D | nx-sha512.c | 51 struct sha512_state *sctx = shash_desc_ctx(desc); in nx_sha512_init() local 53 memset(sctx, 0, sizeof *sctx); in nx_sha512_init() 55 sctx->state[0] = __cpu_to_be64(SHA512_H0); in nx_sha512_init() 56 sctx->state[1] = __cpu_to_be64(SHA512_H1); in nx_sha512_init() 57 sctx->state[2] = __cpu_to_be64(SHA512_H2); in nx_sha512_init() 58 sctx->state[3] = __cpu_to_be64(SHA512_H3); in nx_sha512_init() 59 sctx->state[4] = __cpu_to_be64(SHA512_H4); in nx_sha512_init() 60 sctx->state[5] = __cpu_to_be64(SHA512_H5); in nx_sha512_init() 61 sctx->state[6] = __cpu_to_be64(SHA512_H6); in nx_sha512_init() 62 sctx->state[7] = __cpu_to_be64(SHA512_H7); in nx_sha512_init() [all …]
|
D | nx-aes-xcbc.c | 172 struct xcbc_state *sctx = shash_desc_ctx(desc); in nx_xcbc_init() local 174 memset(sctx, 0, sizeof *sctx); in nx_xcbc_init() 183 struct xcbc_state *sctx = shash_desc_ctx(desc); in nx_xcbc_update() local 197 total = sctx->count + len; in nx_xcbc_update() 204 memcpy(sctx->buffer + sctx->count, data, len); in nx_xcbc_update() 205 sctx->count += len; in nx_xcbc_update() 216 out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *)sctx->state, in nx_xcbc_update() 242 if (sctx->count) { in nx_xcbc_update() 243 data_len = sctx->count; in nx_xcbc_update() 245 (u8 *) sctx->buffer, in nx_xcbc_update() [all …]
|
/drivers/staging/skein/ |
D | skein_generic.c | 44 struct skein_256_ctx *sctx = shash_desc_ctx(desc); in skein256_export() local 46 memcpy(out, sctx, sizeof(*sctx)); in skein256_export() 52 struct skein_256_ctx *sctx = shash_desc_ctx(desc); in skein256_import() local 54 memcpy(sctx, in, sizeof(*sctx)); in skein256_import() 79 struct skein_512_ctx *sctx = shash_desc_ctx(desc); in skein512_export() local 81 memcpy(out, sctx, sizeof(*sctx)); in skein512_export() 87 struct skein_512_ctx *sctx = shash_desc_ctx(desc); in skein512_import() local 89 memcpy(sctx, in, sizeof(*sctx)); in skein512_import() 114 struct skein_1024_ctx *sctx = shash_desc_ctx(desc); in skein1024_export() local 116 memcpy(out, sctx, sizeof(*sctx)); in skein1024_export() [all …]
|
/drivers/crypto/ |
D | padlock-sha.c | 296 struct sha1_state *sctx = shash_desc_ctx(desc); in padlock_sha1_init_nano() local 298 *sctx = (struct sha1_state){ in padlock_sha1_init_nano() 308 struct sha1_state *sctx = shash_desc_ctx(desc); in padlock_sha1_update_nano() local 317 partial = sctx->count & 0x3f; in padlock_sha1_update_nano() 318 sctx->count += len; in padlock_sha1_update_nano() 321 memcpy(dst, (u8 *)(sctx->state), SHA1_DIGEST_SIZE); in padlock_sha1_update_nano() 328 memcpy(sctx->buffer + partial, data, in padlock_sha1_update_nano() 330 src = sctx->buffer; in padlock_sha1_update_nano() 353 memcpy((u8 *)(sctx->state), dst, SHA1_DIGEST_SIZE); in padlock_sha1_update_nano() 354 memcpy(sctx->buffer + partial, src, len - done); in padlock_sha1_update_nano() [all …]
|
/drivers/staging/rtl8188eu/core/ |
D | rtw_xmit.c | 1212 if (pxmitbuf->sctx) { in rtw_alloc_xmitbuf_ext() 1214 rtw_sctx_done_err(&pxmitbuf->sctx, RTW_SCTX_DONE_BUF_ALLOC); in rtw_alloc_xmitbuf_ext() 1259 if (pxmitbuf->sctx) { in rtw_alloc_xmitbuf() 1261 rtw_sctx_done_err(&pxmitbuf->sctx, RTW_SCTX_DONE_BUF_ALLOC); in rtw_alloc_xmitbuf() 1277 if (pxmitbuf->sctx) { in rtw_free_xmitbuf() 1279 rtw_sctx_done_err(&pxmitbuf->sctx, RTW_SCTX_DONE_BUF_FREE); in rtw_free_xmitbuf() 2069 void rtw_sctx_init(struct submit_ctx *sctx, int timeout_ms) in rtw_sctx_init() argument 2071 sctx->timeout_ms = timeout_ms; in rtw_sctx_init() 2072 sctx->submit_time = jiffies; in rtw_sctx_init() 2073 init_completion(&sctx->done); in rtw_sctx_init() [all …]
|
D | rtw_mlme_ext.c | 244 struct submit_ctx sctx; in dump_mgntframe_and_wait() local 249 rtw_sctx_init(&sctx, timeout_ms); in dump_mgntframe_and_wait() 250 pxmitbuf->sctx = &sctx; in dump_mgntframe_and_wait() 255 ret = rtw_sctx_wait(&sctx); in dump_mgntframe_and_wait()
|
/drivers/staging/rtl8188eu/include/ |
D | rtw_xmit.h | 191 void rtw_sctx_init(struct submit_ctx *sctx, int timeout_ms); 192 int rtw_sctx_wait(struct submit_ctx *sctx); 193 void rtw_sctx_done_err(struct submit_ctx **sctx, int status); 205 struct submit_ctx *sctx; member
|
/drivers/staging/rtl8188eu/os_dep/ |
D | usb_ops_linux.c | 683 rtw_sctx_done_err(&pxmitbuf->sctx, in usb_write_port_complete() 711 rtw_sctx_done_err(&xmitbuf->sctx, RTW_SCTX_DONE_TX_DENY); in usb_write_port() 757 rtw_sctx_done_err(&xmitbuf->sctx, RTW_SCTX_DONE_WRITE_PORT_ERR); in usb_write_port()
|
/drivers/staging/rtl8188eu/hal/ |
D | rtl8188eu_xmit.c | 399 rtw_sctx_done_err(&pxmitbuf->sctx, RTW_SCTX_DONE_UNKNOWN); in rtw_dump_xframe()
|