/arch/powerpc/crypto/ |
D | sha1.c | 46 unsigned int partial, done; in sha1_update() local 49 partial = sctx->count & 0x3f; in sha1_update() 54 if ((partial + len) > 63) { in sha1_update() 57 if (partial) { in sha1_update() 58 done = -partial; in sha1_update() 59 memcpy(sctx->buffer + partial, data, done + 64); in sha1_update() 70 partial = 0; in sha1_update() 72 memcpy(sctx->buffer + partial, src, len - done); in sha1_update()
|
/arch/sparc/crypto/ |
D | sha1_glue.c | 41 unsigned int len, unsigned int partial) in __sha1_sparc64_update() argument 46 if (partial) { in __sha1_sparc64_update() 47 done = SHA1_BLOCK_SIZE - partial; in __sha1_sparc64_update() 48 memcpy(sctx->buffer + partial, data, done); in __sha1_sparc64_update() 65 unsigned int partial = sctx->count % SHA1_BLOCK_SIZE; in sha1_sparc64_update() local 68 if (partial + len < SHA1_BLOCK_SIZE) { in sha1_sparc64_update() 70 memcpy(sctx->buffer + partial, data, len); in sha1_sparc64_update() 72 __sha1_sparc64_update(sctx, data, len, partial); in sha1_sparc64_update()
|
D | md5_glue.c | 46 unsigned int len, unsigned int partial) in __md5_sparc64_update() argument 51 if (partial) { in __md5_sparc64_update() 52 done = MD5_HMAC_BLOCK_SIZE - partial; in __md5_sparc64_update() 53 memcpy((u8 *)sctx->block + partial, data, done); in __md5_sparc64_update() 70 unsigned int partial = sctx->byte_count % MD5_HMAC_BLOCK_SIZE; in md5_sparc64_update() local 73 if (partial + len < MD5_HMAC_BLOCK_SIZE) { in md5_sparc64_update() 75 memcpy((u8 *)sctx->block + partial, data, len); in md5_sparc64_update() 77 __md5_sparc64_update(sctx, data, len, partial); in md5_sparc64_update()
|
D | sha256_glue.c | 62 unsigned int len, unsigned int partial) in __sha256_sparc64_update() argument 67 if (partial) { in __sha256_sparc64_update() 68 done = SHA256_BLOCK_SIZE - partial; in __sha256_sparc64_update() 69 memcpy(sctx->buf + partial, data, done); in __sha256_sparc64_update() 86 unsigned int partial = sctx->count % SHA256_BLOCK_SIZE; in sha256_sparc64_update() local 89 if (partial + len < SHA256_BLOCK_SIZE) { in sha256_sparc64_update() 91 memcpy(sctx->buf + partial, data, len); in sha256_sparc64_update() 93 __sha256_sparc64_update(sctx, data, len, partial); in sha256_sparc64_update()
|
D | sha512_glue.c | 61 unsigned int len, unsigned int partial) in __sha512_sparc64_update() argument 67 if (partial) { in __sha512_sparc64_update() 68 done = SHA512_BLOCK_SIZE - partial; in __sha512_sparc64_update() 69 memcpy(sctx->buf + partial, data, done); in __sha512_sparc64_update() 86 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_sparc64_update() local 89 if (partial + len < SHA512_BLOCK_SIZE) { in sha512_sparc64_update() 92 memcpy(sctx->buf + partial, data, len); in sha512_sparc64_update() 94 __sha512_sparc64_update(sctx, data, len, partial); in sha512_sparc64_update()
|
/arch/x86/include/asm/ |
D | unwind.h | 68 bool *partial) in unwind_get_entry_regs() argument 73 if (partial) { in unwind_get_entry_regs() 75 *partial = !state->full_regs; in unwind_get_entry_regs() 77 *partial = false; in unwind_get_entry_regs() 85 bool *partial) in unwind_get_entry_regs() argument
|
/arch/mips/cavium-octeon/crypto/ |
D | octeon-sha1.c | 95 unsigned int partial; in __octeon_sha1_update() local 99 partial = sctx->count % SHA1_BLOCK_SIZE; in __octeon_sha1_update() 104 if ((partial + len) >= SHA1_BLOCK_SIZE) { in __octeon_sha1_update() 105 if (partial) { in __octeon_sha1_update() 106 done = -partial; in __octeon_sha1_update() 107 memcpy(sctx->buffer + partial, data, in __octeon_sha1_update() 118 partial = 0; in __octeon_sha1_update() 120 memcpy(sctx->buffer + partial, src, len - done); in __octeon_sha1_update()
|
D | octeon-sha256.c | 107 unsigned int partial; in __octeon_sha256_update() local 111 partial = sctx->count % SHA256_BLOCK_SIZE; in __octeon_sha256_update() 116 if ((partial + len) >= SHA256_BLOCK_SIZE) { in __octeon_sha256_update() 117 if (partial) { in __octeon_sha256_update() 118 done = -partial; in __octeon_sha256_update() 119 memcpy(sctx->buf + partial, data, in __octeon_sha256_update() 130 partial = 0; in __octeon_sha256_update() 132 memcpy(sctx->buf + partial, src, len - done); in __octeon_sha256_update()
|
/arch/arm/crypto/ |
D | ghash-ce-glue.c | 69 unsigned int partial = ctx->count % GHASH_BLOCK_SIZE; in ghash_update() local 73 if ((partial + len) >= GHASH_BLOCK_SIZE) { in ghash_update() 77 if (partial) { in ghash_update() 78 int p = GHASH_BLOCK_SIZE - partial; in ghash_update() 80 memcpy(ctx->buf + partial, src, p); in ghash_update() 90 partial ? ctx->buf : NULL); in ghash_update() 93 partial = 0; in ghash_update() 96 memcpy(ctx->buf + partial, src, len); in ghash_update() 103 unsigned int partial = ctx->count % GHASH_BLOCK_SIZE; in ghash_final() local 105 if (partial) { in ghash_final() [all …]
|
/arch/x86/kernel/ |
D | dumpstack.c | 81 bool partial) in show_regs_if_on_stack() argument 92 if (!partial && on_stack(info, regs, sizeof(*regs))) { in show_regs_if_on_stack() 95 } else if (partial && on_stack(info, (void *)regs + IRET_FRAME_OFFSET, in show_regs_if_on_stack() 113 bool partial = false; in show_trace_log_lvl() local 119 regs = unwind_get_entry_regs(&state, &partial); in show_trace_log_lvl() 157 show_regs_if_on_stack(&stack_info, regs, partial); in show_trace_log_lvl() 216 regs = unwind_get_entry_regs(&state, &partial); in show_trace_log_lvl() 218 show_regs_if_on_stack(&stack_info, regs, partial); in show_trace_log_lvl()
|
/arch/arm64/crypto/ |
D | ghash-ce-glue.c | 119 unsigned int partial = ctx->count % GHASH_BLOCK_SIZE; in ghash_update() local 123 if ((partial + len) >= GHASH_BLOCK_SIZE) { in ghash_update() 127 if (partial) { in ghash_update() 128 int p = GHASH_BLOCK_SIZE - partial; in ghash_update() 130 memcpy(ctx->buf + partial, src, p); in ghash_update() 139 partial ? ctx->buf : NULL); in ghash_update() 142 partial = 0; in ghash_update() 145 memcpy(ctx->buf + partial, src, len); in ghash_update() 152 unsigned int partial = ctx->count % GHASH_BLOCK_SIZE; in ghash_final() local 154 if (partial) { in ghash_final() [all …]
|
/arch/x86/purgatory/ |
D | sha256.c | 228 unsigned int partial, done; in sha256_update() local 231 partial = sctx->count & 0x3f; in sha256_update() 236 if ((partial + len) > 63) { in sha256_update() 237 if (partial) { in sha256_update() 238 done = -partial; in sha256_update() 239 memcpy(sctx->buf + partial, data, done + 64); in sha256_update() 249 partial = 0; in sha256_update() 251 memcpy(sctx->buf + partial, src, len - done); in sha256_update()
|
/arch/c6x/lib/ |
D | mpyll.S | 23 ;; First compute partial results using 32-bit parts of x and y:
|
/arch/alpha/lib/ |
D | ev6-memset.S | 81 ldq_u $4,0($16) # L : Fetch first partial 259 ldq_u $4,0($16) # L : Fetch first partial 447 ldq_u $4,0($16) # L : Fetch first partial
|
D | memchr.S | 121 # last quad may or may not be partial).
|
D | stxncpy.S | 175 cmpbge zero, t2, t8 # e0 : find nulls in second partial 231 or t0, t1, t0 # e0 : first (partial) source word complete
|
D | ev6-stxncpy.S | 217 cmpbge zero, t2, t8 # E : find nulls in second partial 275 or t0, t1, t0 # E : first (partial) source word complete
|
D | ev6-memchr.S | 140 # last quad may or may not be partial).
|
D | stxcpy.S | 200 or t0, t1, t1 # e1 : first (partial) source word complete
|
D | ev6-stxcpy.S | 228 or t0, t1, t1 # E : first (partial) source word complete (stall)
|
/arch/mips/kernel/ |
D | entry.S | 95 restore_partial: # restore partial frame
|
/arch/m68k/fpsp040/ |
D | decbin.S | 153 mulul #TEN,%d1 |mul partial product by one digit place
|
D | setox.S | 473 movew 4(%a0),%d0 | ...expo. and partial sig. of |X| 640 movew 4(%a0),%d0 | ...expo. and partial sig. of |X|
|
/arch/x86/crypto/ |
D | poly1305-avx2-x86_64.S | 326 # Now do a partial reduction mod (2^130)-5, carrying h0 -> h1 -> h2 ->
|
D | poly1305-sse2-x86_64.S | 523 # Now do a partial reduction mod (2^130)-5, carrying h0 -> h1 -> h2 ->
|