/arch/arc/kernel/ |
D | disasm.c | 24 void __kprobes disasm_instr(unsigned long addr, struct disasm_state *state, in disasm_instr() argument 35 memset(state, 0, sizeof(struct disasm_state)); in disasm_instr() 51 state->major_opcode = (word1 >> 11) & 0x1F; in disasm_instr() 54 if (state->major_opcode < 0x0B) { in disasm_instr() 57 state->instr_len = 4; in disasm_instr() 59 state->words[0] = (word1 << 16) | word0; in disasm_instr() 61 state->instr_len = 2; in disasm_instr() 62 state->words[0] = word1; in disasm_instr() 66 word1 = *((uint16_t *)(addr + state->instr_len)); in disasm_instr() 67 word0 = *((uint16_t *)(addr + state->instr_len + 2)); in disasm_instr() [all …]
|
D | unaligned.c | 136 static void fixup_load(struct disasm_state *state, struct pt_regs *regs, in fixup_load() argument 142 if ((state->aa == 1) || (state->aa == 2)) { in fixup_load() 143 set_reg(state->wb_reg, state->src1 + state->src2, regs, cregs); in fixup_load() 145 if (state->aa == 2) in fixup_load() 146 state->src2 = 0; in fixup_load() 149 if (state->zz == 0) { in fixup_load() 150 get32_unaligned_check(val, state->src1 + state->src2); in fixup_load() 152 get16_unaligned_check(val, state->src1 + state->src2); in fixup_load() 154 if (state->x) in fixup_load() 158 if (state->pref == 0) in fixup_load() [all …]
|
D | unwind.c | 667 static int advance_loc(unsigned long delta, struct unwind_state *state) in advance_loc() argument 669 state->loc += delta * state->codeAlign; in advance_loc() 674 unw_debug("delta %3lu => loc 0x%lx: ", delta, state->loc); in advance_loc() 679 struct unwind_state *state) in set_rule() argument 681 if (reg < ARRAY_SIZE(state->regs)) { in set_rule() 682 state->regs[reg].where = where; in set_rule() 683 state->regs[reg].value = value; in set_rule() 708 signed ptrType, struct unwind_state *state) in processCFI() argument 718 if (start != state->cieStart) { in processCFI() 719 state->loc = state->org; in processCFI() [all …]
|
/arch/x86/kernel/ |
D | unwind_frame.c | 13 unsigned long unwind_get_return_address(struct unwind_state *state) in unwind_get_return_address() argument 15 if (unwind_done(state)) in unwind_get_return_address() 18 return __kernel_text_address(state->ip) ? state->ip : 0; in unwind_get_return_address() 22 unsigned long *unwind_get_return_address_ptr(struct unwind_state *state) in unwind_get_return_address_ptr() argument 24 if (unwind_done(state)) in unwind_get_return_address_ptr() 27 return state->regs ? &state->regs->ip : state->bp + 1; in unwind_get_return_address_ptr() 30 static void unwind_dump(struct unwind_state *state) in unwind_dump() argument 44 state->stack_info.type, state->stack_info.next_sp, in unwind_dump() 45 state->stack_mask, state->graph_idx); in unwind_dump() 47 for (sp = PTR_ALIGN(state->orig_sp, sizeof(long)); sp; in unwind_dump() [all …]
|
D | unwind_guess.c | 8 unsigned long unwind_get_return_address(struct unwind_state *state) in unwind_get_return_address() argument 12 if (unwind_done(state)) in unwind_get_return_address() 15 addr = READ_ONCE_NOCHECK(*state->sp); in unwind_get_return_address() 17 return ftrace_graph_ret_addr(state->task, &state->graph_idx, in unwind_get_return_address() 18 addr, state->sp); in unwind_get_return_address() 22 unsigned long *unwind_get_return_address_ptr(struct unwind_state *state) in unwind_get_return_address_ptr() argument 27 bool unwind_next_frame(struct unwind_state *state) in unwind_next_frame() argument 29 struct stack_info *info = &state->stack_info; in unwind_next_frame() 31 if (unwind_done(state)) in unwind_next_frame() 35 for (state->sp++; state->sp < info->end; state->sp++) { in unwind_next_frame() [all …]
|
D | unwind_orc.c | 250 unsigned long unwind_get_return_address(struct unwind_state *state) in unwind_get_return_address() argument 252 if (unwind_done(state)) in unwind_get_return_address() 255 return __kernel_text_address(state->ip) ? state->ip : 0; in unwind_get_return_address() 259 unsigned long *unwind_get_return_address_ptr(struct unwind_state *state) in unwind_get_return_address_ptr() argument 261 if (unwind_done(state)) in unwind_get_return_address_ptr() 264 if (state->regs) in unwind_get_return_address_ptr() 265 return &state->regs->ip; in unwind_get_return_address_ptr() 267 if (state->sp) in unwind_get_return_address_ptr() 268 return (unsigned long *)state->sp - 1; in unwind_get_return_address_ptr() 273 static bool stack_access_ok(struct unwind_state *state, unsigned long _addr, in stack_access_ok() argument [all …]
|
/arch/s390/kernel/ |
D | fpu.c | 16 void __kernel_fpu_begin(struct kernel_fpu *state, u32 flags) in __kernel_fpu_begin() argument 22 flags &= state->mask; in __kernel_fpu_begin() 26 asm volatile("stfpc %0" : "=m" (state->fpc)); in __kernel_fpu_begin() 31 asm volatile("std 0,%0" : "=Q" (state->fprs[0])); in __kernel_fpu_begin() 32 asm volatile("std 1,%0" : "=Q" (state->fprs[1])); in __kernel_fpu_begin() 33 asm volatile("std 2,%0" : "=Q" (state->fprs[2])); in __kernel_fpu_begin() 34 asm volatile("std 3,%0" : "=Q" (state->fprs[3])); in __kernel_fpu_begin() 35 asm volatile("std 4,%0" : "=Q" (state->fprs[4])); in __kernel_fpu_begin() 36 asm volatile("std 5,%0" : "=Q" (state->fprs[5])); in __kernel_fpu_begin() 37 asm volatile("std 6,%0" : "=Q" (state->fprs[6])); in __kernel_fpu_begin() [all …]
|
/arch/arm64/include/asm/ |
D | fpsimdmacros.h | 20 .macro fpsimd_save state, tmpnr 21 stp q0, q1, [\state, #16 * 0] 22 stp q2, q3, [\state, #16 * 2] 23 stp q4, q5, [\state, #16 * 4] 24 stp q6, q7, [\state, #16 * 6] 25 stp q8, q9, [\state, #16 * 8] 26 stp q10, q11, [\state, #16 * 10] 27 stp q12, q13, [\state, #16 * 12] 28 stp q14, q15, [\state, #16 * 14] 29 stp q16, q17, [\state, #16 * 16] [all …]
|
/arch/mips/cavium-octeon/crypto/ |
D | octeon-sha512.c | 37 write_octeon_64bit_hash_sha512(sctx->state[0], 0); in octeon_sha512_store_hash() 38 write_octeon_64bit_hash_sha512(sctx->state[1], 1); in octeon_sha512_store_hash() 39 write_octeon_64bit_hash_sha512(sctx->state[2], 2); in octeon_sha512_store_hash() 40 write_octeon_64bit_hash_sha512(sctx->state[3], 3); in octeon_sha512_store_hash() 41 write_octeon_64bit_hash_sha512(sctx->state[4], 4); in octeon_sha512_store_hash() 42 write_octeon_64bit_hash_sha512(sctx->state[5], 5); in octeon_sha512_store_hash() 43 write_octeon_64bit_hash_sha512(sctx->state[6], 6); in octeon_sha512_store_hash() 44 write_octeon_64bit_hash_sha512(sctx->state[7], 7); in octeon_sha512_store_hash() 49 sctx->state[0] = read_octeon_64bit_hash_sha512(0); in octeon_sha512_read_hash() 50 sctx->state[1] = read_octeon_64bit_hash_sha512(1); in octeon_sha512_read_hash() [all …]
|
D | octeon-sha256.c | 38 u64 *hash = (u64 *)sctx->state; in octeon_sha256_store_hash() 48 u64 *hash = (u64 *)sctx->state; in octeon_sha256_read_hash() 74 sctx->state[0] = SHA224_H0; in octeon_sha224_init() 75 sctx->state[1] = SHA224_H1; in octeon_sha224_init() 76 sctx->state[2] = SHA224_H2; in octeon_sha224_init() 77 sctx->state[3] = SHA224_H3; in octeon_sha224_init() 78 sctx->state[4] = SHA224_H4; in octeon_sha224_init() 79 sctx->state[5] = SHA224_H5; in octeon_sha224_init() 80 sctx->state[6] = SHA224_H6; in octeon_sha224_init() 81 sctx->state[7] = SHA224_H7; in octeon_sha224_init() [all …]
|
/arch/powerpc/kvm/ |
D | book3s_xive.c | 291 struct kvmppc_xive_irq_state *state) in xive_lock_and_mask() argument 304 old_prio = state->guest_priority; in xive_lock_and_mask() 305 state->guest_priority = MASKED; in xive_lock_and_mask() 307 if (!state->in_eoi) in xive_lock_and_mask() 309 state->guest_priority = old_prio; in xive_lock_and_mask() 318 kvmppc_xive_select_irq(state, &hw_num, &xd); in xive_lock_and_mask() 334 xive->vp_base + state->act_server, in xive_lock_and_mask() 335 MASKED, state->number); in xive_lock_and_mask() 337 state->old_p = true; in xive_lock_and_mask() 338 state->old_q = false; in xive_lock_and_mask() [all …]
|
/arch/x86/crypto/sha512-mb/ |
D | sha512_mb_mgr_submit_avx2.S | 70 #define state arg1 macro 98 # JOB* sha512_mb_mgr_submit_avx2(MB_MGR *state, JOB *job) 99 # arg 1 : rcx : state 106 mov _unused_lanes(state), unused_lanes 111 lea _ldata(state, lane_data), lane_data 112 mov unused_lanes, _unused_lanes(state) 116 movl DWORD_len,_lens+4(state , lane, 8) 124 vmovq %xmm0, _args_digest(state, lane, 8) 125 vpextrq $1, %xmm0, _args_digest+1*32(state , lane, 8) 126 vmovq %xmm1, _args_digest+2*32(state , lane, 8) [all …]
|
D | sha512_mb_mgr_flush_avx2.S | 68 #define state arg1 macro 108 # JOB* sha512_mb_mgr_flush_avx2(MB_MGR *state) 109 # arg 1 : rcx : state 115 mov _unused_lanes(state), unused_lanes 122 cmpq $0, offset(state) 125 cmpq $0, offset(state) 128 cmpq $0, offset(state) 134 mov offset(state,idx,8), tmp 139 cmpq $0, offset(state) 143 mov tmp, offset(state) [all …]
|
D | sha512_mb_mgr_init_avx2.c | 56 void sha512_mb_mgr_init_avx2(struct sha512_mb_mgr *state) in sha512_mb_mgr_init_avx2() argument 61 state->lens[0] = 0xFFFFFFFF00000000; in sha512_mb_mgr_init_avx2() 62 state->lens[1] = 0xFFFFFFFF00000001; in sha512_mb_mgr_init_avx2() 63 state->lens[2] = 0xFFFFFFFF00000002; in sha512_mb_mgr_init_avx2() 64 state->lens[3] = 0xFFFFFFFF00000003; in sha512_mb_mgr_init_avx2() 66 state->unused_lanes = 0xFF03020100; in sha512_mb_mgr_init_avx2() 68 state->ldata[j].job_in_lane = NULL; in sha512_mb_mgr_init_avx2()
|
/arch/s390/crypto/ |
D | sha512_s390.c | 30 *(__u64 *)&ctx->state[0] = 0x6a09e667f3bcc908ULL; in sha512_init() 31 *(__u64 *)&ctx->state[2] = 0xbb67ae8584caa73bULL; in sha512_init() 32 *(__u64 *)&ctx->state[4] = 0x3c6ef372fe94f82bULL; in sha512_init() 33 *(__u64 *)&ctx->state[6] = 0xa54ff53a5f1d36f1ULL; in sha512_init() 34 *(__u64 *)&ctx->state[8] = 0x510e527fade682d1ULL; in sha512_init() 35 *(__u64 *)&ctx->state[10] = 0x9b05688c2b3e6c1fULL; in sha512_init() 36 *(__u64 *)&ctx->state[12] = 0x1f83d9abfb41bd6bULL; in sha512_init() 37 *(__u64 *)&ctx->state[14] = 0x5be0cd19137e2179ULL; in sha512_init() 51 memcpy(octx->state, sctx->state, sizeof(octx->state)); in sha512_export() 65 memcpy(sctx->state, ictx->state, sizeof(ictx->state)); in sha512_import() [all …]
|
D | sha256_s390.c | 29 sctx->state[0] = SHA256_H0; in sha256_init() 30 sctx->state[1] = SHA256_H1; in sha256_init() 31 sctx->state[2] = SHA256_H2; in sha256_init() 32 sctx->state[3] = SHA256_H3; in sha256_init() 33 sctx->state[4] = SHA256_H4; in sha256_init() 34 sctx->state[5] = SHA256_H5; in sha256_init() 35 sctx->state[6] = SHA256_H6; in sha256_init() 36 sctx->state[7] = SHA256_H7; in sha256_init() 49 memcpy(octx->state, sctx->state, sizeof(octx->state)); in sha256_export() 60 memcpy(sctx->state, ictx->state, sizeof(ictx->state)); in sha256_import() [all …]
|
/arch/x86/crypto/sha256-mb/ |
D | sha256_mb_mgr_submit_avx2.S | 68 #define state arg1 macro 96 # JOB* sha256_mb_mgr_submit_avx2(MB_MGR *state, JOB_SHA256 *job) 97 # arg 1 : rcx : state 104 mov _unused_lanes(state), unused_lanes 110 lea _ldata(state, lane_data), lane_data 111 mov unused_lanes, _unused_lanes(state) 118 movl DWORD_len, _lens(state , lane, 4) 123 vmovd %xmm0, _args_digest(state, lane, 4) 124 vpextrd $1, %xmm0, _args_digest+1*32(state , lane, 4) 125 vpextrd $2, %xmm0, _args_digest+2*32(state , lane, 4) [all …]
|
D | sha256_mb_mgr_flush_avx2.S | 64 #define state arg1 macro 102 # JOB_SHA256* sha256_mb_mgr_flush_avx2(MB_MGR *state) 103 # arg 1 : rcx : state 109 mov _unused_lanes(state), unused_lanes 116 cmpq $0, offset(state) 119 cmpq $0, offset(state) 122 cmpq $0, offset(state) 125 cmpq $0, offset(state) 128 cmpq $0, offset(state) 131 cmpq $0, offset(state) [all …]
|
/arch/powerpc/crypto/ |
D | sha256-spe-glue.c | 39 extern void ppc_spe_sha256_transform(u32 *state, const u8 *src, u32 blocks); 69 sctx->state[0] = SHA256_H0; in ppc_spe_sha256_init() 70 sctx->state[1] = SHA256_H1; in ppc_spe_sha256_init() 71 sctx->state[2] = SHA256_H2; in ppc_spe_sha256_init() 72 sctx->state[3] = SHA256_H3; in ppc_spe_sha256_init() 73 sctx->state[4] = SHA256_H4; in ppc_spe_sha256_init() 74 sctx->state[5] = SHA256_H5; in ppc_spe_sha256_init() 75 sctx->state[6] = SHA256_H6; in ppc_spe_sha256_init() 76 sctx->state[7] = SHA256_H7; in ppc_spe_sha256_init() 86 sctx->state[0] = SHA224_H0; in ppc_spe_sha224_init() [all …]
|
/arch/x86/include/asm/ |
D | unwind.h | 38 void __unwind_start(struct unwind_state *state, struct task_struct *task, 40 bool unwind_next_frame(struct unwind_state *state); 41 unsigned long unwind_get_return_address(struct unwind_state *state); 42 unsigned long *unwind_get_return_address_ptr(struct unwind_state *state); 44 static inline bool unwind_done(struct unwind_state *state) in unwind_done() argument 46 return state->stack_info.type == STACK_TYPE_UNKNOWN; in unwind_done() 49 static inline bool unwind_error(struct unwind_state *state) in unwind_error() argument 51 return state->error; in unwind_error() 55 void unwind_start(struct unwind_state *state, struct task_struct *task, in unwind_start() argument 60 __unwind_start(state, task, regs, first_frame); in unwind_start() [all …]
|
/arch/x86/boot/ |
D | cmdline.c | 42 } state = st_wordstart; in __cmdline_find_option() local 51 switch (state) { in __cmdline_find_option() 57 state = st_wordcmp; in __cmdline_find_option() 65 state = st_bufcpy; in __cmdline_find_option() 67 state = st_wordstart; in __cmdline_find_option() 69 state = st_wordskip; in __cmdline_find_option() 75 state = st_wordstart; in __cmdline_find_option() 80 state = st_wordstart; in __cmdline_find_option() 112 } state = st_wordstart; in __cmdline_find_option_bool() local 124 switch (state) { in __cmdline_find_option_bool() [all …]
|
/arch/x86/crypto/sha1-mb/ |
D | sha1_mb_mgr_flush_avx2.S | 66 #define state arg1 macro 104 # JOB* sha1_mb_mgr_flush_avx2(MB_MGR *state) 105 # arg 1 : rcx : state 111 mov _unused_lanes(state), unused_lanes 118 cmpq $0, offset(state) 121 cmpq $0, offset(state) 124 cmpq $0, offset(state) 127 cmpq $0, offset(state) 130 cmpq $0, offset(state) 133 cmpq $0, offset(state) [all …]
|
D | sha1_mb_mgr_submit_avx2.S | 70 #define state arg1 macro 98 # JOB* submit_mb_mgr_submit_avx2(MB_MGR *state, job_sha1 *job) 99 # arg 1 : rcx : state 106 mov _unused_lanes(state), unused_lanes 112 lea _ldata(state, lane_data), lane_data 113 mov unused_lanes, _unused_lanes(state) 120 movl DWORD_len, _lens(state , lane, 4) 125 vmovd %xmm0, _args_digest(state, lane, 4) 126 vpextrd $1, %xmm0, _args_digest+1*32(state , lane, 4) 127 vpextrd $2, %xmm0, _args_digest+2*32(state , lane, 4) [all …]
|
/arch/s390/include/asm/fpu/ |
D | api.h | 90 void __kernel_fpu_begin(struct kernel_fpu *state, u32 flags); 91 void __kernel_fpu_end(struct kernel_fpu *state, u32 flags); 94 static inline void kernel_fpu_begin(struct kernel_fpu *state, u32 flags) in kernel_fpu_begin() argument 97 state->mask = S390_lowcore.fpu_flags; in kernel_fpu_begin() 101 else if (state->mask & flags) in kernel_fpu_begin() 103 __kernel_fpu_begin(state, flags); in kernel_fpu_begin() 107 static inline void kernel_fpu_end(struct kernel_fpu *state, u32 flags) in kernel_fpu_end() argument 109 S390_lowcore.fpu_flags = state->mask; in kernel_fpu_end() 110 if (state->mask & flags) in kernel_fpu_end() 112 __kernel_fpu_end(state, flags); in kernel_fpu_end()
|
/arch/x86/lib/ |
D | cmdline.c | 39 } state = st_wordstart; in __cmdline_find_option_bool() local 52 switch (state) { in __cmdline_find_option_bool() 59 state = st_wordcmp; in __cmdline_find_option_bool() 92 state = st_wordskip; in __cmdline_find_option_bool() 99 state = st_wordstart; in __cmdline_find_option_bool() 134 } state = st_wordstart; in __cmdline_find_option() local 148 switch (state) { in __cmdline_find_option() 153 state = st_wordcmp; in __cmdline_find_option() 166 state = st_bufcpy; in __cmdline_find_option() 175 state = st_wordskip; in __cmdline_find_option() [all …]
|