/arch/arc/kernel/ |
D | disasm.c | 21 void __kprobes disasm_instr(unsigned long addr, struct disasm_state *state, in disasm_instr() argument 32 memset(state, 0, sizeof(struct disasm_state)); in disasm_instr() 48 state->major_opcode = (word1 >> 11) & 0x1F; in disasm_instr() 51 if (state->major_opcode < 0x0B) { in disasm_instr() 54 state->instr_len = 4; in disasm_instr() 56 state->words[0] = (word1 << 16) | word0; in disasm_instr() 58 state->instr_len = 2; in disasm_instr() 59 state->words[0] = word1; in disasm_instr() 63 word1 = *((uint16_t *)(addr + state->instr_len)); in disasm_instr() 64 word0 = *((uint16_t *)(addr + state->instr_len + 2)); in disasm_instr() [all …]
|
D | unaligned.c | 132 static void fixup_load(struct disasm_state *state, struct pt_regs *regs, in fixup_load() argument 138 if ((state->aa == 1) || (state->aa == 2)) { in fixup_load() 139 set_reg(state->wb_reg, state->src1 + state->src2, regs, cregs); in fixup_load() 141 if (state->aa == 2) in fixup_load() 142 state->src2 = 0; in fixup_load() 145 if (state->zz == 0) { in fixup_load() 146 get32_unaligned_check(val, state->src1 + state->src2); in fixup_load() 148 get16_unaligned_check(val, state->src1 + state->src2); in fixup_load() 150 if (state->x) in fixup_load() 154 if (state->pref == 0) in fixup_load() [all …]
|
D | unwind.c | 664 static int advance_loc(unsigned long delta, struct unwind_state *state) in advance_loc() argument 666 state->loc += delta * state->codeAlign; in advance_loc() 671 unw_debug("delta %3lu => loc 0x%lx: ", delta, state->loc); in advance_loc() 676 struct unwind_state *state) in set_rule() argument 678 if (reg < ARRAY_SIZE(state->regs)) { in set_rule() 679 state->regs[reg].where = where; in set_rule() 680 state->regs[reg].value = value; in set_rule() 705 signed ptrType, struct unwind_state *state) in processCFI() argument 715 if (start != state->cieStart) { in processCFI() 716 state->loc = state->org; in processCFI() [all …]
|
/arch/x86/kernel/ |
D | unwind_frame.c | 14 unsigned long unwind_get_return_address(struct unwind_state *state) in unwind_get_return_address() argument 16 if (unwind_done(state)) in unwind_get_return_address() 19 return __kernel_text_address(state->ip) ? state->ip : 0; in unwind_get_return_address() 23 unsigned long *unwind_get_return_address_ptr(struct unwind_state *state) in unwind_get_return_address_ptr() argument 25 if (unwind_done(state)) in unwind_get_return_address_ptr() 28 return state->regs ? &state->regs->ip : state->bp + 1; in unwind_get_return_address_ptr() 31 static void unwind_dump(struct unwind_state *state) in unwind_dump() argument 45 state->stack_info.type, state->stack_info.next_sp, in unwind_dump() 46 state->stack_mask, state->graph_idx); in unwind_dump() 48 for (sp = PTR_ALIGN(state->orig_sp, sizeof(long)); sp; in unwind_dump() [all …]
|
D | unwind_guess.c | 9 unsigned long unwind_get_return_address(struct unwind_state *state) in unwind_get_return_address() argument 13 if (unwind_done(state)) in unwind_get_return_address() 16 addr = READ_ONCE_NOCHECK(*state->sp); in unwind_get_return_address() 18 return ftrace_graph_ret_addr(state->task, &state->graph_idx, in unwind_get_return_address() 19 addr, state->sp); in unwind_get_return_address() 23 unsigned long *unwind_get_return_address_ptr(struct unwind_state *state) in unwind_get_return_address_ptr() argument 28 bool unwind_next_frame(struct unwind_state *state) in unwind_next_frame() argument 30 struct stack_info *info = &state->stack_info; in unwind_next_frame() 32 if (unwind_done(state)) in unwind_next_frame() 36 for (state->sp++; state->sp < info->end; state->sp++) { in unwind_next_frame() [all …]
|
D | unwind_orc.c | 16 if (state->task == current) \ 318 unsigned long unwind_get_return_address(struct unwind_state *state) in unwind_get_return_address() argument 320 if (unwind_done(state)) in unwind_get_return_address() 323 return __kernel_text_address(state->ip) ? state->ip : 0; in unwind_get_return_address() 327 unsigned long *unwind_get_return_address_ptr(struct unwind_state *state) in unwind_get_return_address_ptr() argument 329 if (unwind_done(state)) in unwind_get_return_address_ptr() 332 if (state->regs) in unwind_get_return_address_ptr() 333 return &state->regs->ip; in unwind_get_return_address_ptr() 335 if (state->sp) in unwind_get_return_address_ptr() 336 return (unsigned long *)state->sp - 1; in unwind_get_return_address_ptr() [all …]
|
D | stacktrace.c | 18 struct unwind_state state; in arch_stack_walk() local 24 for (unwind_start(&state, task, regs, NULL); !unwind_done(&state); in arch_stack_walk() 25 unwind_next_frame(&state)) { in arch_stack_walk() 26 addr = unwind_get_return_address(&state); in arch_stack_walk() 41 struct unwind_state state; in arch_stack_walk_reliable() local 45 for (unwind_start(&state, task, NULL, NULL); in arch_stack_walk_reliable() 46 !unwind_done(&state) && !unwind_error(&state); in arch_stack_walk_reliable() 47 unwind_next_frame(&state)) { in arch_stack_walk_reliable() 49 regs = unwind_get_entry_regs(&state, NULL); in arch_stack_walk_reliable() 65 addr = unwind_get_return_address(&state); in arch_stack_walk_reliable() [all …]
|
/arch/s390/kernel/ |
D | fpu.c | 16 void __kernel_fpu_begin(struct kernel_fpu *state, u32 flags) in __kernel_fpu_begin() argument 22 flags &= state->mask; in __kernel_fpu_begin() 26 asm volatile("stfpc %0" : "=Q" (state->fpc)); in __kernel_fpu_begin() 31 asm volatile("std 0,%0" : "=Q" (state->fprs[0])); in __kernel_fpu_begin() 32 asm volatile("std 1,%0" : "=Q" (state->fprs[1])); in __kernel_fpu_begin() 33 asm volatile("std 2,%0" : "=Q" (state->fprs[2])); in __kernel_fpu_begin() 34 asm volatile("std 3,%0" : "=Q" (state->fprs[3])); in __kernel_fpu_begin() 35 asm volatile("std 4,%0" : "=Q" (state->fprs[4])); in __kernel_fpu_begin() 36 asm volatile("std 5,%0" : "=Q" (state->fprs[5])); in __kernel_fpu_begin() 37 asm volatile("std 6,%0" : "=Q" (state->fprs[6])); in __kernel_fpu_begin() [all …]
|
D | unwind_bc.c | 12 unsigned long unwind_get_return_address(struct unwind_state *state) in unwind_get_return_address() argument 14 if (unwind_done(state)) in unwind_get_return_address() 16 return __kernel_text_address(state->ip) ? state->ip : 0; in unwind_get_return_address() 20 static bool outside_of_stack(struct unwind_state *state, unsigned long sp) in outside_of_stack() argument 22 return (sp <= state->sp) || in outside_of_stack() 23 (sp > state->stack_info.end - sizeof(struct stack_frame)); in outside_of_stack() 26 static bool update_stack_info(struct unwind_state *state, unsigned long sp) in update_stack_info() argument 28 struct stack_info *info = &state->stack_info; in update_stack_info() 29 unsigned long *mask = &state->stack_mask; in update_stack_info() 32 if (get_stack_info(sp, state->task, info, mask) != 0 || in update_stack_info() [all …]
|
D | stacktrace.c | 17 struct unwind_state state; in arch_stack_walk() local 20 unwind_for_each_frame(&state, task, regs, 0) { in arch_stack_walk() 21 addr = unwind_get_return_address(&state); in arch_stack_walk() 36 struct unwind_state state; in arch_stack_walk_reliable() local 39 unwind_for_each_frame(&state, task, NULL, 0) { in arch_stack_walk_reliable() 40 if (state.stack_info.type != STACK_TYPE_TASK) in arch_stack_walk_reliable() 43 if (state.regs) in arch_stack_walk_reliable() 46 addr = unwind_get_return_address(&state); in arch_stack_walk_reliable() 55 if (state.ip == (unsigned long)kretprobe_trampoline) in arch_stack_walk_reliable() 64 if (unwind_error(&state)) in arch_stack_walk_reliable()
|
/arch/mips/cavium-octeon/crypto/ |
D | octeon-sha512.c | 33 write_octeon_64bit_hash_sha512(sctx->state[0], 0); in octeon_sha512_store_hash() 34 write_octeon_64bit_hash_sha512(sctx->state[1], 1); in octeon_sha512_store_hash() 35 write_octeon_64bit_hash_sha512(sctx->state[2], 2); in octeon_sha512_store_hash() 36 write_octeon_64bit_hash_sha512(sctx->state[3], 3); in octeon_sha512_store_hash() 37 write_octeon_64bit_hash_sha512(sctx->state[4], 4); in octeon_sha512_store_hash() 38 write_octeon_64bit_hash_sha512(sctx->state[5], 5); in octeon_sha512_store_hash() 39 write_octeon_64bit_hash_sha512(sctx->state[6], 6); in octeon_sha512_store_hash() 40 write_octeon_64bit_hash_sha512(sctx->state[7], 7); in octeon_sha512_store_hash() 45 sctx->state[0] = read_octeon_64bit_hash_sha512(0); in octeon_sha512_read_hash() 46 sctx->state[1] = read_octeon_64bit_hash_sha512(1); in octeon_sha512_read_hash() [all …]
|
D | octeon-sha256.c | 34 u64 *hash = (u64 *)sctx->state; in octeon_sha256_store_hash() 44 u64 *hash = (u64 *)sctx->state; in octeon_sha256_read_hash() 70 sctx->state[0] = SHA224_H0; in octeon_sha224_init() 71 sctx->state[1] = SHA224_H1; in octeon_sha224_init() 72 sctx->state[2] = SHA224_H2; in octeon_sha224_init() 73 sctx->state[3] = SHA224_H3; in octeon_sha224_init() 74 sctx->state[4] = SHA224_H4; in octeon_sha224_init() 75 sctx->state[5] = SHA224_H5; in octeon_sha224_init() 76 sctx->state[6] = SHA224_H6; in octeon_sha224_init() 77 sctx->state[7] = SHA224_H7; in octeon_sha224_init() [all …]
|
/arch/s390/include/asm/ |
D | unwind.h | 43 void __unwind_start(struct unwind_state *state, struct task_struct *task, 45 bool unwind_next_frame(struct unwind_state *state); 46 unsigned long unwind_get_return_address(struct unwind_state *state); 48 static inline bool unwind_done(struct unwind_state *state) in unwind_done() argument 50 return state->stack_info.type == STACK_TYPE_UNKNOWN; in unwind_done() 53 static inline bool unwind_error(struct unwind_state *state) in unwind_error() argument 55 return state->error; in unwind_error() 58 static inline void unwind_start(struct unwind_state *state, in unwind_start() argument 65 __unwind_start(state, task, regs, first_frame); in unwind_start() 68 static inline struct pt_regs *unwind_get_entry_regs(struct unwind_state *state) in unwind_get_entry_regs() argument [all …]
|
D | cpu_mcf.h | 35 static inline void ctr_set_enable(u64 *state, int ctr_set) in ctr_set_enable() argument 37 *state |= cpumf_ctr_ctl[ctr_set] << CPUMF_LCCTL_ENABLE_SHIFT; in ctr_set_enable() 39 static inline void ctr_set_disable(u64 *state, int ctr_set) in ctr_set_disable() argument 41 *state &= ~(cpumf_ctr_ctl[ctr_set] << CPUMF_LCCTL_ENABLE_SHIFT); in ctr_set_disable() 43 static inline void ctr_set_start(u64 *state, int ctr_set) in ctr_set_start() argument 45 *state |= cpumf_ctr_ctl[ctr_set] << CPUMF_LCCTL_ACTCTL_SHIFT; in ctr_set_start() 47 static inline void ctr_set_stop(u64 *state, int ctr_set) in ctr_set_stop() argument 49 *state &= ~(cpumf_ctr_ctl[ctr_set] << CPUMF_LCCTL_ACTCTL_SHIFT); in ctr_set_stop() 52 static inline void ctr_set_multiple_enable(u64 *state, u64 ctrsets) in ctr_set_multiple_enable() argument 54 *state |= ctrsets << CPUMF_LCCTL_ENABLE_SHIFT; in ctr_set_multiple_enable() [all …]
|
/arch/arm64/include/asm/ |
D | fpsimdmacros.h | 11 .macro fpsimd_save state, tmpnr 12 stp q0, q1, [\state, #16 * 0] 13 stp q2, q3, [\state, #16 * 2] 14 stp q4, q5, [\state, #16 * 4] 15 stp q6, q7, [\state, #16 * 6] 16 stp q8, q9, [\state, #16 * 8] 17 stp q10, q11, [\state, #16 * 10] 18 stp q12, q13, [\state, #16 * 12] 19 stp q14, q15, [\state, #16 * 14] 20 stp q16, q17, [\state, #16 * 16] [all …]
|
/arch/x86/crypto/ |
D | chacha_glue.c | 18 asmlinkage void chacha_block_xor_ssse3(u32 *state, u8 *dst, const u8 *src, 20 asmlinkage void chacha_4block_xor_ssse3(u32 *state, u8 *dst, const u8 *src, 22 asmlinkage void hchacha_block_ssse3(const u32 *state, u32 *out, int nrounds); 24 asmlinkage void chacha_2block_xor_avx2(u32 *state, u8 *dst, const u8 *src, 26 asmlinkage void chacha_4block_xor_avx2(u32 *state, u8 *dst, const u8 *src, 28 asmlinkage void chacha_8block_xor_avx2(u32 *state, u8 *dst, const u8 *src, 31 asmlinkage void chacha_2block_xor_avx512vl(u32 *state, u8 *dst, const u8 *src, 33 asmlinkage void chacha_4block_xor_avx512vl(u32 *state, u8 *dst, const u8 *src, 35 asmlinkage void chacha_8block_xor_avx512vl(u32 *state, u8 *dst, const u8 *src, 48 static void chacha_dosimd(u32 *state, u8 *dst, const u8 *src, in chacha_dosimd() argument [all …]
|
D | poly1305_glue.c | 61 struct poly1305_arch_internal *state = ctx; in convert_to_base2_64() local 64 if (!state->is_base2_26) in convert_to_base2_64() 67 cy = state->h[0] >> 26; state->h[0] &= 0x3ffffff; state->h[1] += cy; in convert_to_base2_64() 68 cy = state->h[1] >> 26; state->h[1] &= 0x3ffffff; state->h[2] += cy; in convert_to_base2_64() 69 cy = state->h[2] >> 26; state->h[2] &= 0x3ffffff; state->h[3] += cy; in convert_to_base2_64() 70 cy = state->h[3] >> 26; state->h[3] &= 0x3ffffff; state->h[4] += cy; in convert_to_base2_64() 71 state->hs[0] = ((u64)state->h[2] << 52) | ((u64)state->h[1] << 26) | state->h[0]; in convert_to_base2_64() 72 state->hs[1] = ((u64)state->h[4] << 40) | ((u64)state->h[3] << 14) | (state->h[2] >> 12); in convert_to_base2_64() 73 state->hs[2] = state->h[4] >> 24; in convert_to_base2_64() 75 cy = (state->hs[2] >> 2) + (state->hs[2] & ~3ULL); in convert_to_base2_64() [all …]
|
/arch/x86/kernel/cpu/mce/ |
D | therm_throt.c | 240 struct _thermal_state *state = container_of(to_delayed_work(work), in throttle_active_work() local 247 get_therm_status(state->level, &hot, &temp); in throttle_active_work() 249 if (!hot && temp > state->baseline_temp) { in throttle_active_work() 250 if (state->rate_control_active) in throttle_active_work() 253 state->level == CORE_LEVEL ? "Core" : "Package", in throttle_active_work() 254 state->count); in throttle_active_work() 256 state->rate_control_active = false; in throttle_active_work() 260 if (time_before64(now, state->next_check) && in throttle_active_work() 261 state->rate_control_active) in throttle_active_work() 264 state->next_check = now + CHECK_INTERVAL; in throttle_active_work() [all …]
|
/arch/s390/crypto/ |
D | sha512_s390.c | 25 *(__u64 *)&ctx->state[0] = 0x6a09e667f3bcc908ULL; in sha512_init() 26 *(__u64 *)&ctx->state[2] = 0xbb67ae8584caa73bULL; in sha512_init() 27 *(__u64 *)&ctx->state[4] = 0x3c6ef372fe94f82bULL; in sha512_init() 28 *(__u64 *)&ctx->state[6] = 0xa54ff53a5f1d36f1ULL; in sha512_init() 29 *(__u64 *)&ctx->state[8] = 0x510e527fade682d1ULL; in sha512_init() 30 *(__u64 *)&ctx->state[10] = 0x9b05688c2b3e6c1fULL; in sha512_init() 31 *(__u64 *)&ctx->state[12] = 0x1f83d9abfb41bd6bULL; in sha512_init() 32 *(__u64 *)&ctx->state[14] = 0x5be0cd19137e2179ULL; in sha512_init() 46 memcpy(octx->state, sctx->state, sizeof(octx->state)); in sha512_export() 60 memcpy(sctx->state, ictx->state, sizeof(ictx->state)); in sha512_import() [all …]
|
D | sha256_s390.c | 24 sctx->state[0] = SHA256_H0; in s390_sha256_init() 25 sctx->state[1] = SHA256_H1; in s390_sha256_init() 26 sctx->state[2] = SHA256_H2; in s390_sha256_init() 27 sctx->state[3] = SHA256_H3; in s390_sha256_init() 28 sctx->state[4] = SHA256_H4; in s390_sha256_init() 29 sctx->state[5] = SHA256_H5; in s390_sha256_init() 30 sctx->state[6] = SHA256_H6; in s390_sha256_init() 31 sctx->state[7] = SHA256_H7; in s390_sha256_init() 44 memcpy(octx->state, sctx->state, sizeof(octx->state)); in sha256_export() 55 memcpy(sctx->state, ictx->state, sizeof(ictx->state)); in sha256_import() [all …]
|
/arch/powerpc/kvm/ |
D | book3s_xive.c | 393 struct kvmppc_xive_irq_state *state) in xive_lock_and_mask() argument 406 old_prio = state->guest_priority; in xive_lock_and_mask() 407 state->guest_priority = MASKED; in xive_lock_and_mask() 409 if (!state->in_eoi) in xive_lock_and_mask() 411 state->guest_priority = old_prio; in xive_lock_and_mask() 420 kvmppc_xive_select_irq(state, &hw_num, &xd); in xive_lock_and_mask() 436 kvmppc_xive_vp(xive, state->act_server), in xive_lock_and_mask() 437 MASKED, state->number); in xive_lock_and_mask() 439 state->old_p = true; in xive_lock_and_mask() 440 state->old_q = false; in xive_lock_and_mask() [all …]
|
/arch/powerpc/crypto/ |
D | sha256-spe-glue.c | 33 extern void ppc_spe_sha256_transform(u32 *state, const u8 *src, u32 blocks); 63 sctx->state[0] = SHA256_H0; in ppc_spe_sha256_init() 64 sctx->state[1] = SHA256_H1; in ppc_spe_sha256_init() 65 sctx->state[2] = SHA256_H2; in ppc_spe_sha256_init() 66 sctx->state[3] = SHA256_H3; in ppc_spe_sha256_init() 67 sctx->state[4] = SHA256_H4; in ppc_spe_sha256_init() 68 sctx->state[5] = SHA256_H5; in ppc_spe_sha256_init() 69 sctx->state[6] = SHA256_H6; in ppc_spe_sha256_init() 70 sctx->state[7] = SHA256_H7; in ppc_spe_sha256_init() 80 sctx->state[0] = SHA224_H0; in ppc_spe_sha224_init() [all …]
|
/arch/x86/include/asm/ |
D | unwind.h | 38 void __unwind_start(struct unwind_state *state, struct task_struct *task, 40 bool unwind_next_frame(struct unwind_state *state); 41 unsigned long unwind_get_return_address(struct unwind_state *state); 42 unsigned long *unwind_get_return_address_ptr(struct unwind_state *state); 44 static inline bool unwind_done(struct unwind_state *state) in unwind_done() argument 46 return state->stack_info.type == STACK_TYPE_UNKNOWN; in unwind_done() 49 static inline bool unwind_error(struct unwind_state *state) in unwind_error() argument 51 return state->error; in unwind_error() 55 void unwind_start(struct unwind_state *state, struct task_struct *task, in unwind_start() argument 60 __unwind_start(state, task, regs, first_frame); in unwind_start() [all …]
|
/arch/x86/boot/ |
D | cmdline.c | 40 } state = st_wordstart; in __cmdline_find_option() local 49 switch (state) { in __cmdline_find_option() 55 state = st_wordcmp; in __cmdline_find_option() 63 state = st_bufcpy; in __cmdline_find_option() 65 state = st_wordstart; in __cmdline_find_option() 67 state = st_wordskip; in __cmdline_find_option() 73 state = st_wordstart; in __cmdline_find_option() 78 state = st_wordstart; in __cmdline_find_option() 110 } state = st_wordstart; in __cmdline_find_option_bool() local 122 switch (state) { in __cmdline_find_option_bool() [all …]
|
/arch/arm/mm/ |
D | proc-v7-bugs.c | 119 unsigned int state, method = 0; in cpu_v7_spectre_v2_init() local 128 state = SPECTRE_MITIGATED; in cpu_v7_spectre_v2_init() 134 state = SPECTRE_MITIGATED; in cpu_v7_spectre_v2_init() 140 state = SPECTRE_UNAFFECTED; in cpu_v7_spectre_v2_init() 146 state = SPECTRE_UNAFFECTED; in cpu_v7_spectre_v2_init() 155 state = spectre_v2_get_cpu_fw_mitigation_state(); in cpu_v7_spectre_v2_init() 156 if (state != SPECTRE_MITIGATED) in cpu_v7_spectre_v2_init() 169 state = SPECTRE_VULNERABLE; in cpu_v7_spectre_v2_init() 174 if (state == SPECTRE_MITIGATED) in cpu_v7_spectre_v2_init() 175 state = spectre_v2_install_workaround(method); in cpu_v7_spectre_v2_init() [all …]
|