Home
last modified time | relevance | path

Searched refs:state (Results 1 – 25 of 841) sorted by relevance

12345678910>>...34

/arch/arc/kernel/
Ddisasm.c21 void __kprobes disasm_instr(unsigned long addr, struct disasm_state *state, in disasm_instr() argument
32 memset(state, 0, sizeof(struct disasm_state)); in disasm_instr()
48 state->major_opcode = (word1 >> 11) & 0x1F; in disasm_instr()
51 if (state->major_opcode < 0x0B) { in disasm_instr()
54 state->instr_len = 4; in disasm_instr()
56 state->words[0] = (word1 << 16) | word0; in disasm_instr()
58 state->instr_len = 2; in disasm_instr()
59 state->words[0] = word1; in disasm_instr()
63 word1 = *((uint16_t *)(addr + state->instr_len)); in disasm_instr()
64 word0 = *((uint16_t *)(addr + state->instr_len + 2)); in disasm_instr()
[all …]
Dunaligned.c132 static void fixup_load(struct disasm_state *state, struct pt_regs *regs, in fixup_load() argument
138 if ((state->aa == 1) || (state->aa == 2)) { in fixup_load()
139 set_reg(state->wb_reg, state->src1 + state->src2, regs, cregs); in fixup_load()
141 if (state->aa == 2) in fixup_load()
142 state->src2 = 0; in fixup_load()
145 if (state->zz == 0) { in fixup_load()
146 get32_unaligned_check(val, state->src1 + state->src2); in fixup_load()
148 get16_unaligned_check(val, state->src1 + state->src2); in fixup_load()
150 if (state->x) in fixup_load()
154 if (state->pref == 0) in fixup_load()
[all …]
Dunwind.c664 static int advance_loc(unsigned long delta, struct unwind_state *state) in advance_loc() argument
666 state->loc += delta * state->codeAlign; in advance_loc()
671 unw_debug("delta %3lu => loc 0x%lx: ", delta, state->loc); in advance_loc()
676 struct unwind_state *state) in set_rule() argument
678 if (reg < ARRAY_SIZE(state->regs)) { in set_rule()
679 state->regs[reg].where = where; in set_rule()
680 state->regs[reg].value = value; in set_rule()
705 signed ptrType, struct unwind_state *state) in processCFI() argument
715 if (start != state->cieStart) { in processCFI()
716 state->loc = state->org; in processCFI()
[all …]
/arch/x86/kernel/
Dunwind_frame.c14 unsigned long unwind_get_return_address(struct unwind_state *state) in unwind_get_return_address() argument
16 if (unwind_done(state)) in unwind_get_return_address()
19 return __kernel_text_address(state->ip) ? state->ip : 0; in unwind_get_return_address()
23 unsigned long *unwind_get_return_address_ptr(struct unwind_state *state) in unwind_get_return_address_ptr() argument
25 if (unwind_done(state)) in unwind_get_return_address_ptr()
28 return state->regs ? &state->regs->ip : state->bp + 1; in unwind_get_return_address_ptr()
31 static void unwind_dump(struct unwind_state *state) in unwind_dump() argument
45 state->stack_info.type, state->stack_info.next_sp, in unwind_dump()
46 state->stack_mask, state->graph_idx); in unwind_dump()
48 for (sp = PTR_ALIGN(state->orig_sp, sizeof(long)); sp; in unwind_dump()
[all …]
Dunwind_guess.c9 unsigned long unwind_get_return_address(struct unwind_state *state) in unwind_get_return_address() argument
13 if (unwind_done(state)) in unwind_get_return_address()
16 addr = READ_ONCE_NOCHECK(*state->sp); in unwind_get_return_address()
18 return ftrace_graph_ret_addr(state->task, &state->graph_idx, in unwind_get_return_address()
19 addr, state->sp); in unwind_get_return_address()
23 unsigned long *unwind_get_return_address_ptr(struct unwind_state *state) in unwind_get_return_address_ptr() argument
28 bool unwind_next_frame(struct unwind_state *state) in unwind_next_frame() argument
30 struct stack_info *info = &state->stack_info; in unwind_next_frame()
32 if (unwind_done(state)) in unwind_next_frame()
36 for (state->sp++; state->sp < info->end; state->sp++) { in unwind_next_frame()
[all …]
Dunwind_orc.c308 unsigned long unwind_get_return_address(struct unwind_state *state) in unwind_get_return_address() argument
310 if (unwind_done(state)) in unwind_get_return_address()
313 return __kernel_text_address(state->ip) ? state->ip : 0; in unwind_get_return_address()
317 unsigned long *unwind_get_return_address_ptr(struct unwind_state *state) in unwind_get_return_address_ptr() argument
319 if (unwind_done(state)) in unwind_get_return_address_ptr()
322 if (state->regs) in unwind_get_return_address_ptr()
323 return &state->regs->ip; in unwind_get_return_address_ptr()
325 if (state->sp) in unwind_get_return_address_ptr()
326 return (unsigned long *)state->sp - 1; in unwind_get_return_address_ptr()
331 static bool stack_access_ok(struct unwind_state *state, unsigned long _addr, in stack_access_ok() argument
[all …]
Dstacktrace.c18 struct unwind_state state; in arch_stack_walk() local
24 for (unwind_start(&state, task, regs, NULL); !unwind_done(&state); in arch_stack_walk()
25 unwind_next_frame(&state)) { in arch_stack_walk()
26 addr = unwind_get_return_address(&state); in arch_stack_walk()
41 struct unwind_state state; in arch_stack_walk_reliable() local
45 for (unwind_start(&state, task, NULL, NULL); in arch_stack_walk_reliable()
46 !unwind_done(&state) && !unwind_error(&state); in arch_stack_walk_reliable()
47 unwind_next_frame(&state)) { in arch_stack_walk_reliable()
49 regs = unwind_get_entry_regs(&state, NULL); in arch_stack_walk_reliable()
65 addr = unwind_get_return_address(&state); in arch_stack_walk_reliable()
[all …]
/arch/s390/kernel/
Dfpu.c16 void __kernel_fpu_begin(struct kernel_fpu *state, u32 flags) in __kernel_fpu_begin() argument
22 flags &= state->mask; in __kernel_fpu_begin()
26 asm volatile("stfpc %0" : "=Q" (state->fpc)); in __kernel_fpu_begin()
31 asm volatile("std 0,%0" : "=Q" (state->fprs[0])); in __kernel_fpu_begin()
32 asm volatile("std 1,%0" : "=Q" (state->fprs[1])); in __kernel_fpu_begin()
33 asm volatile("std 2,%0" : "=Q" (state->fprs[2])); in __kernel_fpu_begin()
34 asm volatile("std 3,%0" : "=Q" (state->fprs[3])); in __kernel_fpu_begin()
35 asm volatile("std 4,%0" : "=Q" (state->fprs[4])); in __kernel_fpu_begin()
36 asm volatile("std 5,%0" : "=Q" (state->fprs[5])); in __kernel_fpu_begin()
37 asm volatile("std 6,%0" : "=Q" (state->fprs[6])); in __kernel_fpu_begin()
[all …]
Dunwind_bc.c12 unsigned long unwind_get_return_address(struct unwind_state *state) in unwind_get_return_address() argument
14 if (unwind_done(state)) in unwind_get_return_address()
16 return __kernel_text_address(state->ip) ? state->ip : 0; in unwind_get_return_address()
20 static bool outside_of_stack(struct unwind_state *state, unsigned long sp) in outside_of_stack() argument
22 return (sp <= state->sp) || in outside_of_stack()
23 (sp > state->stack_info.end - sizeof(struct stack_frame)); in outside_of_stack()
26 static bool update_stack_info(struct unwind_state *state, unsigned long sp) in update_stack_info() argument
28 struct stack_info *info = &state->stack_info; in update_stack_info()
29 unsigned long *mask = &state->stack_mask; in update_stack_info()
32 if (get_stack_info(sp, state->task, info, mask) != 0 || in update_stack_info()
[all …]
/arch/mips/cavium-octeon/crypto/
Docteon-sha512.c33 write_octeon_64bit_hash_sha512(sctx->state[0], 0); in octeon_sha512_store_hash()
34 write_octeon_64bit_hash_sha512(sctx->state[1], 1); in octeon_sha512_store_hash()
35 write_octeon_64bit_hash_sha512(sctx->state[2], 2); in octeon_sha512_store_hash()
36 write_octeon_64bit_hash_sha512(sctx->state[3], 3); in octeon_sha512_store_hash()
37 write_octeon_64bit_hash_sha512(sctx->state[4], 4); in octeon_sha512_store_hash()
38 write_octeon_64bit_hash_sha512(sctx->state[5], 5); in octeon_sha512_store_hash()
39 write_octeon_64bit_hash_sha512(sctx->state[6], 6); in octeon_sha512_store_hash()
40 write_octeon_64bit_hash_sha512(sctx->state[7], 7); in octeon_sha512_store_hash()
45 sctx->state[0] = read_octeon_64bit_hash_sha512(0); in octeon_sha512_read_hash()
46 sctx->state[1] = read_octeon_64bit_hash_sha512(1); in octeon_sha512_read_hash()
[all …]
Docteon-sha256.c34 u64 *hash = (u64 *)sctx->state; in octeon_sha256_store_hash()
44 u64 *hash = (u64 *)sctx->state; in octeon_sha256_read_hash()
70 sctx->state[0] = SHA224_H0; in octeon_sha224_init()
71 sctx->state[1] = SHA224_H1; in octeon_sha224_init()
72 sctx->state[2] = SHA224_H2; in octeon_sha224_init()
73 sctx->state[3] = SHA224_H3; in octeon_sha224_init()
74 sctx->state[4] = SHA224_H4; in octeon_sha224_init()
75 sctx->state[5] = SHA224_H5; in octeon_sha224_init()
76 sctx->state[6] = SHA224_H6; in octeon_sha224_init()
77 sctx->state[7] = SHA224_H7; in octeon_sha224_init()
[all …]
/arch/s390/include/asm/
Dunwind.h44 void __unwind_start(struct unwind_state *state, struct task_struct *task,
46 bool unwind_next_frame(struct unwind_state *state);
47 unsigned long unwind_get_return_address(struct unwind_state *state);
49 static inline bool unwind_done(struct unwind_state *state) in unwind_done() argument
51 return state->stack_info.type == STACK_TYPE_UNKNOWN; in unwind_done()
54 static inline bool unwind_error(struct unwind_state *state) in unwind_error() argument
56 return state->error; in unwind_error()
59 static inline void unwind_start(struct unwind_state *state, in unwind_start() argument
65 __unwind_start(state, task, regs, sp); in unwind_start()
68 static inline struct pt_regs *unwind_get_entry_regs(struct unwind_state *state) in unwind_get_entry_regs() argument
[all …]
Dcpu_mcf.h35 static inline void ctr_set_enable(u64 *state, int ctr_set) in ctr_set_enable() argument
37 *state |= cpumf_ctr_ctl[ctr_set] << CPUMF_LCCTL_ENABLE_SHIFT; in ctr_set_enable()
39 static inline void ctr_set_disable(u64 *state, int ctr_set) in ctr_set_disable() argument
41 *state &= ~(cpumf_ctr_ctl[ctr_set] << CPUMF_LCCTL_ENABLE_SHIFT); in ctr_set_disable()
43 static inline void ctr_set_start(u64 *state, int ctr_set) in ctr_set_start() argument
45 *state |= cpumf_ctr_ctl[ctr_set] << CPUMF_LCCTL_ACTCTL_SHIFT; in ctr_set_start()
47 static inline void ctr_set_stop(u64 *state, int ctr_set) in ctr_set_stop() argument
49 *state &= ~(cpumf_ctr_ctl[ctr_set] << CPUMF_LCCTL_ACTCTL_SHIFT); in ctr_set_stop()
52 static inline void ctr_set_multiple_enable(u64 *state, u64 ctrsets) in ctr_set_multiple_enable() argument
54 *state |= ctrsets << CPUMF_LCCTL_ENABLE_SHIFT; in ctr_set_multiple_enable()
[all …]
/arch/arm64/include/asm/
Dfpsimdmacros.h9 .macro fpsimd_save state, tmpnr
10 stp q0, q1, [\state, #16 * 0]
11 stp q2, q3, [\state, #16 * 2]
12 stp q4, q5, [\state, #16 * 4]
13 stp q6, q7, [\state, #16 * 6]
14 stp q8, q9, [\state, #16 * 8]
15 stp q10, q11, [\state, #16 * 10]
16 stp q12, q13, [\state, #16 * 12]
17 stp q14, q15, [\state, #16 * 14]
18 stp q16, q17, [\state, #16 * 16]
[all …]
/arch/s390/crypto/
Dsha256_s390.c24 sctx->state[0] = SHA256_H0; in s390_sha256_init()
25 sctx->state[1] = SHA256_H1; in s390_sha256_init()
26 sctx->state[2] = SHA256_H2; in s390_sha256_init()
27 sctx->state[3] = SHA256_H3; in s390_sha256_init()
28 sctx->state[4] = SHA256_H4; in s390_sha256_init()
29 sctx->state[5] = SHA256_H5; in s390_sha256_init()
30 sctx->state[6] = SHA256_H6; in s390_sha256_init()
31 sctx->state[7] = SHA256_H7; in s390_sha256_init()
44 memcpy(octx->state, sctx->state, sizeof(octx->state)); in sha256_export()
55 memcpy(sctx->state, ictx->state, sizeof(ictx->state)); in sha256_import()
[all …]
Dsha512_s390.c25 *(__u64 *)&ctx->state[0] = 0x6a09e667f3bcc908ULL; in sha512_init()
26 *(__u64 *)&ctx->state[2] = 0xbb67ae8584caa73bULL; in sha512_init()
27 *(__u64 *)&ctx->state[4] = 0x3c6ef372fe94f82bULL; in sha512_init()
28 *(__u64 *)&ctx->state[6] = 0xa54ff53a5f1d36f1ULL; in sha512_init()
29 *(__u64 *)&ctx->state[8] = 0x510e527fade682d1ULL; in sha512_init()
30 *(__u64 *)&ctx->state[10] = 0x9b05688c2b3e6c1fULL; in sha512_init()
31 *(__u64 *)&ctx->state[12] = 0x1f83d9abfb41bd6bULL; in sha512_init()
32 *(__u64 *)&ctx->state[14] = 0x5be0cd19137e2179ULL; in sha512_init()
46 memcpy(octx->state, sctx->state, sizeof(octx->state)); in sha512_export()
60 memcpy(sctx->state, ictx->state, sizeof(ictx->state)); in sha512_import()
[all …]
/arch/powerpc/kvm/
Dbook3s_xive.c393 struct kvmppc_xive_irq_state *state) in xive_lock_and_mask() argument
406 old_prio = state->guest_priority; in xive_lock_and_mask()
407 state->guest_priority = MASKED; in xive_lock_and_mask()
409 if (!state->in_eoi) in xive_lock_and_mask()
411 state->guest_priority = old_prio; in xive_lock_and_mask()
420 kvmppc_xive_select_irq(state, &hw_num, &xd); in xive_lock_and_mask()
436 kvmppc_xive_vp(xive, state->act_server), in xive_lock_and_mask()
437 MASKED, state->number); in xive_lock_and_mask()
439 state->old_p = true; in xive_lock_and_mask()
440 state->old_q = false; in xive_lock_and_mask()
[all …]
/arch/x86/crypto/
Dchacha_glue.c19 asmlinkage void chacha_block_xor_ssse3(u32 *state, u8 *dst, const u8 *src,
21 asmlinkage void chacha_4block_xor_ssse3(u32 *state, u8 *dst, const u8 *src,
23 asmlinkage void hchacha_block_ssse3(const u32 *state, u32 *out, int nrounds);
25 asmlinkage void chacha_2block_xor_avx2(u32 *state, u8 *dst, const u8 *src,
27 asmlinkage void chacha_4block_xor_avx2(u32 *state, u8 *dst, const u8 *src,
29 asmlinkage void chacha_8block_xor_avx2(u32 *state, u8 *dst, const u8 *src,
33 asmlinkage void chacha_2block_xor_avx512vl(u32 *state, u8 *dst, const u8 *src,
35 asmlinkage void chacha_4block_xor_avx512vl(u32 *state, u8 *dst, const u8 *src,
37 asmlinkage void chacha_8block_xor_avx512vl(u32 *state, u8 *dst, const u8 *src,
49 static void chacha_dosimd(u32 *state, u8 *dst, const u8 *src, in chacha_dosimd() argument
[all …]
/arch/powerpc/crypto/
Dsha256-spe-glue.c34 extern void ppc_spe_sha256_transform(u32 *state, const u8 *src, u32 blocks);
64 sctx->state[0] = SHA256_H0; in ppc_spe_sha256_init()
65 sctx->state[1] = SHA256_H1; in ppc_spe_sha256_init()
66 sctx->state[2] = SHA256_H2; in ppc_spe_sha256_init()
67 sctx->state[3] = SHA256_H3; in ppc_spe_sha256_init()
68 sctx->state[4] = SHA256_H4; in ppc_spe_sha256_init()
69 sctx->state[5] = SHA256_H5; in ppc_spe_sha256_init()
70 sctx->state[6] = SHA256_H6; in ppc_spe_sha256_init()
71 sctx->state[7] = SHA256_H7; in ppc_spe_sha256_init()
81 sctx->state[0] = SHA224_H0; in ppc_spe_sha224_init()
[all …]
/arch/x86/include/asm/
Dunwind.h38 void __unwind_start(struct unwind_state *state, struct task_struct *task,
40 bool unwind_next_frame(struct unwind_state *state);
41 unsigned long unwind_get_return_address(struct unwind_state *state);
42 unsigned long *unwind_get_return_address_ptr(struct unwind_state *state);
44 static inline bool unwind_done(struct unwind_state *state) in unwind_done() argument
46 return state->stack_info.type == STACK_TYPE_UNKNOWN; in unwind_done()
49 static inline bool unwind_error(struct unwind_state *state) in unwind_error() argument
51 return state->error; in unwind_error()
55 void unwind_start(struct unwind_state *state, struct task_struct *task, in unwind_start() argument
60 __unwind_start(state, task, regs, first_frame); in unwind_start()
[all …]
/arch/x86/boot/
Dcmdline.c40 } state = st_wordstart; in __cmdline_find_option() local
49 switch (state) { in __cmdline_find_option()
55 state = st_wordcmp; in __cmdline_find_option()
63 state = st_bufcpy; in __cmdline_find_option()
65 state = st_wordstart; in __cmdline_find_option()
67 state = st_wordskip; in __cmdline_find_option()
73 state = st_wordstart; in __cmdline_find_option()
78 state = st_wordstart; in __cmdline_find_option()
110 } state = st_wordstart; in __cmdline_find_option_bool() local
122 switch (state) { in __cmdline_find_option_bool()
[all …]
/arch/s390/include/asm/fpu/
Dapi.h90 void __kernel_fpu_begin(struct kernel_fpu *state, u32 flags);
91 void __kernel_fpu_end(struct kernel_fpu *state, u32 flags);
94 static inline void kernel_fpu_begin(struct kernel_fpu *state, u32 flags) in kernel_fpu_begin() argument
97 state->mask = S390_lowcore.fpu_flags; in kernel_fpu_begin()
101 else if (state->mask & flags) in kernel_fpu_begin()
103 __kernel_fpu_begin(state, flags); in kernel_fpu_begin()
107 static inline void kernel_fpu_end(struct kernel_fpu *state, u32 flags) in kernel_fpu_end() argument
109 S390_lowcore.fpu_flags = state->mask; in kernel_fpu_end()
110 if (state->mask & flags) in kernel_fpu_end()
112 __kernel_fpu_end(state, flags); in kernel_fpu_end()
/arch/x86/lib/
Dcmdline.c38 } state = st_wordstart; in __cmdline_find_option_bool() local
51 switch (state) { in __cmdline_find_option_bool()
58 state = st_wordcmp; in __cmdline_find_option_bool()
91 state = st_wordskip; in __cmdline_find_option_bool()
98 state = st_wordstart; in __cmdline_find_option_bool()
133 } state = st_wordstart; in __cmdline_find_option() local
147 switch (state) { in __cmdline_find_option()
152 state = st_wordcmp; in __cmdline_find_option()
165 state = st_bufcpy; in __cmdline_find_option()
174 state = st_wordskip; in __cmdline_find_option()
[all …]
/arch/um/drivers/
Dhostaudio_kern.c73 struct hostaudio_state *state = file->private_data; in hostaudio_read() local
85 err = os_read_file(state->fd, kbuf, count); in hostaudio_read()
100 struct hostaudio_state *state = file->private_data; in hostaudio_write() local
112 err = os_write_file(state->fd, kbuf, count); in hostaudio_write()
137 struct hostaudio_state *state = file->private_data; in hostaudio_ioctl() local
158 err = os_ioctl_generic(state->fd, cmd, (unsigned long) &data); in hostaudio_ioctl()
179 struct hostaudio_state *state; in hostaudio_open() local
189 state = kmalloc(sizeof(struct hostaudio_state), GFP_KERNEL); in hostaudio_open()
190 if (state == NULL) in hostaudio_open()
205 kfree(state); in hostaudio_open()
[all …]
/arch/arm/mm/
Dproc-v7-bugs.c120 unsigned int state, method = 0; in cpu_v7_spectre_v2_init() local
129 state = SPECTRE_MITIGATED; in cpu_v7_spectre_v2_init()
135 state = SPECTRE_MITIGATED; in cpu_v7_spectre_v2_init()
141 state = SPECTRE_UNAFFECTED; in cpu_v7_spectre_v2_init()
147 state = SPECTRE_UNAFFECTED; in cpu_v7_spectre_v2_init()
156 state = spectre_v2_get_cpu_fw_mitigation_state(); in cpu_v7_spectre_v2_init()
157 if (state != SPECTRE_MITIGATED) in cpu_v7_spectre_v2_init()
181 state = SPECTRE_VULNERABLE; in cpu_v7_spectre_v2_init()
187 if (state == SPECTRE_MITIGATED) in cpu_v7_spectre_v2_init()
188 state = spectre_v2_install_workaround(method); in cpu_v7_spectre_v2_init()
[all …]

12345678910>>...34