Home
last modified time | relevance | path

Searched refs:sve_state (Results 1 – 14 of 14) sorted by relevance

/arch/arm64/kvm/
Dreset.c102 vcpu->arch.sve_state = buf; in alloc_sve_state()
161 void *sve_state = vcpu->arch.sve_state; in kvm_arm_vcpu_destroy() local
165 if (sve_state) { in kvm_arm_vcpu_destroy()
171 kvm_unshare_hyp(sve_state, sve_state + reg_sz); in kvm_arm_vcpu_destroy()
172 free_pages_exact(sve_state, reg_sz); in kvm_arm_vcpu_destroy()
181 memset(vcpu->arch.sve_state, 0, vcpu_sve_state_size(vcpu)); in kvm_vcpu_reset_sve()
Dfpsimd.c136 fp_state.sve_state = vcpu->arch.sve_state; in kvm_arch_vcpu_ctxsync_fp()
Dguest.c351 if (WARN_ON(vcpu->arch.sve_state)) in set_sve_vls()
506 if (copy_to_user(uptr, vcpu->arch.sve_state + region.koffset, in get_sve_reg()
532 if (copy_from_user(vcpu->arch.sve_state + region.koffset, uptr, in set_sve_reg()
Dpkvm.c244 vcpu->arch.sve_state = NULL; in __pkvm_vcpu_hyp_created()
/arch/arm64/kernel/
Dfpsimd.c517 sve_save_state((char *)last->sve_state + in fpsimd_save()
632 #define ZREG(sve_state, vq, n) ((char *)(sve_state) + \ argument
679 void *sst = task->thread.sve_state; in fpsimd_to_sve()
703 void const *sst = task->thread.sve_state; in sve_to_fpsimd()
726 kfree(task->thread.sve_state); in __sve_free()
727 task->thread.sve_state = NULL; in __sve_free()
765 if (task->thread.sve_state) { in sve_alloc()
767 memset(task->thread.sve_state, 0, in sve_alloc()
773 task->thread.sve_state = in sve_alloc()
835 void *sst = task->thread.sve_state; in sve_sync_from_fpsimd_zeropad()
[all …]
Dprocess.c308 dst->thread.sve_state = NULL; in arch_dup_task_struct()
320 dst->thread.sve_state = kzalloc(sve_state_size(src), in arch_dup_task_struct()
322 if (!dst->thread.sve_state) in arch_dup_task_struct()
329 kfree(dst->thread.sve_state); in arch_dup_task_struct()
330 dst->thread.sve_state = NULL; in arch_dup_task_struct()
Dptrace.c808 membuf_write(&to, target->thread.sve_state, end - start); in sve_get_common()
932 if (!target->thread.sve_state) { in sve_set_common()
955 target->thread.sve_state, in sve_set_common()
1101 if (!target->thread.sve_state) { in za_set()
1103 if (!target->thread.sve_state) { in za_set()
1183 if (!target->thread.sve_state) in zt_set()
Dsignal.c268 current->thread.sve_state, in preserve_sve_context()
333 if (!current->thread.sve_state) { in restore_sve_fpsimd_context()
338 err = __copy_from_user(current->thread.sve_state, in restore_sve_fpsimd_context()
/arch/arm64/kvm/hyp/nvhe/
Dpkvm.c594 void *sve_state; in unpin_host_sve_state() local
599 sve_state = kern_hyp_va(hyp_vcpu->vcpu.arch.sve_state); in unpin_host_sve_state()
600 hyp_unpin_shared_mem(sve_state, in unpin_host_sve_state()
601 sve_state + vcpu_sve_state_size(&hyp_vcpu->vcpu)); in unpin_host_sve_state()
607 void *sve_state = hyp_vcpu->vcpu.arch.sve_state; in teardown_sve_state() local
609 if (sve_state) in teardown_sve_state()
610 hyp_free_account(sve_state, hyp_vm->host_kvm); in teardown_sve_state()
654 void *sve_state = kern_hyp_va(READ_ONCE(host_vcpu->arch.sve_state)); in init_pkvm_hyp_vcpu_sve() local
659 if (!sve_state && !pkvm_hyp_vcpu_is_protected(hyp_vcpu)) { in init_pkvm_hyp_vcpu_sve()
672 sve_state = hyp_alloc_account(sve_state_size, in init_pkvm_hyp_vcpu_sve()
[all …]
Dswitch.c273 struct kvm_host_sve_state *sve_state = get_host_sve_state(vcpu); in kvm_hyp_handle_fpsimd_host() local
276 sve_state->zcr_el1 = read_sysreg_el1(SYS_ZCR); in kvm_hyp_handle_fpsimd_host()
278 __sve_save_state(sve_state->sve_regs + in kvm_hyp_handle_fpsimd_host()
280 &sve_state->fpsr); in kvm_hyp_handle_fpsimd_host()
Dhyp-main.c779 struct kvm_host_sve_state *sve_state = get_host_sve_state(vcpu); in __hyp_sve_restore_host() local
782 write_sysreg_el1(sve_state->zcr_el1, SYS_ZCR); in __hyp_sve_restore_host()
784 __sve_restore_state(sve_state->sve_regs + in __hyp_sve_restore_host()
786 &sve_state->fpsr); in __hyp_sve_restore_host()
/arch/arm64/include/asm/
Dfpsimd.h63 void *sve_state; member
105 return (char *)thread->sve_state + sve_ffr_offset(vl); in sve_pffr()
Dprocessor.h164 void *sve_state; /* SVE registers, if any */ member
Dkvm_host.h577 void *sve_state; member
866 #define vcpu_sve_pffr(vcpu) (kern_hyp_va((vcpu)->arch.sve_state) + \