Searched refs:amr (Results 1 – 16 of 16) sorted by relevance
/arch/powerpc/include/asm/book3s/64/ |
D | kup.h | 218 return current->thread.regs->amr; in current_thread_amr() 235 unsigned long amr, iamr; in kuap_user_restore() local 241 amr = mfspr(SPRN_AMR); in kuap_user_restore() 242 if (amr != regs->amr) in kuap_user_restore() 260 mtspr(SPRN_AMR, regs->amr); in kuap_user_restore() 272 unsigned long amr) in kuap_kernel_restore() argument 275 if (unlikely(regs->amr != amr)) { in kuap_kernel_restore() 277 mtspr(SPRN_AMR, regs->amr); in kuap_kernel_restore() 293 unsigned long amr = mfspr(SPRN_AMR); in kuap_get_and_assert_locked() local 295 WARN_ON_ONCE(amr != AMR_KUAP_BLOCKED); in kuap_get_and_assert_locked() [all …]
|
/arch/powerpc/mm/book3s64/ |
D | pkeys.c | 318 current->thread.regs->amr = old_amr | new_amr_bits; in init_amr() 416 u64 amr; in pkey_access_permitted() local 422 amr = current_thread_amr(); in pkey_access_permitted() 424 return !(amr & (AMR_WR_BIT << pkey_shift)); in pkey_access_permitted() 426 return !(amr & (AMR_RD_BIT << pkey_shift)); in pkey_access_permitted()
|
/arch/powerpc/kernel/ |
D | interrupt.c | 102 unsigned long amr, iamr; in system_call_exception() local 108 amr = mfspr(SPRN_AMR); in system_call_exception() 110 regs->amr = amr; in system_call_exception()
|
D | asm-offsets.c | 301 STACK_PT_REGS_OFFSET(STACK_REGS_AMR, amr); in main() 443 OFFSET(VCPU_AMR, kvm_vcpu, arch.amr); in main()
|
D | process.c | 1566 current->thread.regs->amr = default_amr; in arch_setup_new_exec() 1788 kregs->amr = AMR_KUAP_BLOCKED; in copy_thread()
|
/arch/powerpc/include/asm/ |
D | kvm_book3s_64.h | 473 static inline int hpte_get_skey_perm(unsigned long hpte_r, unsigned long amr) in hpte_get_skey_perm() argument 479 return (amr >> (62 - 2 * skey)) & 3; in hpte_get_skey_perm() 603 vcpu->arch.amr = vcpu->arch.amr_tm; in copy_from_checkpoint() 620 vcpu->arch.amr_tm = vcpu->arch.amr; in copy_to_checkpoint()
|
D | kup.h | 55 static inline void kuap_kernel_restore(struct pt_regs *regs, unsigned long amr) { } in kuap_kernel_restore() argument
|
D | ptrace.h | 70 unsigned long amr; member
|
D | kvm_host.h | 568 ulong amr; member
|
/arch/powerpc/platforms/powernv/ |
D | idle.c | 307 u64 amr; member 379 sprs.amr = mfspr(SPRN_AMR); in power7_idle_insn() 398 mtspr(SPRN_AMR, sprs.amr); in power7_idle_insn() 601 u64 amr; member 688 sprs.amr = mfspr(SPRN_AMR); in power9_idle_stop() 709 mtspr(SPRN_AMR, sprs.amr); in power9_idle_stop()
|
/arch/powerpc/kernel/ptrace/ |
D | ptrace-view.c | 475 membuf_store(&to, target->thread.regs->amr); in pkey_get() 509 target->thread.regs->amr = (new_amr & default_uamor) | in pkey_set() 510 (target->thread.regs->amr & ~default_uamor); in pkey_set()
|
/arch/powerpc/kvm/ |
D | book3s_emulate.c | 94 vcpu->arch.amr_tm = vcpu->arch.amr; in kvmppc_copyto_vcpu_tm() 113 vcpu->arch.amr = vcpu->arch.amr_tm; in kvmppc_copyfrom_vcpu_tm()
|
D | book3s_hv.c | 2041 *val = get_reg_val(id, vcpu->arch.amr); in kvmppc_get_one_reg_hv() 2286 vcpu->arch.amr = set_reg_val(id, *val); in kvmppc_set_one_reg_hv() 3783 mtspr(SPRN_AMR, vcpu->arch.amr); in load_spr_state() 3809 vcpu->arch.amr = mfspr(SPRN_AMR); in store_spr_state() 3821 unsigned long amr; member 3830 host_os_sprs->amr = mfspr(SPRN_AMR); in save_p9_host_os_sprs() 3845 if (host_os_sprs->amr != vcpu->arch.amr) in restore_p9_host_os_sprs() 3846 mtspr(SPRN_AMR, host_os_sprs->amr); in restore_p9_host_os_sprs()
|
D | book3s_64_mmu_radix.c | 291 if (vcpu->arch.amr & (1ul << 62)) in kvmppc_mmu_radix_xlate() 293 if (vcpu->arch.amr & (1ul << 63)) in kvmppc_mmu_radix_xlate()
|
D | book3s_hv_rm_mmu.c | 1256 unsigned int perm = hpte_get_skey_perm(gr, vcpu->arch.amr); in kvmppc_hpte_hv_fault()
|
D | book3s_64_mmu_hv.c | 390 int amrfield = hpte_get_skey_perm(gr, vcpu->arch.amr); in kvmppc_mmu_book3s_64_hv_xlate()
|