/arch/arm64/mm/ |
D | context.c | 137 u64 asid = atomic64_read(&mm->context.id); in new_context() 138 u64 generation = atomic64_read(&asid_generation); in new_context() 191 asid = atomic64_read(&mm->context.id); in check_and_switch_context() 207 old_active_asid = atomic64_read(&per_cpu(active_asids, cpu)); in check_and_switch_context() 209 !((asid ^ atomic64_read(&asid_generation)) >> asid_bits) && in check_and_switch_context() 216 asid = atomic64_read(&mm->context.id); in check_and_switch_context() 217 if ((asid ^ atomic64_read(&asid_generation)) >> asid_bits) { in check_and_switch_context()
|
/arch/arm/mm/ |
D | context.c | 192 u64 asid = atomic64_read(&mm->context.id); in new_context() 193 u64 generation = atomic64_read(&asid_generation); in new_context() 253 asid = atomic64_read(&mm->context.id); in check_and_switch_context() 254 if (!((asid ^ atomic64_read(&asid_generation)) >> ASID_BITS) in check_and_switch_context() 260 asid = atomic64_read(&mm->context.id); in check_and_switch_context() 261 if ((asid ^ atomic64_read(&asid_generation)) >> ASID_BITS) { in check_and_switch_context()
|
/arch/csky/include/asm/ |
D | asid.h | 48 asid = atomic64_read(pasid); in asid_check_context() 64 old_active_asid = atomic64_read(&active_asid(info, cpu)); in asid_check_context() 66 !((asid ^ atomic64_read(&info->generation)) >> info->bits) && in asid_check_context()
|
D | mmu_context.h | 24 #define cpu_asid(mm) (atomic64_read(&mm->context.asid) & ASID_MASK)
|
/arch/csky/mm/ |
D | asid.c | 83 u64 asid = atomic64_read(pasid); in new_context() 84 u64 generation = atomic64_read(&info->generation); in new_context() 145 asid = atomic64_read(pasid); in asid_new_context() 146 if ((asid ^ atomic64_read(&info->generation)) >> info->bits) { in asid_new_context()
|
/arch/mips/mm/ |
D | context.c | 130 version = atomic64_read(&mmid_version); in get_new_mmid() 212 !asid_versions_eq(cpu, ctx, atomic64_read(&mmid_version)) || in check_switch_mmu_context() 217 if (!asid_versions_eq(cpu, ctx, atomic64_read(&mmid_version))) in check_switch_mmu_context()
|
/arch/s390/kernel/ |
D | vtime.c | 63 return elapsed >= atomic64_read(&virt_timer_current); in virt_timer_forward() 283 elapsed = atomic64_read(&virt_timer_elapsed); in virt_timer_expire() 306 atomic64_read(&virt_timer_elapsed); in virt_timer_expire() 335 timer->expires += atomic64_read(&virt_timer_elapsed); in internal_add_vtimer() 337 (s64) atomic64_read(&virt_timer_current))) in internal_add_vtimer()
|
D | perf_cpum_cf_common.c | 133 alert = atomic64_read(&cpuhw->alert); in kernel_cpumcf_alert()
|
/arch/ia64/include/asm/ |
D | atomic.h | 26 #define atomic64_read(v) READ_ONCE((v)->counter) macro 135 old = atomic64_read(v); \ 150 old = atomic64_read(v); \
|
/arch/x86/mm/ |
D | tlb.c | 367 next_tlb_gen = atomic64_read(&next->context.tlb_gen); in switch_mm_irqs_off() 411 next_tlb_gen = atomic64_read(&next->context.tlb_gen); in switch_mm_irqs_off() 492 u64 tlb_gen = atomic64_read(&init_mm.context.tlb_gen); in initialize_tlbstate_and_flush() 541 u64 mm_tlb_gen = atomic64_read(&loaded_mm->context.tlb_gen); in flush_tlb_func_common()
|
/arch/sparc/include/asm/ |
D | atomic_64.h | 19 #define atomic64_read(v) READ_ONCE((v)->counter) macro
|
/arch/x86/kernel/ |
D | pvclock.c | 103 last = atomic64_read(&last_value); in pvclock_clocksource_read()
|
/arch/mips/include/asm/ |
D | mmu_context.h | 109 return atomic64_read(&mm->context.mmid); in cpu_context()
|
D | atomic.h | 248 #define atomic64_read(v) READ_ONCE((v)->counter) macro
|
/arch/s390/include/asm/ |
D | atomic.h | 87 static inline s64 atomic64_read(const atomic64_t *v) in atomic64_read() function
|
/arch/s390/pci/ |
D | pci_debug.c | 78 atomic64_read(counter)); in pci_sw_counter_show()
|
/arch/arm/include/asm/ |
D | atomic.h | 255 static inline s64 atomic64_read(const atomic64_t *v) in atomic64_read() function 277 static inline s64 atomic64_read(const atomic64_t *v) in atomic64_read() function
|
/arch/parisc/include/asm/ |
D | atomic.h | 218 atomic64_read(const atomic64_t *v) in atomic64_read() function
|
/arch/alpha/include/asm/ |
D | atomic.h | 31 #define atomic64_read(v) READ_ONCE((v)->counter) macro
|
/arch/arc/include/asm/ |
D | atomic.h | 329 static inline s64 atomic64_read(const atomic64_t *v)
|
/arch/powerpc/include/asm/ |
D | atomic.h | 300 static __inline__ s64 atomic64_read(const atomic64_t *v) in atomic64_read() function
|
/arch/riscv/include/asm/ |
D | atomic.h | 41 static __always_inline s64 atomic64_read(const atomic64_t *v) in atomic64_read() function
|
/arch/powerpc/platforms/pseries/ |
D | lparcfg.c | 66 return atomic64_read(&purr); in get_purr()
|
/arch/x86/kvm/ |
D | hyperv.c | 44 return atomic64_read(&synic->sint[sint]); in synic_read_sint() 297 *pdata = atomic64_read(&synic->sint[msr - HV_X64_MSR_SINT0]); in synic_get_msr()
|
/arch/powerpc/kvm/ |
D | book3s_hv_rm_mmu.c | 1225 mmio_update = atomic64_read(&kvm->arch.mmio_update); in kvmppc_hpte_hv_fault()
|