Home
last modified time | relevance | path

Searched refs:atomic64_read (Results 1 – 25 of 27) sorted by relevance

12

/arch/arm64/mm/
Dcontext.c137 u64 asid = atomic64_read(&mm->context.id); in new_context()
138 u64 generation = atomic64_read(&asid_generation); in new_context()
191 asid = atomic64_read(&mm->context.id); in check_and_switch_context()
207 old_active_asid = atomic64_read(&per_cpu(active_asids, cpu)); in check_and_switch_context()
209 !((asid ^ atomic64_read(&asid_generation)) >> asid_bits) && in check_and_switch_context()
216 asid = atomic64_read(&mm->context.id); in check_and_switch_context()
217 if ((asid ^ atomic64_read(&asid_generation)) >> asid_bits) { in check_and_switch_context()
/arch/arm/mm/
Dcontext.c192 u64 asid = atomic64_read(&mm->context.id); in new_context()
193 u64 generation = atomic64_read(&asid_generation); in new_context()
253 asid = atomic64_read(&mm->context.id); in check_and_switch_context()
254 if (!((asid ^ atomic64_read(&asid_generation)) >> ASID_BITS) in check_and_switch_context()
260 asid = atomic64_read(&mm->context.id); in check_and_switch_context()
261 if ((asid ^ atomic64_read(&asid_generation)) >> ASID_BITS) { in check_and_switch_context()
/arch/csky/include/asm/
Dasid.h48 asid = atomic64_read(pasid); in asid_check_context()
64 old_active_asid = atomic64_read(&active_asid(info, cpu)); in asid_check_context()
66 !((asid ^ atomic64_read(&info->generation)) >> info->bits) && in asid_check_context()
Dmmu_context.h24 #define cpu_asid(mm) (atomic64_read(&mm->context.asid) & ASID_MASK)
/arch/csky/mm/
Dasid.c83 u64 asid = atomic64_read(pasid); in new_context()
84 u64 generation = atomic64_read(&info->generation); in new_context()
145 asid = atomic64_read(pasid); in asid_new_context()
146 if ((asid ^ atomic64_read(&info->generation)) >> info->bits) { in asid_new_context()
/arch/mips/mm/
Dcontext.c130 version = atomic64_read(&mmid_version); in get_new_mmid()
212 !asid_versions_eq(cpu, ctx, atomic64_read(&mmid_version)) || in check_switch_mmu_context()
217 if (!asid_versions_eq(cpu, ctx, atomic64_read(&mmid_version))) in check_switch_mmu_context()
/arch/s390/kernel/
Dvtime.c63 return elapsed >= atomic64_read(&virt_timer_current); in virt_timer_forward()
283 elapsed = atomic64_read(&virt_timer_elapsed); in virt_timer_expire()
306 atomic64_read(&virt_timer_elapsed); in virt_timer_expire()
335 timer->expires += atomic64_read(&virt_timer_elapsed); in internal_add_vtimer()
337 (s64) atomic64_read(&virt_timer_current))) in internal_add_vtimer()
Dperf_cpum_cf_common.c133 alert = atomic64_read(&cpuhw->alert); in kernel_cpumcf_alert()
/arch/ia64/include/asm/
Datomic.h26 #define atomic64_read(v) READ_ONCE((v)->counter) macro
135 old = atomic64_read(v); \
150 old = atomic64_read(v); \
/arch/x86/mm/
Dtlb.c367 next_tlb_gen = atomic64_read(&next->context.tlb_gen); in switch_mm_irqs_off()
411 next_tlb_gen = atomic64_read(&next->context.tlb_gen); in switch_mm_irqs_off()
492 u64 tlb_gen = atomic64_read(&init_mm.context.tlb_gen); in initialize_tlbstate_and_flush()
541 u64 mm_tlb_gen = atomic64_read(&loaded_mm->context.tlb_gen); in flush_tlb_func_common()
/arch/sparc/include/asm/
Datomic_64.h19 #define atomic64_read(v) READ_ONCE((v)->counter) macro
/arch/x86/kernel/
Dpvclock.c103 last = atomic64_read(&last_value); in pvclock_clocksource_read()
/arch/mips/include/asm/
Dmmu_context.h109 return atomic64_read(&mm->context.mmid); in cpu_context()
Datomic.h248 #define atomic64_read(v) READ_ONCE((v)->counter) macro
/arch/s390/include/asm/
Datomic.h87 static inline s64 atomic64_read(const atomic64_t *v) in atomic64_read() function
/arch/s390/pci/
Dpci_debug.c78 atomic64_read(counter)); in pci_sw_counter_show()
/arch/arm/include/asm/
Datomic.h255 static inline s64 atomic64_read(const atomic64_t *v) in atomic64_read() function
277 static inline s64 atomic64_read(const atomic64_t *v) in atomic64_read() function
/arch/parisc/include/asm/
Datomic.h218 atomic64_read(const atomic64_t *v) in atomic64_read() function
/arch/alpha/include/asm/
Datomic.h31 #define atomic64_read(v) READ_ONCE((v)->counter) macro
/arch/arc/include/asm/
Datomic.h329 static inline s64 atomic64_read(const atomic64_t *v)
/arch/powerpc/include/asm/
Datomic.h300 static __inline__ s64 atomic64_read(const atomic64_t *v) in atomic64_read() function
/arch/riscv/include/asm/
Datomic.h41 static __always_inline s64 atomic64_read(const atomic64_t *v) in atomic64_read() function
/arch/powerpc/platforms/pseries/
Dlparcfg.c66 return atomic64_read(&purr); in get_purr()
/arch/x86/kvm/
Dhyperv.c44 return atomic64_read(&synic->sint[sint]); in synic_read_sint()
297 *pdata = atomic64_read(&synic->sint[msr - HV_X64_MSR_SINT0]); in synic_get_msr()
/arch/powerpc/kvm/
Dbook3s_hv_rm_mmu.c1225 mmio_update = atomic64_read(&kvm->arch.mmio_update); in kvmppc_hpte_hv_fault()

12