/kernel/linux/linux-5.10/arch/sparc/kernel/ |
D | nmi.c | 102 if (__this_cpu_read(nmi_touch)) { in perfctr_irq() 106 if (!touched && __this_cpu_read(last_irq_sum) == sum) { in perfctr_irq() 108 if (__this_cpu_read(alert_counter) == 30 * nmi_hz) in perfctr_irq() 115 if (__this_cpu_read(wd_enabled)) { in perfctr_irq() 155 if (!__this_cpu_read(wd_enabled)) in stop_nmi_watchdog() 212 if (__this_cpu_read(wd_enabled)) in start_nmi_watchdog() 226 if (!__this_cpu_read(wd_enabled)) in nmi_adjust_hz_one()
|
/kernel/linux/linux-5.10/arch/mips/kernel/ |
D | mips-r2-to-r6-emul.c | 2243 (unsigned long)__this_cpu_read(mipsr2emustats.movs), in mipsr2_emul_show() 2244 (unsigned long)__this_cpu_read(mipsr2bdemustats.movs)); in mipsr2_emul_show() 2246 (unsigned long)__this_cpu_read(mipsr2emustats.hilo), in mipsr2_emul_show() 2247 (unsigned long)__this_cpu_read(mipsr2bdemustats.hilo)); in mipsr2_emul_show() 2249 (unsigned long)__this_cpu_read(mipsr2emustats.muls), in mipsr2_emul_show() 2250 (unsigned long)__this_cpu_read(mipsr2bdemustats.muls)); in mipsr2_emul_show() 2252 (unsigned long)__this_cpu_read(mipsr2emustats.divs), in mipsr2_emul_show() 2253 (unsigned long)__this_cpu_read(mipsr2bdemustats.divs)); in mipsr2_emul_show() 2255 (unsigned long)__this_cpu_read(mipsr2emustats.dsps), in mipsr2_emul_show() 2256 (unsigned long)__this_cpu_read(mipsr2bdemustats.dsps)); in mipsr2_emul_show() [all …]
|
/kernel/linux/linux-5.10/arch/x86/include/asm/ |
D | irq_stack.h | 12 return __this_cpu_read(irq_count) != -1; in irqstack_active() 23 void *tos = __this_cpu_read(hardirq_stack_ptr); in __run_on_irqstack() 34 void *tos = __this_cpu_read(hardirq_stack_ptr); in __run_sysvec_on_irqstack() 45 void *tos = __this_cpu_read(hardirq_stack_ptr); in __run_irq_on_irqstack()
|
D | cpu_entry_area.h | 148 CEA_ESTACK_TOP(__this_cpu_read(cea_exception_stacks), name) 151 CEA_ESTACK_BOT(__this_cpu_read(cea_exception_stacks), name)
|
/kernel/linux/linux-5.10/kernel/ |
D | context_tracking.c | 70 if ( __this_cpu_read(context_tracking.state) != state) { in __context_tracking_enter() 71 if (__this_cpu_read(context_tracking.active)) { in __context_tracking_enter() 151 if (__this_cpu_read(context_tracking.state) == state) { in __context_tracking_exit() 152 if (__this_cpu_read(context_tracking.active)) { in __context_tracking_exit()
|
D | watchdog_hld.c | 80 delta = now - __this_cpu_read(last_timestamp); in watchdog_check_timestamp() 117 if (__this_cpu_read(watchdog_nmi_touch) == true) { in watchdog_overflow_callback() 135 if (__this_cpu_read(hard_watchdog_warn) == true) in watchdog_overflow_callback()
|
D | softirq.c | 73 struct task_struct *tsk = __this_cpu_read(ksoftirqd); in wakeup_softirqd() 87 struct task_struct *tsk = __this_cpu_read(ksoftirqd); in ksoftirqd_running() 310 if (__this_cpu_read(ksoftirqd) == current) in __do_softirq() 701 *__this_cpu_read(tasklet_vec.tail) = per_cpu(tasklet_vec, cpu).head; in takeover_tasklets() 709 *__this_cpu_read(tasklet_hi_vec.tail) = per_cpu(tasklet_hi_vec, cpu).head; in takeover_tasklets()
|
D | watchdog.c | 308 unsigned long hrint = __this_cpu_read(hrtimer_interrupts); in is_hardlockup() 310 if (__this_cpu_read(hrtimer_interrupts_saved) == hrint) in is_hardlockup() 344 unsigned long touch_ts = __this_cpu_read(watchdog_touch_ts); in watchdog_timer_fn() 367 if (unlikely(__this_cpu_read(softlockup_touch_sync))) { in watchdog_timer_fn()
|
/kernel/linux/linux-5.10/kernel/time/ |
D | tick-oneshot.c | 25 struct clock_event_device *dev = __this_cpu_read(tick_cpu_device.evtdev); in tick_program_event() 52 struct clock_event_device *dev = __this_cpu_read(tick_cpu_device.evtdev); in tick_resume_oneshot() 112 ret = __this_cpu_read(tick_cpu_device.mode) == TICKDEV_MODE_ONESHOT; in tick_oneshot_mode_active()
|
/kernel/linux/linux-5.10/arch/x86/kernel/cpu/mce/ |
D | intel.c | 132 if (__this_cpu_read(cmci_storm_state) == CMCI_STORM_NONE) in mce_intel_cmci_poll() 179 (__this_cpu_read(cmci_storm_state) == CMCI_STORM_ACTIVE)) { in cmci_intel_adjust_timer() 184 switch (__this_cpu_read(cmci_storm_state)) { in cmci_intel_adjust_timer() 218 unsigned int cnt = __this_cpu_read(cmci_storm_cnt); in cmci_storm_detect() 219 unsigned long ts = __this_cpu_read(cmci_time_stamp); in cmci_storm_detect() 223 if (__this_cpu_read(cmci_storm_state) != CMCI_STORM_NONE) in cmci_storm_detect()
|
/kernel/linux/linux-5.10/include/asm-generic/ |
D | irq_regs.h | 21 return __this_cpu_read(__irq_regs); in get_irq_regs() 28 old_regs = __this_cpu_read(__irq_regs); in set_irq_regs()
|
/kernel/linux/linux-5.10/include/linux/ |
D | context_tracking_state.h | 41 return context_tracking_enabled() && __this_cpu_read(context_tracking.active); in context_tracking_enabled_this_cpu() 46 return __this_cpu_read(context_tracking.state) == CONTEXT_USER; in context_tracking_in_user()
|
/kernel/linux/linux-5.10/lib/ |
D | percpu_test.c | 11 WARN(__this_cpu_read(pcp) != (expected), \ 13 __this_cpu_read(pcp), __this_cpu_read(pcp), \
|
/kernel/linux/linux-5.10/arch/x86/kernel/ |
D | hw_breakpoint.c | 485 set_debugreg(__this_cpu_read(cpu_debugreg[0]), 0); in hw_breakpoint_restore() 486 set_debugreg(__this_cpu_read(cpu_debugreg[1]), 1); in hw_breakpoint_restore() 487 set_debugreg(__this_cpu_read(cpu_debugreg[2]), 2); in hw_breakpoint_restore() 488 set_debugreg(__this_cpu_read(cpu_debugreg[3]), 3); in hw_breakpoint_restore() 490 set_debugreg(__this_cpu_read(cpu_dr7), 7); in hw_breakpoint_restore()
|
D | irq_32.c | 79 irqstk = __this_cpu_read(hardirq_stack_ptr); in execute_on_irq_stack() 139 irqstk = __this_cpu_read(softirq_stack_ptr); in do_softirq_own_stack()
|
D | irq.c | 247 desc = __this_cpu_read(vector_irq[vector]); in DEFINE_IDTENTRY_IRQ() 358 if (IS_ERR_OR_NULL(__this_cpu_read(vector_irq[vector]))) in fixup_irqs() 363 desc = __this_cpu_read(vector_irq[vector]); in fixup_irqs() 374 if (__this_cpu_read(vector_irq[vector]) != VECTOR_RETRIGGERED) in fixup_irqs()
|
D | kvm.c | 231 if (__this_cpu_read(apf_reason.enabled)) { in kvm_read_and_reset_apf_flags() 232 flags = __this_cpu_read(apf_reason.flags); in kvm_read_and_reset_apf_flags() 282 if (__this_cpu_read(apf_reason.enabled)) { in DEFINE_IDTENTRY_SYSVEC() 283 token = __this_cpu_read(apf_reason.token); in DEFINE_IDTENTRY_SYSVEC() 370 if (!__this_cpu_read(apf_reason.enabled)) in kvm_pv_disable_apf()
|
/kernel/linux/linux-5.10/drivers/cpuidle/ |
D | cpuidle-psci.c | 47 return __this_cpu_read(domain_state); in psci_get_domain_state() 88 struct device *pd_dev = __this_cpu_read(psci_cpuidle_data.dev); in psci_idle_cpuhp_up() 98 struct device *pd_dev = __this_cpu_read(psci_cpuidle_data.dev); in psci_idle_cpuhp_down() 127 u32 *state = __this_cpu_read(psci_cpuidle_data.psci_states); in psci_enter_idle_state()
|
/kernel/linux/linux-5.10/drivers/xen/events/ |
D | events_2l.c | 123 struct vcpu_info *vcpu_info = __this_cpu_read(xen_vcpu); in evtchn_2l_unmask() 173 struct vcpu_info *vcpu_info = __this_cpu_read(xen_vcpu); in evtchn_2l_handle_events() 192 start_word_idx = __this_cpu_read(current_word_idx); in evtchn_2l_handle_events() 193 start_bit_idx = __this_cpu_read(current_bit_idx); in evtchn_2l_handle_events()
|
/kernel/linux/linux-5.10/arch/arm64/kvm/hyp/vhe/ |
D | switch.c | 69 write_sysreg(__this_cpu_read(kvm_hyp_vector), vbar_el1); in __activate_traps() 91 host_vectors = __this_cpu_read(this_cpu_vector); in __deactivate_traps()
|
/kernel/linux/linux-5.10/drivers/irqchip/ |
D | irq-xtensa-mx.c | 82 mask = __this_cpu_read(cached_irq_mask) & ~mask; in xtensa_mx_irq_mask() 100 mask |= __this_cpu_read(cached_irq_mask); in xtensa_mx_irq_unmask()
|
/kernel/linux/linux-5.10/arch/arm64/include/asm/ |
D | arch_timer.h | 27 __wa = __this_cpu_read(timer_unstable_counter_workaround); \ 34 __wa = __this_cpu_read(timer_unstable_counter_workaround); \
|
/kernel/linux/linux-5.10/arch/powerpc/kernel/ |
D | mce.c | 194 int index = __this_cpu_read(mce_nest_count) - 1; in get_mce_event() 292 while (__this_cpu_read(mce_ue_count) > 0) { in machine_process_ue_event() 293 index = __this_cpu_read(mce_ue_count) - 1; in machine_process_ue_event() 341 while (__this_cpu_read(mce_queue_count) > 0) { in machine_check_process_queued_event() 342 index = __this_cpu_read(mce_queue_count) - 1; in machine_check_process_queued_event()
|
/kernel/linux/linux-5.10/mm/ |
D | vmstat.c | 324 x = delta + __this_cpu_read(*p); in __mod_zone_page_state() 326 t = __this_cpu_read(pcp->stat_threshold); in __mod_zone_page_state() 349 x = delta + __this_cpu_read(*p); in __mod_node_page_state() 351 t = __this_cpu_read(pcp->stat_threshold); in __mod_node_page_state() 391 t = __this_cpu_read(pcp->stat_threshold); in __inc_zone_state() 409 t = __this_cpu_read(pcp->stat_threshold); in __inc_node_state() 437 t = __this_cpu_read(pcp->stat_threshold); in __dec_zone_state() 455 t = __this_cpu_read(pcp->stat_threshold); in __dec_node_state() 810 if (!__this_cpu_read(p->expire) || in refresh_cpu_vm_stats() 811 !__this_cpu_read(p->pcp.count)) in refresh_cpu_vm_stats() [all …]
|
/kernel/linux/linux-5.10/arch/x86/oprofile/ |
D | op_model_ppro.c | 87 __this_cpu_read(cpu_info.x86) == 6 && in ppro_setup_ctrs() 88 __this_cpu_read(cpu_info.x86_model) == 15)) { in ppro_setup_ctrs()
|