Home
last modified time | relevance | path

Searched defs:cpu (Results 1 – 25 of 144) sorted by relevance

123456

/kernel/
Dsmpboot.c30 struct task_struct *idle_thread_get(unsigned int cpu) in idle_thread_get()
50 static __always_inline void idle_init(unsigned int cpu) in idle_init()
68 unsigned int cpu, boot_cpu; in idle_threads_init() local
85 unsigned int cpu; member
170 __smpboot_create_thread(struct smp_hotplug_thread *ht, unsigned int cpu) in __smpboot_create_thread()
213 int smpboot_create_threads(unsigned int cpu) in smpboot_create_threads()
228 static void smpboot_unpark_thread(struct smp_hotplug_thread *ht, unsigned int cpu) in smpboot_unpark_thread()
236 int smpboot_unpark_threads(unsigned int cpu) in smpboot_unpark_threads()
247 static void smpboot_park_thread(struct smp_hotplug_thread *ht, unsigned int cpu) in smpboot_park_thread()
255 int smpboot_park_threads(unsigned int cpu) in smpboot_park_threads()
[all …]
Dcpu.c169 static int cpuhp_invoke_callback(unsigned int cpu, enum cpuhp_state state, in cpuhp_invoke_callback()
308 static bool cpuhp_wait_for_sync_state(unsigned int cpu, enum cpuhp_sync_state state, in cpuhp_wait_for_sync_state()
354 void __weak arch_cpuhp_cleanup_dead_cpu(unsigned int cpu) { } in arch_cpuhp_cleanup_dead_cpu()
360 static void cpuhp_bp_sync_dead(unsigned int cpu) in cpuhp_bp_sync_dead()
381 static inline void cpuhp_bp_sync_dead(unsigned int cpu) { } in cpuhp_bp_sync_dead()
402 static bool cpuhp_can_boot_ap(unsigned int cpu) in cpuhp_can_boot_ap()
430 void __weak arch_cpuhp_cleanup_kick_cpu(unsigned int cpu) { } in arch_cpuhp_cleanup_kick_cpu()
436 static int cpuhp_bp_sync_alive(unsigned int cpu) in cpuhp_bp_sync_alive()
453 static inline int cpuhp_bp_sync_alive(unsigned int cpu) { return 0; } in cpuhp_bp_sync_alive()
454 static inline bool cpuhp_can_boot_ap(unsigned int cpu) { return true; } in cpuhp_can_boot_ap()
[all …]
Dstop_machine.c91 static bool cpu_stop_queue_work(unsigned int cpu, struct cpu_stop_work *work) in cpu_stop_queue_work()
137 int stop_one_cpu(unsigned int cpu, cpu_stop_fn_t fn, void *arg) in stop_one_cpu()
206 int cpu = smp_processor_id(), err = 0; in multi_cpu_stop() local
385 bool stop_one_cpu_nowait(unsigned int cpu, cpu_stop_fn_t fn, void *arg, in stop_one_cpu_nowait()
398 unsigned int cpu; in queue_stop_cpus_work() local
476 static int cpu_stop_should_run(unsigned int cpu) in cpu_stop_should_run()
488 static void cpu_stopper_thread(unsigned int cpu) in cpu_stopper_thread()
528 void stop_machine_park(int cpu) in stop_machine_park()
540 static void cpu_stop_create(unsigned int cpu) in cpu_stop_create()
545 static void cpu_stop_park(unsigned int cpu) in cpu_stop_park()
[all …]
Dsmp.c54 int smpcfd_prepare_cpu(unsigned int cpu) in smpcfd_prepare_cpu()
76 int smpcfd_dead_cpu(unsigned int cpu) in smpcfd_dead_cpu()
86 int smpcfd_dying_cpu(unsigned int cpu) in smpcfd_dying_cpu()
113 send_call_function_single_ipi(int cpu) in send_call_function_single_ipi()
231 int cpu = -1; in csd_lock_wait_toolong() local
378 void __smp_call_single_queue(int cpu, struct llist_node *node) in __smp_call_single_queue()
419 static int generic_exec_single(int cpu, call_single_data_t *csd) in generic_exec_single()
630 int smp_call_function_single(int cpu, smp_call_func_t func, void *info, in smp_call_function_single()
710 int smp_call_function_single_async(int cpu, call_single_data_t *csd) in smp_call_function_single_async()
750 unsigned int cpu; in smp_call_function_any() local
[all …]
Dwatchdog_perf.c119 unsigned int cpu; in hardlockup_detector_event_create() local
155 void watchdog_hardlockup_enable(unsigned int cpu) in watchdog_hardlockup_enable()
174 void watchdog_hardlockup_disable(unsigned int cpu) in watchdog_hardlockup_disable()
195 int cpu; in hardlockup_detector_perf_stop() local
214 int cpu; in hardlockup_detector_perf_restart() local
Dwatchdog_buddy.c11 static unsigned int watchdog_next_cpu(unsigned int cpu) in watchdog_next_cpu()
30 void watchdog_hardlockup_enable(unsigned int cpu) in watchdog_hardlockup_enable()
64 void watchdog_hardlockup_disable(unsigned int cpu) in watchdog_hardlockup_disable()
Dwatchdog.c128 void watchdog_hardlockup_touch_cpu(unsigned int cpu) in watchdog_hardlockup_touch_cpu()
133 static bool is_hardlockup(unsigned int cpu) in is_hardlockup()
158 void watchdog_hardlockup_check(unsigned int cpu, struct pt_regs *regs) in watchdog_hardlockup_check()
240 void __weak watchdog_hardlockup_enable(unsigned int cpu) { } in watchdog_hardlockup_enable()
242 void __weak watchdog_hardlockup_disable(unsigned int cpu) { } in watchdog_hardlockup_disable()
608 int cpu; in touch_all_softlockup_watchdogs() local
780 static void watchdog_enable(unsigned int cpu) in watchdog_enable()
806 static void watchdog_disable(unsigned int cpu) in watchdog_disable()
830 int cpu; in softlockup_stop_all() local
849 int cpu; in softlockup_start_all() local
[all …]
/kernel/time/
Dtick-broadcast.c47 static inline void tick_broadcast_clear_oneshot(int cpu) { } in tick_broadcast_clear_oneshot()
50 static inline void tick_broadcast_oneshot_offline(unsigned int cpu) { } in tick_broadcast_oneshot_offline()
69 const struct clock_event_device *tick_get_wakeup_device(int cpu) in tick_get_wakeup_device()
102 static struct clock_event_device *tick_get_oneshot_wakeup_device(int cpu) in tick_get_oneshot_wakeup_device()
117 int cpu) in tick_set_oneshot_wakeup_device()
148 static struct clock_event_device *tick_get_oneshot_wakeup_device(int cpu) in tick_get_oneshot_wakeup_device()
154 int cpu) in tick_set_oneshot_wakeup_device()
163 void tick_install_broadcast_device(struct clock_event_device *dev, int cpu) in tick_install_broadcast_device()
247 int tick_device_uses_broadcast(struct clock_event_device *dev, int cpu) in tick_device_uses_broadcast()
348 int cpu = smp_processor_id(); in tick_do_broadcast() local
[all …]
Dtick-sched.c43 struct tick_sched *tick_get_tick_sched(int cpu) in tick_get_tick_sched()
209 int tick_cpu, cpu = smp_processor_id(); in tick_sched_do_timer() local
369 static bool can_stop_full_tick(int cpu, struct tick_sched *ts) in can_stop_full_tick()
417 void tick_nohz_full_kick_cpu(int cpu) in tick_nohz_full_kick_cpu()
427 int cpu; in tick_nohz_kick_task() local
477 int cpu; in tick_nohz_full_kick_all() local
516 void tick_nohz_dep_set_cpu(int cpu, enum tick_dep_bits bit) in tick_nohz_dep_set_cpu()
539 void tick_nohz_dep_clear_cpu(int cpu, enum tick_dep_bits bit) in tick_nohz_dep_clear_cpu()
618 bool tick_nohz_cpu_hotpluggable(unsigned int cpu) in tick_nohz_cpu_hotpluggable()
630 static int tick_nohz_cpu_down(unsigned int cpu) in tick_nohz_cpu_down()
[all …]
Dtick-common.c65 struct tick_device *tick_get_device(int cpu) in tick_get_device()
87 static void tick_periodic(int cpu) in tick_periodic()
112 int cpu = smp_processor_id(); in tick_handle_periodic() local
187 struct clock_event_device *newdev, int cpu, in tick_setup_device()
266 int cpu = smp_processor_id(); in tick_install_replacement() local
275 struct clock_event_device *newdev, int cpu) in tick_check_percpu()
331 int cpu; in tick_check_new_device() local
422 void tick_shutdown(unsigned int cpu) in tick_shutdown()
/kernel/cgroup/
Drstat.c17 static struct cgroup_rstat_cpu *cgroup_rstat_cpu(struct cgroup *cgrp, int cpu) in cgroup_rstat_cpu()
31 unsigned long _cgroup_rstat_cpu_lock(raw_spinlock_t *cpu_lock, int cpu, in _cgroup_rstat_cpu_lock()
64 void _cgroup_rstat_cpu_unlock(raw_spinlock_t *cpu_lock, int cpu, in _cgroup_rstat_cpu_unlock()
85 __bpf_kfunc void cgroup_rstat_updated(struct cgroup *cgrp, int cpu) in cgroup_rstat_updated()
145 struct cgroup *child, int cpu) in cgroup_rstat_push_children()
203 static struct cgroup *cgroup_rstat_updated_list(struct cgroup *root, int cpu) in cgroup_rstat_updated_list()
267 struct cgroup *parent, int cpu) in bpf_rstat_flush()
306 int cpu; in cgroup_rstat_flush_locked() local
385 int cpu; in cgroup_rstat_init() local
407 int cpu; in cgroup_rstat_exit() local
[all …]
/kernel/sched/
Dtopology.c39 static int sched_domain_debug_one(struct sched_domain *sd, int cpu, int level, in sched_domain_debug_one()
135 static void sched_domain_debug(struct sched_domain *sd, int cpu) in sched_domain_debug()
161 # define sched_domain_debug(sd, cpu) do { } while (0) argument
356 static struct perf_domain *find_pd(struct perf_domain *pd, int cpu) in find_pd()
367 static struct perf_domain *pd_init(int cpu) in pd_init()
438 int cpu = cpumask_first(cpu_map); in build_perf_domains() local
696 static void update_top_cache_domain(int cpu) in update_top_cache_domain()
741 cpu_attach_domain(struct sched_domain *sd, struct root_domain *rd, int cpu) in cpu_attach_domain()
973 build_group_from_child_sched_domain(struct sched_domain *sd, int cpu) in build_group_from_child_sched_domain()
1002 int cpu; in init_overlap_sched_group() local
[all …]
Dcpudeadline.c124 int cpu, max_cpu = -1; in cpudl_find() local
173 void cpudl_clear(struct cpudl *cp, int cpu) in cpudl_clear()
213 void cpudl_set(struct cpudl *cp, int cpu, u64 dl) in cpudl_set()
244 void cpudl_set_freecpu(struct cpudl *cp, int cpu) in cpudl_set_freecpu()
254 void cpudl_clear_freecpu(struct cpudl *cp, int cpu) in cpudl_clear_freecpu()
Dcpuacct.c94 static u64 cpuacct_cpuusage_read(struct cpuacct *ca, int cpu, in cpuacct_cpuusage_read()
135 static void cpuacct_cpuusage_write(struct cpuacct *ca, int cpu) in cpuacct_cpuusage_write()
195 int cpu; in cpuusage_write() local
243 int cpu; in cpuacct_all_seq_show() local
265 int cpu; in cpuacct_stats_show() local
336 unsigned int cpu = task_cpu(tsk); in cpuacct_charge() local
Dcore.c348 static int sched_task_is_throttled(struct task_struct *p, int cpu) in sched_task_is_throttled()
359 int cpu = task_cpu(p); in sched_core_next() local
412 static void sched_core_lock(int cpu, unsigned long *flags) in sched_core_lock()
422 static void sched_core_unlock(int cpu, unsigned long *flags) in sched_core_unlock()
435 int cpu, t; in __sched_core_flip() local
469 int cpu; in sched_core_assert_empty() local
1139 int cpu, need_lazy = 0; in resched_curr() local
1165 void resched_cpu(int cpu) in resched_cpu()
1188 int i, cpu = smp_processor_id(), default_cpu = -1; in get_nohz_timer_target() local
1233 static void wake_up_idle_cpu(int cpu) in wake_up_idle_cpu()
[all …]
/kernel/trace/
Dring_buffer.c291 #define for_each_buffer_cpu(buffer, cpu) \ argument
294 #define for_each_online_buffer_cpu(buffer, cpu) \ argument
439 int cpu; member
701 size_t ring_buffer_nr_dirty_pages(struct trace_buffer *buffer, int cpu) in ring_buffer_nr_dirty_pages()
725 static __always_inline bool full_hit(struct trace_buffer *buffer, int cpu, int full) in full_hit()
785 void ring_buffer_wake_waiters(struct trace_buffer *buffer, int cpu) in ring_buffer_wake_waiters()
817 static bool rb_watermark_hit(struct trace_buffer *buffer, int cpu, int full) in rb_watermark_hit()
850 int cpu, int full, ring_buffer_cond_fn cond, void *data) in rb_wait_cond()
915 int ring_buffer_wait(struct trace_buffer *buffer, int cpu, int full, in ring_buffer_wait()
974 __poll_t ring_buffer_poll_wait(struct trace_buffer *buffer, int cpu, in ring_buffer_poll_wait()
[all …]
Dtrace_functions_graph.c37 int cpu; member
140 int cpu; in trace_graph_entry() local
250 int cpu; in trace_graph_return() local
360 static void print_graph_cpu(struct trace_seq *s, int cpu) in print_graph_cpu()
412 verif_pid(struct trace_seq *s, pid_t pid, int cpu, struct fgraph_data *data) in verif_pid()
541 enum trace_type type, int cpu, pid_t pid, u32 flags) in print_graph_irq()
716 int cpu = iter->cpu; in print_graph_entry_leaf() local
769 struct trace_seq *s, int cpu, u32 flags) in print_graph_entry_nested()
779 int cpu = iter->cpu; in print_graph_entry_nested() local
818 int cpu = iter->cpu; in print_graph_prologue() local
[all …]
/kernel/rcu/
Dtree_stall.h370 int cpu; in rcu_dump_cpu_stacks() local
423 int cpu; in rcu_is_rcuc_kthread_starving() local
442 static void print_cpu_stat_info(int cpu) in print_cpu_stat_info()
482 static void print_cpu_stall_info(int cpu) in print_cpu_stall_info()
535 int cpu; in rcu_check_gp_kthread_starvation() local
571 int cpu; in rcu_check_gp_kthread_expired_fqs_timer() local
597 int cpu; in print_other_cpu_stall() local
675 int cpu; in print_cpu_stall() local
853 int cpu; in rcu_check_boost_fail() local
897 int cpu; in show_rcu_gp_kthreads() local
[all …]
Dtree.c241 static long rcu_get_n_cbs_cpu(int cpu) in rcu_get_n_cbs_cpu()
344 bool rcu_watching_zero_in_eqs(int cpu, int *vp) in rcu_watching_zero_in_eqs()
752 int cpu; in rcu_request_urgent_qs_task() local
806 #define arch_irq_stat_cpu(cpu) 0 argument
934 int cpu = rdp->cpu; in rcu_watching_snap_recheck() local
1077 int cpu = get_cpu(); in swake_up_one_online() local
2110 int cpu; in rcu_gp_cleanup() local
2694 int cpu; in force_qs_rnp() local
2887 static void rcu_cpu_kthread_park(unsigned int cpu) in rcu_cpu_kthread_park()
2892 static int rcu_cpu_kthread_should_run(unsigned int cpu) in rcu_cpu_kthread_should_run()
[all …]
/kernel/irq/
Dcpuhotplug.c23 unsigned int cpu = smp_processor_id(); in irq_needs_fixup() local
191 static bool hk_should_isolate(struct irq_data *data, unsigned int cpu) in hk_should_isolate()
205 static void irq_restore_affinity_of_irq(struct irq_desc *desc, unsigned int cpu) in irq_restore_affinity_of_irq()
232 int irq_affinity_online_cpu(unsigned int cpu) in irq_affinity_online_cpu()
Dmatrix.c51 unsigned int cpu, matrix_size = BITS_TO_LONGS(matrix_bits); in irq_alloc_matrix() local
138 unsigned int cpu, best_cpu, maxavl = 0; in matrix_find_best_cpu() local
159 unsigned int cpu, best_cpu, allocated = UINT_MAX; in matrix_find_best_cpu_managed() local
218 unsigned int cpu, failed_cpu; in irq_matrix_reserve_managed() local
259 unsigned int cpu; in irq_matrix_remove_managed() local
295 unsigned int bit, cpu, end; in irq_matrix_alloc_managed() local
386 unsigned int cpu, bit; in irq_matrix_alloc() local
424 void irq_matrix_free(struct irq_matrix *m, unsigned int cpu, in irq_matrix_free()
499 int cpu; in irq_matrix_debug_show() local
Dipi.c163 irq_hw_number_t ipi_get_hwirq(unsigned int irq, unsigned int cpu) in ipi_get_hwirq()
189 const struct cpumask *dest, unsigned int cpu) in ipi_send_verify()
227 int __ipi_send_single(struct irq_desc *desc, unsigned int cpu) in __ipi_send_single()
273 unsigned int cpu; in __ipi_send_mask() local
313 int ipi_send_single(unsigned int virq, unsigned int cpu) in ipi_send_single()
/kernel/bpf/
Dpercpu_freelist.c8 int cpu; in pcpu_freelist_init() local
59 int cpu, orig_cpu; in ___pcpu_freelist_push_nmi() local
103 unsigned int cpu, cpu_idx, i, j, n, m; in pcpu_freelist_populate() local
125 int cpu; in ___pcpu_freelist_pop() local
157 int cpu; in ___pcpu_freelist_pop_nmi() local
Dcpumask.c32 static bool cpu_valid(u32 cpu) in cpu_valid()
150 __bpf_kfunc void bpf_cpumask_set_cpu(u32 cpu, struct bpf_cpumask *cpumask) in bpf_cpumask_set_cpu()
163 __bpf_kfunc void bpf_cpumask_clear_cpu(u32 cpu, struct bpf_cpumask *cpumask) in bpf_cpumask_clear_cpu()
180 __bpf_kfunc bool bpf_cpumask_test_cpu(u32 cpu, const struct cpumask *cpumask) in bpf_cpumask_test_cpu()
197 __bpf_kfunc bool bpf_cpumask_test_and_set_cpu(u32 cpu, struct bpf_cpumask *cpumask) in bpf_cpumask_test_and_set_cpu()
215 __bpf_kfunc bool bpf_cpumask_test_and_clear_cpu(u32 cpu, struct bpf_cpumask *cpumask) in bpf_cpumask_test_and_clear_cpu()
/kernel/livepatch/
Dtransition.c95 unsigned int cpu; in klp_complete_transition() local
453 unsigned int cpu; in klp_try_complete_transition() local
533 unsigned int cpu; in klp_start_transition() local
576 unsigned int cpu; in klp_init_transition() local
650 unsigned int cpu; in klp_reverse_transition() local
731 unsigned int cpu; in klp_force_transition() local

123456