| /kernel/ |
| D | smpboot.c | 30 struct task_struct *idle_thread_get(unsigned int cpu) in idle_thread_get() 50 static __always_inline void idle_init(unsigned int cpu) in idle_init() 68 unsigned int cpu, boot_cpu; in idle_threads_init() local 85 unsigned int cpu; member 170 __smpboot_create_thread(struct smp_hotplug_thread *ht, unsigned int cpu) in __smpboot_create_thread() 213 int smpboot_create_threads(unsigned int cpu) in smpboot_create_threads() 228 static void smpboot_unpark_thread(struct smp_hotplug_thread *ht, unsigned int cpu) in smpboot_unpark_thread() 236 int smpboot_unpark_threads(unsigned int cpu) in smpboot_unpark_threads() 247 static void smpboot_park_thread(struct smp_hotplug_thread *ht, unsigned int cpu) in smpboot_park_thread() 255 int smpboot_park_threads(unsigned int cpu) in smpboot_park_threads() [all …]
|
| D | cpu.c | 169 static int cpuhp_invoke_callback(unsigned int cpu, enum cpuhp_state state, in cpuhp_invoke_callback() 308 static bool cpuhp_wait_for_sync_state(unsigned int cpu, enum cpuhp_sync_state state, in cpuhp_wait_for_sync_state() 354 void __weak arch_cpuhp_cleanup_dead_cpu(unsigned int cpu) { } in arch_cpuhp_cleanup_dead_cpu() 360 static void cpuhp_bp_sync_dead(unsigned int cpu) in cpuhp_bp_sync_dead() 381 static inline void cpuhp_bp_sync_dead(unsigned int cpu) { } in cpuhp_bp_sync_dead() 402 static bool cpuhp_can_boot_ap(unsigned int cpu) in cpuhp_can_boot_ap() 430 void __weak arch_cpuhp_cleanup_kick_cpu(unsigned int cpu) { } in arch_cpuhp_cleanup_kick_cpu() 436 static int cpuhp_bp_sync_alive(unsigned int cpu) in cpuhp_bp_sync_alive() 453 static inline int cpuhp_bp_sync_alive(unsigned int cpu) { return 0; } in cpuhp_bp_sync_alive() 454 static inline bool cpuhp_can_boot_ap(unsigned int cpu) { return true; } in cpuhp_can_boot_ap() [all …]
|
| D | stop_machine.c | 91 static bool cpu_stop_queue_work(unsigned int cpu, struct cpu_stop_work *work) in cpu_stop_queue_work() 137 int stop_one_cpu(unsigned int cpu, cpu_stop_fn_t fn, void *arg) in stop_one_cpu() 206 int cpu = smp_processor_id(), err = 0; in multi_cpu_stop() local 385 bool stop_one_cpu_nowait(unsigned int cpu, cpu_stop_fn_t fn, void *arg, in stop_one_cpu_nowait() 398 unsigned int cpu; in queue_stop_cpus_work() local 476 static int cpu_stop_should_run(unsigned int cpu) in cpu_stop_should_run() 488 static void cpu_stopper_thread(unsigned int cpu) in cpu_stopper_thread() 528 void stop_machine_park(int cpu) in stop_machine_park() 540 static void cpu_stop_create(unsigned int cpu) in cpu_stop_create() 545 static void cpu_stop_park(unsigned int cpu) in cpu_stop_park() [all …]
|
| D | smp.c | 54 int smpcfd_prepare_cpu(unsigned int cpu) in smpcfd_prepare_cpu() 76 int smpcfd_dead_cpu(unsigned int cpu) in smpcfd_dead_cpu() 86 int smpcfd_dying_cpu(unsigned int cpu) in smpcfd_dying_cpu() 113 send_call_function_single_ipi(int cpu) in send_call_function_single_ipi() 231 int cpu = -1; in csd_lock_wait_toolong() local 378 void __smp_call_single_queue(int cpu, struct llist_node *node) in __smp_call_single_queue() 419 static int generic_exec_single(int cpu, call_single_data_t *csd) in generic_exec_single() 630 int smp_call_function_single(int cpu, smp_call_func_t func, void *info, in smp_call_function_single() 710 int smp_call_function_single_async(int cpu, call_single_data_t *csd) in smp_call_function_single_async() 750 unsigned int cpu; in smp_call_function_any() local [all …]
|
| D | watchdog_perf.c | 119 unsigned int cpu; in hardlockup_detector_event_create() local 155 void watchdog_hardlockup_enable(unsigned int cpu) in watchdog_hardlockup_enable() 174 void watchdog_hardlockup_disable(unsigned int cpu) in watchdog_hardlockup_disable() 195 int cpu; in hardlockup_detector_perf_stop() local 214 int cpu; in hardlockup_detector_perf_restart() local
|
| D | watchdog_buddy.c | 11 static unsigned int watchdog_next_cpu(unsigned int cpu) in watchdog_next_cpu() 30 void watchdog_hardlockup_enable(unsigned int cpu) in watchdog_hardlockup_enable() 64 void watchdog_hardlockup_disable(unsigned int cpu) in watchdog_hardlockup_disable()
|
| D | watchdog.c | 128 void watchdog_hardlockup_touch_cpu(unsigned int cpu) in watchdog_hardlockup_touch_cpu() 133 static bool is_hardlockup(unsigned int cpu) in is_hardlockup() 158 void watchdog_hardlockup_check(unsigned int cpu, struct pt_regs *regs) in watchdog_hardlockup_check() 240 void __weak watchdog_hardlockup_enable(unsigned int cpu) { } in watchdog_hardlockup_enable() 242 void __weak watchdog_hardlockup_disable(unsigned int cpu) { } in watchdog_hardlockup_disable() 608 int cpu; in touch_all_softlockup_watchdogs() local 780 static void watchdog_enable(unsigned int cpu) in watchdog_enable() 806 static void watchdog_disable(unsigned int cpu) in watchdog_disable() 830 int cpu; in softlockup_stop_all() local 849 int cpu; in softlockup_start_all() local [all …]
|
| /kernel/time/ |
| D | tick-broadcast.c | 47 static inline void tick_broadcast_clear_oneshot(int cpu) { } in tick_broadcast_clear_oneshot() 50 static inline void tick_broadcast_oneshot_offline(unsigned int cpu) { } in tick_broadcast_oneshot_offline() 69 const struct clock_event_device *tick_get_wakeup_device(int cpu) in tick_get_wakeup_device() 102 static struct clock_event_device *tick_get_oneshot_wakeup_device(int cpu) in tick_get_oneshot_wakeup_device() 117 int cpu) in tick_set_oneshot_wakeup_device() 148 static struct clock_event_device *tick_get_oneshot_wakeup_device(int cpu) in tick_get_oneshot_wakeup_device() 154 int cpu) in tick_set_oneshot_wakeup_device() 163 void tick_install_broadcast_device(struct clock_event_device *dev, int cpu) in tick_install_broadcast_device() 247 int tick_device_uses_broadcast(struct clock_event_device *dev, int cpu) in tick_device_uses_broadcast() 348 int cpu = smp_processor_id(); in tick_do_broadcast() local [all …]
|
| D | tick-sched.c | 43 struct tick_sched *tick_get_tick_sched(int cpu) in tick_get_tick_sched() 209 int tick_cpu, cpu = smp_processor_id(); in tick_sched_do_timer() local 369 static bool can_stop_full_tick(int cpu, struct tick_sched *ts) in can_stop_full_tick() 417 void tick_nohz_full_kick_cpu(int cpu) in tick_nohz_full_kick_cpu() 427 int cpu; in tick_nohz_kick_task() local 477 int cpu; in tick_nohz_full_kick_all() local 516 void tick_nohz_dep_set_cpu(int cpu, enum tick_dep_bits bit) in tick_nohz_dep_set_cpu() 539 void tick_nohz_dep_clear_cpu(int cpu, enum tick_dep_bits bit) in tick_nohz_dep_clear_cpu() 618 bool tick_nohz_cpu_hotpluggable(unsigned int cpu) in tick_nohz_cpu_hotpluggable() 630 static int tick_nohz_cpu_down(unsigned int cpu) in tick_nohz_cpu_down() [all …]
|
| D | tick-common.c | 65 struct tick_device *tick_get_device(int cpu) in tick_get_device() 87 static void tick_periodic(int cpu) in tick_periodic() 112 int cpu = smp_processor_id(); in tick_handle_periodic() local 187 struct clock_event_device *newdev, int cpu, in tick_setup_device() 266 int cpu = smp_processor_id(); in tick_install_replacement() local 275 struct clock_event_device *newdev, int cpu) in tick_check_percpu() 331 int cpu; in tick_check_new_device() local 422 void tick_shutdown(unsigned int cpu) in tick_shutdown()
|
| /kernel/cgroup/ |
| D | rstat.c | 17 static struct cgroup_rstat_cpu *cgroup_rstat_cpu(struct cgroup *cgrp, int cpu) in cgroup_rstat_cpu() 31 unsigned long _cgroup_rstat_cpu_lock(raw_spinlock_t *cpu_lock, int cpu, in _cgroup_rstat_cpu_lock() 64 void _cgroup_rstat_cpu_unlock(raw_spinlock_t *cpu_lock, int cpu, in _cgroup_rstat_cpu_unlock() 85 __bpf_kfunc void cgroup_rstat_updated(struct cgroup *cgrp, int cpu) in cgroup_rstat_updated() 145 struct cgroup *child, int cpu) in cgroup_rstat_push_children() 203 static struct cgroup *cgroup_rstat_updated_list(struct cgroup *root, int cpu) in cgroup_rstat_updated_list() 267 struct cgroup *parent, int cpu) in bpf_rstat_flush() 306 int cpu; in cgroup_rstat_flush_locked() local 385 int cpu; in cgroup_rstat_init() local 407 int cpu; in cgroup_rstat_exit() local [all …]
|
| /kernel/sched/ |
| D | topology.c | 39 static int sched_domain_debug_one(struct sched_domain *sd, int cpu, int level, in sched_domain_debug_one() 135 static void sched_domain_debug(struct sched_domain *sd, int cpu) in sched_domain_debug() 161 # define sched_domain_debug(sd, cpu) do { } while (0) argument 356 static struct perf_domain *find_pd(struct perf_domain *pd, int cpu) in find_pd() 367 static struct perf_domain *pd_init(int cpu) in pd_init() 438 int cpu = cpumask_first(cpu_map); in build_perf_domains() local 696 static void update_top_cache_domain(int cpu) in update_top_cache_domain() 741 cpu_attach_domain(struct sched_domain *sd, struct root_domain *rd, int cpu) in cpu_attach_domain() 973 build_group_from_child_sched_domain(struct sched_domain *sd, int cpu) in build_group_from_child_sched_domain() 1002 int cpu; in init_overlap_sched_group() local [all …]
|
| D | cpudeadline.c | 124 int cpu, max_cpu = -1; in cpudl_find() local 173 void cpudl_clear(struct cpudl *cp, int cpu) in cpudl_clear() 213 void cpudl_set(struct cpudl *cp, int cpu, u64 dl) in cpudl_set() 244 void cpudl_set_freecpu(struct cpudl *cp, int cpu) in cpudl_set_freecpu() 254 void cpudl_clear_freecpu(struct cpudl *cp, int cpu) in cpudl_clear_freecpu()
|
| D | cpuacct.c | 94 static u64 cpuacct_cpuusage_read(struct cpuacct *ca, int cpu, in cpuacct_cpuusage_read() 135 static void cpuacct_cpuusage_write(struct cpuacct *ca, int cpu) in cpuacct_cpuusage_write() 195 int cpu; in cpuusage_write() local 243 int cpu; in cpuacct_all_seq_show() local 265 int cpu; in cpuacct_stats_show() local 336 unsigned int cpu = task_cpu(tsk); in cpuacct_charge() local
|
| D | core.c | 348 static int sched_task_is_throttled(struct task_struct *p, int cpu) in sched_task_is_throttled() 359 int cpu = task_cpu(p); in sched_core_next() local 412 static void sched_core_lock(int cpu, unsigned long *flags) in sched_core_lock() 422 static void sched_core_unlock(int cpu, unsigned long *flags) in sched_core_unlock() 435 int cpu, t; in __sched_core_flip() local 469 int cpu; in sched_core_assert_empty() local 1139 int cpu, need_lazy = 0; in resched_curr() local 1165 void resched_cpu(int cpu) in resched_cpu() 1188 int i, cpu = smp_processor_id(), default_cpu = -1; in get_nohz_timer_target() local 1233 static void wake_up_idle_cpu(int cpu) in wake_up_idle_cpu() [all …]
|
| /kernel/trace/ |
| D | ring_buffer.c | 291 #define for_each_buffer_cpu(buffer, cpu) \ argument 294 #define for_each_online_buffer_cpu(buffer, cpu) \ argument 439 int cpu; member 701 size_t ring_buffer_nr_dirty_pages(struct trace_buffer *buffer, int cpu) in ring_buffer_nr_dirty_pages() 725 static __always_inline bool full_hit(struct trace_buffer *buffer, int cpu, int full) in full_hit() 785 void ring_buffer_wake_waiters(struct trace_buffer *buffer, int cpu) in ring_buffer_wake_waiters() 817 static bool rb_watermark_hit(struct trace_buffer *buffer, int cpu, int full) in rb_watermark_hit() 850 int cpu, int full, ring_buffer_cond_fn cond, void *data) in rb_wait_cond() 915 int ring_buffer_wait(struct trace_buffer *buffer, int cpu, int full, in ring_buffer_wait() 974 __poll_t ring_buffer_poll_wait(struct trace_buffer *buffer, int cpu, in ring_buffer_poll_wait() [all …]
|
| D | trace_functions_graph.c | 37 int cpu; member 140 int cpu; in trace_graph_entry() local 250 int cpu; in trace_graph_return() local 360 static void print_graph_cpu(struct trace_seq *s, int cpu) in print_graph_cpu() 412 verif_pid(struct trace_seq *s, pid_t pid, int cpu, struct fgraph_data *data) in verif_pid() 541 enum trace_type type, int cpu, pid_t pid, u32 flags) in print_graph_irq() 716 int cpu = iter->cpu; in print_graph_entry_leaf() local 769 struct trace_seq *s, int cpu, u32 flags) in print_graph_entry_nested() 779 int cpu = iter->cpu; in print_graph_entry_nested() local 818 int cpu = iter->cpu; in print_graph_prologue() local [all …]
|
| /kernel/rcu/ |
| D | tree_stall.h | 370 int cpu; in rcu_dump_cpu_stacks() local 423 int cpu; in rcu_is_rcuc_kthread_starving() local 442 static void print_cpu_stat_info(int cpu) in print_cpu_stat_info() 482 static void print_cpu_stall_info(int cpu) in print_cpu_stall_info() 535 int cpu; in rcu_check_gp_kthread_starvation() local 571 int cpu; in rcu_check_gp_kthread_expired_fqs_timer() local 597 int cpu; in print_other_cpu_stall() local 675 int cpu; in print_cpu_stall() local 853 int cpu; in rcu_check_boost_fail() local 897 int cpu; in show_rcu_gp_kthreads() local [all …]
|
| D | tree.c | 241 static long rcu_get_n_cbs_cpu(int cpu) in rcu_get_n_cbs_cpu() 344 bool rcu_watching_zero_in_eqs(int cpu, int *vp) in rcu_watching_zero_in_eqs() 752 int cpu; in rcu_request_urgent_qs_task() local 806 #define arch_irq_stat_cpu(cpu) 0 argument 934 int cpu = rdp->cpu; in rcu_watching_snap_recheck() local 1077 int cpu = get_cpu(); in swake_up_one_online() local 2110 int cpu; in rcu_gp_cleanup() local 2694 int cpu; in force_qs_rnp() local 2887 static void rcu_cpu_kthread_park(unsigned int cpu) in rcu_cpu_kthread_park() 2892 static int rcu_cpu_kthread_should_run(unsigned int cpu) in rcu_cpu_kthread_should_run() [all …]
|
| /kernel/irq/ |
| D | cpuhotplug.c | 23 unsigned int cpu = smp_processor_id(); in irq_needs_fixup() local 191 static bool hk_should_isolate(struct irq_data *data, unsigned int cpu) in hk_should_isolate() 205 static void irq_restore_affinity_of_irq(struct irq_desc *desc, unsigned int cpu) in irq_restore_affinity_of_irq() 232 int irq_affinity_online_cpu(unsigned int cpu) in irq_affinity_online_cpu()
|
| D | matrix.c | 51 unsigned int cpu, matrix_size = BITS_TO_LONGS(matrix_bits); in irq_alloc_matrix() local 138 unsigned int cpu, best_cpu, maxavl = 0; in matrix_find_best_cpu() local 159 unsigned int cpu, best_cpu, allocated = UINT_MAX; in matrix_find_best_cpu_managed() local 218 unsigned int cpu, failed_cpu; in irq_matrix_reserve_managed() local 259 unsigned int cpu; in irq_matrix_remove_managed() local 295 unsigned int bit, cpu, end; in irq_matrix_alloc_managed() local 386 unsigned int cpu, bit; in irq_matrix_alloc() local 424 void irq_matrix_free(struct irq_matrix *m, unsigned int cpu, in irq_matrix_free() 499 int cpu; in irq_matrix_debug_show() local
|
| D | ipi.c | 163 irq_hw_number_t ipi_get_hwirq(unsigned int irq, unsigned int cpu) in ipi_get_hwirq() 189 const struct cpumask *dest, unsigned int cpu) in ipi_send_verify() 227 int __ipi_send_single(struct irq_desc *desc, unsigned int cpu) in __ipi_send_single() 273 unsigned int cpu; in __ipi_send_mask() local 313 int ipi_send_single(unsigned int virq, unsigned int cpu) in ipi_send_single()
|
| /kernel/bpf/ |
| D | percpu_freelist.c | 8 int cpu; in pcpu_freelist_init() local 59 int cpu, orig_cpu; in ___pcpu_freelist_push_nmi() local 103 unsigned int cpu, cpu_idx, i, j, n, m; in pcpu_freelist_populate() local 125 int cpu; in ___pcpu_freelist_pop() local 157 int cpu; in ___pcpu_freelist_pop_nmi() local
|
| D | cpumask.c | 32 static bool cpu_valid(u32 cpu) in cpu_valid() 150 __bpf_kfunc void bpf_cpumask_set_cpu(u32 cpu, struct bpf_cpumask *cpumask) in bpf_cpumask_set_cpu() 163 __bpf_kfunc void bpf_cpumask_clear_cpu(u32 cpu, struct bpf_cpumask *cpumask) in bpf_cpumask_clear_cpu() 180 __bpf_kfunc bool bpf_cpumask_test_cpu(u32 cpu, const struct cpumask *cpumask) in bpf_cpumask_test_cpu() 197 __bpf_kfunc bool bpf_cpumask_test_and_set_cpu(u32 cpu, struct bpf_cpumask *cpumask) in bpf_cpumask_test_and_set_cpu() 215 __bpf_kfunc bool bpf_cpumask_test_and_clear_cpu(u32 cpu, struct bpf_cpumask *cpumask) in bpf_cpumask_test_and_clear_cpu()
|
| /kernel/livepatch/ |
| D | transition.c | 95 unsigned int cpu; in klp_complete_transition() local 453 unsigned int cpu; in klp_try_complete_transition() local 533 unsigned int cpu; in klp_start_transition() local 576 unsigned int cpu; in klp_init_transition() local 650 unsigned int cpu; in klp_reverse_transition() local 731 unsigned int cpu; in klp_force_transition() local
|