/arch/arm/common/ |
D | mcpm_entry.c | 125 mcpm_sync.clusters[cluster].cpus[cpu].cpu = CPU_GOING_DOWN; in __mcpm_cpu_going_down() 126 sync_cache_w(&mcpm_sync.clusters[cluster].cpus[cpu].cpu); in __mcpm_cpu_going_down() 139 mcpm_sync.clusters[cluster].cpus[cpu].cpu = CPU_DOWN; in __mcpm_cpu_down() 140 sync_cache_w(&mcpm_sync.clusters[cluster].cpus[cpu].cpu); in __mcpm_cpu_down() 192 sync_cache_r(&c->cpus); in __mcpm_outbound_enter_critical() 200 cpustate = c->cpus[i].cpu; in __mcpm_outbound_enter_critical() 205 sync_cache_r(&c->cpus[i].cpu); in __mcpm_outbound_enter_critical() 248 mcpm_sync.clusters[i].cpus[j].cpu = CPU_DOWN; in mcpm_sync_init() 253 mcpm_sync.clusters[this_cluster].cpus[i].cpu = CPU_UP; in mcpm_sync_init()
|
/arch/x86/kernel/ |
D | tsc_sync.c | 126 int cpus = 2; in check_tsc_sync_source() local 150 while (atomic_read(&start_count) != cpus-1) in check_tsc_sync_source() 159 while (atomic_read(&stop_count) != cpus-1) in check_tsc_sync_source() 192 int cpus = 2; in check_tsc_sync_target() local 202 while (atomic_read(&start_count) != cpus) in check_tsc_sync_target() 215 while (atomic_read(&stop_count) != cpus) in check_tsc_sync_target()
|
/arch/mips/cavium-octeon/ |
D | smp.c | 99 int cpus; in octeon_smp_setup() local 116 cpus = 1; in octeon_smp_setup() 119 set_cpu_possible(cpus, true); in octeon_smp_setup() 120 set_cpu_present(cpus, true); in octeon_smp_setup() 121 __cpu_number_map[id] = cpus; in octeon_smp_setup() 122 __cpu_logical_map[cpus] = id; in octeon_smp_setup() 123 cpus++; in octeon_smp_setup() 135 set_cpu_possible(cpus, true); in octeon_smp_setup() 136 __cpu_number_map[id] = cpus; in octeon_smp_setup() 137 __cpu_logical_map[cpus] = id; in octeon_smp_setup() [all …]
|
/arch/ia64/kernel/ |
D | smp.c | 295 cpumask_var_t cpus; in smp_flush_tlb_mm() local 304 if (!alloc_cpumask_var(&cpus, GFP_ATOMIC)) { in smp_flush_tlb_mm() 308 cpumask_copy(cpus, mm_cpumask(mm)); in smp_flush_tlb_mm() 309 smp_call_function_many(cpus, in smp_flush_tlb_mm() 311 free_cpumask_var(cpus); in smp_flush_tlb_mm()
|
/arch/arm/kernel/ |
D | devtree.c | 81 struct device_node *cpu, *cpus; in arm_dt_init_cpu_maps() local 87 cpus = of_find_node_by_path("/cpus"); in arm_dt_init_cpu_maps() 89 if (!cpus) in arm_dt_init_cpu_maps() 92 for_each_child_of_node(cpus, cpu) { in arm_dt_init_cpu_maps()
|
/arch/powerpc/platforms/powermac/ |
D | cpufreq_64.c | 364 cpumask_copy(policy->cpus, cpu_online_mask); in g5_cpufreq_cpu_init() 386 static int __init g5_neo2_cpufreq_init(struct device_node *cpus) in g5_neo2_cpufreq_init() argument 410 (cpunode = of_get_next_child(cpus, cpunode)) != NULL;) { in g5_neo2_cpufreq_init() 540 static int __init g5_pm72_cpufreq_init(struct device_node *cpus) in g5_pm72_cpufreq_init() argument 553 (cpunode = of_get_next_child(cpus, cpunode)) != NULL;) { in g5_pm72_cpufreq_init() 721 struct device_node *cpus; in g5_cpufreq_init() local 724 cpus = of_find_node_by_path("/cpus"); in g5_cpufreq_init() 725 if (cpus == NULL) { in g5_cpufreq_init() 733 rc = g5_pm72_cpufreq_init(cpus); in g5_cpufreq_init() 736 rc = g5_neo2_cpufreq_init(cpus); in g5_cpufreq_init() [all …]
|
D | smp.c | 637 struct device_node *cpus; in smp_core99_pfunc_tb_freeze() local 640 cpus = of_find_node_by_path("/cpus"); in smp_core99_pfunc_tb_freeze() 641 BUG_ON(cpus == NULL); in smp_core99_pfunc_tb_freeze() 644 pmf_call_function(cpus, "cpu-timebase", &args); in smp_core99_pfunc_tb_freeze() 645 of_node_put(cpus); in smp_core99_pfunc_tb_freeze() 715 struct device_node *cpus = in smp_core99_setup() local 717 if (cpus && in smp_core99_setup() 718 of_get_property(cpus, "platform-cpu-timebase", NULL)) { in smp_core99_setup() 771 struct device_node *cpus; in smp_core99_probe() local 777 for (cpus = NULL; (cpus = of_find_node_by_type(cpus, "cpu")) != NULL;) in smp_core99_probe()
|
/arch/powerpc/kernel/ |
D | rtas.c | 818 cpumask_var_t cpus) in rtas_cpu_state_change_mask() argument 820 if (!cpumask_empty(cpus)) { in rtas_cpu_state_change_mask() 821 cpumask_clear(cpus); in rtas_cpu_state_change_mask() 831 cpumask_var_t cpus) in rtas_cpu_state_change_mask() argument 837 if (cpumask_empty(cpus)) in rtas_cpu_state_change_mask() 840 for_each_cpu(cpu, cpus) { in rtas_cpu_state_change_mask() 858 cpumask_shift_right(cpus, cpus, cpu); in rtas_cpu_state_change_mask() 859 cpumask_shift_left(cpus, cpus, cpu); in rtas_cpu_state_change_mask() 863 cpumask_clear_cpu(cpu, cpus); in rtas_cpu_state_change_mask() 872 int rtas_online_cpus_mask(cpumask_var_t cpus) in rtas_online_cpus_mask() argument [all …]
|
/arch/um/kernel/ |
D | smp.c | 209 int cpus = num_online_cpus() - 1; in smp_call_function() local 212 if (!cpus) in smp_call_function() 227 while (atomic_read(&scf_started) != cpus) in smp_call_function() 231 while (atomic_read(&scf_finished) != cpus) in smp_call_function()
|
/arch/mips/mti-sead3/ |
D | sead3.dts | 12 cpus {
|
/arch/mips/include/asm/ |
D | smtc.h | 51 extern void smtc_prepare_cpus(int cpus);
|
/arch/powerpc/platforms/cell/ |
D | cpufreq_spudemand.c | 112 for_each_cpu(i, policy->cpus) { in spu_gov_govern() 129 for_each_cpu (i, policy->cpus) { in spu_gov_govern()
|
/arch/arm/boot/dts/ |
D | sh7372.dtsi | 16 cpus {
|
D | r8a7740.dtsi | 16 cpus {
|
D | omap443x.dtsi | 14 cpus {
|
D | omap34xx.dtsi | 14 cpus {
|
D | omap4460.dtsi | 13 cpus {
|
D | omap36xx.dtsi | 18 cpus {
|
D | r8a7778.dtsi | 22 cpus {
|
/arch/s390/hypfs/ |
D | hypfs_diag.c | 120 __u8 cpus; member 127 __u8 cpus; member 154 return ((struct part_hdr *)hdr)->cpus; in part_hdr__rcpus() 256 __u8 cpus; member 263 __u8 cpus; member 280 return ((struct phys_hdr *)hdr)->cpus; in phys_hdr__cpus() 282 return ((struct x_phys_hdr *)hdr)->cpus; in phys_hdr__cpus()
|
/arch/powerpc/boot/dts/ |
D | mpc8536ds.dts | 18 cpus { 19 #cpus = <1>;
|
D | mpc8536ds_36b.dts | 18 cpus { 19 #cpus = <1>;
|
/arch/arc/boot/dts/ |
D | skeleton.dtsi | 22 cpus {
|
/arch/mips/ralink/dts/ |
D | rt2880.dtsi | 6 cpus {
|
/arch/arm/include/asm/ |
D | mcpm.h | 146 } cpus[MAX_CPUS_PER_CLUSTER]; member
|