/arch/x86/kernel/apic/ |
D | x2apic_cluster.c | 31 unsigned int cpu, this_cpu; in __x2apic_send_IPI_mask() local 39 this_cpu = smp_processor_id(); in __x2apic_send_IPI_mask() 59 if (apic_dest == APIC_DEST_ALLINC || i != this_cpu) in __x2apic_send_IPI_mask() 133 unsigned int this_cpu = smp_processor_id(); in init_x2apic_ldr() local 136 per_cpu(x86_cpu_to_logical_apicid, this_cpu) = apic_read(APIC_LDR); in init_x2apic_ldr() 138 __cpu_set(this_cpu, per_cpu(cpus_in_cluster, this_cpu)); in init_x2apic_ldr() 140 if (x2apic_cluster(this_cpu) != x2apic_cluster(cpu)) in init_x2apic_ldr() 142 __cpu_set(this_cpu, per_cpu(cpus_in_cluster, cpu)); in init_x2apic_ldr() 143 __cpu_set(cpu, per_cpu(cpus_in_cluster, this_cpu)); in init_x2apic_ldr() 153 unsigned int this_cpu = (unsigned long)hcpu; in update_clusterinfo() local [all …]
|
D | ipi.c | 42 unsigned int this_cpu = smp_processor_id(); in default_send_IPI_mask_allbutself_phys() local 50 if (query_cpu == this_cpu) in default_send_IPI_mask_allbutself_phys() 85 unsigned int this_cpu = smp_processor_id(); in default_send_IPI_mask_allbutself_logical() local 91 if (query_cpu == this_cpu) in default_send_IPI_mask_allbutself_logical()
|
D | x2apic_phys.c | 41 unsigned long this_cpu; in __x2apic_send_IPI_mask() local 48 this_cpu = smp_processor_id(); in __x2apic_send_IPI_mask() 50 if (apic_dest == APIC_DEST_ALLBUT && this_cpu == query_cpu) in __x2apic_send_IPI_mask()
|
D | apic_numachip.c | 121 unsigned int this_cpu = smp_processor_id(); in numachip_send_IPI_mask_allbutself() local 125 if (cpu != this_cpu) in numachip_send_IPI_mask_allbutself() 132 unsigned int this_cpu = smp_processor_id(); in numachip_send_IPI_allbutself() local 136 if (cpu != this_cpu) in numachip_send_IPI_allbutself()
|
D | x2apic_uv_x.c | 248 unsigned int this_cpu = smp_processor_id(); in uv_send_IPI_mask_allbutself() local 252 if (cpu != this_cpu) in uv_send_IPI_mask_allbutself() 259 unsigned int this_cpu = smp_processor_id(); in uv_send_IPI_allbutself() local 263 if (cpu != this_cpu) in uv_send_IPI_allbutself()
|
/arch/parisc/kernel/ |
D | smp.c | 123 int this_cpu = smp_processor_id(); in ipi_interrupt() local 124 struct cpuinfo_parisc *p = &per_cpu(cpu_data, this_cpu); in ipi_interrupt() 129 spinlock_t *lock = &per_cpu(ipi_lock, this_cpu); in ipi_interrupt() 147 smp_debug(100, KERN_DEBUG "CPU%d IPI_NOP\n", this_cpu); in ipi_interrupt() 151 smp_debug(100, KERN_DEBUG "CPU%d IPI_RESCHEDULE\n", this_cpu); in ipi_interrupt() 157 smp_debug(100, KERN_DEBUG "CPU%d IPI_CALL_FUNC\n", this_cpu); in ipi_interrupt() 162 smp_debug(100, KERN_DEBUG "CPU%d IPI_CPU_START\n", this_cpu); in ipi_interrupt() 166 smp_debug(100, KERN_DEBUG "CPU%d IPI_CPU_STOP\n", this_cpu); in ipi_interrupt() 171 smp_debug(100, KERN_DEBUG "CPU%d is alive!\n", this_cpu); in ipi_interrupt() 176 this_cpu, which); in ipi_interrupt()
|
/arch/x86/lib/ |
D | msr-smp.c | 10 int this_cpu = raw_smp_processor_id(); in __rdmsr_on_cpu() local 13 reg = per_cpu_ptr(rv->msrs, this_cpu); in __rdmsr_on_cpu() 24 int this_cpu = raw_smp_processor_id(); in __wrmsr_on_cpu() local 27 reg = per_cpu_ptr(rv->msrs, this_cpu); in __wrmsr_on_cpu() 102 int this_cpu; in __rwmsr_on_cpus() local 109 this_cpu = get_cpu(); in __rwmsr_on_cpus() 111 if (cpumask_test_cpu(this_cpu, mask)) in __rwmsr_on_cpus()
|
/arch/alpha/kernel/ |
D | smp.c | 527 int this_cpu = smp_processor_id(); in handle_ipi() local 528 unsigned long *pending_ipis = &ipi_data[this_cpu].bits; in handle_ipi() 533 this_cpu, *pending_ipis, regs->pc)); in handle_ipi() 564 this_cpu, which); in handle_ipi() 572 cpu_data[this_cpu].ipi_count++; in handle_ipi() 663 int cpu, this_cpu = smp_processor_id(); in flush_tlb_mm() local 665 if (!cpu_online(cpu) || cpu == this_cpu) in flush_tlb_mm() 712 int cpu, this_cpu = smp_processor_id(); in flush_tlb_page() local 714 if (!cpu_online(cpu) || cpu == this_cpu) in flush_tlb_page() 768 int cpu, this_cpu = smp_processor_id(); in flush_icache_user_range() local [all …]
|
/arch/blackfin/kernel/ |
D | nmi.c | 146 unsigned int this_cpu = smp_processor_id(); in check_nmi_wdt_touched() local 151 if (!atomic_read(&nmi_touched[this_cpu])) in check_nmi_wdt_touched() 154 atomic_set(&nmi_touched[this_cpu], 0); in check_nmi_wdt_touched() 156 cpumask_clear_cpu(this_cpu, &mask); in check_nmi_wdt_touched()
|
/arch/arm/common/ |
D | bL_switcher.c | 151 unsigned int mpidr, this_cpu, that_cpu; in bL_switch_to() local 159 this_cpu = smp_processor_id(); in bL_switch_to() 163 BUG_ON(cpu_logical_map(this_cpu) != ob_mpidr); in bL_switch_to() 168 that_cpu = bL_switcher_cpu_pairing[this_cpu]; in bL_switch_to() 174 this_cpu, ob_mpidr, ib_mpidr); in bL_switch_to() 176 this_cpu = smp_processor_id(); in bL_switch_to() 184 ipi_nr = register_ipi_completion(&inbound_alive, this_cpu); in bL_switch_to() 222 tdev = tick_get_device(this_cpu); in bL_switch_to() 223 if (tdev && !cpumask_equal(tdev->evtdev->cpumask, cpumask_of(this_cpu))) in bL_switch_to() 237 cpu_logical_map(this_cpu) = ib_mpidr; in bL_switch_to() [all …]
|
/arch/s390/kernel/ |
D | machine_kexec.c | 58 int cpu, this_cpu; in setup_regs() local 62 this_cpu = smp_find_processor_id(stap()); in setup_regs() 63 add_elf_notes(this_cpu); in setup_regs() 65 if (cpu == this_cpu) in setup_regs()
|
/arch/sparc/kernel/ |
D | process_64.c | 195 int this_cpu) in __global_reg_self() argument 201 rp = &global_cpu_snapshot[this_cpu].reg; in __global_reg_self() 247 int this_cpu, cpu; in arch_trigger_all_cpu_backtrace() local 254 this_cpu = raw_smp_processor_id(); in arch_trigger_all_cpu_backtrace() 259 __global_reg_self(tp, regs, this_cpu); in arch_trigger_all_cpu_backtrace() 266 if (!include_self && cpu == this_cpu) in arch_trigger_all_cpu_backtrace() 275 (cpu == this_cpu ? '*' : ' '), cpu, in arch_trigger_all_cpu_backtrace() 312 static void __global_pmu_self(int this_cpu) in __global_pmu_self() argument 320 pp = &global_cpu_snapshot[this_cpu].pmu; in __global_pmu_self() 346 int this_cpu, cpu; in pmu_snapshot_all_cpus() local [all …]
|
D | smp_64.c | 619 int retries, this_cpu, prev_sent, i, saw_cpu_error; in hypervisor_xcall_deliver() local 623 this_cpu = smp_processor_id(); in hypervisor_xcall_deliver() 706 this_cpu, saw_cpu_error - 1); in hypervisor_xcall_deliver() 712 this_cpu, retries); in hypervisor_xcall_deliver() 717 this_cpu, status); in hypervisor_xcall_deliver() 720 this_cpu, cnt, tb->cpu_list_pa, tb->cpu_mondo_block_pa); in hypervisor_xcall_deliver() 723 printk(KERN_CRIT "CPU[%d]: CPU list [ ", this_cpu); in hypervisor_xcall_deliver() 734 int this_cpu, i, cnt; in xcall_deliver() local 751 this_cpu = smp_processor_id(); in xcall_deliver() 752 tb = &trap_block[this_cpu]; in xcall_deliver() [all …]
|
D | nmi.c | 71 int this_cpu = smp_processor_id(); in die_nmi() local 78 panic("Watchdog detected hard LOCKUP on cpu %d", this_cpu); in die_nmi() 80 WARN(1, "Watchdog detected hard LOCKUP on cpu %d", this_cpu); in die_nmi()
|
/arch/x86/kernel/cpu/ |
D | common.c | 92 static const struct cpu_dev *this_cpu = &default_cpu; variable 372 if (!this_cpu) in table_lookup_model() 375 info = this_cpu->legacy_models; in table_lookup_model() 472 if (this_cpu->legacy_cache_size) in cpu_detect_cache_sizes() 473 l2size = this_cpu->legacy_cache_size(c, l2size); in cpu_detect_cache_sizes() 496 if (this_cpu->c_detect_tlb) in cpu_detect_tlb() 497 this_cpu->c_detect_tlb(c); in cpu_detect_tlb() 571 this_cpu = cpu_devs[i]; in get_cpu_vendor() 572 c->x86_vendor = this_cpu->c_x86_vendor; in get_cpu_vendor() 582 this_cpu = &default_cpu; in get_cpu_vendor() [all …]
|
/arch/arm/kernel/ |
D | smp_tlb.c | 126 int this_cpu; in broadcast_tlb_mm_a15_erratum() local 132 this_cpu = get_cpu(); in broadcast_tlb_mm_a15_erratum() 133 a15_erratum_get_cpumask(this_cpu, mm, &mask); in broadcast_tlb_mm_a15_erratum()
|
D | smp.c | 572 unsigned int this_cpu = smp_processor_id(); in smp_send_all_cpu_backtrace() local 583 cpu_clear(this_cpu, backtrace_mask); in smp_send_all_cpu_backtrace() 585 pr_info("Backtrace for cpu %d (current):\n", this_cpu); in smp_send_all_cpu_backtrace()
|
/arch/arm/include/asm/ |
D | mmu_context.h | 32 void a15_erratum_get_cpumask(int this_cpu, struct mm_struct *mm, 35 static inline void a15_erratum_get_cpumask(int this_cpu, struct mm_struct *mm, in a15_erratum_get_cpumask() argument
|
D | mcpm.h | 219 bool __mcpm_outbound_enter_critical(unsigned int this_cpu, unsigned int cluster);
|
/arch/x86/kernel/ |
D | irq.c | 292 unsigned int this_cpu, vector, this_count, count; in check_irq_vectors_for_cpu_disable() local 296 this_cpu = smp_processor_id(); in check_irq_vectors_for_cpu_disable() 298 cpu_clear(this_cpu, online_new); in check_irq_vectors_for_cpu_disable() 307 cpu_clear(this_cpu, affinity_new); in check_irq_vectors_for_cpu_disable() 336 if (cpu == this_cpu) in check_irq_vectors_for_cpu_disable() 354 this_cpu, this_count, count); in check_irq_vectors_for_cpu_disable()
|
/arch/x86/kernel/cpu/mcheck/ |
D | therm_throt.c | 154 unsigned int this_cpu = smp_processor_id(); in therm_throt_process() local 157 struct thermal_state *pstate = &per_cpu(thermal_state, this_cpu); in therm_throt_process() 194 this_cpu, in therm_throt_process() 202 this_cpu, in therm_throt_process() 213 unsigned int this_cpu = smp_processor_id(); in thresh_event_valid() local 214 struct thermal_state *pstate = &per_cpu(thermal_state, this_cpu); in thresh_event_valid()
|
/arch/blackfin/mach-common/ |
D | dpmc.c | 87 unsigned int this_cpu = smp_processor_id(); in bfin_wakeup_cpu() local 91 cpumask_clear_cpu(this_cpu, &mask); in bfin_wakeup_cpu()
|
/arch/ia64/kernel/ |
D | smp.c | 99 int this_cpu = get_cpu(); in handle_IPI() local 129 this_cpu, which); in handle_IPI()
|
D | perfmon.c | 5273 int this_cpu = smp_processor_id(); in pfm_overflow_handler() local 5308 pfm_stats[this_cpu].pfm_smpl_handler_calls++; in pfm_overflow_handler() 5331 pfm_stats[this_cpu].pfm_smpl_handler_cycles += end_cycles - start_cycles; in pfm_overflow_handler() 5473 int this_cpu = smp_processor_id(); in pfm_do_interrupt_handler() local 5476 pfm_stats[this_cpu].pfm_ovfl_intr_count++; in pfm_do_interrupt_handler() 5508 pfm_stats[this_cpu].pfm_spurious_ovfl_intr_count++; in pfm_do_interrupt_handler() 5520 this_cpu, task_pid_nr(task)); in pfm_do_interrupt_handler() 5525 this_cpu, in pfm_do_interrupt_handler() 5536 int this_cpu; in pfm_interrupt_handler() local 5540 this_cpu = get_cpu(); in pfm_interrupt_handler() [all …]
|
/arch/arm/mm/ |
D | context.c | 54 void a15_erratum_get_cpumask(int this_cpu, struct mm_struct *mm, in a15_erratum_get_cpumask() argument 64 if (cpu == this_cpu) in a15_erratum_get_cpumask()
|