/arch/x86/kernel/apic/ |
D | x2apic_cluster.c | 32 unsigned int cpu, this_cpu; in __x2apic_send_IPI_mask() local 40 this_cpu = smp_processor_id(); in __x2apic_send_IPI_mask() 60 if (apic_dest == APIC_DEST_ALLINC || i != this_cpu) in __x2apic_send_IPI_mask() 133 unsigned int this_cpu = smp_processor_id(); in init_x2apic_ldr() local 136 per_cpu(x86_cpu_to_logical_apicid, this_cpu) = apic_read(APIC_LDR); in init_x2apic_ldr() 138 __cpu_set(this_cpu, per_cpu(cpus_in_cluster, this_cpu)); in init_x2apic_ldr() 140 if (x2apic_cluster(this_cpu) != x2apic_cluster(cpu)) in init_x2apic_ldr() 142 __cpu_set(this_cpu, per_cpu(cpus_in_cluster, cpu)); in init_x2apic_ldr() 143 __cpu_set(cpu, per_cpu(cpus_in_cluster, this_cpu)); in init_x2apic_ldr() 153 unsigned int this_cpu = (unsigned long)hcpu; in update_clusterinfo() local [all …]
|
D | ipi.c | 43 unsigned int this_cpu = smp_processor_id(); in default_send_IPI_mask_allbutself_phys() local 51 if (query_cpu == this_cpu) in default_send_IPI_mask_allbutself_phys() 86 unsigned int this_cpu = smp_processor_id(); in default_send_IPI_mask_allbutself_logical() local 92 if (query_cpu == this_cpu) in default_send_IPI_mask_allbutself_logical()
|
D | x2apic_phys.c | 42 unsigned long this_cpu; in __x2apic_send_IPI_mask() local 49 this_cpu = smp_processor_id(); in __x2apic_send_IPI_mask() 51 if (apic_dest == APIC_DEST_ALLBUT && this_cpu == query_cpu) in __x2apic_send_IPI_mask()
|
D | apic_numachip.c | 130 unsigned int this_cpu = smp_processor_id(); in numachip_send_IPI_mask_allbutself() local 134 if (cpu != this_cpu) in numachip_send_IPI_mask_allbutself() 141 unsigned int this_cpu = smp_processor_id(); in numachip_send_IPI_allbutself() local 145 if (cpu != this_cpu) in numachip_send_IPI_allbutself()
|
D | x2apic_uv_x.c | 244 unsigned int this_cpu = smp_processor_id(); in uv_send_IPI_mask_allbutself() local 248 if (cpu != this_cpu) in uv_send_IPI_mask_allbutself() 255 unsigned int this_cpu = smp_processor_id(); in uv_send_IPI_allbutself() local 259 if (cpu != this_cpu) in uv_send_IPI_allbutself()
|
/arch/arm/mach-omap2/ |
D | omap-hotplug.c | 38 unsigned int this_cpu; in platform_cpu_die() local 54 this_cpu = smp_processor_id(); in platform_cpu_die() 55 if (omap_read_auxcoreboot0() == this_cpu) { in platform_cpu_die()
|
/arch/parisc/kernel/ |
D | smp.c | 124 int this_cpu = smp_processor_id(); in ipi_interrupt() local 125 struct cpuinfo_parisc *p = &per_cpu(cpu_data, this_cpu); in ipi_interrupt() 135 spinlock_t *lock = &per_cpu(ipi_lock, this_cpu); in ipi_interrupt() 153 smp_debug(100, KERN_DEBUG "CPU%d IPI_NOP\n", this_cpu); in ipi_interrupt() 157 smp_debug(100, KERN_DEBUG "CPU%d IPI_RESCHEDULE\n", this_cpu); in ipi_interrupt() 162 smp_debug(100, KERN_DEBUG "CPU%d IPI_CALL_FUNC\n", this_cpu); in ipi_interrupt() 167 smp_debug(100, KERN_DEBUG "CPU%d IPI_CALL_FUNC_SINGLE\n", this_cpu); in ipi_interrupt() 172 smp_debug(100, KERN_DEBUG "CPU%d IPI_CPU_START\n", this_cpu); in ipi_interrupt() 176 smp_debug(100, KERN_DEBUG "CPU%d IPI_CPU_STOP\n", this_cpu); in ipi_interrupt() 181 smp_debug(100, KERN_DEBUG "CPU%d is alive!\n", this_cpu); in ipi_interrupt() [all …]
|
/arch/x86/lib/ |
D | msr-smp.c | 10 int this_cpu = raw_smp_processor_id(); in __rdmsr_on_cpu() local 13 reg = per_cpu_ptr(rv->msrs, this_cpu); in __rdmsr_on_cpu() 24 int this_cpu = raw_smp_processor_id(); in __wrmsr_on_cpu() local 27 reg = per_cpu_ptr(rv->msrs, this_cpu); in __wrmsr_on_cpu() 71 int this_cpu; in __rwmsr_on_cpus() local 78 this_cpu = get_cpu(); in __rwmsr_on_cpus() 80 if (cpumask_test_cpu(this_cpu, mask)) in __rwmsr_on_cpus()
|
/arch/s390/kernel/ |
D | machine_kexec.c | 57 int cpu, this_cpu; in setup_regs() local 59 this_cpu = smp_find_processor_id(stap()); in setup_regs() 60 add_elf_notes(this_cpu); in setup_regs() 62 if (cpu == this_cpu) in setup_regs()
|
D | smp.c | 298 unsigned short this_cpu; in pcpu_delegate() local 301 this_cpu = stap(); in pcpu_delegate() 302 if (pcpu->address == this_cpu) in pcpu_delegate() 310 lc->restart_source = (unsigned long) this_cpu; in pcpu_delegate() 316 : : "d" (pcpu->address), "d" (this_cpu) : "0", "1", "cc"); in pcpu_delegate()
|
/arch/x86/kernel/cpu/mcheck/ |
D | therm_throt.c | 141 unsigned int this_cpu = smp_processor_id(); in therm_throt_process() local 144 struct thermal_state *pstate = &per_cpu(thermal_state, this_cpu); in therm_throt_process() 181 this_cpu, in therm_throt_process() 186 this_cpu, in therm_throt_process() 194 this_cpu, in therm_throt_process() 198 this_cpu, in therm_throt_process() 209 unsigned int this_cpu = smp_processor_id(); in thresh_event_valid() local 210 struct thermal_state *pstate = &per_cpu(thermal_state, this_cpu); in thresh_event_valid()
|
/arch/sparc/kernel/ |
D | process_64.c | 211 int this_cpu) in __global_reg_self() argument 215 global_reg_snapshot[this_cpu].tstate = regs->tstate; in __global_reg_self() 216 global_reg_snapshot[this_cpu].tpc = regs->tpc; in __global_reg_self() 217 global_reg_snapshot[this_cpu].tnpc = regs->tnpc; in __global_reg_self() 218 global_reg_snapshot[this_cpu].o7 = regs->u_regs[UREG_I7]; in __global_reg_self() 226 global_reg_snapshot[this_cpu].i7 = rw->ins[7]; in __global_reg_self() 230 global_reg_snapshot[this_cpu].rpc = rw->ins[7]; in __global_reg_self() 233 global_reg_snapshot[this_cpu].i7 = 0; in __global_reg_self() 234 global_reg_snapshot[this_cpu].rpc = 0; in __global_reg_self() 236 global_reg_snapshot[this_cpu].thread = tp; in __global_reg_self() [all …]
|
D | smp_64.c | 629 int retries, this_cpu, prev_sent, i, saw_cpu_error; in hypervisor_xcall_deliver() local 633 this_cpu = smp_processor_id(); in hypervisor_xcall_deliver() 716 this_cpu, saw_cpu_error - 1); in hypervisor_xcall_deliver() 722 this_cpu, retries); in hypervisor_xcall_deliver() 727 this_cpu, status); in hypervisor_xcall_deliver() 730 this_cpu, cnt, tb->cpu_list_pa, tb->cpu_mondo_block_pa); in hypervisor_xcall_deliver() 733 printk(KERN_CRIT "CPU[%d]: CPU list [ ", this_cpu); in hypervisor_xcall_deliver() 744 int this_cpu, i, cnt; in xcall_deliver() local 761 this_cpu = smp_processor_id(); in xcall_deliver() 762 tb = &trap_block[this_cpu]; in xcall_deliver() [all …]
|
D | chmc.c | 591 unsigned long ret, this_cpu; in chmc_read_mcreg() local 595 this_cpu = real_hard_smp_processor_id(); in chmc_read_mcreg() 597 if (p->portid == this_cpu) { in chmc_read_mcreg()
|
D | entry.h | 246 extern void __cpuinit sun4v_register_mondo_queues(int this_cpu);
|
/arch/blackfin/kernel/ |
D | nmi.c | 146 unsigned int this_cpu = smp_processor_id(); in check_nmi_wdt_touched() local 151 if (!atomic_read(&nmi_touched[this_cpu])) in check_nmi_wdt_touched() 154 atomic_set(&nmi_touched[this_cpu], 0); in check_nmi_wdt_touched() 156 cpumask_clear_cpu(this_cpu, &mask); in check_nmi_wdt_touched()
|
/arch/alpha/kernel/ |
D | smp.c | 567 int this_cpu = smp_processor_id(); in handle_ipi() local 568 unsigned long *pending_ipis = &ipi_data[this_cpu].bits; in handle_ipi() 573 this_cpu, *pending_ipis, regs->pc)); in handle_ipi() 604 this_cpu, which); in handle_ipi() 612 cpu_data[this_cpu].ipi_count++; in handle_ipi() 703 int cpu, this_cpu = smp_processor_id(); in flush_tlb_mm() local 705 if (!cpu_online(cpu) || cpu == this_cpu) in flush_tlb_mm() 752 int cpu, this_cpu = smp_processor_id(); in flush_tlb_page() local 754 if (!cpu_online(cpu) || cpu == this_cpu) in flush_tlb_page() 808 int cpu, this_cpu = smp_processor_id(); in flush_icache_user_range() local [all …]
|
/arch/x86/kernel/cpu/ |
D | common.c | 87 static const struct cpu_dev *this_cpu __cpuinitdata = &default_cpu; 342 if (!this_cpu) in table_lookup_model() 345 info = this_cpu->c_models; in table_lookup_model() 441 if (this_cpu->c_size_cache) in cpu_detect_cache_sizes() 442 l2size = this_cpu->c_size_cache(c, l2size); in cpu_detect_cache_sizes() 519 this_cpu = cpu_devs[i]; in get_cpu_vendor() 520 c->x86_vendor = this_cpu->c_x86_vendor; in get_cpu_vendor() 530 this_cpu = &default_cpu; in get_cpu_vendor() 678 if (this_cpu->c_early_init) in early_identify_cpu() 679 this_cpu->c_early_init(c); in early_identify_cpu() [all …]
|
/arch/blackfin/mach-common/ |
D | dpmc.c | 87 unsigned int this_cpu = smp_processor_id(); in bfin_wakeup_cpu() local 91 cpumask_clear_cpu(this_cpu, &mask); in bfin_wakeup_cpu()
|
/arch/ia64/kernel/ |
D | smp.c | 99 int this_cpu = get_cpu(); in handle_IPI() local 129 this_cpu, which); in handle_IPI()
|
D | perfmon.c | 5291 int this_cpu = smp_processor_id(); in pfm_overflow_handler() local 5326 pfm_stats[this_cpu].pfm_smpl_handler_calls++; in pfm_overflow_handler() 5349 pfm_stats[this_cpu].pfm_smpl_handler_cycles += end_cycles - start_cycles; in pfm_overflow_handler() 5491 int this_cpu = smp_processor_id(); in pfm_do_interrupt_handler() local 5494 pfm_stats[this_cpu].pfm_ovfl_intr_count++; in pfm_do_interrupt_handler() 5526 pfm_stats[this_cpu].pfm_spurious_ovfl_intr_count++; in pfm_do_interrupt_handler() 5538 this_cpu, task_pid_nr(task)); in pfm_do_interrupt_handler() 5543 this_cpu, in pfm_do_interrupt_handler() 5554 int this_cpu; in pfm_interrupt_handler() local 5558 this_cpu = get_cpu(); in pfm_interrupt_handler() [all …]
|
D | process.c | 255 unsigned int this_cpu = smp_processor_id(); in play_dead() local 263 ia64_jump_to_sal(&sal_boot_rendez_state[this_cpu]); in play_dead()
|
/arch/arm/kernel/ |
D | smp.c | 533 unsigned int this_cpu = smp_processor_id(); in smp_send_all_cpu_backtrace() local 544 cpu_clear(this_cpu, backtrace_mask); in smp_send_all_cpu_backtrace() 546 pr_info("Backtrace for cpu %d (current):\n", this_cpu); in smp_send_all_cpu_backtrace()
|
/arch/arm/common/ |
D | fiq_debugger.c | 828 int this_cpu, void *regs, void *svc_sp) in debug_handle_uart_interrupt() argument 835 if (this_cpu != state->current_cpu) { in debug_handle_uart_interrupt() 845 this_cpu); in debug_handle_uart_interrupt() 848 switch_cpu(state, this_cpu); in debug_handle_uart_interrupt() 920 unsigned int this_cpu = THREAD_INFO(svc_sp)->cpu; in debug_fiq() local 923 need_irq = debug_handle_uart_interrupt(state, this_cpu, regs, svc_sp); in debug_fiq()
|
/arch/sparc/mm/ |
D | init_64.c | 204 static inline void set_dcache_dirty(struct page *page, int this_cpu) in set_dcache_dirty() argument 206 unsigned long mask = this_cpu; in set_dcache_dirty() 274 int this_cpu = get_cpu(); in flush_dcache() local 279 if (cpu == this_cpu) in flush_dcache() 337 int this_cpu; in flush_dcache_page() local 349 this_cpu = get_cpu(); in flush_dcache_page() 357 if (dirty_cpu == this_cpu) in flush_dcache_page() 361 set_dcache_dirty(page, this_cpu); in flush_dcache_page()
|