/arch/powerpc/mm/ |
D | hugetlbpage-book3e.c | 18 struct paca_struct *paca = get_paca(); in tlb1_next() local 22 tcd = paca->tcd_ptr; in tlb1_next() 62 struct paca_struct *paca = get_paca(); in book3e_tlb_lock() local 86 : "r" (&paca->tcd_ptr->lock), "r" (token) in book3e_tlb_lock() 92 struct paca_struct *paca = get_paca(); in book3e_tlb_unlock() local 98 paca->tcd_ptr->lock = 0; in book3e_tlb_unlock()
|
D | tlb-radix.c | 577 if (paca[sib].kvm_hstate.kvm_vcpu) in radix_kvm_prefetch_workaround()
|
/arch/powerpc/kernel/ |
D | paca.c | 153 struct paca_struct *paca; variable 154 EXPORT_SYMBOL(paca); 224 paca = __va(memblock_alloc_base(paca_size, PAGE_SIZE, limit)); in allocate_pacas() 225 memset(paca, 0, paca_size); in allocate_pacas() 228 paca_size, nr_cpu_ids, paca); in allocate_pacas() 236 initialise_paca(&paca[cpu], cpu); in allocate_pacas() 248 memblock_free(__pa(paca) + new_size, paca_size - new_size); in free_unused_pacas()
|
D | setup_64.c | 112 paca[cpu].tcd_ptr = &paca[first].tcd; in setup_tlb_core_data() 304 setup_paca(&paca[boot_cpuid]); in early_setup() 607 paca[i].crit_kstack = __va(sp + THREAD_SIZE); in exc_lvl_early_init() 611 paca[i].dbg_kstack = __va(sp + THREAD_SIZE); in exc_lvl_early_init() 615 paca[i].mc_kstack = __va(sp + THREAD_SIZE); in exc_lvl_early_init() 672 paca[i].emergency_sp = (void *)ti + THREAD_SIZE; in emergency_stack_init() 679 paca[i].nmi_emergency_sp = (void *)ti + THREAD_SIZE; in emergency_stack_init() 685 paca[i].mc_emergency_sp = (void *)ti + THREAD_SIZE; in emergency_stack_init() 741 paca[cpu].data_offset = __per_cpu_offset[cpu]; in setup_per_cpu_areas() 855 paca[cpu].rfi_flush_fallback_area = l1d_flush_fallback_area; in init_fallback_flush() [all …]
|
D | machine_kexec_64.c | 183 while (paca[i].kexec_state < wait_state) { in kexec_prepare_cpus_wait() 188 i, paca[i].hw_cpu_id, wait_state); in kexec_prepare_cpus_wait() 330 paca = (struct paca_struct *)RELOC_HIDE(&kexec_paca, 0) - in default_machine_kexec()
|
D | smp.c | 126 if (!paca[nr].cpu_start) { in smp_generic_kick_cpu() 127 paca[nr].cpu_start = 1; in smp_generic_kick_cpu() 642 paca[boot_cpuid].__current = current; in smp_prepare_boot_cpu() 733 paca[cpu].__current = idle; in cpu_idle_thread_init() 734 paca[cpu].kstack = (unsigned long)ti + THREAD_SIZE - STACK_FRAME_OVERHEAD; in cpu_idle_thread_init()
|
D | crash.c | 233 while (paca[i].kexec_state < KEXEC_STATE_REAL_MODE) { in crash_kexec_wait_realmode()
|
D | head_64.S | 389 LOAD_REG_ADDR(r13, paca) /* Load paca pointer */ 753 LOAD_REG_ADDR(r4,paca) /* Load paca pointer */
|
D | Makefile | 43 paca.o nvram_64.o firmware.o
|
/arch/powerpc/platforms/cell/ |
D | smp.c | 86 task_thread_info(paca[lcpu].__current)->preempt_count = 0; in smp_startup_cpu() 129 paca[nr].cpu_start = 1; in smp_cell_kick_cpu()
|
/arch/powerpc/platforms/powernv/ |
D | idle.c | 83 uint64_t hsprg0_val = (uint64_t)&paca[cpu]; in pnv_save_sprs_for_deep_states() 176 paca[cpu].core_idle_state_ptr = core_idle_state; in pnv_alloc_idle_core_states() 177 paca[cpu].thread_idle_state = PNV_THREAD_RUNNING; in pnv_alloc_idle_core_states() 178 paca[cpu].thread_mask = 1 << j; in pnv_alloc_idle_core_states() 181 paca[cpu].thread_sibling_pacas = in pnv_alloc_idle_core_states() 752 paca[j].thread_sibling_pacas[idx] = &paca[cpu]; in pnv_init_idle_states()
|
D | setup.c | 303 i, paca[i].hw_cpu_id); in pnv_kexec_wait_secondaries_down() 315 i, paca[i].hw_cpu_id); in pnv_kexec_wait_secondaries_down()
|
D | smp.c | 74 if (paca[nr].cpu_start || !firmware_has_feature(FW_FEATURE_OPAL)) in pnv_smp_kick_cpu()
|
D | subcore.c | 283 paca[cpu].subcore_sibling_mask = mask; in update_subcore_sibling_mask()
|
/arch/powerpc/platforms/85xx/ |
D | smp.c | 150 paca[cpu].cpu_start = 0; in qoriq_cpu_kill() 331 paca[nr].cpu_start = 1; in smp_85xx_kick_cpu() 412 while (paca[disable_cpu].kexec_state < KEXEC_STATE_REAL_MODE) { in mpc85xx_smp_kexec_cpu_down() 419 paca[disable_cpu].kexec_state); in mpc85xx_smp_kexec_cpu_down()
|
/arch/powerpc/include/asm/ |
D | smp.h | 173 return paca[cpu].hw_cpu_id; in get_hard_smp_processor_id() 178 paca[cpu].hw_cpu_id = phys; in set_hard_smp_processor_id()
|
D | lppaca.h | 106 #define lppaca_of(cpu) (*paca[cpu].lppaca_ptr)
|
D | paca.h | 246 extern struct paca_struct *paca;
|
D | kvm_ppc.h | 431 paca[cpu].kvm_hstate.xics_phys = (void __iomem *)addr; in kvmppc_set_xics_phys() 438 paca[cpu].kvm_hstate.xive_tima_phys = (void __iomem *)phys_addr; in kvmppc_set_xive_tima() 439 paca[cpu].kvm_hstate.xive_tima_virt = virt_addr; in kvmppc_set_xive_tima() 453 paca[cpu].kvm_hstate.host_ipi = host_ipi; in kvmppc_set_host_ipi()
|
/arch/powerpc/platforms/pseries/ |
D | smp.c | 113 task_thread_info(paca[lcpu].__current)->preempt_count = 0; in smp_startup_cpu() 168 paca[nr].cpu_start = 1; in smp_pSeries_kick_cpu()
|
D | lpar.c | 103 addr = __pa(paca[cpu].slb_shadow_ptr); in vpa_init() 115 pp = &paca[cpu]; in vpa_init()
|
D | setup.c | 250 pp = &paca[cpu]; in alloc_dispatch_logs()
|
/arch/powerpc/sysdev/xics/ |
D | icp-native.c | 167 xics_phys = paca[cpu].kvm_hstate.xics_phys; in icp_native_cause_ipi_rm()
|
/arch/powerpc/kvm/ |
D | book3s_hv_builtin.c | 254 xics_phys = paca[cpu].kvm_hstate.xics_phys; in kvmhv_rm_send_ipi()
|
D | book3s_hv.c | 166 if (paca[cpu].kvm_hstate.xics_phys) { in kvmppc_ipi_thread() 2146 tpaca = &paca[cpu]; in kvmppc_grab_hwthread() 2179 tpaca = &paca[cpu]; in kvmppc_release_hwthread() 2247 tpaca = &paca[cpu]; in kvmppc_start_thread() 2273 if (paca[cpu + i].kvm_hstate.kvm_vcore) in kvmppc_wait_for_nap() 2283 if (paca[cpu + i].kvm_hstate.kvm_vcore) in kvmppc_wait_for_nap() 2777 paca[pcpu + thr].kvm_hstate.kvm_split_mode = sip; in kvmppc_run_core() 4307 if (paca[first_cpu].sibling_subcore_state) in kvm_init_subcore_bitmap() 4322 paca[cpu].sibling_subcore_state = sibling_subcore_state; in kvm_init_subcore_bitmap()
|