/arch/sparc/mm/ |
D | tlb.c | 25 struct tlb_batch *tb = &get_cpu_var(tlb_batch); in flush_tlb_pending() local 26 struct mm_struct *mm = tb->mm; in flush_tlb_pending() 28 if (!tb->tlb_nr) in flush_tlb_pending() 31 flush_tsb_user(tb); in flush_tlb_pending() 34 if (tb->tlb_nr == 1) { in flush_tlb_pending() 35 global_flush_tlb_page(mm, tb->vaddrs[0]); in flush_tlb_pending() 38 smp_flush_tlb_pending(tb->mm, tb->tlb_nr, in flush_tlb_pending() 39 &tb->vaddrs[0]); in flush_tlb_pending() 41 __flush_tlb_pending(CTX_HWBITS(tb->mm->context), in flush_tlb_pending() 42 tb->tlb_nr, &tb->vaddrs[0]); in flush_tlb_pending() [all …]
|
D | tsb.c | 80 static void __flush_tsb_one(struct tlb_batch *tb, unsigned long hash_shift, in __flush_tsb_one() argument 85 for (i = 0; i < tb->tlb_nr; i++) in __flush_tsb_one() 86 __flush_tsb_one_entry(tsb, tb->vaddrs[i], hash_shift, nentries); in __flush_tsb_one() 89 void flush_tsb_user(struct tlb_batch *tb) in flush_tsb_user() argument 91 struct mm_struct *mm = tb->mm; in flush_tsb_user() 96 if (!tb->huge) { in flush_tsb_user() 101 __flush_tsb_one(tb, PAGE_SHIFT, base, nentries); in flush_tsb_user() 104 if (tb->huge && mm->context.tsb_block[MM_TSB_HUGE].tsb) { in flush_tsb_user() 109 __flush_tsb_one(tb, REAL_HPAGE_SHIFT, base, nentries); in flush_tsb_user()
|
/arch/powerpc/kernel/ |
D | smp-tbsync.c | 24 volatile u64 tb; member 47 u64 tb; in smp_generic_take_timebase() local 62 tb = tbsync->tb; in smp_generic_take_timebase() 71 set_tb(tb >> 32, tb & 0xfffffffful); in smp_generic_take_timebase() 80 u64 tb; in start_contest() local 87 tb = get_tb() + 400; in start_contest() 88 tbsync->tb = tb + offset; in start_contest() 89 tbsync->mark = mark = tb + 400; in start_contest() 97 while (get_tb() <= tb) in start_contest()
|
D | time.c | 200 static unsigned long read_spurr(unsigned long tb) in read_spurr() argument 206 return tb; in read_spurr()
|
/arch/m68k/coldfire/ |
D | intc-simr.c | 132 u16 pa, tb; in intc_irq_set_type() local 136 tb = 0x1; in intc_irq_set_type() 139 tb = 0x2; in intc_irq_set_type() 142 tb = 0x3; in intc_irq_set_type() 146 tb = 0; in intc_irq_set_type() 150 if (tb) in intc_irq_set_type() 155 pa = (pa & ~(0x3 << ebit)) | (tb << ebit); in intc_irq_set_type()
|
D | intc-2.c | 148 u16 pa, tb; in intc_irq_set_type() local 152 tb = 0x1; in intc_irq_set_type() 155 tb = 0x2; in intc_irq_set_type() 158 tb = 0x3; in intc_irq_set_type() 162 tb = 0; in intc_irq_set_type() 166 if (tb) in intc_irq_set_type() 171 pa = (pa & ~(0x3 << (irq * 2))) | (tb << (irq * 2)); in intc_irq_set_type()
|
/arch/sparc/kernel/ |
D | irq_64.c | 1002 struct trap_per_cpu *tb = &trap_block[this_cpu]; in sun4v_register_mondo_queues() local 1004 register_one_mondo(tb->cpu_mondo_pa, HV_CPU_QUEUE_CPU_MONDO, in sun4v_register_mondo_queues() 1005 tb->cpu_mondo_qmask); in sun4v_register_mondo_queues() 1006 register_one_mondo(tb->dev_mondo_pa, HV_CPU_QUEUE_DEVICE_MONDO, in sun4v_register_mondo_queues() 1007 tb->dev_mondo_qmask); in sun4v_register_mondo_queues() 1008 register_one_mondo(tb->resum_mondo_pa, HV_CPU_QUEUE_RES_ERROR, in sun4v_register_mondo_queues() 1009 tb->resum_qmask); in sun4v_register_mondo_queues() 1010 register_one_mondo(tb->nonresum_mondo_pa, HV_CPU_QUEUE_NONRES_ERROR, in sun4v_register_mondo_queues() 1011 tb->nonresum_qmask); in sun4v_register_mondo_queues() 1033 static void __init init_cpu_send_mondo_info(struct trap_per_cpu *tb) in init_cpu_send_mondo_info() argument [all …]
|
D | smp_64.c | 293 struct trap_per_cpu *tb; in ldom_startcpu_cpuid() local 312 tb = &trap_block[cpu]; in ldom_startcpu_cpuid() 314 hdesc->fault_info_va = (unsigned long) &tb->fault_info; in ldom_startcpu_cpuid() 315 hdesc->fault_info_pa = kimage_addr_to_ra(&tb->fault_info); in ldom_startcpu_cpuid() 460 static void spitfire_xcall_deliver(struct trap_per_cpu *tb, int cnt) in spitfire_xcall_deliver() argument 468 cpu_list = __va(tb->cpu_list_pa); in spitfire_xcall_deliver() 469 mondo = __va(tb->cpu_mondo_block_pa); in spitfire_xcall_deliver() 481 static void cheetah_xcall_deliver(struct trap_per_cpu *tb, int cnt) in cheetah_xcall_deliver() argument 487 cpu_list = __va(tb->cpu_list_pa); in cheetah_xcall_deliver() 488 mondo = __va(tb->cpu_mondo_block_pa); in cheetah_xcall_deliver() [all …]
|
D | mdesc.c | 837 struct trap_per_cpu *tb) in get_mondo_data() argument 843 get_one_mondo_bits(val, &tb->cpu_mondo_qmask, 7, ilog2(max_cpus * 2)); in get_mondo_data() 846 get_one_mondo_bits(val, &tb->dev_mondo_qmask, 7, 8); in get_mondo_data() 849 get_one_mondo_bits(val, &tb->resum_qmask, 6, 7); in get_mondo_data() 852 get_one_mondo_bits(val, &tb->nonresum_qmask, 2, 2); in get_mondo_data() 856 tb->cpu_mondo_qmask + 1, in get_mondo_data() 857 tb->dev_mondo_qmask + 1, in get_mondo_data() 858 tb->resum_qmask + 1, in get_mondo_data() 859 tb->nonresum_qmask + 1); in get_mondo_data() 940 struct trap_per_cpu *tb; in fill_in_one_cpu() local [all …]
|
D | traps_64.c | 2008 struct trap_per_cpu *tb; in sun4v_resum_error() local 2014 tb = &trap_block[cpu]; in sun4v_resum_error() 2015 paddr = tb->resum_kernel_buf_pa + offset; in sun4v_resum_error() 2127 struct trap_per_cpu *tb; in sun4v_nonresum_error() local 2133 tb = &trap_block[cpu]; in sun4v_nonresum_error() 2134 paddr = tb->nonresum_kernel_buf_pa + offset; in sun4v_nonresum_error()
|
/arch/powerpc/boot/ |
D | 4xx.c | 339 u32 cpu, plb, opb, ebc, tb, uart0, uart1, m; in ibm440gp_fixup_clocks() local 363 tb = sys_clk; in ibm440gp_fixup_clocks() 366 tb = cpu; in ibm440gp_fixup_clocks() 385 dt_fixup_cpu_clocks(cpu, tb, 0); in ibm440gp_fixup_clocks() 426 u32 ccr1, tb = tmr_clk; in __ibm440eplike_fixup_clocks() local 469 if (tb == 0) { in __ibm440eplike_fixup_clocks() 474 tb = cpu; in __ibm440eplike_fixup_clocks() 476 dt_fixup_cpu_clocks(cpu, tb, 0); in __ibm440eplike_fixup_clocks() 558 u32 cpu, plb, opb, ebc, tb, uart0, uart1, m; in ibm405gp_fixup_clocks() local 614 tb = cpu; in ibm405gp_fixup_clocks() [all …]
|
D | devtree.c | 62 void dt_fixup_cpu_clocks(u32 cpu, u32 tb, u32 bus) in dt_fixup_cpu_clocks() argument 67 printf("CPU timebase-frequency <- 0x%x (%dMHz)\n\r", tb, MHZ(tb)); in dt_fixup_cpu_clocks() 73 setprop_val(devp, "timebase-frequency", tb); in dt_fixup_cpu_clocks() 78 timebase_period_ns = 1000000000 / tb; in dt_fixup_cpu_clocks()
|
/arch/x86/crypto/ |
D | fpu.c | 102 static struct crypto_instance *crypto_fpu_alloc(struct rtattr **tb) in crypto_fpu_alloc() argument 108 err = crypto_check_attr_type(tb, CRYPTO_ALG_TYPE_BLKCIPHER); in crypto_fpu_alloc() 112 alg = crypto_get_attr_alg(tb, CRYPTO_ALG_TYPE_BLKCIPHER, in crypto_fpu_alloc()
|
/arch/powerpc/kvm/ |
D | book3s_64_mmu.c | 58 if (vcpu->arch.slb[i].tb) in kvmppc_mmu_book3s_64_find_slbe() 72 vcpu->arch.slb[i].tb ? 't' : ' ', in kvmppc_mmu_book3s_64_find_slbe() 82 return slbe->tb ? SID_SHIFT_1T : SID_SHIFT; in kvmppc_slb_sid_shift() 146 ssize = slbe->tb ? MMU_SEGSIZE_1T : MMU_SEGSIZE_256M; in kvmppc_mmu_book3s_64_get_pteg() 249 if (slbe->tb) in kvmppc_mmu_book3s_64_xlate() 396 slbe->tb = (rs & SLB_VSID_B_1T) ? 1 : 0; in kvmppc_mmu_book3s_64_slbmte() 397 slbe->esid = slbe->tb ? esid_1t : esid; in kvmppc_mmu_book3s_64_slbmte() 589 if (slb->tb) { in kvmppc_mmu_book3s_64_esid_to_vsid()
|
D | emulate.c | 81 u32 kvmppc_get_dec(struct kvm_vcpu *vcpu, u64 tb) in kvmppc_get_dec() argument 83 u64 jd = tb - vcpu->arch.dec_jiffies; in kvmppc_get_dec()
|
D | booke.c | 548 u64 tb, wdt_tb, wdt_ticks = 0; in watchdog_next_timeout() local 553 tb = get_tb(); in watchdog_next_timeout() 558 if (tb & wdt_tb) in watchdog_next_timeout() 561 wdt_ticks += wdt_tb - (tb & (wdt_tb - 1)); in watchdog_next_timeout() 1485 u64 tb = get_tb(); in get_sregs_base() local 1496 sregs->u.e.dec = kvmppc_get_dec(vcpu, tb); in get_sregs_base() 1497 sregs->u.e.tb = tb; in get_sregs_base()
|
/arch/um/kernel/ |
D | dyn.lds.S | 41 .rel.tbss : { *(.rel.tbss .rel.tbss.* .rel.gnu.linkonce.tb.*) } 42 .rela.tbss : { *(.rela.tbss .rela.tbss.* .rela.gnu.linkonce.tb.*) } 116 .tbss : { *(.tbss .tbss.* .gnu.linkonce.tb.*) *(.tcommon) }
|
D | uml.lds.S | 91 .tbss : { *(.tbss .tbss.* .gnu.linkonce.tb.*) *(.tcommon) }
|
/arch/arm/crypto/ |
D | sha2-ce-core.S | 39 sha256h.32 dg0, dg1, tb\ev 40 sha256h2.32 dg1, dg2, tb\ev
|
D | sha1-ce-core.S | 40 vadd.u32 tb\ev, q\s0, \rc
|
/arch/sparc/include/asm/ |
D | tlbflush_64.h | 19 void flush_tsb_user(struct tlb_batch *tb);
|
/arch/ia64/include/asm/ |
D | kprobes.h | 49 unsigned long long tb : 1; member
|
D | processor.h | 107 __u64 tb : 1; member
|
/arch/ia64/kernel/ |
D | brl_emu.c | 206 } else if (ia64_psr(regs)->tb) { in ia64_emulate_brl()
|
/arch/powerpc/include/uapi/asm/ |
D | kvm.h | 226 __u64 tb; member
|