/arch/m68k/hp300/ |
D | time.c | 78 unsigned char lsb, msb, msb_new; in hp300_read_clk() local 87 lsb = in_8(CLOCKBASE + CLKLSB1); in hp300_read_clk() 94 ticks = INTVAL - ((msb << 8) | lsb); in hp300_read_clk()
|
/arch/sparc/lib/ |
D | udivdi3.S | 24 addxcc %o1,%o1,%o1 ! shift n1n0 and a q-bit in lsb 31 addxcc %o1,%o1,%o1 ! shift n1n0 and a q-bit in lsb 40 addcc %o1,%o1,%o1 ! shift n1n0 and a 0-bit in lsb 62 addxcc %o2,%o2,%o2 ! shift n1n0 and a q-bit in lsb 69 addxcc %o2,%o2,%o2 ! shift n1n0 and a q-bit in lsb 78 addcc %o2,%o2,%o2 ! shift n1n0 and a 0-bit in lsb 88 addxcc %o1,%o1,%o1 ! shift n1n0 and a q-bit in lsb 95 addxcc %o1,%o1,%o1 ! shift n1n0 and a q-bit in lsb 104 addcc %o1,%o1,%o1 ! shift n1n0 and a 0-bit in lsb 167 addxcc %o5,%o5,%o5 ! shift n1n0 and a q-bit in lsb [all …]
|
D | divdi3.S | 48 addxcc %o1,%o1,%o1 ! shift n1n0 and a q-bit in lsb 55 addxcc %o1,%o1,%o1 ! shift n1n0 and a q-bit in lsb 64 addcc %o1,%o1,%o1 ! shift n1n0 and a 0-bit in lsb 84 addxcc %o2,%o2,%o2 ! shift n1n0 and a q-bit in lsb 91 addxcc %o2,%o2,%o2 ! shift n1n0 and a q-bit in lsb 100 addcc %o2,%o2,%o2 ! shift n1n0 and a 0-bit in lsb 108 addxcc %o1,%o1,%o1 ! shift n1n0 and a q-bit in lsb 115 addxcc %o1,%o1,%o1 ! shift n1n0 and a q-bit in lsb 124 addcc %o1,%o1,%o1 ! shift n1n0 and a 0-bit in lsb 183 addxcc %o1,%o1,%o1 ! shift n1n0 and a q-bit in lsb [all …]
|
/arch/parisc/mm/ |
D | fault.c | 392 unsigned int lsb = 0; in do_page_fault() local 402 lsb = hstate_index_to_shift(VM_FAULT_GET_HINDEX(fault)); in do_page_fault() 404 lsb = PAGE_SHIFT; in do_page_fault() 407 lsb); in do_page_fault()
|
/arch/arm64/kernel/ |
D | module.c | 153 int lsb, enum aarch64_insn_movw_imm_type imm_type) in reloc_insn_movw() argument 160 imm = sval >> lsb; in reloc_insn_movw() 194 int lsb, int len, enum aarch64_insn_imm_type imm_type) in reloc_insn_imm() argument 202 sval >>= lsb; in reloc_insn_imm() 205 imm_mask = (BIT(lsb + len) - 1) >> lsb; in reloc_insn_imm()
|
D | insn.c | 1637 u8 lsb) in aarch64_insn_gen_extr() argument 1645 if (lsb > 31) in aarch64_insn_gen_extr() 1649 if (lsb > 63) in aarch64_insn_gen_extr() 1659 insn = aarch64_insn_encode_immediate(AARCH64_INSN_IMM_S, insn, lsb); in aarch64_insn_gen_extr()
|
D | traps.c | 242 void arm64_force_sig_mceerr(int code, void __user *addr, short lsb, in arm64_force_sig_mceerr() argument 246 force_sig_mceerr(code, addr, lsb); in arm64_force_sig_mceerr()
|
/arch/x86/platform/ts5500/ |
D | ts5500.c | 250 u8 lsb, msb; in ts5500_adc_convert() local 265 lsb = inb(TS5500_ADC_CONV_INIT_LSB_ADDR); in ts5500_adc_convert() 268 return (msb << 8) | lsb; in ts5500_adc_convert()
|
/arch/powerpc/mm/ |
D | fault.c | 151 unsigned int lsb = 0; /* shutup gcc */ in do_sigbus() local 157 lsb = hstate_index_to_shift(VM_FAULT_GET_HINDEX(fault)); in do_sigbus() 159 lsb = PAGE_SHIFT; in do_sigbus() 161 force_sig_mceerr(BUS_MCEERR_AR, (void __user *)address, lsb); in do_sigbus()
|
/arch/powerpc/include/asm/ |
D | kvm_ppc.h | 363 static inline u32 kvmppc_get_field(u64 inst, int msb, int lsb) in kvmppc_get_field() argument 368 BUG_ON(msb > lsb); in kvmppc_get_field() 370 mask = (1 << (lsb - msb + 1)) - 1; in kvmppc_get_field() 371 r = (inst >> (63 - lsb)) & mask; in kvmppc_get_field() 379 static inline u32 kvmppc_set_field(u64 inst, int msb, int lsb, int value) in kvmppc_set_field() argument 384 BUG_ON(msb > lsb); in kvmppc_set_field() 386 mask = ((1 << (lsb - msb + 1)) - 1) << (63 - lsb); in kvmppc_set_field() 387 r = (inst & ~mask) | ((value << (63 - lsb)) & mask); in kvmppc_set_field()
|
/arch/arm64/include/asm/ |
D | traps.h | 30 void arm64_force_sig_mceerr(int code, void __user *addr, short lsb, const char *str);
|
D | insn.h | 452 u8 lsb);
|
/arch/arm64/mm/ |
D | fault.c | 601 unsigned int lsb; in do_page_fault() local 603 lsb = PAGE_SHIFT; in do_page_fault() 605 lsb = hstate_index_to_shift(VM_FAULT_GET_HINDEX(fault)); in do_page_fault() 607 arm64_force_sig_mceerr(BUS_MCEERR_AR, (void __user *)addr, lsb, in do_page_fault()
|
/arch/m68k/fpsp040/ |
D | scale.S | 205 | Result is zero. Check for rounding mode to set lsb. If the 217 movel #1,FPTEMP_LO(%a6) |set lsb 222 movel #1,FPTEMP_LO(%a6) |set lsb
|
D | res_func.S | 478 | to determine if the lsb of the result should be set. 497 | Check rmode for rp or rm and set lsb accordingly. 505 bra cu_dpdr |load double pos zero w/lsb 509 bra cu_dndr |load double neg zero w/lsb 512 | Check for rp or rm and set lsb accordingly. 520 bra cu_spdr |load single pos zero w/lsb 524 bra cu_sndr |load single neg zero w/lsb 537 | cu_dpdr and cu_dndr set the double lsb. 549 movel #0x800,LOCAL_LO(%a0) |with lsb set 563 movel #0x800,LOCAL_LO(%a0) |with lsb set [all …]
|
D | bindec.S | 75 | the lsb of Y.
|
/arch/x86/mm/ |
D | fault.c | 1016 unsigned lsb = 0; in do_sigbus() local 1022 lsb = hstate_index_to_shift(VM_FAULT_GET_HINDEX(fault)); in do_sigbus() 1024 lsb = PAGE_SHIFT; in do_sigbus() 1025 force_sig_mceerr(BUS_MCEERR_AR, (void __user *)address, lsb); in do_sigbus()
|
/arch/x86/kernel/cpu/mce/ |
D | amd.c | 892 u8 lsb = (m.addr >> 56) & 0x3f; in __log_error() local 894 m.addr &= GENMASK_ULL(55, lsb); in __log_error()
|
D | core.c | 661 u8 lsb = (m->addr >> 56) & 0x3f; in mce_read_aux() local 663 m->addr &= GENMASK_ULL(55, lsb); in mce_read_aux()
|
/arch/powerpc/kvm/ |
D | book3s_paired_singles.c | 339 static inline u32 inst_get_field(u32 inst, int msb, int lsb) in inst_get_field() argument 341 return kvmppc_get_field(inst, msb + 32, lsb + 32); in inst_get_field()
|
/arch/arm/boot/dts/ |
D | at91sam9x5.dtsi | 213 pinctrl_ebi_data_0_7: ebi-data-lsb-0 {
|
D | sama5d4.dtsi | 1069 pinctrl_ebi_data_0_7: ebi-data-lsb-0 {
|
/arch/x86/math-emu/ |
D | README | 233 these never exceeds 1/2 an lsb. The fprem and fprem1 instructions
|