/arch/riscv/kernel/ |
D | kexec_relocate.S | 33 mv s4, zero 37 mv s8, zero 38 mv s9, zero 39 mv s10, zero 40 mv s11, zero 43 csrw CSR_SIE, zero 44 csrw CSR_SIP, zero 78 csrw CSR_SATP, zero 79 jalr zero, s8, 0 113 mv a3, zero [all …]
|
D | head.S | 134 csrw CSR_IE, zero 135 csrw CSR_IP, zero 186 csrw CSR_SCRATCH, zero 199 csrw CSR_IE, zero 200 csrw CSR_IP, zero 288 REG_S zero, (a3) 398 fmv.s.x f0, zero 399 fmv.s.x f1, zero 400 fmv.s.x f2, zero 401 fmv.s.x f3, zero [all …]
|
D | suspend_entry.S | 71 add t0, a0, zero 72 add t1, a1, zero 80 add a0, t0, zero 81 add a1, t1, zero 85 add a0, a1, zero 94 add a0, zero, zero
|
/arch/loongarch/lib/ |
D | clear_user.S | 33 1: st.b zero, a0, 0 55 0: st.d zero, a0, 0 59 bstrins.d a0, zero, 2, 0 66 1: st.d zero, a0, 0 67 2: st.d zero, a0, 8 68 3: st.d zero, a0, 16 69 4: st.d zero, a0, 24 70 5: st.d zero, a0, 32 71 6: st.d zero, a0, 40 72 7: st.d zero, a0, 48 [all …]
|
/arch/powerpc/crypto/ |
D | ghashp10-ppc.pl | 57 my ($zero,$t0,$t1,$t2,$xC2,$H,$Hh,$Hl,$lemask)=map("v$_",(4..12)); 85 vxor $zero,$zero,$zero 87 vsldoi $xC2,$xC2,$zero,15 # 0xe1... 88 vsldoi $t1,$zero,$t0,1 # ...1 99 vsldoi $xC2,$zero,$xC2,8 # 0xc2.0 100 vsldoi $Hl,$zero,$H,8 # ... and split 101 vsldoi $Hh,$H,$zero,8 127 vxor $zero,$zero,$zero 129 vsldoi $xC2,$xC2,$zero,15 # 0xe1... 130 vsldoi $t1,$zero,$t0,1 # ...1 [all …]
|
D | aesp10-ppc.pl | 121 my ($zero,$in0,$in1,$key,$rcon,$mask,$tmp)=map("v$_",(0..6)); 185 vxor $zero,$zero,$zero 191 ?vperm $outmask,$zero,$outmask,$outperm 202 vsldoi $tmp,$zero,$in0,12 # >>32 211 vsldoi $tmp,$zero,$tmp,12 # >>32 213 vsldoi $tmp,$zero,$tmp,12 # >>32 222 vsldoi $tmp,$zero,$in0,12 # >>32 231 vsldoi $tmp,$zero,$tmp,12 # >>32 233 vsldoi $tmp,$zero,$tmp,12 # >>32 239 vsldoi $tmp,$zero,$in0,12 # >>32 [all …]
|
/arch/loongarch/mm/ |
D | page.S | 16 st.d zero, a0, 0 17 st.d zero, a0, 8 18 st.d zero, a0, 16 19 st.d zero, a0, 24 20 st.d zero, a0, 32 21 st.d zero, a0, 40 22 st.d zero, a0, 48 23 st.d zero, a0, 56 25 st.d zero, a0, -64 26 st.d zero, a0, -56 [all …]
|
/arch/xtensa/lib/ |
D | strnlen_user.S | 74 bnone a9, a5, .Lz0 # if byte 0 is zero 75 bnone a9, a6, .Lz1 # if byte 1 is zero 76 bnone a9, a7, .Lz2 # if byte 2 is zero 77 bnone a9, a8, .Lz3 # if byte 3 is zero 88 bnone a9, a5, .Lz0 # if byte 0 is zero 89 bnone a9, a6, .Lz1 # if byte 1 is zero 102 # the zero byte in order to include the NULL terminator in the count. 104 .Lz3: # byte 3 is zero 105 addi a4, a4, 3 # point to zero byte 106 .Lz0: # byte 0 is zero [all …]
|
D | strncpy_user.S | 55 beqz a4, .Lret # if len is zero 72 beqz a9, .Lret # if byte 0 is zero 75 beqz a4, .Lret # if len is zero 82 beqz a9, .Lret # if byte 0 is zero 85 beqz a4, .Lret # if len is zero 89 beqz a9, .Lret # if byte 0 is zero 113 bnone a9, a5, .Lz0 # if byte 0 is zero 114 bnone a9, a6, .Lz1 # if byte 1 is zero 115 bnone a9, a7, .Lz2 # if byte 2 is zero 117 bnone a9, a8, .Lz3 # if byte 3 is zero [all …]
|
/arch/nios2/kernel/ |
D | insnemu.S | 115 stw zero, 0(sp) /* Save zero on stack to avoid special case for r0. */ 212 bne r7, zero, multiply 262 bne r7, zero, unsigned_division 270 bge r3,zero,dividend_is_nonnegative 271 sub r3, zero, r3 /* -r3 */ 273 bge r5, zero, divisor_is_nonnegative 274 sub r5, zero, r5 /* -r5 */ 304 cmplt r7, r3, zero /* r7 = MSB of r3 */ 325 bne r14, zero, divide_loop 340 bge r17, zero, quotient_is_nonnegative [all …]
|
/arch/arm/mach-zynq/ |
D | platsmp.c | 40 static u8 __iomem *zero; in zynq_cpun_start() local 47 zero = ioremap(0, trampoline_code_size); in zynq_cpun_start() 48 if (!zero) { in zynq_cpun_start() 53 zero = (__force u8 __iomem *)PAGE_OFFSET; in zynq_cpun_start() 62 memcpy_toio(zero, &zynq_secondary_trampoline, in zynq_cpun_start() 64 writel(address, zero + trampoline_size); in zynq_cpun_start() 71 iounmap(zero); in zynq_cpun_start()
|
/arch/riscv/crypto/ |
D | aes-riscv64-zvkned-zvbb-zvkg.S | 92 vsetivli zero, 4, e32, m1, ta, ma 113 vsetvli zero, t0, e32, m1, ta, ma 119 vsetvli zero, t0, e64, m2, ta, ma 123 vsetvli zero, t1, e32, m2, ta, ma 131 vsetvli zero, LEN32, e32, m4, ta, ma 143 vsetivli zero, 2, e64, m1, ta, ma 145 vsetivli zero, 1, e64, m1, tu, ma 148 vsetvli zero, LEN32, e32, m4, ta, ma 163 vsetivli zero, 4, e32, m1, ta, ma 165 vsetivli zero, 1, e8, m1, tu, ma [all …]
|
/arch/arm/include/asm/ |
D | tlbflush.h | 318 const int zero = 0; in __local_flush_tlb_all() local 321 tlb_op(TLB_V4_U_FULL | TLB_V6_U_FULL, "c8, c7, 0", zero); in __local_flush_tlb_all() 322 tlb_op(TLB_V4_D_FULL | TLB_V6_D_FULL, "c8, c6, 0", zero); in __local_flush_tlb_all() 323 tlb_op(TLB_V4_I_FULL | TLB_V6_I_FULL, "c8, c5, 0", zero); in __local_flush_tlb_all() 328 const int zero = 0; in local_flush_tlb_all() local 335 tlb_op(TLB_V7_UIS_FULL, "c8, c7, 0", zero); in local_flush_tlb_all() 345 const int zero = 0; in __flush_tlb_all() local 352 tlb_op(TLB_V7_UIS_FULL, "c8, c3, 0", zero); in __flush_tlb_all() 362 const int zero = 0; in __local_flush_tlb_mm() local 368 tlb_op(TLB_V4_U_FULL, "c8, c7, 0", zero); in __local_flush_tlb_mm() [all …]
|
/arch/alpha/lib/ |
D | strrchr.S | 22 zapnot a1, 1, a1 # e0 : zero extend our test character 23 mov zero, t6 # .. e1 : t6 is last match aligned addr 25 mov zero, t8 # .. e1 : t8 is last match byte compare mask 33 cmpbge zero, t0, t1 # .. e1 : bits set iff byte == zero 36 xor t0, a1, t2 # e0 : make bytes == c zero 37 cmpbge zero, t4, t4 # .. e1 : bits set iff byte is garbage 38 cmpbge zero, t2, t3 # e0 : bits set iff byte == c 50 cmpbge zero, t0, t1 # .. e1 : bits set iff byte == zero 51 cmpbge zero, t2, t3 # e0 : bits set iff byte == c 85 mov zero, v0 # e0 :
|
D | strchr.S | 22 zapnot a1, 1, a1 # e0 : zero extend the search character 29 cmpbge zero, t0, t2 # .. e1 : bits set iff byte == zero 33 cmpbge zero, t4, t4 # .. e1 : bits set iff byte is garbage 35 xor t0, a1, t1 # .. e1 : make bytes == c zero 36 cmpbge zero, t1, t3 # e0 : bits set iff byte == c 37 or t2, t3, t0 # e1 : bits set iff char match or zero match 45 cmpbge zero, t0, t2 # e0 : bits set iff byte == 0 46 cmpbge zero, t1, t3 # .. e1 : bits set iff byte == c 68 mov zero, v0 # e0 :
|
D | ev67-strrchr.S | 40 mov zero, t6 # E : t6 is last match aligned addr 43 mov zero, t8 # E : t8 is last match byte compare mask 55 cmpbge zero, t0, t1 # E : bits set iff byte == zero 57 xor t0, a1, t2 # E : make bytes == c zero 58 cmpbge zero, t4, t4 # E : bits set iff byte is garbage 60 cmpbge zero, t2, t3 # E : bits set iff byte == c 77 cmpbge zero, t0, t1 # E : bits set iff byte == zero 78 cmpbge zero, t2, t3 # E : bits set iff byte == c 103 subq t5, t2, t5 # E : Normalize leading zero count
|
D | ev67-strchr.S | 49 cmpbge zero, t0, t2 # E : bits set iff byte == zero 50 cmpbge zero, t4, t4 # E : bits set iff byte is garbage 54 xor t0, a1, t1 # E : make bytes == c zero 55 cmpbge zero, t1, t3 # E : bits set iff byte == c 56 or t2, t3, t0 # E : bits set iff char match or zero match 72 cmpbge zero, t0, t2 # E : bits set iff byte == 0 74 cmpbge zero, t1, t3 # E : bits set iff byte == c
|
D | stxncpy.S | 55 lda t2, -1 # e1 : build a mask against false zero 60 cmpbge zero, t2, t8 # .. e1 : bits set iff null found 74 cmpbge zero, t0, t8 # .. e1 (stall) 129 addq zero, 1, t10 # .. e1 : 165 or t0, t6, t6 # e1 : mask original data for zero test 166 cmpbge zero, t6, t8 # e0 : 175 cmpbge zero, t2, t8 # e0 : find nulls in second partial 188 cmpbge zero, t2, t8 # .. e1 : 217 cmpbge zero, t2, t8 # e1 : test new word for eos (stall) 233 cmpbge zero, t0, t8 # e0 : is the null in this first bit? [all …]
|
D | stxcpy.S | 47 lda t2, -1 # e1 : build a mask against false zero 52 cmpbge zero, t2, t8 # .. e1 : bits set iff null found 65 cmpbge zero, t1, t8 # e0 (stall) 143 cmpbge zero, t6, t8 # .. e1 : 149 cmpbge zero, t2, t8 # e0 : testing eos 160 cmpbge zero, t2, t8 # .. e1 : 187 cmpbge zero, t2, t8 # e0 : test new word for eos 202 cmpbge zero, t1, t8 # e0 : is the null in this first bit? 209 cmpbge zero, t1, t8 # .. e1 : 244 mov zero, t0 # .. e1 : [all …]
|
D | ev6-stxcpy.S | 58 lda t2, -1 # E : build a mask against false zero 64 cmpbge zero, t2, t8 # E : bits set iff null found 81 cmpbge zero, t1, t8 # E : (3 cycle stall) 166 cmpbge zero, t6, t8 # E : (stall) 172 cmpbge zero, t2, t8 # E : testing eos (stall) 184 cmpbge zero, t2, t8 # E : (stall for t2) 213 cmpbge zero, t2, t8 # E : test new word for eos 229 cmpbge zero, t1, t8 # E : is the null in this first bit? (stall) 236 cmpbge zero, t1, t8 # E : (stall) 272 mov zero, t0 # E : [all …]
|
/arch/m68k/fpsp040/ |
D | do_func.S | 8 | tbldo.sa. Cases of zero, infinity and NaN are handled in 101 | Load a signed zero to fp0 and set inex2/ainex 110 bsr ld_mzero |if neg, load neg zero, return here 113 | Load a signed zero to fp0; do not set inex2/ainex 118 bne ld_mzero |if neg, load neg zero 119 bra ld_pzero |load positive zero 169 fbeq t_dz2 |if = -1, divide by zero exception 183 | Test for 1.0 as an input argument, returning +zero. Also check 256 .long smod_oper | 00,01 norm,zero = nan with operr 259 .long smod_zro | 01,00 zero,norm = +-zero [all …]
|
/arch/mips/kernel/ |
D | relocate_kernel.S | 37 beq s3, zero, 1f 44 beq s3, zero, 1f 51 beq s3, zero, 1f 56 beq s3, zero, process_entry 67 beq s6, zero, process_entry 83 LONG_S zero,(t0) 128 bne s0, zero,1b
|
D | head.S | 42 sll zero,3 # ehb 97 LONG_S zero, (t0) 101 LONG_S zero, (t0) 109 MTC0 zero, CP0_CONTEXT # clear context register 111 MTC0 zero, CP0_XCONTEXT
|
/arch/arm64/include/asm/ |
D | asm-extable.h | 40 #define _ASM_EXTABLE_UACCESS_ERR_ZERO(insn, fixup, err, zero) \ 45 EX_DATA_REG(ZERO, zero) \ 88 #define _ASM_EXTABLE_UACCESS_ERR_ZERO(insn, fixup, err, zero) \ argument 94 EX_DATA_REG(ZERO, zero) \ 97 #define _ASM_EXTABLE_KACCESS_ERR_ZERO(insn, fixup, err, zero) \ argument 103 EX_DATA_REG(ZERO, zero) \
|
/arch/mips/alchemy/common/ |
D | sleeper.S | 73 sw zero, 0x0078(t3) /* sys_slppwr */ 75 sw zero, 0x007c(t3) /* sys_sleep */ 102 sw zero, 0x001c(a0) /* Precharge */ 104 sw zero, 0x0020(a0) /* Auto Refresh */ 106 sw zero, 0x0030(a0) /* Sleep */ 128 sw zero, 0x08c0(a0) /* Precharge */ 130 sw zero, 0x08d0(a0) /* Self Refresh */ 137 beq t2, zero, 2b 174 sw zero, 0x868(a0) /* mem_sdportcfga */ 190 sw zero, 0x08c8(a0) /* mem_sdautoref */
|