/arch/hexagon/mm/ |
D | copy_user_template.S | 32 p0 = cmp.gtu(bytes,#0) 38 p1 = cmp.gtu(bytes,#15) 45 loopcount = lsr(bytes,#3) 59 bytes -= asl(loopcount,#3) 71 p1 = cmp.gtu(bytes,#7) 76 loopcount = lsr(bytes,#2) 89 bytes -= asl(loopcount,#2) 97 p1 = cmp.gtu(bytes,#3) 102 loopcount = lsr(bytes,#1) 115 bytes -= asl(loopcount,#1) [all …]
|
/arch/sparc/include/asm/ |
D | xor_32.h | 24 sparc_2(unsigned long bytes, unsigned long *p1, unsigned long *p2) in sparc_2() argument 26 int lines = bytes / (sizeof (long)) / 8; in sparc_2() 61 sparc_3(unsigned long bytes, unsigned long *p1, unsigned long *p2, in sparc_3() argument 64 int lines = bytes / (sizeof (long)) / 8; in sparc_3() 112 sparc_4(unsigned long bytes, unsigned long *p1, unsigned long *p2, in sparc_4() argument 115 int lines = bytes / (sizeof (long)) / 8; in sparc_4() 176 sparc_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, in sparc_5() argument 179 int lines = bytes / (sizeof (long)) / 8; in sparc_5()
|
/arch/alpha/lib/ |
D | ev6-memcpy.S | 67 ldq $1, 0($17) # L : get 8 bytes 84 wh64 ($7) # L1 : memory subsystem hint: 64 bytes at 86 ldq $6, 0($17) # L0 : bytes 0..7 90 ldq $4, 8($17) # L : bytes 8..15 91 ldq $5, 16($17) # L : bytes 16..23 95 ldq $3, 24($17) # L : bytes 24..31 100 addq $17, 32, $17 # E : src += 32 bytes 101 stq $6, 0($16) # L : bytes 0..7 105 stq $4, 8($16) # L : bytes 8..15 106 stq $5, 16($16) # L : bytes 16..23 [all …]
|
/arch/arm/plat-omap/include/plat/ |
D | iommu.h | 157 #define iopgsz_max(bytes) \ argument 158 (((bytes) >= SZ_16M) ? SZ_16M : \ 159 ((bytes) >= SZ_1M) ? SZ_1M : \ 160 ((bytes) >= SZ_64K) ? SZ_64K : \ 161 ((bytes) >= SZ_4K) ? SZ_4K : 0) 163 #define bytes_to_iopgsz(bytes) \ argument 164 (((bytes) == SZ_16M) ? MMU_CAM_PGSZ_16M : \ 165 ((bytes) == SZ_1M) ? MMU_CAM_PGSZ_1M : \ 166 ((bytes) == SZ_64K) ? MMU_CAM_PGSZ_64K : \ 167 ((bytes) == SZ_4K) ? MMU_CAM_PGSZ_4K : -1) [all …]
|
/arch/s390/crypto/ |
D | ghash_s390.c | 25 u32 bytes; member 62 if (dctx->bytes) { in ghash_update() 63 u8 *pos = buf + (GHASH_BLOCK_SIZE - dctx->bytes); in ghash_update() 65 n = min(srclen, dctx->bytes); in ghash_update() 66 dctx->bytes -= n; in ghash_update() 72 if (!dctx->bytes) { in ghash_update() 88 dctx->bytes = GHASH_BLOCK_SIZE - srclen; in ghash_update() 100 if (dctx->bytes) { in ghash_flush() 101 u8 *pos = buf + (GHASH_BLOCK_SIZE - dctx->bytes); in ghash_flush() 103 memset(pos, 0, dctx->bytes); in ghash_flush() [all …]
|
/arch/x86/oprofile/ |
D | backtrace.c | 47 unsigned long bytes; in dump_user_backtrace_32() local 49 bytes = copy_from_user_nmi(bufhead, head, sizeof(bufhead)); in dump_user_backtrace_32() 50 if (bytes != sizeof(bufhead)) in dump_user_backtrace_32() 93 unsigned long bytes; in dump_user_backtrace() local 95 bytes = copy_from_user_nmi(bufhead, head, sizeof(bufhead)); in dump_user_backtrace() 96 if (bytes != sizeof(bufhead)) in dump_user_backtrace()
|
/arch/arm/include/asm/ |
D | xor.h | 47 xor_arm4regs_2(unsigned long bytes, unsigned long *p1, unsigned long *p2) in xor_arm4regs_2() argument 49 unsigned int lines = bytes / sizeof(unsigned long) / 4; in xor_arm4regs_2() 67 xor_arm4regs_3(unsigned long bytes, unsigned long *p1, unsigned long *p2, in xor_arm4regs_3() argument 70 unsigned int lines = bytes / sizeof(unsigned long) / 4; in xor_arm4regs_3() 89 xor_arm4regs_4(unsigned long bytes, unsigned long *p1, unsigned long *p2, in xor_arm4regs_4() argument 92 unsigned int lines = bytes / sizeof(unsigned long) / 2; in xor_arm4regs_4() 108 xor_arm4regs_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, in xor_arm4regs_5() argument 111 unsigned int lines = bytes / sizeof(unsigned long) / 2; in xor_arm4regs_5()
|
/arch/x86/include/asm/ |
D | xor_32.h | 32 xor_pII_mmx_2(unsigned long bytes, unsigned long *p1, unsigned long *p2) in xor_pII_mmx_2() argument 34 unsigned long lines = bytes >> 7; in xor_pII_mmx_2() 75 xor_pII_mmx_3(unsigned long bytes, unsigned long *p1, unsigned long *p2, in xor_pII_mmx_3() argument 78 unsigned long lines = bytes >> 7; in xor_pII_mmx_3() 124 xor_pII_mmx_4(unsigned long bytes, unsigned long *p1, unsigned long *p2, in xor_pII_mmx_4() argument 127 unsigned long lines = bytes >> 7; in xor_pII_mmx_4() 179 xor_pII_mmx_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, in xor_pII_mmx_5() argument 182 unsigned long lines = bytes >> 7; in xor_pII_mmx_5() 259 xor_p5_mmx_2(unsigned long bytes, unsigned long *p1, unsigned long *p2) in xor_p5_mmx_2() argument 261 unsigned long lines = bytes >> 6; in xor_p5_mmx_2() [all …]
|
D | insn.h | 29 insn_byte_t bytes[4]; member 146 return X86_VEX3_M(insn->vex_prefix.bytes[1]); in insn_vex_m_bits() 152 return X86_VEX_P(insn->vex_prefix.bytes[1]); in insn_vex_p_bits() 154 return X86_VEX_P(insn->vex_prefix.bytes[2]); in insn_vex_p_bits() 163 if (insn->prefixes.bytes[3]) in insn_last_prefix_id() 164 return inat_get_last_prefix_id(insn->prefixes.bytes[3]); in insn_last_prefix_id()
|
D | xor_64.h | 91 xor_sse_2(unsigned long bytes, unsigned long *p1, unsigned long *p2) in xor_sse_2() argument 93 unsigned int lines = bytes >> 8; in xor_sse_2() 142 xor_sse_3(unsigned long bytes, unsigned long *p1, unsigned long *p2, in xor_sse_3() argument 145 unsigned int lines = bytes >> 8; in xor_sse_3() 201 xor_sse_4(unsigned long bytes, unsigned long *p1, unsigned long *p2, in xor_sse_4() argument 204 unsigned int lines = bytes >> 8; in xor_sse_4() 268 xor_sse_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, in xor_sse_5() argument 271 unsigned int lines = bytes >> 8; in xor_sse_5()
|
D | kvm_emulate.h | 97 unsigned int bytes, 108 unsigned long addr, void *val, unsigned int bytes, 118 unsigned long addr, void *val, unsigned int bytes, 128 unsigned long addr, void *val, unsigned int bytes, 140 unsigned int bytes, 155 unsigned int bytes, 204 unsigned int bytes; member
|
/arch/s390/include/asm/ |
D | bitops.h | 412 unsigned long bytes = 0; in __ffz_word_loop() local 434 : "+&a" (bytes), "+&d" (size) in __ffz_word_loop() 437 return bytes; in __ffz_word_loop() 449 unsigned long bytes = 0; in __ffs_word_loop() local 471 : "+&a" (bytes), "+&a" (size) in __ffs_word_loop() 474 return bytes; in __ffs_word_loop() 616 unsigned long bytes, bits; in find_first_zero_bit() local 620 bytes = __ffz_word_loop(addr, size); in find_first_zero_bit() 621 bits = __ffz_word(bytes*8, __load_ulong_be(addr, bytes)); in find_first_zero_bit() 637 unsigned long bytes, bits; in find_first_bit() local [all …]
|
/arch/ia64/sn/kernel/sn2/ |
D | cache.c | 24 sn_flush_all_caches(long flush_addr, long bytes) in sn_flush_all_caches() argument 32 flush_icache_range(addr, addr + bytes); in sn_flush_all_caches() 38 flush_icache_range(addr, addr + bytes); in sn_flush_all_caches()
|
/arch/arm/lib/ |
D | memzero.S | 22 blt 5f @ 1 bytes to align with? 41 blt 4f @ 1 have < 16 bytes 53 3: subs r1, r1, #64 @ 1 write 32 bytes out per loop 66 tst r1, #16 @ 1 16 bytes or more? 111 4: tst r1, #8 @ 1 8 bytes or more? 113 tst r1, #4 @ 1 4 bytes or more? 119 5: tst r1, #2 @ 1 2 bytes or more?
|
/arch/x86/kernel/ |
D | ioport.c | 27 unsigned int i, max_long, bytes, bytes_updated; in sys_ioperm() local 73 bytes = (max_long + 1) * sizeof(unsigned long); in sys_ioperm() 74 bytes_updated = max(bytes, t->io_bitmap_max); in sys_ioperm() 76 t->io_bitmap_max = bytes; in sys_ioperm()
|
D | msr.c | 76 ssize_t bytes = 0; in msr_read() local 90 bytes += 8; in msr_read() 93 return bytes ? bytes : err; in msr_read() 104 ssize_t bytes = 0; in msr_write() local 118 bytes += 8; in msr_write() 121 return bytes ? bytes : err; in msr_write()
|
/arch/x86/crypto/ |
D | ghash-clmulni-intel_glue.c | 45 u32 bytes; member 80 if (dctx->bytes) { in ghash_update() 81 int n = min(srclen, dctx->bytes); in ghash_update() 82 u8 *pos = dst + (GHASH_BLOCK_SIZE - dctx->bytes); in ghash_update() 84 dctx->bytes -= n; in ghash_update() 90 if (!dctx->bytes) in ghash_update() 100 dctx->bytes = GHASH_BLOCK_SIZE - srclen; in ghash_update() 112 if (dctx->bytes) { in ghash_flush() 113 u8 *tmp = dst + (GHASH_BLOCK_SIZE - dctx->bytes); in ghash_flush() 115 while (dctx->bytes--) in ghash_flush() [all …]
|
/arch/powerpc/platforms/powermac/ |
D | pfunc_core.c | 81 static void print_blob(const char *title, const void *blob, int bytes) 84 while(bytes--) { 271 u32 bytes = pmf_next32(cmd); in pmf_parser_read_i2c() local 273 LOG_PARSE("pmf: read_i2c(bytes: %ud)\n", bytes); in pmf_parser_read_i2c() 275 PMF_PARSE_CALL(read_i2c, cmd, h, bytes); in pmf_parser_read_i2c() 280 u32 bytes = pmf_next32(cmd); in pmf_parser_write_i2c() local 281 const void *blob = pmf_next_blob(cmd, bytes); in pmf_parser_write_i2c() 283 LOG_PARSE("pmf: write_i2c(bytes: %ud) ...\n", bytes); in pmf_parser_write_i2c() 284 LOG_BLOB("pmf: data: \n", blob, bytes); in pmf_parser_write_i2c() 286 PMF_PARSE_CALL(write_i2c, cmd, h, bytes, blob); in pmf_parser_write_i2c() [all …]
|
/arch/x86/lib/ |
D | insn.c | 87 if (prefixes->bytes[i] == b) in insn_get_prefixes() 92 prefixes->bytes[nb++] = b; in insn_get_prefixes() 111 if (lb && lb != insn->prefixes.bytes[3]) { in insn_get_prefixes() 112 if (unlikely(insn->prefixes.bytes[3])) { in insn_get_prefixes() 114 b = insn->prefixes.bytes[3]; in insn_get_prefixes() 116 if (prefixes->bytes[i] == lb) in insn_get_prefixes() 117 prefixes->bytes[i] = b; in insn_get_prefixes() 119 insn->prefixes.bytes[3] = lb; in insn_get_prefixes() 151 insn->vex_prefix.bytes[0] = b; in insn_get_prefixes() 152 insn->vex_prefix.bytes[1] = b2; in insn_get_prefixes() [all …]
|
/arch/xtensa/lib/ |
D | memset.S | 48 slli a7, a3, 8 # duplicate character in all bytes of word 64 # set 16 bytes per iteration for word-aligned dst 86 # set 8 bytes 92 # set 4 bytes 97 # set 2 bytes 124 # set 2 bytes
|
D | usercopy.S | 113 # copy 2 bytes 146 movi a2, 0 # return success for len bytes copied 152 # copy 16 bytes per iteration for word-aligned dst and word-aligned src 179 # copy 8 bytes 188 # copy 4 bytes 195 # copy 2 bytes 206 movi a2, 0 # return success for len bytes copied 217 # copy 16 bytes per iteration for word-aligned dst and unaligned src 248 # copy 8 bytes 260 # copy 4 bytes [all …]
|
D | memcopy.S | 132 # copy 2 bytes 165 # copy 16 bytes per iteration for word-aligned dst and word-aligned src 189 # copy 8 bytes 202 # copy 4 bytes 211 # copy 2 bytes 231 # copy 16 bytes per iteration for word-aligned dst and unaligned src 267 # copy 8 bytes 279 # copy 4 bytes 294 # copy 2 bytes
|
/arch/powerpc/sysdev/xics/ |
D | icp-native.c | 33 u8 bytes[4]; member 37 u8 bytes[4]; member 42 u8 bytes[4]; member 69 out_8(&icp_native_regs[cpu]->xirr.bytes[0], value); in icp_native_set_cppr() 74 out_8(&icp_native_regs[n_cpu]->qirr.bytes[0], value); in icp_native_set_qirr()
|
/arch/x86/tools/ |
D | test_get_len.c | 64 indent, field->value, field->bytes[0], field->bytes[1], in dump_field() 65 field->bytes[2], field->bytes[3]); in dump_field()
|
/arch/tile/lib/ |
D | usercopy_64.S | 39 #define __get_user_N(bytes, LOAD) \ argument 40 STD_ENTRY(__get_user_##bytes); \ 43 STD_ENDPROC(__get_user_##bytes); \ 57 #define __put_user_N(bytes, STORE) \ argument 58 STD_ENTRY(__put_user_##bytes); \ 61 STD_ENDPROC(__put_user_##bytes); \
|