/arch/sparc/lib/ |
D | memcpy.S | 17 #define MOVE_BIGCHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, t7) \ argument 31 #define MOVE_BIGALIGNCHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, t7) \ argument 62 #define RMOVE_BIGCHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, t7) \ argument 76 #define RMOVE_BIGALIGNCHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, t7) \ argument 100 #define SMOVE_CHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, prev, shil, shir, offset2) \ argument 118 #define SMOVE_ALIGNCHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, prev, shil, shir, offset2) \ argument
|
D | blockops.S | 26 #define MIRROR_BLOCK(dst, src, offset, t0, t1, t2, t3, t4, t5, t6, t7) \ argument
|
D | copy_user.S | 67 #define MOVE_BIGCHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, t7) \ argument 81 #define MOVE_BIGALIGNCHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, t7) \ argument
|
D | checksum_32.S | 191 #define CSUMCOPY_BIGCHUNK_ALIGNED(src, dst, sum, off, t0, t1, t2, t3, t4, t5, t6, t7) \ argument 213 #define CSUMCOPY_BIGCHUNK(src, dst, sum, off, t0, t1, t2, t3, t4, t5, t6, t7) \ argument
|
/arch/ia64/lib/ |
D | copy_page_mck.S | 80 #define t6 t2 // alias! macro
|
D | memcpy_mck.S | 48 #define t6 t2 // alias! macro
|
/arch/alpha/include/uapi/asm/ |
D | regdef.h | 12 #define t6 $7 macro
|
/arch/mips/cavium-octeon/ |
D | octeon-memcpy.S | 114 #define t6 $14 macro
|
/arch/tile/kernel/ |
D | hvglue_trace.c | 159 #define __HV_DECL6(t6, a6, ...) t6 a6, __HV_DECL5(__VA_ARGS__) argument 168 #define __HV_PASS6(t6, a6, ...) a6, __HV_PASS5(__VA_ARGS__) argument
|
/arch/mips/include/asm/ |
D | regdef.h | 38 #define t6 $14 macro
|
/arch/mips/kernel/ |
D | pm-cps.c | 82 t0, t1, t2, t3, t4, t5, t6, t7, enumerator
|
/arch/mips/lib/ |
D | memcpy.S | 176 #define t6 $14 macro
|
D | csum_partial.S | 34 #define t6 $14 macro
|
/arch/x86/crypto/ |
D | camellia-aesni-avx-asm_64.S | 50 #define roundsm16(x0, x1, x2, x3, x4, x5, x6, x7, t0, t1, t2, t3, t4, t5, t6, \ argument
|
D | camellia-aesni-avx2-asm_64.S | 67 #define roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, t0, t1, t2, t3, t4, t5, t6, \ argument
|