/lib/raid6/ |
D | avx2.c | 40 int d, z, z0; in raid6_avx21_gen_syndrome() local 57 for (z = z0-2; z >= 0; z--) { in raid6_avx21_gen_syndrome() 58 asm volatile("prefetchnta %0" : : "m" (dptr[z][d])); in raid6_avx21_gen_syndrome() 65 asm volatile("vmovdqa %0,%%ymm6" : : "m" (dptr[z][d])); in raid6_avx21_gen_syndrome() 89 int d, z, z0; in raid6_avx21_xor_syndrome() local 104 for (z = z0-1 ; z >= start ; z--) { in raid6_avx21_xor_syndrome() 110 asm volatile("vmovdqa %0,%%ymm5" :: "m" (dptr[z][d])); in raid6_avx21_xor_syndrome() 115 for (z = start-1 ; z >= 0 ; z--) { in raid6_avx21_xor_syndrome() 147 int d, z, z0; in raid6_avx22_gen_syndrome() local 166 for (z = z0-1; z >= 0; z--) { in raid6_avx22_gen_syndrome() [all …]
|
D | sse2.c | 40 int d, z, z0; in raid6_sse21_gen_syndrome() local 57 for ( z = z0-2 ; z >= 0 ; z-- ) { in raid6_sse21_gen_syndrome() 58 asm volatile("prefetchnta %0" : : "m" (dptr[z][d])); in raid6_sse21_gen_syndrome() 66 asm volatile("movdqa %0,%%xmm6" : : "m" (dptr[z][d])); in raid6_sse21_gen_syndrome() 92 int d, z, z0; in raid6_sse21_xor_syndrome() local 107 for ( z = z0-1 ; z >= start ; z-- ) { in raid6_sse21_xor_syndrome() 113 asm volatile("movdqa %0,%%xmm5" :: "m" (dptr[z][d])); in raid6_sse21_xor_syndrome() 118 for ( z = start-1 ; z >= 0 ; z-- ) { in raid6_sse21_xor_syndrome() 150 int d, z, z0; in raid6_sse22_gen_syndrome() local 169 for ( z = z0-1 ; z >= 0 ; z-- ) { in raid6_sse22_gen_syndrome() [all …]
|
D | avx512.c | 48 int d, z, z0; in raid6_avx5121_gen_syndrome() local 69 for (z = z0-2; z >= 0; z--) { in raid6_avx5121_gen_syndrome() 80 : "m" (dptr[z][d])); in raid6_avx5121_gen_syndrome() 106 int d, z, z0; in raid6_avx5121_xor_syndrome() local 124 for (z = z0-1 ; z >= start ; z--) { in raid6_avx5121_xor_syndrome() 135 : "m" (dptr[z][d])); in raid6_avx5121_xor_syndrome() 138 for (z = start-1 ; z >= 0 ; z--) { in raid6_avx5121_xor_syndrome() 175 int d, z, z0; in raid6_avx5122_gen_syndrome() local 198 for (z = z0-1; z >= 0; z--) { in raid6_avx5122_gen_syndrome() 218 : "m" (dptr[z][d]), "m" (dptr[z][d+64])); in raid6_avx5122_gen_syndrome() [all …]
|
D | sse1.c | 44 int d, z, z0; in raid6_sse11_gen_syndrome() local 61 for ( z = z0-2 ; z >= 0 ; z-- ) { in raid6_sse11_gen_syndrome() 62 asm volatile("prefetchnta %0" : : "m" (dptr[z][d])); in raid6_sse11_gen_syndrome() 70 asm volatile("movq %0,%%mm6" : : "m" (dptr[z][d])); in raid6_sse11_gen_syndrome() 103 int d, z, z0; in raid6_sse12_gen_syndrome() local 122 for ( z = z0-1 ; z >= 0 ; z-- ) { in raid6_sse12_gen_syndrome() 123 asm volatile("prefetchnta %0" : : "m" (dptr[z][d])); in raid6_sse12_gen_syndrome() 132 asm volatile("movq %0,%%mm5" : : "m" (dptr[z][d])); in raid6_sse12_gen_syndrome() 133 asm volatile("movq %0,%%mm7" : : "m" (dptr[z][d+8])); in raid6_sse12_gen_syndrome()
|
D | mmx.c | 39 int d, z, z0; in raid6_mmx1_gen_syndrome() local 53 for ( z = z0-1 ; z >= 0 ; z-- ) { in raid6_mmx1_gen_syndrome() 54 asm volatile("movq %0,%%mm6" : : "m" (dptr[z][d])); in raid6_mmx1_gen_syndrome() 87 int d, z, z0; in raid6_mmx2_gen_syndrome() local 104 for ( z = z0-1 ; z >= 0 ; z-- ) { in raid6_mmx2_gen_syndrome() 113 asm volatile("movq %0,%%mm5" : : "m" (dptr[z][d])); in raid6_mmx2_gen_syndrome() 114 asm volatile("movq %0,%%mm7" : : "m" (dptr[z][d+8])); in raid6_mmx2_gen_syndrome()
|
D | s390vx.uc | 48 static inline void AND(int x, int y, int z) 50 asm volatile ("VN %0,%1,%2" : : "i" (x), "i" (y), "i" (z)); 53 static inline void XOR(int x, int y, int z) 55 asm volatile ("VX %0,%1,%2" : : "i" (x), "i" (y), "i" (z)); 87 int d, z, z0; 100 for (z = z0 - 1; z >= 0; z--) { 105 LOAD_DATA(16,&dptr[z][d]); 120 int d, z, z0; 134 for (z = z0 - 1; z >= start; z--) { 139 LOAD_DATA(16,&dptr[z][d]); [all …]
|
D | neon.uc | 60 int d, z, z0; 71 for ( z = z0-1 ; z >= 0 ; z-- ) { 72 wd$$ = vld1q_u8(&dptr[z][d+$$*NSIZE]); 91 int d, z, z0; 105 for ( z = z0-1 ; z >= start ; z-- ) { 106 wd$$ = vld1q_u8(&dptr[z][d+$$*NSIZE]); 116 for ( z = start-1 ; z >= 3 ; z -= 4 ) { 124 switch (z) {
|
D | int.uc | 86 int d, z, z0; 96 for ( z = z0-1 ; z >= 0 ; z-- ) { 97 wd$$ = *(unative_t *)&dptr[z][d+$$*NSIZE]; 115 int d, z, z0; 126 for ( z = z0-1 ; z >= start ; z-- ) { 127 wd$$ = *(unative_t *)&dptr[z][d+$$*NSIZE]; 136 for ( z = start-1 ; z >= 0 ; z-- ) {
|
D | vpermxor.uc | 48 int d, z, z0; 58 for (z = z0-1; z>=0; z--) { 59 wd$$ = *(unative_t *)&dptr[z][d+$$*NSIZE];
|
D | altivec.uc | 75 int d, z, z0; 86 for ( z = z0-1 ; z >= 0 ; z-- ) { 87 wd$$ = *(unative_t *)&dptr[z][d+$$*NSIZE];
|
/lib/zlib_inflate/ |
D | inflate.c | 769 int zlib_inflateIncomp(z_stream *z) in zlib_inflateIncomp() argument 771 struct inflate_state *state = (struct inflate_state *)z->state; in zlib_inflateIncomp() 772 Byte *saved_no = z->next_out; in zlib_inflateIncomp() 773 uInt saved_ao = z->avail_out; in zlib_inflateIncomp() 779 z->avail_out = 0; in zlib_inflateIncomp() 780 z->next_out = (unsigned char*)z->next_in + z->avail_in; in zlib_inflateIncomp() 782 zlib_updatewindow(z, z->avail_in); in zlib_inflateIncomp() 785 z->avail_out = saved_ao; in zlib_inflateIncomp() 786 z->next_out = saved_no; in zlib_inflateIncomp() 788 z->adler = state->check = in zlib_inflateIncomp() [all …]
|
D | infutil.h | 23 #define WS(z) ((struct inflate_workspace *)(z->workspace)) argument
|
/lib/ |
D | inflate.c | 351 unsigned z; /* number of entries in current table */ in huft_build() local 450 z = 0; /* ditto */ in huft_build() 470 z = (z = g - w) > (unsigned)l ? l : z; /* upper limit on table size */ in huft_build() 476 if (j < z) in huft_build() 477 while (++j < z) /* try smaller tables up to z bits */ in huft_build() 485 z = 1 << j; /* table entries for j-bit table */ in huft_build() 488 if ((q = (struct huft *)malloc((z + 1)*sizeof(struct huft))) == in huft_build() 497 hufts += z + 1; /* track memory usage */ in huft_build() 536 for (j = i >> w; j < z; j += f) in huft_build()
|
D | bch.c | 806 const struct gf_poly *f, struct gf_poly *z, in compute_trace_bk_mod() argument 813 z->deg = 1; in compute_trace_bk_mod() 814 z->c[0] = 0; in compute_trace_bk_mod() 815 z->c[1] = bch->a_pow_tab[k]; in compute_trace_bk_mod() 825 for (j = z->deg; j >= 0; j--) { in compute_trace_bk_mod() 826 out->c[j] ^= z->c[j]; in compute_trace_bk_mod() 827 z->c[2*j] = gf_sqr(bch, z->c[j]); in compute_trace_bk_mod() 828 z->c[2*j+1] = 0; in compute_trace_bk_mod() 830 if (z->deg > out->deg) in compute_trace_bk_mod() 831 out->deg = z->deg; in compute_trace_bk_mod() [all …]
|
/lib/mpi/ |
D | mpicoder.c | 334 int x, j, z, lzeros, ents; in mpi_read_raw_from_sgl() local 395 z = BYTES_PER_MPI_LIMB - nbytes % BYTES_PER_MPI_LIMB; in mpi_read_raw_from_sgl() 396 z %= BYTES_PER_MPI_LIMB; in mpi_read_raw_from_sgl() 405 if (((z + x + 1) % BYTES_PER_MPI_LIMB) == 0) { in mpi_read_raw_from_sgl() 410 z += x; in mpi_read_raw_from_sgl()
|
/lib/crypto/ |
D | sha256.c | 21 static inline u32 Ch(u32 x, u32 y, u32 z) in Ch() argument 23 return z ^ (x & (y ^ z)); in Ch() 26 static inline u32 Maj(u32 x, u32 y, u32 z) in Maj() argument 28 return (x & y) | (z & (x | y)); in Maj()
|
D | aes.c | 104 u32 z = w & 0x40404040; in mul_by_x2() local 107 return (x << 2) ^ (y >> 7) * 0x36 ^ (z >> 6) * 0x1b; in mul_by_x2()
|