/lib/raid6/ |
D | sse2.c | 39 u8 *p, *q; in raid6_sse21_gen_syndrome() local 44 q = dptr[z0+2]; /* RS syndrome */ in raid6_sse21_gen_syndrome() 78 asm volatile("movntdq %%xmm4,%0" : "=m" (q[d])); in raid6_sse21_gen_syndrome() 91 u8 *p, *q; in raid6_sse21_xor_syndrome() local 96 q = dptr[disks-1]; /* RS syndrome */ in raid6_sse21_xor_syndrome() 125 asm volatile("pxor %0,%%xmm4" : : "m" (q[d])); in raid6_sse21_xor_syndrome() 127 asm volatile("movdqa %%xmm4,%0" : "=m" (q[d])); in raid6_sse21_xor_syndrome() 149 u8 *p, *q; in raid6_sse22_gen_syndrome() local 154 q = dptr[z0+2]; /* RS syndrome */ in raid6_sse22_gen_syndrome() 190 asm volatile("movntdq %%xmm4,%0" : "=m" (q[d])); in raid6_sse22_gen_syndrome() [all …]
|
D | avx2.c | 39 u8 *p, *q; in raid6_avx21_gen_syndrome() local 44 q = dptr[z0+2]; /* RS syndrome */ in raid6_avx21_gen_syndrome() 76 asm volatile("vmovntdq %%ymm4,%0" : "=m" (q[d])); in raid6_avx21_gen_syndrome() 88 u8 *p, *q; in raid6_avx21_xor_syndrome() local 93 q = dptr[disks-1]; /* RS syndrome */ in raid6_avx21_xor_syndrome() 122 asm volatile("vpxor %0,%%ymm4,%%ymm4" : : "m" (q[d])); in raid6_avx21_xor_syndrome() 124 asm volatile("vmovdqa %%ymm4,%0" : "=m" (q[d])); in raid6_avx21_xor_syndrome() 146 u8 *p, *q; in raid6_avx22_gen_syndrome() local 151 q = dptr[z0+2]; /* RS syndrome */ in raid6_avx22_gen_syndrome() 186 asm volatile("vmovntdq %%ymm4,%0" : "=m" (q[d])); in raid6_avx22_gen_syndrome() [all …]
|
D | avx512.c | 47 u8 *p, *q; in raid6_avx5121_gen_syndrome() local 52 q = dptr[z0+2]; /* RS syndrome */ in raid6_avx5121_gen_syndrome() 94 : "m" (p[d]), "m" (q[d])); in raid6_avx5121_gen_syndrome() 105 u8 *p, *q; in raid6_avx5121_xor_syndrome() local 110 q = dptr[disks-1]; /* RS syndrome */ in raid6_avx5121_xor_syndrome() 153 : "m" (q[d]), "m" (p[d])); in raid6_avx5121_xor_syndrome() 174 u8 *p, *q; in raid6_avx5122_gen_syndrome() local 179 q = dptr[z0+2]; /* RS syndrome */ in raid6_avx5122_gen_syndrome() 225 : "m" (p[d]), "m" (p[d+64]), "m" (q[d]), in raid6_avx5122_gen_syndrome() 226 "m" (q[d+64])); in raid6_avx5122_gen_syndrome() [all …]
|
D | recov_avx2.c | 21 u8 *p, *q, *dp, *dq; in raid6_2data_recov_avx2() local 27 q = (u8 *)ptrs[disks-1]; in raid6_2data_recov_avx2() 45 ptrs[disks-1] = q; in raid6_2data_recov_avx2() 59 asm volatile("vmovdqa %0, %%ymm1" : : "m" (q[0])); in raid6_2data_recov_avx2() 60 asm volatile("vmovdqa %0, %%ymm9" : : "m" (q[32])); in raid6_2data_recov_avx2() 132 q += 64; in raid6_2data_recov_avx2() 136 asm volatile("vmovdqa %0, %%ymm1" : : "m" (*q)); in raid6_2data_recov_avx2() 179 q += 32; in raid6_2data_recov_avx2() 191 u8 *p, *q, *dq; in raid6_datap_recov_avx2() local 196 q = (u8 *)ptrs[disks-1]; in raid6_datap_recov_avx2() [all …]
|
D | recov_avx512.c | 27 u8 *p, *q, *dp, *dq; in raid6_2data_recov_avx512() local 33 q = (u8 *)ptrs[disks-1]; in raid6_2data_recov_avx512() 54 ptrs[disks-1] = q; in raid6_2data_recov_avx512() 77 : "m" (q[0]), "m" (q[64]), "m" (p[0]), in raid6_2data_recov_avx512() 157 q += 128; in raid6_2data_recov_avx512() 166 : "m" (*q), "m" (*p), "m"(*dq), "m" (*dp)); in raid6_2data_recov_avx512() 218 q += 64; in raid6_2data_recov_avx512() 230 u8 *p, *q, *dq; in raid6_datap_recov_avx512() local 235 q = (u8 *)ptrs[disks-1]; in raid6_datap_recov_avx512() 250 ptrs[disks-1] = q; in raid6_datap_recov_avx512() [all …]
|
D | recov_ssse3.c | 21 u8 *p, *q, *dp, *dq; in raid6_2data_recov_ssse3() local 29 q = (u8 *)ptrs[disks-1]; in raid6_2data_recov_ssse3() 47 ptrs[disks-1] = q; in raid6_2data_recov_ssse3() 69 asm volatile("movdqa %0,%%xmm1" : : "m" (q[0])); in raid6_2data_recov_ssse3() 70 asm volatile("movdqa %0,%%xmm9" : : "m" (q[16])); in raid6_2data_recov_ssse3() 136 q += 32; in raid6_2data_recov_ssse3() 140 asm volatile("movdqa %0,%%xmm1" : : "m" (*q)); in raid6_2data_recov_ssse3() 183 q += 16; in raid6_2data_recov_ssse3() 196 u8 *p, *q, *dq; in raid6_datap_recov_ssse3() local 203 q = (u8 *)ptrs[disks-1]; in raid6_datap_recov_ssse3() [all …]
|
D | recov_neon.c | 22 void __raid6_2data_recov_neon(int bytes, uint8_t *p, uint8_t *q, uint8_t *dp, 26 void __raid6_datap_recov_neon(int bytes, uint8_t *p, uint8_t *q, uint8_t *dq, 32 u8 *p, *q, *dp, *dq; in raid6_2data_recov_neon() local 37 q = (u8 *)ptrs[disks - 1]; in raid6_2data_recov_neon() 57 ptrs[disks - 1] = q; in raid6_2data_recov_neon() 65 __raid6_2data_recov_neon(bytes, p, q, dp, dq, pbmul, qmul); in raid6_2data_recov_neon() 72 u8 *p, *q, *dq; in raid6_datap_recov_neon() local 76 q = (u8 *)ptrs[disks - 1]; in raid6_datap_recov_neon() 90 ptrs[disks - 1] = q; in raid6_datap_recov_neon() 96 __raid6_datap_recov_neon(bytes, p, q, dq, qmul); in raid6_datap_recov_neon()
|
D | recov.c | 23 u8 *p, *q, *dp, *dq; in raid6_2data_recov_intx1() local 29 q = (u8 *)ptrs[disks-1]; in raid6_2data_recov_intx1() 47 ptrs[disks-1] = q; in raid6_2data_recov_intx1() 56 qx = qmul[*q ^ *dq]; in raid6_2data_recov_intx1() 59 p++; q++; in raid6_2data_recov_intx1() 67 u8 *p, *q, *dq; in raid6_datap_recov_intx1() local 71 q = (u8 *)ptrs[disks-1]; in raid6_datap_recov_intx1() 83 ptrs[disks-1] = q; in raid6_datap_recov_intx1() 90 *p++ ^= *dq = qmul[*q ^ *dq]; in raid6_datap_recov_intx1() 91 q++; dq++; in raid6_datap_recov_intx1()
|
D | recov_s390xc.c | 26 u8 *p, *q, *dp, *dq; in raid6_2data_recov_s390xc() local 32 q = (u8 *)ptrs[disks-1]; in raid6_2data_recov_s390xc() 50 ptrs[disks-1] = q; in raid6_2data_recov_s390xc() 59 xor_block(dq, q); in raid6_2data_recov_s390xc() 64 q += 256; in raid6_2data_recov_s390xc() 75 u8 *p, *q, *dq; in raid6_datap_recov_s390xc() local 80 q = (u8 *)ptrs[disks-1]; in raid6_datap_recov_s390xc() 92 ptrs[disks-1] = q; in raid6_datap_recov_s390xc() 99 xor_block(dq, q); in raid6_datap_recov_s390xc() 104 q += 256; in raid6_datap_recov_s390xc()
|
D | mmx.c | 38 u8 *p, *q; in raid6_mmx1_gen_syndrome() local 43 q = dptr[z0+2]; /* RS syndrome */ in raid6_mmx1_gen_syndrome() 65 asm volatile("movq %%mm4,%0" : "=m" (q[d])); in raid6_mmx1_gen_syndrome() 86 u8 *p, *q; in raid6_mmx2_gen_syndrome() local 91 q = dptr[z0+2]; /* RS syndrome */ in raid6_mmx2_gen_syndrome() 124 asm volatile("movq %%mm4,%0" : "=m" (q[d])); in raid6_mmx2_gen_syndrome() 125 asm volatile("movq %%mm6,%0" : "=m" (q[d+8])); in raid6_mmx2_gen_syndrome()
|
D | sse1.c | 43 u8 *p, *q; in raid6_sse11_gen_syndrome() local 48 q = dptr[z0+2]; /* RS syndrome */ in raid6_sse11_gen_syndrome() 81 asm volatile("movntq %%mm4,%0" : "=m" (q[d])); in raid6_sse11_gen_syndrome() 102 u8 *p, *q; in raid6_sse12_gen_syndrome() local 107 q = dptr[z0+2]; /* RS syndrome */ in raid6_sse12_gen_syndrome() 143 asm volatile("movntq %%mm4,%0" : "=m" (q[d])); in raid6_sse12_gen_syndrome() 144 asm volatile("movntq %%mm6,%0" : "=m" (q[d+8])); in raid6_sse12_gen_syndrome()
|
D | recov_neon_inner.c | 27 void __raid6_2data_recov_neon(int bytes, uint8_t *p, uint8_t *q, uint8_t *dp, in __raid6_2data_recov_neon() argument 53 vx = veorq_u8(vld1q_u8(q), vld1q_u8(dq)); in __raid6_2data_recov_neon() 71 q += 16; in __raid6_2data_recov_neon() 77 void __raid6_datap_recov_neon(int bytes, uint8_t *p, uint8_t *q, uint8_t *dq, in __raid6_datap_recov_neon() argument 94 vx = veorq_u8(vld1q_u8(q), vld1q_u8(dq)); in __raid6_datap_recov_neon() 107 q += 16; in __raid6_datap_recov_neon()
|
D | neon.uc | 59 uint8_t *p, *q; 67 q = dptr[z0+2]; /* RS syndrome */ 82 vst1q_u8(&q[d+NSIZE*$$], wq$$); 90 uint8_t *p, *q; 98 q = dptr[disks-1]; /* RS syndrome */ 146 w1$$ = vld1q_u8(&q[d+NSIZE*$$]); 150 vst1q_u8(&q[d+NSIZE*$$], wq$$);
|
D | s390vx.uc | 86 u8 **dptr, *p, *q; 95 q = dptr[z0 + 2]; /* RS syndrome */ 110 STORE_DATA(8,&q[d]); 119 u8 **dptr, *p, *q; 125 q = dptr[disks - 1]; /* RS syndrome */ 153 LOAD_DATA(16,&q[d]); 155 STORE_DATA(16,&q[d]);
|
D | int.uc | 85 u8 *p, *q; 92 q = dptr[z0+2]; /* RS syndrome */ 106 *(unative_t *)&q[d+NSIZE*$$] = wq$$; 114 u8 *p, *q; 121 q = dptr[disks-1]; /* RS syndrome */ 143 *(unative_t *)&q[d+NSIZE*$$] ^= wq$$;
|
D | vpermxor.uc | 47 u8 *p, *q; 53 q = dptr[z0+2]; /* RS syndrome */ 68 *(unative_t *)&q[d+NSIZE*$$] = wq$$;
|
/lib/ |
D | ts_kmp.c | 45 unsigned int i, q = 0, text_len, consumed = state->offset; in kmp_find() local 56 while (q > 0 && kmp->pattern[q] in kmp_find() 58 q = kmp->prefix_tbl[q - 1]; in kmp_find() 59 if (kmp->pattern[q] in kmp_find() 61 q++; in kmp_find() 62 if (unlikely(q == kmp->pattern_len)) { in kmp_find() 77 unsigned int k, q; in compute_prefix_tbl() local 80 for (k = 0, q = 1; q < len; q++) { in compute_prefix_tbl() 82 != (icase ? toupper(pattern[q]) : pattern[q])) in compute_prefix_tbl() 85 == (icase ? toupper(pattern[q]) : pattern[q])) in compute_prefix_tbl() [all …]
|
D | crc32.c | 62 # define DO_CRC4 (t3[(q) & 255] ^ t2[(q >> 8) & 255] ^ \ 63 t1[(q >> 16) & 255] ^ t0[(q >> 24) & 255]) 64 # define DO_CRC8 (t7[(q) & 255] ^ t6[(q >> 8) & 255] ^ \ 65 t5[(q >> 16) & 255] ^ t4[(q >> 24) & 255]) 68 # define DO_CRC4 (t0[(q) & 255] ^ t1[(q >> 8) & 255] ^ \ 69 t2[(q >> 16) & 255] ^ t3[(q >> 24) & 255]) 70 # define DO_CRC8 (t4[(q) & 255] ^ t5[(q >> 8) & 255] ^ \ 71 t6[(q >> 16) & 255] ^ t7[(q >> 24) & 255]) 82 u32 q; local 106 q = crc ^ *++b; /* use pre increment for speed */ [all …]
|
D | string_helpers.c | 135 char *p = *dst, *q = *src; in unescape_space() local 137 switch (*q) { in unescape_space() 163 char *p = *dst, *q = *src; in unescape_octal() local 166 if (isodigit(*q) == 0) in unescape_octal() 169 num = (*q++) & 7; in unescape_octal() 170 while (num < 32 && isodigit(*q) && (q - *src < 3)) { in unescape_octal() 172 num += (*q++) & 7; in unescape_octal() 176 *src = q; in unescape_octal() 182 char *p = *dst, *q = *src; in unescape_hex() local 186 if (*q++ != 'x') in unescape_hex() [all …]
|
D | vsprintf.c | 187 unsigned q; in put_dec_trunc8() local 194 q = (r * (u64)0x28f5c29) >> 32; in put_dec_trunc8() 195 *((u16 *)buf) = decpair[r - 100*q]; in put_dec_trunc8() 199 if (q < 100) in put_dec_trunc8() 203 r = (q * (u64)0x28f5c29) >> 32; in put_dec_trunc8() 204 *((u16 *)buf) = decpair[q - 100*r]; in put_dec_trunc8() 212 q = (r * 0x147b) >> 19; in put_dec_trunc8() 213 *((u16 *)buf) = decpair[r - 100*q]; in put_dec_trunc8() 217 r = q; in put_dec_trunc8() 229 unsigned q; in put_dec_full8() local [all …]
|
D | inflate.c | 346 register struct huft *q; /* points to current table */ in huft_build() local 449 q = (struct huft *)NULL; /* ditto */ in huft_build() 488 if ((q = (struct huft *)malloc((z + 1)*sizeof(struct huft))) == in huft_build() 498 *t = q + 1; /* link to list for huft_free() */ in huft_build() 499 *(t = &(q->v.t)) = (struct huft *)NULL; in huft_build() 500 u[h] = ++q; /* table starts after link */ in huft_build() 509 r.v.t = q; /* pointer to this table */ in huft_build() 537 q[j] = r; in huft_build() 574 register struct huft *p, *q; in huft_free() local 581 q = (--p)->v.t; in huft_free() [all …]
|
D | test_hexdump.c | 99 const char *q = *result++; in test_hexdump_prepare_test() local 100 size_t amount = strlen(q); in test_hexdump_prepare_test() 102 memcpy(p, q, amount); in test_hexdump_prepare_test()
|
/lib/mpi/ |
D | mpih-div.c | 104 mpi_limb_t q; in mpihelp_divrem() local 116 q = ~(mpi_limb_t) 0; in mpihelp_divrem() 122 qp[i] = q; in mpihelp_divrem() 128 udiv_qrnnd(q, r, n1, n0, d1); in mpihelp_divrem() 129 umul_ppmm(n1, n0, d0, q); in mpihelp_divrem() 136 q--; in mpihelp_divrem() 143 qp[i] = q; in mpihelp_divrem() 171 mpi_limb_t q; in mpihelp_divrem() local 187 q = ~(mpi_limb_t) 0; in mpihelp_divrem() 191 udiv_qrnnd(q, r, n0, np[dsize - 1], dX); in mpihelp_divrem() [all …]
|
D | longlong.h | 144 #define udiv_qrnnd(q, r, n1, n0, d) \ argument 146 : "=r" ((USItype)(q)), \ 162 #define udiv_qrnnd(q, r, n1, n0, d) \ argument 164 (q) = __udiv_qrnnd(&__r, (n1), (n0), (d)); \ 287 #define udiv_qrnnd(q, r, nh, nl, d) \ argument 289 : "=g" ((USItype)(q)), \ 339 #define udiv_qrnnd(q, r, n1, n0, d) \ 341 (q) = __udiv_qrnnd(&__r, (n1), (n0), (d)); \ 379 #define sdiv_qrnnd(q, r, n1, n0, d) \ argument 388 (q) = __xx.__i.__l; (r) = __xx.__i.__h; \ [all …]
|
/lib/math/ |
D | cordic.c | 55 coord.q = 0; in cordic_calc_iq() 73 valtmp = coord.i - (coord.q >> iter); in cordic_calc_iq() 74 coord.q += (coord.i >> iter); in cordic_calc_iq() 77 valtmp = coord.i + (coord.q >> iter); in cordic_calc_iq() 78 coord.q -= (coord.i >> iter); in cordic_calc_iq() 85 coord.q *= signx; in cordic_calc_iq()
|