/lib/mpi/ |
D | mpi-cmp.c | 23 int mpi_cmp_ui(MPI u, unsigned long v) in mpi_cmp_ui() argument 27 mpi_normalize(u); in mpi_cmp_ui() 28 if (u->nlimbs == 0) { in mpi_cmp_ui() 34 if (u->sign) in mpi_cmp_ui() 36 if (u->nlimbs > 1) in mpi_cmp_ui() 39 if (u->d[0] == limb) in mpi_cmp_ui() 41 else if (u->d[0] > limb) in mpi_cmp_ui() 48 static int do_mpi_cmp(MPI u, MPI v, int absmode) in do_mpi_cmp() argument 56 mpi_normalize(u); in do_mpi_cmp() 59 usize = u->nlimbs; in do_mpi_cmp() [all …]
|
D | mpi-sub-ui.c | 37 int mpi_sub_ui(MPI w, MPI u, unsigned long vval) in mpi_sub_ui() argument 39 if (u->nlimbs == 0) { in mpi_sub_ui() 49 if (mpi_resize(w, u->nlimbs + 1)) in mpi_sub_ui() 52 if (u->sign) { in mpi_sub_ui() 55 cy = mpihelp_add_1(w->d, u->d, u->nlimbs, (mpi_limb_t) vval); in mpi_sub_ui() 56 w->d[u->nlimbs] = cy; in mpi_sub_ui() 57 w->nlimbs = u->nlimbs + cy; in mpi_sub_ui() 63 if (u->nlimbs == 1 && u->d[0] < vval) { in mpi_sub_ui() 64 w->d[0] = vval - u->d[0]; in mpi_sub_ui() 68 mpihelp_sub_1(w->d, u->d, u->nlimbs, (mpi_limb_t) vval); in mpi_sub_ui() [all …]
|
D | mpi-inv.c | 33 MPI u, v, u1, u2 = NULL, u3, v1, v2 = NULL, v3, t1, t2 = NULL, t3; in mpi_invm() local 43 u = mpi_copy(a); in mpi_invm() 46 for (k = 0; !mpi_test_bit(u, 0) && !mpi_test_bit(v, 0); k++) { in mpi_invm() 47 mpi_rshift(u, u, 1); in mpi_invm() 55 u3 = mpi_copy(u); in mpi_invm() 58 v2 = mpi_alloc(mpi_get_nlimbs(u)); in mpi_invm() 59 mpi_sub(v2, u1, u); /* U is used as const 1 */ in mpi_invm() 62 if (mpi_test_bit(u, 0)) { /* u is odd */ in mpi_invm() 75 t3 = mpi_copy(u); in mpi_invm() 84 mpi_sub(t2, t2, u); in mpi_invm() [all …]
|
D | mpi-add.c | 20 void mpi_add_ui(MPI w, MPI u, unsigned long v) in mpi_add_ui() argument 26 usize = u->nlimbs; in mpi_add_ui() 27 usign = u->sign; in mpi_add_ui() 36 up = u->d; in mpi_add_ui() 67 void mpi_add(MPI w, MPI u, MPI v) in mpi_add() argument 73 if (u->nlimbs < v->nlimbs) { /* Swap U and V. */ in mpi_add() 76 vsize = u->nlimbs; in mpi_add() 77 vsign = u->sign; in mpi_add() 82 vp = u->d; in mpi_add() 84 usize = u->nlimbs; in mpi_add() [all …]
|
D | mpiutil.c | 235 void mpi_snatch(MPI w, MPI u) in mpi_snatch() argument 238 mpi_assign_limb_space(w, u->d, u->alloced); in mpi_snatch() 239 w->nlimbs = u->nlimbs; in mpi_snatch() 240 w->sign = u->sign; in mpi_snatch() 241 w->flags = u->flags; in mpi_snatch() 242 u->alloced = 0; in mpi_snatch() 243 u->nlimbs = 0; in mpi_snatch() 244 u->d = NULL; in mpi_snatch() 246 mpi_free(u); in mpi_snatch() 250 MPI mpi_set(MPI w, MPI u) in mpi_set() argument [all …]
|
D | mpi-mul.c | 16 void mpi_mul(MPI w, MPI u, MPI v) in mpi_mul() argument 25 if (u->nlimbs < v->nlimbs) { in mpi_mul() 30 vsize = u->nlimbs; in mpi_mul() 31 vsign = u->sign; in mpi_mul() 32 vp = u->d; in mpi_mul() 34 usize = u->nlimbs; in mpi_mul() 35 usign = u->sign; in mpi_mul() 36 up = u->d; in mpi_mul() 86 void mpi_mulm(MPI w, MPI u, MPI v, MPI m) in mpi_mulm() argument 88 mpi_mul(w, u, v); in mpi_mulm()
|
D | longlong.h | 232 #define umul_ppmm(w1, w0, u, v) \ argument 238 : "%0" ((USItype)(u)), \ 241 #define smul_ppmm(w1, w0, u, v) \ argument 247 : "%0" ((SItype)(u)), \ 250 #define __umulsidi3(u, v) \ argument 254 : "%0" ((USItype)(u)), \ 320 #define umul_ppmm(wh, wl, u, v) \ 327 : "*f" ((USItype)(u)), \ 415 #define umul_ppmm(w1, w0, u, v) \ argument 419 : "%0" ((USItype)(u)), \ [all …]
|
D | ec.c | 122 static void ec_addm(MPI w, MPI u, MPI v, struct mpi_ec_ctx *ctx) in ec_addm() argument 124 mpi_add(w, u, v); in ec_addm() 128 static void ec_subm(MPI w, MPI u, MPI v, struct mpi_ec_ctx *ec) in ec_subm() argument 130 mpi_sub(w, u, v); in ec_subm() 136 static void ec_mulm(MPI w, MPI u, MPI v, struct mpi_ec_ctx *ctx) in ec_mulm() argument 138 mpi_mul(w, u, v); in ec_mulm() 143 static void ec_mul2(MPI w, MPI u, struct mpi_ec_ctx *ctx) in ec_mul2() argument 145 mpi_lshift(w, u, 1); in ec_mul2() 199 static void ec_addm_25519(MPI w, MPI u, MPI v, struct mpi_ec_ctx *ctx) in ec_addm_25519() argument 206 if (w->nlimbs != wsize || u->nlimbs != wsize || v->nlimbs != wsize) in ec_addm_25519() [all …]
|
D | mpi-mod.c | 153 void mpi_mul_barrett(MPI w, MPI u, MPI v, mpi_barrett_t ctx) in mpi_mul_barrett() argument 155 mpi_mul(w, u, v); in mpi_mul_barrett()
|
/lib/zstd/ |
D | zstd_opt.h | 40 unsigned u; in ZSTD_rescaleFreqs() local 50 for (u = 0; u <= MaxLit; u++) in ZSTD_rescaleFreqs() 51 ssPtr->litFreq[u] = 0; in ZSTD_rescaleFreqs() 52 for (u = 0; u < srcSize; u++) in ZSTD_rescaleFreqs() 53 ssPtr->litFreq[src[u]]++; in ZSTD_rescaleFreqs() 61 for (u = 0; u <= MaxLit; u++) { in ZSTD_rescaleFreqs() 62 ssPtr->litFreq[u] = 1 + (ssPtr->litFreq[u] >> ZSTD_FREQ_DIV); in ZSTD_rescaleFreqs() 63 ssPtr->litSum += ssPtr->litFreq[u]; in ZSTD_rescaleFreqs() 65 for (u = 0; u <= MaxLL; u++) in ZSTD_rescaleFreqs() 66 ssPtr->litLengthFreq[u] = 1; in ZSTD_rescaleFreqs() [all …]
|
D | fse_compress.c | 126 U32 u; in FSE_buildCTable_wksp() local 128 for (u = 1; u <= maxSymbolValue + 1; u++) { in FSE_buildCTable_wksp() 129 if (normalizedCounter[u - 1] == -1) { /* Low proba symbol */ in FSE_buildCTable_wksp() 130 cumul[u] = cumul[u - 1] + 1; in FSE_buildCTable_wksp() 131 tableSymbol[highThreshold--] = (FSE_FUNCTION_TYPE)(u - 1); in FSE_buildCTable_wksp() 133 cumul[u] = cumul[u - 1] + normalizedCounter[u - 1]; in FSE_buildCTable_wksp() 159 U32 u; in FSE_buildCTable_wksp() local 160 for (u = 0; u < tableSize; u++) { in FSE_buildCTable_wksp() 161 …FSE_FUNCTION_TYPE s = tableSymbol[u]; /* note : static analyzer may not understand tableSymbol is … in FSE_buildCTable_wksp() 162 …tableU16[cumul[s]++] = (U16)(tableSize + u); /* TableU16 : sorted by symbol order; gives next stat… in FSE_buildCTable_wksp()
|
D | fse_decompress.c | 148 U32 u; in FSE_buildDTable_wksp() local 149 for (u = 0; u < tableSize; u++) { in FSE_buildDTable_wksp() 150 FSE_FUNCTION_TYPE const symbol = (FSE_FUNCTION_TYPE)(tableDecode[u].symbol); in FSE_buildDTable_wksp() 152 tableDecode[u].nbBits = (BYTE)(tableLog - BIT_highbit32((U32)nextState)); in FSE_buildDTable_wksp() 153 tableDecode[u].newState = (U16)((nextState << tableDecode[u].nbBits) - tableSize); in FSE_buildDTable_wksp()
|
D | huf_decompress.c | 145 U32 u; in HUF_readDTableX2_wksp() local 149 for (u = rankVal[w]; u < rankVal[w] + length; u++) in HUF_readDTableX2_wksp() 150 dt[u] = D; in HUF_readDTableX2_wksp() 473 U32 u; in HUF_fillDTableX4() local 474 for (u = start; u < end; u++) in HUF_fillDTableX4() 475 DTable[u] = DElt; in HUF_fillDTableX4()
|
D | compress.c | 399 U32 u; in ZSTD_reduceTable() local 400 for (u = 0; u < size; u++) { in ZSTD_reduceTable() 401 if (table[u] < reducerValue) in ZSTD_reduceTable() 402 table[u] = 0; in ZSTD_reduceTable() 404 table[u] -= reducerValue; in ZSTD_reduceTable() 572 U32 u; in ZSTD_seqToCodes() local 573 for (u = 0; u < nbSeq; u++) { in ZSTD_seqToCodes() 574 U32 const llv = sequences[u].litLength; in ZSTD_seqToCodes() 575 U32 const mlv = sequences[u].matchLength; in ZSTD_seqToCodes() 576 llCodeTable[u] = (llv > 63) ? (BYTE)ZSTD_highbit32(llv) + LL_deltaCode : LL_Code[llv]; in ZSTD_seqToCodes() [all …]
|
/lib/ |
D | test_bpf.c | 72 } u; member 101 self->u.ptr.insns = insn; in bpf_fill_maxinsns1() 102 self->u.ptr.len = len; in bpf_fill_maxinsns1() 120 self->u.ptr.insns = insn; in bpf_fill_maxinsns2() 121 self->u.ptr.len = len; in bpf_fill_maxinsns2() 147 self->u.ptr.insns = insn; in bpf_fill_maxinsns3() 148 self->u.ptr.len = len; in bpf_fill_maxinsns3() 166 self->u.ptr.insns = insn; in bpf_fill_maxinsns4() 167 self->u.ptr.len = len; in bpf_fill_maxinsns4() 189 self->u.ptr.insns = insn; in bpf_fill_maxinsns5() [all …]
|
D | muldi3.c | 16 #define umul_ppmm(w1, w0, u, v) \ argument 21 __ul = __ll_lowpart(u); \ 22 __uh = __ll_highpart(u); \ 42 #define __umulsidi3(u, v) ({ \ argument 44 umul_ppmm(__w.s.high, __w.s.low, u, v); \ 49 long long notrace __muldi3(long long u, long long v) in __muldi3() argument 51 const DWunion uu = {.ll = u}; in __muldi3()
|
D | ashldi3.c | 9 long long notrace __ashldi3(long long u, word_type b) in __ashldi3() argument 15 return u; in __ashldi3() 17 uu.ll = u; in __ashldi3()
|
D | lshrdi3.c | 9 long long notrace __lshrdi3(long long u, word_type b) in __lshrdi3() argument 15 return u; in __lshrdi3() 17 uu.ll = u; in __lshrdi3()
|
D | uuid.c | 123 int guid_parse(const char *uuid, guid_t *u) in guid_parse() argument 125 return __uuid_parse(uuid, u->b, guid_index); in guid_parse() 129 int uuid_parse(const char *uuid, uuid_t *u) in uuid_parse() argument 131 return __uuid_parse(uuid, u->b, uuid_index); in uuid_parse()
|
D | ashrdi3.c | 9 long long notrace __ashrdi3(long long u, word_type b) in __ashrdi3() argument 15 return u; in __ashrdi3() 17 uu.ll = u; in __ashrdi3()
|
D | bitfield_kunit.c | 13 u##tp _res; \ 15 _res = u##tp##_encode_bits(v, field); \ 20 u##tp##_get_bits(_res, field) != v); \
|
D | inflate.c | 354 struct huft *u[BMAX]; /* table stack */ in huft_build() member 359 struct huft **u; in huft_build() local 371 u = stk->u; in huft_build() 448 u[0] = (struct huft *)NULL; /* just to keep compilers happy */ in huft_build() 492 huft_free(u[0]); in huft_build() 500 u[h] = ++q; /* table starts after link */ in huft_build() 511 u[h-1][j] = r; /* connect to last table */ in huft_build()
|
D | atomic64.c | 177 s64 atomic64_fetch_add_unless(atomic64_t *v, s64 a, s64 u) in atomic64_fetch_add_unless() argument 185 if (val != u) in atomic64_fetch_add_unless()
|
/lib/reed_solomon/ |
D | decode_rs.c | 23 uint16_t u, q, tmp, num1, num2, den, discr_r, syn_error; variable 114 u = rs_modnn(rs, prim * (nn - 1 - (eras_pos[i] + pad))); 119 alpha_to[rs_modnn(rs, u + tmp)];
|
/lib/raid6/ |
D | neon.uc | 51 static inline unative_t PMUL(unative_t v, unative_t u) 53 return (unative_t)vmulq_p8((poly8x16_t)v, (poly8x16_t)u);
|