/lib/ |
D | atomic64.c | 36 static inline raw_spinlock_t *lock_addr(const atomic64_t *v) in lock_addr() argument 38 unsigned long addr = (unsigned long) v; in lock_addr() 45 s64 atomic64_read(const atomic64_t *v) in atomic64_read() argument 48 raw_spinlock_t *lock = lock_addr(v); in atomic64_read() 52 val = v->counter; in atomic64_read() 58 void atomic64_set(atomic64_t *v, s64 i) in atomic64_set() argument 61 raw_spinlock_t *lock = lock_addr(v); in atomic64_set() 64 v->counter = i; in atomic64_set() 70 void atomic64_##op(s64 a, atomic64_t *v) \ 73 raw_spinlock_t *lock = lock_addr(v); \ [all …]
|
D | iov_iter.c | 77 #define iterate_all_kinds(i, n, v, I, B, K) { \ argument 81 struct bio_vec v; \ 83 iterate_bvec(i, n, v, __bi, skip, (B)) \ 86 struct kvec v; \ 87 iterate_kvec(i, n, v, kvec, skip, (K)) \ 91 struct iovec v; \ 92 iterate_iovec(i, n, v, iov, skip, (I)) \ 97 #define iterate_and_advance(i, n, v, I, B, K) { \ argument 104 struct bio_vec v; \ 106 iterate_bvec(i, n, v, __bi, skip, (B)) \ [all …]
|
D | atomic64_test.c | 22 atomic##bit##_set(&v, v0); \ 24 atomic##bit##_##op(val, &v); \ 26 WARN(atomic##bit##_read(&v) != r, "%Lx != %Lx\n", \ 27 (unsigned long long)atomic##bit##_read(&v), \ 46 atomic##bit##_set(&v, v0); \ 49 BUG_ON(atomic##bit##_##op(val, &v) != r); \ 50 BUG_ON(atomic##bit##_read(&v) != r); \ 55 atomic##bit##_set(&v, v0); \ 58 BUG_ON(atomic##bit##_##op(val, &v) != v0); \ 59 BUG_ON(atomic##bit##_read(&v) != r); \ [all …]
|
D | test_bitfield.c | 12 #define CHECK_ENC_GET_U(tp, v, field, res) do { \ argument 16 _res = u##tp##_encode_bits(v, field); \ 18 pr_warn("u" #tp "_encode_bits(" #v ", " #field ") is 0x%llx != " #res "\n",\ 22 if (u##tp##_get_bits(_res, field) != v) \ 27 #define CHECK_ENC_GET_LE(tp, v, field, res) do { \ argument 31 _res = le##tp##_encode_bits(v, field); \ 33 pr_warn("le" #tp "_encode_bits(" #v ", " #field ") is 0x%llx != 0x%llx\n",\ 38 if (le##tp##_get_bits(_res, field) != v) \ 43 #define CHECK_ENC_GET_BE(tp, v, field, res) do { \ argument 47 _res = be##tp##_encode_bits(v, field); \ [all …]
|
D | win_minmax.c | 63 return m->s[0].v; in minmax_subwin_update() 69 struct minmax_sample val = { .t = t, .v = meas }; in minmax_running_max() 71 if (unlikely(val.v >= m->s[0].v) || /* found new max? */ in minmax_running_max() 75 if (unlikely(val.v >= m->s[1].v)) in minmax_running_max() 77 else if (unlikely(val.v >= m->s[2].v)) in minmax_running_max() 87 struct minmax_sample val = { .t = t, .v = meas }; in minmax_running_min() 89 if (unlikely(val.v <= m->s[0].v) || /* found new min? */ in minmax_running_min() 93 if (unlikely(val.v <= m->s[1].v)) in minmax_running_min() 95 else if (unlikely(val.v <= m->s[2].v)) in minmax_running_min()
|
D | test_string.c | 10 u16 v, *p; in memset16_selftest() local 18 memset(p, 0xa1, 256 * 2 * sizeof(v)); in memset16_selftest() 21 v = p[k]; in memset16_selftest() 23 if (v != 0xa1a1) in memset16_selftest() 26 if (v != 0xb1b2) in memset16_selftest() 29 if (v != 0xa1a1) in memset16_selftest() 46 u32 v, *p; in memset32_selftest() local 54 memset(p, 0xa1, 256 * 2 * sizeof(v)); in memset32_selftest() 57 v = p[k]; in memset32_selftest() 59 if (v != 0xa1a1a1a1) in memset32_selftest() [all …]
|
D | muldi3.c | 16 #define umul_ppmm(w1, w0, u, v) \ argument 23 __vl = __ll_lowpart(v); \ 24 __vh = __ll_highpart(v); \ 42 #define __umulsidi3(u, v) ({ \ argument 44 umul_ppmm(__w.s.high, __w.s.low, u, v); \ 49 long long notrace __muldi3(long long u, long long v) in __muldi3() argument 52 const DWunion vv = {.ll = v}; in __muldi3()
|
D | oid_registry.c | 108 const unsigned char *v = data, *end = v + datasize; in sprint_oid() local 114 if (v >= end) in sprint_oid() 117 n = *v++; in sprint_oid() 124 while (v < end) { in sprint_oid() 126 n = *v++; in sprint_oid() 132 if (v >= end) in sprint_oid() 134 n = *v++; in sprint_oid()
|
D | earlycpio.c | 65 unsigned int ch[C_NFIELDS], *chp, v; in find_cpio_data() local 83 v = 0; in find_cpio_data() 85 v <<= 4; in find_cpio_data() 90 v += x; in find_cpio_data() 96 v += x + 10; in find_cpio_data() 102 *chp++ = v; in find_cpio_data()
|
D | inflate.c | 145 } v; member 233 #define NEXTBYTE() ({ int v = get_byte(); if (v < 0) goto underrun; (uch)v; }) 355 unsigned v[N_MAX]; /* values in order of bit length */ in huft_build() member 358 unsigned *c, *v, *x; in huft_build() local 369 v = stk->v; in huft_build() 437 v[x[j]++] = i; in huft_build() 445 p = v; /* grab values in bit order */ in huft_build() 499 *(t = &(q->v.t)) = (struct huft *)NULL; in huft_build() 509 r.v.t = q; /* pointer to this table */ in huft_build() 519 if (p >= v + n) in huft_build() [all …]
|
D | error-inject.c | 164 static void ei_seq_stop(struct seq_file *m, void *v) in ei_seq_stop() argument 169 static void *ei_seq_next(struct seq_file *m, void *v, loff_t *pos) in ei_seq_next() argument 171 return seq_list_next(v, &error_injection_list, pos); in ei_seq_next() 188 static int ei_seq_show(struct seq_file *m, void *v) in ei_seq_show() argument 190 struct ei_entry *ent = list_entry(v, struct ei_entry, list); in ei_seq_show()
|
D | generic-radix-tree.c | 107 struct genradix_root *v = READ_ONCE(radix->root); in __genradix_ptr_alloc() local 113 struct genradix_root *r = v, *new_root; in __genradix_ptr_alloc() 131 if ((v = cmpxchg_release(&radix->root, r, new_root)) == r) { in __genradix_ptr_alloc() 132 v = new_root; in __genradix_ptr_alloc()
|
D | bch.c | 266 static inline int modulo(struct bch_control *bch, unsigned int v) in modulo() argument 269 while (v >= n) { in modulo() 270 v -= n; in modulo() 271 v = (v & n) + (v >> GF_M(bch)); in modulo() 273 return v; in modulo() 279 static inline int mod_s(struct bch_control *bch, unsigned int v) in mod_s() argument 282 return (v < n) ? v : v-n; in mod_s() 574 unsigned int u, v, r; in find_poly_deg2_roots() local 591 v = u; in find_poly_deg2_roots() 592 while (v) { in find_poly_deg2_roots() [all …]
|
D | idr.c | 497 unsigned long v = xa_to_value(bitmap); in ida_free() local 500 if (!(v & (1UL << bit))) in ida_free() 502 v &= ~(1UL << bit); in ida_free() 503 if (!v) in ida_free() 505 xas_store(&xas, xa_mk_value(v)); in ida_free()
|
D | string.c | 762 void *memset16(uint16_t *s, uint16_t v, size_t count) in memset16() argument 767 *xs++ = v; in memset16() 784 void *memset32(uint32_t *s, uint32_t v, size_t count) in memset32() argument 789 *xs++ = v; in memset32() 806 void *memset64(uint64_t *s, uint64_t v, size_t count) in memset64() argument 811 *xs++ = v; in memset64()
|
/lib/raid6/ |
D | mktables.c | 23 uint8_t v = 0; in gfmul() local 27 v ^= a; in gfmul() 32 return v; in gfmul() 37 uint8_t v = 1; in gfpow() local 45 v = gfmul(v, a); in gfpow() 50 return v; in gfpow() 56 uint8_t v; in main() local 107 v = 1; in main() 113 exptbl[i + j] = v; in main() 114 printf("0x%02x,%c", v, (j == 7) ? '\n' : ' '); in main() [all …]
|
D | neon.uc | 37 static inline unative_t SHLBYTE(unative_t v) 39 return vshlq_n_u8(v, 1); 46 static inline unative_t MASK(unative_t v) 48 return (unative_t)vshrq_n_s8((int8x16_t)v, 7); 51 static inline unative_t PMUL(unative_t v, unative_t u) 53 return (unative_t)vmulq_p8((poly8x16_t)v, (poly8x16_t)u);
|
D | altivec.uc | 50 static inline __attribute_const__ unative_t SHLBYTE(unative_t v) 52 return vec_add(v,v); 59 static inline __attribute_const__ unative_t MASK(unative_t v) 64 return (unative_t)vec_cmpgt(zv, v);
|
D | vpermxor.uc | 64 asm(VPERMXOR(%0,%1,%2,%3):"=v"(wq$$):"v"(gf_high), "v"(gf_low), "v"(wq$$));
|
D | int.uc | 60 static inline __attribute_const__ unative_t SHLBYTE(unative_t v) 64 vv = (v << 1) & NBYTES(0xfe); 72 static inline __attribute_const__ unative_t MASK(unative_t v) 76 vv = v & NBYTES(0x80);
|
/lib/mpi/ |
D | mpi-cmp.c | 23 int mpi_cmp_ui(MPI u, unsigned long v) in mpi_cmp_ui() argument 25 mpi_limb_t limb = v; in mpi_cmp_ui() 44 int mpi_cmp(MPI u, MPI v) in mpi_cmp() argument 50 mpi_normalize(v); in mpi_cmp() 52 vsize = v->nlimbs; in mpi_cmp() 53 if (!u->sign && v->sign) in mpi_cmp() 55 if (u->sign && !v->sign) in mpi_cmp() 57 if (usize != vsize && !u->sign && !v->sign) in mpi_cmp() 59 if (usize != vsize && u->sign && v->sign) in mpi_cmp() 63 cmp = mpihelp_cmp(u->d, v->d, usize); in mpi_cmp()
|
D | longlong.h | 232 #define umul_ppmm(w1, w0, u, v) \ argument 239 "r" ((USItype)(v))); \ 241 #define smul_ppmm(w1, w0, u, v) \ argument 248 "r" ((SItype)(v))); \ 250 #define __umulsidi3(u, v) \ argument 255 "r" ((USItype)(v))); \ 320 #define umul_ppmm(wh, wl, u, v) \ 328 "*f" ((USItype)(v))); \ 415 #define umul_ppmm(w1, w0, u, v) \ argument 420 "rm" ((USItype)(v))) [all …]
|
/lib/lzo/ |
D | lzo1x_compress.c | 167 u64 v; in lzo1x_1_do_compress() local 168 v = get_unaligned((const u64 *) (ip + m_len)) ^ in lzo1x_1_do_compress() 170 if (unlikely(v == 0)) { in lzo1x_1_do_compress() 173 v = get_unaligned((const u64 *) (ip + m_len)) ^ in lzo1x_1_do_compress() 177 } while (v == 0); in lzo1x_1_do_compress() 180 m_len += (unsigned) __builtin_ctzll(v) / 8; in lzo1x_1_do_compress() 182 m_len += (unsigned) __builtin_clzll(v) / 8; in lzo1x_1_do_compress() 187 u32 v; in lzo1x_1_do_compress() 188 v = get_unaligned((const u32 *) (ip + m_len)) ^ in lzo1x_1_do_compress() 190 if (unlikely(v == 0)) { in lzo1x_1_do_compress() [all …]
|
/lib/reed_solomon/ |
D | test_rslib.c | 32 __param(int, v, V_PROGRESS, "Verbosity level"); 326 if (v >= V_PROGRESS) in ex_rs_helper() 333 if (v >= V_CSUMMARY) { in ex_rs_helper() 343 if (retval && v >= V_PROGRESS) in ex_rs_helper() 356 if (v >= V_PROGRESS) in exercise_rs() 414 if (v >= V_PROGRESS) in exercise_rs_bc() 427 if (v >= V_CSUMMARY) { in exercise_rs_bc() 436 if (stat.noncw && v >= V_PROGRESS) in exercise_rs_bc() 470 if (v >= V_PROGRESS) { in run_exercise()
|
/lib/842/ |
D | 842_decompress.c | 133 u64 v; in do_data() local 139 ret = next_bits(p, &v, n * 8); in do_data() 145 put_unaligned(cpu_to_be16((u16)v), (__be16 *)p->out); in do_data() 148 put_unaligned(cpu_to_be32((u32)v), (__be32 *)p->out); in do_data() 151 put_unaligned(cpu_to_be64((u64)v), (__be64 *)p->out); in do_data()
|