Home
last modified time | relevance | path

Searched refs:v (Results 1 – 25 of 31) sorted by relevance

12

/lib/
Datomic64.c36 static inline raw_spinlock_t *lock_addr(const atomic64_t *v) in lock_addr() argument
38 unsigned long addr = (unsigned long) v; in lock_addr()
45 s64 atomic64_read(const atomic64_t *v) in atomic64_read() argument
48 raw_spinlock_t *lock = lock_addr(v); in atomic64_read()
52 val = v->counter; in atomic64_read()
58 void atomic64_set(atomic64_t *v, s64 i) in atomic64_set() argument
61 raw_spinlock_t *lock = lock_addr(v); in atomic64_set()
64 v->counter = i; in atomic64_set()
70 void atomic64_##op(s64 a, atomic64_t *v) \
73 raw_spinlock_t *lock = lock_addr(v); \
[all …]
Diov_iter.c77 #define iterate_all_kinds(i, n, v, I, B, K) { \ argument
81 struct bio_vec v; \
83 iterate_bvec(i, n, v, __bi, skip, (B)) \
86 struct kvec v; \
87 iterate_kvec(i, n, v, kvec, skip, (K)) \
91 struct iovec v; \
92 iterate_iovec(i, n, v, iov, skip, (I)) \
97 #define iterate_and_advance(i, n, v, I, B, K) { \ argument
104 struct bio_vec v; \
106 iterate_bvec(i, n, v, __bi, skip, (B)) \
[all …]
Datomic64_test.c22 atomic##bit##_set(&v, v0); \
24 atomic##bit##_##op(val, &v); \
26 WARN(atomic##bit##_read(&v) != r, "%Lx != %Lx\n", \
27 (unsigned long long)atomic##bit##_read(&v), \
46 atomic##bit##_set(&v, v0); \
49 BUG_ON(atomic##bit##_##op(val, &v) != r); \
50 BUG_ON(atomic##bit##_read(&v) != r); \
55 atomic##bit##_set(&v, v0); \
58 BUG_ON(atomic##bit##_##op(val, &v) != v0); \
59 BUG_ON(atomic##bit##_read(&v) != r); \
[all …]
Dtest_bitfield.c12 #define CHECK_ENC_GET_U(tp, v, field, res) do { \ argument
16 _res = u##tp##_encode_bits(v, field); \
18 pr_warn("u" #tp "_encode_bits(" #v ", " #field ") is 0x%llx != " #res "\n",\
22 if (u##tp##_get_bits(_res, field) != v) \
27 #define CHECK_ENC_GET_LE(tp, v, field, res) do { \ argument
31 _res = le##tp##_encode_bits(v, field); \
33 pr_warn("le" #tp "_encode_bits(" #v ", " #field ") is 0x%llx != 0x%llx\n",\
38 if (le##tp##_get_bits(_res, field) != v) \
43 #define CHECK_ENC_GET_BE(tp, v, field, res) do { \ argument
47 _res = be##tp##_encode_bits(v, field); \
[all …]
Dwin_minmax.c63 return m->s[0].v; in minmax_subwin_update()
69 struct minmax_sample val = { .t = t, .v = meas }; in minmax_running_max()
71 if (unlikely(val.v >= m->s[0].v) || /* found new max? */ in minmax_running_max()
75 if (unlikely(val.v >= m->s[1].v)) in minmax_running_max()
77 else if (unlikely(val.v >= m->s[2].v)) in minmax_running_max()
87 struct minmax_sample val = { .t = t, .v = meas }; in minmax_running_min()
89 if (unlikely(val.v <= m->s[0].v) || /* found new min? */ in minmax_running_min()
93 if (unlikely(val.v <= m->s[1].v)) in minmax_running_min()
95 else if (unlikely(val.v <= m->s[2].v)) in minmax_running_min()
Dtest_string.c10 u16 v, *p; in memset16_selftest() local
18 memset(p, 0xa1, 256 * 2 * sizeof(v)); in memset16_selftest()
21 v = p[k]; in memset16_selftest()
23 if (v != 0xa1a1) in memset16_selftest()
26 if (v != 0xb1b2) in memset16_selftest()
29 if (v != 0xa1a1) in memset16_selftest()
46 u32 v, *p; in memset32_selftest() local
54 memset(p, 0xa1, 256 * 2 * sizeof(v)); in memset32_selftest()
57 v = p[k]; in memset32_selftest()
59 if (v != 0xa1a1a1a1) in memset32_selftest()
[all …]
Dmuldi3.c16 #define umul_ppmm(w1, w0, u, v) \ argument
23 __vl = __ll_lowpart(v); \
24 __vh = __ll_highpart(v); \
42 #define __umulsidi3(u, v) ({ \ argument
44 umul_ppmm(__w.s.high, __w.s.low, u, v); \
49 long long notrace __muldi3(long long u, long long v) in __muldi3() argument
52 const DWunion vv = {.ll = v}; in __muldi3()
Doid_registry.c108 const unsigned char *v = data, *end = v + datasize; in sprint_oid() local
114 if (v >= end) in sprint_oid()
117 n = *v++; in sprint_oid()
124 while (v < end) { in sprint_oid()
126 n = *v++; in sprint_oid()
132 if (v >= end) in sprint_oid()
134 n = *v++; in sprint_oid()
Dearlycpio.c65 unsigned int ch[C_NFIELDS], *chp, v; in find_cpio_data() local
83 v = 0; in find_cpio_data()
85 v <<= 4; in find_cpio_data()
90 v += x; in find_cpio_data()
96 v += x + 10; in find_cpio_data()
102 *chp++ = v; in find_cpio_data()
Dinflate.c145 } v; member
233 #define NEXTBYTE() ({ int v = get_byte(); if (v < 0) goto underrun; (uch)v; })
355 unsigned v[N_MAX]; /* values in order of bit length */ in huft_build() member
358 unsigned *c, *v, *x; in huft_build() local
369 v = stk->v; in huft_build()
437 v[x[j]++] = i; in huft_build()
445 p = v; /* grab values in bit order */ in huft_build()
499 *(t = &(q->v.t)) = (struct huft *)NULL; in huft_build()
509 r.v.t = q; /* pointer to this table */ in huft_build()
519 if (p >= v + n) in huft_build()
[all …]
Derror-inject.c164 static void ei_seq_stop(struct seq_file *m, void *v) in ei_seq_stop() argument
169 static void *ei_seq_next(struct seq_file *m, void *v, loff_t *pos) in ei_seq_next() argument
171 return seq_list_next(v, &error_injection_list, pos); in ei_seq_next()
188 static int ei_seq_show(struct seq_file *m, void *v) in ei_seq_show() argument
190 struct ei_entry *ent = list_entry(v, struct ei_entry, list); in ei_seq_show()
Dgeneric-radix-tree.c107 struct genradix_root *v = READ_ONCE(radix->root); in __genradix_ptr_alloc() local
113 struct genradix_root *r = v, *new_root; in __genradix_ptr_alloc()
131 if ((v = cmpxchg_release(&radix->root, r, new_root)) == r) { in __genradix_ptr_alloc()
132 v = new_root; in __genradix_ptr_alloc()
Dbch.c266 static inline int modulo(struct bch_control *bch, unsigned int v) in modulo() argument
269 while (v >= n) { in modulo()
270 v -= n; in modulo()
271 v = (v & n) + (v >> GF_M(bch)); in modulo()
273 return v; in modulo()
279 static inline int mod_s(struct bch_control *bch, unsigned int v) in mod_s() argument
282 return (v < n) ? v : v-n; in mod_s()
574 unsigned int u, v, r; in find_poly_deg2_roots() local
591 v = u; in find_poly_deg2_roots()
592 while (v) { in find_poly_deg2_roots()
[all …]
Didr.c497 unsigned long v = xa_to_value(bitmap); in ida_free() local
500 if (!(v & (1UL << bit))) in ida_free()
502 v &= ~(1UL << bit); in ida_free()
503 if (!v) in ida_free()
505 xas_store(&xas, xa_mk_value(v)); in ida_free()
Dstring.c762 void *memset16(uint16_t *s, uint16_t v, size_t count) in memset16() argument
767 *xs++ = v; in memset16()
784 void *memset32(uint32_t *s, uint32_t v, size_t count) in memset32() argument
789 *xs++ = v; in memset32()
806 void *memset64(uint64_t *s, uint64_t v, size_t count) in memset64() argument
811 *xs++ = v; in memset64()
/lib/raid6/
Dmktables.c23 uint8_t v = 0; in gfmul() local
27 v ^= a; in gfmul()
32 return v; in gfmul()
37 uint8_t v = 1; in gfpow() local
45 v = gfmul(v, a); in gfpow()
50 return v; in gfpow()
56 uint8_t v; in main() local
107 v = 1; in main()
113 exptbl[i + j] = v; in main()
114 printf("0x%02x,%c", v, (j == 7) ? '\n' : ' '); in main()
[all …]
Dneon.uc37 static inline unative_t SHLBYTE(unative_t v)
39 return vshlq_n_u8(v, 1);
46 static inline unative_t MASK(unative_t v)
48 return (unative_t)vshrq_n_s8((int8x16_t)v, 7);
51 static inline unative_t PMUL(unative_t v, unative_t u)
53 return (unative_t)vmulq_p8((poly8x16_t)v, (poly8x16_t)u);
Daltivec.uc50 static inline __attribute_const__ unative_t SHLBYTE(unative_t v)
52 return vec_add(v,v);
59 static inline __attribute_const__ unative_t MASK(unative_t v)
64 return (unative_t)vec_cmpgt(zv, v);
Dvpermxor.uc64 asm(VPERMXOR(%0,%1,%2,%3):"=v"(wq$$):"v"(gf_high), "v"(gf_low), "v"(wq$$));
Dint.uc60 static inline __attribute_const__ unative_t SHLBYTE(unative_t v)
64 vv = (v << 1) & NBYTES(0xfe);
72 static inline __attribute_const__ unative_t MASK(unative_t v)
76 vv = v & NBYTES(0x80);
/lib/mpi/
Dmpi-cmp.c23 int mpi_cmp_ui(MPI u, unsigned long v) in mpi_cmp_ui() argument
25 mpi_limb_t limb = v; in mpi_cmp_ui()
44 int mpi_cmp(MPI u, MPI v) in mpi_cmp() argument
50 mpi_normalize(v); in mpi_cmp()
52 vsize = v->nlimbs; in mpi_cmp()
53 if (!u->sign && v->sign) in mpi_cmp()
55 if (u->sign && !v->sign) in mpi_cmp()
57 if (usize != vsize && !u->sign && !v->sign) in mpi_cmp()
59 if (usize != vsize && u->sign && v->sign) in mpi_cmp()
63 cmp = mpihelp_cmp(u->d, v->d, usize); in mpi_cmp()
Dlonglong.h232 #define umul_ppmm(w1, w0, u, v) \ argument
239 "r" ((USItype)(v))); \
241 #define smul_ppmm(w1, w0, u, v) \ argument
248 "r" ((SItype)(v))); \
250 #define __umulsidi3(u, v) \ argument
255 "r" ((USItype)(v))); \
320 #define umul_ppmm(wh, wl, u, v) \
328 "*f" ((USItype)(v))); \
415 #define umul_ppmm(w1, w0, u, v) \ argument
420 "rm" ((USItype)(v)))
[all …]
/lib/lzo/
Dlzo1x_compress.c167 u64 v; in lzo1x_1_do_compress() local
168 v = get_unaligned((const u64 *) (ip + m_len)) ^ in lzo1x_1_do_compress()
170 if (unlikely(v == 0)) { in lzo1x_1_do_compress()
173 v = get_unaligned((const u64 *) (ip + m_len)) ^ in lzo1x_1_do_compress()
177 } while (v == 0); in lzo1x_1_do_compress()
180 m_len += (unsigned) __builtin_ctzll(v) / 8; in lzo1x_1_do_compress()
182 m_len += (unsigned) __builtin_clzll(v) / 8; in lzo1x_1_do_compress()
187 u32 v; in lzo1x_1_do_compress()
188 v = get_unaligned((const u32 *) (ip + m_len)) ^ in lzo1x_1_do_compress()
190 if (unlikely(v == 0)) { in lzo1x_1_do_compress()
[all …]
/lib/reed_solomon/
Dtest_rslib.c32 __param(int, v, V_PROGRESS, "Verbosity level");
326 if (v >= V_PROGRESS) in ex_rs_helper()
333 if (v >= V_CSUMMARY) { in ex_rs_helper()
343 if (retval && v >= V_PROGRESS) in ex_rs_helper()
356 if (v >= V_PROGRESS) in exercise_rs()
414 if (v >= V_PROGRESS) in exercise_rs_bc()
427 if (v >= V_CSUMMARY) { in exercise_rs_bc()
436 if (stat.noncw && v >= V_PROGRESS) in exercise_rs_bc()
470 if (v >= V_PROGRESS) { in run_exercise()
/lib/842/
D842_decompress.c133 u64 v; in do_data() local
139 ret = next_bits(p, &v, n * 8); in do_data()
145 put_unaligned(cpu_to_be16((u16)v), (__be16 *)p->out); in do_data()
148 put_unaligned(cpu_to_be32((u32)v), (__be32 *)p->out); in do_data()
151 put_unaligned(cpu_to_be64((u64)v), (__be64 *)p->out); in do_data()

12