Home
last modified time | relevance | path

Searched refs:size (Results 1 – 25 of 112) sorted by relevance

12345

/lib/
Dusercopy_kunit.c38 size_t size; member
41 static bool is_zeroed(void *from, size_t size) in is_zeroed() argument
43 return memchr_inv(from, 0x0, size) == NULL; in is_zeroed()
53 size_t size = priv->size; in usercopy_test_check_nonzero_user() local
55 KUNIT_ASSERT_GE_MSG(test, size, 2 * PAGE_SIZE, "buffer too small"); in usercopy_test_check_nonzero_user()
63 size = 1024; in usercopy_test_check_nonzero_user()
64 start = PAGE_SIZE - (size / 2); in usercopy_test_check_nonzero_user()
69 zero_start = size / 4; in usercopy_test_check_nonzero_user()
70 zero_end = size - zero_start; in usercopy_test_check_nonzero_user()
83 memset(kmem, 0x0, size); in usercopy_test_check_nonzero_user()
[all …]
Dtest_meminit.c29 static int __init count_nonzero_bytes(void *ptr, size_t size) in count_nonzero_bytes() argument
34 for (i = 0; i < size; i++) in count_nonzero_bytes()
41 static void __init fill_with_garbage_skip(void *ptr, int size, size_t skip) in fill_with_garbage_skip() argument
46 WARN_ON(skip > size); in fill_with_garbage_skip()
47 size -= skip; in fill_with_garbage_skip()
49 while (size >= sizeof(*p)) { in fill_with_garbage_skip()
52 size -= sizeof(*p); in fill_with_garbage_skip()
54 if (size) in fill_with_garbage_skip()
55 memset(&p[i], GARBAGE_BYTE, size); in fill_with_garbage_skip()
58 static void __init fill_with_garbage(void *ptr, size_t size) in fill_with_garbage() argument
[all …]
Dsort.c31 static bool is_aligned(const void *base, size_t size, unsigned char align) in is_aligned() argument
33 unsigned char lsbits = (unsigned char)size; in is_aligned()
137 static void do_swap(void *a, void *b, size_t size, swap_r_func_t swap_func, const void *priv) in do_swap() argument
140 ((const struct wrapper *)priv)->swap(a, b, (int)size); in do_swap()
145 swap_words_64(a, b, size); in do_swap()
147 swap_words_32(a, b, size); in do_swap()
149 swap_bytes(a, b, size); in do_swap()
151 swap_func(a, b, (int)size, priv); in do_swap()
182 static size_t parent(size_t i, unsigned int lsbit, size_t size) in parent() argument
184 i -= size; in parent()
[all …]
Dfind_bit.c28 #define FIND_FIRST_BIT(FETCH, MUNGE, size) \ argument
30 unsigned long idx, val, sz = (size); \
50 #define FIND_NEXT_BIT(FETCH, MUNGE, size, start) \ argument
52 unsigned long mask, idx, tmp, sz = (size), __start = (start); \
71 #define FIND_NTH_BIT(FETCH, size, num) \ argument
73 unsigned long sz = (size), nr = (num), idx, w, tmp; \
99 unsigned long _find_first_bit(const unsigned long *addr, unsigned long size) in _find_first_bit() argument
101 return FIND_FIRST_BIT(addr[idx], /* nop */, size); in _find_first_bit()
112 unsigned long size) in _find_first_and_bit() argument
114 return FIND_FIRST_BIT(addr1[idx] & addr2[idx], /* nop */, size); in _find_first_and_bit()
[all …]
Dgenalloc.c88 const unsigned long size = start + nr; in bitmap_set_ll() local
101 mask_to_set &= BITMAP_LAST_WORD_MASK(size); in bitmap_set_ll()
124 const unsigned long size = start + nr; in bitmap_clear_ll() local
137 mask_to_clear &= BITMAP_LAST_WORD_MASK(size); in bitmap_clear_ll()
185 size_t size, int nid, void *owner) in gen_pool_add_owner() argument
188 unsigned long nbits = size >> pool->min_alloc_order; in gen_pool_add_owner()
198 chunk->end_addr = virt + size - 1; in gen_pool_add_owner()
200 atomic_long_set(&chunk->avail, size); in gen_pool_add_owner()
277 unsigned long gen_pool_alloc_algo_owner(struct gen_pool *pool, size_t size, in gen_pool_alloc_algo_owner() argument
292 if (size == 0) in gen_pool_alloc_algo_owner()
[all …]
Ddevres.c30 resource_size_t size, in __devm_ioremap() argument
42 addr = ioremap(offset, size); in __devm_ioremap()
45 addr = ioremap_uc(offset, size); in __devm_ioremap()
48 addr = ioremap_wc(offset, size); in __devm_ioremap()
51 addr = ioremap_np(offset, size); in __devm_ioremap()
73 resource_size_t size) in devm_ioremap() argument
75 return __devm_ioremap(dev, offset, size, DEVM_IOREMAP); in devm_ioremap()
88 resource_size_t size) in devm_ioremap_uc() argument
90 return __devm_ioremap(dev, offset, size, DEVM_IOREMAP_UC); in devm_ioremap_uc()
103 resource_size_t size) in devm_ioremap_wc() argument
[all …]
Ddecompress_unlz4.c43 long size = in_len; in unlz4() local
84 size = fill(inp, 4); in unlz4()
85 if (size < 4) { in unlz4()
95 size -= 4; in unlz4()
108 size = fill(inp, 4); in unlz4()
109 if (size == 0) in unlz4()
111 if (size < 4) { in unlz4()
115 } else if (size < 4) { in unlz4()
124 size -= 4; in unlz4()
141 size -= 4; in unlz4()
[all …]
Dkunit_iov_iter.c82 size_t size = 0; in iov_kunit_load_kvec() local
92 size += pr->to - pr->from; in iov_kunit_load_kvec()
94 KUNIT_ASSERT_LE(test, size, bufsize); in iov_kunit_load_kvec()
96 iov_iter_kvec(iter, dir, kvec, i, size); in iov_kunit_load_kvec()
109 size_t bufsize, npages, size, copied; in iov_kunit_copy_to_kvec() local
124 size = iter.count; in iov_kunit_copy_to_kvec()
126 copied = copy_to_iter(scratch, size, &iter); in iov_kunit_copy_to_kvec()
128 KUNIT_EXPECT_EQ(test, copied, size); in iov_kunit_copy_to_kvec()
159 size_t bufsize, npages, size, copied; in iov_kunit_copy_from_kvec() local
174 size = min(iter.count, bufsize); in iov_kunit_copy_from_kvec()
[all …]
Dlogic_iomem.c72 static void __iomem *real_ioremap(phys_addr_t offset, size_t size) in real_ioremap() argument
75 (unsigned long long)offset, size); in real_ioremap()
86 void __iomem *ioremap(phys_addr_t offset, size_t size) in ioremap() argument
96 if (rreg->res->end < offset + size - 1) in ioremap()
112 size, &mapped_areas[i].ops, in ioremap()
131 return real_ioremap(offset, size); in ioremap()
195 static void real_memset_io(volatile void __iomem *addr, int value, size_t size) in real_memset_io() argument
202 size_t size) in real_memcpy_fromio() argument
207 memset(buffer, 0xff, size); in real_memcpy_fromio()
211 size_t size) in real_memcpy_toio() argument
[all …]
Dusercopy.c45 int check_zeroed_user(const void __user *from, size_t size) in check_zeroed_user() argument
50 if (unlikely(size == 0)) in check_zeroed_user()
54 size += align; in check_zeroed_user()
56 if (!user_read_access_begin(from, size)) in check_zeroed_user()
63 while (size > sizeof(unsigned long)) { in check_zeroed_user()
68 size -= sizeof(unsigned long); in check_zeroed_user()
73 if (size < sizeof(unsigned long)) in check_zeroed_user()
74 val &= aligned_byte_mask(size); in check_zeroed_user()
Diov_iter.c90 size_t fault_in_iov_iter_readable(const struct iov_iter *i, size_t size) in fault_in_iov_iter_readable() argument
93 size_t n = min(size, iov_iter_count(i)); in fault_in_iov_iter_readable()
95 return size - n; in fault_in_iov_iter_readable()
97 size_t count = min(size, iov_iter_count(i)); in fault_in_iov_iter_readable()
101 size -= count; in fault_in_iov_iter_readable()
113 return count + size; in fault_in_iov_iter_readable()
133 size_t fault_in_iov_iter_writeable(const struct iov_iter *i, size_t size) in fault_in_iov_iter_writeable() argument
136 size_t n = min(size, iov_iter_count(i)); in fault_in_iov_iter_writeable()
138 return size - n; in fault_in_iov_iter_writeable()
140 size_t count = min(size, iov_iter_count(i)); in fault_in_iov_iter_writeable()
[all …]
Ddecompress_unxz.c164 #define kmalloc(size, flags) malloc(size) argument
166 #define vmalloc(size) malloc(size) argument
184 static bool memeq(const void *a, const void *b, size_t size) in memeq() argument
190 for (i = 0; i < size; ++i) in memeq()
199 static void memzero(void *buf, size_t size) in memzero() argument
202 uint8_t *e = b + size; in memzero()
211 void *memmove(void *dest, const void *src, size_t size) in memmove() argument
218 for (i = 0; i < size; ++i) in memmove()
221 i = size; in memmove()
259 long (*fill)(void *dest, unsigned long size), in unxz() argument
[all …]
Dseq_buf.c40 return s->len + len <= s->size; in seq_buf_can_fit()
71 WARN_ON(s->size == 0); in seq_buf_vprintf()
73 if (s->len < s->size) { in seq_buf_vprintf()
74 len = vsnprintf(s->buffer + s->len, s->size - s->len, fmt, args); in seq_buf_vprintf()
75 if (s->len + len < s->size) { in seq_buf_vprintf()
119 if (s->size == 0 || s->len == 0) in seq_buf_do_printk()
159 WARN_ON(s->size == 0); in seq_buf_bprintf()
161 if (s->len < s->size) { in seq_buf_bprintf()
163 if (s->len + ret < s->size) { in seq_buf_bprintf()
186 WARN_ON(s->size == 0); in seq_buf_puts()
[all …]
Dkfifo.c25 int __kfifo_alloc(struct __kfifo *fifo, unsigned int size, in __kfifo_alloc() argument
32 size = roundup_pow_of_two(size); in __kfifo_alloc()
38 if (size < 2) { in __kfifo_alloc()
44 fifo->data = kmalloc_array(esize, size, gfp_mask); in __kfifo_alloc()
50 fifo->mask = size - 1; in __kfifo_alloc()
68 unsigned int size, size_t esize) in __kfifo_init() argument
70 size /= esize; in __kfifo_init()
72 if (!is_power_of_2(size)) in __kfifo_init()
73 size = rounddown_pow_of_two(size); in __kfifo_init()
80 if (size < 2) { in __kfifo_init()
[all …]
Dlogic_pio.c42 if (!new_range || !new_range->fwnode || !new_range->size || in logic_pio_register_range()
47 end = new_range->hw_start + new_range->size; in logic_pio_register_range()
59 if (start >= range->hw_start + range->size || in logic_pio_register_range()
61 mmio_end = range->io_start + range->size; in logic_pio_register_range()
68 iio_sz += range->size; in logic_pio_register_range()
74 if (mmio_end + new_range->size - 1 > MMIO_UPPER_LIMIT) { in logic_pio_register_range()
80 new_range->size = SZ_64K; in logic_pio_register_range()
85 if (iio_sz + new_range->size - 1 > IO_SPACE_LIMIT) { in logic_pio_register_range()
148 if (in_range(pio, range->io_start, range->size)) { in find_io_range()
190 resource_size_t addr, resource_size_t size) in logic_pio_trans_hwaddr() argument
[all …]
Drhashtable.c75 static void nested_table_free(union nested_table *ntbl, unsigned int size) in nested_table_free() argument
85 if (size > len) { in nested_table_free()
86 size >>= shift; in nested_table_free()
88 nested_table_free(ntbl + i, size); in nested_table_free()
96 unsigned int size = tbl->size >> tbl->nest; in nested_bucket_table_free() local
104 nested_table_free(ntbl + i, size); in nested_bucket_table_free()
154 size_t size; in nested_bucket_table_alloc() local
159 size = sizeof(*tbl) + sizeof(tbl->buckets[0]); in nested_bucket_table_alloc()
162 kmalloc_noprof(size, gfp|__GFP_ZERO)); in nested_bucket_table_alloc()
182 size_t size; in bucket_table_alloc() local
[all …]
Dbucket_locks.c19 unsigned int i, size; in __alloc_bucket_spinlocks() local
28 size = min_t(unsigned int, nr_pcpus * cpu_mult, max_size); in __alloc_bucket_spinlocks()
30 size = max_size; in __alloc_bucket_spinlocks()
34 tlocks = kvmalloc_array(size, sizeof(spinlock_t), gfp); in __alloc_bucket_spinlocks()
37 for (i = 0; i < size; i++) { in __alloc_bucket_spinlocks()
44 *locks_mask = size - 1; in __alloc_bucket_spinlocks()
Dtest_static_keys.c55 static void invert_keys(struct test_key *keys, int size) in invert_keys() argument
60 for (i = 0; i < size; i++) { in invert_keys()
68 static int verify_keys(struct test_key *keys, int size, bool invert) in verify_keys() argument
73 for (i = 0; i < size; i++) { in verify_keys()
112 int size; in test_key_func() local
211 size = ARRAY_SIZE(static_key_tests); in test_key_func()
213 ret = verify_keys(static_key_tests, size, false); in test_key_func()
217 invert_keys(static_key_tests, size); in test_key_func()
218 ret = verify_keys(static_key_tests, size, true); in test_key_func()
222 invert_keys(static_key_tests, size); in test_key_func()
[all …]
Dstring_helpers.c39 int string_get_size(u64 size, u64 blk_size, const enum string_size_units units, in string_get_size() argument
66 size = 0; in string_get_size()
67 if (size == 0) in string_get_size()
86 while (size >> 32) { in string_get_size()
87 do_div(size, divisor[units_base]); in string_get_size()
93 size *= blk_size; in string_get_size()
96 while (size >= divisor[units_base]) { in string_get_size()
97 remainder = do_div(size, divisor[units_base]); in string_get_size()
103 sf_cap = size; in string_get_size()
120 size += 1; in string_get_size()
[all …]
Dbuildid.c163 static int parse_build_id(struct freader *r, unsigned char *build_id, __u32 *size, in parse_build_id() argument
202 if (size) in parse_build_id()
203 *size = desc_sz; in parse_build_id()
214 static int get_build_id_32(struct freader *r, unsigned char *build_id, __u32 *size) in get_build_id_32() argument
242 !parse_build_id(r, build_id, size, READ_ONCE(phdr->p_offset), in get_build_id_32()
250 static int get_build_id_64(struct freader *r, unsigned char *build_id, __u32 *size) in get_build_id_64() argument
279 !parse_build_id(r, build_id, size, READ_ONCE(phdr->p_offset), in get_build_id_64()
291 __u32 *size, bool may_fault) in __build_id_parse() argument
322 ret = get_build_id_32(&r, build_id, size); in __build_id_parse()
324 ret = get_build_id_64(&r, build_id, size); in __build_id_parse()
[all …]
Dsg_split.c26 size_t size = sizes[0], len; in sg_calculate_split() local
42 len = min_t(size_t, size, sglen - skip); in sg_calculate_split()
47 size -= len; in sg_calculate_split()
51 while (!size && (skip + len < sglen) && (--nb_splits > 0)) { in sg_calculate_split()
53 size = *(++sizes); in sg_calculate_split()
55 len = min_t(size_t, size, sglen - skip); in sg_calculate_split()
61 size -= len; in sg_calculate_split()
65 if (!size && --nb_splits > 0) { in sg_calculate_split()
67 size = *(++sizes); in sg_calculate_split()
74 return (size || !splitters[0].in_sg0) ? -EINVAL : 0; in sg_calculate_split()
/lib/crypto/mpi/
Dmpih-mul.c21 #define MPN_MUL_N_RECURSE(prodp, up, vp, size, tspace) \ argument
23 if ((size) < KARATSUBA_THRESHOLD) \
24 mul_n_basecase(prodp, up, vp, size); \
26 mul_n(prodp, up, vp, size, tspace); \
29 #define MPN_SQR_N_RECURSE(prodp, up, size, tspace) \ argument
31 if ((size) < KARATSUBA_THRESHOLD) \
32 mpih_sqr_n_basecase(prodp, up, size); \
34 mpih_sqr_n(prodp, up, size, tspace); \
55 mul_n_basecase(mpi_ptr_t prodp, mpi_ptr_t up, mpi_ptr_t vp, mpi_size_t size) in mul_n_basecase() argument
66 MPN_COPY(prodp, up, size); in mul_n_basecase()
[all …]
/lib/xz/
Dxz_dec_bcj.c61 size_t size; member
89 static size_t bcj_x86(struct xz_dec_bcj *s, uint8_t *buf, size_t size) in bcj_x86() argument
104 if (size <= 4) in bcj_x86()
107 size -= 4; in bcj_x86()
108 for (i = 0; i < size; ++i) { in bcj_x86()
161 static size_t bcj_powerpc(struct xz_dec_bcj *s, uint8_t *buf, size_t size) in bcj_powerpc() argument
166 size &= ~(size_t)3; in bcj_powerpc()
168 for (i = 0; i < size; i += 4) { in bcj_powerpc()
184 static size_t bcj_ia64(struct xz_dec_bcj *s, uint8_t *buf, size_t size) in bcj_ia64() argument
225 size &= ~(size_t)15; in bcj_ia64()
[all …]
/lib/zstd/common/
Dzstd_common.c56 void* ZSTD_customMalloc(size_t size, ZSTD_customMem customMem) in ZSTD_customMalloc() argument
59 return customMem.customAlloc(customMem.opaque, size); in ZSTD_customMalloc()
60 return ZSTD_malloc(size); in ZSTD_customMalloc()
63 void* ZSTD_customCalloc(size_t size, ZSTD_customMem customMem) in ZSTD_customCalloc() argument
68 void* const ptr = customMem.customAlloc(customMem.opaque, size); in ZSTD_customCalloc()
69 ZSTD_memset(ptr, 0, size); in ZSTD_customCalloc()
72 return ZSTD_calloc(1, size); in ZSTD_customCalloc()
/lib/crypto/
Dmemneq.c68 __crypto_memneq_generic(const void *a, const void *b, size_t size) in __crypto_memneq_generic() argument
73 while (size >= sizeof(unsigned long)) { in __crypto_memneq_generic()
79 size -= sizeof(unsigned long); in __crypto_memneq_generic()
82 while (size > 0) { in __crypto_memneq_generic()
87 size -= 1; in __crypto_memneq_generic()
164 size_t size) in __crypto_memneq() argument
166 switch (size) { in __crypto_memneq()
170 return __crypto_memneq_generic(a, b, size); in __crypto_memneq()

12345