| /lib/ |
| D | usercopy_kunit.c | 38 size_t size; member 41 static bool is_zeroed(void *from, size_t size) in is_zeroed() argument 43 return memchr_inv(from, 0x0, size) == NULL; in is_zeroed() 53 size_t size = priv->size; in usercopy_test_check_nonzero_user() local 55 KUNIT_ASSERT_GE_MSG(test, size, 2 * PAGE_SIZE, "buffer too small"); in usercopy_test_check_nonzero_user() 63 size = 1024; in usercopy_test_check_nonzero_user() 64 start = PAGE_SIZE - (size / 2); in usercopy_test_check_nonzero_user() 69 zero_start = size / 4; in usercopy_test_check_nonzero_user() 70 zero_end = size - zero_start; in usercopy_test_check_nonzero_user() 83 memset(kmem, 0x0, size); in usercopy_test_check_nonzero_user() [all …]
|
| D | test_meminit.c | 29 static int __init count_nonzero_bytes(void *ptr, size_t size) in count_nonzero_bytes() argument 34 for (i = 0; i < size; i++) in count_nonzero_bytes() 41 static void __init fill_with_garbage_skip(void *ptr, int size, size_t skip) in fill_with_garbage_skip() argument 46 WARN_ON(skip > size); in fill_with_garbage_skip() 47 size -= skip; in fill_with_garbage_skip() 49 while (size >= sizeof(*p)) { in fill_with_garbage_skip() 52 size -= sizeof(*p); in fill_with_garbage_skip() 54 if (size) in fill_with_garbage_skip() 55 memset(&p[i], GARBAGE_BYTE, size); in fill_with_garbage_skip() 58 static void __init fill_with_garbage(void *ptr, size_t size) in fill_with_garbage() argument [all …]
|
| D | sort.c | 31 static bool is_aligned(const void *base, size_t size, unsigned char align) in is_aligned() argument 33 unsigned char lsbits = (unsigned char)size; in is_aligned() 137 static void do_swap(void *a, void *b, size_t size, swap_r_func_t swap_func, const void *priv) in do_swap() argument 140 ((const struct wrapper *)priv)->swap(a, b, (int)size); in do_swap() 145 swap_words_64(a, b, size); in do_swap() 147 swap_words_32(a, b, size); in do_swap() 149 swap_bytes(a, b, size); in do_swap() 151 swap_func(a, b, (int)size, priv); in do_swap() 182 static size_t parent(size_t i, unsigned int lsbit, size_t size) in parent() argument 184 i -= size; in parent() [all …]
|
| D | find_bit.c | 28 #define FIND_FIRST_BIT(FETCH, MUNGE, size) \ argument 30 unsigned long idx, val, sz = (size); \ 50 #define FIND_NEXT_BIT(FETCH, MUNGE, size, start) \ argument 52 unsigned long mask, idx, tmp, sz = (size), __start = (start); \ 71 #define FIND_NTH_BIT(FETCH, size, num) \ argument 73 unsigned long sz = (size), nr = (num), idx, w, tmp; \ 99 unsigned long _find_first_bit(const unsigned long *addr, unsigned long size) in _find_first_bit() argument 101 return FIND_FIRST_BIT(addr[idx], /* nop */, size); in _find_first_bit() 112 unsigned long size) in _find_first_and_bit() argument 114 return FIND_FIRST_BIT(addr1[idx] & addr2[idx], /* nop */, size); in _find_first_and_bit() [all …]
|
| D | genalloc.c | 88 const unsigned long size = start + nr; in bitmap_set_ll() local 101 mask_to_set &= BITMAP_LAST_WORD_MASK(size); in bitmap_set_ll() 124 const unsigned long size = start + nr; in bitmap_clear_ll() local 137 mask_to_clear &= BITMAP_LAST_WORD_MASK(size); in bitmap_clear_ll() 185 size_t size, int nid, void *owner) in gen_pool_add_owner() argument 188 unsigned long nbits = size >> pool->min_alloc_order; in gen_pool_add_owner() 198 chunk->end_addr = virt + size - 1; in gen_pool_add_owner() 200 atomic_long_set(&chunk->avail, size); in gen_pool_add_owner() 277 unsigned long gen_pool_alloc_algo_owner(struct gen_pool *pool, size_t size, in gen_pool_alloc_algo_owner() argument 292 if (size == 0) in gen_pool_alloc_algo_owner() [all …]
|
| D | devres.c | 30 resource_size_t size, in __devm_ioremap() argument 42 addr = ioremap(offset, size); in __devm_ioremap() 45 addr = ioremap_uc(offset, size); in __devm_ioremap() 48 addr = ioremap_wc(offset, size); in __devm_ioremap() 51 addr = ioremap_np(offset, size); in __devm_ioremap() 73 resource_size_t size) in devm_ioremap() argument 75 return __devm_ioremap(dev, offset, size, DEVM_IOREMAP); in devm_ioremap() 88 resource_size_t size) in devm_ioremap_uc() argument 90 return __devm_ioremap(dev, offset, size, DEVM_IOREMAP_UC); in devm_ioremap_uc() 103 resource_size_t size) in devm_ioremap_wc() argument [all …]
|
| D | decompress_unlz4.c | 43 long size = in_len; in unlz4() local 84 size = fill(inp, 4); in unlz4() 85 if (size < 4) { in unlz4() 95 size -= 4; in unlz4() 108 size = fill(inp, 4); in unlz4() 109 if (size == 0) in unlz4() 111 if (size < 4) { in unlz4() 115 } else if (size < 4) { in unlz4() 124 size -= 4; in unlz4() 141 size -= 4; in unlz4() [all …]
|
| D | kunit_iov_iter.c | 82 size_t size = 0; in iov_kunit_load_kvec() local 92 size += pr->to - pr->from; in iov_kunit_load_kvec() 94 KUNIT_ASSERT_LE(test, size, bufsize); in iov_kunit_load_kvec() 96 iov_iter_kvec(iter, dir, kvec, i, size); in iov_kunit_load_kvec() 109 size_t bufsize, npages, size, copied; in iov_kunit_copy_to_kvec() local 124 size = iter.count; in iov_kunit_copy_to_kvec() 126 copied = copy_to_iter(scratch, size, &iter); in iov_kunit_copy_to_kvec() 128 KUNIT_EXPECT_EQ(test, copied, size); in iov_kunit_copy_to_kvec() 159 size_t bufsize, npages, size, copied; in iov_kunit_copy_from_kvec() local 174 size = min(iter.count, bufsize); in iov_kunit_copy_from_kvec() [all …]
|
| D | logic_iomem.c | 72 static void __iomem *real_ioremap(phys_addr_t offset, size_t size) in real_ioremap() argument 75 (unsigned long long)offset, size); in real_ioremap() 86 void __iomem *ioremap(phys_addr_t offset, size_t size) in ioremap() argument 96 if (rreg->res->end < offset + size - 1) in ioremap() 112 size, &mapped_areas[i].ops, in ioremap() 131 return real_ioremap(offset, size); in ioremap() 195 static void real_memset_io(volatile void __iomem *addr, int value, size_t size) in real_memset_io() argument 202 size_t size) in real_memcpy_fromio() argument 207 memset(buffer, 0xff, size); in real_memcpy_fromio() 211 size_t size) in real_memcpy_toio() argument [all …]
|
| D | usercopy.c | 45 int check_zeroed_user(const void __user *from, size_t size) in check_zeroed_user() argument 50 if (unlikely(size == 0)) in check_zeroed_user() 54 size += align; in check_zeroed_user() 56 if (!user_read_access_begin(from, size)) in check_zeroed_user() 63 while (size > sizeof(unsigned long)) { in check_zeroed_user() 68 size -= sizeof(unsigned long); in check_zeroed_user() 73 if (size < sizeof(unsigned long)) in check_zeroed_user() 74 val &= aligned_byte_mask(size); in check_zeroed_user()
|
| D | iov_iter.c | 90 size_t fault_in_iov_iter_readable(const struct iov_iter *i, size_t size) in fault_in_iov_iter_readable() argument 93 size_t n = min(size, iov_iter_count(i)); in fault_in_iov_iter_readable() 95 return size - n; in fault_in_iov_iter_readable() 97 size_t count = min(size, iov_iter_count(i)); in fault_in_iov_iter_readable() 101 size -= count; in fault_in_iov_iter_readable() 113 return count + size; in fault_in_iov_iter_readable() 133 size_t fault_in_iov_iter_writeable(const struct iov_iter *i, size_t size) in fault_in_iov_iter_writeable() argument 136 size_t n = min(size, iov_iter_count(i)); in fault_in_iov_iter_writeable() 138 return size - n; in fault_in_iov_iter_writeable() 140 size_t count = min(size, iov_iter_count(i)); in fault_in_iov_iter_writeable() [all …]
|
| D | decompress_unxz.c | 164 #define kmalloc(size, flags) malloc(size) argument 166 #define vmalloc(size) malloc(size) argument 184 static bool memeq(const void *a, const void *b, size_t size) in memeq() argument 190 for (i = 0; i < size; ++i) in memeq() 199 static void memzero(void *buf, size_t size) in memzero() argument 202 uint8_t *e = b + size; in memzero() 211 void *memmove(void *dest, const void *src, size_t size) in memmove() argument 218 for (i = 0; i < size; ++i) in memmove() 221 i = size; in memmove() 259 long (*fill)(void *dest, unsigned long size), in unxz() argument [all …]
|
| D | seq_buf.c | 40 return s->len + len <= s->size; in seq_buf_can_fit() 71 WARN_ON(s->size == 0); in seq_buf_vprintf() 73 if (s->len < s->size) { in seq_buf_vprintf() 74 len = vsnprintf(s->buffer + s->len, s->size - s->len, fmt, args); in seq_buf_vprintf() 75 if (s->len + len < s->size) { in seq_buf_vprintf() 119 if (s->size == 0 || s->len == 0) in seq_buf_do_printk() 159 WARN_ON(s->size == 0); in seq_buf_bprintf() 161 if (s->len < s->size) { in seq_buf_bprintf() 163 if (s->len + ret < s->size) { in seq_buf_bprintf() 186 WARN_ON(s->size == 0); in seq_buf_puts() [all …]
|
| D | kfifo.c | 25 int __kfifo_alloc(struct __kfifo *fifo, unsigned int size, in __kfifo_alloc() argument 32 size = roundup_pow_of_two(size); in __kfifo_alloc() 38 if (size < 2) { in __kfifo_alloc() 44 fifo->data = kmalloc_array(esize, size, gfp_mask); in __kfifo_alloc() 50 fifo->mask = size - 1; in __kfifo_alloc() 68 unsigned int size, size_t esize) in __kfifo_init() argument 70 size /= esize; in __kfifo_init() 72 if (!is_power_of_2(size)) in __kfifo_init() 73 size = rounddown_pow_of_two(size); in __kfifo_init() 80 if (size < 2) { in __kfifo_init() [all …]
|
| D | logic_pio.c | 42 if (!new_range || !new_range->fwnode || !new_range->size || in logic_pio_register_range() 47 end = new_range->hw_start + new_range->size; in logic_pio_register_range() 59 if (start >= range->hw_start + range->size || in logic_pio_register_range() 61 mmio_end = range->io_start + range->size; in logic_pio_register_range() 68 iio_sz += range->size; in logic_pio_register_range() 74 if (mmio_end + new_range->size - 1 > MMIO_UPPER_LIMIT) { in logic_pio_register_range() 80 new_range->size = SZ_64K; in logic_pio_register_range() 85 if (iio_sz + new_range->size - 1 > IO_SPACE_LIMIT) { in logic_pio_register_range() 148 if (in_range(pio, range->io_start, range->size)) { in find_io_range() 190 resource_size_t addr, resource_size_t size) in logic_pio_trans_hwaddr() argument [all …]
|
| D | rhashtable.c | 75 static void nested_table_free(union nested_table *ntbl, unsigned int size) in nested_table_free() argument 85 if (size > len) { in nested_table_free() 86 size >>= shift; in nested_table_free() 88 nested_table_free(ntbl + i, size); in nested_table_free() 96 unsigned int size = tbl->size >> tbl->nest; in nested_bucket_table_free() local 104 nested_table_free(ntbl + i, size); in nested_bucket_table_free() 154 size_t size; in nested_bucket_table_alloc() local 159 size = sizeof(*tbl) + sizeof(tbl->buckets[0]); in nested_bucket_table_alloc() 162 kmalloc_noprof(size, gfp|__GFP_ZERO)); in nested_bucket_table_alloc() 182 size_t size; in bucket_table_alloc() local [all …]
|
| D | bucket_locks.c | 19 unsigned int i, size; in __alloc_bucket_spinlocks() local 28 size = min_t(unsigned int, nr_pcpus * cpu_mult, max_size); in __alloc_bucket_spinlocks() 30 size = max_size; in __alloc_bucket_spinlocks() 34 tlocks = kvmalloc_array(size, sizeof(spinlock_t), gfp); in __alloc_bucket_spinlocks() 37 for (i = 0; i < size; i++) { in __alloc_bucket_spinlocks() 44 *locks_mask = size - 1; in __alloc_bucket_spinlocks()
|
| D | test_static_keys.c | 55 static void invert_keys(struct test_key *keys, int size) in invert_keys() argument 60 for (i = 0; i < size; i++) { in invert_keys() 68 static int verify_keys(struct test_key *keys, int size, bool invert) in verify_keys() argument 73 for (i = 0; i < size; i++) { in verify_keys() 112 int size; in test_key_func() local 211 size = ARRAY_SIZE(static_key_tests); in test_key_func() 213 ret = verify_keys(static_key_tests, size, false); in test_key_func() 217 invert_keys(static_key_tests, size); in test_key_func() 218 ret = verify_keys(static_key_tests, size, true); in test_key_func() 222 invert_keys(static_key_tests, size); in test_key_func() [all …]
|
| D | string_helpers.c | 39 int string_get_size(u64 size, u64 blk_size, const enum string_size_units units, in string_get_size() argument 66 size = 0; in string_get_size() 67 if (size == 0) in string_get_size() 86 while (size >> 32) { in string_get_size() 87 do_div(size, divisor[units_base]); in string_get_size() 93 size *= blk_size; in string_get_size() 96 while (size >= divisor[units_base]) { in string_get_size() 97 remainder = do_div(size, divisor[units_base]); in string_get_size() 103 sf_cap = size; in string_get_size() 120 size += 1; in string_get_size() [all …]
|
| D | buildid.c | 163 static int parse_build_id(struct freader *r, unsigned char *build_id, __u32 *size, in parse_build_id() argument 202 if (size) in parse_build_id() 203 *size = desc_sz; in parse_build_id() 214 static int get_build_id_32(struct freader *r, unsigned char *build_id, __u32 *size) in get_build_id_32() argument 242 !parse_build_id(r, build_id, size, READ_ONCE(phdr->p_offset), in get_build_id_32() 250 static int get_build_id_64(struct freader *r, unsigned char *build_id, __u32 *size) in get_build_id_64() argument 279 !parse_build_id(r, build_id, size, READ_ONCE(phdr->p_offset), in get_build_id_64() 291 __u32 *size, bool may_fault) in __build_id_parse() argument 322 ret = get_build_id_32(&r, build_id, size); in __build_id_parse() 324 ret = get_build_id_64(&r, build_id, size); in __build_id_parse() [all …]
|
| D | sg_split.c | 26 size_t size = sizes[0], len; in sg_calculate_split() local 42 len = min_t(size_t, size, sglen - skip); in sg_calculate_split() 47 size -= len; in sg_calculate_split() 51 while (!size && (skip + len < sglen) && (--nb_splits > 0)) { in sg_calculate_split() 53 size = *(++sizes); in sg_calculate_split() 55 len = min_t(size_t, size, sglen - skip); in sg_calculate_split() 61 size -= len; in sg_calculate_split() 65 if (!size && --nb_splits > 0) { in sg_calculate_split() 67 size = *(++sizes); in sg_calculate_split() 74 return (size || !splitters[0].in_sg0) ? -EINVAL : 0; in sg_calculate_split()
|
| /lib/crypto/mpi/ |
| D | mpih-mul.c | 21 #define MPN_MUL_N_RECURSE(prodp, up, vp, size, tspace) \ argument 23 if ((size) < KARATSUBA_THRESHOLD) \ 24 mul_n_basecase(prodp, up, vp, size); \ 26 mul_n(prodp, up, vp, size, tspace); \ 29 #define MPN_SQR_N_RECURSE(prodp, up, size, tspace) \ argument 31 if ((size) < KARATSUBA_THRESHOLD) \ 32 mpih_sqr_n_basecase(prodp, up, size); \ 34 mpih_sqr_n(prodp, up, size, tspace); \ 55 mul_n_basecase(mpi_ptr_t prodp, mpi_ptr_t up, mpi_ptr_t vp, mpi_size_t size) in mul_n_basecase() argument 66 MPN_COPY(prodp, up, size); in mul_n_basecase() [all …]
|
| /lib/xz/ |
| D | xz_dec_bcj.c | 61 size_t size; member 89 static size_t bcj_x86(struct xz_dec_bcj *s, uint8_t *buf, size_t size) in bcj_x86() argument 104 if (size <= 4) in bcj_x86() 107 size -= 4; in bcj_x86() 108 for (i = 0; i < size; ++i) { in bcj_x86() 161 static size_t bcj_powerpc(struct xz_dec_bcj *s, uint8_t *buf, size_t size) in bcj_powerpc() argument 166 size &= ~(size_t)3; in bcj_powerpc() 168 for (i = 0; i < size; i += 4) { in bcj_powerpc() 184 static size_t bcj_ia64(struct xz_dec_bcj *s, uint8_t *buf, size_t size) in bcj_ia64() argument 225 size &= ~(size_t)15; in bcj_ia64() [all …]
|
| /lib/zstd/common/ |
| D | zstd_common.c | 56 void* ZSTD_customMalloc(size_t size, ZSTD_customMem customMem) in ZSTD_customMalloc() argument 59 return customMem.customAlloc(customMem.opaque, size); in ZSTD_customMalloc() 60 return ZSTD_malloc(size); in ZSTD_customMalloc() 63 void* ZSTD_customCalloc(size_t size, ZSTD_customMem customMem) in ZSTD_customCalloc() argument 68 void* const ptr = customMem.customAlloc(customMem.opaque, size); in ZSTD_customCalloc() 69 ZSTD_memset(ptr, 0, size); in ZSTD_customCalloc() 72 return ZSTD_calloc(1, size); in ZSTD_customCalloc()
|
| /lib/crypto/ |
| D | memneq.c | 68 __crypto_memneq_generic(const void *a, const void *b, size_t size) in __crypto_memneq_generic() argument 73 while (size >= sizeof(unsigned long)) { in __crypto_memneq_generic() 79 size -= sizeof(unsigned long); in __crypto_memneq_generic() 82 while (size > 0) { in __crypto_memneq_generic() 87 size -= 1; in __crypto_memneq_generic() 164 size_t size) in __crypto_memneq() argument 166 switch (size) { in __crypto_memneq() 170 return __crypto_memneq_generic(a, b, size); in __crypto_memneq()
|