| /lib/ |
| D | generic-radix-tree.c | 12 void *__genradix_ptr(struct __genradix *radix, size_t offset) in __genradix_ptr() argument 14 return __genradix_ptr_inlined(radix, offset); in __genradix_ptr() 22 void *__genradix_ptr_alloc(struct __genradix *radix, size_t offset, in __genradix_ptr_alloc() argument 40 if (n && ilog2(offset) < genradix_depth_shift(level)) in __genradix_ptr_alloc() 63 &n->children[offset >> genradix_depth_shift(level)]; in __genradix_ptr_alloc() 64 offset &= genradix_depth_size(level) - 1; in __genradix_ptr_alloc() 82 return &n->data[offset]; in __genradix_ptr_alloc() 94 if (iter->offset == SIZE_MAX) in __genradix_iter_peek() 105 if (ilog2(iter->offset) >= genradix_depth_shift(level)) in __genradix_iter_peek() 111 i = (iter->offset >> genradix_depth_shift(level)) & in __genradix_iter_peek() [all …]
|
| D | radix-tree.c | 88 unsigned int offset = (index >> parent->shift) & RADIX_TREE_MAP_MASK; in radix_tree_descend() local 89 void __rcu **entry = rcu_dereference_raw(parent->slots[offset]); in radix_tree_descend() 92 return offset; in radix_tree_descend() 101 int offset) in tag_set() argument 103 __set_bit(offset, node->tags[tag]); in tag_set() 107 int offset) in tag_clear() argument 109 __clear_bit(offset, node->tags[tag]); in tag_clear() 113 int offset) in tag_get() argument 115 return test_bit(offset, node->tags[tag]); in tag_get() 181 unsigned long offset) in radix_tree_find_next_bit() argument [all …]
|
| D | maple_tree.c | 717 mas_safe_min(struct ma_state *mas, unsigned long *pivots, unsigned char offset) in mas_safe_min() argument 719 if (likely(offset)) in mas_safe_min() 720 return pivots[offset - 1] + 1; in mas_safe_min() 787 void __rcu **slots, unsigned char offset) in mt_slot() argument 789 return rcu_dereference_check(slots[offset], mt_locked(mt)); in mt_slot() 793 void __rcu **slots, unsigned char offset) in mt_slot_locked() argument 795 return rcu_dereference_protected(slots[offset], mt_write_locked(mt)); in mt_slot_locked() 806 void __rcu **slots, unsigned char offset) in mas_slot_locked() argument 808 return mt_slot_locked(mas->tree, slots, offset); in mas_slot_locked() 820 unsigned char offset) in mas_slot() argument [all …]
|
| D | cmdline_kunit.c | 41 static void cmdline_do_one_test(struct kunit *test, const char *in, int rc, int offset) in cmdline_do_one_test() argument 51 KUNIT_EXPECT_PTR_EQ_MSG(test, out, in + offset, fmt, in); in cmdline_do_one_test() 61 int offset; in cmdline_test_noint() local 64 offset = !!(*str == '-'); in cmdline_test_noint() 65 cmdline_do_one_test(test, str, rc, offset); in cmdline_test_noint() 77 int offset; in cmdline_test_lead_int() local 81 offset = strlen(in) - strlen(str) + !!(rc == 2); in cmdline_test_lead_int() 82 cmdline_do_one_test(test, in, rc, offset); in cmdline_test_lead_int() 95 int offset; in cmdline_test_tail_int() local 102 offset = rc ? strlen(in) : !!(*str == '-'); in cmdline_test_tail_int() [all …]
|
| D | xarray.c | 86 unsigned int offset, xa_mark_t mark) in node_get_mark() argument 88 return test_bit(offset, node_marks(node, mark)); in node_get_mark() 92 static inline bool node_set_mark(struct xa_node *node, unsigned int offset, in node_set_mark() argument 95 return __test_and_set_bit(offset, node_marks(node, mark)); in node_set_mark() 99 static inline bool node_clear_mark(struct xa_node *node, unsigned int offset, in node_clear_mark() argument 102 return __test_and_clear_bit(offset, node_marks(node, mark)); in node_clear_mark() 155 static void xas_move_index(struct xa_state *xas, unsigned long offset) in xas_move_index() argument 159 xas->xa_index += offset << shift; in xas_move_index() 206 unsigned int offset = get_offset(xas->xa_index, node); in xas_descend() local 207 void *entry = xa_entry(xas->xa, node, offset); in xas_descend() [all …]
|
| D | iov_iter.c | 326 static inline bool page_copy_sane(struct page *page, size_t offset, size_t n) in page_copy_sane() argument 329 size_t v = n + offset; in page_copy_sane() 349 size_t copy_page_to_iter(struct page *page, size_t offset, size_t bytes, in copy_page_to_iter() argument 353 if (!page_copy_sane(page, offset, bytes)) in copy_page_to_iter() 357 page += offset / PAGE_SIZE; // first subpage in copy_page_to_iter() 358 offset %= PAGE_SIZE; in copy_page_to_iter() 361 size_t n = min(bytes, (size_t)PAGE_SIZE - offset); in copy_page_to_iter() 362 n = _copy_to_iter(kaddr + offset, n, i); in copy_page_to_iter() 368 offset += n; in copy_page_to_iter() 369 if (offset == PAGE_SIZE) { in copy_page_to_iter() [all …]
|
| D | find_bit.c | 236 unsigned long size, unsigned long offset) in find_next_clump8() argument 238 offset = find_next_bit(addr, size, offset); in find_next_clump8() 239 if (offset == size) in find_next_clump8() 242 offset = round_down(offset, 8); in find_next_clump8() 243 *clump = bitmap_get_value8(addr, offset); in find_next_clump8() 245 return offset; in find_next_clump8() 265 unsigned long size, unsigned long offset) in _find_next_zero_bit_le() argument 267 return FIND_NEXT_BIT(~addr[idx], swab, size, offset); in _find_next_zero_bit_le() 274 unsigned long size, unsigned long offset) in _find_next_bit_le() argument 276 return FIND_NEXT_BIT(addr[idx], swab, size, offset); in _find_next_bit_le()
|
| D | scatterlist.c | 456 struct page **pages, unsigned int n_pages, unsigned int offset, in sg_alloc_append_table_from_pages() argument 478 sgt_append->prv->offset + sgt_append->prv->length) / PAGE_SIZE; in sg_alloc_append_table_from_pages() 480 if (WARN_ON(offset)) in sg_alloc_append_table_from_pages() 538 chunk_size = ((j - cur_page) << PAGE_SHIFT) - offset; in sg_alloc_append_table_from_pages() 540 min_t(unsigned long, size, chunk_size), offset); in sg_alloc_append_table_from_pages() 543 offset = 0; in sg_alloc_append_table_from_pages() 580 unsigned int n_pages, unsigned int offset, in sg_alloc_table_from_pages_segment() argument 587 err = sg_alloc_append_table_from_pages(&append, pages, n_pages, offset, in sg_alloc_table_from_pages_segment() 740 return PAGE_ALIGN(sg->offset + sg->length) >> PAGE_SHIFT; in sg_page_count() 764 return PAGE_ALIGN(sg->offset + sg_dma_len(sg)) >> PAGE_SHIFT; in sg_dma_page_count() [all …]
|
| D | packing.c | 12 static int get_le_offset(int offset) in get_le_offset() argument 16 closest_multiple_of_4 = (offset / 4) * 4; in get_le_offset() 17 offset -= closest_multiple_of_4; in get_le_offset() 18 return closest_multiple_of_4 + (3 - offset); in get_le_offset() 21 static int get_reverse_lsw32_offset(int offset, size_t len) in get_reverse_lsw32_offset() argument 26 word_index = offset / 4; in get_reverse_lsw32_offset() 28 offset -= closest_multiple_of_4; in get_reverse_lsw32_offset() 30 return word_index * 4 + offset; in get_reverse_lsw32_offset()
|
| D | devres.c | 29 static void __iomem *__devm_ioremap(struct device *dev, resource_size_t offset, in __devm_ioremap() argument 42 addr = ioremap(offset, size); in __devm_ioremap() 45 addr = ioremap_uc(offset, size); in __devm_ioremap() 48 addr = ioremap_wc(offset, size); in __devm_ioremap() 51 addr = ioremap_np(offset, size); in __devm_ioremap() 72 void __iomem *devm_ioremap(struct device *dev, resource_size_t offset, in devm_ioremap() argument 75 return __devm_ioremap(dev, offset, size, DEVM_IOREMAP); in devm_ioremap() 87 void __iomem *devm_ioremap_uc(struct device *dev, resource_size_t offset, in devm_ioremap_uc() argument 90 return __devm_ioremap(dev, offset, size, DEVM_IOREMAP_UC); in devm_ioremap_uc() 102 void __iomem *devm_ioremap_wc(struct device *dev, resource_size_t offset, in devm_ioremap_wc() argument [all …]
|
| D | memcpy_kunit.c | 334 for (int offset = 0; offset < ARRAY_SIZE(large_src); offset++) { in copy_large_test() local 335 int right_zero_pos = offset + bytes; in copy_large_test() 340 memmove(large_dst + offset, large_src, bytes); in copy_large_test() 342 memcpy(large_dst + offset, large_src, bytes); in copy_large_test() 346 memcmp(large_dst, large_zero, offset), 0, in copy_large_test() 347 "with size %d at offset %d", bytes, offset); in copy_large_test() 351 "with size %d at offset %d", bytes, offset); in copy_large_test() 355 memcmp(large_dst + offset, large_src, bytes), 0, in copy_large_test() 356 "with size %d at offset %d", bytes, offset); in copy_large_test() 359 memset(large_dst + offset, 0, bytes); in copy_large_test()
|
| D | sbitmap.c | 376 static inline void emit_byte(struct seq_file *m, unsigned int offset, u8 byte) in emit_byte() argument 378 if ((offset & 0xf) == 0) { in emit_byte() 379 if (offset != 0) in emit_byte() 381 seq_printf(m, "%08x:", offset); in emit_byte() 383 if ((offset & 0x1) == 0) in emit_byte() 392 unsigned int offset = 0; in sbitmap_bitmap_show() local 408 emit_byte(m, offset, byte); in sbitmap_bitmap_show() 411 offset++; in sbitmap_bitmap_show() 418 emit_byte(m, offset, byte); in sbitmap_bitmap_show() 419 offset++; in sbitmap_bitmap_show() [all …]
|
| D | alloc_tag.c | 482 unsigned long offset; in reserve_module_tags() local 510 offset = mas.index; in reserve_module_tags() 511 offset += prepend; in reserve_module_tags() 512 offset = ALIGN(offset, align); in reserve_module_tags() 513 if (offset != mas.index) { in reserve_module_tags() 516 mas.last = offset - 1; in reserve_module_tags() 522 mas.index = offset; in reserve_module_tags() 523 mas.last = offset + size - 1; in reserve_module_tags() 531 mas.last = offset + size - 1; in reserve_module_tags() 542 if (module_tags.size < offset + size) { in reserve_module_tags() [all …]
|
| D | logic_iomem.c | 72 static void __iomem *real_ioremap(phys_addr_t offset, size_t size) in real_ioremap() argument 75 (unsigned long long)offset, size); in real_ioremap() 86 void __iomem *ioremap(phys_addr_t offset, size_t size) in ioremap() argument 94 if (rreg->res->start > offset) in ioremap() 96 if (rreg->res->end < offset + size - 1) in ioremap() 111 offs = rreg->ops->map(offset - found->res->start, in ioremap() 131 return real_ioremap(offset, size); in ioremap()
|
| D | ts_kmp.c | 45 unsigned int i, q = 0, text_len, consumed = state->offset; in kmp_find() 63 state->offset = consumed + i + 1; in kmp_find() 64 return state->offset - kmp->pattern_len; in kmp_find()
|
| D | slub_kunit.c | 57 ptr_addr = (unsigned long *)(p + s->offset); in test_next_pointer() 59 p[s->offset] = ~p[s->offset]; in test_next_pointer()
|
| /lib/lzo/ |
| D | lzo1x_decompress_safe.c | 79 size_t offset; in lzo1x_decompress_safe() local 86 offset = ip - ip_last; in lzo1x_decompress_safe() 87 if (unlikely(offset > MAX_255_COUNT)) in lzo1x_decompress_safe() 90 offset = (offset << 8) - offset; in lzo1x_decompress_safe() 91 t += offset + 15 + *ip++; in lzo1x_decompress_safe() 147 size_t offset; in lzo1x_decompress_safe() local 154 offset = ip - ip_last; in lzo1x_decompress_safe() 155 if (unlikely(offset > MAX_255_COUNT)) in lzo1x_decompress_safe() 158 offset = (offset << 8) - offset; in lzo1x_decompress_safe() 159 t += offset + 31 + *ip++; in lzo1x_decompress_safe() [all …]
|
| /lib/pldmfw/ |
| D | pldmfw.c | 25 size_t offset; member 65 pldm_check_fw_space(struct pldmfw_priv *data, size_t offset, size_t length) in pldm_check_fw_space() argument 67 size_t expected_size = offset + length; in pldm_check_fw_space() 95 err = pldm_check_fw_space(data, data->offset, bytes_to_move); in pldm_move_fw_offset() 99 data->offset += bytes_to_move; in pldm_move_fw_offset() 153 err = pldm_check_fw_space(data, data->offset, header_size); in pldm_parse_header() 176 data->offset); in pldm_parse_header() 265 desc_start = data->fw->data + data->offset; in pldm_parse_desc_tlvs() 348 bitmap_ptr = data->fw->data + data->offset; in pldm_parse_one_record() 364 record->version_string = data->fw->data + data->offset; in pldm_parse_one_record() [all …]
|
| /lib/zstd/compress/ |
| D | zstd_ldm.c | 185 unsigned const offset = *pOffset; in ZSTD_ldm_insertEntry() local 187 *(ZSTD_ldm_getBucket(ldmState, hash, ldmParams) + offset) = entry; in ZSTD_ldm_insertEntry() 188 *pOffset = (BYTE)((offset + 1) & ((1u << ldmParams.bucketSizeLog) - 1)); in ZSTD_ldm_insertEntry() 296 entry.offset = (U32)(split - base); in ZSTD_ldm_fillHashTable() 383 U32 offset; in ZSTD_ldm_generateSequences_internal() local 392 newEntry.offset = (U32)(split - base); in ZSTD_ldm_generateSequences_internal() 406 if (cur->checksum != checksum || cur->offset <= lowestIndex) { in ZSTD_ldm_generateSequences_internal() 411 cur->offset < dictLimit ? dictBase : base; in ZSTD_ldm_generateSequences_internal() 412 BYTE const* const pMatch = curMatchBase + cur->offset; in ZSTD_ldm_generateSequences_internal() 414 cur->offset < dictLimit ? dictEnd : iend; in ZSTD_ldm_generateSequences_internal() [all …]
|
| D | zstd_double_fast.c | 73 U32 offset; in ZSTD_compressBlock_doubleFast_noDict_generic() local 144 offset = (U32)(ip-matchl0); in ZSTD_compressBlock_doubleFast_noDict_generic() 192 offset = (U32)(ip-matchl1); in ZSTD_compressBlock_doubleFast_noDict_generic() 200 offset = (U32)(ip - matchs0); in ZSTD_compressBlock_doubleFast_noDict_generic() 207 offset_1 = offset; in ZSTD_compressBlock_doubleFast_noDict_generic() 220 ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, STORE_OFFSET(offset), mLength); in ZSTD_compressBlock_doubleFast_noDict_generic() 309 U32 offset; in ZSTD_compressBlock_doubleFast_dictMatchState_generic() local 339 offset = (U32)(ip-matchLong); in ZSTD_compressBlock_doubleFast_dictMatchState_generic() 351 offset = (U32)(curr - dictMatchIndexL - dictIndexDelta); in ZSTD_compressBlock_doubleFast_dictMatchState_generic() 390 offset = (U32)(ip-matchL3); in ZSTD_compressBlock_doubleFast_dictMatchState_generic() [all …]
|
| /lib/kunit/ |
| D | user_alloc.c | 22 unsigned long offset; member 64 vres.addr = vm_mmap(p->file, p->addr, p->len, p->prot, p->flag, p->offset); in kunit_vm_mmap_init() 94 unsigned long offset) in kunit_vm_mmap() argument 102 .offset = offset, in kunit_vm_mmap()
|
| D | string-stream-test.c | 224 size_t offset, total_len; in string_stream_variable_length_line_test() local 238 offset = prandom_u32_state(&rnd) % (sizeof(line) - 1); in string_stream_variable_length_line_test() 239 string_stream_add(stream, "%s\n", &line[offset]); in string_stream_variable_length_line_test() 240 total_len += sizeof(line) - offset; in string_stream_variable_length_line_test() 261 offset = prandom_u32_state(&rnd) % (sizeof(line) - 1); in string_stream_variable_length_line_test() 262 KUNIT_EXPECT_STREQ(test, pos, &line[offset]); in string_stream_variable_length_line_test() 462 int offset, i; in string_stream_performance_test() local 472 offset = i % (sizeof(test_line) - 1); in string_stream_performance_test() 473 string_stream_add(stream, "%s: %d\n", &test_line[offset], i); in string_stream_performance_test()
|
| /lib/842/ |
| D | 842_decompress.c | 165 u64 index, offset, total = round_down(p->out - p->ostart, 8); in __do_index() local 172 offset = index * size; in __do_index() 184 if (offset >= pos) in __do_index() 187 offset += section; in __do_index() 190 if (offset + size > total) { in __do_index() 192 (unsigned long)offset, (unsigned long)total); in __do_index() 201 (unsigned long)(index * size), (unsigned long)offset, in __do_index() 203 (unsigned long)beN_to_cpu(&p->ostart[offset], size)); in __do_index() 205 memcpy(p->out, &p->ostart[offset], size); in __do_index()
|
| /lib/zstd/decompress/ |
| D | zstd_decompress_block.c | 736 size_t offset; member 759 HINT_INLINE void ZSTD_overlapCopy8(BYTE** op, BYTE const** ip, size_t offset) { in ZSTD_overlapCopy8() argument 761 if (offset < 8) { in ZSTD_overlapCopy8() 765 int const sub2 = dec64table[offset]; in ZSTD_overlapCopy8() 770 *ip += dec32table[offset]; in ZSTD_overlapCopy8() 869 const BYTE* match = oLitEnd - sequence.offset; in ZSTD_execSequenceEnd() 884 if (sequence.offset > (size_t)(oLitEnd - prefixStart)) { in ZSTD_execSequenceEnd() 886 … RETURN_ERROR_IF(sequence.offset > (size_t)(oLitEnd - virtualStart), corruption_detected, ""); in ZSTD_execSequenceEnd() 916 const BYTE* match = oLitEnd - sequence.offset; in ZSTD_execSequenceEndSplitLitBuffer() 932 if (sequence.offset > (size_t)(oLitEnd - prefixStart)) { in ZSTD_execSequenceEndSplitLitBuffer() [all …]
|
| /lib/lz4/ |
| D | lz4_decompress.c | 124 size_t offset; in LZ4_decompress_generic() local 165 offset = LZ4_readLE16(ip); in LZ4_decompress_generic() 167 match = op - offset; in LZ4_decompress_generic() 172 (offset >= 8) && in LZ4_decompress_generic() 289 offset = LZ4_readLE16(ip); in LZ4_decompress_generic() 291 match = op - offset; in LZ4_decompress_generic() 311 LZ4_write32(op, (U32)offset); in LZ4_decompress_generic() 405 if (unlikely(offset < 8)) { in LZ4_decompress_generic() 410 match += inc32table[offset]; in LZ4_decompress_generic() 412 match -= dec64table[offset]; in LZ4_decompress_generic()
|