/lib/ |
D | radix-tree.c | 86 unsigned int offset = (index >> parent->shift) & RADIX_TREE_MAP_MASK; in radix_tree_descend() local 87 void __rcu **entry = rcu_dereference_raw(parent->slots[offset]); in radix_tree_descend() 90 return offset; in radix_tree_descend() 99 int offset) in tag_set() argument 101 __set_bit(offset, node->tags[tag]); in tag_set() 105 int offset) in tag_clear() argument 107 __clear_bit(offset, node->tags[tag]); in tag_clear() 111 int offset) in tag_get() argument 113 return test_bit(offset, node->tags[tag]); in tag_get() 179 unsigned long offset) in radix_tree_find_next_bit() argument [all …]
|
D | generic-radix-tree.c | 54 void *__genradix_ptr(struct __genradix *radix, size_t offset) in __genradix_ptr() argument 60 if (ilog2(offset) >= genradix_depth_shift(level)) in __genradix_ptr() 71 n = n->children[offset >> genradix_depth_shift(level)]; in __genradix_ptr() 72 offset &= genradix_depth_size(level) - 1; in __genradix_ptr() 75 return &n->data[offset]; in __genradix_ptr() 104 void *__genradix_ptr_alloc(struct __genradix *radix, size_t offset, in __genradix_ptr_alloc() argument 118 if (n && ilog2(offset) < genradix_depth_shift(level)) in __genradix_ptr_alloc() 139 &n->children[offset >> genradix_depth_shift(level)]; in __genradix_ptr_alloc() 140 offset &= genradix_depth_size(level) - 1; in __genradix_ptr_alloc() 158 return &n->data[offset]; in __genradix_ptr_alloc() [all …]
|
D | cmdline_kunit.c | 41 static void cmdline_do_one_test(struct kunit *test, const char *in, int rc, int offset) in cmdline_do_one_test() argument 51 KUNIT_EXPECT_PTR_EQ_MSG(test, out, in + offset, fmt, in); in cmdline_do_one_test() 61 int offset; in cmdline_test_noint() local 64 offset = !!(*str == '-'); in cmdline_test_noint() 65 cmdline_do_one_test(test, str, rc, offset); in cmdline_test_noint() 77 int offset; in cmdline_test_lead_int() local 81 offset = strlen(in) - strlen(str) + !!(rc == 2); in cmdline_test_lead_int() 82 cmdline_do_one_test(test, in, rc, offset); in cmdline_test_lead_int() 95 int offset; in cmdline_test_tail_int() local 102 offset = rc ? strlen(in) : !!(*str == '-'); in cmdline_test_tail_int() [all …]
|
D | pci_iomap.c | 30 unsigned long offset, in pci_iomap_range() argument 37 if (len <= offset || !start) in pci_iomap_range() 39 len -= offset; in pci_iomap_range() 40 start += offset; in pci_iomap_range() 70 unsigned long offset, in pci_iomap_wc_range() argument 81 if (len <= offset || !start) in pci_iomap_wc_range() 84 len -= offset; in pci_iomap_wc_range() 85 start += offset; in pci_iomap_wc_range()
|
D | xarray.c | 84 unsigned int offset, xa_mark_t mark) in node_get_mark() argument 86 return test_bit(offset, node_marks(node, mark)); in node_get_mark() 90 static inline bool node_set_mark(struct xa_node *node, unsigned int offset, in node_set_mark() argument 93 return __test_and_set_bit(offset, node_marks(node, mark)); in node_set_mark() 97 static inline bool node_clear_mark(struct xa_node *node, unsigned int offset, in node_clear_mark() argument 100 return __test_and_clear_bit(offset, node_marks(node, mark)); in node_clear_mark() 153 static void xas_move_index(struct xa_state *xas, unsigned long offset) in xas_move_index() argument 157 xas->xa_index += offset << shift; in xas_move_index() 203 unsigned int offset = get_offset(xas->xa_index, node); in xas_descend() local 204 void *entry = xa_entry(xas->xa, node, offset); in xas_descend() [all …]
|
D | find_bit.c | 131 unsigned long size, unsigned long offset) in find_next_clump8() argument 133 offset = find_next_bit(addr, size, offset); in find_next_clump8() 134 if (offset == size) in find_next_clump8() 137 offset = round_down(offset, 8); in find_next_clump8() 138 *clump = bitmap_get_value8(addr, offset); in find_next_clump8() 140 return offset; in find_next_clump8()
|
D | scatterlist.c | 443 struct page **pages, unsigned int n_pages, unsigned int offset, in sg_alloc_append_table_from_pages() argument 465 sgt_append->prv->offset + sgt_append->prv->length) / in sg_alloc_append_table_from_pages() 468 if (WARN_ON(offset)) in sg_alloc_append_table_from_pages() 524 chunk_size = ((j - cur_page) << PAGE_SHIFT) - offset; in sg_alloc_append_table_from_pages() 526 min_t(unsigned long, size, chunk_size), offset); in sg_alloc_append_table_from_pages() 529 offset = 0; in sg_alloc_append_table_from_pages() 566 unsigned int n_pages, unsigned int offset, in sg_alloc_table_from_pages_segment() argument 573 err = sg_alloc_append_table_from_pages(&append, pages, n_pages, offset, in sg_alloc_table_from_pages_segment() 726 return PAGE_ALIGN(sg->offset + sg->length) >> PAGE_SHIFT; in sg_page_count() 750 return PAGE_ALIGN(sg->offset + sg_dma_len(sg)) >> PAGE_SHIFT; in sg_dma_page_count() [all …]
|
D | packing.c | 11 static int get_le_offset(int offset) in get_le_offset() argument 15 closest_multiple_of_4 = (offset / 4) * 4; in get_le_offset() 16 offset -= closest_multiple_of_4; in get_le_offset() 17 return closest_multiple_of_4 + (3 - offset); in get_le_offset() 20 static int get_reverse_lsw32_offset(int offset, size_t len) in get_reverse_lsw32_offset() argument 25 word_index = offset / 4; in get_reverse_lsw32_offset() 27 offset -= closest_multiple_of_4; in get_reverse_lsw32_offset() 29 return word_index * 4 + offset; in get_reverse_lsw32_offset()
|
D | devres.c | 26 static void __iomem *__devm_ioremap(struct device *dev, resource_size_t offset, in __devm_ioremap() argument 38 addr = ioremap(offset, size); in __devm_ioremap() 41 addr = ioremap_uc(offset, size); in __devm_ioremap() 44 addr = ioremap_wc(offset, size); in __devm_ioremap() 47 addr = ioremap_np(offset, size); in __devm_ioremap() 68 void __iomem *devm_ioremap(struct device *dev, resource_size_t offset, in devm_ioremap() argument 71 return __devm_ioremap(dev, offset, size, DEVM_IOREMAP); in devm_ioremap() 83 void __iomem *devm_ioremap_uc(struct device *dev, resource_size_t offset, in devm_ioremap_uc() argument 86 return __devm_ioremap(dev, offset, size, DEVM_IOREMAP_UC); in devm_ioremap_uc() 98 void __iomem *devm_ioremap_wc(struct device *dev, resource_size_t offset, in devm_ioremap_wc() argument [all …]
|
D | iov_iter.c | 45 unsigned offset = p->bv_offset + skip; \ 48 offset / PAGE_SIZE); \ 49 base = kaddr + offset % PAGE_SIZE; \ 51 (size_t)(PAGE_SIZE - offset % PAGE_SIZE)); \ 74 unsigned offset = start % PAGE_SIZE; \ 92 base = kaddr + offset; \ 93 len = PAGE_SIZE - offset; \ 102 offset = 0; \ 173 static size_t copy_page_to_iter_iovec(struct page *page, size_t offset, size_t bytes, in copy_page_to_iter_iovec() argument 196 from = kaddr + offset; in copy_page_to_iter_iovec() [all …]
|
D | logic_iomem.c | 72 static void __iomem *real_ioremap(phys_addr_t offset, size_t size) in real_ioremap() argument 75 (unsigned long long)offset, size); in real_ioremap() 86 void __iomem *ioremap(phys_addr_t offset, size_t size) in ioremap() argument 94 if (rreg->res->start > offset) in ioremap() 96 if (rreg->res->end < offset + size - 1) in ioremap() 111 offs = rreg->ops->map(offset - found->res->start, in ioremap() 131 return real_ioremap(offset, size); in ioremap()
|
D | stackdepot.c | 55 u32 offset : STACK_ALLOC_OFFSET_BITS; member 136 stack->handle.offset = depot_offset >> STACK_ALLOC_ALIGN; in depot_alloc_stack() 230 size_t offset = parts.offset << STACK_ALLOC_ALIGN; in stack_depot_fetch() local 242 stack = slab + offset; in stack_depot_fetch()
|
D | sbitmap.c | 348 static inline void emit_byte(struct seq_file *m, unsigned int offset, u8 byte) in emit_byte() argument 350 if ((offset & 0xf) == 0) { in emit_byte() 351 if (offset != 0) in emit_byte() 353 seq_printf(m, "%08x:", offset); in emit_byte() 355 if ((offset & 0x1) == 0) in emit_byte() 364 unsigned int offset = 0; in sbitmap_bitmap_show() local 380 emit_byte(m, offset, byte); in sbitmap_bitmap_show() 383 offset++; in sbitmap_bitmap_show() 390 emit_byte(m, offset, byte); in sbitmap_bitmap_show() 391 offset++; in sbitmap_bitmap_show() [all …]
|
D | ubsan.c | 359 unsigned long offset); 362 unsigned long offset) in __ubsan_handle_alignment_assumption() argument 372 if (offset) in __ubsan_handle_alignment_assumption() 374 align, offset, data->type->type_name); in __ubsan_handle_alignment_assumption() 379 real_ptr = ptr - offset; in __ubsan_handle_alignment_assumption() 381 offset ? "offset " : "", BIT(real_ptr ? __ffs(real_ptr) : 0), in __ubsan_handle_alignment_assumption()
|
D | ts_kmp.c | 45 unsigned int i, q = 0, text_len, consumed = state->offset; in kmp_find() 63 state->offset = consumed + i + 1; in kmp_find() 64 return state->offset - kmp->pattern_len; in kmp_find()
|
D | test_rhashtable.c | 490 int offset = 0; in print_ht() local 505 offset += sprintf(buff + offset, "\nbucket[%d] -> ", i); in print_ht() 510 offset += sprintf(buff + offset, "[["); in print_ht() 516 offset += sprintf(buff + offset, " val %d (tid=%d)%s", p->value.id, p->value.tid, in print_ht() 525 offset += sprintf(buff + offset, "]]%s", !rht_is_a_nulls(pos) ? " -> " : ""); in print_ht()
|
D | slub_kunit.c | 40 ptr_addr = (unsigned long *)(p + s->offset); in test_next_pointer() 42 p[s->offset] = 0x12; in test_next_pointer()
|
/lib/lzo/ |
D | lzo1x_decompress_safe.c | 79 size_t offset; in lzo1x_decompress_safe() local 86 offset = ip - ip_last; in lzo1x_decompress_safe() 87 if (unlikely(offset > MAX_255_COUNT)) in lzo1x_decompress_safe() 90 offset = (offset << 8) - offset; in lzo1x_decompress_safe() 91 t += offset + 15 + *ip++; in lzo1x_decompress_safe() 147 size_t offset; in lzo1x_decompress_safe() local 154 offset = ip - ip_last; in lzo1x_decompress_safe() 155 if (unlikely(offset > MAX_255_COUNT)) in lzo1x_decompress_safe() 158 offset = (offset << 8) - offset; in lzo1x_decompress_safe() 159 t += offset + 31 + *ip++; in lzo1x_decompress_safe() [all …]
|
/lib/pldmfw/ |
D | pldmfw.c | 25 size_t offset; member 65 pldm_check_fw_space(struct pldmfw_priv *data, size_t offset, size_t length) in pldm_check_fw_space() argument 67 size_t expected_size = offset + length; in pldm_check_fw_space() 95 err = pldm_check_fw_space(data, data->offset, bytes_to_move); in pldm_move_fw_offset() 99 data->offset += bytes_to_move; in pldm_move_fw_offset() 153 err = pldm_check_fw_space(data, data->offset, header_size); in pldm_parse_header() 176 data->offset); in pldm_parse_header() 265 desc_start = data->fw->data + data->offset; in pldm_parse_desc_tlvs() 348 bitmap_ptr = data->fw->data + data->offset; in pldm_parse_one_record() 364 record->version_string = data->fw->data + data->offset; in pldm_parse_one_record() [all …]
|
/lib/842/ |
D | 842_decompress.c | 165 u64 index, offset, total = round_down(p->out - p->ostart, 8); in __do_index() local 172 offset = index * size; in __do_index() 184 if (offset >= pos) in __do_index() 187 offset += section; in __do_index() 190 if (offset + size > total) { in __do_index() 192 (unsigned long)offset, (unsigned long)total); in __do_index() 201 (unsigned long)(index * size), (unsigned long)offset, in __do_index() 203 (unsigned long)beN_to_cpu(&p->ostart[offset], size)); in __do_index() 205 memcpy(p->out, &p->ostart[offset], size); in __do_index()
|
/lib/zstd/ |
D | zstd_opt.h | 142 …tPrice(seqStore_t *seqStorePtr, U32 litLength, const BYTE *literals, U32 offset, U32 matchLength, … in ZSTD_getPrice() argument 146 BYTE const offCode = (BYTE)ZSTD_highbit32(offset + 1); in ZSTD_getPrice() 165 …atePrice(seqStore_t *seqStorePtr, U32 litLength, const BYTE *literals, U32 offset, U32 matchLength) in ZSTD_updatePrice() argument 184 BYTE const offCode = (BYTE)ZSTD_highbit32(offset + 1); in ZSTD_updatePrice() 426 U32 offset, rep[ZSTD_REP_NUM]; in ZSTD_compressBlock_opt_generic() local 636 offset = opt[cur].off; in ZSTD_compressBlock_opt_generic() 640 best_off = offset; in ZSTD_compressBlock_opt_generic() 657 offset = opt[cur].off; in ZSTD_compressBlock_opt_generic() 661 if (offset > ZSTD_REP_MOVE_OPT) { in ZSTD_compressBlock_opt_generic() 664 rep[0] = offset - ZSTD_REP_MOVE_OPT; in ZSTD_compressBlock_opt_generic() [all …]
|
D | decompress.c | 864 size_t offset; member 888 const BYTE *match = oLitEnd - sequence.offset; in ZSTD_execSequenceLast7() 908 if (sequence.offset > (size_t)(oLitEnd - base)) { in ZSTD_execSequenceLast7() 910 if (sequence.offset > (size_t)(oLitEnd - vBase)) in ZSTD_execSequenceLast7() 957 size_t offset; in ZSTD_decodeSequence() local 959 offset = 0; in ZSTD_decodeSequence() 961 …offset = OF_base[ofCode] + BIT_readBitsFast(&seqState->DStream, ofBits); /* <= (ZSTD_WINDOWLOG_MA… in ZSTD_decodeSequence() 967 offset += (llCode == 0); in ZSTD_decodeSequence() 968 if (offset) { in ZSTD_decodeSequence() 969 size_t temp = (offset == 3) ? seqState->prevOffset[0] - 1 : seqState->prevOffset[offset]; in ZSTD_decodeSequence() [all …]
|
D | compress.c | 577 ofCodeTable[u] = (BYTE)ZSTD_highbit32(sequences[u].offset); in ZSTD_seqToCodes() 772 BIT_addBits(&blockStream, sequences[nbSeq - 1].offset, extraBits); in ZSTD_compressSequences_internal() 775 BIT_addBits(&blockStream, sequences[nbSeq - 1].offset >> extraBits, ofBits - extraBits); in ZSTD_compressSequences_internal() 777 BIT_addBits(&blockStream, sequences[nbSeq - 1].offset, ofCodeTable[nbSeq - 1]); in ZSTD_compressSequences_internal() 807 BIT_addBits(&blockStream, sequences[n].offset, extraBits); in ZSTD_compressSequences_internal() 810 BIT_addBits(&blockStream, sequences[n].offset >> extraBits, ofBits - extraBits); /* 31 */ in ZSTD_compressSequences_internal() 812 BIT_addBits(&blockStream, sequences[n].offset, ofBits); /* 31 */ in ZSTD_compressSequences_internal() 875 seqStorePtr->sequences[0].offset = offsetCode + 1; in ZSTD_storeSeq() 1048 U32 offset; in ZSTD_compressBlock_fast_generic() local 1054 offset = (U32)(ip - match); in ZSTD_compressBlock_fast_generic() [all …]
|
/lib/lz4/ |
D | lz4_decompress.c | 124 size_t offset; in LZ4_decompress_generic() local 165 offset = LZ4_readLE16(ip); in LZ4_decompress_generic() 167 match = op - offset; in LZ4_decompress_generic() 172 (offset >= 8) && in LZ4_decompress_generic() 289 offset = LZ4_readLE16(ip); in LZ4_decompress_generic() 291 match = op - offset; in LZ4_decompress_generic() 311 LZ4_write32(op, (U32)offset); in LZ4_decompress_generic() 405 if (unlikely(offset < 8)) { in LZ4_decompress_generic() 410 match += inc32table[offset]; in LZ4_decompress_generic() 412 match -= dec64table[offset]; in LZ4_decompress_generic()
|
/lib/zlib_inflate/ |
D | inflate.c | 649 state->offset = (unsigned)this.val; in zlib_inflate() 656 state->offset += BITS(state->extra); in zlib_inflate() 660 if (state->offset > state->dmax) { in zlib_inflate() 666 if (state->offset > state->whave + out - left) { in zlib_inflate() 676 if (state->offset > copy) { /* copy from window */ in zlib_inflate() 677 copy = state->offset - copy; in zlib_inflate() 687 from = put - state->offset; in zlib_inflate()
|