/lib/ |
D | cpu_rmap.c | 56 rmap->near[cpu].index = cpu % size; in alloc_cpu_rmap() 105 rmap->near[cpu].index = rmap->near[neigh].index; in cpu_rmap_copy_neigh() 116 unsigned index; in debug_print_rmap() local 122 index = rmap->near[cpu].index; in debug_print_rmap() 124 cpu, index, rmap->near[cpu].dist); in debug_print_rmap() 143 u16 index; in cpu_rmap_add() local 146 index = rmap->used++; in cpu_rmap_add() 147 rmap->obj[index] = obj; in cpu_rmap_add() 148 return index; in cpu_rmap_add() 158 int cpu_rmap_update(struct cpu_rmap *rmap, u16 index, in cpu_rmap_update() argument [all …]
|
D | iommu-helper.c | 9 int iommu_is_span_boundary(unsigned int index, unsigned int nr, in iommu_is_span_boundary() argument 15 shift = (shift + index) & (boundary_size - 1); in iommu_is_span_boundary() 24 unsigned long index; in iommu_area_alloc() local 29 index = bitmap_find_next_zero_area(map, size, start, nr, align_mask); in iommu_area_alloc() 30 if (index < size) { in iommu_area_alloc() 31 if (iommu_is_span_boundary(index, nr, shift, boundary_size)) { in iommu_area_alloc() 33 start = index + 1; in iommu_area_alloc() 36 bitmap_set(map, index, nr); in iommu_area_alloc() 37 return index; in iommu_area_alloc()
|
D | radix-tree.c | 321 static int radix_tree_extend(struct radix_tree_root *root, unsigned long index) in radix_tree_extend() argument 330 while (index > radix_tree_maxindex(height)) in radix_tree_extend() 385 int __radix_tree_create(struct radix_tree_root *root, unsigned long index, in __radix_tree_create() argument 393 if (index > radix_tree_maxindex(root->height)) { in __radix_tree_create() 394 error = radix_tree_extend(root, index); in __radix_tree_create() 421 offset = (index >> shift) & RADIX_TREE_MAP_MASK; in __radix_tree_create() 444 unsigned long index, void *item) in radix_tree_insert() argument 452 error = __radix_tree_create(root, index, &node, &slot); in radix_tree_insert() 461 BUG_ON(tag_get(node, 0, index & RADIX_TREE_MAP_MASK)); in radix_tree_insert() 462 BUG_ON(tag_get(node, 1, index & RADIX_TREE_MAP_MASK)); in radix_tree_insert() [all …]
|
D | proportions.c | 83 pd->index = 0; in prop_descriptor_init() 107 int index; in prop_change_shift() local 117 index = pd->index ^ 1; in prop_change_shift() 118 offset = pd->pg[pd->index].shift - shift; in prop_change_shift() 122 pd->pg[index].shift = shift; in prop_change_shift() 125 events = percpu_counter_sum(&pd->pg[pd->index].events); in prop_change_shift() 130 percpu_counter_set(&pd->pg[index].events, events); in prop_change_shift() 136 pd->index = index; in prop_change_shift() 152 int index; in prop_get_global() local 155 index = pd->index; in prop_get_global() [all …]
|
D | swiotlb.c | 432 unsigned int nslots, stride, index, wrap; in swiotlb_tbl_map_single() local 471 index = ALIGN(io_tlb_index, stride); in swiotlb_tbl_map_single() 472 if (index >= io_tlb_nslabs) in swiotlb_tbl_map_single() 473 index = 0; in swiotlb_tbl_map_single() 474 wrap = index; in swiotlb_tbl_map_single() 477 while (iommu_is_span_boundary(index, nslots, offset_slots, in swiotlb_tbl_map_single() 479 index += stride; in swiotlb_tbl_map_single() 480 if (index >= io_tlb_nslabs) in swiotlb_tbl_map_single() 481 index = 0; in swiotlb_tbl_map_single() 482 if (index == wrap) in swiotlb_tbl_map_single() [all …]
|
D | genalloc.c | 549 unsigned long index; in gen_pool_best_fit() local 551 index = bitmap_find_next_zero_area(map, size, start, nr, 0); in gen_pool_best_fit() 553 while (index < size) { in gen_pool_best_fit() 554 int next_bit = find_next_bit(map, size, index + nr); in gen_pool_best_fit() 555 if ((next_bit - index) < len) { in gen_pool_best_fit() 556 len = next_bit - index; in gen_pool_best_fit() 557 start_bit = index; in gen_pool_best_fit() 561 index = bitmap_find_next_zero_area(map, size, in gen_pool_best_fit() 632 const char *propname, int index) in of_get_named_gen_pool() argument 637 np_pool = of_parse_phandle(np, propname, index); in of_get_named_gen_pool()
|
D | bitmap.c | 349 unsigned long index, end, i; in bitmap_find_next_zero_area_off() local 351 index = find_next_zero_bit(map, size, start); in bitmap_find_next_zero_area_off() 354 index = __ALIGN_MASK(index + align_offset, align_mask) - align_offset; in bitmap_find_next_zero_area_off() 356 end = index + nr; in bitmap_find_next_zero_area_off() 359 i = find_next_bit(map, end, index); in bitmap_find_next_zero_area_off() 364 return index; in bitmap_find_next_zero_area_off() 1061 int index; /* index first long of region in bitmap */ in __reg_op() local 1074 index = pos / BITS_PER_LONG; in __reg_op() 1075 offset = pos - (index * BITS_PER_LONG); in __reg_op() 1090 if (bitmap[index + i] & mask) in __reg_op() [all …]
|
D | btree.c | 679 unsigned long *key, size_t index, in __btree_for_each() argument 703 size_t index, void *func2) in empty() argument 708 size_t index, void *__func) in visitorl() argument 712 func(elem, opaque, *key, index); in visitorl() 717 size_t index, void *__func) in visitor32() argument 722 func(elem, opaque, *key, index); in visitor32() 727 size_t index, void *__func) in visitor64() argument 732 func(elem, opaque, *key, index); in visitor64() 737 size_t index, void *__func) in visitor128() argument 742 func(elem, opaque, key[0], key[1], index); in visitor128() [all …]
|
D | lru_cache.c | 619 void lc_set(struct lru_cache *lc, unsigned int enr, int index) in lc_set() argument 624 if (index < 0 || index >= lc->nr_elements) in lc_set() 627 e = lc_element_by_index(lc, index); in lc_set()
|
D | vsprintf.c | 838 int index; in ip4_string() local 844 index = 0; in ip4_string() 847 index = 3; in ip4_string() 852 index = 3; in ip4_string() 858 index = 0; in ip4_string() 864 int digits = put_dec_trunc8(temp, addr[index]) - temp; in ip4_string() 876 index += step; in ip4_string() 1169 const u8 *index = be; in uuid_string() local 1176 index = le; in uuid_string() 1184 p = hex_byte_pack(p, addr[index[i]]); in uuid_string()
|
D | dma-debug.c | 352 struct dma_debug_entry *entry, index = *ref; in bucket_find_contain() local 356 entry = __hash_bucket_find(*bucket, &index, containing_match); in bucket_find_contain() 366 index.dev_addr -= (1 << HASH_FN_SHIFT); in bucket_find_contain() 367 *bucket = get_hash_bucket(&index, flags); in bucket_find_contain()
|
/lib/xz/ |
D | xz_dec_stream.c | 116 } index; member 284 s->index.size += in_used; in index_update() 307 switch (s->index.sequence) { in dec_index() 309 s->index.count = s->vli; in dec_index() 316 if (s->index.count != s->block.count) in dec_index() 319 s->index.sequence = SEQ_INDEX_UNPADDED; in dec_index() 323 s->index.hash.unpadded += s->vli; in dec_index() 324 s->index.sequence = SEQ_INDEX_UNCOMPRESSED; in dec_index() 328 s->index.hash.uncompressed += s->vli; in dec_index() 329 s->index.hash.crc32 = xz_crc32( in dec_index() [all …]
|