Home
last modified time | relevance | path

Searched refs:index (Results 1 – 12 of 12) sorted by relevance

/lib/
Dcpu_rmap.c56 rmap->near[cpu].index = cpu % size; in alloc_cpu_rmap()
105 rmap->near[cpu].index = rmap->near[neigh].index; in cpu_rmap_copy_neigh()
116 unsigned index; in debug_print_rmap() local
122 index = rmap->near[cpu].index; in debug_print_rmap()
124 cpu, index, rmap->near[cpu].dist); in debug_print_rmap()
143 u16 index; in cpu_rmap_add() local
146 index = rmap->used++; in cpu_rmap_add()
147 rmap->obj[index] = obj; in cpu_rmap_add()
148 return index; in cpu_rmap_add()
158 int cpu_rmap_update(struct cpu_rmap *rmap, u16 index, in cpu_rmap_update() argument
[all …]
Diommu-helper.c9 int iommu_is_span_boundary(unsigned int index, unsigned int nr, in iommu_is_span_boundary() argument
15 shift = (shift + index) & (boundary_size - 1); in iommu_is_span_boundary()
24 unsigned long index; in iommu_area_alloc() local
29 index = bitmap_find_next_zero_area(map, size, start, nr, align_mask); in iommu_area_alloc()
30 if (index < size) { in iommu_area_alloc()
31 if (iommu_is_span_boundary(index, nr, shift, boundary_size)) { in iommu_area_alloc()
33 start = index + 1; in iommu_area_alloc()
36 bitmap_set(map, index, nr); in iommu_area_alloc()
37 return index; in iommu_area_alloc()
Dradix-tree.c321 static int radix_tree_extend(struct radix_tree_root *root, unsigned long index) in radix_tree_extend() argument
330 while (index > radix_tree_maxindex(height)) in radix_tree_extend()
385 int __radix_tree_create(struct radix_tree_root *root, unsigned long index, in __radix_tree_create() argument
393 if (index > radix_tree_maxindex(root->height)) { in __radix_tree_create()
394 error = radix_tree_extend(root, index); in __radix_tree_create()
421 offset = (index >> shift) & RADIX_TREE_MAP_MASK; in __radix_tree_create()
444 unsigned long index, void *item) in radix_tree_insert() argument
452 error = __radix_tree_create(root, index, &node, &slot); in radix_tree_insert()
461 BUG_ON(tag_get(node, 0, index & RADIX_TREE_MAP_MASK)); in radix_tree_insert()
462 BUG_ON(tag_get(node, 1, index & RADIX_TREE_MAP_MASK)); in radix_tree_insert()
[all …]
Dproportions.c83 pd->index = 0; in prop_descriptor_init()
107 int index; in prop_change_shift() local
117 index = pd->index ^ 1; in prop_change_shift()
118 offset = pd->pg[pd->index].shift - shift; in prop_change_shift()
122 pd->pg[index].shift = shift; in prop_change_shift()
125 events = percpu_counter_sum(&pd->pg[pd->index].events); in prop_change_shift()
130 percpu_counter_set(&pd->pg[index].events, events); in prop_change_shift()
136 pd->index = index; in prop_change_shift()
152 int index; in prop_get_global() local
155 index = pd->index; in prop_get_global()
[all …]
Dswiotlb.c432 unsigned int nslots, stride, index, wrap; in swiotlb_tbl_map_single() local
471 index = ALIGN(io_tlb_index, stride); in swiotlb_tbl_map_single()
472 if (index >= io_tlb_nslabs) in swiotlb_tbl_map_single()
473 index = 0; in swiotlb_tbl_map_single()
474 wrap = index; in swiotlb_tbl_map_single()
477 while (iommu_is_span_boundary(index, nslots, offset_slots, in swiotlb_tbl_map_single()
479 index += stride; in swiotlb_tbl_map_single()
480 if (index >= io_tlb_nslabs) in swiotlb_tbl_map_single()
481 index = 0; in swiotlb_tbl_map_single()
482 if (index == wrap) in swiotlb_tbl_map_single()
[all …]
Dgenalloc.c549 unsigned long index; in gen_pool_best_fit() local
551 index = bitmap_find_next_zero_area(map, size, start, nr, 0); in gen_pool_best_fit()
553 while (index < size) { in gen_pool_best_fit()
554 int next_bit = find_next_bit(map, size, index + nr); in gen_pool_best_fit()
555 if ((next_bit - index) < len) { in gen_pool_best_fit()
556 len = next_bit - index; in gen_pool_best_fit()
557 start_bit = index; in gen_pool_best_fit()
561 index = bitmap_find_next_zero_area(map, size, in gen_pool_best_fit()
632 const char *propname, int index) in of_get_named_gen_pool() argument
637 np_pool = of_parse_phandle(np, propname, index); in of_get_named_gen_pool()
Dbitmap.c349 unsigned long index, end, i; in bitmap_find_next_zero_area_off() local
351 index = find_next_zero_bit(map, size, start); in bitmap_find_next_zero_area_off()
354 index = __ALIGN_MASK(index + align_offset, align_mask) - align_offset; in bitmap_find_next_zero_area_off()
356 end = index + nr; in bitmap_find_next_zero_area_off()
359 i = find_next_bit(map, end, index); in bitmap_find_next_zero_area_off()
364 return index; in bitmap_find_next_zero_area_off()
1061 int index; /* index first long of region in bitmap */ in __reg_op() local
1074 index = pos / BITS_PER_LONG; in __reg_op()
1075 offset = pos - (index * BITS_PER_LONG); in __reg_op()
1090 if (bitmap[index + i] & mask) in __reg_op()
[all …]
Dbtree.c679 unsigned long *key, size_t index, in __btree_for_each() argument
703 size_t index, void *func2) in empty() argument
708 size_t index, void *__func) in visitorl() argument
712 func(elem, opaque, *key, index); in visitorl()
717 size_t index, void *__func) in visitor32() argument
722 func(elem, opaque, *key, index); in visitor32()
727 size_t index, void *__func) in visitor64() argument
732 func(elem, opaque, *key, index); in visitor64()
737 size_t index, void *__func) in visitor128() argument
742 func(elem, opaque, key[0], key[1], index); in visitor128()
[all …]
Dlru_cache.c619 void lc_set(struct lru_cache *lc, unsigned int enr, int index) in lc_set() argument
624 if (index < 0 || index >= lc->nr_elements) in lc_set()
627 e = lc_element_by_index(lc, index); in lc_set()
Dvsprintf.c838 int index; in ip4_string() local
844 index = 0; in ip4_string()
847 index = 3; in ip4_string()
852 index = 3; in ip4_string()
858 index = 0; in ip4_string()
864 int digits = put_dec_trunc8(temp, addr[index]) - temp; in ip4_string()
876 index += step; in ip4_string()
1169 const u8 *index = be; in uuid_string() local
1176 index = le; in uuid_string()
1184 p = hex_byte_pack(p, addr[index[i]]); in uuid_string()
Ddma-debug.c352 struct dma_debug_entry *entry, index = *ref; in bucket_find_contain() local
356 entry = __hash_bucket_find(*bucket, &index, containing_match); in bucket_find_contain()
366 index.dev_addr -= (1 << HASH_FN_SHIFT); in bucket_find_contain()
367 *bucket = get_hash_bucket(&index, flags); in bucket_find_contain()
/lib/xz/
Dxz_dec_stream.c116 } index; member
284 s->index.size += in_used; in index_update()
307 switch (s->index.sequence) { in dec_index()
309 s->index.count = s->vli; in dec_index()
316 if (s->index.count != s->block.count) in dec_index()
319 s->index.sequence = SEQ_INDEX_UNPADDED; in dec_index()
323 s->index.hash.unpadded += s->vli; in dec_index()
324 s->index.sequence = SEQ_INDEX_UNCOMPRESSED; in dec_index()
328 s->index.hash.uncompressed += s->vli; in dec_index()
329 s->index.hash.crc32 = xz_crc32( in dec_index()
[all …]