/lib/ |
D | memcat_p.c | 13 int nr; in __memcat_p() local 16 for (nr = 0, p = a; *p; nr++, p++) in __memcat_p() 18 for (p = b; *p; nr++, p++) in __memcat_p() 21 nr++; in __memcat_p() 23 new = kmalloc_array(nr, sizeof(void *), GFP_KERNEL); in __memcat_p() 28 for (nr--; nr >= 0; nr--, p = p == b ? &a[nr] : p - 1) in __memcat_p() 29 new[nr] = *p; in __memcat_p()
|
D | sbitmap.c | 46 unsigned int nr) in update_alloc_hint_after_get() argument 48 if (nr == -1) { in update_alloc_hint_after_get() 51 } else if (nr == hint || unlikely(sb->round_robin)) { in update_alloc_hint_after_get() 53 hint = nr + 1; in update_alloc_hint_after_get() 139 int nr; in __sbitmap_get_word() local 145 nr = find_next_zero_bit(word, depth, hint); in __sbitmap_get_word() 146 if (unlikely(nr >= depth)) { in __sbitmap_get_word() 159 if (!test_and_set_bit_lock(nr, word)) in __sbitmap_get_word() 162 hint = nr + 1; in __sbitmap_get_word() 167 return nr; in __sbitmap_get_word() [all …]
|
D | genalloc.c | 85 bitmap_set_ll(unsigned long *map, unsigned long start, unsigned long nr) in bitmap_set_ll() argument 88 const unsigned long size = start + nr; in bitmap_set_ll() 92 while (nr >= bits_to_set) { in bitmap_set_ll() 94 return nr; in bitmap_set_ll() 95 nr -= bits_to_set; in bitmap_set_ll() 100 if (nr) { in bitmap_set_ll() 103 return nr; in bitmap_set_ll() 121 bitmap_clear_ll(unsigned long *map, unsigned long start, unsigned long nr) in bitmap_clear_ll() argument 124 const unsigned long size = start + nr; in bitmap_clear_ll() 128 while (nr >= bits_to_clear) { in bitmap_clear_ll() [all …]
|
D | iommu-helper.c | 10 unsigned long start, unsigned int nr, in iommu_area_alloc() argument 19 index = bitmap_find_next_zero_area(map, size, start, nr, align_mask); in iommu_area_alloc() 21 if (iommu_is_span_boundary(index, nr, shift, boundary_size)) { in iommu_area_alloc() 25 bitmap_set(map, index, nr); in iommu_area_alloc()
|
D | test_min_heap.c | 42 while (heap->nr > 0) { in pop_verify_heap() 68 .nr = ARRAY_SIZE(values), in test_heapify_all() 84 heap.nr = ARRAY_SIZE(values); in test_heapify_all() 85 for (i = 0; i < heap.nr; i++) in test_heapify_all() 101 .nr = 0, in test_heap_push() 118 while (heap.nr < heap.size) { in test_heap_push() 134 .nr = 0, in test_heap_pop_push() 155 heap.nr = 0; in test_heap_pop_push()
|
D | flex_proportions.c | 217 long nr) in __fprop_add_percpu() argument 220 percpu_counter_add_batch(&pl->events, nr, PROP_BATCH); in __fprop_add_percpu() 221 percpu_counter_add(&p->events, nr); in __fprop_add_percpu() 257 struct fprop_local_percpu *pl, int max_frac, long nr) in __fprop_add_percpu_max() argument 270 } else if (tmp < nr * (FPROP_FRAC_BASE - max_frac)) { in __fprop_add_percpu_max() 272 nr = div_u64(tmp + FPROP_FRAC_BASE - max_frac - 1, in __fprop_add_percpu_max() 277 __fprop_add_percpu(p, pl, nr); in __fprop_add_percpu_max()
|
D | syscall.c | 16 info->data.nr = -1; in collect_syscall() 29 info->data.nr = syscall_get_nr(target, regs); in collect_syscall() 30 if (info->data.nr != -1L) in collect_syscall()
|
D | locking-selftest.c | 1505 #define DO_TESTCASE_1(desc, name, nr) \ argument 1506 print_testname(desc"/"#nr); \ 1507 dotest(name##_##nr, SUCCESS, LOCKTYPE_RWLOCK); \ 1510 #define DO_TESTCASE_1B(desc, name, nr) \ argument 1511 print_testname(desc"/"#nr); \ 1512 dotest(name##_##nr, FAILURE, LOCKTYPE_RWLOCK); \ 1515 #define DO_TESTCASE_1RR(desc, name, nr) \ argument 1516 print_testname(desc"/"#nr); \ 1518 dotest(name##_##nr, SUCCESS, LOCKTYPE_RWLOCK); \ 1521 #define DO_TESTCASE_1RRB(desc, name, nr) \ argument [all …]
|
D | find_bit.c | 73 unsigned long sz = (size), nr = (num), idx, w, tmp; \ 76 if (idx * BITS_PER_LONG + nr >= sz) \ 81 if (w > nr) \ 84 nr -= w; \ 90 sz = min(idx * BITS_PER_LONG + fns(tmp, nr), sz); \
|
D | radix-tree.c | 264 if (rtp->nr) { in radix_tree_node_alloc() 267 rtp->nr--; in radix_tree_node_alloc() 322 static __must_check int __radix_tree_preload(gfp_t gfp_mask, unsigned nr) in __radix_tree_preload() argument 336 while (rtp->nr < nr) { in __radix_tree_preload() 343 if (rtp->nr < nr) { in __radix_tree_preload() 346 rtp->nr++; in __radix_tree_preload() 1585 while (rtp->nr) { in radix_tree_cpu_dead() 1589 rtp->nr--; in radix_tree_cpu_dead()
|
D | bitmap.c | 260 unsigned int nr = BITS_TO_LONGS(bits); in __bitmap_or() local 262 for (k = 0; k < nr; k++) in __bitmap_or() 271 unsigned int nr = BITS_TO_LONGS(bits); in __bitmap_xor() local 273 for (k = 0; k < nr; k++) in __bitmap_xor() 299 unsigned int nr = BITS_TO_LONGS(nbits); in __bitmap_replace() local 301 for (k = 0; k < nr; k++) in __bitmap_replace() 420 unsigned int nr, in bitmap_find_next_zero_area_off() argument 431 end = index + nr; in bitmap_find_next_zero_area_off()
|
D | percpu_counter.c | 226 int nr = num_online_cpus(); in compute_batch_value() local 228 percpu_counter_batch = max(32, nr*2); in compute_batch_value()
|
D | iov_iter.c | 1036 unsigned nr, offset, count; in iter_xarray_get_pages() local 1048 nr = iter_xarray_populate_pages(*pages, i->xarray, index, count); in iter_xarray_get_pages() 1049 if (nr == 0) in iter_xarray_get_pages() 1052 maxsize = min_t(size_t, nr * PAGE_SIZE - offset, maxsize); in iter_xarray_get_pages() 1611 unsigned int nr = 0, offset; in iov_iter_extract_xarray_pages() local 1635 p[nr++] = find_subpage(page, xas.xa_index); in iov_iter_extract_xarray_pages() 1636 if (nr == maxpages) in iov_iter_extract_xarray_pages() 1641 maxsize = min_t(size_t, nr * PAGE_SIZE - offset, maxsize); in iov_iter_extract_xarray_pages()
|
D | devres.c | 277 unsigned int nr) in devm_ioport_map() argument 286 addr = ioport_map(port, nr); in devm_ioport_map()
|
D | test_xarray.c | 1417 unsigned long nr = 1UL << order; in check_create_range_2() local 1419 for (i = 0; i < nr * nr; i += nr) in check_create_range_2() 1421 for (i = 0; i < nr * nr; i++) in check_create_range_2()
|
D | iomap.c | 406 void __iomem *ioport_map(unsigned long port, unsigned int nr) in ioport_map() argument
|