Home
last modified time | relevance | path

Searched refs:nr (Results 1 – 15 of 15) sorted by relevance

/lib/
Dmemcat_p.c13 int nr; in __memcat_p() local
16 for (nr = 0, p = a; *p; nr++, p++) in __memcat_p()
18 for (p = b; *p; nr++, p++) in __memcat_p()
21 nr++; in __memcat_p()
23 new = kmalloc_array(nr, sizeof(void *), GFP_KERNEL); in __memcat_p()
28 for (nr--; nr >= 0; nr--, p = p == b ? &a[nr] : p - 1) in __memcat_p()
29 new[nr] = *p; in __memcat_p()
Dsbitmap.c46 unsigned int nr) in update_alloc_hint_after_get() argument
48 if (nr == -1) { in update_alloc_hint_after_get()
51 } else if (nr == hint || unlikely(sb->round_robin)) { in update_alloc_hint_after_get()
53 hint = nr + 1; in update_alloc_hint_after_get()
149 int nr; in __sbitmap_get_word() local
155 nr = find_next_zero_bit(word, depth, hint); in __sbitmap_get_word()
156 if (unlikely(nr >= depth)) { in __sbitmap_get_word()
169 if (!test_and_set_bit_lock(nr, word)) in __sbitmap_get_word()
172 hint = nr + 1; in __sbitmap_get_word()
177 return nr; in __sbitmap_get_word()
[all …]
Dgenalloc.c85 bitmap_set_ll(unsigned long *map, unsigned long start, unsigned long nr) in bitmap_set_ll() argument
88 const unsigned long size = start + nr; in bitmap_set_ll()
92 while (nr >= bits_to_set) { in bitmap_set_ll()
94 return nr; in bitmap_set_ll()
95 nr -= bits_to_set; in bitmap_set_ll()
100 if (nr) { in bitmap_set_ll()
103 return nr; in bitmap_set_ll()
121 bitmap_clear_ll(unsigned long *map, unsigned long start, unsigned long nr) in bitmap_clear_ll() argument
124 const unsigned long size = start + nr; in bitmap_clear_ll()
128 while (nr >= bits_to_clear) { in bitmap_clear_ll()
[all …]
Diommu-helper.c10 unsigned long start, unsigned int nr, in iommu_area_alloc() argument
19 index = bitmap_find_next_zero_area(map, size, start, nr, align_mask); in iommu_area_alloc()
21 if (iommu_is_span_boundary(index, nr, shift, boundary_size)) { in iommu_area_alloc()
25 bitmap_set(map, index, nr); in iommu_area_alloc()
Dtest_min_heap.c42 while (heap->nr > 0) { in pop_verify_heap()
68 .nr = ARRAY_SIZE(values), in test_heapify_all()
84 heap.nr = ARRAY_SIZE(values); in test_heapify_all()
85 for (i = 0; i < heap.nr; i++) in test_heapify_all()
101 .nr = 0, in test_heap_push()
118 while (heap.nr < heap.size) { in test_heap_push()
134 .nr = 0, in test_heap_pop_push()
155 heap.nr = 0; in test_heap_pop_push()
Dsyscall.c16 info->data.nr = -1; in collect_syscall()
29 info->data.nr = syscall_get_nr(target, regs); in collect_syscall()
30 if (info->data.nr != -1L) in collect_syscall()
Dlocking-selftest.c1456 #define DO_TESTCASE_1(desc, name, nr) \ argument
1457 print_testname(desc"/"#nr); \
1458 dotest(name##_##nr, SUCCESS, LOCKTYPE_RWLOCK); \
1461 #define DO_TESTCASE_1B(desc, name, nr) \ argument
1462 print_testname(desc"/"#nr); \
1463 dotest(name##_##nr, FAILURE, LOCKTYPE_RWLOCK); \
1466 #define DO_TESTCASE_1RR(desc, name, nr) \ argument
1467 print_testname(desc"/"#nr); \
1469 dotest(name##_##nr, SUCCESS, LOCKTYPE_RWLOCK); \
1472 #define DO_TESTCASE_1RRB(desc, name, nr) \ argument
[all …]
Dtest_kasan.c988 static void kasan_bitops_modify(struct kunit *test, int nr, void *addr) in kasan_bitops_modify() argument
990 KUNIT_EXPECT_KASAN_FAIL(test, set_bit(nr, addr)); in kasan_bitops_modify()
991 KUNIT_EXPECT_KASAN_FAIL(test, __set_bit(nr, addr)); in kasan_bitops_modify()
992 KUNIT_EXPECT_KASAN_FAIL(test, clear_bit(nr, addr)); in kasan_bitops_modify()
993 KUNIT_EXPECT_KASAN_FAIL(test, __clear_bit(nr, addr)); in kasan_bitops_modify()
994 KUNIT_EXPECT_KASAN_FAIL(test, clear_bit_unlock(nr, addr)); in kasan_bitops_modify()
995 KUNIT_EXPECT_KASAN_FAIL(test, __clear_bit_unlock(nr, addr)); in kasan_bitops_modify()
996 KUNIT_EXPECT_KASAN_FAIL(test, change_bit(nr, addr)); in kasan_bitops_modify()
997 KUNIT_EXPECT_KASAN_FAIL(test, __change_bit(nr, addr)); in kasan_bitops_modify()
1000 static void kasan_bitops_test_and_modify(struct kunit *test, int nr, void *addr) in kasan_bitops_test_and_modify() argument
[all …]
Dradix-tree.c262 if (rtp->nr) { in radix_tree_node_alloc()
265 rtp->nr--; in radix_tree_node_alloc()
320 static __must_check int __radix_tree_preload(gfp_t gfp_mask, unsigned nr) in __radix_tree_preload() argument
334 while (rtp->nr < nr) { in __radix_tree_preload()
341 if (rtp->nr < nr) { in __radix_tree_preload()
344 rtp->nr++; in __radix_tree_preload()
1583 while (rtp->nr) { in radix_tree_cpu_dead()
1587 rtp->nr--; in radix_tree_cpu_dead()
Dbitmap.c260 unsigned int nr = BITS_TO_LONGS(bits); in __bitmap_or() local
262 for (k = 0; k < nr; k++) in __bitmap_or()
271 unsigned int nr = BITS_TO_LONGS(bits); in __bitmap_xor() local
273 for (k = 0; k < nr; k++) in __bitmap_xor()
299 unsigned int nr = BITS_TO_LONGS(nbits); in __bitmap_replace() local
301 for (k = 0; k < nr; k++) in __bitmap_replace()
409 unsigned int nr, in bitmap_find_next_zero_area_off() argument
420 end = index + nr; in bitmap_find_next_zero_area_off()
Dpercpu_counter.c189 int nr = num_online_cpus(); in compute_batch_value() local
191 percpu_counter_batch = max(32, nr*2); in compute_batch_value()
Ddevres.c291 unsigned int nr) in devm_ioport_map() argument
299 addr = ioport_map(port, nr); in devm_ioport_map()
Diov_iter.c1444 unsigned nr, offset; in iter_xarray_get_pages() local
1469 nr = iter_xarray_populate_pages(pages, i->xarray, index, count); in iter_xarray_get_pages()
1470 if (nr == 0) in iter_xarray_get_pages()
1473 return min_t(size_t, nr * PAGE_SIZE - offset, maxsize); in iter_xarray_get_pages()
1606 unsigned nr, offset; in iter_xarray_get_pages_alloc() local
1633 nr = iter_xarray_populate_pages(p, i->xarray, index, count); in iter_xarray_get_pages_alloc()
1634 if (nr == 0) in iter_xarray_get_pages_alloc()
1637 return min_t(size_t, nr * PAGE_SIZE - offset, maxsize); in iter_xarray_get_pages_alloc()
Dtest_xarray.c1417 unsigned long nr = 1UL << order; in check_create_range_2() local
1419 for (i = 0; i < nr * nr; i += nr) in check_create_range_2()
1421 for (i = 0; i < nr * nr; i++) in check_create_range_2()
Diomap.c362 void __iomem *ioport_map(unsigned long port, unsigned int nr) in ioport_map() argument