/lib/ |
D | memcat_p.c | 13 int nr; in __memcat_p() local 16 for (nr = 0, p = a; *p; nr++, p++) in __memcat_p() 18 for (p = b; *p; nr++, p++) in __memcat_p() 21 nr++; in __memcat_p() 23 new = kmalloc_array(nr, sizeof(void *), GFP_KERNEL); in __memcat_p() 28 for (nr--; nr >= 0; nr--, p = p == b ? &a[nr] : p - 1) in __memcat_p() 29 new[nr] = *p; in __memcat_p()
|
D | sbitmap.c | 46 unsigned int nr) in update_alloc_hint_after_get() argument 48 if (nr == -1) { in update_alloc_hint_after_get() 51 } else if (nr == hint || unlikely(sb->round_robin)) { in update_alloc_hint_after_get() 53 hint = nr + 1; in update_alloc_hint_after_get() 149 int nr; in __sbitmap_get_word() local 155 nr = find_next_zero_bit(word, depth, hint); in __sbitmap_get_word() 156 if (unlikely(nr >= depth)) { in __sbitmap_get_word() 169 if (!test_and_set_bit_lock(nr, word)) in __sbitmap_get_word() 172 hint = nr + 1; in __sbitmap_get_word() 177 return nr; in __sbitmap_get_word() [all …]
|
D | genalloc.c | 85 bitmap_set_ll(unsigned long *map, unsigned long start, unsigned long nr) in bitmap_set_ll() argument 88 const unsigned long size = start + nr; in bitmap_set_ll() 92 while (nr >= bits_to_set) { in bitmap_set_ll() 94 return nr; in bitmap_set_ll() 95 nr -= bits_to_set; in bitmap_set_ll() 100 if (nr) { in bitmap_set_ll() 103 return nr; in bitmap_set_ll() 121 bitmap_clear_ll(unsigned long *map, unsigned long start, unsigned long nr) in bitmap_clear_ll() argument 124 const unsigned long size = start + nr; in bitmap_clear_ll() 128 while (nr >= bits_to_clear) { in bitmap_clear_ll() [all …]
|
D | iommu-helper.c | 10 unsigned long start, unsigned int nr, in iommu_area_alloc() argument 19 index = bitmap_find_next_zero_area(map, size, start, nr, align_mask); in iommu_area_alloc() 21 if (iommu_is_span_boundary(index, nr, shift, boundary_size)) { in iommu_area_alloc() 25 bitmap_set(map, index, nr); in iommu_area_alloc()
|
D | test_min_heap.c | 42 while (heap->nr > 0) { in pop_verify_heap() 68 .nr = ARRAY_SIZE(values), in test_heapify_all() 84 heap.nr = ARRAY_SIZE(values); in test_heapify_all() 85 for (i = 0; i < heap.nr; i++) in test_heapify_all() 101 .nr = 0, in test_heap_push() 118 while (heap.nr < heap.size) { in test_heap_push() 134 .nr = 0, in test_heap_pop_push() 155 heap.nr = 0; in test_heap_pop_push()
|
D | syscall.c | 16 info->data.nr = -1; in collect_syscall() 29 info->data.nr = syscall_get_nr(target, regs); in collect_syscall() 30 if (info->data.nr != -1L) in collect_syscall()
|
D | locking-selftest.c | 1456 #define DO_TESTCASE_1(desc, name, nr) \ argument 1457 print_testname(desc"/"#nr); \ 1458 dotest(name##_##nr, SUCCESS, LOCKTYPE_RWLOCK); \ 1461 #define DO_TESTCASE_1B(desc, name, nr) \ argument 1462 print_testname(desc"/"#nr); \ 1463 dotest(name##_##nr, FAILURE, LOCKTYPE_RWLOCK); \ 1466 #define DO_TESTCASE_1RR(desc, name, nr) \ argument 1467 print_testname(desc"/"#nr); \ 1469 dotest(name##_##nr, SUCCESS, LOCKTYPE_RWLOCK); \ 1472 #define DO_TESTCASE_1RRB(desc, name, nr) \ argument [all …]
|
D | test_kasan.c | 988 static void kasan_bitops_modify(struct kunit *test, int nr, void *addr) in kasan_bitops_modify() argument 990 KUNIT_EXPECT_KASAN_FAIL(test, set_bit(nr, addr)); in kasan_bitops_modify() 991 KUNIT_EXPECT_KASAN_FAIL(test, __set_bit(nr, addr)); in kasan_bitops_modify() 992 KUNIT_EXPECT_KASAN_FAIL(test, clear_bit(nr, addr)); in kasan_bitops_modify() 993 KUNIT_EXPECT_KASAN_FAIL(test, __clear_bit(nr, addr)); in kasan_bitops_modify() 994 KUNIT_EXPECT_KASAN_FAIL(test, clear_bit_unlock(nr, addr)); in kasan_bitops_modify() 995 KUNIT_EXPECT_KASAN_FAIL(test, __clear_bit_unlock(nr, addr)); in kasan_bitops_modify() 996 KUNIT_EXPECT_KASAN_FAIL(test, change_bit(nr, addr)); in kasan_bitops_modify() 997 KUNIT_EXPECT_KASAN_FAIL(test, __change_bit(nr, addr)); in kasan_bitops_modify() 1000 static void kasan_bitops_test_and_modify(struct kunit *test, int nr, void *addr) in kasan_bitops_test_and_modify() argument [all …]
|
D | radix-tree.c | 262 if (rtp->nr) { in radix_tree_node_alloc() 265 rtp->nr--; in radix_tree_node_alloc() 320 static __must_check int __radix_tree_preload(gfp_t gfp_mask, unsigned nr) in __radix_tree_preload() argument 334 while (rtp->nr < nr) { in __radix_tree_preload() 341 if (rtp->nr < nr) { in __radix_tree_preload() 344 rtp->nr++; in __radix_tree_preload() 1583 while (rtp->nr) { in radix_tree_cpu_dead() 1587 rtp->nr--; in radix_tree_cpu_dead()
|
D | bitmap.c | 260 unsigned int nr = BITS_TO_LONGS(bits); in __bitmap_or() local 262 for (k = 0; k < nr; k++) in __bitmap_or() 271 unsigned int nr = BITS_TO_LONGS(bits); in __bitmap_xor() local 273 for (k = 0; k < nr; k++) in __bitmap_xor() 299 unsigned int nr = BITS_TO_LONGS(nbits); in __bitmap_replace() local 301 for (k = 0; k < nr; k++) in __bitmap_replace() 409 unsigned int nr, in bitmap_find_next_zero_area_off() argument 420 end = index + nr; in bitmap_find_next_zero_area_off()
|
D | percpu_counter.c | 189 int nr = num_online_cpus(); in compute_batch_value() local 191 percpu_counter_batch = max(32, nr*2); in compute_batch_value()
|
D | devres.c | 291 unsigned int nr) in devm_ioport_map() argument 299 addr = ioport_map(port, nr); in devm_ioport_map()
|
D | iov_iter.c | 1444 unsigned nr, offset; in iter_xarray_get_pages() local 1469 nr = iter_xarray_populate_pages(pages, i->xarray, index, count); in iter_xarray_get_pages() 1470 if (nr == 0) in iter_xarray_get_pages() 1473 return min_t(size_t, nr * PAGE_SIZE - offset, maxsize); in iter_xarray_get_pages() 1606 unsigned nr, offset; in iter_xarray_get_pages_alloc() local 1633 nr = iter_xarray_populate_pages(p, i->xarray, index, count); in iter_xarray_get_pages_alloc() 1634 if (nr == 0) in iter_xarray_get_pages_alloc() 1637 return min_t(size_t, nr * PAGE_SIZE - offset, maxsize); in iter_xarray_get_pages_alloc()
|
D | test_xarray.c | 1417 unsigned long nr = 1UL << order; in check_create_range_2() local 1419 for (i = 0; i < nr * nr; i += nr) in check_create_range_2() 1421 for (i = 0; i < nr * nr; i++) in check_create_range_2()
|
D | iomap.c | 362 void __iomem *ioport_map(unsigned long port, unsigned int nr) in ioport_map() argument
|