/mm/ |
D | slab.c | 246 struct kmem_cache_node *n, int tofree); 698 struct kmem_cache_node *n = get_node(cachep, numa_mem_id()); in recheck_pfmemalloc_active() local 705 spin_lock_irqsave(&n->list_lock, flags); in recheck_pfmemalloc_active() 706 list_for_each_entry(page, &n->slabs_full, lru) in recheck_pfmemalloc_active() 710 list_for_each_entry(page, &n->slabs_partial, lru) in recheck_pfmemalloc_active() 714 list_for_each_entry(page, &n->slabs_free, lru) in recheck_pfmemalloc_active() 720 spin_unlock_irqrestore(&n->list_lock, flags); in recheck_pfmemalloc_active() 731 struct kmem_cache_node *n; in __ac_get_obj() local 753 n = get_node(cachep, numa_mem_id()); in __ac_get_obj() 754 if (!list_empty(&n->slabs_free) && force_refill) { in __ac_get_obj() [all …]
|
D | usercopy.c | 80 static bool overlaps(const void *ptr, unsigned long n, unsigned long low, in overlaps() argument 84 unsigned long check_high = check_low + n; in overlaps() 95 unsigned long n) in check_kernel_text_object() argument 101 if (overlaps(ptr, n, textlow, texthigh)) in check_kernel_text_object() 119 if (overlaps(ptr, n, textlow_linear, texthigh_linear)) in check_kernel_text_object() 125 static inline const char *check_bogus_address(const void *ptr, unsigned long n) in check_bogus_address() argument 128 if ((unsigned long)ptr + n < (unsigned long)ptr) in check_bogus_address() 139 static inline const char *check_page_span(const void *ptr, unsigned long n, in check_page_span() argument 143 const void *end = ptr + n - 1; in check_page_span() 202 static inline const char *check_heap_object(const void *ptr, unsigned long n, in check_heap_object() argument [all …]
|
D | slub.c | 359 const char *n) in __cmpxchg_double_slab() argument 387 pr_info("%s %s: cmpxchg double redo ", n, s->name); in __cmpxchg_double_slab() 396 const char *n) in cmpxchg_double_slab() argument 428 pr_info("%s %s: cmpxchg double redo ", n, s->name); in cmpxchg_double_slab() 977 struct kmem_cache_node *n, struct page *page) in add_full() argument 982 lockdep_assert_held(&n->list_lock); in add_full() 983 list_add(&page->lru, &n->full); in add_full() 986 static void remove_full(struct kmem_cache *s, struct kmem_cache_node *n, struct page *page) in remove_full() argument 991 lockdep_assert_held(&n->list_lock); in remove_full() 998 struct kmem_cache_node *n = get_node(s, node); in slabs_node() local [all …]
|
D | Makefile | 5 KASAN_SANITIZE_slab_common.o := n 6 KASAN_SANITIZE_slub.o := n 11 KCOV_INSTRUMENT_slab_common.o := n 12 KCOV_INSTRUMENT_slob.o := n 13 KCOV_INSTRUMENT_slab.o := n 14 KCOV_INSTRUMENT_slub.o := n 15 KCOV_INSTRUMENT_page_alloc.o := n 16 KCOV_INSTRUMENT_debug-pagealloc.o := n 17 KCOV_INSTRUMENT_kmemleak.o := n 18 KCOV_INSTRUMENT_kmemcheck.o := n [all …]
|
D | vmalloc.c | 289 struct rb_node *n = vmap_area_root.rb_node; in __find_vmap_area() local 291 while (n) { in __find_vmap_area() 294 va = rb_entry(n, struct vmap_area, rb_node); in __find_vmap_area() 296 n = n->rb_left; in __find_vmap_area() 298 n = n->rb_right; in __find_vmap_area() 350 struct rb_node *n; in alloc_vmap_area() local 407 n = vmap_area_root.rb_node; in alloc_vmap_area() 410 while (n) { in alloc_vmap_area() 412 tmp = rb_entry(n, struct vmap_area, rb_node); in alloc_vmap_area() 417 n = n->rb_left; in alloc_vmap_area() [all …]
|
D | iov_iter.c | 412 size_t n; in alignment_iovec() local 418 n = iov->iov_len - i->iov_offset; in alignment_iovec() 419 if (n >= size) in alignment_iovec() 421 size -= n; in alignment_iovec() 422 res |= n; in alignment_iovec() 454 int n; in get_pages_iovec() local 467 n = (len + PAGE_SIZE - 1) / PAGE_SIZE; in get_pages_iovec() 468 res = get_user_pages_fast(addr, n, (i->type & WRITE) != WRITE, pages); in get_pages_iovec() 471 return (res == n ? len : res * PAGE_SIZE) - *start; in get_pages_iovec() 483 int n; in get_pages_alloc_iovec() local [all …]
|
D | mempolicy.c | 2172 struct rb_node *n = sp->root.rb_node; in sp_lookup() local 2174 while (n) { in sp_lookup() 2175 struct sp_node *p = rb_entry(n, struct sp_node, nd); in sp_lookup() 2178 n = n->rb_right; in sp_lookup() 2180 n = n->rb_left; in sp_lookup() 2184 if (!n) in sp_lookup() 2188 struct rb_node *prev = rb_prev(n); in sp_lookup() 2194 n = prev; in sp_lookup() 2196 return rb_entry(n, struct sp_node, nd); in sp_lookup() 2242 static void sp_free(struct sp_node *n) in sp_free() argument [all …]
|
D | debug-pagealloc.c | 33 static void poison_pages(struct page *page, int n) in poison_pages() argument 37 for (i = 0; i < n; i++) in poison_pages() 88 static void unpoison_pages(struct page *page, int n) in unpoison_pages() argument 92 for (i = 0; i < n; i++) in unpoison_pages()
|
D | slab_common.c | 750 char *n; in create_kmalloc_caches() local 753 n = kasprintf(GFP_NOWAIT, "kmalloc-%d", kmalloc_size(i)); in create_kmalloc_caches() 755 BUG_ON(!n); in create_kmalloc_caches() 756 s->name = n; in create_kmalloc_caches() 766 char *n = kasprintf(GFP_NOWAIT, in create_kmalloc_caches() local 769 BUG_ON(!n); in create_kmalloc_caches() 770 kmalloc_dma_caches[i] = create_kmalloc_cache(n, in create_kmalloc_caches() 839 loff_t n = *pos; in s_start() local 842 if (!n) in s_start()
|
D | vmstat.c | 321 long o, n, t, z; in mod_state() local 339 n = delta + o; in mod_state() 341 if (n > t || n < -t) { in mod_state() 345 z = n + os; in mod_state() 346 n = -os; in mod_state() 348 } while (this_cpu_cmpxchg(*p, o, n) != o); in mod_state()
|
D | util.c | 117 char *strndup_user(const char __user *s, long n) in strndup_user() argument 122 length = strnlen_user(s, n); in strndup_user() 127 if (length > n) in strndup_user()
|
D | Kconfig | 158 default n 183 def_bool n 297 may say n to override this. 457 default n 481 default n 508 If unsure, say "n". 547 default n 564 default n 571 default n 582 default n
|
D | list_lru.c | 72 struct list_head *item, *n; in list_lru_walk_node() local 77 list_for_each_safe(item, n, &nlru->list) { in list_lru_walk_node()
|
D | swapfile.c | 229 unsigned int n) in cluster_set_next() argument 231 info->data = n; in cluster_set_next() 235 unsigned int n, unsigned int f) in cluster_set_next_flag() argument 238 info->data = n; in cluster_set_next_flag() 883 int count, tmp_count, n; in swp_swapcount() local 898 n = SWAP_MAP_MAX + 1; in swp_swapcount() 911 count += (tmp_count & ~COUNT_CONTINUED) * n; in swp_swapcount() 912 n *= (SWAP_CONT_MAX + 1); in swp_swapcount() 1095 unsigned int n = 0; in count_swap_pages() local 1103 n = sis->pages; in count_swap_pages() [all …]
|
D | nommu.c | 1153 unsigned long total, point, n; in do_mmap_private() local 1199 n = 1 << order; in do_mmap_private() 1200 kdebug("shave %lu/%lu @%lu", n, total - point, total); in do_mmap_private() 1201 atomic_long_sub(n, &mmap_pages_allocated); in do_mmap_private() 1202 total -= n; in do_mmap_private()
|
D | page_alloc.c | 1867 int n; /* node that zone *z is on */ in zlc_zone_worth_trying() local 1874 n = zlc->z_to_n[i]; in zlc_zone_worth_trying() 1877 return node_isset(n, *allowednodes) && !test_bit(i, zlc->fullzones); in zlc_zone_worth_trying() 3546 int n, val; in find_next_best_node() local 3557 for_each_node_state(n, N_MEMORY) { in find_next_best_node() 3560 if (node_isset(n, *used_node_mask)) in find_next_best_node() 3564 val = node_distance(node, n); in find_next_best_node() 3567 val += (n < node); in find_next_best_node() 3570 tmp = cpumask_of_node(n); in find_next_best_node() 3576 val += node_load[n]; in find_next_best_node() [all …]
|
D | zswap.c | 804 struct zswap_entry *entry, *n; in zswap_frontswap_invalidate_area() local 811 rbtree_postorder_for_each_entry_safe(entry, n, &tree->rbroot, rbnode) in zswap_frontswap_invalidate_area()
|
D | hugetlb.c | 1559 static char * __init memfmt(char *buf, unsigned long n) in memfmt() argument 1561 if (n >= (1UL << 30)) in memfmt() 1562 sprintf(buf, "%lu GB", n >> 30); in memfmt() 1563 else if (n >= (1UL << 20)) in memfmt() 1564 sprintf(buf, "%lu MB", n >> 20); in memfmt() 1566 sprintf(buf, "%lu KB", n >> 10); in memfmt()
|
D | filemap.c | 645 int n; in __page_cache_alloc() local 652 n = cpuset_mem_spread_node(); in __page_cache_alloc() 653 page = alloc_pages_exact_node(n, gfp, 0); in __page_cache_alloc()
|
D | kmemleak.c | 1503 loff_t n = *pos; in kmemleak_seq_start() local 1512 if (n-- > 0) in kmemleak_seq_start()
|
/mm/kasan/ |
D | Makefile | 1 KASAN_SANITIZE := n 2 UBSAN_SANITIZE_kasan.o := n 3 KCOV_INSTRUMENT := n
|