/mm/ |
D | cma_debug.c | 24 static int cma_debugfs_get(void *data, u64 *val) in cma_debugfs_get() argument 26 unsigned long *p = data; in cma_debugfs_get() 34 static int cma_used_get(void *data, u64 *val) in cma_used_get() argument 36 struct cma *cma = data; in cma_used_get() 49 static int cma_maxchunk_get(void *data, u64 *val) in cma_maxchunk_get() argument 51 struct cma *cma = data; in cma_maxchunk_get() 122 static int cma_free_write(void *data, u64 val) in cma_free_write() argument 125 struct cma *cma = data; in cma_free_write() 154 static int cma_alloc_write(void *data, u64 val) in cma_alloc_write() argument 157 struct cma *cma = data; in cma_alloc_write()
|
D | pgsize_migration.c | 286 struct vm_area_struct *data; in get_data_vma() local 291 data = kzalloc(sizeof(struct vm_area_struct), GFP_KERNEL); in get_data_vma() 293 *data = *vma; in get_data_vma() 296 data->vm_end = VMA_PAD_START(data); in get_data_vma() 298 return data; in get_data_vma()
|
D | percpu-km.c | 74 chunk->data = pages; in pcpu_create_chunk() 97 if (chunk->data) in pcpu_destroy_chunk() 98 __free_pages(chunk->data, order_base_2(nr_pages)); in pcpu_destroy_chunk()
|
D | page_pinner.c | 360 static int failure_tracking_set(void *data, u64 val) in failure_tracking_set() argument 372 static int failure_tracking_get(void *data, u64 *val) in failure_tracking_get() argument 381 static int buffer_size_set(void *data, u64 val) in buffer_size_set() argument 401 static int buffer_size_get(void *data, u64 *val) in buffer_size_get() argument
|
D | hwpoison-inject.c | 14 static int hwpoison_inject(void *data, u64 val) in hwpoison_inject() argument 55 static int hwpoison_unpoison(void *data, u64 val) in hwpoison_unpoison() argument
|
D | percpu-vm.c | 349 chunk->data = vms; in pcpu_create_chunk() 366 if (chunk->data) in pcpu_destroy_chunk() 367 pcpu_free_vm_areas(chunk->data, pcpu_nr_groups); in pcpu_destroy_chunk()
|
D | dmapool.c | 349 u8 *data = retval; in dma_pool_alloc() local 352 if (data[i] == POOL_POISON_FREED) in dma_pool_alloc() 366 data, pool->size, 1); in dma_pool_alloc()
|
D | cma.c | 631 int cma_for_each_area(int (*it)(struct cma *cma, void *data), void *data) in cma_for_each_area() argument 636 int ret = it(&cma_areas[i], data); in cma_for_each_area()
|
D | memory.c | 2584 pte_fn_t fn, void *data, bool create, in apply_to_pte_range() argument 2610 err = fn(pte++, addr, data); in apply_to_pte_range() 2627 pte_fn_t fn, void *data, bool create, in apply_to_pmd_range() argument 2655 fn, data, create, mask); in apply_to_pmd_range() 2665 pte_fn_t fn, void *data, bool create, in apply_to_pud_range() argument 2691 fn, data, create, mask); in apply_to_pud_range() 2701 pte_fn_t fn, void *data, bool create, in apply_to_p4d_range() argument 2727 fn, data, create, mask); in apply_to_p4d_range() 2737 void *data, bool create) in __apply_to_page_range() argument 2761 fn, data, create, &mask); in __apply_to_page_range() [all …]
|
D | memremap.c | 161 static void devm_memremap_pages_release(void *data) in devm_memremap_pages_release() argument 163 memunmap_pages(data); in devm_memremap_pages_release()
|
D | readahead.c | 91 int (*filler)(void *, struct page *), void *data) in read_cache_pages() argument 106 ret = filler(data, page); in read_cache_pages()
|
D | percpu-internal.h | 49 void *data; /* chunk data */ member
|
D | vmalloc.c | 2810 struct vmap_pfn_data *data = private; in vmap_pfn_apply() local 2812 if (WARN_ON_ONCE(pfn_valid(data->pfns[data->idx]))) in vmap_pfn_apply() 2814 *pte = pte_mkspecial(pfn_pte(data->pfns[data->idx++], data->prot)); in vmap_pfn_apply() 2829 struct vmap_pfn_data data = { .pfns = pfns, .prot = pgprot_nx(prot) }; in vmap_pfn() local 2837 count * PAGE_SIZE, vmap_pfn_apply, &data)) { in vmap_pfn()
|
D | swapfile.c | 287 return info->data; in cluster_count() 293 info->data = c; in cluster_set_count() 300 info->data = c; in cluster_set_count_flag() 305 return info->data; in cluster_next() 311 info->data = n; in cluster_set_next() 318 info->data = n; in cluster_set_next_flag() 334 info->data = 0; in cluster_set_null()
|
D | page-writeback.c | 2179 void *data) in write_cache_pages() argument 2255 error = (*writepage)(page, wbc, data); in write_cache_pages() 2317 void *data) in __writepage() argument 2319 struct address_space *mapping = data; in __writepage()
|
D | slub.c | 549 (*(int *)resource->data)++; in slab_add_kunit_errors() 921 static void restore_bytes(struct kmem_cache *s, char *message, u8 data, in restore_bytes() argument 924 slab_fix(s, "Restoring %s 0x%p-0x%p=0x%x", message, from, to - 1, data); in restore_bytes() 925 memset(from, data, to - from); in restore_bytes() 5672 int *data = kmalloc_array(nr_cpu_ids, sizeof(int), GFP_KERNEL); in show_stat() local 5674 if (!data) in show_stat() 5680 data[cpu] = x; in show_stat() 5688 if (data[cpu]) in show_stat() 5690 cpu, data[cpu]); in show_stat() 5693 kfree(data); in show_stat()
|
D | compaction.c | 1725 unsigned long data) in compaction_alloc() argument 1727 struct compact_control_ext *cc_ext = (struct compact_control_ext *)data; in compaction_alloc() 1750 static void compaction_free(struct page *page, unsigned long data) in compaction_free() argument 1752 struct compact_control_ext *cc_ext = (struct compact_control_ext *)data; in compaction_free()
|
D | migrate.c | 2094 unsigned long data) in alloc_misplaced_dst_page() argument 2096 int nid = (int) data; in alloc_misplaced_dst_page() 2109 unsigned long data) in alloc_misplaced_dst_page_thp() argument 2111 int nid = (int) data; in alloc_misplaced_dst_page_thp()
|
D | filemap.c | 3513 void *data, in do_read_cache_page() argument 3535 err = filler(data, page); in do_read_cache_page() 3537 err = mapping->a_ops->readpage(data, page); in do_read_cache_page() 3636 void *data) in read_cache_page() argument 3638 return do_read_cache_page(mapping, index, filler, data, in read_cache_page()
|
D | Kconfig.debug | 86 reduce the risk of information leaks from freed data. This does
|
/mm/damon/ |
D | Kconfig | 32 This builds the default data access monitoring primitives for DAMON 40 This builds the default data access monitoring primitives for DAMON 60 can use the interface for arbitrary data access monitoring.
|
D | core.c | 370 static int kdamond_fn(void *data); 1000 static int kdamond_fn(void *data) in kdamond_fn() argument 1002 struct damon_ctx *ctx = (struct damon_ctx *)data; in kdamond_fn()
|
/mm/kasan/ |
D | shadow.c | 181 unsigned long action, void *data) in kasan_mem_notifier() argument 183 struct memory_notify *mem_data = data; in kasan_mem_notifier()
|
D | kasan_test_module.c | 120 ((volatile struct work_struct *)work)->data; in kasan_workqueue_uaf()
|
D | report.c | 165 status = (struct kunit_kasan_status *)resource->data; in update_kunit_status()
|