/mm/ |
D | debug_vm_pgtable.c | 94 static void __init pte_basic_tests(struct pgtable_debug_args *args, int idx) in pte_basic_tests() argument 97 pte_t pte = pfn_pte(args->fixed_pte_pfn, prot); in pte_basic_tests() 122 static void __init pte_advanced_tests(struct pgtable_debug_args *args) in pte_advanced_tests() argument 138 page = (args->pte_pfn != ULONG_MAX) ? pfn_to_page(args->pte_pfn) : NULL; in pte_advanced_tests() 143 pte = pfn_pte(args->pte_pfn, args->page_prot); in pte_advanced_tests() 144 set_pte_at(args->mm, args->vaddr, args->ptep, pte); in pte_advanced_tests() 146 ptep_set_wrprotect(args->mm, args->vaddr, args->ptep); in pte_advanced_tests() 147 pte = ptep_get(args->ptep); in pte_advanced_tests() 149 ptep_get_and_clear(args->mm, args->vaddr, args->ptep); in pte_advanced_tests() 150 pte = ptep_get(args->ptep); in pte_advanced_tests() [all …]
|
D | migrate.c | 2775 int migrate_vma_setup(struct migrate_vma *args) in migrate_vma_setup() argument 2777 long nr_pages = (args->end - args->start) >> PAGE_SHIFT; in migrate_vma_setup() 2779 args->start &= PAGE_MASK; in migrate_vma_setup() 2780 args->end &= PAGE_MASK; in migrate_vma_setup() 2781 if (!args->vma || is_vm_hugetlb_page(args->vma) || in migrate_vma_setup() 2782 (args->vma->vm_flags & VM_SPECIAL) || vma_is_dax(args->vma)) in migrate_vma_setup() 2786 if (args->start < args->vma->vm_start || in migrate_vma_setup() 2787 args->start >= args->vma->vm_end) in migrate_vma_setup() 2789 if (args->end <= args->vma->vm_start || args->end > args->vma->vm_end) in migrate_vma_setup() 2791 if (!args->src || !args->dst) in migrate_vma_setup() [all …]
|
D | vmscan.c | 3558 static int should_skip_vma(unsigned long start, unsigned long end, struct mm_walk *args) in should_skip_vma() argument 3561 struct vm_area_struct *vma = args->vma; in should_skip_vma() 3562 struct lru_gen_mm_walk *walk = args->private; in should_skip_vma() 3598 static bool get_next_vma(unsigned long mask, unsigned long size, struct mm_walk *args, in get_next_vma() argument 3607 while (args->vma) { in get_next_vma() 3608 if (start >= args->vma->vm_end) { in get_next_vma() 3609 args->vma = args->vma->vm_next; in get_next_vma() 3613 if (end && end <= args->vma->vm_start) in get_next_vma() 3616 if (should_skip_vma(args->vma->vm_start, args->vma->vm_end, args)) { in get_next_vma() 3617 args->vma = args->vma->vm_next; in get_next_vma() [all …]
|
D | backing-dev.c | 875 int bdi_register_va(struct backing_dev_info *bdi, const char *fmt, va_list args) in bdi_register_va() argument 883 vsnprintf(bdi->dev_name, sizeof(bdi->dev_name), fmt, args); in bdi_register_va() 912 va_list args; in bdi_register() local 915 va_start(args, fmt); in bdi_register() 916 ret = bdi_register_va(bdi, fmt, args); in bdi_register() 917 va_end(args); in bdi_register()
|
D | rmap.c | 2167 struct make_exclusive_args *args = priv; in page_make_device_exclusive_one() local 2177 address + page_size(page)), args->owner); in page_make_device_exclusive_one() 2205 if (args->mm == mm && args->address == address && in page_make_device_exclusive_one() 2207 args->valid = true; in page_make_device_exclusive_one() 2257 struct make_exclusive_args args = { in page_make_device_exclusive() local 2267 .arg = &args, in page_make_device_exclusive() 2280 return args.valid && !page_mapcount(page); in page_make_device_exclusive()
|
D | vmpressure.c | 382 struct eventfd_ctx *eventfd, const char *args) in vmpressure_register_event() argument 392 spec_orig = spec = kstrndup(args, MAX_VMPRESSURE_ARGS_LEN, GFP_KERNEL); in vmpressure_register_event()
|
D | slub.c | 798 va_list args; in slab_bug() local 800 va_start(args, fmt); in slab_bug() 802 vaf.va = &args; in slab_bug() 806 va_end(args); in slab_bug() 813 va_list args; in slab_fix() local 818 va_start(args, fmt); in slab_fix() 820 vaf.va = &args; in slab_fix() 822 va_end(args); in slab_fix() 892 va_list args; in slab_err() local 898 va_start(args, fmt); in slab_err() [all …]
|
D | memcontrol.c | 158 struct eventfd_ctx *eventfd, const char *args); 4264 struct eventfd_ctx *eventfd, const char *args, enum res_type type) in __mem_cgroup_usage_register_event() argument 4272 ret = page_counter_memparse(args, "-1", &threshold); in __mem_cgroup_usage_register_event() 4344 struct eventfd_ctx *eventfd, const char *args) in mem_cgroup_usage_register_event() argument 4346 return __mem_cgroup_usage_register_event(memcg, eventfd, args, _MEM); in mem_cgroup_usage_register_event() 4350 struct eventfd_ctx *eventfd, const char *args) in memsw_cgroup_usage_register_event() argument 4352 return __mem_cgroup_usage_register_event(memcg, eventfd, args, _MEMSWAP); in memsw_cgroup_usage_register_event() 4453 struct eventfd_ctx *eventfd, const char *args) in mem_cgroup_oom_register_event() argument
|
D | page_alloc.c | 4539 va_list args; in warn_alloc() local 4547 va_start(args, fmt); in warn_alloc() 4549 vaf.va = &args; in warn_alloc() 4553 va_end(args); in warn_alloc()
|
D | Kconfig | 847 of up to eight pages (selected by command line args) within the
|
/mm/kfence/ |
D | report.c | 35 va_list args; in seq_con_printf() local 37 va_start(args, fmt); in seq_con_printf() 39 seq_vprintf(seq, fmt, args); in seq_con_printf() 41 vprintk(fmt, args); in seq_con_printf() 42 va_end(args); in seq_con_printf()
|