Home
last modified time | relevance | path

Searched refs:mode (Results 1 – 18 of 18) sorted by relevance

/mm/
Dmempolicy.c124 .mode = MPOL_PREFERRED,
168 if (pol->mode) in get_task_policy()
242 if (pol->mode == MPOL_PREFERRED && nodes_empty(*nodes)) in mpol_set_nodemask()
258 ret = mpol_ops[pol->mode].create(pol, &nsc->mask2); in mpol_set_nodemask()
260 ret = mpol_ops[pol->mode].create(pol, NULL); in mpol_set_nodemask()
268 static struct mempolicy *mpol_new(unsigned short mode, unsigned short flags, in mpol_new() argument
274 mode, flags, nodes ? nodes_addr(*nodes)[0] : NUMA_NO_NODE); in mpol_new()
276 if (mode == MPOL_DEFAULT) { in mpol_new()
288 if (mode == MPOL_PREFERRED) { in mpol_new()
294 } else if (mode == MPOL_LOCAL) { in mpol_new()
[all …]
Duserfaultfd.c285 enum mcopy_atomic_mode mode) in __mcopy_atomic_hugetlb() argument
305 if (mode == MCOPY_ATOMIC_ZEROPAGE) { in __mcopy_atomic_hugetlb()
373 if (mode != MCOPY_ATOMIC_CONTINUE && in __mcopy_atomic_hugetlb()
382 dst_addr, src_addr, mode, &page); in __mcopy_atomic_hugetlb()
494 enum mcopy_atomic_mode mode);
503 enum mcopy_atomic_mode mode, in mfill_atomic_pte() argument
508 if (mode == MCOPY_ATOMIC_CONTINUE) { in mfill_atomic_pte()
524 if (mode == MCOPY_ATOMIC_NORMAL) in mfill_atomic_pte()
535 mode != MCOPY_ATOMIC_NORMAL, in mfill_atomic_pte()
548 __u64 mode) in __mcopy_atomic() argument
[all …]
Dmigrate.c63 int isolate_movable_page(struct page *page, isolate_mode_t mode) in isolate_movable_page() argument
106 if (!mapping->a_ops->isolate_page(page, mode)) in isolate_movable_page()
683 enum migrate_mode mode) in migrate_page() argument
694 if (mode != MIGRATE_SYNC_NO_COPY) in migrate_page()
705 enum migrate_mode mode) in buffer_migrate_lock_buffers() argument
710 if (mode != MIGRATE_ASYNC) { in buffer_migrate_lock_buffers()
742 struct page *newpage, struct page *page, enum migrate_mode mode, in __buffer_migrate_page() argument
750 return migrate_page(mapping, newpage, page, mode); in __buffer_migrate_page()
758 if (!buffer_migrate_lock_buffers(head, mode)) in __buffer_migrate_page()
801 if (mode != MIGRATE_SYNC_NO_COPY) in __buffer_migrate_page()
[all …]
Dvmpressure.c156 enum vmpressure_modes mode; member
169 if (ancestor && ev->mode == VMPRESSURE_LOCAL) in vmpressure_event()
171 if (signalled && ev->mode == VMPRESSURE_NO_PASSTHROUGH) in vmpressure_event()
379 enum vmpressure_modes mode = VMPRESSURE_NO_PASSTHROUGH; in vmpressure_register_event() local
402 mode = ret; in vmpressure_register_event()
413 ev->mode = mode; in vmpressure_register_event()
Dfailslab.c46 umode_t mode = S_IFREG | 0600; in failslab_debugfs_init() local
52 debugfs_create_bool("ignore-gfp-wait", mode, dir, in failslab_debugfs_init()
54 debugfs_create_bool("cache-filter", mode, dir, in failslab_debugfs_init()
Dballoon_compaction.c206 bool balloon_page_isolate(struct page *page, isolate_mode_t mode) in balloon_page_isolate() argument
235 enum migrate_mode mode) in balloon_page_migrate() argument
244 if (mode == MIGRATE_SYNC_NO_COPY) in balloon_page_migrate()
250 return balloon->migratepage(balloon, newpage, page, mode); in balloon_page_migrate()
Dcompaction.c441 if (cc->mode != MIGRATE_ASYNC && in update_cached_migrate()
509 if (cc->mode == MIGRATE_ASYNC && !cc->contended) { in compact_lock_irqsave()
855 if (cc->mode == MIGRATE_ASYNC) in isolate_migratepages_block()
866 if (cc->direct_compaction && (cc->mode == MIGRATE_ASYNC)) { in isolate_migratepages_block()
1187 if ((cc->mode != MIGRATE_ASYNC) || !cc->direct_compaction) in suitable_migration_source()
1288 if (cc->direct_compaction && cc->mode == MIGRATE_ASYNC) in fast_isolate_around()
1521 stride = cc->mode == MIGRATE_ASYNC ? COMPACT_CLUSTER_MAX : 1; in isolate_freepages()
1809 (cc->mode != MIGRATE_SYNC ? ISOLATE_ASYNC_MIGRATE : 0); in isolate_migratepages()
2080 if (cc->mode == MIGRATE_ASYNC || in __compact_finished()
2238 const bool sync = cc->mode != MIGRATE_ASYNC; in compact_zone()
[all …]
Dshmem.c123 umode_t mode; member
1483 if (!mpol || mpol->mode == MPOL_DEFAULT) in shmem_show_mpol()
2058 static int synchronous_wake_function(wait_queue_entry_t *wait, unsigned mode, int sync, void *key) in synchronous_wake_function() argument
2060 int ret = default_wake_function(wait, mode, sync, key); in synchronous_wake_function()
2306 umode_t mode, dev_t dev, unsigned long flags) in shmem_get_inode() argument
2319 inode_init_owner(inode, dir, mode); in shmem_get_inode()
2334 switch (mode & S_IFMT) { in shmem_get_inode()
2337 init_special_inode(inode, mode, dev); in shmem_get_inode()
2737 static long shmem_fallocate(struct file *file, int mode, loff_t offset, in shmem_fallocate() argument
2747 if (mode & ~(FALLOC_FL_KEEP_SIZE | FALLOC_FL_PUNCH_HOLE)) in shmem_fallocate()
[all …]
Dvmscan.c1611 int __isolate_lru_page(struct page *page, isolate_mode_t mode) in __isolate_lru_page() argument
1620 if (PageUnevictable(page) && !(mode & ISOLATE_UNEVICTABLE)) in __isolate_lru_page()
1633 if (mode & ISOLATE_ASYNC_MIGRATE) { in __isolate_lru_page()
1662 if ((mode & ISOLATE_UNMAPPED) && page_mapped(page)) in __isolate_lru_page()
1728 isolate_mode_t mode = (sc->may_unmap ? 0 : ISOLATE_UNMAPPED); in isolate_lru_pages() local
1760 switch (__isolate_lru_page(page, mode)) { in isolate_lru_pages()
1799 total_scan, skipped, nr_taken, mode, lru); in isolate_lru_pages()
Dinternal.h255 enum migrate_mode mode; /* Async or sync migration mode */ member
Dzsmalloc.c1899 static bool zs_page_isolate(struct page *page, isolate_mode_t mode) in zs_page_isolate() argument
1956 struct page *page, enum migrate_mode mode) in zs_page_migrate() argument
1976 if (mode == MIGRATE_SYNC_NO_COPY) in zs_page_migrate()
Dz3fold.c1577 static bool z3fold_page_isolate(struct page *page, isolate_mode_t mode) in z3fold_page_isolate() argument
1617 struct page *page, enum migrate_mode mode) in z3fold_page_migrate() argument
Dpage_alloc.c3706 umode_t mode = S_IFREG | 0600; in fail_page_alloc_debugfs() local
3712 debugfs_create_bool("ignore-gfp-wait", mode, dir, in fail_page_alloc_debugfs()
3714 debugfs_create_bool("ignore-gfp-highmem", mode, dir, in fail_page_alloc_debugfs()
3716 debugfs_create_u32("min-order", mode, dir, &fail_page_alloc.min_order); in fail_page_alloc_debugfs()
8757 if (cc->alloc_contig && cc->mode == MIGRATE_ASYNC) in __alloc_contig_migrate_range()
8790 NULL, (unsigned long)&mtc, cc->mode, MR_CONTIG_RANGE); in __alloc_contig_migrate_range()
8848 .mode = gfp_mask & __GFP_NORETRY ? MIGRATE_ASYNC : MIGRATE_SYNC, in alloc_contig_range()
Dhugetlb.c2183 enum vma_resv_mode mode) in __vma_reservation_common() argument
2195 switch (mode) { in __vma_reservation_common()
4747 enum mcopy_atomic_mode mode, in hugetlb_mcopy_atomic_pte() argument
4750 bool is_continue = (mode == MCOPY_ATOMIC_CONTINUE); in hugetlb_mcopy_atomic_pte()
Dfilemap.c1054 static int wake_page_function(wait_queue_entry_t *wait, unsigned mode, int sync, void *arg) in wake_page_function() argument
1089 wake_up_state(wait->private, mode); in wake_page_function()
Dmemcontrol.c1920 unsigned mode, int sync, void *arg) in memcg_oom_wake_function() argument
1932 return autoremove_wake_function(wait, mode, sync, arg); in memcg_oom_wake_function()
4849 static int memcg_event_wake(wait_queue_entry_t *wait, unsigned mode, in memcg_event_wake() argument
Dslub.c1963 int mode, int *objects) in acquire_slab() argument
1980 if (mode) { in acquire_slab()
Dmemory.c5118 if (pol && pol->mode == MPOL_INTERLEAVE) { in ___handle_speculative_fault()