/mm/damon/ |
D | reclaim.c | 30 static bool enabled __read_mostly; variable 38 static unsigned long min_age __read_mostly = 120000000; variable 51 static unsigned long quota_ms __read_mostly = 10; variable 64 static unsigned long quota_sz __read_mostly = 128 * 1024 * 1024; variable 77 static unsigned long quota_reset_interval_ms __read_mostly = 1000; variable 86 static unsigned long wmarks_interval __read_mostly = 5000000; variable 96 static unsigned long wmarks_high __read_mostly = 500; variable 106 static unsigned long wmarks_mid __read_mostly = 400; variable 117 static unsigned long wmarks_low __read_mostly = 200; variable 126 static unsigned long sample_interval __read_mostly = 5000; variable [all …]
|
/mm/ |
D | util.c | 857 int sysctl_overcommit_memory __read_mostly = OVERCOMMIT_GUESS; variable 858 int sysctl_overcommit_ratio __read_mostly = 50; variable 859 unsigned long sysctl_overcommit_kbytes __read_mostly; variable 860 int sysctl_max_map_count __read_mostly = DEFAULT_MAX_MAP_COUNT; variable 861 unsigned long sysctl_user_reserve_kbytes __read_mostly = 1UL << 17; /* 128MB */ variable 862 unsigned long sysctl_admin_reserve_kbytes __read_mostly = 1UL << 13; /* 8MB */ variable
|
D | frontswap.c | 30 static struct frontswap_ops *frontswap_ops __read_mostly; variable 43 static bool frontswap_writethrough_enabled __read_mostly; variable 50 static bool frontswap_tmem_exclusive_gets_enabled __read_mostly; variable
|
D | khugepaged.c | 59 static struct task_struct *khugepaged_thread __read_mostly; variable 63 static unsigned int khugepaged_pages_to_scan __read_mostly; variable 66 static unsigned int khugepaged_scan_sleep_millisecs __read_mostly = 10000; variable 68 static unsigned int khugepaged_alloc_sleep_millisecs __read_mostly = 60000; variable 77 static unsigned int khugepaged_max_ptes_none __read_mostly; variable 78 static unsigned int khugepaged_max_ptes_swap __read_mostly; variable 79 static unsigned int khugepaged_max_ptes_shared __read_mostly; variable 82 static __read_mostly DEFINE_HASHTABLE(mm_slots_hash, MM_SLOTS_HASH_BITS); 84 static struct kmem_cache *mm_slot_cache __read_mostly; variable
|
D | swap_state.c | 39 struct address_space *swapper_spaces[MAX_SWAPFILES] __read_mostly; 40 static unsigned int nr_swapper_spaces[MAX_SWAPFILES] __read_mostly; 41 static bool enable_vma_readahead __read_mostly = true; variable
|
D | cleancache.c | 24 static const struct cleancache_ops *cleancache_ops __read_mostly; variable
|
D | page_alloc.c | 248 nodemask_t node_states[NR_NODE_STATES] __read_mostly = { 262 atomic_long_t _totalram_pages __read_mostly; variable 264 unsigned long totalreserve_pages __read_mostly; variable 265 unsigned long totalcma_pages __read_mostly; variable 268 gfp_t gfp_allowed_mask __read_mostly = GFP_BOOT_MASK; variable 275 static bool _init_on_alloc_enabled_early __read_mostly variable 284 static bool _init_on_free_enabled_early __read_mostly variable 349 unsigned int pageblock_order __read_mostly; variable 423 int watermark_boost_factor __read_mostly = 15000; variable 444 unsigned int nr_node_ids __read_mostly = MAX_NUMNODES; variable [all …]
|
D | debug.c | 271 static bool page_init_poisoning __read_mostly = true; variable
|
D | page_reporting.c | 20 static struct page_reporting_dev_info __rcu *pr_dev_info __read_mostly; variable
|
D | memory-failure.c | 64 int sysctl_memory_failure_early_kill __read_mostly = 0; variable 66 int sysctl_memory_failure_recovery __read_mostly = 1; variable 68 atomic_long_t num_poisoned_pages __read_mostly = ATOMIC_LONG_INIT(0); variable
|
D | memory_hotplug.c | 88 static int online_policy __read_mostly = ONLINE_POLICY_CONTIG_ZONES; variable 105 static unsigned int auto_movable_ratio __read_mostly = 301; variable 115 static bool auto_movable_numa_aware __read_mostly = true; variable
|
D | workingset.c | 185 static unsigned int bucket_order __read_mostly; variable
|
D | compaction.c | 1771 int sysctl_compact_unevictable_allowed __read_mostly = 0; variable 1773 int sysctl_compact_unevictable_allowed __read_mostly = 1; variable 2787 unsigned int __read_mostly sysctl_compaction_proactiveness = 20;
|
D | highmem.c | 104 atomic_long_t _totalhigh_pages __read_mostly; variable
|
D | oom_kill.c | 487 static bool oom_killer_disabled __read_mostly; variable
|
D | hugetlb_cgroup.c | 33 static struct hugetlb_cgroup *root_h_cgroup __read_mostly; variable
|
D | huge_memory.c | 50 unsigned long transparent_hugepage_flags __read_mostly = variable 64 struct page *huge_zero_page __read_mostly; variable 65 unsigned long huge_zero_pfn __read_mostly = ~0UL; variable
|
D | ksm.c | 274 static unsigned int zero_checksum __read_mostly; variable 277 static bool ksm_use_zero_pages __read_mostly; variable
|
D | percpu.c | 202 static bool pcpu_async_enabled __read_mostly; variable 3279 unsigned long __per_cpu_offset[NR_CPUS] __read_mostly;
|
D | memory.c | 120 int randomize_va_space __read_mostly = variable 146 unsigned long zero_pfn __read_mostly; variable 149 unsigned long highest_memmap_pfn __read_mostly; variable 4383 static unsigned long fault_around_bytes __read_mostly = variable
|
D | vmstat.c | 1866 int sysctl_stat_interval __read_mostly = HZ; variable
|
D | mmap.c | 72 int mmap_rnd_bits __read_mostly = CONFIG_ARCH_MMAP_RND_BITS; variable 77 int mmap_rnd_compat_bits __read_mostly = CONFIG_ARCH_MMAP_RND_COMPAT_BITS; variable
|
D | memcontrol.c | 77 struct cgroup_subsys memory_cgrp_subsys __read_mostly; variable 80 struct mem_cgroup *root_mem_cgroup __read_mostly; variable 127 static struct mem_cgroup_tree soft_limit_tree __read_mostly; variable
|
D | migrate.c | 1149 static int node_demotion[MAX_NUMNODES] __read_mostly =
|
/mm/kfence/ |
D | core.c | 48 static bool kfence_enabled __read_mostly; variable 50 static unsigned long kfence_sample_interval __read_mostly = CONFIG_KFENCE_SAMPLE_INTERVAL; variable 89 static unsigned long kfence_skip_covered_thresh __read_mostly = 75; variable
|