/mm/ |
D | slab.c | 257 #define OBJFREELIST_SLAB(x) ((x)->flags & CFLGS_OBJFREELIST_SLAB) argument 258 #define OFF_SLAB(x) ((x)->flags & CFLGS_OFF_SLAB) argument 272 #define STATS_INC_ACTIVE(x) ((x)->num_active++) argument 273 #define STATS_DEC_ACTIVE(x) ((x)->num_active--) argument 274 #define STATS_INC_ALLOCED(x) ((x)->num_allocations++) argument 275 #define STATS_INC_GROWN(x) ((x)->grown++) argument 276 #define STATS_ADD_REAPED(x, y) ((x)->reaped += (y)) argument 277 #define STATS_SET_HIGH(x) \ argument 282 #define STATS_INC_ERR(x) ((x)->errors++) argument 283 #define STATS_INC_NODEALLOCS(x) ((x)->node_allocs++) argument [all …]
|
D | page-writeback.c | 302 unsigned long x = 0; in highmem_dirtyable_memory() local 358 unsigned long x; in global_dirtyable_memory() local 805 long x; in pos_ratio_polynom() local 904 long x; in wb_position_ratio() local 1190 unsigned long x; in wb_update_dirty_ratelimit() local
|
D | slub.c | 389 struct kmem_cache_order_objects x = { in oo_make() local 396 static inline unsigned int oo_order(struct kmem_cache_order_objects x) in oo_order() 401 static inline unsigned int oo_objects(struct kmem_cache_order_objects x) in oo_objects() 1682 static __always_inline void kfree_hook(void *x) in kfree_hook() 1689 void *x, bool init) in slab_free_hook() 2777 unsigned long x = 0; in count_partial() local 3520 void ___cache_free(struct kmem_cache *cache, void *x, unsigned long addr) in ___cache_free() 3526 void kmem_cache_free(struct kmem_cache *s, void *x) in kmem_cache_free() 4560 void kfree(const void *x) in kfree() 5248 int x; in show_slab_objects() local [all …]
|
D | swap_state.c | 61 #define INC_CACHE_INFO(x) data_race(swap_cache_info.x++) argument 62 #define ADD_CACHE_INFO(x, nr) data_race(swap_cache_info.x += (nr)) argument
|
D | ksm.c | 46 #define NUMA(x) (x) argument 47 #define DO_NUMA(x) do { (x); } while (0) argument 49 #define NUMA(x) (0) argument 50 #define DO_NUMA(x) do { } while (0) argument
|
D | vmstat.c | 319 long x; in __mod_zone_page_state() local 352 long x; in __mod_node_page_state() local 1027 long x = atomic_long_read(&pgdat->vm_stat[item]); in node_page_state_pages() local
|
D | readahead.c | 118 gfp_t readahead_gfp_mask(struct address_space *x) in readahead_gfp_mask()
|
D | vmalloc.c | 74 bool is_vmalloc_addr(const void *x) in is_vmalloc_addr() 631 int is_vmalloc_or_module_addr(const void *x) in is_vmalloc_or_module_addr() 1824 #define VMAP_MIN(x, y) ((x) < (y) ? (x) : (y)) /* can't use min() */ argument 1825 #define VMAP_MAX(x, y) ((x) > (y) ? (x) : (y)) /* can't use max() */ argument
|
D | oom_kill.c | 489 #define K(x) ((x) << (PAGE_SHIFT-10)) argument
|
D | kmemleak.c | 241 #define kmemleak_warn(x...) do { \ argument 252 #define kmemleak_stop(x...) do { \ argument
|
D | backing-dev.c | 36 #define K(x) ((x) << (PAGE_SHIFT - 10)) argument
|
D | hugetlb_cgroup.c | 26 #define MEMFILE_PRIVATE(x, val) (((x) << 16) | (val)) argument
|
D | memcontrol.c | 219 #define MEMFILE_PRIVATE(x, val) ((x) << 16 | (val)) argument 672 unsigned int x; in memcg_rstat_updated() local 732 long x = 0; in memcg_page_state_local() local 871 long x = 0; in memcg_events_local() local 1547 #define K(x) ((x) << (PAGE_SHIFT-10)) argument
|
D | slab_common.c | 134 void *x = p[i] = kmem_cache_alloc(s, flags); in __kmem_cache_alloc_bulk() local
|
D | util.c | 41 void kfree_const(const void *x) in kfree_const()
|
D | slab.h | 471 static inline struct kmem_cache *cache_from_obj(struct kmem_cache *s, void *x) in cache_from_obj()
|
D | percpu.c | 619 static inline bool pcpu_region_overlap(int a, int b, int x, int y) in pcpu_region_overlap()
|
D | page_alloc.c | 6170 #define K(x) ((x) << (PAGE_SHIFT-10)) argument
|