/mm/ |
D | vmstat.c | 1365 static void *frag_start(struct seq_file *m, loff_t *pos) in frag_start() argument 1378 static void *frag_next(struct seq_file *m, void *arg, loff_t *pos) in frag_next() argument 1386 static void frag_stop(struct seq_file *m, void *arg) in frag_stop() argument 1394 static void walk_zones_in_node(struct seq_file *m, pg_data_t *pgdat, in walk_zones_in_node() argument 1396 void (*print)(struct seq_file *m, pg_data_t *, struct zone *)) in walk_zones_in_node() argument 1408 print(m, pgdat, zone); in walk_zones_in_node() 1416 static void frag_show_print(struct seq_file *m, pg_data_t *pgdat, in frag_show_print() argument 1421 seq_printf(m, "Node %d, zone %8s ", pgdat->node_id, zone->name); in frag_show_print() 1423 seq_printf(m, "%6lu ", zone->free_area[order].nr_free); in frag_show_print() 1424 seq_putc(m, '\n'); in frag_show_print() [all …]
|
D | percpu-stats.c | 20 seq_printf(m, " %-20s: %12lld\n", X, (long long int)Y) 56 static void chunk_map_stats(struct seq_file *m, struct pcpu_chunk *chunk, in chunk_map_stats() argument 139 seq_putc(m, '\n'); in chunk_map_stats() 142 static int percpu_stats_show(struct seq_file *m, void *v) in percpu_stats_show() argument 174 seq_printf(m, " %-20s: %12lld\n", #X, (long long int)pcpu_stats_ai.X) in percpu_stats_show() 176 seq_printf(m, in percpu_stats_show() 186 seq_putc(m, '\n'); in percpu_stats_show() 191 seq_printf(m, " %-20s: %12llu\n", #X, (unsigned long long)pcpu_stats.X) in percpu_stats_show() 193 seq_printf(m, in percpu_stats_show() 205 seq_putc(m, '\n'); in percpu_stats_show() [all …]
|
D | slab_common.c | 934 static void print_slabinfo_header(struct seq_file *m) in print_slabinfo_header() argument 941 seq_puts(m, "slabinfo - version: 2.1 (statistics)\n"); in print_slabinfo_header() 943 seq_puts(m, "slabinfo - version: 2.1\n"); in print_slabinfo_header() 945 seq_puts(m, "# name <active_objs> <num_objs> <objsize> <objperslab> <pagesperslab>"); in print_slabinfo_header() 946 seq_puts(m, " : tunables <limit> <batchcount> <sharedfactor>"); in print_slabinfo_header() 947 seq_puts(m, " : slabdata <active_slabs> <num_slabs> <sharedavail>"); in print_slabinfo_header() 949 …seq_puts(m, " : globalstat <listallocs> <maxobjs> <grown> <reaped> <error> <maxfreeable> <nodeallo… in print_slabinfo_header() 950 seq_puts(m, " : cpustat <allochit> <allocmiss> <freehit> <freemiss>"); in print_slabinfo_header() 952 seq_putc(m, '\n'); in print_slabinfo_header() 955 void *slab_start(struct seq_file *m, loff_t *pos) in slab_start() argument [all …]
|
D | slob.c | 471 unsigned int *m; in __do_kmalloc_node() local 495 m = slob_alloc(size + minalign, gfp, align, node, minalign); in __do_kmalloc_node() 497 if (!m) in __do_kmalloc_node() 499 *m = size; in __do_kmalloc_node() 500 ret = (void *)m + minalign; in __do_kmalloc_node() 555 unsigned int *m = (unsigned int *)(block - align); in kfree() local 557 slob_free(m, *m + align); in kfree() 573 unsigned int *m; in __ksize() local 585 m = (unsigned int *)(block - align); in __ksize() 586 return SLOB_UNITS(*m) * SLOB_UNIT; in __ksize()
|
D | memblock.c | 941 struct memblock_region *m, in should_skip_region() argument 944 int m_nid = memblock_get_region_node(m); in should_skip_region() 955 if (movable_node_is_enabled() && memblock_is_hotpluggable(m) && in should_skip_region() 960 if ((flags & MEMBLOCK_MIRROR) && !memblock_is_mirror(m)) in should_skip_region() 964 if (!(flags & MEMBLOCK_NOMAP) && memblock_is_nomap(m)) in should_skip_region() 1009 struct memblock_region *m = &type_a->regions[idx_a]; in __next_mem_range() local 1011 phys_addr_t m_start = m->base; in __next_mem_range() 1012 phys_addr_t m_end = m->base + m->size; in __next_mem_range() 1013 int m_nid = memblock_get_region_node(m); in __next_mem_range() 1015 if (should_skip_region(type_a, m, nid, flags)) in __next_mem_range() [all …]
|
D | memcontrol.c | 4038 static int memcg_numa_stat_show(struct seq_file *m, void *v) in memcg_numa_stat_show() argument 4053 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); in memcg_numa_stat_show() 4056 seq_printf(m, "%s=%lu", stat->name, in memcg_numa_stat_show() 4060 seq_printf(m, " N%d=%lu", nid, in memcg_numa_stat_show() 4063 seq_putc(m, '\n'); in memcg_numa_stat_show() 4068 seq_printf(m, "hierarchical_%s=%lu", stat->name, in memcg_numa_stat_show() 4072 seq_printf(m, " N%d=%lu", nid, in memcg_numa_stat_show() 4075 seq_putc(m, '\n'); in memcg_numa_stat_show() 4116 static int memcg_stat_show(struct seq_file *m, void *v) in memcg_stat_show() argument 4118 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); in memcg_stat_show() [all …]
|
D | vmalloc.c | 3451 static void *s_start(struct seq_file *m, loff_t *pos) in s_start() argument 3461 static void *s_next(struct seq_file *m, void *p, loff_t *pos) in s_next() argument 3466 static void s_stop(struct seq_file *m, void *p) in s_stop() argument 3474 static void show_numa_info(struct seq_file *m, struct vm_struct *v) in show_numa_info() argument 3477 unsigned int nr, *counters = m->private; in show_numa_info() 3494 seq_printf(m, " N%u=%u", nr, counters[nr]); in show_numa_info() 3498 static void show_purge_info(struct seq_file *m) in show_purge_info() argument 3508 seq_printf(m, "0x%pK-0x%pK %7ld unpurged vm_area\n", in show_purge_info() 3514 static int s_show(struct seq_file *m, void *p) in s_show() argument 3526 seq_printf(m, "0x%pK-0x%pK %7ld vm_map_ram\n", in s_show() [all …]
|
D | slab.h | 214 void slabinfo_show_stats(struct seq_file *m, struct kmem_cache *s); 642 void *slab_start(struct seq_file *m, loff_t *pos); 643 void *slab_next(struct seq_file *m, void *p, loff_t *pos); 644 void slab_stop(struct seq_file *m, void *p); 645 int memcg_slab_show(struct seq_file *m, void *p);
|
D | page_owner.c | 300 void pagetypeinfo_showmixedcount_print(struct seq_file *m, in pagetypeinfo_showmixedcount_print() argument 380 seq_printf(m, "Node %d, zone %8s ", pgdat->node_id, zone->name); in pagetypeinfo_showmixedcount_print() 382 seq_printf(m, "%12lu ", count[i]); in pagetypeinfo_showmixedcount_print() 383 seq_putc(m, '\n'); in pagetypeinfo_showmixedcount_print()
|
D | hugetlb.c | 2453 struct huge_bootmem_page *m; in __alloc_bootmem_huge_page() local 2468 m = addr; in __alloc_bootmem_huge_page() 2475 BUG_ON(!IS_ALIGNED(virt_to_phys(m), huge_page_size(h))); in __alloc_bootmem_huge_page() 2477 INIT_LIST_HEAD(&m->list); in __alloc_bootmem_huge_page() 2478 list_add(&m->list, &huge_boot_pages); in __alloc_bootmem_huge_page() 2479 m->hstate = h; in __alloc_bootmem_huge_page() 2489 struct huge_bootmem_page *m; in gather_bootmem_prealloc() local 2491 list_for_each_entry(m, &huge_boot_pages, list) { in gather_bootmem_prealloc() 2492 struct page *page = virt_to_page(m); in gather_bootmem_prealloc() 2493 struct hstate *h = m->hstate; in gather_bootmem_prealloc() [all …]
|
D | backing-dev.c | 46 static int bdi_debug_stats_show(struct seq_file *m, void *v) in bdi_debug_stats_show() argument 48 struct backing_dev_info *bdi = m->private; in bdi_debug_stats_show() 73 seq_printf(m, in bdi_debug_stats_show()
|
D | slub.c | 2214 enum slab_modes l = M_NONE, m = M_NONE; in deactivate_slab() local 2293 m = M_FREE; in deactivate_slab() 2295 m = M_PARTIAL; in deactivate_slab() 2306 m = M_FULL; in deactivate_slab() 2320 if (l != m) { in deactivate_slab() 2326 if (m == M_PARTIAL) in deactivate_slab() 2328 else if (m == M_FULL) in deactivate_slab() 2332 l = m; in deactivate_slab() 2342 if (m == M_PARTIAL) in deactivate_slab() 2344 else if (m == M_FULL) in deactivate_slab() [all …]
|
D | slab.c | 4068 void slabinfo_show_stats(struct seq_file *m, struct kmem_cache *cachep) in slabinfo_show_stats() argument 4082 seq_printf(m, " : globalstat %7lu %6lu %5lu %4lu %4lu %4lu %4lu %4lu %4lu", in slabinfo_show_stats() 4094 seq_printf(m, " : cpustat %6lu %6lu %6lu %6lu", in slabinfo_show_stats()
|