/mm/ |
D | vmstat.c | 1414 static void *frag_start(struct seq_file *m, loff_t *pos) in frag_start() argument 1427 static void *frag_next(struct seq_file *m, void *arg, loff_t *pos) in frag_next() argument 1435 static void frag_stop(struct seq_file *m, void *arg) in frag_stop() argument 1443 static void walk_zones_in_node(struct seq_file *m, pg_data_t *pgdat, in walk_zones_in_node() argument 1445 void (*print)(struct seq_file *m, pg_data_t *, struct zone *)) in walk_zones_in_node() argument 1457 print(m, pgdat, zone); in walk_zones_in_node() 1465 static void frag_show_print(struct seq_file *m, pg_data_t *pgdat, in frag_show_print() argument 1470 seq_printf(m, "Node %d, zone %8s ", pgdat->node_id, zone->name); in frag_show_print() 1472 seq_printf(m, "%6lu ", zone->free_area[order].nr_free); in frag_show_print() 1473 seq_putc(m, '\n'); in frag_show_print() [all …]
|
D | percpu-stats.c | 20 seq_printf(m, " %-20s: %12lld\n", X, (long long int)Y) 52 static void chunk_map_stats(struct seq_file *m, struct pcpu_chunk *chunk, in chunk_map_stats() argument 132 seq_putc(m, '\n'); in chunk_map_stats() 135 static int percpu_stats_show(struct seq_file *m, void *v) in percpu_stats_show() argument 161 seq_printf(m, " %-20s: %12lld\n", #X, (long long int)pcpu_stats_ai.X) in percpu_stats_show() 163 seq_printf(m, in percpu_stats_show() 173 seq_putc(m, '\n'); in percpu_stats_show() 178 seq_printf(m, " %-20s: %12llu\n", #X, (unsigned long long)pcpu_stats.X) in percpu_stats_show() 180 seq_printf(m, in percpu_stats_show() 192 seq_putc(m, '\n'); in percpu_stats_show() [all …]
|
D | slab_common.c | 1055 static void print_slabinfo_header(struct seq_file *m) in print_slabinfo_header() argument 1062 seq_puts(m, "slabinfo - version: 2.1 (statistics)\n"); in print_slabinfo_header() 1064 seq_puts(m, "slabinfo - version: 2.1\n"); in print_slabinfo_header() 1066 seq_puts(m, "# name <active_objs> <num_objs> <objsize> <objperslab> <pagesperslab>"); in print_slabinfo_header() 1067 seq_puts(m, " : tunables <limit> <batchcount> <sharedfactor>"); in print_slabinfo_header() 1068 seq_puts(m, " : slabdata <active_slabs> <num_slabs> <sharedavail>"); in print_slabinfo_header() 1070 …seq_puts(m, " : globalstat <listallocs> <maxobjs> <grown> <reaped> <error> <maxfreeable> <nodeallo… in print_slabinfo_header() 1071 seq_puts(m, " : cpustat <allochit> <allocmiss> <freehit> <freemiss>"); in print_slabinfo_header() 1073 trace_android_vh_print_slabinfo_header(m); in print_slabinfo_header() 1074 seq_putc(m, '\n'); in print_slabinfo_header() [all …]
|
D | slob.c | 479 unsigned int *m; in __do_kmalloc_node() local 502 m = slob_alloc(size + minalign, gfp, align, node, minalign); in __do_kmalloc_node() 504 if (!m) in __do_kmalloc_node() 506 *m = size; in __do_kmalloc_node() 507 ret = (void *)m + minalign; in __do_kmalloc_node() 562 unsigned int *m = (unsigned int *)(block - align); in kfree() local 564 slob_free(m, *m + align); in kfree() 580 unsigned int *m; in __ksize() local 592 m = (unsigned int *)(block - align); in __ksize() 593 return SLOB_UNITS(*m) * SLOB_UNIT; in __ksize()
|
D | memcontrol.c | 3980 static int memcg_numa_stat_show(struct seq_file *m, void *v) in memcg_numa_stat_show() argument 3995 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); in memcg_numa_stat_show() 4000 seq_printf(m, "%s=%lu", stat->name, in memcg_numa_stat_show() 4004 seq_printf(m, " N%d=%lu", nid, in memcg_numa_stat_show() 4007 seq_putc(m, '\n'); in memcg_numa_stat_show() 4012 seq_printf(m, "hierarchical_%s=%lu", stat->name, in memcg_numa_stat_show() 4016 seq_printf(m, " N%d=%lu", nid, in memcg_numa_stat_show() 4019 seq_putc(m, '\n'); in memcg_numa_stat_show() 4060 static int memcg_stat_show(struct seq_file *m, void *v) in memcg_stat_show() argument 4062 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); in memcg_stat_show() [all …]
|
D | memblock.c | 968 struct memblock_region *m, in should_skip_region() argument 971 int m_nid = memblock_get_region_node(m); in should_skip_region() 982 if (movable_node_is_enabled() && memblock_is_hotpluggable(m) && in should_skip_region() 987 if ((flags & MEMBLOCK_MIRROR) && !memblock_is_mirror(m)) in should_skip_region() 991 if (!(flags & MEMBLOCK_NOMAP) && memblock_is_nomap(m)) in should_skip_region() 1036 struct memblock_region *m = &type_a->regions[idx_a]; in __next_mem_range() local 1038 phys_addr_t m_start = m->base; in __next_mem_range() 1039 phys_addr_t m_end = m->base + m->size; in __next_mem_range() 1040 int m_nid = memblock_get_region_node(m); in __next_mem_range() 1042 if (should_skip_region(type_a, m, nid, flags)) in __next_mem_range() [all …]
|
D | vmalloc.c | 3938 static void *s_start(struct seq_file *m, loff_t *pos) in s_start() argument 3948 static void *s_next(struct seq_file *m, void *p, loff_t *pos) in s_next() argument 3953 static void s_stop(struct seq_file *m, void *p) in s_stop() argument 3961 static void show_numa_info(struct seq_file *m, struct vm_struct *v) in show_numa_info() argument 3964 unsigned int nr, *counters = m->private; in show_numa_info() 3981 seq_printf(m, " N%u=%u", nr, counters[nr]); in show_numa_info() 3985 static void show_purge_info(struct seq_file *m) in show_purge_info() argument 3991 seq_printf(m, "0x%pK-0x%pK %7ld unpurged vm_area\n", in show_purge_info() 3998 static int s_show(struct seq_file *m, void *p) in s_show() argument 4010 seq_printf(m, "0x%pK-0x%pK %7ld vm_map_ram\n", in s_show() [all …]
|
D | slab.h | 210 void slabinfo_show_stats(struct seq_file *m, struct kmem_cache *s); 609 void *slab_start(struct seq_file *m, loff_t *pos); 610 void *slab_next(struct seq_file *m, void *p, loff_t *pos); 611 void slab_stop(struct seq_file *m, void *p); 612 int memcg_slab_show(struct seq_file *m, void *p);
|
D | page_owner.c | 283 void pagetypeinfo_showmixedcount_print(struct seq_file *m, in pagetypeinfo_showmixedcount_print() argument 360 seq_printf(m, "Node %d, zone %8s ", pgdat->node_id, zone->name); in pagetypeinfo_showmixedcount_print() 362 seq_printf(m, "%12lu ", count[i]); in pagetypeinfo_showmixedcount_print() 363 seq_putc(m, '\n'); in pagetypeinfo_showmixedcount_print()
|
D | vmscan.c | 5278 static void *lru_gen_seq_start(struct seq_file *m, loff_t *pos) in lru_gen_seq_start() argument 5283 m->private = kvmalloc(PATH_MAX, GFP_KERNEL); in lru_gen_seq_start() 5284 if (!m->private) in lru_gen_seq_start() 5300 static void lru_gen_seq_stop(struct seq_file *m, void *v) in lru_gen_seq_stop() argument 5305 kvfree(m->private); in lru_gen_seq_stop() 5306 m->private = NULL; in lru_gen_seq_stop() 5309 static void *lru_gen_seq_next(struct seq_file *m, void *v, loff_t *pos) in lru_gen_seq_next() argument 5328 static void lru_gen_seq_show_full(struct seq_file *m, struct lruvec *lruvec, in lru_gen_seq_show_full() argument 5338 seq_printf(m, " %10d", tier); in lru_gen_seq_show_full() 5356 seq_printf(m, " %10lu%c", n[i], s[i]); in lru_gen_seq_show_full() [all …]
|
D | hugetlb.c | 2876 struct huge_bootmem_page *m; in __alloc_bootmem_huge_page() local 2891 m = addr; in __alloc_bootmem_huge_page() 2898 BUG_ON(!IS_ALIGNED(virt_to_phys(m), huge_page_size(h))); in __alloc_bootmem_huge_page() 2900 INIT_LIST_HEAD(&m->list); in __alloc_bootmem_huge_page() 2901 list_add(&m->list, &huge_boot_pages); in __alloc_bootmem_huge_page() 2902 m->hstate = h; in __alloc_bootmem_huge_page() 2912 struct huge_bootmem_page *m; in gather_bootmem_prealloc() local 2914 list_for_each_entry(m, &huge_boot_pages, list) { in gather_bootmem_prealloc() 2915 struct page *page = virt_to_page(m); in gather_bootmem_prealloc() 2916 struct hstate *h = m->hstate; in gather_bootmem_prealloc() [all …]
|
D | backing-dev.c | 49 static int bdi_debug_stats_show(struct seq_file *m, void *v) in bdi_debug_stats_show() argument 51 struct backing_dev_info *bdi = m->private; in bdi_debug_stats_show() 75 seq_printf(m, in bdi_debug_stats_show()
|
D | slub.c | 2344 enum slab_modes l = M_NONE, m = M_NONE; in deactivate_slab() local 2413 m = M_FREE; in deactivate_slab() 2415 m = M_PARTIAL; in deactivate_slab() 2426 m = M_FULL; in deactivate_slab() 2438 if (l != m) { in deactivate_slab() 2444 if (m == M_PARTIAL) in deactivate_slab() 2446 else if (m == M_FULL) in deactivate_slab() 2450 l = m; in deactivate_slab() 2460 if (m == M_PARTIAL) in deactivate_slab() 2462 else if (m == M_FULL) in deactivate_slab() [all …]
|
D | slab.c | 4091 void slabinfo_show_stats(struct seq_file *m, struct kmem_cache *cachep) in slabinfo_show_stats() argument 4105 seq_printf(m, " : globalstat %7lu %6lu %5lu %4lu %4lu %4lu %4lu %4lu %4lu", in slabinfo_show_stats() 4117 seq_printf(m, " : cpustat %6lu %6lu %6lu %6lu", in slabinfo_show_stats()
|