Home
last modified time | relevance | path

Searched refs:pos (Results 1 – 18 of 18) sorted by relevance

/mm/
Dpage_idle.c123 loff_t pos, size_t count) in page_idle_bitmap_read() argument
130 if (pos % BITMAP_CHUNK_SIZE || count % BITMAP_CHUNK_SIZE) in page_idle_bitmap_read()
133 pfn = pos * BITS_PER_BYTE; in page_idle_bitmap_read()
168 loff_t pos, size_t count) in page_idle_bitmap_write() argument
175 if (pos % BITMAP_CHUNK_SIZE || count % BITMAP_CHUNK_SIZE) in page_idle_bitmap_write()
178 pfn = pos * BITS_PER_BYTE; in page_idle_bitmap_write()
Dfilemap.c2905 static int generic_write_check_limits(struct file *file, loff_t pos, in generic_write_check_limits() argument
2913 if (pos >= limit) { in generic_write_check_limits()
2917 *count = min(*count, limit - pos); in generic_write_check_limits()
2923 if (unlikely(pos >= max_size)) in generic_write_check_limits()
2926 *count = min(*count, max_size - pos); in generic_write_check_limits()
3123 loff_t pos, unsigned len, unsigned flags, in pagecache_write_begin() argument
3128 return aops->write_begin(file, mapping, pos, len, flags, in pagecache_write_begin()
3134 loff_t pos, unsigned len, unsigned copied, in pagecache_write_end() argument
3139 return aops->write_end(file, mapping, pos, len, copied, page, fsdata); in pagecache_write_end()
3149 loff_t pos = iocb->ki_pos; in generic_file_direct_write() local
[all …]
Dvmstat.c1303 static void *frag_start(struct seq_file *m, loff_t *pos) in frag_start() argument
1306 loff_t node = *pos; in frag_start()
1316 static void *frag_next(struct seq_file *m, void *arg, loff_t *pos) in frag_next() argument
1320 (*pos)++; in frag_next()
1670 static void *vmstat_start(struct seq_file *m, loff_t *pos) in vmstat_start() argument
1675 if (*pos >= ARRAY_SIZE(vmstat_text)) in vmstat_start()
1715 return (unsigned long *)m->private + *pos; in vmstat_start()
1718 static void *vmstat_next(struct seq_file *m, void *arg, loff_t *pos) in vmstat_next() argument
1720 (*pos)++; in vmstat_next()
1721 if (*pos >= ARRAY_SIZE(vmstat_text)) in vmstat_next()
[all …]
Dslab_common.c1415 void *slab_start(struct seq_file *m, loff_t *pos) in slab_start() argument
1418 return seq_list_start(&slab_root_caches, *pos); in slab_start()
1421 void *slab_next(struct seq_file *m, void *p, loff_t *pos) in slab_next() argument
1423 return seq_list_next(p, &slab_root_caches, pos); in slab_next()
1518 void *memcg_slab_start(struct seq_file *m, loff_t *pos) in memcg_slab_start() argument
1523 return seq_list_start(&memcg->kmem_caches, *pos); in memcg_slab_start()
1526 void *memcg_slab_next(struct seq_file *m, void *p, loff_t *pos) in memcg_slab_next() argument
1530 return seq_list_next(p, &memcg->kmem_caches, pos); in memcg_slab_next()
Dslab.h643 void *slab_start(struct seq_file *m, loff_t *pos);
644 void *slab_next(struct seq_file *m, void *p, loff_t *pos);
646 void *memcg_slab_start(struct seq_file *m, loff_t *pos);
647 void *memcg_slab_next(struct seq_file *m, void *p, loff_t *pos);
Dzswap.c963 unsigned int pos; in zswap_is_page_same_filled() local
967 for (pos = 1; pos < PAGE_SIZE / sizeof(*page); pos++) { in zswap_is_page_same_filled()
968 if (page[pos] != page[0]) in zswap_is_page_same_filled()
Dslub.c1550 unsigned long *pos, void *start, in next_freelist_entry() argument
1561 idx = s->random_seq[*pos]; in next_freelist_entry()
1562 *pos += 1; in next_freelist_entry()
1563 if (*pos >= freelist_count) in next_freelist_entry()
1564 *pos = 0; in next_freelist_entry()
1576 unsigned long idx, pos, page_limit, freelist_count; in shuffle_freelist() local
1582 pos = get_random_int() % freelist_count; in shuffle_freelist()
1588 cur = next_freelist_entry(s, page, &pos, start, page_limit, in shuffle_freelist()
1594 next = next_freelist_entry(s, page, &pos, start, page_limit, in shuffle_freelist()
4523 long start, end, pos; in add_location() local
[all …]
Dshmem.c463 LIST_HEAD(list), *pos, *next; in shmem_unused_huge_shrink()
475 list_for_each_safe(pos, next, &sbinfo->shrinklist) { in shmem_unused_huge_shrink()
476 info = list_entry(pos, struct shmem_inode_info, shrinklist); in shmem_unused_huge_shrink()
503 list_for_each_safe(pos, next, &to_remove) { in shmem_unused_huge_shrink()
504 info = list_entry(pos, struct shmem_inode_info, shrinklist); in shmem_unused_huge_shrink()
510 list_for_each_safe(pos, next, &list) { in shmem_unused_huge_shrink()
513 info = list_entry(pos, struct shmem_inode_info, shrinklist); in shmem_unused_huge_shrink()
2471 loff_t pos, unsigned len, unsigned flags, in shmem_write_begin() argument
2476 pgoff_t index = pos >> PAGE_SHIFT; in shmem_write_begin()
2483 if ((info->seals & F_SEAL_GROW) && pos + len > inode->i_size) in shmem_write_begin()
[all …]
Dzsmalloc.c1981 int offset, pos; in zs_page_migrate() local
2016 pos = offset; in zs_page_migrate()
2018 while (pos < PAGE_SIZE) { in zs_page_migrate()
2019 head = obj_to_head(page, s_addr + pos); in zs_page_migrate()
2025 pos += class->size; in zs_page_migrate()
2035 for (addr = s_addr + offset; addr < s_addr + pos; in zs_page_migrate()
2083 for (addr = s_addr + offset; addr < s_addr + pos; in zs_page_migrate()
Dkmemleak.c1614 static void *kmemleak_seq_start(struct seq_file *seq, loff_t *pos) in kmemleak_seq_start() argument
1617 loff_t n = *pos; in kmemleak_seq_start()
1640 static void *kmemleak_seq_next(struct seq_file *seq, void *v, loff_t *pos) in kmemleak_seq_next() argument
1646 ++(*pos); in kmemleak_seq_next()
Dmemcontrol.c1032 struct mem_cgroup *pos = NULL; in mem_cgroup_iter() local
1041 pos = prev; in mem_cgroup_iter()
1061 pos = READ_ONCE(iter->position); in mem_cgroup_iter()
1062 if (!pos || css_tryget(&pos->css)) in mem_cgroup_iter()
1072 (void)cmpxchg(&iter->position, pos, NULL); in mem_cgroup_iter()
1076 if (pos) in mem_cgroup_iter()
1077 css = &pos->css; in mem_cgroup_iter()
1115 (void)cmpxchg(&iter->position, pos, memcg); in mem_cgroup_iter()
1117 if (pos) in mem_cgroup_iter()
1118 css_put(&pos->css); in mem_cgroup_iter()
Dz3fold.c1117 struct list_head *pos; in z3fold_reclaim_page() local
1131 list_for_each_prev(pos, &pool->lru) { in z3fold_reclaim_page()
1132 page = list_entry(pos, struct page, lru); in z3fold_reclaim_page()
Dhuge_memory.c2916 LIST_HEAD(list), *pos, *next; in deferred_split_scan()
2927 list_for_each_safe(pos, next, &ds_queue->split_queue) { in deferred_split_scan()
2928 page = list_entry((void *)pos, struct page, mapping); in deferred_split_scan()
2942 list_for_each_safe(pos, next, &list) { in deferred_split_scan()
2943 page = list_entry((void *)pos, struct page, mapping); in deferred_split_scan()
Dvmalloc.c3419 static void *s_start(struct seq_file *m, loff_t *pos) in s_start() argument
3423 return seq_list_start(&vmap_area_list, *pos); in s_start()
3426 static void *s_next(struct seq_file *m, void *p, loff_t *pos) in s_next() argument
3428 return seq_list_next(p, &vmap_area_list, pos); in s_next()
Dpage_alloc.c7494 void *pos; in free_reserved_area() local
7499 for (pos = start; pos < end; pos += PAGE_SIZE, pages++) { in free_reserved_area()
7500 struct page *page = virt_to_page(pos); in free_reserved_area()
7555 #define adj_init_size(start, end, size, pos, adj) \ in mem_init_print_info() argument
7557 if (start <= pos && pos < end && size > adj) \ in mem_init_print_info()
Dslab.c2378 unsigned int pos; member
2406 state->pos = rand % count; in freelist_state_initialize()
2415 if (state->pos >= state->count) in next_random_slot()
2416 state->pos = 0; in next_random_slot()
2417 return state->list[state->pos++]; in next_random_slot()
Dpercpu.c1979 struct pcpu_chunk *pos; in free_percpu() local
1981 list_for_each_entry(pos, &pcpu_slot[pcpu_nr_slots - 1], list) in free_percpu()
1982 if (pos != chunk) { in free_percpu()
Dswapfile.c2709 static void *swap_start(struct seq_file *swap, loff_t *pos) in swap_start() argument
2713 loff_t l = *pos; in swap_start()
2730 static void *swap_next(struct seq_file *swap, void *v, loff_t *pos) in swap_next() argument
2743 ++*pos; in swap_next()