Home
last modified time | relevance | path

Searched refs:r (Results 1 – 13 of 13) sorted by relevance

/mm/damon/
Dcore-test.h19 struct damon_region *r; in damon_test_regions() local
22 r = damon_new_region(1, 2); in damon_test_regions()
23 KUNIT_EXPECT_EQ(test, 1ul, r->ar.start); in damon_test_regions()
24 KUNIT_EXPECT_EQ(test, 2ul, r->ar.end); in damon_test_regions()
25 KUNIT_EXPECT_EQ(test, 0u, r->nr_accesses); in damon_test_regions()
30 damon_add_region(r, t); in damon_test_regions()
33 damon_del_region(r, t); in damon_test_regions()
86 struct damon_region *r; in damon_test_aggregate() local
94 r = damon_new_region(saddr[it][ir], eaddr[it][ir]); in damon_test_aggregate()
95 r->nr_accesses = accesses[it][ir]; in damon_test_aggregate()
[all …]
Dcore.c52 void damon_add_region(struct damon_region *r, struct damon_target *t) in damon_add_region() argument
54 list_add_tail(&r->list, &t->regions_list); in damon_add_region()
58 static void damon_del_region(struct damon_region *r, struct damon_target *t) in damon_del_region() argument
60 list_del(&r->list); in damon_del_region()
64 static void damon_free_region(struct damon_region *r) in damon_free_region() argument
66 kfree(r); in damon_free_region()
69 void damon_destroy_region(struct damon_region *r, struct damon_target *t) in damon_destroy_region() argument
71 damon_del_region(r, t); in damon_destroy_region()
72 damon_free_region(r); in damon_destroy_region()
180 struct damon_region *r, *next; in damon_free_target() local
[all …]
Dvaddr.c65 struct damon_region *r, unsigned int nr_pieces) in damon_va_evenly_split_region() argument
71 if (!r || !nr_pieces) in damon_va_evenly_split_region()
74 orig_end = r->ar.end; in damon_va_evenly_split_region()
75 sz_orig = r->ar.end - r->ar.start; in damon_va_evenly_split_region()
81 r->ar.end = r->ar.start + sz_piece; in damon_va_evenly_split_region()
82 next = damon_next_region(r); in damon_va_evenly_split_region()
83 for (start = r->ar.end; start + sz_piece <= orig_end; in damon_va_evenly_split_region()
88 damon_insert_region(n, r, next, t); in damon_va_evenly_split_region()
89 r = n; in damon_va_evenly_split_region()
98 static unsigned long sz_range(struct damon_addr_range *r) in sz_range() argument
[all …]
Dpaddr.c69 struct damon_region *r) in __damon_pa_prepare_access_check() argument
71 r->sampling_addr = damon_rand(r->ar.start, r->ar.end); in __damon_pa_prepare_access_check()
73 damon_pa_mkold(r->sampling_addr); in __damon_pa_prepare_access_check()
79 struct damon_region *r; in damon_pa_prepare_access_checks() local
82 damon_for_each_region(r, t) in damon_pa_prepare_access_checks()
83 __damon_pa_prepare_access_check(ctx, r); in damon_pa_prepare_access_checks()
174 struct damon_region *r) in __damon_pa_check_access() argument
182 ALIGN_DOWN(r->sampling_addr, last_page_sz)) { in __damon_pa_check_access()
184 r->nr_accesses++; in __damon_pa_check_access()
188 last_accessed = damon_pa_young(r->sampling_addr, &last_page_sz); in __damon_pa_check_access()
[all …]
Dvaddr-test.h100 struct damon_region *r; in __nth_region_of() local
103 damon_for_each_region(r, t) { in __nth_region_of()
105 return r; in __nth_region_of()
139 struct damon_region *r; in damon_do_test_apply_three_regions() local
144 r = damon_new_region(regions[i * 2], regions[i * 2 + 1]); in damon_do_test_apply_three_regions()
145 damon_add_region(r, t); in damon_do_test_apply_three_regions()
151 r = __nth_region_of(t, i); in damon_do_test_apply_three_regions()
152 KUNIT_EXPECT_EQ(test, r->ar.start, expected[i * 2]); in damon_do_test_apply_three_regions()
153 KUNIT_EXPECT_EQ(test, r->ar.end, expected[i * 2 + 1]); in damon_do_test_apply_three_regions()
257 struct damon_region *r = damon_new_region(start, end); in damon_test_split_evenly_fail() local
[all …]
Ddbgfs.c70 unsigned long s, a, r, minr, maxr; in dbgfs_attrs_write() local
79 &s, &a, &r, &minr, &maxr) != 5) { in dbgfs_attrs_write()
90 ret = damon_set_attrs(ctx, s, a, r, minr, maxr); in dbgfs_attrs_write()
442 struct damon_region *r; in sprint_init_regions() local
447 damon_for_each_region(r, t) { in sprint_init_regions()
450 t->id, r->ar.start, r->ar.end); in sprint_init_regions()
492 struct damon_region *r, *prev; in add_init_region() local
504 r = damon_new_region(ar->start, ar->end); in add_init_region()
505 if (!r) in add_init_region()
507 damon_add_region(r, t); in add_init_region()
[all …]
Dprmtv-common.c92 int damon_pageout_score(struct damon_ctx *c, struct damon_region *r, in damon_pageout_score() argument
104 freq_subscore = r->nr_accesses * DAMON_MAX_SUBSCORE / max_nr_accesses; in damon_pageout_score()
106 age_in_sec = (unsigned long)r->age * c->aggr_interval / 1000000; in damon_pageout_score()
Dprmtv-common.h15 int damon_pageout_score(struct damon_ctx *c, struct damon_region *r,
/mm/
Dmemblock.c340 static void __init_memblock memblock_remove_region(struct memblock_type *type, unsigned long r) in memblock_remove_region() argument
342 type->total_size -= type->regions[r].size; in memblock_remove_region()
343 memmove(&type->regions[r], &type->regions[r + 1], in memblock_remove_region()
344 (type->cnt - (r + 1)) * sizeof(type->regions[r])); in memblock_remove_region()
866 struct memblock_region *r = &type->regions[i]; in memblock_setclr_flag() local
869 r->flags |= flag; in memblock_setclr_flag()
871 r->flags &= ~flag; in memblock_setclr_flag()
1032 struct memblock_region *r; in __next_mem_range() local
1036 r = &type_b->regions[idx_b]; in __next_mem_range()
1037 r_start = idx_b ? r[-1].base + r[-1].size : 0; in __next_mem_range()
[all …]
Dhmm.c385 int r; in hmm_vma_walk_pmd() local
387 r = hmm_vma_handle_pte(walk, addr, end, pmdp, ptep, hmm_pfns); in hmm_vma_walk_pmd()
388 if (r) { in hmm_vma_walk_pmd()
390 return r; in hmm_vma_walk_pmd()
Dpage_alloc.c6310 static struct memblock_region *r; in overlap_memmap_init() local
6313 if (!r || *pfn >= memblock_region_memory_end_pfn(r)) { in overlap_memmap_init()
6314 for_each_mem_region(r) { in overlap_memmap_init()
6315 if (*pfn < memblock_region_memory_end_pfn(r)) in overlap_memmap_init()
6319 if (*pfn >= memblock_region_memory_base_pfn(r) && in overlap_memmap_init()
6320 memblock_is_mirror(r)) { in overlap_memmap_init()
6321 *pfn = memblock_region_memory_end_pfn(r); in overlap_memmap_init()
6984 struct memblock_region *r; in zone_absent_pages_in_node() local
6986 for_each_mem_region(r) { in zone_absent_pages_in_node()
6987 start_pfn = clamp(memblock_region_memory_base_pfn(r), in zone_absent_pages_in_node()
[all …]
Dmigrate.c2991 int r; in migrate_vma_pages() local
3039 r = migrate_page(mapping, newpage, page, MIGRATE_SYNC_NO_COPY); in migrate_vma_pages()
3040 if (r != MIGRATEPAGE_SUCCESS) in migrate_vma_pages()
/mm/kfence/
Dkfence_test.c77 static const char *get_access_type(const struct expect_report *r) in get_access_type() argument
79 return r->is_write ? "write" : "read"; in get_access_type()
83 static bool report_matches(const struct expect_report *r) in report_matches() argument
100 switch (r->type) { in report_matches()
103 get_access_type(r)); in report_matches()
107 get_access_type(r)); in report_matches()
114 get_access_type(r)); in report_matches()
121 scnprintf(cur, end - cur, " in %pS", r->fn); in report_matches()
131 switch (r->type) { in report_matches()
133 cur += scnprintf(cur, end - cur, "Out-of-bounds %s at", get_access_type(r)); in report_matches()
[all …]