/mm/damon/ |
D | core-test.h | 19 struct damon_region *r; in damon_test_regions() local 22 r = damon_new_region(1, 2); in damon_test_regions() 23 KUNIT_EXPECT_EQ(test, 1ul, r->ar.start); in damon_test_regions() 24 KUNIT_EXPECT_EQ(test, 2ul, r->ar.end); in damon_test_regions() 25 KUNIT_EXPECT_EQ(test, 0u, r->nr_accesses); in damon_test_regions() 30 damon_add_region(r, t); in damon_test_regions() 33 damon_del_region(r, t); in damon_test_regions() 86 struct damon_region *r; in damon_test_aggregate() local 94 r = damon_new_region(saddr[it][ir], eaddr[it][ir]); in damon_test_aggregate() 95 r->nr_accesses = accesses[it][ir]; in damon_test_aggregate() [all …]
|
D | core.c | 52 void damon_add_region(struct damon_region *r, struct damon_target *t) in damon_add_region() argument 54 list_add_tail(&r->list, &t->regions_list); in damon_add_region() 58 static void damon_del_region(struct damon_region *r, struct damon_target *t) in damon_del_region() argument 60 list_del(&r->list); in damon_del_region() 64 static void damon_free_region(struct damon_region *r) in damon_free_region() argument 66 kfree(r); in damon_free_region() 69 void damon_destroy_region(struct damon_region *r, struct damon_target *t) in damon_destroy_region() argument 71 damon_del_region(r, t); in damon_destroy_region() 72 damon_free_region(r); in damon_destroy_region() 180 struct damon_region *r, *next; in damon_free_target() local [all …]
|
D | vaddr.c | 65 struct damon_region *r, unsigned int nr_pieces) in damon_va_evenly_split_region() argument 71 if (!r || !nr_pieces) in damon_va_evenly_split_region() 74 orig_end = r->ar.end; in damon_va_evenly_split_region() 75 sz_orig = r->ar.end - r->ar.start; in damon_va_evenly_split_region() 81 r->ar.end = r->ar.start + sz_piece; in damon_va_evenly_split_region() 82 next = damon_next_region(r); in damon_va_evenly_split_region() 83 for (start = r->ar.end; start + sz_piece <= orig_end; in damon_va_evenly_split_region() 88 damon_insert_region(n, r, next, t); in damon_va_evenly_split_region() 89 r = n; in damon_va_evenly_split_region() 98 static unsigned long sz_range(struct damon_addr_range *r) in sz_range() argument [all …]
|
D | paddr.c | 69 struct damon_region *r) in __damon_pa_prepare_access_check() argument 71 r->sampling_addr = damon_rand(r->ar.start, r->ar.end); in __damon_pa_prepare_access_check() 73 damon_pa_mkold(r->sampling_addr); in __damon_pa_prepare_access_check() 79 struct damon_region *r; in damon_pa_prepare_access_checks() local 82 damon_for_each_region(r, t) in damon_pa_prepare_access_checks() 83 __damon_pa_prepare_access_check(ctx, r); in damon_pa_prepare_access_checks() 174 struct damon_region *r) in __damon_pa_check_access() argument 182 ALIGN_DOWN(r->sampling_addr, last_page_sz)) { in __damon_pa_check_access() 184 r->nr_accesses++; in __damon_pa_check_access() 188 last_accessed = damon_pa_young(r->sampling_addr, &last_page_sz); in __damon_pa_check_access() [all …]
|
D | vaddr-test.h | 100 struct damon_region *r; in __nth_region_of() local 103 damon_for_each_region(r, t) { in __nth_region_of() 105 return r; in __nth_region_of() 139 struct damon_region *r; in damon_do_test_apply_three_regions() local 144 r = damon_new_region(regions[i * 2], regions[i * 2 + 1]); in damon_do_test_apply_three_regions() 145 damon_add_region(r, t); in damon_do_test_apply_three_regions() 151 r = __nth_region_of(t, i); in damon_do_test_apply_three_regions() 152 KUNIT_EXPECT_EQ(test, r->ar.start, expected[i * 2]); in damon_do_test_apply_three_regions() 153 KUNIT_EXPECT_EQ(test, r->ar.end, expected[i * 2 + 1]); in damon_do_test_apply_three_regions() 257 struct damon_region *r = damon_new_region(start, end); in damon_test_split_evenly_fail() local [all …]
|
D | dbgfs.c | 70 unsigned long s, a, r, minr, maxr; in dbgfs_attrs_write() local 79 &s, &a, &r, &minr, &maxr) != 5) { in dbgfs_attrs_write() 90 ret = damon_set_attrs(ctx, s, a, r, minr, maxr); in dbgfs_attrs_write() 442 struct damon_region *r; in sprint_init_regions() local 447 damon_for_each_region(r, t) { in sprint_init_regions() 450 t->id, r->ar.start, r->ar.end); in sprint_init_regions() 492 struct damon_region *r, *prev; in add_init_region() local 504 r = damon_new_region(ar->start, ar->end); in add_init_region() 505 if (!r) in add_init_region() 507 damon_add_region(r, t); in add_init_region() [all …]
|
D | prmtv-common.c | 92 int damon_pageout_score(struct damon_ctx *c, struct damon_region *r, in damon_pageout_score() argument 104 freq_subscore = r->nr_accesses * DAMON_MAX_SUBSCORE / max_nr_accesses; in damon_pageout_score() 106 age_in_sec = (unsigned long)r->age * c->aggr_interval / 1000000; in damon_pageout_score()
|
D | prmtv-common.h | 15 int damon_pageout_score(struct damon_ctx *c, struct damon_region *r,
|
/mm/ |
D | memblock.c | 340 static void __init_memblock memblock_remove_region(struct memblock_type *type, unsigned long r) in memblock_remove_region() argument 342 type->total_size -= type->regions[r].size; in memblock_remove_region() 343 memmove(&type->regions[r], &type->regions[r + 1], in memblock_remove_region() 344 (type->cnt - (r + 1)) * sizeof(type->regions[r])); in memblock_remove_region() 866 struct memblock_region *r = &type->regions[i]; in memblock_setclr_flag() local 869 r->flags |= flag; in memblock_setclr_flag() 871 r->flags &= ~flag; in memblock_setclr_flag() 1032 struct memblock_region *r; in __next_mem_range() local 1036 r = &type_b->regions[idx_b]; in __next_mem_range() 1037 r_start = idx_b ? r[-1].base + r[-1].size : 0; in __next_mem_range() [all …]
|
D | hmm.c | 385 int r; in hmm_vma_walk_pmd() local 387 r = hmm_vma_handle_pte(walk, addr, end, pmdp, ptep, hmm_pfns); in hmm_vma_walk_pmd() 388 if (r) { in hmm_vma_walk_pmd() 390 return r; in hmm_vma_walk_pmd()
|
D | page_alloc.c | 6310 static struct memblock_region *r; in overlap_memmap_init() local 6313 if (!r || *pfn >= memblock_region_memory_end_pfn(r)) { in overlap_memmap_init() 6314 for_each_mem_region(r) { in overlap_memmap_init() 6315 if (*pfn < memblock_region_memory_end_pfn(r)) in overlap_memmap_init() 6319 if (*pfn >= memblock_region_memory_base_pfn(r) && in overlap_memmap_init() 6320 memblock_is_mirror(r)) { in overlap_memmap_init() 6321 *pfn = memblock_region_memory_end_pfn(r); in overlap_memmap_init() 6984 struct memblock_region *r; in zone_absent_pages_in_node() local 6986 for_each_mem_region(r) { in zone_absent_pages_in_node() 6987 start_pfn = clamp(memblock_region_memory_base_pfn(r), in zone_absent_pages_in_node() [all …]
|
D | migrate.c | 2991 int r; in migrate_vma_pages() local 3039 r = migrate_page(mapping, newpage, page, MIGRATE_SYNC_NO_COPY); in migrate_vma_pages() 3040 if (r != MIGRATEPAGE_SUCCESS) in migrate_vma_pages()
|
/mm/kfence/ |
D | kfence_test.c | 77 static const char *get_access_type(const struct expect_report *r) in get_access_type() argument 79 return r->is_write ? "write" : "read"; in get_access_type() 83 static bool report_matches(const struct expect_report *r) in report_matches() argument 100 switch (r->type) { in report_matches() 103 get_access_type(r)); in report_matches() 107 get_access_type(r)); in report_matches() 114 get_access_type(r)); in report_matches() 121 scnprintf(cur, end - cur, " in %pS", r->fn); in report_matches() 131 switch (r->type) { in report_matches() 133 cur += scnprintf(cur, end - cur, "Out-of-bounds %s at", get_access_type(r)); in report_matches() [all …]
|