Home
last modified time | relevance | path

Searched refs:pgmap (Results 1 – 9 of 9) sorted by relevance

/mm/
Dmemremap.c44 static void devmap_managed_enable_put(struct dev_pagemap *pgmap) in devmap_managed_enable_put() argument
46 if (pgmap->type == MEMORY_DEVICE_PRIVATE || in devmap_managed_enable_put()
47 pgmap->type == MEMORY_DEVICE_FS_DAX) in devmap_managed_enable_put()
51 static void devmap_managed_enable_get(struct dev_pagemap *pgmap) in devmap_managed_enable_get() argument
53 if (pgmap->type == MEMORY_DEVICE_PRIVATE || in devmap_managed_enable_get()
54 pgmap->type == MEMORY_DEVICE_FS_DAX) in devmap_managed_enable_get()
58 static void devmap_managed_enable_get(struct dev_pagemap *pgmap) in devmap_managed_enable_get() argument
61 static void devmap_managed_enable_put(struct dev_pagemap *pgmap) in devmap_managed_enable_put() argument
73 static unsigned long pfn_first(struct dev_pagemap *pgmap, int range_id) in pfn_first() argument
75 struct range *range = &pgmap->ranges[range_id]; in pfn_first()
[all …]
Dgup.c29 struct dev_pagemap *pgmap; member
427 struct dev_pagemap **pgmap) in follow_page_pte() argument
489 *pgmap = get_dev_pagemap(pte_pfn(pte), *pgmap); in follow_page_pte()
490 if (*pgmap) in follow_page_pte()
638 page = follow_devmap_pmd(vma, address, pmd, flags, &ctx->pgmap); in follow_pmd_mask()
644 return follow_page_pte(vma, address, pmd, flags, &ctx->pgmap); in follow_pmd_mask()
664 return follow_page_pte(vma, address, pmd, flags, &ctx->pgmap); in follow_pmd_mask()
694 follow_page_pte(vma, address, pmd, flags, &ctx->pgmap); in follow_pmd_mask()
731 page = follow_devmap_pud(vma, address, pud, flags, &ctx->pgmap); in follow_pud_mask()
834 if (ctx.pgmap) in follow_page()
[all …]
Dmemory-failure.c1223 struct dev_pagemap *pgmap) in memory_failure_dev_pagemap() argument
1241 if (!pgmap_pfn_valid(pgmap, pfn)) { in memory_failure_dev_pagemap()
1262 if (pgmap->type == MEMORY_DEVICE_PRIVATE) { in memory_failure_dev_pagemap()
1304 put_dev_pagemap(pgmap); in memory_failure_dev_pagemap()
1331 struct dev_pagemap *pgmap; in memory_failure() local
1341 pgmap = get_dev_pagemap(pfn, NULL); in memory_failure()
1342 if (pgmap) in memory_failure()
1344 pgmap); in memory_failure()
Dhuge_memory.c965 pmd_t *pmd, int flags, struct dev_pagemap **pgmap) in follow_devmap_pmd() argument
1003 *pgmap = get_dev_pagemap(pfn, *pgmap); in follow_devmap_pmd()
1004 if (!*pgmap) in follow_devmap_pmd()
1137 pud_t *pud, int flags, struct dev_pagemap **pgmap) in follow_devmap_pud() argument
1171 *pgmap = get_dev_pagemap(pfn, *pgmap); in follow_devmap_pud()
1172 if (!*pgmap) in follow_devmap_pud()
Dhmm.c217 device_private_entry_to_page(entry)->pgmap->owner == in hmm_is_device_private_entry()
Dswap.c121 put_dev_pagemap(page->pgmap); in __put_page()
Dpage_alloc.c6400 struct dev_pagemap *pgmap) in memmap_init_zone_device() argument
6404 struct vmem_altmap *altmap = pgmap_altmap(pgmap); in memmap_init_zone_device()
6409 if (WARN_ON_ONCE(!pgmap || zone_idx(zone) != ZONE_DEVICE)) in memmap_init_zone_device()
6441 page->pgmap = pgmap; in memmap_init_zone_device()
Dmigrate.c2380 page->pgmap->owner != migrate->pgmap_owner) in migrate_vma_collect_pmd()
Dmemory.c3661 ret = vmf->page->pgmap->ops->migrate_to_ram(vmf); in do_swap_page()