/drivers/base/ |
D | memory.c | 193 static bool pages_correctly_reserved(unsigned long start_pfn) in pages_correctly_reserved() argument 197 unsigned long pfn = start_pfn; in pages_correctly_reserved() 232 unsigned long start_pfn; in memory_block_action() local 236 start_pfn = section_nr_to_pfn(phys_index); in memory_block_action() 240 if (!pages_correctly_reserved(start_pfn)) in memory_block_action() 243 ret = online_pages(start_pfn, nr_pages, online_type); in memory_block_action() 246 ret = offline_pages(start_pfn, nr_pages); in memory_block_action() 392 static void print_allowed_zone(char *buf, int nid, unsigned long start_pfn, in print_allowed_zone() argument 398 zone = zone_for_pfn_range(online_type, nid, start_pfn, nr_pages); in print_allowed_zone() 409 unsigned long start_pfn = section_nr_to_pfn(mem->start_section_nr); in show_valid_zones() local [all …]
|
D | node.c | 486 int link_mem_sections(int nid, unsigned long start_pfn, unsigned long nr_pages) in link_mem_sections() argument 488 unsigned long end_pfn = start_pfn + nr_pages; in link_mem_sections() 493 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in link_mem_sections()
|
/drivers/hv/ |
D | hv_balloon.c | 439 unsigned long start_pfn; member 452 unsigned long start_pfn; member 630 pfn_to_page(gap->start_pfn); in hv_page_online_one() 649 unsigned long start_pfn, unsigned long size) in hv_bring_pgs_online() argument 653 pr_debug("Online %lu pages starting at pfn 0x%lx\n", size, start_pfn); in hv_bring_pgs_online() 655 hv_page_online_one(has, pfn_to_page(start_pfn + i)); in hv_bring_pgs_online() 664 unsigned long start_pfn; in hv_mem_hot_add() local 670 start_pfn = start + (i * HA_CHUNK); in hv_mem_hot_add() 689 nid = memory_add_physaddr_to_nid(PFN_PHYS(start_pfn)); in hv_mem_hot_add() 690 ret = add_memory(nid, PFN_PHYS((start_pfn)), in hv_mem_hot_add() [all …]
|
/drivers/iommu/ |
D | intel-iommu.c | 1077 unsigned long start_pfn, in dma_pte_clear_range() argument 1083 BUG_ON(!domain_pfn_supported(domain, start_pfn)); in dma_pte_clear_range() 1085 BUG_ON(start_pfn > last_pfn); in dma_pte_clear_range() 1090 first_pte = pte = dma_pfn_level_pte(domain, start_pfn, 1, &large_page); in dma_pte_clear_range() 1092 start_pfn = align_to_level(start_pfn + 1, large_page + 1); in dma_pte_clear_range() 1097 start_pfn += lvl_to_nr_pages(large_page); in dma_pte_clear_range() 1099 } while (start_pfn <= last_pfn && !first_pte_in_page(pte)); in dma_pte_clear_range() 1104 } while (start_pfn && start_pfn <= last_pfn); in dma_pte_clear_range() 1109 unsigned long pfn, unsigned long start_pfn, in dma_pte_free_level() argument 1112 pfn = max(start_pfn, pfn); in dma_pte_free_level() [all …]
|
D | iova.c | 40 unsigned long start_pfn, unsigned long pfn_32bit) in init_iova_domain() argument 53 iovad->start_pfn = start_pfn; in init_iova_domain() 239 if ((iovad->start_pfn + size + pad_size) > limit_pfn) { in __alloc_and_insert_iova_range()
|
D | dma-iommu.c | 309 if (iovad->start_pfn) { in iommu_dma_init_domain() 311 base_pfn != iovad->start_pfn) { in iommu_dma_init_domain()
|
/drivers/xen/ |
D | balloon.c | 729 static void __init balloon_add_region(unsigned long start_pfn, in balloon_add_region() argument 740 extra_pfn_end = min(max_pfn, start_pfn + pages); in balloon_add_region() 742 for (pfn = start_pfn; pfn < extra_pfn_end; pfn++) { in balloon_add_region() 750 balloon_stats.total_pages += extra_pfn_end - start_pfn; in balloon_add_region() 794 balloon_add_region(xen_extra_mem[i].start_pfn, in balloon_init()
|
/drivers/net/ethernet/ibm/ehea/ |
D | ehea_qmr.c | 709 unsigned long pfn, start_pfn, end_pfn, nr_pages; in ehea_create_busmap_callback() local 716 start_pfn = initial_pfn; in ehea_create_busmap_callback() 718 pfn = start_pfn; in ehea_create_busmap_callback() 723 nr_pages = pfn - start_pfn; in ehea_create_busmap_callback() 724 ret = ehea_update_busmap(start_pfn, nr_pages, in ehea_create_busmap_callback() 731 start_pfn = pfn; in ehea_create_busmap_callback() 737 nr_pages = pfn - start_pfn; in ehea_create_busmap_callback() 738 return ehea_update_busmap(start_pfn, nr_pages, EHEA_BUSMAP_ADD_SECT); in ehea_create_busmap_callback()
|
D | ehea_main.c | 3289 if (ehea_add_sect_bmap(arg->start_pfn, arg->nr_pages)) in ehea_mem_notifier() 3297 if (ehea_rem_sect_bmap(arg->start_pfn, arg->nr_pages)) in ehea_mem_notifier()
|
/drivers/s390/char/ |
D | sclp_cmd.c | 180 int arch_get_memory_phys_device(unsigned long start_pfn) in arch_get_memory_phys_device() argument 184 return PFN_PHYS(start_pfn) >> ilog2(sclp.rzm); in arch_get_memory_phys_device() 327 start = arg->start_pfn << PAGE_SHIFT; in sclp_mem_notifier()
|
/drivers/gpu/drm/gma500/ |
D | mmu.h | 78 uint32_t start_pfn,
|
D | mmu.c | 660 int psb_mmu_insert_pfn_sequence(struct psb_mmu_pd *pd, uint32_t start_pfn, in psb_mmu_insert_pfn_sequence() argument 685 pte = psb_mmu_mask_pte(start_pfn++, type); in psb_mmu_insert_pfn_sequence()
|