Home
last modified time | relevance | path

Searched full:pmd (Results 1 – 25 of 841) sorted by relevance

12345678910>>...34

/kernel/linux/linux-4.19/drivers/md/
Ddm-thin-metadata.c214 struct dm_pool_metadata *pmd; member
370 static int superblock_lock_zero(struct dm_pool_metadata *pmd, in superblock_lock_zero() argument
373 return dm_bm_write_lock_zero(pmd->bm, THIN_SUPERBLOCK_LOCATION, in superblock_lock_zero()
377 static int superblock_lock(struct dm_pool_metadata *pmd, in superblock_lock() argument
380 return dm_bm_write_lock(pmd->bm, THIN_SUPERBLOCK_LOCATION, in superblock_lock()
413 static void __setup_btree_details(struct dm_pool_metadata *pmd) in __setup_btree_details() argument
415 pmd->info.tm = pmd->tm; in __setup_btree_details()
416 pmd->info.levels = 2; in __setup_btree_details()
417 pmd->info.value_type.context = pmd->data_sm; in __setup_btree_details()
418 pmd->info.value_type.size = sizeof(__le64); in __setup_btree_details()
[all …]
Ddm-thin-metadata.h48 int dm_pool_metadata_close(struct dm_pool_metadata *pmd);
61 int dm_pool_create_thin(struct dm_pool_metadata *pmd, dm_thin_id dev);
69 int dm_pool_create_snap(struct dm_pool_metadata *pmd, dm_thin_id dev,
77 int dm_pool_delete_thin_device(struct dm_pool_metadata *pmd,
84 int dm_pool_commit_metadata(struct dm_pool_metadata *pmd);
94 int dm_pool_abort_metadata(struct dm_pool_metadata *pmd);
99 int dm_pool_set_metadata_transaction_id(struct dm_pool_metadata *pmd,
103 int dm_pool_get_metadata_transaction_id(struct dm_pool_metadata *pmd,
115 int dm_pool_reserve_metadata_snap(struct dm_pool_metadata *pmd);
116 int dm_pool_release_metadata_snap(struct dm_pool_metadata *pmd);
[all …]
/kernel/linux/linux-5.10/drivers/md/
Ddm-thin-metadata.c230 struct dm_pool_metadata *pmd; member
390 static inline void pmd_write_lock_in_core(struct dm_pool_metadata *pmd) in pmd_write_lock_in_core() argument
391 __acquires(pmd->root_lock) in pmd_write_lock_in_core()
393 down_write(&pmd->root_lock); in pmd_write_lock_in_core()
396 static inline void pmd_write_lock(struct dm_pool_metadata *pmd) in pmd_write_lock() argument
398 pmd_write_lock_in_core(pmd); in pmd_write_lock()
399 if (unlikely(!pmd->in_service)) in pmd_write_lock()
400 pmd->in_service = true; in pmd_write_lock()
403 static inline void pmd_write_unlock(struct dm_pool_metadata *pmd) in pmd_write_unlock() argument
404 __releases(pmd->root_lock) in pmd_write_unlock()
[all …]
Ddm-thin-metadata.h48 int dm_pool_metadata_close(struct dm_pool_metadata *pmd);
61 int dm_pool_create_thin(struct dm_pool_metadata *pmd, dm_thin_id dev);
69 int dm_pool_create_snap(struct dm_pool_metadata *pmd, dm_thin_id dev,
77 int dm_pool_delete_thin_device(struct dm_pool_metadata *pmd,
84 int dm_pool_commit_metadata(struct dm_pool_metadata *pmd);
94 int dm_pool_abort_metadata(struct dm_pool_metadata *pmd);
99 int dm_pool_set_metadata_transaction_id(struct dm_pool_metadata *pmd,
103 int dm_pool_get_metadata_transaction_id(struct dm_pool_metadata *pmd,
115 int dm_pool_reserve_metadata_snap(struct dm_pool_metadata *pmd);
116 int dm_pool_release_metadata_snap(struct dm_pool_metadata *pmd);
[all …]
/kernel/linux/linux-5.10/arch/arc/include/asm/
Dhugepage.h13 static inline pte_t pmd_pte(pmd_t pmd) in pmd_pte() argument
15 return __pte(pmd_val(pmd)); in pmd_pte()
23 #define pmd_wrprotect(pmd) pte_pmd(pte_wrprotect(pmd_pte(pmd))) argument
24 #define pmd_mkwrite(pmd) pte_pmd(pte_mkwrite(pmd_pte(pmd))) argument
25 #define pmd_mkdirty(pmd) pte_pmd(pte_mkdirty(pmd_pte(pmd))) argument
26 #define pmd_mkold(pmd) pte_pmd(pte_mkold(pmd_pte(pmd))) argument
27 #define pmd_mkyoung(pmd) pte_pmd(pte_mkyoung(pmd_pte(pmd))) argument
28 #define pmd_mkhuge(pmd) pte_pmd(pte_mkhuge(pmd_pte(pmd))) argument
29 #define pmd_mkinvalid(pmd) pte_pmd(pte_mknotpresent(pmd_pte(pmd))) argument
30 #define pmd_mkclean(pmd) pte_pmd(pte_mkclean(pmd_pte(pmd))) argument
[all …]
/kernel/linux/linux-4.19/arch/arc/include/asm/
Dhugepage.h17 static inline pte_t pmd_pte(pmd_t pmd) in pmd_pte() argument
19 return __pte(pmd_val(pmd)); in pmd_pte()
27 #define pmd_wrprotect(pmd) pte_pmd(pte_wrprotect(pmd_pte(pmd))) argument
28 #define pmd_mkwrite(pmd) pte_pmd(pte_mkwrite(pmd_pte(pmd))) argument
29 #define pmd_mkdirty(pmd) pte_pmd(pte_mkdirty(pmd_pte(pmd))) argument
30 #define pmd_mkold(pmd) pte_pmd(pte_mkold(pmd_pte(pmd))) argument
31 #define pmd_mkyoung(pmd) pte_pmd(pte_mkyoung(pmd_pte(pmd))) argument
32 #define pmd_mkhuge(pmd) pte_pmd(pte_mkhuge(pmd_pte(pmd))) argument
33 #define pmd_mknotpresent(pmd) pte_pmd(pte_mknotpresent(pmd_pte(pmd))) argument
34 #define pmd_mkclean(pmd) pte_pmd(pte_mkclean(pmd_pte(pmd))) argument
[all …]
/kernel/linux/linux-5.10/arch/arm/include/asm/
Dpgtable-3level.h114 #define pmd_table(pmd) ((pmd_val(pmd) & PMD_TYPE_MASK) == \ argument
116 #define pmd_sect(pmd) ((pmd_val(pmd) & PMD_TYPE_MASK) == \ argument
118 #define pmd_large(pmd) pmd_sect(pmd) argument
119 #define pmd_leaf(pmd) pmd_sect(pmd) argument
138 #define pmd_bad(pmd) (!(pmd_val(pmd) & 2)) argument
174 #define pmd_isset(pmd, val) ((u32)(val) == (val) ? pmd_val(pmd) & (val) \ argument
175 : !!(pmd_val(pmd) & (val)))
176 #define pmd_isclear(pmd, val) (!(pmd_val(pmd) & (val))) argument
178 #define pmd_present(pmd) (pmd_isset((pmd), L_PMD_SECT_VALID)) argument
179 #define pmd_young(pmd) (pmd_isset((pmd), PMD_SECT_AF)) argument
[all …]
/kernel/linux/linux-4.19/arch/parisc/include/asm/
Dpgalloc.h16 * allocate the first pmd adjacent to the pgd. This means that we can
17 * subtract a constant offset to get to it. The pmd and pgd sizes are
18 * arranged so that a single pmd covers 4GB (giving a full 64-bit
32 /* Populate first pmd with allocated memory. We mark it in pgd_alloc()
34 * pmd entry may not be cleared. */ in pgd_alloc()
39 /* The first pmd entry also is marked with PxD_FLAG_ATTACHED as in pgd_alloc()
40 * a signal that this pmd may not be freed */ in pgd_alloc()
57 /* Three Level Page Table Support for pmd's */
59 static inline void pgd_populate(struct mm_struct *mm, pgd_t *pgd, pmd_t *pmd) in pgd_populate() argument
62 (__u32)(__pa((unsigned long)pmd) >> PxD_VALUE_SHIFT)); in pgd_populate()
[all …]
/kernel/linux/linux-4.19/arch/arm/include/asm/
Dpgtable-3level.h144 #define pmd_table(pmd) ((pmd_val(pmd) & PMD_TYPE_MASK) == \ argument
146 #define pmd_sect(pmd) ((pmd_val(pmd) & PMD_TYPE_MASK) == \ argument
148 #define pmd_large(pmd) pmd_sect(pmd) argument
174 #define pmd_bad(pmd) (!(pmd_val(pmd) & 2)) argument
210 #define pmd_isset(pmd, val) ((u32)(val) == (val) ? pmd_val(pmd) & (val) \ argument
211 : !!(pmd_val(pmd) & (val)))
212 #define pmd_isclear(pmd, val) (!(pmd_val(pmd) & (val))) argument
214 #define pmd_present(pmd) (pmd_isset((pmd), L_PMD_SECT_VALID)) argument
215 #define pmd_young(pmd) (pmd_isset((pmd), PMD_SECT_AF)) argument
223 #define pmd_write(pmd) (pmd_isclear((pmd), L_PMD_SECT_RDONLY)) argument
[all …]
/kernel/linux/linux-5.10/arch/parisc/include/asm/
Dpgalloc.h21 * allocate the first pmd adjacent to the pgd. This means that we can
22 * subtract a constant offset to get to it. The pmd and pgd sizes are
23 * arranged so that a single pmd covers 4GB (giving a full 64-bit
37 /* Populate first pmd with allocated memory. We mark it in pgd_alloc()
39 * pmd entry may not be cleared. */ in pgd_alloc()
44 /* The first pmd entry also is marked with PxD_FLAG_ATTACHED as in pgd_alloc()
45 * a signal that this pmd may not be freed */ in pgd_alloc()
63 /* Three Level Page Table Support for pmd's */
65 static inline void pud_populate(struct mm_struct *mm, pud_t *pud, pmd_t *pmd) in pud_populate() argument
68 (__u32)(__pa((unsigned long)pmd) >> PxD_VALUE_SHIFT))); in pud_populate()
[all …]
/kernel/linux/linux-5.10/arch/mips/include/asm/
Dpgtable.h92 #define pmd_phys(pmd) virt_to_phys((void *)pmd_val(pmd)) argument
94 #define __pmd_page(pmd) (pfn_to_page(pmd_phys(pmd) >> PAGE_SHIFT)) argument
96 #define pmd_page(pmd) __pmd_page(pmd) argument
99 #define pmd_page_vaddr(pmd) pmd_val(pmd) argument
585 static inline int pmd_trans_huge(pmd_t pmd) argument
587 return !!(pmd_val(pmd) & _PAGE_HUGE);
590 static inline pmd_t pmd_mkhuge(pmd_t pmd) argument
592 pmd_val(pmd) |= _PAGE_HUGE;
594 return pmd;
598 pmd_t *pmdp, pmd_t pmd);
[all …]
/kernel/linux/linux-5.10/arch/arm64/include/asm/
Dpgtable.h139 #define pmd_access_permitted(pmd, write) \ argument
140 (pte_access_permitted(pmd_pte(pmd), (write)))
156 static inline pmd_t clear_pmd_bit(pmd_t pmd, pgprot_t prot) in clear_pmd_bit() argument
158 pmd_val(pmd) &= ~pgprot_val(prot); in clear_pmd_bit()
159 return pmd; in clear_pmd_bit()
162 static inline pmd_t set_pmd_bit(pmd_t pmd, pgprot_t prot) in set_pmd_bit() argument
164 pmd_val(pmd) |= pgprot_val(prot); in set_pmd_bit()
165 return pmd; in set_pmd_bit()
238 static inline pmd_t pmd_mkcont(pmd_t pmd) in pmd_mkcont() argument
240 return __pmd(pmd_val(pmd) | PMD_SECT_CONT); in pmd_mkcont()
[all …]
/kernel/linux/linux-5.10/mm/
Ddebug_vm_pgtable.c150 pmd_t pmd; in pmd_basic_tests() local
155 pr_debug("Validating PMD basic (%pGv)\n", ptr); in pmd_basic_tests()
156 pmd = pfn_pmd(pfn, prot); in pmd_basic_tests()
165 WARN_ON(pmd_dirty(pmd_wrprotect(pmd))); in pmd_basic_tests()
168 WARN_ON(!pmd_same(pmd, pmd)); in pmd_basic_tests()
169 WARN_ON(!pmd_young(pmd_mkyoung(pmd_mkold(pmd)))); in pmd_basic_tests()
170 WARN_ON(!pmd_dirty(pmd_mkdirty(pmd_mkclean(pmd)))); in pmd_basic_tests()
171 WARN_ON(!pmd_write(pmd_mkwrite(pmd_wrprotect(pmd)))); in pmd_basic_tests()
172 WARN_ON(pmd_young(pmd_mkold(pmd_mkyoung(pmd)))); in pmd_basic_tests()
173 WARN_ON(pmd_dirty(pmd_mkclean(pmd_mkdirty(pmd)))); in pmd_basic_tests()
[all …]
Dhuge_memory.c389 * DAX PMD support. in hugepage_init()
479 pmd_t maybe_pmd_mkwrite(pmd_t pmd, struct vm_area_struct *vma) in maybe_pmd_mkwrite() argument
482 pmd = pmd_mkwrite(pmd); in maybe_pmd_mkwrite()
483 return pmd; in maybe_pmd_mkwrite()
613 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in __do_huge_pmd_anonymous_page()
614 if (unlikely(!pmd_none(*vmf->pmd))) { in __do_huge_pmd_anonymous_page()
639 pgtable_trans_huge_deposit(vma->vm_mm, vmf->pmd, pgtable); in __do_huge_pmd_anonymous_page()
640 set_pmd_at(vma->vm_mm, haddr, vmf->pmd, entry); in __do_huge_pmd_anonymous_page()
696 struct vm_area_struct *vma, unsigned long haddr, pmd_t *pmd, in set_huge_zero_page() argument
700 if (!pmd_none(*pmd)) in set_huge_zero_page()
[all …]
/kernel/linux/linux-4.19/arch/mips/include/asm/
Dpgtable.h92 #define pmd_phys(pmd) virt_to_phys((void *)pmd_val(pmd)) argument
94 #define __pmd_page(pmd) (pfn_to_page(pmd_phys(pmd) >> PAGE_SHIFT)) argument
96 #define pmd_page(pmd) __pmd_page(pmd) argument
99 #define pmd_page_vaddr(pmd) pmd_val(pmd) argument
543 static inline int pmd_trans_huge(pmd_t pmd) argument
545 return !!(pmd_val(pmd) & _PAGE_HUGE);
548 static inline pmd_t pmd_mkhuge(pmd_t pmd) argument
550 pmd_val(pmd) |= _PAGE_HUGE;
552 return pmd;
556 pmd_t *pmdp, pmd_t pmd);
[all …]
Dpgalloc.h16 static inline void pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, in pmd_populate_kernel() argument
19 set_pmd(pmd, __pmd((unsigned long)pte)); in pmd_populate_kernel()
22 static inline void pmd_populate(struct mm_struct *mm, pmd_t *pmd, in pmd_populate() argument
25 set_pmd(pmd, __pmd((unsigned long)page_address(pte))); in pmd_populate()
27 #define pmd_pgtable(pmd) pmd_page(pmd) argument
30 * Initialize a new pmd table with invalid pointers.
36 static inline void pud_populate(struct mm_struct *mm, pud_t *pud, pmd_t *pmd) in pud_populate() argument
38 set_pud(pud, __pud((unsigned long)pmd)); in pud_populate()
43 * Initialize a new pgd / pmd table with invalid pointers.
96 pmd_t *pmd; in pmd_alloc_one() local
[all …]
/kernel/linux/linux-4.19/mm/
Dhuge_memory.c480 pmd_t maybe_pmd_mkwrite(pmd_t pmd, struct vm_area_struct *vma) in maybe_pmd_mkwrite() argument
483 pmd = pmd_mkwrite(pmd); in maybe_pmd_mkwrite()
484 return pmd; in maybe_pmd_mkwrite()
588 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in __do_huge_pmd_anonymous_page()
589 if (unlikely(!pmd_none(*vmf->pmd))) { in __do_huge_pmd_anonymous_page()
616 pgtable_trans_huge_deposit(vma->vm_mm, vmf->pmd, pgtable); in __do_huge_pmd_anonymous_page()
617 set_pmd_at(vma->vm_mm, haddr, vmf->pmd, entry); in __do_huge_pmd_anonymous_page()
664 struct vm_area_struct *vma, unsigned long haddr, pmd_t *pmd, in set_huge_zero_page() argument
668 if (!pmd_none(*pmd)) in set_huge_zero_page()
673 pgtable_trans_huge_deposit(mm, pmd, pgtable); in set_huge_zero_page()
[all …]
/kernel/linux/linux-5.10/arch/x86/include/asm/
Dpgtable.h31 bool __init __early_make_pgtable(unsigned long address, pmdval_t pmd);
70 #define set_pmd(pmdp, pmd) native_set_pmd(pmdp, pmd) argument
94 #define pmd_clear(pmd) native_pmd_clear(pmd) argument
163 static inline int pmd_dirty(pmd_t pmd) in pmd_dirty() argument
165 return pmd_flags(pmd) & _PAGE_DIRTY; in pmd_dirty()
168 static inline int pmd_young(pmd_t pmd) in pmd_young() argument
170 return pmd_flags(pmd) & _PAGE_ACCESSED; in pmd_young()
219 static inline unsigned long pmd_pfn(pmd_t pmd) in pmd_pfn() argument
221 phys_addr_t pfn = pmd_val(pmd); in pmd_pfn()
223 return (pfn & pmd_pfn_mask(pmd)) >> PAGE_SHIFT; in pmd_pfn()
[all …]
/kernel/linux/linux-4.19/arch/arm64/include/asm/
Dpgtable.h130 #define pmd_access_permitted(pmd, write) \ argument
131 (pte_access_permitted(pmd_pte(pmd), (write)))
210 static inline pmd_t pmd_mkcont(pmd_t pmd) in pmd_mkcont() argument
212 return __pmd(pmd_val(pmd) | PMD_SECT_CONT); in pmd_mkcont()
303 static inline pte_t pmd_pte(pmd_t pmd) in pmd_pte() argument
305 return __pte(pmd_val(pmd)); in pmd_pte()
327 static inline int pmd_protnone(pmd_t pmd) in pmd_protnone() argument
329 return pte_protnone(pmd_pte(pmd)); in pmd_protnone()
338 #define pmd_trans_huge(pmd) (pmd_val(pmd) && !(pmd_val(pmd) & PMD_TABLE_BIT)) argument
341 #define pmd_present(pmd) pte_present(pmd_pte(pmd)) argument
[all …]
/kernel/linux/linux-5.10/arch/x86/mm/
Dpgtable.c61 void ___pmd_free_tlb(struct mmu_gather *tlb, pmd_t *pmd) in ___pmd_free_tlb() argument
63 struct page *page = virt_to_page(pmd); in ___pmd_free_tlb()
64 paravirt_release_pmd(__pa(pmd) >> PAGE_SHIFT); in ___pmd_free_tlb()
126 ptes in non-PAE, or shared PMD in PAE), then just copy the in pgd_ctor()
156 * kernel pmd is shared. If PAE were not to share the pmd a similar
172 * Also, if we're in a paravirt environment where the kernel pmd is
188 void pud_populate(struct mm_struct *mm, pud_t *pudp, pmd_t *pmd) in pud_populate() argument
190 paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); in pud_populate()
193 reserved at the pmd (PDPT) level. */ in pud_populate()
194 set_pud(pudp, __pud(__pa(pmd) | _PAGE_PRESENT)); in pud_populate()
[all …]
/kernel/linux/linux-4.19/arch/x86/mm/
Dpgtable.c70 void ___pmd_free_tlb(struct mmu_gather *tlb, pmd_t *pmd) in ___pmd_free_tlb() argument
72 struct page *page = virt_to_page(pmd); in ___pmd_free_tlb()
73 paravirt_release_pmd(__pa(pmd) >> PAGE_SHIFT); in ___pmd_free_tlb()
135 ptes in non-PAE, or shared PMD in PAE), then just copy the in pgd_ctor()
165 * kernel pmd is shared. If PAE were not to share the pmd a similar
181 * Also, if we're in a paravirt environment where the kernel pmd is
197 void pud_populate(struct mm_struct *mm, pud_t *pudp, pmd_t *pmd) in pud_populate() argument
199 paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); in pud_populate()
202 reserved at the pmd (PDPT) level. */ in pud_populate()
203 set_pud(pudp, __pud(__pa(pmd) | _PAGE_PRESENT)); in pud_populate()
[all …]
/kernel/linux/linux-5.10/mm/kasan/
Dinit.c74 static inline bool kasan_pte_table(pmd_t pmd) in kasan_pte_table() argument
76 return pmd_page(pmd) == virt_to_page(lm_alias(kasan_early_shadow_pte)); in kasan_pte_table()
96 static void __ref zero_pte_populate(pmd_t *pmd, unsigned long addr, in zero_pte_populate() argument
99 pte_t *pte = pte_offset_kernel(pmd, addr); in zero_pte_populate()
109 pte = pte_offset_kernel(pmd, addr); in zero_pte_populate()
116 pmd_t *pmd = pmd_offset(pud, addr); in zero_pmd_populate() local
123 pmd_populate_kernel(&init_mm, pmd, in zero_pmd_populate()
128 if (pmd_none(*pmd)) { in zero_pmd_populate()
138 pmd_populate_kernel(&init_mm, pmd, p); in zero_pmd_populate()
140 zero_pte_populate(pmd, addr, next); in zero_pmd_populate()
[all …]
/kernel/linux/linux-4.19/mm/kasan/
Dkasan_init.c74 static inline bool kasan_pte_table(pmd_t pmd) in kasan_pte_table() argument
76 return pmd_page(pmd) == virt_to_page(lm_alias(kasan_zero_pte)); in kasan_pte_table()
90 static void __ref zero_pte_populate(pmd_t *pmd, unsigned long addr, in zero_pte_populate() argument
93 pte_t *pte = pte_offset_kernel(pmd, addr); in zero_pte_populate()
102 pte = pte_offset_kernel(pmd, addr); in zero_pte_populate()
109 pmd_t *pmd = pmd_offset(pud, addr); in zero_pmd_populate() local
116 pmd_populate_kernel(&init_mm, pmd, lm_alias(kasan_zero_pte)); in zero_pmd_populate()
120 if (pmd_none(*pmd)) { in zero_pmd_populate()
130 pmd_populate_kernel(&init_mm, pmd, p); in zero_pmd_populate()
132 zero_pte_populate(pmd, addr, next); in zero_pmd_populate()
[all …]
/kernel/linux/linux-5.10/arch/powerpc/include/asm/book3s/64/
Dpgtable.h47 * We need to mark a pmd pte invalid while splitting. We can do that by clearing
194 /* pmd table use page table fragments */
249 /* Bits to mask out from a PMD to get to the PTE page */
251 /* Bits to mask out from a PUD to get to the PMD page */
707 * This is potentially called with a pmd as the argument, in which case it's not
752 #define __pmd_to_swp_entry(pmd) (__pte_to_swp_entry(pmd_pte(pmd))) argument
891 static inline int pmd_none(pmd_t pmd) in pmd_none() argument
893 return !pmd_raw(pmd); in pmd_none()
896 static inline int pmd_present(pmd_t pmd) in pmd_present() argument
899 * A pmd is considerent present if _PAGE_PRESENT is set. in pmd_present()
[all …]
/kernel/linux/linux-4.19/arch/s390/include/asm/
Dpgtable.h71 printk("%s:%d: bad pmd %p.\n", __FILE__, __LINE__, (void *) pmd_val(e))
586 * pgd/p4d/pud/pmd/pte query functions
683 static inline int pmd_large(pmd_t pmd) in pmd_large() argument
685 return (pmd_val(pmd) & _SEGMENT_ENTRY_LARGE) != 0; in pmd_large()
688 static inline int pmd_bad(pmd_t pmd) in pmd_bad() argument
690 if (pmd_large(pmd)) in pmd_bad()
691 return (pmd_val(pmd) & ~_SEGMENT_ENTRY_BITS_LARGE) != 0; in pmd_bad()
692 return (pmd_val(pmd) & ~_SEGMENT_ENTRY_BITS) != 0; in pmd_bad()
711 static inline int pmd_present(pmd_t pmd) in pmd_present() argument
713 return pmd_val(pmd) != _SEGMENT_ENTRY_EMPTY; in pmd_present()
[all …]

12345678910>>...34