| /include/trace/events/ |
| D | compaction.h | 168 int order, 172 TP_ARGS(order, gfp_mask, prio), 175 __field(int, order) 181 __entry->order = order; 187 __entry->order, 195 int order, 198 TP_ARGS(zone, order, ret), 203 __field(int, order) 210 __entry->order = order; 217 __entry->order, [all …]
|
| D | vmscan.h | 68 TP_PROTO(int nid, int zid, int order), 70 TP_ARGS(nid, zid, order), 75 __field( int, order ) 81 __entry->order = order; 86 __entry->order) 91 TP_PROTO(int nid, int zid, int order, gfp_t gfp_flags), 93 TP_ARGS(nid, zid, order, gfp_flags), 98 __field( int, order ) 105 __entry->order = order; 111 __entry->order, [all …]
|
| D | migrate.h | 115 TP_PROTO(unsigned long addr, unsigned long pte, int order), 117 TP_ARGS(addr, pte, order), 122 __field(int, order) 128 __entry->order = order; 131 TP_printk("addr=%lx, pte=%lx order=%d", __entry->addr, __entry->pte, __entry->order) 135 TP_PROTO(unsigned long addr, unsigned long pte, int order), 136 TP_ARGS(addr, pte, order) 140 TP_PROTO(unsigned long addr, unsigned long pte, int order), 141 TP_ARGS(addr, pte, order)
|
| D | kmem.h | 138 TP_PROTO(struct page *page, unsigned int order), 140 TP_ARGS(page, order), 144 __field( unsigned int, order ) 149 __entry->order = order; 155 __entry->order) 179 TP_PROTO(struct page *page, unsigned int order, 182 TP_ARGS(page, order, gfp_flags, migratetype), 186 __field( unsigned int, order ) 193 __entry->order = order; 201 __entry->order, [all …]
|
| D | oom.h | 37 int order, 44 TP_ARGS(zoneref, order, reclaimable, available, min_wmark, no_progress_loops, wmark_check), 49 __field( int, order) 60 __entry->order = order; 70 __entry->order, 185 TP_PROTO(int order, 192 TP_ARGS(order, priority, result, retries, max_retries, ret), 195 __field( int, order) 204 __entry->order = order; 213 __entry->order,
|
| D | filemap.h | 27 __field(unsigned char, order) 38 __entry->order = folio_order(folio); 46 __entry->order)
|
| /include/linux/ |
| D | compaction.h | 65 static inline unsigned long compact_gap(unsigned int order) in compact_gap() argument 80 return 2UL << order; in compact_gap() 90 extern unsigned int extfrag_for_order(struct zone *zone, unsigned int order); 91 extern int fragmentation_index(struct zone *zone, unsigned int order); 93 unsigned int order, unsigned int alloc_flags, 97 extern bool compaction_suitable(struct zone *zone, int order, 100 extern void compaction_defer_reset(struct zone *zone, int order, 103 bool compaction_zonelist_suitable(struct alloc_context *ac, int order, 108 extern void wakeup_kcompactd(pg_data_t *pgdat, int order, int highest_zoneidx); 116 static inline bool compaction_suitable(struct zone *zone, int order, in compaction_suitable() argument [all …]
|
| D | gfp.h | 199 static inline void arch_free_page(struct page *page, int order) { } in arch_free_page() argument 202 static inline void arch_alloc_page(struct page *page, int order) { } in arch_alloc_page() argument 205 struct page *__alloc_pages_noprof(gfp_t gfp, unsigned int order, int preferred_nid, 209 struct folio *__folio_alloc_noprof(gfp_t gfp, unsigned int order, int preferred_nid, 264 __alloc_pages_node_noprof(int nid, gfp_t gfp_mask, unsigned int order) in __alloc_pages_node_noprof() argument 269 return __alloc_pages_noprof(gfp_mask, order, nid, NULL); in __alloc_pages_node_noprof() 275 struct folio *__folio_alloc_node_noprof(gfp_t gfp, unsigned int order, int nid) in __folio_alloc_node_noprof() argument 280 return __folio_alloc_noprof(gfp, order, nid, NULL); in __folio_alloc_node_noprof() 291 unsigned int order) in alloc_pages_node_noprof() argument 296 return __alloc_pages_node_noprof(nid, gfp_mask, order); in alloc_pages_node_noprof() [all …]
|
| D | page_owner.h | 14 extern void __reset_page_owner(struct page *page, unsigned short order); 16 unsigned short order, gfp_t gfp_mask); 25 static inline void reset_page_owner(struct page *page, unsigned short order) in reset_page_owner() argument 28 __reset_page_owner(page, order); in reset_page_owner() 32 unsigned short order, gfp_t gfp_mask) in set_page_owner() argument 35 __set_page_owner(page, order, gfp_mask); in set_page_owner() 60 static inline void reset_page_owner(struct page *page, unsigned short order) in reset_page_owner() argument 64 unsigned short order, gfp_t gfp_mask) in set_page_owner() argument
|
| D | page_pinner.h | 12 extern void __free_page_pinner(struct page *page, unsigned int order); 16 static inline void free_page_pinner(struct page *page, unsigned int order) in free_page_pinner() argument 19 __free_page_pinner(page, order); in free_page_pinner() 38 static inline void free_page_pinner(struct page *page, unsigned int order) in free_page_pinner() argument
|
| D | page_table_check.h | 16 void __page_table_check_zero(struct page *page, unsigned int order); 28 static inline void page_table_check_alloc(struct page *page, unsigned int order) in page_table_check_alloc() argument 33 __page_table_check_zero(page, order); in page_table_check_alloc() 36 static inline void page_table_check_free(struct page *page, unsigned int order) in page_table_check_free() argument 41 __page_table_check_zero(page, order); in page_table_check_free() 107 static inline void page_table_check_alloc(struct page *page, unsigned int order) in page_table_check_alloc() argument 111 static inline void page_table_check_free(struct page *page, unsigned int order) in page_table_check_free() argument
|
| D | kasan.h | 112 void __kasan_poison_pages(struct page *page, unsigned int order, bool init); 114 unsigned int order, bool init) in kasan_poison_pages() argument 117 __kasan_poison_pages(page, order, init); in kasan_poison_pages() 120 bool __kasan_unpoison_pages(struct page *page, unsigned int order, bool init); 122 unsigned int order, bool init) in kasan_unpoison_pages() argument 125 return __kasan_unpoison_pages(page, order, init); in kasan_unpoison_pages() 284 bool __kasan_mempool_poison_pages(struct page *page, unsigned int order, 303 unsigned int order) in kasan_mempool_poison_pages() argument 306 return __kasan_mempool_poison_pages(page, order, _RET_IP_); in kasan_mempool_poison_pages() 310 void __kasan_mempool_unpoison_pages(struct page *page, unsigned int order, [all …]
|
| D | huge_mm.h | 96 #define thp_vma_allowable_order(vma, vm_flags, tva_flags, order) \ argument 97 (!!thp_vma_allowable_orders(vma, vm_flags, tva_flags, BIT(order))) 143 static inline void mod_mthp_stat(int order, enum mthp_stat_item item, int delta) in mod_mthp_stat() argument 145 if (order <= 0 || order > PMD_ORDER) in mod_mthp_stat() 148 this_cpu_add(mthp_stats.stats[order][item], delta); in mod_mthp_stat() 151 static inline void count_mthp_stat(int order, enum mthp_stat_item item) in count_mthp_stat() argument 153 mod_mthp_stat(order, item, 1); in count_mthp_stat() 157 static inline void mod_mthp_stat(int order, enum mthp_stat_item item, int delta) in mod_mthp_stat() argument 161 static inline void count_mthp_stat(int order, enum mthp_stat_item item) in count_mthp_stat() argument 209 unsigned long addr, int order) in thp_vma_suitable_order() argument [all …]
|
| D | kmsan.h | 58 unsigned int order); 69 void kmsan_alloc_page(struct page *page, unsigned int order, gfp_t flags); 78 void kmsan_free_page(struct page *page, unsigned int order); 305 unsigned int order) in kmsan_memblock_free_pages() argument 318 static inline void kmsan_alloc_page(struct page *page, unsigned int order, in kmsan_alloc_page() argument 323 static inline void kmsan_free_page(struct page *page, unsigned int order) in kmsan_free_page() argument
|
| D | page_reporting.h | 23 unsigned int order; member
|
| D | xarray.h | 1404 #define XA_STATE_ORDER(name, array, index, order) \ argument 1406 (index >> order) << order, \ 1407 order - (order % XA_CHUNK_SHIFT), \ 1408 (1U << (order % XA_CHUNK_SHIFT)) - 1) 1556 void xas_split(struct xa_state *, void *entry, unsigned int order); 1557 void xas_split_alloc(struct xa_state *, void *entry, unsigned int order, gfp_t); 1570 unsigned int order) in xas_split() argument 1576 unsigned int order, gfp_t gfp) in xas_split_alloc() argument 1655 unsigned int order) in xas_set_order() argument 1658 xas->xa_index = order < BITS_PER_LONG ? (index >> order) << order : 0; in xas_set_order() [all …]
|
| D | fault-inject.h | 117 bool should_fail_alloc_page(gfp_t gfp_mask, unsigned int order); 119 static inline bool should_fail_alloc_page(gfp_t gfp_mask, unsigned int order) in should_fail_alloc_page() argument
|
| D | cma.h | 60 struct folio *cma_alloc_folio(struct cma *cma, int order, gfp_t gfp); 63 static inline struct folio *cma_alloc_folio(struct cma *cma, int order, gfp_t gfp) in cma_alloc_folio() argument
|
| D | dma-map-ops.h | 120 unsigned int order, bool no_warn); 142 size_t count, unsigned int order, bool no_warn) in dma_alloc_from_contiguous() argument 173 int dma_release_from_dev_coherent(struct device *dev, int order, void *vaddr); 184 #define dma_release_from_dev_coherent(dev, order, vaddr) (0) argument 185 #define dma_mmap_from_dev_coherent(dev, vma, vaddr, order, ret) (0) argument 192 int dma_release_from_global_coherent(int order, void *vaddr); 202 static inline int dma_release_from_global_coherent(int order, void *vaddr) in dma_release_from_global_coherent() argument
|
| D | ring_buffer.h | 139 int order, unsigned long start, 162 #define ring_buffer_alloc_range(size, flags, order, start, range_size) \ argument 165 __ring_buffer_alloc_range((size), (flags), (order), (start), \ 278 int ring_buffer_subbuf_order_set(struct trace_buffer *buffer, int order);
|
| /include/trace/hooks/ |
| D | mm.h | 21 TP_PROTO(struct shmem_inode_info *info, struct folio **folio, int order), 22 TP_ARGS(info, folio, order), 3); 30 TP_PROTO(struct page **page, unsigned int order, 32 TP_ARGS(page, order, gfp, highest_zoneidx), 1); 116 TP_PROTO(struct page **page, struct zone *zone, int order, int migratetype), 117 TP_ARGS(page, zone, order, migratetype)); 119 TP_PROTO(struct page *page, struct zone *zone, int order, int migratetype, 121 TP_ARGS(page, zone, order, migratetype, fpi_flags, bypass)); 144 unsigned int order, gfp_t gfp_flags, 146 TP_ARGS(preferred_zone, zone, order, [all …]
|
| D | vmscan.h | 48 unsigned int order, bool *bypass), 49 TP_ARGS(nr_to_reclaim, nr_reclaimed, order, bypass)); 57 gfp_t gfp_mask, int order, bool *bypass), 58 TP_ARGS(ext, partial, nr_to_reclaim, nr_reclaimed, gfp_mask, order, bypass)); 120 TP_PROTO(unsigned int may_swap, int order, int highest_zoneidx, bool *file_is_tiny), 121 TP_ARGS(may_swap, order, highest_zoneidx, file_is_tiny)); 123 TP_PROTO(int order, int highest_zoneidx, bool *balanced, bool *customized), 124 TP_ARGS(order, highest_zoneidx, balanced, customized)); 126 TP_PROTO(int order, gfp_t gfp, s8 *reclaim_idx, enum zone_type *highest_zoneidx), 127 TP_ARGS(order, gfp, reclaim_idx, highest_zoneidx));
|
| D | compaction.h | 16 TP_PROTO(int node_id, int order, const int highest_zoneidx), 17 TP_ARGS(node_id, order, highest_zoneidx));
|
| D | iommu.h | 39 TP_PROTO(unsigned int order, gfp_t *alloc_flags), 40 TP_ARGS(order, alloc_flags));
|
| /include/drm/ttm/ |
| D | ttm_pool.h | 52 unsigned int order; member
|