Home
last modified time | relevance | path

Searched refs:order (Results 1 – 25 of 55) sorted by relevance

123

/include/trace/events/
Dcompaction.h168 int order,
172 TP_ARGS(order, gfp_mask, prio),
175 __field(int, order)
181 __entry->order = order;
187 __entry->order,
195 int order,
198 TP_ARGS(zone, order, ret),
203 __field(int, order)
210 __entry->order = order;
217 __entry->order,
[all …]
Dvmscan.h68 TP_PROTO(int nid, int zid, int order),
70 TP_ARGS(nid, zid, order),
75 __field( int, order )
81 __entry->order = order;
86 __entry->order)
91 TP_PROTO(int nid, int zid, int order, gfp_t gfp_flags),
93 TP_ARGS(nid, zid, order, gfp_flags),
98 __field( int, order )
105 __entry->order = order;
111 __entry->order,
[all …]
Dmigrate.h115 TP_PROTO(unsigned long addr, unsigned long pte, int order),
117 TP_ARGS(addr, pte, order),
122 __field(int, order)
128 __entry->order = order;
131 TP_printk("addr=%lx, pte=%lx order=%d", __entry->addr, __entry->pte, __entry->order)
135 TP_PROTO(unsigned long addr, unsigned long pte, int order),
136 TP_ARGS(addr, pte, order)
140 TP_PROTO(unsigned long addr, unsigned long pte, int order),
141 TP_ARGS(addr, pte, order)
Dkmem.h138 TP_PROTO(struct page *page, unsigned int order),
140 TP_ARGS(page, order),
144 __field( unsigned int, order )
149 __entry->order = order;
155 __entry->order)
179 TP_PROTO(struct page *page, unsigned int order,
182 TP_ARGS(page, order, gfp_flags, migratetype),
186 __field( unsigned int, order )
193 __entry->order = order;
201 __entry->order,
[all …]
Doom.h37 int order,
44 TP_ARGS(zoneref, order, reclaimable, available, min_wmark, no_progress_loops, wmark_check),
49 __field( int, order)
60 __entry->order = order;
70 __entry->order,
185 TP_PROTO(int order,
192 TP_ARGS(order, priority, result, retries, max_retries, ret),
195 __field( int, order)
204 __entry->order = order;
213 __entry->order,
Dfilemap.h27 __field(unsigned char, order)
38 __entry->order = folio_order(folio);
46 __entry->order)
/include/linux/
Dcompaction.h65 static inline unsigned long compact_gap(unsigned int order) in compact_gap() argument
80 return 2UL << order; in compact_gap()
90 extern unsigned int extfrag_for_order(struct zone *zone, unsigned int order);
91 extern int fragmentation_index(struct zone *zone, unsigned int order);
93 unsigned int order, unsigned int alloc_flags,
97 extern bool compaction_suitable(struct zone *zone, int order,
100 extern void compaction_defer_reset(struct zone *zone, int order,
103 bool compaction_zonelist_suitable(struct alloc_context *ac, int order,
108 extern void wakeup_kcompactd(pg_data_t *pgdat, int order, int highest_zoneidx);
116 static inline bool compaction_suitable(struct zone *zone, int order, in compaction_suitable() argument
[all …]
Dgfp.h199 static inline void arch_free_page(struct page *page, int order) { } in arch_free_page() argument
202 static inline void arch_alloc_page(struct page *page, int order) { } in arch_alloc_page() argument
205 struct page *__alloc_pages_noprof(gfp_t gfp, unsigned int order, int preferred_nid,
209 struct folio *__folio_alloc_noprof(gfp_t gfp, unsigned int order, int preferred_nid,
264 __alloc_pages_node_noprof(int nid, gfp_t gfp_mask, unsigned int order) in __alloc_pages_node_noprof() argument
269 return __alloc_pages_noprof(gfp_mask, order, nid, NULL); in __alloc_pages_node_noprof()
275 struct folio *__folio_alloc_node_noprof(gfp_t gfp, unsigned int order, int nid) in __folio_alloc_node_noprof() argument
280 return __folio_alloc_noprof(gfp, order, nid, NULL); in __folio_alloc_node_noprof()
291 unsigned int order) in alloc_pages_node_noprof() argument
296 return __alloc_pages_node_noprof(nid, gfp_mask, order); in alloc_pages_node_noprof()
[all …]
Dpage_owner.h14 extern void __reset_page_owner(struct page *page, unsigned short order);
16 unsigned short order, gfp_t gfp_mask);
25 static inline void reset_page_owner(struct page *page, unsigned short order) in reset_page_owner() argument
28 __reset_page_owner(page, order); in reset_page_owner()
32 unsigned short order, gfp_t gfp_mask) in set_page_owner() argument
35 __set_page_owner(page, order, gfp_mask); in set_page_owner()
60 static inline void reset_page_owner(struct page *page, unsigned short order) in reset_page_owner() argument
64 unsigned short order, gfp_t gfp_mask) in set_page_owner() argument
Dpage_pinner.h12 extern void __free_page_pinner(struct page *page, unsigned int order);
16 static inline void free_page_pinner(struct page *page, unsigned int order) in free_page_pinner() argument
19 __free_page_pinner(page, order); in free_page_pinner()
38 static inline void free_page_pinner(struct page *page, unsigned int order) in free_page_pinner() argument
Dpage_table_check.h16 void __page_table_check_zero(struct page *page, unsigned int order);
28 static inline void page_table_check_alloc(struct page *page, unsigned int order) in page_table_check_alloc() argument
33 __page_table_check_zero(page, order); in page_table_check_alloc()
36 static inline void page_table_check_free(struct page *page, unsigned int order) in page_table_check_free() argument
41 __page_table_check_zero(page, order); in page_table_check_free()
107 static inline void page_table_check_alloc(struct page *page, unsigned int order) in page_table_check_alloc() argument
111 static inline void page_table_check_free(struct page *page, unsigned int order) in page_table_check_free() argument
Dkasan.h112 void __kasan_poison_pages(struct page *page, unsigned int order, bool init);
114 unsigned int order, bool init) in kasan_poison_pages() argument
117 __kasan_poison_pages(page, order, init); in kasan_poison_pages()
120 bool __kasan_unpoison_pages(struct page *page, unsigned int order, bool init);
122 unsigned int order, bool init) in kasan_unpoison_pages() argument
125 return __kasan_unpoison_pages(page, order, init); in kasan_unpoison_pages()
284 bool __kasan_mempool_poison_pages(struct page *page, unsigned int order,
303 unsigned int order) in kasan_mempool_poison_pages() argument
306 return __kasan_mempool_poison_pages(page, order, _RET_IP_); in kasan_mempool_poison_pages()
310 void __kasan_mempool_unpoison_pages(struct page *page, unsigned int order,
[all …]
Dhuge_mm.h96 #define thp_vma_allowable_order(vma, vm_flags, tva_flags, order) \ argument
97 (!!thp_vma_allowable_orders(vma, vm_flags, tva_flags, BIT(order)))
143 static inline void mod_mthp_stat(int order, enum mthp_stat_item item, int delta) in mod_mthp_stat() argument
145 if (order <= 0 || order > PMD_ORDER) in mod_mthp_stat()
148 this_cpu_add(mthp_stats.stats[order][item], delta); in mod_mthp_stat()
151 static inline void count_mthp_stat(int order, enum mthp_stat_item item) in count_mthp_stat() argument
153 mod_mthp_stat(order, item, 1); in count_mthp_stat()
157 static inline void mod_mthp_stat(int order, enum mthp_stat_item item, int delta) in mod_mthp_stat() argument
161 static inline void count_mthp_stat(int order, enum mthp_stat_item item) in count_mthp_stat() argument
209 unsigned long addr, int order) in thp_vma_suitable_order() argument
[all …]
Dkmsan.h58 unsigned int order);
69 void kmsan_alloc_page(struct page *page, unsigned int order, gfp_t flags);
78 void kmsan_free_page(struct page *page, unsigned int order);
305 unsigned int order) in kmsan_memblock_free_pages() argument
318 static inline void kmsan_alloc_page(struct page *page, unsigned int order, in kmsan_alloc_page() argument
323 static inline void kmsan_free_page(struct page *page, unsigned int order) in kmsan_free_page() argument
Dpage_reporting.h23 unsigned int order; member
Dxarray.h1404 #define XA_STATE_ORDER(name, array, index, order) \ argument
1406 (index >> order) << order, \
1407 order - (order % XA_CHUNK_SHIFT), \
1408 (1U << (order % XA_CHUNK_SHIFT)) - 1)
1556 void xas_split(struct xa_state *, void *entry, unsigned int order);
1557 void xas_split_alloc(struct xa_state *, void *entry, unsigned int order, gfp_t);
1570 unsigned int order) in xas_split() argument
1576 unsigned int order, gfp_t gfp) in xas_split_alloc() argument
1655 unsigned int order) in xas_set_order() argument
1658 xas->xa_index = order < BITS_PER_LONG ? (index >> order) << order : 0; in xas_set_order()
[all …]
Dfault-inject.h117 bool should_fail_alloc_page(gfp_t gfp_mask, unsigned int order);
119 static inline bool should_fail_alloc_page(gfp_t gfp_mask, unsigned int order) in should_fail_alloc_page() argument
Dcma.h60 struct folio *cma_alloc_folio(struct cma *cma, int order, gfp_t gfp);
63 static inline struct folio *cma_alloc_folio(struct cma *cma, int order, gfp_t gfp) in cma_alloc_folio() argument
Ddma-map-ops.h120 unsigned int order, bool no_warn);
142 size_t count, unsigned int order, bool no_warn) in dma_alloc_from_contiguous() argument
173 int dma_release_from_dev_coherent(struct device *dev, int order, void *vaddr);
184 #define dma_release_from_dev_coherent(dev, order, vaddr) (0) argument
185 #define dma_mmap_from_dev_coherent(dev, vma, vaddr, order, ret) (0) argument
192 int dma_release_from_global_coherent(int order, void *vaddr);
202 static inline int dma_release_from_global_coherent(int order, void *vaddr) in dma_release_from_global_coherent() argument
Dring_buffer.h139 int order, unsigned long start,
162 #define ring_buffer_alloc_range(size, flags, order, start, range_size) \ argument
165 __ring_buffer_alloc_range((size), (flags), (order), (start), \
278 int ring_buffer_subbuf_order_set(struct trace_buffer *buffer, int order);
/include/trace/hooks/
Dmm.h21 TP_PROTO(struct shmem_inode_info *info, struct folio **folio, int order),
22 TP_ARGS(info, folio, order), 3);
30 TP_PROTO(struct page **page, unsigned int order,
32 TP_ARGS(page, order, gfp, highest_zoneidx), 1);
116 TP_PROTO(struct page **page, struct zone *zone, int order, int migratetype),
117 TP_ARGS(page, zone, order, migratetype));
119 TP_PROTO(struct page *page, struct zone *zone, int order, int migratetype,
121 TP_ARGS(page, zone, order, migratetype, fpi_flags, bypass));
144 unsigned int order, gfp_t gfp_flags,
146 TP_ARGS(preferred_zone, zone, order,
[all …]
Dvmscan.h48 unsigned int order, bool *bypass),
49 TP_ARGS(nr_to_reclaim, nr_reclaimed, order, bypass));
57 gfp_t gfp_mask, int order, bool *bypass),
58 TP_ARGS(ext, partial, nr_to_reclaim, nr_reclaimed, gfp_mask, order, bypass));
120 TP_PROTO(unsigned int may_swap, int order, int highest_zoneidx, bool *file_is_tiny),
121 TP_ARGS(may_swap, order, highest_zoneidx, file_is_tiny));
123 TP_PROTO(int order, int highest_zoneidx, bool *balanced, bool *customized),
124 TP_ARGS(order, highest_zoneidx, balanced, customized));
126 TP_PROTO(int order, gfp_t gfp, s8 *reclaim_idx, enum zone_type *highest_zoneidx),
127 TP_ARGS(order, gfp, reclaim_idx, highest_zoneidx));
Dcompaction.h16 TP_PROTO(int node_id, int order, const int highest_zoneidx),
17 TP_ARGS(node_id, order, highest_zoneidx));
Diommu.h39 TP_PROTO(unsigned int order, gfp_t *alloc_flags),
40 TP_ARGS(order, alloc_flags));
/include/drm/ttm/
Dttm_pool.h52 unsigned int order; member

123