/include/trace/events/ |
D | compaction.h | 173 int order, 177 TP_ARGS(order, gfp_mask, prio), 180 __field(int, order) 186 __entry->order = order; 192 __entry->order, 200 int order, 203 TP_ARGS(zone, order, ret), 208 __field(int, order) 215 __entry->order = order; 222 __entry->order, [all …]
|
D | kmem.h | 154 TP_PROTO(struct page *page, unsigned int order), 156 TP_ARGS(page, order), 160 __field( unsigned int, order ) 165 __entry->order = order; 171 __entry->order) 195 TP_PROTO(struct page *page, unsigned int order, 198 TP_ARGS(page, order, gfp_flags, migratetype), 202 __field( unsigned int, order ) 209 __entry->order = order; 217 __entry->order, [all …]
|
D | vmscan.h | 54 TP_PROTO(int nid, int zid, int order), 56 TP_ARGS(nid, zid, order), 61 __field( int, order ) 67 __entry->order = order; 72 __entry->order) 77 TP_PROTO(int nid, int zid, int order, gfp_t gfp_flags), 79 TP_ARGS(nid, zid, order, gfp_flags), 84 __field( int, order ) 91 __entry->order = order; 97 __entry->order, [all …]
|
D | oom.h | 35 int order, 42 TP_ARGS(zoneref, order, reclaimable, available, min_wmark, no_progress_loops, wmark_check), 47 __field( int, order) 58 __entry->order = order; 68 __entry->order, 157 TP_PROTO(int order, 164 TP_ARGS(order, priority, result, retries, max_retries, ret), 167 __field( int, order) 176 __entry->order = order; 185 __entry->order,
|
/include/linux/ |
D | compaction.h | 68 static inline unsigned long compact_gap(unsigned int order) in compact_gap() argument 83 return 2UL << order; in compact_gap() 93 extern int fragmentation_index(struct zone *zone, unsigned int order); 95 unsigned int order, unsigned int alloc_flags, 99 extern enum compact_result compaction_suitable(struct zone *zone, int order, 102 extern void defer_compaction(struct zone *zone, int order); 103 extern bool compaction_deferred(struct zone *zone, int order); 104 extern void compaction_defer_reset(struct zone *zone, int order, 106 extern bool compaction_restarting(struct zone *zone, int order); 180 bool compaction_zonelist_suitable(struct alloc_context *ac, int order, [all …]
|
D | page_owner.h | 11 extern void __reset_page_owner(struct page *page, unsigned int order); 13 unsigned int order, gfp_t gfp_mask); 21 static inline void reset_page_owner(struct page *page, unsigned int order) in reset_page_owner() argument 24 __reset_page_owner(page, order); in reset_page_owner() 28 unsigned int order, gfp_t gfp_mask) in set_page_owner() argument 31 __set_page_owner(page, order, gfp_mask); in set_page_owner() 55 static inline void reset_page_owner(struct page *page, unsigned int order) in reset_page_owner() argument 59 unsigned int order, gfp_t gfp_mask) in set_page_owner() argument 63 unsigned int order) in split_page_owner() argument
|
D | gfp.h | 490 static inline void arch_free_page(struct page *page, int order) { } in arch_free_page() argument 493 static inline void arch_alloc_page(struct page *page, int order) { } in arch_alloc_page() argument 497 __alloc_pages_nodemask(gfp_t gfp_mask, unsigned int order, int preferred_nid, 501 __alloc_pages(gfp_t gfp_mask, unsigned int order, int preferred_nid) in __alloc_pages() argument 503 return __alloc_pages_nodemask(gfp_mask, order, preferred_nid, NULL); in __alloc_pages() 511 __alloc_pages_node(int nid, gfp_t gfp_mask, unsigned int order) in __alloc_pages_node() argument 516 return __alloc_pages(gfp_mask, order, nid); in __alloc_pages_node() 525 unsigned int order) in alloc_pages_node() argument 530 return __alloc_pages_node(nid, gfp_mask, order); in alloc_pages_node() 534 extern struct page *alloc_pages_current(gfp_t gfp_mask, unsigned order); [all …]
|
D | bitops.h | 53 int order; in get_bitmask_order() local 55 order = fls(count); in get_bitmask_order() 56 return order; /* We could be slightly more clever with -1 here... */ in get_bitmask_order() 177 int order; in get_count_order() local 179 order = fls(count) - 1; in get_count_order() 181 order++; in get_count_order() 182 return order; in get_count_order()
|
D | mempool.h | 99 static inline int mempool_init_page_pool(mempool_t *pool, int min_nr, int order) in mempool_init_page_pool() argument 102 mempool_free_pages, (void *)(long)order); in mempool_init_page_pool() 105 static inline mempool_t *mempool_create_page_pool(int min_nr, int order) in mempool_create_page_pool() argument 108 (void *)(long)order); in mempool_create_page_pool()
|
D | io-pgtable.h | 58 void *(*alloc_pgtable)(void *cookie, int order, gfp_t gfp_mask); 59 void (*free_pgtable)(void *cookie, void *virt, int order); 200 int order, gfp_t gfp_mask); 214 int order);
|
D | kasan.h | 43 void kasan_alloc_pages(struct page *page, unsigned int order); 44 void kasan_free_pages(struct page *page, unsigned int order); 99 static inline void kasan_alloc_pages(struct page *page, unsigned int order) {} in kasan_alloc_pages() argument 100 static inline void kasan_free_pages(struct page *page, unsigned int order) {} in kasan_free_pages() argument
|
D | xarray.h | 1325 #define XA_STATE_ORDER(name, array, index, order) \ argument 1327 (index >> order) << order, \ 1328 order - (order % XA_CHUNK_SHIFT), \ 1329 (1U << (order % XA_CHUNK_SHIFT)) - 1) 1475 void xas_split(struct xa_state *, void *entry, unsigned int order); 1476 void xas_split_alloc(struct xa_state *, void *entry, unsigned int order, gfp_t); 1484 unsigned int order) in xas_split() argument 1490 unsigned int order, gfp_t gfp) in xas_split_alloc() argument 1540 unsigned int order) in xas_set_order() argument 1543 xas->xa_index = order < BITS_PER_LONG ? (index >> order) << order : 0; in xas_set_order() [all …]
|
D | memcontrol.h | 783 unsigned long mem_cgroup_soft_limit_reclaim(pg_data_t *pgdat, int order, 1178 unsigned long mem_cgroup_soft_limit_reclaim(pg_data_t *pgdat, int order, in mem_cgroup_soft_limit_reclaim() argument 1416 int __memcg_kmem_charge(struct page *page, gfp_t gfp, int order); 1417 void __memcg_kmem_uncharge(struct page *page, int order); 1418 int __memcg_kmem_charge_memcg(struct page *page, gfp_t gfp, int order, 1443 static inline int memcg_kmem_charge(struct page *page, gfp_t gfp, int order) in memcg_kmem_charge() argument 1446 return __memcg_kmem_charge(page, gfp, order); in memcg_kmem_charge() 1450 static inline void memcg_kmem_uncharge(struct page *page, int order) in memcg_kmem_uncharge() argument 1453 __memcg_kmem_uncharge(page, order); in memcg_kmem_uncharge() 1457 int order, struct mem_cgroup *memcg) in memcg_kmem_charge_memcg() argument [all …]
|
D | migrate.h | 38 unsigned int order = 0; in new_page_nodemask() local 47 order = HPAGE_PMD_ORDER; in new_page_nodemask() 53 new_page = __alloc_pages_nodemask(gfp_mask, order, in new_page_nodemask()
|
D | dma-contiguous.h | 112 unsigned int order, bool no_warn); 147 unsigned int order, bool no_warn) in dma_alloc_from_contiguous() argument
|
D | slab.h | 470 extern void *kmalloc_order(size_t size, gfp_t flags, unsigned int order) __assume_page_alignment __… 473 extern void *kmalloc_order_trace(size_t size, gfp_t flags, unsigned int order) __assume_page_alignm… 476 kmalloc_order_trace(size_t size, gfp_t flags, unsigned int order) in kmalloc_order_trace() argument 478 return kmalloc_order(size, flags, order); in kmalloc_order_trace() 484 unsigned int order = get_order(size); in kmalloc_large() local 485 return kmalloc_order_trace(size, flags, order); in kmalloc_large()
|
D | hugetlb.h | 336 unsigned int order; member 378 void __init hugetlb_add_hstate(unsigned order); 413 return (unsigned long)PAGE_SIZE << h->order; in huge_page_size() 427 return h->order; in huge_page_order() 432 return h->order + PAGE_SHIFT; in huge_page_shift() 442 return 1 << h->order; in pages_per_huge_page() 468 return hstates[index].order + PAGE_SHIFT; in hstate_index_to_shift()
|
D | oom.h | 46 const int order; member
|
D | mmzone.h | 88 #define for_each_migratetype_order(order, type) \ argument 89 for (order = 0; order < MAX_ORDER; order++) \ 830 void wakeup_kswapd(struct zone *zone, gfp_t gfp_mask, int order, 832 bool __zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark, 835 bool zone_watermark_ok(struct zone *z, unsigned int order, 838 bool zone_watermark_ok_safe(struct zone *z, unsigned int order,
|
D | scatterlist.h | 299 unsigned int order, bool chainable, 303 void sgl_free_n_order(struct scatterlist *sgl, int nents, int order); 304 void sgl_free_order(struct scatterlist *sgl, int order);
|
/include/drm/ |
D | drm_hashtab.h | 49 u8 order; member 52 int drm_ht_create(struct drm_open_hash *ht, unsigned int order);
|
/include/xen/ |
D | xen-ops.h | 46 int xen_create_contiguous_region(phys_addr_t pstart, unsigned int order, 50 void xen_destroy_contiguous_region(phys_addr_t pstart, unsigned int order); 53 unsigned int order, in xen_create_contiguous_region() argument 61 unsigned int order) { } in xen_destroy_contiguous_region() argument
|
/include/media/drv-intf/ |
D | soc_mediabus.h | 78 enum soc_mbus_order order; member
|
/include/sound/ |
D | soc-component.h | 26 #define for_each_comp_order(order) \ argument 27 for (order = SND_SOC_COMP_ORDER_FIRST; \ 28 order <= SND_SOC_COMP_ORDER_LAST; \ 29 order++)
|
/include/xen/arm/ |
D | page.h | 111 unsigned long xen_get_swiotlb_free_pages(unsigned int order);
|