Searched refs:gfp_mask (Results 1 – 8 of 8) sorted by relevance
/kernel/power/ |
D | snapshot.c | 97 static void *get_image_page(gfp_t gfp_mask, int safe_needed) in get_image_page() argument 101 res = (void *)get_zeroed_page(gfp_mask); in get_image_page() 107 res = (void *)get_zeroed_page(gfp_mask); in get_image_page() 116 unsigned long get_safe_page(gfp_t gfp_mask) in get_safe_page() argument 118 return (unsigned long)get_image_page(gfp_mask, PG_SAFE); in get_safe_page() 121 static struct page *alloc_image_page(gfp_t gfp_mask) in alloc_image_page() argument 125 page = alloc_page(gfp_mask); in alloc_image_page() 191 gfp_t gfp_mask; /* mask for allocating pages */ member 196 chain_init(struct chain_allocator *ca, gfp_t gfp_mask, int safe_needed) in chain_init() argument 200 ca->gfp_mask = gfp_mask; in chain_init() [all …]
|
/kernel/ |
D | audit.c | 181 gfp_t gfp_mask; member 450 static void kauditd_send_multicast_skb(struct sk_buff *skb, gfp_t gfp_mask) in kauditd_send_multicast_skb() argument 469 copy = skb_copy(skb, gfp_mask); in kauditd_send_multicast_skb() 473 nlmsg_multicast(sock, copy, 0, AUDIT_NLGRP_READLOG, gfp_mask); in kauditd_send_multicast_skb() 1261 gfp_t gfp_mask, int type) in audit_buffer_alloc() argument 1277 ab = kmalloc(sizeof(*ab), gfp_mask); in audit_buffer_alloc() 1283 ab->gfp_mask = gfp_mask; in audit_buffer_alloc() 1285 ab->skb = nlmsg_new(AUDIT_BUFSIZ, gfp_mask); in audit_buffer_alloc() 1370 struct audit_buffer *audit_log_start(struct audit_context *ctx, gfp_t gfp_mask, in audit_log_start() argument 1386 if (gfp_mask & __GFP_DIRECT_RECLAIM) { in audit_log_start() [all …]
|
D | kmod.c | 529 char **envp, gfp_t gfp_mask, in call_usermodehelper_setup() argument 535 sub_info = kzalloc(sizeof(struct subprocess_info), gfp_mask); in call_usermodehelper_setup() 628 gfp_t gfp_mask = (wait == UMH_NO_WAIT) ? GFP_ATOMIC : GFP_KERNEL; in call_usermodehelper() local 630 info = call_usermodehelper_setup(path, argv, envp, gfp_mask, in call_usermodehelper()
|
D | kexec_core.c | 143 gfp_t gfp_mask, 282 static struct page *kimage_alloc_pages(gfp_t gfp_mask, unsigned int order) in kimage_alloc_pages() argument 286 pages = alloc_pages(gfp_mask, order); in kimage_alloc_pages() 608 gfp_t gfp_mask, in kimage_alloc_page() argument 648 page = kimage_alloc_pages(gfp_mask, 0); in kimage_alloc_page() 688 if (!(gfp_mask & __GFP_HIGHMEM) && in kimage_alloc_page()
|
D | cpuset.c | 2582 int __cpuset_node_allowed(int node, gfp_t gfp_mask) in __cpuset_node_allowed() argument 2598 if (gfp_mask & __GFP_HARDWALL) /* If hardwall request, stop here */ in __cpuset_node_allowed()
|
D | workqueue.c | 3043 struct workqueue_attrs *alloc_workqueue_attrs(gfp_t gfp_mask) in alloc_workqueue_attrs() argument 3047 attrs = kzalloc(sizeof(*attrs), gfp_mask); in alloc_workqueue_attrs() 3050 if (!alloc_cpumask_var(&attrs->cpumask, gfp_mask)) in alloc_workqueue_attrs()
|
D | cgroup.c | 305 gfp_t gfp_mask) in cgroup_idr_alloc() argument 309 idr_preload(gfp_mask); in cgroup_idr_alloc() 311 ret = idr_alloc(idr, ptr, start, end, gfp_mask & ~__GFP_DIRECT_RECLAIM); in cgroup_idr_alloc()
|
/kernel/locking/ |
D | lockdep.c | 2737 static void __lockdep_trace_alloc(gfp_t gfp_mask, unsigned long flags) in __lockdep_trace_alloc() argument 2745 if (!(gfp_mask & __GFP_DIRECT_RECLAIM)) in __lockdep_trace_alloc() 2749 if ((curr->flags & PF_MEMALLOC) && !(gfp_mask & __GFP_NOMEMALLOC)) in __lockdep_trace_alloc() 2753 if (!(gfp_mask & __GFP_FS)) in __lockdep_trace_alloc() 2767 void lockdep_trace_alloc(gfp_t gfp_mask) in lockdep_trace_alloc() argument 2777 __lockdep_trace_alloc(gfp_mask, flags); in lockdep_trace_alloc() 2893 void lockdep_trace_alloc(gfp_t gfp_mask) in lockdep_trace_alloc() argument 3679 void lockdep_set_current_reclaim_state(gfp_t gfp_mask) in lockdep_set_current_reclaim_state() argument 3681 current->lockdep_reclaim_gfp = gfp_mask; in lockdep_set_current_reclaim_state()
|