Home
last modified time | relevance | path

Searched refs:gfp (Results 1 – 25 of 134) sorted by relevance

123456

/drivers/staging/rtl8723bs/include/
Dioctl_cfg80211.h122 #define rtw_cfg80211_rx_mgmt(adapter, freq, sig_dbm, buf, len, gfp) cfg80211_rx_mgmt((adapter)->rtw… argument
124 …tatus(adapter, cookie, buf, len, ack, gfp) cfg80211_mgmt_tx_status((adapter)->rtw_wdev, cookie, bu… argument
125 …okie, chan, channel_type, duration, gfp) cfg80211_ready_on_channel((adapter)->rtw_wdev, cookie, c… argument
126 …ed(adapter, cookie, chan, chan_type, gfp) cfg80211_remain_on_channel_expired((adapter)->rtw_wdev, … argument
/drivers/gpu/drm/i915/
Di915_sw_fence.c290 wait_queue_entry_t *wq, gfp_t gfp) in __i915_sw_fence_await_sw_fence() argument
308 wq = kmalloc(sizeof(*wq), gfp); in __i915_sw_fence_await_sw_fence()
310 if (!gfpflags_allow_blocking(gfp)) in __i915_sw_fence_await_sw_fence()
349 gfp_t gfp) in i915_sw_fence_await_sw_fence_gfp() argument
351 return __i915_sw_fence_await_sw_fence(fence, signaler, NULL, gfp); in i915_sw_fence_await_sw_fence_gfp()
392 gfp_t gfp) in i915_sw_fence_await_dma_fence() argument
402 cb = kmalloc(sizeof(*cb), gfp); in i915_sw_fence_await_dma_fence()
404 if (!gfpflags_allow_blocking(gfp)) in i915_sw_fence_await_dma_fence()
439 gfp_t gfp) in i915_sw_fence_await_reservation() argument
462 gfp); in i915_sw_fence_await_reservation()
[all …]
Di915_gem_internal.c55 gfp_t gfp; in i915_gem_object_get_pages_internal() local
71 gfp = GFP_KERNEL | __GFP_HIGHMEM | __GFP_RECLAIMABLE; in i915_gem_object_get_pages_internal()
74 gfp &= ~__GFP_HIGHMEM; in i915_gem_object_get_pages_internal()
75 gfp |= __GFP_DMA32; in i915_gem_object_get_pages_internal()
97 page = alloc_pages(gfp | (order ? QUIET : 0), order); in i915_gem_object_get_pages_internal()
Di915_sw_fence.h71 gfp_t gfp);
75 gfp_t gfp);
81 gfp_t gfp);
/drivers/base/
Ddevres.c92 size_t size, gfp_t gfp, int nid) in alloc_dr() argument
97 dr = kmalloc_node_track_caller(tot_size, gfp, nid); in alloc_dr()
116 void * __devres_alloc_node(dr_release_t release, size_t size, gfp_t gfp, int nid, in __devres_alloc_node() argument
121 dr = alloc_dr(release, size, gfp | __GFP_ZERO, nid); in __devres_alloc_node()
143 void * devres_alloc_node(dr_release_t release, size_t size, gfp_t gfp, int nid) in devres_alloc_node() argument
147 dr = alloc_dr(release, size, gfp | __GFP_ZERO, nid); in devres_alloc_node()
540 void * devres_open_group(struct device *dev, void *id, gfp_t gfp) in devres_open_group() argument
545 grp = kmalloc(sizeof(*grp), gfp); in devres_open_group()
782 void * devm_kmalloc(struct device *dev, size_t size, gfp_t gfp) in devm_kmalloc() argument
787 dr = alloc_dr(devm_kmalloc_release, size, gfp, dev_to_node(dev)); in devm_kmalloc()
[all …]
Ddevcoredump.c210 gfp_t gfp) in dev_coredumpv() argument
212 dev_coredumpm(dev, NULL, data, datalen, gfp, devcd_readv, devcd_freev); in dev_coredumpv()
277 void *data, size_t datalen, gfp_t gfp, in dev_coredumpm() argument
299 devcd = kzalloc(sizeof(*devcd), gfp); in dev_coredumpm()
354 size_t datalen, gfp_t gfp) in dev_coredumpsg() argument
356 dev_coredumpm(dev, NULL, table, datalen, gfp, devcd_read_from_sgtable, in dev_coredumpsg()
Ddma-mapping.c62 dma_addr_t *dma_handle, gfp_t gfp) in dmam_alloc_coherent() argument
67 dr = devres_alloc(dmam_release, sizeof(*dr), gfp); in dmam_alloc_coherent()
71 vaddr = dma_alloc_coherent(dev, size, dma_handle, gfp); in dmam_alloc_coherent()
121 gfp_t gfp, unsigned long attrs) in dmam_alloc_attrs() argument
126 dr = devres_alloc(dmam_release, sizeof(*dr), gfp); in dmam_alloc_attrs()
130 vaddr = dma_alloc_attrs(dev, size, dma_handle, gfp, attrs); in dmam_alloc_attrs()
/drivers/staging/greybus/
Doperation.h143 gfp_t gfp);
148 size_t response_size, gfp_t gfp) in gb_operation_create() argument
151 response_size, 0, gfp); in gb_operation_create()
158 gfp_t gfp);
164 size_t response_size, gfp_t gfp);
169 gfp_t gfp);
Doperation.c216 static int gb_message_send(struct gb_message *message, gfp_t gfp) in gb_message_send() argument
224 gfp); in gb_message_send()
469 size_t response_size, gfp_t gfp) in gb_operation_response_alloc() argument
477 response = gb_operation_message_alloc(hd, type, response_size, gfp); in gb_operation_response_alloc()
579 gfp_t gfp) in gb_operation_create_flags() argument
593 flags, gfp); in gb_operation_create_flags()
605 gfp_t gfp) in gb_operation_create_core() argument
613 flags, gfp); in gb_operation_create_core()
723 gfp_t gfp) in gb_operation_request_send() argument
768 ret = gb_message_send(operation->request, gfp); in gb_operation_request_send()
/drivers/usb/wusbcore/
Dwa-rpipe.c194 gfp_t gfp) in rpipe_get_idle() argument
201 rpipe = kzalloc(sizeof(*rpipe), gfp); in rpipe_get_idle()
314 struct usb_host_endpoint *ep, struct urb *urb, gfp_t gfp) in rpipe_aim() argument
398 const struct urb *urb, gfp_t gfp) in rpipe_check_aim() argument
438 struct urb *urb, gfp_t gfp) in rpipe_get_by_ep() argument
449 result = rpipe_check_aim(rpipe, wa, ep, urb, gfp); in rpipe_get_by_ep()
461 result = rpipe_get_idle(&rpipe, wa, 1 << eptype, gfp); in rpipe_get_by_ep()
464 result = rpipe_aim(rpipe, wa, ep, urb, gfp); in rpipe_get_by_ep()
/drivers/staging/unisys/visorbus/
Dvisorchannel.c375 static struct visorchannel *visorchannel_create_guts(u64 physaddr, gfp_t gfp, in visorchannel_create_guts() argument
386 channel = kzalloc(sizeof(*channel), gfp); in visorchannel_create_guts()
445 struct visorchannel *visorchannel_create(u64 physaddr, gfp_t gfp, in visorchannel_create() argument
448 return visorchannel_create_guts(physaddr, gfp, guid, false); in visorchannel_create()
451 struct visorchannel *visorchannel_create_with_lock(u64 physaddr, gfp_t gfp, in visorchannel_create_with_lock() argument
454 return visorchannel_create_guts(physaddr, gfp, guid, true); in visorchannel_create_with_lock()
Dvisorbus_private.h41 struct visorchannel *visorchannel_create(u64 physaddr, gfp_t gfp,
43 struct visorchannel *visorchannel_create_with_lock(u64 physaddr, gfp_t gfp,
/drivers/virtio/
Dvirtio_ring.c239 unsigned int total_sg, gfp_t gfp) in alloc_indirect() argument
249 gfp &= ~__GFP_HIGHMEM; in alloc_indirect()
251 desc = kmalloc(total_sg * sizeof(struct vring_desc), gfp); in alloc_indirect()
267 gfp_t gfp) in virtqueue_add() argument
306 desc = alloc_indirect(_vq, total_sg, gfp); in virtqueue_add()
457 gfp_t gfp) in virtqueue_add_sgs() argument
468 data, NULL, gfp); in virtqueue_add_sgs()
488 gfp_t gfp) in virtqueue_add_outbuf() argument
490 return virtqueue_add(vq, &sg, num, 1, 0, data, NULL, gfp); in virtqueue_add_outbuf()
510 gfp_t gfp) in virtqueue_add_inbuf() argument
[all …]
/drivers/md/bcache/
Dutil.h41 #define init_heap(heap, _size, gfp) \ argument
47 (heap)->data = kvmalloc(_bytes, (gfp) & GFP_KERNEL); \
126 #define __init_fifo(fifo, gfp) \ argument
137 (fifo)->data = kvmalloc(_bytes, (gfp) & GFP_KERNEL); \
141 #define init_fifo_exact(fifo, _size, gfp) \ argument
144 __init_fifo(fifo, gfp); \
147 #define init_fifo(fifo, _size, gfp) \ argument
152 __init_fifo(fifo, gfp); \
/drivers/misc/mic/scif/
Dscif_map.h26 gfp_t gfp) in scif_alloc_coherent() argument
31 va = kmalloc(size, gfp); in scif_alloc_coherent()
36 size, dma_handle, gfp); in scif_alloc_coherent()
/drivers/mmc/core/
Dqueue.c128 static struct scatterlist *mmc_alloc_sg(int sg_len, gfp_t gfp) in mmc_alloc_sg() argument
132 sg = kmalloc_array(sg_len, sizeof(*sg), gfp); in mmc_alloc_sg()
165 gfp_t gfp) in mmc_init_request() argument
172 mq_rq->sg = mmc_alloc_sg(host->max_segs, gfp); in mmc_init_request()
/drivers/net/ethernet/mellanox/mlx5/core/lib/
Dmpfs.h67 #define l2addr_hash_add(hash, mac, type, gfp) ({ \ argument
71 ptr = kzalloc(sizeof(type), gfp); \
/drivers/usb/storage/
Duas.c435 static struct urb *uas_alloc_data_urb(struct uas_dev_info *devinfo, gfp_t gfp, in uas_alloc_data_urb() argument
441 struct urb *urb = usb_alloc_urb(0, gfp); in uas_alloc_data_urb()
459 static struct urb *uas_alloc_sense_urb(struct uas_dev_info *devinfo, gfp_t gfp, in uas_alloc_sense_urb() argument
464 struct urb *urb = usb_alloc_urb(0, gfp); in uas_alloc_sense_urb()
470 iu = kzalloc(sizeof(*iu), gfp); in uas_alloc_sense_urb()
486 static struct urb *uas_alloc_cmd_urb(struct uas_dev_info *devinfo, gfp_t gfp, in uas_alloc_cmd_urb() argument
492 struct urb *urb = usb_alloc_urb(0, gfp); in uas_alloc_cmd_urb()
503 iu = kzalloc(sizeof(*iu) + len, gfp); in uas_alloc_cmd_urb()
530 static struct urb *uas_submit_sense_urb(struct scsi_cmnd *cmnd, gfp_t gfp) in uas_submit_sense_urb() argument
536 urb = uas_alloc_sense_urb(devinfo, gfp, cmnd); in uas_submit_sense_urb()
[all …]
/drivers/crypto/cavium/nitrox/
Dnitrox_reqmgr.c153 sgcomp = kzalloc(sz_comp, sr->gfp); in create_sg_component()
206 glist = kzalloc(sz, sr->gfp); in dma_map_inbufs()
274 slist = kzalloc(sz, sr->gfp); in dma_map_outbufs()
542 sr = kzalloc(sizeof(*sr), req->gfp); in nitrox_process_se_request()
548 sr->gfp = req->gfp; in nitrox_process_se_request()
/drivers/vhost/
Dvringh.c182 static int resize_iovec(struct vringh_kiov *iov, gfp_t gfp) in resize_iovec() argument
192 new = krealloc(iov->iov, new_num * sizeof(struct iovec), gfp); in resize_iovec()
194 new = kmalloc(new_num * sizeof(struct iovec), gfp); in resize_iovec()
263 gfp_t gfp, in __vringh_iov() argument
359 err = resize_iovec(iov, gfp); in __vringh_iov()
913 gfp_t gfp) in vringh_getdesc_kern() argument
927 gfp, copydesc_kern); in vringh_getdesc_kern()
/drivers/iommu/
Ddma-iommu.c432 unsigned long order_mask, gfp_t gfp) in __iommu_dma_alloc_pages() argument
449 gfp |= __GFP_NOWARN | __GFP_HIGHMEM; in __iommu_dma_alloc_pages()
466 gfp | __GFP_NORETRY : gfp, order); in __iommu_dma_alloc_pages()
526 struct page **iommu_dma_alloc(struct device *dev, size_t size, gfp_t gfp, in iommu_dma_alloc() argument
551 pages = __iommu_dma_alloc_pages(count, alloc_sizes >> PAGE_SHIFT, gfp); in iommu_dma_alloc()
/drivers/scsi/cxgbi/
Dlibcxgbi.h348 static inline struct sk_buff *alloc_wr(int wrlen, int dlen, gfp_t gfp) in alloc_wr() argument
350 struct sk_buff *skb = alloc_skb(wrlen + dlen, gfp); in alloc_wr()
538 gfp_t gfp) in cxgbi_alloc_big_mem() argument
540 void *p = kzalloc(size, gfp | __GFP_NOWARN); in cxgbi_alloc_big_mem()
/drivers/crypto/ccp/
Dccp-crypto-main.c283 gfp_t gfp; in ccp_crypto_enqueue_request() local
285 gfp = req->flags & CRYPTO_TFM_REQ_MAY_SLEEP ? GFP_KERNEL : GFP_ATOMIC; in ccp_crypto_enqueue_request()
287 crypto_cmd = kzalloc(sizeof(*crypto_cmd), gfp); in ccp_crypto_enqueue_request()
/drivers/staging/lustre/lnet/libcfs/
Dtracefile.c110 static struct cfs_trace_page *cfs_tage_alloc(gfp_t gfp) in cfs_tage_alloc() argument
123 gfp |= __GFP_NOWARN; in cfs_tage_alloc()
124 page = alloc_page(gfp); in cfs_tage_alloc()
128 tage = kmalloc(sizeof(*tage), gfp); in cfs_tage_alloc()
152 int cfs_trace_refill_stock(struct cfs_trace_cpu_data *tcd, gfp_t gfp, in cfs_trace_refill_stock() argument
165 tage = cfs_tage_alloc(gfp); in cfs_trace_refill_stock()
/drivers/infiniband/hw/cxgb3/
Diwch_cm.c109 static struct sk_buff *get_skb(struct sk_buff *skb, int len, gfp_t gfp);
266 static void *alloc_ep(int size, gfp_t gfp) in alloc_ep() argument
270 epc = kzalloc(size, gfp); in alloc_ep()
325 static struct sk_buff *get_skb(struct sk_buff *skb, int len, gfp_t gfp) in get_skb() argument
331 skb = alloc_skb(len, gfp); in get_skb()
388 static int send_halfclose(struct iwch_ep *ep, gfp_t gfp) in send_halfclose() argument
394 skb = get_skb(NULL, sizeof(*req), gfp); in send_halfclose()
408 static int send_abort(struct iwch_ep *ep, struct sk_buff *skb, gfp_t gfp) in send_abort() argument
413 skb = get_skb(skb, sizeof(*req), gfp); in send_abort()
659 static void abort_connection(struct iwch_ep *ep, struct sk_buff *skb, gfp_t gfp) in abort_connection() argument
[all …]

123456