/drivers/staging/rtl8723bs/include/ |
D | ioctl_cfg80211.h | 122 #define rtw_cfg80211_rx_mgmt(adapter, freq, sig_dbm, buf, len, gfp) cfg80211_rx_mgmt((adapter)->rtw… argument 124 …tatus(adapter, cookie, buf, len, ack, gfp) cfg80211_mgmt_tx_status((adapter)->rtw_wdev, cookie, bu… argument 125 …okie, chan, channel_type, duration, gfp) cfg80211_ready_on_channel((adapter)->rtw_wdev, cookie, c… argument 126 …ed(adapter, cookie, chan, chan_type, gfp) cfg80211_remain_on_channel_expired((adapter)->rtw_wdev, … argument
|
/drivers/gpu/drm/i915/ |
D | i915_sw_fence.c | 290 wait_queue_entry_t *wq, gfp_t gfp) in __i915_sw_fence_await_sw_fence() argument 308 wq = kmalloc(sizeof(*wq), gfp); in __i915_sw_fence_await_sw_fence() 310 if (!gfpflags_allow_blocking(gfp)) in __i915_sw_fence_await_sw_fence() 349 gfp_t gfp) in i915_sw_fence_await_sw_fence_gfp() argument 351 return __i915_sw_fence_await_sw_fence(fence, signaler, NULL, gfp); in i915_sw_fence_await_sw_fence_gfp() 392 gfp_t gfp) in i915_sw_fence_await_dma_fence() argument 402 cb = kmalloc(sizeof(*cb), gfp); in i915_sw_fence_await_dma_fence() 404 if (!gfpflags_allow_blocking(gfp)) in i915_sw_fence_await_dma_fence() 439 gfp_t gfp) in i915_sw_fence_await_reservation() argument 462 gfp); in i915_sw_fence_await_reservation() [all …]
|
D | i915_gem_internal.c | 55 gfp_t gfp; in i915_gem_object_get_pages_internal() local 71 gfp = GFP_KERNEL | __GFP_HIGHMEM | __GFP_RECLAIMABLE; in i915_gem_object_get_pages_internal() 74 gfp &= ~__GFP_HIGHMEM; in i915_gem_object_get_pages_internal() 75 gfp |= __GFP_DMA32; in i915_gem_object_get_pages_internal() 97 page = alloc_pages(gfp | (order ? QUIET : 0), order); in i915_gem_object_get_pages_internal()
|
D | i915_sw_fence.h | 71 gfp_t gfp); 75 gfp_t gfp); 81 gfp_t gfp);
|
/drivers/base/ |
D | devres.c | 92 size_t size, gfp_t gfp, int nid) in alloc_dr() argument 97 dr = kmalloc_node_track_caller(tot_size, gfp, nid); in alloc_dr() 116 void * __devres_alloc_node(dr_release_t release, size_t size, gfp_t gfp, int nid, in __devres_alloc_node() argument 121 dr = alloc_dr(release, size, gfp | __GFP_ZERO, nid); in __devres_alloc_node() 143 void * devres_alloc_node(dr_release_t release, size_t size, gfp_t gfp, int nid) in devres_alloc_node() argument 147 dr = alloc_dr(release, size, gfp | __GFP_ZERO, nid); in devres_alloc_node() 540 void * devres_open_group(struct device *dev, void *id, gfp_t gfp) in devres_open_group() argument 545 grp = kmalloc(sizeof(*grp), gfp); in devres_open_group() 782 void * devm_kmalloc(struct device *dev, size_t size, gfp_t gfp) in devm_kmalloc() argument 787 dr = alloc_dr(devm_kmalloc_release, size, gfp, dev_to_node(dev)); in devm_kmalloc() [all …]
|
D | devcoredump.c | 210 gfp_t gfp) in dev_coredumpv() argument 212 dev_coredumpm(dev, NULL, data, datalen, gfp, devcd_readv, devcd_freev); in dev_coredumpv() 277 void *data, size_t datalen, gfp_t gfp, in dev_coredumpm() argument 299 devcd = kzalloc(sizeof(*devcd), gfp); in dev_coredumpm() 354 size_t datalen, gfp_t gfp) in dev_coredumpsg() argument 356 dev_coredumpm(dev, NULL, table, datalen, gfp, devcd_read_from_sgtable, in dev_coredumpsg()
|
D | dma-mapping.c | 62 dma_addr_t *dma_handle, gfp_t gfp) in dmam_alloc_coherent() argument 67 dr = devres_alloc(dmam_release, sizeof(*dr), gfp); in dmam_alloc_coherent() 71 vaddr = dma_alloc_coherent(dev, size, dma_handle, gfp); in dmam_alloc_coherent() 121 gfp_t gfp, unsigned long attrs) in dmam_alloc_attrs() argument 126 dr = devres_alloc(dmam_release, sizeof(*dr), gfp); in dmam_alloc_attrs() 130 vaddr = dma_alloc_attrs(dev, size, dma_handle, gfp, attrs); in dmam_alloc_attrs()
|
/drivers/staging/greybus/ |
D | operation.h | 143 gfp_t gfp); 148 size_t response_size, gfp_t gfp) in gb_operation_create() argument 151 response_size, 0, gfp); in gb_operation_create() 158 gfp_t gfp); 164 size_t response_size, gfp_t gfp); 169 gfp_t gfp);
|
D | operation.c | 216 static int gb_message_send(struct gb_message *message, gfp_t gfp) in gb_message_send() argument 224 gfp); in gb_message_send() 469 size_t response_size, gfp_t gfp) in gb_operation_response_alloc() argument 477 response = gb_operation_message_alloc(hd, type, response_size, gfp); in gb_operation_response_alloc() 579 gfp_t gfp) in gb_operation_create_flags() argument 593 flags, gfp); in gb_operation_create_flags() 605 gfp_t gfp) in gb_operation_create_core() argument 613 flags, gfp); in gb_operation_create_core() 723 gfp_t gfp) in gb_operation_request_send() argument 768 ret = gb_message_send(operation->request, gfp); in gb_operation_request_send()
|
/drivers/usb/wusbcore/ |
D | wa-rpipe.c | 194 gfp_t gfp) in rpipe_get_idle() argument 201 rpipe = kzalloc(sizeof(*rpipe), gfp); in rpipe_get_idle() 314 struct usb_host_endpoint *ep, struct urb *urb, gfp_t gfp) in rpipe_aim() argument 398 const struct urb *urb, gfp_t gfp) in rpipe_check_aim() argument 438 struct urb *urb, gfp_t gfp) in rpipe_get_by_ep() argument 449 result = rpipe_check_aim(rpipe, wa, ep, urb, gfp); in rpipe_get_by_ep() 461 result = rpipe_get_idle(&rpipe, wa, 1 << eptype, gfp); in rpipe_get_by_ep() 464 result = rpipe_aim(rpipe, wa, ep, urb, gfp); in rpipe_get_by_ep()
|
/drivers/staging/unisys/visorbus/ |
D | visorchannel.c | 375 static struct visorchannel *visorchannel_create_guts(u64 physaddr, gfp_t gfp, in visorchannel_create_guts() argument 386 channel = kzalloc(sizeof(*channel), gfp); in visorchannel_create_guts() 445 struct visorchannel *visorchannel_create(u64 physaddr, gfp_t gfp, in visorchannel_create() argument 448 return visorchannel_create_guts(physaddr, gfp, guid, false); in visorchannel_create() 451 struct visorchannel *visorchannel_create_with_lock(u64 physaddr, gfp_t gfp, in visorchannel_create_with_lock() argument 454 return visorchannel_create_guts(physaddr, gfp, guid, true); in visorchannel_create_with_lock()
|
D | visorbus_private.h | 41 struct visorchannel *visorchannel_create(u64 physaddr, gfp_t gfp, 43 struct visorchannel *visorchannel_create_with_lock(u64 physaddr, gfp_t gfp,
|
/drivers/virtio/ |
D | virtio_ring.c | 239 unsigned int total_sg, gfp_t gfp) in alloc_indirect() argument 249 gfp &= ~__GFP_HIGHMEM; in alloc_indirect() 251 desc = kmalloc(total_sg * sizeof(struct vring_desc), gfp); in alloc_indirect() 267 gfp_t gfp) in virtqueue_add() argument 306 desc = alloc_indirect(_vq, total_sg, gfp); in virtqueue_add() 457 gfp_t gfp) in virtqueue_add_sgs() argument 468 data, NULL, gfp); in virtqueue_add_sgs() 488 gfp_t gfp) in virtqueue_add_outbuf() argument 490 return virtqueue_add(vq, &sg, num, 1, 0, data, NULL, gfp); in virtqueue_add_outbuf() 510 gfp_t gfp) in virtqueue_add_inbuf() argument [all …]
|
/drivers/md/bcache/ |
D | util.h | 41 #define init_heap(heap, _size, gfp) \ argument 47 (heap)->data = kvmalloc(_bytes, (gfp) & GFP_KERNEL); \ 126 #define __init_fifo(fifo, gfp) \ argument 137 (fifo)->data = kvmalloc(_bytes, (gfp) & GFP_KERNEL); \ 141 #define init_fifo_exact(fifo, _size, gfp) \ argument 144 __init_fifo(fifo, gfp); \ 147 #define init_fifo(fifo, _size, gfp) \ argument 152 __init_fifo(fifo, gfp); \
|
/drivers/misc/mic/scif/ |
D | scif_map.h | 26 gfp_t gfp) in scif_alloc_coherent() argument 31 va = kmalloc(size, gfp); in scif_alloc_coherent() 36 size, dma_handle, gfp); in scif_alloc_coherent()
|
/drivers/mmc/core/ |
D | queue.c | 128 static struct scatterlist *mmc_alloc_sg(int sg_len, gfp_t gfp) in mmc_alloc_sg() argument 132 sg = kmalloc_array(sg_len, sizeof(*sg), gfp); in mmc_alloc_sg() 165 gfp_t gfp) in mmc_init_request() argument 172 mq_rq->sg = mmc_alloc_sg(host->max_segs, gfp); in mmc_init_request()
|
/drivers/net/ethernet/mellanox/mlx5/core/lib/ |
D | mpfs.h | 67 #define l2addr_hash_add(hash, mac, type, gfp) ({ \ argument 71 ptr = kzalloc(sizeof(type), gfp); \
|
/drivers/usb/storage/ |
D | uas.c | 435 static struct urb *uas_alloc_data_urb(struct uas_dev_info *devinfo, gfp_t gfp, in uas_alloc_data_urb() argument 441 struct urb *urb = usb_alloc_urb(0, gfp); in uas_alloc_data_urb() 459 static struct urb *uas_alloc_sense_urb(struct uas_dev_info *devinfo, gfp_t gfp, in uas_alloc_sense_urb() argument 464 struct urb *urb = usb_alloc_urb(0, gfp); in uas_alloc_sense_urb() 470 iu = kzalloc(sizeof(*iu), gfp); in uas_alloc_sense_urb() 486 static struct urb *uas_alloc_cmd_urb(struct uas_dev_info *devinfo, gfp_t gfp, in uas_alloc_cmd_urb() argument 492 struct urb *urb = usb_alloc_urb(0, gfp); in uas_alloc_cmd_urb() 503 iu = kzalloc(sizeof(*iu) + len, gfp); in uas_alloc_cmd_urb() 530 static struct urb *uas_submit_sense_urb(struct scsi_cmnd *cmnd, gfp_t gfp) in uas_submit_sense_urb() argument 536 urb = uas_alloc_sense_urb(devinfo, gfp, cmnd); in uas_submit_sense_urb() [all …]
|
/drivers/crypto/cavium/nitrox/ |
D | nitrox_reqmgr.c | 153 sgcomp = kzalloc(sz_comp, sr->gfp); in create_sg_component() 206 glist = kzalloc(sz, sr->gfp); in dma_map_inbufs() 274 slist = kzalloc(sz, sr->gfp); in dma_map_outbufs() 542 sr = kzalloc(sizeof(*sr), req->gfp); in nitrox_process_se_request() 548 sr->gfp = req->gfp; in nitrox_process_se_request()
|
/drivers/vhost/ |
D | vringh.c | 182 static int resize_iovec(struct vringh_kiov *iov, gfp_t gfp) in resize_iovec() argument 192 new = krealloc(iov->iov, new_num * sizeof(struct iovec), gfp); in resize_iovec() 194 new = kmalloc(new_num * sizeof(struct iovec), gfp); in resize_iovec() 263 gfp_t gfp, in __vringh_iov() argument 359 err = resize_iovec(iov, gfp); in __vringh_iov() 913 gfp_t gfp) in vringh_getdesc_kern() argument 927 gfp, copydesc_kern); in vringh_getdesc_kern()
|
/drivers/iommu/ |
D | dma-iommu.c | 432 unsigned long order_mask, gfp_t gfp) in __iommu_dma_alloc_pages() argument 449 gfp |= __GFP_NOWARN | __GFP_HIGHMEM; in __iommu_dma_alloc_pages() 466 gfp | __GFP_NORETRY : gfp, order); in __iommu_dma_alloc_pages() 526 struct page **iommu_dma_alloc(struct device *dev, size_t size, gfp_t gfp, in iommu_dma_alloc() argument 551 pages = __iommu_dma_alloc_pages(count, alloc_sizes >> PAGE_SHIFT, gfp); in iommu_dma_alloc()
|
/drivers/scsi/cxgbi/ |
D | libcxgbi.h | 348 static inline struct sk_buff *alloc_wr(int wrlen, int dlen, gfp_t gfp) in alloc_wr() argument 350 struct sk_buff *skb = alloc_skb(wrlen + dlen, gfp); in alloc_wr() 538 gfp_t gfp) in cxgbi_alloc_big_mem() argument 540 void *p = kzalloc(size, gfp | __GFP_NOWARN); in cxgbi_alloc_big_mem()
|
/drivers/crypto/ccp/ |
D | ccp-crypto-main.c | 283 gfp_t gfp; in ccp_crypto_enqueue_request() local 285 gfp = req->flags & CRYPTO_TFM_REQ_MAY_SLEEP ? GFP_KERNEL : GFP_ATOMIC; in ccp_crypto_enqueue_request() 287 crypto_cmd = kzalloc(sizeof(*crypto_cmd), gfp); in ccp_crypto_enqueue_request()
|
/drivers/staging/lustre/lnet/libcfs/ |
D | tracefile.c | 110 static struct cfs_trace_page *cfs_tage_alloc(gfp_t gfp) in cfs_tage_alloc() argument 123 gfp |= __GFP_NOWARN; in cfs_tage_alloc() 124 page = alloc_page(gfp); in cfs_tage_alloc() 128 tage = kmalloc(sizeof(*tage), gfp); in cfs_tage_alloc() 152 int cfs_trace_refill_stock(struct cfs_trace_cpu_data *tcd, gfp_t gfp, in cfs_trace_refill_stock() argument 165 tage = cfs_tage_alloc(gfp); in cfs_trace_refill_stock()
|
/drivers/infiniband/hw/cxgb3/ |
D | iwch_cm.c | 109 static struct sk_buff *get_skb(struct sk_buff *skb, int len, gfp_t gfp); 266 static void *alloc_ep(int size, gfp_t gfp) in alloc_ep() argument 270 epc = kzalloc(size, gfp); in alloc_ep() 325 static struct sk_buff *get_skb(struct sk_buff *skb, int len, gfp_t gfp) in get_skb() argument 331 skb = alloc_skb(len, gfp); in get_skb() 388 static int send_halfclose(struct iwch_ep *ep, gfp_t gfp) in send_halfclose() argument 394 skb = get_skb(NULL, sizeof(*req), gfp); in send_halfclose() 408 static int send_abort(struct iwch_ep *ep, struct sk_buff *skb, gfp_t gfp) in send_abort() argument 413 skb = get_skb(skb, sizeof(*req), gfp); in send_abort() 659 static void abort_connection(struct iwch_ep *ep, struct sk_buff *skb, gfp_t gfp) in abort_connection() argument [all …]
|