/net/ceph/ |
D | pagelist.c | 12 struct page *page = list_entry(pl->head.prev, struct page, lru); in ceph_pagelist_unmap_tail() local 13 kunmap(page); in ceph_pagelist_unmap_tail() 24 struct page *page = list_first_entry(&pl->head, struct page, in ceph_pagelist_release() local 26 list_del(&page->lru); in ceph_pagelist_release() 27 __free_page(page); in ceph_pagelist_release() 36 struct page *page; in ceph_pagelist_addpage() local 39 page = __page_cache_alloc(GFP_NOFS); in ceph_pagelist_addpage() 41 page = list_first_entry(&pl->free_list, struct page, lru); in ceph_pagelist_addpage() 42 list_del(&page->lru); in ceph_pagelist_addpage() 45 if (!page) in ceph_pagelist_addpage() [all …]
|
D | pagevec.c | 16 struct page **ceph_get_direct_page_vector(const void __user *data, in ceph_get_direct_page_vector() 19 struct page **pages; in ceph_get_direct_page_vector() 46 void ceph_put_page_vector(struct page **pages, int num_pages, bool dirty) in ceph_put_page_vector() 59 void ceph_release_page_vector(struct page **pages, int num_pages) in ceph_release_page_vector() 72 struct page **ceph_alloc_page_vector(int num_pages, gfp_t flags) in ceph_alloc_page_vector() 74 struct page **pages; in ceph_alloc_page_vector() 94 int ceph_copy_user_to_page_vector(struct page **pages, in ceph_copy_user_to_page_vector() 120 void ceph_copy_to_page_vector(struct page **pages, in ceph_copy_to_page_vector() 143 void ceph_copy_from_page_vector(struct page **pages, in ceph_copy_from_page_vector() 170 void ceph_zero_page_vector_range(int off, int len, struct page **pages) in ceph_zero_page_vector_range()
|
D | messenger.c | 194 static struct page *zero_page; /* used in certain error cases */ 537 static int ceph_tcp_recvpage(struct socket *sock, struct page *page, in ceph_tcp_recvpage() argument 541 .bv_page = page, in ceph_tcp_recvpage() 577 static int __ceph_tcp_sendpage(struct socket *sock, struct page *page, in __ceph_tcp_sendpage() argument 583 ret = kernel_sendpage(sock, page, offset, size, flags); in __ceph_tcp_sendpage() 590 static int ceph_tcp_sendpage(struct socket *sock, struct page *page, in ceph_tcp_sendpage() argument 605 if (page_count(page) >= 1 && !PageSlab(page)) in ceph_tcp_sendpage() 606 return __ceph_tcp_sendpage(sock, page, offset, size, more); in ceph_tcp_sendpage() 608 bvec.bv_page = page; in ceph_tcp_sendpage() 861 static struct page *ceph_msg_data_bio_next(struct ceph_msg_data_cursor *cursor, in ceph_msg_data_bio_next() [all …]
|
D | crypto.c | 191 struct page *page; in setup_sgtable() local 195 page = vmalloc_to_page(buf); in setup_sgtable() 197 page = virt_to_page(buf); in setup_sgtable() 199 sg_set_page(sg, page, len, off); in setup_sgtable()
|
D | cls_lock_client.c | 34 struct page *lock_op_page; in ceph_cls_lock() 97 struct page *unlock_op_page; in ceph_cls_unlock() 146 struct page *break_op_page; in ceph_cls_break_lock() 195 struct page *cookie_op_page; in ceph_cls_set_cookie() 332 struct page *get_info_op_page, *reply_page; in ceph_cls_lock_info()
|
/net/rds/ |
D | page.c | 41 struct page *r_page; 73 struct page *page; in rds_page_remainder_alloc() local 80 page = alloc_page(gfp); in rds_page_remainder_alloc() 81 if (!page) { in rds_page_remainder_alloc() 84 sg_set_page(scat, page, PAGE_SIZE, 0); in rds_page_remainder_alloc() 122 page = alloc_page(gfp); in rds_page_remainder_alloc() 127 if (!page) { in rds_page_remainder_alloc() 134 __free_page(page); in rds_page_remainder_alloc() 139 rem->r_page = page; in rds_page_remainder_alloc()
|
D | rdma.c | 157 struct page **pages, int write) in rds_pin_pages() 177 struct page **pages = NULL; in __rds_rdma_map() 214 pages = kcalloc(nr_pages, sizeof(struct page *), GFP_KERNEL); in __rds_rdma_map() 457 struct page *page = sg_page(&ro->op_sg[i]); in rds_rdma_free_op() local 463 WARN_ON(!page->mapping && irqs_disabled()); in rds_rdma_free_op() 464 set_page_dirty(page); in rds_rdma_free_op() 466 put_page(page); in rds_rdma_free_op() 476 struct page *page = sg_page(ao->op_sg); in rds_atomic_free_op() local 481 set_page_dirty(page); in rds_atomic_free_op() 482 put_page(page); in rds_atomic_free_op() [all …]
|
D | info.c | 65 struct page **pages; 167 struct page **pages = NULL; in rds_info_getsockopt() 191 pages = kmalloc(nr_pages * sizeof(struct page *), GFP_KERNEL); in rds_info_getsockopt()
|
/net/sunrpc/xprtrdma/ |
D | svc_rdma_recvfrom.c | 117 struct page *page; in rdma_build_arg_xdr() local 122 page = ctxt->pages[0]; in rdma_build_arg_xdr() 124 rqstp->rq_pages[0] = page; in rdma_build_arg_xdr() 127 rqstp->rq_arg.head[0].iov_base = page_address(page); in rdma_build_arg_xdr() 142 page = ctxt->pages[sge_no]; in rdma_build_arg_xdr() 144 rqstp->rq_pages[sge_no] = page; in rdma_build_arg_xdr() 154 page = ctxt->pages[sge_no++]; in rdma_build_arg_xdr() 155 put_page(page); in rdma_build_arg_xdr() 400 struct page *page; in svc_rdma_send_error() local 407 page = alloc_page(GFP_KERNEL); in svc_rdma_send_error() [all …]
|
D | svc_rdma_backchannel.c | 164 struct page *page; in xprt_rdma_bc_allocate() local 172 page = alloc_page(RPCRDMA_DEF_GFP); in xprt_rdma_bc_allocate() 173 if (!page) in xprt_rdma_bc_allocate() 175 rqst->rq_buffer = page_address(page); in xprt_rdma_bc_allocate() 179 put_page(page); in xprt_rdma_bc_allocate()
|
D | svc_rdma_rw.c | 372 struct page **page; in svc_rdma_pagelist_to_sg() local 377 page = xdr->pages + page_no; in svc_rdma_pagelist_to_sg() 384 sg_set_page(sg, *page, sge_bytes, page_off); in svc_rdma_pagelist_to_sg() 390 page++; in svc_rdma_pagelist_to_sg() 819 struct page **page; in svc_rdma_recv_read_chunk() local 862 for (page = rqstp->rq_pages; page < rqstp->rq_respages; page++) in svc_rdma_recv_read_chunk() 863 *page = NULL; in svc_rdma_recv_read_chunk()
|
D | svc_rdma_sendto.c | 332 struct page *page, in svc_rdma_dma_map_page() argument 339 dma_addr = ib_dma_map_page(dev, page, offset, len, DMA_TO_DEVICE); in svc_rdma_dma_map_page() 387 struct page **ppages; in svc_rdma_map_reply_msg() 628 struct page *res_page; in svc_rdma_sendto()
|
/net/atm/ |
D | mpoa_proc.c | 210 char *page, *p; in proc_mpc_write() local 219 page = (char *)__get_free_page(GFP_KERNEL); in proc_mpc_write() 220 if (!page) in proc_mpc_write() 223 for (p = page, len = 0; len < nbytes; p++, len++) { in proc_mpc_write() 225 free_page((unsigned long)page); in proc_mpc_write() 234 if (!parse_qos(page)) in proc_mpc_write() 235 printk("mpoa: proc_mpc_write: could not parse '%s'\n", page); in proc_mpc_write() 237 free_page((unsigned long)page); in proc_mpc_write()
|
D | proc.c | 381 unsigned long page; in proc_dev_atm_read() local 386 page = get_zeroed_page(GFP_KERNEL); in proc_dev_atm_read() 387 if (!page) in proc_dev_atm_read() 393 length = dev->ops->proc_read(dev, pos, (char *)page); in proc_dev_atm_read() 398 if (copy_to_user(buf, (char *)page, length)) in proc_dev_atm_read() 402 free_page(page); in proc_dev_atm_read()
|
/net/core/ |
D | skbuff.c | 326 struct page_frag_cache page; member 366 return page_frag_alloc(&nc->page, fragsz, gfp_mask); in __napi_alloc_frag() 482 data = page_frag_alloc(&nc->page, len, gfp_mask); in __napi_alloc_skb() 493 if (nc->page.pfmemalloc) in __napi_alloc_skb() 506 void skb_add_rx_frag(struct sk_buff *skb, int i, struct page *page, int off, in skb_add_rx_frag() argument 509 skb_fill_page_desc(skb, i, page, off, size); in skb_add_rx_frag() 1179 struct page *page, *head = NULL; in skb_copy_ubufs() local 1191 page = alloc_page(gfp_mask); in skb_copy_ubufs() 1192 if (!page) { in skb_copy_ubufs() 1194 struct page *next = (struct page *)page_private(head); in skb_copy_ubufs() [all …]
|
/net/mac802154/ |
D | trace.h | 79 TP_PROTO(struct ieee802154_local *local, u8 page, u8 channel), 80 TP_ARGS(local, page, channel), 83 __field(u8, page) 88 __entry->page = page; 92 __entry->page, __entry->channel)
|
/net/ieee802154/ |
D | trace.h | 103 TP_PROTO(struct wpan_phy *wpan_phy, u8 page, u8 channel), 104 TP_ARGS(wpan_phy, page, channel), 107 __field(u8, page) 112 __entry->page = page; 116 __entry->page, __entry->channel)
|
/net/sunrpc/ |
D | backchannel_rqst.c | 74 struct page *page; in xprt_alloc_xdr_buf() local 76 page = alloc_page(gfp_flags); in xprt_alloc_xdr_buf() 77 if (page == NULL) in xprt_alloc_xdr_buf() 79 xdr_buf_init(buf, page_address(page), PAGE_SIZE); in xprt_alloc_xdr_buf()
|
D | sunrpc.h | 60 struct page *headpage, unsigned long headoffset, 61 struct page *tailpage, unsigned long tailoffset);
|
D | xdr.c | 133 struct page **pages, unsigned int base, unsigned int len) in xdr_inline_pages() 172 _shift_data_right_pages(struct page **pages, size_t pgto_base, in _shift_data_right_pages() 175 struct page **pgfrom, **pgto; in _shift_data_right_pages() 233 _copy_to_pages(struct page **pages, size_t pgbase, const char *p, size_t len) in _copy_to_pages() 235 struct page **pgto; in _copy_to_pages() 277 _copy_from_pages(char *p, struct page **pages, size_t pgbase, size_t len) in _copy_from_pages() 279 struct page **pgfrom; in _copy_from_pages() 501 void *page; in xdr_commit_encode() local 505 page = page_address(*xdr->page_ptr); in xdr_commit_encode() 506 memcpy(xdr->scratch.iov_base, page, shift); in xdr_commit_encode() [all …]
|
/net/ipv6/ |
D | esp6.c | 224 struct page *page; in esp6_output_head() local 252 page = pfrag->page; in esp6_output_head() 253 get_page(page); in esp6_output_head() 255 vaddr = kmap_atomic(page); in esp6_output_head() 265 __skb_fill_page_desc(skb, nfrags, page, pfrag->offset, in esp6_output_head() 309 struct page *page; in esp6_output_tail() local 365 page = pfrag->page; in esp6_output_tail() 366 get_page(page); in esp6_output_tail() 368 __skb_fill_page_desc(skb, 0, page, pfrag->offset, skb->data_len); in esp6_output_tail()
|
/net/ipv4/ |
D | esp4.c | 258 struct page *page; in esp_output_head() local 294 page = pfrag->page; in esp_output_head() 295 get_page(page); in esp_output_head() 297 vaddr = kmap_atomic(page); in esp_output_head() 307 __skb_fill_page_desc(skb, nfrags, page, pfrag->offset, in esp_output_head() 353 struct page *page; in esp_output_tail() local 411 page = pfrag->page; in esp_output_tail() 412 get_page(page); in esp_output_tail() 414 __skb_fill_page_desc(skb, 0, page, pfrag->offset, skb->data_len); in esp_output_tail()
|
D | udp_impl.h | 27 int udp_sendpage(struct sock *sk, struct page *page, int offset, size_t size,
|
/net/xfrm/ |
D | xfrm_ipcomp.c | 73 struct page *page; in ipcomp_decompress() local 80 page = alloc_page(GFP_ATOMIC); in ipcomp_decompress() 83 if (!page) in ipcomp_decompress() 86 __skb_frag_set_page(frag, page); in ipcomp_decompress()
|
/net/tls/ |
D | tls_sw.c | 139 if (num_elem > first_coalesce && sg_page(sge) == pfrag->page && in alloc_sg() 145 sg_set_page(sge, pfrag->page, use, orig_offset); in alloc_sg() 146 get_page(pfrag->page); in alloc_sg() 307 struct page *pages[MAX_SKB_FRAGS]; in zerocopy_from_iter() 538 int tls_sw_sendpage(struct sock *sk, struct page *page, in tls_sw_sendpage() argument 602 get_page(page); in tls_sw_sendpage() 604 sg_set_page(sg, page, copy, offset); in tls_sw_sendpage()
|