Home
last modified time | relevance | path

Searched refs:page (Results 1 – 25 of 70) sorted by relevance

123

/net/ceph/
Dpagelist.c11 struct page *page = list_entry(pl->head.prev, struct page, lru); in ceph_pagelist_unmap_tail() local
12 kunmap(page); in ceph_pagelist_unmap_tail()
23 struct page *page = list_first_entry(&pl->head, struct page, in ceph_pagelist_release() local
25 list_del(&page->lru); in ceph_pagelist_release()
26 __free_page(page); in ceph_pagelist_release()
35 struct page *page; in ceph_pagelist_addpage() local
38 page = __page_cache_alloc(GFP_NOFS); in ceph_pagelist_addpage()
40 page = list_first_entry(&pl->free_list, struct page, lru); in ceph_pagelist_addpage()
41 list_del(&page->lru); in ceph_pagelist_addpage()
44 if (!page) in ceph_pagelist_addpage()
[all …]
Dpagevec.c15 struct page **ceph_get_direct_page_vector(const void __user *data, in ceph_get_direct_page_vector()
18 struct page **pages; in ceph_get_direct_page_vector()
45 void ceph_put_page_vector(struct page **pages, int num_pages, bool dirty) in ceph_put_page_vector()
58 void ceph_release_page_vector(struct page **pages, int num_pages) in ceph_release_page_vector()
71 struct page **ceph_alloc_page_vector(int num_pages, gfp_t flags) in ceph_alloc_page_vector()
73 struct page **pages; in ceph_alloc_page_vector()
93 int ceph_copy_user_to_page_vector(struct page **pages, in ceph_copy_user_to_page_vector()
119 void ceph_copy_to_page_vector(struct page **pages, in ceph_copy_to_page_vector()
142 void ceph_copy_from_page_vector(struct page **pages, in ceph_copy_from_page_vector()
169 void ceph_zero_page_vector_range(int off, int len, struct page **pages) in ceph_zero_page_vector_range()
Dmessenger.c196 static struct page *zero_page; /* used in certain error cases */
544 static int ceph_tcp_recvpage(struct socket *sock, struct page *page, in ceph_tcp_recvpage() argument
552 kaddr = kmap(page); in ceph_tcp_recvpage()
555 kunmap(page); in ceph_tcp_recvpage()
581 static int __ceph_tcp_sendpage(struct socket *sock, struct page *page, in __ceph_tcp_sendpage() argument
587 ret = kernel_sendpage(sock, page, offset, size, flags); in __ceph_tcp_sendpage()
594 static int ceph_tcp_sendpage(struct socket *sock, struct page *page, in ceph_tcp_sendpage() argument
602 if (page_count(page) >= 1) in ceph_tcp_sendpage()
603 return __ceph_tcp_sendpage(sock, page, offset, size, more); in ceph_tcp_sendpage()
605 iov.iov_base = kmap(page) + offset; in ceph_tcp_sendpage()
[all …]
/net/rds/
Dpage.c41 struct page *r_page;
56 int rds_page_copy_user(struct page *page, unsigned long offset, in rds_page_copy_user() argument
63 addr = kmap(page); in rds_page_copy_user()
71 kunmap(page); in rds_page_copy_user()
102 struct page *page; in rds_page_remainder_alloc() local
109 page = alloc_page(gfp); in rds_page_remainder_alloc()
110 if (!page) { in rds_page_remainder_alloc()
113 sg_set_page(scat, page, PAGE_SIZE, 0); in rds_page_remainder_alloc()
151 page = alloc_page(gfp); in rds_page_remainder_alloc()
156 if (!page) { in rds_page_remainder_alloc()
[all …]
Drdma.c158 struct page **pages, int write) in rds_pin_pages()
178 struct page **pages = NULL; in __rds_rdma_map()
207 pages = kcalloc(nr_pages, sizeof(struct page *), GFP_KERNEL); in __rds_rdma_map()
449 struct page *page = sg_page(&ro->op_sg[i]); in rds_rdma_free_op() local
455 WARN_ON(!page->mapping && irqs_disabled()); in rds_rdma_free_op()
456 set_page_dirty(page); in rds_rdma_free_op()
458 put_page(page); in rds_rdma_free_op()
468 struct page *page = sg_page(ao->op_sg); in rds_atomic_free_op() local
473 set_page_dirty(page); in rds_atomic_free_op()
474 put_page(page); in rds_atomic_free_op()
[all …]
Dinfo.c65 struct page **pages;
167 struct page **pages = NULL; in rds_info_getsockopt()
191 pages = kmalloc(nr_pages * sizeof(struct page *), GFP_KERNEL); in rds_info_getsockopt()
Drds.h715 int rds_page_copy_user(struct page *page, unsigned long offset,
718 #define rds_page_copy_to_user(page, offset, ptr, bytes) \ argument
719 rds_page_copy_user(page, offset, ptr, bytes, 1)
720 #define rds_page_copy_from_user(page, offset, ptr, bytes) \ argument
721 rds_page_copy_user(page, offset, ptr, bytes, 0)
/net/atm/
Dmpoa_proc.c209 char *page, *p; in proc_mpc_write() local
218 page = (char *)__get_free_page(GFP_KERNEL); in proc_mpc_write()
219 if (!page) in proc_mpc_write()
222 for (p = page, len = 0; len < nbytes; p++, len++) { in proc_mpc_write()
224 free_page((unsigned long)page); in proc_mpc_write()
233 if (!parse_qos(page)) in proc_mpc_write()
234 printk("mpoa: proc_mpc_write: could not parse '%s'\n", page); in proc_mpc_write()
236 free_page((unsigned long)page); in proc_mpc_write()
Dproc.c380 unsigned long page; in proc_dev_atm_read() local
385 page = get_zeroed_page(GFP_KERNEL); in proc_dev_atm_read()
386 if (!page) in proc_dev_atm_read()
392 length = dev->ops->proc_read(dev, pos, (char *)page); in proc_dev_atm_read()
397 if (copy_to_user(buf, (char *)page, length)) in proc_dev_atm_read()
401 free_page(page); in proc_dev_atm_read()
/net/core/
Dskbuff.c536 void skb_add_rx_frag(struct sk_buff *skb, int i, struct page *page, int off, in skb_add_rx_frag() argument
539 skb_fill_page_desc(skb, i, page, off, size); in skb_add_rx_frag()
897 struct page *page, *head = NULL; in skb_copy_ubufs() local
904 page = alloc_page(gfp_mask); in skb_copy_ubufs()
905 if (!page) { in skb_copy_ubufs()
907 struct page *next = (struct page *)page_private(head); in skb_copy_ubufs()
914 memcpy(page_address(page), in skb_copy_ubufs()
917 set_page_private(page, (unsigned long)head); in skb_copy_ubufs()
918 head = page; in skb_copy_ubufs()
931 head = (struct page *)page_private(head); in skb_copy_ubufs()
[all …]
Dtso.c58 tso->data = page_address(frag->page.p) + frag->page_offset; in tso_build_data()
82 tso->data = page_address(frag->page.p) + frag->page_offset; in tso_start()
/net/sunrpc/xprtrdma/
Dsvc_rdma_sendto.c113 struct page *page; in dma_map_xdr() local
118 page = virt_to_page(xdr->head[0].iov_base); in dma_map_xdr()
124 page = xdr->pages[xdr_off >> PAGE_SHIFT]; in dma_map_xdr()
131 page = virt_to_page(xdr->tail[0].iov_base); in dma_map_xdr()
134 dma_addr = ib_dma_map_page(xprt->sc_cm_id->device, page, xdr_off, in dma_map_xdr()
453 struct page *page, in send_reply() argument
479 ctxt->pages[0] = page; in send_reply()
486 ib_dma_map_page(rdma->sc_cm_id->device, page, 0, in send_reply()
580 struct page *res_page; in svc_rdma_sendto()
Dsvc_rdma_recvfrom.c63 struct page *page; in rdma_build_arg_xdr() local
68 page = ctxt->pages[0]; in rdma_build_arg_xdr()
70 rqstp->rq_pages[0] = page; in rdma_build_arg_xdr()
73 rqstp->rq_arg.head[0].iov_base = page_address(page); in rdma_build_arg_xdr()
95 page = ctxt->pages[sge_no]; in rdma_build_arg_xdr()
97 rqstp->rq_pages[sge_no] = page; in rdma_build_arg_xdr()
108 page = ctxt->pages[sge_no++]; in rdma_build_arg_xdr()
109 put_page(page); in rdma_build_arg_xdr()
/net/mac802154/
Dtrace.h78 TP_PROTO(struct ieee802154_local *local, u8 page, u8 channel),
79 TP_ARGS(local, page, channel),
82 __field(u8, page)
87 __entry->page = page;
91 __entry->page, __entry->channel)
Dmib.c29 void mac802154_dev_set_page_channel(struct net_device *dev, u8 page, u8 chan) in mac802154_dev_set_page_channel() argument
39 res = drv_set_channel(local, page, chan); in mac802154_dev_set_page_channel()
44 local->phy->current_page = page; in mac802154_dev_set_page_channel()
Ddriver-ops.h61 drv_set_channel(struct ieee802154_local *local, u8 page, u8 channel) in drv_set_channel() argument
67 trace_802154_drv_set_channel(local, page, channel); in drv_set_channel()
68 ret = local->ops->set_channel(&local->hw, page, channel); in drv_set_channel()
Dcfg.c112 ieee802154_set_channel(struct wpan_phy *wpan_phy, u8 page, u8 channel) in ieee802154_set_channel() argument
119 if (wpan_phy->current_page == page && in ieee802154_set_channel()
123 ret = drv_set_channel(local, page, channel); in ieee802154_set_channel()
125 wpan_phy->current_page = page; in ieee802154_set_channel()
/net/ieee802154/
Dtrace.h102 TP_PROTO(struct wpan_phy *wpan_phy, u8 page, u8 channel),
103 TP_ARGS(wpan_phy, page, channel),
106 __field(u8, page)
111 __entry->page = page;
115 __entry->page, __entry->channel)
Dnl-mac.c183 u8 page; in ieee802154_associate_req() local
212 page = nla_get_u8(info->attrs[IEEE802154_ATTR_PAGE]); in ieee802154_associate_req()
214 page = 0; in ieee802154_associate_req()
218 page, in ieee802154_associate_req()
307 u8 page; in ieee802154_start_req() local
348 page = nla_get_u8(info->attrs[IEEE802154_ATTR_PAGE]); in ieee802154_start_req()
350 page = 0; in ieee802154_start_req()
359 ret = ieee802154_mlme_ops(dev)->start_req(dev, &addr, channel, page, in ieee802154_start_req()
380 u8 page; in ieee802154_scan_req() local
398 page = nla_get_u8(info->attrs[IEEE802154_ATTR_PAGE]); in ieee802154_scan_req()
[all …]
/net/sunrpc/
Dbackchannel_rqst.c74 struct page *page; in xprt_alloc_xdr_buf() local
76 page = alloc_page(gfp_flags); in xprt_alloc_xdr_buf()
77 if (page == NULL) in xprt_alloc_xdr_buf()
79 buf->head[0].iov_base = page_address(page); in xprt_alloc_xdr_buf()
Dsunrpc.h60 struct page *headpage, unsigned long headoffset,
61 struct page *tailpage, unsigned long tailoffset);
Dxdr.c133 struct page **pages, unsigned int base, unsigned int len) in xdr_inline_pages()
172 _shift_data_right_pages(struct page **pages, size_t pgto_base, in _shift_data_right_pages()
175 struct page **pgfrom, **pgto; in _shift_data_right_pages()
233 _copy_to_pages(struct page **pages, size_t pgbase, const char *p, size_t len) in _copy_to_pages()
235 struct page **pgto; in _copy_to_pages()
277 _copy_from_pages(char *p, struct page **pages, size_t pgbase, size_t len) in _copy_from_pages()
279 struct page **pgfrom; in _copy_from_pages()
501 void *page; in xdr_commit_encode() local
505 page = page_address(*xdr->page_ptr); in xdr_commit_encode()
506 memcpy(xdr->scratch.iov_base, page, shift); in xdr_commit_encode()
[all …]
/net/xfrm/
Dxfrm_ipcomp.c73 struct page *page; in ipcomp_decompress() local
80 page = alloc_page(GFP_ATOMIC); in ipcomp_decompress()
83 if (!page) in ipcomp_decompress()
86 __skb_frag_set_page(frag, page); in ipcomp_decompress()
/net/ipv4/
Dudp_impl.h26 int udp_sendpage(struct sock *sk, struct page *page, int offset, size_t size,
Dtcp.c895 static ssize_t do_tcp_sendpages(struct sock *sk, struct page *page, int offset, in do_tcp_sendpages() argument
947 can_coalesce = skb_can_coalesce(skb, i, page, offset); in do_tcp_sendpages()
958 get_page(page); in do_tcp_sendpages()
959 skb_fill_page_desc(skb, i, page, offset, copy); in do_tcp_sendpages()
1022 int tcp_sendpage(struct sock *sk, struct page *page, int offset, in tcp_sendpage() argument
1029 return sock_no_sendpage(sk->sk_socket, page, offset, size, in tcp_sendpage()
1033 res = do_tcp_sendpages(sk, page, offset, size, flags); in tcp_sendpage()
1222 if (!skb_can_coalesce(skb, i, pfrag->page, in tcp_sendmsg()
1237 pfrag->page, in tcp_sendmsg()
1247 skb_fill_page_desc(skb, i, pfrag->page, in tcp_sendmsg()
[all …]

123