Home
last modified time | relevance | path

Searched refs:iov (Results 1 – 25 of 43) sorted by relevance

12

/net/core/
Diovec.c38 int verify_iovec(struct msghdr *m, struct iovec *iov, struct sockaddr_storage *address, int mode) in verify_iovec() argument
58 if (copy_from_user(iov, (void __user __force *) m->msg_iov, size)) in verify_iovec()
61 m->msg_iov = iov; in verify_iovec()
65 size_t len = iov[ct].iov_len; in verify_iovec()
69 iov[ct].iov_len = len; in verify_iovec()
85 int csum_partial_copy_fromiovecend(unsigned char *kdata, struct iovec *iov, in csum_partial_copy_fromiovecend() argument
92 while (offset >= iov->iov_len) { in csum_partial_copy_fromiovecend()
93 offset -= iov->iov_len; in csum_partial_copy_fromiovecend()
94 iov++; in csum_partial_copy_fromiovecend()
98 u8 __user *base = iov->iov_base + offset; in csum_partial_copy_fromiovecend()
[all …]
Ddatagram.c830 int hlen, struct iovec *iov) in skb_copy_and_csum_datagram_iovec() argument
841 while (!iov->iov_len) in skb_copy_and_csum_datagram_iovec()
842 iov++; in skb_copy_and_csum_datagram_iovec()
844 if (iov->iov_len < chunk) { in skb_copy_and_csum_datagram_iovec()
847 if (skb_copy_datagram_iovec(skb, hlen, iov, chunk)) in skb_copy_and_csum_datagram_iovec()
851 if (skb_copy_and_csum_datagram(skb, hlen, iov->iov_base, in skb_copy_and_csum_datagram_iovec()
858 iov->iov_len -= chunk; in skb_copy_and_csum_datagram_iovec()
859 iov->iov_base += chunk; in skb_copy_and_csum_datagram_iovec()
/net/rds/
Dmessage.c273 struct iovec *iov; in rds_message_copy_from_user() local
283 iov = first_iov; in rds_message_copy_from_user()
297 while (iov_off == iov->iov_len) { in rds_message_copy_from_user()
299 iov++; in rds_message_copy_from_user()
302 to_copy = min(iov->iov_len - iov_off, sg->length - sg_off); in rds_message_copy_from_user()
307 to_copy, iov->iov_base, iov->iov_len, iov_off, in rds_message_copy_from_user()
311 iov->iov_base + iov_off, in rds_message_copy_from_user()
332 struct iovec *iov; in rds_message_inc_copy_to_user() local
344 iov = first_iov; in rds_message_inc_copy_to_user()
351 while (iov_off == iov->iov_len) { in rds_message_inc_copy_to_user()
[all …]
Drdma.c484 static int rds_rdma_pages(struct rds_iovec iov[], int nr_iovecs) in rds_rdma_pages() argument
492 nr_pages = rds_pages_in_vec(&iov[i]); in rds_rdma_pages()
661 struct rds_iovec *iov = &iovs[i]; in rds_cmsg_rdma_args() local
663 unsigned int nr = rds_pages_in_vec(iov); in rds_cmsg_rdma_args()
665 rs->rs_user_addr = iov->addr; in rds_cmsg_rdma_args()
666 rs->rs_user_bytes = iov->bytes; in rds_cmsg_rdma_args()
671 ret = rds_pin_pages(iov->addr, nr, pages, !op->op_write); in rds_cmsg_rdma_args()
676 nr_bytes, nr, iov->bytes, iov->addr); in rds_cmsg_rdma_args()
678 nr_bytes += iov->bytes; in rds_cmsg_rdma_args()
681 unsigned int offset = iov->addr & ~PAGE_MASK; in rds_cmsg_rdma_args()
[all …]
Dtcp_recv.c66 struct iovec *iov, tmp; in rds_tcp_inc_copy_to_user() local
75 iov = first_iov; in rds_tcp_inc_copy_to_user()
76 tmp = *iov; in rds_tcp_inc_copy_to_user()
82 iov++; in rds_tcp_inc_copy_to_user()
83 tmp = *iov; in rds_tcp_inc_copy_to_user()
Diw_recv.c311 struct iovec *iov = first_iov; in rds_iw_inc_copy_to_user() local
329 while (iov_off == iov->iov_len) { in rds_iw_inc_copy_to_user()
331 iov++; in rds_iw_inc_copy_to_user()
334 to_copy = min(iov->iov_len - iov_off, RDS_FRAG_SIZE - frag_off); in rds_iw_inc_copy_to_user()
340 to_copy, iov->iov_base, iov->iov_len, iov_off, in rds_iw_inc_copy_to_user()
346 iov->iov_base + iov_off, in rds_iw_inc_copy_to_user()
Dib_recv.c480 struct iovec *iov = first_iov; in rds_ib_inc_copy_to_user() local
498 while (iov_off == iov->iov_len) { in rds_ib_inc_copy_to_user()
500 iov++; in rds_ib_inc_copy_to_user()
503 to_copy = min(iov->iov_len - iov_off, RDS_FRAG_SIZE - frag_off); in rds_ib_inc_copy_to_user()
509 to_copy, iov->iov_base, iov->iov_len, iov_off, in rds_ib_inc_copy_to_user()
515 iov->iov_base + iov_off, in rds_ib_inc_copy_to_user()
/net/rxrpc/
Dar-connevent.c65 struct kvec iov[2]; in rxrpc_abort_connection() local
105 iov[0].iov_base = &hdr; in rxrpc_abort_connection()
106 iov[0].iov_len = sizeof(hdr); in rxrpc_abort_connection()
107 iov[1].iov_base = &word; in rxrpc_abort_connection()
108 iov[1].iov_len = sizeof(word); in rxrpc_abort_connection()
110 len = iov[0].iov_len + iov[1].iov_len; in rxrpc_abort_connection()
115 ret = kernel_sendmsg(conn->trans->local->socket, &msg, iov, 2, len); in rxrpc_abort_connection()
342 struct kvec iov[2]; in rxrpc_reject_packets() local
351 iov[0].iov_base = &hdr; in rxrpc_reject_packets()
352 iov[0].iov_len = sizeof(hdr); in rxrpc_reject_packets()
[all …]
Dar-ack.c885 struct kvec iov[5]; in rxrpc_process_call() local
922 memset(iov, 0, sizeof(iov)); in rxrpc_process_call()
923 iov[0].iov_base = &hdr; in rxrpc_process_call()
924 iov[0].iov_len = sizeof(hdr); in rxrpc_process_call()
978 iov[1].iov_base = &data; in rxrpc_process_call()
979 iov[1].iov_len = sizeof(data); in rxrpc_process_call()
1002 iov[1].iov_base = &ack; in rxrpc_process_call()
1003 iov[1].iov_len = sizeof(ack); in rxrpc_process_call()
1004 iov[2].iov_base = &pad; in rxrpc_process_call()
1005 iov[2].iov_len = 3; in rxrpc_process_call()
[all …]
Dar-output.c340 struct kvec iov[1]; in rxrpc_send_packet() local
346 iov[0].iov_base = skb->head; in rxrpc_send_packet()
347 iov[0].iov_len = skb->len; in rxrpc_send_packet()
365 ret = kernel_sendmsg(trans->local->socket, &msg, iov, 1, in rxrpc_send_packet()
366 iov[0].iov_len); in rxrpc_send_packet()
385 ret = kernel_sendmsg(trans->local->socket, &msg, iov, 1, in rxrpc_send_packet()
386 iov[0].iov_len); in rxrpc_send_packet()
534 struct iovec *iov; in rxrpc_send_data() local
550 iov = msg->msg_iov; in rxrpc_send_data()
552 from = iov->iov_base; in rxrpc_send_data()
[all …]
Drxkad.c572 struct kvec iov[2]; in rxkad_issue_challenge() local
606 iov[0].iov_base = &hdr; in rxkad_issue_challenge()
607 iov[0].iov_len = sizeof(hdr); in rxkad_issue_challenge()
608 iov[1].iov_base = &challenge; in rxkad_issue_challenge()
609 iov[1].iov_len = sizeof(challenge); in rxkad_issue_challenge()
611 len = iov[0].iov_len + iov[1].iov_len; in rxkad_issue_challenge()
616 ret = kernel_sendmsg(conn->trans->local->socket, &msg, iov, 2, len); in rxkad_issue_challenge()
635 struct kvec iov[3]; in rxkad_send_response() local
654 iov[0].iov_base = hdr; in rxkad_send_response()
655 iov[0].iov_len = sizeof(*hdr); in rxkad_send_response()
[all …]
Dar-accept.c33 struct kvec iov[1]; in rxrpc_busy() local
51 iov[0].iov_base = hdr; in rxrpc_busy()
52 iov[0].iov_len = sizeof(*hdr); in rxrpc_busy()
54 len = iov[0].iov_len; in rxrpc_busy()
59 ret = kernel_sendmsg(local->socket, &msg, iov, 1, len); in rxrpc_busy()
/net/sunrpc/
Dxdr.c462 struct kvec *iov = buf->head; in xdr_init_encode() local
468 xdr->iov = iov; in xdr_init_encode()
469 xdr->p = (__be32 *)((char *)iov->iov_base + iov->iov_len); in xdr_init_encode()
470 xdr->end = (__be32 *)((char *)iov->iov_base + scratch_len); in xdr_init_encode()
471 BUG_ON(iov->iov_len > scratch_len); in xdr_init_encode()
480 iov->iov_len += len; in xdr_init_encode()
525 if (xdr->iov) in xdr_get_next_encode_buffer()
526 xdr->iov->iov_len += frag1bytes; in xdr_get_next_encode_buffer()
530 xdr->iov = NULL; in xdr_get_next_encode_buffer()
575 if (xdr->iov) in xdr_reserve_space()
[all …]
Dsvcsock.c317 static int svc_recvfrom(struct svc_rqst *rqstp, struct kvec *iov, int nr, in svc_recvfrom() argument
330 len = kernel_recvmsg(svsk->sk_sock, &msg, iov, nr, buflen, in svc_recvfrom()
339 svsk, iov[0].iov_base, iov[0].iov_len, len); in svc_recvfrom()
344 struct kvec *iov, int nr, in svc_partial_recvfrom() argument
353 return svc_recvfrom(rqstp, iov, nr, buflen); in svc_partial_recvfrom()
356 if (iov[i].iov_len > base) in svc_partial_recvfrom()
358 base -= iov[i].iov_len; in svc_partial_recvfrom()
360 save_iovlen = iov[i].iov_len; in svc_partial_recvfrom()
361 save_iovbase = iov[i].iov_base; in svc_partial_recvfrom()
362 iov[i].iov_len -= base; in svc_partial_recvfrom()
[all …]
/net/tipc/
Dserver.c79 struct kvec iov; member
238 struct kvec iov; in tipc_receive_from_sock() local
248 iov.iov_base = buf; in tipc_receive_from_sock()
249 iov.iov_len = s->max_rcvbuf_size; in tipc_receive_from_sock()
251 ret = kernel_recvmsg(con->sock, &msg, &iov, 1, iov.iov_len, in tipc_receive_from_sock()
383 entry->iov.iov_base = buf; in tipc_alloc_entry()
384 entry->iov.iov_len = len; in tipc_alloc_entry()
391 kfree(e->iov.iov_base); in tipc_free_entry()
473 ret = kernel_sendmsg(con->sock, &msg, &e->iov, 1, in tipc_send_to_sock()
474 e->iov.iov_len); in tipc_send_to_sock()
Dmsg.c172 int tipc_msg_build(struct tipc_msg *mhdr, struct iovec const *iov, in tipc_msg_build() argument
197 if (!dsz || !memcpy_fromiovecend(pktpos, iov, offset, dsz)) in tipc_msg_build()
226 if (memcpy_fromiovecend(pktpos, iov, offset, pktrem)) { in tipc_msg_build()
/net/
Dsocket.c117 static ssize_t sock_aio_read(struct kiocb *iocb, const struct iovec *iov,
119 static ssize_t sock_aio_write(struct kiocb *iocb, const struct iovec *iov,
915 struct file *file, const struct iovec *iov, in do_sock_read() argument
923 size += iov[i].iov_len; in do_sock_read()
929 msg->msg_iov = (struct iovec *)iov; in do_sock_read()
936 static ssize_t sock_aio_read(struct kiocb *iocb, const struct iovec *iov, in sock_aio_read() argument
951 return do_sock_read(&x->async_msg, iocb, iocb->ki_filp, iov, nr_segs); in sock_aio_read()
955 struct file *file, const struct iovec *iov, in do_sock_write() argument
963 size += iov[i].iov_len; in do_sock_write()
969 msg->msg_iov = (struct iovec *)iov; in do_sock_write()
[all …]
/net/ipv6/
Draw.c678 struct iovec *iov; in rawv6_probe_proto_opt() local
689 iov = &msg->msg_iov[i]; in rawv6_probe_proto_opt()
690 if (!iov) in rawv6_probe_proto_opt()
696 if (iov->iov_base && iov->iov_len < 1) in rawv6_probe_proto_opt()
700 type = iov->iov_base; in rawv6_probe_proto_opt()
702 if (iov->iov_len > 1) in rawv6_probe_proto_opt()
705 code = iov->iov_base; in rawv6_probe_proto_opt()
715 if (iov->iov_base && iov->iov_len < 1) in rawv6_probe_proto_opt()
718 if (iov->iov_len > 2 - len) { in rawv6_probe_proto_opt()
719 u8 __user *p = iov->iov_base; in rawv6_probe_proto_opt()
[all …]
/net/sunrpc/auth_gss/
Dauth_gss.c1492 struct kvec iov; in gss_marshal() local
1513 iov.iov_base = xprt_skip_transport_header(req->rq_xprt, in gss_marshal()
1515 iov.iov_len = (u8 *)p - (u8 *)iov.iov_base; in gss_marshal()
1516 xdr_buf_from_iov(&iov, &verf_buf); in gss_marshal()
1615 struct kvec iov; in gss_validate() local
1630 iov.iov_base = &seq; in gss_validate()
1631 iov.iov_len = sizeof(seq); in gss_validate()
1632 xdr_buf_from_iov(&iov, &verf_buf); in gss_validate()
1679 struct kvec *iov; in gss_wrap_req_integ() local
1696 iov = snd_buf->tail; in gss_wrap_req_integ()
[all …]
Dgss_krb5_wrap.c53 struct kvec *iov; in gss_krb5_add_padding() local
56 iov = &buf->tail[0]; in gss_krb5_add_padding()
58 iov = &buf->head[0]; in gss_krb5_add_padding()
59 p = iov->iov_base + iov->iov_len; in gss_krb5_add_padding()
60 iov->iov_len += padding; in gss_krb5_add_padding()
/net/sunrpc/xprtrdma/
Dverbs.c1599 struct ib_mr **mrp, struct ib_sge *iov) in rpcrdma_register_internal() argument
1608 iov->addr = ib_dma_map_single(ia->ri_id->device, in rpcrdma_register_internal()
1610 if (ib_dma_mapping_error(ia->ri_id->device, iov->addr)) in rpcrdma_register_internal()
1613 iov->length = len; in rpcrdma_register_internal()
1617 iov->lkey = ia->ri_dma_lkey; in rpcrdma_register_internal()
1621 iov->lkey = ia->ri_bind_mem->lkey; in rpcrdma_register_internal()
1625 ipb.addr = iov->addr; in rpcrdma_register_internal()
1626 ipb.size = iov->length; in rpcrdma_register_internal()
1628 IB_ACCESS_LOCAL_WRITE, &iov->addr); in rpcrdma_register_internal()
1633 (unsigned long long)iov->addr, len); in rpcrdma_register_internal()
[all …]
/net/ipv4/
Draw.c428 struct iovec *iov; in raw_probe_proto_opt() local
438 iov = &msg->msg_iov[i]; in raw_probe_proto_opt()
439 if (!iov) in raw_probe_proto_opt()
445 if (iov->iov_base && iov->iov_len < 1) in raw_probe_proto_opt()
449 type = iov->iov_base; in raw_probe_proto_opt()
451 if (iov->iov_len > 1) in raw_probe_proto_opt()
454 code = iov->iov_base; in raw_probe_proto_opt()
Dtcp_ipv4.c627 arg.iov[0].iov_base = (unsigned char *)&rep; in tcp_v4_send_reset()
628 arg.iov[0].iov_len = sizeof(rep.th); in tcp_v4_send_reset()
669 arg.iov[0].iov_len += TCPOLEN_MD5SIG_ALIGNED; in tcp_v4_send_reset()
670 rep.th.doff = arg.iov[0].iov_len / 4; in tcp_v4_send_reset()
679 arg.iov[0].iov_len, IPPROTO_TCP, 0); in tcp_v4_send_reset()
694 &arg, arg.iov[0].iov_len); in tcp_v4_send_reset()
733 arg.iov[0].iov_base = (unsigned char *)&rep; in tcp_v4_send_ack()
734 arg.iov[0].iov_len = sizeof(rep.th); in tcp_v4_send_ack()
741 arg.iov[0].iov_len += TCPOLEN_TSTAMP_ALIGNED; in tcp_v4_send_ack()
747 rep.th.doff = arg.iov[0].iov_len / 4; in tcp_v4_send_ack()
[all …]
/net/ipx/
Dipx_route.c168 struct iovec *iov, size_t len, int noblock) in ipxrtr_route_packet() argument
232 rc = memcpy_fromiovec(skb_put(skb, len), iov, len); in ipxrtr_route_packet()
/net/netfilter/ipvs/
Dip_vs_sync.c1545 struct kvec iov; in ip_vs_send_async() local
1549 iov.iov_base = (void *)buffer; in ip_vs_send_async()
1550 iov.iov_len = length; in ip_vs_send_async()
1552 len = kernel_sendmsg(sock, &msg, &iov, 1, (size_t)(length)); in ip_vs_send_async()
1577 struct kvec iov; in ip_vs_receive() local
1583 iov.iov_base = buffer; in ip_vs_receive()
1584 iov.iov_len = (size_t)buflen; in ip_vs_receive()
1586 len = kernel_recvmsg(sock, &msg, &iov, 1, buflen, MSG_DONTWAIT); in ip_vs_receive()

12