/net/core/ |
D | iovec.c | 38 int verify_iovec(struct msghdr *m, struct iovec *iov, struct sockaddr_storage *address, int mode) in verify_iovec() argument 58 if (copy_from_user(iov, (void __user __force *) m->msg_iov, size)) in verify_iovec() 61 m->msg_iov = iov; in verify_iovec() 65 size_t len = iov[ct].iov_len; in verify_iovec() 69 iov[ct].iov_len = len; in verify_iovec() 85 int csum_partial_copy_fromiovecend(unsigned char *kdata, struct iovec *iov, in csum_partial_copy_fromiovecend() argument 92 while (offset >= iov->iov_len) { in csum_partial_copy_fromiovecend() 93 offset -= iov->iov_len; in csum_partial_copy_fromiovecend() 94 iov++; in csum_partial_copy_fromiovecend() 98 u8 __user *base = iov->iov_base + offset; in csum_partial_copy_fromiovecend() [all …]
|
D | datagram.c | 830 int hlen, struct iovec *iov) in skb_copy_and_csum_datagram_iovec() argument 841 while (!iov->iov_len) in skb_copy_and_csum_datagram_iovec() 842 iov++; in skb_copy_and_csum_datagram_iovec() 844 if (iov->iov_len < chunk) { in skb_copy_and_csum_datagram_iovec() 847 if (skb_copy_datagram_iovec(skb, hlen, iov, chunk)) in skb_copy_and_csum_datagram_iovec() 851 if (skb_copy_and_csum_datagram(skb, hlen, iov->iov_base, in skb_copy_and_csum_datagram_iovec() 858 iov->iov_len -= chunk; in skb_copy_and_csum_datagram_iovec() 859 iov->iov_base += chunk; in skb_copy_and_csum_datagram_iovec()
|
/net/rds/ |
D | message.c | 273 struct iovec *iov; in rds_message_copy_from_user() local 283 iov = first_iov; in rds_message_copy_from_user() 297 while (iov_off == iov->iov_len) { in rds_message_copy_from_user() 299 iov++; in rds_message_copy_from_user() 302 to_copy = min(iov->iov_len - iov_off, sg->length - sg_off); in rds_message_copy_from_user() 307 to_copy, iov->iov_base, iov->iov_len, iov_off, in rds_message_copy_from_user() 311 iov->iov_base + iov_off, in rds_message_copy_from_user() 332 struct iovec *iov; in rds_message_inc_copy_to_user() local 344 iov = first_iov; in rds_message_inc_copy_to_user() 351 while (iov_off == iov->iov_len) { in rds_message_inc_copy_to_user() [all …]
|
D | rdma.c | 484 static int rds_rdma_pages(struct rds_iovec iov[], int nr_iovecs) in rds_rdma_pages() argument 492 nr_pages = rds_pages_in_vec(&iov[i]); in rds_rdma_pages() 661 struct rds_iovec *iov = &iovs[i]; in rds_cmsg_rdma_args() local 663 unsigned int nr = rds_pages_in_vec(iov); in rds_cmsg_rdma_args() 665 rs->rs_user_addr = iov->addr; in rds_cmsg_rdma_args() 666 rs->rs_user_bytes = iov->bytes; in rds_cmsg_rdma_args() 671 ret = rds_pin_pages(iov->addr, nr, pages, !op->op_write); in rds_cmsg_rdma_args() 676 nr_bytes, nr, iov->bytes, iov->addr); in rds_cmsg_rdma_args() 678 nr_bytes += iov->bytes; in rds_cmsg_rdma_args() 681 unsigned int offset = iov->addr & ~PAGE_MASK; in rds_cmsg_rdma_args() [all …]
|
D | tcp_recv.c | 66 struct iovec *iov, tmp; in rds_tcp_inc_copy_to_user() local 75 iov = first_iov; in rds_tcp_inc_copy_to_user() 76 tmp = *iov; in rds_tcp_inc_copy_to_user() 82 iov++; in rds_tcp_inc_copy_to_user() 83 tmp = *iov; in rds_tcp_inc_copy_to_user()
|
D | iw_recv.c | 311 struct iovec *iov = first_iov; in rds_iw_inc_copy_to_user() local 329 while (iov_off == iov->iov_len) { in rds_iw_inc_copy_to_user() 331 iov++; in rds_iw_inc_copy_to_user() 334 to_copy = min(iov->iov_len - iov_off, RDS_FRAG_SIZE - frag_off); in rds_iw_inc_copy_to_user() 340 to_copy, iov->iov_base, iov->iov_len, iov_off, in rds_iw_inc_copy_to_user() 346 iov->iov_base + iov_off, in rds_iw_inc_copy_to_user()
|
D | ib_recv.c | 480 struct iovec *iov = first_iov; in rds_ib_inc_copy_to_user() local 498 while (iov_off == iov->iov_len) { in rds_ib_inc_copy_to_user() 500 iov++; in rds_ib_inc_copy_to_user() 503 to_copy = min(iov->iov_len - iov_off, RDS_FRAG_SIZE - frag_off); in rds_ib_inc_copy_to_user() 509 to_copy, iov->iov_base, iov->iov_len, iov_off, in rds_ib_inc_copy_to_user() 515 iov->iov_base + iov_off, in rds_ib_inc_copy_to_user()
|
/net/rxrpc/ |
D | ar-connevent.c | 65 struct kvec iov[2]; in rxrpc_abort_connection() local 105 iov[0].iov_base = &hdr; in rxrpc_abort_connection() 106 iov[0].iov_len = sizeof(hdr); in rxrpc_abort_connection() 107 iov[1].iov_base = &word; in rxrpc_abort_connection() 108 iov[1].iov_len = sizeof(word); in rxrpc_abort_connection() 110 len = iov[0].iov_len + iov[1].iov_len; in rxrpc_abort_connection() 115 ret = kernel_sendmsg(conn->trans->local->socket, &msg, iov, 2, len); in rxrpc_abort_connection() 342 struct kvec iov[2]; in rxrpc_reject_packets() local 351 iov[0].iov_base = &hdr; in rxrpc_reject_packets() 352 iov[0].iov_len = sizeof(hdr); in rxrpc_reject_packets() [all …]
|
D | ar-ack.c | 885 struct kvec iov[5]; in rxrpc_process_call() local 922 memset(iov, 0, sizeof(iov)); in rxrpc_process_call() 923 iov[0].iov_base = &hdr; in rxrpc_process_call() 924 iov[0].iov_len = sizeof(hdr); in rxrpc_process_call() 978 iov[1].iov_base = &data; in rxrpc_process_call() 979 iov[1].iov_len = sizeof(data); in rxrpc_process_call() 1002 iov[1].iov_base = &ack; in rxrpc_process_call() 1003 iov[1].iov_len = sizeof(ack); in rxrpc_process_call() 1004 iov[2].iov_base = &pad; in rxrpc_process_call() 1005 iov[2].iov_len = 3; in rxrpc_process_call() [all …]
|
D | ar-output.c | 340 struct kvec iov[1]; in rxrpc_send_packet() local 346 iov[0].iov_base = skb->head; in rxrpc_send_packet() 347 iov[0].iov_len = skb->len; in rxrpc_send_packet() 365 ret = kernel_sendmsg(trans->local->socket, &msg, iov, 1, in rxrpc_send_packet() 366 iov[0].iov_len); in rxrpc_send_packet() 385 ret = kernel_sendmsg(trans->local->socket, &msg, iov, 1, in rxrpc_send_packet() 386 iov[0].iov_len); in rxrpc_send_packet() 534 struct iovec *iov; in rxrpc_send_data() local 550 iov = msg->msg_iov; in rxrpc_send_data() 552 from = iov->iov_base; in rxrpc_send_data() [all …]
|
D | rxkad.c | 572 struct kvec iov[2]; in rxkad_issue_challenge() local 606 iov[0].iov_base = &hdr; in rxkad_issue_challenge() 607 iov[0].iov_len = sizeof(hdr); in rxkad_issue_challenge() 608 iov[1].iov_base = &challenge; in rxkad_issue_challenge() 609 iov[1].iov_len = sizeof(challenge); in rxkad_issue_challenge() 611 len = iov[0].iov_len + iov[1].iov_len; in rxkad_issue_challenge() 616 ret = kernel_sendmsg(conn->trans->local->socket, &msg, iov, 2, len); in rxkad_issue_challenge() 635 struct kvec iov[3]; in rxkad_send_response() local 654 iov[0].iov_base = hdr; in rxkad_send_response() 655 iov[0].iov_len = sizeof(*hdr); in rxkad_send_response() [all …]
|
D | ar-accept.c | 33 struct kvec iov[1]; in rxrpc_busy() local 51 iov[0].iov_base = hdr; in rxrpc_busy() 52 iov[0].iov_len = sizeof(*hdr); in rxrpc_busy() 54 len = iov[0].iov_len; in rxrpc_busy() 59 ret = kernel_sendmsg(local->socket, &msg, iov, 1, len); in rxrpc_busy()
|
/net/sunrpc/ |
D | xdr.c | 462 struct kvec *iov = buf->head; in xdr_init_encode() local 468 xdr->iov = iov; in xdr_init_encode() 469 xdr->p = (__be32 *)((char *)iov->iov_base + iov->iov_len); in xdr_init_encode() 470 xdr->end = (__be32 *)((char *)iov->iov_base + scratch_len); in xdr_init_encode() 471 BUG_ON(iov->iov_len > scratch_len); in xdr_init_encode() 480 iov->iov_len += len; in xdr_init_encode() 525 if (xdr->iov) in xdr_get_next_encode_buffer() 526 xdr->iov->iov_len += frag1bytes; in xdr_get_next_encode_buffer() 530 xdr->iov = NULL; in xdr_get_next_encode_buffer() 575 if (xdr->iov) in xdr_reserve_space() [all …]
|
D | svcsock.c | 317 static int svc_recvfrom(struct svc_rqst *rqstp, struct kvec *iov, int nr, in svc_recvfrom() argument 330 len = kernel_recvmsg(svsk->sk_sock, &msg, iov, nr, buflen, in svc_recvfrom() 339 svsk, iov[0].iov_base, iov[0].iov_len, len); in svc_recvfrom() 344 struct kvec *iov, int nr, in svc_partial_recvfrom() argument 353 return svc_recvfrom(rqstp, iov, nr, buflen); in svc_partial_recvfrom() 356 if (iov[i].iov_len > base) in svc_partial_recvfrom() 358 base -= iov[i].iov_len; in svc_partial_recvfrom() 360 save_iovlen = iov[i].iov_len; in svc_partial_recvfrom() 361 save_iovbase = iov[i].iov_base; in svc_partial_recvfrom() 362 iov[i].iov_len -= base; in svc_partial_recvfrom() [all …]
|
/net/tipc/ |
D | server.c | 79 struct kvec iov; member 238 struct kvec iov; in tipc_receive_from_sock() local 248 iov.iov_base = buf; in tipc_receive_from_sock() 249 iov.iov_len = s->max_rcvbuf_size; in tipc_receive_from_sock() 251 ret = kernel_recvmsg(con->sock, &msg, &iov, 1, iov.iov_len, in tipc_receive_from_sock() 383 entry->iov.iov_base = buf; in tipc_alloc_entry() 384 entry->iov.iov_len = len; in tipc_alloc_entry() 391 kfree(e->iov.iov_base); in tipc_free_entry() 473 ret = kernel_sendmsg(con->sock, &msg, &e->iov, 1, in tipc_send_to_sock() 474 e->iov.iov_len); in tipc_send_to_sock()
|
D | msg.c | 172 int tipc_msg_build(struct tipc_msg *mhdr, struct iovec const *iov, in tipc_msg_build() argument 197 if (!dsz || !memcpy_fromiovecend(pktpos, iov, offset, dsz)) in tipc_msg_build() 226 if (memcpy_fromiovecend(pktpos, iov, offset, pktrem)) { in tipc_msg_build()
|
/net/ |
D | socket.c | 117 static ssize_t sock_aio_read(struct kiocb *iocb, const struct iovec *iov, 119 static ssize_t sock_aio_write(struct kiocb *iocb, const struct iovec *iov, 915 struct file *file, const struct iovec *iov, in do_sock_read() argument 923 size += iov[i].iov_len; in do_sock_read() 929 msg->msg_iov = (struct iovec *)iov; in do_sock_read() 936 static ssize_t sock_aio_read(struct kiocb *iocb, const struct iovec *iov, in sock_aio_read() argument 951 return do_sock_read(&x->async_msg, iocb, iocb->ki_filp, iov, nr_segs); in sock_aio_read() 955 struct file *file, const struct iovec *iov, in do_sock_write() argument 963 size += iov[i].iov_len; in do_sock_write() 969 msg->msg_iov = (struct iovec *)iov; in do_sock_write() [all …]
|
/net/ipv6/ |
D | raw.c | 678 struct iovec *iov; in rawv6_probe_proto_opt() local 689 iov = &msg->msg_iov[i]; in rawv6_probe_proto_opt() 690 if (!iov) in rawv6_probe_proto_opt() 696 if (iov->iov_base && iov->iov_len < 1) in rawv6_probe_proto_opt() 700 type = iov->iov_base; in rawv6_probe_proto_opt() 702 if (iov->iov_len > 1) in rawv6_probe_proto_opt() 705 code = iov->iov_base; in rawv6_probe_proto_opt() 715 if (iov->iov_base && iov->iov_len < 1) in rawv6_probe_proto_opt() 718 if (iov->iov_len > 2 - len) { in rawv6_probe_proto_opt() 719 u8 __user *p = iov->iov_base; in rawv6_probe_proto_opt() [all …]
|
/net/sunrpc/auth_gss/ |
D | auth_gss.c | 1492 struct kvec iov; in gss_marshal() local 1513 iov.iov_base = xprt_skip_transport_header(req->rq_xprt, in gss_marshal() 1515 iov.iov_len = (u8 *)p - (u8 *)iov.iov_base; in gss_marshal() 1516 xdr_buf_from_iov(&iov, &verf_buf); in gss_marshal() 1615 struct kvec iov; in gss_validate() local 1630 iov.iov_base = &seq; in gss_validate() 1631 iov.iov_len = sizeof(seq); in gss_validate() 1632 xdr_buf_from_iov(&iov, &verf_buf); in gss_validate() 1679 struct kvec *iov; in gss_wrap_req_integ() local 1696 iov = snd_buf->tail; in gss_wrap_req_integ() [all …]
|
D | gss_krb5_wrap.c | 53 struct kvec *iov; in gss_krb5_add_padding() local 56 iov = &buf->tail[0]; in gss_krb5_add_padding() 58 iov = &buf->head[0]; in gss_krb5_add_padding() 59 p = iov->iov_base + iov->iov_len; in gss_krb5_add_padding() 60 iov->iov_len += padding; in gss_krb5_add_padding()
|
/net/sunrpc/xprtrdma/ |
D | verbs.c | 1599 struct ib_mr **mrp, struct ib_sge *iov) in rpcrdma_register_internal() argument 1608 iov->addr = ib_dma_map_single(ia->ri_id->device, in rpcrdma_register_internal() 1610 if (ib_dma_mapping_error(ia->ri_id->device, iov->addr)) in rpcrdma_register_internal() 1613 iov->length = len; in rpcrdma_register_internal() 1617 iov->lkey = ia->ri_dma_lkey; in rpcrdma_register_internal() 1621 iov->lkey = ia->ri_bind_mem->lkey; in rpcrdma_register_internal() 1625 ipb.addr = iov->addr; in rpcrdma_register_internal() 1626 ipb.size = iov->length; in rpcrdma_register_internal() 1628 IB_ACCESS_LOCAL_WRITE, &iov->addr); in rpcrdma_register_internal() 1633 (unsigned long long)iov->addr, len); in rpcrdma_register_internal() [all …]
|
/net/ipv4/ |
D | raw.c | 428 struct iovec *iov; in raw_probe_proto_opt() local 438 iov = &msg->msg_iov[i]; in raw_probe_proto_opt() 439 if (!iov) in raw_probe_proto_opt() 445 if (iov->iov_base && iov->iov_len < 1) in raw_probe_proto_opt() 449 type = iov->iov_base; in raw_probe_proto_opt() 451 if (iov->iov_len > 1) in raw_probe_proto_opt() 454 code = iov->iov_base; in raw_probe_proto_opt()
|
D | tcp_ipv4.c | 627 arg.iov[0].iov_base = (unsigned char *)&rep; in tcp_v4_send_reset() 628 arg.iov[0].iov_len = sizeof(rep.th); in tcp_v4_send_reset() 669 arg.iov[0].iov_len += TCPOLEN_MD5SIG_ALIGNED; in tcp_v4_send_reset() 670 rep.th.doff = arg.iov[0].iov_len / 4; in tcp_v4_send_reset() 679 arg.iov[0].iov_len, IPPROTO_TCP, 0); in tcp_v4_send_reset() 694 &arg, arg.iov[0].iov_len); in tcp_v4_send_reset() 733 arg.iov[0].iov_base = (unsigned char *)&rep; in tcp_v4_send_ack() 734 arg.iov[0].iov_len = sizeof(rep.th); in tcp_v4_send_ack() 741 arg.iov[0].iov_len += TCPOLEN_TSTAMP_ALIGNED; in tcp_v4_send_ack() 747 rep.th.doff = arg.iov[0].iov_len / 4; in tcp_v4_send_ack() [all …]
|
/net/ipx/ |
D | ipx_route.c | 168 struct iovec *iov, size_t len, int noblock) in ipxrtr_route_packet() argument 232 rc = memcpy_fromiovec(skb_put(skb, len), iov, len); in ipxrtr_route_packet()
|
/net/netfilter/ipvs/ |
D | ip_vs_sync.c | 1545 struct kvec iov; in ip_vs_send_async() local 1549 iov.iov_base = (void *)buffer; in ip_vs_send_async() 1550 iov.iov_len = length; in ip_vs_send_async() 1552 len = kernel_sendmsg(sock, &msg, &iov, 1, (size_t)(length)); in ip_vs_send_async() 1577 struct kvec iov; in ip_vs_receive() local 1583 iov.iov_base = buffer; in ip_vs_receive() 1584 iov.iov_len = (size_t)buflen; in ip_vs_receive() 1586 len = kernel_recvmsg(sock, &msg, &iov, 1, buflen, MSG_DONTWAIT); in ip_vs_receive()
|