• Home
  • Raw
  • Download

Lines Matching +full:default +full:- +full:tap

1 // SPDX-License-Identifier: GPL-2.0-only
35 return q->flags & TAP_VNET_BE ? false : in tap_legacy_is_little_endian()
41 int s = !!(q->flags & TAP_VNET_BE); in tap_get_vnet_be()
44 return -EFAULT; in tap_get_vnet_be()
54 return -EFAULT; in tap_set_vnet_be()
57 q->flags |= TAP_VNET_BE; in tap_set_vnet_be()
59 q->flags &= ~TAP_VNET_BE; in tap_set_vnet_be()
71 return -EINVAL; in tap_get_vnet_be()
76 return -EINVAL; in tap_set_vnet_be()
82 return q->flags & TAP_VNET_LE || in tap_is_little_endian()
97 .name = "tap",
124 return rcu_dereference(dev->rx_handler_data); in tap_dev_get_rcu()
134 * through sock_hold(&q->sk). When the macvlan_dev goes away first,
135 * q->vlan becomes inaccessible. When the files gets closed,
144 static int tap_enable_queue(struct tap_dev *tap, struct file *file, in tap_enable_queue() argument
147 int err = -EINVAL; in tap_enable_queue()
151 if (q->enabled) in tap_enable_queue()
155 rcu_assign_pointer(tap->taps[tap->numvtaps], q); in tap_enable_queue()
156 q->queue_index = tap->numvtaps; in tap_enable_queue()
157 q->enabled = true; in tap_enable_queue()
159 tap->numvtaps++; in tap_enable_queue()
165 static int tap_set_queue(struct tap_dev *tap, struct file *file, in tap_set_queue() argument
168 if (tap->numqueues == MAX_TAP_QUEUES) in tap_set_queue()
169 return -EBUSY; in tap_set_queue()
171 rcu_assign_pointer(q->tap, tap); in tap_set_queue()
172 rcu_assign_pointer(tap->taps[tap->numvtaps], q); in tap_set_queue()
173 sock_hold(&q->sk); in tap_set_queue()
175 q->file = file; in tap_set_queue()
176 q->queue_index = tap->numvtaps; in tap_set_queue()
177 q->enabled = true; in tap_set_queue()
178 file->private_data = q; in tap_set_queue()
179 list_add_tail(&q->next, &tap->queue_list); in tap_set_queue()
181 tap->numvtaps++; in tap_set_queue()
182 tap->numqueues++; in tap_set_queue()
189 struct tap_dev *tap; in tap_disable_queue() local
193 if (!q->enabled) in tap_disable_queue()
194 return -EINVAL; in tap_disable_queue()
196 tap = rtnl_dereference(q->tap); in tap_disable_queue()
198 if (tap) { in tap_disable_queue()
199 int index = q->queue_index; in tap_disable_queue()
200 BUG_ON(index >= tap->numvtaps); in tap_disable_queue()
201 nq = rtnl_dereference(tap->taps[tap->numvtaps - 1]); in tap_disable_queue()
202 nq->queue_index = index; in tap_disable_queue()
204 rcu_assign_pointer(tap->taps[index], nq); in tap_disable_queue()
205 RCU_INIT_POINTER(tap->taps[tap->numvtaps - 1], NULL); in tap_disable_queue()
206 q->enabled = false; in tap_disable_queue()
208 tap->numvtaps--; in tap_disable_queue()
224 struct tap_dev *tap; in tap_put_queue() local
227 tap = rtnl_dereference(q->tap); in tap_put_queue()
229 if (tap) { in tap_put_queue()
230 if (q->enabled) in tap_put_queue()
233 tap->numqueues--; in tap_put_queue()
234 RCU_INIT_POINTER(q->tap, NULL); in tap_put_queue()
235 sock_put(&q->sk); in tap_put_queue()
236 list_del_init(&q->next); in tap_put_queue()
242 sock_put(&q->sk); in tap_put_queue()
249 * Cache vlan->numvtaps since it can become zero during the execution
252 static struct tap_queue *tap_get_queue(struct tap_dev *tap, in tap_get_queue() argument
258 * and validate that the result isn't NULL - in case we are in tap_get_queue()
261 int numvtaps = READ_ONCE(tap->numvtaps); in tap_get_queue()
273 queue = rcu_dereference(tap->taps[rxq % numvtaps]); in tap_get_queue()
281 rxq -= numvtaps; in tap_get_queue()
283 queue = rcu_dereference(tap->taps[rxq]); in tap_get_queue()
288 queue = rcu_dereference(tap->taps[0]); in tap_get_queue()
298 void tap_del_queues(struct tap_dev *tap) in tap_del_queues() argument
303 list_for_each_entry_safe(q, tmp, &tap->queue_list, next) { in tap_del_queues()
304 list_del_init(&q->next); in tap_del_queues()
305 RCU_INIT_POINTER(q->tap, NULL); in tap_del_queues()
306 if (q->enabled) in tap_del_queues()
307 tap->numvtaps--; in tap_del_queues()
308 tap->numqueues--; in tap_del_queues()
309 sock_put(&q->sk); in tap_del_queues()
311 BUG_ON(tap->numvtaps); in tap_del_queues()
312 BUG_ON(tap->numqueues); in tap_del_queues()
314 tap->numvtaps = MAX_TAP_QUEUES; in tap_del_queues()
321 struct net_device *dev = skb->dev; in tap_handle_frame()
322 struct tap_dev *tap; in tap_handle_frame() local
326 tap = tap_dev_get_rcu(dev); in tap_handle_frame()
327 if (!tap) in tap_handle_frame()
330 q = tap_get_queue(tap, skb); in tap_handle_frame()
340 if (q->flags & IFF_VNET_HDR) in tap_handle_frame()
341 features |= tap->tap_features; in tap_handle_frame()
350 if (ptr_ring_produce(&q->ring, skb)) in tap_handle_frame()
358 if (ptr_ring_produce(&q->ring, skb)) { in tap_handle_frame()
365 /* If we receive a partial checksum and the tap side in tap_handle_frame()
370 if (skb->ip_summed == CHECKSUM_PARTIAL && in tap_handle_frame()
374 if (ptr_ring_produce(&q->ring, skb)) in tap_handle_frame()
379 wake_up_interruptible_poll(sk_sleep(&q->sk), EPOLLIN | EPOLLRDNORM | EPOLLRDBAND); in tap_handle_frame()
384 if (tap->count_rx_dropped) in tap_handle_frame()
385 tap->count_rx_dropped(tap); in tap_handle_frame()
396 if (tap_major->major == major) in tap_get_major()
403 int tap_get_minor(dev_t major, struct tap_dev *tap) in tap_get_minor() argument
405 int retval = -ENOMEM; in tap_get_minor()
411 retval = -EINVAL; in tap_get_minor()
415 spin_lock(&tap_major->minor_lock); in tap_get_minor()
416 retval = idr_alloc(&tap_major->minor_idr, tap, 1, TAP_NUM_DEVS, GFP_ATOMIC); in tap_get_minor()
418 tap->minor = retval; in tap_get_minor()
419 } else if (retval == -ENOSPC) { in tap_get_minor()
420 netdev_err(tap->dev, "Too many tap devices\n"); in tap_get_minor()
421 retval = -EINVAL; in tap_get_minor()
423 spin_unlock(&tap_major->minor_lock); in tap_get_minor()
431 void tap_free_minor(dev_t major, struct tap_dev *tap) in tap_free_minor() argument
441 spin_lock(&tap_major->minor_lock); in tap_free_minor()
442 if (tap->minor) { in tap_free_minor()
443 idr_remove(&tap_major->minor_idr, tap->minor); in tap_free_minor()
444 tap->minor = 0; in tap_free_minor()
446 spin_unlock(&tap_major->minor_lock); in tap_free_minor()
456 struct tap_dev *tap; in dev_get_by_tap_file() local
462 tap = NULL; in dev_get_by_tap_file()
466 spin_lock(&tap_major->minor_lock); in dev_get_by_tap_file()
467 tap = idr_find(&tap_major->minor_idr, minor); in dev_get_by_tap_file()
468 if (tap) { in dev_get_by_tap_file()
469 dev = tap->dev; in dev_get_by_tap_file()
472 spin_unlock(&tap_major->minor_lock); in dev_get_by_tap_file()
476 return tap; in dev_get_by_tap_file()
484 !test_and_clear_bit(SOCKWQ_ASYNC_NOSPACE, &sk->sk_socket->flags)) in tap_sock_write_space()
496 ptr_ring_cleanup(&q->ring, __skb_array_destroy_skb); in tap_sock_destruct()
501 struct net *net = current->nsproxy->net_ns; in tap_open()
502 struct tap_dev *tap; in tap_open() local
504 int err = -ENODEV; in tap_open()
507 tap = dev_get_by_tap_file(imajor(inode), iminor(inode)); in tap_open()
508 if (!tap) in tap_open()
511 err = -ENOMEM; in tap_open()
516 if (ptr_ring_init(&q->ring, tap->dev->tx_queue_len, GFP_KERNEL)) { in tap_open()
517 sk_free(&q->sk); in tap_open()
521 init_waitqueue_head(&q->sock.wq.wait); in tap_open()
522 q->sock.type = SOCK_RAW; in tap_open()
523 q->sock.state = SS_CONNECTED; in tap_open()
524 q->sock.file = file; in tap_open()
525 q->sock.ops = &tap_socket_ops; in tap_open()
526 sock_init_data_uid(&q->sock, &q->sk, current_fsuid()); in tap_open()
527 q->sk.sk_write_space = tap_sock_write_space; in tap_open()
528 q->sk.sk_destruct = tap_sock_destruct; in tap_open()
529 q->flags = IFF_VNET_HDR | IFF_NO_PI | IFF_TAP; in tap_open()
530 q->vnet_hdr_sz = sizeof(struct virtio_net_hdr); in tap_open()
533 * so far only KVM virtio_net uses tap, enable zero copy between in tap_open()
539 if ((tap->dev->features & NETIF_F_HIGHDMA) && (tap->dev->features & NETIF_F_SG)) in tap_open()
540 sock_set_flag(&q->sk, SOCK_ZEROCOPY); in tap_open()
542 err = tap_set_queue(tap, file, q); in tap_open()
548 dev_put(tap->dev); in tap_open()
554 sock_put(&q->sk); in tap_open()
556 if (tap) in tap_open()
557 dev_put(tap->dev); in tap_open()
565 struct tap_queue *q = file->private_data; in tap_release()
572 struct tap_queue *q = file->private_data; in tap_poll()
579 poll_wait(file, &q->sock.wq.wait, wait); in tap_poll()
581 if (!ptr_ring_empty(&q->ring)) in tap_poll()
584 if (sock_writeable(&q->sk) || in tap_poll()
585 (!test_and_set_bit(SOCKWQ_ASYNC_NOSPACE, &q->sock.flags) && in tap_poll()
586 sock_writeable(&q->sk))) in tap_poll()
603 skb = sock_alloc_send_pskb(sk, prepad + linear, len - linear, noblock, in tap_alloc_skb()
610 skb->data_len = len - linear; in tap_alloc_skb()
611 skb->len += len - linear; in tap_alloc_skb()
625 struct tap_dev *tap; in tap_get_user() local
636 if (q->flags & IFF_VNET_HDR) { in tap_get_user()
637 vnet_hdr_len = READ_ONCE(q->vnet_hdr_sz); in tap_get_user()
639 err = -EINVAL; in tap_get_user()
642 len -= vnet_hdr_len; in tap_get_user()
644 err = -EFAULT; in tap_get_user()
647 iov_iter_advance(from, vnet_hdr_len - sizeof(vnet_hdr)); in tap_get_user()
655 err = -EINVAL; in tap_get_user()
660 err = -EINVAL; in tap_get_user()
664 if (msg_control && sock_flag(&q->sk, SOCK_ZEROCOPY)) { in tap_get_user()
689 skb = tap_alloc_skb(&q->sk, TAP_RESERVE, copylen, in tap_get_user()
704 skb->protocol = eth_hdr(skb)->h_proto; in tap_get_user()
716 if (eth_type_vlan(skb->protocol) && in tap_get_user()
717 vlan_get_protocol_and_depth(skb, skb->protocol, &depth) != 0) in tap_get_user()
721 tap = rcu_dereference(q->tap); in tap_get_user()
724 skb_shinfo(skb)->destructor_arg = msg_control; in tap_get_user()
725 skb_shinfo(skb)->tx_flags |= SKBTX_DEV_ZEROCOPY; in tap_get_user()
726 skb_shinfo(skb)->tx_flags |= SKBTX_SHARED_FRAG; in tap_get_user()
729 uarg->callback(uarg, false); in tap_get_user()
732 if (tap) { in tap_get_user()
733 skb->dev = tap->dev; in tap_get_user()
747 tap = rcu_dereference(q->tap); in tap_get_user()
748 if (tap && tap->count_tx_dropped) in tap_get_user()
749 tap->count_tx_dropped(tap); in tap_get_user()
757 struct file *file = iocb->ki_filp; in tap_write_iter()
758 struct tap_queue *q = file->private_data; in tap_write_iter()
760 return tap_get_user(q, NULL, from, file->f_flags & O_NONBLOCK); in tap_write_iter()
773 if (q->flags & IFF_VNET_HDR) { in tap_put_user()
777 vnet_hdr_len = READ_ONCE(q->vnet_hdr_sz); in tap_put_user()
779 return -EINVAL; in tap_put_user()
788 return -EFAULT; in tap_put_user()
790 iov_iter_advance(iter, vnet_hdr_len - sizeof(vnet_hdr)); in tap_put_user()
793 total += skb->len; in tap_put_user()
800 veth.h_vlan_proto = skb->vlan_proto; in tap_put_user()
816 skb->len - vlan_offset); in tap_put_user()
839 prepare_to_wait(sk_sleep(&q->sk), &wait, in tap_do_read()
843 skb = ptr_ring_consume(&q->ring); in tap_do_read()
847 ret = -EAGAIN; in tap_do_read()
851 ret = -ERESTARTSYS; in tap_do_read()
858 finish_wait(sk_sleep(&q->sk), &wait); in tap_do_read()
873 struct file *file = iocb->ki_filp; in tap_read_iter()
874 struct tap_queue *q = file->private_data; in tap_read_iter()
877 ret = tap_do_read(q, to, file->f_flags & O_NONBLOCK, NULL); in tap_read_iter()
880 iocb->ki_pos = ret; in tap_read_iter()
886 struct tap_dev *tap; in tap_get_tap_dev() local
889 tap = rtnl_dereference(q->tap); in tap_get_tap_dev()
890 if (tap) in tap_get_tap_dev()
891 dev_hold(tap->dev); in tap_get_tap_dev()
893 return tap; in tap_get_tap_dev()
896 static void tap_put_tap_dev(struct tap_dev *tap) in tap_put_tap_dev() argument
898 dev_put(tap->dev); in tap_put_tap_dev()
903 struct tap_queue *q = file->private_data; in tap_ioctl_set_queue()
904 struct tap_dev *tap; in tap_ioctl_set_queue() local
907 tap = tap_get_tap_dev(q); in tap_ioctl_set_queue()
908 if (!tap) in tap_ioctl_set_queue()
909 return -EINVAL; in tap_ioctl_set_queue()
912 ret = tap_enable_queue(tap, file, q); in tap_ioctl_set_queue()
916 ret = -EINVAL; in tap_ioctl_set_queue()
918 tap_put_tap_dev(tap); in tap_ioctl_set_queue()
924 struct tap_dev *tap; in set_offload() local
928 tap = rtnl_dereference(q->tap); in set_offload()
929 if (!tap) in set_offload()
930 return -ENOLINK; in set_offload()
932 features = tap->dev->features; in set_offload()
947 /* tun/tap driver inverts the usage for TSO offloads, where in set_offload()
951 * For tap, we have to invert it to mean the same thing. in set_offload()
953 * user-space will not receive TSO frames. in set_offload()
960 /* tap_features are the same as features on tun/tap and in set_offload()
963 tap->tap_features = feature_mask; in set_offload()
964 if (tap->update_features) in set_offload()
965 tap->update_features(tap, features); in set_offload()
971 * provide compatibility with generic tun/tap interface
976 struct tap_queue *q = file->private_data; in tap_ioctl()
977 struct tap_dev *tap; in tap_ioctl() local
990 if (get_user(u, &ifr->ifr_flags)) in tap_ioctl()
991 return -EFAULT; in tap_ioctl()
995 ret = -EINVAL; in tap_ioctl()
997 q->flags = (q->flags & ~TAP_IFFEATURES) | u; in tap_ioctl()
1003 tap = tap_get_tap_dev(q); in tap_ioctl()
1004 if (!tap) { in tap_ioctl()
1006 return -ENOLINK; in tap_ioctl()
1010 u = q->flags; in tap_ioctl()
1011 if (copy_to_user(&ifr->ifr_name, tap->dev->name, IFNAMSIZ) || in tap_ioctl()
1012 put_user(u, &ifr->ifr_flags)) in tap_ioctl()
1013 ret = -EFAULT; in tap_ioctl()
1014 tap_put_tap_dev(tap); in tap_ioctl()
1019 if (get_user(u, &ifr->ifr_flags)) in tap_ioctl()
1020 return -EFAULT; in tap_ioctl()
1028 return -EFAULT; in tap_ioctl()
1033 return -EFAULT; in tap_ioctl()
1035 return -EINVAL; in tap_ioctl()
1037 q->sk.sk_sndbuf = s; in tap_ioctl()
1041 s = q->vnet_hdr_sz; in tap_ioctl()
1043 return -EFAULT; in tap_ioctl()
1048 return -EFAULT; in tap_ioctl()
1050 return -EINVAL; in tap_ioctl()
1052 q->vnet_hdr_sz = s; in tap_ioctl()
1056 s = !!(q->flags & TAP_VNET_LE); in tap_ioctl()
1058 return -EFAULT; in tap_ioctl()
1063 return -EFAULT; in tap_ioctl()
1065 q->flags |= TAP_VNET_LE; in tap_ioctl()
1067 q->flags &= ~TAP_VNET_LE; in tap_ioctl()
1080 return -EINVAL; in tap_ioctl()
1089 tap = tap_get_tap_dev(q); in tap_ioctl()
1090 if (!tap) { in tap_ioctl()
1092 return -ENOLINK; in tap_ioctl()
1095 dev_get_mac_address(&sa, dev_net(tap->dev), tap->dev->name); in tap_ioctl()
1096 if (copy_to_user(&ifr->ifr_name, tap->dev->name, IFNAMSIZ) || in tap_ioctl()
1097 copy_to_user(&ifr->ifr_hwaddr, &sa, sizeof(sa))) in tap_ioctl()
1098 ret = -EFAULT; in tap_ioctl()
1099 tap_put_tap_dev(tap); in tap_ioctl()
1104 if (copy_from_user(&sa, &ifr->ifr_hwaddr, sizeof(sa))) in tap_ioctl()
1105 return -EFAULT; in tap_ioctl()
1107 tap = tap_get_tap_dev(q); in tap_ioctl()
1108 if (!tap) { in tap_ioctl()
1110 return -ENOLINK; in tap_ioctl()
1112 ret = dev_set_mac_address_user(tap->dev, &sa, NULL); in tap_ioctl()
1113 tap_put_tap_dev(tap); in tap_ioctl()
1117 default: in tap_ioctl()
1118 return -EINVAL; in tap_ioctl()
1136 struct tun_xdp_hdr *hdr = xdp->data_hard_start; in tap_get_user_xdp()
1137 struct virtio_net_hdr *gso = &hdr->gso; in tap_get_user_xdp()
1138 int buflen = hdr->buflen; in tap_get_user_xdp()
1140 struct tap_dev *tap; in tap_get_user_xdp() local
1144 if (unlikely(xdp->data_end - xdp->data < ETH_HLEN)) { in tap_get_user_xdp()
1145 err = -EINVAL; in tap_get_user_xdp()
1149 if (q->flags & IFF_VNET_HDR) in tap_get_user_xdp()
1150 vnet_hdr_len = READ_ONCE(q->vnet_hdr_sz); in tap_get_user_xdp()
1152 skb = build_skb(xdp->data_hard_start, buflen); in tap_get_user_xdp()
1154 err = -ENOMEM; in tap_get_user_xdp()
1158 skb_reserve(skb, xdp->data - xdp->data_hard_start); in tap_get_user_xdp()
1159 skb_put(skb, xdp->data_end - xdp->data); in tap_get_user_xdp()
1163 skb->protocol = eth_hdr(skb)->h_proto; in tap_get_user_xdp()
1172 if (eth_type_vlan(skb->protocol) && in tap_get_user_xdp()
1173 vlan_get_protocol_and_depth(skb, skb->protocol, &depth) != 0) in tap_get_user_xdp()
1177 tap = rcu_dereference(q->tap); in tap_get_user_xdp()
1178 if (tap) { in tap_get_user_xdp()
1179 skb->dev = tap->dev; in tap_get_user_xdp()
1193 tap = rcu_dereference(q->tap); in tap_get_user_xdp()
1194 if (tap && tap->count_tx_dropped) in tap_get_user_xdp()
1195 tap->count_tx_dropped(tap); in tap_get_user_xdp()
1204 struct tun_msg_ctl *ctl = m->msg_control; in tap_sendmsg()
1208 if (m->msg_controllen == sizeof(struct tun_msg_ctl) && in tap_sendmsg()
1209 ctl && ctl->type == TUN_MSG_PTR) { in tap_sendmsg()
1210 for (i = 0; i < ctl->num; i++) { in tap_sendmsg()
1211 xdp = &((struct xdp_buff *)ctl->ptr)[i]; in tap_sendmsg()
1217 return tap_get_user(q, ctl ? ctl->ptr : NULL, &m->msg_iter, in tap_sendmsg()
1218 m->msg_flags & MSG_DONTWAIT); in tap_sendmsg()
1225 struct sk_buff *skb = m->msg_control; in tap_recvmsg()
1229 return -EINVAL; in tap_recvmsg()
1231 ret = tap_do_read(q, &m->msg_iter, flags & MSG_DONTWAIT, skb); in tap_recvmsg()
1233 m->msg_flags |= MSG_TRUNC; in tap_recvmsg()
1243 return PTR_RING_PEEK_CALL(&q->ring, __skb_array_len_with_tag); in tap_peek_len()
1260 if (file->f_op != &tap_fops) in tap_get_socket()
1261 return ERR_PTR(-EINVAL); in tap_get_socket()
1262 q = file->private_data; in tap_get_socket()
1264 return ERR_PTR(-EBADFD); in tap_get_socket()
1265 return &q->sock; in tap_get_socket()
1273 if (file->f_op != &tap_fops) in tap_get_ptr_ring()
1274 return ERR_PTR(-EINVAL); in tap_get_ptr_ring()
1275 q = file->private_data; in tap_get_ptr_ring()
1277 return ERR_PTR(-EBADFD); in tap_get_ptr_ring()
1278 return &q->ring; in tap_get_ptr_ring()
1282 int tap_queue_resize(struct tap_dev *tap) in tap_queue_resize() argument
1284 struct net_device *dev = tap->dev; in tap_queue_resize()
1287 int n = tap->numqueues; in tap_queue_resize()
1292 return -ENOMEM; in tap_queue_resize()
1294 list_for_each_entry(q, &tap->queue_list, next) in tap_queue_resize()
1295 rings[i++] = &q->ring; in tap_queue_resize()
1298 dev->tx_queue_len, GFP_KERNEL, in tap_queue_resize()
1312 return -ENOMEM; in tap_list_add()
1314 tap_major->major = MAJOR(major); in tap_list_add()
1316 idr_init(&tap_major->minor_idr); in tap_list_add()
1317 spin_lock_init(&tap_major->minor_lock); in tap_list_add()
1319 tap_major->device_name = device_name; in tap_list_add()
1321 list_add_tail_rcu(&tap_major->next, &major_list); in tap_list_add()
1335 tap_cdev->owner = module; in tap_create_cdev()
1362 if (tap_major->major == MAJOR(major)) { in tap_destroy_cdev()
1363 idr_destroy(&tap_major->minor_idr); in tap_destroy_cdev()
1364 list_del_rcu(&tap_major->next); in tap_destroy_cdev()