• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  * Copyright (C) 2003-2008 Takahiro Hirofuchi
3  *
4  * This is free software; you can redistribute it and/or modify
5  * it under the terms of the GNU General Public License as published by
6  * the Free Software Foundation; either version 2 of the License, or
7  * (at your option) any later version.
8  *
9  * This is distributed in the hope that it will be useful,
10  * but WITHOUT ANY WARRANTY; without even the implied warranty of
11  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
12  * GNU General Public License for more details.
13  *
14  * You should have received a copy of the GNU General Public License
15  * along with this program; if not, write to the Free Software
16  * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307,
17  * USA.
18  */
19 
20 #include <asm/byteorder.h>
21 #include <linux/kthread.h>
22 #include <linux/usb.h>
23 #include <linux/usb/hcd.h>
24 #include <linux/scatterlist.h>
25 
26 #include "usbip_common.h"
27 #include "stub.h"
28 
is_clear_halt_cmd(struct urb * urb)29 static int is_clear_halt_cmd(struct urb *urb)
30 {
31 	struct usb_ctrlrequest *req;
32 
33 	req = (struct usb_ctrlrequest *) urb->setup_packet;
34 
35 	 return (req->bRequest == USB_REQ_CLEAR_FEATURE) &&
36 		 (req->bRequestType == USB_RECIP_ENDPOINT) &&
37 		 (req->wValue == USB_ENDPOINT_HALT);
38 }
39 
is_set_interface_cmd(struct urb * urb)40 static int is_set_interface_cmd(struct urb *urb)
41 {
42 	struct usb_ctrlrequest *req;
43 
44 	req = (struct usb_ctrlrequest *) urb->setup_packet;
45 
46 	return (req->bRequest == USB_REQ_SET_INTERFACE) &&
47 		(req->bRequestType == USB_RECIP_INTERFACE);
48 }
49 
is_set_configuration_cmd(struct urb * urb)50 static int is_set_configuration_cmd(struct urb *urb)
51 {
52 	struct usb_ctrlrequest *req;
53 
54 	req = (struct usb_ctrlrequest *) urb->setup_packet;
55 
56 	return (req->bRequest == USB_REQ_SET_CONFIGURATION) &&
57 		(req->bRequestType == USB_RECIP_DEVICE);
58 }
59 
is_reset_device_cmd(struct urb * urb)60 static int is_reset_device_cmd(struct urb *urb)
61 {
62 	struct usb_ctrlrequest *req;
63 	__u16 value;
64 	__u16 index;
65 
66 	req = (struct usb_ctrlrequest *) urb->setup_packet;
67 	value = le16_to_cpu(req->wValue);
68 	index = le16_to_cpu(req->wIndex);
69 
70 	if ((req->bRequest == USB_REQ_SET_FEATURE) &&
71 	    (req->bRequestType == USB_RT_PORT) &&
72 	    (value == USB_PORT_FEAT_RESET)) {
73 		usbip_dbg_stub_rx("reset_device_cmd, port %u\n", index);
74 		return 1;
75 	} else
76 		return 0;
77 }
78 
tweak_clear_halt_cmd(struct urb * urb)79 static int tweak_clear_halt_cmd(struct urb *urb)
80 {
81 	struct usb_ctrlrequest *req;
82 	int target_endp;
83 	int target_dir;
84 	int target_pipe;
85 	int ret;
86 
87 	req = (struct usb_ctrlrequest *) urb->setup_packet;
88 
89 	/*
90 	 * The stalled endpoint is specified in the wIndex value. The endpoint
91 	 * of the urb is the target of this clear_halt request (i.e., control
92 	 * endpoint).
93 	 */
94 	target_endp = le16_to_cpu(req->wIndex) & 0x000f;
95 
96 	/* the stalled endpoint direction is IN or OUT?. USB_DIR_IN is 0x80.  */
97 	target_dir = le16_to_cpu(req->wIndex) & 0x0080;
98 
99 	if (target_dir)
100 		target_pipe = usb_rcvctrlpipe(urb->dev, target_endp);
101 	else
102 		target_pipe = usb_sndctrlpipe(urb->dev, target_endp);
103 
104 	ret = usb_clear_halt(urb->dev, target_pipe);
105 	if (ret < 0)
106 		dev_err(&urb->dev->dev,
107 			"usb_clear_halt error: devnum %d endp %d ret %d\n",
108 			urb->dev->devnum, target_endp, ret);
109 	else
110 		dev_info(&urb->dev->dev,
111 			 "usb_clear_halt done: devnum %d endp %d\n",
112 			 urb->dev->devnum, target_endp);
113 
114 	return ret;
115 }
116 
tweak_set_interface_cmd(struct urb * urb)117 static int tweak_set_interface_cmd(struct urb *urb)
118 {
119 	struct usb_ctrlrequest *req;
120 	__u16 alternate;
121 	__u16 interface;
122 	int ret;
123 
124 	req = (struct usb_ctrlrequest *) urb->setup_packet;
125 	alternate = le16_to_cpu(req->wValue);
126 	interface = le16_to_cpu(req->wIndex);
127 
128 	usbip_dbg_stub_rx("set_interface: inf %u alt %u\n",
129 			  interface, alternate);
130 
131 	ret = usb_set_interface(urb->dev, interface, alternate);
132 	if (ret < 0)
133 		dev_err(&urb->dev->dev,
134 			"usb_set_interface error: inf %u alt %u ret %d\n",
135 			interface, alternate, ret);
136 	else
137 		dev_info(&urb->dev->dev,
138 			"usb_set_interface done: inf %u alt %u\n",
139 			interface, alternate);
140 
141 	return ret;
142 }
143 
tweak_set_configuration_cmd(struct urb * urb)144 static int tweak_set_configuration_cmd(struct urb *urb)
145 {
146 	struct stub_priv *priv = (struct stub_priv *) urb->context;
147 	struct stub_device *sdev = priv->sdev;
148 	struct usb_ctrlrequest *req;
149 	__u16 config;
150 	int err;
151 
152 	req = (struct usb_ctrlrequest *) urb->setup_packet;
153 	config = le16_to_cpu(req->wValue);
154 
155 	err = usb_set_configuration(sdev->udev, config);
156 	if (err && err != -ENODEV)
157 		dev_err(&sdev->udev->dev, "can't set config #%d, error %d\n",
158 			config, err);
159 	return 0;
160 }
161 
tweak_reset_device_cmd(struct urb * urb)162 static int tweak_reset_device_cmd(struct urb *urb)
163 {
164 	struct stub_priv *priv = (struct stub_priv *) urb->context;
165 	struct stub_device *sdev = priv->sdev;
166 
167 	dev_info(&urb->dev->dev, "usb_queue_reset_device\n");
168 
169 	if (usb_lock_device_for_reset(sdev->udev, NULL) < 0) {
170 		dev_err(&urb->dev->dev, "could not obtain lock to reset device\n");
171 		return 0;
172 	}
173 	usb_reset_device(sdev->udev);
174 	usb_unlock_device(sdev->udev);
175 
176 	return 0;
177 }
178 
179 /*
180  * clear_halt, set_interface, and set_configuration require special tricks.
181  */
tweak_special_requests(struct urb * urb)182 static void tweak_special_requests(struct urb *urb)
183 {
184 	if (!urb || !urb->setup_packet)
185 		return;
186 
187 	if (usb_pipetype(urb->pipe) != PIPE_CONTROL)
188 		return;
189 
190 	if (is_clear_halt_cmd(urb))
191 		/* tweak clear_halt */
192 		 tweak_clear_halt_cmd(urb);
193 
194 	else if (is_set_interface_cmd(urb))
195 		/* tweak set_interface */
196 		tweak_set_interface_cmd(urb);
197 
198 	else if (is_set_configuration_cmd(urb))
199 		/* tweak set_configuration */
200 		tweak_set_configuration_cmd(urb);
201 
202 	else if (is_reset_device_cmd(urb))
203 		tweak_reset_device_cmd(urb);
204 	else
205 		usbip_dbg_stub_rx("no need to tweak\n");
206 }
207 
208 /*
209  * stub_recv_unlink() unlinks the URB by a call to usb_unlink_urb().
210  * By unlinking the urb asynchronously, stub_rx can continuously
211  * process coming urbs.  Even if the urb is unlinked, its completion
212  * handler will be called and stub_tx will send a return pdu.
213  *
214  * See also comments about unlinking strategy in vhci_hcd.c.
215  */
stub_recv_cmd_unlink(struct stub_device * sdev,struct usbip_header * pdu)216 static int stub_recv_cmd_unlink(struct stub_device *sdev,
217 				struct usbip_header *pdu)
218 {
219 	int ret, i;
220 	unsigned long flags;
221 	struct stub_priv *priv;
222 
223 	spin_lock_irqsave(&sdev->priv_lock, flags);
224 
225 	list_for_each_entry(priv, &sdev->priv_init, list) {
226 		if (priv->seqnum != pdu->u.cmd_unlink.seqnum)
227 			continue;
228 
229 		/*
230 		 * This matched urb is not completed yet (i.e., be in
231 		 * flight in usb hcd hardware/driver). Now we are
232 		 * cancelling it. The unlinking flag means that we are
233 		 * now not going to return the normal result pdu of a
234 		 * submission request, but going to return a result pdu
235 		 * of the unlink request.
236 		 */
237 		priv->unlinking = 1;
238 
239 		/*
240 		 * In the case that unlinking flag is on, prev->seqnum
241 		 * is changed from the seqnum of the cancelling urb to
242 		 * the seqnum of the unlink request. This will be used
243 		 * to make the result pdu of the unlink request.
244 		 */
245 		priv->seqnum = pdu->base.seqnum;
246 
247 		spin_unlock_irqrestore(&sdev->priv_lock, flags);
248 
249 		/*
250 		 * usb_unlink_urb() is now out of spinlocking to avoid
251 		 * spinlock recursion since stub_complete() is
252 		 * sometimes called in this context but not in the
253 		 * interrupt context.  If stub_complete() is executed
254 		 * before we call usb_unlink_urb(), usb_unlink_urb()
255 		 * will return an error value. In this case, stub_tx
256 		 * will return the result pdu of this unlink request
257 		 * though submission is completed and actual unlinking
258 		 * is not executed. OK?
259 		 */
260 		/* In the above case, urb->status is not -ECONNRESET,
261 		 * so a driver in a client host will know the failure
262 		 * of the unlink request ?
263 		 */
264 		for (i = priv->completed_urbs; i < priv->num_urbs; i++) {
265 			ret = usb_unlink_urb(priv->urbs[i]);
266 			if (ret != -EINPROGRESS)
267 				dev_err(&priv->urbs[i]->dev->dev,
268 					"failed to unlink %d/%d urb of seqnum %lu, ret %d\n",
269 					i + 1, priv->num_urbs,
270 					priv->seqnum, ret);
271 		}
272 		return 0;
273 	}
274 
275 	usbip_dbg_stub_rx("seqnum %d is not pending\n",
276 			  pdu->u.cmd_unlink.seqnum);
277 
278 	/*
279 	 * The urb of the unlink target is not found in priv_init queue. It was
280 	 * already completed and its results is/was going to be sent by a
281 	 * CMD_RET pdu. In this case, usb_unlink_urb() is not needed. We only
282 	 * return the completeness of this unlink request to vhci_hcd.
283 	 */
284 	stub_enqueue_ret_unlink(sdev, pdu->base.seqnum, 0);
285 
286 	spin_unlock_irqrestore(&sdev->priv_lock, flags);
287 
288 	return 0;
289 }
290 
valid_request(struct stub_device * sdev,struct usbip_header * pdu)291 static int valid_request(struct stub_device *sdev, struct usbip_header *pdu)
292 {
293 	struct usbip_device *ud = &sdev->ud;
294 	int valid = 0;
295 
296 	if (pdu->base.devid == sdev->devid) {
297 		spin_lock_irq(&ud->lock);
298 		if (ud->status == SDEV_ST_USED) {
299 			/* A request is valid. */
300 			valid = 1;
301 		}
302 		spin_unlock_irq(&ud->lock);
303 	}
304 
305 	return valid;
306 }
307 
stub_priv_alloc(struct stub_device * sdev,struct usbip_header * pdu)308 static struct stub_priv *stub_priv_alloc(struct stub_device *sdev,
309 					 struct usbip_header *pdu)
310 {
311 	struct stub_priv *priv;
312 	struct usbip_device *ud = &sdev->ud;
313 	unsigned long flags;
314 
315 	spin_lock_irqsave(&sdev->priv_lock, flags);
316 
317 	priv = kmem_cache_zalloc(stub_priv_cache, GFP_ATOMIC);
318 	if (!priv) {
319 		dev_err(&sdev->udev->dev, "alloc stub_priv\n");
320 		spin_unlock_irqrestore(&sdev->priv_lock, flags);
321 		usbip_event_add(ud, SDEV_EVENT_ERROR_MALLOC);
322 		return NULL;
323 	}
324 
325 	priv->seqnum = pdu->base.seqnum;
326 	priv->sdev = sdev;
327 
328 	/*
329 	 * After a stub_priv is linked to a list_head,
330 	 * our error handler can free allocated data.
331 	 */
332 	list_add_tail(&priv->list, &sdev->priv_init);
333 
334 	spin_unlock_irqrestore(&sdev->priv_lock, flags);
335 
336 	return priv;
337 }
338 
get_pipe(struct stub_device * sdev,struct usbip_header * pdu)339 static int get_pipe(struct stub_device *sdev, struct usbip_header *pdu)
340 {
341 	struct usb_device *udev = sdev->udev;
342 	struct usb_host_endpoint *ep;
343 	struct usb_endpoint_descriptor *epd = NULL;
344 	int epnum = pdu->base.ep;
345 	int dir = pdu->base.direction;
346 
347 	if (epnum < 0 || epnum > 15)
348 		goto err_ret;
349 
350 	if (dir == USBIP_DIR_IN)
351 		ep = udev->ep_in[epnum & 0x7f];
352 	else
353 		ep = udev->ep_out[epnum & 0x7f];
354 	if (!ep)
355 		goto err_ret;
356 
357 	epd = &ep->desc;
358 
359 	if (usb_endpoint_xfer_control(epd)) {
360 		if (dir == USBIP_DIR_OUT)
361 			return usb_sndctrlpipe(udev, epnum);
362 		else
363 			return usb_rcvctrlpipe(udev, epnum);
364 	}
365 
366 	if (usb_endpoint_xfer_bulk(epd)) {
367 		if (dir == USBIP_DIR_OUT)
368 			return usb_sndbulkpipe(udev, epnum);
369 		else
370 			return usb_rcvbulkpipe(udev, epnum);
371 	}
372 
373 	if (usb_endpoint_xfer_int(epd)) {
374 		if (dir == USBIP_DIR_OUT)
375 			return usb_sndintpipe(udev, epnum);
376 		else
377 			return usb_rcvintpipe(udev, epnum);
378 	}
379 
380 	if (usb_endpoint_xfer_isoc(epd)) {
381 		/* validate number of packets */
382 		if (pdu->u.cmd_submit.number_of_packets < 0 ||
383 		    pdu->u.cmd_submit.number_of_packets >
384 		    USBIP_MAX_ISO_PACKETS) {
385 			dev_err(&sdev->udev->dev,
386 				"CMD_SUBMIT: isoc invalid num packets %d\n",
387 				pdu->u.cmd_submit.number_of_packets);
388 			return -1;
389 		}
390 		if (dir == USBIP_DIR_OUT)
391 			return usb_sndisocpipe(udev, epnum);
392 		else
393 			return usb_rcvisocpipe(udev, epnum);
394 	}
395 
396 err_ret:
397 	/* NOT REACHED */
398 	dev_err(&sdev->udev->dev, "CMD_SUBMIT: invalid epnum %d\n", epnum);
399 	return -1;
400 }
401 
masking_bogus_flags(struct urb * urb)402 static void masking_bogus_flags(struct urb *urb)
403 {
404 	int				xfertype;
405 	struct usb_device		*dev;
406 	struct usb_host_endpoint	*ep;
407 	int				is_out;
408 	unsigned int	allowed;
409 
410 	if (!urb || urb->hcpriv || !urb->complete)
411 		return;
412 	dev = urb->dev;
413 	if ((!dev) || (dev->state < USB_STATE_UNAUTHENTICATED))
414 		return;
415 
416 	ep = (usb_pipein(urb->pipe) ? dev->ep_in : dev->ep_out)
417 		[usb_pipeendpoint(urb->pipe)];
418 	if (!ep)
419 		return;
420 
421 	xfertype = usb_endpoint_type(&ep->desc);
422 	if (xfertype == USB_ENDPOINT_XFER_CONTROL) {
423 		struct usb_ctrlrequest *setup =
424 			(struct usb_ctrlrequest *) urb->setup_packet;
425 
426 		if (!setup)
427 			return;
428 		is_out = !(setup->bRequestType & USB_DIR_IN) ||
429 			!setup->wLength;
430 	} else {
431 		is_out = usb_endpoint_dir_out(&ep->desc);
432 	}
433 
434 	/* enforce simple/standard policy */
435 	allowed = (URB_NO_TRANSFER_DMA_MAP | URB_NO_INTERRUPT |
436 		   URB_DIR_MASK | URB_FREE_BUFFER);
437 	switch (xfertype) {
438 	case USB_ENDPOINT_XFER_BULK:
439 		if (is_out)
440 			allowed |= URB_ZERO_PACKET;
441 		/* FALLTHROUGH */
442 	case USB_ENDPOINT_XFER_CONTROL:
443 		allowed |= URB_NO_FSBR;	/* only affects UHCI */
444 		/* FALLTHROUGH */
445 	default:			/* all non-iso endpoints */
446 		if (!is_out)
447 			allowed |= URB_SHORT_NOT_OK;
448 		break;
449 	case USB_ENDPOINT_XFER_ISOC:
450 		allowed |= URB_ISO_ASAP;
451 		break;
452 	}
453 	urb->transfer_flags &= allowed;
454 }
455 
stub_recv_xbuff(struct usbip_device * ud,struct stub_priv * priv)456 static int stub_recv_xbuff(struct usbip_device *ud, struct stub_priv *priv)
457 {
458 	int ret;
459 	int i;
460 
461 	for (i = 0; i < priv->num_urbs; i++) {
462 		ret = usbip_recv_xbuff(ud, priv->urbs[i]);
463 		if (ret < 0)
464 			break;
465 	}
466 
467 	return ret;
468 }
469 
stub_recv_cmd_submit(struct stub_device * sdev,struct usbip_header * pdu)470 static void stub_recv_cmd_submit(struct stub_device *sdev,
471 				 struct usbip_header *pdu)
472 {
473 	struct stub_priv *priv;
474 	struct usbip_device *ud = &sdev->ud;
475 	struct usb_device *udev = sdev->udev;
476 	struct scatterlist *sgl = NULL, *sg;
477 	void *buffer = NULL;
478 	unsigned long long buf_len;
479 	int nents;
480 	int num_urbs = 1;
481 	int pipe = get_pipe(sdev, pdu);
482 	int use_sg = pdu->u.cmd_submit.transfer_flags & URB_DMA_MAP_SG;
483 	int support_sg = 1;
484 	int np = 0;
485 	int ret, i;
486 
487 	if (pipe == -1)
488 		return;
489 
490 	/*
491 	 * Smatch reported the error case where use_sg is true and buf_len is 0.
492 	 * In this case, It adds SDEV_EVENT_ERROR_MALLOC and stub_priv will be
493 	 * released by stub event handler and connection will be shut down.
494 	 */
495 	priv = stub_priv_alloc(sdev, pdu);
496 	if (!priv)
497 		return;
498 
499 	buf_len = (unsigned long long)pdu->u.cmd_submit.transfer_buffer_length;
500 
501 	if (use_sg && !buf_len) {
502 		dev_err(&udev->dev, "sg buffer with zero length\n");
503 		goto err_malloc;
504 	}
505 
506 	/* allocate urb transfer buffer, if needed */
507 	if (buf_len) {
508 		if (use_sg) {
509 			sgl = sgl_alloc(buf_len, GFP_KERNEL, &nents);
510 			if (!sgl)
511 				goto err_malloc;
512 
513 			/* Check if the server's HCD supports SG */
514 			if (!udev->bus->sg_tablesize) {
515 				/*
516 				 * If the server's HCD doesn't support SG, break
517 				 * a single SG request into several URBs and map
518 				 * each SG list entry to corresponding URB
519 				 * buffer. The previously allocated SG list is
520 				 * stored in priv->sgl (If the server's HCD
521 				 * support SG, SG list is stored only in
522 				 * urb->sg) and it is used as an indicator that
523 				 * the server split single SG request into
524 				 * several URBs. Later, priv->sgl is used by
525 				 * stub_complete() and stub_send_ret_submit() to
526 				 * reassemble the divied URBs.
527 				 */
528 				support_sg = 0;
529 				num_urbs = nents;
530 				priv->completed_urbs = 0;
531 				pdu->u.cmd_submit.transfer_flags &=
532 								~URB_DMA_MAP_SG;
533 			}
534 		} else {
535 			buffer = kzalloc(buf_len, GFP_KERNEL);
536 			if (!buffer)
537 				goto err_malloc;
538 		}
539 	}
540 
541 	/* allocate urb array */
542 	priv->num_urbs = num_urbs;
543 	priv->urbs = kmalloc_array(num_urbs, sizeof(*priv->urbs), GFP_KERNEL);
544 	if (!priv->urbs)
545 		goto err_urbs;
546 
547 	/* setup a urb */
548 	if (support_sg) {
549 		if (usb_pipeisoc(pipe))
550 			np = pdu->u.cmd_submit.number_of_packets;
551 
552 		priv->urbs[0] = usb_alloc_urb(np, GFP_KERNEL);
553 		if (!priv->urbs[0])
554 			goto err_urb;
555 
556 		if (buf_len) {
557 			if (use_sg) {
558 				priv->urbs[0]->sg = sgl;
559 				priv->urbs[0]->num_sgs = nents;
560 				priv->urbs[0]->transfer_buffer = NULL;
561 			} else {
562 				priv->urbs[0]->transfer_buffer = buffer;
563 			}
564 		}
565 
566 		/* copy urb setup packet */
567 		priv->urbs[0]->setup_packet = kmemdup(&pdu->u.cmd_submit.setup,
568 					8, GFP_KERNEL);
569 		if (!priv->urbs[0]->setup_packet) {
570 			usbip_event_add(ud, SDEV_EVENT_ERROR_MALLOC);
571 			return;
572 		}
573 
574 		usbip_pack_pdu(pdu, priv->urbs[0], USBIP_CMD_SUBMIT, 0);
575 	} else {
576 		for_each_sg(sgl, sg, nents, i) {
577 			priv->urbs[i] = usb_alloc_urb(0, GFP_KERNEL);
578 			/* The URBs which is previously allocated will be freed
579 			 * in stub_device_cleanup_urbs() if error occurs.
580 			 */
581 			if (!priv->urbs[i])
582 				goto err_urb;
583 
584 			usbip_pack_pdu(pdu, priv->urbs[i], USBIP_CMD_SUBMIT, 0);
585 			priv->urbs[i]->transfer_buffer = sg_virt(sg);
586 			priv->urbs[i]->transfer_buffer_length = sg->length;
587 		}
588 		priv->sgl = sgl;
589 	}
590 
591 	for (i = 0; i < num_urbs; i++) {
592 		/* set other members from the base header of pdu */
593 		priv->urbs[i]->context = (void *) priv;
594 		priv->urbs[i]->dev = udev;
595 		priv->urbs[i]->pipe = pipe;
596 		priv->urbs[i]->complete = stub_complete;
597 
598 		/* no need to submit an intercepted request, but harmless? */
599 		tweak_special_requests(priv->urbs[i]);
600 
601 		masking_bogus_flags(priv->urbs[i]);
602 	}
603 
604 	if (stub_recv_xbuff(ud, priv) < 0)
605 		return;
606 
607 	if (usbip_recv_iso(ud, priv->urbs[0]) < 0)
608 		return;
609 
610 	/* urb is now ready to submit */
611 	for (i = 0; i < priv->num_urbs; i++) {
612 		ret = usb_submit_urb(priv->urbs[i], GFP_KERNEL);
613 
614 		if (ret == 0)
615 			usbip_dbg_stub_rx("submit urb ok, seqnum %u\n",
616 					pdu->base.seqnum);
617 		else {
618 			dev_err(&udev->dev, "submit_urb error, %d\n", ret);
619 			usbip_dump_header(pdu);
620 			usbip_dump_urb(priv->urbs[i]);
621 
622 			/*
623 			 * Pessimistic.
624 			 * This connection will be discarded.
625 			 */
626 			usbip_event_add(ud, SDEV_EVENT_ERROR_SUBMIT);
627 			break;
628 		}
629 	}
630 
631 	usbip_dbg_stub_rx("Leave\n");
632 	return;
633 
634 err_urb:
635 	kfree(priv->urbs);
636 err_urbs:
637 	kfree(buffer);
638 	sgl_free(sgl);
639 err_malloc:
640 	usbip_event_add(ud, SDEV_EVENT_ERROR_MALLOC);
641 }
642 
643 /* recv a pdu */
stub_rx_pdu(struct usbip_device * ud)644 static void stub_rx_pdu(struct usbip_device *ud)
645 {
646 	int ret;
647 	struct usbip_header pdu;
648 	struct stub_device *sdev = container_of(ud, struct stub_device, ud);
649 	struct device *dev = &sdev->udev->dev;
650 
651 	usbip_dbg_stub_rx("Enter\n");
652 
653 	memset(&pdu, 0, sizeof(pdu));
654 
655 	/* receive a pdu header */
656 	ret = usbip_recv(ud->tcp_socket, &pdu, sizeof(pdu));
657 	if (ret != sizeof(pdu)) {
658 		dev_err(dev, "recv a header, %d\n", ret);
659 		usbip_event_add(ud, SDEV_EVENT_ERROR_TCP);
660 		return;
661 	}
662 
663 	usbip_header_correct_endian(&pdu, 0);
664 
665 	if (usbip_dbg_flag_stub_rx)
666 		usbip_dump_header(&pdu);
667 
668 	if (!valid_request(sdev, &pdu)) {
669 		dev_err(dev, "recv invalid request\n");
670 		usbip_event_add(ud, SDEV_EVENT_ERROR_TCP);
671 		return;
672 	}
673 
674 	switch (pdu.base.command) {
675 	case USBIP_CMD_UNLINK:
676 		stub_recv_cmd_unlink(sdev, &pdu);
677 		break;
678 
679 	case USBIP_CMD_SUBMIT:
680 		stub_recv_cmd_submit(sdev, &pdu);
681 		break;
682 
683 	default:
684 		/* NOTREACHED */
685 		dev_err(dev, "unknown pdu\n");
686 		usbip_event_add(ud, SDEV_EVENT_ERROR_TCP);
687 		break;
688 	}
689 }
690 
stub_rx_loop(void * data)691 int stub_rx_loop(void *data)
692 {
693 	struct usbip_device *ud = data;
694 
695 	while (!kthread_should_stop()) {
696 		if (usbip_event_happened(ud))
697 			break;
698 
699 		stub_rx_pdu(ud);
700 	}
701 
702 	return 0;
703 }
704