Linux kernel mirror (for testing)
git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel
os
linux
1#include <linux/kernel.h>
2#include <linux/errno.h>
3#include <linux/init.h>
4#include <linux/slab.h>
5#include <linux/mm.h>
6#include <linux/module.h>
7#include <linux/moduleparam.h>
8#include <linux/scatterlist.h>
9#include <linux/mutex.h>
10
11#include <linux/usb.h>
12
13
14/*-------------------------------------------------------------------------*/
15
16static int override_alt = -1;
17module_param_named(alt, override_alt, int, 0644);
18MODULE_PARM_DESC(alt, ">= 0 to override altsetting selection");
19
20/*-------------------------------------------------------------------------*/
21
22/* FIXME make these public somewhere; usbdevfs.h? */
23struct usbtest_param {
24 /* inputs */
25 unsigned test_num; /* 0..(TEST_CASES-1) */
26 unsigned iterations;
27 unsigned length;
28 unsigned vary;
29 unsigned sglen;
30
31 /* outputs */
32 struct timeval duration;
33};
34#define USBTEST_REQUEST _IOWR('U', 100, struct usbtest_param)
35
36/*-------------------------------------------------------------------------*/
37
38#define GENERIC /* let probe() bind using module params */
39
40/* Some devices that can be used for testing will have "real" drivers.
41 * Entries for those need to be enabled here by hand, after disabling
42 * that "real" driver.
43 */
44//#define IBOT2 /* grab iBOT2 webcams */
45//#define KEYSPAN_19Qi /* grab un-renumerated serial adapter */
46
47/*-------------------------------------------------------------------------*/
48
49struct usbtest_info {
50 const char *name;
51 u8 ep_in; /* bulk/intr source */
52 u8 ep_out; /* bulk/intr sink */
53 unsigned autoconf:1;
54 unsigned ctrl_out:1;
55 unsigned iso:1; /* try iso in/out */
56 int alt;
57};
58
59/* this is accessed only through usbfs ioctl calls.
60 * one ioctl to issue a test ... one lock per device.
61 * tests create other threads if they need them.
62 * urbs and buffers are allocated dynamically,
63 * and data generated deterministically.
64 */
65struct usbtest_dev {
66 struct usb_interface *intf;
67 struct usbtest_info *info;
68 int in_pipe;
69 int out_pipe;
70 int in_iso_pipe;
71 int out_iso_pipe;
72 struct usb_endpoint_descriptor *iso_in, *iso_out;
73 struct mutex lock;
74
75#define TBUF_SIZE 256
76 u8 *buf;
77};
78
79static struct usb_device *testdev_to_usbdev(struct usbtest_dev *test)
80{
81 return interface_to_usbdev(test->intf);
82}
83
84/* set up all urbs so they can be used with either bulk or interrupt */
85#define INTERRUPT_RATE 1 /* msec/transfer */
86
87#define ERROR(tdev, fmt, args...) \
88 dev_err(&(tdev)->intf->dev , fmt , ## args)
89#define WARNING(tdev, fmt, args...) \
90 dev_warn(&(tdev)->intf->dev , fmt , ## args)
91
92#define GUARD_BYTE 0xA5
93
94/*-------------------------------------------------------------------------*/
95
96static int
97get_endpoints(struct usbtest_dev *dev, struct usb_interface *intf)
98{
99 int tmp;
100 struct usb_host_interface *alt;
101 struct usb_host_endpoint *in, *out;
102 struct usb_host_endpoint *iso_in, *iso_out;
103 struct usb_device *udev;
104
105 for (tmp = 0; tmp < intf->num_altsetting; tmp++) {
106 unsigned ep;
107
108 in = out = NULL;
109 iso_in = iso_out = NULL;
110 alt = intf->altsetting + tmp;
111
112 if (override_alt >= 0 &&
113 override_alt != alt->desc.bAlternateSetting)
114 continue;
115
116 /* take the first altsetting with in-bulk + out-bulk;
117 * ignore other endpoints and altsettings.
118 */
119 for (ep = 0; ep < alt->desc.bNumEndpoints; ep++) {
120 struct usb_host_endpoint *e;
121
122 e = alt->endpoint + ep;
123 switch (usb_endpoint_type(&e->desc)) {
124 case USB_ENDPOINT_XFER_BULK:
125 break;
126 case USB_ENDPOINT_XFER_ISOC:
127 if (dev->info->iso)
128 goto try_iso;
129 /* FALLTHROUGH */
130 default:
131 continue;
132 }
133 if (usb_endpoint_dir_in(&e->desc)) {
134 if (!in)
135 in = e;
136 } else {
137 if (!out)
138 out = e;
139 }
140 continue;
141try_iso:
142 if (usb_endpoint_dir_in(&e->desc)) {
143 if (!iso_in)
144 iso_in = e;
145 } else {
146 if (!iso_out)
147 iso_out = e;
148 }
149 }
150 if ((in && out) || iso_in || iso_out)
151 goto found;
152 }
153 return -EINVAL;
154
155found:
156 udev = testdev_to_usbdev(dev);
157 dev->info->alt = alt->desc.bAlternateSetting;
158 if (alt->desc.bAlternateSetting != 0) {
159 tmp = usb_set_interface(udev,
160 alt->desc.bInterfaceNumber,
161 alt->desc.bAlternateSetting);
162 if (tmp < 0)
163 return tmp;
164 }
165
166 if (in) {
167 dev->in_pipe = usb_rcvbulkpipe(udev,
168 in->desc.bEndpointAddress & USB_ENDPOINT_NUMBER_MASK);
169 dev->out_pipe = usb_sndbulkpipe(udev,
170 out->desc.bEndpointAddress & USB_ENDPOINT_NUMBER_MASK);
171 }
172 if (iso_in) {
173 dev->iso_in = &iso_in->desc;
174 dev->in_iso_pipe = usb_rcvisocpipe(udev,
175 iso_in->desc.bEndpointAddress
176 & USB_ENDPOINT_NUMBER_MASK);
177 }
178
179 if (iso_out) {
180 dev->iso_out = &iso_out->desc;
181 dev->out_iso_pipe = usb_sndisocpipe(udev,
182 iso_out->desc.bEndpointAddress
183 & USB_ENDPOINT_NUMBER_MASK);
184 }
185 return 0;
186}
187
188/*-------------------------------------------------------------------------*/
189
190/* Support for testing basic non-queued I/O streams.
191 *
192 * These just package urbs as requests that can be easily canceled.
193 * Each urb's data buffer is dynamically allocated; callers can fill
194 * them with non-zero test data (or test for it) when appropriate.
195 */
196
197static void simple_callback(struct urb *urb)
198{
199 complete(urb->context);
200}
201
202static struct urb *usbtest_alloc_urb(
203 struct usb_device *udev,
204 int pipe,
205 unsigned long bytes,
206 unsigned transfer_flags,
207 unsigned offset)
208{
209 struct urb *urb;
210
211 urb = usb_alloc_urb(0, GFP_KERNEL);
212 if (!urb)
213 return urb;
214 usb_fill_bulk_urb(urb, udev, pipe, NULL, bytes, simple_callback, NULL);
215 urb->interval = (udev->speed == USB_SPEED_HIGH)
216 ? (INTERRUPT_RATE << 3)
217 : INTERRUPT_RATE;
218 urb->transfer_flags = transfer_flags;
219 if (usb_pipein(pipe))
220 urb->transfer_flags |= URB_SHORT_NOT_OK;
221
222 if (urb->transfer_flags & URB_NO_TRANSFER_DMA_MAP)
223 urb->transfer_buffer = usb_alloc_coherent(udev, bytes + offset,
224 GFP_KERNEL, &urb->transfer_dma);
225 else
226 urb->transfer_buffer = kmalloc(bytes + offset, GFP_KERNEL);
227
228 if (!urb->transfer_buffer) {
229 usb_free_urb(urb);
230 return NULL;
231 }
232
233 /* To test unaligned transfers add an offset and fill the
234 unused memory with a guard value */
235 if (offset) {
236 memset(urb->transfer_buffer, GUARD_BYTE, offset);
237 urb->transfer_buffer += offset;
238 if (urb->transfer_flags & URB_NO_TRANSFER_DMA_MAP)
239 urb->transfer_dma += offset;
240 }
241
242 /* For inbound transfers use guard byte so that test fails if
243 data not correctly copied */
244 memset(urb->transfer_buffer,
245 usb_pipein(urb->pipe) ? GUARD_BYTE : 0,
246 bytes);
247 return urb;
248}
249
250static struct urb *simple_alloc_urb(
251 struct usb_device *udev,
252 int pipe,
253 unsigned long bytes)
254{
255 return usbtest_alloc_urb(udev, pipe, bytes, URB_NO_TRANSFER_DMA_MAP, 0);
256}
257
258static unsigned pattern;
259static unsigned mod_pattern;
260module_param_named(pattern, mod_pattern, uint, S_IRUGO | S_IWUSR);
261MODULE_PARM_DESC(mod_pattern, "i/o pattern (0 == zeroes)");
262
263static inline void simple_fill_buf(struct urb *urb)
264{
265 unsigned i;
266 u8 *buf = urb->transfer_buffer;
267 unsigned len = urb->transfer_buffer_length;
268
269 switch (pattern) {
270 default:
271 /* FALLTHROUGH */
272 case 0:
273 memset(buf, 0, len);
274 break;
275 case 1: /* mod63 */
276 for (i = 0; i < len; i++)
277 *buf++ = (u8) (i % 63);
278 break;
279 }
280}
281
282static inline unsigned long buffer_offset(void *buf)
283{
284 return (unsigned long)buf & (ARCH_KMALLOC_MINALIGN - 1);
285}
286
287static int check_guard_bytes(struct usbtest_dev *tdev, struct urb *urb)
288{
289 u8 *buf = urb->transfer_buffer;
290 u8 *guard = buf - buffer_offset(buf);
291 unsigned i;
292
293 for (i = 0; guard < buf; i++, guard++) {
294 if (*guard != GUARD_BYTE) {
295 ERROR(tdev, "guard byte[%d] %d (not %d)\n",
296 i, *guard, GUARD_BYTE);
297 return -EINVAL;
298 }
299 }
300 return 0;
301}
302
303static int simple_check_buf(struct usbtest_dev *tdev, struct urb *urb)
304{
305 unsigned i;
306 u8 expected;
307 u8 *buf = urb->transfer_buffer;
308 unsigned len = urb->actual_length;
309
310 int ret = check_guard_bytes(tdev, urb);
311 if (ret)
312 return ret;
313
314 for (i = 0; i < len; i++, buf++) {
315 switch (pattern) {
316 /* all-zeroes has no synchronization issues */
317 case 0:
318 expected = 0;
319 break;
320 /* mod63 stays in sync with short-terminated transfers,
321 * or otherwise when host and gadget agree on how large
322 * each usb transfer request should be. resync is done
323 * with set_interface or set_config.
324 */
325 case 1: /* mod63 */
326 expected = i % 63;
327 break;
328 /* always fail unsupported patterns */
329 default:
330 expected = !*buf;
331 break;
332 }
333 if (*buf == expected)
334 continue;
335 ERROR(tdev, "buf[%d] = %d (not %d)\n", i, *buf, expected);
336 return -EINVAL;
337 }
338 return 0;
339}
340
341static void simple_free_urb(struct urb *urb)
342{
343 unsigned long offset = buffer_offset(urb->transfer_buffer);
344
345 if (urb->transfer_flags & URB_NO_TRANSFER_DMA_MAP)
346 usb_free_coherent(
347 urb->dev,
348 urb->transfer_buffer_length + offset,
349 urb->transfer_buffer - offset,
350 urb->transfer_dma - offset);
351 else
352 kfree(urb->transfer_buffer - offset);
353 usb_free_urb(urb);
354}
355
356static int simple_io(
357 struct usbtest_dev *tdev,
358 struct urb *urb,
359 int iterations,
360 int vary,
361 int expected,
362 const char *label
363)
364{
365 struct usb_device *udev = urb->dev;
366 int max = urb->transfer_buffer_length;
367 struct completion completion;
368 int retval = 0;
369
370 urb->context = &completion;
371 while (retval == 0 && iterations-- > 0) {
372 init_completion(&completion);
373 if (usb_pipeout(urb->pipe)) {
374 simple_fill_buf(urb);
375 urb->transfer_flags |= URB_ZERO_PACKET;
376 }
377 retval = usb_submit_urb(urb, GFP_KERNEL);
378 if (retval != 0)
379 break;
380
381 /* NOTE: no timeouts; can't be broken out of by interrupt */
382 wait_for_completion(&completion);
383 retval = urb->status;
384 urb->dev = udev;
385 if (retval == 0 && usb_pipein(urb->pipe))
386 retval = simple_check_buf(tdev, urb);
387
388 if (vary) {
389 int len = urb->transfer_buffer_length;
390
391 len += vary;
392 len %= max;
393 if (len == 0)
394 len = (vary < max) ? vary : max;
395 urb->transfer_buffer_length = len;
396 }
397
398 /* FIXME if endpoint halted, clear halt (and log) */
399 }
400 urb->transfer_buffer_length = max;
401
402 if (expected != retval)
403 dev_err(&udev->dev,
404 "%s failed, iterations left %d, status %d (not %d)\n",
405 label, iterations, retval, expected);
406 return retval;
407}
408
409
410/*-------------------------------------------------------------------------*/
411
412/* We use scatterlist primitives to test queued I/O.
413 * Yes, this also tests the scatterlist primitives.
414 */
415
416static void free_sglist(struct scatterlist *sg, int nents)
417{
418 unsigned i;
419
420 if (!sg)
421 return;
422 for (i = 0; i < nents; i++) {
423 if (!sg_page(&sg[i]))
424 continue;
425 kfree(sg_virt(&sg[i]));
426 }
427 kfree(sg);
428}
429
430static struct scatterlist *
431alloc_sglist(int nents, int max, int vary)
432{
433 struct scatterlist *sg;
434 unsigned i;
435 unsigned size = max;
436
437 if (max == 0)
438 return NULL;
439
440 sg = kmalloc_array(nents, sizeof(*sg), GFP_KERNEL);
441 if (!sg)
442 return NULL;
443 sg_init_table(sg, nents);
444
445 for (i = 0; i < nents; i++) {
446 char *buf;
447 unsigned j;
448
449 buf = kzalloc(size, GFP_KERNEL);
450 if (!buf) {
451 free_sglist(sg, i);
452 return NULL;
453 }
454
455 /* kmalloc pages are always physically contiguous! */
456 sg_set_buf(&sg[i], buf, size);
457
458 switch (pattern) {
459 case 0:
460 /* already zeroed */
461 break;
462 case 1:
463 for (j = 0; j < size; j++)
464 *buf++ = (u8) (j % 63);
465 break;
466 }
467
468 if (vary) {
469 size += vary;
470 size %= max;
471 if (size == 0)
472 size = (vary < max) ? vary : max;
473 }
474 }
475
476 return sg;
477}
478
479static int perform_sglist(
480 struct usbtest_dev *tdev,
481 unsigned iterations,
482 int pipe,
483 struct usb_sg_request *req,
484 struct scatterlist *sg,
485 int nents
486)
487{
488 struct usb_device *udev = testdev_to_usbdev(tdev);
489 int retval = 0;
490
491 while (retval == 0 && iterations-- > 0) {
492 retval = usb_sg_init(req, udev, pipe,
493 (udev->speed == USB_SPEED_HIGH)
494 ? (INTERRUPT_RATE << 3)
495 : INTERRUPT_RATE,
496 sg, nents, 0, GFP_KERNEL);
497
498 if (retval)
499 break;
500 usb_sg_wait(req);
501 retval = req->status;
502
503 /* FIXME check resulting data pattern */
504
505 /* FIXME if endpoint halted, clear halt (and log) */
506 }
507
508 /* FIXME for unlink or fault handling tests, don't report
509 * failure if retval is as we expected ...
510 */
511 if (retval)
512 ERROR(tdev, "perform_sglist failed, "
513 "iterations left %d, status %d\n",
514 iterations, retval);
515 return retval;
516}
517
518
519/*-------------------------------------------------------------------------*/
520
521/* unqueued control message testing
522 *
523 * there's a nice set of device functional requirements in chapter 9 of the
524 * usb 2.0 spec, which we can apply to ANY device, even ones that don't use
525 * special test firmware.
526 *
527 * we know the device is configured (or suspended) by the time it's visible
528 * through usbfs. we can't change that, so we won't test enumeration (which
529 * worked 'well enough' to get here, this time), power management (ditto),
530 * or remote wakeup (which needs human interaction).
531 */
532
533static unsigned realworld = 1;
534module_param(realworld, uint, 0);
535MODULE_PARM_DESC(realworld, "clear to demand stricter spec compliance");
536
537static int get_altsetting(struct usbtest_dev *dev)
538{
539 struct usb_interface *iface = dev->intf;
540 struct usb_device *udev = interface_to_usbdev(iface);
541 int retval;
542
543 retval = usb_control_msg(udev, usb_rcvctrlpipe(udev, 0),
544 USB_REQ_GET_INTERFACE, USB_DIR_IN|USB_RECIP_INTERFACE,
545 0, iface->altsetting[0].desc.bInterfaceNumber,
546 dev->buf, 1, USB_CTRL_GET_TIMEOUT);
547 switch (retval) {
548 case 1:
549 return dev->buf[0];
550 case 0:
551 retval = -ERANGE;
552 /* FALLTHROUGH */
553 default:
554 return retval;
555 }
556}
557
558static int set_altsetting(struct usbtest_dev *dev, int alternate)
559{
560 struct usb_interface *iface = dev->intf;
561 struct usb_device *udev;
562
563 if (alternate < 0 || alternate >= 256)
564 return -EINVAL;
565
566 udev = interface_to_usbdev(iface);
567 return usb_set_interface(udev,
568 iface->altsetting[0].desc.bInterfaceNumber,
569 alternate);
570}
571
572static int is_good_config(struct usbtest_dev *tdev, int len)
573{
574 struct usb_config_descriptor *config;
575
576 if (len < sizeof(*config))
577 return 0;
578 config = (struct usb_config_descriptor *) tdev->buf;
579
580 switch (config->bDescriptorType) {
581 case USB_DT_CONFIG:
582 case USB_DT_OTHER_SPEED_CONFIG:
583 if (config->bLength != 9) {
584 ERROR(tdev, "bogus config descriptor length\n");
585 return 0;
586 }
587 /* this bit 'must be 1' but often isn't */
588 if (!realworld && !(config->bmAttributes & 0x80)) {
589 ERROR(tdev, "high bit of config attributes not set\n");
590 return 0;
591 }
592 if (config->bmAttributes & 0x1f) { /* reserved == 0 */
593 ERROR(tdev, "reserved config bits set\n");
594 return 0;
595 }
596 break;
597 default:
598 return 0;
599 }
600
601 if (le16_to_cpu(config->wTotalLength) == len) /* read it all */
602 return 1;
603 if (le16_to_cpu(config->wTotalLength) >= TBUF_SIZE) /* max partial read */
604 return 1;
605 ERROR(tdev, "bogus config descriptor read size\n");
606 return 0;
607}
608
609static int is_good_ext(struct usbtest_dev *tdev, u8 *buf)
610{
611 struct usb_ext_cap_descriptor *ext;
612 u32 attr;
613
614 ext = (struct usb_ext_cap_descriptor *) buf;
615
616 if (ext->bLength != USB_DT_USB_EXT_CAP_SIZE) {
617 ERROR(tdev, "bogus usb 2.0 extension descriptor length\n");
618 return 0;
619 }
620
621 attr = le32_to_cpu(ext->bmAttributes);
622 /* bits[1:4] is used and others are reserved */
623 if (attr & ~0x1e) { /* reserved == 0 */
624 ERROR(tdev, "reserved bits set\n");
625 return 0;
626 }
627
628 return 1;
629}
630
631static int is_good_ss_cap(struct usbtest_dev *tdev, u8 *buf)
632{
633 struct usb_ss_cap_descriptor *ss;
634
635 ss = (struct usb_ss_cap_descriptor *) buf;
636
637 if (ss->bLength != USB_DT_USB_SS_CAP_SIZE) {
638 ERROR(tdev, "bogus superspeed device capability descriptor length\n");
639 return 0;
640 }
641
642 /*
643 * only bit[1] of bmAttributes is used for LTM and others are
644 * reserved
645 */
646 if (ss->bmAttributes & ~0x02) { /* reserved == 0 */
647 ERROR(tdev, "reserved bits set in bmAttributes\n");
648 return 0;
649 }
650
651 /* bits[0:3] of wSpeedSupported is used and others are reserved */
652 if (le16_to_cpu(ss->wSpeedSupported) & ~0x0f) { /* reserved == 0 */
653 ERROR(tdev, "reserved bits set in wSpeedSupported\n");
654 return 0;
655 }
656
657 return 1;
658}
659
660static int is_good_con_id(struct usbtest_dev *tdev, u8 *buf)
661{
662 struct usb_ss_container_id_descriptor *con_id;
663
664 con_id = (struct usb_ss_container_id_descriptor *) buf;
665
666 if (con_id->bLength != USB_DT_USB_SS_CONTN_ID_SIZE) {
667 ERROR(tdev, "bogus container id descriptor length\n");
668 return 0;
669 }
670
671 if (con_id->bReserved) { /* reserved == 0 */
672 ERROR(tdev, "reserved bits set\n");
673 return 0;
674 }
675
676 return 1;
677}
678
679/* sanity test for standard requests working with usb_control_mesg() and some
680 * of the utility functions which use it.
681 *
682 * this doesn't test how endpoint halts behave or data toggles get set, since
683 * we won't do I/O to bulk/interrupt endpoints here (which is how to change
684 * halt or toggle). toggle testing is impractical without support from hcds.
685 *
686 * this avoids failing devices linux would normally work with, by not testing
687 * config/altsetting operations for devices that only support their defaults.
688 * such devices rarely support those needless operations.
689 *
690 * NOTE that since this is a sanity test, it's not examining boundary cases
691 * to see if usbcore, hcd, and device all behave right. such testing would
692 * involve varied read sizes and other operation sequences.
693 */
694static int ch9_postconfig(struct usbtest_dev *dev)
695{
696 struct usb_interface *iface = dev->intf;
697 struct usb_device *udev = interface_to_usbdev(iface);
698 int i, alt, retval;
699
700 /* [9.2.3] if there's more than one altsetting, we need to be able to
701 * set and get each one. mostly trusts the descriptors from usbcore.
702 */
703 for (i = 0; i < iface->num_altsetting; i++) {
704
705 /* 9.2.3 constrains the range here */
706 alt = iface->altsetting[i].desc.bAlternateSetting;
707 if (alt < 0 || alt >= iface->num_altsetting) {
708 dev_err(&iface->dev,
709 "invalid alt [%d].bAltSetting = %d\n",
710 i, alt);
711 }
712
713 /* [real world] get/set unimplemented if there's only one */
714 if (realworld && iface->num_altsetting == 1)
715 continue;
716
717 /* [9.4.10] set_interface */
718 retval = set_altsetting(dev, alt);
719 if (retval) {
720 dev_err(&iface->dev, "can't set_interface = %d, %d\n",
721 alt, retval);
722 return retval;
723 }
724
725 /* [9.4.4] get_interface always works */
726 retval = get_altsetting(dev);
727 if (retval != alt) {
728 dev_err(&iface->dev, "get alt should be %d, was %d\n",
729 alt, retval);
730 return (retval < 0) ? retval : -EDOM;
731 }
732
733 }
734
735 /* [real world] get_config unimplemented if there's only one */
736 if (!realworld || udev->descriptor.bNumConfigurations != 1) {
737 int expected = udev->actconfig->desc.bConfigurationValue;
738
739 /* [9.4.2] get_configuration always works
740 * ... although some cheap devices (like one TI Hub I've got)
741 * won't return config descriptors except before set_config.
742 */
743 retval = usb_control_msg(udev, usb_rcvctrlpipe(udev, 0),
744 USB_REQ_GET_CONFIGURATION,
745 USB_DIR_IN | USB_RECIP_DEVICE,
746 0, 0, dev->buf, 1, USB_CTRL_GET_TIMEOUT);
747 if (retval != 1 || dev->buf[0] != expected) {
748 dev_err(&iface->dev, "get config --> %d %d (1 %d)\n",
749 retval, dev->buf[0], expected);
750 return (retval < 0) ? retval : -EDOM;
751 }
752 }
753
754 /* there's always [9.4.3] a device descriptor [9.6.1] */
755 retval = usb_get_descriptor(udev, USB_DT_DEVICE, 0,
756 dev->buf, sizeof(udev->descriptor));
757 if (retval != sizeof(udev->descriptor)) {
758 dev_err(&iface->dev, "dev descriptor --> %d\n", retval);
759 return (retval < 0) ? retval : -EDOM;
760 }
761
762 /*
763 * there's always [9.4.3] a bos device descriptor [9.6.2] in USB
764 * 3.0 spec
765 */
766 if (le16_to_cpu(udev->descriptor.bcdUSB) >= 0x0300) {
767 struct usb_bos_descriptor *bos = NULL;
768 struct usb_dev_cap_header *header = NULL;
769 unsigned total, num, length;
770 u8 *buf;
771
772 retval = usb_get_descriptor(udev, USB_DT_BOS, 0, dev->buf,
773 sizeof(*udev->bos->desc));
774 if (retval != sizeof(*udev->bos->desc)) {
775 dev_err(&iface->dev, "bos descriptor --> %d\n", retval);
776 return (retval < 0) ? retval : -EDOM;
777 }
778
779 bos = (struct usb_bos_descriptor *)dev->buf;
780 total = le16_to_cpu(bos->wTotalLength);
781 num = bos->bNumDeviceCaps;
782
783 if (total > TBUF_SIZE)
784 total = TBUF_SIZE;
785
786 /*
787 * get generic device-level capability descriptors [9.6.2]
788 * in USB 3.0 spec
789 */
790 retval = usb_get_descriptor(udev, USB_DT_BOS, 0, dev->buf,
791 total);
792 if (retval != total) {
793 dev_err(&iface->dev, "bos descriptor set --> %d\n",
794 retval);
795 return (retval < 0) ? retval : -EDOM;
796 }
797
798 length = sizeof(*udev->bos->desc);
799 buf = dev->buf;
800 for (i = 0; i < num; i++) {
801 buf += length;
802 if (buf + sizeof(struct usb_dev_cap_header) >
803 dev->buf + total)
804 break;
805
806 header = (struct usb_dev_cap_header *)buf;
807 length = header->bLength;
808
809 if (header->bDescriptorType !=
810 USB_DT_DEVICE_CAPABILITY) {
811 dev_warn(&udev->dev, "not device capability descriptor, skip\n");
812 continue;
813 }
814
815 switch (header->bDevCapabilityType) {
816 case USB_CAP_TYPE_EXT:
817 if (buf + USB_DT_USB_EXT_CAP_SIZE >
818 dev->buf + total ||
819 !is_good_ext(dev, buf)) {
820 dev_err(&iface->dev, "bogus usb 2.0 extension descriptor\n");
821 return -EDOM;
822 }
823 break;
824 case USB_SS_CAP_TYPE:
825 if (buf + USB_DT_USB_SS_CAP_SIZE >
826 dev->buf + total ||
827 !is_good_ss_cap(dev, buf)) {
828 dev_err(&iface->dev, "bogus superspeed device capability descriptor\n");
829 return -EDOM;
830 }
831 break;
832 case CONTAINER_ID_TYPE:
833 if (buf + USB_DT_USB_SS_CONTN_ID_SIZE >
834 dev->buf + total ||
835 !is_good_con_id(dev, buf)) {
836 dev_err(&iface->dev, "bogus container id descriptor\n");
837 return -EDOM;
838 }
839 break;
840 default:
841 break;
842 }
843 }
844 }
845
846 /* there's always [9.4.3] at least one config descriptor [9.6.3] */
847 for (i = 0; i < udev->descriptor.bNumConfigurations; i++) {
848 retval = usb_get_descriptor(udev, USB_DT_CONFIG, i,
849 dev->buf, TBUF_SIZE);
850 if (!is_good_config(dev, retval)) {
851 dev_err(&iface->dev,
852 "config [%d] descriptor --> %d\n",
853 i, retval);
854 return (retval < 0) ? retval : -EDOM;
855 }
856
857 /* FIXME cross-checking udev->config[i] to make sure usbcore
858 * parsed it right (etc) would be good testing paranoia
859 */
860 }
861
862 /* and sometimes [9.2.6.6] speed dependent descriptors */
863 if (le16_to_cpu(udev->descriptor.bcdUSB) == 0x0200) {
864 struct usb_qualifier_descriptor *d = NULL;
865
866 /* device qualifier [9.6.2] */
867 retval = usb_get_descriptor(udev,
868 USB_DT_DEVICE_QUALIFIER, 0, dev->buf,
869 sizeof(struct usb_qualifier_descriptor));
870 if (retval == -EPIPE) {
871 if (udev->speed == USB_SPEED_HIGH) {
872 dev_err(&iface->dev,
873 "hs dev qualifier --> %d\n",
874 retval);
875 return (retval < 0) ? retval : -EDOM;
876 }
877 /* usb2.0 but not high-speed capable; fine */
878 } else if (retval != sizeof(struct usb_qualifier_descriptor)) {
879 dev_err(&iface->dev, "dev qualifier --> %d\n", retval);
880 return (retval < 0) ? retval : -EDOM;
881 } else
882 d = (struct usb_qualifier_descriptor *) dev->buf;
883
884 /* might not have [9.6.2] any other-speed configs [9.6.4] */
885 if (d) {
886 unsigned max = d->bNumConfigurations;
887 for (i = 0; i < max; i++) {
888 retval = usb_get_descriptor(udev,
889 USB_DT_OTHER_SPEED_CONFIG, i,
890 dev->buf, TBUF_SIZE);
891 if (!is_good_config(dev, retval)) {
892 dev_err(&iface->dev,
893 "other speed config --> %d\n",
894 retval);
895 return (retval < 0) ? retval : -EDOM;
896 }
897 }
898 }
899 }
900 /* FIXME fetch strings from at least the device descriptor */
901
902 /* [9.4.5] get_status always works */
903 retval = usb_get_status(udev, USB_RECIP_DEVICE, 0, dev->buf);
904 if (retval) {
905 dev_err(&iface->dev, "get dev status --> %d\n", retval);
906 return retval;
907 }
908
909 /* FIXME configuration.bmAttributes says if we could try to set/clear
910 * the device's remote wakeup feature ... if we can, test that here
911 */
912
913 retval = usb_get_status(udev, USB_RECIP_INTERFACE,
914 iface->altsetting[0].desc.bInterfaceNumber, dev->buf);
915 if (retval) {
916 dev_err(&iface->dev, "get interface status --> %d\n", retval);
917 return retval;
918 }
919 /* FIXME get status for each endpoint in the interface */
920
921 return 0;
922}
923
924/*-------------------------------------------------------------------------*/
925
926/* use ch9 requests to test whether:
927 * (a) queues work for control, keeping N subtests queued and
928 * active (auto-resubmit) for M loops through the queue.
929 * (b) protocol stalls (control-only) will autorecover.
930 * it's not like bulk/intr; no halt clearing.
931 * (c) short control reads are reported and handled.
932 * (d) queues are always processed in-order
933 */
934
935struct ctrl_ctx {
936 spinlock_t lock;
937 struct usbtest_dev *dev;
938 struct completion complete;
939 unsigned count;
940 unsigned pending;
941 int status;
942 struct urb **urb;
943 struct usbtest_param *param;
944 int last;
945};
946
947#define NUM_SUBCASES 15 /* how many test subcases here? */
948
949struct subcase {
950 struct usb_ctrlrequest setup;
951 int number;
952 int expected;
953};
954
955static void ctrl_complete(struct urb *urb)
956{
957 struct ctrl_ctx *ctx = urb->context;
958 struct usb_ctrlrequest *reqp;
959 struct subcase *subcase;
960 int status = urb->status;
961
962 reqp = (struct usb_ctrlrequest *)urb->setup_packet;
963 subcase = container_of(reqp, struct subcase, setup);
964
965 spin_lock(&ctx->lock);
966 ctx->count--;
967 ctx->pending--;
968
969 /* queue must transfer and complete in fifo order, unless
970 * usb_unlink_urb() is used to unlink something not at the
971 * physical queue head (not tested).
972 */
973 if (subcase->number > 0) {
974 if ((subcase->number - ctx->last) != 1) {
975 ERROR(ctx->dev,
976 "subcase %d completed out of order, last %d\n",
977 subcase->number, ctx->last);
978 status = -EDOM;
979 ctx->last = subcase->number;
980 goto error;
981 }
982 }
983 ctx->last = subcase->number;
984
985 /* succeed or fault in only one way? */
986 if (status == subcase->expected)
987 status = 0;
988
989 /* async unlink for cleanup? */
990 else if (status != -ECONNRESET) {
991
992 /* some faults are allowed, not required */
993 if (subcase->expected > 0 && (
994 ((status == -subcase->expected /* happened */
995 || status == 0)))) /* didn't */
996 status = 0;
997 /* sometimes more than one fault is allowed */
998 else if (subcase->number == 12 && status == -EPIPE)
999 status = 0;
1000 else
1001 ERROR(ctx->dev, "subtest %d error, status %d\n",
1002 subcase->number, status);
1003 }
1004
1005 /* unexpected status codes mean errors; ideally, in hardware */
1006 if (status) {
1007error:
1008 if (ctx->status == 0) {
1009 int i;
1010
1011 ctx->status = status;
1012 ERROR(ctx->dev, "control queue %02x.%02x, err %d, "
1013 "%d left, subcase %d, len %d/%d\n",
1014 reqp->bRequestType, reqp->bRequest,
1015 status, ctx->count, subcase->number,
1016 urb->actual_length,
1017 urb->transfer_buffer_length);
1018
1019 /* FIXME this "unlink everything" exit route should
1020 * be a separate test case.
1021 */
1022
1023 /* unlink whatever's still pending */
1024 for (i = 1; i < ctx->param->sglen; i++) {
1025 struct urb *u = ctx->urb[
1026 (i + subcase->number)
1027 % ctx->param->sglen];
1028
1029 if (u == urb || !u->dev)
1030 continue;
1031 spin_unlock(&ctx->lock);
1032 status = usb_unlink_urb(u);
1033 spin_lock(&ctx->lock);
1034 switch (status) {
1035 case -EINPROGRESS:
1036 case -EBUSY:
1037 case -EIDRM:
1038 continue;
1039 default:
1040 ERROR(ctx->dev, "urb unlink --> %d\n",
1041 status);
1042 }
1043 }
1044 status = ctx->status;
1045 }
1046 }
1047
1048 /* resubmit if we need to, else mark this as done */
1049 if ((status == 0) && (ctx->pending < ctx->count)) {
1050 status = usb_submit_urb(urb, GFP_ATOMIC);
1051 if (status != 0) {
1052 ERROR(ctx->dev,
1053 "can't resubmit ctrl %02x.%02x, err %d\n",
1054 reqp->bRequestType, reqp->bRequest, status);
1055 urb->dev = NULL;
1056 } else
1057 ctx->pending++;
1058 } else
1059 urb->dev = NULL;
1060
1061 /* signal completion when nothing's queued */
1062 if (ctx->pending == 0)
1063 complete(&ctx->complete);
1064 spin_unlock(&ctx->lock);
1065}
1066
1067static int
1068test_ctrl_queue(struct usbtest_dev *dev, struct usbtest_param *param)
1069{
1070 struct usb_device *udev = testdev_to_usbdev(dev);
1071 struct urb **urb;
1072 struct ctrl_ctx context;
1073 int i;
1074
1075 if (param->sglen == 0 || param->iterations > UINT_MAX / param->sglen)
1076 return -EOPNOTSUPP;
1077
1078 spin_lock_init(&context.lock);
1079 context.dev = dev;
1080 init_completion(&context.complete);
1081 context.count = param->sglen * param->iterations;
1082 context.pending = 0;
1083 context.status = -ENOMEM;
1084 context.param = param;
1085 context.last = -1;
1086
1087 /* allocate and init the urbs we'll queue.
1088 * as with bulk/intr sglists, sglen is the queue depth; it also
1089 * controls which subtests run (more tests than sglen) or rerun.
1090 */
1091 urb = kcalloc(param->sglen, sizeof(struct urb *), GFP_KERNEL);
1092 if (!urb)
1093 return -ENOMEM;
1094 for (i = 0; i < param->sglen; i++) {
1095 int pipe = usb_rcvctrlpipe(udev, 0);
1096 unsigned len;
1097 struct urb *u;
1098 struct usb_ctrlrequest req;
1099 struct subcase *reqp;
1100
1101 /* sign of this variable means:
1102 * -: tested code must return this (negative) error code
1103 * +: tested code may return this (negative too) error code
1104 */
1105 int expected = 0;
1106
1107 /* requests here are mostly expected to succeed on any
1108 * device, but some are chosen to trigger protocol stalls
1109 * or short reads.
1110 */
1111 memset(&req, 0, sizeof(req));
1112 req.bRequest = USB_REQ_GET_DESCRIPTOR;
1113 req.bRequestType = USB_DIR_IN|USB_RECIP_DEVICE;
1114
1115 switch (i % NUM_SUBCASES) {
1116 case 0: /* get device descriptor */
1117 req.wValue = cpu_to_le16(USB_DT_DEVICE << 8);
1118 len = sizeof(struct usb_device_descriptor);
1119 break;
1120 case 1: /* get first config descriptor (only) */
1121 req.wValue = cpu_to_le16((USB_DT_CONFIG << 8) | 0);
1122 len = sizeof(struct usb_config_descriptor);
1123 break;
1124 case 2: /* get altsetting (OFTEN STALLS) */
1125 req.bRequest = USB_REQ_GET_INTERFACE;
1126 req.bRequestType = USB_DIR_IN|USB_RECIP_INTERFACE;
1127 /* index = 0 means first interface */
1128 len = 1;
1129 expected = EPIPE;
1130 break;
1131 case 3: /* get interface status */
1132 req.bRequest = USB_REQ_GET_STATUS;
1133 req.bRequestType = USB_DIR_IN|USB_RECIP_INTERFACE;
1134 /* interface 0 */
1135 len = 2;
1136 break;
1137 case 4: /* get device status */
1138 req.bRequest = USB_REQ_GET_STATUS;
1139 req.bRequestType = USB_DIR_IN|USB_RECIP_DEVICE;
1140 len = 2;
1141 break;
1142 case 5: /* get device qualifier (MAY STALL) */
1143 req.wValue = cpu_to_le16 (USB_DT_DEVICE_QUALIFIER << 8);
1144 len = sizeof(struct usb_qualifier_descriptor);
1145 if (udev->speed != USB_SPEED_HIGH)
1146 expected = EPIPE;
1147 break;
1148 case 6: /* get first config descriptor, plus interface */
1149 req.wValue = cpu_to_le16((USB_DT_CONFIG << 8) | 0);
1150 len = sizeof(struct usb_config_descriptor);
1151 len += sizeof(struct usb_interface_descriptor);
1152 break;
1153 case 7: /* get interface descriptor (ALWAYS STALLS) */
1154 req.wValue = cpu_to_le16 (USB_DT_INTERFACE << 8);
1155 /* interface == 0 */
1156 len = sizeof(struct usb_interface_descriptor);
1157 expected = -EPIPE;
1158 break;
1159 /* NOTE: two consecutive stalls in the queue here.
1160 * that tests fault recovery a bit more aggressively. */
1161 case 8: /* clear endpoint halt (MAY STALL) */
1162 req.bRequest = USB_REQ_CLEAR_FEATURE;
1163 req.bRequestType = USB_RECIP_ENDPOINT;
1164 /* wValue 0 == ep halt */
1165 /* wIndex 0 == ep0 (shouldn't halt!) */
1166 len = 0;
1167 pipe = usb_sndctrlpipe(udev, 0);
1168 expected = EPIPE;
1169 break;
1170 case 9: /* get endpoint status */
1171 req.bRequest = USB_REQ_GET_STATUS;
1172 req.bRequestType = USB_DIR_IN|USB_RECIP_ENDPOINT;
1173 /* endpoint 0 */
1174 len = 2;
1175 break;
1176 case 10: /* trigger short read (EREMOTEIO) */
1177 req.wValue = cpu_to_le16((USB_DT_CONFIG << 8) | 0);
1178 len = 1024;
1179 expected = -EREMOTEIO;
1180 break;
1181 /* NOTE: two consecutive _different_ faults in the queue. */
1182 case 11: /* get endpoint descriptor (ALWAYS STALLS) */
1183 req.wValue = cpu_to_le16(USB_DT_ENDPOINT << 8);
1184 /* endpoint == 0 */
1185 len = sizeof(struct usb_interface_descriptor);
1186 expected = EPIPE;
1187 break;
1188 /* NOTE: sometimes even a third fault in the queue! */
1189 case 12: /* get string 0 descriptor (MAY STALL) */
1190 req.wValue = cpu_to_le16(USB_DT_STRING << 8);
1191 /* string == 0, for language IDs */
1192 len = sizeof(struct usb_interface_descriptor);
1193 /* may succeed when > 4 languages */
1194 expected = EREMOTEIO; /* or EPIPE, if no strings */
1195 break;
1196 case 13: /* short read, resembling case 10 */
1197 req.wValue = cpu_to_le16((USB_DT_CONFIG << 8) | 0);
1198 /* last data packet "should" be DATA1, not DATA0 */
1199 if (udev->speed == USB_SPEED_SUPER)
1200 len = 1024 - 512;
1201 else
1202 len = 1024 - udev->descriptor.bMaxPacketSize0;
1203 expected = -EREMOTEIO;
1204 break;
1205 case 14: /* short read; try to fill the last packet */
1206 req.wValue = cpu_to_le16((USB_DT_DEVICE << 8) | 0);
1207 /* device descriptor size == 18 bytes */
1208 len = udev->descriptor.bMaxPacketSize0;
1209 if (udev->speed == USB_SPEED_SUPER)
1210 len = 512;
1211 switch (len) {
1212 case 8:
1213 len = 24;
1214 break;
1215 case 16:
1216 len = 32;
1217 break;
1218 }
1219 expected = -EREMOTEIO;
1220 break;
1221 default:
1222 ERROR(dev, "bogus number of ctrl queue testcases!\n");
1223 context.status = -EINVAL;
1224 goto cleanup;
1225 }
1226 req.wLength = cpu_to_le16(len);
1227 urb[i] = u = simple_alloc_urb(udev, pipe, len);
1228 if (!u)
1229 goto cleanup;
1230
1231 reqp = kmalloc(sizeof(*reqp), GFP_KERNEL);
1232 if (!reqp)
1233 goto cleanup;
1234 reqp->setup = req;
1235 reqp->number = i % NUM_SUBCASES;
1236 reqp->expected = expected;
1237 u->setup_packet = (char *) &reqp->setup;
1238
1239 u->context = &context;
1240 u->complete = ctrl_complete;
1241 }
1242
1243 /* queue the urbs */
1244 context.urb = urb;
1245 spin_lock_irq(&context.lock);
1246 for (i = 0; i < param->sglen; i++) {
1247 context.status = usb_submit_urb(urb[i], GFP_ATOMIC);
1248 if (context.status != 0) {
1249 ERROR(dev, "can't submit urb[%d], status %d\n",
1250 i, context.status);
1251 context.count = context.pending;
1252 break;
1253 }
1254 context.pending++;
1255 }
1256 spin_unlock_irq(&context.lock);
1257
1258 /* FIXME set timer and time out; provide a disconnect hook */
1259
1260 /* wait for the last one to complete */
1261 if (context.pending > 0)
1262 wait_for_completion(&context.complete);
1263
1264cleanup:
1265 for (i = 0; i < param->sglen; i++) {
1266 if (!urb[i])
1267 continue;
1268 urb[i]->dev = udev;
1269 kfree(urb[i]->setup_packet);
1270 simple_free_urb(urb[i]);
1271 }
1272 kfree(urb);
1273 return context.status;
1274}
1275#undef NUM_SUBCASES
1276
1277
1278/*-------------------------------------------------------------------------*/
1279
1280static void unlink1_callback(struct urb *urb)
1281{
1282 int status = urb->status;
1283
1284 /* we "know" -EPIPE (stall) never happens */
1285 if (!status)
1286 status = usb_submit_urb(urb, GFP_ATOMIC);
1287 if (status) {
1288 urb->status = status;
1289 complete(urb->context);
1290 }
1291}
1292
1293static int unlink1(struct usbtest_dev *dev, int pipe, int size, int async)
1294{
1295 struct urb *urb;
1296 struct completion completion;
1297 int retval = 0;
1298
1299 init_completion(&completion);
1300 urb = simple_alloc_urb(testdev_to_usbdev(dev), pipe, size);
1301 if (!urb)
1302 return -ENOMEM;
1303 urb->context = &completion;
1304 urb->complete = unlink1_callback;
1305
1306 /* keep the endpoint busy. there are lots of hc/hcd-internal
1307 * states, and testing should get to all of them over time.
1308 *
1309 * FIXME want additional tests for when endpoint is STALLing
1310 * due to errors, or is just NAKing requests.
1311 */
1312 retval = usb_submit_urb(urb, GFP_KERNEL);
1313 if (retval != 0) {
1314 dev_err(&dev->intf->dev, "submit fail %d\n", retval);
1315 return retval;
1316 }
1317
1318 /* unlinking that should always work. variable delay tests more
1319 * hcd states and code paths, even with little other system load.
1320 */
1321 msleep(jiffies % (2 * INTERRUPT_RATE));
1322 if (async) {
1323 while (!completion_done(&completion)) {
1324 retval = usb_unlink_urb(urb);
1325
1326 switch (retval) {
1327 case -EBUSY:
1328 case -EIDRM:
1329 /* we can't unlink urbs while they're completing
1330 * or if they've completed, and we haven't
1331 * resubmitted. "normal" drivers would prevent
1332 * resubmission, but since we're testing unlink
1333 * paths, we can't.
1334 */
1335 ERROR(dev, "unlink retry\n");
1336 continue;
1337 case 0:
1338 case -EINPROGRESS:
1339 break;
1340
1341 default:
1342 dev_err(&dev->intf->dev,
1343 "unlink fail %d\n", retval);
1344 return retval;
1345 }
1346
1347 break;
1348 }
1349 } else
1350 usb_kill_urb(urb);
1351
1352 wait_for_completion(&completion);
1353 retval = urb->status;
1354 simple_free_urb(urb);
1355
1356 if (async)
1357 return (retval == -ECONNRESET) ? 0 : retval - 1000;
1358 else
1359 return (retval == -ENOENT || retval == -EPERM) ?
1360 0 : retval - 2000;
1361}
1362
1363static int unlink_simple(struct usbtest_dev *dev, int pipe, int len)
1364{
1365 int retval = 0;
1366
1367 /* test sync and async paths */
1368 retval = unlink1(dev, pipe, len, 1);
1369 if (!retval)
1370 retval = unlink1(dev, pipe, len, 0);
1371 return retval;
1372}
1373
1374/*-------------------------------------------------------------------------*/
1375
1376struct queued_ctx {
1377 struct completion complete;
1378 atomic_t pending;
1379 unsigned num;
1380 int status;
1381 struct urb **urbs;
1382};
1383
1384static void unlink_queued_callback(struct urb *urb)
1385{
1386 int status = urb->status;
1387 struct queued_ctx *ctx = urb->context;
1388
1389 if (ctx->status)
1390 goto done;
1391 if (urb == ctx->urbs[ctx->num - 4] || urb == ctx->urbs[ctx->num - 2]) {
1392 if (status == -ECONNRESET)
1393 goto done;
1394 /* What error should we report if the URB completed normally? */
1395 }
1396 if (status != 0)
1397 ctx->status = status;
1398
1399 done:
1400 if (atomic_dec_and_test(&ctx->pending))
1401 complete(&ctx->complete);
1402}
1403
1404static int unlink_queued(struct usbtest_dev *dev, int pipe, unsigned num,
1405 unsigned size)
1406{
1407 struct queued_ctx ctx;
1408 struct usb_device *udev = testdev_to_usbdev(dev);
1409 void *buf;
1410 dma_addr_t buf_dma;
1411 int i;
1412 int retval = -ENOMEM;
1413
1414 init_completion(&ctx.complete);
1415 atomic_set(&ctx.pending, 1); /* One more than the actual value */
1416 ctx.num = num;
1417 ctx.status = 0;
1418
1419 buf = usb_alloc_coherent(udev, size, GFP_KERNEL, &buf_dma);
1420 if (!buf)
1421 return retval;
1422 memset(buf, 0, size);
1423
1424 /* Allocate and init the urbs we'll queue */
1425 ctx.urbs = kcalloc(num, sizeof(struct urb *), GFP_KERNEL);
1426 if (!ctx.urbs)
1427 goto free_buf;
1428 for (i = 0; i < num; i++) {
1429 ctx.urbs[i] = usb_alloc_urb(0, GFP_KERNEL);
1430 if (!ctx.urbs[i])
1431 goto free_urbs;
1432 usb_fill_bulk_urb(ctx.urbs[i], udev, pipe, buf, size,
1433 unlink_queued_callback, &ctx);
1434 ctx.urbs[i]->transfer_dma = buf_dma;
1435 ctx.urbs[i]->transfer_flags = URB_NO_TRANSFER_DMA_MAP;
1436 }
1437
1438 /* Submit all the URBs and then unlink URBs num - 4 and num - 2. */
1439 for (i = 0; i < num; i++) {
1440 atomic_inc(&ctx.pending);
1441 retval = usb_submit_urb(ctx.urbs[i], GFP_KERNEL);
1442 if (retval != 0) {
1443 dev_err(&dev->intf->dev, "submit urbs[%d] fail %d\n",
1444 i, retval);
1445 atomic_dec(&ctx.pending);
1446 ctx.status = retval;
1447 break;
1448 }
1449 }
1450 if (i == num) {
1451 usb_unlink_urb(ctx.urbs[num - 4]);
1452 usb_unlink_urb(ctx.urbs[num - 2]);
1453 } else {
1454 while (--i >= 0)
1455 usb_unlink_urb(ctx.urbs[i]);
1456 }
1457
1458 if (atomic_dec_and_test(&ctx.pending)) /* The extra count */
1459 complete(&ctx.complete);
1460 wait_for_completion(&ctx.complete);
1461 retval = ctx.status;
1462
1463 free_urbs:
1464 for (i = 0; i < num; i++)
1465 usb_free_urb(ctx.urbs[i]);
1466 kfree(ctx.urbs);
1467 free_buf:
1468 usb_free_coherent(udev, size, buf, buf_dma);
1469 return retval;
1470}
1471
1472/*-------------------------------------------------------------------------*/
1473
1474static int verify_not_halted(struct usbtest_dev *tdev, int ep, struct urb *urb)
1475{
1476 int retval;
1477 u16 status;
1478
1479 /* shouldn't look or act halted */
1480 retval = usb_get_status(urb->dev, USB_RECIP_ENDPOINT, ep, &status);
1481 if (retval < 0) {
1482 ERROR(tdev, "ep %02x couldn't get no-halt status, %d\n",
1483 ep, retval);
1484 return retval;
1485 }
1486 if (status != 0) {
1487 ERROR(tdev, "ep %02x bogus status: %04x != 0\n", ep, status);
1488 return -EINVAL;
1489 }
1490 retval = simple_io(tdev, urb, 1, 0, 0, __func__);
1491 if (retval != 0)
1492 return -EINVAL;
1493 return 0;
1494}
1495
1496static int verify_halted(struct usbtest_dev *tdev, int ep, struct urb *urb)
1497{
1498 int retval;
1499 u16 status;
1500
1501 /* should look and act halted */
1502 retval = usb_get_status(urb->dev, USB_RECIP_ENDPOINT, ep, &status);
1503 if (retval < 0) {
1504 ERROR(tdev, "ep %02x couldn't get halt status, %d\n",
1505 ep, retval);
1506 return retval;
1507 }
1508 if (status != 1) {
1509 ERROR(tdev, "ep %02x bogus status: %04x != 1\n", ep, status);
1510 return -EINVAL;
1511 }
1512 retval = simple_io(tdev, urb, 1, 0, -EPIPE, __func__);
1513 if (retval != -EPIPE)
1514 return -EINVAL;
1515 retval = simple_io(tdev, urb, 1, 0, -EPIPE, "verify_still_halted");
1516 if (retval != -EPIPE)
1517 return -EINVAL;
1518 return 0;
1519}
1520
1521static int test_halt(struct usbtest_dev *tdev, int ep, struct urb *urb)
1522{
1523 int retval;
1524
1525 /* shouldn't look or act halted now */
1526 retval = verify_not_halted(tdev, ep, urb);
1527 if (retval < 0)
1528 return retval;
1529
1530 /* set halt (protocol test only), verify it worked */
1531 retval = usb_control_msg(urb->dev, usb_sndctrlpipe(urb->dev, 0),
1532 USB_REQ_SET_FEATURE, USB_RECIP_ENDPOINT,
1533 USB_ENDPOINT_HALT, ep,
1534 NULL, 0, USB_CTRL_SET_TIMEOUT);
1535 if (retval < 0) {
1536 ERROR(tdev, "ep %02x couldn't set halt, %d\n", ep, retval);
1537 return retval;
1538 }
1539 retval = verify_halted(tdev, ep, urb);
1540 if (retval < 0)
1541 return retval;
1542
1543 /* clear halt (tests API + protocol), verify it worked */
1544 retval = usb_clear_halt(urb->dev, urb->pipe);
1545 if (retval < 0) {
1546 ERROR(tdev, "ep %02x couldn't clear halt, %d\n", ep, retval);
1547 return retval;
1548 }
1549 retval = verify_not_halted(tdev, ep, urb);
1550 if (retval < 0)
1551 return retval;
1552
1553 /* NOTE: could also verify SET_INTERFACE clear halts ... */
1554
1555 return 0;
1556}
1557
1558static int halt_simple(struct usbtest_dev *dev)
1559{
1560 int ep;
1561 int retval = 0;
1562 struct urb *urb;
1563 struct usb_device *udev = testdev_to_usbdev(dev);
1564
1565 if (udev->speed == USB_SPEED_SUPER)
1566 urb = simple_alloc_urb(udev, 0, 1024);
1567 else
1568 urb = simple_alloc_urb(udev, 0, 512);
1569 if (urb == NULL)
1570 return -ENOMEM;
1571
1572 if (dev->in_pipe) {
1573 ep = usb_pipeendpoint(dev->in_pipe) | USB_DIR_IN;
1574 urb->pipe = dev->in_pipe;
1575 retval = test_halt(dev, ep, urb);
1576 if (retval < 0)
1577 goto done;
1578 }
1579
1580 if (dev->out_pipe) {
1581 ep = usb_pipeendpoint(dev->out_pipe);
1582 urb->pipe = dev->out_pipe;
1583 retval = test_halt(dev, ep, urb);
1584 }
1585done:
1586 simple_free_urb(urb);
1587 return retval;
1588}
1589
1590/*-------------------------------------------------------------------------*/
1591
1592/* Control OUT tests use the vendor control requests from Intel's
1593 * USB 2.0 compliance test device: write a buffer, read it back.
1594 *
1595 * Intel's spec only _requires_ that it work for one packet, which
1596 * is pretty weak. Some HCDs place limits here; most devices will
1597 * need to be able to handle more than one OUT data packet. We'll
1598 * try whatever we're told to try.
1599 */
1600static int ctrl_out(struct usbtest_dev *dev,
1601 unsigned count, unsigned length, unsigned vary, unsigned offset)
1602{
1603 unsigned i, j, len;
1604 int retval;
1605 u8 *buf;
1606 char *what = "?";
1607 struct usb_device *udev;
1608
1609 if (length < 1 || length > 0xffff || vary >= length)
1610 return -EINVAL;
1611
1612 buf = kmalloc(length + offset, GFP_KERNEL);
1613 if (!buf)
1614 return -ENOMEM;
1615
1616 buf += offset;
1617 udev = testdev_to_usbdev(dev);
1618 len = length;
1619 retval = 0;
1620
1621 /* NOTE: hardware might well act differently if we pushed it
1622 * with lots back-to-back queued requests.
1623 */
1624 for (i = 0; i < count; i++) {
1625 /* write patterned data */
1626 for (j = 0; j < len; j++)
1627 buf[j] = i + j;
1628 retval = usb_control_msg(udev, usb_sndctrlpipe(udev, 0),
1629 0x5b, USB_DIR_OUT|USB_TYPE_VENDOR,
1630 0, 0, buf, len, USB_CTRL_SET_TIMEOUT);
1631 if (retval != len) {
1632 what = "write";
1633 if (retval >= 0) {
1634 ERROR(dev, "ctrl_out, wlen %d (expected %d)\n",
1635 retval, len);
1636 retval = -EBADMSG;
1637 }
1638 break;
1639 }
1640
1641 /* read it back -- assuming nothing intervened!! */
1642 retval = usb_control_msg(udev, usb_rcvctrlpipe(udev, 0),
1643 0x5c, USB_DIR_IN|USB_TYPE_VENDOR,
1644 0, 0, buf, len, USB_CTRL_GET_TIMEOUT);
1645 if (retval != len) {
1646 what = "read";
1647 if (retval >= 0) {
1648 ERROR(dev, "ctrl_out, rlen %d (expected %d)\n",
1649 retval, len);
1650 retval = -EBADMSG;
1651 }
1652 break;
1653 }
1654
1655 /* fail if we can't verify */
1656 for (j = 0; j < len; j++) {
1657 if (buf[j] != (u8) (i + j)) {
1658 ERROR(dev, "ctrl_out, byte %d is %d not %d\n",
1659 j, buf[j], (u8) i + j);
1660 retval = -EBADMSG;
1661 break;
1662 }
1663 }
1664 if (retval < 0) {
1665 what = "verify";
1666 break;
1667 }
1668
1669 len += vary;
1670
1671 /* [real world] the "zero bytes IN" case isn't really used.
1672 * hardware can easily trip up in this weird case, since its
1673 * status stage is IN, not OUT like other ep0in transfers.
1674 */
1675 if (len > length)
1676 len = realworld ? 1 : 0;
1677 }
1678
1679 if (retval < 0)
1680 ERROR(dev, "ctrl_out %s failed, code %d, count %d\n",
1681 what, retval, i);
1682
1683 kfree(buf - offset);
1684 return retval;
1685}
1686
1687/*-------------------------------------------------------------------------*/
1688
1689/* ISO tests ... mimics common usage
1690 * - buffer length is split into N packets (mostly maxpacket sized)
1691 * - multi-buffers according to sglen
1692 */
1693
1694struct iso_context {
1695 unsigned count;
1696 unsigned pending;
1697 spinlock_t lock;
1698 struct completion done;
1699 int submit_error;
1700 unsigned long errors;
1701 unsigned long packet_count;
1702 struct usbtest_dev *dev;
1703};
1704
1705static void iso_callback(struct urb *urb)
1706{
1707 struct iso_context *ctx = urb->context;
1708
1709 spin_lock(&ctx->lock);
1710 ctx->count--;
1711
1712 ctx->packet_count += urb->number_of_packets;
1713 if (urb->error_count > 0)
1714 ctx->errors += urb->error_count;
1715 else if (urb->status != 0)
1716 ctx->errors += urb->number_of_packets;
1717 else if (urb->actual_length != urb->transfer_buffer_length)
1718 ctx->errors++;
1719 else if (check_guard_bytes(ctx->dev, urb) != 0)
1720 ctx->errors++;
1721
1722 if (urb->status == 0 && ctx->count > (ctx->pending - 1)
1723 && !ctx->submit_error) {
1724 int status = usb_submit_urb(urb, GFP_ATOMIC);
1725 switch (status) {
1726 case 0:
1727 goto done;
1728 default:
1729 dev_err(&ctx->dev->intf->dev,
1730 "iso resubmit err %d\n",
1731 status);
1732 /* FALLTHROUGH */
1733 case -ENODEV: /* disconnected */
1734 case -ESHUTDOWN: /* endpoint disabled */
1735 ctx->submit_error = 1;
1736 break;
1737 }
1738 }
1739
1740 ctx->pending--;
1741 if (ctx->pending == 0) {
1742 if (ctx->errors)
1743 dev_err(&ctx->dev->intf->dev,
1744 "iso test, %lu errors out of %lu\n",
1745 ctx->errors, ctx->packet_count);
1746 complete(&ctx->done);
1747 }
1748done:
1749 spin_unlock(&ctx->lock);
1750}
1751
1752static struct urb *iso_alloc_urb(
1753 struct usb_device *udev,
1754 int pipe,
1755 struct usb_endpoint_descriptor *desc,
1756 long bytes,
1757 unsigned offset
1758)
1759{
1760 struct urb *urb;
1761 unsigned i, maxp, packets;
1762
1763 if (bytes < 0 || !desc)
1764 return NULL;
1765 maxp = 0x7ff & usb_endpoint_maxp(desc);
1766 maxp *= 1 + (0x3 & (usb_endpoint_maxp(desc) >> 11));
1767 packets = DIV_ROUND_UP(bytes, maxp);
1768
1769 urb = usb_alloc_urb(packets, GFP_KERNEL);
1770 if (!urb)
1771 return urb;
1772 urb->dev = udev;
1773 urb->pipe = pipe;
1774
1775 urb->number_of_packets = packets;
1776 urb->transfer_buffer_length = bytes;
1777 urb->transfer_buffer = usb_alloc_coherent(udev, bytes + offset,
1778 GFP_KERNEL,
1779 &urb->transfer_dma);
1780 if (!urb->transfer_buffer) {
1781 usb_free_urb(urb);
1782 return NULL;
1783 }
1784 if (offset) {
1785 memset(urb->transfer_buffer, GUARD_BYTE, offset);
1786 urb->transfer_buffer += offset;
1787 urb->transfer_dma += offset;
1788 }
1789 /* For inbound transfers use guard byte so that test fails if
1790 data not correctly copied */
1791 memset(urb->transfer_buffer,
1792 usb_pipein(urb->pipe) ? GUARD_BYTE : 0,
1793 bytes);
1794
1795 for (i = 0; i < packets; i++) {
1796 /* here, only the last packet will be short */
1797 urb->iso_frame_desc[i].length = min((unsigned) bytes, maxp);
1798 bytes -= urb->iso_frame_desc[i].length;
1799
1800 urb->iso_frame_desc[i].offset = maxp * i;
1801 }
1802
1803 urb->complete = iso_callback;
1804 /* urb->context = SET BY CALLER */
1805 urb->interval = 1 << (desc->bInterval - 1);
1806 urb->transfer_flags = URB_ISO_ASAP | URB_NO_TRANSFER_DMA_MAP;
1807 return urb;
1808}
1809
1810static int
1811test_iso_queue(struct usbtest_dev *dev, struct usbtest_param *param,
1812 int pipe, struct usb_endpoint_descriptor *desc, unsigned offset)
1813{
1814 struct iso_context context;
1815 struct usb_device *udev;
1816 unsigned i;
1817 unsigned long packets = 0;
1818 int status = 0;
1819 struct urb *urbs[10]; /* FIXME no limit */
1820
1821 if (param->sglen > 10)
1822 return -EDOM;
1823
1824 memset(&context, 0, sizeof(context));
1825 context.count = param->iterations * param->sglen;
1826 context.dev = dev;
1827 init_completion(&context.done);
1828 spin_lock_init(&context.lock);
1829
1830 memset(urbs, 0, sizeof(urbs));
1831 udev = testdev_to_usbdev(dev);
1832 dev_info(&dev->intf->dev,
1833 "... iso period %d %sframes, wMaxPacket %04x\n",
1834 1 << (desc->bInterval - 1),
1835 (udev->speed == USB_SPEED_HIGH) ? "micro" : "",
1836 usb_endpoint_maxp(desc));
1837
1838 for (i = 0; i < param->sglen; i++) {
1839 urbs[i] = iso_alloc_urb(udev, pipe, desc,
1840 param->length, offset);
1841 if (!urbs[i]) {
1842 status = -ENOMEM;
1843 goto fail;
1844 }
1845 packets += urbs[i]->number_of_packets;
1846 urbs[i]->context = &context;
1847 }
1848 packets *= param->iterations;
1849 dev_info(&dev->intf->dev,
1850 "... total %lu msec (%lu packets)\n",
1851 (packets * (1 << (desc->bInterval - 1)))
1852 / ((udev->speed == USB_SPEED_HIGH) ? 8 : 1),
1853 packets);
1854
1855 spin_lock_irq(&context.lock);
1856 for (i = 0; i < param->sglen; i++) {
1857 ++context.pending;
1858 status = usb_submit_urb(urbs[i], GFP_ATOMIC);
1859 if (status < 0) {
1860 ERROR(dev, "submit iso[%d], error %d\n", i, status);
1861 if (i == 0) {
1862 spin_unlock_irq(&context.lock);
1863 goto fail;
1864 }
1865
1866 simple_free_urb(urbs[i]);
1867 urbs[i] = NULL;
1868 context.pending--;
1869 context.submit_error = 1;
1870 break;
1871 }
1872 }
1873 spin_unlock_irq(&context.lock);
1874
1875 wait_for_completion(&context.done);
1876
1877 for (i = 0; i < param->sglen; i++) {
1878 if (urbs[i])
1879 simple_free_urb(urbs[i]);
1880 }
1881 /*
1882 * Isochronous transfers are expected to fail sometimes. As an
1883 * arbitrary limit, we will report an error if any submissions
1884 * fail or if the transfer failure rate is > 10%.
1885 */
1886 if (status != 0)
1887 ;
1888 else if (context.submit_error)
1889 status = -EACCES;
1890 else if (context.errors > context.packet_count / 10)
1891 status = -EIO;
1892 return status;
1893
1894fail:
1895 for (i = 0; i < param->sglen; i++) {
1896 if (urbs[i])
1897 simple_free_urb(urbs[i]);
1898 }
1899 return status;
1900}
1901
1902static int test_unaligned_bulk(
1903 struct usbtest_dev *tdev,
1904 int pipe,
1905 unsigned length,
1906 int iterations,
1907 unsigned transfer_flags,
1908 const char *label)
1909{
1910 int retval;
1911 struct urb *urb = usbtest_alloc_urb(
1912 testdev_to_usbdev(tdev), pipe, length, transfer_flags, 1);
1913
1914 if (!urb)
1915 return -ENOMEM;
1916
1917 retval = simple_io(tdev, urb, iterations, 0, 0, label);
1918 simple_free_urb(urb);
1919 return retval;
1920}
1921
1922/*-------------------------------------------------------------------------*/
1923
1924/* We only have this one interface to user space, through usbfs.
1925 * User mode code can scan usbfs to find N different devices (maybe on
1926 * different busses) to use when testing, and allocate one thread per
1927 * test. So discovery is simplified, and we have no device naming issues.
1928 *
1929 * Don't use these only as stress/load tests. Use them along with with
1930 * other USB bus activity: plugging, unplugging, mousing, mp3 playback,
1931 * video capture, and so on. Run different tests at different times, in
1932 * different sequences. Nothing here should interact with other devices,
1933 * except indirectly by consuming USB bandwidth and CPU resources for test
1934 * threads and request completion. But the only way to know that for sure
1935 * is to test when HC queues are in use by many devices.
1936 *
1937 * WARNING: Because usbfs grabs udev->dev.sem before calling this ioctl(),
1938 * it locks out usbcore in certain code paths. Notably, if you disconnect
1939 * the device-under-test, khubd will wait block forever waiting for the
1940 * ioctl to complete ... so that usb_disconnect() can abort the pending
1941 * urbs and then call usbtest_disconnect(). To abort a test, you're best
1942 * off just killing the userspace task and waiting for it to exit.
1943 */
1944
1945static int
1946usbtest_ioctl(struct usb_interface *intf, unsigned int code, void *buf)
1947{
1948 struct usbtest_dev *dev = usb_get_intfdata(intf);
1949 struct usb_device *udev = testdev_to_usbdev(dev);
1950 struct usbtest_param *param = buf;
1951 int retval = -EOPNOTSUPP;
1952 struct urb *urb;
1953 struct scatterlist *sg;
1954 struct usb_sg_request req;
1955 struct timeval start;
1956 unsigned i;
1957
1958 /* FIXME USBDEVFS_CONNECTINFO doesn't say how fast the device is. */
1959
1960 pattern = mod_pattern;
1961
1962 if (code != USBTEST_REQUEST)
1963 return -EOPNOTSUPP;
1964
1965 if (param->iterations <= 0)
1966 return -EINVAL;
1967
1968 if (mutex_lock_interruptible(&dev->lock))
1969 return -ERESTARTSYS;
1970
1971 /* FIXME: What if a system sleep starts while a test is running? */
1972
1973 /* some devices, like ez-usb default devices, need a non-default
1974 * altsetting to have any active endpoints. some tests change
1975 * altsettings; force a default so most tests don't need to check.
1976 */
1977 if (dev->info->alt >= 0) {
1978 int res;
1979
1980 if (intf->altsetting->desc.bInterfaceNumber) {
1981 mutex_unlock(&dev->lock);
1982 return -ENODEV;
1983 }
1984 res = set_altsetting(dev, dev->info->alt);
1985 if (res) {
1986 dev_err(&intf->dev,
1987 "set altsetting to %d failed, %d\n",
1988 dev->info->alt, res);
1989 mutex_unlock(&dev->lock);
1990 return res;
1991 }
1992 }
1993
1994 /*
1995 * Just a bunch of test cases that every HCD is expected to handle.
1996 *
1997 * Some may need specific firmware, though it'd be good to have
1998 * one firmware image to handle all the test cases.
1999 *
2000 * FIXME add more tests! cancel requests, verify the data, control
2001 * queueing, concurrent read+write threads, and so on.
2002 */
2003 do_gettimeofday(&start);
2004 switch (param->test_num) {
2005
2006 case 0:
2007 dev_info(&intf->dev, "TEST 0: NOP\n");
2008 retval = 0;
2009 break;
2010
2011 /* Simple non-queued bulk I/O tests */
2012 case 1:
2013 if (dev->out_pipe == 0)
2014 break;
2015 dev_info(&intf->dev,
2016 "TEST 1: write %d bytes %u times\n",
2017 param->length, param->iterations);
2018 urb = simple_alloc_urb(udev, dev->out_pipe, param->length);
2019 if (!urb) {
2020 retval = -ENOMEM;
2021 break;
2022 }
2023 /* FIRMWARE: bulk sink (maybe accepts short writes) */
2024 retval = simple_io(dev, urb, param->iterations, 0, 0, "test1");
2025 simple_free_urb(urb);
2026 break;
2027 case 2:
2028 if (dev->in_pipe == 0)
2029 break;
2030 dev_info(&intf->dev,
2031 "TEST 2: read %d bytes %u times\n",
2032 param->length, param->iterations);
2033 urb = simple_alloc_urb(udev, dev->in_pipe, param->length);
2034 if (!urb) {
2035 retval = -ENOMEM;
2036 break;
2037 }
2038 /* FIRMWARE: bulk source (maybe generates short writes) */
2039 retval = simple_io(dev, urb, param->iterations, 0, 0, "test2");
2040 simple_free_urb(urb);
2041 break;
2042 case 3:
2043 if (dev->out_pipe == 0 || param->vary == 0)
2044 break;
2045 dev_info(&intf->dev,
2046 "TEST 3: write/%d 0..%d bytes %u times\n",
2047 param->vary, param->length, param->iterations);
2048 urb = simple_alloc_urb(udev, dev->out_pipe, param->length);
2049 if (!urb) {
2050 retval = -ENOMEM;
2051 break;
2052 }
2053 /* FIRMWARE: bulk sink (maybe accepts short writes) */
2054 retval = simple_io(dev, urb, param->iterations, param->vary,
2055 0, "test3");
2056 simple_free_urb(urb);
2057 break;
2058 case 4:
2059 if (dev->in_pipe == 0 || param->vary == 0)
2060 break;
2061 dev_info(&intf->dev,
2062 "TEST 4: read/%d 0..%d bytes %u times\n",
2063 param->vary, param->length, param->iterations);
2064 urb = simple_alloc_urb(udev, dev->in_pipe, param->length);
2065 if (!urb) {
2066 retval = -ENOMEM;
2067 break;
2068 }
2069 /* FIRMWARE: bulk source (maybe generates short writes) */
2070 retval = simple_io(dev, urb, param->iterations, param->vary,
2071 0, "test4");
2072 simple_free_urb(urb);
2073 break;
2074
2075 /* Queued bulk I/O tests */
2076 case 5:
2077 if (dev->out_pipe == 0 || param->sglen == 0)
2078 break;
2079 dev_info(&intf->dev,
2080 "TEST 5: write %d sglists %d entries of %d bytes\n",
2081 param->iterations,
2082 param->sglen, param->length);
2083 sg = alloc_sglist(param->sglen, param->length, 0);
2084 if (!sg) {
2085 retval = -ENOMEM;
2086 break;
2087 }
2088 /* FIRMWARE: bulk sink (maybe accepts short writes) */
2089 retval = perform_sglist(dev, param->iterations, dev->out_pipe,
2090 &req, sg, param->sglen);
2091 free_sglist(sg, param->sglen);
2092 break;
2093
2094 case 6:
2095 if (dev->in_pipe == 0 || param->sglen == 0)
2096 break;
2097 dev_info(&intf->dev,
2098 "TEST 6: read %d sglists %d entries of %d bytes\n",
2099 param->iterations,
2100 param->sglen, param->length);
2101 sg = alloc_sglist(param->sglen, param->length, 0);
2102 if (!sg) {
2103 retval = -ENOMEM;
2104 break;
2105 }
2106 /* FIRMWARE: bulk source (maybe generates short writes) */
2107 retval = perform_sglist(dev, param->iterations, dev->in_pipe,
2108 &req, sg, param->sglen);
2109 free_sglist(sg, param->sglen);
2110 break;
2111 case 7:
2112 if (dev->out_pipe == 0 || param->sglen == 0 || param->vary == 0)
2113 break;
2114 dev_info(&intf->dev,
2115 "TEST 7: write/%d %d sglists %d entries 0..%d bytes\n",
2116 param->vary, param->iterations,
2117 param->sglen, param->length);
2118 sg = alloc_sglist(param->sglen, param->length, param->vary);
2119 if (!sg) {
2120 retval = -ENOMEM;
2121 break;
2122 }
2123 /* FIRMWARE: bulk sink (maybe accepts short writes) */
2124 retval = perform_sglist(dev, param->iterations, dev->out_pipe,
2125 &req, sg, param->sglen);
2126 free_sglist(sg, param->sglen);
2127 break;
2128 case 8:
2129 if (dev->in_pipe == 0 || param->sglen == 0 || param->vary == 0)
2130 break;
2131 dev_info(&intf->dev,
2132 "TEST 8: read/%d %d sglists %d entries 0..%d bytes\n",
2133 param->vary, param->iterations,
2134 param->sglen, param->length);
2135 sg = alloc_sglist(param->sglen, param->length, param->vary);
2136 if (!sg) {
2137 retval = -ENOMEM;
2138 break;
2139 }
2140 /* FIRMWARE: bulk source (maybe generates short writes) */
2141 retval = perform_sglist(dev, param->iterations, dev->in_pipe,
2142 &req, sg, param->sglen);
2143 free_sglist(sg, param->sglen);
2144 break;
2145
2146 /* non-queued sanity tests for control (chapter 9 subset) */
2147 case 9:
2148 retval = 0;
2149 dev_info(&intf->dev,
2150 "TEST 9: ch9 (subset) control tests, %d times\n",
2151 param->iterations);
2152 for (i = param->iterations; retval == 0 && i--; /* NOP */)
2153 retval = ch9_postconfig(dev);
2154 if (retval)
2155 dev_err(&intf->dev, "ch9 subset failed, "
2156 "iterations left %d\n", i);
2157 break;
2158
2159 /* queued control messaging */
2160 case 10:
2161 retval = 0;
2162 dev_info(&intf->dev,
2163 "TEST 10: queue %d control calls, %d times\n",
2164 param->sglen,
2165 param->iterations);
2166 retval = test_ctrl_queue(dev, param);
2167 break;
2168
2169 /* simple non-queued unlinks (ring with one urb) */
2170 case 11:
2171 if (dev->in_pipe == 0 || !param->length)
2172 break;
2173 retval = 0;
2174 dev_info(&intf->dev, "TEST 11: unlink %d reads of %d\n",
2175 param->iterations, param->length);
2176 for (i = param->iterations; retval == 0 && i--; /* NOP */)
2177 retval = unlink_simple(dev, dev->in_pipe,
2178 param->length);
2179 if (retval)
2180 dev_err(&intf->dev, "unlink reads failed %d, "
2181 "iterations left %d\n", retval, i);
2182 break;
2183 case 12:
2184 if (dev->out_pipe == 0 || !param->length)
2185 break;
2186 retval = 0;
2187 dev_info(&intf->dev, "TEST 12: unlink %d writes of %d\n",
2188 param->iterations, param->length);
2189 for (i = param->iterations; retval == 0 && i--; /* NOP */)
2190 retval = unlink_simple(dev, dev->out_pipe,
2191 param->length);
2192 if (retval)
2193 dev_err(&intf->dev, "unlink writes failed %d, "
2194 "iterations left %d\n", retval, i);
2195 break;
2196
2197 /* ep halt tests */
2198 case 13:
2199 if (dev->out_pipe == 0 && dev->in_pipe == 0)
2200 break;
2201 retval = 0;
2202 dev_info(&intf->dev, "TEST 13: set/clear %d halts\n",
2203 param->iterations);
2204 for (i = param->iterations; retval == 0 && i--; /* NOP */)
2205 retval = halt_simple(dev);
2206
2207 if (retval)
2208 ERROR(dev, "halts failed, iterations left %d\n", i);
2209 break;
2210
2211 /* control write tests */
2212 case 14:
2213 if (!dev->info->ctrl_out)
2214 break;
2215 dev_info(&intf->dev, "TEST 14: %d ep0out, %d..%d vary %d\n",
2216 param->iterations,
2217 realworld ? 1 : 0, param->length,
2218 param->vary);
2219 retval = ctrl_out(dev, param->iterations,
2220 param->length, param->vary, 0);
2221 break;
2222
2223 /* iso write tests */
2224 case 15:
2225 if (dev->out_iso_pipe == 0 || param->sglen == 0)
2226 break;
2227 dev_info(&intf->dev,
2228 "TEST 15: write %d iso, %d entries of %d bytes\n",
2229 param->iterations,
2230 param->sglen, param->length);
2231 /* FIRMWARE: iso sink */
2232 retval = test_iso_queue(dev, param,
2233 dev->out_iso_pipe, dev->iso_out, 0);
2234 break;
2235
2236 /* iso read tests */
2237 case 16:
2238 if (dev->in_iso_pipe == 0 || param->sglen == 0)
2239 break;
2240 dev_info(&intf->dev,
2241 "TEST 16: read %d iso, %d entries of %d bytes\n",
2242 param->iterations,
2243 param->sglen, param->length);
2244 /* FIRMWARE: iso source */
2245 retval = test_iso_queue(dev, param,
2246 dev->in_iso_pipe, dev->iso_in, 0);
2247 break;
2248
2249 /* FIXME scatterlist cancel (needs helper thread) */
2250
2251 /* Tests for bulk I/O using DMA mapping by core and odd address */
2252 case 17:
2253 if (dev->out_pipe == 0)
2254 break;
2255 dev_info(&intf->dev,
2256 "TEST 17: write odd addr %d bytes %u times core map\n",
2257 param->length, param->iterations);
2258
2259 retval = test_unaligned_bulk(
2260 dev, dev->out_pipe,
2261 param->length, param->iterations,
2262 0, "test17");
2263 break;
2264
2265 case 18:
2266 if (dev->in_pipe == 0)
2267 break;
2268 dev_info(&intf->dev,
2269 "TEST 18: read odd addr %d bytes %u times core map\n",
2270 param->length, param->iterations);
2271
2272 retval = test_unaligned_bulk(
2273 dev, dev->in_pipe,
2274 param->length, param->iterations,
2275 0, "test18");
2276 break;
2277
2278 /* Tests for bulk I/O using premapped coherent buffer and odd address */
2279 case 19:
2280 if (dev->out_pipe == 0)
2281 break;
2282 dev_info(&intf->dev,
2283 "TEST 19: write odd addr %d bytes %u times premapped\n",
2284 param->length, param->iterations);
2285
2286 retval = test_unaligned_bulk(
2287 dev, dev->out_pipe,
2288 param->length, param->iterations,
2289 URB_NO_TRANSFER_DMA_MAP, "test19");
2290 break;
2291
2292 case 20:
2293 if (dev->in_pipe == 0)
2294 break;
2295 dev_info(&intf->dev,
2296 "TEST 20: read odd addr %d bytes %u times premapped\n",
2297 param->length, param->iterations);
2298
2299 retval = test_unaligned_bulk(
2300 dev, dev->in_pipe,
2301 param->length, param->iterations,
2302 URB_NO_TRANSFER_DMA_MAP, "test20");
2303 break;
2304
2305 /* control write tests with unaligned buffer */
2306 case 21:
2307 if (!dev->info->ctrl_out)
2308 break;
2309 dev_info(&intf->dev,
2310 "TEST 21: %d ep0out odd addr, %d..%d vary %d\n",
2311 param->iterations,
2312 realworld ? 1 : 0, param->length,
2313 param->vary);
2314 retval = ctrl_out(dev, param->iterations,
2315 param->length, param->vary, 1);
2316 break;
2317
2318 /* unaligned iso tests */
2319 case 22:
2320 if (dev->out_iso_pipe == 0 || param->sglen == 0)
2321 break;
2322 dev_info(&intf->dev,
2323 "TEST 22: write %d iso odd, %d entries of %d bytes\n",
2324 param->iterations,
2325 param->sglen, param->length);
2326 retval = test_iso_queue(dev, param,
2327 dev->out_iso_pipe, dev->iso_out, 1);
2328 break;
2329
2330 case 23:
2331 if (dev->in_iso_pipe == 0 || param->sglen == 0)
2332 break;
2333 dev_info(&intf->dev,
2334 "TEST 23: read %d iso odd, %d entries of %d bytes\n",
2335 param->iterations,
2336 param->sglen, param->length);
2337 retval = test_iso_queue(dev, param,
2338 dev->in_iso_pipe, dev->iso_in, 1);
2339 break;
2340
2341 /* unlink URBs from a bulk-OUT queue */
2342 case 24:
2343 if (dev->out_pipe == 0 || !param->length || param->sglen < 4)
2344 break;
2345 retval = 0;
2346 dev_info(&intf->dev, "TEST 24: unlink from %d queues of "
2347 "%d %d-byte writes\n",
2348 param->iterations, param->sglen, param->length);
2349 for (i = param->iterations; retval == 0 && i > 0; --i) {
2350 retval = unlink_queued(dev, dev->out_pipe,
2351 param->sglen, param->length);
2352 if (retval) {
2353 dev_err(&intf->dev,
2354 "unlink queued writes failed %d, "
2355 "iterations left %d\n", retval, i);
2356 break;
2357 }
2358 }
2359 break;
2360
2361 }
2362 do_gettimeofday(¶m->duration);
2363 param->duration.tv_sec -= start.tv_sec;
2364 param->duration.tv_usec -= start.tv_usec;
2365 if (param->duration.tv_usec < 0) {
2366 param->duration.tv_usec += 1000 * 1000;
2367 param->duration.tv_sec -= 1;
2368 }
2369 mutex_unlock(&dev->lock);
2370 return retval;
2371}
2372
2373/*-------------------------------------------------------------------------*/
2374
2375static unsigned force_interrupt;
2376module_param(force_interrupt, uint, 0);
2377MODULE_PARM_DESC(force_interrupt, "0 = test default; else interrupt");
2378
2379#ifdef GENERIC
2380static unsigned short vendor;
2381module_param(vendor, ushort, 0);
2382MODULE_PARM_DESC(vendor, "vendor code (from usb-if)");
2383
2384static unsigned short product;
2385module_param(product, ushort, 0);
2386MODULE_PARM_DESC(product, "product code (from vendor)");
2387#endif
2388
2389static int
2390usbtest_probe(struct usb_interface *intf, const struct usb_device_id *id)
2391{
2392 struct usb_device *udev;
2393 struct usbtest_dev *dev;
2394 struct usbtest_info *info;
2395 char *rtest, *wtest;
2396 char *irtest, *iwtest;
2397
2398 udev = interface_to_usbdev(intf);
2399
2400#ifdef GENERIC
2401 /* specify devices by module parameters? */
2402 if (id->match_flags == 0) {
2403 /* vendor match required, product match optional */
2404 if (!vendor || le16_to_cpu(udev->descriptor.idVendor) != (u16)vendor)
2405 return -ENODEV;
2406 if (product && le16_to_cpu(udev->descriptor.idProduct) != (u16)product)
2407 return -ENODEV;
2408 dev_info(&intf->dev, "matched module params, "
2409 "vend=0x%04x prod=0x%04x\n",
2410 le16_to_cpu(udev->descriptor.idVendor),
2411 le16_to_cpu(udev->descriptor.idProduct));
2412 }
2413#endif
2414
2415 dev = kzalloc(sizeof(*dev), GFP_KERNEL);
2416 if (!dev)
2417 return -ENOMEM;
2418 info = (struct usbtest_info *) id->driver_info;
2419 dev->info = info;
2420 mutex_init(&dev->lock);
2421
2422 dev->intf = intf;
2423
2424 /* cacheline-aligned scratch for i/o */
2425 dev->buf = kmalloc(TBUF_SIZE, GFP_KERNEL);
2426 if (dev->buf == NULL) {
2427 kfree(dev);
2428 return -ENOMEM;
2429 }
2430
2431 /* NOTE this doesn't yet test the handful of difference that are
2432 * visible with high speed interrupts: bigger maxpacket (1K) and
2433 * "high bandwidth" modes (up to 3 packets/uframe).
2434 */
2435 rtest = wtest = "";
2436 irtest = iwtest = "";
2437 if (force_interrupt || udev->speed == USB_SPEED_LOW) {
2438 if (info->ep_in) {
2439 dev->in_pipe = usb_rcvintpipe(udev, info->ep_in);
2440 rtest = " intr-in";
2441 }
2442 if (info->ep_out) {
2443 dev->out_pipe = usb_sndintpipe(udev, info->ep_out);
2444 wtest = " intr-out";
2445 }
2446 } else {
2447 if (override_alt >= 0 || info->autoconf) {
2448 int status;
2449
2450 status = get_endpoints(dev, intf);
2451 if (status < 0) {
2452 WARNING(dev, "couldn't get endpoints, %d\n",
2453 status);
2454 kfree(dev->buf);
2455 kfree(dev);
2456 return status;
2457 }
2458 /* may find bulk or ISO pipes */
2459 } else {
2460 if (info->ep_in)
2461 dev->in_pipe = usb_rcvbulkpipe(udev,
2462 info->ep_in);
2463 if (info->ep_out)
2464 dev->out_pipe = usb_sndbulkpipe(udev,
2465 info->ep_out);
2466 }
2467 if (dev->in_pipe)
2468 rtest = " bulk-in";
2469 if (dev->out_pipe)
2470 wtest = " bulk-out";
2471 if (dev->in_iso_pipe)
2472 irtest = " iso-in";
2473 if (dev->out_iso_pipe)
2474 iwtest = " iso-out";
2475 }
2476
2477 usb_set_intfdata(intf, dev);
2478 dev_info(&intf->dev, "%s\n", info->name);
2479 dev_info(&intf->dev, "%s {control%s%s%s%s%s} tests%s\n",
2480 usb_speed_string(udev->speed),
2481 info->ctrl_out ? " in/out" : "",
2482 rtest, wtest,
2483 irtest, iwtest,
2484 info->alt >= 0 ? " (+alt)" : "");
2485 return 0;
2486}
2487
2488static int usbtest_suspend(struct usb_interface *intf, pm_message_t message)
2489{
2490 return 0;
2491}
2492
2493static int usbtest_resume(struct usb_interface *intf)
2494{
2495 return 0;
2496}
2497
2498
2499static void usbtest_disconnect(struct usb_interface *intf)
2500{
2501 struct usbtest_dev *dev = usb_get_intfdata(intf);
2502
2503 usb_set_intfdata(intf, NULL);
2504 dev_dbg(&intf->dev, "disconnect\n");
2505 kfree(dev);
2506}
2507
2508/* Basic testing only needs a device that can source or sink bulk traffic.
2509 * Any device can test control transfers (default with GENERIC binding).
2510 *
2511 * Several entries work with the default EP0 implementation that's built
2512 * into EZ-USB chips. There's a default vendor ID which can be overridden
2513 * by (very) small config EEPROMS, but otherwise all these devices act
2514 * identically until firmware is loaded: only EP0 works. It turns out
2515 * to be easy to make other endpoints work, without modifying that EP0
2516 * behavior. For now, we expect that kind of firmware.
2517 */
2518
2519/* an21xx or fx versions of ez-usb */
2520static struct usbtest_info ez1_info = {
2521 .name = "EZ-USB device",
2522 .ep_in = 2,
2523 .ep_out = 2,
2524 .alt = 1,
2525};
2526
2527/* fx2 version of ez-usb */
2528static struct usbtest_info ez2_info = {
2529 .name = "FX2 device",
2530 .ep_in = 6,
2531 .ep_out = 2,
2532 .alt = 1,
2533};
2534
2535/* ezusb family device with dedicated usb test firmware,
2536 */
2537static struct usbtest_info fw_info = {
2538 .name = "usb test device",
2539 .ep_in = 2,
2540 .ep_out = 2,
2541 .alt = 1,
2542 .autoconf = 1, /* iso and ctrl_out need autoconf */
2543 .ctrl_out = 1,
2544 .iso = 1, /* iso_ep's are #8 in/out */
2545};
2546
2547/* peripheral running Linux and 'zero.c' test firmware, or
2548 * its user-mode cousin. different versions of this use
2549 * different hardware with the same vendor/product codes.
2550 * host side MUST rely on the endpoint descriptors.
2551 */
2552static struct usbtest_info gz_info = {
2553 .name = "Linux gadget zero",
2554 .autoconf = 1,
2555 .ctrl_out = 1,
2556 .iso = 1,
2557 .alt = 0,
2558};
2559
2560static struct usbtest_info um_info = {
2561 .name = "Linux user mode test driver",
2562 .autoconf = 1,
2563 .alt = -1,
2564};
2565
2566static struct usbtest_info um2_info = {
2567 .name = "Linux user mode ISO test driver",
2568 .autoconf = 1,
2569 .iso = 1,
2570 .alt = -1,
2571};
2572
2573#ifdef IBOT2
2574/* this is a nice source of high speed bulk data;
2575 * uses an FX2, with firmware provided in the device
2576 */
2577static struct usbtest_info ibot2_info = {
2578 .name = "iBOT2 webcam",
2579 .ep_in = 2,
2580 .alt = -1,
2581};
2582#endif
2583
2584#ifdef GENERIC
2585/* we can use any device to test control traffic */
2586static struct usbtest_info generic_info = {
2587 .name = "Generic USB device",
2588 .alt = -1,
2589};
2590#endif
2591
2592
2593static const struct usb_device_id id_table[] = {
2594
2595 /*-------------------------------------------------------------*/
2596
2597 /* EZ-USB devices which download firmware to replace (or in our
2598 * case augment) the default device implementation.
2599 */
2600
2601 /* generic EZ-USB FX controller */
2602 { USB_DEVICE(0x0547, 0x2235),
2603 .driver_info = (unsigned long) &ez1_info,
2604 },
2605
2606 /* CY3671 development board with EZ-USB FX */
2607 { USB_DEVICE(0x0547, 0x0080),
2608 .driver_info = (unsigned long) &ez1_info,
2609 },
2610
2611 /* generic EZ-USB FX2 controller (or development board) */
2612 { USB_DEVICE(0x04b4, 0x8613),
2613 .driver_info = (unsigned long) &ez2_info,
2614 },
2615
2616 /* re-enumerated usb test device firmware */
2617 { USB_DEVICE(0xfff0, 0xfff0),
2618 .driver_info = (unsigned long) &fw_info,
2619 },
2620
2621 /* "Gadget Zero" firmware runs under Linux */
2622 { USB_DEVICE(0x0525, 0xa4a0),
2623 .driver_info = (unsigned long) &gz_info,
2624 },
2625
2626 /* so does a user-mode variant */
2627 { USB_DEVICE(0x0525, 0xa4a4),
2628 .driver_info = (unsigned long) &um_info,
2629 },
2630
2631 /* ... and a user-mode variant that talks iso */
2632 { USB_DEVICE(0x0525, 0xa4a3),
2633 .driver_info = (unsigned long) &um2_info,
2634 },
2635
2636#ifdef KEYSPAN_19Qi
2637 /* Keyspan 19qi uses an21xx (original EZ-USB) */
2638 /* this does not coexist with the real Keyspan 19qi driver! */
2639 { USB_DEVICE(0x06cd, 0x010b),
2640 .driver_info = (unsigned long) &ez1_info,
2641 },
2642#endif
2643
2644 /*-------------------------------------------------------------*/
2645
2646#ifdef IBOT2
2647 /* iBOT2 makes a nice source of high speed bulk-in data */
2648 /* this does not coexist with a real iBOT2 driver! */
2649 { USB_DEVICE(0x0b62, 0x0059),
2650 .driver_info = (unsigned long) &ibot2_info,
2651 },
2652#endif
2653
2654 /*-------------------------------------------------------------*/
2655
2656#ifdef GENERIC
2657 /* module params can specify devices to use for control tests */
2658 { .driver_info = (unsigned long) &generic_info, },
2659#endif
2660
2661 /*-------------------------------------------------------------*/
2662
2663 { }
2664};
2665MODULE_DEVICE_TABLE(usb, id_table);
2666
2667static struct usb_driver usbtest_driver = {
2668 .name = "usbtest",
2669 .id_table = id_table,
2670 .probe = usbtest_probe,
2671 .unlocked_ioctl = usbtest_ioctl,
2672 .disconnect = usbtest_disconnect,
2673 .suspend = usbtest_suspend,
2674 .resume = usbtest_resume,
2675};
2676
2677/*-------------------------------------------------------------------------*/
2678
2679static int __init usbtest_init(void)
2680{
2681#ifdef GENERIC
2682 if (vendor)
2683 pr_debug("params: vend=0x%04x prod=0x%04x\n", vendor, product);
2684#endif
2685 return usb_register(&usbtest_driver);
2686}
2687module_init(usbtest_init);
2688
2689static void __exit usbtest_exit(void)
2690{
2691 usb_deregister(&usbtest_driver);
2692}
2693module_exit(usbtest_exit);
2694
2695MODULE_DESCRIPTION("USB Core/HCD Testing Driver");
2696MODULE_LICENSE("GPL");
2697