blob: dd72e75beb82eeb06909b5721325daddd1cf1aa5 [file] [log] [blame]
b.liue9582032025-04-17 19:18:16 +08001// SPDX-License-Identifier: GPL-2.0+
2/*
3 * inode.c -- user mode filesystem api for usb gadget controllers
4 *
5 * Copyright (C) 2003-2004 David Brownell
6 * Copyright (C) 2003 Agilent Technologies
7 */
8
9
10/* #define VERBOSE_DEBUG */
11
12#include <linux/init.h>
13#include <linux/module.h>
14#include <linux/fs.h>
15#include <linux/fs_context.h>
16#include <linux/pagemap.h>
17#include <linux/uts.h>
18#include <linux/wait.h>
19#include <linux/compiler.h>
20#include <linux/uaccess.h>
21#include <linux/sched.h>
22#include <linux/slab.h>
23#include <linux/poll.h>
24#include <linux/mmu_context.h>
25#include <linux/aio.h>
26#include <linux/uio.h>
27#include <linux/refcount.h>
28#include <linux/delay.h>
29#include <linux/device.h>
30#include <linux/moduleparam.h>
31
32#include <linux/usb/gadgetfs.h>
33#include <linux/usb/gadget.h>
34
35
36/*
37 * The gadgetfs API maps each endpoint to a file descriptor so that you
38 * can use standard synchronous read/write calls for I/O. There's some
39 * O_NONBLOCK and O_ASYNC/FASYNC style i/o support. Example usermode
40 * drivers show how this works in practice. You can also use AIO to
41 * eliminate I/O gaps between requests, to help when streaming data.
42 *
43 * Key parts that must be USB-specific are protocols defining how the
44 * read/write operations relate to the hardware state machines. There
45 * are two types of files. One type is for the device, implementing ep0.
46 * The other type is for each IN or OUT endpoint. In both cases, the
47 * user mode driver must configure the hardware before using it.
48 *
49 * - First, dev_config() is called when /dev/gadget/$CHIP is configured
50 * (by writing configuration and device descriptors). Afterwards it
51 * may serve as a source of device events, used to handle all control
52 * requests other than basic enumeration.
53 *
54 * - Then, after a SET_CONFIGURATION control request, ep_config() is
55 * called when each /dev/gadget/ep* file is configured (by writing
56 * endpoint descriptors). Afterwards these files are used to write()
57 * IN data or to read() OUT data. To halt the endpoint, a "wrong
58 * direction" request is issued (like reading an IN endpoint).
59 *
60 * Unlike "usbfs" the only ioctl()s are for things that are rare, and maybe
61 * not possible on all hardware. For example, precise fault handling with
62 * respect to data left in endpoint fifos after aborted operations; or
63 * selective clearing of endpoint halts, to implement SET_INTERFACE.
64 */
65
66#define DRIVER_DESC "USB Gadget filesystem"
67#define DRIVER_VERSION "24 Aug 2004"
68
69static const char driver_desc [] = DRIVER_DESC;
70static const char shortname [] = "gadgetfs";
71
72MODULE_DESCRIPTION (DRIVER_DESC);
73MODULE_AUTHOR ("David Brownell");
74MODULE_LICENSE ("GPL");
75
76static int ep_open(struct inode *, struct file *);
77
78
79/*----------------------------------------------------------------------*/
80
81#define GADGETFS_MAGIC 0xaee71ee7
82
83/* /dev/gadget/$CHIP represents ep0 and the whole device */
84enum ep0_state {
85 /* DISABLED is the initial state. */
86 STATE_DEV_DISABLED = 0,
87
88 /* Only one open() of /dev/gadget/$CHIP; only one file tracks
89 * ep0/device i/o modes and binding to the controller. Driver
90 * must always write descriptors to initialize the device, then
91 * the device becomes UNCONNECTED until enumeration.
92 */
93 STATE_DEV_OPENED,
94
95 /* From then on, ep0 fd is in either of two basic modes:
96 * - (UN)CONNECTED: read usb_gadgetfs_event(s) from it
97 * - SETUP: read/write will transfer control data and succeed;
98 * or if "wrong direction", performs protocol stall
99 */
100 STATE_DEV_UNCONNECTED,
101 STATE_DEV_CONNECTED,
102 STATE_DEV_SETUP,
103
104 /* UNBOUND means the driver closed ep0, so the device won't be
105 * accessible again (DEV_DISABLED) until all fds are closed.
106 */
107 STATE_DEV_UNBOUND,
108};
109
110/* enough for the whole queue: most events invalidate others */
111#define N_EVENT 5
112
113#define RBUF_SIZE 256
114
115struct dev_data {
116 spinlock_t lock;
117 refcount_t count;
118 int udc_usage;
119 enum ep0_state state; /* P: lock */
120 struct usb_gadgetfs_event event [N_EVENT];
121 unsigned ev_next;
122 struct fasync_struct *fasync;
123 u8 current_config;
124
125 /* drivers reading ep0 MUST handle control requests (SETUP)
126 * reported that way; else the host will time out.
127 */
128 unsigned usermode_setup : 1,
129 setup_in : 1,
130 setup_can_stall : 1,
131 setup_out_ready : 1,
132 setup_out_error : 1,
133 setup_abort : 1,
134 gadget_registered : 1;
135 unsigned setup_wLength;
136
137 /* the rest is basically write-once */
138 struct usb_config_descriptor *config, *hs_config;
139 struct usb_device_descriptor *dev;
140 struct usb_request *req;
141 struct usb_gadget *gadget;
142 struct list_head epfiles;
143 void *buf;
144 wait_queue_head_t wait;
145 struct super_block *sb;
146 struct dentry *dentry;
147
148 /* except this scratch i/o buffer for ep0 */
149 u8 rbuf[RBUF_SIZE];
150};
151
152static inline void get_dev (struct dev_data *data)
153{
154 refcount_inc (&data->count);
155}
156
157static void put_dev (struct dev_data *data)
158{
159 if (likely (!refcount_dec_and_test (&data->count)))
160 return;
161 /* needs no more cleanup */
162 BUG_ON (waitqueue_active (&data->wait));
163 kfree (data);
164}
165
166static struct dev_data *dev_new (void)
167{
168 struct dev_data *dev;
169
170 dev = kzalloc(sizeof(*dev), GFP_KERNEL);
171 if (!dev)
172 return NULL;
173 dev->state = STATE_DEV_DISABLED;
174 refcount_set (&dev->count, 1);
175 spin_lock_init (&dev->lock);
176 INIT_LIST_HEAD (&dev->epfiles);
177 init_waitqueue_head (&dev->wait);
178 return dev;
179}
180
181/*----------------------------------------------------------------------*/
182
183/* other /dev/gadget/$ENDPOINT files represent endpoints */
184enum ep_state {
185 STATE_EP_DISABLED = 0,
186 STATE_EP_READY,
187 STATE_EP_ENABLED,
188 STATE_EP_UNBOUND,
189};
190
191struct ep_data {
192 struct mutex lock;
193 enum ep_state state;
194 refcount_t count;
195 struct dev_data *dev;
196 /* must hold dev->lock before accessing ep or req */
197 struct usb_ep *ep;
198 struct usb_request *req;
199 ssize_t status;
200 char name [16];
201 struct usb_endpoint_descriptor desc, hs_desc;
202 struct list_head epfiles;
203 wait_queue_head_t wait;
204 struct dentry *dentry;
205};
206
207static inline void get_ep (struct ep_data *data)
208{
209 refcount_inc (&data->count);
210}
211
212static void put_ep (struct ep_data *data)
213{
214 if (likely (!refcount_dec_and_test (&data->count)))
215 return;
216 put_dev (data->dev);
217 /* needs no more cleanup */
218 BUG_ON (!list_empty (&data->epfiles));
219 BUG_ON (waitqueue_active (&data->wait));
220 kfree (data);
221}
222
223/*----------------------------------------------------------------------*/
224
225/* most "how to use the hardware" policy choices are in userspace:
226 * mapping endpoint roles (which the driver needs) to the capabilities
227 * which the usb controller has. most of those capabilities are exposed
228 * implicitly, starting with the driver name and then endpoint names.
229 */
230
231static const char *CHIP;
232static DEFINE_MUTEX(sb_mutex); /* Serialize superblock operations */
233
234/*----------------------------------------------------------------------*/
235
236/* NOTE: don't use dev_printk calls before binding to the gadget
237 * at the end of ep0 configuration, or after unbind.
238 */
239
240/* too wordy: dev_printk(level , &(d)->gadget->dev , fmt , ## args) */
241#define xprintk(d,level,fmt,args...) \
242 printk(level "%s: " fmt , shortname , ## args)
243
244#ifdef DEBUG
245#define DBG(dev,fmt,args...) \
246 xprintk(dev , KERN_DEBUG , fmt , ## args)
247#else
248#define DBG(dev,fmt,args...) \
249 do { } while (0)
250#endif /* DEBUG */
251
252#ifdef VERBOSE_DEBUG
253#define VDEBUG DBG
254#else
255#define VDEBUG(dev,fmt,args...) \
256 do { } while (0)
257#endif /* DEBUG */
258
259#define ERROR(dev,fmt,args...) \
260 xprintk(dev , KERN_ERR , fmt , ## args)
261#define INFO(dev,fmt,args...) \
262 xprintk(dev , KERN_INFO , fmt , ## args)
263
264
265/*----------------------------------------------------------------------*/
266
267/* SYNCHRONOUS ENDPOINT OPERATIONS (bulk/intr/iso)
268 *
269 * After opening, configure non-control endpoints. Then use normal
270 * stream read() and write() requests; and maybe ioctl() to get more
271 * precise FIFO status when recovering from cancellation.
272 */
273
274static void epio_complete (struct usb_ep *ep, struct usb_request *req)
275{
276 struct ep_data *epdata = ep->driver_data;
277
278 if (!req->context)
279 return;
280 if (req->status)
281 epdata->status = req->status;
282 else
283 epdata->status = req->actual;
284 complete ((struct completion *)req->context);
285}
286
287/* tasklock endpoint, returning when it's connected.
288 * still need dev->lock to use epdata->ep.
289 */
290static int
291get_ready_ep (unsigned f_flags, struct ep_data *epdata, bool is_write)
292{
293 int val;
294
295 if (f_flags & O_NONBLOCK) {
296 if (!mutex_trylock(&epdata->lock))
297 goto nonblock;
298 if (epdata->state != STATE_EP_ENABLED &&
299 (!is_write || epdata->state != STATE_EP_READY)) {
300 mutex_unlock(&epdata->lock);
301nonblock:
302 val = -EAGAIN;
303 } else
304 val = 0;
305 return val;
306 }
307
308 val = mutex_lock_interruptible(&epdata->lock);
309 if (val < 0)
310 return val;
311
312 switch (epdata->state) {
313 case STATE_EP_ENABLED:
314 return 0;
315 case STATE_EP_READY: /* not configured yet */
316 if (is_write)
317 return 0;
318 // FALLTHRU
319 case STATE_EP_UNBOUND: /* clean disconnect */
320 break;
321 // case STATE_EP_DISABLED: /* "can't happen" */
322 default: /* error! */
323 pr_debug ("%s: ep %p not available, state %d\n",
324 shortname, epdata, epdata->state);
325 }
326 mutex_unlock(&epdata->lock);
327 return -ENODEV;
328}
329
330static ssize_t
331ep_io (struct ep_data *epdata, void *buf, unsigned len)
332{
333 DECLARE_COMPLETION_ONSTACK (done);
334 int value;
335
336 spin_lock_irq (&epdata->dev->lock);
337 if (likely (epdata->ep != NULL)) {
338 struct usb_request *req = epdata->req;
339
340 req->context = &done;
341 req->complete = epio_complete;
342 req->buf = buf;
343 req->length = len;
344 value = usb_ep_queue (epdata->ep, req, GFP_ATOMIC);
345 } else
346 value = -ENODEV;
347 spin_unlock_irq (&epdata->dev->lock);
348
349 if (likely (value == 0)) {
350 value = wait_event_interruptible (done.wait, done.done);
351 if (value != 0) {
352 spin_lock_irq (&epdata->dev->lock);
353 if (likely (epdata->ep != NULL)) {
354 DBG (epdata->dev, "%s i/o interrupted\n",
355 epdata->name);
356 usb_ep_dequeue (epdata->ep, epdata->req);
357 spin_unlock_irq (&epdata->dev->lock);
358
359 wait_event (done.wait, done.done);
360 if (epdata->status == -ECONNRESET)
361 epdata->status = -EINTR;
362 } else {
363 spin_unlock_irq (&epdata->dev->lock);
364
365 DBG (epdata->dev, "endpoint gone\n");
366 wait_for_completion(&done);
367 epdata->status = -ENODEV;
368 }
369 }
370 return epdata->status;
371 }
372 return value;
373}
374
375static int
376ep_release (struct inode *inode, struct file *fd)
377{
378 struct ep_data *data = fd->private_data;
379 int value;
380
381 value = mutex_lock_interruptible(&data->lock);
382 if (value < 0)
383 return value;
384
385 /* clean up if this can be reopened */
386 if (data->state != STATE_EP_UNBOUND) {
387 data->state = STATE_EP_DISABLED;
388 data->desc.bDescriptorType = 0;
389 data->hs_desc.bDescriptorType = 0;
390 usb_ep_disable(data->ep);
391 }
392 mutex_unlock(&data->lock);
393 put_ep (data);
394 return 0;
395}
396
397static long ep_ioctl(struct file *fd, unsigned code, unsigned long value)
398{
399 struct ep_data *data = fd->private_data;
400 int status;
401
402 if ((status = get_ready_ep (fd->f_flags, data, false)) < 0)
403 return status;
404
405 spin_lock_irq (&data->dev->lock);
406 if (likely (data->ep != NULL)) {
407 switch (code) {
408 case GADGETFS_FIFO_STATUS:
409 status = usb_ep_fifo_status (data->ep);
410 break;
411 case GADGETFS_FIFO_FLUSH:
412 usb_ep_fifo_flush (data->ep);
413 break;
414 case GADGETFS_CLEAR_HALT:
415 status = usb_ep_clear_halt (data->ep);
416 break;
417 default:
418 status = -ENOTTY;
419 }
420 } else
421 status = -ENODEV;
422 spin_unlock_irq (&data->dev->lock);
423 mutex_unlock(&data->lock);
424 return status;
425}
426
427/*----------------------------------------------------------------------*/
428
429/* ASYNCHRONOUS ENDPOINT I/O OPERATIONS (bulk/intr/iso) */
430
431struct kiocb_priv {
432 struct usb_request *req;
433 struct ep_data *epdata;
434 struct kiocb *iocb;
435 struct mm_struct *mm;
436 struct work_struct work;
437 void *buf;
438 struct iov_iter to;
439 const void *to_free;
440 unsigned actual;
441};
442
443static int ep_aio_cancel(struct kiocb *iocb)
444{
445 struct kiocb_priv *priv = iocb->private;
446 struct ep_data *epdata;
447 int value;
448
449 local_irq_disable();
450 epdata = priv->epdata;
451 // spin_lock(&epdata->dev->lock);
452 if (likely(epdata && epdata->ep && priv->req))
453 value = usb_ep_dequeue (epdata->ep, priv->req);
454 else
455 value = -EINVAL;
456 // spin_unlock(&epdata->dev->lock);
457 local_irq_enable();
458
459 return value;
460}
461
462static void ep_user_copy_worker(struct work_struct *work)
463{
464 struct kiocb_priv *priv = container_of(work, struct kiocb_priv, work);
465 struct mm_struct *mm = priv->mm;
466 struct kiocb *iocb = priv->iocb;
467 size_t ret;
468
469 use_mm(mm);
470 ret = copy_to_iter(priv->buf, priv->actual, &priv->to);
471 unuse_mm(mm);
472 if (!ret)
473 ret = -EFAULT;
474
475 /* completing the iocb can drop the ctx and mm, don't touch mm after */
476 iocb->ki_complete(iocb, ret, ret);
477
478 kfree(priv->buf);
479 kfree(priv->to_free);
480 kfree(priv);
481}
482
483static void ep_aio_complete(struct usb_ep *ep, struct usb_request *req)
484{
485 struct kiocb *iocb = req->context;
486 struct kiocb_priv *priv = iocb->private;
487 struct ep_data *epdata = priv->epdata;
488
489 /* lock against disconnect (and ideally, cancel) */
490 spin_lock(&epdata->dev->lock);
491 priv->req = NULL;
492 priv->epdata = NULL;
493
494 /* if this was a write or a read returning no data then we
495 * don't need to copy anything to userspace, so we can
496 * complete the aio request immediately.
497 */
498 if (priv->to_free == NULL || unlikely(req->actual == 0)) {
499 kfree(req->buf);
500 kfree(priv->to_free);
501 kfree(priv);
502 iocb->private = NULL;
503 /* aio_complete() reports bytes-transferred _and_ faults */
504
505 iocb->ki_complete(iocb, req->actual ? req->actual : req->status,
506 req->status);
507 } else {
508 /* ep_copy_to_user() won't report both; we hide some faults */
509 if (unlikely(0 != req->status))
510 DBG(epdata->dev, "%s fault %d len %d\n",
511 ep->name, req->status, req->actual);
512
513 priv->buf = req->buf;
514 priv->actual = req->actual;
515 INIT_WORK(&priv->work, ep_user_copy_worker);
516 schedule_work(&priv->work);
517 }
518
519 usb_ep_free_request(ep, req);
520 spin_unlock(&epdata->dev->lock);
521 put_ep(epdata);
522}
523
524static ssize_t ep_aio(struct kiocb *iocb,
525 struct kiocb_priv *priv,
526 struct ep_data *epdata,
527 char *buf,
528 size_t len)
529{
530 struct usb_request *req;
531 ssize_t value;
532
533 iocb->private = priv;
534 priv->iocb = iocb;
535
536 kiocb_set_cancel_fn(iocb, ep_aio_cancel);
537 get_ep(epdata);
538 priv->epdata = epdata;
539 priv->actual = 0;
540 priv->mm = current->mm; /* mm teardown waits for iocbs in exit_aio() */
541
542 /* each kiocb is coupled to one usb_request, but we can't
543 * allocate or submit those if the host disconnected.
544 */
545 spin_lock_irq(&epdata->dev->lock);
546 value = -ENODEV;
547 if (unlikely(epdata->ep == NULL))
548 goto fail;
549
550 req = usb_ep_alloc_request(epdata->ep, GFP_ATOMIC);
551 value = -ENOMEM;
552 if (unlikely(!req))
553 goto fail;
554
555 priv->req = req;
556 req->buf = buf;
557 req->length = len;
558 req->complete = ep_aio_complete;
559 req->context = iocb;
560 value = usb_ep_queue(epdata->ep, req, GFP_ATOMIC);
561 if (unlikely(0 != value)) {
562 usb_ep_free_request(epdata->ep, req);
563 goto fail;
564 }
565 spin_unlock_irq(&epdata->dev->lock);
566 return -EIOCBQUEUED;
567
568fail:
569 spin_unlock_irq(&epdata->dev->lock);
570 kfree(priv->to_free);
571 kfree(priv);
572 put_ep(epdata);
573 return value;
574}
575
576static ssize_t
577ep_read_iter(struct kiocb *iocb, struct iov_iter *to)
578{
579 struct file *file = iocb->ki_filp;
580 struct ep_data *epdata = file->private_data;
581 size_t len = iov_iter_count(to);
582 ssize_t value;
583 char *buf;
584
585 if ((value = get_ready_ep(file->f_flags, epdata, false)) < 0)
586 return value;
587
588 /* halt any endpoint by doing a "wrong direction" i/o call */
589 if (usb_endpoint_dir_in(&epdata->desc)) {
590 if (usb_endpoint_xfer_isoc(&epdata->desc) ||
591 !is_sync_kiocb(iocb)) {
592 mutex_unlock(&epdata->lock);
593 return -EINVAL;
594 }
595 DBG (epdata->dev, "%s halt\n", epdata->name);
596 spin_lock_irq(&epdata->dev->lock);
597 if (likely(epdata->ep != NULL))
598 usb_ep_set_halt(epdata->ep);
599 spin_unlock_irq(&epdata->dev->lock);
600 mutex_unlock(&epdata->lock);
601 return -EBADMSG;
602 }
603
604 buf = kmalloc(len, GFP_KERNEL);
605 if (unlikely(!buf)) {
606 mutex_unlock(&epdata->lock);
607 return -ENOMEM;
608 }
609 if (is_sync_kiocb(iocb)) {
610 value = ep_io(epdata, buf, len);
611 if (value >= 0 && (copy_to_iter(buf, value, to) != value))
612 value = -EFAULT;
613 } else {
614 struct kiocb_priv *priv = kzalloc(sizeof *priv, GFP_KERNEL);
615 value = -ENOMEM;
616 if (!priv)
617 goto fail;
618 priv->to_free = dup_iter(&priv->to, to, GFP_KERNEL);
619 if (!priv->to_free) {
620 kfree(priv);
621 goto fail;
622 }
623 value = ep_aio(iocb, priv, epdata, buf, len);
624 if (value == -EIOCBQUEUED)
625 buf = NULL;
626 }
627fail:
628 kfree(buf);
629 mutex_unlock(&epdata->lock);
630 return value;
631}
632
633static ssize_t ep_config(struct ep_data *, const char *, size_t);
634
635static ssize_t
636ep_write_iter(struct kiocb *iocb, struct iov_iter *from)
637{
638 struct file *file = iocb->ki_filp;
639 struct ep_data *epdata = file->private_data;
640 size_t len = iov_iter_count(from);
641 bool configured;
642 ssize_t value;
643 char *buf;
644
645 if ((value = get_ready_ep(file->f_flags, epdata, true)) < 0)
646 return value;
647
648 configured = epdata->state == STATE_EP_ENABLED;
649
650 /* halt any endpoint by doing a "wrong direction" i/o call */
651 if (configured && !usb_endpoint_dir_in(&epdata->desc)) {
652 if (usb_endpoint_xfer_isoc(&epdata->desc) ||
653 !is_sync_kiocb(iocb)) {
654 mutex_unlock(&epdata->lock);
655 return -EINVAL;
656 }
657 DBG (epdata->dev, "%s halt\n", epdata->name);
658 spin_lock_irq(&epdata->dev->lock);
659 if (likely(epdata->ep != NULL))
660 usb_ep_set_halt(epdata->ep);
661 spin_unlock_irq(&epdata->dev->lock);
662 mutex_unlock(&epdata->lock);
663 return -EBADMSG;
664 }
665
666 buf = kmalloc(len, GFP_KERNEL);
667 if (unlikely(!buf)) {
668 mutex_unlock(&epdata->lock);
669 return -ENOMEM;
670 }
671
672 if (unlikely(!copy_from_iter_full(buf, len, from))) {
673 value = -EFAULT;
674 goto out;
675 }
676
677 if (unlikely(!configured)) {
678 value = ep_config(epdata, buf, len);
679 } else if (is_sync_kiocb(iocb)) {
680 value = ep_io(epdata, buf, len);
681 } else {
682 struct kiocb_priv *priv = kzalloc(sizeof *priv, GFP_KERNEL);
683 value = -ENOMEM;
684 if (priv) {
685 value = ep_aio(iocb, priv, epdata, buf, len);
686 if (value == -EIOCBQUEUED)
687 buf = NULL;
688 }
689 }
690out:
691 kfree(buf);
692 mutex_unlock(&epdata->lock);
693 return value;
694}
695
696/*----------------------------------------------------------------------*/
697
698/* used after endpoint configuration */
699static const struct file_operations ep_io_operations = {
700 .owner = THIS_MODULE,
701
702 .open = ep_open,
703 .release = ep_release,
704 .llseek = no_llseek,
705 .unlocked_ioctl = ep_ioctl,
706 .read_iter = ep_read_iter,
707 .write_iter = ep_write_iter,
708};
709
710/* ENDPOINT INITIALIZATION
711 *
712 * fd = open ("/dev/gadget/$ENDPOINT", O_RDWR)
713 * status = write (fd, descriptors, sizeof descriptors)
714 *
715 * That write establishes the endpoint configuration, configuring
716 * the controller to process bulk, interrupt, or isochronous transfers
717 * at the right maxpacket size, and so on.
718 *
719 * The descriptors are message type 1, identified by a host order u32
720 * at the beginning of what's written. Descriptor order is: full/low
721 * speed descriptor, then optional high speed descriptor.
722 */
723static ssize_t
724ep_config (struct ep_data *data, const char *buf, size_t len)
725{
726 struct usb_ep *ep;
727 u32 tag;
728 int value, length = len;
729
730 if (data->state != STATE_EP_READY) {
731 value = -EL2HLT;
732 goto fail;
733 }
734
735 value = len;
736 if (len < USB_DT_ENDPOINT_SIZE + 4)
737 goto fail0;
738
739 /* we might need to change message format someday */
740 memcpy(&tag, buf, 4);
741 if (tag != 1) {
742 DBG(data->dev, "config %s, bad tag %d\n", data->name, tag);
743 goto fail0;
744 }
745 buf += 4;
746 len -= 4;
747
748 /* NOTE: audio endpoint extensions not accepted here;
749 * just don't include the extra bytes.
750 */
751
752 /* full/low speed descriptor, then high speed */
753 memcpy(&data->desc, buf, USB_DT_ENDPOINT_SIZE);
754 if (data->desc.bLength != USB_DT_ENDPOINT_SIZE
755 || data->desc.bDescriptorType != USB_DT_ENDPOINT)
756 goto fail0;
757 if (len != USB_DT_ENDPOINT_SIZE) {
758 if (len != 2 * USB_DT_ENDPOINT_SIZE)
759 goto fail0;
760 memcpy(&data->hs_desc, buf + USB_DT_ENDPOINT_SIZE,
761 USB_DT_ENDPOINT_SIZE);
762 if (data->hs_desc.bLength != USB_DT_ENDPOINT_SIZE
763 || data->hs_desc.bDescriptorType
764 != USB_DT_ENDPOINT) {
765 DBG(data->dev, "config %s, bad hs length or type\n",
766 data->name);
767 goto fail0;
768 }
769 }
770
771 spin_lock_irq (&data->dev->lock);
772 if (data->dev->state == STATE_DEV_UNBOUND) {
773 value = -ENOENT;
774 goto gone;
775 } else {
776 ep = data->ep;
777 if (ep == NULL) {
778 value = -ENODEV;
779 goto gone;
780 }
781 }
782 switch (data->dev->gadget->speed) {
783 case USB_SPEED_LOW:
784 case USB_SPEED_FULL:
785 ep->desc = &data->desc;
786 break;
787 case USB_SPEED_HIGH:
788 /* fails if caller didn't provide that descriptor... */
789 ep->desc = &data->hs_desc;
790 break;
791 default:
792 DBG(data->dev, "unconnected, %s init abandoned\n",
793 data->name);
794 value = -EINVAL;
795 goto gone;
796 }
797 value = usb_ep_enable(ep);
798 if (value == 0) {
799 data->state = STATE_EP_ENABLED;
800 value = length;
801 }
802gone:
803 spin_unlock_irq (&data->dev->lock);
804 if (value < 0) {
805fail:
806 data->desc.bDescriptorType = 0;
807 data->hs_desc.bDescriptorType = 0;
808 }
809 return value;
810fail0:
811 value = -EINVAL;
812 goto fail;
813}
814
815static int
816ep_open (struct inode *inode, struct file *fd)
817{
818 struct ep_data *data = inode->i_private;
819 int value = -EBUSY;
820
821 if (mutex_lock_interruptible(&data->lock) != 0)
822 return -EINTR;
823 spin_lock_irq (&data->dev->lock);
824 if (data->dev->state == STATE_DEV_UNBOUND)
825 value = -ENOENT;
826 else if (data->state == STATE_EP_DISABLED) {
827 value = 0;
828 data->state = STATE_EP_READY;
829 get_ep (data);
830 fd->private_data = data;
831 VDEBUG (data->dev, "%s ready\n", data->name);
832 } else
833 DBG (data->dev, "%s state %d\n",
834 data->name, data->state);
835 spin_unlock_irq (&data->dev->lock);
836 mutex_unlock(&data->lock);
837 return value;
838}
839
840/*----------------------------------------------------------------------*/
841
842/* EP0 IMPLEMENTATION can be partly in userspace.
843 *
844 * Drivers that use this facility receive various events, including
845 * control requests the kernel doesn't handle. Drivers that don't
846 * use this facility may be too simple-minded for real applications.
847 */
848
849static inline void ep0_readable (struct dev_data *dev)
850{
851 wake_up (&dev->wait);
852 kill_fasync (&dev->fasync, SIGIO, POLL_IN);
853}
854
855static void clean_req (struct usb_ep *ep, struct usb_request *req)
856{
857 struct dev_data *dev = ep->driver_data;
858
859 if (req->buf != dev->rbuf) {
860 kfree(req->buf);
861 req->buf = dev->rbuf;
862 }
863 req->complete = epio_complete;
864 dev->setup_out_ready = 0;
865}
866
867static void ep0_complete (struct usb_ep *ep, struct usb_request *req)
868{
869 struct dev_data *dev = ep->driver_data;
870 unsigned long flags;
871 int free = 1;
872
873 /* for control OUT, data must still get to userspace */
874 spin_lock_irqsave(&dev->lock, flags);
875 if (!dev->setup_in) {
876 dev->setup_out_error = (req->status != 0);
877 if (!dev->setup_out_error)
878 free = 0;
879 dev->setup_out_ready = 1;
880 ep0_readable (dev);
881 }
882
883 /* clean up as appropriate */
884 if (free && req->buf != &dev->rbuf)
885 clean_req (ep, req);
886 req->complete = epio_complete;
887 spin_unlock_irqrestore(&dev->lock, flags);
888}
889
890static int setup_req (struct usb_ep *ep, struct usb_request *req, u16 len)
891{
892 struct dev_data *dev = ep->driver_data;
893
894 if (dev->setup_out_ready) {
895 DBG (dev, "ep0 request busy!\n");
896 return -EBUSY;
897 }
898 if (len > sizeof (dev->rbuf))
899 req->buf = kmalloc(len, GFP_ATOMIC);
900 if (req->buf == NULL) {
901 req->buf = dev->rbuf;
902 return -ENOMEM;
903 }
904 req->complete = ep0_complete;
905 req->length = len;
906 req->zero = 0;
907 return 0;
908}
909
910static ssize_t
911ep0_read (struct file *fd, char __user *buf, size_t len, loff_t *ptr)
912{
913 struct dev_data *dev = fd->private_data;
914 ssize_t retval;
915 enum ep0_state state;
916
917 spin_lock_irq (&dev->lock);
918 if (dev->state <= STATE_DEV_OPENED) {
919 retval = -EINVAL;
920 goto done;
921 }
922
923 /* report fd mode change before acting on it */
924 if (dev->setup_abort) {
925 dev->setup_abort = 0;
926 retval = -EIDRM;
927 goto done;
928 }
929
930 /* control DATA stage */
931 if ((state = dev->state) == STATE_DEV_SETUP) {
932
933 if (dev->setup_in) { /* stall IN */
934 VDEBUG(dev, "ep0in stall\n");
935 (void) usb_ep_set_halt (dev->gadget->ep0);
936 retval = -EL2HLT;
937 dev->state = STATE_DEV_CONNECTED;
938
939 } else if (len == 0) { /* ack SET_CONFIGURATION etc */
940 struct usb_ep *ep = dev->gadget->ep0;
941 struct usb_request *req = dev->req;
942
943 if ((retval = setup_req (ep, req, 0)) == 0) {
944 ++dev->udc_usage;
945 spin_unlock_irq (&dev->lock);
946 retval = usb_ep_queue (ep, req, GFP_KERNEL);
947 spin_lock_irq (&dev->lock);
948 --dev->udc_usage;
949 }
950 dev->state = STATE_DEV_CONNECTED;
951
952 /* assume that was SET_CONFIGURATION */
953 if (dev->current_config) {
954 unsigned power;
955
956 if (gadget_is_dualspeed(dev->gadget)
957 && (dev->gadget->speed
958 == USB_SPEED_HIGH))
959 power = dev->hs_config->bMaxPower;
960 else
961 power = dev->config->bMaxPower;
962 usb_gadget_vbus_draw(dev->gadget, 2 * power);
963 }
964
965 } else { /* collect OUT data */
966 if ((fd->f_flags & O_NONBLOCK) != 0
967 && !dev->setup_out_ready) {
968 retval = -EAGAIN;
969 goto done;
970 }
971 spin_unlock_irq (&dev->lock);
972 retval = wait_event_interruptible (dev->wait,
973 dev->setup_out_ready != 0);
974
975 /* FIXME state could change from under us */
976 spin_lock_irq (&dev->lock);
977 if (retval)
978 goto done;
979
980 if (dev->state != STATE_DEV_SETUP) {
981 retval = -ECANCELED;
982 goto done;
983 }
984 dev->state = STATE_DEV_CONNECTED;
985
986 if (dev->setup_out_error)
987 retval = -EIO;
988 else {
989 len = min (len, (size_t)dev->req->actual);
990 ++dev->udc_usage;
991 spin_unlock_irq(&dev->lock);
992 if (copy_to_user (buf, dev->req->buf, len))
993 retval = -EFAULT;
994 else
995 retval = len;
996 spin_lock_irq(&dev->lock);
997 --dev->udc_usage;
998 clean_req (dev->gadget->ep0, dev->req);
999 /* NOTE userspace can't yet choose to stall */
1000 }
1001 }
1002 goto done;
1003 }
1004
1005 /* else normal: return event data */
1006 if (len < sizeof dev->event [0]) {
1007 retval = -EINVAL;
1008 goto done;
1009 }
1010 len -= len % sizeof (struct usb_gadgetfs_event);
1011 dev->usermode_setup = 1;
1012
1013scan:
1014 /* return queued events right away */
1015 if (dev->ev_next != 0) {
1016 unsigned i, n;
1017
1018 n = len / sizeof (struct usb_gadgetfs_event);
1019 if (dev->ev_next < n)
1020 n = dev->ev_next;
1021
1022 /* ep0 i/o has special semantics during STATE_DEV_SETUP */
1023 for (i = 0; i < n; i++) {
1024 if (dev->event [i].type == GADGETFS_SETUP) {
1025 dev->state = STATE_DEV_SETUP;
1026 n = i + 1;
1027 break;
1028 }
1029 }
1030 spin_unlock_irq (&dev->lock);
1031 len = n * sizeof (struct usb_gadgetfs_event);
1032 if (copy_to_user (buf, &dev->event, len))
1033 retval = -EFAULT;
1034 else
1035 retval = len;
1036 if (len > 0) {
1037 /* NOTE this doesn't guard against broken drivers;
1038 * concurrent ep0 readers may lose events.
1039 */
1040 spin_lock_irq (&dev->lock);
1041 if (dev->ev_next > n) {
1042 memmove(&dev->event[0], &dev->event[n],
1043 sizeof (struct usb_gadgetfs_event)
1044 * (dev->ev_next - n));
1045 }
1046 dev->ev_next -= n;
1047 spin_unlock_irq (&dev->lock);
1048 }
1049 return retval;
1050 }
1051 if (fd->f_flags & O_NONBLOCK) {
1052 retval = -EAGAIN;
1053 goto done;
1054 }
1055
1056 switch (state) {
1057 default:
1058 DBG (dev, "fail %s, state %d\n", __func__, state);
1059 retval = -ESRCH;
1060 break;
1061 case STATE_DEV_UNCONNECTED:
1062 case STATE_DEV_CONNECTED:
1063 spin_unlock_irq (&dev->lock);
1064 DBG (dev, "%s wait\n", __func__);
1065
1066 /* wait for events */
1067 retval = wait_event_interruptible (dev->wait,
1068 dev->ev_next != 0);
1069 if (retval < 0)
1070 return retval;
1071 spin_lock_irq (&dev->lock);
1072 goto scan;
1073 }
1074
1075done:
1076 spin_unlock_irq (&dev->lock);
1077 return retval;
1078}
1079
1080static struct usb_gadgetfs_event *
1081next_event (struct dev_data *dev, enum usb_gadgetfs_event_type type)
1082{
1083 struct usb_gadgetfs_event *event;
1084 unsigned i;
1085
1086 switch (type) {
1087 /* these events purge the queue */
1088 case GADGETFS_DISCONNECT:
1089 if (dev->state == STATE_DEV_SETUP)
1090 dev->setup_abort = 1;
1091 // FALL THROUGH
1092 case GADGETFS_CONNECT:
1093 dev->ev_next = 0;
1094 break;
1095 case GADGETFS_SETUP: /* previous request timed out */
1096 case GADGETFS_SUSPEND: /* same effect */
1097 /* these events can't be repeated */
1098 for (i = 0; i != dev->ev_next; i++) {
1099 if (dev->event [i].type != type)
1100 continue;
1101 DBG(dev, "discard old event[%d] %d\n", i, type);
1102 dev->ev_next--;
1103 if (i == dev->ev_next)
1104 break;
1105 /* indices start at zero, for simplicity */
1106 memmove (&dev->event [i], &dev->event [i + 1],
1107 sizeof (struct usb_gadgetfs_event)
1108 * (dev->ev_next - i));
1109 }
1110 break;
1111 default:
1112 BUG ();
1113 }
1114 VDEBUG(dev, "event[%d] = %d\n", dev->ev_next, type);
1115 event = &dev->event [dev->ev_next++];
1116 BUG_ON (dev->ev_next > N_EVENT);
1117 memset (event, 0, sizeof *event);
1118 event->type = type;
1119 return event;
1120}
1121
1122static ssize_t
1123ep0_write (struct file *fd, const char __user *buf, size_t len, loff_t *ptr)
1124{
1125 struct dev_data *dev = fd->private_data;
1126 ssize_t retval = -ESRCH;
1127
1128 /* report fd mode change before acting on it */
1129 if (dev->setup_abort) {
1130 dev->setup_abort = 0;
1131 retval = -EIDRM;
1132
1133 /* data and/or status stage for control request */
1134 } else if (dev->state == STATE_DEV_SETUP) {
1135
1136 len = min_t(size_t, len, dev->setup_wLength);
1137 if (dev->setup_in) {
1138 retval = setup_req (dev->gadget->ep0, dev->req, len);
1139 if (retval == 0) {
1140 dev->state = STATE_DEV_CONNECTED;
1141 ++dev->udc_usage;
1142 spin_unlock_irq (&dev->lock);
1143 if (copy_from_user (dev->req->buf, buf, len))
1144 retval = -EFAULT;
1145 else {
1146 if (len < dev->setup_wLength)
1147 dev->req->zero = 1;
1148 retval = usb_ep_queue (
1149 dev->gadget->ep0, dev->req,
1150 GFP_KERNEL);
1151 }
1152 spin_lock_irq(&dev->lock);
1153 --dev->udc_usage;
1154 if (retval < 0) {
1155 clean_req (dev->gadget->ep0, dev->req);
1156 } else
1157 retval = len;
1158
1159 return retval;
1160 }
1161
1162 /* can stall some OUT transfers */
1163 } else if (dev->setup_can_stall) {
1164 VDEBUG(dev, "ep0out stall\n");
1165 (void) usb_ep_set_halt (dev->gadget->ep0);
1166 retval = -EL2HLT;
1167 dev->state = STATE_DEV_CONNECTED;
1168 } else {
1169 DBG(dev, "bogus ep0out stall!\n");
1170 }
1171 } else
1172 DBG (dev, "fail %s, state %d\n", __func__, dev->state);
1173
1174 return retval;
1175}
1176
1177static int
1178ep0_fasync (int f, struct file *fd, int on)
1179{
1180 struct dev_data *dev = fd->private_data;
1181 // caller must F_SETOWN before signal delivery happens
1182 VDEBUG (dev, "%s %s\n", __func__, on ? "on" : "off");
1183 return fasync_helper (f, fd, on, &dev->fasync);
1184}
1185
1186static struct usb_gadget_driver gadgetfs_driver;
1187
1188static int
1189dev_release (struct inode *inode, struct file *fd)
1190{
1191 struct dev_data *dev = fd->private_data;
1192
1193 /* closing ep0 === shutdown all */
1194
1195 if (dev->gadget_registered) {
1196 usb_gadget_unregister_driver (&gadgetfs_driver);
1197 dev->gadget_registered = false;
1198 }
1199
1200 /* at this point "good" hardware has disconnected the
1201 * device from USB; the host won't see it any more.
1202 * alternatively, all host requests will time out.
1203 */
1204
1205 kfree (dev->buf);
1206 dev->buf = NULL;
1207
1208 /* other endpoints were all decoupled from this device */
1209 spin_lock_irq(&dev->lock);
1210 dev->state = STATE_DEV_DISABLED;
1211 spin_unlock_irq(&dev->lock);
1212
1213 put_dev (dev);
1214 return 0;
1215}
1216
1217static __poll_t
1218ep0_poll (struct file *fd, poll_table *wait)
1219{
1220 struct dev_data *dev = fd->private_data;
1221 __poll_t mask = 0;
1222
1223 if (dev->state <= STATE_DEV_OPENED)
1224 return DEFAULT_POLLMASK;
1225
1226 poll_wait(fd, &dev->wait, wait);
1227
1228 spin_lock_irq(&dev->lock);
1229
1230 /* report fd mode change before acting on it */
1231 if (dev->setup_abort) {
1232 dev->setup_abort = 0;
1233 mask = EPOLLHUP;
1234 goto out;
1235 }
1236
1237 if (dev->state == STATE_DEV_SETUP) {
1238 if (dev->setup_in || dev->setup_can_stall)
1239 mask = EPOLLOUT;
1240 } else {
1241 if (dev->ev_next != 0)
1242 mask = EPOLLIN;
1243 }
1244out:
1245 spin_unlock_irq(&dev->lock);
1246 return mask;
1247}
1248
1249static long dev_ioctl (struct file *fd, unsigned code, unsigned long value)
1250{
1251 struct dev_data *dev = fd->private_data;
1252 struct usb_gadget *gadget = dev->gadget;
1253 long ret = -ENOTTY;
1254
1255 spin_lock_irq(&dev->lock);
1256 if (dev->state == STATE_DEV_OPENED ||
1257 dev->state == STATE_DEV_UNBOUND) {
1258 /* Not bound to a UDC */
1259 } else if (gadget->ops->ioctl) {
1260 ++dev->udc_usage;
1261 spin_unlock_irq(&dev->lock);
1262
1263 ret = gadget->ops->ioctl (gadget, code, value);
1264
1265 spin_lock_irq(&dev->lock);
1266 --dev->udc_usage;
1267 }
1268 spin_unlock_irq(&dev->lock);
1269
1270 return ret;
1271}
1272
1273/*----------------------------------------------------------------------*/
1274
1275/* The in-kernel gadget driver handles most ep0 issues, in particular
1276 * enumerating the single configuration (as provided from user space).
1277 *
1278 * Unrecognized ep0 requests may be handled in user space.
1279 */
1280
1281static void make_qualifier (struct dev_data *dev)
1282{
1283 struct usb_qualifier_descriptor qual;
1284 struct usb_device_descriptor *desc;
1285
1286 qual.bLength = sizeof qual;
1287 qual.bDescriptorType = USB_DT_DEVICE_QUALIFIER;
1288 qual.bcdUSB = cpu_to_le16 (0x0200);
1289
1290 desc = dev->dev;
1291 qual.bDeviceClass = desc->bDeviceClass;
1292 qual.bDeviceSubClass = desc->bDeviceSubClass;
1293 qual.bDeviceProtocol = desc->bDeviceProtocol;
1294
1295 /* assumes ep0 uses the same value for both speeds ... */
1296 qual.bMaxPacketSize0 = dev->gadget->ep0->maxpacket;
1297
1298 qual.bNumConfigurations = 1;
1299 qual.bRESERVED = 0;
1300
1301 memcpy (dev->rbuf, &qual, sizeof qual);
1302}
1303
1304static int
1305config_buf (struct dev_data *dev, u8 type, unsigned index)
1306{
1307 int len;
1308 int hs = 0;
1309
1310 /* only one configuration */
1311 if (index > 0)
1312 return -EINVAL;
1313
1314 if (gadget_is_dualspeed(dev->gadget)) {
1315 hs = (dev->gadget->speed == USB_SPEED_HIGH);
1316 if (type == USB_DT_OTHER_SPEED_CONFIG)
1317 hs = !hs;
1318 }
1319 if (hs) {
1320 dev->req->buf = dev->hs_config;
1321 len = le16_to_cpu(dev->hs_config->wTotalLength);
1322 } else {
1323 dev->req->buf = dev->config;
1324 len = le16_to_cpu(dev->config->wTotalLength);
1325 }
1326 ((u8 *)dev->req->buf) [1] = type;
1327 return len;
1328}
1329
1330static int
1331gadgetfs_setup (struct usb_gadget *gadget, const struct usb_ctrlrequest *ctrl)
1332{
1333 struct dev_data *dev = get_gadget_data (gadget);
1334 struct usb_request *req = dev->req;
1335 int value = -EOPNOTSUPP;
1336 struct usb_gadgetfs_event *event;
1337 u16 w_value = le16_to_cpu(ctrl->wValue);
1338 u16 w_length = le16_to_cpu(ctrl->wLength);
1339
1340 if (w_length > RBUF_SIZE) {
1341 if (ctrl->bRequestType & USB_DIR_IN) {
1342 /* Cast away the const, we are going to overwrite on purpose. */
1343 __le16 *temp = (__le16 *)&ctrl->wLength;
1344
1345 *temp = cpu_to_le16(RBUF_SIZE);
1346 w_length = RBUF_SIZE;
1347 } else {
1348 return value;
1349 }
1350 }
1351
1352 spin_lock (&dev->lock);
1353 dev->setup_abort = 0;
1354 if (dev->state == STATE_DEV_UNCONNECTED) {
1355 if (gadget_is_dualspeed(gadget)
1356 && gadget->speed == USB_SPEED_HIGH
1357 && dev->hs_config == NULL) {
1358 spin_unlock(&dev->lock);
1359 ERROR (dev, "no high speed config??\n");
1360 return -EINVAL;
1361 }
1362
1363 dev->state = STATE_DEV_CONNECTED;
1364
1365 INFO (dev, "connected\n");
1366 event = next_event (dev, GADGETFS_CONNECT);
1367 event->u.speed = gadget->speed;
1368 ep0_readable (dev);
1369
1370 /* host may have given up waiting for response. we can miss control
1371 * requests handled lower down (device/endpoint status and features);
1372 * then ep0_{read,write} will report the wrong status. controller
1373 * driver will have aborted pending i/o.
1374 */
1375 } else if (dev->state == STATE_DEV_SETUP)
1376 dev->setup_abort = 1;
1377
1378 req->buf = dev->rbuf;
1379 req->context = NULL;
1380 switch (ctrl->bRequest) {
1381
1382 case USB_REQ_GET_DESCRIPTOR:
1383 if (ctrl->bRequestType != USB_DIR_IN)
1384 goto unrecognized;
1385 switch (w_value >> 8) {
1386
1387 case USB_DT_DEVICE:
1388 value = min (w_length, (u16) sizeof *dev->dev);
1389 dev->dev->bMaxPacketSize0 = dev->gadget->ep0->maxpacket;
1390 req->buf = dev->dev;
1391 break;
1392 case USB_DT_DEVICE_QUALIFIER:
1393 if (!dev->hs_config)
1394 break;
1395 value = min (w_length, (u16)
1396 sizeof (struct usb_qualifier_descriptor));
1397 make_qualifier (dev);
1398 break;
1399 case USB_DT_OTHER_SPEED_CONFIG:
1400 // FALLTHROUGH
1401 case USB_DT_CONFIG:
1402 value = config_buf (dev,
1403 w_value >> 8,
1404 w_value & 0xff);
1405 if (value >= 0)
1406 value = min (w_length, (u16) value);
1407 break;
1408 case USB_DT_STRING:
1409 goto unrecognized;
1410
1411 default: // all others are errors
1412 break;
1413 }
1414 break;
1415
1416 /* currently one config, two speeds */
1417 case USB_REQ_SET_CONFIGURATION:
1418 if (ctrl->bRequestType != 0)
1419 goto unrecognized;
1420 if (0 == (u8) w_value) {
1421 value = 0;
1422 dev->current_config = 0;
1423 usb_gadget_vbus_draw(gadget, 8 /* mA */ );
1424 // user mode expected to disable endpoints
1425 } else {
1426 u8 config, power;
1427
1428 if (gadget_is_dualspeed(gadget)
1429 && gadget->speed == USB_SPEED_HIGH) {
1430 config = dev->hs_config->bConfigurationValue;
1431 power = dev->hs_config->bMaxPower;
1432 } else {
1433 config = dev->config->bConfigurationValue;
1434 power = dev->config->bMaxPower;
1435 }
1436
1437 if (config == (u8) w_value) {
1438 value = 0;
1439 dev->current_config = config;
1440 usb_gadget_vbus_draw(gadget, 2 * power);
1441 }
1442 }
1443
1444 /* report SET_CONFIGURATION like any other control request,
1445 * except that usermode may not stall this. the next
1446 * request mustn't be allowed start until this finishes:
1447 * endpoints and threads set up, etc.
1448 *
1449 * NOTE: older PXA hardware (before PXA 255: without UDCCFR)
1450 * has bad/racey automagic that prevents synchronizing here.
1451 * even kernel mode drivers often miss them.
1452 */
1453 if (value == 0) {
1454 INFO (dev, "configuration #%d\n", dev->current_config);
1455 usb_gadget_set_state(gadget, USB_STATE_CONFIGURED);
1456 if (dev->usermode_setup) {
1457 dev->setup_can_stall = 0;
1458 goto delegate;
1459 }
1460 }
1461 break;
1462
1463#ifndef CONFIG_USB_PXA25X
1464 /* PXA automagically handles this request too */
1465 case USB_REQ_GET_CONFIGURATION:
1466 if (ctrl->bRequestType != 0x80)
1467 goto unrecognized;
1468 *(u8 *)req->buf = dev->current_config;
1469 value = min (w_length, (u16) 1);
1470 break;
1471#endif
1472
1473 default:
1474unrecognized:
1475 VDEBUG (dev, "%s req%02x.%02x v%04x i%04x l%d\n",
1476 dev->usermode_setup ? "delegate" : "fail",
1477 ctrl->bRequestType, ctrl->bRequest,
1478 w_value, le16_to_cpu(ctrl->wIndex), w_length);
1479
1480 /* if there's an ep0 reader, don't stall */
1481 if (dev->usermode_setup) {
1482 dev->setup_can_stall = 1;
1483delegate:
1484 dev->setup_in = (ctrl->bRequestType & USB_DIR_IN)
1485 ? 1 : 0;
1486 dev->setup_wLength = w_length;
1487 dev->setup_out_ready = 0;
1488 dev->setup_out_error = 0;
1489
1490 /* read DATA stage for OUT right away */
1491 if (unlikely (!dev->setup_in && w_length)) {
1492 value = setup_req (gadget->ep0, dev->req,
1493 w_length);
1494 if (value < 0)
1495 break;
1496
1497 ++dev->udc_usage;
1498 spin_unlock (&dev->lock);
1499 value = usb_ep_queue (gadget->ep0, dev->req,
1500 GFP_KERNEL);
1501 spin_lock (&dev->lock);
1502 --dev->udc_usage;
1503 if (value < 0) {
1504 clean_req (gadget->ep0, dev->req);
1505 break;
1506 }
1507
1508 /* we can't currently stall these */
1509 dev->setup_can_stall = 0;
1510 }
1511
1512 /* state changes when reader collects event */
1513 event = next_event (dev, GADGETFS_SETUP);
1514 event->u.setup = *ctrl;
1515 ep0_readable (dev);
1516 spin_unlock (&dev->lock);
1517 return 0;
1518 }
1519 }
1520
1521 /* proceed with data transfer and status phases? */
1522 if (value >= 0 && dev->state != STATE_DEV_SETUP) {
1523 req->length = value;
1524 req->zero = value < w_length;
1525
1526 ++dev->udc_usage;
1527 spin_unlock (&dev->lock);
1528 value = usb_ep_queue (gadget->ep0, req, GFP_KERNEL);
1529 spin_lock(&dev->lock);
1530 --dev->udc_usage;
1531 spin_unlock(&dev->lock);
1532 if (value < 0) {
1533 DBG (dev, "ep_queue --> %d\n", value);
1534 req->status = 0;
1535 }
1536 return value;
1537 }
1538
1539 /* device stalls when value < 0 */
1540 spin_unlock (&dev->lock);
1541 return value;
1542}
1543
1544static void destroy_ep_files (struct dev_data *dev)
1545{
1546 DBG (dev, "%s %d\n", __func__, dev->state);
1547
1548 /* dev->state must prevent interference */
1549 spin_lock_irq (&dev->lock);
1550 while (!list_empty(&dev->epfiles)) {
1551 struct ep_data *ep;
1552 struct inode *parent;
1553 struct dentry *dentry;
1554
1555 /* break link to FS */
1556 ep = list_first_entry (&dev->epfiles, struct ep_data, epfiles);
1557 list_del_init (&ep->epfiles);
1558 spin_unlock_irq (&dev->lock);
1559
1560 dentry = ep->dentry;
1561 ep->dentry = NULL;
1562 parent = d_inode(dentry->d_parent);
1563
1564 /* break link to controller */
1565 mutex_lock(&ep->lock);
1566 if (ep->state == STATE_EP_ENABLED)
1567 (void) usb_ep_disable (ep->ep);
1568 ep->state = STATE_EP_UNBOUND;
1569 usb_ep_free_request (ep->ep, ep->req);
1570 ep->ep = NULL;
1571 mutex_unlock(&ep->lock);
1572
1573 wake_up (&ep->wait);
1574 put_ep (ep);
1575
1576 /* break link to dcache */
1577 inode_lock(parent);
1578 d_delete (dentry);
1579 dput (dentry);
1580 inode_unlock(parent);
1581
1582 spin_lock_irq (&dev->lock);
1583 }
1584 spin_unlock_irq (&dev->lock);
1585}
1586
1587
1588static struct dentry *
1589gadgetfs_create_file (struct super_block *sb, char const *name,
1590 void *data, const struct file_operations *fops);
1591
1592static int activate_ep_files (struct dev_data *dev)
1593{
1594 struct usb_ep *ep;
1595 struct ep_data *data;
1596
1597 gadget_for_each_ep (ep, dev->gadget) {
1598
1599 data = kzalloc(sizeof(*data), GFP_KERNEL);
1600 if (!data)
1601 goto enomem0;
1602 data->state = STATE_EP_DISABLED;
1603 mutex_init(&data->lock);
1604 init_waitqueue_head (&data->wait);
1605
1606 strncpy (data->name, ep->name, sizeof (data->name) - 1);
1607 refcount_set (&data->count, 1);
1608 data->dev = dev;
1609 get_dev (dev);
1610
1611 data->ep = ep;
1612 ep->driver_data = data;
1613
1614 data->req = usb_ep_alloc_request (ep, GFP_KERNEL);
1615 if (!data->req)
1616 goto enomem1;
1617
1618 data->dentry = gadgetfs_create_file (dev->sb, data->name,
1619 data, &ep_io_operations);
1620 if (!data->dentry)
1621 goto enomem2;
1622 list_add_tail (&data->epfiles, &dev->epfiles);
1623 }
1624 return 0;
1625
1626enomem2:
1627 usb_ep_free_request (ep, data->req);
1628enomem1:
1629 put_dev (dev);
1630 kfree (data);
1631enomem0:
1632 DBG (dev, "%s enomem\n", __func__);
1633 destroy_ep_files (dev);
1634 return -ENOMEM;
1635}
1636
1637static void
1638gadgetfs_unbind (struct usb_gadget *gadget)
1639{
1640 struct dev_data *dev = get_gadget_data (gadget);
1641
1642 DBG (dev, "%s\n", __func__);
1643
1644 spin_lock_irq (&dev->lock);
1645 dev->state = STATE_DEV_UNBOUND;
1646 while (dev->udc_usage > 0) {
1647 spin_unlock_irq(&dev->lock);
1648 usleep_range(1000, 2000);
1649 spin_lock_irq(&dev->lock);
1650 }
1651 spin_unlock_irq (&dev->lock);
1652
1653 destroy_ep_files (dev);
1654 gadget->ep0->driver_data = NULL;
1655 set_gadget_data (gadget, NULL);
1656
1657 /* we've already been disconnected ... no i/o is active */
1658 if (dev->req)
1659 usb_ep_free_request (gadget->ep0, dev->req);
1660 DBG (dev, "%s done\n", __func__);
1661 put_dev (dev);
1662}
1663
1664static struct dev_data *the_device;
1665
1666static int gadgetfs_bind(struct usb_gadget *gadget,
1667 struct usb_gadget_driver *driver)
1668{
1669 struct dev_data *dev = the_device;
1670
1671 if (!dev)
1672 return -ESRCH;
1673 if (0 != strcmp (CHIP, gadget->name)) {
1674 pr_err("%s expected %s controller not %s\n",
1675 shortname, CHIP, gadget->name);
1676 return -ENODEV;
1677 }
1678
1679 set_gadget_data (gadget, dev);
1680 dev->gadget = gadget;
1681 gadget->ep0->driver_data = dev;
1682
1683 /* preallocate control response and buffer */
1684 dev->req = usb_ep_alloc_request (gadget->ep0, GFP_KERNEL);
1685 if (!dev->req)
1686 goto enomem;
1687 dev->req->context = NULL;
1688 dev->req->complete = epio_complete;
1689
1690 if (activate_ep_files (dev) < 0)
1691 goto enomem;
1692
1693 INFO (dev, "bound to %s driver\n", gadget->name);
1694 spin_lock_irq(&dev->lock);
1695 dev->state = STATE_DEV_UNCONNECTED;
1696 spin_unlock_irq(&dev->lock);
1697 get_dev (dev);
1698 return 0;
1699
1700enomem:
1701 gadgetfs_unbind (gadget);
1702 return -ENOMEM;
1703}
1704
1705static void
1706gadgetfs_disconnect (struct usb_gadget *gadget)
1707{
1708 struct dev_data *dev = get_gadget_data (gadget);
1709 unsigned long flags;
1710
1711 spin_lock_irqsave (&dev->lock, flags);
1712 if (dev->state == STATE_DEV_UNCONNECTED)
1713 goto exit;
1714 dev->state = STATE_DEV_UNCONNECTED;
1715
1716 INFO (dev, "disconnected\n");
1717 next_event (dev, GADGETFS_DISCONNECT);
1718 ep0_readable (dev);
1719exit:
1720 spin_unlock_irqrestore (&dev->lock, flags);
1721}
1722
1723static void
1724gadgetfs_suspend (struct usb_gadget *gadget)
1725{
1726 struct dev_data *dev = get_gadget_data (gadget);
1727 unsigned long flags;
1728
1729 INFO (dev, "suspended from state %d\n", dev->state);
1730 spin_lock_irqsave(&dev->lock, flags);
1731 switch (dev->state) {
1732 case STATE_DEV_SETUP: // VERY odd... host died??
1733 case STATE_DEV_CONNECTED:
1734 case STATE_DEV_UNCONNECTED:
1735 next_event (dev, GADGETFS_SUSPEND);
1736 ep0_readable (dev);
1737 /* FALLTHROUGH */
1738 default:
1739 break;
1740 }
1741 spin_unlock_irqrestore(&dev->lock, flags);
1742}
1743
1744static struct usb_gadget_driver gadgetfs_driver = {
1745 .function = (char *) driver_desc,
1746 .bind = gadgetfs_bind,
1747 .unbind = gadgetfs_unbind,
1748 .setup = gadgetfs_setup,
1749 .reset = gadgetfs_disconnect,
1750 .disconnect = gadgetfs_disconnect,
1751 .suspend = gadgetfs_suspend,
1752
1753 .driver = {
1754 .name = (char *) shortname,
1755 },
1756};
1757
1758/*----------------------------------------------------------------------*/
1759/* DEVICE INITIALIZATION
1760 *
1761 * fd = open ("/dev/gadget/$CHIP", O_RDWR)
1762 * status = write (fd, descriptors, sizeof descriptors)
1763 *
1764 * That write establishes the device configuration, so the kernel can
1765 * bind to the controller ... guaranteeing it can handle enumeration
1766 * at all necessary speeds. Descriptor order is:
1767 *
1768 * . message tag (u32, host order) ... for now, must be zero; it
1769 * would change to support features like multi-config devices
1770 * . full/low speed config ... all wTotalLength bytes (with interface,
1771 * class, altsetting, endpoint, and other descriptors)
1772 * . high speed config ... all descriptors, for high speed operation;
1773 * this one's optional except for high-speed hardware
1774 * . device descriptor
1775 *
1776 * Endpoints are not yet enabled. Drivers must wait until device
1777 * configuration and interface altsetting changes create
1778 * the need to configure (or unconfigure) them.
1779 *
1780 * After initialization, the device stays active for as long as that
1781 * $CHIP file is open. Events must then be read from that descriptor,
1782 * such as configuration notifications.
1783 */
1784
1785static int is_valid_config(struct usb_config_descriptor *config,
1786 unsigned int total)
1787{
1788 return config->bDescriptorType == USB_DT_CONFIG
1789 && config->bLength == USB_DT_CONFIG_SIZE
1790 && total >= USB_DT_CONFIG_SIZE
1791 && config->bConfigurationValue != 0
1792 && (config->bmAttributes & USB_CONFIG_ATT_ONE) != 0
1793 && (config->bmAttributes & USB_CONFIG_ATT_WAKEUP) == 0;
1794 /* FIXME if gadget->is_otg, _must_ include an otg descriptor */
1795 /* FIXME check lengths: walk to end */
1796}
1797
1798static ssize_t
1799dev_config (struct file *fd, const char __user *buf, size_t len, loff_t *ptr)
1800{
1801 struct dev_data *dev = fd->private_data;
1802 ssize_t value, length = len;
1803 unsigned total;
1804 u32 tag;
1805 char *kbuf;
1806
1807 spin_lock_irq(&dev->lock);
1808 if (dev->state > STATE_DEV_OPENED) {
1809 value = ep0_write(fd, buf, len, ptr);
1810 spin_unlock_irq(&dev->lock);
1811 return value;
1812 }
1813 spin_unlock_irq(&dev->lock);
1814
1815 if ((len < (USB_DT_CONFIG_SIZE + USB_DT_DEVICE_SIZE + 4)) ||
1816 (len > PAGE_SIZE * 4))
1817 return -EINVAL;
1818
1819 /* we might need to change message format someday */
1820 if (copy_from_user (&tag, buf, 4))
1821 return -EFAULT;
1822 if (tag != 0)
1823 return -EINVAL;
1824 buf += 4;
1825 length -= 4;
1826
1827 kbuf = memdup_user(buf, length);
1828 if (IS_ERR(kbuf))
1829 return PTR_ERR(kbuf);
1830
1831 spin_lock_irq (&dev->lock);
1832 value = -EINVAL;
1833 if (dev->buf) {
1834 spin_unlock_irq(&dev->lock);
1835 kfree(kbuf);
1836 return value;
1837 }
1838 dev->buf = kbuf;
1839
1840 /* full or low speed config */
1841 dev->config = (void *) kbuf;
1842 total = le16_to_cpu(dev->config->wTotalLength);
1843 if (!is_valid_config(dev->config, total) ||
1844 total > length - USB_DT_DEVICE_SIZE)
1845 goto fail;
1846 kbuf += total;
1847 length -= total;
1848
1849 /* optional high speed config */
1850 if (kbuf [1] == USB_DT_CONFIG) {
1851 dev->hs_config = (void *) kbuf;
1852 total = le16_to_cpu(dev->hs_config->wTotalLength);
1853 if (!is_valid_config(dev->hs_config, total) ||
1854 total > length - USB_DT_DEVICE_SIZE)
1855 goto fail;
1856 kbuf += total;
1857 length -= total;
1858 } else {
1859 dev->hs_config = NULL;
1860 }
1861
1862 /* could support multiple configs, using another encoding! */
1863
1864 /* device descriptor (tweaked for paranoia) */
1865 if (length != USB_DT_DEVICE_SIZE)
1866 goto fail;
1867 dev->dev = (void *)kbuf;
1868 if (dev->dev->bLength != USB_DT_DEVICE_SIZE
1869 || dev->dev->bDescriptorType != USB_DT_DEVICE
1870 || dev->dev->bNumConfigurations != 1)
1871 goto fail;
1872 dev->dev->bcdUSB = cpu_to_le16 (0x0200);
1873
1874 /* triggers gadgetfs_bind(); then we can enumerate. */
1875 spin_unlock_irq (&dev->lock);
1876 if (dev->hs_config)
1877 gadgetfs_driver.max_speed = USB_SPEED_HIGH;
1878 else
1879 gadgetfs_driver.max_speed = USB_SPEED_FULL;
1880
1881 value = usb_gadget_probe_driver(&gadgetfs_driver);
1882 if (value != 0) {
1883 spin_lock_irq(&dev->lock);
1884 goto fail;
1885 } else {
1886 /* at this point "good" hardware has for the first time
1887 * let the USB the host see us. alternatively, if users
1888 * unplug/replug that will clear all the error state.
1889 *
1890 * note: everything running before here was guaranteed
1891 * to choke driver model style diagnostics. from here
1892 * on, they can work ... except in cleanup paths that
1893 * kick in after the ep0 descriptor is closed.
1894 */
1895 value = len;
1896 dev->gadget_registered = true;
1897 }
1898 return value;
1899
1900fail:
1901 dev->config = NULL;
1902 dev->hs_config = NULL;
1903 dev->dev = NULL;
1904 spin_unlock_irq (&dev->lock);
1905 pr_debug ("%s: %s fail %zd, %p\n", shortname, __func__, value, dev);
1906 kfree (dev->buf);
1907 dev->buf = NULL;
1908 return value;
1909}
1910
1911static int
1912dev_open (struct inode *inode, struct file *fd)
1913{
1914 struct dev_data *dev = inode->i_private;
1915 int value = -EBUSY;
1916
1917 spin_lock_irq(&dev->lock);
1918 if (dev->state == STATE_DEV_DISABLED) {
1919 dev->ev_next = 0;
1920 dev->state = STATE_DEV_OPENED;
1921 fd->private_data = dev;
1922 get_dev (dev);
1923 value = 0;
1924 }
1925 spin_unlock_irq(&dev->lock);
1926 return value;
1927}
1928
1929static const struct file_operations ep0_operations = {
1930 .llseek = no_llseek,
1931
1932 .open = dev_open,
1933 .read = ep0_read,
1934 .write = dev_config,
1935 .fasync = ep0_fasync,
1936 .poll = ep0_poll,
1937 .unlocked_ioctl = dev_ioctl,
1938 .release = dev_release,
1939};
1940
1941/*----------------------------------------------------------------------*/
1942
1943/* FILESYSTEM AND SUPERBLOCK OPERATIONS
1944 *
1945 * Mounting the filesystem creates a controller file, used first for
1946 * device configuration then later for event monitoring.
1947 */
1948
1949
1950/* FIXME PAM etc could set this security policy without mount options
1951 * if epfiles inherited ownership and permissons from ep0 ...
1952 */
1953
1954static unsigned default_uid;
1955static unsigned default_gid;
1956static unsigned default_perm = S_IRUSR | S_IWUSR;
1957
1958module_param (default_uid, uint, 0644);
1959module_param (default_gid, uint, 0644);
1960module_param (default_perm, uint, 0644);
1961
1962
1963static struct inode *
1964gadgetfs_make_inode (struct super_block *sb,
1965 void *data, const struct file_operations *fops,
1966 int mode)
1967{
1968 struct inode *inode = new_inode (sb);
1969
1970 if (inode) {
1971 inode->i_ino = get_next_ino();
1972 inode->i_mode = mode;
1973 inode->i_uid = make_kuid(&init_user_ns, default_uid);
1974 inode->i_gid = make_kgid(&init_user_ns, default_gid);
1975 inode->i_atime = inode->i_mtime = inode->i_ctime
1976 = current_time(inode);
1977 inode->i_private = data;
1978 inode->i_fop = fops;
1979 }
1980 return inode;
1981}
1982
1983/* creates in fs root directory, so non-renamable and non-linkable.
1984 * so inode and dentry are paired, until device reconfig.
1985 */
1986static struct dentry *
1987gadgetfs_create_file (struct super_block *sb, char const *name,
1988 void *data, const struct file_operations *fops)
1989{
1990 struct dentry *dentry;
1991 struct inode *inode;
1992
1993 dentry = d_alloc_name(sb->s_root, name);
1994 if (!dentry)
1995 return NULL;
1996
1997 inode = gadgetfs_make_inode (sb, data, fops,
1998 S_IFREG | (default_perm & S_IRWXUGO));
1999 if (!inode) {
2000 dput(dentry);
2001 return NULL;
2002 }
2003 d_add (dentry, inode);
2004 return dentry;
2005}
2006
2007static const struct super_operations gadget_fs_operations = {
2008 .statfs = simple_statfs,
2009 .drop_inode = generic_delete_inode,
2010};
2011
2012static int
2013gadgetfs_fill_super (struct super_block *sb, struct fs_context *fc)
2014{
2015 struct inode *inode;
2016 struct dev_data *dev;
2017 int rc;
2018
2019 mutex_lock(&sb_mutex);
2020
2021 if (the_device) {
2022 rc = -ESRCH;
2023 goto Done;
2024 }
2025
2026 CHIP = usb_get_gadget_udc_name();
2027 if (!CHIP) {
2028 rc = -ENODEV;
2029 goto Done;
2030 }
2031
2032 /* superblock */
2033 sb->s_blocksize = PAGE_SIZE;
2034 sb->s_blocksize_bits = PAGE_SHIFT;
2035 sb->s_magic = GADGETFS_MAGIC;
2036 sb->s_op = &gadget_fs_operations;
2037 sb->s_time_gran = 1;
2038
2039 /* root inode */
2040 inode = gadgetfs_make_inode (sb,
2041 NULL, &simple_dir_operations,
2042 S_IFDIR | S_IRUGO | S_IXUGO);
2043 if (!inode)
2044 goto Enomem;
2045 inode->i_op = &simple_dir_inode_operations;
2046 if (!(sb->s_root = d_make_root (inode)))
2047 goto Enomem;
2048
2049 /* the ep0 file is named after the controller we expect;
2050 * user mode code can use it for sanity checks, like we do.
2051 */
2052 dev = dev_new ();
2053 if (!dev)
2054 goto Enomem;
2055
2056 dev->sb = sb;
2057 dev->dentry = gadgetfs_create_file(sb, CHIP, dev, &ep0_operations);
2058 if (!dev->dentry) {
2059 put_dev(dev);
2060 goto Enomem;
2061 }
2062
2063 /* other endpoint files are available after hardware setup,
2064 * from binding to a controller.
2065 */
2066 the_device = dev;
2067 rc = 0;
2068 goto Done;
2069
2070 Enomem:
2071 kfree(CHIP);
2072 CHIP = NULL;
2073 rc = -ENOMEM;
2074
2075 Done:
2076 mutex_unlock(&sb_mutex);
2077 return rc;
2078}
2079
2080/* "mount -t gadgetfs path /dev/gadget" ends up here */
2081static int gadgetfs_get_tree(struct fs_context *fc)
2082{
2083 return get_tree_single(fc, gadgetfs_fill_super);
2084}
2085
2086static const struct fs_context_operations gadgetfs_context_ops = {
2087 .get_tree = gadgetfs_get_tree,
2088};
2089
2090static int gadgetfs_init_fs_context(struct fs_context *fc)
2091{
2092 fc->ops = &gadgetfs_context_ops;
2093 return 0;
2094}
2095
2096static void
2097gadgetfs_kill_sb (struct super_block *sb)
2098{
2099 mutex_lock(&sb_mutex);
2100 kill_litter_super (sb);
2101 if (the_device) {
2102 put_dev (the_device);
2103 the_device = NULL;
2104 }
2105 kfree(CHIP);
2106 CHIP = NULL;
2107 mutex_unlock(&sb_mutex);
2108}
2109
2110/*----------------------------------------------------------------------*/
2111
2112static struct file_system_type gadgetfs_type = {
2113 .owner = THIS_MODULE,
2114 .name = shortname,
2115 .init_fs_context = gadgetfs_init_fs_context,
2116 .kill_sb = gadgetfs_kill_sb,
2117};
2118MODULE_ALIAS_FS("gadgetfs");
2119
2120/*----------------------------------------------------------------------*/
2121
2122static int __init init (void)
2123{
2124 int status;
2125
2126 status = register_filesystem (&gadgetfs_type);
2127 if (status == 0)
2128 pr_info ("%s: %s, version " DRIVER_VERSION "\n",
2129 shortname, driver_desc);
2130 return status;
2131}
2132module_init (init);
2133
2134static void __exit cleanup (void)
2135{
2136 pr_debug ("unregister %s\n", shortname);
2137 unregister_filesystem (&gadgetfs_type);
2138}
2139module_exit (cleanup);
2140