| lh | 9ed821d | 2023-04-07 01:36:19 -0700 | [diff] [blame] | 1 | /* | 
|  | 2 | *  History: | 
|  | 3 | *  Started: Aug 9 by Lawrence Foard (entropy@world.std.com), | 
|  | 4 | *           to allow user process control of SCSI devices. | 
|  | 5 | *  Development Sponsored by Killy Corp. NY NY | 
|  | 6 | * | 
|  | 7 | * Original driver (sg.c): | 
|  | 8 | *        Copyright (C) 1992 Lawrence Foard | 
|  | 9 | * Version 2 and 3 extensions to driver: | 
|  | 10 | *        Copyright (C) 1998 - 2005 Douglas Gilbert | 
|  | 11 | * | 
|  | 12 | *  Modified  19-JAN-1998  Richard Gooch <rgooch@atnf.csiro.au>  Devfs support | 
|  | 13 | * | 
|  | 14 | * This program is free software; you can redistribute it and/or modify | 
|  | 15 | * it under the terms of the GNU General Public License as published by | 
|  | 16 | * the Free Software Foundation; either version 2, or (at your option) | 
|  | 17 | * any later version. | 
|  | 18 | * | 
|  | 19 | */ | 
|  | 20 |  | 
|  | 21 | static int sg_version_num = 30534;	/* 2 digits for each component */ | 
|  | 22 | #define SG_VERSION_STR "3.5.34" | 
|  | 23 |  | 
|  | 24 | /* | 
|  | 25 | *  D. P. Gilbert (dgilbert@interlog.com, dougg@triode.net.au), notes: | 
|  | 26 | *      - scsi logging is available via SCSI_LOG_TIMEOUT macros. First | 
|  | 27 | *        the kernel/module needs to be built with CONFIG_SCSI_LOGGING | 
|  | 28 | *        (otherwise the macros compile to empty statements). | 
|  | 29 | * | 
|  | 30 | */ | 
|  | 31 | #include <linux/module.h> | 
|  | 32 |  | 
|  | 33 | #include <linux/fs.h> | 
|  | 34 | #include <linux/kernel.h> | 
|  | 35 | #include <linux/sched.h> | 
|  | 36 | #include <linux/string.h> | 
|  | 37 | #include <linux/mm.h> | 
|  | 38 | #include <linux/errno.h> | 
|  | 39 | #include <linux/mtio.h> | 
|  | 40 | #include <linux/ioctl.h> | 
|  | 41 | #include <linux/slab.h> | 
|  | 42 | #include <linux/fcntl.h> | 
|  | 43 | #include <linux/init.h> | 
|  | 44 | #include <linux/poll.h> | 
|  | 45 | #include <linux/moduleparam.h> | 
|  | 46 | #include <linux/cdev.h> | 
|  | 47 | #include <linux/idr.h> | 
|  | 48 | #include <linux/seq_file.h> | 
|  | 49 | #include <linux/blkdev.h> | 
|  | 50 | #include <linux/delay.h> | 
|  | 51 | #include <linux/blktrace_api.h> | 
|  | 52 | #include <linux/mutex.h> | 
|  | 53 | #include <linux/ratelimit.h> | 
|  | 54 |  | 
|  | 55 | #include "scsi.h" | 
|  | 56 | #include <scsi/scsi_dbg.h> | 
|  | 57 | #include <scsi/scsi_host.h> | 
|  | 58 | #include <scsi/scsi_driver.h> | 
|  | 59 | #include <scsi/scsi_ioctl.h> | 
|  | 60 | #include <scsi/sg.h> | 
|  | 61 |  | 
|  | 62 | #include "scsi_logging.h" | 
|  | 63 |  | 
|  | 64 | #ifdef CONFIG_SCSI_PROC_FS | 
|  | 65 | #include <linux/proc_fs.h> | 
|  | 66 | static char *sg_version_date = "20061027"; | 
|  | 67 |  | 
|  | 68 | static int sg_proc_init(void); | 
|  | 69 | static void sg_proc_cleanup(void); | 
|  | 70 | #endif | 
|  | 71 |  | 
|  | 72 | #define SG_ALLOW_DIO_DEF 0 | 
|  | 73 |  | 
|  | 74 | #define SG_MAX_DEVS 32768 | 
|  | 75 |  | 
|  | 76 | /* | 
|  | 77 | * Suppose you want to calculate the formula muldiv(x,m,d)=int(x * m / d) | 
|  | 78 | * Then when using 32 bit integers x * m may overflow during the calculation. | 
|  | 79 | * Replacing muldiv(x) by muldiv(x)=((x % d) * m) / d + int(x / d) * m | 
|  | 80 | * calculates the same, but prevents the overflow when both m and d | 
|  | 81 | * are "small" numbers (like HZ and USER_HZ). | 
|  | 82 | * Of course an overflow is inavoidable if the result of muldiv doesn't fit | 
|  | 83 | * in 32 bits. | 
|  | 84 | */ | 
|  | 85 | #define MULDIV(X,MUL,DIV) ((((X % DIV) * MUL) / DIV) + ((X / DIV) * MUL)) | 
|  | 86 |  | 
|  | 87 | #define SG_DEFAULT_TIMEOUT MULDIV(SG_DEFAULT_TIMEOUT_USER, HZ, USER_HZ) | 
|  | 88 |  | 
|  | 89 | int sg_big_buff = SG_DEF_RESERVED_SIZE; | 
|  | 90 | /* N.B. This variable is readable and writeable via | 
|  | 91 | /proc/scsi/sg/def_reserved_size . Each time sg_open() is called a buffer | 
|  | 92 | of this size (or less if there is not enough memory) will be reserved | 
|  | 93 | for use by this file descriptor. [Deprecated usage: this variable is also | 
|  | 94 | readable via /proc/sys/kernel/sg-big-buff if the sg driver is built into | 
|  | 95 | the kernel (i.e. it is not a module).] */ | 
|  | 96 | static int def_reserved_size = -1;	/* picks up init parameter */ | 
|  | 97 | static int sg_allow_dio = SG_ALLOW_DIO_DEF; | 
|  | 98 |  | 
|  | 99 | static int scatter_elem_sz = SG_SCATTER_SZ; | 
|  | 100 | static int scatter_elem_sz_prev = SG_SCATTER_SZ; | 
|  | 101 |  | 
|  | 102 | #define SG_SECTOR_SZ 512 | 
|  | 103 |  | 
|  | 104 | static int sg_add(struct device *, struct class_interface *); | 
|  | 105 | static void sg_remove(struct device *, struct class_interface *); | 
|  | 106 |  | 
|  | 107 | static DEFINE_MUTEX(sg_mutex); | 
|  | 108 |  | 
|  | 109 | static DEFINE_IDR(sg_index_idr); | 
|  | 110 | static DEFINE_RWLOCK(sg_index_lock);	/* Also used to lock | 
|  | 111 | file descriptor list for device */ | 
|  | 112 |  | 
|  | 113 | static struct class_interface sg_interface = { | 
|  | 114 | .add_dev	= sg_add, | 
|  | 115 | .remove_dev	= sg_remove, | 
|  | 116 | }; | 
|  | 117 |  | 
|  | 118 | typedef struct sg_scatter_hold { /* holding area for scsi scatter gather info */ | 
|  | 119 | unsigned short k_use_sg; /* Count of kernel scatter-gather pieces */ | 
|  | 120 | unsigned sglist_len; /* size of malloc'd scatter-gather list ++ */ | 
|  | 121 | unsigned bufflen;	/* Size of (aggregate) data buffer */ | 
|  | 122 | struct page **pages; | 
|  | 123 | int page_order; | 
|  | 124 | char dio_in_use;	/* 0->indirect IO (or mmap), 1->dio */ | 
|  | 125 | unsigned char cmd_opcode; /* first byte of command */ | 
|  | 126 | } Sg_scatter_hold; | 
|  | 127 |  | 
|  | 128 | struct sg_device;		/* forward declarations */ | 
|  | 129 | struct sg_fd; | 
|  | 130 |  | 
|  | 131 | typedef struct sg_request {	/* SG_MAX_QUEUE requests outstanding per file */ | 
|  | 132 | struct sg_request *nextrp;	/* NULL -> tail request (slist) */ | 
|  | 133 | struct sg_fd *parentfp;	/* NULL -> not in use */ | 
|  | 134 | Sg_scatter_hold data;	/* hold buffer, perhaps scatter list */ | 
|  | 135 | sg_io_hdr_t header;	/* scsi command+info, see <scsi/sg.h> */ | 
|  | 136 | unsigned char sense_b[SCSI_SENSE_BUFFERSIZE]; | 
|  | 137 | char res_used;		/* 1 -> using reserve buffer, 0 -> not ... */ | 
|  | 138 | char orphan;		/* 1 -> drop on sight, 0 -> normal */ | 
|  | 139 | char sg_io_owned;	/* 1 -> packet belongs to SG_IO */ | 
|  | 140 | volatile char done;	/* 0->before bh, 1->before read, 2->read */ | 
|  | 141 | struct request *rq; | 
|  | 142 | struct bio *bio; | 
|  | 143 | struct execute_work ew; | 
|  | 144 | } Sg_request; | 
|  | 145 |  | 
|  | 146 | typedef struct sg_fd {		/* holds the state of a file descriptor */ | 
|  | 147 | struct list_head sfd_siblings; | 
|  | 148 | struct sg_device *parentdp;	/* owning device */ | 
|  | 149 | wait_queue_head_t read_wait;	/* queue read until command done */ | 
|  | 150 | rwlock_t rq_list_lock;	/* protect access to list in req_arr */ | 
|  | 151 | int timeout;		/* defaults to SG_DEFAULT_TIMEOUT      */ | 
|  | 152 | int timeout_user;	/* defaults to SG_DEFAULT_TIMEOUT_USER */ | 
|  | 153 | Sg_scatter_hold reserve;	/* buffer held for this file descriptor */ | 
|  | 154 | unsigned save_scat_len;	/* original length of trunc. scat. element */ | 
|  | 155 | Sg_request *headrp;	/* head of request slist, NULL->empty */ | 
|  | 156 | struct fasync_struct *async_qp;	/* used by asynchronous notification */ | 
|  | 157 | Sg_request req_arr[SG_MAX_QUEUE];	/* used as singly-linked list */ | 
|  | 158 | char low_dma;		/* as in parent but possibly overridden to 1 */ | 
|  | 159 | char force_packid;	/* 1 -> pack_id input to read(), 0 -> ignored */ | 
|  | 160 | volatile char closed;	/* 1 -> fd closed but request(s) outstanding */ | 
|  | 161 | char cmd_q;		/* 1 -> allow command queuing, 0 -> don't */ | 
|  | 162 | char next_cmd_len;	/* 0 -> automatic (def), >0 -> use on next write() */ | 
|  | 163 | char keep_orphan;	/* 0 -> drop orphan (def), 1 -> keep for read() */ | 
|  | 164 | char mmap_called;	/* 0 -> mmap() never called on this fd */ | 
|  | 165 | struct kref f_ref; | 
|  | 166 | struct execute_work ew; | 
|  | 167 | } Sg_fd; | 
|  | 168 |  | 
|  | 169 | typedef struct sg_device { /* holds the state of each scsi generic device */ | 
|  | 170 | struct scsi_device *device; | 
|  | 171 | wait_queue_head_t o_excl_wait;	/* queue open() when O_EXCL in use */ | 
|  | 172 | int sg_tablesize;	/* adapter's max scatter-gather table size */ | 
|  | 173 | u32 index;		/* device index number */ | 
|  | 174 | struct list_head sfds; | 
|  | 175 | volatile char detached;	/* 0->attached, 1->detached pending removal */ | 
|  | 176 | volatile char exclude;	/* opened for exclusive access */ | 
|  | 177 | char sgdebug;		/* 0->off, 1->sense, 9->dump dev, 10-> all devs */ | 
|  | 178 | struct gendisk *disk; | 
|  | 179 | struct cdev * cdev;	/* char_dev [sysfs: /sys/cdev/major/sg<n>] */ | 
|  | 180 | struct kref d_ref; | 
|  | 181 | } Sg_device; | 
|  | 182 |  | 
|  | 183 | /* tasklet or soft irq callback */ | 
|  | 184 | static void sg_rq_end_io(struct request *rq, int uptodate); | 
|  | 185 | static int sg_start_req(Sg_request *srp, unsigned char *cmd); | 
|  | 186 | static int sg_finish_rem_req(Sg_request * srp); | 
|  | 187 | static int sg_build_indirect(Sg_scatter_hold * schp, Sg_fd * sfp, int buff_size); | 
|  | 188 | static ssize_t sg_new_read(Sg_fd * sfp, char __user *buf, size_t count, | 
|  | 189 | Sg_request * srp); | 
|  | 190 | static ssize_t sg_new_write(Sg_fd *sfp, struct file *file, | 
|  | 191 | const char __user *buf, size_t count, int blocking, | 
|  | 192 | int read_only, int sg_io_owned, Sg_request **o_srp); | 
|  | 193 | static int sg_common_write(Sg_fd * sfp, Sg_request * srp, | 
|  | 194 | unsigned char *cmnd, int timeout, int blocking); | 
|  | 195 | static int sg_read_oxfer(Sg_request * srp, char __user *outp, int num_read_xfer); | 
|  | 196 | static void sg_remove_scat(Sg_scatter_hold * schp); | 
|  | 197 | static void sg_build_reserve(Sg_fd * sfp, int req_size); | 
|  | 198 | static void sg_link_reserve(Sg_fd * sfp, Sg_request * srp, int size); | 
|  | 199 | static void sg_unlink_reserve(Sg_fd * sfp, Sg_request * srp); | 
|  | 200 | static Sg_fd *sg_add_sfp(Sg_device * sdp, int dev); | 
|  | 201 | static void sg_remove_sfp(struct kref *); | 
|  | 202 | static Sg_request *sg_get_rq_mark(Sg_fd * sfp, int pack_id); | 
|  | 203 | static Sg_request *sg_add_request(Sg_fd * sfp); | 
|  | 204 | static int sg_remove_request(Sg_fd * sfp, Sg_request * srp); | 
|  | 205 | static int sg_res_in_use(Sg_fd * sfp); | 
|  | 206 | static Sg_device *sg_get_dev(int dev); | 
|  | 207 | static void sg_put_dev(Sg_device *sdp); | 
|  | 208 |  | 
|  | 209 | #define SZ_SG_HEADER sizeof(struct sg_header) | 
|  | 210 | #define SZ_SG_IO_HDR sizeof(sg_io_hdr_t) | 
|  | 211 | #define SZ_SG_IOVEC sizeof(sg_iovec_t) | 
|  | 212 | #define SZ_SG_REQ_INFO sizeof(sg_req_info_t) | 
|  | 213 |  | 
|  | 214 | static int sg_allow_access(struct file *filp, unsigned char *cmd) | 
|  | 215 | { | 
|  | 216 | struct sg_fd *sfp = filp->private_data; | 
|  | 217 |  | 
|  | 218 | if (sfp->parentdp->device->type == TYPE_SCANNER) | 
|  | 219 | return 0; | 
|  | 220 |  | 
|  | 221 | return blk_verify_command(cmd, filp->f_mode & FMODE_WRITE); | 
|  | 222 | } | 
|  | 223 |  | 
|  | 224 | static int | 
|  | 225 | sg_open(struct inode *inode, struct file *filp) | 
|  | 226 | { | 
|  | 227 | int dev = iminor(inode); | 
|  | 228 | int flags = filp->f_flags; | 
|  | 229 | struct request_queue *q; | 
|  | 230 | Sg_device *sdp; | 
|  | 231 | Sg_fd *sfp; | 
|  | 232 | int res; | 
|  | 233 | int retval; | 
|  | 234 |  | 
|  | 235 | mutex_lock(&sg_mutex); | 
|  | 236 | nonseekable_open(inode, filp); | 
|  | 237 | SCSI_LOG_TIMEOUT(3, printk("sg_open: dev=%d, flags=0x%x\n", dev, flags)); | 
|  | 238 | sdp = sg_get_dev(dev); | 
|  | 239 | if (IS_ERR(sdp)) { | 
|  | 240 | retval = PTR_ERR(sdp); | 
|  | 241 | sdp = NULL; | 
|  | 242 | goto sg_put; | 
|  | 243 | } | 
|  | 244 |  | 
|  | 245 | /* This driver's module count bumped by fops_get in <linux/fs.h> */ | 
|  | 246 | /* Prevent the device driver from vanishing while we sleep */ | 
|  | 247 | retval = scsi_device_get(sdp->device); | 
|  | 248 | if (retval) | 
|  | 249 | goto sg_put; | 
|  | 250 |  | 
|  | 251 | retval = scsi_autopm_get_device(sdp->device); | 
|  | 252 | if (retval) | 
|  | 253 | goto sdp_put; | 
|  | 254 |  | 
|  | 255 | if (!((flags & O_NONBLOCK) || | 
|  | 256 | scsi_block_when_processing_errors(sdp->device))) { | 
|  | 257 | retval = -ENXIO; | 
|  | 258 | /* we are in error recovery for this device */ | 
|  | 259 | goto error_out; | 
|  | 260 | } | 
|  | 261 |  | 
|  | 262 | if (flags & O_EXCL) { | 
|  | 263 | if (O_RDONLY == (flags & O_ACCMODE)) { | 
|  | 264 | retval = -EPERM; /* Can't lock it with read only access */ | 
|  | 265 | goto error_out; | 
|  | 266 | } | 
|  | 267 | if (!list_empty(&sdp->sfds) && (flags & O_NONBLOCK)) { | 
|  | 268 | retval = -EBUSY; | 
|  | 269 | goto error_out; | 
|  | 270 | } | 
|  | 271 | res = 0; | 
|  | 272 | __wait_event_interruptible(sdp->o_excl_wait, | 
|  | 273 | ((!list_empty(&sdp->sfds) || sdp->exclude) ? 0 : (sdp->exclude = 1)), res); | 
|  | 274 | if (res) { | 
|  | 275 | retval = res;	/* -ERESTARTSYS because signal hit process */ | 
|  | 276 | goto error_out; | 
|  | 277 | } | 
|  | 278 | } else if (sdp->exclude) {	/* some other fd has an exclusive lock on dev */ | 
|  | 279 | if (flags & O_NONBLOCK) { | 
|  | 280 | retval = -EBUSY; | 
|  | 281 | goto error_out; | 
|  | 282 | } | 
|  | 283 | res = 0; | 
|  | 284 | __wait_event_interruptible(sdp->o_excl_wait, (!sdp->exclude), | 
|  | 285 | res); | 
|  | 286 | if (res) { | 
|  | 287 | retval = res;	/* -ERESTARTSYS because signal hit process */ | 
|  | 288 | goto error_out; | 
|  | 289 | } | 
|  | 290 | } | 
|  | 291 | if (sdp->detached) { | 
|  | 292 | retval = -ENODEV; | 
|  | 293 | goto error_out; | 
|  | 294 | } | 
|  | 295 | if (list_empty(&sdp->sfds)) {	/* no existing opens on this device */ | 
|  | 296 | sdp->sgdebug = 0; | 
|  | 297 | q = sdp->device->request_queue; | 
|  | 298 | sdp->sg_tablesize = queue_max_segments(q); | 
|  | 299 | } | 
|  | 300 | if ((sfp = sg_add_sfp(sdp, dev))) | 
|  | 301 | filp->private_data = sfp; | 
|  | 302 | else { | 
|  | 303 | if (flags & O_EXCL) { | 
|  | 304 | sdp->exclude = 0;	/* undo if error */ | 
|  | 305 | wake_up_interruptible(&sdp->o_excl_wait); | 
|  | 306 | } | 
|  | 307 | retval = -ENOMEM; | 
|  | 308 | goto error_out; | 
|  | 309 | } | 
|  | 310 | retval = 0; | 
|  | 311 | error_out: | 
|  | 312 | if (retval) { | 
|  | 313 | scsi_autopm_put_device(sdp->device); | 
|  | 314 | sdp_put: | 
|  | 315 | scsi_device_put(sdp->device); | 
|  | 316 | } | 
|  | 317 | sg_put: | 
|  | 318 | if (sdp) | 
|  | 319 | sg_put_dev(sdp); | 
|  | 320 | mutex_unlock(&sg_mutex); | 
|  | 321 | return retval; | 
|  | 322 | } | 
|  | 323 |  | 
|  | 324 | /* Following function was formerly called 'sg_close' */ | 
|  | 325 | static int | 
|  | 326 | sg_release(struct inode *inode, struct file *filp) | 
|  | 327 | { | 
|  | 328 | Sg_device *sdp; | 
|  | 329 | Sg_fd *sfp; | 
|  | 330 |  | 
|  | 331 | if ((!(sfp = (Sg_fd *) filp->private_data)) || (!(sdp = sfp->parentdp))) | 
|  | 332 | return -ENXIO; | 
|  | 333 | SCSI_LOG_TIMEOUT(3, printk("sg_release: %s\n", sdp->disk->disk_name)); | 
|  | 334 |  | 
|  | 335 | sfp->closed = 1; | 
|  | 336 |  | 
|  | 337 | sdp->exclude = 0; | 
|  | 338 | wake_up_interruptible(&sdp->o_excl_wait); | 
|  | 339 |  | 
|  | 340 | scsi_autopm_put_device(sdp->device); | 
|  | 341 | kref_put(&sfp->f_ref, sg_remove_sfp); | 
|  | 342 | return 0; | 
|  | 343 | } | 
|  | 344 |  | 
|  | 345 | static ssize_t | 
|  | 346 | sg_read(struct file *filp, char __user *buf, size_t count, loff_t * ppos) | 
|  | 347 | { | 
|  | 348 | Sg_device *sdp; | 
|  | 349 | Sg_fd *sfp; | 
|  | 350 | Sg_request *srp; | 
|  | 351 | int req_pack_id = -1; | 
|  | 352 | sg_io_hdr_t *hp; | 
|  | 353 | struct sg_header *old_hdr = NULL; | 
|  | 354 | int retval = 0; | 
|  | 355 |  | 
|  | 356 | if ((!(sfp = (Sg_fd *) filp->private_data)) || (!(sdp = sfp->parentdp))) | 
|  | 357 | return -ENXIO; | 
|  | 358 | SCSI_LOG_TIMEOUT(3, printk("sg_read: %s, count=%d\n", | 
|  | 359 | sdp->disk->disk_name, (int) count)); | 
|  | 360 |  | 
|  | 361 | if (!access_ok(VERIFY_WRITE, buf, count)) | 
|  | 362 | return -EFAULT; | 
|  | 363 | if (sfp->force_packid && (count >= SZ_SG_HEADER)) { | 
|  | 364 | old_hdr = kmalloc(SZ_SG_HEADER, GFP_KERNEL); | 
|  | 365 | if (!old_hdr) | 
|  | 366 | return -ENOMEM; | 
|  | 367 | if (__copy_from_user(old_hdr, buf, SZ_SG_HEADER)) { | 
|  | 368 | retval = -EFAULT; | 
|  | 369 | goto free_old_hdr; | 
|  | 370 | } | 
|  | 371 | if (old_hdr->reply_len < 0) { | 
|  | 372 | if (count >= SZ_SG_IO_HDR) { | 
|  | 373 | sg_io_hdr_t *new_hdr; | 
|  | 374 | new_hdr = kmalloc(SZ_SG_IO_HDR, GFP_KERNEL); | 
|  | 375 | if (!new_hdr) { | 
|  | 376 | retval = -ENOMEM; | 
|  | 377 | goto free_old_hdr; | 
|  | 378 | } | 
|  | 379 | retval =__copy_from_user | 
|  | 380 | (new_hdr, buf, SZ_SG_IO_HDR); | 
|  | 381 | req_pack_id = new_hdr->pack_id; | 
|  | 382 | kfree(new_hdr); | 
|  | 383 | if (retval) { | 
|  | 384 | retval = -EFAULT; | 
|  | 385 | goto free_old_hdr; | 
|  | 386 | } | 
|  | 387 | } | 
|  | 388 | } else | 
|  | 389 | req_pack_id = old_hdr->pack_id; | 
|  | 390 | } | 
|  | 391 | srp = sg_get_rq_mark(sfp, req_pack_id); | 
|  | 392 | if (!srp) {		/* now wait on packet to arrive */ | 
|  | 393 | if (sdp->detached) { | 
|  | 394 | retval = -ENODEV; | 
|  | 395 | goto free_old_hdr; | 
|  | 396 | } | 
|  | 397 | if (filp->f_flags & O_NONBLOCK) { | 
|  | 398 | retval = -EAGAIN; | 
|  | 399 | goto free_old_hdr; | 
|  | 400 | } | 
|  | 401 | while (1) { | 
|  | 402 | retval = 0; /* following macro beats race condition */ | 
|  | 403 | __wait_event_interruptible(sfp->read_wait, | 
|  | 404 | (sdp->detached || | 
|  | 405 | (srp = sg_get_rq_mark(sfp, req_pack_id))), | 
|  | 406 | retval); | 
|  | 407 | if (sdp->detached) { | 
|  | 408 | retval = -ENODEV; | 
|  | 409 | goto free_old_hdr; | 
|  | 410 | } | 
|  | 411 | if (0 == retval) | 
|  | 412 | break; | 
|  | 413 |  | 
|  | 414 | /* -ERESTARTSYS as signal hit process */ | 
|  | 415 | goto free_old_hdr; | 
|  | 416 | } | 
|  | 417 | } | 
|  | 418 | if (srp->header.interface_id != '\0') { | 
|  | 419 | retval = sg_new_read(sfp, buf, count, srp); | 
|  | 420 | goto free_old_hdr; | 
|  | 421 | } | 
|  | 422 |  | 
|  | 423 | hp = &srp->header; | 
|  | 424 | if (old_hdr == NULL) { | 
|  | 425 | old_hdr = kmalloc(SZ_SG_HEADER, GFP_KERNEL); | 
|  | 426 | if (! old_hdr) { | 
|  | 427 | retval = -ENOMEM; | 
|  | 428 | goto free_old_hdr; | 
|  | 429 | } | 
|  | 430 | } | 
|  | 431 | memset(old_hdr, 0, SZ_SG_HEADER); | 
|  | 432 | old_hdr->reply_len = (int) hp->timeout; | 
|  | 433 | old_hdr->pack_len = old_hdr->reply_len; /* old, strange behaviour */ | 
|  | 434 | old_hdr->pack_id = hp->pack_id; | 
|  | 435 | old_hdr->twelve_byte = | 
|  | 436 | ((srp->data.cmd_opcode >= 0xc0) && (12 == hp->cmd_len)) ? 1 : 0; | 
|  | 437 | old_hdr->target_status = hp->masked_status; | 
|  | 438 | old_hdr->host_status = hp->host_status; | 
|  | 439 | old_hdr->driver_status = hp->driver_status; | 
|  | 440 | if ((CHECK_CONDITION & hp->masked_status) || | 
|  | 441 | (DRIVER_SENSE & hp->driver_status)) | 
|  | 442 | memcpy(old_hdr->sense_buffer, srp->sense_b, | 
|  | 443 | sizeof (old_hdr->sense_buffer)); | 
|  | 444 | switch (hp->host_status) { | 
|  | 445 | /* This setup of 'result' is for backward compatibility and is best | 
|  | 446 | ignored by the user who should use target, host + driver status */ | 
|  | 447 | case DID_OK: | 
|  | 448 | case DID_PASSTHROUGH: | 
|  | 449 | case DID_SOFT_ERROR: | 
|  | 450 | old_hdr->result = 0; | 
|  | 451 | break; | 
|  | 452 | case DID_NO_CONNECT: | 
|  | 453 | case DID_BUS_BUSY: | 
|  | 454 | case DID_TIME_OUT: | 
|  | 455 | old_hdr->result = EBUSY; | 
|  | 456 | break; | 
|  | 457 | case DID_BAD_TARGET: | 
|  | 458 | case DID_ABORT: | 
|  | 459 | case DID_PARITY: | 
|  | 460 | case DID_RESET: | 
|  | 461 | case DID_BAD_INTR: | 
|  | 462 | old_hdr->result = EIO; | 
|  | 463 | break; | 
|  | 464 | case DID_ERROR: | 
|  | 465 | old_hdr->result = (srp->sense_b[0] == 0 && | 
|  | 466 | hp->masked_status == GOOD) ? 0 : EIO; | 
|  | 467 | break; | 
|  | 468 | default: | 
|  | 469 | old_hdr->result = EIO; | 
|  | 470 | break; | 
|  | 471 | } | 
|  | 472 |  | 
|  | 473 | /* Now copy the result back to the user buffer.  */ | 
|  | 474 | if (count >= SZ_SG_HEADER) { | 
|  | 475 | if (__copy_to_user(buf, old_hdr, SZ_SG_HEADER)) { | 
|  | 476 | retval = -EFAULT; | 
|  | 477 | goto free_old_hdr; | 
|  | 478 | } | 
|  | 479 | buf += SZ_SG_HEADER; | 
|  | 480 | if (count > old_hdr->reply_len) | 
|  | 481 | count = old_hdr->reply_len; | 
|  | 482 | if (count > SZ_SG_HEADER) { | 
|  | 483 | if (sg_read_oxfer(srp, buf, count - SZ_SG_HEADER)) { | 
|  | 484 | retval = -EFAULT; | 
|  | 485 | goto free_old_hdr; | 
|  | 486 | } | 
|  | 487 | } | 
|  | 488 | } else | 
|  | 489 | count = (old_hdr->result == 0) ? 0 : -EIO; | 
|  | 490 | sg_finish_rem_req(srp); | 
|  | 491 | retval = count; | 
|  | 492 | free_old_hdr: | 
|  | 493 | kfree(old_hdr); | 
|  | 494 | return retval; | 
|  | 495 | } | 
|  | 496 |  | 
|  | 497 | static ssize_t | 
|  | 498 | sg_new_read(Sg_fd * sfp, char __user *buf, size_t count, Sg_request * srp) | 
|  | 499 | { | 
|  | 500 | sg_io_hdr_t *hp = &srp->header; | 
|  | 501 | int err = 0, err2; | 
|  | 502 | int len; | 
|  | 503 |  | 
|  | 504 | if (count < SZ_SG_IO_HDR) { | 
|  | 505 | err = -EINVAL; | 
|  | 506 | goto err_out; | 
|  | 507 | } | 
|  | 508 | hp->sb_len_wr = 0; | 
|  | 509 | if ((hp->mx_sb_len > 0) && hp->sbp) { | 
|  | 510 | if ((CHECK_CONDITION & hp->masked_status) || | 
|  | 511 | (DRIVER_SENSE & hp->driver_status)) { | 
|  | 512 | int sb_len = SCSI_SENSE_BUFFERSIZE; | 
|  | 513 | sb_len = (hp->mx_sb_len > sb_len) ? sb_len : hp->mx_sb_len; | 
|  | 514 | len = 8 + (int) srp->sense_b[7];	/* Additional sense length field */ | 
|  | 515 | len = (len > sb_len) ? sb_len : len; | 
|  | 516 | if (copy_to_user(hp->sbp, srp->sense_b, len)) { | 
|  | 517 | err = -EFAULT; | 
|  | 518 | goto err_out; | 
|  | 519 | } | 
|  | 520 | hp->sb_len_wr = len; | 
|  | 521 | } | 
|  | 522 | } | 
|  | 523 | if (hp->masked_status || hp->host_status || hp->driver_status) | 
|  | 524 | hp->info |= SG_INFO_CHECK; | 
|  | 525 | if (copy_to_user(buf, hp, SZ_SG_IO_HDR)) { | 
|  | 526 | err = -EFAULT; | 
|  | 527 | goto err_out; | 
|  | 528 | } | 
|  | 529 | err_out: | 
|  | 530 | err2 = sg_finish_rem_req(srp); | 
|  | 531 | return err ? : err2 ? : count; | 
|  | 532 | } | 
|  | 533 |  | 
|  | 534 | static ssize_t | 
|  | 535 | sg_write(struct file *filp, const char __user *buf, size_t count, loff_t * ppos) | 
|  | 536 | { | 
|  | 537 | int mxsize, cmd_size, k; | 
|  | 538 | int input_size, blocking; | 
|  | 539 | unsigned char opcode; | 
|  | 540 | Sg_device *sdp; | 
|  | 541 | Sg_fd *sfp; | 
|  | 542 | Sg_request *srp; | 
|  | 543 | struct sg_header old_hdr; | 
|  | 544 | sg_io_hdr_t *hp; | 
|  | 545 | unsigned char cmnd[MAX_COMMAND_SIZE]; | 
|  | 546 |  | 
|  | 547 | if(unlikely(segment_eq(get_fs(), KERNEL_DS))) | 
|  | 548 | return -EINVAL; | 
|  | 549 |  | 
|  | 550 | if ((!(sfp = (Sg_fd *) filp->private_data)) || (!(sdp = sfp->parentdp))) | 
|  | 551 | return -ENXIO; | 
|  | 552 | SCSI_LOG_TIMEOUT(3, printk("sg_write: %s, count=%d\n", | 
|  | 553 | sdp->disk->disk_name, (int) count)); | 
|  | 554 | if (sdp->detached) | 
|  | 555 | return -ENODEV; | 
|  | 556 | if (!((filp->f_flags & O_NONBLOCK) || | 
|  | 557 | scsi_block_when_processing_errors(sdp->device))) | 
|  | 558 | return -ENXIO; | 
|  | 559 |  | 
|  | 560 | if (!access_ok(VERIFY_READ, buf, count)) | 
|  | 561 | return -EFAULT;	/* protects following copy_from_user()s + get_user()s */ | 
|  | 562 | if (count < SZ_SG_HEADER) | 
|  | 563 | return -EIO; | 
|  | 564 | if (__copy_from_user(&old_hdr, buf, SZ_SG_HEADER)) | 
|  | 565 | return -EFAULT; | 
|  | 566 | blocking = !(filp->f_flags & O_NONBLOCK); | 
|  | 567 | if (old_hdr.reply_len < 0) | 
|  | 568 | return sg_new_write(sfp, filp, buf, count, | 
|  | 569 | blocking, 0, 0, NULL); | 
|  | 570 | if (count < (SZ_SG_HEADER + 6)) | 
|  | 571 | return -EIO;	/* The minimum scsi command length is 6 bytes. */ | 
|  | 572 |  | 
|  | 573 | if (!(srp = sg_add_request(sfp))) { | 
|  | 574 | SCSI_LOG_TIMEOUT(1, printk("sg_write: queue full\n")); | 
|  | 575 | return -EDOM; | 
|  | 576 | } | 
|  | 577 | buf += SZ_SG_HEADER; | 
|  | 578 | __get_user(opcode, buf); | 
|  | 579 | if (sfp->next_cmd_len > 0) { | 
|  | 580 | if (sfp->next_cmd_len > MAX_COMMAND_SIZE) { | 
|  | 581 | SCSI_LOG_TIMEOUT(1, printk("sg_write: command length too long\n")); | 
|  | 582 | sfp->next_cmd_len = 0; | 
|  | 583 | sg_remove_request(sfp, srp); | 
|  | 584 | return -EIO; | 
|  | 585 | } | 
|  | 586 | cmd_size = sfp->next_cmd_len; | 
|  | 587 | sfp->next_cmd_len = 0;	/* reset so only this write() effected */ | 
|  | 588 | } else { | 
|  | 589 | cmd_size = COMMAND_SIZE(opcode);	/* based on SCSI command group */ | 
|  | 590 | if ((opcode >= 0xc0) && old_hdr.twelve_byte) | 
|  | 591 | cmd_size = 12; | 
|  | 592 | } | 
|  | 593 | SCSI_LOG_TIMEOUT(4, printk( | 
|  | 594 | "sg_write:   scsi opcode=0x%02x, cmd_size=%d\n", (int) opcode, cmd_size)); | 
|  | 595 | /* Determine buffer size.  */ | 
|  | 596 | input_size = count - cmd_size; | 
|  | 597 | mxsize = (input_size > old_hdr.reply_len) ? input_size : old_hdr.reply_len; | 
|  | 598 | mxsize -= SZ_SG_HEADER; | 
|  | 599 | input_size -= SZ_SG_HEADER; | 
|  | 600 | if (input_size < 0) { | 
|  | 601 | sg_remove_request(sfp, srp); | 
|  | 602 | return -EIO;	/* User did not pass enough bytes for this command. */ | 
|  | 603 | } | 
|  | 604 | hp = &srp->header; | 
|  | 605 | hp->interface_id = '\0';	/* indicator of old interface tunnelled */ | 
|  | 606 | hp->cmd_len = (unsigned char) cmd_size; | 
|  | 607 | hp->iovec_count = 0; | 
|  | 608 | hp->mx_sb_len = 0; | 
|  | 609 | if (input_size > 0) | 
|  | 610 | hp->dxfer_direction = (old_hdr.reply_len > SZ_SG_HEADER) ? | 
|  | 611 | SG_DXFER_TO_FROM_DEV : SG_DXFER_TO_DEV; | 
|  | 612 | else | 
|  | 613 | hp->dxfer_direction = (mxsize > 0) ? SG_DXFER_FROM_DEV : SG_DXFER_NONE; | 
|  | 614 | hp->dxfer_len = mxsize; | 
|  | 615 | if (hp->dxfer_direction == SG_DXFER_TO_DEV) | 
|  | 616 | hp->dxferp = (char __user *)buf + cmd_size; | 
|  | 617 | else | 
|  | 618 | hp->dxferp = NULL; | 
|  | 619 | hp->sbp = NULL; | 
|  | 620 | hp->timeout = old_hdr.reply_len;	/* structure abuse ... */ | 
|  | 621 | hp->flags = input_size;	/* structure abuse ... */ | 
|  | 622 | hp->pack_id = old_hdr.pack_id; | 
|  | 623 | hp->usr_ptr = NULL; | 
|  | 624 | if (__copy_from_user(cmnd, buf, cmd_size)) | 
|  | 625 | return -EFAULT; | 
|  | 626 | /* | 
|  | 627 | * SG_DXFER_TO_FROM_DEV is functionally equivalent to SG_DXFER_FROM_DEV, | 
|  | 628 | * but is is possible that the app intended SG_DXFER_TO_DEV, because there | 
|  | 629 | * is a non-zero input_size, so emit a warning. | 
|  | 630 | */ | 
|  | 631 | if (hp->dxfer_direction == SG_DXFER_TO_FROM_DEV) { | 
|  | 632 | static char cmd[TASK_COMM_LEN]; | 
|  | 633 | if (strcmp(current->comm, cmd)) { | 
|  | 634 | printk_ratelimited(KERN_WARNING | 
|  | 635 | "sg_write: data in/out %d/%d bytes " | 
|  | 636 | "for SCSI command 0x%x-- guessing " | 
|  | 637 | "data in;\n   program %s not setting " | 
|  | 638 | "count and/or reply_len properly\n", | 
|  | 639 | old_hdr.reply_len - (int)SZ_SG_HEADER, | 
|  | 640 | input_size, (unsigned int) cmnd[0], | 
|  | 641 | current->comm); | 
|  | 642 | strcpy(cmd, current->comm); | 
|  | 643 | } | 
|  | 644 | } | 
|  | 645 | k = sg_common_write(sfp, srp, cmnd, sfp->timeout, blocking); | 
|  | 646 | return (k < 0) ? k : count; | 
|  | 647 | } | 
|  | 648 |  | 
|  | 649 | static ssize_t | 
|  | 650 | sg_new_write(Sg_fd *sfp, struct file *file, const char __user *buf, | 
|  | 651 | size_t count, int blocking, int read_only, int sg_io_owned, | 
|  | 652 | Sg_request **o_srp) | 
|  | 653 | { | 
|  | 654 | int k; | 
|  | 655 | Sg_request *srp; | 
|  | 656 | sg_io_hdr_t *hp; | 
|  | 657 | unsigned char cmnd[MAX_COMMAND_SIZE]; | 
|  | 658 | int timeout; | 
|  | 659 | unsigned long ul_timeout; | 
|  | 660 |  | 
|  | 661 | if (count < SZ_SG_IO_HDR) | 
|  | 662 | return -EINVAL; | 
|  | 663 | if (!access_ok(VERIFY_READ, buf, count)) | 
|  | 664 | return -EFAULT; /* protects following copy_from_user()s + get_user()s */ | 
|  | 665 |  | 
|  | 666 | sfp->cmd_q = 1;	/* when sg_io_hdr seen, set command queuing on */ | 
|  | 667 | if (!(srp = sg_add_request(sfp))) { | 
|  | 668 | SCSI_LOG_TIMEOUT(1, printk("sg_new_write: queue full\n")); | 
|  | 669 | return -EDOM; | 
|  | 670 | } | 
|  | 671 | srp->sg_io_owned = sg_io_owned; | 
|  | 672 | hp = &srp->header; | 
|  | 673 | if (__copy_from_user(hp, buf, SZ_SG_IO_HDR)) { | 
|  | 674 | sg_remove_request(sfp, srp); | 
|  | 675 | return -EFAULT; | 
|  | 676 | } | 
|  | 677 | if (hp->interface_id != 'S') { | 
|  | 678 | sg_remove_request(sfp, srp); | 
|  | 679 | return -ENOSYS; | 
|  | 680 | } | 
|  | 681 | if (hp->flags & SG_FLAG_MMAP_IO) { | 
|  | 682 | if (hp->dxfer_len > sfp->reserve.bufflen) { | 
|  | 683 | sg_remove_request(sfp, srp); | 
|  | 684 | return -ENOMEM;	/* MMAP_IO size must fit in reserve buffer */ | 
|  | 685 | } | 
|  | 686 | if (hp->flags & SG_FLAG_DIRECT_IO) { | 
|  | 687 | sg_remove_request(sfp, srp); | 
|  | 688 | return -EINVAL;	/* either MMAP_IO or DIRECT_IO (not both) */ | 
|  | 689 | } | 
|  | 690 | if (sg_res_in_use(sfp)) { | 
|  | 691 | sg_remove_request(sfp, srp); | 
|  | 692 | return -EBUSY;	/* reserve buffer already being used */ | 
|  | 693 | } | 
|  | 694 | } | 
|  | 695 | ul_timeout = msecs_to_jiffies(srp->header.timeout); | 
|  | 696 | timeout = (ul_timeout < INT_MAX) ? ul_timeout : INT_MAX; | 
|  | 697 | if ((!hp->cmdp) || (hp->cmd_len < 6) || (hp->cmd_len > sizeof (cmnd))) { | 
|  | 698 | sg_remove_request(sfp, srp); | 
|  | 699 | return -EMSGSIZE; | 
|  | 700 | } | 
|  | 701 | if (!access_ok(VERIFY_READ, hp->cmdp, hp->cmd_len)) { | 
|  | 702 | sg_remove_request(sfp, srp); | 
|  | 703 | return -EFAULT;	/* protects following copy_from_user()s + get_user()s */ | 
|  | 704 | } | 
|  | 705 | if (__copy_from_user(cmnd, hp->cmdp, hp->cmd_len)) { | 
|  | 706 | sg_remove_request(sfp, srp); | 
|  | 707 | return -EFAULT; | 
|  | 708 | } | 
|  | 709 | if (read_only && sg_allow_access(file, cmnd)) { | 
|  | 710 | sg_remove_request(sfp, srp); | 
|  | 711 | return -EPERM; | 
|  | 712 | } | 
|  | 713 | k = sg_common_write(sfp, srp, cmnd, timeout, blocking); | 
|  | 714 | if (k < 0) | 
|  | 715 | return k; | 
|  | 716 | if (o_srp) | 
|  | 717 | *o_srp = srp; | 
|  | 718 | return count; | 
|  | 719 | } | 
|  | 720 |  | 
|  | 721 | static int | 
|  | 722 | sg_common_write(Sg_fd * sfp, Sg_request * srp, | 
|  | 723 | unsigned char *cmnd, int timeout, int blocking) | 
|  | 724 | { | 
|  | 725 | int k, data_dir; | 
|  | 726 | Sg_device *sdp = sfp->parentdp; | 
|  | 727 | sg_io_hdr_t *hp = &srp->header; | 
|  | 728 |  | 
|  | 729 | srp->data.cmd_opcode = cmnd[0];	/* hold opcode of command */ | 
|  | 730 | hp->status = 0; | 
|  | 731 | hp->masked_status = 0; | 
|  | 732 | hp->msg_status = 0; | 
|  | 733 | hp->info = 0; | 
|  | 734 | hp->host_status = 0; | 
|  | 735 | hp->driver_status = 0; | 
|  | 736 | hp->resid = 0; | 
|  | 737 | SCSI_LOG_TIMEOUT(4, printk("sg_common_write:  scsi opcode=0x%02x, cmd_size=%d\n", | 
|  | 738 | (int) cmnd[0], (int) hp->cmd_len)); | 
|  | 739 |  | 
|  | 740 | k = sg_start_req(srp, cmnd); | 
|  | 741 | if (k) { | 
|  | 742 | SCSI_LOG_TIMEOUT(1, printk("sg_common_write: start_req err=%d\n", k)); | 
|  | 743 | sg_finish_rem_req(srp); | 
|  | 744 | return k;	/* probably out of space --> ENOMEM */ | 
|  | 745 | } | 
|  | 746 | if (sdp->detached) { | 
|  | 747 | if (srp->bio) | 
|  | 748 | blk_end_request_all(srp->rq, -EIO); | 
|  | 749 | sg_finish_rem_req(srp); | 
|  | 750 | return -ENODEV; | 
|  | 751 | } | 
|  | 752 |  | 
|  | 753 | switch (hp->dxfer_direction) { | 
|  | 754 | case SG_DXFER_TO_FROM_DEV: | 
|  | 755 | case SG_DXFER_FROM_DEV: | 
|  | 756 | data_dir = DMA_FROM_DEVICE; | 
|  | 757 | break; | 
|  | 758 | case SG_DXFER_TO_DEV: | 
|  | 759 | data_dir = DMA_TO_DEVICE; | 
|  | 760 | break; | 
|  | 761 | case SG_DXFER_UNKNOWN: | 
|  | 762 | data_dir = DMA_BIDIRECTIONAL; | 
|  | 763 | break; | 
|  | 764 | default: | 
|  | 765 | data_dir = DMA_NONE; | 
|  | 766 | break; | 
|  | 767 | } | 
|  | 768 | hp->duration = jiffies_to_msecs(jiffies); | 
|  | 769 |  | 
|  | 770 | srp->rq->timeout = timeout; | 
|  | 771 | kref_get(&sfp->f_ref); /* sg_rq_end_io() does kref_put(). */ | 
|  | 772 | blk_execute_rq_nowait(sdp->device->request_queue, sdp->disk, | 
|  | 773 | srp->rq, 1, sg_rq_end_io); | 
|  | 774 | return 0; | 
|  | 775 | } | 
|  | 776 |  | 
|  | 777 | static int | 
|  | 778 | sg_ioctl(struct file *filp, unsigned int cmd_in, unsigned long arg) | 
|  | 779 | { | 
|  | 780 | void __user *p = (void __user *)arg; | 
|  | 781 | int __user *ip = p; | 
|  | 782 | int result, val, read_only; | 
|  | 783 | Sg_device *sdp; | 
|  | 784 | Sg_fd *sfp; | 
|  | 785 | Sg_request *srp; | 
|  | 786 | unsigned long iflags; | 
|  | 787 |  | 
|  | 788 | if ((!(sfp = (Sg_fd *) filp->private_data)) || (!(sdp = sfp->parentdp))) | 
|  | 789 | return -ENXIO; | 
|  | 790 |  | 
|  | 791 | SCSI_LOG_TIMEOUT(3, printk("sg_ioctl: %s, cmd=0x%x\n", | 
|  | 792 | sdp->disk->disk_name, (int) cmd_in)); | 
|  | 793 | read_only = (O_RDWR != (filp->f_flags & O_ACCMODE)); | 
|  | 794 |  | 
|  | 795 | switch (cmd_in) { | 
|  | 796 | case SG_IO: | 
|  | 797 | { | 
|  | 798 | int blocking = 1;	/* ignore O_NONBLOCK flag */ | 
|  | 799 |  | 
|  | 800 | if (sdp->detached) | 
|  | 801 | return -ENODEV; | 
|  | 802 | if (!scsi_block_when_processing_errors(sdp->device)) | 
|  | 803 | return -ENXIO; | 
|  | 804 | if (!access_ok(VERIFY_WRITE, p, SZ_SG_IO_HDR)) | 
|  | 805 | return -EFAULT; | 
|  | 806 | result = | 
|  | 807 | sg_new_write(sfp, filp, p, SZ_SG_IO_HDR, | 
|  | 808 | blocking, read_only, 1, &srp); | 
|  | 809 | if (result < 0) | 
|  | 810 | return result; | 
|  | 811 | while (1) { | 
|  | 812 | result = 0;	/* following macro to beat race condition */ | 
|  | 813 | __wait_event_interruptible(sfp->read_wait, | 
|  | 814 | (srp->done || sdp->detached), | 
|  | 815 | result); | 
|  | 816 | if (sdp->detached) | 
|  | 817 | return -ENODEV; | 
|  | 818 | write_lock_irq(&sfp->rq_list_lock); | 
|  | 819 | if (srp->done) { | 
|  | 820 | srp->done = 2; | 
|  | 821 | write_unlock_irq(&sfp->rq_list_lock); | 
|  | 822 | break; | 
|  | 823 | } | 
|  | 824 | srp->orphan = 1; | 
|  | 825 | write_unlock_irq(&sfp->rq_list_lock); | 
|  | 826 | return result;	/* -ERESTARTSYS because signal hit process */ | 
|  | 827 | } | 
|  | 828 | result = sg_new_read(sfp, p, SZ_SG_IO_HDR, srp); | 
|  | 829 | return (result < 0) ? result : 0; | 
|  | 830 | } | 
|  | 831 | case SG_SET_TIMEOUT: | 
|  | 832 | result = get_user(val, ip); | 
|  | 833 | if (result) | 
|  | 834 | return result; | 
|  | 835 | if (val < 0) | 
|  | 836 | return -EIO; | 
|  | 837 | if (val >= MULDIV (INT_MAX, USER_HZ, HZ)) | 
|  | 838 | val = MULDIV (INT_MAX, USER_HZ, HZ); | 
|  | 839 | sfp->timeout_user = val; | 
|  | 840 | sfp->timeout = MULDIV (val, HZ, USER_HZ); | 
|  | 841 |  | 
|  | 842 | return 0; | 
|  | 843 | case SG_GET_TIMEOUT:	/* N.B. User receives timeout as return value */ | 
|  | 844 | /* strange ..., for backward compatibility */ | 
|  | 845 | return sfp->timeout_user; | 
|  | 846 | case SG_SET_FORCE_LOW_DMA: | 
|  | 847 | result = get_user(val, ip); | 
|  | 848 | if (result) | 
|  | 849 | return result; | 
|  | 850 | if (val) { | 
|  | 851 | sfp->low_dma = 1; | 
|  | 852 | if ((0 == sfp->low_dma) && (0 == sg_res_in_use(sfp))) { | 
|  | 853 | val = (int) sfp->reserve.bufflen; | 
|  | 854 | sg_remove_scat(&sfp->reserve); | 
|  | 855 | sg_build_reserve(sfp, val); | 
|  | 856 | } | 
|  | 857 | } else { | 
|  | 858 | if (sdp->detached) | 
|  | 859 | return -ENODEV; | 
|  | 860 | sfp->low_dma = sdp->device->host->unchecked_isa_dma; | 
|  | 861 | } | 
|  | 862 | return 0; | 
|  | 863 | case SG_GET_LOW_DMA: | 
|  | 864 | return put_user((int) sfp->low_dma, ip); | 
|  | 865 | case SG_GET_SCSI_ID: | 
|  | 866 | if (!access_ok(VERIFY_WRITE, p, sizeof (sg_scsi_id_t))) | 
|  | 867 | return -EFAULT; | 
|  | 868 | else { | 
|  | 869 | sg_scsi_id_t __user *sg_idp = p; | 
|  | 870 |  | 
|  | 871 | if (sdp->detached) | 
|  | 872 | return -ENODEV; | 
|  | 873 | __put_user((int) sdp->device->host->host_no, | 
|  | 874 | &sg_idp->host_no); | 
|  | 875 | __put_user((int) sdp->device->channel, | 
|  | 876 | &sg_idp->channel); | 
|  | 877 | __put_user((int) sdp->device->id, &sg_idp->scsi_id); | 
|  | 878 | __put_user((int) sdp->device->lun, &sg_idp->lun); | 
|  | 879 | __put_user((int) sdp->device->type, &sg_idp->scsi_type); | 
|  | 880 | __put_user((short) sdp->device->host->cmd_per_lun, | 
|  | 881 | &sg_idp->h_cmd_per_lun); | 
|  | 882 | __put_user((short) sdp->device->queue_depth, | 
|  | 883 | &sg_idp->d_queue_depth); | 
|  | 884 | __put_user(0, &sg_idp->unused[0]); | 
|  | 885 | __put_user(0, &sg_idp->unused[1]); | 
|  | 886 | return 0; | 
|  | 887 | } | 
|  | 888 | case SG_SET_FORCE_PACK_ID: | 
|  | 889 | result = get_user(val, ip); | 
|  | 890 | if (result) | 
|  | 891 | return result; | 
|  | 892 | sfp->force_packid = val ? 1 : 0; | 
|  | 893 | return 0; | 
|  | 894 | case SG_GET_PACK_ID: | 
|  | 895 | if (!access_ok(VERIFY_WRITE, ip, sizeof (int))) | 
|  | 896 | return -EFAULT; | 
|  | 897 | read_lock_irqsave(&sfp->rq_list_lock, iflags); | 
|  | 898 | for (srp = sfp->headrp; srp; srp = srp->nextrp) { | 
|  | 899 | if ((1 == srp->done) && (!srp->sg_io_owned)) { | 
|  | 900 | read_unlock_irqrestore(&sfp->rq_list_lock, | 
|  | 901 | iflags); | 
|  | 902 | __put_user(srp->header.pack_id, ip); | 
|  | 903 | return 0; | 
|  | 904 | } | 
|  | 905 | } | 
|  | 906 | read_unlock_irqrestore(&sfp->rq_list_lock, iflags); | 
|  | 907 | __put_user(-1, ip); | 
|  | 908 | return 0; | 
|  | 909 | case SG_GET_NUM_WAITING: | 
|  | 910 | read_lock_irqsave(&sfp->rq_list_lock, iflags); | 
|  | 911 | for (val = 0, srp = sfp->headrp; srp; srp = srp->nextrp) { | 
|  | 912 | if ((1 == srp->done) && (!srp->sg_io_owned)) | 
|  | 913 | ++val; | 
|  | 914 | } | 
|  | 915 | read_unlock_irqrestore(&sfp->rq_list_lock, iflags); | 
|  | 916 | return put_user(val, ip); | 
|  | 917 | case SG_GET_SG_TABLESIZE: | 
|  | 918 | return put_user(sdp->sg_tablesize, ip); | 
|  | 919 | case SG_SET_RESERVED_SIZE: | 
|  | 920 | result = get_user(val, ip); | 
|  | 921 | if (result) | 
|  | 922 | return result; | 
|  | 923 | if (val < 0) | 
|  | 924 | return -EINVAL; | 
|  | 925 | val = min_t(int, val, | 
|  | 926 | queue_max_sectors(sdp->device->request_queue) * 512); | 
|  | 927 | if (val != sfp->reserve.bufflen) { | 
|  | 928 | if (sg_res_in_use(sfp) || sfp->mmap_called) | 
|  | 929 | return -EBUSY; | 
|  | 930 | sg_remove_scat(&sfp->reserve); | 
|  | 931 | sg_build_reserve(sfp, val); | 
|  | 932 | } | 
|  | 933 | return 0; | 
|  | 934 | case SG_GET_RESERVED_SIZE: | 
|  | 935 | val = min_t(int, sfp->reserve.bufflen, | 
|  | 936 | queue_max_sectors(sdp->device->request_queue) * 512); | 
|  | 937 | return put_user(val, ip); | 
|  | 938 | case SG_SET_COMMAND_Q: | 
|  | 939 | result = get_user(val, ip); | 
|  | 940 | if (result) | 
|  | 941 | return result; | 
|  | 942 | sfp->cmd_q = val ? 1 : 0; | 
|  | 943 | return 0; | 
|  | 944 | case SG_GET_COMMAND_Q: | 
|  | 945 | return put_user((int) sfp->cmd_q, ip); | 
|  | 946 | case SG_SET_KEEP_ORPHAN: | 
|  | 947 | result = get_user(val, ip); | 
|  | 948 | if (result) | 
|  | 949 | return result; | 
|  | 950 | sfp->keep_orphan = val; | 
|  | 951 | return 0; | 
|  | 952 | case SG_GET_KEEP_ORPHAN: | 
|  | 953 | return put_user((int) sfp->keep_orphan, ip); | 
|  | 954 | case SG_NEXT_CMD_LEN: | 
|  | 955 | result = get_user(val, ip); | 
|  | 956 | if (result) | 
|  | 957 | return result; | 
|  | 958 | sfp->next_cmd_len = (val > 0) ? val : 0; | 
|  | 959 | return 0; | 
|  | 960 | case SG_GET_VERSION_NUM: | 
|  | 961 | return put_user(sg_version_num, ip); | 
|  | 962 | case SG_GET_ACCESS_COUNT: | 
|  | 963 | /* faked - we don't have a real access count anymore */ | 
|  | 964 | val = (sdp->device ? 1 : 0); | 
|  | 965 | return put_user(val, ip); | 
|  | 966 | case SG_GET_REQUEST_TABLE: | 
|  | 967 | if (!access_ok(VERIFY_WRITE, p, SZ_SG_REQ_INFO * SG_MAX_QUEUE)) | 
|  | 968 | return -EFAULT; | 
|  | 969 | else { | 
|  | 970 | sg_req_info_t *rinfo; | 
|  | 971 | unsigned int ms; | 
|  | 972 |  | 
|  | 973 | rinfo = kmalloc(SZ_SG_REQ_INFO * SG_MAX_QUEUE, | 
|  | 974 | GFP_KERNEL); | 
|  | 975 | if (!rinfo) | 
|  | 976 | return -ENOMEM; | 
|  | 977 | read_lock_irqsave(&sfp->rq_list_lock, iflags); | 
|  | 978 | for (srp = sfp->headrp, val = 0; val < SG_MAX_QUEUE; | 
|  | 979 | ++val, srp = srp ? srp->nextrp : srp) { | 
|  | 980 | memset(&rinfo[val], 0, SZ_SG_REQ_INFO); | 
|  | 981 | if (srp) { | 
|  | 982 | rinfo[val].req_state = srp->done + 1; | 
|  | 983 | rinfo[val].problem = | 
|  | 984 | srp->header.masked_status & | 
|  | 985 | srp->header.host_status & | 
|  | 986 | srp->header.driver_status; | 
|  | 987 | if (srp->done) | 
|  | 988 | rinfo[val].duration = | 
|  | 989 | srp->header.duration; | 
|  | 990 | else { | 
|  | 991 | ms = jiffies_to_msecs(jiffies); | 
|  | 992 | rinfo[val].duration = | 
|  | 993 | (ms > srp->header.duration) ? | 
|  | 994 | (ms - srp->header.duration) : 0; | 
|  | 995 | } | 
|  | 996 | rinfo[val].orphan = srp->orphan; | 
|  | 997 | rinfo[val].sg_io_owned = | 
|  | 998 | srp->sg_io_owned; | 
|  | 999 | rinfo[val].pack_id = | 
|  | 1000 | srp->header.pack_id; | 
|  | 1001 | rinfo[val].usr_ptr = | 
|  | 1002 | srp->header.usr_ptr; | 
|  | 1003 | } | 
|  | 1004 | } | 
|  | 1005 | read_unlock_irqrestore(&sfp->rq_list_lock, iflags); | 
|  | 1006 | result = __copy_to_user(p, rinfo, | 
|  | 1007 | SZ_SG_REQ_INFO * SG_MAX_QUEUE); | 
|  | 1008 | result = result ? -EFAULT : 0; | 
|  | 1009 | kfree(rinfo); | 
|  | 1010 | return result; | 
|  | 1011 | } | 
|  | 1012 | case SG_EMULATED_HOST: | 
|  | 1013 | if (sdp->detached) | 
|  | 1014 | return -ENODEV; | 
|  | 1015 | return put_user(sdp->device->host->hostt->emulated, ip); | 
|  | 1016 | case SG_SCSI_RESET: | 
|  | 1017 | if (sdp->detached) | 
|  | 1018 | return -ENODEV; | 
|  | 1019 | if (filp->f_flags & O_NONBLOCK) { | 
|  | 1020 | if (scsi_host_in_recovery(sdp->device->host)) | 
|  | 1021 | return -EBUSY; | 
|  | 1022 | } else if (!scsi_block_when_processing_errors(sdp->device)) | 
|  | 1023 | return -EBUSY; | 
|  | 1024 | result = get_user(val, ip); | 
|  | 1025 | if (result) | 
|  | 1026 | return result; | 
|  | 1027 | if (SG_SCSI_RESET_NOTHING == val) | 
|  | 1028 | return 0; | 
|  | 1029 | switch (val) { | 
|  | 1030 | case SG_SCSI_RESET_DEVICE: | 
|  | 1031 | val = SCSI_TRY_RESET_DEVICE; | 
|  | 1032 | break; | 
|  | 1033 | case SG_SCSI_RESET_TARGET: | 
|  | 1034 | val = SCSI_TRY_RESET_TARGET; | 
|  | 1035 | break; | 
|  | 1036 | case SG_SCSI_RESET_BUS: | 
|  | 1037 | val = SCSI_TRY_RESET_BUS; | 
|  | 1038 | break; | 
|  | 1039 | case SG_SCSI_RESET_HOST: | 
|  | 1040 | val = SCSI_TRY_RESET_HOST; | 
|  | 1041 | break; | 
|  | 1042 | default: | 
|  | 1043 | return -EINVAL; | 
|  | 1044 | } | 
|  | 1045 | if (!capable(CAP_SYS_ADMIN) || !capable(CAP_SYS_RAWIO)) | 
|  | 1046 | return -EACCES; | 
|  | 1047 | return (scsi_reset_provider(sdp->device, val) == | 
|  | 1048 | SUCCESS) ? 0 : -EIO; | 
|  | 1049 | case SCSI_IOCTL_SEND_COMMAND: | 
|  | 1050 | if (sdp->detached) | 
|  | 1051 | return -ENODEV; | 
|  | 1052 | if (read_only) { | 
|  | 1053 | unsigned char opcode = WRITE_6; | 
|  | 1054 | Scsi_Ioctl_Command __user *siocp = p; | 
|  | 1055 |  | 
|  | 1056 | if (copy_from_user(&opcode, siocp->data, 1)) | 
|  | 1057 | return -EFAULT; | 
|  | 1058 | if (sg_allow_access(filp, &opcode)) | 
|  | 1059 | return -EPERM; | 
|  | 1060 | } | 
|  | 1061 | return sg_scsi_ioctl(sdp->device->request_queue, NULL, filp->f_mode, p); | 
|  | 1062 | case SG_SET_DEBUG: | 
|  | 1063 | result = get_user(val, ip); | 
|  | 1064 | if (result) | 
|  | 1065 | return result; | 
|  | 1066 | sdp->sgdebug = (char) val; | 
|  | 1067 | return 0; | 
|  | 1068 | case SCSI_IOCTL_GET_IDLUN: | 
|  | 1069 | case SCSI_IOCTL_GET_BUS_NUMBER: | 
|  | 1070 | case SCSI_IOCTL_PROBE_HOST: | 
|  | 1071 | case SG_GET_TRANSFORM: | 
|  | 1072 | if (sdp->detached) | 
|  | 1073 | return -ENODEV; | 
|  | 1074 | return scsi_ioctl(sdp->device, cmd_in, p); | 
|  | 1075 | case BLKSECTGET: | 
|  | 1076 | return put_user(queue_max_sectors(sdp->device->request_queue) * 512, | 
|  | 1077 | ip); | 
|  | 1078 | case BLKTRACESETUP: | 
|  | 1079 | return blk_trace_setup(sdp->device->request_queue, | 
|  | 1080 | sdp->disk->disk_name, | 
|  | 1081 | MKDEV(SCSI_GENERIC_MAJOR, sdp->index), | 
|  | 1082 | NULL, | 
|  | 1083 | (char *)arg); | 
|  | 1084 | case BLKTRACESTART: | 
|  | 1085 | return blk_trace_startstop(sdp->device->request_queue, 1); | 
|  | 1086 | case BLKTRACESTOP: | 
|  | 1087 | return blk_trace_startstop(sdp->device->request_queue, 0); | 
|  | 1088 | case BLKTRACETEARDOWN: | 
|  | 1089 | return blk_trace_remove(sdp->device->request_queue); | 
|  | 1090 | default: | 
|  | 1091 | if (read_only) | 
|  | 1092 | return -EPERM;	/* don't know so take safe approach */ | 
|  | 1093 | return scsi_ioctl(sdp->device, cmd_in, p); | 
|  | 1094 | } | 
|  | 1095 | } | 
|  | 1096 |  | 
|  | 1097 | static long | 
|  | 1098 | sg_unlocked_ioctl(struct file *filp, unsigned int cmd_in, unsigned long arg) | 
|  | 1099 | { | 
|  | 1100 | int ret; | 
|  | 1101 |  | 
|  | 1102 | mutex_lock(&sg_mutex); | 
|  | 1103 | ret = sg_ioctl(filp, cmd_in, arg); | 
|  | 1104 | mutex_unlock(&sg_mutex); | 
|  | 1105 |  | 
|  | 1106 | return ret; | 
|  | 1107 | } | 
|  | 1108 |  | 
|  | 1109 | #ifdef CONFIG_COMPAT | 
|  | 1110 | static long sg_compat_ioctl(struct file *filp, unsigned int cmd_in, unsigned long arg) | 
|  | 1111 | { | 
|  | 1112 | Sg_device *sdp; | 
|  | 1113 | Sg_fd *sfp; | 
|  | 1114 | struct scsi_device *sdev; | 
|  | 1115 |  | 
|  | 1116 | if ((!(sfp = (Sg_fd *) filp->private_data)) || (!(sdp = sfp->parentdp))) | 
|  | 1117 | return -ENXIO; | 
|  | 1118 |  | 
|  | 1119 | sdev = sdp->device; | 
|  | 1120 | if (sdev->host->hostt->compat_ioctl) { | 
|  | 1121 | int ret; | 
|  | 1122 |  | 
|  | 1123 | ret = sdev->host->hostt->compat_ioctl(sdev, cmd_in, (void __user *)arg); | 
|  | 1124 |  | 
|  | 1125 | return ret; | 
|  | 1126 | } | 
|  | 1127 |  | 
|  | 1128 | return -ENOIOCTLCMD; | 
|  | 1129 | } | 
|  | 1130 | #endif | 
|  | 1131 |  | 
|  | 1132 | static unsigned int | 
|  | 1133 | sg_poll(struct file *filp, poll_table * wait) | 
|  | 1134 | { | 
|  | 1135 | unsigned int res = 0; | 
|  | 1136 | Sg_device *sdp; | 
|  | 1137 | Sg_fd *sfp; | 
|  | 1138 | Sg_request *srp; | 
|  | 1139 | int count = 0; | 
|  | 1140 | unsigned long iflags; | 
|  | 1141 |  | 
|  | 1142 | if ((!(sfp = (Sg_fd *) filp->private_data)) || (!(sdp = sfp->parentdp)) | 
|  | 1143 | || sfp->closed) | 
|  | 1144 | return POLLERR; | 
|  | 1145 | poll_wait(filp, &sfp->read_wait, wait); | 
|  | 1146 | read_lock_irqsave(&sfp->rq_list_lock, iflags); | 
|  | 1147 | for (srp = sfp->headrp; srp; srp = srp->nextrp) { | 
|  | 1148 | /* if any read waiting, flag it */ | 
|  | 1149 | if ((0 == res) && (1 == srp->done) && (!srp->sg_io_owned)) | 
|  | 1150 | res = POLLIN | POLLRDNORM; | 
|  | 1151 | ++count; | 
|  | 1152 | } | 
|  | 1153 | read_unlock_irqrestore(&sfp->rq_list_lock, iflags); | 
|  | 1154 |  | 
|  | 1155 | if (sdp->detached) | 
|  | 1156 | res |= POLLHUP; | 
|  | 1157 | else if (!sfp->cmd_q) { | 
|  | 1158 | if (0 == count) | 
|  | 1159 | res |= POLLOUT | POLLWRNORM; | 
|  | 1160 | } else if (count < SG_MAX_QUEUE) | 
|  | 1161 | res |= POLLOUT | POLLWRNORM; | 
|  | 1162 | SCSI_LOG_TIMEOUT(3, printk("sg_poll: %s, res=0x%x\n", | 
|  | 1163 | sdp->disk->disk_name, (int) res)); | 
|  | 1164 | return res; | 
|  | 1165 | } | 
|  | 1166 |  | 
|  | 1167 | static int | 
|  | 1168 | sg_fasync(int fd, struct file *filp, int mode) | 
|  | 1169 | { | 
|  | 1170 | Sg_device *sdp; | 
|  | 1171 | Sg_fd *sfp; | 
|  | 1172 |  | 
|  | 1173 | if ((!(sfp = (Sg_fd *) filp->private_data)) || (!(sdp = sfp->parentdp))) | 
|  | 1174 | return -ENXIO; | 
|  | 1175 | SCSI_LOG_TIMEOUT(3, printk("sg_fasync: %s, mode=%d\n", | 
|  | 1176 | sdp->disk->disk_name, mode)); | 
|  | 1177 |  | 
|  | 1178 | return fasync_helper(fd, filp, mode, &sfp->async_qp); | 
|  | 1179 | } | 
|  | 1180 |  | 
|  | 1181 | static int | 
|  | 1182 | sg_vma_fault(struct vm_area_struct *vma, struct vm_fault *vmf) | 
|  | 1183 | { | 
|  | 1184 | Sg_fd *sfp; | 
|  | 1185 | unsigned long offset, len, sa; | 
|  | 1186 | Sg_scatter_hold *rsv_schp; | 
|  | 1187 | int k, length; | 
|  | 1188 |  | 
|  | 1189 | if ((NULL == vma) || (!(sfp = (Sg_fd *) vma->vm_private_data))) | 
|  | 1190 | return VM_FAULT_SIGBUS; | 
|  | 1191 | rsv_schp = &sfp->reserve; | 
|  | 1192 | offset = vmf->pgoff << PAGE_SHIFT; | 
|  | 1193 | if (offset >= rsv_schp->bufflen) | 
|  | 1194 | return VM_FAULT_SIGBUS; | 
|  | 1195 | SCSI_LOG_TIMEOUT(3, printk("sg_vma_fault: offset=%lu, scatg=%d\n", | 
|  | 1196 | offset, rsv_schp->k_use_sg)); | 
|  | 1197 | sa = vma->vm_start; | 
|  | 1198 | length = 1 << (PAGE_SHIFT + rsv_schp->page_order); | 
|  | 1199 | for (k = 0; k < rsv_schp->k_use_sg && sa < vma->vm_end; k++) { | 
|  | 1200 | len = vma->vm_end - sa; | 
|  | 1201 | len = (len < length) ? len : length; | 
|  | 1202 | if (offset < len) { | 
|  | 1203 | struct page *page = nth_page(rsv_schp->pages[k], | 
|  | 1204 | offset >> PAGE_SHIFT); | 
|  | 1205 | get_page(page);	/* increment page count */ | 
|  | 1206 | vmf->page = page; | 
|  | 1207 | return 0; /* success */ | 
|  | 1208 | } | 
|  | 1209 | sa += len; | 
|  | 1210 | offset -= len; | 
|  | 1211 | } | 
|  | 1212 |  | 
|  | 1213 | return VM_FAULT_SIGBUS; | 
|  | 1214 | } | 
|  | 1215 |  | 
|  | 1216 | static const struct vm_operations_struct sg_mmap_vm_ops = { | 
|  | 1217 | .fault = sg_vma_fault, | 
|  | 1218 | }; | 
|  | 1219 |  | 
|  | 1220 | static int | 
|  | 1221 | sg_mmap(struct file *filp, struct vm_area_struct *vma) | 
|  | 1222 | { | 
|  | 1223 | Sg_fd *sfp; | 
|  | 1224 | unsigned long req_sz, len, sa; | 
|  | 1225 | Sg_scatter_hold *rsv_schp; | 
|  | 1226 | int k, length; | 
|  | 1227 |  | 
|  | 1228 | if ((!filp) || (!vma) || (!(sfp = (Sg_fd *) filp->private_data))) | 
|  | 1229 | return -ENXIO; | 
|  | 1230 | req_sz = vma->vm_end - vma->vm_start; | 
|  | 1231 | SCSI_LOG_TIMEOUT(3, printk("sg_mmap starting, vm_start=%p, len=%d\n", | 
|  | 1232 | (void *) vma->vm_start, (int) req_sz)); | 
|  | 1233 | if (vma->vm_pgoff) | 
|  | 1234 | return -EINVAL;	/* want no offset */ | 
|  | 1235 | rsv_schp = &sfp->reserve; | 
|  | 1236 | if (req_sz > rsv_schp->bufflen) | 
|  | 1237 | return -ENOMEM;	/* cannot map more than reserved buffer */ | 
|  | 1238 |  | 
|  | 1239 | sa = vma->vm_start; | 
|  | 1240 | length = 1 << (PAGE_SHIFT + rsv_schp->page_order); | 
|  | 1241 | for (k = 0; k < rsv_schp->k_use_sg && sa < vma->vm_end; k++) { | 
|  | 1242 | len = vma->vm_end - sa; | 
|  | 1243 | len = (len < length) ? len : length; | 
|  | 1244 | sa += len; | 
|  | 1245 | } | 
|  | 1246 |  | 
|  | 1247 | sfp->mmap_called = 1; | 
|  | 1248 | vma->vm_flags |= VM_RESERVED; | 
|  | 1249 | vma->vm_private_data = sfp; | 
|  | 1250 | vma->vm_ops = &sg_mmap_vm_ops; | 
|  | 1251 | return 0; | 
|  | 1252 | } | 
|  | 1253 |  | 
|  | 1254 | static void sg_rq_end_io_usercontext(struct work_struct *work) | 
|  | 1255 | { | 
|  | 1256 | struct sg_request *srp = container_of(work, struct sg_request, ew.work); | 
|  | 1257 | struct sg_fd *sfp = srp->parentfp; | 
|  | 1258 |  | 
|  | 1259 | sg_finish_rem_req(srp); | 
|  | 1260 | kref_put(&sfp->f_ref, sg_remove_sfp); | 
|  | 1261 | } | 
|  | 1262 |  | 
|  | 1263 | /* | 
|  | 1264 | * This function is a "bottom half" handler that is called by the mid | 
|  | 1265 | * level when a command is completed (or has failed). | 
|  | 1266 | */ | 
|  | 1267 | static void sg_rq_end_io(struct request *rq, int uptodate) | 
|  | 1268 | { | 
|  | 1269 | struct sg_request *srp = rq->end_io_data; | 
|  | 1270 | Sg_device *sdp; | 
|  | 1271 | Sg_fd *sfp; | 
|  | 1272 | unsigned long iflags; | 
|  | 1273 | unsigned int ms; | 
|  | 1274 | char *sense; | 
|  | 1275 | int result, resid, done = 1; | 
|  | 1276 |  | 
|  | 1277 | if (WARN_ON(srp->done != 0)) | 
|  | 1278 | return; | 
|  | 1279 |  | 
|  | 1280 | sfp = srp->parentfp; | 
|  | 1281 | if (WARN_ON(sfp == NULL)) | 
|  | 1282 | return; | 
|  | 1283 |  | 
|  | 1284 | sdp = sfp->parentdp; | 
|  | 1285 | if (unlikely(sdp->detached)) | 
|  | 1286 | printk(KERN_INFO "sg_rq_end_io: device detached\n"); | 
|  | 1287 |  | 
|  | 1288 | sense = rq->sense; | 
|  | 1289 | result = rq->errors; | 
|  | 1290 | resid = rq->resid_len; | 
|  | 1291 |  | 
|  | 1292 | SCSI_LOG_TIMEOUT(4, printk("sg_cmd_done: %s, pack_id=%d, res=0x%x\n", | 
|  | 1293 | sdp->disk->disk_name, srp->header.pack_id, result)); | 
|  | 1294 | srp->header.resid = resid; | 
|  | 1295 | ms = jiffies_to_msecs(jiffies); | 
|  | 1296 | srp->header.duration = (ms > srp->header.duration) ? | 
|  | 1297 | (ms - srp->header.duration) : 0; | 
|  | 1298 | if (0 != result) { | 
|  | 1299 | struct scsi_sense_hdr sshdr; | 
|  | 1300 |  | 
|  | 1301 | srp->header.status = 0xff & result; | 
|  | 1302 | srp->header.masked_status = status_byte(result); | 
|  | 1303 | srp->header.msg_status = msg_byte(result); | 
|  | 1304 | srp->header.host_status = host_byte(result); | 
|  | 1305 | srp->header.driver_status = driver_byte(result); | 
|  | 1306 | if ((sdp->sgdebug > 0) && | 
|  | 1307 | ((CHECK_CONDITION == srp->header.masked_status) || | 
|  | 1308 | (COMMAND_TERMINATED == srp->header.masked_status))) | 
|  | 1309 | __scsi_print_sense("sg_cmd_done", sense, | 
|  | 1310 | SCSI_SENSE_BUFFERSIZE); | 
|  | 1311 |  | 
|  | 1312 | /* Following if statement is a patch supplied by Eric Youngdale */ | 
|  | 1313 | if (driver_byte(result) != 0 | 
|  | 1314 | && scsi_normalize_sense(sense, SCSI_SENSE_BUFFERSIZE, &sshdr) | 
|  | 1315 | && !scsi_sense_is_deferred(&sshdr) | 
|  | 1316 | && sshdr.sense_key == UNIT_ATTENTION | 
|  | 1317 | && sdp->device->removable) { | 
|  | 1318 | /* Detected possible disc change. Set the bit - this */ | 
|  | 1319 | /* may be used if there are filesystems using this device */ | 
|  | 1320 | sdp->device->changed = 1; | 
|  | 1321 | } | 
|  | 1322 | } | 
|  | 1323 | /* Rely on write phase to clean out srp status values, so no "else" */ | 
|  | 1324 |  | 
|  | 1325 | write_lock_irqsave(&sfp->rq_list_lock, iflags); | 
|  | 1326 | if (unlikely(srp->orphan)) { | 
|  | 1327 | if (sfp->keep_orphan) | 
|  | 1328 | srp->sg_io_owned = 0; | 
|  | 1329 | else | 
|  | 1330 | done = 0; | 
|  | 1331 | } | 
|  | 1332 | srp->done = done; | 
|  | 1333 | write_unlock_irqrestore(&sfp->rq_list_lock, iflags); | 
|  | 1334 |  | 
|  | 1335 | if (likely(done)) { | 
|  | 1336 | /* Now wake up any sg_read() that is waiting for this | 
|  | 1337 | * packet. | 
|  | 1338 | */ | 
|  | 1339 | wake_up_interruptible(&sfp->read_wait); | 
|  | 1340 | kill_fasync(&sfp->async_qp, SIGPOLL, POLL_IN); | 
|  | 1341 | kref_put(&sfp->f_ref, sg_remove_sfp); | 
|  | 1342 | } else { | 
|  | 1343 | INIT_WORK(&srp->ew.work, sg_rq_end_io_usercontext); | 
|  | 1344 | schedule_work(&srp->ew.work); | 
|  | 1345 | } | 
|  | 1346 | } | 
|  | 1347 |  | 
|  | 1348 | static const struct file_operations sg_fops = { | 
|  | 1349 | .owner = THIS_MODULE, | 
|  | 1350 | .read = sg_read, | 
|  | 1351 | .write = sg_write, | 
|  | 1352 | .poll = sg_poll, | 
|  | 1353 | .unlocked_ioctl = sg_unlocked_ioctl, | 
|  | 1354 | #ifdef CONFIG_COMPAT | 
|  | 1355 | .compat_ioctl = sg_compat_ioctl, | 
|  | 1356 | #endif | 
|  | 1357 | .open = sg_open, | 
|  | 1358 | .mmap = sg_mmap, | 
|  | 1359 | .release = sg_release, | 
|  | 1360 | .fasync = sg_fasync, | 
|  | 1361 | .llseek = no_llseek, | 
|  | 1362 | }; | 
|  | 1363 |  | 
|  | 1364 | static struct class *sg_sysfs_class; | 
|  | 1365 |  | 
|  | 1366 | static int sg_sysfs_valid = 0; | 
|  | 1367 |  | 
|  | 1368 | static Sg_device *sg_alloc(struct gendisk *disk, struct scsi_device *scsidp) | 
|  | 1369 | { | 
|  | 1370 | struct request_queue *q = scsidp->request_queue; | 
|  | 1371 | Sg_device *sdp; | 
|  | 1372 | unsigned long iflags; | 
|  | 1373 | int error; | 
|  | 1374 | u32 k; | 
|  | 1375 |  | 
|  | 1376 | sdp = kzalloc(sizeof(Sg_device), GFP_KERNEL); | 
|  | 1377 | if (!sdp) { | 
|  | 1378 | printk(KERN_WARNING "kmalloc Sg_device failure\n"); | 
|  | 1379 | return ERR_PTR(-ENOMEM); | 
|  | 1380 | } | 
|  | 1381 |  | 
|  | 1382 | if (!idr_pre_get(&sg_index_idr, GFP_KERNEL)) { | 
|  | 1383 | printk(KERN_WARNING "idr expansion Sg_device failure\n"); | 
|  | 1384 | error = -ENOMEM; | 
|  | 1385 | goto out; | 
|  | 1386 | } | 
|  | 1387 |  | 
|  | 1388 | write_lock_irqsave(&sg_index_lock, iflags); | 
|  | 1389 |  | 
|  | 1390 | error = idr_get_new(&sg_index_idr, sdp, &k); | 
|  | 1391 | if (error) { | 
|  | 1392 | write_unlock_irqrestore(&sg_index_lock, iflags); | 
|  | 1393 | printk(KERN_WARNING "idr allocation Sg_device failure: %d\n", | 
|  | 1394 | error); | 
|  | 1395 | goto out; | 
|  | 1396 | } | 
|  | 1397 |  | 
|  | 1398 | if (unlikely(k >= SG_MAX_DEVS)) | 
|  | 1399 | goto overflow; | 
|  | 1400 |  | 
|  | 1401 | SCSI_LOG_TIMEOUT(3, printk("sg_alloc: dev=%d \n", k)); | 
|  | 1402 | sprintf(disk->disk_name, "sg%d", k); | 
|  | 1403 | disk->first_minor = k; | 
|  | 1404 | sdp->disk = disk; | 
|  | 1405 | sdp->device = scsidp; | 
|  | 1406 | INIT_LIST_HEAD(&sdp->sfds); | 
|  | 1407 | init_waitqueue_head(&sdp->o_excl_wait); | 
|  | 1408 | sdp->sg_tablesize = queue_max_segments(q); | 
|  | 1409 | sdp->index = k; | 
|  | 1410 | kref_init(&sdp->d_ref); | 
|  | 1411 |  | 
|  | 1412 | write_unlock_irqrestore(&sg_index_lock, iflags); | 
|  | 1413 |  | 
|  | 1414 | error = 0; | 
|  | 1415 | out: | 
|  | 1416 | if (error) { | 
|  | 1417 | kfree(sdp); | 
|  | 1418 | return ERR_PTR(error); | 
|  | 1419 | } | 
|  | 1420 | return sdp; | 
|  | 1421 |  | 
|  | 1422 | overflow: | 
|  | 1423 | idr_remove(&sg_index_idr, k); | 
|  | 1424 | write_unlock_irqrestore(&sg_index_lock, iflags); | 
|  | 1425 | sdev_printk(KERN_WARNING, scsidp, | 
|  | 1426 | "Unable to attach sg device type=%d, minor " | 
|  | 1427 | "number exceeds %d\n", scsidp->type, SG_MAX_DEVS - 1); | 
|  | 1428 | error = -ENODEV; | 
|  | 1429 | goto out; | 
|  | 1430 | } | 
|  | 1431 |  | 
|  | 1432 | static int | 
|  | 1433 | sg_add(struct device *cl_dev, struct class_interface *cl_intf) | 
|  | 1434 | { | 
|  | 1435 | struct scsi_device *scsidp = to_scsi_device(cl_dev->parent); | 
|  | 1436 | struct gendisk *disk; | 
|  | 1437 | Sg_device *sdp = NULL; | 
|  | 1438 | struct cdev * cdev = NULL; | 
|  | 1439 | int error; | 
|  | 1440 | unsigned long iflags; | 
|  | 1441 |  | 
|  | 1442 | disk = alloc_disk(1); | 
|  | 1443 | if (!disk) { | 
|  | 1444 | printk(KERN_WARNING "alloc_disk failed\n"); | 
|  | 1445 | return -ENOMEM; | 
|  | 1446 | } | 
|  | 1447 | disk->major = SCSI_GENERIC_MAJOR; | 
|  | 1448 |  | 
|  | 1449 | error = -ENOMEM; | 
|  | 1450 | cdev = cdev_alloc(); | 
|  | 1451 | if (!cdev) { | 
|  | 1452 | printk(KERN_WARNING "cdev_alloc failed\n"); | 
|  | 1453 | goto out; | 
|  | 1454 | } | 
|  | 1455 | cdev->owner = THIS_MODULE; | 
|  | 1456 | cdev->ops = &sg_fops; | 
|  | 1457 |  | 
|  | 1458 | sdp = sg_alloc(disk, scsidp); | 
|  | 1459 | if (IS_ERR(sdp)) { | 
|  | 1460 | printk(KERN_WARNING "sg_alloc failed\n"); | 
|  | 1461 | error = PTR_ERR(sdp); | 
|  | 1462 | goto out; | 
|  | 1463 | } | 
|  | 1464 |  | 
|  | 1465 | error = cdev_add(cdev, MKDEV(SCSI_GENERIC_MAJOR, sdp->index), 1); | 
|  | 1466 | if (error) | 
|  | 1467 | goto cdev_add_err; | 
|  | 1468 |  | 
|  | 1469 | sdp->cdev = cdev; | 
|  | 1470 | if (sg_sysfs_valid) { | 
|  | 1471 | struct device *sg_class_member; | 
|  | 1472 |  | 
|  | 1473 | sg_class_member = device_create(sg_sysfs_class, cl_dev->parent, | 
|  | 1474 | MKDEV(SCSI_GENERIC_MAJOR, | 
|  | 1475 | sdp->index), | 
|  | 1476 | sdp, "%s", disk->disk_name); | 
|  | 1477 | if (IS_ERR(sg_class_member)) { | 
|  | 1478 | printk(KERN_ERR "sg_add: " | 
|  | 1479 | "device_create failed\n"); | 
|  | 1480 | error = PTR_ERR(sg_class_member); | 
|  | 1481 | goto cdev_add_err; | 
|  | 1482 | } | 
|  | 1483 | error = sysfs_create_link(&scsidp->sdev_gendev.kobj, | 
|  | 1484 | &sg_class_member->kobj, "generic"); | 
|  | 1485 | if (error) | 
|  | 1486 | printk(KERN_ERR "sg_add: unable to make symlink " | 
|  | 1487 | "'generic' back to sg%d\n", sdp->index); | 
|  | 1488 | } else | 
|  | 1489 | printk(KERN_WARNING "sg_add: sg_sys Invalid\n"); | 
|  | 1490 |  | 
|  | 1491 | sdev_printk(KERN_NOTICE, scsidp, | 
|  | 1492 | "Attached scsi generic sg%d type %d\n", sdp->index, | 
|  | 1493 | scsidp->type); | 
|  | 1494 |  | 
|  | 1495 | dev_set_drvdata(cl_dev, sdp); | 
|  | 1496 |  | 
|  | 1497 | return 0; | 
|  | 1498 |  | 
|  | 1499 | cdev_add_err: | 
|  | 1500 | write_lock_irqsave(&sg_index_lock, iflags); | 
|  | 1501 | idr_remove(&sg_index_idr, sdp->index); | 
|  | 1502 | write_unlock_irqrestore(&sg_index_lock, iflags); | 
|  | 1503 | kfree(sdp); | 
|  | 1504 |  | 
|  | 1505 | out: | 
|  | 1506 | put_disk(disk); | 
|  | 1507 | if (cdev) | 
|  | 1508 | cdev_del(cdev); | 
|  | 1509 | return error; | 
|  | 1510 | } | 
|  | 1511 |  | 
|  | 1512 | static void sg_device_destroy(struct kref *kref) | 
|  | 1513 | { | 
|  | 1514 | struct sg_device *sdp = container_of(kref, struct sg_device, d_ref); | 
|  | 1515 | unsigned long flags; | 
|  | 1516 |  | 
|  | 1517 | /* CAUTION!  Note that the device can still be found via idr_find() | 
|  | 1518 | * even though the refcount is 0.  Therefore, do idr_remove() BEFORE | 
|  | 1519 | * any other cleanup. | 
|  | 1520 | */ | 
|  | 1521 |  | 
|  | 1522 | write_lock_irqsave(&sg_index_lock, flags); | 
|  | 1523 | idr_remove(&sg_index_idr, sdp->index); | 
|  | 1524 | write_unlock_irqrestore(&sg_index_lock, flags); | 
|  | 1525 |  | 
|  | 1526 | SCSI_LOG_TIMEOUT(3, | 
|  | 1527 | printk("sg_device_destroy: %s\n", | 
|  | 1528 | sdp->disk->disk_name)); | 
|  | 1529 |  | 
|  | 1530 | put_disk(sdp->disk); | 
|  | 1531 | kfree(sdp); | 
|  | 1532 | } | 
|  | 1533 |  | 
|  | 1534 | static void sg_remove(struct device *cl_dev, struct class_interface *cl_intf) | 
|  | 1535 | { | 
|  | 1536 | struct scsi_device *scsidp = to_scsi_device(cl_dev->parent); | 
|  | 1537 | Sg_device *sdp = dev_get_drvdata(cl_dev); | 
|  | 1538 | unsigned long iflags; | 
|  | 1539 | Sg_fd *sfp; | 
|  | 1540 |  | 
|  | 1541 | if (!sdp || sdp->detached) | 
|  | 1542 | return; | 
|  | 1543 |  | 
|  | 1544 | SCSI_LOG_TIMEOUT(3, printk("sg_remove: %s\n", sdp->disk->disk_name)); | 
|  | 1545 |  | 
|  | 1546 | /* Need a write lock to set sdp->detached. */ | 
|  | 1547 | write_lock_irqsave(&sg_index_lock, iflags); | 
|  | 1548 | sdp->detached = 1; | 
|  | 1549 | list_for_each_entry(sfp, &sdp->sfds, sfd_siblings) { | 
|  | 1550 | wake_up_interruptible(&sfp->read_wait); | 
|  | 1551 | kill_fasync(&sfp->async_qp, SIGPOLL, POLL_HUP); | 
|  | 1552 | } | 
|  | 1553 | write_unlock_irqrestore(&sg_index_lock, iflags); | 
|  | 1554 |  | 
|  | 1555 | sysfs_remove_link(&scsidp->sdev_gendev.kobj, "generic"); | 
|  | 1556 | device_destroy(sg_sysfs_class, MKDEV(SCSI_GENERIC_MAJOR, sdp->index)); | 
|  | 1557 | cdev_del(sdp->cdev); | 
|  | 1558 | sdp->cdev = NULL; | 
|  | 1559 |  | 
|  | 1560 | sg_put_dev(sdp); | 
|  | 1561 | } | 
|  | 1562 |  | 
|  | 1563 | module_param_named(scatter_elem_sz, scatter_elem_sz, int, S_IRUGO | S_IWUSR); | 
|  | 1564 | module_param_named(def_reserved_size, def_reserved_size, int, | 
|  | 1565 | S_IRUGO | S_IWUSR); | 
|  | 1566 | module_param_named(allow_dio, sg_allow_dio, int, S_IRUGO | S_IWUSR); | 
|  | 1567 |  | 
|  | 1568 | MODULE_AUTHOR("Douglas Gilbert"); | 
|  | 1569 | MODULE_DESCRIPTION("SCSI generic (sg) driver"); | 
|  | 1570 | MODULE_LICENSE("GPL"); | 
|  | 1571 | MODULE_VERSION(SG_VERSION_STR); | 
|  | 1572 | MODULE_ALIAS_CHARDEV_MAJOR(SCSI_GENERIC_MAJOR); | 
|  | 1573 |  | 
|  | 1574 | MODULE_PARM_DESC(scatter_elem_sz, "scatter gather element " | 
|  | 1575 | "size (default: max(SG_SCATTER_SZ, PAGE_SIZE))"); | 
|  | 1576 | MODULE_PARM_DESC(def_reserved_size, "size of buffer reserved for each fd"); | 
|  | 1577 | MODULE_PARM_DESC(allow_dio, "allow direct I/O (default: 0 (disallow))"); | 
|  | 1578 |  | 
|  | 1579 | static int __init | 
|  | 1580 | init_sg(void) | 
|  | 1581 | { | 
|  | 1582 | int rc; | 
|  | 1583 |  | 
|  | 1584 | if (scatter_elem_sz < PAGE_SIZE) { | 
|  | 1585 | scatter_elem_sz = PAGE_SIZE; | 
|  | 1586 | scatter_elem_sz_prev = scatter_elem_sz; | 
|  | 1587 | } | 
|  | 1588 | if (def_reserved_size >= 0) | 
|  | 1589 | sg_big_buff = def_reserved_size; | 
|  | 1590 | else | 
|  | 1591 | def_reserved_size = sg_big_buff; | 
|  | 1592 |  | 
|  | 1593 | rc = register_chrdev_region(MKDEV(SCSI_GENERIC_MAJOR, 0), | 
|  | 1594 | SG_MAX_DEVS, "sg"); | 
|  | 1595 | if (rc) | 
|  | 1596 | return rc; | 
|  | 1597 | sg_sysfs_class = class_create(THIS_MODULE, "scsi_generic"); | 
|  | 1598 | if ( IS_ERR(sg_sysfs_class) ) { | 
|  | 1599 | rc = PTR_ERR(sg_sysfs_class); | 
|  | 1600 | goto err_out; | 
|  | 1601 | } | 
|  | 1602 | sg_sysfs_valid = 1; | 
|  | 1603 | rc = scsi_register_interface(&sg_interface); | 
|  | 1604 | if (0 == rc) { | 
|  | 1605 | #ifdef CONFIG_SCSI_PROC_FS | 
|  | 1606 | sg_proc_init(); | 
|  | 1607 | #endif				/* CONFIG_SCSI_PROC_FS */ | 
|  | 1608 | return 0; | 
|  | 1609 | } | 
|  | 1610 | class_destroy(sg_sysfs_class); | 
|  | 1611 | err_out: | 
|  | 1612 | unregister_chrdev_region(MKDEV(SCSI_GENERIC_MAJOR, 0), SG_MAX_DEVS); | 
|  | 1613 | return rc; | 
|  | 1614 | } | 
|  | 1615 |  | 
|  | 1616 | static void __exit | 
|  | 1617 | exit_sg(void) | 
|  | 1618 | { | 
|  | 1619 | #ifdef CONFIG_SCSI_PROC_FS | 
|  | 1620 | sg_proc_cleanup(); | 
|  | 1621 | #endif				/* CONFIG_SCSI_PROC_FS */ | 
|  | 1622 | scsi_unregister_interface(&sg_interface); | 
|  | 1623 | class_destroy(sg_sysfs_class); | 
|  | 1624 | sg_sysfs_valid = 0; | 
|  | 1625 | unregister_chrdev_region(MKDEV(SCSI_GENERIC_MAJOR, 0), | 
|  | 1626 | SG_MAX_DEVS); | 
|  | 1627 | idr_destroy(&sg_index_idr); | 
|  | 1628 | } | 
|  | 1629 |  | 
|  | 1630 | static int sg_start_req(Sg_request *srp, unsigned char *cmd) | 
|  | 1631 | { | 
|  | 1632 | int res; | 
|  | 1633 | struct request *rq; | 
|  | 1634 | Sg_fd *sfp = srp->parentfp; | 
|  | 1635 | sg_io_hdr_t *hp = &srp->header; | 
|  | 1636 | int dxfer_len = (int) hp->dxfer_len; | 
|  | 1637 | int dxfer_dir = hp->dxfer_direction; | 
|  | 1638 | unsigned int iov_count = hp->iovec_count; | 
|  | 1639 | Sg_scatter_hold *req_schp = &srp->data; | 
|  | 1640 | Sg_scatter_hold *rsv_schp = &sfp->reserve; | 
|  | 1641 | struct request_queue *q = sfp->parentdp->device->request_queue; | 
|  | 1642 | struct rq_map_data *md, map_data; | 
|  | 1643 | int rw = hp->dxfer_direction == SG_DXFER_TO_DEV ? WRITE : READ; | 
|  | 1644 |  | 
|  | 1645 | SCSI_LOG_TIMEOUT(4, printk(KERN_INFO "sg_start_req: dxfer_len=%d\n", | 
|  | 1646 | dxfer_len)); | 
|  | 1647 |  | 
|  | 1648 | rq = blk_get_request(q, rw, GFP_ATOMIC); | 
|  | 1649 | if (!rq) | 
|  | 1650 | return -ENOMEM; | 
|  | 1651 |  | 
|  | 1652 | memcpy(rq->cmd, cmd, hp->cmd_len); | 
|  | 1653 |  | 
|  | 1654 | rq->cmd_len = hp->cmd_len; | 
|  | 1655 | rq->cmd_type = REQ_TYPE_BLOCK_PC; | 
|  | 1656 |  | 
|  | 1657 | srp->rq = rq; | 
|  | 1658 | rq->end_io_data = srp; | 
|  | 1659 | rq->sense = srp->sense_b; | 
|  | 1660 | rq->retries = SG_DEFAULT_RETRIES; | 
|  | 1661 |  | 
|  | 1662 | if ((dxfer_len <= 0) || (dxfer_dir == SG_DXFER_NONE)) | 
|  | 1663 | return 0; | 
|  | 1664 |  | 
|  | 1665 | if (sg_allow_dio && hp->flags & SG_FLAG_DIRECT_IO && | 
|  | 1666 | dxfer_dir != SG_DXFER_UNKNOWN && !iov_count && | 
|  | 1667 | !sfp->parentdp->device->host->unchecked_isa_dma && | 
|  | 1668 | blk_rq_aligned(q, (unsigned long)hp->dxferp, dxfer_len)) | 
|  | 1669 | md = NULL; | 
|  | 1670 | else | 
|  | 1671 | md = &map_data; | 
|  | 1672 |  | 
|  | 1673 | if (md) { | 
|  | 1674 | if (!sg_res_in_use(sfp) && dxfer_len <= rsv_schp->bufflen) | 
|  | 1675 | sg_link_reserve(sfp, srp, dxfer_len); | 
|  | 1676 | else { | 
|  | 1677 | res = sg_build_indirect(req_schp, sfp, dxfer_len); | 
|  | 1678 | if (res) | 
|  | 1679 | return res; | 
|  | 1680 | } | 
|  | 1681 |  | 
|  | 1682 | md->pages = req_schp->pages; | 
|  | 1683 | md->page_order = req_schp->page_order; | 
|  | 1684 | md->nr_entries = req_schp->k_use_sg; | 
|  | 1685 | md->offset = 0; | 
|  | 1686 | md->null_mapped = hp->dxferp ? 0 : 1; | 
|  | 1687 | if (dxfer_dir == SG_DXFER_TO_FROM_DEV) | 
|  | 1688 | md->from_user = 1; | 
|  | 1689 | else | 
|  | 1690 | md->from_user = 0; | 
|  | 1691 | } | 
|  | 1692 |  | 
|  | 1693 | if (iov_count) { | 
|  | 1694 | int len, size = sizeof(struct sg_iovec) * iov_count; | 
|  | 1695 | struct iovec *iov; | 
|  | 1696 |  | 
|  | 1697 | iov = memdup_user(hp->dxferp, size); | 
|  | 1698 | if (IS_ERR(iov)) | 
|  | 1699 | return PTR_ERR(iov); | 
|  | 1700 |  | 
|  | 1701 | len = iov_length(iov, iov_count); | 
|  | 1702 | if (hp->dxfer_len < len) { | 
|  | 1703 | iov_count = iov_shorten(iov, iov_count, hp->dxfer_len); | 
|  | 1704 | len = hp->dxfer_len; | 
|  | 1705 | } | 
|  | 1706 |  | 
|  | 1707 | res = blk_rq_map_user_iov(q, rq, md, (struct sg_iovec *)iov, | 
|  | 1708 | iov_count, | 
|  | 1709 | len, GFP_ATOMIC); | 
|  | 1710 | kfree(iov); | 
|  | 1711 | } else | 
|  | 1712 | res = blk_rq_map_user(q, rq, md, hp->dxferp, | 
|  | 1713 | hp->dxfer_len, GFP_ATOMIC); | 
|  | 1714 |  | 
|  | 1715 | if (!res) { | 
|  | 1716 | srp->bio = rq->bio; | 
|  | 1717 |  | 
|  | 1718 | if (!md) { | 
|  | 1719 | req_schp->dio_in_use = 1; | 
|  | 1720 | hp->info |= SG_INFO_DIRECT_IO; | 
|  | 1721 | } | 
|  | 1722 | } | 
|  | 1723 | return res; | 
|  | 1724 | } | 
|  | 1725 |  | 
|  | 1726 | static int sg_finish_rem_req(Sg_request * srp) | 
|  | 1727 | { | 
|  | 1728 | int ret = 0; | 
|  | 1729 |  | 
|  | 1730 | Sg_fd *sfp = srp->parentfp; | 
|  | 1731 | Sg_scatter_hold *req_schp = &srp->data; | 
|  | 1732 |  | 
|  | 1733 | SCSI_LOG_TIMEOUT(4, printk("sg_finish_rem_req: res_used=%d\n", (int) srp->res_used)); | 
|  | 1734 | if (srp->rq) { | 
|  | 1735 | if (srp->bio) | 
|  | 1736 | ret = blk_rq_unmap_user(srp->bio); | 
|  | 1737 |  | 
|  | 1738 | blk_put_request(srp->rq); | 
|  | 1739 | } | 
|  | 1740 |  | 
|  | 1741 | if (srp->res_used) | 
|  | 1742 | sg_unlink_reserve(sfp, srp); | 
|  | 1743 | else | 
|  | 1744 | sg_remove_scat(req_schp); | 
|  | 1745 |  | 
|  | 1746 | sg_remove_request(sfp, srp); | 
|  | 1747 |  | 
|  | 1748 | return ret; | 
|  | 1749 | } | 
|  | 1750 |  | 
|  | 1751 | static int | 
|  | 1752 | sg_build_sgat(Sg_scatter_hold * schp, const Sg_fd * sfp, int tablesize) | 
|  | 1753 | { | 
|  | 1754 | int sg_bufflen = tablesize * sizeof(struct page *); | 
|  | 1755 | gfp_t gfp_flags = GFP_ATOMIC | __GFP_NOWARN; | 
|  | 1756 |  | 
|  | 1757 | schp->pages = kzalloc(sg_bufflen, gfp_flags); | 
|  | 1758 | if (!schp->pages) | 
|  | 1759 | return -ENOMEM; | 
|  | 1760 | schp->sglist_len = sg_bufflen; | 
|  | 1761 | return tablesize;	/* number of scat_gath elements allocated */ | 
|  | 1762 | } | 
|  | 1763 |  | 
|  | 1764 | static int | 
|  | 1765 | sg_build_indirect(Sg_scatter_hold * schp, Sg_fd * sfp, int buff_size) | 
|  | 1766 | { | 
|  | 1767 | int ret_sz = 0, i, k, rem_sz, num, mx_sc_elems; | 
|  | 1768 | int sg_tablesize = sfp->parentdp->sg_tablesize; | 
|  | 1769 | int blk_size = buff_size, order; | 
|  | 1770 | gfp_t gfp_mask = GFP_ATOMIC | __GFP_COMP | __GFP_NOWARN; | 
|  | 1771 |  | 
|  | 1772 | if (blk_size < 0) | 
|  | 1773 | return -EFAULT; | 
|  | 1774 | if (0 == blk_size) | 
|  | 1775 | ++blk_size;	/* don't know why */ | 
|  | 1776 | /* round request up to next highest SG_SECTOR_SZ byte boundary */ | 
|  | 1777 | blk_size = ALIGN(blk_size, SG_SECTOR_SZ); | 
|  | 1778 | SCSI_LOG_TIMEOUT(4, printk("sg_build_indirect: buff_size=%d, blk_size=%d\n", | 
|  | 1779 | buff_size, blk_size)); | 
|  | 1780 |  | 
|  | 1781 | /* N.B. ret_sz carried into this block ... */ | 
|  | 1782 | mx_sc_elems = sg_build_sgat(schp, sfp, sg_tablesize); | 
|  | 1783 | if (mx_sc_elems < 0) | 
|  | 1784 | return mx_sc_elems;	/* most likely -ENOMEM */ | 
|  | 1785 |  | 
|  | 1786 | num = scatter_elem_sz; | 
|  | 1787 | if (unlikely(num != scatter_elem_sz_prev)) { | 
|  | 1788 | if (num < PAGE_SIZE) { | 
|  | 1789 | scatter_elem_sz = PAGE_SIZE; | 
|  | 1790 | scatter_elem_sz_prev = PAGE_SIZE; | 
|  | 1791 | } else | 
|  | 1792 | scatter_elem_sz_prev = num; | 
|  | 1793 | } | 
|  | 1794 |  | 
|  | 1795 | if (sfp->low_dma) | 
|  | 1796 | gfp_mask |= GFP_DMA; | 
|  | 1797 |  | 
|  | 1798 | if (!capable(CAP_SYS_ADMIN) || !capable(CAP_SYS_RAWIO)) | 
|  | 1799 | gfp_mask |= __GFP_ZERO; | 
|  | 1800 |  | 
|  | 1801 | order = get_order(num); | 
|  | 1802 | retry: | 
|  | 1803 | ret_sz = 1 << (PAGE_SHIFT + order); | 
|  | 1804 |  | 
|  | 1805 | for (k = 0, rem_sz = blk_size; rem_sz > 0 && k < mx_sc_elems; | 
|  | 1806 | k++, rem_sz -= ret_sz) { | 
|  | 1807 |  | 
|  | 1808 | num = (rem_sz > scatter_elem_sz_prev) ? | 
|  | 1809 | scatter_elem_sz_prev : rem_sz; | 
|  | 1810 |  | 
|  | 1811 | schp->pages[k] = alloc_pages(gfp_mask, order); | 
|  | 1812 | if (!schp->pages[k]) | 
|  | 1813 | goto out; | 
|  | 1814 |  | 
|  | 1815 | if (num == scatter_elem_sz_prev) { | 
|  | 1816 | if (unlikely(ret_sz > scatter_elem_sz_prev)) { | 
|  | 1817 | scatter_elem_sz = ret_sz; | 
|  | 1818 | scatter_elem_sz_prev = ret_sz; | 
|  | 1819 | } | 
|  | 1820 | } | 
|  | 1821 |  | 
|  | 1822 | SCSI_LOG_TIMEOUT(5, printk("sg_build_indirect: k=%d, num=%d, " | 
|  | 1823 | "ret_sz=%d\n", k, num, ret_sz)); | 
|  | 1824 | }		/* end of for loop */ | 
|  | 1825 |  | 
|  | 1826 | schp->page_order = order; | 
|  | 1827 | schp->k_use_sg = k; | 
|  | 1828 | SCSI_LOG_TIMEOUT(5, printk("sg_build_indirect: k_use_sg=%d, " | 
|  | 1829 | "rem_sz=%d\n", k, rem_sz)); | 
|  | 1830 |  | 
|  | 1831 | schp->bufflen = blk_size; | 
|  | 1832 | if (rem_sz > 0)	/* must have failed */ | 
|  | 1833 | return -ENOMEM; | 
|  | 1834 | return 0; | 
|  | 1835 | out: | 
|  | 1836 | for (i = 0; i < k; i++) | 
|  | 1837 | __free_pages(schp->pages[i], order); | 
|  | 1838 |  | 
|  | 1839 | if (--order >= 0) | 
|  | 1840 | goto retry; | 
|  | 1841 |  | 
|  | 1842 | return -ENOMEM; | 
|  | 1843 | } | 
|  | 1844 |  | 
|  | 1845 | static void | 
|  | 1846 | sg_remove_scat(Sg_scatter_hold * schp) | 
|  | 1847 | { | 
|  | 1848 | SCSI_LOG_TIMEOUT(4, printk("sg_remove_scat: k_use_sg=%d\n", schp->k_use_sg)); | 
|  | 1849 | if (schp->pages && schp->sglist_len > 0) { | 
|  | 1850 | if (!schp->dio_in_use) { | 
|  | 1851 | int k; | 
|  | 1852 |  | 
|  | 1853 | for (k = 0; k < schp->k_use_sg && schp->pages[k]; k++) { | 
|  | 1854 | SCSI_LOG_TIMEOUT(5, printk( | 
|  | 1855 | "sg_remove_scat: k=%d, pg=0x%p\n", | 
|  | 1856 | k, schp->pages[k])); | 
|  | 1857 | __free_pages(schp->pages[k], schp->page_order); | 
|  | 1858 | } | 
|  | 1859 |  | 
|  | 1860 | kfree(schp->pages); | 
|  | 1861 | } | 
|  | 1862 | } | 
|  | 1863 | memset(schp, 0, sizeof (*schp)); | 
|  | 1864 | } | 
|  | 1865 |  | 
|  | 1866 | static int | 
|  | 1867 | sg_read_oxfer(Sg_request * srp, char __user *outp, int num_read_xfer) | 
|  | 1868 | { | 
|  | 1869 | Sg_scatter_hold *schp = &srp->data; | 
|  | 1870 | int k, num; | 
|  | 1871 |  | 
|  | 1872 | SCSI_LOG_TIMEOUT(4, printk("sg_read_oxfer: num_read_xfer=%d\n", | 
|  | 1873 | num_read_xfer)); | 
|  | 1874 | if ((!outp) || (num_read_xfer <= 0)) | 
|  | 1875 | return 0; | 
|  | 1876 |  | 
|  | 1877 | num = 1 << (PAGE_SHIFT + schp->page_order); | 
|  | 1878 | for (k = 0; k < schp->k_use_sg && schp->pages[k]; k++) { | 
|  | 1879 | if (num > num_read_xfer) { | 
|  | 1880 | if (__copy_to_user(outp, page_address(schp->pages[k]), | 
|  | 1881 | num_read_xfer)) | 
|  | 1882 | return -EFAULT; | 
|  | 1883 | break; | 
|  | 1884 | } else { | 
|  | 1885 | if (__copy_to_user(outp, page_address(schp->pages[k]), | 
|  | 1886 | num)) | 
|  | 1887 | return -EFAULT; | 
|  | 1888 | num_read_xfer -= num; | 
|  | 1889 | if (num_read_xfer <= 0) | 
|  | 1890 | break; | 
|  | 1891 | outp += num; | 
|  | 1892 | } | 
|  | 1893 | } | 
|  | 1894 |  | 
|  | 1895 | return 0; | 
|  | 1896 | } | 
|  | 1897 |  | 
|  | 1898 | static void | 
|  | 1899 | sg_build_reserve(Sg_fd * sfp, int req_size) | 
|  | 1900 | { | 
|  | 1901 | Sg_scatter_hold *schp = &sfp->reserve; | 
|  | 1902 |  | 
|  | 1903 | SCSI_LOG_TIMEOUT(4, printk("sg_build_reserve: req_size=%d\n", req_size)); | 
|  | 1904 | do { | 
|  | 1905 | if (req_size < PAGE_SIZE) | 
|  | 1906 | req_size = PAGE_SIZE; | 
|  | 1907 | if (0 == sg_build_indirect(schp, sfp, req_size)) | 
|  | 1908 | return; | 
|  | 1909 | else | 
|  | 1910 | sg_remove_scat(schp); | 
|  | 1911 | req_size >>= 1;	/* divide by 2 */ | 
|  | 1912 | } while (req_size > (PAGE_SIZE / 2)); | 
|  | 1913 | } | 
|  | 1914 |  | 
|  | 1915 | static void | 
|  | 1916 | sg_link_reserve(Sg_fd * sfp, Sg_request * srp, int size) | 
|  | 1917 | { | 
|  | 1918 | Sg_scatter_hold *req_schp = &srp->data; | 
|  | 1919 | Sg_scatter_hold *rsv_schp = &sfp->reserve; | 
|  | 1920 | int k, num, rem; | 
|  | 1921 |  | 
|  | 1922 | srp->res_used = 1; | 
|  | 1923 | SCSI_LOG_TIMEOUT(4, printk("sg_link_reserve: size=%d\n", size)); | 
|  | 1924 | rem = size; | 
|  | 1925 |  | 
|  | 1926 | num = 1 << (PAGE_SHIFT + rsv_schp->page_order); | 
|  | 1927 | for (k = 0; k < rsv_schp->k_use_sg; k++) { | 
|  | 1928 | if (rem <= num) { | 
|  | 1929 | req_schp->k_use_sg = k + 1; | 
|  | 1930 | req_schp->sglist_len = rsv_schp->sglist_len; | 
|  | 1931 | req_schp->pages = rsv_schp->pages; | 
|  | 1932 |  | 
|  | 1933 | req_schp->bufflen = size; | 
|  | 1934 | req_schp->page_order = rsv_schp->page_order; | 
|  | 1935 | break; | 
|  | 1936 | } else | 
|  | 1937 | rem -= num; | 
|  | 1938 | } | 
|  | 1939 |  | 
|  | 1940 | if (k >= rsv_schp->k_use_sg) | 
|  | 1941 | SCSI_LOG_TIMEOUT(1, printk("sg_link_reserve: BAD size\n")); | 
|  | 1942 | } | 
|  | 1943 |  | 
|  | 1944 | static void | 
|  | 1945 | sg_unlink_reserve(Sg_fd * sfp, Sg_request * srp) | 
|  | 1946 | { | 
|  | 1947 | Sg_scatter_hold *req_schp = &srp->data; | 
|  | 1948 |  | 
|  | 1949 | SCSI_LOG_TIMEOUT(4, printk("sg_unlink_reserve: req->k_use_sg=%d\n", | 
|  | 1950 | (int) req_schp->k_use_sg)); | 
|  | 1951 | req_schp->k_use_sg = 0; | 
|  | 1952 | req_schp->bufflen = 0; | 
|  | 1953 | req_schp->pages = NULL; | 
|  | 1954 | req_schp->page_order = 0; | 
|  | 1955 | req_schp->sglist_len = 0; | 
|  | 1956 | sfp->save_scat_len = 0; | 
|  | 1957 | srp->res_used = 0; | 
|  | 1958 | } | 
|  | 1959 |  | 
|  | 1960 | static Sg_request * | 
|  | 1961 | sg_get_rq_mark(Sg_fd * sfp, int pack_id) | 
|  | 1962 | { | 
|  | 1963 | Sg_request *resp; | 
|  | 1964 | unsigned long iflags; | 
|  | 1965 |  | 
|  | 1966 | write_lock_irqsave(&sfp->rq_list_lock, iflags); | 
|  | 1967 | for (resp = sfp->headrp; resp; resp = resp->nextrp) { | 
|  | 1968 | /* look for requests that are ready + not SG_IO owned */ | 
|  | 1969 | if ((1 == resp->done) && (!resp->sg_io_owned) && | 
|  | 1970 | ((-1 == pack_id) || (resp->header.pack_id == pack_id))) { | 
|  | 1971 | resp->done = 2;	/* guard against other readers */ | 
|  | 1972 | break; | 
|  | 1973 | } | 
|  | 1974 | } | 
|  | 1975 | write_unlock_irqrestore(&sfp->rq_list_lock, iflags); | 
|  | 1976 | return resp; | 
|  | 1977 | } | 
|  | 1978 |  | 
|  | 1979 | /* always adds to end of list */ | 
|  | 1980 | static Sg_request * | 
|  | 1981 | sg_add_request(Sg_fd * sfp) | 
|  | 1982 | { | 
|  | 1983 | int k; | 
|  | 1984 | unsigned long iflags; | 
|  | 1985 | Sg_request *resp; | 
|  | 1986 | Sg_request *rp = sfp->req_arr; | 
|  | 1987 |  | 
|  | 1988 | write_lock_irqsave(&sfp->rq_list_lock, iflags); | 
|  | 1989 | resp = sfp->headrp; | 
|  | 1990 | if (!resp) { | 
|  | 1991 | memset(rp, 0, sizeof (Sg_request)); | 
|  | 1992 | rp->parentfp = sfp; | 
|  | 1993 | resp = rp; | 
|  | 1994 | sfp->headrp = resp; | 
|  | 1995 | } else { | 
|  | 1996 | if (0 == sfp->cmd_q) | 
|  | 1997 | resp = NULL;	/* command queuing disallowed */ | 
|  | 1998 | else { | 
|  | 1999 | for (k = 0; k < SG_MAX_QUEUE; ++k, ++rp) { | 
|  | 2000 | if (!rp->parentfp) | 
|  | 2001 | break; | 
|  | 2002 | } | 
|  | 2003 | if (k < SG_MAX_QUEUE) { | 
|  | 2004 | memset(rp, 0, sizeof (Sg_request)); | 
|  | 2005 | rp->parentfp = sfp; | 
|  | 2006 | while (resp->nextrp) | 
|  | 2007 | resp = resp->nextrp; | 
|  | 2008 | resp->nextrp = rp; | 
|  | 2009 | resp = rp; | 
|  | 2010 | } else | 
|  | 2011 | resp = NULL; | 
|  | 2012 | } | 
|  | 2013 | } | 
|  | 2014 | if (resp) { | 
|  | 2015 | resp->nextrp = NULL; | 
|  | 2016 | resp->header.duration = jiffies_to_msecs(jiffies); | 
|  | 2017 | } | 
|  | 2018 | write_unlock_irqrestore(&sfp->rq_list_lock, iflags); | 
|  | 2019 | return resp; | 
|  | 2020 | } | 
|  | 2021 |  | 
|  | 2022 | /* Return of 1 for found; 0 for not found */ | 
|  | 2023 | static int | 
|  | 2024 | sg_remove_request(Sg_fd * sfp, Sg_request * srp) | 
|  | 2025 | { | 
|  | 2026 | Sg_request *prev_rp; | 
|  | 2027 | Sg_request *rp; | 
|  | 2028 | unsigned long iflags; | 
|  | 2029 | int res = 0; | 
|  | 2030 |  | 
|  | 2031 | if ((!sfp) || (!srp) || (!sfp->headrp)) | 
|  | 2032 | return res; | 
|  | 2033 | write_lock_irqsave(&sfp->rq_list_lock, iflags); | 
|  | 2034 | prev_rp = sfp->headrp; | 
|  | 2035 | if (srp == prev_rp) { | 
|  | 2036 | sfp->headrp = prev_rp->nextrp; | 
|  | 2037 | prev_rp->parentfp = NULL; | 
|  | 2038 | res = 1; | 
|  | 2039 | } else { | 
|  | 2040 | while ((rp = prev_rp->nextrp)) { | 
|  | 2041 | if (srp == rp) { | 
|  | 2042 | prev_rp->nextrp = rp->nextrp; | 
|  | 2043 | rp->parentfp = NULL; | 
|  | 2044 | res = 1; | 
|  | 2045 | break; | 
|  | 2046 | } | 
|  | 2047 | prev_rp = rp; | 
|  | 2048 | } | 
|  | 2049 | } | 
|  | 2050 | write_unlock_irqrestore(&sfp->rq_list_lock, iflags); | 
|  | 2051 | return res; | 
|  | 2052 | } | 
|  | 2053 |  | 
|  | 2054 | static Sg_fd * | 
|  | 2055 | sg_add_sfp(Sg_device * sdp, int dev) | 
|  | 2056 | { | 
|  | 2057 | Sg_fd *sfp; | 
|  | 2058 | unsigned long iflags; | 
|  | 2059 | int bufflen; | 
|  | 2060 |  | 
|  | 2061 | sfp = kzalloc(sizeof(*sfp), GFP_ATOMIC | __GFP_NOWARN); | 
|  | 2062 | if (!sfp) | 
|  | 2063 | return NULL; | 
|  | 2064 |  | 
|  | 2065 | init_waitqueue_head(&sfp->read_wait); | 
|  | 2066 | rwlock_init(&sfp->rq_list_lock); | 
|  | 2067 |  | 
|  | 2068 | kref_init(&sfp->f_ref); | 
|  | 2069 | sfp->timeout = SG_DEFAULT_TIMEOUT; | 
|  | 2070 | sfp->timeout_user = SG_DEFAULT_TIMEOUT_USER; | 
|  | 2071 | sfp->force_packid = SG_DEF_FORCE_PACK_ID; | 
|  | 2072 | sfp->low_dma = (SG_DEF_FORCE_LOW_DMA == 0) ? | 
|  | 2073 | sdp->device->host->unchecked_isa_dma : 1; | 
|  | 2074 | sfp->cmd_q = SG_DEF_COMMAND_Q; | 
|  | 2075 | sfp->keep_orphan = SG_DEF_KEEP_ORPHAN; | 
|  | 2076 | sfp->parentdp = sdp; | 
|  | 2077 | write_lock_irqsave(&sg_index_lock, iflags); | 
|  | 2078 | list_add_tail(&sfp->sfd_siblings, &sdp->sfds); | 
|  | 2079 | write_unlock_irqrestore(&sg_index_lock, iflags); | 
|  | 2080 | SCSI_LOG_TIMEOUT(3, printk("sg_add_sfp: sfp=0x%p\n", sfp)); | 
|  | 2081 | if (unlikely(sg_big_buff != def_reserved_size)) | 
|  | 2082 | sg_big_buff = def_reserved_size; | 
|  | 2083 |  | 
|  | 2084 | bufflen = min_t(int, sg_big_buff, | 
|  | 2085 | queue_max_sectors(sdp->device->request_queue) * 512); | 
|  | 2086 | sg_build_reserve(sfp, bufflen); | 
|  | 2087 | SCSI_LOG_TIMEOUT(3, printk("sg_add_sfp:   bufflen=%d, k_use_sg=%d\n", | 
|  | 2088 | sfp->reserve.bufflen, sfp->reserve.k_use_sg)); | 
|  | 2089 |  | 
|  | 2090 | kref_get(&sdp->d_ref); | 
|  | 2091 | __module_get(THIS_MODULE); | 
|  | 2092 | return sfp; | 
|  | 2093 | } | 
|  | 2094 |  | 
|  | 2095 | static void sg_remove_sfp_usercontext(struct work_struct *work) | 
|  | 2096 | { | 
|  | 2097 | struct sg_fd *sfp = container_of(work, struct sg_fd, ew.work); | 
|  | 2098 | struct sg_device *sdp = sfp->parentdp; | 
|  | 2099 |  | 
|  | 2100 | /* Cleanup any responses which were never read(). */ | 
|  | 2101 | while (sfp->headrp) | 
|  | 2102 | sg_finish_rem_req(sfp->headrp); | 
|  | 2103 |  | 
|  | 2104 | if (sfp->reserve.bufflen > 0) { | 
|  | 2105 | SCSI_LOG_TIMEOUT(6, | 
|  | 2106 | printk("sg_remove_sfp:    bufflen=%d, k_use_sg=%d\n", | 
|  | 2107 | (int) sfp->reserve.bufflen, | 
|  | 2108 | (int) sfp->reserve.k_use_sg)); | 
|  | 2109 | sg_remove_scat(&sfp->reserve); | 
|  | 2110 | } | 
|  | 2111 |  | 
|  | 2112 | SCSI_LOG_TIMEOUT(6, | 
|  | 2113 | printk("sg_remove_sfp: %s, sfp=0x%p\n", | 
|  | 2114 | sdp->disk->disk_name, | 
|  | 2115 | sfp)); | 
|  | 2116 | kfree(sfp); | 
|  | 2117 |  | 
|  | 2118 | scsi_device_put(sdp->device); | 
|  | 2119 | sg_put_dev(sdp); | 
|  | 2120 | module_put(THIS_MODULE); | 
|  | 2121 | } | 
|  | 2122 |  | 
|  | 2123 | static void sg_remove_sfp(struct kref *kref) | 
|  | 2124 | { | 
|  | 2125 | struct sg_fd *sfp = container_of(kref, struct sg_fd, f_ref); | 
|  | 2126 | struct sg_device *sdp = sfp->parentdp; | 
|  | 2127 | unsigned long iflags; | 
|  | 2128 |  | 
|  | 2129 | write_lock_irqsave(&sg_index_lock, iflags); | 
|  | 2130 | list_del(&sfp->sfd_siblings); | 
|  | 2131 | write_unlock_irqrestore(&sg_index_lock, iflags); | 
|  | 2132 | wake_up_interruptible(&sdp->o_excl_wait); | 
|  | 2133 |  | 
|  | 2134 | INIT_WORK(&sfp->ew.work, sg_remove_sfp_usercontext); | 
|  | 2135 | schedule_work(&sfp->ew.work); | 
|  | 2136 | } | 
|  | 2137 |  | 
|  | 2138 | static int | 
|  | 2139 | sg_res_in_use(Sg_fd * sfp) | 
|  | 2140 | { | 
|  | 2141 | const Sg_request *srp; | 
|  | 2142 | unsigned long iflags; | 
|  | 2143 |  | 
|  | 2144 | read_lock_irqsave(&sfp->rq_list_lock, iflags); | 
|  | 2145 | for (srp = sfp->headrp; srp; srp = srp->nextrp) | 
|  | 2146 | if (srp->res_used) | 
|  | 2147 | break; | 
|  | 2148 | read_unlock_irqrestore(&sfp->rq_list_lock, iflags); | 
|  | 2149 | return srp ? 1 : 0; | 
|  | 2150 | } | 
|  | 2151 |  | 
|  | 2152 | #ifdef CONFIG_SCSI_PROC_FS | 
|  | 2153 | static int | 
|  | 2154 | sg_idr_max_id(int id, void *p, void *data) | 
|  | 2155 | { | 
|  | 2156 | int *k = data; | 
|  | 2157 |  | 
|  | 2158 | if (*k < id) | 
|  | 2159 | *k = id; | 
|  | 2160 |  | 
|  | 2161 | return 0; | 
|  | 2162 | } | 
|  | 2163 |  | 
|  | 2164 | static int | 
|  | 2165 | sg_last_dev(void) | 
|  | 2166 | { | 
|  | 2167 | int k = -1; | 
|  | 2168 | unsigned long iflags; | 
|  | 2169 |  | 
|  | 2170 | read_lock_irqsave(&sg_index_lock, iflags); | 
|  | 2171 | idr_for_each(&sg_index_idr, sg_idr_max_id, &k); | 
|  | 2172 | read_unlock_irqrestore(&sg_index_lock, iflags); | 
|  | 2173 | return k + 1;		/* origin 1 */ | 
|  | 2174 | } | 
|  | 2175 | #endif | 
|  | 2176 |  | 
|  | 2177 | /* must be called with sg_index_lock held */ | 
|  | 2178 | static Sg_device *sg_lookup_dev(int dev) | 
|  | 2179 | { | 
|  | 2180 | return idr_find(&sg_index_idr, dev); | 
|  | 2181 | } | 
|  | 2182 |  | 
|  | 2183 | static Sg_device *sg_get_dev(int dev) | 
|  | 2184 | { | 
|  | 2185 | struct sg_device *sdp; | 
|  | 2186 | unsigned long flags; | 
|  | 2187 |  | 
|  | 2188 | read_lock_irqsave(&sg_index_lock, flags); | 
|  | 2189 | sdp = sg_lookup_dev(dev); | 
|  | 2190 | if (!sdp) | 
|  | 2191 | sdp = ERR_PTR(-ENXIO); | 
|  | 2192 | else if (sdp->detached) { | 
|  | 2193 | /* If sdp->detached, then the refcount may already be 0, in | 
|  | 2194 | * which case it would be a bug to do kref_get(). | 
|  | 2195 | */ | 
|  | 2196 | sdp = ERR_PTR(-ENODEV); | 
|  | 2197 | } else | 
|  | 2198 | kref_get(&sdp->d_ref); | 
|  | 2199 | read_unlock_irqrestore(&sg_index_lock, flags); | 
|  | 2200 |  | 
|  | 2201 | return sdp; | 
|  | 2202 | } | 
|  | 2203 |  | 
|  | 2204 | static void sg_put_dev(struct sg_device *sdp) | 
|  | 2205 | { | 
|  | 2206 | kref_put(&sdp->d_ref, sg_device_destroy); | 
|  | 2207 | } | 
|  | 2208 |  | 
|  | 2209 | #ifdef CONFIG_SCSI_PROC_FS | 
|  | 2210 |  | 
|  | 2211 | static struct proc_dir_entry *sg_proc_sgp = NULL; | 
|  | 2212 |  | 
|  | 2213 | static char sg_proc_sg_dirname[] = "scsi/sg"; | 
|  | 2214 |  | 
|  | 2215 | static int sg_proc_seq_show_int(struct seq_file *s, void *v); | 
|  | 2216 |  | 
|  | 2217 | static int sg_proc_single_open_adio(struct inode *inode, struct file *file); | 
|  | 2218 | static ssize_t sg_proc_write_adio(struct file *filp, const char __user *buffer, | 
|  | 2219 | size_t count, loff_t *off); | 
|  | 2220 | static const struct file_operations adio_fops = { | 
|  | 2221 | .owner = THIS_MODULE, | 
|  | 2222 | .open = sg_proc_single_open_adio, | 
|  | 2223 | .read = seq_read, | 
|  | 2224 | .llseek = seq_lseek, | 
|  | 2225 | .write = sg_proc_write_adio, | 
|  | 2226 | .release = single_release, | 
|  | 2227 | }; | 
|  | 2228 |  | 
|  | 2229 | static int sg_proc_single_open_dressz(struct inode *inode, struct file *file); | 
|  | 2230 | static ssize_t sg_proc_write_dressz(struct file *filp, | 
|  | 2231 | const char __user *buffer, size_t count, loff_t *off); | 
|  | 2232 | static const struct file_operations dressz_fops = { | 
|  | 2233 | .owner = THIS_MODULE, | 
|  | 2234 | .open = sg_proc_single_open_dressz, | 
|  | 2235 | .read = seq_read, | 
|  | 2236 | .llseek = seq_lseek, | 
|  | 2237 | .write = sg_proc_write_dressz, | 
|  | 2238 | .release = single_release, | 
|  | 2239 | }; | 
|  | 2240 |  | 
|  | 2241 | static int sg_proc_seq_show_version(struct seq_file *s, void *v); | 
|  | 2242 | static int sg_proc_single_open_version(struct inode *inode, struct file *file); | 
|  | 2243 | static const struct file_operations version_fops = { | 
|  | 2244 | .owner = THIS_MODULE, | 
|  | 2245 | .open = sg_proc_single_open_version, | 
|  | 2246 | .read = seq_read, | 
|  | 2247 | .llseek = seq_lseek, | 
|  | 2248 | .release = single_release, | 
|  | 2249 | }; | 
|  | 2250 |  | 
|  | 2251 | static int sg_proc_seq_show_devhdr(struct seq_file *s, void *v); | 
|  | 2252 | static int sg_proc_single_open_devhdr(struct inode *inode, struct file *file); | 
|  | 2253 | static const struct file_operations devhdr_fops = { | 
|  | 2254 | .owner = THIS_MODULE, | 
|  | 2255 | .open = sg_proc_single_open_devhdr, | 
|  | 2256 | .read = seq_read, | 
|  | 2257 | .llseek = seq_lseek, | 
|  | 2258 | .release = single_release, | 
|  | 2259 | }; | 
|  | 2260 |  | 
|  | 2261 | static int sg_proc_seq_show_dev(struct seq_file *s, void *v); | 
|  | 2262 | static int sg_proc_open_dev(struct inode *inode, struct file *file); | 
|  | 2263 | static void * dev_seq_start(struct seq_file *s, loff_t *pos); | 
|  | 2264 | static void * dev_seq_next(struct seq_file *s, void *v, loff_t *pos); | 
|  | 2265 | static void dev_seq_stop(struct seq_file *s, void *v); | 
|  | 2266 | static const struct file_operations dev_fops = { | 
|  | 2267 | .owner = THIS_MODULE, | 
|  | 2268 | .open = sg_proc_open_dev, | 
|  | 2269 | .read = seq_read, | 
|  | 2270 | .llseek = seq_lseek, | 
|  | 2271 | .release = seq_release, | 
|  | 2272 | }; | 
|  | 2273 | static const struct seq_operations dev_seq_ops = { | 
|  | 2274 | .start = dev_seq_start, | 
|  | 2275 | .next  = dev_seq_next, | 
|  | 2276 | .stop  = dev_seq_stop, | 
|  | 2277 | .show  = sg_proc_seq_show_dev, | 
|  | 2278 | }; | 
|  | 2279 |  | 
|  | 2280 | static int sg_proc_seq_show_devstrs(struct seq_file *s, void *v); | 
|  | 2281 | static int sg_proc_open_devstrs(struct inode *inode, struct file *file); | 
|  | 2282 | static const struct file_operations devstrs_fops = { | 
|  | 2283 | .owner = THIS_MODULE, | 
|  | 2284 | .open = sg_proc_open_devstrs, | 
|  | 2285 | .read = seq_read, | 
|  | 2286 | .llseek = seq_lseek, | 
|  | 2287 | .release = seq_release, | 
|  | 2288 | }; | 
|  | 2289 | static const struct seq_operations devstrs_seq_ops = { | 
|  | 2290 | .start = dev_seq_start, | 
|  | 2291 | .next  = dev_seq_next, | 
|  | 2292 | .stop  = dev_seq_stop, | 
|  | 2293 | .show  = sg_proc_seq_show_devstrs, | 
|  | 2294 | }; | 
|  | 2295 |  | 
|  | 2296 | static int sg_proc_seq_show_debug(struct seq_file *s, void *v); | 
|  | 2297 | static int sg_proc_open_debug(struct inode *inode, struct file *file); | 
|  | 2298 | static const struct file_operations debug_fops = { | 
|  | 2299 | .owner = THIS_MODULE, | 
|  | 2300 | .open = sg_proc_open_debug, | 
|  | 2301 | .read = seq_read, | 
|  | 2302 | .llseek = seq_lseek, | 
|  | 2303 | .release = seq_release, | 
|  | 2304 | }; | 
|  | 2305 | static const struct seq_operations debug_seq_ops = { | 
|  | 2306 | .start = dev_seq_start, | 
|  | 2307 | .next  = dev_seq_next, | 
|  | 2308 | .stop  = dev_seq_stop, | 
|  | 2309 | .show  = sg_proc_seq_show_debug, | 
|  | 2310 | }; | 
|  | 2311 |  | 
|  | 2312 |  | 
|  | 2313 | struct sg_proc_leaf { | 
|  | 2314 | const char * name; | 
|  | 2315 | const struct file_operations * fops; | 
|  | 2316 | }; | 
|  | 2317 |  | 
|  | 2318 | static struct sg_proc_leaf sg_proc_leaf_arr[] = { | 
|  | 2319 | {"allow_dio", &adio_fops}, | 
|  | 2320 | {"debug", &debug_fops}, | 
|  | 2321 | {"def_reserved_size", &dressz_fops}, | 
|  | 2322 | {"device_hdr", &devhdr_fops}, | 
|  | 2323 | {"devices", &dev_fops}, | 
|  | 2324 | {"device_strs", &devstrs_fops}, | 
|  | 2325 | {"version", &version_fops} | 
|  | 2326 | }; | 
|  | 2327 |  | 
|  | 2328 | static int | 
|  | 2329 | sg_proc_init(void) | 
|  | 2330 | { | 
|  | 2331 | int num_leaves = ARRAY_SIZE(sg_proc_leaf_arr); | 
|  | 2332 | int k; | 
|  | 2333 |  | 
|  | 2334 | sg_proc_sgp = proc_mkdir(sg_proc_sg_dirname, NULL); | 
|  | 2335 | if (!sg_proc_sgp) | 
|  | 2336 | return 1; | 
|  | 2337 | for (k = 0; k < num_leaves; ++k) { | 
|  | 2338 | struct sg_proc_leaf *leaf = &sg_proc_leaf_arr[k]; | 
|  | 2339 | umode_t mask = leaf->fops->write ? S_IRUGO | S_IWUSR : S_IRUGO; | 
|  | 2340 | proc_create(leaf->name, mask, sg_proc_sgp, leaf->fops); | 
|  | 2341 | } | 
|  | 2342 | return 0; | 
|  | 2343 | } | 
|  | 2344 |  | 
|  | 2345 | static void | 
|  | 2346 | sg_proc_cleanup(void) | 
|  | 2347 | { | 
|  | 2348 | int k; | 
|  | 2349 | int num_leaves = ARRAY_SIZE(sg_proc_leaf_arr); | 
|  | 2350 |  | 
|  | 2351 | if (!sg_proc_sgp) | 
|  | 2352 | return; | 
|  | 2353 | for (k = 0; k < num_leaves; ++k) | 
|  | 2354 | remove_proc_entry(sg_proc_leaf_arr[k].name, sg_proc_sgp); | 
|  | 2355 | remove_proc_entry(sg_proc_sg_dirname, NULL); | 
|  | 2356 | } | 
|  | 2357 |  | 
|  | 2358 |  | 
|  | 2359 | static int sg_proc_seq_show_int(struct seq_file *s, void *v) | 
|  | 2360 | { | 
|  | 2361 | seq_printf(s, "%d\n", *((int *)s->private)); | 
|  | 2362 | return 0; | 
|  | 2363 | } | 
|  | 2364 |  | 
|  | 2365 | static int sg_proc_single_open_adio(struct inode *inode, struct file *file) | 
|  | 2366 | { | 
|  | 2367 | return single_open(file, sg_proc_seq_show_int, &sg_allow_dio); | 
|  | 2368 | } | 
|  | 2369 |  | 
|  | 2370 | static ssize_t | 
|  | 2371 | sg_proc_write_adio(struct file *filp, const char __user *buffer, | 
|  | 2372 | size_t count, loff_t *off) | 
|  | 2373 | { | 
|  | 2374 | int err; | 
|  | 2375 | unsigned long num; | 
|  | 2376 |  | 
|  | 2377 | if (!capable(CAP_SYS_ADMIN) || !capable(CAP_SYS_RAWIO)) | 
|  | 2378 | return -EACCES; | 
|  | 2379 | err = kstrtoul_from_user(buffer, count, 0, &num); | 
|  | 2380 | if (err) | 
|  | 2381 | return err; | 
|  | 2382 | sg_allow_dio = num ? 1 : 0; | 
|  | 2383 | return count; | 
|  | 2384 | } | 
|  | 2385 |  | 
|  | 2386 | static int sg_proc_single_open_dressz(struct inode *inode, struct file *file) | 
|  | 2387 | { | 
|  | 2388 | return single_open(file, sg_proc_seq_show_int, &sg_big_buff); | 
|  | 2389 | } | 
|  | 2390 |  | 
|  | 2391 | static ssize_t | 
|  | 2392 | sg_proc_write_dressz(struct file *filp, const char __user *buffer, | 
|  | 2393 | size_t count, loff_t *off) | 
|  | 2394 | { | 
|  | 2395 | int err; | 
|  | 2396 | unsigned long k = ULONG_MAX; | 
|  | 2397 |  | 
|  | 2398 | if (!capable(CAP_SYS_ADMIN) || !capable(CAP_SYS_RAWIO)) | 
|  | 2399 | return -EACCES; | 
|  | 2400 |  | 
|  | 2401 | err = kstrtoul_from_user(buffer, count, 0, &k); | 
|  | 2402 | if (err) | 
|  | 2403 | return err; | 
|  | 2404 | if (k <= 1048576) {	/* limit "big buff" to 1 MB */ | 
|  | 2405 | sg_big_buff = k; | 
|  | 2406 | return count; | 
|  | 2407 | } | 
|  | 2408 | return -ERANGE; | 
|  | 2409 | } | 
|  | 2410 |  | 
|  | 2411 | static int sg_proc_seq_show_version(struct seq_file *s, void *v) | 
|  | 2412 | { | 
|  | 2413 | seq_printf(s, "%d\t%s [%s]\n", sg_version_num, SG_VERSION_STR, | 
|  | 2414 | sg_version_date); | 
|  | 2415 | return 0; | 
|  | 2416 | } | 
|  | 2417 |  | 
|  | 2418 | static int sg_proc_single_open_version(struct inode *inode, struct file *file) | 
|  | 2419 | { | 
|  | 2420 | return single_open(file, sg_proc_seq_show_version, NULL); | 
|  | 2421 | } | 
|  | 2422 |  | 
|  | 2423 | static int sg_proc_seq_show_devhdr(struct seq_file *s, void *v) | 
|  | 2424 | { | 
|  | 2425 | seq_printf(s, "host\tchan\tid\tlun\ttype\topens\tqdepth\tbusy\t" | 
|  | 2426 | "online\n"); | 
|  | 2427 | return 0; | 
|  | 2428 | } | 
|  | 2429 |  | 
|  | 2430 | static int sg_proc_single_open_devhdr(struct inode *inode, struct file *file) | 
|  | 2431 | { | 
|  | 2432 | return single_open(file, sg_proc_seq_show_devhdr, NULL); | 
|  | 2433 | } | 
|  | 2434 |  | 
|  | 2435 | struct sg_proc_deviter { | 
|  | 2436 | loff_t	index; | 
|  | 2437 | size_t	max; | 
|  | 2438 | }; | 
|  | 2439 |  | 
|  | 2440 | static void * dev_seq_start(struct seq_file *s, loff_t *pos) | 
|  | 2441 | { | 
|  | 2442 | struct sg_proc_deviter * it = kmalloc(sizeof(*it), GFP_KERNEL); | 
|  | 2443 |  | 
|  | 2444 | s->private = it; | 
|  | 2445 | if (! it) | 
|  | 2446 | return NULL; | 
|  | 2447 |  | 
|  | 2448 | it->index = *pos; | 
|  | 2449 | it->max = sg_last_dev(); | 
|  | 2450 | if (it->index >= it->max) | 
|  | 2451 | return NULL; | 
|  | 2452 | return it; | 
|  | 2453 | } | 
|  | 2454 |  | 
|  | 2455 | static void * dev_seq_next(struct seq_file *s, void *v, loff_t *pos) | 
|  | 2456 | { | 
|  | 2457 | struct sg_proc_deviter * it = s->private; | 
|  | 2458 |  | 
|  | 2459 | *pos = ++it->index; | 
|  | 2460 | return (it->index < it->max) ? it : NULL; | 
|  | 2461 | } | 
|  | 2462 |  | 
|  | 2463 | static void dev_seq_stop(struct seq_file *s, void *v) | 
|  | 2464 | { | 
|  | 2465 | kfree(s->private); | 
|  | 2466 | } | 
|  | 2467 |  | 
|  | 2468 | static int sg_proc_open_dev(struct inode *inode, struct file *file) | 
|  | 2469 | { | 
|  | 2470 | return seq_open(file, &dev_seq_ops); | 
|  | 2471 | } | 
|  | 2472 |  | 
|  | 2473 | static int sg_proc_seq_show_dev(struct seq_file *s, void *v) | 
|  | 2474 | { | 
|  | 2475 | struct sg_proc_deviter * it = (struct sg_proc_deviter *) v; | 
|  | 2476 | Sg_device *sdp; | 
|  | 2477 | struct scsi_device *scsidp; | 
|  | 2478 | unsigned long iflags; | 
|  | 2479 |  | 
|  | 2480 | read_lock_irqsave(&sg_index_lock, iflags); | 
|  | 2481 | sdp = it ? sg_lookup_dev(it->index) : NULL; | 
|  | 2482 | if (sdp && (scsidp = sdp->device) && (!sdp->detached)) | 
|  | 2483 | seq_printf(s, "%d\t%d\t%d\t%d\t%d\t%d\t%d\t%d\t%d\n", | 
|  | 2484 | scsidp->host->host_no, scsidp->channel, | 
|  | 2485 | scsidp->id, scsidp->lun, (int) scsidp->type, | 
|  | 2486 | 1, | 
|  | 2487 | (int) scsidp->queue_depth, | 
|  | 2488 | (int) scsidp->device_busy, | 
|  | 2489 | (int) scsi_device_online(scsidp)); | 
|  | 2490 | else | 
|  | 2491 | seq_printf(s, "-1\t-1\t-1\t-1\t-1\t-1\t-1\t-1\t-1\n"); | 
|  | 2492 | read_unlock_irqrestore(&sg_index_lock, iflags); | 
|  | 2493 | return 0; | 
|  | 2494 | } | 
|  | 2495 |  | 
|  | 2496 | static int sg_proc_open_devstrs(struct inode *inode, struct file *file) | 
|  | 2497 | { | 
|  | 2498 | return seq_open(file, &devstrs_seq_ops); | 
|  | 2499 | } | 
|  | 2500 |  | 
|  | 2501 | static int sg_proc_seq_show_devstrs(struct seq_file *s, void *v) | 
|  | 2502 | { | 
|  | 2503 | struct sg_proc_deviter * it = (struct sg_proc_deviter *) v; | 
|  | 2504 | Sg_device *sdp; | 
|  | 2505 | struct scsi_device *scsidp; | 
|  | 2506 | unsigned long iflags; | 
|  | 2507 |  | 
|  | 2508 | read_lock_irqsave(&sg_index_lock, iflags); | 
|  | 2509 | sdp = it ? sg_lookup_dev(it->index) : NULL; | 
|  | 2510 | if (sdp && (scsidp = sdp->device) && (!sdp->detached)) | 
|  | 2511 | seq_printf(s, "%8.8s\t%16.16s\t%4.4s\n", | 
|  | 2512 | scsidp->vendor, scsidp->model, scsidp->rev); | 
|  | 2513 | else | 
|  | 2514 | seq_printf(s, "<no active device>\n"); | 
|  | 2515 | read_unlock_irqrestore(&sg_index_lock, iflags); | 
|  | 2516 | return 0; | 
|  | 2517 | } | 
|  | 2518 |  | 
|  | 2519 | /* must be called while holding sg_index_lock */ | 
|  | 2520 | static void sg_proc_debug_helper(struct seq_file *s, Sg_device * sdp) | 
|  | 2521 | { | 
|  | 2522 | int k, m, new_interface, blen, usg; | 
|  | 2523 | Sg_request *srp; | 
|  | 2524 | Sg_fd *fp; | 
|  | 2525 | const sg_io_hdr_t *hp; | 
|  | 2526 | const char * cp; | 
|  | 2527 | unsigned int ms; | 
|  | 2528 |  | 
|  | 2529 | k = 0; | 
|  | 2530 | list_for_each_entry(fp, &sdp->sfds, sfd_siblings) { | 
|  | 2531 | k++; | 
|  | 2532 | read_lock(&fp->rq_list_lock); /* irqs already disabled */ | 
|  | 2533 | seq_printf(s, "   FD(%d): timeout=%dms bufflen=%d " | 
|  | 2534 | "(res)sgat=%d low_dma=%d\n", k, | 
|  | 2535 | jiffies_to_msecs(fp->timeout), | 
|  | 2536 | fp->reserve.bufflen, | 
|  | 2537 | (int) fp->reserve.k_use_sg, | 
|  | 2538 | (int) fp->low_dma); | 
|  | 2539 | seq_printf(s, "   cmd_q=%d f_packid=%d k_orphan=%d closed=%d\n", | 
|  | 2540 | (int) fp->cmd_q, (int) fp->force_packid, | 
|  | 2541 | (int) fp->keep_orphan, (int) fp->closed); | 
|  | 2542 | for (m = 0, srp = fp->headrp; | 
|  | 2543 | srp != NULL; | 
|  | 2544 | ++m, srp = srp->nextrp) { | 
|  | 2545 | hp = &srp->header; | 
|  | 2546 | new_interface = (hp->interface_id == '\0') ? 0 : 1; | 
|  | 2547 | if (srp->res_used) { | 
|  | 2548 | if (new_interface && | 
|  | 2549 | (SG_FLAG_MMAP_IO & hp->flags)) | 
|  | 2550 | cp = "     mmap>> "; | 
|  | 2551 | else | 
|  | 2552 | cp = "     rb>> "; | 
|  | 2553 | } else { | 
|  | 2554 | if (SG_INFO_DIRECT_IO_MASK & hp->info) | 
|  | 2555 | cp = "     dio>> "; | 
|  | 2556 | else | 
|  | 2557 | cp = "     "; | 
|  | 2558 | } | 
|  | 2559 | seq_printf(s, cp); | 
|  | 2560 | blen = srp->data.bufflen; | 
|  | 2561 | usg = srp->data.k_use_sg; | 
|  | 2562 | seq_printf(s, srp->done ? | 
|  | 2563 | ((1 == srp->done) ?  "rcv:" : "fin:") | 
|  | 2564 | : "act:"); | 
|  | 2565 | seq_printf(s, " id=%d blen=%d", | 
|  | 2566 | srp->header.pack_id, blen); | 
|  | 2567 | if (srp->done) | 
|  | 2568 | seq_printf(s, " dur=%d", hp->duration); | 
|  | 2569 | else { | 
|  | 2570 | ms = jiffies_to_msecs(jiffies); | 
|  | 2571 | seq_printf(s, " t_o/elap=%d/%d", | 
|  | 2572 | (new_interface ? hp->timeout : | 
|  | 2573 | jiffies_to_msecs(fp->timeout)), | 
|  | 2574 | (ms > hp->duration ? ms - hp->duration : 0)); | 
|  | 2575 | } | 
|  | 2576 | seq_printf(s, "ms sgat=%d op=0x%02x\n", usg, | 
|  | 2577 | (int) srp->data.cmd_opcode); | 
|  | 2578 | } | 
|  | 2579 | if (0 == m) | 
|  | 2580 | seq_printf(s, "     No requests active\n"); | 
|  | 2581 | read_unlock(&fp->rq_list_lock); | 
|  | 2582 | } | 
|  | 2583 | } | 
|  | 2584 |  | 
|  | 2585 | static int sg_proc_open_debug(struct inode *inode, struct file *file) | 
|  | 2586 | { | 
|  | 2587 | return seq_open(file, &debug_seq_ops); | 
|  | 2588 | } | 
|  | 2589 |  | 
|  | 2590 | static int sg_proc_seq_show_debug(struct seq_file *s, void *v) | 
|  | 2591 | { | 
|  | 2592 | struct sg_proc_deviter * it = (struct sg_proc_deviter *) v; | 
|  | 2593 | Sg_device *sdp; | 
|  | 2594 | unsigned long iflags; | 
|  | 2595 |  | 
|  | 2596 | if (it && (0 == it->index)) { | 
|  | 2597 | seq_printf(s, "max_active_device=%d(origin 1)\n", | 
|  | 2598 | (int)it->max); | 
|  | 2599 | seq_printf(s, " def_reserved_size=%d\n", sg_big_buff); | 
|  | 2600 | } | 
|  | 2601 |  | 
|  | 2602 | read_lock_irqsave(&sg_index_lock, iflags); | 
|  | 2603 | sdp = it ? sg_lookup_dev(it->index) : NULL; | 
|  | 2604 | if (sdp && !list_empty(&sdp->sfds)) { | 
|  | 2605 | struct scsi_device *scsidp = sdp->device; | 
|  | 2606 |  | 
|  | 2607 | seq_printf(s, " >>> device=%s ", sdp->disk->disk_name); | 
|  | 2608 | if (sdp->detached) | 
|  | 2609 | seq_printf(s, "detached pending close "); | 
|  | 2610 | else | 
|  | 2611 | seq_printf | 
|  | 2612 | (s, "scsi%d chan=%d id=%d lun=%d   em=%d", | 
|  | 2613 | scsidp->host->host_no, | 
|  | 2614 | scsidp->channel, scsidp->id, | 
|  | 2615 | scsidp->lun, | 
|  | 2616 | scsidp->host->hostt->emulated); | 
|  | 2617 | seq_printf(s, " sg_tablesize=%d excl=%d\n", | 
|  | 2618 | sdp->sg_tablesize, sdp->exclude); | 
|  | 2619 | sg_proc_debug_helper(s, sdp); | 
|  | 2620 | } | 
|  | 2621 | read_unlock_irqrestore(&sg_index_lock, iflags); | 
|  | 2622 | return 0; | 
|  | 2623 | } | 
|  | 2624 |  | 
|  | 2625 | #endif				/* CONFIG_SCSI_PROC_FS */ | 
|  | 2626 |  | 
|  | 2627 | module_init(init_sg); | 
|  | 2628 | module_exit(exit_sg); |