Linux kernel mirror (for testing)
git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel
os
linux
fork
Configure Feed
Select the types of activity you want to include in your feed.
1/*
2 * scsi_sysfs.c
3 *
4 * SCSI sysfs interface routines.
5 *
6 * Created to pull SCSI mid layer sysfs routines into one file.
7 */
8
9#include <linux/module.h>
10#include <linux/slab.h>
11#include <linux/init.h>
12#include <linux/blkdev.h>
13#include <linux/device.h>
14#include <linux/pm_runtime.h>
15
16#include <scsi/scsi.h>
17#include <scsi/scsi_device.h>
18#include <scsi/scsi_host.h>
19#include <scsi/scsi_tcq.h>
20#include <scsi/scsi_transport.h>
21#include <scsi/scsi_driver.h>
22
23#include "scsi_priv.h"
24#include "scsi_logging.h"
25
26static struct device_type scsi_dev_type;
27
28static const struct {
29 enum scsi_device_state value;
30 char *name;
31} sdev_states[] = {
32 { SDEV_CREATED, "created" },
33 { SDEV_RUNNING, "running" },
34 { SDEV_CANCEL, "cancel" },
35 { SDEV_DEL, "deleted" },
36 { SDEV_QUIESCE, "quiesce" },
37 { SDEV_OFFLINE, "offline" },
38 { SDEV_BLOCK, "blocked" },
39 { SDEV_CREATED_BLOCK, "created-blocked" },
40};
41
42const char *scsi_device_state_name(enum scsi_device_state state)
43{
44 int i;
45 char *name = NULL;
46
47 for (i = 0; i < ARRAY_SIZE(sdev_states); i++) {
48 if (sdev_states[i].value == state) {
49 name = sdev_states[i].name;
50 break;
51 }
52 }
53 return name;
54}
55
56static const struct {
57 enum scsi_host_state value;
58 char *name;
59} shost_states[] = {
60 { SHOST_CREATED, "created" },
61 { SHOST_RUNNING, "running" },
62 { SHOST_CANCEL, "cancel" },
63 { SHOST_DEL, "deleted" },
64 { SHOST_RECOVERY, "recovery" },
65 { SHOST_CANCEL_RECOVERY, "cancel/recovery" },
66 { SHOST_DEL_RECOVERY, "deleted/recovery", },
67};
68const char *scsi_host_state_name(enum scsi_host_state state)
69{
70 int i;
71 char *name = NULL;
72
73 for (i = 0; i < ARRAY_SIZE(shost_states); i++) {
74 if (shost_states[i].value == state) {
75 name = shost_states[i].name;
76 break;
77 }
78 }
79 return name;
80}
81
82static int check_set(unsigned int *val, char *src)
83{
84 char *last;
85
86 if (strncmp(src, "-", 20) == 0) {
87 *val = SCAN_WILD_CARD;
88 } else {
89 /*
90 * Doesn't check for int overflow
91 */
92 *val = simple_strtoul(src, &last, 0);
93 if (*last != '\0')
94 return 1;
95 }
96 return 0;
97}
98
99static int scsi_scan(struct Scsi_Host *shost, const char *str)
100{
101 char s1[15], s2[15], s3[15], junk;
102 unsigned int channel, id, lun;
103 int res;
104
105 res = sscanf(str, "%10s %10s %10s %c", s1, s2, s3, &junk);
106 if (res != 3)
107 return -EINVAL;
108 if (check_set(&channel, s1))
109 return -EINVAL;
110 if (check_set(&id, s2))
111 return -EINVAL;
112 if (check_set(&lun, s3))
113 return -EINVAL;
114 if (shost->transportt->user_scan)
115 res = shost->transportt->user_scan(shost, channel, id, lun);
116 else
117 res = scsi_scan_host_selected(shost, channel, id, lun, 1);
118 return res;
119}
120
121/*
122 * shost_show_function: macro to create an attr function that can be used to
123 * show a non-bit field.
124 */
125#define shost_show_function(name, field, format_string) \
126static ssize_t \
127show_##name (struct device *dev, struct device_attribute *attr, \
128 char *buf) \
129{ \
130 struct Scsi_Host *shost = class_to_shost(dev); \
131 return snprintf (buf, 20, format_string, shost->field); \
132}
133
134/*
135 * shost_rd_attr: macro to create a function and attribute variable for a
136 * read only field.
137 */
138#define shost_rd_attr2(name, field, format_string) \
139 shost_show_function(name, field, format_string) \
140static DEVICE_ATTR(name, S_IRUGO, show_##name, NULL);
141
142#define shost_rd_attr(field, format_string) \
143shost_rd_attr2(field, field, format_string)
144
145/*
146 * Create the actual show/store functions and data structures.
147 */
148
149static ssize_t
150store_scan(struct device *dev, struct device_attribute *attr,
151 const char *buf, size_t count)
152{
153 struct Scsi_Host *shost = class_to_shost(dev);
154 int res;
155
156 res = scsi_scan(shost, buf);
157 if (res == 0)
158 res = count;
159 return res;
160};
161static DEVICE_ATTR(scan, S_IWUSR, NULL, store_scan);
162
163static ssize_t
164store_shost_state(struct device *dev, struct device_attribute *attr,
165 const char *buf, size_t count)
166{
167 int i;
168 struct Scsi_Host *shost = class_to_shost(dev);
169 enum scsi_host_state state = 0;
170
171 for (i = 0; i < ARRAY_SIZE(shost_states); i++) {
172 const int len = strlen(shost_states[i].name);
173 if (strncmp(shost_states[i].name, buf, len) == 0 &&
174 buf[len] == '\n') {
175 state = shost_states[i].value;
176 break;
177 }
178 }
179 if (!state)
180 return -EINVAL;
181
182 if (scsi_host_set_state(shost, state))
183 return -EINVAL;
184 return count;
185}
186
187static ssize_t
188show_shost_state(struct device *dev, struct device_attribute *attr, char *buf)
189{
190 struct Scsi_Host *shost = class_to_shost(dev);
191 const char *name = scsi_host_state_name(shost->shost_state);
192
193 if (!name)
194 return -EINVAL;
195
196 return snprintf(buf, 20, "%s\n", name);
197}
198
199/* DEVICE_ATTR(state) clashes with dev_attr_state for sdev */
200struct device_attribute dev_attr_hstate =
201 __ATTR(state, S_IRUGO | S_IWUSR, show_shost_state, store_shost_state);
202
203static ssize_t
204show_shost_mode(unsigned int mode, char *buf)
205{
206 ssize_t len = 0;
207
208 if (mode & MODE_INITIATOR)
209 len = sprintf(buf, "%s", "Initiator");
210
211 if (mode & MODE_TARGET)
212 len += sprintf(buf + len, "%s%s", len ? ", " : "", "Target");
213
214 len += sprintf(buf + len, "\n");
215
216 return len;
217}
218
219static ssize_t
220show_shost_supported_mode(struct device *dev, struct device_attribute *attr,
221 char *buf)
222{
223 struct Scsi_Host *shost = class_to_shost(dev);
224 unsigned int supported_mode = shost->hostt->supported_mode;
225
226 if (supported_mode == MODE_UNKNOWN)
227 /* by default this should be initiator */
228 supported_mode = MODE_INITIATOR;
229
230 return show_shost_mode(supported_mode, buf);
231}
232
233static DEVICE_ATTR(supported_mode, S_IRUGO | S_IWUSR, show_shost_supported_mode, NULL);
234
235static ssize_t
236show_shost_active_mode(struct device *dev,
237 struct device_attribute *attr, char *buf)
238{
239 struct Scsi_Host *shost = class_to_shost(dev);
240
241 if (shost->active_mode == MODE_UNKNOWN)
242 return snprintf(buf, 20, "unknown\n");
243 else
244 return show_shost_mode(shost->active_mode, buf);
245}
246
247static DEVICE_ATTR(active_mode, S_IRUGO | S_IWUSR, show_shost_active_mode, NULL);
248
249shost_rd_attr(unique_id, "%u\n");
250shost_rd_attr(host_busy, "%hu\n");
251shost_rd_attr(cmd_per_lun, "%hd\n");
252shost_rd_attr(can_queue, "%hd\n");
253shost_rd_attr(sg_tablesize, "%hu\n");
254shost_rd_attr(unchecked_isa_dma, "%d\n");
255shost_rd_attr(prot_capabilities, "%u\n");
256shost_rd_attr(prot_guard_type, "%hd\n");
257shost_rd_attr2(proc_name, hostt->proc_name, "%s\n");
258
259static struct attribute *scsi_sysfs_shost_attrs[] = {
260 &dev_attr_unique_id.attr,
261 &dev_attr_host_busy.attr,
262 &dev_attr_cmd_per_lun.attr,
263 &dev_attr_can_queue.attr,
264 &dev_attr_sg_tablesize.attr,
265 &dev_attr_unchecked_isa_dma.attr,
266 &dev_attr_proc_name.attr,
267 &dev_attr_scan.attr,
268 &dev_attr_hstate.attr,
269 &dev_attr_supported_mode.attr,
270 &dev_attr_active_mode.attr,
271 &dev_attr_prot_capabilities.attr,
272 &dev_attr_prot_guard_type.attr,
273 NULL
274};
275
276struct attribute_group scsi_shost_attr_group = {
277 .attrs = scsi_sysfs_shost_attrs,
278};
279
280const struct attribute_group *scsi_sysfs_shost_attr_groups[] = {
281 &scsi_shost_attr_group,
282 NULL
283};
284
285static void scsi_device_cls_release(struct device *class_dev)
286{
287 struct scsi_device *sdev;
288
289 sdev = class_to_sdev(class_dev);
290 put_device(&sdev->sdev_gendev);
291}
292
293static void scsi_device_dev_release_usercontext(struct work_struct *work)
294{
295 struct scsi_device *sdev;
296 struct device *parent;
297 struct scsi_target *starget;
298 struct list_head *this, *tmp;
299 unsigned long flags;
300
301 sdev = container_of(work, struct scsi_device, ew.work);
302
303 parent = sdev->sdev_gendev.parent;
304 starget = to_scsi_target(parent);
305
306 spin_lock_irqsave(sdev->host->host_lock, flags);
307 starget->reap_ref++;
308 list_del(&sdev->siblings);
309 list_del(&sdev->same_target_siblings);
310 list_del(&sdev->starved_entry);
311 spin_unlock_irqrestore(sdev->host->host_lock, flags);
312
313 cancel_work_sync(&sdev->event_work);
314
315 list_for_each_safe(this, tmp, &sdev->event_list) {
316 struct scsi_event *evt;
317
318 evt = list_entry(this, struct scsi_event, node);
319 list_del(&evt->node);
320 kfree(evt);
321 }
322
323 if (sdev->request_queue) {
324 sdev->request_queue->queuedata = NULL;
325 /* user context needed to free queue */
326 scsi_free_queue(sdev->request_queue);
327 /* temporary expedient, try to catch use of queue lock
328 * after free of sdev */
329 sdev->request_queue = NULL;
330 }
331
332 scsi_target_reap(scsi_target(sdev));
333
334 kfree(sdev->inquiry);
335 kfree(sdev);
336
337 if (parent)
338 put_device(parent);
339}
340
341static void scsi_device_dev_release(struct device *dev)
342{
343 struct scsi_device *sdp = to_scsi_device(dev);
344 execute_in_process_context(scsi_device_dev_release_usercontext,
345 &sdp->ew);
346}
347
348static struct class sdev_class = {
349 .name = "scsi_device",
350 .dev_release = scsi_device_cls_release,
351};
352
353/* all probing is done in the individual ->probe routines */
354static int scsi_bus_match(struct device *dev, struct device_driver *gendrv)
355{
356 struct scsi_device *sdp;
357
358 if (dev->type != &scsi_dev_type)
359 return 0;
360
361 sdp = to_scsi_device(dev);
362 if (sdp->no_uld_attach)
363 return 0;
364 return (sdp->inq_periph_qual == SCSI_INQ_PQ_CON)? 1: 0;
365}
366
367static int scsi_bus_uevent(struct device *dev, struct kobj_uevent_env *env)
368{
369 struct scsi_device *sdev;
370
371 if (dev->type != &scsi_dev_type)
372 return 0;
373
374 sdev = to_scsi_device(dev);
375
376 add_uevent_var(env, "MODALIAS=" SCSI_DEVICE_MODALIAS_FMT, sdev->type);
377 return 0;
378}
379
380struct bus_type scsi_bus_type = {
381 .name = "scsi",
382 .match = scsi_bus_match,
383 .uevent = scsi_bus_uevent,
384#ifdef CONFIG_PM_OPS
385 .pm = &scsi_bus_pm_ops,
386#endif
387};
388EXPORT_SYMBOL_GPL(scsi_bus_type);
389
390int scsi_sysfs_register(void)
391{
392 int error;
393
394 error = bus_register(&scsi_bus_type);
395 if (!error) {
396 error = class_register(&sdev_class);
397 if (error)
398 bus_unregister(&scsi_bus_type);
399 }
400
401 return error;
402}
403
404void scsi_sysfs_unregister(void)
405{
406 class_unregister(&sdev_class);
407 bus_unregister(&scsi_bus_type);
408}
409
410/*
411 * sdev_show_function: macro to create an attr function that can be used to
412 * show a non-bit field.
413 */
414#define sdev_show_function(field, format_string) \
415static ssize_t \
416sdev_show_##field (struct device *dev, struct device_attribute *attr, \
417 char *buf) \
418{ \
419 struct scsi_device *sdev; \
420 sdev = to_scsi_device(dev); \
421 return snprintf (buf, 20, format_string, sdev->field); \
422} \
423
424/*
425 * sdev_rd_attr: macro to create a function and attribute variable for a
426 * read only field.
427 */
428#define sdev_rd_attr(field, format_string) \
429 sdev_show_function(field, format_string) \
430static DEVICE_ATTR(field, S_IRUGO, sdev_show_##field, NULL);
431
432
433/*
434 * sdev_rw_attr: create a function and attribute variable for a
435 * read/write field.
436 */
437#define sdev_rw_attr(field, format_string) \
438 sdev_show_function(field, format_string) \
439 \
440static ssize_t \
441sdev_store_##field (struct device *dev, struct device_attribute *attr, \
442 const char *buf, size_t count) \
443{ \
444 struct scsi_device *sdev; \
445 sdev = to_scsi_device(dev); \
446 sscanf (buf, format_string, &sdev->field); \
447 return count; \
448} \
449static DEVICE_ATTR(field, S_IRUGO | S_IWUSR, sdev_show_##field, sdev_store_##field);
450
451/* Currently we don't export bit fields, but we might in future,
452 * so leave this code in */
453#if 0
454/*
455 * sdev_rd_attr: create a function and attribute variable for a
456 * read/write bit field.
457 */
458#define sdev_rw_attr_bit(field) \
459 sdev_show_function(field, "%d\n") \
460 \
461static ssize_t \
462sdev_store_##field (struct device *dev, struct device_attribute *attr, \
463 const char *buf, size_t count) \
464{ \
465 int ret; \
466 struct scsi_device *sdev; \
467 ret = scsi_sdev_check_buf_bit(buf); \
468 if (ret >= 0) { \
469 sdev = to_scsi_device(dev); \
470 sdev->field = ret; \
471 ret = count; \
472 } \
473 return ret; \
474} \
475static DEVICE_ATTR(field, S_IRUGO | S_IWUSR, sdev_show_##field, sdev_store_##field);
476
477/*
478 * scsi_sdev_check_buf_bit: return 0 if buf is "0", return 1 if buf is "1",
479 * else return -EINVAL.
480 */
481static int scsi_sdev_check_buf_bit(const char *buf)
482{
483 if ((buf[1] == '\0') || ((buf[1] == '\n') && (buf[2] == '\0'))) {
484 if (buf[0] == '1')
485 return 1;
486 else if (buf[0] == '0')
487 return 0;
488 else
489 return -EINVAL;
490 } else
491 return -EINVAL;
492}
493#endif
494/*
495 * Create the actual show/store functions and data structures.
496 */
497sdev_rd_attr (device_blocked, "%d\n");
498sdev_rd_attr (queue_depth, "%d\n");
499sdev_rd_attr (type, "%d\n");
500sdev_rd_attr (scsi_level, "%d\n");
501sdev_rd_attr (vendor, "%.8s\n");
502sdev_rd_attr (model, "%.16s\n");
503sdev_rd_attr (rev, "%.4s\n");
504
505/*
506 * TODO: can we make these symlinks to the block layer ones?
507 */
508static ssize_t
509sdev_show_timeout (struct device *dev, struct device_attribute *attr, char *buf)
510{
511 struct scsi_device *sdev;
512 sdev = to_scsi_device(dev);
513 return snprintf(buf, 20, "%d\n", sdev->request_queue->rq_timeout / HZ);
514}
515
516static ssize_t
517sdev_store_timeout (struct device *dev, struct device_attribute *attr,
518 const char *buf, size_t count)
519{
520 struct scsi_device *sdev;
521 int timeout;
522 sdev = to_scsi_device(dev);
523 sscanf (buf, "%d\n", &timeout);
524 blk_queue_rq_timeout(sdev->request_queue, timeout * HZ);
525 return count;
526}
527static DEVICE_ATTR(timeout, S_IRUGO | S_IWUSR, sdev_show_timeout, sdev_store_timeout);
528
529static ssize_t
530store_rescan_field (struct device *dev, struct device_attribute *attr,
531 const char *buf, size_t count)
532{
533 scsi_rescan_device(dev);
534 return count;
535}
536static DEVICE_ATTR(rescan, S_IWUSR, NULL, store_rescan_field);
537
538static void sdev_store_delete_callback(struct device *dev)
539{
540 scsi_remove_device(to_scsi_device(dev));
541}
542
543static ssize_t
544sdev_store_delete(struct device *dev, struct device_attribute *attr,
545 const char *buf, size_t count)
546{
547 int rc;
548
549 /* An attribute cannot be unregistered by one of its own methods,
550 * so we have to use this roundabout approach.
551 */
552 rc = device_schedule_callback(dev, sdev_store_delete_callback);
553 if (rc)
554 count = rc;
555 return count;
556};
557static DEVICE_ATTR(delete, S_IWUSR, NULL, sdev_store_delete);
558
559static ssize_t
560store_state_field(struct device *dev, struct device_attribute *attr,
561 const char *buf, size_t count)
562{
563 int i;
564 struct scsi_device *sdev = to_scsi_device(dev);
565 enum scsi_device_state state = 0;
566
567 for (i = 0; i < ARRAY_SIZE(sdev_states); i++) {
568 const int len = strlen(sdev_states[i].name);
569 if (strncmp(sdev_states[i].name, buf, len) == 0 &&
570 buf[len] == '\n') {
571 state = sdev_states[i].value;
572 break;
573 }
574 }
575 if (!state)
576 return -EINVAL;
577
578 if (scsi_device_set_state(sdev, state))
579 return -EINVAL;
580 return count;
581}
582
583static ssize_t
584show_state_field(struct device *dev, struct device_attribute *attr, char *buf)
585{
586 struct scsi_device *sdev = to_scsi_device(dev);
587 const char *name = scsi_device_state_name(sdev->sdev_state);
588
589 if (!name)
590 return -EINVAL;
591
592 return snprintf(buf, 20, "%s\n", name);
593}
594
595static DEVICE_ATTR(state, S_IRUGO | S_IWUSR, show_state_field, store_state_field);
596
597static ssize_t
598show_queue_type_field(struct device *dev, struct device_attribute *attr,
599 char *buf)
600{
601 struct scsi_device *sdev = to_scsi_device(dev);
602 const char *name = "none";
603
604 if (sdev->ordered_tags)
605 name = "ordered";
606 else if (sdev->simple_tags)
607 name = "simple";
608
609 return snprintf(buf, 20, "%s\n", name);
610}
611
612static DEVICE_ATTR(queue_type, S_IRUGO, show_queue_type_field, NULL);
613
614static ssize_t
615show_iostat_counterbits(struct device *dev, struct device_attribute *attr, char *buf)
616{
617 return snprintf(buf, 20, "%d\n", (int)sizeof(atomic_t) * 8);
618}
619
620static DEVICE_ATTR(iocounterbits, S_IRUGO, show_iostat_counterbits, NULL);
621
622#define show_sdev_iostat(field) \
623static ssize_t \
624show_iostat_##field(struct device *dev, struct device_attribute *attr, \
625 char *buf) \
626{ \
627 struct scsi_device *sdev = to_scsi_device(dev); \
628 unsigned long long count = atomic_read(&sdev->field); \
629 return snprintf(buf, 20, "0x%llx\n", count); \
630} \
631static DEVICE_ATTR(field, S_IRUGO, show_iostat_##field, NULL)
632
633show_sdev_iostat(iorequest_cnt);
634show_sdev_iostat(iodone_cnt);
635show_sdev_iostat(ioerr_cnt);
636
637static ssize_t
638sdev_show_modalias(struct device *dev, struct device_attribute *attr, char *buf)
639{
640 struct scsi_device *sdev;
641 sdev = to_scsi_device(dev);
642 return snprintf (buf, 20, SCSI_DEVICE_MODALIAS_FMT "\n", sdev->type);
643}
644static DEVICE_ATTR(modalias, S_IRUGO, sdev_show_modalias, NULL);
645
646#define DECLARE_EVT_SHOW(name, Cap_name) \
647static ssize_t \
648sdev_show_evt_##name(struct device *dev, struct device_attribute *attr, \
649 char *buf) \
650{ \
651 struct scsi_device *sdev = to_scsi_device(dev); \
652 int val = test_bit(SDEV_EVT_##Cap_name, sdev->supported_events);\
653 return snprintf(buf, 20, "%d\n", val); \
654}
655
656#define DECLARE_EVT_STORE(name, Cap_name) \
657static ssize_t \
658sdev_store_evt_##name(struct device *dev, struct device_attribute *attr,\
659 const char *buf, size_t count) \
660{ \
661 struct scsi_device *sdev = to_scsi_device(dev); \
662 int val = simple_strtoul(buf, NULL, 0); \
663 if (val == 0) \
664 clear_bit(SDEV_EVT_##Cap_name, sdev->supported_events); \
665 else if (val == 1) \
666 set_bit(SDEV_EVT_##Cap_name, sdev->supported_events); \
667 else \
668 return -EINVAL; \
669 return count; \
670}
671
672#define DECLARE_EVT(name, Cap_name) \
673 DECLARE_EVT_SHOW(name, Cap_name) \
674 DECLARE_EVT_STORE(name, Cap_name) \
675 static DEVICE_ATTR(evt_##name, S_IRUGO, sdev_show_evt_##name, \
676 sdev_store_evt_##name);
677#define REF_EVT(name) &dev_attr_evt_##name.attr
678
679DECLARE_EVT(media_change, MEDIA_CHANGE)
680
681/* Default template for device attributes. May NOT be modified */
682static struct attribute *scsi_sdev_attrs[] = {
683 &dev_attr_device_blocked.attr,
684 &dev_attr_type.attr,
685 &dev_attr_scsi_level.attr,
686 &dev_attr_vendor.attr,
687 &dev_attr_model.attr,
688 &dev_attr_rev.attr,
689 &dev_attr_rescan.attr,
690 &dev_attr_delete.attr,
691 &dev_attr_state.attr,
692 &dev_attr_timeout.attr,
693 &dev_attr_iocounterbits.attr,
694 &dev_attr_iorequest_cnt.attr,
695 &dev_attr_iodone_cnt.attr,
696 &dev_attr_ioerr_cnt.attr,
697 &dev_attr_modalias.attr,
698 REF_EVT(media_change),
699 NULL
700};
701
702static struct attribute_group scsi_sdev_attr_group = {
703 .attrs = scsi_sdev_attrs,
704};
705
706static const struct attribute_group *scsi_sdev_attr_groups[] = {
707 &scsi_sdev_attr_group,
708 NULL
709};
710
711static ssize_t
712sdev_store_queue_depth_rw(struct device *dev, struct device_attribute *attr,
713 const char *buf, size_t count)
714{
715 int depth, retval;
716 struct scsi_device *sdev = to_scsi_device(dev);
717 struct scsi_host_template *sht = sdev->host->hostt;
718
719 if (!sht->change_queue_depth)
720 return -EINVAL;
721
722 depth = simple_strtoul(buf, NULL, 0);
723
724 if (depth < 1)
725 return -EINVAL;
726
727 retval = sht->change_queue_depth(sdev, depth,
728 SCSI_QDEPTH_DEFAULT);
729 if (retval < 0)
730 return retval;
731
732 sdev->max_queue_depth = sdev->queue_depth;
733
734 return count;
735}
736
737static struct device_attribute sdev_attr_queue_depth_rw =
738 __ATTR(queue_depth, S_IRUGO | S_IWUSR, sdev_show_queue_depth,
739 sdev_store_queue_depth_rw);
740
741static ssize_t
742sdev_show_queue_ramp_up_period(struct device *dev,
743 struct device_attribute *attr,
744 char *buf)
745{
746 struct scsi_device *sdev;
747 sdev = to_scsi_device(dev);
748 return snprintf(buf, 20, "%u\n",
749 jiffies_to_msecs(sdev->queue_ramp_up_period));
750}
751
752static ssize_t
753sdev_store_queue_ramp_up_period(struct device *dev,
754 struct device_attribute *attr,
755 const char *buf, size_t count)
756{
757 struct scsi_device *sdev = to_scsi_device(dev);
758 unsigned long period;
759
760 if (strict_strtoul(buf, 10, &period))
761 return -EINVAL;
762
763 sdev->queue_ramp_up_period = msecs_to_jiffies(period);
764 return period;
765}
766
767static struct device_attribute sdev_attr_queue_ramp_up_period =
768 __ATTR(queue_ramp_up_period, S_IRUGO | S_IWUSR,
769 sdev_show_queue_ramp_up_period,
770 sdev_store_queue_ramp_up_period);
771
772static ssize_t
773sdev_store_queue_type_rw(struct device *dev, struct device_attribute *attr,
774 const char *buf, size_t count)
775{
776 struct scsi_device *sdev = to_scsi_device(dev);
777 struct scsi_host_template *sht = sdev->host->hostt;
778 int tag_type = 0, retval;
779 int prev_tag_type = scsi_get_tag_type(sdev);
780
781 if (!sdev->tagged_supported || !sht->change_queue_type)
782 return -EINVAL;
783
784 if (strncmp(buf, "ordered", 7) == 0)
785 tag_type = MSG_ORDERED_TAG;
786 else if (strncmp(buf, "simple", 6) == 0)
787 tag_type = MSG_SIMPLE_TAG;
788 else if (strncmp(buf, "none", 4) != 0)
789 return -EINVAL;
790
791 if (tag_type == prev_tag_type)
792 return count;
793
794 retval = sht->change_queue_type(sdev, tag_type);
795 if (retval < 0)
796 return retval;
797
798 return count;
799}
800
801static int scsi_target_add(struct scsi_target *starget)
802{
803 int error;
804
805 if (starget->state != STARGET_CREATED)
806 return 0;
807
808 error = device_add(&starget->dev);
809 if (error) {
810 dev_err(&starget->dev, "target device_add failed, error %d\n", error);
811 return error;
812 }
813 transport_add_device(&starget->dev);
814 starget->state = STARGET_RUNNING;
815
816 pm_runtime_set_active(&starget->dev);
817 pm_runtime_enable(&starget->dev);
818 device_enable_async_suspend(&starget->dev);
819
820 return 0;
821}
822
823static struct device_attribute sdev_attr_queue_type_rw =
824 __ATTR(queue_type, S_IRUGO | S_IWUSR, show_queue_type_field,
825 sdev_store_queue_type_rw);
826
827/**
828 * scsi_sysfs_add_sdev - add scsi device to sysfs
829 * @sdev: scsi_device to add
830 *
831 * Return value:
832 * 0 on Success / non-zero on Failure
833 **/
834int scsi_sysfs_add_sdev(struct scsi_device *sdev)
835{
836 int error, i;
837 struct request_queue *rq = sdev->request_queue;
838 struct scsi_target *starget = sdev->sdev_target;
839
840 error = scsi_device_set_state(sdev, SDEV_RUNNING);
841 if (error)
842 return error;
843
844 error = scsi_target_add(starget);
845 if (error)
846 return error;
847
848 transport_configure_device(&starget->dev);
849
850 device_enable_async_suspend(&sdev->sdev_gendev);
851 scsi_autopm_get_target(starget);
852 pm_runtime_set_active(&sdev->sdev_gendev);
853 pm_runtime_forbid(&sdev->sdev_gendev);
854 pm_runtime_enable(&sdev->sdev_gendev);
855 scsi_autopm_put_target(starget);
856
857 /* The following call will keep sdev active indefinitely, until
858 * its driver does a corresponding scsi_autopm_pm_device(). Only
859 * drivers supporting autosuspend will do this.
860 */
861 scsi_autopm_get_device(sdev);
862
863 error = device_add(&sdev->sdev_gendev);
864 if (error) {
865 printk(KERN_INFO "error 1\n");
866 return error;
867 }
868 device_enable_async_suspend(&sdev->sdev_dev);
869 error = device_add(&sdev->sdev_dev);
870 if (error) {
871 printk(KERN_INFO "error 2\n");
872 device_del(&sdev->sdev_gendev);
873 return error;
874 }
875 transport_add_device(&sdev->sdev_gendev);
876 sdev->is_visible = 1;
877
878 /* create queue files, which may be writable, depending on the host */
879 if (sdev->host->hostt->change_queue_depth) {
880 error = device_create_file(&sdev->sdev_gendev,
881 &sdev_attr_queue_depth_rw);
882 error = device_create_file(&sdev->sdev_gendev,
883 &sdev_attr_queue_ramp_up_period);
884 }
885 else
886 error = device_create_file(&sdev->sdev_gendev, &dev_attr_queue_depth);
887 if (error)
888 return error;
889
890 if (sdev->host->hostt->change_queue_type)
891 error = device_create_file(&sdev->sdev_gendev, &sdev_attr_queue_type_rw);
892 else
893 error = device_create_file(&sdev->sdev_gendev, &dev_attr_queue_type);
894 if (error)
895 return error;
896
897 error = bsg_register_queue(rq, &sdev->sdev_gendev, NULL, NULL);
898
899 if (error)
900 /* we're treating error on bsg register as non-fatal,
901 * so pretend nothing went wrong */
902 sdev_printk(KERN_INFO, sdev,
903 "Failed to register bsg queue, errno=%d\n", error);
904
905 /* add additional host specific attributes */
906 if (sdev->host->hostt->sdev_attrs) {
907 for (i = 0; sdev->host->hostt->sdev_attrs[i]; i++) {
908 error = device_create_file(&sdev->sdev_gendev,
909 sdev->host->hostt->sdev_attrs[i]);
910 if (error)
911 return error;
912 }
913 }
914
915 return error;
916}
917
918void __scsi_remove_device(struct scsi_device *sdev)
919{
920 struct device *dev = &sdev->sdev_gendev;
921
922 if (sdev->is_visible) {
923 if (scsi_device_set_state(sdev, SDEV_CANCEL) != 0)
924 return;
925
926 bsg_unregister_queue(sdev->request_queue);
927 device_unregister(&sdev->sdev_dev);
928 transport_remove_device(dev);
929 device_del(dev);
930 } else
931 put_device(&sdev->sdev_dev);
932 scsi_device_set_state(sdev, SDEV_DEL);
933 if (sdev->host->hostt->slave_destroy)
934 sdev->host->hostt->slave_destroy(sdev);
935 transport_destroy_device(dev);
936 put_device(dev);
937}
938
939/**
940 * scsi_remove_device - unregister a device from the scsi bus
941 * @sdev: scsi_device to unregister
942 **/
943void scsi_remove_device(struct scsi_device *sdev)
944{
945 struct Scsi_Host *shost = sdev->host;
946
947 mutex_lock(&shost->scan_mutex);
948 __scsi_remove_device(sdev);
949 mutex_unlock(&shost->scan_mutex);
950}
951EXPORT_SYMBOL(scsi_remove_device);
952
953static void __scsi_remove_target(struct scsi_target *starget)
954{
955 struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
956 unsigned long flags;
957 struct scsi_device *sdev;
958
959 spin_lock_irqsave(shost->host_lock, flags);
960 starget->reap_ref++;
961 restart:
962 list_for_each_entry(sdev, &shost->__devices, siblings) {
963 if (sdev->channel != starget->channel ||
964 sdev->id != starget->id ||
965 sdev->sdev_state == SDEV_DEL)
966 continue;
967 spin_unlock_irqrestore(shost->host_lock, flags);
968 scsi_remove_device(sdev);
969 spin_lock_irqsave(shost->host_lock, flags);
970 goto restart;
971 }
972 spin_unlock_irqrestore(shost->host_lock, flags);
973 scsi_target_reap(starget);
974}
975
976static int __remove_child (struct device * dev, void * data)
977{
978 if (scsi_is_target_device(dev))
979 __scsi_remove_target(to_scsi_target(dev));
980 return 0;
981}
982
983/**
984 * scsi_remove_target - try to remove a target and all its devices
985 * @dev: generic starget or parent of generic stargets to be removed
986 *
987 * Note: This is slightly racy. It is possible that if the user
988 * requests the addition of another device then the target won't be
989 * removed.
990 */
991void scsi_remove_target(struct device *dev)
992{
993 struct device *rdev;
994
995 if (scsi_is_target_device(dev)) {
996 __scsi_remove_target(to_scsi_target(dev));
997 return;
998 }
999
1000 rdev = get_device(dev);
1001 device_for_each_child(dev, NULL, __remove_child);
1002 put_device(rdev);
1003}
1004EXPORT_SYMBOL(scsi_remove_target);
1005
1006int scsi_register_driver(struct device_driver *drv)
1007{
1008 drv->bus = &scsi_bus_type;
1009
1010 return driver_register(drv);
1011}
1012EXPORT_SYMBOL(scsi_register_driver);
1013
1014int scsi_register_interface(struct class_interface *intf)
1015{
1016 intf->class = &sdev_class;
1017
1018 return class_interface_register(intf);
1019}
1020EXPORT_SYMBOL(scsi_register_interface);
1021
1022/**
1023 * scsi_sysfs_add_host - add scsi host to subsystem
1024 * @shost: scsi host struct to add to subsystem
1025 **/
1026int scsi_sysfs_add_host(struct Scsi_Host *shost)
1027{
1028 int error, i;
1029
1030 /* add host specific attributes */
1031 if (shost->hostt->shost_attrs) {
1032 for (i = 0; shost->hostt->shost_attrs[i]; i++) {
1033 error = device_create_file(&shost->shost_dev,
1034 shost->hostt->shost_attrs[i]);
1035 if (error)
1036 return error;
1037 }
1038 }
1039
1040 transport_register_device(&shost->shost_gendev);
1041 transport_configure_device(&shost->shost_gendev);
1042 return 0;
1043}
1044
1045static struct device_type scsi_dev_type = {
1046 .name = "scsi_device",
1047 .release = scsi_device_dev_release,
1048 .groups = scsi_sdev_attr_groups,
1049};
1050
1051void scsi_sysfs_device_initialize(struct scsi_device *sdev)
1052{
1053 unsigned long flags;
1054 struct Scsi_Host *shost = sdev->host;
1055 struct scsi_target *starget = sdev->sdev_target;
1056
1057 device_initialize(&sdev->sdev_gendev);
1058 sdev->sdev_gendev.bus = &scsi_bus_type;
1059 sdev->sdev_gendev.type = &scsi_dev_type;
1060 dev_set_name(&sdev->sdev_gendev, "%d:%d:%d:%d",
1061 sdev->host->host_no, sdev->channel, sdev->id, sdev->lun);
1062
1063 device_initialize(&sdev->sdev_dev);
1064 sdev->sdev_dev.parent = get_device(&sdev->sdev_gendev);
1065 sdev->sdev_dev.class = &sdev_class;
1066 dev_set_name(&sdev->sdev_dev, "%d:%d:%d:%d",
1067 sdev->host->host_no, sdev->channel, sdev->id, sdev->lun);
1068 sdev->scsi_level = starget->scsi_level;
1069 transport_setup_device(&sdev->sdev_gendev);
1070 spin_lock_irqsave(shost->host_lock, flags);
1071 list_add_tail(&sdev->same_target_siblings, &starget->devices);
1072 list_add_tail(&sdev->siblings, &shost->__devices);
1073 spin_unlock_irqrestore(shost->host_lock, flags);
1074}
1075
1076int scsi_is_sdev_device(const struct device *dev)
1077{
1078 return dev->type == &scsi_dev_type;
1079}
1080EXPORT_SYMBOL(scsi_is_sdev_device);
1081
1082/* A blank transport template that is used in drivers that don't
1083 * yet implement Transport Attributes */
1084struct scsi_transport_template blank_transport_template = { { { {NULL, }, }, }, };