6 * Copyright IBM Corp. 2008, 2010
9 #define KMSG_COMPONENT "zfcp"
10 #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt
12 #include <linux/slab.h>
15 #define ZFCP_DEV_ATTR(_feat, _name, _mode, _show, _store) \
16 struct device_attribute dev_attr_##_feat##_##_name = __ATTR(_name, _mode,\
18 #define ZFCP_DEFINE_ATTR(_feat_def, _feat, _name, _format, _value) \
19 static ssize_t zfcp_sysfs_##_feat##_##_name##_show(struct device *dev, \
20 struct device_attribute *at,\
23 struct _feat_def *_feat = container_of(dev, struct _feat_def, dev); \
25 return sprintf(buf, _format, _value); \
27 static ZFCP_DEV_ATTR(_feat, _name, S_IRUGO, \
28 zfcp_sysfs_##_feat##_##_name##_show, NULL);
30 #define ZFCP_DEFINE_ATTR_CONST(_feat, _name, _format, _value) \
31 static ssize_t zfcp_sysfs_##_feat##_##_name##_show(struct device *dev, \
32 struct device_attribute *at,\
35 return sprintf(buf, _format, _value); \
37 static ZFCP_DEV_ATTR(_feat, _name, S_IRUGO, \
38 zfcp_sysfs_##_feat##_##_name##_show, NULL);
40 #define ZFCP_DEFINE_A_ATTR(_name, _format, _value) \
41 static ssize_t zfcp_sysfs_adapter_##_name##_show(struct device *dev, \
42 struct device_attribute *at,\
45 struct ccw_device *cdev = to_ccwdev(dev); \
46 struct zfcp_adapter *adapter = zfcp_ccw_adapter_by_cdev(cdev); \
52 i = sprintf(buf, _format, _value); \
53 zfcp_ccw_adapter_put(adapter); \
56 static ZFCP_DEV_ATTR(adapter, _name, S_IRUGO, \
57 zfcp_sysfs_adapter_##_name##_show, NULL);
59 ZFCP_DEFINE_A_ATTR(status, "0x%08x\n", atomic_read(&adapter->status));
60 ZFCP_DEFINE_A_ATTR(peer_wwnn, "0x%016llx\n",
61 (unsigned long long) adapter->peer_wwnn);
62 ZFCP_DEFINE_A_ATTR(peer_wwpn, "0x%016llx\n",
63 (unsigned long long) adapter->peer_wwpn);
64 ZFCP_DEFINE_A_ATTR(peer_d_id, "0x%06x\n", adapter->peer_d_id);
65 ZFCP_DEFINE_A_ATTR(card_version, "0x%04x\n", adapter->hydra_version);
66 ZFCP_DEFINE_A_ATTR(lic_version, "0x%08x\n", adapter->fsf_lic_version);
67 ZFCP_DEFINE_A_ATTR(hardware_version, "0x%08x\n", adapter->hardware_version);
68 ZFCP_DEFINE_A_ATTR(in_recovery, "%d\n", (atomic_read(&adapter->status) &
69 ZFCP_STATUS_COMMON_ERP_INUSE) != 0);
71 ZFCP_DEFINE_ATTR(zfcp_port, port, status, "0x%08x\n",
72 atomic_read(&port->status));
73 ZFCP_DEFINE_ATTR(zfcp_port, port, in_recovery, "%d\n",
74 (atomic_read(&port->status) &
75 ZFCP_STATUS_COMMON_ERP_INUSE) != 0);
76 ZFCP_DEFINE_ATTR_CONST(port, access_denied, "%d\n", 0);
78 ZFCP_DEFINE_ATTR(zfcp_unit, unit, status, "0x%08x\n",
79 zfcp_unit_sdev_status(unit));
80 ZFCP_DEFINE_ATTR(zfcp_unit, unit, in_recovery, "%d\n",
81 (zfcp_unit_sdev_status(unit) &
82 ZFCP_STATUS_COMMON_ERP_INUSE) != 0);
83 ZFCP_DEFINE_ATTR(zfcp_unit, unit, access_denied, "%d\n",
84 (zfcp_unit_sdev_status(unit) &
85 ZFCP_STATUS_COMMON_ACCESS_DENIED) != 0);
86 ZFCP_DEFINE_ATTR_CONST(unit, access_shared, "%d\n", 0);
87 ZFCP_DEFINE_ATTR_CONST(unit, access_readonly, "%d\n", 0);
89 static ssize_t zfcp_sysfs_port_failed_show(struct device *dev,
90 struct device_attribute *attr,
93 struct zfcp_port *port = container_of(dev, struct zfcp_port, dev);
95 if (atomic_read(&port->status) & ZFCP_STATUS_COMMON_ERP_FAILED)
96 return sprintf(buf, "1\n");
98 return sprintf(buf, "0\n");
101 static ssize_t zfcp_sysfs_port_failed_store(struct device *dev,
102 struct device_attribute *attr,
103 const char *buf, size_t count)
105 struct zfcp_port *port = container_of(dev, struct zfcp_port, dev);
108 if (kstrtoul(buf, 0, &val) || val != 0)
111 zfcp_erp_set_port_status(port, ZFCP_STATUS_COMMON_RUNNING);
112 zfcp_erp_port_reopen(port, ZFCP_STATUS_COMMON_ERP_FAILED, "sypfai2");
113 zfcp_erp_wait(port->adapter);
117 static ZFCP_DEV_ATTR(port, failed, S_IWUSR | S_IRUGO,
118 zfcp_sysfs_port_failed_show,
119 zfcp_sysfs_port_failed_store);
121 static ssize_t zfcp_sysfs_unit_failed_show(struct device *dev,
122 struct device_attribute *attr,
125 struct zfcp_unit *unit = container_of(dev, struct zfcp_unit, dev);
126 struct scsi_device *sdev;
127 unsigned int status, failed = 1;
129 sdev = zfcp_unit_sdev(unit);
131 status = atomic_read(&sdev_to_zfcp(sdev)->status);
132 failed = status & ZFCP_STATUS_COMMON_ERP_FAILED ? 1 : 0;
133 scsi_device_put(sdev);
136 return sprintf(buf, "%d\n", failed);
139 static ssize_t zfcp_sysfs_unit_failed_store(struct device *dev,
140 struct device_attribute *attr,
141 const char *buf, size_t count)
143 struct zfcp_unit *unit = container_of(dev, struct zfcp_unit, dev);
145 struct scsi_device *sdev;
147 if (kstrtoul(buf, 0, &val) || val != 0)
150 sdev = zfcp_unit_sdev(unit);
152 zfcp_erp_set_lun_status(sdev, ZFCP_STATUS_COMMON_RUNNING);
153 zfcp_erp_lun_reopen(sdev, ZFCP_STATUS_COMMON_ERP_FAILED,
155 zfcp_erp_wait(unit->port->adapter);
157 zfcp_unit_scsi_scan(unit);
161 static ZFCP_DEV_ATTR(unit, failed, S_IWUSR | S_IRUGO,
162 zfcp_sysfs_unit_failed_show,
163 zfcp_sysfs_unit_failed_store);
165 static ssize_t zfcp_sysfs_adapter_failed_show(struct device *dev,
166 struct device_attribute *attr,
169 struct ccw_device *cdev = to_ccwdev(dev);
170 struct zfcp_adapter *adapter = zfcp_ccw_adapter_by_cdev(cdev);
176 if (atomic_read(&adapter->status) & ZFCP_STATUS_COMMON_ERP_FAILED)
177 i = sprintf(buf, "1\n");
179 i = sprintf(buf, "0\n");
181 zfcp_ccw_adapter_put(adapter);
185 static ssize_t zfcp_sysfs_adapter_failed_store(struct device *dev,
186 struct device_attribute *attr,
187 const char *buf, size_t count)
189 struct ccw_device *cdev = to_ccwdev(dev);
190 struct zfcp_adapter *adapter = zfcp_ccw_adapter_by_cdev(cdev);
197 if (kstrtoul(buf, 0, &val) || val != 0) {
202 zfcp_erp_set_adapter_status(adapter, ZFCP_STATUS_COMMON_RUNNING);
203 zfcp_erp_adapter_reopen(adapter, ZFCP_STATUS_COMMON_ERP_FAILED,
205 zfcp_erp_wait(adapter);
207 zfcp_ccw_adapter_put(adapter);
208 return retval ? retval : (ssize_t) count;
210 static ZFCP_DEV_ATTR(adapter, failed, S_IWUSR | S_IRUGO,
211 zfcp_sysfs_adapter_failed_show,
212 zfcp_sysfs_adapter_failed_store);
214 static ssize_t zfcp_sysfs_port_rescan_store(struct device *dev,
215 struct device_attribute *attr,
216 const char *buf, size_t count)
218 struct ccw_device *cdev = to_ccwdev(dev);
219 struct zfcp_adapter *adapter = zfcp_ccw_adapter_by_cdev(cdev);
225 * Users wish is our command: immediately schedule and flush a
226 * worker to conduct a synchronous port scan, that is, neither
227 * a random delay nor a rate limit is applied here.
229 queue_delayed_work(adapter->work_queue, &adapter->scan_work, 0);
230 flush_delayed_work(&adapter->scan_work);
231 zfcp_ccw_adapter_put(adapter);
233 return (ssize_t) count;
235 static ZFCP_DEV_ATTR(adapter, port_rescan, S_IWUSR, NULL,
236 zfcp_sysfs_port_rescan_store);
238 DEFINE_MUTEX(zfcp_sysfs_port_units_mutex);
240 static void zfcp_sysfs_port_set_removing(struct zfcp_port *const port)
242 lockdep_assert_held(&zfcp_sysfs_port_units_mutex);
243 atomic_set(&port->units, -1);
246 bool zfcp_sysfs_port_is_removing(const struct zfcp_port *const port)
248 lockdep_assert_held(&zfcp_sysfs_port_units_mutex);
249 return atomic_read(&port->units) == -1;
252 static bool zfcp_sysfs_port_in_use(struct zfcp_port *const port)
254 struct zfcp_adapter *const adapter = port->adapter;
256 struct scsi_device *sdev;
259 mutex_lock(&zfcp_sysfs_port_units_mutex);
260 if (atomic_read(&port->units) > 0)
261 goto unlock_port_units_mutex; /* zfcp_unit(s) under port */
263 spin_lock_irqsave(adapter->scsi_host->host_lock, flags);
264 __shost_for_each_device(sdev, adapter->scsi_host) {
265 const struct zfcp_scsi_dev *zsdev = sdev_to_zfcp(sdev);
267 if (sdev->sdev_state == SDEV_DEL ||
268 sdev->sdev_state == SDEV_CANCEL)
270 if (zsdev->port != port)
272 /* alive scsi_device under port of interest */
273 goto unlock_host_lock;
276 /* port is about to be removed, so no more unit_add or slave_alloc */
277 zfcp_sysfs_port_set_removing(port);
281 spin_unlock_irqrestore(adapter->scsi_host->host_lock, flags);
282 unlock_port_units_mutex:
283 mutex_unlock(&zfcp_sysfs_port_units_mutex);
287 static ssize_t zfcp_sysfs_port_remove_store(struct device *dev,
288 struct device_attribute *attr,
289 const char *buf, size_t count)
291 struct ccw_device *cdev = to_ccwdev(dev);
292 struct zfcp_adapter *adapter = zfcp_ccw_adapter_by_cdev(cdev);
293 struct zfcp_port *port;
295 int retval = -EINVAL;
300 if (kstrtoull(buf, 0, (unsigned long long *) &wwpn))
303 port = zfcp_get_port_by_wwpn(adapter, wwpn);
309 if (zfcp_sysfs_port_in_use(port)) {
311 put_device(&port->dev); /* undo zfcp_get_port_by_wwpn() */
315 write_lock_irq(&adapter->port_list_lock);
316 list_del(&port->list);
317 write_unlock_irq(&adapter->port_list_lock);
319 put_device(&port->dev);
321 zfcp_erp_port_shutdown(port, 0, "syprs_1");
322 device_unregister(&port->dev);
324 zfcp_ccw_adapter_put(adapter);
325 return retval ? retval : (ssize_t) count;
327 static ZFCP_DEV_ATTR(adapter, port_remove, S_IWUSR, NULL,
328 zfcp_sysfs_port_remove_store);
330 static struct attribute *zfcp_adapter_attrs[] = {
331 &dev_attr_adapter_failed.attr,
332 &dev_attr_adapter_in_recovery.attr,
333 &dev_attr_adapter_port_remove.attr,
334 &dev_attr_adapter_port_rescan.attr,
335 &dev_attr_adapter_peer_wwnn.attr,
336 &dev_attr_adapter_peer_wwpn.attr,
337 &dev_attr_adapter_peer_d_id.attr,
338 &dev_attr_adapter_card_version.attr,
339 &dev_attr_adapter_lic_version.attr,
340 &dev_attr_adapter_status.attr,
341 &dev_attr_adapter_hardware_version.attr,
345 struct attribute_group zfcp_sysfs_adapter_attrs = {
346 .attrs = zfcp_adapter_attrs,
349 static ssize_t zfcp_sysfs_unit_add_store(struct device *dev,
350 struct device_attribute *attr,
351 const char *buf, size_t count)
353 struct zfcp_port *port = container_of(dev, struct zfcp_port, dev);
357 if (kstrtoull(buf, 0, (unsigned long long *) &fcp_lun))
360 retval = zfcp_unit_add(port, fcp_lun);
366 static DEVICE_ATTR(unit_add, S_IWUSR, NULL, zfcp_sysfs_unit_add_store);
368 static ssize_t zfcp_sysfs_unit_remove_store(struct device *dev,
369 struct device_attribute *attr,
370 const char *buf, size_t count)
372 struct zfcp_port *port = container_of(dev, struct zfcp_port, dev);
375 if (kstrtoull(buf, 0, (unsigned long long *) &fcp_lun))
378 if (zfcp_unit_remove(port, fcp_lun))
383 static DEVICE_ATTR(unit_remove, S_IWUSR, NULL, zfcp_sysfs_unit_remove_store);
385 static struct attribute *zfcp_port_attrs[] = {
386 &dev_attr_unit_add.attr,
387 &dev_attr_unit_remove.attr,
388 &dev_attr_port_failed.attr,
389 &dev_attr_port_in_recovery.attr,
390 &dev_attr_port_status.attr,
391 &dev_attr_port_access_denied.attr,
394 static struct attribute_group zfcp_port_attr_group = {
395 .attrs = zfcp_port_attrs,
397 const struct attribute_group *zfcp_port_attr_groups[] = {
398 &zfcp_port_attr_group,
402 static struct attribute *zfcp_unit_attrs[] = {
403 &dev_attr_unit_failed.attr,
404 &dev_attr_unit_in_recovery.attr,
405 &dev_attr_unit_status.attr,
406 &dev_attr_unit_access_denied.attr,
407 &dev_attr_unit_access_shared.attr,
408 &dev_attr_unit_access_readonly.attr,
411 static struct attribute_group zfcp_unit_attr_group = {
412 .attrs = zfcp_unit_attrs,
414 const struct attribute_group *zfcp_unit_attr_groups[] = {
415 &zfcp_unit_attr_group,
419 #define ZFCP_DEFINE_LATENCY_ATTR(_name) \
421 zfcp_sysfs_unit_##_name##_latency_show(struct device *dev, \
422 struct device_attribute *attr, \
424 struct scsi_device *sdev = to_scsi_device(dev); \
425 struct zfcp_scsi_dev *zfcp_sdev = sdev_to_zfcp(sdev); \
426 struct zfcp_latencies *lat = &zfcp_sdev->latencies; \
427 struct zfcp_adapter *adapter = zfcp_sdev->port->adapter; \
428 unsigned long long fsum, fmin, fmax, csum, cmin, cmax, cc; \
430 spin_lock_bh(&lat->lock); \
431 fsum = lat->_name.fabric.sum * adapter->timer_ticks; \
432 fmin = lat->_name.fabric.min * adapter->timer_ticks; \
433 fmax = lat->_name.fabric.max * adapter->timer_ticks; \
434 csum = lat->_name.channel.sum * adapter->timer_ticks; \
435 cmin = lat->_name.channel.min * adapter->timer_ticks; \
436 cmax = lat->_name.channel.max * adapter->timer_ticks; \
437 cc = lat->_name.counter; \
438 spin_unlock_bh(&lat->lock); \
440 do_div(fsum, 1000); \
441 do_div(fmin, 1000); \
442 do_div(fmax, 1000); \
443 do_div(csum, 1000); \
444 do_div(cmin, 1000); \
445 do_div(cmax, 1000); \
447 return sprintf(buf, "%llu %llu %llu %llu %llu %llu %llu\n", \
448 fmin, fmax, fsum, cmin, cmax, csum, cc); \
451 zfcp_sysfs_unit_##_name##_latency_store(struct device *dev, \
452 struct device_attribute *attr, \
453 const char *buf, size_t count) \
455 struct scsi_device *sdev = to_scsi_device(dev); \
456 struct zfcp_scsi_dev *zfcp_sdev = sdev_to_zfcp(sdev); \
457 struct zfcp_latencies *lat = &zfcp_sdev->latencies; \
458 unsigned long flags; \
460 spin_lock_irqsave(&lat->lock, flags); \
461 lat->_name.fabric.sum = 0; \
462 lat->_name.fabric.min = 0xFFFFFFFF; \
463 lat->_name.fabric.max = 0; \
464 lat->_name.channel.sum = 0; \
465 lat->_name.channel.min = 0xFFFFFFFF; \
466 lat->_name.channel.max = 0; \
467 lat->_name.counter = 0; \
468 spin_unlock_irqrestore(&lat->lock, flags); \
470 return (ssize_t) count; \
472 static DEVICE_ATTR(_name##_latency, S_IWUSR | S_IRUGO, \
473 zfcp_sysfs_unit_##_name##_latency_show, \
474 zfcp_sysfs_unit_##_name##_latency_store);
476 ZFCP_DEFINE_LATENCY_ATTR(read);
477 ZFCP_DEFINE_LATENCY_ATTR(write);
478 ZFCP_DEFINE_LATENCY_ATTR(cmd);
480 #define ZFCP_DEFINE_SCSI_ATTR(_name, _format, _value) \
481 static ssize_t zfcp_sysfs_scsi_##_name##_show(struct device *dev, \
482 struct device_attribute *attr,\
485 struct scsi_device *sdev = to_scsi_device(dev); \
486 struct zfcp_scsi_dev *zfcp_sdev = sdev_to_zfcp(sdev); \
488 return sprintf(buf, _format, _value); \
490 static DEVICE_ATTR(_name, S_IRUGO, zfcp_sysfs_scsi_##_name##_show, NULL);
492 ZFCP_DEFINE_SCSI_ATTR(hba_id, "%s\n",
493 dev_name(&zfcp_sdev->port->adapter->ccw_device->dev));
494 ZFCP_DEFINE_SCSI_ATTR(wwpn, "0x%016llx\n",
495 (unsigned long long) zfcp_sdev->port->wwpn);
497 static ssize_t zfcp_sysfs_scsi_fcp_lun_show(struct device *dev,
498 struct device_attribute *attr,
501 struct scsi_device *sdev = to_scsi_device(dev);
503 return sprintf(buf, "0x%016llx\n", zfcp_scsi_dev_lun(sdev));
505 static DEVICE_ATTR(fcp_lun, S_IRUGO, zfcp_sysfs_scsi_fcp_lun_show, NULL);
507 ZFCP_DEFINE_SCSI_ATTR(zfcp_access_denied, "%d\n",
508 (atomic_read(&zfcp_sdev->status) &
509 ZFCP_STATUS_COMMON_ACCESS_DENIED) != 0);
511 static ssize_t zfcp_sysfs_scsi_zfcp_failed_show(struct device *dev,
512 struct device_attribute *attr,
515 struct scsi_device *sdev = to_scsi_device(dev);
516 unsigned int status = atomic_read(&sdev_to_zfcp(sdev)->status);
517 unsigned int failed = status & ZFCP_STATUS_COMMON_ERP_FAILED ? 1 : 0;
519 return sprintf(buf, "%d\n", failed);
522 static ssize_t zfcp_sysfs_scsi_zfcp_failed_store(struct device *dev,
523 struct device_attribute *attr,
524 const char *buf, size_t count)
526 struct scsi_device *sdev = to_scsi_device(dev);
529 if (kstrtoul(buf, 0, &val) || val != 0)
532 zfcp_erp_set_lun_status(sdev, ZFCP_STATUS_COMMON_RUNNING);
533 zfcp_erp_lun_reopen(sdev, ZFCP_STATUS_COMMON_ERP_FAILED,
535 zfcp_erp_wait(sdev_to_zfcp(sdev)->port->adapter);
539 static DEVICE_ATTR(zfcp_failed, S_IWUSR | S_IRUGO,
540 zfcp_sysfs_scsi_zfcp_failed_show,
541 zfcp_sysfs_scsi_zfcp_failed_store);
543 ZFCP_DEFINE_SCSI_ATTR(zfcp_in_recovery, "%d\n",
544 (atomic_read(&zfcp_sdev->status) &
545 ZFCP_STATUS_COMMON_ERP_INUSE) != 0);
547 ZFCP_DEFINE_SCSI_ATTR(zfcp_status, "0x%08x\n",
548 atomic_read(&zfcp_sdev->status));
550 struct device_attribute *zfcp_sysfs_sdev_attrs[] = {
554 &dev_attr_read_latency,
555 &dev_attr_write_latency,
556 &dev_attr_cmd_latency,
557 &dev_attr_zfcp_access_denied,
558 &dev_attr_zfcp_failed,
559 &dev_attr_zfcp_in_recovery,
560 &dev_attr_zfcp_status,
564 static ssize_t zfcp_sysfs_adapter_util_show(struct device *dev,
565 struct device_attribute *attr,
568 struct Scsi_Host *scsi_host = dev_to_shost(dev);
569 struct fsf_qtcb_bottom_port *qtcb_port;
570 struct zfcp_adapter *adapter;
573 adapter = (struct zfcp_adapter *) scsi_host->hostdata[0];
574 if (!(adapter->adapter_features & FSF_FEATURE_MEASUREMENT_DATA))
577 qtcb_port = kzalloc(sizeof(struct fsf_qtcb_bottom_port), GFP_KERNEL);
581 retval = zfcp_fsf_exchange_port_data_sync(adapter->qdio, qtcb_port);
583 retval = sprintf(buf, "%u %u %u\n", qtcb_port->cp_util,
584 qtcb_port->cb_util, qtcb_port->a_util);
588 static DEVICE_ATTR(utilization, S_IRUGO, zfcp_sysfs_adapter_util_show, NULL);
590 static int zfcp_sysfs_adapter_ex_config(struct device *dev,
591 struct fsf_statistics_info *stat_inf)
593 struct Scsi_Host *scsi_host = dev_to_shost(dev);
594 struct fsf_qtcb_bottom_config *qtcb_config;
595 struct zfcp_adapter *adapter;
598 adapter = (struct zfcp_adapter *) scsi_host->hostdata[0];
599 if (!(adapter->adapter_features & FSF_FEATURE_MEASUREMENT_DATA))
602 qtcb_config = kzalloc(sizeof(struct fsf_qtcb_bottom_config),
607 retval = zfcp_fsf_exchange_config_data_sync(adapter->qdio, qtcb_config);
609 *stat_inf = qtcb_config->stat_info;
615 #define ZFCP_SHOST_ATTR(_name, _format, _arg...) \
616 static ssize_t zfcp_sysfs_adapter_##_name##_show(struct device *dev, \
617 struct device_attribute *attr,\
620 struct fsf_statistics_info stat_info; \
623 retval = zfcp_sysfs_adapter_ex_config(dev, &stat_info); \
627 return sprintf(buf, _format, ## _arg); \
629 static DEVICE_ATTR(_name, S_IRUGO, zfcp_sysfs_adapter_##_name##_show, NULL);
631 ZFCP_SHOST_ATTR(requests, "%llu %llu %llu\n",
632 (unsigned long long) stat_info.input_req,
633 (unsigned long long) stat_info.output_req,
634 (unsigned long long) stat_info.control_req);
636 ZFCP_SHOST_ATTR(megabytes, "%llu %llu\n",
637 (unsigned long long) stat_info.input_mb,
638 (unsigned long long) stat_info.output_mb);
640 ZFCP_SHOST_ATTR(seconds_active, "%llu\n",
641 (unsigned long long) stat_info.seconds_act);
643 static ssize_t zfcp_sysfs_adapter_q_full_show(struct device *dev,
644 struct device_attribute *attr,
647 struct Scsi_Host *scsi_host = class_to_shost(dev);
648 struct zfcp_qdio *qdio =
649 ((struct zfcp_adapter *) scsi_host->hostdata[0])->qdio;
652 spin_lock_bh(&qdio->stat_lock);
653 util = qdio->req_q_util;
654 spin_unlock_bh(&qdio->stat_lock);
656 return sprintf(buf, "%d %llu\n", atomic_read(&qdio->req_q_full),
657 (unsigned long long)util);
659 static DEVICE_ATTR(queue_full, S_IRUGO, zfcp_sysfs_adapter_q_full_show, NULL);
661 struct device_attribute *zfcp_sysfs_shost_attrs[] = {
662 &dev_attr_utilization,
665 &dev_attr_seconds_active,
666 &dev_attr_queue_full,