Loading...
1/*
2 * FiberChannel transport specific attributes exported to sysfs.
3 *
4 * Copyright (c) 2003 Silicon Graphics, Inc. All rights reserved.
5 *
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License as published by
8 * the Free Software Foundation; either version 2 of the License, or
9 * (at your option) any later version.
10 *
11 * This program is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 * GNU General Public License for more details.
15 *
16 * You should have received a copy of the GNU General Public License
17 * along with this program; if not, write to the Free Software
18 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
19 *
20 * ========
21 *
22 * Copyright (C) 2004-2007 James Smart, Emulex Corporation
23 * Rewrite for host, target, device, and remote port attributes,
24 * statistics, and service functions...
25 * Add vports, etc
26 *
27 */
28#include <linux/module.h>
29#include <linux/init.h>
30#include <linux/slab.h>
31#include <linux/delay.h>
32#include <linux/kernel.h>
33#include <scsi/scsi_device.h>
34#include <scsi/scsi_host.h>
35#include <scsi/scsi_transport.h>
36#include <scsi/scsi_transport_fc.h>
37#include <scsi/scsi_cmnd.h>
38#include <net/netlink.h>
39#include <scsi/scsi_netlink_fc.h>
40#include <scsi/scsi_bsg_fc.h>
41#include "scsi_priv.h"
42#include "scsi_transport_fc_internal.h"
43
44static int fc_queue_work(struct Scsi_Host *, struct work_struct *);
45static void fc_vport_sched_delete(struct work_struct *work);
46static int fc_vport_setup(struct Scsi_Host *shost, int channel,
47 struct device *pdev, struct fc_vport_identifiers *ids,
48 struct fc_vport **vport);
49static int fc_bsg_hostadd(struct Scsi_Host *, struct fc_host_attrs *);
50static int fc_bsg_rportadd(struct Scsi_Host *, struct fc_rport *);
51static void fc_bsg_remove(struct request_queue *);
52static void fc_bsg_goose_queue(struct fc_rport *);
53
54/*
55 * Module Parameters
56 */
57
58/*
59 * dev_loss_tmo: the default number of seconds that the FC transport
60 * should insulate the loss of a remote port.
61 * The maximum will be capped by the value of SCSI_DEVICE_BLOCK_MAX_TIMEOUT.
62 */
63static unsigned int fc_dev_loss_tmo = 60; /* seconds */
64
65module_param_named(dev_loss_tmo, fc_dev_loss_tmo, uint, S_IRUGO|S_IWUSR);
66MODULE_PARM_DESC(dev_loss_tmo,
67 "Maximum number of seconds that the FC transport should"
68 " insulate the loss of a remote port. Once this value is"
69 " exceeded, the scsi target is removed. Value should be"
70 " between 1 and SCSI_DEVICE_BLOCK_MAX_TIMEOUT if"
71 " fast_io_fail_tmo is not set.");
72
73/*
74 * Redefine so that we can have same named attributes in the
75 * sdev/starget/host objects.
76 */
77#define FC_DEVICE_ATTR(_prefix,_name,_mode,_show,_store) \
78struct device_attribute device_attr_##_prefix##_##_name = \
79 __ATTR(_name,_mode,_show,_store)
80
81#define fc_enum_name_search(title, table_type, table) \
82static const char *get_fc_##title##_name(enum table_type table_key) \
83{ \
84 int i; \
85 char *name = NULL; \
86 \
87 for (i = 0; i < ARRAY_SIZE(table); i++) { \
88 if (table[i].value == table_key) { \
89 name = table[i].name; \
90 break; \
91 } \
92 } \
93 return name; \
94}
95
96#define fc_enum_name_match(title, table_type, table) \
97static int get_fc_##title##_match(const char *table_key, \
98 enum table_type *value) \
99{ \
100 int i; \
101 \
102 for (i = 0; i < ARRAY_SIZE(table); i++) { \
103 if (strncmp(table_key, table[i].name, \
104 table[i].matchlen) == 0) { \
105 *value = table[i].value; \
106 return 0; /* success */ \
107 } \
108 } \
109 return 1; /* failure */ \
110}
111
112
113/* Convert fc_port_type values to ascii string name */
114static struct {
115 enum fc_port_type value;
116 char *name;
117} fc_port_type_names[] = {
118 { FC_PORTTYPE_UNKNOWN, "Unknown" },
119 { FC_PORTTYPE_OTHER, "Other" },
120 { FC_PORTTYPE_NOTPRESENT, "Not Present" },
121 { FC_PORTTYPE_NPORT, "NPort (fabric via point-to-point)" },
122 { FC_PORTTYPE_NLPORT, "NLPort (fabric via loop)" },
123 { FC_PORTTYPE_LPORT, "LPort (private loop)" },
124 { FC_PORTTYPE_PTP, "Point-To-Point (direct nport connection)" },
125 { FC_PORTTYPE_NPIV, "NPIV VPORT" },
126};
127fc_enum_name_search(port_type, fc_port_type, fc_port_type_names)
128#define FC_PORTTYPE_MAX_NAMELEN 50
129
130/* Reuse fc_port_type enum function for vport_type */
131#define get_fc_vport_type_name get_fc_port_type_name
132
133
134/* Convert fc_host_event_code values to ascii string name */
135static const struct {
136 enum fc_host_event_code value;
137 char *name;
138} fc_host_event_code_names[] = {
139 { FCH_EVT_LIP, "lip" },
140 { FCH_EVT_LINKUP, "link_up" },
141 { FCH_EVT_LINKDOWN, "link_down" },
142 { FCH_EVT_LIPRESET, "lip_reset" },
143 { FCH_EVT_RSCN, "rscn" },
144 { FCH_EVT_ADAPTER_CHANGE, "adapter_chg" },
145 { FCH_EVT_PORT_UNKNOWN, "port_unknown" },
146 { FCH_EVT_PORT_ONLINE, "port_online" },
147 { FCH_EVT_PORT_OFFLINE, "port_offline" },
148 { FCH_EVT_PORT_FABRIC, "port_fabric" },
149 { FCH_EVT_LINK_UNKNOWN, "link_unknown" },
150 { FCH_EVT_VENDOR_UNIQUE, "vendor_unique" },
151};
152fc_enum_name_search(host_event_code, fc_host_event_code,
153 fc_host_event_code_names)
154#define FC_HOST_EVENT_CODE_MAX_NAMELEN 30
155
156
157/* Convert fc_port_state values to ascii string name */
158static struct {
159 enum fc_port_state value;
160 char *name;
161} fc_port_state_names[] = {
162 { FC_PORTSTATE_UNKNOWN, "Unknown" },
163 { FC_PORTSTATE_NOTPRESENT, "Not Present" },
164 { FC_PORTSTATE_ONLINE, "Online" },
165 { FC_PORTSTATE_OFFLINE, "Offline" },
166 { FC_PORTSTATE_BLOCKED, "Blocked" },
167 { FC_PORTSTATE_BYPASSED, "Bypassed" },
168 { FC_PORTSTATE_DIAGNOSTICS, "Diagnostics" },
169 { FC_PORTSTATE_LINKDOWN, "Linkdown" },
170 { FC_PORTSTATE_ERROR, "Error" },
171 { FC_PORTSTATE_LOOPBACK, "Loopback" },
172 { FC_PORTSTATE_DELETED, "Deleted" },
173};
174fc_enum_name_search(port_state, fc_port_state, fc_port_state_names)
175#define FC_PORTSTATE_MAX_NAMELEN 20
176
177
178/* Convert fc_vport_state values to ascii string name */
179static struct {
180 enum fc_vport_state value;
181 char *name;
182} fc_vport_state_names[] = {
183 { FC_VPORT_UNKNOWN, "Unknown" },
184 { FC_VPORT_ACTIVE, "Active" },
185 { FC_VPORT_DISABLED, "Disabled" },
186 { FC_VPORT_LINKDOWN, "Linkdown" },
187 { FC_VPORT_INITIALIZING, "Initializing" },
188 { FC_VPORT_NO_FABRIC_SUPP, "No Fabric Support" },
189 { FC_VPORT_NO_FABRIC_RSCS, "No Fabric Resources" },
190 { FC_VPORT_FABRIC_LOGOUT, "Fabric Logout" },
191 { FC_VPORT_FABRIC_REJ_WWN, "Fabric Rejected WWN" },
192 { FC_VPORT_FAILED, "VPort Failed" },
193};
194fc_enum_name_search(vport_state, fc_vport_state, fc_vport_state_names)
195#define FC_VPORTSTATE_MAX_NAMELEN 24
196
197/* Reuse fc_vport_state enum function for vport_last_state */
198#define get_fc_vport_last_state_name get_fc_vport_state_name
199
200
201/* Convert fc_tgtid_binding_type values to ascii string name */
202static const struct {
203 enum fc_tgtid_binding_type value;
204 char *name;
205 int matchlen;
206} fc_tgtid_binding_type_names[] = {
207 { FC_TGTID_BIND_NONE, "none", 4 },
208 { FC_TGTID_BIND_BY_WWPN, "wwpn (World Wide Port Name)", 4 },
209 { FC_TGTID_BIND_BY_WWNN, "wwnn (World Wide Node Name)", 4 },
210 { FC_TGTID_BIND_BY_ID, "port_id (FC Address)", 7 },
211};
212fc_enum_name_search(tgtid_bind_type, fc_tgtid_binding_type,
213 fc_tgtid_binding_type_names)
214fc_enum_name_match(tgtid_bind_type, fc_tgtid_binding_type,
215 fc_tgtid_binding_type_names)
216#define FC_BINDTYPE_MAX_NAMELEN 30
217
218
219#define fc_bitfield_name_search(title, table) \
220static ssize_t \
221get_fc_##title##_names(u32 table_key, char *buf) \
222{ \
223 char *prefix = ""; \
224 ssize_t len = 0; \
225 int i; \
226 \
227 for (i = 0; i < ARRAY_SIZE(table); i++) { \
228 if (table[i].value & table_key) { \
229 len += sprintf(buf + len, "%s%s", \
230 prefix, table[i].name); \
231 prefix = ", "; \
232 } \
233 } \
234 len += sprintf(buf + len, "\n"); \
235 return len; \
236}
237
238
239/* Convert FC_COS bit values to ascii string name */
240static const struct {
241 u32 value;
242 char *name;
243} fc_cos_names[] = {
244 { FC_COS_CLASS1, "Class 1" },
245 { FC_COS_CLASS2, "Class 2" },
246 { FC_COS_CLASS3, "Class 3" },
247 { FC_COS_CLASS4, "Class 4" },
248 { FC_COS_CLASS6, "Class 6" },
249};
250fc_bitfield_name_search(cos, fc_cos_names)
251
252
253/* Convert FC_PORTSPEED bit values to ascii string name */
254static const struct {
255 u32 value;
256 char *name;
257} fc_port_speed_names[] = {
258 { FC_PORTSPEED_1GBIT, "1 Gbit" },
259 { FC_PORTSPEED_2GBIT, "2 Gbit" },
260 { FC_PORTSPEED_4GBIT, "4 Gbit" },
261 { FC_PORTSPEED_10GBIT, "10 Gbit" },
262 { FC_PORTSPEED_8GBIT, "8 Gbit" },
263 { FC_PORTSPEED_16GBIT, "16 Gbit" },
264 { FC_PORTSPEED_32GBIT, "32 Gbit" },
265 { FC_PORTSPEED_NOT_NEGOTIATED, "Not Negotiated" },
266};
267fc_bitfield_name_search(port_speed, fc_port_speed_names)
268
269
270static int
271show_fc_fc4s (char *buf, u8 *fc4_list)
272{
273 int i, len=0;
274
275 for (i = 0; i < FC_FC4_LIST_SIZE; i++, fc4_list++)
276 len += sprintf(buf + len , "0x%02x ", *fc4_list);
277 len += sprintf(buf + len, "\n");
278 return len;
279}
280
281
282/* Convert FC_PORT_ROLE bit values to ascii string name */
283static const struct {
284 u32 value;
285 char *name;
286} fc_port_role_names[] = {
287 { FC_PORT_ROLE_FCP_TARGET, "FCP Target" },
288 { FC_PORT_ROLE_FCP_INITIATOR, "FCP Initiator" },
289 { FC_PORT_ROLE_IP_PORT, "IP Port" },
290};
291fc_bitfield_name_search(port_roles, fc_port_role_names)
292
293/*
294 * Define roles that are specific to port_id. Values are relative to ROLE_MASK.
295 */
296#define FC_WELLKNOWN_PORTID_MASK 0xfffff0
297#define FC_WELLKNOWN_ROLE_MASK 0x00000f
298#define FC_FPORT_PORTID 0x00000e
299#define FC_FABCTLR_PORTID 0x00000d
300#define FC_DIRSRVR_PORTID 0x00000c
301#define FC_TIMESRVR_PORTID 0x00000b
302#define FC_MGMTSRVR_PORTID 0x00000a
303
304
305static void fc_timeout_deleted_rport(struct work_struct *work);
306static void fc_timeout_fail_rport_io(struct work_struct *work);
307static void fc_scsi_scan_rport(struct work_struct *work);
308
309/*
310 * Attribute counts pre object type...
311 * Increase these values if you add attributes
312 */
313#define FC_STARGET_NUM_ATTRS 3
314#define FC_RPORT_NUM_ATTRS 10
315#define FC_VPORT_NUM_ATTRS 9
316#define FC_HOST_NUM_ATTRS 29
317
318struct fc_internal {
319 struct scsi_transport_template t;
320 struct fc_function_template *f;
321
322 /*
323 * For attributes : each object has :
324 * An array of the actual attributes structures
325 * An array of null-terminated pointers to the attribute
326 * structures - used for mid-layer interaction.
327 *
328 * The attribute containers for the starget and host are are
329 * part of the midlayer. As the remote port is specific to the
330 * fc transport, we must provide the attribute container.
331 */
332 struct device_attribute private_starget_attrs[
333 FC_STARGET_NUM_ATTRS];
334 struct device_attribute *starget_attrs[FC_STARGET_NUM_ATTRS + 1];
335
336 struct device_attribute private_host_attrs[FC_HOST_NUM_ATTRS];
337 struct device_attribute *host_attrs[FC_HOST_NUM_ATTRS + 1];
338
339 struct transport_container rport_attr_cont;
340 struct device_attribute private_rport_attrs[FC_RPORT_NUM_ATTRS];
341 struct device_attribute *rport_attrs[FC_RPORT_NUM_ATTRS + 1];
342
343 struct transport_container vport_attr_cont;
344 struct device_attribute private_vport_attrs[FC_VPORT_NUM_ATTRS];
345 struct device_attribute *vport_attrs[FC_VPORT_NUM_ATTRS + 1];
346};
347
348#define to_fc_internal(tmpl) container_of(tmpl, struct fc_internal, t)
349
350static int fc_target_setup(struct transport_container *tc, struct device *dev,
351 struct device *cdev)
352{
353 struct scsi_target *starget = to_scsi_target(dev);
354 struct fc_rport *rport = starget_to_rport(starget);
355
356 /*
357 * if parent is remote port, use values from remote port.
358 * Otherwise, this host uses the fc_transport, but not the
359 * remote port interface. As such, initialize to known non-values.
360 */
361 if (rport) {
362 fc_starget_node_name(starget) = rport->node_name;
363 fc_starget_port_name(starget) = rport->port_name;
364 fc_starget_port_id(starget) = rport->port_id;
365 } else {
366 fc_starget_node_name(starget) = -1;
367 fc_starget_port_name(starget) = -1;
368 fc_starget_port_id(starget) = -1;
369 }
370
371 return 0;
372}
373
374static DECLARE_TRANSPORT_CLASS(fc_transport_class,
375 "fc_transport",
376 fc_target_setup,
377 NULL,
378 NULL);
379
380static int fc_host_setup(struct transport_container *tc, struct device *dev,
381 struct device *cdev)
382{
383 struct Scsi_Host *shost = dev_to_shost(dev);
384 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
385
386 /*
387 * Set default values easily detected by the midlayer as
388 * failure cases. The scsi lldd is responsible for initializing
389 * all transport attributes to valid values per host.
390 */
391 fc_host->node_name = -1;
392 fc_host->port_name = -1;
393 fc_host->permanent_port_name = -1;
394 fc_host->supported_classes = FC_COS_UNSPECIFIED;
395 memset(fc_host->supported_fc4s, 0,
396 sizeof(fc_host->supported_fc4s));
397 fc_host->supported_speeds = FC_PORTSPEED_UNKNOWN;
398 fc_host->maxframe_size = -1;
399 fc_host->max_npiv_vports = 0;
400 memset(fc_host->serial_number, 0,
401 sizeof(fc_host->serial_number));
402 memset(fc_host->manufacturer, 0,
403 sizeof(fc_host->manufacturer));
404 memset(fc_host->model, 0,
405 sizeof(fc_host->model));
406 memset(fc_host->model_description, 0,
407 sizeof(fc_host->model_description));
408 memset(fc_host->hardware_version, 0,
409 sizeof(fc_host->hardware_version));
410 memset(fc_host->driver_version, 0,
411 sizeof(fc_host->driver_version));
412 memset(fc_host->firmware_version, 0,
413 sizeof(fc_host->firmware_version));
414 memset(fc_host->optionrom_version, 0,
415 sizeof(fc_host->optionrom_version));
416
417 fc_host->port_id = -1;
418 fc_host->port_type = FC_PORTTYPE_UNKNOWN;
419 fc_host->port_state = FC_PORTSTATE_UNKNOWN;
420 memset(fc_host->active_fc4s, 0,
421 sizeof(fc_host->active_fc4s));
422 fc_host->speed = FC_PORTSPEED_UNKNOWN;
423 fc_host->fabric_name = -1;
424 memset(fc_host->symbolic_name, 0, sizeof(fc_host->symbolic_name));
425 memset(fc_host->system_hostname, 0, sizeof(fc_host->system_hostname));
426
427 fc_host->tgtid_bind_type = FC_TGTID_BIND_BY_WWPN;
428
429 INIT_LIST_HEAD(&fc_host->rports);
430 INIT_LIST_HEAD(&fc_host->rport_bindings);
431 INIT_LIST_HEAD(&fc_host->vports);
432 fc_host->next_rport_number = 0;
433 fc_host->next_target_id = 0;
434 fc_host->next_vport_number = 0;
435 fc_host->npiv_vports_inuse = 0;
436
437 snprintf(fc_host->work_q_name, sizeof(fc_host->work_q_name),
438 "fc_wq_%d", shost->host_no);
439 fc_host->work_q = alloc_workqueue("%s", 0, 0, fc_host->work_q_name);
440 if (!fc_host->work_q)
441 return -ENOMEM;
442
443 fc_host->dev_loss_tmo = fc_dev_loss_tmo;
444 snprintf(fc_host->devloss_work_q_name,
445 sizeof(fc_host->devloss_work_q_name),
446 "fc_dl_%d", shost->host_no);
447 fc_host->devloss_work_q = alloc_workqueue("%s", 0, 0,
448 fc_host->devloss_work_q_name);
449 if (!fc_host->devloss_work_q) {
450 destroy_workqueue(fc_host->work_q);
451 fc_host->work_q = NULL;
452 return -ENOMEM;
453 }
454
455 fc_bsg_hostadd(shost, fc_host);
456 /* ignore any bsg add error - we just can't do sgio */
457
458 return 0;
459}
460
461static int fc_host_remove(struct transport_container *tc, struct device *dev,
462 struct device *cdev)
463{
464 struct Scsi_Host *shost = dev_to_shost(dev);
465 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
466
467 fc_bsg_remove(fc_host->rqst_q);
468 return 0;
469}
470
471static DECLARE_TRANSPORT_CLASS(fc_host_class,
472 "fc_host",
473 fc_host_setup,
474 fc_host_remove,
475 NULL);
476
477/*
478 * Setup and Remove actions for remote ports are handled
479 * in the service functions below.
480 */
481static DECLARE_TRANSPORT_CLASS(fc_rport_class,
482 "fc_remote_ports",
483 NULL,
484 NULL,
485 NULL);
486
487/*
488 * Setup and Remove actions for virtual ports are handled
489 * in the service functions below.
490 */
491static DECLARE_TRANSPORT_CLASS(fc_vport_class,
492 "fc_vports",
493 NULL,
494 NULL,
495 NULL);
496
497/*
498 * Netlink Infrastructure
499 */
500
501static atomic_t fc_event_seq;
502
503/**
504 * fc_get_event_number - Obtain the next sequential FC event number
505 *
506 * Notes:
507 * We could have inlined this, but it would have required fc_event_seq to
508 * be exposed. For now, live with the subroutine call.
509 * Atomic used to avoid lock/unlock...
510 */
511u32
512fc_get_event_number(void)
513{
514 return atomic_add_return(1, &fc_event_seq);
515}
516EXPORT_SYMBOL(fc_get_event_number);
517
518
519/**
520 * fc_host_post_event - called to post an even on an fc_host.
521 * @shost: host the event occurred on
522 * @event_number: fc event number obtained from get_fc_event_number()
523 * @event_code: fc_host event being posted
524 * @event_data: 32bits of data for the event being posted
525 *
526 * Notes:
527 * This routine assumes no locks are held on entry.
528 */
529void
530fc_host_post_event(struct Scsi_Host *shost, u32 event_number,
531 enum fc_host_event_code event_code, u32 event_data)
532{
533 struct sk_buff *skb;
534 struct nlmsghdr *nlh;
535 struct fc_nl_event *event;
536 const char *name;
537 u32 len;
538 int err;
539
540 if (!scsi_nl_sock) {
541 err = -ENOENT;
542 goto send_fail;
543 }
544
545 len = FC_NL_MSGALIGN(sizeof(*event));
546
547 skb = nlmsg_new(len, GFP_KERNEL);
548 if (!skb) {
549 err = -ENOBUFS;
550 goto send_fail;
551 }
552
553 nlh = nlmsg_put(skb, 0, 0, SCSI_TRANSPORT_MSG, len, 0);
554 if (!nlh) {
555 err = -ENOBUFS;
556 goto send_fail_skb;
557 }
558 event = nlmsg_data(nlh);
559
560 INIT_SCSI_NL_HDR(&event->snlh, SCSI_NL_TRANSPORT_FC,
561 FC_NL_ASYNC_EVENT, len);
562 event->seconds = get_seconds();
563 event->vendor_id = 0;
564 event->host_no = shost->host_no;
565 event->event_datalen = sizeof(u32); /* bytes */
566 event->event_num = event_number;
567 event->event_code = event_code;
568 event->event_data = event_data;
569
570 nlmsg_multicast(scsi_nl_sock, skb, 0, SCSI_NL_GRP_FC_EVENTS,
571 GFP_KERNEL);
572 return;
573
574send_fail_skb:
575 kfree_skb(skb);
576send_fail:
577 name = get_fc_host_event_code_name(event_code);
578 printk(KERN_WARNING
579 "%s: Dropped Event : host %d %s data 0x%08x - err %d\n",
580 __func__, shost->host_no,
581 (name) ? name : "<unknown>", event_data, err);
582 return;
583}
584EXPORT_SYMBOL(fc_host_post_event);
585
586
587/**
588 * fc_host_post_vendor_event - called to post a vendor unique event on an fc_host
589 * @shost: host the event occurred on
590 * @event_number: fc event number obtained from get_fc_event_number()
591 * @data_len: amount, in bytes, of vendor unique data
592 * @data_buf: pointer to vendor unique data
593 * @vendor_id: Vendor id
594 *
595 * Notes:
596 * This routine assumes no locks are held on entry.
597 */
598void
599fc_host_post_vendor_event(struct Scsi_Host *shost, u32 event_number,
600 u32 data_len, char * data_buf, u64 vendor_id)
601{
602 struct sk_buff *skb;
603 struct nlmsghdr *nlh;
604 struct fc_nl_event *event;
605 u32 len;
606 int err;
607
608 if (!scsi_nl_sock) {
609 err = -ENOENT;
610 goto send_vendor_fail;
611 }
612
613 len = FC_NL_MSGALIGN(sizeof(*event) + data_len);
614
615 skb = nlmsg_new(len, GFP_KERNEL);
616 if (!skb) {
617 err = -ENOBUFS;
618 goto send_vendor_fail;
619 }
620
621 nlh = nlmsg_put(skb, 0, 0, SCSI_TRANSPORT_MSG, len, 0);
622 if (!nlh) {
623 err = -ENOBUFS;
624 goto send_vendor_fail_skb;
625 }
626 event = nlmsg_data(nlh);
627
628 INIT_SCSI_NL_HDR(&event->snlh, SCSI_NL_TRANSPORT_FC,
629 FC_NL_ASYNC_EVENT, len);
630 event->seconds = get_seconds();
631 event->vendor_id = vendor_id;
632 event->host_no = shost->host_no;
633 event->event_datalen = data_len; /* bytes */
634 event->event_num = event_number;
635 event->event_code = FCH_EVT_VENDOR_UNIQUE;
636 memcpy(&event->event_data, data_buf, data_len);
637
638 nlmsg_multicast(scsi_nl_sock, skb, 0, SCSI_NL_GRP_FC_EVENTS,
639 GFP_KERNEL);
640 return;
641
642send_vendor_fail_skb:
643 kfree_skb(skb);
644send_vendor_fail:
645 printk(KERN_WARNING
646 "%s: Dropped Event : host %d vendor_unique - err %d\n",
647 __func__, shost->host_no, err);
648 return;
649}
650EXPORT_SYMBOL(fc_host_post_vendor_event);
651
652
653
654static __init int fc_transport_init(void)
655{
656 int error;
657
658 atomic_set(&fc_event_seq, 0);
659
660 error = transport_class_register(&fc_host_class);
661 if (error)
662 return error;
663 error = transport_class_register(&fc_vport_class);
664 if (error)
665 goto unreg_host_class;
666 error = transport_class_register(&fc_rport_class);
667 if (error)
668 goto unreg_vport_class;
669 error = transport_class_register(&fc_transport_class);
670 if (error)
671 goto unreg_rport_class;
672 return 0;
673
674unreg_rport_class:
675 transport_class_unregister(&fc_rport_class);
676unreg_vport_class:
677 transport_class_unregister(&fc_vport_class);
678unreg_host_class:
679 transport_class_unregister(&fc_host_class);
680 return error;
681}
682
683static void __exit fc_transport_exit(void)
684{
685 transport_class_unregister(&fc_transport_class);
686 transport_class_unregister(&fc_rport_class);
687 transport_class_unregister(&fc_host_class);
688 transport_class_unregister(&fc_vport_class);
689}
690
691/*
692 * FC Remote Port Attribute Management
693 */
694
695#define fc_rport_show_function(field, format_string, sz, cast) \
696static ssize_t \
697show_fc_rport_##field (struct device *dev, \
698 struct device_attribute *attr, char *buf) \
699{ \
700 struct fc_rport *rport = transport_class_to_rport(dev); \
701 struct Scsi_Host *shost = rport_to_shost(rport); \
702 struct fc_internal *i = to_fc_internal(shost->transportt); \
703 if ((i->f->get_rport_##field) && \
704 !((rport->port_state == FC_PORTSTATE_BLOCKED) || \
705 (rport->port_state == FC_PORTSTATE_DELETED) || \
706 (rport->port_state == FC_PORTSTATE_NOTPRESENT))) \
707 i->f->get_rport_##field(rport); \
708 return snprintf(buf, sz, format_string, cast rport->field); \
709}
710
711#define fc_rport_store_function(field) \
712static ssize_t \
713store_fc_rport_##field(struct device *dev, \
714 struct device_attribute *attr, \
715 const char *buf, size_t count) \
716{ \
717 int val; \
718 struct fc_rport *rport = transport_class_to_rport(dev); \
719 struct Scsi_Host *shost = rport_to_shost(rport); \
720 struct fc_internal *i = to_fc_internal(shost->transportt); \
721 char *cp; \
722 if ((rport->port_state == FC_PORTSTATE_BLOCKED) || \
723 (rport->port_state == FC_PORTSTATE_DELETED) || \
724 (rport->port_state == FC_PORTSTATE_NOTPRESENT)) \
725 return -EBUSY; \
726 val = simple_strtoul(buf, &cp, 0); \
727 if (*cp && (*cp != '\n')) \
728 return -EINVAL; \
729 i->f->set_rport_##field(rport, val); \
730 return count; \
731}
732
733#define fc_rport_rd_attr(field, format_string, sz) \
734 fc_rport_show_function(field, format_string, sz, ) \
735static FC_DEVICE_ATTR(rport, field, S_IRUGO, \
736 show_fc_rport_##field, NULL)
737
738#define fc_rport_rd_attr_cast(field, format_string, sz, cast) \
739 fc_rport_show_function(field, format_string, sz, (cast)) \
740static FC_DEVICE_ATTR(rport, field, S_IRUGO, \
741 show_fc_rport_##field, NULL)
742
743#define fc_rport_rw_attr(field, format_string, sz) \
744 fc_rport_show_function(field, format_string, sz, ) \
745 fc_rport_store_function(field) \
746static FC_DEVICE_ATTR(rport, field, S_IRUGO | S_IWUSR, \
747 show_fc_rport_##field, \
748 store_fc_rport_##field)
749
750
751#define fc_private_rport_show_function(field, format_string, sz, cast) \
752static ssize_t \
753show_fc_rport_##field (struct device *dev, \
754 struct device_attribute *attr, char *buf) \
755{ \
756 struct fc_rport *rport = transport_class_to_rport(dev); \
757 return snprintf(buf, sz, format_string, cast rport->field); \
758}
759
760#define fc_private_rport_rd_attr(field, format_string, sz) \
761 fc_private_rport_show_function(field, format_string, sz, ) \
762static FC_DEVICE_ATTR(rport, field, S_IRUGO, \
763 show_fc_rport_##field, NULL)
764
765#define fc_private_rport_rd_attr_cast(field, format_string, sz, cast) \
766 fc_private_rport_show_function(field, format_string, sz, (cast)) \
767static FC_DEVICE_ATTR(rport, field, S_IRUGO, \
768 show_fc_rport_##field, NULL)
769
770
771#define fc_private_rport_rd_enum_attr(title, maxlen) \
772static ssize_t \
773show_fc_rport_##title (struct device *dev, \
774 struct device_attribute *attr, char *buf) \
775{ \
776 struct fc_rport *rport = transport_class_to_rport(dev); \
777 const char *name; \
778 name = get_fc_##title##_name(rport->title); \
779 if (!name) \
780 return -EINVAL; \
781 return snprintf(buf, maxlen, "%s\n", name); \
782} \
783static FC_DEVICE_ATTR(rport, title, S_IRUGO, \
784 show_fc_rport_##title, NULL)
785
786
787#define SETUP_RPORT_ATTRIBUTE_RD(field) \
788 i->private_rport_attrs[count] = device_attr_rport_##field; \
789 i->private_rport_attrs[count].attr.mode = S_IRUGO; \
790 i->private_rport_attrs[count].store = NULL; \
791 i->rport_attrs[count] = &i->private_rport_attrs[count]; \
792 if (i->f->show_rport_##field) \
793 count++
794
795#define SETUP_PRIVATE_RPORT_ATTRIBUTE_RD(field) \
796 i->private_rport_attrs[count] = device_attr_rport_##field; \
797 i->private_rport_attrs[count].attr.mode = S_IRUGO; \
798 i->private_rport_attrs[count].store = NULL; \
799 i->rport_attrs[count] = &i->private_rport_attrs[count]; \
800 count++
801
802#define SETUP_RPORT_ATTRIBUTE_RW(field) \
803 i->private_rport_attrs[count] = device_attr_rport_##field; \
804 if (!i->f->set_rport_##field) { \
805 i->private_rport_attrs[count].attr.mode = S_IRUGO; \
806 i->private_rport_attrs[count].store = NULL; \
807 } \
808 i->rport_attrs[count] = &i->private_rport_attrs[count]; \
809 if (i->f->show_rport_##field) \
810 count++
811
812#define SETUP_PRIVATE_RPORT_ATTRIBUTE_RW(field) \
813{ \
814 i->private_rport_attrs[count] = device_attr_rport_##field; \
815 i->rport_attrs[count] = &i->private_rport_attrs[count]; \
816 count++; \
817}
818
819
820/* The FC Transport Remote Port Attributes: */
821
822/* Fixed Remote Port Attributes */
823
824fc_private_rport_rd_attr(maxframe_size, "%u bytes\n", 20);
825
826static ssize_t
827show_fc_rport_supported_classes (struct device *dev,
828 struct device_attribute *attr, char *buf)
829{
830 struct fc_rport *rport = transport_class_to_rport(dev);
831 if (rport->supported_classes == FC_COS_UNSPECIFIED)
832 return snprintf(buf, 20, "unspecified\n");
833 return get_fc_cos_names(rport->supported_classes, buf);
834}
835static FC_DEVICE_ATTR(rport, supported_classes, S_IRUGO,
836 show_fc_rport_supported_classes, NULL);
837
838/* Dynamic Remote Port Attributes */
839
840/*
841 * dev_loss_tmo attribute
842 */
843static int fc_str_to_dev_loss(const char *buf, unsigned long *val)
844{
845 char *cp;
846
847 *val = simple_strtoul(buf, &cp, 0);
848 if ((*cp && (*cp != '\n')) || (*val < 0))
849 return -EINVAL;
850 /*
851 * Check for overflow; dev_loss_tmo is u32
852 */
853 if (*val > UINT_MAX)
854 return -EINVAL;
855
856 return 0;
857}
858
859static int fc_rport_set_dev_loss_tmo(struct fc_rport *rport,
860 unsigned long val)
861{
862 struct Scsi_Host *shost = rport_to_shost(rport);
863 struct fc_internal *i = to_fc_internal(shost->transportt);
864
865 if ((rport->port_state == FC_PORTSTATE_BLOCKED) ||
866 (rport->port_state == FC_PORTSTATE_DELETED) ||
867 (rport->port_state == FC_PORTSTATE_NOTPRESENT))
868 return -EBUSY;
869 /*
870 * Check for overflow; dev_loss_tmo is u32
871 */
872 if (val > UINT_MAX)
873 return -EINVAL;
874
875 /*
876 * If fast_io_fail is off we have to cap
877 * dev_loss_tmo at SCSI_DEVICE_BLOCK_MAX_TIMEOUT
878 */
879 if (rport->fast_io_fail_tmo == -1 &&
880 val > SCSI_DEVICE_BLOCK_MAX_TIMEOUT)
881 return -EINVAL;
882
883 i->f->set_rport_dev_loss_tmo(rport, val);
884 return 0;
885}
886
887fc_rport_show_function(dev_loss_tmo, "%d\n", 20, )
888static ssize_t
889store_fc_rport_dev_loss_tmo(struct device *dev, struct device_attribute *attr,
890 const char *buf, size_t count)
891{
892 struct fc_rport *rport = transport_class_to_rport(dev);
893 unsigned long val;
894 int rc;
895
896 rc = fc_str_to_dev_loss(buf, &val);
897 if (rc)
898 return rc;
899
900 rc = fc_rport_set_dev_loss_tmo(rport, val);
901 if (rc)
902 return rc;
903 return count;
904}
905static FC_DEVICE_ATTR(rport, dev_loss_tmo, S_IRUGO | S_IWUSR,
906 show_fc_rport_dev_loss_tmo, store_fc_rport_dev_loss_tmo);
907
908
909/* Private Remote Port Attributes */
910
911fc_private_rport_rd_attr_cast(node_name, "0x%llx\n", 20, unsigned long long);
912fc_private_rport_rd_attr_cast(port_name, "0x%llx\n", 20, unsigned long long);
913fc_private_rport_rd_attr(port_id, "0x%06x\n", 20);
914
915static ssize_t
916show_fc_rport_roles (struct device *dev, struct device_attribute *attr,
917 char *buf)
918{
919 struct fc_rport *rport = transport_class_to_rport(dev);
920
921 /* identify any roles that are port_id specific */
922 if ((rport->port_id != -1) &&
923 (rport->port_id & FC_WELLKNOWN_PORTID_MASK) ==
924 FC_WELLKNOWN_PORTID_MASK) {
925 switch (rport->port_id & FC_WELLKNOWN_ROLE_MASK) {
926 case FC_FPORT_PORTID:
927 return snprintf(buf, 30, "Fabric Port\n");
928 case FC_FABCTLR_PORTID:
929 return snprintf(buf, 30, "Fabric Controller\n");
930 case FC_DIRSRVR_PORTID:
931 return snprintf(buf, 30, "Directory Server\n");
932 case FC_TIMESRVR_PORTID:
933 return snprintf(buf, 30, "Time Server\n");
934 case FC_MGMTSRVR_PORTID:
935 return snprintf(buf, 30, "Management Server\n");
936 default:
937 return snprintf(buf, 30, "Unknown Fabric Entity\n");
938 }
939 } else {
940 if (rport->roles == FC_PORT_ROLE_UNKNOWN)
941 return snprintf(buf, 20, "unknown\n");
942 return get_fc_port_roles_names(rport->roles, buf);
943 }
944}
945static FC_DEVICE_ATTR(rport, roles, S_IRUGO,
946 show_fc_rport_roles, NULL);
947
948fc_private_rport_rd_enum_attr(port_state, FC_PORTSTATE_MAX_NAMELEN);
949fc_private_rport_rd_attr(scsi_target_id, "%d\n", 20);
950
951/*
952 * fast_io_fail_tmo attribute
953 */
954static ssize_t
955show_fc_rport_fast_io_fail_tmo (struct device *dev,
956 struct device_attribute *attr, char *buf)
957{
958 struct fc_rport *rport = transport_class_to_rport(dev);
959
960 if (rport->fast_io_fail_tmo == -1)
961 return snprintf(buf, 5, "off\n");
962 return snprintf(buf, 20, "%d\n", rport->fast_io_fail_tmo);
963}
964
965static ssize_t
966store_fc_rport_fast_io_fail_tmo(struct device *dev,
967 struct device_attribute *attr, const char *buf,
968 size_t count)
969{
970 int val;
971 char *cp;
972 struct fc_rport *rport = transport_class_to_rport(dev);
973
974 if ((rport->port_state == FC_PORTSTATE_BLOCKED) ||
975 (rport->port_state == FC_PORTSTATE_DELETED) ||
976 (rport->port_state == FC_PORTSTATE_NOTPRESENT))
977 return -EBUSY;
978 if (strncmp(buf, "off", 3) == 0)
979 rport->fast_io_fail_tmo = -1;
980 else {
981 val = simple_strtoul(buf, &cp, 0);
982 if ((*cp && (*cp != '\n')) || (val < 0))
983 return -EINVAL;
984 /*
985 * Cap fast_io_fail by dev_loss_tmo or
986 * SCSI_DEVICE_BLOCK_MAX_TIMEOUT.
987 */
988 if ((val >= rport->dev_loss_tmo) ||
989 (val > SCSI_DEVICE_BLOCK_MAX_TIMEOUT))
990 return -EINVAL;
991
992 rport->fast_io_fail_tmo = val;
993 }
994 return count;
995}
996static FC_DEVICE_ATTR(rport, fast_io_fail_tmo, S_IRUGO | S_IWUSR,
997 show_fc_rport_fast_io_fail_tmo, store_fc_rport_fast_io_fail_tmo);
998
999
1000/*
1001 * FC SCSI Target Attribute Management
1002 */
1003
1004/*
1005 * Note: in the target show function we recognize when the remote
1006 * port is in the hierarchy and do not allow the driver to get
1007 * involved in sysfs functions. The driver only gets involved if
1008 * it's the "old" style that doesn't use rports.
1009 */
1010#define fc_starget_show_function(field, format_string, sz, cast) \
1011static ssize_t \
1012show_fc_starget_##field (struct device *dev, \
1013 struct device_attribute *attr, char *buf) \
1014{ \
1015 struct scsi_target *starget = transport_class_to_starget(dev); \
1016 struct Scsi_Host *shost = dev_to_shost(starget->dev.parent); \
1017 struct fc_internal *i = to_fc_internal(shost->transportt); \
1018 struct fc_rport *rport = starget_to_rport(starget); \
1019 if (rport) \
1020 fc_starget_##field(starget) = rport->field; \
1021 else if (i->f->get_starget_##field) \
1022 i->f->get_starget_##field(starget); \
1023 return snprintf(buf, sz, format_string, \
1024 cast fc_starget_##field(starget)); \
1025}
1026
1027#define fc_starget_rd_attr(field, format_string, sz) \
1028 fc_starget_show_function(field, format_string, sz, ) \
1029static FC_DEVICE_ATTR(starget, field, S_IRUGO, \
1030 show_fc_starget_##field, NULL)
1031
1032#define fc_starget_rd_attr_cast(field, format_string, sz, cast) \
1033 fc_starget_show_function(field, format_string, sz, (cast)) \
1034static FC_DEVICE_ATTR(starget, field, S_IRUGO, \
1035 show_fc_starget_##field, NULL)
1036
1037#define SETUP_STARGET_ATTRIBUTE_RD(field) \
1038 i->private_starget_attrs[count] = device_attr_starget_##field; \
1039 i->private_starget_attrs[count].attr.mode = S_IRUGO; \
1040 i->private_starget_attrs[count].store = NULL; \
1041 i->starget_attrs[count] = &i->private_starget_attrs[count]; \
1042 if (i->f->show_starget_##field) \
1043 count++
1044
1045#define SETUP_STARGET_ATTRIBUTE_RW(field) \
1046 i->private_starget_attrs[count] = device_attr_starget_##field; \
1047 if (!i->f->set_starget_##field) { \
1048 i->private_starget_attrs[count].attr.mode = S_IRUGO; \
1049 i->private_starget_attrs[count].store = NULL; \
1050 } \
1051 i->starget_attrs[count] = &i->private_starget_attrs[count]; \
1052 if (i->f->show_starget_##field) \
1053 count++
1054
1055/* The FC Transport SCSI Target Attributes: */
1056fc_starget_rd_attr_cast(node_name, "0x%llx\n", 20, unsigned long long);
1057fc_starget_rd_attr_cast(port_name, "0x%llx\n", 20, unsigned long long);
1058fc_starget_rd_attr(port_id, "0x%06x\n", 20);
1059
1060
1061/*
1062 * FC Virtual Port Attribute Management
1063 */
1064
1065#define fc_vport_show_function(field, format_string, sz, cast) \
1066static ssize_t \
1067show_fc_vport_##field (struct device *dev, \
1068 struct device_attribute *attr, char *buf) \
1069{ \
1070 struct fc_vport *vport = transport_class_to_vport(dev); \
1071 struct Scsi_Host *shost = vport_to_shost(vport); \
1072 struct fc_internal *i = to_fc_internal(shost->transportt); \
1073 if ((i->f->get_vport_##field) && \
1074 !(vport->flags & (FC_VPORT_DEL | FC_VPORT_CREATING))) \
1075 i->f->get_vport_##field(vport); \
1076 return snprintf(buf, sz, format_string, cast vport->field); \
1077}
1078
1079#define fc_vport_store_function(field) \
1080static ssize_t \
1081store_fc_vport_##field(struct device *dev, \
1082 struct device_attribute *attr, \
1083 const char *buf, size_t count) \
1084{ \
1085 int val; \
1086 struct fc_vport *vport = transport_class_to_vport(dev); \
1087 struct Scsi_Host *shost = vport_to_shost(vport); \
1088 struct fc_internal *i = to_fc_internal(shost->transportt); \
1089 char *cp; \
1090 if (vport->flags & (FC_VPORT_DEL | FC_VPORT_CREATING)) \
1091 return -EBUSY; \
1092 val = simple_strtoul(buf, &cp, 0); \
1093 if (*cp && (*cp != '\n')) \
1094 return -EINVAL; \
1095 i->f->set_vport_##field(vport, val); \
1096 return count; \
1097}
1098
1099#define fc_vport_store_str_function(field, slen) \
1100static ssize_t \
1101store_fc_vport_##field(struct device *dev, \
1102 struct device_attribute *attr, \
1103 const char *buf, size_t count) \
1104{ \
1105 struct fc_vport *vport = transport_class_to_vport(dev); \
1106 struct Scsi_Host *shost = vport_to_shost(vport); \
1107 struct fc_internal *i = to_fc_internal(shost->transportt); \
1108 unsigned int cnt=count; \
1109 \
1110 /* count may include a LF at end of string */ \
1111 if (buf[cnt-1] == '\n') \
1112 cnt--; \
1113 if (cnt > ((slen) - 1)) \
1114 return -EINVAL; \
1115 memcpy(vport->field, buf, cnt); \
1116 i->f->set_vport_##field(vport); \
1117 return count; \
1118}
1119
1120#define fc_vport_rd_attr(field, format_string, sz) \
1121 fc_vport_show_function(field, format_string, sz, ) \
1122static FC_DEVICE_ATTR(vport, field, S_IRUGO, \
1123 show_fc_vport_##field, NULL)
1124
1125#define fc_vport_rd_attr_cast(field, format_string, sz, cast) \
1126 fc_vport_show_function(field, format_string, sz, (cast)) \
1127static FC_DEVICE_ATTR(vport, field, S_IRUGO, \
1128 show_fc_vport_##field, NULL)
1129
1130#define fc_vport_rw_attr(field, format_string, sz) \
1131 fc_vport_show_function(field, format_string, sz, ) \
1132 fc_vport_store_function(field) \
1133static FC_DEVICE_ATTR(vport, field, S_IRUGO | S_IWUSR, \
1134 show_fc_vport_##field, \
1135 store_fc_vport_##field)
1136
1137#define fc_private_vport_show_function(field, format_string, sz, cast) \
1138static ssize_t \
1139show_fc_vport_##field (struct device *dev, \
1140 struct device_attribute *attr, char *buf) \
1141{ \
1142 struct fc_vport *vport = transport_class_to_vport(dev); \
1143 return snprintf(buf, sz, format_string, cast vport->field); \
1144}
1145
1146#define fc_private_vport_store_u32_function(field) \
1147static ssize_t \
1148store_fc_vport_##field(struct device *dev, \
1149 struct device_attribute *attr, \
1150 const char *buf, size_t count) \
1151{ \
1152 u32 val; \
1153 struct fc_vport *vport = transport_class_to_vport(dev); \
1154 char *cp; \
1155 if (vport->flags & (FC_VPORT_DEL | FC_VPORT_CREATING)) \
1156 return -EBUSY; \
1157 val = simple_strtoul(buf, &cp, 0); \
1158 if (*cp && (*cp != '\n')) \
1159 return -EINVAL; \
1160 vport->field = val; \
1161 return count; \
1162}
1163
1164
1165#define fc_private_vport_rd_attr(field, format_string, sz) \
1166 fc_private_vport_show_function(field, format_string, sz, ) \
1167static FC_DEVICE_ATTR(vport, field, S_IRUGO, \
1168 show_fc_vport_##field, NULL)
1169
1170#define fc_private_vport_rd_attr_cast(field, format_string, sz, cast) \
1171 fc_private_vport_show_function(field, format_string, sz, (cast)) \
1172static FC_DEVICE_ATTR(vport, field, S_IRUGO, \
1173 show_fc_vport_##field, NULL)
1174
1175#define fc_private_vport_rw_u32_attr(field, format_string, sz) \
1176 fc_private_vport_show_function(field, format_string, sz, ) \
1177 fc_private_vport_store_u32_function(field) \
1178static FC_DEVICE_ATTR(vport, field, S_IRUGO | S_IWUSR, \
1179 show_fc_vport_##field, \
1180 store_fc_vport_##field)
1181
1182
1183#define fc_private_vport_rd_enum_attr(title, maxlen) \
1184static ssize_t \
1185show_fc_vport_##title (struct device *dev, \
1186 struct device_attribute *attr, \
1187 char *buf) \
1188{ \
1189 struct fc_vport *vport = transport_class_to_vport(dev); \
1190 const char *name; \
1191 name = get_fc_##title##_name(vport->title); \
1192 if (!name) \
1193 return -EINVAL; \
1194 return snprintf(buf, maxlen, "%s\n", name); \
1195} \
1196static FC_DEVICE_ATTR(vport, title, S_IRUGO, \
1197 show_fc_vport_##title, NULL)
1198
1199
1200#define SETUP_VPORT_ATTRIBUTE_RD(field) \
1201 i->private_vport_attrs[count] = device_attr_vport_##field; \
1202 i->private_vport_attrs[count].attr.mode = S_IRUGO; \
1203 i->private_vport_attrs[count].store = NULL; \
1204 i->vport_attrs[count] = &i->private_vport_attrs[count]; \
1205 if (i->f->get_##field) \
1206 count++
1207 /* NOTE: Above MACRO differs: checks function not show bit */
1208
1209#define SETUP_PRIVATE_VPORT_ATTRIBUTE_RD(field) \
1210 i->private_vport_attrs[count] = device_attr_vport_##field; \
1211 i->private_vport_attrs[count].attr.mode = S_IRUGO; \
1212 i->private_vport_attrs[count].store = NULL; \
1213 i->vport_attrs[count] = &i->private_vport_attrs[count]; \
1214 count++
1215
1216#define SETUP_VPORT_ATTRIBUTE_WR(field) \
1217 i->private_vport_attrs[count] = device_attr_vport_##field; \
1218 i->vport_attrs[count] = &i->private_vport_attrs[count]; \
1219 if (i->f->field) \
1220 count++
1221 /* NOTE: Above MACRO differs: checks function */
1222
1223#define SETUP_VPORT_ATTRIBUTE_RW(field) \
1224 i->private_vport_attrs[count] = device_attr_vport_##field; \
1225 if (!i->f->set_vport_##field) { \
1226 i->private_vport_attrs[count].attr.mode = S_IRUGO; \
1227 i->private_vport_attrs[count].store = NULL; \
1228 } \
1229 i->vport_attrs[count] = &i->private_vport_attrs[count]; \
1230 count++
1231 /* NOTE: Above MACRO differs: does not check show bit */
1232
1233#define SETUP_PRIVATE_VPORT_ATTRIBUTE_RW(field) \
1234{ \
1235 i->private_vport_attrs[count] = device_attr_vport_##field; \
1236 i->vport_attrs[count] = &i->private_vport_attrs[count]; \
1237 count++; \
1238}
1239
1240
1241/* The FC Transport Virtual Port Attributes: */
1242
1243/* Fixed Virtual Port Attributes */
1244
1245/* Dynamic Virtual Port Attributes */
1246
1247/* Private Virtual Port Attributes */
1248
1249fc_private_vport_rd_enum_attr(vport_state, FC_VPORTSTATE_MAX_NAMELEN);
1250fc_private_vport_rd_enum_attr(vport_last_state, FC_VPORTSTATE_MAX_NAMELEN);
1251fc_private_vport_rd_attr_cast(node_name, "0x%llx\n", 20, unsigned long long);
1252fc_private_vport_rd_attr_cast(port_name, "0x%llx\n", 20, unsigned long long);
1253
1254static ssize_t
1255show_fc_vport_roles (struct device *dev, struct device_attribute *attr,
1256 char *buf)
1257{
1258 struct fc_vport *vport = transport_class_to_vport(dev);
1259
1260 if (vport->roles == FC_PORT_ROLE_UNKNOWN)
1261 return snprintf(buf, 20, "unknown\n");
1262 return get_fc_port_roles_names(vport->roles, buf);
1263}
1264static FC_DEVICE_ATTR(vport, roles, S_IRUGO, show_fc_vport_roles, NULL);
1265
1266fc_private_vport_rd_enum_attr(vport_type, FC_PORTTYPE_MAX_NAMELEN);
1267
1268fc_private_vport_show_function(symbolic_name, "%s\n",
1269 FC_VPORT_SYMBOLIC_NAMELEN + 1, )
1270fc_vport_store_str_function(symbolic_name, FC_VPORT_SYMBOLIC_NAMELEN)
1271static FC_DEVICE_ATTR(vport, symbolic_name, S_IRUGO | S_IWUSR,
1272 show_fc_vport_symbolic_name, store_fc_vport_symbolic_name);
1273
1274static ssize_t
1275store_fc_vport_delete(struct device *dev, struct device_attribute *attr,
1276 const char *buf, size_t count)
1277{
1278 struct fc_vport *vport = transport_class_to_vport(dev);
1279 struct Scsi_Host *shost = vport_to_shost(vport);
1280 unsigned long flags;
1281
1282 spin_lock_irqsave(shost->host_lock, flags);
1283 if (vport->flags & (FC_VPORT_DEL | FC_VPORT_CREATING)) {
1284 spin_unlock_irqrestore(shost->host_lock, flags);
1285 return -EBUSY;
1286 }
1287 vport->flags |= FC_VPORT_DELETING;
1288 spin_unlock_irqrestore(shost->host_lock, flags);
1289
1290 fc_queue_work(shost, &vport->vport_delete_work);
1291 return count;
1292}
1293static FC_DEVICE_ATTR(vport, vport_delete, S_IWUSR,
1294 NULL, store_fc_vport_delete);
1295
1296
1297/*
1298 * Enable/Disable vport
1299 * Write "1" to disable, write "0" to enable
1300 */
1301static ssize_t
1302store_fc_vport_disable(struct device *dev, struct device_attribute *attr,
1303 const char *buf,
1304 size_t count)
1305{
1306 struct fc_vport *vport = transport_class_to_vport(dev);
1307 struct Scsi_Host *shost = vport_to_shost(vport);
1308 struct fc_internal *i = to_fc_internal(shost->transportt);
1309 int stat;
1310
1311 if (vport->flags & (FC_VPORT_DEL | FC_VPORT_CREATING))
1312 return -EBUSY;
1313
1314 if (*buf == '0') {
1315 if (vport->vport_state != FC_VPORT_DISABLED)
1316 return -EALREADY;
1317 } else if (*buf == '1') {
1318 if (vport->vport_state == FC_VPORT_DISABLED)
1319 return -EALREADY;
1320 } else
1321 return -EINVAL;
1322
1323 stat = i->f->vport_disable(vport, ((*buf == '0') ? false : true));
1324 return stat ? stat : count;
1325}
1326static FC_DEVICE_ATTR(vport, vport_disable, S_IWUSR,
1327 NULL, store_fc_vport_disable);
1328
1329
1330/*
1331 * Host Attribute Management
1332 */
1333
1334#define fc_host_show_function(field, format_string, sz, cast) \
1335static ssize_t \
1336show_fc_host_##field (struct device *dev, \
1337 struct device_attribute *attr, char *buf) \
1338{ \
1339 struct Scsi_Host *shost = transport_class_to_shost(dev); \
1340 struct fc_internal *i = to_fc_internal(shost->transportt); \
1341 if (i->f->get_host_##field) \
1342 i->f->get_host_##field(shost); \
1343 return snprintf(buf, sz, format_string, cast fc_host_##field(shost)); \
1344}
1345
1346#define fc_host_store_function(field) \
1347static ssize_t \
1348store_fc_host_##field(struct device *dev, \
1349 struct device_attribute *attr, \
1350 const char *buf, size_t count) \
1351{ \
1352 int val; \
1353 struct Scsi_Host *shost = transport_class_to_shost(dev); \
1354 struct fc_internal *i = to_fc_internal(shost->transportt); \
1355 char *cp; \
1356 \
1357 val = simple_strtoul(buf, &cp, 0); \
1358 if (*cp && (*cp != '\n')) \
1359 return -EINVAL; \
1360 i->f->set_host_##field(shost, val); \
1361 return count; \
1362}
1363
1364#define fc_host_store_str_function(field, slen) \
1365static ssize_t \
1366store_fc_host_##field(struct device *dev, \
1367 struct device_attribute *attr, \
1368 const char *buf, size_t count) \
1369{ \
1370 struct Scsi_Host *shost = transport_class_to_shost(dev); \
1371 struct fc_internal *i = to_fc_internal(shost->transportt); \
1372 unsigned int cnt=count; \
1373 \
1374 /* count may include a LF at end of string */ \
1375 if (buf[cnt-1] == '\n') \
1376 cnt--; \
1377 if (cnt > ((slen) - 1)) \
1378 return -EINVAL; \
1379 memcpy(fc_host_##field(shost), buf, cnt); \
1380 i->f->set_host_##field(shost); \
1381 return count; \
1382}
1383
1384#define fc_host_rd_attr(field, format_string, sz) \
1385 fc_host_show_function(field, format_string, sz, ) \
1386static FC_DEVICE_ATTR(host, field, S_IRUGO, \
1387 show_fc_host_##field, NULL)
1388
1389#define fc_host_rd_attr_cast(field, format_string, sz, cast) \
1390 fc_host_show_function(field, format_string, sz, (cast)) \
1391static FC_DEVICE_ATTR(host, field, S_IRUGO, \
1392 show_fc_host_##field, NULL)
1393
1394#define fc_host_rw_attr(field, format_string, sz) \
1395 fc_host_show_function(field, format_string, sz, ) \
1396 fc_host_store_function(field) \
1397static FC_DEVICE_ATTR(host, field, S_IRUGO | S_IWUSR, \
1398 show_fc_host_##field, \
1399 store_fc_host_##field)
1400
1401#define fc_host_rd_enum_attr(title, maxlen) \
1402static ssize_t \
1403show_fc_host_##title (struct device *dev, \
1404 struct device_attribute *attr, char *buf) \
1405{ \
1406 struct Scsi_Host *shost = transport_class_to_shost(dev); \
1407 struct fc_internal *i = to_fc_internal(shost->transportt); \
1408 const char *name; \
1409 if (i->f->get_host_##title) \
1410 i->f->get_host_##title(shost); \
1411 name = get_fc_##title##_name(fc_host_##title(shost)); \
1412 if (!name) \
1413 return -EINVAL; \
1414 return snprintf(buf, maxlen, "%s\n", name); \
1415} \
1416static FC_DEVICE_ATTR(host, title, S_IRUGO, show_fc_host_##title, NULL)
1417
1418#define SETUP_HOST_ATTRIBUTE_RD(field) \
1419 i->private_host_attrs[count] = device_attr_host_##field; \
1420 i->private_host_attrs[count].attr.mode = S_IRUGO; \
1421 i->private_host_attrs[count].store = NULL; \
1422 i->host_attrs[count] = &i->private_host_attrs[count]; \
1423 if (i->f->show_host_##field) \
1424 count++
1425
1426#define SETUP_HOST_ATTRIBUTE_RD_NS(field) \
1427 i->private_host_attrs[count] = device_attr_host_##field; \
1428 i->private_host_attrs[count].attr.mode = S_IRUGO; \
1429 i->private_host_attrs[count].store = NULL; \
1430 i->host_attrs[count] = &i->private_host_attrs[count]; \
1431 count++
1432
1433#define SETUP_HOST_ATTRIBUTE_RW(field) \
1434 i->private_host_attrs[count] = device_attr_host_##field; \
1435 if (!i->f->set_host_##field) { \
1436 i->private_host_attrs[count].attr.mode = S_IRUGO; \
1437 i->private_host_attrs[count].store = NULL; \
1438 } \
1439 i->host_attrs[count] = &i->private_host_attrs[count]; \
1440 if (i->f->show_host_##field) \
1441 count++
1442
1443
1444#define fc_private_host_show_function(field, format_string, sz, cast) \
1445static ssize_t \
1446show_fc_host_##field (struct device *dev, \
1447 struct device_attribute *attr, char *buf) \
1448{ \
1449 struct Scsi_Host *shost = transport_class_to_shost(dev); \
1450 return snprintf(buf, sz, format_string, cast fc_host_##field(shost)); \
1451}
1452
1453#define fc_private_host_rd_attr(field, format_string, sz) \
1454 fc_private_host_show_function(field, format_string, sz, ) \
1455static FC_DEVICE_ATTR(host, field, S_IRUGO, \
1456 show_fc_host_##field, NULL)
1457
1458#define fc_private_host_rd_attr_cast(field, format_string, sz, cast) \
1459 fc_private_host_show_function(field, format_string, sz, (cast)) \
1460static FC_DEVICE_ATTR(host, field, S_IRUGO, \
1461 show_fc_host_##field, NULL)
1462
1463#define SETUP_PRIVATE_HOST_ATTRIBUTE_RD(field) \
1464 i->private_host_attrs[count] = device_attr_host_##field; \
1465 i->private_host_attrs[count].attr.mode = S_IRUGO; \
1466 i->private_host_attrs[count].store = NULL; \
1467 i->host_attrs[count] = &i->private_host_attrs[count]; \
1468 count++
1469
1470#define SETUP_PRIVATE_HOST_ATTRIBUTE_RW(field) \
1471{ \
1472 i->private_host_attrs[count] = device_attr_host_##field; \
1473 i->host_attrs[count] = &i->private_host_attrs[count]; \
1474 count++; \
1475}
1476
1477
1478/* Fixed Host Attributes */
1479
1480static ssize_t
1481show_fc_host_supported_classes (struct device *dev,
1482 struct device_attribute *attr, char *buf)
1483{
1484 struct Scsi_Host *shost = transport_class_to_shost(dev);
1485
1486 if (fc_host_supported_classes(shost) == FC_COS_UNSPECIFIED)
1487 return snprintf(buf, 20, "unspecified\n");
1488
1489 return get_fc_cos_names(fc_host_supported_classes(shost), buf);
1490}
1491static FC_DEVICE_ATTR(host, supported_classes, S_IRUGO,
1492 show_fc_host_supported_classes, NULL);
1493
1494static ssize_t
1495show_fc_host_supported_fc4s (struct device *dev,
1496 struct device_attribute *attr, char *buf)
1497{
1498 struct Scsi_Host *shost = transport_class_to_shost(dev);
1499 return (ssize_t)show_fc_fc4s(buf, fc_host_supported_fc4s(shost));
1500}
1501static FC_DEVICE_ATTR(host, supported_fc4s, S_IRUGO,
1502 show_fc_host_supported_fc4s, NULL);
1503
1504static ssize_t
1505show_fc_host_supported_speeds (struct device *dev,
1506 struct device_attribute *attr, char *buf)
1507{
1508 struct Scsi_Host *shost = transport_class_to_shost(dev);
1509
1510 if (fc_host_supported_speeds(shost) == FC_PORTSPEED_UNKNOWN)
1511 return snprintf(buf, 20, "unknown\n");
1512
1513 return get_fc_port_speed_names(fc_host_supported_speeds(shost), buf);
1514}
1515static FC_DEVICE_ATTR(host, supported_speeds, S_IRUGO,
1516 show_fc_host_supported_speeds, NULL);
1517
1518
1519fc_private_host_rd_attr_cast(node_name, "0x%llx\n", 20, unsigned long long);
1520fc_private_host_rd_attr_cast(port_name, "0x%llx\n", 20, unsigned long long);
1521fc_private_host_rd_attr_cast(permanent_port_name, "0x%llx\n", 20,
1522 unsigned long long);
1523fc_private_host_rd_attr(maxframe_size, "%u bytes\n", 20);
1524fc_private_host_rd_attr(max_npiv_vports, "%u\n", 20);
1525fc_private_host_rd_attr(serial_number, "%s\n", (FC_SERIAL_NUMBER_SIZE +1));
1526fc_private_host_rd_attr(manufacturer, "%s\n", FC_SERIAL_NUMBER_SIZE + 1);
1527fc_private_host_rd_attr(model, "%s\n", FC_SYMBOLIC_NAME_SIZE + 1);
1528fc_private_host_rd_attr(model_description, "%s\n", FC_SYMBOLIC_NAME_SIZE + 1);
1529fc_private_host_rd_attr(hardware_version, "%s\n", FC_VERSION_STRING_SIZE + 1);
1530fc_private_host_rd_attr(driver_version, "%s\n", FC_VERSION_STRING_SIZE + 1);
1531fc_private_host_rd_attr(firmware_version, "%s\n", FC_VERSION_STRING_SIZE + 1);
1532fc_private_host_rd_attr(optionrom_version, "%s\n", FC_VERSION_STRING_SIZE + 1);
1533
1534
1535/* Dynamic Host Attributes */
1536
1537static ssize_t
1538show_fc_host_active_fc4s (struct device *dev,
1539 struct device_attribute *attr, char *buf)
1540{
1541 struct Scsi_Host *shost = transport_class_to_shost(dev);
1542 struct fc_internal *i = to_fc_internal(shost->transportt);
1543
1544 if (i->f->get_host_active_fc4s)
1545 i->f->get_host_active_fc4s(shost);
1546
1547 return (ssize_t)show_fc_fc4s(buf, fc_host_active_fc4s(shost));
1548}
1549static FC_DEVICE_ATTR(host, active_fc4s, S_IRUGO,
1550 show_fc_host_active_fc4s, NULL);
1551
1552static ssize_t
1553show_fc_host_speed (struct device *dev,
1554 struct device_attribute *attr, char *buf)
1555{
1556 struct Scsi_Host *shost = transport_class_to_shost(dev);
1557 struct fc_internal *i = to_fc_internal(shost->transportt);
1558
1559 if (i->f->get_host_speed)
1560 i->f->get_host_speed(shost);
1561
1562 if (fc_host_speed(shost) == FC_PORTSPEED_UNKNOWN)
1563 return snprintf(buf, 20, "unknown\n");
1564
1565 return get_fc_port_speed_names(fc_host_speed(shost), buf);
1566}
1567static FC_DEVICE_ATTR(host, speed, S_IRUGO,
1568 show_fc_host_speed, NULL);
1569
1570
1571fc_host_rd_attr(port_id, "0x%06x\n", 20);
1572fc_host_rd_enum_attr(port_type, FC_PORTTYPE_MAX_NAMELEN);
1573fc_host_rd_enum_attr(port_state, FC_PORTSTATE_MAX_NAMELEN);
1574fc_host_rd_attr_cast(fabric_name, "0x%llx\n", 20, unsigned long long);
1575fc_host_rd_attr(symbolic_name, "%s\n", FC_SYMBOLIC_NAME_SIZE + 1);
1576
1577fc_private_host_show_function(system_hostname, "%s\n",
1578 FC_SYMBOLIC_NAME_SIZE + 1, )
1579fc_host_store_str_function(system_hostname, FC_SYMBOLIC_NAME_SIZE)
1580static FC_DEVICE_ATTR(host, system_hostname, S_IRUGO | S_IWUSR,
1581 show_fc_host_system_hostname, store_fc_host_system_hostname);
1582
1583
1584/* Private Host Attributes */
1585
1586static ssize_t
1587show_fc_private_host_tgtid_bind_type(struct device *dev,
1588 struct device_attribute *attr, char *buf)
1589{
1590 struct Scsi_Host *shost = transport_class_to_shost(dev);
1591 const char *name;
1592
1593 name = get_fc_tgtid_bind_type_name(fc_host_tgtid_bind_type(shost));
1594 if (!name)
1595 return -EINVAL;
1596 return snprintf(buf, FC_BINDTYPE_MAX_NAMELEN, "%s\n", name);
1597}
1598
1599#define get_list_head_entry(pos, head, member) \
1600 pos = list_entry((head)->next, typeof(*pos), member)
1601
1602static ssize_t
1603store_fc_private_host_tgtid_bind_type(struct device *dev,
1604 struct device_attribute *attr, const char *buf, size_t count)
1605{
1606 struct Scsi_Host *shost = transport_class_to_shost(dev);
1607 struct fc_rport *rport;
1608 enum fc_tgtid_binding_type val;
1609 unsigned long flags;
1610
1611 if (get_fc_tgtid_bind_type_match(buf, &val))
1612 return -EINVAL;
1613
1614 /* if changing bind type, purge all unused consistent bindings */
1615 if (val != fc_host_tgtid_bind_type(shost)) {
1616 spin_lock_irqsave(shost->host_lock, flags);
1617 while (!list_empty(&fc_host_rport_bindings(shost))) {
1618 get_list_head_entry(rport,
1619 &fc_host_rport_bindings(shost), peers);
1620 list_del(&rport->peers);
1621 rport->port_state = FC_PORTSTATE_DELETED;
1622 fc_queue_work(shost, &rport->rport_delete_work);
1623 }
1624 spin_unlock_irqrestore(shost->host_lock, flags);
1625 }
1626
1627 fc_host_tgtid_bind_type(shost) = val;
1628 return count;
1629}
1630
1631static FC_DEVICE_ATTR(host, tgtid_bind_type, S_IRUGO | S_IWUSR,
1632 show_fc_private_host_tgtid_bind_type,
1633 store_fc_private_host_tgtid_bind_type);
1634
1635static ssize_t
1636store_fc_private_host_issue_lip(struct device *dev,
1637 struct device_attribute *attr, const char *buf, size_t count)
1638{
1639 struct Scsi_Host *shost = transport_class_to_shost(dev);
1640 struct fc_internal *i = to_fc_internal(shost->transportt);
1641 int ret;
1642
1643 /* ignore any data value written to the attribute */
1644 if (i->f->issue_fc_host_lip) {
1645 ret = i->f->issue_fc_host_lip(shost);
1646 return ret ? ret: count;
1647 }
1648
1649 return -ENOENT;
1650}
1651
1652static FC_DEVICE_ATTR(host, issue_lip, S_IWUSR, NULL,
1653 store_fc_private_host_issue_lip);
1654
1655static ssize_t
1656store_fc_private_host_dev_loss_tmo(struct device *dev,
1657 struct device_attribute *attr,
1658 const char *buf, size_t count)
1659{
1660 struct Scsi_Host *shost = transport_class_to_shost(dev);
1661 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
1662 struct fc_rport *rport;
1663 unsigned long val, flags;
1664 int rc;
1665
1666 rc = fc_str_to_dev_loss(buf, &val);
1667 if (rc)
1668 return rc;
1669
1670 fc_host_dev_loss_tmo(shost) = val;
1671 spin_lock_irqsave(shost->host_lock, flags);
1672 list_for_each_entry(rport, &fc_host->rports, peers)
1673 fc_rport_set_dev_loss_tmo(rport, val);
1674 spin_unlock_irqrestore(shost->host_lock, flags);
1675 return count;
1676}
1677
1678fc_private_host_show_function(dev_loss_tmo, "%d\n", 20, );
1679static FC_DEVICE_ATTR(host, dev_loss_tmo, S_IRUGO | S_IWUSR,
1680 show_fc_host_dev_loss_tmo,
1681 store_fc_private_host_dev_loss_tmo);
1682
1683fc_private_host_rd_attr(npiv_vports_inuse, "%u\n", 20);
1684
1685/*
1686 * Host Statistics Management
1687 */
1688
1689/* Show a given an attribute in the statistics group */
1690static ssize_t
1691fc_stat_show(const struct device *dev, char *buf, unsigned long offset)
1692{
1693 struct Scsi_Host *shost = transport_class_to_shost(dev);
1694 struct fc_internal *i = to_fc_internal(shost->transportt);
1695 struct fc_host_statistics *stats;
1696 ssize_t ret = -ENOENT;
1697
1698 if (offset > sizeof(struct fc_host_statistics) ||
1699 offset % sizeof(u64) != 0)
1700 WARN_ON(1);
1701
1702 if (i->f->get_fc_host_stats) {
1703 stats = (i->f->get_fc_host_stats)(shost);
1704 if (stats)
1705 ret = snprintf(buf, 20, "0x%llx\n",
1706 (unsigned long long)*(u64 *)(((u8 *) stats) + offset));
1707 }
1708 return ret;
1709}
1710
1711
1712/* generate a read-only statistics attribute */
1713#define fc_host_statistic(name) \
1714static ssize_t show_fcstat_##name(struct device *cd, \
1715 struct device_attribute *attr, \
1716 char *buf) \
1717{ \
1718 return fc_stat_show(cd, buf, \
1719 offsetof(struct fc_host_statistics, name)); \
1720} \
1721static FC_DEVICE_ATTR(host, name, S_IRUGO, show_fcstat_##name, NULL)
1722
1723fc_host_statistic(seconds_since_last_reset);
1724fc_host_statistic(tx_frames);
1725fc_host_statistic(tx_words);
1726fc_host_statistic(rx_frames);
1727fc_host_statistic(rx_words);
1728fc_host_statistic(lip_count);
1729fc_host_statistic(nos_count);
1730fc_host_statistic(error_frames);
1731fc_host_statistic(dumped_frames);
1732fc_host_statistic(link_failure_count);
1733fc_host_statistic(loss_of_sync_count);
1734fc_host_statistic(loss_of_signal_count);
1735fc_host_statistic(prim_seq_protocol_err_count);
1736fc_host_statistic(invalid_tx_word_count);
1737fc_host_statistic(invalid_crc_count);
1738fc_host_statistic(fcp_input_requests);
1739fc_host_statistic(fcp_output_requests);
1740fc_host_statistic(fcp_control_requests);
1741fc_host_statistic(fcp_input_megabytes);
1742fc_host_statistic(fcp_output_megabytes);
1743fc_host_statistic(fcp_packet_alloc_failures);
1744fc_host_statistic(fcp_packet_aborts);
1745fc_host_statistic(fcp_frame_alloc_failures);
1746fc_host_statistic(fc_no_free_exch);
1747fc_host_statistic(fc_no_free_exch_xid);
1748fc_host_statistic(fc_xid_not_found);
1749fc_host_statistic(fc_xid_busy);
1750fc_host_statistic(fc_seq_not_found);
1751fc_host_statistic(fc_non_bls_resp);
1752
1753static ssize_t
1754fc_reset_statistics(struct device *dev, struct device_attribute *attr,
1755 const char *buf, size_t count)
1756{
1757 struct Scsi_Host *shost = transport_class_to_shost(dev);
1758 struct fc_internal *i = to_fc_internal(shost->transportt);
1759
1760 /* ignore any data value written to the attribute */
1761 if (i->f->reset_fc_host_stats) {
1762 i->f->reset_fc_host_stats(shost);
1763 return count;
1764 }
1765
1766 return -ENOENT;
1767}
1768static FC_DEVICE_ATTR(host, reset_statistics, S_IWUSR, NULL,
1769 fc_reset_statistics);
1770
1771static struct attribute *fc_statistics_attrs[] = {
1772 &device_attr_host_seconds_since_last_reset.attr,
1773 &device_attr_host_tx_frames.attr,
1774 &device_attr_host_tx_words.attr,
1775 &device_attr_host_rx_frames.attr,
1776 &device_attr_host_rx_words.attr,
1777 &device_attr_host_lip_count.attr,
1778 &device_attr_host_nos_count.attr,
1779 &device_attr_host_error_frames.attr,
1780 &device_attr_host_dumped_frames.attr,
1781 &device_attr_host_link_failure_count.attr,
1782 &device_attr_host_loss_of_sync_count.attr,
1783 &device_attr_host_loss_of_signal_count.attr,
1784 &device_attr_host_prim_seq_protocol_err_count.attr,
1785 &device_attr_host_invalid_tx_word_count.attr,
1786 &device_attr_host_invalid_crc_count.attr,
1787 &device_attr_host_fcp_input_requests.attr,
1788 &device_attr_host_fcp_output_requests.attr,
1789 &device_attr_host_fcp_control_requests.attr,
1790 &device_attr_host_fcp_input_megabytes.attr,
1791 &device_attr_host_fcp_output_megabytes.attr,
1792 &device_attr_host_fcp_packet_alloc_failures.attr,
1793 &device_attr_host_fcp_packet_aborts.attr,
1794 &device_attr_host_fcp_frame_alloc_failures.attr,
1795 &device_attr_host_fc_no_free_exch.attr,
1796 &device_attr_host_fc_no_free_exch_xid.attr,
1797 &device_attr_host_fc_xid_not_found.attr,
1798 &device_attr_host_fc_xid_busy.attr,
1799 &device_attr_host_fc_seq_not_found.attr,
1800 &device_attr_host_fc_non_bls_resp.attr,
1801 &device_attr_host_reset_statistics.attr,
1802 NULL
1803};
1804
1805static struct attribute_group fc_statistics_group = {
1806 .name = "statistics",
1807 .attrs = fc_statistics_attrs,
1808};
1809
1810
1811/* Host Vport Attributes */
1812
1813static int
1814fc_parse_wwn(const char *ns, u64 *nm)
1815{
1816 unsigned int i, j;
1817 u8 wwn[8];
1818
1819 memset(wwn, 0, sizeof(wwn));
1820
1821 /* Validate and store the new name */
1822 for (i=0, j=0; i < 16; i++) {
1823 int value;
1824
1825 value = hex_to_bin(*ns++);
1826 if (value >= 0)
1827 j = (j << 4) | value;
1828 else
1829 return -EINVAL;
1830 if (i % 2) {
1831 wwn[i/2] = j & 0xff;
1832 j = 0;
1833 }
1834 }
1835
1836 *nm = wwn_to_u64(wwn);
1837
1838 return 0;
1839}
1840
1841
1842/*
1843 * "Short-cut" sysfs variable to create a new vport on a FC Host.
1844 * Input is a string of the form "<WWPN>:<WWNN>". Other attributes
1845 * will default to a NPIV-based FCP_Initiator; The WWNs are specified
1846 * as hex characters, and may *not* contain any prefixes (e.g. 0x, x, etc)
1847 */
1848static ssize_t
1849store_fc_host_vport_create(struct device *dev, struct device_attribute *attr,
1850 const char *buf, size_t count)
1851{
1852 struct Scsi_Host *shost = transport_class_to_shost(dev);
1853 struct fc_vport_identifiers vid;
1854 struct fc_vport *vport;
1855 unsigned int cnt=count;
1856 int stat;
1857
1858 memset(&vid, 0, sizeof(vid));
1859
1860 /* count may include a LF at end of string */
1861 if (buf[cnt-1] == '\n')
1862 cnt--;
1863
1864 /* validate we have enough characters for WWPN */
1865 if ((cnt != (16+1+16)) || (buf[16] != ':'))
1866 return -EINVAL;
1867
1868 stat = fc_parse_wwn(&buf[0], &vid.port_name);
1869 if (stat)
1870 return stat;
1871
1872 stat = fc_parse_wwn(&buf[17], &vid.node_name);
1873 if (stat)
1874 return stat;
1875
1876 vid.roles = FC_PORT_ROLE_FCP_INITIATOR;
1877 vid.vport_type = FC_PORTTYPE_NPIV;
1878 /* vid.symbolic_name is already zero/NULL's */
1879 vid.disable = false; /* always enabled */
1880
1881 /* we only allow support on Channel 0 !!! */
1882 stat = fc_vport_setup(shost, 0, &shost->shost_gendev, &vid, &vport);
1883 return stat ? stat : count;
1884}
1885static FC_DEVICE_ATTR(host, vport_create, S_IWUSR, NULL,
1886 store_fc_host_vport_create);
1887
1888
1889/*
1890 * "Short-cut" sysfs variable to delete a vport on a FC Host.
1891 * Vport is identified by a string containing "<WWPN>:<WWNN>".
1892 * The WWNs are specified as hex characters, and may *not* contain
1893 * any prefixes (e.g. 0x, x, etc)
1894 */
1895static ssize_t
1896store_fc_host_vport_delete(struct device *dev, struct device_attribute *attr,
1897 const char *buf, size_t count)
1898{
1899 struct Scsi_Host *shost = transport_class_to_shost(dev);
1900 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
1901 struct fc_vport *vport;
1902 u64 wwpn, wwnn;
1903 unsigned long flags;
1904 unsigned int cnt=count;
1905 int stat, match;
1906
1907 /* count may include a LF at end of string */
1908 if (buf[cnt-1] == '\n')
1909 cnt--;
1910
1911 /* validate we have enough characters for WWPN */
1912 if ((cnt != (16+1+16)) || (buf[16] != ':'))
1913 return -EINVAL;
1914
1915 stat = fc_parse_wwn(&buf[0], &wwpn);
1916 if (stat)
1917 return stat;
1918
1919 stat = fc_parse_wwn(&buf[17], &wwnn);
1920 if (stat)
1921 return stat;
1922
1923 spin_lock_irqsave(shost->host_lock, flags);
1924 match = 0;
1925 /* we only allow support on Channel 0 !!! */
1926 list_for_each_entry(vport, &fc_host->vports, peers) {
1927 if ((vport->channel == 0) &&
1928 (vport->port_name == wwpn) && (vport->node_name == wwnn)) {
1929 if (vport->flags & (FC_VPORT_DEL | FC_VPORT_CREATING))
1930 break;
1931 vport->flags |= FC_VPORT_DELETING;
1932 match = 1;
1933 break;
1934 }
1935 }
1936 spin_unlock_irqrestore(shost->host_lock, flags);
1937
1938 if (!match)
1939 return -ENODEV;
1940
1941 stat = fc_vport_terminate(vport);
1942 return stat ? stat : count;
1943}
1944static FC_DEVICE_ATTR(host, vport_delete, S_IWUSR, NULL,
1945 store_fc_host_vport_delete);
1946
1947
1948static int fc_host_match(struct attribute_container *cont,
1949 struct device *dev)
1950{
1951 struct Scsi_Host *shost;
1952 struct fc_internal *i;
1953
1954 if (!scsi_is_host_device(dev))
1955 return 0;
1956
1957 shost = dev_to_shost(dev);
1958 if (!shost->transportt || shost->transportt->host_attrs.ac.class
1959 != &fc_host_class.class)
1960 return 0;
1961
1962 i = to_fc_internal(shost->transportt);
1963
1964 return &i->t.host_attrs.ac == cont;
1965}
1966
1967static int fc_target_match(struct attribute_container *cont,
1968 struct device *dev)
1969{
1970 struct Scsi_Host *shost;
1971 struct fc_internal *i;
1972
1973 if (!scsi_is_target_device(dev))
1974 return 0;
1975
1976 shost = dev_to_shost(dev->parent);
1977 if (!shost->transportt || shost->transportt->host_attrs.ac.class
1978 != &fc_host_class.class)
1979 return 0;
1980
1981 i = to_fc_internal(shost->transportt);
1982
1983 return &i->t.target_attrs.ac == cont;
1984}
1985
1986static void fc_rport_dev_release(struct device *dev)
1987{
1988 struct fc_rport *rport = dev_to_rport(dev);
1989 put_device(dev->parent);
1990 kfree(rport);
1991}
1992
1993int scsi_is_fc_rport(const struct device *dev)
1994{
1995 return dev->release == fc_rport_dev_release;
1996}
1997EXPORT_SYMBOL(scsi_is_fc_rport);
1998
1999static int fc_rport_match(struct attribute_container *cont,
2000 struct device *dev)
2001{
2002 struct Scsi_Host *shost;
2003 struct fc_internal *i;
2004
2005 if (!scsi_is_fc_rport(dev))
2006 return 0;
2007
2008 shost = dev_to_shost(dev->parent);
2009 if (!shost->transportt || shost->transportt->host_attrs.ac.class
2010 != &fc_host_class.class)
2011 return 0;
2012
2013 i = to_fc_internal(shost->transportt);
2014
2015 return &i->rport_attr_cont.ac == cont;
2016}
2017
2018
2019static void fc_vport_dev_release(struct device *dev)
2020{
2021 struct fc_vport *vport = dev_to_vport(dev);
2022 put_device(dev->parent); /* release kobj parent */
2023 kfree(vport);
2024}
2025
2026int scsi_is_fc_vport(const struct device *dev)
2027{
2028 return dev->release == fc_vport_dev_release;
2029}
2030EXPORT_SYMBOL(scsi_is_fc_vport);
2031
2032static int fc_vport_match(struct attribute_container *cont,
2033 struct device *dev)
2034{
2035 struct fc_vport *vport;
2036 struct Scsi_Host *shost;
2037 struct fc_internal *i;
2038
2039 if (!scsi_is_fc_vport(dev))
2040 return 0;
2041 vport = dev_to_vport(dev);
2042
2043 shost = vport_to_shost(vport);
2044 if (!shost->transportt || shost->transportt->host_attrs.ac.class
2045 != &fc_host_class.class)
2046 return 0;
2047
2048 i = to_fc_internal(shost->transportt);
2049 return &i->vport_attr_cont.ac == cont;
2050}
2051
2052
2053/**
2054 * fc_timed_out - FC Transport I/O timeout intercept handler
2055 * @scmd: The SCSI command which timed out
2056 *
2057 * This routine protects against error handlers getting invoked while a
2058 * rport is in a blocked state, typically due to a temporarily loss of
2059 * connectivity. If the error handlers are allowed to proceed, requests
2060 * to abort i/o, reset the target, etc will likely fail as there is no way
2061 * to communicate with the device to perform the requested function. These
2062 * failures may result in the midlayer taking the device offline, requiring
2063 * manual intervention to restore operation.
2064 *
2065 * This routine, called whenever an i/o times out, validates the state of
2066 * the underlying rport. If the rport is blocked, it returns
2067 * EH_RESET_TIMER, which will continue to reschedule the timeout.
2068 * Eventually, either the device will return, or devloss_tmo will fire,
2069 * and when the timeout then fires, it will be handled normally.
2070 * If the rport is not blocked, normal error handling continues.
2071 *
2072 * Notes:
2073 * This routine assumes no locks are held on entry.
2074 */
2075static enum blk_eh_timer_return
2076fc_timed_out(struct scsi_cmnd *scmd)
2077{
2078 struct fc_rport *rport = starget_to_rport(scsi_target(scmd->device));
2079
2080 if (rport->port_state == FC_PORTSTATE_BLOCKED)
2081 return BLK_EH_RESET_TIMER;
2082
2083 return BLK_EH_NOT_HANDLED;
2084}
2085
2086/*
2087 * Called by fc_user_scan to locate an rport on the shost that
2088 * matches the channel and target id, and invoke scsi_scan_target()
2089 * on the rport.
2090 */
2091static void
2092fc_user_scan_tgt(struct Scsi_Host *shost, uint channel, uint id, uint lun)
2093{
2094 struct fc_rport *rport;
2095 unsigned long flags;
2096
2097 spin_lock_irqsave(shost->host_lock, flags);
2098
2099 list_for_each_entry(rport, &fc_host_rports(shost), peers) {
2100 if (rport->scsi_target_id == -1)
2101 continue;
2102
2103 if (rport->port_state != FC_PORTSTATE_ONLINE)
2104 continue;
2105
2106 if ((channel == rport->channel) &&
2107 (id == rport->scsi_target_id)) {
2108 spin_unlock_irqrestore(shost->host_lock, flags);
2109 scsi_scan_target(&rport->dev, channel, id, lun, 1);
2110 return;
2111 }
2112 }
2113
2114 spin_unlock_irqrestore(shost->host_lock, flags);
2115}
2116
2117/*
2118 * Called via sysfs scan routines. Necessary, as the FC transport
2119 * wants to place all target objects below the rport object. So this
2120 * routine must invoke the scsi_scan_target() routine with the rport
2121 * object as the parent.
2122 */
2123static int
2124fc_user_scan(struct Scsi_Host *shost, uint channel, uint id, uint lun)
2125{
2126 uint chlo, chhi;
2127 uint tgtlo, tgthi;
2128
2129 if (((channel != SCAN_WILD_CARD) && (channel > shost->max_channel)) ||
2130 ((id != SCAN_WILD_CARD) && (id >= shost->max_id)) ||
2131 ((lun != SCAN_WILD_CARD) && (lun > shost->max_lun)))
2132 return -EINVAL;
2133
2134 if (channel == SCAN_WILD_CARD) {
2135 chlo = 0;
2136 chhi = shost->max_channel + 1;
2137 } else {
2138 chlo = channel;
2139 chhi = channel + 1;
2140 }
2141
2142 if (id == SCAN_WILD_CARD) {
2143 tgtlo = 0;
2144 tgthi = shost->max_id;
2145 } else {
2146 tgtlo = id;
2147 tgthi = id + 1;
2148 }
2149
2150 for ( ; chlo < chhi; chlo++)
2151 for ( ; tgtlo < tgthi; tgtlo++)
2152 fc_user_scan_tgt(shost, chlo, tgtlo, lun);
2153
2154 return 0;
2155}
2156
2157static int fc_tsk_mgmt_response(struct Scsi_Host *shost, u64 nexus, u64 tm_id,
2158 int result)
2159{
2160 struct fc_internal *i = to_fc_internal(shost->transportt);
2161 return i->f->tsk_mgmt_response(shost, nexus, tm_id, result);
2162}
2163
2164static int fc_it_nexus_response(struct Scsi_Host *shost, u64 nexus, int result)
2165{
2166 struct fc_internal *i = to_fc_internal(shost->transportt);
2167 return i->f->it_nexus_response(shost, nexus, result);
2168}
2169
2170struct scsi_transport_template *
2171fc_attach_transport(struct fc_function_template *ft)
2172{
2173 int count;
2174 struct fc_internal *i = kzalloc(sizeof(struct fc_internal),
2175 GFP_KERNEL);
2176
2177 if (unlikely(!i))
2178 return NULL;
2179
2180 i->t.target_attrs.ac.attrs = &i->starget_attrs[0];
2181 i->t.target_attrs.ac.class = &fc_transport_class.class;
2182 i->t.target_attrs.ac.match = fc_target_match;
2183 i->t.target_size = sizeof(struct fc_starget_attrs);
2184 transport_container_register(&i->t.target_attrs);
2185
2186 i->t.host_attrs.ac.attrs = &i->host_attrs[0];
2187 i->t.host_attrs.ac.class = &fc_host_class.class;
2188 i->t.host_attrs.ac.match = fc_host_match;
2189 i->t.host_size = sizeof(struct fc_host_attrs);
2190 if (ft->get_fc_host_stats)
2191 i->t.host_attrs.statistics = &fc_statistics_group;
2192 transport_container_register(&i->t.host_attrs);
2193
2194 i->rport_attr_cont.ac.attrs = &i->rport_attrs[0];
2195 i->rport_attr_cont.ac.class = &fc_rport_class.class;
2196 i->rport_attr_cont.ac.match = fc_rport_match;
2197 transport_container_register(&i->rport_attr_cont);
2198
2199 i->vport_attr_cont.ac.attrs = &i->vport_attrs[0];
2200 i->vport_attr_cont.ac.class = &fc_vport_class.class;
2201 i->vport_attr_cont.ac.match = fc_vport_match;
2202 transport_container_register(&i->vport_attr_cont);
2203
2204 i->f = ft;
2205
2206 /* Transport uses the shost workq for scsi scanning */
2207 i->t.create_work_queue = 1;
2208
2209 i->t.eh_timed_out = fc_timed_out;
2210
2211 i->t.user_scan = fc_user_scan;
2212
2213 /* target-mode drivers' functions */
2214 i->t.tsk_mgmt_response = fc_tsk_mgmt_response;
2215 i->t.it_nexus_response = fc_it_nexus_response;
2216
2217 /*
2218 * Setup SCSI Target Attributes.
2219 */
2220 count = 0;
2221 SETUP_STARGET_ATTRIBUTE_RD(node_name);
2222 SETUP_STARGET_ATTRIBUTE_RD(port_name);
2223 SETUP_STARGET_ATTRIBUTE_RD(port_id);
2224
2225 BUG_ON(count > FC_STARGET_NUM_ATTRS);
2226
2227 i->starget_attrs[count] = NULL;
2228
2229
2230 /*
2231 * Setup SCSI Host Attributes.
2232 */
2233 count=0;
2234 SETUP_HOST_ATTRIBUTE_RD(node_name);
2235 SETUP_HOST_ATTRIBUTE_RD(port_name);
2236 SETUP_HOST_ATTRIBUTE_RD(permanent_port_name);
2237 SETUP_HOST_ATTRIBUTE_RD(supported_classes);
2238 SETUP_HOST_ATTRIBUTE_RD(supported_fc4s);
2239 SETUP_HOST_ATTRIBUTE_RD(supported_speeds);
2240 SETUP_HOST_ATTRIBUTE_RD(maxframe_size);
2241 if (ft->vport_create) {
2242 SETUP_HOST_ATTRIBUTE_RD_NS(max_npiv_vports);
2243 SETUP_HOST_ATTRIBUTE_RD_NS(npiv_vports_inuse);
2244 }
2245 SETUP_HOST_ATTRIBUTE_RD(serial_number);
2246 SETUP_HOST_ATTRIBUTE_RD(manufacturer);
2247 SETUP_HOST_ATTRIBUTE_RD(model);
2248 SETUP_HOST_ATTRIBUTE_RD(model_description);
2249 SETUP_HOST_ATTRIBUTE_RD(hardware_version);
2250 SETUP_HOST_ATTRIBUTE_RD(driver_version);
2251 SETUP_HOST_ATTRIBUTE_RD(firmware_version);
2252 SETUP_HOST_ATTRIBUTE_RD(optionrom_version);
2253
2254 SETUP_HOST_ATTRIBUTE_RD(port_id);
2255 SETUP_HOST_ATTRIBUTE_RD(port_type);
2256 SETUP_HOST_ATTRIBUTE_RD(port_state);
2257 SETUP_HOST_ATTRIBUTE_RD(active_fc4s);
2258 SETUP_HOST_ATTRIBUTE_RD(speed);
2259 SETUP_HOST_ATTRIBUTE_RD(fabric_name);
2260 SETUP_HOST_ATTRIBUTE_RD(symbolic_name);
2261 SETUP_HOST_ATTRIBUTE_RW(system_hostname);
2262
2263 /* Transport-managed attributes */
2264 SETUP_PRIVATE_HOST_ATTRIBUTE_RW(dev_loss_tmo);
2265 SETUP_PRIVATE_HOST_ATTRIBUTE_RW(tgtid_bind_type);
2266 if (ft->issue_fc_host_lip)
2267 SETUP_PRIVATE_HOST_ATTRIBUTE_RW(issue_lip);
2268 if (ft->vport_create)
2269 SETUP_PRIVATE_HOST_ATTRIBUTE_RW(vport_create);
2270 if (ft->vport_delete)
2271 SETUP_PRIVATE_HOST_ATTRIBUTE_RW(vport_delete);
2272
2273 BUG_ON(count > FC_HOST_NUM_ATTRS);
2274
2275 i->host_attrs[count] = NULL;
2276
2277 /*
2278 * Setup Remote Port Attributes.
2279 */
2280 count=0;
2281 SETUP_RPORT_ATTRIBUTE_RD(maxframe_size);
2282 SETUP_RPORT_ATTRIBUTE_RD(supported_classes);
2283 SETUP_RPORT_ATTRIBUTE_RW(dev_loss_tmo);
2284 SETUP_PRIVATE_RPORT_ATTRIBUTE_RD(node_name);
2285 SETUP_PRIVATE_RPORT_ATTRIBUTE_RD(port_name);
2286 SETUP_PRIVATE_RPORT_ATTRIBUTE_RD(port_id);
2287 SETUP_PRIVATE_RPORT_ATTRIBUTE_RD(roles);
2288 SETUP_PRIVATE_RPORT_ATTRIBUTE_RD(port_state);
2289 SETUP_PRIVATE_RPORT_ATTRIBUTE_RD(scsi_target_id);
2290 SETUP_PRIVATE_RPORT_ATTRIBUTE_RW(fast_io_fail_tmo);
2291
2292 BUG_ON(count > FC_RPORT_NUM_ATTRS);
2293
2294 i->rport_attrs[count] = NULL;
2295
2296 /*
2297 * Setup Virtual Port Attributes.
2298 */
2299 count=0;
2300 SETUP_PRIVATE_VPORT_ATTRIBUTE_RD(vport_state);
2301 SETUP_PRIVATE_VPORT_ATTRIBUTE_RD(vport_last_state);
2302 SETUP_PRIVATE_VPORT_ATTRIBUTE_RD(node_name);
2303 SETUP_PRIVATE_VPORT_ATTRIBUTE_RD(port_name);
2304 SETUP_PRIVATE_VPORT_ATTRIBUTE_RD(roles);
2305 SETUP_PRIVATE_VPORT_ATTRIBUTE_RD(vport_type);
2306 SETUP_VPORT_ATTRIBUTE_RW(symbolic_name);
2307 SETUP_VPORT_ATTRIBUTE_WR(vport_delete);
2308 SETUP_VPORT_ATTRIBUTE_WR(vport_disable);
2309
2310 BUG_ON(count > FC_VPORT_NUM_ATTRS);
2311
2312 i->vport_attrs[count] = NULL;
2313
2314 return &i->t;
2315}
2316EXPORT_SYMBOL(fc_attach_transport);
2317
2318void fc_release_transport(struct scsi_transport_template *t)
2319{
2320 struct fc_internal *i = to_fc_internal(t);
2321
2322 transport_container_unregister(&i->t.target_attrs);
2323 transport_container_unregister(&i->t.host_attrs);
2324 transport_container_unregister(&i->rport_attr_cont);
2325 transport_container_unregister(&i->vport_attr_cont);
2326
2327 kfree(i);
2328}
2329EXPORT_SYMBOL(fc_release_transport);
2330
2331/**
2332 * fc_queue_work - Queue work to the fc_host workqueue.
2333 * @shost: Pointer to Scsi_Host bound to fc_host.
2334 * @work: Work to queue for execution.
2335 *
2336 * Return value:
2337 * 1 - work queued for execution
2338 * 0 - work is already queued
2339 * -EINVAL - work queue doesn't exist
2340 */
2341static int
2342fc_queue_work(struct Scsi_Host *shost, struct work_struct *work)
2343{
2344 if (unlikely(!fc_host_work_q(shost))) {
2345 printk(KERN_ERR
2346 "ERROR: FC host '%s' attempted to queue work, "
2347 "when no workqueue created.\n", shost->hostt->name);
2348 dump_stack();
2349
2350 return -EINVAL;
2351 }
2352
2353 return queue_work(fc_host_work_q(shost), work);
2354}
2355
2356/**
2357 * fc_flush_work - Flush a fc_host's workqueue.
2358 * @shost: Pointer to Scsi_Host bound to fc_host.
2359 */
2360static void
2361fc_flush_work(struct Scsi_Host *shost)
2362{
2363 if (!fc_host_work_q(shost)) {
2364 printk(KERN_ERR
2365 "ERROR: FC host '%s' attempted to flush work, "
2366 "when no workqueue created.\n", shost->hostt->name);
2367 dump_stack();
2368 return;
2369 }
2370
2371 flush_workqueue(fc_host_work_q(shost));
2372}
2373
2374/**
2375 * fc_queue_devloss_work - Schedule work for the fc_host devloss workqueue.
2376 * @shost: Pointer to Scsi_Host bound to fc_host.
2377 * @work: Work to queue for execution.
2378 * @delay: jiffies to delay the work queuing
2379 *
2380 * Return value:
2381 * 1 on success / 0 already queued / < 0 for error
2382 */
2383static int
2384fc_queue_devloss_work(struct Scsi_Host *shost, struct delayed_work *work,
2385 unsigned long delay)
2386{
2387 if (unlikely(!fc_host_devloss_work_q(shost))) {
2388 printk(KERN_ERR
2389 "ERROR: FC host '%s' attempted to queue work, "
2390 "when no workqueue created.\n", shost->hostt->name);
2391 dump_stack();
2392
2393 return -EINVAL;
2394 }
2395
2396 return queue_delayed_work(fc_host_devloss_work_q(shost), work, delay);
2397}
2398
2399/**
2400 * fc_flush_devloss - Flush a fc_host's devloss workqueue.
2401 * @shost: Pointer to Scsi_Host bound to fc_host.
2402 */
2403static void
2404fc_flush_devloss(struct Scsi_Host *shost)
2405{
2406 if (!fc_host_devloss_work_q(shost)) {
2407 printk(KERN_ERR
2408 "ERROR: FC host '%s' attempted to flush work, "
2409 "when no workqueue created.\n", shost->hostt->name);
2410 dump_stack();
2411 return;
2412 }
2413
2414 flush_workqueue(fc_host_devloss_work_q(shost));
2415}
2416
2417
2418/**
2419 * fc_remove_host - called to terminate any fc_transport-related elements for a scsi host.
2420 * @shost: Which &Scsi_Host
2421 *
2422 * This routine is expected to be called immediately preceding the
2423 * a driver's call to scsi_remove_host().
2424 *
2425 * WARNING: A driver utilizing the fc_transport, which fails to call
2426 * this routine prior to scsi_remove_host(), will leave dangling
2427 * objects in /sys/class/fc_remote_ports. Access to any of these
2428 * objects can result in a system crash !!!
2429 *
2430 * Notes:
2431 * This routine assumes no locks are held on entry.
2432 */
2433void
2434fc_remove_host(struct Scsi_Host *shost)
2435{
2436 struct fc_vport *vport = NULL, *next_vport = NULL;
2437 struct fc_rport *rport = NULL, *next_rport = NULL;
2438 struct workqueue_struct *work_q;
2439 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
2440 unsigned long flags;
2441
2442 spin_lock_irqsave(shost->host_lock, flags);
2443
2444 /* Remove any vports */
2445 list_for_each_entry_safe(vport, next_vport, &fc_host->vports, peers)
2446 fc_queue_work(shost, &vport->vport_delete_work);
2447
2448 /* Remove any remote ports */
2449 list_for_each_entry_safe(rport, next_rport,
2450 &fc_host->rports, peers) {
2451 list_del(&rport->peers);
2452 rport->port_state = FC_PORTSTATE_DELETED;
2453 fc_queue_work(shost, &rport->rport_delete_work);
2454 }
2455
2456 list_for_each_entry_safe(rport, next_rport,
2457 &fc_host->rport_bindings, peers) {
2458 list_del(&rport->peers);
2459 rport->port_state = FC_PORTSTATE_DELETED;
2460 fc_queue_work(shost, &rport->rport_delete_work);
2461 }
2462
2463 spin_unlock_irqrestore(shost->host_lock, flags);
2464
2465 /* flush all scan work items */
2466 scsi_flush_work(shost);
2467
2468 /* flush all stgt delete, and rport delete work items, then kill it */
2469 if (fc_host->work_q) {
2470 work_q = fc_host->work_q;
2471 fc_host->work_q = NULL;
2472 destroy_workqueue(work_q);
2473 }
2474
2475 /* flush all devloss work items, then kill it */
2476 if (fc_host->devloss_work_q) {
2477 work_q = fc_host->devloss_work_q;
2478 fc_host->devloss_work_q = NULL;
2479 destroy_workqueue(work_q);
2480 }
2481}
2482EXPORT_SYMBOL(fc_remove_host);
2483
2484static void fc_terminate_rport_io(struct fc_rport *rport)
2485{
2486 struct Scsi_Host *shost = rport_to_shost(rport);
2487 struct fc_internal *i = to_fc_internal(shost->transportt);
2488
2489 /* Involve the LLDD if possible to terminate all io on the rport. */
2490 if (i->f->terminate_rport_io)
2491 i->f->terminate_rport_io(rport);
2492
2493 /*
2494 * Must unblock to flush queued IO. scsi-ml will fail incoming reqs.
2495 */
2496 scsi_target_unblock(&rport->dev, SDEV_TRANSPORT_OFFLINE);
2497}
2498
2499/**
2500 * fc_starget_delete - called to delete the scsi descendants of an rport
2501 * @work: remote port to be operated on.
2502 *
2503 * Deletes target and all sdevs.
2504 */
2505static void
2506fc_starget_delete(struct work_struct *work)
2507{
2508 struct fc_rport *rport =
2509 container_of(work, struct fc_rport, stgt_delete_work);
2510
2511 fc_terminate_rport_io(rport);
2512 scsi_remove_target(&rport->dev);
2513}
2514
2515
2516/**
2517 * fc_rport_final_delete - finish rport termination and delete it.
2518 * @work: remote port to be deleted.
2519 */
2520static void
2521fc_rport_final_delete(struct work_struct *work)
2522{
2523 struct fc_rport *rport =
2524 container_of(work, struct fc_rport, rport_delete_work);
2525 struct device *dev = &rport->dev;
2526 struct Scsi_Host *shost = rport_to_shost(rport);
2527 struct fc_internal *i = to_fc_internal(shost->transportt);
2528 unsigned long flags;
2529 int do_callback = 0;
2530
2531 fc_terminate_rport_io(rport);
2532
2533 /*
2534 * if a scan is pending, flush the SCSI Host work_q so that
2535 * that we can reclaim the rport scan work element.
2536 */
2537 if (rport->flags & FC_RPORT_SCAN_PENDING)
2538 scsi_flush_work(shost);
2539
2540 /*
2541 * Cancel any outstanding timers. These should really exist
2542 * only when rmmod'ing the LLDD and we're asking for
2543 * immediate termination of the rports
2544 */
2545 spin_lock_irqsave(shost->host_lock, flags);
2546 if (rport->flags & FC_RPORT_DEVLOSS_PENDING) {
2547 spin_unlock_irqrestore(shost->host_lock, flags);
2548 if (!cancel_delayed_work(&rport->fail_io_work))
2549 fc_flush_devloss(shost);
2550 if (!cancel_delayed_work(&rport->dev_loss_work))
2551 fc_flush_devloss(shost);
2552 spin_lock_irqsave(shost->host_lock, flags);
2553 rport->flags &= ~FC_RPORT_DEVLOSS_PENDING;
2554 }
2555 spin_unlock_irqrestore(shost->host_lock, flags);
2556
2557 /* Delete SCSI target and sdevs */
2558 if (rport->scsi_target_id != -1)
2559 fc_starget_delete(&rport->stgt_delete_work);
2560
2561 /*
2562 * Notify the driver that the rport is now dead. The LLDD will
2563 * also guarantee that any communication to the rport is terminated
2564 *
2565 * Avoid this call if we already called it when we preserved the
2566 * rport for the binding.
2567 */
2568 spin_lock_irqsave(shost->host_lock, flags);
2569 if (!(rport->flags & FC_RPORT_DEVLOSS_CALLBK_DONE) &&
2570 (i->f->dev_loss_tmo_callbk)) {
2571 rport->flags |= FC_RPORT_DEVLOSS_CALLBK_DONE;
2572 do_callback = 1;
2573 }
2574 spin_unlock_irqrestore(shost->host_lock, flags);
2575
2576 if (do_callback)
2577 i->f->dev_loss_tmo_callbk(rport);
2578
2579 fc_bsg_remove(rport->rqst_q);
2580
2581 transport_remove_device(dev);
2582 device_del(dev);
2583 transport_destroy_device(dev);
2584 put_device(&shost->shost_gendev); /* for fc_host->rport list */
2585 put_device(dev); /* for self-reference */
2586}
2587
2588
2589/**
2590 * fc_rport_create - allocates and creates a remote FC port.
2591 * @shost: scsi host the remote port is connected to.
2592 * @channel: Channel on shost port connected to.
2593 * @ids: The world wide names, fc address, and FC4 port
2594 * roles for the remote port.
2595 *
2596 * Allocates and creates the remoter port structure, including the
2597 * class and sysfs creation.
2598 *
2599 * Notes:
2600 * This routine assumes no locks are held on entry.
2601 */
2602static struct fc_rport *
2603fc_rport_create(struct Scsi_Host *shost, int channel,
2604 struct fc_rport_identifiers *ids)
2605{
2606 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
2607 struct fc_internal *fci = to_fc_internal(shost->transportt);
2608 struct fc_rport *rport;
2609 struct device *dev;
2610 unsigned long flags;
2611 int error;
2612 size_t size;
2613
2614 size = (sizeof(struct fc_rport) + fci->f->dd_fcrport_size);
2615 rport = kzalloc(size, GFP_KERNEL);
2616 if (unlikely(!rport)) {
2617 printk(KERN_ERR "%s: allocation failure\n", __func__);
2618 return NULL;
2619 }
2620
2621 rport->maxframe_size = -1;
2622 rport->supported_classes = FC_COS_UNSPECIFIED;
2623 rport->dev_loss_tmo = fc_host->dev_loss_tmo;
2624 memcpy(&rport->node_name, &ids->node_name, sizeof(rport->node_name));
2625 memcpy(&rport->port_name, &ids->port_name, sizeof(rport->port_name));
2626 rport->port_id = ids->port_id;
2627 rport->roles = ids->roles;
2628 rport->port_state = FC_PORTSTATE_ONLINE;
2629 if (fci->f->dd_fcrport_size)
2630 rport->dd_data = &rport[1];
2631 rport->channel = channel;
2632 rport->fast_io_fail_tmo = -1;
2633
2634 INIT_DELAYED_WORK(&rport->dev_loss_work, fc_timeout_deleted_rport);
2635 INIT_DELAYED_WORK(&rport->fail_io_work, fc_timeout_fail_rport_io);
2636 INIT_WORK(&rport->scan_work, fc_scsi_scan_rport);
2637 INIT_WORK(&rport->stgt_delete_work, fc_starget_delete);
2638 INIT_WORK(&rport->rport_delete_work, fc_rport_final_delete);
2639
2640 spin_lock_irqsave(shost->host_lock, flags);
2641
2642 rport->number = fc_host->next_rport_number++;
2643 if (rport->roles & FC_PORT_ROLE_FCP_TARGET)
2644 rport->scsi_target_id = fc_host->next_target_id++;
2645 else
2646 rport->scsi_target_id = -1;
2647 list_add_tail(&rport->peers, &fc_host->rports);
2648 get_device(&shost->shost_gendev); /* for fc_host->rport list */
2649
2650 spin_unlock_irqrestore(shost->host_lock, flags);
2651
2652 dev = &rport->dev;
2653 device_initialize(dev); /* takes self reference */
2654 dev->parent = get_device(&shost->shost_gendev); /* parent reference */
2655 dev->release = fc_rport_dev_release;
2656 dev_set_name(dev, "rport-%d:%d-%d",
2657 shost->host_no, channel, rport->number);
2658 transport_setup_device(dev);
2659
2660 error = device_add(dev);
2661 if (error) {
2662 printk(KERN_ERR "FC Remote Port device_add failed\n");
2663 goto delete_rport;
2664 }
2665 transport_add_device(dev);
2666 transport_configure_device(dev);
2667
2668 fc_bsg_rportadd(shost, rport);
2669 /* ignore any bsg add error - we just can't do sgio */
2670
2671 if (rport->roles & FC_PORT_ROLE_FCP_TARGET) {
2672 /* initiate a scan of the target */
2673 rport->flags |= FC_RPORT_SCAN_PENDING;
2674 scsi_queue_work(shost, &rport->scan_work);
2675 }
2676
2677 return rport;
2678
2679delete_rport:
2680 transport_destroy_device(dev);
2681 spin_lock_irqsave(shost->host_lock, flags);
2682 list_del(&rport->peers);
2683 put_device(&shost->shost_gendev); /* for fc_host->rport list */
2684 spin_unlock_irqrestore(shost->host_lock, flags);
2685 put_device(dev->parent);
2686 kfree(rport);
2687 return NULL;
2688}
2689
2690/**
2691 * fc_remote_port_add - notify fc transport of the existence of a remote FC port.
2692 * @shost: scsi host the remote port is connected to.
2693 * @channel: Channel on shost port connected to.
2694 * @ids: The world wide names, fc address, and FC4 port
2695 * roles for the remote port.
2696 *
2697 * The LLDD calls this routine to notify the transport of the existence
2698 * of a remote port. The LLDD provides the unique identifiers (wwpn,wwn)
2699 * of the port, it's FC address (port_id), and the FC4 roles that are
2700 * active for the port.
2701 *
2702 * For ports that are FCP targets (aka scsi targets), the FC transport
2703 * maintains consistent target id bindings on behalf of the LLDD.
2704 * A consistent target id binding is an assignment of a target id to
2705 * a remote port identifier, which persists while the scsi host is
2706 * attached. The remote port can disappear, then later reappear, and
2707 * it's target id assignment remains the same. This allows for shifts
2708 * in FC addressing (if binding by wwpn or wwnn) with no apparent
2709 * changes to the scsi subsystem which is based on scsi host number and
2710 * target id values. Bindings are only valid during the attachment of
2711 * the scsi host. If the host detaches, then later re-attaches, target
2712 * id bindings may change.
2713 *
2714 * This routine is responsible for returning a remote port structure.
2715 * The routine will search the list of remote ports it maintains
2716 * internally on behalf of consistent target id mappings. If found, the
2717 * remote port structure will be reused. Otherwise, a new remote port
2718 * structure will be allocated.
2719 *
2720 * Whenever a remote port is allocated, a new fc_remote_port class
2721 * device is created.
2722 *
2723 * Should not be called from interrupt context.
2724 *
2725 * Notes:
2726 * This routine assumes no locks are held on entry.
2727 */
2728struct fc_rport *
2729fc_remote_port_add(struct Scsi_Host *shost, int channel,
2730 struct fc_rport_identifiers *ids)
2731{
2732 struct fc_internal *fci = to_fc_internal(shost->transportt);
2733 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
2734 struct fc_rport *rport;
2735 unsigned long flags;
2736 int match = 0;
2737
2738 /* ensure any stgt delete functions are done */
2739 fc_flush_work(shost);
2740
2741 /*
2742 * Search the list of "active" rports, for an rport that has been
2743 * deleted, but we've held off the real delete while the target
2744 * is in a "blocked" state.
2745 */
2746 spin_lock_irqsave(shost->host_lock, flags);
2747
2748 list_for_each_entry(rport, &fc_host->rports, peers) {
2749
2750 if ((rport->port_state == FC_PORTSTATE_BLOCKED) &&
2751 (rport->channel == channel)) {
2752
2753 switch (fc_host->tgtid_bind_type) {
2754 case FC_TGTID_BIND_BY_WWPN:
2755 case FC_TGTID_BIND_NONE:
2756 if (rport->port_name == ids->port_name)
2757 match = 1;
2758 break;
2759 case FC_TGTID_BIND_BY_WWNN:
2760 if (rport->node_name == ids->node_name)
2761 match = 1;
2762 break;
2763 case FC_TGTID_BIND_BY_ID:
2764 if (rport->port_id == ids->port_id)
2765 match = 1;
2766 break;
2767 }
2768
2769 if (match) {
2770
2771 memcpy(&rport->node_name, &ids->node_name,
2772 sizeof(rport->node_name));
2773 memcpy(&rport->port_name, &ids->port_name,
2774 sizeof(rport->port_name));
2775 rport->port_id = ids->port_id;
2776
2777 rport->port_state = FC_PORTSTATE_ONLINE;
2778 rport->roles = ids->roles;
2779
2780 spin_unlock_irqrestore(shost->host_lock, flags);
2781
2782 if (fci->f->dd_fcrport_size)
2783 memset(rport->dd_data, 0,
2784 fci->f->dd_fcrport_size);
2785
2786 /*
2787 * If we were not a target, cancel the
2788 * io terminate and rport timers, and
2789 * we're done.
2790 *
2791 * If we were a target, but our new role
2792 * doesn't indicate a target, leave the
2793 * timers running expecting the role to
2794 * change as the target fully logs in. If
2795 * it doesn't, the target will be torn down.
2796 *
2797 * If we were a target, and our role shows
2798 * we're still a target, cancel the timers
2799 * and kick off a scan.
2800 */
2801
2802 /* was a target, not in roles */
2803 if ((rport->scsi_target_id != -1) &&
2804 (!(ids->roles & FC_PORT_ROLE_FCP_TARGET)))
2805 return rport;
2806
2807 /*
2808 * Stop the fail io and dev_loss timers.
2809 * If they flush, the port_state will
2810 * be checked and will NOOP the function.
2811 */
2812 if (!cancel_delayed_work(&rport->fail_io_work))
2813 fc_flush_devloss(shost);
2814 if (!cancel_delayed_work(&rport->dev_loss_work))
2815 fc_flush_devloss(shost);
2816
2817 spin_lock_irqsave(shost->host_lock, flags);
2818
2819 rport->flags &= ~(FC_RPORT_FAST_FAIL_TIMEDOUT |
2820 FC_RPORT_DEVLOSS_PENDING |
2821 FC_RPORT_DEVLOSS_CALLBK_DONE);
2822
2823 spin_unlock_irqrestore(shost->host_lock, flags);
2824
2825 /* if target, initiate a scan */
2826 if (rport->scsi_target_id != -1) {
2827 scsi_target_unblock(&rport->dev,
2828 SDEV_RUNNING);
2829 spin_lock_irqsave(shost->host_lock,
2830 flags);
2831 rport->flags |= FC_RPORT_SCAN_PENDING;
2832 scsi_queue_work(shost,
2833 &rport->scan_work);
2834 spin_unlock_irqrestore(shost->host_lock,
2835 flags);
2836 }
2837
2838 fc_bsg_goose_queue(rport);
2839
2840 return rport;
2841 }
2842 }
2843 }
2844
2845 /*
2846 * Search the bindings array
2847 * Note: if never a FCP target, you won't be on this list
2848 */
2849 if (fc_host->tgtid_bind_type != FC_TGTID_BIND_NONE) {
2850
2851 /* search for a matching consistent binding */
2852
2853 list_for_each_entry(rport, &fc_host->rport_bindings,
2854 peers) {
2855 if (rport->channel != channel)
2856 continue;
2857
2858 switch (fc_host->tgtid_bind_type) {
2859 case FC_TGTID_BIND_BY_WWPN:
2860 if (rport->port_name == ids->port_name)
2861 match = 1;
2862 break;
2863 case FC_TGTID_BIND_BY_WWNN:
2864 if (rport->node_name == ids->node_name)
2865 match = 1;
2866 break;
2867 case FC_TGTID_BIND_BY_ID:
2868 if (rport->port_id == ids->port_id)
2869 match = 1;
2870 break;
2871 case FC_TGTID_BIND_NONE: /* to keep compiler happy */
2872 break;
2873 }
2874
2875 if (match) {
2876 list_move_tail(&rport->peers, &fc_host->rports);
2877 break;
2878 }
2879 }
2880
2881 if (match) {
2882 memcpy(&rport->node_name, &ids->node_name,
2883 sizeof(rport->node_name));
2884 memcpy(&rport->port_name, &ids->port_name,
2885 sizeof(rport->port_name));
2886 rport->port_id = ids->port_id;
2887 rport->roles = ids->roles;
2888 rport->port_state = FC_PORTSTATE_ONLINE;
2889 rport->flags &= ~FC_RPORT_FAST_FAIL_TIMEDOUT;
2890
2891 if (fci->f->dd_fcrport_size)
2892 memset(rport->dd_data, 0,
2893 fci->f->dd_fcrport_size);
2894 spin_unlock_irqrestore(shost->host_lock, flags);
2895
2896 if (ids->roles & FC_PORT_ROLE_FCP_TARGET) {
2897 scsi_target_unblock(&rport->dev, SDEV_RUNNING);
2898
2899 /* initiate a scan of the target */
2900 spin_lock_irqsave(shost->host_lock, flags);
2901 rport->flags |= FC_RPORT_SCAN_PENDING;
2902 scsi_queue_work(shost, &rport->scan_work);
2903 spin_unlock_irqrestore(shost->host_lock, flags);
2904 }
2905 return rport;
2906 }
2907 }
2908
2909 spin_unlock_irqrestore(shost->host_lock, flags);
2910
2911 /* No consistent binding found - create new remote port entry */
2912 rport = fc_rport_create(shost, channel, ids);
2913
2914 return rport;
2915}
2916EXPORT_SYMBOL(fc_remote_port_add);
2917
2918
2919/**
2920 * fc_remote_port_delete - notifies the fc transport that a remote port is no longer in existence.
2921 * @rport: The remote port that no longer exists
2922 *
2923 * The LLDD calls this routine to notify the transport that a remote
2924 * port is no longer part of the topology. Note: Although a port
2925 * may no longer be part of the topology, it may persist in the remote
2926 * ports displayed by the fc_host. We do this under 2 conditions:
2927 * 1) If the port was a scsi target, we delay its deletion by "blocking" it.
2928 * This allows the port to temporarily disappear, then reappear without
2929 * disrupting the SCSI device tree attached to it. During the "blocked"
2930 * period the port will still exist.
2931 * 2) If the port was a scsi target and disappears for longer than we
2932 * expect, we'll delete the port and the tear down the SCSI device tree
2933 * attached to it. However, we want to semi-persist the target id assigned
2934 * to that port if it eventually does exist. The port structure will
2935 * remain (although with minimal information) so that the target id
2936 * bindings remails.
2937 *
2938 * If the remote port is not an FCP Target, it will be fully torn down
2939 * and deallocated, including the fc_remote_port class device.
2940 *
2941 * If the remote port is an FCP Target, the port will be placed in a
2942 * temporary blocked state. From the LLDD's perspective, the rport no
2943 * longer exists. From the SCSI midlayer's perspective, the SCSI target
2944 * exists, but all sdevs on it are blocked from further I/O. The following
2945 * is then expected.
2946 *
2947 * If the remote port does not return (signaled by a LLDD call to
2948 * fc_remote_port_add()) within the dev_loss_tmo timeout, then the
2949 * scsi target is removed - killing all outstanding i/o and removing the
2950 * scsi devices attached ot it. The port structure will be marked Not
2951 * Present and be partially cleared, leaving only enough information to
2952 * recognize the remote port relative to the scsi target id binding if
2953 * it later appears. The port will remain as long as there is a valid
2954 * binding (e.g. until the user changes the binding type or unloads the
2955 * scsi host with the binding).
2956 *
2957 * If the remote port returns within the dev_loss_tmo value (and matches
2958 * according to the target id binding type), the port structure will be
2959 * reused. If it is no longer a SCSI target, the target will be torn
2960 * down. If it continues to be a SCSI target, then the target will be
2961 * unblocked (allowing i/o to be resumed), and a scan will be activated
2962 * to ensure that all luns are detected.
2963 *
2964 * Called from normal process context only - cannot be called from interrupt.
2965 *
2966 * Notes:
2967 * This routine assumes no locks are held on entry.
2968 */
2969void
2970fc_remote_port_delete(struct fc_rport *rport)
2971{
2972 struct Scsi_Host *shost = rport_to_shost(rport);
2973 unsigned long timeout = rport->dev_loss_tmo;
2974 unsigned long flags;
2975
2976 /*
2977 * No need to flush the fc_host work_q's, as all adds are synchronous.
2978 *
2979 * We do need to reclaim the rport scan work element, so eventually
2980 * (in fc_rport_final_delete()) we'll flush the scsi host work_q if
2981 * there's still a scan pending.
2982 */
2983
2984 spin_lock_irqsave(shost->host_lock, flags);
2985
2986 if (rport->port_state != FC_PORTSTATE_ONLINE) {
2987 spin_unlock_irqrestore(shost->host_lock, flags);
2988 return;
2989 }
2990
2991 /*
2992 * In the past, we if this was not an FCP-Target, we would
2993 * unconditionally just jump to deleting the rport.
2994 * However, rports can be used as node containers by the LLDD,
2995 * and its not appropriate to just terminate the rport at the
2996 * first sign of a loss in connectivity. The LLDD may want to
2997 * send ELS traffic to re-validate the login. If the rport is
2998 * immediately deleted, it makes it inappropriate for a node
2999 * container.
3000 * So... we now unconditionally wait dev_loss_tmo before
3001 * destroying an rport.
3002 */
3003
3004 rport->port_state = FC_PORTSTATE_BLOCKED;
3005
3006 rport->flags |= FC_RPORT_DEVLOSS_PENDING;
3007
3008 spin_unlock_irqrestore(shost->host_lock, flags);
3009
3010 if (rport->roles & FC_PORT_ROLE_FCP_INITIATOR &&
3011 shost->active_mode & MODE_TARGET)
3012 fc_tgt_it_nexus_destroy(shost, (unsigned long)rport);
3013
3014 scsi_target_block(&rport->dev);
3015
3016 /* see if we need to kill io faster than waiting for device loss */
3017 if ((rport->fast_io_fail_tmo != -1) &&
3018 (rport->fast_io_fail_tmo < timeout))
3019 fc_queue_devloss_work(shost, &rport->fail_io_work,
3020 rport->fast_io_fail_tmo * HZ);
3021
3022 /* cap the length the devices can be blocked until they are deleted */
3023 fc_queue_devloss_work(shost, &rport->dev_loss_work, timeout * HZ);
3024}
3025EXPORT_SYMBOL(fc_remote_port_delete);
3026
3027/**
3028 * fc_remote_port_rolechg - notifies the fc transport that the roles on a remote may have changed.
3029 * @rport: The remote port that changed.
3030 * @roles: New roles for this port.
3031 *
3032 * Description: The LLDD calls this routine to notify the transport that the
3033 * roles on a remote port may have changed. The largest effect of this is
3034 * if a port now becomes a FCP Target, it must be allocated a
3035 * scsi target id. If the port is no longer a FCP target, any
3036 * scsi target id value assigned to it will persist in case the
3037 * role changes back to include FCP Target. No changes in the scsi
3038 * midlayer will be invoked if the role changes (in the expectation
3039 * that the role will be resumed. If it doesn't normal error processing
3040 * will take place).
3041 *
3042 * Should not be called from interrupt context.
3043 *
3044 * Notes:
3045 * This routine assumes no locks are held on entry.
3046 */
3047void
3048fc_remote_port_rolechg(struct fc_rport *rport, u32 roles)
3049{
3050 struct Scsi_Host *shost = rport_to_shost(rport);
3051 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
3052 unsigned long flags;
3053 int create = 0;
3054 int ret;
3055
3056 spin_lock_irqsave(shost->host_lock, flags);
3057 if (roles & FC_PORT_ROLE_FCP_TARGET) {
3058 if (rport->scsi_target_id == -1) {
3059 rport->scsi_target_id = fc_host->next_target_id++;
3060 create = 1;
3061 } else if (!(rport->roles & FC_PORT_ROLE_FCP_TARGET))
3062 create = 1;
3063 } else if (shost->active_mode & MODE_TARGET) {
3064 ret = fc_tgt_it_nexus_create(shost, (unsigned long)rport,
3065 (char *)&rport->node_name);
3066 if (ret)
3067 printk(KERN_ERR "FC Remore Port tgt nexus failed %d\n",
3068 ret);
3069 }
3070
3071 rport->roles = roles;
3072
3073 spin_unlock_irqrestore(shost->host_lock, flags);
3074
3075 if (create) {
3076 /*
3077 * There may have been a delete timer running on the
3078 * port. Ensure that it is cancelled as we now know
3079 * the port is an FCP Target.
3080 * Note: we know the rport is exists and in an online
3081 * state as the LLDD would not have had an rport
3082 * reference to pass us.
3083 *
3084 * Take no action on the del_timer failure as the state
3085 * machine state change will validate the
3086 * transaction.
3087 */
3088 if (!cancel_delayed_work(&rport->fail_io_work))
3089 fc_flush_devloss(shost);
3090 if (!cancel_delayed_work(&rport->dev_loss_work))
3091 fc_flush_devloss(shost);
3092
3093 spin_lock_irqsave(shost->host_lock, flags);
3094 rport->flags &= ~(FC_RPORT_FAST_FAIL_TIMEDOUT |
3095 FC_RPORT_DEVLOSS_PENDING |
3096 FC_RPORT_DEVLOSS_CALLBK_DONE);
3097 spin_unlock_irqrestore(shost->host_lock, flags);
3098
3099 /* ensure any stgt delete functions are done */
3100 fc_flush_work(shost);
3101
3102 scsi_target_unblock(&rport->dev, SDEV_RUNNING);
3103 /* initiate a scan of the target */
3104 spin_lock_irqsave(shost->host_lock, flags);
3105 rport->flags |= FC_RPORT_SCAN_PENDING;
3106 scsi_queue_work(shost, &rport->scan_work);
3107 spin_unlock_irqrestore(shost->host_lock, flags);
3108 }
3109}
3110EXPORT_SYMBOL(fc_remote_port_rolechg);
3111
3112/**
3113 * fc_timeout_deleted_rport - Timeout handler for a deleted remote port.
3114 * @work: rport target that failed to reappear in the allotted time.
3115 *
3116 * Description: An attempt to delete a remote port blocks, and if it fails
3117 * to return in the allotted time this gets called.
3118 */
3119static void
3120fc_timeout_deleted_rport(struct work_struct *work)
3121{
3122 struct fc_rport *rport =
3123 container_of(work, struct fc_rport, dev_loss_work.work);
3124 struct Scsi_Host *shost = rport_to_shost(rport);
3125 struct fc_internal *i = to_fc_internal(shost->transportt);
3126 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
3127 unsigned long flags;
3128 int do_callback = 0;
3129
3130 spin_lock_irqsave(shost->host_lock, flags);
3131
3132 rport->flags &= ~FC_RPORT_DEVLOSS_PENDING;
3133
3134 /*
3135 * If the port is ONLINE, then it came back. If it was a SCSI
3136 * target, validate it still is. If not, tear down the
3137 * scsi_target on it.
3138 */
3139 if ((rport->port_state == FC_PORTSTATE_ONLINE) &&
3140 (rport->scsi_target_id != -1) &&
3141 !(rport->roles & FC_PORT_ROLE_FCP_TARGET)) {
3142 dev_printk(KERN_ERR, &rport->dev,
3143 "blocked FC remote port time out: no longer"
3144 " a FCP target, removing starget\n");
3145 spin_unlock_irqrestore(shost->host_lock, flags);
3146 scsi_target_unblock(&rport->dev, SDEV_TRANSPORT_OFFLINE);
3147 fc_queue_work(shost, &rport->stgt_delete_work);
3148 return;
3149 }
3150
3151 /* NOOP state - we're flushing workq's */
3152 if (rport->port_state != FC_PORTSTATE_BLOCKED) {
3153 spin_unlock_irqrestore(shost->host_lock, flags);
3154 dev_printk(KERN_ERR, &rport->dev,
3155 "blocked FC remote port time out: leaving"
3156 " rport%s alone\n",
3157 (rport->scsi_target_id != -1) ? " and starget" : "");
3158 return;
3159 }
3160
3161 if ((fc_host->tgtid_bind_type == FC_TGTID_BIND_NONE) ||
3162 (rport->scsi_target_id == -1)) {
3163 list_del(&rport->peers);
3164 rport->port_state = FC_PORTSTATE_DELETED;
3165 dev_printk(KERN_ERR, &rport->dev,
3166 "blocked FC remote port time out: removing"
3167 " rport%s\n",
3168 (rport->scsi_target_id != -1) ? " and starget" : "");
3169 fc_queue_work(shost, &rport->rport_delete_work);
3170 spin_unlock_irqrestore(shost->host_lock, flags);
3171 return;
3172 }
3173
3174 dev_printk(KERN_ERR, &rport->dev,
3175 "blocked FC remote port time out: removing target and "
3176 "saving binding\n");
3177
3178 list_move_tail(&rport->peers, &fc_host->rport_bindings);
3179
3180 /*
3181 * Note: We do not remove or clear the hostdata area. This allows
3182 * host-specific target data to persist along with the
3183 * scsi_target_id. It's up to the host to manage it's hostdata area.
3184 */
3185
3186 /*
3187 * Reinitialize port attributes that may change if the port comes back.
3188 */
3189 rport->maxframe_size = -1;
3190 rport->supported_classes = FC_COS_UNSPECIFIED;
3191 rport->roles = FC_PORT_ROLE_UNKNOWN;
3192 rport->port_state = FC_PORTSTATE_NOTPRESENT;
3193 rport->flags &= ~FC_RPORT_FAST_FAIL_TIMEDOUT;
3194
3195 /*
3196 * Pre-emptively kill I/O rather than waiting for the work queue
3197 * item to teardown the starget. (FCOE libFC folks prefer this
3198 * and to have the rport_port_id still set when it's done).
3199 */
3200 spin_unlock_irqrestore(shost->host_lock, flags);
3201 fc_terminate_rport_io(rport);
3202
3203 spin_lock_irqsave(shost->host_lock, flags);
3204
3205 if (rport->port_state == FC_PORTSTATE_NOTPRESENT) { /* still missing */
3206
3207 /* remove the identifiers that aren't used in the consisting binding */
3208 switch (fc_host->tgtid_bind_type) {
3209 case FC_TGTID_BIND_BY_WWPN:
3210 rport->node_name = -1;
3211 rport->port_id = -1;
3212 break;
3213 case FC_TGTID_BIND_BY_WWNN:
3214 rport->port_name = -1;
3215 rport->port_id = -1;
3216 break;
3217 case FC_TGTID_BIND_BY_ID:
3218 rport->node_name = -1;
3219 rport->port_name = -1;
3220 break;
3221 case FC_TGTID_BIND_NONE: /* to keep compiler happy */
3222 break;
3223 }
3224
3225 /*
3226 * As this only occurs if the remote port (scsi target)
3227 * went away and didn't come back - we'll remove
3228 * all attached scsi devices.
3229 */
3230 rport->flags |= FC_RPORT_DEVLOSS_CALLBK_DONE;
3231 fc_queue_work(shost, &rport->stgt_delete_work);
3232
3233 do_callback = 1;
3234 }
3235
3236 spin_unlock_irqrestore(shost->host_lock, flags);
3237
3238 /*
3239 * Notify the driver that the rport is now dead. The LLDD will
3240 * also guarantee that any communication to the rport is terminated
3241 *
3242 * Note: we set the CALLBK_DONE flag above to correspond
3243 */
3244 if (do_callback && i->f->dev_loss_tmo_callbk)
3245 i->f->dev_loss_tmo_callbk(rport);
3246}
3247
3248
3249/**
3250 * fc_timeout_fail_rport_io - Timeout handler for a fast io failing on a disconnected SCSI target.
3251 * @work: rport to terminate io on.
3252 *
3253 * Notes: Only requests the failure of the io, not that all are flushed
3254 * prior to returning.
3255 */
3256static void
3257fc_timeout_fail_rport_io(struct work_struct *work)
3258{
3259 struct fc_rport *rport =
3260 container_of(work, struct fc_rport, fail_io_work.work);
3261
3262 if (rport->port_state != FC_PORTSTATE_BLOCKED)
3263 return;
3264
3265 rport->flags |= FC_RPORT_FAST_FAIL_TIMEDOUT;
3266 fc_terminate_rport_io(rport);
3267}
3268
3269/**
3270 * fc_scsi_scan_rport - called to perform a scsi scan on a remote port.
3271 * @work: remote port to be scanned.
3272 */
3273static void
3274fc_scsi_scan_rport(struct work_struct *work)
3275{
3276 struct fc_rport *rport =
3277 container_of(work, struct fc_rport, scan_work);
3278 struct Scsi_Host *shost = rport_to_shost(rport);
3279 struct fc_internal *i = to_fc_internal(shost->transportt);
3280 unsigned long flags;
3281
3282 if ((rport->port_state == FC_PORTSTATE_ONLINE) &&
3283 (rport->roles & FC_PORT_ROLE_FCP_TARGET) &&
3284 !(i->f->disable_target_scan)) {
3285 scsi_scan_target(&rport->dev, rport->channel,
3286 rport->scsi_target_id, SCAN_WILD_CARD, 1);
3287 }
3288
3289 spin_lock_irqsave(shost->host_lock, flags);
3290 rport->flags &= ~FC_RPORT_SCAN_PENDING;
3291 spin_unlock_irqrestore(shost->host_lock, flags);
3292}
3293
3294/**
3295 * fc_block_scsi_eh - Block SCSI eh thread for blocked fc_rport
3296 * @cmnd: SCSI command that scsi_eh is trying to recover
3297 *
3298 * This routine can be called from a FC LLD scsi_eh callback. It
3299 * blocks the scsi_eh thread until the fc_rport leaves the
3300 * FC_PORTSTATE_BLOCKED, or the fast_io_fail_tmo fires. This is
3301 * necessary to avoid the scsi_eh failing recovery actions for blocked
3302 * rports which would lead to offlined SCSI devices.
3303 *
3304 * Returns: 0 if the fc_rport left the state FC_PORTSTATE_BLOCKED.
3305 * FAST_IO_FAIL if the fast_io_fail_tmo fired, this should be
3306 * passed back to scsi_eh.
3307 */
3308int fc_block_scsi_eh(struct scsi_cmnd *cmnd)
3309{
3310 struct Scsi_Host *shost = cmnd->device->host;
3311 struct fc_rport *rport = starget_to_rport(scsi_target(cmnd->device));
3312 unsigned long flags;
3313
3314 spin_lock_irqsave(shost->host_lock, flags);
3315 while (rport->port_state == FC_PORTSTATE_BLOCKED &&
3316 !(rport->flags & FC_RPORT_FAST_FAIL_TIMEDOUT)) {
3317 spin_unlock_irqrestore(shost->host_lock, flags);
3318 msleep(1000);
3319 spin_lock_irqsave(shost->host_lock, flags);
3320 }
3321 spin_unlock_irqrestore(shost->host_lock, flags);
3322
3323 if (rport->flags & FC_RPORT_FAST_FAIL_TIMEDOUT)
3324 return FAST_IO_FAIL;
3325
3326 return 0;
3327}
3328EXPORT_SYMBOL(fc_block_scsi_eh);
3329
3330/**
3331 * fc_vport_setup - allocates and creates a FC virtual port.
3332 * @shost: scsi host the virtual port is connected to.
3333 * @channel: Channel on shost port connected to.
3334 * @pdev: parent device for vport
3335 * @ids: The world wide names, FC4 port roles, etc for
3336 * the virtual port.
3337 * @ret_vport: The pointer to the created vport.
3338 *
3339 * Allocates and creates the vport structure, calls the parent host
3340 * to instantiate the vport, the completes w/ class and sysfs creation.
3341 *
3342 * Notes:
3343 * This routine assumes no locks are held on entry.
3344 */
3345static int
3346fc_vport_setup(struct Scsi_Host *shost, int channel, struct device *pdev,
3347 struct fc_vport_identifiers *ids, struct fc_vport **ret_vport)
3348{
3349 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
3350 struct fc_internal *fci = to_fc_internal(shost->transportt);
3351 struct fc_vport *vport;
3352 struct device *dev;
3353 unsigned long flags;
3354 size_t size;
3355 int error;
3356
3357 *ret_vport = NULL;
3358
3359 if ( ! fci->f->vport_create)
3360 return -ENOENT;
3361
3362 size = (sizeof(struct fc_vport) + fci->f->dd_fcvport_size);
3363 vport = kzalloc(size, GFP_KERNEL);
3364 if (unlikely(!vport)) {
3365 printk(KERN_ERR "%s: allocation failure\n", __func__);
3366 return -ENOMEM;
3367 }
3368
3369 vport->vport_state = FC_VPORT_UNKNOWN;
3370 vport->vport_last_state = FC_VPORT_UNKNOWN;
3371 vport->node_name = ids->node_name;
3372 vport->port_name = ids->port_name;
3373 vport->roles = ids->roles;
3374 vport->vport_type = ids->vport_type;
3375 if (fci->f->dd_fcvport_size)
3376 vport->dd_data = &vport[1];
3377 vport->shost = shost;
3378 vport->channel = channel;
3379 vport->flags = FC_VPORT_CREATING;
3380 INIT_WORK(&vport->vport_delete_work, fc_vport_sched_delete);
3381
3382 spin_lock_irqsave(shost->host_lock, flags);
3383
3384 if (fc_host->npiv_vports_inuse >= fc_host->max_npiv_vports) {
3385 spin_unlock_irqrestore(shost->host_lock, flags);
3386 kfree(vport);
3387 return -ENOSPC;
3388 }
3389 fc_host->npiv_vports_inuse++;
3390 vport->number = fc_host->next_vport_number++;
3391 list_add_tail(&vport->peers, &fc_host->vports);
3392 get_device(&shost->shost_gendev); /* for fc_host->vport list */
3393
3394 spin_unlock_irqrestore(shost->host_lock, flags);
3395
3396 dev = &vport->dev;
3397 device_initialize(dev); /* takes self reference */
3398 dev->parent = get_device(pdev); /* takes parent reference */
3399 dev->release = fc_vport_dev_release;
3400 dev_set_name(dev, "vport-%d:%d-%d",
3401 shost->host_no, channel, vport->number);
3402 transport_setup_device(dev);
3403
3404 error = device_add(dev);
3405 if (error) {
3406 printk(KERN_ERR "FC Virtual Port device_add failed\n");
3407 goto delete_vport;
3408 }
3409 transport_add_device(dev);
3410 transport_configure_device(dev);
3411
3412 error = fci->f->vport_create(vport, ids->disable);
3413 if (error) {
3414 printk(KERN_ERR "FC Virtual Port LLDD Create failed\n");
3415 goto delete_vport_all;
3416 }
3417
3418 /*
3419 * if the parent isn't the physical adapter's Scsi_Host, ensure
3420 * the Scsi_Host at least contains ia symlink to the vport.
3421 */
3422 if (pdev != &shost->shost_gendev) {
3423 error = sysfs_create_link(&shost->shost_gendev.kobj,
3424 &dev->kobj, dev_name(dev));
3425 if (error)
3426 printk(KERN_ERR
3427 "%s: Cannot create vport symlinks for "
3428 "%s, err=%d\n",
3429 __func__, dev_name(dev), error);
3430 }
3431 spin_lock_irqsave(shost->host_lock, flags);
3432 vport->flags &= ~FC_VPORT_CREATING;
3433 spin_unlock_irqrestore(shost->host_lock, flags);
3434
3435 dev_printk(KERN_NOTICE, pdev,
3436 "%s created via shost%d channel %d\n", dev_name(dev),
3437 shost->host_no, channel);
3438
3439 *ret_vport = vport;
3440
3441 return 0;
3442
3443delete_vport_all:
3444 transport_remove_device(dev);
3445 device_del(dev);
3446delete_vport:
3447 transport_destroy_device(dev);
3448 spin_lock_irqsave(shost->host_lock, flags);
3449 list_del(&vport->peers);
3450 put_device(&shost->shost_gendev); /* for fc_host->vport list */
3451 fc_host->npiv_vports_inuse--;
3452 spin_unlock_irqrestore(shost->host_lock, flags);
3453 put_device(dev->parent);
3454 kfree(vport);
3455
3456 return error;
3457}
3458
3459/**
3460 * fc_vport_create - Admin App or LLDD requests creation of a vport
3461 * @shost: scsi host the virtual port is connected to.
3462 * @channel: channel on shost port connected to.
3463 * @ids: The world wide names, FC4 port roles, etc for
3464 * the virtual port.
3465 *
3466 * Notes:
3467 * This routine assumes no locks are held on entry.
3468 */
3469struct fc_vport *
3470fc_vport_create(struct Scsi_Host *shost, int channel,
3471 struct fc_vport_identifiers *ids)
3472{
3473 int stat;
3474 struct fc_vport *vport;
3475
3476 stat = fc_vport_setup(shost, channel, &shost->shost_gendev,
3477 ids, &vport);
3478 return stat ? NULL : vport;
3479}
3480EXPORT_SYMBOL(fc_vport_create);
3481
3482/**
3483 * fc_vport_terminate - Admin App or LLDD requests termination of a vport
3484 * @vport: fc_vport to be terminated
3485 *
3486 * Calls the LLDD vport_delete() function, then deallocates and removes
3487 * the vport from the shost and object tree.
3488 *
3489 * Notes:
3490 * This routine assumes no locks are held on entry.
3491 */
3492int
3493fc_vport_terminate(struct fc_vport *vport)
3494{
3495 struct Scsi_Host *shost = vport_to_shost(vport);
3496 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
3497 struct fc_internal *i = to_fc_internal(shost->transportt);
3498 struct device *dev = &vport->dev;
3499 unsigned long flags;
3500 int stat;
3501
3502 if (i->f->vport_delete)
3503 stat = i->f->vport_delete(vport);
3504 else
3505 stat = -ENOENT;
3506
3507 spin_lock_irqsave(shost->host_lock, flags);
3508 vport->flags &= ~FC_VPORT_DELETING;
3509 if (!stat) {
3510 vport->flags |= FC_VPORT_DELETED;
3511 list_del(&vport->peers);
3512 fc_host->npiv_vports_inuse--;
3513 put_device(&shost->shost_gendev); /* for fc_host->vport list */
3514 }
3515 spin_unlock_irqrestore(shost->host_lock, flags);
3516
3517 if (stat)
3518 return stat;
3519
3520 if (dev->parent != &shost->shost_gendev)
3521 sysfs_remove_link(&shost->shost_gendev.kobj, dev_name(dev));
3522 transport_remove_device(dev);
3523 device_del(dev);
3524 transport_destroy_device(dev);
3525
3526 /*
3527 * Removing our self-reference should mean our
3528 * release function gets called, which will drop the remaining
3529 * parent reference and free the data structure.
3530 */
3531 put_device(dev); /* for self-reference */
3532
3533 return 0; /* SUCCESS */
3534}
3535EXPORT_SYMBOL(fc_vport_terminate);
3536
3537/**
3538 * fc_vport_sched_delete - workq-based delete request for a vport
3539 * @work: vport to be deleted.
3540 */
3541static void
3542fc_vport_sched_delete(struct work_struct *work)
3543{
3544 struct fc_vport *vport =
3545 container_of(work, struct fc_vport, vport_delete_work);
3546 int stat;
3547
3548 stat = fc_vport_terminate(vport);
3549 if (stat)
3550 dev_printk(KERN_ERR, vport->dev.parent,
3551 "%s: %s could not be deleted created via "
3552 "shost%d channel %d - error %d\n", __func__,
3553 dev_name(&vport->dev), vport->shost->host_no,
3554 vport->channel, stat);
3555}
3556
3557
3558/*
3559 * BSG support
3560 */
3561
3562
3563/**
3564 * fc_destroy_bsgjob - routine to teardown/delete a fc bsg job
3565 * @job: fc_bsg_job that is to be torn down
3566 */
3567static void
3568fc_destroy_bsgjob(struct fc_bsg_job *job)
3569{
3570 unsigned long flags;
3571
3572 spin_lock_irqsave(&job->job_lock, flags);
3573 if (job->ref_cnt) {
3574 spin_unlock_irqrestore(&job->job_lock, flags);
3575 return;
3576 }
3577 spin_unlock_irqrestore(&job->job_lock, flags);
3578
3579 put_device(job->dev); /* release reference for the request */
3580
3581 kfree(job->request_payload.sg_list);
3582 kfree(job->reply_payload.sg_list);
3583 kfree(job);
3584}
3585
3586/**
3587 * fc_bsg_jobdone - completion routine for bsg requests that the LLD has
3588 * completed
3589 * @job: fc_bsg_job that is complete
3590 */
3591static void
3592fc_bsg_jobdone(struct fc_bsg_job *job)
3593{
3594 struct request *req = job->req;
3595 struct request *rsp = req->next_rq;
3596 int err;
3597
3598 err = job->req->errors = job->reply->result;
3599
3600 if (err < 0)
3601 /* we're only returning the result field in the reply */
3602 job->req->sense_len = sizeof(uint32_t);
3603 else
3604 job->req->sense_len = job->reply_len;
3605
3606 /* we assume all request payload was transferred, residual == 0 */
3607 req->resid_len = 0;
3608
3609 if (rsp) {
3610 WARN_ON(job->reply->reply_payload_rcv_len > rsp->resid_len);
3611
3612 /* set reply (bidi) residual */
3613 rsp->resid_len -= min(job->reply->reply_payload_rcv_len,
3614 rsp->resid_len);
3615 }
3616 blk_complete_request(req);
3617}
3618
3619/**
3620 * fc_bsg_softirq_done - softirq done routine for destroying the bsg requests
3621 * @rq: BSG request that holds the job to be destroyed
3622 */
3623static void fc_bsg_softirq_done(struct request *rq)
3624{
3625 struct fc_bsg_job *job = rq->special;
3626 unsigned long flags;
3627
3628 spin_lock_irqsave(&job->job_lock, flags);
3629 job->state_flags |= FC_RQST_STATE_DONE;
3630 job->ref_cnt--;
3631 spin_unlock_irqrestore(&job->job_lock, flags);
3632
3633 blk_end_request_all(rq, rq->errors);
3634 fc_destroy_bsgjob(job);
3635}
3636
3637/**
3638 * fc_bsg_job_timeout - handler for when a bsg request timesout
3639 * @req: request that timed out
3640 */
3641static enum blk_eh_timer_return
3642fc_bsg_job_timeout(struct request *req)
3643{
3644 struct fc_bsg_job *job = (void *) req->special;
3645 struct Scsi_Host *shost = job->shost;
3646 struct fc_internal *i = to_fc_internal(shost->transportt);
3647 unsigned long flags;
3648 int err = 0, done = 0;
3649
3650 if (job->rport && job->rport->port_state == FC_PORTSTATE_BLOCKED)
3651 return BLK_EH_RESET_TIMER;
3652
3653 spin_lock_irqsave(&job->job_lock, flags);
3654 if (job->state_flags & FC_RQST_STATE_DONE)
3655 done = 1;
3656 else
3657 job->ref_cnt++;
3658 spin_unlock_irqrestore(&job->job_lock, flags);
3659
3660 if (!done && i->f->bsg_timeout) {
3661 /* call LLDD to abort the i/o as it has timed out */
3662 err = i->f->bsg_timeout(job);
3663 if (err == -EAGAIN) {
3664 job->ref_cnt--;
3665 return BLK_EH_RESET_TIMER;
3666 } else if (err)
3667 printk(KERN_ERR "ERROR: FC BSG request timeout - LLD "
3668 "abort failed with status %d\n", err);
3669 }
3670
3671 /* the blk_end_sync_io() doesn't check the error */
3672 if (done)
3673 return BLK_EH_NOT_HANDLED;
3674 else
3675 return BLK_EH_HANDLED;
3676}
3677
3678static int
3679fc_bsg_map_buffer(struct fc_bsg_buffer *buf, struct request *req)
3680{
3681 size_t sz = (sizeof(struct scatterlist) * req->nr_phys_segments);
3682
3683 BUG_ON(!req->nr_phys_segments);
3684
3685 buf->sg_list = kzalloc(sz, GFP_KERNEL);
3686 if (!buf->sg_list)
3687 return -ENOMEM;
3688 sg_init_table(buf->sg_list, req->nr_phys_segments);
3689 buf->sg_cnt = blk_rq_map_sg(req->q, req, buf->sg_list);
3690 buf->payload_len = blk_rq_bytes(req);
3691 return 0;
3692}
3693
3694
3695/**
3696 * fc_req_to_bsgjob - Allocate/create the fc_bsg_job structure for the
3697 * bsg request
3698 * @shost: SCSI Host corresponding to the bsg object
3699 * @rport: (optional) FC Remote Port corresponding to the bsg object
3700 * @req: BSG request that needs a job structure
3701 */
3702static int
3703fc_req_to_bsgjob(struct Scsi_Host *shost, struct fc_rport *rport,
3704 struct request *req)
3705{
3706 struct fc_internal *i = to_fc_internal(shost->transportt);
3707 struct request *rsp = req->next_rq;
3708 struct fc_bsg_job *job;
3709 int ret;
3710
3711 BUG_ON(req->special);
3712
3713 job = kzalloc(sizeof(struct fc_bsg_job) + i->f->dd_bsg_size,
3714 GFP_KERNEL);
3715 if (!job)
3716 return -ENOMEM;
3717
3718 /*
3719 * Note: this is a bit silly.
3720 * The request gets formatted as a SGIO v4 ioctl request, which
3721 * then gets reformatted as a blk request, which then gets
3722 * reformatted as a fc bsg request. And on completion, we have
3723 * to wrap return results such that SGIO v4 thinks it was a scsi
3724 * status. I hope this was all worth it.
3725 */
3726
3727 req->special = job;
3728 job->shost = shost;
3729 job->rport = rport;
3730 job->req = req;
3731 if (i->f->dd_bsg_size)
3732 job->dd_data = (void *)&job[1];
3733 spin_lock_init(&job->job_lock);
3734 job->request = (struct fc_bsg_request *)req->cmd;
3735 job->request_len = req->cmd_len;
3736 job->reply = req->sense;
3737 job->reply_len = SCSI_SENSE_BUFFERSIZE; /* Size of sense buffer
3738 * allocated */
3739 if (req->bio) {
3740 ret = fc_bsg_map_buffer(&job->request_payload, req);
3741 if (ret)
3742 goto failjob_rls_job;
3743 }
3744 if (rsp && rsp->bio) {
3745 ret = fc_bsg_map_buffer(&job->reply_payload, rsp);
3746 if (ret)
3747 goto failjob_rls_rqst_payload;
3748 }
3749 job->job_done = fc_bsg_jobdone;
3750 if (rport)
3751 job->dev = &rport->dev;
3752 else
3753 job->dev = &shost->shost_gendev;
3754 get_device(job->dev); /* take a reference for the request */
3755
3756 job->ref_cnt = 1;
3757
3758 return 0;
3759
3760
3761failjob_rls_rqst_payload:
3762 kfree(job->request_payload.sg_list);
3763failjob_rls_job:
3764 kfree(job);
3765 return -ENOMEM;
3766}
3767
3768
3769enum fc_dispatch_result {
3770 FC_DISPATCH_BREAK, /* on return, q is locked, break from q loop */
3771 FC_DISPATCH_LOCKED, /* on return, q is locked, continue on */
3772 FC_DISPATCH_UNLOCKED, /* on return, q is unlocked, continue on */
3773};
3774
3775
3776/**
3777 * fc_bsg_host_dispatch - process fc host bsg requests and dispatch to LLDD
3778 * @q: fc host request queue
3779 * @shost: scsi host rport attached to
3780 * @job: bsg job to be processed
3781 */
3782static enum fc_dispatch_result
3783fc_bsg_host_dispatch(struct request_queue *q, struct Scsi_Host *shost,
3784 struct fc_bsg_job *job)
3785{
3786 struct fc_internal *i = to_fc_internal(shost->transportt);
3787 int cmdlen = sizeof(uint32_t); /* start with length of msgcode */
3788 int ret;
3789
3790 /* Validate the host command */
3791 switch (job->request->msgcode) {
3792 case FC_BSG_HST_ADD_RPORT:
3793 cmdlen += sizeof(struct fc_bsg_host_add_rport);
3794 break;
3795
3796 case FC_BSG_HST_DEL_RPORT:
3797 cmdlen += sizeof(struct fc_bsg_host_del_rport);
3798 break;
3799
3800 case FC_BSG_HST_ELS_NOLOGIN:
3801 cmdlen += sizeof(struct fc_bsg_host_els);
3802 /* there better be a xmt and rcv payloads */
3803 if ((!job->request_payload.payload_len) ||
3804 (!job->reply_payload.payload_len)) {
3805 ret = -EINVAL;
3806 goto fail_host_msg;
3807 }
3808 break;
3809
3810 case FC_BSG_HST_CT:
3811 cmdlen += sizeof(struct fc_bsg_host_ct);
3812 /* there better be xmt and rcv payloads */
3813 if ((!job->request_payload.payload_len) ||
3814 (!job->reply_payload.payload_len)) {
3815 ret = -EINVAL;
3816 goto fail_host_msg;
3817 }
3818 break;
3819
3820 case FC_BSG_HST_VENDOR:
3821 cmdlen += sizeof(struct fc_bsg_host_vendor);
3822 if ((shost->hostt->vendor_id == 0L) ||
3823 (job->request->rqst_data.h_vendor.vendor_id !=
3824 shost->hostt->vendor_id)) {
3825 ret = -ESRCH;
3826 goto fail_host_msg;
3827 }
3828 break;
3829
3830 default:
3831 ret = -EBADR;
3832 goto fail_host_msg;
3833 }
3834
3835 /* check if we really have all the request data needed */
3836 if (job->request_len < cmdlen) {
3837 ret = -ENOMSG;
3838 goto fail_host_msg;
3839 }
3840
3841 ret = i->f->bsg_request(job);
3842 if (!ret)
3843 return FC_DISPATCH_UNLOCKED;
3844
3845fail_host_msg:
3846 /* return the errno failure code as the only status */
3847 BUG_ON(job->reply_len < sizeof(uint32_t));
3848 job->reply->reply_payload_rcv_len = 0;
3849 job->reply->result = ret;
3850 job->reply_len = sizeof(uint32_t);
3851 fc_bsg_jobdone(job);
3852 return FC_DISPATCH_UNLOCKED;
3853}
3854
3855
3856/*
3857 * fc_bsg_goose_queue - restart rport queue in case it was stopped
3858 * @rport: rport to be restarted
3859 */
3860static void
3861fc_bsg_goose_queue(struct fc_rport *rport)
3862{
3863 if (!rport->rqst_q)
3864 return;
3865
3866 /*
3867 * This get/put dance makes no sense
3868 */
3869 get_device(&rport->dev);
3870 blk_run_queue_async(rport->rqst_q);
3871 put_device(&rport->dev);
3872}
3873
3874/**
3875 * fc_bsg_rport_dispatch - process rport bsg requests and dispatch to LLDD
3876 * @q: rport request queue
3877 * @shost: scsi host rport attached to
3878 * @rport: rport request destined to
3879 * @job: bsg job to be processed
3880 */
3881static enum fc_dispatch_result
3882fc_bsg_rport_dispatch(struct request_queue *q, struct Scsi_Host *shost,
3883 struct fc_rport *rport, struct fc_bsg_job *job)
3884{
3885 struct fc_internal *i = to_fc_internal(shost->transportt);
3886 int cmdlen = sizeof(uint32_t); /* start with length of msgcode */
3887 int ret;
3888
3889 /* Validate the rport command */
3890 switch (job->request->msgcode) {
3891 case FC_BSG_RPT_ELS:
3892 cmdlen += sizeof(struct fc_bsg_rport_els);
3893 goto check_bidi;
3894
3895 case FC_BSG_RPT_CT:
3896 cmdlen += sizeof(struct fc_bsg_rport_ct);
3897check_bidi:
3898 /* there better be xmt and rcv payloads */
3899 if ((!job->request_payload.payload_len) ||
3900 (!job->reply_payload.payload_len)) {
3901 ret = -EINVAL;
3902 goto fail_rport_msg;
3903 }
3904 break;
3905 default:
3906 ret = -EBADR;
3907 goto fail_rport_msg;
3908 }
3909
3910 /* check if we really have all the request data needed */
3911 if (job->request_len < cmdlen) {
3912 ret = -ENOMSG;
3913 goto fail_rport_msg;
3914 }
3915
3916 ret = i->f->bsg_request(job);
3917 if (!ret)
3918 return FC_DISPATCH_UNLOCKED;
3919
3920fail_rport_msg:
3921 /* return the errno failure code as the only status */
3922 BUG_ON(job->reply_len < sizeof(uint32_t));
3923 job->reply->reply_payload_rcv_len = 0;
3924 job->reply->result = ret;
3925 job->reply_len = sizeof(uint32_t);
3926 fc_bsg_jobdone(job);
3927 return FC_DISPATCH_UNLOCKED;
3928}
3929
3930
3931/**
3932 * fc_bsg_request_handler - generic handler for bsg requests
3933 * @q: request queue to manage
3934 * @shost: Scsi_Host related to the bsg object
3935 * @rport: FC remote port related to the bsg object (optional)
3936 * @dev: device structure for bsg object
3937 */
3938static void
3939fc_bsg_request_handler(struct request_queue *q, struct Scsi_Host *shost,
3940 struct fc_rport *rport, struct device *dev)
3941{
3942 struct request *req;
3943 struct fc_bsg_job *job;
3944 enum fc_dispatch_result ret;
3945
3946 if (!get_device(dev))
3947 return;
3948
3949 while (1) {
3950 if (rport && (rport->port_state == FC_PORTSTATE_BLOCKED) &&
3951 !(rport->flags & FC_RPORT_FAST_FAIL_TIMEDOUT))
3952 break;
3953
3954 req = blk_fetch_request(q);
3955 if (!req)
3956 break;
3957
3958 if (rport && (rport->port_state != FC_PORTSTATE_ONLINE)) {
3959 req->errors = -ENXIO;
3960 spin_unlock_irq(q->queue_lock);
3961 blk_end_request_all(req, -ENXIO);
3962 spin_lock_irq(q->queue_lock);
3963 continue;
3964 }
3965
3966 spin_unlock_irq(q->queue_lock);
3967
3968 ret = fc_req_to_bsgjob(shost, rport, req);
3969 if (ret) {
3970 req->errors = ret;
3971 blk_end_request_all(req, ret);
3972 spin_lock_irq(q->queue_lock);
3973 continue;
3974 }
3975
3976 job = req->special;
3977
3978 /* check if we have the msgcode value at least */
3979 if (job->request_len < sizeof(uint32_t)) {
3980 BUG_ON(job->reply_len < sizeof(uint32_t));
3981 job->reply->reply_payload_rcv_len = 0;
3982 job->reply->result = -ENOMSG;
3983 job->reply_len = sizeof(uint32_t);
3984 fc_bsg_jobdone(job);
3985 spin_lock_irq(q->queue_lock);
3986 continue;
3987 }
3988
3989 /* the dispatch routines will unlock the queue_lock */
3990 if (rport)
3991 ret = fc_bsg_rport_dispatch(q, shost, rport, job);
3992 else
3993 ret = fc_bsg_host_dispatch(q, shost, job);
3994
3995 /* did dispatcher hit state that can't process any more */
3996 if (ret == FC_DISPATCH_BREAK)
3997 break;
3998
3999 /* did dispatcher had released the lock */
4000 if (ret == FC_DISPATCH_UNLOCKED)
4001 spin_lock_irq(q->queue_lock);
4002 }
4003
4004 spin_unlock_irq(q->queue_lock);
4005 put_device(dev);
4006 spin_lock_irq(q->queue_lock);
4007}
4008
4009
4010/**
4011 * fc_bsg_host_handler - handler for bsg requests for a fc host
4012 * @q: fc host request queue
4013 */
4014static void
4015fc_bsg_host_handler(struct request_queue *q)
4016{
4017 struct Scsi_Host *shost = q->queuedata;
4018
4019 fc_bsg_request_handler(q, shost, NULL, &shost->shost_gendev);
4020}
4021
4022
4023/**
4024 * fc_bsg_rport_handler - handler for bsg requests for a fc rport
4025 * @q: rport request queue
4026 */
4027static void
4028fc_bsg_rport_handler(struct request_queue *q)
4029{
4030 struct fc_rport *rport = q->queuedata;
4031 struct Scsi_Host *shost = rport_to_shost(rport);
4032
4033 fc_bsg_request_handler(q, shost, rport, &rport->dev);
4034}
4035
4036
4037/**
4038 * fc_bsg_hostadd - Create and add the bsg hooks so we can receive requests
4039 * @shost: shost for fc_host
4040 * @fc_host: fc_host adding the structures to
4041 */
4042static int
4043fc_bsg_hostadd(struct Scsi_Host *shost, struct fc_host_attrs *fc_host)
4044{
4045 struct device *dev = &shost->shost_gendev;
4046 struct fc_internal *i = to_fc_internal(shost->transportt);
4047 struct request_queue *q;
4048 int err;
4049 char bsg_name[20];
4050
4051 fc_host->rqst_q = NULL;
4052
4053 if (!i->f->bsg_request)
4054 return -ENOTSUPP;
4055
4056 snprintf(bsg_name, sizeof(bsg_name),
4057 "fc_host%d", shost->host_no);
4058
4059 q = __scsi_alloc_queue(shost, fc_bsg_host_handler);
4060 if (!q) {
4061 printk(KERN_ERR "fc_host%d: bsg interface failed to "
4062 "initialize - no request queue\n",
4063 shost->host_no);
4064 return -ENOMEM;
4065 }
4066
4067 q->queuedata = shost;
4068 queue_flag_set_unlocked(QUEUE_FLAG_BIDI, q);
4069 blk_queue_softirq_done(q, fc_bsg_softirq_done);
4070 blk_queue_rq_timed_out(q, fc_bsg_job_timeout);
4071 blk_queue_rq_timeout(q, FC_DEFAULT_BSG_TIMEOUT);
4072
4073 err = bsg_register_queue(q, dev, bsg_name, NULL);
4074 if (err) {
4075 printk(KERN_ERR "fc_host%d: bsg interface failed to "
4076 "initialize - register queue\n",
4077 shost->host_no);
4078 blk_cleanup_queue(q);
4079 return err;
4080 }
4081
4082 fc_host->rqst_q = q;
4083 return 0;
4084}
4085
4086
4087/**
4088 * fc_bsg_rportadd - Create and add the bsg hooks so we can receive requests
4089 * @shost: shost that rport is attached to
4090 * @rport: rport that the bsg hooks are being attached to
4091 */
4092static int
4093fc_bsg_rportadd(struct Scsi_Host *shost, struct fc_rport *rport)
4094{
4095 struct device *dev = &rport->dev;
4096 struct fc_internal *i = to_fc_internal(shost->transportt);
4097 struct request_queue *q;
4098 int err;
4099
4100 rport->rqst_q = NULL;
4101
4102 if (!i->f->bsg_request)
4103 return -ENOTSUPP;
4104
4105 q = __scsi_alloc_queue(shost, fc_bsg_rport_handler);
4106 if (!q) {
4107 printk(KERN_ERR "%s: bsg interface failed to "
4108 "initialize - no request queue\n",
4109 dev->kobj.name);
4110 return -ENOMEM;
4111 }
4112
4113 q->queuedata = rport;
4114 queue_flag_set_unlocked(QUEUE_FLAG_BIDI, q);
4115 blk_queue_softirq_done(q, fc_bsg_softirq_done);
4116 blk_queue_rq_timed_out(q, fc_bsg_job_timeout);
4117 blk_queue_rq_timeout(q, BLK_DEFAULT_SG_TIMEOUT);
4118
4119 err = bsg_register_queue(q, dev, NULL, NULL);
4120 if (err) {
4121 printk(KERN_ERR "%s: bsg interface failed to "
4122 "initialize - register queue\n",
4123 dev->kobj.name);
4124 blk_cleanup_queue(q);
4125 return err;
4126 }
4127
4128 rport->rqst_q = q;
4129 return 0;
4130}
4131
4132
4133/**
4134 * fc_bsg_remove - Deletes the bsg hooks on fchosts/rports
4135 * @q: the request_queue that is to be torn down.
4136 *
4137 * Notes:
4138 * Before unregistering the queue empty any requests that are blocked
4139 *
4140 *
4141 */
4142static void
4143fc_bsg_remove(struct request_queue *q)
4144{
4145 if (q) {
4146 bsg_unregister_queue(q);
4147 blk_cleanup_queue(q);
4148 }
4149}
4150
4151
4152/* Original Author: Martin Hicks */
4153MODULE_AUTHOR("James Smart");
4154MODULE_DESCRIPTION("FC Transport Attributes");
4155MODULE_LICENSE("GPL");
4156
4157module_init(fc_transport_init);
4158module_exit(fc_transport_exit);
1/*
2 * FiberChannel transport specific attributes exported to sysfs.
3 *
4 * Copyright (c) 2003 Silicon Graphics, Inc. All rights reserved.
5 *
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License as published by
8 * the Free Software Foundation; either version 2 of the License, or
9 * (at your option) any later version.
10 *
11 * This program is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 * GNU General Public License for more details.
15 *
16 * You should have received a copy of the GNU General Public License
17 * along with this program; if not, write to the Free Software
18 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
19 *
20 * ========
21 *
22 * Copyright (C) 2004-2007 James Smart, Emulex Corporation
23 * Rewrite for host, target, device, and remote port attributes,
24 * statistics, and service functions...
25 * Add vports, etc
26 *
27 */
28#include <linux/module.h>
29#include <linux/init.h>
30#include <linux/slab.h>
31#include <linux/delay.h>
32#include <linux/kernel.h>
33#include <scsi/scsi_device.h>
34#include <scsi/scsi_host.h>
35#include <scsi/scsi_transport.h>
36#include <scsi/scsi_transport_fc.h>
37#include <scsi/scsi_cmnd.h>
38#include <linux/netlink.h>
39#include <net/netlink.h>
40#include <scsi/scsi_netlink_fc.h>
41#include <scsi/scsi_bsg_fc.h>
42#include "scsi_priv.h"
43#include "scsi_transport_fc_internal.h"
44
45static int fc_queue_work(struct Scsi_Host *, struct work_struct *);
46static void fc_vport_sched_delete(struct work_struct *work);
47static int fc_vport_setup(struct Scsi_Host *shost, int channel,
48 struct device *pdev, struct fc_vport_identifiers *ids,
49 struct fc_vport **vport);
50static int fc_bsg_hostadd(struct Scsi_Host *, struct fc_host_attrs *);
51static int fc_bsg_rportadd(struct Scsi_Host *, struct fc_rport *);
52static void fc_bsg_remove(struct request_queue *);
53static void fc_bsg_goose_queue(struct fc_rport *);
54
55/*
56 * Module Parameters
57 */
58
59/*
60 * dev_loss_tmo: the default number of seconds that the FC transport
61 * should insulate the loss of a remote port.
62 * The maximum will be capped by the value of SCSI_DEVICE_BLOCK_MAX_TIMEOUT.
63 */
64static unsigned int fc_dev_loss_tmo = 60; /* seconds */
65
66module_param_named(dev_loss_tmo, fc_dev_loss_tmo, uint, S_IRUGO|S_IWUSR);
67MODULE_PARM_DESC(dev_loss_tmo,
68 "Maximum number of seconds that the FC transport should"
69 " insulate the loss of a remote port. Once this value is"
70 " exceeded, the scsi target is removed. Value should be"
71 " between 1 and SCSI_DEVICE_BLOCK_MAX_TIMEOUT if"
72 " fast_io_fail_tmo is not set.");
73
74/*
75 * Redefine so that we can have same named attributes in the
76 * sdev/starget/host objects.
77 */
78#define FC_DEVICE_ATTR(_prefix,_name,_mode,_show,_store) \
79struct device_attribute device_attr_##_prefix##_##_name = \
80 __ATTR(_name,_mode,_show,_store)
81
82#define fc_enum_name_search(title, table_type, table) \
83static const char *get_fc_##title##_name(enum table_type table_key) \
84{ \
85 int i; \
86 char *name = NULL; \
87 \
88 for (i = 0; i < ARRAY_SIZE(table); i++) { \
89 if (table[i].value == table_key) { \
90 name = table[i].name; \
91 break; \
92 } \
93 } \
94 return name; \
95}
96
97#define fc_enum_name_match(title, table_type, table) \
98static int get_fc_##title##_match(const char *table_key, \
99 enum table_type *value) \
100{ \
101 int i; \
102 \
103 for (i = 0; i < ARRAY_SIZE(table); i++) { \
104 if (strncmp(table_key, table[i].name, \
105 table[i].matchlen) == 0) { \
106 *value = table[i].value; \
107 return 0; /* success */ \
108 } \
109 } \
110 return 1; /* failure */ \
111}
112
113
114/* Convert fc_port_type values to ascii string name */
115static struct {
116 enum fc_port_type value;
117 char *name;
118} fc_port_type_names[] = {
119 { FC_PORTTYPE_UNKNOWN, "Unknown" },
120 { FC_PORTTYPE_OTHER, "Other" },
121 { FC_PORTTYPE_NOTPRESENT, "Not Present" },
122 { FC_PORTTYPE_NPORT, "NPort (fabric via point-to-point)" },
123 { FC_PORTTYPE_NLPORT, "NLPort (fabric via loop)" },
124 { FC_PORTTYPE_LPORT, "LPort (private loop)" },
125 { FC_PORTTYPE_PTP, "Point-To-Point (direct nport connection)" },
126 { FC_PORTTYPE_NPIV, "NPIV VPORT" },
127};
128fc_enum_name_search(port_type, fc_port_type, fc_port_type_names)
129#define FC_PORTTYPE_MAX_NAMELEN 50
130
131/* Reuse fc_port_type enum function for vport_type */
132#define get_fc_vport_type_name get_fc_port_type_name
133
134
135/* Convert fc_host_event_code values to ascii string name */
136static const struct {
137 enum fc_host_event_code value;
138 char *name;
139} fc_host_event_code_names[] = {
140 { FCH_EVT_LIP, "lip" },
141 { FCH_EVT_LINKUP, "link_up" },
142 { FCH_EVT_LINKDOWN, "link_down" },
143 { FCH_EVT_LIPRESET, "lip_reset" },
144 { FCH_EVT_RSCN, "rscn" },
145 { FCH_EVT_ADAPTER_CHANGE, "adapter_chg" },
146 { FCH_EVT_PORT_UNKNOWN, "port_unknown" },
147 { FCH_EVT_PORT_ONLINE, "port_online" },
148 { FCH_EVT_PORT_OFFLINE, "port_offline" },
149 { FCH_EVT_PORT_FABRIC, "port_fabric" },
150 { FCH_EVT_LINK_UNKNOWN, "link_unknown" },
151 { FCH_EVT_VENDOR_UNIQUE, "vendor_unique" },
152};
153fc_enum_name_search(host_event_code, fc_host_event_code,
154 fc_host_event_code_names)
155#define FC_HOST_EVENT_CODE_MAX_NAMELEN 30
156
157
158/* Convert fc_port_state values to ascii string name */
159static struct {
160 enum fc_port_state value;
161 char *name;
162} fc_port_state_names[] = {
163 { FC_PORTSTATE_UNKNOWN, "Unknown" },
164 { FC_PORTSTATE_NOTPRESENT, "Not Present" },
165 { FC_PORTSTATE_ONLINE, "Online" },
166 { FC_PORTSTATE_OFFLINE, "Offline" },
167 { FC_PORTSTATE_BLOCKED, "Blocked" },
168 { FC_PORTSTATE_BYPASSED, "Bypassed" },
169 { FC_PORTSTATE_DIAGNOSTICS, "Diagnostics" },
170 { FC_PORTSTATE_LINKDOWN, "Linkdown" },
171 { FC_PORTSTATE_ERROR, "Error" },
172 { FC_PORTSTATE_LOOPBACK, "Loopback" },
173 { FC_PORTSTATE_DELETED, "Deleted" },
174};
175fc_enum_name_search(port_state, fc_port_state, fc_port_state_names)
176#define FC_PORTSTATE_MAX_NAMELEN 20
177
178
179/* Convert fc_vport_state values to ascii string name */
180static struct {
181 enum fc_vport_state value;
182 char *name;
183} fc_vport_state_names[] = {
184 { FC_VPORT_UNKNOWN, "Unknown" },
185 { FC_VPORT_ACTIVE, "Active" },
186 { FC_VPORT_DISABLED, "Disabled" },
187 { FC_VPORT_LINKDOWN, "Linkdown" },
188 { FC_VPORT_INITIALIZING, "Initializing" },
189 { FC_VPORT_NO_FABRIC_SUPP, "No Fabric Support" },
190 { FC_VPORT_NO_FABRIC_RSCS, "No Fabric Resources" },
191 { FC_VPORT_FABRIC_LOGOUT, "Fabric Logout" },
192 { FC_VPORT_FABRIC_REJ_WWN, "Fabric Rejected WWN" },
193 { FC_VPORT_FAILED, "VPort Failed" },
194};
195fc_enum_name_search(vport_state, fc_vport_state, fc_vport_state_names)
196#define FC_VPORTSTATE_MAX_NAMELEN 24
197
198/* Reuse fc_vport_state enum function for vport_last_state */
199#define get_fc_vport_last_state_name get_fc_vport_state_name
200
201
202/* Convert fc_tgtid_binding_type values to ascii string name */
203static const struct {
204 enum fc_tgtid_binding_type value;
205 char *name;
206 int matchlen;
207} fc_tgtid_binding_type_names[] = {
208 { FC_TGTID_BIND_NONE, "none", 4 },
209 { FC_TGTID_BIND_BY_WWPN, "wwpn (World Wide Port Name)", 4 },
210 { FC_TGTID_BIND_BY_WWNN, "wwnn (World Wide Node Name)", 4 },
211 { FC_TGTID_BIND_BY_ID, "port_id (FC Address)", 7 },
212};
213fc_enum_name_search(tgtid_bind_type, fc_tgtid_binding_type,
214 fc_tgtid_binding_type_names)
215fc_enum_name_match(tgtid_bind_type, fc_tgtid_binding_type,
216 fc_tgtid_binding_type_names)
217#define FC_BINDTYPE_MAX_NAMELEN 30
218
219
220#define fc_bitfield_name_search(title, table) \
221static ssize_t \
222get_fc_##title##_names(u32 table_key, char *buf) \
223{ \
224 char *prefix = ""; \
225 ssize_t len = 0; \
226 int i; \
227 \
228 for (i = 0; i < ARRAY_SIZE(table); i++) { \
229 if (table[i].value & table_key) { \
230 len += sprintf(buf + len, "%s%s", \
231 prefix, table[i].name); \
232 prefix = ", "; \
233 } \
234 } \
235 len += sprintf(buf + len, "\n"); \
236 return len; \
237}
238
239
240/* Convert FC_COS bit values to ascii string name */
241static const struct {
242 u32 value;
243 char *name;
244} fc_cos_names[] = {
245 { FC_COS_CLASS1, "Class 1" },
246 { FC_COS_CLASS2, "Class 2" },
247 { FC_COS_CLASS3, "Class 3" },
248 { FC_COS_CLASS4, "Class 4" },
249 { FC_COS_CLASS6, "Class 6" },
250};
251fc_bitfield_name_search(cos, fc_cos_names)
252
253
254/* Convert FC_PORTSPEED bit values to ascii string name */
255static const struct {
256 u32 value;
257 char *name;
258} fc_port_speed_names[] = {
259 { FC_PORTSPEED_1GBIT, "1 Gbit" },
260 { FC_PORTSPEED_2GBIT, "2 Gbit" },
261 { FC_PORTSPEED_4GBIT, "4 Gbit" },
262 { FC_PORTSPEED_10GBIT, "10 Gbit" },
263 { FC_PORTSPEED_8GBIT, "8 Gbit" },
264 { FC_PORTSPEED_16GBIT, "16 Gbit" },
265 { FC_PORTSPEED_NOT_NEGOTIATED, "Not Negotiated" },
266};
267fc_bitfield_name_search(port_speed, fc_port_speed_names)
268
269
270static int
271show_fc_fc4s (char *buf, u8 *fc4_list)
272{
273 int i, len=0;
274
275 for (i = 0; i < FC_FC4_LIST_SIZE; i++, fc4_list++)
276 len += sprintf(buf + len , "0x%02x ", *fc4_list);
277 len += sprintf(buf + len, "\n");
278 return len;
279}
280
281
282/* Convert FC_PORT_ROLE bit values to ascii string name */
283static const struct {
284 u32 value;
285 char *name;
286} fc_port_role_names[] = {
287 { FC_PORT_ROLE_FCP_TARGET, "FCP Target" },
288 { FC_PORT_ROLE_FCP_INITIATOR, "FCP Initiator" },
289 { FC_PORT_ROLE_IP_PORT, "IP Port" },
290};
291fc_bitfield_name_search(port_roles, fc_port_role_names)
292
293/*
294 * Define roles that are specific to port_id. Values are relative to ROLE_MASK.
295 */
296#define FC_WELLKNOWN_PORTID_MASK 0xfffff0
297#define FC_WELLKNOWN_ROLE_MASK 0x00000f
298#define FC_FPORT_PORTID 0x00000e
299#define FC_FABCTLR_PORTID 0x00000d
300#define FC_DIRSRVR_PORTID 0x00000c
301#define FC_TIMESRVR_PORTID 0x00000b
302#define FC_MGMTSRVR_PORTID 0x00000a
303
304
305static void fc_timeout_deleted_rport(struct work_struct *work);
306static void fc_timeout_fail_rport_io(struct work_struct *work);
307static void fc_scsi_scan_rport(struct work_struct *work);
308
309/*
310 * Attribute counts pre object type...
311 * Increase these values if you add attributes
312 */
313#define FC_STARGET_NUM_ATTRS 3
314#define FC_RPORT_NUM_ATTRS 10
315#define FC_VPORT_NUM_ATTRS 9
316#define FC_HOST_NUM_ATTRS 29
317
318struct fc_internal {
319 struct scsi_transport_template t;
320 struct fc_function_template *f;
321
322 /*
323 * For attributes : each object has :
324 * An array of the actual attributes structures
325 * An array of null-terminated pointers to the attribute
326 * structures - used for mid-layer interaction.
327 *
328 * The attribute containers for the starget and host are are
329 * part of the midlayer. As the remote port is specific to the
330 * fc transport, we must provide the attribute container.
331 */
332 struct device_attribute private_starget_attrs[
333 FC_STARGET_NUM_ATTRS];
334 struct device_attribute *starget_attrs[FC_STARGET_NUM_ATTRS + 1];
335
336 struct device_attribute private_host_attrs[FC_HOST_NUM_ATTRS];
337 struct device_attribute *host_attrs[FC_HOST_NUM_ATTRS + 1];
338
339 struct transport_container rport_attr_cont;
340 struct device_attribute private_rport_attrs[FC_RPORT_NUM_ATTRS];
341 struct device_attribute *rport_attrs[FC_RPORT_NUM_ATTRS + 1];
342
343 struct transport_container vport_attr_cont;
344 struct device_attribute private_vport_attrs[FC_VPORT_NUM_ATTRS];
345 struct device_attribute *vport_attrs[FC_VPORT_NUM_ATTRS + 1];
346};
347
348#define to_fc_internal(tmpl) container_of(tmpl, struct fc_internal, t)
349
350static int fc_target_setup(struct transport_container *tc, struct device *dev,
351 struct device *cdev)
352{
353 struct scsi_target *starget = to_scsi_target(dev);
354 struct fc_rport *rport = starget_to_rport(starget);
355
356 /*
357 * if parent is remote port, use values from remote port.
358 * Otherwise, this host uses the fc_transport, but not the
359 * remote port interface. As such, initialize to known non-values.
360 */
361 if (rport) {
362 fc_starget_node_name(starget) = rport->node_name;
363 fc_starget_port_name(starget) = rport->port_name;
364 fc_starget_port_id(starget) = rport->port_id;
365 } else {
366 fc_starget_node_name(starget) = -1;
367 fc_starget_port_name(starget) = -1;
368 fc_starget_port_id(starget) = -1;
369 }
370
371 return 0;
372}
373
374static DECLARE_TRANSPORT_CLASS(fc_transport_class,
375 "fc_transport",
376 fc_target_setup,
377 NULL,
378 NULL);
379
380static int fc_host_setup(struct transport_container *tc, struct device *dev,
381 struct device *cdev)
382{
383 struct Scsi_Host *shost = dev_to_shost(dev);
384 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
385
386 /*
387 * Set default values easily detected by the midlayer as
388 * failure cases. The scsi lldd is responsible for initializing
389 * all transport attributes to valid values per host.
390 */
391 fc_host->node_name = -1;
392 fc_host->port_name = -1;
393 fc_host->permanent_port_name = -1;
394 fc_host->supported_classes = FC_COS_UNSPECIFIED;
395 memset(fc_host->supported_fc4s, 0,
396 sizeof(fc_host->supported_fc4s));
397 fc_host->supported_speeds = FC_PORTSPEED_UNKNOWN;
398 fc_host->maxframe_size = -1;
399 fc_host->max_npiv_vports = 0;
400 memset(fc_host->serial_number, 0,
401 sizeof(fc_host->serial_number));
402 memset(fc_host->manufacturer, 0,
403 sizeof(fc_host->manufacturer));
404 memset(fc_host->model, 0,
405 sizeof(fc_host->model));
406 memset(fc_host->model_description, 0,
407 sizeof(fc_host->model_description));
408 memset(fc_host->hardware_version, 0,
409 sizeof(fc_host->hardware_version));
410 memset(fc_host->driver_version, 0,
411 sizeof(fc_host->driver_version));
412 memset(fc_host->firmware_version, 0,
413 sizeof(fc_host->firmware_version));
414 memset(fc_host->optionrom_version, 0,
415 sizeof(fc_host->optionrom_version));
416
417 fc_host->port_id = -1;
418 fc_host->port_type = FC_PORTTYPE_UNKNOWN;
419 fc_host->port_state = FC_PORTSTATE_UNKNOWN;
420 memset(fc_host->active_fc4s, 0,
421 sizeof(fc_host->active_fc4s));
422 fc_host->speed = FC_PORTSPEED_UNKNOWN;
423 fc_host->fabric_name = -1;
424 memset(fc_host->symbolic_name, 0, sizeof(fc_host->symbolic_name));
425 memset(fc_host->system_hostname, 0, sizeof(fc_host->system_hostname));
426
427 fc_host->tgtid_bind_type = FC_TGTID_BIND_BY_WWPN;
428
429 INIT_LIST_HEAD(&fc_host->rports);
430 INIT_LIST_HEAD(&fc_host->rport_bindings);
431 INIT_LIST_HEAD(&fc_host->vports);
432 fc_host->next_rport_number = 0;
433 fc_host->next_target_id = 0;
434 fc_host->next_vport_number = 0;
435 fc_host->npiv_vports_inuse = 0;
436
437 snprintf(fc_host->work_q_name, sizeof(fc_host->work_q_name),
438 "fc_wq_%d", shost->host_no);
439 fc_host->work_q = alloc_workqueue(fc_host->work_q_name, 0, 0);
440 if (!fc_host->work_q)
441 return -ENOMEM;
442
443 fc_host->dev_loss_tmo = fc_dev_loss_tmo;
444 snprintf(fc_host->devloss_work_q_name,
445 sizeof(fc_host->devloss_work_q_name),
446 "fc_dl_%d", shost->host_no);
447 fc_host->devloss_work_q =
448 alloc_workqueue(fc_host->devloss_work_q_name, 0, 0);
449 if (!fc_host->devloss_work_q) {
450 destroy_workqueue(fc_host->work_q);
451 fc_host->work_q = NULL;
452 return -ENOMEM;
453 }
454
455 fc_bsg_hostadd(shost, fc_host);
456 /* ignore any bsg add error - we just can't do sgio */
457
458 return 0;
459}
460
461static int fc_host_remove(struct transport_container *tc, struct device *dev,
462 struct device *cdev)
463{
464 struct Scsi_Host *shost = dev_to_shost(dev);
465 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
466
467 fc_bsg_remove(fc_host->rqst_q);
468 return 0;
469}
470
471static DECLARE_TRANSPORT_CLASS(fc_host_class,
472 "fc_host",
473 fc_host_setup,
474 fc_host_remove,
475 NULL);
476
477/*
478 * Setup and Remove actions for remote ports are handled
479 * in the service functions below.
480 */
481static DECLARE_TRANSPORT_CLASS(fc_rport_class,
482 "fc_remote_ports",
483 NULL,
484 NULL,
485 NULL);
486
487/*
488 * Setup and Remove actions for virtual ports are handled
489 * in the service functions below.
490 */
491static DECLARE_TRANSPORT_CLASS(fc_vport_class,
492 "fc_vports",
493 NULL,
494 NULL,
495 NULL);
496
497/*
498 * Netlink Infrastructure
499 */
500
501static atomic_t fc_event_seq;
502
503/**
504 * fc_get_event_number - Obtain the next sequential FC event number
505 *
506 * Notes:
507 * We could have inlined this, but it would have required fc_event_seq to
508 * be exposed. For now, live with the subroutine call.
509 * Atomic used to avoid lock/unlock...
510 */
511u32
512fc_get_event_number(void)
513{
514 return atomic_add_return(1, &fc_event_seq);
515}
516EXPORT_SYMBOL(fc_get_event_number);
517
518
519/**
520 * fc_host_post_event - called to post an even on an fc_host.
521 * @shost: host the event occurred on
522 * @event_number: fc event number obtained from get_fc_event_number()
523 * @event_code: fc_host event being posted
524 * @event_data: 32bits of data for the event being posted
525 *
526 * Notes:
527 * This routine assumes no locks are held on entry.
528 */
529void
530fc_host_post_event(struct Scsi_Host *shost, u32 event_number,
531 enum fc_host_event_code event_code, u32 event_data)
532{
533 struct sk_buff *skb;
534 struct nlmsghdr *nlh;
535 struct fc_nl_event *event;
536 const char *name;
537 u32 len, skblen;
538 int err;
539
540 if (!scsi_nl_sock) {
541 err = -ENOENT;
542 goto send_fail;
543 }
544
545 len = FC_NL_MSGALIGN(sizeof(*event));
546 skblen = NLMSG_SPACE(len);
547
548 skb = alloc_skb(skblen, GFP_KERNEL);
549 if (!skb) {
550 err = -ENOBUFS;
551 goto send_fail;
552 }
553
554 nlh = nlmsg_put(skb, 0, 0, SCSI_TRANSPORT_MSG,
555 skblen - sizeof(*nlh), 0);
556 if (!nlh) {
557 err = -ENOBUFS;
558 goto send_fail_skb;
559 }
560 event = NLMSG_DATA(nlh);
561
562 INIT_SCSI_NL_HDR(&event->snlh, SCSI_NL_TRANSPORT_FC,
563 FC_NL_ASYNC_EVENT, len);
564 event->seconds = get_seconds();
565 event->vendor_id = 0;
566 event->host_no = shost->host_no;
567 event->event_datalen = sizeof(u32); /* bytes */
568 event->event_num = event_number;
569 event->event_code = event_code;
570 event->event_data = event_data;
571
572 nlmsg_multicast(scsi_nl_sock, skb, 0, SCSI_NL_GRP_FC_EVENTS,
573 GFP_KERNEL);
574 return;
575
576send_fail_skb:
577 kfree_skb(skb);
578send_fail:
579 name = get_fc_host_event_code_name(event_code);
580 printk(KERN_WARNING
581 "%s: Dropped Event : host %d %s data 0x%08x - err %d\n",
582 __func__, shost->host_no,
583 (name) ? name : "<unknown>", event_data, err);
584 return;
585}
586EXPORT_SYMBOL(fc_host_post_event);
587
588
589/**
590 * fc_host_post_vendor_event - called to post a vendor unique event on an fc_host
591 * @shost: host the event occurred on
592 * @event_number: fc event number obtained from get_fc_event_number()
593 * @data_len: amount, in bytes, of vendor unique data
594 * @data_buf: pointer to vendor unique data
595 * @vendor_id: Vendor id
596 *
597 * Notes:
598 * This routine assumes no locks are held on entry.
599 */
600void
601fc_host_post_vendor_event(struct Scsi_Host *shost, u32 event_number,
602 u32 data_len, char * data_buf, u64 vendor_id)
603{
604 struct sk_buff *skb;
605 struct nlmsghdr *nlh;
606 struct fc_nl_event *event;
607 u32 len, skblen;
608 int err;
609
610 if (!scsi_nl_sock) {
611 err = -ENOENT;
612 goto send_vendor_fail;
613 }
614
615 len = FC_NL_MSGALIGN(sizeof(*event) + data_len);
616 skblen = NLMSG_SPACE(len);
617
618 skb = alloc_skb(skblen, GFP_KERNEL);
619 if (!skb) {
620 err = -ENOBUFS;
621 goto send_vendor_fail;
622 }
623
624 nlh = nlmsg_put(skb, 0, 0, SCSI_TRANSPORT_MSG,
625 skblen - sizeof(*nlh), 0);
626 if (!nlh) {
627 err = -ENOBUFS;
628 goto send_vendor_fail_skb;
629 }
630 event = NLMSG_DATA(nlh);
631
632 INIT_SCSI_NL_HDR(&event->snlh, SCSI_NL_TRANSPORT_FC,
633 FC_NL_ASYNC_EVENT, len);
634 event->seconds = get_seconds();
635 event->vendor_id = vendor_id;
636 event->host_no = shost->host_no;
637 event->event_datalen = data_len; /* bytes */
638 event->event_num = event_number;
639 event->event_code = FCH_EVT_VENDOR_UNIQUE;
640 memcpy(&event->event_data, data_buf, data_len);
641
642 nlmsg_multicast(scsi_nl_sock, skb, 0, SCSI_NL_GRP_FC_EVENTS,
643 GFP_KERNEL);
644 return;
645
646send_vendor_fail_skb:
647 kfree_skb(skb);
648send_vendor_fail:
649 printk(KERN_WARNING
650 "%s: Dropped Event : host %d vendor_unique - err %d\n",
651 __func__, shost->host_no, err);
652 return;
653}
654EXPORT_SYMBOL(fc_host_post_vendor_event);
655
656
657
658static __init int fc_transport_init(void)
659{
660 int error;
661
662 atomic_set(&fc_event_seq, 0);
663
664 error = transport_class_register(&fc_host_class);
665 if (error)
666 return error;
667 error = transport_class_register(&fc_vport_class);
668 if (error)
669 goto unreg_host_class;
670 error = transport_class_register(&fc_rport_class);
671 if (error)
672 goto unreg_vport_class;
673 error = transport_class_register(&fc_transport_class);
674 if (error)
675 goto unreg_rport_class;
676 return 0;
677
678unreg_rport_class:
679 transport_class_unregister(&fc_rport_class);
680unreg_vport_class:
681 transport_class_unregister(&fc_vport_class);
682unreg_host_class:
683 transport_class_unregister(&fc_host_class);
684 return error;
685}
686
687static void __exit fc_transport_exit(void)
688{
689 transport_class_unregister(&fc_transport_class);
690 transport_class_unregister(&fc_rport_class);
691 transport_class_unregister(&fc_host_class);
692 transport_class_unregister(&fc_vport_class);
693}
694
695/*
696 * FC Remote Port Attribute Management
697 */
698
699#define fc_rport_show_function(field, format_string, sz, cast) \
700static ssize_t \
701show_fc_rport_##field (struct device *dev, \
702 struct device_attribute *attr, char *buf) \
703{ \
704 struct fc_rport *rport = transport_class_to_rport(dev); \
705 struct Scsi_Host *shost = rport_to_shost(rport); \
706 struct fc_internal *i = to_fc_internal(shost->transportt); \
707 if ((i->f->get_rport_##field) && \
708 !((rport->port_state == FC_PORTSTATE_BLOCKED) || \
709 (rport->port_state == FC_PORTSTATE_DELETED) || \
710 (rport->port_state == FC_PORTSTATE_NOTPRESENT))) \
711 i->f->get_rport_##field(rport); \
712 return snprintf(buf, sz, format_string, cast rport->field); \
713}
714
715#define fc_rport_store_function(field) \
716static ssize_t \
717store_fc_rport_##field(struct device *dev, \
718 struct device_attribute *attr, \
719 const char *buf, size_t count) \
720{ \
721 int val; \
722 struct fc_rport *rport = transport_class_to_rport(dev); \
723 struct Scsi_Host *shost = rport_to_shost(rport); \
724 struct fc_internal *i = to_fc_internal(shost->transportt); \
725 char *cp; \
726 if ((rport->port_state == FC_PORTSTATE_BLOCKED) || \
727 (rport->port_state == FC_PORTSTATE_DELETED) || \
728 (rport->port_state == FC_PORTSTATE_NOTPRESENT)) \
729 return -EBUSY; \
730 val = simple_strtoul(buf, &cp, 0); \
731 if (*cp && (*cp != '\n')) \
732 return -EINVAL; \
733 i->f->set_rport_##field(rport, val); \
734 return count; \
735}
736
737#define fc_rport_rd_attr(field, format_string, sz) \
738 fc_rport_show_function(field, format_string, sz, ) \
739static FC_DEVICE_ATTR(rport, field, S_IRUGO, \
740 show_fc_rport_##field, NULL)
741
742#define fc_rport_rd_attr_cast(field, format_string, sz, cast) \
743 fc_rport_show_function(field, format_string, sz, (cast)) \
744static FC_DEVICE_ATTR(rport, field, S_IRUGO, \
745 show_fc_rport_##field, NULL)
746
747#define fc_rport_rw_attr(field, format_string, sz) \
748 fc_rport_show_function(field, format_string, sz, ) \
749 fc_rport_store_function(field) \
750static FC_DEVICE_ATTR(rport, field, S_IRUGO | S_IWUSR, \
751 show_fc_rport_##field, \
752 store_fc_rport_##field)
753
754
755#define fc_private_rport_show_function(field, format_string, sz, cast) \
756static ssize_t \
757show_fc_rport_##field (struct device *dev, \
758 struct device_attribute *attr, char *buf) \
759{ \
760 struct fc_rport *rport = transport_class_to_rport(dev); \
761 return snprintf(buf, sz, format_string, cast rport->field); \
762}
763
764#define fc_private_rport_rd_attr(field, format_string, sz) \
765 fc_private_rport_show_function(field, format_string, sz, ) \
766static FC_DEVICE_ATTR(rport, field, S_IRUGO, \
767 show_fc_rport_##field, NULL)
768
769#define fc_private_rport_rd_attr_cast(field, format_string, sz, cast) \
770 fc_private_rport_show_function(field, format_string, sz, (cast)) \
771static FC_DEVICE_ATTR(rport, field, S_IRUGO, \
772 show_fc_rport_##field, NULL)
773
774
775#define fc_private_rport_rd_enum_attr(title, maxlen) \
776static ssize_t \
777show_fc_rport_##title (struct device *dev, \
778 struct device_attribute *attr, char *buf) \
779{ \
780 struct fc_rport *rport = transport_class_to_rport(dev); \
781 const char *name; \
782 name = get_fc_##title##_name(rport->title); \
783 if (!name) \
784 return -EINVAL; \
785 return snprintf(buf, maxlen, "%s\n", name); \
786} \
787static FC_DEVICE_ATTR(rport, title, S_IRUGO, \
788 show_fc_rport_##title, NULL)
789
790
791#define SETUP_RPORT_ATTRIBUTE_RD(field) \
792 i->private_rport_attrs[count] = device_attr_rport_##field; \
793 i->private_rport_attrs[count].attr.mode = S_IRUGO; \
794 i->private_rport_attrs[count].store = NULL; \
795 i->rport_attrs[count] = &i->private_rport_attrs[count]; \
796 if (i->f->show_rport_##field) \
797 count++
798
799#define SETUP_PRIVATE_RPORT_ATTRIBUTE_RD(field) \
800 i->private_rport_attrs[count] = device_attr_rport_##field; \
801 i->private_rport_attrs[count].attr.mode = S_IRUGO; \
802 i->private_rport_attrs[count].store = NULL; \
803 i->rport_attrs[count] = &i->private_rport_attrs[count]; \
804 count++
805
806#define SETUP_RPORT_ATTRIBUTE_RW(field) \
807 i->private_rport_attrs[count] = device_attr_rport_##field; \
808 if (!i->f->set_rport_##field) { \
809 i->private_rport_attrs[count].attr.mode = S_IRUGO; \
810 i->private_rport_attrs[count].store = NULL; \
811 } \
812 i->rport_attrs[count] = &i->private_rport_attrs[count]; \
813 if (i->f->show_rport_##field) \
814 count++
815
816#define SETUP_PRIVATE_RPORT_ATTRIBUTE_RW(field) \
817{ \
818 i->private_rport_attrs[count] = device_attr_rport_##field; \
819 i->rport_attrs[count] = &i->private_rport_attrs[count]; \
820 count++; \
821}
822
823
824/* The FC Transport Remote Port Attributes: */
825
826/* Fixed Remote Port Attributes */
827
828fc_private_rport_rd_attr(maxframe_size, "%u bytes\n", 20);
829
830static ssize_t
831show_fc_rport_supported_classes (struct device *dev,
832 struct device_attribute *attr, char *buf)
833{
834 struct fc_rport *rport = transport_class_to_rport(dev);
835 if (rport->supported_classes == FC_COS_UNSPECIFIED)
836 return snprintf(buf, 20, "unspecified\n");
837 return get_fc_cos_names(rport->supported_classes, buf);
838}
839static FC_DEVICE_ATTR(rport, supported_classes, S_IRUGO,
840 show_fc_rport_supported_classes, NULL);
841
842/* Dynamic Remote Port Attributes */
843
844/*
845 * dev_loss_tmo attribute
846 */
847static int fc_str_to_dev_loss(const char *buf, unsigned long *val)
848{
849 char *cp;
850
851 *val = simple_strtoul(buf, &cp, 0);
852 if ((*cp && (*cp != '\n')) || (*val < 0))
853 return -EINVAL;
854 /*
855 * Check for overflow; dev_loss_tmo is u32
856 */
857 if (*val > UINT_MAX)
858 return -EINVAL;
859
860 return 0;
861}
862
863static int fc_rport_set_dev_loss_tmo(struct fc_rport *rport,
864 unsigned long val)
865{
866 struct Scsi_Host *shost = rport_to_shost(rport);
867 struct fc_internal *i = to_fc_internal(shost->transportt);
868
869 if ((rport->port_state == FC_PORTSTATE_BLOCKED) ||
870 (rport->port_state == FC_PORTSTATE_DELETED) ||
871 (rport->port_state == FC_PORTSTATE_NOTPRESENT))
872 return -EBUSY;
873 /*
874 * Check for overflow; dev_loss_tmo is u32
875 */
876 if (val > UINT_MAX)
877 return -EINVAL;
878
879 /*
880 * If fast_io_fail is off we have to cap
881 * dev_loss_tmo at SCSI_DEVICE_BLOCK_MAX_TIMEOUT
882 */
883 if (rport->fast_io_fail_tmo == -1 &&
884 val > SCSI_DEVICE_BLOCK_MAX_TIMEOUT)
885 return -EINVAL;
886
887 i->f->set_rport_dev_loss_tmo(rport, val);
888 return 0;
889}
890
891fc_rport_show_function(dev_loss_tmo, "%d\n", 20, )
892static ssize_t
893store_fc_rport_dev_loss_tmo(struct device *dev, struct device_attribute *attr,
894 const char *buf, size_t count)
895{
896 struct fc_rport *rport = transport_class_to_rport(dev);
897 unsigned long val;
898 int rc;
899
900 rc = fc_str_to_dev_loss(buf, &val);
901 if (rc)
902 return rc;
903
904 rc = fc_rport_set_dev_loss_tmo(rport, val);
905 if (rc)
906 return rc;
907 return count;
908}
909static FC_DEVICE_ATTR(rport, dev_loss_tmo, S_IRUGO | S_IWUSR,
910 show_fc_rport_dev_loss_tmo, store_fc_rport_dev_loss_tmo);
911
912
913/* Private Remote Port Attributes */
914
915fc_private_rport_rd_attr_cast(node_name, "0x%llx\n", 20, unsigned long long);
916fc_private_rport_rd_attr_cast(port_name, "0x%llx\n", 20, unsigned long long);
917fc_private_rport_rd_attr(port_id, "0x%06x\n", 20);
918
919static ssize_t
920show_fc_rport_roles (struct device *dev, struct device_attribute *attr,
921 char *buf)
922{
923 struct fc_rport *rport = transport_class_to_rport(dev);
924
925 /* identify any roles that are port_id specific */
926 if ((rport->port_id != -1) &&
927 (rport->port_id & FC_WELLKNOWN_PORTID_MASK) ==
928 FC_WELLKNOWN_PORTID_MASK) {
929 switch (rport->port_id & FC_WELLKNOWN_ROLE_MASK) {
930 case FC_FPORT_PORTID:
931 return snprintf(buf, 30, "Fabric Port\n");
932 case FC_FABCTLR_PORTID:
933 return snprintf(buf, 30, "Fabric Controller\n");
934 case FC_DIRSRVR_PORTID:
935 return snprintf(buf, 30, "Directory Server\n");
936 case FC_TIMESRVR_PORTID:
937 return snprintf(buf, 30, "Time Server\n");
938 case FC_MGMTSRVR_PORTID:
939 return snprintf(buf, 30, "Management Server\n");
940 default:
941 return snprintf(buf, 30, "Unknown Fabric Entity\n");
942 }
943 } else {
944 if (rport->roles == FC_PORT_ROLE_UNKNOWN)
945 return snprintf(buf, 20, "unknown\n");
946 return get_fc_port_roles_names(rport->roles, buf);
947 }
948}
949static FC_DEVICE_ATTR(rport, roles, S_IRUGO,
950 show_fc_rport_roles, NULL);
951
952fc_private_rport_rd_enum_attr(port_state, FC_PORTSTATE_MAX_NAMELEN);
953fc_private_rport_rd_attr(scsi_target_id, "%d\n", 20);
954
955/*
956 * fast_io_fail_tmo attribute
957 */
958static ssize_t
959show_fc_rport_fast_io_fail_tmo (struct device *dev,
960 struct device_attribute *attr, char *buf)
961{
962 struct fc_rport *rport = transport_class_to_rport(dev);
963
964 if (rport->fast_io_fail_tmo == -1)
965 return snprintf(buf, 5, "off\n");
966 return snprintf(buf, 20, "%d\n", rport->fast_io_fail_tmo);
967}
968
969static ssize_t
970store_fc_rport_fast_io_fail_tmo(struct device *dev,
971 struct device_attribute *attr, const char *buf,
972 size_t count)
973{
974 int val;
975 char *cp;
976 struct fc_rport *rport = transport_class_to_rport(dev);
977
978 if ((rport->port_state == FC_PORTSTATE_BLOCKED) ||
979 (rport->port_state == FC_PORTSTATE_DELETED) ||
980 (rport->port_state == FC_PORTSTATE_NOTPRESENT))
981 return -EBUSY;
982 if (strncmp(buf, "off", 3) == 0)
983 rport->fast_io_fail_tmo = -1;
984 else {
985 val = simple_strtoul(buf, &cp, 0);
986 if ((*cp && (*cp != '\n')) || (val < 0))
987 return -EINVAL;
988 /*
989 * Cap fast_io_fail by dev_loss_tmo or
990 * SCSI_DEVICE_BLOCK_MAX_TIMEOUT.
991 */
992 if ((val >= rport->dev_loss_tmo) ||
993 (val > SCSI_DEVICE_BLOCK_MAX_TIMEOUT))
994 return -EINVAL;
995
996 rport->fast_io_fail_tmo = val;
997 }
998 return count;
999}
1000static FC_DEVICE_ATTR(rport, fast_io_fail_tmo, S_IRUGO | S_IWUSR,
1001 show_fc_rport_fast_io_fail_tmo, store_fc_rport_fast_io_fail_tmo);
1002
1003
1004/*
1005 * FC SCSI Target Attribute Management
1006 */
1007
1008/*
1009 * Note: in the target show function we recognize when the remote
1010 * port is in the hierarchy and do not allow the driver to get
1011 * involved in sysfs functions. The driver only gets involved if
1012 * it's the "old" style that doesn't use rports.
1013 */
1014#define fc_starget_show_function(field, format_string, sz, cast) \
1015static ssize_t \
1016show_fc_starget_##field (struct device *dev, \
1017 struct device_attribute *attr, char *buf) \
1018{ \
1019 struct scsi_target *starget = transport_class_to_starget(dev); \
1020 struct Scsi_Host *shost = dev_to_shost(starget->dev.parent); \
1021 struct fc_internal *i = to_fc_internal(shost->transportt); \
1022 struct fc_rport *rport = starget_to_rport(starget); \
1023 if (rport) \
1024 fc_starget_##field(starget) = rport->field; \
1025 else if (i->f->get_starget_##field) \
1026 i->f->get_starget_##field(starget); \
1027 return snprintf(buf, sz, format_string, \
1028 cast fc_starget_##field(starget)); \
1029}
1030
1031#define fc_starget_rd_attr(field, format_string, sz) \
1032 fc_starget_show_function(field, format_string, sz, ) \
1033static FC_DEVICE_ATTR(starget, field, S_IRUGO, \
1034 show_fc_starget_##field, NULL)
1035
1036#define fc_starget_rd_attr_cast(field, format_string, sz, cast) \
1037 fc_starget_show_function(field, format_string, sz, (cast)) \
1038static FC_DEVICE_ATTR(starget, field, S_IRUGO, \
1039 show_fc_starget_##field, NULL)
1040
1041#define SETUP_STARGET_ATTRIBUTE_RD(field) \
1042 i->private_starget_attrs[count] = device_attr_starget_##field; \
1043 i->private_starget_attrs[count].attr.mode = S_IRUGO; \
1044 i->private_starget_attrs[count].store = NULL; \
1045 i->starget_attrs[count] = &i->private_starget_attrs[count]; \
1046 if (i->f->show_starget_##field) \
1047 count++
1048
1049#define SETUP_STARGET_ATTRIBUTE_RW(field) \
1050 i->private_starget_attrs[count] = device_attr_starget_##field; \
1051 if (!i->f->set_starget_##field) { \
1052 i->private_starget_attrs[count].attr.mode = S_IRUGO; \
1053 i->private_starget_attrs[count].store = NULL; \
1054 } \
1055 i->starget_attrs[count] = &i->private_starget_attrs[count]; \
1056 if (i->f->show_starget_##field) \
1057 count++
1058
1059/* The FC Transport SCSI Target Attributes: */
1060fc_starget_rd_attr_cast(node_name, "0x%llx\n", 20, unsigned long long);
1061fc_starget_rd_attr_cast(port_name, "0x%llx\n", 20, unsigned long long);
1062fc_starget_rd_attr(port_id, "0x%06x\n", 20);
1063
1064
1065/*
1066 * FC Virtual Port Attribute Management
1067 */
1068
1069#define fc_vport_show_function(field, format_string, sz, cast) \
1070static ssize_t \
1071show_fc_vport_##field (struct device *dev, \
1072 struct device_attribute *attr, char *buf) \
1073{ \
1074 struct fc_vport *vport = transport_class_to_vport(dev); \
1075 struct Scsi_Host *shost = vport_to_shost(vport); \
1076 struct fc_internal *i = to_fc_internal(shost->transportt); \
1077 if ((i->f->get_vport_##field) && \
1078 !(vport->flags & (FC_VPORT_DEL | FC_VPORT_CREATING))) \
1079 i->f->get_vport_##field(vport); \
1080 return snprintf(buf, sz, format_string, cast vport->field); \
1081}
1082
1083#define fc_vport_store_function(field) \
1084static ssize_t \
1085store_fc_vport_##field(struct device *dev, \
1086 struct device_attribute *attr, \
1087 const char *buf, size_t count) \
1088{ \
1089 int val; \
1090 struct fc_vport *vport = transport_class_to_vport(dev); \
1091 struct Scsi_Host *shost = vport_to_shost(vport); \
1092 struct fc_internal *i = to_fc_internal(shost->transportt); \
1093 char *cp; \
1094 if (vport->flags & (FC_VPORT_DEL | FC_VPORT_CREATING)) \
1095 return -EBUSY; \
1096 val = simple_strtoul(buf, &cp, 0); \
1097 if (*cp && (*cp != '\n')) \
1098 return -EINVAL; \
1099 i->f->set_vport_##field(vport, val); \
1100 return count; \
1101}
1102
1103#define fc_vport_store_str_function(field, slen) \
1104static ssize_t \
1105store_fc_vport_##field(struct device *dev, \
1106 struct device_attribute *attr, \
1107 const char *buf, size_t count) \
1108{ \
1109 struct fc_vport *vport = transport_class_to_vport(dev); \
1110 struct Scsi_Host *shost = vport_to_shost(vport); \
1111 struct fc_internal *i = to_fc_internal(shost->transportt); \
1112 unsigned int cnt=count; \
1113 \
1114 /* count may include a LF at end of string */ \
1115 if (buf[cnt-1] == '\n') \
1116 cnt--; \
1117 if (cnt > ((slen) - 1)) \
1118 return -EINVAL; \
1119 memcpy(vport->field, buf, cnt); \
1120 i->f->set_vport_##field(vport); \
1121 return count; \
1122}
1123
1124#define fc_vport_rd_attr(field, format_string, sz) \
1125 fc_vport_show_function(field, format_string, sz, ) \
1126static FC_DEVICE_ATTR(vport, field, S_IRUGO, \
1127 show_fc_vport_##field, NULL)
1128
1129#define fc_vport_rd_attr_cast(field, format_string, sz, cast) \
1130 fc_vport_show_function(field, format_string, sz, (cast)) \
1131static FC_DEVICE_ATTR(vport, field, S_IRUGO, \
1132 show_fc_vport_##field, NULL)
1133
1134#define fc_vport_rw_attr(field, format_string, sz) \
1135 fc_vport_show_function(field, format_string, sz, ) \
1136 fc_vport_store_function(field) \
1137static FC_DEVICE_ATTR(vport, field, S_IRUGO | S_IWUSR, \
1138 show_fc_vport_##field, \
1139 store_fc_vport_##field)
1140
1141#define fc_private_vport_show_function(field, format_string, sz, cast) \
1142static ssize_t \
1143show_fc_vport_##field (struct device *dev, \
1144 struct device_attribute *attr, char *buf) \
1145{ \
1146 struct fc_vport *vport = transport_class_to_vport(dev); \
1147 return snprintf(buf, sz, format_string, cast vport->field); \
1148}
1149
1150#define fc_private_vport_store_u32_function(field) \
1151static ssize_t \
1152store_fc_vport_##field(struct device *dev, \
1153 struct device_attribute *attr, \
1154 const char *buf, size_t count) \
1155{ \
1156 u32 val; \
1157 struct fc_vport *vport = transport_class_to_vport(dev); \
1158 char *cp; \
1159 if (vport->flags & (FC_VPORT_DEL | FC_VPORT_CREATING)) \
1160 return -EBUSY; \
1161 val = simple_strtoul(buf, &cp, 0); \
1162 if (*cp && (*cp != '\n')) \
1163 return -EINVAL; \
1164 vport->field = val; \
1165 return count; \
1166}
1167
1168
1169#define fc_private_vport_rd_attr(field, format_string, sz) \
1170 fc_private_vport_show_function(field, format_string, sz, ) \
1171static FC_DEVICE_ATTR(vport, field, S_IRUGO, \
1172 show_fc_vport_##field, NULL)
1173
1174#define fc_private_vport_rd_attr_cast(field, format_string, sz, cast) \
1175 fc_private_vport_show_function(field, format_string, sz, (cast)) \
1176static FC_DEVICE_ATTR(vport, field, S_IRUGO, \
1177 show_fc_vport_##field, NULL)
1178
1179#define fc_private_vport_rw_u32_attr(field, format_string, sz) \
1180 fc_private_vport_show_function(field, format_string, sz, ) \
1181 fc_private_vport_store_u32_function(field) \
1182static FC_DEVICE_ATTR(vport, field, S_IRUGO | S_IWUSR, \
1183 show_fc_vport_##field, \
1184 store_fc_vport_##field)
1185
1186
1187#define fc_private_vport_rd_enum_attr(title, maxlen) \
1188static ssize_t \
1189show_fc_vport_##title (struct device *dev, \
1190 struct device_attribute *attr, \
1191 char *buf) \
1192{ \
1193 struct fc_vport *vport = transport_class_to_vport(dev); \
1194 const char *name; \
1195 name = get_fc_##title##_name(vport->title); \
1196 if (!name) \
1197 return -EINVAL; \
1198 return snprintf(buf, maxlen, "%s\n", name); \
1199} \
1200static FC_DEVICE_ATTR(vport, title, S_IRUGO, \
1201 show_fc_vport_##title, NULL)
1202
1203
1204#define SETUP_VPORT_ATTRIBUTE_RD(field) \
1205 i->private_vport_attrs[count] = device_attr_vport_##field; \
1206 i->private_vport_attrs[count].attr.mode = S_IRUGO; \
1207 i->private_vport_attrs[count].store = NULL; \
1208 i->vport_attrs[count] = &i->private_vport_attrs[count]; \
1209 if (i->f->get_##field) \
1210 count++
1211 /* NOTE: Above MACRO differs: checks function not show bit */
1212
1213#define SETUP_PRIVATE_VPORT_ATTRIBUTE_RD(field) \
1214 i->private_vport_attrs[count] = device_attr_vport_##field; \
1215 i->private_vport_attrs[count].attr.mode = S_IRUGO; \
1216 i->private_vport_attrs[count].store = NULL; \
1217 i->vport_attrs[count] = &i->private_vport_attrs[count]; \
1218 count++
1219
1220#define SETUP_VPORT_ATTRIBUTE_WR(field) \
1221 i->private_vport_attrs[count] = device_attr_vport_##field; \
1222 i->vport_attrs[count] = &i->private_vport_attrs[count]; \
1223 if (i->f->field) \
1224 count++
1225 /* NOTE: Above MACRO differs: checks function */
1226
1227#define SETUP_VPORT_ATTRIBUTE_RW(field) \
1228 i->private_vport_attrs[count] = device_attr_vport_##field; \
1229 if (!i->f->set_vport_##field) { \
1230 i->private_vport_attrs[count].attr.mode = S_IRUGO; \
1231 i->private_vport_attrs[count].store = NULL; \
1232 } \
1233 i->vport_attrs[count] = &i->private_vport_attrs[count]; \
1234 count++
1235 /* NOTE: Above MACRO differs: does not check show bit */
1236
1237#define SETUP_PRIVATE_VPORT_ATTRIBUTE_RW(field) \
1238{ \
1239 i->private_vport_attrs[count] = device_attr_vport_##field; \
1240 i->vport_attrs[count] = &i->private_vport_attrs[count]; \
1241 count++; \
1242}
1243
1244
1245/* The FC Transport Virtual Port Attributes: */
1246
1247/* Fixed Virtual Port Attributes */
1248
1249/* Dynamic Virtual Port Attributes */
1250
1251/* Private Virtual Port Attributes */
1252
1253fc_private_vport_rd_enum_attr(vport_state, FC_VPORTSTATE_MAX_NAMELEN);
1254fc_private_vport_rd_enum_attr(vport_last_state, FC_VPORTSTATE_MAX_NAMELEN);
1255fc_private_vport_rd_attr_cast(node_name, "0x%llx\n", 20, unsigned long long);
1256fc_private_vport_rd_attr_cast(port_name, "0x%llx\n", 20, unsigned long long);
1257
1258static ssize_t
1259show_fc_vport_roles (struct device *dev, struct device_attribute *attr,
1260 char *buf)
1261{
1262 struct fc_vport *vport = transport_class_to_vport(dev);
1263
1264 if (vport->roles == FC_PORT_ROLE_UNKNOWN)
1265 return snprintf(buf, 20, "unknown\n");
1266 return get_fc_port_roles_names(vport->roles, buf);
1267}
1268static FC_DEVICE_ATTR(vport, roles, S_IRUGO, show_fc_vport_roles, NULL);
1269
1270fc_private_vport_rd_enum_attr(vport_type, FC_PORTTYPE_MAX_NAMELEN);
1271
1272fc_private_vport_show_function(symbolic_name, "%s\n",
1273 FC_VPORT_SYMBOLIC_NAMELEN + 1, )
1274fc_vport_store_str_function(symbolic_name, FC_VPORT_SYMBOLIC_NAMELEN)
1275static FC_DEVICE_ATTR(vport, symbolic_name, S_IRUGO | S_IWUSR,
1276 show_fc_vport_symbolic_name, store_fc_vport_symbolic_name);
1277
1278static ssize_t
1279store_fc_vport_delete(struct device *dev, struct device_attribute *attr,
1280 const char *buf, size_t count)
1281{
1282 struct fc_vport *vport = transport_class_to_vport(dev);
1283 struct Scsi_Host *shost = vport_to_shost(vport);
1284 unsigned long flags;
1285
1286 spin_lock_irqsave(shost->host_lock, flags);
1287 if (vport->flags & (FC_VPORT_DEL | FC_VPORT_CREATING)) {
1288 spin_unlock_irqrestore(shost->host_lock, flags);
1289 return -EBUSY;
1290 }
1291 vport->flags |= FC_VPORT_DELETING;
1292 spin_unlock_irqrestore(shost->host_lock, flags);
1293
1294 fc_queue_work(shost, &vport->vport_delete_work);
1295 return count;
1296}
1297static FC_DEVICE_ATTR(vport, vport_delete, S_IWUSR,
1298 NULL, store_fc_vport_delete);
1299
1300
1301/*
1302 * Enable/Disable vport
1303 * Write "1" to disable, write "0" to enable
1304 */
1305static ssize_t
1306store_fc_vport_disable(struct device *dev, struct device_attribute *attr,
1307 const char *buf,
1308 size_t count)
1309{
1310 struct fc_vport *vport = transport_class_to_vport(dev);
1311 struct Scsi_Host *shost = vport_to_shost(vport);
1312 struct fc_internal *i = to_fc_internal(shost->transportt);
1313 int stat;
1314
1315 if (vport->flags & (FC_VPORT_DEL | FC_VPORT_CREATING))
1316 return -EBUSY;
1317
1318 if (*buf == '0') {
1319 if (vport->vport_state != FC_VPORT_DISABLED)
1320 return -EALREADY;
1321 } else if (*buf == '1') {
1322 if (vport->vport_state == FC_VPORT_DISABLED)
1323 return -EALREADY;
1324 } else
1325 return -EINVAL;
1326
1327 stat = i->f->vport_disable(vport, ((*buf == '0') ? false : true));
1328 return stat ? stat : count;
1329}
1330static FC_DEVICE_ATTR(vport, vport_disable, S_IWUSR,
1331 NULL, store_fc_vport_disable);
1332
1333
1334/*
1335 * Host Attribute Management
1336 */
1337
1338#define fc_host_show_function(field, format_string, sz, cast) \
1339static ssize_t \
1340show_fc_host_##field (struct device *dev, \
1341 struct device_attribute *attr, char *buf) \
1342{ \
1343 struct Scsi_Host *shost = transport_class_to_shost(dev); \
1344 struct fc_internal *i = to_fc_internal(shost->transportt); \
1345 if (i->f->get_host_##field) \
1346 i->f->get_host_##field(shost); \
1347 return snprintf(buf, sz, format_string, cast fc_host_##field(shost)); \
1348}
1349
1350#define fc_host_store_function(field) \
1351static ssize_t \
1352store_fc_host_##field(struct device *dev, \
1353 struct device_attribute *attr, \
1354 const char *buf, size_t count) \
1355{ \
1356 int val; \
1357 struct Scsi_Host *shost = transport_class_to_shost(dev); \
1358 struct fc_internal *i = to_fc_internal(shost->transportt); \
1359 char *cp; \
1360 \
1361 val = simple_strtoul(buf, &cp, 0); \
1362 if (*cp && (*cp != '\n')) \
1363 return -EINVAL; \
1364 i->f->set_host_##field(shost, val); \
1365 return count; \
1366}
1367
1368#define fc_host_store_str_function(field, slen) \
1369static ssize_t \
1370store_fc_host_##field(struct device *dev, \
1371 struct device_attribute *attr, \
1372 const char *buf, size_t count) \
1373{ \
1374 struct Scsi_Host *shost = transport_class_to_shost(dev); \
1375 struct fc_internal *i = to_fc_internal(shost->transportt); \
1376 unsigned int cnt=count; \
1377 \
1378 /* count may include a LF at end of string */ \
1379 if (buf[cnt-1] == '\n') \
1380 cnt--; \
1381 if (cnt > ((slen) - 1)) \
1382 return -EINVAL; \
1383 memcpy(fc_host_##field(shost), buf, cnt); \
1384 i->f->set_host_##field(shost); \
1385 return count; \
1386}
1387
1388#define fc_host_rd_attr(field, format_string, sz) \
1389 fc_host_show_function(field, format_string, sz, ) \
1390static FC_DEVICE_ATTR(host, field, S_IRUGO, \
1391 show_fc_host_##field, NULL)
1392
1393#define fc_host_rd_attr_cast(field, format_string, sz, cast) \
1394 fc_host_show_function(field, format_string, sz, (cast)) \
1395static FC_DEVICE_ATTR(host, field, S_IRUGO, \
1396 show_fc_host_##field, NULL)
1397
1398#define fc_host_rw_attr(field, format_string, sz) \
1399 fc_host_show_function(field, format_string, sz, ) \
1400 fc_host_store_function(field) \
1401static FC_DEVICE_ATTR(host, field, S_IRUGO | S_IWUSR, \
1402 show_fc_host_##field, \
1403 store_fc_host_##field)
1404
1405#define fc_host_rd_enum_attr(title, maxlen) \
1406static ssize_t \
1407show_fc_host_##title (struct device *dev, \
1408 struct device_attribute *attr, char *buf) \
1409{ \
1410 struct Scsi_Host *shost = transport_class_to_shost(dev); \
1411 struct fc_internal *i = to_fc_internal(shost->transportt); \
1412 const char *name; \
1413 if (i->f->get_host_##title) \
1414 i->f->get_host_##title(shost); \
1415 name = get_fc_##title##_name(fc_host_##title(shost)); \
1416 if (!name) \
1417 return -EINVAL; \
1418 return snprintf(buf, maxlen, "%s\n", name); \
1419} \
1420static FC_DEVICE_ATTR(host, title, S_IRUGO, show_fc_host_##title, NULL)
1421
1422#define SETUP_HOST_ATTRIBUTE_RD(field) \
1423 i->private_host_attrs[count] = device_attr_host_##field; \
1424 i->private_host_attrs[count].attr.mode = S_IRUGO; \
1425 i->private_host_attrs[count].store = NULL; \
1426 i->host_attrs[count] = &i->private_host_attrs[count]; \
1427 if (i->f->show_host_##field) \
1428 count++
1429
1430#define SETUP_HOST_ATTRIBUTE_RD_NS(field) \
1431 i->private_host_attrs[count] = device_attr_host_##field; \
1432 i->private_host_attrs[count].attr.mode = S_IRUGO; \
1433 i->private_host_attrs[count].store = NULL; \
1434 i->host_attrs[count] = &i->private_host_attrs[count]; \
1435 count++
1436
1437#define SETUP_HOST_ATTRIBUTE_RW(field) \
1438 i->private_host_attrs[count] = device_attr_host_##field; \
1439 if (!i->f->set_host_##field) { \
1440 i->private_host_attrs[count].attr.mode = S_IRUGO; \
1441 i->private_host_attrs[count].store = NULL; \
1442 } \
1443 i->host_attrs[count] = &i->private_host_attrs[count]; \
1444 if (i->f->show_host_##field) \
1445 count++
1446
1447
1448#define fc_private_host_show_function(field, format_string, sz, cast) \
1449static ssize_t \
1450show_fc_host_##field (struct device *dev, \
1451 struct device_attribute *attr, char *buf) \
1452{ \
1453 struct Scsi_Host *shost = transport_class_to_shost(dev); \
1454 return snprintf(buf, sz, format_string, cast fc_host_##field(shost)); \
1455}
1456
1457#define fc_private_host_rd_attr(field, format_string, sz) \
1458 fc_private_host_show_function(field, format_string, sz, ) \
1459static FC_DEVICE_ATTR(host, field, S_IRUGO, \
1460 show_fc_host_##field, NULL)
1461
1462#define fc_private_host_rd_attr_cast(field, format_string, sz, cast) \
1463 fc_private_host_show_function(field, format_string, sz, (cast)) \
1464static FC_DEVICE_ATTR(host, field, S_IRUGO, \
1465 show_fc_host_##field, NULL)
1466
1467#define SETUP_PRIVATE_HOST_ATTRIBUTE_RD(field) \
1468 i->private_host_attrs[count] = device_attr_host_##field; \
1469 i->private_host_attrs[count].attr.mode = S_IRUGO; \
1470 i->private_host_attrs[count].store = NULL; \
1471 i->host_attrs[count] = &i->private_host_attrs[count]; \
1472 count++
1473
1474#define SETUP_PRIVATE_HOST_ATTRIBUTE_RW(field) \
1475{ \
1476 i->private_host_attrs[count] = device_attr_host_##field; \
1477 i->host_attrs[count] = &i->private_host_attrs[count]; \
1478 count++; \
1479}
1480
1481
1482/* Fixed Host Attributes */
1483
1484static ssize_t
1485show_fc_host_supported_classes (struct device *dev,
1486 struct device_attribute *attr, char *buf)
1487{
1488 struct Scsi_Host *shost = transport_class_to_shost(dev);
1489
1490 if (fc_host_supported_classes(shost) == FC_COS_UNSPECIFIED)
1491 return snprintf(buf, 20, "unspecified\n");
1492
1493 return get_fc_cos_names(fc_host_supported_classes(shost), buf);
1494}
1495static FC_DEVICE_ATTR(host, supported_classes, S_IRUGO,
1496 show_fc_host_supported_classes, NULL);
1497
1498static ssize_t
1499show_fc_host_supported_fc4s (struct device *dev,
1500 struct device_attribute *attr, char *buf)
1501{
1502 struct Scsi_Host *shost = transport_class_to_shost(dev);
1503 return (ssize_t)show_fc_fc4s(buf, fc_host_supported_fc4s(shost));
1504}
1505static FC_DEVICE_ATTR(host, supported_fc4s, S_IRUGO,
1506 show_fc_host_supported_fc4s, NULL);
1507
1508static ssize_t
1509show_fc_host_supported_speeds (struct device *dev,
1510 struct device_attribute *attr, char *buf)
1511{
1512 struct Scsi_Host *shost = transport_class_to_shost(dev);
1513
1514 if (fc_host_supported_speeds(shost) == FC_PORTSPEED_UNKNOWN)
1515 return snprintf(buf, 20, "unknown\n");
1516
1517 return get_fc_port_speed_names(fc_host_supported_speeds(shost), buf);
1518}
1519static FC_DEVICE_ATTR(host, supported_speeds, S_IRUGO,
1520 show_fc_host_supported_speeds, NULL);
1521
1522
1523fc_private_host_rd_attr_cast(node_name, "0x%llx\n", 20, unsigned long long);
1524fc_private_host_rd_attr_cast(port_name, "0x%llx\n", 20, unsigned long long);
1525fc_private_host_rd_attr_cast(permanent_port_name, "0x%llx\n", 20,
1526 unsigned long long);
1527fc_private_host_rd_attr(maxframe_size, "%u bytes\n", 20);
1528fc_private_host_rd_attr(max_npiv_vports, "%u\n", 20);
1529fc_private_host_rd_attr(serial_number, "%s\n", (FC_SERIAL_NUMBER_SIZE +1));
1530fc_private_host_rd_attr(manufacturer, "%s\n", FC_SERIAL_NUMBER_SIZE + 1);
1531fc_private_host_rd_attr(model, "%s\n", FC_SYMBOLIC_NAME_SIZE + 1);
1532fc_private_host_rd_attr(model_description, "%s\n", FC_SYMBOLIC_NAME_SIZE + 1);
1533fc_private_host_rd_attr(hardware_version, "%s\n", FC_VERSION_STRING_SIZE + 1);
1534fc_private_host_rd_attr(driver_version, "%s\n", FC_VERSION_STRING_SIZE + 1);
1535fc_private_host_rd_attr(firmware_version, "%s\n", FC_VERSION_STRING_SIZE + 1);
1536fc_private_host_rd_attr(optionrom_version, "%s\n", FC_VERSION_STRING_SIZE + 1);
1537
1538
1539/* Dynamic Host Attributes */
1540
1541static ssize_t
1542show_fc_host_active_fc4s (struct device *dev,
1543 struct device_attribute *attr, char *buf)
1544{
1545 struct Scsi_Host *shost = transport_class_to_shost(dev);
1546 struct fc_internal *i = to_fc_internal(shost->transportt);
1547
1548 if (i->f->get_host_active_fc4s)
1549 i->f->get_host_active_fc4s(shost);
1550
1551 return (ssize_t)show_fc_fc4s(buf, fc_host_active_fc4s(shost));
1552}
1553static FC_DEVICE_ATTR(host, active_fc4s, S_IRUGO,
1554 show_fc_host_active_fc4s, NULL);
1555
1556static ssize_t
1557show_fc_host_speed (struct device *dev,
1558 struct device_attribute *attr, char *buf)
1559{
1560 struct Scsi_Host *shost = transport_class_to_shost(dev);
1561 struct fc_internal *i = to_fc_internal(shost->transportt);
1562
1563 if (i->f->get_host_speed)
1564 i->f->get_host_speed(shost);
1565
1566 if (fc_host_speed(shost) == FC_PORTSPEED_UNKNOWN)
1567 return snprintf(buf, 20, "unknown\n");
1568
1569 return get_fc_port_speed_names(fc_host_speed(shost), buf);
1570}
1571static FC_DEVICE_ATTR(host, speed, S_IRUGO,
1572 show_fc_host_speed, NULL);
1573
1574
1575fc_host_rd_attr(port_id, "0x%06x\n", 20);
1576fc_host_rd_enum_attr(port_type, FC_PORTTYPE_MAX_NAMELEN);
1577fc_host_rd_enum_attr(port_state, FC_PORTSTATE_MAX_NAMELEN);
1578fc_host_rd_attr_cast(fabric_name, "0x%llx\n", 20, unsigned long long);
1579fc_host_rd_attr(symbolic_name, "%s\n", FC_SYMBOLIC_NAME_SIZE + 1);
1580
1581fc_private_host_show_function(system_hostname, "%s\n",
1582 FC_SYMBOLIC_NAME_SIZE + 1, )
1583fc_host_store_str_function(system_hostname, FC_SYMBOLIC_NAME_SIZE)
1584static FC_DEVICE_ATTR(host, system_hostname, S_IRUGO | S_IWUSR,
1585 show_fc_host_system_hostname, store_fc_host_system_hostname);
1586
1587
1588/* Private Host Attributes */
1589
1590static ssize_t
1591show_fc_private_host_tgtid_bind_type(struct device *dev,
1592 struct device_attribute *attr, char *buf)
1593{
1594 struct Scsi_Host *shost = transport_class_to_shost(dev);
1595 const char *name;
1596
1597 name = get_fc_tgtid_bind_type_name(fc_host_tgtid_bind_type(shost));
1598 if (!name)
1599 return -EINVAL;
1600 return snprintf(buf, FC_BINDTYPE_MAX_NAMELEN, "%s\n", name);
1601}
1602
1603#define get_list_head_entry(pos, head, member) \
1604 pos = list_entry((head)->next, typeof(*pos), member)
1605
1606static ssize_t
1607store_fc_private_host_tgtid_bind_type(struct device *dev,
1608 struct device_attribute *attr, const char *buf, size_t count)
1609{
1610 struct Scsi_Host *shost = transport_class_to_shost(dev);
1611 struct fc_rport *rport;
1612 enum fc_tgtid_binding_type val;
1613 unsigned long flags;
1614
1615 if (get_fc_tgtid_bind_type_match(buf, &val))
1616 return -EINVAL;
1617
1618 /* if changing bind type, purge all unused consistent bindings */
1619 if (val != fc_host_tgtid_bind_type(shost)) {
1620 spin_lock_irqsave(shost->host_lock, flags);
1621 while (!list_empty(&fc_host_rport_bindings(shost))) {
1622 get_list_head_entry(rport,
1623 &fc_host_rport_bindings(shost), peers);
1624 list_del(&rport->peers);
1625 rport->port_state = FC_PORTSTATE_DELETED;
1626 fc_queue_work(shost, &rport->rport_delete_work);
1627 }
1628 spin_unlock_irqrestore(shost->host_lock, flags);
1629 }
1630
1631 fc_host_tgtid_bind_type(shost) = val;
1632 return count;
1633}
1634
1635static FC_DEVICE_ATTR(host, tgtid_bind_type, S_IRUGO | S_IWUSR,
1636 show_fc_private_host_tgtid_bind_type,
1637 store_fc_private_host_tgtid_bind_type);
1638
1639static ssize_t
1640store_fc_private_host_issue_lip(struct device *dev,
1641 struct device_attribute *attr, const char *buf, size_t count)
1642{
1643 struct Scsi_Host *shost = transport_class_to_shost(dev);
1644 struct fc_internal *i = to_fc_internal(shost->transportt);
1645 int ret;
1646
1647 /* ignore any data value written to the attribute */
1648 if (i->f->issue_fc_host_lip) {
1649 ret = i->f->issue_fc_host_lip(shost);
1650 return ret ? ret: count;
1651 }
1652
1653 return -ENOENT;
1654}
1655
1656static FC_DEVICE_ATTR(host, issue_lip, S_IWUSR, NULL,
1657 store_fc_private_host_issue_lip);
1658
1659static ssize_t
1660store_fc_private_host_dev_loss_tmo(struct device *dev,
1661 struct device_attribute *attr,
1662 const char *buf, size_t count)
1663{
1664 struct Scsi_Host *shost = transport_class_to_shost(dev);
1665 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
1666 struct fc_rport *rport;
1667 unsigned long val, flags;
1668 int rc;
1669
1670 rc = fc_str_to_dev_loss(buf, &val);
1671 if (rc)
1672 return rc;
1673
1674 fc_host_dev_loss_tmo(shost) = val;
1675 spin_lock_irqsave(shost->host_lock, flags);
1676 list_for_each_entry(rport, &fc_host->rports, peers)
1677 fc_rport_set_dev_loss_tmo(rport, val);
1678 spin_unlock_irqrestore(shost->host_lock, flags);
1679 return count;
1680}
1681
1682fc_private_host_show_function(dev_loss_tmo, "%d\n", 20, );
1683static FC_DEVICE_ATTR(host, dev_loss_tmo, S_IRUGO | S_IWUSR,
1684 show_fc_host_dev_loss_tmo,
1685 store_fc_private_host_dev_loss_tmo);
1686
1687fc_private_host_rd_attr(npiv_vports_inuse, "%u\n", 20);
1688
1689/*
1690 * Host Statistics Management
1691 */
1692
1693/* Show a given an attribute in the statistics group */
1694static ssize_t
1695fc_stat_show(const struct device *dev, char *buf, unsigned long offset)
1696{
1697 struct Scsi_Host *shost = transport_class_to_shost(dev);
1698 struct fc_internal *i = to_fc_internal(shost->transportt);
1699 struct fc_host_statistics *stats;
1700 ssize_t ret = -ENOENT;
1701
1702 if (offset > sizeof(struct fc_host_statistics) ||
1703 offset % sizeof(u64) != 0)
1704 WARN_ON(1);
1705
1706 if (i->f->get_fc_host_stats) {
1707 stats = (i->f->get_fc_host_stats)(shost);
1708 if (stats)
1709 ret = snprintf(buf, 20, "0x%llx\n",
1710 (unsigned long long)*(u64 *)(((u8 *) stats) + offset));
1711 }
1712 return ret;
1713}
1714
1715
1716/* generate a read-only statistics attribute */
1717#define fc_host_statistic(name) \
1718static ssize_t show_fcstat_##name(struct device *cd, \
1719 struct device_attribute *attr, \
1720 char *buf) \
1721{ \
1722 return fc_stat_show(cd, buf, \
1723 offsetof(struct fc_host_statistics, name)); \
1724} \
1725static FC_DEVICE_ATTR(host, name, S_IRUGO, show_fcstat_##name, NULL)
1726
1727fc_host_statistic(seconds_since_last_reset);
1728fc_host_statistic(tx_frames);
1729fc_host_statistic(tx_words);
1730fc_host_statistic(rx_frames);
1731fc_host_statistic(rx_words);
1732fc_host_statistic(lip_count);
1733fc_host_statistic(nos_count);
1734fc_host_statistic(error_frames);
1735fc_host_statistic(dumped_frames);
1736fc_host_statistic(link_failure_count);
1737fc_host_statistic(loss_of_sync_count);
1738fc_host_statistic(loss_of_signal_count);
1739fc_host_statistic(prim_seq_protocol_err_count);
1740fc_host_statistic(invalid_tx_word_count);
1741fc_host_statistic(invalid_crc_count);
1742fc_host_statistic(fcp_input_requests);
1743fc_host_statistic(fcp_output_requests);
1744fc_host_statistic(fcp_control_requests);
1745fc_host_statistic(fcp_input_megabytes);
1746fc_host_statistic(fcp_output_megabytes);
1747
1748static ssize_t
1749fc_reset_statistics(struct device *dev, struct device_attribute *attr,
1750 const char *buf, size_t count)
1751{
1752 struct Scsi_Host *shost = transport_class_to_shost(dev);
1753 struct fc_internal *i = to_fc_internal(shost->transportt);
1754
1755 /* ignore any data value written to the attribute */
1756 if (i->f->reset_fc_host_stats) {
1757 i->f->reset_fc_host_stats(shost);
1758 return count;
1759 }
1760
1761 return -ENOENT;
1762}
1763static FC_DEVICE_ATTR(host, reset_statistics, S_IWUSR, NULL,
1764 fc_reset_statistics);
1765
1766static struct attribute *fc_statistics_attrs[] = {
1767 &device_attr_host_seconds_since_last_reset.attr,
1768 &device_attr_host_tx_frames.attr,
1769 &device_attr_host_tx_words.attr,
1770 &device_attr_host_rx_frames.attr,
1771 &device_attr_host_rx_words.attr,
1772 &device_attr_host_lip_count.attr,
1773 &device_attr_host_nos_count.attr,
1774 &device_attr_host_error_frames.attr,
1775 &device_attr_host_dumped_frames.attr,
1776 &device_attr_host_link_failure_count.attr,
1777 &device_attr_host_loss_of_sync_count.attr,
1778 &device_attr_host_loss_of_signal_count.attr,
1779 &device_attr_host_prim_seq_protocol_err_count.attr,
1780 &device_attr_host_invalid_tx_word_count.attr,
1781 &device_attr_host_invalid_crc_count.attr,
1782 &device_attr_host_fcp_input_requests.attr,
1783 &device_attr_host_fcp_output_requests.attr,
1784 &device_attr_host_fcp_control_requests.attr,
1785 &device_attr_host_fcp_input_megabytes.attr,
1786 &device_attr_host_fcp_output_megabytes.attr,
1787 &device_attr_host_reset_statistics.attr,
1788 NULL
1789};
1790
1791static struct attribute_group fc_statistics_group = {
1792 .name = "statistics",
1793 .attrs = fc_statistics_attrs,
1794};
1795
1796
1797/* Host Vport Attributes */
1798
1799static int
1800fc_parse_wwn(const char *ns, u64 *nm)
1801{
1802 unsigned int i, j;
1803 u8 wwn[8];
1804
1805 memset(wwn, 0, sizeof(wwn));
1806
1807 /* Validate and store the new name */
1808 for (i=0, j=0; i < 16; i++) {
1809 int value;
1810
1811 value = hex_to_bin(*ns++);
1812 if (value >= 0)
1813 j = (j << 4) | value;
1814 else
1815 return -EINVAL;
1816 if (i % 2) {
1817 wwn[i/2] = j & 0xff;
1818 j = 0;
1819 }
1820 }
1821
1822 *nm = wwn_to_u64(wwn);
1823
1824 return 0;
1825}
1826
1827
1828/*
1829 * "Short-cut" sysfs variable to create a new vport on a FC Host.
1830 * Input is a string of the form "<WWPN>:<WWNN>". Other attributes
1831 * will default to a NPIV-based FCP_Initiator; The WWNs are specified
1832 * as hex characters, and may *not* contain any prefixes (e.g. 0x, x, etc)
1833 */
1834static ssize_t
1835store_fc_host_vport_create(struct device *dev, struct device_attribute *attr,
1836 const char *buf, size_t count)
1837{
1838 struct Scsi_Host *shost = transport_class_to_shost(dev);
1839 struct fc_vport_identifiers vid;
1840 struct fc_vport *vport;
1841 unsigned int cnt=count;
1842 int stat;
1843
1844 memset(&vid, 0, sizeof(vid));
1845
1846 /* count may include a LF at end of string */
1847 if (buf[cnt-1] == '\n')
1848 cnt--;
1849
1850 /* validate we have enough characters for WWPN */
1851 if ((cnt != (16+1+16)) || (buf[16] != ':'))
1852 return -EINVAL;
1853
1854 stat = fc_parse_wwn(&buf[0], &vid.port_name);
1855 if (stat)
1856 return stat;
1857
1858 stat = fc_parse_wwn(&buf[17], &vid.node_name);
1859 if (stat)
1860 return stat;
1861
1862 vid.roles = FC_PORT_ROLE_FCP_INITIATOR;
1863 vid.vport_type = FC_PORTTYPE_NPIV;
1864 /* vid.symbolic_name is already zero/NULL's */
1865 vid.disable = false; /* always enabled */
1866
1867 /* we only allow support on Channel 0 !!! */
1868 stat = fc_vport_setup(shost, 0, &shost->shost_gendev, &vid, &vport);
1869 return stat ? stat : count;
1870}
1871static FC_DEVICE_ATTR(host, vport_create, S_IWUSR, NULL,
1872 store_fc_host_vport_create);
1873
1874
1875/*
1876 * "Short-cut" sysfs variable to delete a vport on a FC Host.
1877 * Vport is identified by a string containing "<WWPN>:<WWNN>".
1878 * The WWNs are specified as hex characters, and may *not* contain
1879 * any prefixes (e.g. 0x, x, etc)
1880 */
1881static ssize_t
1882store_fc_host_vport_delete(struct device *dev, struct device_attribute *attr,
1883 const char *buf, size_t count)
1884{
1885 struct Scsi_Host *shost = transport_class_to_shost(dev);
1886 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
1887 struct fc_vport *vport;
1888 u64 wwpn, wwnn;
1889 unsigned long flags;
1890 unsigned int cnt=count;
1891 int stat, match;
1892
1893 /* count may include a LF at end of string */
1894 if (buf[cnt-1] == '\n')
1895 cnt--;
1896
1897 /* validate we have enough characters for WWPN */
1898 if ((cnt != (16+1+16)) || (buf[16] != ':'))
1899 return -EINVAL;
1900
1901 stat = fc_parse_wwn(&buf[0], &wwpn);
1902 if (stat)
1903 return stat;
1904
1905 stat = fc_parse_wwn(&buf[17], &wwnn);
1906 if (stat)
1907 return stat;
1908
1909 spin_lock_irqsave(shost->host_lock, flags);
1910 match = 0;
1911 /* we only allow support on Channel 0 !!! */
1912 list_for_each_entry(vport, &fc_host->vports, peers) {
1913 if ((vport->channel == 0) &&
1914 (vport->port_name == wwpn) && (vport->node_name == wwnn)) {
1915 if (vport->flags & (FC_VPORT_DEL | FC_VPORT_CREATING))
1916 break;
1917 vport->flags |= FC_VPORT_DELETING;
1918 match = 1;
1919 break;
1920 }
1921 }
1922 spin_unlock_irqrestore(shost->host_lock, flags);
1923
1924 if (!match)
1925 return -ENODEV;
1926
1927 stat = fc_vport_terminate(vport);
1928 return stat ? stat : count;
1929}
1930static FC_DEVICE_ATTR(host, vport_delete, S_IWUSR, NULL,
1931 store_fc_host_vport_delete);
1932
1933
1934static int fc_host_match(struct attribute_container *cont,
1935 struct device *dev)
1936{
1937 struct Scsi_Host *shost;
1938 struct fc_internal *i;
1939
1940 if (!scsi_is_host_device(dev))
1941 return 0;
1942
1943 shost = dev_to_shost(dev);
1944 if (!shost->transportt || shost->transportt->host_attrs.ac.class
1945 != &fc_host_class.class)
1946 return 0;
1947
1948 i = to_fc_internal(shost->transportt);
1949
1950 return &i->t.host_attrs.ac == cont;
1951}
1952
1953static int fc_target_match(struct attribute_container *cont,
1954 struct device *dev)
1955{
1956 struct Scsi_Host *shost;
1957 struct fc_internal *i;
1958
1959 if (!scsi_is_target_device(dev))
1960 return 0;
1961
1962 shost = dev_to_shost(dev->parent);
1963 if (!shost->transportt || shost->transportt->host_attrs.ac.class
1964 != &fc_host_class.class)
1965 return 0;
1966
1967 i = to_fc_internal(shost->transportt);
1968
1969 return &i->t.target_attrs.ac == cont;
1970}
1971
1972static void fc_rport_dev_release(struct device *dev)
1973{
1974 struct fc_rport *rport = dev_to_rport(dev);
1975 put_device(dev->parent);
1976 kfree(rport);
1977}
1978
1979int scsi_is_fc_rport(const struct device *dev)
1980{
1981 return dev->release == fc_rport_dev_release;
1982}
1983EXPORT_SYMBOL(scsi_is_fc_rport);
1984
1985static int fc_rport_match(struct attribute_container *cont,
1986 struct device *dev)
1987{
1988 struct Scsi_Host *shost;
1989 struct fc_internal *i;
1990
1991 if (!scsi_is_fc_rport(dev))
1992 return 0;
1993
1994 shost = dev_to_shost(dev->parent);
1995 if (!shost->transportt || shost->transportt->host_attrs.ac.class
1996 != &fc_host_class.class)
1997 return 0;
1998
1999 i = to_fc_internal(shost->transportt);
2000
2001 return &i->rport_attr_cont.ac == cont;
2002}
2003
2004
2005static void fc_vport_dev_release(struct device *dev)
2006{
2007 struct fc_vport *vport = dev_to_vport(dev);
2008 put_device(dev->parent); /* release kobj parent */
2009 kfree(vport);
2010}
2011
2012int scsi_is_fc_vport(const struct device *dev)
2013{
2014 return dev->release == fc_vport_dev_release;
2015}
2016EXPORT_SYMBOL(scsi_is_fc_vport);
2017
2018static int fc_vport_match(struct attribute_container *cont,
2019 struct device *dev)
2020{
2021 struct fc_vport *vport;
2022 struct Scsi_Host *shost;
2023 struct fc_internal *i;
2024
2025 if (!scsi_is_fc_vport(dev))
2026 return 0;
2027 vport = dev_to_vport(dev);
2028
2029 shost = vport_to_shost(vport);
2030 if (!shost->transportt || shost->transportt->host_attrs.ac.class
2031 != &fc_host_class.class)
2032 return 0;
2033
2034 i = to_fc_internal(shost->transportt);
2035 return &i->vport_attr_cont.ac == cont;
2036}
2037
2038
2039/**
2040 * fc_timed_out - FC Transport I/O timeout intercept handler
2041 * @scmd: The SCSI command which timed out
2042 *
2043 * This routine protects against error handlers getting invoked while a
2044 * rport is in a blocked state, typically due to a temporarily loss of
2045 * connectivity. If the error handlers are allowed to proceed, requests
2046 * to abort i/o, reset the target, etc will likely fail as there is no way
2047 * to communicate with the device to perform the requested function. These
2048 * failures may result in the midlayer taking the device offline, requiring
2049 * manual intervention to restore operation.
2050 *
2051 * This routine, called whenever an i/o times out, validates the state of
2052 * the underlying rport. If the rport is blocked, it returns
2053 * EH_RESET_TIMER, which will continue to reschedule the timeout.
2054 * Eventually, either the device will return, or devloss_tmo will fire,
2055 * and when the timeout then fires, it will be handled normally.
2056 * If the rport is not blocked, normal error handling continues.
2057 *
2058 * Notes:
2059 * This routine assumes no locks are held on entry.
2060 */
2061static enum blk_eh_timer_return
2062fc_timed_out(struct scsi_cmnd *scmd)
2063{
2064 struct fc_rport *rport = starget_to_rport(scsi_target(scmd->device));
2065
2066 if (rport->port_state == FC_PORTSTATE_BLOCKED)
2067 return BLK_EH_RESET_TIMER;
2068
2069 return BLK_EH_NOT_HANDLED;
2070}
2071
2072/*
2073 * Called by fc_user_scan to locate an rport on the shost that
2074 * matches the channel and target id, and invoke scsi_scan_target()
2075 * on the rport.
2076 */
2077static void
2078fc_user_scan_tgt(struct Scsi_Host *shost, uint channel, uint id, uint lun)
2079{
2080 struct fc_rport *rport;
2081 unsigned long flags;
2082
2083 spin_lock_irqsave(shost->host_lock, flags);
2084
2085 list_for_each_entry(rport, &fc_host_rports(shost), peers) {
2086 if (rport->scsi_target_id == -1)
2087 continue;
2088
2089 if (rport->port_state != FC_PORTSTATE_ONLINE)
2090 continue;
2091
2092 if ((channel == rport->channel) &&
2093 (id == rport->scsi_target_id)) {
2094 spin_unlock_irqrestore(shost->host_lock, flags);
2095 scsi_scan_target(&rport->dev, channel, id, lun, 1);
2096 return;
2097 }
2098 }
2099
2100 spin_unlock_irqrestore(shost->host_lock, flags);
2101}
2102
2103/*
2104 * Called via sysfs scan routines. Necessary, as the FC transport
2105 * wants to place all target objects below the rport object. So this
2106 * routine must invoke the scsi_scan_target() routine with the rport
2107 * object as the parent.
2108 */
2109static int
2110fc_user_scan(struct Scsi_Host *shost, uint channel, uint id, uint lun)
2111{
2112 uint chlo, chhi;
2113 uint tgtlo, tgthi;
2114
2115 if (((channel != SCAN_WILD_CARD) && (channel > shost->max_channel)) ||
2116 ((id != SCAN_WILD_CARD) && (id >= shost->max_id)) ||
2117 ((lun != SCAN_WILD_CARD) && (lun > shost->max_lun)))
2118 return -EINVAL;
2119
2120 if (channel == SCAN_WILD_CARD) {
2121 chlo = 0;
2122 chhi = shost->max_channel + 1;
2123 } else {
2124 chlo = channel;
2125 chhi = channel + 1;
2126 }
2127
2128 if (id == SCAN_WILD_CARD) {
2129 tgtlo = 0;
2130 tgthi = shost->max_id;
2131 } else {
2132 tgtlo = id;
2133 tgthi = id + 1;
2134 }
2135
2136 for ( ; chlo < chhi; chlo++)
2137 for ( ; tgtlo < tgthi; tgtlo++)
2138 fc_user_scan_tgt(shost, chlo, tgtlo, lun);
2139
2140 return 0;
2141}
2142
2143static int fc_tsk_mgmt_response(struct Scsi_Host *shost, u64 nexus, u64 tm_id,
2144 int result)
2145{
2146 struct fc_internal *i = to_fc_internal(shost->transportt);
2147 return i->f->tsk_mgmt_response(shost, nexus, tm_id, result);
2148}
2149
2150static int fc_it_nexus_response(struct Scsi_Host *shost, u64 nexus, int result)
2151{
2152 struct fc_internal *i = to_fc_internal(shost->transportt);
2153 return i->f->it_nexus_response(shost, nexus, result);
2154}
2155
2156struct scsi_transport_template *
2157fc_attach_transport(struct fc_function_template *ft)
2158{
2159 int count;
2160 struct fc_internal *i = kzalloc(sizeof(struct fc_internal),
2161 GFP_KERNEL);
2162
2163 if (unlikely(!i))
2164 return NULL;
2165
2166 i->t.target_attrs.ac.attrs = &i->starget_attrs[0];
2167 i->t.target_attrs.ac.class = &fc_transport_class.class;
2168 i->t.target_attrs.ac.match = fc_target_match;
2169 i->t.target_size = sizeof(struct fc_starget_attrs);
2170 transport_container_register(&i->t.target_attrs);
2171
2172 i->t.host_attrs.ac.attrs = &i->host_attrs[0];
2173 i->t.host_attrs.ac.class = &fc_host_class.class;
2174 i->t.host_attrs.ac.match = fc_host_match;
2175 i->t.host_size = sizeof(struct fc_host_attrs);
2176 if (ft->get_fc_host_stats)
2177 i->t.host_attrs.statistics = &fc_statistics_group;
2178 transport_container_register(&i->t.host_attrs);
2179
2180 i->rport_attr_cont.ac.attrs = &i->rport_attrs[0];
2181 i->rport_attr_cont.ac.class = &fc_rport_class.class;
2182 i->rport_attr_cont.ac.match = fc_rport_match;
2183 transport_container_register(&i->rport_attr_cont);
2184
2185 i->vport_attr_cont.ac.attrs = &i->vport_attrs[0];
2186 i->vport_attr_cont.ac.class = &fc_vport_class.class;
2187 i->vport_attr_cont.ac.match = fc_vport_match;
2188 transport_container_register(&i->vport_attr_cont);
2189
2190 i->f = ft;
2191
2192 /* Transport uses the shost workq for scsi scanning */
2193 i->t.create_work_queue = 1;
2194
2195 i->t.eh_timed_out = fc_timed_out;
2196
2197 i->t.user_scan = fc_user_scan;
2198
2199 /* target-mode drivers' functions */
2200 i->t.tsk_mgmt_response = fc_tsk_mgmt_response;
2201 i->t.it_nexus_response = fc_it_nexus_response;
2202
2203 /*
2204 * Setup SCSI Target Attributes.
2205 */
2206 count = 0;
2207 SETUP_STARGET_ATTRIBUTE_RD(node_name);
2208 SETUP_STARGET_ATTRIBUTE_RD(port_name);
2209 SETUP_STARGET_ATTRIBUTE_RD(port_id);
2210
2211 BUG_ON(count > FC_STARGET_NUM_ATTRS);
2212
2213 i->starget_attrs[count] = NULL;
2214
2215
2216 /*
2217 * Setup SCSI Host Attributes.
2218 */
2219 count=0;
2220 SETUP_HOST_ATTRIBUTE_RD(node_name);
2221 SETUP_HOST_ATTRIBUTE_RD(port_name);
2222 SETUP_HOST_ATTRIBUTE_RD(permanent_port_name);
2223 SETUP_HOST_ATTRIBUTE_RD(supported_classes);
2224 SETUP_HOST_ATTRIBUTE_RD(supported_fc4s);
2225 SETUP_HOST_ATTRIBUTE_RD(supported_speeds);
2226 SETUP_HOST_ATTRIBUTE_RD(maxframe_size);
2227 if (ft->vport_create) {
2228 SETUP_HOST_ATTRIBUTE_RD_NS(max_npiv_vports);
2229 SETUP_HOST_ATTRIBUTE_RD_NS(npiv_vports_inuse);
2230 }
2231 SETUP_HOST_ATTRIBUTE_RD(serial_number);
2232 SETUP_HOST_ATTRIBUTE_RD(manufacturer);
2233 SETUP_HOST_ATTRIBUTE_RD(model);
2234 SETUP_HOST_ATTRIBUTE_RD(model_description);
2235 SETUP_HOST_ATTRIBUTE_RD(hardware_version);
2236 SETUP_HOST_ATTRIBUTE_RD(driver_version);
2237 SETUP_HOST_ATTRIBUTE_RD(firmware_version);
2238 SETUP_HOST_ATTRIBUTE_RD(optionrom_version);
2239
2240 SETUP_HOST_ATTRIBUTE_RD(port_id);
2241 SETUP_HOST_ATTRIBUTE_RD(port_type);
2242 SETUP_HOST_ATTRIBUTE_RD(port_state);
2243 SETUP_HOST_ATTRIBUTE_RD(active_fc4s);
2244 SETUP_HOST_ATTRIBUTE_RD(speed);
2245 SETUP_HOST_ATTRIBUTE_RD(fabric_name);
2246 SETUP_HOST_ATTRIBUTE_RD(symbolic_name);
2247 SETUP_HOST_ATTRIBUTE_RW(system_hostname);
2248
2249 /* Transport-managed attributes */
2250 SETUP_PRIVATE_HOST_ATTRIBUTE_RW(dev_loss_tmo);
2251 SETUP_PRIVATE_HOST_ATTRIBUTE_RW(tgtid_bind_type);
2252 if (ft->issue_fc_host_lip)
2253 SETUP_PRIVATE_HOST_ATTRIBUTE_RW(issue_lip);
2254 if (ft->vport_create)
2255 SETUP_PRIVATE_HOST_ATTRIBUTE_RW(vport_create);
2256 if (ft->vport_delete)
2257 SETUP_PRIVATE_HOST_ATTRIBUTE_RW(vport_delete);
2258
2259 BUG_ON(count > FC_HOST_NUM_ATTRS);
2260
2261 i->host_attrs[count] = NULL;
2262
2263 /*
2264 * Setup Remote Port Attributes.
2265 */
2266 count=0;
2267 SETUP_RPORT_ATTRIBUTE_RD(maxframe_size);
2268 SETUP_RPORT_ATTRIBUTE_RD(supported_classes);
2269 SETUP_RPORT_ATTRIBUTE_RW(dev_loss_tmo);
2270 SETUP_PRIVATE_RPORT_ATTRIBUTE_RD(node_name);
2271 SETUP_PRIVATE_RPORT_ATTRIBUTE_RD(port_name);
2272 SETUP_PRIVATE_RPORT_ATTRIBUTE_RD(port_id);
2273 SETUP_PRIVATE_RPORT_ATTRIBUTE_RD(roles);
2274 SETUP_PRIVATE_RPORT_ATTRIBUTE_RD(port_state);
2275 SETUP_PRIVATE_RPORT_ATTRIBUTE_RD(scsi_target_id);
2276 SETUP_PRIVATE_RPORT_ATTRIBUTE_RW(fast_io_fail_tmo);
2277
2278 BUG_ON(count > FC_RPORT_NUM_ATTRS);
2279
2280 i->rport_attrs[count] = NULL;
2281
2282 /*
2283 * Setup Virtual Port Attributes.
2284 */
2285 count=0;
2286 SETUP_PRIVATE_VPORT_ATTRIBUTE_RD(vport_state);
2287 SETUP_PRIVATE_VPORT_ATTRIBUTE_RD(vport_last_state);
2288 SETUP_PRIVATE_VPORT_ATTRIBUTE_RD(node_name);
2289 SETUP_PRIVATE_VPORT_ATTRIBUTE_RD(port_name);
2290 SETUP_PRIVATE_VPORT_ATTRIBUTE_RD(roles);
2291 SETUP_PRIVATE_VPORT_ATTRIBUTE_RD(vport_type);
2292 SETUP_VPORT_ATTRIBUTE_RW(symbolic_name);
2293 SETUP_VPORT_ATTRIBUTE_WR(vport_delete);
2294 SETUP_VPORT_ATTRIBUTE_WR(vport_disable);
2295
2296 BUG_ON(count > FC_VPORT_NUM_ATTRS);
2297
2298 i->vport_attrs[count] = NULL;
2299
2300 return &i->t;
2301}
2302EXPORT_SYMBOL(fc_attach_transport);
2303
2304void fc_release_transport(struct scsi_transport_template *t)
2305{
2306 struct fc_internal *i = to_fc_internal(t);
2307
2308 transport_container_unregister(&i->t.target_attrs);
2309 transport_container_unregister(&i->t.host_attrs);
2310 transport_container_unregister(&i->rport_attr_cont);
2311 transport_container_unregister(&i->vport_attr_cont);
2312
2313 kfree(i);
2314}
2315EXPORT_SYMBOL(fc_release_transport);
2316
2317/**
2318 * fc_queue_work - Queue work to the fc_host workqueue.
2319 * @shost: Pointer to Scsi_Host bound to fc_host.
2320 * @work: Work to queue for execution.
2321 *
2322 * Return value:
2323 * 1 - work queued for execution
2324 * 0 - work is already queued
2325 * -EINVAL - work queue doesn't exist
2326 */
2327static int
2328fc_queue_work(struct Scsi_Host *shost, struct work_struct *work)
2329{
2330 if (unlikely(!fc_host_work_q(shost))) {
2331 printk(KERN_ERR
2332 "ERROR: FC host '%s' attempted to queue work, "
2333 "when no workqueue created.\n", shost->hostt->name);
2334 dump_stack();
2335
2336 return -EINVAL;
2337 }
2338
2339 return queue_work(fc_host_work_q(shost), work);
2340}
2341
2342/**
2343 * fc_flush_work - Flush a fc_host's workqueue.
2344 * @shost: Pointer to Scsi_Host bound to fc_host.
2345 */
2346static void
2347fc_flush_work(struct Scsi_Host *shost)
2348{
2349 if (!fc_host_work_q(shost)) {
2350 printk(KERN_ERR
2351 "ERROR: FC host '%s' attempted to flush work, "
2352 "when no workqueue created.\n", shost->hostt->name);
2353 dump_stack();
2354 return;
2355 }
2356
2357 flush_workqueue(fc_host_work_q(shost));
2358}
2359
2360/**
2361 * fc_queue_devloss_work - Schedule work for the fc_host devloss workqueue.
2362 * @shost: Pointer to Scsi_Host bound to fc_host.
2363 * @work: Work to queue for execution.
2364 * @delay: jiffies to delay the work queuing
2365 *
2366 * Return value:
2367 * 1 on success / 0 already queued / < 0 for error
2368 */
2369static int
2370fc_queue_devloss_work(struct Scsi_Host *shost, struct delayed_work *work,
2371 unsigned long delay)
2372{
2373 if (unlikely(!fc_host_devloss_work_q(shost))) {
2374 printk(KERN_ERR
2375 "ERROR: FC host '%s' attempted to queue work, "
2376 "when no workqueue created.\n", shost->hostt->name);
2377 dump_stack();
2378
2379 return -EINVAL;
2380 }
2381
2382 return queue_delayed_work(fc_host_devloss_work_q(shost), work, delay);
2383}
2384
2385/**
2386 * fc_flush_devloss - Flush a fc_host's devloss workqueue.
2387 * @shost: Pointer to Scsi_Host bound to fc_host.
2388 */
2389static void
2390fc_flush_devloss(struct Scsi_Host *shost)
2391{
2392 if (!fc_host_devloss_work_q(shost)) {
2393 printk(KERN_ERR
2394 "ERROR: FC host '%s' attempted to flush work, "
2395 "when no workqueue created.\n", shost->hostt->name);
2396 dump_stack();
2397 return;
2398 }
2399
2400 flush_workqueue(fc_host_devloss_work_q(shost));
2401}
2402
2403
2404/**
2405 * fc_remove_host - called to terminate any fc_transport-related elements for a scsi host.
2406 * @shost: Which &Scsi_Host
2407 *
2408 * This routine is expected to be called immediately preceding the
2409 * a driver's call to scsi_remove_host().
2410 *
2411 * WARNING: A driver utilizing the fc_transport, which fails to call
2412 * this routine prior to scsi_remove_host(), will leave dangling
2413 * objects in /sys/class/fc_remote_ports. Access to any of these
2414 * objects can result in a system crash !!!
2415 *
2416 * Notes:
2417 * This routine assumes no locks are held on entry.
2418 */
2419void
2420fc_remove_host(struct Scsi_Host *shost)
2421{
2422 struct fc_vport *vport = NULL, *next_vport = NULL;
2423 struct fc_rport *rport = NULL, *next_rport = NULL;
2424 struct workqueue_struct *work_q;
2425 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
2426 unsigned long flags;
2427
2428 spin_lock_irqsave(shost->host_lock, flags);
2429
2430 /* Remove any vports */
2431 list_for_each_entry_safe(vport, next_vport, &fc_host->vports, peers)
2432 fc_queue_work(shost, &vport->vport_delete_work);
2433
2434 /* Remove any remote ports */
2435 list_for_each_entry_safe(rport, next_rport,
2436 &fc_host->rports, peers) {
2437 list_del(&rport->peers);
2438 rport->port_state = FC_PORTSTATE_DELETED;
2439 fc_queue_work(shost, &rport->rport_delete_work);
2440 }
2441
2442 list_for_each_entry_safe(rport, next_rport,
2443 &fc_host->rport_bindings, peers) {
2444 list_del(&rport->peers);
2445 rport->port_state = FC_PORTSTATE_DELETED;
2446 fc_queue_work(shost, &rport->rport_delete_work);
2447 }
2448
2449 spin_unlock_irqrestore(shost->host_lock, flags);
2450
2451 /* flush all scan work items */
2452 scsi_flush_work(shost);
2453
2454 /* flush all stgt delete, and rport delete work items, then kill it */
2455 if (fc_host->work_q) {
2456 work_q = fc_host->work_q;
2457 fc_host->work_q = NULL;
2458 destroy_workqueue(work_q);
2459 }
2460
2461 /* flush all devloss work items, then kill it */
2462 if (fc_host->devloss_work_q) {
2463 work_q = fc_host->devloss_work_q;
2464 fc_host->devloss_work_q = NULL;
2465 destroy_workqueue(work_q);
2466 }
2467}
2468EXPORT_SYMBOL(fc_remove_host);
2469
2470static void fc_terminate_rport_io(struct fc_rport *rport)
2471{
2472 struct Scsi_Host *shost = rport_to_shost(rport);
2473 struct fc_internal *i = to_fc_internal(shost->transportt);
2474
2475 /* Involve the LLDD if possible to terminate all io on the rport. */
2476 if (i->f->terminate_rport_io)
2477 i->f->terminate_rport_io(rport);
2478
2479 /*
2480 * must unblock to flush queued IO. The caller will have set
2481 * the port_state or flags, so that fc_remote_port_chkready will
2482 * fail IO.
2483 */
2484 scsi_target_unblock(&rport->dev);
2485}
2486
2487/**
2488 * fc_starget_delete - called to delete the scsi descendants of an rport
2489 * @work: remote port to be operated on.
2490 *
2491 * Deletes target and all sdevs.
2492 */
2493static void
2494fc_starget_delete(struct work_struct *work)
2495{
2496 struct fc_rport *rport =
2497 container_of(work, struct fc_rport, stgt_delete_work);
2498
2499 fc_terminate_rport_io(rport);
2500 scsi_remove_target(&rport->dev);
2501}
2502
2503
2504/**
2505 * fc_rport_final_delete - finish rport termination and delete it.
2506 * @work: remote port to be deleted.
2507 */
2508static void
2509fc_rport_final_delete(struct work_struct *work)
2510{
2511 struct fc_rport *rport =
2512 container_of(work, struct fc_rport, rport_delete_work);
2513 struct device *dev = &rport->dev;
2514 struct Scsi_Host *shost = rport_to_shost(rport);
2515 struct fc_internal *i = to_fc_internal(shost->transportt);
2516 unsigned long flags;
2517 int do_callback = 0;
2518
2519 fc_terminate_rport_io(rport);
2520
2521 /*
2522 * if a scan is pending, flush the SCSI Host work_q so that
2523 * that we can reclaim the rport scan work element.
2524 */
2525 if (rport->flags & FC_RPORT_SCAN_PENDING)
2526 scsi_flush_work(shost);
2527
2528 /*
2529 * Cancel any outstanding timers. These should really exist
2530 * only when rmmod'ing the LLDD and we're asking for
2531 * immediate termination of the rports
2532 */
2533 spin_lock_irqsave(shost->host_lock, flags);
2534 if (rport->flags & FC_RPORT_DEVLOSS_PENDING) {
2535 spin_unlock_irqrestore(shost->host_lock, flags);
2536 if (!cancel_delayed_work(&rport->fail_io_work))
2537 fc_flush_devloss(shost);
2538 if (!cancel_delayed_work(&rport->dev_loss_work))
2539 fc_flush_devloss(shost);
2540 spin_lock_irqsave(shost->host_lock, flags);
2541 rport->flags &= ~FC_RPORT_DEVLOSS_PENDING;
2542 }
2543 spin_unlock_irqrestore(shost->host_lock, flags);
2544
2545 /* Delete SCSI target and sdevs */
2546 if (rport->scsi_target_id != -1)
2547 fc_starget_delete(&rport->stgt_delete_work);
2548
2549 /*
2550 * Notify the driver that the rport is now dead. The LLDD will
2551 * also guarantee that any communication to the rport is terminated
2552 *
2553 * Avoid this call if we already called it when we preserved the
2554 * rport for the binding.
2555 */
2556 spin_lock_irqsave(shost->host_lock, flags);
2557 if (!(rport->flags & FC_RPORT_DEVLOSS_CALLBK_DONE) &&
2558 (i->f->dev_loss_tmo_callbk)) {
2559 rport->flags |= FC_RPORT_DEVLOSS_CALLBK_DONE;
2560 do_callback = 1;
2561 }
2562 spin_unlock_irqrestore(shost->host_lock, flags);
2563
2564 if (do_callback)
2565 i->f->dev_loss_tmo_callbk(rport);
2566
2567 fc_bsg_remove(rport->rqst_q);
2568
2569 transport_remove_device(dev);
2570 device_del(dev);
2571 transport_destroy_device(dev);
2572 put_device(&shost->shost_gendev); /* for fc_host->rport list */
2573 put_device(dev); /* for self-reference */
2574}
2575
2576
2577/**
2578 * fc_rport_create - allocates and creates a remote FC port.
2579 * @shost: scsi host the remote port is connected to.
2580 * @channel: Channel on shost port connected to.
2581 * @ids: The world wide names, fc address, and FC4 port
2582 * roles for the remote port.
2583 *
2584 * Allocates and creates the remoter port structure, including the
2585 * class and sysfs creation.
2586 *
2587 * Notes:
2588 * This routine assumes no locks are held on entry.
2589 */
2590static struct fc_rport *
2591fc_rport_create(struct Scsi_Host *shost, int channel,
2592 struct fc_rport_identifiers *ids)
2593{
2594 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
2595 struct fc_internal *fci = to_fc_internal(shost->transportt);
2596 struct fc_rport *rport;
2597 struct device *dev;
2598 unsigned long flags;
2599 int error;
2600 size_t size;
2601
2602 size = (sizeof(struct fc_rport) + fci->f->dd_fcrport_size);
2603 rport = kzalloc(size, GFP_KERNEL);
2604 if (unlikely(!rport)) {
2605 printk(KERN_ERR "%s: allocation failure\n", __func__);
2606 return NULL;
2607 }
2608
2609 rport->maxframe_size = -1;
2610 rport->supported_classes = FC_COS_UNSPECIFIED;
2611 rport->dev_loss_tmo = fc_host->dev_loss_tmo;
2612 memcpy(&rport->node_name, &ids->node_name, sizeof(rport->node_name));
2613 memcpy(&rport->port_name, &ids->port_name, sizeof(rport->port_name));
2614 rport->port_id = ids->port_id;
2615 rport->roles = ids->roles;
2616 rport->port_state = FC_PORTSTATE_ONLINE;
2617 if (fci->f->dd_fcrport_size)
2618 rport->dd_data = &rport[1];
2619 rport->channel = channel;
2620 rport->fast_io_fail_tmo = -1;
2621
2622 INIT_DELAYED_WORK(&rport->dev_loss_work, fc_timeout_deleted_rport);
2623 INIT_DELAYED_WORK(&rport->fail_io_work, fc_timeout_fail_rport_io);
2624 INIT_WORK(&rport->scan_work, fc_scsi_scan_rport);
2625 INIT_WORK(&rport->stgt_delete_work, fc_starget_delete);
2626 INIT_WORK(&rport->rport_delete_work, fc_rport_final_delete);
2627
2628 spin_lock_irqsave(shost->host_lock, flags);
2629
2630 rport->number = fc_host->next_rport_number++;
2631 if (rport->roles & FC_PORT_ROLE_FCP_TARGET)
2632 rport->scsi_target_id = fc_host->next_target_id++;
2633 else
2634 rport->scsi_target_id = -1;
2635 list_add_tail(&rport->peers, &fc_host->rports);
2636 get_device(&shost->shost_gendev); /* for fc_host->rport list */
2637
2638 spin_unlock_irqrestore(shost->host_lock, flags);
2639
2640 dev = &rport->dev;
2641 device_initialize(dev); /* takes self reference */
2642 dev->parent = get_device(&shost->shost_gendev); /* parent reference */
2643 dev->release = fc_rport_dev_release;
2644 dev_set_name(dev, "rport-%d:%d-%d",
2645 shost->host_no, channel, rport->number);
2646 transport_setup_device(dev);
2647
2648 error = device_add(dev);
2649 if (error) {
2650 printk(KERN_ERR "FC Remote Port device_add failed\n");
2651 goto delete_rport;
2652 }
2653 transport_add_device(dev);
2654 transport_configure_device(dev);
2655
2656 fc_bsg_rportadd(shost, rport);
2657 /* ignore any bsg add error - we just can't do sgio */
2658
2659 if (rport->roles & FC_PORT_ROLE_FCP_TARGET) {
2660 /* initiate a scan of the target */
2661 rport->flags |= FC_RPORT_SCAN_PENDING;
2662 scsi_queue_work(shost, &rport->scan_work);
2663 }
2664
2665 return rport;
2666
2667delete_rport:
2668 transport_destroy_device(dev);
2669 spin_lock_irqsave(shost->host_lock, flags);
2670 list_del(&rport->peers);
2671 put_device(&shost->shost_gendev); /* for fc_host->rport list */
2672 spin_unlock_irqrestore(shost->host_lock, flags);
2673 put_device(dev->parent);
2674 kfree(rport);
2675 return NULL;
2676}
2677
2678/**
2679 * fc_remote_port_add - notify fc transport of the existence of a remote FC port.
2680 * @shost: scsi host the remote port is connected to.
2681 * @channel: Channel on shost port connected to.
2682 * @ids: The world wide names, fc address, and FC4 port
2683 * roles for the remote port.
2684 *
2685 * The LLDD calls this routine to notify the transport of the existence
2686 * of a remote port. The LLDD provides the unique identifiers (wwpn,wwn)
2687 * of the port, it's FC address (port_id), and the FC4 roles that are
2688 * active for the port.
2689 *
2690 * For ports that are FCP targets (aka scsi targets), the FC transport
2691 * maintains consistent target id bindings on behalf of the LLDD.
2692 * A consistent target id binding is an assignment of a target id to
2693 * a remote port identifier, which persists while the scsi host is
2694 * attached. The remote port can disappear, then later reappear, and
2695 * it's target id assignment remains the same. This allows for shifts
2696 * in FC addressing (if binding by wwpn or wwnn) with no apparent
2697 * changes to the scsi subsystem which is based on scsi host number and
2698 * target id values. Bindings are only valid during the attachment of
2699 * the scsi host. If the host detaches, then later re-attaches, target
2700 * id bindings may change.
2701 *
2702 * This routine is responsible for returning a remote port structure.
2703 * The routine will search the list of remote ports it maintains
2704 * internally on behalf of consistent target id mappings. If found, the
2705 * remote port structure will be reused. Otherwise, a new remote port
2706 * structure will be allocated.
2707 *
2708 * Whenever a remote port is allocated, a new fc_remote_port class
2709 * device is created.
2710 *
2711 * Should not be called from interrupt context.
2712 *
2713 * Notes:
2714 * This routine assumes no locks are held on entry.
2715 */
2716struct fc_rport *
2717fc_remote_port_add(struct Scsi_Host *shost, int channel,
2718 struct fc_rport_identifiers *ids)
2719{
2720 struct fc_internal *fci = to_fc_internal(shost->transportt);
2721 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
2722 struct fc_rport *rport;
2723 unsigned long flags;
2724 int match = 0;
2725
2726 /* ensure any stgt delete functions are done */
2727 fc_flush_work(shost);
2728
2729 /*
2730 * Search the list of "active" rports, for an rport that has been
2731 * deleted, but we've held off the real delete while the target
2732 * is in a "blocked" state.
2733 */
2734 spin_lock_irqsave(shost->host_lock, flags);
2735
2736 list_for_each_entry(rport, &fc_host->rports, peers) {
2737
2738 if ((rport->port_state == FC_PORTSTATE_BLOCKED) &&
2739 (rport->channel == channel)) {
2740
2741 switch (fc_host->tgtid_bind_type) {
2742 case FC_TGTID_BIND_BY_WWPN:
2743 case FC_TGTID_BIND_NONE:
2744 if (rport->port_name == ids->port_name)
2745 match = 1;
2746 break;
2747 case FC_TGTID_BIND_BY_WWNN:
2748 if (rport->node_name == ids->node_name)
2749 match = 1;
2750 break;
2751 case FC_TGTID_BIND_BY_ID:
2752 if (rport->port_id == ids->port_id)
2753 match = 1;
2754 break;
2755 }
2756
2757 if (match) {
2758
2759 memcpy(&rport->node_name, &ids->node_name,
2760 sizeof(rport->node_name));
2761 memcpy(&rport->port_name, &ids->port_name,
2762 sizeof(rport->port_name));
2763 rport->port_id = ids->port_id;
2764
2765 rport->port_state = FC_PORTSTATE_ONLINE;
2766 rport->roles = ids->roles;
2767
2768 spin_unlock_irqrestore(shost->host_lock, flags);
2769
2770 if (fci->f->dd_fcrport_size)
2771 memset(rport->dd_data, 0,
2772 fci->f->dd_fcrport_size);
2773
2774 /*
2775 * If we were not a target, cancel the
2776 * io terminate and rport timers, and
2777 * we're done.
2778 *
2779 * If we were a target, but our new role
2780 * doesn't indicate a target, leave the
2781 * timers running expecting the role to
2782 * change as the target fully logs in. If
2783 * it doesn't, the target will be torn down.
2784 *
2785 * If we were a target, and our role shows
2786 * we're still a target, cancel the timers
2787 * and kick off a scan.
2788 */
2789
2790 /* was a target, not in roles */
2791 if ((rport->scsi_target_id != -1) &&
2792 (!(ids->roles & FC_PORT_ROLE_FCP_TARGET)))
2793 return rport;
2794
2795 /*
2796 * Stop the fail io and dev_loss timers.
2797 * If they flush, the port_state will
2798 * be checked and will NOOP the function.
2799 */
2800 if (!cancel_delayed_work(&rport->fail_io_work))
2801 fc_flush_devloss(shost);
2802 if (!cancel_delayed_work(&rport->dev_loss_work))
2803 fc_flush_devloss(shost);
2804
2805 spin_lock_irqsave(shost->host_lock, flags);
2806
2807 rport->flags &= ~(FC_RPORT_FAST_FAIL_TIMEDOUT |
2808 FC_RPORT_DEVLOSS_PENDING |
2809 FC_RPORT_DEVLOSS_CALLBK_DONE);
2810
2811 spin_unlock_irqrestore(shost->host_lock, flags);
2812
2813 /* if target, initiate a scan */
2814 if (rport->scsi_target_id != -1) {
2815 scsi_target_unblock(&rport->dev);
2816
2817 spin_lock_irqsave(shost->host_lock,
2818 flags);
2819 rport->flags |= FC_RPORT_SCAN_PENDING;
2820 scsi_queue_work(shost,
2821 &rport->scan_work);
2822 spin_unlock_irqrestore(shost->host_lock,
2823 flags);
2824 }
2825
2826 fc_bsg_goose_queue(rport);
2827
2828 return rport;
2829 }
2830 }
2831 }
2832
2833 /*
2834 * Search the bindings array
2835 * Note: if never a FCP target, you won't be on this list
2836 */
2837 if (fc_host->tgtid_bind_type != FC_TGTID_BIND_NONE) {
2838
2839 /* search for a matching consistent binding */
2840
2841 list_for_each_entry(rport, &fc_host->rport_bindings,
2842 peers) {
2843 if (rport->channel != channel)
2844 continue;
2845
2846 switch (fc_host->tgtid_bind_type) {
2847 case FC_TGTID_BIND_BY_WWPN:
2848 if (rport->port_name == ids->port_name)
2849 match = 1;
2850 break;
2851 case FC_TGTID_BIND_BY_WWNN:
2852 if (rport->node_name == ids->node_name)
2853 match = 1;
2854 break;
2855 case FC_TGTID_BIND_BY_ID:
2856 if (rport->port_id == ids->port_id)
2857 match = 1;
2858 break;
2859 case FC_TGTID_BIND_NONE: /* to keep compiler happy */
2860 break;
2861 }
2862
2863 if (match) {
2864 list_move_tail(&rport->peers, &fc_host->rports);
2865 break;
2866 }
2867 }
2868
2869 if (match) {
2870 memcpy(&rport->node_name, &ids->node_name,
2871 sizeof(rport->node_name));
2872 memcpy(&rport->port_name, &ids->port_name,
2873 sizeof(rport->port_name));
2874 rport->port_id = ids->port_id;
2875 rport->roles = ids->roles;
2876 rport->port_state = FC_PORTSTATE_ONLINE;
2877 rport->flags &= ~FC_RPORT_FAST_FAIL_TIMEDOUT;
2878
2879 if (fci->f->dd_fcrport_size)
2880 memset(rport->dd_data, 0,
2881 fci->f->dd_fcrport_size);
2882 spin_unlock_irqrestore(shost->host_lock, flags);
2883
2884 if (ids->roles & FC_PORT_ROLE_FCP_TARGET) {
2885 scsi_target_unblock(&rport->dev);
2886
2887 /* initiate a scan of the target */
2888 spin_lock_irqsave(shost->host_lock, flags);
2889 rport->flags |= FC_RPORT_SCAN_PENDING;
2890 scsi_queue_work(shost, &rport->scan_work);
2891 spin_unlock_irqrestore(shost->host_lock, flags);
2892 }
2893 return rport;
2894 }
2895 }
2896
2897 spin_unlock_irqrestore(shost->host_lock, flags);
2898
2899 /* No consistent binding found - create new remote port entry */
2900 rport = fc_rport_create(shost, channel, ids);
2901
2902 return rport;
2903}
2904EXPORT_SYMBOL(fc_remote_port_add);
2905
2906
2907/**
2908 * fc_remote_port_delete - notifies the fc transport that a remote port is no longer in existence.
2909 * @rport: The remote port that no longer exists
2910 *
2911 * The LLDD calls this routine to notify the transport that a remote
2912 * port is no longer part of the topology. Note: Although a port
2913 * may no longer be part of the topology, it may persist in the remote
2914 * ports displayed by the fc_host. We do this under 2 conditions:
2915 * 1) If the port was a scsi target, we delay its deletion by "blocking" it.
2916 * This allows the port to temporarily disappear, then reappear without
2917 * disrupting the SCSI device tree attached to it. During the "blocked"
2918 * period the port will still exist.
2919 * 2) If the port was a scsi target and disappears for longer than we
2920 * expect, we'll delete the port and the tear down the SCSI device tree
2921 * attached to it. However, we want to semi-persist the target id assigned
2922 * to that port if it eventually does exist. The port structure will
2923 * remain (although with minimal information) so that the target id
2924 * bindings remails.
2925 *
2926 * If the remote port is not an FCP Target, it will be fully torn down
2927 * and deallocated, including the fc_remote_port class device.
2928 *
2929 * If the remote port is an FCP Target, the port will be placed in a
2930 * temporary blocked state. From the LLDD's perspective, the rport no
2931 * longer exists. From the SCSI midlayer's perspective, the SCSI target
2932 * exists, but all sdevs on it are blocked from further I/O. The following
2933 * is then expected.
2934 *
2935 * If the remote port does not return (signaled by a LLDD call to
2936 * fc_remote_port_add()) within the dev_loss_tmo timeout, then the
2937 * scsi target is removed - killing all outstanding i/o and removing the
2938 * scsi devices attached ot it. The port structure will be marked Not
2939 * Present and be partially cleared, leaving only enough information to
2940 * recognize the remote port relative to the scsi target id binding if
2941 * it later appears. The port will remain as long as there is a valid
2942 * binding (e.g. until the user changes the binding type or unloads the
2943 * scsi host with the binding).
2944 *
2945 * If the remote port returns within the dev_loss_tmo value (and matches
2946 * according to the target id binding type), the port structure will be
2947 * reused. If it is no longer a SCSI target, the target will be torn
2948 * down. If it continues to be a SCSI target, then the target will be
2949 * unblocked (allowing i/o to be resumed), and a scan will be activated
2950 * to ensure that all luns are detected.
2951 *
2952 * Called from normal process context only - cannot be called from interrupt.
2953 *
2954 * Notes:
2955 * This routine assumes no locks are held on entry.
2956 */
2957void
2958fc_remote_port_delete(struct fc_rport *rport)
2959{
2960 struct Scsi_Host *shost = rport_to_shost(rport);
2961 unsigned long timeout = rport->dev_loss_tmo;
2962 unsigned long flags;
2963
2964 /*
2965 * No need to flush the fc_host work_q's, as all adds are synchronous.
2966 *
2967 * We do need to reclaim the rport scan work element, so eventually
2968 * (in fc_rport_final_delete()) we'll flush the scsi host work_q if
2969 * there's still a scan pending.
2970 */
2971
2972 spin_lock_irqsave(shost->host_lock, flags);
2973
2974 if (rport->port_state != FC_PORTSTATE_ONLINE) {
2975 spin_unlock_irqrestore(shost->host_lock, flags);
2976 return;
2977 }
2978
2979 /*
2980 * In the past, we if this was not an FCP-Target, we would
2981 * unconditionally just jump to deleting the rport.
2982 * However, rports can be used as node containers by the LLDD,
2983 * and its not appropriate to just terminate the rport at the
2984 * first sign of a loss in connectivity. The LLDD may want to
2985 * send ELS traffic to re-validate the login. If the rport is
2986 * immediately deleted, it makes it inappropriate for a node
2987 * container.
2988 * So... we now unconditionally wait dev_loss_tmo before
2989 * destroying an rport.
2990 */
2991
2992 rport->port_state = FC_PORTSTATE_BLOCKED;
2993
2994 rport->flags |= FC_RPORT_DEVLOSS_PENDING;
2995
2996 spin_unlock_irqrestore(shost->host_lock, flags);
2997
2998 if (rport->roles & FC_PORT_ROLE_FCP_INITIATOR &&
2999 shost->active_mode & MODE_TARGET)
3000 fc_tgt_it_nexus_destroy(shost, (unsigned long)rport);
3001
3002 scsi_target_block(&rport->dev);
3003
3004 /* see if we need to kill io faster than waiting for device loss */
3005 if ((rport->fast_io_fail_tmo != -1) &&
3006 (rport->fast_io_fail_tmo < timeout))
3007 fc_queue_devloss_work(shost, &rport->fail_io_work,
3008 rport->fast_io_fail_tmo * HZ);
3009
3010 /* cap the length the devices can be blocked until they are deleted */
3011 fc_queue_devloss_work(shost, &rport->dev_loss_work, timeout * HZ);
3012}
3013EXPORT_SYMBOL(fc_remote_port_delete);
3014
3015/**
3016 * fc_remote_port_rolechg - notifies the fc transport that the roles on a remote may have changed.
3017 * @rport: The remote port that changed.
3018 * @roles: New roles for this port.
3019 *
3020 * Description: The LLDD calls this routine to notify the transport that the
3021 * roles on a remote port may have changed. The largest effect of this is
3022 * if a port now becomes a FCP Target, it must be allocated a
3023 * scsi target id. If the port is no longer a FCP target, any
3024 * scsi target id value assigned to it will persist in case the
3025 * role changes back to include FCP Target. No changes in the scsi
3026 * midlayer will be invoked if the role changes (in the expectation
3027 * that the role will be resumed. If it doesn't normal error processing
3028 * will take place).
3029 *
3030 * Should not be called from interrupt context.
3031 *
3032 * Notes:
3033 * This routine assumes no locks are held on entry.
3034 */
3035void
3036fc_remote_port_rolechg(struct fc_rport *rport, u32 roles)
3037{
3038 struct Scsi_Host *shost = rport_to_shost(rport);
3039 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
3040 unsigned long flags;
3041 int create = 0;
3042 int ret;
3043
3044 spin_lock_irqsave(shost->host_lock, flags);
3045 if (roles & FC_PORT_ROLE_FCP_TARGET) {
3046 if (rport->scsi_target_id == -1) {
3047 rport->scsi_target_id = fc_host->next_target_id++;
3048 create = 1;
3049 } else if (!(rport->roles & FC_PORT_ROLE_FCP_TARGET))
3050 create = 1;
3051 } else if (shost->active_mode & MODE_TARGET) {
3052 ret = fc_tgt_it_nexus_create(shost, (unsigned long)rport,
3053 (char *)&rport->node_name);
3054 if (ret)
3055 printk(KERN_ERR "FC Remore Port tgt nexus failed %d\n",
3056 ret);
3057 }
3058
3059 rport->roles = roles;
3060
3061 spin_unlock_irqrestore(shost->host_lock, flags);
3062
3063 if (create) {
3064 /*
3065 * There may have been a delete timer running on the
3066 * port. Ensure that it is cancelled as we now know
3067 * the port is an FCP Target.
3068 * Note: we know the rport is exists and in an online
3069 * state as the LLDD would not have had an rport
3070 * reference to pass us.
3071 *
3072 * Take no action on the del_timer failure as the state
3073 * machine state change will validate the
3074 * transaction.
3075 */
3076 if (!cancel_delayed_work(&rport->fail_io_work))
3077 fc_flush_devloss(shost);
3078 if (!cancel_delayed_work(&rport->dev_loss_work))
3079 fc_flush_devloss(shost);
3080
3081 spin_lock_irqsave(shost->host_lock, flags);
3082 rport->flags &= ~(FC_RPORT_FAST_FAIL_TIMEDOUT |
3083 FC_RPORT_DEVLOSS_PENDING |
3084 FC_RPORT_DEVLOSS_CALLBK_DONE);
3085 spin_unlock_irqrestore(shost->host_lock, flags);
3086
3087 /* ensure any stgt delete functions are done */
3088 fc_flush_work(shost);
3089
3090 scsi_target_unblock(&rport->dev);
3091 /* initiate a scan of the target */
3092 spin_lock_irqsave(shost->host_lock, flags);
3093 rport->flags |= FC_RPORT_SCAN_PENDING;
3094 scsi_queue_work(shost, &rport->scan_work);
3095 spin_unlock_irqrestore(shost->host_lock, flags);
3096 }
3097}
3098EXPORT_SYMBOL(fc_remote_port_rolechg);
3099
3100/**
3101 * fc_timeout_deleted_rport - Timeout handler for a deleted remote port.
3102 * @work: rport target that failed to reappear in the allotted time.
3103 *
3104 * Description: An attempt to delete a remote port blocks, and if it fails
3105 * to return in the allotted time this gets called.
3106 */
3107static void
3108fc_timeout_deleted_rport(struct work_struct *work)
3109{
3110 struct fc_rport *rport =
3111 container_of(work, struct fc_rport, dev_loss_work.work);
3112 struct Scsi_Host *shost = rport_to_shost(rport);
3113 struct fc_internal *i = to_fc_internal(shost->transportt);
3114 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
3115 unsigned long flags;
3116 int do_callback = 0;
3117
3118 spin_lock_irqsave(shost->host_lock, flags);
3119
3120 rport->flags &= ~FC_RPORT_DEVLOSS_PENDING;
3121
3122 /*
3123 * If the port is ONLINE, then it came back. If it was a SCSI
3124 * target, validate it still is. If not, tear down the
3125 * scsi_target on it.
3126 */
3127 if ((rport->port_state == FC_PORTSTATE_ONLINE) &&
3128 (rport->scsi_target_id != -1) &&
3129 !(rport->roles & FC_PORT_ROLE_FCP_TARGET)) {
3130 dev_printk(KERN_ERR, &rport->dev,
3131 "blocked FC remote port time out: no longer"
3132 " a FCP target, removing starget\n");
3133 spin_unlock_irqrestore(shost->host_lock, flags);
3134 scsi_target_unblock(&rport->dev);
3135 fc_queue_work(shost, &rport->stgt_delete_work);
3136 return;
3137 }
3138
3139 /* NOOP state - we're flushing workq's */
3140 if (rport->port_state != FC_PORTSTATE_BLOCKED) {
3141 spin_unlock_irqrestore(shost->host_lock, flags);
3142 dev_printk(KERN_ERR, &rport->dev,
3143 "blocked FC remote port time out: leaving"
3144 " rport%s alone\n",
3145 (rport->scsi_target_id != -1) ? " and starget" : "");
3146 return;
3147 }
3148
3149 if ((fc_host->tgtid_bind_type == FC_TGTID_BIND_NONE) ||
3150 (rport->scsi_target_id == -1)) {
3151 list_del(&rport->peers);
3152 rport->port_state = FC_PORTSTATE_DELETED;
3153 dev_printk(KERN_ERR, &rport->dev,
3154 "blocked FC remote port time out: removing"
3155 " rport%s\n",
3156 (rport->scsi_target_id != -1) ? " and starget" : "");
3157 fc_queue_work(shost, &rport->rport_delete_work);
3158 spin_unlock_irqrestore(shost->host_lock, flags);
3159 return;
3160 }
3161
3162 dev_printk(KERN_ERR, &rport->dev,
3163 "blocked FC remote port time out: removing target and "
3164 "saving binding\n");
3165
3166 list_move_tail(&rport->peers, &fc_host->rport_bindings);
3167
3168 /*
3169 * Note: We do not remove or clear the hostdata area. This allows
3170 * host-specific target data to persist along with the
3171 * scsi_target_id. It's up to the host to manage it's hostdata area.
3172 */
3173
3174 /*
3175 * Reinitialize port attributes that may change if the port comes back.
3176 */
3177 rport->maxframe_size = -1;
3178 rport->supported_classes = FC_COS_UNSPECIFIED;
3179 rport->roles = FC_PORT_ROLE_UNKNOWN;
3180 rport->port_state = FC_PORTSTATE_NOTPRESENT;
3181 rport->flags &= ~FC_RPORT_FAST_FAIL_TIMEDOUT;
3182
3183 /*
3184 * Pre-emptively kill I/O rather than waiting for the work queue
3185 * item to teardown the starget. (FCOE libFC folks prefer this
3186 * and to have the rport_port_id still set when it's done).
3187 */
3188 spin_unlock_irqrestore(shost->host_lock, flags);
3189 fc_terminate_rport_io(rport);
3190
3191 spin_lock_irqsave(shost->host_lock, flags);
3192
3193 if (rport->port_state == FC_PORTSTATE_NOTPRESENT) { /* still missing */
3194
3195 /* remove the identifiers that aren't used in the consisting binding */
3196 switch (fc_host->tgtid_bind_type) {
3197 case FC_TGTID_BIND_BY_WWPN:
3198 rport->node_name = -1;
3199 rport->port_id = -1;
3200 break;
3201 case FC_TGTID_BIND_BY_WWNN:
3202 rport->port_name = -1;
3203 rport->port_id = -1;
3204 break;
3205 case FC_TGTID_BIND_BY_ID:
3206 rport->node_name = -1;
3207 rport->port_name = -1;
3208 break;
3209 case FC_TGTID_BIND_NONE: /* to keep compiler happy */
3210 break;
3211 }
3212
3213 /*
3214 * As this only occurs if the remote port (scsi target)
3215 * went away and didn't come back - we'll remove
3216 * all attached scsi devices.
3217 */
3218 rport->flags |= FC_RPORT_DEVLOSS_CALLBK_DONE;
3219 fc_queue_work(shost, &rport->stgt_delete_work);
3220
3221 do_callback = 1;
3222 }
3223
3224 spin_unlock_irqrestore(shost->host_lock, flags);
3225
3226 /*
3227 * Notify the driver that the rport is now dead. The LLDD will
3228 * also guarantee that any communication to the rport is terminated
3229 *
3230 * Note: we set the CALLBK_DONE flag above to correspond
3231 */
3232 if (do_callback && i->f->dev_loss_tmo_callbk)
3233 i->f->dev_loss_tmo_callbk(rport);
3234}
3235
3236
3237/**
3238 * fc_timeout_fail_rport_io - Timeout handler for a fast io failing on a disconnected SCSI target.
3239 * @work: rport to terminate io on.
3240 *
3241 * Notes: Only requests the failure of the io, not that all are flushed
3242 * prior to returning.
3243 */
3244static void
3245fc_timeout_fail_rport_io(struct work_struct *work)
3246{
3247 struct fc_rport *rport =
3248 container_of(work, struct fc_rport, fail_io_work.work);
3249
3250 if (rport->port_state != FC_PORTSTATE_BLOCKED)
3251 return;
3252
3253 rport->flags |= FC_RPORT_FAST_FAIL_TIMEDOUT;
3254 fc_terminate_rport_io(rport);
3255}
3256
3257/**
3258 * fc_scsi_scan_rport - called to perform a scsi scan on a remote port.
3259 * @work: remote port to be scanned.
3260 */
3261static void
3262fc_scsi_scan_rport(struct work_struct *work)
3263{
3264 struct fc_rport *rport =
3265 container_of(work, struct fc_rport, scan_work);
3266 struct Scsi_Host *shost = rport_to_shost(rport);
3267 struct fc_internal *i = to_fc_internal(shost->transportt);
3268 unsigned long flags;
3269
3270 if ((rport->port_state == FC_PORTSTATE_ONLINE) &&
3271 (rport->roles & FC_PORT_ROLE_FCP_TARGET) &&
3272 !(i->f->disable_target_scan)) {
3273 scsi_scan_target(&rport->dev, rport->channel,
3274 rport->scsi_target_id, SCAN_WILD_CARD, 1);
3275 }
3276
3277 spin_lock_irqsave(shost->host_lock, flags);
3278 rport->flags &= ~FC_RPORT_SCAN_PENDING;
3279 spin_unlock_irqrestore(shost->host_lock, flags);
3280}
3281
3282/**
3283 * fc_block_scsi_eh - Block SCSI eh thread for blocked fc_rport
3284 * @cmnd: SCSI command that scsi_eh is trying to recover
3285 *
3286 * This routine can be called from a FC LLD scsi_eh callback. It
3287 * blocks the scsi_eh thread until the fc_rport leaves the
3288 * FC_PORTSTATE_BLOCKED, or the fast_io_fail_tmo fires. This is
3289 * necessary to avoid the scsi_eh failing recovery actions for blocked
3290 * rports which would lead to offlined SCSI devices.
3291 *
3292 * Returns: 0 if the fc_rport left the state FC_PORTSTATE_BLOCKED.
3293 * FAST_IO_FAIL if the fast_io_fail_tmo fired, this should be
3294 * passed back to scsi_eh.
3295 */
3296int fc_block_scsi_eh(struct scsi_cmnd *cmnd)
3297{
3298 struct Scsi_Host *shost = cmnd->device->host;
3299 struct fc_rport *rport = starget_to_rport(scsi_target(cmnd->device));
3300 unsigned long flags;
3301
3302 spin_lock_irqsave(shost->host_lock, flags);
3303 while (rport->port_state == FC_PORTSTATE_BLOCKED &&
3304 !(rport->flags & FC_RPORT_FAST_FAIL_TIMEDOUT)) {
3305 spin_unlock_irqrestore(shost->host_lock, flags);
3306 msleep(1000);
3307 spin_lock_irqsave(shost->host_lock, flags);
3308 }
3309 spin_unlock_irqrestore(shost->host_lock, flags);
3310
3311 if (rport->flags & FC_RPORT_FAST_FAIL_TIMEDOUT)
3312 return FAST_IO_FAIL;
3313
3314 return 0;
3315}
3316EXPORT_SYMBOL(fc_block_scsi_eh);
3317
3318/**
3319 * fc_vport_setup - allocates and creates a FC virtual port.
3320 * @shost: scsi host the virtual port is connected to.
3321 * @channel: Channel on shost port connected to.
3322 * @pdev: parent device for vport
3323 * @ids: The world wide names, FC4 port roles, etc for
3324 * the virtual port.
3325 * @ret_vport: The pointer to the created vport.
3326 *
3327 * Allocates and creates the vport structure, calls the parent host
3328 * to instantiate the vport, the completes w/ class and sysfs creation.
3329 *
3330 * Notes:
3331 * This routine assumes no locks are held on entry.
3332 */
3333static int
3334fc_vport_setup(struct Scsi_Host *shost, int channel, struct device *pdev,
3335 struct fc_vport_identifiers *ids, struct fc_vport **ret_vport)
3336{
3337 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
3338 struct fc_internal *fci = to_fc_internal(shost->transportt);
3339 struct fc_vport *vport;
3340 struct device *dev;
3341 unsigned long flags;
3342 size_t size;
3343 int error;
3344
3345 *ret_vport = NULL;
3346
3347 if ( ! fci->f->vport_create)
3348 return -ENOENT;
3349
3350 size = (sizeof(struct fc_vport) + fci->f->dd_fcvport_size);
3351 vport = kzalloc(size, GFP_KERNEL);
3352 if (unlikely(!vport)) {
3353 printk(KERN_ERR "%s: allocation failure\n", __func__);
3354 return -ENOMEM;
3355 }
3356
3357 vport->vport_state = FC_VPORT_UNKNOWN;
3358 vport->vport_last_state = FC_VPORT_UNKNOWN;
3359 vport->node_name = ids->node_name;
3360 vport->port_name = ids->port_name;
3361 vport->roles = ids->roles;
3362 vport->vport_type = ids->vport_type;
3363 if (fci->f->dd_fcvport_size)
3364 vport->dd_data = &vport[1];
3365 vport->shost = shost;
3366 vport->channel = channel;
3367 vport->flags = FC_VPORT_CREATING;
3368 INIT_WORK(&vport->vport_delete_work, fc_vport_sched_delete);
3369
3370 spin_lock_irqsave(shost->host_lock, flags);
3371
3372 if (fc_host->npiv_vports_inuse >= fc_host->max_npiv_vports) {
3373 spin_unlock_irqrestore(shost->host_lock, flags);
3374 kfree(vport);
3375 return -ENOSPC;
3376 }
3377 fc_host->npiv_vports_inuse++;
3378 vport->number = fc_host->next_vport_number++;
3379 list_add_tail(&vport->peers, &fc_host->vports);
3380 get_device(&shost->shost_gendev); /* for fc_host->vport list */
3381
3382 spin_unlock_irqrestore(shost->host_lock, flags);
3383
3384 dev = &vport->dev;
3385 device_initialize(dev); /* takes self reference */
3386 dev->parent = get_device(pdev); /* takes parent reference */
3387 dev->release = fc_vport_dev_release;
3388 dev_set_name(dev, "vport-%d:%d-%d",
3389 shost->host_no, channel, vport->number);
3390 transport_setup_device(dev);
3391
3392 error = device_add(dev);
3393 if (error) {
3394 printk(KERN_ERR "FC Virtual Port device_add failed\n");
3395 goto delete_vport;
3396 }
3397 transport_add_device(dev);
3398 transport_configure_device(dev);
3399
3400 error = fci->f->vport_create(vport, ids->disable);
3401 if (error) {
3402 printk(KERN_ERR "FC Virtual Port LLDD Create failed\n");
3403 goto delete_vport_all;
3404 }
3405
3406 /*
3407 * if the parent isn't the physical adapter's Scsi_Host, ensure
3408 * the Scsi_Host at least contains ia symlink to the vport.
3409 */
3410 if (pdev != &shost->shost_gendev) {
3411 error = sysfs_create_link(&shost->shost_gendev.kobj,
3412 &dev->kobj, dev_name(dev));
3413 if (error)
3414 printk(KERN_ERR
3415 "%s: Cannot create vport symlinks for "
3416 "%s, err=%d\n",
3417 __func__, dev_name(dev), error);
3418 }
3419 spin_lock_irqsave(shost->host_lock, flags);
3420 vport->flags &= ~FC_VPORT_CREATING;
3421 spin_unlock_irqrestore(shost->host_lock, flags);
3422
3423 dev_printk(KERN_NOTICE, pdev,
3424 "%s created via shost%d channel %d\n", dev_name(dev),
3425 shost->host_no, channel);
3426
3427 *ret_vport = vport;
3428
3429 return 0;
3430
3431delete_vport_all:
3432 transport_remove_device(dev);
3433 device_del(dev);
3434delete_vport:
3435 transport_destroy_device(dev);
3436 spin_lock_irqsave(shost->host_lock, flags);
3437 list_del(&vport->peers);
3438 put_device(&shost->shost_gendev); /* for fc_host->vport list */
3439 fc_host->npiv_vports_inuse--;
3440 spin_unlock_irqrestore(shost->host_lock, flags);
3441 put_device(dev->parent);
3442 kfree(vport);
3443
3444 return error;
3445}
3446
3447/**
3448 * fc_vport_create - Admin App or LLDD requests creation of a vport
3449 * @shost: scsi host the virtual port is connected to.
3450 * @channel: channel on shost port connected to.
3451 * @ids: The world wide names, FC4 port roles, etc for
3452 * the virtual port.
3453 *
3454 * Notes:
3455 * This routine assumes no locks are held on entry.
3456 */
3457struct fc_vport *
3458fc_vport_create(struct Scsi_Host *shost, int channel,
3459 struct fc_vport_identifiers *ids)
3460{
3461 int stat;
3462 struct fc_vport *vport;
3463
3464 stat = fc_vport_setup(shost, channel, &shost->shost_gendev,
3465 ids, &vport);
3466 return stat ? NULL : vport;
3467}
3468EXPORT_SYMBOL(fc_vport_create);
3469
3470/**
3471 * fc_vport_terminate - Admin App or LLDD requests termination of a vport
3472 * @vport: fc_vport to be terminated
3473 *
3474 * Calls the LLDD vport_delete() function, then deallocates and removes
3475 * the vport from the shost and object tree.
3476 *
3477 * Notes:
3478 * This routine assumes no locks are held on entry.
3479 */
3480int
3481fc_vport_terminate(struct fc_vport *vport)
3482{
3483 struct Scsi_Host *shost = vport_to_shost(vport);
3484 struct fc_host_attrs *fc_host = shost_to_fc_host(shost);
3485 struct fc_internal *i = to_fc_internal(shost->transportt);
3486 struct device *dev = &vport->dev;
3487 unsigned long flags;
3488 int stat;
3489
3490 if (i->f->vport_delete)
3491 stat = i->f->vport_delete(vport);
3492 else
3493 stat = -ENOENT;
3494
3495 spin_lock_irqsave(shost->host_lock, flags);
3496 vport->flags &= ~FC_VPORT_DELETING;
3497 if (!stat) {
3498 vport->flags |= FC_VPORT_DELETED;
3499 list_del(&vport->peers);
3500 fc_host->npiv_vports_inuse--;
3501 put_device(&shost->shost_gendev); /* for fc_host->vport list */
3502 }
3503 spin_unlock_irqrestore(shost->host_lock, flags);
3504
3505 if (stat)
3506 return stat;
3507
3508 if (dev->parent != &shost->shost_gendev)
3509 sysfs_remove_link(&shost->shost_gendev.kobj, dev_name(dev));
3510 transport_remove_device(dev);
3511 device_del(dev);
3512 transport_destroy_device(dev);
3513
3514 /*
3515 * Removing our self-reference should mean our
3516 * release function gets called, which will drop the remaining
3517 * parent reference and free the data structure.
3518 */
3519 put_device(dev); /* for self-reference */
3520
3521 return 0; /* SUCCESS */
3522}
3523EXPORT_SYMBOL(fc_vport_terminate);
3524
3525/**
3526 * fc_vport_sched_delete - workq-based delete request for a vport
3527 * @work: vport to be deleted.
3528 */
3529static void
3530fc_vport_sched_delete(struct work_struct *work)
3531{
3532 struct fc_vport *vport =
3533 container_of(work, struct fc_vport, vport_delete_work);
3534 int stat;
3535
3536 stat = fc_vport_terminate(vport);
3537 if (stat)
3538 dev_printk(KERN_ERR, vport->dev.parent,
3539 "%s: %s could not be deleted created via "
3540 "shost%d channel %d - error %d\n", __func__,
3541 dev_name(&vport->dev), vport->shost->host_no,
3542 vport->channel, stat);
3543}
3544
3545
3546/*
3547 * BSG support
3548 */
3549
3550
3551/**
3552 * fc_destroy_bsgjob - routine to teardown/delete a fc bsg job
3553 * @job: fc_bsg_job that is to be torn down
3554 */
3555static void
3556fc_destroy_bsgjob(struct fc_bsg_job *job)
3557{
3558 unsigned long flags;
3559
3560 spin_lock_irqsave(&job->job_lock, flags);
3561 if (job->ref_cnt) {
3562 spin_unlock_irqrestore(&job->job_lock, flags);
3563 return;
3564 }
3565 spin_unlock_irqrestore(&job->job_lock, flags);
3566
3567 put_device(job->dev); /* release reference for the request */
3568
3569 kfree(job->request_payload.sg_list);
3570 kfree(job->reply_payload.sg_list);
3571 kfree(job);
3572}
3573
3574/**
3575 * fc_bsg_jobdone - completion routine for bsg requests that the LLD has
3576 * completed
3577 * @job: fc_bsg_job that is complete
3578 */
3579static void
3580fc_bsg_jobdone(struct fc_bsg_job *job)
3581{
3582 struct request *req = job->req;
3583 struct request *rsp = req->next_rq;
3584 int err;
3585
3586 err = job->req->errors = job->reply->result;
3587
3588 if (err < 0)
3589 /* we're only returning the result field in the reply */
3590 job->req->sense_len = sizeof(uint32_t);
3591 else
3592 job->req->sense_len = job->reply_len;
3593
3594 /* we assume all request payload was transferred, residual == 0 */
3595 req->resid_len = 0;
3596
3597 if (rsp) {
3598 WARN_ON(job->reply->reply_payload_rcv_len > rsp->resid_len);
3599
3600 /* set reply (bidi) residual */
3601 rsp->resid_len -= min(job->reply->reply_payload_rcv_len,
3602 rsp->resid_len);
3603 }
3604 blk_complete_request(req);
3605}
3606
3607/**
3608 * fc_bsg_softirq_done - softirq done routine for destroying the bsg requests
3609 * @rq: BSG request that holds the job to be destroyed
3610 */
3611static void fc_bsg_softirq_done(struct request *rq)
3612{
3613 struct fc_bsg_job *job = rq->special;
3614 unsigned long flags;
3615
3616 spin_lock_irqsave(&job->job_lock, flags);
3617 job->state_flags |= FC_RQST_STATE_DONE;
3618 job->ref_cnt--;
3619 spin_unlock_irqrestore(&job->job_lock, flags);
3620
3621 blk_end_request_all(rq, rq->errors);
3622 fc_destroy_bsgjob(job);
3623}
3624
3625/**
3626 * fc_bsg_job_timeout - handler for when a bsg request timesout
3627 * @req: request that timed out
3628 */
3629static enum blk_eh_timer_return
3630fc_bsg_job_timeout(struct request *req)
3631{
3632 struct fc_bsg_job *job = (void *) req->special;
3633 struct Scsi_Host *shost = job->shost;
3634 struct fc_internal *i = to_fc_internal(shost->transportt);
3635 unsigned long flags;
3636 int err = 0, done = 0;
3637
3638 if (job->rport && job->rport->port_state == FC_PORTSTATE_BLOCKED)
3639 return BLK_EH_RESET_TIMER;
3640
3641 spin_lock_irqsave(&job->job_lock, flags);
3642 if (job->state_flags & FC_RQST_STATE_DONE)
3643 done = 1;
3644 else
3645 job->ref_cnt++;
3646 spin_unlock_irqrestore(&job->job_lock, flags);
3647
3648 if (!done && i->f->bsg_timeout) {
3649 /* call LLDD to abort the i/o as it has timed out */
3650 err = i->f->bsg_timeout(job);
3651 if (err == -EAGAIN) {
3652 job->ref_cnt--;
3653 return BLK_EH_RESET_TIMER;
3654 } else if (err)
3655 printk(KERN_ERR "ERROR: FC BSG request timeout - LLD "
3656 "abort failed with status %d\n", err);
3657 }
3658
3659 /* the blk_end_sync_io() doesn't check the error */
3660 if (done)
3661 return BLK_EH_NOT_HANDLED;
3662 else
3663 return BLK_EH_HANDLED;
3664}
3665
3666static int
3667fc_bsg_map_buffer(struct fc_bsg_buffer *buf, struct request *req)
3668{
3669 size_t sz = (sizeof(struct scatterlist) * req->nr_phys_segments);
3670
3671 BUG_ON(!req->nr_phys_segments);
3672
3673 buf->sg_list = kzalloc(sz, GFP_KERNEL);
3674 if (!buf->sg_list)
3675 return -ENOMEM;
3676 sg_init_table(buf->sg_list, req->nr_phys_segments);
3677 buf->sg_cnt = blk_rq_map_sg(req->q, req, buf->sg_list);
3678 buf->payload_len = blk_rq_bytes(req);
3679 return 0;
3680}
3681
3682
3683/**
3684 * fc_req_to_bsgjob - Allocate/create the fc_bsg_job structure for the
3685 * bsg request
3686 * @shost: SCSI Host corresponding to the bsg object
3687 * @rport: (optional) FC Remote Port corresponding to the bsg object
3688 * @req: BSG request that needs a job structure
3689 */
3690static int
3691fc_req_to_bsgjob(struct Scsi_Host *shost, struct fc_rport *rport,
3692 struct request *req)
3693{
3694 struct fc_internal *i = to_fc_internal(shost->transportt);
3695 struct request *rsp = req->next_rq;
3696 struct fc_bsg_job *job;
3697 int ret;
3698
3699 BUG_ON(req->special);
3700
3701 job = kzalloc(sizeof(struct fc_bsg_job) + i->f->dd_bsg_size,
3702 GFP_KERNEL);
3703 if (!job)
3704 return -ENOMEM;
3705
3706 /*
3707 * Note: this is a bit silly.
3708 * The request gets formatted as a SGIO v4 ioctl request, which
3709 * then gets reformatted as a blk request, which then gets
3710 * reformatted as a fc bsg request. And on completion, we have
3711 * to wrap return results such that SGIO v4 thinks it was a scsi
3712 * status. I hope this was all worth it.
3713 */
3714
3715 req->special = job;
3716 job->shost = shost;
3717 job->rport = rport;
3718 job->req = req;
3719 if (i->f->dd_bsg_size)
3720 job->dd_data = (void *)&job[1];
3721 spin_lock_init(&job->job_lock);
3722 job->request = (struct fc_bsg_request *)req->cmd;
3723 job->request_len = req->cmd_len;
3724 job->reply = req->sense;
3725 job->reply_len = SCSI_SENSE_BUFFERSIZE; /* Size of sense buffer
3726 * allocated */
3727 if (req->bio) {
3728 ret = fc_bsg_map_buffer(&job->request_payload, req);
3729 if (ret)
3730 goto failjob_rls_job;
3731 }
3732 if (rsp && rsp->bio) {
3733 ret = fc_bsg_map_buffer(&job->reply_payload, rsp);
3734 if (ret)
3735 goto failjob_rls_rqst_payload;
3736 }
3737 job->job_done = fc_bsg_jobdone;
3738 if (rport)
3739 job->dev = &rport->dev;
3740 else
3741 job->dev = &shost->shost_gendev;
3742 get_device(job->dev); /* take a reference for the request */
3743
3744 job->ref_cnt = 1;
3745
3746 return 0;
3747
3748
3749failjob_rls_rqst_payload:
3750 kfree(job->request_payload.sg_list);
3751failjob_rls_job:
3752 kfree(job);
3753 return -ENOMEM;
3754}
3755
3756
3757enum fc_dispatch_result {
3758 FC_DISPATCH_BREAK, /* on return, q is locked, break from q loop */
3759 FC_DISPATCH_LOCKED, /* on return, q is locked, continue on */
3760 FC_DISPATCH_UNLOCKED, /* on return, q is unlocked, continue on */
3761};
3762
3763
3764/**
3765 * fc_bsg_host_dispatch - process fc host bsg requests and dispatch to LLDD
3766 * @q: fc host request queue
3767 * @shost: scsi host rport attached to
3768 * @job: bsg job to be processed
3769 */
3770static enum fc_dispatch_result
3771fc_bsg_host_dispatch(struct request_queue *q, struct Scsi_Host *shost,
3772 struct fc_bsg_job *job)
3773{
3774 struct fc_internal *i = to_fc_internal(shost->transportt);
3775 int cmdlen = sizeof(uint32_t); /* start with length of msgcode */
3776 int ret;
3777
3778 /* Validate the host command */
3779 switch (job->request->msgcode) {
3780 case FC_BSG_HST_ADD_RPORT:
3781 cmdlen += sizeof(struct fc_bsg_host_add_rport);
3782 break;
3783
3784 case FC_BSG_HST_DEL_RPORT:
3785 cmdlen += sizeof(struct fc_bsg_host_del_rport);
3786 break;
3787
3788 case FC_BSG_HST_ELS_NOLOGIN:
3789 cmdlen += sizeof(struct fc_bsg_host_els);
3790 /* there better be a xmt and rcv payloads */
3791 if ((!job->request_payload.payload_len) ||
3792 (!job->reply_payload.payload_len)) {
3793 ret = -EINVAL;
3794 goto fail_host_msg;
3795 }
3796 break;
3797
3798 case FC_BSG_HST_CT:
3799 cmdlen += sizeof(struct fc_bsg_host_ct);
3800 /* there better be xmt and rcv payloads */
3801 if ((!job->request_payload.payload_len) ||
3802 (!job->reply_payload.payload_len)) {
3803 ret = -EINVAL;
3804 goto fail_host_msg;
3805 }
3806 break;
3807
3808 case FC_BSG_HST_VENDOR:
3809 cmdlen += sizeof(struct fc_bsg_host_vendor);
3810 if ((shost->hostt->vendor_id == 0L) ||
3811 (job->request->rqst_data.h_vendor.vendor_id !=
3812 shost->hostt->vendor_id)) {
3813 ret = -ESRCH;
3814 goto fail_host_msg;
3815 }
3816 break;
3817
3818 default:
3819 ret = -EBADR;
3820 goto fail_host_msg;
3821 }
3822
3823 /* check if we really have all the request data needed */
3824 if (job->request_len < cmdlen) {
3825 ret = -ENOMSG;
3826 goto fail_host_msg;
3827 }
3828
3829 ret = i->f->bsg_request(job);
3830 if (!ret)
3831 return FC_DISPATCH_UNLOCKED;
3832
3833fail_host_msg:
3834 /* return the errno failure code as the only status */
3835 BUG_ON(job->reply_len < sizeof(uint32_t));
3836 job->reply->reply_payload_rcv_len = 0;
3837 job->reply->result = ret;
3838 job->reply_len = sizeof(uint32_t);
3839 fc_bsg_jobdone(job);
3840 return FC_DISPATCH_UNLOCKED;
3841}
3842
3843
3844/*
3845 * fc_bsg_goose_queue - restart rport queue in case it was stopped
3846 * @rport: rport to be restarted
3847 */
3848static void
3849fc_bsg_goose_queue(struct fc_rport *rport)
3850{
3851 if (!rport->rqst_q)
3852 return;
3853
3854 /*
3855 * This get/put dance makes no sense
3856 */
3857 get_device(&rport->dev);
3858 blk_run_queue_async(rport->rqst_q);
3859 put_device(&rport->dev);
3860}
3861
3862/**
3863 * fc_bsg_rport_dispatch - process rport bsg requests and dispatch to LLDD
3864 * @q: rport request queue
3865 * @shost: scsi host rport attached to
3866 * @rport: rport request destined to
3867 * @job: bsg job to be processed
3868 */
3869static enum fc_dispatch_result
3870fc_bsg_rport_dispatch(struct request_queue *q, struct Scsi_Host *shost,
3871 struct fc_rport *rport, struct fc_bsg_job *job)
3872{
3873 struct fc_internal *i = to_fc_internal(shost->transportt);
3874 int cmdlen = sizeof(uint32_t); /* start with length of msgcode */
3875 int ret;
3876
3877 /* Validate the rport command */
3878 switch (job->request->msgcode) {
3879 case FC_BSG_RPT_ELS:
3880 cmdlen += sizeof(struct fc_bsg_rport_els);
3881 goto check_bidi;
3882
3883 case FC_BSG_RPT_CT:
3884 cmdlen += sizeof(struct fc_bsg_rport_ct);
3885check_bidi:
3886 /* there better be xmt and rcv payloads */
3887 if ((!job->request_payload.payload_len) ||
3888 (!job->reply_payload.payload_len)) {
3889 ret = -EINVAL;
3890 goto fail_rport_msg;
3891 }
3892 break;
3893 default:
3894 ret = -EBADR;
3895 goto fail_rport_msg;
3896 }
3897
3898 /* check if we really have all the request data needed */
3899 if (job->request_len < cmdlen) {
3900 ret = -ENOMSG;
3901 goto fail_rport_msg;
3902 }
3903
3904 ret = i->f->bsg_request(job);
3905 if (!ret)
3906 return FC_DISPATCH_UNLOCKED;
3907
3908fail_rport_msg:
3909 /* return the errno failure code as the only status */
3910 BUG_ON(job->reply_len < sizeof(uint32_t));
3911 job->reply->reply_payload_rcv_len = 0;
3912 job->reply->result = ret;
3913 job->reply_len = sizeof(uint32_t);
3914 fc_bsg_jobdone(job);
3915 return FC_DISPATCH_UNLOCKED;
3916}
3917
3918
3919/**
3920 * fc_bsg_request_handler - generic handler for bsg requests
3921 * @q: request queue to manage
3922 * @shost: Scsi_Host related to the bsg object
3923 * @rport: FC remote port related to the bsg object (optional)
3924 * @dev: device structure for bsg object
3925 */
3926static void
3927fc_bsg_request_handler(struct request_queue *q, struct Scsi_Host *shost,
3928 struct fc_rport *rport, struct device *dev)
3929{
3930 struct request *req;
3931 struct fc_bsg_job *job;
3932 enum fc_dispatch_result ret;
3933
3934 if (!get_device(dev))
3935 return;
3936
3937 while (1) {
3938 if (rport && (rport->port_state == FC_PORTSTATE_BLOCKED) &&
3939 !(rport->flags & FC_RPORT_FAST_FAIL_TIMEDOUT))
3940 break;
3941
3942 req = blk_fetch_request(q);
3943 if (!req)
3944 break;
3945
3946 if (rport && (rport->port_state != FC_PORTSTATE_ONLINE)) {
3947 req->errors = -ENXIO;
3948 spin_unlock_irq(q->queue_lock);
3949 blk_end_request_all(req, -ENXIO);
3950 spin_lock_irq(q->queue_lock);
3951 continue;
3952 }
3953
3954 spin_unlock_irq(q->queue_lock);
3955
3956 ret = fc_req_to_bsgjob(shost, rport, req);
3957 if (ret) {
3958 req->errors = ret;
3959 blk_end_request_all(req, ret);
3960 spin_lock_irq(q->queue_lock);
3961 continue;
3962 }
3963
3964 job = req->special;
3965
3966 /* check if we have the msgcode value at least */
3967 if (job->request_len < sizeof(uint32_t)) {
3968 BUG_ON(job->reply_len < sizeof(uint32_t));
3969 job->reply->reply_payload_rcv_len = 0;
3970 job->reply->result = -ENOMSG;
3971 job->reply_len = sizeof(uint32_t);
3972 fc_bsg_jobdone(job);
3973 spin_lock_irq(q->queue_lock);
3974 continue;
3975 }
3976
3977 /* the dispatch routines will unlock the queue_lock */
3978 if (rport)
3979 ret = fc_bsg_rport_dispatch(q, shost, rport, job);
3980 else
3981 ret = fc_bsg_host_dispatch(q, shost, job);
3982
3983 /* did dispatcher hit state that can't process any more */
3984 if (ret == FC_DISPATCH_BREAK)
3985 break;
3986
3987 /* did dispatcher had released the lock */
3988 if (ret == FC_DISPATCH_UNLOCKED)
3989 spin_lock_irq(q->queue_lock);
3990 }
3991
3992 spin_unlock_irq(q->queue_lock);
3993 put_device(dev);
3994 spin_lock_irq(q->queue_lock);
3995}
3996
3997
3998/**
3999 * fc_bsg_host_handler - handler for bsg requests for a fc host
4000 * @q: fc host request queue
4001 */
4002static void
4003fc_bsg_host_handler(struct request_queue *q)
4004{
4005 struct Scsi_Host *shost = q->queuedata;
4006
4007 fc_bsg_request_handler(q, shost, NULL, &shost->shost_gendev);
4008}
4009
4010
4011/**
4012 * fc_bsg_rport_handler - handler for bsg requests for a fc rport
4013 * @q: rport request queue
4014 */
4015static void
4016fc_bsg_rport_handler(struct request_queue *q)
4017{
4018 struct fc_rport *rport = q->queuedata;
4019 struct Scsi_Host *shost = rport_to_shost(rport);
4020
4021 fc_bsg_request_handler(q, shost, rport, &rport->dev);
4022}
4023
4024
4025/**
4026 * fc_bsg_hostadd - Create and add the bsg hooks so we can receive requests
4027 * @shost: shost for fc_host
4028 * @fc_host: fc_host adding the structures to
4029 */
4030static int
4031fc_bsg_hostadd(struct Scsi_Host *shost, struct fc_host_attrs *fc_host)
4032{
4033 struct device *dev = &shost->shost_gendev;
4034 struct fc_internal *i = to_fc_internal(shost->transportt);
4035 struct request_queue *q;
4036 int err;
4037 char bsg_name[20];
4038
4039 fc_host->rqst_q = NULL;
4040
4041 if (!i->f->bsg_request)
4042 return -ENOTSUPP;
4043
4044 snprintf(bsg_name, sizeof(bsg_name),
4045 "fc_host%d", shost->host_no);
4046
4047 q = __scsi_alloc_queue(shost, fc_bsg_host_handler);
4048 if (!q) {
4049 printk(KERN_ERR "fc_host%d: bsg interface failed to "
4050 "initialize - no request queue\n",
4051 shost->host_no);
4052 return -ENOMEM;
4053 }
4054
4055 q->queuedata = shost;
4056 queue_flag_set_unlocked(QUEUE_FLAG_BIDI, q);
4057 blk_queue_softirq_done(q, fc_bsg_softirq_done);
4058 blk_queue_rq_timed_out(q, fc_bsg_job_timeout);
4059 blk_queue_rq_timeout(q, FC_DEFAULT_BSG_TIMEOUT);
4060
4061 err = bsg_register_queue(q, dev, bsg_name, NULL);
4062 if (err) {
4063 printk(KERN_ERR "fc_host%d: bsg interface failed to "
4064 "initialize - register queue\n",
4065 shost->host_no);
4066 blk_cleanup_queue(q);
4067 return err;
4068 }
4069
4070 fc_host->rqst_q = q;
4071 return 0;
4072}
4073
4074
4075/**
4076 * fc_bsg_rportadd - Create and add the bsg hooks so we can receive requests
4077 * @shost: shost that rport is attached to
4078 * @rport: rport that the bsg hooks are being attached to
4079 */
4080static int
4081fc_bsg_rportadd(struct Scsi_Host *shost, struct fc_rport *rport)
4082{
4083 struct device *dev = &rport->dev;
4084 struct fc_internal *i = to_fc_internal(shost->transportt);
4085 struct request_queue *q;
4086 int err;
4087
4088 rport->rqst_q = NULL;
4089
4090 if (!i->f->bsg_request)
4091 return -ENOTSUPP;
4092
4093 q = __scsi_alloc_queue(shost, fc_bsg_rport_handler);
4094 if (!q) {
4095 printk(KERN_ERR "%s: bsg interface failed to "
4096 "initialize - no request queue\n",
4097 dev->kobj.name);
4098 return -ENOMEM;
4099 }
4100
4101 q->queuedata = rport;
4102 queue_flag_set_unlocked(QUEUE_FLAG_BIDI, q);
4103 blk_queue_softirq_done(q, fc_bsg_softirq_done);
4104 blk_queue_rq_timed_out(q, fc_bsg_job_timeout);
4105 blk_queue_rq_timeout(q, BLK_DEFAULT_SG_TIMEOUT);
4106
4107 err = bsg_register_queue(q, dev, NULL, NULL);
4108 if (err) {
4109 printk(KERN_ERR "%s: bsg interface failed to "
4110 "initialize - register queue\n",
4111 dev->kobj.name);
4112 blk_cleanup_queue(q);
4113 return err;
4114 }
4115
4116 rport->rqst_q = q;
4117 return 0;
4118}
4119
4120
4121/**
4122 * fc_bsg_remove - Deletes the bsg hooks on fchosts/rports
4123 * @q: the request_queue that is to be torn down.
4124 *
4125 * Notes:
4126 * Before unregistering the queue empty any requests that are blocked
4127 *
4128 *
4129 */
4130static void
4131fc_bsg_remove(struct request_queue *q)
4132{
4133 struct request *req; /* block request */
4134 int counts; /* totals for request_list count and starved */
4135
4136 if (q) {
4137 /* Stop taking in new requests */
4138 spin_lock_irq(q->queue_lock);
4139 blk_stop_queue(q);
4140
4141 /* drain all requests in the queue */
4142 while (1) {
4143 /* need the lock to fetch a request
4144 * this may fetch the same reqeust as the previous pass
4145 */
4146 req = blk_fetch_request(q);
4147 /* save requests in use and starved */
4148 counts = q->rq.count[0] + q->rq.count[1] +
4149 q->rq.starved[0] + q->rq.starved[1];
4150 spin_unlock_irq(q->queue_lock);
4151 /* any requests still outstanding? */
4152 if (counts == 0)
4153 break;
4154
4155 /* This may be the same req as the previous iteration,
4156 * always send the blk_end_request_all after a prefetch.
4157 * It is not okay to not end the request because the
4158 * prefetch started the request.
4159 */
4160 if (req) {
4161 /* return -ENXIO to indicate that this queue is
4162 * going away
4163 */
4164 req->errors = -ENXIO;
4165 blk_end_request_all(req, -ENXIO);
4166 }
4167
4168 msleep(200); /* allow bsg to possibly finish */
4169 spin_lock_irq(q->queue_lock);
4170 }
4171
4172 bsg_unregister_queue(q);
4173 blk_cleanup_queue(q);
4174 }
4175}
4176
4177
4178/* Original Author: Martin Hicks */
4179MODULE_AUTHOR("James Smart");
4180MODULE_DESCRIPTION("FC Transport Attributes");
4181MODULE_LICENSE("GPL");
4182
4183module_init(fc_transport_init);
4184module_exit(fc_transport_exit);