diff options
Diffstat (limited to 'drivers/scsi/csiostor/csio_attr.c')
| -rw-r--r-- | drivers/scsi/csiostor/csio_attr.c | 796 | 
1 files changed, 796 insertions, 0 deletions
| diff --git a/drivers/scsi/csiostor/csio_attr.c b/drivers/scsi/csiostor/csio_attr.c new file mode 100644 index 000000000000..065a87ace623 --- /dev/null +++ b/drivers/scsi/csiostor/csio_attr.c @@ -0,0 +1,796 @@ +/* + * This file is part of the Chelsio FCoE driver for Linux. + * + * Copyright (c) 2008-2012 Chelsio Communications, Inc. All rights reserved. + * + * This software is available to you under a choice of one of two + * licenses.  You may choose to be licensed under the terms of the GNU + * General Public License (GPL) Version 2, available from the file + * COPYING in the main directory of this source tree, or the + * OpenIB.org BSD license below: + * + *     Redistribution and use in source and binary forms, with or + *     without modification, are permitted provided that the following + *     conditions are met: + * + *      - Redistributions of source code must retain the above + *        copyright notice, this list of conditions and the following + *        disclaimer. + * + *      - Redistributions in binary form must reproduce the above + *        copyright notice, this list of conditions and the following + *        disclaimer in the documentation and/or other materials + *        provided with the distribution. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, + * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF + * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND + * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS + * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN + * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN + * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE + * SOFTWARE. + */ + +#include <linux/kernel.h> +#include <linux/string.h> +#include <linux/delay.h> +#include <linux/module.h> +#include <linux/init.h> +#include <linux/pci.h> +#include <linux/mm.h> +#include <linux/jiffies.h> +#include <scsi/fc/fc_fs.h> + +#include "csio_init.h" + +static void +csio_vport_set_state(struct csio_lnode *ln); + +/* + * csio_reg_rnode - Register a remote port with FC transport. + * @rn: Rnode representing remote port. + * + * Call fc_remote_port_add() to register this remote port with FC transport. + * If remote port is Initiator OR Target OR both, change the role appropriately. + * + */ +void +csio_reg_rnode(struct csio_rnode *rn) +{ +	struct csio_lnode *ln		= csio_rnode_to_lnode(rn); +	struct Scsi_Host *shost		= csio_ln_to_shost(ln); +	struct fc_rport_identifiers ids; +	struct fc_rport  *rport; +	struct csio_service_parms *sp; + +	ids.node_name	= wwn_to_u64(csio_rn_wwnn(rn)); +	ids.port_name	= wwn_to_u64(csio_rn_wwpn(rn)); +	ids.port_id	= rn->nport_id; +	ids.roles	= FC_RPORT_ROLE_UNKNOWN; + +	if (rn->role & CSIO_RNFR_INITIATOR || rn->role & CSIO_RNFR_TARGET) { +		rport = rn->rport; +		CSIO_ASSERT(rport != NULL); +		goto update_role; +	} + +	rn->rport = fc_remote_port_add(shost, 0, &ids); +	if (!rn->rport) { +		csio_ln_err(ln, "Failed to register rport = 0x%x.\n", +					rn->nport_id); +		return; +	} + +	ln->num_reg_rnodes++; +	rport = rn->rport; +	spin_lock_irq(shost->host_lock); +	*((struct csio_rnode **)rport->dd_data) = rn; +	spin_unlock_irq(shost->host_lock); + +	sp = &rn->rn_sparm; +	rport->maxframe_size = ntohs(sp->csp.sp_bb_data); +	if (ntohs(sp->clsp[2].cp_class) & FC_CPC_VALID) +		rport->supported_classes = FC_COS_CLASS3; +	else +		rport->supported_classes = FC_COS_UNSPECIFIED; +update_role: +	if (rn->role & CSIO_RNFR_INITIATOR) +		ids.roles |= FC_RPORT_ROLE_FCP_INITIATOR; +	if (rn->role & CSIO_RNFR_TARGET) +		ids.roles |= FC_RPORT_ROLE_FCP_TARGET; + +	if (ids.roles != FC_RPORT_ROLE_UNKNOWN) +		fc_remote_port_rolechg(rport, ids.roles); + +	rn->scsi_id = rport->scsi_target_id; + +	csio_ln_dbg(ln, "Remote port x%x role 0x%x registered\n", +		rn->nport_id, ids.roles); +} + +/* + * csio_unreg_rnode - Unregister a remote port with FC transport. + * @rn: Rnode representing remote port. + * + * Call fc_remote_port_delete() to unregister this remote port with FC + * transport. + * + */ +void +csio_unreg_rnode(struct csio_rnode *rn) +{ +	struct csio_lnode *ln = csio_rnode_to_lnode(rn); +	struct fc_rport *rport = rn->rport; + +	rn->role &= ~(CSIO_RNFR_INITIATOR | CSIO_RNFR_TARGET); +	fc_remote_port_delete(rport); +	ln->num_reg_rnodes--; + +	csio_ln_dbg(ln, "Remote port x%x un-registered\n", rn->nport_id); +} + +/* + * csio_lnode_async_event - Async events from local port. + * @ln: lnode representing local port. + * + * Async events from local node that FC transport/SCSI ML + * should be made aware of (Eg: RSCN). + */ +void +csio_lnode_async_event(struct csio_lnode *ln, enum csio_ln_fc_evt fc_evt) +{ +	switch (fc_evt) { +	case CSIO_LN_FC_RSCN: +		/* Get payload of rscn from ln */ +		/* For each RSCN entry */ +			/* +			 * fc_host_post_event(shost, +			 *		      fc_get_event_number(), +			 *		      FCH_EVT_RSCN, +			 *		      rscn_entry); +			 */ +		break; +	case CSIO_LN_FC_LINKUP: +		/* send fc_host_post_event */ +		/* set vport state */ +		if (csio_is_npiv_ln(ln)) +			csio_vport_set_state(ln); + +		break; +	case CSIO_LN_FC_LINKDOWN: +		/* send fc_host_post_event */ +		/* set vport state */ +		if (csio_is_npiv_ln(ln)) +			csio_vport_set_state(ln); + +		break; +	case CSIO_LN_FC_ATTRIB_UPDATE: +		csio_fchost_attr_init(ln); +		break; +	default: +		break; +	} +} + +/* + * csio_fchost_attr_init - Initialize FC transport attributes + * @ln: Lnode. + * + */ +void +csio_fchost_attr_init(struct csio_lnode *ln) +{ +	struct Scsi_Host  *shost = csio_ln_to_shost(ln); + +	fc_host_node_name(shost) = wwn_to_u64(csio_ln_wwnn(ln)); +	fc_host_port_name(shost) = wwn_to_u64(csio_ln_wwpn(ln)); + +	fc_host_supported_classes(shost) = FC_COS_CLASS3; +	fc_host_max_npiv_vports(shost) = +			(csio_lnode_to_hw(ln))->fres_info.max_vnps; +	fc_host_supported_speeds(shost) = FC_PORTSPEED_10GBIT | +		FC_PORTSPEED_1GBIT; + +	fc_host_maxframe_size(shost) = ntohs(ln->ln_sparm.csp.sp_bb_data); +	memset(fc_host_supported_fc4s(shost), 0, +		sizeof(fc_host_supported_fc4s(shost))); +	fc_host_supported_fc4s(shost)[7] = 1; + +	memset(fc_host_active_fc4s(shost), 0, +		sizeof(fc_host_active_fc4s(shost))); +	fc_host_active_fc4s(shost)[7] = 1; +} + +/* + * csio_get_host_port_id - sysfs entries for nport_id is + * populated/cached from this function + */ +static void +csio_get_host_port_id(struct Scsi_Host *shost) +{ +	struct csio_lnode *ln	= shost_priv(shost); +	struct csio_hw *hw = csio_lnode_to_hw(ln); + +	spin_lock_irq(&hw->lock); +	fc_host_port_id(shost) = ln->nport_id; +	spin_unlock_irq(&hw->lock); +} + +/* + * csio_get_port_type - Return FC local port type. + * @shost: scsi host. + * + */ +static void +csio_get_host_port_type(struct Scsi_Host *shost) +{ +	struct csio_lnode *ln = shost_priv(shost); +	struct csio_hw *hw = csio_lnode_to_hw(ln); + +	spin_lock_irq(&hw->lock); +	if (csio_is_npiv_ln(ln)) +		fc_host_port_type(shost) = FC_PORTTYPE_NPIV; +	else +		fc_host_port_type(shost) = FC_PORTTYPE_NPORT; +	spin_unlock_irq(&hw->lock); +} + +/* + * csio_get_port_state - Return FC local port state. + * @shost: scsi host. + * + */ +static void +csio_get_host_port_state(struct Scsi_Host *shost) +{ +	struct csio_lnode *ln = shost_priv(shost); +	struct csio_hw *hw = csio_lnode_to_hw(ln); +	char state[16]; + +	spin_lock_irq(&hw->lock); + +	csio_lnode_state_to_str(ln, state); +	if (!strcmp(state, "READY")) +		fc_host_port_state(shost) = FC_PORTSTATE_ONLINE; +	else if (!strcmp(state, "OFFLINE")) +		fc_host_port_state(shost) = FC_PORTSTATE_LINKDOWN; +	else +		fc_host_port_state(shost) = FC_PORTSTATE_UNKNOWN; + +	spin_unlock_irq(&hw->lock); +} + +/* + * csio_get_host_speed - Return link speed to FC transport. + * @shost: scsi host. + * + */ +static void +csio_get_host_speed(struct Scsi_Host *shost) +{ +	struct csio_lnode *ln = shost_priv(shost); +	struct csio_hw *hw = csio_lnode_to_hw(ln); + +	spin_lock_irq(&hw->lock); +	switch (hw->pport[ln->portid].link_speed) { +	case FW_PORT_CAP_SPEED_1G: +		fc_host_speed(shost) = FC_PORTSPEED_1GBIT; +		break; +	case FW_PORT_CAP_SPEED_10G: +		fc_host_speed(shost) = FC_PORTSPEED_10GBIT; +		break; +	default: +		fc_host_speed(shost) = FC_PORTSPEED_UNKNOWN; +		break; +	} +	spin_unlock_irq(&hw->lock); +} + +/* + * csio_get_host_fabric_name - Return fabric name + * @shost: scsi host. + * + */ +static void +csio_get_host_fabric_name(struct Scsi_Host *shost) +{ +	struct csio_lnode *ln = shost_priv(shost); +	struct csio_rnode *rn = NULL; +	struct csio_hw *hw = csio_lnode_to_hw(ln); + +	spin_lock_irq(&hw->lock); +	rn = csio_rnode_lookup_portid(ln, FC_FID_FLOGI); +	if (rn) +		fc_host_fabric_name(shost) = wwn_to_u64(csio_rn_wwnn(rn)); +	else +		fc_host_fabric_name(shost) = 0; +	spin_unlock_irq(&hw->lock); +} + +/* + * csio_get_host_speed - Return FC transport statistics. + * @ln: Lnode. + * + */ +static struct fc_host_statistics * +csio_get_stats(struct Scsi_Host *shost) +{ +	struct csio_lnode *ln = shost_priv(shost); +	struct csio_hw *hw = csio_lnode_to_hw(ln); +	struct fc_host_statistics *fhs = &ln->fch_stats; +	struct fw_fcoe_port_stats fcoe_port_stats; +	uint64_t seconds; + +	memset(&fcoe_port_stats, 0, sizeof(struct fw_fcoe_port_stats)); +	csio_get_phy_port_stats(hw, ln->portid, &fcoe_port_stats); + +	fhs->tx_frames  += (be64_to_cpu(fcoe_port_stats.tx_bcast_frames) + +			    be64_to_cpu(fcoe_port_stats.tx_mcast_frames) + +			    be64_to_cpu(fcoe_port_stats.tx_ucast_frames) + +			    be64_to_cpu(fcoe_port_stats.tx_offload_frames)); +	fhs->tx_words  += (be64_to_cpu(fcoe_port_stats.tx_bcast_bytes) + +			   be64_to_cpu(fcoe_port_stats.tx_mcast_bytes) + +			   be64_to_cpu(fcoe_port_stats.tx_ucast_bytes) + +			   be64_to_cpu(fcoe_port_stats.tx_offload_bytes)) / +							CSIO_WORD_TO_BYTE; +	fhs->rx_frames += (be64_to_cpu(fcoe_port_stats.rx_bcast_frames) + +			   be64_to_cpu(fcoe_port_stats.rx_mcast_frames) + +			   be64_to_cpu(fcoe_port_stats.rx_ucast_frames)); +	fhs->rx_words += (be64_to_cpu(fcoe_port_stats.rx_bcast_bytes) + +			  be64_to_cpu(fcoe_port_stats.rx_mcast_bytes) + +			  be64_to_cpu(fcoe_port_stats.rx_ucast_bytes)) / +							CSIO_WORD_TO_BYTE; +	fhs->error_frames += be64_to_cpu(fcoe_port_stats.rx_err_frames); +	fhs->fcp_input_requests +=  ln->stats.n_input_requests; +	fhs->fcp_output_requests +=  ln->stats.n_output_requests; +	fhs->fcp_control_requests +=  ln->stats.n_control_requests; +	fhs->fcp_input_megabytes +=  ln->stats.n_input_bytes >> 20; +	fhs->fcp_output_megabytes +=  ln->stats.n_output_bytes >> 20; +	fhs->link_failure_count = ln->stats.n_link_down; +	/* Reset stats for the device */ +	seconds = jiffies_to_msecs(jiffies) - hw->stats.n_reset_start; +	do_div(seconds, 1000); +	fhs->seconds_since_last_reset = seconds; + +	return fhs; +} + +/* + * csio_set_rport_loss_tmo - Set the rport dev loss timeout + * @rport: fc rport. + * @timeout: new value for dev loss tmo. + * + * If timeout is non zero set the dev_loss_tmo to timeout, else set + * dev_loss_tmo to one. + */ +static void +csio_set_rport_loss_tmo(struct fc_rport *rport, uint32_t timeout) +{ +	if (timeout) +		rport->dev_loss_tmo = timeout; +	else +		rport->dev_loss_tmo = 1; +} + +static void +csio_vport_set_state(struct csio_lnode *ln) +{ +	struct fc_vport *fc_vport = ln->fc_vport; +	struct csio_lnode  *pln = ln->pln; +	char state[16]; + +	/* Set fc vport state based on phyiscal lnode */ +	csio_lnode_state_to_str(pln, state); +	if (strcmp(state, "READY")) { +		fc_vport_set_state(fc_vport, FC_VPORT_LINKDOWN); +		return; +	} + +	if (!(pln->flags & CSIO_LNF_NPIVSUPP)) { +		fc_vport_set_state(fc_vport, FC_VPORT_NO_FABRIC_SUPP); +		return; +	} + +	/* Set fc vport state based on virtual lnode */ +	csio_lnode_state_to_str(ln, state); +	if (strcmp(state, "READY")) { +		fc_vport_set_state(fc_vport, FC_VPORT_LINKDOWN); +		return; +	} +	fc_vport_set_state(fc_vport, FC_VPORT_ACTIVE); +} + +static int +csio_fcoe_alloc_vnp(struct csio_hw *hw, struct csio_lnode *ln) +{ +	struct csio_lnode *pln; +	struct csio_mb  *mbp; +	struct fw_fcoe_vnp_cmd *rsp; +	int ret = 0; +	int retry = 0; + +	/* Issue VNP cmd to alloc vport */ +	/* Allocate Mbox request */ +	spin_lock_irq(&hw->lock); +	mbp = mempool_alloc(hw->mb_mempool, GFP_ATOMIC); +	if (!mbp) { +		CSIO_INC_STATS(hw, n_err_nomem); +		ret = -ENOMEM; +		goto out; +	} + +	pln = ln->pln; +	ln->fcf_flowid = pln->fcf_flowid; +	ln->portid = pln->portid; + +	csio_fcoe_vnp_alloc_init_mb(ln, mbp, CSIO_MB_DEFAULT_TMO, +				    pln->fcf_flowid, pln->vnp_flowid, 0, +				    csio_ln_wwnn(ln), csio_ln_wwpn(ln), NULL); + +	for (retry = 0; retry < 3; retry++) { +		/* FW is expected to complete vnp cmd in immediate mode +		 * without much delay. +		 * Otherwise, there will be increase in IO latency since HW +		 * lock is held till completion of vnp mbox cmd. +		 */ +		ret = csio_mb_issue(hw, mbp); +		if (ret != -EBUSY) +			break; + +		/* Retry if mbox returns busy */ +		spin_unlock_irq(&hw->lock); +		msleep(2000); +		spin_lock_irq(&hw->lock); +	} + +	if (ret) { +		csio_ln_err(ln, "Failed to issue mbox FCoE VNP command\n"); +		goto out_free; +	} + +	/* Process Mbox response of VNP command */ +	rsp = (struct fw_fcoe_vnp_cmd *)(mbp->mb); +	if (FW_CMD_RETVAL_GET(ntohl(rsp->alloc_to_len16)) != FW_SUCCESS) { +		csio_ln_err(ln, "FCOE VNP ALLOC cmd returned 0x%x!\n", +			    FW_CMD_RETVAL_GET(ntohl(rsp->alloc_to_len16))); +		ret = -EINVAL; +		goto out_free; +	} + +	ln->vnp_flowid = FW_FCOE_VNP_CMD_VNPI_GET( +				ntohl(rsp->gen_wwn_to_vnpi)); +	memcpy(csio_ln_wwnn(ln), rsp->vnport_wwnn, 8); +	memcpy(csio_ln_wwpn(ln), rsp->vnport_wwpn, 8); + +	csio_ln_dbg(ln, "FCOE VNPI: 0x%x\n", ln->vnp_flowid); +	csio_ln_dbg(ln, "\tWWNN: %x%x%x%x%x%x%x%x\n", +		    ln->ln_sparm.wwnn[0], ln->ln_sparm.wwnn[1], +		    ln->ln_sparm.wwnn[2], ln->ln_sparm.wwnn[3], +		    ln->ln_sparm.wwnn[4], ln->ln_sparm.wwnn[5], +		    ln->ln_sparm.wwnn[6], ln->ln_sparm.wwnn[7]); +	csio_ln_dbg(ln, "\tWWPN: %x%x%x%x%x%x%x%x\n", +		    ln->ln_sparm.wwpn[0], ln->ln_sparm.wwpn[1], +		    ln->ln_sparm.wwpn[2], ln->ln_sparm.wwpn[3], +		    ln->ln_sparm.wwpn[4], ln->ln_sparm.wwpn[5], +		    ln->ln_sparm.wwpn[6], ln->ln_sparm.wwpn[7]); + +out_free: +	mempool_free(mbp, hw->mb_mempool); +out: +	spin_unlock_irq(&hw->lock); +	return ret; +} + +static int +csio_fcoe_free_vnp(struct csio_hw *hw, struct csio_lnode *ln) +{ +	struct csio_lnode *pln; +	struct csio_mb  *mbp; +	struct fw_fcoe_vnp_cmd *rsp; +	int ret = 0; +	int retry = 0; + +	/* Issue VNP cmd to free vport */ +	/* Allocate Mbox request */ + +	spin_lock_irq(&hw->lock); +	mbp = mempool_alloc(hw->mb_mempool, GFP_ATOMIC); +	if (!mbp) { +		CSIO_INC_STATS(hw, n_err_nomem); +		ret = -ENOMEM; +		goto out; +	} + +	pln = ln->pln; + +	csio_fcoe_vnp_free_init_mb(ln, mbp, CSIO_MB_DEFAULT_TMO, +				   ln->fcf_flowid, ln->vnp_flowid, +				   NULL); + +	for (retry = 0; retry < 3; retry++) { +		ret = csio_mb_issue(hw, mbp); +		if (ret != -EBUSY) +			break; + +		/* Retry if mbox returns busy */ +		spin_unlock_irq(&hw->lock); +		msleep(2000); +		spin_lock_irq(&hw->lock); +	} + +	if (ret) { +		csio_ln_err(ln, "Failed to issue mbox FCoE VNP command\n"); +		goto out_free; +	} + +	/* Process Mbox response of VNP command */ +	rsp = (struct fw_fcoe_vnp_cmd *)(mbp->mb); +	if (FW_CMD_RETVAL_GET(ntohl(rsp->alloc_to_len16)) != FW_SUCCESS) { +		csio_ln_err(ln, "FCOE VNP FREE cmd returned 0x%x!\n", +			    FW_CMD_RETVAL_GET(ntohl(rsp->alloc_to_len16))); +		ret = -EINVAL; +	} + +out_free: +	mempool_free(mbp, hw->mb_mempool); +out: +	spin_unlock_irq(&hw->lock); +	return ret; +} + +static int +csio_vport_create(struct fc_vport *fc_vport, bool disable) +{ +	struct Scsi_Host *shost = fc_vport->shost; +	struct csio_lnode *pln = shost_priv(shost); +	struct csio_lnode *ln = NULL; +	struct csio_hw *hw = csio_lnode_to_hw(pln); +	uint8_t wwn[8]; +	int ret = -1; + +	ln = csio_shost_init(hw, &fc_vport->dev, false, pln); +	if (!ln) +		goto error; + +	if (fc_vport->node_name != 0) { +		u64_to_wwn(fc_vport->node_name, wwn); + +		if (!CSIO_VALID_WWN(wwn)) { +			csio_ln_err(ln, +				    "vport create failed. Invalid wwnn\n"); +			goto error; +		} +		memcpy(csio_ln_wwnn(ln), wwn, 8); +	} + +	if (fc_vport->port_name != 0) { +		u64_to_wwn(fc_vport->port_name, wwn); + +		if (!CSIO_VALID_WWN(wwn)) { +			csio_ln_err(ln, +				    "vport create failed. Invalid wwpn\n"); +			goto error; +		} + +		if (csio_lnode_lookup_by_wwpn(hw, wwn)) { +			csio_ln_err(ln, +			    "vport create failed. wwpn already exists\n"); +			goto error; +		} +		memcpy(csio_ln_wwpn(ln), wwn, 8); +	} + +	fc_vport_set_state(fc_vport, FC_VPORT_INITIALIZING); + +	if (csio_fcoe_alloc_vnp(hw, ln)) +		goto error; + +	*(struct csio_lnode **)fc_vport->dd_data = ln; +	ln->fc_vport = fc_vport; +	if (!fc_vport->node_name) +		fc_vport->node_name = wwn_to_u64(csio_ln_wwnn(ln)); +	if (!fc_vport->port_name) +		fc_vport->port_name = wwn_to_u64(csio_ln_wwpn(ln)); +	csio_fchost_attr_init(ln); +	return 0; +error: +	if (ln) +		csio_shost_exit(ln); + +	return ret; +} + +static int +csio_vport_delete(struct fc_vport *fc_vport) +{ +	struct csio_lnode *ln = *(struct csio_lnode **)fc_vport->dd_data; +	struct Scsi_Host *shost = csio_ln_to_shost(ln); +	struct csio_hw *hw = csio_lnode_to_hw(ln); +	int rmv; + +	spin_lock_irq(&hw->lock); +	rmv = csio_is_hw_removing(hw); +	spin_unlock_irq(&hw->lock); + +	if (rmv) { +		csio_shost_exit(ln); +		return 0; +	} + +	/* Quiesce ios and send remove event to lnode */ +	scsi_block_requests(shost); +	spin_lock_irq(&hw->lock); +	csio_scsim_cleanup_io_lnode(csio_hw_to_scsim(hw), ln); +	csio_lnode_close(ln); +	spin_unlock_irq(&hw->lock); +	scsi_unblock_requests(shost); + +	/* Free vnp */ +	if (fc_vport->vport_state !=  FC_VPORT_DISABLED) +		csio_fcoe_free_vnp(hw, ln); + +	csio_shost_exit(ln); +	return 0; +} + +static int +csio_vport_disable(struct fc_vport *fc_vport, bool disable) +{ +	struct csio_lnode *ln = *(struct csio_lnode **)fc_vport->dd_data; +	struct Scsi_Host *shost = csio_ln_to_shost(ln); +	struct csio_hw *hw = csio_lnode_to_hw(ln); + +	/* disable vport */ +	if (disable) { +		/* Quiesce ios and send stop event to lnode */ +		scsi_block_requests(shost); +		spin_lock_irq(&hw->lock); +		csio_scsim_cleanup_io_lnode(csio_hw_to_scsim(hw), ln); +		csio_lnode_stop(ln); +		spin_unlock_irq(&hw->lock); +		scsi_unblock_requests(shost); + +		/* Free vnp */ +		csio_fcoe_free_vnp(hw, ln); +		fc_vport_set_state(fc_vport, FC_VPORT_DISABLED); +		csio_ln_err(ln, "vport disabled\n"); +		return 0; +	} else { +		/* enable vport */ +		fc_vport_set_state(fc_vport, FC_VPORT_INITIALIZING); +		if (csio_fcoe_alloc_vnp(hw, ln)) { +			csio_ln_err(ln, "vport enabled failed.\n"); +			return -1; +		} +		csio_ln_err(ln, "vport enabled\n"); +		return 0; +	} +} + +static void +csio_dev_loss_tmo_callbk(struct fc_rport *rport) +{ +	struct csio_rnode *rn; +	struct csio_hw *hw; +	struct csio_lnode *ln; + +	rn = *((struct csio_rnode **)rport->dd_data); +	ln = csio_rnode_to_lnode(rn); +	hw = csio_lnode_to_hw(ln); + +	spin_lock_irq(&hw->lock); + +	/* return if driver is being removed or same rnode comes back online */ +	if (csio_is_hw_removing(hw) || csio_is_rnode_ready(rn)) +		goto out; + +	csio_ln_dbg(ln, "devloss timeout on rnode:%p portid:x%x flowid:x%x\n", +		    rn, rn->nport_id, csio_rn_flowid(rn)); + +	CSIO_INC_STATS(ln, n_dev_loss_tmo); + +	/* +	 * enqueue devloss event to event worker thread to serialize all +	 * rnode events. +	 */ +	if (csio_enqueue_evt(hw, CSIO_EVT_DEV_LOSS, &rn, sizeof(rn))) { +		CSIO_INC_STATS(hw, n_evt_drop); +		goto out; +	} + +	if (!(hw->flags & CSIO_HWF_FWEVT_PENDING)) { +		hw->flags |= CSIO_HWF_FWEVT_PENDING; +		spin_unlock_irq(&hw->lock); +		schedule_work(&hw->evtq_work); +		return; +	} + +out: +	spin_unlock_irq(&hw->lock); +} + +/* FC transport functions template - Physical port */ +struct fc_function_template csio_fc_transport_funcs = { +	.show_host_node_name = 1, +	.show_host_port_name = 1, +	.show_host_supported_classes = 1, +	.show_host_supported_fc4s = 1, +	.show_host_maxframe_size = 1, + +	.get_host_port_id = csio_get_host_port_id, +	.show_host_port_id = 1, + +	.get_host_port_type = csio_get_host_port_type, +	.show_host_port_type = 1, + +	.get_host_port_state = csio_get_host_port_state, +	.show_host_port_state = 1, + +	.show_host_active_fc4s = 1, +	.get_host_speed = csio_get_host_speed, +	.show_host_speed = 1, +	.get_host_fabric_name = csio_get_host_fabric_name, +	.show_host_fabric_name = 1, + +	.get_fc_host_stats = csio_get_stats, + +	.dd_fcrport_size = sizeof(struct csio_rnode *), +	.show_rport_maxframe_size = 1, +	.show_rport_supported_classes = 1, + +	.set_rport_dev_loss_tmo = csio_set_rport_loss_tmo, +	.show_rport_dev_loss_tmo = 1, + +	.show_starget_port_id = 1, +	.show_starget_node_name = 1, +	.show_starget_port_name = 1, + +	.dev_loss_tmo_callbk = csio_dev_loss_tmo_callbk, +	.dd_fcvport_size = sizeof(struct csio_lnode *), + +	.vport_create = csio_vport_create, +	.vport_disable = csio_vport_disable, +	.vport_delete = csio_vport_delete, +}; + +/* FC transport functions template - Virtual  port */ +struct fc_function_template csio_fc_transport_vport_funcs = { +	.show_host_node_name = 1, +	.show_host_port_name = 1, +	.show_host_supported_classes = 1, +	.show_host_supported_fc4s = 1, +	.show_host_maxframe_size = 1, + +	.get_host_port_id = csio_get_host_port_id, +	.show_host_port_id = 1, + +	.get_host_port_type = csio_get_host_port_type, +	.show_host_port_type = 1, + +	.get_host_port_state = csio_get_host_port_state, +	.show_host_port_state = 1, +	.show_host_active_fc4s = 1, + +	.get_host_speed = csio_get_host_speed, +	.show_host_speed = 1, + +	.get_host_fabric_name = csio_get_host_fabric_name, +	.show_host_fabric_name = 1, + +	.get_fc_host_stats = csio_get_stats, + +	.dd_fcrport_size = sizeof(struct csio_rnode *), +	.show_rport_maxframe_size = 1, +	.show_rport_supported_classes = 1, + +	.set_rport_dev_loss_tmo = csio_set_rport_loss_tmo, +	.show_rport_dev_loss_tmo = 1, + +	.show_starget_port_id = 1, +	.show_starget_node_name = 1, +	.show_starget_port_name = 1, + +	.dev_loss_tmo_callbk = csio_dev_loss_tmo_callbk, + +}; |