Message ID | 20241018161409.4442-10-kartilak@cisco.com |
---|---|
State | New |
Headers | show |
Series | Introduce support for Fabric Discovery and... | expand |
Hi Karan, kernel test robot noticed the following build warnings: [auto build test WARNING on mkp-scsi/for-next] [cannot apply to jejb-scsi/for-next linus/master v6.12-rc3 next-20241018] [If your patch is applied to the wrong git tree, kindly drop us a note. And when submitting patch, we suggest to use '--base' as documented in https://git-scm.com/docs/git-format-patch#_base_tree_information] url: https://github.com/intel-lab-lkp/linux/commits/Karan-Tilak-Kumar/scsi-fnic-Replace-shost_printk-with-dev_info-dev_err/20241019-002800 base: https://git.kernel.org/pub/scm/linux/kernel/git/mkp/scsi.git for-next patch link: https://lore.kernel.org/r/20241018161409.4442-10-kartilak%40cisco.com patch subject: [PATCH v5 09/14] scsi: fnic: Modify IO path to use FDLS config: x86_64-kexec (https://download.01.org/0day-ci/archive/20241021/202410210147.fQp7tYeb-lkp@intel.com/config) compiler: clang version 18.1.8 (https://github.com/llvm/llvm-project 3b5b5c1ec4a3095ab096dd780e84d7ab81f3d7ff) reproduce (this is a W=1 build): (https://download.01.org/0day-ci/archive/20241021/202410210147.fQp7tYeb-lkp@intel.com/reproduce) If you fix the issue in a separate patch/commit (i.e. not just a new version of the same patch/commit), kindly add following tags | Reported-by: kernel test robot <lkp@intel.com> | Closes: https://lore.kernel.org/oe-kbuild-all/202410210147.fQp7tYeb-lkp@intel.com/ All warnings (new ones prefixed by >>): >> drivers/scsi/fnic/fnic_scsi.c:269:15: warning: variable 'ioreq_count' set but not used [-Wunused-but-set-variable] 269 | unsigned int ioreq_count; | ^ drivers/scsi/fnic/fnic_scsi.c:928:6: warning: variable 'xfer_len' set but not used [-Wunused-but-set-variable] 928 | u64 xfer_len = 0; | ^ drivers/scsi/fnic/fnic_scsi.c:2506:21: warning: variable 'fnic_stats' set but not used [-Wunused-but-set-variable] 2506 | struct fnic_stats *fnic_stats; | ^ 3 warnings generated. vim +/ioreq_count +269 drivers/scsi/fnic/fnic_scsi.c 258 259 260 /* 261 * fnic_fw_reset_handler 262 * Routine to send reset msg to fw 263 */ 264 int fnic_fw_reset_handler(struct fnic *fnic) 265 { 266 struct vnic_wq_copy *wq = &fnic->hw_copy_wq[0]; 267 int ret = 0; 268 unsigned long flags; > 269 unsigned int ioreq_count; 270 271 /* indicate fwreset to io path */ 272 fnic_set_state_flags(fnic, FNIC_FLAGS_FWRESET); 273 ioreq_count = fnic_count_all_ioreqs(fnic); 274 275 /* wait for io cmpl */ 276 while (atomic_read(&fnic->in_flight)) 277 schedule_timeout(msecs_to_jiffies(1)); 278 279 spin_lock_irqsave(&fnic->wq_copy_lock[0], flags); 280 281 if (vnic_wq_copy_desc_avail(wq) <= fnic->wq_copy_desc_low[0]) 282 free_wq_copy_descs(fnic, wq, 0); 283 284 if (!vnic_wq_copy_desc_avail(wq)) 285 ret = -EAGAIN; 286 else { 287 fnic_queue_wq_copy_desc_fw_reset(wq, SCSI_NO_TAG); 288 atomic64_inc(&fnic->fnic_stats.fw_stats.active_fw_reqs); 289 if (atomic64_read(&fnic->fnic_stats.fw_stats.active_fw_reqs) > 290 atomic64_read(&fnic->fnic_stats.fw_stats.max_fw_reqs)) 291 atomic64_set(&fnic->fnic_stats.fw_stats.max_fw_reqs, 292 atomic64_read( 293 &fnic->fnic_stats.fw_stats.active_fw_reqs)); 294 } 295 296 spin_unlock_irqrestore(&fnic->wq_copy_lock[0], flags); 297 298 if (!ret) { 299 atomic64_inc(&fnic->fnic_stats.reset_stats.fw_resets); 300 FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, 301 "Issued fw reset\n"); 302 } else { 303 fnic_clear_state_flags(fnic, FNIC_FLAGS_FWRESET); 304 FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, 305 "Failed to issue fw reset\n"); 306 } 307 308 return ret; 309 } 310
On 10/18/24 18:14, Karan Tilak Kumar wrote: > Modify IO path to use FDLS. > Add helper functions to process IOs. > Remove unused template functions. > Cleanup obsolete code. > Refactor old function definitions. > > Reviewed-by: Sesidhar Baddela <sebaddel@cisco.com> > Reviewed-by: Arulprabhu Ponnusamy <arulponn@cisco.com> > Reviewed-by: Gian Carlo Boffa <gcboffa@cisco.com> > Reviewed-by: Arun Easi <aeasi@cisco.com> > Signed-off-by: Karan Tilak Kumar <kartilak@cisco.com> > --- > Changes between v2 and v3: > Replace fnic->host with fnic->lport->host to prevent compilation > errors. > > Changes between v1 and v2: > Replace pr_err with printk. > Incorporate review comments by Hannes: > Restore usage of tagset iterators. > --- > drivers/scsi/fnic/fnic.h | 20 +- > drivers/scsi/fnic/fnic_io.h | 3 + > drivers/scsi/fnic/fnic_main.c | 5 +- > drivers/scsi/fnic/fnic_scsi.c | 846 +++++++++++++++++++-------------- > drivers/scsi/fnic/fnic_stats.h | 2 - > 5 files changed, 513 insertions(+), 363 deletions(-) > > diff --git a/drivers/scsi/fnic/fnic.h b/drivers/scsi/fnic/fnic.h > index 4f38cbae1994..c6fe9eec9a0c 100644 > --- a/drivers/scsi/fnic/fnic.h > +++ b/drivers/scsi/fnic/fnic.h > @@ -40,6 +40,7 @@ > #define FNIC_DFLT_IO_REQ 256 /* Default scsi_cmnd tag map entries */ > #define FNIC_DFLT_QUEUE_DEPTH 256 > #define FNIC_STATS_RATE_LIMIT 4 /* limit rate at which stats are pulled up */ > +#define LUN0_DELAY_TIME 9 > > /* > * Tag bits used for special requests. > @@ -472,7 +473,6 @@ int fnic_set_intr_mode_msix(struct fnic *fnic); > void fnic_free_intr(struct fnic *fnic); > int fnic_request_intr(struct fnic *fnic); > > -int fnic_send(struct fc_lport *, struct fc_frame *); > void fnic_free_wq_buf(struct vnic_wq *wq, struct vnic_wq_buf *buf); > void fnic_handle_frame(struct work_struct *work); > void fnic_tport_event_handler(struct work_struct *work); > @@ -489,11 +489,9 @@ int fnic_abort_cmd(struct scsi_cmnd *); > int fnic_device_reset(struct scsi_cmnd *); > int fnic_eh_host_reset_handler(struct scsi_cmnd *sc); > int fnic_host_reset(struct Scsi_Host *shost); > -int fnic_reset(struct Scsi_Host *); > -void fnic_scsi_cleanup(struct fc_lport *); > -void fnic_scsi_abort_io(struct fc_lport *); > -void fnic_empty_scsi_cleanup(struct fc_lport *); > -void fnic_exch_mgr_reset(struct fc_lport *, u32, u32); > +void fnic_reset(struct Scsi_Host *shost); > +int fnic_issue_fc_host_lip(struct Scsi_Host *shost); > +void fnic_scsi_fcpio_reset(struct fnic *fnic); > int fnic_wq_copy_cmpl_handler(struct fnic *fnic, int copy_work_to_do, unsigned int cq_index); > int fnic_wq_cmpl_handler(struct fnic *fnic, int); > int fnic_flogi_reg_handler(struct fnic *fnic, u32); > @@ -505,7 +503,8 @@ const char *fnic_state_to_str(unsigned int state); > void fnic_mq_map_queues_cpus(struct Scsi_Host *host); > void fnic_log_q_error(struct fnic *fnic); > void fnic_handle_link_event(struct fnic *fnic); > - > +void fnic_stats_debugfs_init(struct fnic *fnic); > +void fnic_stats_debugfs_remove(struct fnic *fnic); > int fnic_is_abts_pending(struct fnic *, struct scsi_cmnd *); > > void fnic_handle_fip_frame(struct work_struct *work); > @@ -526,5 +525,12 @@ int fnic_get_desc_by_devid(struct pci_dev *pdev, char **desc, > void fnic_fdls_link_status_change(struct fnic *fnic, int linkup); > void fnic_delete_fcp_tports(struct fnic *fnic); > void fnic_flush_tport_event_list(struct fnic *fnic); > +int fnic_count_ioreqs_wq(struct fnic *fnic, u32 hwq, u32 portid); > +unsigned int fnic_count_ioreqs(struct fnic *fnic, u32 portid); > +unsigned int fnic_count_all_ioreqs(struct fnic *fnic); > +unsigned int fnic_count_lun_ioreqs_wq(struct fnic *fnic, u32 hwq, > + struct scsi_device *device); > +unsigned int fnic_count_lun_ioreqs(struct fnic *fnic, > + struct scsi_device *device); > > #endif /* _FNIC_H_ */ > diff --git a/drivers/scsi/fnic/fnic_io.h b/drivers/scsi/fnic/fnic_io.h > index 6fe642cb387b..0d974e040ab7 100644 > --- a/drivers/scsi/fnic/fnic_io.h > +++ b/drivers/scsi/fnic/fnic_io.h > @@ -7,6 +7,7 @@ > #define _FNIC_IO_H_ > > #include <scsi/fc/fc_fcp.h> > +#include "fnic_fdls.h" > > #define FNIC_DFLT_SG_DESC_CNT 32 > #define FNIC_MAX_SG_DESC_CNT 256 /* Maximum descriptors per sgl */ > @@ -41,6 +42,8 @@ enum fnic_ioreq_state { > }; > > struct fnic_io_req { > + struct fnic_iport_s *iport; > + struct fnic_tport_s *tport; > struct host_sg_desc *sgl_list; /* sgl list */ > void *sgl_list_alloc; /* sgl list address used for free */ > dma_addr_t sense_buf_pa; /* dma address for sense buffer*/ > diff --git a/drivers/scsi/fnic/fnic_main.c b/drivers/scsi/fnic/fnic_main.c > index 33c63d7a9787..dadbe47c0bbb 100644 > --- a/drivers/scsi/fnic/fnic_main.c > +++ b/drivers/scsi/fnic/fnic_main.c > @@ -85,9 +85,6 @@ module_param(fnic_max_qdepth, uint, S_IRUGO|S_IWUSR); > MODULE_PARM_DESC(fnic_max_qdepth, "Queue depth to report for each LUN"); > > static struct libfc_function_template fnic_transport_template = { > - .fcp_abort_io = fnic_empty_scsi_cleanup, > - .fcp_cleanup = fnic_empty_scsi_cleanup, > - .exch_mgr_reset = fnic_exch_mgr_reset > }; > The template is now empty; can't you delete it completely? > struct workqueue_struct *fnic_fip_queue; > @@ -162,7 +159,7 @@ static struct fc_function_template fnic_fc_functions = { > .show_starget_port_id = 1, > .show_rport_dev_loss_tmo = 1, > .set_rport_dev_loss_tmo = fnic_set_rport_dev_loss_tmo, > - .issue_fc_host_lip = fnic_reset, > + .issue_fc_host_lip = fnic_issue_fc_host_lip, > .get_fc_host_stats = fnic_get_stats, > .reset_fc_host_stats = fnic_reset_host_stats, > .dd_fcrport_size = sizeof(struct fc_rport_libfc_priv), > diff --git a/drivers/scsi/fnic/fnic_scsi.c b/drivers/scsi/fnic/fnic_scsi.c > index 74298f9a34e5..af729637ef2d 100644 > --- a/drivers/scsi/fnic/fnic_scsi.c > +++ b/drivers/scsi/fnic/fnic_scsi.c > @@ -25,9 +25,12 @@ > #include <scsi/fc/fc_fcoe.h> > #include <scsi/libfc.h> > #include <scsi/fc_frame.h> > +#include <scsi/scsi_transport_fc.h> > #include "fnic_io.h" > #include "fnic.h" > > +static void fnic_cleanup_io(struct fnic *fnic, int exclude_id); > + > const char *fnic_state_str[] = { > [FNIC_IN_FC_MODE] = "FNIC_IN_FC_MODE", > [FNIC_IN_FC_TRANS_ETH_MODE] = "FNIC_IN_FC_TRANS_ETH_MODE", > @@ -65,6 +68,18 @@ static const char *fcpio_status_str[] = { > [FCPIO_LUNMAP_CHNG_PEND] = "FCPIO_LUNHMAP_CHNG_PEND", > }; > > +enum terminate_io_return { > + TERM_SUCCESS = 0, > + TERM_NO_SC = 1, > + TERM_IO_REQ_NOT_FOUND, > + TERM_ANOTHER_PORT, > + TERM_GSTATE, > + TERM_IO_BLOCKED, > + TERM_OUT_OF_WQ_DESC, > + TERM_TIMED_OUT, > + TERM_MISC, > +}; > + > const char *fnic_state_to_str(unsigned int state) > { > if (state >= ARRAY_SIZE(fnic_state_str) || !fnic_state_str[state]) > @@ -90,8 +105,6 @@ static const char *fnic_fcpio_status_to_str(unsigned int status) > return fcpio_status_str[status]; > } > > -static void fnic_cleanup_io(struct fnic *fnic); > - > /* > * Unmap the data buffer and sense buffer for an io_req, > * also unmap and free the device-private scatter/gather list. > @@ -114,6 +127,80 @@ static void fnic_release_ioreq_buf(struct fnic *fnic, > SCSI_SENSE_BUFFERSIZE, DMA_FROM_DEVICE); > } > > +int fnic_count_ioreqs_wq(struct fnic *fnic, u32 hwq, u32 portid) > +{ > + unsigned long flags = 0; > + int i = 0, count = 0; > + > + spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > + for (i = 0; i != fnic->sw_copy_wq[hwq].ioreq_table_size; ++i) { > + if (fnic->sw_copy_wq[hwq].io_req_table[i] != NULL && > + (!portid > + || fnic->sw_copy_wq[hwq].io_req_table[i]->port_id == portid)) > + count++; > + } > + spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > + > + return count; > +} > + > +unsigned int fnic_count_ioreqs(struct fnic *fnic, u32 portid) > +{ > + int i; > + unsigned int count = 0; > + > + for (i = 0; i < fnic->wq_copy_count; i++) > + count += fnic_count_ioreqs_wq(fnic, i, portid); > + > + return count; > +} > + > +unsigned int fnic_count_all_ioreqs(struct fnic *fnic) > +{ > + return fnic_count_ioreqs(fnic, 0); > +} > + > +unsigned int fnic_count_lun_ioreqs_wq(struct fnic *fnic, u32 hwq, > + struct scsi_device *device) > +{ > + struct fnic_io_req *io_req; > + int i; > + unsigned int count = 0; > + unsigned long flags = 0; > + > + spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > + for (i = 0; i != fnic->sw_copy_wq[hwq].ioreq_table_size; ++i) { > + io_req = fnic->sw_copy_wq[hwq].io_req_table[i]; > + > + if (io_req != NULL) { > + struct scsi_cmnd *sc = > + scsi_host_find_tag(fnic->lport->host, io_req->tag); > + Be careful. 'scsi_host_find_tag()' finds requests on _all_ hwqs. But here you are interested in tags for a specific hwq only. > + if (!sc) > + continue; > + > + if (sc->device == device) > + count++; > + } > + } > + spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > + > + return count; > +} > + And now this is pretty pointless. Please use the block layer blk_mq_tagset_busy_iter() to iterate over busy/active tags. > +unsigned int fnic_count_lun_ioreqs(struct fnic *fnic, > + struct scsi_device *device) > +{ > + int hwq; > + unsigned int count = 0; > + > + /*count if any pending IOs on this lun */ > + for (hwq = 0; hwq < fnic->wq_copy_count; hwq++) > + count += fnic_count_lun_ioreqs_wq(fnic, hwq, device); > + > + return count; > +} > + Same here. > /* Free up Copy Wq descriptors. Called with copy_wq lock held */ > static int free_wq_copy_descs(struct fnic *fnic, struct vnic_wq_copy *wq, unsigned int hwq) > { > @@ -179,12 +266,11 @@ int fnic_fw_reset_handler(struct fnic *fnic) > struct vnic_wq_copy *wq = &fnic->hw_copy_wq[0]; > int ret = 0; > unsigned long flags; > + unsigned int ioreq_count; > > /* indicate fwreset to io path */ > fnic_set_state_flags(fnic, FNIC_FLAGS_FWRESET); > - > - fnic_free_txq(&fnic->frame_queue); > - fnic_free_txq(&fnic->tx_queue); > + ioreq_count = fnic_count_all_ioreqs(fnic); > > /* wait for io cmpl */ > while (atomic_read(&fnic->in_flight)) > @@ -231,10 +317,10 @@ int fnic_flogi_reg_handler(struct fnic *fnic, u32 fc_id) > { > struct vnic_wq_copy *wq = &fnic->hw_copy_wq[0]; > enum fcpio_flogi_reg_format_type format; > - struct fc_lport *lp = fnic->lport; > u8 gw_mac[ETH_ALEN]; > int ret = 0; > unsigned long flags; > + struct fnic_iport_s *iport = &fnic->iport; > > spin_lock_irqsave(&fnic->wq_copy_lock[0], flags); > > @@ -246,28 +332,23 @@ int fnic_flogi_reg_handler(struct fnic *fnic, u32 fc_id) > goto flogi_reg_ioreq_end; > } > > - if (fnic->ctlr.map_dest) { > - eth_broadcast_addr(gw_mac); > - format = FCPIO_FLOGI_REG_DEF_DEST; > - } else { > - memcpy(gw_mac, fnic->ctlr.dest_addr, ETH_ALEN); > - format = FCPIO_FLOGI_REG_GW_DEST; > - } > + memcpy(gw_mac, fnic->iport.fcfmac, ETH_ALEN); > + format = FCPIO_FLOGI_REG_GW_DEST; > > - if ((fnic->config.flags & VFCF_FIP_CAPABLE) && !fnic->ctlr.map_dest) { > + if (fnic->config.flags & VFCF_FIP_CAPABLE) { > fnic_queue_wq_copy_desc_fip_reg(wq, SCSI_NO_TAG, > fc_id, gw_mac, > - fnic->data_src_addr, > - lp->r_a_tov, lp->e_d_tov); > - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > - "FLOGI FIP reg issued fcid %x src %pM dest %pM\n", > - fc_id, fnic->data_src_addr, gw_mac); > + fnic->iport.fpma, > + iport->r_a_tov, iport->e_d_tov); > + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, > + "FLOGI FIP reg issued fcid: 0x%x src %p dest %p\n", > + fc_id, fnic->iport.fpma, gw_mac); > } else { > fnic_queue_wq_copy_desc_flogi_reg(wq, SCSI_NO_TAG, > format, fc_id, gw_mac); > FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, > - "FLOGI reg issued fcid 0x%x map %d dest 0x%p\n", > - fc_id, fnic->ctlr.map_dest, gw_mac); > + "FLOGI reg issued fcid 0x%x dest %p\n", > + fc_id, gw_mac); > } > > atomic64_inc(&fnic->fnic_stats.fw_stats.active_fw_reqs); > @@ -295,13 +376,17 @@ static inline int fnic_queue_wq_copy_desc(struct fnic *fnic, > { > struct scatterlist *sg; > struct fc_rport *rport = starget_to_rport(scsi_target(sc->device)); > - struct fc_rport_libfc_priv *rp = rport->dd_data; > struct host_sg_desc *desc; > struct misc_stats *misc_stats = &fnic->fnic_stats.misc_stats; > unsigned int i; > int flags; > u8 exch_flags; > struct scsi_lun fc_lun; > + struct fnic_tport_s *tport; > + struct rport_dd_data_s *rdd_data; > + > + rdd_data = rport->dd_data; > + tport = rdd_data->tport; > > if (sg_count) { > /* For each SGE, create a device desc entry */ > @@ -356,7 +441,7 @@ static inline int fnic_queue_wq_copy_desc(struct fnic *fnic, > > exch_flags = 0; > if ((fnic->config.flags & VFCF_FCP_SEQ_LVL_ERR) && > - (rp->flags & FC_RP_FLAGS_RETRY)) > + (tport->tgt_flags & FDLS_FC_RP_FLAGS_RETRY)) > exch_flags |= FCPIO_ICMND_SRFLAG_RETRY; > > fnic_queue_wq_copy_desc_icmnd_16(wq, mqtag, > @@ -371,8 +456,8 @@ static inline int fnic_queue_wq_copy_desc(struct fnic *fnic, > sc->cmnd, sc->cmd_len, > scsi_bufflen(sc), > fc_lun.scsi_lun, io_req->port_id, > - rport->maxframe_size, rp->r_a_tov, > - rp->e_d_tov); > + tport->max_payload_size, > + tport->r_a_tov, tport->e_d_tov); > > atomic64_inc(&fnic->fnic_stats.fw_stats.active_fw_reqs); > if (atomic64_read(&fnic->fnic_stats.fw_stats.active_fw_reqs) > > @@ -388,10 +473,10 @@ int fnic_queuecommand(struct Scsi_Host *shost, struct scsi_cmnd *sc) > struct request *const rq = scsi_cmd_to_rq(sc); > uint32_t mqtag = 0; > void (*done)(struct scsi_cmnd *) = scsi_done; > - struct fc_lport *lp = shost_priv(sc->device->host); > struct fc_rport *rport; > struct fnic_io_req *io_req = NULL; > - struct fnic *fnic = lport_priv(lp); > + struct fnic *fnic = *((struct fnic **) shost_priv(sc->device->host)); > + struct fnic_iport_s *iport = NULL; > struct fnic_stats *fnic_stats = &fnic->fnic_stats; > struct vnic_wq_copy *wq; > int ret = 1; > @@ -400,32 +485,14 @@ int fnic_queuecommand(struct Scsi_Host *shost, struct scsi_cmnd *sc) > unsigned long flags = 0; > unsigned long ptr; > int io_lock_acquired = 0; > - struct fc_rport_libfc_priv *rp; > uint16_t hwq = 0; > - > - mqtag = blk_mq_unique_tag(rq); > - spin_lock_irqsave(&fnic->fnic_lock, flags); > - > - if (unlikely(fnic_chk_state_flags_locked(fnic, FNIC_FLAGS_IO_BLOCKED))) { > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > - FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > - "fnic IO blocked flags: 0x%lx. Returning SCSI_MLQUEUE_HOST_BUSY\n", > - fnic->state_flags); > - return SCSI_MLQUEUE_HOST_BUSY; > - } > - > - if (unlikely(fnic_chk_state_flags_locked(fnic, FNIC_FLAGS_FWRESET))) { > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > - FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > - "fnic flags: 0x%lx. Returning SCSI_MLQUEUE_HOST_BUSY\n", > - fnic->state_flags); > - return SCSI_MLQUEUE_HOST_BUSY; > - } > + struct fnic_tport_s *tport = NULL; > + struct rport_dd_data_s *rdd_data; > + uint16_t lun0_delay = 0; > > rport = starget_to_rport(scsi_target(sc->device)); > if (!rport) { > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > "returning DID_NO_CONNECT for IO as rport is NULL\n"); > sc->result = DID_NO_CONNECT << 16; > done(sc); > @@ -434,50 +501,95 @@ int fnic_queuecommand(struct Scsi_Host *shost, struct scsi_cmnd *sc) > > ret = fc_remote_port_chkready(rport); > if (ret) { > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > "rport is not ready\n"); > - atomic64_inc(&fnic_stats->misc_stats.rport_not_ready); > sc->result = ret; > done(sc); > return 0; > } > > - rp = rport->dd_data; > - if (!rp || rp->rp_state == RPORT_ST_DELETE) { > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > - "rport 0x%x removed, returning DID_NO_CONNECT\n", > - rport->port_id); > + mqtag = blk_mq_unique_tag(rq); > + spin_lock_irqsave(&fnic->fnic_lock, flags); > + iport = &fnic->iport; > > - atomic64_inc(&fnic_stats->misc_stats.rport_not_ready); > - sc->result = DID_NO_CONNECT<<16; > + if (iport->state != FNIC_IPORT_STATE_READY) { > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > + "returning DID_NO_CONNECT for IO as iport state: %d\n", > + iport->state); > + sc->result = DID_NO_CONNECT << 16; > done(sc); > return 0; > } > > - if (rp->rp_state != RPORT_ST_READY) { > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > - "rport 0x%x in state 0x%x, returning DID_IMM_RETRY\n", > - rport->port_id, rp->rp_state); > + /* fc_remote_port_add() may have added the tport to > + * fc_transport but dd_data not yet set > + */ > + rdd_data = rport->dd_data; > + tport = rdd_data->tport; > + if (!tport || (rdd_data->iport != iport)) { > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > + "dd_data not yet set in SCSI for rport portid: 0x%x\n", > + rport->port_id); > + tport = fnic_find_tport_by_fcid(iport, rport->port_id); > + if (!tport) { > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > + "returning DID_BUS_BUSY for IO as tport not found for: 0x%x\n", > + rport->port_id); > + sc->result = DID_BUS_BUSY << 16; > + done(sc); > + return 0; > + } > + > + /* Re-assign same params as in fnic_fdls_add_tport */ > + rport->maxframe_size = FNIC_FC_MAX_PAYLOAD_LEN; > + rport->supported_classes = > + FC_COS_CLASS3 | FC_RPORT_ROLE_FCP_TARGET; > + /* the dd_data is allocated by fctransport of size dd_fcrport_size */ > + rdd_data = rport->dd_data; > + rdd_data->tport = tport; > + rdd_data->iport = iport; > + tport->rport = rport; > + tport->flags |= FNIC_FDLS_SCSI_REGISTERED; > + } > > - sc->result = DID_IMM_RETRY << 16; > + if ((tport->state != FDLS_TGT_STATE_READY) > + && (tport->state != FDLS_TGT_STATE_ADISC)) { > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > + "returning DID_NO_CONNECT for IO as tport state: %d\n", > + tport->state); > + sc->result = DID_NO_CONNECT << 16; > done(sc); > return 0; > } > > - if (lp->state != LPORT_ST_READY || !(lp->link_up)) { > + atomic_inc(&fnic->in_flight); > + atomic_inc(&tport->in_flight); > + > + if (unlikely(fnic_chk_state_flags_locked(fnic, FNIC_FLAGS_IO_BLOCKED))) { > + atomic_dec(&fnic->in_flight); > + atomic_dec(&tport->in_flight); > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > + return SCSI_MLQUEUE_HOST_BUSY; > + } > + > + if (unlikely(fnic_chk_state_flags_locked(fnic, FNIC_FLAGS_FWRESET))) { > spin_unlock_irqrestore(&fnic->fnic_lock, flags); > FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > - "state not ready: %d/link not up: %d Returning HOST_BUSY\n", > - lp->state, lp->link_up); > + "fnic flags FW reset: 0x%lx. Returning SCSI_MLQUEUE_HOST_BUSY\n", > + fnic->state_flags); > return SCSI_MLQUEUE_HOST_BUSY; > } > > - atomic_inc(&fnic->in_flight); > + if (!tport->lun0_delay) { > + lun0_delay = 1; > + tport->lun0_delay++; > + } > > spin_unlock_irqrestore(&fnic->fnic_lock, flags); > + > fnic_priv(sc)->state = FNIC_IOREQ_NOT_INITED; > fnic_priv(sc)->flags = FNIC_NO_FLAGS; > > @@ -499,6 +611,7 @@ int fnic_queuecommand(struct Scsi_Host *shost, struct scsi_cmnd *sc) > goto out; > } > > + io_req->tport = tport; > /* Determine the type of scatter/gather list we need */ > io_req->sgl_cnt = sg_count; > io_req->sgl_type = FNIC_SGL_CACHE_DFLT; > @@ -575,6 +688,7 @@ int fnic_queuecommand(struct Scsi_Host *shost, struct scsi_cmnd *sc) > mempool_free(io_req, fnic->io_req_pool); > } > atomic_dec(&fnic->in_flight); > + atomic_dec(&tport->in_flight); > return ret; > } else { > atomic64_inc(&fnic_stats->io_stats.active_ios); > @@ -602,6 +716,14 @@ int fnic_queuecommand(struct Scsi_Host *shost, struct scsi_cmnd *sc) > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > atomic_dec(&fnic->in_flight); > + atomic_dec(&tport->in_flight); > + > + if (lun0_delay) { > + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, > + "LUN0 delay\n"); > + mdelay(LUN0_DELAY_TIME); > + } > + > return ret; > } > > @@ -625,7 +747,7 @@ static int fnic_fcpio_fw_reset_cmpl_handler(struct fnic *fnic, > atomic64_inc(&reset_stats->fw_reset_completions); > > /* Clean up all outstanding io requests */ > - fnic_cleanup_io(fnic); > + fnic_cleanup_io(fnic, SCSI_NO_TAG); > > atomic64_set(&fnic->fnic_stats.fw_stats.active_fw_reqs, 0); > atomic64_set(&fnic->fnic_stats.io_stats.active_ios, 0); > @@ -646,12 +768,6 @@ static int fnic_fcpio_fw_reset_cmpl_handler(struct fnic *fnic, > "reset failed with header status: %s\n", > fnic_fcpio_status_to_str(hdr_status)); > > - /* > - * Unable to change to eth mode, cannot send out flogi > - * Change state to fc mode, so that subsequent Flogi > - * requests from libFC will cause more attempts to > - * reset the firmware. Free the cached flogi > - */ > fnic->state = FNIC_IN_FC_MODE; > atomic64_inc(&reset_stats->fw_reset_failures); > ret = -1; > @@ -664,15 +780,14 @@ static int fnic_fcpio_fw_reset_cmpl_handler(struct fnic *fnic, > ret = -1; > } > > - /* Thread removing device blocks till firmware reset is complete */ > - if (fnic->remove_wait) > - complete(fnic->remove_wait); > + if (fnic->fw_reset_done) > + complete(fnic->fw_reset_done); > > /* > * If fnic is being removed, or fw reset failed > * free the flogi frame. Else, send it out > */ > - if (fnic->remove_wait || ret) { > + if (ret) { > spin_unlock_irqrestore(&fnic->fnic_lock, flags); > fnic_free_txq(&fnic->tx_queue); > goto reset_cmpl_handler_end; > @@ -711,12 +826,12 @@ static int fnic_fcpio_flogi_reg_cmpl_handler(struct fnic *fnic, > /* Check flogi registration completion status */ > if (!hdr_status) { > FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > - "flog reg succeeded\n"); > + "FLOGI reg succeeded\n"); > fnic->state = FNIC_IN_FC_MODE; > } else { > FNIC_SCSI_DBG(KERN_DEBUG, > fnic->lport->host, fnic->fnic_num, > - "fnic flogi reg :failed %s\n", > + "fnic flogi reg failed: %s\n", > fnic_fcpio_status_to_str(hdr_status)); > fnic->state = FNIC_IN_ETH_MODE; > ret = -1; > @@ -1023,15 +1138,6 @@ static void fnic_fcpio_icmnd_cmpl_handler(struct fnic *fnic, unsigned int cq_ind > jiffies_to_msecs(jiffies - start_time)), > desc, cmd_trace, fnic_flags_and_state(sc)); > > - if (sc->sc_data_direction == DMA_FROM_DEVICE) { > - fnic->lport->host_stats.fcp_input_requests++; > - fnic->fcp_input_bytes += xfer_len; > - } else if (sc->sc_data_direction == DMA_TO_DEVICE) { > - fnic->lport->host_stats.fcp_output_requests++; > - fnic->fcp_output_bytes += xfer_len; > - } else > - fnic->lport->host_stats.fcp_control_requests++; > - > /* Call SCSI completion function to complete the IO */ > scsi_done(sc); > > @@ -1414,8 +1520,8 @@ static bool fnic_cleanup_io_iter(struct scsi_cmnd *sc, void *data) > struct request *const rq = scsi_cmd_to_rq(sc); > struct fnic *fnic = data; > struct fnic_io_req *io_req; > - unsigned long flags = 0; > unsigned long start_time = 0; > + unsigned long flags; > struct fnic_stats *fnic_stats = &fnic->fnic_stats; > uint16_t hwq = 0; > int tag; > @@ -1439,7 +1545,7 @@ static bool fnic_cleanup_io_iter(struct scsi_cmnd *sc, void *data) > } > > if ((fnic_priv(sc)->flags & FNIC_DEVICE_RESET) && > - !(fnic_priv(sc)->flags & FNIC_DEV_RST_DONE)) { > + !(fnic_priv(sc)->flags & FNIC_DEV_RST_DONE)) { > /* > * We will be here only when FW completes reset > * without sending completions for outstanding ios. > @@ -1449,6 +1555,7 @@ static bool fnic_cleanup_io_iter(struct scsi_cmnd *sc, void *data) > complete(io_req->dr_done); > else if (io_req && io_req->abts_done) > complete(io_req->abts_done); > + > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > return true; > } else if (fnic_priv(sc)->flags & FNIC_DEVICE_RESET) { > @@ -1458,19 +1565,19 @@ static bool fnic_cleanup_io_iter(struct scsi_cmnd *sc, void *data) > > fnic_priv(sc)->io_req = NULL; > io_req->sc = NULL; > + start_time = io_req->start_time; > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > /* > * If there is a scsi_cmnd associated with this io_req, then > * free the corresponding state > */ > - start_time = io_req->start_time; > fnic_release_ioreq_buf(fnic, io_req, sc); > mempool_free(io_req, fnic->io_req_pool); > > sc->result = DID_TRANSPORT_DISRUPTED << 16; > FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > - "mqtag:0x%x tag: 0x%x sc:0x%p duration = %lu DID_TRANSPORT_DISRUPTED\n", > + "mqtag: 0x%x tag: 0x%x sc: 0x%p duration = %lu DID_TRANSPORT_DISRUPTED\n", > mqtag, tag, sc, (jiffies - start_time)); > > if (atomic64_read(&fnic->io_cmpl_skip)) > @@ -1479,23 +1586,60 @@ static bool fnic_cleanup_io_iter(struct scsi_cmnd *sc, void *data) > atomic64_inc(&fnic_stats->io_stats.io_completions); > > FNIC_TRACE(fnic_cleanup_io, > - sc->device->host->host_no, tag, sc, > - jiffies_to_msecs(jiffies - start_time), > - 0, ((u64)sc->cmnd[0] << 32 | > - (u64)sc->cmnd[2] << 24 | > - (u64)sc->cmnd[3] << 16 | > - (u64)sc->cmnd[4] << 8 | sc->cmnd[5]), > - fnic_flags_and_state(sc)); > - > + sc->device->host->host_no, tag, sc, > + jiffies_to_msecs(jiffies - start_time), > + 0, ((u64) sc->cmnd[0] << 32 | > + (u64) sc->cmnd[2] << 24 | > + (u64) sc->cmnd[3] << 16 | > + (u64) sc->cmnd[4] << 8 | sc->cmnd[5]), > + (((u64) fnic_priv(sc)->flags << 32) | fnic_priv(sc)-> > + state)); > + > + /* Complete the command to SCSI */ > scsi_done(sc); > - > return true; > } > > -static void fnic_cleanup_io(struct fnic *fnic) > +static void fnic_cleanup_io(struct fnic *fnic, int exclude_id) > { > + unsigned int io_count = 0; > + unsigned long flags; > + struct fnic_io_req *io_req = NULL; > + struct scsi_cmnd *sc = NULL; > + > + io_count = fnic_count_all_ioreqs(fnic); > + FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > + "Outstanding ioreq count: %d active io count: %lld Waiting\n", > + io_count, > + atomic64_read(&fnic->fnic_stats.io_stats.active_ios)); > + > scsi_host_busy_iter(fnic->lport->host, > - fnic_cleanup_io_iter, fnic); > + fnic_cleanup_io_iter, fnic); > + > + /* with sg3utils device reset, SC needs to be retrieved from ioreq */ > + spin_lock_irqsave(&fnic->wq_copy_lock[0], flags); > + io_req = fnic->sw_copy_wq[0].io_req_table[fnic->fnic_max_tag_id]; > + if (io_req) { > + sc = io_req->sc; > + if (sc) { > + if ((fnic_priv(sc)->flags & FNIC_DEVICE_RESET) > + && !(fnic_priv(sc)->flags & FNIC_DEV_RST_DONE)) { > + fnic_priv(sc)->flags |= FNIC_DEV_RST_DONE; > + if (io_req && io_req->dr_done) > + complete(io_req->dr_done); > + } > + } > + } > + spin_unlock_irqrestore(&fnic->wq_copy_lock[0], flags); > + > + while ((io_count = fnic_count_all_ioreqs(fnic))) { > + FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > + "Outstanding ioreq count: %d active io count: %lld Waiting\n", > + io_count, > + atomic64_read(&fnic->fnic_stats.io_stats.active_ios)); > + > + schedule_timeout(msecs_to_jiffies(100)); > + } > } > > void fnic_wq_copy_cleanup_handler(struct vnic_wq_copy *wq, > @@ -1567,10 +1711,13 @@ static inline int fnic_queue_abort_io_req(struct fnic *fnic, int tag, > struct vnic_wq_copy *wq = &fnic->hw_copy_wq[hwq]; > struct misc_stats *misc_stats = &fnic->fnic_stats.misc_stats; > unsigned long flags; > + struct fnic_tport_s *tport = io_req->tport; > > spin_lock_irqsave(&fnic->fnic_lock, flags); > if (unlikely(fnic_chk_state_flags_locked(fnic, > FNIC_FLAGS_IO_BLOCKED))) { > + atomic_dec(&fnic->in_flight); > + atomic_dec(&tport->in_flight); > spin_unlock_irqrestore(&fnic->fnic_lock, flags); > return 1; > } else > @@ -1585,6 +1732,7 @@ static inline int fnic_queue_abort_io_req(struct fnic *fnic, int tag, > if (!vnic_wq_copy_desc_avail(wq)) { > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > atomic_dec(&fnic->in_flight); > + atomic_dec(&tport->in_flight); > FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > "fnic_queue_abort_io_req: failure: no descriptors\n"); > atomic64_inc(&misc_stats->abts_cpwq_alloc_failures); > @@ -1619,20 +1767,24 @@ static bool fnic_rport_abort_io_iter(struct scsi_cmnd *sc, void *data) > struct fnic *fnic = iter_data->fnic; > int abt_tag = 0; > struct fnic_io_req *io_req; > - unsigned long flags; > struct reset_stats *reset_stats = &fnic->fnic_stats.reset_stats; > struct terminate_stats *term_stats = &fnic->fnic_stats.term_stats; > struct scsi_lun fc_lun; > enum fnic_ioreq_state old_ioreq_state; > uint16_t hwq = 0; > + unsigned long flags; > > abt_tag = blk_mq_unique_tag(rq); > hwq = blk_mq_unique_tag_to_hwq(abt_tag); > > - spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > + if (!sc) { > + FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > + "sc is NULL abt_tag: 0x%x hwq: %d\n", abt_tag, hwq); > + return true; > + } > > + spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > io_req = fnic_priv(sc)->io_req; > - > if (!io_req || io_req->port_id != iter_data->port_id) { > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > return true; > @@ -1655,37 +1807,38 @@ static bool fnic_rport_abort_io_iter(struct scsi_cmnd *sc, void *data) > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > return true; > } > + > if (io_req->abts_done) { > shost_printk(KERN_ERR, fnic->lport->host, > - "fnic_rport_exch_reset: io_req->abts_done is set " > - "state is %s\n", > + "fnic_rport_exch_reset: io_req->abts_done is set state is %s\n", > fnic_ioreq_state_to_str(fnic_priv(sc)->state)); > } > > if (!(fnic_priv(sc)->flags & FNIC_IO_ISSUED)) { > shost_printk(KERN_ERR, fnic->lport->host, > - "rport_exch_reset " > - "IO not yet issued %p tag 0x%x flags " > - "%x state %d\n", > - sc, abt_tag, fnic_priv(sc)->flags, fnic_priv(sc)->state); > + "rport_exch_reset IO not yet issued %p abt_tag 0x%x", > + sc, abt_tag); > + shost_printk(KERN_ERR, fnic->lport->host, > + "flags %x state %d\n", fnic_priv(sc)->flags, > + fnic_priv(sc)->state); > } > old_ioreq_state = fnic_priv(sc)->state; > fnic_priv(sc)->state = FNIC_IOREQ_ABTS_PENDING; > fnic_priv(sc)->abts_status = FCPIO_INVALID_CODE; > + > if (fnic_priv(sc)->flags & FNIC_DEVICE_RESET) { > atomic64_inc(&reset_stats->device_reset_terminates); > abt_tag |= FNIC_TAG_DEV_RST; > } > FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > - "fnic_rport_exch_reset dev rst sc 0x%p\n", sc); > - BUG_ON(io_req->abts_done); > - > + "fnic_rport_exch_reset: dev rst sc 0x%p\n", sc); > + WARN_ON_ONCE(io_req->abts_done); > FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > "fnic_rport_reset_exch: Issuing abts\n"); > > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > - /* Now queue the abort command to firmware */ > + /* Queue the abort command to firmware */ > int_to_scsilun(sc->device->lun, &fc_lun); > > if (fnic_queue_abort_io_req(fnic, abt_tag, > @@ -1714,11 +1867,14 @@ static bool fnic_rport_abort_io_iter(struct scsi_cmnd *sc, void *data) > atomic64_inc(&term_stats->terminates); > iter_data->term_cnt++; > } > + > return true; > } > > void fnic_rport_exch_reset(struct fnic *fnic, u32 port_id) > { > + unsigned int io_count = 0; > + unsigned long flags; > struct terminate_stats *term_stats = &fnic->fnic_stats.term_stats; > struct fnic_rport_abort_io_iter_data iter_data = { > .fnic = fnic, > @@ -1726,53 +1882,75 @@ void fnic_rport_exch_reset(struct fnic *fnic, u32 port_id) > .term_cnt = 0, > }; > > - FNIC_SCSI_DBG(KERN_DEBUG, > - fnic->lport->host, fnic->fnic_num, > - "fnic_rport_exch_reset called portid 0x%06x\n", > - port_id); > + FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > + "fnic rport exchange reset for tport: 0x%06x\n", > + port_id); > > if (fnic->in_remove) > return; > > + io_count = fnic_count_ioreqs(fnic, port_id); > + FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > + "Starting terminates: rport:0x%x portid-io-count: %d active-io-count: %lld\n", > + port_id, io_count, > + atomic64_read(&fnic->fnic_stats.io_stats.active_ios)); > + > + spin_lock_irqsave(&fnic->fnic_lock, flags); > + atomic_inc(&fnic->in_flight); > + if (unlikely(fnic_chk_state_flags_locked(fnic, FNIC_FLAGS_IO_BLOCKED))) { > + atomic_dec(&fnic->in_flight); > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > + return; > + } > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > + > scsi_host_busy_iter(fnic->lport->host, fnic_rport_abort_io_iter, > &iter_data); > + > if (iter_data.term_cnt > atomic64_read(&term_stats->max_terminates)) > atomic64_set(&term_stats->max_terminates, iter_data.term_cnt); > > + atomic_dec(&fnic->in_flight); > + > + while ((io_count = fnic_count_ioreqs(fnic, port_id))) > + schedule_timeout(msecs_to_jiffies(1000)); > + That really feels overly complex. You send the aborts with the previous 'scsi_host_busy_iter()' call, and now you wait for the I/Os to complete. But you already have an atomic counter for I/Os in flight to the tport; can't you just wait for that counter to drop to zero? > + FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > + "rport: 0x%x remaining portid-io-count: %d ", > + port_id, io_count); > } > > void fnic_terminate_rport_io(struct fc_rport *rport) > { > - struct fc_rport_libfc_priv *rdata; > - struct fc_lport *lport; > - struct fnic *fnic; > + struct fnic_tport_s *tport; > + struct rport_dd_data_s *rdd_data; > + struct fnic_iport_s *iport = NULL; > + struct fnic *fnic = NULL; > > if (!rport) { > - printk(KERN_ERR "fnic_terminate_rport_io: rport is NULL\n"); > + pr_err("rport is NULL\n"); > return; > } > - rdata = rport->dd_data; > > - if (!rdata) { > - printk(KERN_ERR "fnic_terminate_rport_io: rdata is NULL\n"); > - return; > - } > - lport = rdata->local_port; > - > - if (!lport) { > - printk(KERN_ERR "fnic_terminate_rport_io: lport is NULL\n"); > - return; > + rdd_data = rport->dd_data; > + if (rdd_data) { > + tport = rdd_data->tport; > + if (!tport) { > + pr_err( > + "term rport io called after tport is deleted. Returning 0x%8x\n", > + rport->port_id); > + } else { > + pr_err( > + "term rport io called after tport is set 0x%8x\n", > + rport->port_id); > + pr_err( > + "tport maybe rediscovered\n"); > + > + iport = (struct fnic_iport_s *) tport->iport; > + fnic = iport->fnic; > + fnic_rport_exch_reset(fnic, rport->port_id); > + } > } > - fnic = lport_priv(lport); > - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > - "wwpn 0x%llx, wwnn0x%llx, rport 0x%p, portid 0x%06x\n", > - rport->port_name, rport->node_name, rport, > - rport->port_id); > - > - if (fnic->in_remove) > - return; > - > - fnic_rport_exch_reset(fnic, rport->port_id); > } > > /* > @@ -1783,10 +1961,12 @@ void fnic_terminate_rport_io(struct fc_rport *rport) > int fnic_abort_cmd(struct scsi_cmnd *sc) > { > struct request *const rq = scsi_cmd_to_rq(sc); > - struct fc_lport *lp; > + struct fnic_iport_s *iport; > + struct fnic_tport_s *tport; > struct fnic *fnic; > struct fnic_io_req *io_req = NULL; > struct fc_rport *rport; > + struct rport_dd_data_s *rdd_data; > unsigned long flags; > unsigned long start_time = 0; > int ret = SUCCESS; > @@ -1806,11 +1986,11 @@ int fnic_abort_cmd(struct scsi_cmnd *sc) > fc_block_scsi_eh(sc); > > /* Get local-port, check ready and link up */ > - lp = shost_priv(sc->device->host); > - > - fnic = lport_priv(lp); > + fnic = *((struct fnic **) shost_priv(sc->device->host)); > > spin_lock_irqsave(&fnic->fnic_lock, flags); > + iport = &fnic->iport; > + > fnic_stats = &fnic->fnic_stats; > abts_stats = &fnic->fnic_stats.abts_stats; > term_stats = &fnic->fnic_stats.term_stats; > @@ -1821,7 +2001,43 @@ int fnic_abort_cmd(struct scsi_cmnd *sc) > > fnic_priv(sc)->flags = FNIC_NO_FLAGS; > > - if (lp->state != LPORT_ST_READY || !(lp->link_up)) { > + rdd_data = rport->dd_data; > + tport = rdd_data->tport; > + > + if (!tport) { > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > + "Abort cmd called after tport delete! rport fcid: 0x%x", > + rport->port_id); > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > + "lun: %llu hwq: 0x%x mqtag: 0x%x Op: 0x%x flags: 0x%x\n", > + sc->device->lun, hwq, mqtag, > + sc->cmnd[0], fnic_priv(sc)->flags); > + ret = FAILED; > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > + goto fnic_abort_cmd_end; > + } > + > + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, > + "Abort cmd called rport fcid: 0x%x lun: %llu hwq: 0x%x mqtag: 0x%x", > + rport->port_id, sc->device->lun, hwq, mqtag); > + > + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, > + "Op: 0x%x flags: 0x%x\n", > + sc->cmnd[0], > + fnic_priv(sc)->flags); > + > + if (iport->state != FNIC_IPORT_STATE_READY) { > + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, > + "iport NOT in READY state"); > + ret = FAILED; > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > + goto fnic_abort_cmd_end; > + } > + > + if ((tport->state != FDLS_TGT_STATE_READY) && > + (tport->state != FDLS_TGT_STATE_ADISC)) { > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > + "tport state: %d\n", tport->state); > ret = FAILED; > spin_unlock_irqrestore(&fnic->fnic_lock, flags); > goto fnic_abort_cmd_end; > @@ -1843,6 +2059,7 @@ int fnic_abort_cmd(struct scsi_cmnd *sc) > spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > io_req = fnic_priv(sc)->io_req; > if (!io_req) { > + ret = FAILED; > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > goto fnic_abort_cmd_end; > } > @@ -1893,7 +2110,6 @@ int fnic_abort_cmd(struct scsi_cmnd *sc) > if (fc_remote_port_chkready(rport) == 0) > task_req = FCPIO_ITMF_ABT_TASK; > else { > - atomic64_inc(&fnic_stats->misc_stats.rport_not_ready); > task_req = FCPIO_ITMF_ABT_TASK_TERM; > } > > @@ -2027,6 +2243,7 @@ static inline int fnic_queue_dr_io_req(struct fnic *fnic, > unsigned long flags; > uint16_t hwq = 0; > uint32_t tag = 0; > + struct fnic_tport_s *tport = io_req->tport; > > tag = io_req->tag; > hwq = blk_mq_unique_tag_to_hwq(tag); > @@ -2037,8 +2254,10 @@ static inline int fnic_queue_dr_io_req(struct fnic *fnic, > FNIC_FLAGS_IO_BLOCKED))) { > spin_unlock_irqrestore(&fnic->fnic_lock, flags); > return FAILED; > - } else > + } else { > atomic_inc(&fnic->in_flight); > + atomic_inc(&tport->in_flight); > + } > spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > @@ -2072,6 +2291,7 @@ static inline int fnic_queue_dr_io_req(struct fnic *fnic, > lr_io_req_end: > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > atomic_dec(&fnic->in_flight); > + atomic_inc(&tport->in_flight); > > return ret; > } > @@ -2274,11 +2494,11 @@ static int fnic_clean_pending_aborts(struct fnic *fnic, > int fnic_device_reset(struct scsi_cmnd *sc) > { > struct request *rq = scsi_cmd_to_rq(sc); > - struct fc_lport *lp; > struct fnic *fnic; > struct fnic_io_req *io_req = NULL; > struct fc_rport *rport; > int status; > + int count = 0; > int ret = FAILED; > unsigned long flags; > unsigned long start_time = 0; > @@ -2289,31 +2509,61 @@ int fnic_device_reset(struct scsi_cmnd *sc) > DECLARE_COMPLETION_ONSTACK(tm_done); > bool new_sc = 0; > uint16_t hwq = 0; > + struct fnic_iport_s *iport = NULL; > + struct rport_dd_data_s *rdd_data; > + struct fnic_tport_s *tport; > + u32 old_soft_reset_count; > + u32 old_link_down_cnt; > + int exit_dr = 0; > > /* Wait for rport to unblock */ > fc_block_scsi_eh(sc); > > /* Get local-port, check ready and link up */ > - lp = shost_priv(sc->device->host); > + fnic = *((struct fnic **) shost_priv(sc->device->host)); > + iport = &fnic->iport; > > - fnic = lport_priv(lp); > fnic_stats = &fnic->fnic_stats; > reset_stats = &fnic->fnic_stats.reset_stats; > > atomic64_inc(&reset_stats->device_resets); > > rport = starget_to_rport(scsi_target(sc->device)); > + > + spin_lock_irqsave(&fnic->fnic_lock, flags); > FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > - "fcid: 0x%x lun: 0x%llx hwq: %d mqtag: 0x%x flags: 0x%x Device reset\n", > + "fcid: 0x%x lun: %llu hwq: %d mqtag: 0x%x flags: 0x%x Device reset\n", > rport->port_id, sc->device->lun, hwq, mqtag, > fnic_priv(sc)->flags); > > - if (lp->state != LPORT_ST_READY || !(lp->link_up)) > + rdd_data = rport->dd_data; > + tport = rdd_data->tport; > + if (!tport) { > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > + "Dev rst called after tport delete! rport fcid: 0x%x lun: %llu\n", > + rport->port_id, sc->device->lun); > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > + goto fnic_device_reset_end; > + } > + > + if (iport->state != FNIC_IPORT_STATE_READY) { > + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, > + "iport NOT in READY state"); > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > goto fnic_device_reset_end; > + } > + > + if ((tport->state != FDLS_TGT_STATE_READY) && > + (tport->state != FDLS_TGT_STATE_ADISC)) { > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > + "tport state: %d\n", tport->state); > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > + goto fnic_device_reset_end; > + } > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > Please check if returning FAST_IO_FAIL here wouldn't be a better option. Most of the time a device reset is triggered by a command timeout, which typically happens due to a transport issue (eg link is down or something). So returning 'FAILED' will just escalate to host reset, and that can take for a really long time trying to abort all commands. > /* Check if remote port up */ > if (fc_remote_port_chkready(rport)) { > - atomic64_inc(&fnic_stats->misc_stats.rport_not_ready); > goto fnic_device_reset_end; > } > > @@ -2352,6 +2602,7 @@ int fnic_device_reset(struct scsi_cmnd *sc) > io_req->port_id = rport->port_id; > io_req->tag = mqtag; > fnic_priv(sc)->io_req = io_req; > + io_req->tport = tport; > io_req->sc = sc; > > if (fnic->sw_copy_wq[hwq].io_req_table[blk_mq_unique_tag_to_tag(mqtag)] != NULL) > @@ -2383,6 +2634,11 @@ int fnic_device_reset(struct scsi_cmnd *sc) > fnic_priv(sc)->flags |= FNIC_DEV_RST_ISSUED; > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > + spin_lock_irqsave(&fnic->fnic_lock, flags); > + old_link_down_cnt = iport->fnic->link_down_cnt; > + old_soft_reset_count = fnic->soft_reset_count; > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > + > /* > * Wait on the local completion for LUN reset. The io_req may be > * freed while we wait since we hold no lock. > @@ -2390,6 +2646,24 @@ int fnic_device_reset(struct scsi_cmnd *sc) > wait_for_completion_timeout(&tm_done, > msecs_to_jiffies(FNIC_LUN_RESET_TIMEOUT)); > > + /* > + * Wake up can be due to the following reasons: > + * 1) The device reset completed from target. > + * 2) Device reset timed out. > + * 3) A link-down/host_reset may have happened in between. > + * 4) The device reset was aborted and io_req->dr_done was called. > + */ > + > + exit_dr = 0; > + spin_lock_irqsave(&fnic->fnic_lock, flags); > + if ((old_link_down_cnt != fnic->link_down_cnt) || > + (fnic->reset_in_progress) || > + (fnic->soft_reset_count != old_soft_reset_count) || > + (iport->state != FNIC_IPORT_STATE_READY)) > + exit_dr = 1; > + > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > + > spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > io_req = fnic_priv(sc)->io_req; > if (!io_req) { > @@ -2398,6 +2672,13 @@ int fnic_device_reset(struct scsi_cmnd *sc) > "io_req is null mqtag 0x%x sc 0x%p\n", mqtag, sc); > goto fnic_device_reset_end; > } > + > + if (exit_dr) { > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > + "Host reset called for fnic. Exit device reset\n"); > + io_req->dr_done = NULL; > + goto fnic_device_reset_clean; > + } > io_req->dr_done = NULL; > > status = fnic_priv(sc)->lr_status; > @@ -2411,50 +2692,8 @@ int fnic_device_reset(struct scsi_cmnd *sc) > FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > "Device reset timed out\n"); > fnic_priv(sc)->flags |= FNIC_DEV_RST_TIMED_OUT; > - spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > int_to_scsilun(sc->device->lun, &fc_lun); > - /* > - * Issue abort and terminate on device reset request. > - * If q'ing of terminate fails, retry it after a delay. > - */ > - while (1) { > - spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > - if (fnic_priv(sc)->flags & FNIC_DEV_RST_TERM_ISSUED) { > - spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > - break; > - } > - spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > - if (fnic_queue_abort_io_req(fnic, > - mqtag | FNIC_TAG_DEV_RST, > - FCPIO_ITMF_ABT_TASK_TERM, > - fc_lun.scsi_lun, io_req, hwq)) { > - wait_for_completion_timeout(&tm_done, > - msecs_to_jiffies(FNIC_ABT_TERM_DELAY_TIMEOUT)); > - } else { > - spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > - fnic_priv(sc)->flags |= FNIC_DEV_RST_TERM_ISSUED; > - fnic_priv(sc)->state = FNIC_IOREQ_ABTS_PENDING; > - io_req->abts_done = &tm_done; > - spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > - "Abort and terminate issued on Device reset mqtag 0x%x sc 0x%p\n", > - mqtag, sc); > - break; > - } > - } > - while (1) { > - spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > - if (!(fnic_priv(sc)->flags & FNIC_DEV_RST_DONE)) { > - spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > - wait_for_completion_timeout(&tm_done, > - msecs_to_jiffies(FNIC_LUN_RESET_TIMEOUT)); > - break; > - } else { > - io_req = fnic_priv(sc)->io_req; > - io_req->abts_done = NULL; > - goto fnic_device_reset_clean; > - } > - } > + goto fnic_device_reset_clean; > } else { > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > } > @@ -2480,8 +2719,7 @@ int fnic_device_reset(struct scsi_cmnd *sc) > spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > io_req = fnic_priv(sc)->io_req; > FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > - "Device reset failed" > - " since could not abort all IOs\n"); > + "Device reset failed: Cannot abort all IOs\n"); > goto fnic_device_reset_clean; > } > > @@ -2507,6 +2745,15 @@ int fnic_device_reset(struct scsi_cmnd *sc) > mempool_free(io_req, fnic->io_req_pool); > } > > + /* > + * If link-event is seen while LUN reset is issued we need > + * to complete the LUN reset here > + */ > + if (!new_sc) { > + sc->result = DID_RESET << 16; > + scsi_done(sc); > + } > + > fnic_device_reset_end: > FNIC_TRACE(fnic_device_reset, sc->device->host->host_no, rq->tag, sc, > jiffies_to_msecs(jiffies - start_time), > @@ -2520,6 +2767,17 @@ int fnic_device_reset(struct scsi_cmnd *sc) > mutex_unlock(&fnic->sgreset_mutex); > } > > + while ((ret == SUCCESS) && fnic_count_lun_ioreqs(fnic, sc->device)) { > + if (count >= 2) { > + ret = FAILED; > + break; > + } > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > + "Cannot clean up all IOs for the LUN\n"); > + schedule_timeout(msecs_to_jiffies(1000)); > + count++; > + } > + > FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > "Returning from device reset %s\n", > (ret == SUCCESS) ? > @@ -2531,50 +2789,54 @@ int fnic_device_reset(struct scsi_cmnd *sc) > return ret; > } > > -/* Clean up all IOs, clean up libFC local port */ > -int fnic_reset(struct Scsi_Host *shost) > +static void fnic_post_flogo_linkflap(struct fnic *fnic) > +{ > + unsigned long flags; > + > + fnic_fdls_link_status_change(fnic, 0); > + spin_lock_irqsave(&fnic->fnic_lock, flags); > + > + if (fnic->link_status) { > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > + fnic_fdls_link_status_change(fnic, 1); > + return; > + } > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > +} > + > +/* Logout from all the targets and simulate link flap */ > +void fnic_reset(struct Scsi_Host *shost) > { > - struct fc_lport *lp; > struct fnic *fnic; > - int ret = 0; > struct reset_stats *reset_stats; > > - lp = shost_priv(shost); > - fnic = lport_priv(lp); > + fnic = *((struct fnic **) shost_priv(shost)); > reset_stats = &fnic->fnic_stats.reset_stats; > > FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, > - "Issuing fnic reset\n"); > + "Issuing fnic reset\n"); > > atomic64_inc(&reset_stats->fnic_resets); > - > - /* > - * Reset local port, this will clean up libFC exchanges, > - * reset remote port sessions, and if link is up, begin flogi > - */ > - ret = fc_lport_reset(lp); > + fnic_post_flogo_linkflap(fnic); > > FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, > - "Returning from fnic reset with: %s\n", > - (ret == 0) ? "SUCCESS" : "FAILED"); > + "Returning from fnic reset"); > > - if (ret == 0) > - atomic64_inc(&reset_stats->fnic_reset_completions); > - else > - atomic64_inc(&reset_stats->fnic_reset_failures); > + atomic64_inc(&reset_stats->fnic_reset_completions); > +} > + > +int fnic_issue_fc_host_lip(struct Scsi_Host *shost) > +{ > + int ret = 0; > + struct fnic *fnic = *((struct fnic **) shost_priv(shost)); > + > + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, > + "FC host lip issued"); > > + ret = fnic_host_reset(shost); > return ret; > } > > -/* > - * SCSI Error handling calls driver's eh_host_reset if all prior > - * error handling levels return FAILED. If host reset completes > - * successfully, and if link is up, then Fabric login begins. > - * > - * Host Reset is the highest level of error recovery. If this fails, then > - * host is offlined by SCSI. > - * > - */ > int fnic_host_reset(struct Scsi_Host *shost) > { > int ret = SUCCESS; > @@ -2637,122 +2899,6 @@ int fnic_host_reset(struct Scsi_Host *shost) > return ret; > } > > -/* > - * This fxn is called from libFC when host is removed > - */ > -void fnic_scsi_abort_io(struct fc_lport *lp) > -{ > - int err = 0; > - unsigned long flags; > - enum fnic_state old_state; > - struct fnic *fnic = lport_priv(lp); > - DECLARE_COMPLETION_ONSTACK(remove_wait); > - > - /* Issue firmware reset for fnic, wait for reset to complete */ > -retry_fw_reset: > - spin_lock_irqsave(&fnic->fnic_lock, flags); > - if (unlikely(fnic->state == FNIC_IN_FC_TRANS_ETH_MODE) && > - fnic->link_events) { > - /* fw reset is in progress, poll for its completion */ > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > - schedule_timeout(msecs_to_jiffies(100)); > - goto retry_fw_reset; > - } > - > - fnic->remove_wait = &remove_wait; > - old_state = fnic->state; > - fnic->state = FNIC_IN_FC_TRANS_ETH_MODE; > - fnic_update_mac_locked(fnic, fnic->ctlr.ctl_src_addr); > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > - > - err = fnic_fw_reset_handler(fnic); > - if (err) { > - spin_lock_irqsave(&fnic->fnic_lock, flags); > - if (fnic->state == FNIC_IN_FC_TRANS_ETH_MODE) > - fnic->state = old_state; > - fnic->remove_wait = NULL; > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > - return; > - } > - > - /* Wait for firmware reset to complete */ > - wait_for_completion_timeout(&remove_wait, > - msecs_to_jiffies(FNIC_RMDEVICE_TIMEOUT)); > - > - spin_lock_irqsave(&fnic->fnic_lock, flags); > - fnic->remove_wait = NULL; > - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > - "fnic_scsi_abort_io %s\n", > - (fnic->state == FNIC_IN_ETH_MODE) ? > - "SUCCESS" : "FAILED"); > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > - > -} > - > -/* > - * This fxn called from libFC to clean up driver IO state on link down > - */ > -void fnic_scsi_cleanup(struct fc_lport *lp) > -{ > - unsigned long flags; > - enum fnic_state old_state; > - struct fnic *fnic = lport_priv(lp); > - > - /* issue fw reset */ > -retry_fw_reset: > - spin_lock_irqsave(&fnic->fnic_lock, flags); > - if (unlikely(fnic->state == FNIC_IN_FC_TRANS_ETH_MODE)) { > - /* fw reset is in progress, poll for its completion */ > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > - schedule_timeout(msecs_to_jiffies(100)); > - goto retry_fw_reset; > - } > - old_state = fnic->state; > - fnic->state = FNIC_IN_FC_TRANS_ETH_MODE; > - fnic_update_mac_locked(fnic, fnic->ctlr.ctl_src_addr); > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > - > - if (fnic_fw_reset_handler(fnic)) { > - spin_lock_irqsave(&fnic->fnic_lock, flags); > - if (fnic->state == FNIC_IN_FC_TRANS_ETH_MODE) > - fnic->state = old_state; > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > - } > - > -} > - > -void fnic_empty_scsi_cleanup(struct fc_lport *lp) > -{ > -} > - > -void fnic_exch_mgr_reset(struct fc_lport *lp, u32 sid, u32 did) > -{ > - struct fnic *fnic = lport_priv(lp); > - > - /* Non-zero sid, nothing to do */ > - if (sid) > - goto call_fc_exch_mgr_reset; > - > - if (did) { > - fnic_rport_exch_reset(fnic, did); > - goto call_fc_exch_mgr_reset; > - } > - > - /* > - * sid = 0, did = 0 > - * link down or device being removed > - */ > - if (!fnic->in_remove) > - fnic_scsi_cleanup(lp); > - else > - fnic_scsi_abort_io(lp); > - > - /* call libFC exch mgr reset to reset its exchanges */ > -call_fc_exch_mgr_reset: > - fc_exch_mgr_reset(lp, sid, did); > - > -} > - > static bool fnic_abts_pending_iter(struct scsi_cmnd *sc, void *data) > { > struct request *const rq = scsi_cmd_to_rq(sc); > diff --git a/drivers/scsi/fnic/fnic_stats.h b/drivers/scsi/fnic/fnic_stats.h > index 9d7f98c452dd..1f1a1ec90a23 100644 > --- a/drivers/scsi/fnic/fnic_stats.h > +++ b/drivers/scsi/fnic/fnic_stats.h > @@ -127,6 +127,4 @@ struct stats_debug_info { > }; > > int fnic_get_stats_data(struct stats_debug_info *, struct fnic_stats *); > -void fnic_stats_debugfs_init(struct fnic *); > -void fnic_stats_debugfs_remove(struct fnic *); > #endif /* _FNIC_STATS_H_ */ Cheers, Hannes
On Thursday, October 24, 2024 12:05 AM, Hannes Reinecke <hare@suse.de> wrote: > > On 10/18/24 18:14, Karan Tilak Kumar wrote: > > Modify IO path to use FDLS. > > Add helper functions to process IOs. > > Remove unused template functions. > > Cleanup obsolete code. > > Refactor old function definitions. > > > > Reviewed-by: Sesidhar Baddela <sebaddel@cisco.com> > > Reviewed-by: Arulprabhu Ponnusamy <arulponn@cisco.com> > > Reviewed-by: Gian Carlo Boffa <gcboffa@cisco.com> > > Reviewed-by: Arun Easi <aeasi@cisco.com> > > Signed-off-by: Karan Tilak Kumar <kartilak@cisco.com> > > --- > > Changes between v2 and v3: > > Replace fnic->host with fnic->lport->host to prevent compilation > > errors. > > > > Changes between v1 and v2: > > Replace pr_err with printk. > > Incorporate review comments by Hannes: > > Restore usage of tagset iterators. > > --- > > drivers/scsi/fnic/fnic.h | 20 +- > > drivers/scsi/fnic/fnic_io.h | 3 + > > drivers/scsi/fnic/fnic_main.c | 5 +- > > drivers/scsi/fnic/fnic_scsi.c | 846 +++++++++++++++++++-------------- > > drivers/scsi/fnic/fnic_stats.h | 2 - > > 5 files changed, 513 insertions(+), 363 deletions(-) > > > > diff --git a/drivers/scsi/fnic/fnic.h b/drivers/scsi/fnic/fnic.h > > index 4f38cbae1994..c6fe9eec9a0c 100644 > > --- a/drivers/scsi/fnic/fnic.h > > +++ b/drivers/scsi/fnic/fnic.h > > @@ -40,6 +40,7 @@ > > #define FNIC_DFLT_IO_REQ 256 /* Default scsi_cmnd tag map entries */ > > #define FNIC_DFLT_QUEUE_DEPTH 256 > > #define FNIC_STATS_RATE_LIMIT 4 /* limit rate at which stats are pulled up */ > > +#define LUN0_DELAY_TIME 9 > > > > /* > > * Tag bits used for special requests. > > @@ -472,7 +473,6 @@ int fnic_set_intr_mode_msix(struct fnic *fnic); > > void fnic_free_intr(struct fnic *fnic); > > int fnic_request_intr(struct fnic *fnic); > > > > -int fnic_send(struct fc_lport *, struct fc_frame *); > > void fnic_free_wq_buf(struct vnic_wq *wq, struct vnic_wq_buf *buf); > > void fnic_handle_frame(struct work_struct *work); > > void fnic_tport_event_handler(struct work_struct *work); > > @@ -489,11 +489,9 @@ int fnic_abort_cmd(struct scsi_cmnd *); > > int fnic_device_reset(struct scsi_cmnd *); > > int fnic_eh_host_reset_handler(struct scsi_cmnd *sc); > > int fnic_host_reset(struct Scsi_Host *shost); > > -int fnic_reset(struct Scsi_Host *); > > -void fnic_scsi_cleanup(struct fc_lport *); > > -void fnic_scsi_abort_io(struct fc_lport *); > > -void fnic_empty_scsi_cleanup(struct fc_lport *); > > -void fnic_exch_mgr_reset(struct fc_lport *, u32, u32); > > +void fnic_reset(struct Scsi_Host *shost); > > +int fnic_issue_fc_host_lip(struct Scsi_Host *shost); > > +void fnic_scsi_fcpio_reset(struct fnic *fnic); > > int fnic_wq_copy_cmpl_handler(struct fnic *fnic, int copy_work_to_do, unsigned int cq_index); > > int fnic_wq_cmpl_handler(struct fnic *fnic, int); > > int fnic_flogi_reg_handler(struct fnic *fnic, u32); > > @@ -505,7 +503,8 @@ const char *fnic_state_to_str(unsigned int state); > > void fnic_mq_map_queues_cpus(struct Scsi_Host *host); > > void fnic_log_q_error(struct fnic *fnic); > > void fnic_handle_link_event(struct fnic *fnic); > > - > > +void fnic_stats_debugfs_init(struct fnic *fnic); > > +void fnic_stats_debugfs_remove(struct fnic *fnic); > > int fnic_is_abts_pending(struct fnic *, struct scsi_cmnd *); > > > > void fnic_handle_fip_frame(struct work_struct *work); > > @@ -526,5 +525,12 @@ int fnic_get_desc_by_devid(struct pci_dev *pdev, char **desc, > > void fnic_fdls_link_status_change(struct fnic *fnic, int linkup); > > void fnic_delete_fcp_tports(struct fnic *fnic); > > void fnic_flush_tport_event_list(struct fnic *fnic); > > +int fnic_count_ioreqs_wq(struct fnic *fnic, u32 hwq, u32 portid); > > +unsigned int fnic_count_ioreqs(struct fnic *fnic, u32 portid); > > +unsigned int fnic_count_all_ioreqs(struct fnic *fnic); > > +unsigned int fnic_count_lun_ioreqs_wq(struct fnic *fnic, u32 hwq, > > + struct scsi_device *device); > > +unsigned int fnic_count_lun_ioreqs(struct fnic *fnic, > > + struct scsi_device *device); > > > > #endif /* _FNIC_H_ */ > > diff --git a/drivers/scsi/fnic/fnic_io.h b/drivers/scsi/fnic/fnic_io.h > > index 6fe642cb387b..0d974e040ab7 100644 > > --- a/drivers/scsi/fnic/fnic_io.h > > +++ b/drivers/scsi/fnic/fnic_io.h > > @@ -7,6 +7,7 @@ > > #define _FNIC_IO_H_ > > > > #include <scsi/fc/fc_fcp.h> > > +#include "fnic_fdls.h" > > > > #define FNIC_DFLT_SG_DESC_CNT 32 > > #define FNIC_MAX_SG_DESC_CNT 256 /* Maximum descriptors per sgl */ > > @@ -41,6 +42,8 @@ enum fnic_ioreq_state { > > }; > > > > struct fnic_io_req { > > + struct fnic_iport_s *iport; > > + struct fnic_tport_s *tport; > > struct host_sg_desc *sgl_list; /* sgl list */ > > void *sgl_list_alloc; /* sgl list address used for free */ > > dma_addr_t sense_buf_pa; /* dma address for sense buffer*/ > > diff --git a/drivers/scsi/fnic/fnic_main.c b/drivers/scsi/fnic/fnic_main.c > > index 33c63d7a9787..dadbe47c0bbb 100644 > > --- a/drivers/scsi/fnic/fnic_main.c > > +++ b/drivers/scsi/fnic/fnic_main.c > > @@ -85,9 +85,6 @@ module_param(fnic_max_qdepth, uint, S_IRUGO|S_IWUSR); > > MODULE_PARM_DESC(fnic_max_qdepth, "Queue depth to report for each LUN"); > > > > static struct libfc_function_template fnic_transport_template = { > > - .fcp_abort_io = fnic_empty_scsi_cleanup, > > - .fcp_cleanup = fnic_empty_scsi_cleanup, > > - .exch_mgr_reset = fnic_exch_mgr_reset > > }; > > > The template is now empty; can't you delete it completely? A patch later on in the series does code cleanup. Sure, I can move this to an earlier patch. > > struct workqueue_struct *fnic_fip_queue; > > @@ -162,7 +159,7 @@ static struct fc_function_template fnic_fc_functions = { > > .show_starget_port_id = 1, > > .show_rport_dev_loss_tmo = 1, > > .set_rport_dev_loss_tmo = fnic_set_rport_dev_loss_tmo, > > - .issue_fc_host_lip = fnic_reset, > > + .issue_fc_host_lip = fnic_issue_fc_host_lip, > > .get_fc_host_stats = fnic_get_stats, > > .reset_fc_host_stats = fnic_reset_host_stats, > > .dd_fcrport_size = sizeof(struct fc_rport_libfc_priv), > > diff --git a/drivers/scsi/fnic/fnic_scsi.c b/drivers/scsi/fnic/fnic_scsi.c > > index 74298f9a34e5..af729637ef2d 100644 > > --- a/drivers/scsi/fnic/fnic_scsi.c > > +++ b/drivers/scsi/fnic/fnic_scsi.c > > @@ -25,9 +25,12 @@ > > #include <scsi/fc/fc_fcoe.h> > > #include <scsi/libfc.h> > > #include <scsi/fc_frame.h> > > +#include <scsi/scsi_transport_fc.h> > > #include "fnic_io.h" > > #include "fnic.h" > > > > +static void fnic_cleanup_io(struct fnic *fnic, int exclude_id); > > + > > const char *fnic_state_str[] = { > > [FNIC_IN_FC_MODE] = "FNIC_IN_FC_MODE", > > [FNIC_IN_FC_TRANS_ETH_MODE] = "FNIC_IN_FC_TRANS_ETH_MODE", > > @@ -65,6 +68,18 @@ static const char *fcpio_status_str[] = { > > [FCPIO_LUNMAP_CHNG_PEND] = "FCPIO_LUNHMAP_CHNG_PEND", > > }; > > > > +enum terminate_io_return { > > + TERM_SUCCESS = 0, > > + TERM_NO_SC = 1, > > + TERM_IO_REQ_NOT_FOUND, > > + TERM_ANOTHER_PORT, > > + TERM_GSTATE, > > + TERM_IO_BLOCKED, > > + TERM_OUT_OF_WQ_DESC, > > + TERM_TIMED_OUT, > > + TERM_MISC, > > +}; > > + > > const char *fnic_state_to_str(unsigned int state) > > { > > if (state >= ARRAY_SIZE(fnic_state_str) || !fnic_state_str[state]) > > @@ -90,8 +105,6 @@ static const char *fnic_fcpio_status_to_str(unsigned int status) > > return fcpio_status_str[status]; > > } > > > > -static void fnic_cleanup_io(struct fnic *fnic); > > - > > /* > > * Unmap the data buffer and sense buffer for an io_req, > > * also unmap and free the device-private scatter/gather list. > > @@ -114,6 +127,80 @@ static void fnic_release_ioreq_buf(struct fnic *fnic, > > SCSI_SENSE_BUFFERSIZE, DMA_FROM_DEVICE); > > } > > > > +int fnic_count_ioreqs_wq(struct fnic *fnic, u32 hwq, u32 portid) > > +{ > > + unsigned long flags = 0; > > + int i = 0, count = 0; > > + > > + spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > > + for (i = 0; i != fnic->sw_copy_wq[hwq].ioreq_table_size; ++i) { > > + if (fnic->sw_copy_wq[hwq].io_req_table[i] != NULL && > > + (!portid > > + || fnic->sw_copy_wq[hwq].io_req_table[i]->port_id == portid)) > > + count++; > > + } > > + spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > + > > + return count; > > +} > > + > > +unsigned int fnic_count_ioreqs(struct fnic *fnic, u32 portid) > > +{ > > + int i; > > + unsigned int count = 0; > > + > > + for (i = 0; i < fnic->wq_copy_count; i++) > > + count += fnic_count_ioreqs_wq(fnic, i, portid); > > + > > + return count; > > +} > > + > > +unsigned int fnic_count_all_ioreqs(struct fnic *fnic) > > +{ > > + return fnic_count_ioreqs(fnic, 0); > > +} > > + > > +unsigned int fnic_count_lun_ioreqs_wq(struct fnic *fnic, u32 hwq, > > + struct scsi_device *device) > > +{ > > + struct fnic_io_req *io_req; > > + int i; > > + unsigned int count = 0; > > + unsigned long flags = 0; > > + > > + spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > > + for (i = 0; i != fnic->sw_copy_wq[hwq].ioreq_table_size; ++i) { > > + io_req = fnic->sw_copy_wq[hwq].io_req_table[i]; > > + > > + if (io_req != NULL) { > > + struct scsi_cmnd *sc = > > + scsi_host_find_tag(fnic->lport->host, io_req->tag); > > + > Be careful. 'scsi_host_find_tag()' finds requests on _all_ hwqs. > But here you are interested in tags for a specific hwq only. Thanks Hannes. We'll look at this closer. > > + if (!sc) > > + continue; > > + > > + if (sc->device == device) > > + count++; > > + } > > + } > > + spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > + > > + return count; > > +} > > + > And now this is pretty pointless. Please use the block layer > blk_mq_tagset_busy_iter() to iterate over busy/active tags. Sure Hannes. I'll fix it in the next version. > > +unsigned int fnic_count_lun_ioreqs(struct fnic *fnic, > > + struct scsi_device *device) > > +{ > > + int hwq; > > + unsigned int count = 0; > > + > > + /*count if any pending IOs on this lun */ > > + for (hwq = 0; hwq < fnic->wq_copy_count; hwq++) > > + count += fnic_count_lun_ioreqs_wq(fnic, hwq, device); > > + > > + return count; > > +} > > + > Same here. Sure Hannes. I'll fix it in the next version. > > /* Free up Copy Wq descriptors. Called with copy_wq lock held */ > > static int free_wq_copy_descs(struct fnic *fnic, struct vnic_wq_copy *wq, unsigned int hwq) > > { > > @@ -179,12 +266,11 @@ int fnic_fw_reset_handler(struct fnic *fnic) > > struct vnic_wq_copy *wq = &fnic->hw_copy_wq[0]; > > int ret = 0; > > unsigned long flags; > > + unsigned int ioreq_count; > > > > /* indicate fwreset to io path */ > > fnic_set_state_flags(fnic, FNIC_FLAGS_FWRESET); > > - > > - fnic_free_txq(&fnic->frame_queue); > > - fnic_free_txq(&fnic->tx_queue); > > + ioreq_count = fnic_count_all_ioreqs(fnic); > > > > /* wait for io cmpl */ > > while (atomic_read(&fnic->in_flight)) > > @@ -231,10 +317,10 @@ int fnic_flogi_reg_handler(struct fnic *fnic, u32 fc_id) > > { > > struct vnic_wq_copy *wq = &fnic->hw_copy_wq[0]; > > enum fcpio_flogi_reg_format_type format; > > - struct fc_lport *lp = fnic->lport; > > u8 gw_mac[ETH_ALEN]; > > int ret = 0; > > unsigned long flags; > > + struct fnic_iport_s *iport = &fnic->iport; > > > > spin_lock_irqsave(&fnic->wq_copy_lock[0], flags); > > > > @@ -246,28 +332,23 @@ int fnic_flogi_reg_handler(struct fnic *fnic, u32 fc_id) > > goto flogi_reg_ioreq_end; > > } > > > > - if (fnic->ctlr.map_dest) { > > - eth_broadcast_addr(gw_mac); > > - format = FCPIO_FLOGI_REG_DEF_DEST; > > - } else { > > - memcpy(gw_mac, fnic->ctlr.dest_addr, ETH_ALEN); > > - format = FCPIO_FLOGI_REG_GW_DEST; > > - } > > + memcpy(gw_mac, fnic->iport.fcfmac, ETH_ALEN); > > + format = FCPIO_FLOGI_REG_GW_DEST; > > > > - if ((fnic->config.flags & VFCF_FIP_CAPABLE) && !fnic->ctlr.map_dest) { > > + if (fnic->config.flags & VFCF_FIP_CAPABLE) { > > fnic_queue_wq_copy_desc_fip_reg(wq, SCSI_NO_TAG, > > fc_id, gw_mac, > > - fnic->data_src_addr, > > - lp->r_a_tov, lp->e_d_tov); > > - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > - "FLOGI FIP reg issued fcid %x src %pM dest %pM\n", > > - fc_id, fnic->data_src_addr, gw_mac); > > + fnic->iport.fpma, > > + iport->r_a_tov, iport->e_d_tov); > > + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, > > + "FLOGI FIP reg issued fcid: 0x%x src %p dest %p\n", > > + fc_id, fnic->iport.fpma, gw_mac); > > } else { > > fnic_queue_wq_copy_desc_flogi_reg(wq, SCSI_NO_TAG, > > format, fc_id, gw_mac); > > FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, > > - "FLOGI reg issued fcid 0x%x map %d dest 0x%p\n", > > - fc_id, fnic->ctlr.map_dest, gw_mac); > > + "FLOGI reg issued fcid 0x%x dest %p\n", > > + fc_id, gw_mac); > > } > > > > atomic64_inc(&fnic->fnic_stats.fw_stats.active_fw_reqs); > > @@ -295,13 +376,17 @@ static inline int fnic_queue_wq_copy_desc(struct fnic *fnic, > > { > > struct scatterlist *sg; > > struct fc_rport *rport = starget_to_rport(scsi_target(sc->device)); > > - struct fc_rport_libfc_priv *rp = rport->dd_data; > > struct host_sg_desc *desc; > > struct misc_stats *misc_stats = &fnic->fnic_stats.misc_stats; > > unsigned int i; > > int flags; > > u8 exch_flags; > > struct scsi_lun fc_lun; > > + struct fnic_tport_s *tport; > > + struct rport_dd_data_s *rdd_data; > > + > > + rdd_data = rport->dd_data; > > + tport = rdd_data->tport; > > > > if (sg_count) { > > /* For each SGE, create a device desc entry */ > > @@ -356,7 +441,7 @@ static inline int fnic_queue_wq_copy_desc(struct fnic *fnic, > > > > exch_flags = 0; > > if ((fnic->config.flags & VFCF_FCP_SEQ_LVL_ERR) && > > - (rp->flags & FC_RP_FLAGS_RETRY)) > > + (tport->tgt_flags & FDLS_FC_RP_FLAGS_RETRY)) > > exch_flags |= FCPIO_ICMND_SRFLAG_RETRY; > > > > fnic_queue_wq_copy_desc_icmnd_16(wq, mqtag, > > @@ -371,8 +456,8 @@ static inline int fnic_queue_wq_copy_desc(struct fnic *fnic, > > sc->cmnd, sc->cmd_len, > > scsi_bufflen(sc), > > fc_lun.scsi_lun, io_req->port_id, > > - rport->maxframe_size, rp->r_a_tov, > > - rp->e_d_tov); > > + tport->max_payload_size, > > + tport->r_a_tov, tport->e_d_tov); > > > > atomic64_inc(&fnic->fnic_stats.fw_stats.active_fw_reqs); > > if (atomic64_read(&fnic->fnic_stats.fw_stats.active_fw_reqs) > > > @@ -388,10 +473,10 @@ int fnic_queuecommand(struct Scsi_Host *shost, struct scsi_cmnd *sc) > > struct request *const rq = scsi_cmd_to_rq(sc); > > uint32_t mqtag = 0; > > void (*done)(struct scsi_cmnd *) = scsi_done; > > - struct fc_lport *lp = shost_priv(sc->device->host); > > struct fc_rport *rport; > > struct fnic_io_req *io_req = NULL; > > - struct fnic *fnic = lport_priv(lp); > > + struct fnic *fnic = *((struct fnic **) shost_priv(sc->device->host)); > > + struct fnic_iport_s *iport = NULL; > > struct fnic_stats *fnic_stats = &fnic->fnic_stats; > > struct vnic_wq_copy *wq; > > int ret = 1; > > @@ -400,32 +485,14 @@ int fnic_queuecommand(struct Scsi_Host *shost, struct scsi_cmnd *sc) > > unsigned long flags = 0; > > unsigned long ptr; > > int io_lock_acquired = 0; > > - struct fc_rport_libfc_priv *rp; > > uint16_t hwq = 0; > > - > > - mqtag = blk_mq_unique_tag(rq); > > - spin_lock_irqsave(&fnic->fnic_lock, flags); > > - > > - if (unlikely(fnic_chk_state_flags_locked(fnic, FNIC_FLAGS_IO_BLOCKED))) { > > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > - FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > > - "fnic IO blocked flags: 0x%lx. Returning SCSI_MLQUEUE_HOST_BUSY\n", > > - fnic->state_flags); > > - return SCSI_MLQUEUE_HOST_BUSY; > > - } > > - > > - if (unlikely(fnic_chk_state_flags_locked(fnic, FNIC_FLAGS_FWRESET))) { > > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > - FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > > - "fnic flags: 0x%lx. Returning SCSI_MLQUEUE_HOST_BUSY\n", > > - fnic->state_flags); > > - return SCSI_MLQUEUE_HOST_BUSY; > > - } > > + struct fnic_tport_s *tport = NULL; > > + struct rport_dd_data_s *rdd_data; > > + uint16_t lun0_delay = 0; > > > > rport = starget_to_rport(scsi_target(sc->device)); > > if (!rport) { > > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > > "returning DID_NO_CONNECT for IO as rport is NULL\n"); > > sc->result = DID_NO_CONNECT << 16; > > done(sc); > > @@ -434,50 +501,95 @@ int fnic_queuecommand(struct Scsi_Host *shost, struct scsi_cmnd *sc) > > > > ret = fc_remote_port_chkready(rport); > > if (ret) { > > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > > "rport is not ready\n"); > > - atomic64_inc(&fnic_stats->misc_stats.rport_not_ready); > > sc->result = ret; > > done(sc); > > return 0; > > } > > > > - rp = rport->dd_data; > > - if (!rp || rp->rp_state == RPORT_ST_DELETE) { > > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > - "rport 0x%x removed, returning DID_NO_CONNECT\n", > > - rport->port_id); > > + mqtag = blk_mq_unique_tag(rq); > > + spin_lock_irqsave(&fnic->fnic_lock, flags); > > + iport = &fnic->iport; > > > > - atomic64_inc(&fnic_stats->misc_stats.rport_not_ready); > > - sc->result = DID_NO_CONNECT<<16; > > + if (iport->state != FNIC_IPORT_STATE_READY) { > > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > > + "returning DID_NO_CONNECT for IO as iport state: %d\n", > > + iport->state); > > + sc->result = DID_NO_CONNECT << 16; > > done(sc); > > return 0; > > } > > > > - if (rp->rp_state != RPORT_ST_READY) { > > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > - "rport 0x%x in state 0x%x, returning DID_IMM_RETRY\n", > > - rport->port_id, rp->rp_state); > > + /* fc_remote_port_add() may have added the tport to > > + * fc_transport but dd_data not yet set > > + */ > > + rdd_data = rport->dd_data; > > + tport = rdd_data->tport; > > + if (!tport || (rdd_data->iport != iport)) { > > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > > + "dd_data not yet set in SCSI for rport portid: 0x%x\n", > > + rport->port_id); > > + tport = fnic_find_tport_by_fcid(iport, rport->port_id); > > + if (!tport) { > > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > > + "returning DID_BUS_BUSY for IO as tport not found for: 0x%x\n", > > + rport->port_id); > > + sc->result = DID_BUS_BUSY << 16; > > + done(sc); > > + return 0; > > + } > > + > > + /* Re-assign same params as in fnic_fdls_add_tport */ > > + rport->maxframe_size = FNIC_FC_MAX_PAYLOAD_LEN; > > + rport->supported_classes = > > + FC_COS_CLASS3 | FC_RPORT_ROLE_FCP_TARGET; > > + /* the dd_data is allocated by fctransport of size dd_fcrport_size */ > > + rdd_data = rport->dd_data; > > + rdd_data->tport = tport; > > + rdd_data->iport = iport; > > + tport->rport = rport; > > + tport->flags |= FNIC_FDLS_SCSI_REGISTERED; > > + } > > > > - sc->result = DID_IMM_RETRY << 16; > > + if ((tport->state != FDLS_TGT_STATE_READY) > > + && (tport->state != FDLS_TGT_STATE_ADISC)) { > > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > > + "returning DID_NO_CONNECT for IO as tport state: %d\n", > > + tport->state); > > + sc->result = DID_NO_CONNECT << 16; > > done(sc); > > return 0; > > } > > > > - if (lp->state != LPORT_ST_READY || !(lp->link_up)) { > > + atomic_inc(&fnic->in_flight); > > + atomic_inc(&tport->in_flight); > > + > > + if (unlikely(fnic_chk_state_flags_locked(fnic, FNIC_FLAGS_IO_BLOCKED))) { > > + atomic_dec(&fnic->in_flight); > > + atomic_dec(&tport->in_flight); > > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > + return SCSI_MLQUEUE_HOST_BUSY; > > + } > > + > > + if (unlikely(fnic_chk_state_flags_locked(fnic, FNIC_FLAGS_FWRESET))) { > > spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > > - "state not ready: %d/link not up: %d Returning HOST_BUSY\n", > > - lp->state, lp->link_up); > > + "fnic flags FW reset: 0x%lx. Returning SCSI_MLQUEUE_HOST_BUSY\n", > > + fnic->state_flags); > > return SCSI_MLQUEUE_HOST_BUSY; > > } > > > > - atomic_inc(&fnic->in_flight); > > + if (!tport->lun0_delay) { > > + lun0_delay = 1; > > + tport->lun0_delay++; > > + } > > > > spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > + > > fnic_priv(sc)->state = FNIC_IOREQ_NOT_INITED; > > fnic_priv(sc)->flags = FNIC_NO_FLAGS; > > > > @@ -499,6 +611,7 @@ int fnic_queuecommand(struct Scsi_Host *shost, struct scsi_cmnd *sc) > > goto out; > > } > > > > + io_req->tport = tport; > > /* Determine the type of scatter/gather list we need */ > > io_req->sgl_cnt = sg_count; > > io_req->sgl_type = FNIC_SGL_CACHE_DFLT; > > @@ -575,6 +688,7 @@ int fnic_queuecommand(struct Scsi_Host *shost, struct scsi_cmnd *sc) > > mempool_free(io_req, fnic->io_req_pool); > > } > > atomic_dec(&fnic->in_flight); > > + atomic_dec(&tport->in_flight); > > return ret; > > } else { > > atomic64_inc(&fnic_stats->io_stats.active_ios); > > @@ -602,6 +716,14 @@ int fnic_queuecommand(struct Scsi_Host *shost, struct scsi_cmnd *sc) > > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > > > atomic_dec(&fnic->in_flight); > > + atomic_dec(&tport->in_flight); > > + > > + if (lun0_delay) { > > + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, > > + "LUN0 delay\n"); > > + mdelay(LUN0_DELAY_TIME); > > + } > > + > > return ret; > > } > > > > @@ -625,7 +747,7 @@ static int fnic_fcpio_fw_reset_cmpl_handler(struct fnic *fnic, > > atomic64_inc(&reset_stats->fw_reset_completions); > > > > /* Clean up all outstanding io requests */ > > - fnic_cleanup_io(fnic); > > + fnic_cleanup_io(fnic, SCSI_NO_TAG); > > > > atomic64_set(&fnic->fnic_stats.fw_stats.active_fw_reqs, 0); > > atomic64_set(&fnic->fnic_stats.io_stats.active_ios, 0); > > @@ -646,12 +768,6 @@ static int fnic_fcpio_fw_reset_cmpl_handler(struct fnic *fnic, > > "reset failed with header status: %s\n", > > fnic_fcpio_status_to_str(hdr_status)); > > > > - /* > > - * Unable to change to eth mode, cannot send out flogi > > - * Change state to fc mode, so that subsequent Flogi > > - * requests from libFC will cause more attempts to > > - * reset the firmware. Free the cached flogi > > - */ > > fnic->state = FNIC_IN_FC_MODE; > > atomic64_inc(&reset_stats->fw_reset_failures); > > ret = -1; > > @@ -664,15 +780,14 @@ static int fnic_fcpio_fw_reset_cmpl_handler(struct fnic *fnic, > > ret = -1; > > } > > > > - /* Thread removing device blocks till firmware reset is complete */ > > - if (fnic->remove_wait) > > - complete(fnic->remove_wait); > > + if (fnic->fw_reset_done) > > + complete(fnic->fw_reset_done); > > > > /* > > * If fnic is being removed, or fw reset failed > > * free the flogi frame. Else, send it out > > */ > > - if (fnic->remove_wait || ret) { > > + if (ret) { > > spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > fnic_free_txq(&fnic->tx_queue); > > goto reset_cmpl_handler_end; > > @@ -711,12 +826,12 @@ static int fnic_fcpio_flogi_reg_cmpl_handler(struct fnic *fnic, > > /* Check flogi registration completion status */ > > if (!hdr_status) { > > FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > - "flog reg succeeded\n"); > > + "FLOGI reg succeeded\n"); > > fnic->state = FNIC_IN_FC_MODE; > > } else { > > FNIC_SCSI_DBG(KERN_DEBUG, > > fnic->lport->host, fnic->fnic_num, > > - "fnic flogi reg :failed %s\n", > > + "fnic flogi reg failed: %s\n", > > fnic_fcpio_status_to_str(hdr_status)); > > fnic->state = FNIC_IN_ETH_MODE; > > ret = -1; > > @@ -1023,15 +1138,6 @@ static void fnic_fcpio_icmnd_cmpl_handler(struct fnic *fnic, unsigned int cq_ind > > jiffies_to_msecs(jiffies - start_time)), > > desc, cmd_trace, fnic_flags_and_state(sc)); > > > > - if (sc->sc_data_direction == DMA_FROM_DEVICE) { > > - fnic->lport->host_stats.fcp_input_requests++; > > - fnic->fcp_input_bytes += xfer_len; > > - } else if (sc->sc_data_direction == DMA_TO_DEVICE) { > > - fnic->lport->host_stats.fcp_output_requests++; > > - fnic->fcp_output_bytes += xfer_len; > > - } else > > - fnic->lport->host_stats.fcp_control_requests++; > > - > > /* Call SCSI completion function to complete the IO */ > > scsi_done(sc); > > > > @@ -1414,8 +1520,8 @@ static bool fnic_cleanup_io_iter(struct scsi_cmnd *sc, void *data) > > struct request *const rq = scsi_cmd_to_rq(sc); > > struct fnic *fnic = data; > > struct fnic_io_req *io_req; > > - unsigned long flags = 0; > > unsigned long start_time = 0; > > + unsigned long flags; > > struct fnic_stats *fnic_stats = &fnic->fnic_stats; > > uint16_t hwq = 0; > > int tag; > > @@ -1439,7 +1545,7 @@ static bool fnic_cleanup_io_iter(struct scsi_cmnd *sc, void *data) > > } > > > > if ((fnic_priv(sc)->flags & FNIC_DEVICE_RESET) && > > - !(fnic_priv(sc)->flags & FNIC_DEV_RST_DONE)) { > > + !(fnic_priv(sc)->flags & FNIC_DEV_RST_DONE)) { > > /* > > * We will be here only when FW completes reset > > * without sending completions for outstanding ios. > > @@ -1449,6 +1555,7 @@ static bool fnic_cleanup_io_iter(struct scsi_cmnd *sc, void *data) > > complete(io_req->dr_done); > > else if (io_req && io_req->abts_done) > > complete(io_req->abts_done); > > + > > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > return true; > > } else if (fnic_priv(sc)->flags & FNIC_DEVICE_RESET) { > > @@ -1458,19 +1565,19 @@ static bool fnic_cleanup_io_iter(struct scsi_cmnd *sc, void *data) > > > > fnic_priv(sc)->io_req = NULL; > > io_req->sc = NULL; > > + start_time = io_req->start_time; > > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > > > /* > > * If there is a scsi_cmnd associated with this io_req, then > > * free the corresponding state > > */ > > - start_time = io_req->start_time; > > fnic_release_ioreq_buf(fnic, io_req, sc); > > mempool_free(io_req, fnic->io_req_pool); > > > > sc->result = DID_TRANSPORT_DISRUPTED << 16; > > FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > > - "mqtag:0x%x tag: 0x%x sc:0x%p duration = %lu DID_TRANSPORT_DISRUPTED\n", > > + "mqtag: 0x%x tag: 0x%x sc: 0x%p duration = %lu DID_TRANSPORT_DISRUPTED\n", > > mqtag, tag, sc, (jiffies - start_time)); > > > > if (atomic64_read(&fnic->io_cmpl_skip)) > > @@ -1479,23 +1586,60 @@ static bool fnic_cleanup_io_iter(struct scsi_cmnd *sc, void *data) > > atomic64_inc(&fnic_stats->io_stats.io_completions); > > > > FNIC_TRACE(fnic_cleanup_io, > > - sc->device->host->host_no, tag, sc, > > - jiffies_to_msecs(jiffies - start_time), > > - 0, ((u64)sc->cmnd[0] << 32 | > > - (u64)sc->cmnd[2] << 24 | > > - (u64)sc->cmnd[3] << 16 | > > - (u64)sc->cmnd[4] << 8 | sc->cmnd[5]), > > - fnic_flags_and_state(sc)); > > - > > + sc->device->host->host_no, tag, sc, > > + jiffies_to_msecs(jiffies - start_time), > > + 0, ((u64) sc->cmnd[0] << 32 | > > + (u64) sc->cmnd[2] << 24 | > > + (u64) sc->cmnd[3] << 16 | > > + (u64) sc->cmnd[4] << 8 | sc->cmnd[5]), > > + (((u64) fnic_priv(sc)->flags << 32) | fnic_priv(sc)-> > > + state)); > > + > > + /* Complete the command to SCSI */ > > scsi_done(sc); > > - > > return true; > > } > > > > -static void fnic_cleanup_io(struct fnic *fnic) > > +static void fnic_cleanup_io(struct fnic *fnic, int exclude_id) > > { > > + unsigned int io_count = 0; > > + unsigned long flags; > > + struct fnic_io_req *io_req = NULL; > > + struct scsi_cmnd *sc = NULL; > > + > > + io_count = fnic_count_all_ioreqs(fnic); > > + FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > + "Outstanding ioreq count: %d active io count: %lld Waiting\n", > > + io_count, > > + atomic64_read(&fnic->fnic_stats.io_stats.active_ios)); > > + > > scsi_host_busy_iter(fnic->lport->host, > > - fnic_cleanup_io_iter, fnic); > > + fnic_cleanup_io_iter, fnic); > > + > > + /* with sg3utils device reset, SC needs to be retrieved from ioreq */ > > + spin_lock_irqsave(&fnic->wq_copy_lock[0], flags); > > + io_req = fnic->sw_copy_wq[0].io_req_table[fnic->fnic_max_tag_id]; > > + if (io_req) { > > + sc = io_req->sc; > > + if (sc) { > > + if ((fnic_priv(sc)->flags & FNIC_DEVICE_RESET) > > + && !(fnic_priv(sc)->flags & FNIC_DEV_RST_DONE)) { > > + fnic_priv(sc)->flags |= FNIC_DEV_RST_DONE; > > + if (io_req && io_req->dr_done) > > + complete(io_req->dr_done); > > + } > > + } > > + } > > + spin_unlock_irqrestore(&fnic->wq_copy_lock[0], flags); > > + > > + while ((io_count = fnic_count_all_ioreqs(fnic))) { > > + FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > + "Outstanding ioreq count: %d active io count: %lld Waiting\n", > > + io_count, > > + atomic64_read(&fnic->fnic_stats.io_stats.active_ios)); > > + > > + schedule_timeout(msecs_to_jiffies(100)); > > + } > > } > > > > void fnic_wq_copy_cleanup_handler(struct vnic_wq_copy *wq, > > @@ -1567,10 +1711,13 @@ static inline int fnic_queue_abort_io_req(struct fnic *fnic, int tag, > > struct vnic_wq_copy *wq = &fnic->hw_copy_wq[hwq]; > > struct misc_stats *misc_stats = &fnic->fnic_stats.misc_stats; > > unsigned long flags; > > + struct fnic_tport_s *tport = io_req->tport; > > > > spin_lock_irqsave(&fnic->fnic_lock, flags); > > if (unlikely(fnic_chk_state_flags_locked(fnic, > > FNIC_FLAGS_IO_BLOCKED))) { > > + atomic_dec(&fnic->in_flight); > > + atomic_dec(&tport->in_flight); > > spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > return 1; > > } else > > @@ -1585,6 +1732,7 @@ static inline int fnic_queue_abort_io_req(struct fnic *fnic, int tag, > > if (!vnic_wq_copy_desc_avail(wq)) { > > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > atomic_dec(&fnic->in_flight); > > + atomic_dec(&tport->in_flight); > > FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > "fnic_queue_abort_io_req: failure: no descriptors\n"); > > atomic64_inc(&misc_stats->abts_cpwq_alloc_failures); > > @@ -1619,20 +1767,24 @@ static bool fnic_rport_abort_io_iter(struct scsi_cmnd *sc, void *data) > > struct fnic *fnic = iter_data->fnic; > > int abt_tag = 0; > > struct fnic_io_req *io_req; > > - unsigned long flags; > > struct reset_stats *reset_stats = &fnic->fnic_stats.reset_stats; > > struct terminate_stats *term_stats = &fnic->fnic_stats.term_stats; > > struct scsi_lun fc_lun; > > enum fnic_ioreq_state old_ioreq_state; > > uint16_t hwq = 0; > > + unsigned long flags; > > > > abt_tag = blk_mq_unique_tag(rq); > > hwq = blk_mq_unique_tag_to_hwq(abt_tag); > > > > - spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > > + if (!sc) { > > + FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > + "sc is NULL abt_tag: 0x%x hwq: %d\n", abt_tag, hwq); > > + return true; > > + } > > > > + spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > > io_req = fnic_priv(sc)->io_req; > > - > > if (!io_req || io_req->port_id != iter_data->port_id) { > > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > return true; > > @@ -1655,37 +1807,38 @@ static bool fnic_rport_abort_io_iter(struct scsi_cmnd *sc, void *data) > > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > return true; > > } > > + > > if (io_req->abts_done) { > > shost_printk(KERN_ERR, fnic->lport->host, > > - "fnic_rport_exch_reset: io_req->abts_done is set " > > - "state is %s\n", > > + "fnic_rport_exch_reset: io_req->abts_done is set state is %s\n", > > fnic_ioreq_state_to_str(fnic_priv(sc)->state)); > > } > > > > if (!(fnic_priv(sc)->flags & FNIC_IO_ISSUED)) { > > shost_printk(KERN_ERR, fnic->lport->host, > > - "rport_exch_reset " > > - "IO not yet issued %p tag 0x%x flags " > > - "%x state %d\n", > > - sc, abt_tag, fnic_priv(sc)->flags, fnic_priv(sc)->state); > > + "rport_exch_reset IO not yet issued %p abt_tag 0x%x", > > + sc, abt_tag); > > + shost_printk(KERN_ERR, fnic->lport->host, > > + "flags %x state %d\n", fnic_priv(sc)->flags, > > + fnic_priv(sc)->state); > > } > > old_ioreq_state = fnic_priv(sc)->state; > > fnic_priv(sc)->state = FNIC_IOREQ_ABTS_PENDING; > > fnic_priv(sc)->abts_status = FCPIO_INVALID_CODE; > > + > > if (fnic_priv(sc)->flags & FNIC_DEVICE_RESET) { > > atomic64_inc(&reset_stats->device_reset_terminates); > > abt_tag |= FNIC_TAG_DEV_RST; > > } > > FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > - "fnic_rport_exch_reset dev rst sc 0x%p\n", sc); > > - BUG_ON(io_req->abts_done); > > - > > + "fnic_rport_exch_reset: dev rst sc 0x%p\n", sc); > > + WARN_ON_ONCE(io_req->abts_done); > > FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > "fnic_rport_reset_exch: Issuing abts\n"); > > > > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > > > - /* Now queue the abort command to firmware */ > > + /* Queue the abort command to firmware */ > > int_to_scsilun(sc->device->lun, &fc_lun); > > > > if (fnic_queue_abort_io_req(fnic, abt_tag, > > @@ -1714,11 +1867,14 @@ static bool fnic_rport_abort_io_iter(struct scsi_cmnd *sc, void *data) > > atomic64_inc(&term_stats->terminates); > > iter_data->term_cnt++; > > } > > + > > return true; > > } > > > > void fnic_rport_exch_reset(struct fnic *fnic, u32 port_id) > > { > > + unsigned int io_count = 0; > > + unsigned long flags; > > struct terminate_stats *term_stats = &fnic->fnic_stats.term_stats; > > struct fnic_rport_abort_io_iter_data iter_data = { > > .fnic = fnic, > > @@ -1726,53 +1882,75 @@ void fnic_rport_exch_reset(struct fnic *fnic, u32 port_id) > > .term_cnt = 0, > > }; > > > > - FNIC_SCSI_DBG(KERN_DEBUG, > > - fnic->lport->host, fnic->fnic_num, > > - "fnic_rport_exch_reset called portid 0x%06x\n", > > - port_id); > > + FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > + "fnic rport exchange reset for tport: 0x%06x\n", > > + port_id); > > > > if (fnic->in_remove) > > return; > > > > + io_count = fnic_count_ioreqs(fnic, port_id); > > + FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > + "Starting terminates: rport:0x%x portid-io-count: %d active-io-count: %lld\n", > > + port_id, io_count, > > + atomic64_read(&fnic->fnic_stats.io_stats.active_ios)); > > + > > + spin_lock_irqsave(&fnic->fnic_lock, flags); > > + atomic_inc(&fnic->in_flight); > > + if (unlikely(fnic_chk_state_flags_locked(fnic, FNIC_FLAGS_IO_BLOCKED))) { > > + atomic_dec(&fnic->in_flight); > > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > + return; > > + } > > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > + > > scsi_host_busy_iter(fnic->lport->host, fnic_rport_abort_io_iter, > > &iter_data); > > + > > if (iter_data.term_cnt > atomic64_read(&term_stats->max_terminates)) > > atomic64_set(&term_stats->max_terminates, iter_data.term_cnt); > > > > + atomic_dec(&fnic->in_flight); > > + > > + while ((io_count = fnic_count_ioreqs(fnic, port_id))) > > + schedule_timeout(msecs_to_jiffies(1000)); > > + > > That really feels overly complex. > You send the aborts with the previous 'scsi_host_busy_iter()' call, and > now you wait for the I/Os to complete. > But you already have an atomic counter for I/Os in flight to the tport; > can't you just wait for that counter to drop to zero? Thanks for your comment Hannes. To clarify, the in_flight variable is to track IOs which have not left fnic yet. It is _not_ to track IOs which are on the wire to the tport. I can add a comment regarding its use, if that's acceptable. > > + FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > + "rport: 0x%x remaining portid-io-count: %d ", > > + port_id, io_count); > > } > > > > void fnic_terminate_rport_io(struct fc_rport *rport) > > { > > - struct fc_rport_libfc_priv *rdata; > > - struct fc_lport *lport; > > - struct fnic *fnic; > > + struct fnic_tport_s *tport; > > + struct rport_dd_data_s *rdd_data; > > + struct fnic_iport_s *iport = NULL; > > + struct fnic *fnic = NULL; > > > > if (!rport) { > > - printk(KERN_ERR "fnic_terminate_rport_io: rport is NULL\n"); > > + pr_err("rport is NULL\n"); > > return; > > } > > - rdata = rport->dd_data; > > > > - if (!rdata) { > > - printk(KERN_ERR "fnic_terminate_rport_io: rdata is NULL\n"); > > - return; > > - } > > - lport = rdata->local_port; > > - > > - if (!lport) { > > - printk(KERN_ERR "fnic_terminate_rport_io: lport is NULL\n"); > > - return; > > + rdd_data = rport->dd_data; > > + if (rdd_data) { > > + tport = rdd_data->tport; > > + if (!tport) { > > + pr_err( > > + "term rport io called after tport is deleted. Returning 0x%8x\n", > > + rport->port_id); > > + } else { > > + pr_err( > > + "term rport io called after tport is set 0x%8x\n", > > + rport->port_id); > > + pr_err( > > + "tport maybe rediscovered\n"); > > + > > + iport = (struct fnic_iport_s *) tport->iport; > > + fnic = iport->fnic; > > + fnic_rport_exch_reset(fnic, rport->port_id); > > + } > > } > > - fnic = lport_priv(lport); > > - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > - "wwpn 0x%llx, wwnn0x%llx, rport 0x%p, portid 0x%06x\n", > > - rport->port_name, rport->node_name, rport, > > - rport->port_id); > > - > > - if (fnic->in_remove) > > - return; > > - > > - fnic_rport_exch_reset(fnic, rport->port_id); > > } > > > > /* > > @@ -1783,10 +1961,12 @@ void fnic_terminate_rport_io(struct fc_rport *rport) > > int fnic_abort_cmd(struct scsi_cmnd *sc) > > { > > struct request *const rq = scsi_cmd_to_rq(sc); > > - struct fc_lport *lp; > > + struct fnic_iport_s *iport; > > + struct fnic_tport_s *tport; > > struct fnic *fnic; > > struct fnic_io_req *io_req = NULL; > > struct fc_rport *rport; > > + struct rport_dd_data_s *rdd_data; > > unsigned long flags; > > unsigned long start_time = 0; > > int ret = SUCCESS; > > @@ -1806,11 +1986,11 @@ int fnic_abort_cmd(struct scsi_cmnd *sc) > > fc_block_scsi_eh(sc); > > > > /* Get local-port, check ready and link up */ > > - lp = shost_priv(sc->device->host); > > - > > - fnic = lport_priv(lp); > > + fnic = *((struct fnic **) shost_priv(sc->device->host)); > > > > spin_lock_irqsave(&fnic->fnic_lock, flags); > > + iport = &fnic->iport; > > + > > fnic_stats = &fnic->fnic_stats; > > abts_stats = &fnic->fnic_stats.abts_stats; > > term_stats = &fnic->fnic_stats.term_stats; > > @@ -1821,7 +2001,43 @@ int fnic_abort_cmd(struct scsi_cmnd *sc) > > > > fnic_priv(sc)->flags = FNIC_NO_FLAGS; > > > > - if (lp->state != LPORT_ST_READY || !(lp->link_up)) { > > + rdd_data = rport->dd_data; > > + tport = rdd_data->tport; > > + > > + if (!tport) { > > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > > + "Abort cmd called after tport delete! rport fcid: 0x%x", > > + rport->port_id); > > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > > + "lun: %llu hwq: 0x%x mqtag: 0x%x Op: 0x%x flags: 0x%x\n", > > + sc->device->lun, hwq, mqtag, > > + sc->cmnd[0], fnic_priv(sc)->flags); > > + ret = FAILED; > > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > + goto fnic_abort_cmd_end; > > + } > > + > > + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, > > + "Abort cmd called rport fcid: 0x%x lun: %llu hwq: 0x%x mqtag: 0x%x", > > + rport->port_id, sc->device->lun, hwq, mqtag); > > + > > + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, > > + "Op: 0x%x flags: 0x%x\n", > > + sc->cmnd[0], > > + fnic_priv(sc)->flags); > > + > > + if (iport->state != FNIC_IPORT_STATE_READY) { > > + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, > > + "iport NOT in READY state"); > > + ret = FAILED; > > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > + goto fnic_abort_cmd_end; > > + } > > + > > + if ((tport->state != FDLS_TGT_STATE_READY) && > > + (tport->state != FDLS_TGT_STATE_ADISC)) { > > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > > + "tport state: %d\n", tport->state); > > ret = FAILED; > > spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > goto fnic_abort_cmd_end; > > @@ -1843,6 +2059,7 @@ int fnic_abort_cmd(struct scsi_cmnd *sc) > > spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > > io_req = fnic_priv(sc)->io_req; > > if (!io_req) { > > + ret = FAILED; > > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > goto fnic_abort_cmd_end; > > } > > @@ -1893,7 +2110,6 @@ int fnic_abort_cmd(struct scsi_cmnd *sc) > > if (fc_remote_port_chkready(rport) == 0) > > task_req = FCPIO_ITMF_ABT_TASK; > > else { > > - atomic64_inc(&fnic_stats->misc_stats.rport_not_ready); > > task_req = FCPIO_ITMF_ABT_TASK_TERM; > > } > > > > @@ -2027,6 +2243,7 @@ static inline int fnic_queue_dr_io_req(struct fnic *fnic, > > unsigned long flags; > > uint16_t hwq = 0; > > uint32_t tag = 0; > > + struct fnic_tport_s *tport = io_req->tport; > > > > tag = io_req->tag; > > hwq = blk_mq_unique_tag_to_hwq(tag); > > @@ -2037,8 +2254,10 @@ static inline int fnic_queue_dr_io_req(struct fnic *fnic, > > FNIC_FLAGS_IO_BLOCKED))) { > > spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > return FAILED; > > - } else > > + } else { > > atomic_inc(&fnic->in_flight); > > + atomic_inc(&tport->in_flight); > > + } > > spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > > > spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > > @@ -2072,6 +2291,7 @@ static inline int fnic_queue_dr_io_req(struct fnic *fnic, > > lr_io_req_end: > > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > atomic_dec(&fnic->in_flight); > > + atomic_inc(&tport->in_flight); > > > > return ret; > > } > > @@ -2274,11 +2494,11 @@ static int fnic_clean_pending_aborts(struct fnic *fnic, > > int fnic_device_reset(struct scsi_cmnd *sc) > > { > > struct request *rq = scsi_cmd_to_rq(sc); > > - struct fc_lport *lp; > > struct fnic *fnic; > > struct fnic_io_req *io_req = NULL; > > struct fc_rport *rport; > > int status; > > + int count = 0; > > int ret = FAILED; > > unsigned long flags; > > unsigned long start_time = 0; > > @@ -2289,31 +2509,61 @@ int fnic_device_reset(struct scsi_cmnd *sc) > > DECLARE_COMPLETION_ONSTACK(tm_done); > > bool new_sc = 0; > > uint16_t hwq = 0; > > + struct fnic_iport_s *iport = NULL; > > + struct rport_dd_data_s *rdd_data; > > + struct fnic_tport_s *tport; > > + u32 old_soft_reset_count; > > + u32 old_link_down_cnt; > > + int exit_dr = 0; > > > > /* Wait for rport to unblock */ > > fc_block_scsi_eh(sc); > > > > /* Get local-port, check ready and link up */ > > - lp = shost_priv(sc->device->host); > > + fnic = *((struct fnic **) shost_priv(sc->device->host)); > > + iport = &fnic->iport; > > > > - fnic = lport_priv(lp); > > fnic_stats = &fnic->fnic_stats; > > reset_stats = &fnic->fnic_stats.reset_stats; > > > > atomic64_inc(&reset_stats->device_resets); > > > > rport = starget_to_rport(scsi_target(sc->device)); > > + > > + spin_lock_irqsave(&fnic->fnic_lock, flags); > > FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > - "fcid: 0x%x lun: 0x%llx hwq: %d mqtag: 0x%x flags: 0x%x Device reset\n", > > + "fcid: 0x%x lun: %llu hwq: %d mqtag: 0x%x flags: 0x%x Device reset\n", > > rport->port_id, sc->device->lun, hwq, mqtag, > > fnic_priv(sc)->flags); > > > > - if (lp->state != LPORT_ST_READY || !(lp->link_up)) > > + rdd_data = rport->dd_data; > > + tport = rdd_data->tport; > > + if (!tport) { > > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > > + "Dev rst called after tport delete! rport fcid: 0x%x lun: %llu\n", > > + rport->port_id, sc->device->lun); > > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > + goto fnic_device_reset_end; > > + } > > + > > + if (iport->state != FNIC_IPORT_STATE_READY) { > > + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, > > + "iport NOT in READY state"); > > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > goto fnic_device_reset_end; > > + } > > + > > + if ((tport->state != FDLS_TGT_STATE_READY) && > > + (tport->state != FDLS_TGT_STATE_ADISC)) { > > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > > + "tport state: %d\n", tport->state); > > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > + goto fnic_device_reset_end; > > + } > > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > > Please check if returning FAST_IO_FAIL here wouldn't be a better option. > Most of the time a device reset is triggered by a command timeout, which > typically happens due to a transport issue (eg link is down or something). > So returning 'FAILED' will just escalate to host reset, and that can > take for a really long time trying to abort all commands. Thanks for your insights Hannes. We're investing this currently. > > /* Check if remote port up */ > > if (fc_remote_port_chkready(rport)) { > > - atomic64_inc(&fnic_stats->misc_stats.rport_not_ready); > > goto fnic_device_reset_end; > > } > > > > @@ -2352,6 +2602,7 @@ int fnic_device_reset(struct scsi_cmnd *sc) > > io_req->port_id = rport->port_id; > > io_req->tag = mqtag; > > fnic_priv(sc)->io_req = io_req; > > + io_req->tport = tport; > > io_req->sc = sc; > > > > if (fnic->sw_copy_wq[hwq].io_req_table[blk_mq_unique_tag_to_tag(mqtag)] != NULL) > > @@ -2383,6 +2634,11 @@ int fnic_device_reset(struct scsi_cmnd *sc) > > fnic_priv(sc)->flags |= FNIC_DEV_RST_ISSUED; > > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > > > + spin_lock_irqsave(&fnic->fnic_lock, flags); > > + old_link_down_cnt = iport->fnic->link_down_cnt; > > + old_soft_reset_count = fnic->soft_reset_count; > > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > + > > /* > > * Wait on the local completion for LUN reset. The io_req may be > > * freed while we wait since we hold no lock. > > @@ -2390,6 +2646,24 @@ int fnic_device_reset(struct scsi_cmnd *sc) > > wait_for_completion_timeout(&tm_done, > > msecs_to_jiffies(FNIC_LUN_RESET_TIMEOUT)); > > > > + /* > > + * Wake up can be due to the following reasons: > > + * 1) The device reset completed from target. > > + * 2) Device reset timed out. > > + * 3) A link-down/host_reset may have happened in between. > > + * 4) The device reset was aborted and io_req->dr_done was called. > > + */ > > + > > + exit_dr = 0; > > + spin_lock_irqsave(&fnic->fnic_lock, flags); > > + if ((old_link_down_cnt != fnic->link_down_cnt) || > > + (fnic->reset_in_progress) || > > + (fnic->soft_reset_count != old_soft_reset_count) || > > + (iport->state != FNIC_IPORT_STATE_READY)) > > + exit_dr = 1; > > + > > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > + > > spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > > io_req = fnic_priv(sc)->io_req; > > if (!io_req) { > > @@ -2398,6 +2672,13 @@ int fnic_device_reset(struct scsi_cmnd *sc) > > "io_req is null mqtag 0x%x sc 0x%p\n", mqtag, sc); > > goto fnic_device_reset_end; > > } > > + > > + if (exit_dr) { > > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > > + "Host reset called for fnic. Exit device reset\n"); > > + io_req->dr_done = NULL; > > + goto fnic_device_reset_clean; > > + } > > io_req->dr_done = NULL; > > > > status = fnic_priv(sc)->lr_status; > > @@ -2411,50 +2692,8 @@ int fnic_device_reset(struct scsi_cmnd *sc) > > FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > "Device reset timed out\n"); > > fnic_priv(sc)->flags |= FNIC_DEV_RST_TIMED_OUT; > > - spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > int_to_scsilun(sc->device->lun, &fc_lun); > > - /* > > - * Issue abort and terminate on device reset request. > > - * If q'ing of terminate fails, retry it after a delay. > > - */ > > - while (1) { > > - spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > > - if (fnic_priv(sc)->flags & FNIC_DEV_RST_TERM_ISSUED) { > > - spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > - break; > > - } > > - spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > - if (fnic_queue_abort_io_req(fnic, > > - mqtag | FNIC_TAG_DEV_RST, > > - FCPIO_ITMF_ABT_TASK_TERM, > > - fc_lun.scsi_lun, io_req, hwq)) { > > - wait_for_completion_timeout(&tm_done, > > - msecs_to_jiffies(FNIC_ABT_TERM_DELAY_TIMEOUT)); > > - } else { > > - spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > > - fnic_priv(sc)->flags |= FNIC_DEV_RST_TERM_ISSUED; > > - fnic_priv(sc)->state = FNIC_IOREQ_ABTS_PENDING; > > - io_req->abts_done = &tm_done; > > - spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > - "Abort and terminate issued on Device reset mqtag 0x%x sc 0x%p\n", > > - mqtag, sc); > > - break; > > - } > > - } > > - while (1) { > > - spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > > - if (!(fnic_priv(sc)->flags & FNIC_DEV_RST_DONE)) { > > - spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > - wait_for_completion_timeout(&tm_done, > > - msecs_to_jiffies(FNIC_LUN_RESET_TIMEOUT)); > > - break; > > - } else { > > - io_req = fnic_priv(sc)->io_req; > > - io_req->abts_done = NULL; > > - goto fnic_device_reset_clean; > > - } > > - } > > + goto fnic_device_reset_clean; > > } else { > > spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); > > } > > @@ -2480,8 +2719,7 @@ int fnic_device_reset(struct scsi_cmnd *sc) > > spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); > > io_req = fnic_priv(sc)->io_req; > > FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > - "Device reset failed" > > - " since could not abort all IOs\n"); > > + "Device reset failed: Cannot abort all IOs\n"); > > goto fnic_device_reset_clean; > > } > > > > @@ -2507,6 +2745,15 @@ int fnic_device_reset(struct scsi_cmnd *sc) > > mempool_free(io_req, fnic->io_req_pool); > > } > > > > + /* > > + * If link-event is seen while LUN reset is issued we need > > + * to complete the LUN reset here > > + */ > > + if (!new_sc) { > > + sc->result = DID_RESET << 16; > > + scsi_done(sc); > > + } > > + > > fnic_device_reset_end: > > FNIC_TRACE(fnic_device_reset, sc->device->host->host_no, rq->tag, sc, > > jiffies_to_msecs(jiffies - start_time), > > @@ -2520,6 +2767,17 @@ int fnic_device_reset(struct scsi_cmnd *sc) > > mutex_unlock(&fnic->sgreset_mutex); > > } > > > > + while ((ret == SUCCESS) && fnic_count_lun_ioreqs(fnic, sc->device)) { > > + if (count >= 2) { > > + ret = FAILED; > > + break; > > + } > > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > > + "Cannot clean up all IOs for the LUN\n"); > > + schedule_timeout(msecs_to_jiffies(1000)); > > + count++; > > + } > > + > > FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > "Returning from device reset %s\n", > > (ret == SUCCESS) ? > > @@ -2531,50 +2789,54 @@ int fnic_device_reset(struct scsi_cmnd *sc) > > return ret; > > } > > > > -/* Clean up all IOs, clean up libFC local port */ > > -int fnic_reset(struct Scsi_Host *shost) > > +static void fnic_post_flogo_linkflap(struct fnic *fnic) > > +{ > > + unsigned long flags; > > + > > + fnic_fdls_link_status_change(fnic, 0); > > + spin_lock_irqsave(&fnic->fnic_lock, flags); > > + > > + if (fnic->link_status) { > > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > + fnic_fdls_link_status_change(fnic, 1); > > + return; > > + } > > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > +} > > + > > +/* Logout from all the targets and simulate link flap */ > > +void fnic_reset(struct Scsi_Host *shost) > > { > > - struct fc_lport *lp; > > struct fnic *fnic; > > - int ret = 0; > > struct reset_stats *reset_stats; > > > > - lp = shost_priv(shost); > > - fnic = lport_priv(lp); > > + fnic = *((struct fnic **) shost_priv(shost)); > > reset_stats = &fnic->fnic_stats.reset_stats; > > > > FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, > > - "Issuing fnic reset\n"); > > + "Issuing fnic reset\n"); > > > > atomic64_inc(&reset_stats->fnic_resets); > > - > > - /* > > - * Reset local port, this will clean up libFC exchanges, > > - * reset remote port sessions, and if link is up, begin flogi > > - */ > > - ret = fc_lport_reset(lp); > > + fnic_post_flogo_linkflap(fnic); > > > > FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, > > - "Returning from fnic reset with: %s\n", > > - (ret == 0) ? "SUCCESS" : "FAILED"); > > + "Returning from fnic reset"); > > > > - if (ret == 0) > > - atomic64_inc(&reset_stats->fnic_reset_completions); > > - else > > - atomic64_inc(&reset_stats->fnic_reset_failures); > > + atomic64_inc(&reset_stats->fnic_reset_completions); > > +} > > + > > +int fnic_issue_fc_host_lip(struct Scsi_Host *shost) > > +{ > > + int ret = 0; > > + struct fnic *fnic = *((struct fnic **) shost_priv(shost)); > > + > > + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, > > + "FC host lip issued"); > > > > + ret = fnic_host_reset(shost); > > return ret; > > } > > > > -/* > > - * SCSI Error handling calls driver's eh_host_reset if all prior > > - * error handling levels return FAILED. If host reset completes > > - * successfully, and if link is up, then Fabric login begins. > > - * > > - * Host Reset is the highest level of error recovery. If this fails, then > > - * host is offlined by SCSI. > > - * > > - */ > > int fnic_host_reset(struct Scsi_Host *shost) > > { > > int ret = SUCCESS; > > @@ -2637,122 +2899,6 @@ int fnic_host_reset(struct Scsi_Host *shost) > > return ret; > > } > > > > -/* > > - * This fxn is called from libFC when host is removed > > - */ > > -void fnic_scsi_abort_io(struct fc_lport *lp) > > -{ > > - int err = 0; > > - unsigned long flags; > > - enum fnic_state old_state; > > - struct fnic *fnic = lport_priv(lp); > > - DECLARE_COMPLETION_ONSTACK(remove_wait); > > - > > - /* Issue firmware reset for fnic, wait for reset to complete */ > > -retry_fw_reset: > > - spin_lock_irqsave(&fnic->fnic_lock, flags); > > - if (unlikely(fnic->state == FNIC_IN_FC_TRANS_ETH_MODE) && > > - fnic->link_events) { > > - /* fw reset is in progress, poll for its completion */ > > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > - schedule_timeout(msecs_to_jiffies(100)); > > - goto retry_fw_reset; > > - } > > - > > - fnic->remove_wait = &remove_wait; > > - old_state = fnic->state; > > - fnic->state = FNIC_IN_FC_TRANS_ETH_MODE; > > - fnic_update_mac_locked(fnic, fnic->ctlr.ctl_src_addr); > > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > - > > - err = fnic_fw_reset_handler(fnic); > > - if (err) { > > - spin_lock_irqsave(&fnic->fnic_lock, flags); > > - if (fnic->state == FNIC_IN_FC_TRANS_ETH_MODE) > > - fnic->state = old_state; > > - fnic->remove_wait = NULL; > > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > - return; > > - } > > - > > - /* Wait for firmware reset to complete */ > > - wait_for_completion_timeout(&remove_wait, > > - msecs_to_jiffies(FNIC_RMDEVICE_TIMEOUT)); > > - > > - spin_lock_irqsave(&fnic->fnic_lock, flags); > > - fnic->remove_wait = NULL; > > - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > - "fnic_scsi_abort_io %s\n", > > - (fnic->state == FNIC_IN_ETH_MODE) ? > > - "SUCCESS" : "FAILED"); > > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > - > > -} > > - > > -/* > > - * This fxn called from libFC to clean up driver IO state on link down > > - */ > > -void fnic_scsi_cleanup(struct fc_lport *lp) > > -{ > > - unsigned long flags; > > - enum fnic_state old_state; > > - struct fnic *fnic = lport_priv(lp); > > - > > - /* issue fw reset */ > > -retry_fw_reset: > > - spin_lock_irqsave(&fnic->fnic_lock, flags); > > - if (unlikely(fnic->state == FNIC_IN_FC_TRANS_ETH_MODE)) { > > - /* fw reset is in progress, poll for its completion */ > > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > - schedule_timeout(msecs_to_jiffies(100)); > > - goto retry_fw_reset; > > - } > > - old_state = fnic->state; > > - fnic->state = FNIC_IN_FC_TRANS_ETH_MODE; > > - fnic_update_mac_locked(fnic, fnic->ctlr.ctl_src_addr); > > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > - > > - if (fnic_fw_reset_handler(fnic)) { > > - spin_lock_irqsave(&fnic->fnic_lock, flags); > > - if (fnic->state == FNIC_IN_FC_TRANS_ETH_MODE) > > - fnic->state = old_state; > > - spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > - } > > - > > -} > > - > > -void fnic_empty_scsi_cleanup(struct fc_lport *lp) > > -{ > > -} > > - > > -void fnic_exch_mgr_reset(struct fc_lport *lp, u32 sid, u32 did) > > -{ > > - struct fnic *fnic = lport_priv(lp); > > - > > - /* Non-zero sid, nothing to do */ > > - if (sid) > > - goto call_fc_exch_mgr_reset; > > - > > - if (did) { > > - fnic_rport_exch_reset(fnic, did); > > - goto call_fc_exch_mgr_reset; > > - } > > - > > - /* > > - * sid = 0, did = 0 > > - * link down or device being removed > > - */ > > - if (!fnic->in_remove) > > - fnic_scsi_cleanup(lp); > > - else > > - fnic_scsi_abort_io(lp); > > - > > - /* call libFC exch mgr reset to reset its exchanges */ > > -call_fc_exch_mgr_reset: > > - fc_exch_mgr_reset(lp, sid, did); > > - > > -} > > - > > static bool fnic_abts_pending_iter(struct scsi_cmnd *sc, void *data) > > { > > struct request *const rq = scsi_cmd_to_rq(sc); > > diff --git a/drivers/scsi/fnic/fnic_stats.h b/drivers/scsi/fnic/fnic_stats.h > > index 9d7f98c452dd..1f1a1ec90a23 100644 > > --- a/drivers/scsi/fnic/fnic_stats.h > > +++ b/drivers/scsi/fnic/fnic_stats.h > > @@ -127,6 +127,4 @@ struct stats_debug_info { > > }; > > > > int fnic_get_stats_data(struct stats_debug_info *, struct fnic_stats *); > > -void fnic_stats_debugfs_init(struct fnic *); > > -void fnic_stats_debugfs_remove(struct fnic *); > > #endif /* _FNIC_STATS_H_ */ > > Cheers, > > Hannes > -- > Dr. Hannes Reinecke Kernel Storage Architect > hare@suse.de +49 911 74053 688 > SUSE Software Solutions GmbH, Frankenstr. 146, 90461 Nürnberg > HRB 36809 (AG Nürnberg), GF: I. Totev, A. McDonald, W. Knoblich Regards, Karan
On Monday, October 28, 2024 5:55 PM, Karan Tilak Kumar (kartilak) wrote: > > On Thursday, October 24, 2024 12:05 AM, Hannes Reinecke <hare@suse.de> wrote: > > > > On 10/18/24 18:14, Karan Tilak Kumar wrote: > > > Modify IO path to use FDLS. > > > Add helper functions to process IOs. > > > Remove unused template functions. > > > Cleanup obsolete code. > > > Refactor old function definitions. > > > > > > Reviewed-by: Sesidhar Baddela <sebaddel@cisco.com> > > > Reviewed-by: Arulprabhu Ponnusamy <arulponn@cisco.com> > > > Reviewed-by: Gian Carlo Boffa <gcboffa@cisco.com> > > > Reviewed-by: Arun Easi <aeasi@cisco.com> > > > Signed-off-by: Karan Tilak Kumar <kartilak@cisco.com> > > > @@ -2274,11 +2494,11 @@ static int fnic_clean_pending_aborts(struct fnic *fnic, > > > int fnic_device_reset(struct scsi_cmnd *sc) > > > { > > > struct request *rq = scsi_cmd_to_rq(sc); > > > - struct fc_lport *lp; > > > struct fnic *fnic; > > > struct fnic_io_req *io_req = NULL; > > > struct fc_rport *rport; > > > int status; > > > + int count = 0; > > > int ret = FAILED; > > > unsigned long flags; > > > unsigned long start_time = 0; > > > @@ -2289,31 +2509,61 @@ int fnic_device_reset(struct scsi_cmnd *sc) > > > DECLARE_COMPLETION_ONSTACK(tm_done); > > > bool new_sc = 0; > > > uint16_t hwq = 0; > > > + struct fnic_iport_s *iport = NULL; > > > + struct rport_dd_data_s *rdd_data; > > > + struct fnic_tport_s *tport; > > > + u32 old_soft_reset_count; > > > + u32 old_link_down_cnt; > > > + int exit_dr = 0; > > > > > > /* Wait for rport to unblock */ > > > fc_block_scsi_eh(sc); > > > > > > /* Get local-port, check ready and link up */ > > > - lp = shost_priv(sc->device->host); > > > + fnic = *((struct fnic **) shost_priv(sc->device->host)); > > > + iport = &fnic->iport; > > > > > > - fnic = lport_priv(lp); > > > fnic_stats = &fnic->fnic_stats; > > > reset_stats = &fnic->fnic_stats.reset_stats; > > > > > > atomic64_inc(&reset_stats->device_resets); > > > > > > rport = starget_to_rport(scsi_target(sc->device)); > > > + > > > + spin_lock_irqsave(&fnic->fnic_lock, flags); > > > FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, > > > - "fcid: 0x%x lun: 0x%llx hwq: %d mqtag: 0x%x flags: 0x%x Device reset\n", > > > + "fcid: 0x%x lun: %llu hwq: %d mqtag: 0x%x flags: 0x%x Device reset\n", > > > rport->port_id, sc->device->lun, hwq, mqtag, > > > fnic_priv(sc)->flags); > > > > > > - if (lp->state != LPORT_ST_READY || !(lp->link_up)) > > > + rdd_data = rport->dd_data; > > > + tport = rdd_data->tport; > > > + if (!tport) { > > > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > > > + "Dev rst called after tport delete! rport fcid: 0x%x lun: %llu\n", > > > + rport->port_id, sc->device->lun); > > > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > > + goto fnic_device_reset_end; > > > + } > > > + > > > + if (iport->state != FNIC_IPORT_STATE_READY) { > > > + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, > > > + "iport NOT in READY state"); > > > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > > goto fnic_device_reset_end; > > > + } > > > + > > > + if ((tport->state != FDLS_TGT_STATE_READY) && > > > + (tport->state != FDLS_TGT_STATE_ADISC)) { > > > + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, > > > + "tport state: %d\n", tport->state); > > > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > > + goto fnic_device_reset_end; > > > + } > > > + spin_unlock_irqrestore(&fnic->fnic_lock, flags); > > > > > Please check if returning FAST_IO_FAIL here wouldn't be a better option. > > Most of the time a device reset is triggered by a command timeout, which > > typically happens due to a transport issue (eg link is down or something). > > So returning 'FAILED' will just escalate to host reset, and that can > > take for a really long time trying to abort all commands. > > Thanks for your insights Hannes. > We're investing this currently. > Thanks again for this comment, Hannes. We reviewed this internally and as we understand it, returning a FAST_IO_FAIL status would complete the scsi_cmnd. This would cause issues if the IO is still active in firmware. We do take care of cleaning up such I/Os in the higher error escalation path (host reset) though, which would be induced by returning the 'FAILED' return status here. Regards, Karan
diff --git a/drivers/scsi/fnic/fnic.h b/drivers/scsi/fnic/fnic.h index 4f38cbae1994..c6fe9eec9a0c 100644 --- a/drivers/scsi/fnic/fnic.h +++ b/drivers/scsi/fnic/fnic.h @@ -40,6 +40,7 @@ #define FNIC_DFLT_IO_REQ 256 /* Default scsi_cmnd tag map entries */ #define FNIC_DFLT_QUEUE_DEPTH 256 #define FNIC_STATS_RATE_LIMIT 4 /* limit rate at which stats are pulled up */ +#define LUN0_DELAY_TIME 9 /* * Tag bits used for special requests. @@ -472,7 +473,6 @@ int fnic_set_intr_mode_msix(struct fnic *fnic); void fnic_free_intr(struct fnic *fnic); int fnic_request_intr(struct fnic *fnic); -int fnic_send(struct fc_lport *, struct fc_frame *); void fnic_free_wq_buf(struct vnic_wq *wq, struct vnic_wq_buf *buf); void fnic_handle_frame(struct work_struct *work); void fnic_tport_event_handler(struct work_struct *work); @@ -489,11 +489,9 @@ int fnic_abort_cmd(struct scsi_cmnd *); int fnic_device_reset(struct scsi_cmnd *); int fnic_eh_host_reset_handler(struct scsi_cmnd *sc); int fnic_host_reset(struct Scsi_Host *shost); -int fnic_reset(struct Scsi_Host *); -void fnic_scsi_cleanup(struct fc_lport *); -void fnic_scsi_abort_io(struct fc_lport *); -void fnic_empty_scsi_cleanup(struct fc_lport *); -void fnic_exch_mgr_reset(struct fc_lport *, u32, u32); +void fnic_reset(struct Scsi_Host *shost); +int fnic_issue_fc_host_lip(struct Scsi_Host *shost); +void fnic_scsi_fcpio_reset(struct fnic *fnic); int fnic_wq_copy_cmpl_handler(struct fnic *fnic, int copy_work_to_do, unsigned int cq_index); int fnic_wq_cmpl_handler(struct fnic *fnic, int); int fnic_flogi_reg_handler(struct fnic *fnic, u32); @@ -505,7 +503,8 @@ const char *fnic_state_to_str(unsigned int state); void fnic_mq_map_queues_cpus(struct Scsi_Host *host); void fnic_log_q_error(struct fnic *fnic); void fnic_handle_link_event(struct fnic *fnic); - +void fnic_stats_debugfs_init(struct fnic *fnic); +void fnic_stats_debugfs_remove(struct fnic *fnic); int fnic_is_abts_pending(struct fnic *, struct scsi_cmnd *); void fnic_handle_fip_frame(struct work_struct *work); @@ -526,5 +525,12 @@ int fnic_get_desc_by_devid(struct pci_dev *pdev, char **desc, void fnic_fdls_link_status_change(struct fnic *fnic, int linkup); void fnic_delete_fcp_tports(struct fnic *fnic); void fnic_flush_tport_event_list(struct fnic *fnic); +int fnic_count_ioreqs_wq(struct fnic *fnic, u32 hwq, u32 portid); +unsigned int fnic_count_ioreqs(struct fnic *fnic, u32 portid); +unsigned int fnic_count_all_ioreqs(struct fnic *fnic); +unsigned int fnic_count_lun_ioreqs_wq(struct fnic *fnic, u32 hwq, + struct scsi_device *device); +unsigned int fnic_count_lun_ioreqs(struct fnic *fnic, + struct scsi_device *device); #endif /* _FNIC_H_ */ diff --git a/drivers/scsi/fnic/fnic_io.h b/drivers/scsi/fnic/fnic_io.h index 6fe642cb387b..0d974e040ab7 100644 --- a/drivers/scsi/fnic/fnic_io.h +++ b/drivers/scsi/fnic/fnic_io.h @@ -7,6 +7,7 @@ #define _FNIC_IO_H_ #include <scsi/fc/fc_fcp.h> +#include "fnic_fdls.h" #define FNIC_DFLT_SG_DESC_CNT 32 #define FNIC_MAX_SG_DESC_CNT 256 /* Maximum descriptors per sgl */ @@ -41,6 +42,8 @@ enum fnic_ioreq_state { }; struct fnic_io_req { + struct fnic_iport_s *iport; + struct fnic_tport_s *tport; struct host_sg_desc *sgl_list; /* sgl list */ void *sgl_list_alloc; /* sgl list address used for free */ dma_addr_t sense_buf_pa; /* dma address for sense buffer*/ diff --git a/drivers/scsi/fnic/fnic_main.c b/drivers/scsi/fnic/fnic_main.c index 33c63d7a9787..dadbe47c0bbb 100644 --- a/drivers/scsi/fnic/fnic_main.c +++ b/drivers/scsi/fnic/fnic_main.c @@ -85,9 +85,6 @@ module_param(fnic_max_qdepth, uint, S_IRUGO|S_IWUSR); MODULE_PARM_DESC(fnic_max_qdepth, "Queue depth to report for each LUN"); static struct libfc_function_template fnic_transport_template = { - .fcp_abort_io = fnic_empty_scsi_cleanup, - .fcp_cleanup = fnic_empty_scsi_cleanup, - .exch_mgr_reset = fnic_exch_mgr_reset }; struct workqueue_struct *fnic_fip_queue; @@ -162,7 +159,7 @@ static struct fc_function_template fnic_fc_functions = { .show_starget_port_id = 1, .show_rport_dev_loss_tmo = 1, .set_rport_dev_loss_tmo = fnic_set_rport_dev_loss_tmo, - .issue_fc_host_lip = fnic_reset, + .issue_fc_host_lip = fnic_issue_fc_host_lip, .get_fc_host_stats = fnic_get_stats, .reset_fc_host_stats = fnic_reset_host_stats, .dd_fcrport_size = sizeof(struct fc_rport_libfc_priv), diff --git a/drivers/scsi/fnic/fnic_scsi.c b/drivers/scsi/fnic/fnic_scsi.c index 74298f9a34e5..af729637ef2d 100644 --- a/drivers/scsi/fnic/fnic_scsi.c +++ b/drivers/scsi/fnic/fnic_scsi.c @@ -25,9 +25,12 @@ #include <scsi/fc/fc_fcoe.h> #include <scsi/libfc.h> #include <scsi/fc_frame.h> +#include <scsi/scsi_transport_fc.h> #include "fnic_io.h" #include "fnic.h" +static void fnic_cleanup_io(struct fnic *fnic, int exclude_id); + const char *fnic_state_str[] = { [FNIC_IN_FC_MODE] = "FNIC_IN_FC_MODE", [FNIC_IN_FC_TRANS_ETH_MODE] = "FNIC_IN_FC_TRANS_ETH_MODE", @@ -65,6 +68,18 @@ static const char *fcpio_status_str[] = { [FCPIO_LUNMAP_CHNG_PEND] = "FCPIO_LUNHMAP_CHNG_PEND", }; +enum terminate_io_return { + TERM_SUCCESS = 0, + TERM_NO_SC = 1, + TERM_IO_REQ_NOT_FOUND, + TERM_ANOTHER_PORT, + TERM_GSTATE, + TERM_IO_BLOCKED, + TERM_OUT_OF_WQ_DESC, + TERM_TIMED_OUT, + TERM_MISC, +}; + const char *fnic_state_to_str(unsigned int state) { if (state >= ARRAY_SIZE(fnic_state_str) || !fnic_state_str[state]) @@ -90,8 +105,6 @@ static const char *fnic_fcpio_status_to_str(unsigned int status) return fcpio_status_str[status]; } -static void fnic_cleanup_io(struct fnic *fnic); - /* * Unmap the data buffer and sense buffer for an io_req, * also unmap and free the device-private scatter/gather list. @@ -114,6 +127,80 @@ static void fnic_release_ioreq_buf(struct fnic *fnic, SCSI_SENSE_BUFFERSIZE, DMA_FROM_DEVICE); } +int fnic_count_ioreqs_wq(struct fnic *fnic, u32 hwq, u32 portid) +{ + unsigned long flags = 0; + int i = 0, count = 0; + + spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); + for (i = 0; i != fnic->sw_copy_wq[hwq].ioreq_table_size; ++i) { + if (fnic->sw_copy_wq[hwq].io_req_table[i] != NULL && + (!portid + || fnic->sw_copy_wq[hwq].io_req_table[i]->port_id == portid)) + count++; + } + spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); + + return count; +} + +unsigned int fnic_count_ioreqs(struct fnic *fnic, u32 portid) +{ + int i; + unsigned int count = 0; + + for (i = 0; i < fnic->wq_copy_count; i++) + count += fnic_count_ioreqs_wq(fnic, i, portid); + + return count; +} + +unsigned int fnic_count_all_ioreqs(struct fnic *fnic) +{ + return fnic_count_ioreqs(fnic, 0); +} + +unsigned int fnic_count_lun_ioreqs_wq(struct fnic *fnic, u32 hwq, + struct scsi_device *device) +{ + struct fnic_io_req *io_req; + int i; + unsigned int count = 0; + unsigned long flags = 0; + + spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); + for (i = 0; i != fnic->sw_copy_wq[hwq].ioreq_table_size; ++i) { + io_req = fnic->sw_copy_wq[hwq].io_req_table[i]; + + if (io_req != NULL) { + struct scsi_cmnd *sc = + scsi_host_find_tag(fnic->lport->host, io_req->tag); + + if (!sc) + continue; + + if (sc->device == device) + count++; + } + } + spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); + + return count; +} + +unsigned int fnic_count_lun_ioreqs(struct fnic *fnic, + struct scsi_device *device) +{ + int hwq; + unsigned int count = 0; + + /*count if any pending IOs on this lun */ + for (hwq = 0; hwq < fnic->wq_copy_count; hwq++) + count += fnic_count_lun_ioreqs_wq(fnic, hwq, device); + + return count; +} + /* Free up Copy Wq descriptors. Called with copy_wq lock held */ static int free_wq_copy_descs(struct fnic *fnic, struct vnic_wq_copy *wq, unsigned int hwq) { @@ -179,12 +266,11 @@ int fnic_fw_reset_handler(struct fnic *fnic) struct vnic_wq_copy *wq = &fnic->hw_copy_wq[0]; int ret = 0; unsigned long flags; + unsigned int ioreq_count; /* indicate fwreset to io path */ fnic_set_state_flags(fnic, FNIC_FLAGS_FWRESET); - - fnic_free_txq(&fnic->frame_queue); - fnic_free_txq(&fnic->tx_queue); + ioreq_count = fnic_count_all_ioreqs(fnic); /* wait for io cmpl */ while (atomic_read(&fnic->in_flight)) @@ -231,10 +317,10 @@ int fnic_flogi_reg_handler(struct fnic *fnic, u32 fc_id) { struct vnic_wq_copy *wq = &fnic->hw_copy_wq[0]; enum fcpio_flogi_reg_format_type format; - struct fc_lport *lp = fnic->lport; u8 gw_mac[ETH_ALEN]; int ret = 0; unsigned long flags; + struct fnic_iport_s *iport = &fnic->iport; spin_lock_irqsave(&fnic->wq_copy_lock[0], flags); @@ -246,28 +332,23 @@ int fnic_flogi_reg_handler(struct fnic *fnic, u32 fc_id) goto flogi_reg_ioreq_end; } - if (fnic->ctlr.map_dest) { - eth_broadcast_addr(gw_mac); - format = FCPIO_FLOGI_REG_DEF_DEST; - } else { - memcpy(gw_mac, fnic->ctlr.dest_addr, ETH_ALEN); - format = FCPIO_FLOGI_REG_GW_DEST; - } + memcpy(gw_mac, fnic->iport.fcfmac, ETH_ALEN); + format = FCPIO_FLOGI_REG_GW_DEST; - if ((fnic->config.flags & VFCF_FIP_CAPABLE) && !fnic->ctlr.map_dest) { + if (fnic->config.flags & VFCF_FIP_CAPABLE) { fnic_queue_wq_copy_desc_fip_reg(wq, SCSI_NO_TAG, fc_id, gw_mac, - fnic->data_src_addr, - lp->r_a_tov, lp->e_d_tov); - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, - "FLOGI FIP reg issued fcid %x src %pM dest %pM\n", - fc_id, fnic->data_src_addr, gw_mac); + fnic->iport.fpma, + iport->r_a_tov, iport->e_d_tov); + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, + "FLOGI FIP reg issued fcid: 0x%x src %p dest %p\n", + fc_id, fnic->iport.fpma, gw_mac); } else { fnic_queue_wq_copy_desc_flogi_reg(wq, SCSI_NO_TAG, format, fc_id, gw_mac); FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, - "FLOGI reg issued fcid 0x%x map %d dest 0x%p\n", - fc_id, fnic->ctlr.map_dest, gw_mac); + "FLOGI reg issued fcid 0x%x dest %p\n", + fc_id, gw_mac); } atomic64_inc(&fnic->fnic_stats.fw_stats.active_fw_reqs); @@ -295,13 +376,17 @@ static inline int fnic_queue_wq_copy_desc(struct fnic *fnic, { struct scatterlist *sg; struct fc_rport *rport = starget_to_rport(scsi_target(sc->device)); - struct fc_rport_libfc_priv *rp = rport->dd_data; struct host_sg_desc *desc; struct misc_stats *misc_stats = &fnic->fnic_stats.misc_stats; unsigned int i; int flags; u8 exch_flags; struct scsi_lun fc_lun; + struct fnic_tport_s *tport; + struct rport_dd_data_s *rdd_data; + + rdd_data = rport->dd_data; + tport = rdd_data->tport; if (sg_count) { /* For each SGE, create a device desc entry */ @@ -356,7 +441,7 @@ static inline int fnic_queue_wq_copy_desc(struct fnic *fnic, exch_flags = 0; if ((fnic->config.flags & VFCF_FCP_SEQ_LVL_ERR) && - (rp->flags & FC_RP_FLAGS_RETRY)) + (tport->tgt_flags & FDLS_FC_RP_FLAGS_RETRY)) exch_flags |= FCPIO_ICMND_SRFLAG_RETRY; fnic_queue_wq_copy_desc_icmnd_16(wq, mqtag, @@ -371,8 +456,8 @@ static inline int fnic_queue_wq_copy_desc(struct fnic *fnic, sc->cmnd, sc->cmd_len, scsi_bufflen(sc), fc_lun.scsi_lun, io_req->port_id, - rport->maxframe_size, rp->r_a_tov, - rp->e_d_tov); + tport->max_payload_size, + tport->r_a_tov, tport->e_d_tov); atomic64_inc(&fnic->fnic_stats.fw_stats.active_fw_reqs); if (atomic64_read(&fnic->fnic_stats.fw_stats.active_fw_reqs) > @@ -388,10 +473,10 @@ int fnic_queuecommand(struct Scsi_Host *shost, struct scsi_cmnd *sc) struct request *const rq = scsi_cmd_to_rq(sc); uint32_t mqtag = 0; void (*done)(struct scsi_cmnd *) = scsi_done; - struct fc_lport *lp = shost_priv(sc->device->host); struct fc_rport *rport; struct fnic_io_req *io_req = NULL; - struct fnic *fnic = lport_priv(lp); + struct fnic *fnic = *((struct fnic **) shost_priv(sc->device->host)); + struct fnic_iport_s *iport = NULL; struct fnic_stats *fnic_stats = &fnic->fnic_stats; struct vnic_wq_copy *wq; int ret = 1; @@ -400,32 +485,14 @@ int fnic_queuecommand(struct Scsi_Host *shost, struct scsi_cmnd *sc) unsigned long flags = 0; unsigned long ptr; int io_lock_acquired = 0; - struct fc_rport_libfc_priv *rp; uint16_t hwq = 0; - - mqtag = blk_mq_unique_tag(rq); - spin_lock_irqsave(&fnic->fnic_lock, flags); - - if (unlikely(fnic_chk_state_flags_locked(fnic, FNIC_FLAGS_IO_BLOCKED))) { - spin_unlock_irqrestore(&fnic->fnic_lock, flags); - FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, - "fnic IO blocked flags: 0x%lx. Returning SCSI_MLQUEUE_HOST_BUSY\n", - fnic->state_flags); - return SCSI_MLQUEUE_HOST_BUSY; - } - - if (unlikely(fnic_chk_state_flags_locked(fnic, FNIC_FLAGS_FWRESET))) { - spin_unlock_irqrestore(&fnic->fnic_lock, flags); - FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, - "fnic flags: 0x%lx. Returning SCSI_MLQUEUE_HOST_BUSY\n", - fnic->state_flags); - return SCSI_MLQUEUE_HOST_BUSY; - } + struct fnic_tport_s *tport = NULL; + struct rport_dd_data_s *rdd_data; + uint16_t lun0_delay = 0; rport = starget_to_rport(scsi_target(sc->device)); if (!rport) { - spin_unlock_irqrestore(&fnic->fnic_lock, flags); - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, "returning DID_NO_CONNECT for IO as rport is NULL\n"); sc->result = DID_NO_CONNECT << 16; done(sc); @@ -434,50 +501,95 @@ int fnic_queuecommand(struct Scsi_Host *shost, struct scsi_cmnd *sc) ret = fc_remote_port_chkready(rport); if (ret) { - spin_unlock_irqrestore(&fnic->fnic_lock, flags); - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, "rport is not ready\n"); - atomic64_inc(&fnic_stats->misc_stats.rport_not_ready); sc->result = ret; done(sc); return 0; } - rp = rport->dd_data; - if (!rp || rp->rp_state == RPORT_ST_DELETE) { - spin_unlock_irqrestore(&fnic->fnic_lock, flags); - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, - "rport 0x%x removed, returning DID_NO_CONNECT\n", - rport->port_id); + mqtag = blk_mq_unique_tag(rq); + spin_lock_irqsave(&fnic->fnic_lock, flags); + iport = &fnic->iport; - atomic64_inc(&fnic_stats->misc_stats.rport_not_ready); - sc->result = DID_NO_CONNECT<<16; + if (iport->state != FNIC_IPORT_STATE_READY) { + spin_unlock_irqrestore(&fnic->fnic_lock, flags); + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, + "returning DID_NO_CONNECT for IO as iport state: %d\n", + iport->state); + sc->result = DID_NO_CONNECT << 16; done(sc); return 0; } - if (rp->rp_state != RPORT_ST_READY) { - spin_unlock_irqrestore(&fnic->fnic_lock, flags); - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, - "rport 0x%x in state 0x%x, returning DID_IMM_RETRY\n", - rport->port_id, rp->rp_state); + /* fc_remote_port_add() may have added the tport to + * fc_transport but dd_data not yet set + */ + rdd_data = rport->dd_data; + tport = rdd_data->tport; + if (!tport || (rdd_data->iport != iport)) { + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, + "dd_data not yet set in SCSI for rport portid: 0x%x\n", + rport->port_id); + tport = fnic_find_tport_by_fcid(iport, rport->port_id); + if (!tport) { + spin_unlock_irqrestore(&fnic->fnic_lock, flags); + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, + "returning DID_BUS_BUSY for IO as tport not found for: 0x%x\n", + rport->port_id); + sc->result = DID_BUS_BUSY << 16; + done(sc); + return 0; + } + + /* Re-assign same params as in fnic_fdls_add_tport */ + rport->maxframe_size = FNIC_FC_MAX_PAYLOAD_LEN; + rport->supported_classes = + FC_COS_CLASS3 | FC_RPORT_ROLE_FCP_TARGET; + /* the dd_data is allocated by fctransport of size dd_fcrport_size */ + rdd_data = rport->dd_data; + rdd_data->tport = tport; + rdd_data->iport = iport; + tport->rport = rport; + tport->flags |= FNIC_FDLS_SCSI_REGISTERED; + } - sc->result = DID_IMM_RETRY << 16; + if ((tport->state != FDLS_TGT_STATE_READY) + && (tport->state != FDLS_TGT_STATE_ADISC)) { + spin_unlock_irqrestore(&fnic->fnic_lock, flags); + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, + "returning DID_NO_CONNECT for IO as tport state: %d\n", + tport->state); + sc->result = DID_NO_CONNECT << 16; done(sc); return 0; } - if (lp->state != LPORT_ST_READY || !(lp->link_up)) { + atomic_inc(&fnic->in_flight); + atomic_inc(&tport->in_flight); + + if (unlikely(fnic_chk_state_flags_locked(fnic, FNIC_FLAGS_IO_BLOCKED))) { + atomic_dec(&fnic->in_flight); + atomic_dec(&tport->in_flight); + spin_unlock_irqrestore(&fnic->fnic_lock, flags); + return SCSI_MLQUEUE_HOST_BUSY; + } + + if (unlikely(fnic_chk_state_flags_locked(fnic, FNIC_FLAGS_FWRESET))) { spin_unlock_irqrestore(&fnic->fnic_lock, flags); FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, - "state not ready: %d/link not up: %d Returning HOST_BUSY\n", - lp->state, lp->link_up); + "fnic flags FW reset: 0x%lx. Returning SCSI_MLQUEUE_HOST_BUSY\n", + fnic->state_flags); return SCSI_MLQUEUE_HOST_BUSY; } - atomic_inc(&fnic->in_flight); + if (!tport->lun0_delay) { + lun0_delay = 1; + tport->lun0_delay++; + } spin_unlock_irqrestore(&fnic->fnic_lock, flags); + fnic_priv(sc)->state = FNIC_IOREQ_NOT_INITED; fnic_priv(sc)->flags = FNIC_NO_FLAGS; @@ -499,6 +611,7 @@ int fnic_queuecommand(struct Scsi_Host *shost, struct scsi_cmnd *sc) goto out; } + io_req->tport = tport; /* Determine the type of scatter/gather list we need */ io_req->sgl_cnt = sg_count; io_req->sgl_type = FNIC_SGL_CACHE_DFLT; @@ -575,6 +688,7 @@ int fnic_queuecommand(struct Scsi_Host *shost, struct scsi_cmnd *sc) mempool_free(io_req, fnic->io_req_pool); } atomic_dec(&fnic->in_flight); + atomic_dec(&tport->in_flight); return ret; } else { atomic64_inc(&fnic_stats->io_stats.active_ios); @@ -602,6 +716,14 @@ int fnic_queuecommand(struct Scsi_Host *shost, struct scsi_cmnd *sc) spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); atomic_dec(&fnic->in_flight); + atomic_dec(&tport->in_flight); + + if (lun0_delay) { + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, + "LUN0 delay\n"); + mdelay(LUN0_DELAY_TIME); + } + return ret; } @@ -625,7 +747,7 @@ static int fnic_fcpio_fw_reset_cmpl_handler(struct fnic *fnic, atomic64_inc(&reset_stats->fw_reset_completions); /* Clean up all outstanding io requests */ - fnic_cleanup_io(fnic); + fnic_cleanup_io(fnic, SCSI_NO_TAG); atomic64_set(&fnic->fnic_stats.fw_stats.active_fw_reqs, 0); atomic64_set(&fnic->fnic_stats.io_stats.active_ios, 0); @@ -646,12 +768,6 @@ static int fnic_fcpio_fw_reset_cmpl_handler(struct fnic *fnic, "reset failed with header status: %s\n", fnic_fcpio_status_to_str(hdr_status)); - /* - * Unable to change to eth mode, cannot send out flogi - * Change state to fc mode, so that subsequent Flogi - * requests from libFC will cause more attempts to - * reset the firmware. Free the cached flogi - */ fnic->state = FNIC_IN_FC_MODE; atomic64_inc(&reset_stats->fw_reset_failures); ret = -1; @@ -664,15 +780,14 @@ static int fnic_fcpio_fw_reset_cmpl_handler(struct fnic *fnic, ret = -1; } - /* Thread removing device blocks till firmware reset is complete */ - if (fnic->remove_wait) - complete(fnic->remove_wait); + if (fnic->fw_reset_done) + complete(fnic->fw_reset_done); /* * If fnic is being removed, or fw reset failed * free the flogi frame. Else, send it out */ - if (fnic->remove_wait || ret) { + if (ret) { spin_unlock_irqrestore(&fnic->fnic_lock, flags); fnic_free_txq(&fnic->tx_queue); goto reset_cmpl_handler_end; @@ -711,12 +826,12 @@ static int fnic_fcpio_flogi_reg_cmpl_handler(struct fnic *fnic, /* Check flogi registration completion status */ if (!hdr_status) { FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, - "flog reg succeeded\n"); + "FLOGI reg succeeded\n"); fnic->state = FNIC_IN_FC_MODE; } else { FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, - "fnic flogi reg :failed %s\n", + "fnic flogi reg failed: %s\n", fnic_fcpio_status_to_str(hdr_status)); fnic->state = FNIC_IN_ETH_MODE; ret = -1; @@ -1023,15 +1138,6 @@ static void fnic_fcpio_icmnd_cmpl_handler(struct fnic *fnic, unsigned int cq_ind jiffies_to_msecs(jiffies - start_time)), desc, cmd_trace, fnic_flags_and_state(sc)); - if (sc->sc_data_direction == DMA_FROM_DEVICE) { - fnic->lport->host_stats.fcp_input_requests++; - fnic->fcp_input_bytes += xfer_len; - } else if (sc->sc_data_direction == DMA_TO_DEVICE) { - fnic->lport->host_stats.fcp_output_requests++; - fnic->fcp_output_bytes += xfer_len; - } else - fnic->lport->host_stats.fcp_control_requests++; - /* Call SCSI completion function to complete the IO */ scsi_done(sc); @@ -1414,8 +1520,8 @@ static bool fnic_cleanup_io_iter(struct scsi_cmnd *sc, void *data) struct request *const rq = scsi_cmd_to_rq(sc); struct fnic *fnic = data; struct fnic_io_req *io_req; - unsigned long flags = 0; unsigned long start_time = 0; + unsigned long flags; struct fnic_stats *fnic_stats = &fnic->fnic_stats; uint16_t hwq = 0; int tag; @@ -1439,7 +1545,7 @@ static bool fnic_cleanup_io_iter(struct scsi_cmnd *sc, void *data) } if ((fnic_priv(sc)->flags & FNIC_DEVICE_RESET) && - !(fnic_priv(sc)->flags & FNIC_DEV_RST_DONE)) { + !(fnic_priv(sc)->flags & FNIC_DEV_RST_DONE)) { /* * We will be here only when FW completes reset * without sending completions for outstanding ios. @@ -1449,6 +1555,7 @@ static bool fnic_cleanup_io_iter(struct scsi_cmnd *sc, void *data) complete(io_req->dr_done); else if (io_req && io_req->abts_done) complete(io_req->abts_done); + spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); return true; } else if (fnic_priv(sc)->flags & FNIC_DEVICE_RESET) { @@ -1458,19 +1565,19 @@ static bool fnic_cleanup_io_iter(struct scsi_cmnd *sc, void *data) fnic_priv(sc)->io_req = NULL; io_req->sc = NULL; + start_time = io_req->start_time; spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); /* * If there is a scsi_cmnd associated with this io_req, then * free the corresponding state */ - start_time = io_req->start_time; fnic_release_ioreq_buf(fnic, io_req, sc); mempool_free(io_req, fnic->io_req_pool); sc->result = DID_TRANSPORT_DISRUPTED << 16; FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, - "mqtag:0x%x tag: 0x%x sc:0x%p duration = %lu DID_TRANSPORT_DISRUPTED\n", + "mqtag: 0x%x tag: 0x%x sc: 0x%p duration = %lu DID_TRANSPORT_DISRUPTED\n", mqtag, tag, sc, (jiffies - start_time)); if (atomic64_read(&fnic->io_cmpl_skip)) @@ -1479,23 +1586,60 @@ static bool fnic_cleanup_io_iter(struct scsi_cmnd *sc, void *data) atomic64_inc(&fnic_stats->io_stats.io_completions); FNIC_TRACE(fnic_cleanup_io, - sc->device->host->host_no, tag, sc, - jiffies_to_msecs(jiffies - start_time), - 0, ((u64)sc->cmnd[0] << 32 | - (u64)sc->cmnd[2] << 24 | - (u64)sc->cmnd[3] << 16 | - (u64)sc->cmnd[4] << 8 | sc->cmnd[5]), - fnic_flags_and_state(sc)); - + sc->device->host->host_no, tag, sc, + jiffies_to_msecs(jiffies - start_time), + 0, ((u64) sc->cmnd[0] << 32 | + (u64) sc->cmnd[2] << 24 | + (u64) sc->cmnd[3] << 16 | + (u64) sc->cmnd[4] << 8 | sc->cmnd[5]), + (((u64) fnic_priv(sc)->flags << 32) | fnic_priv(sc)-> + state)); + + /* Complete the command to SCSI */ scsi_done(sc); - return true; } -static void fnic_cleanup_io(struct fnic *fnic) +static void fnic_cleanup_io(struct fnic *fnic, int exclude_id) { + unsigned int io_count = 0; + unsigned long flags; + struct fnic_io_req *io_req = NULL; + struct scsi_cmnd *sc = NULL; + + io_count = fnic_count_all_ioreqs(fnic); + FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, + "Outstanding ioreq count: %d active io count: %lld Waiting\n", + io_count, + atomic64_read(&fnic->fnic_stats.io_stats.active_ios)); + scsi_host_busy_iter(fnic->lport->host, - fnic_cleanup_io_iter, fnic); + fnic_cleanup_io_iter, fnic); + + /* with sg3utils device reset, SC needs to be retrieved from ioreq */ + spin_lock_irqsave(&fnic->wq_copy_lock[0], flags); + io_req = fnic->sw_copy_wq[0].io_req_table[fnic->fnic_max_tag_id]; + if (io_req) { + sc = io_req->sc; + if (sc) { + if ((fnic_priv(sc)->flags & FNIC_DEVICE_RESET) + && !(fnic_priv(sc)->flags & FNIC_DEV_RST_DONE)) { + fnic_priv(sc)->flags |= FNIC_DEV_RST_DONE; + if (io_req && io_req->dr_done) + complete(io_req->dr_done); + } + } + } + spin_unlock_irqrestore(&fnic->wq_copy_lock[0], flags); + + while ((io_count = fnic_count_all_ioreqs(fnic))) { + FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, + "Outstanding ioreq count: %d active io count: %lld Waiting\n", + io_count, + atomic64_read(&fnic->fnic_stats.io_stats.active_ios)); + + schedule_timeout(msecs_to_jiffies(100)); + } } void fnic_wq_copy_cleanup_handler(struct vnic_wq_copy *wq, @@ -1567,10 +1711,13 @@ static inline int fnic_queue_abort_io_req(struct fnic *fnic, int tag, struct vnic_wq_copy *wq = &fnic->hw_copy_wq[hwq]; struct misc_stats *misc_stats = &fnic->fnic_stats.misc_stats; unsigned long flags; + struct fnic_tport_s *tport = io_req->tport; spin_lock_irqsave(&fnic->fnic_lock, flags); if (unlikely(fnic_chk_state_flags_locked(fnic, FNIC_FLAGS_IO_BLOCKED))) { + atomic_dec(&fnic->in_flight); + atomic_dec(&tport->in_flight); spin_unlock_irqrestore(&fnic->fnic_lock, flags); return 1; } else @@ -1585,6 +1732,7 @@ static inline int fnic_queue_abort_io_req(struct fnic *fnic, int tag, if (!vnic_wq_copy_desc_avail(wq)) { spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); atomic_dec(&fnic->in_flight); + atomic_dec(&tport->in_flight); FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, "fnic_queue_abort_io_req: failure: no descriptors\n"); atomic64_inc(&misc_stats->abts_cpwq_alloc_failures); @@ -1619,20 +1767,24 @@ static bool fnic_rport_abort_io_iter(struct scsi_cmnd *sc, void *data) struct fnic *fnic = iter_data->fnic; int abt_tag = 0; struct fnic_io_req *io_req; - unsigned long flags; struct reset_stats *reset_stats = &fnic->fnic_stats.reset_stats; struct terminate_stats *term_stats = &fnic->fnic_stats.term_stats; struct scsi_lun fc_lun; enum fnic_ioreq_state old_ioreq_state; uint16_t hwq = 0; + unsigned long flags; abt_tag = blk_mq_unique_tag(rq); hwq = blk_mq_unique_tag_to_hwq(abt_tag); - spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); + if (!sc) { + FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, + "sc is NULL abt_tag: 0x%x hwq: %d\n", abt_tag, hwq); + return true; + } + spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); io_req = fnic_priv(sc)->io_req; - if (!io_req || io_req->port_id != iter_data->port_id) { spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); return true; @@ -1655,37 +1807,38 @@ static bool fnic_rport_abort_io_iter(struct scsi_cmnd *sc, void *data) spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); return true; } + if (io_req->abts_done) { shost_printk(KERN_ERR, fnic->lport->host, - "fnic_rport_exch_reset: io_req->abts_done is set " - "state is %s\n", + "fnic_rport_exch_reset: io_req->abts_done is set state is %s\n", fnic_ioreq_state_to_str(fnic_priv(sc)->state)); } if (!(fnic_priv(sc)->flags & FNIC_IO_ISSUED)) { shost_printk(KERN_ERR, fnic->lport->host, - "rport_exch_reset " - "IO not yet issued %p tag 0x%x flags " - "%x state %d\n", - sc, abt_tag, fnic_priv(sc)->flags, fnic_priv(sc)->state); + "rport_exch_reset IO not yet issued %p abt_tag 0x%x", + sc, abt_tag); + shost_printk(KERN_ERR, fnic->lport->host, + "flags %x state %d\n", fnic_priv(sc)->flags, + fnic_priv(sc)->state); } old_ioreq_state = fnic_priv(sc)->state; fnic_priv(sc)->state = FNIC_IOREQ_ABTS_PENDING; fnic_priv(sc)->abts_status = FCPIO_INVALID_CODE; + if (fnic_priv(sc)->flags & FNIC_DEVICE_RESET) { atomic64_inc(&reset_stats->device_reset_terminates); abt_tag |= FNIC_TAG_DEV_RST; } FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, - "fnic_rport_exch_reset dev rst sc 0x%p\n", sc); - BUG_ON(io_req->abts_done); - + "fnic_rport_exch_reset: dev rst sc 0x%p\n", sc); + WARN_ON_ONCE(io_req->abts_done); FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, "fnic_rport_reset_exch: Issuing abts\n"); spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); - /* Now queue the abort command to firmware */ + /* Queue the abort command to firmware */ int_to_scsilun(sc->device->lun, &fc_lun); if (fnic_queue_abort_io_req(fnic, abt_tag, @@ -1714,11 +1867,14 @@ static bool fnic_rport_abort_io_iter(struct scsi_cmnd *sc, void *data) atomic64_inc(&term_stats->terminates); iter_data->term_cnt++; } + return true; } void fnic_rport_exch_reset(struct fnic *fnic, u32 port_id) { + unsigned int io_count = 0; + unsigned long flags; struct terminate_stats *term_stats = &fnic->fnic_stats.term_stats; struct fnic_rport_abort_io_iter_data iter_data = { .fnic = fnic, @@ -1726,53 +1882,75 @@ void fnic_rport_exch_reset(struct fnic *fnic, u32 port_id) .term_cnt = 0, }; - FNIC_SCSI_DBG(KERN_DEBUG, - fnic->lport->host, fnic->fnic_num, - "fnic_rport_exch_reset called portid 0x%06x\n", - port_id); + FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, + "fnic rport exchange reset for tport: 0x%06x\n", + port_id); if (fnic->in_remove) return; + io_count = fnic_count_ioreqs(fnic, port_id); + FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, + "Starting terminates: rport:0x%x portid-io-count: %d active-io-count: %lld\n", + port_id, io_count, + atomic64_read(&fnic->fnic_stats.io_stats.active_ios)); + + spin_lock_irqsave(&fnic->fnic_lock, flags); + atomic_inc(&fnic->in_flight); + if (unlikely(fnic_chk_state_flags_locked(fnic, FNIC_FLAGS_IO_BLOCKED))) { + atomic_dec(&fnic->in_flight); + spin_unlock_irqrestore(&fnic->fnic_lock, flags); + return; + } + spin_unlock_irqrestore(&fnic->fnic_lock, flags); + scsi_host_busy_iter(fnic->lport->host, fnic_rport_abort_io_iter, &iter_data); + if (iter_data.term_cnt > atomic64_read(&term_stats->max_terminates)) atomic64_set(&term_stats->max_terminates, iter_data.term_cnt); + atomic_dec(&fnic->in_flight); + + while ((io_count = fnic_count_ioreqs(fnic, port_id))) + schedule_timeout(msecs_to_jiffies(1000)); + + FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, + "rport: 0x%x remaining portid-io-count: %d ", + port_id, io_count); } void fnic_terminate_rport_io(struct fc_rport *rport) { - struct fc_rport_libfc_priv *rdata; - struct fc_lport *lport; - struct fnic *fnic; + struct fnic_tport_s *tport; + struct rport_dd_data_s *rdd_data; + struct fnic_iport_s *iport = NULL; + struct fnic *fnic = NULL; if (!rport) { - printk(KERN_ERR "fnic_terminate_rport_io: rport is NULL\n"); + pr_err("rport is NULL\n"); return; } - rdata = rport->dd_data; - if (!rdata) { - printk(KERN_ERR "fnic_terminate_rport_io: rdata is NULL\n"); - return; - } - lport = rdata->local_port; - - if (!lport) { - printk(KERN_ERR "fnic_terminate_rport_io: lport is NULL\n"); - return; + rdd_data = rport->dd_data; + if (rdd_data) { + tport = rdd_data->tport; + if (!tport) { + pr_err( + "term rport io called after tport is deleted. Returning 0x%8x\n", + rport->port_id); + } else { + pr_err( + "term rport io called after tport is set 0x%8x\n", + rport->port_id); + pr_err( + "tport maybe rediscovered\n"); + + iport = (struct fnic_iport_s *) tport->iport; + fnic = iport->fnic; + fnic_rport_exch_reset(fnic, rport->port_id); + } } - fnic = lport_priv(lport); - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, - "wwpn 0x%llx, wwnn0x%llx, rport 0x%p, portid 0x%06x\n", - rport->port_name, rport->node_name, rport, - rport->port_id); - - if (fnic->in_remove) - return; - - fnic_rport_exch_reset(fnic, rport->port_id); } /* @@ -1783,10 +1961,12 @@ void fnic_terminate_rport_io(struct fc_rport *rport) int fnic_abort_cmd(struct scsi_cmnd *sc) { struct request *const rq = scsi_cmd_to_rq(sc); - struct fc_lport *lp; + struct fnic_iport_s *iport; + struct fnic_tport_s *tport; struct fnic *fnic; struct fnic_io_req *io_req = NULL; struct fc_rport *rport; + struct rport_dd_data_s *rdd_data; unsigned long flags; unsigned long start_time = 0; int ret = SUCCESS; @@ -1806,11 +1986,11 @@ int fnic_abort_cmd(struct scsi_cmnd *sc) fc_block_scsi_eh(sc); /* Get local-port, check ready and link up */ - lp = shost_priv(sc->device->host); - - fnic = lport_priv(lp); + fnic = *((struct fnic **) shost_priv(sc->device->host)); spin_lock_irqsave(&fnic->fnic_lock, flags); + iport = &fnic->iport; + fnic_stats = &fnic->fnic_stats; abts_stats = &fnic->fnic_stats.abts_stats; term_stats = &fnic->fnic_stats.term_stats; @@ -1821,7 +2001,43 @@ int fnic_abort_cmd(struct scsi_cmnd *sc) fnic_priv(sc)->flags = FNIC_NO_FLAGS; - if (lp->state != LPORT_ST_READY || !(lp->link_up)) { + rdd_data = rport->dd_data; + tport = rdd_data->tport; + + if (!tport) { + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, + "Abort cmd called after tport delete! rport fcid: 0x%x", + rport->port_id); + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, + "lun: %llu hwq: 0x%x mqtag: 0x%x Op: 0x%x flags: 0x%x\n", + sc->device->lun, hwq, mqtag, + sc->cmnd[0], fnic_priv(sc)->flags); + ret = FAILED; + spin_unlock_irqrestore(&fnic->fnic_lock, flags); + goto fnic_abort_cmd_end; + } + + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, + "Abort cmd called rport fcid: 0x%x lun: %llu hwq: 0x%x mqtag: 0x%x", + rport->port_id, sc->device->lun, hwq, mqtag); + + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, + "Op: 0x%x flags: 0x%x\n", + sc->cmnd[0], + fnic_priv(sc)->flags); + + if (iport->state != FNIC_IPORT_STATE_READY) { + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, + "iport NOT in READY state"); + ret = FAILED; + spin_unlock_irqrestore(&fnic->fnic_lock, flags); + goto fnic_abort_cmd_end; + } + + if ((tport->state != FDLS_TGT_STATE_READY) && + (tport->state != FDLS_TGT_STATE_ADISC)) { + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, + "tport state: %d\n", tport->state); ret = FAILED; spin_unlock_irqrestore(&fnic->fnic_lock, flags); goto fnic_abort_cmd_end; @@ -1843,6 +2059,7 @@ int fnic_abort_cmd(struct scsi_cmnd *sc) spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); io_req = fnic_priv(sc)->io_req; if (!io_req) { + ret = FAILED; spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); goto fnic_abort_cmd_end; } @@ -1893,7 +2110,6 @@ int fnic_abort_cmd(struct scsi_cmnd *sc) if (fc_remote_port_chkready(rport) == 0) task_req = FCPIO_ITMF_ABT_TASK; else { - atomic64_inc(&fnic_stats->misc_stats.rport_not_ready); task_req = FCPIO_ITMF_ABT_TASK_TERM; } @@ -2027,6 +2243,7 @@ static inline int fnic_queue_dr_io_req(struct fnic *fnic, unsigned long flags; uint16_t hwq = 0; uint32_t tag = 0; + struct fnic_tport_s *tport = io_req->tport; tag = io_req->tag; hwq = blk_mq_unique_tag_to_hwq(tag); @@ -2037,8 +2254,10 @@ static inline int fnic_queue_dr_io_req(struct fnic *fnic, FNIC_FLAGS_IO_BLOCKED))) { spin_unlock_irqrestore(&fnic->fnic_lock, flags); return FAILED; - } else + } else { atomic_inc(&fnic->in_flight); + atomic_inc(&tport->in_flight); + } spin_unlock_irqrestore(&fnic->fnic_lock, flags); spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); @@ -2072,6 +2291,7 @@ static inline int fnic_queue_dr_io_req(struct fnic *fnic, lr_io_req_end: spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); atomic_dec(&fnic->in_flight); + atomic_inc(&tport->in_flight); return ret; } @@ -2274,11 +2494,11 @@ static int fnic_clean_pending_aborts(struct fnic *fnic, int fnic_device_reset(struct scsi_cmnd *sc) { struct request *rq = scsi_cmd_to_rq(sc); - struct fc_lport *lp; struct fnic *fnic; struct fnic_io_req *io_req = NULL; struct fc_rport *rport; int status; + int count = 0; int ret = FAILED; unsigned long flags; unsigned long start_time = 0; @@ -2289,31 +2509,61 @@ int fnic_device_reset(struct scsi_cmnd *sc) DECLARE_COMPLETION_ONSTACK(tm_done); bool new_sc = 0; uint16_t hwq = 0; + struct fnic_iport_s *iport = NULL; + struct rport_dd_data_s *rdd_data; + struct fnic_tport_s *tport; + u32 old_soft_reset_count; + u32 old_link_down_cnt; + int exit_dr = 0; /* Wait for rport to unblock */ fc_block_scsi_eh(sc); /* Get local-port, check ready and link up */ - lp = shost_priv(sc->device->host); + fnic = *((struct fnic **) shost_priv(sc->device->host)); + iport = &fnic->iport; - fnic = lport_priv(lp); fnic_stats = &fnic->fnic_stats; reset_stats = &fnic->fnic_stats.reset_stats; atomic64_inc(&reset_stats->device_resets); rport = starget_to_rport(scsi_target(sc->device)); + + spin_lock_irqsave(&fnic->fnic_lock, flags); FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, - "fcid: 0x%x lun: 0x%llx hwq: %d mqtag: 0x%x flags: 0x%x Device reset\n", + "fcid: 0x%x lun: %llu hwq: %d mqtag: 0x%x flags: 0x%x Device reset\n", rport->port_id, sc->device->lun, hwq, mqtag, fnic_priv(sc)->flags); - if (lp->state != LPORT_ST_READY || !(lp->link_up)) + rdd_data = rport->dd_data; + tport = rdd_data->tport; + if (!tport) { + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, + "Dev rst called after tport delete! rport fcid: 0x%x lun: %llu\n", + rport->port_id, sc->device->lun); + spin_unlock_irqrestore(&fnic->fnic_lock, flags); + goto fnic_device_reset_end; + } + + if (iport->state != FNIC_IPORT_STATE_READY) { + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, + "iport NOT in READY state"); + spin_unlock_irqrestore(&fnic->fnic_lock, flags); goto fnic_device_reset_end; + } + + if ((tport->state != FDLS_TGT_STATE_READY) && + (tport->state != FDLS_TGT_STATE_ADISC)) { + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, + "tport state: %d\n", tport->state); + spin_unlock_irqrestore(&fnic->fnic_lock, flags); + goto fnic_device_reset_end; + } + spin_unlock_irqrestore(&fnic->fnic_lock, flags); /* Check if remote port up */ if (fc_remote_port_chkready(rport)) { - atomic64_inc(&fnic_stats->misc_stats.rport_not_ready); goto fnic_device_reset_end; } @@ -2352,6 +2602,7 @@ int fnic_device_reset(struct scsi_cmnd *sc) io_req->port_id = rport->port_id; io_req->tag = mqtag; fnic_priv(sc)->io_req = io_req; + io_req->tport = tport; io_req->sc = sc; if (fnic->sw_copy_wq[hwq].io_req_table[blk_mq_unique_tag_to_tag(mqtag)] != NULL) @@ -2383,6 +2634,11 @@ int fnic_device_reset(struct scsi_cmnd *sc) fnic_priv(sc)->flags |= FNIC_DEV_RST_ISSUED; spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); + spin_lock_irqsave(&fnic->fnic_lock, flags); + old_link_down_cnt = iport->fnic->link_down_cnt; + old_soft_reset_count = fnic->soft_reset_count; + spin_unlock_irqrestore(&fnic->fnic_lock, flags); + /* * Wait on the local completion for LUN reset. The io_req may be * freed while we wait since we hold no lock. @@ -2390,6 +2646,24 @@ int fnic_device_reset(struct scsi_cmnd *sc) wait_for_completion_timeout(&tm_done, msecs_to_jiffies(FNIC_LUN_RESET_TIMEOUT)); + /* + * Wake up can be due to the following reasons: + * 1) The device reset completed from target. + * 2) Device reset timed out. + * 3) A link-down/host_reset may have happened in between. + * 4) The device reset was aborted and io_req->dr_done was called. + */ + + exit_dr = 0; + spin_lock_irqsave(&fnic->fnic_lock, flags); + if ((old_link_down_cnt != fnic->link_down_cnt) || + (fnic->reset_in_progress) || + (fnic->soft_reset_count != old_soft_reset_count) || + (iport->state != FNIC_IPORT_STATE_READY)) + exit_dr = 1; + + spin_unlock_irqrestore(&fnic->fnic_lock, flags); + spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); io_req = fnic_priv(sc)->io_req; if (!io_req) { @@ -2398,6 +2672,13 @@ int fnic_device_reset(struct scsi_cmnd *sc) "io_req is null mqtag 0x%x sc 0x%p\n", mqtag, sc); goto fnic_device_reset_end; } + + if (exit_dr) { + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, + "Host reset called for fnic. Exit device reset\n"); + io_req->dr_done = NULL; + goto fnic_device_reset_clean; + } io_req->dr_done = NULL; status = fnic_priv(sc)->lr_status; @@ -2411,50 +2692,8 @@ int fnic_device_reset(struct scsi_cmnd *sc) FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, "Device reset timed out\n"); fnic_priv(sc)->flags |= FNIC_DEV_RST_TIMED_OUT; - spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); int_to_scsilun(sc->device->lun, &fc_lun); - /* - * Issue abort and terminate on device reset request. - * If q'ing of terminate fails, retry it after a delay. - */ - while (1) { - spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); - if (fnic_priv(sc)->flags & FNIC_DEV_RST_TERM_ISSUED) { - spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); - break; - } - spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); - if (fnic_queue_abort_io_req(fnic, - mqtag | FNIC_TAG_DEV_RST, - FCPIO_ITMF_ABT_TASK_TERM, - fc_lun.scsi_lun, io_req, hwq)) { - wait_for_completion_timeout(&tm_done, - msecs_to_jiffies(FNIC_ABT_TERM_DELAY_TIMEOUT)); - } else { - spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); - fnic_priv(sc)->flags |= FNIC_DEV_RST_TERM_ISSUED; - fnic_priv(sc)->state = FNIC_IOREQ_ABTS_PENDING; - io_req->abts_done = &tm_done; - spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, - "Abort and terminate issued on Device reset mqtag 0x%x sc 0x%p\n", - mqtag, sc); - break; - } - } - while (1) { - spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); - if (!(fnic_priv(sc)->flags & FNIC_DEV_RST_DONE)) { - spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); - wait_for_completion_timeout(&tm_done, - msecs_to_jiffies(FNIC_LUN_RESET_TIMEOUT)); - break; - } else { - io_req = fnic_priv(sc)->io_req; - io_req->abts_done = NULL; - goto fnic_device_reset_clean; - } - } + goto fnic_device_reset_clean; } else { spin_unlock_irqrestore(&fnic->wq_copy_lock[hwq], flags); } @@ -2480,8 +2719,7 @@ int fnic_device_reset(struct scsi_cmnd *sc) spin_lock_irqsave(&fnic->wq_copy_lock[hwq], flags); io_req = fnic_priv(sc)->io_req; FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, - "Device reset failed" - " since could not abort all IOs\n"); + "Device reset failed: Cannot abort all IOs\n"); goto fnic_device_reset_clean; } @@ -2507,6 +2745,15 @@ int fnic_device_reset(struct scsi_cmnd *sc) mempool_free(io_req, fnic->io_req_pool); } + /* + * If link-event is seen while LUN reset is issued we need + * to complete the LUN reset here + */ + if (!new_sc) { + sc->result = DID_RESET << 16; + scsi_done(sc); + } + fnic_device_reset_end: FNIC_TRACE(fnic_device_reset, sc->device->host->host_no, rq->tag, sc, jiffies_to_msecs(jiffies - start_time), @@ -2520,6 +2767,17 @@ int fnic_device_reset(struct scsi_cmnd *sc) mutex_unlock(&fnic->sgreset_mutex); } + while ((ret == SUCCESS) && fnic_count_lun_ioreqs(fnic, sc->device)) { + if (count >= 2) { + ret = FAILED; + break; + } + FNIC_SCSI_DBG(KERN_ERR, fnic->lport->host, fnic->fnic_num, + "Cannot clean up all IOs for the LUN\n"); + schedule_timeout(msecs_to_jiffies(1000)); + count++; + } + FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, "Returning from device reset %s\n", (ret == SUCCESS) ? @@ -2531,50 +2789,54 @@ int fnic_device_reset(struct scsi_cmnd *sc) return ret; } -/* Clean up all IOs, clean up libFC local port */ -int fnic_reset(struct Scsi_Host *shost) +static void fnic_post_flogo_linkflap(struct fnic *fnic) +{ + unsigned long flags; + + fnic_fdls_link_status_change(fnic, 0); + spin_lock_irqsave(&fnic->fnic_lock, flags); + + if (fnic->link_status) { + spin_unlock_irqrestore(&fnic->fnic_lock, flags); + fnic_fdls_link_status_change(fnic, 1); + return; + } + spin_unlock_irqrestore(&fnic->fnic_lock, flags); +} + +/* Logout from all the targets and simulate link flap */ +void fnic_reset(struct Scsi_Host *shost) { - struct fc_lport *lp; struct fnic *fnic; - int ret = 0; struct reset_stats *reset_stats; - lp = shost_priv(shost); - fnic = lport_priv(lp); + fnic = *((struct fnic **) shost_priv(shost)); reset_stats = &fnic->fnic_stats.reset_stats; FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, - "Issuing fnic reset\n"); + "Issuing fnic reset\n"); atomic64_inc(&reset_stats->fnic_resets); - - /* - * Reset local port, this will clean up libFC exchanges, - * reset remote port sessions, and if link is up, begin flogi - */ - ret = fc_lport_reset(lp); + fnic_post_flogo_linkflap(fnic); FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, - "Returning from fnic reset with: %s\n", - (ret == 0) ? "SUCCESS" : "FAILED"); + "Returning from fnic reset"); - if (ret == 0) - atomic64_inc(&reset_stats->fnic_reset_completions); - else - atomic64_inc(&reset_stats->fnic_reset_failures); + atomic64_inc(&reset_stats->fnic_reset_completions); +} + +int fnic_issue_fc_host_lip(struct Scsi_Host *shost) +{ + int ret = 0; + struct fnic *fnic = *((struct fnic **) shost_priv(shost)); + + FNIC_SCSI_DBG(KERN_INFO, fnic->lport->host, fnic->fnic_num, + "FC host lip issued"); + ret = fnic_host_reset(shost); return ret; } -/* - * SCSI Error handling calls driver's eh_host_reset if all prior - * error handling levels return FAILED. If host reset completes - * successfully, and if link is up, then Fabric login begins. - * - * Host Reset is the highest level of error recovery. If this fails, then - * host is offlined by SCSI. - * - */ int fnic_host_reset(struct Scsi_Host *shost) { int ret = SUCCESS; @@ -2637,122 +2899,6 @@ int fnic_host_reset(struct Scsi_Host *shost) return ret; } -/* - * This fxn is called from libFC when host is removed - */ -void fnic_scsi_abort_io(struct fc_lport *lp) -{ - int err = 0; - unsigned long flags; - enum fnic_state old_state; - struct fnic *fnic = lport_priv(lp); - DECLARE_COMPLETION_ONSTACK(remove_wait); - - /* Issue firmware reset for fnic, wait for reset to complete */ -retry_fw_reset: - spin_lock_irqsave(&fnic->fnic_lock, flags); - if (unlikely(fnic->state == FNIC_IN_FC_TRANS_ETH_MODE) && - fnic->link_events) { - /* fw reset is in progress, poll for its completion */ - spin_unlock_irqrestore(&fnic->fnic_lock, flags); - schedule_timeout(msecs_to_jiffies(100)); - goto retry_fw_reset; - } - - fnic->remove_wait = &remove_wait; - old_state = fnic->state; - fnic->state = FNIC_IN_FC_TRANS_ETH_MODE; - fnic_update_mac_locked(fnic, fnic->ctlr.ctl_src_addr); - spin_unlock_irqrestore(&fnic->fnic_lock, flags); - - err = fnic_fw_reset_handler(fnic); - if (err) { - spin_lock_irqsave(&fnic->fnic_lock, flags); - if (fnic->state == FNIC_IN_FC_TRANS_ETH_MODE) - fnic->state = old_state; - fnic->remove_wait = NULL; - spin_unlock_irqrestore(&fnic->fnic_lock, flags); - return; - } - - /* Wait for firmware reset to complete */ - wait_for_completion_timeout(&remove_wait, - msecs_to_jiffies(FNIC_RMDEVICE_TIMEOUT)); - - spin_lock_irqsave(&fnic->fnic_lock, flags); - fnic->remove_wait = NULL; - FNIC_SCSI_DBG(KERN_DEBUG, fnic->lport->host, fnic->fnic_num, - "fnic_scsi_abort_io %s\n", - (fnic->state == FNIC_IN_ETH_MODE) ? - "SUCCESS" : "FAILED"); - spin_unlock_irqrestore(&fnic->fnic_lock, flags); - -} - -/* - * This fxn called from libFC to clean up driver IO state on link down - */ -void fnic_scsi_cleanup(struct fc_lport *lp) -{ - unsigned long flags; - enum fnic_state old_state; - struct fnic *fnic = lport_priv(lp); - - /* issue fw reset */ -retry_fw_reset: - spin_lock_irqsave(&fnic->fnic_lock, flags); - if (unlikely(fnic->state == FNIC_IN_FC_TRANS_ETH_MODE)) { - /* fw reset is in progress, poll for its completion */ - spin_unlock_irqrestore(&fnic->fnic_lock, flags); - schedule_timeout(msecs_to_jiffies(100)); - goto retry_fw_reset; - } - old_state = fnic->state; - fnic->state = FNIC_IN_FC_TRANS_ETH_MODE; - fnic_update_mac_locked(fnic, fnic->ctlr.ctl_src_addr); - spin_unlock_irqrestore(&fnic->fnic_lock, flags); - - if (fnic_fw_reset_handler(fnic)) { - spin_lock_irqsave(&fnic->fnic_lock, flags); - if (fnic->state == FNIC_IN_FC_TRANS_ETH_MODE) - fnic->state = old_state; - spin_unlock_irqrestore(&fnic->fnic_lock, flags); - } - -} - -void fnic_empty_scsi_cleanup(struct fc_lport *lp) -{ -} - -void fnic_exch_mgr_reset(struct fc_lport *lp, u32 sid, u32 did) -{ - struct fnic *fnic = lport_priv(lp); - - /* Non-zero sid, nothing to do */ - if (sid) - goto call_fc_exch_mgr_reset; - - if (did) { - fnic_rport_exch_reset(fnic, did); - goto call_fc_exch_mgr_reset; - } - - /* - * sid = 0, did = 0 - * link down or device being removed - */ - if (!fnic->in_remove) - fnic_scsi_cleanup(lp); - else - fnic_scsi_abort_io(lp); - - /* call libFC exch mgr reset to reset its exchanges */ -call_fc_exch_mgr_reset: - fc_exch_mgr_reset(lp, sid, did); - -} - static bool fnic_abts_pending_iter(struct scsi_cmnd *sc, void *data) { struct request *const rq = scsi_cmd_to_rq(sc); diff --git a/drivers/scsi/fnic/fnic_stats.h b/drivers/scsi/fnic/fnic_stats.h index 9d7f98c452dd..1f1a1ec90a23 100644 --- a/drivers/scsi/fnic/fnic_stats.h +++ b/drivers/scsi/fnic/fnic_stats.h @@ -127,6 +127,4 @@ struct stats_debug_info { }; int fnic_get_stats_data(struct stats_debug_info *, struct fnic_stats *); -void fnic_stats_debugfs_init(struct fnic *); -void fnic_stats_debugfs_remove(struct fnic *); #endif /* _FNIC_STATS_H_ */