Merge remote-tracking branches 'asoc/topic/sta529', 'asoc/topic/sti', 'asoc/topic...
[sfrench/cifs-2.6.git] / drivers / scsi / qla2xxx / qla_isr.c
1 /*
2  * QLogic Fibre Channel HBA Driver
3  * Copyright (c)  2003-2014 QLogic Corporation
4  *
5  * See LICENSE.qla2xxx for copyright and licensing details.
6  */
7 #include "qla_def.h"
8 #include "qla_target.h"
9
10 #include <linux/delay.h>
11 #include <linux/slab.h>
12 #include <linux/t10-pi.h>
13 #include <scsi/scsi_tcq.h>
14 #include <scsi/scsi_bsg_fc.h>
15 #include <scsi/scsi_eh.h>
16
17 static void qla2x00_mbx_completion(scsi_qla_host_t *, uint16_t);
18 static void qla2x00_status_entry(scsi_qla_host_t *, struct rsp_que *, void *);
19 static void qla2x00_status_cont_entry(struct rsp_que *, sts_cont_entry_t *);
20 static void qla2x00_error_entry(scsi_qla_host_t *, struct rsp_que *,
21         sts_entry_t *);
22
23 /**
24  * qla2100_intr_handler() - Process interrupts for the ISP2100 and ISP2200.
25  * @irq:
26  * @dev_id: SCSI driver HA context
27  *
28  * Called by system whenever the host adapter generates an interrupt.
29  *
30  * Returns handled flag.
31  */
32 irqreturn_t
33 qla2100_intr_handler(int irq, void *dev_id)
34 {
35         scsi_qla_host_t *vha;
36         struct qla_hw_data *ha;
37         struct device_reg_2xxx __iomem *reg;
38         int             status;
39         unsigned long   iter;
40         uint16_t        hccr;
41         uint16_t        mb[4];
42         struct rsp_que *rsp;
43         unsigned long   flags;
44
45         rsp = (struct rsp_que *) dev_id;
46         if (!rsp) {
47                 ql_log(ql_log_info, NULL, 0x505d,
48                     "%s: NULL response queue pointer.\n", __func__);
49                 return (IRQ_NONE);
50         }
51
52         ha = rsp->hw;
53         reg = &ha->iobase->isp;
54         status = 0;
55
56         spin_lock_irqsave(&ha->hardware_lock, flags);
57         vha = pci_get_drvdata(ha->pdev);
58         for (iter = 50; iter--; ) {
59                 hccr = RD_REG_WORD(&reg->hccr);
60                 if (qla2x00_check_reg16_for_disconnect(vha, hccr))
61                         break;
62                 if (hccr & HCCR_RISC_PAUSE) {
63                         if (pci_channel_offline(ha->pdev))
64                                 break;
65
66                         /*
67                          * Issue a "HARD" reset in order for the RISC interrupt
68                          * bit to be cleared.  Schedule a big hammer to get
69                          * out of the RISC PAUSED state.
70                          */
71                         WRT_REG_WORD(&reg->hccr, HCCR_RESET_RISC);
72                         RD_REG_WORD(&reg->hccr);
73
74                         ha->isp_ops->fw_dump(vha, 1);
75                         set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
76                         break;
77                 } else if ((RD_REG_WORD(&reg->istatus) & ISR_RISC_INT) == 0)
78                         break;
79
80                 if (RD_REG_WORD(&reg->semaphore) & BIT_0) {
81                         WRT_REG_WORD(&reg->hccr, HCCR_CLR_RISC_INT);
82                         RD_REG_WORD(&reg->hccr);
83
84                         /* Get mailbox data. */
85                         mb[0] = RD_MAILBOX_REG(ha, reg, 0);
86                         if (mb[0] > 0x3fff && mb[0] < 0x8000) {
87                                 qla2x00_mbx_completion(vha, mb[0]);
88                                 status |= MBX_INTERRUPT;
89                         } else if (mb[0] > 0x7fff && mb[0] < 0xc000) {
90                                 mb[1] = RD_MAILBOX_REG(ha, reg, 1);
91                                 mb[2] = RD_MAILBOX_REG(ha, reg, 2);
92                                 mb[3] = RD_MAILBOX_REG(ha, reg, 3);
93                                 qla2x00_async_event(vha, rsp, mb);
94                         } else {
95                                 /*EMPTY*/
96                                 ql_dbg(ql_dbg_async, vha, 0x5025,
97                                     "Unrecognized interrupt type (%d).\n",
98                                     mb[0]);
99                         }
100                         /* Release mailbox registers. */
101                         WRT_REG_WORD(&reg->semaphore, 0);
102                         RD_REG_WORD(&reg->semaphore);
103                 } else {
104                         qla2x00_process_response_queue(rsp);
105
106                         WRT_REG_WORD(&reg->hccr, HCCR_CLR_RISC_INT);
107                         RD_REG_WORD(&reg->hccr);
108                 }
109         }
110         qla2x00_handle_mbx_completion(ha, status);
111         spin_unlock_irqrestore(&ha->hardware_lock, flags);
112
113         return (IRQ_HANDLED);
114 }
115
116 bool
117 qla2x00_check_reg32_for_disconnect(scsi_qla_host_t *vha, uint32_t reg)
118 {
119         /* Check for PCI disconnection */
120         if (reg == 0xffffffff && !pci_channel_offline(vha->hw->pdev)) {
121                 if (!test_and_set_bit(PFLG_DISCONNECTED, &vha->pci_flags) &&
122                     !test_bit(PFLG_DRIVER_REMOVING, &vha->pci_flags) &&
123                     !test_bit(PFLG_DRIVER_PROBING, &vha->pci_flags)) {
124                         /*
125                          * Schedule this (only once) on the default system
126                          * workqueue so that all the adapter workqueues and the
127                          * DPC thread can be shutdown cleanly.
128                          */
129                         schedule_work(&vha->hw->board_disable);
130                 }
131                 return true;
132         } else
133                 return false;
134 }
135
136 bool
137 qla2x00_check_reg16_for_disconnect(scsi_qla_host_t *vha, uint16_t reg)
138 {
139         return qla2x00_check_reg32_for_disconnect(vha, 0xffff0000 | reg);
140 }
141
142 /**
143  * qla2300_intr_handler() - Process interrupts for the ISP23xx and ISP63xx.
144  * @irq:
145  * @dev_id: SCSI driver HA context
146  *
147  * Called by system whenever the host adapter generates an interrupt.
148  *
149  * Returns handled flag.
150  */
151 irqreturn_t
152 qla2300_intr_handler(int irq, void *dev_id)
153 {
154         scsi_qla_host_t *vha;
155         struct device_reg_2xxx __iomem *reg;
156         int             status;
157         unsigned long   iter;
158         uint32_t        stat;
159         uint16_t        hccr;
160         uint16_t        mb[4];
161         struct rsp_que *rsp;
162         struct qla_hw_data *ha;
163         unsigned long   flags;
164
165         rsp = (struct rsp_que *) dev_id;
166         if (!rsp) {
167                 ql_log(ql_log_info, NULL, 0x5058,
168                     "%s: NULL response queue pointer.\n", __func__);
169                 return (IRQ_NONE);
170         }
171
172         ha = rsp->hw;
173         reg = &ha->iobase->isp;
174         status = 0;
175
176         spin_lock_irqsave(&ha->hardware_lock, flags);
177         vha = pci_get_drvdata(ha->pdev);
178         for (iter = 50; iter--; ) {
179                 stat = RD_REG_DWORD(&reg->u.isp2300.host_status);
180                 if (qla2x00_check_reg32_for_disconnect(vha, stat))
181                         break;
182                 if (stat & HSR_RISC_PAUSED) {
183                         if (unlikely(pci_channel_offline(ha->pdev)))
184                                 break;
185
186                         hccr = RD_REG_WORD(&reg->hccr);
187
188                         if (hccr & (BIT_15 | BIT_13 | BIT_11 | BIT_8))
189                                 ql_log(ql_log_warn, vha, 0x5026,
190                                     "Parity error -- HCCR=%x, Dumping "
191                                     "firmware.\n", hccr);
192                         else
193                                 ql_log(ql_log_warn, vha, 0x5027,
194                                     "RISC paused -- HCCR=%x, Dumping "
195                                     "firmware.\n", hccr);
196
197                         /*
198                          * Issue a "HARD" reset in order for the RISC
199                          * interrupt bit to be cleared.  Schedule a big
200                          * hammer to get out of the RISC PAUSED state.
201                          */
202                         WRT_REG_WORD(&reg->hccr, HCCR_RESET_RISC);
203                         RD_REG_WORD(&reg->hccr);
204
205                         ha->isp_ops->fw_dump(vha, 1);
206                         set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
207                         break;
208                 } else if ((stat & HSR_RISC_INT) == 0)
209                         break;
210
211                 switch (stat & 0xff) {
212                 case 0x1:
213                 case 0x2:
214                 case 0x10:
215                 case 0x11:
216                         qla2x00_mbx_completion(vha, MSW(stat));
217                         status |= MBX_INTERRUPT;
218
219                         /* Release mailbox registers. */
220                         WRT_REG_WORD(&reg->semaphore, 0);
221                         break;
222                 case 0x12:
223                         mb[0] = MSW(stat);
224                         mb[1] = RD_MAILBOX_REG(ha, reg, 1);
225                         mb[2] = RD_MAILBOX_REG(ha, reg, 2);
226                         mb[3] = RD_MAILBOX_REG(ha, reg, 3);
227                         qla2x00_async_event(vha, rsp, mb);
228                         break;
229                 case 0x13:
230                         qla2x00_process_response_queue(rsp);
231                         break;
232                 case 0x15:
233                         mb[0] = MBA_CMPLT_1_16BIT;
234                         mb[1] = MSW(stat);
235                         qla2x00_async_event(vha, rsp, mb);
236                         break;
237                 case 0x16:
238                         mb[0] = MBA_SCSI_COMPLETION;
239                         mb[1] = MSW(stat);
240                         mb[2] = RD_MAILBOX_REG(ha, reg, 2);
241                         qla2x00_async_event(vha, rsp, mb);
242                         break;
243                 default:
244                         ql_dbg(ql_dbg_async, vha, 0x5028,
245                             "Unrecognized interrupt type (%d).\n", stat & 0xff);
246                         break;
247                 }
248                 WRT_REG_WORD(&reg->hccr, HCCR_CLR_RISC_INT);
249                 RD_REG_WORD_RELAXED(&reg->hccr);
250         }
251         qla2x00_handle_mbx_completion(ha, status);
252         spin_unlock_irqrestore(&ha->hardware_lock, flags);
253
254         return (IRQ_HANDLED);
255 }
256
257 /**
258  * qla2x00_mbx_completion() - Process mailbox command completions.
259  * @ha: SCSI driver HA context
260  * @mb0: Mailbox0 register
261  */
262 static void
263 qla2x00_mbx_completion(scsi_qla_host_t *vha, uint16_t mb0)
264 {
265         uint16_t        cnt;
266         uint32_t        mboxes;
267         uint16_t __iomem *wptr;
268         struct qla_hw_data *ha = vha->hw;
269         struct device_reg_2xxx __iomem *reg = &ha->iobase->isp;
270
271         /* Read all mbox registers? */
272         mboxes = (1 << ha->mbx_count) - 1;
273         if (!ha->mcp)
274                 ql_dbg(ql_dbg_async, vha, 0x5001, "MBX pointer ERROR.\n");
275         else
276                 mboxes = ha->mcp->in_mb;
277
278         /* Load return mailbox registers. */
279         ha->flags.mbox_int = 1;
280         ha->mailbox_out[0] = mb0;
281         mboxes >>= 1;
282         wptr = (uint16_t __iomem *)MAILBOX_REG(ha, reg, 1);
283
284         for (cnt = 1; cnt < ha->mbx_count; cnt++) {
285                 if (IS_QLA2200(ha) && cnt == 8)
286                         wptr = (uint16_t __iomem *)MAILBOX_REG(ha, reg, 8);
287                 if ((cnt == 4 || cnt == 5) && (mboxes & BIT_0))
288                         ha->mailbox_out[cnt] = qla2x00_debounce_register(wptr);
289                 else if (mboxes & BIT_0)
290                         ha->mailbox_out[cnt] = RD_REG_WORD(wptr);
291
292                 wptr++;
293                 mboxes >>= 1;
294         }
295 }
296
297 static void
298 qla81xx_idc_event(scsi_qla_host_t *vha, uint16_t aen, uint16_t descr)
299 {
300         static char *event[] =
301                 { "Complete", "Request Notification", "Time Extension" };
302         int rval;
303         struct device_reg_24xx __iomem *reg24 = &vha->hw->iobase->isp24;
304         struct device_reg_82xx __iomem *reg82 = &vha->hw->iobase->isp82;
305         uint16_t __iomem *wptr;
306         uint16_t cnt, timeout, mb[QLA_IDC_ACK_REGS];
307
308         /* Seed data -- mailbox1 -> mailbox7. */
309         if (IS_QLA81XX(vha->hw) || IS_QLA83XX(vha->hw))
310                 wptr = (uint16_t __iomem *)&reg24->mailbox1;
311         else if (IS_QLA8044(vha->hw))
312                 wptr = (uint16_t __iomem *)&reg82->mailbox_out[1];
313         else
314                 return;
315
316         for (cnt = 0; cnt < QLA_IDC_ACK_REGS; cnt++, wptr++)
317                 mb[cnt] = RD_REG_WORD(wptr);
318
319         ql_dbg(ql_dbg_async, vha, 0x5021,
320             "Inter-Driver Communication %s -- "
321             "%04x %04x %04x %04x %04x %04x %04x.\n",
322             event[aen & 0xff], mb[0], mb[1], mb[2], mb[3],
323             mb[4], mb[5], mb[6]);
324         switch (aen) {
325         /* Handle IDC Error completion case. */
326         case MBA_IDC_COMPLETE:
327                 if (mb[1] >> 15) {
328                         vha->hw->flags.idc_compl_status = 1;
329                         if (vha->hw->notify_dcbx_comp && !vha->vp_idx)
330                                 complete(&vha->hw->dcbx_comp);
331                 }
332                 break;
333
334         case MBA_IDC_NOTIFY:
335                 /* Acknowledgement needed? [Notify && non-zero timeout]. */
336                 timeout = (descr >> 8) & 0xf;
337                 ql_dbg(ql_dbg_async, vha, 0x5022,
338                     "%lu Inter-Driver Communication %s -- ACK timeout=%d.\n",
339                     vha->host_no, event[aen & 0xff], timeout);
340
341                 if (!timeout)
342                         return;
343                 rval = qla2x00_post_idc_ack_work(vha, mb);
344                 if (rval != QLA_SUCCESS)
345                         ql_log(ql_log_warn, vha, 0x5023,
346                             "IDC failed to post ACK.\n");
347                 break;
348         case MBA_IDC_TIME_EXT:
349                 vha->hw->idc_extend_tmo = descr;
350                 ql_dbg(ql_dbg_async, vha, 0x5087,
351                     "%lu Inter-Driver Communication %s -- "
352                     "Extend timeout by=%d.\n",
353                     vha->host_no, event[aen & 0xff], vha->hw->idc_extend_tmo);
354                 break;
355         }
356 }
357
358 #define LS_UNKNOWN      2
359 const char *
360 qla2x00_get_link_speed_str(struct qla_hw_data *ha, uint16_t speed)
361 {
362         static const char *const link_speeds[] = {
363                 "1", "2", "?", "4", "8", "16", "32", "10"
364         };
365 #define QLA_LAST_SPEED  7
366
367         if (IS_QLA2100(ha) || IS_QLA2200(ha))
368                 return link_speeds[0];
369         else if (speed == 0x13)
370                 return link_speeds[QLA_LAST_SPEED];
371         else if (speed < QLA_LAST_SPEED)
372                 return link_speeds[speed];
373         else
374                 return link_speeds[LS_UNKNOWN];
375 }
376
377 static void
378 qla83xx_handle_8200_aen(scsi_qla_host_t *vha, uint16_t *mb)
379 {
380         struct qla_hw_data *ha = vha->hw;
381
382         /*
383          * 8200 AEN Interpretation:
384          * mb[0] = AEN code
385          * mb[1] = AEN Reason code
386          * mb[2] = LSW of Peg-Halt Status-1 Register
387          * mb[6] = MSW of Peg-Halt Status-1 Register
388          * mb[3] = LSW of Peg-Halt Status-2 register
389          * mb[7] = MSW of Peg-Halt Status-2 register
390          * mb[4] = IDC Device-State Register value
391          * mb[5] = IDC Driver-Presence Register value
392          */
393         ql_dbg(ql_dbg_async, vha, 0x506b, "AEN Code: mb[0] = 0x%x AEN reason: "
394             "mb[1] = 0x%x PH-status1: mb[2] = 0x%x PH-status1: mb[6] = 0x%x.\n",
395             mb[0], mb[1], mb[2], mb[6]);
396         ql_dbg(ql_dbg_async, vha, 0x506c, "PH-status2: mb[3] = 0x%x "
397             "PH-status2: mb[7] = 0x%x Device-State: mb[4] = 0x%x "
398             "Drv-Presence: mb[5] = 0x%x.\n", mb[3], mb[7], mb[4], mb[5]);
399
400         if (mb[1] & (IDC_PEG_HALT_STATUS_CHANGE | IDC_NIC_FW_REPORTED_FAILURE |
401                                 IDC_HEARTBEAT_FAILURE)) {
402                 ha->flags.nic_core_hung = 1;
403                 ql_log(ql_log_warn, vha, 0x5060,
404                     "83XX: F/W Error Reported: Check if reset required.\n");
405
406                 if (mb[1] & IDC_PEG_HALT_STATUS_CHANGE) {
407                         uint32_t protocol_engine_id, fw_err_code, err_level;
408
409                         /*
410                          * IDC_PEG_HALT_STATUS_CHANGE interpretation:
411                          *  - PEG-Halt Status-1 Register:
412                          *      (LSW = mb[2], MSW = mb[6])
413                          *      Bits 0-7   = protocol-engine ID
414                          *      Bits 8-28  = f/w error code
415                          *      Bits 29-31 = Error-level
416                          *          Error-level 0x1 = Non-Fatal error
417                          *          Error-level 0x2 = Recoverable Fatal error
418                          *          Error-level 0x4 = UnRecoverable Fatal error
419                          *  - PEG-Halt Status-2 Register:
420                          *      (LSW = mb[3], MSW = mb[7])
421                          */
422                         protocol_engine_id = (mb[2] & 0xff);
423                         fw_err_code = (((mb[2] & 0xff00) >> 8) |
424                             ((mb[6] & 0x1fff) << 8));
425                         err_level = ((mb[6] & 0xe000) >> 13);
426                         ql_log(ql_log_warn, vha, 0x5061, "PegHalt Status-1 "
427                             "Register: protocol_engine_id=0x%x "
428                             "fw_err_code=0x%x err_level=0x%x.\n",
429                             protocol_engine_id, fw_err_code, err_level);
430                         ql_log(ql_log_warn, vha, 0x5062, "PegHalt Status-2 "
431                             "Register: 0x%x%x.\n", mb[7], mb[3]);
432                         if (err_level == ERR_LEVEL_NON_FATAL) {
433                                 ql_log(ql_log_warn, vha, 0x5063,
434                                     "Not a fatal error, f/w has recovered "
435                                     "iteself.\n");
436                         } else if (err_level == ERR_LEVEL_RECOVERABLE_FATAL) {
437                                 ql_log(ql_log_fatal, vha, 0x5064,
438                                     "Recoverable Fatal error: Chip reset "
439                                     "required.\n");
440                                 qla83xx_schedule_work(vha,
441                                     QLA83XX_NIC_CORE_RESET);
442                         } else if (err_level == ERR_LEVEL_UNRECOVERABLE_FATAL) {
443                                 ql_log(ql_log_fatal, vha, 0x5065,
444                                     "Unrecoverable Fatal error: Set FAILED "
445                                     "state, reboot required.\n");
446                                 qla83xx_schedule_work(vha,
447                                     QLA83XX_NIC_CORE_UNRECOVERABLE);
448                         }
449                 }
450
451                 if (mb[1] & IDC_NIC_FW_REPORTED_FAILURE) {
452                         uint16_t peg_fw_state, nw_interface_link_up;
453                         uint16_t nw_interface_signal_detect, sfp_status;
454                         uint16_t htbt_counter, htbt_monitor_enable;
455                         uint16_t sfp_additonal_info, sfp_multirate;
456                         uint16_t sfp_tx_fault, link_speed, dcbx_status;
457
458                         /*
459                          * IDC_NIC_FW_REPORTED_FAILURE interpretation:
460                          *  - PEG-to-FC Status Register:
461                          *      (LSW = mb[2], MSW = mb[6])
462                          *      Bits 0-7   = Peg-Firmware state
463                          *      Bit 8      = N/W Interface Link-up
464                          *      Bit 9      = N/W Interface signal detected
465                          *      Bits 10-11 = SFP Status
466                          *        SFP Status 0x0 = SFP+ transceiver not expected
467                          *        SFP Status 0x1 = SFP+ transceiver not present
468                          *        SFP Status 0x2 = SFP+ transceiver invalid
469                          *        SFP Status 0x3 = SFP+ transceiver present and
470                          *        valid
471                          *      Bits 12-14 = Heartbeat Counter
472                          *      Bit 15     = Heartbeat Monitor Enable
473                          *      Bits 16-17 = SFP Additional Info
474                          *        SFP info 0x0 = Unregocnized transceiver for
475                          *        Ethernet
476                          *        SFP info 0x1 = SFP+ brand validation failed
477                          *        SFP info 0x2 = SFP+ speed validation failed
478                          *        SFP info 0x3 = SFP+ access error
479                          *      Bit 18     = SFP Multirate
480                          *      Bit 19     = SFP Tx Fault
481                          *      Bits 20-22 = Link Speed
482                          *      Bits 23-27 = Reserved
483                          *      Bits 28-30 = DCBX Status
484                          *        DCBX Status 0x0 = DCBX Disabled
485                          *        DCBX Status 0x1 = DCBX Enabled
486                          *        DCBX Status 0x2 = DCBX Exchange error
487                          *      Bit 31     = Reserved
488                          */
489                         peg_fw_state = (mb[2] & 0x00ff);
490                         nw_interface_link_up = ((mb[2] & 0x0100) >> 8);
491                         nw_interface_signal_detect = ((mb[2] & 0x0200) >> 9);
492                         sfp_status = ((mb[2] & 0x0c00) >> 10);
493                         htbt_counter = ((mb[2] & 0x7000) >> 12);
494                         htbt_monitor_enable = ((mb[2] & 0x8000) >> 15);
495                         sfp_additonal_info = (mb[6] & 0x0003);
496                         sfp_multirate = ((mb[6] & 0x0004) >> 2);
497                         sfp_tx_fault = ((mb[6] & 0x0008) >> 3);
498                         link_speed = ((mb[6] & 0x0070) >> 4);
499                         dcbx_status = ((mb[6] & 0x7000) >> 12);
500
501                         ql_log(ql_log_warn, vha, 0x5066,
502                             "Peg-to-Fc Status Register:\n"
503                             "peg_fw_state=0x%x, nw_interface_link_up=0x%x, "
504                             "nw_interface_signal_detect=0x%x"
505                             "\nsfp_statis=0x%x.\n ", peg_fw_state,
506                             nw_interface_link_up, nw_interface_signal_detect,
507                             sfp_status);
508                         ql_log(ql_log_warn, vha, 0x5067,
509                             "htbt_counter=0x%x, htbt_monitor_enable=0x%x, "
510                             "sfp_additonal_info=0x%x, sfp_multirate=0x%x.\n ",
511                             htbt_counter, htbt_monitor_enable,
512                             sfp_additonal_info, sfp_multirate);
513                         ql_log(ql_log_warn, vha, 0x5068,
514                             "sfp_tx_fault=0x%x, link_state=0x%x, "
515                             "dcbx_status=0x%x.\n", sfp_tx_fault, link_speed,
516                             dcbx_status);
517
518                         qla83xx_schedule_work(vha, QLA83XX_NIC_CORE_RESET);
519                 }
520
521                 if (mb[1] & IDC_HEARTBEAT_FAILURE) {
522                         ql_log(ql_log_warn, vha, 0x5069,
523                             "Heartbeat Failure encountered, chip reset "
524                             "required.\n");
525
526                         qla83xx_schedule_work(vha, QLA83XX_NIC_CORE_RESET);
527                 }
528         }
529
530         if (mb[1] & IDC_DEVICE_STATE_CHANGE) {
531                 ql_log(ql_log_info, vha, 0x506a,
532                     "IDC Device-State changed = 0x%x.\n", mb[4]);
533                 if (ha->flags.nic_core_reset_owner)
534                         return;
535                 qla83xx_schedule_work(vha, MBA_IDC_AEN);
536         }
537 }
538
539 int
540 qla2x00_is_a_vp_did(scsi_qla_host_t *vha, uint32_t rscn_entry)
541 {
542         struct qla_hw_data *ha = vha->hw;
543         scsi_qla_host_t *vp;
544         uint32_t vp_did;
545         unsigned long flags;
546         int ret = 0;
547
548         if (!ha->num_vhosts)
549                 return ret;
550
551         spin_lock_irqsave(&ha->vport_slock, flags);
552         list_for_each_entry(vp, &ha->vp_list, list) {
553                 vp_did = vp->d_id.b24;
554                 if (vp_did == rscn_entry) {
555                         ret = 1;
556                         break;
557                 }
558         }
559         spin_unlock_irqrestore(&ha->vport_slock, flags);
560
561         return ret;
562 }
563
564 fc_port_t *
565 qla2x00_find_fcport_by_loopid(scsi_qla_host_t *vha, uint16_t loop_id)
566 {
567         fc_port_t *f, *tf;
568
569         f = tf = NULL;
570         list_for_each_entry_safe(f, tf, &vha->vp_fcports, list)
571                 if (f->loop_id == loop_id)
572                         return f;
573         return NULL;
574 }
575
576 fc_port_t *
577 qla2x00_find_fcport_by_wwpn(scsi_qla_host_t *vha, u8 *wwpn, u8 incl_deleted)
578 {
579         fc_port_t *f, *tf;
580
581         f = tf = NULL;
582         list_for_each_entry_safe(f, tf, &vha->vp_fcports, list) {
583                 if (memcmp(f->port_name, wwpn, WWN_SIZE) == 0) {
584                         if (incl_deleted)
585                                 return f;
586                         else if (f->deleted == 0)
587                                 return f;
588                 }
589         }
590         return NULL;
591 }
592
593 fc_port_t *
594 qla2x00_find_fcport_by_nportid(scsi_qla_host_t *vha, port_id_t *id,
595         u8 incl_deleted)
596 {
597         fc_port_t *f, *tf;
598
599         f = tf = NULL;
600         list_for_each_entry_safe(f, tf, &vha->vp_fcports, list) {
601                 if (f->d_id.b24 == id->b24) {
602                         if (incl_deleted)
603                                 return f;
604                         else if (f->deleted == 0)
605                                 return f;
606                 }
607         }
608         return NULL;
609 }
610
611 /**
612  * qla2x00_async_event() - Process aynchronous events.
613  * @ha: SCSI driver HA context
614  * @mb: Mailbox registers (0 - 3)
615  */
616 void
617 qla2x00_async_event(scsi_qla_host_t *vha, struct rsp_que *rsp, uint16_t *mb)
618 {
619         uint16_t        handle_cnt;
620         uint16_t        cnt, mbx;
621         uint32_t        handles[5];
622         struct qla_hw_data *ha = vha->hw;
623         struct device_reg_2xxx __iomem *reg = &ha->iobase->isp;
624         struct device_reg_24xx __iomem *reg24 = &ha->iobase->isp24;
625         struct device_reg_82xx __iomem *reg82 = &ha->iobase->isp82;
626         uint32_t        rscn_entry, host_pid;
627         unsigned long   flags;
628         fc_port_t       *fcport = NULL;
629
630         /* Setup to process RIO completion. */
631         handle_cnt = 0;
632         if (IS_CNA_CAPABLE(ha))
633                 goto skip_rio;
634         switch (mb[0]) {
635         case MBA_SCSI_COMPLETION:
636                 handles[0] = le32_to_cpu((uint32_t)((mb[2] << 16) | mb[1]));
637                 handle_cnt = 1;
638                 break;
639         case MBA_CMPLT_1_16BIT:
640                 handles[0] = mb[1];
641                 handle_cnt = 1;
642                 mb[0] = MBA_SCSI_COMPLETION;
643                 break;
644         case MBA_CMPLT_2_16BIT:
645                 handles[0] = mb[1];
646                 handles[1] = mb[2];
647                 handle_cnt = 2;
648                 mb[0] = MBA_SCSI_COMPLETION;
649                 break;
650         case MBA_CMPLT_3_16BIT:
651                 handles[0] = mb[1];
652                 handles[1] = mb[2];
653                 handles[2] = mb[3];
654                 handle_cnt = 3;
655                 mb[0] = MBA_SCSI_COMPLETION;
656                 break;
657         case MBA_CMPLT_4_16BIT:
658                 handles[0] = mb[1];
659                 handles[1] = mb[2];
660                 handles[2] = mb[3];
661                 handles[3] = (uint32_t)RD_MAILBOX_REG(ha, reg, 6);
662                 handle_cnt = 4;
663                 mb[0] = MBA_SCSI_COMPLETION;
664                 break;
665         case MBA_CMPLT_5_16BIT:
666                 handles[0] = mb[1];
667                 handles[1] = mb[2];
668                 handles[2] = mb[3];
669                 handles[3] = (uint32_t)RD_MAILBOX_REG(ha, reg, 6);
670                 handles[4] = (uint32_t)RD_MAILBOX_REG(ha, reg, 7);
671                 handle_cnt = 5;
672                 mb[0] = MBA_SCSI_COMPLETION;
673                 break;
674         case MBA_CMPLT_2_32BIT:
675                 handles[0] = le32_to_cpu((uint32_t)((mb[2] << 16) | mb[1]));
676                 handles[1] = le32_to_cpu(
677                     ((uint32_t)(RD_MAILBOX_REG(ha, reg, 7) << 16)) |
678                     RD_MAILBOX_REG(ha, reg, 6));
679                 handle_cnt = 2;
680                 mb[0] = MBA_SCSI_COMPLETION;
681                 break;
682         default:
683                 break;
684         }
685 skip_rio:
686         switch (mb[0]) {
687         case MBA_SCSI_COMPLETION:       /* Fast Post */
688                 if (!vha->flags.online)
689                         break;
690
691                 for (cnt = 0; cnt < handle_cnt; cnt++)
692                         qla2x00_process_completed_request(vha, rsp->req,
693                                 handles[cnt]);
694                 break;
695
696         case MBA_RESET:                 /* Reset */
697                 ql_dbg(ql_dbg_async, vha, 0x5002,
698                     "Asynchronous RESET.\n");
699
700                 set_bit(RESET_MARKER_NEEDED, &vha->dpc_flags);
701                 break;
702
703         case MBA_SYSTEM_ERR:            /* System Error */
704                 mbx = (IS_QLA81XX(ha) || IS_QLA83XX(ha) || IS_QLA27XX(ha)) ?
705                         RD_REG_WORD(&reg24->mailbox7) : 0;
706                 ql_log(ql_log_warn, vha, 0x5003,
707                     "ISP System Error - mbx1=%xh mbx2=%xh mbx3=%xh "
708                     "mbx7=%xh.\n", mb[1], mb[2], mb[3], mbx);
709
710                 ha->isp_ops->fw_dump(vha, 1);
711                 ha->flags.fw_init_done = 0;
712                 ha->flags.fw_started = 0;
713
714                 if (IS_FWI2_CAPABLE(ha)) {
715                         if (mb[1] == 0 && mb[2] == 0) {
716                                 ql_log(ql_log_fatal, vha, 0x5004,
717                                     "Unrecoverable Hardware Error: adapter "
718                                     "marked OFFLINE!\n");
719                                 vha->flags.online = 0;
720                                 vha->device_flags |= DFLG_DEV_FAILED;
721                         } else {
722                                 /* Check to see if MPI timeout occurred */
723                                 if ((mbx & MBX_3) && (ha->port_no == 0))
724                                         set_bit(MPI_RESET_NEEDED,
725                                             &vha->dpc_flags);
726
727                                 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
728                         }
729                 } else if (mb[1] == 0) {
730                         ql_log(ql_log_fatal, vha, 0x5005,
731                             "Unrecoverable Hardware Error: adapter marked "
732                             "OFFLINE!\n");
733                         vha->flags.online = 0;
734                         vha->device_flags |= DFLG_DEV_FAILED;
735                 } else
736                         set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
737                 break;
738
739         case MBA_REQ_TRANSFER_ERR:      /* Request Transfer Error */
740                 ql_log(ql_log_warn, vha, 0x5006,
741                     "ISP Request Transfer Error (%x).\n",  mb[1]);
742
743                 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
744                 break;
745
746         case MBA_RSP_TRANSFER_ERR:      /* Response Transfer Error */
747                 ql_log(ql_log_warn, vha, 0x5007,
748                     "ISP Response Transfer Error (%x).\n", mb[1]);
749
750                 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
751                 break;
752
753         case MBA_WAKEUP_THRES:          /* Request Queue Wake-up */
754                 ql_dbg(ql_dbg_async, vha, 0x5008,
755                     "Asynchronous WAKEUP_THRES (%x).\n", mb[1]);
756                 break;
757
758         case MBA_LOOP_INIT_ERR:
759                 ql_log(ql_log_warn, vha, 0x5090,
760                     "LOOP INIT ERROR (%x).\n", mb[1]);
761                 ha->isp_ops->fw_dump(vha, 1);
762                 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
763                 break;
764
765         case MBA_LIP_OCCURRED:          /* Loop Initialization Procedure */
766                 ha->flags.lip_ae = 1;
767                 ha->flags.n2n_ae = 0;
768
769                 ql_dbg(ql_dbg_async, vha, 0x5009,
770                     "LIP occurred (%x).\n", mb[1]);
771
772                 if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
773                         atomic_set(&vha->loop_state, LOOP_DOWN);
774                         atomic_set(&vha->loop_down_timer, LOOP_DOWN_TIME);
775                         qla2x00_mark_all_devices_lost(vha, 1);
776                 }
777
778                 if (vha->vp_idx) {
779                         atomic_set(&vha->vp_state, VP_FAILED);
780                         fc_vport_set_state(vha->fc_vport, FC_VPORT_FAILED);
781                 }
782
783                 set_bit(REGISTER_FC4_NEEDED, &vha->dpc_flags);
784                 set_bit(REGISTER_FDMI_NEEDED, &vha->dpc_flags);
785
786                 vha->flags.management_server_logged_in = 0;
787                 qla2x00_post_aen_work(vha, FCH_EVT_LIP, mb[1]);
788                 break;
789
790         case MBA_LOOP_UP:               /* Loop Up Event */
791                 if (IS_QLA2100(ha) || IS_QLA2200(ha))
792                         ha->link_data_rate = PORT_SPEED_1GB;
793                 else
794                         ha->link_data_rate = mb[1];
795
796                 ql_log(ql_log_info, vha, 0x500a,
797                     "LOOP UP detected (%s Gbps).\n",
798                     qla2x00_get_link_speed_str(ha, ha->link_data_rate));
799
800                 vha->flags.management_server_logged_in = 0;
801                 qla2x00_post_aen_work(vha, FCH_EVT_LINKUP, ha->link_data_rate);
802                 break;
803
804         case MBA_LOOP_DOWN:             /* Loop Down Event */
805                 ha->flags.n2n_ae = 0;
806                 ha->flags.lip_ae = 0;
807                 ha->current_topology = 0;
808
809                 mbx = (IS_QLA81XX(ha) || IS_QLA8031(ha))
810                         ? RD_REG_WORD(&reg24->mailbox4) : 0;
811                 mbx = (IS_P3P_TYPE(ha)) ? RD_REG_WORD(&reg82->mailbox_out[4])
812                         : mbx;
813                 ql_log(ql_log_info, vha, 0x500b,
814                     "LOOP DOWN detected (%x %x %x %x).\n",
815                     mb[1], mb[2], mb[3], mbx);
816
817                 if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
818                         atomic_set(&vha->loop_state, LOOP_DOWN);
819                         atomic_set(&vha->loop_down_timer, LOOP_DOWN_TIME);
820                         /*
821                          * In case of loop down, restore WWPN from
822                          * NVRAM in case of FA-WWPN capable ISP
823                          * Restore for Physical Port only
824                          */
825                         if (!vha->vp_idx) {
826                                 if (ha->flags.fawwpn_enabled) {
827                                         void *wwpn = ha->init_cb->port_name;
828                                         memcpy(vha->port_name, wwpn, WWN_SIZE);
829                                         fc_host_port_name(vha->host) =
830                                             wwn_to_u64(vha->port_name);
831                                         ql_dbg(ql_dbg_init + ql_dbg_verbose,
832                                             vha, 0x0144, "LOOP DOWN detected,"
833                                             "restore WWPN %016llx\n",
834                                             wwn_to_u64(vha->port_name));
835                                 }
836
837                                 clear_bit(VP_CONFIG_OK, &vha->vp_flags);
838                         }
839
840                         vha->device_flags |= DFLG_NO_CABLE;
841                         qla2x00_mark_all_devices_lost(vha, 1);
842                 }
843
844                 if (vha->vp_idx) {
845                         atomic_set(&vha->vp_state, VP_FAILED);
846                         fc_vport_set_state(vha->fc_vport, FC_VPORT_FAILED);
847                 }
848
849                 vha->flags.management_server_logged_in = 0;
850                 ha->link_data_rate = PORT_SPEED_UNKNOWN;
851                 qla2x00_post_aen_work(vha, FCH_EVT_LINKDOWN, 0);
852                 break;
853
854         case MBA_LIP_RESET:             /* LIP reset occurred */
855                 ql_dbg(ql_dbg_async, vha, 0x500c,
856                     "LIP reset occurred (%x).\n", mb[1]);
857
858                 if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
859                         atomic_set(&vha->loop_state, LOOP_DOWN);
860                         atomic_set(&vha->loop_down_timer, LOOP_DOWN_TIME);
861                         qla2x00_mark_all_devices_lost(vha, 1);
862                 }
863
864                 if (vha->vp_idx) {
865                         atomic_set(&vha->vp_state, VP_FAILED);
866                         fc_vport_set_state(vha->fc_vport, FC_VPORT_FAILED);
867                 }
868
869                 set_bit(RESET_MARKER_NEEDED, &vha->dpc_flags);
870
871                 ha->operating_mode = LOOP;
872                 vha->flags.management_server_logged_in = 0;
873                 qla2x00_post_aen_work(vha, FCH_EVT_LIPRESET, mb[1]);
874                 break;
875
876         /* case MBA_DCBX_COMPLETE: */
877         case MBA_POINT_TO_POINT:        /* Point-to-Point */
878                 ha->flags.lip_ae = 0;
879                 ha->flags.n2n_ae = 1;
880
881                 if (IS_QLA2100(ha))
882                         break;
883
884                 if (IS_CNA_CAPABLE(ha)) {
885                         ql_dbg(ql_dbg_async, vha, 0x500d,
886                             "DCBX Completed -- %04x %04x %04x.\n",
887                             mb[1], mb[2], mb[3]);
888                         if (ha->notify_dcbx_comp && !vha->vp_idx)
889                                 complete(&ha->dcbx_comp);
890
891                 } else
892                         ql_dbg(ql_dbg_async, vha, 0x500e,
893                             "Asynchronous P2P MODE received.\n");
894
895                 /*
896                  * Until there's a transition from loop down to loop up, treat
897                  * this as loop down only.
898                  */
899                 if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
900                         atomic_set(&vha->loop_state, LOOP_DOWN);
901                         if (!atomic_read(&vha->loop_down_timer))
902                                 atomic_set(&vha->loop_down_timer,
903                                     LOOP_DOWN_TIME);
904                         qla2x00_mark_all_devices_lost(vha, 1);
905                 }
906
907                 if (vha->vp_idx) {
908                         atomic_set(&vha->vp_state, VP_FAILED);
909                         fc_vport_set_state(vha->fc_vport, FC_VPORT_FAILED);
910                 }
911
912                 if (!(test_bit(ABORT_ISP_ACTIVE, &vha->dpc_flags)))
913                         set_bit(RESET_MARKER_NEEDED, &vha->dpc_flags);
914
915                 set_bit(REGISTER_FC4_NEEDED, &vha->dpc_flags);
916                 set_bit(REGISTER_FDMI_NEEDED, &vha->dpc_flags);
917
918                 ha->flags.gpsc_supported = 1;
919                 vha->flags.management_server_logged_in = 0;
920                 break;
921
922         case MBA_CHG_IN_CONNECTION:     /* Change in connection mode */
923                 if (IS_QLA2100(ha))
924                         break;
925
926                 ql_dbg(ql_dbg_async, vha, 0x500f,
927                     "Configuration change detected: value=%x.\n", mb[1]);
928
929                 if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
930                         atomic_set(&vha->loop_state, LOOP_DOWN);
931                         if (!atomic_read(&vha->loop_down_timer))
932                                 atomic_set(&vha->loop_down_timer,
933                                     LOOP_DOWN_TIME);
934                         qla2x00_mark_all_devices_lost(vha, 1);
935                 }
936
937                 if (vha->vp_idx) {
938                         atomic_set(&vha->vp_state, VP_FAILED);
939                         fc_vport_set_state(vha->fc_vport, FC_VPORT_FAILED);
940                 }
941
942                 set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags);
943                 set_bit(LOCAL_LOOP_UPDATE, &vha->dpc_flags);
944                 break;
945
946         case MBA_PORT_UPDATE:           /* Port database update */
947                 /*
948                  * Handle only global and vn-port update events
949                  *
950                  * Relevant inputs:
951                  * mb[1] = N_Port handle of changed port
952                  * OR 0xffff for global event
953                  * mb[2] = New login state
954                  * 7 = Port logged out
955                  * mb[3] = LSB is vp_idx, 0xff = all vps
956                  *
957                  * Skip processing if:
958                  *       Event is global, vp_idx is NOT all vps,
959                  *           vp_idx does not match
960                  *       Event is not global, vp_idx does not match
961                  */
962                 if (IS_QLA2XXX_MIDTYPE(ha) &&
963                     ((mb[1] == 0xffff && (mb[3] & 0xff) != 0xff) ||
964                         (mb[1] != 0xffff)) && vha->vp_idx != (mb[3] & 0xff))
965                         break;
966
967                 if (mb[2] == 0x7) {
968                         ql_dbg(ql_dbg_async, vha, 0x5010,
969                             "Port %s %04x %04x %04x.\n",
970                             mb[1] == 0xffff ? "unavailable" : "logout",
971                             mb[1], mb[2], mb[3]);
972
973                         if (mb[1] == 0xffff)
974                                 goto global_port_update;
975
976                         /* Port logout */
977                         fcport = qla2x00_find_fcport_by_loopid(vha, mb[1]);
978                         if (!fcport)
979                                 break;
980                         if (atomic_read(&fcport->state) != FCS_ONLINE)
981                                 break;
982                         ql_dbg(ql_dbg_async, vha, 0x508a,
983                             "Marking port lost loopid=%04x portid=%06x.\n",
984                             fcport->loop_id, fcport->d_id.b24);
985                         if (qla_ini_mode_enabled(vha)) {
986                                 qla2x00_mark_device_lost(fcport->vha, fcport, 1, 1);
987                                 fcport->logout_on_delete = 0;
988                                 qlt_schedule_sess_for_deletion_lock(fcport);
989                         }
990                         break;
991
992 global_port_update:
993                         if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
994                                 atomic_set(&vha->loop_state, LOOP_DOWN);
995                                 atomic_set(&vha->loop_down_timer,
996                                     LOOP_DOWN_TIME);
997                                 vha->device_flags |= DFLG_NO_CABLE;
998                                 qla2x00_mark_all_devices_lost(vha, 1);
999                         }
1000
1001                         if (vha->vp_idx) {
1002                                 atomic_set(&vha->vp_state, VP_FAILED);
1003                                 fc_vport_set_state(vha->fc_vport,
1004                                     FC_VPORT_FAILED);
1005                                 qla2x00_mark_all_devices_lost(vha, 1);
1006                         }
1007
1008                         vha->flags.management_server_logged_in = 0;
1009                         ha->link_data_rate = PORT_SPEED_UNKNOWN;
1010                         break;
1011                 }
1012
1013                 /*
1014                  * If PORT UPDATE is global (received LIP_OCCURRED/LIP_RESET
1015                  * event etc. earlier indicating loop is down) then process
1016                  * it.  Otherwise ignore it and Wait for RSCN to come in.
1017                  */
1018                 atomic_set(&vha->loop_down_timer, 0);
1019                 if (atomic_read(&vha->loop_state) != LOOP_DOWN &&
1020                     atomic_read(&vha->loop_state) != LOOP_DEAD) {
1021                         ql_dbg(ql_dbg_async, vha, 0x5011,
1022                             "Asynchronous PORT UPDATE ignored %04x/%04x/%04x.\n",
1023                             mb[1], mb[2], mb[3]);
1024
1025                         qlt_async_event(mb[0], vha, mb);
1026                         break;
1027                 }
1028
1029                 ql_dbg(ql_dbg_async, vha, 0x5012,
1030                     "Port database changed %04x %04x %04x.\n",
1031                     mb[1], mb[2], mb[3]);
1032
1033                 /*
1034                  * Mark all devices as missing so we will login again.
1035                  */
1036                 atomic_set(&vha->loop_state, LOOP_UP);
1037
1038                 qla2x00_mark_all_devices_lost(vha, 1);
1039
1040                 set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags);
1041                 set_bit(LOCAL_LOOP_UPDATE, &vha->dpc_flags);
1042                 set_bit(VP_CONFIG_OK, &vha->vp_flags);
1043
1044                 qlt_async_event(mb[0], vha, mb);
1045                 break;
1046
1047         case MBA_RSCN_UPDATE:           /* State Change Registration */
1048                 /* Check if the Vport has issued a SCR */
1049                 if (vha->vp_idx && test_bit(VP_SCR_NEEDED, &vha->vp_flags))
1050                         break;
1051                 /* Only handle SCNs for our Vport index. */
1052                 if (ha->flags.npiv_supported && vha->vp_idx != (mb[3] & 0xff))
1053                         break;
1054
1055                 ql_dbg(ql_dbg_async, vha, 0x5013,
1056                     "RSCN database changed -- %04x %04x %04x.\n",
1057                     mb[1], mb[2], mb[3]);
1058
1059                 rscn_entry = ((mb[1] & 0xff) << 16) | mb[2];
1060                 host_pid = (vha->d_id.b.domain << 16) | (vha->d_id.b.area << 8)
1061                                 | vha->d_id.b.al_pa;
1062                 if (rscn_entry == host_pid) {
1063                         ql_dbg(ql_dbg_async, vha, 0x5014,
1064                             "Ignoring RSCN update to local host "
1065                             "port ID (%06x).\n", host_pid);
1066                         break;
1067                 }
1068
1069                 /* Ignore reserved bits from RSCN-payload. */
1070                 rscn_entry = ((mb[1] & 0x3ff) << 16) | mb[2];
1071
1072                 /* Skip RSCNs for virtual ports on the same physical port */
1073                 if (qla2x00_is_a_vp_did(vha, rscn_entry))
1074                         break;
1075
1076                 atomic_set(&vha->loop_down_timer, 0);
1077                 vha->flags.management_server_logged_in = 0;
1078                 {
1079                         struct event_arg ea;
1080
1081                         memset(&ea, 0, sizeof(ea));
1082                         ea.event = FCME_RSCN;
1083                         ea.id.b24 = rscn_entry;
1084                         ea.id.b.rsvd_1 = rscn_entry >> 24;
1085                         qla2x00_fcport_event_handler(vha, &ea);
1086                         qla2x00_post_aen_work(vha, FCH_EVT_RSCN, rscn_entry);
1087                 }
1088                 break;
1089         /* case MBA_RIO_RESPONSE: */
1090         case MBA_ZIO_RESPONSE:
1091                 ql_dbg(ql_dbg_async, vha, 0x5015,
1092                     "[R|Z]IO update completion.\n");
1093
1094                 if (IS_FWI2_CAPABLE(ha))
1095                         qla24xx_process_response_queue(vha, rsp);
1096                 else
1097                         qla2x00_process_response_queue(rsp);
1098                 break;
1099
1100         case MBA_DISCARD_RND_FRAME:
1101                 ql_dbg(ql_dbg_async, vha, 0x5016,
1102                     "Discard RND Frame -- %04x %04x %04x.\n",
1103                     mb[1], mb[2], mb[3]);
1104                 break;
1105
1106         case MBA_TRACE_NOTIFICATION:
1107                 ql_dbg(ql_dbg_async, vha, 0x5017,
1108                     "Trace Notification -- %04x %04x.\n", mb[1], mb[2]);
1109                 break;
1110
1111         case MBA_ISP84XX_ALERT:
1112                 ql_dbg(ql_dbg_async, vha, 0x5018,
1113                     "ISP84XX Alert Notification -- %04x %04x %04x.\n",
1114                     mb[1], mb[2], mb[3]);
1115
1116                 spin_lock_irqsave(&ha->cs84xx->access_lock, flags);
1117                 switch (mb[1]) {
1118                 case A84_PANIC_RECOVERY:
1119                         ql_log(ql_log_info, vha, 0x5019,
1120                             "Alert 84XX: panic recovery %04x %04x.\n",
1121                             mb[2], mb[3]);
1122                         break;
1123                 case A84_OP_LOGIN_COMPLETE:
1124                         ha->cs84xx->op_fw_version = mb[3] << 16 | mb[2];
1125                         ql_log(ql_log_info, vha, 0x501a,
1126                             "Alert 84XX: firmware version %x.\n",
1127                             ha->cs84xx->op_fw_version);
1128                         break;
1129                 case A84_DIAG_LOGIN_COMPLETE:
1130                         ha->cs84xx->diag_fw_version = mb[3] << 16 | mb[2];
1131                         ql_log(ql_log_info, vha, 0x501b,
1132                             "Alert 84XX: diagnostic firmware version %x.\n",
1133                             ha->cs84xx->diag_fw_version);
1134                         break;
1135                 case A84_GOLD_LOGIN_COMPLETE:
1136                         ha->cs84xx->diag_fw_version = mb[3] << 16 | mb[2];
1137                         ha->cs84xx->fw_update = 1;
1138                         ql_log(ql_log_info, vha, 0x501c,
1139                             "Alert 84XX: gold firmware version %x.\n",
1140                             ha->cs84xx->gold_fw_version);
1141                         break;
1142                 default:
1143                         ql_log(ql_log_warn, vha, 0x501d,
1144                             "Alert 84xx: Invalid Alert %04x %04x %04x.\n",
1145                             mb[1], mb[2], mb[3]);
1146                 }
1147                 spin_unlock_irqrestore(&ha->cs84xx->access_lock, flags);
1148                 break;
1149         case MBA_DCBX_START:
1150                 ql_dbg(ql_dbg_async, vha, 0x501e,
1151                     "DCBX Started -- %04x %04x %04x.\n",
1152                     mb[1], mb[2], mb[3]);
1153                 break;
1154         case MBA_DCBX_PARAM_UPDATE:
1155                 ql_dbg(ql_dbg_async, vha, 0x501f,
1156                     "DCBX Parameters Updated -- %04x %04x %04x.\n",
1157                     mb[1], mb[2], mb[3]);
1158                 break;
1159         case MBA_FCF_CONF_ERR:
1160                 ql_dbg(ql_dbg_async, vha, 0x5020,
1161                     "FCF Configuration Error -- %04x %04x %04x.\n",
1162                     mb[1], mb[2], mb[3]);
1163                 break;
1164         case MBA_IDC_NOTIFY:
1165                 if (IS_QLA8031(vha->hw) || IS_QLA8044(ha)) {
1166                         mb[4] = RD_REG_WORD(&reg24->mailbox4);
1167                         if (((mb[2] & 0x7fff) == MBC_PORT_RESET ||
1168                             (mb[2] & 0x7fff) == MBC_SET_PORT_CONFIG) &&
1169                             (mb[4] & INTERNAL_LOOPBACK_MASK) != 0) {
1170                                 set_bit(ISP_QUIESCE_NEEDED, &vha->dpc_flags);
1171                                 /*
1172                                  * Extend loop down timer since port is active.
1173                                  */
1174                                 if (atomic_read(&vha->loop_state) == LOOP_DOWN)
1175                                         atomic_set(&vha->loop_down_timer,
1176                                             LOOP_DOWN_TIME);
1177                                 qla2xxx_wake_dpc(vha);
1178                         }
1179                 }
1180         case MBA_IDC_COMPLETE:
1181                 if (ha->notify_lb_portup_comp && !vha->vp_idx)
1182                         complete(&ha->lb_portup_comp);
1183                 /* Fallthru */
1184         case MBA_IDC_TIME_EXT:
1185                 if (IS_QLA81XX(vha->hw) || IS_QLA8031(vha->hw) ||
1186                     IS_QLA8044(ha))
1187                         qla81xx_idc_event(vha, mb[0], mb[1]);
1188                 break;
1189
1190         case MBA_IDC_AEN:
1191                 mb[4] = RD_REG_WORD(&reg24->mailbox4);
1192                 mb[5] = RD_REG_WORD(&reg24->mailbox5);
1193                 mb[6] = RD_REG_WORD(&reg24->mailbox6);
1194                 mb[7] = RD_REG_WORD(&reg24->mailbox7);
1195                 qla83xx_handle_8200_aen(vha, mb);
1196                 break;
1197
1198         case MBA_DPORT_DIAGNOSTICS:
1199                 ql_dbg(ql_dbg_async, vha, 0x5052,
1200                     "D-Port Diagnostics: %04x result=%s\n",
1201                     mb[0],
1202                     mb[1] == 0 ? "start" :
1203                     mb[1] == 1 ? "done (pass)" :
1204                     mb[1] == 2 ? "done (error)" : "other");
1205                 break;
1206
1207         case MBA_TEMPERATURE_ALERT:
1208                 ql_dbg(ql_dbg_async, vha, 0x505e,
1209                     "TEMPERATURE ALERT: %04x %04x %04x\n", mb[1], mb[2], mb[3]);
1210                 if (mb[1] == 0x12)
1211                         schedule_work(&ha->board_disable);
1212                 break;
1213
1214         default:
1215                 ql_dbg(ql_dbg_async, vha, 0x5057,
1216                     "Unknown AEN:%04x %04x %04x %04x\n",
1217                     mb[0], mb[1], mb[2], mb[3]);
1218         }
1219
1220         qlt_async_event(mb[0], vha, mb);
1221
1222         if (!vha->vp_idx && ha->num_vhosts)
1223                 qla2x00_alert_all_vps(rsp, mb);
1224 }
1225
1226 /**
1227  * qla2x00_process_completed_request() - Process a Fast Post response.
1228  * @ha: SCSI driver HA context
1229  * @index: SRB index
1230  */
1231 void
1232 qla2x00_process_completed_request(struct scsi_qla_host *vha,
1233                                   struct req_que *req, uint32_t index)
1234 {
1235         srb_t *sp;
1236         struct qla_hw_data *ha = vha->hw;
1237
1238         /* Validate handle. */
1239         if (index >= req->num_outstanding_cmds) {
1240                 ql_log(ql_log_warn, vha, 0x3014,
1241                     "Invalid SCSI command index (%x).\n", index);
1242
1243                 if (IS_P3P_TYPE(ha))
1244                         set_bit(FCOE_CTX_RESET_NEEDED, &vha->dpc_flags);
1245                 else
1246                         set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
1247                 return;
1248         }
1249
1250         sp = req->outstanding_cmds[index];
1251         if (sp) {
1252                 /* Free outstanding command slot. */
1253                 req->outstanding_cmds[index] = NULL;
1254
1255                 /* Save ISP completion status */
1256                 sp->done(sp, DID_OK << 16);
1257         } else {
1258                 ql_log(ql_log_warn, vha, 0x3016, "Invalid SCSI SRB.\n");
1259
1260                 if (IS_P3P_TYPE(ha))
1261                         set_bit(FCOE_CTX_RESET_NEEDED, &vha->dpc_flags);
1262                 else
1263                         set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
1264         }
1265 }
1266
1267 srb_t *
1268 qla2x00_get_sp_from_handle(scsi_qla_host_t *vha, const char *func,
1269     struct req_que *req, void *iocb)
1270 {
1271         struct qla_hw_data *ha = vha->hw;
1272         sts_entry_t *pkt = iocb;
1273         srb_t *sp = NULL;
1274         uint16_t index;
1275
1276         index = LSW(pkt->handle);
1277         if (index >= req->num_outstanding_cmds) {
1278                 ql_log(ql_log_warn, vha, 0x5031,
1279                            "Invalid command index (%x) type %8ph.\n",
1280                            index, iocb);
1281                 if (IS_P3P_TYPE(ha))
1282                         set_bit(FCOE_CTX_RESET_NEEDED, &vha->dpc_flags);
1283                 else
1284                         set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
1285                 goto done;
1286         }
1287         sp = req->outstanding_cmds[index];
1288         if (!sp) {
1289                 ql_log(ql_log_warn, vha, 0x5032,
1290                     "Invalid completion handle (%x) -- timed-out.\n", index);
1291                 return sp;
1292         }
1293         if (sp->handle != index) {
1294                 ql_log(ql_log_warn, vha, 0x5033,
1295                     "SRB handle (%x) mismatch %x.\n", sp->handle, index);
1296                 return NULL;
1297         }
1298
1299         req->outstanding_cmds[index] = NULL;
1300
1301 done:
1302         return sp;
1303 }
1304
1305 static void
1306 qla2x00_mbx_iocb_entry(scsi_qla_host_t *vha, struct req_que *req,
1307     struct mbx_entry *mbx)
1308 {
1309         const char func[] = "MBX-IOCB";
1310         const char *type;
1311         fc_port_t *fcport;
1312         srb_t *sp;
1313         struct srb_iocb *lio;
1314         uint16_t *data;
1315         uint16_t status;
1316
1317         sp = qla2x00_get_sp_from_handle(vha, func, req, mbx);
1318         if (!sp)
1319                 return;
1320
1321         lio = &sp->u.iocb_cmd;
1322         type = sp->name;
1323         fcport = sp->fcport;
1324         data = lio->u.logio.data;
1325
1326         data[0] = MBS_COMMAND_ERROR;
1327         data[1] = lio->u.logio.flags & SRB_LOGIN_RETRIED ?
1328             QLA_LOGIO_LOGIN_RETRIED : 0;
1329         if (mbx->entry_status) {
1330                 ql_dbg(ql_dbg_async, vha, 0x5043,
1331                     "Async-%s error entry - hdl=%x portid=%02x%02x%02x "
1332                     "entry-status=%x status=%x state-flag=%x "
1333                     "status-flags=%x.\n", type, sp->handle,
1334                     fcport->d_id.b.domain, fcport->d_id.b.area,
1335                     fcport->d_id.b.al_pa, mbx->entry_status,
1336                     le16_to_cpu(mbx->status), le16_to_cpu(mbx->state_flags),
1337                     le16_to_cpu(mbx->status_flags));
1338
1339                 ql_dump_buffer(ql_dbg_async + ql_dbg_buffer, vha, 0x5029,
1340                     (uint8_t *)mbx, sizeof(*mbx));
1341
1342                 goto logio_done;
1343         }
1344
1345         status = le16_to_cpu(mbx->status);
1346         if (status == 0x30 && sp->type == SRB_LOGIN_CMD &&
1347             le16_to_cpu(mbx->mb0) == MBS_COMMAND_COMPLETE)
1348                 status = 0;
1349         if (!status && le16_to_cpu(mbx->mb0) == MBS_COMMAND_COMPLETE) {
1350                 ql_dbg(ql_dbg_async, vha, 0x5045,
1351                     "Async-%s complete - hdl=%x portid=%02x%02x%02x mbx1=%x.\n",
1352                     type, sp->handle, fcport->d_id.b.domain,
1353                     fcport->d_id.b.area, fcport->d_id.b.al_pa,
1354                     le16_to_cpu(mbx->mb1));
1355
1356                 data[0] = MBS_COMMAND_COMPLETE;
1357                 if (sp->type == SRB_LOGIN_CMD) {
1358                         fcport->port_type = FCT_TARGET;
1359                         if (le16_to_cpu(mbx->mb1) & BIT_0)
1360                                 fcport->port_type = FCT_INITIATOR;
1361                         else if (le16_to_cpu(mbx->mb1) & BIT_1)
1362                                 fcport->flags |= FCF_FCP2_DEVICE;
1363                 }
1364                 goto logio_done;
1365         }
1366
1367         data[0] = le16_to_cpu(mbx->mb0);
1368         switch (data[0]) {
1369         case MBS_PORT_ID_USED:
1370                 data[1] = le16_to_cpu(mbx->mb1);
1371                 break;
1372         case MBS_LOOP_ID_USED:
1373                 break;
1374         default:
1375                 data[0] = MBS_COMMAND_ERROR;
1376                 break;
1377         }
1378
1379         ql_log(ql_log_warn, vha, 0x5046,
1380             "Async-%s failed - hdl=%x portid=%02x%02x%02x status=%x "
1381             "mb0=%x mb1=%x mb2=%x mb6=%x mb7=%x.\n", type, sp->handle,
1382             fcport->d_id.b.domain, fcport->d_id.b.area, fcport->d_id.b.al_pa,
1383             status, le16_to_cpu(mbx->mb0), le16_to_cpu(mbx->mb1),
1384             le16_to_cpu(mbx->mb2), le16_to_cpu(mbx->mb6),
1385             le16_to_cpu(mbx->mb7));
1386
1387 logio_done:
1388         sp->done(sp, 0);
1389 }
1390
1391 static void
1392 qla24xx_mbx_iocb_entry(scsi_qla_host_t *vha, struct req_que *req,
1393     struct mbx_24xx_entry *pkt)
1394 {
1395         const char func[] = "MBX-IOCB2";
1396         srb_t *sp;
1397         struct srb_iocb *si;
1398         u16 sz, i;
1399         int res;
1400
1401         sp = qla2x00_get_sp_from_handle(vha, func, req, pkt);
1402         if (!sp)
1403                 return;
1404
1405         si = &sp->u.iocb_cmd;
1406         sz = min(ARRAY_SIZE(pkt->mb), ARRAY_SIZE(sp->u.iocb_cmd.u.mbx.in_mb));
1407
1408         for (i = 0; i < sz; i++)
1409                 si->u.mbx.in_mb[i] = le16_to_cpu(pkt->mb[i]);
1410
1411         res = (si->u.mbx.in_mb[0] & MBS_MASK);
1412
1413         sp->done(sp, res);
1414 }
1415
1416 static void
1417 qla24xxx_nack_iocb_entry(scsi_qla_host_t *vha, struct req_que *req,
1418     struct nack_to_isp *pkt)
1419 {
1420         const char func[] = "nack";
1421         srb_t *sp;
1422         int res = 0;
1423
1424         sp = qla2x00_get_sp_from_handle(vha, func, req, pkt);
1425         if (!sp)
1426                 return;
1427
1428         if (pkt->u.isp2x.status != cpu_to_le16(NOTIFY_ACK_SUCCESS))
1429                 res = QLA_FUNCTION_FAILED;
1430
1431         sp->done(sp, res);
1432 }
1433
1434 static void
1435 qla2x00_ct_entry(scsi_qla_host_t *vha, struct req_que *req,
1436     sts_entry_t *pkt, int iocb_type)
1437 {
1438         const char func[] = "CT_IOCB";
1439         const char *type;
1440         srb_t *sp;
1441         struct bsg_job *bsg_job;
1442         struct fc_bsg_reply *bsg_reply;
1443         uint16_t comp_status;
1444         int res = 0;
1445
1446         sp = qla2x00_get_sp_from_handle(vha, func, req, pkt);
1447         if (!sp)
1448                 return;
1449
1450         switch (sp->type) {
1451         case SRB_CT_CMD:
1452             bsg_job = sp->u.bsg_job;
1453             bsg_reply = bsg_job->reply;
1454
1455             type = "ct pass-through";
1456
1457             comp_status = le16_to_cpu(pkt->comp_status);
1458
1459             /*
1460              * return FC_CTELS_STATUS_OK and leave the decoding of the ELS/CT
1461              * fc payload  to the caller
1462              */
1463             bsg_reply->reply_data.ctels_reply.status = FC_CTELS_STATUS_OK;
1464             bsg_job->reply_len = sizeof(struct fc_bsg_reply);
1465
1466             if (comp_status != CS_COMPLETE) {
1467                     if (comp_status == CS_DATA_UNDERRUN) {
1468                             res = DID_OK << 16;
1469                             bsg_reply->reply_payload_rcv_len =
1470                                 le16_to_cpu(((sts_entry_t *)pkt)->rsp_info_len);
1471
1472                             ql_log(ql_log_warn, vha, 0x5048,
1473                                 "CT pass-through-%s error comp_status=0x%x total_byte=0x%x.\n",
1474                                 type, comp_status,
1475                                 bsg_reply->reply_payload_rcv_len);
1476                     } else {
1477                             ql_log(ql_log_warn, vha, 0x5049,
1478                                 "CT pass-through-%s error comp_status=0x%x.\n",
1479                                 type, comp_status);
1480                             res = DID_ERROR << 16;
1481                             bsg_reply->reply_payload_rcv_len = 0;
1482                     }
1483                     ql_dump_buffer(ql_dbg_async + ql_dbg_buffer, vha, 0x5035,
1484                         (uint8_t *)pkt, sizeof(*pkt));
1485             } else {
1486                     res = DID_OK << 16;
1487                     bsg_reply->reply_payload_rcv_len =
1488                         bsg_job->reply_payload.payload_len;
1489                     bsg_job->reply_len = 0;
1490             }
1491             break;
1492         case SRB_CT_PTHRU_CMD:
1493             /*
1494              * borrowing sts_entry_24xx.comp_status.
1495              * same location as ct_entry_24xx.comp_status
1496              */
1497              res = qla2x00_chk_ms_status(vha, (ms_iocb_entry_t *)pkt,
1498                  (struct ct_sns_rsp *)sp->u.iocb_cmd.u.ctarg.rsp,
1499                  sp->name);
1500              break;
1501         }
1502
1503         sp->done(sp, res);
1504 }
1505
1506 static void
1507 qla24xx_els_ct_entry(scsi_qla_host_t *vha, struct req_que *req,
1508     struct sts_entry_24xx *pkt, int iocb_type)
1509 {
1510         const char func[] = "ELS_CT_IOCB";
1511         const char *type;
1512         srb_t *sp;
1513         struct bsg_job *bsg_job;
1514         struct fc_bsg_reply *bsg_reply;
1515         uint16_t comp_status;
1516         uint32_t fw_status[3];
1517         uint8_t* fw_sts_ptr;
1518         int res;
1519
1520         sp = qla2x00_get_sp_from_handle(vha, func, req, pkt);
1521         if (!sp)
1522                 return;
1523         bsg_job = sp->u.bsg_job;
1524         bsg_reply = bsg_job->reply;
1525
1526         type = NULL;
1527         switch (sp->type) {
1528         case SRB_ELS_CMD_RPT:
1529         case SRB_ELS_CMD_HST:
1530                 type = "els";
1531                 break;
1532         case SRB_CT_CMD:
1533                 type = "ct pass-through";
1534                 break;
1535         case SRB_ELS_DCMD:
1536                 type = "Driver ELS logo";
1537                 ql_dbg(ql_dbg_user, vha, 0x5047,
1538                     "Completing %s: (%p) type=%d.\n", type, sp, sp->type);
1539                 sp->done(sp, 0);
1540                 return;
1541         case SRB_CT_PTHRU_CMD:
1542                 /* borrowing sts_entry_24xx.comp_status.
1543                    same location as ct_entry_24xx.comp_status
1544                  */
1545                 res = qla2x00_chk_ms_status(vha, (ms_iocb_entry_t *)pkt,
1546                         (struct ct_sns_rsp *)sp->u.iocb_cmd.u.ctarg.rsp,
1547                         sp->name);
1548                 sp->done(sp, res);
1549                 return;
1550         default:
1551                 ql_dbg(ql_dbg_user, vha, 0x503e,
1552                     "Unrecognized SRB: (%p) type=%d.\n", sp, sp->type);
1553                 return;
1554         }
1555
1556         comp_status = fw_status[0] = le16_to_cpu(pkt->comp_status);
1557         fw_status[1] = le16_to_cpu(((struct els_sts_entry_24xx*)pkt)->error_subcode_1);
1558         fw_status[2] = le16_to_cpu(((struct els_sts_entry_24xx*)pkt)->error_subcode_2);
1559
1560         /* return FC_CTELS_STATUS_OK and leave the decoding of the ELS/CT
1561          * fc payload  to the caller
1562          */
1563         bsg_reply->reply_data.ctels_reply.status = FC_CTELS_STATUS_OK;
1564         bsg_job->reply_len = sizeof(struct fc_bsg_reply) + sizeof(fw_status);
1565
1566         if (comp_status != CS_COMPLETE) {
1567                 if (comp_status == CS_DATA_UNDERRUN) {
1568                         res = DID_OK << 16;
1569                         bsg_reply->reply_payload_rcv_len =
1570                             le16_to_cpu(((struct els_sts_entry_24xx *)pkt)->total_byte_count);
1571
1572                         ql_dbg(ql_dbg_user, vha, 0x503f,
1573                             "ELS-CT pass-through-%s error hdl=%x comp_status-status=0x%x "
1574                             "error subcode 1=0x%x error subcode 2=0x%x total_byte = 0x%x.\n",
1575                             type, sp->handle, comp_status, fw_status[1], fw_status[2],
1576                             le16_to_cpu(((struct els_sts_entry_24xx *)
1577                                 pkt)->total_byte_count));
1578                         fw_sts_ptr = ((uint8_t*)scsi_req(bsg_job->req)->sense) +
1579                                 sizeof(struct fc_bsg_reply);
1580                         memcpy( fw_sts_ptr, fw_status, sizeof(fw_status));
1581                 }
1582                 else {
1583                         ql_dbg(ql_dbg_user, vha, 0x5040,
1584                             "ELS-CT pass-through-%s error hdl=%x comp_status-status=0x%x "
1585                             "error subcode 1=0x%x error subcode 2=0x%x.\n",
1586                             type, sp->handle, comp_status,
1587                             le16_to_cpu(((struct els_sts_entry_24xx *)
1588                                 pkt)->error_subcode_1),
1589                             le16_to_cpu(((struct els_sts_entry_24xx *)
1590                                     pkt)->error_subcode_2));
1591                         res = DID_ERROR << 16;
1592                         bsg_reply->reply_payload_rcv_len = 0;
1593                         fw_sts_ptr = ((uint8_t*)scsi_req(bsg_job->req)->sense) +
1594                                         sizeof(struct fc_bsg_reply);
1595                         memcpy( fw_sts_ptr, fw_status, sizeof(fw_status));
1596                 }
1597                 ql_dump_buffer(ql_dbg_user + ql_dbg_buffer, vha, 0x5056,
1598                                 (uint8_t *)pkt, sizeof(*pkt));
1599         }
1600         else {
1601                 res =  DID_OK << 16;
1602                 bsg_reply->reply_payload_rcv_len = bsg_job->reply_payload.payload_len;
1603                 bsg_job->reply_len = 0;
1604         }
1605
1606         sp->done(sp, res);
1607 }
1608
1609 static void
1610 qla24xx_logio_entry(scsi_qla_host_t *vha, struct req_que *req,
1611     struct logio_entry_24xx *logio)
1612 {
1613         const char func[] = "LOGIO-IOCB";
1614         const char *type;
1615         fc_port_t *fcport;
1616         srb_t *sp;
1617         struct srb_iocb *lio;
1618         uint16_t *data;
1619         uint32_t iop[2];
1620
1621         sp = qla2x00_get_sp_from_handle(vha, func, req, logio);
1622         if (!sp)
1623                 return;
1624
1625         lio = &sp->u.iocb_cmd;
1626         type = sp->name;
1627         fcport = sp->fcport;
1628         data = lio->u.logio.data;
1629
1630         data[0] = MBS_COMMAND_ERROR;
1631         data[1] = lio->u.logio.flags & SRB_LOGIN_RETRIED ?
1632                 QLA_LOGIO_LOGIN_RETRIED : 0;
1633         if (logio->entry_status) {
1634                 ql_log(ql_log_warn, fcport->vha, 0x5034,
1635                     "Async-%s error entry - %8phC hdl=%x"
1636                     "portid=%02x%02x%02x entry-status=%x.\n",
1637                     type, fcport->port_name, sp->handle, fcport->d_id.b.domain,
1638                     fcport->d_id.b.area, fcport->d_id.b.al_pa,
1639                     logio->entry_status);
1640                 ql_dump_buffer(ql_dbg_async + ql_dbg_buffer, vha, 0x504d,
1641                     (uint8_t *)logio, sizeof(*logio));
1642
1643                 goto logio_done;
1644         }
1645
1646         if (le16_to_cpu(logio->comp_status) == CS_COMPLETE) {
1647                 ql_dbg(ql_dbg_async, fcport->vha, 0x5036,
1648                     "Async-%s complete - %8phC hdl=%x portid=%02x%02x%02x "
1649                     "iop0=%x.\n", type, fcport->port_name, sp->handle,
1650                     fcport->d_id.b.domain,
1651                     fcport->d_id.b.area, fcport->d_id.b.al_pa,
1652                     le32_to_cpu(logio->io_parameter[0]));
1653
1654                 vha->hw->exch_starvation = 0;
1655                 data[0] = MBS_COMMAND_COMPLETE;
1656                 if (sp->type != SRB_LOGIN_CMD)
1657                         goto logio_done;
1658
1659                 iop[0] = le32_to_cpu(logio->io_parameter[0]);
1660                 if (iop[0] & BIT_4) {
1661                         fcport->port_type = FCT_TARGET;
1662                         if (iop[0] & BIT_8)
1663                                 fcport->flags |= FCF_FCP2_DEVICE;
1664                 } else if (iop[0] & BIT_5)
1665                         fcport->port_type = FCT_INITIATOR;
1666
1667                 if (iop[0] & BIT_7)
1668                         fcport->flags |= FCF_CONF_COMP_SUPPORTED;
1669
1670                 if (logio->io_parameter[7] || logio->io_parameter[8])
1671                         fcport->supported_classes |= FC_COS_CLASS2;
1672                 if (logio->io_parameter[9] || logio->io_parameter[10])
1673                         fcport->supported_classes |= FC_COS_CLASS3;
1674
1675                 goto logio_done;
1676         }
1677
1678         iop[0] = le32_to_cpu(logio->io_parameter[0]);
1679         iop[1] = le32_to_cpu(logio->io_parameter[1]);
1680         lio->u.logio.iop[0] = iop[0];
1681         lio->u.logio.iop[1] = iop[1];
1682         switch (iop[0]) {
1683         case LSC_SCODE_PORTID_USED:
1684                 data[0] = MBS_PORT_ID_USED;
1685                 data[1] = LSW(iop[1]);
1686                 break;
1687         case LSC_SCODE_NPORT_USED:
1688                 data[0] = MBS_LOOP_ID_USED;
1689                 break;
1690         case LSC_SCODE_CMD_FAILED:
1691                 if (iop[1] == 0x0606) {
1692                         /*
1693                          * PLOGI/PRLI Completed. We must have Recv PLOGI/PRLI,
1694                          * Target side acked.
1695                          */
1696                         data[0] = MBS_COMMAND_COMPLETE;
1697                         goto logio_done;
1698                 }
1699                 data[0] = MBS_COMMAND_ERROR;
1700                 break;
1701         case LSC_SCODE_NOXCB:
1702                 vha->hw->exch_starvation++;
1703                 if (vha->hw->exch_starvation > 5) {
1704                         ql_log(ql_log_warn, vha, 0xffff,
1705                             "Exchange starvation. Resetting RISC\n");
1706
1707                         vha->hw->exch_starvation = 0;
1708
1709                         if (IS_P3P_TYPE(vha->hw))
1710                                 set_bit(FCOE_CTX_RESET_NEEDED, &vha->dpc_flags);
1711                         else
1712                                 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
1713                         qla2xxx_wake_dpc(vha);
1714                 }
1715                 /* drop through */
1716         default:
1717                 data[0] = MBS_COMMAND_ERROR;
1718                 break;
1719         }
1720
1721         ql_dbg(ql_dbg_async, fcport->vha, 0x5037,
1722             "Async-%s failed - %8phC hdl=%x portid=%02x%02x%02x comp=%x "
1723             "iop0=%x iop1=%x.\n", type, fcport->port_name,
1724                 sp->handle, fcport->d_id.b.domain,
1725             fcport->d_id.b.area, fcport->d_id.b.al_pa,
1726             le16_to_cpu(logio->comp_status),
1727             le32_to_cpu(logio->io_parameter[0]),
1728             le32_to_cpu(logio->io_parameter[1]));
1729
1730 logio_done:
1731         sp->done(sp, 0);
1732 }
1733
1734 static void
1735 qla24xx_tm_iocb_entry(scsi_qla_host_t *vha, struct req_que *req, void *tsk)
1736 {
1737         const char func[] = "TMF-IOCB";
1738         const char *type;
1739         fc_port_t *fcport;
1740         srb_t *sp;
1741         struct srb_iocb *iocb;
1742         struct sts_entry_24xx *sts = (struct sts_entry_24xx *)tsk;
1743
1744         sp = qla2x00_get_sp_from_handle(vha, func, req, tsk);
1745         if (!sp)
1746                 return;
1747
1748         iocb = &sp->u.iocb_cmd;
1749         type = sp->name;
1750         fcport = sp->fcport;
1751         iocb->u.tmf.data = QLA_SUCCESS;
1752
1753         if (sts->entry_status) {
1754                 ql_log(ql_log_warn, fcport->vha, 0x5038,
1755                     "Async-%s error - hdl=%x entry-status(%x).\n",
1756                     type, sp->handle, sts->entry_status);
1757                 iocb->u.tmf.data = QLA_FUNCTION_FAILED;
1758         } else if (sts->comp_status != cpu_to_le16(CS_COMPLETE)) {
1759                 ql_log(ql_log_warn, fcport->vha, 0x5039,
1760                     "Async-%s error - hdl=%x completion status(%x).\n",
1761                     type, sp->handle, sts->comp_status);
1762                 iocb->u.tmf.data = QLA_FUNCTION_FAILED;
1763         } else if ((le16_to_cpu(sts->scsi_status) &
1764             SS_RESPONSE_INFO_LEN_VALID)) {
1765                 if (le32_to_cpu(sts->rsp_data_len) < 4) {
1766                         ql_log(ql_log_warn, fcport->vha, 0x503b,
1767                             "Async-%s error - hdl=%x not enough response(%d).\n",
1768                             type, sp->handle, sts->rsp_data_len);
1769                 } else if (sts->data[3]) {
1770                         ql_log(ql_log_warn, fcport->vha, 0x503c,
1771                             "Async-%s error - hdl=%x response(%x).\n",
1772                             type, sp->handle, sts->data[3]);
1773                         iocb->u.tmf.data = QLA_FUNCTION_FAILED;
1774                 }
1775         }
1776
1777         if (iocb->u.tmf.data != QLA_SUCCESS)
1778                 ql_dump_buffer(ql_dbg_async + ql_dbg_buffer, vha, 0x5055,
1779                     (uint8_t *)sts, sizeof(*sts));
1780
1781         sp->done(sp, 0);
1782 }
1783
1784 /**
1785  * qla2x00_process_response_queue() - Process response queue entries.
1786  * @ha: SCSI driver HA context
1787  */
1788 void
1789 qla2x00_process_response_queue(struct rsp_que *rsp)
1790 {
1791         struct scsi_qla_host *vha;
1792         struct qla_hw_data *ha = rsp->hw;
1793         struct device_reg_2xxx __iomem *reg = &ha->iobase->isp;
1794         sts_entry_t     *pkt;
1795         uint16_t        handle_cnt;
1796         uint16_t        cnt;
1797
1798         vha = pci_get_drvdata(ha->pdev);
1799
1800         if (!vha->flags.online)
1801                 return;
1802
1803         while (rsp->ring_ptr->signature != RESPONSE_PROCESSED) {
1804                 pkt = (sts_entry_t *)rsp->ring_ptr;
1805
1806                 rsp->ring_index++;
1807                 if (rsp->ring_index == rsp->length) {
1808                         rsp->ring_index = 0;
1809                         rsp->ring_ptr = rsp->ring;
1810                 } else {
1811                         rsp->ring_ptr++;
1812                 }
1813
1814                 if (pkt->entry_status != 0) {
1815                         qla2x00_error_entry(vha, rsp, pkt);
1816                         ((response_t *)pkt)->signature = RESPONSE_PROCESSED;
1817                         wmb();
1818                         continue;
1819                 }
1820
1821                 switch (pkt->entry_type) {
1822                 case STATUS_TYPE:
1823                         qla2x00_status_entry(vha, rsp, pkt);
1824                         break;
1825                 case STATUS_TYPE_21:
1826                         handle_cnt = ((sts21_entry_t *)pkt)->handle_count;
1827                         for (cnt = 0; cnt < handle_cnt; cnt++) {
1828                                 qla2x00_process_completed_request(vha, rsp->req,
1829                                     ((sts21_entry_t *)pkt)->handle[cnt]);
1830                         }
1831                         break;
1832                 case STATUS_TYPE_22:
1833                         handle_cnt = ((sts22_entry_t *)pkt)->handle_count;
1834                         for (cnt = 0; cnt < handle_cnt; cnt++) {
1835                                 qla2x00_process_completed_request(vha, rsp->req,
1836                                     ((sts22_entry_t *)pkt)->handle[cnt]);
1837                         }
1838                         break;
1839                 case STATUS_CONT_TYPE:
1840                         qla2x00_status_cont_entry(rsp, (sts_cont_entry_t *)pkt);
1841                         break;
1842                 case MBX_IOCB_TYPE:
1843                         qla2x00_mbx_iocb_entry(vha, rsp->req,
1844                             (struct mbx_entry *)pkt);
1845                         break;
1846                 case CT_IOCB_TYPE:
1847                         qla2x00_ct_entry(vha, rsp->req, pkt, CT_IOCB_TYPE);
1848                         break;
1849                 default:
1850                         /* Type Not Supported. */
1851                         ql_log(ql_log_warn, vha, 0x504a,
1852                             "Received unknown response pkt type %x "
1853                             "entry status=%x.\n",
1854                             pkt->entry_type, pkt->entry_status);
1855                         break;
1856                 }
1857                 ((response_t *)pkt)->signature = RESPONSE_PROCESSED;
1858                 wmb();
1859         }
1860
1861         /* Adjust ring index */
1862         WRT_REG_WORD(ISP_RSP_Q_OUT(ha, reg), rsp->ring_index);
1863 }
1864
1865 static inline void
1866 qla2x00_handle_sense(srb_t *sp, uint8_t *sense_data, uint32_t par_sense_len,
1867                      uint32_t sense_len, struct rsp_que *rsp, int res)
1868 {
1869         struct scsi_qla_host *vha = sp->vha;
1870         struct scsi_cmnd *cp = GET_CMD_SP(sp);
1871         uint32_t track_sense_len;
1872
1873         if (sense_len >= SCSI_SENSE_BUFFERSIZE)
1874                 sense_len = SCSI_SENSE_BUFFERSIZE;
1875
1876         SET_CMD_SENSE_LEN(sp, sense_len);
1877         SET_CMD_SENSE_PTR(sp, cp->sense_buffer);
1878         track_sense_len = sense_len;
1879
1880         if (sense_len > par_sense_len)
1881                 sense_len = par_sense_len;
1882
1883         memcpy(cp->sense_buffer, sense_data, sense_len);
1884
1885         SET_CMD_SENSE_PTR(sp, cp->sense_buffer + sense_len);
1886         track_sense_len -= sense_len;
1887         SET_CMD_SENSE_LEN(sp, track_sense_len);
1888
1889         if (track_sense_len != 0) {
1890                 rsp->status_srb = sp;
1891                 cp->result = res;
1892         }
1893
1894         if (sense_len) {
1895                 ql_dbg(ql_dbg_io + ql_dbg_buffer, vha, 0x301c,
1896                     "Check condition Sense data, nexus%ld:%d:%llu cmd=%p.\n",
1897                     sp->vha->host_no, cp->device->id, cp->device->lun,
1898                     cp);
1899                 ql_dump_buffer(ql_dbg_io + ql_dbg_buffer, vha, 0x302b,
1900                     cp->sense_buffer, sense_len);
1901         }
1902 }
1903
1904 struct scsi_dif_tuple {
1905         __be16 guard;       /* Checksum */
1906         __be16 app_tag;         /* APPL identifier */
1907         __be32 ref_tag;         /* Target LBA or indirect LBA */
1908 };
1909
1910 /*
1911  * Checks the guard or meta-data for the type of error
1912  * detected by the HBA. In case of errors, we set the
1913  * ASC/ASCQ fields in the sense buffer with ILLEGAL_REQUEST
1914  * to indicate to the kernel that the HBA detected error.
1915  */
1916 static inline int
1917 qla2x00_handle_dif_error(srb_t *sp, struct sts_entry_24xx *sts24)
1918 {
1919         struct scsi_qla_host *vha = sp->vha;
1920         struct scsi_cmnd *cmd = GET_CMD_SP(sp);
1921         uint8_t         *ap = &sts24->data[12];
1922         uint8_t         *ep = &sts24->data[20];
1923         uint32_t        e_ref_tag, a_ref_tag;
1924         uint16_t        e_app_tag, a_app_tag;
1925         uint16_t        e_guard, a_guard;
1926
1927         /*
1928          * swab32 of the "data" field in the beginning of qla2x00_status_entry()
1929          * would make guard field appear at offset 2
1930          */
1931         a_guard   = le16_to_cpu(*(uint16_t *)(ap + 2));
1932         a_app_tag = le16_to_cpu(*(uint16_t *)(ap + 0));
1933         a_ref_tag = le32_to_cpu(*(uint32_t *)(ap + 4));
1934         e_guard   = le16_to_cpu(*(uint16_t *)(ep + 2));
1935         e_app_tag = le16_to_cpu(*(uint16_t *)(ep + 0));
1936         e_ref_tag = le32_to_cpu(*(uint32_t *)(ep + 4));
1937
1938         ql_dbg(ql_dbg_io, vha, 0x3023,
1939             "iocb(s) %p Returned STATUS.\n", sts24);
1940
1941         ql_dbg(ql_dbg_io, vha, 0x3024,
1942             "DIF ERROR in cmd 0x%x lba 0x%llx act ref"
1943             " tag=0x%x, exp ref_tag=0x%x, act app tag=0x%x, exp app"
1944             " tag=0x%x, act guard=0x%x, exp guard=0x%x.\n",
1945             cmd->cmnd[0], (u64)scsi_get_lba(cmd), a_ref_tag, e_ref_tag,
1946             a_app_tag, e_app_tag, a_guard, e_guard);
1947
1948         /*
1949          * Ignore sector if:
1950          * For type     3: ref & app tag is all 'f's
1951          * For type 0,1,2: app tag is all 'f's
1952          */
1953         if ((a_app_tag == 0xffff) &&
1954             ((scsi_get_prot_type(cmd) != SCSI_PROT_DIF_TYPE3) ||
1955              (a_ref_tag == 0xffffffff))) {
1956                 uint32_t blocks_done, resid;
1957                 sector_t lba_s = scsi_get_lba(cmd);
1958
1959                 /* 2TB boundary case covered automatically with this */
1960                 blocks_done = e_ref_tag - (uint32_t)lba_s + 1;
1961
1962                 resid = scsi_bufflen(cmd) - (blocks_done *
1963                     cmd->device->sector_size);
1964
1965                 scsi_set_resid(cmd, resid);
1966                 cmd->result = DID_OK << 16;
1967
1968                 /* Update protection tag */
1969                 if (scsi_prot_sg_count(cmd)) {
1970                         uint32_t i, j = 0, k = 0, num_ent;
1971                         struct scatterlist *sg;
1972                         struct t10_pi_tuple *spt;
1973
1974                         /* Patch the corresponding protection tags */
1975                         scsi_for_each_prot_sg(cmd, sg,
1976                             scsi_prot_sg_count(cmd), i) {
1977                                 num_ent = sg_dma_len(sg) / 8;
1978                                 if (k + num_ent < blocks_done) {
1979                                         k += num_ent;
1980                                         continue;
1981                                 }
1982                                 j = blocks_done - k - 1;
1983                                 k = blocks_done;
1984                                 break;
1985                         }
1986
1987                         if (k != blocks_done) {
1988                                 ql_log(ql_log_warn, vha, 0x302f,
1989                                     "unexpected tag values tag:lba=%x:%llx)\n",
1990                                     e_ref_tag, (unsigned long long)lba_s);
1991                                 return 1;
1992                         }
1993
1994                         spt = page_address(sg_page(sg)) + sg->offset;
1995                         spt += j;
1996
1997                         spt->app_tag = 0xffff;
1998                         if (scsi_get_prot_type(cmd) == SCSI_PROT_DIF_TYPE3)
1999                                 spt->ref_tag = 0xffffffff;
2000                 }
2001
2002                 return 0;
2003         }
2004
2005         /* check guard */
2006         if (e_guard != a_guard) {
2007                 scsi_build_sense_buffer(1, cmd->sense_buffer, ILLEGAL_REQUEST,
2008                     0x10, 0x1);
2009                 set_driver_byte(cmd, DRIVER_SENSE);
2010                 set_host_byte(cmd, DID_ABORT);
2011                 cmd->result |= SAM_STAT_CHECK_CONDITION << 1;
2012                 return 1;
2013         }
2014
2015         /* check ref tag */
2016         if (e_ref_tag != a_ref_tag) {
2017                 scsi_build_sense_buffer(1, cmd->sense_buffer, ILLEGAL_REQUEST,
2018                     0x10, 0x3);
2019                 set_driver_byte(cmd, DRIVER_SENSE);
2020                 set_host_byte(cmd, DID_ABORT);
2021                 cmd->result |= SAM_STAT_CHECK_CONDITION << 1;
2022                 return 1;
2023         }
2024
2025         /* check appl tag */
2026         if (e_app_tag != a_app_tag) {
2027                 scsi_build_sense_buffer(1, cmd->sense_buffer, ILLEGAL_REQUEST,
2028                     0x10, 0x2);
2029                 set_driver_byte(cmd, DRIVER_SENSE);
2030                 set_host_byte(cmd, DID_ABORT);
2031                 cmd->result |= SAM_STAT_CHECK_CONDITION << 1;
2032                 return 1;
2033         }
2034
2035         return 1;
2036 }
2037
2038 static void
2039 qla25xx_process_bidir_status_iocb(scsi_qla_host_t *vha, void *pkt,
2040                                   struct req_que *req, uint32_t index)
2041 {
2042         struct qla_hw_data *ha = vha->hw;
2043         srb_t *sp;
2044         uint16_t        comp_status;
2045         uint16_t        scsi_status;
2046         uint16_t thread_id;
2047         uint32_t rval = EXT_STATUS_OK;
2048         struct bsg_job *bsg_job = NULL;
2049         struct fc_bsg_request *bsg_request;
2050         struct fc_bsg_reply *bsg_reply;
2051         sts_entry_t *sts;
2052         struct sts_entry_24xx *sts24;
2053         sts = (sts_entry_t *) pkt;
2054         sts24 = (struct sts_entry_24xx *) pkt;
2055
2056         /* Validate handle. */
2057         if (index >= req->num_outstanding_cmds) {
2058                 ql_log(ql_log_warn, vha, 0x70af,
2059                     "Invalid SCSI completion handle 0x%x.\n", index);
2060                 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
2061                 return;
2062         }
2063
2064         sp = req->outstanding_cmds[index];
2065         if (!sp) {
2066                 ql_log(ql_log_warn, vha, 0x70b0,
2067                     "Req:%d: Invalid ISP SCSI completion handle(0x%x)\n",
2068                     req->id, index);
2069
2070                 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
2071                 return;
2072         }
2073
2074         /* Free outstanding command slot. */
2075         req->outstanding_cmds[index] = NULL;
2076         bsg_job = sp->u.bsg_job;
2077         bsg_request = bsg_job->request;
2078         bsg_reply = bsg_job->reply;
2079
2080         if (IS_FWI2_CAPABLE(ha)) {
2081                 comp_status = le16_to_cpu(sts24->comp_status);
2082                 scsi_status = le16_to_cpu(sts24->scsi_status) & SS_MASK;
2083         } else {
2084                 comp_status = le16_to_cpu(sts->comp_status);
2085                 scsi_status = le16_to_cpu(sts->scsi_status) & SS_MASK;
2086         }
2087
2088         thread_id = bsg_request->rqst_data.h_vendor.vendor_cmd[1];
2089         switch (comp_status) {
2090         case CS_COMPLETE:
2091                 if (scsi_status == 0) {
2092                         bsg_reply->reply_payload_rcv_len =
2093                                         bsg_job->reply_payload.payload_len;
2094                         vha->qla_stats.input_bytes +=
2095                                 bsg_reply->reply_payload_rcv_len;
2096                         vha->qla_stats.input_requests++;
2097                         rval = EXT_STATUS_OK;
2098                 }
2099                 goto done;
2100
2101         case CS_DATA_OVERRUN:
2102                 ql_dbg(ql_dbg_user, vha, 0x70b1,
2103                     "Command completed with date overrun thread_id=%d\n",
2104                     thread_id);
2105                 rval = EXT_STATUS_DATA_OVERRUN;
2106                 break;
2107
2108         case CS_DATA_UNDERRUN:
2109                 ql_dbg(ql_dbg_user, vha, 0x70b2,
2110                     "Command completed with date underrun thread_id=%d\n",
2111                     thread_id);
2112                 rval = EXT_STATUS_DATA_UNDERRUN;
2113                 break;
2114         case CS_BIDIR_RD_OVERRUN:
2115                 ql_dbg(ql_dbg_user, vha, 0x70b3,
2116                     "Command completed with read data overrun thread_id=%d\n",
2117                     thread_id);
2118                 rval = EXT_STATUS_DATA_OVERRUN;
2119                 break;
2120
2121         case CS_BIDIR_RD_WR_OVERRUN:
2122                 ql_dbg(ql_dbg_user, vha, 0x70b4,
2123                     "Command completed with read and write data overrun "
2124                     "thread_id=%d\n", thread_id);
2125                 rval = EXT_STATUS_DATA_OVERRUN;
2126                 break;
2127
2128         case CS_BIDIR_RD_OVERRUN_WR_UNDERRUN:
2129                 ql_dbg(ql_dbg_user, vha, 0x70b5,
2130                     "Command completed with read data over and write data "
2131                     "underrun thread_id=%d\n", thread_id);
2132                 rval = EXT_STATUS_DATA_OVERRUN;
2133                 break;
2134
2135         case CS_BIDIR_RD_UNDERRUN:
2136                 ql_dbg(ql_dbg_user, vha, 0x70b6,
2137                     "Command completed with read data data underrun "
2138                     "thread_id=%d\n", thread_id);
2139                 rval = EXT_STATUS_DATA_UNDERRUN;
2140                 break;
2141
2142         case CS_BIDIR_RD_UNDERRUN_WR_OVERRUN:
2143                 ql_dbg(ql_dbg_user, vha, 0x70b7,
2144                     "Command completed with read data under and write data "
2145                     "overrun thread_id=%d\n", thread_id);
2146                 rval = EXT_STATUS_DATA_UNDERRUN;
2147                 break;
2148
2149         case CS_BIDIR_RD_WR_UNDERRUN:
2150                 ql_dbg(ql_dbg_user, vha, 0x70b8,
2151                     "Command completed with read and write data underrun "
2152                     "thread_id=%d\n", thread_id);
2153                 rval = EXT_STATUS_DATA_UNDERRUN;
2154                 break;
2155
2156         case CS_BIDIR_DMA:
2157                 ql_dbg(ql_dbg_user, vha, 0x70b9,
2158                     "Command completed with data DMA error thread_id=%d\n",
2159                     thread_id);
2160                 rval = EXT_STATUS_DMA_ERR;
2161                 break;
2162
2163         case CS_TIMEOUT:
2164                 ql_dbg(ql_dbg_user, vha, 0x70ba,
2165                     "Command completed with timeout thread_id=%d\n",
2166                     thread_id);
2167                 rval = EXT_STATUS_TIMEOUT;
2168                 break;
2169         default:
2170                 ql_dbg(ql_dbg_user, vha, 0x70bb,
2171                     "Command completed with completion status=0x%x "
2172                     "thread_id=%d\n", comp_status, thread_id);
2173                 rval = EXT_STATUS_ERR;
2174                 break;
2175         }
2176         bsg_reply->reply_payload_rcv_len = 0;
2177
2178 done:
2179         /* Return the vendor specific reply to API */
2180         bsg_reply->reply_data.vendor_reply.vendor_rsp[0] = rval;
2181         bsg_job->reply_len = sizeof(struct fc_bsg_reply);
2182         /* Always return DID_OK, bsg will send the vendor specific response
2183          * in this case only */
2184         sp->done(sp, DID_OK << 6);
2185
2186 }
2187
2188 /**
2189  * qla2x00_status_entry() - Process a Status IOCB entry.
2190  * @ha: SCSI driver HA context
2191  * @pkt: Entry pointer
2192  */
2193 static void
2194 qla2x00_status_entry(scsi_qla_host_t *vha, struct rsp_que *rsp, void *pkt)
2195 {
2196         srb_t           *sp;
2197         fc_port_t       *fcport;
2198         struct scsi_cmnd *cp;
2199         sts_entry_t *sts;
2200         struct sts_entry_24xx *sts24;
2201         uint16_t        comp_status;
2202         uint16_t        scsi_status;
2203         uint16_t        ox_id;
2204         uint8_t         lscsi_status;
2205         int32_t         resid;
2206         uint32_t sense_len, par_sense_len, rsp_info_len, resid_len,
2207             fw_resid_len;
2208         uint8_t         *rsp_info, *sense_data;
2209         struct qla_hw_data *ha = vha->hw;
2210         uint32_t handle;
2211         uint16_t que;
2212         struct req_que *req;
2213         int logit = 1;
2214         int res = 0;
2215         uint16_t state_flags = 0;
2216         uint16_t retry_delay = 0;
2217         uint8_t no_logout = 0;
2218
2219         sts = (sts_entry_t *) pkt;
2220         sts24 = (struct sts_entry_24xx *) pkt;
2221         if (IS_FWI2_CAPABLE(ha)) {
2222                 comp_status = le16_to_cpu(sts24->comp_status);
2223                 scsi_status = le16_to_cpu(sts24->scsi_status) & SS_MASK;
2224                 state_flags = le16_to_cpu(sts24->state_flags);
2225         } else {
2226                 comp_status = le16_to_cpu(sts->comp_status);
2227                 scsi_status = le16_to_cpu(sts->scsi_status) & SS_MASK;
2228         }
2229         handle = (uint32_t) LSW(sts->handle);
2230         que = MSW(sts->handle);
2231         req = ha->req_q_map[que];
2232
2233         /* Check for invalid queue pointer */
2234         if (req == NULL ||
2235             que >= find_first_zero_bit(ha->req_qid_map, ha->max_req_queues)) {
2236                 ql_dbg(ql_dbg_io, vha, 0x3059,
2237                     "Invalid status handle (0x%x): Bad req pointer. req=%p, "
2238                     "que=%u.\n", sts->handle, req, que);
2239                 return;
2240         }
2241
2242         /* Validate handle. */
2243         if (handle < req->num_outstanding_cmds) {
2244                 sp = req->outstanding_cmds[handle];
2245                 if (!sp) {
2246                         ql_dbg(ql_dbg_io, vha, 0x3075,
2247                             "%s(%ld): Already returned command for status handle (0x%x).\n",
2248                             __func__, vha->host_no, sts->handle);
2249                         return;
2250                 }
2251         } else {
2252                 ql_dbg(ql_dbg_io, vha, 0x3017,
2253                     "Invalid status handle, out of range (0x%x).\n",
2254                     sts->handle);
2255
2256                 if (!test_bit(ABORT_ISP_ACTIVE, &vha->dpc_flags)) {
2257                         if (IS_P3P_TYPE(ha))
2258                                 set_bit(FCOE_CTX_RESET_NEEDED, &vha->dpc_flags);
2259                         else
2260                                 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
2261                         qla2xxx_wake_dpc(vha);
2262                 }
2263                 return;
2264         }
2265
2266         if (unlikely((state_flags & BIT_1) && (sp->type == SRB_BIDI_CMD))) {
2267                 qla25xx_process_bidir_status_iocb(vha, pkt, req, handle);
2268                 return;
2269         }
2270
2271         /* Task Management completion. */
2272         if (sp->type == SRB_TM_CMD) {
2273                 qla24xx_tm_iocb_entry(vha, req, pkt);
2274                 return;
2275         }
2276
2277         /* Fast path completion. */
2278         if (comp_status == CS_COMPLETE && scsi_status == 0) {
2279                 qla2x00_process_completed_request(vha, req, handle);
2280
2281                 return;
2282         }
2283
2284         req->outstanding_cmds[handle] = NULL;
2285         cp = GET_CMD_SP(sp);
2286         if (cp == NULL) {
2287                 ql_dbg(ql_dbg_io, vha, 0x3018,
2288                     "Command already returned (0x%x/%p).\n",
2289                     sts->handle, sp);
2290
2291                 return;
2292         }
2293
2294         lscsi_status = scsi_status & STATUS_MASK;
2295
2296         fcport = sp->fcport;
2297
2298         ox_id = 0;
2299         sense_len = par_sense_len = rsp_info_len = resid_len =
2300             fw_resid_len = 0;
2301         if (IS_FWI2_CAPABLE(ha)) {
2302                 if (scsi_status & SS_SENSE_LEN_VALID)
2303                         sense_len = le32_to_cpu(sts24->sense_len);
2304                 if (scsi_status & SS_RESPONSE_INFO_LEN_VALID)
2305                         rsp_info_len = le32_to_cpu(sts24->rsp_data_len);
2306                 if (scsi_status & (SS_RESIDUAL_UNDER | SS_RESIDUAL_OVER))
2307                         resid_len = le32_to_cpu(sts24->rsp_residual_count);
2308                 if (comp_status == CS_DATA_UNDERRUN)
2309                         fw_resid_len = le32_to_cpu(sts24->residual_len);
2310                 rsp_info = sts24->data;
2311                 sense_data = sts24->data;
2312                 host_to_fcp_swap(sts24->data, sizeof(sts24->data));
2313                 ox_id = le16_to_cpu(sts24->ox_id);
2314                 par_sense_len = sizeof(sts24->data);
2315                 /* Valid values of the retry delay timer are 0x1-0xffef */
2316                 if (sts24->retry_delay > 0 && sts24->retry_delay < 0xfff1)
2317                         retry_delay = sts24->retry_delay;
2318         } else {
2319                 if (scsi_status & SS_SENSE_LEN_VALID)
2320                         sense_len = le16_to_cpu(sts->req_sense_length);
2321                 if (scsi_status & SS_RESPONSE_INFO_LEN_VALID)
2322                         rsp_info_len = le16_to_cpu(sts->rsp_info_len);
2323                 resid_len = le32_to_cpu(sts->residual_length);
2324                 rsp_info = sts->rsp_info;
2325                 sense_data = sts->req_sense_data;
2326                 par_sense_len = sizeof(sts->req_sense_data);
2327         }
2328
2329         /* Check for any FCP transport errors. */
2330         if (scsi_status & SS_RESPONSE_INFO_LEN_VALID) {
2331                 /* Sense data lies beyond any FCP RESPONSE data. */
2332                 if (IS_FWI2_CAPABLE(ha)) {
2333                         sense_data += rsp_info_len;
2334                         par_sense_len -= rsp_info_len;
2335                 }
2336                 if (rsp_info_len > 3 && rsp_info[3]) {
2337                         ql_dbg(ql_dbg_io, fcport->vha, 0x3019,
2338                             "FCP I/O protocol failure (0x%x/0x%x).\n",
2339                             rsp_info_len, rsp_info[3]);
2340
2341                         res = DID_BUS_BUSY << 16;
2342                         goto out;
2343                 }
2344         }
2345
2346         /* Check for overrun. */
2347         if (IS_FWI2_CAPABLE(ha) && comp_status == CS_COMPLETE &&
2348             scsi_status & SS_RESIDUAL_OVER)
2349                 comp_status = CS_DATA_OVERRUN;
2350
2351         /*
2352          * Check retry_delay_timer value if we receive a busy or
2353          * queue full.
2354          */
2355         if (lscsi_status == SAM_STAT_TASK_SET_FULL ||
2356             lscsi_status == SAM_STAT_BUSY)
2357                 qla2x00_set_retry_delay_timestamp(fcport, retry_delay);
2358
2359         /*
2360          * Based on Host and scsi status generate status code for Linux
2361          */
2362         switch (comp_status) {
2363         case CS_COMPLETE:
2364         case CS_QUEUE_FULL:
2365                 if (scsi_status == 0) {
2366                         res = DID_OK << 16;
2367                         break;
2368                 }
2369                 if (scsi_status & (SS_RESIDUAL_UNDER | SS_RESIDUAL_OVER)) {
2370                         resid = resid_len;
2371                         scsi_set_resid(cp, resid);
2372
2373                         if (!lscsi_status &&
2374                             ((unsigned)(scsi_bufflen(cp) - resid) <
2375                              cp->underflow)) {
2376                                 ql_dbg(ql_dbg_io, fcport->vha, 0x301a,
2377                                     "Mid-layer underflow "
2378                                     "detected (0x%x of 0x%x bytes).\n",
2379                                     resid, scsi_bufflen(cp));
2380
2381                                 res = DID_ERROR << 16;
2382                                 break;
2383                         }
2384                 }
2385                 res = DID_OK << 16 | lscsi_status;
2386
2387                 if (lscsi_status == SAM_STAT_TASK_SET_FULL) {
2388                         ql_dbg(ql_dbg_io, fcport->vha, 0x301b,
2389                             "QUEUE FULL detected.\n");
2390                         break;
2391                 }
2392                 logit = 0;
2393                 if (lscsi_status != SS_CHECK_CONDITION)
2394                         break;
2395
2396                 memset(cp->sense_buffer, 0, SCSI_SENSE_BUFFERSIZE);
2397                 if (!(scsi_status & SS_SENSE_LEN_VALID))
2398                         break;
2399
2400                 qla2x00_handle_sense(sp, sense_data, par_sense_len, sense_len,
2401                     rsp, res);
2402                 break;
2403
2404         case CS_DATA_UNDERRUN:
2405                 /* Use F/W calculated residual length. */
2406                 resid = IS_FWI2_CAPABLE(ha) ? fw_resid_len : resid_len;
2407                 scsi_set_resid(cp, resid);
2408                 if (scsi_status & SS_RESIDUAL_UNDER) {
2409                         if (IS_FWI2_CAPABLE(ha) && fw_resid_len != resid_len) {
2410                                 ql_dbg(ql_dbg_io, fcport->vha, 0x301d,
2411                                     "Dropped frame(s) detected "
2412                                     "(0x%x of 0x%x bytes).\n",
2413                                     resid, scsi_bufflen(cp));
2414
2415                                 res = DID_ERROR << 16 | lscsi_status;
2416                                 goto check_scsi_status;
2417                         }
2418
2419                         if (!lscsi_status &&
2420                             ((unsigned)(scsi_bufflen(cp) - resid) <
2421                             cp->underflow)) {
2422                                 ql_dbg(ql_dbg_io, fcport->vha, 0x301e,
2423                                     "Mid-layer underflow "
2424                                     "detected (0x%x of 0x%x bytes).\n",
2425                                     resid, scsi_bufflen(cp));
2426
2427                                 res = DID_ERROR << 16;
2428                                 break;
2429                         }
2430                 } else if (lscsi_status != SAM_STAT_TASK_SET_FULL &&
2431                             lscsi_status != SAM_STAT_BUSY) {
2432                         /*
2433                          * scsi status of task set and busy are considered to be
2434                          * task not completed.
2435                          */
2436
2437                         ql_dbg(ql_dbg_io, fcport->vha, 0x301f,
2438                             "Dropped frame(s) detected (0x%x "
2439                             "of 0x%x bytes).\n", resid,
2440                             scsi_bufflen(cp));
2441
2442                         res = DID_ERROR << 16 | lscsi_status;
2443                         goto check_scsi_status;
2444                 } else {
2445                         ql_dbg(ql_dbg_io, fcport->vha, 0x3030,
2446                             "scsi_status: 0x%x, lscsi_status: 0x%x\n",
2447                             scsi_status, lscsi_status);
2448                 }
2449
2450                 res = DID_OK << 16 | lscsi_status;
2451                 logit = 0;
2452
2453 check_scsi_status:
2454                 /*
2455                  * Check to see if SCSI Status is non zero. If so report SCSI
2456                  * Status.
2457                  */
2458                 if (lscsi_status != 0) {
2459                         if (lscsi_status == SAM_STAT_TASK_SET_FULL) {
2460                                 ql_dbg(ql_dbg_io, fcport->vha, 0x3020,
2461                                     "QUEUE FULL detected.\n");
2462                                 logit = 1;
2463                                 break;
2464                         }
2465                         if (lscsi_status != SS_CHECK_CONDITION)
2466                                 break;
2467
2468                         memset(cp->sense_buffer, 0, SCSI_SENSE_BUFFERSIZE);
2469                         if (!(scsi_status & SS_SENSE_LEN_VALID))
2470                                 break;
2471
2472                         qla2x00_handle_sense(sp, sense_data, par_sense_len,
2473                             sense_len, rsp, res);
2474                 }
2475                 break;
2476
2477         case CS_PORT_LOGGED_OUT:
2478                 no_logout = 1;
2479         case CS_PORT_CONFIG_CHG:
2480         case CS_PORT_BUSY:
2481         case CS_INCOMPLETE:
2482         case CS_PORT_UNAVAILABLE:
2483         case CS_TIMEOUT:
2484         case CS_RESET:
2485
2486                 /*
2487                  * We are going to have the fc class block the rport
2488                  * while we try to recover so instruct the mid layer
2489                  * to requeue until the class decides how to handle this.
2490                  */
2491                 res = DID_TRANSPORT_DISRUPTED << 16;
2492
2493                 if (comp_status == CS_TIMEOUT) {
2494                         if (IS_FWI2_CAPABLE(ha))
2495                                 break;
2496                         else if ((le16_to_cpu(sts->status_flags) &
2497                             SF_LOGOUT_SENT) == 0)
2498                                 break;
2499                 }
2500
2501                 if (atomic_read(&fcport->state) == FCS_ONLINE) {
2502                         ql_dbg(ql_dbg_disc, fcport->vha, 0x3021,
2503                                 "Port to be marked lost on fcport=%02x%02x%02x, current "
2504                                 "port state= %s comp_status %x.\n", fcport->d_id.b.domain,
2505                                 fcport->d_id.b.area, fcport->d_id.b.al_pa,
2506                                 port_state_str[atomic_read(&fcport->state)],
2507                                 comp_status);
2508
2509                         if (no_logout)
2510                                 fcport->logout_on_delete = 0;
2511
2512                         qla2x00_mark_device_lost(fcport->vha, fcport, 1, 1);
2513                         qlt_schedule_sess_for_deletion_lock(fcport);
2514                 }
2515
2516                 break;
2517
2518         case CS_ABORTED:
2519                 res = DID_RESET << 16;
2520                 break;
2521
2522         case CS_DIF_ERROR:
2523                 logit = qla2x00_handle_dif_error(sp, sts24);
2524                 res = cp->result;
2525                 break;
2526
2527         case CS_TRANSPORT:
2528                 res = DID_ERROR << 16;
2529
2530                 if (!IS_PI_SPLIT_DET_CAPABLE(ha))
2531                         break;
2532
2533                 if (state_flags & BIT_4)
2534                         scmd_printk(KERN_WARNING, cp,
2535                             "Unsupported device '%s' found.\n",
2536                             cp->device->vendor);
2537                 break;
2538
2539         default:
2540                 res = DID_ERROR << 16;
2541                 break;
2542         }
2543
2544 out:
2545         if (logit)
2546                 ql_dbg(ql_dbg_io, fcport->vha, 0x3022,
2547                     "FCP command status: 0x%x-0x%x (0x%x) nexus=%ld:%d:%llu "
2548                     "portid=%02x%02x%02x oxid=0x%x cdb=%10phN len=0x%x "
2549                     "rsp_info=0x%x resid=0x%x fw_resid=0x%x sp=%p cp=%p.\n",
2550                     comp_status, scsi_status, res, vha->host_no,
2551                     cp->device->id, cp->device->lun, fcport->d_id.b.domain,
2552                     fcport->d_id.b.area, fcport->d_id.b.al_pa, ox_id,
2553                     cp->cmnd, scsi_bufflen(cp), rsp_info_len,
2554                     resid_len, fw_resid_len, sp, cp);
2555
2556         if (rsp->status_srb == NULL)
2557                 sp->done(sp, res);
2558 }
2559
2560 /**
2561  * qla2x00_status_cont_entry() - Process a Status Continuations entry.
2562  * @ha: SCSI driver HA context
2563  * @pkt: Entry pointer
2564  *
2565  * Extended sense data.
2566  */
2567 static void
2568 qla2x00_status_cont_entry(struct rsp_que *rsp, sts_cont_entry_t *pkt)
2569 {
2570         uint8_t sense_sz = 0;
2571         struct qla_hw_data *ha = rsp->hw;
2572         struct scsi_qla_host *vha = pci_get_drvdata(ha->pdev);
2573         srb_t *sp = rsp->status_srb;
2574         struct scsi_cmnd *cp;
2575         uint32_t sense_len;
2576         uint8_t *sense_ptr;
2577
2578         if (!sp || !GET_CMD_SENSE_LEN(sp))
2579                 return;
2580
2581         sense_len = GET_CMD_SENSE_LEN(sp);
2582         sense_ptr = GET_CMD_SENSE_PTR(sp);
2583
2584         cp = GET_CMD_SP(sp);
2585         if (cp == NULL) {
2586                 ql_log(ql_log_warn, vha, 0x3025,
2587                     "cmd is NULL: already returned to OS (sp=%p).\n", sp);
2588
2589                 rsp->status_srb = NULL;
2590                 return;
2591         }
2592
2593         if (sense_len > sizeof(pkt->data))
2594                 sense_sz = sizeof(pkt->data);
2595         else
2596                 sense_sz = sense_len;
2597
2598         /* Move sense data. */
2599         if (IS_FWI2_CAPABLE(ha))
2600                 host_to_fcp_swap(pkt->data, sizeof(pkt->data));
2601         memcpy(sense_ptr, pkt->data, sense_sz);
2602         ql_dump_buffer(ql_dbg_io + ql_dbg_buffer, vha, 0x302c,
2603                 sense_ptr, sense_sz);
2604
2605         sense_len -= sense_sz;
2606         sense_ptr += sense_sz;
2607
2608         SET_CMD_SENSE_PTR(sp, sense_ptr);
2609         SET_CMD_SENSE_LEN(sp, sense_len);
2610
2611         /* Place command on done queue. */
2612         if (sense_len == 0) {
2613                 rsp->status_srb = NULL;
2614                 sp->done(sp, cp->result);
2615         }
2616 }
2617
2618 /**
2619  * qla2x00_error_entry() - Process an error entry.
2620  * @ha: SCSI driver HA context
2621  * @pkt: Entry pointer
2622  */
2623 static void
2624 qla2x00_error_entry(scsi_qla_host_t *vha, struct rsp_que *rsp, sts_entry_t *pkt)
2625 {
2626         srb_t *sp;
2627         struct qla_hw_data *ha = vha->hw;
2628         const char func[] = "ERROR-IOCB";
2629         uint16_t que = MSW(pkt->handle);
2630         struct req_que *req = NULL;
2631         int res = DID_ERROR << 16;
2632
2633         ql_dbg(ql_dbg_async, vha, 0x502a,
2634             "type of error status in response: 0x%x\n", pkt->entry_status);
2635
2636         if (que >= ha->max_req_queues || !ha->req_q_map[que])
2637                 goto fatal;
2638
2639         req = ha->req_q_map[que];
2640
2641         if (pkt->entry_status & RF_BUSY)
2642                 res = DID_BUS_BUSY << 16;
2643
2644         if (pkt->entry_type == NOTIFY_ACK_TYPE &&
2645             pkt->handle == QLA_TGT_SKIP_HANDLE)
2646                 return;
2647
2648         sp = qla2x00_get_sp_from_handle(vha, func, req, pkt);
2649         if (sp) {
2650                 sp->done(sp, res);
2651                 return;
2652         }
2653 fatal:
2654         ql_log(ql_log_warn, vha, 0x5030,
2655             "Error entry - invalid handle/queue (%04x).\n", que);
2656 }
2657
2658 /**
2659  * qla24xx_mbx_completion() - Process mailbox command completions.
2660  * @ha: SCSI driver HA context
2661  * @mb0: Mailbox0 register
2662  */
2663 static void
2664 qla24xx_mbx_completion(scsi_qla_host_t *vha, uint16_t mb0)
2665 {
2666         uint16_t        cnt;
2667         uint32_t        mboxes;
2668         uint16_t __iomem *wptr;
2669         struct qla_hw_data *ha = vha->hw;
2670         struct device_reg_24xx __iomem *reg = &ha->iobase->isp24;
2671
2672         /* Read all mbox registers? */
2673         mboxes = (1 << ha->mbx_count) - 1;
2674         if (!ha->mcp)
2675                 ql_dbg(ql_dbg_async, vha, 0x504e, "MBX pointer ERROR.\n");
2676         else
2677                 mboxes = ha->mcp->in_mb;
2678
2679         /* Load return mailbox registers. */
2680         ha->flags.mbox_int = 1;
2681         ha->mailbox_out[0] = mb0;
2682         mboxes >>= 1;
2683         wptr = (uint16_t __iomem *)&reg->mailbox1;
2684
2685         for (cnt = 1; cnt < ha->mbx_count; cnt++) {
2686                 if (mboxes & BIT_0)
2687                         ha->mailbox_out[cnt] = RD_REG_WORD(wptr);
2688
2689                 mboxes >>= 1;
2690                 wptr++;
2691         }
2692 }
2693
2694 static void
2695 qla24xx_abort_iocb_entry(scsi_qla_host_t *vha, struct req_que *req,
2696         struct abort_entry_24xx *pkt)
2697 {
2698         const char func[] = "ABT_IOCB";
2699         srb_t *sp;
2700         struct srb_iocb *abt;
2701
2702         sp = qla2x00_get_sp_from_handle(vha, func, req, pkt);
2703         if (!sp)
2704                 return;
2705
2706         abt = &sp->u.iocb_cmd;
2707         abt->u.abt.comp_status = le16_to_cpu(pkt->nport_handle);
2708         sp->done(sp, 0);
2709 }
2710
2711 /**
2712  * qla24xx_process_response_queue() - Process response queue entries.
2713  * @ha: SCSI driver HA context
2714  */
2715 void qla24xx_process_response_queue(struct scsi_qla_host *vha,
2716         struct rsp_que *rsp)
2717 {
2718         struct sts_entry_24xx *pkt;
2719         struct qla_hw_data *ha = vha->hw;
2720
2721         if (!ha->flags.fw_started)
2722                 return;
2723
2724         while (rsp->ring_ptr->signature != RESPONSE_PROCESSED) {
2725                 pkt = (struct sts_entry_24xx *)rsp->ring_ptr;
2726
2727                 rsp->ring_index++;
2728                 if (rsp->ring_index == rsp->length) {
2729                         rsp->ring_index = 0;
2730                         rsp->ring_ptr = rsp->ring;
2731                 } else {
2732                         rsp->ring_ptr++;
2733                 }
2734
2735                 if (pkt->entry_status != 0) {
2736                         qla2x00_error_entry(vha, rsp, (sts_entry_t *) pkt);
2737
2738                         if (qlt_24xx_process_response_error(vha, pkt))
2739                                 goto process_err;
2740
2741                         ((response_t *)pkt)->signature = RESPONSE_PROCESSED;
2742                         wmb();
2743                         continue;
2744                 }
2745 process_err:
2746
2747                 switch (pkt->entry_type) {
2748                 case STATUS_TYPE:
2749                         qla2x00_status_entry(vha, rsp, pkt);
2750                         break;
2751                 case STATUS_CONT_TYPE:
2752                         qla2x00_status_cont_entry(rsp, (sts_cont_entry_t *)pkt);
2753                         break;
2754                 case VP_RPT_ID_IOCB_TYPE:
2755                         qla24xx_report_id_acquisition(vha,
2756                             (struct vp_rpt_id_entry_24xx *)pkt);
2757                         break;
2758                 case LOGINOUT_PORT_IOCB_TYPE:
2759                         qla24xx_logio_entry(vha, rsp->req,
2760                             (struct logio_entry_24xx *)pkt);
2761                         break;
2762                 case CT_IOCB_TYPE:
2763                         qla24xx_els_ct_entry(vha, rsp->req, pkt, CT_IOCB_TYPE);
2764                         break;
2765                 case ELS_IOCB_TYPE:
2766                         qla24xx_els_ct_entry(vha, rsp->req, pkt, ELS_IOCB_TYPE);
2767                         break;
2768                 case ABTS_RECV_24XX:
2769                         if (IS_QLA83XX(ha) || IS_QLA27XX(ha)) {
2770                                 /* ensure that the ATIO queue is empty */
2771                                 qlt_handle_abts_recv(vha, (response_t *)pkt);
2772                                 break;
2773                         } else {
2774                                 /* drop through */
2775                                 qlt_24xx_process_atio_queue(vha, 1);
2776                         }
2777                 case ABTS_RESP_24XX:
2778                 case CTIO_TYPE7:
2779                 case CTIO_CRC2:
2780                         qlt_response_pkt_all_vps(vha, (response_t *)pkt);
2781                         break;
2782                 case NOTIFY_ACK_TYPE:
2783                         if (pkt->handle == QLA_TGT_SKIP_HANDLE)
2784                                 qlt_response_pkt_all_vps(vha, (response_t *)pkt);
2785                         else
2786                                 qla24xxx_nack_iocb_entry(vha, rsp->req,
2787                                         (struct nack_to_isp *)pkt);
2788                         break;
2789                 case MARKER_TYPE:
2790                         /* Do nothing in this case, this check is to prevent it
2791                          * from falling into default case
2792                          */
2793                         break;
2794                 case ABORT_IOCB_TYPE:
2795                         qla24xx_abort_iocb_entry(vha, rsp->req,
2796                             (struct abort_entry_24xx *)pkt);
2797                         break;
2798                 case MBX_IOCB_TYPE:
2799                         qla24xx_mbx_iocb_entry(vha, rsp->req,
2800                             (struct mbx_24xx_entry *)pkt);
2801                         break;
2802                 default:
2803                         /* Type Not Supported. */
2804                         ql_dbg(ql_dbg_async, vha, 0x5042,
2805                             "Received unknown response pkt type %x "
2806                             "entry status=%x.\n",
2807                             pkt->entry_type, pkt->entry_status);
2808                         break;
2809                 }
2810                 ((response_t *)pkt)->signature = RESPONSE_PROCESSED;
2811                 wmb();
2812         }
2813
2814         /* Adjust ring index */
2815         if (IS_P3P_TYPE(ha)) {
2816                 struct device_reg_82xx __iomem *reg = &ha->iobase->isp82;
2817                 WRT_REG_DWORD(&reg->rsp_q_out[0], rsp->ring_index);
2818         } else {
2819                 WRT_REG_DWORD(rsp->rsp_q_out, rsp->ring_index);
2820         }
2821 }
2822
2823 static void
2824 qla2xxx_check_risc_status(scsi_qla_host_t *vha)
2825 {
2826         int rval;
2827         uint32_t cnt;
2828         struct qla_hw_data *ha = vha->hw;
2829         struct device_reg_24xx __iomem *reg = &ha->iobase->isp24;
2830
2831         if (!IS_QLA25XX(ha) && !IS_QLA81XX(ha) && !IS_QLA83XX(ha) &&
2832             !IS_QLA27XX(ha))
2833                 return;
2834
2835         rval = QLA_SUCCESS;
2836         WRT_REG_DWORD(&reg->iobase_addr, 0x7C00);
2837         RD_REG_DWORD(&reg->iobase_addr);
2838         WRT_REG_DWORD(&reg->iobase_window, 0x0001);
2839         for (cnt = 10000; (RD_REG_DWORD(&reg->iobase_window) & BIT_0) == 0 &&
2840             rval == QLA_SUCCESS; cnt--) {
2841                 if (cnt) {
2842                         WRT_REG_DWORD(&reg->iobase_window, 0x0001);
2843                         udelay(10);
2844                 } else
2845                         rval = QLA_FUNCTION_TIMEOUT;
2846         }
2847         if (rval == QLA_SUCCESS)
2848                 goto next_test;
2849
2850         rval = QLA_SUCCESS;
2851         WRT_REG_DWORD(&reg->iobase_window, 0x0003);
2852         for (cnt = 100; (RD_REG_DWORD(&reg->iobase_window) & BIT_0) == 0 &&
2853             rval == QLA_SUCCESS; cnt--) {
2854                 if (cnt) {
2855                         WRT_REG_DWORD(&reg->iobase_window, 0x0003);
2856                         udelay(10);
2857                 } else
2858                         rval = QLA_FUNCTION_TIMEOUT;
2859         }
2860         if (rval != QLA_SUCCESS)
2861                 goto done;
2862
2863 next_test:
2864         if (RD_REG_DWORD(&reg->iobase_c8) & BIT_3)
2865                 ql_log(ql_log_info, vha, 0x504c,
2866                     "Additional code -- 0x55AA.\n");
2867
2868 done:
2869         WRT_REG_DWORD(&reg->iobase_window, 0x0000);
2870         RD_REG_DWORD(&reg->iobase_window);
2871 }
2872
2873 /**
2874  * qla24xx_intr_handler() - Process interrupts for the ISP23xx and ISP24xx.
2875  * @irq:
2876  * @dev_id: SCSI driver HA context
2877  *
2878  * Called by system whenever the host adapter generates an interrupt.
2879  *
2880  * Returns handled flag.
2881  */
2882 irqreturn_t
2883 qla24xx_intr_handler(int irq, void *dev_id)
2884 {
2885         scsi_qla_host_t *vha;
2886         struct qla_hw_data *ha;
2887         struct device_reg_24xx __iomem *reg;
2888         int             status;
2889         unsigned long   iter;
2890         uint32_t        stat;
2891         uint32_t        hccr;
2892         uint16_t        mb[8];
2893         struct rsp_que *rsp;
2894         unsigned long   flags;
2895
2896         rsp = (struct rsp_que *) dev_id;
2897         if (!rsp) {
2898                 ql_log(ql_log_info, NULL, 0x5059,
2899                     "%s: NULL response queue pointer.\n", __func__);
2900                 return IRQ_NONE;
2901         }
2902
2903         ha = rsp->hw;
2904         reg = &ha->iobase->isp24;
2905         status = 0;
2906
2907         if (unlikely(pci_channel_offline(ha->pdev)))
2908                 return IRQ_HANDLED;
2909
2910         spin_lock_irqsave(&ha->hardware_lock, flags);
2911         vha = pci_get_drvdata(ha->pdev);
2912         for (iter = 50; iter--; ) {
2913                 stat = RD_REG_DWORD(&reg->host_status);
2914                 if (qla2x00_check_reg32_for_disconnect(vha, stat))
2915                         break;
2916                 if (stat & HSRX_RISC_PAUSED) {
2917                         if (unlikely(pci_channel_offline(ha->pdev)))
2918                                 break;
2919
2920                         hccr = RD_REG_DWORD(&reg->hccr);
2921
2922                         ql_log(ql_log_warn, vha, 0x504b,
2923                             "RISC paused -- HCCR=%x, Dumping firmware.\n",
2924                             hccr);
2925
2926                         qla2xxx_check_risc_status(vha);
2927
2928                         ha->isp_ops->fw_dump(vha, 1);
2929                         set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
2930                         break;
2931                 } else if ((stat & HSRX_RISC_INT) == 0)
2932                         break;
2933
2934                 switch (stat & 0xff) {
2935                 case INTR_ROM_MB_SUCCESS:
2936                 case INTR_ROM_MB_FAILED:
2937                 case INTR_MB_SUCCESS:
2938                 case INTR_MB_FAILED:
2939                         qla24xx_mbx_completion(vha, MSW(stat));
2940                         status |= MBX_INTERRUPT;
2941
2942                         break;
2943                 case INTR_ASYNC_EVENT:
2944                         mb[0] = MSW(stat);
2945                         mb[1] = RD_REG_WORD(&reg->mailbox1);
2946                         mb[2] = RD_REG_WORD(&reg->mailbox2);
2947                         mb[3] = RD_REG_WORD(&reg->mailbox3);
2948                         qla2x00_async_event(vha, rsp, mb);
2949                         break;
2950                 case INTR_RSP_QUE_UPDATE:
2951                 case INTR_RSP_QUE_UPDATE_83XX:
2952                         qla24xx_process_response_queue(vha, rsp);
2953                         break;
2954                 case INTR_ATIO_QUE_UPDATE:{
2955                         unsigned long flags2;
2956                         spin_lock_irqsave(&ha->tgt.atio_lock, flags2);
2957                         qlt_24xx_process_atio_queue(vha, 1);
2958                         spin_unlock_irqrestore(&ha->tgt.atio_lock, flags2);
2959                         break;
2960                 }
2961                 case INTR_ATIO_RSP_QUE_UPDATE: {
2962                         unsigned long flags2;
2963                         spin_lock_irqsave(&ha->tgt.atio_lock, flags2);
2964                         qlt_24xx_process_atio_queue(vha, 1);
2965                         spin_unlock_irqrestore(&ha->tgt.atio_lock, flags2);
2966
2967                         qla24xx_process_response_queue(vha, rsp);
2968                         break;
2969                 }
2970                 default:
2971                         ql_dbg(ql_dbg_async, vha, 0x504f,
2972                             "Unrecognized interrupt type (%d).\n", stat * 0xff);
2973                         break;
2974                 }
2975                 WRT_REG_DWORD(&reg->hccr, HCCRX_CLR_RISC_INT);
2976                 RD_REG_DWORD_RELAXED(&reg->hccr);
2977                 if (unlikely(IS_QLA83XX(ha) && (ha->pdev->revision == 1)))
2978                         ndelay(3500);
2979         }
2980         qla2x00_handle_mbx_completion(ha, status);
2981         spin_unlock_irqrestore(&ha->hardware_lock, flags);
2982
2983         return IRQ_HANDLED;
2984 }
2985
2986 static irqreturn_t
2987 qla24xx_msix_rsp_q(int irq, void *dev_id)
2988 {
2989         struct qla_hw_data *ha;
2990         struct rsp_que *rsp;
2991         struct device_reg_24xx __iomem *reg;
2992         struct scsi_qla_host *vha;
2993         unsigned long flags;
2994         uint32_t stat = 0;
2995
2996         rsp = (struct rsp_que *) dev_id;
2997         if (!rsp) {
2998                 ql_log(ql_log_info, NULL, 0x505a,
2999                     "%s: NULL response queue pointer.\n", __func__);
3000                 return IRQ_NONE;
3001         }
3002         ha = rsp->hw;
3003         reg = &ha->iobase->isp24;
3004
3005         spin_lock_irqsave(&ha->hardware_lock, flags);
3006
3007         vha = pci_get_drvdata(ha->pdev);
3008         /*
3009          * Use host_status register to check to PCI disconnection before we
3010          * we process the response queue.
3011          */
3012         stat = RD_REG_DWORD(&reg->host_status);
3013         if (qla2x00_check_reg32_for_disconnect(vha, stat))
3014                 goto out;
3015         qla24xx_process_response_queue(vha, rsp);
3016         if (!ha->flags.disable_msix_handshake) {
3017                 WRT_REG_DWORD(&reg->hccr, HCCRX_CLR_RISC_INT);
3018                 RD_REG_DWORD_RELAXED(&reg->hccr);
3019         }
3020 out:
3021         spin_unlock_irqrestore(&ha->hardware_lock, flags);
3022
3023         return IRQ_HANDLED;
3024 }
3025
3026 static irqreturn_t
3027 qla24xx_msix_default(int irq, void *dev_id)
3028 {
3029         scsi_qla_host_t *vha;
3030         struct qla_hw_data *ha;
3031         struct rsp_que *rsp;
3032         struct device_reg_24xx __iomem *reg;
3033         int             status;
3034         uint32_t        stat;
3035         uint32_t        hccr;
3036         uint16_t        mb[8];
3037         unsigned long flags;
3038
3039         rsp = (struct rsp_que *) dev_id;
3040         if (!rsp) {
3041                 ql_log(ql_log_info, NULL, 0x505c,
3042                     "%s: NULL response queue pointer.\n", __func__);
3043                 return IRQ_NONE;
3044         }
3045         ha = rsp->hw;
3046         reg = &ha->iobase->isp24;
3047         status = 0;
3048
3049         spin_lock_irqsave(&ha->hardware_lock, flags);
3050         vha = pci_get_drvdata(ha->pdev);
3051         do {
3052                 stat = RD_REG_DWORD(&reg->host_status);
3053                 if (qla2x00_check_reg32_for_disconnect(vha, stat))
3054                         break;
3055                 if (stat & HSRX_RISC_PAUSED) {
3056                         if (unlikely(pci_channel_offline(ha->pdev)))
3057                                 break;
3058
3059                         hccr = RD_REG_DWORD(&reg->hccr);
3060
3061                         ql_log(ql_log_info, vha, 0x5050,
3062                             "RISC paused -- HCCR=%x, Dumping firmware.\n",
3063                             hccr);
3064
3065                         qla2xxx_check_risc_status(vha);
3066
3067                         ha->isp_ops->fw_dump(vha, 1);
3068                         set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
3069                         break;
3070                 } else if ((stat & HSRX_RISC_INT) == 0)
3071                         break;
3072
3073                 switch (stat & 0xff) {
3074                 case INTR_ROM_MB_SUCCESS:
3075                 case INTR_ROM_MB_FAILED:
3076                 case INTR_MB_SUCCESS:
3077                 case INTR_MB_FAILED:
3078                         qla24xx_mbx_completion(vha, MSW(stat));
3079                         status |= MBX_INTERRUPT;
3080
3081                         break;
3082                 case INTR_ASYNC_EVENT:
3083                         mb[0] = MSW(stat);
3084                         mb[1] = RD_REG_WORD(&reg->mailbox1);
3085                         mb[2] = RD_REG_WORD(&reg->mailbox2);
3086                         mb[3] = RD_REG_WORD(&reg->mailbox3);
3087                         qla2x00_async_event(vha, rsp, mb);
3088                         break;
3089                 case INTR_RSP_QUE_UPDATE:
3090                 case INTR_RSP_QUE_UPDATE_83XX:
3091                         qla24xx_process_response_queue(vha, rsp);
3092                         break;
3093                 case INTR_ATIO_QUE_UPDATE:{
3094                         unsigned long flags2;
3095                         spin_lock_irqsave(&ha->tgt.atio_lock, flags2);
3096                         qlt_24xx_process_atio_queue(vha, 1);
3097                         spin_unlock_irqrestore(&ha->tgt.atio_lock, flags2);
3098                         break;
3099                 }
3100                 case INTR_ATIO_RSP_QUE_UPDATE: {
3101                         unsigned long flags2;
3102                         spin_lock_irqsave(&ha->tgt.atio_lock, flags2);
3103                         qlt_24xx_process_atio_queue(vha, 1);
3104                         spin_unlock_irqrestore(&ha->tgt.atio_lock, flags2);
3105
3106                         qla24xx_process_response_queue(vha, rsp);
3107                         break;
3108                 }
3109                 default:
3110                         ql_dbg(ql_dbg_async, vha, 0x5051,
3111                             "Unrecognized interrupt type (%d).\n", stat & 0xff);
3112                         break;
3113                 }
3114                 WRT_REG_DWORD(&reg->hccr, HCCRX_CLR_RISC_INT);
3115         } while (0);
3116         qla2x00_handle_mbx_completion(ha, status);
3117         spin_unlock_irqrestore(&ha->hardware_lock, flags);
3118
3119         return IRQ_HANDLED;
3120 }
3121
3122 irqreturn_t
3123 qla2xxx_msix_rsp_q(int irq, void *dev_id)
3124 {
3125         struct qla_hw_data *ha;
3126         struct qla_qpair *qpair;
3127         struct device_reg_24xx __iomem *reg;
3128         unsigned long flags;
3129
3130         qpair = dev_id;
3131         if (!qpair) {
3132                 ql_log(ql_log_info, NULL, 0x505b,
3133                     "%s: NULL response queue pointer.\n", __func__);
3134                 return IRQ_NONE;
3135         }
3136         ha = qpair->hw;
3137
3138         /* Clear the interrupt, if enabled, for this response queue */
3139         if (unlikely(!ha->flags.disable_msix_handshake)) {
3140                 reg = &ha->iobase->isp24;
3141                 spin_lock_irqsave(&ha->hardware_lock, flags);
3142                 WRT_REG_DWORD(&reg->hccr, HCCRX_CLR_RISC_INT);
3143                 spin_unlock_irqrestore(&ha->hardware_lock, flags);
3144         }
3145
3146         queue_work(ha->wq, &qpair->q_work);
3147
3148         return IRQ_HANDLED;
3149 }
3150
3151 /* Interrupt handling helpers. */
3152
3153 struct qla_init_msix_entry {
3154         const char *name;
3155         irq_handler_t handler;
3156 };
3157
3158 static const struct qla_init_msix_entry msix_entries[] = {
3159         { "qla2xxx (default)", qla24xx_msix_default },
3160         { "qla2xxx (rsp_q)", qla24xx_msix_rsp_q },
3161         { "qla2xxx (atio_q)", qla83xx_msix_atio_q },
3162         { "qla2xxx (qpair_multiq)", qla2xxx_msix_rsp_q },
3163 };
3164
3165 static const struct qla_init_msix_entry qla82xx_msix_entries[] = {
3166         { "qla2xxx (default)", qla82xx_msix_default },
3167         { "qla2xxx (rsp_q)", qla82xx_msix_rsp_q },
3168 };
3169
3170 static int
3171 qla24xx_enable_msix(struct qla_hw_data *ha, struct rsp_que *rsp)
3172 {
3173         int i, ret;
3174         struct qla_msix_entry *qentry;
3175         scsi_qla_host_t *vha = pci_get_drvdata(ha->pdev);
3176         int min_vecs = QLA_BASE_VECTORS;
3177         struct irq_affinity desc = {
3178                 .pre_vectors = QLA_BASE_VECTORS,
3179         };
3180
3181         if (QLA_TGT_MODE_ENABLED() && IS_ATIO_MSIX_CAPABLE(ha)) {
3182                 desc.pre_vectors++;
3183                 min_vecs++;
3184         }
3185
3186         ret = pci_alloc_irq_vectors_affinity(ha->pdev, min_vecs,
3187                         ha->msix_count, PCI_IRQ_MSIX | PCI_IRQ_AFFINITY,
3188                         &desc);
3189
3190         if (ret < 0) {
3191                 ql_log(ql_log_fatal, vha, 0x00c7,
3192                     "MSI-X: Failed to enable support, "
3193                     "giving   up -- %d/%d.\n",
3194                     ha->msix_count, ret);
3195                 goto msix_out;
3196         } else if (ret < ha->msix_count) {
3197                 ql_log(ql_log_warn, vha, 0x00c6,
3198                     "MSI-X: Failed to enable support "
3199                      "with %d vectors, using %d vectors.\n",
3200                     ha->msix_count, ret);
3201                 ha->msix_count = ret;
3202                 /* Recalculate queue values */
3203                 if (ha->mqiobase && ql2xmqsupport) {
3204                         ha->max_req_queues = ha->msix_count - 1;
3205
3206                         /* ATIOQ needs 1 vector. That's 1 less QPair */
3207                         if (QLA_TGT_MODE_ENABLED())
3208                                 ha->max_req_queues--;
3209
3210                         ha->max_rsp_queues = ha->max_req_queues;
3211
3212                         ha->max_qpairs = ha->max_req_queues - 1;
3213                         ql_dbg_pci(ql_dbg_init, ha->pdev, 0x0190,
3214                             "Adjusted Max no of queues pairs: %d.\n", ha->max_qpairs);
3215                 }
3216         }
3217         ha->msix_entries = kzalloc(sizeof(struct qla_msix_entry) *
3218                                 ha->msix_count, GFP_KERNEL);
3219         if (!ha->msix_entries) {
3220                 ql_log(ql_log_fatal, vha, 0x00c8,
3221                     "Failed to allocate memory for ha->msix_entries.\n");
3222                 ret = -ENOMEM;
3223                 goto msix_out;
3224         }
3225         ha->flags.msix_enabled = 1;
3226
3227         for (i = 0; i < ha->msix_count; i++) {
3228                 qentry = &ha->msix_entries[i];
3229                 qentry->vector = pci_irq_vector(ha->pdev, i);
3230                 qentry->entry = i;
3231                 qentry->have_irq = 0;
3232                 qentry->in_use = 0;
3233                 qentry->handle = NULL;
3234         }
3235
3236         /* Enable MSI-X vectors for the base queue */
3237         for (i = 0; i < QLA_BASE_VECTORS; i++) {
3238                 qentry = &ha->msix_entries[i];
3239                 qentry->handle = rsp;
3240                 rsp->msix = qentry;
3241                 scnprintf(qentry->name, sizeof(qentry->name),
3242                     "%s", msix_entries[i].name);
3243                 if (IS_P3P_TYPE(ha))
3244                         ret = request_irq(qentry->vector,
3245                                 qla82xx_msix_entries[i].handler,
3246                                 0, qla82xx_msix_entries[i].name, rsp);
3247                 else
3248                         ret = request_irq(qentry->vector,
3249                                 msix_entries[i].handler,
3250                                 0, msix_entries[i].name, rsp);
3251                 if (ret)
3252                         goto msix_register_fail;
3253                 qentry->have_irq = 1;
3254                 qentry->in_use = 1;
3255         }
3256
3257         /*
3258          * If target mode is enable, also request the vector for the ATIO
3259          * queue.
3260          */
3261         if (QLA_TGT_MODE_ENABLED() && IS_ATIO_MSIX_CAPABLE(ha)) {
3262                 qentry = &ha->msix_entries[QLA_ATIO_VECTOR];
3263                 rsp->msix = qentry;
3264                 qentry->handle = rsp;
3265                 scnprintf(qentry->name, sizeof(qentry->name),
3266                     "%s", msix_entries[QLA_ATIO_VECTOR].name);
3267                 qentry->in_use = 1;
3268                 ret = request_irq(qentry->vector,
3269                         msix_entries[QLA_ATIO_VECTOR].handler,
3270                         0, msix_entries[QLA_ATIO_VECTOR].name, rsp);
3271                 qentry->have_irq = 1;
3272         }
3273
3274 msix_register_fail:
3275         if (ret) {
3276                 ql_log(ql_log_fatal, vha, 0x00cb,
3277                     "MSI-X: unable to register handler -- %x/%d.\n",
3278                     qentry->vector, ret);
3279                 qla2x00_free_irqs(vha);
3280                 ha->mqenable = 0;
3281                 goto msix_out;
3282         }
3283
3284         /* Enable MSI-X vector for response queue update for queue 0 */
3285         if (IS_QLA83XX(ha) || IS_QLA27XX(ha)) {
3286                 if (ha->msixbase && ha->mqiobase &&
3287                     (ha->max_rsp_queues > 1 || ha->max_req_queues > 1 ||
3288                      ql2xmqsupport))
3289                         ha->mqenable = 1;
3290         } else
3291                 if (ha->mqiobase &&
3292                     (ha->max_rsp_queues > 1 || ha->max_req_queues > 1 ||
3293                      ql2xmqsupport))
3294                         ha->mqenable = 1;
3295         ql_dbg(ql_dbg_multiq, vha, 0xc005,
3296             "mqiobase=%p, max_rsp_queues=%d, max_req_queues=%d.\n",
3297             ha->mqiobase, ha->max_rsp_queues, ha->max_req_queues);
3298         ql_dbg(ql_dbg_init, vha, 0x0055,
3299             "mqiobase=%p, max_rsp_queues=%d, max_req_queues=%d.\n",
3300             ha->mqiobase, ha->max_rsp_queues, ha->max_req_queues);
3301
3302 msix_out:
3303         return ret;
3304 }
3305
3306 int
3307 qla2x00_request_irqs(struct qla_hw_data *ha, struct rsp_que *rsp)
3308 {
3309         int ret = QLA_FUNCTION_FAILED;
3310         device_reg_t *reg = ha->iobase;
3311         scsi_qla_host_t *vha = pci_get_drvdata(ha->pdev);
3312
3313         /* If possible, enable MSI-X. */
3314         if (!IS_QLA2432(ha) && !IS_QLA2532(ha) && !IS_QLA8432(ha) &&
3315             !IS_CNA_CAPABLE(ha) && !IS_QLA2031(ha) && !IS_QLAFX00(ha) &&
3316             !IS_QLA27XX(ha))
3317                 goto skip_msi;
3318
3319         if (ha->pdev->subsystem_vendor == PCI_VENDOR_ID_HP &&
3320                 (ha->pdev->subsystem_device == 0x7040 ||
3321                 ha->pdev->subsystem_device == 0x7041 ||
3322                 ha->pdev->subsystem_device == 0x1705)) {
3323                 ql_log(ql_log_warn, vha, 0x0034,
3324                     "MSI-X: Unsupported ISP 2432 SSVID/SSDID (0x%X,0x%X).\n",
3325                         ha->pdev->subsystem_vendor,
3326                         ha->pdev->subsystem_device);
3327                 goto skip_msi;
3328         }
3329
3330         if (IS_QLA2432(ha) && (ha->pdev->revision < QLA_MSIX_CHIP_REV_24XX)) {
3331                 ql_log(ql_log_warn, vha, 0x0035,
3332                     "MSI-X; Unsupported ISP2432 (0x%X, 0x%X).\n",
3333                     ha->pdev->revision, QLA_MSIX_CHIP_REV_24XX);
3334                 goto skip_msix;
3335         }
3336
3337         ret = qla24xx_enable_msix(ha, rsp);
3338         if (!ret) {
3339                 ql_dbg(ql_dbg_init, vha, 0x0036,
3340                     "MSI-X: Enabled (0x%X, 0x%X).\n",
3341                     ha->chip_revision, ha->fw_attributes);
3342                 goto clear_risc_ints;
3343         }
3344
3345 skip_msix:
3346
3347         ql_log(ql_log_info, vha, 0x0037,
3348             "Falling back-to MSI mode -%d.\n", ret);
3349
3350         if (!IS_QLA24XX(ha) && !IS_QLA2532(ha) && !IS_QLA8432(ha) &&
3351             !IS_QLA8001(ha) && !IS_P3P_TYPE(ha) && !IS_QLAFX00(ha) &&
3352             !IS_QLA27XX(ha))
3353                 goto skip_msi;
3354
3355         ret = pci_alloc_irq_vectors(ha->pdev, 1, 1, PCI_IRQ_MSI);
3356         if (!ret) {
3357                 ql_dbg(ql_dbg_init, vha, 0x0038,
3358                     "MSI: Enabled.\n");
3359                 ha->flags.msi_enabled = 1;
3360         } else
3361                 ql_log(ql_log_warn, vha, 0x0039,
3362                     "Falling back-to INTa mode -- %d.\n", ret);
3363 skip_msi:
3364
3365         /* Skip INTx on ISP82xx. */
3366         if (!ha->flags.msi_enabled && IS_QLA82XX(ha))
3367                 return QLA_FUNCTION_FAILED;
3368
3369         ret = request_irq(ha->pdev->irq, ha->isp_ops->intr_handler,
3370             ha->flags.msi_enabled ? 0 : IRQF_SHARED,
3371             QLA2XXX_DRIVER_NAME, rsp);
3372         if (ret) {
3373                 ql_log(ql_log_warn, vha, 0x003a,
3374                     "Failed to reserve interrupt %d already in use.\n",
3375                     ha->pdev->irq);
3376                 goto fail;
3377         } else if (!ha->flags.msi_enabled) {
3378                 ql_dbg(ql_dbg_init, vha, 0x0125,
3379                     "INTa mode: Enabled.\n");
3380                 ha->flags.mr_intr_valid = 1;
3381         }
3382
3383 clear_risc_ints:
3384         if (IS_FWI2_CAPABLE(ha) || IS_QLAFX00(ha))
3385                 goto fail;
3386
3387         spin_lock_irq(&ha->hardware_lock);
3388         WRT_REG_WORD(&reg->isp.semaphore, 0);
3389         spin_unlock_irq(&ha->hardware_lock);
3390
3391 fail:
3392         return ret;
3393 }
3394
3395 void
3396 qla2x00_free_irqs(scsi_qla_host_t *vha)
3397 {
3398         struct qla_hw_data *ha = vha->hw;
3399         struct rsp_que *rsp;
3400         struct qla_msix_entry *qentry;
3401         int i;
3402
3403         /*
3404          * We need to check that ha->rsp_q_map is valid in case we are called
3405          * from a probe failure context.
3406          */
3407         if (!ha->rsp_q_map || !ha->rsp_q_map[0])
3408                 goto free_irqs;
3409         rsp = ha->rsp_q_map[0];
3410
3411         if (ha->flags.msix_enabled) {
3412                 for (i = 0; i < ha->msix_count; i++) {
3413                         qentry = &ha->msix_entries[i];
3414                         if (qentry->have_irq) {
3415                                 irq_set_affinity_notifier(qentry->vector, NULL);
3416                                 free_irq(pci_irq_vector(ha->pdev, i), qentry->handle);
3417                         }
3418                 }
3419                 kfree(ha->msix_entries);
3420                 ha->msix_entries = NULL;
3421                 ha->flags.msix_enabled = 0;
3422                 ql_dbg(ql_dbg_init, vha, 0x0042,
3423                         "Disabled MSI-X.\n");
3424         } else {
3425                 free_irq(pci_irq_vector(ha->pdev, 0), rsp);
3426         }
3427
3428 free_irqs:
3429         pci_free_irq_vectors(ha->pdev);
3430 }
3431
3432 int qla25xx_request_irq(struct qla_hw_data *ha, struct qla_qpair *qpair,
3433         struct qla_msix_entry *msix, int vector_type)
3434 {
3435         const struct qla_init_msix_entry *intr = &msix_entries[vector_type];
3436         scsi_qla_host_t *vha = pci_get_drvdata(ha->pdev);
3437         int ret;
3438
3439         scnprintf(msix->name, sizeof(msix->name),
3440             "qla2xxx%lu_qpair%d", vha->host_no, qpair->id);
3441         ret = request_irq(msix->vector, intr->handler, 0, msix->name, qpair);
3442         if (ret) {
3443                 ql_log(ql_log_fatal, vha, 0x00e6,
3444                     "MSI-X: Unable to register handler -- %x/%d.\n",
3445                     msix->vector, ret);
3446                 return ret;
3447         }
3448         msix->have_irq = 1;
3449         msix->handle = qpair;
3450         return ret;
3451 }