[SCSI] bfa: replace bfa_assert with WARN_ON
[linux-2.6/linux-acpi-2.6/ibm-acpi-2.6.git] / drivers / scsi / bfa / bfa_ioc_ct.c
blobccfecfca07451232c74fff3690caced66723e0cc
1 /*
2 * Copyright (c) 2005-2010 Brocade Communications Systems, Inc.
3 * All rights reserved
4 * www.brocade.com
6 * Linux driver for Brocade Fibre Channel Host Bus Adapter.
8 * This program is free software; you can redistribute it and/or modify it
9 * under the terms of the GNU General Public License (GPL) Version 2 as
10 * published by the Free Software Foundation
12 * This program is distributed in the hope that it will be useful, but
13 * WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * General Public License for more details.
18 #include "bfad_drv.h"
19 #include "bfa_ioc.h"
20 #include "bfi_ctreg.h"
21 #include "bfa_defs.h"
23 BFA_TRC_FILE(CNA, IOC_CT);
25 #define bfa_ioc_ct_sync_pos(__ioc) \
26 ((uint32_t) (1 << bfa_ioc_pcifn(__ioc)))
27 #define BFA_IOC_SYNC_REQD_SH 16
28 #define bfa_ioc_ct_get_sync_ackd(__val) (__val & 0x0000ffff)
29 #define bfa_ioc_ct_clear_sync_ackd(__val) (__val & 0xffff0000)
30 #define bfa_ioc_ct_get_sync_reqd(__val) (__val >> BFA_IOC_SYNC_REQD_SH)
31 #define bfa_ioc_ct_sync_reqd_pos(__ioc) \
32 (bfa_ioc_ct_sync_pos(__ioc) << BFA_IOC_SYNC_REQD_SH)
35 * forward declarations
37 static bfa_boolean_t bfa_ioc_ct_firmware_lock(struct bfa_ioc_s *ioc);
38 static void bfa_ioc_ct_firmware_unlock(struct bfa_ioc_s *ioc);
39 static void bfa_ioc_ct_reg_init(struct bfa_ioc_s *ioc);
40 static void bfa_ioc_ct_map_port(struct bfa_ioc_s *ioc);
41 static void bfa_ioc_ct_isr_mode_set(struct bfa_ioc_s *ioc, bfa_boolean_t msix);
42 static void bfa_ioc_ct_notify_fail(struct bfa_ioc_s *ioc);
43 static void bfa_ioc_ct_ownership_reset(struct bfa_ioc_s *ioc);
44 static void bfa_ioc_ct_sync_join(struct bfa_ioc_s *ioc);
45 static void bfa_ioc_ct_sync_leave(struct bfa_ioc_s *ioc);
46 static void bfa_ioc_ct_sync_ack(struct bfa_ioc_s *ioc);
47 static bfa_boolean_t bfa_ioc_ct_sync_complete(struct bfa_ioc_s *ioc);
49 static struct bfa_ioc_hwif_s hwif_ct;
52 * Called from bfa_ioc_attach() to map asic specific calls.
54 void
55 bfa_ioc_set_ct_hwif(struct bfa_ioc_s *ioc)
57 hwif_ct.ioc_pll_init = bfa_ioc_ct_pll_init;
58 hwif_ct.ioc_firmware_lock = bfa_ioc_ct_firmware_lock;
59 hwif_ct.ioc_firmware_unlock = bfa_ioc_ct_firmware_unlock;
60 hwif_ct.ioc_reg_init = bfa_ioc_ct_reg_init;
61 hwif_ct.ioc_map_port = bfa_ioc_ct_map_port;
62 hwif_ct.ioc_isr_mode_set = bfa_ioc_ct_isr_mode_set;
63 hwif_ct.ioc_notify_fail = bfa_ioc_ct_notify_fail;
64 hwif_ct.ioc_ownership_reset = bfa_ioc_ct_ownership_reset;
65 hwif_ct.ioc_sync_join = bfa_ioc_ct_sync_join;
66 hwif_ct.ioc_sync_leave = bfa_ioc_ct_sync_leave;
67 hwif_ct.ioc_sync_ack = bfa_ioc_ct_sync_ack;
68 hwif_ct.ioc_sync_complete = bfa_ioc_ct_sync_complete;
70 ioc->ioc_hwif = &hwif_ct;
74 * Return true if firmware of current driver matches the running firmware.
76 static bfa_boolean_t
77 bfa_ioc_ct_firmware_lock(struct bfa_ioc_s *ioc)
79 enum bfi_ioc_state ioc_fwstate;
80 u32 usecnt;
81 struct bfi_ioc_image_hdr_s fwhdr;
84 * Firmware match check is relevant only for CNA.
86 if (!ioc->cna)
87 return BFA_TRUE;
90 * If bios boot (flash based) -- do not increment usage count
92 if (bfa_cb_image_get_size(BFA_IOC_FWIMG_TYPE(ioc)) <
93 BFA_IOC_FWIMG_MINSZ)
94 return BFA_TRUE;
96 bfa_ioc_sem_get(ioc->ioc_regs.ioc_usage_sem_reg);
97 usecnt = readl(ioc->ioc_regs.ioc_usage_reg);
100 * If usage count is 0, always return TRUE.
102 if (usecnt == 0) {
103 writel(1, ioc->ioc_regs.ioc_usage_reg);
104 writel(1, ioc->ioc_regs.ioc_usage_sem_reg);
105 writel(0, ioc->ioc_regs.ioc_fail_sync);
106 bfa_trc(ioc, usecnt);
107 return BFA_TRUE;
110 ioc_fwstate = readl(ioc->ioc_regs.ioc_fwstate);
111 bfa_trc(ioc, ioc_fwstate);
114 * Use count cannot be non-zero and chip in uninitialized state.
116 WARN_ON(ioc_fwstate == BFI_IOC_UNINIT);
119 * Check if another driver with a different firmware is active
121 bfa_ioc_fwver_get(ioc, &fwhdr);
122 if (!bfa_ioc_fwver_cmp(ioc, &fwhdr)) {
123 writel(1, ioc->ioc_regs.ioc_usage_sem_reg);
124 bfa_trc(ioc, usecnt);
125 return BFA_FALSE;
129 * Same firmware version. Increment the reference count.
131 usecnt++;
132 writel(usecnt, ioc->ioc_regs.ioc_usage_reg);
133 writel(1, ioc->ioc_regs.ioc_usage_sem_reg);
134 bfa_trc(ioc, usecnt);
135 return BFA_TRUE;
138 static void
139 bfa_ioc_ct_firmware_unlock(struct bfa_ioc_s *ioc)
141 u32 usecnt;
144 * Firmware lock is relevant only for CNA.
146 if (!ioc->cna)
147 return;
150 * If bios boot (flash based) -- do not decrement usage count
152 if (bfa_cb_image_get_size(BFA_IOC_FWIMG_TYPE(ioc)) <
153 BFA_IOC_FWIMG_MINSZ)
154 return;
157 * decrement usage count
159 bfa_ioc_sem_get(ioc->ioc_regs.ioc_usage_sem_reg);
160 usecnt = readl(ioc->ioc_regs.ioc_usage_reg);
161 WARN_ON(usecnt <= 0);
163 usecnt--;
164 writel(usecnt, ioc->ioc_regs.ioc_usage_reg);
165 bfa_trc(ioc, usecnt);
167 writel(1, ioc->ioc_regs.ioc_usage_sem_reg);
171 * Notify other functions on HB failure.
173 static void
174 bfa_ioc_ct_notify_fail(struct bfa_ioc_s *ioc)
176 if (ioc->cna) {
177 writel(__FW_INIT_HALT_P, ioc->ioc_regs.ll_halt);
178 writel(__FW_INIT_HALT_P, ioc->ioc_regs.alt_ll_halt);
179 /* Wait for halt to take effect */
180 readl(ioc->ioc_regs.ll_halt);
181 readl(ioc->ioc_regs.alt_ll_halt);
182 } else {
183 writel(__PSS_ERR_STATUS_SET, ioc->ioc_regs.err_set);
184 readl(ioc->ioc_regs.err_set);
189 * Host to LPU mailbox message addresses
191 static struct { u32 hfn_mbox, lpu_mbox, hfn_pgn; } iocreg_fnreg[] = {
192 { HOSTFN0_LPU_MBOX0_0, LPU_HOSTFN0_MBOX0_0, HOST_PAGE_NUM_FN0 },
193 { HOSTFN1_LPU_MBOX0_8, LPU_HOSTFN1_MBOX0_8, HOST_PAGE_NUM_FN1 },
194 { HOSTFN2_LPU_MBOX0_0, LPU_HOSTFN2_MBOX0_0, HOST_PAGE_NUM_FN2 },
195 { HOSTFN3_LPU_MBOX0_8, LPU_HOSTFN3_MBOX0_8, HOST_PAGE_NUM_FN3 }
199 * Host <-> LPU mailbox command/status registers - port 0
201 static struct { u32 hfn, lpu; } iocreg_mbcmd_p0[] = {
202 { HOSTFN0_LPU0_MBOX0_CMD_STAT, LPU0_HOSTFN0_MBOX0_CMD_STAT },
203 { HOSTFN1_LPU0_MBOX0_CMD_STAT, LPU0_HOSTFN1_MBOX0_CMD_STAT },
204 { HOSTFN2_LPU0_MBOX0_CMD_STAT, LPU0_HOSTFN2_MBOX0_CMD_STAT },
205 { HOSTFN3_LPU0_MBOX0_CMD_STAT, LPU0_HOSTFN3_MBOX0_CMD_STAT }
209 * Host <-> LPU mailbox command/status registers - port 1
211 static struct { u32 hfn, lpu; } iocreg_mbcmd_p1[] = {
212 { HOSTFN0_LPU1_MBOX0_CMD_STAT, LPU1_HOSTFN0_MBOX0_CMD_STAT },
213 { HOSTFN1_LPU1_MBOX0_CMD_STAT, LPU1_HOSTFN1_MBOX0_CMD_STAT },
214 { HOSTFN2_LPU1_MBOX0_CMD_STAT, LPU1_HOSTFN2_MBOX0_CMD_STAT },
215 { HOSTFN3_LPU1_MBOX0_CMD_STAT, LPU1_HOSTFN3_MBOX0_CMD_STAT }
218 static void
219 bfa_ioc_ct_reg_init(struct bfa_ioc_s *ioc)
221 void __iomem *rb;
222 int pcifn = bfa_ioc_pcifn(ioc);
224 rb = bfa_ioc_bar0(ioc);
226 ioc->ioc_regs.hfn_mbox = rb + iocreg_fnreg[pcifn].hfn_mbox;
227 ioc->ioc_regs.lpu_mbox = rb + iocreg_fnreg[pcifn].lpu_mbox;
228 ioc->ioc_regs.host_page_num_fn = rb + iocreg_fnreg[pcifn].hfn_pgn;
230 if (ioc->port_id == 0) {
231 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG;
232 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG;
233 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG;
234 ioc->ioc_regs.hfn_mbox_cmd = rb + iocreg_mbcmd_p0[pcifn].hfn;
235 ioc->ioc_regs.lpu_mbox_cmd = rb + iocreg_mbcmd_p0[pcifn].lpu;
236 ioc->ioc_regs.ll_halt = rb + FW_INIT_HALT_P0;
237 ioc->ioc_regs.alt_ll_halt = rb + FW_INIT_HALT_P1;
238 } else {
239 ioc->ioc_regs.heartbeat = (rb + BFA_IOC1_HBEAT_REG);
240 ioc->ioc_regs.ioc_fwstate = (rb + BFA_IOC1_STATE_REG);
241 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC0_STATE_REG;
242 ioc->ioc_regs.hfn_mbox_cmd = rb + iocreg_mbcmd_p1[pcifn].hfn;
243 ioc->ioc_regs.lpu_mbox_cmd = rb + iocreg_mbcmd_p1[pcifn].lpu;
244 ioc->ioc_regs.ll_halt = rb + FW_INIT_HALT_P1;
245 ioc->ioc_regs.alt_ll_halt = rb + FW_INIT_HALT_P0;
249 * PSS control registers
251 ioc->ioc_regs.pss_ctl_reg = (rb + PSS_CTL_REG);
252 ioc->ioc_regs.pss_err_status_reg = (rb + PSS_ERR_STATUS_REG);
253 ioc->ioc_regs.app_pll_fast_ctl_reg = (rb + APP_PLL_425_CTL_REG);
254 ioc->ioc_regs.app_pll_slow_ctl_reg = (rb + APP_PLL_312_CTL_REG);
257 * IOC semaphore registers and serialization
259 ioc->ioc_regs.ioc_sem_reg = (rb + HOST_SEM0_REG);
260 ioc->ioc_regs.ioc_usage_sem_reg = (rb + HOST_SEM1_REG);
261 ioc->ioc_regs.ioc_init_sem_reg = (rb + HOST_SEM2_REG);
262 ioc->ioc_regs.ioc_usage_reg = (rb + BFA_FW_USE_COUNT);
263 ioc->ioc_regs.ioc_fail_sync = (rb + BFA_IOC_FAIL_SYNC);
266 * sram memory access
268 ioc->ioc_regs.smem_page_start = (rb + PSS_SMEM_PAGE_START);
269 ioc->ioc_regs.smem_pg0 = BFI_IOC_SMEM_PG0_CT;
272 * err set reg : for notification of hb failure in fcmode
274 ioc->ioc_regs.err_set = (rb + ERR_SET_REG);
278 * Initialize IOC to port mapping.
281 #define FNC_PERS_FN_SHIFT(__fn) ((__fn) * 8)
282 static void
283 bfa_ioc_ct_map_port(struct bfa_ioc_s *ioc)
285 void __iomem *rb = ioc->pcidev.pci_bar_kva;
286 u32 r32;
289 * For catapult, base port id on personality register and IOC type
291 r32 = readl(rb + FNC_PERS_REG);
292 r32 >>= FNC_PERS_FN_SHIFT(bfa_ioc_pcifn(ioc));
293 ioc->port_id = (r32 & __F0_PORT_MAP_MK) >> __F0_PORT_MAP_SH;
295 bfa_trc(ioc, bfa_ioc_pcifn(ioc));
296 bfa_trc(ioc, ioc->port_id);
300 * Set interrupt mode for a function: INTX or MSIX
302 static void
303 bfa_ioc_ct_isr_mode_set(struct bfa_ioc_s *ioc, bfa_boolean_t msix)
305 void __iomem *rb = ioc->pcidev.pci_bar_kva;
306 u32 r32, mode;
308 r32 = readl(rb + FNC_PERS_REG);
309 bfa_trc(ioc, r32);
311 mode = (r32 >> FNC_PERS_FN_SHIFT(bfa_ioc_pcifn(ioc))) &
312 __F0_INTX_STATUS;
315 * If already in desired mode, do not change anything
317 if (!msix && mode)
318 return;
320 if (msix)
321 mode = __F0_INTX_STATUS_MSIX;
322 else
323 mode = __F0_INTX_STATUS_INTA;
325 r32 &= ~(__F0_INTX_STATUS << FNC_PERS_FN_SHIFT(bfa_ioc_pcifn(ioc)));
326 r32 |= (mode << FNC_PERS_FN_SHIFT(bfa_ioc_pcifn(ioc)));
327 bfa_trc(ioc, r32);
329 writel(r32, rb + FNC_PERS_REG);
333 * Cleanup hw semaphore and usecnt registers
335 static void
336 bfa_ioc_ct_ownership_reset(struct bfa_ioc_s *ioc)
339 if (ioc->cna) {
340 bfa_ioc_sem_get(ioc->ioc_regs.ioc_usage_sem_reg);
341 writel(0, ioc->ioc_regs.ioc_usage_reg);
342 writel(1, ioc->ioc_regs.ioc_usage_sem_reg);
346 * Read the hw sem reg to make sure that it is locked
347 * before we clear it. If it is not locked, writing 1
348 * will lock it instead of clearing it.
350 readl(ioc->ioc_regs.ioc_sem_reg);
351 writel(1, ioc->ioc_regs.ioc_sem_reg);
355 * Synchronized IOC failure processing routines
357 static void
358 bfa_ioc_ct_sync_join(struct bfa_ioc_s *ioc)
360 uint32_t r32 = readl(ioc->ioc_regs.ioc_fail_sync);
361 uint32_t sync_pos = bfa_ioc_ct_sync_reqd_pos(ioc);
363 writel((r32 | sync_pos), ioc->ioc_regs.ioc_fail_sync);
366 static void
367 bfa_ioc_ct_sync_leave(struct bfa_ioc_s *ioc)
369 uint32_t r32 = readl(ioc->ioc_regs.ioc_fail_sync);
370 uint32_t sync_msk = bfa_ioc_ct_sync_reqd_pos(ioc) |
371 bfa_ioc_ct_sync_pos(ioc);
373 writel((r32 & ~sync_msk), ioc->ioc_regs.ioc_fail_sync);
376 static void
377 bfa_ioc_ct_sync_ack(struct bfa_ioc_s *ioc)
379 uint32_t r32 = readl(ioc->ioc_regs.ioc_fail_sync);
381 writel((r32 | bfa_ioc_ct_sync_pos(ioc)),
382 ioc->ioc_regs.ioc_fail_sync);
385 static bfa_boolean_t
386 bfa_ioc_ct_sync_complete(struct bfa_ioc_s *ioc)
388 uint32_t r32 = readl(ioc->ioc_regs.ioc_fail_sync);
389 uint32_t sync_reqd = bfa_ioc_ct_get_sync_reqd(r32);
390 uint32_t sync_ackd = bfa_ioc_ct_get_sync_ackd(r32);
391 uint32_t tmp_ackd;
393 if (sync_ackd == 0)
394 return BFA_TRUE;
397 * The check below is to see whether any other PCI fn
398 * has reinitialized the ASIC (reset sync_ackd bits)
399 * and failed again while this IOC was waiting for hw
400 * semaphore (in bfa_iocpf_sm_semwait()).
402 tmp_ackd = sync_ackd;
403 if ((sync_reqd & bfa_ioc_ct_sync_pos(ioc)) &&
404 !(sync_ackd & bfa_ioc_ct_sync_pos(ioc)))
405 sync_ackd |= bfa_ioc_ct_sync_pos(ioc);
407 if (sync_reqd == sync_ackd) {
408 writel(bfa_ioc_ct_clear_sync_ackd(r32),
409 ioc->ioc_regs.ioc_fail_sync);
410 writel(BFI_IOC_FAIL, ioc->ioc_regs.ioc_fwstate);
411 writel(BFI_IOC_FAIL, ioc->ioc_regs.alt_ioc_fwstate);
412 return BFA_TRUE;
416 * If another PCI fn reinitialized and failed again while
417 * this IOC was waiting for hw sem, the sync_ackd bit for
418 * this IOC need to be set again to allow reinitialization.
420 if (tmp_ackd != sync_ackd)
421 writel((r32 | sync_ackd), ioc->ioc_regs.ioc_fail_sync);
423 return BFA_FALSE;
427 * Check the firmware state to know if pll_init has been completed already
429 bfa_boolean_t
430 bfa_ioc_ct_pll_init_complete(void __iomem *rb)
432 if ((readl(rb + BFA_IOC0_STATE_REG) == BFI_IOC_OP) ||
433 (readl(rb + BFA_IOC1_STATE_REG) == BFI_IOC_OP))
434 return BFA_TRUE;
436 return BFA_FALSE;
439 bfa_status_t
440 bfa_ioc_ct_pll_init(void __iomem *rb, bfa_boolean_t fcmode)
442 u32 pll_sclk, pll_fclk, r32;
444 pll_sclk = __APP_PLL_312_LRESETN | __APP_PLL_312_ENARST |
445 __APP_PLL_312_RSEL200500 | __APP_PLL_312_P0_1(3U) |
446 __APP_PLL_312_JITLMT0_1(3U) |
447 __APP_PLL_312_CNTLMT0_1(1U);
448 pll_fclk = __APP_PLL_425_LRESETN | __APP_PLL_425_ENARST |
449 __APP_PLL_425_RSEL200500 | __APP_PLL_425_P0_1(3U) |
450 __APP_PLL_425_JITLMT0_1(3U) |
451 __APP_PLL_425_CNTLMT0_1(1U);
452 if (fcmode) {
453 writel(0, (rb + OP_MODE));
454 writel(__APP_EMS_CMLCKSEL | __APP_EMS_REFCKBUFEN2 |
455 __APP_EMS_CHANNEL_SEL, (rb + ETH_MAC_SER_REG));
456 } else {
457 writel(__GLOBAL_FCOE_MODE, (rb + OP_MODE));
458 writel(__APP_EMS_REFCKBUFEN1, (rb + ETH_MAC_SER_REG));
460 writel(BFI_IOC_UNINIT, (rb + BFA_IOC0_STATE_REG));
461 writel(BFI_IOC_UNINIT, (rb + BFA_IOC1_STATE_REG));
462 writel(0xffffffffU, (rb + HOSTFN0_INT_MSK));
463 writel(0xffffffffU, (rb + HOSTFN1_INT_MSK));
464 writel(0xffffffffU, (rb + HOSTFN0_INT_STATUS));
465 writel(0xffffffffU, (rb + HOSTFN1_INT_STATUS));
466 writel(0xffffffffU, (rb + HOSTFN0_INT_MSK));
467 writel(0xffffffffU, (rb + HOSTFN1_INT_MSK));
468 writel(pll_sclk | __APP_PLL_312_LOGIC_SOFT_RESET,
469 rb + APP_PLL_312_CTL_REG);
470 writel(pll_fclk | __APP_PLL_425_LOGIC_SOFT_RESET,
471 rb + APP_PLL_425_CTL_REG);
472 writel(pll_sclk | __APP_PLL_312_LOGIC_SOFT_RESET | __APP_PLL_312_ENABLE,
473 rb + APP_PLL_312_CTL_REG);
474 writel(pll_fclk | __APP_PLL_425_LOGIC_SOFT_RESET | __APP_PLL_425_ENABLE,
475 rb + APP_PLL_425_CTL_REG);
476 readl(rb + HOSTFN0_INT_MSK);
477 udelay(2000);
478 writel(0xffffffffU, (rb + HOSTFN0_INT_STATUS));
479 writel(0xffffffffU, (rb + HOSTFN1_INT_STATUS));
480 writel(pll_sclk | __APP_PLL_312_ENABLE, rb + APP_PLL_312_CTL_REG);
481 writel(pll_fclk | __APP_PLL_425_ENABLE, rb + APP_PLL_425_CTL_REG);
482 if (!fcmode) {
483 writel(__PMM_1T_RESET_P, (rb + PMM_1T_RESET_REG_P0));
484 writel(__PMM_1T_RESET_P, (rb + PMM_1T_RESET_REG_P1));
486 r32 = readl((rb + PSS_CTL_REG));
487 r32 &= ~__PSS_LMEM_RESET;
488 writel(r32, (rb + PSS_CTL_REG));
489 udelay(1000);
490 if (!fcmode) {
491 writel(0, (rb + PMM_1T_RESET_REG_P0));
492 writel(0, (rb + PMM_1T_RESET_REG_P1));
495 writel(__EDRAM_BISTR_START, (rb + MBIST_CTL_REG));
496 udelay(1000);
497 r32 = readl((rb + MBIST_STAT_REG));
498 writel(0, (rb + MBIST_CTL_REG));
499 return BFA_STATUS_OK;