pciehp: Mask hotplug interrupt at controller release
[linux-2.6/mini2440.git] / drivers / scsi / mesh.c
blobfd63b06d9ef15d64351f235144ba982486cc5c50
1 /*
2 * SCSI low-level driver for the MESH (Macintosh Enhanced SCSI Hardware)
3 * bus adaptor found on Power Macintosh computers.
4 * We assume the MESH is connected to a DBDMA (descriptor-based DMA)
5 * controller.
7 * Paul Mackerras, August 1996.
8 * Copyright (C) 1996 Paul Mackerras.
10 * Apr. 21 2002 - BenH Rework bus reset code for new error handler
11 * Add delay after initial bus reset
12 * Add module parameters
14 * Sep. 27 2003 - BenH Move to new driver model, fix some write posting
15 * issues
16 * To do:
17 * - handle aborts correctly
18 * - retry arbitration if lost (unless higher levels do this for us)
19 * - power down the chip when no device is detected
21 #include <linux/module.h>
22 #include <linux/kernel.h>
23 #include <linux/delay.h>
24 #include <linux/types.h>
25 #include <linux/string.h>
26 #include <linux/slab.h>
27 #include <linux/blkdev.h>
28 #include <linux/proc_fs.h>
29 #include <linux/stat.h>
30 #include <linux/interrupt.h>
31 #include <linux/reboot.h>
32 #include <linux/spinlock.h>
33 #include <asm/dbdma.h>
34 #include <asm/io.h>
35 #include <asm/pgtable.h>
36 #include <asm/prom.h>
37 #include <asm/system.h>
38 #include <asm/irq.h>
39 #include <asm/hydra.h>
40 #include <asm/processor.h>
41 #include <asm/machdep.h>
42 #include <asm/pmac_feature.h>
43 #include <asm/pci-bridge.h>
44 #include <asm/macio.h>
46 #include <scsi/scsi.h>
47 #include <scsi/scsi_cmnd.h>
48 #include <scsi/scsi_device.h>
49 #include <scsi/scsi_host.h>
51 #include "mesh.h"
53 #if 1
54 #undef KERN_DEBUG
55 #define KERN_DEBUG KERN_WARNING
56 #endif
58 MODULE_AUTHOR("Paul Mackerras (paulus@samba.org)");
59 MODULE_DESCRIPTION("PowerMac MESH SCSI driver");
60 MODULE_LICENSE("GPL");
62 static int sync_rate = CONFIG_SCSI_MESH_SYNC_RATE;
63 static int sync_targets = 0xff;
64 static int resel_targets = 0xff;
65 static int debug_targets = 0; /* print debug for these targets */
66 static int init_reset_delay = CONFIG_SCSI_MESH_RESET_DELAY_MS;
68 module_param(sync_rate, int, 0);
69 MODULE_PARM_DESC(sync_rate, "Synchronous rate (0..10, 0=async)");
70 module_param(sync_targets, int, 0);
71 MODULE_PARM_DESC(sync_targets, "Bitmask of targets allowed to set synchronous");
72 module_param(resel_targets, int, 0);
73 MODULE_PARM_DESC(resel_targets, "Bitmask of targets allowed to set disconnect");
74 module_param(debug_targets, int, 0644);
75 MODULE_PARM_DESC(debug_targets, "Bitmask of debugged targets");
76 module_param(init_reset_delay, int, 0);
77 MODULE_PARM_DESC(init_reset_delay, "Initial bus reset delay (0=no reset)");
79 static int mesh_sync_period = 100;
80 static int mesh_sync_offset = 0;
81 static unsigned char use_active_neg = 0; /* bit mask for SEQ_ACTIVE_NEG if used */
83 #define ALLOW_SYNC(tgt) ((sync_targets >> (tgt)) & 1)
84 #define ALLOW_RESEL(tgt) ((resel_targets >> (tgt)) & 1)
85 #define ALLOW_DEBUG(tgt) ((debug_targets >> (tgt)) & 1)
86 #define DEBUG_TARGET(cmd) ((cmd) && ALLOW_DEBUG((cmd)->device->id))
88 #undef MESH_DBG
89 #define N_DBG_LOG 50
90 #define N_DBG_SLOG 20
91 #define NUM_DBG_EVENTS 13
92 #undef DBG_USE_TB /* bombs on 601 */
94 struct dbglog {
95 char *fmt;
96 u32 tb;
97 u8 phase;
98 u8 bs0;
99 u8 bs1;
100 u8 tgt;
101 int d;
104 enum mesh_phase {
105 idle,
106 arbitrating,
107 selecting,
108 commanding,
109 dataing,
110 statusing,
111 busfreeing,
112 disconnecting,
113 reselecting,
114 sleeping
117 enum msg_phase {
118 msg_none,
119 msg_out,
120 msg_out_xxx,
121 msg_out_last,
122 msg_in,
123 msg_in_bad,
126 enum sdtr_phase {
127 do_sdtr,
128 sdtr_sent,
129 sdtr_done
132 struct mesh_target {
133 enum sdtr_phase sdtr_state;
134 int sync_params;
135 int data_goes_out; /* guess as to data direction */
136 struct scsi_cmnd *current_req;
137 u32 saved_ptr;
138 #ifdef MESH_DBG
139 int log_ix;
140 int n_log;
141 struct dbglog log[N_DBG_LOG];
142 #endif
145 struct mesh_state {
146 volatile struct mesh_regs __iomem *mesh;
147 int meshintr;
148 volatile struct dbdma_regs __iomem *dma;
149 int dmaintr;
150 struct Scsi_Host *host;
151 struct mesh_state *next;
152 struct scsi_cmnd *request_q;
153 struct scsi_cmnd *request_qtail;
154 enum mesh_phase phase; /* what we're currently trying to do */
155 enum msg_phase msgphase;
156 int conn_tgt; /* target we're connected to */
157 struct scsi_cmnd *current_req; /* req we're currently working on */
158 int data_ptr;
159 int dma_started;
160 int dma_count;
161 int stat;
162 int aborting;
163 int expect_reply;
164 int n_msgin;
165 u8 msgin[16];
166 int n_msgout;
167 int last_n_msgout;
168 u8 msgout[16];
169 struct dbdma_cmd *dma_cmds; /* space for dbdma commands, aligned */
170 dma_addr_t dma_cmd_bus;
171 void *dma_cmd_space;
172 int dma_cmd_size;
173 int clk_freq;
174 struct mesh_target tgts[8];
175 struct macio_dev *mdev;
176 struct pci_dev* pdev;
177 #ifdef MESH_DBG
178 int log_ix;
179 int n_log;
180 struct dbglog log[N_DBG_SLOG];
181 #endif
185 * Driver is too messy, we need a few prototypes...
187 static void mesh_done(struct mesh_state *ms, int start_next);
188 static void mesh_interrupt(struct mesh_state *ms);
189 static void cmd_complete(struct mesh_state *ms);
190 static void set_dma_cmds(struct mesh_state *ms, struct scsi_cmnd *cmd);
191 static void halt_dma(struct mesh_state *ms);
192 static void phase_mismatch(struct mesh_state *ms);
196 * Some debugging & logging routines
199 #ifdef MESH_DBG
201 static inline u32 readtb(void)
203 u32 tb;
205 #ifdef DBG_USE_TB
206 /* Beware: if you enable this, it will crash on 601s. */
207 asm ("mftb %0" : "=r" (tb) : );
208 #else
209 tb = 0;
210 #endif
211 return tb;
214 static void dlog(struct mesh_state *ms, char *fmt, int a)
216 struct mesh_target *tp = &ms->tgts[ms->conn_tgt];
217 struct dbglog *tlp, *slp;
219 tlp = &tp->log[tp->log_ix];
220 slp = &ms->log[ms->log_ix];
221 tlp->fmt = fmt;
222 tlp->tb = readtb();
223 tlp->phase = (ms->msgphase << 4) + ms->phase;
224 tlp->bs0 = ms->mesh->bus_status0;
225 tlp->bs1 = ms->mesh->bus_status1;
226 tlp->tgt = ms->conn_tgt;
227 tlp->d = a;
228 *slp = *tlp;
229 if (++tp->log_ix >= N_DBG_LOG)
230 tp->log_ix = 0;
231 if (tp->n_log < N_DBG_LOG)
232 ++tp->n_log;
233 if (++ms->log_ix >= N_DBG_SLOG)
234 ms->log_ix = 0;
235 if (ms->n_log < N_DBG_SLOG)
236 ++ms->n_log;
239 static void dumplog(struct mesh_state *ms, int t)
241 struct mesh_target *tp = &ms->tgts[t];
242 struct dbglog *lp;
243 int i;
245 if (tp->n_log == 0)
246 return;
247 i = tp->log_ix - tp->n_log;
248 if (i < 0)
249 i += N_DBG_LOG;
250 tp->n_log = 0;
251 do {
252 lp = &tp->log[i];
253 printk(KERN_DEBUG "mesh log %d: bs=%.2x%.2x ph=%.2x ",
254 t, lp->bs1, lp->bs0, lp->phase);
255 #ifdef DBG_USE_TB
256 printk("tb=%10u ", lp->tb);
257 #endif
258 printk(lp->fmt, lp->d);
259 printk("\n");
260 if (++i >= N_DBG_LOG)
261 i = 0;
262 } while (i != tp->log_ix);
265 static void dumpslog(struct mesh_state *ms)
267 struct dbglog *lp;
268 int i;
270 if (ms->n_log == 0)
271 return;
272 i = ms->log_ix - ms->n_log;
273 if (i < 0)
274 i += N_DBG_SLOG;
275 ms->n_log = 0;
276 do {
277 lp = &ms->log[i];
278 printk(KERN_DEBUG "mesh log: bs=%.2x%.2x ph=%.2x t%d ",
279 lp->bs1, lp->bs0, lp->phase, lp->tgt);
280 #ifdef DBG_USE_TB
281 printk("tb=%10u ", lp->tb);
282 #endif
283 printk(lp->fmt, lp->d);
284 printk("\n");
285 if (++i >= N_DBG_SLOG)
286 i = 0;
287 } while (i != ms->log_ix);
290 #else
292 static inline void dlog(struct mesh_state *ms, char *fmt, int a)
294 static inline void dumplog(struct mesh_state *ms, int tgt)
296 static inline void dumpslog(struct mesh_state *ms)
299 #endif /* MESH_DBG */
301 #define MKWORD(a, b, c, d) (((a) << 24) + ((b) << 16) + ((c) << 8) + (d))
303 static void
304 mesh_dump_regs(struct mesh_state *ms)
306 volatile struct mesh_regs __iomem *mr = ms->mesh;
307 volatile struct dbdma_regs __iomem *md = ms->dma;
308 int t;
309 struct mesh_target *tp;
311 printk(KERN_DEBUG "mesh: state at %p, regs at %p, dma at %p\n",
312 ms, mr, md);
313 printk(KERN_DEBUG " ct=%4x seq=%2x bs=%4x fc=%2x "
314 "exc=%2x err=%2x im=%2x int=%2x sp=%2x\n",
315 (mr->count_hi << 8) + mr->count_lo, mr->sequence,
316 (mr->bus_status1 << 8) + mr->bus_status0, mr->fifo_count,
317 mr->exception, mr->error, mr->intr_mask, mr->interrupt,
318 mr->sync_params);
319 while(in_8(&mr->fifo_count))
320 printk(KERN_DEBUG " fifo data=%.2x\n",in_8(&mr->fifo));
321 printk(KERN_DEBUG " dma stat=%x cmdptr=%x\n",
322 in_le32(&md->status), in_le32(&md->cmdptr));
323 printk(KERN_DEBUG " phase=%d msgphase=%d conn_tgt=%d data_ptr=%d\n",
324 ms->phase, ms->msgphase, ms->conn_tgt, ms->data_ptr);
325 printk(KERN_DEBUG " dma_st=%d dma_ct=%d n_msgout=%d\n",
326 ms->dma_started, ms->dma_count, ms->n_msgout);
327 for (t = 0; t < 8; ++t) {
328 tp = &ms->tgts[t];
329 if (tp->current_req == NULL)
330 continue;
331 printk(KERN_DEBUG " target %d: req=%p goes_out=%d saved_ptr=%d\n",
332 t, tp->current_req, tp->data_goes_out, tp->saved_ptr);
338 * Flush write buffers on the bus path to the mesh
340 static inline void mesh_flush_io(volatile struct mesh_regs __iomem *mr)
342 (void)in_8(&mr->mesh_id);
347 * Complete a SCSI command
349 static void mesh_completed(struct mesh_state *ms, struct scsi_cmnd *cmd)
351 (*cmd->scsi_done)(cmd);
355 /* Called with meshinterrupt disabled, initialize the chipset
356 * and eventually do the initial bus reset. The lock must not be
357 * held since we can schedule.
359 static void mesh_init(struct mesh_state *ms)
361 volatile struct mesh_regs __iomem *mr = ms->mesh;
362 volatile struct dbdma_regs __iomem *md = ms->dma;
364 mesh_flush_io(mr);
365 udelay(100);
367 /* Reset controller */
368 out_le32(&md->control, (RUN|PAUSE|FLUSH|WAKE) << 16); /* stop dma */
369 out_8(&mr->exception, 0xff); /* clear all exception bits */
370 out_8(&mr->error, 0xff); /* clear all error bits */
371 out_8(&mr->sequence, SEQ_RESETMESH);
372 mesh_flush_io(mr);
373 udelay(10);
374 out_8(&mr->intr_mask, INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
375 out_8(&mr->source_id, ms->host->this_id);
376 out_8(&mr->sel_timeout, 25); /* 250ms */
377 out_8(&mr->sync_params, ASYNC_PARAMS);
379 if (init_reset_delay) {
380 printk(KERN_INFO "mesh: performing initial bus reset...\n");
382 /* Reset bus */
383 out_8(&mr->bus_status1, BS1_RST); /* assert RST */
384 mesh_flush_io(mr);
385 udelay(30); /* leave it on for >= 25us */
386 out_8(&mr->bus_status1, 0); /* negate RST */
387 mesh_flush_io(mr);
389 /* Wait for bus to come back */
390 msleep(init_reset_delay);
393 /* Reconfigure controller */
394 out_8(&mr->interrupt, 0xff); /* clear all interrupt bits */
395 out_8(&mr->sequence, SEQ_FLUSHFIFO);
396 mesh_flush_io(mr);
397 udelay(1);
398 out_8(&mr->sync_params, ASYNC_PARAMS);
399 out_8(&mr->sequence, SEQ_ENBRESEL);
401 ms->phase = idle;
402 ms->msgphase = msg_none;
406 static void mesh_start_cmd(struct mesh_state *ms, struct scsi_cmnd *cmd)
408 volatile struct mesh_regs __iomem *mr = ms->mesh;
409 int t, id;
411 id = cmd->device->id;
412 ms->current_req = cmd;
413 ms->tgts[id].data_goes_out = cmd->sc_data_direction == DMA_TO_DEVICE;
414 ms->tgts[id].current_req = cmd;
416 #if 1
417 if (DEBUG_TARGET(cmd)) {
418 int i;
419 printk(KERN_DEBUG "mesh_start: %p ser=%lu tgt=%d cmd=",
420 cmd, cmd->serial_number, id);
421 for (i = 0; i < cmd->cmd_len; ++i)
422 printk(" %x", cmd->cmnd[i]);
423 printk(" use_sg=%d buffer=%p bufflen=%u\n",
424 scsi_sg_count(cmd), scsi_sglist(cmd), scsi_bufflen(cmd));
426 #endif
427 if (ms->dma_started)
428 panic("mesh: double DMA start !\n");
430 ms->phase = arbitrating;
431 ms->msgphase = msg_none;
432 ms->data_ptr = 0;
433 ms->dma_started = 0;
434 ms->n_msgout = 0;
435 ms->last_n_msgout = 0;
436 ms->expect_reply = 0;
437 ms->conn_tgt = id;
438 ms->tgts[id].saved_ptr = 0;
439 ms->stat = DID_OK;
440 ms->aborting = 0;
441 #ifdef MESH_DBG
442 ms->tgts[id].n_log = 0;
443 dlog(ms, "start cmd=%x", (int) cmd);
444 #endif
446 /* Off we go */
447 dlog(ms, "about to arb, intr/exc/err/fc=%.8x",
448 MKWORD(mr->interrupt, mr->exception, mr->error, mr->fifo_count));
449 out_8(&mr->interrupt, INT_CMDDONE);
450 out_8(&mr->sequence, SEQ_ENBRESEL);
451 mesh_flush_io(mr);
452 udelay(1);
454 if (in_8(&mr->bus_status1) & (BS1_BSY | BS1_SEL)) {
456 * Some other device has the bus or is arbitrating for it -
457 * probably a target which is about to reselect us.
459 dlog(ms, "busy b4 arb, intr/exc/err/fc=%.8x",
460 MKWORD(mr->interrupt, mr->exception,
461 mr->error, mr->fifo_count));
462 for (t = 100; t > 0; --t) {
463 if ((in_8(&mr->bus_status1) & (BS1_BSY | BS1_SEL)) == 0)
464 break;
465 if (in_8(&mr->interrupt) != 0) {
466 dlog(ms, "intr b4 arb, intr/exc/err/fc=%.8x",
467 MKWORD(mr->interrupt, mr->exception,
468 mr->error, mr->fifo_count));
469 mesh_interrupt(ms);
470 if (ms->phase != arbitrating)
471 return;
473 udelay(1);
475 if (in_8(&mr->bus_status1) & (BS1_BSY | BS1_SEL)) {
476 /* XXX should try again in a little while */
477 ms->stat = DID_BUS_BUSY;
478 ms->phase = idle;
479 mesh_done(ms, 0);
480 return;
485 * Apparently the mesh has a bug where it will assert both its
486 * own bit and the target's bit on the bus during arbitration.
488 out_8(&mr->dest_id, mr->source_id);
491 * There appears to be a race with reselection sometimes,
492 * where a target reselects us just as we issue the
493 * arbitrate command. It seems that then the arbitrate
494 * command just hangs waiting for the bus to be free
495 * without giving us a reselection exception.
496 * The only way I have found to get it to respond correctly
497 * is this: disable reselection before issuing the arbitrate
498 * command, then after issuing it, if it looks like a target
499 * is trying to reselect us, reset the mesh and then enable
500 * reselection.
502 out_8(&mr->sequence, SEQ_DISRESEL);
503 if (in_8(&mr->interrupt) != 0) {
504 dlog(ms, "intr after disresel, intr/exc/err/fc=%.8x",
505 MKWORD(mr->interrupt, mr->exception,
506 mr->error, mr->fifo_count));
507 mesh_interrupt(ms);
508 if (ms->phase != arbitrating)
509 return;
510 dlog(ms, "after intr after disresel, intr/exc/err/fc=%.8x",
511 MKWORD(mr->interrupt, mr->exception,
512 mr->error, mr->fifo_count));
515 out_8(&mr->sequence, SEQ_ARBITRATE);
517 for (t = 230; t > 0; --t) {
518 if (in_8(&mr->interrupt) != 0)
519 break;
520 udelay(1);
522 dlog(ms, "after arb, intr/exc/err/fc=%.8x",
523 MKWORD(mr->interrupt, mr->exception, mr->error, mr->fifo_count));
524 if (in_8(&mr->interrupt) == 0 && (in_8(&mr->bus_status1) & BS1_SEL)
525 && (in_8(&mr->bus_status0) & BS0_IO)) {
526 /* looks like a reselection - try resetting the mesh */
527 dlog(ms, "resel? after arb, intr/exc/err/fc=%.8x",
528 MKWORD(mr->interrupt, mr->exception, mr->error, mr->fifo_count));
529 out_8(&mr->sequence, SEQ_RESETMESH);
530 mesh_flush_io(mr);
531 udelay(10);
532 out_8(&mr->interrupt, INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
533 out_8(&mr->intr_mask, INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
534 out_8(&mr->sequence, SEQ_ENBRESEL);
535 mesh_flush_io(mr);
536 for (t = 10; t > 0 && in_8(&mr->interrupt) == 0; --t)
537 udelay(1);
538 dlog(ms, "tried reset after arb, intr/exc/err/fc=%.8x",
539 MKWORD(mr->interrupt, mr->exception, mr->error, mr->fifo_count));
540 #ifndef MESH_MULTIPLE_HOSTS
541 if (in_8(&mr->interrupt) == 0 && (in_8(&mr->bus_status1) & BS1_SEL)
542 && (in_8(&mr->bus_status0) & BS0_IO)) {
543 printk(KERN_ERR "mesh: controller not responding"
544 " to reselection!\n");
546 * If this is a target reselecting us, and the
547 * mesh isn't responding, the higher levels of
548 * the scsi code will eventually time out and
549 * reset the bus.
552 #endif
557 * Start the next command for a MESH.
558 * Should be called with interrupts disabled.
560 static void mesh_start(struct mesh_state *ms)
562 struct scsi_cmnd *cmd, *prev, *next;
564 if (ms->phase != idle || ms->current_req != NULL) {
565 printk(KERN_ERR "inappropriate mesh_start (phase=%d, ms=%p)",
566 ms->phase, ms);
567 return;
570 while (ms->phase == idle) {
571 prev = NULL;
572 for (cmd = ms->request_q; ; cmd = (struct scsi_cmnd *) cmd->host_scribble) {
573 if (cmd == NULL)
574 return;
575 if (ms->tgts[cmd->device->id].current_req == NULL)
576 break;
577 prev = cmd;
579 next = (struct scsi_cmnd *) cmd->host_scribble;
580 if (prev == NULL)
581 ms->request_q = next;
582 else
583 prev->host_scribble = (void *) next;
584 if (next == NULL)
585 ms->request_qtail = prev;
587 mesh_start_cmd(ms, cmd);
591 static void mesh_done(struct mesh_state *ms, int start_next)
593 struct scsi_cmnd *cmd;
594 struct mesh_target *tp = &ms->tgts[ms->conn_tgt];
596 cmd = ms->current_req;
597 ms->current_req = NULL;
598 tp->current_req = NULL;
599 if (cmd) {
600 cmd->result = (ms->stat << 16) + cmd->SCp.Status;
601 if (ms->stat == DID_OK)
602 cmd->result += (cmd->SCp.Message << 8);
603 if (DEBUG_TARGET(cmd)) {
604 printk(KERN_DEBUG "mesh_done: result = %x, data_ptr=%d, buflen=%d\n",
605 cmd->result, ms->data_ptr, scsi_bufflen(cmd));
606 #if 0
607 /* needs to use sg? */
608 if ((cmd->cmnd[0] == 0 || cmd->cmnd[0] == 0x12 || cmd->cmnd[0] == 3)
609 && cmd->request_buffer != 0) {
610 unsigned char *b = cmd->request_buffer;
611 printk(KERN_DEBUG "buffer = %x %x %x %x %x %x %x %x\n",
612 b[0], b[1], b[2], b[3], b[4], b[5], b[6], b[7]);
614 #endif
616 cmd->SCp.this_residual -= ms->data_ptr;
617 mesh_completed(ms, cmd);
619 if (start_next) {
620 out_8(&ms->mesh->sequence, SEQ_ENBRESEL);
621 mesh_flush_io(ms->mesh);
622 udelay(1);
623 ms->phase = idle;
624 mesh_start(ms);
628 static inline void add_sdtr_msg(struct mesh_state *ms)
630 int i = ms->n_msgout;
632 ms->msgout[i] = EXTENDED_MESSAGE;
633 ms->msgout[i+1] = 3;
634 ms->msgout[i+2] = EXTENDED_SDTR;
635 ms->msgout[i+3] = mesh_sync_period/4;
636 ms->msgout[i+4] = (ALLOW_SYNC(ms->conn_tgt)? mesh_sync_offset: 0);
637 ms->n_msgout = i + 5;
640 static void set_sdtr(struct mesh_state *ms, int period, int offset)
642 struct mesh_target *tp = &ms->tgts[ms->conn_tgt];
643 volatile struct mesh_regs __iomem *mr = ms->mesh;
644 int v, tr;
646 tp->sdtr_state = sdtr_done;
647 if (offset == 0) {
648 /* asynchronous */
649 if (SYNC_OFF(tp->sync_params))
650 printk(KERN_INFO "mesh: target %d now asynchronous\n",
651 ms->conn_tgt);
652 tp->sync_params = ASYNC_PARAMS;
653 out_8(&mr->sync_params, ASYNC_PARAMS);
654 return;
657 * We need to compute ceil(clk_freq * period / 500e6) - 2
658 * without incurring overflow.
660 v = (ms->clk_freq / 5000) * period;
661 if (v <= 250000) {
662 /* special case: sync_period == 5 * clk_period */
663 v = 0;
664 /* units of tr are 100kB/s */
665 tr = (ms->clk_freq + 250000) / 500000;
666 } else {
667 /* sync_period == (v + 2) * 2 * clk_period */
668 v = (v + 99999) / 100000 - 2;
669 if (v > 15)
670 v = 15; /* oops */
671 tr = ((ms->clk_freq / (v + 2)) + 199999) / 200000;
673 if (offset > 15)
674 offset = 15; /* can't happen */
675 tp->sync_params = SYNC_PARAMS(offset, v);
676 out_8(&mr->sync_params, tp->sync_params);
677 printk(KERN_INFO "mesh: target %d synchronous at %d.%d MB/s\n",
678 ms->conn_tgt, tr/10, tr%10);
681 static void start_phase(struct mesh_state *ms)
683 int i, seq, nb;
684 volatile struct mesh_regs __iomem *mr = ms->mesh;
685 volatile struct dbdma_regs __iomem *md = ms->dma;
686 struct scsi_cmnd *cmd = ms->current_req;
687 struct mesh_target *tp = &ms->tgts[ms->conn_tgt];
689 dlog(ms, "start_phase nmo/exc/fc/seq = %.8x",
690 MKWORD(ms->n_msgout, mr->exception, mr->fifo_count, mr->sequence));
691 out_8(&mr->interrupt, INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
692 seq = use_active_neg + (ms->n_msgout? SEQ_ATN: 0);
693 switch (ms->msgphase) {
694 case msg_none:
695 break;
697 case msg_in:
698 out_8(&mr->count_hi, 0);
699 out_8(&mr->count_lo, 1);
700 out_8(&mr->sequence, SEQ_MSGIN + seq);
701 ms->n_msgin = 0;
702 return;
704 case msg_out:
706 * To make sure ATN drops before we assert ACK for
707 * the last byte of the message, we have to do the
708 * last byte specially.
710 if (ms->n_msgout <= 0) {
711 printk(KERN_ERR "mesh: msg_out but n_msgout=%d\n",
712 ms->n_msgout);
713 mesh_dump_regs(ms);
714 ms->msgphase = msg_none;
715 break;
717 if (ALLOW_DEBUG(ms->conn_tgt)) {
718 printk(KERN_DEBUG "mesh: sending %d msg bytes:",
719 ms->n_msgout);
720 for (i = 0; i < ms->n_msgout; ++i)
721 printk(" %x", ms->msgout[i]);
722 printk("\n");
724 dlog(ms, "msgout msg=%.8x", MKWORD(ms->n_msgout, ms->msgout[0],
725 ms->msgout[1], ms->msgout[2]));
726 out_8(&mr->count_hi, 0);
727 out_8(&mr->sequence, SEQ_FLUSHFIFO);
728 mesh_flush_io(mr);
729 udelay(1);
731 * If ATN is not already asserted, we assert it, then
732 * issue a SEQ_MSGOUT to get the mesh to drop ACK.
734 if ((in_8(&mr->bus_status0) & BS0_ATN) == 0) {
735 dlog(ms, "bus0 was %.2x explicitly asserting ATN", mr->bus_status0);
736 out_8(&mr->bus_status0, BS0_ATN); /* explicit ATN */
737 mesh_flush_io(mr);
738 udelay(1);
739 out_8(&mr->count_lo, 1);
740 out_8(&mr->sequence, SEQ_MSGOUT + seq);
741 out_8(&mr->bus_status0, 0); /* release explicit ATN */
742 dlog(ms,"hace: after explicit ATN bus0=%.2x",mr->bus_status0);
744 if (ms->n_msgout == 1) {
746 * We can't issue the SEQ_MSGOUT without ATN
747 * until the target has asserted REQ. The logic
748 * in cmd_complete handles both situations:
749 * REQ already asserted or not.
751 cmd_complete(ms);
752 } else {
753 out_8(&mr->count_lo, ms->n_msgout - 1);
754 out_8(&mr->sequence, SEQ_MSGOUT + seq);
755 for (i = 0; i < ms->n_msgout - 1; ++i)
756 out_8(&mr->fifo, ms->msgout[i]);
758 return;
760 default:
761 printk(KERN_ERR "mesh bug: start_phase msgphase=%d\n",
762 ms->msgphase);
765 switch (ms->phase) {
766 case selecting:
767 out_8(&mr->dest_id, ms->conn_tgt);
768 out_8(&mr->sequence, SEQ_SELECT + SEQ_ATN);
769 break;
770 case commanding:
771 out_8(&mr->sync_params, tp->sync_params);
772 out_8(&mr->count_hi, 0);
773 if (cmd) {
774 out_8(&mr->count_lo, cmd->cmd_len);
775 out_8(&mr->sequence, SEQ_COMMAND + seq);
776 for (i = 0; i < cmd->cmd_len; ++i)
777 out_8(&mr->fifo, cmd->cmnd[i]);
778 } else {
779 out_8(&mr->count_lo, 6);
780 out_8(&mr->sequence, SEQ_COMMAND + seq);
781 for (i = 0; i < 6; ++i)
782 out_8(&mr->fifo, 0);
784 break;
785 case dataing:
786 /* transfer data, if any */
787 if (!ms->dma_started) {
788 set_dma_cmds(ms, cmd);
789 out_le32(&md->cmdptr, virt_to_phys(ms->dma_cmds));
790 out_le32(&md->control, (RUN << 16) | RUN);
791 ms->dma_started = 1;
793 nb = ms->dma_count;
794 if (nb > 0xfff0)
795 nb = 0xfff0;
796 ms->dma_count -= nb;
797 ms->data_ptr += nb;
798 out_8(&mr->count_lo, nb);
799 out_8(&mr->count_hi, nb >> 8);
800 out_8(&mr->sequence, (tp->data_goes_out?
801 SEQ_DATAOUT: SEQ_DATAIN) + SEQ_DMA_MODE + seq);
802 break;
803 case statusing:
804 out_8(&mr->count_hi, 0);
805 out_8(&mr->count_lo, 1);
806 out_8(&mr->sequence, SEQ_STATUS + seq);
807 break;
808 case busfreeing:
809 case disconnecting:
810 out_8(&mr->sequence, SEQ_ENBRESEL);
811 mesh_flush_io(mr);
812 udelay(1);
813 dlog(ms, "enbresel intr/exc/err/fc=%.8x",
814 MKWORD(mr->interrupt, mr->exception, mr->error,
815 mr->fifo_count));
816 out_8(&mr->sequence, SEQ_BUSFREE);
817 break;
818 default:
819 printk(KERN_ERR "mesh: start_phase called with phase=%d\n",
820 ms->phase);
821 dumpslog(ms);
826 static inline void get_msgin(struct mesh_state *ms)
828 volatile struct mesh_regs __iomem *mr = ms->mesh;
829 int i, n;
831 n = mr->fifo_count;
832 if (n != 0) {
833 i = ms->n_msgin;
834 ms->n_msgin = i + n;
835 for (; n > 0; --n)
836 ms->msgin[i++] = in_8(&mr->fifo);
840 static inline int msgin_length(struct mesh_state *ms)
842 int b, n;
844 n = 1;
845 if (ms->n_msgin > 0) {
846 b = ms->msgin[0];
847 if (b == 1) {
848 /* extended message */
849 n = ms->n_msgin < 2? 2: ms->msgin[1] + 2;
850 } else if (0x20 <= b && b <= 0x2f) {
851 /* 2-byte message */
852 n = 2;
855 return n;
858 static void reselected(struct mesh_state *ms)
860 volatile struct mesh_regs __iomem *mr = ms->mesh;
861 struct scsi_cmnd *cmd;
862 struct mesh_target *tp;
863 int b, t, prev;
865 switch (ms->phase) {
866 case idle:
867 break;
868 case arbitrating:
869 if ((cmd = ms->current_req) != NULL) {
870 /* put the command back on the queue */
871 cmd->host_scribble = (void *) ms->request_q;
872 if (ms->request_q == NULL)
873 ms->request_qtail = cmd;
874 ms->request_q = cmd;
875 tp = &ms->tgts[cmd->device->id];
876 tp->current_req = NULL;
878 break;
879 case busfreeing:
880 ms->phase = reselecting;
881 mesh_done(ms, 0);
882 break;
883 case disconnecting:
884 break;
885 default:
886 printk(KERN_ERR "mesh: reselected in phase %d/%d tgt %d\n",
887 ms->msgphase, ms->phase, ms->conn_tgt);
888 dumplog(ms, ms->conn_tgt);
889 dumpslog(ms);
892 if (ms->dma_started) {
893 printk(KERN_ERR "mesh: reselected with DMA started !\n");
894 halt_dma(ms);
896 ms->current_req = NULL;
897 ms->phase = dataing;
898 ms->msgphase = msg_in;
899 ms->n_msgout = 0;
900 ms->last_n_msgout = 0;
901 prev = ms->conn_tgt;
904 * We seem to get abortive reselections sometimes.
906 while ((in_8(&mr->bus_status1) & BS1_BSY) == 0) {
907 static int mesh_aborted_resels;
908 mesh_aborted_resels++;
909 out_8(&mr->interrupt, INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
910 mesh_flush_io(mr);
911 udelay(1);
912 out_8(&mr->sequence, SEQ_ENBRESEL);
913 mesh_flush_io(mr);
914 udelay(5);
915 dlog(ms, "extra resel err/exc/fc = %.6x",
916 MKWORD(0, mr->error, mr->exception, mr->fifo_count));
918 out_8(&mr->interrupt, INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
919 mesh_flush_io(mr);
920 udelay(1);
921 out_8(&mr->sequence, SEQ_ENBRESEL);
922 mesh_flush_io(mr);
923 udelay(1);
924 out_8(&mr->sync_params, ASYNC_PARAMS);
927 * Find out who reselected us.
929 if (in_8(&mr->fifo_count) == 0) {
930 printk(KERN_ERR "mesh: reselection but nothing in fifo?\n");
931 ms->conn_tgt = ms->host->this_id;
932 goto bogus;
934 /* get the last byte in the fifo */
935 do {
936 b = in_8(&mr->fifo);
937 dlog(ms, "reseldata %x", b);
938 } while (in_8(&mr->fifo_count));
939 for (t = 0; t < 8; ++t)
940 if ((b & (1 << t)) != 0 && t != ms->host->this_id)
941 break;
942 if (b != (1 << t) + (1 << ms->host->this_id)) {
943 printk(KERN_ERR "mesh: bad reselection data %x\n", b);
944 ms->conn_tgt = ms->host->this_id;
945 goto bogus;
950 * Set up to continue with that target's transfer.
952 ms->conn_tgt = t;
953 tp = &ms->tgts[t];
954 out_8(&mr->sync_params, tp->sync_params);
955 if (ALLOW_DEBUG(t)) {
956 printk(KERN_DEBUG "mesh: reselected by target %d\n", t);
957 printk(KERN_DEBUG "mesh: saved_ptr=%x goes_out=%d cmd=%p\n",
958 tp->saved_ptr, tp->data_goes_out, tp->current_req);
960 ms->current_req = tp->current_req;
961 if (tp->current_req == NULL) {
962 printk(KERN_ERR "mesh: reselected by tgt %d but no cmd!\n", t);
963 goto bogus;
965 ms->data_ptr = tp->saved_ptr;
966 dlog(ms, "resel prev tgt=%d", prev);
967 dlog(ms, "resel err/exc=%.4x", MKWORD(0, 0, mr->error, mr->exception));
968 start_phase(ms);
969 return;
971 bogus:
972 dumplog(ms, ms->conn_tgt);
973 dumpslog(ms);
974 ms->data_ptr = 0;
975 ms->aborting = 1;
976 start_phase(ms);
979 static void do_abort(struct mesh_state *ms)
981 ms->msgout[0] = ABORT;
982 ms->n_msgout = 1;
983 ms->aborting = 1;
984 ms->stat = DID_ABORT;
985 dlog(ms, "abort", 0);
988 static void handle_reset(struct mesh_state *ms)
990 int tgt;
991 struct mesh_target *tp;
992 struct scsi_cmnd *cmd;
993 volatile struct mesh_regs __iomem *mr = ms->mesh;
995 for (tgt = 0; tgt < 8; ++tgt) {
996 tp = &ms->tgts[tgt];
997 if ((cmd = tp->current_req) != NULL) {
998 cmd->result = DID_RESET << 16;
999 tp->current_req = NULL;
1000 mesh_completed(ms, cmd);
1002 ms->tgts[tgt].sdtr_state = do_sdtr;
1003 ms->tgts[tgt].sync_params = ASYNC_PARAMS;
1005 ms->current_req = NULL;
1006 while ((cmd = ms->request_q) != NULL) {
1007 ms->request_q = (struct scsi_cmnd *) cmd->host_scribble;
1008 cmd->result = DID_RESET << 16;
1009 mesh_completed(ms, cmd);
1011 ms->phase = idle;
1012 ms->msgphase = msg_none;
1013 out_8(&mr->interrupt, INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
1014 out_8(&mr->sequence, SEQ_FLUSHFIFO);
1015 mesh_flush_io(mr);
1016 udelay(1);
1017 out_8(&mr->sync_params, ASYNC_PARAMS);
1018 out_8(&mr->sequence, SEQ_ENBRESEL);
1021 static irqreturn_t do_mesh_interrupt(int irq, void *dev_id)
1023 unsigned long flags;
1024 struct mesh_state *ms = dev_id;
1025 struct Scsi_Host *dev = ms->host;
1027 spin_lock_irqsave(dev->host_lock, flags);
1028 mesh_interrupt(ms);
1029 spin_unlock_irqrestore(dev->host_lock, flags);
1030 return IRQ_HANDLED;
1033 static void handle_error(struct mesh_state *ms)
1035 int err, exc, count;
1036 volatile struct mesh_regs __iomem *mr = ms->mesh;
1038 err = in_8(&mr->error);
1039 exc = in_8(&mr->exception);
1040 out_8(&mr->interrupt, INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
1041 dlog(ms, "error err/exc/fc/cl=%.8x",
1042 MKWORD(err, exc, mr->fifo_count, mr->count_lo));
1043 if (err & ERR_SCSIRESET) {
1044 /* SCSI bus was reset */
1045 printk(KERN_INFO "mesh: SCSI bus reset detected: "
1046 "waiting for end...");
1047 while ((in_8(&mr->bus_status1) & BS1_RST) != 0)
1048 udelay(1);
1049 printk("done\n");
1050 handle_reset(ms);
1051 /* request_q is empty, no point in mesh_start() */
1052 return;
1054 if (err & ERR_UNEXPDISC) {
1055 /* Unexpected disconnect */
1056 if (exc & EXC_RESELECTED) {
1057 reselected(ms);
1058 return;
1060 if (!ms->aborting) {
1061 printk(KERN_WARNING "mesh: target %d aborted\n",
1062 ms->conn_tgt);
1063 dumplog(ms, ms->conn_tgt);
1064 dumpslog(ms);
1066 out_8(&mr->interrupt, INT_CMDDONE);
1067 ms->stat = DID_ABORT;
1068 mesh_done(ms, 1);
1069 return;
1071 if (err & ERR_PARITY) {
1072 if (ms->msgphase == msg_in) {
1073 printk(KERN_ERR "mesh: msg parity error, target %d\n",
1074 ms->conn_tgt);
1075 ms->msgout[0] = MSG_PARITY_ERROR;
1076 ms->n_msgout = 1;
1077 ms->msgphase = msg_in_bad;
1078 cmd_complete(ms);
1079 return;
1081 if (ms->stat == DID_OK) {
1082 printk(KERN_ERR "mesh: parity error, target %d\n",
1083 ms->conn_tgt);
1084 ms->stat = DID_PARITY;
1086 count = (mr->count_hi << 8) + mr->count_lo;
1087 if (count == 0) {
1088 cmd_complete(ms);
1089 } else {
1090 /* reissue the data transfer command */
1091 out_8(&mr->sequence, mr->sequence);
1093 return;
1095 if (err & ERR_SEQERR) {
1096 if (exc & EXC_RESELECTED) {
1097 /* This can happen if we issue a command to
1098 get the bus just after the target reselects us. */
1099 static int mesh_resel_seqerr;
1100 mesh_resel_seqerr++;
1101 reselected(ms);
1102 return;
1104 if (exc == EXC_PHASEMM) {
1105 static int mesh_phasemm_seqerr;
1106 mesh_phasemm_seqerr++;
1107 phase_mismatch(ms);
1108 return;
1110 printk(KERN_ERR "mesh: sequence error (err=%x exc=%x)\n",
1111 err, exc);
1112 } else {
1113 printk(KERN_ERR "mesh: unknown error %x (exc=%x)\n", err, exc);
1115 mesh_dump_regs(ms);
1116 dumplog(ms, ms->conn_tgt);
1117 if (ms->phase > selecting && (in_8(&mr->bus_status1) & BS1_BSY)) {
1118 /* try to do what the target wants */
1119 do_abort(ms);
1120 phase_mismatch(ms);
1121 return;
1123 ms->stat = DID_ERROR;
1124 mesh_done(ms, 1);
1127 static void handle_exception(struct mesh_state *ms)
1129 int exc;
1130 volatile struct mesh_regs __iomem *mr = ms->mesh;
1132 exc = in_8(&mr->exception);
1133 out_8(&mr->interrupt, INT_EXCEPTION | INT_CMDDONE);
1134 if (exc & EXC_RESELECTED) {
1135 static int mesh_resel_exc;
1136 mesh_resel_exc++;
1137 reselected(ms);
1138 } else if (exc == EXC_ARBLOST) {
1139 printk(KERN_DEBUG "mesh: lost arbitration\n");
1140 ms->stat = DID_BUS_BUSY;
1141 mesh_done(ms, 1);
1142 } else if (exc == EXC_SELTO) {
1143 /* selection timed out */
1144 ms->stat = DID_BAD_TARGET;
1145 mesh_done(ms, 1);
1146 } else if (exc == EXC_PHASEMM) {
1147 /* target wants to do something different:
1148 find out what it wants and do it. */
1149 phase_mismatch(ms);
1150 } else {
1151 printk(KERN_ERR "mesh: can't cope with exception %x\n", exc);
1152 mesh_dump_regs(ms);
1153 dumplog(ms, ms->conn_tgt);
1154 do_abort(ms);
1155 phase_mismatch(ms);
1159 static void handle_msgin(struct mesh_state *ms)
1161 int i, code;
1162 struct scsi_cmnd *cmd = ms->current_req;
1163 struct mesh_target *tp = &ms->tgts[ms->conn_tgt];
1165 if (ms->n_msgin == 0)
1166 return;
1167 code = ms->msgin[0];
1168 if (ALLOW_DEBUG(ms->conn_tgt)) {
1169 printk(KERN_DEBUG "got %d message bytes:", ms->n_msgin);
1170 for (i = 0; i < ms->n_msgin; ++i)
1171 printk(" %x", ms->msgin[i]);
1172 printk("\n");
1174 dlog(ms, "msgin msg=%.8x",
1175 MKWORD(ms->n_msgin, code, ms->msgin[1], ms->msgin[2]));
1177 ms->expect_reply = 0;
1178 ms->n_msgout = 0;
1179 if (ms->n_msgin < msgin_length(ms))
1180 goto reject;
1181 if (cmd)
1182 cmd->SCp.Message = code;
1183 switch (code) {
1184 case COMMAND_COMPLETE:
1185 break;
1186 case EXTENDED_MESSAGE:
1187 switch (ms->msgin[2]) {
1188 case EXTENDED_MODIFY_DATA_POINTER:
1189 ms->data_ptr += (ms->msgin[3] << 24) + ms->msgin[6]
1190 + (ms->msgin[4] << 16) + (ms->msgin[5] << 8);
1191 break;
1192 case EXTENDED_SDTR:
1193 if (tp->sdtr_state != sdtr_sent) {
1194 /* reply with an SDTR */
1195 add_sdtr_msg(ms);
1196 /* limit period to at least his value,
1197 offset to no more than his */
1198 if (ms->msgout[3] < ms->msgin[3])
1199 ms->msgout[3] = ms->msgin[3];
1200 if (ms->msgout[4] > ms->msgin[4])
1201 ms->msgout[4] = ms->msgin[4];
1202 set_sdtr(ms, ms->msgout[3], ms->msgout[4]);
1203 ms->msgphase = msg_out;
1204 } else {
1205 set_sdtr(ms, ms->msgin[3], ms->msgin[4]);
1207 break;
1208 default:
1209 goto reject;
1211 break;
1212 case SAVE_POINTERS:
1213 tp->saved_ptr = ms->data_ptr;
1214 break;
1215 case RESTORE_POINTERS:
1216 ms->data_ptr = tp->saved_ptr;
1217 break;
1218 case DISCONNECT:
1219 ms->phase = disconnecting;
1220 break;
1221 case ABORT:
1222 break;
1223 case MESSAGE_REJECT:
1224 if (tp->sdtr_state == sdtr_sent)
1225 set_sdtr(ms, 0, 0);
1226 break;
1227 case NOP:
1228 break;
1229 default:
1230 if (IDENTIFY_BASE <= code && code <= IDENTIFY_BASE + 7) {
1231 if (cmd == NULL) {
1232 do_abort(ms);
1233 ms->msgphase = msg_out;
1234 } else if (code != cmd->device->lun + IDENTIFY_BASE) {
1235 printk(KERN_WARNING "mesh: lun mismatch "
1236 "(%d != %d) on reselection from "
1237 "target %d\n", code - IDENTIFY_BASE,
1238 cmd->device->lun, ms->conn_tgt);
1240 break;
1242 goto reject;
1244 return;
1246 reject:
1247 printk(KERN_WARNING "mesh: rejecting message from target %d:",
1248 ms->conn_tgt);
1249 for (i = 0; i < ms->n_msgin; ++i)
1250 printk(" %x", ms->msgin[i]);
1251 printk("\n");
1252 ms->msgout[0] = MESSAGE_REJECT;
1253 ms->n_msgout = 1;
1254 ms->msgphase = msg_out;
1258 * Set up DMA commands for transferring data.
1260 static void set_dma_cmds(struct mesh_state *ms, struct scsi_cmnd *cmd)
1262 int i, dma_cmd, total, off, dtot;
1263 struct scatterlist *scl;
1264 struct dbdma_cmd *dcmds;
1266 dma_cmd = ms->tgts[ms->conn_tgt].data_goes_out?
1267 OUTPUT_MORE: INPUT_MORE;
1268 dcmds = ms->dma_cmds;
1269 dtot = 0;
1270 if (cmd) {
1271 int nseg;
1273 cmd->SCp.this_residual = scsi_bufflen(cmd);
1275 nseg = scsi_dma_map(cmd);
1276 BUG_ON(nseg < 0);
1278 if (nseg) {
1279 total = 0;
1280 off = ms->data_ptr;
1282 scsi_for_each_sg(cmd, scl, nseg, i) {
1283 u32 dma_addr = sg_dma_address(scl);
1284 u32 dma_len = sg_dma_len(scl);
1286 total += scl->length;
1287 if (off >= dma_len) {
1288 off -= dma_len;
1289 continue;
1291 if (dma_len > 0xffff)
1292 panic("mesh: scatterlist element >= 64k");
1293 st_le16(&dcmds->req_count, dma_len - off);
1294 st_le16(&dcmds->command, dma_cmd);
1295 st_le32(&dcmds->phy_addr, dma_addr + off);
1296 dcmds->xfer_status = 0;
1297 ++dcmds;
1298 dtot += dma_len - off;
1299 off = 0;
1303 if (dtot == 0) {
1304 /* Either the target has overrun our buffer,
1305 or the caller didn't provide a buffer. */
1306 static char mesh_extra_buf[64];
1308 dtot = sizeof(mesh_extra_buf);
1309 st_le16(&dcmds->req_count, dtot);
1310 st_le32(&dcmds->phy_addr, virt_to_phys(mesh_extra_buf));
1311 dcmds->xfer_status = 0;
1312 ++dcmds;
1314 dma_cmd += OUTPUT_LAST - OUTPUT_MORE;
1315 st_le16(&dcmds[-1].command, dma_cmd);
1316 memset(dcmds, 0, sizeof(*dcmds));
1317 st_le16(&dcmds->command, DBDMA_STOP);
1318 ms->dma_count = dtot;
1321 static void halt_dma(struct mesh_state *ms)
1323 volatile struct dbdma_regs __iomem *md = ms->dma;
1324 volatile struct mesh_regs __iomem *mr = ms->mesh;
1325 struct scsi_cmnd *cmd = ms->current_req;
1326 int t, nb;
1328 if (!ms->tgts[ms->conn_tgt].data_goes_out) {
1329 /* wait a little while until the fifo drains */
1330 t = 50;
1331 while (t > 0 && in_8(&mr->fifo_count) != 0
1332 && (in_le32(&md->status) & ACTIVE) != 0) {
1333 --t;
1334 udelay(1);
1337 out_le32(&md->control, RUN << 16); /* turn off RUN bit */
1338 nb = (mr->count_hi << 8) + mr->count_lo;
1339 dlog(ms, "halt_dma fc/count=%.6x",
1340 MKWORD(0, mr->fifo_count, 0, nb));
1341 if (ms->tgts[ms->conn_tgt].data_goes_out)
1342 nb += mr->fifo_count;
1343 /* nb is the number of bytes not yet transferred
1344 to/from the target. */
1345 ms->data_ptr -= nb;
1346 dlog(ms, "data_ptr %x", ms->data_ptr);
1347 if (ms->data_ptr < 0) {
1348 printk(KERN_ERR "mesh: halt_dma: data_ptr=%d (nb=%d, ms=%p)\n",
1349 ms->data_ptr, nb, ms);
1350 ms->data_ptr = 0;
1351 #ifdef MESH_DBG
1352 dumplog(ms, ms->conn_tgt);
1353 dumpslog(ms);
1354 #endif /* MESH_DBG */
1355 } else if (cmd && scsi_bufflen(cmd) &&
1356 ms->data_ptr > scsi_bufflen(cmd)) {
1357 printk(KERN_DEBUG "mesh: target %d overrun, "
1358 "data_ptr=%x total=%x goes_out=%d\n",
1359 ms->conn_tgt, ms->data_ptr, scsi_bufflen(cmd),
1360 ms->tgts[ms->conn_tgt].data_goes_out);
1362 scsi_dma_unmap(cmd);
1363 ms->dma_started = 0;
1366 static void phase_mismatch(struct mesh_state *ms)
1368 volatile struct mesh_regs __iomem *mr = ms->mesh;
1369 int phase;
1371 dlog(ms, "phasemm ch/cl/seq/fc=%.8x",
1372 MKWORD(mr->count_hi, mr->count_lo, mr->sequence, mr->fifo_count));
1373 phase = in_8(&mr->bus_status0) & BS0_PHASE;
1374 if (ms->msgphase == msg_out_xxx && phase == BP_MSGOUT) {
1375 /* output the last byte of the message, without ATN */
1376 out_8(&mr->count_lo, 1);
1377 out_8(&mr->sequence, SEQ_MSGOUT + use_active_neg);
1378 mesh_flush_io(mr);
1379 udelay(1);
1380 out_8(&mr->fifo, ms->msgout[ms->n_msgout-1]);
1381 ms->msgphase = msg_out_last;
1382 return;
1385 if (ms->msgphase == msg_in) {
1386 get_msgin(ms);
1387 if (ms->n_msgin)
1388 handle_msgin(ms);
1391 if (ms->dma_started)
1392 halt_dma(ms);
1393 if (mr->fifo_count) {
1394 out_8(&mr->sequence, SEQ_FLUSHFIFO);
1395 mesh_flush_io(mr);
1396 udelay(1);
1399 ms->msgphase = msg_none;
1400 switch (phase) {
1401 case BP_DATAIN:
1402 ms->tgts[ms->conn_tgt].data_goes_out = 0;
1403 ms->phase = dataing;
1404 break;
1405 case BP_DATAOUT:
1406 ms->tgts[ms->conn_tgt].data_goes_out = 1;
1407 ms->phase = dataing;
1408 break;
1409 case BP_COMMAND:
1410 ms->phase = commanding;
1411 break;
1412 case BP_STATUS:
1413 ms->phase = statusing;
1414 break;
1415 case BP_MSGIN:
1416 ms->msgphase = msg_in;
1417 ms->n_msgin = 0;
1418 break;
1419 case BP_MSGOUT:
1420 ms->msgphase = msg_out;
1421 if (ms->n_msgout == 0) {
1422 if (ms->aborting) {
1423 do_abort(ms);
1424 } else {
1425 if (ms->last_n_msgout == 0) {
1426 printk(KERN_DEBUG
1427 "mesh: no msg to repeat\n");
1428 ms->msgout[0] = NOP;
1429 ms->last_n_msgout = 1;
1431 ms->n_msgout = ms->last_n_msgout;
1434 break;
1435 default:
1436 printk(KERN_DEBUG "mesh: unknown scsi phase %x\n", phase);
1437 ms->stat = DID_ERROR;
1438 mesh_done(ms, 1);
1439 return;
1442 start_phase(ms);
1445 static void cmd_complete(struct mesh_state *ms)
1447 volatile struct mesh_regs __iomem *mr = ms->mesh;
1448 struct scsi_cmnd *cmd = ms->current_req;
1449 struct mesh_target *tp = &ms->tgts[ms->conn_tgt];
1450 int seq, n, t;
1452 dlog(ms, "cmd_complete fc=%x", mr->fifo_count);
1453 seq = use_active_neg + (ms->n_msgout? SEQ_ATN: 0);
1454 switch (ms->msgphase) {
1455 case msg_out_xxx:
1456 /* huh? we expected a phase mismatch */
1457 ms->n_msgin = 0;
1458 ms->msgphase = msg_in;
1459 /* fall through */
1461 case msg_in:
1462 /* should have some message bytes in fifo */
1463 get_msgin(ms);
1464 n = msgin_length(ms);
1465 if (ms->n_msgin < n) {
1466 out_8(&mr->count_lo, n - ms->n_msgin);
1467 out_8(&mr->sequence, SEQ_MSGIN + seq);
1468 } else {
1469 ms->msgphase = msg_none;
1470 handle_msgin(ms);
1471 start_phase(ms);
1473 break;
1475 case msg_in_bad:
1476 out_8(&mr->sequence, SEQ_FLUSHFIFO);
1477 mesh_flush_io(mr);
1478 udelay(1);
1479 out_8(&mr->count_lo, 1);
1480 out_8(&mr->sequence, SEQ_MSGIN + SEQ_ATN + use_active_neg);
1481 break;
1483 case msg_out:
1485 * To get the right timing on ATN wrt ACK, we have
1486 * to get the MESH to drop ACK, wait until REQ gets
1487 * asserted, then drop ATN. To do this we first
1488 * issue a SEQ_MSGOUT with ATN and wait for REQ,
1489 * then change the command to a SEQ_MSGOUT w/o ATN.
1490 * If we don't see REQ in a reasonable time, we
1491 * change the command to SEQ_MSGIN with ATN,
1492 * wait for the phase mismatch interrupt, then
1493 * issue the SEQ_MSGOUT without ATN.
1495 out_8(&mr->count_lo, 1);
1496 out_8(&mr->sequence, SEQ_MSGOUT + use_active_neg + SEQ_ATN);
1497 t = 30; /* wait up to 30us */
1498 while ((in_8(&mr->bus_status0) & BS0_REQ) == 0 && --t >= 0)
1499 udelay(1);
1500 dlog(ms, "last_mbyte err/exc/fc/cl=%.8x",
1501 MKWORD(mr->error, mr->exception,
1502 mr->fifo_count, mr->count_lo));
1503 if (in_8(&mr->interrupt) & (INT_ERROR | INT_EXCEPTION)) {
1504 /* whoops, target didn't do what we expected */
1505 ms->last_n_msgout = ms->n_msgout;
1506 ms->n_msgout = 0;
1507 if (in_8(&mr->interrupt) & INT_ERROR) {
1508 printk(KERN_ERR "mesh: error %x in msg_out\n",
1509 in_8(&mr->error));
1510 handle_error(ms);
1511 return;
1513 if (in_8(&mr->exception) != EXC_PHASEMM)
1514 printk(KERN_ERR "mesh: exc %x in msg_out\n",
1515 in_8(&mr->exception));
1516 else
1517 printk(KERN_DEBUG "mesh: bs0=%x in msg_out\n",
1518 in_8(&mr->bus_status0));
1519 handle_exception(ms);
1520 return;
1522 if (in_8(&mr->bus_status0) & BS0_REQ) {
1523 out_8(&mr->sequence, SEQ_MSGOUT + use_active_neg);
1524 mesh_flush_io(mr);
1525 udelay(1);
1526 out_8(&mr->fifo, ms->msgout[ms->n_msgout-1]);
1527 ms->msgphase = msg_out_last;
1528 } else {
1529 out_8(&mr->sequence, SEQ_MSGIN + use_active_neg + SEQ_ATN);
1530 ms->msgphase = msg_out_xxx;
1532 break;
1534 case msg_out_last:
1535 ms->last_n_msgout = ms->n_msgout;
1536 ms->n_msgout = 0;
1537 ms->msgphase = ms->expect_reply? msg_in: msg_none;
1538 start_phase(ms);
1539 break;
1541 case msg_none:
1542 switch (ms->phase) {
1543 case idle:
1544 printk(KERN_ERR "mesh: interrupt in idle phase?\n");
1545 dumpslog(ms);
1546 return;
1547 case selecting:
1548 dlog(ms, "Selecting phase at command completion",0);
1549 ms->msgout[0] = IDENTIFY(ALLOW_RESEL(ms->conn_tgt),
1550 (cmd? cmd->device->lun: 0));
1551 ms->n_msgout = 1;
1552 ms->expect_reply = 0;
1553 if (ms->aborting) {
1554 ms->msgout[0] = ABORT;
1555 ms->n_msgout++;
1556 } else if (tp->sdtr_state == do_sdtr) {
1557 /* add SDTR message */
1558 add_sdtr_msg(ms);
1559 ms->expect_reply = 1;
1560 tp->sdtr_state = sdtr_sent;
1562 ms->msgphase = msg_out;
1564 * We need to wait for REQ before dropping ATN.
1565 * We wait for at most 30us, then fall back to
1566 * a scheme where we issue a SEQ_COMMAND with ATN,
1567 * which will give us a phase mismatch interrupt
1568 * when REQ does come, and then we send the message.
1570 t = 230; /* wait up to 230us */
1571 while ((in_8(&mr->bus_status0) & BS0_REQ) == 0) {
1572 if (--t < 0) {
1573 dlog(ms, "impatient for req", ms->n_msgout);
1574 ms->msgphase = msg_none;
1575 break;
1577 udelay(1);
1579 break;
1580 case dataing:
1581 if (ms->dma_count != 0) {
1582 start_phase(ms);
1583 return;
1586 * We can get a phase mismatch here if the target
1587 * changes to the status phase, even though we have
1588 * had a command complete interrupt. Then, if we
1589 * issue the SEQ_STATUS command, we'll get a sequence
1590 * error interrupt. Which isn't so bad except that
1591 * occasionally the mesh actually executes the
1592 * SEQ_STATUS *as well as* giving us the sequence
1593 * error and phase mismatch exception.
1595 out_8(&mr->sequence, 0);
1596 out_8(&mr->interrupt,
1597 INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
1598 halt_dma(ms);
1599 break;
1600 case statusing:
1601 if (cmd) {
1602 cmd->SCp.Status = mr->fifo;
1603 if (DEBUG_TARGET(cmd))
1604 printk(KERN_DEBUG "mesh: status is %x\n",
1605 cmd->SCp.Status);
1607 ms->msgphase = msg_in;
1608 break;
1609 case busfreeing:
1610 mesh_done(ms, 1);
1611 return;
1612 case disconnecting:
1613 ms->current_req = NULL;
1614 ms->phase = idle;
1615 mesh_start(ms);
1616 return;
1617 default:
1618 break;
1620 ++ms->phase;
1621 start_phase(ms);
1622 break;
1628 * Called by midlayer with host locked to queue a new
1629 * request
1631 static int mesh_queue(struct scsi_cmnd *cmd, void (*done)(struct scsi_cmnd *))
1633 struct mesh_state *ms;
1635 cmd->scsi_done = done;
1636 cmd->host_scribble = NULL;
1638 ms = (struct mesh_state *) cmd->device->host->hostdata;
1640 if (ms->request_q == NULL)
1641 ms->request_q = cmd;
1642 else
1643 ms->request_qtail->host_scribble = (void *) cmd;
1644 ms->request_qtail = cmd;
1646 if (ms->phase == idle)
1647 mesh_start(ms);
1649 return 0;
1653 * Called to handle interrupts, either call by the interrupt
1654 * handler (do_mesh_interrupt) or by other functions in
1655 * exceptional circumstances
1657 static void mesh_interrupt(struct mesh_state *ms)
1659 volatile struct mesh_regs __iomem *mr = ms->mesh;
1660 int intr;
1662 #if 0
1663 if (ALLOW_DEBUG(ms->conn_tgt))
1664 printk(KERN_DEBUG "mesh_intr, bs0=%x int=%x exc=%x err=%x "
1665 "phase=%d msgphase=%d\n", mr->bus_status0,
1666 mr->interrupt, mr->exception, mr->error,
1667 ms->phase, ms->msgphase);
1668 #endif
1669 while ((intr = in_8(&mr->interrupt)) != 0) {
1670 dlog(ms, "interrupt intr/err/exc/seq=%.8x",
1671 MKWORD(intr, mr->error, mr->exception, mr->sequence));
1672 if (intr & INT_ERROR) {
1673 handle_error(ms);
1674 } else if (intr & INT_EXCEPTION) {
1675 handle_exception(ms);
1676 } else if (intr & INT_CMDDONE) {
1677 out_8(&mr->interrupt, INT_CMDDONE);
1678 cmd_complete(ms);
1683 /* Todo: here we can at least try to remove the command from the
1684 * queue if it isn't connected yet, and for pending command, assert
1685 * ATN until the bus gets freed.
1687 static int mesh_abort(struct scsi_cmnd *cmd)
1689 struct mesh_state *ms = (struct mesh_state *) cmd->device->host->hostdata;
1691 printk(KERN_DEBUG "mesh_abort(%p)\n", cmd);
1692 mesh_dump_regs(ms);
1693 dumplog(ms, cmd->device->id);
1694 dumpslog(ms);
1695 return FAILED;
1699 * Called by the midlayer with the lock held to reset the
1700 * SCSI host and bus.
1701 * The midlayer will wait for devices to come back, we don't need
1702 * to do that ourselves
1704 static int mesh_host_reset(struct scsi_cmnd *cmd)
1706 struct mesh_state *ms = (struct mesh_state *) cmd->device->host->hostdata;
1707 volatile struct mesh_regs __iomem *mr = ms->mesh;
1708 volatile struct dbdma_regs __iomem *md = ms->dma;
1709 unsigned long flags;
1711 printk(KERN_DEBUG "mesh_host_reset\n");
1713 spin_lock_irqsave(ms->host->host_lock, flags);
1715 /* Reset the controller & dbdma channel */
1716 out_le32(&md->control, (RUN|PAUSE|FLUSH|WAKE) << 16); /* stop dma */
1717 out_8(&mr->exception, 0xff); /* clear all exception bits */
1718 out_8(&mr->error, 0xff); /* clear all error bits */
1719 out_8(&mr->sequence, SEQ_RESETMESH);
1720 mesh_flush_io(mr);
1721 udelay(1);
1722 out_8(&mr->intr_mask, INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
1723 out_8(&mr->source_id, ms->host->this_id);
1724 out_8(&mr->sel_timeout, 25); /* 250ms */
1725 out_8(&mr->sync_params, ASYNC_PARAMS);
1727 /* Reset the bus */
1728 out_8(&mr->bus_status1, BS1_RST); /* assert RST */
1729 mesh_flush_io(mr);
1730 udelay(30); /* leave it on for >= 25us */
1731 out_8(&mr->bus_status1, 0); /* negate RST */
1733 /* Complete pending commands */
1734 handle_reset(ms);
1736 spin_unlock_irqrestore(ms->host->host_lock, flags);
1737 return SUCCESS;
1740 static void set_mesh_power(struct mesh_state *ms, int state)
1742 if (!machine_is(powermac))
1743 return;
1744 if (state) {
1745 pmac_call_feature(PMAC_FTR_MESH_ENABLE, macio_get_of_node(ms->mdev), 0, 1);
1746 msleep(200);
1747 } else {
1748 pmac_call_feature(PMAC_FTR_MESH_ENABLE, macio_get_of_node(ms->mdev), 0, 0);
1749 msleep(10);
1754 #ifdef CONFIG_PM
1755 static int mesh_suspend(struct macio_dev *mdev, pm_message_t mesg)
1757 struct mesh_state *ms = (struct mesh_state *)macio_get_drvdata(mdev);
1758 unsigned long flags;
1760 switch (mesg.event) {
1761 case PM_EVENT_SUSPEND:
1762 case PM_EVENT_HIBERNATE:
1763 case PM_EVENT_FREEZE:
1764 break;
1765 default:
1766 return 0;
1768 if (mesg.event == mdev->ofdev.dev.power.power_state.event)
1769 return 0;
1771 scsi_block_requests(ms->host);
1772 spin_lock_irqsave(ms->host->host_lock, flags);
1773 while(ms->phase != idle) {
1774 spin_unlock_irqrestore(ms->host->host_lock, flags);
1775 msleep(10);
1776 spin_lock_irqsave(ms->host->host_lock, flags);
1778 ms->phase = sleeping;
1779 spin_unlock_irqrestore(ms->host->host_lock, flags);
1780 disable_irq(ms->meshintr);
1781 set_mesh_power(ms, 0);
1783 mdev->ofdev.dev.power.power_state = mesg;
1785 return 0;
1788 static int mesh_resume(struct macio_dev *mdev)
1790 struct mesh_state *ms = (struct mesh_state *)macio_get_drvdata(mdev);
1791 unsigned long flags;
1793 if (mdev->ofdev.dev.power.power_state.event == PM_EVENT_ON)
1794 return 0;
1796 set_mesh_power(ms, 1);
1797 mesh_init(ms);
1798 spin_lock_irqsave(ms->host->host_lock, flags);
1799 mesh_start(ms);
1800 spin_unlock_irqrestore(ms->host->host_lock, flags);
1801 enable_irq(ms->meshintr);
1802 scsi_unblock_requests(ms->host);
1804 mdev->ofdev.dev.power.power_state.event = PM_EVENT_ON;
1806 return 0;
1809 #endif /* CONFIG_PM */
1812 * If we leave drives set for synchronous transfers (especially
1813 * CDROMs), and reboot to MacOS, it gets confused, poor thing.
1814 * So, on reboot we reset the SCSI bus.
1816 static int mesh_shutdown(struct macio_dev *mdev)
1818 struct mesh_state *ms = (struct mesh_state *)macio_get_drvdata(mdev);
1819 volatile struct mesh_regs __iomem *mr;
1820 unsigned long flags;
1822 printk(KERN_INFO "resetting MESH scsi bus(es)\n");
1823 spin_lock_irqsave(ms->host->host_lock, flags);
1824 mr = ms->mesh;
1825 out_8(&mr->intr_mask, 0);
1826 out_8(&mr->interrupt, INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
1827 out_8(&mr->bus_status1, BS1_RST);
1828 mesh_flush_io(mr);
1829 udelay(30);
1830 out_8(&mr->bus_status1, 0);
1831 spin_unlock_irqrestore(ms->host->host_lock, flags);
1833 return 0;
1836 static struct scsi_host_template mesh_template = {
1837 .proc_name = "mesh",
1838 .name = "MESH",
1839 .queuecommand = mesh_queue,
1840 .eh_abort_handler = mesh_abort,
1841 .eh_host_reset_handler = mesh_host_reset,
1842 .can_queue = 20,
1843 .this_id = 7,
1844 .sg_tablesize = SG_ALL,
1845 .cmd_per_lun = 2,
1846 .use_clustering = DISABLE_CLUSTERING,
1849 static int mesh_probe(struct macio_dev *mdev, const struct of_device_id *match)
1851 struct device_node *mesh = macio_get_of_node(mdev);
1852 struct pci_dev* pdev = macio_get_pci_dev(mdev);
1853 int tgt, minper;
1854 const int *cfp;
1855 struct mesh_state *ms;
1856 struct Scsi_Host *mesh_host;
1857 void *dma_cmd_space;
1858 dma_addr_t dma_cmd_bus;
1860 switch (mdev->bus->chip->type) {
1861 case macio_heathrow:
1862 case macio_gatwick:
1863 case macio_paddington:
1864 use_active_neg = 0;
1865 break;
1866 default:
1867 use_active_neg = SEQ_ACTIVE_NEG;
1870 if (macio_resource_count(mdev) != 2 || macio_irq_count(mdev) != 2) {
1871 printk(KERN_ERR "mesh: expected 2 addrs and 2 intrs"
1872 " (got %d,%d)\n", macio_resource_count(mdev),
1873 macio_irq_count(mdev));
1874 return -ENODEV;
1877 if (macio_request_resources(mdev, "mesh") != 0) {
1878 printk(KERN_ERR "mesh: unable to request memory resources");
1879 return -EBUSY;
1881 mesh_host = scsi_host_alloc(&mesh_template, sizeof(struct mesh_state));
1882 if (mesh_host == NULL) {
1883 printk(KERN_ERR "mesh: couldn't register host");
1884 goto out_release;
1887 /* Old junk for root discovery, that will die ultimately */
1888 #if !defined(MODULE)
1889 note_scsi_host(mesh, mesh_host);
1890 #endif
1892 mesh_host->base = macio_resource_start(mdev, 0);
1893 mesh_host->irq = macio_irq(mdev, 0);
1894 ms = (struct mesh_state *) mesh_host->hostdata;
1895 macio_set_drvdata(mdev, ms);
1896 ms->host = mesh_host;
1897 ms->mdev = mdev;
1898 ms->pdev = pdev;
1900 ms->mesh = ioremap(macio_resource_start(mdev, 0), 0x1000);
1901 if (ms->mesh == NULL) {
1902 printk(KERN_ERR "mesh: can't map registers\n");
1903 goto out_free;
1905 ms->dma = ioremap(macio_resource_start(mdev, 1), 0x1000);
1906 if (ms->dma == NULL) {
1907 printk(KERN_ERR "mesh: can't map registers\n");
1908 iounmap(ms->mesh);
1909 goto out_free;
1912 ms->meshintr = macio_irq(mdev, 0);
1913 ms->dmaintr = macio_irq(mdev, 1);
1915 /* Space for dma command list: +1 for stop command,
1916 * +1 to allow for aligning.
1918 ms->dma_cmd_size = (mesh_host->sg_tablesize + 2) * sizeof(struct dbdma_cmd);
1920 /* We use the PCI APIs for now until the generic one gets fixed
1921 * enough or until we get some macio-specific versions
1923 dma_cmd_space = pci_alloc_consistent(macio_get_pci_dev(mdev),
1924 ms->dma_cmd_size,
1925 &dma_cmd_bus);
1926 if (dma_cmd_space == NULL) {
1927 printk(KERN_ERR "mesh: can't allocate DMA table\n");
1928 goto out_unmap;
1930 memset(dma_cmd_space, 0, ms->dma_cmd_size);
1932 ms->dma_cmds = (struct dbdma_cmd *) DBDMA_ALIGN(dma_cmd_space);
1933 ms->dma_cmd_space = dma_cmd_space;
1934 ms->dma_cmd_bus = dma_cmd_bus + ((unsigned long)ms->dma_cmds)
1935 - (unsigned long)dma_cmd_space;
1936 ms->current_req = NULL;
1937 for (tgt = 0; tgt < 8; ++tgt) {
1938 ms->tgts[tgt].sdtr_state = do_sdtr;
1939 ms->tgts[tgt].sync_params = ASYNC_PARAMS;
1940 ms->tgts[tgt].current_req = NULL;
1943 if ((cfp = of_get_property(mesh, "clock-frequency", NULL)))
1944 ms->clk_freq = *cfp;
1945 else {
1946 printk(KERN_INFO "mesh: assuming 50MHz clock frequency\n");
1947 ms->clk_freq = 50000000;
1950 /* The maximum sync rate is clock / 5; increase
1951 * mesh_sync_period if necessary.
1953 minper = 1000000000 / (ms->clk_freq / 5); /* ns */
1954 if (mesh_sync_period < minper)
1955 mesh_sync_period = minper;
1957 /* Power up the chip */
1958 set_mesh_power(ms, 1);
1960 /* Set it up */
1961 mesh_init(ms);
1963 /* Request interrupt */
1964 if (request_irq(ms->meshintr, do_mesh_interrupt, 0, "MESH", ms)) {
1965 printk(KERN_ERR "MESH: can't get irq %d\n", ms->meshintr);
1966 goto out_shutdown;
1969 /* Add scsi host & scan */
1970 if (scsi_add_host(mesh_host, &mdev->ofdev.dev))
1971 goto out_release_irq;
1972 scsi_scan_host(mesh_host);
1974 return 0;
1976 out_release_irq:
1977 free_irq(ms->meshintr, ms);
1978 out_shutdown:
1979 /* shutdown & reset bus in case of error or macos can be confused
1980 * at reboot if the bus was set to synchronous mode already
1982 mesh_shutdown(mdev);
1983 set_mesh_power(ms, 0);
1984 pci_free_consistent(macio_get_pci_dev(mdev), ms->dma_cmd_size,
1985 ms->dma_cmd_space, ms->dma_cmd_bus);
1986 out_unmap:
1987 iounmap(ms->dma);
1988 iounmap(ms->mesh);
1989 out_free:
1990 scsi_host_put(mesh_host);
1991 out_release:
1992 macio_release_resources(mdev);
1994 return -ENODEV;
1997 static int mesh_remove(struct macio_dev *mdev)
1999 struct mesh_state *ms = (struct mesh_state *)macio_get_drvdata(mdev);
2000 struct Scsi_Host *mesh_host = ms->host;
2002 scsi_remove_host(mesh_host);
2004 free_irq(ms->meshintr, ms);
2006 /* Reset scsi bus */
2007 mesh_shutdown(mdev);
2009 /* Shut down chip & termination */
2010 set_mesh_power(ms, 0);
2012 /* Unmap registers & dma controller */
2013 iounmap(ms->mesh);
2014 iounmap(ms->dma);
2016 /* Free DMA commands memory */
2017 pci_free_consistent(macio_get_pci_dev(mdev), ms->dma_cmd_size,
2018 ms->dma_cmd_space, ms->dma_cmd_bus);
2020 /* Release memory resources */
2021 macio_release_resources(mdev);
2023 scsi_host_put(mesh_host);
2025 return 0;
2029 static struct of_device_id mesh_match[] =
2032 .name = "mesh",
2035 .type = "scsi",
2036 .compatible = "chrp,mesh0"
2040 MODULE_DEVICE_TABLE (of, mesh_match);
2042 static struct macio_driver mesh_driver =
2044 .name = "mesh",
2045 .match_table = mesh_match,
2046 .probe = mesh_probe,
2047 .remove = mesh_remove,
2048 .shutdown = mesh_shutdown,
2049 #ifdef CONFIG_PM
2050 .suspend = mesh_suspend,
2051 .resume = mesh_resume,
2052 #endif
2056 static int __init init_mesh(void)
2059 /* Calculate sync rate from module parameters */
2060 if (sync_rate > 10)
2061 sync_rate = 10;
2062 if (sync_rate > 0) {
2063 printk(KERN_INFO "mesh: configured for synchronous %d MB/s\n", sync_rate);
2064 mesh_sync_period = 1000 / sync_rate; /* ns */
2065 mesh_sync_offset = 15;
2066 } else
2067 printk(KERN_INFO "mesh: configured for asynchronous\n");
2069 return macio_register_driver(&mesh_driver);
2072 static void __exit exit_mesh(void)
2074 return macio_unregister_driver(&mesh_driver);
2077 module_init(init_mesh);
2078 module_exit(exit_mesh);