2 * Copyright (c) 1998 - 2006 Søren Schmidt <sos@FreeBSD.org>
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
8 * 1. Redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer,
10 * without modification, immediately at the beginning of the file.
11 * 2. Redistributions in binary form must reproduce the above copyright
12 * notice, this list of conditions and the following disclaimer in the
13 * documentation and/or other materials provided with the distribution.
15 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
16 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
17 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
18 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
19 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
20 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
21 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
22 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
23 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
24 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
26 * $FreeBSD: src/sys/dev/ata/ata-queue.c,v 1.67 2007/01/27 21:15:58 remko Exp $
31 #include <sys/param.h>
33 #include <sys/callout.h>
35 #include <sys/queue.h>
38 #include <sys/systm.h>
39 #include <sys/taskqueue.h>
45 static void ata_completed(void *, int);
46 static void ata_sort_queue(struct ata_channel
*ch
, struct ata_request
*request
);
47 static void atawritereorder(struct ata_channel
*ch
);
48 static char *ata_skey2str(u_int8_t
);
51 ata_queue_init(struct ata_channel
*ch
)
53 TAILQ_INIT(&ch
->ata_queue
);
55 ch
->transition
= NULL
;
59 * Rudely drop all requests queued to the channel of specified device.
60 * XXX: The requests are leaked, use only in fatal case.
63 ata_drop_requests(device_t dev
)
65 struct ata_channel
*ch
= device_get_softc(device_get_parent(dev
));
66 struct ata_request
*request
, *tmp
;
68 lockmgr(&ch
->queue_mtx
, LK_EXCLUSIVE
);
69 TAILQ_FOREACH_MUTABLE(request
, &ch
->ata_queue
, chain
, tmp
) {
70 TAILQ_REMOVE(&ch
->ata_queue
, request
, chain
);
71 request
->result
= ENXIO
;
73 lockmgr(&ch
->queue_mtx
, LK_RELEASE
);
77 ata_queue_request(struct ata_request
*request
)
79 struct ata_channel
*ch
;
81 /* treat request as virgin (this might be an ATA_R_REQUEUE) */
82 request
->result
= request
->status
= request
->error
= 0;
84 /* check that that the device is still valid */
85 if (!(request
->parent
= device_get_parent(request
->dev
))) {
86 request
->result
= ENXIO
;
87 if (request
->callback
)
88 (request
->callback
)(request
);
91 ch
= device_get_softc(request
->parent
);
92 /* serialization done via state_mtx */
93 callout_init_lk(&request
->callout
, &ch
->state_mtx
);
94 if (!request
->callback
&& !(request
->flags
& ATA_R_REQUEUE
))
95 lockinit(&request
->done
, "ataqueuerqdone", 0, 0);
97 /* in ATA_STALL_QUEUE state we call HW directly */
98 if ((ch
->state
& ATA_STALL_QUEUE
) && (request
->flags
& ATA_R_CONTROL
)) {
99 lockmgr(&ch
->state_mtx
, LK_EXCLUSIVE
);
100 ch
->running
= request
;
101 if (ch
->hw
.begin_transaction(request
) == ATA_OP_FINISHED
) {
103 if (!request
->callback
)
104 lockuninit(&request
->done
);
105 lockmgr(&ch
->state_mtx
, LK_RELEASE
);
108 /* interlock against interrupt */
109 request
->flags
|= ATA_R_HWCMDQUEUED
;
110 lockmgr(&ch
->state_mtx
, LK_RELEASE
);
112 /* otherwise put request on the locked queue at the specified location */
114 lockmgr(&ch
->queue_mtx
, LK_EXCLUSIVE
);
115 if (request
->flags
& ATA_R_AT_HEAD
) {
116 TAILQ_INSERT_HEAD(&ch
->ata_queue
, request
, chain
);
117 } else if (request
->flags
& ATA_R_ORDERED
) {
118 ata_sort_queue(ch
, request
);
120 TAILQ_INSERT_TAIL(&ch
->ata_queue
, request
, chain
);
121 ch
->transition
= NULL
;
123 lockmgr(&ch
->queue_mtx
, LK_RELEASE
);
124 ATA_DEBUG_RQ(request
, "queued");
128 /* if this is a requeued request callback/sleep we're done */
129 if (request
->flags
& ATA_R_REQUEUE
)
132 /* if this is not a callback wait until request is completed */
133 if (!request
->callback
) {
134 ATA_DEBUG_RQ(request
, "wait for completion");
136 /* interlock against wakeup */
137 lockmgr(&request
->done
, LK_EXCLUSIVE
);
138 /* check if the request was completed already */
139 if (!(request
->flags
& ATA_R_COMPLETED
))
140 lksleep(request
, &request
->done
, 0, "ATA request completion "
141 "wait", request
->timeout
* hz
* 4);
142 lockmgr(&request
->done
, LK_RELEASE
);
143 /* check if the request was completed while sleeping */
144 if (!(request
->flags
& ATA_R_COMPLETED
)) {
146 device_printf(request
->dev
, "WARNING - %s taskqueue timeout - "
147 "completing request directly\n",
148 ata_cmd2str(request
));
149 request
->flags
|= ATA_R_DANGER1
;
150 ata_completed(request
, 0);
153 lockuninit(&request
->done
);
158 ata_controlcmd(device_t dev
, u_int8_t command
, u_int16_t feature
,
159 u_int64_t lba
, u_int16_t count
)
161 struct ata_request
*request
= ata_alloc_request();
166 request
->u
.ata
.command
= command
;
167 request
->u
.ata
.lba
= lba
;
168 request
->u
.ata
.count
= count
;
169 request
->u
.ata
.feature
= feature
;
170 request
->flags
= ATA_R_CONTROL
;
171 request
->timeout
= ATA_DEFAULT_TIMEOUT
;
172 request
->retries
= 0;
173 ata_queue_request(request
);
174 error
= request
->result
;
175 ata_free_request(request
);
181 ata_atapicmd(device_t dev
, u_int8_t
*ccb
, caddr_t data
,
182 int count
, int flags
, int timeout
)
184 struct ata_request
*request
= ata_alloc_request();
185 struct ata_device
*atadev
= device_get_softc(dev
);
190 if ((atadev
->param
.config
& ATA_PROTO_MASK
) == ATA_PROTO_ATAPI_12
)
191 bcopy(ccb
, request
->u
.atapi
.ccb
, 12);
193 bcopy(ccb
, request
->u
.atapi
.ccb
, 16);
194 request
->data
= data
;
195 request
->bytecount
= count
;
196 request
->transfersize
= min(request
->bytecount
, 65534);
197 request
->flags
= flags
| ATA_R_ATAPI
;
198 request
->timeout
= timeout
;
199 request
->retries
= 0;
200 ata_queue_request(request
);
201 error
= request
->result
;
202 ata_free_request(request
);
208 ata_start(device_t dev
)
210 struct ata_channel
*ch
= device_get_softc(dev
);
211 struct ata_request
*request
;
212 struct ata_composite
*cptr
;
213 int dependencies
= 0;
215 /* if we have a request on the queue try to get it running */
216 lockmgr(&ch
->queue_mtx
, LK_EXCLUSIVE
);
217 if ((request
= TAILQ_FIRST(&ch
->ata_queue
))) {
219 /* we need the locking function to get the lock for this channel */
220 if (ATA_LOCKING(dev
, ATA_LF_LOCK
) == ch
->unit
) {
222 /* check for composite dependencies */
223 if ((cptr
= request
->composite
)) {
224 lockmgr(&cptr
->lock
, LK_EXCLUSIVE
);
225 if ((request
->flags
& ATA_R_WRITE
) &&
226 (cptr
->wr_depend
& cptr
->rd_done
) != cptr
->wr_depend
) {
229 lockmgr(&cptr
->lock
, LK_RELEASE
);
232 /* check we are in the right state and has no dependencies */
233 lockmgr(&ch
->state_mtx
, LK_EXCLUSIVE
);
234 if (ch
->state
== ATA_IDLE
&& !dependencies
) {
235 ATA_DEBUG_RQ(request
, "starting");
237 if (ch
->transition
== request
)
238 ch
->transition
= TAILQ_NEXT(request
, chain
);
239 TAILQ_REMOVE(&ch
->ata_queue
, request
, chain
);
240 ch
->running
= request
;
241 ch
->state
= ATA_ACTIVE
;
243 if (ch
->hw
.begin_transaction(request
) == ATA_OP_FINISHED
) {
245 ch
->state
= ATA_IDLE
;
246 lockmgr(&ch
->state_mtx
, LK_RELEASE
);
247 lockmgr(&ch
->queue_mtx
, LK_RELEASE
);
248 ATA_LOCKING(dev
, ATA_LF_UNLOCK
);
253 /* interlock against interrupt */
254 request
->flags
|= ATA_R_HWCMDQUEUED
;
257 lockmgr(&ch
->state_mtx
, LK_RELEASE
);
258 lockmgr(&ch
->queue_mtx
, LK_RELEASE
);
259 while (!ata_interrupt(ch
))
264 lockmgr(&ch
->state_mtx
, LK_RELEASE
);
267 lockmgr(&ch
->queue_mtx
, LK_RELEASE
);
271 ata_finish(struct ata_request
*request
)
273 struct ata_channel
*ch
= device_get_softc(request
->parent
);
276 * if in ATA_STALL_QUEUE state or request has ATA_R_DIRECT flags set
277 * we need to call ata_complete() directly here (no taskqueue involvement)
280 (ch
->state
& ATA_STALL_QUEUE
) || (request
->flags
& ATA_R_DIRECT
)) {
281 ATA_DEBUG_RQ(request
, "finish directly");
282 ata_completed(request
, 0);
285 /* put request on the proper taskqueue for completion */
286 /* XXX FreeBSD has some sort of bio_taskqueue code here */
287 TASK_INIT(&request
->task
, 0, ata_completed
, request
);
288 ATA_DEBUG_RQ(request
, "finish taskqueue_swi_mp");
289 taskqueue_enqueue(taskqueue_swi_mp
, &request
->task
);
294 ata_completed(void *context
, int dummy
)
296 struct ata_request
*request
= (struct ata_request
*)context
;
297 struct ata_channel
*ch
= device_get_softc(request
->parent
);
298 struct ata_device
*atadev
= device_get_softc(request
->dev
);
299 struct ata_composite
*composite
;
301 if (request
->flags
& ATA_R_DANGER2
) {
302 device_printf(request
->dev
,
303 "WARNING - %s freeing taskqueue zombie request\n",
304 ata_cmd2str(request
));
305 request
->flags
&= ~(ATA_R_DANGER1
| ATA_R_DANGER2
);
306 ata_free_request(request
);
309 if (request
->flags
& ATA_R_DANGER1
)
310 request
->flags
|= ATA_R_DANGER2
;
312 ATA_DEBUG_RQ(request
, "completed entered");
314 /* if we had a timeout, reinit channel and deal with the falldown */
315 if (request
->flags
& ATA_R_TIMEOUT
) {
317 * if the channel is still present and
318 * reinit succeeds and
319 * the device doesn't get detached and
320 * there are retries left we reinject this request
322 if (ch
&& !ata_reinit(ch
->dev
) && !request
->result
&&
323 (request
->retries
-- > 0)) {
324 if (!(request
->flags
& ATA_R_QUIET
)) {
325 device_printf(request
->dev
,
326 "TIMEOUT - %s retrying (%d retr%s left)",
327 ata_cmd2str(request
), request
->retries
,
328 request
->retries
== 1 ? "y" : "ies");
329 if (!(request
->flags
& (ATA_R_ATAPI
| ATA_R_CONTROL
)))
330 kprintf(" LBA=%ju", request
->u
.ata
.lba
);
333 request
->flags
&= ~(ATA_R_TIMEOUT
| ATA_R_DEBUG
);
334 request
->flags
|= (ATA_R_AT_HEAD
| ATA_R_REQUEUE
);
335 ATA_DEBUG_RQ(request
, "completed reinject");
336 ata_queue_request(request
);
340 /* ran out of good intentions so finish with error */
341 if (!request
->result
) {
342 if (!(request
->flags
& ATA_R_QUIET
)) {
344 device_printf(request
->dev
, "FAILURE - %s timed out",
345 ata_cmd2str(request
));
346 if (!(request
->flags
& (ATA_R_ATAPI
| ATA_R_CONTROL
)))
347 kprintf(" LBA=%ju", request
->u
.ata
.lba
);
351 request
->result
= EIO
;
354 else if (!(request
->flags
& ATA_R_ATAPI
) ){
355 /* if this is a soft ECC error warn about it */
356 /* XXX SOS we could do WARF here */
357 if ((request
->status
& (ATA_S_CORR
| ATA_S_ERROR
)) == ATA_S_CORR
) {
358 device_printf(request
->dev
,
359 "WARNING - %s soft error (ECC corrected)",
360 ata_cmd2str(request
));
361 if (!(request
->flags
& (ATA_R_ATAPI
| ATA_R_CONTROL
)))
362 kprintf(" LBA=%ju", request
->u
.ata
.lba
);
366 /* if this is a UDMA CRC error we reinject if there are retries left */
367 if (request
->flags
& ATA_R_DMA
&& request
->error
& ATA_E_ICRC
) {
368 if (request
->retries
-- > 0) {
369 device_printf(request
->dev
,
370 "WARNING - %s UDMA ICRC error (retrying request)",
371 ata_cmd2str(request
));
372 if (!(request
->flags
& (ATA_R_ATAPI
| ATA_R_CONTROL
)))
373 kprintf(" LBA=%ju", request
->u
.ata
.lba
);
375 request
->flags
|= (ATA_R_AT_HEAD
| ATA_R_REQUEUE
);
376 ata_queue_request(request
);
382 switch (request
->flags
& ATA_R_ATAPI
) {
386 if (!request
->result
&& request
->status
& ATA_S_ERROR
) {
387 if (!(request
->flags
& ATA_R_QUIET
)) {
388 device_printf(request
->dev
,
389 "FAILURE - %s status=%b error=%b",
390 ata_cmd2str(request
),
391 request
->status
, "\20\10BUSY\7READY\6DMA_READY"
392 "\5DSC\4DRQ\3CORRECTABLE\2INDEX\1ERROR",
393 request
->error
, "\20\10ICRC\7UNCORRECTABLE"
394 "\6MEDIA_CHANGED\5NID_NOT_FOUND"
395 "\4MEDIA_CHANGE_REQEST"
396 "\3ABORTED\2NO_MEDIA\1ILLEGAL_LENGTH");
397 if ((request
->flags
& ATA_R_DMA
) &&
398 (request
->dmastat
& ATA_BMSTAT_ERROR
))
399 kprintf(" dma=0x%02x", request
->dmastat
);
400 if (!(request
->flags
& (ATA_R_ATAPI
| ATA_R_CONTROL
)))
401 kprintf(" LBA=%ju", request
->u
.ata
.lba
);
404 request
->result
= EIO
;
410 /* skip if result already set */
414 /* if we have a sensekey -> request sense from device */
415 if ((request
->error
& ATA_E_ATAPI_SENSE_MASK
) &&
416 (request
->u
.atapi
.ccb
[0] != ATAPI_REQUEST_SENSE
)) {
417 static u_int8_t ccb
[16] = { ATAPI_REQUEST_SENSE
, 0, 0, 0,
418 sizeof(struct atapi_sense
),
419 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 };
421 request
->u
.atapi
.saved_cmd
= request
->u
.atapi
.ccb
[0];
422 bcopy(ccb
, request
->u
.atapi
.ccb
, 16);
423 request
->data
= (caddr_t
)&request
->u
.atapi
.sense
;
424 request
->bytecount
= sizeof(struct atapi_sense
);
425 request
->donecount
= 0;
426 request
->transfersize
= sizeof(struct atapi_sense
);
427 request
->timeout
= ATA_DEFAULT_TIMEOUT
;
428 request
->flags
&= (ATA_R_ATAPI
| ATA_R_QUIET
);
429 request
->flags
|= (ATA_R_READ
| ATA_R_AT_HEAD
| ATA_R_REQUEUE
);
430 ATA_DEBUG_RQ(request
, "autoissue request sense");
431 ata_queue_request(request
);
435 switch (request
->u
.atapi
.sense
.key
& ATA_SENSE_KEY_MASK
) {
436 case ATA_SENSE_RECOVERED_ERROR
:
437 device_printf(request
->dev
, "WARNING - %s recovered error\n",
438 ata_cmd2str(request
));
441 case ATA_SENSE_NO_SENSE
:
445 case ATA_SENSE_NOT_READY
:
446 request
->result
= EBUSY
;
449 case ATA_SENSE_UNIT_ATTENTION
:
450 atadev
->flags
|= ATA_D_MEDIA_CHANGED
;
451 request
->result
= EIO
;
455 request
->result
= EIO
;
456 if (request
->flags
& ATA_R_QUIET
)
459 device_printf(request
->dev
,
460 "FAILURE - %s %s asc=0x%02x ascq=0x%02x ",
461 ata_cmd2str(request
), ata_skey2str(
462 (request
->u
.atapi
.sense
.key
& ATA_SENSE_KEY_MASK
)),
463 request
->u
.atapi
.sense
.asc
,
464 request
->u
.atapi
.sense
.ascq
);
465 if (request
->u
.atapi
.sense
.specific
& ATA_SENSE_SPEC_VALID
)
466 kprintf("sks=0x%02x 0x%02x 0x%02x\n",
467 request
->u
.atapi
.sense
.specific
& ATA_SENSE_SPEC_MASK
,
468 request
->u
.atapi
.sense
.specific1
,
469 request
->u
.atapi
.sense
.specific2
);
474 if ((request
->u
.atapi
.sense
.key
& ATA_SENSE_KEY_MASK
?
475 request
->u
.atapi
.sense
.key
& ATA_SENSE_KEY_MASK
:
477 request
->result
= EIO
;
480 ATA_DEBUG_RQ(request
, "completed callback/wakeup");
482 /* if we are part of a composite operation we need to maintain progress */
483 if ((composite
= request
->composite
)) {
486 lockmgr(&composite
->lock
, LK_EXCLUSIVE
);
488 /* update whats done */
489 if (request
->flags
& ATA_R_READ
)
490 composite
->rd_done
|= (1 << request
->this);
491 if (request
->flags
& ATA_R_WRITE
)
492 composite
->wr_done
|= (1 << request
->this);
494 /* find ready to go dependencies */
495 if (composite
->wr_depend
&&
496 (composite
->rd_done
& composite
->wr_depend
)==composite
->wr_depend
&&
497 (composite
->wr_needed
& (~composite
->wr_done
))) {
498 index
= composite
->wr_needed
& ~composite
->wr_done
;
501 lockmgr(&composite
->lock
, LK_RELEASE
);
503 /* if we have any ready candidates kick them off */
507 for (bit
= 0; bit
< MAX_COMPOSITES
; bit
++) {
508 if (index
& (1 << bit
))
509 ata_start(device_get_parent(composite
->request
[bit
]->dev
));
514 /* get results back to the initiator for this request */
515 if (request
->callback
)
516 (request
->callback
)(request
);
518 lockmgr(&request
->done
, LK_EXCLUSIVE
);
519 request
->flags
|= ATA_R_COMPLETED
;
520 lockmgr(&request
->done
, LK_RELEASE
);
524 /* only call ata_start if channel is present */
530 ata_timeout(struct ata_request
*request
)
532 struct ata_channel
*ch
= device_get_softc(request
->parent
);
535 * NOTE: callout acquired ch->state_mtx for us.
537 /*request->flags |= ATA_R_DEBUG;*/
538 ATA_DEBUG_RQ(request
, "timeout");
541 * if we have an ATA_ACTIVE request running, we flag the request
542 * ATA_R_TIMEOUT so ata_finish will handle it correctly
543 * also NULL out the running request so we wont loose
544 * the race with an eventual interrupt arriving late
546 if (ch
->state
== ATA_ACTIVE
) {
547 request
->flags
|= ATA_R_TIMEOUT
;
548 lockmgr(&ch
->state_mtx
, LK_RELEASE
);
549 ATA_LOCKING(ch
->dev
, ATA_LF_UNLOCK
);
551 lockmgr(&ch
->state_mtx
, LK_EXCLUSIVE
);
556 ata_fail_requests(device_t dev
)
558 struct ata_channel
*ch
= device_get_softc(device_get_parent(dev
));
559 struct ata_request
*request
, *tmp
;
560 TAILQ_HEAD(, ata_request
) fail_requests
;
561 TAILQ_INIT(&fail_requests
);
563 /* grap all channel locks to avoid races */
564 lockmgr(&ch
->queue_mtx
, LK_EXCLUSIVE
);
565 lockmgr(&ch
->state_mtx
, LK_EXCLUSIVE
);
567 /* do we have any running request to care about ? */
568 if ((request
= ch
->running
) && (!dev
|| request
->dev
== dev
)) {
569 callout_stop_sync(&request
->callout
);
571 request
->result
= ENXIO
;
572 TAILQ_INSERT_TAIL(&fail_requests
, request
, chain
);
575 /* fail all requests queued on this channel for device dev if !NULL */
576 TAILQ_FOREACH_MUTABLE(request
, &ch
->ata_queue
, chain
, tmp
) {
577 if (!dev
|| request
->dev
== dev
) {
578 if (ch
->transition
== request
)
579 ch
->transition
= TAILQ_NEXT(request
, chain
);
580 TAILQ_REMOVE(&ch
->ata_queue
, request
, chain
);
581 request
->result
= ENXIO
;
582 TAILQ_INSERT_TAIL(&fail_requests
, request
, chain
);
586 lockmgr(&ch
->state_mtx
, LK_RELEASE
);
587 lockmgr(&ch
->queue_mtx
, LK_RELEASE
);
589 /* finish up all requests collected above */
590 TAILQ_FOREACH_MUTABLE(request
, &fail_requests
, chain
, tmp
) {
591 TAILQ_REMOVE(&fail_requests
, request
, chain
);
597 ata_get_lba(struct ata_request
*request
)
599 if (request
->flags
& ATA_R_ATAPI
) {
600 switch (request
->u
.atapi
.ccb
[0]) {
602 case ATAPI_WRITE_BIG
:
604 return (request
->u
.atapi
.ccb
[5]) | (request
->u
.atapi
.ccb
[4]<<8) |
605 (request
->u
.atapi
.ccb
[3]<<16)|(request
->u
.atapi
.ccb
[2]<<24);
608 return (request
->u
.atapi
.ccb
[4]) | (request
->u
.atapi
.ccb
[3]<<8) |
609 (request
->u
.atapi
.ccb
[2]<<16);
615 return request
->u
.ata
.lba
;
619 * This implements exactly bioqdisksort() in the DragonFly kernel.
620 * The short description is: Because megabytes and megabytes worth of
621 * writes can be queued there needs to be a read-prioritization mechanism
622 * or reads get completely starved out.
625 ata_sort_queue(struct ata_channel
*ch
, struct ata_request
*request
)
627 if ((request
->flags
& ATA_R_WRITE
) == 0) {
628 if (ch
->transition
) {
630 * Insert before the first write
632 TAILQ_INSERT_BEFORE(ch
->transition
, request
, chain
);
633 if (++ch
->reorder
>= bioq_reorder_minor_interval
) {
639 * No writes queued (or ordering was forced),
642 TAILQ_INSERT_TAIL(&ch
->ata_queue
, request
, chain
);
646 * Writes are always appended. If no writes were previously
647 * queued or an ordered tail insertion occured the transition
648 * field will be NULL.
650 TAILQ_INSERT_TAIL(&ch
->ata_queue
, request
, chain
);
651 if (ch
->transition
== NULL
)
652 ch
->transition
= request
;
654 if (request
->composite
) {
655 ch
->transition
= NULL
;
661 * Move the transition point to prevent reads from completely
662 * starving our writes. This brings a number of writes into
663 * the fold every N reads.
666 atawritereorder(struct ata_channel
*ch
)
668 struct ata_request
*req
;
669 u_int64_t next_offset
;
670 size_t left
= (size_t)bioq_reorder_minor_bytes
;
673 next_offset
= ata_get_lba(ch
->transition
);
674 while ((req
= ch
->transition
) != NULL
&&
675 next_offset
== ata_get_lba(req
)) {
676 n
= req
->u
.ata
.count
;
677 next_offset
= ata_get_lba(req
);
678 ch
->transition
= TAILQ_NEXT(req
, chain
);
686 ata_cmd2str(struct ata_request
*request
)
688 static char buffer
[20];
690 if (request
->flags
& ATA_R_ATAPI
) {
691 switch (request
->u
.atapi
.sense
.key
?
692 request
->u
.atapi
.saved_cmd
: request
->u
.atapi
.ccb
[0]) {
693 case 0x00: return ("TEST_UNIT_READY");
694 case 0x01: return ("REZERO");
695 case 0x03: return ("REQUEST_SENSE");
696 case 0x04: return ("FORMAT");
697 case 0x08: return ("READ");
698 case 0x0a: return ("WRITE");
699 case 0x10: return ("WEOF");
700 case 0x11: return ("SPACE");
701 case 0x12: return ("INQUIRY");
702 case 0x15: return ("MODE_SELECT");
703 case 0x19: return ("ERASE");
704 case 0x1a: return ("MODE_SENSE");
705 case 0x1b: return ("START_STOP");
706 case 0x1e: return ("PREVENT_ALLOW");
707 case 0x23: return ("ATAPI_READ_FORMAT_CAPACITIES");
708 case 0x25: return ("READ_CAPACITY");
709 case 0x28: return ("READ_BIG");
710 case 0x2a: return ("WRITE_BIG");
711 case 0x2b: return ("LOCATE");
712 case 0x34: return ("READ_POSITION");
713 case 0x35: return ("SYNCHRONIZE_CACHE");
714 case 0x3b: return ("WRITE_BUFFER");
715 case 0x3c: return ("READ_BUFFER");
716 case 0x42: return ("READ_SUBCHANNEL");
717 case 0x43: return ("READ_TOC");
718 case 0x45: return ("PLAY_10");
719 case 0x47: return ("PLAY_MSF");
720 case 0x48: return ("PLAY_TRACK");
721 case 0x4b: return ("PAUSE");
722 case 0x51: return ("READ_DISK_INFO");
723 case 0x52: return ("READ_TRACK_INFO");
724 case 0x53: return ("RESERVE_TRACK");
725 case 0x54: return ("SEND_OPC_INFO");
726 case 0x55: return ("MODE_SELECT_BIG");
727 case 0x58: return ("REPAIR_TRACK");
728 case 0x59: return ("READ_MASTER_CUE");
729 case 0x5a: return ("MODE_SENSE_BIG");
730 case 0x5b: return ("CLOSE_TRACK/SESSION");
731 case 0x5c: return ("READ_BUFFER_CAPACITY");
732 case 0x5d: return ("SEND_CUE_SHEET");
733 case 0x96: return ("READ_CAPACITY_16");
734 case 0xa1: return ("BLANK_CMD");
735 case 0xa3: return ("SEND_KEY");
736 case 0xa4: return ("REPORT_KEY");
737 case 0xa5: return ("PLAY_12");
738 case 0xa6: return ("LOAD_UNLOAD");
739 case 0xad: return ("READ_DVD_STRUCTURE");
740 case 0xb4: return ("PLAY_CD");
741 case 0xbb: return ("SET_SPEED");
742 case 0xbd: return ("MECH_STATUS");
743 case 0xbe: return ("READ_CD");
744 case 0xff: return ("POLL_DSC");
748 switch (request
->u
.ata
.command
) {
749 case 0x00: return ("NOP");
750 case 0x08: return ("DEVICE_RESET");
751 case 0x20: return ("READ");
752 case 0x24: return ("READ48");
753 case 0x25: return ("READ_DMA48");
754 case 0x26: return ("READ_DMA_QUEUED48");
755 case 0x29: return ("READ_MUL48");
756 case 0x30: return ("WRITE");
757 case 0x34: return ("WRITE48");
758 case 0x35: return ("WRITE_DMA48");
759 case 0x36: return ("WRITE_DMA_QUEUED48");
760 case 0x39: return ("WRITE_MUL48");
761 case 0x70: return ("SEEK");
762 case 0xa0: return ("PACKET_CMD");
763 case 0xa1: return ("ATAPI_IDENTIFY");
764 case 0xa2: return ("SERVICE");
765 case 0xb0: return ("SMART");
766 case 0xc0: return ("CFA ERASE");
767 case 0xc4: return ("READ_MUL");
768 case 0xc5: return ("WRITE_MUL");
769 case 0xc6: return ("SET_MULTI");
770 case 0xc7: return ("READ_DMA_QUEUED");
771 case 0xc8: return ("READ_DMA");
772 case 0xca: return ("WRITE_DMA");
773 case 0xcc: return ("WRITE_DMA_QUEUED");
774 case 0xe6: return ("SLEEP");
775 case 0xe7: return ("FLUSHCACHE");
776 case 0xea: return ("FLUSHCACHE48");
777 case 0xec: return ("ATA_IDENTIFY");
779 switch (request
->u
.ata
.feature
) {
780 case 0x03: return ("SETFEATURES SET TRANSFER MODE");
781 case 0x02: return ("SETFEATURES ENABLE WCACHE");
782 case 0x82: return ("SETFEATURES DISABLE WCACHE");
783 case 0xaa: return ("SETFEATURES ENABLE RCACHE");
784 case 0x55: return ("SETFEATURES DISABLE RCACHE");
786 ksprintf(buffer
, "SETFEATURES 0x%02x", request
->u
.ata
.feature
);
790 ksprintf(buffer
, "unknown CMD (0x%02x)", request
->u
.ata
.command
);
795 ata_skey2str(u_int8_t skey
)
798 case 0x00: return ("NO SENSE");
799 case 0x01: return ("RECOVERED ERROR");
800 case 0x02: return ("NOT READY");
801 case 0x03: return ("MEDIUM ERROR");
802 case 0x04: return ("HARDWARE ERROR");
803 case 0x05: return ("ILLEGAL REQUEST");
804 case 0x06: return ("UNIT ATTENTION");
805 case 0x07: return ("DATA PROTECT");
806 case 0x08: return ("BLANK CHECK");
807 case 0x09: return ("VENDOR SPECIFIC");
808 case 0x0a: return ("COPY ABORTED");
809 case 0x0b: return ("ABORTED COMMAND");
810 case 0x0c: return ("EQUAL");
811 case 0x0d: return ("VOLUME OVERFLOW");
812 case 0x0e: return ("MISCOMPARE");
813 case 0x0f: return ("RESERVED");
814 default: return("UNKNOWN");