2 * QEMU PowerPC pSeries Logical Partition (aka sPAPR) hardware System Emulator
4 * PAPR Virtual SCSI, aka ibmvscsi
6 * Copyright (c) 2010,2011 Benjamin Herrenschmidt, IBM Corporation.
8 * Permission is hereby granted, free of charge, to any person obtaining a copy
9 * of this software and associated documentation files (the "Software"), to deal
10 * in the Software without restriction, including without limitation the rights
11 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
12 * copies of the Software, and to permit persons to whom the Software is
13 * furnished to do so, subject to the following conditions:
15 * The above copyright notice and this permission notice shall be included in
16 * all copies or substantial portions of the Software.
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
19 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
20 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
21 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
22 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
23 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
29 * - Sort out better how to assign devices to VSCSI instances
30 * - Fix residual counts
31 * - Add indirect descriptors support
32 * - Maybe do autosense (PAPR seems to mandate it, linux doesn't care)
35 #include "qemu/osdep.h"
36 #include "qemu/module.h"
38 #include "hw/scsi/scsi.h"
39 #include "migration/vmstate.h"
40 #include "scsi/constants.h"
42 #include "hw/ppc/spapr.h"
43 #include "hw/ppc/spapr_vio.h"
44 #include "hw/qdev-properties.h"
49 #include "qom/object.h"
56 #define VSCSI_MAX_SECTORS 4096
57 #define VSCSI_REQ_LIMIT 24
59 /* Maximum size of a IU payload */
60 #define SRP_MAX_IU_DATA_LEN (SRP_MAX_IU_LEN - sizeof(union srp_iu))
61 #define SRP_RSP_SENSE_DATA_LEN 18
63 #define SRP_REPORT_LUNS_WLUN 0xc10100000000000ULL
65 typedef union vscsi_crq
{
70 typedef struct vscsi_req
{
72 uint8_t viosrp_iu_buf
[SRP_MAX_IU_LEN
];
74 /* SCSI request tracking */
76 uint32_t qtag
; /* qemu tag != srp tag */
82 uint8_t sense
[SCSI_SENSE_BUF_SIZE
];
84 /* RDMA related bits */
89 uint16_t cur_desc_num
;
90 uint16_t cur_desc_offset
;
93 #define TYPE_VIO_SPAPR_VSCSI_DEVICE "spapr-vscsi"
94 OBJECT_DECLARE_SIMPLE_TYPE(VSCSIState
, VIO_SPAPR_VSCSI_DEVICE
)
99 vscsi_req reqs
[VSCSI_REQ_LIMIT
];
102 static union viosrp_iu
*req_iu(vscsi_req
*req
)
104 return (union viosrp_iu
*)req
->viosrp_iu_buf
;
107 static struct vscsi_req
*vscsi_get_req(VSCSIState
*s
)
112 for (i
= 0; i
< VSCSI_REQ_LIMIT
; i
++) {
115 memset(req
, 0, sizeof(*req
));
124 static struct vscsi_req
*vscsi_find_req(VSCSIState
*s
, uint64_t srp_tag
)
129 for (i
= 0; i
< VSCSI_REQ_LIMIT
; i
++) {
131 if (req_iu(req
)->srp
.cmd
.tag
== srp_tag
) {
138 static void vscsi_put_req(vscsi_req
*req
)
140 if (req
->sreq
!= NULL
) {
141 scsi_req_unref(req
->sreq
);
147 static SCSIDevice
*vscsi_device_find(SCSIBus
*bus
, uint64_t srp_lun
, int *lun
)
149 int channel
= 0, id
= 0;
152 switch (srp_lun
>> 62) {
154 if ((srp_lun
>> 56) != 0) {
155 channel
= (srp_lun
>> 56) & 0x3f;
156 id
= (srp_lun
>> 48) & 0xff;
160 *lun
= (srp_lun
>> 48) & 0xff;
164 *lun
= (srp_lun
>> 48) & 0x3fff;
167 channel
= (srp_lun
>> 53) & 0x7;
168 id
= (srp_lun
>> 56) & 0x3f;
169 *lun
= (srp_lun
>> 48) & 0x1f;
178 return scsi_device_find(bus
, channel
, id
, *lun
);
181 static int vscsi_send_iu(VSCSIState
*s
, vscsi_req
*req
,
182 uint64_t length
, uint8_t format
)
186 assert(length
<= SRP_MAX_IU_LEN
);
188 /* First copy the SRP */
189 rc
= spapr_vio_dma_write(&s
->vdev
, req
->crq
.s
.IU_data_ptr
,
190 &req
->viosrp_iu_buf
, length
);
192 fprintf(stderr
, "vscsi_send_iu: DMA write failure !\n");
195 req
->crq
.s
.valid
= 0x80;
196 req
->crq
.s
.format
= format
;
197 req
->crq
.s
.reserved
= 0x00;
198 req
->crq
.s
.timeout
= cpu_to_be16(0x0000);
199 req
->crq
.s
.IU_length
= cpu_to_be16(length
);
200 req
->crq
.s
.IU_data_ptr
= req_iu(req
)->srp
.rsp
.tag
; /* right byte order */
203 req
->crq
.s
.status
= VIOSRP_OK
;
205 req
->crq
.s
.status
= VIOSRP_ADAPTER_FAIL
;
208 rc1
= spapr_vio_send_crq(&s
->vdev
, req
->crq
.raw
);
210 fprintf(stderr
, "vscsi_send_iu: Error sending response\n");
217 static void vscsi_makeup_sense(VSCSIState
*s
, vscsi_req
*req
,
218 uint8_t key
, uint8_t asc
, uint8_t ascq
)
220 req
->senselen
= SRP_RSP_SENSE_DATA_LEN
;
222 /* Valid bit and 'current errors' */
223 req
->sense
[0] = (0x1 << 7 | 0x70);
226 /* Additional sense length */
227 req
->sense
[7] = 0xa; /* 10 bytes */
228 /* Additional sense code */
229 req
->sense
[12] = asc
;
230 req
->sense
[13] = ascq
;
233 static int vscsi_send_rsp(VSCSIState
*s
, vscsi_req
*req
,
234 uint8_t status
, int32_t res_in
, int32_t res_out
)
236 union viosrp_iu
*iu
= req_iu(req
);
237 uint64_t tag
= iu
->srp
.rsp
.tag
;
238 int total_len
= sizeof(iu
->srp
.rsp
);
239 uint8_t sol_not
= iu
->srp
.cmd
.sol_not
;
241 trace_spapr_vscsi_send_rsp(status
, res_in
, res_out
);
243 memset(iu
, 0, sizeof(struct srp_rsp
));
244 iu
->srp
.rsp
.opcode
= SRP_RSP
;
245 iu
->srp
.rsp
.req_lim_delta
= cpu_to_be32(1);
246 iu
->srp
.rsp
.tag
= tag
;
248 /* Handle residuals */
250 iu
->srp
.rsp
.flags
|= SRP_RSP_FLAG_DIUNDER
;
253 iu
->srp
.rsp
.flags
|= SRP_RSP_FLAG_DIOVER
;
256 iu
->srp
.rsp
.flags
|= SRP_RSP_FLAG_DOUNDER
;
258 } else if (res_out
) {
259 iu
->srp
.rsp
.flags
|= SRP_RSP_FLAG_DOOVER
;
261 iu
->srp
.rsp
.data_in_res_cnt
= cpu_to_be32(res_in
);
262 iu
->srp
.rsp
.data_out_res_cnt
= cpu_to_be32(res_out
);
264 /* We don't do response data */
265 /* iu->srp.rsp.flags &= ~SRP_RSP_FLAG_RSPVALID; */
266 iu
->srp
.rsp
.resp_data_len
= cpu_to_be32(0);
268 /* Handle success vs. failure */
269 iu
->srp
.rsp
.status
= status
;
271 iu
->srp
.rsp
.sol_not
= (sol_not
& 0x04) >> 2;
273 int sense_data_len
= MIN(req
->senselen
, SRP_MAX_IU_DATA_LEN
);
275 iu
->srp
.rsp
.flags
|= SRP_RSP_FLAG_SNSVALID
;
276 iu
->srp
.rsp
.sense_data_len
= cpu_to_be32(sense_data_len
);
277 memcpy(iu
->srp
.rsp
.data
, req
->sense
, sense_data_len
);
278 total_len
+= sense_data_len
;
281 iu
->srp
.rsp
.sol_not
= (sol_not
& 0x02) >> 1;
284 vscsi_send_iu(s
, req
, total_len
, VIOSRP_SRP_FORMAT
);
288 static inline struct srp_direct_buf
vscsi_swap_desc(struct srp_direct_buf desc
)
290 desc
.va
= be64_to_cpu(desc
.va
);
291 desc
.len
= be32_to_cpu(desc
.len
);
295 static int vscsi_fetch_desc(VSCSIState
*s
, struct vscsi_req
*req
,
296 unsigned n
, unsigned buf_offset
,
297 struct srp_direct_buf
*ret
)
299 struct srp_cmd
*cmd
= &req_iu(req
)->srp
.cmd
;
301 switch (req
->dma_fmt
) {
302 case SRP_NO_DATA_DESC
: {
303 trace_spapr_vscsi_fetch_desc_no_data();
306 case SRP_DATA_DESC_DIRECT
: {
307 memcpy(ret
, cmd
->add_data
+ req
->cdb_offset
, sizeof(*ret
));
308 assert(req
->cur_desc_num
== 0);
309 trace_spapr_vscsi_fetch_desc_direct();
312 case SRP_DATA_DESC_INDIRECT
: {
313 struct srp_indirect_buf
*tmp
= (struct srp_indirect_buf
*)
314 (cmd
->add_data
+ req
->cdb_offset
);
315 if (n
< req
->local_desc
) {
316 *ret
= tmp
->desc_list
[n
];
317 trace_spapr_vscsi_fetch_desc_indirect(req
->qtag
, n
,
319 } else if (n
< req
->total_desc
) {
321 struct srp_direct_buf tbl_desc
= vscsi_swap_desc(tmp
->table_desc
);
322 unsigned desc_offset
= n
* sizeof(struct srp_direct_buf
);
324 if (desc_offset
>= tbl_desc
.len
) {
325 trace_spapr_vscsi_fetch_desc_out_of_range(n
, desc_offset
);
328 rc
= spapr_vio_dma_read(&s
->vdev
, tbl_desc
.va
+ desc_offset
,
329 ret
, sizeof(struct srp_direct_buf
));
331 trace_spapr_vscsi_fetch_desc_dma_read_error(rc
);
334 trace_spapr_vscsi_fetch_desc_indirect_seg_ext(req
->qtag
, n
,
339 trace_spapr_vscsi_fetch_desc_out_of_desc();
345 fprintf(stderr
, "VSCSI: Unknown format %x\n", req
->dma_fmt
);
349 *ret
= vscsi_swap_desc(*ret
);
350 if (buf_offset
> ret
->len
) {
351 trace_spapr_vscsi_fetch_desc_out_of_desc_boundary(buf_offset
,
356 ret
->va
+= buf_offset
;
357 ret
->len
-= buf_offset
;
359 trace_spapr_vscsi_fetch_desc_done(req
->cur_desc_num
, req
->cur_desc_offset
,
362 return ret
->len
? 1 : 0;
365 static int vscsi_srp_direct_data(VSCSIState
*s
, vscsi_req
*req
,
366 uint8_t *buf
, uint32_t len
)
368 struct srp_direct_buf md
;
372 rc
= vscsi_fetch_desc(s
, req
, req
->cur_desc_num
, req
->cur_desc_offset
, &md
);
375 } else if (rc
== 0) {
379 llen
= MIN(len
, md
.len
);
381 if (req
->writing
) { /* writing = to device = reading from memory */
382 rc
= spapr_vio_dma_read(&s
->vdev
, md
.va
, buf
, llen
);
384 rc
= spapr_vio_dma_write(&s
->vdev
, md
.va
, buf
, llen
);
391 req
->cur_desc_offset
+= llen
;
396 static int vscsi_srp_indirect_data(VSCSIState
*s
, vscsi_req
*req
,
397 uint8_t *buf
, uint32_t len
)
399 struct srp_direct_buf md
;
401 uint32_t llen
, total
= 0;
403 trace_spapr_vscsi_srp_indirect_data(len
);
405 /* While we have data ... */
407 rc
= vscsi_fetch_desc(s
, req
, req
->cur_desc_num
, req
->cur_desc_offset
, &md
);
410 } else if (rc
== 0) {
414 /* Perform transfer */
415 llen
= MIN(len
, md
.len
);
416 if (req
->writing
) { /* writing = to device = reading from memory */
417 rc
= spapr_vio_dma_read(&s
->vdev
, md
.va
, buf
, llen
);
419 rc
= spapr_vio_dma_write(&s
->vdev
, md
.va
, buf
, llen
);
422 trace_spapr_vscsi_srp_indirect_data_rw(req
->writing
, rc
);
425 trace_spapr_vscsi_srp_indirect_data_buf(buf
[0], buf
[1], buf
[2], buf
[3]);
432 /* Update current position in the current descriptor */
433 req
->cur_desc_offset
+= llen
;
434 if (md
.len
== llen
) {
435 /* Go to the next descriptor if the current one finished */
437 req
->cur_desc_offset
= 0;
441 return rc
? -1 : total
;
444 static int vscsi_srp_transfer_data(VSCSIState
*s
, vscsi_req
*req
,
445 int writing
, uint8_t *buf
, uint32_t len
)
449 switch (req
->dma_fmt
) {
450 case SRP_NO_DATA_DESC
:
451 trace_spapr_vscsi_srp_transfer_data(len
);
453 case SRP_DATA_DESC_DIRECT
:
454 err
= vscsi_srp_direct_data(s
, req
, buf
, len
);
456 case SRP_DATA_DESC_INDIRECT
:
457 err
= vscsi_srp_indirect_data(s
, req
, buf
, len
);
463 /* Bits from linux srp */
464 static int data_out_desc_size(struct srp_cmd
*cmd
)
467 uint8_t fmt
= cmd
->buf_fmt
>> 4;
470 case SRP_NO_DATA_DESC
:
472 case SRP_DATA_DESC_DIRECT
:
473 size
= sizeof(struct srp_direct_buf
);
475 case SRP_DATA_DESC_INDIRECT
:
476 size
= sizeof(struct srp_indirect_buf
) +
477 sizeof(struct srp_direct_buf
)*cmd
->data_out_desc_cnt
;
485 static int vscsi_preprocess_desc(vscsi_req
*req
)
487 struct srp_cmd
*cmd
= &req_iu(req
)->srp
.cmd
;
489 req
->cdb_offset
= cmd
->add_cdb_len
& ~3;
492 req
->dma_fmt
= cmd
->buf_fmt
>> 4;
494 req
->cdb_offset
+= data_out_desc_size(cmd
);
495 req
->dma_fmt
= cmd
->buf_fmt
& ((1U << 4) - 1);
498 switch (req
->dma_fmt
) {
499 case SRP_NO_DATA_DESC
:
501 case SRP_DATA_DESC_DIRECT
:
502 req
->total_desc
= req
->local_desc
= 1;
504 case SRP_DATA_DESC_INDIRECT
: {
505 struct srp_indirect_buf
*ind_tmp
= (struct srp_indirect_buf
*)
506 (cmd
->add_data
+ req
->cdb_offset
);
508 req
->total_desc
= be32_to_cpu(ind_tmp
->table_desc
.len
) /
509 sizeof(struct srp_direct_buf
);
510 req
->local_desc
= req
->writing
? cmd
->data_out_desc_cnt
:
511 cmd
->data_in_desc_cnt
;
516 "vscsi_preprocess_desc: Unknown format %x\n", req
->dma_fmt
);
523 /* Callback to indicate that the SCSI layer has completed a transfer. */
524 static void vscsi_transfer_data(SCSIRequest
*sreq
, uint32_t len
)
526 VSCSIState
*s
= VIO_SPAPR_VSCSI_DEVICE(sreq
->bus
->qbus
.parent
);
527 vscsi_req
*req
= sreq
->hba_private
;
531 trace_spapr_vscsi_transfer_data(sreq
->tag
, len
, req
);
533 fprintf(stderr
, "VSCSI: Can't find request for tag 0x%x\n", sreq
->tag
);
538 buf
= scsi_req_get_buf(sreq
);
539 rc
= vscsi_srp_transfer_data(s
, req
, req
->writing
, buf
, len
);
542 fprintf(stderr
, "VSCSI: RDMA error rc=%d!\n", rc
);
543 req
->dma_error
= true;
544 scsi_req_cancel(req
->sreq
);
548 /* Start next chunk */
550 scsi_req_continue(sreq
);
553 /* Callback to indicate that the SCSI layer has completed a transfer. */
554 static void vscsi_command_complete(SCSIRequest
*sreq
, size_t resid
)
556 VSCSIState
*s
= VIO_SPAPR_VSCSI_DEVICE(sreq
->bus
->qbus
.parent
);
557 vscsi_req
*req
= sreq
->hba_private
;
558 int32_t res_in
= 0, res_out
= 0;
560 trace_spapr_vscsi_command_complete(sreq
->tag
, sreq
->status
, req
);
562 fprintf(stderr
, "VSCSI: Can't find request for tag 0x%x\n", sreq
->tag
);
566 if (sreq
->status
== CHECK_CONDITION
) {
567 req
->senselen
= scsi_req_get_sense(req
->sreq
, req
->sense
,
569 trace_spapr_vscsi_command_complete_sense_data1(req
->senselen
,
570 req
->sense
[0], req
->sense
[1], req
->sense
[2], req
->sense
[3],
571 req
->sense
[4], req
->sense
[5], req
->sense
[6], req
->sense
[7]);
572 trace_spapr_vscsi_command_complete_sense_data2(
573 req
->sense
[8], req
->sense
[9], req
->sense
[10], req
->sense
[11],
574 req
->sense
[12], req
->sense
[13], req
->sense
[14], req
->sense
[15]);
577 trace_spapr_vscsi_command_complete_status(sreq
->status
);
578 if (sreq
->status
== 0) {
579 /* We handle overflows, not underflows for normal commands,
580 * but hopefully nobody cares
583 res_out
= req
->data_len
;
585 res_in
= req
->data_len
;
588 vscsi_send_rsp(s
, req
, sreq
->status
, res_in
, res_out
);
592 static void vscsi_request_cancelled(SCSIRequest
*sreq
)
594 vscsi_req
*req
= sreq
->hba_private
;
596 if (req
->dma_error
) {
597 VSCSIState
*s
= VIO_SPAPR_VSCSI_DEVICE(sreq
->bus
->qbus
.parent
);
599 vscsi_makeup_sense(s
, req
, HARDWARE_ERROR
, 0, 0);
600 vscsi_send_rsp(s
, req
, CHECK_CONDITION
, 0, 0);
605 static const VMStateDescription vmstate_spapr_vscsi_req
= {
606 .name
= "spapr_vscsi_req",
608 .minimum_version_id
= 1,
609 .fields
= (VMStateField
[]) {
610 VMSTATE_BUFFER(crq
.raw
, vscsi_req
),
611 VMSTATE_BUFFER(viosrp_iu_buf
, vscsi_req
),
612 VMSTATE_UINT32(qtag
, vscsi_req
),
613 VMSTATE_BOOL(active
, vscsi_req
),
614 VMSTATE_UINT32(data_len
, vscsi_req
),
615 VMSTATE_BOOL(writing
, vscsi_req
),
616 VMSTATE_UINT32(senselen
, vscsi_req
),
617 VMSTATE_BUFFER(sense
, vscsi_req
),
618 VMSTATE_UINT8(dma_fmt
, vscsi_req
),
619 VMSTATE_UINT16(local_desc
, vscsi_req
),
620 VMSTATE_UINT16(total_desc
, vscsi_req
),
621 VMSTATE_UINT16(cdb_offset
, vscsi_req
),
622 /*Restart SCSI request from the beginning for now */
623 /*VMSTATE_UINT16(cur_desc_num, vscsi_req),
624 VMSTATE_UINT16(cur_desc_offset, vscsi_req),*/
625 VMSTATE_END_OF_LIST()
629 static void vscsi_save_request(QEMUFile
*f
, SCSIRequest
*sreq
)
631 vscsi_req
*req
= sreq
->hba_private
;
634 vmstate_save_state(f
, &vmstate_spapr_vscsi_req
, req
, NULL
);
636 trace_spapr_vscsi_save_request(req
->qtag
, req
->cur_desc_num
,
637 req
->cur_desc_offset
);
640 static void *vscsi_load_request(QEMUFile
*f
, SCSIRequest
*sreq
)
642 SCSIBus
*bus
= sreq
->bus
;
643 VSCSIState
*s
= VIO_SPAPR_VSCSI_DEVICE(bus
->qbus
.parent
);
647 assert(sreq
->tag
< VSCSI_REQ_LIMIT
);
648 req
= &s
->reqs
[sreq
->tag
];
649 assert(!req
->active
);
651 memset(req
, 0, sizeof(*req
));
652 rc
= vmstate_load_state(f
, &vmstate_spapr_vscsi_req
, req
, 1);
654 fprintf(stderr
, "VSCSI: failed loading request tag#%u\n", sreq
->tag
);
659 req
->sreq
= scsi_req_ref(sreq
);
661 trace_spapr_vscsi_load_request(req
->qtag
, req
->cur_desc_num
,
662 req
->cur_desc_offset
);
667 static void vscsi_process_login(VSCSIState
*s
, vscsi_req
*req
)
669 union viosrp_iu
*iu
= req_iu(req
);
670 struct srp_login_rsp
*rsp
= &iu
->srp
.login_rsp
;
671 uint64_t tag
= iu
->srp
.rsp
.tag
;
673 trace_spapr_vscsi_process_login();
675 /* TODO handle case that requested size is wrong and
676 * buffer format is wrong
678 memset(iu
, 0, sizeof(struct srp_login_rsp
));
679 rsp
->opcode
= SRP_LOGIN_RSP
;
680 /* Don't advertise quite as many request as we support to
681 * keep room for management stuff etc...
683 rsp
->req_lim_delta
= cpu_to_be32(VSCSI_REQ_LIMIT
-2);
685 rsp
->max_it_iu_len
= cpu_to_be32(SRP_MAX_IU_LEN
);
686 rsp
->max_ti_iu_len
= cpu_to_be32(SRP_MAX_IU_LEN
);
687 /* direct and indirect */
688 rsp
->buf_fmt
= cpu_to_be16(SRP_BUF_FORMAT_DIRECT
| SRP_BUF_FORMAT_INDIRECT
);
690 vscsi_send_iu(s
, req
, sizeof(*rsp
), VIOSRP_SRP_FORMAT
);
693 static void vscsi_inquiry_no_target(VSCSIState
*s
, vscsi_req
*req
)
695 uint8_t *cdb
= req_iu(req
)->srp
.cmd
.cdb
;
696 uint8_t resp_data
[36];
699 /* We don't do EVPD. Also check that page_code is 0 */
700 if ((cdb
[1] & 0x01) || cdb
[2] != 0) {
701 /* Send INVALID FIELD IN CDB */
702 vscsi_makeup_sense(s
, req
, ILLEGAL_REQUEST
, 0x24, 0);
703 vscsi_send_rsp(s
, req
, CHECK_CONDITION
, 0, 0);
707 alen
= (alen
<< 8) | cdb
[4];
710 /* Fake up inquiry using PQ=3 */
711 memset(resp_data
, 0, 36);
712 resp_data
[0] = 0x7f; /* Not capable of supporting a device here */
713 resp_data
[2] = 0x06; /* SPS-4 */
714 resp_data
[3] = 0x02; /* Resp data format */
715 resp_data
[4] = 36 - 5; /* Additional length */
716 resp_data
[7] = 0x10; /* Sync transfers */
717 memcpy(&resp_data
[16], "QEMU EMPTY ", 16);
718 memcpy(&resp_data
[8], "QEMU ", 8);
721 vscsi_preprocess_desc(req
);
722 rc
= vscsi_srp_transfer_data(s
, req
, 0, resp_data
, len
);
724 vscsi_makeup_sense(s
, req
, HARDWARE_ERROR
, 0, 0);
725 vscsi_send_rsp(s
, req
, CHECK_CONDITION
, 0, 0);
727 vscsi_send_rsp(s
, req
, 0, 36 - rc
, 0);
731 static void vscsi_report_luns(VSCSIState
*s
, vscsi_req
*req
)
740 QTAILQ_FOREACH(kid
, &s
->bus
.qbus
.children
, sibling
) {
741 SCSIDevice
*dev
= SCSI_DEVICE(kid
->child
);
744 if (dev
->channel
== 0 && dev
->id
== 0 && dev
->lun
== 0) {
753 resp_data
= g_malloc0(len
);
754 stl_be_p(resp_data
, n
);
755 i
= found_lun0
? 8 : 16;
756 QTAILQ_FOREACH(kid
, &s
->bus
.qbus
.children
, sibling
) {
757 DeviceState
*qdev
= kid
->child
;
758 SCSIDevice
*dev
= SCSI_DEVICE(qdev
);
760 if (dev
->id
== 0 && dev
->channel
== 0) {
761 resp_data
[i
] = 0; /* Use simple LUN for 0 (SAM5 4.7.7.1) */
763 resp_data
[i
] = (2 << 6); /* Otherwise LUN addressing (4.7.7.4) */
765 resp_data
[i
] |= dev
->id
;
766 resp_data
[i
+1] = (dev
->channel
<< 5);
767 resp_data
[i
+1] |= dev
->lun
;
771 vscsi_preprocess_desc(req
);
772 rc
= vscsi_srp_transfer_data(s
, req
, 0, resp_data
, len
);
775 vscsi_makeup_sense(s
, req
, HARDWARE_ERROR
, 0, 0);
776 vscsi_send_rsp(s
, req
, CHECK_CONDITION
, 0, 0);
778 vscsi_send_rsp(s
, req
, 0, len
- rc
, 0);
782 static int vscsi_queue_cmd(VSCSIState
*s
, vscsi_req
*req
)
784 union srp_iu
*srp
= &req_iu(req
)->srp
;
788 if ((srp
->cmd
.lun
== 0 || be64_to_cpu(srp
->cmd
.lun
) == SRP_REPORT_LUNS_WLUN
)
789 && srp
->cmd
.cdb
[0] == REPORT_LUNS
) {
790 vscsi_report_luns(s
, req
);
794 sdev
= vscsi_device_find(&s
->bus
, be64_to_cpu(srp
->cmd
.lun
), &lun
);
796 trace_spapr_vscsi_queue_cmd_no_drive(be64_to_cpu(srp
->cmd
.lun
));
797 if (srp
->cmd
.cdb
[0] == INQUIRY
) {
798 vscsi_inquiry_no_target(s
, req
);
800 vscsi_makeup_sense(s
, req
, ILLEGAL_REQUEST
, 0x24, 0x00);
801 vscsi_send_rsp(s
, req
, CHECK_CONDITION
, 0, 0);
805 req
->sreq
= scsi_req_new(sdev
, req
->qtag
, lun
, srp
->cmd
.cdb
, req
);
806 n
= scsi_req_enqueue(req
->sreq
);
808 trace_spapr_vscsi_queue_cmd(req
->qtag
, srp
->cmd
.cdb
[0],
809 scsi_command_name(srp
->cmd
.cdb
[0]), lun
, n
);
812 /* Transfer direction must be set before preprocessing the
815 req
->writing
= (n
< 1);
817 /* Preprocess RDMA descriptors */
818 vscsi_preprocess_desc(req
);
820 /* Get transfer direction and initiate transfer */
826 scsi_req_continue(req
->sreq
);
828 /* Don't touch req here, it may have been recycled already */
833 static int vscsi_process_tsk_mgmt(VSCSIState
*s
, vscsi_req
*req
)
835 union viosrp_iu
*iu
= req_iu(req
);
837 int i
, lun
= 0, resp
= SRP_TSK_MGMT_COMPLETE
;
839 uint64_t tag
= iu
->srp
.rsp
.tag
;
840 uint8_t sol_not
= iu
->srp
.cmd
.sol_not
;
842 trace_spapr_vscsi_process_tsk_mgmt(iu
->srp
.tsk_mgmt
.tsk_mgmt_func
);
843 d
= vscsi_device_find(&s
->bus
,
844 be64_to_cpu(req_iu(req
)->srp
.tsk_mgmt
.lun
), &lun
);
846 resp
= SRP_TSK_MGMT_FIELDS_INVALID
;
848 switch (iu
->srp
.tsk_mgmt
.tsk_mgmt_func
) {
849 case SRP_TSK_ABORT_TASK
:
851 resp
= SRP_TSK_MGMT_FIELDS_INVALID
;
855 tmpreq
= vscsi_find_req(s
, req_iu(req
)->srp
.tsk_mgmt
.task_tag
);
856 if (tmpreq
&& tmpreq
->sreq
) {
857 assert(tmpreq
->sreq
->hba_private
);
858 scsi_req_cancel(tmpreq
->sreq
);
862 case SRP_TSK_LUN_RESET
:
864 resp
= SRP_TSK_MGMT_FIELDS_INVALID
;
868 qdev_reset_all(&d
->qdev
);
871 case SRP_TSK_ABORT_TASK_SET
:
872 case SRP_TSK_CLEAR_TASK_SET
:
874 resp
= SRP_TSK_MGMT_FIELDS_INVALID
;
878 for (i
= 0; i
< VSCSI_REQ_LIMIT
; i
++) {
879 tmpreq
= &s
->reqs
[i
];
880 if (req_iu(tmpreq
)->srp
.cmd
.lun
881 != req_iu(req
)->srp
.tsk_mgmt
.lun
) {
884 if (!tmpreq
->active
|| !tmpreq
->sreq
) {
887 assert(tmpreq
->sreq
->hba_private
);
888 scsi_req_cancel(tmpreq
->sreq
);
892 case SRP_TSK_CLEAR_ACA
:
893 resp
= SRP_TSK_MGMT_NOT_SUPPORTED
;
897 resp
= SRP_TSK_MGMT_FIELDS_INVALID
;
902 /* Compose the response here as */
903 QEMU_BUILD_BUG_ON(SRP_MAX_IU_DATA_LEN
< 4);
904 memset(iu
, 0, sizeof(struct srp_rsp
) + 4);
905 iu
->srp
.rsp
.opcode
= SRP_RSP
;
906 iu
->srp
.rsp
.req_lim_delta
= cpu_to_be32(1);
907 iu
->srp
.rsp
.tag
= tag
;
908 iu
->srp
.rsp
.flags
|= SRP_RSP_FLAG_RSPVALID
;
909 iu
->srp
.rsp
.resp_data_len
= cpu_to_be32(4);
911 iu
->srp
.rsp
.sol_not
= (sol_not
& 0x04) >> 2;
913 iu
->srp
.rsp
.sol_not
= (sol_not
& 0x02) >> 1;
916 iu
->srp
.rsp
.status
= GOOD
;
917 iu
->srp
.rsp
.data
[3] = resp
;
919 vscsi_send_iu(s
, req
, sizeof(iu
->srp
.rsp
) + 4, VIOSRP_SRP_FORMAT
);
924 static int vscsi_handle_srp_req(VSCSIState
*s
, vscsi_req
*req
)
926 union srp_iu
*srp
= &req_iu(req
)->srp
;
928 uint8_t opcode
= srp
->rsp
.opcode
;
932 vscsi_process_login(s
, req
);
935 done
= vscsi_process_tsk_mgmt(s
, req
);
938 done
= vscsi_queue_cmd(s
, req
);
948 fprintf(stderr
, "VSCSI: Unsupported opcode %02x\n", opcode
);
951 fprintf(stderr
, "VSCSI: Unknown type %02x\n", opcode
);
957 static int vscsi_send_adapter_info(VSCSIState
*s
, vscsi_req
*req
)
959 struct viosrp_adapter_info
*sinfo
;
960 struct mad_adapter_info_data info
;
963 sinfo
= &req_iu(req
)->mad
.adapter_info
;
965 #if 0 /* What for ? */
966 rc
= spapr_vio_dma_read(&s
->vdev
, be64_to_cpu(sinfo
->buffer
),
967 &info
, be16_to_cpu(sinfo
->common
.length
));
969 fprintf(stderr
, "vscsi_send_adapter_info: DMA read failure !\n");
972 memset(&info
, 0, sizeof(info
));
973 strcpy(info
.srp_version
, SRP_VERSION
);
974 memcpy(info
.partition_name
, "qemu", sizeof("qemu"));
975 info
.partition_number
= cpu_to_be32(0);
976 info
.mad_version
= cpu_to_be32(1);
977 info
.os_type
= cpu_to_be32(2);
978 info
.port_max_txu
[0] = cpu_to_be32(VSCSI_MAX_SECTORS
<< 9);
980 rc
= spapr_vio_dma_write(&s
->vdev
, be64_to_cpu(sinfo
->buffer
),
981 &info
, be16_to_cpu(sinfo
->common
.length
));
983 fprintf(stderr
, "vscsi_send_adapter_info: DMA write failure !\n");
986 sinfo
->common
.status
= rc
? cpu_to_be32(1) : 0;
988 return vscsi_send_iu(s
, req
, sizeof(*sinfo
), VIOSRP_MAD_FORMAT
);
991 static int vscsi_send_capabilities(VSCSIState
*s
, vscsi_req
*req
)
993 struct viosrp_capabilities
*vcap
;
994 struct capabilities cap
= { };
995 uint16_t len
, req_len
;
999 vcap
= &req_iu(req
)->mad
.capabilities
;
1000 req_len
= len
= be16_to_cpu(vcap
->common
.length
);
1001 buffer
= be64_to_cpu(vcap
->buffer
);
1002 if (len
> sizeof(cap
)) {
1003 fprintf(stderr
, "vscsi_send_capabilities: capabilities size mismatch !\n");
1006 * Just read and populate the structure that is known.
1007 * Zero rest of the structure.
1011 rc
= spapr_vio_dma_read(&s
->vdev
, buffer
, &cap
, len
);
1013 fprintf(stderr
, "vscsi_send_capabilities: DMA read failure !\n");
1017 * Current implementation does not suppport any migration or
1018 * reservation capabilities. Construct the response telling the
1019 * guest not to use them.
1022 cap
.migration
.ecl
= 0;
1023 cap
.reserve
.type
= 0;
1024 cap
.migration
.common
.server_support
= 0;
1025 cap
.reserve
.common
.server_support
= 0;
1027 rc
= spapr_vio_dma_write(&s
->vdev
, buffer
, &cap
, len
);
1029 fprintf(stderr
, "vscsi_send_capabilities: DMA write failure !\n");
1031 if (req_len
> len
) {
1033 * Being paranoid and lets not worry about the error code
1034 * here. Actual write of the cap is done above.
1036 spapr_vio_dma_set(&s
->vdev
, (buffer
+ len
), 0, (req_len
- len
));
1038 vcap
->common
.status
= rc
? cpu_to_be32(1) : 0;
1039 return vscsi_send_iu(s
, req
, sizeof(*vcap
), VIOSRP_MAD_FORMAT
);
1042 static int vscsi_handle_mad_req(VSCSIState
*s
, vscsi_req
*req
)
1044 union mad_iu
*mad
= &req_iu(req
)->mad
;
1045 bool request_handled
= false;
1046 uint64_t retlen
= 0;
1048 switch (be32_to_cpu(mad
->empty_iu
.common
.type
)) {
1049 case VIOSRP_EMPTY_IU_TYPE
:
1050 fprintf(stderr
, "Unsupported EMPTY MAD IU\n");
1051 retlen
= sizeof(mad
->empty_iu
);
1053 case VIOSRP_ERROR_LOG_TYPE
:
1054 fprintf(stderr
, "Unsupported ERROR LOG MAD IU\n");
1055 retlen
= sizeof(mad
->error_log
);
1057 case VIOSRP_ADAPTER_INFO_TYPE
:
1058 vscsi_send_adapter_info(s
, req
);
1059 request_handled
= true;
1061 case VIOSRP_HOST_CONFIG_TYPE
:
1062 retlen
= sizeof(mad
->host_config
);
1064 case VIOSRP_CAPABILITIES_TYPE
:
1065 vscsi_send_capabilities(s
, req
);
1066 request_handled
= true;
1069 fprintf(stderr
, "VSCSI: Unknown MAD type %02x\n",
1070 be32_to_cpu(mad
->empty_iu
.common
.type
));
1072 * PAPR+ says that "The length field is set to the length
1073 * of the data structure(s) used in the command".
1074 * As we did not recognize the request type, put zero there.
1079 if (!request_handled
) {
1080 mad
->empty_iu
.common
.status
= cpu_to_be16(VIOSRP_MAD_NOT_SUPPORTED
);
1081 vscsi_send_iu(s
, req
, retlen
, VIOSRP_MAD_FORMAT
);
1087 static void vscsi_got_payload(VSCSIState
*s
, vscsi_crq
*crq
)
1092 req
= vscsi_get_req(s
);
1094 fprintf(stderr
, "VSCSI: Failed to get a request !\n");
1098 /* We only support a limited number of descriptors, we know
1099 * the ibmvscsi driver uses up to 10 max, so it should fit
1100 * in our 256 bytes IUs. If not we'll have to increase the size
1103 if (crq
->s
.IU_length
> SRP_MAX_IU_LEN
) {
1104 fprintf(stderr
, "VSCSI: SRP IU too long (%d bytes) !\n",
1110 /* XXX Handle failure differently ? */
1111 if (spapr_vio_dma_read(&s
->vdev
, crq
->s
.IU_data_ptr
, &req
->viosrp_iu_buf
,
1112 crq
->s
.IU_length
)) {
1113 fprintf(stderr
, "vscsi_got_payload: DMA read failure !\n");
1117 memcpy(&req
->crq
, crq
, sizeof(vscsi_crq
));
1119 if (crq
->s
.format
== VIOSRP_MAD_FORMAT
) {
1120 done
= vscsi_handle_mad_req(s
, req
);
1122 done
= vscsi_handle_srp_req(s
, req
);
1131 static int vscsi_do_crq(struct SpaprVioDevice
*dev
, uint8_t *crq_data
)
1133 VSCSIState
*s
= VIO_SPAPR_VSCSI_DEVICE(dev
);
1136 memcpy(crq
.raw
, crq_data
, 16);
1137 crq
.s
.timeout
= be16_to_cpu(crq
.s
.timeout
);
1138 crq
.s
.IU_length
= be16_to_cpu(crq
.s
.IU_length
);
1139 crq
.s
.IU_data_ptr
= be64_to_cpu(crq
.s
.IU_data_ptr
);
1141 trace_spapr_vscsi_do_crq(crq
.raw
[0], crq
.raw
[1]);
1143 switch (crq
.s
.valid
) {
1144 case 0xc0: /* Init command/response */
1146 /* Respond to initialization request */
1147 if (crq
.s
.format
== 0x01) {
1148 memset(crq
.raw
, 0, 16);
1150 crq
.s
.format
= 0x02;
1151 spapr_vio_send_crq(dev
, crq
.raw
);
1154 /* Note that in hotplug cases, we might get a 0x02
1155 * as a result of us emitting the init request
1159 case 0xff: /* Link event */
1161 /* Not handled for now */
1164 case 0x80: /* Payloads */
1165 switch (crq
.s
.format
) {
1166 case VIOSRP_SRP_FORMAT
: /* AKA VSCSI request */
1167 case VIOSRP_MAD_FORMAT
: /* AKA VSCSI response */
1168 vscsi_got_payload(s
, &crq
);
1170 case VIOSRP_OS400_FORMAT
:
1171 case VIOSRP_AIX_FORMAT
:
1172 case VIOSRP_LINUX_FORMAT
:
1173 case VIOSRP_INLINE_FORMAT
:
1174 fprintf(stderr
, "vscsi_do_srq: Unsupported payload format %02x\n",
1178 fprintf(stderr
, "vscsi_do_srq: Unknown payload format %02x\n",
1183 fprintf(stderr
, "vscsi_do_crq: unknown CRQ %02x %02x ...\n",
1184 crq
.raw
[0], crq
.raw
[1]);
1190 static const struct SCSIBusInfo vscsi_scsi_info
= {
1192 .max_channel
= 7, /* logical unit addressing format */
1196 .transfer_data
= vscsi_transfer_data
,
1197 .complete
= vscsi_command_complete
,
1198 .cancel
= vscsi_request_cancelled
,
1199 .save_request
= vscsi_save_request
,
1200 .load_request
= vscsi_load_request
,
1203 static void spapr_vscsi_reset(SpaprVioDevice
*dev
)
1205 VSCSIState
*s
= VIO_SPAPR_VSCSI_DEVICE(dev
);
1208 memset(s
->reqs
, 0, sizeof(s
->reqs
));
1209 for (i
= 0; i
< VSCSI_REQ_LIMIT
; i
++) {
1210 s
->reqs
[i
].qtag
= i
;
1214 static void spapr_vscsi_realize(SpaprVioDevice
*dev
, Error
**errp
)
1216 VSCSIState
*s
= VIO_SPAPR_VSCSI_DEVICE(dev
);
1218 dev
->crq
.SendFunc
= vscsi_do_crq
;
1220 scsi_bus_new(&s
->bus
, sizeof(s
->bus
), DEVICE(dev
),
1221 &vscsi_scsi_info
, NULL
);
1223 /* ibmvscsi SCSI bus does not allow hotplug. */
1224 qbus_set_hotplug_handler(BUS(&s
->bus
), NULL
);
1227 void spapr_vscsi_create(SpaprVioBus
*bus
)
1231 dev
= qdev_new("spapr-vscsi");
1233 qdev_realize_and_unref(dev
, &bus
->bus
, &error_fatal
);
1234 scsi_bus_legacy_handle_cmdline(&VIO_SPAPR_VSCSI_DEVICE(dev
)->bus
);
1237 static int spapr_vscsi_devnode(SpaprVioDevice
*dev
, void *fdt
, int node_off
)
1241 ret
= fdt_setprop_cell(fdt
, node_off
, "#address-cells", 2);
1246 ret
= fdt_setprop_cell(fdt
, node_off
, "#size-cells", 0);
1254 static Property spapr_vscsi_properties
[] = {
1255 DEFINE_SPAPR_PROPERTIES(VSCSIState
, vdev
),
1256 DEFINE_PROP_END_OF_LIST(),
1259 static const VMStateDescription vmstate_spapr_vscsi
= {
1260 .name
= "spapr_vscsi",
1262 .minimum_version_id
= 1,
1263 .fields
= (VMStateField
[]) {
1264 VMSTATE_SPAPR_VIO(vdev
, VSCSIState
),
1268 VMSTATE_END_OF_LIST()
1272 static void spapr_vscsi_class_init(ObjectClass
*klass
, void *data
)
1274 DeviceClass
*dc
= DEVICE_CLASS(klass
);
1275 SpaprVioDeviceClass
*k
= VIO_SPAPR_DEVICE_CLASS(klass
);
1277 k
->realize
= spapr_vscsi_realize
;
1278 k
->reset
= spapr_vscsi_reset
;
1279 k
->devnode
= spapr_vscsi_devnode
;
1280 k
->dt_name
= "v-scsi";
1281 k
->dt_type
= "vscsi";
1282 k
->dt_compatible
= "IBM,v-scsi";
1283 k
->signal_mask
= 0x00000001;
1284 set_bit(DEVICE_CATEGORY_STORAGE
, dc
->categories
);
1285 device_class_set_props(dc
, spapr_vscsi_properties
);
1286 k
->rtce_window_size
= 0x10000000;
1287 dc
->vmsd
= &vmstate_spapr_vscsi
;
1290 static const TypeInfo spapr_vscsi_info
= {
1291 .name
= TYPE_VIO_SPAPR_VSCSI_DEVICE
,
1292 .parent
= TYPE_VIO_SPAPR_DEVICE
,
1293 .instance_size
= sizeof(VSCSIState
),
1294 .class_init
= spapr_vscsi_class_init
,
1297 static void spapr_vscsi_register_types(void)
1299 type_register_static(&spapr_vscsi_info
);
1302 type_init(spapr_vscsi_register_types
)