vfio/pci: Cache vendor and device ID
[qemu/ar7.git] / hw / scsi / spapr_vscsi.c
blob891424fae9e2f8afe65a14651dac0da7aa083642
1 /*
2 * QEMU PowerPC pSeries Logical Partition (aka sPAPR) hardware System Emulator
4 * PAPR Virtual SCSI, aka ibmvscsi
6 * Copyright (c) 2010,2011 Benjamin Herrenschmidt, IBM Corporation.
8 * Permission is hereby granted, free of charge, to any person obtaining a copy
9 * of this software and associated documentation files (the "Software"), to deal
10 * in the Software without restriction, including without limitation the rights
11 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
12 * copies of the Software, and to permit persons to whom the Software is
13 * furnished to do so, subject to the following conditions:
15 * The above copyright notice and this permission notice shall be included in
16 * all copies or substantial portions of the Software.
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
19 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
20 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
21 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
22 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
23 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
24 * THE SOFTWARE.
26 * TODO:
28 * - Cleanups :-)
29 * - Sort out better how to assign devices to VSCSI instances
30 * - Fix residual counts
31 * - Add indirect descriptors support
32 * - Maybe do autosense (PAPR seems to mandate it, linux doesn't care)
34 #include "hw/hw.h"
35 #include "hw/scsi/scsi.h"
36 #include "block/scsi.h"
37 #include "srp.h"
38 #include "hw/qdev.h"
39 #include "hw/ppc/spapr.h"
40 #include "hw/ppc/spapr_vio.h"
41 #include "viosrp.h"
43 #include <libfdt.h>
45 /*#define DEBUG_VSCSI*/
47 #ifdef DEBUG_VSCSI
48 #define DPRINTF(fmt, ...) \
49 do { fprintf(stderr, fmt, ## __VA_ARGS__); } while (0)
50 #else
51 #define DPRINTF(fmt, ...) \
52 do { } while (0)
53 #endif
56 * Virtual SCSI device
59 /* Random numbers */
60 #define VSCSI_MAX_SECTORS 4096
61 #define VSCSI_REQ_LIMIT 24
63 #define SRP_RSP_SENSE_DATA_LEN 18
65 #define SRP_REPORT_LUNS_WLUN 0xc10100000000000ULL
67 typedef union vscsi_crq {
68 struct viosrp_crq s;
69 uint8_t raw[16];
70 } vscsi_crq;
72 typedef struct vscsi_req {
73 vscsi_crq crq;
74 union viosrp_iu iu;
76 /* SCSI request tracking */
77 SCSIRequest *sreq;
78 uint32_t qtag; /* qemu tag != srp tag */
79 bool active;
80 bool writing;
81 bool dma_error;
82 uint32_t data_len;
83 uint32_t senselen;
84 uint8_t sense[SCSI_SENSE_BUF_SIZE];
86 /* RDMA related bits */
87 uint8_t dma_fmt;
88 uint16_t local_desc;
89 uint16_t total_desc;
90 uint16_t cdb_offset;
91 uint16_t cur_desc_num;
92 uint16_t cur_desc_offset;
93 } vscsi_req;
95 #define TYPE_VIO_SPAPR_VSCSI_DEVICE "spapr-vscsi"
96 #define VIO_SPAPR_VSCSI_DEVICE(obj) \
97 OBJECT_CHECK(VSCSIState, (obj), TYPE_VIO_SPAPR_VSCSI_DEVICE)
99 typedef struct {
100 VIOsPAPRDevice vdev;
101 SCSIBus bus;
102 vscsi_req reqs[VSCSI_REQ_LIMIT];
103 } VSCSIState;
105 static struct vscsi_req *vscsi_get_req(VSCSIState *s)
107 vscsi_req *req;
108 int i;
110 for (i = 0; i < VSCSI_REQ_LIMIT; i++) {
111 req = &s->reqs[i];
112 if (!req->active) {
113 memset(req, 0, sizeof(*req));
114 req->qtag = i;
115 req->active = 1;
116 return req;
119 return NULL;
122 static struct vscsi_req *vscsi_find_req(VSCSIState *s, uint64_t srp_tag)
124 vscsi_req *req;
125 int i;
127 for (i = 0; i < VSCSI_REQ_LIMIT; i++) {
128 req = &s->reqs[i];
129 if (req->iu.srp.cmd.tag == srp_tag) {
130 return req;
133 return NULL;
136 static void vscsi_put_req(vscsi_req *req)
138 if (req->sreq != NULL) {
139 scsi_req_unref(req->sreq);
141 req->sreq = NULL;
142 req->active = 0;
145 static SCSIDevice *vscsi_device_find(SCSIBus *bus, uint64_t srp_lun, int *lun)
147 int channel = 0, id = 0;
149 retry:
150 switch (srp_lun >> 62) {
151 case 0:
152 if ((srp_lun >> 56) != 0) {
153 channel = (srp_lun >> 56) & 0x3f;
154 id = (srp_lun >> 48) & 0xff;
155 srp_lun <<= 16;
156 goto retry;
158 *lun = (srp_lun >> 48) & 0xff;
159 break;
161 case 1:
162 *lun = (srp_lun >> 48) & 0x3fff;
163 break;
164 case 2:
165 channel = (srp_lun >> 53) & 0x7;
166 id = (srp_lun >> 56) & 0x3f;
167 *lun = (srp_lun >> 48) & 0x1f;
168 break;
169 case 3:
170 *lun = -1;
171 return NULL;
172 default:
173 abort();
176 return scsi_device_find(bus, channel, id, *lun);
179 static int vscsi_send_iu(VSCSIState *s, vscsi_req *req,
180 uint64_t length, uint8_t format)
182 long rc, rc1;
184 /* First copy the SRP */
185 rc = spapr_vio_dma_write(&s->vdev, req->crq.s.IU_data_ptr,
186 &req->iu, length);
187 if (rc) {
188 fprintf(stderr, "vscsi_send_iu: DMA write failure !\n");
191 req->crq.s.valid = 0x80;
192 req->crq.s.format = format;
193 req->crq.s.reserved = 0x00;
194 req->crq.s.timeout = cpu_to_be16(0x0000);
195 req->crq.s.IU_length = cpu_to_be16(length);
196 req->crq.s.IU_data_ptr = req->iu.srp.rsp.tag; /* right byte order */
198 if (rc == 0) {
199 req->crq.s.status = VIOSRP_OK;
200 } else {
201 req->crq.s.status = VIOSRP_ADAPTER_FAIL;
204 rc1 = spapr_vio_send_crq(&s->vdev, req->crq.raw);
205 if (rc1) {
206 fprintf(stderr, "vscsi_send_iu: Error sending response\n");
207 return rc1;
210 return rc;
213 static void vscsi_makeup_sense(VSCSIState *s, vscsi_req *req,
214 uint8_t key, uint8_t asc, uint8_t ascq)
216 req->senselen = SRP_RSP_SENSE_DATA_LEN;
218 /* Valid bit and 'current errors' */
219 req->sense[0] = (0x1 << 7 | 0x70);
220 /* Sense key */
221 req->sense[2] = key;
222 /* Additional sense length */
223 req->sense[7] = 0xa; /* 10 bytes */
224 /* Additional sense code */
225 req->sense[12] = asc;
226 req->sense[13] = ascq;
229 static int vscsi_send_rsp(VSCSIState *s, vscsi_req *req,
230 uint8_t status, int32_t res_in, int32_t res_out)
232 union viosrp_iu *iu = &req->iu;
233 uint64_t tag = iu->srp.rsp.tag;
234 int total_len = sizeof(iu->srp.rsp);
235 uint8_t sol_not = iu->srp.cmd.sol_not;
237 DPRINTF("VSCSI: Sending resp status: 0x%x, "
238 "res_in: %d, res_out: %d\n", status, res_in, res_out);
240 memset(iu, 0, sizeof(struct srp_rsp));
241 iu->srp.rsp.opcode = SRP_RSP;
242 iu->srp.rsp.req_lim_delta = cpu_to_be32(1);
243 iu->srp.rsp.tag = tag;
245 /* Handle residuals */
246 if (res_in < 0) {
247 iu->srp.rsp.flags |= SRP_RSP_FLAG_DIUNDER;
248 res_in = -res_in;
249 } else if (res_in) {
250 iu->srp.rsp.flags |= SRP_RSP_FLAG_DIOVER;
252 if (res_out < 0) {
253 iu->srp.rsp.flags |= SRP_RSP_FLAG_DOUNDER;
254 res_out = -res_out;
255 } else if (res_out) {
256 iu->srp.rsp.flags |= SRP_RSP_FLAG_DOOVER;
258 iu->srp.rsp.data_in_res_cnt = cpu_to_be32(res_in);
259 iu->srp.rsp.data_out_res_cnt = cpu_to_be32(res_out);
261 /* We don't do response data */
262 /* iu->srp.rsp.flags &= ~SRP_RSP_FLAG_RSPVALID; */
263 iu->srp.rsp.resp_data_len = cpu_to_be32(0);
265 /* Handle success vs. failure */
266 iu->srp.rsp.status = status;
267 if (status) {
268 iu->srp.rsp.sol_not = (sol_not & 0x04) >> 2;
269 if (req->senselen) {
270 req->iu.srp.rsp.flags |= SRP_RSP_FLAG_SNSVALID;
271 req->iu.srp.rsp.sense_data_len = cpu_to_be32(req->senselen);
272 memcpy(req->iu.srp.rsp.data, req->sense, req->senselen);
273 total_len += req->senselen;
275 } else {
276 iu->srp.rsp.sol_not = (sol_not & 0x02) >> 1;
279 vscsi_send_iu(s, req, total_len, VIOSRP_SRP_FORMAT);
280 return 0;
283 static inline struct srp_direct_buf vscsi_swap_desc(struct srp_direct_buf desc)
285 desc.va = be64_to_cpu(desc.va);
286 desc.len = be32_to_cpu(desc.len);
287 return desc;
290 static int vscsi_fetch_desc(VSCSIState *s, struct vscsi_req *req,
291 unsigned n, unsigned buf_offset,
292 struct srp_direct_buf *ret)
294 struct srp_cmd *cmd = &req->iu.srp.cmd;
296 switch (req->dma_fmt) {
297 case SRP_NO_DATA_DESC: {
298 DPRINTF("VSCSI: no data descriptor\n");
299 return 0;
301 case SRP_DATA_DESC_DIRECT: {
302 memcpy(ret, cmd->add_data + req->cdb_offset, sizeof(*ret));
303 assert(req->cur_desc_num == 0);
304 DPRINTF("VSCSI: direct segment\n");
305 break;
307 case SRP_DATA_DESC_INDIRECT: {
308 struct srp_indirect_buf *tmp = (struct srp_indirect_buf *)
309 (cmd->add_data + req->cdb_offset);
310 if (n < req->local_desc) {
311 *ret = tmp->desc_list[n];
312 DPRINTF("VSCSI: indirect segment local tag=0x%x desc#%d/%d\n",
313 req->qtag, n, req->local_desc);
315 } else if (n < req->total_desc) {
316 int rc;
317 struct srp_direct_buf tbl_desc = vscsi_swap_desc(tmp->table_desc);
318 unsigned desc_offset = n * sizeof(struct srp_direct_buf);
320 if (desc_offset >= tbl_desc.len) {
321 DPRINTF("VSCSI: #%d is ouf of range (%d bytes)\n",
322 n, desc_offset);
323 return -1;
325 rc = spapr_vio_dma_read(&s->vdev, tbl_desc.va + desc_offset,
326 ret, sizeof(struct srp_direct_buf));
327 if (rc) {
328 DPRINTF("VSCSI: spapr_vio_dma_read -> %d reading ext_desc\n",
329 rc);
330 return -1;
332 DPRINTF("VSCSI: indirect segment ext. tag=0x%x desc#%d/%d { va=%"PRIx64" len=%x }\n",
333 req->qtag, n, req->total_desc, tbl_desc.va, tbl_desc.len);
334 } else {
335 DPRINTF("VSCSI: Out of descriptors !\n");
336 return 0;
338 break;
340 default:
341 fprintf(stderr, "VSCSI: Unknown format %x\n", req->dma_fmt);
342 return -1;
345 *ret = vscsi_swap_desc(*ret);
346 if (buf_offset > ret->len) {
347 DPRINTF(" offset=%x is out of a descriptor #%d boundary=%x\n",
348 buf_offset, req->cur_desc_num, ret->len);
349 return -1;
351 ret->va += buf_offset;
352 ret->len -= buf_offset;
354 DPRINTF(" cur=%d offs=%x ret { va=%"PRIx64" len=%x }\n",
355 req->cur_desc_num, req->cur_desc_offset, ret->va, ret->len);
357 return ret->len ? 1 : 0;
360 static int vscsi_srp_direct_data(VSCSIState *s, vscsi_req *req,
361 uint8_t *buf, uint32_t len)
363 struct srp_direct_buf md;
364 uint32_t llen;
365 int rc = 0;
367 rc = vscsi_fetch_desc(s, req, req->cur_desc_num, req->cur_desc_offset, &md);
368 if (rc < 0) {
369 return -1;
370 } else if (rc == 0) {
371 return 0;
374 llen = MIN(len, md.len);
375 if (llen) {
376 if (req->writing) { /* writing = to device = reading from memory */
377 rc = spapr_vio_dma_read(&s->vdev, md.va, buf, llen);
378 } else {
379 rc = spapr_vio_dma_write(&s->vdev, md.va, buf, llen);
383 if (rc) {
384 return -1;
386 req->cur_desc_offset += llen;
388 return llen;
391 static int vscsi_srp_indirect_data(VSCSIState *s, vscsi_req *req,
392 uint8_t *buf, uint32_t len)
394 struct srp_direct_buf md;
395 int rc = 0;
396 uint32_t llen, total = 0;
398 DPRINTF("VSCSI: indirect segment 0x%x bytes\n", len);
400 /* While we have data ... */
401 while (len) {
402 rc = vscsi_fetch_desc(s, req, req->cur_desc_num, req->cur_desc_offset, &md);
403 if (rc < 0) {
404 return -1;
405 } else if (rc == 0) {
406 break;
409 /* Perform transfer */
410 llen = MIN(len, md.len);
411 if (req->writing) { /* writing = to device = reading from memory */
412 rc = spapr_vio_dma_read(&s->vdev, md.va, buf, llen);
413 } else {
414 rc = spapr_vio_dma_write(&s->vdev, md.va, buf, llen);
416 if (rc) {
417 DPRINTF("VSCSI: spapr_vio_dma_r/w(%d) -> %d\n", req->writing, rc);
418 break;
420 DPRINTF("VSCSI: data: %02x %02x %02x %02x...\n",
421 buf[0], buf[1], buf[2], buf[3]);
423 len -= llen;
424 buf += llen;
426 total += llen;
428 /* Update current position in the current descriptor */
429 req->cur_desc_offset += llen;
430 if (md.len == llen) {
431 /* Go to the next descriptor if the current one finished */
432 ++req->cur_desc_num;
433 req->cur_desc_offset = 0;
437 return rc ? -1 : total;
440 static int vscsi_srp_transfer_data(VSCSIState *s, vscsi_req *req,
441 int writing, uint8_t *buf, uint32_t len)
443 int err = 0;
445 switch (req->dma_fmt) {
446 case SRP_NO_DATA_DESC:
447 DPRINTF("VSCSI: no data desc transfer, skipping 0x%x bytes\n", len);
448 break;
449 case SRP_DATA_DESC_DIRECT:
450 err = vscsi_srp_direct_data(s, req, buf, len);
451 break;
452 case SRP_DATA_DESC_INDIRECT:
453 err = vscsi_srp_indirect_data(s, req, buf, len);
454 break;
456 return err;
459 /* Bits from linux srp */
460 static int data_out_desc_size(struct srp_cmd *cmd)
462 int size = 0;
463 uint8_t fmt = cmd->buf_fmt >> 4;
465 switch (fmt) {
466 case SRP_NO_DATA_DESC:
467 break;
468 case SRP_DATA_DESC_DIRECT:
469 size = sizeof(struct srp_direct_buf);
470 break;
471 case SRP_DATA_DESC_INDIRECT:
472 size = sizeof(struct srp_indirect_buf) +
473 sizeof(struct srp_direct_buf)*cmd->data_out_desc_cnt;
474 break;
475 default:
476 break;
478 return size;
481 static int vscsi_preprocess_desc(vscsi_req *req)
483 struct srp_cmd *cmd = &req->iu.srp.cmd;
485 req->cdb_offset = cmd->add_cdb_len & ~3;
487 if (req->writing) {
488 req->dma_fmt = cmd->buf_fmt >> 4;
489 } else {
490 req->cdb_offset += data_out_desc_size(cmd);
491 req->dma_fmt = cmd->buf_fmt & ((1U << 4) - 1);
494 switch (req->dma_fmt) {
495 case SRP_NO_DATA_DESC:
496 break;
497 case SRP_DATA_DESC_DIRECT:
498 req->total_desc = req->local_desc = 1;
499 break;
500 case SRP_DATA_DESC_INDIRECT: {
501 struct srp_indirect_buf *ind_tmp = (struct srp_indirect_buf *)
502 (cmd->add_data + req->cdb_offset);
504 req->total_desc = be32_to_cpu(ind_tmp->table_desc.len) /
505 sizeof(struct srp_direct_buf);
506 req->local_desc = req->writing ? cmd->data_out_desc_cnt :
507 cmd->data_in_desc_cnt;
508 break;
510 default:
511 fprintf(stderr,
512 "vscsi_preprocess_desc: Unknown format %x\n", req->dma_fmt);
513 return -1;
516 return 0;
519 /* Callback to indicate that the SCSI layer has completed a transfer. */
520 static void vscsi_transfer_data(SCSIRequest *sreq, uint32_t len)
522 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(sreq->bus->qbus.parent);
523 vscsi_req *req = sreq->hba_private;
524 uint8_t *buf;
525 int rc = 0;
527 DPRINTF("VSCSI: SCSI xfer complete tag=0x%x len=0x%x, req=%p\n",
528 sreq->tag, len, req);
529 if (req == NULL) {
530 fprintf(stderr, "VSCSI: Can't find request for tag 0x%x\n", sreq->tag);
531 return;
534 if (len) {
535 buf = scsi_req_get_buf(sreq);
536 rc = vscsi_srp_transfer_data(s, req, req->writing, buf, len);
538 if (rc < 0) {
539 fprintf(stderr, "VSCSI: RDMA error rc=%d!\n", rc);
540 req->dma_error = true;
541 scsi_req_cancel(req->sreq);
542 return;
545 /* Start next chunk */
546 req->data_len -= rc;
547 scsi_req_continue(sreq);
550 /* Callback to indicate that the SCSI layer has completed a transfer. */
551 static void vscsi_command_complete(SCSIRequest *sreq, uint32_t status, size_t resid)
553 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(sreq->bus->qbus.parent);
554 vscsi_req *req = sreq->hba_private;
555 int32_t res_in = 0, res_out = 0;
557 DPRINTF("VSCSI: SCSI cmd complete, tag=0x%x status=0x%x, req=%p\n",
558 sreq->tag, status, req);
559 if (req == NULL) {
560 fprintf(stderr, "VSCSI: Can't find request for tag 0x%x\n", sreq->tag);
561 return;
564 if (status == CHECK_CONDITION) {
565 req->senselen = scsi_req_get_sense(req->sreq, req->sense,
566 sizeof(req->sense));
567 DPRINTF("VSCSI: Sense data, %d bytes:\n", req->senselen);
568 DPRINTF(" %02x %02x %02x %02x %02x %02x %02x %02x\n",
569 req->sense[0], req->sense[1], req->sense[2], req->sense[3],
570 req->sense[4], req->sense[5], req->sense[6], req->sense[7]);
571 DPRINTF(" %02x %02x %02x %02x %02x %02x %02x %02x\n",
572 req->sense[8], req->sense[9], req->sense[10], req->sense[11],
573 req->sense[12], req->sense[13], req->sense[14], req->sense[15]);
576 DPRINTF("VSCSI: Command complete err=%d\n", status);
577 if (status == 0) {
578 /* We handle overflows, not underflows for normal commands,
579 * but hopefully nobody cares
581 if (req->writing) {
582 res_out = req->data_len;
583 } else {
584 res_in = req->data_len;
587 vscsi_send_rsp(s, req, status, res_in, res_out);
588 vscsi_put_req(req);
591 static void vscsi_request_cancelled(SCSIRequest *sreq)
593 vscsi_req *req = sreq->hba_private;
595 if (req->dma_error) {
596 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(sreq->bus->qbus.parent);
598 vscsi_makeup_sense(s, req, HARDWARE_ERROR, 0, 0);
599 vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0);
601 vscsi_put_req(req);
604 static const VMStateDescription vmstate_spapr_vscsi_req = {
605 .name = "spapr_vscsi_req",
606 .version_id = 1,
607 .minimum_version_id = 1,
608 .fields = (VMStateField[]) {
609 VMSTATE_BUFFER(crq.raw, vscsi_req),
610 VMSTATE_BUFFER(iu.srp.reserved, vscsi_req),
611 VMSTATE_UINT32(qtag, vscsi_req),
612 VMSTATE_BOOL(active, vscsi_req),
613 VMSTATE_UINT32(data_len, vscsi_req),
614 VMSTATE_BOOL(writing, vscsi_req),
615 VMSTATE_UINT32(senselen, vscsi_req),
616 VMSTATE_BUFFER(sense, vscsi_req),
617 VMSTATE_UINT8(dma_fmt, vscsi_req),
618 VMSTATE_UINT16(local_desc, vscsi_req),
619 VMSTATE_UINT16(total_desc, vscsi_req),
620 VMSTATE_UINT16(cdb_offset, vscsi_req),
621 /*Restart SCSI request from the beginning for now */
622 /*VMSTATE_UINT16(cur_desc_num, vscsi_req),
623 VMSTATE_UINT16(cur_desc_offset, vscsi_req),*/
624 VMSTATE_END_OF_LIST()
628 static void vscsi_save_request(QEMUFile *f, SCSIRequest *sreq)
630 vscsi_req *req = sreq->hba_private;
631 assert(req->active);
633 vmstate_save_state(f, &vmstate_spapr_vscsi_req, req, NULL);
635 DPRINTF("VSCSI: saving tag=%u, current desc#%d, offset=%x\n",
636 req->qtag, req->cur_desc_num, req->cur_desc_offset);
639 static void *vscsi_load_request(QEMUFile *f, SCSIRequest *sreq)
641 SCSIBus *bus = sreq->bus;
642 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(bus->qbus.parent);
643 vscsi_req *req;
644 int rc;
646 assert(sreq->tag < VSCSI_REQ_LIMIT);
647 req = &s->reqs[sreq->tag];
648 assert(!req->active);
650 memset(req, 0, sizeof(*req));
651 rc = vmstate_load_state(f, &vmstate_spapr_vscsi_req, req, 1);
652 if (rc) {
653 fprintf(stderr, "VSCSI: failed loading request tag#%u\n", sreq->tag);
654 return NULL;
656 assert(req->active);
658 req->sreq = scsi_req_ref(sreq);
660 DPRINTF("VSCSI: restoring tag=%u, current desc#%d, offset=%x\n",
661 req->qtag, req->cur_desc_num, req->cur_desc_offset);
663 return req;
666 static void vscsi_process_login(VSCSIState *s, vscsi_req *req)
668 union viosrp_iu *iu = &req->iu;
669 struct srp_login_rsp *rsp = &iu->srp.login_rsp;
670 uint64_t tag = iu->srp.rsp.tag;
672 DPRINTF("VSCSI: Got login, sendin response !\n");
674 /* TODO handle case that requested size is wrong and
675 * buffer format is wrong
677 memset(iu, 0, sizeof(struct srp_login_rsp));
678 rsp->opcode = SRP_LOGIN_RSP;
679 /* Don't advertise quite as many request as we support to
680 * keep room for management stuff etc...
682 rsp->req_lim_delta = cpu_to_be32(VSCSI_REQ_LIMIT-2);
683 rsp->tag = tag;
684 rsp->max_it_iu_len = cpu_to_be32(sizeof(union srp_iu));
685 rsp->max_ti_iu_len = cpu_to_be32(sizeof(union srp_iu));
686 /* direct and indirect */
687 rsp->buf_fmt = cpu_to_be16(SRP_BUF_FORMAT_DIRECT | SRP_BUF_FORMAT_INDIRECT);
689 vscsi_send_iu(s, req, sizeof(*rsp), VIOSRP_SRP_FORMAT);
692 static void vscsi_inquiry_no_target(VSCSIState *s, vscsi_req *req)
694 uint8_t *cdb = req->iu.srp.cmd.cdb;
695 uint8_t resp_data[36];
696 int rc, len, alen;
698 /* We dont do EVPD. Also check that page_code is 0 */
699 if ((cdb[1] & 0x01) || cdb[2] != 0) {
700 /* Send INVALID FIELD IN CDB */
701 vscsi_makeup_sense(s, req, ILLEGAL_REQUEST, 0x24, 0);
702 vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0);
703 return;
705 alen = cdb[3];
706 alen = (alen << 8) | cdb[4];
707 len = MIN(alen, 36);
709 /* Fake up inquiry using PQ=3 */
710 memset(resp_data, 0, 36);
711 resp_data[0] = 0x7f; /* Not capable of supporting a device here */
712 resp_data[2] = 0x06; /* SPS-4 */
713 resp_data[3] = 0x02; /* Resp data format */
714 resp_data[4] = 36 - 5; /* Additional length */
715 resp_data[7] = 0x10; /* Sync transfers */
716 memcpy(&resp_data[16], "QEMU EMPTY ", 16);
717 memcpy(&resp_data[8], "QEMU ", 8);
719 req->writing = 0;
720 vscsi_preprocess_desc(req);
721 rc = vscsi_srp_transfer_data(s, req, 0, resp_data, len);
722 if (rc < 0) {
723 vscsi_makeup_sense(s, req, HARDWARE_ERROR, 0, 0);
724 vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0);
725 } else {
726 vscsi_send_rsp(s, req, 0, 36 - rc, 0);
730 static void vscsi_report_luns(VSCSIState *s, vscsi_req *req)
732 BusChild *kid;
733 int i, len, n, rc;
734 uint8_t *resp_data;
735 bool found_lun0;
737 n = 0;
738 found_lun0 = false;
739 QTAILQ_FOREACH(kid, &s->bus.qbus.children, sibling) {
740 SCSIDevice *dev = SCSI_DEVICE(kid->child);
742 n += 8;
743 if (dev->channel == 0 && dev->id == 0 && dev->lun == 0) {
744 found_lun0 = true;
747 if (!found_lun0) {
748 n += 8;
750 len = n+8;
752 resp_data = g_malloc0(len);
753 memset(resp_data, 0, len);
754 stl_be_p(resp_data, n);
755 i = found_lun0 ? 8 : 16;
756 QTAILQ_FOREACH(kid, &s->bus.qbus.children, sibling) {
757 DeviceState *qdev = kid->child;
758 SCSIDevice *dev = SCSI_DEVICE(qdev);
760 if (dev->id == 0 && dev->channel == 0) {
761 resp_data[i] = 0; /* Use simple LUN for 0 (SAM5 4.7.7.1) */
762 } else {
763 resp_data[i] = (2 << 6); /* Otherwise LUN addressing (4.7.7.4) */
765 resp_data[i] |= dev->id;
766 resp_data[i+1] = (dev->channel << 5);
767 resp_data[i+1] |= dev->lun;
768 i += 8;
771 vscsi_preprocess_desc(req);
772 rc = vscsi_srp_transfer_data(s, req, 0, resp_data, len);
773 g_free(resp_data);
774 if (rc < 0) {
775 vscsi_makeup_sense(s, req, HARDWARE_ERROR, 0, 0);
776 vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0);
777 } else {
778 vscsi_send_rsp(s, req, 0, len - rc, 0);
782 static int vscsi_queue_cmd(VSCSIState *s, vscsi_req *req)
784 union srp_iu *srp = &req->iu.srp;
785 SCSIDevice *sdev;
786 int n, lun;
788 if ((srp->cmd.lun == 0 || be64_to_cpu(srp->cmd.lun) == SRP_REPORT_LUNS_WLUN)
789 && srp->cmd.cdb[0] == REPORT_LUNS) {
790 vscsi_report_luns(s, req);
791 return 0;
794 sdev = vscsi_device_find(&s->bus, be64_to_cpu(srp->cmd.lun), &lun);
795 if (!sdev) {
796 DPRINTF("VSCSI: Command for lun %08" PRIx64 " with no drive\n",
797 be64_to_cpu(srp->cmd.lun));
798 if (srp->cmd.cdb[0] == INQUIRY) {
799 vscsi_inquiry_no_target(s, req);
800 } else {
801 vscsi_makeup_sense(s, req, ILLEGAL_REQUEST, 0x24, 0x00);
802 vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0);
803 } return 1;
806 req->sreq = scsi_req_new(sdev, req->qtag, lun, srp->cmd.cdb, req);
807 n = scsi_req_enqueue(req->sreq);
809 DPRINTF("VSCSI: Queued command tag 0x%x CMD 0x%x=%s LUN %d ret: %d\n",
810 req->qtag, srp->cmd.cdb[0], scsi_command_name(srp->cmd.cdb[0]),
811 lun, n);
813 if (n) {
814 /* Transfer direction must be set before preprocessing the
815 * descriptors
817 req->writing = (n < 1);
819 /* Preprocess RDMA descriptors */
820 vscsi_preprocess_desc(req);
822 /* Get transfer direction and initiate transfer */
823 if (n > 0) {
824 req->data_len = n;
825 } else if (n < 0) {
826 req->data_len = -n;
828 scsi_req_continue(req->sreq);
830 /* Don't touch req here, it may have been recycled already */
832 return 0;
835 static int vscsi_process_tsk_mgmt(VSCSIState *s, vscsi_req *req)
837 union viosrp_iu *iu = &req->iu;
838 vscsi_req *tmpreq;
839 int i, lun = 0, resp = SRP_TSK_MGMT_COMPLETE;
840 SCSIDevice *d;
841 uint64_t tag = iu->srp.rsp.tag;
842 uint8_t sol_not = iu->srp.cmd.sol_not;
844 fprintf(stderr, "vscsi_process_tsk_mgmt %02x\n",
845 iu->srp.tsk_mgmt.tsk_mgmt_func);
847 d = vscsi_device_find(&s->bus, be64_to_cpu(req->iu.srp.tsk_mgmt.lun), &lun);
848 if (!d) {
849 resp = SRP_TSK_MGMT_FIELDS_INVALID;
850 } else {
851 switch (iu->srp.tsk_mgmt.tsk_mgmt_func) {
852 case SRP_TSK_ABORT_TASK:
853 if (d->lun != lun) {
854 resp = SRP_TSK_MGMT_FIELDS_INVALID;
855 break;
858 tmpreq = vscsi_find_req(s, req->iu.srp.tsk_mgmt.task_tag);
859 if (tmpreq && tmpreq->sreq) {
860 assert(tmpreq->sreq->hba_private);
861 scsi_req_cancel(tmpreq->sreq);
863 break;
865 case SRP_TSK_LUN_RESET:
866 if (d->lun != lun) {
867 resp = SRP_TSK_MGMT_FIELDS_INVALID;
868 break;
871 qdev_reset_all(&d->qdev);
872 break;
874 case SRP_TSK_ABORT_TASK_SET:
875 case SRP_TSK_CLEAR_TASK_SET:
876 if (d->lun != lun) {
877 resp = SRP_TSK_MGMT_FIELDS_INVALID;
878 break;
881 for (i = 0; i < VSCSI_REQ_LIMIT; i++) {
882 tmpreq = &s->reqs[i];
883 if (tmpreq->iu.srp.cmd.lun != req->iu.srp.tsk_mgmt.lun) {
884 continue;
886 if (!tmpreq->active || !tmpreq->sreq) {
887 continue;
889 assert(tmpreq->sreq->hba_private);
890 scsi_req_cancel(tmpreq->sreq);
892 break;
894 case SRP_TSK_CLEAR_ACA:
895 resp = SRP_TSK_MGMT_NOT_SUPPORTED;
896 break;
898 default:
899 resp = SRP_TSK_MGMT_FIELDS_INVALID;
900 break;
904 /* Compose the response here as */
905 memset(iu, 0, sizeof(struct srp_rsp) + 4);
906 iu->srp.rsp.opcode = SRP_RSP;
907 iu->srp.rsp.req_lim_delta = cpu_to_be32(1);
908 iu->srp.rsp.tag = tag;
909 iu->srp.rsp.flags |= SRP_RSP_FLAG_RSPVALID;
910 iu->srp.rsp.resp_data_len = cpu_to_be32(4);
911 if (resp) {
912 iu->srp.rsp.sol_not = (sol_not & 0x04) >> 2;
913 } else {
914 iu->srp.rsp.sol_not = (sol_not & 0x02) >> 1;
917 iu->srp.rsp.status = GOOD;
918 iu->srp.rsp.data[3] = resp;
920 vscsi_send_iu(s, req, sizeof(iu->srp.rsp) + 4, VIOSRP_SRP_FORMAT);
922 return 1;
925 static int vscsi_handle_srp_req(VSCSIState *s, vscsi_req *req)
927 union srp_iu *srp = &req->iu.srp;
928 int done = 1;
929 uint8_t opcode = srp->rsp.opcode;
931 switch (opcode) {
932 case SRP_LOGIN_REQ:
933 vscsi_process_login(s, req);
934 break;
935 case SRP_TSK_MGMT:
936 done = vscsi_process_tsk_mgmt(s, req);
937 break;
938 case SRP_CMD:
939 done = vscsi_queue_cmd(s, req);
940 break;
941 case SRP_LOGIN_RSP:
942 case SRP_I_LOGOUT:
943 case SRP_T_LOGOUT:
944 case SRP_RSP:
945 case SRP_CRED_REQ:
946 case SRP_CRED_RSP:
947 case SRP_AER_REQ:
948 case SRP_AER_RSP:
949 fprintf(stderr, "VSCSI: Unsupported opcode %02x\n", opcode);
950 break;
951 default:
952 fprintf(stderr, "VSCSI: Unknown type %02x\n", opcode);
955 return done;
958 static int vscsi_send_adapter_info(VSCSIState *s, vscsi_req *req)
960 struct viosrp_adapter_info *sinfo;
961 struct mad_adapter_info_data info;
962 int rc;
964 sinfo = &req->iu.mad.adapter_info;
966 #if 0 /* What for ? */
967 rc = spapr_vio_dma_read(&s->vdev, be64_to_cpu(sinfo->buffer),
968 &info, be16_to_cpu(sinfo->common.length));
969 if (rc) {
970 fprintf(stderr, "vscsi_send_adapter_info: DMA read failure !\n");
972 #endif
973 memset(&info, 0, sizeof(info));
974 strcpy(info.srp_version, SRP_VERSION);
975 memcpy(info.partition_name, "qemu", sizeof("qemu"));
976 info.partition_number = cpu_to_be32(0);
977 info.mad_version = cpu_to_be32(1);
978 info.os_type = cpu_to_be32(2);
979 info.port_max_txu[0] = cpu_to_be32(VSCSI_MAX_SECTORS << 9);
981 rc = spapr_vio_dma_write(&s->vdev, be64_to_cpu(sinfo->buffer),
982 &info, be16_to_cpu(sinfo->common.length));
983 if (rc) {
984 fprintf(stderr, "vscsi_send_adapter_info: DMA write failure !\n");
987 sinfo->common.status = rc ? cpu_to_be32(1) : 0;
989 return vscsi_send_iu(s, req, sizeof(*sinfo), VIOSRP_MAD_FORMAT);
992 static int vscsi_send_capabilities(VSCSIState *s, vscsi_req *req)
994 struct viosrp_capabilities *vcap;
995 struct capabilities cap = { };
996 uint16_t len, req_len;
997 uint64_t buffer;
998 int rc;
1000 vcap = &req->iu.mad.capabilities;
1001 req_len = len = be16_to_cpu(vcap->common.length);
1002 buffer = be64_to_cpu(vcap->buffer);
1003 if (len > sizeof(cap)) {
1004 fprintf(stderr, "vscsi_send_capabilities: capabilities size mismatch !\n");
1007 * Just read and populate the structure that is known.
1008 * Zero rest of the structure.
1010 len = sizeof(cap);
1012 rc = spapr_vio_dma_read(&s->vdev, buffer, &cap, len);
1013 if (rc) {
1014 fprintf(stderr, "vscsi_send_capabilities: DMA read failure !\n");
1018 * Current implementation does not suppport any migration or
1019 * reservation capabilities. Construct the response telling the
1020 * guest not to use them.
1022 cap.flags = 0;
1023 cap.migration.ecl = 0;
1024 cap.reserve.type = 0;
1025 cap.migration.common.server_support = 0;
1026 cap.reserve.common.server_support = 0;
1028 rc = spapr_vio_dma_write(&s->vdev, buffer, &cap, len);
1029 if (rc) {
1030 fprintf(stderr, "vscsi_send_capabilities: DMA write failure !\n");
1032 if (req_len > len) {
1034 * Being paranoid and lets not worry about the error code
1035 * here. Actual write of the cap is done above.
1037 spapr_vio_dma_set(&s->vdev, (buffer + len), 0, (req_len - len));
1039 vcap->common.status = rc ? cpu_to_be32(1) : 0;
1040 return vscsi_send_iu(s, req, sizeof(*vcap), VIOSRP_MAD_FORMAT);
1043 static int vscsi_handle_mad_req(VSCSIState *s, vscsi_req *req)
1045 union mad_iu *mad = &req->iu.mad;
1046 bool request_handled = false;
1047 uint64_t retlen = 0;
1049 switch (be32_to_cpu(mad->empty_iu.common.type)) {
1050 case VIOSRP_EMPTY_IU_TYPE:
1051 fprintf(stderr, "Unsupported EMPTY MAD IU\n");
1052 retlen = sizeof(mad->empty_iu);
1053 break;
1054 case VIOSRP_ERROR_LOG_TYPE:
1055 fprintf(stderr, "Unsupported ERROR LOG MAD IU\n");
1056 retlen = sizeof(mad->error_log);
1057 break;
1058 case VIOSRP_ADAPTER_INFO_TYPE:
1059 vscsi_send_adapter_info(s, req);
1060 request_handled = true;
1061 break;
1062 case VIOSRP_HOST_CONFIG_TYPE:
1063 retlen = sizeof(mad->host_config);
1064 break;
1065 case VIOSRP_CAPABILITIES_TYPE:
1066 vscsi_send_capabilities(s, req);
1067 request_handled = true;
1068 break;
1069 default:
1070 fprintf(stderr, "VSCSI: Unknown MAD type %02x\n",
1071 be32_to_cpu(mad->empty_iu.common.type));
1073 * PAPR+ says that "The length field is set to the length
1074 * of the data structure(s) used in the command".
1075 * As we did not recognize the request type, put zero there.
1077 retlen = 0;
1080 if (!request_handled) {
1081 mad->empty_iu.common.status = cpu_to_be16(VIOSRP_MAD_NOT_SUPPORTED);
1082 vscsi_send_iu(s, req, retlen, VIOSRP_MAD_FORMAT);
1085 return 1;
1088 static void vscsi_got_payload(VSCSIState *s, vscsi_crq *crq)
1090 vscsi_req *req;
1091 int done;
1093 req = vscsi_get_req(s);
1094 if (req == NULL) {
1095 fprintf(stderr, "VSCSI: Failed to get a request !\n");
1096 return;
1099 /* We only support a limited number of descriptors, we know
1100 * the ibmvscsi driver uses up to 10 max, so it should fit
1101 * in our 256 bytes IUs. If not we'll have to increase the size
1102 * of the structure.
1104 if (crq->s.IU_length > sizeof(union viosrp_iu)) {
1105 fprintf(stderr, "VSCSI: SRP IU too long (%d bytes) !\n",
1106 crq->s.IU_length);
1107 vscsi_put_req(req);
1108 return;
1111 /* XXX Handle failure differently ? */
1112 if (spapr_vio_dma_read(&s->vdev, crq->s.IU_data_ptr, &req->iu,
1113 crq->s.IU_length)) {
1114 fprintf(stderr, "vscsi_got_payload: DMA read failure !\n");
1115 vscsi_put_req(req);
1116 return;
1118 memcpy(&req->crq, crq, sizeof(vscsi_crq));
1120 if (crq->s.format == VIOSRP_MAD_FORMAT) {
1121 done = vscsi_handle_mad_req(s, req);
1122 } else {
1123 done = vscsi_handle_srp_req(s, req);
1126 if (done) {
1127 vscsi_put_req(req);
1132 static int vscsi_do_crq(struct VIOsPAPRDevice *dev, uint8_t *crq_data)
1134 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(dev);
1135 vscsi_crq crq;
1137 memcpy(crq.raw, crq_data, 16);
1138 crq.s.timeout = be16_to_cpu(crq.s.timeout);
1139 crq.s.IU_length = be16_to_cpu(crq.s.IU_length);
1140 crq.s.IU_data_ptr = be64_to_cpu(crq.s.IU_data_ptr);
1142 DPRINTF("VSCSI: do_crq %02x %02x ...\n", crq.raw[0], crq.raw[1]);
1144 switch (crq.s.valid) {
1145 case 0xc0: /* Init command/response */
1147 /* Respond to initialization request */
1148 if (crq.s.format == 0x01) {
1149 memset(crq.raw, 0, 16);
1150 crq.s.valid = 0xc0;
1151 crq.s.format = 0x02;
1152 spapr_vio_send_crq(dev, crq.raw);
1155 /* Note that in hotplug cases, we might get a 0x02
1156 * as a result of us emitting the init request
1159 break;
1160 case 0xff: /* Link event */
1162 /* Not handled for now */
1164 break;
1165 case 0x80: /* Payloads */
1166 switch (crq.s.format) {
1167 case VIOSRP_SRP_FORMAT: /* AKA VSCSI request */
1168 case VIOSRP_MAD_FORMAT: /* AKA VSCSI response */
1169 vscsi_got_payload(s, &crq);
1170 break;
1171 case VIOSRP_OS400_FORMAT:
1172 case VIOSRP_AIX_FORMAT:
1173 case VIOSRP_LINUX_FORMAT:
1174 case VIOSRP_INLINE_FORMAT:
1175 fprintf(stderr, "vscsi_do_srq: Unsupported payload format %02x\n",
1176 crq.s.format);
1177 break;
1178 default:
1179 fprintf(stderr, "vscsi_do_srq: Unknown payload format %02x\n",
1180 crq.s.format);
1182 break;
1183 default:
1184 fprintf(stderr, "vscsi_do_crq: unknown CRQ %02x %02x ...\n",
1185 crq.raw[0], crq.raw[1]);
1188 return 0;
1191 static const struct SCSIBusInfo vscsi_scsi_info = {
1192 .tcq = true,
1193 .max_channel = 7, /* logical unit addressing format */
1194 .max_target = 63,
1195 .max_lun = 31,
1197 .transfer_data = vscsi_transfer_data,
1198 .complete = vscsi_command_complete,
1199 .cancel = vscsi_request_cancelled,
1200 .save_request = vscsi_save_request,
1201 .load_request = vscsi_load_request,
1204 static void spapr_vscsi_reset(VIOsPAPRDevice *dev)
1206 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(dev);
1207 int i;
1209 memset(s->reqs, 0, sizeof(s->reqs));
1210 for (i = 0; i < VSCSI_REQ_LIMIT; i++) {
1211 s->reqs[i].qtag = i;
1215 static void spapr_vscsi_realize(VIOsPAPRDevice *dev, Error **errp)
1217 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(dev);
1219 dev->crq.SendFunc = vscsi_do_crq;
1221 scsi_bus_new(&s->bus, sizeof(s->bus), DEVICE(dev),
1222 &vscsi_scsi_info, NULL);
1223 if (!dev->qdev.hotplugged) {
1224 scsi_bus_legacy_handle_cmdline(&s->bus, errp);
1228 void spapr_vscsi_create(VIOsPAPRBus *bus)
1230 DeviceState *dev;
1232 dev = qdev_create(&bus->bus, "spapr-vscsi");
1234 qdev_init_nofail(dev);
1237 static int spapr_vscsi_devnode(VIOsPAPRDevice *dev, void *fdt, int node_off)
1239 int ret;
1241 ret = fdt_setprop_cell(fdt, node_off, "#address-cells", 2);
1242 if (ret < 0) {
1243 return ret;
1246 ret = fdt_setprop_cell(fdt, node_off, "#size-cells", 0);
1247 if (ret < 0) {
1248 return ret;
1251 return 0;
1254 static Property spapr_vscsi_properties[] = {
1255 DEFINE_SPAPR_PROPERTIES(VSCSIState, vdev),
1256 DEFINE_PROP_END_OF_LIST(),
1259 static const VMStateDescription vmstate_spapr_vscsi = {
1260 .name = "spapr_vscsi",
1261 .version_id = 1,
1262 .minimum_version_id = 1,
1263 .fields = (VMStateField[]) {
1264 VMSTATE_SPAPR_VIO(vdev, VSCSIState),
1265 /* VSCSI state */
1266 /* ???? */
1268 VMSTATE_END_OF_LIST()
1272 static void spapr_vscsi_class_init(ObjectClass *klass, void *data)
1274 DeviceClass *dc = DEVICE_CLASS(klass);
1275 VIOsPAPRDeviceClass *k = VIO_SPAPR_DEVICE_CLASS(klass);
1277 k->realize = spapr_vscsi_realize;
1278 k->reset = spapr_vscsi_reset;
1279 k->devnode = spapr_vscsi_devnode;
1280 k->dt_name = "v-scsi";
1281 k->dt_type = "vscsi";
1282 k->dt_compatible = "IBM,v-scsi";
1283 k->signal_mask = 0x00000001;
1284 set_bit(DEVICE_CATEGORY_STORAGE, dc->categories);
1285 dc->props = spapr_vscsi_properties;
1286 k->rtce_window_size = 0x10000000;
1287 dc->vmsd = &vmstate_spapr_vscsi;
1290 static const TypeInfo spapr_vscsi_info = {
1291 .name = TYPE_VIO_SPAPR_VSCSI_DEVICE,
1292 .parent = TYPE_VIO_SPAPR_DEVICE,
1293 .instance_size = sizeof(VSCSIState),
1294 .class_init = spapr_vscsi_class_init,
1297 static void spapr_vscsi_register_types(void)
1299 type_register_static(&spapr_vscsi_info);
1302 type_init(spapr_vscsi_register_types)