fe9d8927e2
When a SCSI command is received from the guest, the CDB length implied by the first byte might exceed the number of bytes the guest sent. In this case scsi_req_new() will read uninitialized data, causing unpredictable behavior. Adds the buf_len parameter to scsi_req_new() and plumbs it through the call stack. Signed-off-by: John Millikin <john@john-millikin.com> Resolves: https://gitlab.com/qemu-project/qemu/-/issues/1127 Message-Id: <20220817053458.698416-1-john@john-millikin.com> [Fill in correct length for adapters other than ESP. - Paolo] Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
1302 lines
38 KiB
C
1302 lines
38 KiB
C
/*
|
|
* QEMU PowerPC pSeries Logical Partition (aka sPAPR) hardware System Emulator
|
|
*
|
|
* PAPR Virtual SCSI, aka ibmvscsi
|
|
*
|
|
* Copyright (c) 2010,2011 Benjamin Herrenschmidt, IBM Corporation.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
* of this software and associated documentation files (the "Software"), to deal
|
|
* in the Software without restriction, including without limitation the rights
|
|
* to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
|
* copies of the Software, and to permit persons to whom the Software is
|
|
* furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
|
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
|
|
* THE SOFTWARE.
|
|
*
|
|
* TODO:
|
|
*
|
|
* - Cleanups :-)
|
|
* - Sort out better how to assign devices to VSCSI instances
|
|
* - Fix residual counts
|
|
* - Add indirect descriptors support
|
|
* - Maybe do autosense (PAPR seems to mandate it, linux doesn't care)
|
|
*/
|
|
|
|
#include "qemu/osdep.h"
|
|
#include "qemu/module.h"
|
|
#include "hw/scsi/scsi.h"
|
|
#include "migration/vmstate.h"
|
|
#include "scsi/constants.h"
|
|
#include "srp.h"
|
|
#include "hw/ppc/spapr.h"
|
|
#include "hw/ppc/spapr_vio.h"
|
|
#include "hw/qdev-properties.h"
|
|
#include "viosrp.h"
|
|
#include "trace.h"
|
|
|
|
#include <libfdt.h>
|
|
#include "qom/object.h"
|
|
|
|
/*
|
|
* Virtual SCSI device
|
|
*/
|
|
|
|
/* Random numbers */
|
|
#define VSCSI_MAX_SECTORS 4096
|
|
#define VSCSI_REQ_LIMIT 24
|
|
|
|
/* Maximum size of a IU payload */
|
|
#define SRP_MAX_IU_DATA_LEN (SRP_MAX_IU_LEN - sizeof(union srp_iu))
|
|
#define SRP_RSP_SENSE_DATA_LEN 18
|
|
|
|
#define SRP_REPORT_LUNS_WLUN 0xc10100000000000ULL
|
|
|
|
typedef union vscsi_crq {
|
|
struct viosrp_crq s;
|
|
uint8_t raw[16];
|
|
} vscsi_crq;
|
|
|
|
typedef struct vscsi_req {
|
|
vscsi_crq crq;
|
|
uint8_t viosrp_iu_buf[SRP_MAX_IU_LEN];
|
|
|
|
/* SCSI request tracking */
|
|
SCSIRequest *sreq;
|
|
uint32_t qtag; /* qemu tag != srp tag */
|
|
bool active;
|
|
bool writing;
|
|
bool dma_error;
|
|
uint32_t data_len;
|
|
uint32_t senselen;
|
|
uint8_t sense[SCSI_SENSE_BUF_SIZE];
|
|
|
|
/* RDMA related bits */
|
|
uint8_t dma_fmt;
|
|
uint16_t local_desc;
|
|
uint16_t total_desc;
|
|
uint16_t cdb_offset;
|
|
uint16_t cur_desc_num;
|
|
uint16_t cur_desc_offset;
|
|
} vscsi_req;
|
|
|
|
#define TYPE_VIO_SPAPR_VSCSI_DEVICE "spapr-vscsi"
|
|
OBJECT_DECLARE_SIMPLE_TYPE(VSCSIState, VIO_SPAPR_VSCSI_DEVICE)
|
|
|
|
struct VSCSIState {
|
|
SpaprVioDevice vdev;
|
|
SCSIBus bus;
|
|
vscsi_req reqs[VSCSI_REQ_LIMIT];
|
|
};
|
|
|
|
static union viosrp_iu *req_iu(vscsi_req *req)
|
|
{
|
|
return (union viosrp_iu *)req->viosrp_iu_buf;
|
|
}
|
|
|
|
static struct vscsi_req *vscsi_get_req(VSCSIState *s)
|
|
{
|
|
vscsi_req *req;
|
|
int i;
|
|
|
|
for (i = 0; i < VSCSI_REQ_LIMIT; i++) {
|
|
req = &s->reqs[i];
|
|
if (!req->active) {
|
|
memset(req, 0, sizeof(*req));
|
|
req->qtag = i;
|
|
req->active = 1;
|
|
return req;
|
|
}
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
static struct vscsi_req *vscsi_find_req(VSCSIState *s, uint64_t srp_tag)
|
|
{
|
|
vscsi_req *req;
|
|
int i;
|
|
|
|
for (i = 0; i < VSCSI_REQ_LIMIT; i++) {
|
|
req = &s->reqs[i];
|
|
if (req_iu(req)->srp.cmd.tag == srp_tag) {
|
|
return req;
|
|
}
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
static void vscsi_put_req(vscsi_req *req)
|
|
{
|
|
if (req->sreq != NULL) {
|
|
scsi_req_unref(req->sreq);
|
|
}
|
|
req->sreq = NULL;
|
|
req->active = 0;
|
|
}
|
|
|
|
static SCSIDevice *vscsi_device_find(SCSIBus *bus, uint64_t srp_lun, int *lun)
|
|
{
|
|
int channel = 0, id = 0;
|
|
|
|
retry:
|
|
switch (srp_lun >> 62) {
|
|
case 0:
|
|
if ((srp_lun >> 56) != 0) {
|
|
channel = (srp_lun >> 56) & 0x3f;
|
|
id = (srp_lun >> 48) & 0xff;
|
|
srp_lun <<= 16;
|
|
goto retry;
|
|
}
|
|
*lun = (srp_lun >> 48) & 0xff;
|
|
break;
|
|
|
|
case 1:
|
|
*lun = (srp_lun >> 48) & 0x3fff;
|
|
break;
|
|
case 2:
|
|
channel = (srp_lun >> 53) & 0x7;
|
|
id = (srp_lun >> 56) & 0x3f;
|
|
*lun = (srp_lun >> 48) & 0x1f;
|
|
break;
|
|
case 3:
|
|
*lun = -1;
|
|
return NULL;
|
|
default:
|
|
abort();
|
|
}
|
|
|
|
return scsi_device_find(bus, channel, id, *lun);
|
|
}
|
|
|
|
static int vscsi_send_iu(VSCSIState *s, vscsi_req *req,
|
|
uint64_t length, uint8_t format)
|
|
{
|
|
long rc, rc1;
|
|
|
|
assert(length <= SRP_MAX_IU_LEN);
|
|
|
|
/* First copy the SRP */
|
|
rc = spapr_vio_dma_write(&s->vdev, req->crq.s.IU_data_ptr,
|
|
&req->viosrp_iu_buf, length);
|
|
if (rc) {
|
|
fprintf(stderr, "vscsi_send_iu: DMA write failure !\n");
|
|
}
|
|
|
|
req->crq.s.valid = 0x80;
|
|
req->crq.s.format = format;
|
|
req->crq.s.reserved = 0x00;
|
|
req->crq.s.timeout = cpu_to_be16(0x0000);
|
|
req->crq.s.IU_length = cpu_to_be16(length);
|
|
req->crq.s.IU_data_ptr = req_iu(req)->srp.rsp.tag; /* right byte order */
|
|
|
|
if (rc == 0) {
|
|
req->crq.s.status = VIOSRP_OK;
|
|
} else {
|
|
req->crq.s.status = VIOSRP_ADAPTER_FAIL;
|
|
}
|
|
|
|
rc1 = spapr_vio_send_crq(&s->vdev, req->crq.raw);
|
|
if (rc1) {
|
|
fprintf(stderr, "vscsi_send_iu: Error sending response\n");
|
|
return rc1;
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
static void vscsi_makeup_sense(VSCSIState *s, vscsi_req *req,
|
|
uint8_t key, uint8_t asc, uint8_t ascq)
|
|
{
|
|
req->senselen = SRP_RSP_SENSE_DATA_LEN;
|
|
|
|
/* Valid bit and 'current errors' */
|
|
req->sense[0] = (0x1 << 7 | 0x70);
|
|
/* Sense key */
|
|
req->sense[2] = key;
|
|
/* Additional sense length */
|
|
req->sense[7] = 0xa; /* 10 bytes */
|
|
/* Additional sense code */
|
|
req->sense[12] = asc;
|
|
req->sense[13] = ascq;
|
|
}
|
|
|
|
static int vscsi_send_rsp(VSCSIState *s, vscsi_req *req,
|
|
uint8_t status, int32_t res_in, int32_t res_out)
|
|
{
|
|
union viosrp_iu *iu = req_iu(req);
|
|
uint64_t tag = iu->srp.rsp.tag;
|
|
int total_len = sizeof(iu->srp.rsp);
|
|
uint8_t sol_not = iu->srp.cmd.sol_not;
|
|
|
|
trace_spapr_vscsi_send_rsp(status, res_in, res_out);
|
|
|
|
memset(iu, 0, sizeof(struct srp_rsp));
|
|
iu->srp.rsp.opcode = SRP_RSP;
|
|
iu->srp.rsp.req_lim_delta = cpu_to_be32(1);
|
|
iu->srp.rsp.tag = tag;
|
|
|
|
/* Handle residuals */
|
|
if (res_in < 0) {
|
|
iu->srp.rsp.flags |= SRP_RSP_FLAG_DIUNDER;
|
|
res_in = -res_in;
|
|
} else if (res_in) {
|
|
iu->srp.rsp.flags |= SRP_RSP_FLAG_DIOVER;
|
|
}
|
|
if (res_out < 0) {
|
|
iu->srp.rsp.flags |= SRP_RSP_FLAG_DOUNDER;
|
|
res_out = -res_out;
|
|
} else if (res_out) {
|
|
iu->srp.rsp.flags |= SRP_RSP_FLAG_DOOVER;
|
|
}
|
|
iu->srp.rsp.data_in_res_cnt = cpu_to_be32(res_in);
|
|
iu->srp.rsp.data_out_res_cnt = cpu_to_be32(res_out);
|
|
|
|
/* We don't do response data */
|
|
/* iu->srp.rsp.flags &= ~SRP_RSP_FLAG_RSPVALID; */
|
|
iu->srp.rsp.resp_data_len = cpu_to_be32(0);
|
|
|
|
/* Handle success vs. failure */
|
|
iu->srp.rsp.status = status;
|
|
if (status) {
|
|
iu->srp.rsp.sol_not = (sol_not & 0x04) >> 2;
|
|
if (req->senselen) {
|
|
int sense_data_len = MIN(req->senselen, SRP_MAX_IU_DATA_LEN);
|
|
|
|
iu->srp.rsp.flags |= SRP_RSP_FLAG_SNSVALID;
|
|
iu->srp.rsp.sense_data_len = cpu_to_be32(sense_data_len);
|
|
memcpy(iu->srp.rsp.data, req->sense, sense_data_len);
|
|
total_len += sense_data_len;
|
|
}
|
|
} else {
|
|
iu->srp.rsp.sol_not = (sol_not & 0x02) >> 1;
|
|
}
|
|
|
|
vscsi_send_iu(s, req, total_len, VIOSRP_SRP_FORMAT);
|
|
return 0;
|
|
}
|
|
|
|
static inline struct srp_direct_buf vscsi_swap_desc(struct srp_direct_buf desc)
|
|
{
|
|
desc.va = be64_to_cpu(desc.va);
|
|
desc.len = be32_to_cpu(desc.len);
|
|
return desc;
|
|
}
|
|
|
|
static int vscsi_fetch_desc(VSCSIState *s, struct vscsi_req *req,
|
|
unsigned n, unsigned buf_offset,
|
|
struct srp_direct_buf *ret)
|
|
{
|
|
struct srp_cmd *cmd = &req_iu(req)->srp.cmd;
|
|
|
|
switch (req->dma_fmt) {
|
|
case SRP_NO_DATA_DESC: {
|
|
trace_spapr_vscsi_fetch_desc_no_data();
|
|
return 0;
|
|
}
|
|
case SRP_DATA_DESC_DIRECT: {
|
|
memcpy(ret, cmd->add_data + req->cdb_offset, sizeof(*ret));
|
|
assert(req->cur_desc_num == 0);
|
|
trace_spapr_vscsi_fetch_desc_direct();
|
|
break;
|
|
}
|
|
case SRP_DATA_DESC_INDIRECT: {
|
|
struct srp_indirect_buf *tmp = (struct srp_indirect_buf *)
|
|
(cmd->add_data + req->cdb_offset);
|
|
if (n < req->local_desc) {
|
|
*ret = tmp->desc_list[n];
|
|
trace_spapr_vscsi_fetch_desc_indirect(req->qtag, n,
|
|
req->local_desc);
|
|
} else if (n < req->total_desc) {
|
|
int rc;
|
|
struct srp_direct_buf tbl_desc = vscsi_swap_desc(tmp->table_desc);
|
|
unsigned desc_offset = n * sizeof(struct srp_direct_buf);
|
|
|
|
if (desc_offset >= tbl_desc.len) {
|
|
trace_spapr_vscsi_fetch_desc_out_of_range(n, desc_offset);
|
|
return -1;
|
|
}
|
|
rc = spapr_vio_dma_read(&s->vdev, tbl_desc.va + desc_offset,
|
|
ret, sizeof(struct srp_direct_buf));
|
|
if (rc) {
|
|
trace_spapr_vscsi_fetch_desc_dma_read_error(rc);
|
|
return -1;
|
|
}
|
|
trace_spapr_vscsi_fetch_desc_indirect_seg_ext(req->qtag, n,
|
|
req->total_desc,
|
|
tbl_desc.va,
|
|
tbl_desc.len);
|
|
} else {
|
|
trace_spapr_vscsi_fetch_desc_out_of_desc();
|
|
return 0;
|
|
}
|
|
break;
|
|
}
|
|
default:
|
|
fprintf(stderr, "VSCSI: Unknown format %x\n", req->dma_fmt);
|
|
return -1;
|
|
}
|
|
|
|
*ret = vscsi_swap_desc(*ret);
|
|
if (buf_offset > ret->len) {
|
|
trace_spapr_vscsi_fetch_desc_out_of_desc_boundary(buf_offset,
|
|
req->cur_desc_num,
|
|
ret->len);
|
|
return -1;
|
|
}
|
|
ret->va += buf_offset;
|
|
ret->len -= buf_offset;
|
|
|
|
trace_spapr_vscsi_fetch_desc_done(req->cur_desc_num, req->cur_desc_offset,
|
|
ret->va, ret->len);
|
|
|
|
return ret->len ? 1 : 0;
|
|
}
|
|
|
|
static int vscsi_srp_direct_data(VSCSIState *s, vscsi_req *req,
|
|
uint8_t *buf, uint32_t len)
|
|
{
|
|
struct srp_direct_buf md;
|
|
uint32_t llen;
|
|
int rc = 0;
|
|
|
|
rc = vscsi_fetch_desc(s, req, req->cur_desc_num, req->cur_desc_offset, &md);
|
|
if (rc < 0) {
|
|
return -1;
|
|
} else if (rc == 0) {
|
|
return 0;
|
|
}
|
|
|
|
llen = MIN(len, md.len);
|
|
if (llen) {
|
|
if (req->writing) { /* writing = to device = reading from memory */
|
|
rc = spapr_vio_dma_read(&s->vdev, md.va, buf, llen);
|
|
} else {
|
|
rc = spapr_vio_dma_write(&s->vdev, md.va, buf, llen);
|
|
}
|
|
}
|
|
|
|
if (rc) {
|
|
return -1;
|
|
}
|
|
req->cur_desc_offset += llen;
|
|
|
|
return llen;
|
|
}
|
|
|
|
static int vscsi_srp_indirect_data(VSCSIState *s, vscsi_req *req,
|
|
uint8_t *buf, uint32_t len)
|
|
{
|
|
struct srp_direct_buf md;
|
|
int rc = 0;
|
|
uint32_t llen, total = 0;
|
|
|
|
trace_spapr_vscsi_srp_indirect_data(len);
|
|
|
|
/* While we have data ... */
|
|
while (len) {
|
|
rc = vscsi_fetch_desc(s, req, req->cur_desc_num, req->cur_desc_offset, &md);
|
|
if (rc < 0) {
|
|
return -1;
|
|
} else if (rc == 0) {
|
|
break;
|
|
}
|
|
|
|
/* Perform transfer */
|
|
llen = MIN(len, md.len);
|
|
if (req->writing) { /* writing = to device = reading from memory */
|
|
rc = spapr_vio_dma_read(&s->vdev, md.va, buf, llen);
|
|
} else {
|
|
rc = spapr_vio_dma_write(&s->vdev, md.va, buf, llen);
|
|
}
|
|
if (rc) {
|
|
trace_spapr_vscsi_srp_indirect_data_rw(req->writing, rc);
|
|
break;
|
|
}
|
|
trace_spapr_vscsi_srp_indirect_data_buf(buf[0], buf[1], buf[2], buf[3]);
|
|
|
|
len -= llen;
|
|
buf += llen;
|
|
|
|
total += llen;
|
|
|
|
/* Update current position in the current descriptor */
|
|
req->cur_desc_offset += llen;
|
|
if (md.len == llen) {
|
|
/* Go to the next descriptor if the current one finished */
|
|
++req->cur_desc_num;
|
|
req->cur_desc_offset = 0;
|
|
}
|
|
}
|
|
|
|
return rc ? -1 : total;
|
|
}
|
|
|
|
static int vscsi_srp_transfer_data(VSCSIState *s, vscsi_req *req,
|
|
int writing, uint8_t *buf, uint32_t len)
|
|
{
|
|
int err = 0;
|
|
|
|
switch (req->dma_fmt) {
|
|
case SRP_NO_DATA_DESC:
|
|
trace_spapr_vscsi_srp_transfer_data(len);
|
|
break;
|
|
case SRP_DATA_DESC_DIRECT:
|
|
err = vscsi_srp_direct_data(s, req, buf, len);
|
|
break;
|
|
case SRP_DATA_DESC_INDIRECT:
|
|
err = vscsi_srp_indirect_data(s, req, buf, len);
|
|
break;
|
|
}
|
|
return err;
|
|
}
|
|
|
|
/* Bits from linux srp */
|
|
static int data_out_desc_size(struct srp_cmd *cmd)
|
|
{
|
|
int size = 0;
|
|
uint8_t fmt = cmd->buf_fmt >> 4;
|
|
|
|
switch (fmt) {
|
|
case SRP_NO_DATA_DESC:
|
|
break;
|
|
case SRP_DATA_DESC_DIRECT:
|
|
size = sizeof(struct srp_direct_buf);
|
|
break;
|
|
case SRP_DATA_DESC_INDIRECT:
|
|
size = sizeof(struct srp_indirect_buf) +
|
|
sizeof(struct srp_direct_buf)*cmd->data_out_desc_cnt;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
return size;
|
|
}
|
|
|
|
static int vscsi_preprocess_desc(vscsi_req *req)
|
|
{
|
|
struct srp_cmd *cmd = &req_iu(req)->srp.cmd;
|
|
|
|
req->cdb_offset = cmd->add_cdb_len & ~3;
|
|
|
|
if (req->writing) {
|
|
req->dma_fmt = cmd->buf_fmt >> 4;
|
|
} else {
|
|
req->cdb_offset += data_out_desc_size(cmd);
|
|
req->dma_fmt = cmd->buf_fmt & ((1U << 4) - 1);
|
|
}
|
|
|
|
switch (req->dma_fmt) {
|
|
case SRP_NO_DATA_DESC:
|
|
break;
|
|
case SRP_DATA_DESC_DIRECT:
|
|
req->total_desc = req->local_desc = 1;
|
|
break;
|
|
case SRP_DATA_DESC_INDIRECT: {
|
|
struct srp_indirect_buf *ind_tmp = (struct srp_indirect_buf *)
|
|
(cmd->add_data + req->cdb_offset);
|
|
|
|
req->total_desc = be32_to_cpu(ind_tmp->table_desc.len) /
|
|
sizeof(struct srp_direct_buf);
|
|
req->local_desc = req->writing ? cmd->data_out_desc_cnt :
|
|
cmd->data_in_desc_cnt;
|
|
break;
|
|
}
|
|
default:
|
|
fprintf(stderr,
|
|
"vscsi_preprocess_desc: Unknown format %x\n", req->dma_fmt);
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Callback to indicate that the SCSI layer has completed a transfer. */
|
|
static void vscsi_transfer_data(SCSIRequest *sreq, uint32_t len)
|
|
{
|
|
VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(sreq->bus->qbus.parent);
|
|
vscsi_req *req = sreq->hba_private;
|
|
uint8_t *buf;
|
|
int rc = 0;
|
|
|
|
trace_spapr_vscsi_transfer_data(sreq->tag, len, req);
|
|
if (req == NULL) {
|
|
fprintf(stderr, "VSCSI: Can't find request for tag 0x%x\n", sreq->tag);
|
|
return;
|
|
}
|
|
|
|
if (len) {
|
|
buf = scsi_req_get_buf(sreq);
|
|
rc = vscsi_srp_transfer_data(s, req, req->writing, buf, len);
|
|
}
|
|
if (rc < 0) {
|
|
fprintf(stderr, "VSCSI: RDMA error rc=%d!\n", rc);
|
|
req->dma_error = true;
|
|
scsi_req_cancel(req->sreq);
|
|
return;
|
|
}
|
|
|
|
/* Start next chunk */
|
|
req->data_len -= rc;
|
|
scsi_req_continue(sreq);
|
|
}
|
|
|
|
/* Callback to indicate that the SCSI layer has completed a transfer. */
|
|
static void vscsi_command_complete(SCSIRequest *sreq, size_t resid)
|
|
{
|
|
VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(sreq->bus->qbus.parent);
|
|
vscsi_req *req = sreq->hba_private;
|
|
int32_t res_in = 0, res_out = 0;
|
|
|
|
trace_spapr_vscsi_command_complete(sreq->tag, sreq->status, req);
|
|
if (req == NULL) {
|
|
fprintf(stderr, "VSCSI: Can't find request for tag 0x%x\n", sreq->tag);
|
|
return;
|
|
}
|
|
|
|
if (sreq->status == CHECK_CONDITION) {
|
|
req->senselen = scsi_req_get_sense(req->sreq, req->sense,
|
|
sizeof(req->sense));
|
|
trace_spapr_vscsi_command_complete_sense_data1(req->senselen,
|
|
req->sense[0], req->sense[1], req->sense[2], req->sense[3],
|
|
req->sense[4], req->sense[5], req->sense[6], req->sense[7]);
|
|
trace_spapr_vscsi_command_complete_sense_data2(
|
|
req->sense[8], req->sense[9], req->sense[10], req->sense[11],
|
|
req->sense[12], req->sense[13], req->sense[14], req->sense[15]);
|
|
}
|
|
|
|
trace_spapr_vscsi_command_complete_status(sreq->status);
|
|
if (sreq->status == 0) {
|
|
/* We handle overflows, not underflows for normal commands,
|
|
* but hopefully nobody cares
|
|
*/
|
|
if (req->writing) {
|
|
res_out = req->data_len;
|
|
} else {
|
|
res_in = req->data_len;
|
|
}
|
|
}
|
|
vscsi_send_rsp(s, req, sreq->status, res_in, res_out);
|
|
vscsi_put_req(req);
|
|
}
|
|
|
|
static void vscsi_request_cancelled(SCSIRequest *sreq)
|
|
{
|
|
vscsi_req *req = sreq->hba_private;
|
|
|
|
if (req->dma_error) {
|
|
VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(sreq->bus->qbus.parent);
|
|
|
|
vscsi_makeup_sense(s, req, HARDWARE_ERROR, 0, 0);
|
|
vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0);
|
|
}
|
|
vscsi_put_req(req);
|
|
}
|
|
|
|
static const VMStateDescription vmstate_spapr_vscsi_req = {
|
|
.name = "spapr_vscsi_req",
|
|
.version_id = 1,
|
|
.minimum_version_id = 1,
|
|
.fields = (VMStateField[]) {
|
|
VMSTATE_BUFFER(crq.raw, vscsi_req),
|
|
VMSTATE_BUFFER(viosrp_iu_buf, vscsi_req),
|
|
VMSTATE_UINT32(qtag, vscsi_req),
|
|
VMSTATE_BOOL(active, vscsi_req),
|
|
VMSTATE_UINT32(data_len, vscsi_req),
|
|
VMSTATE_BOOL(writing, vscsi_req),
|
|
VMSTATE_UINT32(senselen, vscsi_req),
|
|
VMSTATE_BUFFER(sense, vscsi_req),
|
|
VMSTATE_UINT8(dma_fmt, vscsi_req),
|
|
VMSTATE_UINT16(local_desc, vscsi_req),
|
|
VMSTATE_UINT16(total_desc, vscsi_req),
|
|
VMSTATE_UINT16(cdb_offset, vscsi_req),
|
|
/*Restart SCSI request from the beginning for now */
|
|
/*VMSTATE_UINT16(cur_desc_num, vscsi_req),
|
|
VMSTATE_UINT16(cur_desc_offset, vscsi_req),*/
|
|
VMSTATE_END_OF_LIST()
|
|
},
|
|
};
|
|
|
|
static void vscsi_save_request(QEMUFile *f, SCSIRequest *sreq)
|
|
{
|
|
vscsi_req *req = sreq->hba_private;
|
|
assert(req->active);
|
|
|
|
vmstate_save_state(f, &vmstate_spapr_vscsi_req, req, NULL);
|
|
|
|
trace_spapr_vscsi_save_request(req->qtag, req->cur_desc_num,
|
|
req->cur_desc_offset);
|
|
}
|
|
|
|
static void *vscsi_load_request(QEMUFile *f, SCSIRequest *sreq)
|
|
{
|
|
SCSIBus *bus = sreq->bus;
|
|
VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(bus->qbus.parent);
|
|
vscsi_req *req;
|
|
int rc;
|
|
|
|
assert(sreq->tag < VSCSI_REQ_LIMIT);
|
|
req = &s->reqs[sreq->tag];
|
|
assert(!req->active);
|
|
|
|
memset(req, 0, sizeof(*req));
|
|
rc = vmstate_load_state(f, &vmstate_spapr_vscsi_req, req, 1);
|
|
if (rc) {
|
|
fprintf(stderr, "VSCSI: failed loading request tag#%u\n", sreq->tag);
|
|
return NULL;
|
|
}
|
|
assert(req->active);
|
|
|
|
req->sreq = scsi_req_ref(sreq);
|
|
|
|
trace_spapr_vscsi_load_request(req->qtag, req->cur_desc_num,
|
|
req->cur_desc_offset);
|
|
|
|
return req;
|
|
}
|
|
|
|
static void vscsi_process_login(VSCSIState *s, vscsi_req *req)
|
|
{
|
|
union viosrp_iu *iu = req_iu(req);
|
|
struct srp_login_rsp *rsp = &iu->srp.login_rsp;
|
|
uint64_t tag = iu->srp.rsp.tag;
|
|
|
|
trace_spapr_vscsi_process_login();
|
|
|
|
/* TODO handle case that requested size is wrong and
|
|
* buffer format is wrong
|
|
*/
|
|
memset(iu, 0, sizeof(struct srp_login_rsp));
|
|
rsp->opcode = SRP_LOGIN_RSP;
|
|
/* Don't advertise quite as many request as we support to
|
|
* keep room for management stuff etc...
|
|
*/
|
|
rsp->req_lim_delta = cpu_to_be32(VSCSI_REQ_LIMIT-2);
|
|
rsp->tag = tag;
|
|
rsp->max_it_iu_len = cpu_to_be32(SRP_MAX_IU_LEN);
|
|
rsp->max_ti_iu_len = cpu_to_be32(SRP_MAX_IU_LEN);
|
|
/* direct and indirect */
|
|
rsp->buf_fmt = cpu_to_be16(SRP_BUF_FORMAT_DIRECT | SRP_BUF_FORMAT_INDIRECT);
|
|
|
|
vscsi_send_iu(s, req, sizeof(*rsp), VIOSRP_SRP_FORMAT);
|
|
}
|
|
|
|
static void vscsi_inquiry_no_target(VSCSIState *s, vscsi_req *req)
|
|
{
|
|
uint8_t *cdb = req_iu(req)->srp.cmd.cdb;
|
|
uint8_t resp_data[36];
|
|
int rc, len, alen;
|
|
|
|
/* We don't do EVPD. Also check that page_code is 0 */
|
|
if ((cdb[1] & 0x01) || cdb[2] != 0) {
|
|
/* Send INVALID FIELD IN CDB */
|
|
vscsi_makeup_sense(s, req, ILLEGAL_REQUEST, 0x24, 0);
|
|
vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0);
|
|
return;
|
|
}
|
|
alen = cdb[3];
|
|
alen = (alen << 8) | cdb[4];
|
|
len = MIN(alen, 36);
|
|
|
|
/* Fake up inquiry using PQ=3 */
|
|
memset(resp_data, 0, 36);
|
|
resp_data[0] = 0x7f; /* Not capable of supporting a device here */
|
|
resp_data[2] = 0x06; /* SPS-4 */
|
|
resp_data[3] = 0x02; /* Resp data format */
|
|
resp_data[4] = 36 - 5; /* Additional length */
|
|
resp_data[7] = 0x10; /* Sync transfers */
|
|
memcpy(&resp_data[16], "QEMU EMPTY ", 16);
|
|
memcpy(&resp_data[8], "QEMU ", 8);
|
|
|
|
req->writing = 0;
|
|
vscsi_preprocess_desc(req);
|
|
rc = vscsi_srp_transfer_data(s, req, 0, resp_data, len);
|
|
if (rc < 0) {
|
|
vscsi_makeup_sense(s, req, HARDWARE_ERROR, 0, 0);
|
|
vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0);
|
|
} else {
|
|
vscsi_send_rsp(s, req, 0, 36 - rc, 0);
|
|
}
|
|
}
|
|
|
|
static void vscsi_report_luns(VSCSIState *s, vscsi_req *req)
|
|
{
|
|
BusChild *kid;
|
|
int i, len, n, rc;
|
|
uint8_t *resp_data;
|
|
bool found_lun0;
|
|
|
|
n = 0;
|
|
found_lun0 = false;
|
|
QTAILQ_FOREACH(kid, &s->bus.qbus.children, sibling) {
|
|
SCSIDevice *dev = SCSI_DEVICE(kid->child);
|
|
|
|
n += 8;
|
|
if (dev->channel == 0 && dev->id == 0 && dev->lun == 0) {
|
|
found_lun0 = true;
|
|
}
|
|
}
|
|
if (!found_lun0) {
|
|
n += 8;
|
|
}
|
|
len = n+8;
|
|
|
|
resp_data = g_malloc0(len);
|
|
stl_be_p(resp_data, n);
|
|
i = found_lun0 ? 8 : 16;
|
|
QTAILQ_FOREACH(kid, &s->bus.qbus.children, sibling) {
|
|
DeviceState *qdev = kid->child;
|
|
SCSIDevice *dev = SCSI_DEVICE(qdev);
|
|
|
|
if (dev->id == 0 && dev->channel == 0) {
|
|
resp_data[i] = 0; /* Use simple LUN for 0 (SAM5 4.7.7.1) */
|
|
} else {
|
|
resp_data[i] = (2 << 6); /* Otherwise LUN addressing (4.7.7.4) */
|
|
}
|
|
resp_data[i] |= dev->id;
|
|
resp_data[i+1] = (dev->channel << 5);
|
|
resp_data[i+1] |= dev->lun;
|
|
i += 8;
|
|
}
|
|
|
|
vscsi_preprocess_desc(req);
|
|
rc = vscsi_srp_transfer_data(s, req, 0, resp_data, len);
|
|
g_free(resp_data);
|
|
if (rc < 0) {
|
|
vscsi_makeup_sense(s, req, HARDWARE_ERROR, 0, 0);
|
|
vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0);
|
|
} else {
|
|
vscsi_send_rsp(s, req, 0, len - rc, 0);
|
|
}
|
|
}
|
|
|
|
static int vscsi_queue_cmd(VSCSIState *s, vscsi_req *req)
|
|
{
|
|
union srp_iu *srp = &req_iu(req)->srp;
|
|
SCSIDevice *sdev;
|
|
int n, lun;
|
|
size_t cdb_len = sizeof (srp->cmd.cdb) + (srp->cmd.add_cdb_len & ~3);
|
|
|
|
if ((srp->cmd.lun == 0 || be64_to_cpu(srp->cmd.lun) == SRP_REPORT_LUNS_WLUN)
|
|
&& srp->cmd.cdb[0] == REPORT_LUNS) {
|
|
vscsi_report_luns(s, req);
|
|
return 0;
|
|
}
|
|
|
|
sdev = vscsi_device_find(&s->bus, be64_to_cpu(srp->cmd.lun), &lun);
|
|
if (!sdev) {
|
|
trace_spapr_vscsi_queue_cmd_no_drive(be64_to_cpu(srp->cmd.lun));
|
|
if (srp->cmd.cdb[0] == INQUIRY) {
|
|
vscsi_inquiry_no_target(s, req);
|
|
} else {
|
|
vscsi_makeup_sense(s, req, ILLEGAL_REQUEST, 0x24, 0x00);
|
|
vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0);
|
|
} return 1;
|
|
}
|
|
|
|
req->sreq = scsi_req_new(sdev, req->qtag, lun, srp->cmd.cdb, cdb_len, req);
|
|
n = scsi_req_enqueue(req->sreq);
|
|
|
|
trace_spapr_vscsi_queue_cmd(req->qtag, srp->cmd.cdb[0],
|
|
scsi_command_name(srp->cmd.cdb[0]), lun, n);
|
|
|
|
if (n) {
|
|
/* Transfer direction must be set before preprocessing the
|
|
* descriptors
|
|
*/
|
|
req->writing = (n < 1);
|
|
|
|
/* Preprocess RDMA descriptors */
|
|
vscsi_preprocess_desc(req);
|
|
|
|
/* Get transfer direction and initiate transfer */
|
|
if (n > 0) {
|
|
req->data_len = n;
|
|
} else if (n < 0) {
|
|
req->data_len = -n;
|
|
}
|
|
scsi_req_continue(req->sreq);
|
|
}
|
|
/* Don't touch req here, it may have been recycled already */
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int vscsi_process_tsk_mgmt(VSCSIState *s, vscsi_req *req)
|
|
{
|
|
union viosrp_iu *iu = req_iu(req);
|
|
vscsi_req *tmpreq;
|
|
int i, lun = 0, resp = SRP_TSK_MGMT_COMPLETE;
|
|
SCSIDevice *d;
|
|
uint64_t tag = iu->srp.rsp.tag;
|
|
uint8_t sol_not = iu->srp.cmd.sol_not;
|
|
|
|
trace_spapr_vscsi_process_tsk_mgmt(iu->srp.tsk_mgmt.tsk_mgmt_func);
|
|
d = vscsi_device_find(&s->bus,
|
|
be64_to_cpu(req_iu(req)->srp.tsk_mgmt.lun), &lun);
|
|
if (!d) {
|
|
resp = SRP_TSK_MGMT_FIELDS_INVALID;
|
|
} else {
|
|
switch (iu->srp.tsk_mgmt.tsk_mgmt_func) {
|
|
case SRP_TSK_ABORT_TASK:
|
|
if (d->lun != lun) {
|
|
resp = SRP_TSK_MGMT_FIELDS_INVALID;
|
|
break;
|
|
}
|
|
|
|
tmpreq = vscsi_find_req(s, req_iu(req)->srp.tsk_mgmt.task_tag);
|
|
if (tmpreq && tmpreq->sreq) {
|
|
assert(tmpreq->sreq->hba_private);
|
|
scsi_req_cancel(tmpreq->sreq);
|
|
}
|
|
break;
|
|
|
|
case SRP_TSK_LUN_RESET:
|
|
if (d->lun != lun) {
|
|
resp = SRP_TSK_MGMT_FIELDS_INVALID;
|
|
break;
|
|
}
|
|
|
|
qdev_reset_all(&d->qdev);
|
|
break;
|
|
|
|
case SRP_TSK_ABORT_TASK_SET:
|
|
case SRP_TSK_CLEAR_TASK_SET:
|
|
if (d->lun != lun) {
|
|
resp = SRP_TSK_MGMT_FIELDS_INVALID;
|
|
break;
|
|
}
|
|
|
|
for (i = 0; i < VSCSI_REQ_LIMIT; i++) {
|
|
tmpreq = &s->reqs[i];
|
|
if (req_iu(tmpreq)->srp.cmd.lun
|
|
!= req_iu(req)->srp.tsk_mgmt.lun) {
|
|
continue;
|
|
}
|
|
if (!tmpreq->active || !tmpreq->sreq) {
|
|
continue;
|
|
}
|
|
assert(tmpreq->sreq->hba_private);
|
|
scsi_req_cancel(tmpreq->sreq);
|
|
}
|
|
break;
|
|
|
|
case SRP_TSK_CLEAR_ACA:
|
|
resp = SRP_TSK_MGMT_NOT_SUPPORTED;
|
|
break;
|
|
|
|
default:
|
|
resp = SRP_TSK_MGMT_FIELDS_INVALID;
|
|
break;
|
|
}
|
|
}
|
|
|
|
/* Compose the response here as */
|
|
QEMU_BUILD_BUG_ON(SRP_MAX_IU_DATA_LEN < 4);
|
|
memset(iu, 0, sizeof(struct srp_rsp) + 4);
|
|
iu->srp.rsp.opcode = SRP_RSP;
|
|
iu->srp.rsp.req_lim_delta = cpu_to_be32(1);
|
|
iu->srp.rsp.tag = tag;
|
|
iu->srp.rsp.flags |= SRP_RSP_FLAG_RSPVALID;
|
|
iu->srp.rsp.resp_data_len = cpu_to_be32(4);
|
|
if (resp) {
|
|
iu->srp.rsp.sol_not = (sol_not & 0x04) >> 2;
|
|
} else {
|
|
iu->srp.rsp.sol_not = (sol_not & 0x02) >> 1;
|
|
}
|
|
|
|
iu->srp.rsp.status = GOOD;
|
|
iu->srp.rsp.data[3] = resp;
|
|
|
|
vscsi_send_iu(s, req, sizeof(iu->srp.rsp) + 4, VIOSRP_SRP_FORMAT);
|
|
|
|
return 1;
|
|
}
|
|
|
|
static int vscsi_handle_srp_req(VSCSIState *s, vscsi_req *req)
|
|
{
|
|
union srp_iu *srp = &req_iu(req)->srp;
|
|
int done = 1;
|
|
uint8_t opcode = srp->rsp.opcode;
|
|
|
|
switch (opcode) {
|
|
case SRP_LOGIN_REQ:
|
|
vscsi_process_login(s, req);
|
|
break;
|
|
case SRP_TSK_MGMT:
|
|
done = vscsi_process_tsk_mgmt(s, req);
|
|
break;
|
|
case SRP_CMD:
|
|
done = vscsi_queue_cmd(s, req);
|
|
break;
|
|
case SRP_LOGIN_RSP:
|
|
case SRP_I_LOGOUT:
|
|
case SRP_T_LOGOUT:
|
|
case SRP_RSP:
|
|
case SRP_CRED_REQ:
|
|
case SRP_CRED_RSP:
|
|
case SRP_AER_REQ:
|
|
case SRP_AER_RSP:
|
|
fprintf(stderr, "VSCSI: Unsupported opcode %02x\n", opcode);
|
|
break;
|
|
default:
|
|
fprintf(stderr, "VSCSI: Unknown type %02x\n", opcode);
|
|
}
|
|
|
|
return done;
|
|
}
|
|
|
|
static int vscsi_send_adapter_info(VSCSIState *s, vscsi_req *req)
|
|
{
|
|
struct viosrp_adapter_info *sinfo;
|
|
struct mad_adapter_info_data info;
|
|
int rc;
|
|
|
|
sinfo = &req_iu(req)->mad.adapter_info;
|
|
|
|
#if 0 /* What for ? */
|
|
rc = spapr_vio_dma_read(&s->vdev, be64_to_cpu(sinfo->buffer),
|
|
&info, be16_to_cpu(sinfo->common.length));
|
|
if (rc) {
|
|
fprintf(stderr, "vscsi_send_adapter_info: DMA read failure !\n");
|
|
}
|
|
#endif
|
|
memset(&info, 0, sizeof(info));
|
|
strcpy(info.srp_version, SRP_VERSION);
|
|
memcpy(info.partition_name, "qemu", sizeof("qemu"));
|
|
info.partition_number = cpu_to_be32(0);
|
|
info.mad_version = cpu_to_be32(1);
|
|
info.os_type = cpu_to_be32(2);
|
|
info.port_max_txu[0] = cpu_to_be32(VSCSI_MAX_SECTORS << 9);
|
|
|
|
rc = spapr_vio_dma_write(&s->vdev, be64_to_cpu(sinfo->buffer),
|
|
&info, be16_to_cpu(sinfo->common.length));
|
|
if (rc) {
|
|
fprintf(stderr, "vscsi_send_adapter_info: DMA write failure !\n");
|
|
}
|
|
|
|
sinfo->common.status = rc ? cpu_to_be32(1) : 0;
|
|
|
|
return vscsi_send_iu(s, req, sizeof(*sinfo), VIOSRP_MAD_FORMAT);
|
|
}
|
|
|
|
static int vscsi_send_capabilities(VSCSIState *s, vscsi_req *req)
|
|
{
|
|
struct viosrp_capabilities *vcap;
|
|
struct capabilities cap = { };
|
|
uint16_t len, req_len;
|
|
uint64_t buffer;
|
|
int rc;
|
|
|
|
vcap = &req_iu(req)->mad.capabilities;
|
|
req_len = len = be16_to_cpu(vcap->common.length);
|
|
buffer = be64_to_cpu(vcap->buffer);
|
|
if (len > sizeof(cap)) {
|
|
fprintf(stderr, "vscsi_send_capabilities: capabilities size mismatch !\n");
|
|
|
|
/*
|
|
* Just read and populate the structure that is known.
|
|
* Zero rest of the structure.
|
|
*/
|
|
len = sizeof(cap);
|
|
}
|
|
rc = spapr_vio_dma_read(&s->vdev, buffer, &cap, len);
|
|
if (rc) {
|
|
fprintf(stderr, "vscsi_send_capabilities: DMA read failure !\n");
|
|
}
|
|
|
|
/*
|
|
* Current implementation does not support any migration or
|
|
* reservation capabilities. Construct the response telling the
|
|
* guest not to use them.
|
|
*/
|
|
cap.flags = 0;
|
|
cap.migration.ecl = 0;
|
|
cap.reserve.type = 0;
|
|
cap.migration.common.server_support = 0;
|
|
cap.reserve.common.server_support = 0;
|
|
|
|
rc = spapr_vio_dma_write(&s->vdev, buffer, &cap, len);
|
|
if (rc) {
|
|
fprintf(stderr, "vscsi_send_capabilities: DMA write failure !\n");
|
|
}
|
|
if (req_len > len) {
|
|
/*
|
|
* Being paranoid and lets not worry about the error code
|
|
* here. Actual write of the cap is done above.
|
|
*/
|
|
spapr_vio_dma_set(&s->vdev, (buffer + len), 0, (req_len - len));
|
|
}
|
|
vcap->common.status = rc ? cpu_to_be32(1) : 0;
|
|
return vscsi_send_iu(s, req, sizeof(*vcap), VIOSRP_MAD_FORMAT);
|
|
}
|
|
|
|
static int vscsi_handle_mad_req(VSCSIState *s, vscsi_req *req)
|
|
{
|
|
union mad_iu *mad = &req_iu(req)->mad;
|
|
bool request_handled = false;
|
|
uint64_t retlen = 0;
|
|
|
|
switch (be32_to_cpu(mad->empty_iu.common.type)) {
|
|
case VIOSRP_EMPTY_IU_TYPE:
|
|
fprintf(stderr, "Unsupported EMPTY MAD IU\n");
|
|
retlen = sizeof(mad->empty_iu);
|
|
break;
|
|
case VIOSRP_ERROR_LOG_TYPE:
|
|
fprintf(stderr, "Unsupported ERROR LOG MAD IU\n");
|
|
retlen = sizeof(mad->error_log);
|
|
break;
|
|
case VIOSRP_ADAPTER_INFO_TYPE:
|
|
vscsi_send_adapter_info(s, req);
|
|
request_handled = true;
|
|
break;
|
|
case VIOSRP_HOST_CONFIG_TYPE:
|
|
retlen = sizeof(mad->host_config);
|
|
break;
|
|
case VIOSRP_CAPABILITIES_TYPE:
|
|
vscsi_send_capabilities(s, req);
|
|
request_handled = true;
|
|
break;
|
|
default:
|
|
fprintf(stderr, "VSCSI: Unknown MAD type %02x\n",
|
|
be32_to_cpu(mad->empty_iu.common.type));
|
|
/*
|
|
* PAPR+ says that "The length field is set to the length
|
|
* of the data structure(s) used in the command".
|
|
* As we did not recognize the request type, put zero there.
|
|
*/
|
|
retlen = 0;
|
|
}
|
|
|
|
if (!request_handled) {
|
|
mad->empty_iu.common.status = cpu_to_be16(VIOSRP_MAD_NOT_SUPPORTED);
|
|
vscsi_send_iu(s, req, retlen, VIOSRP_MAD_FORMAT);
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
static void vscsi_got_payload(VSCSIState *s, vscsi_crq *crq)
|
|
{
|
|
vscsi_req *req;
|
|
int done;
|
|
|
|
req = vscsi_get_req(s);
|
|
if (req == NULL) {
|
|
fprintf(stderr, "VSCSI: Failed to get a request !\n");
|
|
return;
|
|
}
|
|
|
|
/* We only support a limited number of descriptors, we know
|
|
* the ibmvscsi driver uses up to 10 max, so it should fit
|
|
* in our 256 bytes IUs. If not we'll have to increase the size
|
|
* of the structure.
|
|
*/
|
|
if (crq->s.IU_length > SRP_MAX_IU_LEN) {
|
|
fprintf(stderr, "VSCSI: SRP IU too long (%d bytes) !\n",
|
|
crq->s.IU_length);
|
|
vscsi_put_req(req);
|
|
return;
|
|
}
|
|
|
|
/* XXX Handle failure differently ? */
|
|
if (spapr_vio_dma_read(&s->vdev, crq->s.IU_data_ptr, &req->viosrp_iu_buf,
|
|
crq->s.IU_length)) {
|
|
fprintf(stderr, "vscsi_got_payload: DMA read failure !\n");
|
|
vscsi_put_req(req);
|
|
return;
|
|
}
|
|
memcpy(&req->crq, crq, sizeof(vscsi_crq));
|
|
|
|
if (crq->s.format == VIOSRP_MAD_FORMAT) {
|
|
done = vscsi_handle_mad_req(s, req);
|
|
} else {
|
|
done = vscsi_handle_srp_req(s, req);
|
|
}
|
|
|
|
if (done) {
|
|
vscsi_put_req(req);
|
|
}
|
|
}
|
|
|
|
|
|
static int vscsi_do_crq(struct SpaprVioDevice *dev, uint8_t *crq_data)
|
|
{
|
|
VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(dev);
|
|
vscsi_crq crq;
|
|
|
|
memcpy(crq.raw, crq_data, 16);
|
|
crq.s.timeout = be16_to_cpu(crq.s.timeout);
|
|
crq.s.IU_length = be16_to_cpu(crq.s.IU_length);
|
|
crq.s.IU_data_ptr = be64_to_cpu(crq.s.IU_data_ptr);
|
|
|
|
trace_spapr_vscsi_do_crq(crq.raw[0], crq.raw[1]);
|
|
|
|
switch (crq.s.valid) {
|
|
case 0xc0: /* Init command/response */
|
|
|
|
/* Respond to initialization request */
|
|
if (crq.s.format == 0x01) {
|
|
memset(crq.raw, 0, 16);
|
|
crq.s.valid = 0xc0;
|
|
crq.s.format = 0x02;
|
|
spapr_vio_send_crq(dev, crq.raw);
|
|
}
|
|
|
|
/* Note that in hotplug cases, we might get a 0x02
|
|
* as a result of us emitting the init request
|
|
*/
|
|
|
|
break;
|
|
case 0xff: /* Link event */
|
|
|
|
/* Not handled for now */
|
|
|
|
break;
|
|
case 0x80: /* Payloads */
|
|
switch (crq.s.format) {
|
|
case VIOSRP_SRP_FORMAT: /* AKA VSCSI request */
|
|
case VIOSRP_MAD_FORMAT: /* AKA VSCSI response */
|
|
vscsi_got_payload(s, &crq);
|
|
break;
|
|
case VIOSRP_OS400_FORMAT:
|
|
case VIOSRP_AIX_FORMAT:
|
|
case VIOSRP_LINUX_FORMAT:
|
|
case VIOSRP_INLINE_FORMAT:
|
|
fprintf(stderr, "vscsi_do_srq: Unsupported payload format %02x\n",
|
|
crq.s.format);
|
|
break;
|
|
default:
|
|
fprintf(stderr, "vscsi_do_srq: Unknown payload format %02x\n",
|
|
crq.s.format);
|
|
}
|
|
break;
|
|
default:
|
|
fprintf(stderr, "vscsi_do_crq: unknown CRQ %02x %02x ...\n",
|
|
crq.raw[0], crq.raw[1]);
|
|
};
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct SCSIBusInfo vscsi_scsi_info = {
|
|
.tcq = true,
|
|
.max_channel = 7, /* logical unit addressing format */
|
|
.max_target = 63,
|
|
.max_lun = 31,
|
|
|
|
.transfer_data = vscsi_transfer_data,
|
|
.complete = vscsi_command_complete,
|
|
.cancel = vscsi_request_cancelled,
|
|
.save_request = vscsi_save_request,
|
|
.load_request = vscsi_load_request,
|
|
};
|
|
|
|
static void spapr_vscsi_reset(SpaprVioDevice *dev)
|
|
{
|
|
VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(dev);
|
|
int i;
|
|
|
|
memset(s->reqs, 0, sizeof(s->reqs));
|
|
for (i = 0; i < VSCSI_REQ_LIMIT; i++) {
|
|
s->reqs[i].qtag = i;
|
|
}
|
|
}
|
|
|
|
static void spapr_vscsi_realize(SpaprVioDevice *dev, Error **errp)
|
|
{
|
|
VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(dev);
|
|
|
|
dev->crq.SendFunc = vscsi_do_crq;
|
|
|
|
scsi_bus_init(&s->bus, sizeof(s->bus), DEVICE(dev), &vscsi_scsi_info);
|
|
|
|
/* ibmvscsi SCSI bus does not allow hotplug. */
|
|
qbus_set_hotplug_handler(BUS(&s->bus), NULL);
|
|
}
|
|
|
|
void spapr_vscsi_create(SpaprVioBus *bus)
|
|
{
|
|
DeviceState *dev;
|
|
|
|
dev = qdev_new("spapr-vscsi");
|
|
|
|
qdev_realize_and_unref(dev, &bus->bus, &error_fatal);
|
|
scsi_bus_legacy_handle_cmdline(&VIO_SPAPR_VSCSI_DEVICE(dev)->bus);
|
|
}
|
|
|
|
static int spapr_vscsi_devnode(SpaprVioDevice *dev, void *fdt, int node_off)
|
|
{
|
|
int ret;
|
|
|
|
ret = fdt_setprop_cell(fdt, node_off, "#address-cells", 2);
|
|
if (ret < 0) {
|
|
return ret;
|
|
}
|
|
|
|
ret = fdt_setprop_cell(fdt, node_off, "#size-cells", 0);
|
|
if (ret < 0) {
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static Property spapr_vscsi_properties[] = {
|
|
DEFINE_SPAPR_PROPERTIES(VSCSIState, vdev),
|
|
DEFINE_PROP_END_OF_LIST(),
|
|
};
|
|
|
|
static const VMStateDescription vmstate_spapr_vscsi = {
|
|
.name = "spapr_vscsi",
|
|
.version_id = 1,
|
|
.minimum_version_id = 1,
|
|
.fields = (VMStateField[]) {
|
|
VMSTATE_SPAPR_VIO(vdev, VSCSIState),
|
|
/* VSCSI state */
|
|
/* ???? */
|
|
|
|
VMSTATE_END_OF_LIST()
|
|
},
|
|
};
|
|
|
|
static void spapr_vscsi_class_init(ObjectClass *klass, void *data)
|
|
{
|
|
DeviceClass *dc = DEVICE_CLASS(klass);
|
|
SpaprVioDeviceClass *k = VIO_SPAPR_DEVICE_CLASS(klass);
|
|
|
|
k->realize = spapr_vscsi_realize;
|
|
k->reset = spapr_vscsi_reset;
|
|
k->devnode = spapr_vscsi_devnode;
|
|
k->dt_name = "v-scsi";
|
|
k->dt_type = "vscsi";
|
|
k->dt_compatible = "IBM,v-scsi";
|
|
k->signal_mask = 0x00000001;
|
|
set_bit(DEVICE_CATEGORY_STORAGE, dc->categories);
|
|
device_class_set_props(dc, spapr_vscsi_properties);
|
|
k->rtce_window_size = 0x10000000;
|
|
dc->vmsd = &vmstate_spapr_vscsi;
|
|
}
|
|
|
|
static const TypeInfo spapr_vscsi_info = {
|
|
.name = TYPE_VIO_SPAPR_VSCSI_DEVICE,
|
|
.parent = TYPE_VIO_SPAPR_DEVICE,
|
|
.instance_size = sizeof(VSCSIState),
|
|
.class_init = spapr_vscsi_class_init,
|
|
};
|
|
|
|
static void spapr_vscsi_register_types(void)
|
|
{
|
|
type_register_static(&spapr_vscsi_info);
|
|
}
|
|
|
|
type_init(spapr_vscsi_register_types)
|