hw/pvrdma: Fill all CQE fields

Add ability to pass specific WC attributes to CQE such as GRH_BIT flag.

Signed-off-by: Yuval Shaia <yuval.shaia@oracle.com>
Reviewed-by: Marcel Apfelbaum <marcel.apfelbaum@gmail.com>
Signed-off-by: Marcel Apfelbaum <marcel.apfelbaum@gmail.com>
This commit is contained in:
Yuval Shaia 2018-12-21 16:40:30 +02:00 committed by Marcel Apfelbaum
parent e976ebc87c
commit eaac01005d
3 changed files with 58 additions and 36 deletions

View File

@ -60,13 +60,24 @@ struct backend_umad {
char mad[RDMA_MAX_PRIVATE_DATA]; char mad[RDMA_MAX_PRIVATE_DATA];
}; };
static void (*comp_handler)(int status, unsigned int vendor_err, void *ctx); static void (*comp_handler)(void *ctx, struct ibv_wc *wc);
static void dummy_comp_handler(int status, unsigned int vendor_err, void *ctx) static void dummy_comp_handler(void *ctx, struct ibv_wc *wc)
{ {
pr_err("No completion handler is registered\n"); pr_err("No completion handler is registered\n");
} }
static inline void complete_work(enum ibv_wc_status status, uint32_t vendor_err,
void *ctx)
{
struct ibv_wc wc = {0};
wc.status = status;
wc.vendor_err = vendor_err;
comp_handler(ctx, &wc);
}
static void poll_cq(RdmaDeviceResources *rdma_dev_res, struct ibv_cq *ibcq) static void poll_cq(RdmaDeviceResources *rdma_dev_res, struct ibv_cq *ibcq)
{ {
int i, ne; int i, ne;
@ -91,7 +102,7 @@ static void poll_cq(RdmaDeviceResources *rdma_dev_res, struct ibv_cq *ibcq)
} }
pr_dbg("Processing %s CQE\n", bctx->is_tx_req ? "send" : "recv"); pr_dbg("Processing %s CQE\n", bctx->is_tx_req ? "send" : "recv");
comp_handler(wc[i].status, wc[i].vendor_err, bctx->up_ctx); comp_handler(bctx->up_ctx, &wc[i]);
rdma_rm_dealloc_cqe_ctx(rdma_dev_res, wc[i].wr_id); rdma_rm_dealloc_cqe_ctx(rdma_dev_res, wc[i].wr_id);
g_free(bctx); g_free(bctx);
@ -256,8 +267,8 @@ static void start_comp_thread(RdmaBackendDev *backend_dev)
comp_handler_thread, backend_dev, QEMU_THREAD_DETACHED); comp_handler_thread, backend_dev, QEMU_THREAD_DETACHED);
} }
void rdma_backend_register_comp_handler(void (*handler)(int status, void rdma_backend_register_comp_handler(void (*handler)(void *ctx,
unsigned int vendor_err, void *ctx)) struct ibv_wc *wc))
{ {
comp_handler = handler; comp_handler = handler;
} }
@ -451,14 +462,14 @@ void rdma_backend_post_send(RdmaBackendDev *backend_dev,
if (!qp->ibqp) { /* This field does not get initialized for QP0 and QP1 */ if (!qp->ibqp) { /* This field does not get initialized for QP0 and QP1 */
if (qp_type == IBV_QPT_SMI) { if (qp_type == IBV_QPT_SMI) {
pr_dbg("QP0 unsupported\n"); pr_dbg("QP0 unsupported\n");
comp_handler(IBV_WC_GENERAL_ERR, VENDOR_ERR_QP0, ctx); complete_work(IBV_WC_GENERAL_ERR, VENDOR_ERR_QP0, ctx);
} else if (qp_type == IBV_QPT_GSI) { } else if (qp_type == IBV_QPT_GSI) {
pr_dbg("QP1\n"); pr_dbg("QP1\n");
rc = mad_send(backend_dev, sgid_idx, sgid, sge, num_sge); rc = mad_send(backend_dev, sgid_idx, sgid, sge, num_sge);
if (rc) { if (rc) {
comp_handler(IBV_WC_GENERAL_ERR, VENDOR_ERR_MAD_SEND, ctx); complete_work(IBV_WC_GENERAL_ERR, VENDOR_ERR_MAD_SEND, ctx);
} else { } else {
comp_handler(IBV_WC_SUCCESS, 0, ctx); complete_work(IBV_WC_SUCCESS, 0, ctx);
} }
} }
return; return;
@ -467,7 +478,7 @@ void rdma_backend_post_send(RdmaBackendDev *backend_dev,
pr_dbg("num_sge=%d\n", num_sge); pr_dbg("num_sge=%d\n", num_sge);
if (!num_sge) { if (!num_sge) {
pr_dbg("num_sge=0\n"); pr_dbg("num_sge=0\n");
comp_handler(IBV_WC_GENERAL_ERR, VENDOR_ERR_NO_SGE, ctx); complete_work(IBV_WC_GENERAL_ERR, VENDOR_ERR_NO_SGE, ctx);
return; return;
} }
@ -478,21 +489,21 @@ void rdma_backend_post_send(RdmaBackendDev *backend_dev,
rc = rdma_rm_alloc_cqe_ctx(backend_dev->rdma_dev_res, &bctx_id, bctx); rc = rdma_rm_alloc_cqe_ctx(backend_dev->rdma_dev_res, &bctx_id, bctx);
if (unlikely(rc)) { if (unlikely(rc)) {
pr_dbg("Failed to allocate cqe_ctx\n"); pr_dbg("Failed to allocate cqe_ctx\n");
comp_handler(IBV_WC_GENERAL_ERR, VENDOR_ERR_NOMEM, ctx); complete_work(IBV_WC_GENERAL_ERR, VENDOR_ERR_NOMEM, ctx);
goto out_free_bctx; goto out_free_bctx;
} }
rc = build_host_sge_array(backend_dev->rdma_dev_res, new_sge, sge, num_sge); rc = build_host_sge_array(backend_dev->rdma_dev_res, new_sge, sge, num_sge);
if (rc) { if (rc) {
pr_dbg("Error: Failed to build host SGE array\n"); pr_dbg("Error: Failed to build host SGE array\n");
comp_handler(IBV_WC_GENERAL_ERR, rc, ctx); complete_work(IBV_WC_GENERAL_ERR, rc, ctx);
goto out_dealloc_cqe_ctx; goto out_dealloc_cqe_ctx;
} }
if (qp_type == IBV_QPT_UD) { if (qp_type == IBV_QPT_UD) {
wr.wr.ud.ah = create_ah(backend_dev, qp->ibpd, sgid_idx, dgid); wr.wr.ud.ah = create_ah(backend_dev, qp->ibpd, sgid_idx, dgid);
if (!wr.wr.ud.ah) { if (!wr.wr.ud.ah) {
comp_handler(IBV_WC_GENERAL_ERR, VENDOR_ERR_FAIL_BACKEND, ctx); complete_work(IBV_WC_GENERAL_ERR, VENDOR_ERR_FAIL_BACKEND, ctx);
goto out_dealloc_cqe_ctx; goto out_dealloc_cqe_ctx;
} }
wr.wr.ud.remote_qpn = dqpn; wr.wr.ud.remote_qpn = dqpn;
@ -510,7 +521,7 @@ void rdma_backend_post_send(RdmaBackendDev *backend_dev,
if (rc) { if (rc) {
pr_dbg("Fail (%d, %d) to post send WQE to qpn %d\n", rc, errno, pr_dbg("Fail (%d, %d) to post send WQE to qpn %d\n", rc, errno,
qp->ibqp->qp_num); qp->ibqp->qp_num);
comp_handler(IBV_WC_GENERAL_ERR, VENDOR_ERR_FAIL_BACKEND, ctx); complete_work(IBV_WC_GENERAL_ERR, VENDOR_ERR_FAIL_BACKEND, ctx);
goto out_dealloc_cqe_ctx; goto out_dealloc_cqe_ctx;
} }
@ -579,13 +590,13 @@ void rdma_backend_post_recv(RdmaBackendDev *backend_dev,
if (!qp->ibqp) { /* This field does not get initialized for QP0 and QP1 */ if (!qp->ibqp) { /* This field does not get initialized for QP0 and QP1 */
if (qp_type == IBV_QPT_SMI) { if (qp_type == IBV_QPT_SMI) {
pr_dbg("QP0 unsupported\n"); pr_dbg("QP0 unsupported\n");
comp_handler(IBV_WC_GENERAL_ERR, VENDOR_ERR_QP0, ctx); complete_work(IBV_WC_GENERAL_ERR, VENDOR_ERR_QP0, ctx);
} }
if (qp_type == IBV_QPT_GSI) { if (qp_type == IBV_QPT_GSI) {
pr_dbg("QP1\n"); pr_dbg("QP1\n");
rc = save_mad_recv_buffer(backend_dev, sge, num_sge, ctx); rc = save_mad_recv_buffer(backend_dev, sge, num_sge, ctx);
if (rc) { if (rc) {
comp_handler(IBV_WC_GENERAL_ERR, rc, ctx); complete_work(IBV_WC_GENERAL_ERR, rc, ctx);
} }
} }
return; return;
@ -594,7 +605,7 @@ void rdma_backend_post_recv(RdmaBackendDev *backend_dev,
pr_dbg("num_sge=%d\n", num_sge); pr_dbg("num_sge=%d\n", num_sge);
if (!num_sge) { if (!num_sge) {
pr_dbg("num_sge=0\n"); pr_dbg("num_sge=0\n");
comp_handler(IBV_WC_GENERAL_ERR, VENDOR_ERR_NO_SGE, ctx); complete_work(IBV_WC_GENERAL_ERR, VENDOR_ERR_NO_SGE, ctx);
return; return;
} }
@ -605,14 +616,14 @@ void rdma_backend_post_recv(RdmaBackendDev *backend_dev,
rc = rdma_rm_alloc_cqe_ctx(rdma_dev_res, &bctx_id, bctx); rc = rdma_rm_alloc_cqe_ctx(rdma_dev_res, &bctx_id, bctx);
if (unlikely(rc)) { if (unlikely(rc)) {
pr_dbg("Failed to allocate cqe_ctx\n"); pr_dbg("Failed to allocate cqe_ctx\n");
comp_handler(IBV_WC_GENERAL_ERR, VENDOR_ERR_NOMEM, ctx); complete_work(IBV_WC_GENERAL_ERR, VENDOR_ERR_NOMEM, ctx);
goto out_free_bctx; goto out_free_bctx;
} }
rc = build_host_sge_array(rdma_dev_res, new_sge, sge, num_sge); rc = build_host_sge_array(rdma_dev_res, new_sge, sge, num_sge);
if (rc) { if (rc) {
pr_dbg("Error: Failed to build host SGE array\n"); pr_dbg("Error: Failed to build host SGE array\n");
comp_handler(IBV_WC_GENERAL_ERR, rc, ctx); complete_work(IBV_WC_GENERAL_ERR, rc, ctx);
goto out_dealloc_cqe_ctx; goto out_dealloc_cqe_ctx;
} }
@ -624,7 +635,7 @@ void rdma_backend_post_recv(RdmaBackendDev *backend_dev,
if (rc) { if (rc) {
pr_dbg("Fail (%d, %d) to post recv WQE to qpn %d\n", rc, errno, pr_dbg("Fail (%d, %d) to post recv WQE to qpn %d\n", rc, errno,
qp->ibqp->qp_num); qp->ibqp->qp_num);
comp_handler(IBV_WC_GENERAL_ERR, VENDOR_ERR_FAIL_BACKEND, ctx); complete_work(IBV_WC_GENERAL_ERR, VENDOR_ERR_FAIL_BACKEND, ctx);
goto out_dealloc_cqe_ctx; goto out_dealloc_cqe_ctx;
} }
@ -998,9 +1009,10 @@ static void process_incoming_mad_req(RdmaBackendDev *backend_dev,
mad = rdma_pci_dma_map(backend_dev->dev, bctx->sge.addr, mad = rdma_pci_dma_map(backend_dev->dev, bctx->sge.addr,
bctx->sge.length); bctx->sge.length);
if (!mad || bctx->sge.length < msg->umad_len + MAD_HDR_SIZE) { if (!mad || bctx->sge.length < msg->umad_len + MAD_HDR_SIZE) {
comp_handler(IBV_WC_GENERAL_ERR, VENDOR_ERR_INV_MAD_BUFF, complete_work(IBV_WC_GENERAL_ERR, VENDOR_ERR_INV_MAD_BUFF,
bctx->up_ctx); bctx->up_ctx);
} else { } else {
struct ibv_wc wc = {0};
pr_dbg_buf("mad", msg->umad.mad, msg->umad_len); pr_dbg_buf("mad", msg->umad.mad, msg->umad_len);
memset(mad, 0, bctx->sge.length); memset(mad, 0, bctx->sge.length);
build_mad_hdr((struct ibv_grh *)mad, build_mad_hdr((struct ibv_grh *)mad,
@ -1009,7 +1021,10 @@ static void process_incoming_mad_req(RdmaBackendDev *backend_dev,
memcpy(&mad[MAD_HDR_SIZE], msg->umad.mad, msg->umad_len); memcpy(&mad[MAD_HDR_SIZE], msg->umad.mad, msg->umad_len);
rdma_pci_dma_unmap(backend_dev->dev, mad, bctx->sge.length); rdma_pci_dma_unmap(backend_dev->dev, mad, bctx->sge.length);
comp_handler(IBV_WC_SUCCESS, 0, bctx->up_ctx); wc.byte_len = msg->umad_len;
wc.status = IBV_WC_SUCCESS;
wc.wc_flags = IBV_WC_GRH;
comp_handler(bctx->up_ctx, &wc);
} }
g_free(bctx); g_free(bctx);

View File

@ -57,8 +57,8 @@ int rdma_backend_get_gid_index(RdmaBackendDev *backend_dev,
union ibv_gid *gid); union ibv_gid *gid);
void rdma_backend_start(RdmaBackendDev *backend_dev); void rdma_backend_start(RdmaBackendDev *backend_dev);
void rdma_backend_stop(RdmaBackendDev *backend_dev); void rdma_backend_stop(RdmaBackendDev *backend_dev);
void rdma_backend_register_comp_handler(void (*handler)(int status, void rdma_backend_register_comp_handler(void (*handler)(void *ctx,
unsigned int vendor_err, void *ctx)); struct ibv_wc *wc));
void rdma_backend_unregister_comp_handler(void); void rdma_backend_unregister_comp_handler(void);
int rdma_backend_query_port(RdmaBackendDev *backend_dev, int rdma_backend_query_port(RdmaBackendDev *backend_dev,

View File

@ -47,7 +47,7 @@ typedef struct PvrdmaRqWqe {
* 3. Interrupt host * 3. Interrupt host
*/ */
static int pvrdma_post_cqe(PVRDMADev *dev, uint32_t cq_handle, static int pvrdma_post_cqe(PVRDMADev *dev, uint32_t cq_handle,
struct pvrdma_cqe *cqe) struct pvrdma_cqe *cqe, struct ibv_wc *wc)
{ {
struct pvrdma_cqe *cqe1; struct pvrdma_cqe *cqe1;
struct pvrdma_cqne *cqne; struct pvrdma_cqne *cqne;
@ -66,6 +66,7 @@ static int pvrdma_post_cqe(PVRDMADev *dev, uint32_t cq_handle,
pr_dbg("Writing CQE\n"); pr_dbg("Writing CQE\n");
cqe1 = pvrdma_ring_next_elem_write(ring); cqe1 = pvrdma_ring_next_elem_write(ring);
if (unlikely(!cqe1)) { if (unlikely(!cqe1)) {
pr_dbg("No CQEs in ring\n");
return -EINVAL; return -EINVAL;
} }
@ -73,8 +74,20 @@ static int pvrdma_post_cqe(PVRDMADev *dev, uint32_t cq_handle,
cqe1->wr_id = cqe->wr_id; cqe1->wr_id = cqe->wr_id;
cqe1->qp = cqe->qp; cqe1->qp = cqe->qp;
cqe1->opcode = cqe->opcode; cqe1->opcode = cqe->opcode;
cqe1->status = cqe->status; cqe1->status = wc->status;
cqe1->vendor_err = cqe->vendor_err; cqe1->byte_len = wc->byte_len;
cqe1->src_qp = wc->src_qp;
cqe1->wc_flags = wc->wc_flags;
cqe1->vendor_err = wc->vendor_err;
pr_dbg("wr_id=%" PRIx64 "\n", cqe1->wr_id);
pr_dbg("qp=0x%lx\n", cqe1->qp);
pr_dbg("opcode=%d\n", cqe1->opcode);
pr_dbg("status=%d\n", cqe1->status);
pr_dbg("byte_len=%d\n", cqe1->byte_len);
pr_dbg("src_qp=%d\n", cqe1->src_qp);
pr_dbg("wc_flags=%d\n", cqe1->wc_flags);
pr_dbg("vendor_err=%d\n", cqe1->vendor_err);
pvrdma_ring_write_inc(ring); pvrdma_ring_write_inc(ring);
@ -99,18 +112,12 @@ static int pvrdma_post_cqe(PVRDMADev *dev, uint32_t cq_handle,
return 0; return 0;
} }
static void pvrdma_qp_ops_comp_handler(int status, unsigned int vendor_err, static void pvrdma_qp_ops_comp_handler(void *ctx, struct ibv_wc *wc)
void *ctx)
{ {
CompHandlerCtx *comp_ctx = (CompHandlerCtx *)ctx; CompHandlerCtx *comp_ctx = (CompHandlerCtx *)ctx;
pr_dbg("cq_handle=%d\n", comp_ctx->cq_handle); pvrdma_post_cqe(comp_ctx->dev, comp_ctx->cq_handle, &comp_ctx->cqe, wc);
pr_dbg("wr_id=%" PRIx64 "\n", comp_ctx->cqe.wr_id);
pr_dbg("status=%d\n", status);
pr_dbg("vendor_err=0x%x\n", vendor_err);
comp_ctx->cqe.status = status;
comp_ctx->cqe.vendor_err = vendor_err;
pvrdma_post_cqe(comp_ctx->dev, comp_ctx->cq_handle, &comp_ctx->cqe);
g_free(ctx); g_free(ctx);
} }