| // SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB |
| /* |
| * Copyright (c) 2022, Microsoft Corporation. All rights reserved. |
| */ |
| |
| #include <stddef.h> |
| #include <stdio.h> |
| #include <stdlib.h> |
| #include <stdarg.h> |
| #include <stdatomic.h> |
| #include <util/compiler.h> |
| #include <util/util.h> |
| #include <sys/mman.h> |
| |
| #include <infiniband/driver.h> |
| |
| #include <infiniband/kern-abi.h> |
| #include <rdma/mana-abi.h> |
| #include <kernel-abi/mana-abi.h> |
| |
| #include "mana.h" |
| |
| DECLARE_DRV_CMD(mana_alloc_ucontext, IB_USER_VERBS_CMD_GET_CONTEXT, empty, |
| empty); |
| |
| DECLARE_DRV_CMD(mana_alloc_pd, IB_USER_VERBS_CMD_ALLOC_PD, empty, empty); |
| |
| DECLARE_DRV_CMD(mana_create_cq, IB_USER_VERBS_CMD_CREATE_CQ, mana_ib_create_cq, |
| empty); |
| |
| static const struct verbs_match_ent hca_table[] = { |
| VERBS_DRIVER_ID(RDMA_DRIVER_MANA), |
| {}, |
| }; |
| |
| struct mana_context *to_mctx(struct ibv_context *ibctx) |
| { |
| return container_of(ibctx, struct mana_context, ibv_ctx.context); |
| } |
| |
| int mana_query_device_ex(struct ibv_context *context, |
| const struct ibv_query_device_ex_input *input, |
| struct ibv_device_attr_ex *attr, size_t attr_size) |
| { |
| struct ib_uverbs_ex_query_device_resp resp; |
| size_t resp_size = sizeof(resp); |
| int ret; |
| |
| ret = ibv_cmd_query_device_any(context, input, attr, attr_size, &resp, |
| &resp_size); |
| |
| verbs_debug(verbs_get_ctx(context), |
| "device attr max_qp %d max_qp_wr %d max_cqe %d\n", |
| attr->orig_attr.max_qp, attr->orig_attr.max_qp_wr, |
| attr->orig_attr.max_cqe); |
| |
| return ret; |
| } |
| |
| int mana_query_port(struct ibv_context *context, uint8_t port, |
| struct ibv_port_attr *attr) |
| { |
| struct ibv_query_port cmd; |
| |
| return ibv_cmd_query_port(context, port, attr, &cmd, sizeof(cmd)); |
| } |
| |
| struct ibv_pd *mana_alloc_pd(struct ibv_context *context) |
| { |
| struct ibv_alloc_pd cmd; |
| struct mana_alloc_pd_resp resp; |
| struct mana_pd *pd; |
| int ret; |
| |
| pd = calloc(1, sizeof(*pd)); |
| if (!pd) |
| return NULL; |
| |
| ret = ibv_cmd_alloc_pd(context, &pd->ibv_pd, &cmd, sizeof(cmd), |
| &resp.ibv_resp, sizeof(resp)); |
| if (ret) { |
| verbs_err(verbs_get_ctx(context), "Failed to allocate PD\n"); |
| errno = ret; |
| free(pd); |
| return NULL; |
| } |
| |
| return &pd->ibv_pd; |
| } |
| |
| struct ibv_pd * |
| mana_alloc_parent_domain(struct ibv_context *context, |
| struct ibv_parent_domain_init_attr *attr) |
| { |
| struct mana_parent_domain *mparent_domain; |
| |
| if (ibv_check_alloc_parent_domain(attr)) { |
| errno = EINVAL; |
| return NULL; |
| } |
| |
| if (!check_comp_mask(attr->comp_mask, |
| IBV_PARENT_DOMAIN_INIT_ATTR_PD_CONTEXT)) { |
| verbs_err( |
| verbs_get_ctx(context), |
| "This driver supports IBV_PARENT_DOMAIN_INIT_ATTR_PD_CONTEXT only\n"); |
| errno = EOPNOTSUPP; |
| return NULL; |
| } |
| |
| mparent_domain = calloc(1, sizeof(*mparent_domain)); |
| if (!mparent_domain) { |
| errno = ENOMEM; |
| return NULL; |
| } |
| |
| mparent_domain->mpd.mprotection_domain = |
| container_of(attr->pd, struct mana_pd, ibv_pd); |
| ibv_initialize_parent_domain(&mparent_domain->mpd.ibv_pd, attr->pd); |
| |
| if (attr->comp_mask & IBV_PARENT_DOMAIN_INIT_ATTR_PD_CONTEXT) |
| mparent_domain->pd_context = attr->pd_context; |
| |
| return &mparent_domain->mpd.ibv_pd; |
| } |
| |
| int mana_dealloc_pd(struct ibv_pd *ibpd) |
| { |
| int ret; |
| struct mana_pd *pd = container_of(ibpd, struct mana_pd, ibv_pd); |
| |
| if (pd->mprotection_domain) { |
| struct mana_parent_domain *parent_domain = |
| container_of(pd, struct mana_parent_domain, mpd); |
| |
| free(parent_domain); |
| return 0; |
| } |
| |
| ret = ibv_cmd_dealloc_pd(ibpd); |
| if (ret) { |
| verbs_err(verbs_get_ctx(ibpd->context), |
| "Failed to deallocate PD\n"); |
| return ret; |
| } |
| |
| free(pd); |
| |
| return 0; |
| } |
| |
| struct ibv_mr *mana_reg_mr(struct ibv_pd *pd, void *addr, size_t length, |
| uint64_t hca_va, int access) |
| { |
| struct verbs_mr *vmr; |
| struct ibv_reg_mr cmd; |
| struct ib_uverbs_reg_mr_resp resp; |
| int ret; |
| |
| vmr = malloc(sizeof(*vmr)); |
| if (!vmr) |
| return NULL; |
| |
| ret = ibv_cmd_reg_mr(pd, addr, length, hca_va, access, vmr, &cmd, |
| sizeof(cmd), &resp, sizeof(resp)); |
| if (ret) { |
| verbs_err(verbs_get_ctx(pd->context), |
| "Failed to register MR\n"); |
| errno = ret; |
| free(vmr); |
| return NULL; |
| } |
| |
| return &vmr->ibv_mr; |
| } |
| |
| int mana_dereg_mr(struct verbs_mr *vmr) |
| { |
| int ret; |
| |
| ret = ibv_cmd_dereg_mr(vmr); |
| if (ret) { |
| verbs_err(verbs_get_ctx(vmr->ibv_mr.context), |
| "Failed to deregister MR\n"); |
| return ret; |
| } |
| |
| free(vmr); |
| return 0; |
| } |
| |
| struct ibv_cq *mana_create_cq(struct ibv_context *context, int cqe, |
| struct ibv_comp_channel *channel, int comp_vector) |
| { |
| struct mana_context *ctx = to_mctx(context); |
| struct mana_cq *cq; |
| struct mana_create_cq cmd = {}; |
| struct mana_create_cq_resp resp = {}; |
| struct mana_ib_create_cq *cmd_drv; |
| int cq_size; |
| int ret; |
| |
| if (cqe > MAX_SEND_BUFFERS_PER_QUEUE) { |
| verbs_err(verbs_get_ctx(context), "CQE %d exceeding limit\n", |
| cqe); |
| errno = EINVAL; |
| return NULL; |
| } |
| |
| if (!ctx->extern_alloc.alloc || !ctx->extern_alloc.free) { |
| /* |
| * This version of driver doesn't support allocating buffers |
| * in rdma-core. |
| */ |
| verbs_err(verbs_get_ctx(context), |
| "Allocating core buffers for CQ is not supported\n"); |
| errno = EINVAL; |
| return NULL; |
| } |
| |
| cq = calloc(1, sizeof(*cq)); |
| if (!cq) |
| return NULL; |
| |
| cq_size = align_hw_size(cqe * COMP_ENTRY_SIZE); |
| |
| cq->buf = ctx->extern_alloc.alloc(cq_size, ctx->extern_alloc.data); |
| if (!cq->buf) { |
| errno = ENOMEM; |
| goto free_cq; |
| } |
| cq->cqe = cqe; |
| |
| cmd_drv = &cmd.drv_payload; |
| cmd_drv->buf_addr = (uintptr_t)cq->buf; |
| |
| ret = ibv_cmd_create_cq(context, cq->cqe, channel, comp_vector, |
| &cq->ibcq, &cmd.ibv_cmd, sizeof(cmd), |
| &resp.ibv_resp, sizeof(resp)); |
| |
| if (ret) { |
| verbs_err(verbs_get_ctx(context), "Failed to Create CQ\n"); |
| ctx->extern_alloc.free(cq->buf, ctx->extern_alloc.data); |
| errno = ret; |
| goto free_cq; |
| } |
| |
| return &cq->ibcq; |
| |
| free_cq: |
| free(cq); |
| return NULL; |
| } |
| |
| int mana_destroy_cq(struct ibv_cq *ibcq) |
| { |
| int ret; |
| struct mana_cq *cq = container_of(ibcq, struct mana_cq, ibcq); |
| struct mana_context *ctx = to_mctx(ibcq->context); |
| |
| if (!ctx->extern_alloc.free) { |
| /* |
| * This version of driver doesn't support allocating buffers |
| * in rdma-core. It's not possible to reach the code here. |
| */ |
| verbs_err(verbs_get_ctx(ibcq->context), |
| "Invalid external context in destroy CQ\n"); |
| return -EINVAL; |
| } |
| |
| ret = ibv_cmd_destroy_cq(ibcq); |
| if (ret) { |
| verbs_err(verbs_get_ctx(ibcq->context), |
| "Failed to Destroy CQ\n"); |
| return ret; |
| } |
| |
| ctx->extern_alloc.free(cq->buf, ctx->extern_alloc.data); |
| free(cq); |
| |
| return ret; |
| } |
| |
| static int mana_poll_cq(struct ibv_cq *ibcq, int nwc, struct ibv_wc *wc) |
| { |
| /* This version of driver supports RAW QP only. |
| * Polling CQ is done directly in the application. |
| */ |
| return EOPNOTSUPP; |
| } |
| |
| static int mana_post_recv(struct ibv_qp *ibqp, struct ibv_recv_wr *wr, |
| struct ibv_recv_wr **bad) |
| { |
| /* This version of driver supports RAW QP only. |
| * Posting WR is done directly in the application. |
| */ |
| return EOPNOTSUPP; |
| } |
| |
| static int mana_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr, |
| struct ibv_send_wr **bad) |
| { |
| /* This version of driver supports RAW QP only. |
| * Posting WR is done directly in the application. |
| */ |
| return EOPNOTSUPP; |
| } |
| |
| static void mana_free_context(struct ibv_context *ibctx) |
| { |
| struct mana_context *context = to_mctx(ibctx); |
| |
| munmap(context->db_page, DOORBELL_PAGE_SIZE); |
| verbs_uninit_context(&context->ibv_ctx); |
| free(context); |
| } |
| |
| static const struct verbs_context_ops mana_ctx_ops = { |
| .alloc_pd = mana_alloc_pd, |
| .alloc_parent_domain = mana_alloc_parent_domain, |
| .create_cq = mana_create_cq, |
| .create_qp = mana_create_qp, |
| .create_qp_ex = mana_create_qp_ex, |
| .create_rwq_ind_table = mana_create_rwq_ind_table, |
| .create_wq = mana_create_wq, |
| .dealloc_pd = mana_dealloc_pd, |
| .dereg_mr = mana_dereg_mr, |
| .destroy_cq = mana_destroy_cq, |
| .destroy_qp = mana_destroy_qp, |
| .destroy_rwq_ind_table = mana_destroy_rwq_ind_table, |
| .destroy_wq = mana_destroy_wq, |
| .free_context = mana_free_context, |
| .modify_wq = mana_modify_wq, |
| .modify_qp = mana_modify_qp, |
| .poll_cq = mana_poll_cq, |
| .post_recv = mana_post_recv, |
| .post_send = mana_post_send, |
| .query_device_ex = mana_query_device_ex, |
| .query_port = mana_query_port, |
| .reg_mr = mana_reg_mr, |
| }; |
| |
| static struct verbs_device *mana_device_alloc(struct verbs_sysfs_dev *sysfs_dev) |
| { |
| struct mana_device *dev; |
| |
| dev = calloc(1, sizeof(*dev)); |
| if (!dev) |
| return NULL; |
| |
| return &dev->verbs_dev; |
| } |
| |
| static void mana_uninit_device(struct verbs_device *verbs_device) |
| { |
| struct mana_device *dev = |
| container_of(verbs_device, struct mana_device, verbs_dev); |
| |
| free(dev); |
| } |
| |
| static struct verbs_context *mana_alloc_context(struct ibv_device *ibdev, |
| int cmd_fd, void *private_data) |
| { |
| int ret; |
| struct mana_context *context; |
| struct mana_alloc_ucontext_resp resp; |
| struct ibv_get_context cmd; |
| |
| context = verbs_init_and_alloc_context(ibdev, cmd_fd, context, ibv_ctx, |
| RDMA_DRIVER_MANA); |
| if (!context) |
| return NULL; |
| |
| ret = ibv_cmd_get_context(&context->ibv_ctx, &cmd, sizeof(cmd), |
| &resp.ibv_resp, sizeof(resp)); |
| if (ret) { |
| verbs_err(&context->ibv_ctx, "Failed to get ucontext\n"); |
| errno = ret; |
| goto free_ctx; |
| } |
| |
| verbs_set_ops(&context->ibv_ctx, &mana_ctx_ops); |
| |
| context->db_page = mmap(NULL, DOORBELL_PAGE_SIZE, PROT_WRITE, |
| MAP_SHARED, context->ibv_ctx.context.cmd_fd, 0); |
| if (context->db_page == MAP_FAILED) { |
| verbs_err(&context->ibv_ctx, "Failed to map doorbell page\n"); |
| errno = ENOENT; |
| goto free_ctx; |
| } |
| verbs_debug(&context->ibv_ctx, "Mapped db_page=%p\n", context->db_page); |
| |
| return &context->ibv_ctx; |
| |
| free_ctx: |
| verbs_uninit_context(&context->ibv_ctx); |
| free(context); |
| return NULL; |
| } |
| |
| static const struct verbs_device_ops mana_dev_ops = { |
| .name = "mana", |
| .match_min_abi_version = MANA_IB_UVERBS_ABI_VERSION, |
| .match_max_abi_version = MANA_IB_UVERBS_ABI_VERSION, |
| .match_table = hca_table, |
| .alloc_device = mana_device_alloc, |
| .uninit_device = mana_uninit_device, |
| .alloc_context = mana_alloc_context, |
| }; |
| |
| PROVIDER_DRIVER(mana, mana_dev_ops); |