kernel_samsung_a34x-permissive/drivers/infiniband/sw/rdmavt/vt.c
2024-04-28 15:51:13 +02:00

889 lines
22 KiB
C

/*
* Copyright(c) 2016 - 2018 Intel Corporation.
*
* This file is provided under a dual BSD/GPLv2 license. When using or
* redistributing this file, you may do so under either license.
*
* GPL LICENSE SUMMARY
*
* This program is free software; you can redistribute it and/or modify
* it under the terms of version 2 of the GNU General Public License as
* published by the Free Software Foundation.
*
* This program is distributed in the hope that it will be useful, but
* WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* General Public License for more details.
*
* BSD LICENSE
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
*
* - Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* - Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in
* the documentation and/or other materials provided with the
* distribution.
* - Neither the name of Intel Corporation nor the names of its
* contributors may be used to endorse or promote products derived
* from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*
*/
#include <linux/module.h>
#include <linux/kernel.h>
#include <linux/dma-mapping.h>
#include "vt.h"
#include "cq.h"
#include "trace.h"
#define RVT_UVERBS_ABI_VERSION 2
MODULE_LICENSE("Dual BSD/GPL");
MODULE_DESCRIPTION("RDMA Verbs Transport Library");
static int rvt_init(void)
{
int ret = rvt_driver_cq_init();
if (ret)
pr_err("Error in driver CQ init.\n");
return ret;
}
module_init(rvt_init);
static void rvt_cleanup(void)
{
rvt_cq_exit();
}
module_exit(rvt_cleanup);
/**
* rvt_alloc_device - allocate rdi
* @size: how big of a structure to allocate
* @nports: number of ports to allocate array slots for
*
* Use IB core device alloc to allocate space for the rdi which is assumed to be
* inside of the ib_device. Any extra space that drivers require should be
* included in size.
*
* We also allocate a port array based on the number of ports.
*
* Return: pointer to allocated rdi
*/
struct rvt_dev_info *rvt_alloc_device(size_t size, int nports)
{
struct rvt_dev_info *rdi;
rdi = (struct rvt_dev_info *)ib_alloc_device(size);
if (!rdi)
return rdi;
rdi->ports = kcalloc(nports, sizeof(*rdi->ports), GFP_KERNEL);
if (!rdi->ports)
ib_dealloc_device(&rdi->ibdev);
return rdi;
}
EXPORT_SYMBOL(rvt_alloc_device);
/**
* rvt_dealloc_device - deallocate rdi
* @rdi: structure to free
*
* Free a structure allocated with rvt_alloc_device()
*/
void rvt_dealloc_device(struct rvt_dev_info *rdi)
{
kfree(rdi->ports);
ib_dealloc_device(&rdi->ibdev);
}
EXPORT_SYMBOL(rvt_dealloc_device);
static int rvt_query_device(struct ib_device *ibdev,
struct ib_device_attr *props,
struct ib_udata *uhw)
{
struct rvt_dev_info *rdi = ib_to_rvt(ibdev);
if (uhw->inlen || uhw->outlen)
return -EINVAL;
/*
* Return rvt_dev_info.dparms.props contents
*/
*props = rdi->dparms.props;
return 0;
}
static int rvt_modify_device(struct ib_device *device,
int device_modify_mask,
struct ib_device_modify *device_modify)
{
/*
* There is currently no need to supply this based on qib and hfi1.
* Future drivers may need to implement this though.
*/
return -EOPNOTSUPP;
}
/**
* rvt_query_port: Passes the query port call to the driver
* @ibdev: Verbs IB dev
* @port_num: port number, 1 based from ib core
* @props: structure to hold returned properties
*
* Return: 0 on success
*/
static int rvt_query_port(struct ib_device *ibdev, u8 port_num,
struct ib_port_attr *props)
{
struct rvt_dev_info *rdi = ib_to_rvt(ibdev);
struct rvt_ibport *rvp;
int port_index = ibport_num_to_idx(ibdev, port_num);
if (port_index < 0)
return -EINVAL;
rvp = rdi->ports[port_index];
/* props being zeroed by the caller, avoid zeroing it here */
props->sm_lid = rvp->sm_lid;
props->sm_sl = rvp->sm_sl;
props->port_cap_flags = rvp->port_cap_flags;
props->max_msg_sz = 0x80000000;
props->pkey_tbl_len = rvt_get_npkeys(rdi);
props->bad_pkey_cntr = rvp->pkey_violations;
props->qkey_viol_cntr = rvp->qkey_violations;
props->subnet_timeout = rvp->subnet_timeout;
props->init_type_reply = 0;
/* Populate the remaining ib_port_attr elements */
return rdi->driver_f.query_port_state(rdi, port_num, props);
}
/**
* rvt_modify_port
* @ibdev: Verbs IB dev
* @port_num: Port number, 1 based from ib core
* @port_modify_mask: How to change the port
* @props: Structure to fill in
*
* Return: 0 on success
*/
static int rvt_modify_port(struct ib_device *ibdev, u8 port_num,
int port_modify_mask, struct ib_port_modify *props)
{
struct rvt_dev_info *rdi = ib_to_rvt(ibdev);
struct rvt_ibport *rvp;
int ret = 0;
int port_index = ibport_num_to_idx(ibdev, port_num);
if (port_index < 0)
return -EINVAL;
rvp = rdi->ports[port_index];
if (port_modify_mask & IB_PORT_OPA_MASK_CHG) {
rvp->port_cap3_flags |= props->set_port_cap_mask;
rvp->port_cap3_flags &= ~props->clr_port_cap_mask;
} else {
rvp->port_cap_flags |= props->set_port_cap_mask;
rvp->port_cap_flags &= ~props->clr_port_cap_mask;
}
if (props->set_port_cap_mask || props->clr_port_cap_mask)
rdi->driver_f.cap_mask_chg(rdi, port_num);
if (port_modify_mask & IB_PORT_SHUTDOWN)
ret = rdi->driver_f.shut_down_port(rdi, port_num);
if (port_modify_mask & IB_PORT_RESET_QKEY_CNTR)
rvp->qkey_violations = 0;
return ret;
}
/**
* rvt_query_pkey - Return a pkey from the table at a given index
* @ibdev: Verbs IB dev
* @port_num: Port number, 1 based from ib core
* @index: Index into pkey table
* @pkey: returned pkey from the port pkey table
*
* Return: 0 on failure pkey otherwise
*/
static int rvt_query_pkey(struct ib_device *ibdev, u8 port_num, u16 index,
u16 *pkey)
{
/*
* Driver will be responsible for keeping rvt_dev_info.pkey_table up to
* date. This function will just return that value. There is no need to
* lock, if a stale value is read and sent to the user so be it there is
* no way to protect against that anyway.
*/
struct rvt_dev_info *rdi = ib_to_rvt(ibdev);
int port_index;
port_index = ibport_num_to_idx(ibdev, port_num);
if (port_index < 0)
return -EINVAL;
if (index >= rvt_get_npkeys(rdi))
return -EINVAL;
*pkey = rvt_get_pkey(rdi, port_index, index);
return 0;
}
/**
* rvt_query_gid - Return a gid from the table
* @ibdev: Verbs IB dev
* @port_num: Port number, 1 based from ib core
* @guid_index: Index in table
* @gid: Gid to return
*
* Return: 0 on success
*/
static int rvt_query_gid(struct ib_device *ibdev, u8 port_num,
int guid_index, union ib_gid *gid)
{
struct rvt_dev_info *rdi;
struct rvt_ibport *rvp;
int port_index;
/*
* Driver is responsible for updating the guid table. Which will be used
* to craft the return value. This will work similar to how query_pkey()
* is being done.
*/
port_index = ibport_num_to_idx(ibdev, port_num);
if (port_index < 0)
return -EINVAL;
rdi = ib_to_rvt(ibdev);
rvp = rdi->ports[port_index];
gid->global.subnet_prefix = rvp->gid_prefix;
return rdi->driver_f.get_guid_be(rdi, rvp, guid_index,
&gid->global.interface_id);
}
struct rvt_ucontext {
struct ib_ucontext ibucontext;
};
static inline struct rvt_ucontext *to_iucontext(struct ib_ucontext
*ibucontext)
{
return container_of(ibucontext, struct rvt_ucontext, ibucontext);
}
/**
* rvt_alloc_ucontext - Allocate a user context
* @ibdev: Verbs IB dev
* @udata: User data allocated
*/
static struct ib_ucontext *rvt_alloc_ucontext(struct ib_device *ibdev,
struct ib_udata *udata)
{
struct rvt_ucontext *context;
context = kmalloc(sizeof(*context), GFP_KERNEL);
if (!context)
return ERR_PTR(-ENOMEM);
return &context->ibucontext;
}
/**
*rvt_dealloc_ucontext - Free a user context
*@context - Free this
*/
static int rvt_dealloc_ucontext(struct ib_ucontext *context)
{
kfree(to_iucontext(context));
return 0;
}
static int rvt_get_port_immutable(struct ib_device *ibdev, u8 port_num,
struct ib_port_immutable *immutable)
{
struct rvt_dev_info *rdi = ib_to_rvt(ibdev);
struct ib_port_attr attr;
int err, port_index;
port_index = ibport_num_to_idx(ibdev, port_num);
if (port_index < 0)
return -EINVAL;
immutable->core_cap_flags = rdi->dparms.core_cap_flags;
err = ib_query_port(ibdev, port_num, &attr);
if (err)
return err;
immutable->pkey_tbl_len = attr.pkey_tbl_len;
immutable->gid_tbl_len = attr.gid_tbl_len;
immutable->max_mad_size = rdi->dparms.max_mad_size;
return 0;
}
enum {
MISC,
QUERY_DEVICE,
MODIFY_DEVICE,
QUERY_PORT,
MODIFY_PORT,
QUERY_PKEY,
QUERY_GID,
ALLOC_UCONTEXT,
DEALLOC_UCONTEXT,
GET_PORT_IMMUTABLE,
CREATE_QP,
MODIFY_QP,
DESTROY_QP,
QUERY_QP,
POST_SEND,
POST_RECV,
POST_SRQ_RECV,
CREATE_AH,
DESTROY_AH,
MODIFY_AH,
QUERY_AH,
CREATE_SRQ,
MODIFY_SRQ,
DESTROY_SRQ,
QUERY_SRQ,
ATTACH_MCAST,
DETACH_MCAST,
GET_DMA_MR,
REG_USER_MR,
DEREG_MR,
ALLOC_MR,
MAP_MR_SG,
ALLOC_FMR,
MAP_PHYS_FMR,
UNMAP_FMR,
DEALLOC_FMR,
MMAP,
CREATE_CQ,
DESTROY_CQ,
POLL_CQ,
REQ_NOTFIY_CQ,
RESIZE_CQ,
ALLOC_PD,
DEALLOC_PD,
_VERB_IDX_MAX /* Must always be last! */
};
static inline int check_driver_override(struct rvt_dev_info *rdi,
size_t offset, void *func)
{
if (!*(void **)((void *)&rdi->ibdev + offset)) {
*(void **)((void *)&rdi->ibdev + offset) = func;
return 0;
}
return 1;
}
static noinline int check_support(struct rvt_dev_info *rdi, int verb)
{
switch (verb) {
case MISC:
/*
* These functions are not part of verbs specifically but are
* required for rdmavt to function.
*/
if ((!rdi->driver_f.port_callback) ||
(!rdi->driver_f.get_pci_dev))
return -EINVAL;
break;
case QUERY_DEVICE:
check_driver_override(rdi, offsetof(struct ib_device,
query_device),
rvt_query_device);
break;
case MODIFY_DEVICE:
/*
* rdmavt does not support modify device currently drivers must
* provide.
*/
if (!check_driver_override(rdi, offsetof(struct ib_device,
modify_device),
rvt_modify_device))
return -EOPNOTSUPP;
break;
case QUERY_PORT:
if (!check_driver_override(rdi, offsetof(struct ib_device,
query_port),
rvt_query_port))
if (!rdi->driver_f.query_port_state)
return -EINVAL;
break;
case MODIFY_PORT:
if (!check_driver_override(rdi, offsetof(struct ib_device,
modify_port),
rvt_modify_port))
if (!rdi->driver_f.cap_mask_chg ||
!rdi->driver_f.shut_down_port)
return -EINVAL;
break;
case QUERY_PKEY:
check_driver_override(rdi, offsetof(struct ib_device,
query_pkey),
rvt_query_pkey);
break;
case QUERY_GID:
if (!check_driver_override(rdi, offsetof(struct ib_device,
query_gid),
rvt_query_gid))
if (!rdi->driver_f.get_guid_be)
return -EINVAL;
break;
case ALLOC_UCONTEXT:
check_driver_override(rdi, offsetof(struct ib_device,
alloc_ucontext),
rvt_alloc_ucontext);
break;
case DEALLOC_UCONTEXT:
check_driver_override(rdi, offsetof(struct ib_device,
dealloc_ucontext),
rvt_dealloc_ucontext);
break;
case GET_PORT_IMMUTABLE:
check_driver_override(rdi, offsetof(struct ib_device,
get_port_immutable),
rvt_get_port_immutable);
break;
case CREATE_QP:
if (!check_driver_override(rdi, offsetof(struct ib_device,
create_qp),
rvt_create_qp))
if (!rdi->driver_f.qp_priv_alloc ||
!rdi->driver_f.qp_priv_free ||
!rdi->driver_f.notify_qp_reset ||
!rdi->driver_f.flush_qp_waiters ||
!rdi->driver_f.stop_send_queue ||
!rdi->driver_f.quiesce_qp)
return -EINVAL;
break;
case MODIFY_QP:
if (!check_driver_override(rdi, offsetof(struct ib_device,
modify_qp),
rvt_modify_qp))
if (!rdi->driver_f.notify_qp_reset ||
!rdi->driver_f.schedule_send ||
!rdi->driver_f.get_pmtu_from_attr ||
!rdi->driver_f.flush_qp_waiters ||
!rdi->driver_f.stop_send_queue ||
!rdi->driver_f.quiesce_qp ||
!rdi->driver_f.notify_error_qp ||
!rdi->driver_f.mtu_from_qp ||
!rdi->driver_f.mtu_to_path_mtu)
return -EINVAL;
break;
case DESTROY_QP:
if (!check_driver_override(rdi, offsetof(struct ib_device,
destroy_qp),
rvt_destroy_qp))
if (!rdi->driver_f.qp_priv_free ||
!rdi->driver_f.notify_qp_reset ||
!rdi->driver_f.flush_qp_waiters ||
!rdi->driver_f.stop_send_queue ||
!rdi->driver_f.quiesce_qp)
return -EINVAL;
break;
case QUERY_QP:
check_driver_override(rdi, offsetof(struct ib_device,
query_qp),
rvt_query_qp);
break;
case POST_SEND:
if (!check_driver_override(rdi, offsetof(struct ib_device,
post_send),
rvt_post_send))
if (!rdi->driver_f.schedule_send ||
!rdi->driver_f.do_send ||
!rdi->post_parms)
return -EINVAL;
break;
case POST_RECV:
check_driver_override(rdi, offsetof(struct ib_device,
post_recv),
rvt_post_recv);
break;
case POST_SRQ_RECV:
check_driver_override(rdi, offsetof(struct ib_device,
post_srq_recv),
rvt_post_srq_recv);
break;
case CREATE_AH:
check_driver_override(rdi, offsetof(struct ib_device,
create_ah),
rvt_create_ah);
break;
case DESTROY_AH:
check_driver_override(rdi, offsetof(struct ib_device,
destroy_ah),
rvt_destroy_ah);
break;
case MODIFY_AH:
check_driver_override(rdi, offsetof(struct ib_device,
modify_ah),
rvt_modify_ah);
break;
case QUERY_AH:
check_driver_override(rdi, offsetof(struct ib_device,
query_ah),
rvt_query_ah);
break;
case CREATE_SRQ:
check_driver_override(rdi, offsetof(struct ib_device,
create_srq),
rvt_create_srq);
break;
case MODIFY_SRQ:
check_driver_override(rdi, offsetof(struct ib_device,
modify_srq),
rvt_modify_srq);
break;
case DESTROY_SRQ:
check_driver_override(rdi, offsetof(struct ib_device,
destroy_srq),
rvt_destroy_srq);
break;
case QUERY_SRQ:
check_driver_override(rdi, offsetof(struct ib_device,
query_srq),
rvt_query_srq);
break;
case ATTACH_MCAST:
check_driver_override(rdi, offsetof(struct ib_device,
attach_mcast),
rvt_attach_mcast);
break;
case DETACH_MCAST:
check_driver_override(rdi, offsetof(struct ib_device,
detach_mcast),
rvt_detach_mcast);
break;
case GET_DMA_MR:
check_driver_override(rdi, offsetof(struct ib_device,
get_dma_mr),
rvt_get_dma_mr);
break;
case REG_USER_MR:
check_driver_override(rdi, offsetof(struct ib_device,
reg_user_mr),
rvt_reg_user_mr);
break;
case DEREG_MR:
check_driver_override(rdi, offsetof(struct ib_device,
dereg_mr),
rvt_dereg_mr);
break;
case ALLOC_FMR:
check_driver_override(rdi, offsetof(struct ib_device,
alloc_fmr),
rvt_alloc_fmr);
break;
case ALLOC_MR:
check_driver_override(rdi, offsetof(struct ib_device,
alloc_mr),
rvt_alloc_mr);
break;
case MAP_MR_SG:
check_driver_override(rdi, offsetof(struct ib_device,
map_mr_sg),
rvt_map_mr_sg);
break;
case MAP_PHYS_FMR:
check_driver_override(rdi, offsetof(struct ib_device,
map_phys_fmr),
rvt_map_phys_fmr);
break;
case UNMAP_FMR:
check_driver_override(rdi, offsetof(struct ib_device,
unmap_fmr),
rvt_unmap_fmr);
break;
case DEALLOC_FMR:
check_driver_override(rdi, offsetof(struct ib_device,
dealloc_fmr),
rvt_dealloc_fmr);
break;
case MMAP:
check_driver_override(rdi, offsetof(struct ib_device,
mmap),
rvt_mmap);
break;
case CREATE_CQ:
check_driver_override(rdi, offsetof(struct ib_device,
create_cq),
rvt_create_cq);
break;
case DESTROY_CQ:
check_driver_override(rdi, offsetof(struct ib_device,
destroy_cq),
rvt_destroy_cq);
break;
case POLL_CQ:
check_driver_override(rdi, offsetof(struct ib_device,
poll_cq),
rvt_poll_cq);
break;
case REQ_NOTFIY_CQ:
check_driver_override(rdi, offsetof(struct ib_device,
req_notify_cq),
rvt_req_notify_cq);
break;
case RESIZE_CQ:
check_driver_override(rdi, offsetof(struct ib_device,
resize_cq),
rvt_resize_cq);
break;
case ALLOC_PD:
check_driver_override(rdi, offsetof(struct ib_device,
alloc_pd),
rvt_alloc_pd);
break;
case DEALLOC_PD:
check_driver_override(rdi, offsetof(struct ib_device,
dealloc_pd),
rvt_dealloc_pd);
break;
default:
return -EINVAL;
}
return 0;
}
/**
* rvt_register_device - register a driver
* @rdi: main dev structure for all of rdmavt operations
*
* It is up to drivers to allocate the rdi and fill in the appropriate
* information.
*
* Return: 0 on success otherwise an errno.
*/
int rvt_register_device(struct rvt_dev_info *rdi, u32 driver_id)
{
int ret = 0, i;
if (!rdi)
return -EINVAL;
/*
* Check to ensure drivers have setup the required helpers for the verbs
* they want rdmavt to handle
*/
for (i = 0; i < _VERB_IDX_MAX; i++)
if (check_support(rdi, i)) {
pr_err("Driver support req not met at %d\n", i);
return -EINVAL;
}
/* Once we get past here we can use rvt_pr macros and tracepoints */
trace_rvt_dbg(rdi, "Driver attempting registration");
rvt_mmap_init(rdi);
/* Queue Pairs */
ret = rvt_driver_qp_init(rdi);
if (ret) {
pr_err("Error in driver QP init.\n");
return -EINVAL;
}
/* Address Handle */
spin_lock_init(&rdi->n_ahs_lock);
rdi->n_ahs_allocated = 0;
/* Shared Receive Queue */
rvt_driver_srq_init(rdi);
/* Multicast */
rvt_driver_mcast_init(rdi);
/* Mem Region */
ret = rvt_driver_mr_init(rdi);
if (ret) {
pr_err("Error in driver MR init.\n");
goto bail_no_mr;
}
/* Completion queues */
spin_lock_init(&rdi->n_cqs_lock);
/* DMA Operations */
rdi->ibdev.dev.dma_ops = rdi->ibdev.dev.dma_ops ? : &dma_virt_ops;
/* Protection Domain */
spin_lock_init(&rdi->n_pds_lock);
rdi->n_pds_allocated = 0;
/*
* There are some things which could be set by underlying drivers but
* really should be up to rdmavt to set. For instance drivers can't know
* exactly which functions rdmavt supports, nor do they know the ABI
* version, so we do all of this sort of stuff here.
*/
rdi->ibdev.uverbs_abi_ver = RVT_UVERBS_ABI_VERSION;
rdi->ibdev.uverbs_cmd_mask =
(1ull << IB_USER_VERBS_CMD_GET_CONTEXT) |
(1ull << IB_USER_VERBS_CMD_QUERY_DEVICE) |
(1ull << IB_USER_VERBS_CMD_QUERY_PORT) |
(1ull << IB_USER_VERBS_CMD_ALLOC_PD) |
(1ull << IB_USER_VERBS_CMD_DEALLOC_PD) |
(1ull << IB_USER_VERBS_CMD_CREATE_AH) |
(1ull << IB_USER_VERBS_CMD_MODIFY_AH) |
(1ull << IB_USER_VERBS_CMD_QUERY_AH) |
(1ull << IB_USER_VERBS_CMD_DESTROY_AH) |
(1ull << IB_USER_VERBS_CMD_REG_MR) |
(1ull << IB_USER_VERBS_CMD_DEREG_MR) |
(1ull << IB_USER_VERBS_CMD_CREATE_COMP_CHANNEL) |
(1ull << IB_USER_VERBS_CMD_CREATE_CQ) |
(1ull << IB_USER_VERBS_CMD_RESIZE_CQ) |
(1ull << IB_USER_VERBS_CMD_DESTROY_CQ) |
(1ull << IB_USER_VERBS_CMD_POLL_CQ) |
(1ull << IB_USER_VERBS_CMD_REQ_NOTIFY_CQ) |
(1ull << IB_USER_VERBS_CMD_CREATE_QP) |
(1ull << IB_USER_VERBS_CMD_QUERY_QP) |
(1ull << IB_USER_VERBS_CMD_MODIFY_QP) |
(1ull << IB_USER_VERBS_CMD_DESTROY_QP) |
(1ull << IB_USER_VERBS_CMD_POST_SEND) |
(1ull << IB_USER_VERBS_CMD_POST_RECV) |
(1ull << IB_USER_VERBS_CMD_ATTACH_MCAST) |
(1ull << IB_USER_VERBS_CMD_DETACH_MCAST) |
(1ull << IB_USER_VERBS_CMD_CREATE_SRQ) |
(1ull << IB_USER_VERBS_CMD_MODIFY_SRQ) |
(1ull << IB_USER_VERBS_CMD_QUERY_SRQ) |
(1ull << IB_USER_VERBS_CMD_DESTROY_SRQ) |
(1ull << IB_USER_VERBS_CMD_POST_SRQ_RECV);
rdi->ibdev.node_type = RDMA_NODE_IB_CA;
if (!rdi->ibdev.num_comp_vectors)
rdi->ibdev.num_comp_vectors = 1;
rdi->ibdev.driver_id = driver_id;
/* We are now good to announce we exist */
ret = ib_register_device(&rdi->ibdev, rdi->driver_f.port_callback);
if (ret) {
rvt_pr_err(rdi, "Failed to register driver with ib core.\n");
goto bail_mr;
}
rvt_create_mad_agents(rdi);
rvt_pr_info(rdi, "Registration with rdmavt done.\n");
return ret;
bail_mr:
rvt_mr_exit(rdi);
bail_no_mr:
rvt_qp_exit(rdi);
return ret;
}
EXPORT_SYMBOL(rvt_register_device);
/**
* rvt_unregister_device - remove a driver
* @rdi: rvt dev struct
*/
void rvt_unregister_device(struct rvt_dev_info *rdi)
{
trace_rvt_dbg(rdi, "Driver is unregistering.");
if (!rdi)
return;
rvt_free_mad_agents(rdi);
ib_unregister_device(&rdi->ibdev);
rvt_mr_exit(rdi);
rvt_qp_exit(rdi);
}
EXPORT_SYMBOL(rvt_unregister_device);
/**
* rvt_init_port - init internal data for driver port
* @rdi: rvt dev strut
* @port: rvt port
* @port_index: 0 based index of ports, different from IB core port num
*
* Keep track of a list of ports. No need to have a detach port.
* They persist until the driver goes away.
*
* Return: always 0
*/
int rvt_init_port(struct rvt_dev_info *rdi, struct rvt_ibport *port,
int port_index, u16 *pkey_table)
{
rdi->ports[port_index] = port;
rdi->ports[port_index]->pkey_table = pkey_table;
return 0;
}
EXPORT_SYMBOL(rvt_init_port);