2 * Copyright(c) 2016 Intel Corporation.
4 * This file is provided under a dual BSD/GPLv2 license. When using or
5 * redistributing this file, you may do so under either license.
9 * This program is free software; you can redistribute it and/or modify
10 * it under the terms of version 2 of the GNU General Public License as
11 * published by the Free Software Foundation.
13 * This program is distributed in the hope that it will be useful, but
14 * WITHOUT ANY WARRANTY; without even the implied warranty of
15 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
16 * General Public License for more details.
20 * Redistribution and use in source and binary forms, with or without
21 * modification, are permitted provided that the following conditions
24 * - Redistributions of source code must retain the above copyright
25 * notice, this list of conditions and the following disclaimer.
26 * - Redistributions in binary form must reproduce the above copyright
27 * notice, this list of conditions and the following disclaimer in
28 * the documentation and/or other materials provided with the
30 * - Neither the name of Intel Corporation nor the names of its
31 * contributors may be used to endorse or promote products derived
32 * from this software without specific prior written permission.
34 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
35 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
36 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
37 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
38 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
39 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
40 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
41 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
42 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
43 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
44 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
48 #include <linux/module.h>
49 #include <linux/kernel.h>
50 #include <linux/dma-mapping.h>
54 #define RVT_UVERBS_ABI_VERSION 2
56 MODULE_LICENSE("Dual BSD/GPL");
57 MODULE_DESCRIPTION("RDMA Verbs Transport Library");
59 static int rvt_init(void)
62 * rdmavt does not need to do anything special when it starts up. All it
63 * needs to do is sit and wait until a driver attempts registration.
67 module_init(rvt_init);
69 static void rvt_cleanup(void)
72 * Nothing to do at exit time either. The module won't be able to be
73 * removed until all drivers are gone which means all the dev structs
74 * are gone so there is really nothing to do.
77 module_exit(rvt_cleanup);
80 * rvt_alloc_device - allocate rdi
81 * @size: how big of a structure to allocate
82 * @nports: number of ports to allocate array slots for
84 * Use IB core device alloc to allocate space for the rdi which is assumed to be
85 * inside of the ib_device. Any extra space that drivers require should be
88 * We also allocate a port array based on the number of ports.
90 * Return: pointer to allocated rdi
92 struct rvt_dev_info *rvt_alloc_device(size_t size, int nports)
94 struct rvt_dev_info *rdi = ERR_PTR(-ENOMEM);
96 rdi = (struct rvt_dev_info *)ib_alloc_device(size);
100 rdi->ports = kcalloc(nports, sizeof(*rdi->ports), GFP_KERNEL);
102 ib_dealloc_device(&rdi->ibdev);
106 EXPORT_SYMBOL(rvt_alloc_device);
109 * rvt_dealloc_device - deallocate rdi
110 * @rdi: structure to free
112 * Free a structure allocated with rvt_alloc_device()
114 void rvt_dealloc_device(struct rvt_dev_info *rdi)
117 ib_dealloc_device(&rdi->ibdev);
119 EXPORT_SYMBOL(rvt_dealloc_device);
121 static int rvt_query_device(struct ib_device *ibdev,
122 struct ib_device_attr *props,
123 struct ib_udata *uhw)
125 struct rvt_dev_info *rdi = ib_to_rvt(ibdev);
127 if (uhw->inlen || uhw->outlen)
130 * Return rvt_dev_info.dparms.props contents
132 *props = rdi->dparms.props;
136 static int rvt_modify_device(struct ib_device *device,
137 int device_modify_mask,
138 struct ib_device_modify *device_modify)
141 * There is currently no need to supply this based on qib and hfi1.
142 * Future drivers may need to implement this though.
149 * rvt_query_port: Passes the query port call to the driver
150 * @ibdev: Verbs IB dev
151 * @port_num: port number, 1 based from ib core
152 * @props: structure to hold returned properties
154 * Return: 0 on success
156 static int rvt_query_port(struct ib_device *ibdev, u8 port_num,
157 struct ib_port_attr *props)
159 struct rvt_dev_info *rdi = ib_to_rvt(ibdev);
160 struct rvt_ibport *rvp;
161 int port_index = ibport_num_to_idx(ibdev, port_num);
166 rvp = rdi->ports[port_index];
167 /* props being zeroed by the caller, avoid zeroing it here */
168 props->sm_lid = rvp->sm_lid;
169 props->sm_sl = rvp->sm_sl;
170 props->port_cap_flags = rvp->port_cap_flags;
171 props->max_msg_sz = 0x80000000;
172 props->pkey_tbl_len = rvt_get_npkeys(rdi);
173 props->bad_pkey_cntr = rvp->pkey_violations;
174 props->qkey_viol_cntr = rvp->qkey_violations;
175 props->subnet_timeout = rvp->subnet_timeout;
176 props->init_type_reply = 0;
178 /* Populate the remaining ib_port_attr elements */
179 return rdi->driver_f.query_port_state(rdi, port_num, props);
184 * @ibdev: Verbs IB dev
185 * @port_num: Port number, 1 based from ib core
186 * @port_modify_mask: How to change the port
187 * @props: Structure to fill in
189 * Return: 0 on success
191 static int rvt_modify_port(struct ib_device *ibdev, u8 port_num,
192 int port_modify_mask, struct ib_port_modify *props)
194 struct rvt_dev_info *rdi = ib_to_rvt(ibdev);
195 struct rvt_ibport *rvp;
197 int port_index = ibport_num_to_idx(ibdev, port_num);
202 rvp = rdi->ports[port_index];
203 if (port_modify_mask & IB_PORT_OPA_MASK_CHG) {
204 rvp->port_cap3_flags |= props->set_port_cap_mask;
205 rvp->port_cap3_flags &= ~props->clr_port_cap_mask;
207 rvp->port_cap_flags |= props->set_port_cap_mask;
208 rvp->port_cap_flags &= ~props->clr_port_cap_mask;
211 if (props->set_port_cap_mask || props->clr_port_cap_mask)
212 rdi->driver_f.cap_mask_chg(rdi, port_num);
213 if (port_modify_mask & IB_PORT_SHUTDOWN)
214 ret = rdi->driver_f.shut_down_port(rdi, port_num);
215 if (port_modify_mask & IB_PORT_RESET_QKEY_CNTR)
216 rvp->qkey_violations = 0;
222 * rvt_query_pkey - Return a pkey from the table at a given index
223 * @ibdev: Verbs IB dev
224 * @port_num: Port number, 1 based from ib core
225 * @intex: Index into pkey table
227 * Return: 0 on failure pkey otherwise
229 static int rvt_query_pkey(struct ib_device *ibdev, u8 port_num, u16 index,
233 * Driver will be responsible for keeping rvt_dev_info.pkey_table up to
234 * date. This function will just return that value. There is no need to
235 * lock, if a stale value is read and sent to the user so be it there is
236 * no way to protect against that anyway.
238 struct rvt_dev_info *rdi = ib_to_rvt(ibdev);
241 port_index = ibport_num_to_idx(ibdev, port_num);
245 if (index >= rvt_get_npkeys(rdi))
248 *pkey = rvt_get_pkey(rdi, port_index, index);
253 * rvt_query_gid - Return a gid from the table
254 * @ibdev: Verbs IB dev
255 * @port_num: Port number, 1 based from ib core
256 * @index: = Index in table
257 * @gid: Gid to return
259 * Return: 0 on success
261 static int rvt_query_gid(struct ib_device *ibdev, u8 port_num,
262 int guid_index, union ib_gid *gid)
264 struct rvt_dev_info *rdi;
265 struct rvt_ibport *rvp;
269 * Driver is responsible for updating the guid table. Which will be used
270 * to craft the return value. This will work similar to how query_pkey()
273 port_index = ibport_num_to_idx(ibdev, port_num);
277 rdi = ib_to_rvt(ibdev);
278 rvp = rdi->ports[port_index];
280 gid->global.subnet_prefix = rvp->gid_prefix;
282 return rdi->driver_f.get_guid_be(rdi, rvp, guid_index,
283 &gid->global.interface_id);
286 struct rvt_ucontext {
287 struct ib_ucontext ibucontext;
290 static inline struct rvt_ucontext *to_iucontext(struct ib_ucontext
293 return container_of(ibucontext, struct rvt_ucontext, ibucontext);
297 * rvt_alloc_ucontext - Allocate a user context
298 * @ibdev: Vers IB dev
299 * @data: User data allocated
301 static struct ib_ucontext *rvt_alloc_ucontext(struct ib_device *ibdev,
302 struct ib_udata *udata)
304 struct rvt_ucontext *context;
306 context = kmalloc(sizeof(*context), GFP_KERNEL);
308 return ERR_PTR(-ENOMEM);
309 return &context->ibucontext;
313 *rvt_dealloc_ucontext - Free a user context
314 *@context - Free this
316 static int rvt_dealloc_ucontext(struct ib_ucontext *context)
318 kfree(to_iucontext(context));
322 static int rvt_get_port_immutable(struct ib_device *ibdev, u8 port_num,
323 struct ib_port_immutable *immutable)
325 struct rvt_dev_info *rdi = ib_to_rvt(ibdev);
326 struct ib_port_attr attr;
329 port_index = ibport_num_to_idx(ibdev, port_num);
333 immutable->core_cap_flags = rdi->dparms.core_cap_flags;
335 err = ib_query_port(ibdev, port_num, &attr);
339 immutable->pkey_tbl_len = attr.pkey_tbl_len;
340 immutable->gid_tbl_len = attr.gid_tbl_len;
341 immutable->max_mad_size = rdi->dparms.max_mad_size;
391 _VERB_IDX_MAX /* Must always be last! */
394 static inline int check_driver_override(struct rvt_dev_info *rdi,
395 size_t offset, void *func)
397 if (!*(void **)((void *)&rdi->ibdev + offset)) {
398 *(void **)((void *)&rdi->ibdev + offset) = func;
405 static noinline int check_support(struct rvt_dev_info *rdi, int verb)
410 * These functions are not part of verbs specifically but are
411 * required for rdmavt to function.
413 if ((!rdi->driver_f.port_callback) ||
414 (!rdi->driver_f.get_card_name) ||
415 (!rdi->driver_f.get_pci_dev))
420 check_driver_override(rdi, offsetof(struct ib_device,
427 * rdmavt does not support modify device currently drivers must
430 if (!check_driver_override(rdi, offsetof(struct ib_device,
437 if (!check_driver_override(rdi, offsetof(struct ib_device,
440 if (!rdi->driver_f.query_port_state)
445 if (!check_driver_override(rdi, offsetof(struct ib_device,
448 if (!rdi->driver_f.cap_mask_chg ||
449 !rdi->driver_f.shut_down_port)
454 check_driver_override(rdi, offsetof(struct ib_device,
460 if (!check_driver_override(rdi, offsetof(struct ib_device,
463 if (!rdi->driver_f.get_guid_be)
468 check_driver_override(rdi, offsetof(struct ib_device,
473 case DEALLOC_UCONTEXT:
474 check_driver_override(rdi, offsetof(struct ib_device,
476 rvt_dealloc_ucontext);
479 case GET_PORT_IMMUTABLE:
480 check_driver_override(rdi, offsetof(struct ib_device,
482 rvt_get_port_immutable);
486 if (!check_driver_override(rdi, offsetof(struct ib_device,
489 if (!rdi->driver_f.qp_priv_alloc ||
490 !rdi->driver_f.qp_priv_free ||
491 !rdi->driver_f.notify_qp_reset ||
492 !rdi->driver_f.flush_qp_waiters ||
493 !rdi->driver_f.stop_send_queue ||
494 !rdi->driver_f.quiesce_qp)
499 if (!check_driver_override(rdi, offsetof(struct ib_device,
502 if (!rdi->driver_f.notify_qp_reset ||
503 !rdi->driver_f.schedule_send ||
504 !rdi->driver_f.get_pmtu_from_attr ||
505 !rdi->driver_f.flush_qp_waiters ||
506 !rdi->driver_f.stop_send_queue ||
507 !rdi->driver_f.quiesce_qp ||
508 !rdi->driver_f.notify_error_qp ||
509 !rdi->driver_f.mtu_from_qp ||
510 !rdi->driver_f.mtu_to_path_mtu)
515 if (!check_driver_override(rdi, offsetof(struct ib_device,
518 if (!rdi->driver_f.qp_priv_free ||
519 !rdi->driver_f.notify_qp_reset ||
520 !rdi->driver_f.flush_qp_waiters ||
521 !rdi->driver_f.stop_send_queue ||
522 !rdi->driver_f.quiesce_qp)
527 check_driver_override(rdi, offsetof(struct ib_device,
533 if (!check_driver_override(rdi, offsetof(struct ib_device,
536 if (!rdi->driver_f.schedule_send ||
537 !rdi->driver_f.do_send ||
543 check_driver_override(rdi, offsetof(struct ib_device,
548 check_driver_override(rdi, offsetof(struct ib_device,
554 check_driver_override(rdi, offsetof(struct ib_device,
560 check_driver_override(rdi, offsetof(struct ib_device,
566 check_driver_override(rdi, offsetof(struct ib_device,
572 check_driver_override(rdi, offsetof(struct ib_device,
578 check_driver_override(rdi, offsetof(struct ib_device,
584 check_driver_override(rdi, offsetof(struct ib_device,
590 check_driver_override(rdi, offsetof(struct ib_device,
596 check_driver_override(rdi, offsetof(struct ib_device,
602 check_driver_override(rdi, offsetof(struct ib_device,
608 check_driver_override(rdi, offsetof(struct ib_device,
614 check_driver_override(rdi, offsetof(struct ib_device,
620 check_driver_override(rdi, offsetof(struct ib_device,
626 check_driver_override(rdi, offsetof(struct ib_device,
632 check_driver_override(rdi, offsetof(struct ib_device,
638 check_driver_override(rdi, offsetof(struct ib_device,
644 check_driver_override(rdi, offsetof(struct ib_device,
650 check_driver_override(rdi, offsetof(struct ib_device,
656 check_driver_override(rdi, offsetof(struct ib_device,
662 check_driver_override(rdi, offsetof(struct ib_device,
668 check_driver_override(rdi, offsetof(struct ib_device,
674 check_driver_override(rdi, offsetof(struct ib_device,
680 check_driver_override(rdi, offsetof(struct ib_device,
686 check_driver_override(rdi, offsetof(struct ib_device,
692 check_driver_override(rdi, offsetof(struct ib_device,
698 check_driver_override(rdi, offsetof(struct ib_device,
704 check_driver_override(rdi, offsetof(struct ib_device,
710 check_driver_override(rdi, offsetof(struct ib_device,
723 * rvt_register_device - register a driver
724 * @rdi: main dev structure for all of rdmavt operations
726 * It is up to drivers to allocate the rdi and fill in the appropriate
729 * Return: 0 on success otherwise an errno.
731 int rvt_register_device(struct rvt_dev_info *rdi)
739 * Check to ensure drivers have setup the required helpers for the verbs
740 * they want rdmavt to handle
742 for (i = 0; i < _VERB_IDX_MAX; i++)
743 if (check_support(rdi, i)) {
744 pr_err("Driver support req not met at %d\n", i);
749 /* Once we get past here we can use rvt_pr macros and tracepoints */
750 trace_rvt_dbg(rdi, "Driver attempting registration");
754 ret = rvt_driver_qp_init(rdi);
756 pr_err("Error in driver QP init.\n");
761 spin_lock_init(&rdi->n_ahs_lock);
762 rdi->n_ahs_allocated = 0;
764 /* Shared Receive Queue */
765 rvt_driver_srq_init(rdi);
768 rvt_driver_mcast_init(rdi);
771 ret = rvt_driver_mr_init(rdi);
773 pr_err("Error in driver MR init.\n");
777 /* Completion queues */
778 ret = rvt_driver_cq_init(rdi);
780 pr_err("Error in driver CQ init.\n");
785 rdi->ibdev.dev.dma_ops = rdi->ibdev.dev.dma_ops ? : &dma_virt_ops;
787 /* Protection Domain */
788 spin_lock_init(&rdi->n_pds_lock);
789 rdi->n_pds_allocated = 0;
792 * There are some things which could be set by underlying drivers but
793 * really should be up to rdmavt to set. For instance drivers can't know
794 * exactly which functions rdmavt supports, nor do they know the ABI
795 * version, so we do all of this sort of stuff here.
797 rdi->ibdev.uverbs_abi_ver = RVT_UVERBS_ABI_VERSION;
798 rdi->ibdev.uverbs_cmd_mask =
799 (1ull << IB_USER_VERBS_CMD_GET_CONTEXT) |
800 (1ull << IB_USER_VERBS_CMD_QUERY_DEVICE) |
801 (1ull << IB_USER_VERBS_CMD_QUERY_PORT) |
802 (1ull << IB_USER_VERBS_CMD_ALLOC_PD) |
803 (1ull << IB_USER_VERBS_CMD_DEALLOC_PD) |
804 (1ull << IB_USER_VERBS_CMD_CREATE_AH) |
805 (1ull << IB_USER_VERBS_CMD_MODIFY_AH) |
806 (1ull << IB_USER_VERBS_CMD_QUERY_AH) |
807 (1ull << IB_USER_VERBS_CMD_DESTROY_AH) |
808 (1ull << IB_USER_VERBS_CMD_REG_MR) |
809 (1ull << IB_USER_VERBS_CMD_DEREG_MR) |
810 (1ull << IB_USER_VERBS_CMD_CREATE_COMP_CHANNEL) |
811 (1ull << IB_USER_VERBS_CMD_CREATE_CQ) |
812 (1ull << IB_USER_VERBS_CMD_RESIZE_CQ) |
813 (1ull << IB_USER_VERBS_CMD_DESTROY_CQ) |
814 (1ull << IB_USER_VERBS_CMD_POLL_CQ) |
815 (1ull << IB_USER_VERBS_CMD_REQ_NOTIFY_CQ) |
816 (1ull << IB_USER_VERBS_CMD_CREATE_QP) |
817 (1ull << IB_USER_VERBS_CMD_QUERY_QP) |
818 (1ull << IB_USER_VERBS_CMD_MODIFY_QP) |
819 (1ull << IB_USER_VERBS_CMD_DESTROY_QP) |
820 (1ull << IB_USER_VERBS_CMD_POST_SEND) |
821 (1ull << IB_USER_VERBS_CMD_POST_RECV) |
822 (1ull << IB_USER_VERBS_CMD_ATTACH_MCAST) |
823 (1ull << IB_USER_VERBS_CMD_DETACH_MCAST) |
824 (1ull << IB_USER_VERBS_CMD_CREATE_SRQ) |
825 (1ull << IB_USER_VERBS_CMD_MODIFY_SRQ) |
826 (1ull << IB_USER_VERBS_CMD_QUERY_SRQ) |
827 (1ull << IB_USER_VERBS_CMD_DESTROY_SRQ) |
828 (1ull << IB_USER_VERBS_CMD_POST_SRQ_RECV);
829 rdi->ibdev.node_type = RDMA_NODE_IB_CA;
830 rdi->ibdev.num_comp_vectors = 1;
832 /* We are now good to announce we exist */
833 ret = ib_register_device(&rdi->ibdev, rdi->driver_f.port_callback);
835 rvt_pr_err(rdi, "Failed to register driver with ib core.\n");
839 rvt_create_mad_agents(rdi);
841 rvt_pr_info(rdi, "Registration with rdmavt done.\n");
855 EXPORT_SYMBOL(rvt_register_device);
858 * rvt_unregister_device - remove a driver
859 * @rdi: rvt dev struct
861 void rvt_unregister_device(struct rvt_dev_info *rdi)
863 trace_rvt_dbg(rdi, "Driver is unregistering.");
867 rvt_free_mad_agents(rdi);
869 ib_unregister_device(&rdi->ibdev);
874 EXPORT_SYMBOL(rvt_unregister_device);
877 * rvt_init_port - init internal data for driver port
878 * @rdi: rvt dev strut
880 * @port_index: 0 based index of ports, different from IB core port num
882 * Keep track of a list of ports. No need to have a detach port.
883 * They persist until the driver goes away.
887 int rvt_init_port(struct rvt_dev_info *rdi, struct rvt_ibport *port,
888 int port_index, u16 *pkey_table)
891 rdi->ports[port_index] = port;
892 rdi->ports[port_index]->pkey_table = pkey_table;
896 EXPORT_SYMBOL(rvt_init_port);