Mention branches and keyring.
[releases.git] / hw / irdma / main.c
1 // SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB
2 /* Copyright (c) 2015 - 2021 Intel Corporation */
3 #include "main.h"
4 #include "../../../net/ethernet/intel/ice/ice.h"
5
6 MODULE_ALIAS("i40iw");
7 MODULE_AUTHOR("Intel Corporation, <e1000-rdma@lists.sourceforge.net>");
8 MODULE_DESCRIPTION("Intel(R) Ethernet Protocol Driver for RDMA");
9 MODULE_LICENSE("Dual BSD/GPL");
10
11 static struct notifier_block irdma_inetaddr_notifier = {
12         .notifier_call = irdma_inetaddr_event
13 };
14
15 static struct notifier_block irdma_inetaddr6_notifier = {
16         .notifier_call = irdma_inet6addr_event
17 };
18
19 static struct notifier_block irdma_net_notifier = {
20         .notifier_call = irdma_net_event
21 };
22
23 static struct notifier_block irdma_netdevice_notifier = {
24         .notifier_call = irdma_netdevice_event
25 };
26
27 static void irdma_register_notifiers(void)
28 {
29         register_inetaddr_notifier(&irdma_inetaddr_notifier);
30         register_inet6addr_notifier(&irdma_inetaddr6_notifier);
31         register_netevent_notifier(&irdma_net_notifier);
32         register_netdevice_notifier(&irdma_netdevice_notifier);
33 }
34
35 static void irdma_unregister_notifiers(void)
36 {
37         unregister_netevent_notifier(&irdma_net_notifier);
38         unregister_inetaddr_notifier(&irdma_inetaddr_notifier);
39         unregister_inet6addr_notifier(&irdma_inetaddr6_notifier);
40         unregister_netdevice_notifier(&irdma_netdevice_notifier);
41 }
42
43 static void irdma_prep_tc_change(struct irdma_device *iwdev)
44 {
45         iwdev->vsi.tc_change_pending = true;
46         irdma_sc_suspend_resume_qps(&iwdev->vsi, IRDMA_OP_SUSPEND);
47
48         /* Wait for all qp's to suspend */
49         wait_event_timeout(iwdev->suspend_wq,
50                            !atomic_read(&iwdev->vsi.qp_suspend_reqs),
51                            msecs_to_jiffies(IRDMA_EVENT_TIMEOUT_MS));
52         irdma_ws_reset(&iwdev->vsi);
53 }
54
55 static void irdma_log_invalid_mtu(u16 mtu, struct irdma_sc_dev *dev)
56 {
57         if (mtu < IRDMA_MIN_MTU_IPV4)
58                 ibdev_warn(to_ibdev(dev), "MTU setting [%d] too low for RDMA traffic. Minimum MTU is 576 for IPv4\n", mtu);
59         else if (mtu < IRDMA_MIN_MTU_IPV6)
60                 ibdev_warn(to_ibdev(dev), "MTU setting [%d] too low for RDMA traffic. Minimum MTU is 1280 for IPv6\\n", mtu);
61 }
62
63 static void irdma_fill_qos_info(struct irdma_l2params *l2params,
64                                 struct iidc_qos_params *qos_info)
65 {
66         int i;
67
68         l2params->num_tc = qos_info->num_tc;
69         l2params->vsi_prio_type = qos_info->vport_priority_type;
70         l2params->vsi_rel_bw = qos_info->vport_relative_bw;
71         for (i = 0; i < l2params->num_tc; i++) {
72                 l2params->tc_info[i].egress_virt_up =
73                         qos_info->tc_info[i].egress_virt_up;
74                 l2params->tc_info[i].ingress_virt_up =
75                         qos_info->tc_info[i].ingress_virt_up;
76                 l2params->tc_info[i].prio_type = qos_info->tc_info[i].prio_type;
77                 l2params->tc_info[i].rel_bw = qos_info->tc_info[i].rel_bw;
78                 l2params->tc_info[i].tc_ctx = qos_info->tc_info[i].tc_ctx;
79         }
80         for (i = 0; i < IIDC_MAX_USER_PRIORITY; i++)
81                 l2params->up2tc[i] = qos_info->up2tc[i];
82         if (qos_info->pfc_mode == IIDC_DSCP_PFC_MODE) {
83                 l2params->dscp_mode = true;
84                 memcpy(l2params->dscp_map, qos_info->dscp_map, sizeof(l2params->dscp_map));
85         }
86 }
87
88 static void irdma_iidc_event_handler(struct ice_pf *pf, struct iidc_event *event)
89 {
90         struct irdma_device *iwdev = dev_get_drvdata(&pf->adev->dev);
91         struct irdma_l2params l2params = {};
92
93         if (*event->type & BIT(IIDC_EVENT_AFTER_MTU_CHANGE)) {
94                 ibdev_dbg(&iwdev->ibdev, "CLNT: new MTU = %d\n", iwdev->netdev->mtu);
95                 if (iwdev->vsi.mtu != iwdev->netdev->mtu) {
96                         l2params.mtu = iwdev->netdev->mtu;
97                         l2params.mtu_changed = true;
98                         irdma_log_invalid_mtu(l2params.mtu, &iwdev->rf->sc_dev);
99                         irdma_change_l2params(&iwdev->vsi, &l2params);
100                 }
101         } else if (*event->type & BIT(IIDC_EVENT_BEFORE_TC_CHANGE)) {
102                 if (iwdev->vsi.tc_change_pending)
103                         return;
104
105                 irdma_prep_tc_change(iwdev);
106         } else if (*event->type & BIT(IIDC_EVENT_AFTER_TC_CHANGE)) {
107                 struct iidc_qos_params qos_info = {};
108
109                 if (!iwdev->vsi.tc_change_pending)
110                         return;
111
112                 l2params.tc_changed = true;
113                 ibdev_dbg(&iwdev->ibdev, "CLNT: TC Change\n");
114                 ice_get_qos_params(pf, &qos_info);
115                 irdma_fill_qos_info(&l2params, &qos_info);
116                 if (iwdev->rf->protocol_used != IRDMA_IWARP_PROTOCOL_ONLY)
117                         iwdev->dcb_vlan_mode = qos_info.num_tc > 1 && !l2params.dscp_mode;
118                 irdma_change_l2params(&iwdev->vsi, &l2params);
119         } else if (*event->type & BIT(IIDC_EVENT_CRIT_ERR)) {
120                 ibdev_warn(&iwdev->ibdev, "ICE OICR event notification: oicr = 0x%08x\n",
121                            event->reg);
122                 if (event->reg & IRDMAPFINT_OICR_PE_CRITERR_M) {
123                         u32 pe_criterr;
124
125                         pe_criterr = readl(iwdev->rf->sc_dev.hw_regs[IRDMA_GLPE_CRITERR]);
126 #define IRDMA_Q1_RESOURCE_ERR 0x0001024d
127                         if (pe_criterr != IRDMA_Q1_RESOURCE_ERR) {
128                                 ibdev_err(&iwdev->ibdev, "critical PE Error, GLPE_CRITERR=0x%08x\n",
129                                           pe_criterr);
130                                 iwdev->rf->reset = true;
131                         } else {
132                                 ibdev_warn(&iwdev->ibdev, "Q1 Resource Check\n");
133                         }
134                 }
135                 if (event->reg & IRDMAPFINT_OICR_HMC_ERR_M) {
136                         ibdev_err(&iwdev->ibdev, "HMC Error\n");
137                         iwdev->rf->reset = true;
138                 }
139                 if (event->reg & IRDMAPFINT_OICR_PE_PUSH_M) {
140                         ibdev_err(&iwdev->ibdev, "PE Push Error\n");
141                         iwdev->rf->reset = true;
142                 }
143                 if (iwdev->rf->reset)
144                         iwdev->rf->gen_ops.request_reset(iwdev->rf);
145         }
146 }
147
148 /**
149  * irdma_request_reset - Request a reset
150  * @rf: RDMA PCI function
151  */
152 static void irdma_request_reset(struct irdma_pci_f *rf)
153 {
154         struct ice_pf *pf = rf->cdev;
155
156         ibdev_warn(&rf->iwdev->ibdev, "Requesting a reset\n");
157         ice_rdma_request_reset(pf, IIDC_PFR);
158 }
159
160 /**
161  * irdma_lan_register_qset - Register qset with LAN driver
162  * @vsi: vsi structure
163  * @tc_node: Traffic class node
164  */
165 static int irdma_lan_register_qset(struct irdma_sc_vsi *vsi,
166                                    struct irdma_ws_node *tc_node)
167 {
168         struct irdma_device *iwdev = vsi->back_vsi;
169         struct ice_pf *pf = iwdev->rf->cdev;
170         struct iidc_rdma_qset_params qset = {};
171         int ret;
172
173         qset.qs_handle = tc_node->qs_handle;
174         qset.tc = tc_node->traffic_class;
175         qset.vport_id = vsi->vsi_idx;
176         ret = ice_add_rdma_qset(pf, &qset);
177         if (ret) {
178                 ibdev_dbg(&iwdev->ibdev, "WS: LAN alloc_res for rdma qset failed.\n");
179                 return ret;
180         }
181
182         tc_node->l2_sched_node_id = qset.teid;
183         vsi->qos[tc_node->user_pri].l2_sched_node_id = qset.teid;
184
185         return 0;
186 }
187
188 /**
189  * irdma_lan_unregister_qset - Unregister qset with LAN driver
190  * @vsi: vsi structure
191  * @tc_node: Traffic class node
192  */
193 static void irdma_lan_unregister_qset(struct irdma_sc_vsi *vsi,
194                                       struct irdma_ws_node *tc_node)
195 {
196         struct irdma_device *iwdev = vsi->back_vsi;
197         struct ice_pf *pf = iwdev->rf->cdev;
198         struct iidc_rdma_qset_params qset = {};
199
200         qset.qs_handle = tc_node->qs_handle;
201         qset.tc = tc_node->traffic_class;
202         qset.vport_id = vsi->vsi_idx;
203         qset.teid = tc_node->l2_sched_node_id;
204
205         if (ice_del_rdma_qset(pf, &qset))
206                 ibdev_dbg(&iwdev->ibdev, "WS: LAN free_res for rdma qset failed.\n");
207 }
208
209 static void irdma_remove(struct auxiliary_device *aux_dev)
210 {
211         struct iidc_auxiliary_dev *iidc_adev = container_of(aux_dev,
212                                                             struct iidc_auxiliary_dev,
213                                                             adev);
214         struct ice_pf *pf = iidc_adev->pf;
215         struct irdma_device *iwdev = auxiliary_get_drvdata(aux_dev);
216
217         irdma_ib_unregister_device(iwdev);
218         ice_rdma_update_vsi_filter(pf, iwdev->vsi_num, false);
219
220         pr_debug("INIT: Gen2 PF[%d] device remove success\n", PCI_FUNC(pf->pdev->devfn));
221 }
222
223 static void irdma_fill_device_info(struct irdma_device *iwdev, struct ice_pf *pf,
224                                    struct ice_vsi *vsi)
225 {
226         struct irdma_pci_f *rf = iwdev->rf;
227
228         rf->cdev = pf;
229         rf->gen_ops.register_qset = irdma_lan_register_qset;
230         rf->gen_ops.unregister_qset = irdma_lan_unregister_qset;
231         rf->hw.hw_addr = pf->hw.hw_addr;
232         rf->pcidev = pf->pdev;
233         rf->msix_count =  pf->num_rdma_msix;
234         rf->pf_id = pf->hw.pf_id;
235         rf->msix_entries = &pf->msix_entries[pf->rdma_base_vector];
236         rf->default_vsi.vsi_idx = vsi->vsi_num;
237         rf->protocol_used = pf->rdma_mode & IIDC_RDMA_PROTOCOL_ROCEV2 ?
238                             IRDMA_ROCE_PROTOCOL_ONLY : IRDMA_IWARP_PROTOCOL_ONLY;
239         rf->rdma_ver = IRDMA_GEN_2;
240         rf->rsrc_profile = IRDMA_HMC_PROFILE_DEFAULT;
241         rf->rst_to = IRDMA_RST_TIMEOUT_HZ;
242         rf->gen_ops.request_reset = irdma_request_reset;
243         rf->limits_sel = 7;
244         rf->iwdev = iwdev;
245         mutex_init(&iwdev->ah_tbl_lock);
246         iwdev->netdev = vsi->netdev;
247         iwdev->vsi_num = vsi->vsi_num;
248         iwdev->init_state = INITIAL_STATE;
249         iwdev->roce_cwnd = IRDMA_ROCE_CWND_DEFAULT;
250         iwdev->roce_ackcreds = IRDMA_ROCE_ACKCREDS_DEFAULT;
251         iwdev->rcv_wnd = IRDMA_CM_DEFAULT_RCV_WND_SCALED;
252         iwdev->rcv_wscale = IRDMA_CM_DEFAULT_RCV_WND_SCALE;
253         if (rf->protocol_used == IRDMA_ROCE_PROTOCOL_ONLY)
254                 iwdev->roce_mode = true;
255 }
256
257 static int irdma_probe(struct auxiliary_device *aux_dev, const struct auxiliary_device_id *id)
258 {
259         struct iidc_auxiliary_dev *iidc_adev = container_of(aux_dev,
260                                                             struct iidc_auxiliary_dev,
261                                                             adev);
262         struct ice_pf *pf = iidc_adev->pf;
263         struct ice_vsi *vsi = ice_get_main_vsi(pf);
264         struct iidc_qos_params qos_info = {};
265         struct irdma_device *iwdev;
266         struct irdma_pci_f *rf;
267         struct irdma_l2params l2params = {};
268         int err;
269
270         if (!vsi)
271                 return -EIO;
272         iwdev = ib_alloc_device(irdma_device, ibdev);
273         if (!iwdev)
274                 return -ENOMEM;
275         iwdev->rf = kzalloc(sizeof(*rf), GFP_KERNEL);
276         if (!iwdev->rf) {
277                 ib_dealloc_device(&iwdev->ibdev);
278                 return -ENOMEM;
279         }
280
281         irdma_fill_device_info(iwdev, pf, vsi);
282         rf = iwdev->rf;
283
284         err = irdma_ctrl_init_hw(rf);
285         if (err)
286                 goto err_ctrl_init;
287
288         l2params.mtu = iwdev->netdev->mtu;
289         ice_get_qos_params(pf, &qos_info);
290         irdma_fill_qos_info(&l2params, &qos_info);
291         if (iwdev->rf->protocol_used != IRDMA_IWARP_PROTOCOL_ONLY)
292                 iwdev->dcb_vlan_mode = l2params.num_tc > 1 && !l2params.dscp_mode;
293
294         err = irdma_rt_init_hw(iwdev, &l2params);
295         if (err)
296                 goto err_rt_init;
297
298         err = irdma_ib_register_device(iwdev);
299         if (err)
300                 goto err_ibreg;
301
302         ice_rdma_update_vsi_filter(pf, iwdev->vsi_num, true);
303
304         ibdev_dbg(&iwdev->ibdev, "INIT: Gen2 PF[%d] device probe success\n", PCI_FUNC(rf->pcidev->devfn));
305         auxiliary_set_drvdata(aux_dev, iwdev);
306
307         return 0;
308
309 err_ibreg:
310         irdma_rt_deinit_hw(iwdev);
311 err_rt_init:
312         irdma_ctrl_deinit_hw(rf);
313 err_ctrl_init:
314         kfree(iwdev->rf);
315         ib_dealloc_device(&iwdev->ibdev);
316
317         return err;
318 }
319
320 static const struct auxiliary_device_id irdma_auxiliary_id_table[] = {
321         {.name = "ice.iwarp", },
322         {.name = "ice.roce", },
323         {},
324 };
325
326 MODULE_DEVICE_TABLE(auxiliary, irdma_auxiliary_id_table);
327
328 static struct iidc_auxiliary_drv irdma_auxiliary_drv = {
329         .adrv = {
330             .id_table = irdma_auxiliary_id_table,
331             .probe = irdma_probe,
332             .remove = irdma_remove,
333         },
334         .event_handler = irdma_iidc_event_handler,
335 };
336
337 static int __init irdma_init_module(void)
338 {
339         int ret;
340
341         ret = auxiliary_driver_register(&i40iw_auxiliary_drv);
342         if (ret) {
343                 pr_err("Failed i40iw(gen_1) auxiliary_driver_register() ret=%d\n",
344                        ret);
345                 return ret;
346         }
347
348         ret = auxiliary_driver_register(&irdma_auxiliary_drv.adrv);
349         if (ret) {
350                 auxiliary_driver_unregister(&i40iw_auxiliary_drv);
351                 pr_err("Failed irdma auxiliary_driver_register() ret=%d\n",
352                        ret);
353                 return ret;
354         }
355
356         irdma_register_notifiers();
357
358         return 0;
359 }
360
361 static void __exit irdma_exit_module(void)
362 {
363         irdma_unregister_notifiers();
364         auxiliary_driver_unregister(&irdma_auxiliary_drv.adrv);
365         auxiliary_driver_unregister(&i40iw_auxiliary_drv);
366 }
367
368 module_init(irdma_init_module);
369 module_exit(irdma_exit_module);