2 * Copyright 2018 Advanced Micro Devices, Inc.
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
23 #include <linux/kconfig.h>
25 #if IS_REACHABLE(CONFIG_AMD_IOMMU_V2)
27 #include <linux/printk.h>
28 #include <linux/device.h>
29 #include <linux/slab.h>
30 #include <linux/pci.h>
31 #include <linux/amd-iommu.h>
33 #include "kfd_dbgmgr.h"
34 #include "kfd_topology.h"
35 #include "kfd_iommu.h"
37 static const u32 required_iommu_flags = AMD_IOMMU_DEVICE_FLAG_ATS_SUP |
38 AMD_IOMMU_DEVICE_FLAG_PRI_SUP |
39 AMD_IOMMU_DEVICE_FLAG_PASID_SUP;
41 /** kfd_iommu_check_device - Check whether IOMMU is available for device
43 int kfd_iommu_check_device(struct kfd_dev *kfd)
45 struct amd_iommu_device_info iommu_info;
48 if (!kfd->device_info->needs_iommu_device)
52 err = amd_iommu_device_info(kfd->pdev, &iommu_info);
56 if ((iommu_info.flags & required_iommu_flags) != required_iommu_flags)
62 /** kfd_iommu_device_init - Initialize IOMMU for device
64 int kfd_iommu_device_init(struct kfd_dev *kfd)
66 struct amd_iommu_device_info iommu_info;
67 unsigned int pasid_limit;
69 struct kfd_topology_device *top_dev;
71 top_dev = kfd_topology_device_by_id(kfd->id);
74 * Overwrite ATS capability according to needs_iommu_device to fix
75 * potential missing corresponding bit in CRAT of BIOS.
77 if (!kfd->device_info->needs_iommu_device) {
78 top_dev->node_props.capability &= ~HSA_CAP_ATS_PRESENT;
82 top_dev->node_props.capability |= HSA_CAP_ATS_PRESENT;
85 err = amd_iommu_device_info(kfd->pdev, &iommu_info);
88 "error getting iommu info. is the iommu enabled?\n");
92 if ((iommu_info.flags & required_iommu_flags) != required_iommu_flags) {
94 "error required iommu flags ats %i, pri %i, pasid %i\n",
95 (iommu_info.flags & AMD_IOMMU_DEVICE_FLAG_ATS_SUP) != 0,
96 (iommu_info.flags & AMD_IOMMU_DEVICE_FLAG_PRI_SUP) != 0,
97 (iommu_info.flags & AMD_IOMMU_DEVICE_FLAG_PASID_SUP)
102 pasid_limit = min_t(unsigned int,
103 (unsigned int)(1 << kfd->device_info->max_pasid_bits),
104 iommu_info.max_pasids);
106 if (!kfd_set_pasid_limit(pasid_limit)) {
107 dev_err(kfd_device, "error setting pasid limit\n");
114 /** kfd_iommu_bind_process_to_device - Have the IOMMU bind a process
116 * Binds the given process to the given device using its PASID. This
117 * enables IOMMUv2 address translation for the process on the device.
119 * This function assumes that the process mutex is held.
121 int kfd_iommu_bind_process_to_device(struct kfd_process_device *pdd)
123 struct kfd_dev *dev = pdd->dev;
124 struct kfd_process *p = pdd->process;
127 if (!dev->device_info->needs_iommu_device || pdd->bound == PDD_BOUND)
130 if (unlikely(pdd->bound == PDD_BOUND_SUSPENDED)) {
131 pr_err("Binding PDD_BOUND_SUSPENDED pdd is unexpected!\n");
135 err = amd_iommu_bind_pasid(dev->pdev, p->pasid, p->lead_thread);
137 pdd->bound = PDD_BOUND;
142 /** kfd_iommu_unbind_process - Unbind process from all devices
144 * This removes all IOMMU device bindings of the process. To be used
145 * before process termination.
147 void kfd_iommu_unbind_process(struct kfd_process *p)
149 struct kfd_process_device *pdd;
151 list_for_each_entry(pdd, &p->per_device_data, per_device_list)
152 if (pdd->bound == PDD_BOUND)
153 amd_iommu_unbind_pasid(pdd->dev->pdev, p->pasid);
156 /* Callback for process shutdown invoked by the IOMMU driver */
157 static void iommu_pasid_shutdown_callback(struct pci_dev *pdev, int pasid)
159 struct kfd_dev *dev = kfd_device_by_pci_dev(pdev);
160 struct kfd_process *p;
161 struct kfd_process_device *pdd;
167 * Look for the process that matches the pasid. If there is no such
168 * process, we either released it in amdkfd's own notifier, or there
169 * is a bug. Unfortunately, there is no way to tell...
171 p = kfd_lookup_process_by_pasid(pasid);
175 pr_debug("Unbinding process %d from IOMMU\n", pasid);
177 mutex_lock(kfd_get_dbgmgr_mutex());
179 if (dev->dbgmgr && dev->dbgmgr->pasid == p->pasid) {
180 if (!kfd_dbgmgr_unregister(dev->dbgmgr, p)) {
181 kfd_dbgmgr_destroy(dev->dbgmgr);
186 mutex_unlock(kfd_get_dbgmgr_mutex());
188 mutex_lock(&p->mutex);
190 pdd = kfd_get_process_device_data(dev, p);
192 /* For GPU relying on IOMMU, we need to dequeue here
193 * when PASID is still bound.
195 kfd_process_dequeue_from_device(pdd);
197 mutex_unlock(&p->mutex);
199 kfd_unref_process(p);
202 /* This function called by IOMMU driver on PPR failure */
203 static int iommu_invalid_ppr_cb(struct pci_dev *pdev, int pasid,
204 unsigned long address, u16 flags)
208 dev_warn_ratelimited(kfd_device,
209 "Invalid PPR device %x:%x.%x pasid %d address 0x%lX flags 0x%X",
210 PCI_BUS_NUM(pdev->devfn),
211 PCI_SLOT(pdev->devfn),
212 PCI_FUNC(pdev->devfn),
217 dev = kfd_device_by_pci_dev(pdev);
219 kfd_signal_iommu_event(dev, pasid, address,
220 flags & PPR_FAULT_WRITE, flags & PPR_FAULT_EXEC);
222 return AMD_IOMMU_INV_PRI_RSP_INVALID;
226 * Bind processes do the device that have been temporarily unbound
227 * (PDD_BOUND_SUSPENDED) in kfd_unbind_processes_from_device.
229 static int kfd_bind_processes_to_device(struct kfd_dev *kfd)
231 struct kfd_process_device *pdd;
232 struct kfd_process *p;
236 int idx = srcu_read_lock(&kfd_processes_srcu);
238 hash_for_each_rcu(kfd_processes_table, temp, p, kfd_processes) {
239 mutex_lock(&p->mutex);
240 pdd = kfd_get_process_device_data(kfd, p);
242 if (WARN_ON(!pdd) || pdd->bound != PDD_BOUND_SUSPENDED) {
243 mutex_unlock(&p->mutex);
247 err = amd_iommu_bind_pasid(kfd->pdev, p->pasid,
250 pr_err("Unexpected pasid %d binding failure\n",
252 mutex_unlock(&p->mutex);
256 pdd->bound = PDD_BOUND;
257 mutex_unlock(&p->mutex);
260 srcu_read_unlock(&kfd_processes_srcu, idx);
266 * Mark currently bound processes as PDD_BOUND_SUSPENDED. These
267 * processes will be restored to PDD_BOUND state in
268 * kfd_bind_processes_to_device.
270 static void kfd_unbind_processes_from_device(struct kfd_dev *kfd)
272 struct kfd_process_device *pdd;
273 struct kfd_process *p;
276 int idx = srcu_read_lock(&kfd_processes_srcu);
278 hash_for_each_rcu(kfd_processes_table, temp, p, kfd_processes) {
279 mutex_lock(&p->mutex);
280 pdd = kfd_get_process_device_data(kfd, p);
283 mutex_unlock(&p->mutex);
287 if (pdd->bound == PDD_BOUND)
288 pdd->bound = PDD_BOUND_SUSPENDED;
289 mutex_unlock(&p->mutex);
292 srcu_read_unlock(&kfd_processes_srcu, idx);
295 /** kfd_iommu_suspend - Prepare IOMMU for suspend
297 * This unbinds processes from the device and disables the IOMMU for
300 void kfd_iommu_suspend(struct kfd_dev *kfd)
302 if (!kfd->device_info->needs_iommu_device)
305 kfd_unbind_processes_from_device(kfd);
307 amd_iommu_set_invalidate_ctx_cb(kfd->pdev, NULL);
308 amd_iommu_set_invalid_ppr_cb(kfd->pdev, NULL);
309 amd_iommu_free_device(kfd->pdev);
312 /** kfd_iommu_resume - Restore IOMMU after resume
314 * This reinitializes the IOMMU for the device and re-binds previously
315 * suspended processes to the device.
317 int kfd_iommu_resume(struct kfd_dev *kfd)
319 unsigned int pasid_limit;
322 if (!kfd->device_info->needs_iommu_device)
325 pasid_limit = kfd_get_pasid_limit();
327 err = amd_iommu_init_device(kfd->pdev, pasid_limit);
331 amd_iommu_set_invalidate_ctx_cb(kfd->pdev,
332 iommu_pasid_shutdown_callback);
333 amd_iommu_set_invalid_ppr_cb(kfd->pdev,
334 iommu_invalid_ppr_cb);
336 err = kfd_bind_processes_to_device(kfd);
338 amd_iommu_set_invalidate_ctx_cb(kfd->pdev, NULL);
339 amd_iommu_set_invalid_ppr_cb(kfd->pdev, NULL);
340 amd_iommu_free_device(kfd->pdev);
347 extern bool amd_iommu_pc_supported(void);
348 extern u8 amd_iommu_pc_get_max_banks(u16 devid);
349 extern u8 amd_iommu_pc_get_max_counters(u16 devid);
351 /** kfd_iommu_add_perf_counters - Add IOMMU performance counters to topology
353 int kfd_iommu_add_perf_counters(struct kfd_topology_device *kdev)
355 struct kfd_perf_properties *props;
357 if (!(kdev->node_props.capability & HSA_CAP_ATS_PRESENT))
360 if (!amd_iommu_pc_supported())
363 props = kfd_alloc_struct(props);
366 strcpy(props->block_name, "iommu");
367 props->max_concurrent = amd_iommu_pc_get_max_banks(0) *
368 amd_iommu_pc_get_max_counters(0); /* assume one iommu */
369 list_add_tail(&props->list, &kdev->perf_props);