1 // SPDX-License-Identifier: GPL-2.0 OR MIT
2 /*
3  * Copyright 2018-2022 Advanced Micro Devices, Inc.
4  *
5  * Permission is hereby granted, free of charge, to any person obtaining a
6  * copy of this software and associated documentation files (the "Software"),
7  * to deal in the Software without restriction, including without limitation
8  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
9  * and/or sell copies of the Software, and to permit persons to whom the
10  * Software is furnished to do so, subject to the following conditions:
11  *
12  * The above copyright notice and this permission notice shall be included in
13  * all copies or substantial portions of the Software.
14  *
15  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
18  * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
19  * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
20  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
21  * OTHER DEALINGS IN THE SOFTWARE.
22  */
23 
24 #include <linux/kconfig.h>
25 
26 #if IS_REACHABLE(CONFIG_AMD_IOMMU_V2)
27 
28 #include <linux/printk.h>
29 #include <linux/device.h>
30 #include <linux/slab.h>
31 #include <linux/pci.h>
32 #include <linux/amd-iommu.h>
33 #include "kfd_priv.h"
34 #include "kfd_topology.h"
35 #include "kfd_iommu.h"
36 
37 static const u32 required_iommu_flags = AMD_IOMMU_DEVICE_FLAG_ATS_SUP |
38 					AMD_IOMMU_DEVICE_FLAG_PRI_SUP |
39 					AMD_IOMMU_DEVICE_FLAG_PASID_SUP;
40 
41 /** kfd_iommu_check_device - Check whether IOMMU is available for device
42  */
kfd_iommu_check_device(struct kfd_dev * kfd)43 int kfd_iommu_check_device(struct kfd_dev *kfd)
44 {
45 	struct amd_iommu_device_info iommu_info;
46 	int err;
47 
48 	if (!kfd->use_iommu_v2)
49 		return -ENODEV;
50 
51 	iommu_info.flags = 0;
52 	err = amd_iommu_device_info(kfd->adev->pdev, &iommu_info);
53 	if (err)
54 		return err;
55 
56 	if ((iommu_info.flags & required_iommu_flags) != required_iommu_flags)
57 		return -ENODEV;
58 
59 	return 0;
60 }
61 
62 /** kfd_iommu_device_init - Initialize IOMMU for device
63  */
kfd_iommu_device_init(struct kfd_dev * kfd)64 int kfd_iommu_device_init(struct kfd_dev *kfd)
65 {
66 	struct amd_iommu_device_info iommu_info;
67 	unsigned int pasid_limit;
68 	int err;
69 
70 	if (!kfd->use_iommu_v2)
71 		return 0;
72 
73 	iommu_info.flags = 0;
74 	err = amd_iommu_device_info(kfd->adev->pdev, &iommu_info);
75 	if (err < 0) {
76 		dev_err(kfd_device,
77 			"error getting iommu info. is the iommu enabled?\n");
78 		return -ENODEV;
79 	}
80 
81 	if ((iommu_info.flags & required_iommu_flags) != required_iommu_flags) {
82 		dev_err(kfd_device,
83 			"error required iommu flags ats %i, pri %i, pasid %i\n",
84 		       (iommu_info.flags & AMD_IOMMU_DEVICE_FLAG_ATS_SUP) != 0,
85 		       (iommu_info.flags & AMD_IOMMU_DEVICE_FLAG_PRI_SUP) != 0,
86 		       (iommu_info.flags & AMD_IOMMU_DEVICE_FLAG_PASID_SUP)
87 									!= 0);
88 		return -ENODEV;
89 	}
90 
91 	pasid_limit = min_t(unsigned int,
92 			(unsigned int)(1 << kfd->device_info.max_pasid_bits),
93 			iommu_info.max_pasids);
94 
95 	if (!kfd_set_pasid_limit(pasid_limit)) {
96 		dev_err(kfd_device, "error setting pasid limit\n");
97 		return -EBUSY;
98 	}
99 
100 	return 0;
101 }
102 
103 /** kfd_iommu_bind_process_to_device - Have the IOMMU bind a process
104  *
105  * Binds the given process to the given device using its PASID. This
106  * enables IOMMUv2 address translation for the process on the device.
107  *
108  * This function assumes that the process mutex is held.
109  */
kfd_iommu_bind_process_to_device(struct kfd_process_device * pdd)110 int kfd_iommu_bind_process_to_device(struct kfd_process_device *pdd)
111 {
112 	struct kfd_dev *dev = pdd->dev;
113 	struct kfd_process *p = pdd->process;
114 	int err;
115 
116 	if (!dev->use_iommu_v2 || pdd->bound == PDD_BOUND)
117 		return 0;
118 
119 	if (unlikely(pdd->bound == PDD_BOUND_SUSPENDED)) {
120 		pr_err("Binding PDD_BOUND_SUSPENDED pdd is unexpected!\n");
121 		return -EINVAL;
122 	}
123 
124 	err = amd_iommu_bind_pasid(dev->adev->pdev, p->pasid, p->lead_thread);
125 	if (!err)
126 		pdd->bound = PDD_BOUND;
127 
128 	return err;
129 }
130 
131 /** kfd_iommu_unbind_process - Unbind process from all devices
132  *
133  * This removes all IOMMU device bindings of the process. To be used
134  * before process termination.
135  */
kfd_iommu_unbind_process(struct kfd_process * p)136 void kfd_iommu_unbind_process(struct kfd_process *p)
137 {
138 	int i;
139 
140 	for (i = 0; i < p->n_pdds; i++)
141 		if (p->pdds[i]->bound == PDD_BOUND)
142 			amd_iommu_unbind_pasid(p->pdds[i]->dev->adev->pdev,
143 					       p->pasid);
144 }
145 
146 /* Callback for process shutdown invoked by the IOMMU driver */
iommu_pasid_shutdown_callback(struct pci_dev * pdev,u32 pasid)147 static void iommu_pasid_shutdown_callback(struct pci_dev *pdev, u32 pasid)
148 {
149 	struct kfd_dev *dev = kfd_device_by_pci_dev(pdev);
150 	struct kfd_process *p;
151 	struct kfd_process_device *pdd;
152 
153 	if (!dev)
154 		return;
155 
156 	/*
157 	 * Look for the process that matches the pasid. If there is no such
158 	 * process, we either released it in amdkfd's own notifier, or there
159 	 * is a bug. Unfortunately, there is no way to tell...
160 	 */
161 	p = kfd_lookup_process_by_pasid(pasid);
162 	if (!p)
163 		return;
164 
165 	pr_debug("Unbinding process 0x%x from IOMMU\n", pasid);
166 
167 	mutex_lock(&p->mutex);
168 
169 	pdd = kfd_get_process_device_data(dev, p);
170 	if (pdd)
171 		/* For GPU relying on IOMMU, we need to dequeue here
172 		 * when PASID is still bound.
173 		 */
174 		kfd_process_dequeue_from_device(pdd);
175 
176 	mutex_unlock(&p->mutex);
177 
178 	kfd_unref_process(p);
179 }
180 
181 /* This function called by IOMMU driver on PPR failure */
iommu_invalid_ppr_cb(struct pci_dev * pdev,u32 pasid,unsigned long address,u16 flags)182 static int iommu_invalid_ppr_cb(struct pci_dev *pdev, u32 pasid,
183 				unsigned long address, u16 flags)
184 {
185 	struct kfd_dev *dev;
186 
187 	dev_warn_ratelimited(kfd_device,
188 			"Invalid PPR device %x:%x.%x pasid 0x%x address 0x%lX flags 0x%X",
189 			pdev->bus->number,
190 			PCI_SLOT(pdev->devfn),
191 			PCI_FUNC(pdev->devfn),
192 			pasid,
193 			address,
194 			flags);
195 
196 	dev = kfd_device_by_pci_dev(pdev);
197 	if (!WARN_ON(!dev))
198 		kfd_signal_iommu_event(dev, pasid, address,
199 			flags & PPR_FAULT_WRITE, flags & PPR_FAULT_EXEC);
200 
201 	return AMD_IOMMU_INV_PRI_RSP_INVALID;
202 }
203 
204 /*
205  * Bind processes do the device that have been temporarily unbound
206  * (PDD_BOUND_SUSPENDED) in kfd_unbind_processes_from_device.
207  */
kfd_bind_processes_to_device(struct kfd_dev * kfd)208 static int kfd_bind_processes_to_device(struct kfd_dev *kfd)
209 {
210 	struct kfd_process_device *pdd;
211 	struct kfd_process *p;
212 	unsigned int temp;
213 	int err = 0;
214 
215 	int idx = srcu_read_lock(&kfd_processes_srcu);
216 
217 	hash_for_each_rcu(kfd_processes_table, temp, p, kfd_processes) {
218 		mutex_lock(&p->mutex);
219 		pdd = kfd_get_process_device_data(kfd, p);
220 
221 		if (WARN_ON(!pdd) || pdd->bound != PDD_BOUND_SUSPENDED) {
222 			mutex_unlock(&p->mutex);
223 			continue;
224 		}
225 
226 		err = amd_iommu_bind_pasid(kfd->adev->pdev, p->pasid,
227 				p->lead_thread);
228 		if (err < 0) {
229 			pr_err("Unexpected pasid 0x%x binding failure\n",
230 					p->pasid);
231 			mutex_unlock(&p->mutex);
232 			break;
233 		}
234 
235 		pdd->bound = PDD_BOUND;
236 		mutex_unlock(&p->mutex);
237 	}
238 
239 	srcu_read_unlock(&kfd_processes_srcu, idx);
240 
241 	return err;
242 }
243 
244 /*
245  * Mark currently bound processes as PDD_BOUND_SUSPENDED. These
246  * processes will be restored to PDD_BOUND state in
247  * kfd_bind_processes_to_device.
248  */
kfd_unbind_processes_from_device(struct kfd_dev * kfd)249 static void kfd_unbind_processes_from_device(struct kfd_dev *kfd)
250 {
251 	struct kfd_process_device *pdd;
252 	struct kfd_process *p;
253 	unsigned int temp;
254 
255 	int idx = srcu_read_lock(&kfd_processes_srcu);
256 
257 	hash_for_each_rcu(kfd_processes_table, temp, p, kfd_processes) {
258 		mutex_lock(&p->mutex);
259 		pdd = kfd_get_process_device_data(kfd, p);
260 
261 		if (WARN_ON(!pdd)) {
262 			mutex_unlock(&p->mutex);
263 			continue;
264 		}
265 
266 		if (pdd->bound == PDD_BOUND)
267 			pdd->bound = PDD_BOUND_SUSPENDED;
268 		mutex_unlock(&p->mutex);
269 	}
270 
271 	srcu_read_unlock(&kfd_processes_srcu, idx);
272 }
273 
274 /** kfd_iommu_suspend - Prepare IOMMU for suspend
275  *
276  * This unbinds processes from the device and disables the IOMMU for
277  * the device.
278  */
kfd_iommu_suspend(struct kfd_dev * kfd)279 void kfd_iommu_suspend(struct kfd_dev *kfd)
280 {
281 	if (!kfd->use_iommu_v2)
282 		return;
283 
284 	kfd_unbind_processes_from_device(kfd);
285 
286 	amd_iommu_set_invalidate_ctx_cb(kfd->adev->pdev, NULL);
287 	amd_iommu_set_invalid_ppr_cb(kfd->adev->pdev, NULL);
288 	amd_iommu_free_device(kfd->adev->pdev);
289 }
290 
291 /** kfd_iommu_resume - Restore IOMMU after resume
292  *
293  * This reinitializes the IOMMU for the device and re-binds previously
294  * suspended processes to the device.
295  */
kfd_iommu_resume(struct kfd_dev * kfd)296 int kfd_iommu_resume(struct kfd_dev *kfd)
297 {
298 	unsigned int pasid_limit;
299 	int err;
300 
301 	if (!kfd->use_iommu_v2)
302 		return 0;
303 
304 	pasid_limit = kfd_get_pasid_limit();
305 
306 	err = amd_iommu_init_device(kfd->adev->pdev, pasid_limit);
307 	if (err)
308 		return -ENXIO;
309 
310 	amd_iommu_set_invalidate_ctx_cb(kfd->adev->pdev,
311 					iommu_pasid_shutdown_callback);
312 	amd_iommu_set_invalid_ppr_cb(kfd->adev->pdev,
313 				     iommu_invalid_ppr_cb);
314 
315 	err = kfd_bind_processes_to_device(kfd);
316 	if (err) {
317 		amd_iommu_set_invalidate_ctx_cb(kfd->adev->pdev, NULL);
318 		amd_iommu_set_invalid_ppr_cb(kfd->adev->pdev, NULL);
319 		amd_iommu_free_device(kfd->adev->pdev);
320 		return err;
321 	}
322 
323 	return 0;
324 }
325 
326 /** kfd_iommu_add_perf_counters - Add IOMMU performance counters to topology
327  */
kfd_iommu_add_perf_counters(struct kfd_topology_device * kdev)328 int kfd_iommu_add_perf_counters(struct kfd_topology_device *kdev)
329 {
330 	struct kfd_perf_properties *props;
331 
332 	if (!(kdev->node_props.capability & HSA_CAP_ATS_PRESENT))
333 		return 0;
334 
335 	if (!amd_iommu_pc_supported())
336 		return 0;
337 
338 	props = kfd_alloc_struct(props);
339 	if (!props)
340 		return -ENOMEM;
341 	strcpy(props->block_name, "iommu");
342 	props->max_concurrent = amd_iommu_pc_get_max_banks(0) *
343 		amd_iommu_pc_get_max_counters(0); /* assume one iommu */
344 	list_add_tail(&props->list, &kdev->perf_props);
345 
346 	return 0;
347 }
348 
349 #endif
350