Linux kernel mirror (for testing)
git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel
os
linux
1// SPDX-License-Identifier: GPL-2.0-only
2#include <stdio.h>
3#include <sys/mman.h>
4#include <unistd.h>
5
6#include <linux/iommufd.h>
7#include <linux/limits.h>
8#include <linux/mman.h>
9#include <linux/sizes.h>
10#include <linux/vfio.h>
11
12#include <libvfio.h>
13
14#include "kselftest_harness.h"
15
16static const char *device_bdf;
17
18struct iommu_mapping {
19 u64 pgd;
20 u64 p4d;
21 u64 pud;
22 u64 pmd;
23 u64 pte;
24};
25
26static void parse_next_value(char **line, u64 *value)
27{
28 char *token;
29
30 token = strtok_r(*line, " \t|\n", line);
31 if (!token)
32 return;
33
34 /* Caller verifies `value`. No need to check return value. */
35 sscanf(token, "0x%lx", value);
36}
37
38static int intel_iommu_mapping_get(const char *bdf, u64 iova,
39 struct iommu_mapping *mapping)
40{
41 char iommu_mapping_path[PATH_MAX], line[PATH_MAX];
42 u64 line_iova = -1;
43 int ret = -ENOENT;
44 FILE *file;
45 char *rest;
46
47 snprintf(iommu_mapping_path, sizeof(iommu_mapping_path),
48 "/sys/kernel/debug/iommu/intel/%s/domain_translation_struct",
49 bdf);
50
51 printf("Searching for IOVA 0x%lx in %s\n", iova, iommu_mapping_path);
52
53 file = fopen(iommu_mapping_path, "r");
54 VFIO_ASSERT_NOT_NULL(file, "fopen(%s) failed", iommu_mapping_path);
55
56 while (fgets(line, sizeof(line), file)) {
57 rest = line;
58
59 parse_next_value(&rest, &line_iova);
60 if (line_iova != (iova / getpagesize()))
61 continue;
62
63 /*
64 * Ensure each struct field is initialized in case of empty
65 * page table values.
66 */
67 memset(mapping, 0, sizeof(*mapping));
68 parse_next_value(&rest, &mapping->pgd);
69 parse_next_value(&rest, &mapping->p4d);
70 parse_next_value(&rest, &mapping->pud);
71 parse_next_value(&rest, &mapping->pmd);
72 parse_next_value(&rest, &mapping->pte);
73
74 ret = 0;
75 break;
76 }
77
78 fclose(file);
79
80 if (ret)
81 printf("IOVA not found\n");
82
83 return ret;
84}
85
86static int iommu_mapping_get(const char *bdf, u64 iova,
87 struct iommu_mapping *mapping)
88{
89 if (!access("/sys/kernel/debug/iommu/intel", F_OK))
90 return intel_iommu_mapping_get(bdf, iova, mapping);
91
92 return -EOPNOTSUPP;
93}
94
95FIXTURE(vfio_dma_mapping_test) {
96 struct iommu *iommu;
97 struct vfio_pci_device *device;
98 struct iova_allocator *iova_allocator;
99};
100
101FIXTURE_VARIANT(vfio_dma_mapping_test) {
102 const char *iommu_mode;
103 u64 size;
104 int mmap_flags;
105};
106
107#define FIXTURE_VARIANT_ADD_IOMMU_MODE(_iommu_mode, _name, _size, _mmap_flags) \
108FIXTURE_VARIANT_ADD(vfio_dma_mapping_test, _iommu_mode ## _ ## _name) { \
109 .iommu_mode = #_iommu_mode, \
110 .size = (_size), \
111 .mmap_flags = MAP_ANONYMOUS | MAP_PRIVATE | (_mmap_flags), \
112}
113
114FIXTURE_VARIANT_ADD_ALL_IOMMU_MODES(anonymous, 0, 0);
115FIXTURE_VARIANT_ADD_ALL_IOMMU_MODES(anonymous_hugetlb_2mb, SZ_2M, MAP_HUGETLB | MAP_HUGE_2MB);
116FIXTURE_VARIANT_ADD_ALL_IOMMU_MODES(anonymous_hugetlb_1gb, SZ_1G, MAP_HUGETLB | MAP_HUGE_1GB);
117
118#undef FIXTURE_VARIANT_ADD_IOMMU_MODE
119
120FIXTURE_SETUP(vfio_dma_mapping_test)
121{
122 self->iommu = iommu_init(variant->iommu_mode);
123 self->device = vfio_pci_device_init(device_bdf, self->iommu);
124 self->iova_allocator = iova_allocator_init(self->iommu);
125}
126
127FIXTURE_TEARDOWN(vfio_dma_mapping_test)
128{
129 iova_allocator_cleanup(self->iova_allocator);
130 vfio_pci_device_cleanup(self->device);
131 iommu_cleanup(self->iommu);
132}
133
134TEST_F(vfio_dma_mapping_test, dma_map_unmap)
135{
136 const u64 size = variant->size ?: getpagesize();
137 const int flags = variant->mmap_flags;
138 struct dma_region region;
139 struct iommu_mapping mapping;
140 u64 mapping_size = size;
141 u64 unmapped;
142 int rc;
143
144 region.vaddr = mmap(NULL, size, PROT_READ | PROT_WRITE, flags, -1, 0);
145
146 /* Skip the test if there aren't enough HugeTLB pages available. */
147 if (flags & MAP_HUGETLB && region.vaddr == MAP_FAILED)
148 SKIP(return, "mmap() failed: %s (%d)\n", strerror(errno), errno);
149 else
150 ASSERT_NE(region.vaddr, MAP_FAILED);
151
152 region.iova = iova_allocator_alloc(self->iova_allocator, size);
153 region.size = size;
154
155 iommu_map(self->iommu, ®ion);
156 printf("Mapped HVA %p (size 0x%lx) at IOVA 0x%lx\n", region.vaddr, size, region.iova);
157
158 ASSERT_EQ(region.iova, to_iova(self->device, region.vaddr));
159
160 rc = iommu_mapping_get(device_bdf, region.iova, &mapping);
161 if (rc == -EOPNOTSUPP)
162 goto unmap;
163
164 /*
165 * IOMMUFD compatibility-mode does not support huge mappings when
166 * using VFIO_TYPE1_IOMMU.
167 */
168 if (!strcmp(variant->iommu_mode, "iommufd_compat_type1"))
169 mapping_size = SZ_4K;
170
171 ASSERT_EQ(0, rc);
172 printf("Found IOMMU mappings for IOVA 0x%lx:\n", region.iova);
173 printf("PGD: 0x%016lx\n", mapping.pgd);
174 printf("P4D: 0x%016lx\n", mapping.p4d);
175 printf("PUD: 0x%016lx\n", mapping.pud);
176 printf("PMD: 0x%016lx\n", mapping.pmd);
177 printf("PTE: 0x%016lx\n", mapping.pte);
178
179 switch (mapping_size) {
180 case SZ_4K:
181 ASSERT_NE(0, mapping.pte);
182 break;
183 case SZ_2M:
184 ASSERT_EQ(0, mapping.pte);
185 ASSERT_NE(0, mapping.pmd);
186 break;
187 case SZ_1G:
188 ASSERT_EQ(0, mapping.pte);
189 ASSERT_EQ(0, mapping.pmd);
190 ASSERT_NE(0, mapping.pud);
191 break;
192 default:
193 VFIO_FAIL("Unrecognized size: 0x%lx\n", mapping_size);
194 }
195
196unmap:
197 rc = __iommu_unmap(self->iommu, ®ion, &unmapped);
198 ASSERT_EQ(rc, 0);
199 ASSERT_EQ(unmapped, region.size);
200 printf("Unmapped IOVA 0x%lx\n", region.iova);
201 ASSERT_NE(0, __to_iova(self->device, region.vaddr, NULL));
202 ASSERT_NE(0, iommu_mapping_get(device_bdf, region.iova, &mapping));
203
204 ASSERT_TRUE(!munmap(region.vaddr, size));
205}
206
207FIXTURE(vfio_dma_map_limit_test) {
208 struct iommu *iommu;
209 struct vfio_pci_device *device;
210 struct dma_region region;
211 size_t mmap_size;
212};
213
214FIXTURE_VARIANT(vfio_dma_map_limit_test) {
215 const char *iommu_mode;
216};
217
218#define FIXTURE_VARIANT_ADD_IOMMU_MODE(_iommu_mode) \
219FIXTURE_VARIANT_ADD(vfio_dma_map_limit_test, _iommu_mode) { \
220 .iommu_mode = #_iommu_mode, \
221}
222
223FIXTURE_VARIANT_ADD_ALL_IOMMU_MODES();
224
225#undef FIXTURE_VARIANT_ADD_IOMMU_MODE
226
227FIXTURE_SETUP(vfio_dma_map_limit_test)
228{
229 struct dma_region *region = &self->region;
230 struct iommu_iova_range *ranges;
231 u64 region_size = getpagesize();
232 iova_t last_iova;
233 u32 nranges;
234
235 /*
236 * Over-allocate mmap by double the size to provide enough backing vaddr
237 * for overflow tests
238 */
239 self->mmap_size = 2 * region_size;
240
241 self->iommu = iommu_init(variant->iommu_mode);
242 self->device = vfio_pci_device_init(device_bdf, self->iommu);
243 region->vaddr = mmap(NULL, self->mmap_size, PROT_READ | PROT_WRITE,
244 MAP_ANONYMOUS | MAP_PRIVATE, -1, 0);
245 ASSERT_NE(region->vaddr, MAP_FAILED);
246
247 ranges = iommu_iova_ranges(self->iommu, &nranges);
248 VFIO_ASSERT_NOT_NULL(ranges);
249 last_iova = ranges[nranges - 1].last;
250 free(ranges);
251
252 /* One page prior to the last iova */
253 region->iova = last_iova & ~(region_size - 1);
254 region->size = region_size;
255}
256
257FIXTURE_TEARDOWN(vfio_dma_map_limit_test)
258{
259 vfio_pci_device_cleanup(self->device);
260 iommu_cleanup(self->iommu);
261 ASSERT_EQ(munmap(self->region.vaddr, self->mmap_size), 0);
262}
263
264TEST_F(vfio_dma_map_limit_test, unmap_range)
265{
266 struct dma_region *region = &self->region;
267 u64 unmapped;
268 int rc;
269
270 iommu_map(self->iommu, region);
271 ASSERT_EQ(region->iova, to_iova(self->device, region->vaddr));
272
273 rc = __iommu_unmap(self->iommu, region, &unmapped);
274 ASSERT_EQ(rc, 0);
275 ASSERT_EQ(unmapped, region->size);
276}
277
278TEST_F(vfio_dma_map_limit_test, unmap_all)
279{
280 struct dma_region *region = &self->region;
281 u64 unmapped;
282 int rc;
283
284 iommu_map(self->iommu, region);
285 ASSERT_EQ(region->iova, to_iova(self->device, region->vaddr));
286
287 rc = __iommu_unmap_all(self->iommu, &unmapped);
288 ASSERT_EQ(rc, 0);
289 ASSERT_EQ(unmapped, region->size);
290}
291
292TEST_F(vfio_dma_map_limit_test, overflow)
293{
294 struct dma_region *region = &self->region;
295 int rc;
296
297 region->iova = ~(iova_t)0 & ~(region->size - 1);
298 region->size = self->mmap_size;
299
300 rc = __iommu_map(self->iommu, region);
301 ASSERT_EQ(rc, -EOVERFLOW);
302
303 rc = __iommu_unmap(self->iommu, region, NULL);
304 ASSERT_EQ(rc, -EOVERFLOW);
305}
306
307int main(int argc, char *argv[])
308{
309 device_bdf = vfio_selftests_get_bdf(&argc, argv);
310 return test_harness_run(argc, argv);
311}