Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

drm/suballoc: Extract amdgpu_sa.c as generic suballocation helper

Suballocating a buffer object is something that is not driver-specific
and useful for many drivers.

Use a slightly modified version of amdgpu_sa.c

v2:
- Style cleanups.
- Added / Modified documentation.
- Use u64 for the sizes and offset. The code dates back to 2012 and
using unsigned int will probably soon come back to bite us.
We can consider size_t as well for better 32-bit efficiency.
- Add and document gfp, intr and align arguments to drm_suballoc_new().
- Use drm_printer for debug output.

v3:
- Remove stale author info (Christian König)

v4:
- Avoid 64-bit integer divisions (kernel test robot <lkp@intel.com>)
- Use size_t rather than u64 for the managed range. (Thomas)

Signed-off-by: Maarten Lankhorst <maarten.lankhorst@linux.intel.com>
Co-developed-by: Thomas Hellström <thomas.hellstrom@linux.intel.com>
Signed-off-by: Thomas Hellström <thomas.hellstrom@linux.intel.com>
Reviewed-by: Christian König <christian.koenig@amd.com>
Acked-by: Maarten Lankhorst <maarten.lankhorst@linux.intel.com>
Acked-by: Alex Deucher <alexdeucher@gmail.com>
Acked-by: Christian König <christian.koenig@amd.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20230224095152.30134-2-thomas.hellstrom@linux.intel.com

authored by

Maarten Lankhorst and committed by
Thomas Hellström
849ee8a2 ca62297b

+572
+4
drivers/gpu/drm/Kconfig
··· 231 231 help 232 232 Choose this if you need the GEM shmem helper functions 233 233 234 + config DRM_SUBALLOC_HELPER 235 + tristate 236 + depends on DRM 237 + 234 238 config DRM_SCHED 235 239 tristate 236 240 depends on DRM
+3
drivers/gpu/drm/Makefile
··· 88 88 drm_shmem_helper-y := drm_gem_shmem_helper.o 89 89 obj-$(CONFIG_DRM_GEM_SHMEM_HELPER) += drm_shmem_helper.o 90 90 91 + drm_suballoc_helper-y := drm_suballoc.o 92 + obj-$(CONFIG_DRM_SUBALLOC_HELPER) += drm_suballoc_helper.o 93 + 91 94 drm_vram_helper-y := drm_gem_vram_helper.o 92 95 obj-$(CONFIG_DRM_VRAM_HELPER) += drm_vram_helper.o 93 96
+457
drivers/gpu/drm/drm_suballoc.c
··· 1 + // SPDX-License-Identifier: GPL-2.0 OR MIT 2 + /* 3 + * Copyright 2011 Red Hat Inc. 4 + * Copyright 2023 Intel Corporation. 5 + * All Rights Reserved. 6 + * 7 + * Permission is hereby granted, free of charge, to any person obtaining a 8 + * copy of this software and associated documentation files (the 9 + * "Software"), to deal in the Software without restriction, including 10 + * without limitation the rights to use, copy, modify, merge, publish, 11 + * distribute, sub license, and/or sell copies of the Software, and to 12 + * permit persons to whom the Software is furnished to do so, subject to 13 + * the following conditions: 14 + * 15 + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 16 + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 17 + * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL 18 + * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, 19 + * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR 20 + * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE 21 + * USE OR OTHER DEALINGS IN THE SOFTWARE. 22 + * 23 + * The above copyright notice and this permission notice (including the 24 + * next paragraph) shall be included in all copies or substantial portions 25 + * of the Software. 26 + * 27 + */ 28 + /* Algorithm: 29 + * 30 + * We store the last allocated bo in "hole", we always try to allocate 31 + * after the last allocated bo. Principle is that in a linear GPU ring 32 + * progression was is after last is the oldest bo we allocated and thus 33 + * the first one that should no longer be in use by the GPU. 34 + * 35 + * If it's not the case we skip over the bo after last to the closest 36 + * done bo if such one exist. If none exist and we are not asked to 37 + * block we report failure to allocate. 38 + * 39 + * If we are asked to block we wait on all the oldest fence of all 40 + * rings. We just wait for any of those fence to complete. 41 + */ 42 + 43 + #include <drm/drm_suballoc.h> 44 + #include <drm/drm_print.h> 45 + #include <linux/slab.h> 46 + #include <linux/sched.h> 47 + #include <linux/wait.h> 48 + #include <linux/dma-fence.h> 49 + 50 + static void drm_suballoc_remove_locked(struct drm_suballoc *sa); 51 + static void drm_suballoc_try_free(struct drm_suballoc_manager *sa_manager); 52 + 53 + /** 54 + * drm_suballoc_manager_init() - Initialise the drm_suballoc_manager 55 + * @sa_manager: pointer to the sa_manager 56 + * @size: number of bytes we want to suballocate 57 + * @align: alignment for each suballocated chunk 58 + * 59 + * Prepares the suballocation manager for suballocations. 60 + */ 61 + void drm_suballoc_manager_init(struct drm_suballoc_manager *sa_manager, 62 + size_t size, size_t align) 63 + { 64 + unsigned int i; 65 + 66 + BUILD_BUG_ON(!is_power_of_2(DRM_SUBALLOC_MAX_QUEUES)); 67 + 68 + if (!align) 69 + align = 1; 70 + 71 + /* alignment must be a power of 2 */ 72 + if (WARN_ON_ONCE(align & (align - 1))) 73 + align = roundup_pow_of_two(align); 74 + 75 + init_waitqueue_head(&sa_manager->wq); 76 + sa_manager->size = size; 77 + sa_manager->align = align; 78 + sa_manager->hole = &sa_manager->olist; 79 + INIT_LIST_HEAD(&sa_manager->olist); 80 + for (i = 0; i < DRM_SUBALLOC_MAX_QUEUES; ++i) 81 + INIT_LIST_HEAD(&sa_manager->flist[i]); 82 + } 83 + EXPORT_SYMBOL(drm_suballoc_manager_init); 84 + 85 + /** 86 + * drm_suballoc_manager_fini() - Destroy the drm_suballoc_manager 87 + * @sa_manager: pointer to the sa_manager 88 + * 89 + * Cleans up the suballocation manager after use. All fences added 90 + * with drm_suballoc_free() must be signaled, or we cannot clean up 91 + * the entire manager. 92 + */ 93 + void drm_suballoc_manager_fini(struct drm_suballoc_manager *sa_manager) 94 + { 95 + struct drm_suballoc *sa, *tmp; 96 + 97 + if (!sa_manager->size) 98 + return; 99 + 100 + if (!list_empty(&sa_manager->olist)) { 101 + sa_manager->hole = &sa_manager->olist; 102 + drm_suballoc_try_free(sa_manager); 103 + if (!list_empty(&sa_manager->olist)) 104 + DRM_ERROR("sa_manager is not empty, clearing anyway\n"); 105 + } 106 + list_for_each_entry_safe(sa, tmp, &sa_manager->olist, olist) { 107 + drm_suballoc_remove_locked(sa); 108 + } 109 + 110 + sa_manager->size = 0; 111 + } 112 + EXPORT_SYMBOL(drm_suballoc_manager_fini); 113 + 114 + static void drm_suballoc_remove_locked(struct drm_suballoc *sa) 115 + { 116 + struct drm_suballoc_manager *sa_manager = sa->manager; 117 + 118 + if (sa_manager->hole == &sa->olist) 119 + sa_manager->hole = sa->olist.prev; 120 + 121 + list_del_init(&sa->olist); 122 + list_del_init(&sa->flist); 123 + dma_fence_put(sa->fence); 124 + kfree(sa); 125 + } 126 + 127 + static void drm_suballoc_try_free(struct drm_suballoc_manager *sa_manager) 128 + { 129 + struct drm_suballoc *sa, *tmp; 130 + 131 + if (sa_manager->hole->next == &sa_manager->olist) 132 + return; 133 + 134 + sa = list_entry(sa_manager->hole->next, struct drm_suballoc, olist); 135 + list_for_each_entry_safe_from(sa, tmp, &sa_manager->olist, olist) { 136 + if (!sa->fence || !dma_fence_is_signaled(sa->fence)) 137 + return; 138 + 139 + drm_suballoc_remove_locked(sa); 140 + } 141 + } 142 + 143 + static size_t drm_suballoc_hole_soffset(struct drm_suballoc_manager *sa_manager) 144 + { 145 + struct list_head *hole = sa_manager->hole; 146 + 147 + if (hole != &sa_manager->olist) 148 + return list_entry(hole, struct drm_suballoc, olist)->eoffset; 149 + 150 + return 0; 151 + } 152 + 153 + static size_t drm_suballoc_hole_eoffset(struct drm_suballoc_manager *sa_manager) 154 + { 155 + struct list_head *hole = sa_manager->hole; 156 + 157 + if (hole->next != &sa_manager->olist) 158 + return list_entry(hole->next, struct drm_suballoc, olist)->soffset; 159 + return sa_manager->size; 160 + } 161 + 162 + static bool drm_suballoc_try_alloc(struct drm_suballoc_manager *sa_manager, 163 + struct drm_suballoc *sa, 164 + size_t size, size_t align) 165 + { 166 + size_t soffset, eoffset, wasted; 167 + 168 + soffset = drm_suballoc_hole_soffset(sa_manager); 169 + eoffset = drm_suballoc_hole_eoffset(sa_manager); 170 + wasted = round_up(soffset, align) - soffset; 171 + 172 + if ((eoffset - soffset) >= (size + wasted)) { 173 + soffset += wasted; 174 + 175 + sa->manager = sa_manager; 176 + sa->soffset = soffset; 177 + sa->eoffset = soffset + size; 178 + list_add(&sa->olist, sa_manager->hole); 179 + INIT_LIST_HEAD(&sa->flist); 180 + sa_manager->hole = &sa->olist; 181 + return true; 182 + } 183 + return false; 184 + } 185 + 186 + static bool __drm_suballoc_event(struct drm_suballoc_manager *sa_manager, 187 + size_t size, size_t align) 188 + { 189 + size_t soffset, eoffset, wasted; 190 + unsigned int i; 191 + 192 + for (i = 0; i < DRM_SUBALLOC_MAX_QUEUES; ++i) 193 + if (!list_empty(&sa_manager->flist[i])) 194 + return true; 195 + 196 + soffset = drm_suballoc_hole_soffset(sa_manager); 197 + eoffset = drm_suballoc_hole_eoffset(sa_manager); 198 + wasted = round_up(soffset, align) - soffset; 199 + 200 + return ((eoffset - soffset) >= (size + wasted)); 201 + } 202 + 203 + /** 204 + * drm_suballoc_event() - Check if we can stop waiting 205 + * @sa_manager: pointer to the sa_manager 206 + * @size: number of bytes we want to allocate 207 + * @align: alignment we need to match 208 + * 209 + * Return: true if either there is a fence we can wait for or 210 + * enough free memory to satisfy the allocation directly. 211 + * false otherwise. 212 + */ 213 + static bool drm_suballoc_event(struct drm_suballoc_manager *sa_manager, 214 + size_t size, size_t align) 215 + { 216 + bool ret; 217 + 218 + spin_lock(&sa_manager->wq.lock); 219 + ret = __drm_suballoc_event(sa_manager, size, align); 220 + spin_unlock(&sa_manager->wq.lock); 221 + return ret; 222 + } 223 + 224 + static bool drm_suballoc_next_hole(struct drm_suballoc_manager *sa_manager, 225 + struct dma_fence **fences, 226 + unsigned int *tries) 227 + { 228 + struct drm_suballoc *best_bo = NULL; 229 + unsigned int i, best_idx; 230 + size_t soffset, best, tmp; 231 + 232 + /* if hole points to the end of the buffer */ 233 + if (sa_manager->hole->next == &sa_manager->olist) { 234 + /* try again with its beginning */ 235 + sa_manager->hole = &sa_manager->olist; 236 + return true; 237 + } 238 + 239 + soffset = drm_suballoc_hole_soffset(sa_manager); 240 + /* to handle wrap around we add sa_manager->size */ 241 + best = sa_manager->size * 2; 242 + /* go over all fence list and try to find the closest sa 243 + * of the current last 244 + */ 245 + for (i = 0; i < DRM_SUBALLOC_MAX_QUEUES; ++i) { 246 + struct drm_suballoc *sa; 247 + 248 + fences[i] = NULL; 249 + 250 + if (list_empty(&sa_manager->flist[i])) 251 + continue; 252 + 253 + sa = list_first_entry(&sa_manager->flist[i], 254 + struct drm_suballoc, flist); 255 + 256 + if (!dma_fence_is_signaled(sa->fence)) { 257 + fences[i] = sa->fence; 258 + continue; 259 + } 260 + 261 + /* limit the number of tries each freelist gets */ 262 + if (tries[i] > 2) 263 + continue; 264 + 265 + tmp = sa->soffset; 266 + if (tmp < soffset) { 267 + /* wrap around, pretend it's after */ 268 + tmp += sa_manager->size; 269 + } 270 + tmp -= soffset; 271 + if (tmp < best) { 272 + /* this sa bo is the closest one */ 273 + best = tmp; 274 + best_idx = i; 275 + best_bo = sa; 276 + } 277 + } 278 + 279 + if (best_bo) { 280 + ++tries[best_idx]; 281 + sa_manager->hole = best_bo->olist.prev; 282 + 283 + /* 284 + * We know that this one is signaled, 285 + * so it's safe to remove it. 286 + */ 287 + drm_suballoc_remove_locked(best_bo); 288 + return true; 289 + } 290 + return false; 291 + } 292 + 293 + /** 294 + * drm_suballoc_new() - Make a suballocation. 295 + * @sa_manager: pointer to the sa_manager 296 + * @size: number of bytes we want to suballocate. 297 + * @gfp: gfp flags used for memory allocation. Typically GFP_KERNEL but 298 + * the argument is provided for suballocations from reclaim context or 299 + * where the caller wants to avoid pipelining rather than wait for 300 + * reclaim. 301 + * @intr: Whether to perform waits interruptible. This should typically 302 + * always be true, unless the caller needs to propagate a 303 + * non-interruptible context from above layers. 304 + * @align: Alignment. Must not exceed the default manager alignment. 305 + * If @align is zero, then the manager alignment is used. 306 + * 307 + * Try to make a suballocation of size @size, which will be rounded 308 + * up to the alignment specified in specified in drm_suballoc_manager_init(). 309 + * 310 + * Return: a new suballocated bo, or an ERR_PTR. 311 + */ 312 + struct drm_suballoc * 313 + drm_suballoc_new(struct drm_suballoc_manager *sa_manager, size_t size, 314 + gfp_t gfp, bool intr, size_t align) 315 + { 316 + struct dma_fence *fences[DRM_SUBALLOC_MAX_QUEUES]; 317 + unsigned int tries[DRM_SUBALLOC_MAX_QUEUES]; 318 + unsigned int count; 319 + int i, r; 320 + struct drm_suballoc *sa; 321 + 322 + if (WARN_ON_ONCE(align > sa_manager->align)) 323 + return ERR_PTR(-EINVAL); 324 + if (WARN_ON_ONCE(size > sa_manager->size || !size)) 325 + return ERR_PTR(-EINVAL); 326 + 327 + if (!align) 328 + align = sa_manager->align; 329 + 330 + sa = kmalloc(sizeof(*sa), gfp); 331 + if (!sa) 332 + return ERR_PTR(-ENOMEM); 333 + sa->manager = sa_manager; 334 + sa->fence = NULL; 335 + INIT_LIST_HEAD(&sa->olist); 336 + INIT_LIST_HEAD(&sa->flist); 337 + 338 + spin_lock(&sa_manager->wq.lock); 339 + do { 340 + for (i = 0; i < DRM_SUBALLOC_MAX_QUEUES; ++i) 341 + tries[i] = 0; 342 + 343 + do { 344 + drm_suballoc_try_free(sa_manager); 345 + 346 + if (drm_suballoc_try_alloc(sa_manager, sa, 347 + size, align)) { 348 + spin_unlock(&sa_manager->wq.lock); 349 + return sa; 350 + } 351 + 352 + /* see if we can skip over some allocations */ 353 + } while (drm_suballoc_next_hole(sa_manager, fences, tries)); 354 + 355 + for (i = 0, count = 0; i < DRM_SUBALLOC_MAX_QUEUES; ++i) 356 + if (fences[i]) 357 + fences[count++] = dma_fence_get(fences[i]); 358 + 359 + if (count) { 360 + long t; 361 + 362 + spin_unlock(&sa_manager->wq.lock); 363 + t = dma_fence_wait_any_timeout(fences, count, intr, 364 + MAX_SCHEDULE_TIMEOUT, 365 + NULL); 366 + for (i = 0; i < count; ++i) 367 + dma_fence_put(fences[i]); 368 + 369 + r = (t > 0) ? 0 : t; 370 + spin_lock(&sa_manager->wq.lock); 371 + } else if (intr) { 372 + /* if we have nothing to wait for block */ 373 + r = wait_event_interruptible_locked 374 + (sa_manager->wq, 375 + __drm_suballoc_event(sa_manager, size, align)); 376 + } else { 377 + spin_unlock(&sa_manager->wq.lock); 378 + wait_event(sa_manager->wq, 379 + drm_suballoc_event(sa_manager, size, align)); 380 + r = 0; 381 + spin_lock(&sa_manager->wq.lock); 382 + } 383 + } while (!r); 384 + 385 + spin_unlock(&sa_manager->wq.lock); 386 + kfree(sa); 387 + return ERR_PTR(r); 388 + } 389 + EXPORT_SYMBOL(drm_suballoc_new); 390 + 391 + /** 392 + * drm_suballoc_free - Free a suballocation 393 + * @suballoc: pointer to the suballocation 394 + * @fence: fence that signals when suballocation is idle 395 + * 396 + * Free the suballocation. The suballocation can be re-used after @fence signals. 397 + */ 398 + void drm_suballoc_free(struct drm_suballoc *suballoc, 399 + struct dma_fence *fence) 400 + { 401 + struct drm_suballoc_manager *sa_manager; 402 + 403 + if (!suballoc) 404 + return; 405 + 406 + sa_manager = suballoc->manager; 407 + 408 + spin_lock(&sa_manager->wq.lock); 409 + if (fence && !dma_fence_is_signaled(fence)) { 410 + u32 idx; 411 + 412 + suballoc->fence = dma_fence_get(fence); 413 + idx = fence->context & (DRM_SUBALLOC_MAX_QUEUES - 1); 414 + list_add_tail(&suballoc->flist, &sa_manager->flist[idx]); 415 + } else { 416 + drm_suballoc_remove_locked(suballoc); 417 + } 418 + wake_up_all_locked(&sa_manager->wq); 419 + spin_unlock(&sa_manager->wq.lock); 420 + } 421 + EXPORT_SYMBOL(drm_suballoc_free); 422 + 423 + #ifdef CONFIG_DEBUG_FS 424 + void drm_suballoc_dump_debug_info(struct drm_suballoc_manager *sa_manager, 425 + struct drm_printer *p, 426 + unsigned long long suballoc_base) 427 + { 428 + struct drm_suballoc *i; 429 + 430 + spin_lock(&sa_manager->wq.lock); 431 + list_for_each_entry(i, &sa_manager->olist, olist) { 432 + unsigned long long soffset = i->soffset; 433 + unsigned long long eoffset = i->eoffset; 434 + 435 + if (&i->olist == sa_manager->hole) 436 + drm_puts(p, ">"); 437 + else 438 + drm_puts(p, " "); 439 + 440 + drm_printf(p, "[0x%010llx 0x%010llx] size %8lld", 441 + suballoc_base + soffset, suballoc_base + eoffset, 442 + eoffset - soffset); 443 + 444 + if (i->fence) 445 + drm_printf(p, " protected by 0x%016llx on context %llu", 446 + (unsigned long long)i->fence->seqno, 447 + (unsigned long long)i->fence->context); 448 + 449 + drm_puts(p, "\n"); 450 + } 451 + spin_unlock(&sa_manager->wq.lock); 452 + } 453 + EXPORT_SYMBOL(drm_suballoc_dump_debug_info); 454 + #endif 455 + MODULE_AUTHOR("Multiple"); 456 + MODULE_DESCRIPTION("Range suballocator helper"); 457 + MODULE_LICENSE("Dual MIT/GPL");
+108
include/drm/drm_suballoc.h
··· 1 + /* SPDX-License-Identifier: GPL-2.0 OR MIT */ 2 + /* 3 + * Copyright 2011 Red Hat Inc. 4 + * Copyright © 2022 Intel Corporation 5 + */ 6 + #ifndef _DRM_SUBALLOC_H_ 7 + #define _DRM_SUBALLOC_H_ 8 + 9 + #include <drm/drm_mm.h> 10 + 11 + #include <linux/dma-fence.h> 12 + #include <linux/types.h> 13 + 14 + #define DRM_SUBALLOC_MAX_QUEUES 32 15 + /** 16 + * struct drm_suballoc_manager - fenced range allocations 17 + * @wq: Wait queue for sleeping allocations on contention. 18 + * @hole: Pointer to first hole node. 19 + * @olist: List of allocated ranges. 20 + * @flist: Array[fence context hash] of queues of fenced allocated ranges. 21 + * @size: Size of the managed range. 22 + * @align: Default alignment for the managed range. 23 + */ 24 + struct drm_suballoc_manager { 25 + wait_queue_head_t wq; 26 + struct list_head *hole; 27 + struct list_head olist; 28 + struct list_head flist[DRM_SUBALLOC_MAX_QUEUES]; 29 + size_t size; 30 + size_t align; 31 + }; 32 + 33 + /** 34 + * struct drm_suballoc - Sub-allocated range 35 + * @olist: List link for list of allocated ranges. 36 + * @flist: List linkk for the manager fenced allocated ranges queues. 37 + * @manager: The drm_suballoc_manager. 38 + * @soffset: Start offset. 39 + * @eoffset: End offset + 1 so that @eoffset - @soffset = size. 40 + * @dma_fence: The fence protecting the allocation. 41 + */ 42 + struct drm_suballoc { 43 + struct list_head olist; 44 + struct list_head flist; 45 + struct drm_suballoc_manager *manager; 46 + size_t soffset; 47 + size_t eoffset; 48 + struct dma_fence *fence; 49 + }; 50 + 51 + void drm_suballoc_manager_init(struct drm_suballoc_manager *sa_manager, 52 + size_t size, size_t align); 53 + 54 + void drm_suballoc_manager_fini(struct drm_suballoc_manager *sa_manager); 55 + 56 + struct drm_suballoc * 57 + drm_suballoc_new(struct drm_suballoc_manager *sa_manager, size_t size, 58 + gfp_t gfp, bool intr, size_t align); 59 + 60 + void drm_suballoc_free(struct drm_suballoc *sa, struct dma_fence *fence); 61 + 62 + /** 63 + * drm_suballoc_soffset - Range start. 64 + * @sa: The struct drm_suballoc. 65 + * 66 + * Return: The start of the allocated range. 67 + */ 68 + static inline size_t drm_suballoc_soffset(struct drm_suballoc *sa) 69 + { 70 + return sa->soffset; 71 + } 72 + 73 + /** 74 + * drm_suballoc_eoffset - Range end. 75 + * @sa: The struct drm_suballoc. 76 + * 77 + * Return: The end of the allocated range + 1. 78 + */ 79 + static inline size_t drm_suballoc_eoffset(struct drm_suballoc *sa) 80 + { 81 + return sa->eoffset; 82 + } 83 + 84 + /** 85 + * drm_suballoc_size - Range size. 86 + * @sa: The struct drm_suballoc. 87 + * 88 + * Return: The size of the allocated range. 89 + */ 90 + static inline size_t drm_suballoc_size(struct drm_suballoc *sa) 91 + { 92 + return sa->eoffset - sa->soffset; 93 + } 94 + 95 + #ifdef CONFIG_DEBUG_FS 96 + void drm_suballoc_dump_debug_info(struct drm_suballoc_manager *sa_manager, 97 + struct drm_printer *p, 98 + unsigned long long suballoc_base); 99 + #else 100 + static inline void 101 + drm_suballoc_dump_debug_info(struct drm_suballoc_manager *sa_manager, 102 + struct drm_printer *p, 103 + unsigned long long suballoc_base) 104 + { } 105 + 106 + #endif 107 + 108 + #endif /* _DRM_SUBALLOC_H_ */