Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

drm/cma-helper: Pass GEM CMA object in public interfaces

Change all GEM CMA object functions that receive a GEM object
of type struct drm_gem_object to expect an object of type
struct drm_gem_cma_object instead.

This change reduces the number of upcasts from struct drm_gem_object
by moving them into callers. The C compiler can now verify that the
GEM CMA functions are called with the correct type.

For consistency, the patch also renames drm_gem_cma_free_object to
drm_gem_cma_free. It further updates documentation for a number of
functions.

Signed-off-by: Thomas Zimmermann <tzimmermann@suse.de>
Reviewed-by: Maxime Ripard <maxime@cerno.tech>
Link: https://patchwork.freedesktop.org/patch/msgid/20211115120148.21766-4-tzimmermann@suse.de

+48 -47
+24 -28
drivers/gpu/drm/drm_gem_cma_helper.c
··· 32 32 * The DRM GEM/CMA helpers use this allocator as a means to provide buffer 33 33 * objects that are physically contiguous in memory. This is useful for 34 34 * display drivers that are unable to map scattered buffers via an IOMMU. 35 + * 36 + * For GEM callback helpers in struct &drm_gem_object functions, see likewise 37 + * named functions with an _object_ infix (e.g., drm_gem_cma_object_vmap() wraps 38 + * drm_gem_cma_vmap()). These helpers perform the necessary type conversion. 35 39 */ 36 40 37 41 static const struct drm_gem_object_funcs drm_gem_cma_default_funcs = { ··· 196 192 } 197 193 198 194 /** 199 - * drm_gem_cma_free_object - free resources associated with a CMA GEM object 200 - * @gem_obj: GEM object to free 195 + * drm_gem_cma_free - free resources associated with a CMA GEM object 196 + * @cma_obj: CMA GEM object to free 201 197 * 202 198 * This function frees the backing memory of the CMA GEM object, cleans up the 203 199 * GEM object state and frees the memory used to store the object itself. 204 200 * If the buffer is imported and the virtual address is set, it is released. 205 201 */ 206 - void drm_gem_cma_free_object(struct drm_gem_object *gem_obj) 202 + void drm_gem_cma_free(struct drm_gem_cma_object *cma_obj) 207 203 { 208 - struct drm_gem_cma_object *cma_obj = to_drm_gem_cma_obj(gem_obj); 204 + struct drm_gem_object *gem_obj = &cma_obj->base; 209 205 struct dma_buf_map map = DMA_BUF_MAP_INIT_VADDR(cma_obj->vaddr); 210 206 211 207 if (gem_obj->import_attach) { ··· 221 217 222 218 kfree(cma_obj); 223 219 } 224 - EXPORT_SYMBOL_GPL(drm_gem_cma_free_object); 220 + EXPORT_SYMBOL_GPL(drm_gem_cma_free); 225 221 226 222 /** 227 223 * drm_gem_cma_dumb_create_internal - create a dumb buffer object ··· 368 364 369 365 /** 370 366 * drm_gem_cma_print_info() - Print &drm_gem_cma_object info for debugfs 367 + * @cma_obj: CMA GEM object 371 368 * @p: DRM printer 372 369 * @indent: Tab indentation level 373 - * @obj: GEM object 374 370 * 375 - * This function can be used as the &drm_driver->gem_print_info callback. 376 - * It prints paddr and vaddr for use in e.g. debugfs output. 371 + * This function prints paddr and vaddr for use in e.g. debugfs output. 377 372 */ 378 - void drm_gem_cma_print_info(struct drm_printer *p, unsigned int indent, 379 - const struct drm_gem_object *obj) 373 + void drm_gem_cma_print_info(const struct drm_gem_cma_object *cma_obj, 374 + struct drm_printer *p, unsigned int indent) 380 375 { 381 - const struct drm_gem_cma_object *cma_obj = to_drm_gem_cma_obj(obj); 382 - 383 376 drm_printf_indent(p, indent, "paddr=%pad\n", &cma_obj->paddr); 384 377 drm_printf_indent(p, indent, "vaddr=%p\n", cma_obj->vaddr); 385 378 } ··· 385 384 /** 386 385 * drm_gem_cma_get_sg_table - provide a scatter/gather table of pinned 387 386 * pages for a CMA GEM object 388 - * @obj: GEM object 387 + * @cma_obj: CMA GEM object 389 388 * 390 389 * This function exports a scatter/gather table by calling the standard 391 390 * DMA mapping API. ··· 393 392 * Returns: 394 393 * A pointer to the scatter/gather table of pinned pages or NULL on failure. 395 394 */ 396 - struct sg_table *drm_gem_cma_get_sg_table(struct drm_gem_object *obj) 395 + struct sg_table *drm_gem_cma_get_sg_table(struct drm_gem_cma_object *cma_obj) 397 396 { 398 - struct drm_gem_cma_object *cma_obj = to_drm_gem_cma_obj(obj); 397 + struct drm_gem_object *obj = &cma_obj->base; 399 398 struct sg_table *sgt; 400 399 int ret; 401 400 ··· 461 460 /** 462 461 * drm_gem_cma_vmap - map a CMA GEM object into the kernel's virtual 463 462 * address space 464 - * @obj: GEM object 463 + * @cma_obj: CMA GEM object 465 464 * @map: Returns the kernel virtual address of the CMA GEM object's backing 466 465 * store. 467 466 * 468 - * This function maps a buffer into the kernel's 469 - * virtual address space. Since the CMA buffers are already mapped into the 470 - * kernel virtual address space this simply returns the cached virtual 471 - * address. 467 + * This function maps a buffer into the kernel's virtual address space. 468 + * Since the CMA buffers are already mapped into the kernel virtual address 469 + * space this simply returns the cached virtual address. 472 470 * 473 471 * Returns: 474 472 * 0 on success, or a negative error code otherwise. 475 473 */ 476 - int drm_gem_cma_vmap(struct drm_gem_object *obj, struct dma_buf_map *map) 474 + int drm_gem_cma_vmap(struct drm_gem_cma_object *cma_obj, struct dma_buf_map *map) 477 475 { 478 - struct drm_gem_cma_object *cma_obj = to_drm_gem_cma_obj(obj); 479 - 480 476 dma_buf_map_set_vaddr(map, cma_obj->vaddr); 481 477 482 478 return 0; ··· 482 484 483 485 /** 484 486 * drm_gem_cma_mmap - memory-map an exported CMA GEM object 485 - * @obj: GEM object 487 + * @cma_obj: CMA GEM object 486 488 * @vma: VMA for the area to be mapped 487 489 * 488 490 * This function maps a buffer into a userspace process's address space. ··· 492 494 * Returns: 493 495 * 0 on success or a negative error code on failure. 494 496 */ 495 - int drm_gem_cma_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma) 497 + int drm_gem_cma_mmap(struct drm_gem_cma_object *cma_obj, struct vm_area_struct *vma) 496 498 { 497 - struct drm_gem_cma_object *cma_obj; 499 + struct drm_gem_object *obj = &cma_obj->base; 498 500 int ret; 499 501 500 502 /* ··· 504 506 */ 505 507 vma->vm_pgoff -= drm_vma_node_start(&obj->vma_node); 506 508 vma->vm_flags &= ~VM_PFNMAP; 507 - 508 - cma_obj = to_drm_gem_cma_obj(obj); 509 509 510 510 if (cma_obj->map_noncoherent) { 511 511 vma->vm_page_prot = vm_get_page_prot(vma->vm_flags);
+2 -2
drivers/gpu/drm/vc4/vc4_bo.c
··· 177 177 bo->validated_shader = NULL; 178 178 } 179 179 180 - drm_gem_cma_free_object(obj); 180 + drm_gem_cma_free(&bo->base); 181 181 } 182 182 183 183 static void vc4_bo_remove_from_cache(struct vc4_bo *bo) ··· 720 720 return -EINVAL; 721 721 } 722 722 723 - return drm_gem_cma_mmap(obj, vma); 723 + return drm_gem_cma_mmap(&bo->base, vma); 724 724 } 725 725 726 726 static const struct vm_operations_struct vc4_vm_ops = {
+22 -17
include/drm/drm_gem_cma_helper.h
··· 32 32 #define to_drm_gem_cma_obj(gem_obj) \ 33 33 container_of(gem_obj, struct drm_gem_cma_object, base) 34 34 35 - /* free GEM object */ 36 - void drm_gem_cma_free_object(struct drm_gem_object *gem_obj); 37 - 38 - /* allocate physical memory */ 39 35 struct drm_gem_cma_object *drm_gem_cma_create(struct drm_device *drm, 40 36 size_t size); 37 + void drm_gem_cma_free(struct drm_gem_cma_object *cma_obj); 38 + void drm_gem_cma_print_info(const struct drm_gem_cma_object *cma_obj, 39 + struct drm_printer *p, unsigned int indent); 40 + struct sg_table *drm_gem_cma_get_sg_table(struct drm_gem_cma_object *cma_obj); 41 + int drm_gem_cma_vmap(struct drm_gem_cma_object *cma_obj, struct dma_buf_map *map); 42 + int drm_gem_cma_mmap(struct drm_gem_cma_object *cma_obj, struct vm_area_struct *vma); 41 43 42 44 extern const struct vm_operations_struct drm_gem_cma_vm_ops; 43 - 44 - void drm_gem_cma_print_info(struct drm_printer *p, unsigned int indent, 45 - const struct drm_gem_object *obj); 46 - 47 - struct sg_table *drm_gem_cma_get_sg_table(struct drm_gem_object *obj); 48 - int drm_gem_cma_vmap(struct drm_gem_object *obj, struct dma_buf_map *map); 49 - int drm_gem_cma_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma); 50 45 51 46 /* 52 47 * GEM object functions 53 48 */ 54 49 55 50 /** 56 - * drm_gem_cma_object_free - GEM object function for drm_gem_cma_free_object() 51 + * drm_gem_cma_object_free - GEM object function for drm_gem_cma_free() 57 52 * @obj: GEM object to free 58 53 * 59 54 * This function wraps drm_gem_cma_free_object(). Drivers that employ the CMA helpers ··· 56 61 */ 57 62 static inline void drm_gem_cma_object_free(struct drm_gem_object *obj) 58 63 { 59 - drm_gem_cma_free_object(obj); 64 + struct drm_gem_cma_object *cma_obj = to_drm_gem_cma_obj(obj); 65 + 66 + drm_gem_cma_free(cma_obj); 60 67 } 61 68 62 69 /** ··· 73 76 static inline void drm_gem_cma_object_print_info(struct drm_printer *p, unsigned int indent, 74 77 const struct drm_gem_object *obj) 75 78 { 76 - drm_gem_cma_print_info(p, indent, obj); 79 + const struct drm_gem_cma_object *cma_obj = to_drm_gem_cma_obj(obj); 80 + 81 + drm_gem_cma_print_info(cma_obj, p, indent); 77 82 } 78 83 79 84 /** ··· 90 91 */ 91 92 static inline struct sg_table *drm_gem_cma_object_get_sg_table(struct drm_gem_object *obj) 92 93 { 93 - return drm_gem_cma_get_sg_table(obj); 94 + struct drm_gem_cma_object *cma_obj = to_drm_gem_cma_obj(obj); 95 + 96 + return drm_gem_cma_get_sg_table(cma_obj); 94 97 } 95 98 96 99 /* ··· 108 107 */ 109 108 static inline int drm_gem_cma_object_vmap(struct drm_gem_object *obj, struct dma_buf_map *map) 110 109 { 111 - return drm_gem_cma_vmap(obj, map); 110 + struct drm_gem_cma_object *cma_obj = to_drm_gem_cma_obj(obj); 111 + 112 + return drm_gem_cma_vmap(cma_obj, map); 112 113 } 113 114 114 115 /** ··· 126 123 */ 127 124 static inline int drm_gem_cma_object_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma) 128 125 { 129 - return drm_gem_cma_mmap(obj, vma); 126 + struct drm_gem_cma_object *cma_obj = to_drm_gem_cma_obj(obj); 127 + 128 + return drm_gem_cma_mmap(cma_obj, vma); 130 129 } 131 130 132 131 /*