mirror of
https://github.com/torvalds/linux.git
synced 2026-04-19 15:24:02 -04:00
clangd reports many "unused header" warnings throughout the Xe driver. Start working to clean this up by removing unnecessary includes in our .c files and/or replacing them with explicit includes of other headers that were previously being included indirectly. By far the most common offender here was unnecessary inclusion of xe_gt.h. That likely originates from the early days of xe.ko when xe_mmio did not exist and all register accesses, including those unrelated to GTs, were done with GT functions. There's still a lot of additional #include cleanup that can be done in the headers themselves; that will come as a followup series. v2: - Squash the 79-patch series down to a single patch. (MattB) Reviewed-by: Matthew Brost <matthew.brost@intel.com> Link: https://patch.msgid.link/20260115032803.4067824-2-matthew.d.roper@intel.com Signed-off-by: Matt Roper <matthew.d.roper@intel.com>
217 lines
5.7 KiB
C
217 lines
5.7 KiB
C
// SPDX-License-Identifier: MIT
|
|
/*
|
|
* Copyright © 2022 Intel Corporation
|
|
*/
|
|
|
|
#include "xe_sa.h"
|
|
|
|
#include <linux/kernel.h>
|
|
|
|
#include <drm/drm_managed.h>
|
|
|
|
#include "xe_bo.h"
|
|
#include "xe_device_types.h"
|
|
#include "xe_map.h"
|
|
|
|
static void xe_sa_bo_manager_fini(struct drm_device *drm, void *arg)
|
|
{
|
|
struct xe_sa_manager *sa_manager = arg;
|
|
struct xe_bo *bo = sa_manager->bo;
|
|
|
|
if (!bo) {
|
|
drm_err(drm, "no bo for sa manager\n");
|
|
return;
|
|
}
|
|
|
|
drm_suballoc_manager_fini(&sa_manager->base);
|
|
|
|
if (sa_manager->is_iomem)
|
|
kvfree(sa_manager->cpu_ptr);
|
|
|
|
sa_manager->bo = NULL;
|
|
sa_manager->shadow = NULL;
|
|
}
|
|
|
|
/**
|
|
* __xe_sa_bo_manager_init() - Create and initialize the suballocator
|
|
* @tile: the &xe_tile where allocate
|
|
* @size: number of bytes to allocate
|
|
* @guard: number of bytes to exclude from suballocations
|
|
* @align: alignment for each suballocated chunk
|
|
* @flags: flags for suballocator
|
|
*
|
|
* Prepares the suballocation manager for suballocations.
|
|
*
|
|
* Return: a pointer to the &xe_sa_manager or an ERR_PTR on failure.
|
|
*/
|
|
struct xe_sa_manager *__xe_sa_bo_manager_init(struct xe_tile *tile, u32 size,
|
|
u32 guard, u32 align, u32 flags)
|
|
{
|
|
struct xe_device *xe = tile_to_xe(tile);
|
|
struct xe_sa_manager *sa_manager;
|
|
u32 managed_size;
|
|
struct xe_bo *bo;
|
|
int ret;
|
|
|
|
xe_tile_assert(tile, size > guard);
|
|
managed_size = size - guard;
|
|
|
|
sa_manager = drmm_kzalloc(&xe->drm, sizeof(*sa_manager), GFP_KERNEL);
|
|
if (!sa_manager)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
bo = xe_managed_bo_create_pin_map(xe, tile, size,
|
|
XE_BO_FLAG_VRAM_IF_DGFX(tile) |
|
|
XE_BO_FLAG_GGTT |
|
|
XE_BO_FLAG_GGTT_INVALIDATE |
|
|
XE_BO_FLAG_PINNED_NORESTORE);
|
|
if (IS_ERR(bo)) {
|
|
drm_err(&xe->drm, "Failed to prepare %uKiB BO for SA manager (%pe)\n",
|
|
size / SZ_1K, bo);
|
|
return ERR_CAST(bo);
|
|
}
|
|
sa_manager->bo = bo;
|
|
sa_manager->is_iomem = bo->vmap.is_iomem;
|
|
|
|
if (bo->vmap.is_iomem) {
|
|
sa_manager->cpu_ptr = kvzalloc(managed_size, GFP_KERNEL);
|
|
if (!sa_manager->cpu_ptr)
|
|
return ERR_PTR(-ENOMEM);
|
|
} else {
|
|
sa_manager->cpu_ptr = bo->vmap.vaddr;
|
|
memset(sa_manager->cpu_ptr, 0, bo->ttm.base.size);
|
|
}
|
|
|
|
if (flags & XE_SA_BO_MANAGER_FLAG_SHADOW) {
|
|
struct xe_bo *shadow;
|
|
|
|
ret = drmm_mutex_init(&xe->drm, &sa_manager->swap_guard);
|
|
if (ret)
|
|
return ERR_PTR(ret);
|
|
|
|
shadow = xe_managed_bo_create_pin_map(xe, tile, size,
|
|
XE_BO_FLAG_VRAM_IF_DGFX(tile) |
|
|
XE_BO_FLAG_GGTT |
|
|
XE_BO_FLAG_GGTT_INVALIDATE |
|
|
XE_BO_FLAG_PINNED_NORESTORE);
|
|
if (IS_ERR(shadow)) {
|
|
drm_err(&xe->drm, "Failed to prepare %uKiB BO for SA manager (%pe)\n",
|
|
size / SZ_1K, shadow);
|
|
return ERR_CAST(shadow);
|
|
}
|
|
sa_manager->shadow = shadow;
|
|
}
|
|
|
|
drm_suballoc_manager_init(&sa_manager->base, managed_size, align);
|
|
ret = drmm_add_action_or_reset(&xe->drm, xe_sa_bo_manager_fini,
|
|
sa_manager);
|
|
if (ret)
|
|
return ERR_PTR(ret);
|
|
|
|
return sa_manager;
|
|
}
|
|
|
|
/**
|
|
* xe_sa_bo_swap_shadow() - Swap the SA BO with shadow BO.
|
|
* @sa_manager: the XE sub allocator manager
|
|
*
|
|
* Swaps the sub-allocator primary buffer object with shadow buffer object.
|
|
*
|
|
* Return: None.
|
|
*/
|
|
void xe_sa_bo_swap_shadow(struct xe_sa_manager *sa_manager)
|
|
{
|
|
struct xe_device *xe = tile_to_xe(sa_manager->bo->tile);
|
|
|
|
xe_assert(xe, sa_manager->shadow);
|
|
lockdep_assert_held(&sa_manager->swap_guard);
|
|
|
|
swap(sa_manager->bo, sa_manager->shadow);
|
|
if (!sa_manager->bo->vmap.is_iomem)
|
|
sa_manager->cpu_ptr = sa_manager->bo->vmap.vaddr;
|
|
}
|
|
|
|
/**
|
|
* xe_sa_bo_sync_shadow() - Sync the SA Shadow BO with primary BO.
|
|
* @sa_bo: the sub-allocator buffer object.
|
|
*
|
|
* Synchronize sub-allocator shadow buffer object with primary buffer object.
|
|
*
|
|
* Return: None.
|
|
*/
|
|
void xe_sa_bo_sync_shadow(struct drm_suballoc *sa_bo)
|
|
{
|
|
struct xe_sa_manager *sa_manager = to_xe_sa_manager(sa_bo->manager);
|
|
struct xe_device *xe = tile_to_xe(sa_manager->bo->tile);
|
|
|
|
xe_assert(xe, sa_manager->shadow);
|
|
lockdep_assert_held(&sa_manager->swap_guard);
|
|
|
|
xe_map_memcpy_to(xe, &sa_manager->shadow->vmap,
|
|
drm_suballoc_soffset(sa_bo),
|
|
xe_sa_bo_cpu_addr(sa_bo),
|
|
drm_suballoc_size(sa_bo));
|
|
}
|
|
|
|
/**
|
|
* __xe_sa_bo_new() - Make a suballocation but use custom gfp flags.
|
|
* @sa_manager: the &xe_sa_manager
|
|
* @size: number of bytes we want to suballocate
|
|
* @gfp: gfp flags used for memory allocation. Typically GFP_KERNEL.
|
|
*
|
|
* Try to make a suballocation of size @size.
|
|
*
|
|
* Return: a &drm_suballoc, or an ERR_PTR.
|
|
*/
|
|
struct drm_suballoc *__xe_sa_bo_new(struct xe_sa_manager *sa_manager, u32 size, gfp_t gfp)
|
|
{
|
|
/*
|
|
* BB to large, return -ENOBUFS indicating user should split
|
|
* array of binds into smaller chunks.
|
|
*/
|
|
if (size > sa_manager->base.size)
|
|
return ERR_PTR(-ENOBUFS);
|
|
|
|
return drm_suballoc_new(&sa_manager->base, size, gfp, true, 0);
|
|
}
|
|
|
|
/**
|
|
* xe_sa_bo_flush_write() - Copy the data from the sub-allocation to the GPU memory.
|
|
* @sa_bo: the &drm_suballoc to flush
|
|
*/
|
|
void xe_sa_bo_flush_write(struct drm_suballoc *sa_bo)
|
|
{
|
|
struct xe_sa_manager *sa_manager = to_xe_sa_manager(sa_bo->manager);
|
|
struct xe_device *xe = tile_to_xe(sa_manager->bo->tile);
|
|
|
|
if (!sa_manager->bo->vmap.is_iomem)
|
|
return;
|
|
|
|
xe_map_memcpy_to(xe, &sa_manager->bo->vmap, drm_suballoc_soffset(sa_bo),
|
|
xe_sa_bo_cpu_addr(sa_bo),
|
|
drm_suballoc_size(sa_bo));
|
|
}
|
|
|
|
/**
|
|
* xe_sa_bo_sync_read() - Copy the data from GPU memory to the sub-allocation.
|
|
* @sa_bo: the &drm_suballoc to sync
|
|
*/
|
|
void xe_sa_bo_sync_read(struct drm_suballoc *sa_bo)
|
|
{
|
|
struct xe_sa_manager *sa_manager = to_xe_sa_manager(sa_bo->manager);
|
|
struct xe_device *xe = tile_to_xe(sa_manager->bo->tile);
|
|
|
|
if (!sa_manager->bo->vmap.is_iomem)
|
|
return;
|
|
|
|
xe_map_memcpy_from(xe, xe_sa_bo_cpu_addr(sa_bo), &sa_manager->bo->vmap,
|
|
drm_suballoc_soffset(sa_bo),
|
|
drm_suballoc_size(sa_bo));
|
|
}
|
|
|
|
void xe_sa_bo_free(struct drm_suballoc *sa_bo,
|
|
struct dma_fence *fence)
|
|
{
|
|
drm_suballoc_free(sa_bo, fence);
|
|
}
|