mirror of
https://github.com/torvalds/linux.git
synced 2026-04-20 15:53:59 -04:00
When mapping an IO region, the pseudo-file offset is dependent on the
userspace architecture. panthor_device_mmio_offset() abstracts that
away for us by turning a userspace MMIO offset into its kernel
equivalent, but we were not updating vm_area_struct::vm_pgoff
accordingly, leading us to attach the MMIO region to the wrong file
offset.
This has implications when we start mixing 64 bit and 32 bit apps, but
that's only really a problem when we start having more that 2^43 bytes of
memory allocated, which is very unlikely to happen.
What's more problematic is the fact this turns our
unmap_mapping_range(DRM_PANTHOR_USER_MMIO_OFFSET) calls, which are
supposed to kill the MMIO mapping when entering suspend, into NOPs.
Which means we either keep the dummy flush_id mapping active at all
times, or we risk a BUS_FAULT if the MMIO region was mapped, and the
GPU is suspended after that.
Solve that by patching vm_pgoff early in panthor_mmap(). With
this in place, we no longer need the panthor_device_mmio_offset()
helper.
v3:
- No changes
v2:
- Kill panthor_device_mmio_offset()
Fixes: 5fe909cae1 ("drm/panthor: Add the device logical block")
Reported-by: Adrián Larumbe <adrian.larumbe@collabora.com>
Reported-by: Lukas F. Hartmann <lukas@mntmn.com>
Closes: https://gitlab.freedesktop.org/mesa/mesa/-/issues/10835
Signed-off-by: Boris Brezillon <boris.brezillon@collabora.com>
Reviewed-by: Steven Price <steven.price@arm.com>
Reviewed-by: Liviu Dudau <liviu.dudau@arm.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20240326111205.510019-1-boris.brezillon@collabora.com
562 lines
15 KiB
C
562 lines
15 KiB
C
// SPDX-License-Identifier: GPL-2.0 or MIT
|
|
/* Copyright 2018 Marty E. Plummer <hanetzer@startmail.com> */
|
|
/* Copyright 2019 Linaro, Ltd, Rob Herring <robh@kernel.org> */
|
|
/* Copyright 2023 Collabora ltd. */
|
|
|
|
#include <linux/clk.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/pm_domain.h>
|
|
#include <linux/pm_runtime.h>
|
|
#include <linux/regulator/consumer.h>
|
|
#include <linux/reset.h>
|
|
|
|
#include <drm/drm_drv.h>
|
|
#include <drm/drm_managed.h>
|
|
|
|
#include "panthor_devfreq.h"
|
|
#include "panthor_device.h"
|
|
#include "panthor_fw.h"
|
|
#include "panthor_gpu.h"
|
|
#include "panthor_mmu.h"
|
|
#include "panthor_regs.h"
|
|
#include "panthor_sched.h"
|
|
|
|
static int panthor_clk_init(struct panthor_device *ptdev)
|
|
{
|
|
ptdev->clks.core = devm_clk_get(ptdev->base.dev, NULL);
|
|
if (IS_ERR(ptdev->clks.core))
|
|
return dev_err_probe(ptdev->base.dev,
|
|
PTR_ERR(ptdev->clks.core),
|
|
"get 'core' clock failed");
|
|
|
|
ptdev->clks.stacks = devm_clk_get_optional(ptdev->base.dev, "stacks");
|
|
if (IS_ERR(ptdev->clks.stacks))
|
|
return dev_err_probe(ptdev->base.dev,
|
|
PTR_ERR(ptdev->clks.stacks),
|
|
"get 'stacks' clock failed");
|
|
|
|
ptdev->clks.coregroup = devm_clk_get_optional(ptdev->base.dev, "coregroup");
|
|
if (IS_ERR(ptdev->clks.coregroup))
|
|
return dev_err_probe(ptdev->base.dev,
|
|
PTR_ERR(ptdev->clks.coregroup),
|
|
"get 'coregroup' clock failed");
|
|
|
|
drm_info(&ptdev->base, "clock rate = %lu\n", clk_get_rate(ptdev->clks.core));
|
|
return 0;
|
|
}
|
|
|
|
void panthor_device_unplug(struct panthor_device *ptdev)
|
|
{
|
|
/* This function can be called from two different path: the reset work
|
|
* and the platform device remove callback. drm_dev_unplug() doesn't
|
|
* deal with concurrent callers, so we have to protect drm_dev_unplug()
|
|
* calls with our own lock, and bail out if the device is already
|
|
* unplugged.
|
|
*/
|
|
mutex_lock(&ptdev->unplug.lock);
|
|
if (drm_dev_is_unplugged(&ptdev->base)) {
|
|
/* Someone beat us, release the lock and wait for the unplug
|
|
* operation to be reported as done.
|
|
**/
|
|
mutex_unlock(&ptdev->unplug.lock);
|
|
wait_for_completion(&ptdev->unplug.done);
|
|
return;
|
|
}
|
|
|
|
/* Call drm_dev_unplug() so any access to HW blocks happening after
|
|
* that point get rejected.
|
|
*/
|
|
drm_dev_unplug(&ptdev->base);
|
|
|
|
/* We do the rest of the unplug with the unplug lock released,
|
|
* future callers will wait on ptdev->unplug.done anyway.
|
|
*/
|
|
mutex_unlock(&ptdev->unplug.lock);
|
|
|
|
drm_WARN_ON(&ptdev->base, pm_runtime_get_sync(ptdev->base.dev) < 0);
|
|
|
|
/* Now, try to cleanly shutdown the GPU before the device resources
|
|
* get reclaimed.
|
|
*/
|
|
panthor_sched_unplug(ptdev);
|
|
panthor_fw_unplug(ptdev);
|
|
panthor_mmu_unplug(ptdev);
|
|
panthor_gpu_unplug(ptdev);
|
|
|
|
pm_runtime_dont_use_autosuspend(ptdev->base.dev);
|
|
pm_runtime_put_sync_suspend(ptdev->base.dev);
|
|
|
|
/* If PM is disabled, we need to call the suspend handler manually. */
|
|
if (!IS_ENABLED(CONFIG_PM))
|
|
panthor_device_suspend(ptdev->base.dev);
|
|
|
|
/* Report the unplug operation as done to unblock concurrent
|
|
* panthor_device_unplug() callers.
|
|
*/
|
|
complete_all(&ptdev->unplug.done);
|
|
}
|
|
|
|
static void panthor_device_reset_cleanup(struct drm_device *ddev, void *data)
|
|
{
|
|
struct panthor_device *ptdev = container_of(ddev, struct panthor_device, base);
|
|
|
|
cancel_work_sync(&ptdev->reset.work);
|
|
destroy_workqueue(ptdev->reset.wq);
|
|
}
|
|
|
|
static void panthor_device_reset_work(struct work_struct *work)
|
|
{
|
|
struct panthor_device *ptdev = container_of(work, struct panthor_device, reset.work);
|
|
int ret = 0, cookie;
|
|
|
|
if (atomic_read(&ptdev->pm.state) != PANTHOR_DEVICE_PM_STATE_ACTIVE) {
|
|
/*
|
|
* No need for a reset as the device has been (or will be)
|
|
* powered down
|
|
*/
|
|
atomic_set(&ptdev->reset.pending, 0);
|
|
return;
|
|
}
|
|
|
|
if (!drm_dev_enter(&ptdev->base, &cookie))
|
|
return;
|
|
|
|
panthor_sched_pre_reset(ptdev);
|
|
panthor_fw_pre_reset(ptdev, true);
|
|
panthor_mmu_pre_reset(ptdev);
|
|
panthor_gpu_soft_reset(ptdev);
|
|
panthor_gpu_l2_power_on(ptdev);
|
|
panthor_mmu_post_reset(ptdev);
|
|
ret = panthor_fw_post_reset(ptdev);
|
|
if (ret)
|
|
goto out_dev_exit;
|
|
|
|
atomic_set(&ptdev->reset.pending, 0);
|
|
panthor_sched_post_reset(ptdev);
|
|
|
|
out_dev_exit:
|
|
drm_dev_exit(cookie);
|
|
|
|
if (ret) {
|
|
panthor_device_unplug(ptdev);
|
|
drm_err(&ptdev->base, "Failed to boot MCU after reset, making device unusable.");
|
|
}
|
|
}
|
|
|
|
static bool panthor_device_is_initialized(struct panthor_device *ptdev)
|
|
{
|
|
return !!ptdev->scheduler;
|
|
}
|
|
|
|
static void panthor_device_free_page(struct drm_device *ddev, void *data)
|
|
{
|
|
__free_page(data);
|
|
}
|
|
|
|
int panthor_device_init(struct panthor_device *ptdev)
|
|
{
|
|
u32 *dummy_page_virt;
|
|
struct resource *res;
|
|
struct page *p;
|
|
int ret;
|
|
|
|
ptdev->coherent = device_get_dma_attr(ptdev->base.dev) == DEV_DMA_COHERENT;
|
|
|
|
init_completion(&ptdev->unplug.done);
|
|
ret = drmm_mutex_init(&ptdev->base, &ptdev->unplug.lock);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = drmm_mutex_init(&ptdev->base, &ptdev->pm.mmio_lock);
|
|
if (ret)
|
|
return ret;
|
|
|
|
atomic_set(&ptdev->pm.state, PANTHOR_DEVICE_PM_STATE_SUSPENDED);
|
|
p = alloc_page(GFP_KERNEL | __GFP_ZERO);
|
|
if (!p)
|
|
return -ENOMEM;
|
|
|
|
ptdev->pm.dummy_latest_flush = p;
|
|
dummy_page_virt = page_address(p);
|
|
ret = drmm_add_action_or_reset(&ptdev->base, panthor_device_free_page,
|
|
ptdev->pm.dummy_latest_flush);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/*
|
|
* Set the dummy page holding the latest flush to 1. This will cause the
|
|
* flush to avoided as we know it isn't necessary if the submission
|
|
* happens while the dummy page is mapped. Zero cannot be used because
|
|
* that means 'always flush'.
|
|
*/
|
|
*dummy_page_virt = 1;
|
|
|
|
INIT_WORK(&ptdev->reset.work, panthor_device_reset_work);
|
|
ptdev->reset.wq = alloc_ordered_workqueue("panthor-reset-wq", 0);
|
|
if (!ptdev->reset.wq)
|
|
return -ENOMEM;
|
|
|
|
ret = drmm_add_action_or_reset(&ptdev->base, panthor_device_reset_cleanup, NULL);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = panthor_clk_init(ptdev);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = panthor_devfreq_init(ptdev);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ptdev->iomem = devm_platform_get_and_ioremap_resource(to_platform_device(ptdev->base.dev),
|
|
0, &res);
|
|
if (IS_ERR(ptdev->iomem))
|
|
return PTR_ERR(ptdev->iomem);
|
|
|
|
ptdev->phys_addr = res->start;
|
|
|
|
ret = devm_pm_runtime_enable(ptdev->base.dev);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = pm_runtime_resume_and_get(ptdev->base.dev);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/* If PM is disabled, we need to call panthor_device_resume() manually. */
|
|
if (!IS_ENABLED(CONFIG_PM)) {
|
|
ret = panthor_device_resume(ptdev->base.dev);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
ret = panthor_gpu_init(ptdev);
|
|
if (ret)
|
|
goto err_rpm_put;
|
|
|
|
ret = panthor_mmu_init(ptdev);
|
|
if (ret)
|
|
goto err_unplug_gpu;
|
|
|
|
ret = panthor_fw_init(ptdev);
|
|
if (ret)
|
|
goto err_unplug_mmu;
|
|
|
|
ret = panthor_sched_init(ptdev);
|
|
if (ret)
|
|
goto err_unplug_fw;
|
|
|
|
/* ~3 frames */
|
|
pm_runtime_set_autosuspend_delay(ptdev->base.dev, 50);
|
|
pm_runtime_use_autosuspend(ptdev->base.dev);
|
|
|
|
ret = drm_dev_register(&ptdev->base, 0);
|
|
if (ret)
|
|
goto err_disable_autosuspend;
|
|
|
|
pm_runtime_put_autosuspend(ptdev->base.dev);
|
|
return 0;
|
|
|
|
err_disable_autosuspend:
|
|
pm_runtime_dont_use_autosuspend(ptdev->base.dev);
|
|
panthor_sched_unplug(ptdev);
|
|
|
|
err_unplug_fw:
|
|
panthor_fw_unplug(ptdev);
|
|
|
|
err_unplug_mmu:
|
|
panthor_mmu_unplug(ptdev);
|
|
|
|
err_unplug_gpu:
|
|
panthor_gpu_unplug(ptdev);
|
|
|
|
err_rpm_put:
|
|
pm_runtime_put_sync_suspend(ptdev->base.dev);
|
|
return ret;
|
|
}
|
|
|
|
#define PANTHOR_EXCEPTION(id) \
|
|
[DRM_PANTHOR_EXCEPTION_ ## id] = { \
|
|
.name = #id, \
|
|
}
|
|
|
|
struct panthor_exception_info {
|
|
const char *name;
|
|
};
|
|
|
|
static const struct panthor_exception_info panthor_exception_infos[] = {
|
|
PANTHOR_EXCEPTION(OK),
|
|
PANTHOR_EXCEPTION(TERMINATED),
|
|
PANTHOR_EXCEPTION(KABOOM),
|
|
PANTHOR_EXCEPTION(EUREKA),
|
|
PANTHOR_EXCEPTION(ACTIVE),
|
|
PANTHOR_EXCEPTION(CS_RES_TERM),
|
|
PANTHOR_EXCEPTION(CS_CONFIG_FAULT),
|
|
PANTHOR_EXCEPTION(CS_ENDPOINT_FAULT),
|
|
PANTHOR_EXCEPTION(CS_BUS_FAULT),
|
|
PANTHOR_EXCEPTION(CS_INSTR_INVALID),
|
|
PANTHOR_EXCEPTION(CS_CALL_STACK_OVERFLOW),
|
|
PANTHOR_EXCEPTION(CS_INHERIT_FAULT),
|
|
PANTHOR_EXCEPTION(INSTR_INVALID_PC),
|
|
PANTHOR_EXCEPTION(INSTR_INVALID_ENC),
|
|
PANTHOR_EXCEPTION(INSTR_BARRIER_FAULT),
|
|
PANTHOR_EXCEPTION(DATA_INVALID_FAULT),
|
|
PANTHOR_EXCEPTION(TILE_RANGE_FAULT),
|
|
PANTHOR_EXCEPTION(ADDR_RANGE_FAULT),
|
|
PANTHOR_EXCEPTION(IMPRECISE_FAULT),
|
|
PANTHOR_EXCEPTION(OOM),
|
|
PANTHOR_EXCEPTION(CSF_FW_INTERNAL_ERROR),
|
|
PANTHOR_EXCEPTION(CSF_RES_EVICTION_TIMEOUT),
|
|
PANTHOR_EXCEPTION(GPU_BUS_FAULT),
|
|
PANTHOR_EXCEPTION(GPU_SHAREABILITY_FAULT),
|
|
PANTHOR_EXCEPTION(SYS_SHAREABILITY_FAULT),
|
|
PANTHOR_EXCEPTION(GPU_CACHEABILITY_FAULT),
|
|
PANTHOR_EXCEPTION(TRANSLATION_FAULT_0),
|
|
PANTHOR_EXCEPTION(TRANSLATION_FAULT_1),
|
|
PANTHOR_EXCEPTION(TRANSLATION_FAULT_2),
|
|
PANTHOR_EXCEPTION(TRANSLATION_FAULT_3),
|
|
PANTHOR_EXCEPTION(TRANSLATION_FAULT_4),
|
|
PANTHOR_EXCEPTION(PERM_FAULT_0),
|
|
PANTHOR_EXCEPTION(PERM_FAULT_1),
|
|
PANTHOR_EXCEPTION(PERM_FAULT_2),
|
|
PANTHOR_EXCEPTION(PERM_FAULT_3),
|
|
PANTHOR_EXCEPTION(ACCESS_FLAG_1),
|
|
PANTHOR_EXCEPTION(ACCESS_FLAG_2),
|
|
PANTHOR_EXCEPTION(ACCESS_FLAG_3),
|
|
PANTHOR_EXCEPTION(ADDR_SIZE_FAULT_IN),
|
|
PANTHOR_EXCEPTION(ADDR_SIZE_FAULT_OUT0),
|
|
PANTHOR_EXCEPTION(ADDR_SIZE_FAULT_OUT1),
|
|
PANTHOR_EXCEPTION(ADDR_SIZE_FAULT_OUT2),
|
|
PANTHOR_EXCEPTION(ADDR_SIZE_FAULT_OUT3),
|
|
PANTHOR_EXCEPTION(MEM_ATTR_FAULT_0),
|
|
PANTHOR_EXCEPTION(MEM_ATTR_FAULT_1),
|
|
PANTHOR_EXCEPTION(MEM_ATTR_FAULT_2),
|
|
PANTHOR_EXCEPTION(MEM_ATTR_FAULT_3),
|
|
};
|
|
|
|
const char *panthor_exception_name(struct panthor_device *ptdev, u32 exception_code)
|
|
{
|
|
if (exception_code >= ARRAY_SIZE(panthor_exception_infos) ||
|
|
!panthor_exception_infos[exception_code].name)
|
|
return "Unknown exception type";
|
|
|
|
return panthor_exception_infos[exception_code].name;
|
|
}
|
|
|
|
static vm_fault_t panthor_mmio_vm_fault(struct vm_fault *vmf)
|
|
{
|
|
struct vm_area_struct *vma = vmf->vma;
|
|
struct panthor_device *ptdev = vma->vm_private_data;
|
|
u64 offset = (u64)vma->vm_pgoff << PAGE_SHIFT;
|
|
unsigned long pfn;
|
|
pgprot_t pgprot;
|
|
vm_fault_t ret;
|
|
bool active;
|
|
int cookie;
|
|
|
|
if (!drm_dev_enter(&ptdev->base, &cookie))
|
|
return VM_FAULT_SIGBUS;
|
|
|
|
mutex_lock(&ptdev->pm.mmio_lock);
|
|
active = atomic_read(&ptdev->pm.state) == PANTHOR_DEVICE_PM_STATE_ACTIVE;
|
|
|
|
switch (offset) {
|
|
case DRM_PANTHOR_USER_FLUSH_ID_MMIO_OFFSET:
|
|
if (active)
|
|
pfn = __phys_to_pfn(ptdev->phys_addr + CSF_GPU_LATEST_FLUSH_ID);
|
|
else
|
|
pfn = page_to_pfn(ptdev->pm.dummy_latest_flush);
|
|
break;
|
|
|
|
default:
|
|
ret = VM_FAULT_SIGBUS;
|
|
goto out_unlock;
|
|
}
|
|
|
|
pgprot = vma->vm_page_prot;
|
|
if (active)
|
|
pgprot = pgprot_noncached(pgprot);
|
|
|
|
ret = vmf_insert_pfn_prot(vma, vmf->address, pfn, pgprot);
|
|
|
|
out_unlock:
|
|
mutex_unlock(&ptdev->pm.mmio_lock);
|
|
drm_dev_exit(cookie);
|
|
return ret;
|
|
}
|
|
|
|
static const struct vm_operations_struct panthor_mmio_vm_ops = {
|
|
.fault = panthor_mmio_vm_fault,
|
|
};
|
|
|
|
int panthor_device_mmap_io(struct panthor_device *ptdev, struct vm_area_struct *vma)
|
|
{
|
|
u64 offset = (u64)vma->vm_pgoff << PAGE_SHIFT;
|
|
|
|
switch (offset) {
|
|
case DRM_PANTHOR_USER_FLUSH_ID_MMIO_OFFSET:
|
|
if (vma->vm_end - vma->vm_start != PAGE_SIZE ||
|
|
(vma->vm_flags & (VM_WRITE | VM_EXEC)))
|
|
return -EINVAL;
|
|
|
|
break;
|
|
|
|
default:
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* Defer actual mapping to the fault handler. */
|
|
vma->vm_private_data = ptdev;
|
|
vma->vm_ops = &panthor_mmio_vm_ops;
|
|
vm_flags_set(vma,
|
|
VM_IO | VM_DONTCOPY | VM_DONTEXPAND |
|
|
VM_NORESERVE | VM_DONTDUMP | VM_PFNMAP);
|
|
return 0;
|
|
}
|
|
|
|
int panthor_device_resume(struct device *dev)
|
|
{
|
|
struct panthor_device *ptdev = dev_get_drvdata(dev);
|
|
int ret, cookie;
|
|
|
|
if (atomic_read(&ptdev->pm.state) != PANTHOR_DEVICE_PM_STATE_SUSPENDED)
|
|
return -EINVAL;
|
|
|
|
atomic_set(&ptdev->pm.state, PANTHOR_DEVICE_PM_STATE_RESUMING);
|
|
|
|
ret = clk_prepare_enable(ptdev->clks.core);
|
|
if (ret)
|
|
goto err_set_suspended;
|
|
|
|
ret = clk_prepare_enable(ptdev->clks.stacks);
|
|
if (ret)
|
|
goto err_disable_core_clk;
|
|
|
|
ret = clk_prepare_enable(ptdev->clks.coregroup);
|
|
if (ret)
|
|
goto err_disable_stacks_clk;
|
|
|
|
ret = panthor_devfreq_resume(ptdev);
|
|
if (ret)
|
|
goto err_disable_coregroup_clk;
|
|
|
|
if (panthor_device_is_initialized(ptdev) &&
|
|
drm_dev_enter(&ptdev->base, &cookie)) {
|
|
panthor_gpu_resume(ptdev);
|
|
panthor_mmu_resume(ptdev);
|
|
ret = drm_WARN_ON(&ptdev->base, panthor_fw_resume(ptdev));
|
|
if (!ret) {
|
|
panthor_sched_resume(ptdev);
|
|
} else {
|
|
panthor_mmu_suspend(ptdev);
|
|
panthor_gpu_suspend(ptdev);
|
|
}
|
|
|
|
drm_dev_exit(cookie);
|
|
|
|
if (ret)
|
|
goto err_suspend_devfreq;
|
|
}
|
|
|
|
if (atomic_read(&ptdev->reset.pending))
|
|
queue_work(ptdev->reset.wq, &ptdev->reset.work);
|
|
|
|
/* Clear all IOMEM mappings pointing to this device after we've
|
|
* resumed. This way the fake mappings pointing to the dummy pages
|
|
* are removed and the real iomem mapping will be restored on next
|
|
* access.
|
|
*/
|
|
mutex_lock(&ptdev->pm.mmio_lock);
|
|
unmap_mapping_range(ptdev->base.anon_inode->i_mapping,
|
|
DRM_PANTHOR_USER_MMIO_OFFSET, 0, 1);
|
|
atomic_set(&ptdev->pm.state, PANTHOR_DEVICE_PM_STATE_ACTIVE);
|
|
mutex_unlock(&ptdev->pm.mmio_lock);
|
|
return 0;
|
|
|
|
err_suspend_devfreq:
|
|
panthor_devfreq_suspend(ptdev);
|
|
|
|
err_disable_coregroup_clk:
|
|
clk_disable_unprepare(ptdev->clks.coregroup);
|
|
|
|
err_disable_stacks_clk:
|
|
clk_disable_unprepare(ptdev->clks.stacks);
|
|
|
|
err_disable_core_clk:
|
|
clk_disable_unprepare(ptdev->clks.core);
|
|
|
|
err_set_suspended:
|
|
atomic_set(&ptdev->pm.state, PANTHOR_DEVICE_PM_STATE_SUSPENDED);
|
|
return ret;
|
|
}
|
|
|
|
int panthor_device_suspend(struct device *dev)
|
|
{
|
|
struct panthor_device *ptdev = dev_get_drvdata(dev);
|
|
int ret, cookie;
|
|
|
|
if (atomic_read(&ptdev->pm.state) != PANTHOR_DEVICE_PM_STATE_ACTIVE)
|
|
return -EINVAL;
|
|
|
|
/* Clear all IOMEM mappings pointing to this device before we
|
|
* shutdown the power-domain and clocks. Failing to do that results
|
|
* in external aborts when the process accesses the iomem region.
|
|
* We change the state and call unmap_mapping_range() with the
|
|
* mmio_lock held to make sure the vm_fault handler won't set up
|
|
* invalid mappings.
|
|
*/
|
|
mutex_lock(&ptdev->pm.mmio_lock);
|
|
atomic_set(&ptdev->pm.state, PANTHOR_DEVICE_PM_STATE_SUSPENDING);
|
|
unmap_mapping_range(ptdev->base.anon_inode->i_mapping,
|
|
DRM_PANTHOR_USER_MMIO_OFFSET, 0, 1);
|
|
mutex_unlock(&ptdev->pm.mmio_lock);
|
|
|
|
if (panthor_device_is_initialized(ptdev) &&
|
|
drm_dev_enter(&ptdev->base, &cookie)) {
|
|
cancel_work_sync(&ptdev->reset.work);
|
|
|
|
/* We prepare everything as if we were resetting the GPU.
|
|
* The end of the reset will happen in the resume path though.
|
|
*/
|
|
panthor_sched_suspend(ptdev);
|
|
panthor_fw_suspend(ptdev);
|
|
panthor_mmu_suspend(ptdev);
|
|
panthor_gpu_suspend(ptdev);
|
|
drm_dev_exit(cookie);
|
|
}
|
|
|
|
ret = panthor_devfreq_suspend(ptdev);
|
|
if (ret) {
|
|
if (panthor_device_is_initialized(ptdev) &&
|
|
drm_dev_enter(&ptdev->base, &cookie)) {
|
|
panthor_gpu_resume(ptdev);
|
|
panthor_mmu_resume(ptdev);
|
|
drm_WARN_ON(&ptdev->base, panthor_fw_resume(ptdev));
|
|
panthor_sched_resume(ptdev);
|
|
drm_dev_exit(cookie);
|
|
}
|
|
|
|
goto err_set_active;
|
|
}
|
|
|
|
clk_disable_unprepare(ptdev->clks.coregroup);
|
|
clk_disable_unprepare(ptdev->clks.stacks);
|
|
clk_disable_unprepare(ptdev->clks.core);
|
|
atomic_set(&ptdev->pm.state, PANTHOR_DEVICE_PM_STATE_SUSPENDED);
|
|
return 0;
|
|
|
|
err_set_active:
|
|
/* If something failed and we have to revert back to an
|
|
* active state, we also need to clear the MMIO userspace
|
|
* mappings, so any dumb pages that were mapped while we
|
|
* were trying to suspend gets invalidated.
|
|
*/
|
|
mutex_lock(&ptdev->pm.mmio_lock);
|
|
atomic_set(&ptdev->pm.state, PANTHOR_DEVICE_PM_STATE_ACTIVE);
|
|
unmap_mapping_range(ptdev->base.anon_inode->i_mapping,
|
|
DRM_PANTHOR_USER_MMIO_OFFSET, 0, 1);
|
|
mutex_unlock(&ptdev->pm.mmio_lock);
|
|
return ret;
|
|
}
|