mirror of
https://github.com/torvalds/linux.git
synced 2026-04-18 06:44:00 -04:00
bpf: fix mm lifecycle in open-coded task_vma iterator
The open-coded task_vma iterator reads task->mm locklessly and acquires
mmap_read_trylock() but never calls mmget(). If the task exits
concurrently, the mm_struct can be freed as it is not
SLAB_TYPESAFE_BY_RCU, resulting in a use-after-free.
Safely read task->mm with a trylock on alloc_lock and acquire an mm
reference. Drop the reference via bpf_iter_mmput_async() in _destroy()
and error paths. bpf_iter_mmput_async() is a local wrapper around
mmput_async() with a fallback to mmput() on !CONFIG_MMU.
Reject irqs-disabled contexts (including NMI) up front. Operations used
by _next() and _destroy() (mmap_read_unlock, bpf_iter_mmput_async)
take spinlocks with IRQs disabled (pool->lock, pi_lock). Running from
NMI or from a tracepoint that fires with those locks held could
deadlock.
A trylock on alloc_lock is used instead of the blocking task_lock()
(get_task_mm) to avoid a deadlock when a softirq BPF program iterates
a task that already holds its alloc_lock on the same CPU.
Fixes: 4ac4546821 ("bpf: Introduce task_vma open-coded iterator kfuncs")
Signed-off-by: Puranjay Mohan <puranjay@kernel.org>
Link: https://lore.kernel.org/r/20260408154539.3832150-2-puranjay@kernel.org
Signed-off-by: Alexei Starovoitov <ast@kernel.org>
This commit is contained in:
committed by
Alexei Starovoitov
parent
a0c584fc18
commit
d8e27d2d22
@@ -9,6 +9,7 @@
|
||||
#include <linux/bpf_mem_alloc.h>
|
||||
#include <linux/btf_ids.h>
|
||||
#include <linux/mm_types.h>
|
||||
#include <linux/sched/mm.h>
|
||||
#include "mmap_unlock_work.h"
|
||||
|
||||
static const char * const iter_task_type_names[] = {
|
||||
@@ -794,6 +795,15 @@ const struct bpf_func_proto bpf_find_vma_proto = {
|
||||
.arg5_type = ARG_ANYTHING,
|
||||
};
|
||||
|
||||
static inline void bpf_iter_mmput_async(struct mm_struct *mm)
|
||||
{
|
||||
#ifdef CONFIG_MMU
|
||||
mmput_async(mm);
|
||||
#else
|
||||
mmput(mm);
|
||||
#endif
|
||||
}
|
||||
|
||||
struct bpf_iter_task_vma_kern_data {
|
||||
struct task_struct *task;
|
||||
struct mm_struct *mm;
|
||||
@@ -825,6 +835,24 @@ __bpf_kfunc int bpf_iter_task_vma_new(struct bpf_iter_task_vma *it,
|
||||
BUILD_BUG_ON(sizeof(struct bpf_iter_task_vma_kern) != sizeof(struct bpf_iter_task_vma));
|
||||
BUILD_BUG_ON(__alignof__(struct bpf_iter_task_vma_kern) != __alignof__(struct bpf_iter_task_vma));
|
||||
|
||||
/* bpf_iter_mmput_async() needs mmput_async() which requires CONFIG_MMU */
|
||||
if (!IS_ENABLED(CONFIG_MMU)) {
|
||||
kit->data = NULL;
|
||||
return -EOPNOTSUPP;
|
||||
}
|
||||
|
||||
/*
|
||||
* Reject irqs-disabled contexts including NMI. Operations used
|
||||
* by _next() and _destroy() (mmap_read_unlock, bpf_iter_mmput_async)
|
||||
* can take spinlocks with IRQs disabled (pi_lock, pool->lock).
|
||||
* Running from NMI or from a tracepoint that fires with those
|
||||
* locks held could deadlock.
|
||||
*/
|
||||
if (irqs_disabled()) {
|
||||
kit->data = NULL;
|
||||
return -EBUSY;
|
||||
}
|
||||
|
||||
/* is_iter_reg_valid_uninit guarantees that kit hasn't been initialized
|
||||
* before, so non-NULL kit->data doesn't point to previously
|
||||
* bpf_mem_alloc'd bpf_iter_task_vma_kern_data
|
||||
@@ -834,7 +862,25 @@ __bpf_kfunc int bpf_iter_task_vma_new(struct bpf_iter_task_vma *it,
|
||||
return -ENOMEM;
|
||||
|
||||
kit->data->task = get_task_struct(task);
|
||||
/*
|
||||
* Safely read task->mm and acquire an mm reference.
|
||||
*
|
||||
* Cannot use get_task_mm() because its task_lock() is a
|
||||
* blocking spin_lock that would deadlock if the target task
|
||||
* already holds alloc_lock on this CPU (e.g. a softirq BPF
|
||||
* program iterating a task interrupted while holding its
|
||||
* alloc_lock).
|
||||
*/
|
||||
if (!spin_trylock(&task->alloc_lock)) {
|
||||
err = -EBUSY;
|
||||
goto err_cleanup_iter;
|
||||
}
|
||||
kit->data->mm = task->mm;
|
||||
if (kit->data->mm && !(task->flags & PF_KTHREAD))
|
||||
mmget(kit->data->mm);
|
||||
else
|
||||
kit->data->mm = NULL;
|
||||
spin_unlock(&task->alloc_lock);
|
||||
if (!kit->data->mm) {
|
||||
err = -ENOENT;
|
||||
goto err_cleanup_iter;
|
||||
@@ -844,15 +890,16 @@ __bpf_kfunc int bpf_iter_task_vma_new(struct bpf_iter_task_vma *it,
|
||||
irq_work_busy = bpf_mmap_unlock_get_irq_work(&kit->data->work);
|
||||
if (irq_work_busy || !mmap_read_trylock(kit->data->mm)) {
|
||||
err = -EBUSY;
|
||||
goto err_cleanup_iter;
|
||||
goto err_cleanup_mmget;
|
||||
}
|
||||
|
||||
vma_iter_init(&kit->data->vmi, kit->data->mm, addr);
|
||||
return 0;
|
||||
|
||||
err_cleanup_mmget:
|
||||
bpf_iter_mmput_async(kit->data->mm);
|
||||
err_cleanup_iter:
|
||||
if (kit->data->task)
|
||||
put_task_struct(kit->data->task);
|
||||
put_task_struct(kit->data->task);
|
||||
bpf_mem_free(&bpf_global_ma, kit->data);
|
||||
/* NULL kit->data signals failed bpf_iter_task_vma initialization */
|
||||
kit->data = NULL;
|
||||
@@ -875,6 +922,7 @@ __bpf_kfunc void bpf_iter_task_vma_destroy(struct bpf_iter_task_vma *it)
|
||||
if (kit->data) {
|
||||
bpf_mmap_unlock_mm(kit->data->work, kit->data->mm);
|
||||
put_task_struct(kit->data->task);
|
||||
bpf_iter_mmput_async(kit->data->mm);
|
||||
bpf_mem_free(&bpf_global_ma, kit->data);
|
||||
}
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user