diff options
author | Alexei Starovoitov <ast@kernel.org> | 2021-11-07 11:54:51 -0800 |
---|---|---|
committer | Alexei Starovoitov <ast@kernel.org> | 2021-11-07 11:54:51 -0800 |
commit | 0cc78dcca36de43f6ec4a94ddf64ddfa9bb36024 (patch) | |
tree | 0adf410fb0caebdb0252f340a7fd0b27ff6c56e3 /kernel/bpf/mmap_unlock_work.h | |
parent | 5fd79ed9bed1934de0ba0024e62eacaedb3c4c8a (diff) | |
parent | f108662b27c96cdadfadd39f0c0d650704cd593d (diff) |
Merge branch 'introduce bpf_find_vma'
Song Liu says:
====================
Changes v4 => v5:
1. Clean up and style change in 2/2. (Andrii)
Changes v3 => v4:
1. Move mmap_unlock_work to task_iter.c to fix build for .config without
!CONFIG_PERF_EVENTS. (kernel test robot <lkp@intel.com>)
Changes v2 => v3:
1. Avoid using x86 only function in selftests. (Yonghong)
2. Add struct file and struct vm_area_struct to btf_task_struct_ids, and
use it in bpf_find_vma and stackmap.c. (Yonghong)
3. Fix inaccurate comments. (Yonghong)
Changes v1 => v2:
1. Share irq_work with stackmap.c. (Daniel)
2. Add tests for illegal writes to task/vma from the callback function.
(Daniel)
3. Other small fixes.
Add helper bpf_find_vma. This can be used in some profiling use cases. It
might also be useful for LSM.
====================
Signed-off-by: Alexei Starovoitov <ast@kernel.org>
Diffstat (limited to 'kernel/bpf/mmap_unlock_work.h')
-rw-r--r-- | kernel/bpf/mmap_unlock_work.h | 65 |
1 files changed, 65 insertions, 0 deletions
diff --git a/kernel/bpf/mmap_unlock_work.h b/kernel/bpf/mmap_unlock_work.h new file mode 100644 index 000000000000..5d18d7d85bef --- /dev/null +++ b/kernel/bpf/mmap_unlock_work.h @@ -0,0 +1,65 @@ +/* SPDX-License-Identifier: GPL-2.0-only */ +/* Copyright (c) 2021 Facebook + */ + +#ifndef __MMAP_UNLOCK_WORK_H__ +#define __MMAP_UNLOCK_WORK_H__ +#include <linux/irq_work.h> + +/* irq_work to run mmap_read_unlock() in irq_work */ +struct mmap_unlock_irq_work { + struct irq_work irq_work; + struct mm_struct *mm; +}; + +DECLARE_PER_CPU(struct mmap_unlock_irq_work, mmap_unlock_work); + +/* + * We cannot do mmap_read_unlock() when the irq is disabled, because of + * risk to deadlock with rq_lock. To look up vma when the irqs are + * disabled, we need to run mmap_read_unlock() in irq_work. We use a + * percpu variable to do the irq_work. If the irq_work is already used + * by another lookup, we fall over. + */ +static inline bool bpf_mmap_unlock_get_irq_work(struct mmap_unlock_irq_work **work_ptr) +{ + struct mmap_unlock_irq_work *work = NULL; + bool irq_work_busy = false; + + if (irqs_disabled()) { + if (!IS_ENABLED(CONFIG_PREEMPT_RT)) { + work = this_cpu_ptr(&mmap_unlock_work); + if (irq_work_is_busy(&work->irq_work)) { + /* cannot queue more up_read, fallback */ + irq_work_busy = true; + } + } else { + /* + * PREEMPT_RT does not allow to trylock mmap sem in + * interrupt disabled context. Force the fallback code. + */ + irq_work_busy = true; + } + } + + *work_ptr = work; + return irq_work_busy; +} + +static inline void bpf_mmap_unlock_mm(struct mmap_unlock_irq_work *work, struct mm_struct *mm) +{ + if (!work) { + mmap_read_unlock(mm); + } else { + work->mm = mm; + + /* The lock will be released once we're out of interrupt + * context. Tell lockdep that we've released it now so + * it doesn't complain that we forgot to release it. + */ + rwsem_release(&mm->mmap_lock.dep_map, _RET_IP_); + irq_work_queue(&work->irq_work); + } +} + +#endif /* __MMAP_UNLOCK_WORK_H__ */ |