mirror of
https://github.com/torvalds/linux.git
synced 2024-11-10 14:11:52 +00:00
14c4c8e415
If rcu_read_lock_sched tracing is enabled, the tracing subsystem can
perform a jump which needs to be checked by CFI. For example, stm_ftrace
source is enabled as a module and hooks into enabled ftrace events. This
can cause an recursive loop where find_shadow_check_fn ->
rcu_read_lock_sched -> (call to stm_ftrace generates cfi slowpath) ->
find_shadow_check_fn -> rcu_read_lock_sched -> ...
To avoid the recursion, either the ftrace codes needs to be marked with
__no_cfi or CFI should not trace. Use the "_notrace" in CFI to avoid
tracing so that CFI can guard ftrace.
Signed-off-by: Elliot Berman <quic_eberman@quicinc.com>
Reviewed-by: Sami Tolvanen <samitolvanen@google.com>
Cc: stable@vger.kernel.org
Fixes: cf68fffb66
("add support for Clang CFI")
Signed-off-by: Kees Cook <keescook@chromium.org>
Link: https://lore.kernel.org/r/20210811155914.19550-1-quic_eberman@quicinc.com
330 lines
8.1 KiB
C
330 lines
8.1 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Clang Control Flow Integrity (CFI) error and slowpath handling.
|
|
*
|
|
* Copyright (C) 2021 Google LLC
|
|
*/
|
|
|
|
#include <linux/hardirq.h>
|
|
#include <linux/kallsyms.h>
|
|
#include <linux/module.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/printk.h>
|
|
#include <linux/ratelimit.h>
|
|
#include <linux/rcupdate.h>
|
|
#include <linux/vmalloc.h>
|
|
#include <asm/cacheflush.h>
|
|
#include <asm/set_memory.h>
|
|
|
|
/* Compiler-defined handler names */
|
|
#ifdef CONFIG_CFI_PERMISSIVE
|
|
#define cfi_failure_handler __ubsan_handle_cfi_check_fail
|
|
#else
|
|
#define cfi_failure_handler __ubsan_handle_cfi_check_fail_abort
|
|
#endif
|
|
|
|
static inline void handle_cfi_failure(void *ptr)
|
|
{
|
|
if (IS_ENABLED(CONFIG_CFI_PERMISSIVE))
|
|
WARN_RATELIMIT(1, "CFI failure (target: %pS):\n", ptr);
|
|
else
|
|
panic("CFI failure (target: %pS)\n", ptr);
|
|
}
|
|
|
|
#ifdef CONFIG_MODULES
|
|
#ifdef CONFIG_CFI_CLANG_SHADOW
|
|
/*
|
|
* Index type. A 16-bit index can address at most (2^16)-2 pages (taking
|
|
* into account SHADOW_INVALID), i.e. ~256M with 4k pages.
|
|
*/
|
|
typedef u16 shadow_t;
|
|
#define SHADOW_INVALID ((shadow_t)~0UL)
|
|
|
|
struct cfi_shadow {
|
|
/* Page index for the beginning of the shadow */
|
|
unsigned long base;
|
|
/* An array of __cfi_check locations (as indices to the shadow) */
|
|
shadow_t shadow[1];
|
|
} __packed;
|
|
|
|
/*
|
|
* The shadow covers ~128M from the beginning of the module region. If
|
|
* the region is larger, we fall back to __module_address for the rest.
|
|
*/
|
|
#define __SHADOW_RANGE (_UL(SZ_128M) >> PAGE_SHIFT)
|
|
|
|
/* The in-memory size of struct cfi_shadow, always at least one page */
|
|
#define __SHADOW_PAGES ((__SHADOW_RANGE * sizeof(shadow_t)) >> PAGE_SHIFT)
|
|
#define SHADOW_PAGES max(1UL, __SHADOW_PAGES)
|
|
#define SHADOW_SIZE (SHADOW_PAGES << PAGE_SHIFT)
|
|
|
|
/* The actual size of the shadow array, minus metadata */
|
|
#define SHADOW_ARR_SIZE (SHADOW_SIZE - offsetof(struct cfi_shadow, shadow))
|
|
#define SHADOW_ARR_SLOTS (SHADOW_ARR_SIZE / sizeof(shadow_t))
|
|
|
|
static DEFINE_MUTEX(shadow_update_lock);
|
|
static struct cfi_shadow __rcu *cfi_shadow __read_mostly;
|
|
|
|
/* Returns the index in the shadow for the given address */
|
|
static inline int ptr_to_shadow(const struct cfi_shadow *s, unsigned long ptr)
|
|
{
|
|
unsigned long index;
|
|
unsigned long page = ptr >> PAGE_SHIFT;
|
|
|
|
if (unlikely(page < s->base))
|
|
return -1; /* Outside of module area */
|
|
|
|
index = page - s->base;
|
|
|
|
if (index >= SHADOW_ARR_SLOTS)
|
|
return -1; /* Cannot be addressed with shadow */
|
|
|
|
return (int)index;
|
|
}
|
|
|
|
/* Returns the page address for an index in the shadow */
|
|
static inline unsigned long shadow_to_ptr(const struct cfi_shadow *s,
|
|
int index)
|
|
{
|
|
if (unlikely(index < 0 || index >= SHADOW_ARR_SLOTS))
|
|
return 0;
|
|
|
|
return (s->base + index) << PAGE_SHIFT;
|
|
}
|
|
|
|
/* Returns the __cfi_check function address for the given shadow location */
|
|
static inline unsigned long shadow_to_check_fn(const struct cfi_shadow *s,
|
|
int index)
|
|
{
|
|
if (unlikely(index < 0 || index >= SHADOW_ARR_SLOTS))
|
|
return 0;
|
|
|
|
if (unlikely(s->shadow[index] == SHADOW_INVALID))
|
|
return 0;
|
|
|
|
/* __cfi_check is always page aligned */
|
|
return (s->base + s->shadow[index]) << PAGE_SHIFT;
|
|
}
|
|
|
|
static void prepare_next_shadow(const struct cfi_shadow __rcu *prev,
|
|
struct cfi_shadow *next)
|
|
{
|
|
int i, index, check;
|
|
|
|
/* Mark everything invalid */
|
|
memset(next->shadow, 0xFF, SHADOW_ARR_SIZE);
|
|
|
|
if (!prev)
|
|
return; /* No previous shadow */
|
|
|
|
/* If the base address didn't change, an update is not needed */
|
|
if (prev->base == next->base) {
|
|
memcpy(next->shadow, prev->shadow, SHADOW_ARR_SIZE);
|
|
return;
|
|
}
|
|
|
|
/* Convert the previous shadow to the new address range */
|
|
for (i = 0; i < SHADOW_ARR_SLOTS; ++i) {
|
|
if (prev->shadow[i] == SHADOW_INVALID)
|
|
continue;
|
|
|
|
index = ptr_to_shadow(next, shadow_to_ptr(prev, i));
|
|
if (index < 0)
|
|
continue;
|
|
|
|
check = ptr_to_shadow(next,
|
|
shadow_to_check_fn(prev, prev->shadow[i]));
|
|
if (check < 0)
|
|
continue;
|
|
|
|
next->shadow[index] = (shadow_t)check;
|
|
}
|
|
}
|
|
|
|
static void add_module_to_shadow(struct cfi_shadow *s, struct module *mod,
|
|
unsigned long min_addr, unsigned long max_addr)
|
|
{
|
|
int check_index;
|
|
unsigned long check = (unsigned long)mod->cfi_check;
|
|
unsigned long ptr;
|
|
|
|
if (unlikely(!PAGE_ALIGNED(check))) {
|
|
pr_warn("cfi: not using shadow for module %s\n", mod->name);
|
|
return;
|
|
}
|
|
|
|
check_index = ptr_to_shadow(s, check);
|
|
if (check_index < 0)
|
|
return; /* Module not addressable with shadow */
|
|
|
|
/* For each page, store the check function index in the shadow */
|
|
for (ptr = min_addr; ptr <= max_addr; ptr += PAGE_SIZE) {
|
|
int index = ptr_to_shadow(s, ptr);
|
|
|
|
if (index >= 0) {
|
|
/* Each page must only contain one module */
|
|
WARN_ON_ONCE(s->shadow[index] != SHADOW_INVALID);
|
|
s->shadow[index] = (shadow_t)check_index;
|
|
}
|
|
}
|
|
}
|
|
|
|
static void remove_module_from_shadow(struct cfi_shadow *s, struct module *mod,
|
|
unsigned long min_addr, unsigned long max_addr)
|
|
{
|
|
unsigned long ptr;
|
|
|
|
for (ptr = min_addr; ptr <= max_addr; ptr += PAGE_SIZE) {
|
|
int index = ptr_to_shadow(s, ptr);
|
|
|
|
if (index >= 0)
|
|
s->shadow[index] = SHADOW_INVALID;
|
|
}
|
|
}
|
|
|
|
typedef void (*update_shadow_fn)(struct cfi_shadow *, struct module *,
|
|
unsigned long min_addr, unsigned long max_addr);
|
|
|
|
static void update_shadow(struct module *mod, unsigned long base_addr,
|
|
update_shadow_fn fn)
|
|
{
|
|
struct cfi_shadow *prev;
|
|
struct cfi_shadow *next;
|
|
unsigned long min_addr, max_addr;
|
|
|
|
next = vmalloc(SHADOW_SIZE);
|
|
|
|
mutex_lock(&shadow_update_lock);
|
|
prev = rcu_dereference_protected(cfi_shadow,
|
|
mutex_is_locked(&shadow_update_lock));
|
|
|
|
if (next) {
|
|
next->base = base_addr >> PAGE_SHIFT;
|
|
prepare_next_shadow(prev, next);
|
|
|
|
min_addr = (unsigned long)mod->core_layout.base;
|
|
max_addr = min_addr + mod->core_layout.text_size;
|
|
fn(next, mod, min_addr & PAGE_MASK, max_addr & PAGE_MASK);
|
|
|
|
set_memory_ro((unsigned long)next, SHADOW_PAGES);
|
|
}
|
|
|
|
rcu_assign_pointer(cfi_shadow, next);
|
|
mutex_unlock(&shadow_update_lock);
|
|
synchronize_rcu();
|
|
|
|
if (prev) {
|
|
set_memory_rw((unsigned long)prev, SHADOW_PAGES);
|
|
vfree(prev);
|
|
}
|
|
}
|
|
|
|
void cfi_module_add(struct module *mod, unsigned long base_addr)
|
|
{
|
|
update_shadow(mod, base_addr, add_module_to_shadow);
|
|
}
|
|
|
|
void cfi_module_remove(struct module *mod, unsigned long base_addr)
|
|
{
|
|
update_shadow(mod, base_addr, remove_module_from_shadow);
|
|
}
|
|
|
|
static inline cfi_check_fn ptr_to_check_fn(const struct cfi_shadow __rcu *s,
|
|
unsigned long ptr)
|
|
{
|
|
int index;
|
|
|
|
if (unlikely(!s))
|
|
return NULL; /* No shadow available */
|
|
|
|
index = ptr_to_shadow(s, ptr);
|
|
if (index < 0)
|
|
return NULL; /* Cannot be addressed with shadow */
|
|
|
|
return (cfi_check_fn)shadow_to_check_fn(s, index);
|
|
}
|
|
|
|
static inline cfi_check_fn find_shadow_check_fn(unsigned long ptr)
|
|
{
|
|
cfi_check_fn fn;
|
|
|
|
rcu_read_lock_sched_notrace();
|
|
fn = ptr_to_check_fn(rcu_dereference_sched(cfi_shadow), ptr);
|
|
rcu_read_unlock_sched_notrace();
|
|
|
|
return fn;
|
|
}
|
|
|
|
#else /* !CONFIG_CFI_CLANG_SHADOW */
|
|
|
|
static inline cfi_check_fn find_shadow_check_fn(unsigned long ptr)
|
|
{
|
|
return NULL;
|
|
}
|
|
|
|
#endif /* CONFIG_CFI_CLANG_SHADOW */
|
|
|
|
static inline cfi_check_fn find_module_check_fn(unsigned long ptr)
|
|
{
|
|
cfi_check_fn fn = NULL;
|
|
struct module *mod;
|
|
|
|
rcu_read_lock_sched_notrace();
|
|
mod = __module_address(ptr);
|
|
if (mod)
|
|
fn = mod->cfi_check;
|
|
rcu_read_unlock_sched_notrace();
|
|
|
|
return fn;
|
|
}
|
|
|
|
static inline cfi_check_fn find_check_fn(unsigned long ptr)
|
|
{
|
|
cfi_check_fn fn = NULL;
|
|
|
|
if (is_kernel_text(ptr))
|
|
return __cfi_check;
|
|
|
|
/*
|
|
* Indirect call checks can happen when RCU is not watching. Both
|
|
* the shadow and __module_address use RCU, so we need to wake it
|
|
* up if necessary.
|
|
*/
|
|
RCU_NONIDLE({
|
|
if (IS_ENABLED(CONFIG_CFI_CLANG_SHADOW))
|
|
fn = find_shadow_check_fn(ptr);
|
|
|
|
if (!fn)
|
|
fn = find_module_check_fn(ptr);
|
|
});
|
|
|
|
return fn;
|
|
}
|
|
|
|
void __cfi_slowpath_diag(uint64_t id, void *ptr, void *diag)
|
|
{
|
|
cfi_check_fn fn = find_check_fn((unsigned long)ptr);
|
|
|
|
if (likely(fn))
|
|
fn(id, ptr, diag);
|
|
else /* Don't allow unchecked modules */
|
|
handle_cfi_failure(ptr);
|
|
}
|
|
EXPORT_SYMBOL(__cfi_slowpath_diag);
|
|
|
|
#else /* !CONFIG_MODULES */
|
|
|
|
void __cfi_slowpath_diag(uint64_t id, void *ptr, void *diag)
|
|
{
|
|
handle_cfi_failure(ptr); /* No modules */
|
|
}
|
|
EXPORT_SYMBOL(__cfi_slowpath_diag);
|
|
|
|
#endif /* CONFIG_MODULES */
|
|
|
|
void cfi_failure_handler(void *data, void *ptr, void *vtable)
|
|
{
|
|
handle_cfi_failure(ptr);
|
|
}
|
|
EXPORT_SYMBOL(cfi_failure_handler);
|