// SPDX-License-Identifier: GPL-2.0 /* * Clang Control Flow Integrity (CFI) error and slowpath handling. * * Copyright (C) 2021 Google LLC */ #include #include #include #include #include #include #include #include #include #include /* Compiler-defined handler names */ #ifdef CONFIG_CFI_PERMISSIVE #define cfi_failure_handler __ubsan_handle_cfi_check_fail #else #define cfi_failure_handler __ubsan_handle_cfi_check_fail_abort #endif static inline void handle_cfi_failure(void *ptr) { if (IS_ENABLED(CONFIG_CFI_PERMISSIVE)) WARN_RATELIMIT(1, "CFI failure (target: %pS):\n", ptr); else panic("CFI failure (target: %pS)\n", ptr); } #ifdef CONFIG_MODULES #ifdef CONFIG_CFI_CLANG_SHADOW /* * Index type. A 16-bit index can address at most (2^16)-2 pages (taking * into account SHADOW_INVALID), i.e. ~256M with 4k pages. */ typedef u16 shadow_t; #define SHADOW_INVALID ((shadow_t)~0UL) struct cfi_shadow { /* Page index for the beginning of the shadow */ unsigned long base; /* An array of __cfi_check locations (as indices to the shadow) */ shadow_t shadow[1]; } __packed; /* * The shadow covers ~128M from the beginning of the module region. If * the region is larger, we fall back to __module_address for the rest. */ #define __SHADOW_RANGE (_UL(SZ_128M) >> PAGE_SHIFT) /* The in-memory size of struct cfi_shadow, always at least one page */ #define __SHADOW_PAGES ((__SHADOW_RANGE * sizeof(shadow_t)) >> PAGE_SHIFT) #define SHADOW_PAGES max(1UL, __SHADOW_PAGES) #define SHADOW_SIZE (SHADOW_PAGES << PAGE_SHIFT) /* The actual size of the shadow array, minus metadata */ #define SHADOW_ARR_SIZE (SHADOW_SIZE - offsetof(struct cfi_shadow, shadow)) #define SHADOW_ARR_SLOTS (SHADOW_ARR_SIZE / sizeof(shadow_t)) static DEFINE_MUTEX(shadow_update_lock); static struct cfi_shadow __rcu *cfi_shadow __read_mostly; /* Returns the index in the shadow for the given address */ static inline int ptr_to_shadow(const struct cfi_shadow *s, unsigned long ptr) { unsigned long index; unsigned long page = ptr >> PAGE_SHIFT; if (unlikely(page < s->base)) return -1; /* Outside of module area */ index = page - s->base; if (index >= SHADOW_ARR_SLOTS) return -1; /* Cannot be addressed with shadow */ return (int)index; } /* Returns the page address for an index in the shadow */ static inline unsigned long shadow_to_ptr(const struct cfi_shadow *s, int index) { if (unlikely(index < 0 || index >= SHADOW_ARR_SLOTS)) return 0; return (s->base + index) << PAGE_SHIFT; } /* Returns the __cfi_check function address for the given shadow location */ static inline unsigned long shadow_to_check_fn(const struct cfi_shadow *s, int index) { if (unlikely(index < 0 || index >= SHADOW_ARR_SLOTS)) return 0; if (unlikely(s->shadow[index] == SHADOW_INVALID)) return 0; /* __cfi_check is always page aligned */ return (s->base + s->shadow[index]) << PAGE_SHIFT; } static void prepare_next_shadow(const struct cfi_shadow __rcu *prev, struct cfi_shadow *next) { int i, index, check; /* Mark everything invalid */ memset(next->shadow, 0xFF, SHADOW_ARR_SIZE); if (!prev) return; /* No previous shadow */ /* If the base address didn't change, an update is not needed */ if (prev->base == next->base) { memcpy(next->shadow, prev->shadow, SHADOW_ARR_SIZE); return; } /* Convert the previous shadow to the new address range */ for (i = 0; i < SHADOW_ARR_SLOTS; ++i) { if (prev->shadow[i] == SHADOW_INVALID) continue; index = ptr_to_shadow(next, shadow_to_ptr(prev, i)); if (index < 0) continue; check = ptr_to_shadow(next, shadow_to_check_fn(prev, prev->shadow[i])); if (check < 0) continue; next->shadow[index] = (shadow_t)check; } } static void add_module_to_shadow(struct cfi_shadow *s, struct module *mod, unsigned long min_addr, unsigned long max_addr) { int check_index; unsigned long check = (unsigned long)mod->cfi_check; unsigned long ptr; if (unlikely(!PAGE_ALIGNED(check))) { pr_warn("cfi: not using shadow for module %s\n", mod->name); return; } check_index = ptr_to_shadow(s, check); if (check_index < 0) return; /* Module not addressable with shadow */ /* For each page, store the check function index in the shadow */ for (ptr = min_addr; ptr <= max_addr; ptr += PAGE_SIZE) { int index = ptr_to_shadow(s, ptr); if (index >= 0) { /* Each page must only contain one module */ WARN_ON_ONCE(s->shadow[index] != SHADOW_INVALID); s->shadow[index] = (shadow_t)check_index; } } } static void remove_module_from_shadow(struct cfi_shadow *s, struct module *mod, unsigned long min_addr, unsigned long max_addr) { unsigned long ptr; for (ptr = min_addr; ptr <= max_addr; ptr += PAGE_SIZE) { int index = ptr_to_shadow(s, ptr); if (index >= 0) s->shadow[index] = SHADOW_INVALID; } } typedef void (*update_shadow_fn)(struct cfi_shadow *, struct module *, unsigned long min_addr, unsigned long max_addr); static void update_shadow(struct module *mod, unsigned long base_addr, update_shadow_fn fn) { struct cfi_shadow *prev; struct cfi_shadow *next; unsigned long min_addr, max_addr; next = vmalloc(SHADOW_SIZE); mutex_lock(&shadow_update_lock); prev = rcu_dereference_protected(cfi_shadow, mutex_is_locked(&shadow_update_lock)); if (next) { next->base = base_addr >> PAGE_SHIFT; prepare_next_shadow(prev, next); min_addr = (unsigned long)mod->core_layout.base; max_addr = min_addr + mod->core_layout.text_size; fn(next, mod, min_addr & PAGE_MASK, max_addr & PAGE_MASK); set_memory_ro((unsigned long)next, SHADOW_PAGES); } rcu_assign_pointer(cfi_shadow, next); mutex_unlock(&shadow_update_lock); synchronize_rcu(); if (prev) { set_memory_rw((unsigned long)prev, SHADOW_PAGES); vfree(prev); } } void cfi_module_add(struct module *mod, unsigned long base_addr) { update_shadow(mod, base_addr, add_module_to_shadow); } void cfi_module_remove(struct module *mod, unsigned long base_addr) { update_shadow(mod, base_addr, remove_module_from_shadow); } static inline cfi_check_fn ptr_to_check_fn(const struct cfi_shadow __rcu *s, unsigned long ptr) { int index; if (unlikely(!s)) return NULL; /* No shadow available */ index = ptr_to_shadow(s, ptr); if (index < 0) return NULL; /* Cannot be addressed with shadow */ return (cfi_check_fn)shadow_to_check_fn(s, index); } static inline cfi_check_fn find_shadow_check_fn(unsigned long ptr) { cfi_check_fn fn; rcu_read_lock_sched_notrace(); fn = ptr_to_check_fn(rcu_dereference_sched(cfi_shadow), ptr); rcu_read_unlock_sched_notrace(); return fn; } #else /* !CONFIG_CFI_CLANG_SHADOW */ static inline cfi_check_fn find_shadow_check_fn(unsigned long ptr) { return NULL; } #endif /* CONFIG_CFI_CLANG_SHADOW */ static inline cfi_check_fn find_module_check_fn(unsigned long ptr) { cfi_check_fn fn = NULL; struct module *mod; rcu_read_lock_sched_notrace(); mod = __module_address(ptr); if (mod) fn = mod->cfi_check; rcu_read_unlock_sched_notrace(); return fn; } static inline cfi_check_fn find_check_fn(unsigned long ptr) { cfi_check_fn fn = NULL; if (is_kernel_text(ptr)) return __cfi_check; /* * Indirect call checks can happen when RCU is not watching. Both * the shadow and __module_address use RCU, so we need to wake it * up if necessary. */ RCU_NONIDLE({ if (IS_ENABLED(CONFIG_CFI_CLANG_SHADOW)) fn = find_shadow_check_fn(ptr); if (!fn) fn = find_module_check_fn(ptr); }); return fn; } void __cfi_slowpath_diag(uint64_t id, void *ptr, void *diag) { cfi_check_fn fn = find_check_fn((unsigned long)ptr); if (likely(fn)) fn(id, ptr, diag); else /* Don't allow unchecked modules */ handle_cfi_failure(ptr); } EXPORT_SYMBOL(__cfi_slowpath_diag); #else /* !CONFIG_MODULES */ void __cfi_slowpath_diag(uint64_t id, void *ptr, void *diag) { handle_cfi_failure(ptr); /* No modules */ } EXPORT_SYMBOL(__cfi_slowpath_diag); #endif /* CONFIG_MODULES */ void cfi_failure_handler(void *data, void *ptr, void *vtable) { handle_cfi_failure(ptr); } EXPORT_SYMBOL(cfi_failure_handler);