/* SPDX-License-Identifier: GPL-2.0 */ /* * Shadow Call Stack support. * * Copyright (C) 2019 Google LLC */ #ifndef _LINUX_SCS_H #define _LINUX_SCS_H #include #include #include #include #ifdef CONFIG_SHADOW_CALL_STACK /* * In testing, 1 KiB shadow stack size (i.e. 128 stack frames on a 64-bit * architecture) provided ~40% safety margin on stack usage while keeping * memory allocation overhead reasonable. */ #define SCS_SIZE SZ_1K #define GFP_SCS (GFP_KERNEL | __GFP_ZERO) /* An illegal pointer value to mark the end of the shadow stack. */ #define SCS_END_MAGIC (0x5f6UL + POISON_POINTER_DELTA) /* Allocate a static per-CPU shadow stack */ #define DEFINE_SCS(name) \ DEFINE_PER_CPU(unsigned long [SCS_SIZE/sizeof(long)], name) \ #define task_scs(tsk) (task_thread_info(tsk)->scs_base) #define task_scs_sp(tsk) (task_thread_info(tsk)->scs_sp) void scs_init(void); int scs_prepare(struct task_struct *tsk, int node); void scs_release(struct task_struct *tsk); static inline void scs_task_reset(struct task_struct *tsk) { /* * Reset the shadow stack to the base address in case the task * is reused. */ task_scs_sp(tsk) = task_scs(tsk); } static inline unsigned long *__scs_magic(void *s) { return (unsigned long *)(s + SCS_SIZE) - 1; } static inline bool task_scs_end_corrupted(struct task_struct *tsk) { unsigned long *magic = __scs_magic(task_scs(tsk)); unsigned long sz = task_scs_sp(tsk) - task_scs(tsk); return sz >= SCS_SIZE - 1 || READ_ONCE_NOCHECK(*magic) != SCS_END_MAGIC; } #else /* CONFIG_SHADOW_CALL_STACK */ static inline void scs_init(void) {} static inline void scs_task_reset(struct task_struct *tsk) {} static inline int scs_prepare(struct task_struct *tsk, int node) { return 0; } static inline void scs_release(struct task_struct *tsk) {} static inline bool task_scs_end_corrupted(struct task_struct *tsk) { return false; } #endif /* CONFIG_SHADOW_CALL_STACK */ #endif /* _LINUX_SCS_H */