/* SPDX-License-Identifier: GPL-2.0-only */ /* * Copyright (C) 2004, 2007-2010, 2011-2012 Synopsys, Inc. (www.synopsys.com) */ /* * page table flags for software walked/managed MMUv3 (ARC700) and MMUv4 (HS) * There correspond to the corresponding bits in the TLB */ #ifndef _ASM_ARC_PGTABLE_BITS_ARCV2_H #define _ASM_ARC_PGTABLE_BITS_ARCV2_H #ifdef CONFIG_ARC_CACHE_PAGES #define _PAGE_CACHEABLE (1 << 0) /* Cached (H) */ #else #define _PAGE_CACHEABLE 0 #endif #define _PAGE_EXECUTE (1 << 1) /* User Execute (H) */ #define _PAGE_WRITE (1 << 2) /* User Write (H) */ #define _PAGE_READ (1 << 3) /* User Read (H) */ #define _PAGE_ACCESSED (1 << 4) /* Accessed (s) */ #define _PAGE_DIRTY (1 << 5) /* Modified (s) */ #define _PAGE_SPECIAL (1 << 6) #define _PAGE_GLOBAL (1 << 8) /* ASID agnostic (H) */ #define _PAGE_PRESENT (1 << 9) /* PTE/TLB Valid (H) */ #ifdef CONFIG_ARC_MMU_V4 #define _PAGE_HW_SZ (1 << 10) /* Normal/super (H) */ #else #define _PAGE_HW_SZ 0 #endif /* Defaults for every user page */ #define ___DEF (_PAGE_PRESENT | _PAGE_CACHEABLE) /* Set of bits not changed in pte_modify */ #define _PAGE_CHG_MASK (PAGE_MASK_PHYS | _PAGE_ACCESSED | _PAGE_DIRTY | \ _PAGE_SPECIAL) /* More Abbrevaited helpers */ #define PAGE_U_NONE __pgprot(___DEF) #define PAGE_U_R __pgprot(___DEF | _PAGE_READ) #define PAGE_U_W_R __pgprot(___DEF | _PAGE_READ | _PAGE_WRITE) #define PAGE_U_X_R __pgprot(___DEF | _PAGE_READ | _PAGE_EXECUTE) #define PAGE_U_X_W_R __pgprot(___DEF \ | _PAGE_READ | _PAGE_WRITE | _PAGE_EXECUTE) #define PAGE_KERNEL __pgprot(___DEF | _PAGE_GLOBAL \ | _PAGE_READ | _PAGE_WRITE | _PAGE_EXECUTE) #define PAGE_SHARED PAGE_U_W_R #define pgprot_noncached(prot) (__pgprot(pgprot_val(prot) & ~_PAGE_CACHEABLE)) /* * Mapping of vm_flags (Generic VM) to PTE flags (arch specific) * * Certain cases have 1:1 mapping * e.g. __P101 means VM_READ, VM_EXEC and !VM_SHARED * which directly corresponds to PAGE_U_X_R * * Other rules which cause the divergence from 1:1 mapping * * 1. Although ARC700 can do exclusive execute/write protection (meaning R * can be tracked independet of X/W unlike some other CPUs), still to * keep things consistent with other archs: * -Write implies Read: W => R * -Execute implies Read: X => R * * 2. Pvt Writable doesn't have Write Enabled initially: Pvt-W => !W * This is to enable COW mechanism */ /* xwr */ #ifndef __ASSEMBLY__ #define pte_write(pte) (pte_val(pte) & _PAGE_WRITE) #define pte_dirty(pte) (pte_val(pte) & _PAGE_DIRTY) #define pte_young(pte) (pte_val(pte) & _PAGE_ACCESSED) #define pte_special(pte) (pte_val(pte) & _PAGE_SPECIAL) #define PTE_BIT_FUNC(fn, op) \ static inline pte_t pte_##fn(pte_t pte) { pte_val(pte) op; return pte; } PTE_BIT_FUNC(mknotpresent, &= ~(_PAGE_PRESENT)); PTE_BIT_FUNC(wrprotect, &= ~(_PAGE_WRITE)); PTE_BIT_FUNC(mkwrite, |= (_PAGE_WRITE)); PTE_BIT_FUNC(mkclean, &= ~(_PAGE_DIRTY)); PTE_BIT_FUNC(mkdirty, |= (_PAGE_DIRTY)); PTE_BIT_FUNC(mkold, &= ~(_PAGE_ACCESSED)); PTE_BIT_FUNC(mkyoung, |= (_PAGE_ACCESSED)); PTE_BIT_FUNC(mkspecial, |= (_PAGE_SPECIAL)); PTE_BIT_FUNC(mkhuge, |= (_PAGE_HW_SZ)); static inline pte_t pte_modify(pte_t pte, pgprot_t newprot) { return __pte((pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot)); } static inline void set_pte_at(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t pteval) { set_pte(ptep, pteval); } void update_mmu_cache(struct vm_area_struct *vma, unsigned long address, pte_t *ptep); /* Encode swap {type,off} tuple into PTE * We reserve 13 bits for 5-bit @type, keeping bits 12-5 zero, ensuring that * PAGE_PRESENT is zero in a PTE holding swap "identifier" */ #define __swp_entry(type, off) ((swp_entry_t) \ { ((type) & 0x1f) | ((off) << 13) }) /* Decode a PTE containing swap "identifier "into constituents */ #define __swp_type(pte_lookalike) (((pte_lookalike).val) & 0x1f) #define __swp_offset(pte_lookalike) ((pte_lookalike).val >> 13) #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) #define __swp_entry_to_pte(x) ((pte_t) { (x).val }) #define kern_addr_valid(addr) (1) #ifdef CONFIG_TRANSPARENT_HUGEPAGE #include #endif #endif /* __ASSEMBLY__ */ #endif