/* * arch/sh/mm/tlb-pteaex.c * * TLB operations for SH-X3 CPUs featuring PTE ASID Extensions. * * Copyright (C) 2009 Paul Mundt * * This file is subject to the terms and conditions of the GNU General Public * License. See the file "COPYING" in the main directory of this archive * for more details. */ #include #include #include #include #include #include void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t pte) { unsigned long flags, pteval, vpn; /* * Handle debugger faulting in for debugee. */ if (vma && current->active_mm != vma->vm_mm) return; local_irq_save(flags); /* Set PTEH register */ vpn = address & MMU_VPN_MASK; __raw_writel(vpn, MMU_PTEH); /* Set PTEAEX */ __raw_writel(get_asid(), MMU_PTEAEX); pteval = pte.pte_low; /* Set PTEA register */ #ifdef CONFIG_X2TLB /* * For the extended mode TLB this is trivial, only the ESZ and * EPR bits need to be written out to PTEA, with the remainder of * the protection bits (with the exception of the compat-mode SZ * and PR bits, which are cleared) being written out in PTEL. */ __raw_writel(pte.pte_high, MMU_PTEA); #endif /* Set PTEL register */ pteval &= _PAGE_FLAGS_HARDWARE_MASK; /* drop software flags */ #ifdef CONFIG_CACHE_WRITETHROUGH pteval |= _PAGE_WT; #endif /* conveniently, we want all the software flags to be 0 anyway */ __raw_writel(pteval, MMU_PTEL); /* Load the TLB */ asm volatile("ldtlb": /* no output */ : /* no input */ : "memory"); local_irq_restore(flags); } /* * While SH-X2 extended TLB mode splits out the memory-mapped I/UTLB * data arrays, SH-X3 cores with PTEAEX split out the memory-mapped * address arrays. In compat mode the second array is inaccessible, while * in extended mode, the legacy 8-bit ASID field in address array 1 has * undefined behaviour. */ void __uses_jump_to_uncached local_flush_tlb_one(unsigned long asid, unsigned long page) { jump_to_uncached(); __raw_writel(page, MMU_UTLB_ADDRESS_ARRAY | MMU_PAGE_ASSOC_BIT); __raw_writel(asid, MMU_UTLB_ADDRESS_ARRAY2 | MMU_PAGE_ASSOC_BIT); back_to_cached(); } /* * Load the entry for 'addr' into the TLB and wire the entry. */ void tlb_wire_entry(struct vm_area_struct *vma, unsigned long addr, pte_t pte) { unsigned long status, flags; int urb; local_irq_save(flags); /* Load the entry into the TLB */ __update_tlb(vma, addr, pte); /* ... and wire it up. */ status = ctrl_inl(MMUCR); urb = (status & MMUCR_URB) >> MMUCR_URB_SHIFT; status &= ~MMUCR_URB; /* * Make sure we're not trying to wire the last TLB entry slot. */ BUG_ON(!--urb); urb = urb % MMUCR_URB_NENTRIES; status |= (urb << MMUCR_URB_SHIFT); ctrl_outl(status, MMUCR); ctrl_barrier(); local_irq_restore(flags); } /* * Unwire the last wired TLB entry. * * It should also be noted that it is not possible to wire and unwire * TLB entries in an arbitrary order. If you wire TLB entry N, followed * by entry N+1, you must unwire entry N+1 first, then entry N. In this * respect, it works like a stack or LIFO queue. */ void tlb_unwire_entry(void) { unsigned long status, flags; int urb; local_irq_save(flags); status = ctrl_inl(MMUCR); urb = (status & MMUCR_URB) >> MMUCR_URB_SHIFT; status &= ~MMUCR_URB; /* * Make sure we're not trying to unwire a TLB entry when none * have been wired. */ BUG_ON(urb++ == MMUCR_URB_NENTRIES); urb = urb % MMUCR_URB_NENTRIES; status |= (urb << MMUCR_URB_SHIFT); ctrl_outl(status, MMUCR); ctrl_barrier(); local_irq_restore(flags); }