#ifndef _ASM_POWERPC_PTE_WALK_H #define _ASM_POWERPC_PTE_WALK_H #include /* Don't use this directly */ extern pte_t *__find_linux_pte(pgd_t *pgdir, unsigned long ea, bool *is_thp, unsigned *hshift); static inline pte_t *find_linux_pte(pgd_t *pgdir, unsigned long ea, bool *is_thp, unsigned *hshift) { VM_WARN(!arch_irqs_disabled(), "%s called with irq enabled\n", __func__); return __find_linux_pte(pgdir, ea, is_thp, hshift); } static inline pte_t *find_init_mm_pte(unsigned long ea, unsigned *hshift) { pgd_t *pgdir = init_mm.pgd; return __find_linux_pte(pgdir, ea, NULL, hshift); } /* * This is what we should always use. Any other lockless page table lookup needs * careful audit against THP split. */ static inline pte_t *find_current_mm_pte(pgd_t *pgdir, unsigned long ea, bool *is_thp, unsigned *hshift) { VM_WARN(!arch_irqs_disabled(), "%s called with irq enabled\n", __func__); VM_WARN(pgdir != current->mm->pgd, "%s lock less page table lookup called on wrong mm\n", __func__); return __find_linux_pte(pgdir, ea, is_thp, hshift); } #endif /* _ASM_POWERPC_PTE_WALK_H */