Searched refs:_PAGE_PRIVILEGED (Results 1 – 8 of 8) sorted by relevance
177 VM_WARN_ON_ONCE(oldval & _PAGE_PRIVILEGED); in __pte_flags_need_flush()178 VM_WARN_ON_ONCE(newval & _PAGE_PRIVILEGED); in __pte_flags_need_flush()
20 #define _PAGE_PRIVILEGED 0x00008 /* kernel access only */ macro116 #define _PAGE_KERNEL_RW (_PAGE_PRIVILEGED | _PAGE_RW | _PAGE_DIRTY)117 #define _PAGE_KERNEL_RO (_PAGE_PRIVILEGED | _PAGE_READ)118 #define _PAGE_KERNEL_ROX (_PAGE_PRIVILEGED | _PAGE_READ | _PAGE_EXEC)119 #define _PAGE_KERNEL_RWX (_PAGE_PRIVILEGED | _PAGE_DIRTY | _PAGE_RW | _PAGE_EXEC)524 return !(pte_raw(pte) & cpu_to_be64(_PAGE_PRIVILEGED)); in pte_user()716 if ((access & _PAGE_PRIVILEGED) != (ptev & _PAGE_PRIVILEGED)) in check_pte_access()
14 .mask = _PAGE_PRIVILEGED,
168 access |= _PAGE_PRIVILEGED; in cxl_handle_mm_fault()
303 if (pteflags & _PAGE_PRIVILEGED) { in htab_convert_pte_flags()1723 access |= _PAGE_PRIVILEGED; in DEFINE_INTERRUPT_HANDLER()1725 access &= ~_PAGE_PRIVILEGED; in DEFINE_INTERRUPT_HANDLER()
295 if (pte & _PAGE_PRIVILEGED) { in kvmppc_mmu_radix_xlate()301 if (!(pte & _PAGE_PRIVILEGED)) { in kvmppc_mmu_radix_xlate()
176 access |= _PAGE_PRIVILEGED; in xsl_fault_handler_bh()
187 _PAGE_PRESENT | _PAGE_READ | _PAGE_PRIVILEGED, in __spu_trap_data_map()