Home
last modified time | relevance | path

Searched refs:pteval (Results 1 – 25 of 32) sorted by relevance

12

/linux/mm/
H A Drmap.c742 static void set_tlb_ubc_flush_pending(struct mm_struct *mm, pte_t pteval, in set_tlb_ubc_flush_pending() argument
747 bool writable = pte_dirty(pteval); in set_tlb_ubc_flush_pending()
749 if (!pte_accessible(mm, pteval)) in set_tlb_ubc_flush_pending()
827 static void set_tlb_ubc_flush_pending(struct mm_struct *mm, pte_t pteval, in set_tlb_ubc_flush_pending() argument
971 pte_t pteval = ptep_get(pvmw.pte); in folio_referenced_one() local
973 nr = folio_pte_batch(folio, pvmw.pte, pteval, max_nr); in folio_referenced_one()
1990 pte_t pteval; in try_to_unmap_one() local
2100 pteval = ptep_get(pvmw.pte); in try_to_unmap_one()
2101 if (likely(pte_present(pteval))) { in try_to_unmap_one()
2102 pfn = pte_pfn(pteval); in try_to_unmap_one()
[all …]
H A Dmemory.c1595 struct zap_details *details, pte_t pteval) in zap_install_uffd_wp_if_needed() argument
1611 if (pte_install_uffd_wp_if_needed(vma, addr, pte, pteval)) in zap_install_uffd_wp_if_needed()
2340 pte_t pteval = ptep_get(pte); in insert_page_into_pte_locked() local
2342 if (!pte_none(pteval)) { in insert_page_into_pte_locked()
2347 if (pte_pfn(pteval) != page_to_pfn(page)) { in insert_page_into_pte_locked()
2348 WARN_ON_ONCE(!is_zero_pfn(pte_pfn(pteval))); in insert_page_into_pte_locked()
2351 pteval = maybe_mkwrite(pteval, vma); in insert_page_into_pte_locked()
2352 pteval = pte_mkyoung(pteval); in insert_page_into_pte_locked()
2353 if (ptep_set_access_flags(vma, addr, pte, pteval, 1)) in insert_page_into_pte_locked()
2359 pteval = mk_pte(page, prot); in insert_page_into_pte_locked()
[all …]
/linux/drivers/staging/media/ipu3/
H A Dipu3-mmu.c121 static u32 *imgu_mmu_alloc_page_table(u32 pteval) in imgu_mmu_alloc_page_table() argument
131 pt[pte] = pteval; in imgu_mmu_alloc_page_table()
172 u32 pteval; in imgu_mmu_get_l2pt() local
203 pteval = IPU3_ADDR2PTE(virt_to_phys(new_l2pt)); in imgu_mmu_get_l2pt()
204 mmu->l1pt[l1pt_idx] = pteval; in imgu_mmu_get_l2pt()
430 u32 pteval; in imgu_mmu_init() local
450 pteval = IPU3_ADDR2PTE(virt_to_phys(mmu->dummy_page)); in imgu_mmu_init()
451 mmu->dummy_page_pteval = pteval; in imgu_mmu_init()
457 mmu->dummy_l2pt = imgu_mmu_alloc_page_table(pteval); in imgu_mmu_init()
460 pteval = IPU3_ADDR2PTE(virt_to_phys(mmu->dummy_l2pt)); in imgu_mmu_init()
[all …]
/linux/arch/sh/mm/
H A Dtlb-sh4.c18 unsigned long flags, pteval, vpn; in __update_tlb() local
32 pteval = pte.pte_low; in __update_tlb()
48 __raw_writel(copy_ptea_attributes(pteval), MMU_PTEA); in __update_tlb()
53 pteval &= _PAGE_FLAGS_HARDWARE_MASK; /* drop software flags */ in __update_tlb()
55 pteval |= _PAGE_WT; in __update_tlb()
58 __raw_writel(pteval, MMU_PTEL); in __update_tlb()
H A Dtlb-pteaex.c20 unsigned long flags, pteval, vpn; in __update_tlb() local
37 pteval = pte.pte_low; in __update_tlb()
51 pteval &= _PAGE_FLAGS_HARDWARE_MASK; /* drop software flags */ in __update_tlb()
53 pteval |= _PAGE_WT; in __update_tlb()
56 __raw_writel(pteval, MMU_PTEL); in __update_tlb()
H A Dtlb-sh3.c29 unsigned long flags, pteval, vpn; in __update_tlb() local
43 pteval = pte_val(pte); in __update_tlb()
46 pteval &= _PAGE_FLAGS_HARDWARE_MASK; /* drop software flags */ in __update_tlb()
48 __raw_writel(pteval, MMU_PTEL); in __update_tlb()
/linux/include/trace/events/
H A Dxen.h135 TP_PROTO(pte_t *ptep, pte_t pteval),
136 TP_ARGS(ptep, pteval),
139 __field(pteval_t, pteval)
142 __entry->pteval = pteval.pte),
145 (int)sizeof(pteval_t) * 2, (unsigned long long)pte_val(native_make_pte(__entry->pteval)),
146 (int)sizeof(pteval_t) * 2, (unsigned long long)__entry->pteval)
151 TP_PROTO(pte_t *ptep, pte_t pteval), \
152 TP_ARGS(ptep, pteval))
260 pte_t *ptep, pte_t pteval),
261 TP_ARGS(mm, addr, ptep, pteval),
[all …]
/linux/arch/x86/mm/
H A Dpgtable_32.c27 void set_pte_vaddr(unsigned long vaddr, pte_t pteval) in set_pte_vaddr() argument
56 if (!pte_none(pteval)) in set_pte_vaddr()
57 set_pte_at(&init_mm, vaddr, pte, pteval); in set_pte_vaddr()
/linux/arch/sparc/include/asm/
H A Dhighmem.h54 #define arch_kmap_local_pre_map(vaddr, pteval) flush_cache_all() argument
56 #define arch_kmap_local_post_map(vaddr, pteval) flush_tlb_all() argument
/linux/arch/powerpc/mm/ptdump/
H A Dhashpagetable.c383 unsigned long addr, pteval, psize; in walk_pte() local
388 pteval = pte_val(*pte); in walk_pte()
397 ((pteval & H_PAGE_COMBO) == H_PAGE_COMBO || in walk_pte()
398 (pteval & H_PAGE_4K_PFN) == H_PAGE_4K_PFN)) in walk_pte()
404 if (((pteval & H_PAGE_HASHPTE) != H_PAGE_HASHPTE) in walk_pte()
409 addr, pteval); in walk_pte()
/linux/arch/arm/mm/
H A Dflush.c285 void __sync_icache_dcache(pte_t pteval) in __sync_icache_dcache() argument
291 if (cache_is_vipt_nonaliasing() && !pte_exec(pteval)) in __sync_icache_dcache()
294 pfn = pte_pfn(pteval); in __sync_icache_dcache()
312 if (pte_exec(pteval)) in __sync_icache_dcache()
/linux/arch/riscv/include/asm/
H A Dpgtable.h605 static inline void set_pte(pte_t *ptep, pte_t pteval) in set_pte() argument
607 WRITE_ONCE(*ptep, pteval); in set_pte()
612 static inline void __set_pte_at(struct mm_struct *mm, pte_t *ptep, pte_t pteval) in __set_pte_at() argument
614 if (pte_present(pteval) && pte_exec(pteval)) in __set_pte_at()
615 flush_icache_pte(mm, pteval); in __set_pte_at()
617 set_pte(ptep, pteval); in __set_pte_at()
623 pte_t *ptep, pte_t pteval, unsigned int nr) in set_ptes() argument
625 page_table_check_ptes_set(mm, addr, ptep, pteval, nr); in set_ptes()
628 __set_pte_at(mm, ptep, pteval); in set_ptes()
632 pte_val(pteval) += 1 << _PAGE_PFN_SHIFT; in set_ptes()
/linux/arch/xtensa/mm/
H A Dtlb.c183 unsigned int pteval; in get_pte_for_vaddr() local
202 pteval = pte_val(*pte); in get_pte_for_vaddr()
204 return pteval; in get_pte_for_vaddr()
/linux/arch/loongarch/include/asm/
H A Dhighmem.h38 #define arch_kmap_local_post_map(vaddr, pteval) local_flush_tlb_one(vaddr) argument
H A Dpgtable.h364 static inline void set_pte(pte_t *ptep, pte_t pteval) in set_pte() argument
366 WRITE_ONCE(*ptep, pteval); in set_pte()
369 if (pte_val(pteval) & _PAGE_GLOBAL) in set_pte()
/linux/arch/csky/include/asm/
H A Dhighmem.h37 #define arch_kmap_local_post_map(vaddr, pteval) kmap_flush_tlb(vaddr) argument
/linux/arch/microblaze/include/asm/
H A Dhighmem.h54 #define arch_kmap_local_post_map(vaddr, pteval) \ argument
/linux/arch/mips/include/asm/
H A Dhighmem.h55 #define arch_kmap_local_post_map(vaddr, pteval) local_flush_tlb_one(vaddr) argument
/linux/arch/powerpc/include/asm/
H A Dhighmem.h63 #define arch_kmap_local_post_map(vaddr, pteval) \ argument
/linux/arch/arm/include/asm/
H A Dhighmem.h66 #define arch_kmap_local_post_map(vaddr, pteval) \ argument
/linux/arch/x86/xen/
H A Dmmu_pv.c315 static bool xen_batched_set_pte(pte_t *ptep, pte_t pteval) in xen_batched_set_pte() argument
325 u.val = pte_val_ma(pteval); in xen_batched_set_pte()
333 static inline void __xen_set_pte(pte_t *ptep, pte_t pteval) in __xen_set_pte() argument
335 if (!xen_batched_set_pte(ptep, pteval)) { in __xen_set_pte()
343 u.val = pte_val_ma(pteval); in __xen_set_pte()
348 static void xen_set_pte(pte_t *ptep, pte_t pteval) in xen_set_pte() argument
350 trace_xen_mmu_set_pte(ptep, pteval); in xen_set_pte()
351 __xen_set_pte(ptep, pteval); in xen_set_pte()
423 pteval_t pteval = pte.pte; in xen_pte_val() local
425 return pte_mfn_to_pfn(pteval); in xen_pte_val()
/linux/arch/sh/include/asm/
H A Dpgtable_32.h307 #define set_pte(pteptr, pteval) (*(pteptr) = pteval) argument
/linux/arch/parisc/include/asm/
H A Dpgtable.h70 #define set_pte(pteptr, pteval) \ argument
72 *(pteptr) = (pteval); \
/linux/arch/s390/include/asm/
H A Dpgtable.h1856 unsigned long pteval; in mk_swap_pte() local
1858 pteval = _PAGE_INVALID | _PAGE_PROTECT; in mk_swap_pte()
1859 pteval |= (offset & __SWP_OFFSET_MASK) << __SWP_OFFSET_SHIFT; in mk_swap_pte()
1860 pteval |= (type & __SWP_TYPE_MASK) << __SWP_TYPE_SHIFT; in mk_swap_pte()
1861 return __pte(pteval); in mk_swap_pte()
/linux/drivers/media/pci/intel/ipu6/
H A Dipu6-mmu.c284 phys_addr_t pteval = TBL_PHYS_ADDR(l2_pt[l2_idx]); in l2_unmap() local
288 l2_idx, &pteval); in l2_unmap()

12