Searched refs:pmd_none (Results 1 – 25 of 64) sorted by relevance
123
66 if (!pmd_none(*pmd)) { in hash__flush_range()104 if (!pmd_none(*pmd)) in hash__flush_tlb_page()
237 if (pmd_none(*pmd)) in modify_pmd_table()252 } else if (pmd_none(*pmd)) { in modify_pmd_table()309 if (!pmd_none(*pmd)) in try_free_pmd_table()614 if (pmd_none(*pmd)) { in vmem_get_alloc_pte()
47 ((unlikely(pmd_none(*(pmd))) && \
148 if (pmd_none(*pmdp)) { in migrate_vma_collect_huge_pmd()234 if (pmd_none(pmdp_get_lockless(pmdp))) in migrate_vma_collect_huge_pmd()805 VM_WARN_ON_ONCE(!pmd_none(*pmdp) && !is_huge_zero_pmd(*pmdp)); in migrate_vma_insert_huge_pmd_page()862 if (!pmd_none(*pmdp)) { in migrate_vma_insert_huge_pmd_page()866 } else if (!pmd_none(*pmdp)) in migrate_vma_insert_huge_pmd_page()1008 if (!pmd_none(*pmdp)) { in migrate_vma_insert_page()
130 if (pmd_none(*pmd)) { in walk_pmd_range()958 if (pmd_none(pmd)) in folio_walk_start()970 if (pmd_none(pmd)) { in folio_walk_start()
298 if (unlikely(pmd_none(dst_pmdval)) && in mfill_establish_pmd()1390 if (pmd_none(*dst_pmd) || pmd_none(*src_pmd) || in move_pages_ptes()1945 !pmd_none(dst_pmdval)) { in move_pages()1971 if (pmd_none(*src_pmd)) { in move_pages()
51 if (pmd_none(*pmd)) { in set_pte_vaddr()
277 if (pmd_none(*pmd)) { in pti_user_pagetable_walk_pte()355 if (pmd_none(*pmd)) { in pti_clone_pgtable()
51 return (!pmd || pmd_none(pmdp_get(pmd))) ? NULL : (pte_t *) pmd; in huge_pte_offset()
46 BUG_ON(!pmd_none(*pmd)); in init_pmd()
30 BUG_ON(!pmd_none(*pmd)); in kasan_early_init()
174 if (pmd_none(READ_ONCE(*dst_pmdp))) { in temp_pgtable_map_pte()219 if (pmd_none(pmd)) in temp_pgtable_map_pmd()
152 #define pmd_none(pmd) (!pmd_val(pmd)) macro
65 if (pmd_none(READ_ONCE(*pmdp))) { in kasan_pte_offset()141 } while (pmdp++, addr = next, addr != end && pmd_none(READ_ONCE(*pmdp))); in kasan_pmd_populate()
61 if (pmd_none(*pmd)) { in resume_one_page_table_init()
61 if (pmd_none(*pmd_k)) in handle_kernel_vaddr_fault()
71 if (!pmd_none(*pmd)) in huge_pte_offset()
59 if (pmd_none(*pmdp)) { in early_pte_alloc_kernel()
155 #define pmd_none(x) (!pmd_val(x)) macro
82 if (pmd_none(*pmd)) in show_pte()294 if (pmd_none(pmd_k[index])) in vmalloc_fault()
100 if (pmd_none(*pmdp)) { in kasan_pmd_populate()
113 #define pmd_none(pmd) pmd_none2(&(pmd)) macro
137 if (pmd_none(*pmd)) in load_ksp_mmu()
114 #define pmd_none(pmd) (!pmd_val(pmd)) macro
34 if (pmd_none(pmdp_get(pmd))) { in kasan_populate_pte()66 if (pmd_none(pmdp_get(pmdp)) && IS_ALIGNED(vaddr, PMD_SIZE) && in kasan_populate_pmd()