Lines Matching refs:prot
245 vm_prot_t prot; in linux_get_user_pages_internal() local
249 prot = write ? (VM_PROT_READ | VM_PROT_WRITE) : VM_PROT_READ; in linux_get_user_pages_internal()
251 count = vm_fault_quick_hold_pages(map, start, len, prot, pages, nr_pages); in linux_get_user_pages_internal()
263 vm_prot_t prot; in __get_user_pages_fast() local
274 prot = write ? (VM_PROT_READ | VM_PROT_WRITE) : VM_PROT_READ; in __get_user_pages_fast()
277 *mp = pmap_extract_and_hold(map->pmap, va, prot); in __get_user_pages_fast()
281 if ((prot & VM_PROT_WRITE) != 0 && in __get_user_pages_fast()
416 vmap(struct page **pages, unsigned int count, unsigned long flags, int prot) in vmap() argument
434 linuxkpi_vmap_pfn(unsigned long *pfns, unsigned int count, int prot) in linuxkpi_vmap_pfn() argument
450 attr = pgprot2cachemode(prot); in linuxkpi_vmap_pfn()
501 unsigned long pfn, pgprot_t prot) in lkpi_vmf_insert_pfn_prot_locked() argument
559 pmap_page_set_memattr(page, pgprot2cachemode(prot)); in lkpi_vmf_insert_pfn_prot_locked()
567 unsigned long start_pfn, unsigned long size, pgprot_t prot) in lkpi_remap_pfn_range() argument
581 ret = lkpi_vmf_insert_pfn_prot_locked(vma, addr, pfn, prot); in lkpi_remap_pfn_range()
613 pgprot_t prot; in lkpi_io_mapping_map_user() local
616 prot = cachemode2protval(iomap->attr); in lkpi_io_mapping_map_user()
617 ret = lkpi_remap_pfn_range(vma, addr, pfn, size, prot); in lkpi_io_mapping_map_user()