Searched refs:vma_pages (Results 1 – 25 of 35) sorted by relevance
12
3 /// Use vma_pages function on vma object instead of explicit computation.6 // Keywords: vma_pages vma33 + vma_pages(vma)51 msg="WARNING: Consider using vma_pages helper on %s" % (x)59 msg="WARNING: Consider using vma_pages helper on %s" % (x)
32 if (vma->vm_pgoff + vma_pages(vma) > size) in pci_mmap_resource_range()67 nr = vma_pages(vma); in pci_mmap_fits()
167 if (vma_pages(vma) != 1) in check_mmap_afu_irq()190 if ((vma_pages(vma) + vma->vm_pgoff) > in check_mmap_mmio()
133 if ((vma_pages(vma) + vma->vm_pgoff) > in global_mmio_mmap()
40 unsigned long user_count = vma_pages(vma); in dma_common_mmap()
772 if (vma->vm_pgoff >= count || vma_pages(vma) > count - vma->vm_pgoff) in dma_mmap_pages()776 vma_pages(vma) << PAGE_SHIFT, vma->vm_page_prot); in dma_mmap_pages()
137 unsigned long count = vma_pages(vma); in privcmd_buf_mmap()
496 int count = vma_pages(vma); in gntalloc_mmap()
1058 int count = vma_pages(vma); in gntdev_mmap()
52 unsigned long vma_nrpages = vma_pages(vma); in memtrace_mmap()
242 if (vma->vm_pgoff + vma_pages(vma) > in dma_buf_mmap_internal()1590 if (pgoff + vma_pages(vma) < pgoff) in dma_buf_mmap()1594 if (pgoff + vma_pages(vma) > in dma_buf_mmap()
113 if (vma->vm_pgoff + vma_pages(vma) > ctrl->mem_size >> PAGE_SHIFT) in aspeed_p2a_mmap()
316 if (vma_pages(vma) > 4) in kvmppc_xive_native_mmap()320 if (vma_pages(vma) > KVMPPC_XIVE_NR_IRQS * 2) in kvmppc_xive_native_mmap()
222 unsigned long user_count = vma_pages(vma); in rockchip_drm_gem_object_mmap_iommu()
367 unsigned long num_pages = vma_pages(vma); in amdxdna_insert_pages()450 unsigned long num_pages = vma_pages(vma); in amdxdna_gem_dmabuf_mmap()
579 if (vma->vm_pgoff + vma_pages(vma) > shm->size >> PAGE_SHIFT) in tee_shm_fop_mmap()
224 if (vmg->prev->vm_pgoff + vma_pages(vmg->prev) == vmg->pgoff) in can_vma_merge_after()1459 nrpages = vma_pages(next); in vms_gather_munmap_vmas()3299 unsigned long charged = vma_pages(vma); in insert_vm_struct()
817 vea = vba + vma_pages(vma); in walk_page_mapping()
1416 len = vma_pages(vma) << PAGE_SHIFT; in mmap_action_finish()
725 if (vma->vm_pgoff + vma_pages(vma) > size) in cdx_mmap_resource()
1088 vma_pages(vma)); in i915_gem_mmap()
1095 if (vma->vm_pgoff >= count || vma_pages(vma) > count - vma->vm_pgoff) in iommu_dma_mmap_noncontiguous()1705 if (off >= nr_pages || vma_pages(vma) > nr_pages - off) in iommu_dma_mmap()
1389 vma_pages(vma)); in drm_gem_mmap()
1311 static inline unsigned long vma_pages(const struct vm_area_struct *vma) in vma_pages() function
1215 if (vma->vm_pgoff + vma_pages(vma) > pg_size) in bcm_vk_mmap()