Lines Matching refs:r
43 static inline unsigned long sgt_pfn(const struct remap_pfn *r) in sgt_pfn() argument
45 if (use_dma(r->iobase)) in sgt_pfn()
46 return (r->sgt.dma + r->sgt.curr + r->iobase) >> PAGE_SHIFT; in sgt_pfn()
48 return r->sgt.pfn + (r->sgt.curr >> PAGE_SHIFT); in sgt_pfn()
53 struct remap_pfn *r = data; in remap_sg() local
55 if (GEM_WARN_ON(!r->sgt.sgp)) in remap_sg()
59 set_pte_at(r->mm, addr, pte, in remap_sg()
60 pte_mkspecial(pfn_pte(sgt_pfn(r), r->prot))); in remap_sg()
61 r->pfn++; /* track insertions in case we need to unwind later */ in remap_sg()
63 r->sgt.curr += PAGE_SIZE; in remap_sg()
64 if (r->sgt.curr >= r->sgt.max) in remap_sg()
65 r->sgt = __sgt_iter(__sg_next(r->sgt.sgp), use_dma(r->iobase)); in remap_sg()
75 struct remap_pfn *r = data; in remap_pfn() local
78 set_pte_at(r->mm, addr, pte, pte_mkspecial(pfn_pte(r->pfn, r->prot))); in remap_pfn()
79 r->pfn++; in remap_pfn()
98 struct remap_pfn r; in remap_io_mapping() local
104 r.mm = vma->vm_mm; in remap_io_mapping()
105 r.pfn = pfn; in remap_io_mapping()
106 r.prot = __pgprot((pgprot_val(iomap->prot) & _PAGE_CACHE_MASK) | in remap_io_mapping()
109 err = apply_to_page_range(r.mm, addr, size, remap_pfn, &r); in remap_io_mapping()
111 zap_vma_ptes(vma, addr, (r.pfn - pfn) << PAGE_SHIFT); in remap_io_mapping()
135 struct remap_pfn r = { in remap_io_sg() local
146 while (offset >= r.sgt.max >> PAGE_SHIFT) { in remap_io_sg()
147 offset -= r.sgt.max >> PAGE_SHIFT; in remap_io_sg()
148 r.sgt = __sgt_iter(__sg_next(r.sgt.sgp), use_dma(iobase)); in remap_io_sg()
149 if (!r.sgt.sgp) in remap_io_sg()
152 r.sgt.curr = offset << PAGE_SHIFT; in remap_io_sg()
157 err = apply_to_page_range(r.mm, addr, size, remap_sg, &r); in remap_io_sg()
159 zap_vma_ptes(vma, addr, r.pfn << PAGE_SHIFT); in remap_io_sg()