Lines Matching refs:attrs

37 	unsigned long	attrs;
45 this->attrs);
85 * @attrs: Flags in the DMA_ATTR_* namespace.
94 gfp_t gfp, unsigned long attrs)
103 vaddr = dma_alloc_attrs(dev, size, dma_handle, gfp, attrs);
112 dr->attrs = attrs;
156 enum dma_data_direction dir, unsigned long attrs)
159 bool is_mmio = attrs & DMA_ATTR_MMIO;
167 if (!dev_is_dma_coherent(dev) && (attrs & DMA_ATTR_REQUIRE_COHERENT))
172 addr = dma_direct_map_phys(dev, phys, size, dir, attrs);
174 addr = iommu_dma_map_phys(dev, phys, size, dir, attrs);
176 addr = ops->map_phys(dev, phys, size, dir, attrs);
180 trace_dma_map_phys(dev, phys, addr, size, dir, attrs);
181 debug_dma_map_phys(dev, phys, size, dir, addr, attrs);
189 unsigned long attrs)
193 if (unlikely(attrs & DMA_ATTR_MMIO))
200 return dma_map_phys(dev, phys, size, dir, attrs);
205 enum dma_data_direction dir, unsigned long attrs)
208 bool is_mmio = attrs & DMA_ATTR_MMIO;
213 dma_direct_unmap_phys(dev, addr, size, dir, attrs);
215 iommu_dma_unmap_phys(dev, addr, size, dir, attrs);
217 ops->unmap_phys(dev, addr, size, dir, attrs);
218 trace_dma_unmap_phys(dev, addr, size, dir, attrs);
224 enum dma_data_direction dir, unsigned long attrs)
226 if (unlikely(attrs & DMA_ATTR_MMIO))
229 dma_unmap_phys(dev, addr, size, dir, attrs);
234 int nents, enum dma_data_direction dir, unsigned long attrs)
241 if (!dev_is_dma_coherent(dev) && (attrs & DMA_ATTR_REQUIRE_COHERENT))
249 ents = dma_direct_map_sg(dev, sg, nents, dir, attrs);
251 ents = iommu_dma_map_sg(dev, sg, nents, dir, attrs);
253 ents = ops->map_sg(dev, sg, nents, dir, attrs);
257 trace_dma_map_sg(dev, sg, nents, ents, dir, attrs);
258 debug_dma_map_sg(dev, sg, nents, ents, dir, attrs);
261 trace_dma_map_sg_err(dev, sg, nents, ents, dir, attrs);
274 * @attrs: Optional DMA attributes for the map operation
286 int nents, enum dma_data_direction dir, unsigned long attrs)
290 ret = __dma_map_sg_attrs(dev, sg, nents, dir, attrs);
302 * @attrs: Optional DMA attributes for the map operation
325 enum dma_data_direction dir, unsigned long attrs)
329 nents = __dma_map_sg_attrs(dev, sgt->sgl, sgt->orig_nents, dir, attrs);
339 unsigned long attrs)
344 trace_dma_unmap_sg(dev, sg, nents, dir, attrs);
348 dma_direct_unmap_sg(dev, sg, nents, dir, attrs);
350 iommu_dma_unmap_sg(dev, sg, nents, dir, attrs);
352 ops->unmap_sg(dev, sg, nents, dir, attrs);
357 size_t size, enum dma_data_direction dir, unsigned long attrs)
363 return dma_map_phys(dev, phys_addr, size, dir, attrs | DMA_ATTR_MMIO);
368 enum dma_data_direction dir, unsigned long attrs)
370 dma_unmap_phys(dev, addr, size, dir, attrs | DMA_ATTR_MMIO);
514 unsigned long attrs)
520 size, attrs);
523 size, attrs);
526 return ops->get_sgtable(dev, sgt, cpu_addr, dma_addr, size, attrs);
535 pgprot_t dma_pgprot(struct device *dev, pgprot_t prot, unsigned long attrs)
540 if (attrs & DMA_ATTR_WRITE_COMBINE)
573 * @attrs: attributes of mapping properties requested in dma_alloc_attrs
581 unsigned long attrs)
587 attrs);
590 attrs);
593 return ops->mmap(dev, vma, cpu_addr, dma_addr, size, attrs);
623 gfp_t flag, unsigned long attrs)
640 DMA_BIDIRECTIONAL, flag, attrs);
648 cpu_addr = dma_direct_alloc(dev, size, dma_handle, flag, attrs);
650 cpu_addr = iommu_dma_alloc(dev, size, dma_handle, flag, attrs);
652 cpu_addr = ops->alloc(dev, size, dma_handle, flag, attrs);
655 attrs);
660 flag, attrs);
661 debug_dma_alloc_coherent(dev, size, *dma_handle, cpu_addr, attrs);
667 dma_addr_t dma_handle, unsigned long attrs)
683 attrs);
689 dma_direct_free(dev, size, cpu_addr, dma_handle, attrs);
691 iommu_dma_free(dev, size, cpu_addr, dma_handle, attrs);
693 ops->free(dev, size, cpu_addr, dma_handle, attrs);
796 enum dma_data_direction dir, gfp_t gfp, unsigned long attrs)
800 if (WARN_ON_ONCE(attrs & ~DMA_ATTR_ALLOC_SINGLE_PAGES))
806 sgt = iommu_dma_alloc_noncontiguous(dev, size, dir, gfp, attrs);
812 trace_dma_alloc_sgt(dev, sgt, size, dir, gfp, attrs);
813 debug_dma_map_sg(dev, sgt->sgl, sgt->orig_nents, 1, dir, attrs);
815 trace_dma_alloc_sgt_err(dev, NULL, 0, size, dir, gfp, attrs);