Home
last modified time | relevance | path

Searched refs:attach (Results 1 – 25 of 448) sorted by relevance

12345678910>>...18

/linux/drivers/gpu/drm/virtio/
H A Dvirtgpu_prime.c49 virtgpu_gem_map_dma_buf(struct dma_buf_attachment *attach, in virtgpu_gem_map_dma_buf() argument
52 struct drm_gem_object *obj = attach->dmabuf->priv; in virtgpu_gem_map_dma_buf()
56 return virtio_gpu_vram_map_dma_buf(bo, attach->dev, dir); in virtgpu_gem_map_dma_buf()
58 return drm_gem_map_dma_buf(attach, dir); in virtgpu_gem_map_dma_buf()
61 static void virtgpu_gem_unmap_dma_buf(struct dma_buf_attachment *attach, in virtgpu_gem_unmap_dma_buf() argument
65 struct drm_gem_object *obj = attach->dmabuf->priv; in virtgpu_gem_unmap_dma_buf()
69 virtio_gpu_vram_unmap_dma_buf(attach->dev, sgt, dir); in virtgpu_gem_unmap_dma_buf()
73 drm_gem_unmap_dma_buf(attach, sgt, dir); in virtgpu_gem_unmap_dma_buf()
78 .attach = virtio_dma_buf_attach,
149 struct dma_buf_attachment *attach) in virtgpu_dma_buf_import_sgt() argument
[all …]
/linux/drivers/gpu/drm/xe/
H A Dxe_dma_buf.c26 struct dma_buf_attachment *attach) in xe_dma_buf_attach() argument
28 struct drm_gem_object *obj = attach->dmabuf->priv; in xe_dma_buf_attach()
30 if (attach->peer2peer && in xe_dma_buf_attach()
31 pci_p2pdma_distance(to_pci_dev(obj->dev->dev), attach->dev, false) < 0) in xe_dma_buf_attach()
32 attach->peer2peer = false; in xe_dma_buf_attach()
34 if (!attach->peer2peer && !xe_bo_can_migrate(gem_to_xe_bo(obj), XE_PL_TT)) in xe_dma_buf_attach()
42 struct dma_buf_attachment *attach) in xe_dma_buf_detach() argument
44 struct drm_gem_object *obj = attach->dmabuf->priv; in xe_dma_buf_detach()
49 static int xe_dma_buf_pin(struct dma_buf_attachment *attach) in xe_dma_buf_pin() argument
51 struct dma_buf *dmabuf = attach->dmabuf; in xe_dma_buf_pin()
[all …]
/linux/drivers/dma-buf/
H A Ddma-buf.c910 dma_buf_attachment_is_dynamic(struct dma_buf_attachment *attach) in dma_buf_attachment_is_dynamic() argument
912 return !!attach->importer_ops; in dma_buf_attachment_is_dynamic()
916 dma_buf_pin_on_map(struct dma_buf_attachment *attach) in dma_buf_pin_on_map() argument
918 return attach->dmabuf->ops->pin && in dma_buf_pin_on_map()
919 (!dma_buf_attachment_is_dynamic(attach) || in dma_buf_pin_on_map()
1014 struct dma_buf_attachment *attach; in dma_buf_dynamic_attach() local
1023 attach = kzalloc_obj(*attach); in dma_buf_dynamic_attach()
1024 if (!attach) in dma_buf_dynamic_attach()
1027 attach->dev = dev; in dma_buf_dynamic_attach()
1028 attach->dmabuf = dmabuf; in dma_buf_dynamic_attach()
[all …]
H A Ddma-buf-mapping.c92 struct sg_table *dma_buf_phys_vec_to_sgt(struct dma_buf_attachment *attach, in dma_buf_phys_vec_to_sgt() argument
105 dma_resv_assert_held(attach->dmabuf->resv); in dma_buf_phys_vec_to_sgt()
107 if (WARN_ON(!attach || !attach->dmabuf || !provider)) in dma_buf_phys_vec_to_sgt()
115 switch (pci_p2pdma_map_type(provider, attach->dev)) { in dma_buf_phys_vec_to_sgt()
128 dma_iova_try_alloc(attach->dev, dma->state, 0, size); in dma_buf_phys_vec_to_sgt()
147 ret = dma_iova_link(attach->dev, dma->state, in dma_buf_phys_vec_to_sgt()
156 addr = dma_map_phys(attach->dev, phys_vec[i].paddr, in dma_buf_phys_vec_to_sgt()
159 ret = dma_mapping_error(attach->dev, addr); in dma_buf_phys_vec_to_sgt()
170 ret = dma_iova_sync(attach->dev, dma->state, 0, mapped_len); in dma_buf_phys_vec_to_sgt()
197 dma_iova_destroy(attach->dev, dma->state, mapped_len, dir, in dma_buf_phys_vec_to_sgt()
[all …]
/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_dma_buf.c58 static struct amdgpu_device *dma_buf_attach_adev(struct dma_buf_attachment *attach) in dma_buf_attach_adev() argument
60 if (attach->importer_ops == &amdgpu_dma_buf_attach_ops) { in dma_buf_attach_adev()
61 struct drm_gem_object *obj = attach->importer_priv; in dma_buf_attach_adev()
79 struct dma_buf_attachment *attach) in amdgpu_dma_buf_attach() argument
81 struct amdgpu_device *attach_adev = dma_buf_attach_adev(attach); in amdgpu_dma_buf_attach()
97 attach->peer2peer = false; in amdgpu_dma_buf_attach()
100 pci_p2pdma_distance(adev->pdev, attach->dev, false) < 0) in amdgpu_dma_buf_attach()
101 attach->peer2peer = false; in amdgpu_dma_buf_attach()
121 static int amdgpu_dma_buf_pin(struct dma_buf_attachment *attach) in amdgpu_dma_buf_pin() argument
123 struct dma_buf *dmabuf = attach->dmabuf; in amdgpu_dma_buf_pin()
[all …]
/linux/drivers/infiniband/core/
H A Dumem_dmabuf.c24 dma_resv_assert_held(umem_dmabuf->attach->dmabuf->resv); in ib_umem_dmabuf_map_pages()
32 sgt = dma_buf_map_attachment(umem_dmabuf->attach, in ib_umem_dmabuf_map_pages()
75 ret = dma_resv_wait_timeout(umem_dmabuf->attach->dmabuf->resv, in ib_umem_dmabuf_map_pages()
88 dma_resv_assert_held(umem_dmabuf->attach->dmabuf->resv); in ib_umem_dmabuf_unmap_pages()
109 dma_buf_unmap_attachment(umem_dmabuf->attach, umem_dmabuf->sgt, in ib_umem_dmabuf_unmap_pages()
155 umem_dmabuf->attach = dma_buf_dynamic_attach( in ib_umem_dmabuf_get_with_dma_device()
160 if (IS_ERR(umem_dmabuf->attach)) { in ib_umem_dmabuf_get_with_dma_device()
161 ret = ERR_CAST(umem_dmabuf->attach); in ib_umem_dmabuf_get_with_dma_device()
185 ib_umem_dmabuf_unsupported_move_notify(struct dma_buf_attachment *attach) in ib_umem_dmabuf_unsupported_move_notify() argument
187 struct ib_umem_dmabuf *umem_dmabuf = attach->importer_priv; in ib_umem_dmabuf_unsupported_move_notify()
[all …]
/linux/drivers/media/common/videobuf2/
H A Dvideobuf2-vmalloc.c215 struct vb2_vmalloc_attachment *attach; in vb2_vmalloc_dmabuf_ops_attach() local
224 attach = kzalloc_obj(*attach); in vb2_vmalloc_dmabuf_ops_attach()
225 if (!attach) in vb2_vmalloc_dmabuf_ops_attach()
228 sgt = &attach->sgt; in vb2_vmalloc_dmabuf_ops_attach()
231 kfree(attach); in vb2_vmalloc_dmabuf_ops_attach()
239 kfree(attach); in vb2_vmalloc_dmabuf_ops_attach()
246 attach->dma_dir = DMA_NONE; in vb2_vmalloc_dmabuf_ops_attach()
247 dbuf_attach->priv = attach; in vb2_vmalloc_dmabuf_ops_attach()
254 struct vb2_vmalloc_attachment *attach = db_attach->priv; in vb2_vmalloc_dmabuf_ops_detach() local
257 if (!attach) in vb2_vmalloc_dmabuf_ops_detach()
[all …]
H A Dvideobuf2-dma-sg.c371 struct vb2_dma_sg_attachment *attach; in vb2_dma_sg_dmabuf_ops_attach() local
378 attach = kzalloc_obj(*attach); in vb2_dma_sg_dmabuf_ops_attach()
379 if (!attach) in vb2_dma_sg_dmabuf_ops_attach()
382 sgt = &attach->sgt; in vb2_dma_sg_dmabuf_ops_attach()
388 kfree(attach); in vb2_dma_sg_dmabuf_ops_attach()
400 attach->dma_dir = DMA_NONE; in vb2_dma_sg_dmabuf_ops_attach()
401 dbuf_attach->priv = attach; in vb2_dma_sg_dmabuf_ops_attach()
409 struct vb2_dma_sg_attachment *attach = db_attach->priv; in vb2_dma_sg_dmabuf_ops_detach() local
412 if (!attach) in vb2_dma_sg_dmabuf_ops_detach()
415 sgt = &attach->sgt; in vb2_dma_sg_dmabuf_ops_detach()
[all …]
H A Dvideobuf2-dma-contig.c321 struct vb2_dc_attachment *attach; in vb2_dc_dmabuf_ops_attach() local
328 attach = kzalloc_obj(*attach); in vb2_dc_dmabuf_ops_attach()
329 if (!attach) in vb2_dc_dmabuf_ops_attach()
332 sgt = &attach->sgt; in vb2_dc_dmabuf_ops_attach()
338 kfree(attach); in vb2_dc_dmabuf_ops_attach()
350 attach->dma_dir = DMA_NONE; in vb2_dc_dmabuf_ops_attach()
351 dbuf_attach->priv = attach; in vb2_dc_dmabuf_ops_attach()
359 struct vb2_dc_attachment *attach = db_attach->priv; in vb2_dc_dmabuf_ops_detach() local
362 if (!attach) in vb2_dc_dmabuf_ops_detach()
365 sgt = &attach->sgt; in vb2_dc_dmabuf_ops_detach()
[all …]
/linux/drivers/xen/
H A Dgntdev-dmabuf.c48 struct dma_buf_attachment *attach; member
221 struct dma_buf_attachment *attach) in dmabuf_exp_ops_attach() argument
230 attach->priv = gntdev_dmabuf_attach; in dmabuf_exp_ops_attach()
235 struct dma_buf_attachment *attach) in dmabuf_exp_ops_detach() argument
237 struct gntdev_dmabuf_attachment *gntdev_dmabuf_attach = attach->priv; in dmabuf_exp_ops_detach()
244 dma_unmap_sgtable(attach->dev, sgt, in dmabuf_exp_ops_detach()
252 attach->priv = NULL; in dmabuf_exp_ops_detach()
257 dmabuf_exp_ops_map_dma_buf(struct dma_buf_attachment *attach, in dmabuf_exp_ops_map_dma_buf() argument
260 struct gntdev_dmabuf_attachment *gntdev_dmabuf_attach = attach->priv; in dmabuf_exp_ops_map_dma_buf()
261 struct gntdev_dmabuf *gntdev_dmabuf = attach->dmabuf->priv; in dmabuf_exp_ops_map_dma_buf()
[all …]
/linux/drivers/iommu/iommufd/
H A Ddevice.c355 struct iommufd_attach *attach; in iommufd_group_device_num() local
362 attach = xa_load(&igroup->pasid_attach, pasid); in iommufd_group_device_num()
363 if (attach) in iommufd_group_device_num()
364 xa_for_each(&attach->device_array, index, idev) in iommufd_group_device_num()
443 struct iommufd_attach *attach; in iommufd_device_is_attached() local
445 attach = xa_load(&idev->igroup->pasid_attach, pasid); in iommufd_device_is_attached()
446 return xa_load(&attach->device_array, idev->obj.id); in iommufd_device_is_attached()
465 struct iommufd_attach *attach; in iommufd_hwpt_pasid_compat() local
470 attach = xa_load(&igroup->pasid_attach, IOMMU_NO_PASID); in iommufd_hwpt_pasid_compat()
471 if (attach && attach->hwpt && !attach->hwpt->pasid_compat) in iommufd_hwpt_pasid_compat()
[all …]
/linux/drivers/gpu/drm/omapdrm/
H A Domap_gem_dmabuf.c100 struct dma_buf_attachment *attach; in omap_gem_prime_import() local
117 attach = dma_buf_attach(dma_buf, dev->dev); in omap_gem_prime_import()
118 if (IS_ERR(attach)) in omap_gem_prime_import()
119 return ERR_CAST(attach); in omap_gem_prime_import()
123 sgt = dma_buf_map_attachment_unlocked(attach, DMA_TO_DEVICE); in omap_gem_prime_import()
135 obj->import_attach = attach; in omap_gem_prime_import()
140 dma_buf_unmap_attachment_unlocked(attach, sgt, DMA_TO_DEVICE); in omap_gem_prime_import()
142 dma_buf_detach(dma_buf, attach); in omap_gem_prime_import()
/linux/drivers/vfio/
H A Ddevice_cdev.c197 struct vfio_device_attach_iommufd_pt attach; in vfio_df_ioctl_attach_pt() local
204 if (copy_from_user(&attach, arg, minsz)) in vfio_df_ioctl_attach_pt()
207 if (attach.argsz < minsz) in vfio_df_ioctl_attach_pt()
210 if (attach.flags & ~VFIO_DEVICE_ATTACH_PASID) in vfio_df_ioctl_attach_pt()
213 if (attach.flags & VFIO_DEVICE_ATTACH_PASID) { in vfio_df_ioctl_attach_pt()
220 if (attach.argsz < xend) in vfio_df_ioctl_attach_pt()
223 if (copy_from_user((void *)&attach + minsz, in vfio_df_ioctl_attach_pt()
229 if (attach.flags & VFIO_DEVICE_ATTACH_PASID) in vfio_df_ioctl_attach_pt()
231 attach.pasid, in vfio_df_ioctl_attach_pt()
232 &attach.pt_id); in vfio_df_ioctl_attach_pt()
[all …]
/linux/drivers/gpu/drm/
H A Ddrm_prime.c602 struct dma_buf_attachment *attach) in drm_gem_map_attach() argument
638 struct dma_buf_attachment *attach) in drm_gem_map_detach() argument
666 struct sg_table *drm_gem_map_dma_buf(struct dma_buf_attachment *attach, in drm_gem_map_dma_buf() argument
669 struct drm_gem_object *obj = attach->dmabuf->priv; in drm_gem_map_dma_buf()
683 ret = dma_map_sgtable(attach->dev, sgt, dir, in drm_gem_map_dma_buf()
703 void drm_gem_unmap_dma_buf(struct dma_buf_attachment *attach, in drm_gem_unmap_dma_buf() argument
710 dma_unmap_sgtable(attach->dev, sgt, dir, DMA_ATTR_SKIP_CPU_SYNC); in drm_gem_unmap_dma_buf()
828 .attach = drm_gem_map_attach,
972 struct dma_buf_attachment *attach; in drm_gem_prime_import_dev() local
990 attach = dma_buf_attach(dma_buf, attach_dev); in drm_gem_prime_import_dev()
[all …]
/linux/drivers/gpu/drm/i915/gem/
H A Di915_gem_dmabuf.c27 static struct sg_table *i915_gem_map_dma_buf(struct dma_buf_attachment *attach, in i915_gem_map_dma_buf() argument
30 struct drm_i915_gem_object *obj = dma_buf_to_obj(attach->dmabuf); in i915_gem_map_dma_buf()
55 ret = dma_map_sgtable(attach->dev, sgt, dir, DMA_ATTR_SKIP_CPU_SYNC); in i915_gem_map_dma_buf()
167 struct dma_buf_attachment *attach) in i915_gem_dmabuf_attach() argument
196 struct dma_buf_attachment *attach) in i915_gem_dmabuf_detach() argument
204 .attach = i915_gem_dmabuf_attach,
285 struct dma_buf_attachment *attach; in i915_gem_prime_import() local
307 attach = dma_buf_attach(dma_buf, dev->dev); in i915_gem_prime_import()
308 if (IS_ERR(attach)) in i915_gem_prime_import()
309 return ERR_CAST(attach); in i915_gem_prime_import()
[all …]
/linux/drivers/gpu/drm/tegra/
H A Dgem.c82 map->attach = dma_buf_attach(buf, dev); in tegra_bo_pin()
83 if (IS_ERR(map->attach)) { in tegra_bo_pin()
84 err = PTR_ERR(map->attach); in tegra_bo_pin()
88 map->sgt = dma_buf_map_attachment_unlocked(map->attach, direction); in tegra_bo_pin()
90 dma_buf_detach(buf, map->attach); in tegra_bo_pin()
163 if (map->attach) { in tegra_bo_unpin()
164 dma_buf_unmap_attachment_unlocked(map->attach, map->sgt, in tegra_bo_unpin()
166 dma_buf_detach(map->attach->dmabuf, map->attach); in tegra_bo_unpin()
460 struct dma_buf_attachment *attach; in tegra_bo_import() local
473 attach = dma_buf_attach(buf, drm->dev); in tegra_bo_import()
[all …]
/linux/tools/bpf/bpftool/Documentation/
H A Dbpftool-cgroup.rst22 { **show** | **list** | **tree** | **attach** | **detach** | **help** }
29 | **bpftool** **cgroup attach** *CGROUP* *ATTACH_TYPE* *PROG* [*ATTACH_FLAGS*]
55 Output will start with program ID followed by attach type, attach flags and
67 with absolute cgroup path, followed by program ID, attach type, attach
74 bpftool cgroup attach *CGROUP* *ATTACH_TYPE* *PROG* [*ATTACH_FLAGS*]
75 Attach program *PROG* to the cgroup *CGROUP* with attach type *ATTACH_TYPE*
83 Only one program is allowed to be attached to a cgroup with no attach flags
85 program and attach the new one.
125 Detach *PROG* from the cgroup *CGROUP* and attach type *ATTACH_TYPE*.
143 | **# bpftool cgroup attach /sys/fs/cgroup/test.slice/ device id 1 allow_multi**
H A Dbpftool-net.rst21 *COMMANDS* := { **show** | **list** | **attach** | **detach** | **help** }
27 | **bpftool** **net attach** *ATTACH_TYPE* *PROG* **dev** *NAME* [ **overwrite** | **prepend** ]
61 bpftool net attach *ATTACH_TYPE* *PROG* dev *NAME* [ overwrite | prepend ]
73 For XDP-related attach types (**xdp**, **xdpgeneric**, **xdpdrv**,
77 For **tcx_ingress** and **tcx_egress** attach types, the **prepend** option
78 can be used to attach the program at the beginning of the chain instead of
84 for attach must be specified.
162 | **# bpftool net attach xdpdrv id 16 dev enp6s0np0**
171 | **# bpftool net attach xdpdrv id 16 dev enp6s0np0**
172 | **# bpftool net attach xdpdr
[all...]
/linux/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_prime.c45 struct dma_buf_attachment *attach) in vmw_prime_map_attach() argument
51 struct dma_buf_attachment *attach) in vmw_prime_map_detach() argument
55 static struct sg_table *vmw_prime_map_dma_buf(struct dma_buf_attachment *attach, in vmw_prime_map_dma_buf() argument
61 static void vmw_prime_unmap_dma_buf(struct dma_buf_attachment *attach, in vmw_prime_unmap_dma_buf() argument
68 .attach = vmw_prime_map_attach,
/linux/drivers/iio/
H A Dindustrialio-buffer.c53 struct dma_buf_attachment *attach; member
1562 struct dma_buf_attachment *attach = priv->attach; in iio_buffer_dmabuf_release() local
1564 struct dma_buf *dmabuf = attach->dmabuf; in iio_buffer_dmabuf_release()
1566 dma_buf_unmap_attachment_unlocked(attach, priv->sgt, priv->dir); in iio_buffer_dmabuf_release()
1570 dma_buf_detach(attach->dmabuf, attach); in iio_buffer_dmabuf_release()
1575 static void iio_buffer_dmabuf_get(struct dma_buf_attachment *attach) in iio_buffer_dmabuf_get() argument
1577 struct iio_dmabuf_priv *priv = attach->importer_priv; in iio_buffer_dmabuf_get()
1582 static void iio_buffer_dmabuf_put(struct dma_buf_attachment *attach) in iio_buffer_dmabuf_put() argument
1584 struct iio_dmabuf_priv *priv = attach->importer_priv; in iio_buffer_dmabuf_put()
1603 iio_buffer_dmabuf_put(priv->attach); in iio_buffer_chrdev_release()
[all …]
/linux/drivers/gpu/drm/panfrost/
H A Dpanfrost_gem.c281 panfrost_gem_prime_map_dma_buf(struct dma_buf_attachment *attach, in panfrost_gem_prime_map_dma_buf() argument
284 struct sg_table *sgt = drm_gem_map_dma_buf(attach, dir); in panfrost_gem_prime_map_dma_buf()
287 attach->priv = sgt; in panfrost_gem_prime_map_dma_buf()
293 panfrost_gem_prime_unmap_dma_buf(struct dma_buf_attachment *attach, in panfrost_gem_prime_unmap_dma_buf() argument
297 attach->priv = NULL; in panfrost_gem_prime_unmap_dma_buf()
298 drm_gem_unmap_dma_buf(attach, sgt, dir); in panfrost_gem_prime_unmap_dma_buf()
308 struct dma_buf_attachment *attach; in panfrost_gem_prime_begin_cpu_access() local
317 list_for_each_entry(attach, &dma_buf->attachments, node) { in panfrost_gem_prime_begin_cpu_access()
318 struct sg_table *sgt = attach->priv; in panfrost_gem_prime_begin_cpu_access()
321 dma_sync_sgtable_for_cpu(attach->dev, sgt, dir); in panfrost_gem_prime_begin_cpu_access()
[all …]
/linux/Documentation/bpf/
H A Dmap_cgroup_storage.rst10 attach to cgroups; the programs are made available by the same Kconfig. The
34 ``attach_type`` is the program's attach type.
37 When this key type is used, then all attach types of the particular cgroup and
39 ``struct bpf_cgroup_storage_key``, then programs of different attach types
133 multiple attach types, and each attach creates a fresh zeroed storage. The
145 does not already contain an entry for the cgroup and attach type pair, or else
146 the old storage is reused for the new attachment. If the map is attach type
147 shared, then attach type is simply ignored during comparison. Storage is freed
158 In all versions, userspace may use the attach parameters of cgroup and
159 attach type pair in ``struct bpf_cgroup_storage_key`` as the key to the BPF map
[all …]
/linux/drivers/gpu/drm/armada/
H A Darmada_gem.c387 armada_gem_prime_map_dma_buf(struct dma_buf_attachment *attach, in armada_gem_prime_map_dma_buf() argument
390 struct drm_gem_object *obj = attach->dmabuf->priv; in armada_gem_prime_map_dma_buf()
420 if (dma_map_sgtable(attach->dev, sgt, dir, 0)) in armada_gem_prime_map_dma_buf()
429 if (dma_map_sgtable(attach->dev, sgt, dir, 0)) in armada_gem_prime_map_dma_buf()
453 static void armada_gem_prime_unmap_dma_buf(struct dma_buf_attachment *attach, in armada_gem_prime_unmap_dma_buf() argument
456 struct drm_gem_object *obj = attach->dmabuf->priv; in armada_gem_prime_unmap_dma_buf()
461 dma_unmap_sgtable(attach->dev, sgt, dir, 0); in armada_gem_prime_unmap_dma_buf()
503 struct dma_buf_attachment *attach; in armada_gem_prime_import() local
518 attach = dma_buf_attach(buf, dev->dev); in armada_gem_prime_import()
519 if (IS_ERR(attach)) in armada_gem_prime_import()
[all …]
/linux/drivers/virtio/
H A Dvirtio_dma_buf.c28 exp_info->ops->attach != &virtio_dma_buf_attach || in virtio_dma_buf_export()
43 struct dma_buf_attachment *attach) in virtio_dma_buf_attach() argument
51 ret = ops->device_attach(dma_buf, attach); in virtio_dma_buf_attach()
65 return dma_buf->ops->attach == &virtio_dma_buf_attach; in is_virtio_dma_buf()
/linux/drivers/gpu/drm/panthor/
H A Dpanthor_gem.c240 panthor_gem_prime_map_dma_buf(struct dma_buf_attachment *attach, in panthor_gem_prime_map_dma_buf() argument
243 struct sg_table *sgt = drm_gem_map_dma_buf(attach, dir); in panthor_gem_prime_map_dma_buf()
246 attach->priv = sgt; in panthor_gem_prime_map_dma_buf()
252 panthor_gem_prime_unmap_dma_buf(struct dma_buf_attachment *attach, in panthor_gem_prime_unmap_dma_buf() argument
256 attach->priv = NULL; in panthor_gem_prime_unmap_dma_buf()
257 drm_gem_unmap_dma_buf(attach, sgt, dir); in panthor_gem_prime_unmap_dma_buf()
267 struct dma_buf_attachment *attach; in panthor_gem_prime_begin_cpu_access() local
276 list_for_each_entry(attach, &dma_buf->attachments, node) { in panthor_gem_prime_begin_cpu_access()
277 struct sg_table *sgt = attach->priv; in panthor_gem_prime_begin_cpu_access()
280 dma_sync_sgtable_for_cpu(attach->dev, sgt, dir); in panthor_gem_prime_begin_cpu_access()
[all …]

12345678910>>...18