| /linux/drivers/gpu/drm/virtio/ |
| H A D | virtgpu_prime.c | 49 virtgpu_gem_map_dma_buf(struct dma_buf_attachment *attach, in virtgpu_gem_map_dma_buf() argument 52 struct drm_gem_object *obj = attach->dmabuf->priv; in virtgpu_gem_map_dma_buf() 56 return virtio_gpu_vram_map_dma_buf(bo, attach->dev, dir); in virtgpu_gem_map_dma_buf() 58 return drm_gem_map_dma_buf(attach, dir); in virtgpu_gem_map_dma_buf() 61 static void virtgpu_gem_unmap_dma_buf(struct dma_buf_attachment *attach, in virtgpu_gem_unmap_dma_buf() argument 65 struct drm_gem_object *obj = attach->dmabuf->priv; in virtgpu_gem_unmap_dma_buf() 69 virtio_gpu_vram_unmap_dma_buf(attach->dev, sgt, dir); in virtgpu_gem_unmap_dma_buf() 73 drm_gem_unmap_dma_buf(attach, sgt, dir); in virtgpu_gem_unmap_dma_buf() 78 .attach = virtio_dma_buf_attach, 149 struct dma_buf_attachment *attach) in virtgpu_dma_buf_import_sgt() argument [all …]
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_dma_buf.c | 26 struct dma_buf_attachment *attach) in xe_dma_buf_attach() argument 28 struct drm_gem_object *obj = attach->dmabuf->priv; in xe_dma_buf_attach() 30 if (attach->peer2peer && in xe_dma_buf_attach() 31 pci_p2pdma_distance(to_pci_dev(obj->dev->dev), attach->dev, false) < 0) in xe_dma_buf_attach() 32 attach->peer2peer = false; in xe_dma_buf_attach() 34 if (!attach->peer2peer && !xe_bo_can_migrate(gem_to_xe_bo(obj), XE_PL_TT)) in xe_dma_buf_attach() 42 struct dma_buf_attachment *attach) in xe_dma_buf_detach() argument 44 struct drm_gem_object *obj = attach->dmabuf->priv; in xe_dma_buf_detach() 49 static int xe_dma_buf_pin(struct dma_buf_attachment *attach) in xe_dma_buf_pin() argument 51 struct dma_buf *dmabuf = attach->dmabuf; in xe_dma_buf_pin() [all …]
|
| /linux/drivers/dma-buf/ |
| H A D | dma-buf.c | 910 dma_buf_attachment_is_dynamic(struct dma_buf_attachment *attach) in dma_buf_attachment_is_dynamic() argument 912 return !!attach->importer_ops; in dma_buf_attachment_is_dynamic() 916 dma_buf_pin_on_map(struct dma_buf_attachment *attach) in dma_buf_pin_on_map() argument 918 return attach->dmabuf->ops->pin && in dma_buf_pin_on_map() 919 (!dma_buf_attachment_is_dynamic(attach) || in dma_buf_pin_on_map() 1014 struct dma_buf_attachment *attach; in dma_buf_dynamic_attach() local 1023 attach = kzalloc_obj(*attach); in dma_buf_dynamic_attach() 1024 if (!attach) in dma_buf_dynamic_attach() 1027 attach->dev = dev; in dma_buf_dynamic_attach() 1028 attach->dmabuf = dmabuf; in dma_buf_dynamic_attach() [all …]
|
| H A D | dma-buf-mapping.c | 92 struct sg_table *dma_buf_phys_vec_to_sgt(struct dma_buf_attachment *attach, in dma_buf_phys_vec_to_sgt() argument 105 dma_resv_assert_held(attach->dmabuf->resv); in dma_buf_phys_vec_to_sgt() 107 if (WARN_ON(!attach || !attach->dmabuf || !provider)) in dma_buf_phys_vec_to_sgt() 115 switch (pci_p2pdma_map_type(provider, attach->dev)) { in dma_buf_phys_vec_to_sgt() 128 dma_iova_try_alloc(attach->dev, dma->state, 0, size); in dma_buf_phys_vec_to_sgt() 147 ret = dma_iova_link(attach->dev, dma->state, in dma_buf_phys_vec_to_sgt() 156 addr = dma_map_phys(attach->dev, phys_vec[i].paddr, in dma_buf_phys_vec_to_sgt() 159 ret = dma_mapping_error(attach->dev, addr); in dma_buf_phys_vec_to_sgt() 170 ret = dma_iova_sync(attach->dev, dma->state, 0, mapped_len); in dma_buf_phys_vec_to_sgt() 197 dma_iova_destroy(attach->dev, dma->state, mapped_len, dir, in dma_buf_phys_vec_to_sgt() [all …]
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_dma_buf.c | 58 static struct amdgpu_device *dma_buf_attach_adev(struct dma_buf_attachment *attach) in dma_buf_attach_adev() argument 60 if (attach->importer_ops == &amdgpu_dma_buf_attach_ops) { in dma_buf_attach_adev() 61 struct drm_gem_object *obj = attach->importer_priv; in dma_buf_attach_adev() 79 struct dma_buf_attachment *attach) in amdgpu_dma_buf_attach() argument 81 struct amdgpu_device *attach_adev = dma_buf_attach_adev(attach); in amdgpu_dma_buf_attach() 97 attach->peer2peer = false; in amdgpu_dma_buf_attach() 100 pci_p2pdma_distance(adev->pdev, attach->dev, false) < 0) in amdgpu_dma_buf_attach() 101 attach->peer2peer = false; in amdgpu_dma_buf_attach() 121 static int amdgpu_dma_buf_pin(struct dma_buf_attachment *attach) in amdgpu_dma_buf_pin() argument 123 struct dma_buf *dmabuf = attach->dmabuf; in amdgpu_dma_buf_pin() [all …]
|
| /linux/drivers/infiniband/core/ |
| H A D | umem_dmabuf.c | 24 dma_resv_assert_held(umem_dmabuf->attach->dmabuf->resv); in ib_umem_dmabuf_map_pages() 32 sgt = dma_buf_map_attachment(umem_dmabuf->attach, in ib_umem_dmabuf_map_pages() 75 ret = dma_resv_wait_timeout(umem_dmabuf->attach->dmabuf->resv, in ib_umem_dmabuf_map_pages() 88 dma_resv_assert_held(umem_dmabuf->attach->dmabuf->resv); in ib_umem_dmabuf_unmap_pages() 109 dma_buf_unmap_attachment(umem_dmabuf->attach, umem_dmabuf->sgt, in ib_umem_dmabuf_unmap_pages() 155 umem_dmabuf->attach = dma_buf_dynamic_attach( in ib_umem_dmabuf_get_with_dma_device() 160 if (IS_ERR(umem_dmabuf->attach)) { in ib_umem_dmabuf_get_with_dma_device() 161 ret = ERR_CAST(umem_dmabuf->attach); in ib_umem_dmabuf_get_with_dma_device() 185 ib_umem_dmabuf_unsupported_move_notify(struct dma_buf_attachment *attach) in ib_umem_dmabuf_unsupported_move_notify() argument 187 struct ib_umem_dmabuf *umem_dmabuf = attach->importer_priv; in ib_umem_dmabuf_unsupported_move_notify() [all …]
|
| /linux/drivers/media/common/videobuf2/ |
| H A D | videobuf2-vmalloc.c | 215 struct vb2_vmalloc_attachment *attach; in vb2_vmalloc_dmabuf_ops_attach() local 224 attach = kzalloc_obj(*attach); in vb2_vmalloc_dmabuf_ops_attach() 225 if (!attach) in vb2_vmalloc_dmabuf_ops_attach() 228 sgt = &attach->sgt; in vb2_vmalloc_dmabuf_ops_attach() 231 kfree(attach); in vb2_vmalloc_dmabuf_ops_attach() 239 kfree(attach); in vb2_vmalloc_dmabuf_ops_attach() 246 attach->dma_dir = DMA_NONE; in vb2_vmalloc_dmabuf_ops_attach() 247 dbuf_attach->priv = attach; in vb2_vmalloc_dmabuf_ops_attach() 254 struct vb2_vmalloc_attachment *attach = db_attach->priv; in vb2_vmalloc_dmabuf_ops_detach() local 257 if (!attach) in vb2_vmalloc_dmabuf_ops_detach() [all …]
|
| H A D | videobuf2-dma-sg.c | 371 struct vb2_dma_sg_attachment *attach; in vb2_dma_sg_dmabuf_ops_attach() local 378 attach = kzalloc_obj(*attach); in vb2_dma_sg_dmabuf_ops_attach() 379 if (!attach) in vb2_dma_sg_dmabuf_ops_attach() 382 sgt = &attach->sgt; in vb2_dma_sg_dmabuf_ops_attach() 388 kfree(attach); in vb2_dma_sg_dmabuf_ops_attach() 400 attach->dma_dir = DMA_NONE; in vb2_dma_sg_dmabuf_ops_attach() 401 dbuf_attach->priv = attach; in vb2_dma_sg_dmabuf_ops_attach() 409 struct vb2_dma_sg_attachment *attach = db_attach->priv; in vb2_dma_sg_dmabuf_ops_detach() local 412 if (!attach) in vb2_dma_sg_dmabuf_ops_detach() 415 sgt = &attach->sgt; in vb2_dma_sg_dmabuf_ops_detach() [all …]
|
| H A D | videobuf2-dma-contig.c | 321 struct vb2_dc_attachment *attach; in vb2_dc_dmabuf_ops_attach() local 328 attach = kzalloc_obj(*attach); in vb2_dc_dmabuf_ops_attach() 329 if (!attach) in vb2_dc_dmabuf_ops_attach() 332 sgt = &attach->sgt; in vb2_dc_dmabuf_ops_attach() 338 kfree(attach); in vb2_dc_dmabuf_ops_attach() 350 attach->dma_dir = DMA_NONE; in vb2_dc_dmabuf_ops_attach() 351 dbuf_attach->priv = attach; in vb2_dc_dmabuf_ops_attach() 359 struct vb2_dc_attachment *attach = db_attach->priv; in vb2_dc_dmabuf_ops_detach() local 362 if (!attach) in vb2_dc_dmabuf_ops_detach() 365 sgt = &attach->sgt; in vb2_dc_dmabuf_ops_detach() [all …]
|
| /linux/drivers/xen/ |
| H A D | gntdev-dmabuf.c | 48 struct dma_buf_attachment *attach; member 221 struct dma_buf_attachment *attach) in dmabuf_exp_ops_attach() argument 230 attach->priv = gntdev_dmabuf_attach; in dmabuf_exp_ops_attach() 235 struct dma_buf_attachment *attach) in dmabuf_exp_ops_detach() argument 237 struct gntdev_dmabuf_attachment *gntdev_dmabuf_attach = attach->priv; in dmabuf_exp_ops_detach() 244 dma_unmap_sgtable(attach->dev, sgt, in dmabuf_exp_ops_detach() 252 attach->priv = NULL; in dmabuf_exp_ops_detach() 257 dmabuf_exp_ops_map_dma_buf(struct dma_buf_attachment *attach, in dmabuf_exp_ops_map_dma_buf() argument 260 struct gntdev_dmabuf_attachment *gntdev_dmabuf_attach = attach->priv; in dmabuf_exp_ops_map_dma_buf() 261 struct gntdev_dmabuf *gntdev_dmabuf = attach->dmabuf->priv; in dmabuf_exp_ops_map_dma_buf() [all …]
|
| /linux/drivers/iommu/iommufd/ |
| H A D | device.c | 355 struct iommufd_attach *attach; in iommufd_group_device_num() local 362 attach = xa_load(&igroup->pasid_attach, pasid); in iommufd_group_device_num() 363 if (attach) in iommufd_group_device_num() 364 xa_for_each(&attach->device_array, index, idev) in iommufd_group_device_num() 443 struct iommufd_attach *attach; in iommufd_device_is_attached() local 445 attach = xa_load(&idev->igroup->pasid_attach, pasid); in iommufd_device_is_attached() 446 return xa_load(&attach->device_array, idev->obj.id); in iommufd_device_is_attached() 465 struct iommufd_attach *attach; in iommufd_hwpt_pasid_compat() local 470 attach = xa_load(&igroup->pasid_attach, IOMMU_NO_PASID); in iommufd_hwpt_pasid_compat() 471 if (attach && attach->hwpt && !attach->hwpt->pasid_compat) in iommufd_hwpt_pasid_compat() [all …]
|
| /linux/drivers/gpu/drm/omapdrm/ |
| H A D | omap_gem_dmabuf.c | 100 struct dma_buf_attachment *attach; in omap_gem_prime_import() local 117 attach = dma_buf_attach(dma_buf, dev->dev); in omap_gem_prime_import() 118 if (IS_ERR(attach)) in omap_gem_prime_import() 119 return ERR_CAST(attach); in omap_gem_prime_import() 123 sgt = dma_buf_map_attachment_unlocked(attach, DMA_TO_DEVICE); in omap_gem_prime_import() 135 obj->import_attach = attach; in omap_gem_prime_import() 140 dma_buf_unmap_attachment_unlocked(attach, sgt, DMA_TO_DEVICE); in omap_gem_prime_import() 142 dma_buf_detach(dma_buf, attach); in omap_gem_prime_import()
|
| /linux/drivers/vfio/ |
| H A D | device_cdev.c | 197 struct vfio_device_attach_iommufd_pt attach; in vfio_df_ioctl_attach_pt() local 204 if (copy_from_user(&attach, arg, minsz)) in vfio_df_ioctl_attach_pt() 207 if (attach.argsz < minsz) in vfio_df_ioctl_attach_pt() 210 if (attach.flags & ~VFIO_DEVICE_ATTACH_PASID) in vfio_df_ioctl_attach_pt() 213 if (attach.flags & VFIO_DEVICE_ATTACH_PASID) { in vfio_df_ioctl_attach_pt() 220 if (attach.argsz < xend) in vfio_df_ioctl_attach_pt() 223 if (copy_from_user((void *)&attach + minsz, in vfio_df_ioctl_attach_pt() 229 if (attach.flags & VFIO_DEVICE_ATTACH_PASID) in vfio_df_ioctl_attach_pt() 231 attach.pasid, in vfio_df_ioctl_attach_pt() 232 &attach.pt_id); in vfio_df_ioctl_attach_pt() [all …]
|
| /linux/drivers/gpu/drm/ |
| H A D | drm_prime.c | 602 struct dma_buf_attachment *attach) in drm_gem_map_attach() argument 638 struct dma_buf_attachment *attach) in drm_gem_map_detach() argument 666 struct sg_table *drm_gem_map_dma_buf(struct dma_buf_attachment *attach, in drm_gem_map_dma_buf() argument 669 struct drm_gem_object *obj = attach->dmabuf->priv; in drm_gem_map_dma_buf() 683 ret = dma_map_sgtable(attach->dev, sgt, dir, in drm_gem_map_dma_buf() 703 void drm_gem_unmap_dma_buf(struct dma_buf_attachment *attach, in drm_gem_unmap_dma_buf() argument 710 dma_unmap_sgtable(attach->dev, sgt, dir, DMA_ATTR_SKIP_CPU_SYNC); in drm_gem_unmap_dma_buf() 828 .attach = drm_gem_map_attach, 972 struct dma_buf_attachment *attach; in drm_gem_prime_import_dev() local 990 attach = dma_buf_attach(dma_buf, attach_dev); in drm_gem_prime_import_dev() [all …]
|
| /linux/drivers/gpu/drm/i915/gem/ |
| H A D | i915_gem_dmabuf.c | 27 static struct sg_table *i915_gem_map_dma_buf(struct dma_buf_attachment *attach, in i915_gem_map_dma_buf() argument 30 struct drm_i915_gem_object *obj = dma_buf_to_obj(attach->dmabuf); in i915_gem_map_dma_buf() 55 ret = dma_map_sgtable(attach->dev, sgt, dir, DMA_ATTR_SKIP_CPU_SYNC); in i915_gem_map_dma_buf() 167 struct dma_buf_attachment *attach) in i915_gem_dmabuf_attach() argument 196 struct dma_buf_attachment *attach) in i915_gem_dmabuf_detach() argument 204 .attach = i915_gem_dmabuf_attach, 285 struct dma_buf_attachment *attach; in i915_gem_prime_import() local 307 attach = dma_buf_attach(dma_buf, dev->dev); in i915_gem_prime_import() 308 if (IS_ERR(attach)) in i915_gem_prime_import() 309 return ERR_CAST(attach); in i915_gem_prime_import() [all …]
|
| /linux/drivers/gpu/drm/tegra/ |
| H A D | gem.c | 82 map->attach = dma_buf_attach(buf, dev); in tegra_bo_pin() 83 if (IS_ERR(map->attach)) { in tegra_bo_pin() 84 err = PTR_ERR(map->attach); in tegra_bo_pin() 88 map->sgt = dma_buf_map_attachment_unlocked(map->attach, direction); in tegra_bo_pin() 90 dma_buf_detach(buf, map->attach); in tegra_bo_pin() 163 if (map->attach) { in tegra_bo_unpin() 164 dma_buf_unmap_attachment_unlocked(map->attach, map->sgt, in tegra_bo_unpin() 166 dma_buf_detach(map->attach->dmabuf, map->attach); in tegra_bo_unpin() 460 struct dma_buf_attachment *attach; in tegra_bo_import() local 473 attach = dma_buf_attach(buf, drm->dev); in tegra_bo_import() [all …]
|
| /linux/tools/bpf/bpftool/Documentation/ |
| H A D | bpftool-cgroup.rst | 22 { **show** | **list** | **tree** | **attach** | **detach** | **help** } 29 | **bpftool** **cgroup attach** *CGROUP* *ATTACH_TYPE* *PROG* [*ATTACH_FLAGS*] 55 Output will start with program ID followed by attach type, attach flags and 67 with absolute cgroup path, followed by program ID, attach type, attach 74 bpftool cgroup attach *CGROUP* *ATTACH_TYPE* *PROG* [*ATTACH_FLAGS*] 75 Attach program *PROG* to the cgroup *CGROUP* with attach type *ATTACH_TYPE* 83 Only one program is allowed to be attached to a cgroup with no attach flags 85 program and attach the new one. 125 Detach *PROG* from the cgroup *CGROUP* and attach type *ATTACH_TYPE*. 143 | **# bpftool cgroup attach /sys/fs/cgroup/test.slice/ device id 1 allow_multi**
|
| H A D | bpftool-net.rst | 21 *COMMANDS* := { **show** | **list** | **attach** | **detach** | **help** } 27 | **bpftool** **net attach** *ATTACH_TYPE* *PROG* **dev** *NAME* [ **overwrite** | **prepend** ] 61 bpftool net attach *ATTACH_TYPE* *PROG* dev *NAME* [ overwrite | prepend ] 73 For XDP-related attach types (**xdp**, **xdpgeneric**, **xdpdrv**, 77 For **tcx_ingress** and **tcx_egress** attach types, the **prepend** option 78 can be used to attach the program at the beginning of the chain instead of 84 for attach must be specified. 162 | **# bpftool net attach xdpdrv id 16 dev enp6s0np0** 171 | **# bpftool net attach xdpdrv id 16 dev enp6s0np0** 172 | **# bpftool net attach xdpdr [all...] |
| /linux/drivers/gpu/drm/vmwgfx/ |
| H A D | vmwgfx_prime.c | 45 struct dma_buf_attachment *attach) in vmw_prime_map_attach() argument 51 struct dma_buf_attachment *attach) in vmw_prime_map_detach() argument 55 static struct sg_table *vmw_prime_map_dma_buf(struct dma_buf_attachment *attach, in vmw_prime_map_dma_buf() argument 61 static void vmw_prime_unmap_dma_buf(struct dma_buf_attachment *attach, in vmw_prime_unmap_dma_buf() argument 68 .attach = vmw_prime_map_attach,
|
| /linux/drivers/iio/ |
| H A D | industrialio-buffer.c | 53 struct dma_buf_attachment *attach; member 1562 struct dma_buf_attachment *attach = priv->attach; in iio_buffer_dmabuf_release() local 1564 struct dma_buf *dmabuf = attach->dmabuf; in iio_buffer_dmabuf_release() 1566 dma_buf_unmap_attachment_unlocked(attach, priv->sgt, priv->dir); in iio_buffer_dmabuf_release() 1570 dma_buf_detach(attach->dmabuf, attach); in iio_buffer_dmabuf_release() 1575 static void iio_buffer_dmabuf_get(struct dma_buf_attachment *attach) in iio_buffer_dmabuf_get() argument 1577 struct iio_dmabuf_priv *priv = attach->importer_priv; in iio_buffer_dmabuf_get() 1582 static void iio_buffer_dmabuf_put(struct dma_buf_attachment *attach) in iio_buffer_dmabuf_put() argument 1584 struct iio_dmabuf_priv *priv = attach->importer_priv; in iio_buffer_dmabuf_put() 1603 iio_buffer_dmabuf_put(priv->attach); in iio_buffer_chrdev_release() [all …]
|
| /linux/drivers/gpu/drm/panfrost/ |
| H A D | panfrost_gem.c | 281 panfrost_gem_prime_map_dma_buf(struct dma_buf_attachment *attach, in panfrost_gem_prime_map_dma_buf() argument 284 struct sg_table *sgt = drm_gem_map_dma_buf(attach, dir); in panfrost_gem_prime_map_dma_buf() 287 attach->priv = sgt; in panfrost_gem_prime_map_dma_buf() 293 panfrost_gem_prime_unmap_dma_buf(struct dma_buf_attachment *attach, in panfrost_gem_prime_unmap_dma_buf() argument 297 attach->priv = NULL; in panfrost_gem_prime_unmap_dma_buf() 298 drm_gem_unmap_dma_buf(attach, sgt, dir); in panfrost_gem_prime_unmap_dma_buf() 308 struct dma_buf_attachment *attach; in panfrost_gem_prime_begin_cpu_access() local 317 list_for_each_entry(attach, &dma_buf->attachments, node) { in panfrost_gem_prime_begin_cpu_access() 318 struct sg_table *sgt = attach->priv; in panfrost_gem_prime_begin_cpu_access() 321 dma_sync_sgtable_for_cpu(attach->dev, sgt, dir); in panfrost_gem_prime_begin_cpu_access() [all …]
|
| /linux/Documentation/bpf/ |
| H A D | map_cgroup_storage.rst | 10 attach to cgroups; the programs are made available by the same Kconfig. The 34 ``attach_type`` is the program's attach type. 37 When this key type is used, then all attach types of the particular cgroup and 39 ``struct bpf_cgroup_storage_key``, then programs of different attach types 133 multiple attach types, and each attach creates a fresh zeroed storage. The 145 does not already contain an entry for the cgroup and attach type pair, or else 146 the old storage is reused for the new attachment. If the map is attach type 147 shared, then attach type is simply ignored during comparison. Storage is freed 158 In all versions, userspace may use the attach parameters of cgroup and 159 attach type pair in ``struct bpf_cgroup_storage_key`` as the key to the BPF map [all …]
|
| /linux/drivers/gpu/drm/armada/ |
| H A D | armada_gem.c | 387 armada_gem_prime_map_dma_buf(struct dma_buf_attachment *attach, in armada_gem_prime_map_dma_buf() argument 390 struct drm_gem_object *obj = attach->dmabuf->priv; in armada_gem_prime_map_dma_buf() 420 if (dma_map_sgtable(attach->dev, sgt, dir, 0)) in armada_gem_prime_map_dma_buf() 429 if (dma_map_sgtable(attach->dev, sgt, dir, 0)) in armada_gem_prime_map_dma_buf() 453 static void armada_gem_prime_unmap_dma_buf(struct dma_buf_attachment *attach, in armada_gem_prime_unmap_dma_buf() argument 456 struct drm_gem_object *obj = attach->dmabuf->priv; in armada_gem_prime_unmap_dma_buf() 461 dma_unmap_sgtable(attach->dev, sgt, dir, 0); in armada_gem_prime_unmap_dma_buf() 503 struct dma_buf_attachment *attach; in armada_gem_prime_import() local 518 attach = dma_buf_attach(buf, dev->dev); in armada_gem_prime_import() 519 if (IS_ERR(attach)) in armada_gem_prime_import() [all …]
|
| /linux/drivers/virtio/ |
| H A D | virtio_dma_buf.c | 28 exp_info->ops->attach != &virtio_dma_buf_attach || in virtio_dma_buf_export() 43 struct dma_buf_attachment *attach) in virtio_dma_buf_attach() argument 51 ret = ops->device_attach(dma_buf, attach); in virtio_dma_buf_attach() 65 return dma_buf->ops->attach == &virtio_dma_buf_attach; in is_virtio_dma_buf()
|
| /linux/drivers/gpu/drm/panthor/ |
| H A D | panthor_gem.c | 240 panthor_gem_prime_map_dma_buf(struct dma_buf_attachment *attach, in panthor_gem_prime_map_dma_buf() argument 243 struct sg_table *sgt = drm_gem_map_dma_buf(attach, dir); in panthor_gem_prime_map_dma_buf() 246 attach->priv = sgt; in panthor_gem_prime_map_dma_buf() 252 panthor_gem_prime_unmap_dma_buf(struct dma_buf_attachment *attach, in panthor_gem_prime_unmap_dma_buf() argument 256 attach->priv = NULL; in panthor_gem_prime_unmap_dma_buf() 257 drm_gem_unmap_dma_buf(attach, sgt, dir); in panthor_gem_prime_unmap_dma_buf() 267 struct dma_buf_attachment *attach; in panthor_gem_prime_begin_cpu_access() local 276 list_for_each_entry(attach, &dma_buf->attachments, node) { in panthor_gem_prime_begin_cpu_access() 277 struct sg_table *sgt = attach->priv; in panthor_gem_prime_begin_cpu_access() 280 dma_sync_sgtable_for_cpu(attach->dev, sgt, dir); in panthor_gem_prime_begin_cpu_access() [all …]
|