| /linux/drivers/dma-buf/ |
| H A D | dma-buf.c | 909 dma_buf_attachment_is_dynamic(struct dma_buf_attachment *attach) in dma_buf_attachment_is_dynamic() argument 911 return !!attach->importer_ops; in dma_buf_attachment_is_dynamic() 915 dma_buf_pin_on_map(struct dma_buf_attachment *attach) in dma_buf_pin_on_map() argument 917 return attach->dmabuf->ops->pin && in dma_buf_pin_on_map() 918 !dma_buf_attachment_is_dynamic(attach); in dma_buf_pin_on_map() 1012 struct dma_buf_attachment *attach; in dma_buf_dynamic_attach() local 1018 attach = kzalloc_obj(*attach); in dma_buf_dynamic_attach() 1019 if (!attach) in dma_buf_dynamic_attach() 1022 attach->dev = dev; in dma_buf_dynamic_attach() 1023 attach->dmabuf = dmabuf; in dma_buf_dynamic_attach() [all …]
|
| H A D | dma-buf-mapping.c | 92 struct sg_table *dma_buf_phys_vec_to_sgt(struct dma_buf_attachment *attach, in dma_buf_phys_vec_to_sgt() argument 105 dma_resv_assert_held(attach->dmabuf->resv); in dma_buf_phys_vec_to_sgt() 107 if (WARN_ON(!attach || !attach->dmabuf || !provider)) in dma_buf_phys_vec_to_sgt() 115 switch (pci_p2pdma_map_type(provider, attach->dev)) { in dma_buf_phys_vec_to_sgt() 128 dma_iova_try_alloc(attach->dev, dma->state, 0, size); in dma_buf_phys_vec_to_sgt() 147 ret = dma_iova_link(attach->dev, dma->state, in dma_buf_phys_vec_to_sgt() 156 addr = dma_map_phys(attach->dev, phys_vec[i].paddr, in dma_buf_phys_vec_to_sgt() 159 ret = dma_mapping_error(attach->dev, addr); in dma_buf_phys_vec_to_sgt() 170 ret = dma_iova_sync(attach->dev, dma->state, 0, mapped_len); in dma_buf_phys_vec_to_sgt() 197 dma_iova_destroy(attach->dev, dma->state, mapped_len, dir, in dma_buf_phys_vec_to_sgt() [all …]
|
| /linux/drivers/infiniband/core/ |
| H A D | umem_dmabuf.c | 24 dma_resv_assert_held(umem_dmabuf->attach->dmabuf->resv); in ib_umem_dmabuf_map_pages() 32 sgt = dma_buf_map_attachment(umem_dmabuf->attach, in ib_umem_dmabuf_map_pages() 75 ret = dma_resv_wait_timeout(umem_dmabuf->attach->dmabuf->resv, in ib_umem_dmabuf_map_pages() 88 dma_resv_assert_held(umem_dmabuf->attach->dmabuf->resv); in ib_umem_dmabuf_unmap_pages() 109 dma_buf_unmap_attachment(umem_dmabuf->attach, umem_dmabuf->sgt, in ib_umem_dmabuf_unmap_pages() 155 umem_dmabuf->attach = dma_buf_dynamic_attach( in ib_umem_dmabuf_get_with_dma_device() 160 if (IS_ERR(umem_dmabuf->attach)) { in ib_umem_dmabuf_get_with_dma_device() 161 ret = ERR_CAST(umem_dmabuf->attach); in ib_umem_dmabuf_get_with_dma_device() 188 static void ib_umem_dmabuf_revoke_locked(struct dma_buf_attachment *attach) in ib_umem_dmabuf_revoke_locked() argument 190 struct ib_umem_dmabuf *umem_dmabuf = attach->importer_priv; in ib_umem_dmabuf_revoke_locked() [all …]
|
| /linux/drivers/media/common/videobuf2/ |
| H A D | videobuf2-vmalloc.c | 215 struct vb2_vmalloc_attachment *attach; in vb2_vmalloc_dmabuf_ops_attach() local 224 attach = kzalloc_obj(*attach); in vb2_vmalloc_dmabuf_ops_attach() 225 if (!attach) in vb2_vmalloc_dmabuf_ops_attach() 228 sgt = &attach->sgt; in vb2_vmalloc_dmabuf_ops_attach() 231 kfree(attach); in vb2_vmalloc_dmabuf_ops_attach() 239 kfree(attach); in vb2_vmalloc_dmabuf_ops_attach() 246 attach->dma_dir = DMA_NONE; in vb2_vmalloc_dmabuf_ops_attach() 247 dbuf_attach->priv = attach; in vb2_vmalloc_dmabuf_ops_attach() 254 struct vb2_vmalloc_attachment *attach = db_attach->priv; in vb2_vmalloc_dmabuf_ops_detach() local 257 if (!attach) in vb2_vmalloc_dmabuf_ops_detach() [all …]
|
| H A D | videobuf2-dma-contig.c | 321 struct vb2_dc_attachment *attach; in vb2_dc_dmabuf_ops_attach() local 328 attach = kzalloc_obj(*attach); in vb2_dc_dmabuf_ops_attach() 329 if (!attach) in vb2_dc_dmabuf_ops_attach() 332 sgt = &attach->sgt; in vb2_dc_dmabuf_ops_attach() 338 kfree(attach); in vb2_dc_dmabuf_ops_attach() 350 attach->dma_dir = DMA_NONE; in vb2_dc_dmabuf_ops_attach() 351 dbuf_attach->priv = attach; in vb2_dc_dmabuf_ops_attach() 359 struct vb2_dc_attachment *attach = db_attach->priv; in vb2_dc_dmabuf_ops_detach() local 362 if (!attach) in vb2_dc_dmabuf_ops_detach() 365 sgt = &attach->sgt; in vb2_dc_dmabuf_ops_detach() [all …]
|
| /linux/drivers/xen/ |
| H A D | gntdev-dmabuf.c | 48 struct dma_buf_attachment *attach; member 221 struct dma_buf_attachment *attach) in dmabuf_exp_ops_attach() argument 230 attach->priv = gntdev_dmabuf_attach; in dmabuf_exp_ops_attach() 235 struct dma_buf_attachment *attach) in dmabuf_exp_ops_detach() argument 237 struct gntdev_dmabuf_attachment *gntdev_dmabuf_attach = attach->priv; in dmabuf_exp_ops_detach() 244 dma_unmap_sgtable(attach->dev, sgt, in dmabuf_exp_ops_detach() 252 attach->priv = NULL; in dmabuf_exp_ops_detach() 257 dmabuf_exp_ops_map_dma_buf(struct dma_buf_attachment *attach, in dmabuf_exp_ops_map_dma_buf() argument 260 struct gntdev_dmabuf_attachment *gntdev_dmabuf_attach = attach->priv; in dmabuf_exp_ops_map_dma_buf() 261 struct gntdev_dmabuf *gntdev_dmabuf = attach->dmabuf->priv; in dmabuf_exp_ops_map_dma_buf() [all …]
|
| /linux/drivers/gpu/drm/omapdrm/ |
| H A D | omap_gem_dmabuf.c | 100 struct dma_buf_attachment *attach; in omap_gem_prime_import() local 117 attach = dma_buf_attach(dma_buf, dev->dev); in omap_gem_prime_import() 118 if (IS_ERR(attach)) in omap_gem_prime_import() 119 return ERR_CAST(attach); in omap_gem_prime_import() 123 sgt = dma_buf_map_attachment_unlocked(attach, DMA_TO_DEVICE); in omap_gem_prime_import() 135 obj->import_attach = attach; in omap_gem_prime_import() 140 dma_buf_unmap_attachment_unlocked(attach, sgt, DMA_TO_DEVICE); in omap_gem_prime_import() 142 dma_buf_detach(dma_buf, attach); in omap_gem_prime_import()
|
| /linux/drivers/gpu/drm/i915/gem/ |
| H A D | i915_gem_dmabuf.c | 27 static struct sg_table *i915_gem_map_dma_buf(struct dma_buf_attachment *attach, in i915_gem_map_dma_buf() argument 30 struct drm_i915_gem_object *obj = dma_buf_to_obj(attach->dmabuf); in i915_gem_map_dma_buf() 55 ret = dma_map_sgtable(attach->dev, sgt, dir, DMA_ATTR_SKIP_CPU_SYNC); in i915_gem_map_dma_buf() 167 struct dma_buf_attachment *attach) in i915_gem_dmabuf_attach() argument 196 struct dma_buf_attachment *attach) in i915_gem_dmabuf_detach() argument 204 .attach = i915_gem_dmabuf_attach, 285 struct dma_buf_attachment *attach; in i915_gem_prime_import() local 307 attach = dma_buf_attach(dma_buf, dev->dev); in i915_gem_prime_import() 308 if (IS_ERR(attach)) in i915_gem_prime_import() 309 return ERR_CAST(attach); in i915_gem_prime_import() [all …]
|
| /linux/drivers/gpu/drm/tegra/ |
| H A D | gem.c | 82 map->attach = dma_buf_attach(buf, dev); in tegra_bo_pin() 83 if (IS_ERR(map->attach)) { in tegra_bo_pin() 84 err = PTR_ERR(map->attach); in tegra_bo_pin() 88 map->sgt = dma_buf_map_attachment_unlocked(map->attach, direction); in tegra_bo_pin() 90 dma_buf_detach(buf, map->attach); in tegra_bo_pin() 163 if (map->attach) { in tegra_bo_unpin() 164 dma_buf_unmap_attachment_unlocked(map->attach, map->sgt, in tegra_bo_unpin() 166 dma_buf_detach(map->attach->dmabuf, map->attach); in tegra_bo_unpin() 460 struct dma_buf_attachment *attach; in tegra_bo_import() local 473 attach = dma_buf_attach(buf, drm->dev); in tegra_bo_import() [all …]
|
| /linux/tools/bpf/bpftool/Documentation/ |
| H A D | bpftool-cgroup.rst | 22 { **show** | **list** | **tree** | **attach** | **detach** | **help** } 29 | **bpftool** **cgroup attach** *CGROUP* *ATTACH_TYPE* *PROG* [*ATTACH_FLAGS*] 55 Output will start with program ID followed by attach type, attach flags and 67 with absolute cgroup path, followed by program ID, attach type, attach 74 bpftool cgroup attach *CGROUP* *ATTACH_TYPE* *PROG* [*ATTACH_FLAGS*] 75 Attach program *PROG* to the cgroup *CGROUP* with attach type *ATTACH_TYPE* 83 Only one program is allowed to be attached to a cgroup with no attach flags 85 program and attach the new one. 125 Detach *PROG* from the cgroup *CGROUP* and attach type *ATTACH_TYPE*. 143 | **# bpftool cgroup attach /sys/fs/cgroup/test.slice/ device id 1 allow_multi**
|
| H A D | bpftool-net.rst | 21 *COMMANDS* := { **show** | **list** | **attach** | **detach** | **help** } 27 | **bpftool** **net attach** *ATTACH_TYPE* *PROG* **dev** *NAME* [ **overwrite** | **prepend** ] 61 bpftool net attach *ATTACH_TYPE* *PROG* dev *NAME* [ overwrite | prepend ] 73 For XDP-related attach types (**xdp**, **xdpgeneric**, **xdpdrv**, 77 For **tcx_ingress** and **tcx_egress** attach types, the **prepend** option 78 can be used to attach the program at the beginning of the chain instead of 84 for attach must be specified. 162 | **# bpftool net attach xdpdrv id 16 dev enp6s0np0** 171 | **# bpftool net attach xdpdrv id 16 dev enp6s0np0** 172 | **# bpftool net attach xdpdr [all...] |
| /linux/drivers/gpu/drm/vmwgfx/ |
| H A D | vmwgfx_prime.c | 45 struct dma_buf_attachment *attach) in vmw_prime_map_attach() argument 51 struct dma_buf_attachment *attach) in vmw_prime_map_detach() argument 55 static struct sg_table *vmw_prime_map_dma_buf(struct dma_buf_attachment *attach, in vmw_prime_map_dma_buf() argument 61 static void vmw_prime_unmap_dma_buf(struct dma_buf_attachment *attach, in vmw_prime_unmap_dma_buf() argument 68 .attach = vmw_prime_map_attach,
|
| /linux/drivers/iio/ |
| H A D | industrialio-buffer.c | 53 struct dma_buf_attachment *attach; member 1564 struct dma_buf_attachment *attach = priv->attach; in iio_buffer_dmabuf_release() local 1566 struct dma_buf *dmabuf = attach->dmabuf; in iio_buffer_dmabuf_release() 1568 dma_buf_unmap_attachment_unlocked(attach, priv->sgt, priv->dir); in iio_buffer_dmabuf_release() 1572 dma_buf_detach(attach->dmabuf, attach); in iio_buffer_dmabuf_release() 1577 static void iio_buffer_dmabuf_get(struct dma_buf_attachment *attach) in iio_buffer_dmabuf_get() argument 1579 struct iio_dmabuf_priv *priv = attach->importer_priv; in iio_buffer_dmabuf_get() 1584 static void iio_buffer_dmabuf_put(struct dma_buf_attachment *attach) in iio_buffer_dmabuf_put() argument 1586 struct iio_dmabuf_priv *priv = attach->importer_priv; in iio_buffer_dmabuf_put() 1605 iio_buffer_dmabuf_put(priv->attach); in iio_buffer_chrdev_release() [all …]
|
| /linux/drivers/virtio/ |
| H A D | virtio_dma_buf.c | 28 exp_info->ops->attach != &virtio_dma_buf_attach || in virtio_dma_buf_export() 43 struct dma_buf_attachment *attach) in virtio_dma_buf_attach() argument 51 ret = ops->device_attach(dma_buf, attach); in virtio_dma_buf_attach() 65 return dma_buf->ops->attach == &virtio_dma_buf_attach; in is_virtio_dma_buf()
|
| /linux/Documentation/bpf/ |
| H A D | map_cgroup_storage.rst | 10 attach to cgroups; the programs are made available by the same Kconfig. The 34 ``attach_type`` is the program's attach type. 37 When this key type is used, then all attach types of the particular cgroup and 39 ``struct bpf_cgroup_storage_key``, then programs of different attach types 133 multiple attach types, and each attach creates a fresh zeroed storage. The 145 does not already contain an entry for the cgroup and attach type pair, or else 146 the old storage is reused for the new attachment. If the map is attach type 147 shared, then attach type is simply ignored during comparison. Storage is freed 158 In all versions, userspace may use the attach parameters of cgroup and 159 attach type pair in ``struct bpf_cgroup_storage_key`` as the key to the BPF map [all …]
|
| /linux/drivers/gpu/drm/armada/ |
| H A D | armada_gem.c | 387 armada_gem_prime_map_dma_buf(struct dma_buf_attachment *attach, in armada_gem_prime_map_dma_buf() argument 390 struct drm_gem_object *obj = attach->dmabuf->priv; in armada_gem_prime_map_dma_buf() 420 if (dma_map_sgtable(attach->dev, sgt, dir, 0)) in armada_gem_prime_map_dma_buf() 429 if (dma_map_sgtable(attach->dev, sgt, dir, 0)) in armada_gem_prime_map_dma_buf() 453 static void armada_gem_prime_unmap_dma_buf(struct dma_buf_attachment *attach, in armada_gem_prime_unmap_dma_buf() argument 456 struct drm_gem_object *obj = attach->dmabuf->priv; in armada_gem_prime_unmap_dma_buf() 461 dma_unmap_sgtable(attach->dev, sgt, dir, 0); in armada_gem_prime_unmap_dma_buf() 503 struct dma_buf_attachment *attach; in armada_gem_prime_import() local 518 attach = dma_buf_attach(buf, dev->dev); in armada_gem_prime_import() 519 if (IS_ERR(attach)) in armada_gem_prime_import() [all …]
|
| /linux/drivers/gpu/drm/panthor/ |
| H A D | panthor_gem.c | 240 panthor_gem_prime_map_dma_buf(struct dma_buf_attachment *attach, in panthor_gem_prime_map_dma_buf() argument 243 struct sg_table *sgt = drm_gem_map_dma_buf(attach, dir); in panthor_gem_prime_map_dma_buf() 246 attach->priv = sgt; in panthor_gem_prime_map_dma_buf() 252 panthor_gem_prime_unmap_dma_buf(struct dma_buf_attachment *attach, in panthor_gem_prime_unmap_dma_buf() argument 256 attach->priv = NULL; in panthor_gem_prime_unmap_dma_buf() 257 drm_gem_unmap_dma_buf(attach, sgt, dir); in panthor_gem_prime_unmap_dma_buf() 267 struct dma_buf_attachment *attach; in panthor_gem_prime_begin_cpu_access() local 276 list_for_each_entry(attach, &dma_buf->attachments, node) { in panthor_gem_prime_begin_cpu_access() 277 struct sg_table *sgt = attach->priv; in panthor_gem_prime_begin_cpu_access() 280 dma_sync_sgtable_for_cpu(attach->dev, sgt, dir); in panthor_gem_prime_begin_cpu_access() [all …]
|
| /linux/samples/bpf/ |
| H A D | tcp_bpf.readme | 2 programs. These programs attach to a cgroupv2. The following commands create 3 a cgroupv2 and attach a bash shell to the group. 12 (attach) one of the tcp_*_kern.o programs: 15 bpftool cgroup attach /tmp/cgroupv2/foo sock_ops pinned /sys/fs/bpf/tcp_prog
|
| H A D | xdp_fwd_user.c | 129 int attach = 1; in main() local 135 attach = 0; in main() 160 if (attach) { in main() 214 if (!attach) { in main()
|
| /linux/tools/testing/selftests/bpf/ |
| H A D | flow_dissector_load.c | 62 bool attach = false; in parse_opts() local 71 attach = true; in parse_opts() 74 if (attach) in parse_opts()
|
| /linux/security/apparmor/ |
| H A D | domain.c | 317 struct aa_attachment *attach = &profile->attach; in aa_xattrs_match() local 318 int size, value_size = 0, ret = attach->xattr_count; in aa_xattrs_match() 320 if (!bprm || !attach->xattr_count) in aa_xattrs_match() 325 state = aa_dfa_outofband_transition(attach->xmatch->dfa, state); in aa_xattrs_match() 328 for (i = 0; i < attach->xattr_count; i++) { in aa_xattrs_match() 329 size = vfs_getxattr_alloc(&nop_mnt_idmap, d, attach->xattrs[i], in aa_xattrs_match() 339 state = aa_dfa_null_transition(attach->xmatch->dfa, in aa_xattrs_match() 342 state = aa_dfa_match_len(attach->xmatch->dfa, state, in aa_xattrs_match() 344 perms = aa_lookup_perms(attach->xmatch, state); in aa_xattrs_match() 351 state = aa_dfa_outofband_transition(attach->xmatch->dfa, state); in aa_xattrs_match() [all …]
|
| /linux/drivers/nvdimm/ |
| H A D | claim.c | 43 bool __nd_attach_ndns(struct device *dev, struct nd_namespace_common *attach, in __nd_attach_ndns() argument 46 struct nvdimm_bus *nvdimm_bus = walk_to_nvdimm_bus(&attach->dev); in __nd_attach_ndns() 48 if (attach->claim) in __nd_attach_ndns() 52 attach->claim = dev; in __nd_attach_ndns() 53 *_ndns = attach; in __nd_attach_ndns() 54 get_device(&attach->dev); in __nd_attach_ndns()
|
| /linux/drivers/gpu/drm/radeon/ |
| H A D | radeon_prime.c | 46 struct dma_buf_attachment *attach, in radeon_gem_prime_import_sg_table() argument 49 struct dma_resv *resv = attach->dmabuf->resv; in radeon_gem_prime_import_sg_table() 55 ret = radeon_bo_create(rdev, attach->dmabuf->size, PAGE_SIZE, false, in radeon_gem_prime_import_sg_table()
|
| /linux/tools/testing/selftests/drivers/usb/usbip/ |
| H A D | usbip_test.sh | 125 src/usbip attach -r localhost -b $busid; 141 src/usbip attach -r localhost -b $busid; 153 src/usbip attach -r localhost -b $busid;
|
| /linux/drivers/accel/amdxdna/ |
| H A D | amdxdna_ubuf.c | 23 static struct sg_table *amdxdna_ubuf_map(struct dma_buf_attachment *attach, in amdxdna_ubuf_map() argument 26 struct amdxdna_ubuf_priv *ubuf = attach->dmabuf->priv; in amdxdna_ubuf_map() 39 ret = dma_map_sgtable(attach->dev, sg, direction, 0); in amdxdna_ubuf_map() 52 static void amdxdna_ubuf_unmap(struct dma_buf_attachment *attach, in amdxdna_ubuf_unmap() argument 56 dma_unmap_sgtable(attach->dev, sg, direction, 0); in amdxdna_ubuf_unmap()
|