| /linux/net/xdp/ |
| H A D | xsk_buff_pool.c | 315 struct xsk_dma_map *dma_map; in xp_find_dma_map() local 317 list_for_each_entry(dma_map, &pool->umem->xsk_dma_list, list) { in xp_find_dma_map() 318 if (dma_map->netdev == pool->netdev) in xp_find_dma_map() 319 return dma_map; in xp_find_dma_map() 328 struct xsk_dma_map *dma_map; in xp_create_dma_map() local 330 dma_map = kzalloc_obj(*dma_map); in xp_create_dma_map() 331 if (!dma_map) in xp_create_dma_map() 334 dma_map->dma_pages = kvzalloc_objs(*dma_map->dma_pages, nr_pages); in xp_create_dma_map() 335 if (!dma_map->dma_pages) { in xp_create_dma_map() 336 kfree(dma_map); in xp_create_dma_map() [all …]
|
| /linux/drivers/misc/genwqe/ |
| H A D | card_dev.c | 121 struct dma_mapping *dma_map) in __genwqe_add_mapping() argument 126 list_add(&dma_map->card_list, &cfile->map_list); in __genwqe_add_mapping() 131 struct dma_mapping *dma_map) in __genwqe_del_mapping() argument 136 list_del(&dma_map->card_list); in __genwqe_del_mapping() 193 struct dma_mapping *dma_map; in genwqe_remove_mappings() local 198 dma_map = list_entry(node, struct dma_mapping, card_list); in genwqe_remove_mappings() 200 list_del_init(&dma_map->card_list); in genwqe_remove_mappings() 211 __func__, i++, dma_map->u_vaddr, in genwqe_remove_mappings() 212 (unsigned long)dma_map->k_vaddr, in genwqe_remove_mappings() 213 (unsigned long)dma_map->dma_addr); in genwqe_remove_mappings() [all …]
|
| /linux/drivers/scsi/ibmvscsi_tgt/ |
| H A D | libsrp.c | 163 int dma_map, int ext_desc) in srp_direct_data() argument 169 if (dma_map) { in srp_direct_data() 186 if (dma_map) in srp_direct_data() 196 int dma_map, int ext_desc) in srp_indirect_data() argument 205 if (dma_map || ext_desc) { in srp_indirect_data() 218 if (ext_desc && dma_map) { in srp_indirect_data() 243 if (dma_map) { in srp_indirect_data() 259 if (dma_map) in srp_indirect_data() 264 if (token && dma_map) { in srp_indirect_data() 298 srp_rdma_t rdma_io, int dma_map, int ext_desc) in srp_transfer_data() argument [all …]
|
| /linux/drivers/gpu/drm/i915/gem/selftests/ |
| H A D | i915_gem_dmabuf.c | 349 void *obj_map, *dma_map; in igt_dmabuf_import() local 380 dma_map = err ? NULL : map.vaddr; in igt_dmabuf_import() 381 if (!dma_map) { in igt_dmabuf_import() 396 memset(dma_map, pattern[i], PAGE_SIZE); in igt_dmabuf_import() 407 if (memchr_inv(dma_map, pattern[i], PAGE_SIZE)) { in igt_dmabuf_import()
|
| /linux/arch/um/drivers/ |
| H A D | vfio_user.c | 57 struct vfio_iommu_type1_dma_map dma_map = { in uml_vfio_user_setup_iommu() local 58 .argsz = sizeof(dma_map), in uml_vfio_user_setup_iommu() 68 if (ioctl(container, VFIO_IOMMU_MAP_DMA, &dma_map) < 0) in uml_vfio_user_setup_iommu()
|
| /linux/net/core/ |
| H A D | page_pool.c | 225 pool->dma_map = true; in page_pool_init() 286 if (!pool->dma_map || !pool->dma_sync) { in page_pool_init() 574 if (pool->dma_map && unlikely(!page_pool_dma_map(pool, page_to_netmem(page), gfp))) { in __page_pool_alloc_page_order() 595 bool dma_map = pool->dma_map; in __page_pool_alloc_netmems_slow() 626 if (dma_map && unlikely(!page_pool_dma_map(pool, netmem, gfp))) { in __page_pool_alloc_netmems_slow() 730 if (!pool->dma_map) in __page_pool_release_netmem_dma() 1160 if (!pool->destroy_cnt++ && pool->dma_map) { in page_pool_scrub() 591 bool dma_map = pool->dma_map; __page_pool_alloc_netmems_slow() local
|
| /linux/drivers/usb/host/ |
| H A D | r8a66597.h | 80 unsigned char dma_map; member 121 unsigned char dma_map; member
|
| H A D | r8a66597-hcd.c | 768 if ((r8a66597->dma_map & (1 << i)) != 0) in enable_r8a66597_pipe_dma() 778 r8a66597->dma_map |= 1 << i; in enable_r8a66597_pipe_dma() 779 dev->dma_map |= 1 << i; in enable_r8a66597_pipe_dma() 883 r8a66597->dma_map &= ~(dev->dma_map); in disable_r8a66597_pipe_all() 884 dev->dma_map = 0; in disable_r8a66597_pipe_all()
|
| /linux/drivers/net/ethernet/qlogic/qed/ |
| H A D | qed_chain.c | 120 entry->dma_map); in qed_chain_free_pbl() 305 addr_tbl[i].dma_map = phys; in qed_chain_alloc_pbl()
|
| /linux/drivers/net/vmxnet3/ |
| H A D | vmxnet3_xdp.c | 117 struct vmxnet3_tx_queue *tq, bool dma_map) in vmxnet3_xdp_xmit_frame() argument 143 if (dma_map) { /* ndo_xdp_xmit */ in vmxnet3_xdp_xmit_frame()
|
| /linux/drivers/vhost/ |
| H A D | vdpa.c | 459 return (!ops->set_map && !ops->dma_map) || ops->reset_map || in vhost_vdpa_has_persistent_map() 913 if (ops->dma_map) { in vhost_vdpa_general_unmap() 1006 if (ops->dma_map) { in vhost_vdpa_map() 1007 r = ops->dma_map(vdpa, asid, iova, size, pa, perm, opaque); in vhost_vdpa_map() 1328 if (ops->set_map || ops->dma_map) in vhost_vdpa_alloc_domain() 1571 if (!ops->set_map && !ops->dma_map && in vhost_vdpa_probe()
|
| /linux/include/net/page_pool/ |
| H A D | types.h | 174 bool dma_map:1; /* Perform DMA mapping */ member
|
| /linux/include/trace/events/ |
| H A D | dma.h | 39 DECLARE_EVENT_CLASS(dma_map, 72 DEFINE_EVENT(dma_map, name, \
|
| /linux/drivers/infiniband/hw/bnxt_re/ |
| H A D | qplib_res.c | 822 stats->dma, stats->dma_map); in bnxt_qplib_free_stats_ctx() 836 &stats->dma_map, GFP_KERNEL); in bnxt_qplib_alloc_stats_ctx()
|
| H A D | main.c | 1011 req.stats_dma_addr = cpu_to_le64(stats->dma_map); in bnxt_re_net_stats_ctx_alloc()
|
| /linux/include/linux/qed/ |
| H A D | qed_chain.h | 72 dma_addr_t dma_map; member
|
| /linux/drivers/infiniband/hw/bng_re/ |
| H A D | bng_dev.c | 202 req.stats_dma_addr = cpu_to_le64(stats->dma_map); in bng_re_stats_ctx_alloc()
|
| /linux/drivers/vdpa/vdpa_sim/ |
| H A D | vdpa_sim.c | 796 .dma_map = vdpasim_dma_map,
|
| /linux/drivers/net/ethernet/mediatek/ |
| H A D | mtk_eth_soc.c | 1936 void *data, u16 headroom, int index, bool dma_map) in mtk_xdp_frame_map() argument 1942 if (dma_map) { /* ndo_xdp_xmit */ in mtk_xdp_frame_map() 1960 tx_buf->type = dma_map ? MTK_TYPE_XDP_NDO : MTK_TYPE_XDP_TX; in mtk_xdp_frame_map() 1971 struct net_device *dev, bool dma_map) in mtk_xdp_submit_frame() argument 2010 data, xdpf->headroom, index, dma_map); in mtk_xdp_submit_frame()
|
| /linux/Documentation/virt/hyperv/ |
| H A D | vpci.rst | 253 operations as part of dma_map/unmap_*() calls.
|
| /linux/io_uring/ |
| H A D | zcrx.c | 1045 if (WARN_ON_ONCE(!pp->dma_map)) in io_pp_zc_init()
|
| /linux/drivers/net/ethernet/marvell/ |
| H A D | mvneta.c | 2092 struct xdp_frame *xdpf, int *nxmit_byte, bool dma_map) in mvneta_xdp_submit_frame() argument 2118 if (dma_map) { in mvneta_xdp_submit_frame()
|
| /linux/drivers/atm/ |
| H A D | fore200e.c | 1724 …oc3_regs_dma_addr = fore200e->bus->dma_map(fore200e, regs, sizeof(struct oc3_regs), DMA_FROM_DEVIC…
|
| /linux/drivers/net/ethernet/marvell/mvpp2/ |
| H A D | mvpp2_main.c | 3668 struct xdp_frame *xdpf, bool dma_map) in mvpp2_xdp_submit_frame() argument 3697 if (dma_map) { in mvpp2_xdp_submit_frame()
|
| /linux/drivers/net/ethernet/stmicro/stmmac/ |
| H A D | stmmac_main.c | 5067 struct xdp_frame *xdpf, bool dma_map) in stmmac_xdp_xmit_xdpf() argument 5094 if (dma_map) { in stmmac_xdp_xmit_xdpf()
|