| /freebsd/sys/dev/smartpqi/ |
| H A D | smartpqi_queue.c | 155 dma_addr_t dma_addr = 0; in pqisrc_allocate_and_init_inbound_q() local 175 dma_mem->tag, dma_mem->size, dma_mem->align, dma_mem->virt_addr, (void*)dma_mem->dma_addr); in pqisrc_allocate_and_init_inbound_q() 179 dma_addr = dma_mem->dma_addr; in pqisrc_allocate_and_init_inbound_q() 181 ASSERT(!(dma_addr & PQI_ADDR_ALIGN_MASK)); in pqisrc_allocate_and_init_inbound_q() 184 ASSERT(!(dma_addr & PQI_ADDR_ALIGN_MASK)); in pqisrc_allocate_and_init_inbound_q() 186 ib_q->array_dma_addr = dma_addr; in pqisrc_allocate_and_init_inbound_q() 191 dma_addr += ib_array_size; in pqisrc_allocate_and_init_inbound_q() 194 ASSERT(!(dma_addr & PQI_CI_PI_ALIGN_MASK)); in pqisrc_allocate_and_init_inbound_q() 196 ib_q->ci_dma_addr = dma_addr; in pqisrc_allocate_and_init_inbound_q() 227 dma_addr_t dma_addr = 0; in pqisrc_allocate_and_init_outbound_q() local [all …]
|
| H A D | smartpqi_mem.c | 109 os_dma_map, &dma_mem->dma_addr, 0)) != 0) { in os_dma_mem_alloc() 145 if(dma_mem->dma_addr) { in os_dma_mem_free() 147 dma_mem->dma_addr = 0; in os_dma_mem_free()
|
| H A D | smartpqi_event.c | 357 request->sg_desc.addr = buff->dma_addr; in pqi_event_configure() 405 DBG_INFO("buf_report_event.dma_addr = %p \n",(void*)buf_report_event.dma_addr); in pqisrc_report_event_config() 467 DBG_INFO("buf_set_event.dma_addr = %p\n",(void*)buf_set_event.dma_addr); in pqisrc_set_event_config()
|
| H A D | smartpqi_ioctl.c | 298 DBG_IO("ioctl_dma_buf.dma_addr = %p\n",(void*)ioctl_dma_buf.dma_addr); in pqisrc_passthru_ioctl() 336 request.sg_descriptors[0].addr = ioctl_dma_buf.dma_addr; in pqisrc_passthru_ioctl() 490 DBG_IO("ioctl_dma_buf.dma_addr = %p\n",(void*)ioctl_dma_buf.dma_addr); in pqisrc_big_passthru_ioctl() 528 request.sg_descriptors[0].addr = ioctl_dma_buf.dma_addr; in pqisrc_big_passthru_ioctl()
|
| /freebsd/sys/dev/iser/ |
| H A D | iser_initiator.c | 108 tx_desc->dma_addr, ISER_HEADERS_LEN, DMA_TO_DEVICE); in iser_create_send_desc() 201 u64 dma_addr; in iser_alloc_rx_descriptors() local 224 dma_addr = ib_dma_map_single(device->ib_device, (void *)rx_desc, in iser_alloc_rx_descriptors() 226 if (ib_dma_mapping_error(device->ib_device, dma_addr)) in iser_alloc_rx_descriptors() 229 rx_desc->dma_addr = dma_addr; in iser_alloc_rx_descriptors() 232 rx_sg->addr = rx_desc->dma_addr; in iser_alloc_rx_descriptors() 244 ib_dma_unmap_single(device->ib_device, rx_desc->dma_addr, in iser_alloc_rx_descriptors() 268 ib_dma_unmap_single(device->ib_device, rx_desc->dma_addr, in iser_free_rx_descriptors() 483 rx_dma = rx_desc->dma_addr; in iser_rcv_completion()
|
| H A D | icl_iser.c | 93 u64 dma_addr; in iser_initialize_headers() local 96 dma_addr = ib_dma_map_single(device->ib_device, (void *)tx_desc, in iser_initialize_headers() 98 if (ib_dma_mapping_error(device->ib_device, dma_addr)) { in iser_initialize_headers() 104 tx_desc->dma_addr = dma_addr; in iser_initialize_headers() 105 tx_desc->tx_sg[0].addr = tx_desc->dma_addr; in iser_initialize_headers() 504 ib_dma_unmap_single(device->ib_device, tx_desc->dma_addr, in iser_conn_task_done()
|
| H A D | icl_iser.h | 276 u64 dma_addr; member 298 u64 dma_addr; member
|
| /freebsd/sys/compat/linuxkpi/common/include/linux/ |
| H A D | dma-mapping.h | 86 int (*mapping_error)(struct device *dev, dma_addr_t dma_addr); 104 void linux_dma_unmap(struct device *dev, dma_addr_t dma_addr, size_t size); /* backward compat */ 179 dma_addr_t dma_addr) in dma_free_coherent() argument 182 lkpi_dma_unmap(dev, dma_addr, size, DMA_BIDIRECTIONAL, 0); in dma_free_coherent() 327 dma_mapping_error(struct device *dev, dma_addr_t dma_addr) in dma_mapping_error() argument 330 if (dma_addr == 0 || dma_addr == DMA_MAPPING_ERROR) in dma_mapping_error() 366 #define dma_unmap_single_attrs(dev, dma_addr, size, dir, attrs) \ argument 367 _dma_unmap_single_attrs(dev, dma_addr, size, dir, 0)
|
| H A D | dmapool.h | 46 dma_addr_t dma_addr); 97 dma_pool_free(struct dma_pool *pool, void *vaddr, dma_addr_t dma_addr) in dma_pool_free() argument 100 linux_dma_pool_free(pool, vaddr, dma_addr); in dma_pool_free()
|
| H A D | usb.h | 299 uint16_t mem_flags, uint8_t *dma_addr); 302 void usb_buffer_free(struct usb_device *dev, usb_size_t size, void *addr, uint8_t dma_addr);
|
| /freebsd/sys/dev/mlx5/mlx5_core/ |
| H A D | mlx5_pagealloc.c | 63 fwp->dma_addr = segs->ds_addr; in mlx5_fwp_load_mem_cb() 194 return ((fwp + index)->dma_addr + (offset % MLX5_ADAPTER_PAGE_SIZE)); in mlx5_fwp_get_dma() 217 if (tfp->dma_addr < nfp->dma_addr) in mlx5_insert_fw_page_locked() 219 else if (tfp->dma_addr > nfp->dma_addr) in mlx5_insert_fw_page_locked() 240 if (tfp->dma_addr < addr) { in mlx5_remove_fw_page_locked() 242 } else if (tfp->dma_addr > addr) { in mlx5_remove_fw_page_locked() 276 *addr = fwp->dma_addr; in alloc_4k() 408 MLX5_ARRAY_SET64(manage_pages_out, out, pas, i, fwp->dma_addr); in reclaim_pages_cmd()
|
| /freebsd/sys/dev/liquidio/base/ |
| H A D | lio_request_manager.c | 716 uint64_t dma_addr; in lio_setup_sc_buffer_pool() local 725 lio_dma_alloc(LIO_SOFT_COMMAND_BUFFER_SIZE, (vm_paddr_t *)&dma_addr); in lio_setup_sc_buffer_pool() 731 sc->dma_addr = dma_addr; in lio_setup_sc_buffer_pool() 770 uint64_t dma_addr; in lio_alloc_soft_command() local 795 dma_addr = sc->dma_addr; in lio_alloc_soft_command() 800 sc->dma_addr = dma_addr; in lio_alloc_soft_command() 813 sc->dmadptr = dma_addr + offset; in lio_alloc_soft_command() 823 sc->dmarptr = dma_addr + offset; in lio_alloc_soft_command()
|
| /freebsd/sys/dev/liquidio/ |
| H A D | lio_network.h | 218 vm_paddr_t dma_addr; in lio_map_ring() local 220 dma_addr = vtophys(((struct mbuf *)buf)->m_data); in lio_map_ring() 221 return ((uint64_t)dma_addr); in lio_map_ring()
|
| /freebsd/sys/ofed/drivers/infiniband/core/ |
| H A D | ib_umem_odp.c | 442 dma_addr_t dma_addr; in ib_umem_odp_map_dma_single_page() local 457 dma_addr = ib_dma_map_page(dev, in ib_umem_odp_map_dma_single_page() 461 if (ib_dma_mapping_error(dev, dma_addr)) { in ib_umem_odp_map_dma_single_page() 465 umem->odp_data->dma_list[page_index] = dma_addr | access_mask; in ib_umem_odp_map_dma_single_page() 642 dma_addr_t dma_addr = dma & ODP_DMA_ADDR_MASK; in ib_umem_odp_unmap_dma_pages() local 644 WARN_ON(!dma_addr); in ib_umem_odp_unmap_dma_pages() 646 ib_dma_unmap_page(dev, dma_addr, PAGE_SIZE, in ib_umem_odp_unmap_dma_pages()
|
| /freebsd/sys/dev/mpi3mr/ |
| H A D | mpi3mr_app.c | 216 sgl_dma = (U64)sc->ioctl_chain_sge.dma_addr; in mpi3mr_app_construct_nvme_sgl() 240 sgl_dma = htole64(dma_buff->dma_desc[i].dma_addr); in mpi3mr_app_construct_nvme_sgl() 268 bus_addr_t prp_entry_dma, prp_page_dma, dma_addr; in mpi3mr_app_build_nvme_prp() local 324 dma_addr = dma_buff->dma_desc[count].dma_addr; in mpi3mr_app_build_nvme_prp() 325 if (dma_addr & page_mask) { in mpi3mr_app_build_nvme_prp() 327 __func__, dma_addr, dev_pgsz); in mpi3mr_app_build_nvme_prp() 332 dma_addr = dma_buff->dma_desc[0].dma_addr; in mpi3mr_app_build_nvme_prp() 410 offset = dma_addr & page_mask; in mpi3mr_app_build_nvme_prp() 418 *prp1_entry = dma_addr; in mpi3mr_app_build_nvme_prp() 463 *prp2_entry = dma_addr; in mpi3mr_app_build_nvme_prp() [all …]
|
| /freebsd/sys/contrib/dev/mediatek/mt76/ |
| H A D | dma.c | 35 t->dma_addr = addr; in mt76_alloc_txwi() 139 dma_unmap_single(dev->dma_dev, t->dma_addr, dev->drv->txwi_size, in mt76_free_pending_txwi() 298 entry->dma_addr[0] = buf->addr; in mt76_dma_add_rx_buf() 338 entry->dma_addr[0] = buf[0].addr; in mt76_dma_add_buf() 346 entry->dma_addr[1] = buf[1].addr; in mt76_dma_add_buf() 386 dma_unmap_single(dev->dma_dev, e->dma_addr[0], e->dma_len[0], in mt76_dma_tx_cleanup_idx() 390 dma_unmap_single(dev->dma_dev, e->dma_addr[1], e->dma_len[1], in mt76_dma_tx_cleanup_idx() 470 dma_sync_single_for_cpu(dev->dma_dev, t->dma_addr, in mt76_dma_get_rxdmad_c_buf() 487 t->dma_addr = 0; in mt76_dma_get_rxdmad_c_buf() 529 dma_sync_single_for_cpu(dev->dma_dev, t->dma_addr, in mt76_dma_get_buf() [all …]
|
| /freebsd/sys/dev/mlx5/mlx5_accel/ |
| H A D | mlx5_ipsec_offload.c | 269 ctrl->va_l = cpu_to_be32(lower_32_bits(aso->dma_addr) | ASO_CTRL_READ_EN); in mlx5e_ipsec_aso_query() 270 ctrl->va_h = cpu_to_be32(upper_32_bits(aso->dma_addr)); in mlx5e_ipsec_aso_query() 452 aso->dma_addr = dma_map_single(pdev, aso->ctx, sizeof(aso->ctx), DMA_BIDIRECTIONAL); in mlx5e_ipsec_aso_init() 453 err = dma_mapping_error(pdev, aso->dma_addr); in mlx5e_ipsec_aso_init() 468 dma_unmap_single(pdev, aso->dma_addr, sizeof(aso->ctx), DMA_BIDIRECTIONAL); in mlx5e_ipsec_aso_init() 484 dma_unmap_single(pdev, aso->dma_addr, sizeof(aso->ctx), DMA_BIDIRECTIONAL); in mlx5e_ipsec_aso_cleanup()
|
| /freebsd/sys/compat/linuxkpi/common/src/ |
| H A D | linux_pci.c | 1553 uint64_t dma_addr; 1599 PCTRIE_DEFINE(LINUX_DMA, linux_dma_obj, dma_addr, linux_dma_trie_alloc, 1654 obj->dma_addr = seg.ds_addr; in linux_dma_map_phys_common() 1665 return (obj->dma_addr); in linux_dma_map_phys_common() 1703 lkpi_dma_unmap(struct device *dev, dma_addr_t dma_addr, size_t len, 1715 obj = LINUX_DMA_PCTRIE_LOOKUP(&priv->ptree, dma_addr); in lkpi_dma_unmap() 1720 LINUX_DMA_PCTRIE_REMOVE(&priv->ptree, dma_addr); in lkpi_dma_unmap() 1750 lkpi_dma_unmap(struct device *dev, dma_addr_t dma_addr, size_t len, in linux_dma_alloc_coherent() 1758 linux_dma_unmap(struct device *dev, dma_addr_t dma_addr, size_t len) in linux_dma_alloc_coherent() 1760 lkpi_dma_unmap(dev, dma_addr, le in linux_dma_alloc_coherent() 1554 uint64_t dma_addr; global() member 1704 lkpi_dma_unmap(struct device * dev,dma_addr_t dma_addr,size_t len,enum dma_data_direction direction,unsigned long attrs) lkpi_dma_unmap() argument 1734 lkpi_dma_unmap(struct device * dev,dma_addr_t dma_addr,size_t len,enum dma_data_direction direction,unsigned long attrs) lkpi_dma_unmap() argument 1742 linux_dma_unmap(struct device * dev,dma_addr_t dma_addr,size_t len) linux_dma_unmap() argument 1860 linuxkpi_dma_sync(struct device * dev,dma_addr_t dma_addr,size_t size,bus_dmasync_op_t op) linuxkpi_dma_sync() argument 2143 linux_dma_pool_free(struct dma_pool * pool,void * vaddr,dma_addr_t dma_addr) linux_dma_pool_free() argument [all...] |
| /freebsd/sys/contrib/dev/mediatek/mt76/mt7996/ |
| H A D | mac.c | 1783 static u32 mt7996_msdu_page_hash_from_addr(dma_addr_t dma_addr) in mt7996_msdu_page_hash_from_addr() argument 1788 while (dma_addr) { in mt7996_msdu_page_hash_from_addr() 1789 val += (u32)((dma_addr & 0xff) + i) % MT7996_RRO_MSDU_PG_HASH_SIZE; in mt7996_msdu_page_hash_from_addr() 1790 dma_addr >>= 8; in mt7996_msdu_page_hash_from_addr() 1798 mt7996_rro_msdu_page_get(struct mt7996_dev *dev, dma_addr_t dma_addr) in mt7996_rro_msdu_page_get() argument 1800 u32 hash = mt7996_msdu_page_hash_from_addr(dma_addr); in mt7996_rro_msdu_page_get() 1807 if (p->dma_addr == dma_addr) { in mt7996_rro_msdu_page_get() 1831 t->dma_addr = 0; in mt7996_rx_token_put() 1862 dma_addr_t dma_addr, void *data) in mt7996_rro_msdu_page_add() argument 1875 p->dma_addr = dma_addr; in mt7996_rro_msdu_page_add() [all …]
|
| /freebsd/sys/dev/sfxge/common/ |
| H A D | ef10_ev.c | 152 efx_qword_t *dma_addr; in efx_mcdi_init_evq() local 231 dma_addr = MCDI_IN2(req, efx_qword_t, INIT_EVQ_IN_DMA_ADDR); in efx_mcdi_init_evq() 235 EFX_POPULATE_QWORD_2(*dma_addr, in efx_mcdi_init_evq() 239 dma_addr++; in efx_mcdi_init_evq() 289 efx_qword_t *dma_addr; in efx_mcdi_init_evq_v2() local 356 dma_addr = MCDI_IN2(req, efx_qword_t, INIT_EVQ_V2_IN_DMA_ADDR); in efx_mcdi_init_evq_v2() 360 EFX_POPULATE_QWORD_2(*dma_addr, in efx_mcdi_init_evq_v2() 364 dma_addr++; in efx_mcdi_init_evq_v2()
|
| H A D | ef10_tx.c | 60 efx_qword_t *dma_addr; in efx_mcdi_init_txq() local 109 dma_addr = MCDI_IN2(req, efx_qword_t, INIT_TXQ_IN_DMA_ADDR); in efx_mcdi_init_txq() 113 EFX_POPULATE_QWORD_2(*dma_addr, in efx_mcdi_init_txq() 117 dma_addr++; in efx_mcdi_init_txq()
|
| /freebsd/sys/dev/sfxge/ |
| H A D | sfxge_tx.c | 951 uint64_t dma_addr; /* DMA address of current position */ member 1155 uint64_t dma_addr = tso->dma_addr; in tso_fill_packet_with_fragment() local 1183 tso->dma_addr += n; in tso_fill_packet_with_fragment() 1195 efx_tx_qdesc_dma_create(txq->common, dma_addr, n, eop, desc); in tso_fill_packet_with_fragment() 1222 uint64_t dma_addr; in tso_start_new_packet() local 1278 dma_addr = (txq->tsoh_buffer[page_index].esm_addr + in tso_start_new_packet() 1293 tso_map_long_header, &dma_addr, in tso_start_new_packet() 1295 if (__predict_false(dma_addr == 0)) { in tso_start_new_packet() 1346 dma_addr, in tso_start_new_packet() 1379 tso.dma_addr = dma_seg->ds_addr + (tso.header_len - skipped); in sfxge_tx_queue_tso() [all …]
|
| /freebsd/sys/dev/mlx5/mlx5_fpga/ |
| H A D | mlx5fpga_conn.c | 56 buf->sg[0].dma_addr = dma_map_single(dma_device, buf->sg[0].data, in mlx5_fpga_conn_map_buf() 58 err = dma_mapping_error(dma_device, buf->sg[0].dma_addr); in mlx5_fpga_conn_map_buf() 68 buf->sg[1].dma_addr = dma_map_single(dma_device, buf->sg[1].data, in mlx5_fpga_conn_map_buf() 70 err = dma_mapping_error(dma_device, buf->sg[1].dma_addr); in mlx5_fpga_conn_map_buf() 73 dma_unmap_single(dma_device, buf->sg[0].dma_addr, in mlx5_fpga_conn_map_buf() 89 dma_unmap_single(dma_device, buf->sg[1].dma_addr, in mlx5_fpga_conn_unmap_buf() 93 dma_unmap_single(dma_device, buf->sg[0].dma_addr, in mlx5_fpga_conn_unmap_buf() 117 data->addr = cpu_to_be64(buf->sg[0].dma_addr); in mlx5_fpga_conn_post_recv() 157 data->addr = cpu_to_be64(buf->sg[sgi].dma_addr); in mlx5_fpga_conn_post_send()
|
| /freebsd/sys/dev/qlnx/qlnxe/ |
| H A D | qlnx_os.h | 143 bus_addr_t dma_addr; member
|
| /freebsd/sys/dev/qat/include/common/ |
| H A D | adf_transport_internal.h | 19 bus_addr_t dma_addr; member
|