| /linux/drivers/crypto/intel/keembay/ |
| H A D | ocs-hcu.c | 444 const struct ocs_hcu_dma_list *dma_list, in ocs_hcu_ll_dma_start() argument 450 if (!dma_list) in ocs_hcu_ll_dma_start() 471 writel(dma_list->dma_addr, hcu_dev->io_base + OCS_HCU_DMA_NEXT_SRC_DESCR); in ocs_hcu_ll_dma_start() 492 struct ocs_hcu_dma_list *dma_list; in ocs_hcu_dma_list_alloc() local 494 dma_list = kmalloc_obj(*dma_list); in ocs_hcu_dma_list_alloc() 495 if (!dma_list) in ocs_hcu_dma_list_alloc() 499 dma_list->head = dma_alloc_coherent(hcu_dev->dev, in ocs_hcu_dma_list_alloc() 500 sizeof(*dma_list->head) * max_nents, in ocs_hcu_dma_list_alloc() 501 &dma_list->dma_addr, GFP_KERNEL); in ocs_hcu_dma_list_alloc() 502 if (!dma_list->head) { in ocs_hcu_dma_list_alloc() [all …]
|
| H A D | ocs-hcu.h | 77 struct ocs_hcu_dma_list *dma_list); 80 struct ocs_hcu_dma_list *dma_list, 87 const struct ocs_hcu_dma_list *dma_list); 91 const struct ocs_hcu_dma_list *dma_list, 103 const struct ocs_hcu_dma_list *dma_list,
|
| /linux/drivers/infiniband/hw/mthca/ |
| H A D | mthca_allocator.c | 195 u64 *dma_list = NULL; in mthca_buf_alloc() local 216 dma_list = kmalloc_array(npages, sizeof(*dma_list), in mthca_buf_alloc() 218 if (!dma_list) in mthca_buf_alloc() 222 dma_list[i] = t + i * (1 << shift); in mthca_buf_alloc() 228 dma_list = kmalloc_array(npages, sizeof(*dma_list), in mthca_buf_alloc() 230 if (!dma_list) in mthca_buf_alloc() 247 dma_list[i] = t; in mthca_buf_alloc() 255 dma_list, shift, npages, in mthca_buf_alloc() 263 kfree(dma_list); in mthca_buf_alloc() 271 kfree(dma_list); in mthca_buf_alloc()
|
| H A D | mthca_eq.c | 471 u64 *dma_list = NULL; in mthca_create_eq() local 489 dma_list = kmalloc_array(npages, sizeof(*dma_list), GFP_KERNEL); in mthca_create_eq() 490 if (!dma_list) in mthca_create_eq() 504 dma_list[i] = t; in mthca_create_eq() 518 dma_list, PAGE_SHIFT, npages, in mthca_create_eq() 550 kfree(dma_list); in mthca_create_eq() 581 kfree(dma_list); in mthca_create_eq()
|
| /linux/drivers/misc/genwqe/ |
| H A D | card_utils.c | 229 static void genwqe_unmap_pages(struct genwqe_dev *cd, dma_addr_t *dma_list, in genwqe_unmap_pages() argument 235 for (i = 0; (i < num_pages) && (dma_list[i] != 0x0); i++) { in genwqe_unmap_pages() 236 dma_unmap_page(&pci_dev->dev, dma_list[i], PAGE_SIZE, in genwqe_unmap_pages() 238 dma_list[i] = 0x0; in genwqe_unmap_pages() 244 dma_addr_t *dma_list) in genwqe_map_pages() argument 253 dma_list[i] = 0x0; in genwqe_map_pages() 266 dma_list[i] = daddr; in genwqe_map_pages() 271 genwqe_unmap_pages(cd, dma_list, num_pages); in genwqe_map_pages() 375 dma_addr_t *dma_list) in genwqe_setup_sgl() argument 410 daddr = dma_list[p] + map_offs; in genwqe_setup_sgl() [all …]
|
| H A D | card_base.h | 171 dma_addr_t *dma_list; /* list of dma addresses per page */ member 372 dma_addr_t *dma_list);
|
| H A D | card_dev.c | 965 &m->dma_list[page_offs]); in ddcb_cmd_fixups()
|
| /linux/arch/powerpc/platforms/cell/spufs/ |
| H A D | spu_utils.h | 38 struct dma_list_elem dma_list[15] __attribute__ ((aligned(8))); variable 92 dma_list[i].size = 16384; in build_dma_list() 93 dma_list[i].ea_low = ea_low; in build_dma_list()
|
| H A D | spu_save.c | 50 unsigned int list = (unsigned int)&dma_list[0]; in save_upper_240kb() 51 unsigned int size = sizeof(dma_list); in save_upper_240kb()
|
| H A D | spu_restore.c | 50 unsigned int list = (unsigned int)&dma_list[0]; in restore_upper_240kb() 51 unsigned int size = sizeof(dma_list); in restore_upper_240kb()
|
| /linux/mm/ |
| H A D | hmm.c | 726 map->dma_list = kvzalloc_objs(*map->dma_list, nr_entries, in hmm_dma_map_alloc() 728 if (!map->dma_list) in hmm_dma_map_alloc() 751 kvfree(map->dma_list); in hmm_dma_map_free() 776 dma_addr_t *dma_addrs = map->dma_list; in hmm_dma_map_pfn() 872 dma_addr_t *dma_addrs = map->dma_list; in hmm_dma_unmap_pfn()
|
| /linux/include/linux/ |
| H A D | hmm-dma.h | 22 dma_addr_t *dma_list; member
|
| /linux/drivers/staging/vme_user/ |
| H A D | vme.c | 890 struct vme_dma_list *dma_list; in vme_new_dma_list() local 897 dma_list = kmalloc_obj(*dma_list); in vme_new_dma_list() 898 if (!dma_list) in vme_new_dma_list() 901 INIT_LIST_HEAD(&dma_list->entries); in vme_new_dma_list() 902 dma_list->parent = list_entry(resource->entry, in vme_new_dma_list() 905 mutex_init(&dma_list->mtx); in vme_new_dma_list() 907 return dma_list; in vme_new_dma_list()
|
| /linux/drivers/vfio/ |
| H A D | vfio_iommu_type1.c | 69 struct rb_root dma_list; member 169 struct rb_node *node = iommu->dma_list.rb_node; in vfio_find_dma() 192 struct rb_node *node = iommu->dma_list.rb_node; in vfio_find_dma_first_node() 217 struct rb_node **link = &iommu->dma_list.rb_node, *parent = NULL; in vfio_link_dma() 233 rb_insert_color(&new->node, &iommu->dma_list); in vfio_link_dma() 238 rb_erase(&old->node, &iommu->dma_list); in vfio_unlink_dma() 285 for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) { in vfio_iommu_populate_bitmap_full() 296 for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) { in vfio_dma_bitmap_alloc_all() 321 for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) { in vfio_dma_bitmap_free_all() 1330 for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) { in vfio_iova_dirty_bitmap() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| H A D | eq.c | 977 u64 *dma_list = NULL; in mlx4_create_eq() local 997 dma_list = kmalloc_array(npages, sizeof(*dma_list), GFP_KERNEL); in mlx4_create_eq() 998 if (!dma_list) in mlx4_create_eq() 1014 dma_list[i] = t; in mlx4_create_eq() 1032 err = mlx4_write_mtt(dev, &eq->mtt, 0, npages, dma_list); in mlx4_create_eq() 1052 kfree(dma_list); in mlx4_create_eq() 1081 kfree(dma_list); in mlx4_create_eq()
|
| /linux/drivers/pci/endpoint/functions/ |
| H A D | pci-epf-mhi.c | 155 struct list_head dma_list; member 461 list_splice_tail_init(&epf_mhi->dma_list, &head); in pci_epf_mhi_dma_worker() 479 list_add_tail(&transfer->node, &epf_mhi->dma_list); in pci_epf_mhi_dma_async_callback() 695 INIT_LIST_HEAD(&epf_mhi->dma_list); in pci_epf_mhi_dma_init()
|