Home
last modified time | relevance | path

Searched refs:dma_list (Results 1 – 16 of 16) sorted by relevance

/linux/drivers/crypto/intel/keembay/
H A Docs-hcu.c444 const struct ocs_hcu_dma_list *dma_list, in ocs_hcu_ll_dma_start() argument
450 if (!dma_list) in ocs_hcu_ll_dma_start()
471 writel(dma_list->dma_addr, hcu_dev->io_base + OCS_HCU_DMA_NEXT_SRC_DESCR); in ocs_hcu_ll_dma_start()
492 struct ocs_hcu_dma_list *dma_list; in ocs_hcu_dma_list_alloc() local
494 dma_list = kmalloc_obj(*dma_list); in ocs_hcu_dma_list_alloc()
495 if (!dma_list) in ocs_hcu_dma_list_alloc()
499 dma_list->head = dma_alloc_coherent(hcu_dev->dev, in ocs_hcu_dma_list_alloc()
500 sizeof(*dma_list->head) * max_nents, in ocs_hcu_dma_list_alloc()
501 &dma_list->dma_addr, GFP_KERNEL); in ocs_hcu_dma_list_alloc()
502 if (!dma_list->head) { in ocs_hcu_dma_list_alloc()
[all …]
H A Docs-hcu.h77 struct ocs_hcu_dma_list *dma_list);
80 struct ocs_hcu_dma_list *dma_list,
87 const struct ocs_hcu_dma_list *dma_list);
91 const struct ocs_hcu_dma_list *dma_list,
103 const struct ocs_hcu_dma_list *dma_list,
/linux/drivers/infiniband/hw/mthca/
H A Dmthca_allocator.c195 u64 *dma_list = NULL; in mthca_buf_alloc() local
216 dma_list = kmalloc_array(npages, sizeof(*dma_list), in mthca_buf_alloc()
218 if (!dma_list) in mthca_buf_alloc()
222 dma_list[i] = t + i * (1 << shift); in mthca_buf_alloc()
228 dma_list = kmalloc_array(npages, sizeof(*dma_list), in mthca_buf_alloc()
230 if (!dma_list) in mthca_buf_alloc()
247 dma_list[i] = t; in mthca_buf_alloc()
255 dma_list, shift, npages, in mthca_buf_alloc()
263 kfree(dma_list); in mthca_buf_alloc()
271 kfree(dma_list); in mthca_buf_alloc()
H A Dmthca_eq.c471 u64 *dma_list = NULL; in mthca_create_eq() local
489 dma_list = kmalloc_array(npages, sizeof(*dma_list), GFP_KERNEL); in mthca_create_eq()
490 if (!dma_list) in mthca_create_eq()
504 dma_list[i] = t; in mthca_create_eq()
518 dma_list, PAGE_SHIFT, npages, in mthca_create_eq()
550 kfree(dma_list); in mthca_create_eq()
581 kfree(dma_list); in mthca_create_eq()
/linux/drivers/misc/genwqe/
H A Dcard_utils.c229 static void genwqe_unmap_pages(struct genwqe_dev *cd, dma_addr_t *dma_list, in genwqe_unmap_pages() argument
235 for (i = 0; (i < num_pages) && (dma_list[i] != 0x0); i++) { in genwqe_unmap_pages()
236 dma_unmap_page(&pci_dev->dev, dma_list[i], PAGE_SIZE, in genwqe_unmap_pages()
238 dma_list[i] = 0x0; in genwqe_unmap_pages()
244 dma_addr_t *dma_list) in genwqe_map_pages() argument
253 dma_list[i] = 0x0; in genwqe_map_pages()
266 dma_list[i] = daddr; in genwqe_map_pages()
271 genwqe_unmap_pages(cd, dma_list, num_pages); in genwqe_map_pages()
375 dma_addr_t *dma_list) in genwqe_setup_sgl() argument
410 daddr = dma_list[p] + map_offs; in genwqe_setup_sgl()
[all …]
H A Dcard_base.h171 dma_addr_t *dma_list; /* list of dma addresses per page */ member
372 dma_addr_t *dma_list);
H A Dcard_dev.c965 &m->dma_list[page_offs]); in ddcb_cmd_fixups()
/linux/arch/powerpc/platforms/cell/spufs/
H A Dspu_utils.h38 struct dma_list_elem dma_list[15] __attribute__ ((aligned(8))); variable
92 dma_list[i].size = 16384; in build_dma_list()
93 dma_list[i].ea_low = ea_low; in build_dma_list()
H A Dspu_save.c50 unsigned int list = (unsigned int)&dma_list[0]; in save_upper_240kb()
51 unsigned int size = sizeof(dma_list); in save_upper_240kb()
H A Dspu_restore.c50 unsigned int list = (unsigned int)&dma_list[0]; in restore_upper_240kb()
51 unsigned int size = sizeof(dma_list); in restore_upper_240kb()
/linux/mm/
H A Dhmm.c726 map->dma_list = kvzalloc_objs(*map->dma_list, nr_entries, in hmm_dma_map_alloc()
728 if (!map->dma_list) in hmm_dma_map_alloc()
751 kvfree(map->dma_list); in hmm_dma_map_free()
776 dma_addr_t *dma_addrs = map->dma_list; in hmm_dma_map_pfn()
872 dma_addr_t *dma_addrs = map->dma_list; in hmm_dma_unmap_pfn()
/linux/include/linux/
H A Dhmm-dma.h22 dma_addr_t *dma_list; member
/linux/drivers/staging/vme_user/
H A Dvme.c890 struct vme_dma_list *dma_list; in vme_new_dma_list() local
897 dma_list = kmalloc_obj(*dma_list); in vme_new_dma_list()
898 if (!dma_list) in vme_new_dma_list()
901 INIT_LIST_HEAD(&dma_list->entries); in vme_new_dma_list()
902 dma_list->parent = list_entry(resource->entry, in vme_new_dma_list()
905 mutex_init(&dma_list->mtx); in vme_new_dma_list()
907 return dma_list; in vme_new_dma_list()
/linux/drivers/vfio/
H A Dvfio_iommu_type1.c69 struct rb_root dma_list; member
169 struct rb_node *node = iommu->dma_list.rb_node; in vfio_find_dma()
192 struct rb_node *node = iommu->dma_list.rb_node; in vfio_find_dma_first_node()
217 struct rb_node **link = &iommu->dma_list.rb_node, *parent = NULL; in vfio_link_dma()
233 rb_insert_color(&new->node, &iommu->dma_list); in vfio_link_dma()
238 rb_erase(&old->node, &iommu->dma_list); in vfio_unlink_dma()
285 for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) { in vfio_iommu_populate_bitmap_full()
296 for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) { in vfio_dma_bitmap_alloc_all()
321 for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) { in vfio_dma_bitmap_free_all()
1330 for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) { in vfio_iova_dirty_bitmap()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx4/
H A Deq.c977 u64 *dma_list = NULL; in mlx4_create_eq() local
997 dma_list = kmalloc_array(npages, sizeof(*dma_list), GFP_KERNEL); in mlx4_create_eq()
998 if (!dma_list) in mlx4_create_eq()
1014 dma_list[i] = t; in mlx4_create_eq()
1032 err = mlx4_write_mtt(dev, &eq->mtt, 0, npages, dma_list); in mlx4_create_eq()
1052 kfree(dma_list); in mlx4_create_eq()
1081 kfree(dma_list); in mlx4_create_eq()
/linux/drivers/pci/endpoint/functions/
H A Dpci-epf-mhi.c155 struct list_head dma_list; member
461 list_splice_tail_init(&epf_mhi->dma_list, &head); in pci_epf_mhi_dma_worker()
479 list_add_tail(&transfer->node, &epf_mhi->dma_list); in pci_epf_mhi_dma_async_callback()
695 INIT_LIST_HEAD(&epf_mhi->dma_list); in pci_epf_mhi_dma_init()