Home
last modified time | relevance | path

Searched refs:unmap (Results 1 – 25 of 201) sorted by relevance

123456789

/linux/crypto/async_tx/
H A Dasync_pq.c37 struct dmaengine_unmap_data *unmap, in do_async_gen_syndrome() argument
76 dma_dest[0] = unmap->addr[disks - 2]; in do_async_gen_syndrome()
77 dma_dest[1] = unmap->addr[disks - 1]; in do_async_gen_syndrome()
79 &unmap->addr[src_off], in do_async_gen_syndrome()
81 &scfs[src_off], unmap->len, in do_async_gen_syndrome()
89 dma_set_unmap(tx, unmap); in do_async_gen_syndrome()
185 struct dmaengine_unmap_data *unmap = NULL; in async_gen_syndrome() local
190 unmap = dmaengine_get_unmap_data(device->dev, disks, GFP_NOWAIT); in async_gen_syndrome()
193 if (unmap && !(submit->flags & ASYNC_TX_PQ_XOR_DST) && in async_gen_syndrome()
209 unmap->len = len; in async_gen_syndrome()
[all …]
H A Dasync_memcpy.c40 struct dmaengine_unmap_data *unmap = NULL; in async_memcpy() local
43 unmap = dmaengine_get_unmap_data(device->dev, 2, GFP_NOWAIT); in async_memcpy()
45 if (unmap && is_dma_copy_aligned(device, src_offset, dest_offset, len)) { in async_memcpy()
53 unmap->to_cnt = 1; in async_memcpy()
54 unmap->addr[0] = dma_map_page(device->dev, src, src_offset, len, in async_memcpy()
56 unmap->from_cnt = 1; in async_memcpy()
57 unmap->addr[1] = dma_map_page(device->dev, dest, dest_offset, len, in async_memcpy()
59 unmap->len = len; in async_memcpy()
61 tx = device->device_prep_dma_memcpy(chan, unmap->addr[1], in async_memcpy()
62 unmap->addr[0], len, in async_memcpy()
[all …]
H A Dasync_raid6_recov.c25 struct dmaengine_unmap_data *unmap = NULL; in async_sum_product() local
31 unmap = dmaengine_get_unmap_data(dma->dev, 3, GFP_NOWAIT); in async_sum_product()
33 if (unmap) { in async_sum_product()
41 unmap->addr[0] = dma_map_page(dev, srcs[0], src_offs[0], in async_sum_product()
43 unmap->addr[1] = dma_map_page(dev, srcs[1], src_offs[1], in async_sum_product()
45 unmap->to_cnt = 2; in async_sum_product()
47 unmap->addr[2] = dma_map_page(dev, dest, d_off, in async_sum_product()
49 unmap->bidi_cnt = 1; in async_sum_product()
51 pq[1] = unmap->addr[2]; in async_sum_product()
53 unmap->len = len; in async_sum_product()
[all …]
/linux/tools/testing/selftests/mm/
H A Dksm_functional_tests.c103 goto unmap; in __mmap_and_merge_range()
112 goto unmap; in __mmap_and_merge_range()
121 goto unmap; in __mmap_and_merge_range()
124 goto unmap; in __mmap_and_merge_range()
130 goto unmap; in __mmap_and_merge_range()
140 goto unmap; in __mmap_and_merge_range()
149 goto unmap; in __mmap_and_merge_range()
153 unmap: in __mmap_and_merge_range()
188 goto unmap; in test_unmerge()
193 unmap: in test_unmerge()
[all …]
/linux/drivers/xen/xenbus/
H A Dxenbus_client.c80 struct gnttab_unmap_grant_ref unmap[XENBUS_MAX_RING_GRANTS]; member
92 int (*unmap)(struct xenbus_device *dev, void *vaddr); member
593 gnttab_set_unmap_op(&info->unmap[j], in __xenbus_map_ring()
600 BUG_ON(HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, info->unmap, j)); in __xenbus_map_ring()
604 if (info->unmap[i].status != GNTST_okay) { in __xenbus_map_ring()
628 struct gnttab_unmap_grant_ref unmap[XENBUS_MAX_RING_GRANTS]; in xenbus_unmap_ring() local
636 gnttab_set_unmap_op(&unmap[i], vaddrs[i], in xenbus_unmap_ring()
639 BUG_ON(HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, unmap, i)); in xenbus_unmap_ring()
643 if (unmap[i].status != GNTST_okay) { in xenbus_unmap_ring()
644 xenbus_dev_error(dev, unmap[i].status, in xenbus_unmap_ring()
[all …]
/linux/arch/arm/xen/
H A Dp2m.c98 struct gnttab_unmap_grant_ref unmap; in set_foreign_p2m_mapping() local
112 unmap.host_addr = map_ops[i].host_addr; in set_foreign_p2m_mapping()
113 unmap.handle = map_ops[i].handle; in set_foreign_p2m_mapping()
116 unmap.dev_bus_addr = map_ops[i].dev_bus_addr; in set_foreign_p2m_mapping()
118 unmap.dev_bus_addr = 0; in set_foreign_p2m_mapping()
124 unmap.status = 1; in set_foreign_p2m_mapping()
127 &unmap, 1); in set_foreign_p2m_mapping()
128 if (rc || unmap.status != GNTST_okay) in set_foreign_p2m_mapping()
130 rc, unmap.status); in set_foreign_p2m_mapping()
/linux/drivers/iommu/iommufd/
H A Dvfio_compat.c208 struct vfio_iommu_type1_dma_unmap unmap; in iommufd_vfio_unmap_dma() local
213 if (copy_from_user(&unmap, arg, minsz)) in iommufd_vfio_unmap_dma()
216 if (unmap.argsz < minsz || unmap.flags & ~supported_flags) in iommufd_vfio_unmap_dma()
223 if (unmap.flags & VFIO_DMA_UNMAP_FLAG_ALL) { in iommufd_vfio_unmap_dma()
224 if (unmap.iova != 0 || unmap.size != 0) { in iommufd_vfio_unmap_dma()
236 unsigned long iovas[] = { unmap.iova + unmap.size - 1, in iommufd_vfio_unmap_dma()
237 unmap.iova - 1 }; in iommufd_vfio_unmap_dma()
240 unmap.iova ? 2 : 1); in iommufd_vfio_unmap_dma()
244 rc = iopt_unmap_iova(&ioas->iopt, unmap.iova, unmap.size, in iommufd_vfio_unmap_dma()
247 unmap.size = unmapped; in iommufd_vfio_unmap_dma()
[all …]
/linux/mm/
H A Dvma.h174 static inline void unmap_all_init(struct unmap_desc *unmap, in unmap_all_init() argument
177 unmap->mas = &vmi->mas; in unmap_all_init()
178 unmap->first = vma; in unmap_all_init()
179 unmap->pg_start = FIRST_USER_ADDRESS; in unmap_all_init()
180 unmap->pg_end = USER_PGTABLES_CEILING; in unmap_all_init()
181 unmap->vma_start = 0; in unmap_all_init()
182 unmap->vma_end = ULONG_MAX; in unmap_all_init()
183 unmap->tree_end = ULONG_MAX; in unmap_all_init()
184 unmap->tree_reset = vma->vm_end; in unmap_all_init()
185 unmap->mm_wr_locked = false; in unmap_all_init()
[all …]
/linux/arch/um/kernel/
H A Dtlb.c24 int (*unmap)(struct mm_id *mm_idp, member
85 ret = ops->unmap(ops->mm_idp, addr, PAGE_SIZE); in update_pte_range()
105 ret = ops->unmap(ops->mm_idp, addr, in update_pmd_range()
128 ret = ops->unmap(ops->mm_idp, addr, in update_pud_range()
151 ret = ops->unmap(ops->mm_idp, addr, in update_p4d_range()
176 ops.unmap = kern_unmap; in um_tlb_sync()
179 ops.unmap = unmap; in um_tlb_sync()
188 ret = ops.unmap(ops.mm_idp, addr, in um_tlb_sync()
/linux/drivers/net/ethernet/brocade/bna/
H A Dbnad.c94 struct bnad_tx_unmap *unmap; in bnad_tx_buff_unmap() local
98 unmap = &unmap_q[index]; in bnad_tx_buff_unmap()
99 nvecs = unmap->nvecs; in bnad_tx_buff_unmap()
101 skb = unmap->skb; in bnad_tx_buff_unmap()
102 unmap->skb = NULL; in bnad_tx_buff_unmap()
103 unmap->nvecs = 0; in bnad_tx_buff_unmap()
105 dma_unmap_addr(&unmap->vectors[0], dma_addr), in bnad_tx_buff_unmap()
107 dma_unmap_addr_set(&unmap->vectors[0], dma_addr, 0); in bnad_tx_buff_unmap()
116 unmap = &unmap_q[index]; in bnad_tx_buff_unmap()
120 dma_unmap_addr(&unmap->vectors[vector], dma_addr), in bnad_tx_buff_unmap()
[all …]
/linux/drivers/dma/
H A Ddmaengine.c1384 struct dmaengine_unmap_data *unmap = container_of(kref, typeof(*unmap), kref); in dmaengine_unmap() local
1385 struct device *dev = unmap->dev; in dmaengine_unmap()
1388 cnt = unmap->to_cnt; in dmaengine_unmap()
1390 dma_unmap_page(dev, unmap->addr[i], unmap->len, in dmaengine_unmap()
1392 cnt += unmap->from_cnt; in dmaengine_unmap()
1394 dma_unmap_page(dev, unmap->addr[i], unmap->len, in dmaengine_unmap()
1396 cnt += unmap->bidi_cnt; in dmaengine_unmap()
1398 if (unmap->addr[i] == 0) in dmaengine_unmap()
1400 dma_unmap_page(dev, unmap->addr[i], unmap->len, in dmaengine_unmap()
1403 cnt = unmap->map_cnt; in dmaengine_unmap()
[all …]
H A Dmv_xor.c769 struct dmaengine_unmap_data *unmap; in mv_chan_memcpy_self_test() local
792 unmap = dmaengine_get_unmap_data(dma_chan->device->dev, 2, GFP_KERNEL); in mv_chan_memcpy_self_test()
793 if (!unmap) { in mv_chan_memcpy_self_test()
801 unmap->addr[0] = src_dma; in mv_chan_memcpy_self_test()
808 unmap->to_cnt = 1; in mv_chan_memcpy_self_test()
813 unmap->addr[1] = dest_dma; in mv_chan_memcpy_self_test()
820 unmap->from_cnt = 1; in mv_chan_memcpy_self_test()
821 unmap->len = PAGE_SIZE; in mv_chan_memcpy_self_test()
862 dmaengine_unmap_put(unmap); in mv_chan_memcpy_self_test()
880 struct dmaengine_unmap_data *unmap; in mv_chan_xor_self_test() local
[all …]
/linux/arch/x86/xen/
H A Dp2m.c701 struct gnttab_unmap_grant_ref unmap[2]; in set_foreign_p2m_mapping() local
728 unmap[0].host_addr = map_ops[i].host_addr; in set_foreign_p2m_mapping()
729 unmap[0].handle = map_ops[i].handle; in set_foreign_p2m_mapping()
732 unmap[0].dev_bus_addr = map_ops[i].dev_bus_addr; in set_foreign_p2m_mapping()
734 unmap[0].dev_bus_addr = 0; in set_foreign_p2m_mapping()
738 unmap[1].host_addr = kmap_ops[i].host_addr; in set_foreign_p2m_mapping()
739 unmap[1].handle = kmap_ops[i].handle; in set_foreign_p2m_mapping()
742 unmap[1].dev_bus_addr = kmap_ops[i].dev_bus_addr; in set_foreign_p2m_mapping()
744 unmap[1].dev_bus_addr = 0; in set_foreign_p2m_mapping()
751 unmap[0].status = 1; in set_foreign_p2m_mapping()
[all …]
/linux/include/xen/
H A Dgrant_table.h179 gnttab_set_unmap_op(struct gnttab_unmap_grant_ref *unmap, phys_addr_t addr, in gnttab_set_unmap_op() argument
183 unmap->host_addr = addr; in gnttab_set_unmap_op()
185 unmap->host_addr = __pa(addr); in gnttab_set_unmap_op()
187 unmap->host_addr = addr; in gnttab_set_unmap_op()
189 unmap->handle = handle; in gnttab_set_unmap_op()
190 unmap->dev_bus_addr = 0; in gnttab_set_unmap_op()
/linux/drivers/acpi/
H A Dnvs.c80 bool unmap; member
139 if (entry->unmap) { in suspend_nvs_free()
141 entry->unmap = false; in suspend_nvs_free()
185 entry->unmap = !!entry->kaddr; in suspend_nvs_save()
/linux/drivers/ntb/
H A Dntb_transport.c1521 struct dmaengine_unmap_data *unmap; in ntb_async_rx_submit() local
1533 unmap = dmaengine_get_unmap_data(device->dev, 2, GFP_NOWAIT); in ntb_async_rx_submit()
1534 if (!unmap) in ntb_async_rx_submit()
1537 unmap->len = len; in ntb_async_rx_submit()
1538 unmap->addr[0] = dma_map_phys(device->dev, virt_to_phys(offset), in ntb_async_rx_submit()
1540 if (dma_mapping_error(device->dev, unmap->addr[0])) in ntb_async_rx_submit()
1543 unmap->to_cnt = 1; in ntb_async_rx_submit()
1545 unmap->addr[1] = dma_map_phys(device->dev, virt_to_phys(buf), in ntb_async_rx_submit()
1547 if (dma_mapping_error(device->dev, unmap->addr[1])) in ntb_async_rx_submit()
1550 unmap->from_cnt = 1; in ntb_async_rx_submit()
[all …]
/linux/drivers/gpu/drm/imagination/
H A Dpvr_mmu.c1446 } unmap; member
1521 op_ctx->curr_page.l1_table->next_free = op_ctx->unmap.l1_free_tables; in pvr_page_table_l2_remove()
1522 op_ctx->unmap.l1_free_tables = op_ctx->curr_page.l1_table; in pvr_page_table_l2_remove()
1584 op_ctx->curr_page.l0_table->next_free = op_ctx->unmap.l0_free_tables; in pvr_page_table_l1_remove()
1585 op_ctx->unmap.l0_free_tables = op_ctx->curr_page.l0_table; in pvr_page_table_l1_remove()
2315 while (op_ctx->unmap.l0_free_tables) { in pvr_mmu_op_context_destroy()
2316 struct pvr_page_table_l0 *tmp = op_ctx->unmap.l0_free_tables; in pvr_mmu_op_context_destroy()
2318 op_ctx->unmap.l0_free_tables = in pvr_mmu_op_context_destroy()
2319 op_ctx->unmap.l0_free_tables->next_free; in pvr_mmu_op_context_destroy()
2323 while (op_ctx->unmap.l1_free_tables) { in pvr_mmu_op_context_destroy()
[all …]
/linux/drivers/gpu/drm/
H A Ddrm_gpuvm.c2320 struct drm_gpuva *va = op->unmap->va; in drm_gpuva_remap()
2373 struct drm_gpuva_op_unmap *unmap) in op_remap_cb() argument
2382 r->unmap = unmap; in op_remap_cb()
2397 op.unmap.va = va; in op_unmap_cb()
2398 op.unmap.keep = merge; in op_unmap_cb()
2622 struct drm_gpuva_op_unmap unmap = { .va = va }; in __drm_gpuvm_sm_unmap() local
2627 &unmap); in __drm_gpuvm_sm_unmap()
2733 if (op->remap.unmap->va->gem.obj) in drm_gpuva_sm_step_lock()
2734 return drm_exec_lock_obj(exec, op->remap.unmap->va->gem.obj); in drm_gpuva_sm_step_lock()
2737 if (op->unmap.va->gem.obj) in drm_gpuva_sm_step_lock()
[all …]
/linux/include/linux/
H A Ddmaengine.h624 struct dmaengine_unmap_data *unmap; member
636 struct dmaengine_unmap_data *unmap) in dma_set_unmap() argument
638 kref_get(&unmap->kref); in dma_set_unmap()
639 tx->unmap = unmap; in dma_set_unmap()
644 void dmaengine_unmap_put(struct dmaengine_unmap_data *unmap);
647 struct dmaengine_unmap_data *unmap) in dma_set_unmap() argument
655 static inline void dmaengine_unmap_put(struct dmaengine_unmap_data *unmap) in dmaengine_unmap_put() argument
662 if (!tx->unmap) in dma_descriptor_unmap()
665 dmaengine_unmap_put(tx->unmap); in dma_descriptor_unmap()
666 tx->unmap = NULL; in dma_descriptor_unmap()
/linux/drivers/irqchip/
H A Dirq-mchp-eic.c225 goto unmap; in mchp_eic_probe()
231 goto unmap; in mchp_eic_probe()
236 goto unmap; in mchp_eic_probe()
273 unmap: in mchp_eic_probe()
/linux/arch/x86/platform/geode/
H A Dnet5501.c66 goto unmap; in net5501_present()
80 unmap: in net5501_present()
/linux/drivers/xen/
H A Darm-device.c70 goto unmap; in xen_map_device_mmio()
94 goto unmap; in xen_map_device_mmio()
99 unmap: in xen_map_device_mmio()
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/
H A Dvmmgm200.c37 .unmap = gf100_vmm_pgt_unmap,
47 .unmap = gf100_vmm_pgt_unmap,
62 .unmap = gf100_vmm_pgt_unmap,
/linux/drivers/gpu/drm/nouveau/
H A Dnouveau_uvmm.c532 struct drm_gpuva *va = r->unmap->va; in nouveau_uvmm_sm_prepare_unwind()
544 op_unmap_prepare_unwind(op->unmap.va); in nouveau_uvmm_sm_prepare_unwind()
570 struct drm_gpuva *va = r->unmap->va; in nouveau_uvmm_sm_prepare_unwind()
587 struct drm_gpuva_op_unmap *u = &op->unmap; in nouveau_uvmm_sm_prepare_unwind()
712 struct drm_gpuva *va = r->unmap->va; in nouveau_uvmm_sm_prepare()
721 op_unmap_prepare(r->unmap); in nouveau_uvmm_sm_prepare()
752 struct drm_gpuva_op_unmap *u = &op->unmap; in nouveau_uvmm_sm_prepare()
835 return op->remap.unmap->va->gem.obj; in op_gem_obj()
837 return op->unmap.va->gem.obj; in op_gem_obj()
879 struct drm_gpuva_op_unmap *u = r->unmap; in op_remap()
[all …]
/linux/drivers/block/xen-blkback/
H A Dblkback.c245 struct gnttab_unmap_grant_ref unmap[BLKIF_MAX_SEGMENTS_PER_REQUEST]; in free_persistent_gnts() local
256 unmap_data.unmap_ops = unmap; in free_persistent_gnts()
262 gnttab_set_unmap_op(&unmap[segs_to_unmap], in free_persistent_gnts()
292 struct gnttab_unmap_grant_ref unmap[BLKIF_MAX_SEGMENTS_PER_REQUEST]; in xen_blkbk_unmap_purged_grants() local
300 unmap_data.unmap_ops = unmap; in xen_blkbk_unmap_purged_grants()
309 gnttab_set_unmap_op(&unmap[segs_to_unmap], in xen_blkbk_unmap_purged_grants()
711 req->unmap, req->unmap_pages); in xen_blkbk_unmap_and_respond()
715 work->unmap_ops = req->unmap; in xen_blkbk_unmap_and_respond()
735 struct gnttab_unmap_grant_ref unmap[BLKIF_MAX_SEGMENTS_PER_REQUEST]; in xen_blkbk_unmap() local
744 unmap, unmap_pages); in xen_blkbk_unmap()
[all …]

123456789