Home
last modified time | relevance | path

Searched refs:dmap (Results 1 – 22 of 22) sorted by relevance

/linux/fs/fuse/
H A Ddax.c130 struct fuse_dax_mapping *dmap; in alloc_dax_mapping() local
133 dmap = list_first_entry_or_null(&fcd->free_ranges, in alloc_dax_mapping()
135 if (dmap) { in alloc_dax_mapping()
136 list_del_init(&dmap->list); in alloc_dax_mapping()
143 return dmap; in alloc_dax_mapping()
148 struct fuse_dax_mapping *dmap) in __dmap_remove_busy_list() argument
150 list_del_init(&dmap->busy_list); in __dmap_remove_busy_list()
156 struct fuse_dax_mapping *dmap) in dmap_remove_busy_list() argument
159 __dmap_remove_busy_list(fcd, dmap); in dmap_remove_busy_list()
165 struct fuse_dax_mapping *dmap) in __dmap_add_to_free_pool() argument
[all …]
/linux/drivers/pmdomain/ti/
H A Domap_prm.c61 const struct omap_prm_domain_map *dmap; member
161 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_reton,
165 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_noinact,
170 .pwrstctrl = 0, .pwrstst = 0x4, .dmap = &omap_prm_all,
174 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_alwon,
178 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_reton,
185 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_noinact,
190 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_onoff_noauto,
194 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_noinact
198 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_onoff_noauto
[all …]
/linux/drivers/md/
H A Ddm-clone-metadata.c127 struct dirty_map dmap[2]; member
468 static int __dirty_map_init(struct dirty_map *dmap, unsigned long nr_words, in __dirty_map_init() argument
471 dmap->changed = 0; in __dirty_map_init()
473 dmap->dirty_words = kvzalloc(bitmap_size(nr_words), GFP_KERNEL); in __dirty_map_init()
474 if (!dmap->dirty_words) in __dirty_map_init()
477 dmap->dirty_regions = kvzalloc(bitmap_size(nr_regions), GFP_KERNEL); in __dirty_map_init()
478 if (!dmap->dirty_regions) { in __dirty_map_init()
479 kvfree(dmap->dirty_words); in __dirty_map_init()
486 static void __dirty_map_exit(struct dirty_map *dmap) in __dirty_map_exit() argument
488 kvfree(dmap->dirty_words); in __dirty_map_exit()
[all …]
H A Ddm-zoned-metadata.c1683 struct dmz_map *dmap; in dmz_load_mapping() local
1701 dmap = dmap_mblk->data; in dmz_load_mapping()
1707 dzone_id = le32_to_cpu(dmap[e].dzone_id); in dmz_load_mapping()
1735 bzone_id = le32_to_cpu(dmap[e].bzone_id); in dmz_load_mapping()
1832 struct dmz_map *dmap = dmap_mblk->data; in dmz_set_chunk_mapping() local
1835 dmap[map_idx].dzone_id = cpu_to_le32(dzone_id); in dmz_set_chunk_mapping()
1836 dmap[map_idx].bzone_id = cpu_to_le32(bzone_id); in dmz_set_chunk_mapping()
2045 struct dmz_map *dmap = dmap_mblk->data; in dmz_get_chunk_mapping() local
2055 dzone_id = le32_to_cpu(dmap[dmap_idx].dzone_id); in dmz_get_chunk_mapping()
/linux/drivers/media/pci/intel/ipu6/
H A Dipu6-mmu.c424 iova = alloc_iova(&mmu->dmap->iovad, n_pages, in allocate_trash_buffer()
425 PHYS_PFN(mmu->dmap->mmu_info->aperture_end), 0); in allocate_trash_buffer()
431 dma = dma_map_page(mmu->dmap->mmu_info->dev, mmu->trash_page, 0, in allocate_trash_buffer()
433 if (dma_mapping_error(mmu->dmap->mmu_info->dev, dma)) { in allocate_trash_buffer()
434 dev_err(mmu->dmap->mmu_info->dev, "Failed to map trash page\n"); in allocate_trash_buffer()
447 ret = ipu6_mmu_map(mmu->dmap->mmu_info, PFN_PHYS(iova_addr), in allocate_trash_buffer()
464 ipu6_mmu_unmap(mmu->dmap->mmu_info, PFN_PHYS(iova->pfn_lo), in allocate_trash_buffer()
466 dma_unmap_page(mmu->dmap->mmu_info->dev, mmu->pci_trash_page, in allocate_trash_buffer()
469 __free_iova(&mmu->dmap->iovad, iova); in allocate_trash_buffer()
479 mmu_info = mmu->dmap->mmu_info; in ipu6_mmu_hw_init()
[all …]
H A Dipu6-dma.c174 iova = alloc_iova(&mmu->dmap->iovad, count, in ipu6_dma_alloc()
175 PHYS_PFN(mmu->dmap->mmu_info->aperture_end), 0); in ipu6_dma_alloc()
196 ret = ipu6_mmu_map(mmu->dmap->mmu_info, in ipu6_dma_alloc()
225 pci_dma_addr = ipu6_mmu_iova_to_phys(mmu->dmap->mmu_info, in ipu6_dma_alloc()
230 ipu6_mmu_unmap(mmu->dmap->mmu_info, ipu6_iova, PAGE_SIZE); in ipu6_dma_alloc()
236 __free_iova(&mmu->dmap->iovad, iova); in ipu6_dma_alloc()
249 struct iova *iova = find_iova(&mmu->dmap->iovad, PHYS_PFN(dma_handle)); in ipu6_dma_free()
278 pci_dma_addr = ipu6_mmu_iova_to_phys(mmu->dmap->mmu_info, in ipu6_dma_free()
284 ipu6_mmu_unmap(mmu->dmap->mmu_info, PFN_PHYS(iova->pfn_lo), in ipu6_dma_free()
291 __free_iova(&mmu->dmap->iovad, iova); in ipu6_dma_free()
[all …]
/linux/drivers/staging/media/ipu7/
H A Dipu7-mmu.c407 iova = alloc_iova(&mmu->dmap->iovad, n_pages, in allocate_trash_buffer()
408 PHYS_PFN(mmu->dmap->mmu_info->aperture_end), 0); in allocate_trash_buffer()
414 dma = dma_map_page(mmu->dmap->mmu_info->dev, mmu->trash_page, 0, in allocate_trash_buffer()
416 if (dma_mapping_error(mmu->dmap->mmu_info->dev, dma)) { in allocate_trash_buffer()
417 dev_err(mmu->dmap->mmu_info->dev, "Failed to map trash page\n"); in allocate_trash_buffer()
430 ret = ipu7_mmu_map(mmu->dmap->mmu_info, PFN_PHYS(iova_addr), in allocate_trash_buffer()
447 ipu7_mmu_unmap(mmu->dmap->mmu_info, PFN_PHYS(iova->pfn_lo), in allocate_trash_buffer()
449 dma_unmap_page(mmu->dmap->mmu_info->dev, mmu->pci_trash_page, in allocate_trash_buffer()
452 __free_iova(&mmu->dmap->iovad, iova); in allocate_trash_buffer()
461 mmu_info = mmu->dmap->mmu_info; in __mmu_at_init()
[all …]
H A Dipu7-dma.c174 iova = alloc_iova(&mmu->dmap->iovad, count, in ipu7_dma_alloc()
175 PHYS_PFN(mmu->dmap->mmu_info->aperture_end), 0); in ipu7_dma_alloc()
196 ret = ipu7_mmu_map(mmu->dmap->mmu_info, in ipu7_dma_alloc()
225 pci_dma_addr = ipu7_mmu_iova_to_phys(mmu->dmap->mmu_info, in ipu7_dma_alloc()
230 ipu7_mmu_unmap(mmu->dmap->mmu_info, ipu7_iova, PAGE_SIZE); in ipu7_dma_alloc()
236 __free_iova(&mmu->dmap->iovad, iova); in ipu7_dma_alloc()
249 struct iova *iova = find_iova(&mmu->dmap->iovad, PHYS_PFN(dma_handle)); in ipu7_dma_free()
278 pci_dma_addr = ipu7_mmu_iova_to_phys(mmu->dmap->mmu_info, in ipu7_dma_free()
284 ipu7_mmu_unmap(mmu->dmap->mmu_info, PFN_PHYS(iova->pfn_lo), in ipu7_dma_free()
291 __free_iova(&mmu->dmap->iovad, iova); in ipu7_dma_free()
[all …]
/linux/sound/soc/tegra/
H A Dtegra_pcm.c80 struct snd_dmaengine_dai_dma_data *dmap; in tegra_pcm_open() local
88 dmap = snd_soc_dai_get_dma_data(cpu_dai, substream); in tegra_pcm_open()
101 chan = dma_request_chan(cpu_dai->dev, dmap->chan_name); in tegra_pcm_open()
105 dmap->chan_name); in tegra_pcm_open()
113 dmap->chan_name); in tegra_pcm_open()
146 struct snd_dmaengine_dai_dma_data *dmap; in tegra_pcm_hw_params() local
154 dmap = snd_soc_dai_get_dma_data(snd_soc_rtd_to_cpu(rtd, 0), substream); in tegra_pcm_hw_params()
155 if (!dmap) in tegra_pcm_hw_params()
169 slave_config.dst_addr = dmap->addr; in tegra_pcm_hw_params()
173 slave_config.src_addr = dmap->addr; in tegra_pcm_hw_params()
/linux/fs/nilfs2/
H A Dpage.c243 int nilfs_copy_dirty_pages(struct address_space *dmap, in nilfs_copy_dirty_pages() argument
264 dfolio = filemap_grab_folio(dmap, folio->index); in nilfs_copy_dirty_pages()
298 void nilfs_copy_back_pages(struct address_space *dmap, in nilfs_copy_back_pages() argument
316 dfolio = filemap_lock_folio(dmap, index); in nilfs_copy_back_pages()
334 xa_lock_irq(&dmap->i_pages); in nilfs_copy_back_pages()
335 f = __xa_store(&dmap->i_pages, index, folio, GFP_NOFS); in nilfs_copy_back_pages()
341 folio->mapping = dmap; in nilfs_copy_back_pages()
342 dmap->nrpages++; in nilfs_copy_back_pages()
344 __xa_set_mark(&dmap->i_pages, index, in nilfs_copy_back_pages()
347 xa_unlock_irq(&dmap->i_pages); in nilfs_copy_back_pages()
/linux/fs/jfs/
H A Djfs_dmap.c64 static void dbAllocBits(struct bmap * bmp, struct dmap * dp, s64 blkno,
73 static int dbAllocNext(struct bmap * bmp, struct dmap * dp, s64 blkno,
75 static int dbAllocNear(struct bmap * bmp, struct dmap * dp, s64 blkno,
78 static int dbAllocDmap(struct bmap * bmp, struct dmap * dp, s64 blkno,
80 static int dbAllocDmapLev(struct bmap * bmp, struct dmap * dp, int nblocks,
91 static int dbFreeBits(struct bmap * bmp, struct dmap * dp, s64 blkno,
93 static int dbFreeDmap(struct bmap * bmp, struct dmap * dp, s64 blkno,
101 static int dbAllocDmapBU(struct bmap * bmp, struct dmap * dp, s64 blkno,
103 static int dbInitDmap(struct dmap * dp, s64 blkno, int nblocks);
104 static int dbInitDmapTree(struct dmap * dp);
[all …]
H A Djfs_dmap.h147 struct dmap { struct
/linux/include/drm/ttm/
H A Dttm_kmap_iter.h32 struct iosys_map *dmap, pgoff_t i);
42 struct iosys_map *dmap);
H A Dttm_resource.h387 struct iosys_map dmap; member
/linux/drivers/misc/sgi-gru/
H A Dgrufault.c143 struct gru_tlb_fault_map *dmap) in get_clear_fault_map() argument
158 dmap->fault_bits[i] = k; in get_clear_fault_map()
522 struct gru_tlb_fault_map imap, dmap; in gru_intr() local
536 get_clear_fault_map(gru, &imap, &dmap); in gru_intr()
541 dmap.fault_bits[0], dmap.fault_bits[1]); in gru_intr()
543 for_each_cbr_in_tfm(cbrnum, dmap.fault_bits) { in gru_intr()
/linux/drivers/iommu/
H A Dtegra-smmu.c586 dma_addr_t *dmap) in tegra_smmu_pte_lookup() argument
596 *dmap = smmu_pde_to_dma(smmu, as->pd->val[pd_index]); in tegra_smmu_pte_lookup()
602 dma_addr_t *dmap, struct tegra_pt *pt) in as_get_pte() argument
629 *dmap = dma; in as_get_pte()
631 *dmap = smmu_pde_to_dma(smmu, as->pd->val[pde]); in as_get_pte()
/linux/drivers/gpu/drm/ttm/
H A Dttm_tt.c488 struct iosys_map *dmap, in ttm_kmap_iter_tt_map_local() argument
494 iosys_map_set_vaddr(dmap, kmap_local_page_prot(iter_tt->tt->pages[i], in ttm_kmap_iter_tt_map_local()
/linux/drivers/net/ethernet/apple/
H A Dbmac.c177 dbdma_continue(volatile struct dbdma_regs __iomem *dmap) in dbdma_continue() argument
179 dbdma_st32(&dmap->control, in dbdma_continue()
185 dbdma_reset(volatile struct dbdma_regs __iomem *dmap) in dbdma_reset() argument
187 dbdma_st32(&dmap->control, in dbdma_reset()
190 while (dbdma_ld32(&dmap->status) & RUN) in dbdma_reset()
/linux/fs/
H A Ddax.c2160 const struct iomap *dmap = &it_dest->iomap; in dax_range_compare_iter() local
2165 len = min(len, min(smap->length, dmap->length)); in dax_range_compare_iter()
2167 if (smap->type == IOMAP_HOLE && dmap->type == IOMAP_HOLE) { in dax_range_compare_iter()
2172 if (smap->type == IOMAP_HOLE || dmap->type == IOMAP_HOLE) { in dax_range_compare_iter()
2183 ret = dax_iomap_direct_access(dmap, pos2, ALIGN(pos2 + len, PAGE_SIZE), in dax_range_compare_iter()
/linux/drivers/android/
H A Dbinder.c1069 struct dbitmap *dmap = &proc->dmap; in get_ref_desc_olocked() local
1076 if (!dbitmap_enabled(dmap)) { in get_ref_desc_olocked()
1081 if (dbitmap_acquire_next_zero_bit(dmap, offset, &bit) == 0) { in get_ref_desc_olocked()
1090 nbits = dbitmap_grow_nbits(dmap); in get_ref_desc_olocked()
1094 dbitmap_grow(dmap, new, nbits); in get_ref_desc_olocked()
1184 struct dbitmap *dmap = &ref->proc->dmap; in binder_cleanup_ref_olocked() local
1192 if (dbitmap_enabled(dmap)) in binder_cleanup_ref_olocked()
1193 dbitmap_clear_bit(dmap, ref->data.desc); in binder_cleanup_ref_olocked()
1368 nbits = dbitmap_shrink_nbits(&proc->dmap); in try_shrink_dmap()
1376 dbitmap_shrink(&proc->dmap, new, nbits); in try_shrink_dmap()
[all …]
/linux/fs/ocfs2/dlm/
H A Ddlmdomain.c49 static inline void byte_copymap(u8 dmap[], unsigned long smap[], in byte_copymap() argument
57 memset(dmap, 0, ((sz + 7) >> 3)); in byte_copymap()
60 byte_set_bit(nn, dmap); in byte_copymap()
/linux/drivers/gpu/drm/vmwgfx/device_include/
H A Dsvga3d_types.h604 uint32 dmap : 1; member