Home
last modified time | relevance | path

Searched refs:iommu (Results 1 – 25 of 191) sorted by relevance

12345678

/linux/drivers/iommu/
H A Dsun50i-iommu.c101 struct iommu_device iommu; member
125 struct sun50i_iommu *iommu; member
138 static u32 iommu_read(struct sun50i_iommu *iommu, u32 offset) in iommu_read() argument
140 return readl(iommu->base + offset); in iommu_read()
143 static void iommu_write(struct sun50i_iommu *iommu, u32 offset, u32 value) in iommu_write() argument
145 writel(value, iommu->base + offset); in iommu_write()
294 struct sun50i_iommu *iommu = sun50i_domain->iommu; in sun50i_table_flush() local
298 dma_sync_single_for_device(iommu->dev, dma, size, DMA_TO_DEVICE); in sun50i_table_flush()
301 static void sun50i_iommu_zap_iova(struct sun50i_iommu *iommu, in sun50i_iommu_zap_iova() argument
307 iommu_write(iommu, IOMMU_TLB_IVLD_ADDR_REG, iova); in sun50i_iommu_zap_iova()
[all …]
H A Drockchip-iommu.c117 struct iommu_device iommu; member
124 struct rk_iommu *iommu; member
347 static void rk_iommu_command(struct rk_iommu *iommu, u32 command) in rk_iommu_command() argument
351 for (i = 0; i < iommu->num_mmu; i++) in rk_iommu_command()
352 writel(command, iommu->bases[i] + RK_MMU_COMMAND); in rk_iommu_command()
359 static void rk_iommu_zap_lines(struct rk_iommu *iommu, dma_addr_t iova_start, in rk_iommu_zap_lines() argument
368 for (i = 0; i < iommu->num_mmu; i++) { in rk_iommu_zap_lines()
372 rk_iommu_write(iommu->bases[i], RK_MMU_ZAP_ONE_LINE, iova); in rk_iommu_zap_lines()
376 static bool rk_iommu_is_stall_active(struct rk_iommu *iommu) in rk_iommu_is_stall_active() argument
381 for (i = 0; i < iommu->num_mmu; i++) in rk_iommu_is_stall_active()
[all …]
H A Dmsm_iommu.c54 static int __enable_clocks(struct msm_iommu_dev *iommu) in __enable_clocks() argument
58 ret = clk_enable(iommu->pclk); in __enable_clocks()
62 if (iommu->clk) { in __enable_clocks()
63 ret = clk_enable(iommu->clk); in __enable_clocks()
65 clk_disable(iommu->pclk); in __enable_clocks()
71 static void __disable_clocks(struct msm_iommu_dev *iommu) in __disable_clocks() argument
73 if (iommu->clk) in __disable_clocks()
74 clk_disable(iommu->clk); in __disable_clocks()
75 clk_disable(iommu->pclk); in __disable_clocks()
120 struct msm_iommu_dev *iommu = NULL; in __flush_iotlb() local
[all …]
H A Diommu-sysfs.c54 int iommu_device_sysfs_add(struct iommu_device *iommu, in iommu_device_sysfs_add() argument
62 iommu->dev = kzalloc_obj(*iommu->dev); in iommu_device_sysfs_add()
63 if (!iommu->dev) in iommu_device_sysfs_add()
66 device_initialize(iommu->dev); in iommu_device_sysfs_add()
68 iommu->dev->class = &iommu_class; in iommu_device_sysfs_add()
69 iommu->dev->parent = parent; in iommu_device_sysfs_add()
70 iommu->dev->groups = groups; in iommu_device_sysfs_add()
73 ret = kobject_set_name_vargs(&iommu->dev->kobj, fmt, vargs); in iommu_device_sysfs_add()
78 ret = device_add(iommu->dev); in iommu_device_sysfs_add()
82 dev_set_drvdata(iommu->dev, iommu); in iommu_device_sysfs_add()
[all …]
/linux/drivers/iommu/amd/
H A Dinit.c235 bool translation_pre_enabled(struct amd_iommu *iommu) in translation_pre_enabled() argument
237 return (iommu->flags & AMD_IOMMU_FLAG_TRANS_PRE_ENABLED); in translation_pre_enabled()
240 static void clear_translation_pre_enabled(struct amd_iommu *iommu) in clear_translation_pre_enabled() argument
242 iommu->flags &= ~AMD_IOMMU_FLAG_TRANS_PRE_ENABLED; in clear_translation_pre_enabled()
245 static void init_translation_status(struct amd_iommu *iommu) in init_translation_status() argument
249 ctrl = readq(iommu->mmio_base + MMIO_CONTROL_OFFSET); in init_translation_status()
251 iommu->flags |= AMD_IOMMU_FLAG_TRANS_PRE_ENABLED; in init_translation_status()
270 struct amd_iommu *iommu; in get_global_efr() local
272 for_each_iommu(iommu) { in get_global_efr()
273 u64 tmp = iommu->features; in get_global_efr()
[all …]
H A Dppr.c20 int __init amd_iommu_alloc_ppr_log(struct amd_iommu *iommu) in amd_iommu_alloc_ppr_log() argument
22 iommu->ppr_log = iommu_alloc_4k_pages(iommu, GFP_KERNEL | __GFP_ZERO, in amd_iommu_alloc_ppr_log()
24 return iommu->ppr_log ? 0 : -ENOMEM; in amd_iommu_alloc_ppr_log()
27 void amd_iommu_enable_ppr_log(struct amd_iommu *iommu) in amd_iommu_enable_ppr_log() argument
31 if (iommu->ppr_log == NULL) in amd_iommu_enable_ppr_log()
34 iommu_feature_enable(iommu, CONTROL_PPR_EN); in amd_iommu_enable_ppr_log()
36 entry = iommu_virt_to_phys(iommu->ppr_log); in amd_iommu_enable_ppr_log()
40 memcpy_toio(iommu->mmio_base + MMIO_PPR_LOG_OFFSET, in amd_iommu_enable_ppr_log()
44 writel(0x00, iommu->mmio_base + MMIO_PPR_HEAD_OFFSET); in amd_iommu_enable_ppr_log()
45 writel(0x00, iommu->mmio_base + MMIO_PPR_TAIL_OFFSET); in amd_iommu_enable_ppr_log()
[all …]
H A Diommu.c75 static void set_dte_entry(struct amd_iommu *iommu,
84 static void iommu_flush_dte_sync(struct amd_iommu *iommu, u16 devid);
86 static struct iommu_dev_data *find_dev_data(struct amd_iommu *iommu, u16 devid);
91 static void clone_aliases(struct amd_iommu *iommu, struct device *dev);
93 static int iommu_completion_wait(struct amd_iommu *iommu);
146 static void update_dte256(struct amd_iommu *iommu, struct iommu_dev_data *dev_data, in update_dte256() argument
150 struct dev_table_entry *dev_table = get_dev_table(iommu); in update_dte256()
159 iommu_flush_dte_sync(iommu, dev_data->devid); in update_dte256()
164 iommu_flush_dte_sync(iommu, dev_data->devid); in update_dte256()
172 iommu_flush_dte_sync(iommu, dev_data->devid); in update_dte256()
[all …]
H A Damd_iommu.h21 void amd_iommu_restart_log(struct amd_iommu *iommu, const char *evt_type,
24 void amd_iommu_restart_event_logging(struct amd_iommu *iommu);
25 void amd_iommu_restart_ga_log(struct amd_iommu *iommu);
26 void amd_iommu_restart_ppr_log(struct amd_iommu *iommu);
27 void amd_iommu_set_rlookup_table(struct amd_iommu *iommu, u16 devid);
28 void iommu_feature_enable(struct amd_iommu *iommu, u8 bit);
29 void *__init iommu_alloc_4k_pages(struct amd_iommu *iommu,
67 int amd_iommu_iopf_init(struct amd_iommu *iommu);
68 void amd_iommu_iopf_uninit(struct amd_iommu *iommu);
71 int amd_iommu_iopf_add_device(struct amd_iommu *iommu,
[all …]
/linux/arch/sparc/kernel/
H A Diommu.c52 struct iommu *iommu = container_of(iommu_map_table, struct iommu, tbl); in iommu_flushall() local
53 if (iommu->iommu_flushinv) { in iommu_flushall()
54 iommu_write(iommu->iommu_flushinv, ~(u64)0); in iommu_flushall()
59 tag = iommu->iommu_tags; in iommu_flushall()
66 (void) iommu_read(iommu->write_complete_reg); in iommu_flushall()
80 #define IOPTE_IS_DUMMY(iommu, iopte) \ argument
81 ((iopte_val(*iopte) & IOPTE_PAGE) == (iommu)->dummy_page_pa)
83 static inline void iopte_make_dummy(struct iommu *iommu, iopte_t *iopte) in iopte_make_dummy() argument
88 val |= iommu->dummy_page_pa; in iopte_make_dummy()
93 int iommu_table_init(struct iommu *iommu, int tsbsize, in iommu_table_init() argument
[all …]
H A Diommu-common.c19 static inline bool need_flush(struct iommu_map_table *iommu) in need_flush() argument
21 return ((iommu->flags & IOMMU_NEED_FLUSH) != 0); in need_flush()
24 static inline void set_flush(struct iommu_map_table *iommu) in set_flush() argument
26 iommu->flags |= IOMMU_NEED_FLUSH; in set_flush()
29 static inline void clear_flush(struct iommu_map_table *iommu) in clear_flush() argument
31 iommu->flags &= ~IOMMU_NEED_FLUSH; in clear_flush()
52 void iommu_tbl_pool_init(struct iommu_map_table *iommu, in iommu_tbl_pool_init() argument
60 struct iommu_pool *p = &(iommu->large_pool); in iommu_tbl_pool_init()
64 iommu->nr_pools = IOMMU_NR_POOLS; in iommu_tbl_pool_init()
66 iommu->nr_pools = npools; in iommu_tbl_pool_init()
[all …]
H A Dsbus.c63 struct iommu *iommu = dev->archdata.iommu; in sbus_set_sbus64() local
78 cfg_reg = iommu->write_complete_reg; in sbus_set_sbus64()
213 struct iommu *iommu = op->dev.archdata.iommu; in sbus_build_irq() local
214 unsigned long reg_base = iommu->write_complete_reg - 0x2000UL; in sbus_build_irq()
275 struct iommu *iommu = op->dev.archdata.iommu; in sysio_ue_handler() local
276 unsigned long reg_base = iommu->write_complete_reg - 0x2000UL; in sysio_ue_handler()
349 struct iommu *iommu = op->dev.archdata.iommu; in sysio_ce_handler() local
350 unsigned long reg_base = iommu->write_complete_reg - 0x2000UL; in sysio_ce_handler()
428 struct iommu *iommu = op->dev.archdata.iommu; in sysio_sbus_error_handler() local
433 reg_base = iommu->write_complete_reg - 0x2000UL; in sysio_sbus_error_handler()
[all …]
H A Dpci_sun4v.c78 static inline bool iommu_use_atu(struct iommu *iommu, u64 mask) in iommu_use_atu() argument
80 return iommu->atu && mask > DMA_BIT_MASK(32); in iommu_use_atu()
102 if (!iommu_use_atu(pbm->iommu, mask)) { in iommu_batch_flush()
119 iotsb_num = pbm->iommu->atu->iotsb->iotsb_num; in iommu_batch_flush()
188 struct iommu *iommu; in dma_4v_alloc_coherent() local
213 iommu = dev->archdata.iommu; in dma_4v_alloc_coherent()
215 if (!iommu_use_atu(iommu, mask)) in dma_4v_alloc_coherent()
216 tbl = &iommu->tbl; in dma_4v_alloc_coherent()
218 tbl = &iommu->atu->tbl; in dma_4v_alloc_coherent()
327 struct iommu *iommu; in dma_4v_free_coherent() local
[all …]
/linux/tools/testing/selftests/vfio/lib/
H A Diommu.c69 int __iommu_hva2iova(struct iommu *iommu, void *vaddr, iova_t *iova) in __iommu_hva2iova() argument
73 list_for_each_entry(region, &iommu->dma_regions, link) { in __iommu_hva2iova()
89 iova_t iommu_hva2iova(struct iommu *iommu, void *vaddr) in iommu_hva2iova() argument
94 ret = __iommu_hva2iova(iommu, vaddr, &iova); in iommu_hva2iova()
100 static int vfio_iommu_map(struct iommu *iommu, struct dma_region *region) in vfio_iommu_map() argument
110 if (ioctl(iommu->container_fd, VFIO_IOMMU_MAP_DMA, &args)) in vfio_iommu_map()
116 static int iommufd_map(struct iommu *iommu, struct dma_region *region) in iommufd_map() argument
126 .ioas_id = iommu->ioas_id, in iommufd_map()
129 if (ioctl(iommu->iommufd, IOMMU_IOAS_MAP, &args)) in iommufd_map()
135 int __iommu_map(struct iommu *iommu, struct dma_region *region) in __iommu_map() argument
[all …]
/linux/drivers/iommu/intel/
H A Ddmar.c68 static void free_iommu(struct intel_iommu *iommu);
462 if (dmaru->iommu) in dmar_free_drhd()
463 free_iommu(dmaru->iommu); in dmar_free_drhd()
502 drhd->iommu->node = node; in dmar_parse_one_rhsa()
939 x86_init.iommu.iommu_init = intel_iommu_init; in detect_intel_iommu()
950 static void unmap_iommu(struct intel_iommu *iommu) in unmap_iommu() argument
952 iounmap(iommu->reg); in unmap_iommu()
953 release_mem_region(iommu->reg_phys, iommu->reg_size); in unmap_iommu()
964 static int map_iommu(struct intel_iommu *iommu, struct dmar_drhd_unit *drhd) in map_iommu() argument
969 iommu->reg_phys = phys_addr; in map_iommu()
[all …]
H A Dperf.c18 bool dmar_latency_enabled(struct intel_iommu *iommu, enum latency_type type) in dmar_latency_enabled() argument
20 struct latency_statistic *lstat = iommu->perf_statistic; in dmar_latency_enabled()
25 int dmar_latency_enable(struct intel_iommu *iommu, enum latency_type type) in dmar_latency_enable() argument
31 if (dmar_latency_enabled(iommu, type)) in dmar_latency_enable()
35 if (!iommu->perf_statistic) { in dmar_latency_enable()
36 iommu->perf_statistic = kzalloc_objs(*lstat, DMAR_LATENCY_NUM, in dmar_latency_enable()
38 if (!iommu->perf_statistic) { in dmar_latency_enable()
44 lstat = iommu->perf_statistic; in dmar_latency_enable()
57 void dmar_latency_disable(struct intel_iommu *iommu, enum latency_type type) in dmar_latency_disable() argument
59 struct latency_statistic *lstat = iommu->perf_statistic; in dmar_latency_disable()
[all …]
H A Dsvm.c28 void intel_svm_check(struct intel_iommu *iommu) in intel_svm_check() argument
30 if (!pasid_supported(iommu)) in intel_svm_check()
34 !cap_fl1gp_support(iommu->cap)) { in intel_svm_check()
36 iommu->name); in intel_svm_check()
41 !cap_fl5lp_support(iommu->cap)) { in intel_svm_check()
43 iommu->name); in intel_svm_check()
47 iommu->flags |= VTD_FLAG_SVM_CAPABLE; in intel_svm_check()
92 intel_pasid_tear_down_entry(info->iommu, dev_pasid->dev, in intel_mm_release()
116 struct intel_iommu *iommu; in intel_iommu_sva_supported() local
121 iommu = info->iommu; in intel_iommu_sva_supported()
[all …]
H A Dperfmon.h39 int alloc_iommu_pmu(struct intel_iommu *iommu);
40 void free_iommu_pmu(struct intel_iommu *iommu);
41 void iommu_pmu_register(struct intel_iommu *iommu);
42 void iommu_pmu_unregister(struct intel_iommu *iommu);
45 alloc_iommu_pmu(struct intel_iommu *iommu) in alloc_iommu_pmu() argument
51 free_iommu_pmu(struct intel_iommu *iommu) in free_iommu_pmu() argument
56 iommu_pmu_register(struct intel_iommu *iommu) in iommu_pmu_register() argument
61 iommu_pmu_unregister(struct intel_iommu *iommu) in iommu_pmu_unregister() argument
/linux/tools/testing/selftests/vfio/lib/include/libvfio/
H A Diommu.h27 struct iommu { struct
35 struct iommu *iommu_init(const char *iommu_mode); argument
36 void iommu_cleanup(struct iommu *iommu);
38 int __iommu_map(struct iommu *iommu, struct dma_region *region);
40 static inline void iommu_map(struct iommu *iommu, struct dma_region *region) in iommu_map() argument
42 VFIO_ASSERT_EQ(__iommu_map(iommu, region), 0); in iommu_map()
45 int __iommu_unmap(struct iommu *iomm
28 modeiommu global() argument
30 iommufdiommu global() argument
47 iommu_unmap(struct iommu * iommu,struct dma_region * region) iommu_unmap() argument
54 iommu_unmap_all(struct iommu * iommu) iommu_unmap_all() argument
[all...]
/linux/arch/sparc/mm/
H A Diommu.c60 struct iommu_struct *iommu; in sbus_iommu_init() local
67 iommu = kmalloc_obj(struct iommu_struct); in sbus_iommu_init()
68 if (!iommu) { in sbus_iommu_init()
73 iommu->regs = of_ioremap(&op->resource[0], 0, PAGE_SIZE * 3, in sbus_iommu_init()
75 if (!iommu->regs) { in sbus_iommu_init()
80 control = sbus_readl(&iommu->regs->control); in sbus_iommu_init()
85 sbus_writel(control, &iommu->regs->control); in sbus_iommu_init()
87 iommu_invalidate(iommu->regs); in sbus_iommu_init()
88 iommu->start = IOMMU_START; in sbus_iommu_init()
89 iommu->end = 0xffffffff; in sbus_iommu_init()
[all …]
/linux/Documentation/ABI/testing/
H A Ddebugfs-amd-iommu1 What: /sys/kernel/debug/iommu/amd/iommu<x>/mmio
6 MMIO register offset for iommu<x>, and the file outputs the corresponding
7 MMIO register value of iommu<x>
11 $ echo "0x18" > /sys/kernel/debug/iommu/amd/iommu00/mmio
12 $ cat /sys/kernel/debug/iommu/amd/iommu00/mmio
18 What: /sys/kernel/debug/iommu/amd/iommu<x>/capability
23 capability register offset for iommu<x>, and the file outputs the
24 corresponding capability register value of iommu<x>.
28 $ echo "0x10" > /sys/kernel/debug/iommu/amd/iommu00/capability
29 $ cat /sys/kernel/debug/iommu/amd/iommu00/capability
[all …]
H A Dsysfs-class-iommu-intel-iommu1 What: /sys/class/iommu/<iommu>/intel-iommu/address
8 intel-iommu with a DMAR DRHD table entry.
10 What: /sys/class/iommu/<iommu>/intel-iommu/cap
18 What: /sys/class/iommu/<iommu>/intel-iommu/ecap
26 What: /sys/class/iommu/<iommu>/intel-iommu/version
/linux/drivers/gpu/drm/msm/
H A Dmsm_iommu.c138 struct msm_iommu *iommu = to_msm_iommu(pagetable->parent); in msm_iommu_pagetable_map_prr() local
139 phys_addr_t phys = page_to_phys(iommu->prr_page); in msm_iommu_pagetable_map_prr()
224 struct msm_iommu *iommu = to_msm_iommu(pagetable->parent); in msm_iommu_pagetable_destroy() local
232 mutex_lock(&iommu->init_lock); in msm_iommu_pagetable_destroy()
233 if (--iommu->pagetables == 0) { in msm_iommu_pagetable_destroy()
238 __free_page(iommu->prr_page); in msm_iommu_pagetable_destroy()
239 iommu->prr_page = NULL; in msm_iommu_pagetable_destroy()
242 mutex_unlock(&iommu->init_lock); in msm_iommu_pagetable_destroy()
269 struct msm_iommu *iommu = to_msm_iommu(mmu); in msm_iommu_get_geometry() local
271 return &iommu->domain->geometry; in msm_iommu_get_geometry()
[all …]
/linux/tools/testing/selftests/vfio/
H A Dvfio_pci_driver_test.c21 static void region_setup(struct iommu *iommu, in region_setup() argument
36 iommu_map(iommu, region); in region_setup()
39 static void region_teardown(struct iommu *iommu, struct dma_region *region) in region_teardown() argument
41 iommu_unmap(iommu, region); in region_teardown()
46 struct iommu *iommu; in FIXTURE() local
76 self->iommu = iommu_init(variant->iommu_mode); in FIXTURE_SETUP()
77 self->device = vfio_pci_device_init(device_bdf, self->iommu); in FIXTURE_SETUP()
78 self->iova_allocator = iova_allocator_init(self->iommu); in FIXTURE_SETUP()
82 region_setup(self->iommu, self->iova_allocator, &self->memcpy_region, SZ_1G); in FIXTURE_SETUP()
83 region_setup(self->iommu, self->iova_allocator, &driver->region, SZ_2M); in FIXTURE_SETUP()
[all …]
H A Dvfio_dma_mapping_test.c96 struct iommu *iommu; in FIXTURE() local
122 self->iommu = iommu_init(variant->iommu_mode); in FIXTURE_SETUP()
123 self->device = vfio_pci_device_init(device_bdf, self->iommu); in FIXTURE_SETUP()
124 self->iova_allocator = iova_allocator_init(self->iommu); in FIXTURE_SETUP()
131 iommu_cleanup(self->iommu); in FIXTURE_TEARDOWN()
155 iommu_map(self->iommu, &region); in TEST_F()
164 if (self->iommu->mode->iommu_type == VFIO_TYPE1_IOMMU) in TEST_F()
193 rc = __iommu_unmap(self->iommu, &region, &unmapped); in TEST_F()
204 struct iommu *iommu; in FIXTURE() local
237 self->iommu = iommu_init(variant->iommu_mode); in FIXTURE_SETUP()
[all …]
/linux/Documentation/devicetree/bindings/iommu/
H A Dti,omap-iommu.txt5 "ti,omap2-iommu" for OMAP2/OMAP3 IOMMU instances
6 "ti,omap4-iommu" for OMAP4/OMAP5 IOMMU instances
7 "ti,dra7-dsp-iommu" for DRA7xx DSP IOMMU instances
8 "ti,dra7-iommu" for DRA7xx IOMMU instances
12 - #iommu-cells : Should be 0. OMAP IOMMUs are all "single-master" devices,
16 Documentation/devicetree/bindings/iommu/iommu.txt
21 - ti,iommu-bus-err-back : Indicates the IOMMU instance supports throwing
34 #iommu-cells = <0>;
35 compatible = "ti,omap2-iommu";
44 compatible = "ti,dra7-dsp-iommu";
[all …]

12345678