xref: /linux/drivers/gpu/drm/exynos/exynos_drm_dma.c (revision 95298d63c67673c654c08952672d016212b26054)
1 // SPDX-License-Identifier: GPL-2.0
2 //
3 // Copyright (c) 2012 Samsung Electronics Co., Ltd.
4 // Author: Inki Dae <inki.dae@samsung.com>
5 // Author: Andrzej Hajda <a.hajda@samsung.com>
6 
7 #include <linux/dma-iommu.h>
8 #include <linux/dma-mapping.h>
9 #include <linux/iommu.h>
10 #include <linux/platform_device.h>
11 
12 #include <drm/drm_print.h>
13 #include <drm/exynos_drm.h>
14 
15 #include "exynos_drm_drv.h"
16 
17 #if defined(CONFIG_ARM_DMA_USE_IOMMU)
18 #include <asm/dma-iommu.h>
19 #else
20 #define arm_iommu_create_mapping(...)	({ NULL; })
21 #define arm_iommu_attach_device(...)	({ -ENODEV; })
22 #define arm_iommu_release_mapping(...)	({ })
23 #define arm_iommu_detach_device(...)	({ })
24 #define to_dma_iommu_mapping(dev) NULL
25 #endif
26 
27 #if !defined(CONFIG_IOMMU_DMA)
28 #define iommu_dma_init_domain(...) ({ -EINVAL; })
29 #endif
30 
31 #define EXYNOS_DEV_ADDR_START	0x20000000
32 #define EXYNOS_DEV_ADDR_SIZE	0x40000000
33 
34 static inline int configure_dma_max_seg_size(struct device *dev)
35 {
36 	if (!dev->dma_parms)
37 		dev->dma_parms = kzalloc(sizeof(*dev->dma_parms), GFP_KERNEL);
38 	if (!dev->dma_parms)
39 		return -ENOMEM;
40 
41 	dma_set_max_seg_size(dev, DMA_BIT_MASK(32));
42 	return 0;
43 }
44 
45 static inline void clear_dma_max_seg_size(struct device *dev)
46 {
47 	kfree(dev->dma_parms);
48 	dev->dma_parms = NULL;
49 }
50 
51 /*
52  * drm_iommu_attach_device- attach device to iommu mapping
53  *
54  * @drm_dev: DRM device
55  * @subdrv_dev: device to be attach
56  *
57  * This function should be called by sub drivers to attach it to iommu
58  * mapping.
59  */
60 static int drm_iommu_attach_device(struct drm_device *drm_dev,
61 				struct device *subdrv_dev, void **dma_priv)
62 {
63 	struct exynos_drm_private *priv = drm_dev->dev_private;
64 	int ret;
65 
66 	if (get_dma_ops(priv->dma_dev) != get_dma_ops(subdrv_dev)) {
67 		DRM_DEV_ERROR(subdrv_dev, "Device %s lacks support for IOMMU\n",
68 			  dev_name(subdrv_dev));
69 		return -EINVAL;
70 	}
71 
72 	ret = configure_dma_max_seg_size(subdrv_dev);
73 	if (ret)
74 		return ret;
75 
76 	if (IS_ENABLED(CONFIG_ARM_DMA_USE_IOMMU)) {
77 		/*
78 		 * Keep the original DMA mapping of the sub-device and
79 		 * restore it on Exynos DRM detach, otherwise the DMA
80 		 * framework considers it as IOMMU-less during the next
81 		 * probe (in case of deferred probe or modular build)
82 		 */
83 		*dma_priv = to_dma_iommu_mapping(subdrv_dev);
84 		if (*dma_priv)
85 			arm_iommu_detach_device(subdrv_dev);
86 
87 		ret = arm_iommu_attach_device(subdrv_dev, priv->mapping);
88 	} else if (IS_ENABLED(CONFIG_IOMMU_DMA)) {
89 		ret = iommu_attach_device(priv->mapping, subdrv_dev);
90 	}
91 
92 	if (ret)
93 		clear_dma_max_seg_size(subdrv_dev);
94 
95 	return 0;
96 }
97 
98 /*
99  * drm_iommu_detach_device -detach device address space mapping from device
100  *
101  * @drm_dev: DRM device
102  * @subdrv_dev: device to be detached
103  *
104  * This function should be called by sub drivers to detach it from iommu
105  * mapping
106  */
107 static void drm_iommu_detach_device(struct drm_device *drm_dev,
108 				    struct device *subdrv_dev, void **dma_priv)
109 {
110 	struct exynos_drm_private *priv = drm_dev->dev_private;
111 
112 	if (IS_ENABLED(CONFIG_ARM_DMA_USE_IOMMU)) {
113 		arm_iommu_detach_device(subdrv_dev);
114 		arm_iommu_attach_device(subdrv_dev, *dma_priv);
115 	} else if (IS_ENABLED(CONFIG_IOMMU_DMA))
116 		iommu_detach_device(priv->mapping, subdrv_dev);
117 
118 	clear_dma_max_seg_size(subdrv_dev);
119 }
120 
121 int exynos_drm_register_dma(struct drm_device *drm, struct device *dev,
122 			    void **dma_priv)
123 {
124 	struct exynos_drm_private *priv = drm->dev_private;
125 
126 	if (!priv->dma_dev) {
127 		priv->dma_dev = dev;
128 		DRM_INFO("Exynos DRM: using %s device for DMA mapping operations\n",
129 			 dev_name(dev));
130 	}
131 
132 	if (!IS_ENABLED(CONFIG_EXYNOS_IOMMU))
133 		return 0;
134 
135 	if (!priv->mapping) {
136 		void *mapping;
137 
138 		if (IS_ENABLED(CONFIG_ARM_DMA_USE_IOMMU))
139 			mapping = arm_iommu_create_mapping(&platform_bus_type,
140 				EXYNOS_DEV_ADDR_START, EXYNOS_DEV_ADDR_SIZE);
141 		else if (IS_ENABLED(CONFIG_IOMMU_DMA))
142 			mapping = iommu_get_domain_for_dev(priv->dma_dev);
143 
144 		if (IS_ERR(mapping))
145 			return PTR_ERR(mapping);
146 		priv->mapping = mapping;
147 	}
148 
149 	return drm_iommu_attach_device(drm, dev, dma_priv);
150 }
151 
152 void exynos_drm_unregister_dma(struct drm_device *drm, struct device *dev,
153 			       void **dma_priv)
154 {
155 	if (IS_ENABLED(CONFIG_EXYNOS_IOMMU))
156 		drm_iommu_detach_device(drm, dev, dma_priv);
157 }
158 
159 void exynos_drm_cleanup_dma(struct drm_device *drm)
160 {
161 	struct exynos_drm_private *priv = drm->dev_private;
162 
163 	if (!IS_ENABLED(CONFIG_EXYNOS_IOMMU))
164 		return;
165 
166 	arm_iommu_release_mapping(priv->mapping);
167 	priv->mapping = NULL;
168 	priv->dma_dev = NULL;
169 }
170