xref: /linux/drivers/gpu/drm/i915/i915_scatterlist.h (revision 0ea5c948cb64bab5bc7a5516774eb8536f05aa0d)
1 /*
2  * SPDX-License-Identifier: MIT
3  *
4  * Copyright © 2016 Intel Corporation
5  */
6 
7 #ifndef I915_SCATTERLIST_H
8 #define I915_SCATTERLIST_H
9 
10 #include <linux/pfn.h>
11 #include <linux/scatterlist.h>
12 #include <linux/dma-mapping.h>
13 #include <xen/xen.h>
14 
15 #include "i915_gem.h"
16 
17 struct drm_mm_node;
18 struct ttm_resource;
19 
20 /*
21  * Optimised SGL iterator for GEM objects
22  */
23 static __always_inline struct sgt_iter {
24 	struct scatterlist *sgp;
25 	union {
26 		unsigned long pfn;
27 		dma_addr_t dma;
28 	};
29 	unsigned int curr;
30 	unsigned int max;
__sgt_iter(struct scatterlist * sgl,bool dma)31 } __sgt_iter(struct scatterlist *sgl, bool dma) {
32 	struct sgt_iter s = { .sgp = sgl };
33 
34 	if (dma && s.sgp && sg_dma_len(s.sgp) == 0) {
35 		s.sgp = NULL;
36 	} else if (s.sgp) {
37 		s.max = s.curr = s.sgp->offset;
38 		if (dma) {
39 			s.dma = sg_dma_address(s.sgp);
40 			s.max += sg_dma_len(s.sgp);
41 		} else {
42 			s.pfn = page_to_pfn(sg_page(s.sgp));
43 			s.max += s.sgp->length;
44 		}
45 	}
46 
47 	return s;
48 }
49 
__sg_page_count(const struct scatterlist * sg)50 static inline int __sg_page_count(const struct scatterlist *sg)
51 {
52 	return sg->length >> PAGE_SHIFT;
53 }
54 
__sg_dma_page_count(const struct scatterlist * sg)55 static inline int __sg_dma_page_count(const struct scatterlist *sg)
56 {
57 	return sg_dma_len(sg) >> PAGE_SHIFT;
58 }
59 
____sg_next(struct scatterlist * sg)60 static inline struct scatterlist *____sg_next(struct scatterlist *sg)
61 {
62 	++sg;
63 	if (unlikely(sg_is_chain(sg)))
64 		sg = sg_chain_ptr(sg);
65 	return sg;
66 }
67 
68 /**
69  * __sg_next - return the next scatterlist entry in a list
70  * @sg:		The current sg entry
71  *
72  * Description:
73  *   If the entry is the last, return NULL; otherwise, step to the next
74  *   element in the array (@sg@+1). If that's a chain pointer, follow it;
75  *   otherwise just return the pointer to the current element.
76  **/
__sg_next(struct scatterlist * sg)77 static inline struct scatterlist *__sg_next(struct scatterlist *sg)
78 {
79 	return sg_is_last(sg) ? NULL : ____sg_next(sg);
80 }
81 
82 /**
83  * __for_each_sgt_daddr - iterate over the device addresses of the given sg_table
84  * @__dp:	Device address (output)
85  * @__iter:	'struct sgt_iter' (iterator state, internal)
86  * @__sgt:	sg_table to iterate over (input)
87  * @__step:	step size
88  */
89 #define __for_each_sgt_daddr(__dp, __iter, __sgt, __step)		\
90 	for ((__iter) = __sgt_iter((__sgt)->sgl, true);			\
91 	     ((__dp) = (__iter).dma + (__iter).curr), (__iter).sgp;	\
92 	     (((__iter).curr += (__step)) >= (__iter).max) ?		\
93 	     (__iter) = __sgt_iter(__sg_next((__iter).sgp), true), 0 : 0)
94 /**
95  * __for_each_daddr_next - iterates over the device addresses with pre-initialized iterator.
96  * @__dp:	Device address (output)
97  * @__iter:	'struct sgt_iter' (iterator state, external)
98  * @__step:	step size
99  */
100 #define __for_each_daddr_next(__dp, __iter, __step)                  \
101 	for (; ((__dp) = (__iter).dma + (__iter).curr), (__iter).sgp;   \
102 	     (((__iter).curr += (__step)) >= (__iter).max) ?            \
103 	     (__iter) = __sgt_iter(__sg_next((__iter).sgp), true), 0 : 0)
104 
105 /**
106  * for_each_sgt_page - iterate over the pages of the given sg_table
107  * @__pp:	page pointer (output)
108  * @__iter:	'struct sgt_iter' (iterator state, internal)
109  * @__sgt:	sg_table to iterate over (input)
110  */
111 #define for_each_sgt_page(__pp, __iter, __sgt)				\
112 	for ((__iter) = __sgt_iter((__sgt)->sgl, false);		\
113 	     ((__pp) = (__iter).pfn == 0 ? NULL :			\
114 	      pfn_to_page((__iter).pfn + ((__iter).curr >> PAGE_SHIFT))); \
115 	     (((__iter).curr += PAGE_SIZE) >= (__iter).max) ?		\
116 	     (__iter) = __sgt_iter(__sg_next((__iter).sgp), false), 0 : 0)
117 
118 /**
119  * i915_sg_dma_sizes - Record the dma segment sizes of a scatterlist
120  * @sg: The scatterlist
121  *
122  * Return: An unsigned int with segment sizes logically or'ed together.
123  * A caller can use this information to determine what hardware page table
124  * entry sizes can be used to map the memory represented by the scatterlist.
125  */
i915_sg_dma_sizes(struct scatterlist * sg)126 static inline unsigned int i915_sg_dma_sizes(struct scatterlist *sg)
127 {
128 	unsigned int page_sizes;
129 
130 	page_sizes = 0;
131 	while (sg && sg_dma_len(sg)) {
132 		GEM_BUG_ON(sg->offset);
133 		GEM_BUG_ON(!IS_ALIGNED(sg_dma_len(sg), PAGE_SIZE));
134 		page_sizes |= sg_dma_len(sg);
135 		sg = __sg_next(sg);
136 	}
137 
138 	return page_sizes;
139 }
140 
i915_sg_segment_size(struct device * dev)141 static inline unsigned int i915_sg_segment_size(struct device *dev)
142 {
143 	size_t max = min_t(size_t, UINT_MAX, dma_max_mapping_size(dev));
144 
145 	/*
146 	 * For Xen PV guests pages aren't contiguous in DMA (machine) address
147 	 * space.  The DMA API takes care of that both in dma_alloc_* (by
148 	 * calling into the hypervisor to make the pages contiguous) and in
149 	 * dma_map_* (by bounce buffering).  But i915 abuses ignores the
150 	 * coherency aspects of the DMA API and thus can't cope with bounce
151 	 * buffering actually happening, so add a hack here to force small
152 	 * allocations and mappings when running in PV mode on Xen.
153 	 *
154 	 * Note this will still break if bounce buffering is required for other
155 	 * reasons, like confidential computing hypervisors or PCIe root ports
156 	 * with addressing limitations.
157 	 */
158 	if (xen_pv_domain())
159 		max = PAGE_SIZE;
160 	return round_down(max, PAGE_SIZE);
161 }
162 
163 bool i915_sg_trim(struct sg_table *orig_st);
164 
165 /**
166  * struct i915_refct_sgt_ops - Operations structure for struct i915_refct_sgt
167  */
168 struct i915_refct_sgt_ops {
169 	/**
170 	 * @release: Free the memory of the struct i915_refct_sgt
171 	 */
172 	void (*release)(struct kref *ref);
173 };
174 
175 /**
176  * struct i915_refct_sgt - A refcounted scatter-gather table
177  * @kref: struct kref for refcounting
178  * @table: struct sg_table holding the scatter-gather table itself. Note that
179  * @table->sgl = NULL can be used to determine whether a scatter-gather table
180  * is present or not.
181  * @size: The size in bytes of the underlying memory buffer
182  * @ops: The operations structure.
183  */
184 struct i915_refct_sgt {
185 	struct kref kref;
186 	struct sg_table table;
187 	size_t size;
188 	const struct i915_refct_sgt_ops *ops;
189 };
190 
191 /**
192  * i915_refct_sgt_put - Put a refcounted sg-table
193  * @rsgt: the struct i915_refct_sgt to put.
194  */
i915_refct_sgt_put(struct i915_refct_sgt * rsgt)195 static inline void i915_refct_sgt_put(struct i915_refct_sgt *rsgt)
196 {
197 	if (rsgt)
198 		kref_put(&rsgt->kref, rsgt->ops->release);
199 }
200 
201 /**
202  * i915_refct_sgt_get - Get a refcounted sg-table
203  * @rsgt: the struct i915_refct_sgt to get.
204  */
205 static inline struct i915_refct_sgt *
i915_refct_sgt_get(struct i915_refct_sgt * rsgt)206 i915_refct_sgt_get(struct i915_refct_sgt *rsgt)
207 {
208 	kref_get(&rsgt->kref);
209 	return rsgt;
210 }
211 
212 /**
213  * __i915_refct_sgt_init - Initialize a refcounted sg-list with a custom
214  * operations structure
215  * @rsgt: The struct i915_refct_sgt to initialize.
216  * @size: Size in bytes of the underlying memory buffer.
217  * @ops: A customized operations structure in case the refcounted sg-list
218  * is embedded into another structure.
219  */
__i915_refct_sgt_init(struct i915_refct_sgt * rsgt,size_t size,const struct i915_refct_sgt_ops * ops)220 static inline void __i915_refct_sgt_init(struct i915_refct_sgt *rsgt,
221 					 size_t size,
222 					 const struct i915_refct_sgt_ops *ops)
223 {
224 	kref_init(&rsgt->kref);
225 	rsgt->table.sgl = NULL;
226 	rsgt->size = size;
227 	rsgt->ops = ops;
228 }
229 
230 void i915_refct_sgt_init(struct i915_refct_sgt *rsgt, size_t size);
231 
232 struct i915_refct_sgt *i915_rsgt_from_mm_node(const struct drm_mm_node *node,
233 					      u64 region_start,
234 					      u32 page_alignment);
235 
236 struct i915_refct_sgt *i915_rsgt_from_buddy_resource(struct ttm_resource *res,
237 						     u64 region_start,
238 						     u32 page_alignment);
239 
240 #endif
241