xref: /freebsd/sys/compat/linuxkpi/common/include/linux/io.h (revision 046b82842c61c2d5d1648b5024958827e0a85725)
1 /*-
2  * Copyright (c) 2010 Isilon Systems, Inc.
3  * Copyright (c) 2010 iX Systems, Inc.
4  * Copyright (c) 2010 Panasas, Inc.
5  * Copyright (c) 2013-2015 Mellanox Technologies, Ltd.
6  * All rights reserved.
7  *
8  * Redistribution and use in source and binary forms, with or without
9  * modification, are permitted provided that the following conditions
10  * are met:
11  * 1. Redistributions of source code must retain the above copyright
12  *    notice unmodified, this list of conditions, and the following
13  *    disclaimer.
14  * 2. Redistributions in binary form must reproduce the above copyright
15  *    notice, this list of conditions and the following disclaimer in the
16  *    documentation and/or other materials provided with the distribution.
17  *
18  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
19  * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
20  * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
21  * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
22  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
23  * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
24  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
25  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
26  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
27  * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
28  *
29  * $FreeBSD$
30  */
31 #ifndef	_LINUXKPI_LINUX_IO_H_
32 #define	_LINUXKPI_LINUX_IO_H_
33 
34 #include <sys/endian.h>
35 #include <sys/types.h>
36 
37 #include <machine/vm.h>
38 
39 #include <linux/compiler.h>
40 #include <linux/types.h>
41 #if !defined(__arm__)
42 #include <asm/set_memory.h>
43 #endif
44 
45 /*
46  * XXX This is all x86 specific.  It should be bus space access.
47  */
48 
49 /* rmb and wmb are declared in machine/atomic.h, so should be included first. */
50 #ifndef __io_br
51 #define	__io_br()	__compiler_membar()
52 #endif
53 
54 #ifndef __io_ar
55 #ifdef rmb
56 #define	__io_ar()	rmb()
57 #else
58 #define	__io_ar()	__compiler_membar()
59 #endif
60 #endif
61 
62 #ifndef __io_bw
63 #ifdef wmb
64 #define	__io_bw()	wmb()
65 #else
66 #define	__io_bw()	__compiler_membar()
67 #endif
68 #endif
69 
70 #ifndef __io_aw
71 #define	__io_aw()	__compiler_membar()
72 #endif
73 
74 /* Access MMIO registers atomically without barriers and byte swapping. */
75 
76 static inline uint8_t
77 __raw_readb(const volatile void *addr)
78 {
79 	return (*(const volatile uint8_t *)addr);
80 }
81 #define	__raw_readb(addr)	__raw_readb(addr)
82 
83 static inline void
84 __raw_writeb(uint8_t v, volatile void *addr)
85 {
86 	*(volatile uint8_t *)addr = v;
87 }
88 #define	__raw_writeb(v, addr)	__raw_writeb(v, addr)
89 
90 static inline uint16_t
91 __raw_readw(const volatile void *addr)
92 {
93 	return (*(const volatile uint16_t *)addr);
94 }
95 #define	__raw_readw(addr)	__raw_readw(addr)
96 
97 static inline void
98 __raw_writew(uint16_t v, volatile void *addr)
99 {
100 	*(volatile uint16_t *)addr = v;
101 }
102 #define	__raw_writew(v, addr)	__raw_writew(v, addr)
103 
104 static inline uint32_t
105 __raw_readl(const volatile void *addr)
106 {
107 	return (*(const volatile uint32_t *)addr);
108 }
109 #define	__raw_readl(addr)	__raw_readl(addr)
110 
111 static inline void
112 __raw_writel(uint32_t v, volatile void *addr)
113 {
114 	*(volatile uint32_t *)addr = v;
115 }
116 #define	__raw_writel(v, addr)	__raw_writel(v, addr)
117 
118 #ifdef __LP64__
119 static inline uint64_t
120 __raw_readq(const volatile void *addr)
121 {
122 	return (*(const volatile uint64_t *)addr);
123 }
124 #define	__raw_readq(addr)	__raw_readq(addr)
125 
126 static inline void
127 __raw_writeq(uint64_t v, volatile void *addr)
128 {
129 	*(volatile uint64_t *)addr = v;
130 }
131 #define	__raw_writeq(v, addr)	__raw_writeq(v, addr)
132 #endif
133 
134 #define	mmiowb()	barrier()
135 
136 /* Access little-endian MMIO registers atomically with memory barriers. */
137 
138 #undef readb
139 static inline uint8_t
140 readb(const volatile void *addr)
141 {
142 	uint8_t v;
143 
144 	__io_br();
145 	v = *(const volatile uint8_t *)addr;
146 	__io_ar();
147 	return (v);
148 }
149 #define	readb(addr)		readb(addr)
150 
151 #undef writeb
152 static inline void
153 writeb(uint8_t v, volatile void *addr)
154 {
155 	__io_bw();
156 	*(volatile uint8_t *)addr = v;
157 	__io_aw();
158 }
159 #define	writeb(v, addr)		writeb(v, addr)
160 
161 #undef readw
162 static inline uint16_t
163 readw(const volatile void *addr)
164 {
165 	uint16_t v;
166 
167 	__io_br();
168 	v = le16toh(__raw_readw(addr));
169 	__io_ar();
170 	return (v);
171 }
172 #define	readw(addr)		readw(addr)
173 
174 #undef writew
175 static inline void
176 writew(uint16_t v, volatile void *addr)
177 {
178 	__io_bw();
179 	__raw_writew(htole16(v), addr);
180 	__io_aw();
181 }
182 #define	writew(v, addr)		writew(v, addr)
183 
184 #undef readl
185 static inline uint32_t
186 readl(const volatile void *addr)
187 {
188 	uint32_t v;
189 
190 	__io_br();
191 	v = le32toh(__raw_readl(addr));
192 	__io_ar();
193 	return (v);
194 }
195 #define	readl(addr)		readl(addr)
196 
197 #undef writel
198 static inline void
199 writel(uint32_t v, volatile void *addr)
200 {
201 	__io_bw();
202 	__raw_writel(htole32(v), addr);
203 	__io_aw();
204 }
205 #define	writel(v, addr)		writel(v, addr)
206 
207 #undef readq
208 #undef writeq
209 #ifdef __LP64__
210 static inline uint64_t
211 readq(const volatile void *addr)
212 {
213 	uint64_t v;
214 
215 	__io_br();
216 	v = le64toh(__raw_readq(addr));
217 	__io_ar();
218 	return (v);
219 }
220 #define	readq(addr)		readq(addr)
221 
222 static inline void
223 writeq(uint64_t v, volatile void *addr)
224 {
225 	__io_bw();
226 	__raw_writeq(htole64(v), addr);
227 	__io_aw();
228 }
229 #define	writeq(v, addr)		writeq(v, addr)
230 #endif
231 
232 /* Access little-endian MMIO registers atomically without memory barriers. */
233 
234 #undef readb_relaxed
235 static inline uint8_t
236 readb_relaxed(const volatile void *addr)
237 {
238 	return (__raw_readb(addr));
239 }
240 #define	readb_relaxed(addr)	readb_relaxed(addr)
241 
242 #undef writeb_relaxed
243 static inline void
244 writeb_relaxed(uint8_t v, volatile void *addr)
245 {
246 	__raw_writeb(v, addr);
247 }
248 #define	writeb_relaxed(v, addr)	writeb_relaxed(v, addr)
249 
250 #undef readw_relaxed
251 static inline uint16_t
252 readw_relaxed(const volatile void *addr)
253 {
254 	return (le16toh(__raw_readw(addr)));
255 }
256 #define	readw_relaxed(addr)	readw_relaxed(addr)
257 
258 #undef writew_relaxed
259 static inline void
260 writew_relaxed(uint16_t v, volatile void *addr)
261 {
262 	__raw_writew(htole16(v), addr);
263 }
264 #define	writew_relaxed(v, addr)	writew_relaxed(v, addr)
265 
266 #undef readl_relaxed
267 static inline uint32_t
268 readl_relaxed(const volatile void *addr)
269 {
270 	return (le32toh(__raw_readl(addr)));
271 }
272 #define	readl_relaxed(addr)	readl_relaxed(addr)
273 
274 #undef writel_relaxed
275 static inline void
276 writel_relaxed(uint32_t v, volatile void *addr)
277 {
278 	__raw_writel(htole32(v), addr);
279 }
280 #define	writel_relaxed(v, addr)	writel_relaxed(v, addr)
281 
282 #undef readq_relaxed
283 #undef writeq_relaxed
284 #ifdef __LP64__
285 static inline uint64_t
286 readq_relaxed(const volatile void *addr)
287 {
288 	return (le64toh(__raw_readq(addr)));
289 }
290 #define	readq_relaxed(addr)	readq_relaxed(addr)
291 
292 static inline void
293 writeq_relaxed(uint64_t v, volatile void *addr)
294 {
295 	__raw_writeq(htole64(v), addr);
296 }
297 #define	writeq_relaxed(v, addr)	writeq_relaxed(v, addr)
298 #endif
299 
300 /* XXX On Linux ioread and iowrite handle both MMIO and port IO. */
301 
302 #undef ioread8
303 static inline uint8_t
304 ioread8(const volatile void *addr)
305 {
306 	return (readb(addr));
307 }
308 #define	ioread8(addr)		ioread8(addr)
309 
310 #undef ioread16
311 static inline uint16_t
312 ioread16(const volatile void *addr)
313 {
314 	return (readw(addr));
315 }
316 #define	ioread16(addr)		ioread16(addr)
317 
318 #undef ioread16be
319 static inline uint16_t
320 ioread16be(const volatile void *addr)
321 {
322 	uint16_t v;
323 
324 	__io_br();
325 	v = (be16toh(__raw_readw(addr)));
326 	__io_ar();
327 
328 	return (v);
329 }
330 #define	ioread16be(addr)	ioread16be(addr)
331 
332 #undef ioread32
333 static inline uint32_t
334 ioread32(const volatile void *addr)
335 {
336 	return (readl(addr));
337 }
338 #define	ioread32(addr)		ioread32(addr)
339 
340 #undef ioread32be
341 static inline uint32_t
342 ioread32be(const volatile void *addr)
343 {
344 	uint32_t v;
345 
346 	__io_br();
347 	v = (be32toh(__raw_readl(addr)));
348 	__io_ar();
349 
350 	return (v);
351 }
352 #define	ioread32be(addr)	ioread32be(addr)
353 
354 #undef iowrite8
355 static inline void
356 iowrite8(uint8_t v, volatile void *addr)
357 {
358 	writeb(v, addr);
359 }
360 #define	iowrite8(v, addr)	iowrite8(v, addr)
361 
362 #undef iowrite16
363 static inline void
364 iowrite16(uint16_t v, volatile void *addr)
365 {
366 	writew(v, addr);
367 }
368 #define	iowrite16	iowrite16
369 
370 #undef iowrite32
371 static inline void
372 iowrite32(uint32_t v, volatile void *addr)
373 {
374 	writel(v, addr);
375 }
376 #define	iowrite32(v, addr)	iowrite32(v, addr)
377 
378 #undef iowrite32be
379 static inline void
380 iowrite32be(uint32_t v, volatile void *addr)
381 {
382 	__io_bw();
383 	__raw_writel(htobe32(v), addr);
384 	__io_aw();
385 }
386 #define	iowrite32be(v, addr)	iowrite32be(v, addr)
387 
388 #if defined(__i386__) || defined(__amd64__)
389 static inline void
390 _outb(u_char data, u_int port)
391 {
392 	__asm __volatile("outb %0, %w1" : : "a" (data), "Nd" (port));
393 }
394 #endif
395 
396 #if defined(__i386__) || defined(__amd64__) || defined(__powerpc__) || defined(__aarch64__) || defined(__riscv)
397 void *_ioremap_attr(vm_paddr_t phys_addr, unsigned long size, int attr);
398 #else
399 static __inline void *
400 _ioremap_attr(vm_paddr_t _phys_addr, unsigned long _size, int _attr)
401 {
402 	return (NULL);
403 }
404 #endif
405 
406 #ifdef VM_MEMATTR_DEVICE
407 #define	ioremap_nocache(addr, size)					\
408     _ioremap_attr((addr), (size), VM_MEMATTR_DEVICE)
409 #define	ioremap_wt(addr, size)						\
410     _ioremap_attr((addr), (size), VM_MEMATTR_DEVICE)
411 #define	ioremap(addr, size)						\
412     _ioremap_attr((addr), (size), VM_MEMATTR_DEVICE)
413 #else
414 #define	ioremap_nocache(addr, size)					\
415     _ioremap_attr((addr), (size), VM_MEMATTR_UNCACHEABLE)
416 #define	ioremap_wt(addr, size)						\
417     _ioremap_attr((addr), (size), VM_MEMATTR_WRITE_THROUGH)
418 #define	ioremap(addr, size)						\
419     _ioremap_attr((addr), (size), VM_MEMATTR_UNCACHEABLE)
420 #endif
421 #ifdef VM_MEMATTR_WRITE_COMBINING
422 #define	ioremap_wc(addr, size)						\
423     _ioremap_attr((addr), (size), VM_MEMATTR_WRITE_COMBINING)
424 #else
425 #define	ioremap_wc(addr, size)	ioremap_nocache(addr, size)
426 #endif
427 #define	ioremap_wb(addr, size)						\
428     _ioremap_attr((addr), (size), VM_MEMATTR_WRITE_BACK)
429 void iounmap(void *addr);
430 
431 #define	memset_io(a, b, c)	memset((a), (b), (c))
432 #define	memcpy_fromio(a, b, c)	memcpy((a), (b), (c))
433 #define	memcpy_toio(a, b, c)	memcpy((a), (b), (c))
434 
435 static inline void
436 __iowrite32_copy(void *to, const void *from, size_t count)
437 {
438 	const uint32_t *src;
439 	uint32_t *dst;
440 	int i;
441 
442 	for (i = 0, src = from, dst = to; i < count; i++, src++, dst++)
443 		__raw_writel(*src, dst);
444 }
445 
446 static inline void
447 __iowrite64_copy(void *to, const void *from, size_t count)
448 {
449 #ifdef __LP64__
450 	const uint64_t *src;
451 	uint64_t *dst;
452 	int i;
453 
454 	for (i = 0, src = from, dst = to; i < count; i++, src++, dst++)
455 		__raw_writeq(*src, dst);
456 #else
457 	__iowrite32_copy(to, from, count * 2);
458 #endif
459 }
460 
461 static inline void
462 __ioread32_copy(void *to, const void *from, size_t count)
463 {
464 	const uint32_t *src;
465 	uint32_t *dst;
466 	int i;
467 
468 	for (i = 0, src = from, dst = to; i < count; i++, src++, dst++)
469 		*dst = __raw_readl(src);
470 }
471 
472 static inline void
473 __ioread64_copy(void *to, const void *from, size_t count)
474 {
475 #ifdef __LP64__
476 	const uint64_t *src;
477 	uint64_t *dst;
478 	int i;
479 
480 	for (i = 0, src = from, dst = to; i < count; i++, src++, dst++)
481 		*dst = __raw_readq(src);
482 #else
483 	__ioread32_copy(to, from, count * 2);
484 #endif
485 }
486 
487 enum {
488 	MEMREMAP_WB = 1 << 0,
489 	MEMREMAP_WT = 1 << 1,
490 	MEMREMAP_WC = 1 << 2,
491 };
492 
493 static inline void *
494 memremap(resource_size_t offset, size_t size, unsigned long flags)
495 {
496 	void *addr = NULL;
497 
498 	if ((flags & MEMREMAP_WB) &&
499 	    (addr = ioremap_wb(offset, size)) != NULL)
500 		goto done;
501 	if ((flags & MEMREMAP_WT) &&
502 	    (addr = ioremap_wt(offset, size)) != NULL)
503 		goto done;
504 	if ((flags & MEMREMAP_WC) &&
505 	    (addr = ioremap_wc(offset, size)) != NULL)
506 		goto done;
507 done:
508 	return (addr);
509 }
510 
511 static inline void
512 memunmap(void *addr)
513 {
514 	/* XXX May need to check if this is RAM */
515 	iounmap(addr);
516 }
517 
518 #define	__MTRR_ID_BASE	1
519 int lkpi_arch_phys_wc_add(unsigned long, unsigned long);
520 void lkpi_arch_phys_wc_del(int);
521 #define	arch_phys_wc_add(...)	lkpi_arch_phys_wc_add(__VA_ARGS__)
522 #define	arch_phys_wc_del(...)	lkpi_arch_phys_wc_del(__VA_ARGS__)
523 #define	arch_phys_wc_index(x)	\
524 	(((x) < __MTRR_ID_BASE) ? -1 : ((x) - __MTRR_ID_BASE))
525 
526 #if defined(__amd64__) || defined(__i386__) || defined(__aarch64__) || defined(__powerpc__) || defined(__riscv)
527 static inline int
528 arch_io_reserve_memtype_wc(resource_size_t start, resource_size_t size)
529 {
530 
531 	return (set_memory_wc(start, size >> PAGE_SHIFT));
532 }
533 
534 static inline void
535 arch_io_free_memtype_wc(resource_size_t start, resource_size_t size)
536 {
537 	set_memory_wb(start, size >> PAGE_SHIFT);
538 }
539 #endif
540 
541 #endif	/* _LINUXKPI_LINUX_IO_H_ */
542