1 /**
2 * \file drm_os_freebsd.h
3 * OS abstraction macros.
4 */
5
6 #include <sys/cdefs.h>
7 #ifndef _DRM_OS_FREEBSD_H_
8 #define _DRM_OS_FREEBSD_H_
9
10 #include <sys/fbio.h>
11 #include <sys/smp.h>
12
13 #if _BYTE_ORDER == _BIG_ENDIAN
14 #define __BIG_ENDIAN 4321
15 #else
16 #define __LITTLE_ENDIAN 1234
17 #endif
18
19 #ifdef __LP64__
20 #define BITS_PER_LONG 64
21 #else
22 #define BITS_PER_LONG 32
23 #endif
24
25 #ifndef __user
26 #define __user
27 #endif
28 #ifndef __iomem
29 #define __iomem
30 #endif
31 #ifndef __always_unused
32 #define __always_unused
33 #endif
34 #ifndef __must_check
35 #define __must_check
36 #endif
37 #ifndef __force
38 #define __force
39 #endif
40 #ifndef uninitialized_var
41 #define uninitialized_var(x) x
42 #endif
43
44 #define cpu_to_le16(x) htole16(x)
45 #define le16_to_cpu(x) le16toh(x)
46 #define cpu_to_le32(x) htole32(x)
47 #define le32_to_cpu(x) le32toh(x)
48
49 #define cpu_to_be16(x) htobe16(x)
50 #define be16_to_cpu(x) be16toh(x)
51 #define cpu_to_be32(x) htobe32(x)
52 #define be32_to_cpu(x) be32toh(x)
53 #define be32_to_cpup(x) be32toh(*x)
54
55 typedef vm_paddr_t dma_addr_t;
56 typedef vm_paddr_t resource_size_t;
57 #define wait_queue_head_t atomic_t
58
59 typedef uint64_t u64;
60 typedef uint32_t u32;
61 typedef uint16_t u16;
62 typedef uint8_t u8;
63 typedef int64_t s64;
64 typedef int32_t s32;
65 typedef int16_t s16;
66 typedef int8_t s8;
67 typedef uint16_t __le16;
68 typedef uint32_t __le32;
69 typedef uint64_t __le64;
70 typedef uint16_t __be16;
71 typedef uint32_t __be32;
72 typedef uint64_t __be64;
73
74 #define DRM_IRQ_ARGS void *arg
75 typedef void irqreturn_t;
76 #define IRQ_HANDLED /* nothing */
77 #define IRQ_NONE /* nothing */
78
79 #define __init
80 #define __exit
81
82 #define BUILD_BUG_ON(x) CTASSERT(!(x))
83 #define BUILD_BUG_ON_NOT_POWER_OF_2(x)
84
85 #ifndef WARN
86 #define WARN(condition, format, ...) ({ \
87 int __ret_warn_on = !!(condition); \
88 if (unlikely(__ret_warn_on)) \
89 DRM_ERROR(format, ##__VA_ARGS__); \
90 unlikely(__ret_warn_on); \
91 })
92 #endif
93 #define WARN_ONCE(condition, format, ...) \
94 WARN(condition, format, ##__VA_ARGS__)
95 #define WARN_ON(cond) WARN(cond, "WARN ON: " #cond)
96 #define WARN_ON_SMP(cond) WARN_ON(cond)
97 #define BUG() panic("BUG")
98 #define BUG_ON(cond) KASSERT(!(cond), ("BUG ON: " #cond " -> 0x%jx", (uintmax_t)(cond)))
99 #define unlikely(x) __builtin_expect(!!(x), 0)
100 #define likely(x) __builtin_expect(!!(x), 1)
101 #define container_of(ptr, type, member) ({ \
102 __typeof( ((type *)0)->member ) *__mptr = (ptr); \
103 (type *)( (char *)__mptr - offsetof(type,member) );})
104
105 #define KHZ2PICOS(a) (1000000000UL/(a))
106
107 #define ARRAY_SIZE(x) (sizeof(x)/sizeof(x[0]))
108
109 #define HZ hz
110 #define DRM_HZ hz
111 #define DRM_CURRENTPID curthread->td_proc->p_pid
112 #define DRM_SUSER(p) (priv_check(p, PRIV_DRIVER) == 0)
113 #define udelay(usecs) DELAY(usecs)
114 #define mdelay(msecs) do { int loops = (msecs); \
115 while (loops--) DELAY(1000); \
116 } while (0)
117 #define DRM_UDELAY(udelay) DELAY(udelay)
118 #define drm_msleep(x, msg) pause((msg), ((int64_t)(x)) * hz / 1000)
119 #define DRM_MSLEEP(msecs) drm_msleep((msecs), "drm_msleep")
120 #define get_seconds() time_second
121
122 #define ioread8(addr) *(volatile uint8_t *)((char *)addr)
123 #define ioread16(addr) *(volatile uint16_t *)((char *)addr)
124 #define ioread32(addr) *(volatile uint32_t *)((char *)addr)
125
126 #define iowrite8(data, addr) *(volatile uint8_t *)((char *)addr) = data;
127 #define iowrite16(data, addr) *(volatile uint16_t *)((char *)addr) = data;
128 #define iowrite32(data, addr) *(volatile uint32_t *)((char *)addr) = data;
129
130 #define DRM_READ8(map, offset) \
131 *(volatile u_int8_t *)(((vm_offset_t)(map)->handle) + \
132 (vm_offset_t)(offset))
133 #define DRM_READ16(map, offset) \
134 le16toh(*(volatile u_int16_t *)(((vm_offset_t)(map)->handle) + \
135 (vm_offset_t)(offset)))
136 #define DRM_READ32(map, offset) \
137 le32toh(*(volatile u_int32_t *)(((vm_offset_t)(map)->handle) + \
138 (vm_offset_t)(offset)))
139 #define DRM_READ64(map, offset) \
140 le64toh(*(volatile u_int64_t *)(((vm_offset_t)(map)->handle) + \
141 (vm_offset_t)(offset)))
142 #define DRM_WRITE8(map, offset, val) \
143 *(volatile u_int8_t *)(((vm_offset_t)(map)->handle) + \
144 (vm_offset_t)(offset)) = val
145 #define DRM_WRITE16(map, offset, val) \
146 *(volatile u_int16_t *)(((vm_offset_t)(map)->handle) + \
147 (vm_offset_t)(offset)) = htole16(val)
148 #define DRM_WRITE32(map, offset, val) \
149 *(volatile u_int32_t *)(((vm_offset_t)(map)->handle) + \
150 (vm_offset_t)(offset)) = htole32(val)
151 #define DRM_WRITE64(map, offset, val) \
152 *(volatile u_int64_t *)(((vm_offset_t)(map)->handle) + \
153 (vm_offset_t)(offset)) = htole64(val)
154
155 #if !defined(__arm__)
156 #if defined(__i386__) || defined(__amd64__) || defined(__powerpc__) || defined(__aarch64__)
157 #define DRM_MSG "This code is deprecated. Install the graphics/drm-kmod pkg\n"
158 #else
159 #define DRM_MSG "This code is deprecated."
160 #endif
161
162 #define DRM_OBSOLETE(dev) \
163 do { \
164 device_printf(dev, "=======================================================\n"); \
165 device_printf(dev, DRM_MSG); \
166 device_printf(dev, "=======================================================\n"); \
167 gone_in_dev(dev, 13, "drm2 drivers"); \
168 } while (0)
169 #endif /* __arm__ */
170
171 /* DRM_READMEMORYBARRIER() prevents reordering of reads.
172 * DRM_WRITEMEMORYBARRIER() prevents reordering of writes.
173 * DRM_MEMORYBARRIER() prevents reordering of reads and writes.
174 */
175 #define DRM_READMEMORYBARRIER() rmb()
176 #define DRM_WRITEMEMORYBARRIER() wmb()
177 #define DRM_MEMORYBARRIER() mb()
178 #define smp_rmb() rmb()
179 #define smp_wmb() wmb()
180 #define smp_mb__before_atomic_inc() mb()
181 #define smp_mb__after_atomic_inc() mb()
182 #define barrier() __compiler_membar()
183
184 #define do_div(a, b) ((a) /= (b))
185 #define div64_u64(a, b) ((a) / (b))
186 #define lower_32_bits(n) ((u32)(n))
187 #define upper_32_bits(n) ((u32)(((n) >> 16) >> 16))
188
189 #define __set_bit(n, s) set_bit((n), (s))
190 #define __clear_bit(n, s) clear_bit((n), (s))
191
192 #define min_t(type, x, y) ({ \
193 type __min1 = (x); \
194 type __min2 = (y); \
195 __min1 < __min2 ? __min1 : __min2; })
196
197 #define max_t(type, x, y) ({ \
198 type __max1 = (x); \
199 type __max2 = (y); \
200 __max1 > __max2 ? __max1 : __max2; })
201
202 #define memset_io(a, b, c) memset((a), (b), (c))
203 #define memcpy_fromio(a, b, c) memcpy((a), (b), (c))
204 #define memcpy_toio(a, b, c) memcpy((a), (b), (c))
205
206 #define VERIFY_READ VM_PROT_READ
207 #define VERIFY_WRITE VM_PROT_WRITE
208 #define access_ok(prot, p, l) useracc((p), (l), (prot))
209
210 /* XXXKIB what is the right code for the FreeBSD ? */
211 /* kib@ used ENXIO here -- dumbbell@ */
212 #define EREMOTEIO EIO
213 #define ERESTARTSYS 512 /* Same value as Linux. */
214
215 #define KTR_DRM KTR_DEV
216 #define KTR_DRM_REG KTR_SPARE3
217
218 #define DRM_AGP_KERN struct agp_info
219 #define DRM_AGP_MEM void
220
221 #define PCI_VENDOR_ID_APPLE 0x106b
222 #define PCI_VENDOR_ID_ASUSTEK 0x1043
223 #define PCI_VENDOR_ID_ATI 0x1002
224 #define PCI_VENDOR_ID_DELL 0x1028
225 #define PCI_VENDOR_ID_HP 0x103c
226 #define PCI_VENDOR_ID_IBM 0x1014
227 #define PCI_VENDOR_ID_INTEL 0x8086
228 #define PCI_VENDOR_ID_SERVERWORKS 0x1166
229 #define PCI_VENDOR_ID_SONY 0x104d
230 #define PCI_VENDOR_ID_VIA 0x1106
231
232 #define DIV_ROUND_UP(n,d) (((n) + (d) - 1) / (d))
233 #define DIV_ROUND_CLOSEST(n,d) (((n) + (d) / 2) / (d))
234 #define div_u64(n, d) ((n) / (d))
235 #define hweight32(i) bitcount32(i)
236
237 /**
238 * ror32 - rotate a 32-bit value right
239 * @word: value to rotate
240 * @shift: bits to roll
241 *
242 * Source: include/linux/bitops.h
243 */
244 static inline uint32_t
ror32(uint32_t word,unsigned int shift)245 ror32(uint32_t word, unsigned int shift)
246 {
247
248 return (word >> shift) | (word << (32 - shift));
249 }
250
251 #define IS_ALIGNED(x, y) (((x) & ((y) - 1)) == 0)
252 #define round_down(x, y) rounddown2((x), (y))
253 #define round_up(x, y) roundup2((x), (y))
254 #define get_unaligned(ptr) \
255 ({ __typeof__(*(ptr)) __tmp; \
256 memcpy(&__tmp, (ptr), sizeof(*(ptr))); __tmp; })
257
258 #if _BYTE_ORDER == _LITTLE_ENDIAN
259 /* Taken from linux/include/linux/unaligned/le_struct.h. */
260 struct __una_u32 { u32 x; } __packed;
261
262 static inline u32
__get_unaligned_cpu32(const void * p)263 __get_unaligned_cpu32(const void *p)
264 {
265 const struct __una_u32 *ptr = (const struct __una_u32 *)p;
266
267 return (ptr->x);
268 }
269
270 static inline u32
get_unaligned_le32(const void * p)271 get_unaligned_le32(const void *p)
272 {
273
274 return (__get_unaligned_cpu32((const u8 *)p));
275 }
276 #else
277 /* Taken from linux/include/linux/unaligned/le_byteshift.h. */
278 static inline u32
__get_unaligned_le32(const u8 * p)279 __get_unaligned_le32(const u8 *p)
280 {
281
282 return (p[0] | p[1] << 8 | p[2] << 16 | p[3] << 24);
283 }
284
285 static inline u32
get_unaligned_le32(const void * p)286 get_unaligned_le32(const void *p)
287 {
288
289 return (__get_unaligned_le32((const u8 *)p));
290 }
291 #endif
292
293 int64_t timeval_to_ns(const struct timeval *tv);
294 struct timeval ns_to_timeval(const int64_t nsec);
295
296 #define PAGE_ALIGN(addr) round_page(addr)
297 #define page_to_phys(x) VM_PAGE_TO_PHYS(x)
298 #define offset_in_page(x) ((x) & PAGE_MASK)
299
300 #define drm_get_device_from_kdev(_kdev) (((struct drm_minor *)(_kdev)->si_drv1)->dev)
301
302 #define DRM_IOC_VOID IOC_VOID
303 #define DRM_IOC_READ IOC_OUT
304 #define DRM_IOC_WRITE IOC_IN
305 #define DRM_IOC_READWRITE IOC_INOUT
306 #define DRM_IOC(dir, group, nr, size) _IOC(dir, group, nr, size)
307
308 static inline long
__copy_to_user(void __user * to,const void * from,unsigned long n)309 __copy_to_user(void __user *to, const void *from, unsigned long n)
310 {
311 return (copyout(from, to, n) != 0 ? n : 0);
312 }
313 #define copy_to_user(to, from, n) __copy_to_user((to), (from), (n))
314
315 static inline int
__put_user(size_t size,void * ptr,void * x)316 __put_user(size_t size, void *ptr, void *x)
317 {
318
319 size = copy_to_user(ptr, x, size);
320
321 return (size ? -EFAULT : size);
322 }
323 #define put_user(x, ptr) __put_user(sizeof(*ptr), (ptr), &(x))
324
325 static inline unsigned long
__copy_from_user(void * to,const void __user * from,unsigned long n)326 __copy_from_user(void *to, const void __user *from, unsigned long n)
327 {
328 return ((copyin(__DECONST(void *, from), to, n) != 0 ? n : 0));
329 }
330 #define copy_from_user(to, from, n) __copy_from_user((to), (from), (n))
331
332 static inline int
__get_user(size_t size,const void * ptr,void * x)333 __get_user(size_t size, const void *ptr, void *x)
334 {
335
336 size = copy_from_user(x, ptr, size);
337
338 return (size ? -EFAULT : size);
339 }
340 #define get_user(x, ptr) __get_user(sizeof(*ptr), (ptr), &(x))
341
342 static inline int
__copy_to_user_inatomic(void __user * to,const void * from,unsigned n)343 __copy_to_user_inatomic(void __user *to, const void *from, unsigned n)
344 {
345
346 return (copyout_nofault(from, to, n) != 0 ? n : 0);
347 }
348 #define __copy_to_user_inatomic_nocache(to, from, n) \
349 __copy_to_user_inatomic((to), (from), (n))
350
351 static inline unsigned long
__copy_from_user_inatomic(void * to,const void __user * from,unsigned long n)352 __copy_from_user_inatomic(void *to, const void __user *from,
353 unsigned long n)
354 {
355
356 /*
357 * XXXKIB. Equivalent Linux function is implemented using
358 * MOVNTI for aligned moves. For unaligned head and tail,
359 * normal move is performed. As such, it is not incorrect, if
360 * only somewhat slower, to use normal copyin. All uses
361 * except shmem_pwrite_fast() have the destination mapped WC.
362 */
363 return ((copyin_nofault(__DECONST(void *, from), to, n) != 0 ? n : 0));
364 }
365 #define __copy_from_user_inatomic_nocache(to, from, n) \
366 __copy_from_user_inatomic((to), (from), (n))
367
368 static inline int
fault_in_multipages_readable(const char __user * uaddr,int size)369 fault_in_multipages_readable(const char __user *uaddr, int size)
370 {
371 char c;
372 int ret = 0;
373 const char __user *end = uaddr + size - 1;
374
375 if (unlikely(size == 0))
376 return ret;
377
378 while (uaddr <= end) {
379 ret = -copyin(uaddr, &c, 1);
380 if (ret != 0)
381 return -EFAULT;
382 uaddr += PAGE_SIZE;
383 }
384
385 /* Check whether the range spilled into the next page. */
386 if (((unsigned long)uaddr & ~PAGE_MASK) ==
387 ((unsigned long)end & ~PAGE_MASK)) {
388 ret = -copyin(end, &c, 1);
389 }
390
391 return ret;
392 }
393
394 static inline int
fault_in_multipages_writeable(char __user * uaddr,int size)395 fault_in_multipages_writeable(char __user *uaddr, int size)
396 {
397 int ret = 0;
398 char __user *end = uaddr + size - 1;
399
400 if (unlikely(size == 0))
401 return ret;
402
403 /*
404 * Writing zeroes into userspace here is OK, because we know that if
405 * the zero gets there, we'll be overwriting it.
406 */
407 while (uaddr <= end) {
408 ret = subyte(uaddr, 0);
409 if (ret != 0)
410 return -EFAULT;
411 uaddr += PAGE_SIZE;
412 }
413
414 /* Check whether the range spilled into the next page. */
415 if (((unsigned long)uaddr & ~PAGE_MASK) ==
416 ((unsigned long)end & ~PAGE_MASK))
417 ret = subyte(end, 0);
418
419 return ret;
420 }
421
422 enum __drm_capabilities {
423 CAP_SYS_ADMIN
424 };
425
426 static inline bool
capable(enum __drm_capabilities cap)427 capable(enum __drm_capabilities cap)
428 {
429
430 switch (cap) {
431 case CAP_SYS_ADMIN:
432 return DRM_SUSER(curthread);
433 default:
434 panic("%s: unhandled capability: %0x", __func__, cap);
435 return (false);
436 }
437 }
438
439 #define to_user_ptr(x) ((void *)(uintptr_t)(x))
440 #define sigemptyset(set) SIGEMPTYSET(set)
441 #define sigaddset(set, sig) SIGADDSET(set, sig)
442
443 #define DRM_LOCK(dev) sx_xlock(&(dev)->dev_struct_lock)
444 #define DRM_UNLOCK(dev) sx_xunlock(&(dev)->dev_struct_lock)
445
446 extern unsigned long drm_linux_timer_hz_mask;
447 #define jiffies ticks
448 #define jiffies_to_msecs(x) (((int64_t)(x)) * 1000 / hz)
449 #define msecs_to_jiffies(x) (((int64_t)(x)) * hz / 1000)
450 #define timespec_to_jiffies(x) (((x)->tv_sec * 1000000 + (x)->tv_nsec) * hz / 1000000)
451 #define time_after(a,b) ((long)(b) - (long)(a) < 0)
452 #define time_after_eq(a,b) ((long)(b) - (long)(a) <= 0)
453 #define round_jiffies(j) ((unsigned long)(((j) + drm_linux_timer_hz_mask) & ~drm_linux_timer_hz_mask))
454 #define round_jiffies_up(j) round_jiffies(j) /* TODO */
455 #define round_jiffies_up_relative(j) round_jiffies_up(j) /* TODO */
456
457 #define getrawmonotonic(ts) getnanouptime(ts)
458
459 #define wake_up(queue) wakeup_one((void *)queue)
460 #define wake_up_interruptible(queue) wakeup_one((void *)queue)
461 #define wake_up_all(queue) wakeup((void *)queue)
462 #define wake_up_interruptible_all(queue) wakeup((void *)queue)
463
464 struct completion {
465 unsigned int done;
466 struct mtx lock;
467 };
468
469 #define INIT_COMPLETION(c) ((c).done = 0);
470
471 static inline void
init_completion(struct completion * c)472 init_completion(struct completion *c)
473 {
474
475 mtx_init(&c->lock, "drmcompl", NULL, MTX_DEF);
476 c->done = 0;
477 }
478
479 static inline void
free_completion(struct completion * c)480 free_completion(struct completion *c)
481 {
482
483 mtx_destroy(&c->lock);
484 }
485
486 static inline void
complete_all(struct completion * c)487 complete_all(struct completion *c)
488 {
489
490 mtx_lock(&c->lock);
491 c->done++;
492 mtx_unlock(&c->lock);
493 wakeup(c);
494 }
495
496 static inline long
wait_for_completion_interruptible_timeout(struct completion * c,unsigned long timeout)497 wait_for_completion_interruptible_timeout(struct completion *c,
498 unsigned long timeout)
499 {
500 unsigned long start_jiffies, elapsed_jiffies;
501 bool timeout_expired = false, awakened = false;
502 long ret = timeout;
503
504 start_jiffies = ticks;
505
506 mtx_lock(&c->lock);
507 while (c->done == 0 && !timeout_expired) {
508 ret = -msleep(c, &c->lock, PCATCH, "drmwco", timeout);
509 switch(ret) {
510 case -EWOULDBLOCK:
511 timeout_expired = true;
512 ret = 0;
513 break;
514 case -EINTR:
515 case -ERESTART:
516 ret = -ERESTARTSYS;
517 break;
518 case 0:
519 awakened = true;
520 break;
521 }
522 }
523 mtx_unlock(&c->lock);
524
525 if (awakened) {
526 elapsed_jiffies = ticks - start_jiffies;
527 ret = timeout > elapsed_jiffies ? timeout - elapsed_jiffies : 1;
528 }
529
530 return (ret);
531 }
532
533 MALLOC_DECLARE(DRM_MEM_DMA);
534 MALLOC_DECLARE(DRM_MEM_SAREA);
535 MALLOC_DECLARE(DRM_MEM_DRIVER);
536 MALLOC_DECLARE(DRM_MEM_MAGIC);
537 MALLOC_DECLARE(DRM_MEM_MINOR);
538 MALLOC_DECLARE(DRM_MEM_IOCTLS);
539 MALLOC_DECLARE(DRM_MEM_MAPS);
540 MALLOC_DECLARE(DRM_MEM_BUFS);
541 MALLOC_DECLARE(DRM_MEM_SEGS);
542 MALLOC_DECLARE(DRM_MEM_PAGES);
543 MALLOC_DECLARE(DRM_MEM_FILES);
544 MALLOC_DECLARE(DRM_MEM_QUEUES);
545 MALLOC_DECLARE(DRM_MEM_CMDS);
546 MALLOC_DECLARE(DRM_MEM_MAPPINGS);
547 MALLOC_DECLARE(DRM_MEM_BUFLISTS);
548 MALLOC_DECLARE(DRM_MEM_AGPLISTS);
549 MALLOC_DECLARE(DRM_MEM_CTXBITMAP);
550 MALLOC_DECLARE(DRM_MEM_SGLISTS);
551 MALLOC_DECLARE(DRM_MEM_MM);
552 MALLOC_DECLARE(DRM_MEM_HASHTAB);
553 MALLOC_DECLARE(DRM_MEM_KMS);
554 MALLOC_DECLARE(DRM_MEM_VBLANK);
555
556 #define simple_strtol(a, b, c) strtol((a), (b), (c))
557
558 typedef struct drm_pci_id_list
559 {
560 int vendor;
561 int device;
562 long driver_private;
563 char *name;
564 } drm_pci_id_list_t;
565
566 #ifdef __i386__
567 #define CONFIG_X86 1
568 #endif
569 #ifdef __amd64__
570 #define CONFIG_X86 1
571 #define CONFIG_X86_64 1
572 #endif
573 #ifdef __ia64__
574 #define CONFIG_IA64 1
575 #endif
576
577 #if defined(__i386__) || defined(__amd64__)
578 #define CONFIG_ACPI
579 #define CONFIG_DRM_I915_KMS
580 #undef CONFIG_INTEL_IOMMU
581 #endif
582
583 #ifdef COMPAT_FREEBSD32
584 #define CONFIG_COMPAT
585 #endif
586
587 #ifndef __arm__
588 #define CONFIG_AGP 1
589 #define CONFIG_MTRR 1
590 #endif
591
592 #define CONFIG_FB 1
593 extern const char *fb_mode_option;
594
595 #undef CONFIG_DEBUG_FS
596 #undef CONFIG_VGA_CONSOLE
597
598 #define EXPORT_SYMBOL(x)
599 #define EXPORT_SYMBOL_GPL(x)
600 #define MODULE_AUTHOR(author)
601 #define MODULE_DESCRIPTION(desc)
602 #define MODULE_LICENSE(license)
603 #define MODULE_PARM_DESC(name, desc)
604 #define MODULE_DEVICE_TABLE(name, list)
605 #define module_param_named(name, var, type, perm)
606
607 #define printk printf
608 #define pr_err DRM_ERROR
609 #define pr_warn DRM_WARNING
610 #define pr_warn_once DRM_WARNING
611 #define KERN_DEBUG ""
612
613 /* I2C compatibility. */
614 #define I2C_M_RD IIC_M_RD
615 #define I2C_M_WR IIC_M_WR
616 #define I2C_M_NOSTART IIC_M_NOSTART
617
618 struct fb_info * framebuffer_alloc(void);
619 void framebuffer_release(struct fb_info *info);
620
621 #define console_lock()
622 #define console_unlock()
623 #define console_trylock() true
624
625 #define PM_EVENT_SUSPEND 0x0002
626 #define PM_EVENT_QUIESCE 0x0008
627 #define PM_EVENT_PRETHAW PM_EVENT_QUIESCE
628
629 typedef struct pm_message {
630 int event;
631 } pm_message_t;
632
633 static inline int
pci_read_config_byte(device_t kdev,int where,u8 * val)634 pci_read_config_byte(device_t kdev, int where, u8 *val)
635 {
636
637 *val = (u8)pci_read_config(kdev, where, 1);
638 return (0);
639 }
640
641 static inline int
pci_write_config_byte(device_t kdev,int where,u8 val)642 pci_write_config_byte(device_t kdev, int where, u8 val)
643 {
644
645 pci_write_config(kdev, where, val, 1);
646 return (0);
647 }
648
649 static inline int
pci_read_config_word(device_t kdev,int where,uint16_t * val)650 pci_read_config_word(device_t kdev, int where, uint16_t *val)
651 {
652
653 *val = (uint16_t)pci_read_config(kdev, where, 2);
654 return (0);
655 }
656
657 static inline int
pci_write_config_word(device_t kdev,int where,uint16_t val)658 pci_write_config_word(device_t kdev, int where, uint16_t val)
659 {
660
661 pci_write_config(kdev, where, val, 2);
662 return (0);
663 }
664
665 static inline int
pci_read_config_dword(device_t kdev,int where,uint32_t * val)666 pci_read_config_dword(device_t kdev, int where, uint32_t *val)
667 {
668
669 *val = (uint32_t)pci_read_config(kdev, where, 4);
670 return (0);
671 }
672
673 static inline int
pci_write_config_dword(device_t kdev,int where,uint32_t val)674 pci_write_config_dword(device_t kdev, int where, uint32_t val)
675 {
676
677 pci_write_config(kdev, where, val, 4);
678 return (0);
679 }
680
681 static inline void
on_each_cpu(void callback (void * data),void * data,int wait)682 on_each_cpu(void callback(void *data), void *data, int wait)
683 {
684
685 smp_rendezvous(NULL, callback, NULL, data);
686 }
687
688 void hex_dump_to_buffer(const void *buf, size_t len, int rowsize,
689 int groupsize, char *linebuf, size_t linebuflen, bool ascii);
690
691 #define KIB_NOTYET() \
692 do { \
693 if (drm_debug && drm_notyet) \
694 printf("NOTYET: %s at %s:%d\n", __func__, __FILE__, __LINE__); \
695 } while (0)
696
697 #endif /* _DRM_OS_FREEBSD_H_ */
698