xref: /linux/arch/arm/include/asm/div64.h (revision e5c86679d5e864947a52fb31e45a425dea3e7fa9)
1 #ifndef __ASM_ARM_DIV64
2 #define __ASM_ARM_DIV64
3 
4 #include <linux/types.h>
5 #include <asm/compiler.h>
6 
7 /*
8  * The semantics of __div64_32() are:
9  *
10  * uint32_t __div64_32(uint64_t *n, uint32_t base)
11  * {
12  * 	uint32_t remainder = *n % base;
13  * 	*n = *n / base;
14  * 	return remainder;
15  * }
16  *
17  * In other words, a 64-bit dividend with a 32-bit divisor producing
18  * a 64-bit result and a 32-bit remainder.  To accomplish this optimally
19  * we override the generic version in lib/div64.c to call our __do_div64
20  * assembly implementation with completely non standard calling convention
21  * for arguments and results (beware).
22  */
23 
24 #ifdef __ARMEB__
25 #define __xh "r0"
26 #define __xl "r1"
27 #else
28 #define __xl "r0"
29 #define __xh "r1"
30 #endif
31 
32 static inline uint32_t __div64_32(uint64_t *n, uint32_t base)
33 {
34 	register unsigned int __base      asm("r4") = base;
35 	register unsigned long long __n   asm("r0") = *n;
36 	register unsigned long long __res asm("r2");
37 	register unsigned int __rem       asm(__xh);
38 	asm(	__asmeq("%0", __xh)
39 		__asmeq("%1", "r2")
40 		__asmeq("%2", "r0")
41 		__asmeq("%3", "r4")
42 		"bl	__do_div64"
43 		: "=r" (__rem), "=r" (__res)
44 		: "r" (__n), "r" (__base)
45 		: "ip", "lr", "cc");
46 	*n = __res;
47 	return __rem;
48 }
49 #define __div64_32 __div64_32
50 
51 #if !defined(CONFIG_AEABI)
52 
53 /*
54  * In OABI configurations, some uses of the do_div function
55  * cause gcc to run out of registers. To work around that,
56  * we can force the use of the out-of-line version for
57  * configurations that build a OABI kernel.
58  */
59 #define do_div(n, base) __div64_32(&(n), base)
60 
61 #else
62 
63 /*
64  * gcc versions earlier than 4.0 are simply too problematic for the
65  * __div64_const32() code in asm-generic/div64.h. First there is
66  * gcc PR 15089 that tend to trig on more complex constructs, spurious
67  * .global __udivsi3 are inserted even if none of those symbols are
68  * referenced in the generated code, and those gcc versions are not able
69  * to do constant propagation on long long values anyway.
70  */
71 
72 #define __div64_const32_is_OK (__GNUC__ >= 4)
73 
74 static inline uint64_t __arch_xprod_64(uint64_t m, uint64_t n, bool bias)
75 {
76 	unsigned long long res;
77 	register unsigned int tmp asm("ip") = 0;
78 
79 	if (!bias) {
80 		asm (	"umull	%Q0, %R0, %Q1, %Q2\n\t"
81 			"mov	%Q0, #0"
82 			: "=&r" (res)
83 			: "r" (m), "r" (n)
84 			: "cc");
85 	} else if (!(m & ((1ULL << 63) | (1ULL << 31)))) {
86 		res = m;
87 		asm (	"umlal	%Q0, %R0, %Q1, %Q2\n\t"
88 			"mov	%Q0, #0"
89 			: "+&r" (res)
90 			: "r" (m), "r" (n)
91 			: "cc");
92 	} else {
93 		asm (	"umull	%Q0, %R0, %Q2, %Q3\n\t"
94 			"cmn	%Q0, %Q2\n\t"
95 			"adcs	%R0, %R0, %R2\n\t"
96 			"adc	%Q0, %1, #0"
97 			: "=&r" (res), "+&r" (tmp)
98 			: "r" (m), "r" (n)
99 			: "cc");
100 	}
101 
102 	if (!(m & ((1ULL << 63) | (1ULL << 31)))) {
103 		asm (	"umlal	%R0, %Q0, %R1, %Q2\n\t"
104 			"umlal	%R0, %Q0, %Q1, %R2\n\t"
105 			"mov	%R0, #0\n\t"
106 			"umlal	%Q0, %R0, %R1, %R2"
107 			: "+&r" (res)
108 			: "r" (m), "r" (n)
109 			: "cc");
110 	} else {
111 		asm (	"umlal	%R0, %Q0, %R2, %Q3\n\t"
112 			"umlal	%R0, %1, %Q2, %R3\n\t"
113 			"mov	%R0, #0\n\t"
114 			"adds	%Q0, %1, %Q0\n\t"
115 			"adc	%R0, %R0, #0\n\t"
116 			"umlal	%Q0, %R0, %R2, %R3"
117 			: "+&r" (res), "+&r" (tmp)
118 			: "r" (m), "r" (n)
119 			: "cc");
120 	}
121 
122 	return res;
123 }
124 #define __arch_xprod_64 __arch_xprod_64
125 
126 #include <asm-generic/div64.h>
127 
128 #endif
129 
130 #endif
131