1 /*
2 * Copyright 2010 Samy Al Bahra.
3 * All rights reserved.
4 *
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
7 * are met:
8 * 1. Redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer.
10 * 2. Redistributions in binary form must reproduce the above copyright
11 * notice, this list of conditions and the following disclaimer in the
12 * documentation and/or other materials provided with the distribution.
13 *
14 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
15 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
16 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
17 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
18 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
19 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
20 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
21 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
22 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
23 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
24 * SUCH DAMAGE.
25 */
26
27 #ifndef CK_PR_GCC_H
28 #define CK_PR_GCC_H
29
30 #ifndef CK_PR_H
31 #error Do not include this file directly, use ck_pr.h
32 #endif
33
34 #include <ck_cc.h>
35
36 CK_CC_INLINE static void
ck_pr_barrier(void)37 ck_pr_barrier(void)
38 {
39
40 __asm__ __volatile__("" ::: "memory");
41 return;
42 }
43
44 #ifndef CK_F_PR
45 #define CK_F_PR
46
47 #include <ck_stdbool.h>
48 #include <ck_stdint.h>
49
50 /*
51 * The following represent supported atomic operations.
52 * These operations may be emulated.
53 */
54 #include "ck_f_pr.h"
55
56 #define CK_PR_ACCESS(x) (*(volatile __typeof__(x) *)&(x))
57
58 #define CK_PR_LOAD(S, M, T) \
59 CK_CC_INLINE static T \
60 ck_pr_md_load_##S(const M *target) \
61 { \
62 T r; \
63 ck_pr_barrier(); \
64 r = CK_PR_ACCESS(*(const T *)target); \
65 ck_pr_barrier(); \
66 return (r); \
67 } \
68 CK_CC_INLINE static void \
69 ck_pr_md_store_##S(M *target, T v) \
70 { \
71 ck_pr_barrier(); \
72 CK_PR_ACCESS(*(T *)target) = v; \
73 ck_pr_barrier(); \
74 return; \
75 }
76
77 CK_CC_INLINE static void *
ck_pr_md_load_ptr(const void * target)78 ck_pr_md_load_ptr(const void *target)
79 {
80 void *r;
81
82 ck_pr_barrier();
83 r = CK_CC_DECONST_PTR(*(volatile void *const*)(target));
84 ck_pr_barrier();
85
86 return r;
87 }
88
89 CK_CC_INLINE static void
ck_pr_md_store_ptr(void * target,const void * v)90 ck_pr_md_store_ptr(void *target, const void *v)
91 {
92
93 ck_pr_barrier();
94 *(volatile void **)target = CK_CC_DECONST_PTR(v);
95 ck_pr_barrier();
96 return;
97 }
98
99 #define CK_PR_LOAD_S(S, T) CK_PR_LOAD(S, T, T)
100
CK_PR_LOAD_S(char,char)101 CK_PR_LOAD_S(char, char)
102 CK_PR_LOAD_S(uint, unsigned int)
103 CK_PR_LOAD_S(int, int)
104 #ifndef CK_PR_DISABLE_DOUBLE
105 CK_PR_LOAD_S(double, double)
106 #endif
107 CK_PR_LOAD_S(64, uint64_t)
108 CK_PR_LOAD_S(32, uint32_t)
109 CK_PR_LOAD_S(16, uint16_t)
110 CK_PR_LOAD_S(8, uint8_t)
111
112 #undef CK_PR_LOAD_S
113 #undef CK_PR_LOAD
114
115 CK_CC_INLINE static void
116 ck_pr_stall(void)
117 {
118
119 ck_pr_barrier();
120 }
121
122 /*
123 * Load and store fences are equivalent to full fences in the GCC port.
124 */
125 #define CK_PR_FENCE(T) \
126 CK_CC_INLINE static void \
127 ck_pr_fence_strict_##T(void) \
128 { \
129 __sync_synchronize(); \
130 }
131
132 CK_PR_FENCE(atomic)
CK_PR_FENCE(atomic_atomic)133 CK_PR_FENCE(atomic_atomic)
134 CK_PR_FENCE(atomic_load)
135 CK_PR_FENCE(atomic_store)
136 CK_PR_FENCE(store_atomic)
137 CK_PR_FENCE(load_atomic)
138 CK_PR_FENCE(load)
139 CK_PR_FENCE(load_load)
140 CK_PR_FENCE(load_store)
141 CK_PR_FENCE(store)
142 CK_PR_FENCE(store_store)
143 CK_PR_FENCE(store_load)
144 CK_PR_FENCE(memory)
145 CK_PR_FENCE(acquire)
146 CK_PR_FENCE(release)
147 CK_PR_FENCE(acqrel)
148 CK_PR_FENCE(lock)
149 CK_PR_FENCE(unlock)
150
151 #undef CK_PR_FENCE
152
153 /*
154 * Atomic compare and swap.
155 */
156 #define CK_PR_CAS(S, M, T) \
157 CK_CC_INLINE static bool \
158 ck_pr_cas_##S(M *target, T compare, T set) \
159 { \
160 bool z; \
161 z = __sync_bool_compare_and_swap((T *)target, compare, set); \
162 return z; \
163 }
164
165 CK_PR_CAS(ptr, void, void *)
166
167 #define CK_PR_CAS_S(S, T) CK_PR_CAS(S, T, T)
168
169 CK_PR_CAS_S(char, char)
170 CK_PR_CAS_S(int, int)
171 CK_PR_CAS_S(uint, unsigned int)
172 CK_PR_CAS_S(64, uint64_t)
173 CK_PR_CAS_S(32, uint32_t)
174 CK_PR_CAS_S(16, uint16_t)
175 CK_PR_CAS_S(8, uint8_t)
176
177 #undef CK_PR_CAS_S
178 #undef CK_PR_CAS
179
180 /*
181 * Compare and swap, set *v to old value of target.
182 */
183 CK_CC_INLINE static bool
184 ck_pr_cas_ptr_value(void *target, void *compare, void *set, void *v)
185 {
186 set = __sync_val_compare_and_swap((void **)target, compare, set);
187 *(void **)v = set;
188 return (set == compare);
189 }
190
191 #define CK_PR_CAS_O(S, T) \
192 CK_CC_INLINE static bool \
193 ck_pr_cas_##S##_value(T *target, T compare, T set, T *v) \
194 { \
195 set = __sync_val_compare_and_swap(target, compare, set);\
196 *v = set; \
197 return (set == compare); \
198 }
199
200 CK_PR_CAS_O(char, char)
201 CK_PR_CAS_O(int, int)
202 CK_PR_CAS_O(uint, unsigned int)
203 CK_PR_CAS_O(64, uint64_t)
204 CK_PR_CAS_O(32, uint32_t)
205 CK_PR_CAS_O(16, uint16_t)
206 CK_PR_CAS_O(8, uint8_t)
207
208 #undef CK_PR_CAS_O
209
210 /*
211 * Atomic fetch-and-add operations.
212 */
213 #define CK_PR_FAA(S, M, T) \
214 CK_CC_INLINE static T \
215 ck_pr_faa_##S(M *target, T d) \
216 { \
217 d = __sync_fetch_and_add((T *)target, d); \
218 return (d); \
219 }
220
221 CK_PR_FAA(ptr, void, void *)
222
223 #define CK_PR_FAA_S(S, T) CK_PR_FAA(S, T, T)
224
225 CK_PR_FAA_S(char, char)
226 CK_PR_FAA_S(uint, unsigned int)
227 CK_PR_FAA_S(int, int)
228 CK_PR_FAA_S(64, uint64_t)
229 CK_PR_FAA_S(32, uint32_t)
230 CK_PR_FAA_S(16, uint16_t)
231 CK_PR_FAA_S(8, uint8_t)
232
233 #undef CK_PR_FAA_S
234 #undef CK_PR_FAA
235
236 /*
237 * Atomic store-only binary operations.
238 */
239 #define CK_PR_BINARY(K, S, M, T) \
240 CK_CC_INLINE static void \
241 ck_pr_##K##_##S(M *target, T d) \
242 { \
243 d = __sync_fetch_and_##K((T *)target, d); \
244 return; \
245 }
246
247 #define CK_PR_BINARY_S(K, S, T) CK_PR_BINARY(K, S, T, T)
248
249 #define CK_PR_GENERATE(K) \
250 CK_PR_BINARY(K, ptr, void, void *) \
251 CK_PR_BINARY_S(K, char, char) \
252 CK_PR_BINARY_S(K, int, int) \
253 CK_PR_BINARY_S(K, uint, unsigned int) \
254 CK_PR_BINARY_S(K, 64, uint64_t) \
255 CK_PR_BINARY_S(K, 32, uint32_t) \
256 CK_PR_BINARY_S(K, 16, uint16_t) \
257 CK_PR_BINARY_S(K, 8, uint8_t)
258
259 CK_PR_GENERATE(add)
260 CK_PR_GENERATE(sub)
261 CK_PR_GENERATE(and)
262 CK_PR_GENERATE(or)
263 CK_PR_GENERATE(xor)
264
265 #undef CK_PR_GENERATE
266 #undef CK_PR_BINARY_S
267 #undef CK_PR_BINARY
268
269 #define CK_PR_UNARY(S, M, T) \
270 CK_CC_INLINE static void \
271 ck_pr_inc_##S(M *target) \
272 { \
273 ck_pr_add_##S(target, (T)1); \
274 return; \
275 } \
276 CK_CC_INLINE static void \
277 ck_pr_dec_##S(M *target) \
278 { \
279 ck_pr_sub_##S(target, (T)1); \
280 return; \
281 }
282
283 #define CK_PR_UNARY_S(S, M) CK_PR_UNARY(S, M, M)
284
285 CK_PR_UNARY(ptr, void, void *)
286 CK_PR_UNARY_S(char, char)
287 CK_PR_UNARY_S(int, int)
288 CK_PR_UNARY_S(uint, unsigned int)
289 CK_PR_UNARY_S(64, uint64_t)
290 CK_PR_UNARY_S(32, uint32_t)
291 CK_PR_UNARY_S(16, uint16_t)
292 CK_PR_UNARY_S(8, uint8_t)
293
294 #undef CK_PR_UNARY_S
295 #undef CK_PR_UNARY
296 #endif /* !CK_F_PR */
297 #endif /* CK_PR_GCC_H */
298