1 /* 2 * Copyright 2010 Samy Al Bahra. 3 * All rights reserved. 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 1. Redistributions of source code must retain the above copyright 9 * notice, this list of conditions and the following disclaimer. 10 * 2. Redistributions in binary form must reproduce the above copyright 11 * notice, this list of conditions and the following disclaimer in the 12 * documentation and/or other materials provided with the distribution. 13 * 14 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 15 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 16 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 17 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 18 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 19 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 20 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 21 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 22 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 23 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 24 * SUCH DAMAGE. 25 */ 26 27 #ifndef CK_PR_GCC_H 28 #define CK_PR_GCC_H 29 30 #ifndef CK_PR_H 31 #error Do not include this file directly, use ck_pr.h 32 #endif 33 34 #include <ck_cc.h> 35 36 CK_CC_INLINE static void 37 ck_pr_barrier(void) 38 { 39 40 __asm__ __volatile__("" ::: "memory"); 41 return; 42 } 43 44 #ifndef CK_F_PR 45 #define CK_F_PR 46 47 #include <ck_stdbool.h> 48 #include <ck_stdint.h> 49 50 /* 51 * The following represent supported atomic operations. 52 * These operations may be emulated. 53 */ 54 #include "ck_f_pr.h" 55 56 #define CK_PR_ACCESS(x) (*(volatile __typeof__(x) *)&(x)) 57 58 #define CK_PR_LOAD(S, M, T) \ 59 CK_CC_INLINE static T \ 60 ck_pr_md_load_##S(const M *target) \ 61 { \ 62 T r; \ 63 ck_pr_barrier(); \ 64 r = CK_PR_ACCESS(*(const T *)target); \ 65 ck_pr_barrier(); \ 66 return (r); \ 67 } \ 68 CK_CC_INLINE static void \ 69 ck_pr_md_store_##S(M *target, T v) \ 70 { \ 71 ck_pr_barrier(); \ 72 CK_PR_ACCESS(*(T *)target) = v; \ 73 ck_pr_barrier(); \ 74 return; \ 75 } 76 77 CK_CC_INLINE static void * 78 ck_pr_md_load_ptr(const void *target) 79 { 80 void *r; 81 82 ck_pr_barrier(); 83 r = CK_CC_DECONST_PTR(*(volatile void *const*)(target)); 84 ck_pr_barrier(); 85 86 return r; 87 } 88 89 CK_CC_INLINE static void 90 ck_pr_md_store_ptr(void *target, const void *v) 91 { 92 93 ck_pr_barrier(); 94 *(volatile void **)target = CK_CC_DECONST_PTR(v); 95 ck_pr_barrier(); 96 return; 97 } 98 99 #define CK_PR_LOAD_S(S, T) CK_PR_LOAD(S, T, T) 100 101 CK_PR_LOAD_S(char, char) 102 CK_PR_LOAD_S(uint, unsigned int) 103 CK_PR_LOAD_S(int, int) 104 #ifndef CK_PR_DISABLE_DOUBLE 105 CK_PR_LOAD_S(double, double) 106 #endif 107 CK_PR_LOAD_S(64, uint64_t) 108 CK_PR_LOAD_S(32, uint32_t) 109 CK_PR_LOAD_S(16, uint16_t) 110 CK_PR_LOAD_S(8, uint8_t) 111 112 #undef CK_PR_LOAD_S 113 #undef CK_PR_LOAD 114 115 CK_CC_INLINE static void 116 ck_pr_stall(void) 117 { 118 119 ck_pr_barrier(); 120 } 121 122 /* 123 * Load and store fences are equivalent to full fences in the GCC port. 124 */ 125 #define CK_PR_FENCE(T) \ 126 CK_CC_INLINE static void \ 127 ck_pr_fence_strict_##T(void) \ 128 { \ 129 __sync_synchronize(); \ 130 } 131 132 CK_PR_FENCE(atomic) 133 CK_PR_FENCE(atomic_atomic) 134 CK_PR_FENCE(atomic_load) 135 CK_PR_FENCE(atomic_store) 136 CK_PR_FENCE(store_atomic) 137 CK_PR_FENCE(load_atomic) 138 CK_PR_FENCE(load) 139 CK_PR_FENCE(load_load) 140 CK_PR_FENCE(load_store) 141 CK_PR_FENCE(store) 142 CK_PR_FENCE(store_store) 143 CK_PR_FENCE(store_load) 144 CK_PR_FENCE(memory) 145 CK_PR_FENCE(acquire) 146 CK_PR_FENCE(release) 147 CK_PR_FENCE(acqrel) 148 CK_PR_FENCE(lock) 149 CK_PR_FENCE(unlock) 150 151 #undef CK_PR_FENCE 152 153 /* 154 * Atomic compare and swap. 155 */ 156 #define CK_PR_CAS(S, M, T) \ 157 CK_CC_INLINE static bool \ 158 ck_pr_cas_##S(M *target, T compare, T set) \ 159 { \ 160 bool z; \ 161 z = __sync_bool_compare_and_swap((T *)target, compare, set); \ 162 return z; \ 163 } 164 165 CK_PR_CAS(ptr, void, void *) 166 167 #define CK_PR_CAS_S(S, T) CK_PR_CAS(S, T, T) 168 169 CK_PR_CAS_S(char, char) 170 CK_PR_CAS_S(int, int) 171 CK_PR_CAS_S(uint, unsigned int) 172 CK_PR_CAS_S(64, uint64_t) 173 CK_PR_CAS_S(32, uint32_t) 174 CK_PR_CAS_S(16, uint16_t) 175 CK_PR_CAS_S(8, uint8_t) 176 177 #undef CK_PR_CAS_S 178 #undef CK_PR_CAS 179 180 /* 181 * Compare and swap, set *v to old value of target. 182 */ 183 CK_CC_INLINE static bool 184 ck_pr_cas_ptr_value(void *target, void *compare, void *set, void *v) 185 { 186 set = __sync_val_compare_and_swap((void **)target, compare, set); 187 *(void **)v = set; 188 return (set == compare); 189 } 190 191 #define CK_PR_CAS_O(S, T) \ 192 CK_CC_INLINE static bool \ 193 ck_pr_cas_##S##_value(T *target, T compare, T set, T *v) \ 194 { \ 195 set = __sync_val_compare_and_swap(target, compare, set);\ 196 *v = set; \ 197 return (set == compare); \ 198 } 199 200 CK_PR_CAS_O(char, char) 201 CK_PR_CAS_O(int, int) 202 CK_PR_CAS_O(uint, unsigned int) 203 CK_PR_CAS_O(64, uint64_t) 204 CK_PR_CAS_O(32, uint32_t) 205 CK_PR_CAS_O(16, uint16_t) 206 CK_PR_CAS_O(8, uint8_t) 207 208 #undef CK_PR_CAS_O 209 210 /* 211 * Atomic fetch-and-add operations. 212 */ 213 #define CK_PR_FAA(S, M, T) \ 214 CK_CC_INLINE static T \ 215 ck_pr_faa_##S(M *target, T d) \ 216 { \ 217 d = __sync_fetch_and_add((T *)target, d); \ 218 return (d); \ 219 } 220 221 CK_PR_FAA(ptr, void, void *) 222 223 #define CK_PR_FAA_S(S, T) CK_PR_FAA(S, T, T) 224 225 CK_PR_FAA_S(char, char) 226 CK_PR_FAA_S(uint, unsigned int) 227 CK_PR_FAA_S(int, int) 228 CK_PR_FAA_S(64, uint64_t) 229 CK_PR_FAA_S(32, uint32_t) 230 CK_PR_FAA_S(16, uint16_t) 231 CK_PR_FAA_S(8, uint8_t) 232 233 #undef CK_PR_FAA_S 234 #undef CK_PR_FAA 235 236 /* 237 * Atomic store-only binary operations. 238 */ 239 #define CK_PR_BINARY(K, S, M, T) \ 240 CK_CC_INLINE static void \ 241 ck_pr_##K##_##S(M *target, T d) \ 242 { \ 243 d = __sync_fetch_and_##K((T *)target, d); \ 244 return; \ 245 } 246 247 #define CK_PR_BINARY_S(K, S, T) CK_PR_BINARY(K, S, T, T) 248 249 #define CK_PR_GENERATE(K) \ 250 CK_PR_BINARY(K, ptr, void, void *) \ 251 CK_PR_BINARY_S(K, char, char) \ 252 CK_PR_BINARY_S(K, int, int) \ 253 CK_PR_BINARY_S(K, uint, unsigned int) \ 254 CK_PR_BINARY_S(K, 64, uint64_t) \ 255 CK_PR_BINARY_S(K, 32, uint32_t) \ 256 CK_PR_BINARY_S(K, 16, uint16_t) \ 257 CK_PR_BINARY_S(K, 8, uint8_t) 258 259 CK_PR_GENERATE(add) 260 CK_PR_GENERATE(sub) 261 CK_PR_GENERATE(and) 262 CK_PR_GENERATE(or) 263 CK_PR_GENERATE(xor) 264 265 #undef CK_PR_GENERATE 266 #undef CK_PR_BINARY_S 267 #undef CK_PR_BINARY 268 269 #define CK_PR_UNARY(S, M, T) \ 270 CK_CC_INLINE static void \ 271 ck_pr_inc_##S(M *target) \ 272 { \ 273 ck_pr_add_##S(target, (T)1); \ 274 return; \ 275 } \ 276 CK_CC_INLINE static void \ 277 ck_pr_dec_##S(M *target) \ 278 { \ 279 ck_pr_sub_##S(target, (T)1); \ 280 return; \ 281 } 282 283 #define CK_PR_UNARY_S(S, M) CK_PR_UNARY(S, M, M) 284 285 CK_PR_UNARY(ptr, void, void *) 286 CK_PR_UNARY_S(char, char) 287 CK_PR_UNARY_S(int, int) 288 CK_PR_UNARY_S(uint, unsigned int) 289 CK_PR_UNARY_S(64, uint64_t) 290 CK_PR_UNARY_S(32, uint32_t) 291 CK_PR_UNARY_S(16, uint16_t) 292 CK_PR_UNARY_S(8, uint8_t) 293 294 #undef CK_PR_UNARY_S 295 #undef CK_PR_UNARY 296 #endif /* !CK_F_PR */ 297 #endif /* CK_PR_GCC_H */ 298