1 /* SPDX-License-Identifier: GPL-2.0-only */ 2 /* 3 * Copyright (C) 2014 Regents of the University of California 4 */ 5 6 #ifndef _ASM_RISCV_CMPXCHG_H 7 #define _ASM_RISCV_CMPXCHG_H 8 9 #include <linux/bug.h> 10 11 #include <asm/barrier.h> 12 #include <asm/fence.h> 13 14 #define __xchg_relaxed(ptr, new, size) \ 15 ({ \ 16 __typeof__(ptr) __ptr = (ptr); \ 17 __typeof__(new) __new = (new); \ 18 __typeof__(*(ptr)) __ret; \ 19 switch (size) { \ 20 case 4: \ 21 __asm__ __volatile__ ( \ 22 " amoswap.w %0, %2, %1\n" \ 23 : "=r" (__ret), "+A" (*__ptr) \ 24 : "r" (__new) \ 25 : "memory"); \ 26 break; \ 27 case 8: \ 28 __asm__ __volatile__ ( \ 29 " amoswap.d %0, %2, %1\n" \ 30 : "=r" (__ret), "+A" (*__ptr) \ 31 : "r" (__new) \ 32 : "memory"); \ 33 break; \ 34 default: \ 35 BUILD_BUG(); \ 36 } \ 37 __ret; \ 38 }) 39 40 #define arch_xchg_relaxed(ptr, x) \ 41 ({ \ 42 __typeof__(*(ptr)) _x_ = (x); \ 43 (__typeof__(*(ptr))) __xchg_relaxed((ptr), \ 44 _x_, sizeof(*(ptr))); \ 45 }) 46 47 #define __xchg_acquire(ptr, new, size) \ 48 ({ \ 49 __typeof__(ptr) __ptr = (ptr); \ 50 __typeof__(new) __new = (new); \ 51 __typeof__(*(ptr)) __ret; \ 52 switch (size) { \ 53 case 4: \ 54 __asm__ __volatile__ ( \ 55 " amoswap.w %0, %2, %1\n" \ 56 RISCV_ACQUIRE_BARRIER \ 57 : "=r" (__ret), "+A" (*__ptr) \ 58 : "r" (__new) \ 59 : "memory"); \ 60 break; \ 61 case 8: \ 62 __asm__ __volatile__ ( \ 63 " amoswap.d %0, %2, %1\n" \ 64 RISCV_ACQUIRE_BARRIER \ 65 : "=r" (__ret), "+A" (*__ptr) \ 66 : "r" (__new) \ 67 : "memory"); \ 68 break; \ 69 default: \ 70 BUILD_BUG(); \ 71 } \ 72 __ret; \ 73 }) 74 75 #define arch_xchg_acquire(ptr, x) \ 76 ({ \ 77 __typeof__(*(ptr)) _x_ = (x); \ 78 (__typeof__(*(ptr))) __xchg_acquire((ptr), \ 79 _x_, sizeof(*(ptr))); \ 80 }) 81 82 #define __xchg_release(ptr, new, size) \ 83 ({ \ 84 __typeof__(ptr) __ptr = (ptr); \ 85 __typeof__(new) __new = (new); \ 86 __typeof__(*(ptr)) __ret; \ 87 switch (size) { \ 88 case 4: \ 89 __asm__ __volatile__ ( \ 90 RISCV_RELEASE_BARRIER \ 91 " amoswap.w %0, %2, %1\n" \ 92 : "=r" (__ret), "+A" (*__ptr) \ 93 : "r" (__new) \ 94 : "memory"); \ 95 break; \ 96 case 8: \ 97 __asm__ __volatile__ ( \ 98 RISCV_RELEASE_BARRIER \ 99 " amoswap.d %0, %2, %1\n" \ 100 : "=r" (__ret), "+A" (*__ptr) \ 101 : "r" (__new) \ 102 : "memory"); \ 103 break; \ 104 default: \ 105 BUILD_BUG(); \ 106 } \ 107 __ret; \ 108 }) 109 110 #define arch_xchg_release(ptr, x) \ 111 ({ \ 112 __typeof__(*(ptr)) _x_ = (x); \ 113 (__typeof__(*(ptr))) __xchg_release((ptr), \ 114 _x_, sizeof(*(ptr))); \ 115 }) 116 117 #define __arch_xchg(ptr, new, size) \ 118 ({ \ 119 __typeof__(ptr) __ptr = (ptr); \ 120 __typeof__(new) __new = (new); \ 121 __typeof__(*(ptr)) __ret; \ 122 switch (size) { \ 123 case 4: \ 124 __asm__ __volatile__ ( \ 125 " amoswap.w.aqrl %0, %2, %1\n" \ 126 : "=r" (__ret), "+A" (*__ptr) \ 127 : "r" (__new) \ 128 : "memory"); \ 129 break; \ 130 case 8: \ 131 __asm__ __volatile__ ( \ 132 " amoswap.d.aqrl %0, %2, %1\n" \ 133 : "=r" (__ret), "+A" (*__ptr) \ 134 : "r" (__new) \ 135 : "memory"); \ 136 break; \ 137 default: \ 138 BUILD_BUG(); \ 139 } \ 140 __ret; \ 141 }) 142 143 #define arch_xchg(ptr, x) \ 144 ({ \ 145 __typeof__(*(ptr)) _x_ = (x); \ 146 (__typeof__(*(ptr))) __arch_xchg((ptr), _x_, sizeof(*(ptr))); \ 147 }) 148 149 #define xchg32(ptr, x) \ 150 ({ \ 151 BUILD_BUG_ON(sizeof(*(ptr)) != 4); \ 152 arch_xchg((ptr), (x)); \ 153 }) 154 155 #define xchg64(ptr, x) \ 156 ({ \ 157 BUILD_BUG_ON(sizeof(*(ptr)) != 8); \ 158 arch_xchg((ptr), (x)); \ 159 }) 160 161 /* 162 * Atomic compare and exchange. Compare OLD with MEM, if identical, 163 * store NEW in MEM. Return the initial value in MEM. Success is 164 * indicated by comparing RETURN with OLD. 165 */ 166 #define __cmpxchg_relaxed(ptr, old, new, size) \ 167 ({ \ 168 __typeof__(ptr) __ptr = (ptr); \ 169 __typeof__(*(ptr)) __old = (old); \ 170 __typeof__(*(ptr)) __new = (new); \ 171 __typeof__(*(ptr)) __ret; \ 172 register unsigned int __rc; \ 173 switch (size) { \ 174 case 4: \ 175 __asm__ __volatile__ ( \ 176 "0: lr.w %0, %2\n" \ 177 " bne %0, %z3, 1f\n" \ 178 " sc.w %1, %z4, %2\n" \ 179 " bnez %1, 0b\n" \ 180 "1:\n" \ 181 : "=&r" (__ret), "=&r" (__rc), "+A" (*__ptr) \ 182 : "rJ" ((long)__old), "rJ" (__new) \ 183 : "memory"); \ 184 break; \ 185 case 8: \ 186 __asm__ __volatile__ ( \ 187 "0: lr.d %0, %2\n" \ 188 " bne %0, %z3, 1f\n" \ 189 " sc.d %1, %z4, %2\n" \ 190 " bnez %1, 0b\n" \ 191 "1:\n" \ 192 : "=&r" (__ret), "=&r" (__rc), "+A" (*__ptr) \ 193 : "rJ" (__old), "rJ" (__new) \ 194 : "memory"); \ 195 break; \ 196 default: \ 197 BUILD_BUG(); \ 198 } \ 199 __ret; \ 200 }) 201 202 #define arch_cmpxchg_relaxed(ptr, o, n) \ 203 ({ \ 204 __typeof__(*(ptr)) _o_ = (o); \ 205 __typeof__(*(ptr)) _n_ = (n); \ 206 (__typeof__(*(ptr))) __cmpxchg_relaxed((ptr), \ 207 _o_, _n_, sizeof(*(ptr))); \ 208 }) 209 210 #define __cmpxchg_acquire(ptr, old, new, size) \ 211 ({ \ 212 __typeof__(ptr) __ptr = (ptr); \ 213 __typeof__(*(ptr)) __old = (old); \ 214 __typeof__(*(ptr)) __new = (new); \ 215 __typeof__(*(ptr)) __ret; \ 216 register unsigned int __rc; \ 217 switch (size) { \ 218 case 4: \ 219 __asm__ __volatile__ ( \ 220 "0: lr.w %0, %2\n" \ 221 " bne %0, %z3, 1f\n" \ 222 " sc.w %1, %z4, %2\n" \ 223 " bnez %1, 0b\n" \ 224 RISCV_ACQUIRE_BARRIER \ 225 "1:\n" \ 226 : "=&r" (__ret), "=&r" (__rc), "+A" (*__ptr) \ 227 : "rJ" ((long)__old), "rJ" (__new) \ 228 : "memory"); \ 229 break; \ 230 case 8: \ 231 __asm__ __volatile__ ( \ 232 "0: lr.d %0, %2\n" \ 233 " bne %0, %z3, 1f\n" \ 234 " sc.d %1, %z4, %2\n" \ 235 " bnez %1, 0b\n" \ 236 RISCV_ACQUIRE_BARRIER \ 237 "1:\n" \ 238 : "=&r" (__ret), "=&r" (__rc), "+A" (*__ptr) \ 239 : "rJ" (__old), "rJ" (__new) \ 240 : "memory"); \ 241 break; \ 242 default: \ 243 BUILD_BUG(); \ 244 } \ 245 __ret; \ 246 }) 247 248 #define arch_cmpxchg_acquire(ptr, o, n) \ 249 ({ \ 250 __typeof__(*(ptr)) _o_ = (o); \ 251 __typeof__(*(ptr)) _n_ = (n); \ 252 (__typeof__(*(ptr))) __cmpxchg_acquire((ptr), \ 253 _o_, _n_, sizeof(*(ptr))); \ 254 }) 255 256 #define __cmpxchg_release(ptr, old, new, size) \ 257 ({ \ 258 __typeof__(ptr) __ptr = (ptr); \ 259 __typeof__(*(ptr)) __old = (old); \ 260 __typeof__(*(ptr)) __new = (new); \ 261 __typeof__(*(ptr)) __ret; \ 262 register unsigned int __rc; \ 263 switch (size) { \ 264 case 4: \ 265 __asm__ __volatile__ ( \ 266 RISCV_RELEASE_BARRIER \ 267 "0: lr.w %0, %2\n" \ 268 " bne %0, %z3, 1f\n" \ 269 " sc.w %1, %z4, %2\n" \ 270 " bnez %1, 0b\n" \ 271 "1:\n" \ 272 : "=&r" (__ret), "=&r" (__rc), "+A" (*__ptr) \ 273 : "rJ" ((long)__old), "rJ" (__new) \ 274 : "memory"); \ 275 break; \ 276 case 8: \ 277 __asm__ __volatile__ ( \ 278 RISCV_RELEASE_BARRIER \ 279 "0: lr.d %0, %2\n" \ 280 " bne %0, %z3, 1f\n" \ 281 " sc.d %1, %z4, %2\n" \ 282 " bnez %1, 0b\n" \ 283 "1:\n" \ 284 : "=&r" (__ret), "=&r" (__rc), "+A" (*__ptr) \ 285 : "rJ" (__old), "rJ" (__new) \ 286 : "memory"); \ 287 break; \ 288 default: \ 289 BUILD_BUG(); \ 290 } \ 291 __ret; \ 292 }) 293 294 #define arch_cmpxchg_release(ptr, o, n) \ 295 ({ \ 296 __typeof__(*(ptr)) _o_ = (o); \ 297 __typeof__(*(ptr)) _n_ = (n); \ 298 (__typeof__(*(ptr))) __cmpxchg_release((ptr), \ 299 _o_, _n_, sizeof(*(ptr))); \ 300 }) 301 302 #define __cmpxchg(ptr, old, new, size) \ 303 ({ \ 304 __typeof__(ptr) __ptr = (ptr); \ 305 __typeof__(*(ptr)) __old = (old); \ 306 __typeof__(*(ptr)) __new = (new); \ 307 __typeof__(*(ptr)) __ret; \ 308 register unsigned int __rc; \ 309 switch (size) { \ 310 case 4: \ 311 __asm__ __volatile__ ( \ 312 "0: lr.w %0, %2\n" \ 313 " bne %0, %z3, 1f\n" \ 314 " sc.w.rl %1, %z4, %2\n" \ 315 " bnez %1, 0b\n" \ 316 " fence rw, rw\n" \ 317 "1:\n" \ 318 : "=&r" (__ret), "=&r" (__rc), "+A" (*__ptr) \ 319 : "rJ" ((long)__old), "rJ" (__new) \ 320 : "memory"); \ 321 break; \ 322 case 8: \ 323 __asm__ __volatile__ ( \ 324 "0: lr.d %0, %2\n" \ 325 " bne %0, %z3, 1f\n" \ 326 " sc.d.rl %1, %z4, %2\n" \ 327 " bnez %1, 0b\n" \ 328 " fence rw, rw\n" \ 329 "1:\n" \ 330 : "=&r" (__ret), "=&r" (__rc), "+A" (*__ptr) \ 331 : "rJ" (__old), "rJ" (__new) \ 332 : "memory"); \ 333 break; \ 334 default: \ 335 BUILD_BUG(); \ 336 } \ 337 __ret; \ 338 }) 339 340 #define arch_cmpxchg(ptr, o, n) \ 341 ({ \ 342 __typeof__(*(ptr)) _o_ = (o); \ 343 __typeof__(*(ptr)) _n_ = (n); \ 344 (__typeof__(*(ptr))) __cmpxchg((ptr), \ 345 _o_, _n_, sizeof(*(ptr))); \ 346 }) 347 348 #define arch_cmpxchg_local(ptr, o, n) \ 349 (__cmpxchg_relaxed((ptr), (o), (n), sizeof(*(ptr)))) 350 351 #define arch_cmpxchg64(ptr, o, n) \ 352 ({ \ 353 BUILD_BUG_ON(sizeof(*(ptr)) != 8); \ 354 arch_cmpxchg((ptr), (o), (n)); \ 355 }) 356 357 #define arch_cmpxchg64_local(ptr, o, n) \ 358 ({ \ 359 BUILD_BUG_ON(sizeof(*(ptr)) != 8); \ 360 arch_cmpxchg_relaxed((ptr), (o), (n)); \ 361 }) 362 363 #endif /* _ASM_RISCV_CMPXCHG_H */ 364