xref: /linux/arch/arc/include/asm/futex.h (revision ca55b2fef3a9373fcfc30f82fd26bc7fccbda732)
1 /*
2  * Copyright (C) 2004, 2007-2010, 2011-2012 Synopsys, Inc. (www.synopsys.com)
3  *
4  * This program is free software; you can redistribute it and/or modify
5  * it under the terms of the GNU General Public License version 2 as
6  * published by the Free Software Foundation.
7  *
8  * Vineetg: August 2010: From Android kernel work
9  */
10 
11 #ifndef _ASM_FUTEX_H
12 #define _ASM_FUTEX_H
13 
14 #include <linux/futex.h>
15 #include <linux/preempt.h>
16 #include <linux/uaccess.h>
17 #include <asm/errno.h>
18 
19 #ifdef CONFIG_ARC_HAS_LLSC
20 
21 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg)\
22 							\
23 	smp_mb();					\
24 	__asm__ __volatile__(				\
25 	"1:	llock	%1, [%2]		\n"	\
26 		insn				"\n"	\
27 	"2:	scond	%0, [%2]		\n"	\
28 	"	bnz	1b			\n"	\
29 	"	mov %0, 0			\n"	\
30 	"3:					\n"	\
31 	"	.section .fixup,\"ax\"		\n"	\
32 	"	.align  4			\n"	\
33 	"4:	mov %0, %4			\n"	\
34 	"	j   3b				\n"	\
35 	"	.previous			\n"	\
36 	"	.section __ex_table,\"a\"	\n"	\
37 	"	.align  4			\n"	\
38 	"	.word   1b, 4b			\n"	\
39 	"	.word   2b, 4b			\n"	\
40 	"	.previous			\n"	\
41 							\
42 	: "=&r" (ret), "=&r" (oldval)			\
43 	: "r" (uaddr), "r" (oparg), "ir" (-EFAULT)	\
44 	: "cc", "memory");				\
45 	smp_mb()					\
46 
47 #else	/* !CONFIG_ARC_HAS_LLSC */
48 
49 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg)\
50 							\
51 	smp_mb();					\
52 	__asm__ __volatile__(				\
53 	"1:	ld	%1, [%2]		\n"	\
54 		insn				"\n"	\
55 	"2:	st	%0, [%2]		\n"	\
56 	"	mov %0, 0			\n"	\
57 	"3:					\n"	\
58 	"	.section .fixup,\"ax\"		\n"	\
59 	"	.align  4			\n"	\
60 	"4:	mov %0, %4			\n"	\
61 	"	j   3b				\n"	\
62 	"	.previous			\n"	\
63 	"	.section __ex_table,\"a\"	\n"	\
64 	"	.align  4			\n"	\
65 	"	.word   1b, 4b			\n"	\
66 	"	.word   2b, 4b			\n"	\
67 	"	.previous			\n"	\
68 							\
69 	: "=&r" (ret), "=&r" (oldval)			\
70 	: "r" (uaddr), "r" (oparg), "ir" (-EFAULT)	\
71 	: "cc", "memory");				\
72 	smp_mb()					\
73 
74 #endif
75 
76 static inline int futex_atomic_op_inuser(int encoded_op, u32 __user *uaddr)
77 {
78 	int op = (encoded_op >> 28) & 7;
79 	int cmp = (encoded_op >> 24) & 15;
80 	int oparg = (encoded_op << 8) >> 20;
81 	int cmparg = (encoded_op << 20) >> 20;
82 	int oldval = 0, ret;
83 
84 	if (encoded_op & (FUTEX_OP_OPARG_SHIFT << 28))
85 		oparg = 1 << oparg;
86 
87 	if (!access_ok(VERIFY_WRITE, uaddr, sizeof(int)))
88 		return -EFAULT;
89 
90 #ifndef CONFIG_ARC_HAS_LLSC
91 	preempt_disable();	/* to guarantee atomic r-m-w of futex op */
92 #endif
93 	pagefault_disable();
94 
95 	switch (op) {
96 	case FUTEX_OP_SET:
97 		__futex_atomic_op("mov %0, %3", ret, oldval, uaddr, oparg);
98 		break;
99 	case FUTEX_OP_ADD:
100 		/* oldval = *uaddr; *uaddr += oparg ; ret = *uaddr */
101 		__futex_atomic_op("add %0, %1, %3", ret, oldval, uaddr, oparg);
102 		break;
103 	case FUTEX_OP_OR:
104 		__futex_atomic_op("or  %0, %1, %3", ret, oldval, uaddr, oparg);
105 		break;
106 	case FUTEX_OP_ANDN:
107 		__futex_atomic_op("bic %0, %1, %3", ret, oldval, uaddr, oparg);
108 		break;
109 	case FUTEX_OP_XOR:
110 		__futex_atomic_op("xor %0, %1, %3", ret, oldval, uaddr, oparg);
111 		break;
112 	default:
113 		ret = -ENOSYS;
114 	}
115 
116 	pagefault_enable();
117 #ifndef CONFIG_ARC_HAS_LLSC
118 	preempt_enable();
119 #endif
120 
121 	if (!ret) {
122 		switch (cmp) {
123 		case FUTEX_OP_CMP_EQ:
124 			ret = (oldval == cmparg);
125 			break;
126 		case FUTEX_OP_CMP_NE:
127 			ret = (oldval != cmparg);
128 			break;
129 		case FUTEX_OP_CMP_LT:
130 			ret = (oldval < cmparg);
131 			break;
132 		case FUTEX_OP_CMP_GE:
133 			ret = (oldval >= cmparg);
134 			break;
135 		case FUTEX_OP_CMP_LE:
136 			ret = (oldval <= cmparg);
137 			break;
138 		case FUTEX_OP_CMP_GT:
139 			ret = (oldval > cmparg);
140 			break;
141 		default:
142 			ret = -ENOSYS;
143 		}
144 	}
145 	return ret;
146 }
147 
148 /*
149  * cmpxchg of futex (pagefaults disabled by caller)
150  * Return 0 for success, -EFAULT otherwise
151  */
152 static inline int
153 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, u32 expval,
154 			      u32 newval)
155 {
156 	int ret = 0;
157 	u32 existval;
158 
159 	if (!access_ok(VERIFY_WRITE, uaddr, sizeof(u32)))
160 		return -EFAULT;
161 
162 #ifndef CONFIG_ARC_HAS_LLSC
163 	preempt_disable();	/* to guarantee atomic r-m-w of futex op */
164 #endif
165 	smp_mb();
166 
167 	__asm__ __volatile__(
168 #ifdef CONFIG_ARC_HAS_LLSC
169 	"1:	llock	%1, [%4]		\n"
170 	"	brne	%1, %2, 3f		\n"
171 	"2:	scond	%3, [%4]		\n"
172 	"	bnz	1b			\n"
173 #else
174 	"1:	ld	%1, [%4]		\n"
175 	"	brne	%1, %2, 3f		\n"
176 	"2:	st	%3, [%4]		\n"
177 #endif
178 	"3:	\n"
179 	"	.section .fixup,\"ax\"	\n"
180 	"4:	mov %0, %5	\n"
181 	"	j   3b	\n"
182 	"	.previous	\n"
183 	"	.section __ex_table,\"a\"	\n"
184 	"	.align  4	\n"
185 	"	.word   1b, 4b	\n"
186 	"	.word   2b, 4b	\n"
187 	"	.previous\n"
188 	: "+&r"(ret), "=&r"(existval)
189 	: "r"(expval), "r"(newval), "r"(uaddr), "ir"(-EFAULT)
190 	: "cc", "memory");
191 
192 	smp_mb();
193 
194 #ifndef CONFIG_ARC_HAS_LLSC
195 	preempt_enable();
196 #endif
197 	*uval = existval;
198 	return ret;
199 }
200 
201 #endif
202