2 * Atomic helper templates
3 * Included from tcg-runtime.c and cputlb.c.
5 * Copyright (c) 2016 Red Hat, Inc
7 * This library is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU Lesser General Public
9 * License as published by the Free Software Foundation; either
10 * version 2.1 of the License, or (at your option) any later version.
12 * This library is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * Lesser General Public License for more details.
17 * You should have received a copy of the GNU Lesser General Public
18 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
21 #include "trace/mem.h"
25 # define DATA_TYPE Int128
26 # define BSWAP bswap128
30 # define DATA_TYPE uint64_t
31 # define SDATA_TYPE int64_t
32 # define BSWAP bswap64
36 # define DATA_TYPE uint32_t
37 # define SDATA_TYPE int32_t
38 # define BSWAP bswap32
42 # define DATA_TYPE uint16_t
43 # define SDATA_TYPE int16_t
44 # define BSWAP bswap16
48 # define DATA_TYPE uint8_t
49 # define SDATA_TYPE int8_t
53 # error unsupported data size
57 # define ABI_TYPE DATA_TYPE
59 # define ABI_TYPE uint32_t
62 #define ATOMIC_TRACE_RMW do { \
63 uint16_t info = glue(trace_mem_build_info_no_se, MEND) \
64 (SHIFT, false, ATOMIC_MMU_IDX); \
66 trace_guest_mem_before_exec(env_cpu(env), addr, info); \
67 trace_guest_mem_before_exec(env_cpu(env), addr, \
68 info | TRACE_MEM_ST); \
71 #define ATOMIC_TRACE_LD do { \
72 uint16_t info = glue(trace_mem_build_info_no_se, MEND) \
73 (SHIFT, false, ATOMIC_MMU_IDX); \
75 trace_guest_mem_before_exec(env_cpu(env), addr, info); \
78 #define ATOMIC_TRACE_ST do { \
79 uint16_t info = glue(trace_mem_build_info_no_se, MEND) \
80 (SHIFT, true, ATOMIC_MMU_IDX); \
82 trace_guest_mem_before_exec(env_cpu(env), addr, info); \
85 /* Define host-endian atomic operations. Note that END is used within
86 the ATOMIC_NAME macro, and redefined below. */
89 # define MEND _be /* either le or be would be fine */
90 #elif defined(HOST_WORDS_BIGENDIAN)
98 ABI_TYPE ATOMIC_NAME(cmpxchg)(CPUArchState *env, target_ulong addr,
99 ABI_TYPE cmpv, ABI_TYPE newv EXTRA_ARGS)
102 DATA_TYPE *haddr = ATOMIC_MMU_LOOKUP;
107 ret = atomic16_cmpxchg(haddr, cmpv, newv);
109 ret = atomic_cmpxchg__nocheck(haddr, cmpv, newv);
117 ABI_TYPE ATOMIC_NAME(ld)(CPUArchState *env, target_ulong addr EXTRA_ARGS)
120 DATA_TYPE val, *haddr = ATOMIC_MMU_LOOKUP;
123 val = atomic16_read(haddr);
128 void ATOMIC_NAME(st)(CPUArchState *env, target_ulong addr,
129 ABI_TYPE val EXTRA_ARGS)
132 DATA_TYPE *haddr = ATOMIC_MMU_LOOKUP;
135 atomic16_set(haddr, val);
140 ABI_TYPE ATOMIC_NAME(xchg)(CPUArchState *env, target_ulong addr,
141 ABI_TYPE val EXTRA_ARGS)
144 DATA_TYPE *haddr = ATOMIC_MMU_LOOKUP;
148 ret = atomic_xchg__nocheck(haddr, val);
153 #define GEN_ATOMIC_HELPER(X) \
154 ABI_TYPE ATOMIC_NAME(X)(CPUArchState *env, target_ulong addr, \
155 ABI_TYPE val EXTRA_ARGS) \
158 DATA_TYPE *haddr = ATOMIC_MMU_LOOKUP; \
162 ret = atomic_##X(haddr, val); \
163 ATOMIC_MMU_CLEANUP; \
167 GEN_ATOMIC_HELPER(fetch_add)
168 GEN_ATOMIC_HELPER(fetch_and)
169 GEN_ATOMIC_HELPER(fetch_or)
170 GEN_ATOMIC_HELPER(fetch_xor)
171 GEN_ATOMIC_HELPER(add_fetch)
172 GEN_ATOMIC_HELPER(and_fetch)
173 GEN_ATOMIC_HELPER(or_fetch)
174 GEN_ATOMIC_HELPER(xor_fetch)
176 #undef GEN_ATOMIC_HELPER
178 /* These helpers are, as a whole, full barriers. Within the helper,
179 * the leading barrier is explicit and the trailing barrier is within
182 * Trace this load + RMW loop as a single RMW op. This way, regardless
183 * of CF_PARALLEL's value, we'll trace just a read and a write.
185 #define GEN_ATOMIC_HELPER_FN(X, FN, XDATA_TYPE, RET) \
186 ABI_TYPE ATOMIC_NAME(X)(CPUArchState *env, target_ulong addr, \
187 ABI_TYPE xval EXTRA_ARGS) \
190 XDATA_TYPE *haddr = ATOMIC_MMU_LOOKUP; \
191 XDATA_TYPE cmp, old, new, val = xval; \
195 cmp = atomic_read__nocheck(haddr); \
197 old = cmp; new = FN(old, val); \
198 cmp = atomic_cmpxchg__nocheck(haddr, old, new); \
199 } while (cmp != old); \
200 ATOMIC_MMU_CLEANUP; \
204 GEN_ATOMIC_HELPER_FN(fetch_smin, MIN, SDATA_TYPE, old)
205 GEN_ATOMIC_HELPER_FN(fetch_umin, MIN, DATA_TYPE, old)
206 GEN_ATOMIC_HELPER_FN(fetch_smax, MAX, SDATA_TYPE, old)
207 GEN_ATOMIC_HELPER_FN(fetch_umax, MAX, DATA_TYPE, old)
209 GEN_ATOMIC_HELPER_FN(smin_fetch, MIN, SDATA_TYPE, new)
210 GEN_ATOMIC_HELPER_FN(umin_fetch, MIN, DATA_TYPE, new)
211 GEN_ATOMIC_HELPER_FN(smax_fetch, MAX, SDATA_TYPE, new)
212 GEN_ATOMIC_HELPER_FN(umax_fetch, MAX, DATA_TYPE, new)
214 #undef GEN_ATOMIC_HELPER_FN
215 #endif /* DATA SIZE >= 16 */
222 /* Define reverse-host-endian atomic operations. Note that END is used
223 within the ATOMIC_NAME macro. */
224 #ifdef HOST_WORDS_BIGENDIAN
232 ABI_TYPE ATOMIC_NAME(cmpxchg)(CPUArchState *env, target_ulong addr,
233 ABI_TYPE cmpv, ABI_TYPE newv EXTRA_ARGS)
236 DATA_TYPE *haddr = ATOMIC_MMU_LOOKUP;
241 ret = atomic16_cmpxchg(haddr, BSWAP(cmpv), BSWAP(newv));
243 ret = atomic_cmpxchg__nocheck(haddr, BSWAP(cmpv), BSWAP(newv));
251 ABI_TYPE ATOMIC_NAME(ld)(CPUArchState *env, target_ulong addr EXTRA_ARGS)
254 DATA_TYPE val, *haddr = ATOMIC_MMU_LOOKUP;
257 val = atomic16_read(haddr);
262 void ATOMIC_NAME(st)(CPUArchState *env, target_ulong addr,
263 ABI_TYPE val EXTRA_ARGS)
266 DATA_TYPE *haddr = ATOMIC_MMU_LOOKUP;
270 atomic16_set(haddr, val);
275 ABI_TYPE ATOMIC_NAME(xchg)(CPUArchState *env, target_ulong addr,
276 ABI_TYPE val EXTRA_ARGS)
279 DATA_TYPE *haddr = ATOMIC_MMU_LOOKUP;
283 ret = atomic_xchg__nocheck(haddr, BSWAP(val));
288 #define GEN_ATOMIC_HELPER(X) \
289 ABI_TYPE ATOMIC_NAME(X)(CPUArchState *env, target_ulong addr, \
290 ABI_TYPE val EXTRA_ARGS) \
293 DATA_TYPE *haddr = ATOMIC_MMU_LOOKUP; \
297 ret = atomic_##X(haddr, BSWAP(val)); \
298 ATOMIC_MMU_CLEANUP; \
302 GEN_ATOMIC_HELPER(fetch_and)
303 GEN_ATOMIC_HELPER(fetch_or)
304 GEN_ATOMIC_HELPER(fetch_xor)
305 GEN_ATOMIC_HELPER(and_fetch)
306 GEN_ATOMIC_HELPER(or_fetch)
307 GEN_ATOMIC_HELPER(xor_fetch)
309 #undef GEN_ATOMIC_HELPER
311 /* These helpers are, as a whole, full barriers. Within the helper,
312 * the leading barrier is explicit and the trailing barrier is within
315 * Trace this load + RMW loop as a single RMW op. This way, regardless
316 * of CF_PARALLEL's value, we'll trace just a read and a write.
318 #define GEN_ATOMIC_HELPER_FN(X, FN, XDATA_TYPE, RET) \
319 ABI_TYPE ATOMIC_NAME(X)(CPUArchState *env, target_ulong addr, \
320 ABI_TYPE xval EXTRA_ARGS) \
323 XDATA_TYPE *haddr = ATOMIC_MMU_LOOKUP; \
324 XDATA_TYPE ldo, ldn, old, new, val = xval; \
328 ldn = atomic_read__nocheck(haddr); \
330 ldo = ldn; old = BSWAP(ldo); new = FN(old, val); \
331 ldn = atomic_cmpxchg__nocheck(haddr, ldo, BSWAP(new)); \
332 } while (ldo != ldn); \
333 ATOMIC_MMU_CLEANUP; \
337 GEN_ATOMIC_HELPER_FN(fetch_smin, MIN, SDATA_TYPE, old)
338 GEN_ATOMIC_HELPER_FN(fetch_umin, MIN, DATA_TYPE, old)
339 GEN_ATOMIC_HELPER_FN(fetch_smax, MAX, SDATA_TYPE, old)
340 GEN_ATOMIC_HELPER_FN(fetch_umax, MAX, DATA_TYPE, old)
342 GEN_ATOMIC_HELPER_FN(smin_fetch, MIN, SDATA_TYPE, new)
343 GEN_ATOMIC_HELPER_FN(umin_fetch, MIN, DATA_TYPE, new)
344 GEN_ATOMIC_HELPER_FN(smax_fetch, MAX, SDATA_TYPE, new)
345 GEN_ATOMIC_HELPER_FN(umax_fetch, MAX, DATA_TYPE, new)
347 /* Note that for addition, we need to use a separate cmpxchg loop instead
348 of bswaps for the reverse-host-endian helpers. */
349 #define ADD(X, Y) (X + Y)
350 GEN_ATOMIC_HELPER_FN(fetch_add, ADD, DATA_TYPE, old)
351 GEN_ATOMIC_HELPER_FN(add_fetch, ADD, DATA_TYPE, new)
354 #undef GEN_ATOMIC_HELPER_FN
355 #endif /* DATA_SIZE >= 16 */
359 #endif /* DATA_SIZE > 1 */
361 #undef ATOMIC_TRACE_ST
362 #undef ATOMIC_TRACE_LD
363 #undef ATOMIC_TRACE_RMW