The goal is to (in the future) allow for per-target compilation of functions in atomic_template.h whilst atomic_mmu_lookup() and cputlb.c are compiled once-per user- or system mode. Signed-off-by: Anton Johansson <anjo@rev.ng> Message-Id: <20230912153428.17816-7-anjo@rev.ng> Reviewed-by: Richard Henderson <richard.henderson@linaro.org> Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> [rth: Use cpu->neg.tlb instead of cpu_tlb()] Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
		
			
				
	
	
		
			284 lines
		
	
	
		
			9.9 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			284 lines
		
	
	
		
			9.9 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 * Atomic helper templates
 | 
						|
 * Included from tcg-runtime.c and cputlb.c.
 | 
						|
 *
 | 
						|
 * Copyright (c) 2016 Red Hat, Inc
 | 
						|
 *
 | 
						|
 * This library is free software; you can redistribute it and/or
 | 
						|
 * modify it under the terms of the GNU Lesser General Public
 | 
						|
 * License as published by the Free Software Foundation; either
 | 
						|
 * version 2.1 of the License, or (at your option) any later version.
 | 
						|
 *
 | 
						|
 * This library is distributed in the hope that it will be useful,
 | 
						|
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
						|
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 | 
						|
 * Lesser General Public License for more details.
 | 
						|
 *
 | 
						|
 * You should have received a copy of the GNU Lesser General Public
 | 
						|
 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
 | 
						|
 */
 | 
						|
 | 
						|
#include "qemu/plugin.h"
 | 
						|
 | 
						|
#if DATA_SIZE == 16
 | 
						|
# define SUFFIX     o
 | 
						|
# define DATA_TYPE  Int128
 | 
						|
# define BSWAP      bswap128
 | 
						|
# define SHIFT      4
 | 
						|
#elif DATA_SIZE == 8
 | 
						|
# define SUFFIX     q
 | 
						|
# define DATA_TYPE  aligned_uint64_t
 | 
						|
# define SDATA_TYPE aligned_int64_t
 | 
						|
# define BSWAP      bswap64
 | 
						|
# define SHIFT      3
 | 
						|
#elif DATA_SIZE == 4
 | 
						|
# define SUFFIX     l
 | 
						|
# define DATA_TYPE  uint32_t
 | 
						|
# define SDATA_TYPE int32_t
 | 
						|
# define BSWAP      bswap32
 | 
						|
# define SHIFT      2
 | 
						|
#elif DATA_SIZE == 2
 | 
						|
# define SUFFIX     w
 | 
						|
# define DATA_TYPE  uint16_t
 | 
						|
# define SDATA_TYPE int16_t
 | 
						|
# define BSWAP      bswap16
 | 
						|
# define SHIFT      1
 | 
						|
#elif DATA_SIZE == 1
 | 
						|
# define SUFFIX     b
 | 
						|
# define DATA_TYPE  uint8_t
 | 
						|
# define SDATA_TYPE int8_t
 | 
						|
# define BSWAP
 | 
						|
# define SHIFT      0
 | 
						|
#else
 | 
						|
# error unsupported data size
 | 
						|
#endif
 | 
						|
 | 
						|
#if DATA_SIZE >= 4
 | 
						|
# define ABI_TYPE  DATA_TYPE
 | 
						|
#else
 | 
						|
# define ABI_TYPE  uint32_t
 | 
						|
#endif
 | 
						|
 | 
						|
/* Define host-endian atomic operations.  Note that END is used within
 | 
						|
   the ATOMIC_NAME macro, and redefined below.  */
 | 
						|
#if DATA_SIZE == 1
 | 
						|
# define END
 | 
						|
#elif HOST_BIG_ENDIAN
 | 
						|
# define END  _be
 | 
						|
#else
 | 
						|
# define END  _le
 | 
						|
#endif
 | 
						|
 | 
						|
ABI_TYPE ATOMIC_NAME(cmpxchg)(CPUArchState *env, abi_ptr addr,
 | 
						|
                              ABI_TYPE cmpv, ABI_TYPE newv,
 | 
						|
                              MemOpIdx oi, uintptr_t retaddr)
 | 
						|
{
 | 
						|
    DATA_TYPE *haddr = atomic_mmu_lookup(env_cpu(env), addr, oi,
 | 
						|
                                         DATA_SIZE, retaddr);
 | 
						|
    DATA_TYPE ret;
 | 
						|
 | 
						|
#if DATA_SIZE == 16
 | 
						|
    ret = atomic16_cmpxchg(haddr, cmpv, newv);
 | 
						|
#else
 | 
						|
    ret = qatomic_cmpxchg__nocheck(haddr, cmpv, newv);
 | 
						|
#endif
 | 
						|
    ATOMIC_MMU_CLEANUP;
 | 
						|
    atomic_trace_rmw_post(env, addr, oi);
 | 
						|
    return ret;
 | 
						|
}
 | 
						|
 | 
						|
#if DATA_SIZE < 16
 | 
						|
ABI_TYPE ATOMIC_NAME(xchg)(CPUArchState *env, abi_ptr addr, ABI_TYPE val,
 | 
						|
                           MemOpIdx oi, uintptr_t retaddr)
 | 
						|
{
 | 
						|
    DATA_TYPE *haddr = atomic_mmu_lookup(env_cpu(env), addr, oi,
 | 
						|
                                         DATA_SIZE, retaddr);
 | 
						|
    DATA_TYPE ret;
 | 
						|
 | 
						|
    ret = qatomic_xchg__nocheck(haddr, val);
 | 
						|
    ATOMIC_MMU_CLEANUP;
 | 
						|
    atomic_trace_rmw_post(env, addr, oi);
 | 
						|
    return ret;
 | 
						|
}
 | 
						|
 | 
						|
#define GEN_ATOMIC_HELPER(X)                                        \
 | 
						|
ABI_TYPE ATOMIC_NAME(X)(CPUArchState *env, abi_ptr addr,            \
 | 
						|
                        ABI_TYPE val, MemOpIdx oi, uintptr_t retaddr) \
 | 
						|
{                                                                   \
 | 
						|
    DATA_TYPE *haddr, ret;                                          \
 | 
						|
    haddr = atomic_mmu_lookup(env_cpu(env), addr, oi, DATA_SIZE, retaddr);   \
 | 
						|
    ret = qatomic_##X(haddr, val);                                  \
 | 
						|
    ATOMIC_MMU_CLEANUP;                                             \
 | 
						|
    atomic_trace_rmw_post(env, addr, oi);                           \
 | 
						|
    return ret;                                                     \
 | 
						|
}
 | 
						|
 | 
						|
GEN_ATOMIC_HELPER(fetch_add)
 | 
						|
GEN_ATOMIC_HELPER(fetch_and)
 | 
						|
GEN_ATOMIC_HELPER(fetch_or)
 | 
						|
GEN_ATOMIC_HELPER(fetch_xor)
 | 
						|
GEN_ATOMIC_HELPER(add_fetch)
 | 
						|
GEN_ATOMIC_HELPER(and_fetch)
 | 
						|
GEN_ATOMIC_HELPER(or_fetch)
 | 
						|
GEN_ATOMIC_HELPER(xor_fetch)
 | 
						|
 | 
						|
#undef GEN_ATOMIC_HELPER
 | 
						|
 | 
						|
/*
 | 
						|
 * These helpers are, as a whole, full barriers.  Within the helper,
 | 
						|
 * the leading barrier is explicit and the trailing barrier is within
 | 
						|
 * cmpxchg primitive.
 | 
						|
 *
 | 
						|
 * Trace this load + RMW loop as a single RMW op. This way, regardless
 | 
						|
 * of CF_PARALLEL's value, we'll trace just a read and a write.
 | 
						|
 */
 | 
						|
#define GEN_ATOMIC_HELPER_FN(X, FN, XDATA_TYPE, RET)                \
 | 
						|
ABI_TYPE ATOMIC_NAME(X)(CPUArchState *env, abi_ptr addr,            \
 | 
						|
                        ABI_TYPE xval, MemOpIdx oi, uintptr_t retaddr) \
 | 
						|
{                                                                   \
 | 
						|
    XDATA_TYPE *haddr, cmp, old, new, val = xval;                   \
 | 
						|
    haddr = atomic_mmu_lookup(env_cpu(env), addr, oi, DATA_SIZE, retaddr);   \
 | 
						|
    smp_mb();                                                       \
 | 
						|
    cmp = qatomic_read__nocheck(haddr);                             \
 | 
						|
    do {                                                            \
 | 
						|
        old = cmp; new = FN(old, val);                              \
 | 
						|
        cmp = qatomic_cmpxchg__nocheck(haddr, old, new);            \
 | 
						|
    } while (cmp != old);                                           \
 | 
						|
    ATOMIC_MMU_CLEANUP;                                             \
 | 
						|
    atomic_trace_rmw_post(env, addr, oi);                           \
 | 
						|
    return RET;                                                     \
 | 
						|
}
 | 
						|
 | 
						|
GEN_ATOMIC_HELPER_FN(fetch_smin, MIN, SDATA_TYPE, old)
 | 
						|
GEN_ATOMIC_HELPER_FN(fetch_umin, MIN,  DATA_TYPE, old)
 | 
						|
GEN_ATOMIC_HELPER_FN(fetch_smax, MAX, SDATA_TYPE, old)
 | 
						|
GEN_ATOMIC_HELPER_FN(fetch_umax, MAX,  DATA_TYPE, old)
 | 
						|
 | 
						|
GEN_ATOMIC_HELPER_FN(smin_fetch, MIN, SDATA_TYPE, new)
 | 
						|
GEN_ATOMIC_HELPER_FN(umin_fetch, MIN,  DATA_TYPE, new)
 | 
						|
GEN_ATOMIC_HELPER_FN(smax_fetch, MAX, SDATA_TYPE, new)
 | 
						|
GEN_ATOMIC_HELPER_FN(umax_fetch, MAX,  DATA_TYPE, new)
 | 
						|
 | 
						|
#undef GEN_ATOMIC_HELPER_FN
 | 
						|
#endif /* DATA SIZE < 16 */
 | 
						|
 | 
						|
#undef END
 | 
						|
 | 
						|
#if DATA_SIZE > 1
 | 
						|
 | 
						|
/* Define reverse-host-endian atomic operations.  Note that END is used
 | 
						|
   within the ATOMIC_NAME macro.  */
 | 
						|
#if HOST_BIG_ENDIAN
 | 
						|
# define END  _le
 | 
						|
#else
 | 
						|
# define END  _be
 | 
						|
#endif
 | 
						|
 | 
						|
ABI_TYPE ATOMIC_NAME(cmpxchg)(CPUArchState *env, abi_ptr addr,
 | 
						|
                              ABI_TYPE cmpv, ABI_TYPE newv,
 | 
						|
                              MemOpIdx oi, uintptr_t retaddr)
 | 
						|
{
 | 
						|
    DATA_TYPE *haddr = atomic_mmu_lookup(env_cpu(env), addr, oi,
 | 
						|
                                         DATA_SIZE, retaddr);
 | 
						|
    DATA_TYPE ret;
 | 
						|
 | 
						|
#if DATA_SIZE == 16
 | 
						|
    ret = atomic16_cmpxchg(haddr, BSWAP(cmpv), BSWAP(newv));
 | 
						|
#else
 | 
						|
    ret = qatomic_cmpxchg__nocheck(haddr, BSWAP(cmpv), BSWAP(newv));
 | 
						|
#endif
 | 
						|
    ATOMIC_MMU_CLEANUP;
 | 
						|
    atomic_trace_rmw_post(env, addr, oi);
 | 
						|
    return BSWAP(ret);
 | 
						|
}
 | 
						|
 | 
						|
#if DATA_SIZE < 16
 | 
						|
ABI_TYPE ATOMIC_NAME(xchg)(CPUArchState *env, abi_ptr addr, ABI_TYPE val,
 | 
						|
                           MemOpIdx oi, uintptr_t retaddr)
 | 
						|
{
 | 
						|
    DATA_TYPE *haddr = atomic_mmu_lookup(env_cpu(env), addr, oi,
 | 
						|
                                         DATA_SIZE, retaddr);
 | 
						|
    ABI_TYPE ret;
 | 
						|
 | 
						|
    ret = qatomic_xchg__nocheck(haddr, BSWAP(val));
 | 
						|
    ATOMIC_MMU_CLEANUP;
 | 
						|
    atomic_trace_rmw_post(env, addr, oi);
 | 
						|
    return BSWAP(ret);
 | 
						|
}
 | 
						|
 | 
						|
#define GEN_ATOMIC_HELPER(X)                                        \
 | 
						|
ABI_TYPE ATOMIC_NAME(X)(CPUArchState *env, abi_ptr addr,            \
 | 
						|
                        ABI_TYPE val, MemOpIdx oi, uintptr_t retaddr) \
 | 
						|
{                                                                   \
 | 
						|
    DATA_TYPE *haddr, ret;                                          \
 | 
						|
    haddr = atomic_mmu_lookup(env_cpu(env), addr, oi, DATA_SIZE, retaddr);   \
 | 
						|
    ret = qatomic_##X(haddr, BSWAP(val));                           \
 | 
						|
    ATOMIC_MMU_CLEANUP;                                             \
 | 
						|
    atomic_trace_rmw_post(env, addr, oi);                           \
 | 
						|
    return BSWAP(ret);                                              \
 | 
						|
}
 | 
						|
 | 
						|
GEN_ATOMIC_HELPER(fetch_and)
 | 
						|
GEN_ATOMIC_HELPER(fetch_or)
 | 
						|
GEN_ATOMIC_HELPER(fetch_xor)
 | 
						|
GEN_ATOMIC_HELPER(and_fetch)
 | 
						|
GEN_ATOMIC_HELPER(or_fetch)
 | 
						|
GEN_ATOMIC_HELPER(xor_fetch)
 | 
						|
 | 
						|
#undef GEN_ATOMIC_HELPER
 | 
						|
 | 
						|
/* These helpers are, as a whole, full barriers.  Within the helper,
 | 
						|
 * the leading barrier is explicit and the trailing barrier is within
 | 
						|
 * cmpxchg primitive.
 | 
						|
 *
 | 
						|
 * Trace this load + RMW loop as a single RMW op. This way, regardless
 | 
						|
 * of CF_PARALLEL's value, we'll trace just a read and a write.
 | 
						|
 */
 | 
						|
#define GEN_ATOMIC_HELPER_FN(X, FN, XDATA_TYPE, RET)                \
 | 
						|
ABI_TYPE ATOMIC_NAME(X)(CPUArchState *env, abi_ptr addr,            \
 | 
						|
                        ABI_TYPE xval, MemOpIdx oi, uintptr_t retaddr) \
 | 
						|
{                                                                   \
 | 
						|
    XDATA_TYPE *haddr, ldo, ldn, old, new, val = xval;              \
 | 
						|
    haddr = atomic_mmu_lookup(env_cpu(env), addr, oi, DATA_SIZE, retaddr);   \
 | 
						|
    smp_mb();                                                       \
 | 
						|
    ldn = qatomic_read__nocheck(haddr);                             \
 | 
						|
    do {                                                            \
 | 
						|
        ldo = ldn; old = BSWAP(ldo); new = FN(old, val);            \
 | 
						|
        ldn = qatomic_cmpxchg__nocheck(haddr, ldo, BSWAP(new));     \
 | 
						|
    } while (ldo != ldn);                                           \
 | 
						|
    ATOMIC_MMU_CLEANUP;                                             \
 | 
						|
    atomic_trace_rmw_post(env, addr, oi);                           \
 | 
						|
    return RET;                                                     \
 | 
						|
}
 | 
						|
 | 
						|
GEN_ATOMIC_HELPER_FN(fetch_smin, MIN, SDATA_TYPE, old)
 | 
						|
GEN_ATOMIC_HELPER_FN(fetch_umin, MIN,  DATA_TYPE, old)
 | 
						|
GEN_ATOMIC_HELPER_FN(fetch_smax, MAX, SDATA_TYPE, old)
 | 
						|
GEN_ATOMIC_HELPER_FN(fetch_umax, MAX,  DATA_TYPE, old)
 | 
						|
 | 
						|
GEN_ATOMIC_HELPER_FN(smin_fetch, MIN, SDATA_TYPE, new)
 | 
						|
GEN_ATOMIC_HELPER_FN(umin_fetch, MIN,  DATA_TYPE, new)
 | 
						|
GEN_ATOMIC_HELPER_FN(smax_fetch, MAX, SDATA_TYPE, new)
 | 
						|
GEN_ATOMIC_HELPER_FN(umax_fetch, MAX,  DATA_TYPE, new)
 | 
						|
 | 
						|
/* Note that for addition, we need to use a separate cmpxchg loop instead
 | 
						|
   of bswaps for the reverse-host-endian helpers.  */
 | 
						|
#define ADD(X, Y)   (X + Y)
 | 
						|
GEN_ATOMIC_HELPER_FN(fetch_add, ADD, DATA_TYPE, old)
 | 
						|
GEN_ATOMIC_HELPER_FN(add_fetch, ADD, DATA_TYPE, new)
 | 
						|
#undef ADD
 | 
						|
 | 
						|
#undef GEN_ATOMIC_HELPER_FN
 | 
						|
#endif /* DATA_SIZE < 16 */
 | 
						|
 | 
						|
#undef END
 | 
						|
#endif /* DATA_SIZE > 1 */
 | 
						|
 | 
						|
#undef BSWAP
 | 
						|
#undef ABI_TYPE
 | 
						|
#undef DATA_TYPE
 | 
						|
#undef SDATA_TYPE
 | 
						|
#undef SUFFIX
 | 
						|
#undef DATA_SIZE
 | 
						|
#undef SHIFT
 |