[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [RFC PATCH v2 11/15] xen/arm64: port Linux's arm64 atomic.h to Xen
From: Ash Wilding <ash.j.wilding@xxxxxxxxx> - Drop atomic64_t helper declarations as we don't currently have an atomic64_t in Xen. - Drop arch_* prefixes. - Swap include of <linux/compiler.h> to just <xen/rwonce.h>. Signed-off-by: Ash Wilding <ash.j.wilding@xxxxxxxxx> --- xen/include/asm-arm/arm64/atomic.h | 256 ++++++++--------------------- 1 file changed, 73 insertions(+), 183 deletions(-) diff --git a/xen/include/asm-arm/arm64/atomic.h b/xen/include/asm-arm/arm64/atomic.h index a2eab9f091..b695cc6e09 100644 --- a/xen/include/asm-arm/arm64/atomic.h +++ b/xen/include/asm-arm/arm64/atomic.h @@ -1,23 +1,23 @@ -/* SPDX-License-Identifier: GPL-2.0-only */ + /* - * Based on arch/arm/include/asm/atomic.h + * Taken from Linux 5.10-rc2 (last commit 3cea11cd5) * * Copyright (C) 1996 Russell King. * Copyright (C) 2002 Deep Blue Solutions Ltd. * Copyright (C) 2012 ARM Ltd. + * SPDX-License-Identifier: GPL-2.0-only */ -#ifndef __ASM_ATOMIC_H -#define __ASM_ATOMIC_H +#ifndef __ASM_ARM_ARM64_ATOMIC_H +#define __ASM_ARM_ARM64_ATOMIC_H -#include <linux/compiler.h> -#include <linux/types.h> +#include <xen/rwonce.h> +#include <xen/types.h> -#include <asm/barrier.h> -#include <asm/cmpxchg.h> -#include <asm/lse.h> +#include "lse.h" +#include "cmpxchg.h" #define ATOMIC_OP(op) \ -static inline void arch_##op(int i, atomic_t *v) \ +static inline void op(int i, atomic_t *v) \ { \ __lse_ll_sc_body(op, i, v); \ } @@ -32,7 +32,7 @@ ATOMIC_OP(atomic_sub) #undef ATOMIC_OP #define ATOMIC_FETCH_OP(name, op) \ -static inline int arch_##op##name(int i, atomic_t *v) \ +static inline int op##name(int i, atomic_t *v) \ { \ return __lse_ll_sc_body(op##name, i, v); \ } @@ -54,175 +54,65 @@ ATOMIC_FETCH_OPS(atomic_sub_return) #undef ATOMIC_FETCH_OP #undef ATOMIC_FETCH_OPS - -#define ATOMIC64_OP(op) \ -static inline void arch_##op(long i, atomic64_t *v) \ -{ \ - __lse_ll_sc_body(op, i, v); \ -} - -ATOMIC64_OP(atomic64_andnot) -ATOMIC64_OP(atomic64_or) -ATOMIC64_OP(atomic64_xor) -ATOMIC64_OP(atomic64_add) -ATOMIC64_OP(atomic64_and) -ATOMIC64_OP(atomic64_sub) - -#undef ATOMIC64_OP - -#define ATOMIC64_FETCH_OP(name, op) \ -static inline long arch_##op##name(long i, atomic64_t *v) \ -{ \ - return __lse_ll_sc_body(op##name, i, v); \ -} - -#define ATOMIC64_FETCH_OPS(op) \ - ATOMIC64_FETCH_OP(_relaxed, op) \ - ATOMIC64_FETCH_OP(_acquire, op) \ - ATOMIC64_FETCH_OP(_release, op) \ - ATOMIC64_FETCH_OP( , op) - -ATOMIC64_FETCH_OPS(atomic64_fetch_andnot) -ATOMIC64_FETCH_OPS(atomic64_fetch_or) -ATOMIC64_FETCH_OPS(atomic64_fetch_xor) -ATOMIC64_FETCH_OPS(atomic64_fetch_add) -ATOMIC64_FETCH_OPS(atomic64_fetch_and) -ATOMIC64_FETCH_OPS(atomic64_fetch_sub) -ATOMIC64_FETCH_OPS(atomic64_add_return) -ATOMIC64_FETCH_OPS(atomic64_sub_return) - -#undef ATOMIC64_FETCH_OP -#undef ATOMIC64_FETCH_OPS - -static inline long arch_atomic64_dec_if_positive(atomic64_t *v) -{ - return __lse_ll_sc_body(atomic64_dec_if_positive, v); -} - -#define arch_atomic_read(v) __READ_ONCE((v)->counter) -#define arch_atomic_set(v, i) __WRITE_ONCE(((v)->counter), (i)) - -#define arch_atomic_add_return_relaxed arch_atomic_add_return_relaxed -#define arch_atomic_add_return_acquire arch_atomic_add_return_acquire -#define arch_atomic_add_return_release arch_atomic_add_return_release -#define arch_atomic_add_return arch_atomic_add_return - -#define arch_atomic_sub_return_relaxed arch_atomic_sub_return_relaxed -#define arch_atomic_sub_return_acquire arch_atomic_sub_return_acquire -#define arch_atomic_sub_return_release arch_atomic_sub_return_release -#define arch_atomic_sub_return arch_atomic_sub_return - -#define arch_atomic_fetch_add_relaxed arch_atomic_fetch_add_relaxed -#define arch_atomic_fetch_add_acquire arch_atomic_fetch_add_acquire -#define arch_atomic_fetch_add_release arch_atomic_fetch_add_release -#define arch_atomic_fetch_add arch_atomic_fetch_add - -#define arch_atomic_fetch_sub_relaxed arch_atomic_fetch_sub_relaxed -#define arch_atomic_fetch_sub_acquire arch_atomic_fetch_sub_acquire -#define arch_atomic_fetch_sub_release arch_atomic_fetch_sub_release -#define arch_atomic_fetch_sub arch_atomic_fetch_sub - -#define arch_atomic_fetch_and_relaxed arch_atomic_fetch_and_relaxed -#define arch_atomic_fetch_and_acquire arch_atomic_fetch_and_acquire -#define arch_atomic_fetch_and_release arch_atomic_fetch_and_release -#define arch_atomic_fetch_and arch_atomic_fetch_and - -#define arch_atomic_fetch_andnot_relaxed arch_atomic_fetch_andnot_relaxed -#define arch_atomic_fetch_andnot_acquire arch_atomic_fetch_andnot_acquire -#define arch_atomic_fetch_andnot_release arch_atomic_fetch_andnot_release -#define arch_atomic_fetch_andnot arch_atomic_fetch_andnot - -#define arch_atomic_fetch_or_relaxed arch_atomic_fetch_or_relaxed -#define arch_atomic_fetch_or_acquire arch_atomic_fetch_or_acquire -#define arch_atomic_fetch_or_release arch_atomic_fetch_or_release -#define arch_atomic_fetch_or arch_atomic_fetch_or - -#define arch_atomic_fetch_xor_relaxed arch_atomic_fetch_xor_relaxed -#define arch_atomic_fetch_xor_acquire arch_atomic_fetch_xor_acquire -#define arch_atomic_fetch_xor_release arch_atomic_fetch_xor_release -#define arch_atomic_fetch_xor arch_atomic_fetch_xor - -#define arch_atomic_xchg_relaxed(v, new) \ - arch_xchg_relaxed(&((v)->counter), (new)) -#define arch_atomic_xchg_acquire(v, new) \ - arch_xchg_acquire(&((v)->counter), (new)) -#define arch_atomic_xchg_release(v, new) \ - arch_xchg_release(&((v)->counter), (new)) -#define arch_atomic_xchg(v, new) \ - arch_xchg(&((v)->counter), (new)) - -#define arch_atomic_cmpxchg_relaxed(v, old, new) \ - arch_cmpxchg_relaxed(&((v)->counter), (old), (new)) -#define arch_atomic_cmpxchg_acquire(v, old, new) \ - arch_cmpxchg_acquire(&((v)->counter), (old), (new)) -#define arch_atomic_cmpxchg_release(v, old, new) \ - arch_cmpxchg_release(&((v)->counter), (old), (new)) -#define arch_atomic_cmpxchg(v, old, new) \ - arch_cmpxchg(&((v)->counter), (old), (new)) - -#define arch_atomic_andnot arch_atomic_andnot - -/* - * 64-bit arch_atomic operations. - */ -#define ATOMIC64_INIT ATOMIC_INIT -#define arch_atomic64_read arch_atomic_read -#define arch_atomic64_set arch_atomic_set - -#define arch_atomic64_add_return_relaxed arch_atomic64_add_return_relaxed -#define arch_atomic64_add_return_acquire arch_atomic64_add_return_acquire -#define arch_atomic64_add_return_release arch_atomic64_add_return_release -#define arch_atomic64_add_return arch_atomic64_add_return - -#define arch_atomic64_sub_return_relaxed arch_atomic64_sub_return_relaxed -#define arch_atomic64_sub_return_acquire arch_atomic64_sub_return_acquire -#define arch_atomic64_sub_return_release arch_atomic64_sub_return_release -#define arch_atomic64_sub_return arch_atomic64_sub_return - -#define arch_atomic64_fetch_add_relaxed arch_atomic64_fetch_add_relaxed -#define arch_atomic64_fetch_add_acquire arch_atomic64_fetch_add_acquire -#define arch_atomic64_fetch_add_release arch_atomic64_fetch_add_release -#define arch_atomic64_fetch_add arch_atomic64_fetch_add - -#define arch_atomic64_fetch_sub_relaxed arch_atomic64_fetch_sub_relaxed -#define arch_atomic64_fetch_sub_acquire arch_atomic64_fetch_sub_acquire -#define arch_atomic64_fetch_sub_release arch_atomic64_fetch_sub_release -#define arch_atomic64_fetch_sub arch_atomic64_fetch_sub - -#define arch_atomic64_fetch_and_relaxed arch_atomic64_fetch_and_relaxed -#define arch_atomic64_fetch_and_acquire arch_atomic64_fetch_and_acquire -#define arch_atomic64_fetch_and_release arch_atomic64_fetch_and_release -#define arch_atomic64_fetch_and arch_atomic64_fetch_and - -#define arch_atomic64_fetch_andnot_relaxed arch_atomic64_fetch_andnot_relaxed -#define arch_atomic64_fetch_andnot_acquire arch_atomic64_fetch_andnot_acquire -#define arch_atomic64_fetch_andnot_release arch_atomic64_fetch_andnot_release -#define arch_atomic64_fetch_andnot arch_atomic64_fetch_andnot - -#define arch_atomic64_fetch_or_relaxed arch_atomic64_fetch_or_relaxed -#define arch_atomic64_fetch_or_acquire arch_atomic64_fetch_or_acquire -#define arch_atomic64_fetch_or_release arch_atomic64_fetch_or_release -#define arch_atomic64_fetch_or arch_atomic64_fetch_or - -#define arch_atomic64_fetch_xor_relaxed arch_atomic64_fetch_xor_relaxed -#define arch_atomic64_fetch_xor_acquire arch_atomic64_fetch_xor_acquire -#define arch_atomic64_fetch_xor_release arch_atomic64_fetch_xor_release -#define arch_atomic64_fetch_xor arch_atomic64_fetch_xor - -#define arch_atomic64_xchg_relaxed arch_atomic_xchg_relaxed -#define arch_atomic64_xchg_acquire arch_atomic_xchg_acquire -#define arch_atomic64_xchg_release arch_atomic_xchg_release -#define arch_atomic64_xchg arch_atomic_xchg - -#define arch_atomic64_cmpxchg_relaxed arch_atomic_cmpxchg_relaxed -#define arch_atomic64_cmpxchg_acquire arch_atomic_cmpxchg_acquire -#define arch_atomic64_cmpxchg_release arch_atomic_cmpxchg_release -#define arch_atomic64_cmpxchg arch_atomic_cmpxchg - -#define arch_atomic64_andnot arch_atomic64_andnot - -#define arch_atomic64_dec_if_positive arch_atomic64_dec_if_positive - -#define ARCH_ATOMIC - -#endif /* __ASM_ATOMIC_H */ \ No newline at end of file +#define atomic_read(v) __READ_ONCE((v)->counter) +#define atomic_set(v, i) __WRITE_ONCE(((v)->counter), (i)) + +#define atomic_add_return_relaxed atomic_add_return_relaxed +#define atomic_add_return_acquire atomic_add_return_acquire +#define atomic_add_return_release atomic_add_return_release +#define atomic_add_return atomic_add_return + +#define atomic_sub_return_relaxed atomic_sub_return_relaxed +#define atomic_sub_return_acquire atomic_sub_return_acquire +#define atomic_sub_return_release atomic_sub_return_release +#define atomic_sub_return atomic_sub_return + +#define atomic_fetch_add_relaxed atomic_fetch_add_relaxed +#define atomic_fetch_add_acquire atomic_fetch_add_acquire +#define atomic_fetch_add_release atomic_fetch_add_release +#define atomic_fetch_add atomic_fetch_add + +#define atomic_fetch_sub_relaxed atomic_fetch_sub_relaxed +#define atomic_fetch_sub_acquire atomic_fetch_sub_acquire +#define atomic_fetch_sub_release atomic_fetch_sub_release +#define atomic_fetch_sub atomic_fetch_sub + +#define atomic_fetch_and_relaxed atomic_fetch_and_relaxed +#define atomic_fetch_and_acquire atomic_fetch_and_acquire +#define atomic_fetch_and_release atomic_fetch_and_release +#define atomic_fetch_and atomic_fetch_and + +#define atomic_fetch_andnot_relaxed atomic_fetch_andnot_relaxed +#define atomic_fetch_andnot_acquire atomic_fetch_andnot_acquire +#define atomic_fetch_andnot_release atomic_fetch_andnot_release +#define atomic_fetch_andnot atomic_fetch_andnot + +#define atomic_fetch_or_relaxed atomic_fetch_or_relaxed +#define atomic_fetch_or_acquire atomic_fetch_or_acquire +#define atomic_fetch_or_release atomic_fetch_or_release +#define atomic_fetch_or atomic_fetch_or + +#define atomic_fetch_xor_relaxed atomic_fetch_xor_relaxed +#define atomic_fetch_xor_acquire atomic_fetch_xor_acquire +#define atomic_fetch_xor_release atomic_fetch_xor_release +#define atomic_fetch_xor atomic_fetch_xor + +#define atomic_xchg_relaxed(v, new) \ + xchg_relaxed(&((v)->counter), (new)) +#define atomic_xchg_acquire(v, new) \ + xchg_acquire(&((v)->counter), (new)) +#define atomic_xchg_release(v, new) \ + xchg_release(&((v)->counter), (new)) +#define atomic_xchg(v, new) \ + xchg(&((v)->counter), (new)) + +#define atomic_cmpxchg_relaxed(v, old, new) \ + cmpxchg_relaxed(&((v)->counter), (old), (new)) +#define atomic_cmpxchg_acquire(v, old, new) \ + cmpxchg_acquire(&((v)->counter), (old), (new)) +#define atomic_cmpxchg_release(v, old, new) \ + cmpxchg_release(&((v)->counter), (old), (new)) + +#define atomic_andnot atomic_andnot + +#endif /* __ASM_ARM_ARM64_ATOMIC_H */ \ No newline at end of file -- 2.24.3 (Apple Git-128)
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |