274 lines
6.0 KiB
C
274 lines
6.0 KiB
C
|
/* SPDX-License-Identifier: GPL-2.0 */
|
||
|
/*
|
||
|
* Copyright IBM Corp. 1999, 2011
|
||
|
*
|
||
|
* Author(s): Martin Schwidefsky <schwidefsky@de.ibm.com>,
|
||
|
*/
|
||
|
|
||
|
#ifndef __ASM_CMPXCHG_H
|
||
|
#define __ASM_CMPXCHG_H
|
||
|
|
||
|
#include <linux/mmdebug.h>
|
||
|
#include <linux/types.h>
|
||
|
#include <linux/bug.h>
|
||
|
#include <asm/asm.h>
|
||
|
|
||
|
void __cmpxchg_called_with_bad_pointer(void);
|
||
|
|
||
|
static __always_inline u32 __cs_asm(u64 ptr, u32 old, u32 new)
|
||
|
{
|
||
|
asm volatile(
|
||
|
" cs %[old],%[new],%[ptr]\n"
|
||
|
: [old] "+d" (old), [ptr] "+Q" (*(u32 *)ptr)
|
||
|
: [new] "d" (new)
|
||
|
: "memory", "cc");
|
||
|
return old;
|
||
|
}
|
||
|
|
||
|
static __always_inline u64 __csg_asm(u64 ptr, u64 old, u64 new)
|
||
|
{
|
||
|
asm volatile(
|
||
|
" csg %[old],%[new],%[ptr]\n"
|
||
|
: [old] "+d" (old), [ptr] "+QS" (*(u64 *)ptr)
|
||
|
: [new] "d" (new)
|
||
|
: "memory", "cc");
|
||
|
return old;
|
||
|
}
|
||
|
|
||
|
static inline u8 __arch_cmpxchg1(u64 ptr, u8 old, u8 new)
|
||
|
{
|
||
|
union {
|
||
|
u8 b[4];
|
||
|
u32 w;
|
||
|
} old32, new32;
|
||
|
u32 prev;
|
||
|
int i;
|
||
|
|
||
|
i = ptr & 3;
|
||
|
ptr &= ~0x3;
|
||
|
prev = READ_ONCE(*(u32 *)ptr);
|
||
|
do {
|
||
|
old32.w = prev;
|
||
|
if (old32.b[i] != old)
|
||
|
return old32.b[i];
|
||
|
new32.w = old32.w;
|
||
|
new32.b[i] = new;
|
||
|
prev = __cs_asm(ptr, old32.w, new32.w);
|
||
|
} while (prev != old32.w);
|
||
|
return old;
|
||
|
}
|
||
|
|
||
|
static inline u16 __arch_cmpxchg2(u64 ptr, u16 old, u16 new)
|
||
|
{
|
||
|
union {
|
||
|
u16 b[2];
|
||
|
u32 w;
|
||
|
} old32, new32;
|
||
|
u32 prev;
|
||
|
int i;
|
||
|
|
||
|
i = (ptr & 3) >> 1;
|
||
|
ptr &= ~0x3;
|
||
|
prev = READ_ONCE(*(u32 *)ptr);
|
||
|
do {
|
||
|
old32.w = prev;
|
||
|
if (old32.b[i] != old)
|
||
|
return old32.b[i];
|
||
|
new32.w = old32.w;
|
||
|
new32.b[i] = new;
|
||
|
prev = __cs_asm(ptr, old32.w, new32.w);
|
||
|
} while (prev != old32.w);
|
||
|
return old;
|
||
|
}
|
||
|
|
||
|
static __always_inline u64 __arch_cmpxchg(u64 ptr, u64 old, u64 new, int size)
|
||
|
{
|
||
|
switch (size) {
|
||
|
case 1: return __arch_cmpxchg1(ptr, old & 0xff, new & 0xff);
|
||
|
case 2: return __arch_cmpxchg2(ptr, old & 0xffff, new & 0xffff);
|
||
|
case 4: return __cs_asm(ptr, old & 0xffffffff, new & 0xffffffff);
|
||
|
case 8: return __csg_asm(ptr, old, new);
|
||
|
default: __cmpxchg_called_with_bad_pointer();
|
||
|
}
|
||
|
return old;
|
||
|
}
|
||
|
|
||
|
#define arch_cmpxchg(ptr, o, n) \
|
||
|
({ \
|
||
|
(__typeof__(*(ptr)))__arch_cmpxchg((unsigned long)(ptr), \
|
||
|
(unsigned long)(o), \
|
||
|
(unsigned long)(n), \
|
||
|
sizeof(*(ptr))); \
|
||
|
})
|
||
|
|
||
|
#define arch_cmpxchg64 arch_cmpxchg
|
||
|
#define arch_cmpxchg_local arch_cmpxchg
|
||
|
#define arch_cmpxchg64_local arch_cmpxchg
|
||
|
|
||
|
#ifdef __HAVE_ASM_FLAG_OUTPUTS__
|
||
|
|
||
|
#define arch_try_cmpxchg(ptr, oldp, new) \
|
||
|
({ \
|
||
|
__typeof__(ptr) __oldp = (__typeof__(ptr))(oldp); \
|
||
|
__typeof__(*(ptr)) __old = *__oldp; \
|
||
|
__typeof__(*(ptr)) __new = (new); \
|
||
|
__typeof__(*(ptr)) __prev; \
|
||
|
int __cc; \
|
||
|
\
|
||
|
switch (sizeof(*(ptr))) { \
|
||
|
case 1: \
|
||
|
case 2: { \
|
||
|
__prev = arch_cmpxchg((ptr), (__old), (__new)); \
|
||
|
__cc = (__prev != __old); \
|
||
|
if (unlikely(__cc)) \
|
||
|
*__oldp = __prev; \
|
||
|
break; \
|
||
|
} \
|
||
|
case 4: { \
|
||
|
asm volatile( \
|
||
|
" cs %[__old],%[__new],%[__ptr]\n" \
|
||
|
: [__old] "+d" (*__oldp), \
|
||
|
[__ptr] "+Q" (*(ptr)), \
|
||
|
"=@cc" (__cc) \
|
||
|
: [__new] "d" (__new) \
|
||
|
: "memory"); \
|
||
|
break; \
|
||
|
} \
|
||
|
case 8: { \
|
||
|
asm volatile( \
|
||
|
" csg %[__old],%[__new],%[__ptr]\n" \
|
||
|
: [__old] "+d" (*__oldp), \
|
||
|
[__ptr] "+QS" (*(ptr)), \
|
||
|
"=@cc" (__cc) \
|
||
|
: [__new] "d" (__new) \
|
||
|
: "memory"); \
|
||
|
break; \
|
||
|
} \
|
||
|
default: \
|
||
|
__cmpxchg_called_with_bad_pointer(); \
|
||
|
} \
|
||
|
likely(__cc == 0); \
|
||
|
})
|
||
|
|
||
|
#else /* __HAVE_ASM_FLAG_OUTPUTS__ */
|
||
|
|
||
|
#define arch_try_cmpxchg(ptr, oldp, new) \
|
||
|
({ \
|
||
|
__typeof__((ptr)) __oldp = (__typeof__(ptr))(oldp); \
|
||
|
__typeof__(*(ptr)) __old = *__oldp; \
|
||
|
__typeof__(*(ptr)) __new = (new); \
|
||
|
__typeof__(*(ptr)) __prev; \
|
||
|
\
|
||
|
__prev = arch_cmpxchg((ptr), (__old), (__new)); \
|
||
|
if (unlikely(__prev != __old)) \
|
||
|
*__oldp = __prev; \
|
||
|
likely(__prev == __old); \
|
||
|
})
|
||
|
|
||
|
#endif /* __HAVE_ASM_FLAG_OUTPUTS__ */
|
||
|
|
||
|
#define arch_try_cmpxchg64 arch_try_cmpxchg
|
||
|
#define arch_try_cmpxchg_local arch_try_cmpxchg
|
||
|
#define arch_try_cmpxchg64_local arch_try_cmpxchg
|
||
|
|
||
|
void __xchg_called_with_bad_pointer(void);
|
||
|
|
||
|
static inline u8 __arch_xchg1(u64 ptr, u8 x)
|
||
|
{
|
||
|
int shift = (3 ^ (ptr & 3)) << 3;
|
||
|
u32 mask, old, new;
|
||
|
|
||
|
ptr &= ~0x3;
|
||
|
mask = ~(0xff << shift);
|
||
|
old = READ_ONCE(*(u32 *)ptr);
|
||
|
do {
|
||
|
new = old & mask;
|
||
|
new |= x << shift;
|
||
|
} while (!arch_try_cmpxchg((u32 *)ptr, &old, new));
|
||
|
return old >> shift;
|
||
|
}
|
||
|
|
||
|
static inline u16 __arch_xchg2(u64 ptr, u16 x)
|
||
|
{
|
||
|
int shift = (2 ^ (ptr & 2)) << 3;
|
||
|
u32 mask, old, new;
|
||
|
|
||
|
ptr &= ~0x3;
|
||
|
mask = ~(0xffff << shift);
|
||
|
old = READ_ONCE(*(u32 *)ptr);
|
||
|
do {
|
||
|
new = old & mask;
|
||
|
new |= x << shift;
|
||
|
} while (!arch_try_cmpxchg((u32 *)ptr, &old, new));
|
||
|
return old >> shift;
|
||
|
}
|
||
|
|
||
|
static __always_inline u64 __arch_xchg(u64 ptr, u64 x, int size)
|
||
|
{
|
||
|
switch (size) {
|
||
|
case 1:
|
||
|
return __arch_xchg1(ptr, x & 0xff);
|
||
|
case 2:
|
||
|
return __arch_xchg2(ptr, x & 0xffff);
|
||
|
case 4: {
|
||
|
u32 old = READ_ONCE(*(u32 *)ptr);
|
||
|
|
||
|
do {
|
||
|
} while (!arch_try_cmpxchg((u32 *)ptr, &old, x & 0xffffffff));
|
||
|
return old;
|
||
|
}
|
||
|
case 8: {
|
||
|
u64 old = READ_ONCE(*(u64 *)ptr);
|
||
|
|
||
|
do {
|
||
|
} while (!arch_try_cmpxchg((u64 *)ptr, &old, x));
|
||
|
return old;
|
||
|
}
|
||
|
}
|
||
|
__xchg_called_with_bad_pointer();
|
||
|
return x;
|
||
|
}
|
||
|
|
||
|
#define arch_xchg(ptr, x) \
|
||
|
({ \
|
||
|
(__typeof__(*(ptr)))__arch_xchg((unsigned long)(ptr), \
|
||
|
(unsigned long)(x), \
|
||
|
sizeof(*(ptr))); \
|
||
|
})
|
||
|
|
||
|
#define system_has_cmpxchg128() 1
|
||
|
|
||
|
static __always_inline u128 arch_cmpxchg128(volatile u128 *ptr, u128 old, u128 new)
|
||
|
{
|
||
|
asm volatile(
|
||
|
" cdsg %[old],%[new],%[ptr]\n"
|
||
|
: [old] "+d" (old), [ptr] "+QS" (*ptr)
|
||
|
: [new] "d" (new)
|
||
|
: "memory", "cc");
|
||
|
return old;
|
||
|
}
|
||
|
|
||
|
#define arch_cmpxchg128 arch_cmpxchg128
|
||
|
#define arch_cmpxchg128_local arch_cmpxchg128
|
||
|
|
||
|
#ifdef __HAVE_ASM_FLAG_OUTPUTS__
|
||
|
|
||
|
static __always_inline bool arch_try_cmpxchg128(volatile u128 *ptr, u128 *oldp, u128 new)
|
||
|
{
|
||
|
int cc;
|
||
|
|
||
|
asm volatile(
|
||
|
" cdsg %[old],%[new],%[ptr]\n"
|
||
|
: [old] "+d" (*oldp), [ptr] "+QS" (*ptr), "=@cc" (cc)
|
||
|
: [new] "d" (new)
|
||
|
: "memory");
|
||
|
return likely(cc == 0);
|
||
|
}
|
||
|
|
||
|
#define arch_try_cmpxchg128 arch_try_cmpxchg128
|
||
|
#define arch_try_cmpxchg128_local arch_try_cmpxchg128
|
||
|
|
||
|
#endif /* __HAVE_ASM_FLAG_OUTPUTS__ */
|
||
|
|
||
|
#endif /* __ASM_CMPXCHG_H */
|