cmpxchg.h (3262B)
1/* SPDX-License-Identifier: GPL-2.0 */ 2 3#ifndef __ASM_CSKY_CMPXCHG_H 4#define __ASM_CSKY_CMPXCHG_H 5 6#ifdef CONFIG_SMP 7#include <asm/barrier.h> 8 9extern void __bad_xchg(void); 10 11#define __xchg_relaxed(new, ptr, size) \ 12({ \ 13 __typeof__(ptr) __ptr = (ptr); \ 14 __typeof__(new) __new = (new); \ 15 __typeof__(*(ptr)) __ret; \ 16 unsigned long tmp; \ 17 switch (size) { \ 18 case 4: \ 19 asm volatile ( \ 20 "1: ldex.w %0, (%3) \n" \ 21 " mov %1, %2 \n" \ 22 " stex.w %1, (%3) \n" \ 23 " bez %1, 1b \n" \ 24 : "=&r" (__ret), "=&r" (tmp) \ 25 : "r" (__new), "r"(__ptr) \ 26 :); \ 27 break; \ 28 default: \ 29 __bad_xchg(); \ 30 } \ 31 __ret; \ 32}) 33 34#define arch_xchg_relaxed(ptr, x) \ 35 (__xchg_relaxed((x), (ptr), sizeof(*(ptr)))) 36 37#define __cmpxchg_relaxed(ptr, old, new, size) \ 38({ \ 39 __typeof__(ptr) __ptr = (ptr); \ 40 __typeof__(new) __new = (new); \ 41 __typeof__(new) __tmp; \ 42 __typeof__(old) __old = (old); \ 43 __typeof__(*(ptr)) __ret; \ 44 switch (size) { \ 45 case 4: \ 46 asm volatile ( \ 47 "1: ldex.w %0, (%3) \n" \ 48 " cmpne %0, %4 \n" \ 49 " bt 2f \n" \ 50 " mov %1, %2 \n" \ 51 " stex.w %1, (%3) \n" \ 52 " bez %1, 1b \n" \ 53 "2: \n" \ 54 : "=&r" (__ret), "=&r" (__tmp) \ 55 : "r" (__new), "r"(__ptr), "r"(__old) \ 56 :); \ 57 break; \ 58 default: \ 59 __bad_xchg(); \ 60 } \ 61 __ret; \ 62}) 63 64#define arch_cmpxchg_relaxed(ptr, o, n) \ 65 (__cmpxchg_relaxed((ptr), (o), (n), sizeof(*(ptr)))) 66 67#define __cmpxchg_acquire(ptr, old, new, size) \ 68({ \ 69 __typeof__(ptr) __ptr = (ptr); \ 70 __typeof__(new) __new = (new); \ 71 __typeof__(new) __tmp; \ 72 __typeof__(old) __old = (old); \ 73 __typeof__(*(ptr)) __ret; \ 74 switch (size) { \ 75 case 4: \ 76 asm volatile ( \ 77 "1: ldex.w %0, (%3) \n" \ 78 " cmpne %0, %4 \n" \ 79 " bt 2f \n" \ 80 " mov %1, %2 \n" \ 81 " stex.w %1, (%3) \n" \ 82 " bez %1, 1b \n" \ 83 ACQUIRE_FENCE \ 84 "2: \n" \ 85 : "=&r" (__ret), "=&r" (__tmp) \ 86 : "r" (__new), "r"(__ptr), "r"(__old) \ 87 :); \ 88 break; \ 89 default: \ 90 __bad_xchg(); \ 91 } \ 92 __ret; \ 93}) 94 95#define arch_cmpxchg_acquire(ptr, o, n) \ 96 (__cmpxchg_acquire((ptr), (o), (n), sizeof(*(ptr)))) 97 98#define __cmpxchg(ptr, old, new, size) \ 99({ \ 100 __typeof__(ptr) __ptr = (ptr); \ 101 __typeof__(new) __new = (new); \ 102 __typeof__(new) __tmp; \ 103 __typeof__(old) __old = (old); \ 104 __typeof__(*(ptr)) __ret; \ 105 switch (size) { \ 106 case 4: \ 107 asm volatile ( \ 108 RELEASE_FENCE \ 109 "1: ldex.w %0, (%3) \n" \ 110 " cmpne %0, %4 \n" \ 111 " bt 2f \n" \ 112 " mov %1, %2 \n" \ 113 " stex.w %1, (%3) \n" \ 114 " bez %1, 1b \n" \ 115 FULL_FENCE \ 116 "2: \n" \ 117 : "=&r" (__ret), "=&r" (__tmp) \ 118 : "r" (__new), "r"(__ptr), "r"(__old) \ 119 :); \ 120 break; \ 121 default: \ 122 __bad_xchg(); \ 123 } \ 124 __ret; \ 125}) 126 127#define arch_cmpxchg(ptr, o, n) \ 128 (__cmpxchg((ptr), (o), (n), sizeof(*(ptr)))) 129 130#define arch_cmpxchg_local(ptr, o, n) \ 131 (__cmpxchg_relaxed((ptr), (o), (n), sizeof(*(ptr)))) 132#else 133#include <asm-generic/cmpxchg.h> 134#endif 135 136#endif /* __ASM_CSKY_CMPXCHG_H */