cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

cmpxchg.h (3188B)


      1/* SPDX-License-Identifier: GPL-2.0-only */
      2/*
      3 * Copyright (C) 2004, 2007-2010, 2011-2012 Synopsys, Inc. (www.synopsys.com)
      4 */
      5
      6#ifndef __ASM_ARC_CMPXCHG_H
      7#define __ASM_ARC_CMPXCHG_H
      8
      9#include <linux/build_bug.h>
     10#include <linux/types.h>
     11
     12#include <asm/barrier.h>
     13#include <asm/smp.h>
     14
     15#ifdef CONFIG_ARC_HAS_LLSC
     16
     17/*
     18 * if (*ptr == @old)
     19 *      *ptr = @new
     20 */
     21#define __cmpxchg(ptr, old, new)					\
     22({									\
     23	__typeof__(*(ptr)) _prev;					\
     24									\
     25	__asm__ __volatile__(						\
     26	"1:	llock  %0, [%1]	\n"					\
     27	"	brne   %0, %2, 2f	\n"				\
     28	"	scond  %3, [%1]	\n"					\
     29	"	bnz     1b		\n"				\
     30	"2:				\n"				\
     31	: "=&r"(_prev)	/* Early clobber prevent reg reuse */		\
     32	: "r"(ptr),	/* Not "m": llock only supports reg */		\
     33	  "ir"(old),							\
     34	  "r"(new)	/* Not "ir": scond can't take LIMM */		\
     35	: "cc",								\
     36	  "memory");	/* gcc knows memory is clobbered */		\
     37									\
     38	_prev;								\
     39})
     40
     41#define arch_cmpxchg_relaxed(ptr, old, new)				\
     42({									\
     43	__typeof__(ptr) _p_ = (ptr);					\
     44	__typeof__(*(ptr)) _o_ = (old);					\
     45	__typeof__(*(ptr)) _n_ = (new);					\
     46	__typeof__(*(ptr)) _prev_;					\
     47									\
     48	switch(sizeof((_p_))) {						\
     49	case 4:								\
     50		_prev_ = __cmpxchg(_p_, _o_, _n_);			\
     51		break;							\
     52	default:							\
     53		BUILD_BUG();						\
     54	}								\
     55	_prev_;								\
     56})
     57
     58#else
     59
     60#define arch_cmpxchg(ptr, old, new)				        \
     61({									\
     62	volatile __typeof__(ptr) _p_ = (ptr);				\
     63	__typeof__(*(ptr)) _o_ = (old);					\
     64	__typeof__(*(ptr)) _n_ = (new);					\
     65	__typeof__(*(ptr)) _prev_;					\
     66	unsigned long __flags;						\
     67									\
     68	BUILD_BUG_ON(sizeof(_p_) != 4);					\
     69									\
     70	/*								\
     71	 * spin lock/unlock provide the needed smp_mb() before/after	\
     72	 */								\
     73	atomic_ops_lock(__flags);					\
     74	_prev_ = *_p_;							\
     75	if (_prev_ == _o_)						\
     76		*_p_ = _n_;						\
     77	atomic_ops_unlock(__flags);					\
     78	_prev_;								\
     79})
     80
     81#endif
     82
     83/*
     84 * xchg
     85 */
     86#ifdef CONFIG_ARC_HAS_LLSC
     87
     88#define __xchg(ptr, val)						\
     89({									\
     90	__asm__ __volatile__(						\
     91	"	ex  %0, [%1]	\n"	/* set new value */	        \
     92	: "+r"(val)							\
     93	: "r"(ptr)							\
     94	: "memory");							\
     95	_val_;		/* get old value */				\
     96})
     97
     98#define arch_xchg_relaxed(ptr, val)					\
     99({									\
    100	__typeof__(ptr) _p_ = (ptr);					\
    101	__typeof__(*(ptr)) _val_ = (val);				\
    102									\
    103	switch(sizeof(*(_p_))) {					\
    104	case 4:								\
    105		_val_ = __xchg(_p_, _val_);				\
    106		break;							\
    107	default:							\
    108		BUILD_BUG();						\
    109	}								\
    110	_val_;								\
    111})
    112
    113#else  /* !CONFIG_ARC_HAS_LLSC */
    114
    115/*
    116 * EX instructions is baseline and present in !LLSC too. But in this
    117 * regime it still needs use @atomic_ops_lock spinlock to allow interop
    118 * with cmpxchg() which uses spinlock in !LLSC
    119 * (llist.h use xchg and cmpxchg on sama data)
    120 */
    121
    122#define arch_xchg(ptr, val)					        \
    123({									\
    124	__typeof__(ptr) _p_ = (ptr);					\
    125	__typeof__(*(ptr)) _val_ = (val);				\
    126									\
    127	unsigned long __flags;						\
    128									\
    129	atomic_ops_lock(__flags);					\
    130									\
    131	__asm__ __volatile__(						\
    132	"	ex  %0, [%1]	\n"					\
    133	: "+r"(_val_)							\
    134	: "r"(_p_)							\
    135	: "memory");							\
    136									\
    137	atomic_ops_unlock(__flags);					\
    138	_val_;								\
    139})
    140
    141#endif
    142
    143#endif