cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

backoff.h (2756B)


      1/* SPDX-License-Identifier: GPL-2.0 */
      2#ifndef _SPARC64_BACKOFF_H
      3#define _SPARC64_BACKOFF_H
      4
      5/* The macros in this file implement an exponential backoff facility
      6 * for atomic operations.
      7 *
      8 * When multiple threads compete on an atomic operation, it is
      9 * possible for one thread to be continually denied a successful
     10 * completion of the compare-and-swap instruction.  Heavily
     11 * threaded cpu implementations like Niagara can compound this
     12 * problem even further.
     13 *
     14 * When an atomic operation fails and needs to be retried, we spin a
     15 * certain number of times.  At each subsequent failure of the same
     16 * operation we double the spin count, realizing an exponential
     17 * backoff.
     18 *
     19 * When we spin, we try to use an operation that will cause the
     20 * current cpu strand to block, and therefore make the core fully
     21 * available to any other runnable strands.  There are two
     22 * options, based upon cpu capabilities.
     23 *
     24 * On all cpus prior to SPARC-T4 we do three dummy reads of the
     25 * condition code register.  Each read blocks the strand for something
     26 * between 40 and 50 cpu cycles.
     27 *
     28 * For SPARC-T4 and later we have a special "pause" instruction
     29 * available.  This is implemented using writes to register %asr27.
     30 * The cpu will block the number of cycles written into the register,
     31 * unless a disrupting trap happens first.  SPARC-T4 specifically
     32 * implements pause with a granularity of 8 cycles.  Each strand has
     33 * an internal pause counter which decrements every 8 cycles.  So the
     34 * chip shifts the %asr27 value down by 3 bits, and writes the result
     35 * into the pause counter.  If a value smaller than 8 is written, the
     36 * chip blocks for 1 cycle.
     37 *
     38 * To achieve the same amount of backoff as the three %ccr reads give
     39 * on earlier chips, we shift the backoff value up by 7 bits.  (Three
     40 * %ccr reads block for about 128 cycles, 1 << 7 == 128) We write the
     41 * whole amount we want to block into the pause register, rather than
     42 * loop writing 128 each time.
     43 */
     44
     45#define BACKOFF_LIMIT	(4 * 1024)
     46
     47#ifdef CONFIG_SMP
     48
     49#define BACKOFF_SETUP(reg)	\
     50	mov	1, reg
     51
     52#define BACKOFF_LABEL(spin_label, continue_label) \
     53	spin_label
     54
     55#define BACKOFF_SPIN(reg, tmp, label)		\
     56	mov		reg, tmp;		\
     5788:	rd		%ccr, %g0;		\
     58	rd		%ccr, %g0;		\
     59	rd		%ccr, %g0;		\
     60	.section	.pause_3insn_patch,"ax";\
     61	.word		88b;			\
     62	sllx		tmp, 7, tmp;		\
     63	wr		tmp, 0, %asr27;		\
     64	clr		tmp;			\
     65	.previous;				\
     66	brnz,pt		tmp, 88b;		\
     67	 sub		tmp, 1, tmp;		\
     68	set		BACKOFF_LIMIT, tmp;	\
     69	cmp		reg, tmp;		\
     70	bg,pn		%xcc, label;		\
     71	 nop;					\
     72	ba,pt		%xcc, label;		\
     73	 sllx		reg, 1, reg;
     74
     75#else
     76
     77#define BACKOFF_SETUP(reg)
     78
     79#define BACKOFF_LABEL(spin_label, continue_label) \
     80	continue_label
     81
     82#define BACKOFF_SPIN(reg, tmp, label)
     83
     84#endif
     85
     86#endif /* _SPARC64_BACKOFF_H */