mmu-8xx.h (9776B)
1/* SPDX-License-Identifier: GPL-2.0 */ 2#ifndef _ASM_POWERPC_MMU_8XX_H_ 3#define _ASM_POWERPC_MMU_8XX_H_ 4/* 5 * PPC8xx support 6 */ 7 8/* Control/status registers for the MPC8xx. 9 * A write operation to these registers causes serialized access. 10 * During software tablewalk, the registers used perform mask/shift-add 11 * operations when written/read. A TLB entry is created when the Mx_RPN 12 * is written, and the contents of several registers are used to 13 * create the entry. 14 */ 15#define SPRN_MI_CTR 784 /* Instruction TLB control register */ 16#define MI_GPM 0x80000000 /* Set domain manager mode */ 17#define MI_PPM 0x40000000 /* Set subpage protection */ 18#define MI_CIDEF 0x20000000 /* Set cache inhibit when MMU dis */ 19#define MI_RSV4I 0x08000000 /* Reserve 4 TLB entries */ 20#define MI_PPCS 0x02000000 /* Use MI_RPN prob/priv state */ 21#define MI_IDXMASK 0x00001f00 /* TLB index to be loaded */ 22 23/* These are the Ks and Kp from the PowerPC books. For proper operation, 24 * Ks = 0, Kp = 1. 25 */ 26#define SPRN_MI_AP 786 27#define MI_Ks 0x80000000 /* Should not be set */ 28#define MI_Kp 0x40000000 /* Should always be set */ 29 30/* 31 * All pages' PP data bits are set to either 001 or 011 by copying _PAGE_EXEC 32 * into bit 21 in the ITLBmiss handler (bit 21 is the middle bit), which means 33 * respectively NA for All or X for Supervisor and no access for User. 34 * Then we use the APG to say whether accesses are according to Page rules or 35 * "all Supervisor" rules (Access to all) 36 * _PAGE_ACCESSED is also managed via APG. When _PAGE_ACCESSED is not set, say 37 * "all User" rules, that will lead to NA for all. 38 * Therefore, we define 4 APG groups. lsb is _PAGE_ACCESSED 39 * 0 => Kernel => 11 (all accesses performed according as user iaw page definition) 40 * 1 => Kernel+Accessed => 01 (all accesses performed according to page definition) 41 * 2 => User => 11 (all accesses performed according as user iaw page definition) 42 * 3 => User+Accessed => 10 (all accesses performed according to swaped page definition) for KUEP 43 * 4-15 => Not Used 44 */ 45#define MI_APG_INIT 0xde000000 46 47/* The effective page number register. When read, contains the information 48 * about the last instruction TLB miss. When MI_RPN is written, bits in 49 * this register are used to create the TLB entry. 50 */ 51#define SPRN_MI_EPN 787 52#define MI_EPNMASK 0xfffff000 /* Effective page number for entry */ 53#define MI_EVALID 0x00000200 /* Entry is valid */ 54#define MI_ASIDMASK 0x0000000f /* ASID match value */ 55 /* Reset value is undefined */ 56 57/* A "level 1" or "segment" or whatever you want to call it register. 58 * For the instruction TLB, it contains bits that get loaded into the 59 * TLB entry when the MI_RPN is written. 60 */ 61#define SPRN_MI_TWC 789 62#define MI_APG 0x000001e0 /* Access protection group (0) */ 63#define MI_GUARDED 0x00000010 /* Guarded storage */ 64#define MI_PSMASK 0x0000000c /* Mask of page size bits */ 65#define MI_PS8MEG 0x0000000c /* 8M page size */ 66#define MI_PS512K 0x00000004 /* 512K page size */ 67#define MI_PS4K_16K 0x00000000 /* 4K or 16K page size */ 68#define MI_SVALID 0x00000001 /* Segment entry is valid */ 69 /* Reset value is undefined */ 70 71/* Real page number. Defined by the pte. Writing this register 72 * causes a TLB entry to be created for the instruction TLB, using 73 * additional information from the MI_EPN, and MI_TWC registers. 74 */ 75#define SPRN_MI_RPN 790 76#define MI_SPS16K 0x00000008 /* Small page size (0 = 4k, 1 = 16k) */ 77 78/* Define an RPN value for mapping kernel memory to large virtual 79 * pages for boot initialization. This has real page number of 0, 80 * large page size, shared page, cache enabled, and valid. 81 * Also mark all subpages valid and write access. 82 */ 83#define MI_BOOTINIT 0x000001fd 84 85#define SPRN_MD_CTR 792 /* Data TLB control register */ 86#define MD_GPM 0x80000000 /* Set domain manager mode */ 87#define MD_PPM 0x40000000 /* Set subpage protection */ 88#define MD_CIDEF 0x20000000 /* Set cache inhibit when MMU dis */ 89#define MD_WTDEF 0x10000000 /* Set writethrough when MMU dis */ 90#define MD_RSV4I 0x08000000 /* Reserve 4 TLB entries */ 91#define MD_TWAM 0x04000000 /* Use 4K page hardware assist */ 92#define MD_PPCS 0x02000000 /* Use MI_RPN prob/priv state */ 93#define MD_IDXMASK 0x00001f00 /* TLB index to be loaded */ 94 95#define SPRN_M_CASID 793 /* Address space ID (context) to match */ 96#define MC_ASIDMASK 0x0000000f /* Bits used for ASID value */ 97 98 99/* These are the Ks and Kp from the PowerPC books. For proper operation, 100 * Ks = 0, Kp = 1. 101 */ 102#define SPRN_MD_AP 794 103#define MD_Ks 0x80000000 /* Should not be set */ 104#define MD_Kp 0x40000000 /* Should always be set */ 105 106/* See explanation above at the definition of MI_APG_INIT */ 107#define MD_APG_INIT 0xdc000000 108#define MD_APG_KUAP 0xde000000 109 110/* The effective page number register. When read, contains the information 111 * about the last instruction TLB miss. When MD_RPN is written, bits in 112 * this register are used to create the TLB entry. 113 */ 114#define SPRN_MD_EPN 795 115#define MD_EPNMASK 0xfffff000 /* Effective page number for entry */ 116#define MD_EVALID 0x00000200 /* Entry is valid */ 117#define MD_ASIDMASK 0x0000000f /* ASID match value */ 118 /* Reset value is undefined */ 119 120/* The pointer to the base address of the first level page table. 121 * During a software tablewalk, reading this register provides the address 122 * of the entry associated with MD_EPN. 123 */ 124#define SPRN_M_TWB 796 125#define M_L1TB 0xfffff000 /* Level 1 table base address */ 126#define M_L1INDX 0x00000ffc /* Level 1 index, when read */ 127 /* Reset value is undefined */ 128 129/* A "level 1" or "segment" or whatever you want to call it register. 130 * For the data TLB, it contains bits that get loaded into the TLB entry 131 * when the MD_RPN is written. It is also provides the hardware assist 132 * for finding the PTE address during software tablewalk. 133 */ 134#define SPRN_MD_TWC 797 135#define MD_L2TB 0xfffff000 /* Level 2 table base address */ 136#define MD_L2INDX 0xfffffe00 /* Level 2 index (*pte), when read */ 137#define MD_APG 0x000001e0 /* Access protection group (0) */ 138#define MD_GUARDED 0x00000010 /* Guarded storage */ 139#define MD_PSMASK 0x0000000c /* Mask of page size bits */ 140#define MD_PS8MEG 0x0000000c /* 8M page size */ 141#define MD_PS512K 0x00000004 /* 512K page size */ 142#define MD_PS4K_16K 0x00000000 /* 4K or 16K page size */ 143#define MD_WT 0x00000002 /* Use writethrough page attribute */ 144#define MD_SVALID 0x00000001 /* Segment entry is valid */ 145 /* Reset value is undefined */ 146 147 148/* Real page number. Defined by the pte. Writing this register 149 * causes a TLB entry to be created for the data TLB, using 150 * additional information from the MD_EPN, and MD_TWC registers. 151 */ 152#define SPRN_MD_RPN 798 153#define MD_SPS16K 0x00000008 /* Small page size (0 = 4k, 1 = 16k) */ 154 155/* This is a temporary storage register that could be used to save 156 * a processor working register during a tablewalk. 157 */ 158#define SPRN_M_TW 799 159 160#if defined(CONFIG_PPC_4K_PAGES) 161#define mmu_virtual_psize MMU_PAGE_4K 162#elif defined(CONFIG_PPC_16K_PAGES) 163#define mmu_virtual_psize MMU_PAGE_16K 164#define PTE_FRAG_NR 4 165#define PTE_FRAG_SIZE_SHIFT 12 166#define PTE_FRAG_SIZE (1UL << 12) 167#else 168#error "Unsupported PAGE_SIZE" 169#endif 170 171#define mmu_linear_psize MMU_PAGE_8M 172 173#define MODULES_VADDR (PAGE_OFFSET - SZ_256M) 174#define MODULES_END PAGE_OFFSET 175 176#ifndef __ASSEMBLY__ 177 178#include <linux/mmdebug.h> 179#include <linux/sizes.h> 180 181void mmu_pin_tlb(unsigned long top, bool readonly); 182 183typedef struct { 184 unsigned int id; 185 unsigned int active; 186 void __user *vdso; 187 void *pte_frag; 188} mm_context_t; 189 190#define PHYS_IMMR_BASE (mfspr(SPRN_IMMR) & 0xfff80000) 191#define VIRT_IMMR_BASE (__fix_to_virt(FIX_IMMR_BASE)) 192 193/* Page size definitions, common between 32 and 64-bit 194 * 195 * shift : is the "PAGE_SHIFT" value for that page size 196 * penc : is the pte encoding mask 197 * 198 */ 199struct mmu_psize_def { 200 unsigned int shift; /* number of bits */ 201 unsigned int enc; /* PTE encoding */ 202 unsigned int ind; /* Corresponding indirect page size shift */ 203 unsigned int flags; 204#define MMU_PAGE_SIZE_DIRECT 0x1 /* Supported as a direct size */ 205#define MMU_PAGE_SIZE_INDIRECT 0x2 /* Supported as an indirect size */ 206}; 207 208extern struct mmu_psize_def mmu_psize_defs[MMU_PAGE_COUNT]; 209 210static inline int shift_to_mmu_psize(unsigned int shift) 211{ 212 int psize; 213 214 for (psize = 0; psize < MMU_PAGE_COUNT; ++psize) 215 if (mmu_psize_defs[psize].shift == shift) 216 return psize; 217 return -1; 218} 219 220static inline unsigned int mmu_psize_to_shift(unsigned int mmu_psize) 221{ 222 if (mmu_psize_defs[mmu_psize].shift) 223 return mmu_psize_defs[mmu_psize].shift; 224 BUG(); 225} 226 227static inline bool arch_vmap_try_size(unsigned long addr, unsigned long end, u64 pfn, 228 unsigned int max_page_shift, unsigned long size) 229{ 230 if (end - addr < size) 231 return false; 232 233 if ((1UL << max_page_shift) < size) 234 return false; 235 236 if (!IS_ALIGNED(addr, size)) 237 return false; 238 239 if (!IS_ALIGNED(PFN_PHYS(pfn), size)) 240 return false; 241 242 return true; 243} 244 245static inline unsigned long arch_vmap_pte_range_map_size(unsigned long addr, unsigned long end, 246 u64 pfn, unsigned int max_page_shift) 247{ 248 if (arch_vmap_try_size(addr, end, pfn, max_page_shift, SZ_512K)) 249 return SZ_512K; 250 if (PAGE_SIZE == SZ_16K) 251 return SZ_16K; 252 if (arch_vmap_try_size(addr, end, pfn, max_page_shift, SZ_16K)) 253 return SZ_16K; 254 return PAGE_SIZE; 255} 256#define arch_vmap_pte_range_map_size arch_vmap_pte_range_map_size 257 258static inline int arch_vmap_pte_supported_shift(unsigned long size) 259{ 260 if (size >= SZ_512K) 261 return 19; 262 else if (size >= SZ_16K) 263 return 14; 264 else 265 return PAGE_SHIFT; 266} 267#define arch_vmap_pte_supported_shift arch_vmap_pte_supported_shift 268 269/* patch sites */ 270extern s32 patch__itlbmiss_exit_1, patch__dtlbmiss_exit_1; 271extern s32 patch__itlbmiss_perf, patch__dtlbmiss_perf; 272 273#endif /* !__ASSEMBLY__ */ 274 275#endif /* _ASM_POWERPC_MMU_8XX_H_ */