cache.S (7369B)
1/* SPDX-License-Identifier: GPL-2.0-or-later */ 2/* 3 * This file contains low-level cache management functions 4 * used for sleep and CPU speed changes on Apple machines. 5 * (In fact the only thing that is Apple-specific is that we assume 6 * that we can read from ROM at physical address 0xfff00000.) 7 * 8 * Copyright (C) 2004 Paul Mackerras (paulus@samba.org) and 9 * Benjamin Herrenschmidt (benh@kernel.crashing.org) 10 */ 11 12#include <asm/processor.h> 13#include <asm/ppc_asm.h> 14#include <asm/cputable.h> 15#include <asm/feature-fixups.h> 16 17/* 18 * Flush and disable all data caches (dL1, L2, L3). This is used 19 * when going to sleep, when doing a PMU based cpufreq transition, 20 * or when "offlining" a CPU on SMP machines. This code is over 21 * paranoid, but I've had enough issues with various CPU revs and 22 * bugs that I decided it was worth being over cautious 23 */ 24 25_GLOBAL(flush_disable_caches) 26#ifndef CONFIG_PPC_BOOK3S_32 27 blr 28#else 29BEGIN_FTR_SECTION 30 b flush_disable_745x 31END_FTR_SECTION_IFSET(CPU_FTR_SPEC7450) 32BEGIN_FTR_SECTION 33 b flush_disable_75x 34END_FTR_SECTION_IFSET(CPU_FTR_L2CR) 35 b __flush_disable_L1 36 37/* This is the code for G3 and 74[01]0 */ 38flush_disable_75x: 39 mflr r10 40 41 /* Turn off EE and DR in MSR */ 42 mfmsr r11 43 rlwinm r0,r11,0,~MSR_EE 44 rlwinm r0,r0,0,~MSR_DR 45 sync 46 mtmsr r0 47 isync 48 49 /* Stop DST streams */ 50BEGIN_FTR_SECTION 51 PPC_DSSALL 52 sync 53END_FTR_SECTION_IFSET(CPU_FTR_ALTIVEC) 54 55 /* Stop DPM */ 56 mfspr r8,SPRN_HID0 /* Save SPRN_HID0 in r8 */ 57 rlwinm r4,r8,0,12,10 /* Turn off HID0[DPM] */ 58 sync 59 mtspr SPRN_HID0,r4 /* Disable DPM */ 60 sync 61 62 /* Disp-flush L1. We have a weird problem here that I never 63 * totally figured out. On 750FX, using the ROM for the flush 64 * results in a non-working flush. We use that workaround for 65 * now until I finally understand what's going on. --BenH 66 */ 67 68 /* ROM base by default */ 69 lis r4,0xfff0 70 mfpvr r3 71 srwi r3,r3,16 72 cmplwi cr0,r3,0x7000 73 bne+ 1f 74 /* RAM base on 750FX */ 75 li r4,0 761: li r4,0x4000 77 mtctr r4 781: lwz r0,0(r4) 79 addi r4,r4,32 80 bdnz 1b 81 sync 82 isync 83 84 /* Disable / invalidate / enable L1 data */ 85 mfspr r3,SPRN_HID0 86 rlwinm r3,r3,0,~(HID0_DCE | HID0_ICE) 87 mtspr SPRN_HID0,r3 88 sync 89 isync 90 ori r3,r3,(HID0_DCE|HID0_DCI|HID0_ICE|HID0_ICFI) 91 sync 92 isync 93 mtspr SPRN_HID0,r3 94 xori r3,r3,(HID0_DCI|HID0_ICFI) 95 mtspr SPRN_HID0,r3 96 sync 97 98 /* Get the current enable bit of the L2CR into r4 */ 99 mfspr r5,SPRN_L2CR 100 /* Set to data-only (pre-745x bit) */ 101 oris r3,r5,L2CR_L2DO@h 102 b 2f 103 /* When disabling L2, code must be in L1 */ 104 .balign 32 1051: mtspr SPRN_L2CR,r3 1063: sync 107 isync 108 b 1f 1092: b 3f 1103: sync 111 isync 112 b 1b 1131: /* disp-flush L2. The interesting thing here is that the L2 can be 114 * up to 2Mb ... so using the ROM, we'll end up wrapping back to memory 115 * but that is probbaly fine. We disp-flush over 4Mb to be safe 116 */ 117 lis r4,2 118 mtctr r4 119 lis r4,0xfff0 1201: lwz r0,0(r4) 121 addi r4,r4,32 122 bdnz 1b 123 sync 124 isync 125 lis r4,2 126 mtctr r4 127 lis r4,0xfff0 1281: dcbf 0,r4 129 addi r4,r4,32 130 bdnz 1b 131 sync 132 isync 133 134 /* now disable L2 */ 135 rlwinm r5,r5,0,~L2CR_L2E 136 b 2f 137 /* When disabling L2, code must be in L1 */ 138 .balign 32 1391: mtspr SPRN_L2CR,r5 1403: sync 141 isync 142 b 1f 1432: b 3f 1443: sync 145 isync 146 b 1b 1471: sync 148 isync 149 /* Invalidate L2. This is pre-745x, we clear the L2I bit ourselves */ 150 oris r4,r5,L2CR_L2I@h 151 mtspr SPRN_L2CR,r4 152 sync 153 isync 154 155 /* Wait for the invalidation to complete */ 1561: mfspr r3,SPRN_L2CR 157 rlwinm. r0,r3,0,31,31 158 bne 1b 159 160 /* Clear L2I */ 161 xoris r4,r4,L2CR_L2I@h 162 sync 163 mtspr SPRN_L2CR,r4 164 sync 165 166 /* now disable the L1 data cache */ 167 mfspr r0,SPRN_HID0 168 rlwinm r0,r0,0,~(HID0_DCE|HID0_ICE) 169 mtspr SPRN_HID0,r0 170 sync 171 isync 172 173 /* Restore HID0[DPM] to whatever it was before */ 174 sync 175 mfspr r0,SPRN_HID0 176 rlwimi r0,r8,0,11,11 /* Turn back HID0[DPM] */ 177 mtspr SPRN_HID0,r0 178 sync 179 180 /* restore DR and EE */ 181 sync 182 mtmsr r11 183 isync 184 185 mtlr r10 186 blr 187_ASM_NOKPROBE_SYMBOL(flush_disable_75x) 188 189/* This code is for 745x processors */ 190flush_disable_745x: 191 /* Turn off EE and DR in MSR */ 192 mfmsr r11 193 rlwinm r0,r11,0,~MSR_EE 194 rlwinm r0,r0,0,~MSR_DR 195 sync 196 mtmsr r0 197 isync 198 199 /* Stop prefetch streams */ 200 PPC_DSSALL 201 sync 202 203 /* Disable L2 prefetching */ 204 mfspr r0,SPRN_MSSCR0 205 rlwinm r0,r0,0,0,29 206 mtspr SPRN_MSSCR0,r0 207 sync 208 isync 209 lis r4,0 210 dcbf 0,r4 211 dcbf 0,r4 212 dcbf 0,r4 213 dcbf 0,r4 214 dcbf 0,r4 215 dcbf 0,r4 216 dcbf 0,r4 217 dcbf 0,r4 218 219 /* Due to a bug with the HW flush on some CPU revs, we occasionally 220 * experience data corruption. I'm adding a displacement flush along 221 * with a dcbf loop over a few Mb to "help". The problem isn't totally 222 * fixed by this in theory, but at least, in practice, I couldn't reproduce 223 * it even with a big hammer... 224 */ 225 226 lis r4,0x0002 227 mtctr r4 228 li r4,0 2291: 230 lwz r0,0(r4) 231 addi r4,r4,32 /* Go to start of next cache line */ 232 bdnz 1b 233 isync 234 235 /* Now, flush the first 4MB of memory */ 236 lis r4,0x0002 237 mtctr r4 238 li r4,0 239 sync 2401: 241 dcbf 0,r4 242 addi r4,r4,32 /* Go to start of next cache line */ 243 bdnz 1b 244 245 /* Flush and disable the L1 data cache */ 246 mfspr r6,SPRN_LDSTCR 247 lis r3,0xfff0 /* read from ROM for displacement flush */ 248 li r4,0xfe /* start with only way 0 unlocked */ 249 li r5,128 /* 128 lines in each way */ 2501: mtctr r5 251 rlwimi r6,r4,0,24,31 252 mtspr SPRN_LDSTCR,r6 253 sync 254 isync 2552: lwz r0,0(r3) /* touch each cache line */ 256 addi r3,r3,32 257 bdnz 2b 258 rlwinm r4,r4,1,24,30 /* move on to the next way */ 259 ori r4,r4,1 260 cmpwi r4,0xff /* all done? */ 261 bne 1b 262 /* now unlock the L1 data cache */ 263 li r4,0 264 rlwimi r6,r4,0,24,31 265 sync 266 mtspr SPRN_LDSTCR,r6 267 sync 268 isync 269 270 /* Flush the L2 cache using the hardware assist */ 271 mfspr r3,SPRN_L2CR 272 cmpwi r3,0 /* check if it is enabled first */ 273 bge 4f 274 oris r0,r3,(L2CR_L2IO_745x|L2CR_L2DO_745x)@h 275 b 2f 276 /* When disabling/locking L2, code must be in L1 */ 277 .balign 32 2781: mtspr SPRN_L2CR,r0 /* lock the L2 cache */ 2793: sync 280 isync 281 b 1f 2822: b 3f 2833: sync 284 isync 285 b 1b 2861: sync 287 isync 288 ori r0,r3,L2CR_L2HWF_745x 289 sync 290 mtspr SPRN_L2CR,r0 /* set the hardware flush bit */ 2913: mfspr r0,SPRN_L2CR /* wait for it to go to 0 */ 292 andi. r0,r0,L2CR_L2HWF_745x 293 bne 3b 294 sync 295 rlwinm r3,r3,0,~L2CR_L2E 296 b 2f 297 /* When disabling L2, code must be in L1 */ 298 .balign 32 2991: mtspr SPRN_L2CR,r3 /* disable the L2 cache */ 3003: sync 301 isync 302 b 1f 3032: b 3f 3043: sync 305 isync 306 b 1b 3071: sync 308 isync 309 oris r4,r3,L2CR_L2I@h 310 mtspr SPRN_L2CR,r4 311 sync 312 isync 3131: mfspr r4,SPRN_L2CR 314 andis. r0,r4,L2CR_L2I@h 315 bne 1b 316 sync 317 318BEGIN_FTR_SECTION 319 /* Flush the L3 cache using the hardware assist */ 3204: mfspr r3,SPRN_L3CR 321 cmpwi r3,0 /* check if it is enabled */ 322 bge 6f 323 oris r0,r3,L3CR_L3IO@h 324 ori r0,r0,L3CR_L3DO 325 sync 326 mtspr SPRN_L3CR,r0 /* lock the L3 cache */ 327 sync 328 isync 329 ori r0,r0,L3CR_L3HWF 330 sync 331 mtspr SPRN_L3CR,r0 /* set the hardware flush bit */ 3325: mfspr r0,SPRN_L3CR /* wait for it to go to zero */ 333 andi. r0,r0,L3CR_L3HWF 334 bne 5b 335 rlwinm r3,r3,0,~L3CR_L3E 336 sync 337 mtspr SPRN_L3CR,r3 /* disable the L3 cache */ 338 sync 339 ori r4,r3,L3CR_L3I 340 mtspr SPRN_L3CR,r4 3411: mfspr r4,SPRN_L3CR 342 andi. r0,r4,L3CR_L3I 343 bne 1b 344 sync 345END_FTR_SECTION_IFSET(CPU_FTR_L3CR) 346 3476: mfspr r0,SPRN_HID0 /* now disable the L1 data cache */ 348 rlwinm r0,r0,0,~HID0_DCE 349 mtspr SPRN_HID0,r0 350 sync 351 isync 352 mtmsr r11 /* restore DR and EE */ 353 isync 354 blr 355_ASM_NOKPROBE_SYMBOL(flush_disable_745x) 356#endif /* CONFIG_PPC_BOOK3S_32 */