8xx-pmu.c (4788B)
1// SPDX-License-Identifier: GPL-2.0-or-later 2/* 3 * Performance event support - PPC 8xx 4 * 5 * Copyright 2016 Christophe Leroy, CS Systemes d'Information 6 */ 7 8#include <linux/kernel.h> 9#include <linux/sched.h> 10#include <linux/perf_event.h> 11#include <linux/percpu.h> 12#include <linux/hardirq.h> 13#include <asm/pmc.h> 14#include <asm/machdep.h> 15#include <asm/firmware.h> 16#include <asm/ptrace.h> 17#include <asm/code-patching.h> 18#include <asm/inst.h> 19 20#define PERF_8xx_ID_CPU_CYCLES 1 21#define PERF_8xx_ID_HW_INSTRUCTIONS 2 22#define PERF_8xx_ID_ITLB_LOAD_MISS 3 23#define PERF_8xx_ID_DTLB_LOAD_MISS 4 24 25#define C(x) PERF_COUNT_HW_CACHE_##x 26#define DTLB_LOAD_MISS (C(DTLB) | (C(OP_READ) << 8) | (C(RESULT_MISS) << 16)) 27#define ITLB_LOAD_MISS (C(ITLB) | (C(OP_READ) << 8) | (C(RESULT_MISS) << 16)) 28 29extern unsigned long itlb_miss_counter, dtlb_miss_counter; 30extern atomic_t instruction_counter; 31 32static atomic_t insn_ctr_ref; 33static atomic_t itlb_miss_ref; 34static atomic_t dtlb_miss_ref; 35 36static s64 get_insn_ctr(void) 37{ 38 int ctr; 39 unsigned long counta; 40 41 do { 42 ctr = atomic_read(&instruction_counter); 43 counta = mfspr(SPRN_COUNTA); 44 } while (ctr != atomic_read(&instruction_counter)); 45 46 return ((s64)ctr << 16) | (counta >> 16); 47} 48 49static int event_type(struct perf_event *event) 50{ 51 switch (event->attr.type) { 52 case PERF_TYPE_HARDWARE: 53 if (event->attr.config == PERF_COUNT_HW_CPU_CYCLES) 54 return PERF_8xx_ID_CPU_CYCLES; 55 if (event->attr.config == PERF_COUNT_HW_INSTRUCTIONS) 56 return PERF_8xx_ID_HW_INSTRUCTIONS; 57 break; 58 case PERF_TYPE_HW_CACHE: 59 if (event->attr.config == ITLB_LOAD_MISS) 60 return PERF_8xx_ID_ITLB_LOAD_MISS; 61 if (event->attr.config == DTLB_LOAD_MISS) 62 return PERF_8xx_ID_DTLB_LOAD_MISS; 63 break; 64 case PERF_TYPE_RAW: 65 break; 66 default: 67 return -ENOENT; 68 } 69 return -EOPNOTSUPP; 70} 71 72static int mpc8xx_pmu_event_init(struct perf_event *event) 73{ 74 int type = event_type(event); 75 76 if (type < 0) 77 return type; 78 return 0; 79} 80 81static int mpc8xx_pmu_add(struct perf_event *event, int flags) 82{ 83 int type = event_type(event); 84 s64 val = 0; 85 86 if (type < 0) 87 return type; 88 89 switch (type) { 90 case PERF_8xx_ID_CPU_CYCLES: 91 val = get_tb(); 92 break; 93 case PERF_8xx_ID_HW_INSTRUCTIONS: 94 if (atomic_inc_return(&insn_ctr_ref) == 1) 95 mtspr(SPRN_ICTRL, 0xc0080007); 96 val = get_insn_ctr(); 97 break; 98 case PERF_8xx_ID_ITLB_LOAD_MISS: 99 if (atomic_inc_return(&itlb_miss_ref) == 1) { 100 unsigned long target = patch_site_addr(&patch__itlbmiss_perf); 101 102 patch_branch_site(&patch__itlbmiss_exit_1, target, 0); 103 } 104 val = itlb_miss_counter; 105 break; 106 case PERF_8xx_ID_DTLB_LOAD_MISS: 107 if (atomic_inc_return(&dtlb_miss_ref) == 1) { 108 unsigned long target = patch_site_addr(&patch__dtlbmiss_perf); 109 110 patch_branch_site(&patch__dtlbmiss_exit_1, target, 0); 111 } 112 val = dtlb_miss_counter; 113 break; 114 } 115 local64_set(&event->hw.prev_count, val); 116 return 0; 117} 118 119static void mpc8xx_pmu_read(struct perf_event *event) 120{ 121 int type = event_type(event); 122 s64 prev, val = 0, delta = 0; 123 124 if (type < 0) 125 return; 126 127 do { 128 prev = local64_read(&event->hw.prev_count); 129 switch (type) { 130 case PERF_8xx_ID_CPU_CYCLES: 131 val = get_tb(); 132 delta = 16 * (val - prev); 133 break; 134 case PERF_8xx_ID_HW_INSTRUCTIONS: 135 val = get_insn_ctr(); 136 delta = prev - val; 137 if (delta < 0) 138 delta += 0x1000000000000LL; 139 break; 140 case PERF_8xx_ID_ITLB_LOAD_MISS: 141 val = itlb_miss_counter; 142 delta = (s64)((s32)val - (s32)prev); 143 break; 144 case PERF_8xx_ID_DTLB_LOAD_MISS: 145 val = dtlb_miss_counter; 146 delta = (s64)((s32)val - (s32)prev); 147 break; 148 } 149 } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); 150 151 local64_add(delta, &event->count); 152} 153 154static void mpc8xx_pmu_del(struct perf_event *event, int flags) 155{ 156 ppc_inst_t insn = ppc_inst(PPC_RAW_MFSPR(10, SPRN_SPRG_SCRATCH2)); 157 158 mpc8xx_pmu_read(event); 159 160 /* If it was the last user, stop counting to avoid useless overhead */ 161 switch (event_type(event)) { 162 case PERF_8xx_ID_CPU_CYCLES: 163 break; 164 case PERF_8xx_ID_HW_INSTRUCTIONS: 165 if (atomic_dec_return(&insn_ctr_ref) == 0) 166 mtspr(SPRN_ICTRL, 7); 167 break; 168 case PERF_8xx_ID_ITLB_LOAD_MISS: 169 if (atomic_dec_return(&itlb_miss_ref) == 0) 170 patch_instruction_site(&patch__itlbmiss_exit_1, insn); 171 break; 172 case PERF_8xx_ID_DTLB_LOAD_MISS: 173 if (atomic_dec_return(&dtlb_miss_ref) == 0) 174 patch_instruction_site(&patch__dtlbmiss_exit_1, insn); 175 break; 176 } 177} 178 179static struct pmu mpc8xx_pmu = { 180 .event_init = mpc8xx_pmu_event_init, 181 .add = mpc8xx_pmu_add, 182 .del = mpc8xx_pmu_del, 183 .read = mpc8xx_pmu_read, 184 .capabilities = PERF_PMU_CAP_NO_INTERRUPT | 185 PERF_PMU_CAP_NO_NMI, 186}; 187 188static int init_mpc8xx_pmu(void) 189{ 190 mtspr(SPRN_ICTRL, 7); 191 mtspr(SPRN_CMPA, 0); 192 mtspr(SPRN_COUNTA, 0xffff); 193 194 return perf_pmu_register(&mpc8xx_pmu, "cpu", PERF_TYPE_RAW); 195} 196 197early_initcall(init_mpc8xx_pmu);