14244065bSChristopher Covington /* 24244065bSChristopher Covington * Test the ARM Performance Monitors Unit (PMU). 34244065bSChristopher Covington * 44244065bSChristopher Covington * Copyright (c) 2015-2016, The Linux Foundation. All rights reserved. 54244065bSChristopher Covington * Copyright (C) 2016, Red Hat Inc, Wei Huang <wei@redhat.com> 64244065bSChristopher Covington * 74244065bSChristopher Covington * This program is free software; you can redistribute it and/or modify it 84244065bSChristopher Covington * under the terms of the GNU Lesser General Public License version 2.1 and 94244065bSChristopher Covington * only version 2.1 as published by the Free Software Foundation. 104244065bSChristopher Covington * 114244065bSChristopher Covington * This program is distributed in the hope that it will be useful, but WITHOUT 124244065bSChristopher Covington * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 134244065bSChristopher Covington * FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public License 144244065bSChristopher Covington * for more details. 154244065bSChristopher Covington */ 164244065bSChristopher Covington #include "libcflat.h" 174c357610SAndrew Jones #include "errata.h" 184244065bSChristopher Covington #include "asm/barrier.h" 194244065bSChristopher Covington #include "asm/sysreg.h" 204244065bSChristopher Covington #include "asm/processor.h" 214870738cSEric Auger #include <bitops.h> 224ce2a804SEric Auger #include <asm/gic.h> 234244065bSChristopher Covington 24d81bb7a3SChristopher Covington #define PMU_PMCR_E (1 << 0) 254ce2a804SEric Auger #define PMU_PMCR_P (1 << 1) 26d81bb7a3SChristopher Covington #define PMU_PMCR_C (1 << 2) 274ce2a804SEric Auger #define PMU_PMCR_D (1 << 3) 284ce2a804SEric Auger #define PMU_PMCR_X (1 << 4) 294ce2a804SEric Auger #define PMU_PMCR_DP (1 << 5) 30d81bb7a3SChristopher Covington #define PMU_PMCR_LC (1 << 6) 31036369c5SRicardo Koller #define PMU_PMCR_LP (1 << 7) 324244065bSChristopher Covington #define PMU_PMCR_N_SHIFT 11 334244065bSChristopher Covington #define PMU_PMCR_N_MASK 0x1f 344244065bSChristopher Covington #define PMU_PMCR_ID_SHIFT 16 354244065bSChristopher Covington #define PMU_PMCR_ID_MASK 0xff 364244065bSChristopher Covington #define PMU_PMCR_IMP_SHIFT 24 374244065bSChristopher Covington #define PMU_PMCR_IMP_MASK 0xff 384244065bSChristopher Covington 39d81bb7a3SChristopher Covington #define PMU_CYCLE_IDX 31 40d81bb7a3SChristopher Covington 41d81bb7a3SChristopher Covington #define NR_SAMPLES 10 42d81bb7a3SChristopher Covington 434870738cSEric Auger /* Some PMU events */ 444870738cSEric Auger #define SW_INCR 0x0 454870738cSEric Auger #define INST_RETIRED 0x8 464870738cSEric Auger #define CPU_CYCLES 0x11 474ce2a804SEric Auger #define MEM_ACCESS 0x13 484870738cSEric Auger #define INST_PREC 0x1B 494870738cSEric Auger #define STALL_FRONTEND 0x23 504870738cSEric Auger #define STALL_BACKEND 0x24 5166fee034SEric Auger #define CHAIN 0x1E 524870738cSEric Auger 534870738cSEric Auger #define COMMON_EVENTS_LOW 0x0 544870738cSEric Auger #define COMMON_EVENTS_HIGH 0x3F 554870738cSEric Auger #define EXT_COMMON_EVENTS_LOW 0x4000 564870738cSEric Auger #define EXT_COMMON_EVENTS_HIGH 0x403F 574870738cSEric Auger 587d1f853aSRicardo Koller #define ALL_SET_32 0x00000000FFFFFFFFULL 5939d1347aSRicardo Koller #define ALL_CLEAR 0x0000000000000000ULL 607d1f853aSRicardo Koller #define PRE_OVERFLOW_32 0x00000000FFFFFFF0ULL 61036369c5SRicardo Koller #define PRE_OVERFLOW_64 0xFFFFFFFFFFFFFFF0ULL 6291806724SEric Auger #define COUNT 20 6391806724SEric Auger #define MARGIN 15 6491806724SEric Auger /* 6591806724SEric Auger * PRE_OVERFLOW2 is set so that 1st @COUNT iterations do not 6691806724SEric Auger * produce 32b overflow and 2nd @COUNT iterations do. To accommodate 6791806724SEric Auger * for some observed variability we take into account a given @MARGIN 6891806724SEric Auger */ 6991806724SEric Auger #define PRE_OVERFLOW2_32 (ALL_SET_32 - COUNT - MARGIN) 70036369c5SRicardo Koller 71036369c5SRicardo Koller #define PRE_OVERFLOW(__overflow_at_64bits) \ 72036369c5SRicardo Koller (__overflow_at_64bits ? PRE_OVERFLOW_64 : PRE_OVERFLOW_32) 734ce2a804SEric Auger 744f5ef94fSEric Auger #define PMU_PPI 23 754f5ef94fSEric Auger 768f747a85SEric Auger struct pmu { 778f747a85SEric Auger unsigned int version; 788f747a85SEric Auger unsigned int nb_implemented_counters; 798f747a85SEric Auger uint32_t pmcr_ro; 808f747a85SEric Auger }; 818f747a85SEric Auger 824f5ef94fSEric Auger struct pmu_stats { 834f5ef94fSEric Auger unsigned long bitmap; 844f5ef94fSEric Auger uint32_t interrupts[32]; 854f5ef94fSEric Auger bool unexpected; 864f5ef94fSEric Auger }; 874f5ef94fSEric Auger 888f747a85SEric Auger static struct pmu pmu; 898f747a85SEric Auger 904244065bSChristopher Covington #if defined(__arm__) 91098add54SAndrew Jones #define ID_DFR0_PERFMON_SHIFT 24 92098add54SAndrew Jones #define ID_DFR0_PERFMON_MASK 0xf 93098add54SAndrew Jones 94784ee933SEric Auger #define ID_DFR0_PMU_NOTIMPL 0b0000 95784ee933SEric Auger #define ID_DFR0_PMU_V1 0b0001 96784ee933SEric Auger #define ID_DFR0_PMU_V2 0b0010 97784ee933SEric Auger #define ID_DFR0_PMU_V3 0b0011 98784ee933SEric Auger #define ID_DFR0_PMU_V3_8_1 0b0100 99784ee933SEric Auger #define ID_DFR0_PMU_V3_8_4 0b0101 100784ee933SEric Auger #define ID_DFR0_PMU_V3_8_5 0b0110 101784ee933SEric Auger #define ID_DFR0_PMU_IMPDEF 0b1111 102784ee933SEric Auger 1034244065bSChristopher Covington #define PMCR __ACCESS_CP15(c9, 0, c12, 0) 1044244065bSChristopher Covington #define ID_DFR0 __ACCESS_CP15(c0, 0, c1, 2) 105d81bb7a3SChristopher Covington #define PMSELR __ACCESS_CP15(c9, 0, c12, 5) 106d81bb7a3SChristopher Covington #define PMXEVTYPER __ACCESS_CP15(c9, 0, c13, 1) 107d81bb7a3SChristopher Covington #define PMCNTENSET __ACCESS_CP15(c9, 0, c12, 1) 108a7326740SRicardo Koller #define PMCNTENCLR __ACCESS_CP15(c9, 0, c12, 2) 109a7326740SRicardo Koller #define PMOVSR __ACCESS_CP15(c9, 0, c12, 3) 110d81bb7a3SChristopher Covington #define PMCCNTR32 __ACCESS_CP15(c9, 0, c13, 0) 111a7326740SRicardo Koller #define PMINTENCLR __ACCESS_CP15(c9, 0, c14, 2) 112d81bb7a3SChristopher Covington #define PMCCNTR64 __ACCESS_CP15_64(0, c9) 1134244065bSChristopher Covington 1144244065bSChristopher Covington static inline uint32_t get_id_dfr0(void) { return read_sysreg(ID_DFR0); } 1154244065bSChristopher Covington static inline uint32_t get_pmcr(void) { return read_sysreg(PMCR); } 116d81bb7a3SChristopher Covington static inline void set_pmcr(uint32_t v) { write_sysreg(v, PMCR); } 117d81bb7a3SChristopher Covington static inline void set_pmcntenset(uint32_t v) { write_sysreg(v, PMCNTENSET); } 118d81bb7a3SChristopher Covington 119098add54SAndrew Jones static inline uint8_t get_pmu_version(void) 120098add54SAndrew Jones { 121098add54SAndrew Jones return (get_id_dfr0() >> ID_DFR0_PERFMON_SHIFT) & ID_DFR0_PERFMON_MASK; 122098add54SAndrew Jones } 123098add54SAndrew Jones 124d81bb7a3SChristopher Covington static inline uint64_t get_pmccntr(void) 125d81bb7a3SChristopher Covington { 126d81bb7a3SChristopher Covington return read_sysreg(PMCCNTR32); 127d81bb7a3SChristopher Covington } 128d81bb7a3SChristopher Covington 1298f76a347SChristopher Covington static inline void set_pmccntr(uint64_t value) 1308f76a347SChristopher Covington { 1318f76a347SChristopher Covington write_sysreg(value & 0xffffffff, PMCCNTR32); 1328f76a347SChristopher Covington } 1338f76a347SChristopher Covington 134d81bb7a3SChristopher Covington /* PMCCFILTR is an obsolete name for PMXEVTYPER31 in ARMv7 */ 135d81bb7a3SChristopher Covington static inline void set_pmccfiltr(uint32_t value) 136d81bb7a3SChristopher Covington { 137d81bb7a3SChristopher Covington write_sysreg(PMU_CYCLE_IDX, PMSELR); 138d81bb7a3SChristopher Covington write_sysreg(value, PMXEVTYPER); 139d81bb7a3SChristopher Covington isb(); 140d81bb7a3SChristopher Covington } 1418f76a347SChristopher Covington 1428f76a347SChristopher Covington /* 1438f76a347SChristopher Covington * Extra instructions inserted by the compiler would be difficult to compensate 1448f76a347SChristopher Covington * for, so hand assemble everything between, and including, the PMCR accesses 1458f76a347SChristopher Covington * to start and stop counting. isb instructions were inserted to make sure 1468f76a347SChristopher Covington * pmccntr read after this function returns the exact instructions executed in 1478f76a347SChristopher Covington * the controlled block. Total instrs = isb + mcr + 2*loop = 2 + 2*loop. 1488f76a347SChristopher Covington */ 1498f76a347SChristopher Covington static inline void precise_instrs_loop(int loop, uint32_t pmcr) 1508f76a347SChristopher Covington { 1518f76a347SChristopher Covington asm volatile( 1528f76a347SChristopher Covington " mcr p15, 0, %[pmcr], c9, c12, 0\n" 1538f76a347SChristopher Covington " isb\n" 1548f76a347SChristopher Covington "1: subs %[loop], %[loop], #1\n" 1558f76a347SChristopher Covington " bgt 1b\n" 1568f76a347SChristopher Covington " mcr p15, 0, %[z], c9, c12, 0\n" 1578f76a347SChristopher Covington " isb\n" 1588f76a347SChristopher Covington : [loop] "+r" (loop) 1598f76a347SChristopher Covington : [pmcr] "r" (pmcr), [z] "r" (0) 1608f76a347SChristopher Covington : "cc"); 1618f76a347SChristopher Covington } 1624870738cSEric Auger 163a7326740SRicardo Koller static void pmu_reset(void) 164a7326740SRicardo Koller { 165a7326740SRicardo Koller /* reset all counters, counting disabled at PMCR level*/ 166a7326740SRicardo Koller set_pmcr(pmu.pmcr_ro | PMU_PMCR_LC | PMU_PMCR_C | PMU_PMCR_P); 167a7326740SRicardo Koller /* Disable all counters */ 1687d1f853aSRicardo Koller write_sysreg(ALL_SET_32, PMCNTENCLR); 169a7326740SRicardo Koller /* clear overflow reg */ 1707d1f853aSRicardo Koller write_sysreg(ALL_SET_32, PMOVSR); 171a7326740SRicardo Koller /* disable overflow interrupts on all counters */ 1727d1f853aSRicardo Koller write_sysreg(ALL_SET_32, PMINTENCLR); 173a7326740SRicardo Koller isb(); 174a7326740SRicardo Koller } 175a7326740SRicardo Koller 1764870738cSEric Auger /* event counter tests only implemented for aarch64 */ 1774870738cSEric Auger static void test_event_introspection(void) {} 1784ce2a804SEric Auger static void test_event_counter_config(void) {} 179041df25bSRicardo Koller static void test_basic_event_count(bool overflow_at_64bits) {} 180041df25bSRicardo Koller static void test_mem_access(bool overflow_at_64bits) {} 181041df25bSRicardo Koller static void test_sw_incr(bool overflow_at_64bits) {} 182041df25bSRicardo Koller static void test_chained_counters(bool unused) {} 183041df25bSRicardo Koller static void test_chained_sw_incr(bool unused) {} 184041df25bSRicardo Koller static void test_chain_promotion(bool unused) {} 185041df25bSRicardo Koller static void test_overflow_interrupt(bool overflow_at_64bits) {} 1864870738cSEric Auger 1874244065bSChristopher Covington #elif defined(__aarch64__) 188098add54SAndrew Jones #define ID_AA64DFR0_PERFMON_SHIFT 8 189098add54SAndrew Jones #define ID_AA64DFR0_PERFMON_MASK 0xf 190098add54SAndrew Jones 191784ee933SEric Auger #define ID_DFR0_PMU_NOTIMPL 0b0000 192784ee933SEric Auger #define ID_DFR0_PMU_V3 0b0001 193784ee933SEric Auger #define ID_DFR0_PMU_V3_8_1 0b0100 194784ee933SEric Auger #define ID_DFR0_PMU_V3_8_4 0b0101 195784ee933SEric Auger #define ID_DFR0_PMU_V3_8_5 0b0110 196784ee933SEric Auger #define ID_DFR0_PMU_IMPDEF 0b1111 197784ee933SEric Auger 198098add54SAndrew Jones static inline uint32_t get_id_aa64dfr0(void) { return read_sysreg(id_aa64dfr0_el1); } 1994244065bSChristopher Covington static inline uint32_t get_pmcr(void) { return read_sysreg(pmcr_el0); } 200d81bb7a3SChristopher Covington static inline void set_pmcr(uint32_t v) { write_sysreg(v, pmcr_el0); } 201d81bb7a3SChristopher Covington static inline uint64_t get_pmccntr(void) { return read_sysreg(pmccntr_el0); } 2028f76a347SChristopher Covington static inline void set_pmccntr(uint64_t v) { write_sysreg(v, pmccntr_el0); } 203d81bb7a3SChristopher Covington static inline void set_pmcntenset(uint32_t v) { write_sysreg(v, pmcntenset_el0); } 204d81bb7a3SChristopher Covington static inline void set_pmccfiltr(uint32_t v) { write_sysreg(v, pmccfiltr_el0); } 2058f76a347SChristopher Covington 206098add54SAndrew Jones static inline uint8_t get_pmu_version(void) 207098add54SAndrew Jones { 208098add54SAndrew Jones uint8_t ver = (get_id_aa64dfr0() >> ID_AA64DFR0_PERFMON_SHIFT) & ID_AA64DFR0_PERFMON_MASK; 209784ee933SEric Auger return ver; 210098add54SAndrew Jones } 211098add54SAndrew Jones 2128f76a347SChristopher Covington /* 2138f76a347SChristopher Covington * Extra instructions inserted by the compiler would be difficult to compensate 2148f76a347SChristopher Covington * for, so hand assemble everything between, and including, the PMCR accesses 2158f76a347SChristopher Covington * to start and stop counting. isb instructions are inserted to make sure 2168f76a347SChristopher Covington * pmccntr read after this function returns the exact instructions executed 2178f76a347SChristopher Covington * in the controlled block. Total instrs = isb + msr + 2*loop = 2 + 2*loop. 2188f76a347SChristopher Covington */ 2198f76a347SChristopher Covington static inline void precise_instrs_loop(int loop, uint32_t pmcr) 2208f76a347SChristopher Covington { 2219e186511SThomas Huth uint64_t pmcr64 = pmcr; 2228f76a347SChristopher Covington asm volatile( 2238f76a347SChristopher Covington " msr pmcr_el0, %[pmcr]\n" 2248f76a347SChristopher Covington " isb\n" 2259e186511SThomas Huth "1: subs %w[loop], %w[loop], #1\n" 2268f76a347SChristopher Covington " b.gt 1b\n" 2278f76a347SChristopher Covington " msr pmcr_el0, xzr\n" 2288f76a347SChristopher Covington " isb\n" 2298f76a347SChristopher Covington : [loop] "+r" (loop) 2309e186511SThomas Huth : [pmcr] "r" (pmcr64) 2318f76a347SChristopher Covington : "cc"); 2328f76a347SChristopher Covington } 2334870738cSEric Auger 2344870738cSEric Auger #define PMCEID1_EL0 sys_reg(3, 3, 9, 12, 7) 2354ce2a804SEric Auger #define PMCNTENSET_EL0 sys_reg(3, 3, 9, 12, 1) 2364ce2a804SEric Auger #define PMCNTENCLR_EL0 sys_reg(3, 3, 9, 12, 2) 2374ce2a804SEric Auger 2384ce2a804SEric Auger #define PMEVTYPER_EXCLUDE_EL1 BIT(31) 2394ce2a804SEric Auger #define PMEVTYPER_EXCLUDE_EL0 BIT(30) 2404870738cSEric Auger 2414870738cSEric Auger static bool is_event_supported(uint32_t n, bool warn) 2424870738cSEric Auger { 2434870738cSEric Auger uint64_t pmceid0 = read_sysreg(pmceid0_el0); 2444870738cSEric Auger uint64_t pmceid1 = read_sysreg_s(PMCEID1_EL0); 2454870738cSEric Auger bool supported; 2464870738cSEric Auger uint64_t reg; 2474870738cSEric Auger 2484870738cSEric Auger /* 2494870738cSEric Auger * The low 32-bits of PMCEID0/1 respectively describe 2504870738cSEric Auger * event support for events 0-31/32-63. Their High 2514870738cSEric Auger * 32-bits describe support for extended events 2524870738cSEric Auger * starting at 0x4000, using the same split. 2534870738cSEric Auger */ 2544870738cSEric Auger assert((n >= COMMON_EVENTS_LOW && n <= COMMON_EVENTS_HIGH) || 2554870738cSEric Auger (n >= EXT_COMMON_EVENTS_LOW && n <= EXT_COMMON_EVENTS_HIGH)); 2564870738cSEric Auger 2574870738cSEric Auger if (n <= COMMON_EVENTS_HIGH) 2584870738cSEric Auger reg = lower_32_bits(pmceid0) | ((u64)lower_32_bits(pmceid1) << 32); 2594870738cSEric Auger else 2604870738cSEric Auger reg = upper_32_bits(pmceid0) | ((u64)upper_32_bits(pmceid1) << 32); 2614870738cSEric Auger 2624870738cSEric Auger supported = reg & (1UL << (n & 0x3F)); 2634870738cSEric Auger 2644870738cSEric Auger if (!supported && warn) 2654870738cSEric Auger report_info("event 0x%x is not supported", n); 2664870738cSEric Auger return supported; 2674870738cSEric Auger } 2684870738cSEric Auger 2694870738cSEric Auger static void test_event_introspection(void) 2704870738cSEric Auger { 2714870738cSEric Auger bool required_events; 2724870738cSEric Auger 2734870738cSEric Auger if (!pmu.nb_implemented_counters) { 2744870738cSEric Auger report_skip("No event counter, skip ..."); 2754870738cSEric Auger return; 2764870738cSEric Auger } 2774870738cSEric Auger 2784870738cSEric Auger /* PMUv3 requires an implementation includes some common events */ 2794870738cSEric Auger required_events = is_event_supported(SW_INCR, true) && 2804870738cSEric Auger is_event_supported(CPU_CYCLES, true) && 2814870738cSEric Auger (is_event_supported(INST_RETIRED, true) || 2824870738cSEric Auger is_event_supported(INST_PREC, true)); 2834870738cSEric Auger 2844870738cSEric Auger if (pmu.version >= ID_DFR0_PMU_V3_8_1) { 2854870738cSEric Auger required_events = required_events && 2864870738cSEric Auger is_event_supported(STALL_FRONTEND, true) && 2874870738cSEric Auger is_event_supported(STALL_BACKEND, true); 2884870738cSEric Auger } 2894870738cSEric Auger 2904870738cSEric Auger report(required_events, "Check required events are implemented"); 2914870738cSEric Auger } 2924870738cSEric Auger 2934ce2a804SEric Auger /* 2944ce2a804SEric Auger * Extra instructions inserted by the compiler would be difficult to compensate 2954ce2a804SEric Auger * for, so hand assemble everything between, and including, the PMCR accesses 2964ce2a804SEric Auger * to start and stop counting. isb instructions are inserted to make sure 2974ce2a804SEric Auger * pmccntr read after this function returns the exact instructions executed 2984ce2a804SEric Auger * in the controlled block. Loads @loop times the data at @address into x9. 2994ce2a804SEric Auger */ 3009e186511SThomas Huth static void mem_access_loop(void *addr, long loop, uint32_t pmcr) 3014ce2a804SEric Auger { 3029e186511SThomas Huth uint64_t pmcr64 = pmcr; 3034ce2a804SEric Auger asm volatile( 3049f2f6819SEric Auger " dsb ish\n" 3054ce2a804SEric Auger " msr pmcr_el0, %[pmcr]\n" 3064ce2a804SEric Auger " isb\n" 3079f2f6819SEric Auger " dsb ish\n" 3084ce2a804SEric Auger " mov x10, %[loop]\n" 3094ce2a804SEric Auger "1: sub x10, x10, #1\n" 3104ce2a804SEric Auger " ldr x9, [%[addr]]\n" 3114ce2a804SEric Auger " cmp x10, #0x0\n" 3124ce2a804SEric Auger " b.gt 1b\n" 3139f2f6819SEric Auger " dsb ish\n" 3144ce2a804SEric Auger " msr pmcr_el0, xzr\n" 3154ce2a804SEric Auger " isb\n" 3164ce2a804SEric Auger : 3179e186511SThomas Huth : [addr] "r" (addr), [pmcr] "r" (pmcr64), [loop] "r" (loop) 3184ce2a804SEric Auger : "x9", "x10", "cc"); 3194ce2a804SEric Auger } 3204ce2a804SEric Auger 3214f5ef94fSEric Auger static struct pmu_stats pmu_stats; 3224f5ef94fSEric Auger 3234f5ef94fSEric Auger static void irq_handler(struct pt_regs *regs) 3244f5ef94fSEric Auger { 3254f5ef94fSEric Auger uint32_t irqstat, irqnr; 3264f5ef94fSEric Auger 3274f5ef94fSEric Auger irqstat = gic_read_iar(); 3284f5ef94fSEric Auger irqnr = gic_iar_irqnr(irqstat); 3294f5ef94fSEric Auger 3304f5ef94fSEric Auger if (irqnr == PMU_PPI) { 3314f5ef94fSEric Auger unsigned long overflows = read_sysreg(pmovsclr_el0); 3324f5ef94fSEric Auger int i; 3334f5ef94fSEric Auger 3344f5ef94fSEric Auger for (i = 0; i < 32; i++) { 3354f5ef94fSEric Auger if (test_and_clear_bit(i, &overflows)) { 3364f5ef94fSEric Auger pmu_stats.interrupts[i]++; 3374f5ef94fSEric Auger pmu_stats.bitmap |= 1 << i; 3384f5ef94fSEric Auger } 3394f5ef94fSEric Auger } 3407d1f853aSRicardo Koller write_sysreg(ALL_SET_32, pmovsclr_el0); 341e0a6e56bSRicardo Koller isb(); 3424f5ef94fSEric Auger } else { 3434f5ef94fSEric Auger pmu_stats.unexpected = true; 3444f5ef94fSEric Auger } 3454f5ef94fSEric Auger gic_write_eoir(irqstat); 3464f5ef94fSEric Auger } 3474f5ef94fSEric Auger 3484f5ef94fSEric Auger static void pmu_reset_stats(void) 3494f5ef94fSEric Auger { 3504f5ef94fSEric Auger int i; 3514f5ef94fSEric Auger 3524f5ef94fSEric Auger for (i = 0; i < 32; i++) 3534f5ef94fSEric Auger pmu_stats.interrupts[i] = 0; 3544f5ef94fSEric Auger 3554f5ef94fSEric Auger pmu_stats.bitmap = 0; 3564f5ef94fSEric Auger pmu_stats.unexpected = false; 3574f5ef94fSEric Auger } 3584f5ef94fSEric Auger 3594ce2a804SEric Auger static void pmu_reset(void) 3604ce2a804SEric Auger { 3614ce2a804SEric Auger /* reset all counters, counting disabled at PMCR level*/ 3624ce2a804SEric Auger set_pmcr(pmu.pmcr_ro | PMU_PMCR_LC | PMU_PMCR_C | PMU_PMCR_P); 3634ce2a804SEric Auger /* Disable all counters */ 3647d1f853aSRicardo Koller write_sysreg_s(ALL_SET_32, PMCNTENCLR_EL0); 3654ce2a804SEric Auger /* clear overflow reg */ 3667d1f853aSRicardo Koller write_sysreg(ALL_SET_32, pmovsclr_el0); 3674ce2a804SEric Auger /* disable overflow interrupts on all counters */ 3687d1f853aSRicardo Koller write_sysreg(ALL_SET_32, pmintenclr_el1); 3694f5ef94fSEric Auger pmu_reset_stats(); 3704ce2a804SEric Auger isb(); 3714ce2a804SEric Auger } 3724ce2a804SEric Auger 3734ce2a804SEric Auger static void test_event_counter_config(void) 3744ce2a804SEric Auger { 3754ce2a804SEric Auger int i; 3764ce2a804SEric Auger 3774ce2a804SEric Auger if (!pmu.nb_implemented_counters) { 3784ce2a804SEric Auger report_skip("No event counter, skip ..."); 3794ce2a804SEric Auger return; 3804ce2a804SEric Auger } 3814ce2a804SEric Auger 3824ce2a804SEric Auger pmu_reset(); 3834ce2a804SEric Auger 3844ce2a804SEric Auger /* 3854ce2a804SEric Auger * Test setting through PMESELR/PMXEVTYPER and PMEVTYPERn read, 3864ce2a804SEric Auger * select counter 0 3874ce2a804SEric Auger */ 3884ce2a804SEric Auger write_sysreg(1, PMSELR_EL0); 3894ce2a804SEric Auger /* program this counter to count unsupported event */ 3904ce2a804SEric Auger write_sysreg(0xEA, PMXEVTYPER_EL0); 3914ce2a804SEric Auger write_sysreg(0xdeadbeef, PMXEVCNTR_EL0); 3924ce2a804SEric Auger report((read_regn_el0(pmevtyper, 1) & 0xFFF) == 0xEA, 3934ce2a804SEric Auger "PMESELR/PMXEVTYPER/PMEVTYPERn"); 3944ce2a804SEric Auger report((read_regn_el0(pmevcntr, 1) == 0xdeadbeef), 3954ce2a804SEric Auger "PMESELR/PMXEVCNTR/PMEVCNTRn"); 3964ce2a804SEric Auger 3974ce2a804SEric Auger /* try to configure an unsupported event within the range [0x0, 0x3F] */ 3984ce2a804SEric Auger for (i = 0; i <= 0x3F; i++) { 3994ce2a804SEric Auger if (!is_event_supported(i, false)) 4004ce2a804SEric Auger break; 4014ce2a804SEric Auger } 4024ce2a804SEric Auger if (i > 0x3F) { 4034ce2a804SEric Auger report_skip("pmevtyper: all events within [0x0, 0x3F] are supported"); 4044ce2a804SEric Auger return; 4054ce2a804SEric Auger } 4064ce2a804SEric Auger 4074ce2a804SEric Auger /* select counter 0 */ 4084ce2a804SEric Auger write_sysreg(0, PMSELR_EL0); 4094ce2a804SEric Auger /* program this counter to count unsupported event */ 4104ce2a804SEric Auger write_sysreg(i, PMXEVCNTR_EL0); 4114ce2a804SEric Auger /* read the counter value */ 4124ce2a804SEric Auger read_sysreg(PMXEVCNTR_EL0); 4134ce2a804SEric Auger report(read_sysreg(PMXEVCNTR_EL0) == i, 4144ce2a804SEric Auger "read of a counter programmed with unsupported event"); 4154ce2a804SEric Auger } 4164ce2a804SEric Auger 4174ce2a804SEric Auger static bool satisfy_prerequisites(uint32_t *events, unsigned int nb_events) 4184ce2a804SEric Auger { 4194ce2a804SEric Auger int i; 4204ce2a804SEric Auger 4214ce2a804SEric Auger if (pmu.nb_implemented_counters < nb_events) { 4224ce2a804SEric Auger report_skip("Skip test as number of counters is too small (%d)", 4234ce2a804SEric Auger pmu.nb_implemented_counters); 4244ce2a804SEric Auger return false; 4254ce2a804SEric Auger } 4264ce2a804SEric Auger 4274ce2a804SEric Auger for (i = 0; i < nb_events; i++) { 4284ce2a804SEric Auger if (!is_event_supported(events[i], false)) { 4294ce2a804SEric Auger report_skip("Skip test as event 0x%x is not supported", 4304ce2a804SEric Auger events[i]); 4314ce2a804SEric Auger return false; 4324ce2a804SEric Auger } 4334ce2a804SEric Auger } 4344ce2a804SEric Auger return true; 4354ce2a804SEric Auger } 4364ce2a804SEric Auger 43739d1347aSRicardo Koller static uint64_t pmevcntr_mask(void) 43839d1347aSRicardo Koller { 43939d1347aSRicardo Koller /* 44039d1347aSRicardo Koller * Bits [63:0] are always incremented for 64-bit counters, 44139d1347aSRicardo Koller * even if the PMU is configured to generate an overflow at 44239d1347aSRicardo Koller * bits [31:0] 44339d1347aSRicardo Koller * 44439d1347aSRicardo Koller * For more details see the AArch64.IncrementEventCounter() 44539d1347aSRicardo Koller * pseudo-code in the ARM ARM DDI 0487I.a, section J1.1.1. 44639d1347aSRicardo Koller */ 44739d1347aSRicardo Koller if (pmu.version >= ID_DFR0_PMU_V3_8_5) 44839d1347aSRicardo Koller return ~0; 44939d1347aSRicardo Koller 45039d1347aSRicardo Koller return (uint32_t)~0; 45139d1347aSRicardo Koller } 45239d1347aSRicardo Koller 453041df25bSRicardo Koller static bool check_overflow_prerequisites(bool overflow_at_64bits) 454041df25bSRicardo Koller { 455041df25bSRicardo Koller if (overflow_at_64bits && pmu.version < ID_DFR0_PMU_V3_8_5) { 456041df25bSRicardo Koller report_skip("Skip test as 64 overflows need FEAT_PMUv3p5"); 457041df25bSRicardo Koller return false; 458041df25bSRicardo Koller } 459041df25bSRicardo Koller 460041df25bSRicardo Koller return true; 461041df25bSRicardo Koller } 462041df25bSRicardo Koller 463041df25bSRicardo Koller static void test_basic_event_count(bool overflow_at_64bits) 4644ce2a804SEric Auger { 4654ce2a804SEric Auger uint32_t implemented_counter_mask, non_implemented_counter_mask; 466036369c5SRicardo Koller uint64_t pre_overflow = PRE_OVERFLOW(overflow_at_64bits); 467036369c5SRicardo Koller uint64_t pmcr_lp = overflow_at_64bits ? PMU_PMCR_LP : 0; 4684ce2a804SEric Auger uint32_t events[] = {CPU_CYCLES, INST_RETIRED}; 469036369c5SRicardo Koller uint32_t counter_mask; 4704ce2a804SEric Auger 471041df25bSRicardo Koller if (!satisfy_prerequisites(events, ARRAY_SIZE(events)) || 472041df25bSRicardo Koller !check_overflow_prerequisites(overflow_at_64bits)) 4734ce2a804SEric Auger return; 4744ce2a804SEric Auger 4754ce2a804SEric Auger implemented_counter_mask = BIT(pmu.nb_implemented_counters) - 1; 4764ce2a804SEric Auger non_implemented_counter_mask = ~(BIT(31) | implemented_counter_mask); 4774ce2a804SEric Auger counter_mask = implemented_counter_mask | non_implemented_counter_mask; 4784ce2a804SEric Auger 4794ce2a804SEric Auger write_regn_el0(pmevtyper, 0, CPU_CYCLES | PMEVTYPER_EXCLUDE_EL0); 4804ce2a804SEric Auger write_regn_el0(pmevtyper, 1, INST_RETIRED | PMEVTYPER_EXCLUDE_EL0); 4814ce2a804SEric Auger 4824ce2a804SEric Auger /* disable all counters */ 4837d1f853aSRicardo Koller write_sysreg_s(ALL_SET_32, PMCNTENCLR_EL0); 4844ce2a804SEric Auger report(!read_sysreg_s(PMCNTENCLR_EL0) && !read_sysreg_s(PMCNTENSET_EL0), 4854ce2a804SEric Auger "pmcntenclr: disable all counters"); 4864ce2a804SEric Auger 4874ce2a804SEric Auger /* 4884ce2a804SEric Auger * clear cycle and all event counters and allow counter enablement 4894ce2a804SEric Auger * through PMCNTENSET. LC is RES1. 4904ce2a804SEric Auger */ 491036369c5SRicardo Koller set_pmcr(pmu.pmcr_ro | PMU_PMCR_LC | PMU_PMCR_C | PMU_PMCR_P | pmcr_lp); 4924ce2a804SEric Auger isb(); 493036369c5SRicardo Koller report(get_pmcr() == (pmu.pmcr_ro | PMU_PMCR_LC | pmcr_lp), "pmcr: reset counters"); 4944ce2a804SEric Auger 4954ce2a804SEric Auger /* Preset counter #0 to pre overflow value to trigger an overflow */ 496036369c5SRicardo Koller write_regn_el0(pmevcntr, 0, pre_overflow); 497036369c5SRicardo Koller report(read_regn_el0(pmevcntr, 0) == pre_overflow, 4984ce2a804SEric Auger "counter #0 preset to pre-overflow value"); 4994ce2a804SEric Auger report(!read_regn_el0(pmevcntr, 1), "counter #1 is 0"); 5004ce2a804SEric Auger 5014ce2a804SEric Auger /* 5024ce2a804SEric Auger * Enable all implemented counters and also attempt to enable 5034ce2a804SEric Auger * not supported counters. Counting still is disabled by !PMCR.E 5044ce2a804SEric Auger */ 5054ce2a804SEric Auger write_sysreg_s(counter_mask, PMCNTENSET_EL0); 5064ce2a804SEric Auger 5074ce2a804SEric Auger /* check only those implemented are enabled */ 5084ce2a804SEric Auger report((read_sysreg_s(PMCNTENSET_EL0) == read_sysreg_s(PMCNTENCLR_EL0)) && 5094ce2a804SEric Auger (read_sysreg_s(PMCNTENSET_EL0) == implemented_counter_mask), 5104ce2a804SEric Auger "pmcntenset: enabled implemented_counters"); 5114ce2a804SEric Auger 5124ce2a804SEric Auger /* Disable all counters but counters #0 and #1 */ 5134ce2a804SEric Auger write_sysreg_s(~0x3, PMCNTENCLR_EL0); 5144ce2a804SEric Auger report((read_sysreg_s(PMCNTENSET_EL0) == read_sysreg_s(PMCNTENCLR_EL0)) && 5154ce2a804SEric Auger (read_sysreg_s(PMCNTENSET_EL0) == 0x3), 5164ce2a804SEric Auger "pmcntenset: just enabled #0 and #1"); 5174ce2a804SEric Auger 5184ce2a804SEric Auger /* clear overflow register */ 5197d1f853aSRicardo Koller write_sysreg(ALL_SET_32, pmovsclr_el0); 5204ce2a804SEric Auger report(!read_sysreg(pmovsclr_el0), "check overflow reg is 0"); 5214ce2a804SEric Auger 5224ce2a804SEric Auger /* disable overflow interrupts on all counters*/ 5237d1f853aSRicardo Koller write_sysreg(ALL_SET_32, pmintenclr_el1); 5244ce2a804SEric Auger report(!read_sysreg(pmintenclr_el1), 5254ce2a804SEric Auger "pmintenclr_el1=0, all interrupts disabled"); 5264ce2a804SEric Auger 5274ce2a804SEric Auger /* enable overflow interrupts on all event counters */ 5284ce2a804SEric Auger write_sysreg(implemented_counter_mask | non_implemented_counter_mask, 5294ce2a804SEric Auger pmintenset_el1); 5304ce2a804SEric Auger report(read_sysreg(pmintenset_el1) == implemented_counter_mask, 5314ce2a804SEric Auger "overflow interrupts enabled on all implemented counters"); 5324ce2a804SEric Auger 5334ce2a804SEric Auger /* Set PMCR.E, execute asm code and unset PMCR.E */ 5344ce2a804SEric Auger precise_instrs_loop(20, pmu.pmcr_ro | PMU_PMCR_E); 5354ce2a804SEric Auger 5364ce2a804SEric Auger report_info("counter #0 is 0x%lx (CPU_CYCLES)", 5374ce2a804SEric Auger read_regn_el0(pmevcntr, 0)); 5384ce2a804SEric Auger report_info("counter #1 is 0x%lx (INST_RETIRED)", 5394ce2a804SEric Auger read_regn_el0(pmevcntr, 1)); 5404ce2a804SEric Auger 5414ce2a804SEric Auger report_info("overflow reg = 0x%lx", read_sysreg(pmovsclr_el0)); 5424ce2a804SEric Auger report(read_sysreg(pmovsclr_el0) & 0x1, 5434ce2a804SEric Auger "check overflow happened on #0 only"); 5444ce2a804SEric Auger } 5454ce2a804SEric Auger 546041df25bSRicardo Koller static void test_mem_access(bool overflow_at_64bits) 5474ce2a804SEric Auger { 5484ce2a804SEric Auger void *addr = malloc(PAGE_SIZE); 5494ce2a804SEric Auger uint32_t events[] = {MEM_ACCESS, MEM_ACCESS}; 550036369c5SRicardo Koller uint64_t pre_overflow = PRE_OVERFLOW(overflow_at_64bits); 551036369c5SRicardo Koller uint64_t pmcr_lp = overflow_at_64bits ? PMU_PMCR_LP : 0; 5524ce2a804SEric Auger 553041df25bSRicardo Koller if (!satisfy_prerequisites(events, ARRAY_SIZE(events)) || 554041df25bSRicardo Koller !check_overflow_prerequisites(overflow_at_64bits)) 5554ce2a804SEric Auger return; 5564ce2a804SEric Auger 5574ce2a804SEric Auger pmu_reset(); 5584ce2a804SEric Auger 5594ce2a804SEric Auger write_regn_el0(pmevtyper, 0, MEM_ACCESS | PMEVTYPER_EXCLUDE_EL0); 5604ce2a804SEric Auger write_regn_el0(pmevtyper, 1, MEM_ACCESS | PMEVTYPER_EXCLUDE_EL0); 5614ce2a804SEric Auger write_sysreg_s(0x3, PMCNTENSET_EL0); 5624ce2a804SEric Auger isb(); 563036369c5SRicardo Koller mem_access_loop(addr, 20, pmu.pmcr_ro | PMU_PMCR_E | pmcr_lp); 564a7509187SRicardo Koller report_info("counter #0 is 0x%lx (MEM_ACCESS)", read_regn_el0(pmevcntr, 0)); 565a7509187SRicardo Koller report_info("counter #1 is 0x%lx (MEM_ACCESS)", read_regn_el0(pmevcntr, 1)); 5664ce2a804SEric Auger /* We may measure more than 20 mem access depending on the core */ 5674ce2a804SEric Auger report((read_regn_el0(pmevcntr, 0) == read_regn_el0(pmevcntr, 1)) && 5684ce2a804SEric Auger (read_regn_el0(pmevcntr, 0) >= 20) && !read_sysreg(pmovsclr_el0), 5694ce2a804SEric Auger "Ran 20 mem accesses"); 5704ce2a804SEric Auger 5714ce2a804SEric Auger pmu_reset(); 5724ce2a804SEric Auger 573036369c5SRicardo Koller write_regn_el0(pmevcntr, 0, pre_overflow); 574036369c5SRicardo Koller write_regn_el0(pmevcntr, 1, pre_overflow); 5754ce2a804SEric Auger write_sysreg_s(0x3, PMCNTENSET_EL0); 5764ce2a804SEric Auger isb(); 577036369c5SRicardo Koller mem_access_loop(addr, 20, pmu.pmcr_ro | PMU_PMCR_E | pmcr_lp); 5784ce2a804SEric Auger report(read_sysreg(pmovsclr_el0) == 0x3, 5794ce2a804SEric Auger "Ran 20 mem accesses with expected overflows on both counters"); 580a7509187SRicardo Koller report_info("cnt#0=0x%lx cnt#1=0x%lx overflow=0x%lx", 5814ce2a804SEric Auger read_regn_el0(pmevcntr, 0), read_regn_el0(pmevcntr, 1), 5824ce2a804SEric Auger read_sysreg(pmovsclr_el0)); 5834ce2a804SEric Auger } 5844ce2a804SEric Auger 585041df25bSRicardo Koller static void test_sw_incr(bool overflow_at_64bits) 586bb9a5adcSEric Auger { 587036369c5SRicardo Koller uint64_t pre_overflow = PRE_OVERFLOW(overflow_at_64bits); 588036369c5SRicardo Koller uint64_t pmcr_lp = overflow_at_64bits ? PMU_PMCR_LP : 0; 589bb9a5adcSEric Auger uint32_t events[] = {SW_INCR, SW_INCR}; 590036369c5SRicardo Koller uint64_t cntr0 = (pre_overflow + 100) & pmevcntr_mask(); 591bb9a5adcSEric Auger int i; 592bb9a5adcSEric Auger 593041df25bSRicardo Koller if (!satisfy_prerequisites(events, ARRAY_SIZE(events)) || 594041df25bSRicardo Koller !check_overflow_prerequisites(overflow_at_64bits)) 595bb9a5adcSEric Auger return; 596bb9a5adcSEric Auger 597bb9a5adcSEric Auger pmu_reset(); 598bb9a5adcSEric Auger 599bb9a5adcSEric Auger write_regn_el0(pmevtyper, 0, SW_INCR | PMEVTYPER_EXCLUDE_EL0); 600bb9a5adcSEric Auger write_regn_el0(pmevtyper, 1, SW_INCR | PMEVTYPER_EXCLUDE_EL0); 601bb9a5adcSEric Auger /* enable counters #0 and #1 */ 602bb9a5adcSEric Auger write_sysreg_s(0x3, PMCNTENSET_EL0); 603bb9a5adcSEric Auger 604036369c5SRicardo Koller write_regn_el0(pmevcntr, 0, pre_overflow); 605e0a6e56bSRicardo Koller isb(); 606bb9a5adcSEric Auger 607bb9a5adcSEric Auger for (i = 0; i < 100; i++) 608bb9a5adcSEric Auger write_sysreg(0x1, pmswinc_el0); 609bb9a5adcSEric Auger 610e0a6e56bSRicardo Koller isb(); 611a7509187SRicardo Koller report_info("SW_INCR counter #0 has value 0x%lx", read_regn_el0(pmevcntr, 0)); 612036369c5SRicardo Koller report(read_regn_el0(pmevcntr, 0) == pre_overflow, 613bb9a5adcSEric Auger "PWSYNC does not increment if PMCR.E is unset"); 614bb9a5adcSEric Auger 615bb9a5adcSEric Auger pmu_reset(); 616bb9a5adcSEric Auger 617036369c5SRicardo Koller write_regn_el0(pmevcntr, 0, pre_overflow); 618bb9a5adcSEric Auger write_sysreg_s(0x3, PMCNTENSET_EL0); 619036369c5SRicardo Koller set_pmcr(pmu.pmcr_ro | PMU_PMCR_E | pmcr_lp); 620e0a6e56bSRicardo Koller isb(); 621bb9a5adcSEric Auger 622bb9a5adcSEric Auger for (i = 0; i < 100; i++) 623bb9a5adcSEric Auger write_sysreg(0x3, pmswinc_el0); 624bb9a5adcSEric Auger 625e0a6e56bSRicardo Koller isb(); 62639d1347aSRicardo Koller report(read_regn_el0(pmevcntr, 0) == cntr0, "counter #0 after + 100 SW_INCR"); 62739d1347aSRicardo Koller report(read_regn_el0(pmevcntr, 1) == 100, "counter #1 after + 100 SW_INCR"); 628a7509187SRicardo Koller report_info("counter values after 100 SW_INCR #0=0x%lx #1=0x%lx", 629bb9a5adcSEric Auger read_regn_el0(pmevcntr, 0), read_regn_el0(pmevcntr, 1)); 630bb9a5adcSEric Auger report(read_sysreg(pmovsclr_el0) == 0x1, 63166fee034SEric Auger "overflow on counter #0 after 100 SW_INCR"); 63266fee034SEric Auger } 63366fee034SEric Auger 634*de62a4b6SEric Auger static void enable_chain_counter(int even) 635*de62a4b6SEric Auger { 636*de62a4b6SEric Auger write_sysreg_s(BIT(even + 1), PMCNTENSET_EL0); /* Enable the high counter first */ 637*de62a4b6SEric Auger isb(); 638*de62a4b6SEric Auger write_sysreg_s(BIT(even), PMCNTENSET_EL0); /* Enable the low counter */ 639*de62a4b6SEric Auger isb(); 640*de62a4b6SEric Auger } 641*de62a4b6SEric Auger 642*de62a4b6SEric Auger static void disable_chain_counter(int even) 643*de62a4b6SEric Auger { 644*de62a4b6SEric Auger write_sysreg_s(BIT(even), PMCNTENCLR_EL0); /* Disable the low counter first*/ 645*de62a4b6SEric Auger isb(); 646*de62a4b6SEric Auger write_sysreg_s(BIT(even + 1), PMCNTENCLR_EL0); /* Disable the high counter */ 647*de62a4b6SEric Auger isb(); 648*de62a4b6SEric Auger } 649*de62a4b6SEric Auger 650041df25bSRicardo Koller static void test_chained_counters(bool unused) 65166fee034SEric Auger { 65266fee034SEric Auger uint32_t events[] = {CPU_CYCLES, CHAIN}; 653036369c5SRicardo Koller uint64_t all_set = pmevcntr_mask(); 65466fee034SEric Auger 65566fee034SEric Auger if (!satisfy_prerequisites(events, ARRAY_SIZE(events))) 65666fee034SEric Auger return; 65766fee034SEric Auger 65866fee034SEric Auger pmu_reset(); 65966fee034SEric Auger 66066fee034SEric Auger write_regn_el0(pmevtyper, 0, CPU_CYCLES | PMEVTYPER_EXCLUDE_EL0); 66166fee034SEric Auger write_regn_el0(pmevtyper, 1, CHAIN | PMEVTYPER_EXCLUDE_EL0); 6627d1f853aSRicardo Koller write_regn_el0(pmevcntr, 0, PRE_OVERFLOW_32); 663*de62a4b6SEric Auger enable_chain_counter(0); 66466fee034SEric Auger 66566fee034SEric Auger precise_instrs_loop(22, pmu.pmcr_ro | PMU_PMCR_E); 66666fee034SEric Auger 66766fee034SEric Auger report(read_regn_el0(pmevcntr, 1) == 1, "CHAIN counter #1 incremented"); 668b5489580SRicardo Koller report(read_sysreg(pmovsclr_el0) == 0x1, "overflow recorded for chained incr #1"); 66966fee034SEric Auger 67066fee034SEric Auger /* test 64b overflow */ 67166fee034SEric Auger 67266fee034SEric Auger pmu_reset(); 67366fee034SEric Auger 6747d1f853aSRicardo Koller write_regn_el0(pmevcntr, 0, PRE_OVERFLOW_32); 67566fee034SEric Auger write_regn_el0(pmevcntr, 1, 0x1); 676*de62a4b6SEric Auger enable_chain_counter(0); 67766fee034SEric Auger precise_instrs_loop(22, pmu.pmcr_ro | PMU_PMCR_E); 67866fee034SEric Auger report_info("overflow reg = 0x%lx", read_sysreg(pmovsclr_el0)); 67966fee034SEric Auger report(read_regn_el0(pmevcntr, 1) == 2, "CHAIN counter #1 set to 2"); 680b5489580SRicardo Koller report(read_sysreg(pmovsclr_el0) == 0x1, "overflow recorded for chained incr #2"); 68166fee034SEric Auger 6827d1f853aSRicardo Koller write_regn_el0(pmevcntr, 0, PRE_OVERFLOW_32); 683036369c5SRicardo Koller write_regn_el0(pmevcntr, 1, all_set); 68466fee034SEric Auger 68566fee034SEric Auger precise_instrs_loop(22, pmu.pmcr_ro | PMU_PMCR_E); 68666fee034SEric Auger report_info("overflow reg = 0x%lx", read_sysreg(pmovsclr_el0)); 687036369c5SRicardo Koller report(read_regn_el0(pmevcntr, 1) == 0, "CHAIN counter #1 wrapped"); 688b5489580SRicardo Koller report(read_sysreg(pmovsclr_el0) == 0x3, "overflow on even and odd counters"); 68966fee034SEric Auger } 69066fee034SEric Auger 691041df25bSRicardo Koller static void test_chained_sw_incr(bool unused) 69266fee034SEric Auger { 69366fee034SEric Auger uint32_t events[] = {SW_INCR, CHAIN}; 6947d1f853aSRicardo Koller uint64_t cntr0 = (PRE_OVERFLOW_32 + 100) & pmevcntr_mask(); 6957d1f853aSRicardo Koller uint64_t cntr1 = (ALL_SET_32 + 1) & pmevcntr_mask(); 69666fee034SEric Auger int i; 69766fee034SEric Auger 69866fee034SEric Auger if (!satisfy_prerequisites(events, ARRAY_SIZE(events))) 69966fee034SEric Auger return; 70066fee034SEric Auger 70166fee034SEric Auger pmu_reset(); 70266fee034SEric Auger 70366fee034SEric Auger write_regn_el0(pmevtyper, 0, SW_INCR | PMEVTYPER_EXCLUDE_EL0); 70466fee034SEric Auger write_regn_el0(pmevtyper, 1, CHAIN | PMEVTYPER_EXCLUDE_EL0); 705*de62a4b6SEric Auger enable_chain_counter(0); 70666fee034SEric Auger 7077d1f853aSRicardo Koller write_regn_el0(pmevcntr, 0, PRE_OVERFLOW_32); 70866fee034SEric Auger set_pmcr(pmu.pmcr_ro | PMU_PMCR_E); 709e0a6e56bSRicardo Koller isb(); 710e0a6e56bSRicardo Koller 71166fee034SEric Auger for (i = 0; i < 100; i++) 71266fee034SEric Auger write_sysreg(0x1, pmswinc_el0); 71366fee034SEric Auger 714e0a6e56bSRicardo Koller isb(); 715b5489580SRicardo Koller report((read_sysreg(pmovsclr_el0) == 0x1) && 716b5489580SRicardo Koller (read_regn_el0(pmevcntr, 1) == 1), 717b5489580SRicardo Koller "overflow and chain counter incremented after 100 SW_INCR/CHAIN"); 718a7509187SRicardo Koller report_info("overflow=0x%lx, #0=0x%lx #1=0x%lx", read_sysreg(pmovsclr_el0), 71966fee034SEric Auger read_regn_el0(pmevcntr, 0), read_regn_el0(pmevcntr, 1)); 72066fee034SEric Auger 72166fee034SEric Auger /* 64b SW_INCR and overflow on CHAIN counter*/ 72266fee034SEric Auger pmu_reset(); 72366fee034SEric Auger 7247d1f853aSRicardo Koller write_regn_el0(pmevcntr, 0, PRE_OVERFLOW_32); 7257d1f853aSRicardo Koller write_regn_el0(pmevcntr, 1, ALL_SET_32); 726*de62a4b6SEric Auger enable_chain_counter(0); 72766fee034SEric Auger set_pmcr(pmu.pmcr_ro | PMU_PMCR_E); 728e0a6e56bSRicardo Koller isb(); 729e0a6e56bSRicardo Koller 73066fee034SEric Auger for (i = 0; i < 100; i++) 73166fee034SEric Auger write_sysreg(0x1, pmswinc_el0); 73266fee034SEric Auger 733e0a6e56bSRicardo Koller isb(); 734b5489580SRicardo Koller report((read_sysreg(pmovsclr_el0) == 0x3) && 73539d1347aSRicardo Koller (read_regn_el0(pmevcntr, 0) == cntr0) && 73639d1347aSRicardo Koller (read_regn_el0(pmevcntr, 1) == cntr1), 737b5489580SRicardo Koller "expected overflows and values after 100 SW_INCR/CHAIN"); 738a7509187SRicardo Koller report_info("overflow=0x%lx, #0=0x%lx #1=0x%lx", read_sysreg(pmovsclr_el0), 73966fee034SEric Auger read_regn_el0(pmevcntr, 0), read_regn_el0(pmevcntr, 1)); 740bb9a5adcSEric Auger } 741e4c27538SEric Auger #define PRINT_REGS(__s) \ 742e4c27538SEric Auger report_info("%s #1=0x%lx #0=0x%lx overflow=0x%lx", __s, \ 743e4c27538SEric Auger read_regn_el0(pmevcntr, 1), \ 744e4c27538SEric Auger read_regn_el0(pmevcntr, 0), \ 745e4c27538SEric Auger read_sysreg(pmovsclr_el0)) 746bb9a5adcSEric Auger 747041df25bSRicardo Koller static void test_chain_promotion(bool unused) 748ca42f29aSEric Auger { 749ca42f29aSEric Auger uint32_t events[] = {MEM_ACCESS, CHAIN}; 750ca42f29aSEric Auger void *addr = malloc(PAGE_SIZE); 751ca42f29aSEric Auger 752ca42f29aSEric Auger if (!satisfy_prerequisites(events, ARRAY_SIZE(events))) 753ca42f29aSEric Auger return; 754ca42f29aSEric Auger 755ca42f29aSEric Auger /* Only enable CHAIN counter */ 756e4c27538SEric Auger report_prefix_push("subtest1"); 757ca42f29aSEric Auger pmu_reset(); 758ca42f29aSEric Auger write_regn_el0(pmevtyper, 0, MEM_ACCESS | PMEVTYPER_EXCLUDE_EL0); 759ca42f29aSEric Auger write_regn_el0(pmevtyper, 1, CHAIN | PMEVTYPER_EXCLUDE_EL0); 760ca42f29aSEric Auger write_sysreg_s(0x2, PMCNTENSET_EL0); 761ca42f29aSEric Auger isb(); 762ca42f29aSEric Auger 76391806724SEric Auger mem_access_loop(addr, COUNT, pmu.pmcr_ro | PMU_PMCR_E); 764e4c27538SEric Auger PRINT_REGS("post"); 765ca42f29aSEric Auger report(!read_regn_el0(pmevcntr, 0), 766ca42f29aSEric Auger "chain counter not counting if even counter is disabled"); 767e4c27538SEric Auger report_prefix_pop(); 768ca42f29aSEric Auger 769ca42f29aSEric Auger /* Only enable even counter */ 770e4c27538SEric Auger report_prefix_push("subtest2"); 771ca42f29aSEric Auger pmu_reset(); 7727d1f853aSRicardo Koller write_regn_el0(pmevcntr, 0, PRE_OVERFLOW_32); 773ca42f29aSEric Auger write_sysreg_s(0x1, PMCNTENSET_EL0); 774ca42f29aSEric Auger isb(); 775ca42f29aSEric Auger 77691806724SEric Auger mem_access_loop(addr, COUNT, pmu.pmcr_ro | PMU_PMCR_E); 777e4c27538SEric Auger PRINT_REGS("post"); 778ca42f29aSEric Auger report(!read_regn_el0(pmevcntr, 1) && (read_sysreg(pmovsclr_el0) == 0x1), 779ca42f29aSEric Auger "odd counter did not increment on overflow if disabled"); 780e4c27538SEric Auger report_prefix_pop(); 781ca42f29aSEric Auger 78291806724SEric Auger /* 1st COUNT with CHAIN enabled, next COUNT with CHAIN disabled */ 783e4c27538SEric Auger report_prefix_push("subtest3"); 784ca42f29aSEric Auger pmu_reset(); 7857d1f853aSRicardo Koller write_regn_el0(pmevcntr, 0, PRE_OVERFLOW2_32); 786*de62a4b6SEric Auger enable_chain_counter(0); 787e4c27538SEric Auger PRINT_REGS("init"); 788ca42f29aSEric Auger 78991806724SEric Auger mem_access_loop(addr, COUNT, pmu.pmcr_ro | PMU_PMCR_E); 790e4c27538SEric Auger PRINT_REGS("After 1st loop"); 791ca42f29aSEric Auger 792ca42f29aSEric Auger /* disable the CHAIN event */ 793*de62a4b6SEric Auger disable_chain_counter(0); 794*de62a4b6SEric Auger write_sysreg_s(0x1, PMCNTENSET_EL0); /* Enable the low counter */ 795*de62a4b6SEric Auger isb(); 79691806724SEric Auger mem_access_loop(addr, COUNT, pmu.pmcr_ro | PMU_PMCR_E); 797e4c27538SEric Auger PRINT_REGS("After 2nd loop"); 798ca42f29aSEric Auger report(read_sysreg(pmovsclr_el0) == 0x1, 799ca42f29aSEric Auger "should have triggered an overflow on #0"); 800ca42f29aSEric Auger report(!read_regn_el0(pmevcntr, 1), 801ca42f29aSEric Auger "CHAIN counter #1 shouldn't have incremented"); 802e4c27538SEric Auger report_prefix_pop(); 803ca42f29aSEric Auger 80491806724SEric Auger /* 1st COUNT with CHAIN disabled, next COUNT with CHAIN enabled */ 805ca42f29aSEric Auger 806e4c27538SEric Auger report_prefix_push("subtest4"); 807ca42f29aSEric Auger pmu_reset(); 808ca42f29aSEric Auger write_sysreg_s(0x1, PMCNTENSET_EL0); 8097d1f853aSRicardo Koller write_regn_el0(pmevcntr, 0, PRE_OVERFLOW2_32); 810ca42f29aSEric Auger isb(); 811e4c27538SEric Auger PRINT_REGS("init"); 812ca42f29aSEric Auger 81391806724SEric Auger mem_access_loop(addr, COUNT, pmu.pmcr_ro | PMU_PMCR_E); 814e4c27538SEric Auger PRINT_REGS("After 1st loop"); 815ca42f29aSEric Auger 816*de62a4b6SEric Auger /* Disable the low counter first and enable the chain counter */ 817*de62a4b6SEric Auger write_sysreg_s(0x1, PMCNTENCLR_EL0); 818ca42f29aSEric Auger isb(); 819*de62a4b6SEric Auger enable_chain_counter(0); 820*de62a4b6SEric Auger 82191806724SEric Auger mem_access_loop(addr, COUNT, pmu.pmcr_ro | PMU_PMCR_E); 822e4c27538SEric Auger 823e4c27538SEric Auger PRINT_REGS("After 2nd loop"); 824ca42f29aSEric Auger 825b5489580SRicardo Koller report((read_regn_el0(pmevcntr, 1) == 1) && 826b5489580SRicardo Koller (read_sysreg(pmovsclr_el0) == 0x1), 827b5489580SRicardo Koller "CHAIN counter enabled: CHAIN counter was incremented and overflow"); 828e4c27538SEric Auger report_prefix_pop(); 829ca42f29aSEric Auger 830ca42f29aSEric Auger /* start as MEM_ACCESS/CPU_CYCLES and move to CHAIN/MEM_ACCESS */ 831e4c27538SEric Auger report_prefix_push("subtest5"); 832ca42f29aSEric Auger pmu_reset(); 833ca42f29aSEric Auger write_regn_el0(pmevtyper, 0, MEM_ACCESS | PMEVTYPER_EXCLUDE_EL0); 834ca42f29aSEric Auger write_regn_el0(pmevtyper, 1, CPU_CYCLES | PMEVTYPER_EXCLUDE_EL0); 835ca42f29aSEric Auger write_sysreg_s(0x3, PMCNTENSET_EL0); 8367d1f853aSRicardo Koller write_regn_el0(pmevcntr, 0, PRE_OVERFLOW2_32); 837ca42f29aSEric Auger isb(); 838e4c27538SEric Auger PRINT_REGS("init"); 839ca42f29aSEric Auger 84091806724SEric Auger mem_access_loop(addr, COUNT, pmu.pmcr_ro | PMU_PMCR_E); 841e4c27538SEric Auger PRINT_REGS("After 1st loop"); 842ca42f29aSEric Auger 843ca42f29aSEric Auger /* 0 becomes CHAINED */ 844*de62a4b6SEric Auger write_sysreg_s(0x3, PMCNTENCLR_EL0); 845ca42f29aSEric Auger write_regn_el0(pmevtyper, 1, CHAIN | PMEVTYPER_EXCLUDE_EL0); 846ca42f29aSEric Auger write_regn_el0(pmevcntr, 1, 0x0); 847*de62a4b6SEric Auger enable_chain_counter(0); 848ca42f29aSEric Auger 84991806724SEric Auger mem_access_loop(addr, COUNT, pmu.pmcr_ro | PMU_PMCR_E); 850e4c27538SEric Auger PRINT_REGS("After 2nd loop"); 851ca42f29aSEric Auger 852b5489580SRicardo Koller report((read_regn_el0(pmevcntr, 1) == 1) && 853b5489580SRicardo Koller (read_sysreg(pmovsclr_el0) == 0x1), 854b5489580SRicardo Koller "32b->64b: CHAIN counter incremented and overflow"); 855e4c27538SEric Auger report_prefix_pop(); 856ca42f29aSEric Auger 857ca42f29aSEric Auger /* start as CHAIN/MEM_ACCESS and move to MEM_ACCESS/CPU_CYCLES */ 858e4c27538SEric Auger report_prefix_push("subtest6"); 859ca42f29aSEric Auger pmu_reset(); 860ca42f29aSEric Auger write_regn_el0(pmevtyper, 0, MEM_ACCESS | PMEVTYPER_EXCLUDE_EL0); 861ca42f29aSEric Auger write_regn_el0(pmevtyper, 1, CHAIN | PMEVTYPER_EXCLUDE_EL0); 8627d1f853aSRicardo Koller write_regn_el0(pmevcntr, 0, PRE_OVERFLOW2_32); 863*de62a4b6SEric Auger enable_chain_counter(0); 864e4c27538SEric Auger PRINT_REGS("init"); 865ca42f29aSEric Auger 86691806724SEric Auger mem_access_loop(addr, COUNT, pmu.pmcr_ro | PMU_PMCR_E); 867e4c27538SEric Auger PRINT_REGS("After 1st loop"); 868ca42f29aSEric Auger 869*de62a4b6SEric Auger disable_chain_counter(0); 870ca42f29aSEric Auger write_regn_el0(pmevtyper, 1, CPU_CYCLES | PMEVTYPER_EXCLUDE_EL0); 871ca42f29aSEric Auger write_sysreg_s(0x3, PMCNTENSET_EL0); 872ca42f29aSEric Auger 87391806724SEric Auger mem_access_loop(addr, COUNT, pmu.pmcr_ro | PMU_PMCR_E); 874e4c27538SEric Auger PRINT_REGS("After 2nd loop"); 875ca42f29aSEric Auger report(read_sysreg(pmovsclr_el0) == 1, 876ca42f29aSEric Auger "overflow is expected on counter 0"); 877e4c27538SEric Auger report_prefix_pop(); 878ca42f29aSEric Auger } 879ca42f29aSEric Auger 8804f5ef94fSEric Auger static bool expect_interrupts(uint32_t bitmap) 8814f5ef94fSEric Auger { 8824f5ef94fSEric Auger int i; 8834f5ef94fSEric Auger 8844f5ef94fSEric Auger if (pmu_stats.bitmap ^ bitmap || pmu_stats.unexpected) 8854f5ef94fSEric Auger return false; 8864f5ef94fSEric Auger 8874f5ef94fSEric Auger for (i = 0; i < 32; i++) { 8884f5ef94fSEric Auger if (test_and_clear_bit(i, &pmu_stats.bitmap)) 8894f5ef94fSEric Auger if (pmu_stats.interrupts[i] != 1) 8904f5ef94fSEric Auger return false; 8914f5ef94fSEric Auger } 8924f5ef94fSEric Auger return true; 8934f5ef94fSEric Auger } 8944f5ef94fSEric Auger 895041df25bSRicardo Koller static void test_overflow_interrupt(bool overflow_at_64bits) 8964f5ef94fSEric Auger { 897036369c5SRicardo Koller uint64_t pre_overflow = PRE_OVERFLOW(overflow_at_64bits); 898036369c5SRicardo Koller uint64_t all_set = pmevcntr_mask(); 899036369c5SRicardo Koller uint64_t pmcr_lp = overflow_at_64bits ? PMU_PMCR_LP : 0; 9004f5ef94fSEric Auger uint32_t events[] = {MEM_ACCESS, SW_INCR}; 9014f5ef94fSEric Auger void *addr = malloc(PAGE_SIZE); 9024f5ef94fSEric Auger int i; 9034f5ef94fSEric Auger 904041df25bSRicardo Koller if (!satisfy_prerequisites(events, ARRAY_SIZE(events)) || 905041df25bSRicardo Koller !check_overflow_prerequisites(overflow_at_64bits)) 9064f5ef94fSEric Auger return; 9074f5ef94fSEric Auger 9084f5ef94fSEric Auger gic_enable_defaults(); 9094f5ef94fSEric Auger install_irq_handler(EL1H_IRQ, irq_handler); 9104f5ef94fSEric Auger local_irq_enable(); 9114f5ef94fSEric Auger gic_enable_irq(23); 9124f5ef94fSEric Auger 9134f5ef94fSEric Auger pmu_reset(); 9144f5ef94fSEric Auger 9154f5ef94fSEric Auger write_regn_el0(pmevtyper, 0, MEM_ACCESS | PMEVTYPER_EXCLUDE_EL0); 9164f5ef94fSEric Auger write_regn_el0(pmevtyper, 1, SW_INCR | PMEVTYPER_EXCLUDE_EL0); 9174f5ef94fSEric Auger write_sysreg_s(0x3, PMCNTENSET_EL0); 918036369c5SRicardo Koller write_regn_el0(pmevcntr, 0, pre_overflow); 919036369c5SRicardo Koller write_regn_el0(pmevcntr, 1, pre_overflow); 9204f5ef94fSEric Auger isb(); 9214f5ef94fSEric Auger 9221a97dad8SRicardo Koller /* interrupts are disabled (PMINTENSET_EL1 == 0) */ 9234f5ef94fSEric Auger 924036369c5SRicardo Koller mem_access_loop(addr, 200, pmu.pmcr_ro | PMU_PMCR_E | pmcr_lp); 9254f5ef94fSEric Auger report(expect_interrupts(0), "no overflow interrupt after preset"); 9264f5ef94fSEric Auger 927036369c5SRicardo Koller set_pmcr(pmu.pmcr_ro | PMU_PMCR_E | pmcr_lp); 928e0a6e56bSRicardo Koller isb(); 929e0a6e56bSRicardo Koller 9304f5ef94fSEric Auger for (i = 0; i < 100; i++) 9314f5ef94fSEric Auger write_sysreg(0x2, pmswinc_el0); 9324f5ef94fSEric Auger 933e0a6e56bSRicardo Koller isb(); 9344f5ef94fSEric Auger set_pmcr(pmu.pmcr_ro); 935e0a6e56bSRicardo Koller isb(); 9364f5ef94fSEric Auger report(expect_interrupts(0), "no overflow interrupt after counting"); 9374f5ef94fSEric Auger 9387d1f853aSRicardo Koller /* enable interrupts (PMINTENSET_EL1 <= ALL_SET_32) */ 9394f5ef94fSEric Auger 9404f5ef94fSEric Auger pmu_reset_stats(); 9414f5ef94fSEric Auger 942036369c5SRicardo Koller write_regn_el0(pmevcntr, 0, pre_overflow); 943036369c5SRicardo Koller write_regn_el0(pmevcntr, 1, pre_overflow); 944cc08ef11SRicardo Koller write_sysreg(ALL_SET_32, pmovsclr_el0); 9457d1f853aSRicardo Koller write_sysreg(ALL_SET_32, pmintenset_el1); 9464f5ef94fSEric Auger isb(); 9474f5ef94fSEric Auger 948036369c5SRicardo Koller mem_access_loop(addr, 200, pmu.pmcr_ro | PMU_PMCR_E | pmcr_lp); 949cc08ef11SRicardo Koller 950cc08ef11SRicardo Koller set_pmcr(pmu.pmcr_ro | PMU_PMCR_E | pmcr_lp); 951cc08ef11SRicardo Koller isb(); 952cc08ef11SRicardo Koller 9534f5ef94fSEric Auger for (i = 0; i < 100; i++) 9544f5ef94fSEric Auger write_sysreg(0x3, pmswinc_el0); 9554f5ef94fSEric Auger 9564f5ef94fSEric Auger mem_access_loop(addr, 200, pmu.pmcr_ro); 9574f5ef94fSEric Auger report_info("overflow=0x%lx", read_sysreg(pmovsclr_el0)); 9584f5ef94fSEric Auger report(expect_interrupts(0x3), 9594f5ef94fSEric Auger "overflow interrupts expected on #0 and #1"); 9604f5ef94fSEric Auger 961036369c5SRicardo Koller /* 962036369c5SRicardo Koller * promote to 64-b: 963036369c5SRicardo Koller * 964036369c5SRicardo Koller * This only applies to the !overflow_at_64bits case, as 965036369c5SRicardo Koller * overflow_at_64bits doesn't implement CHAIN events. The 966036369c5SRicardo Koller * overflow_at_64bits case just checks that chained counters are 967036369c5SRicardo Koller * not incremented when PMCR.LP == 1. 968036369c5SRicardo Koller */ 9694f5ef94fSEric Auger 9704f5ef94fSEric Auger pmu_reset_stats(); 9714f5ef94fSEric Auger 9724f5ef94fSEric Auger write_regn_el0(pmevtyper, 1, CHAIN | PMEVTYPER_EXCLUDE_EL0); 973036369c5SRicardo Koller write_regn_el0(pmevcntr, 0, pre_overflow); 9744f5ef94fSEric Auger isb(); 975036369c5SRicardo Koller mem_access_loop(addr, 200, pmu.pmcr_ro | PMU_PMCR_E | pmcr_lp); 976036369c5SRicardo Koller report(expect_interrupts(0x1), "expect overflow interrupt"); 9774f5ef94fSEric Auger 9784f5ef94fSEric Auger /* overflow on odd counter */ 9794f5ef94fSEric Auger pmu_reset_stats(); 980036369c5SRicardo Koller write_regn_el0(pmevcntr, 0, pre_overflow); 981036369c5SRicardo Koller write_regn_el0(pmevcntr, 1, all_set); 9824f5ef94fSEric Auger isb(); 983036369c5SRicardo Koller mem_access_loop(addr, 400, pmu.pmcr_ro | PMU_PMCR_E | pmcr_lp); 984036369c5SRicardo Koller if (overflow_at_64bits) { 985036369c5SRicardo Koller report(expect_interrupts(0x1), 986036369c5SRicardo Koller "expect overflow interrupt on even counter"); 987036369c5SRicardo Koller report(read_regn_el0(pmevcntr, 1) == all_set, 988036369c5SRicardo Koller "Odd counter did not change"); 989036369c5SRicardo Koller } else { 990b5489580SRicardo Koller report(expect_interrupts(0x3), 991b5489580SRicardo Koller "expect overflow interrupt on even and odd counter"); 992036369c5SRicardo Koller report(read_regn_el0(pmevcntr, 1) != all_set, 993036369c5SRicardo Koller "Odd counter wrapped"); 994036369c5SRicardo Koller } 9954f5ef94fSEric Auger } 9964244065bSChristopher Covington #endif 9974244065bSChristopher Covington 9984244065bSChristopher Covington /* 999d81bb7a3SChristopher Covington * Ensure that the cycle counter progresses between back-to-back reads. 1000d81bb7a3SChristopher Covington */ 1001d81bb7a3SChristopher Covington static bool check_cycles_increase(void) 1002d81bb7a3SChristopher Covington { 1003d81bb7a3SChristopher Covington bool success = true; 1004d81bb7a3SChristopher Covington 1005d81bb7a3SChristopher Covington /* init before event access, this test only cares about cycle count */ 10061a97dad8SRicardo Koller pmu_reset(); 1007d81bb7a3SChristopher Covington set_pmcntenset(1 << PMU_CYCLE_IDX); 1008d81bb7a3SChristopher Covington set_pmccfiltr(0); /* count cycles in EL0, EL1, but not EL2 */ 1009d81bb7a3SChristopher Covington 1010d81bb7a3SChristopher Covington set_pmcr(get_pmcr() | PMU_PMCR_LC | PMU_PMCR_C | PMU_PMCR_E); 1011e0a6e56bSRicardo Koller isb(); 1012d81bb7a3SChristopher Covington 1013d81bb7a3SChristopher Covington for (int i = 0; i < NR_SAMPLES; i++) { 1014d81bb7a3SChristopher Covington uint64_t a, b; 1015d81bb7a3SChristopher Covington 1016d81bb7a3SChristopher Covington a = get_pmccntr(); 1017d81bb7a3SChristopher Covington b = get_pmccntr(); 1018d81bb7a3SChristopher Covington 1019d81bb7a3SChristopher Covington if (a >= b) { 1020d81bb7a3SChristopher Covington printf("Read %"PRId64" then %"PRId64".\n", a, b); 1021d81bb7a3SChristopher Covington success = false; 1022d81bb7a3SChristopher Covington break; 1023d81bb7a3SChristopher Covington } 1024d81bb7a3SChristopher Covington } 1025d81bb7a3SChristopher Covington 1026d81bb7a3SChristopher Covington set_pmcr(get_pmcr() & ~PMU_PMCR_E); 1027e0a6e56bSRicardo Koller isb(); 1028d81bb7a3SChristopher Covington 1029d81bb7a3SChristopher Covington return success; 1030d81bb7a3SChristopher Covington } 1031d81bb7a3SChristopher Covington 10328f76a347SChristopher Covington /* 10338f76a347SChristopher Covington * Execute a known number of guest instructions. Only even instruction counts 10348f76a347SChristopher Covington * greater than or equal to 4 are supported by the in-line assembly code. The 10358f76a347SChristopher Covington * control register (PMCR_EL0) is initialized with the provided value (allowing 10368f76a347SChristopher Covington * for example for the cycle counter or event counters to be reset). At the end 10378f76a347SChristopher Covington * of the exact instruction loop, zero is written to PMCR_EL0 to disable 10388f76a347SChristopher Covington * counting, allowing the cycle counter or event counters to be read at the 10398f76a347SChristopher Covington * leisure of the calling code. 10408f76a347SChristopher Covington */ 10418f76a347SChristopher Covington static void measure_instrs(int num, uint32_t pmcr) 10428f76a347SChristopher Covington { 10438f76a347SChristopher Covington int loop = (num - 2) / 2; 10448f76a347SChristopher Covington 10458f76a347SChristopher Covington assert(num >= 4 && ((num - 2) % 2 == 0)); 10468f76a347SChristopher Covington precise_instrs_loop(loop, pmcr); 10478f76a347SChristopher Covington } 10488f76a347SChristopher Covington 10498f76a347SChristopher Covington /* 10508f76a347SChristopher Covington * Measure cycle counts for various known instruction counts. Ensure that the 10518f76a347SChristopher Covington * cycle counter progresses (similar to check_cycles_increase() but with more 10528f76a347SChristopher Covington * instructions and using reset and stop controls). If supplied a positive, 10538f76a347SChristopher Covington * nonzero CPI parameter, it also strictly checks that every measurement matches 10548f76a347SChristopher Covington * it. Strict CPI checking is used to test -icount mode. 10558f76a347SChristopher Covington */ 10568f76a347SChristopher Covington static bool check_cpi(int cpi) 10578f76a347SChristopher Covington { 10588f76a347SChristopher Covington uint32_t pmcr = get_pmcr() | PMU_PMCR_LC | PMU_PMCR_C | PMU_PMCR_E; 10598f76a347SChristopher Covington 10608f76a347SChristopher Covington /* init before event access, this test only cares about cycle count */ 10611a97dad8SRicardo Koller pmu_reset(); 10628f76a347SChristopher Covington set_pmcntenset(1 << PMU_CYCLE_IDX); 10638f76a347SChristopher Covington set_pmccfiltr(0); /* count cycles in EL0, EL1, but not EL2 */ 10648f76a347SChristopher Covington 10658f76a347SChristopher Covington if (cpi > 0) 10668f76a347SChristopher Covington printf("Checking for CPI=%d.\n", cpi); 10678f76a347SChristopher Covington printf("instrs : cycles0 cycles1 ...\n"); 10688f76a347SChristopher Covington 10698f76a347SChristopher Covington for (unsigned int i = 4; i < 300; i += 32) { 10708f76a347SChristopher Covington uint64_t avg, sum = 0; 10718f76a347SChristopher Covington 10728f76a347SChristopher Covington printf("%4d:", i); 10738f76a347SChristopher Covington for (int j = 0; j < NR_SAMPLES; j++) { 10748f76a347SChristopher Covington uint64_t cycles; 10758f76a347SChristopher Covington 10768f76a347SChristopher Covington set_pmccntr(0); 10778f76a347SChristopher Covington measure_instrs(i, pmcr); 10788f76a347SChristopher Covington cycles = get_pmccntr(); 10798f76a347SChristopher Covington printf(" %4"PRId64"", cycles); 10808f76a347SChristopher Covington 10818f76a347SChristopher Covington if (!cycles) { 10828f76a347SChristopher Covington printf("\ncycles not incrementing!\n"); 10838f76a347SChristopher Covington return false; 10848f76a347SChristopher Covington } else if (cpi > 0 && cycles != i * cpi) { 10858f76a347SChristopher Covington printf("\nunexpected cycle count received!\n"); 10868f76a347SChristopher Covington return false; 10878f76a347SChristopher Covington } else if ((cycles >> 32) != 0) { 10888f76a347SChristopher Covington /* The cycles taken by the loop above should 10898f76a347SChristopher Covington * fit in 32 bits easily. We check the upper 10908f76a347SChristopher Covington * 32 bits of the cycle counter to make sure 10918f76a347SChristopher Covington * there is no supprise. */ 10928f76a347SChristopher Covington printf("\ncycle count bigger than 32bit!\n"); 10938f76a347SChristopher Covington return false; 10948f76a347SChristopher Covington } 10958f76a347SChristopher Covington 10968f76a347SChristopher Covington sum += cycles; 10978f76a347SChristopher Covington } 10988f76a347SChristopher Covington avg = sum / NR_SAMPLES; 10998f76a347SChristopher Covington printf(" avg=%-4"PRId64" %s=%-3"PRId64"\n", avg, 11008f76a347SChristopher Covington (avg >= i) ? "cpi" : "ipc", 11018f76a347SChristopher Covington (avg >= i) ? avg / i : i / avg); 11028f76a347SChristopher Covington } 11038f76a347SChristopher Covington 11048f76a347SChristopher Covington return true; 11058f76a347SChristopher Covington } 11068f76a347SChristopher Covington 11074c357610SAndrew Jones static void pmccntr64_test(void) 11084c357610SAndrew Jones { 11094c357610SAndrew Jones #ifdef __arm__ 1110784ee933SEric Auger if (pmu.version == ID_DFR0_PMU_V3) { 11114c357610SAndrew Jones if (ERRATA(9e3f7a296940)) { 11124c357610SAndrew Jones write_sysreg(0xdead, PMCCNTR64); 1113a299895bSThomas Huth report(read_sysreg(PMCCNTR64) == 0xdead, "pmccntr64"); 11144c357610SAndrew Jones } else 11154c357610SAndrew Jones report_skip("Skipping unsafe pmccntr64 test. Set ERRATA_9e3f7a296940=y to enable."); 11164c357610SAndrew Jones } 11174c357610SAndrew Jones #endif 11184c357610SAndrew Jones } 11194c357610SAndrew Jones 11204244065bSChristopher Covington /* Return FALSE if no PMU found, otherwise return TRUE */ 112123b8916bSThomas Huth static bool pmu_probe(void) 11224244065bSChristopher Covington { 11231e4f5392SAlexandru Elisei uint32_t pmcr; 112446ca10f4SAlexandru Elisei uint8_t implementer; 1125eff8f161SEric Auger 11268f747a85SEric Auger pmu.version = get_pmu_version(); 1127784ee933SEric Auger if (pmu.version == ID_DFR0_PMU_NOTIMPL || pmu.version == ID_DFR0_PMU_IMPDEF) 1128eff8f161SEric Auger return false; 1129eff8f161SEric Auger 1130784ee933SEric Auger report_info("PMU version: 0x%x", pmu.version); 1131eff8f161SEric Auger 11321e4f5392SAlexandru Elisei pmcr = get_pmcr(); 113346ca10f4SAlexandru Elisei implementer = (pmcr >> PMU_PMCR_IMP_SHIFT) & PMU_PMCR_IMP_MASK; 113446ca10f4SAlexandru Elisei report_info("PMU implementer/ID code: %#"PRIx32"(\"%c\")/%#"PRIx32, 1135eff8f161SEric Auger (pmcr >> PMU_PMCR_IMP_SHIFT) & PMU_PMCR_IMP_MASK, 113646ca10f4SAlexandru Elisei implementer ? implementer : ' ', 11378f747a85SEric Auger (pmcr >> PMU_PMCR_ID_SHIFT) & PMU_PMCR_ID_MASK); 11388f747a85SEric Auger 11398f747a85SEric Auger /* store read-only and RES0 fields of the PMCR bottom-half*/ 11408f747a85SEric Auger pmu.pmcr_ro = pmcr & 0xFFFFFF00; 11418f747a85SEric Auger pmu.nb_implemented_counters = 11428f747a85SEric Auger (pmcr >> PMU_PMCR_N_SHIFT) & PMU_PMCR_N_MASK; 11438f747a85SEric Auger report_info("Implements %d event counters", 11448f747a85SEric Auger pmu.nb_implemented_counters); 1145eff8f161SEric Auger 1146eff8f161SEric Auger return true; 11474244065bSChristopher Covington } 11484244065bSChristopher Covington 1149041df25bSRicardo Koller static void run_test(const char *name, const char *prefix, 1150041df25bSRicardo Koller void (*test)(bool), void *arg) 1151041df25bSRicardo Koller { 1152041df25bSRicardo Koller report_prefix_push(name); 1153041df25bSRicardo Koller report_prefix_push(prefix); 1154041df25bSRicardo Koller 1155041df25bSRicardo Koller test(arg); 1156041df25bSRicardo Koller 1157041df25bSRicardo Koller report_prefix_pop(); 1158041df25bSRicardo Koller report_prefix_pop(); 1159041df25bSRicardo Koller } 1160041df25bSRicardo Koller 1161041df25bSRicardo Koller static void run_event_test(char *name, void (*test)(bool), 1162041df25bSRicardo Koller bool overflow_at_64bits) 1163041df25bSRicardo Koller { 1164041df25bSRicardo Koller const char *prefix = overflow_at_64bits ? "64-bit overflows" 1165041df25bSRicardo Koller : "32-bit overflows"; 1166041df25bSRicardo Koller 1167041df25bSRicardo Koller run_test(name, prefix, test, (void *)overflow_at_64bits); 1168041df25bSRicardo Koller } 1169041df25bSRicardo Koller 11708f76a347SChristopher Covington int main(int argc, char *argv[]) 11714244065bSChristopher Covington { 11728f76a347SChristopher Covington int cpi = 0; 11738f76a347SChristopher Covington 11744244065bSChristopher Covington if (!pmu_probe()) { 11754244065bSChristopher Covington printf("No PMU found, test skipped...\n"); 11764244065bSChristopher Covington return report_summary(); 11774244065bSChristopher Covington } 11784244065bSChristopher Covington 117957ec1086SEric Auger if (argc < 2) 118057ec1086SEric Auger report_abort("no test specified"); 118157ec1086SEric Auger 11824244065bSChristopher Covington report_prefix_push("pmu"); 11834244065bSChristopher Covington 118457ec1086SEric Auger if (strcmp(argv[1], "cycle-counter") == 0) { 118557ec1086SEric Auger report_prefix_push(argv[1]); 118657ec1086SEric Auger if (argc > 2) 118757ec1086SEric Auger cpi = atol(argv[2]); 1188a299895bSThomas Huth report(check_cycles_increase(), 1189a299895bSThomas Huth "Monotonically increasing cycle count"); 1190a299895bSThomas Huth report(check_cpi(cpi), "Cycle/instruction ratio"); 11914c357610SAndrew Jones pmccntr64_test(); 119257ec1086SEric Auger report_prefix_pop(); 11934870738cSEric Auger } else if (strcmp(argv[1], "pmu-event-introspection") == 0) { 11944870738cSEric Auger report_prefix_push(argv[1]); 11954870738cSEric Auger test_event_introspection(); 11964870738cSEric Auger report_prefix_pop(); 11974ce2a804SEric Auger } else if (strcmp(argv[1], "pmu-event-counter-config") == 0) { 11984ce2a804SEric Auger report_prefix_push(argv[1]); 11994ce2a804SEric Auger test_event_counter_config(); 12004ce2a804SEric Auger report_prefix_pop(); 12014ce2a804SEric Auger } else if (strcmp(argv[1], "pmu-basic-event-count") == 0) { 1202041df25bSRicardo Koller run_event_test(argv[1], test_basic_event_count, false); 1203036369c5SRicardo Koller run_event_test(argv[1], test_basic_event_count, true); 12044ce2a804SEric Auger } else if (strcmp(argv[1], "pmu-mem-access") == 0) { 1205041df25bSRicardo Koller run_event_test(argv[1], test_mem_access, false); 1206036369c5SRicardo Koller run_event_test(argv[1], test_mem_access, true); 1207bb9a5adcSEric Auger } else if (strcmp(argv[1], "pmu-sw-incr") == 0) { 1208041df25bSRicardo Koller run_event_test(argv[1], test_sw_incr, false); 1209036369c5SRicardo Koller run_event_test(argv[1], test_sw_incr, true); 121066fee034SEric Auger } else if (strcmp(argv[1], "pmu-chained-counters") == 0) { 1211041df25bSRicardo Koller run_event_test(argv[1], test_chained_counters, false); 121266fee034SEric Auger } else if (strcmp(argv[1], "pmu-chained-sw-incr") == 0) { 1213041df25bSRicardo Koller run_event_test(argv[1], test_chained_sw_incr, false); 1214ca42f29aSEric Auger } else if (strcmp(argv[1], "pmu-chain-promotion") == 0) { 1215041df25bSRicardo Koller run_event_test(argv[1], test_chain_promotion, false); 12164f5ef94fSEric Auger } else if (strcmp(argv[1], "pmu-overflow-interrupt") == 0) { 1217041df25bSRicardo Koller run_event_test(argv[1], test_overflow_interrupt, false); 1218036369c5SRicardo Koller run_event_test(argv[1], test_overflow_interrupt, true); 121957ec1086SEric Auger } else { 122057ec1086SEric Auger report_abort("Unknown sub-test '%s'", argv[1]); 122157ec1086SEric Auger } 12224c357610SAndrew Jones 12234244065bSChristopher Covington return report_summary(); 12244244065bSChristopher Covington } 1225