/* * Copyright (c) 2011, Code Aurora Forum. All rights reserved. * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License version 2 and * only version 2 as published by the Free Software Foundation. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. */ #ifdef CONFIG_CPU_HAS_L2_PMU #include #define MAX_BB_L2_PERIOD ((1ULL << 32) - 1) #define MAX_BB_L2_CTRS 5 #define BB_L2CYCLE_CTR_BIT 31 #define BB_L2CYCLE_CTR_EVENT_IDX 4 #define BB_L2CYCLE_CTR_RAW_CODE 0xff #define SCORPIONL2_PMNC_E (1 << 0) /* Enable all counters */ /* * Lock to protect r/m/w sequences to the L2 PMU. */ DEFINE_RAW_SPINLOCK(bb_l2_pmu_lock); static struct platform_device *bb_l2_pmu_device; struct hw_bb_l2_pmu { struct perf_event *events[MAX_BB_L2_CTRS]; unsigned long active_mask[BITS_TO_LONGS(MAX_BB_L2_CTRS)]; raw_spinlock_t lock; }; struct hw_bb_l2_pmu hw_bb_l2_pmu; struct bb_l2_scorp_evt { u32 evt_type; u32 val; u8 grp; u32 evt_type_act; }; enum scorpion_perf_types { SCORPIONL2_TOTAL_BANK_REQ = 0x90, SCORPIONL2_DSIDE_READ = 0x91, SCORPIONL2_DSIDE_WRITE = 0x92, SCORPIONL2_ISIDE_READ = 0x93, SCORPIONL2_L2CACHE_ISIDE_READ = 0x94, SCORPIONL2_L2CACHE_BANK_REQ = 0x95, SCORPIONL2_L2CACHE_DSIDE_READ = 0x96, SCORPIONL2_L2CACHE_DSIDE_WRITE = 0x97, SCORPIONL2_L2NOCACHE_DSIDE_WRITE = 0x98, SCORPIONL2_L2NOCACHE_ISIDE_READ = 0x99, SCORPIONL2_L2NOCACHE_TOTAL_REQ = 0x9a, SCORPIONL2_L2NOCACHE_DSIDE_READ = 0x9b, SCORPIONL2_DSIDE_READ_NOL1 = 0x9c, SCORPIONL2_L2CACHE_WRITETHROUGH = 0x9d, SCORPIONL2_BARRIERS = 0x9e, SCORPIONL2_HARDWARE_TABLE_WALKS = 0x9f, SCORPIONL2_MVA_POC = 0xa0, SCORPIONL2_L2CACHE_HW_TABLE_WALKS = 0xa1, SCORPIONL2_SETWAY_CACHE_OPS = 0xa2, SCORPIONL2_DSIDE_WRITE_HITS = 0xa3, SCORPIONL2_ISIDE_READ_HITS = 0xa4, SCORPIONL2_CACHE_DSIDE_READ_NOL1 = 0xa5, SCORPIONL2_TOTAL_CACHE_HITS = 0xa6, SCORPIONL2_CACHE_MATCH_MISS = 0xa7, SCORPIONL2_DREAD_HIT_L1_DATA = 0xa8, SCORPIONL2_L2LINE_LOCKED = 0xa9, SCORPIONL2_HW_TABLE_WALK_HIT = 0xaa, SCORPIONL2_CACHE_MVA_POC = 0xab, SCORPIONL2_L2ALLOC_DWRITE_MISS = 0xac, SCORPIONL2_CORRECTED_TAG_ARRAY = 0xad, SCORPIONL2_CORRECTED_DATA_ARRAY = 0xae, SCORPIONL2_CORRECTED_REPLACEMENT_ARRAY = 0xaf, SCORPIONL2_PMBUS_MPAAF = 0xb0, SCORPIONL2_PMBUS_MPWDAF = 0xb1, SCORPIONL2_PMBUS_MPBRT = 0xb2, SCORPIONL2_CPU0_GRANT = 0xb3, SCORPIONL2_CPU1_GRANT = 0xb4, SCORPIONL2_CPU0_NOGRANT = 0xb5, SCORPIONL2_CPU1_NOGRANT = 0xb6, SCORPIONL2_CPU0_LOSING_ARB = 0xb7, SCORPIONL2_CPU1_LOSING_ARB = 0xb8, SCORPIONL2_SLAVEPORT_NOGRANT = 0xb9, SCORPIONL2_SLAVEPORT_BPQ_FULL = 0xba, SCORPIONL2_SLAVEPORT_LOSING_ARB = 0xbb, SCORPIONL2_SLAVEPORT_GRANT = 0xbc, SCORPIONL2_SLAVEPORT_GRANTLOCK = 0xbd, SCORPIONL2_L2EM_STREX_PASS = 0xbe, SCORPIONL2_L2EM_STREX_FAIL = 0xbf, SCORPIONL2_LDREX_RESERVE_L2EM = 0xc0, SCORPIONL2_SLAVEPORT_LDREX = 0xc1, SCORPIONL2_CPU0_L2EM_CLEARED = 0xc2, SCORPIONL2_CPU1_L2EM_CLEARED = 0xc3, SCORPIONL2_SLAVEPORT_L2EM_CLEARED = 0xc4, SCORPIONL2_CPU0_CLAMPED = 0xc5, SCORPIONL2_CPU1_CLAMPED = 0xc6, SCORPIONL2_CPU0_WAIT = 0xc7, SCORPIONL2_CPU1_WAIT = 0xc8, SCORPIONL2_CPU0_NONAMBAS_WAIT = 0xc9, SCORPIONL2_CPU1_NONAMBAS_WAIT = 0xca, SCORPIONL2_CPU0_DSB_WAIT = 0xcb, SCORPIONL2_CPU1_DSB_WAIT = 0xcc, SCORPIONL2_AXI_READ = 0xcd, SCORPIONL2_AXI_WRITE = 0xce, SCORPIONL2_1BEAT_WRITE = 0xcf, SCORPIONL2_2BEAT_WRITE = 0xd0, SCORPIONL2_4BEAT_WRITE = 0xd1, SCORPIONL2_8BEAT_WRITE = 0xd2, SCORPIONL2_12BEAT_WRITE = 0xd3, SCORPIONL2_16BEAT_WRITE = 0xd4, SCORPIONL2_1BEAT_DSIDE_READ = 0xd5, SCORPIONL2_2BEAT_DSIDE_READ = 0xd6, SCORPIONL2_4BEAT_DSIDE_READ = 0xd7, SCORPIONL2_8BEAT_DSIDE_READ = 0xd8, SCORPIONL2_CSYS_READ_1BEAT = 0xd9, SCORPIONL2_CSYS_READ_2BEAT = 0xda, SCORPIONL2_CSYS_READ_4BEAT = 0xdb, SCORPIONL2_CSYS_READ_8BEAT = 0xdc, SCORPIONL2_4BEAT_IFETCH_READ = 0xdd, SCORPIONL2_8BEAT_IFETCH_READ = 0xde, SCORPIONL2_CSYS_WRITE_1BEAT = 0xdf, SCORPIONL2_CSYS_WRITE_2BEAT = 0xe0, SCORPIONL2_AXI_READ_DATA_BEAT = 0xe1, SCORPIONL2_AXI_WRITE_EVT1 = 0xe2, SCORPIONL2_AXI_WRITE_EVT2 = 0xe3, SCORPIONL2_LDREX_REQ = 0xe4, SCORPIONL2_STREX_PASS = 0xe5, SCORPIONL2_STREX_FAIL = 0xe6, SCORPIONL2_CPREAD = 0xe7, SCORPIONL2_CPWRITE = 0xe8, SCORPIONL2_BARRIER_REQ = 0xe9, SCORPIONL2_AXI_READ_SLVPORT = 0xea, SCORPIONL2_AXI_WRITE_SLVPORT = 0xeb, SCORPIONL2_AXI_READ_SLVPORT_DATABEAT = 0xec, SCORPIONL2_AXI_WRITE_SLVPORT_DATABEAT = 0xed, SCORPIONL2_SNOOPKILL_PREFILTER = 0xee, SCORPIONL2_SNOOPKILL_FILTEROUT = 0xef, SCORPIONL2_SNOOPED_IC = 0xf0, SCORPIONL2_SNOOPED_BP = 0xf1, SCORPIONL2_SNOOPED_BARRIERS = 0xf2, SCORPIONL2_SNOOPED_TLB = 0xf3, BB_L2_MAX_EVT, }; static const struct bb_l2_scorp_evt sc_evt[] = { {SCORPIONL2_TOTAL_BANK_REQ, 0x80000001, 0, 0x00}, {SCORPIONL2_DSIDE_READ, 0x80000100, 0, 0x01}, {SCORPIONL2_DSIDE_WRITE, 0x80010000, 0, 0x02}, {SCORPIONL2_ISIDE_READ, 0x81000000, 0, 0x03}, {SCORPIONL2_L2CACHE_ISIDE_READ, 0x80000002, 0, 0x00}, {SCORPIONL2_L2CACHE_BANK_REQ, 0x80000200, 0, 0x01}, {SCORPIONL2_L2CACHE_DSIDE_READ, 0x80020000, 0, 0x02}, {SCORPIONL2_L2CACHE_DSIDE_WRITE, 0x82000000, 0, 0x03}, {SCORPIONL2_L2NOCACHE_DSIDE_WRITE, 0x80000003, 0, 0x00}, {SCORPIONL2_L2NOCACHE_ISIDE_READ, 0x80000300, 0, 0x01}, {SCORPIONL2_L2NOCACHE_TOTAL_REQ, 0x80030000, 0, 0x02}, {SCORPIONL2_L2NOCACHE_DSIDE_READ, 0x83000000, 0, 0x03}, {SCORPIONL2_DSIDE_READ_NOL1, 0x80000004, 0, 0x00}, {SCORPIONL2_L2CACHE_WRITETHROUGH, 0x80000400, 0, 0x01}, {SCORPIONL2_BARRIERS, 0x84000000, 0, 0x03}, {SCORPIONL2_HARDWARE_TABLE_WALKS, 0x80000005, 0, 0x00}, {SCORPIONL2_MVA_POC, 0x80000500, 0, 0x01}, {SCORPIONL2_L2CACHE_HW_TABLE_WALKS, 0x80050000, 0, 0x02}, {SCORPIONL2_SETWAY_CACHE_OPS, 0x85000000, 0, 0x03}, {SCORPIONL2_DSIDE_WRITE_HITS, 0x80000006, 0, 0x00}, {SCORPIONL2_ISIDE_READ_HITS, 0x80000600, 0, 0x01}, {SCORPIONL2_CACHE_DSIDE_READ_NOL1, 0x80060000, 0, 0x02}, {SCORPIONL2_TOTAL_CACHE_HITS, 0x86000000, 0, 0x03}, {SCORPIONL2_CACHE_MATCH_MISS, 0x80000007, 0, 0x00}, {SCORPIONL2_DREAD_HIT_L1_DATA, 0x87000000, 0, 0x03}, {SCORPIONL2_L2LINE_LOCKED, 0x80000008, 0, 0x00}, {SCORPIONL2_HW_TABLE_WALK_HIT, 0x80000800, 0, 0x01}, {SCORPIONL2_CACHE_MVA_POC, 0x80080000, 0, 0x02}, {SCORPIONL2_L2ALLOC_DWRITE_MISS, 0x88000000, 0, 0x03}, {SCORPIONL2_CORRECTED_TAG_ARRAY, 0x80001A00, 0, 0x01}, {SCORPIONL2_CORRECTED_DATA_ARRAY, 0x801A0000, 0, 0x02}, {SCORPIONL2_CORRECTED_REPLACEMENT_ARRAY, 0x9A000000, 0, 0x03}, {SCORPIONL2_PMBUS_MPAAF, 0x80001C00, 0, 0x01}, {SCORPIONL2_PMBUS_MPWDAF, 0x801C0000, 0, 0x02}, {SCORPIONL2_PMBUS_MPBRT, 0x9C000000, 0, 0x03}, {SCORPIONL2_CPU0_GRANT, 0x80000001, 1, 0x04}, {SCORPIONL2_CPU1_GRANT, 0x80000100, 1, 0x05}, {SCORPIONL2_CPU0_NOGRANT, 0x80020000, 1, 0x06}, {SCORPIONL2_CPU1_NOGRANT, 0x82000000, 1, 0x07}, {SCORPIONL2_CPU0_LOSING_ARB, 0x80040000, 1, 0x06}, {SCORPIONL2_CPU1_LOSING_ARB, 0x84000000, 1, 0x07}, {SCORPIONL2_SLAVEPORT_NOGRANT, 0x80000007, 1, 0x04}, {SCORPIONL2_SLAVEPORT_BPQ_FULL, 0x80000700, 1, 0x05}, {SCORPIONL2_SLAVEPORT_LOSING_ARB, 0x80070000, 1, 0x06}, {SCORPIONL2_SLAVEPORT_GRANT, 0x87000000, 1, 0x07}, {SCORPIONL2_SLAVEPORT_GRANTLOCK, 0x80000008, 1, 0x04}, {SCORPIONL2_L2EM_STREX_PASS, 0x80000009, 1, 0x04}, {SCORPIONL2_L2EM_STREX_FAIL, 0x80000900, 1, 0x05}, {SCORPIONL2_LDREX_RESERVE_L2EM, 0x80090000, 1, 0x06}, {SCORPIONL2_SLAVEPORT_LDREX, 0x89000000, 1, 0x07}, {SCORPIONL2_CPU0_L2EM_CLEARED, 0x800A0000, 1, 0x06}, {SCORPIONL2_CPU1_L2EM_CLEARED, 0x8A000000, 1, 0x07}, {SCORPIONL2_SLAVEPORT_L2EM_CLEARED, 0x80000B00, 1, 0x05}, {SCORPIONL2_CPU0_CLAMPED, 0x8000000E, 1, 0x04}, {SCORPIONL2_CPU1_CLAMPED, 0x80000E00, 1, 0x05}, {SCORPIONL2_CPU0_WAIT, 0x800F0000, 1, 0x06}, {SCORPIONL2_CPU1_WAIT, 0x8F000000, 1, 0x07}, {SCORPIONL2_CPU0_NONAMBAS_WAIT, 0x80000010, 1, 0x04}, {SCORPIONL2_CPU1_NONAMBAS_WAIT, 0x80001000, 1, 0x05}, {SCORPIONL2_CPU0_DSB_WAIT, 0x80000014, 1, 0x04}, {SCORPIONL2_CPU1_DSB_WAIT, 0x80001400, 1, 0x05}, {SCORPIONL2_AXI_READ, 0x80000001, 2, 0x08}, {SCORPIONL2_AXI_WRITE, 0x80000100, 2, 0x09}, {SCORPIONL2_1BEAT_WRITE, 0x80010000, 2, 0x0a}, {SCORPIONL2_2BEAT_WRITE, 0x80010000, 2, 0x0b}, {SCORPIONL2_4BEAT_WRITE, 0x80000002, 2, 0x08}, {SCORPIONL2_8BEAT_WRITE, 0x80000200, 2, 0x09}, {SCORPIONL2_12BEAT_WRITE, 0x80020000, 2, 0x0a}, {SCORPIONL2_16BEAT_WRITE, 0x82000000, 2, 0x0b}, {SCORPIONL2_1BEAT_DSIDE_READ, 0x80000003, 2, 0x08}, {SCORPIONL2_2BEAT_DSIDE_READ, 0x80000300, 2, 0x09}, {SCORPIONL2_4BEAT_DSIDE_READ, 0x80030000, 2, 0x0a}, {SCORPIONL2_8BEAT_DSIDE_READ, 0x83000000, 2, 0x0b}, {SCORPIONL2_CSYS_READ_1BEAT, 0x80000004, 2, 0x08}, {SCORPIONL2_CSYS_READ_2BEAT, 0x80000400, 2, 0x09}, {SCORPIONL2_CSYS_READ_4BEAT, 0x80040000, 2, 0x0a}, {SCORPIONL2_CSYS_READ_8BEAT, 0x84000000, 2, 0x0b}, {SCORPIONL2_4BEAT_IFETCH_READ, 0x80000005, 2, 0x08}, {SCORPIONL2_8BEAT_IFETCH_READ, 0x80000500, 2, 0x09}, {SCORPIONL2_CSYS_WRITE_1BEAT, 0x80050000, 2, 0x0a}, {SCORPIONL2_CSYS_WRITE_2BEAT, 0x85000000, 2, 0x0b}, {SCORPIONL2_AXI_READ_DATA_BEAT, 0x80000600, 2, 0x09}, {SCORPIONL2_AXI_WRITE_EVT1, 0x80060000, 2, 0x0a}, {SCORPIONL2_AXI_WRITE_EVT2, 0x86000000, 2, 0x0b}, {SCORPIONL2_LDREX_REQ, 0x80000007, 2, 0x08}, {SCORPIONL2_STREX_PASS, 0x80000700, 2, 0x09}, {SCORPIONL2_STREX_FAIL, 0x80070000, 2, 0x0a}, {SCORPIONL2_CPREAD, 0x80000008, 2, 0x08}, {SCORPIONL2_CPWRITE, 0x80000800, 2, 0x09}, {SCORPIONL2_BARRIER_REQ, 0x88000000, 2, 0x0b}, {SCORPIONL2_AXI_READ_SLVPORT, 0x80000001, 3, 0x0c}, {SCORPIONL2_AXI_WRITE_SLVPORT, 0x80000100, 3, 0x0d}, {SCORPIONL2_AXI_READ_SLVPORT_DATABEAT, 0x80010000, 3, 0x0e}, {SCORPIONL2_AXI_WRITE_SLVPORT_DATABEAT, 0x81000000, 3, 0x0f}, {SCORPIONL2_SNOOPKILL_PREFILTER, 0x80000001, 4, 0x10}, {SCORPIONL2_SNOOPKILL_FILTEROUT, 0x80000100, 4, 0x11}, {SCORPIONL2_SNOOPED_IC, 0x80000002, 4, 0x10}, {SCORPIONL2_SNOOPED_BP, 0x80000200, 4, 0x11}, {SCORPIONL2_SNOOPED_BARRIERS, 0x80020000, 4, 0x12}, {SCORPIONL2_SNOOPED_TLB, 0x82000000, 4, 0x13}, }; static u32 bb_l2_read_l2pm0(void) { u32 val; asm volatile ("mrc p15, 3, %0, c15, c7, 0" : "=r" (val)); return val; } static void bb_l2_write_l2pm0(u32 val) { asm volatile ("mcr p15, 3, %0, c15, c7, 0" : : "r" (val)); } static u32 bb_l2_read_l2pm1(void) { u32 val; asm volatile ("mrc p15, 3, %0, c15, c7, 1" : "=r" (val)); return val; } static void bb_l2_write_l2pm1(u32 val) { asm volatile ("mcr p15, 3, %0, c15, c7, 1" : : "r" (val)); } static u32 bb_l2_read_l2pm2(void) { u32 val; asm volatile ("mrc p15, 3, %0, c15, c7, 2" : "=r" (val)); return val; } static void bb_l2_write_l2pm2(u32 val) { asm volatile ("mcr p15, 3, %0, c15, c7, 2" : : "r" (val)); } static u32 bb_l2_read_l2pm3(void) { u32 val; asm volatile ("mrc p15, 3, %0, c15, c7, 3" : "=r" (val)); return val; } static void bb_l2_write_l2pm3(u32 val) { asm volatile ("mcr p15, 3, %0, c15, c7, 3" : : "r" (val)); } static u32 bb_l2_read_l2pm4(void) { u32 val; asm volatile ("mrc p15, 3, %0, c15, c7, 4" : "=r" (val)); return val; } static void bb_l2_write_l2pm4(u32 val) { asm volatile ("mcr p15, 3, %0, c15, c7, 4" : : "r" (val)); } struct bb_scorpion_access_funcs { u32(*read) (void); void (*write) (u32); void (*pre) (void); void (*post) (void); }; struct bb_scorpion_access_funcs bb_l2_func[] = { {bb_l2_read_l2pm0, bb_l2_write_l2pm0, NULL, NULL}, {bb_l2_read_l2pm1, bb_l2_write_l2pm1, NULL, NULL}, {bb_l2_read_l2pm2, bb_l2_write_l2pm2, NULL, NULL}, {bb_l2_read_l2pm3, bb_l2_write_l2pm3, NULL, NULL}, {bb_l2_read_l2pm4, bb_l2_write_l2pm4, NULL, NULL}, }; #define COLMN0MASK 0x000000ff #define COLMN1MASK 0x0000ff00 #define COLMN2MASK 0x00ff0000 static u32 bb_l2_get_columnmask(u32 setval) { if (setval & COLMN0MASK) return 0xffffff00; else if (setval & COLMN1MASK) return 0xffff00ff; else if (setval & COLMN2MASK) return 0xff00ffff; else return 0x80ffffff; } static void bb_l2_evt_setup(u32 gr, u32 setval) { u32 val; if (bb_l2_func[gr].pre) bb_l2_func[gr].pre(); val = bb_l2_get_columnmask(setval) & bb_l2_func[gr].read(); val = val | setval; bb_l2_func[gr].write(val); if (bb_l2_func[gr].post) bb_l2_func[gr].post(); } #define BB_L2_EVT_START_IDX 0x90 #define BB_L2_INV_EVTYPE 0 static unsigned int get_bb_l2_evtinfo(unsigned int evt_type, struct bb_l2_scorp_evt *evtinfo) { u32 idx; if (evt_type < BB_L2_EVT_START_IDX || evt_type >= BB_L2_MAX_EVT) return BB_L2_INV_EVTYPE; idx = evt_type - BB_L2_EVT_START_IDX; if (sc_evt[idx].evt_type == evt_type) { evtinfo->val = sc_evt[idx].val; evtinfo->grp = sc_evt[idx].grp; evtinfo->evt_type_act = sc_evt[idx].evt_type_act; return sc_evt[idx].evt_type_act; } return BB_L2_INV_EVTYPE; } static inline void bb_l2_pmnc_write(unsigned long val) { val &= 0xff; asm volatile ("mcr p15, 3, %0, c15, c4, 0" : : "r" (val)); } static inline unsigned long bb_l2_pmnc_read(void) { u32 val; asm volatile ("mrc p15, 3, %0, c15, c4, 0" : "=r" (val)); return val; } static void bb_l2_set_evcntcr(void) { u32 val = 0x0; asm volatile ("mcr p15, 3, %0, c15, c6, 4" : : "r" (val)); } static inline void bb_l2_set_evtyper(int ctr, int val) { /* select ctr */ asm volatile ("mcr p15, 3, %0, c15, c6, 0" : : "r" (ctr)); /* write into EVTYPER */ asm volatile ("mcr p15, 3, %0, c15, c6, 7" : : "r" (val)); } static void bb_l2_set_evfilter(void) { u32 filter_val = 0x000f0030 | 1 << smp_processor_id(); asm volatile ("mcr p15, 3, %0, c15, c6, 3" : : "r" (filter_val)); } static void bb_l2_enable_intenset(u32 idx) { if (idx == BB_L2CYCLE_CTR_EVENT_IDX) { asm volatile ("mcr p15, 3, %0, c15, c5, 1" : : "r" (1 << BB_L2CYCLE_CTR_BIT)); } else { asm volatile ("mcr p15, 3, %0, c15, c5, 1" : : "r" (1 << idx)); } } static void bb_l2_disable_intenclr(u32 idx) { if (idx == BB_L2CYCLE_CTR_EVENT_IDX) { asm volatile ("mcr p15, 3, %0, c15, c5, 0" : : "r" (1 << BB_L2CYCLE_CTR_BIT)); } else { asm volatile ("mcr p15, 3, %0, c15, c5, 0" : : "r" (1 << idx)); } } static void bb_l2_enable_counter(u32 idx) { if (idx == BB_L2CYCLE_CTR_EVENT_IDX) { asm volatile ("mcr p15, 3, %0, c15, c4, 3" : : "r" (1 << BB_L2CYCLE_CTR_BIT)); } else { asm volatile ("mcr p15, 3, %0, c15, c4, 3" : : "r" (1 << idx)); } } static void bb_l2_disable_counter(u32 idx) { if (idx == BB_L2CYCLE_CTR_EVENT_IDX) { asm volatile ("mcr p15, 3, %0, c15, c4, 2" : : "r" (1 << BB_L2CYCLE_CTR_BIT)); } else { asm volatile ("mcr p15, 3, %0, c15, c4, 2" : : "r" (1 << idx)); } } static u64 bb_l2_read_counter(u32 idx) { u32 val; unsigned long flags; if (idx == BB_L2CYCLE_CTR_EVENT_IDX) { asm volatile ("mrc p15, 3, %0, c15, c4, 5" : "=r" (val)); } else { raw_spin_lock_irqsave(&bb_l2_pmu_lock, flags); asm volatile ("mcr p15, 3, %0, c15, c6, 0" : : "r" (idx)); /* read val from counter */ asm volatile ("mrc p15, 3, %0, c15, c6, 5" : "=r" (val)); raw_spin_unlock_irqrestore(&bb_l2_pmu_lock, flags); } return val; } static void bb_l2_write_counter(u32 idx, u32 val) { unsigned long flags; if (idx == BB_L2CYCLE_CTR_EVENT_IDX) { asm volatile ("mcr p15, 3, %0, c15, c4, 5" : : "r" (val)); } else { raw_spin_lock_irqsave(&bb_l2_pmu_lock, flags); /* select counter */ asm volatile ("mcr p15, 3, %0, c15, c6, 0" : : "r" (idx)); /* write val into counter */ asm volatile ("mcr p15, 3, %0, c15, c6, 5" : : "r" (val)); raw_spin_unlock_irqrestore(&bb_l2_pmu_lock, flags); } } static int bb_pmu_event_set_period(struct perf_event *event, struct hw_perf_event *hwc, int idx) { s64 left = local64_read(&hwc->period_left); s64 period = hwc->sample_period; int ret = 0; if (unlikely(left <= -period)) { left = period; local64_set(&hwc->period_left, left); hwc->last_period = period; ret = 1; } if (unlikely(left <= 0)) { left += period; local64_set(&hwc->period_left, left); hwc->last_period = period; ret = 1; } if (left > (s64) MAX_BB_L2_PERIOD) left = MAX_BB_L2_PERIOD; local64_set(&hwc->prev_count, (u64)-left); bb_l2_write_counter(idx, (u64) (-left) & 0xffffffff); perf_event_update_userpage(event); return ret; } static u64 bb_pmu_event_update(struct perf_event *event, struct hw_perf_event *hwc, int idx, int overflow) { u64 prev_raw_count, new_raw_count; u64 delta; again: prev_raw_count = local64_read(&hwc->prev_count); new_raw_count = bb_l2_read_counter(idx); if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, new_raw_count) != prev_raw_count) goto again; new_raw_count &= MAX_BB_L2_PERIOD; prev_raw_count &= MAX_BB_L2_PERIOD; if (overflow) { delta = MAX_BB_L2_PERIOD - prev_raw_count + new_raw_count; pr_err("%s: delta: %lld\n", __func__, delta); } else delta = new_raw_count - prev_raw_count; local64_add(delta, &event->count); local64_sub(delta, &hwc->period_left); pr_debug("%s: new: %lld, prev: %lld, event: %ld count: %lld\n", __func__, new_raw_count, prev_raw_count, hwc->config_base, local64_read(&event->count)); return new_raw_count; } static void bb_l2_read(struct perf_event *event) { struct hw_perf_event *hwc = &event->hw; bb_pmu_event_update(event, hwc, hwc->idx, 0); } static void bb_l2_stop_counter(struct perf_event *event, int flags) { struct hw_perf_event *hwc = &event->hw; int idx = hwc->idx; if (!(hwc->state & PERF_HES_STOPPED)) { bb_l2_disable_intenclr(idx); bb_l2_disable_counter(idx); bb_pmu_event_update(event, hwc, idx, 0); hwc->state |= PERF_HES_STOPPED | PERF_HES_UPTODATE; } pr_debug("%s: event: %ld ctr: %d stopped\n", __func__, hwc->config_base, idx); } static void bb_l2_start_counter(struct perf_event *event, int flags) { struct hw_perf_event *hwc = &event->hw; int idx = hwc->idx; struct bb_l2_scorp_evt evtinfo; int evtype = hwc->config_base; int ev_typer; unsigned long iflags; int cpu_id = smp_processor_id(); if (flags & PERF_EF_RELOAD) WARN_ON_ONCE(!(hwc->state & PERF_HES_UPTODATE)); hwc->state = 0; bb_pmu_event_set_period(event, hwc, idx); if (hwc->config_base == BB_L2CYCLE_CTR_RAW_CODE) goto out; memset(&evtinfo, 0, sizeof(evtinfo)); ev_typer = get_bb_l2_evtinfo(evtype, &evtinfo); raw_spin_lock_irqsave(&bb_l2_pmu_lock, iflags); bb_l2_set_evtyper(idx, ev_typer); bb_l2_set_evcntcr(); bb_l2_set_evfilter(); bb_l2_evt_setup(evtinfo.grp, evtinfo.val); raw_spin_unlock_irqrestore(&bb_l2_pmu_lock, iflags); out: bb_l2_enable_intenset(idx); bb_l2_enable_counter(idx); pr_debug("%s: idx: %d, event: %d, val: %x, cpu: %d\n", __func__, idx, evtype, evtinfo.val, cpu_id); } static void bb_l2_del_event(struct perf_event *event, int flags) { struct hw_perf_event *hwc = &event->hw; int idx = hwc->idx; unsigned long iflags; raw_spin_lock_irqsave(&hw_bb_l2_pmu.lock, iflags); clear_bit(idx, (long unsigned int *)(&hw_bb_l2_pmu.active_mask)); bb_l2_stop_counter(event, PERF_EF_UPDATE); hw_bb_l2_pmu.events[idx] = NULL; hwc->idx = -1; raw_spin_unlock_irqrestore(&hw_bb_l2_pmu.lock, iflags); pr_debug("%s: event: %ld deleted\n", __func__, hwc->config_base); perf_event_update_userpage(event); } static int bb_l2_add_event(struct perf_event *event, int flags) { int ctr = 0; struct hw_perf_event *hwc = &event->hw; unsigned long iflags; int err = 0; perf_pmu_disable(event->pmu); raw_spin_lock_irqsave(&hw_bb_l2_pmu.lock, iflags); /* Cycle counter has a resrvd index */ if (hwc->config_base == BB_L2CYCLE_CTR_RAW_CODE) { if (hw_bb_l2_pmu.events[BB_L2CYCLE_CTR_EVENT_IDX]) { pr_err("%s: Stale cycle ctr event ptr !\n", __func__); err = -EINVAL; goto out; } hwc->idx = BB_L2CYCLE_CTR_EVENT_IDX; hw_bb_l2_pmu.events[BB_L2CYCLE_CTR_EVENT_IDX] = event; set_bit(BB_L2CYCLE_CTR_EVENT_IDX, (long unsigned int *)&hw_bb_l2_pmu.active_mask); goto skip_ctr_loop; } for (ctr = 0; ctr < MAX_BB_L2_CTRS - 1; ctr++) { if (!hw_bb_l2_pmu.events[ctr]) { hwc->idx = ctr; hw_bb_l2_pmu.events[ctr] = event; set_bit(ctr, (long unsigned int *) &hw_bb_l2_pmu.active_mask); break; } } if (hwc->idx < 0) { err = -ENOSPC; pr_err("%s: No space for event: %llx!!\n", __func__, event->attr.config); goto out; } skip_ctr_loop: bb_l2_disable_counter(hwc->idx); hwc->state = PERF_HES_STOPPED | PERF_HES_UPTODATE; if (flags & PERF_EF_START) bb_l2_start_counter(event, PERF_EF_RELOAD); perf_event_update_userpage(event); pr_debug("%s: event: %ld, ctr: %d added from cpu:%d\n", __func__, hwc->config_base, hwc->idx, smp_processor_id()); out: raw_spin_unlock_irqrestore(&hw_bb_l2_pmu.lock, iflags); /* Resume the PMU even if this event could not be added */ perf_pmu_enable(event->pmu); return err; } static void bb_l2_pmu_enable(struct pmu *pmu) { unsigned long flags; isb(); raw_spin_lock_irqsave(&bb_l2_pmu_lock, flags); /* Enable all counters */ bb_l2_pmnc_write(bb_l2_pmnc_read() | SCORPIONL2_PMNC_E); raw_spin_unlock_irqrestore(&bb_l2_pmu_lock, flags); } static void bb_l2_pmu_disable(struct pmu *pmu) { unsigned long flags; raw_spin_lock_irqsave(&bb_l2_pmu_lock, flags); /* Disable all counters */ bb_l2_pmnc_write(bb_l2_pmnc_read() & ~SCORPIONL2_PMNC_E); raw_spin_unlock_irqrestore(&bb_l2_pmu_lock, flags); isb(); } static inline u32 bb_l2_get_reset_pmovsr(void) { u32 val; /* Read */ asm volatile ("mrc p15, 3, %0, c15, c4, 1" : "=r" (val)); /* Write to clear flags */ val &= 0xffffffff; asm volatile ("mcr p15, 3, %0, c15, c4, 1" : : "r" (val)); return val; } static irqreturn_t bb_l2_handle_irq(int irq_num, void *dev) { unsigned long pmovsr; struct perf_sample_data data; struct pt_regs *regs; struct perf_event *event; struct hw_perf_event *hwc; int bitp; int idx = 0; pmovsr = bb_l2_get_reset_pmovsr(); if (!(pmovsr & 0xffffffff)) return IRQ_NONE; regs = get_irq_regs(); perf_sample_data_init(&data, 0); raw_spin_lock(&hw_bb_l2_pmu.lock); while (pmovsr) { bitp = __ffs(pmovsr); if (bitp == BB_L2CYCLE_CTR_BIT) idx = BB_L2CYCLE_CTR_EVENT_IDX; else idx = bitp; event = hw_bb_l2_pmu.events[idx]; if (!event) goto next; if (!test_bit(idx, hw_bb_l2_pmu.active_mask)) goto next; hwc = &event->hw; bb_pmu_event_update(event, hwc, idx, 1); data.period = event->hw.last_period; if (!bb_pmu_event_set_period(event, hwc, idx)) goto next; if (perf_event_overflow(event, 0, &data, regs)) bb_l2_disable_counter(hwc->idx); next: pmovsr &= (pmovsr - 1); } raw_spin_unlock(&hw_bb_l2_pmu.lock); irq_work_run(); return IRQ_HANDLED; } static atomic_t active_bb_l2_events = ATOMIC_INIT(0); static DEFINE_MUTEX(bb_pmu_reserve_mutex); static int bb_pmu_reserve_hardware(void) { int i, err = -ENODEV, irq; bb_l2_pmu_device = reserve_pmu(ARM_PMU_DEVICE_L2); if (IS_ERR(bb_l2_pmu_device)) { pr_warning("unable to reserve pmu\n"); return PTR_ERR(bb_l2_pmu_device); } if (bb_l2_pmu_device->num_resources < 1) { pr_err("no irqs for PMUs defined\n"); return -ENODEV; } if (strncmp(bb_l2_pmu_device->name, "l2-arm-pmu", 6)) { pr_err("Incorrect pdev reserved !\n"); return -EINVAL; } for (i = 0; i < bb_l2_pmu_device->num_resources; ++i) { irq = platform_get_irq(bb_l2_pmu_device, i); if (irq < 0) continue; err = request_irq(irq, bb_l2_handle_irq, IRQF_DISABLED | IRQF_NOBALANCING, "bb-l2-pmu", NULL); if (err) { pr_warning("unable to request IRQ%d for Krait L2 perf " "counters\n", irq); break; } irq_get_chip(irq)->irq_unmask(irq_get_irq_data(irq)); } if (err) { for (i = i - 1; i >= 0; --i) { irq = platform_get_irq(bb_l2_pmu_device, i); if (irq >= 0) free_irq(irq, NULL); } release_pmu(bb_l2_pmu_device); bb_l2_pmu_device = NULL; } return err; } static void bb_pmu_release_hardware(void) { int i, irq; for (i = bb_l2_pmu_device->num_resources - 1; i >= 0; --i) { irq = platform_get_irq(bb_l2_pmu_device, i); if (irq >= 0) free_irq(irq, NULL); } bb_l2_pmu_disable(NULL); release_pmu(bb_l2_pmu_device); bb_l2_pmu_device = NULL; } static void bb_pmu_perf_event_destroy(struct perf_event *event) { if (atomic_dec_and_mutex_lock (&active_bb_l2_events, &bb_pmu_reserve_mutex)) { bb_pmu_release_hardware(); mutex_unlock(&bb_pmu_reserve_mutex); } } static int bb_l2_event_init(struct perf_event *event) { int err = 0; struct hw_perf_event *hwc = &event->hw; int status = 0; switch (event->attr.type) { case PERF_TYPE_SHARED: break; default: return -ENOENT; } hwc->idx = -1; event->destroy = bb_pmu_perf_event_destroy; if (!atomic_inc_not_zero(&active_bb_l2_events)) { /* 0 active events */ mutex_lock(&bb_pmu_reserve_mutex); err = bb_pmu_reserve_hardware(); mutex_unlock(&bb_pmu_reserve_mutex); if (!err) atomic_inc(&active_bb_l2_events); else return err; } else { if (atomic_read(&active_bb_l2_events) > (MAX_BB_L2_CTRS - 1)) { pr_err("%s: No space left on PMU for event: %llx\n", __func__, event->attr.config); atomic_dec(&active_bb_l2_events); return -ENOSPC; } } hwc->config_base = event->attr.config & 0xff; hwc->config = 0; hwc->event_base = 0; /* Only one CPU can control the cycle counter */ if (hwc->config_base == BB_L2CYCLE_CTR_RAW_CODE) { /* Check if its already running */ asm volatile ("mrc p15, 3, %0, c15, c4, 6" : "=r" (status)); if (status == 0x2) { err = -ENOSPC; goto out; } } if (!hwc->sample_period) { hwc->sample_period = MAX_BB_L2_PERIOD; hwc->last_period = hwc->sample_period; local64_set(&hwc->period_left, hwc->sample_period); } pr_debug("%s: event: %lld init'd\n", __func__, event->attr.config); out: if (err < 0) bb_pmu_perf_event_destroy(event); return err; } static struct pmu bb_l2_pmu = { .pmu_enable = bb_l2_pmu_enable, .pmu_disable = bb_l2_pmu_disable, .event_init = bb_l2_event_init, .add = bb_l2_add_event, .del = bb_l2_del_event, .start = bb_l2_start_counter, .stop = bb_l2_stop_counter, .read = bb_l2_read, }; static const struct arm_pmu *__init scorpionmp_l2_pmu_init(void) { /* Register our own PMU here */ perf_pmu_register(&bb_l2_pmu, "BB L2", PERF_TYPE_SHARED); memset(&hw_bb_l2_pmu, 0, sizeof(hw_bb_l2_pmu)); /* Avoid spurious interrupts at startup */ bb_l2_get_reset_pmovsr(); /* Don't return an arm_pmu here */ return NULL; } #else static const struct arm_pmu *__init scorpionmp_l2_pmu_init(void) { return NULL; } #endif