mirror of
https://github.com/FEX-Emu/linux.git
synced 2024-12-30 13:38:40 +00:00
511f838945
Previously tile was rolling its own method of capturing backtrace data in the NMI handlers, but it was relying on running printk() from the NMI handler, which is not always safe. So adopt the nmi_backtrace model (with the new cpumask extension) instead. So we can call the nmi_backtrace code directly from the nmi handler, move the nmi_enter()/exit() into the top-level tile NMI handler. The semantics of the routine change slightly since it is now synchronous with the remote cores completing the backtraces. Previously it was asynchronous, but with protection to avoid starting a new remote backtrace if the old one was still in progress. Link: http://lkml.kernel.org/r/1472487169-14923-4-git-send-email-cmetcalf@mellanox.com Signed-off-by: Chris Metcalf <cmetcalf@mellanox.com> Cc: Daniel Thompson <daniel.thompson@linaro.org> [arm] Cc: Petr Mladek <pmladek@suse.com> Cc: Aaron Tomlin <atomlin@redhat.com> Cc: Peter Zijlstra (Intel) <peterz@infradead.org> Cc: "Rafael J. Wysocki" <rjw@rjwysocki.net> Cc: Russell King <linux@arm.linux.org.uk> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: Ingo Molnar <mingo@elte.hu> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
119 lines
3.1 KiB
C
119 lines
3.1 KiB
C
/*
|
|
* Copyright 2014 Tilera Corporation. All Rights Reserved.
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License
|
|
* as published by the Free Software Foundation, version 2.
|
|
*
|
|
* This program is distributed in the hope that it will be useful, but
|
|
* WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or
|
|
* NON INFRINGEMENT. See the GNU General Public License for
|
|
* more details.
|
|
*/
|
|
|
|
#include <linux/errno.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/module.h>
|
|
#include <linux/atomic.h>
|
|
|
|
#include <asm/processor.h>
|
|
#include <asm/pmc.h>
|
|
|
|
perf_irq_t perf_irq = NULL;
|
|
int handle_perf_interrupt(struct pt_regs *regs, int fault)
|
|
{
|
|
int retval;
|
|
|
|
if (!perf_irq)
|
|
panic("Unexpected PERF_COUNT interrupt %d\n", fault);
|
|
|
|
retval = perf_irq(regs, fault);
|
|
return retval;
|
|
}
|
|
|
|
/* Reserve PMC hardware if it is available. */
|
|
perf_irq_t reserve_pmc_hardware(perf_irq_t new_perf_irq)
|
|
{
|
|
return cmpxchg(&perf_irq, NULL, new_perf_irq);
|
|
}
|
|
EXPORT_SYMBOL(reserve_pmc_hardware);
|
|
|
|
/* Release PMC hardware. */
|
|
void release_pmc_hardware(void)
|
|
{
|
|
perf_irq = NULL;
|
|
}
|
|
EXPORT_SYMBOL(release_pmc_hardware);
|
|
|
|
|
|
/*
|
|
* Get current overflow status of each performance counter,
|
|
* and auxiliary performance counter.
|
|
*/
|
|
unsigned long
|
|
pmc_get_overflow(void)
|
|
{
|
|
unsigned long status;
|
|
|
|
/*
|
|
* merge base+aux into a single vector
|
|
*/
|
|
status = __insn_mfspr(SPR_PERF_COUNT_STS);
|
|
status |= __insn_mfspr(SPR_AUX_PERF_COUNT_STS) << TILE_BASE_COUNTERS;
|
|
return status;
|
|
}
|
|
|
|
/*
|
|
* Clear the status bit for the corresponding counter, if written
|
|
* with a one.
|
|
*/
|
|
void
|
|
pmc_ack_overflow(unsigned long status)
|
|
{
|
|
/*
|
|
* clear overflow status by writing ones
|
|
*/
|
|
__insn_mtspr(SPR_PERF_COUNT_STS, status);
|
|
__insn_mtspr(SPR_AUX_PERF_COUNT_STS, status >> TILE_BASE_COUNTERS);
|
|
}
|
|
|
|
/*
|
|
* The perf count interrupts are masked and unmasked explicitly,
|
|
* and only here. The normal irq_enable() does not enable them,
|
|
* and irq_disable() does not disable them. That lets these
|
|
* routines drive the perf count interrupts orthogonally.
|
|
*
|
|
* We also mask the perf count interrupts on entry to the perf count
|
|
* interrupt handler in assembly code, and by default unmask them
|
|
* again (with interrupt critical section protection) just before
|
|
* returning from the interrupt. If the perf count handler returns
|
|
* a non-zero error code, then we don't re-enable them before returning.
|
|
*
|
|
* For Pro, we rely on both interrupts being in the same word to update
|
|
* them atomically so we never have one enabled and one disabled.
|
|
*/
|
|
|
|
#if CHIP_HAS_SPLIT_INTR_MASK()
|
|
# if INT_PERF_COUNT < 32 || INT_AUX_PERF_COUNT < 32
|
|
# error Fix assumptions about which word PERF_COUNT interrupts are in
|
|
# endif
|
|
#endif
|
|
|
|
static inline unsigned long long pmc_mask(void)
|
|
{
|
|
unsigned long long mask = 1ULL << INT_PERF_COUNT;
|
|
mask |= 1ULL << INT_AUX_PERF_COUNT;
|
|
return mask;
|
|
}
|
|
|
|
void unmask_pmc_interrupts(void)
|
|
{
|
|
interrupt_mask_reset_mask(pmc_mask());
|
|
}
|
|
|
|
void mask_pmc_interrupts(void)
|
|
{
|
|
interrupt_mask_set_mask(pmc_mask());
|
|
}
|