mirror of
https://github.com/FEX-Emu/linux.git
synced 2024-12-22 09:22:37 +00:00
b39898cd40
On a 68k platform a couple of interrupts are demultiplexed and "polled" from a top level interrupt. Unfortunately there is no way to determine which of the sub interrupts raised the top level interrupt, so all of the demultiplexed interrupt handlers need to be invoked. Given a high enough frequency this can trigger the spurious interrupt detection mechanism, if one of the demultiplex interrupts returns IRQ_NONE continuously. But this is a false positive as the polling causes this behaviour and not buggy hardware/software. Introduce IRQ_POLLED which can be set at interrupt chip setup time via irq_set_status_flags(). The flag excludes the interrupt from the spurious detector and from all core polling activities. Reported-and-tested-by: Michael Schmitz <schmitzmic@gmail.com> Cc: Geert Uytterhoeven <geert@linux-m68k.org> Cc: linux-m68k@vger.kernel.org Link: http://lkml.kernel.org/r/alpine.DEB.2.02.1311061149250.23353@ionos.tec.linutronix.de
157 lines
4.0 KiB
C
157 lines
4.0 KiB
C
/*
|
|
* Internal header to deal with irq_desc->status which will be renamed
|
|
* to irq_desc->settings.
|
|
*/
|
|
enum {
|
|
_IRQ_DEFAULT_INIT_FLAGS = IRQ_DEFAULT_INIT_FLAGS,
|
|
_IRQ_PER_CPU = IRQ_PER_CPU,
|
|
_IRQ_LEVEL = IRQ_LEVEL,
|
|
_IRQ_NOPROBE = IRQ_NOPROBE,
|
|
_IRQ_NOREQUEST = IRQ_NOREQUEST,
|
|
_IRQ_NOTHREAD = IRQ_NOTHREAD,
|
|
_IRQ_NOAUTOEN = IRQ_NOAUTOEN,
|
|
_IRQ_MOVE_PCNTXT = IRQ_MOVE_PCNTXT,
|
|
_IRQ_NO_BALANCING = IRQ_NO_BALANCING,
|
|
_IRQ_NESTED_THREAD = IRQ_NESTED_THREAD,
|
|
_IRQ_PER_CPU_DEVID = IRQ_PER_CPU_DEVID,
|
|
_IRQ_IS_POLLED = IRQ_IS_POLLED,
|
|
_IRQF_MODIFY_MASK = IRQF_MODIFY_MASK,
|
|
};
|
|
|
|
#define IRQ_PER_CPU GOT_YOU_MORON
|
|
#define IRQ_NO_BALANCING GOT_YOU_MORON
|
|
#define IRQ_LEVEL GOT_YOU_MORON
|
|
#define IRQ_NOPROBE GOT_YOU_MORON
|
|
#define IRQ_NOREQUEST GOT_YOU_MORON
|
|
#define IRQ_NOTHREAD GOT_YOU_MORON
|
|
#define IRQ_NOAUTOEN GOT_YOU_MORON
|
|
#define IRQ_NESTED_THREAD GOT_YOU_MORON
|
|
#define IRQ_PER_CPU_DEVID GOT_YOU_MORON
|
|
#define IRQ_IS_POLLED GOT_YOU_MORON
|
|
#undef IRQF_MODIFY_MASK
|
|
#define IRQF_MODIFY_MASK GOT_YOU_MORON
|
|
|
|
static inline void
|
|
irq_settings_clr_and_set(struct irq_desc *desc, u32 clr, u32 set)
|
|
{
|
|
desc->status_use_accessors &= ~(clr & _IRQF_MODIFY_MASK);
|
|
desc->status_use_accessors |= (set & _IRQF_MODIFY_MASK);
|
|
}
|
|
|
|
static inline bool irq_settings_is_per_cpu(struct irq_desc *desc)
|
|
{
|
|
return desc->status_use_accessors & _IRQ_PER_CPU;
|
|
}
|
|
|
|
static inline bool irq_settings_is_per_cpu_devid(struct irq_desc *desc)
|
|
{
|
|
return desc->status_use_accessors & _IRQ_PER_CPU_DEVID;
|
|
}
|
|
|
|
static inline void irq_settings_set_per_cpu(struct irq_desc *desc)
|
|
{
|
|
desc->status_use_accessors |= _IRQ_PER_CPU;
|
|
}
|
|
|
|
static inline void irq_settings_set_no_balancing(struct irq_desc *desc)
|
|
{
|
|
desc->status_use_accessors |= _IRQ_NO_BALANCING;
|
|
}
|
|
|
|
static inline bool irq_settings_has_no_balance_set(struct irq_desc *desc)
|
|
{
|
|
return desc->status_use_accessors & _IRQ_NO_BALANCING;
|
|
}
|
|
|
|
static inline u32 irq_settings_get_trigger_mask(struct irq_desc *desc)
|
|
{
|
|
return desc->status_use_accessors & IRQ_TYPE_SENSE_MASK;
|
|
}
|
|
|
|
static inline void
|
|
irq_settings_set_trigger_mask(struct irq_desc *desc, u32 mask)
|
|
{
|
|
desc->status_use_accessors &= ~IRQ_TYPE_SENSE_MASK;
|
|
desc->status_use_accessors |= mask & IRQ_TYPE_SENSE_MASK;
|
|
}
|
|
|
|
static inline bool irq_settings_is_level(struct irq_desc *desc)
|
|
{
|
|
return desc->status_use_accessors & _IRQ_LEVEL;
|
|
}
|
|
|
|
static inline void irq_settings_clr_level(struct irq_desc *desc)
|
|
{
|
|
desc->status_use_accessors &= ~_IRQ_LEVEL;
|
|
}
|
|
|
|
static inline void irq_settings_set_level(struct irq_desc *desc)
|
|
{
|
|
desc->status_use_accessors |= _IRQ_LEVEL;
|
|
}
|
|
|
|
static inline bool irq_settings_can_request(struct irq_desc *desc)
|
|
{
|
|
return !(desc->status_use_accessors & _IRQ_NOREQUEST);
|
|
}
|
|
|
|
static inline void irq_settings_clr_norequest(struct irq_desc *desc)
|
|
{
|
|
desc->status_use_accessors &= ~_IRQ_NOREQUEST;
|
|
}
|
|
|
|
static inline void irq_settings_set_norequest(struct irq_desc *desc)
|
|
{
|
|
desc->status_use_accessors |= _IRQ_NOREQUEST;
|
|
}
|
|
|
|
static inline bool irq_settings_can_thread(struct irq_desc *desc)
|
|
{
|
|
return !(desc->status_use_accessors & _IRQ_NOTHREAD);
|
|
}
|
|
|
|
static inline void irq_settings_clr_nothread(struct irq_desc *desc)
|
|
{
|
|
desc->status_use_accessors &= ~_IRQ_NOTHREAD;
|
|
}
|
|
|
|
static inline void irq_settings_set_nothread(struct irq_desc *desc)
|
|
{
|
|
desc->status_use_accessors |= _IRQ_NOTHREAD;
|
|
}
|
|
|
|
static inline bool irq_settings_can_probe(struct irq_desc *desc)
|
|
{
|
|
return !(desc->status_use_accessors & _IRQ_NOPROBE);
|
|
}
|
|
|
|
static inline void irq_settings_clr_noprobe(struct irq_desc *desc)
|
|
{
|
|
desc->status_use_accessors &= ~_IRQ_NOPROBE;
|
|
}
|
|
|
|
static inline void irq_settings_set_noprobe(struct irq_desc *desc)
|
|
{
|
|
desc->status_use_accessors |= _IRQ_NOPROBE;
|
|
}
|
|
|
|
static inline bool irq_settings_can_move_pcntxt(struct irq_desc *desc)
|
|
{
|
|
return desc->status_use_accessors & _IRQ_MOVE_PCNTXT;
|
|
}
|
|
|
|
static inline bool irq_settings_can_autoenable(struct irq_desc *desc)
|
|
{
|
|
return !(desc->status_use_accessors & _IRQ_NOAUTOEN);
|
|
}
|
|
|
|
static inline bool irq_settings_is_nested_thread(struct irq_desc *desc)
|
|
{
|
|
return desc->status_use_accessors & _IRQ_NESTED_THREAD;
|
|
}
|
|
|
|
static inline bool irq_settings_is_polled(struct irq_desc *desc)
|
|
{
|
|
return desc->status_use_accessors & _IRQ_IS_POLLED;
|
|
}
|