Release 4.12 drivers/irqchip/irq-i8259.c
  
  
  
/*
 * This file is subject to the terms and conditions of the GNU General Public
 * License.  See the file "COPYING" in the main directory of this archive
 * for more details.
 *
 * Code to handle x86 style IRQs plus some generic interrupt stuff.
 *
 * Copyright (C) 1992 Linus Torvalds
 * Copyright (C) 1994 - 2000 Ralf Baechle
 */
#include <linux/delay.h>
#include <linux/init.h>
#include <linux/ioport.h>
#include <linux/interrupt.h>
#include <linux/irqchip.h>
#include <linux/irqdomain.h>
#include <linux/kernel.h>
#include <linux/of_irq.h>
#include <linux/spinlock.h>
#include <linux/syscore_ops.h>
#include <linux/irq.h>
#include <asm/i8259.h>
#include <asm/io.h>
/*
 * This is the 'legacy' 8259A Programmable Interrupt Controller,
 * present in the majority of PC/AT boxes.
 * plus some generic x86 specific things if generic specifics makes
 * any sense at all.
 * this file should become arch/i386/kernel/irq.c when the old irq.c
 * moves to arch independent land
 */
static int i8259A_auto_eoi = -1;
DEFINE_RAW_SPINLOCK(i8259A_lock);
static void disable_8259A_irq(struct irq_data *d);
static void enable_8259A_irq(struct irq_data *d);
static void mask_and_ack_8259A(struct irq_data *d);
static void init_8259A(int auto_eoi);
static int (*i8259_poll)(void) = i8259_irq;
static struct irq_chip i8259A_chip = {
	.name			= "XT-PIC",
	.irq_mask		= disable_8259A_irq,
	.irq_disable		= disable_8259A_irq,
	.irq_unmask		= enable_8259A_irq,
	.irq_mask_ack		= mask_and_ack_8259A,
};
/*
 * 8259A PIC functions to handle ISA devices:
 */
void i8259_set_poll(int (*poll)(void))
{
	i8259_poll = poll;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Paul Burton | 18 | 100.00% | 1 | 100.00% | 
| Total | 18 | 100.00% | 1 | 100.00% | 
/*
 * This contains the irq mask for both 8259A irq controllers,
 */
static unsigned int cached_irq_mask = 0xffff;
#define cached_master_mask	(cached_irq_mask)
#define cached_slave_mask	(cached_irq_mask >> 8)
static void disable_8259A_irq(struct irq_data *d)
{
	unsigned int mask, irq = d->irq - I8259A_IRQ_BASE;
	unsigned long flags;
	mask = 1 << irq;
	raw_spin_lock_irqsave(&i8259A_lock, flags);
	cached_irq_mask |= mask;
	if (irq & 8)
		outb(cached_slave_mask, PIC_SLAVE_IMR);
	else
		outb(cached_master_mask, PIC_MASTER_IMR);
	raw_spin_unlock_irqrestore(&i8259A_lock, flags);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Linus Torvalds | 49 | 66.22% | 1 | 16.67% | 
| Thomas Gleixner | 12 | 16.22% | 1 | 16.67% | 
| Atsushi Nemoto | 10 | 13.51% | 2 | 33.33% | 
| Ralf Bächle | 2 | 2.70% | 1 | 16.67% | 
| Yoichi Yuasa | 1 | 1.35% | 1 | 16.67% | 
| Total | 74 | 100.00% | 6 | 100.00% | 
static void enable_8259A_irq(struct irq_data *d)
{
	unsigned int mask, irq = d->irq - I8259A_IRQ_BASE;
	unsigned long flags;
	mask = ~(1 << irq);
	raw_spin_lock_irqsave(&i8259A_lock, flags);
	cached_irq_mask &= mask;
	if (irq & 8)
		outb(cached_slave_mask, PIC_SLAVE_IMR);
	else
		outb(cached_master_mask, PIC_MASTER_IMR);
	raw_spin_unlock_irqrestore(&i8259A_lock, flags);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Linus Torvalds | 50 | 64.94% | 1 | 16.67% | 
| Thomas Gleixner | 12 | 15.58% | 1 | 16.67% | 
| Atsushi Nemoto | 12 | 15.58% | 2 | 33.33% | 
| Ralf Bächle | 2 | 2.60% | 1 | 16.67% | 
| Yoichi Yuasa | 1 | 1.30% | 1 | 16.67% | 
| Total | 77 | 100.00% | 6 | 100.00% | 
void make_8259A_irq(unsigned int irq)
{
	disable_irq_nosync(irq);
	irq_set_chip_and_handler(irq, &i8259A_chip, handle_level_irq);
	enable_irq(irq);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Linus Torvalds | 22 | 75.86% | 1 | 25.00% | 
| Atsushi Nemoto | 6 | 20.69% | 2 | 50.00% | 
| Thomas Gleixner | 1 | 3.45% | 1 | 25.00% | 
| Total | 29 | 100.00% | 4 | 100.00% | 
/*
 * This function assumes to be called rarely. Switching between
 * 8259A registers is slow.
 * This has to be protected by the irq controller spinlock
 * before being called.
 */
static inline int i8259A_irq_real(unsigned int irq)
{
	int value;
	int irqmask = 1 << irq;
	if (irq < 8) {
		outb(0x0B, PIC_MASTER_CMD);	/* ISR register */
		value = inb(PIC_MASTER_CMD) & irqmask;
		outb(0x0A, PIC_MASTER_CMD);	/* back to the IRR register */
		return value;
	}
	outb(0x0B, PIC_SLAVE_CMD);	/* ISR register */
	value = inb(PIC_SLAVE_CMD) & (irqmask >> 8);
	outb(0x0A, PIC_SLAVE_CMD);	/* back to the IRR register */
	return value;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Linus Torvalds | 83 | 93.26% | 1 | 50.00% | 
| Atsushi Nemoto | 6 | 6.74% | 1 | 50.00% | 
| Total | 89 | 100.00% | 2 | 100.00% | 
/*
 * Careful! The 8259A is a fragile beast, it pretty
 * much _has_ to be done exactly like this (mask it
 * first, _then_ send the EOI, and the order of EOI
 * to the two 8259s is important!
 */
static void mask_and_ack_8259A(struct irq_data *d)
{
	unsigned int irqmask, irq = d->irq - I8259A_IRQ_BASE;
	unsigned long flags;
	irqmask = 1 << irq;
	raw_spin_lock_irqsave(&i8259A_lock, flags);
	/*
         * Lightweight spurious IRQ detection. We do not want
         * to overdo spurious IRQ handling - it's usually a sign
         * of hardware problems, so we only do the checks we can
         * do without slowing down good hardware unnecessarily.
         *
         * Note that IRQ7 and IRQ15 (the two spurious IRQs
         * usually resulting from the 8259A-1|2 PICs) occur
         * even if the IRQ is masked in the 8259A. Thus we
         * can check spurious 8259A IRQs without doing the
         * quite slow i8259A_irq_real() call for every IRQ.
         * This does not cover 100% of spurious interrupts,
         * but should be enough to warn the user that there
         * is something bad going on ...
         */
	if (cached_irq_mask & irqmask)
		goto spurious_8259A_irq;
	cached_irq_mask |= irqmask;
handle_real_irq:
	if (irq & 8) {
		inb(PIC_SLAVE_IMR);	/* DUMMY - (do we need this?) */
		outb(cached_slave_mask, PIC_SLAVE_IMR);
		outb(0x60+(irq&7), PIC_SLAVE_CMD);/* 'Specific EOI' to slave */
		outb(0x60+PIC_CASCADE_IR, PIC_MASTER_CMD); /* 'Specific EOI' to master-IRQ2 */
	} else {
		inb(PIC_MASTER_IMR);	/* DUMMY - (do we need this?) */
		outb(cached_master_mask, PIC_MASTER_IMR);
		outb(0x60+irq, PIC_MASTER_CMD); /* 'Specific EOI to master */
	}
	raw_spin_unlock_irqrestore(&i8259A_lock, flags);
	return;
spurious_8259A_irq:
	/*
         * this is the slow path - should happen rarely.
         */
	if (i8259A_irq_real(irq))
		/*
                 * oops, the IRQ _is_ in service according to the
                 * 8259A - not spurious, go handle it.
                 */
		goto handle_real_irq;
	{
		static int spurious_irq_mask;
		/*
                 * At this point we can be sure the IRQ is spurious,
                 * lets ACK and report it. [once per IRQ]
                 */
		if (!(spurious_irq_mask & irqmask)) {
			printk(KERN_DEBUG "spurious 8259A interrupt: IRQ%d.\n", irq);
			spurious_irq_mask |= irqmask;
		}
		atomic_inc(&irq_err_count);
		/*
                 * Theoretically we do not have to handle this IRQ,
                 * but in Linux this does not cause problems and is
                 * simpler for us.
                 */
		goto handle_real_irq;
	}
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Linus Torvalds | 151 | 79.06% | 1 | 12.50% | 
| Atsushi Nemoto | 20 | 10.47% | 2 | 25.00% | 
| Thomas Gleixner | 12 | 6.28% | 1 | 12.50% | 
| Ralf Bächle | 7 | 3.66% | 3 | 37.50% | 
| Yoichi Yuasa | 1 | 0.52% | 1 | 12.50% | 
| Total | 191 | 100.00% | 8 | 100.00% | 
static void i8259A_resume(void)
{
	if (i8259A_auto_eoi >= 0)
		init_8259A(i8259A_auto_eoi);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Atsushi Nemoto | 10 | 52.63% | 2 | 50.00% | 
| Ralf Bächle | 7 | 36.84% | 1 | 25.00% | 
| Rafael J. Wysocki | 2 | 10.53% | 1 | 25.00% | 
| Total | 19 | 100.00% | 4 | 100.00% | 
static void i8259A_shutdown(void)
{
	/* Put the i8259A into a quiescent state that
         * the kernel initialization code can get it
         * out of.
         */
	if (i8259A_auto_eoi >= 0) {
		outb(0xff, PIC_MASTER_IMR);	/* mask all of 8259A-1 */
		outb(0xff, PIC_SLAVE_IMR);	/* mask all of 8259A-2 */
	}
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Atsushi Nemoto | 27 | 81.82% | 2 | 40.00% | 
| Ralf Bächle | 3 | 9.09% | 1 | 20.00% | 
| Rafael J. Wysocki | 2 | 6.06% | 1 | 20.00% | 
| Hillf Danton | 1 | 3.03% | 1 | 20.00% | 
| Total | 33 | 100.00% | 5 | 100.00% | 
static struct syscore_ops i8259_syscore_ops = {
	.resume = i8259A_resume,
	.shutdown = i8259A_shutdown,
};
static int __init i8259A_init_sysfs(void)
{
	register_syscore_ops(&i8259_syscore_ops);
	return 0;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Ralf Bächle | 15 | 83.33% | 1 | 50.00% | 
| Rafael J. Wysocki | 3 | 16.67% | 1 | 50.00% | 
| Total | 18 | 100.00% | 2 | 100.00% | 
device_initcall(i8259A_init_sysfs);
static void init_8259A(int auto_eoi)
{
	unsigned long flags;
	i8259A_auto_eoi = auto_eoi;
	raw_spin_lock_irqsave(&i8259A_lock, flags);
	outb(0xff, PIC_MASTER_IMR);	/* mask all of 8259A-1 */
	outb(0xff, PIC_SLAVE_IMR);	/* mask all of 8259A-2 */
	/*
         * outb_p - this has to work on a wide range of PC hardware.
         */
	outb_p(0x11, PIC_MASTER_CMD);	/* ICW1: select 8259A-1 init */
	outb_p(I8259A_IRQ_BASE + 0, PIC_MASTER_IMR);	/* ICW2: 8259A-1 IR0 mapped to I8259A_IRQ_BASE + 0x00 */
	outb_p(1U << PIC_CASCADE_IR, PIC_MASTER_IMR);	/* 8259A-1 (the master) has a slave on IR2 */
	if (auto_eoi)	/* master does Auto EOI */
		outb_p(MASTER_ICW4_DEFAULT | PIC_ICW4_AEOI, PIC_MASTER_IMR);
	else		/* master expects normal EOI */
		outb_p(MASTER_ICW4_DEFAULT, PIC_MASTER_IMR);
	outb_p(0x11, PIC_SLAVE_CMD);	/* ICW1: select 8259A-2 init */
	outb_p(I8259A_IRQ_BASE + 8, PIC_SLAVE_IMR);	/* ICW2: 8259A-2 IR0 mapped to I8259A_IRQ_BASE + 0x08 */
	outb_p(PIC_CASCADE_IR, PIC_SLAVE_IMR);	/* 8259A-2 is a slave on master's IR2 */
	outb_p(SLAVE_ICW4_DEFAULT, PIC_SLAVE_IMR); /* (slave's support for AEOI in flat mode is to be investigated) */
	if (auto_eoi)
		/*
                 * In AEOI mode we just have to mask the interrupt
                 * when acking.
                 */
		i8259A_chip.irq_mask_ack = disable_8259A_irq;
	else
		i8259A_chip.irq_mask_ack = mask_and_ack_8259A;
	udelay(100);		/* wait for 8259A to initialize */
	outb(cached_master_mask, PIC_MASTER_IMR); /* restore master IRQ mask */
	outb(cached_slave_mask, PIC_SLAVE_IMR);	  /* restore slave IRQ mask */
	raw_spin_unlock_irqrestore(&i8259A_lock, flags);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Linus Torvalds | 128 | 73.14% | 1 | 20.00% | 
| Atsushi Nemoto | 42 | 24.00% | 1 | 20.00% | 
| Ralf Bächle | 2 | 1.14% | 1 | 20.00% | 
| Thomas Gleixner | 2 | 1.14% | 1 | 20.00% | 
| Yoichi Yuasa | 1 | 0.57% | 1 | 20.00% | 
| Total | 175 | 100.00% | 5 | 100.00% | 
/*
 * IRQ2 is cascade interrupt to second interrupt controller
 */
static struct irqaction irq2 = {
	.handler = no_action,
	.name = "cascade",
	.flags = IRQF_NO_THREAD,
};
static struct resource pic1_io_resource = {
	.name = "pic1",
	.start = PIC_MASTER_CMD,
	.end = PIC_MASTER_IMR,
	.flags = IORESOURCE_BUSY
};
static struct resource pic2_io_resource = {
	.name = "pic2",
	.start = PIC_SLAVE_CMD,
	.end = PIC_SLAVE_IMR,
	.flags = IORESOURCE_BUSY
};
static int i8259A_irq_domain_map(struct irq_domain *d, unsigned int virq,
				 irq_hw_number_t hw)
{
	irq_set_chip_and_handler(virq, &i8259A_chip, handle_level_irq);
	irq_set_probe(virq);
	return 0;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Andrew Bresticker | 36 | 100.00% | 1 | 100.00% | 
| Total | 36 | 100.00% | 1 | 100.00% | 
static struct irq_domain_ops i8259A_ops = {
	.map = i8259A_irq_domain_map,
	.xlate = irq_domain_xlate_onecell,
};
/*
 * On systems with i8259-style interrupt controllers we assume for
 * driver compatibility reasons interrupts 0 - 15 to be the i8259
 * interrupts even if the hardware uses a different interrupt numbering.
 */
struct irq_domain * __init __init_i8259_irqs(struct device_node *node)
{
	struct irq_domain *domain;
	insert_resource(&ioport_resource, &pic1_io_resource);
	insert_resource(&ioport_resource, &pic2_io_resource);
	init_8259A(0);
	domain = irq_domain_add_legacy(node, 16, I8259A_IRQ_BASE, 0,
				       &i8259A_ops, NULL);
	if (!domain)
		panic("Failed to add i8259 IRQ domain");
	setup_irq(I8259A_IRQ_BASE + PIC_CASCADE_IR, &irq2);
	return domain;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Linus Torvalds | 36 | 43.90% | 2 | 22.22% | 
| Andrew Bresticker | 21 | 25.61% | 1 | 11.11% | 
| Paul Burton | 13 | 15.85% | 1 | 11.11% | 
| Atsushi Nemoto | 7 | 8.54% | 3 | 33.33% | 
| Ralf Bächle | 3 | 3.66% | 1 | 11.11% | 
| Thomas Bogendoerfer | 2 | 2.44% | 1 | 11.11% | 
| Total | 82 | 100.00% | 9 | 100.00% | 
void __init init_i8259_irqs(void)
{
	__init_i8259_irqs(NULL);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Paul Burton | 13 | 100.00% | 1 | 100.00% | 
| Total | 13 | 100.00% | 1 | 100.00% | 
static void i8259_irq_dispatch(struct irq_desc *desc)
{
	struct irq_domain *domain = irq_desc_get_handler_data(desc);
	int hwirq = i8259_poll();
	unsigned int irq;
	if (hwirq < 0)
		return;
	irq = irq_linear_revmap(domain, hwirq);
	generic_handle_irq(irq);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Paul Burton | 46 | 88.46% | 2 | 66.67% | 
| Thomas Gleixner | 6 | 11.54% | 1 | 33.33% | 
| Total | 52 | 100.00% | 3 | 100.00% | 
int __init i8259_of_init(struct device_node *node, struct device_node *parent)
{
	struct irq_domain *domain;
	unsigned int parent_irq;
	domain = __init_i8259_irqs(node);
	parent_irq = irq_of_parse_and_map(node, 0);
	if (!parent_irq) {
		pr_err("Failed to map i8259 parent IRQ\n");
		irq_domain_remove(domain);
		return -ENODEV;
	}
	irq_set_chained_handler_and_data(parent_irq, i8259_irq_dispatch,
					 domain);
	return 0;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Paul Burton | 70 | 94.59% | 2 | 50.00% | 
| Axel Lin | 3 | 4.05% | 1 | 25.00% | 
| Linus Torvalds | 1 | 1.35% | 1 | 25.00% | 
| Total | 74 | 100.00% | 4 | 100.00% | 
IRQCHIP_DECLARE(i8259, "intel,i8259", i8259_of_init);
Overall Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Linus Torvalds | 607 | 48.17% | 2 | 5.71% | 
| Paul Burton | 186 | 14.76% | 3 | 8.57% | 
| Atsushi Nemoto | 162 | 12.86% | 4 | 11.43% | 
| Ralf Bächle | 104 | 8.25% | 11 | 31.43% | 
| Andrew Bresticker | 77 | 6.11% | 1 | 2.86% | 
| Thomas Gleixner | 68 | 5.40% | 4 | 11.43% | 
| Yoichi Yuasa | 25 | 1.98% | 1 | 2.86% | 
| Rafael J. Wysocki | 10 | 0.79% | 1 | 2.86% | 
| Liming Wang | 5 | 0.40% | 1 | 2.86% | 
| Kyungmin Park | 4 | 0.32% | 1 | 2.86% | 
| Axel Lin | 3 | 0.24% | 1 | 2.86% | 
| David Howells | 3 | 0.24% | 1 | 2.86% | 
| Joël Porquet | 2 | 0.16% | 1 | 2.86% | 
| Thomas Bogendoerfer | 2 | 0.16% | 1 | 2.86% | 
| Hillf Danton | 1 | 0.08% | 1 | 2.86% | 
| Kay Sievers | 1 | 0.08% | 1 | 2.86% | 
| Total | 1260 | 100.00% | 35 | 100.00% | 
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.