irq.c 27 KB
Newer Older
Linus Torvalds's avatar
Linus Torvalds committed
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
/* $Id: irq.c,v 1.114 2002/01/11 08:45:38 davem Exp $
 * irq.c: UltraSparc IRQ handling/init/registry.
 *
 * Copyright (C) 1997  David S. Miller  (davem@caip.rutgers.edu)
 * Copyright (C) 1998  Eddie C. Dost    (ecd@skynet.be)
 * Copyright (C) 1998  Jakub Jelinek    (jj@ultra.linux.cz)
 */

#include <linux/config.h>
#include <linux/module.h>
#include <linux/sched.h>
#include <linux/ptrace.h>
#include <linux/errno.h>
#include <linux/kernel_stat.h>
#include <linux/signal.h>
#include <linux/mm.h>
#include <linux/interrupt.h>
#include <linux/slab.h>
#include <linux/random.h>
#include <linux/init.h>
#include <linux/delay.h>
#include <linux/proc_fs.h>
#include <linux/seq_file.h>
24
#include <linux/bootmem.h>
Linus Torvalds's avatar
Linus Torvalds committed
25 26 27 28 29 30

#include <asm/ptrace.h>
#include <asm/processor.h>
#include <asm/atomic.h>
#include <asm/system.h>
#include <asm/irq.h>
31
#include <asm/io.h>
Linus Torvalds's avatar
Linus Torvalds committed
32 33 34 35 36 37 38 39 40 41
#include <asm/sbus.h>
#include <asm/iommu.h>
#include <asm/upa.h>
#include <asm/oplib.h>
#include <asm/timer.h>
#include <asm/smp.h>
#include <asm/starfire.h>
#include <asm/uaccess.h>
#include <asm/cache.h>
#include <asm/cpudata.h>
42
#include <asm/auxio.h>
43
#include <asm/head.h>
Linus Torvalds's avatar
Linus Torvalds committed
44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70

#ifdef CONFIG_SMP
static void distribute_irqs(void);
#endif

/* UPA nodes send interrupt packet to UltraSparc with first data reg
 * value low 5 (7 on Starfire) bits holding the IRQ identifier being
 * delivered.  We must translate this into a non-vector IRQ so we can
 * set the softint on this cpu.
 *
 * To make processing these packets efficient and race free we use
 * an array of irq buckets below.  The interrupt vector handler in
 * entry.S feeds incoming packets into per-cpu pil-indexed lists.
 * The IVEC handler does not need to act atomically, the PIL dispatch
 * code uses CAS to get an atomic snapshot of the list and clear it
 * at the same time.
 */

struct ino_bucket ivector_table[NUM_IVECS] __attribute__ ((aligned (SMP_CACHE_BYTES)));

/* This has to be in the main kernel image, it cannot be
 * turned into per-cpu data.  The reason is that the main
 * kernel image is locked into the TLB and this structure
 * is accessed from the vectored interrupt trap handler.  If
 * access to this structure takes a TLB miss it could cause
 * the 5-level sparc v9 trap stack to overflow.
 */
71
#define irq_work(__cpu)	&(trap_block[(__cpu)].irq_worklist)
Linus Torvalds's avatar
Linus Torvalds committed
72

73
static struct irqaction *irq_action[NR_IRQS+1];
Linus Torvalds's avatar
Linus Torvalds committed
74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89

/* This only synchronizes entities which modify IRQ handler
 * state and some selected user-level spots that want to
 * read things in the table.  IRQ handler processing orders
 * its' accesses such that no locking is needed.
 */
static DEFINE_SPINLOCK(irq_action_lock);

static void register_irq_proc (unsigned int irq);

/*
 * Upper 2b of irqaction->flags holds the ino.
 * irqaction->mask holds the smp affinity information.
 */
#define put_ino_in_irqaction(action, irq) \
	action->flags &= 0xffffffffffffUL; \
90 91
	action->flags |= __irq_ino(irq) << 48;

Linus Torvalds's avatar
Linus Torvalds committed
92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113
#define get_ino_in_irqaction(action)	(action->flags >> 48)

#define put_smpaff_in_irqaction(action, smpaff)	(action)->mask = (smpaff)
#define get_smpaff_in_irqaction(action) 	((action)->mask)

int show_interrupts(struct seq_file *p, void *v)
{
	unsigned long flags;
	int i = *(loff_t *) v;
	struct irqaction *action;
#ifdef CONFIG_SMP
	int j;
#endif

	spin_lock_irqsave(&irq_action_lock, flags);
	if (i <= NR_IRQS) {
		if (!(action = *(i + irq_action)))
			goto out_unlock;
		seq_printf(p, "%3d: ", i);
#ifndef CONFIG_SMP
		seq_printf(p, "%10u ", kstat_irqs(i));
#else
114
		for_each_online_cpu(j) {
Linus Torvalds's avatar
Linus Torvalds committed
115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132
			seq_printf(p, "%10u ",
				   kstat_cpu(j).irqs[i]);
		}
#endif
		seq_printf(p, " %s:%lx", action->name,
			   get_ino_in_irqaction(action));
		for (action = action->next; action; action = action->next) {
			seq_printf(p, ", %s:%lx", action->name,
				   get_ino_in_irqaction(action));
		}
		seq_putc(p, '\n');
	}
out_unlock:
	spin_unlock_irqrestore(&irq_action_lock, flags);

	return 0;
}

133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169
extern unsigned long real_hard_smp_processor_id(void);

static unsigned int sun4u_compute_tid(unsigned long imap, unsigned long cpuid)
{
	unsigned int tid;

	if (this_is_starfire) {
		tid = starfire_translate(imap, cpuid);
		tid <<= IMAP_TID_SHIFT;
		tid &= IMAP_TID_UPA;
	} else {
		if (tlb_type == cheetah || tlb_type == cheetah_plus) {
			unsigned long ver;

			__asm__ ("rdpr %%ver, %0" : "=r" (ver));
			if ((ver >> 32UL) == __JALAPENO_ID ||
			    (ver >> 32UL) == __SERRANO_ID) {
				tid = cpuid << IMAP_TID_SHIFT;
				tid &= IMAP_TID_JBUS;
			} else {
				unsigned int a = cpuid & 0x1f;
				unsigned int n = (cpuid >> 5) & 0x1f;

				tid = ((a << IMAP_AID_SHIFT) |
				       (n << IMAP_NID_SHIFT));
				tid &= (IMAP_AID_SAFARI |
					IMAP_NID_SAFARI);;
			}
		} else {
			tid = cpuid << IMAP_TID_SHIFT;
			tid &= IMAP_TID_UPA;
		}
	}

	return tid;
}

Linus Torvalds's avatar
Linus Torvalds committed
170 171 172 173
/* Now these are always passed a true fully specified sun4u INO. */
void enable_irq(unsigned int irq)
{
	struct ino_bucket *bucket = __bucket(irq);
174
	unsigned long imap, cpuid;
Linus Torvalds's avatar
Linus Torvalds committed
175 176 177 178 179 180 181

	imap = bucket->imap;
	if (imap == 0UL)
		return;

	preempt_disable();

182 183 184 185 186
	/* This gets the physical processor ID, even on uniprocessor,
	 * so we can always program the interrupt target correctly.
	 */
	cpuid = real_hard_smp_processor_id();

187
	if (tlb_type == hypervisor) {
188
		unsigned int ino = __irq_ino(irq);
189
		int err;
190

191
		err = sun4v_intr_settarget(ino, cpuid);
192
		if (err != HV_EOK)
193 194
			printk("sun4v_intr_settarget(%x,%lu): err(%d)\n",
			       ino, cpuid, err);
195
		err = sun4v_intr_setenabled(ino, HV_INTR_ENABLED);
196 197 198
		if (err != HV_EOK)
			printk("sun4v_intr_setenabled(%x): err(%d)\n",
			       ino, err);
199
	} else {
200
		unsigned int tid = sun4u_compute_tid(imap, cpuid);
Linus Torvalds's avatar
Linus Torvalds committed
201

202 203 204 205 206 207 208 209 210 211 212
		/* NOTE NOTE NOTE, IGN and INO are read-only, IGN is a product
		 * of this SYSIO's preconfigured IGN in the SYSIO Control
		 * Register, the hardware just mirrors that value here.
		 * However for Graphics and UPA Slave devices the full
		 * IMAP_INR field can be set by the programmer here.
		 *
		 * Things like FFB can now be handled via the new IRQ
		 * mechanism.
		 */
		upa_writel(tid | IMAP_VALID, imap);
	}
Linus Torvalds's avatar
Linus Torvalds committed
213 214 215 216 217 218 219 220 221 222 223 224

	preempt_enable();
}

/* This now gets passed true ino's as well. */
void disable_irq(unsigned int irq)
{
	struct ino_bucket *bucket = __bucket(irq);
	unsigned long imap;

	imap = bucket->imap;
	if (imap != 0UL) {
225
		if (tlb_type == hypervisor) {
226
			unsigned int ino = __irq_ino(irq);
227
			int err;
228

229 230 231 232
			err = sun4v_intr_setenabled(ino, HV_INTR_DISABLED);
			if (err != HV_EOK)
				printk("sun4v_intr_setenabled(%x): "
				       "err(%d)\n", ino, err);
233 234
		} else {
			u32 tmp;
Linus Torvalds's avatar
Linus Torvalds committed
235

236 237 238 239 240 241 242 243 244
			/* NOTE: We do not want to futz with the IRQ clear registers
			 *       and move the state to IDLE, the SCSI code does call
			 *       disable_irq() to assure atomicity in the queue cmd
			 *       SCSI adapter driver code.  Thus we'd lose interrupts.
			 */
			tmp = upa_readl(imap);
			tmp &= ~IMAP_VALID;
			upa_writel(tmp, imap);
		}
Linus Torvalds's avatar
Linus Torvalds committed
245 246 247
	}
}

248 249 250 251 252 253 254 255 256 257 258
static void build_irq_error(const char *msg, unsigned int ino, int pil, int inofixup,
			    unsigned long iclr, unsigned long imap,
			    struct ino_bucket *bucket)
{
	prom_printf("IRQ: INO %04x (%d:%016lx:%016lx) --> "
		    "(%d:%d:%016lx:%016lx), halting...\n",
		    ino, bucket->pil, bucket->iclr, bucket->imap,
		    pil, inofixup, iclr, imap);
	prom_halt();
}

Linus Torvalds's avatar
Linus Torvalds committed
259 260 261 262 263
unsigned int build_irq(int pil, int inofixup, unsigned long iclr, unsigned long imap)
{
	struct ino_bucket *bucket;
	int ino;

264
	BUG_ON(pil == 0);
265 266
	BUG_ON(tlb_type == hypervisor);

Linus Torvalds's avatar
Linus Torvalds committed
267 268 269 270 271 272 273 274 275 276 277 278 279 280 281
	/* RULE: Both must be specified in all other cases. */
	if (iclr == 0UL || imap == 0UL) {
		prom_printf("Invalid build_irq %d %d %016lx %016lx\n",
			    pil, inofixup, iclr, imap);
		prom_halt();
	}
	
	ino = (upa_readl(imap) & (IMAP_IGN | IMAP_INO)) + inofixup;
	if (ino > NUM_IVECS) {
		prom_printf("Invalid INO %04x (%d:%d:%016lx:%016lx)\n",
			    ino, pil, inofixup, iclr, imap);
		prom_halt();
	}

	bucket = &ivector_table[ino];
282 283 284 285 286 287 288 289 290 291 292 293
	if (bucket->flags & IBF_ACTIVE)
		build_irq_error("IRQ: Trying to build active INO bucket.\n",
				ino, pil, inofixup, iclr, imap, bucket);

	if (bucket->irq_info) {
		if (bucket->imap != imap || bucket->iclr != iclr)
			build_irq_error("IRQ: Trying to reinit INO bucket.\n",
					ino, pil, inofixup, iclr, imap, bucket);

		goto out;
	}

294
	bucket->irq_info = kzalloc(sizeof(struct irq_desc), GFP_ATOMIC);
295 296
	if (!bucket->irq_info) {
		prom_printf("IRQ: Error, kmalloc(irq_desc) failed.\n");
Linus Torvalds's avatar
Linus Torvalds committed
297 298
		prom_halt();
	}
299 300 301 302

	/* Ok, looks good, set it up.  Don't touch the irq_chain or
	 * the pending flag.
	 */
Linus Torvalds's avatar
Linus Torvalds committed
303 304 305 306 307
	bucket->imap  = imap;
	bucket->iclr  = iclr;
	bucket->pil   = pil;
	bucket->flags = 0;

308
out:
Linus Torvalds's avatar
Linus Torvalds committed
309 310 311
	return __irq(bucket);
}

312 313 314 315 316 317 318 319 320 321 322 323
unsigned int sun4v_build_irq(u32 devhandle, unsigned int devino, int pil, unsigned char flags)
{
	struct ino_bucket *bucket;
	unsigned long sysino;

	sysino = sun4v_devino_to_sysino(devhandle, devino);

	bucket = &ivector_table[sysino];

	/* Catch accidental accesses to these things.  IMAP/ICLR handling
	 * is done by hypervisor calls on sun4v platforms, not by direct
	 * register accesses.
324 325 326
	 *
	 * But we need to make them look unique for the disable_irq() logic
	 * in free_irq().
327
	 */
328 329
	bucket->imap = ~0UL - sysino;
	bucket->iclr = ~0UL - sysino;
330 331 332 333

	bucket->pil = pil;
	bucket->flags = flags;

334
	bucket->irq_info = kzalloc(sizeof(struct irq_desc), GFP_ATOMIC);
335 336 337 338 339 340 341 342
	if (!bucket->irq_info) {
		prom_printf("IRQ: Error, kmalloc(irq_desc) failed.\n");
		prom_halt();
	}

	return __irq(bucket);
}

Linus Torvalds's avatar
Linus Torvalds committed
343 344 345 346 347 348 349 350
static void atomic_bucket_insert(struct ino_bucket *bucket)
{
	unsigned long pstate;
	unsigned int *ent;

	__asm__ __volatile__("rdpr %%pstate, %0" : "=r" (pstate));
	__asm__ __volatile__("wrpr %0, %1, %%pstate"
			     : : "r" (pstate), "i" (PSTATE_IE));
351
	ent = irq_work(smp_processor_id());
Linus Torvalds's avatar
Linus Torvalds committed
352 353 354 355 356
	bucket->irq_chain = *ent;
	*ent = __irq(bucket);
	__asm__ __volatile__("wrpr %0, 0x0, %%pstate" : : "r" (pstate));
}

357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402
static int check_irq_sharing(int pil, unsigned long irqflags)
{
	struct irqaction *action, *tmp;

	action = *(irq_action + pil);
	if (action) {
		if ((action->flags & SA_SHIRQ) && (irqflags & SA_SHIRQ)) {
			for (tmp = action; tmp->next; tmp = tmp->next)
				;
		} else {
			return -EBUSY;
		}
	}
	return 0;
}

static void append_irq_action(int pil, struct irqaction *action)
{
	struct irqaction **pp = irq_action + pil;

	while (*pp)
		pp = &((*pp)->next);
	*pp = action;
}

static struct irqaction *get_action_slot(struct ino_bucket *bucket)
{
	struct irq_desc *desc = bucket->irq_info;
	int max_irq, i;

	max_irq = 1;
	if (bucket->flags & IBF_PCI)
		max_irq = MAX_IRQ_DESC_ACTION;
	for (i = 0; i < max_irq; i++) {
		struct irqaction *p = &desc->action[i];
		u32 mask = (1 << i);

		if (desc->action_active_mask & mask)
			continue;

		desc->action_active_mask |= mask;
		return p;
	}
	return NULL;
}

Linus Torvalds's avatar
Linus Torvalds committed
403 404 405
int request_irq(unsigned int irq, irqreturn_t (*handler)(int, void *, struct pt_regs *),
		unsigned long irqflags, const char *name, void *dev_id)
{
406
	struct irqaction *action;
Linus Torvalds's avatar
Linus Torvalds committed
407 408 409 410
	struct ino_bucket *bucket = __bucket(irq);
	unsigned long flags;
	int pending = 0;

411
	if (unlikely(!handler))
Linus Torvalds's avatar
Linus Torvalds committed
412
		return -EINVAL;
413 414 415

	if (unlikely(!bucket->irq_info))
		return -ENODEV;
Linus Torvalds's avatar
Linus Torvalds committed
416

417
	if (irqflags & SA_SAMPLE_RANDOM) {
Linus Torvalds's avatar
Linus Torvalds committed
418 419 420 421 422 423 424 425 426 427 428 429 430 431 432
		/*
	 	 * This function might sleep, we want to call it first,
	 	 * outside of the atomic block. In SA_STATIC_ALLOC case,
		 * random driver's kmalloc will fail, but it is safe.
		 * If already initialized, random driver will not reinit.
	 	 * Yes, this might clear the entropy pool if the wrong
	 	 * driver is attempted to be loaded, without actually
	 	 * installing a new handler, but is this really a problem,
	 	 * only the sysadmin is able to do this.
	 	 */
		rand_initialize_irq(irq);
	}

	spin_lock_irqsave(&irq_action_lock, flags);

433 434 435
	if (check_irq_sharing(bucket->pil, irqflags)) {
		spin_unlock_irqrestore(&irq_action_lock, flags);
		return -EBUSY;
Linus Torvalds's avatar
Linus Torvalds committed
436 437
	}

438
	action = get_action_slot(bucket);
Linus Torvalds's avatar
Linus Torvalds committed
439 440 441 442 443
	if (!action) { 
		spin_unlock_irqrestore(&irq_action_lock, flags);
		return -ENOMEM;
	}

444
	bucket->flags |= IBF_ACTIVE;
445 446 447
	pending = bucket->pending;
	if (pending)
		bucket->pending = 0;
Linus Torvalds's avatar
Linus Torvalds committed
448 449 450 451 452 453 454 455 456

	action->handler = handler;
	action->flags = irqflags;
	action->name = name;
	action->next = NULL;
	action->dev_id = dev_id;
	put_ino_in_irqaction(action, irq);
	put_smpaff_in_irqaction(action, CPU_MASK_NONE);

457
	append_irq_action(bucket->pil, action);
Linus Torvalds's avatar
Linus Torvalds committed
458 459 460 461 462 463

	enable_irq(irq);

	/* We ate the IVEC already, this makes sure it does not get lost. */
	if (pending) {
		atomic_bucket_insert(bucket);
464
		set_softint(1 << PIL_DEVICE_IRQ);
Linus Torvalds's avatar
Linus Torvalds committed
465
	}
466

Linus Torvalds's avatar
Linus Torvalds committed
467
	spin_unlock_irqrestore(&irq_action_lock, flags);
468

469
	register_irq_proc(__irq_ino(irq));
Linus Torvalds's avatar
Linus Torvalds committed
470 471 472 473 474 475 476 477 478

#ifdef CONFIG_SMP
	distribute_irqs();
#endif
	return 0;
}

EXPORT_SYMBOL(request_irq);

479
static struct irqaction *unlink_irq_action(unsigned int irq, void *dev_id)
Linus Torvalds's avatar
Linus Torvalds committed
480
{
481 482
	struct ino_bucket *bucket = __bucket(irq);
	struct irqaction *action, **pp;
Linus Torvalds's avatar
Linus Torvalds committed
483

484 485 486 487
	pp = irq_action + bucket->pil;
	action = *pp;
	if (unlikely(!action))
		return NULL;
Linus Torvalds's avatar
Linus Torvalds committed
488

489
	if (unlikely(!action->handler)) {
Linus Torvalds's avatar
Linus Torvalds committed
490
		printk("Freeing free IRQ %d\n", bucket->pil);
491
		return NULL;
Linus Torvalds's avatar
Linus Torvalds committed
492 493
	}

494 495 496
	while (action && action->dev_id != dev_id) {
		pp = &action->next;
		action = *pp;
Linus Torvalds's avatar
Linus Torvalds committed
497 498
	}

499 500 501 502 503 504 505 506 507 508
	if (likely(action))
		*pp = action->next;

	return action;
}

void free_irq(unsigned int irq, void *dev_id)
{
	struct irqaction *action;
	struct ino_bucket *bucket;
509
	struct irq_desc *desc;
510
	unsigned long flags;
511
	int ent, i;
512 513 514 515

	spin_lock_irqsave(&irq_action_lock, flags);

	action = unlink_irq_action(irq, dev_id);
Linus Torvalds's avatar
Linus Torvalds committed
516 517 518

	spin_unlock_irqrestore(&irq_action_lock, flags);

519 520 521
	if (unlikely(!action))
		return;

Linus Torvalds's avatar
Linus Torvalds committed
522 523 524 525
	synchronize_irq(irq);

	spin_lock_irqsave(&irq_action_lock, flags);

526
	bucket = __bucket(irq);
527
	desc = bucket->irq_info;
Linus Torvalds's avatar
Linus Torvalds committed
528

529 530
	for (i = 0; i < MAX_IRQ_DESC_ACTION; i++) {
		struct irqaction *p = &desc->action[i];
531

532 533 534
		if (p == action) {
			desc->action_active_mask &= ~(1 << i);
			break;
Linus Torvalds's avatar
Linus Torvalds committed
535
		}
536
	}
Linus Torvalds's avatar
Linus Torvalds committed
537

538 539
	if (!desc->action_active_mask) {
		unsigned long imap = bucket->imap;
Linus Torvalds's avatar
Linus Torvalds committed
540

541 542
		/* This unique interrupt source is now inactive. */
		bucket->flags &= ~IBF_ACTIVE;
Linus Torvalds's avatar
Linus Torvalds committed
543

544 545 546 547 548 549 550 551 552
		/* See if any other buckets share this bucket's IMAP
		 * and are still active.
		 */
		for (ent = 0; ent < NUM_IVECS; ent++) {
			struct ino_bucket *bp = &ivector_table[ent];
			if (bp != bucket	&&
			    bp->imap == imap	&&
			    (bp->flags & IBF_ACTIVE) != 0)
				break;
553
		}
554 555 556 557 558 559

		/* Only disable when no other sub-irq levels of
		 * the same IMAP are active.
		 */
		if (ent == NUM_IVECS)
			disable_irq(irq);
Linus Torvalds's avatar
Linus Torvalds committed
560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599
	}

	spin_unlock_irqrestore(&irq_action_lock, flags);
}

EXPORT_SYMBOL(free_irq);

#ifdef CONFIG_SMP
void synchronize_irq(unsigned int irq)
{
	struct ino_bucket *bucket = __bucket(irq);

#if 0
	/* The following is how I wish I could implement this.
	 * Unfortunately the ICLR registers are read-only, you can
	 * only write ICLR_foo values to them.  To get the current
	 * IRQ status you would need to get at the IRQ diag registers
	 * in the PCI/SBUS controller and the layout of those vary
	 * from one controller to the next, sigh... -DaveM
	 */
	unsigned long iclr = bucket->iclr;

	while (1) {
		u32 tmp = upa_readl(iclr);
		
		if (tmp == ICLR_TRANSMIT ||
		    tmp == ICLR_PENDING) {
			cpu_relax();
			continue;
		}
		break;
	}
#else
	/* So we have to do this with a INPROGRESS bit just like x86.  */
	while (bucket->flags & IBF_INPROGRESS)
		cpu_relax();
#endif
}
#endif /* CONFIG_SMP */

600
static void process_bucket(struct ino_bucket *bp, struct pt_regs *regs)
Linus Torvalds's avatar
Linus Torvalds committed
601
{
602 603 604 605
	struct irq_desc *desc = bp->irq_info;
	unsigned char flags = bp->flags;
	u32 action_mask, i;
	int random;
Linus Torvalds's avatar
Linus Torvalds committed
606

607
	bp->flags |= IBF_INPROGRESS;
Linus Torvalds's avatar
Linus Torvalds committed
608

609 610
	if (unlikely(!(flags & IBF_ACTIVE))) {
		bp->pending = 1;
Linus Torvalds's avatar
Linus Torvalds committed
611 612 613
		goto out;
	}

614 615 616 617
	if (desc->pre_handler)
		desc->pre_handler(bp,
				  desc->pre_handler_arg1,
				  desc->pre_handler_arg2);
Linus Torvalds's avatar
Linus Torvalds committed
618

619 620 621 622 623
	action_mask = desc->action_active_mask;
	random = 0;
	for (i = 0; i < MAX_IRQ_DESC_ACTION; i++) {
		struct irqaction *p = &desc->action[i];
		u32 mask = (1 << i);
Linus Torvalds's avatar
Linus Torvalds committed
624

625 626
		if (!(action_mask & mask))
			continue;
Linus Torvalds's avatar
Linus Torvalds committed
627

628
		action_mask &= ~mask;
Linus Torvalds's avatar
Linus Torvalds committed
629

630 631 632 633 634 635 636
		if (p->handler(__irq(bp), p->dev_id, regs) == IRQ_HANDLED)
			random |= p->flags;

		if (!action_mask)
			break;
	}
	if (bp->pil != 0) {
637
		if (tlb_type == hypervisor) {
638
			unsigned int ino = __irq_ino(bp);
639
			int err;
640

641 642 643 644
			err = sun4v_intr_setstate(ino, HV_INTR_STATE_IDLE);
			if (err != HV_EOK)
				printk("sun4v_intr_setstate(%x): "
				       "err(%d)\n", ino, err);
645 646 647
		} else {
			upa_writel(ICLR_IDLE, bp->iclr);
		}
648 649 650

		/* Test and add entropy */
		if (random & SA_SAMPLE_RANDOM)
651
			add_interrupt_randomness(bp->pil);
652
	}
Linus Torvalds's avatar
Linus Torvalds committed
653
out:
654
	bp->flags &= ~IBF_INPROGRESS;
Linus Torvalds's avatar
Linus Torvalds committed
655 656
}

657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677
#ifndef CONFIG_SMP
extern irqreturn_t timer_interrupt(int, void *, struct pt_regs *);

void timer_irq(int irq, struct pt_regs *regs)
{
	unsigned long clr_mask = 1 << irq;
	unsigned long tick_mask = tick_ops->softint_mask;

	if (get_softint() & tick_mask) {
		irq = 0;
		clr_mask = tick_mask;
	}
	clear_softint(clr_mask);

	irq_enter();
	kstat_this_cpu.irqs[irq]++;
	timer_interrupt(irq, NULL, regs);
	irq_exit();
}
#endif

Linus Torvalds's avatar
Linus Torvalds committed
678 679
void handler_irq(int irq, struct pt_regs *regs)
{
680
	struct ino_bucket *bp;
Linus Torvalds's avatar
Linus Torvalds committed
681 682
	int cpu = smp_processor_id();

683 684
	/* XXX at this point we should be able to assert that
	 * XXX irq is PIL_DEVICE_IRQ...
Linus Torvalds's avatar
Linus Torvalds committed
685 686 687 688 689 690
	 */
	clear_softint(1 << irq);

	irq_enter();

	/* Sliiiick... */
691
	bp = __bucket(xchg32(irq_work(cpu), 0));
692 693
	while (bp) {
		struct ino_bucket *nbp = __bucket(bp->irq_chain);
Linus Torvalds's avatar
Linus Torvalds committed
694

695 696
		kstat_this_cpu.irqs[bp->pil]++;

Linus Torvalds's avatar
Linus Torvalds committed
697
		bp->irq_chain = 0;
698
		process_bucket(bp, regs);
699
		bp = nbp;
Linus Torvalds's avatar
Linus Torvalds committed
700 701 702 703 704
	}
	irq_exit();
}

#ifdef CONFIG_BLK_DEV_FD
Alexey Dobriyan's avatar
Alexey Dobriyan committed
705
extern irqreturn_t floppy_interrupt(int, void *, struct pt_regs *);
Linus Torvalds's avatar
Linus Torvalds committed
706

707 708 709 710 711
/* XXX No easy way to include asm/floppy.h XXX */
extern unsigned char *pdma_vaddr;
extern unsigned long pdma_size;
extern volatile int doing_pdma;
extern unsigned long fdc_status;
Linus Torvalds's avatar
Linus Torvalds committed
712

713
irqreturn_t sparc_floppy_irq(int irq, void *dev_cookie, struct pt_regs *regs)
Linus Torvalds's avatar
Linus Torvalds committed
714
{
715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738
	if (likely(doing_pdma)) {
		void __iomem *stat = (void __iomem *) fdc_status;
		unsigned char *vaddr = pdma_vaddr;
		unsigned long size = pdma_size;
		u8 val;

		while (size) {
			val = readb(stat);
			if (unlikely(!(val & 0x80))) {
				pdma_vaddr = vaddr;
				pdma_size = size;
				return IRQ_HANDLED;
			}
			if (unlikely(!(val & 0x20))) {
				pdma_vaddr = vaddr;
				pdma_size = size;
				doing_pdma = 0;
				goto main_interrupt;
			}
			if (val & 0x40) {
				/* read */
				*vaddr++ = readb(stat + 1);
			} else {
				unsigned char data = *vaddr++;
Linus Torvalds's avatar
Linus Torvalds committed
739

740 741 742 743 744
				/* write */
				writeb(data, stat + 1);
			}
			size--;
		}
Linus Torvalds's avatar
Linus Torvalds committed
745

746 747
		pdma_vaddr = vaddr;
		pdma_size = size;
Linus Torvalds's avatar
Linus Torvalds committed
748

749 750 751 752
		/* Send Terminal Count pulse to floppy controller. */
		val = readb(auxio_register);
		val |= AUXIO_AUX1_FTCNT;
		writeb(val, auxio_register);
753
		val &= ~AUXIO_AUX1_FTCNT;
754
		writeb(val, auxio_register);
Linus Torvalds's avatar
Linus Torvalds committed
755

756
		doing_pdma = 0;
Linus Torvalds's avatar
Linus Torvalds committed
757 758
	}

759 760
main_interrupt:
	return floppy_interrupt(irq, dev_cookie, regs);
Linus Torvalds's avatar
Linus Torvalds committed
761
}
762 763
EXPORT_SYMBOL(sparc_floppy_irq);
#endif
Linus Torvalds's avatar
Linus Torvalds committed
764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791

/* We really don't need these at all on the Sparc.  We only have
 * stubs here because they are exported to modules.
 */
unsigned long probe_irq_on(void)
{
	return 0;
}

EXPORT_SYMBOL(probe_irq_on);

int probe_irq_off(unsigned long mask)
{
	return 0;
}

EXPORT_SYMBOL(probe_irq_off);

#ifdef CONFIG_SMP
static int retarget_one_irq(struct irqaction *p, int goal_cpu)
{
	struct ino_bucket *bucket = get_ino_in_irqaction(p) + ivector_table;

	while (!cpu_online(goal_cpu)) {
		if (++goal_cpu >= NR_CPUS)
			goal_cpu = 0;
	}

792
	if (tlb_type == hypervisor) {
793
		unsigned int ino = __irq_ino(bucket);
794

795 796
		sun4v_intr_settarget(ino, goal_cpu);
		sun4v_intr_setenabled(ino, HV_INTR_ENABLED);
Linus Torvalds's avatar
Linus Torvalds committed
797
	} else {
798
		unsigned long imap = bucket->imap;
799
		unsigned int tid = sun4u_compute_tid(imap, goal_cpu);
800 801

		upa_writel(tid | IMAP_VALID, imap);
Linus Torvalds's avatar
Linus Torvalds committed
802 803
	}

804
	do {
Linus Torvalds's avatar
Linus Torvalds committed
805 806
		if (++goal_cpu >= NR_CPUS)
			goal_cpu = 0;
807
	} while (!cpu_online(goal_cpu));
Linus Torvalds's avatar
Linus Torvalds committed
808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826

	return goal_cpu;
}

/* Called from request_irq. */
static void distribute_irqs(void)
{
	unsigned long flags;
	int cpu, level;

	spin_lock_irqsave(&irq_action_lock, flags);
	cpu = 0;

	/*
	 * Skip the timer at [0], and very rare error/power intrs at [15].
	 * Also level [12], it causes problems on Ex000 systems.
	 */
	for (level = 1; level < NR_IRQS; level++) {
		struct irqaction *p = irq_action[level];
827 828 829 830

		if (level == 12)
			continue;

Linus Torvalds's avatar
Linus Torvalds committed
831 832 833 834 835 836 837 838 839
		while(p) {
			cpu = retarget_one_irq(p, cpu);
			p = p->next;
		}
	}
	spin_unlock_irqrestore(&irq_action_lock, flags);
}
#endif

840 841 842 843 844 845
struct sun5_timer {
	u64	count0;
	u64	limit0;
	u64	count1;
	u64	limit1;
};
Linus Torvalds's avatar
Linus Torvalds committed
846

847
static struct sun5_timer *prom_timers;
Linus Torvalds's avatar
Linus Torvalds committed
848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906
static u64 prom_limit0, prom_limit1;

static void map_prom_timers(void)
{
	unsigned int addr[3];
	int tnode, err;

	/* PROM timer node hangs out in the top level of device siblings... */
	tnode = prom_finddevice("/counter-timer");

	/* Assume if node is not present, PROM uses different tick mechanism
	 * which we should not care about.
	 */
	if (tnode == 0 || tnode == -1) {
		prom_timers = (struct sun5_timer *) 0;
		return;
	}

	/* If PROM is really using this, it must be mapped by him. */
	err = prom_getproperty(tnode, "address", (char *)addr, sizeof(addr));
	if (err == -1) {
		prom_printf("PROM does not have timer mapped, trying to continue.\n");
		prom_timers = (struct sun5_timer *) 0;
		return;
	}
	prom_timers = (struct sun5_timer *) ((unsigned long)addr[0]);
}

static void kill_prom_timer(void)
{
	if (!prom_timers)
		return;

	/* Save them away for later. */
	prom_limit0 = prom_timers->limit0;
	prom_limit1 = prom_timers->limit1;

	/* Just as in sun4c/sun4m PROM uses timer which ticks at IRQ 14.
	 * We turn both off here just to be paranoid.
	 */
	prom_timers->limit0 = 0;
	prom_timers->limit1 = 0;

	/* Wheee, eat the interrupt packet too... */
	__asm__ __volatile__(
"	mov	0x40, %%g2\n"
"	ldxa	[%%g0] %0, %%g1\n"
"	ldxa	[%%g2] %1, %%g1\n"
"	stxa	%%g0, [%%g0] %0\n"
"	membar	#Sync\n"
	: /* no outputs */
	: "i" (ASI_INTR_RECEIVE), "i" (ASI_INTR_R)
	: "g1", "g2");
}

void init_irqwork_curcpu(void)
{
	int cpu = hard_smp_processor_id();

907
	trap_block[cpu].irq_worklist = 0;
Linus Torvalds's avatar
Linus Torvalds committed
908 909
}

910
static void __cpuinit register_one_mondo(unsigned long paddr, unsigned long type)
911
{
912 913 914 915 916 917 918
	unsigned long num_entries = 128;
	unsigned long status;

	status = sun4v_cpu_qconf(type, paddr, num_entries);
	if (status != HV_EOK) {
		prom_printf("SUN4V: sun4v_cpu_qconf(%lu:%lx:%lu) failed, "
			    "err %lu\n", type, paddr, num_entries, status);
919 920 921 922
		prom_halt();
	}
}

923
static void __cpuinit sun4v_register_mondo_queues(int this_cpu)
924
{
925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951 952 953 954 955 956 957
	struct trap_per_cpu *tb = &trap_block[this_cpu];

	register_one_mondo(tb->cpu_mondo_pa, HV_CPU_QUEUE_CPU_MONDO);
	register_one_mondo(tb->dev_mondo_pa, HV_CPU_QUEUE_DEVICE_MONDO);
	register_one_mondo(tb->resum_mondo_pa, HV_CPU_QUEUE_RES_ERROR);
	register_one_mondo(tb->nonresum_mondo_pa, HV_CPU_QUEUE_NONRES_ERROR);
}

static void __cpuinit alloc_one_mondo(unsigned long *pa_ptr, int use_bootmem)
{
	void *page;

	if (use_bootmem)
		page = alloc_bootmem_low_pages(PAGE_SIZE);
	else
		page = (void *) get_zeroed_page(GFP_ATOMIC);

	if (!page) {
		prom_printf("SUN4V: Error, cannot allocate mondo queue.\n");
		prom_halt();
	}

	*pa_ptr = __pa(page);
}

static void __cpuinit alloc_one_kbuf(unsigned long *pa_ptr, int use_bootmem)
{
	void *page;

	if (use_bootmem)
		page = alloc_bootmem_low_pages(PAGE_SIZE);
	else
		page = (void *) get_zeroed_page(GFP_ATOMIC);
958 959 960 961 962 963 964 965 966

	if (!page) {
		prom_printf("SUN4V: Error, cannot allocate kbuf page.\n");
		prom_halt();
	}

	*pa_ptr = __pa(page);
}

967
static void __cpuinit init_cpu_send_mondo_info(struct trap_per_cpu *tb, int use_bootmem)
968 969
{
#ifdef CONFIG_SMP
970
	void *page;
971 972 973

	BUILD_BUG_ON((NR_CPUS * sizeof(u16)) > (PAGE_SIZE - 64));

974 975 976 977 978
	if (use_bootmem)
		page = alloc_bootmem_low_pages(PAGE_SIZE);
	else
		page = (void *) get_zeroed_page(GFP_ATOMIC);

979 980 981 982 983 984 985 986 987 988
	if (!page) {
		prom_printf("SUN4V: Error, cannot allocate cpu mondo page.\n");
		prom_halt();
	}

	tb->cpu_mondo_block_pa = __pa(page);
	tb->cpu_list_pa = __pa(page + 64);
#endif
}

989
/* Allocate and register the mondo and error queues for this cpu.  */
990
void __cpuinit sun4v_init_mondo_queues(int use_bootmem, int cpu, int alloc, int load)
991 992 993
{
	struct trap_per_cpu *tb = &trap_block[cpu];

994 995 996 997 998 999 1000
	if (alloc) {
		alloc_one_mondo(&tb->cpu_mondo_pa, use_bootmem);
		alloc_one_mondo(&tb->dev_mondo_pa, use_bootmem);
		alloc_one_mondo(&tb->resum_mondo_pa, use_bootmem);
		alloc_one_kbuf(&tb->resum_kernel_buf_pa, use_bootmem);
		alloc_one_mondo(&tb->nonresum_mondo_pa, use_bootmem);
		alloc_one_kbuf(&tb->nonresum_kernel_buf_pa, use_bootmem);
1001

1002 1003
		init_cpu_send_mondo_info(tb, use_bootmem);
	}
1004

1005 1006 1007 1008 1009 1010 1011 1012
	if (load) {
		if (cpu != hard_smp_processor_id()) {
			prom_printf("SUN4V: init mondo on cpu %d not %d\n",
				    cpu, hard_smp_processor_id());
			prom_halt();
		}
		sun4v_register_mondo_queues(cpu);
	}
1013 1014
}

Linus Torvalds's avatar
Linus Torvalds committed
1015 1016 1017 1018 1019 1020 1021
/* Only invoked on boot processor. */
void __init init_IRQ(void)
{
	map_prom_timers();
	kill_prom_timer();
	memset(&ivector_table[0], 0, sizeof(ivector_table));

1022
	if (tlb_type == hypervisor)
1023
		sun4v_init_mondo_queues(1, hard_smp_processor_id(), 1, 1);
1024

Linus Torvalds's avatar
Linus Torvalds committed
1025 1026 1027 1028 1029 1030 1031 1032 1033 1034 1035 1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052
	/* We need to clear any IRQ's pending in the soft interrupt
	 * registers, a spurious one could be left around from the
	 * PROM timer which we just disabled.
	 */
	clear_softint(get_softint());

	/* Now that ivector table is initialized, it is safe
	 * to receive IRQ vector traps.  We will normally take
	 * one or two right now, in case some device PROM used
	 * to boot us wants to speak to us.  We just ignore them.
	 */
	__asm__ __volatile__("rdpr	%%pstate, %%g1\n\t"
			     "or	%%g1, %0, %%g1\n\t"
			     "wrpr	%%g1, 0x0, %%pstate"
			     : /* No outputs */
			     : "i" (PSTATE_IE)
			     : "g1");
}

static struct proc_dir_entry * root_irq_dir;
static struct proc_dir_entry * irq_dir [NUM_IVECS];

#ifdef CONFIG_SMP

static int irq_affinity_read_proc (char *page, char **start, off_t off,
			int count, int *eof, void *data)
{
	struct ino_bucket *bp = ivector_table + (long)data;
1053 1054
	struct irq_desc *desc = bp->irq_info;
	struct irqaction *ap = desc->action;
Linus Torvalds's avatar
Linus Torvalds committed
1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067 1068 1069 1070 1071
	cpumask_t mask;
	int len;

	mask = get_smpaff_in_irqaction(ap);
	if (cpus_empty(mask))
		mask = cpu_online_map;

	len = cpumask_scnprintf(page, count, mask);
	if (count - len < 2)
		return -EINVAL;
	len += sprintf(page + len, "\n");
	return len;
}

static inline void set_intr_affinity(int irq, cpumask_t hw_aff)
{
	struct ino_bucket *bp = ivector_table + irq;
1072 1073
	struct irq_desc *desc = bp->irq_info;
	struct irqaction *ap = desc->action;
Linus Torvalds's avatar
Linus Torvalds committed
1074 1075 1076 1077

	/* Users specify affinity in terms of hw cpu ids.
	 * As soon as we do this, handler_irq() might see and take action.
	 */
1078
	put_smpaff_in_irqaction(ap, hw_aff);
Linus Torvalds's avatar
Linus Torvalds committed
1079 1080 1081 1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 1102 1103 1104 1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119 1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135 1136 1137 1138 1139 1140 1141 1142 1143 1144 1145 1146 1147

	/* Migration is simply done by the next cpu to service this
	 * interrupt.
	 */
}

static int irq_affinity_write_proc (struct file *file, const char __user *buffer,
					unsigned long count, void *data)
{
	int irq = (long) data, full_count = count, err;
	cpumask_t new_value;

	err = cpumask_parse(buffer, count, new_value);

	/*
	 * Do not allow disabling IRQs completely - it's a too easy
	 * way to make the system unusable accidentally :-) At least
	 * one online CPU still has to be targeted.
	 */
	cpus_and(new_value, new_value, cpu_online_map);
	if (cpus_empty(new_value))
		return -EINVAL;

	set_intr_affinity(irq, new_value);

	return full_count;
}

#endif

#define MAX_NAMELEN 10

static void register_irq_proc (unsigned int irq)
{
	char name [MAX_NAMELEN];

	if (!root_irq_dir || irq_dir[irq])
		return;

	memset(name, 0, MAX_NAMELEN);
	sprintf(name, "%x", irq);

	/* create /proc/irq/1234 */
	irq_dir[irq] = proc_mkdir(name, root_irq_dir);

#ifdef CONFIG_SMP
	/* XXX SMP affinity not supported on starfire yet. */
	if (this_is_starfire == 0) {
		struct proc_dir_entry *entry;

		/* create /proc/irq/1234/smp_affinity */
		entry = create_proc_entry("smp_affinity", 0600, irq_dir[irq]);

		if (entry) {
			entry->nlink = 1;
			entry->data = (void *)(long)irq;
			entry->read_proc = irq_affinity_read_proc;
			entry->write_proc = irq_affinity_write_proc;
		}
	}
#endif
}

void init_irq_proc (void)
{
	/* create /proc/irq */
	root_irq_dir = proc_mkdir("irq", NULL);
}