Commit b5a37e96 authored by David S. Miller's avatar David S. Miller

[SPARC64]: Fix mondo queue allocations.

We have to use bootmem during init_IRQ and page alloc
for sibling cpu calls.

Also, fix incorrect hypervisor call return value
checks in the hypervisor SMP cpu mondo send code.
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent c4bce90e
...@@ -21,6 +21,7 @@ ...@@ -21,6 +21,7 @@
#include <linux/delay.h> #include <linux/delay.h>
#include <linux/proc_fs.h> #include <linux/proc_fs.h>
#include <linux/seq_file.h> #include <linux/seq_file.h>
#include <linux/bootmem.h>
#include <asm/ptrace.h> #include <asm/ptrace.h>
#include <asm/processor.h> #include <asm/processor.h>
...@@ -861,24 +862,16 @@ void init_irqwork_curcpu(void) ...@@ -861,24 +862,16 @@ void init_irqwork_curcpu(void)
memset(__irq_work + cpu, 0, sizeof(struct irq_work_struct)); memset(__irq_work + cpu, 0, sizeof(struct irq_work_struct));
} }
static void __cpuinit init_one_mondo(unsigned long *pa_ptr, unsigned long type) static void __cpuinit register_one_mondo(unsigned long paddr, unsigned long type)
{ {
register unsigned long func __asm__("%o5"); register unsigned long func __asm__("%o5");
register unsigned long arg0 __asm__("%o0"); register unsigned long arg0 __asm__("%o0");
register unsigned long arg1 __asm__("%o1"); register unsigned long arg1 __asm__("%o1");
register unsigned long arg2 __asm__("%o2"); register unsigned long arg2 __asm__("%o2");
unsigned long page = get_zeroed_page(GFP_ATOMIC);
if (!page) {
prom_printf("SUN4V: Error, cannot allocate mondo queue.\n");
prom_halt();
}
*pa_ptr = __pa(page);
func = HV_FAST_CPU_QCONF; func = HV_FAST_CPU_QCONF;
arg0 = type; arg0 = type;
arg1 = *pa_ptr; arg1 = paddr;
arg2 = 128; /* XXX Implied by Niagara queue offsets. XXX */ arg2 = 128; /* XXX Implied by Niagara queue offsets. XXX */
__asm__ __volatile__("ta %8" __asm__ __volatile__("ta %8"
: "=&r" (func), "=&r" (arg0), : "=&r" (func), "=&r" (arg0),
...@@ -887,16 +880,48 @@ static void __cpuinit init_one_mondo(unsigned long *pa_ptr, unsigned long type) ...@@ -887,16 +880,48 @@ static void __cpuinit init_one_mondo(unsigned long *pa_ptr, unsigned long type)
"2" (arg1), "3" (arg2), "2" (arg1), "3" (arg2),
"i" (HV_FAST_TRAP)); "i" (HV_FAST_TRAP));
if (func != HV_EOK) { if (arg0 != HV_EOK) {
prom_printf("SUN4V: cpu_qconf(%lu) failed with error %lu\n", prom_printf("SUN4V: cpu_qconf(%lu) failed with error %lu\n",
type, func); type, func);
prom_halt(); prom_halt();
} }
} }
static void __cpuinit init_one_kbuf(unsigned long *pa_ptr) static void __cpuinit sun4v_register_mondo_queues(int this_cpu)
{ {
unsigned long page = get_zeroed_page(GFP_ATOMIC); struct trap_per_cpu *tb = &trap_block[this_cpu];
register_one_mondo(tb->cpu_mondo_pa, HV_CPU_QUEUE_CPU_MONDO);
register_one_mondo(tb->dev_mondo_pa, HV_CPU_QUEUE_DEVICE_MONDO);
register_one_mondo(tb->resum_mondo_pa, HV_CPU_QUEUE_RES_ERROR);
register_one_mondo(tb->nonresum_mondo_pa, HV_CPU_QUEUE_NONRES_ERROR);
}
static void __cpuinit alloc_one_mondo(unsigned long *pa_ptr, int use_bootmem)
{
void *page;
if (use_bootmem)
page = alloc_bootmem_low_pages(PAGE_SIZE);
else
page = (void *) get_zeroed_page(GFP_ATOMIC);
if (!page) {
prom_printf("SUN4V: Error, cannot allocate mondo queue.\n");
prom_halt();
}
*pa_ptr = __pa(page);
}
static void __cpuinit alloc_one_kbuf(unsigned long *pa_ptr, int use_bootmem)
{
void *page;
if (use_bootmem)
page = alloc_bootmem_low_pages(PAGE_SIZE);
else
page = (void *) get_zeroed_page(GFP_ATOMIC);
if (!page) { if (!page) {
prom_printf("SUN4V: Error, cannot allocate kbuf page.\n"); prom_printf("SUN4V: Error, cannot allocate kbuf page.\n");
...@@ -906,14 +931,18 @@ static void __cpuinit init_one_kbuf(unsigned long *pa_ptr) ...@@ -906,14 +931,18 @@ static void __cpuinit init_one_kbuf(unsigned long *pa_ptr)
*pa_ptr = __pa(page); *pa_ptr = __pa(page);
} }
static void __cpuinit init_cpu_send_mondo_info(struct trap_per_cpu *tb) static void __cpuinit init_cpu_send_mondo_info(struct trap_per_cpu *tb, int use_bootmem)
{ {
#ifdef CONFIG_SMP #ifdef CONFIG_SMP
unsigned long page; void *page;
BUILD_BUG_ON((NR_CPUS * sizeof(u16)) > (PAGE_SIZE - 64)); BUILD_BUG_ON((NR_CPUS * sizeof(u16)) > (PAGE_SIZE - 64));
page = get_zeroed_page(GFP_ATOMIC); if (use_bootmem)
page = alloc_bootmem_low_pages(PAGE_SIZE);
else
page = (void *) get_zeroed_page(GFP_ATOMIC);
if (!page) { if (!page) {
prom_printf("SUN4V: Error, cannot allocate cpu mondo page.\n"); prom_printf("SUN4V: Error, cannot allocate cpu mondo page.\n");
prom_halt(); prom_halt();
...@@ -924,22 +953,22 @@ static void __cpuinit init_cpu_send_mondo_info(struct trap_per_cpu *tb) ...@@ -924,22 +953,22 @@ static void __cpuinit init_cpu_send_mondo_info(struct trap_per_cpu *tb)
#endif #endif
} }
/* Allocate and init the mondo and error queues for this cpu. */ /* Allocate and register the mondo and error queues for this cpu. */
void __cpuinit sun4v_init_mondo_queues(void) void __cpuinit sun4v_init_mondo_queues(int use_bootmem)
{ {
int cpu = hard_smp_processor_id(); int cpu = hard_smp_processor_id();
struct trap_per_cpu *tb = &trap_block[cpu]; struct trap_per_cpu *tb = &trap_block[cpu];
init_one_mondo(&tb->cpu_mondo_pa, HV_CPU_QUEUE_CPU_MONDO); alloc_one_mondo(&tb->cpu_mondo_pa, use_bootmem);
init_one_mondo(&tb->dev_mondo_pa, HV_CPU_QUEUE_DEVICE_MONDO); alloc_one_mondo(&tb->dev_mondo_pa, use_bootmem);
alloc_one_mondo(&tb->resum_mondo_pa, use_bootmem);
init_one_mondo(&tb->resum_mondo_pa, HV_CPU_QUEUE_RES_ERROR); alloc_one_kbuf(&tb->resum_kernel_buf_pa, use_bootmem);
init_one_kbuf(&tb->resum_kernel_buf_pa); alloc_one_mondo(&tb->nonresum_mondo_pa, use_bootmem);
alloc_one_kbuf(&tb->nonresum_kernel_buf_pa, use_bootmem);
init_one_mondo(&tb->nonresum_mondo_pa, HV_CPU_QUEUE_NONRES_ERROR); init_cpu_send_mondo_info(tb, use_bootmem);
init_one_kbuf(&tb->nonresum_kernel_buf_pa);
init_cpu_send_mondo_info(tb); sun4v_register_mondo_queues(cpu);
} }
/* Only invoked on boot processor. */ /* Only invoked on boot processor. */
...@@ -950,7 +979,7 @@ void __init init_IRQ(void) ...@@ -950,7 +979,7 @@ void __init init_IRQ(void)
memset(&ivector_table[0], 0, sizeof(ivector_table)); memset(&ivector_table[0], 0, sizeof(ivector_table));
if (tlb_type == hypervisor) if (tlb_type == hypervisor)
sun4v_init_mondo_queues(); sun4v_init_mondo_queues(1);
/* We need to clear any IRQ's pending in the soft interrupt /* We need to clear any IRQ's pending in the soft interrupt
* registers, a spurious one could be left around from the * registers, a spurious one could be left around from the
......
...@@ -591,7 +591,7 @@ static void hypervisor_xcall_deliver(u64 data0, u64 data1, u64 data2, cpumask_t ...@@ -591,7 +591,7 @@ static void hypervisor_xcall_deliver(u64 data0, u64 data1, u64 data2, cpumask_t
"2" (arg1), "3" (arg2), "2" (arg1), "3" (arg2),
"i" (HV_FAST_TRAP) "i" (HV_FAST_TRAP)
: "memory"); : "memory");
if (likely(func == HV_EOK)) if (likely(arg0 == HV_EOK))
break; break;
if (unlikely(++retries > 100)) { if (unlikely(++retries > 100)) {
...@@ -644,7 +644,7 @@ static void hypervisor_xcall_deliver(u64 data0, u64 data1, u64 data2, cpumask_t ...@@ -644,7 +644,7 @@ static void hypervisor_xcall_deliver(u64 data0, u64 data1, u64 data2, cpumask_t
"2" (arg1), "3" (arg2), "2" (arg1), "3" (arg2),
"i" (HV_FAST_TRAP) "i" (HV_FAST_TRAP)
: "memory"); : "memory");
if (likely(func == HV_EOK)) if (likely(arg0 == HV_EOK))
break; break;
if (unlikely(++retries > 100)) { if (unlikely(++retries > 100)) {
......
...@@ -367,7 +367,7 @@ after_lock_tlb: ...@@ -367,7 +367,7 @@ after_lock_tlb:
nop nop
call sun4v_init_mondo_queues call sun4v_init_mondo_queues
nop mov 0, %o0
1: call init_cur_cpu_trap 1: call init_cur_cpu_trap
nop nop
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment