Commit e4903fb5 authored by Linus Torvalds's avatar Linus Torvalds

Merge branch 'release' of git://git.kernel.org/pub/scm/linux/kernel/git/aegl/linux-2.6

* 'release' of git://git.kernel.org/pub/scm/linux/kernel/git/aegl/linux-2.6:
  [IA64] Nail two more simple section mismatch errors
  [IA64] fix section mismatch warnings
  [IA64] rename partial_page
  [IA64] Ensure that machvec is set up takes place before serial console
  [IA64] vector-domain - fix vector_table
  [IA64] vector-domain - handle assign_irq_vector(AUTO_ASSIGN)
parents 6a28a05f cb2e0912
...@@ -249,10 +249,10 @@ ia32_init (void) ...@@ -249,10 +249,10 @@ ia32_init (void)
#if PAGE_SHIFT > IA32_PAGE_SHIFT #if PAGE_SHIFT > IA32_PAGE_SHIFT
{ {
extern struct kmem_cache *partial_page_cachep; extern struct kmem_cache *ia64_partial_page_cachep;
partial_page_cachep = kmem_cache_create("partial_page_cache", ia64_partial_page_cachep = kmem_cache_create("ia64_partial_page_cache",
sizeof(struct partial_page), sizeof(struct ia64_partial_page),
0, SLAB_PANIC, NULL); 0, SLAB_PANIC, NULL);
} }
#endif #endif
......
...@@ -25,8 +25,8 @@ ...@@ -25,8 +25,8 @@
* partially mapped pages provide precise accounting of which 4k sub pages * partially mapped pages provide precise accounting of which 4k sub pages
* are mapped and which ones are not, thereby improving IA-32 compatibility. * are mapped and which ones are not, thereby improving IA-32 compatibility.
*/ */
struct partial_page { struct ia64_partial_page {
struct partial_page *next; /* linked list, sorted by address */ struct ia64_partial_page *next; /* linked list, sorted by address */
struct rb_node pp_rb; struct rb_node pp_rb;
/* 64K is the largest "normal" page supported by ia64 ABI. So 4K*64 /* 64K is the largest "normal" page supported by ia64 ABI. So 4K*64
* should suffice.*/ * should suffice.*/
...@@ -34,17 +34,17 @@ struct partial_page { ...@@ -34,17 +34,17 @@ struct partial_page {
unsigned int base; unsigned int base;
}; };
struct partial_page_list { struct ia64_partial_page_list {
struct partial_page *pp_head; /* list head, points to the lowest struct ia64_partial_page *pp_head; /* list head, points to the lowest
* addressed partial page */ * addressed partial page */
struct rb_root ppl_rb; struct rb_root ppl_rb;
struct partial_page *pp_hint; /* pp_hint->next is the last struct ia64_partial_page *pp_hint; /* pp_hint->next is the last
* accessed partial page */ * accessed partial page */
atomic_t pp_count; /* reference count */ atomic_t pp_count; /* reference count */
}; };
#if PAGE_SHIFT > IA32_PAGE_SHIFT #if PAGE_SHIFT > IA32_PAGE_SHIFT
struct partial_page_list* ia32_init_pp_list (void); struct ia64_partial_page_list* ia32_init_pp_list (void);
#else #else
# define ia32_init_pp_list() 0 # define ia32_init_pp_list() 0
#endif #endif
......
...@@ -253,17 +253,17 @@ mmap_subpage (struct file *file, unsigned long start, unsigned long end, int pro ...@@ -253,17 +253,17 @@ mmap_subpage (struct file *file, unsigned long start, unsigned long end, int pro
return ret; return ret;
} }
/* SLAB cache for partial_page structures */ /* SLAB cache for ia64_partial_page structures */
struct kmem_cache *partial_page_cachep; struct kmem_cache *ia64_partial_page_cachep;
/* /*
* init partial_page_list. * init ia64_partial_page_list.
* return 0 means kmalloc fail. * return 0 means kmalloc fail.
*/ */
struct partial_page_list* struct ia64_partial_page_list*
ia32_init_pp_list(void) ia32_init_pp_list(void)
{ {
struct partial_page_list *p; struct ia64_partial_page_list *p;
if ((p = kmalloc(sizeof(*p), GFP_KERNEL)) == NULL) if ((p = kmalloc(sizeof(*p), GFP_KERNEL)) == NULL)
return p; return p;
...@@ -280,12 +280,12 @@ ia32_init_pp_list(void) ...@@ -280,12 +280,12 @@ ia32_init_pp_list(void)
* Else, return 0 and provide @pprev, @rb_link, @rb_parent to * Else, return 0 and provide @pprev, @rb_link, @rb_parent to
* be used by later __ia32_insert_pp(). * be used by later __ia32_insert_pp().
*/ */
static struct partial_page * static struct ia64_partial_page *
__ia32_find_pp(struct partial_page_list *ppl, unsigned int start, __ia32_find_pp(struct ia64_partial_page_list *ppl, unsigned int start,
struct partial_page **pprev, struct rb_node ***rb_link, struct ia64_partial_page **pprev, struct rb_node ***rb_link,
struct rb_node **rb_parent) struct rb_node **rb_parent)
{ {
struct partial_page *pp; struct ia64_partial_page *pp;
struct rb_node **__rb_link, *__rb_parent, *rb_prev; struct rb_node **__rb_link, *__rb_parent, *rb_prev;
pp = ppl->pp_hint; pp = ppl->pp_hint;
...@@ -297,7 +297,7 @@ __ia32_find_pp(struct partial_page_list *ppl, unsigned int start, ...@@ -297,7 +297,7 @@ __ia32_find_pp(struct partial_page_list *ppl, unsigned int start,
while (*__rb_link) { while (*__rb_link) {
__rb_parent = *__rb_link; __rb_parent = *__rb_link;
pp = rb_entry(__rb_parent, struct partial_page, pp_rb); pp = rb_entry(__rb_parent, struct ia64_partial_page, pp_rb);
if (pp->base == start) { if (pp->base == start) {
ppl->pp_hint = pp; ppl->pp_hint = pp;
...@@ -314,7 +314,7 @@ __ia32_find_pp(struct partial_page_list *ppl, unsigned int start, ...@@ -314,7 +314,7 @@ __ia32_find_pp(struct partial_page_list *ppl, unsigned int start,
*rb_parent = __rb_parent; *rb_parent = __rb_parent;
*pprev = NULL; *pprev = NULL;
if (rb_prev) if (rb_prev)
*pprev = rb_entry(rb_prev, struct partial_page, pp_rb); *pprev = rb_entry(rb_prev, struct ia64_partial_page, pp_rb);
return NULL; return NULL;
} }
...@@ -322,9 +322,9 @@ __ia32_find_pp(struct partial_page_list *ppl, unsigned int start, ...@@ -322,9 +322,9 @@ __ia32_find_pp(struct partial_page_list *ppl, unsigned int start,
* insert @pp into @ppl. * insert @pp into @ppl.
*/ */
static void static void
__ia32_insert_pp(struct partial_page_list *ppl, struct partial_page *pp, __ia32_insert_pp(struct ia64_partial_page_list *ppl,
struct partial_page *prev, struct rb_node **rb_link, struct ia64_partial_page *pp, struct ia64_partial_page *prev,
struct rb_node *rb_parent) struct rb_node **rb_link, struct rb_node *rb_parent)
{ {
/* link list */ /* link list */
if (prev) { if (prev) {
...@@ -334,7 +334,7 @@ __ia32_insert_pp(struct partial_page_list *ppl, struct partial_page *pp, ...@@ -334,7 +334,7 @@ __ia32_insert_pp(struct partial_page_list *ppl, struct partial_page *pp,
ppl->pp_head = pp; ppl->pp_head = pp;
if (rb_parent) if (rb_parent)
pp->next = rb_entry(rb_parent, pp->next = rb_entry(rb_parent,
struct partial_page, pp_rb); struct ia64_partial_page, pp_rb);
else else
pp->next = NULL; pp->next = NULL;
} }
...@@ -350,8 +350,8 @@ __ia32_insert_pp(struct partial_page_list *ppl, struct partial_page *pp, ...@@ -350,8 +350,8 @@ __ia32_insert_pp(struct partial_page_list *ppl, struct partial_page *pp,
* delete @pp from partial page list @ppl. * delete @pp from partial page list @ppl.
*/ */
static void static void
__ia32_delete_pp(struct partial_page_list *ppl, struct partial_page *pp, __ia32_delete_pp(struct ia64_partial_page_list *ppl,
struct partial_page *prev) struct ia64_partial_page *pp, struct ia64_partial_page *prev)
{ {
if (prev) { if (prev) {
prev->next = pp->next; prev->next = pp->next;
...@@ -363,15 +363,15 @@ __ia32_delete_pp(struct partial_page_list *ppl, struct partial_page *pp, ...@@ -363,15 +363,15 @@ __ia32_delete_pp(struct partial_page_list *ppl, struct partial_page *pp,
ppl->pp_hint = pp->next; ppl->pp_hint = pp->next;
} }
rb_erase(&pp->pp_rb, &ppl->ppl_rb); rb_erase(&pp->pp_rb, &ppl->ppl_rb);
kmem_cache_free(partial_page_cachep, pp); kmem_cache_free(ia64_partial_page_cachep, pp);
} }
static struct partial_page * static struct ia64_partial_page *
__pp_prev(struct partial_page *pp) __pp_prev(struct ia64_partial_page *pp)
{ {
struct rb_node *prev = rb_prev(&pp->pp_rb); struct rb_node *prev = rb_prev(&pp->pp_rb);
if (prev) if (prev)
return rb_entry(prev, struct partial_page, pp_rb); return rb_entry(prev, struct ia64_partial_page, pp_rb);
else else
return NULL; return NULL;
} }
...@@ -383,7 +383,7 @@ __pp_prev(struct partial_page *pp) ...@@ -383,7 +383,7 @@ __pp_prev(struct partial_page *pp)
static void static void
__ia32_delete_pp_range(unsigned int start, unsigned int end) __ia32_delete_pp_range(unsigned int start, unsigned int end)
{ {
struct partial_page *pp, *prev; struct ia64_partial_page *pp, *prev;
struct rb_node **rb_link, *rb_parent; struct rb_node **rb_link, *rb_parent;
if (start >= end) if (start >= end)
...@@ -401,7 +401,7 @@ __ia32_delete_pp_range(unsigned int start, unsigned int end) ...@@ -401,7 +401,7 @@ __ia32_delete_pp_range(unsigned int start, unsigned int end)
} }
while (pp && pp->base < end) { while (pp && pp->base < end) {
struct partial_page *tmp = pp->next; struct ia64_partial_page *tmp = pp->next;
__ia32_delete_pp(current->thread.ppl, pp, prev); __ia32_delete_pp(current->thread.ppl, pp, prev);
pp = tmp; pp = tmp;
} }
...@@ -414,7 +414,7 @@ __ia32_delete_pp_range(unsigned int start, unsigned int end) ...@@ -414,7 +414,7 @@ __ia32_delete_pp_range(unsigned int start, unsigned int end)
static int static int
__ia32_set_pp(unsigned int start, unsigned int end, int flags) __ia32_set_pp(unsigned int start, unsigned int end, int flags)
{ {
struct partial_page *pp, *prev; struct ia64_partial_page *pp, *prev;
struct rb_node ** rb_link, *rb_parent; struct rb_node ** rb_link, *rb_parent;
unsigned int pstart, start_bit, end_bit, i; unsigned int pstart, start_bit, end_bit, i;
...@@ -450,8 +450,8 @@ __ia32_set_pp(unsigned int start, unsigned int end, int flags) ...@@ -450,8 +450,8 @@ __ia32_set_pp(unsigned int start, unsigned int end, int flags)
return 0; return 0;
} }
/* new a partial_page */ /* new a ia64_partial_page */
pp = kmem_cache_alloc(partial_page_cachep, GFP_KERNEL); pp = kmem_cache_alloc(ia64_partial_page_cachep, GFP_KERNEL);
if (!pp) if (!pp)
return -ENOMEM; return -ENOMEM;
pp->base = pstart; pp->base = pstart;
...@@ -504,7 +504,7 @@ ia32_set_pp(unsigned int start, unsigned int end, int flags) ...@@ -504,7 +504,7 @@ ia32_set_pp(unsigned int start, unsigned int end, int flags)
static int static int
__ia32_unset_pp(unsigned int start, unsigned int end) __ia32_unset_pp(unsigned int start, unsigned int end)
{ {
struct partial_page *pp, *prev; struct ia64_partial_page *pp, *prev;
struct rb_node ** rb_link, *rb_parent; struct rb_node ** rb_link, *rb_parent;
unsigned int pstart, start_bit, end_bit, i; unsigned int pstart, start_bit, end_bit, i;
struct vm_area_struct *vma; struct vm_area_struct *vma;
...@@ -532,8 +532,8 @@ __ia32_unset_pp(unsigned int start, unsigned int end) ...@@ -532,8 +532,8 @@ __ia32_unset_pp(unsigned int start, unsigned int end)
return -ENOMEM; return -ENOMEM;
} }
/* new a partial_page */ /* new a ia64_partial_page */
pp = kmem_cache_alloc(partial_page_cachep, GFP_KERNEL); pp = kmem_cache_alloc(ia64_partial_page_cachep, GFP_KERNEL);
if (!pp) if (!pp)
return -ENOMEM; return -ENOMEM;
pp->base = pstart; pp->base = pstart;
...@@ -605,7 +605,7 @@ ia32_unset_pp(unsigned int *startp, unsigned int *endp) ...@@ -605,7 +605,7 @@ ia32_unset_pp(unsigned int *startp, unsigned int *endp)
static int static int
__ia32_compare_pp(unsigned int start, unsigned int end) __ia32_compare_pp(unsigned int start, unsigned int end)
{ {
struct partial_page *pp, *prev; struct ia64_partial_page *pp, *prev;
struct rb_node ** rb_link, *rb_parent; struct rb_node ** rb_link, *rb_parent;
unsigned int pstart, start_bit, end_bit, size; unsigned int pstart, start_bit, end_bit, size;
unsigned int first_bit, next_zero_bit; /* the first range in bitmap */ unsigned int first_bit, next_zero_bit; /* the first range in bitmap */
...@@ -682,13 +682,13 @@ ia32_compare_pp(unsigned int *startp, unsigned int *endp) ...@@ -682,13 +682,13 @@ ia32_compare_pp(unsigned int *startp, unsigned int *endp)
} }
static void static void
__ia32_drop_pp_list(struct partial_page_list *ppl) __ia32_drop_pp_list(struct ia64_partial_page_list *ppl)
{ {
struct partial_page *pp = ppl->pp_head; struct ia64_partial_page *pp = ppl->pp_head;
while (pp) { while (pp) {
struct partial_page *next = pp->next; struct ia64_partial_page *next = pp->next;
kmem_cache_free(partial_page_cachep, pp); kmem_cache_free(ia64_partial_page_cachep, pp);
pp = next; pp = next;
} }
...@@ -696,9 +696,9 @@ __ia32_drop_pp_list(struct partial_page_list *ppl) ...@@ -696,9 +696,9 @@ __ia32_drop_pp_list(struct partial_page_list *ppl)
} }
void void
ia32_drop_partial_page_list(struct task_struct *task) ia32_drop_ia64_partial_page_list(struct task_struct *task)
{ {
struct partial_page_list* ppl = task->thread.ppl; struct ia64_partial_page_list* ppl = task->thread.ppl;
if (ppl && atomic_dec_and_test(&ppl->pp_count)) if (ppl && atomic_dec_and_test(&ppl->pp_count))
__ia32_drop_pp_list(ppl); __ia32_drop_pp_list(ppl);
...@@ -708,9 +708,9 @@ ia32_drop_partial_page_list(struct task_struct *task) ...@@ -708,9 +708,9 @@ ia32_drop_partial_page_list(struct task_struct *task)
* Copy current->thread.ppl to ppl (already initialized). * Copy current->thread.ppl to ppl (already initialized).
*/ */
static int static int
__ia32_copy_pp_list(struct partial_page_list *ppl) __ia32_copy_pp_list(struct ia64_partial_page_list *ppl)
{ {
struct partial_page *pp, *tmp, *prev; struct ia64_partial_page *pp, *tmp, *prev;
struct rb_node **rb_link, *rb_parent; struct rb_node **rb_link, *rb_parent;
ppl->pp_head = NULL; ppl->pp_head = NULL;
...@@ -721,7 +721,7 @@ __ia32_copy_pp_list(struct partial_page_list *ppl) ...@@ -721,7 +721,7 @@ __ia32_copy_pp_list(struct partial_page_list *ppl)
prev = NULL; prev = NULL;
for (pp = current->thread.ppl->pp_head; pp; pp = pp->next) { for (pp = current->thread.ppl->pp_head; pp; pp = pp->next) {
tmp = kmem_cache_alloc(partial_page_cachep, GFP_KERNEL); tmp = kmem_cache_alloc(ia64_partial_page_cachep, GFP_KERNEL);
if (!tmp) if (!tmp)
return -ENOMEM; return -ENOMEM;
*tmp = *pp; *tmp = *pp;
...@@ -734,7 +734,8 @@ __ia32_copy_pp_list(struct partial_page_list *ppl) ...@@ -734,7 +734,8 @@ __ia32_copy_pp_list(struct partial_page_list *ppl)
} }
int int
ia32_copy_partial_page_list(struct task_struct *p, unsigned long clone_flags) ia32_copy_ia64_partial_page_list(struct task_struct *p,
unsigned long clone_flags)
{ {
int retval = 0; int retval = 0;
......
...@@ -178,7 +178,7 @@ swapper_pg_dir: ...@@ -178,7 +178,7 @@ swapper_pg_dir:
halt_msg: halt_msg:
stringz "Halting kernel\n" stringz "Halting kernel\n"
.text .section .text.head,"ax"
.global start_ap .global start_ap
...@@ -392,6 +392,8 @@ self: hint @pause ...@@ -392,6 +392,8 @@ self: hint @pause
br.sptk.many self // endless loop br.sptk.many self // endless loop
END(_start) END(_start)
.text
GLOBAL_ENTRY(ia64_save_debug_regs) GLOBAL_ENTRY(ia64_save_debug_regs)
alloc r16=ar.pfs,1,0,0,0 alloc r16=ar.pfs,1,0,0,0
mov r20=ar.lc // preserve ar.lc mov r20=ar.lc // preserve ar.lc
......
...@@ -85,8 +85,8 @@ DEFINE_PER_CPU(int[IA64_NUM_VECTORS], vector_irq) = { ...@@ -85,8 +85,8 @@ DEFINE_PER_CPU(int[IA64_NUM_VECTORS], vector_irq) = {
[0 ... IA64_NUM_VECTORS - 1] = IA64_SPURIOUS_INT_VECTOR [0 ... IA64_NUM_VECTORS - 1] = IA64_SPURIOUS_INT_VECTOR
}; };
static cpumask_t vector_table[IA64_MAX_DEVICE_VECTORS] = { static cpumask_t vector_table[IA64_NUM_VECTORS] = {
[0 ... IA64_MAX_DEVICE_VECTORS - 1] = CPU_MASK_NONE [0 ... IA64_NUM_VECTORS - 1] = CPU_MASK_NONE
}; };
static int irq_status[NR_IRQS] = { static int irq_status[NR_IRQS] = {
...@@ -123,17 +123,18 @@ static inline int find_unassigned_irq(void) ...@@ -123,17 +123,18 @@ static inline int find_unassigned_irq(void)
static inline int find_unassigned_vector(cpumask_t domain) static inline int find_unassigned_vector(cpumask_t domain)
{ {
cpumask_t mask; cpumask_t mask;
int pos; int pos, vector;
cpus_and(mask, domain, cpu_online_map); cpus_and(mask, domain, cpu_online_map);
if (cpus_empty(mask)) if (cpus_empty(mask))
return -EINVAL; return -EINVAL;
for (pos = 0; pos < IA64_NUM_DEVICE_VECTORS; pos++) { for (pos = 0; pos < IA64_NUM_DEVICE_VECTORS; pos++) {
cpus_and(mask, domain, vector_table[pos]); vector = IA64_FIRST_DEVICE_VECTOR + pos;
cpus_and(mask, domain, vector_table[vector]);
if (!cpus_empty(mask)) if (!cpus_empty(mask))
continue; continue;
return IA64_FIRST_DEVICE_VECTOR + pos; return vector;
} }
return -ENOSPC; return -ENOSPC;
} }
...@@ -141,7 +142,7 @@ static inline int find_unassigned_vector(cpumask_t domain) ...@@ -141,7 +142,7 @@ static inline int find_unassigned_vector(cpumask_t domain)
static int __bind_irq_vector(int irq, int vector, cpumask_t domain) static int __bind_irq_vector(int irq, int vector, cpumask_t domain)
{ {
cpumask_t mask; cpumask_t mask;
int cpu, pos; int cpu;
struct irq_cfg *cfg = &irq_cfg[irq]; struct irq_cfg *cfg = &irq_cfg[irq];
cpus_and(mask, domain, cpu_online_map); cpus_and(mask, domain, cpu_online_map);
...@@ -156,8 +157,7 @@ static int __bind_irq_vector(int irq, int vector, cpumask_t domain) ...@@ -156,8 +157,7 @@ static int __bind_irq_vector(int irq, int vector, cpumask_t domain)
cfg->vector = vector; cfg->vector = vector;
cfg->domain = domain; cfg->domain = domain;
irq_status[irq] = IRQ_USED; irq_status[irq] = IRQ_USED;
pos = vector - IA64_FIRST_DEVICE_VECTOR; cpus_or(vector_table[vector], vector_table[vector], domain);
cpus_or(vector_table[pos], vector_table[pos], domain);
return 0; return 0;
} }
...@@ -174,7 +174,7 @@ int bind_irq_vector(int irq, int vector, cpumask_t domain) ...@@ -174,7 +174,7 @@ int bind_irq_vector(int irq, int vector, cpumask_t domain)
static void __clear_irq_vector(int irq) static void __clear_irq_vector(int irq)
{ {
int vector, cpu, pos; int vector, cpu;
cpumask_t mask; cpumask_t mask;
cpumask_t domain; cpumask_t domain;
struct irq_cfg *cfg = &irq_cfg[irq]; struct irq_cfg *cfg = &irq_cfg[irq];
...@@ -189,8 +189,7 @@ static void __clear_irq_vector(int irq) ...@@ -189,8 +189,7 @@ static void __clear_irq_vector(int irq)
cfg->vector = IRQ_VECTOR_UNASSIGNED; cfg->vector = IRQ_VECTOR_UNASSIGNED;
cfg->domain = CPU_MASK_NONE; cfg->domain = CPU_MASK_NONE;
irq_status[irq] = IRQ_UNUSED; irq_status[irq] = IRQ_UNUSED;
pos = vector - IA64_FIRST_DEVICE_VECTOR; cpus_andnot(vector_table[vector], vector_table[vector], domain);
cpus_andnot(vector_table[pos], vector_table[pos], domain);
} }
static void clear_irq_vector(int irq) static void clear_irq_vector(int irq)
...@@ -212,9 +211,6 @@ assign_irq_vector (int irq) ...@@ -212,9 +211,6 @@ assign_irq_vector (int irq)
vector = -ENOSPC; vector = -ENOSPC;
spin_lock_irqsave(&vector_lock, flags); spin_lock_irqsave(&vector_lock, flags);
if (irq < 0) {
goto out;
}
for_each_online_cpu(cpu) { for_each_online_cpu(cpu) {
domain = vector_allocation_domain(cpu); domain = vector_allocation_domain(cpu);
vector = find_unassigned_vector(domain); vector = find_unassigned_vector(domain);
...@@ -223,6 +219,8 @@ assign_irq_vector (int irq) ...@@ -223,6 +219,8 @@ assign_irq_vector (int irq)
} }
if (vector < 0) if (vector < 0)
goto out; goto out;
if (irq == AUTO_ASSIGN)
irq = vector;
BUG_ON(__bind_irq_vector(irq, vector, domain)); BUG_ON(__bind_irq_vector(irq, vector, domain));
out: out:
spin_unlock_irqrestore(&vector_lock, flags); spin_unlock_irqrestore(&vector_lock, flags);
......
...@@ -13,14 +13,6 @@ ...@@ -13,14 +13,6 @@
struct ia64_machine_vector ia64_mv; struct ia64_machine_vector ia64_mv;
EXPORT_SYMBOL(ia64_mv); EXPORT_SYMBOL(ia64_mv);
static __initdata const char *mvec_name;
static __init int setup_mvec(char *s)
{
mvec_name = s;
return 0;
}
early_param("machvec", setup_mvec);
static struct ia64_machine_vector * __init static struct ia64_machine_vector * __init
lookup_machvec (const char *name) lookup_machvec (const char *name)
{ {
...@@ -41,7 +33,7 @@ machvec_init (const char *name) ...@@ -41,7 +33,7 @@ machvec_init (const char *name)
struct ia64_machine_vector *mv; struct ia64_machine_vector *mv;
if (!name) if (!name)
name = mvec_name ? mvec_name : acpi_get_sysname(); name = acpi_get_sysname();
mv = lookup_machvec(name); mv = lookup_machvec(name);
if (!mv) if (!mv)
panic("generic kernel failed to find machine vector for" panic("generic kernel failed to find machine vector for"
...@@ -51,6 +43,23 @@ machvec_init (const char *name) ...@@ -51,6 +43,23 @@ machvec_init (const char *name)
printk(KERN_INFO "booting generic kernel on platform %s\n", name); printk(KERN_INFO "booting generic kernel on platform %s\n", name);
} }
void __init
machvec_init_from_cmdline(const char *cmdline)
{
char str[64];
const char *start;
char *end;
if (! (start = strstr(cmdline, "machvec=")) )
return machvec_init(NULL);
strlcpy(str, start + strlen("machvec="), sizeof(str));
if ( (end = strchr(str, ' ')) )
*end = '\0';
return machvec_init(str);
}
#endif /* CONFIG_IA64_GENERIC */ #endif /* CONFIG_IA64_GENERIC */
void void
......
...@@ -499,7 +499,8 @@ copy_thread (int nr, unsigned long clone_flags, ...@@ -499,7 +499,8 @@ copy_thread (int nr, unsigned long clone_flags,
/* Copy partially mapped page list */ /* Copy partially mapped page list */
if (!retval) if (!retval)
retval = ia32_copy_partial_page_list(p, clone_flags); retval = ia32_copy_ia64_partial_page_list(p,
clone_flags);
} }
#endif #endif
...@@ -728,7 +729,7 @@ flush_thread (void) ...@@ -728,7 +729,7 @@ flush_thread (void)
ia64_drop_fpu(current); ia64_drop_fpu(current);
#ifdef CONFIG_IA32_SUPPORT #ifdef CONFIG_IA32_SUPPORT
if (IS_IA32_PROCESS(task_pt_regs(current))) { if (IS_IA32_PROCESS(task_pt_regs(current))) {
ia32_drop_partial_page_list(current); ia32_drop_ia64_partial_page_list(current);
current->thread.task_size = IA32_PAGE_OFFSET; current->thread.task_size = IA32_PAGE_OFFSET;
set_fs(USER_DS); set_fs(USER_DS);
} }
...@@ -754,7 +755,7 @@ exit_thread (void) ...@@ -754,7 +755,7 @@ exit_thread (void)
pfm_release_debug_registers(current); pfm_release_debug_registers(current);
#endif #endif
if (IS_IA32_PROCESS(task_pt_regs(current))) if (IS_IA32_PROCESS(task_pt_regs(current)))
ia32_drop_partial_page_list(current); ia32_drop_ia64_partial_page_list(current);
} }
unsigned long unsigned long
......
...@@ -491,12 +491,17 @@ setup_arch (char **cmdline_p) ...@@ -491,12 +491,17 @@ setup_arch (char **cmdline_p)
efi_init(); efi_init();
io_port_init(); io_port_init();
parse_early_param();
#ifdef CONFIG_IA64_GENERIC #ifdef CONFIG_IA64_GENERIC
machvec_init(NULL); /* machvec needs to be parsed from the command line
* before parse_early_param() is called to ensure
* that ia64_mv is initialised before any command line
* settings may cause console setup to occur
*/
machvec_init_from_cmdline(*cmdline_p);
#endif #endif
parse_early_param();
if (early_console_setup(*cmdline_p) == 0) if (early_console_setup(*cmdline_p) == 0)
mark_bsp_online(); mark_bsp_online();
......
...@@ -468,7 +468,7 @@ smp_send_stop (void) ...@@ -468,7 +468,7 @@ smp_send_stop (void)
send_IPI_allbutself(IPI_CPU_STOP); send_IPI_allbutself(IPI_CPU_STOP);
} }
int __init int
setup_profiling_timer (unsigned int multiplier) setup_profiling_timer (unsigned int multiplier)
{ {
return -EINVAL; return -EINVAL;
......
...@@ -487,7 +487,7 @@ struct create_idle { ...@@ -487,7 +487,7 @@ struct create_idle {
int cpu; int cpu;
}; };
void void __cpuinit
do_fork_idle(struct work_struct *work) do_fork_idle(struct work_struct *work)
{ {
struct create_idle *c_idle = struct create_idle *c_idle =
...@@ -497,7 +497,7 @@ do_fork_idle(struct work_struct *work) ...@@ -497,7 +497,7 @@ do_fork_idle(struct work_struct *work)
complete(&c_idle->done); complete(&c_idle->done);
} }
static int __devinit static int __cpuinit
do_boot_cpu (int sapicid, int cpu) do_boot_cpu (int sapicid, int cpu)
{ {
int timeout; int timeout;
...@@ -808,7 +808,7 @@ set_cpu_sibling_map(int cpu) ...@@ -808,7 +808,7 @@ set_cpu_sibling_map(int cpu)
} }
} }
int __devinit int __cpuinit
__cpu_up (unsigned int cpu) __cpu_up (unsigned int cpu)
{ {
int ret; int ret;
......
...@@ -50,6 +50,8 @@ SECTIONS ...@@ -50,6 +50,8 @@ SECTIONS
KPROBES_TEXT KPROBES_TEXT
*(.gnu.linkonce.t*) *(.gnu.linkonce.t*)
} }
.text.head : AT(ADDR(.text.head) - LOAD_OFFSET)
{ *(.text.head) }
.text2 : AT(ADDR(.text2) - LOAD_OFFSET) .text2 : AT(ADDR(.text2) - LOAD_OFFSET)
{ *(.text2) } { *(.text2) }
#ifdef CONFIG_SMP #ifdef CONFIG_SMP
......
...@@ -581,7 +581,7 @@ pcibios_align_resource (void *data, struct resource *res, ...@@ -581,7 +581,7 @@ pcibios_align_resource (void *data, struct resource *res,
/* /*
* PCI BIOS setup, always defaults to SAL interface * PCI BIOS setup, always defaults to SAL interface
*/ */
char * __init char * __devinit
pcibios_setup (char *str) pcibios_setup (char *str)
{ {
return str; return str;
......
...@@ -27,11 +27,12 @@ extern int ia32_clone_tls (struct task_struct *child, struct pt_regs *childregs) ...@@ -27,11 +27,12 @@ extern int ia32_clone_tls (struct task_struct *child, struct pt_regs *childregs)
extern int ia32_setup_frame1 (int sig, struct k_sigaction *ka, siginfo_t *info, extern int ia32_setup_frame1 (int sig, struct k_sigaction *ka, siginfo_t *info,
sigset_t *set, struct pt_regs *regs); sigset_t *set, struct pt_regs *regs);
#if PAGE_SHIFT > IA32_PAGE_SHIFT #if PAGE_SHIFT > IA32_PAGE_SHIFT
extern int ia32_copy_partial_page_list (struct task_struct *, unsigned long); extern int ia32_copy_ia64_partial_page_list(struct task_struct *,
extern void ia32_drop_partial_page_list (struct task_struct *); unsigned long);
extern void ia32_drop_ia64_partial_page_list(struct task_struct *);
#else #else
# define ia32_copy_partial_page_list(a1, a2) 0 # define ia32_copy_ia64_partial_page_list(a1, a2) 0
# define ia32_drop_partial_page_list(a1) do { ; } while (0) # define ia32_drop_ia64_partial_page_list(a1) do { ; } while (0)
#endif #endif
#endif /* !__ASSEMBLY__ */ #endif /* !__ASSEMBLY__ */
......
...@@ -275,6 +275,7 @@ struct ia64_machine_vector { ...@@ -275,6 +275,7 @@ struct ia64_machine_vector {
extern struct ia64_machine_vector ia64_mv; extern struct ia64_machine_vector ia64_mv;
extern void machvec_init (const char *name); extern void machvec_init (const char *name);
extern void machvec_init_from_cmdline(const char *cmdline);
# else # else
# error Unknown configuration. Update asm-ia64/machvec.h. # error Unknown configuration. Update asm-ia64/machvec.h.
......
...@@ -220,7 +220,7 @@ struct desc_struct { ...@@ -220,7 +220,7 @@ struct desc_struct {
#define TLS_SIZE (GDT_ENTRY_TLS_ENTRIES * 8) #define TLS_SIZE (GDT_ENTRY_TLS_ENTRIES * 8)
struct partial_page_list; struct ia64_partial_page_list;
#endif #endif
struct thread_struct { struct thread_struct {
...@@ -242,7 +242,7 @@ struct thread_struct { ...@@ -242,7 +242,7 @@ struct thread_struct {
__u64 fdr; /* IA32 fp except. data reg */ __u64 fdr; /* IA32 fp except. data reg */
__u64 old_k1; /* old value of ar.k1 */ __u64 old_k1; /* old value of ar.k1 */
__u64 old_iob; /* old IOBase value */ __u64 old_iob; /* old IOBase value */
struct partial_page_list *ppl; /* partial page list for 4K page size issue */ struct ia64_partial_page_list *ppl; /* partial page list for 4K page size issue */
/* cached TLS descriptors. */ /* cached TLS descriptors. */
struct desc_struct tls_array[GDT_ENTRY_TLS_ENTRIES]; struct desc_struct tls_array[GDT_ENTRY_TLS_ENTRIES];
......
...@@ -116,7 +116,6 @@ max_xtp (void) ...@@ -116,7 +116,6 @@ max_xtp (void)
extern int __cpu_disable (void); extern int __cpu_disable (void);
extern void __cpu_die (unsigned int cpu); extern void __cpu_die (unsigned int cpu);
extern void cpu_die (void) __attribute__ ((noreturn)); extern void cpu_die (void) __attribute__ ((noreturn));
extern int __cpu_up (unsigned int cpu);
extern void __init smp_build_cpu_map(void); extern void __init smp_build_cpu_map(void);
extern void __init init_smp_config (void); extern void __init init_smp_config (void);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment