Commit 50fbe56c authored by Tony Luck's avatar Tony Luck Committed by Tony Luck

Pull cpumask into release branch

parents 15f7176e 5d8c39f6
...@@ -87,7 +87,7 @@ get_mmu_context (struct mm_struct *mm) ...@@ -87,7 +87,7 @@ get_mmu_context (struct mm_struct *mm)
/* re-check, now that we've got the lock: */ /* re-check, now that we've got the lock: */
context = mm->context; context = mm->context;
if (context == 0) { if (context == 0) {
cpus_clear(mm->cpu_vm_mask); cpumask_clear(mm_cpumask(mm));
if (ia64_ctx.next >= ia64_ctx.limit) { if (ia64_ctx.next >= ia64_ctx.limit) {
ia64_ctx.next = find_next_zero_bit(ia64_ctx.bitmap, ia64_ctx.next = find_next_zero_bit(ia64_ctx.bitmap,
ia64_ctx.max_ctx, ia64_ctx.next); ia64_ctx.max_ctx, ia64_ctx.next);
...@@ -166,8 +166,8 @@ activate_context (struct mm_struct *mm) ...@@ -166,8 +166,8 @@ activate_context (struct mm_struct *mm)
do { do {
context = get_mmu_context(mm); context = get_mmu_context(mm);
if (!cpu_isset(smp_processor_id(), mm->cpu_vm_mask)) if (!cpumask_test_cpu(smp_processor_id(), mm_cpumask(mm)))
cpu_set(smp_processor_id(), mm->cpu_vm_mask); cpumask_set_cpu(smp_processor_id(), mm_cpumask(mm));
reload_context(context); reload_context(context);
/* /*
* in the unlikely event of a TLB-flush by another thread, * in the unlikely event of a TLB-flush by another thread,
......
...@@ -126,7 +126,8 @@ extern void identify_siblings (struct cpuinfo_ia64 *); ...@@ -126,7 +126,8 @@ extern void identify_siblings (struct cpuinfo_ia64 *);
extern int is_multithreading_enabled(void); extern int is_multithreading_enabled(void);
extern void arch_send_call_function_single_ipi(int cpu); extern void arch_send_call_function_single_ipi(int cpu);
extern void arch_send_call_function_ipi(cpumask_t mask); extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
#define arch_send_call_function_ipi_mask arch_send_call_function_ipi_mask
#else /* CONFIG_SMP */ #else /* CONFIG_SMP */
......
...@@ -112,11 +112,6 @@ void build_cpu_to_node_map(void); ...@@ -112,11 +112,6 @@ void build_cpu_to_node_map(void);
extern void arch_fix_phys_package_id(int num, u32 slot); extern void arch_fix_phys_package_id(int num, u32 slot);
#define pcibus_to_cpumask(bus) (pcibus_to_node(bus) == -1 ? \
CPU_MASK_ALL : \
node_to_cpumask(pcibus_to_node(bus)) \
)
#define cpumask_of_pcibus(bus) (pcibus_to_node(bus) == -1 ? \ #define cpumask_of_pcibus(bus) (pcibus_to_node(bus) == -1 ? \
cpu_all_mask : \ cpu_all_mask : \
cpumask_of_node(pcibus_to_node(bus))) cpumask_of_node(pcibus_to_node(bus)))
......
...@@ -890,7 +890,7 @@ __init void prefill_possible_map(void) ...@@ -890,7 +890,7 @@ __init void prefill_possible_map(void)
possible, max((possible - available_cpus), 0)); possible, max((possible - available_cpus), 0));
for (i = 0; i < possible; i++) for (i = 0; i < possible; i++)
cpu_set(i, cpu_possible_map); set_cpu_possible(i, true);
} }
int acpi_map_lsapic(acpi_handle handle, int *pcpu) int acpi_map_lsapic(acpi_handle handle, int *pcpu)
...@@ -928,9 +928,9 @@ int acpi_map_lsapic(acpi_handle handle, int *pcpu) ...@@ -928,9 +928,9 @@ int acpi_map_lsapic(acpi_handle handle, int *pcpu)
buffer.length = ACPI_ALLOCATE_BUFFER; buffer.length = ACPI_ALLOCATE_BUFFER;
buffer.pointer = NULL; buffer.pointer = NULL;
cpus_complement(tmp_map, cpu_present_map); cpumask_complement(&tmp_map, cpu_present_mask);
cpu = first_cpu(tmp_map); cpu = cpumask_first(&tmp_map);
if (cpu >= NR_CPUS) if (cpu >= nr_cpu_ids)
return -EINVAL; return -EINVAL;
acpi_map_cpu2node(handle, cpu, physid); acpi_map_cpu2node(handle, cpu, physid);
......
...@@ -1456,9 +1456,9 @@ ia64_mca_cmc_int_caller(int cmc_irq, void *arg) ...@@ -1456,9 +1456,9 @@ ia64_mca_cmc_int_caller(int cmc_irq, void *arg)
ia64_mca_cmc_int_handler(cmc_irq, arg); ia64_mca_cmc_int_handler(cmc_irq, arg);
for (++cpuid ; cpuid < NR_CPUS && !cpu_online(cpuid) ; cpuid++); cpuid = cpumask_next(cpuid+1, cpu_online_mask);
if (cpuid < NR_CPUS) { if (cpuid < nr_cpu_ids) {
platform_send_ipi(cpuid, IA64_CMCP_VECTOR, IA64_IPI_DM_INT, 0); platform_send_ipi(cpuid, IA64_CMCP_VECTOR, IA64_IPI_DM_INT, 0);
} else { } else {
/* If no log record, switch out of polling mode */ /* If no log record, switch out of polling mode */
...@@ -1525,7 +1525,7 @@ ia64_mca_cpe_int_caller(int cpe_irq, void *arg) ...@@ -1525,7 +1525,7 @@ ia64_mca_cpe_int_caller(int cpe_irq, void *arg)
ia64_mca_cpe_int_handler(cpe_irq, arg); ia64_mca_cpe_int_handler(cpe_irq, arg);
for (++cpuid ; cpuid < NR_CPUS && !cpu_online(cpuid) ; cpuid++); cpuid = cpumask_next(cpuid+1, cpu_online_mask);
if (cpuid < NR_CPUS) { if (cpuid < NR_CPUS) {
platform_send_ipi(cpuid, IA64_CPEP_VECTOR, IA64_IPI_DM_INT, 0); platform_send_ipi(cpuid, IA64_CPEP_VECTOR, IA64_IPI_DM_INT, 0);
......
...@@ -5603,7 +5603,7 @@ pfm_interrupt_handler(int irq, void *arg) ...@@ -5603,7 +5603,7 @@ pfm_interrupt_handler(int irq, void *arg)
* /proc/perfmon interface, for debug only * /proc/perfmon interface, for debug only
*/ */
#define PFM_PROC_SHOW_HEADER ((void *)NR_CPUS+1) #define PFM_PROC_SHOW_HEADER ((void *)nr_cpu_ids+1)
static void * static void *
pfm_proc_start(struct seq_file *m, loff_t *pos) pfm_proc_start(struct seq_file *m, loff_t *pos)
...@@ -5612,7 +5612,7 @@ pfm_proc_start(struct seq_file *m, loff_t *pos) ...@@ -5612,7 +5612,7 @@ pfm_proc_start(struct seq_file *m, loff_t *pos)
return PFM_PROC_SHOW_HEADER; return PFM_PROC_SHOW_HEADER;
} }
while (*pos <= NR_CPUS) { while (*pos <= nr_cpu_ids) {
if (cpu_online(*pos - 1)) { if (cpu_online(*pos - 1)) {
return (void *)*pos; return (void *)*pos;
} }
......
...@@ -317,7 +317,7 @@ retry: ...@@ -317,7 +317,7 @@ retry:
} }
n = data->cpu_check; n = data->cpu_check;
for (i = 0; i < NR_CPUS; i++) { for (i = 0; i < nr_cpu_ids; i++) {
if (cpu_isset(n, data->cpu_event)) { if (cpu_isset(n, data->cpu_event)) {
if (!cpu_online(n)) { if (!cpu_online(n)) {
cpu_clear(n, data->cpu_event); cpu_clear(n, data->cpu_event);
...@@ -326,7 +326,7 @@ retry: ...@@ -326,7 +326,7 @@ retry:
cpu = n; cpu = n;
break; break;
} }
if (++n == NR_CPUS) if (++n == nr_cpu_ids)
n = 0; n = 0;
} }
...@@ -337,7 +337,7 @@ retry: ...@@ -337,7 +337,7 @@ retry:
/* for next read, start checking at next CPU */ /* for next read, start checking at next CPU */
data->cpu_check = cpu; data->cpu_check = cpu;
if (++data->cpu_check == NR_CPUS) if (++data->cpu_check == nr_cpu_ids)
data->cpu_check = 0; data->cpu_check = 0;
snprintf(cmd, sizeof(cmd), "read %d\n", cpu); snprintf(cmd, sizeof(cmd), "read %d\n", cpu);
......
...@@ -730,10 +730,10 @@ static void * ...@@ -730,10 +730,10 @@ static void *
c_start (struct seq_file *m, loff_t *pos) c_start (struct seq_file *m, loff_t *pos)
{ {
#ifdef CONFIG_SMP #ifdef CONFIG_SMP
while (*pos < NR_CPUS && !cpu_isset(*pos, cpu_online_map)) while (*pos < nr_cpu_ids && !cpu_online(*pos))
++*pos; ++*pos;
#endif #endif
return *pos < NR_CPUS ? cpu_data(*pos) : NULL; return *pos < nr_cpu_ids ? cpu_data(*pos) : NULL;
} }
static void * static void *
......
...@@ -166,11 +166,11 @@ send_IPI_allbutself (int op) ...@@ -166,11 +166,11 @@ send_IPI_allbutself (int op)
* Called with preemption disabled. * Called with preemption disabled.
*/ */
static inline void static inline void
send_IPI_mask(cpumask_t mask, int op) send_IPI_mask(const struct cpumask *mask, int op)
{ {
unsigned int cpu; unsigned int cpu;
for_each_cpu_mask(cpu, mask) { for_each_cpu(cpu, mask) {
send_IPI_single(cpu, op); send_IPI_single(cpu, op);
} }
} }
...@@ -316,7 +316,7 @@ void arch_send_call_function_single_ipi(int cpu) ...@@ -316,7 +316,7 @@ void arch_send_call_function_single_ipi(int cpu)
send_IPI_single(cpu, IPI_CALL_FUNC_SINGLE); send_IPI_single(cpu, IPI_CALL_FUNC_SINGLE);
} }
void arch_send_call_function_ipi(cpumask_t mask) void arch_send_call_function_ipi_mask(const struct cpumask *mask)
{ {
send_IPI_mask(mask, IPI_CALL_FUNC); send_IPI_mask(mask, IPI_CALL_FUNC);
} }
......
...@@ -581,14 +581,14 @@ smp_build_cpu_map (void) ...@@ -581,14 +581,14 @@ smp_build_cpu_map (void)
ia64_cpu_to_sapicid[0] = boot_cpu_id; ia64_cpu_to_sapicid[0] = boot_cpu_id;
cpus_clear(cpu_present_map); cpus_clear(cpu_present_map);
cpu_set(0, cpu_present_map); set_cpu_present(0, true);
cpu_set(0, cpu_possible_map); set_cpu_possible(0, true);
for (cpu = 1, i = 0; i < smp_boot_data.cpu_count; i++) { for (cpu = 1, i = 0; i < smp_boot_data.cpu_count; i++) {
sapicid = smp_boot_data.cpu_phys_id[i]; sapicid = smp_boot_data.cpu_phys_id[i];
if (sapicid == boot_cpu_id) if (sapicid == boot_cpu_id)
continue; continue;
cpu_set(cpu, cpu_present_map); set_cpu_present(cpu, true);
cpu_set(cpu, cpu_possible_map); set_cpu_possible(cpu, true);
ia64_cpu_to_sapicid[cpu] = sapicid; ia64_cpu_to_sapicid[cpu] = sapicid;
cpu++; cpu++;
} }
...@@ -626,12 +626,9 @@ smp_prepare_cpus (unsigned int max_cpus) ...@@ -626,12 +626,9 @@ smp_prepare_cpus (unsigned int max_cpus)
*/ */
if (!max_cpus) { if (!max_cpus) {
printk(KERN_INFO "SMP mode deactivated.\n"); printk(KERN_INFO "SMP mode deactivated.\n");
cpus_clear(cpu_online_map); init_cpu_online(cpumask_of(0));
cpus_clear(cpu_present_map); init_cpu_present(cpumask_of(0));
cpus_clear(cpu_possible_map); init_cpu_possible(cpumask_of(0));
cpu_set(0, cpu_online_map);
cpu_set(0, cpu_present_map);
cpu_set(0, cpu_possible_map);
return; return;
} }
} }
......
...@@ -309,7 +309,7 @@ flush_tlb_range (struct vm_area_struct *vma, unsigned long start, ...@@ -309,7 +309,7 @@ flush_tlb_range (struct vm_area_struct *vma, unsigned long start,
preempt_disable(); preempt_disable();
#ifdef CONFIG_SMP #ifdef CONFIG_SMP
if (mm != current->active_mm || cpus_weight(mm->cpu_vm_mask) != 1) { if (mm != current->active_mm || cpumask_weight(mm_cpumask(mm)) != 1) {
platform_global_tlb_purge(mm, start, end, nbits); platform_global_tlb_purge(mm, start, end, nbits);
preempt_enable(); preempt_enable();
return; return;
......
...@@ -750,7 +750,7 @@ nasid_slice_to_cpuid(int nasid, int slice) ...@@ -750,7 +750,7 @@ nasid_slice_to_cpuid(int nasid, int slice)
{ {
long cpu; long cpu;
for (cpu = 0; cpu < NR_CPUS; cpu++) for (cpu = 0; cpu < nr_cpu_ids; cpu++)
if (cpuid_to_nasid(cpu) == nasid && if (cpuid_to_nasid(cpu) == nasid &&
cpuid_to_slice(cpu) == slice) cpuid_to_slice(cpu) == slice)
return cpu; return cpu;
......
...@@ -133,7 +133,7 @@ sn2_ipi_flush_all_tlb(struct mm_struct *mm) ...@@ -133,7 +133,7 @@ sn2_ipi_flush_all_tlb(struct mm_struct *mm)
unsigned long itc; unsigned long itc;
itc = ia64_get_itc(); itc = ia64_get_itc();
smp_flush_tlb_cpumask(mm->cpu_vm_mask); smp_flush_tlb_cpumask(*mm_cpumask(mm));
itc = ia64_get_itc() - itc; itc = ia64_get_itc() - itc;
__get_cpu_var(ptcstats).shub_ipi_flushes_itc_clocks += itc; __get_cpu_var(ptcstats).shub_ipi_flushes_itc_clocks += itc;
__get_cpu_var(ptcstats).shub_ipi_flushes++; __get_cpu_var(ptcstats).shub_ipi_flushes++;
...@@ -182,7 +182,7 @@ sn2_global_tlb_purge(struct mm_struct *mm, unsigned long start, ...@@ -182,7 +182,7 @@ sn2_global_tlb_purge(struct mm_struct *mm, unsigned long start,
nodes_clear(nodes_flushed); nodes_clear(nodes_flushed);
i = 0; i = 0;
for_each_cpu_mask(cpu, mm->cpu_vm_mask) { for_each_cpu(cpu, mm_cpumask(mm)) {
cnode = cpu_to_node(cpu); cnode = cpu_to_node(cpu);
node_set(cnode, nodes_flushed); node_set(cnode, nodes_flushed);
lcpu = cpu; lcpu = cpu;
...@@ -461,7 +461,7 @@ bool sn_cpu_disable_allowed(int cpu) ...@@ -461,7 +461,7 @@ bool sn_cpu_disable_allowed(int cpu)
static void *sn2_ptc_seq_start(struct seq_file *file, loff_t * offset) static void *sn2_ptc_seq_start(struct seq_file *file, loff_t * offset)
{ {
if (*offset < NR_CPUS) if (*offset < nr_cpu_ids)
return offset; return offset;
return NULL; return NULL;
} }
...@@ -469,7 +469,7 @@ static void *sn2_ptc_seq_start(struct seq_file *file, loff_t * offset) ...@@ -469,7 +469,7 @@ static void *sn2_ptc_seq_start(struct seq_file *file, loff_t * offset)
static void *sn2_ptc_seq_next(struct seq_file *file, void *data, loff_t * offset) static void *sn2_ptc_seq_next(struct seq_file *file, void *data, loff_t * offset)
{ {
(*offset)++; (*offset)++;
if (*offset < NR_CPUS) if (*offset < nr_cpu_ids)
return offset; return offset;
return NULL; return NULL;
} }
...@@ -491,7 +491,7 @@ static int sn2_ptc_seq_show(struct seq_file *file, void *data) ...@@ -491,7 +491,7 @@ static int sn2_ptc_seq_show(struct seq_file *file, void *data)
seq_printf(file, "# ptctest %d, flushopt %d\n", sn2_ptctest, sn2_flush_opt); seq_printf(file, "# ptctest %d, flushopt %d\n", sn2_ptctest, sn2_flush_opt);
} }
if (cpu < NR_CPUS && cpu_online(cpu)) { if (cpu < nr_cpu_ids && cpu_online(cpu)) {
stat = &per_cpu(ptcstats, cpu); stat = &per_cpu(ptcstats, cpu);
seq_printf(file, "cpu %d %ld %ld %ld %ld %ld %ld %ld %ld %ld %ld %ld %ld\n", cpu, stat->ptc_l, seq_printf(file, "cpu %d %ld %ld %ld %ld %ld %ld %ld %ld %ld %ld %ld %ld\n", cpu, stat->ptc_l,
stat->change_rid, stat->shub_ptc_flushes, stat->nodes_flushed, stat->change_rid, stat->shub_ptc_flushes, stat->nodes_flushed,
......
...@@ -612,7 +612,7 @@ static int sn_hwperf_op_cpu(struct sn_hwperf_op_info *op_info) ...@@ -612,7 +612,7 @@ static int sn_hwperf_op_cpu(struct sn_hwperf_op_info *op_info)
op_info->a->arg &= SN_HWPERF_ARG_OBJID_MASK; op_info->a->arg &= SN_HWPERF_ARG_OBJID_MASK;
if (cpu != SN_HWPERF_ARG_ANY_CPU) { if (cpu != SN_HWPERF_ARG_ANY_CPU) {
if (cpu >= NR_CPUS || !cpu_online(cpu)) { if (cpu >= nr_cpu_ids || !cpu_online(cpu)) {
r = -EINVAL; r = -EINVAL;
goto out; goto out;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment