Commit 6bb74df4 authored by john stultz's avatar john stultz Committed by Linus Torvalds

[PATCH] clocksource init adjustments (fix bug #7426)

This patch resolves the issue found here:
http://bugme.osdl.org/show_bug.cgi?id=7426

The basic summary is:
Currently we register most of i386/x86_64 clocksources at module_init
time. Then we enable clocksource selection at late_initcall time. This
causes some problems for drivers that use gettimeofday for init
calibration routines (specifically the es1968 driver in this case),
where durring module_init, the only clocksource available is the low-res
jiffies clocksource. This may cause slight calibration errors, due to
the small sampling time used.

It should be noted that drivers that require fine grained time may not
function on architectures that do not have better then jiffies
resolution timekeeping (there are a few). However, this does not
discount the reasonable need for such fine-grained timekeeping at init
time.

Thus the solution here is to register clocksources earlier (ideally when
the hardware is being initialized), and then we enable clocksource
selection at fs_initcall (before device_initcall).

This patch should probably get some testing time in -mm, since
clocksource selection is one of the most important issues for correct
timekeeping, and I've only been able to test this on a few of my own
boxes.
Signed-off-by: default avatarJohn Stultz <johnstul@us.ibm.com>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: "David S. Miller" <davem@davemloft.net>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
parent 45407680
...@@ -200,6 +200,23 @@ static int hpet_next_event(unsigned long delta, ...@@ -200,6 +200,23 @@ static int hpet_next_event(unsigned long delta,
return ((long)(hpet_readl(HPET_COUNTER) - cnt ) > 0); return ((long)(hpet_readl(HPET_COUNTER) - cnt ) > 0);
} }
/*
* Clock source related code
*/
static cycle_t read_hpet(void)
{
return (cycle_t)hpet_readl(HPET_COUNTER);
}
static struct clocksource clocksource_hpet = {
.name = "hpet",
.rating = 250,
.read = read_hpet,
.mask = HPET_MASK,
.shift = HPET_SHIFT,
.flags = CLOCK_SOURCE_IS_CONTINUOUS,
};
/* /*
* Try to setup the HPET timer * Try to setup the HPET timer
*/ */
...@@ -207,6 +224,7 @@ int __init hpet_enable(void) ...@@ -207,6 +224,7 @@ int __init hpet_enable(void)
{ {
unsigned long id; unsigned long id;
uint64_t hpet_freq; uint64_t hpet_freq;
u64 tmp;
if (!is_hpet_capable()) if (!is_hpet_capable())
return 0; return 0;
...@@ -253,6 +271,25 @@ int __init hpet_enable(void) ...@@ -253,6 +271,25 @@ int __init hpet_enable(void)
/* Start the counter */ /* Start the counter */
hpet_start_counter(); hpet_start_counter();
/* Initialize and register HPET clocksource
*
* hpet period is in femto seconds per cycle
* so we need to convert this to ns/cyc units
* aproximated by mult/2^shift
*
* fsec/cyc * 1nsec/1000000fsec = nsec/cyc = mult/2^shift
* fsec/cyc * 1ns/1000000fsec * 2^shift = mult
* fsec/cyc * 2^shift * 1nsec/1000000fsec = mult
* (fsec/cyc << shift)/1000000 = mult
* (hpet_period << shift)/FSEC_PER_NSEC = mult
*/
tmp = (u64)hpet_period << HPET_SHIFT;
do_div(tmp, FSEC_PER_NSEC);
clocksource_hpet.mult = (u32)tmp;
clocksource_register(&clocksource_hpet);
if (id & HPET_ID_LEGSUP) { if (id & HPET_ID_LEGSUP) {
hpet_enable_int(); hpet_enable_int();
hpet_reserve_platform_timers(id); hpet_reserve_platform_timers(id);
...@@ -273,49 +310,6 @@ out_nohpet: ...@@ -273,49 +310,6 @@ out_nohpet:
return 0; return 0;
} }
/*
* Clock source related code
*/
static cycle_t read_hpet(void)
{
return (cycle_t)hpet_readl(HPET_COUNTER);
}
static struct clocksource clocksource_hpet = {
.name = "hpet",
.rating = 250,
.read = read_hpet,
.mask = HPET_MASK,
.shift = HPET_SHIFT,
.flags = CLOCK_SOURCE_IS_CONTINUOUS,
};
static int __init init_hpet_clocksource(void)
{
u64 tmp;
if (!hpet_virt_address)
return -ENODEV;
/*
* hpet period is in femto seconds per cycle
* so we need to convert this to ns/cyc units
* aproximated by mult/2^shift
*
* fsec/cyc * 1nsec/1000000fsec = nsec/cyc = mult/2^shift
* fsec/cyc * 1ns/1000000fsec * 2^shift = mult
* fsec/cyc * 2^shift * 1nsec/1000000fsec = mult
* (fsec/cyc << shift)/1000000 = mult
* (hpet_period << shift)/FSEC_PER_NSEC = mult
*/
tmp = (u64)hpet_period << HPET_SHIFT;
do_div(tmp, FSEC_PER_NSEC);
clocksource_hpet.mult = (u32)tmp;
return clocksource_register(&clocksource_hpet);
}
module_init(init_hpet_clocksource);
#ifdef CONFIG_HPET_EMULATE_RTC #ifdef CONFIG_HPET_EMULATE_RTC
......
...@@ -195,4 +195,4 @@ static int __init init_pit_clocksource(void) ...@@ -195,4 +195,4 @@ static int __init init_pit_clocksource(void)
clocksource_pit.mult = clocksource_hz2mult(CLOCK_TICK_RATE, 20); clocksource_pit.mult = clocksource_hz2mult(CLOCK_TICK_RATE, 20);
return clocksource_register(&clocksource_pit); return clocksource_register(&clocksource_pit);
} }
module_init(init_pit_clocksource); arch_initcall(init_pit_clocksource);
...@@ -657,5 +657,4 @@ void __init setup_arch(char **cmdline_p) ...@@ -657,5 +657,4 @@ void __init setup_arch(char **cmdline_p)
conswitchp = &dummy_con; conswitchp = &dummy_con;
#endif #endif
#endif #endif
tsc_init();
} }
...@@ -279,5 +279,6 @@ void __init hpet_time_init(void) ...@@ -279,5 +279,6 @@ void __init hpet_time_init(void)
*/ */
void __init time_init(void) void __init time_init(void)
{ {
tsc_init();
late_time_init = choose_time_init(); late_time_init = choose_time_init();
} }
...@@ -184,34 +184,6 @@ int recalibrate_cpu_khz(void) ...@@ -184,34 +184,6 @@ int recalibrate_cpu_khz(void)
EXPORT_SYMBOL(recalibrate_cpu_khz); EXPORT_SYMBOL(recalibrate_cpu_khz);
void __init tsc_init(void)
{
if (!cpu_has_tsc || tsc_disable)
goto out_no_tsc;
cpu_khz = calculate_cpu_khz();
tsc_khz = cpu_khz;
if (!cpu_khz)
goto out_no_tsc;
printk("Detected %lu.%03lu MHz processor.\n",
(unsigned long)cpu_khz / 1000,
(unsigned long)cpu_khz % 1000);
set_cyc2ns_scale(cpu_khz);
use_tsc_delay();
return;
out_no_tsc:
/*
* Set the tsc_disable flag if there's no TSC support, this
* makes it a fast flag for the kernel to see whether it
* should be using the TSC.
*/
tsc_disable = 1;
}
#ifdef CONFIG_CPU_FREQ #ifdef CONFIG_CPU_FREQ
/* /*
...@@ -381,28 +353,47 @@ static void __init check_geode_tsc_reliable(void) ...@@ -381,28 +353,47 @@ static void __init check_geode_tsc_reliable(void)
static inline void check_geode_tsc_reliable(void) { } static inline void check_geode_tsc_reliable(void) { }
#endif #endif
static int __init init_tsc_clocksource(void)
void __init tsc_init(void)
{ {
if (!cpu_has_tsc || tsc_disable)
goto out_no_tsc;
if (cpu_has_tsc && tsc_khz && !tsc_disable) { cpu_khz = calculate_cpu_khz();
/* check blacklist */ tsc_khz = cpu_khz;
dmi_check_system(bad_tsc_dmi_table);
unsynchronized_tsc(); if (!cpu_khz)
check_geode_tsc_reliable(); goto out_no_tsc;
current_tsc_khz = tsc_khz;
clocksource_tsc.mult = clocksource_khz2mult(current_tsc_khz,
clocksource_tsc.shift);
/* lower the rating if we already know its unstable: */
if (check_tsc_unstable()) {
clocksource_tsc.rating = 0;
clocksource_tsc.flags &= ~CLOCK_SOURCE_IS_CONTINUOUS;
}
return clocksource_register(&clocksource_tsc); printk("Detected %lu.%03lu MHz processor.\n",
(unsigned long)cpu_khz / 1000,
(unsigned long)cpu_khz % 1000);
set_cyc2ns_scale(cpu_khz);
use_tsc_delay();
/* Check and install the TSC clocksource */
dmi_check_system(bad_tsc_dmi_table);
unsynchronized_tsc();
check_geode_tsc_reliable();
current_tsc_khz = tsc_khz;
clocksource_tsc.mult = clocksource_khz2mult(current_tsc_khz,
clocksource_tsc.shift);
/* lower the rating if we already know its unstable: */
if (check_tsc_unstable()) {
clocksource_tsc.rating = 0;
clocksource_tsc.flags &= ~CLOCK_SOURCE_IS_CONTINUOUS;
} }
clocksource_register(&clocksource_tsc);
return 0; return;
}
module_init(init_tsc_clocksource); out_no_tsc:
/*
* Set the tsc_disable flag if there's no TSC support, this
* makes it a fast flag for the kernel to see whether it
* should be using the TSC.
*/
tsc_disable = 1;
}
...@@ -12,6 +12,12 @@ ...@@ -12,6 +12,12 @@
#include <asm/timex.h> #include <asm/timex.h>
#include <asm/hpet.h> #include <asm/hpet.h>
#define HPET_MASK 0xFFFFFFFF
#define HPET_SHIFT 22
/* FSEC = 10^-15 NSEC = 10^-9 */
#define FSEC_PER_NSEC 1000000
int nohpet __initdata; int nohpet __initdata;
unsigned long hpet_address; unsigned long hpet_address;
...@@ -106,9 +112,31 @@ int hpet_timer_stop_set_go(unsigned long tick) ...@@ -106,9 +112,31 @@ int hpet_timer_stop_set_go(unsigned long tick)
return 0; return 0;
} }
static cycle_t read_hpet(void)
{
return (cycle_t)hpet_readl(HPET_COUNTER);
}
static cycle_t __vsyscall_fn vread_hpet(void)
{
return readl((void __iomem *)fix_to_virt(VSYSCALL_HPET) + 0xf0);
}
struct clocksource clocksource_hpet = {
.name = "hpet",
.rating = 250,
.read = read_hpet,
.mask = (cycle_t)HPET_MASK,
.mult = 0, /* set below */
.shift = HPET_SHIFT,
.flags = CLOCK_SOURCE_IS_CONTINUOUS,
.vread = vread_hpet,
};
int hpet_arch_init(void) int hpet_arch_init(void)
{ {
unsigned int id; unsigned int id;
u64 tmp;
if (!hpet_address) if (!hpet_address)
return -1; return -1;
...@@ -132,6 +160,22 @@ int hpet_arch_init(void) ...@@ -132,6 +160,22 @@ int hpet_arch_init(void)
hpet_use_timer = (id & HPET_ID_LEGSUP); hpet_use_timer = (id & HPET_ID_LEGSUP);
/*
* hpet period is in femto seconds per cycle
* so we need to convert this to ns/cyc units
* aproximated by mult/2^shift
*
* fsec/cyc * 1nsec/1000000fsec = nsec/cyc = mult/2^shift
* fsec/cyc * 1ns/1000000fsec * 2^shift = mult
* fsec/cyc * 2^shift * 1nsec/1000000fsec = mult
* (fsec/cyc << shift)/1000000 = mult
* (hpet_period << shift)/FSEC_PER_NSEC = mult
*/
tmp = (u64)hpet_period << HPET_SHIFT;
do_div(tmp, FSEC_PER_NSEC);
clocksource_hpet.mult = (u32)tmp;
clocksource_register(&clocksource_hpet);
return hpet_timer_stop_set_go(hpet_tick); return hpet_timer_stop_set_go(hpet_tick);
} }
...@@ -444,68 +488,3 @@ static int __init nohpet_setup(char *s) ...@@ -444,68 +488,3 @@ static int __init nohpet_setup(char *s)
} }
__setup("nohpet", nohpet_setup); __setup("nohpet", nohpet_setup);
#define HPET_MASK 0xFFFFFFFF
#define HPET_SHIFT 22
/* FSEC = 10^-15 NSEC = 10^-9 */
#define FSEC_PER_NSEC 1000000
static void *hpet_ptr;
static cycle_t read_hpet(void)
{
return (cycle_t)readl(hpet_ptr);
}
static cycle_t __vsyscall_fn vread_hpet(void)
{
return readl((void __iomem *)fix_to_virt(VSYSCALL_HPET) + 0xf0);
}
struct clocksource clocksource_hpet = {
.name = "hpet",
.rating = 250,
.read = read_hpet,
.mask = (cycle_t)HPET_MASK,
.mult = 0, /* set below */
.shift = HPET_SHIFT,
.flags = CLOCK_SOURCE_IS_CONTINUOUS,
.vread = vread_hpet,
};
static int __init init_hpet_clocksource(void)
{
unsigned long hpet_period;
void __iomem *hpet_base;
u64 tmp;
if (!hpet_address)
return -ENODEV;
/* calculate the hpet address: */
hpet_base = ioremap_nocache(hpet_address, HPET_MMAP_SIZE);
hpet_ptr = hpet_base + HPET_COUNTER;
/* calculate the frequency: */
hpet_period = readl(hpet_base + HPET_PERIOD);
/*
* hpet period is in femto seconds per cycle
* so we need to convert this to ns/cyc units
* aproximated by mult/2^shift
*
* fsec/cyc * 1nsec/1000000fsec = nsec/cyc = mult/2^shift
* fsec/cyc * 1ns/1000000fsec * 2^shift = mult
* fsec/cyc * 2^shift * 1nsec/1000000fsec = mult
* (fsec/cyc << shift)/1000000 = mult
* (hpet_period << shift)/FSEC_PER_NSEC = mult
*/
tmp = (u64)hpet_period << HPET_SHIFT;
do_div(tmp, FSEC_PER_NSEC);
clocksource_hpet.mult = (u32)tmp;
return clocksource_register(&clocksource_hpet);
}
module_init(init_hpet_clocksource);
...@@ -358,6 +358,8 @@ void __init time_init(void) ...@@ -358,6 +358,8 @@ void __init time_init(void)
set_cyc2ns_scale(cpu_khz); set_cyc2ns_scale(cpu_khz);
printk(KERN_INFO "time.c: Detected %d.%03d MHz processor.\n", printk(KERN_INFO "time.c: Detected %d.%03d MHz processor.\n",
cpu_khz / 1000, cpu_khz % 1000); cpu_khz / 1000, cpu_khz % 1000);
init_tsc_clocksource();
setup_irq(0, &irq0); setup_irq(0, &irq0);
} }
......
...@@ -210,7 +210,7 @@ void mark_tsc_unstable(void) ...@@ -210,7 +210,7 @@ void mark_tsc_unstable(void)
} }
EXPORT_SYMBOL_GPL(mark_tsc_unstable); EXPORT_SYMBOL_GPL(mark_tsc_unstable);
static int __init init_tsc_clocksource(void) void __init init_tsc_clocksource(void)
{ {
if (!notsc) { if (!notsc) {
clocksource_tsc.mult = clocksource_khz2mult(cpu_khz, clocksource_tsc.mult = clocksource_khz2mult(cpu_khz,
...@@ -218,9 +218,6 @@ static int __init init_tsc_clocksource(void) ...@@ -218,9 +218,6 @@ static int __init init_tsc_clocksource(void)
if (check_tsc_unstable()) if (check_tsc_unstable())
clocksource_tsc.rating = 0; clocksource_tsc.rating = 0;
return clocksource_register(&clocksource_tsc); clocksource_register(&clocksource_tsc);
} }
return 0;
} }
module_init(init_tsc_clocksource);
...@@ -214,4 +214,7 @@ pm_good: ...@@ -214,4 +214,7 @@ pm_good:
return clocksource_register(&clocksource_acpi_pm); return clocksource_register(&clocksource_acpi_pm);
} }
module_init(init_acpi_pm_clocksource); /* We use fs_initcall because we want the PCI fixups to have run
* but we still need to load before device_initcall
*/
fs_initcall(init_acpi_pm_clocksource);
...@@ -116,4 +116,4 @@ static int __init init_cyclone_clocksource(void) ...@@ -116,4 +116,4 @@ static int __init init_cyclone_clocksource(void)
return clocksource_register(&clocksource_cyclone); return clocksource_register(&clocksource_cyclone);
} }
module_init(init_cyclone_clocksource); arch_initcall(init_cyclone_clocksource);
...@@ -55,6 +55,7 @@ static __always_inline cycles_t get_cycles_sync(void) ...@@ -55,6 +55,7 @@ static __always_inline cycles_t get_cycles_sync(void)
extern void tsc_init(void); extern void tsc_init(void);
extern void mark_tsc_unstable(void); extern void mark_tsc_unstable(void);
extern int unsynchronized_tsc(void); extern int unsynchronized_tsc(void);
extern void init_tsc_clocksource(void);
/* /*
* Boot-time check whether the TSCs are synchronized across * Boot-time check whether the TSCs are synchronized across
......
...@@ -55,16 +55,18 @@ static DEFINE_SPINLOCK(clocksource_lock); ...@@ -55,16 +55,18 @@ static DEFINE_SPINLOCK(clocksource_lock);
static char override_name[32]; static char override_name[32];
static int finished_booting; static int finished_booting;
/* clocksource_done_booting - Called near the end of bootup /* clocksource_done_booting - Called near the end of core bootup
* *
* Hack to avoid lots of clocksource churn at boot time * Hack to avoid lots of clocksource churn at boot time.
* We use fs_initcall because we want this to start before
* device_initcall but after subsys_initcall.
*/ */
static int __init clocksource_done_booting(void) static int __init clocksource_done_booting(void)
{ {
finished_booting = 1; finished_booting = 1;
return 0; return 0;
} }
late_initcall(clocksource_done_booting); fs_initcall(clocksource_done_booting);
#ifdef CONFIG_CLOCKSOURCE_WATCHDOG #ifdef CONFIG_CLOCKSOURCE_WATCHDOG
static LIST_HEAD(watchdog_list); static LIST_HEAD(watchdog_list);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment