Commit 3a143125 authored by Linus Torvalds's avatar Linus Torvalds

Merge branch 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/x86/linux-2.6-x86

* 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/x86/linux-2.6-x86:
  x86: revert assign IRQs to hpet timer
  x86: tsc prevent time going backwards
  xen: Clear PG_pinned in release_{pt,pd}()
  xen: Do not pin/unpin PMD pages
  xen: refactor xen_{alloc,release}_{pt,pd}()
  x86, agpgart: scary messages are fortunately obsolete
  xen: fix grant table bug
  x86: fix breakage of vSMP irq operations
  x86: print message if nmi_watchdog=2 cannot be enabled
  x86: fix nmi_watchdog=2 on Pentium-D CPUs
parents a1aa758d 5761d64b
...@@ -652,9 +652,6 @@ static void probe_nmi_watchdog(void) ...@@ -652,9 +652,6 @@ static void probe_nmi_watchdog(void)
wd_ops = &p6_wd_ops; wd_ops = &p6_wd_ops;
break; break;
case 15: case 15:
if (boot_cpu_data.x86_model > 0x4)
return;
wd_ops = &p4_wd_ops; wd_ops = &p4_wd_ops;
break; break;
default: default:
...@@ -670,8 +667,10 @@ int lapic_watchdog_init(unsigned nmi_hz) ...@@ -670,8 +667,10 @@ int lapic_watchdog_init(unsigned nmi_hz)
{ {
if (!wd_ops) { if (!wd_ops) {
probe_nmi_watchdog(); probe_nmi_watchdog();
if (!wd_ops) if (!wd_ops) {
printk(KERN_INFO "NMI watchdog: CPU not supported\n");
return -1; return -1;
}
if (!wd_ops->reserve()) { if (!wd_ops->reserve()) {
printk(KERN_ERR printk(KERN_ERR
......
...@@ -133,13 +133,16 @@ static void hpet_reserve_platform_timers(unsigned long id) ...@@ -133,13 +133,16 @@ static void hpet_reserve_platform_timers(unsigned long id)
#ifdef CONFIG_HPET_EMULATE_RTC #ifdef CONFIG_HPET_EMULATE_RTC
hpet_reserve_timer(&hd, 1); hpet_reserve_timer(&hd, 1);
#endif #endif
hd.hd_irq[0] = HPET_LEGACY_8254; hd.hd_irq[0] = HPET_LEGACY_8254;
hd.hd_irq[1] = HPET_LEGACY_RTC; hd.hd_irq[1] = HPET_LEGACY_RTC;
for (i = 2; i < nrtimers; timer++, i++) for (i = 2; i < nrtimers; timer++, i++)
hd.hd_irq[i] = (timer->hpet_config & Tn_INT_ROUTE_CNF_MASK) >> hd.hd_irq[i] = (timer->hpet_config & Tn_INT_ROUTE_CNF_MASK) >>
Tn_INT_ROUTE_CNF_SHIFT; Tn_INT_ROUTE_CNF_SHIFT;
hpet_alloc(&hd); hpet_alloc(&hd);
} }
#else #else
static void hpet_reserve_platform_timers(unsigned long id) { } static void hpet_reserve_platform_timers(unsigned long id) { }
......
...@@ -615,8 +615,8 @@ static __init int init_k8_gatt(struct agp_kern_info *info) ...@@ -615,8 +615,8 @@ static __init int init_k8_gatt(struct agp_kern_info *info)
nommu: nommu:
/* Should not happen anymore */ /* Should not happen anymore */
printk(KERN_ERR "PCI-DMA: More than 4GB of RAM and no IOMMU\n" printk(KERN_WARNING "PCI-DMA: More than 4GB of RAM and no IOMMU\n"
KERN_ERR "PCI-DMA: 32bit PCI IO may malfunction.\n"); KERN_WARNING "falling back to iommu=soft.\n");
return -1; return -1;
} }
...@@ -692,9 +692,9 @@ void __init gart_iommu_init(void) ...@@ -692,9 +692,9 @@ void __init gart_iommu_init(void)
!gart_iommu_aperture || !gart_iommu_aperture ||
(no_agp && init_k8_gatt(&info) < 0)) { (no_agp && init_k8_gatt(&info) < 0)) {
if (end_pfn > MAX_DMA32_PFN) { if (end_pfn > MAX_DMA32_PFN) {
printk(KERN_ERR "WARNING more than 4GB of memory " printk(KERN_WARNING "More than 4GB of memory "
"but GART IOMMU not available.\n" "but GART IOMMU not available.\n"
KERN_ERR "WARNING 32bit PCI may malfunction.\n"); KERN_WARNING "falling back to iommu=soft.\n");
} }
return; return;
} }
......
...@@ -287,14 +287,27 @@ core_initcall(cpufreq_tsc); ...@@ -287,14 +287,27 @@ core_initcall(cpufreq_tsc);
/* clock source code */ /* clock source code */
static unsigned long current_tsc_khz = 0; static unsigned long current_tsc_khz = 0;
static struct clocksource clocksource_tsc;
/*
* We compare the TSC to the cycle_last value in the clocksource
* structure to avoid a nasty time-warp issue. This can be observed in
* a very small window right after one CPU updated cycle_last under
* xtime lock and the other CPU reads a TSC value which is smaller
* than the cycle_last reference value due to a TSC which is slighty
* behind. This delta is nowhere else observable, but in that case it
* results in a forward time jump in the range of hours due to the
* unsigned delta calculation of the time keeping core code, which is
* necessary to support wrapping clocksources like pm timer.
*/
static cycle_t read_tsc(void) static cycle_t read_tsc(void)
{ {
cycle_t ret; cycle_t ret;
rdtscll(ret); rdtscll(ret);
return ret; return ret >= clocksource_tsc.cycle_last ?
ret : clocksource_tsc.cycle_last;
} }
static struct clocksource clocksource_tsc = { static struct clocksource clocksource_tsc = {
......
...@@ -11,6 +11,7 @@ ...@@ -11,6 +11,7 @@
#include <asm/hpet.h> #include <asm/hpet.h>
#include <asm/timex.h> #include <asm/timex.h>
#include <asm/timer.h> #include <asm/timer.h>
#include <asm/vgtod.h>
static int notsc __initdata = 0; static int notsc __initdata = 0;
...@@ -290,18 +291,34 @@ int __init notsc_setup(char *s) ...@@ -290,18 +291,34 @@ int __init notsc_setup(char *s)
__setup("notsc", notsc_setup); __setup("notsc", notsc_setup);
static struct clocksource clocksource_tsc;
/* clock source code: */ /*
* We compare the TSC to the cycle_last value in the clocksource
* structure to avoid a nasty time-warp. This can be observed in a
* very small window right after one CPU updated cycle_last under
* xtime/vsyscall_gtod lock and the other CPU reads a TSC value which
* is smaller than the cycle_last reference value due to a TSC which
* is slighty behind. This delta is nowhere else observable, but in
* that case it results in a forward time jump in the range of hours
* due to the unsigned delta calculation of the time keeping core
* code, which is necessary to support wrapping clocksources like pm
* timer.
*/
static cycle_t read_tsc(void) static cycle_t read_tsc(void)
{ {
cycle_t ret = (cycle_t)get_cycles(); cycle_t ret = (cycle_t)get_cycles();
return ret;
return ret >= clocksource_tsc.cycle_last ?
ret : clocksource_tsc.cycle_last;
} }
static cycle_t __vsyscall_fn vread_tsc(void) static cycle_t __vsyscall_fn vread_tsc(void)
{ {
cycle_t ret = (cycle_t)vget_cycles(); cycle_t ret = (cycle_t)vget_cycles();
return ret;
return ret >= __vsyscall_gtod_data.clock.cycle_last ?
ret : __vsyscall_gtod_data.clock.cycle_last;
} }
static struct clocksource clocksource_tsc = { static struct clocksource clocksource_tsc = {
......
...@@ -667,10 +667,10 @@ static void xen_release_pt_init(u32 pfn) ...@@ -667,10 +667,10 @@ static void xen_release_pt_init(u32 pfn)
make_lowmem_page_readwrite(__va(PFN_PHYS(pfn))); make_lowmem_page_readwrite(__va(PFN_PHYS(pfn)));
} }
static void pin_pagetable_pfn(unsigned level, unsigned long pfn) static void pin_pagetable_pfn(unsigned cmd, unsigned long pfn)
{ {
struct mmuext_op op; struct mmuext_op op;
op.cmd = level; op.cmd = cmd;
op.arg1.mfn = pfn_to_mfn(pfn); op.arg1.mfn = pfn_to_mfn(pfn);
if (HYPERVISOR_mmuext_op(&op, 1, NULL, DOMID_SELF)) if (HYPERVISOR_mmuext_op(&op, 1, NULL, DOMID_SELF))
BUG(); BUG();
...@@ -687,7 +687,8 @@ static void xen_alloc_ptpage(struct mm_struct *mm, u32 pfn, unsigned level) ...@@ -687,7 +687,8 @@ static void xen_alloc_ptpage(struct mm_struct *mm, u32 pfn, unsigned level)
if (!PageHighMem(page)) { if (!PageHighMem(page)) {
make_lowmem_page_readonly(__va(PFN_PHYS(pfn))); make_lowmem_page_readonly(__va(PFN_PHYS(pfn)));
pin_pagetable_pfn(level, pfn); if (level == PT_PTE)
pin_pagetable_pfn(MMUEXT_PIN_L1_TABLE, pfn);
} else } else
/* make sure there are no stray mappings of /* make sure there are no stray mappings of
this page */ this page */
...@@ -697,27 +698,39 @@ static void xen_alloc_ptpage(struct mm_struct *mm, u32 pfn, unsigned level) ...@@ -697,27 +698,39 @@ static void xen_alloc_ptpage(struct mm_struct *mm, u32 pfn, unsigned level)
static void xen_alloc_pt(struct mm_struct *mm, u32 pfn) static void xen_alloc_pt(struct mm_struct *mm, u32 pfn)
{ {
xen_alloc_ptpage(mm, pfn, MMUEXT_PIN_L1_TABLE); xen_alloc_ptpage(mm, pfn, PT_PTE);
} }
static void xen_alloc_pd(struct mm_struct *mm, u32 pfn) static void xen_alloc_pd(struct mm_struct *mm, u32 pfn)
{ {
xen_alloc_ptpage(mm, pfn, MMUEXT_PIN_L2_TABLE); xen_alloc_ptpage(mm, pfn, PT_PMD);
} }
/* This should never happen until we're OK to use struct page */ /* This should never happen until we're OK to use struct page */
static void xen_release_pt(u32 pfn) static void xen_release_ptpage(u32 pfn, unsigned level)
{ {
struct page *page = pfn_to_page(pfn); struct page *page = pfn_to_page(pfn);
if (PagePinned(page)) { if (PagePinned(page)) {
if (!PageHighMem(page)) { if (!PageHighMem(page)) {
if (level == PT_PTE)
pin_pagetable_pfn(MMUEXT_UNPIN_TABLE, pfn); pin_pagetable_pfn(MMUEXT_UNPIN_TABLE, pfn);
make_lowmem_page_readwrite(__va(PFN_PHYS(pfn))); make_lowmem_page_readwrite(__va(PFN_PHYS(pfn)));
} }
ClearPagePinned(page);
} }
} }
static void xen_release_pt(u32 pfn)
{
xen_release_ptpage(pfn, PT_PTE);
}
static void xen_release_pd(u32 pfn)
{
xen_release_ptpage(pfn, PT_PMD);
}
#ifdef CONFIG_HIGHPTE #ifdef CONFIG_HIGHPTE
static void *xen_kmap_atomic_pte(struct page *page, enum km_type type) static void *xen_kmap_atomic_pte(struct page *page, enum km_type type)
{ {
...@@ -838,7 +851,7 @@ static __init void xen_pagetable_setup_done(pgd_t *base) ...@@ -838,7 +851,7 @@ static __init void xen_pagetable_setup_done(pgd_t *base)
pv_mmu_ops.alloc_pt = xen_alloc_pt; pv_mmu_ops.alloc_pt = xen_alloc_pt;
pv_mmu_ops.alloc_pd = xen_alloc_pd; pv_mmu_ops.alloc_pd = xen_alloc_pd;
pv_mmu_ops.release_pt = xen_release_pt; pv_mmu_ops.release_pt = xen_release_pt;
pv_mmu_ops.release_pd = xen_release_pt; pv_mmu_ops.release_pd = xen_release_pd;
pv_mmu_ops.set_pte = xen_set_pte; pv_mmu_ops.set_pte = xen_set_pte;
setup_shared_info(); setup_shared_info();
......
...@@ -310,13 +310,6 @@ pgd_t xen_make_pgd(unsigned long pgd) ...@@ -310,13 +310,6 @@ pgd_t xen_make_pgd(unsigned long pgd)
} }
#endif /* CONFIG_X86_PAE */ #endif /* CONFIG_X86_PAE */
enum pt_level {
PT_PGD,
PT_PUD,
PT_PMD,
PT_PTE
};
/* /*
(Yet another) pagetable walker. This one is intended for pinning a (Yet another) pagetable walker. This one is intended for pinning a
pagetable. This means that it walks a pagetable and calls the pagetable. This means that it walks a pagetable and calls the
......
...@@ -3,6 +3,13 @@ ...@@ -3,6 +3,13 @@
#include <linux/linkage.h> #include <linux/linkage.h>
#include <asm/page.h> #include <asm/page.h>
enum pt_level {
PT_PGD,
PT_PUD,
PT_PMD,
PT_PTE
};
/* /*
* Page-directory addresses above 4GB do not fit into architectural %cr3. * Page-directory addresses above 4GB do not fit into architectural %cr3.
* When accessing %cr3, or equivalent field in vcpu_guest_context, guests * When accessing %cr3, or equivalent field in vcpu_guest_context, guests
......
...@@ -731,14 +731,14 @@ static unsigned long hpet_calibrate(struct hpets *hpetp) ...@@ -731,14 +731,14 @@ static unsigned long hpet_calibrate(struct hpets *hpetp)
int hpet_alloc(struct hpet_data *hdp) int hpet_alloc(struct hpet_data *hdp)
{ {
u64 cap, mcfg, hpet_config; u64 cap, mcfg;
struct hpet_dev *devp; struct hpet_dev *devp;
u32 i, ntimer, irq; u32 i, ntimer;
struct hpets *hpetp; struct hpets *hpetp;
size_t siz; size_t siz;
struct hpet __iomem *hpet; struct hpet __iomem *hpet;
static struct hpets *last = NULL; static struct hpets *last = NULL;
unsigned long period, irq_bitmap; unsigned long period;
unsigned long long temp; unsigned long long temp;
/* /*
...@@ -765,47 +765,11 @@ int hpet_alloc(struct hpet_data *hdp) ...@@ -765,47 +765,11 @@ int hpet_alloc(struct hpet_data *hdp)
hpetp->hp_hpet_phys = hdp->hd_phys_address; hpetp->hp_hpet_phys = hdp->hd_phys_address;
hpetp->hp_ntimer = hdp->hd_nirqs; hpetp->hp_ntimer = hdp->hd_nirqs;
hpet = hpetp->hp_hpet;
/* Assign IRQs statically for legacy devices */
hpetp->hp_dev[0].hd_hdwirq = hdp->hd_irq[0];
hpetp->hp_dev[1].hd_hdwirq = hdp->hd_irq[1];
/* Assign IRQs dynamically for the others */
for (i = 2, devp = &hpetp->hp_dev[2]; i < hdp->hd_nirqs; i++, devp++) {
struct hpet_timer __iomem *timer;
timer = &hpet->hpet_timers[devp - hpetp->hp_dev]; for (i = 0; i < hdp->hd_nirqs; i++)
/* Check if there's already an IRQ assigned to the timer */
if (hdp->hd_irq[i]) {
hpetp->hp_dev[i].hd_hdwirq = hdp->hd_irq[i]; hpetp->hp_dev[i].hd_hdwirq = hdp->hd_irq[i];
continue;
}
hpet_config = readq(&timer->hpet_config); hpet = hpetp->hp_hpet;
irq_bitmap = (hpet_config & Tn_INT_ROUTE_CAP_MASK)
>> Tn_INT_ROUTE_CAP_SHIFT;
if (!irq_bitmap)
irq = 0; /* No valid IRQ Assignable */
else {
irq = find_first_bit(&irq_bitmap, 32);
do {
hpet_config |= irq << Tn_INT_ROUTE_CNF_SHIFT;
writeq(hpet_config, &timer->hpet_config);
/*
* Verify whether we have written a valid
* IRQ number by reading it back again
*/
hpet_config = readq(&timer->hpet_config);
if (irq == (hpet_config & Tn_INT_ROUTE_CNF_MASK)
>> Tn_INT_ROUTE_CNF_SHIFT)
break; /* Success */
} while ((irq = (find_next_bit(&irq_bitmap, 32, irq))));
}
hpetp->hp_dev[i].hd_hdwirq = irq;
}
cap = readq(&hpet->hpet_cap); cap = readq(&hpet->hpet_cap);
...@@ -836,8 +800,7 @@ int hpet_alloc(struct hpet_data *hdp) ...@@ -836,8 +800,7 @@ int hpet_alloc(struct hpet_data *hdp)
hpetp->hp_which, hdp->hd_phys_address, hpetp->hp_which, hdp->hd_phys_address,
hpetp->hp_ntimer > 1 ? "s" : ""); hpetp->hp_ntimer > 1 ? "s" : "");
for (i = 0; i < hpetp->hp_ntimer; i++) for (i = 0; i < hpetp->hp_ntimer; i++)
printk("%s %d", i > 0 ? "," : "", printk("%s %d", i > 0 ? "," : "", hdp->hd_irq[i]);
hpetp->hp_dev[i].hd_hdwirq);
printk("\n"); printk("\n");
printk(KERN_INFO "hpet%u: %u %d-bit timers, %Lu Hz\n", printk(KERN_INFO "hpet%u: %u %d-bit timers, %Lu Hz\n",
......
...@@ -381,11 +381,15 @@ EXPORT_SYMBOL_GPL(gnttab_cancel_free_callback); ...@@ -381,11 +381,15 @@ EXPORT_SYMBOL_GPL(gnttab_cancel_free_callback);
static int grow_gnttab_list(unsigned int more_frames) static int grow_gnttab_list(unsigned int more_frames)
{ {
unsigned int new_nr_grant_frames, extra_entries, i; unsigned int new_nr_grant_frames, extra_entries, i;
unsigned int nr_glist_frames, new_nr_glist_frames;
new_nr_grant_frames = nr_grant_frames + more_frames; new_nr_grant_frames = nr_grant_frames + more_frames;
extra_entries = more_frames * GREFS_PER_GRANT_FRAME; extra_entries = more_frames * GREFS_PER_GRANT_FRAME;
for (i = nr_grant_frames; i < new_nr_grant_frames; i++) { nr_glist_frames = (nr_grant_frames * GREFS_PER_GRANT_FRAME + RPP - 1) / RPP;
new_nr_glist_frames =
(new_nr_grant_frames * GREFS_PER_GRANT_FRAME + RPP - 1) / RPP;
for (i = nr_glist_frames; i < new_nr_glist_frames; i++) {
gnttab_list[i] = (grant_ref_t *)__get_free_page(GFP_ATOMIC); gnttab_list[i] = (grant_ref_t *)__get_free_page(GFP_ATOMIC);
if (!gnttab_list[i]) if (!gnttab_list[i])
goto grow_nomem; goto grow_nomem;
...@@ -407,7 +411,7 @@ static int grow_gnttab_list(unsigned int more_frames) ...@@ -407,7 +411,7 @@ static int grow_gnttab_list(unsigned int more_frames)
return 0; return 0;
grow_nomem: grow_nomem:
for ( ; i >= nr_grant_frames; i--) for ( ; i >= nr_glist_frames; i--)
free_page((unsigned long) gnttab_list[i]); free_page((unsigned long) gnttab_list[i]);
return -ENOMEM; return -ENOMEM;
} }
...@@ -530,7 +534,7 @@ static int gnttab_expand(unsigned int req_entries) ...@@ -530,7 +534,7 @@ static int gnttab_expand(unsigned int req_entries)
static int __devinit gnttab_init(void) static int __devinit gnttab_init(void)
{ {
int i; int i;
unsigned int max_nr_glist_frames; unsigned int max_nr_glist_frames, nr_glist_frames;
unsigned int nr_init_grefs; unsigned int nr_init_grefs;
if (!is_running_on_xen()) if (!is_running_on_xen())
...@@ -543,15 +547,15 @@ static int __devinit gnttab_init(void) ...@@ -543,15 +547,15 @@ static int __devinit gnttab_init(void)
* grant reference free list on the current hypervisor. * grant reference free list on the current hypervisor.
*/ */
max_nr_glist_frames = (boot_max_nr_grant_frames * max_nr_glist_frames = (boot_max_nr_grant_frames *
GREFS_PER_GRANT_FRAME / GREFS_PER_GRANT_FRAME / RPP);
(PAGE_SIZE / sizeof(grant_ref_t)));
gnttab_list = kmalloc(max_nr_glist_frames * sizeof(grant_ref_t *), gnttab_list = kmalloc(max_nr_glist_frames * sizeof(grant_ref_t *),
GFP_KERNEL); GFP_KERNEL);
if (gnttab_list == NULL) if (gnttab_list == NULL)
return -ENOMEM; return -ENOMEM;
for (i = 0; i < nr_grant_frames; i++) { nr_glist_frames = (nr_grant_frames * GREFS_PER_GRANT_FRAME + RPP - 1) / RPP;
for (i = 0; i < nr_glist_frames; i++) {
gnttab_list[i] = (grant_ref_t *)__get_free_page(GFP_KERNEL); gnttab_list[i] = (grant_ref_t *)__get_free_page(GFP_KERNEL);
if (gnttab_list[i] == NULL) if (gnttab_list[i] == NULL)
goto ini_nomem; goto ini_nomem;
......
...@@ -70,6 +70,26 @@ static inline void raw_local_irq_restore(unsigned long flags) ...@@ -70,6 +70,26 @@ static inline void raw_local_irq_restore(unsigned long flags)
native_restore_fl(flags); native_restore_fl(flags);
} }
#ifdef CONFIG_X86_VSMP
/*
* Interrupt control for the VSMP architecture:
*/
static inline void raw_local_irq_disable(void)
{
unsigned long flags = __raw_local_save_flags();
raw_local_irq_restore((flags & ~X86_EFLAGS_IF) | X86_EFLAGS_AC);
}
static inline void raw_local_irq_enable(void)
{
unsigned long flags = __raw_local_save_flags();
raw_local_irq_restore((flags | X86_EFLAGS_IF) & (~X86_EFLAGS_AC));
}
#else
static inline void raw_local_irq_disable(void) static inline void raw_local_irq_disable(void)
{ {
native_irq_disable(); native_irq_disable();
...@@ -80,6 +100,8 @@ static inline void raw_local_irq_enable(void) ...@@ -80,6 +100,8 @@ static inline void raw_local_irq_enable(void)
native_irq_enable(); native_irq_enable();
} }
#endif
/* /*
* Used in the idle loop; sti takes one instruction cycle * Used in the idle loop; sti takes one instruction cycle
* to complete: * to complete:
...@@ -137,10 +159,17 @@ static inline unsigned long __raw_local_irq_save(void) ...@@ -137,10 +159,17 @@ static inline unsigned long __raw_local_irq_save(void)
#define raw_local_irq_save(flags) \ #define raw_local_irq_save(flags) \
do { (flags) = __raw_local_irq_save(); } while (0) do { (flags) = __raw_local_irq_save(); } while (0)
#ifdef CONFIG_X86_VSMP
static inline int raw_irqs_disabled_flags(unsigned long flags)
{
return !(flags & X86_EFLAGS_IF) || (flags & X86_EFLAGS_AC);
}
#else
static inline int raw_irqs_disabled_flags(unsigned long flags) static inline int raw_irqs_disabled_flags(unsigned long flags)
{ {
return !(flags & X86_EFLAGS_IF); return !(flags & X86_EFLAGS_IF);
} }
#endif
static inline int raw_irqs_disabled(void) static inline int raw_irqs_disabled(void)
{ {
......
...@@ -64,7 +64,7 @@ struct hpet { ...@@ -64,7 +64,7 @@ struct hpet {
*/ */
#define Tn_INT_ROUTE_CAP_MASK (0xffffffff00000000ULL) #define Tn_INT_ROUTE_CAP_MASK (0xffffffff00000000ULL)
#define Tn_INT_ROUTE_CAP_SHIFT (32UL) #define Tn_INI_ROUTE_CAP_SHIFT (32UL)
#define Tn_FSB_INT_DELCAP_MASK (0x8000UL) #define Tn_FSB_INT_DELCAP_MASK (0x8000UL)
#define Tn_FSB_INT_DELCAP_SHIFT (15) #define Tn_FSB_INT_DELCAP_SHIFT (15)
#define Tn_FSB_EN_CNF_MASK (0x4000UL) #define Tn_FSB_EN_CNF_MASK (0x4000UL)
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment