[v1 8/9] x86/tsc: tsc early
From: Pavel Tatashin
Date: Wed Mar 22 2017 - 16:19:11 EST
tsc_early_init():
Use verious methods to determine the availability of TSC feature and its
frequency early in boot, and if that is possible initialize TSC and also
call sched_clock_early_init() to be able to get timestamps early in boot.
tsc_early_fini()
Implement the finish part of early tsc feature, print message about the
offset, which can be useful to findout how much time was spent in post and
boot manager, and also call sched_clock_early_fini() to let sched clock
know that
sched_clock_early():
TSC based implementation of weak function that is defined in sched clock.
Signed-off-by: Pavel Tatashin <pasha.tatashin@xxxxxxxxxx>
---
arch/x86/include/asm/tsc.h | 4 ++
arch/x86/kernel/tsc.c | 107 ++++++++++++++++++++++++++++++++++++++++++++
2 files changed, 111 insertions(+), 0 deletions(-)
diff --git a/arch/x86/include/asm/tsc.h b/arch/x86/include/asm/tsc.h
index 893de0c..a8c7f2e 100644
--- a/arch/x86/include/asm/tsc.h
+++ b/arch/x86/include/asm/tsc.h
@@ -50,11 +50,15 @@ static inline cycles_t get_cycles(void)
extern void tsc_verify_tsc_adjust(bool resume);
extern void check_tsc_sync_source(int cpu);
extern void check_tsc_sync_target(void);
+void tsc_early_init(void);
+void tsc_early_fini(void);
#else
static inline bool tsc_store_and_check_tsc_adjust(bool bootcpu) { return false; }
static inline void tsc_verify_tsc_adjust(bool resume) { }
static inline void check_tsc_sync_source(int cpu) { }
static inline void check_tsc_sync_target(void) { }
+static inline void tsc_early_init(void) { }
+static inline void tsc_early_fini(void) { }
#endif
extern int notsc_setup(char *);
diff --git a/arch/x86/kernel/tsc.c b/arch/x86/kernel/tsc.c
index 58bd575..67ecddf 100644
--- a/arch/x86/kernel/tsc.c
+++ b/arch/x86/kernel/tsc.c
@@ -25,10 +25,13 @@
#include <asm/geode.h>
#include <asm/apic.h>
#include <asm/intel-family.h>
+#include <asm/cpu.h>
/* CPUID 15H TSC/Crystal ratio, plus optionally Crystal Hz */
#define CPUID_TSC_LEAF 0x15
+static struct cyc2ns_data __read_mostly cyc2ns_early;
+
unsigned int __read_mostly cpu_khz; /* TSC clocks / usec, not used here */
EXPORT_SYMBOL(cpu_khz);
@@ -290,6 +293,16 @@ static void set_cyc2ns_scale(unsigned long khz, int cpu)
sched_clock_idle_wakeup_event(0);
local_irq_restore(flags);
}
+
+u64 sched_clock_early(void)
+{
+ u64 ns;
+
+ ns = mul_u64_u32_shr(rdtsc(), cyc2ns_early.cyc2ns_mul,
+ cyc2ns_early.cyc2ns_shift);
+ return ns + cyc2ns_early.cyc2ns_offset;
+}
+
/*
* Scheduler clock - returns current time in nanosec units.
*/
@@ -1365,6 +1378,100 @@ static int __init init_tsc_clocksource(void)
*/
device_initcall(init_tsc_clocksource);
+#ifdef CONFIG_X86_TSC
+
+/* Determine if tsc is invariant early in boot */
+static bool __init tsc_invariant_early(void)
+{
+ unsigned int ext_cpuid_level, tsc_flag;
+
+ /* Get extended CPUID level */
+ ext_cpuid_level = cpuid_eax(0x80000000);
+ if (ext_cpuid_level < 0x80000007)
+ return false;
+
+ /* get field with invariant TSC flag */
+ tsc_flag = cpuid_edx(0x80000007);
+ if (!(tsc_flag & (1 << 8)))
+ return false;
+
+ return true;
+}
+
+/*
+ * Determine if we can use TSC early in boot. On larger machines early boot can
+ * take a significant amount of time, therefore, for observability reasons, and
+ * also to avoid regressions it is important to have timestamps during the whole
+ * boot process.
+ */
+void __init tsc_early_init(void)
+{
+ int vendor, model, family, cpuid_level;
+ unsigned int sig, khz;
+ u64 tsc_now;
+
+ /*
+ * Should we disable early timestamps on platforms without invariant
+ * TSC?
+ *
+ * On the one hand invariant TSC guarantees, that early timestamps run
+ * only on the latest hardware (Nehalem and later), but on the other
+ * hand accuracy wise, non-invariant timestamps should be OK,
+ * because during early boot power management features are not used.
+ * ---
+ * For now we disable invariant TSC for early boot.
+ */
+ if (!tsc_invariant_early())
+ return;
+
+ cpuid_level = cpuid_eax(0);
+ sig = cpuid_eax(1);
+ model = x86_model(sig);
+ family = x86_family(sig);
+ vendor = get_x86_vendor_early();
+
+ /*
+ * Try several methods to get TSC frequency, if fail, return false
+ * otherwise setup mult and shift values to convert ticks to nanoseconds
+ * efficiently.
+ */
+ khz = 0;
+ if (vendor == X86_VENDOR_INTEL && cpuid_level >= CPUID_TSC_LEAF)
+ khz = calibrate_tsc_early(model);
+
+ if (khz == 0)
+ khz = cpu_khz_from_cpuid_early(vendor, cpuid_level);
+
+ if (khz == 0)
+ khz = cpu_khz_from_msr_early(vendor, family, model);
+
+ if (khz == 0)
+ khz = quick_pit_calibrate(true);
+
+ if (khz == 0)
+ return;
+
+ tsc_now = rdtsc();
+ clocks_calc_mult_shift(&cyc2ns_early.cyc2ns_mul,
+ &cyc2ns_early.cyc2ns_shift,
+ khz, NSEC_PER_MSEC, 0);
+ cyc2ns_early.cyc2ns_offset = -sched_clock_early();
+ sched_clock_early_init();
+}
+
+void __init tsc_early_fini(void)
+{
+ /* We did not have early sched clock if multiplier is 0 */
+ if (cyc2ns_early.cyc2ns_mul == 0)
+ return;
+
+ sched_clock_early_fini();
+ pr_info("sched clock early is finished, offset [%lld.%09llds]\n",
+ -cyc2ns_early.cyc2ns_offset / NSEC_PER_SEC,
+ -cyc2ns_early.cyc2ns_offset % NSEC_PER_SEC);
+}
+#endif /* CONFIG_X86_TSC */
+
void __init tsc_init(void)
{
u64 lpj;
--
1.7.1