Lines Matching +full:calibration +full:- +full:data

1 // SPDX-License-Identifier: GPL-2.0-only
62 struct cyc2ns_data data[2]; /* 0 + 2*16 = 32 */ member
75 __always_inline void __cyc2ns_read(struct cyc2ns_data *data) in __cyc2ns_read() argument
83 data->cyc2ns_offset = this_cpu_read(cyc2ns.data[idx].cyc2ns_offset); in __cyc2ns_read()
84 data->cyc2ns_mul = this_cpu_read(cyc2ns.data[idx].cyc2ns_mul); in __cyc2ns_read()
85 data->cyc2ns_shift = this_cpu_read(cyc2ns.data[idx].cyc2ns_shift); in __cyc2ns_read()
90 __always_inline void cyc2ns_read_begin(struct cyc2ns_data *data) in cyc2ns_read_begin() argument
93 __cyc2ns_read(data); in cyc2ns_read_begin()
116 * cyc2ns_scale needs to be a 32-bit value so that 32-bit multiplication
117 * (64-bit result) can be used.
122 * -[email protected] "math is hard, lets go shopping!"
127 struct cyc2ns_data data; in __cycles_2_ns() local
130 __cyc2ns_read(&data); in __cycles_2_ns()
132 ns = data.cyc2ns_offset; in __cycles_2_ns()
133 ns += mul_u64_u32_shr(cyc, data.cyc2ns_mul, data.cyc2ns_shift); in __cycles_2_ns()
150 struct cyc2ns_data data; in __set_cyc2ns_scale() local
160 clocks_calc_mult_shift(&data.cyc2ns_mul, &data.cyc2ns_shift, khz, in __set_cyc2ns_scale()
166 * conversion algorithm shifting a 32-bit value (now specifies a 64-bit in __set_cyc2ns_scale()
167 * value) - refer perf_event_mmap_page documentation in perf_event.h. in __set_cyc2ns_scale()
169 if (data.cyc2ns_shift == 32) { in __set_cyc2ns_scale()
170 data.cyc2ns_shift = 31; in __set_cyc2ns_scale()
171 data.cyc2ns_mul >>= 1; in __set_cyc2ns_scale()
174 data.cyc2ns_offset = ns_now - in __set_cyc2ns_scale()
175 mul_u64_u32_shr(tsc_now, data.cyc2ns_mul, data.cyc2ns_shift); in __set_cyc2ns_scale()
179 write_seqcount_latch_begin(&c2n->seq); in __set_cyc2ns_scale()
180 c2n->data[0] = data; in __set_cyc2ns_scale()
181 write_seqcount_latch(&c2n->seq); in __set_cyc2ns_scale()
182 c2n->data[1] = data; in __set_cyc2ns_scale()
183 write_seqcount_latch_end(&c2n->seq); in __set_cyc2ns_scale()
207 seqcount_latch_init(&c2n->seq); in cyc2ns_init_boot_cpu()
220 struct cyc2ns_data *data = c2n->data; in cyc2ns_init_secondary_cpus() local
224 seqcount_latch_init(&c2n->seq); in cyc2ns_init_secondary_cpus()
226 c2n->data[0] = data[0]; in cyc2ns_init_secondary_cpus()
227 c2n->data[1] = data[1]; in cyc2ns_init_secondary_cpus()
233 * Scheduler clock - returns current time in nanosec units.
254 return (jiffies_64 - INITIAL_JIFFIES) * (1000000000 / HZ); in native_sched_clock()
370 if ((t2 - t1) < thresh) in tsc_read_refs()
385 hpet2 -= hpet1; in calc_hpet_ref()
405 pm2 -= pm1; in calc_pmtimer_ref()
468 delta = t2 - tsc; in pit_calibrate_tsc()
490 delta = t2 - t1; in pit_calibrate_tsc()
498 * non-virtualized hardware.
502 * - the PIT is running at roughly 1.19MHz
504 * - each IO is going to take about 1us on real hardware,
507 * update - anything else implies a unacceptably slow CPU
508 * or PIT for the fast calibration to work.
510 * - with 256 PIT ticks to read the value, we have 214us to
514 * - We're doing 2 reads per loop (LSB, MSB), and we expect
519 * - if the PIT is stuck, and we see *many* more reads, we
548 *deltap = get_cycles() - prev_tsc; in pit_expect_msb()
580 * Counter 2, mode 0 (one-shot), binary count in quick_pit_calibrate()
584 * final output frequency as a decrement-by-one), in quick_pit_calibrate()
597 * to do that is to just read back the 16-bit counter in quick_pit_calibrate()
604 if (!pit_expect_msb(0xff-i, &delta, &d2)) in quick_pit_calibrate()
607 delta -= tsc; in quick_pit_calibrate()
630 if (!pit_verify_msb(0xfe - i)) in quick_pit_calibrate()
635 pr_info("Fast TSC calibration failed\n"); in quick_pit_calibrate()
648 * kHz = ticks / time-in-seconds / 1000; in quick_pit_calibrate()
649 * kHz = (t2 - t1) / (I * 256 / PIT_TICK_RATE) / 1000 in quick_pit_calibrate()
650 * kHz = ((t2 - t1) * PIT_TICK_RATE) / (I * 256 * 1000) in quick_pit_calibrate()
654 pr_info("Fast TSC calibration using PIT\n"); in quick_pit_calibrate()
659 * native_calibrate_tsc - determine TSC frequency
765 * Run 5 calibration loops to get the lowest frequency value in pit_hpet_ptimer_calibrate_cpu()
766 * (the best estimate). We use two different calibration modes in pit_hpet_ptimer_calibrate_cpu()
785 * calibration delay loop as we have to wait for a certain in pit_hpet_ptimer_calibrate_cpu()
800 * calibration, which will take at least 50ms, and in pit_hpet_ptimer_calibrate_cpu()
809 /* Pick the lowest PIT TSC calibration so far */ in pit_hpet_ptimer_calibrate_cpu()
820 tsc2 = (tsc2 - tsc1) * 1000000LL; in pit_hpet_ptimer_calibrate_cpu()
833 * If both calibration results are inside a 10% window in pit_hpet_ptimer_calibrate_cpu()
834 * then we can be sure, that the calibration in pit_hpet_ptimer_calibrate_cpu()
839 pr_info("PIT calibration matches %s. %d loops\n", in pit_hpet_ptimer_calibrate_cpu()
872 pr_warn("HPET/PMTIMER calibration failed\n"); in pit_hpet_ptimer_calibrate_cpu()
877 pr_info("using %s reference calibration\n", in pit_hpet_ptimer_calibrate_cpu()
883 /* We don't have an alternative source, use the PIT calibration value */ in pit_hpet_ptimer_calibrate_cpu()
885 pr_info("Using PIT calibration value\n"); in pit_hpet_ptimer_calibrate_cpu()
889 /* The alternative source failed, use the PIT calibration value */ in pit_hpet_ptimer_calibrate_cpu()
891 pr_warn("HPET/PMTIMER calibration failed. Using PIT calibration.\n"); in pit_hpet_ptimer_calibrate_cpu()
896 * The calibration values differ too much. In doubt, we use in pit_hpet_ptimer_calibrate_cpu()
900 pr_warn("PIT calibration deviates from %s: %lu %lu\n", in pit_hpet_ptimer_calibrate_cpu()
902 pr_info("Using PIT calibration value\n"); in pit_hpet_ptimer_calibrate_cpu()
907 * native_calibrate_cpu_early - can calibrate the cpu early in boot
925 * native_calibrate_cpu - calibrate the cpu
949 else if (abs(cpu_khz - tsc_khz) * 10 > tsc_khz) in recalibrate_cpu_khz()
990 * data fields. in tsc_restore_sched_clock_state()
993 this_cpu_write(cyc2ns.data[0].cyc2ns_offset, 0); in tsc_restore_sched_clock_state()
994 this_cpu_write(cyc2ns.data[1].cyc2ns_offset, 0); in tsc_restore_sched_clock_state()
996 offset = cyc2ns_suspend - sched_clock(); in tsc_restore_sched_clock_state()
999 per_cpu(cyc2ns.data[0].cyc2ns_offset, cpu) = offset; in tsc_restore_sched_clock_state()
1000 per_cpu(cyc2ns.data[1].cyc2ns_offset, cpu) = offset; in tsc_restore_sched_clock_state()
1023 void *data) in time_cpufreq_notifier() argument
1025 struct cpufreq_freqs *freq = data; in time_cpufreq_notifier()
1033 ref_freq = freq->old; in time_cpufreq_notifier()
1038 if ((val == CPUFREQ_PRECHANGE && freq->old < freq->new) || in time_cpufreq_notifier()
1039 (val == CPUFREQ_POSTCHANGE && freq->old > freq->new)) { in time_cpufreq_notifier()
1041 cpufreq_scale(loops_per_jiffy_ref, ref_freq, freq->new); in time_cpufreq_notifier()
1043 tsc_khz = cpufreq_scale(tsc_khz_ref, ref_freq, freq->new); in time_cpufreq_notifier()
1044 if (!(freq->flags & CPUFREQ_CONST_LOOPS)) in time_cpufreq_notifier()
1047 set_cyc2ns_scale(tsc_khz, freq->policy->cpu, rdtsc()); in time_cpufreq_notifier()
1085 * Don't enable ART in a VM, non-stop TSC and TSC_ADJUST required, in detect_art()
1117 * structure to avoid a nasty time-warp. This can be observed in a
1128 * checking the result of read_tsc() - cycle_last for being negative.
1167 .name = "tsc-early",
1244 /* Geode_LX - the OLPC CPU has a very reliable TSC */ in check_system_tsc_reliable()
1254 * - TSC running at constant frequency in check_system_tsc_reliable()
1255 * - TSC which does not stop in C-States in check_system_tsc_reliable()
1256 * - the TSC_ADJUST register which allows to detect even minimal in check_system_tsc_reliable()
1258 * - not more than four packages in check_system_tsc_reliable()
1302 * tsc_refine_calibration_work - Further refine tsc freq calibration
1308 * process while this longer calibration is done.
1310 * If there are any calibration anomalies (too many SMIs, etc),
1311 * or the refined calibration is off by 1% of the fast early
1312 * calibration, we throw out the new calibration and use the
1313 * early calibration.
1354 delta = tsc_stop - tsc_start; in tsc_refine_calibration_work()
1365 if (abs(tsc_khz - freq) > (tsc_khz >> 11)) { in tsc_refine_calibration_work()
1381 if (abs(tsc_khz - freq) > tsc_khz/100) in tsc_refine_calibration_work()
1385 pr_info("Refined TSC clocksource calibration: %lu.%03lu MHz\n", in tsc_refine_calibration_work()
1425 * the refined calibration and directly register it as a clocksource. in init_tsc_clocksource()
1462 /* We should not be here with non-native cpu calibration */ in determine_cpu_tsc_frequencies()
1468 * Trust non-zero tsc_khz as authoritative, in determine_cpu_tsc_frequencies()
1474 else if (abs(cpu_khz - tsc_khz) * 10 > tsc_khz) in determine_cpu_tsc_frequencies()
1515 /* Don't change UV TSC multi-chassis synchronization */ in tsc_early_init()
1573 * Check whether existing calibration data can be reused.
1583 * sockets then reuse CPU0 calibration. in calibrate_delay_is_known()
1591 * the calibration value of an already online CPU on that socket. in calibrate_delay_is_known()