timekeeper_internal.h 5.1 KB
Newer Older
1
/* SPDX-License-Identifier: GPL-2.0 */
2 3 4 5 6 7 8 9 10 11 12 13
/*
 * You SHOULD NOT be including this unless you're vsyscall
 * handling code or timekeeping internal code!
 */

#ifndef _LINUX_TIMEKEEPER_INTERNAL_H
#define _LINUX_TIMEKEEPER_INTERNAL_H

#include <linux/clocksource.h>
#include <linux/jiffies.h>
#include <linux/time.h>

14 15 16 17 18
/**
 * struct tk_read_base - base structure for timekeeping readout
 * @clock:	Current clocksource used for timekeeping.
 * @mask:	Bitmask for two's complement subtraction of non 64bit clocks
 * @cycle_last: @clock cycle value at last update
19
 * @mult:	(NTP adjusted) multiplier for scaled math conversion
20 21
 * @shift:	Shift value for scaled math conversion
 * @xtime_nsec: Shifted (fractional) nano seconds offset for readout
22
 * @base:	ktime_t (nanoseconds) base time for readout
23
 * @base_real:	Nanoseconds base value for clock REALTIME readout
24
 *
25 26
 * This struct has size 56 byte on 64 bit. Together with a seqcount it
 * occupies a single 64byte cache line.
27
 *
28
 * The struct is separate from struct timekeeper as it is also used
29
 * for a fast NMI safe accessors.
30 31 32
 *
 * @base_real is for the fast NMI safe accessor to allow reading clock
 * realtime from any context.
33
 */
34
struct tk_read_base {
35
	struct clocksource	*clock;
36 37
	u64			mask;
	u64			cycle_last;
38 39
	u32			mult;
	u32			shift;
40
	u64			xtime_nsec;
41
	ktime_t			base;
42
	u64			base_real;
43
};
44

45 46
/**
 * struct timekeeper - Structure holding internal timekeeping values.
47
 * @tkr_mono:		The readout base structure for CLOCK_MONOTONIC
P
Peter Zijlstra 已提交
48
 * @tkr_raw:		The readout base structure for CLOCK_MONOTONIC_RAW
49
 * @xtime_sec:		Current CLOCK_REALTIME time in seconds
50
 * @ktime_sec:		Current CLOCK_MONOTONIC time in seconds
51 52 53 54 55
 * @wall_to_monotonic:	CLOCK_REALTIME to CLOCK_MONOTONIC offset
 * @offs_real:		Offset clock monotonic -> clock realtime
 * @offs_boot:		Offset clock monotonic -> clock boottime
 * @offs_tai:		Offset clock monotonic -> clock tai
 * @tai_offset:		The current UTC to TAI offset in seconds
56
 * @clock_was_set_seq:	The sequence number of clock was set events
57
 * @cs_was_changed_seq:	The sequence number of clocksource change events
58
 * @next_leap_ktime:	CLOCK_MONOTONIC time value of a pending leap-second
59
 * @raw_sec:		CLOCK_MONOTONIC_RAW  time in seconds
60 61 62 63 64
 * @cycle_interval:	Number of clock cycles in one NTP interval
 * @xtime_interval:	Number of clock shifted nano seconds in one NTP
 *			interval.
 * @xtime_remainder:	Shifted nano seconds left over when rounding
 *			@cycle_interval
65
 * @raw_interval:	Shifted raw nano seconds accumulated per NTP interval.
66 67 68 69
 * @ntp_error:		Difference between accumulated time and NTP time in ntp
 *			shifted nano seconds.
 * @ntp_error_shift:	Shift conversion between clock shifted nano seconds and
 *			ntp shifted nano seconds.
70 71 72
 * @last_warning:	Warning ratelimiter (DEBUG_TIMEKEEPING)
 * @underflow_seen:	Underflow warning flag (DEBUG_TIMEKEEPING)
 * @overflow_seen:	Overflow warning flag (DEBUG_TIMEKEEPING)
73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88
 *
 * Note: For timespec(64) based interfaces wall_to_monotonic is what
 * we need to add to xtime (or xtime corrected for sub jiffie times)
 * to get to monotonic time.  Monotonic is pegged at zero at system
 * boot time, so wall_to_monotonic will be negative, however, we will
 * ALWAYS keep the tv_nsec part positive so we can use the usual
 * normalization.
 *
 * wall_to_monotonic is moved after resume from suspend for the
 * monotonic time not to jump. We need to add total_sleep_time to
 * wall_to_monotonic to get the real boot based time offset.
 *
 * wall_to_monotonic is no longer the boot time, getboottime must be
 * used instead.
 */
struct timekeeper {
89
	struct tk_read_base	tkr_mono;
P
Peter Zijlstra 已提交
90
	struct tk_read_base	tkr_raw;
91
	u64			xtime_sec;
92
	unsigned long		ktime_sec;
93 94 95 96 97
	struct timespec64	wall_to_monotonic;
	ktime_t			offs_real;
	ktime_t			offs_boot;
	ktime_t			offs_tai;
	s32			tai_offset;
98
	unsigned int		clock_was_set_seq;
99
	u8			cs_was_changed_seq;
100
	ktime_t			next_leap_ktime;
101
	u64			raw_sec;
102

103
	/* The following members are for timekeeping internal use */
104
	u64			cycle_interval;
105 106
	u64			xtime_interval;
	s64			xtime_remainder;
107
	u64			raw_interval;
108 109 110 111 112 113 114 115 116
	/* The ntp_tick_length() value currently being used.
	 * This cached copy ensures we consistently apply the tick
	 * length for an entire tick, as ntp_tick_length may change
	 * mid-tick, and we don't want to apply that new value to
	 * the tick in progress.
	 */
	u64			ntp_tick;
	/* Difference between accumulated time and NTP time in ntp
	 * shifted nano seconds. */
117
	s64			ntp_error;
118
	u32			ntp_error_shift;
119
	u32			ntp_err_mult;
120 121
	/* Flag used to avoid updating NTP twice with same second */
	u32			skip_second_overflow;
122 123 124 125 126 127 128 129 130 131 132 133
#ifdef CONFIG_DEBUG_TIMEKEEPING
	long			last_warning;
	/*
	 * These simple flag variables are managed
	 * without locks, which is racy, but they are
	 * ok since we don't really care about being
	 * super precise about how many events were
	 * seen, just that a problem was observed.
	 */
	int			underflow_seen;
	int			overflow_seen;
#endif
134
};
135

136 137 138 139
#ifdef CONFIG_GENERIC_TIME_VSYSCALL

extern void update_vsyscall(struct timekeeper *tk);
extern void update_vsyscall_tz(void);
140 141

#else
142 143

static inline void update_vsyscall(struct timekeeper *tk)
144 145 146 147 148 149 150
{
}
static inline void update_vsyscall_tz(void)
{
}
#endif

151
#endif /* _LINUX_TIMEKEEPER_INTERNAL_H */