Merge tag 'ext4_for_linus_stable' of git://git.kernel.org/pub/scm/linux/kernel/git...
[platform/kernel/linux-rpi.git] / kernel / time / vsyscall.c
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright 2019 ARM Ltd.
4  *
5  * Generic implementation of update_vsyscall and update_vsyscall_tz.
6  *
7  * Based on the x86 specific implementation.
8  */
9
10 #include <linux/hrtimer.h>
11 #include <linux/timekeeper_internal.h>
12 #include <vdso/datapage.h>
13 #include <vdso/helpers.h>
14 #include <vdso/vsyscall.h>
15
16 #include "timekeeping_internal.h"
17
18 static inline void update_vdso_data(struct vdso_data *vdata,
19                                     struct timekeeper *tk)
20 {
21         struct vdso_timestamp *vdso_ts;
22         u64 nsec, sec;
23
24         vdata[CS_HRES_COARSE].cycle_last        = tk->tkr_mono.cycle_last;
25         vdata[CS_HRES_COARSE].mask              = tk->tkr_mono.mask;
26         vdata[CS_HRES_COARSE].mult              = tk->tkr_mono.mult;
27         vdata[CS_HRES_COARSE].shift             = tk->tkr_mono.shift;
28         vdata[CS_RAW].cycle_last                = tk->tkr_raw.cycle_last;
29         vdata[CS_RAW].mask                      = tk->tkr_raw.mask;
30         vdata[CS_RAW].mult                      = tk->tkr_raw.mult;
31         vdata[CS_RAW].shift                     = tk->tkr_raw.shift;
32
33         /* CLOCK_MONOTONIC */
34         vdso_ts         = &vdata[CS_HRES_COARSE].basetime[CLOCK_MONOTONIC];
35         vdso_ts->sec    = tk->xtime_sec + tk->wall_to_monotonic.tv_sec;
36
37         nsec = tk->tkr_mono.xtime_nsec;
38         nsec += ((u64)tk->wall_to_monotonic.tv_nsec << tk->tkr_mono.shift);
39         while (nsec >= (((u64)NSEC_PER_SEC) << tk->tkr_mono.shift)) {
40                 nsec -= (((u64)NSEC_PER_SEC) << tk->tkr_mono.shift);
41                 vdso_ts->sec++;
42         }
43         vdso_ts->nsec   = nsec;
44
45         /* Copy MONOTONIC time for BOOTTIME */
46         sec     = vdso_ts->sec;
47         /* Add the boot offset */
48         sec     += tk->monotonic_to_boot.tv_sec;
49         nsec    += (u64)tk->monotonic_to_boot.tv_nsec << tk->tkr_mono.shift;
50
51         /* CLOCK_BOOTTIME */
52         vdso_ts         = &vdata[CS_HRES_COARSE].basetime[CLOCK_BOOTTIME];
53         vdso_ts->sec    = sec;
54
55         while (nsec >= (((u64)NSEC_PER_SEC) << tk->tkr_mono.shift)) {
56                 nsec -= (((u64)NSEC_PER_SEC) << tk->tkr_mono.shift);
57                 vdso_ts->sec++;
58         }
59         vdso_ts->nsec   = nsec;
60
61         /* CLOCK_MONOTONIC_RAW */
62         vdso_ts         = &vdata[CS_RAW].basetime[CLOCK_MONOTONIC_RAW];
63         vdso_ts->sec    = tk->raw_sec;
64         vdso_ts->nsec   = tk->tkr_raw.xtime_nsec;
65
66         /* CLOCK_TAI */
67         vdso_ts         = &vdata[CS_HRES_COARSE].basetime[CLOCK_TAI];
68         vdso_ts->sec    = tk->xtime_sec + (s64)tk->tai_offset;
69         vdso_ts->nsec   = tk->tkr_mono.xtime_nsec;
70 }
71
72 void update_vsyscall(struct timekeeper *tk)
73 {
74         struct vdso_data *vdata = __arch_get_k_vdso_data();
75         struct vdso_timestamp *vdso_ts;
76         s32 clock_mode;
77         u64 nsec;
78
79         /* copy vsyscall data */
80         vdso_write_begin(vdata);
81
82         clock_mode = tk->tkr_mono.clock->vdso_clock_mode;
83         vdata[CS_HRES_COARSE].clock_mode        = clock_mode;
84         vdata[CS_RAW].clock_mode                = clock_mode;
85
86         /* CLOCK_REALTIME also required for time() */
87         vdso_ts         = &vdata[CS_HRES_COARSE].basetime[CLOCK_REALTIME];
88         vdso_ts->sec    = tk->xtime_sec;
89         vdso_ts->nsec   = tk->tkr_mono.xtime_nsec;
90
91         /* CLOCK_REALTIME_COARSE */
92         vdso_ts         = &vdata[CS_HRES_COARSE].basetime[CLOCK_REALTIME_COARSE];
93         vdso_ts->sec    = tk->xtime_sec;
94         vdso_ts->nsec   = tk->tkr_mono.xtime_nsec >> tk->tkr_mono.shift;
95
96         /* CLOCK_MONOTONIC_COARSE */
97         vdso_ts         = &vdata[CS_HRES_COARSE].basetime[CLOCK_MONOTONIC_COARSE];
98         vdso_ts->sec    = tk->xtime_sec + tk->wall_to_monotonic.tv_sec;
99         nsec            = tk->tkr_mono.xtime_nsec >> tk->tkr_mono.shift;
100         nsec            = nsec + tk->wall_to_monotonic.tv_nsec;
101         vdso_ts->sec    += __iter_div_u64_rem(nsec, NSEC_PER_SEC, &vdso_ts->nsec);
102
103         /*
104          * Read without the seqlock held by clock_getres().
105          * Note: No need to have a second copy.
106          */
107         WRITE_ONCE(vdata[CS_HRES_COARSE].hrtimer_res, hrtimer_resolution);
108
109         /*
110          * If the current clocksource is not VDSO capable, then spare the
111          * update of the high resolution parts.
112          */
113         if (clock_mode != VDSO_CLOCKMODE_NONE)
114                 update_vdso_data(vdata, tk);
115
116         __arch_update_vsyscall(vdata, tk);
117
118         vdso_write_end(vdata);
119
120         __arch_sync_vdso_data(vdata);
121 }
122
123 void update_vsyscall_tz(void)
124 {
125         struct vdso_data *vdata = __arch_get_k_vdso_data();
126
127         vdata[CS_HRES_COARSE].tz_minuteswest = sys_tz.tz_minuteswest;
128         vdata[CS_HRES_COARSE].tz_dsttime = sys_tz.tz_dsttime;
129
130         __arch_sync_vdso_data(vdata);
131 }
132
133 /**
134  * vdso_update_begin - Start of a VDSO update section
135  *
136  * Allows architecture code to safely update the architecture specific VDSO
137  * data. Disables interrupts, acquires timekeeper lock to serialize against
138  * concurrent updates from timekeeping and invalidates the VDSO data
139  * sequence counter to prevent concurrent readers from accessing
140  * inconsistent data.
141  *
142  * Returns: Saved interrupt flags which need to be handed in to
143  * vdso_update_end().
144  */
145 unsigned long vdso_update_begin(void)
146 {
147         struct vdso_data *vdata = __arch_get_k_vdso_data();
148         unsigned long flags;
149
150         raw_spin_lock_irqsave(&timekeeper_lock, flags);
151         vdso_write_begin(vdata);
152         return flags;
153 }
154
155 /**
156  * vdso_update_end - End of a VDSO update section
157  * @flags:      Interrupt flags as returned from vdso_update_begin()
158  *
159  * Pairs with vdso_update_begin(). Marks vdso data consistent, invokes data
160  * synchronization if the architecture requires it, drops timekeeper lock
161  * and restores interrupt flags.
162  */
163 void vdso_update_end(unsigned long flags)
164 {
165         struct vdso_data *vdata = __arch_get_k_vdso_data();
166
167         vdso_write_end(vdata);
168         __arch_sync_vdso_data(vdata);
169         raw_spin_unlock_irqrestore(&timekeeper_lock, flags);
170 }