Branch data Line data Source code
1 : : // SPDX-License-Identifier: GPL-2.0
2 : : /*
3 : : * Copyright(C) 2005-2006, Thomas Gleixner <tglx@linutronix.de>
4 : : * Copyright(C) 2005-2007, Red Hat, Inc., Ingo Molnar
5 : : * Copyright(C) 2006-2007 Timesys Corp., Thomas Gleixner
6 : : *
7 : : * High-resolution kernel timers
8 : : *
9 : : * In contrast to the low-resolution timeout API, aka timer wheel,
10 : : * hrtimers provide finer resolution and accuracy depending on system
11 : : * configuration and capabilities.
12 : : *
13 : : * Started by: Thomas Gleixner and Ingo Molnar
14 : : *
15 : : * Credits:
16 : : * Based on the original timer wheel code
17 : : *
18 : : * Help, testing, suggestions, bugfixes, improvements were
19 : : * provided by:
20 : : *
21 : : * George Anzinger, Andrew Morton, Steven Rostedt, Roman Zippel
22 : : * et. al.
23 : : */
24 : :
25 : : #include <linux/cpu.h>
26 : : #include <linux/export.h>
27 : : #include <linux/percpu.h>
28 : : #include <linux/hrtimer.h>
29 : : #include <linux/notifier.h>
30 : : #include <linux/syscalls.h>
31 : : #include <linux/interrupt.h>
32 : : #include <linux/tick.h>
33 : : #include <linux/err.h>
34 : : #include <linux/debugobjects.h>
35 : : #include <linux/sched/signal.h>
36 : : #include <linux/sched/sysctl.h>
37 : : #include <linux/sched/rt.h>
38 : : #include <linux/sched/deadline.h>
39 : : #include <linux/sched/nohz.h>
40 : : #include <linux/sched/debug.h>
41 : : #include <linux/timer.h>
42 : : #include <linux/freezer.h>
43 : : #include <linux/compat.h>
44 : :
45 : : #include <linux/uaccess.h>
46 : :
47 : : #include <trace/events/timer.h>
48 : :
49 : : #include "tick-internal.h"
50 : :
51 : : /*
52 : : * Masks for selecting the soft and hard context timers from
53 : : * cpu_base->active
54 : : */
55 : : #define MASK_SHIFT (HRTIMER_BASE_MONOTONIC_SOFT)
56 : : #define HRTIMER_ACTIVE_HARD ((1U << MASK_SHIFT) - 1)
57 : : #define HRTIMER_ACTIVE_SOFT (HRTIMER_ACTIVE_HARD << MASK_SHIFT)
58 : : #define HRTIMER_ACTIVE_ALL (HRTIMER_ACTIVE_SOFT | HRTIMER_ACTIVE_HARD)
59 : :
60 : : /*
61 : : * The timer bases:
62 : : *
63 : : * There are more clockids than hrtimer bases. Thus, we index
64 : : * into the timer bases by the hrtimer_base_type enum. When trying
65 : : * to reach a base using a clockid, hrtimer_clockid_to_base()
66 : : * is used to convert from clockid to the proper hrtimer_base_type.
67 : : */
68 : : DEFINE_PER_CPU(struct hrtimer_cpu_base, hrtimer_bases) =
69 : : {
70 : : .lock = __RAW_SPIN_LOCK_UNLOCKED(hrtimer_bases.lock),
71 : : .clock_base =
72 : : {
73 : : {
74 : : .index = HRTIMER_BASE_MONOTONIC,
75 : : .clockid = CLOCK_MONOTONIC,
76 : : .get_time = &ktime_get,
77 : : },
78 : : {
79 : : .index = HRTIMER_BASE_REALTIME,
80 : : .clockid = CLOCK_REALTIME,
81 : : .get_time = &ktime_get_real,
82 : : },
83 : : {
84 : : .index = HRTIMER_BASE_BOOTTIME,
85 : : .clockid = CLOCK_BOOTTIME,
86 : : .get_time = &ktime_get_boottime,
87 : : },
88 : : {
89 : : .index = HRTIMER_BASE_TAI,
90 : : .clockid = CLOCK_TAI,
91 : : .get_time = &ktime_get_clocktai,
92 : : },
93 : : {
94 : : .index = HRTIMER_BASE_MONOTONIC_SOFT,
95 : : .clockid = CLOCK_MONOTONIC,
96 : : .get_time = &ktime_get,
97 : : },
98 : : {
99 : : .index = HRTIMER_BASE_REALTIME_SOFT,
100 : : .clockid = CLOCK_REALTIME,
101 : : .get_time = &ktime_get_real,
102 : : },
103 : : {
104 : : .index = HRTIMER_BASE_BOOTTIME_SOFT,
105 : : .clockid = CLOCK_BOOTTIME,
106 : : .get_time = &ktime_get_boottime,
107 : : },
108 : : {
109 : : .index = HRTIMER_BASE_TAI_SOFT,
110 : : .clockid = CLOCK_TAI,
111 : : .get_time = &ktime_get_clocktai,
112 : : },
113 : : }
114 : : };
115 : :
116 : : static const int hrtimer_clock_to_base_table[MAX_CLOCKS] = {
117 : : /* Make sure we catch unsupported clockids */
118 : : [0 ... MAX_CLOCKS - 1] = HRTIMER_MAX_CLOCK_BASES,
119 : :
120 : : [CLOCK_REALTIME] = HRTIMER_BASE_REALTIME,
121 : : [CLOCK_MONOTONIC] = HRTIMER_BASE_MONOTONIC,
122 : : [CLOCK_BOOTTIME] = HRTIMER_BASE_BOOTTIME,
123 : : [CLOCK_TAI] = HRTIMER_BASE_TAI,
124 : : };
125 : :
126 : : /*
127 : : * Functions and macros which are different for UP/SMP systems are kept in a
128 : : * single place
129 : : */
130 : : #ifdef CONFIG_SMP
131 : :
132 : : /*
133 : : * We require the migration_base for lock_hrtimer_base()/switch_hrtimer_base()
134 : : * such that hrtimer_callback_running() can unconditionally dereference
135 : : * timer->base->cpu_base
136 : : */
137 : : static struct hrtimer_cpu_base migration_cpu_base = {
138 : : .clock_base = { { .cpu_base = &migration_cpu_base, }, },
139 : : };
140 : :
141 : : #define migration_base migration_cpu_base.clock_base[0]
142 : :
143 : : static inline bool is_migration_base(struct hrtimer_clock_base *base)
144 : : {
145 : : return base == &migration_base;
146 : : }
147 : :
148 : : /*
149 : : * We are using hashed locking: holding per_cpu(hrtimer_bases)[n].lock
150 : : * means that all timers which are tied to this base via timer->base are
151 : : * locked, and the base itself is locked too.
152 : : *
153 : : * So __run_timers/migrate_timers can safely modify all timers which could
154 : : * be found on the lists/queues.
155 : : *
156 : : * When the timer's base is locked, and the timer removed from list, it is
157 : : * possible to set timer->base = &migration_base and drop the lock: the timer
158 : : * remains locked.
159 : : */
160 : : static
161 : 11431 : struct hrtimer_clock_base *lock_hrtimer_base(const struct hrtimer *timer,
162 : : unsigned long *flags)
163 : : {
164 : 11431 : struct hrtimer_clock_base *base;
165 : :
166 : 11431 : for (;;) {
167 [ + - ]: 11431 : base = READ_ONCE(timer->base);
168 [ + - ]: 11431 : if (likely(base != &migration_base)) {
169 : 11431 : raw_spin_lock_irqsave(&base->cpu_base->lock, *flags);
170 [ + - ]: 11431 : if (likely(base == timer->base))
171 : 11431 : return base;
172 : : /* The timer has migrated to another CPU: */
173 : 0 : raw_spin_unlock_irqrestore(&base->cpu_base->lock, *flags);
174 : : }
175 : 0 : cpu_relax();
176 : : }
177 : : }
178 : :
179 : : /*
180 : : * We do not migrate the timer when it is expiring before the next
181 : : * event on the target cpu. When high resolution is enabled, we cannot
182 : : * reprogram the target cpu hardware and we would cause it to fire
183 : : * late. To keep it simple, we handle the high resolution enabled and
184 : : * disabled case similar.
185 : : *
186 : : * Called with cpu_base->lock of target cpu held.
187 : : */
188 : : static int
189 : 0 : hrtimer_check_target(struct hrtimer *timer, struct hrtimer_clock_base *new_base)
190 : : {
191 : 0 : ktime_t expires;
192 : :
193 : 0 : expires = ktime_sub(hrtimer_get_expires(timer), new_base->offset);
194 : 0 : return expires < new_base->cpu_base->expires_next;
195 : : }
196 : :
197 : : static inline
198 : 6559 : struct hrtimer_cpu_base *get_target_base(struct hrtimer_cpu_base *base,
199 : : int pinned)
200 : : {
201 : : #if defined(CONFIG_SMP) && defined(CONFIG_NO_HZ_COMMON)
202 [ + + + + : 13088 : if (static_branch_likely(&timers_migration_enabled) && !pinned)
+ + ]
203 : 5105 : return &per_cpu(hrtimer_bases, get_nohz_timer_target());
204 : : #endif
205 : : return base;
206 : : }
207 : :
208 : : /*
209 : : * We switch the timer base to a power-optimized selected CPU target,
210 : : * if:
211 : : * - NO_HZ_COMMON is enabled
212 : : * - timer migration is enabled
213 : : * - the timer callback is not running
214 : : * - the timer is not the first expiring timer on the new target
215 : : *
216 : : * If one of the above requirements is not fulfilled we move the timer
217 : : * to the current CPU or leave it on the previously assigned CPU if
218 : : * the timer callback is currently running.
219 : : */
220 : : static inline struct hrtimer_clock_base *
221 : 6559 : switch_hrtimer_base(struct hrtimer *timer, struct hrtimer_clock_base *base,
222 : : int pinned)
223 : : {
224 : 6559 : struct hrtimer_cpu_base *new_cpu_base, *this_cpu_base;
225 : 6559 : struct hrtimer_clock_base *new_base;
226 : 6559 : int basenum = base->index;
227 : :
228 : 6559 : this_cpu_base = this_cpu_ptr(&hrtimer_bases);
229 : 6559 : new_cpu_base = get_target_base(this_cpu_base, pinned);
230 : : again:
231 : 6559 : new_base = &new_cpu_base->clock_base[basenum];
232 : :
233 [ - + ]: 6559 : if (base != new_base) {
234 : : /*
235 : : * We are trying to move timer to new_base.
236 : : * However we can't change timer's base while it is running,
237 : : * so we keep it on the same CPU. No hassle vs. reprogramming
238 : : * the event source in the high resolution case. The softirq
239 : : * code will take care of this when the timer function has
240 : : * completed. There is no conflict as we hold the lock until
241 : : * the timer is enqueued.
242 : : */
243 [ # # ]: 0 : if (unlikely(hrtimer_callback_running(timer)))
244 : : return base;
245 : :
246 : : /* See the comment in lock_hrtimer_base() */
247 : 0 : WRITE_ONCE(timer->base, &migration_base);
248 : 0 : raw_spin_unlock(&base->cpu_base->lock);
249 : 0 : raw_spin_lock(&new_base->cpu_base->lock);
250 : :
251 [ # # # # ]: 0 : if (new_cpu_base != this_cpu_base &&
252 [ # # ]: 0 : hrtimer_check_target(timer, new_base)) {
253 : 0 : raw_spin_unlock(&new_base->cpu_base->lock);
254 : 0 : raw_spin_lock(&base->cpu_base->lock);
255 : 0 : new_cpu_base = this_cpu_base;
256 : 0 : WRITE_ONCE(timer->base, base);
257 : 0 : goto again;
258 : : }
259 : 0 : WRITE_ONCE(timer->base, new_base);
260 : : } else {
261 [ - + - - ]: 6559 : if (new_cpu_base != this_cpu_base &&
262 [ # # ]: 0 : hrtimer_check_target(timer, new_base)) {
263 : 0 : new_cpu_base = this_cpu_base;
264 : 0 : goto again;
265 : : }
266 : : }
267 : : return new_base;
268 : : }
269 : :
270 : : #else /* CONFIG_SMP */
271 : :
272 : : static inline bool is_migration_base(struct hrtimer_clock_base *base)
273 : : {
274 : : return false;
275 : : }
276 : :
277 : : static inline struct hrtimer_clock_base *
278 : : lock_hrtimer_base(const struct hrtimer *timer, unsigned long *flags)
279 : : {
280 : : struct hrtimer_clock_base *base = timer->base;
281 : :
282 : : raw_spin_lock_irqsave(&base->cpu_base->lock, *flags);
283 : :
284 : : return base;
285 : : }
286 : :
287 : : # define switch_hrtimer_base(t, b, p) (b)
288 : :
289 : : #endif /* !CONFIG_SMP */
290 : :
291 : : /*
292 : : * Functions for the union type storage format of ktime_t which are
293 : : * too large for inlining:
294 : : */
295 : : #if BITS_PER_LONG < 64
296 : : /*
297 : : * Divide a ktime value by a nanosecond value
298 : : */
299 : : s64 __ktime_divns(const ktime_t kt, s64 div)
300 : : {
301 : : int sft = 0;
302 : : s64 dclc;
303 : : u64 tmp;
304 : :
305 : : dclc = ktime_to_ns(kt);
306 : : tmp = dclc < 0 ? -dclc : dclc;
307 : :
308 : : /* Make sure the divisor is less than 2^32: */
309 : : while (div >> 32) {
310 : : sft++;
311 : : div >>= 1;
312 : : }
313 : : tmp >>= sft;
314 : : do_div(tmp, (unsigned long) div);
315 : : return dclc < 0 ? -tmp : tmp;
316 : : }
317 : : EXPORT_SYMBOL_GPL(__ktime_divns);
318 : : #endif /* BITS_PER_LONG >= 64 */
319 : :
320 : : /*
321 : : * Add two ktime values and do a safety check for overflow:
322 : : */
323 : 189032 : ktime_t ktime_add_safe(const ktime_t lhs, const ktime_t rhs)
324 : : {
325 : 189032 : ktime_t res = ktime_add_unsafe(lhs, rhs);
326 : :
327 : : /*
328 : : * We use KTIME_SEC_MAX here, the maximum timeout which we can
329 : : * return to user space in a timespec:
330 : : */
331 [ + - - + : 98236 : if (res < 0 || res < lhs || res < rhs)
+ - - + +
- - + + -
- + + - -
+ + - - +
- - - - ]
332 : 0 : res = ktime_set(KTIME_SEC_MAX, 0);
333 : :
334 [ + - ]: 188911 : return res;
335 : : }
336 : :
337 : : EXPORT_SYMBOL_GPL(ktime_add_safe);
338 : :
339 : : #ifdef CONFIG_DEBUG_OBJECTS_TIMERS
340 : :
341 : : static struct debug_obj_descr hrtimer_debug_descr;
342 : :
343 : : static void *hrtimer_debug_hint(void *addr)
344 : : {
345 : : return ((struct hrtimer *) addr)->function;
346 : : }
347 : :
348 : : /*
349 : : * fixup_init is called when:
350 : : * - an active object is initialized
351 : : */
352 : : static bool hrtimer_fixup_init(void *addr, enum debug_obj_state state)
353 : : {
354 : : struct hrtimer *timer = addr;
355 : :
356 : : switch (state) {
357 : : case ODEBUG_STATE_ACTIVE:
358 : : hrtimer_cancel(timer);
359 : : debug_object_init(timer, &hrtimer_debug_descr);
360 : : return true;
361 : : default:
362 : : return false;
363 : : }
364 : : }
365 : :
366 : : /*
367 : : * fixup_activate is called when:
368 : : * - an active object is activated
369 : : * - an unknown non-static object is activated
370 : : */
371 : : static bool hrtimer_fixup_activate(void *addr, enum debug_obj_state state)
372 : : {
373 : : switch (state) {
374 : : case ODEBUG_STATE_ACTIVE:
375 : : WARN_ON(1);
376 : : /* fall through */
377 : : default:
378 : : return false;
379 : : }
380 : : }
381 : :
382 : : /*
383 : : * fixup_free is called when:
384 : : * - an active object is freed
385 : : */
386 : : static bool hrtimer_fixup_free(void *addr, enum debug_obj_state state)
387 : : {
388 : : struct hrtimer *timer = addr;
389 : :
390 : : switch (state) {
391 : : case ODEBUG_STATE_ACTIVE:
392 : : hrtimer_cancel(timer);
393 : : debug_object_free(timer, &hrtimer_debug_descr);
394 : : return true;
395 : : default:
396 : : return false;
397 : : }
398 : : }
399 : :
400 : : static struct debug_obj_descr hrtimer_debug_descr = {
401 : : .name = "hrtimer",
402 : : .debug_hint = hrtimer_debug_hint,
403 : : .fixup_init = hrtimer_fixup_init,
404 : : .fixup_activate = hrtimer_fixup_activate,
405 : : .fixup_free = hrtimer_fixup_free,
406 : : };
407 : :
408 : : static inline void debug_hrtimer_init(struct hrtimer *timer)
409 : : {
410 : : debug_object_init(timer, &hrtimer_debug_descr);
411 : : }
412 : :
413 : : static inline void debug_hrtimer_activate(struct hrtimer *timer,
414 : : enum hrtimer_mode mode)
415 : : {
416 : : debug_object_activate(timer, &hrtimer_debug_descr);
417 : : }
418 : :
419 : : static inline void debug_hrtimer_deactivate(struct hrtimer *timer)
420 : : {
421 : : debug_object_deactivate(timer, &hrtimer_debug_descr);
422 : : }
423 : :
424 : : static inline void debug_hrtimer_free(struct hrtimer *timer)
425 : : {
426 : : debug_object_free(timer, &hrtimer_debug_descr);
427 : : }
428 : :
429 : : static void __hrtimer_init(struct hrtimer *timer, clockid_t clock_id,
430 : : enum hrtimer_mode mode);
431 : :
432 : : void hrtimer_init_on_stack(struct hrtimer *timer, clockid_t clock_id,
433 : : enum hrtimer_mode mode)
434 : : {
435 : : debug_object_init_on_stack(timer, &hrtimer_debug_descr);
436 : : __hrtimer_init(timer, clock_id, mode);
437 : : }
438 : : EXPORT_SYMBOL_GPL(hrtimer_init_on_stack);
439 : :
440 : : static void __hrtimer_init_sleeper(struct hrtimer_sleeper *sl,
441 : : clockid_t clock_id, enum hrtimer_mode mode);
442 : :
443 : : void hrtimer_init_sleeper_on_stack(struct hrtimer_sleeper *sl,
444 : : clockid_t clock_id, enum hrtimer_mode mode)
445 : : {
446 : : debug_object_init_on_stack(&sl->timer, &hrtimer_debug_descr);
447 : : __hrtimer_init_sleeper(sl, clock_id, mode);
448 : : }
449 : : EXPORT_SYMBOL_GPL(hrtimer_init_sleeper_on_stack);
450 : :
451 : : void destroy_hrtimer_on_stack(struct hrtimer *timer)
452 : : {
453 : : debug_object_free(timer, &hrtimer_debug_descr);
454 : : }
455 : : EXPORT_SYMBOL_GPL(destroy_hrtimer_on_stack);
456 : :
457 : : #else
458 : :
459 : 79142 : static inline void debug_hrtimer_init(struct hrtimer *timer) { }
460 : 96698 : static inline void debug_hrtimer_activate(struct hrtimer *timer,
461 : 96698 : enum hrtimer_mode mode) { }
462 : 96362 : static inline void debug_hrtimer_deactivate(struct hrtimer *timer) { }
463 : : #endif
464 : :
465 : : static inline void
466 : 79142 : debug_init(struct hrtimer *timer, clockid_t clockid,
467 : : enum hrtimer_mode mode)
468 : : {
469 : 79142 : debug_hrtimer_init(timer);
470 : 79142 : trace_hrtimer_init(timer, clockid, mode);
471 : : }
472 : :
473 : 96698 : static inline void debug_activate(struct hrtimer *timer,
474 : : enum hrtimer_mode mode)
475 : : {
476 : 96698 : debug_hrtimer_activate(timer, mode);
477 : 96698 : trace_hrtimer_start(timer, mode);
478 : : }
479 : :
480 : 96362 : static inline void debug_deactivate(struct hrtimer *timer)
481 : : {
482 : 96362 : debug_hrtimer_deactivate(timer);
483 : 96362 : trace_hrtimer_cancel(timer);
484 : : }
485 : :
486 : : static struct hrtimer_clock_base *
487 : 695907 : __next_base(struct hrtimer_cpu_base *cpu_base, unsigned int *active)
488 : : {
489 : 695907 : unsigned int idx;
490 : :
491 : 695907 : if (!*active)
492 : : return NULL;
493 : :
494 [ + - + - ]: 414283 : idx = __ffs(*active);
495 : 414283 : *active &= ~(1U << idx);
496 : :
497 : 414283 : return &cpu_base->clock_base[idx];
498 : : }
499 : :
500 : : #define for_each_active_base(base, cpu_base, active) \
501 : : while ((base = __next_base((cpu_base), &(active))))
502 : :
503 : 183718 : static ktime_t __hrtimer_next_event_base(struct hrtimer_cpu_base *cpu_base,
504 : : const struct hrtimer *exclude,
505 : : unsigned int active,
506 : : ktime_t expires_next)
507 : : {
508 : 183718 : struct hrtimer_clock_base *base;
509 : 183718 : ktime_t expires;
510 : :
511 [ + + + - ]: 391573 : for_each_active_base(base, cpu_base, active) {
512 : 207855 : struct timerqueue_node *next;
513 : 207855 : struct hrtimer *timer;
514 : :
515 [ - + ]: 207855 : next = timerqueue_getnext(&base->active);
516 : 207855 : timer = container_of(next, struct hrtimer, node);
517 [ - + ]: 207855 : if (timer == exclude) {
518 : : /* Get to the next timer in the queue. */
519 : 0 : next = timerqueue_iterate_next(next);
520 [ # # ]: 0 : if (!next)
521 : 0 : continue;
522 : :
523 : : timer = container_of(next, struct hrtimer, node);
524 : : }
525 [ + + ]: 207855 : expires = ktime_sub(hrtimer_get_expires(timer), base->offset);
526 [ + + ]: 207855 : if (expires < expires_next) {
527 : 91345 : expires_next = expires;
528 : :
529 : : /* Skip cpu_base update if a timer is being excluded. */
530 [ - + ]: 91345 : if (exclude)
531 : 0 : continue;
532 : :
533 [ - + ]: 91345 : if (timer->is_soft)
534 : 0 : cpu_base->softirq_next_timer = timer;
535 : : else
536 : 91345 : cpu_base->next_timer = timer;
537 : : }
538 : : }
539 : : /*
540 : : * clock_was_set() might have changed base->offset of any of
541 : : * the clock bases so the result might be negative. Fix it up
542 : : * to prevent a false positive in clockevents_program_event().
543 : : */
544 : 183718 : if (expires_next < 0)
545 : : expires_next = 0;
546 : 183718 : return expires_next;
547 : : }
548 : :
549 : : /*
550 : : * Recomputes cpu_base::*next_timer and returns the earliest expires_next but
551 : : * does not set cpu_base::*expires_next, that is done by hrtimer_reprogram.
552 : : *
553 : : * When a softirq is pending, we can ignore the HRTIMER_ACTIVE_SOFT bases,
554 : : * those timers will get run whenever the softirq gets handled, at the end of
555 : : * hrtimer_run_softirq(), hrtimer_update_softirq_timer() will re-add these bases.
556 : : *
557 : : * Therefore softirq values are those from the HRTIMER_ACTIVE_SOFT clock bases.
558 : : * The !softirq values are the minima across HRTIMER_ACTIVE_ALL, unless an actual
559 : : * softirq is pending, in which case they're the minima of HRTIMER_ACTIVE_HARD.
560 : : *
561 : : * @active_mask must be one of:
562 : : * - HRTIMER_ACTIVE_ALL,
563 : : * - HRTIMER_ACTIVE_SOFT, or
564 : : * - HRTIMER_ACTIVE_HARD.
565 : : */
566 : : static ktime_t
567 : 91859 : __hrtimer_get_next_event(struct hrtimer_cpu_base *cpu_base, unsigned int active_mask)
568 : : {
569 : 91859 : unsigned int active;
570 : 91859 : struct hrtimer *next_timer = NULL;
571 : 91859 : ktime_t expires_next = KTIME_MAX;
572 : :
573 [ + - + - ]: 91859 : if (!cpu_base->softirq_activated && (active_mask & HRTIMER_ACTIVE_SOFT)) {
574 : 91859 : active = cpu_base->active_bases & HRTIMER_ACTIVE_SOFT;
575 : 91859 : cpu_base->softirq_next_timer = NULL;
576 : 91859 : expires_next = __hrtimer_next_event_base(cpu_base, NULL,
577 : : active, KTIME_MAX);
578 : :
579 : 91859 : next_timer = cpu_base->softirq_next_timer;
580 : : }
581 : :
582 [ + - ]: 91859 : if (active_mask & HRTIMER_ACTIVE_HARD) {
583 : 91859 : active = cpu_base->active_bases & HRTIMER_ACTIVE_HARD;
584 : 91859 : cpu_base->next_timer = next_timer;
585 : 91859 : expires_next = __hrtimer_next_event_base(cpu_base, NULL, active,
586 : : expires_next);
587 : : }
588 : :
589 : 91859 : return expires_next;
590 : : }
591 : :
592 : 97936 : static inline ktime_t hrtimer_update_base(struct hrtimer_cpu_base *base)
593 : : {
594 : 97936 : ktime_t *offs_real = &base->clock_base[HRTIMER_BASE_REALTIME].offset;
595 : 97936 : ktime_t *offs_boot = &base->clock_base[HRTIMER_BASE_BOOTTIME].offset;
596 : 97936 : ktime_t *offs_tai = &base->clock_base[HRTIMER_BASE_TAI].offset;
597 : :
598 : 97936 : ktime_t now = ktime_get_update_offsets_now(&base->clock_was_set_seq,
599 : : offs_real, offs_boot, offs_tai);
600 : :
601 : 97936 : base->clock_base[HRTIMER_BASE_REALTIME_SOFT].offset = *offs_real;
602 : 97936 : base->clock_base[HRTIMER_BASE_BOOTTIME_SOFT].offset = *offs_boot;
603 : 97936 : base->clock_base[HRTIMER_BASE_TAI_SOFT].offset = *offs_tai;
604 : :
605 : 90960 : return now;
606 : : }
607 : :
608 : : /*
609 : : * Is the high resolution mode active ?
610 : : */
611 : 100951 : static inline int __hrtimer_hres_active(struct hrtimer_cpu_base *cpu_base)
612 : : {
613 : 100951 : return IS_ENABLED(CONFIG_HIGH_RES_TIMERS) ?
614 : 100951 : cpu_base->hres_active : 0;
615 : : }
616 : :
617 : 0 : static inline int hrtimer_hres_active(void)
618 : : {
619 : 0 : return __hrtimer_hres_active(this_cpu_ptr(&hrtimer_bases));
620 : : }
621 : :
622 : : /*
623 : : * Reprogram the event source with checking both queues for the
624 : : * next event
625 : : * Called with interrupts disabled and base->lock held
626 : : */
627 : : static void
628 : 868 : hrtimer_force_reprogram(struct hrtimer_cpu_base *cpu_base, int skip_equal)
629 : : {
630 : 868 : ktime_t expires_next;
631 : :
632 : : /*
633 : : * Find the current next expiration time.
634 : : */
635 : 868 : expires_next = __hrtimer_get_next_event(cpu_base, HRTIMER_ACTIVE_ALL);
636 : :
637 [ + + - + ]: 868 : if (cpu_base->next_timer && cpu_base->next_timer->is_soft) {
638 : : /*
639 : : * When the softirq is activated, hrtimer has to be
640 : : * programmed with the first hard hrtimer because soft
641 : : * timer interrupt could occur too late.
642 : : */
643 [ # # ]: 0 : if (cpu_base->softirq_activated)
644 : 0 : expires_next = __hrtimer_get_next_event(cpu_base,
645 : : HRTIMER_ACTIVE_HARD);
646 : : else
647 : 0 : cpu_base->softirq_expires_next = expires_next;
648 : : }
649 : :
650 [ + + + - ]: 868 : if (skip_equal && expires_next == cpu_base->expires_next)
651 : : return;
652 : :
653 : 868 : cpu_base->expires_next = expires_next;
654 : :
655 : : /*
656 : : * If hres is not active, hardware does not have to be
657 : : * reprogrammed yet.
658 : : *
659 : : * If a hang was detected in the last timer interrupt then we
660 : : * leave the hang delay active in the hardware. We want the
661 : : * system to make progress. That also prevents the following
662 : : * scenario:
663 : : * T1 expires 50ms from now
664 : : * T2 expires 5s from now
665 : : *
666 : : * T1 is removed, so this code is called and would reprogram
667 : : * the hardware to 5s from now. Any hrtimer_start after that
668 : : * will not reprogram the hardware due to hang_detected being
669 : : * set. So we'd effectivly block all timers until the T2 event
670 : : * fires.
671 : : */
672 [ + - + - ]: 868 : if (!__hrtimer_hres_active(cpu_base) || cpu_base->hang_detected)
673 : : return;
674 : :
675 : 868 : tick_program_event(cpu_base->expires_next, 1);
676 : : }
677 : :
678 : : /* High resolution timer related functions */
679 : : #ifdef CONFIG_HIGH_RES_TIMERS
680 : :
681 : : /*
682 : : * High resolution timer enabled ?
683 : : */
684 : : static bool hrtimer_hres_enabled __read_mostly = true;
685 : : unsigned int hrtimer_resolution __read_mostly = LOW_RES_NSEC;
686 : : EXPORT_SYMBOL_GPL(hrtimer_resolution);
687 : :
688 : : /*
689 : : * Enable / Disable high resolution mode
690 : : */
691 : 0 : static int __init setup_hrtimer_hres(char *str)
692 : : {
693 : 0 : return (kstrtobool(str, &hrtimer_hres_enabled) == 0);
694 : : }
695 : :
696 : : __setup("highres=", setup_hrtimer_hres);
697 : :
698 : : /*
699 : : * hrtimer_high_res_enabled - query, if the highres mode is enabled
700 : : */
701 : 6945 : static inline int hrtimer_is_hres_enabled(void)
702 : : {
703 : 6945 : return hrtimer_hres_enabled;
704 : : }
705 : :
706 : : /*
707 : : * Retrigger next event is called after clock was set
708 : : *
709 : : * Called with interrupts disabled via on_each_cpu()
710 : : */
711 : 30 : static void retrigger_next_event(void *arg)
712 : : {
713 : 30 : struct hrtimer_cpu_base *base = this_cpu_ptr(&hrtimer_bases);
714 : :
715 [ + - ]: 30 : if (!__hrtimer_hres_active(base))
716 : : return;
717 : :
718 : 30 : raw_spin_lock(&base->lock);
719 : 30 : hrtimer_update_base(base);
720 : 30 : hrtimer_force_reprogram(base, 0);
721 : 30 : raw_spin_unlock(&base->lock);
722 : : }
723 : :
724 : : /*
725 : : * Switch to high resolution mode
726 : : */
727 : 30 : static void hrtimer_switch_to_hres(void)
728 : : {
729 : 30 : struct hrtimer_cpu_base *base = this_cpu_ptr(&hrtimer_bases);
730 : :
731 [ - + ]: 30 : if (tick_init_highres()) {
732 : 0 : pr_warn("Could not switch to high resolution mode on CPU %u\n",
733 : : base->cpu);
734 : 0 : return;
735 : : }
736 : 30 : base->hres_active = 1;
737 : 30 : hrtimer_resolution = HIGH_RES_NSEC;
738 : :
739 : 30 : tick_setup_sched_timer();
740 : : /* "Retrigger" the interrupt to get things going */
741 : 30 : retrigger_next_event(NULL);
742 : : }
743 : :
744 : 0 : static void clock_was_set_work(struct work_struct *work)
745 : : {
746 : 0 : clock_was_set();
747 : 0 : }
748 : :
749 : : static DECLARE_WORK(hrtimer_work, clock_was_set_work);
750 : :
751 : : /*
752 : : * Called from timekeeping and resume code to reprogram the hrtimer
753 : : * interrupt device on all cpus.
754 : : */
755 : 0 : void clock_was_set_delayed(void)
756 : : {
757 : 0 : schedule_work(&hrtimer_work);
758 : 0 : }
759 : :
760 : : #else
761 : :
762 : : static inline int hrtimer_is_hres_enabled(void) { return 0; }
763 : : static inline void hrtimer_switch_to_hres(void) { }
764 : : static inline void retrigger_next_event(void *arg) { }
765 : :
766 : : #endif /* CONFIG_HIGH_RES_TIMERS */
767 : :
768 : : /*
769 : : * When a timer is enqueued and expires earlier than the already enqueued
770 : : * timers, we have to check, whether it expires earlier than the timer for
771 : : * which the clock event device was armed.
772 : : *
773 : : * Called with interrupts disabled and base->cpu_base.lock held
774 : : */
775 : 1653 : static void hrtimer_reprogram(struct hrtimer *timer, bool reprogram)
776 : : {
777 : 1653 : struct hrtimer_cpu_base *cpu_base = this_cpu_ptr(&hrtimer_bases);
778 : 1653 : struct hrtimer_clock_base *base = timer->base;
779 [ - + ]: 1653 : ktime_t expires = ktime_sub(hrtimer_get_expires(timer), base->offset);
780 : :
781 [ - + ]: 1653 : WARN_ON_ONCE(hrtimer_get_expires_tv64(timer) < 0);
782 : :
783 : : /*
784 : : * CLOCK_REALTIME timer might be requested with an absolute
785 : : * expiry time which is less than base->offset. Set it to 0.
786 : : */
787 : 1653 : if (expires < 0)
788 : : expires = 0;
789 : :
790 [ - + ]: 1653 : if (timer->is_soft) {
791 : : /*
792 : : * soft hrtimer could be started on a remote CPU. In this
793 : : * case softirq_expires_next needs to be updated on the
794 : : * remote CPU. The soft hrtimer will not expire before the
795 : : * first hard hrtimer on the remote CPU -
796 : : * hrtimer_check_target() prevents this case.
797 : : */
798 : 0 : struct hrtimer_cpu_base *timer_cpu_base = base->cpu_base;
799 : :
800 [ # # ]: 0 : if (timer_cpu_base->softirq_activated)
801 : : return;
802 : :
803 [ # # ]: 0 : if (!ktime_before(expires, timer_cpu_base->softirq_expires_next))
804 : : return;
805 : :
806 : 0 : timer_cpu_base->softirq_next_timer = timer;
807 : 0 : timer_cpu_base->softirq_expires_next = expires;
808 : :
809 [ # # # # ]: 0 : if (!ktime_before(expires, timer_cpu_base->expires_next) ||
810 : : !reprogram)
811 : : return;
812 : : }
813 : :
814 : : /*
815 : : * If the timer is not on the current cpu, we cannot reprogram
816 : : * the other cpus clock event device.
817 : : */
818 [ + - ]: 1653 : if (base->cpu_base != cpu_base)
819 : : return;
820 : :
821 : : /*
822 : : * If the hrtimer interrupt is running, then it will
823 : : * reevaluate the clock bases and reprogram the clock event
824 : : * device. The callbacks are always executed in hard interrupt
825 : : * context so we don't need an extra check for a running
826 : : * callback.
827 : : */
828 [ + - ]: 1653 : if (cpu_base->in_hrtirq)
829 : : return;
830 : :
831 [ + + ]: 1653 : if (expires >= cpu_base->expires_next)
832 : : return;
833 : :
834 : : /* Update the pointer to the next expiring timer */
835 : 1563 : cpu_base->next_timer = timer;
836 : 1563 : cpu_base->expires_next = expires;
837 : :
838 : : /*
839 : : * If hres is not active, hardware does not have to be
840 : : * programmed yet.
841 : : *
842 : : * If a hang was detected in the last timer interrupt then we
843 : : * do not schedule a timer which is earlier than the expiry
844 : : * which we enforced in the hang detection. We want the system
845 : : * to make progress.
846 : : */
847 [ + - + - ]: 1563 : if (!__hrtimer_hres_active(cpu_base) || cpu_base->hang_detected)
848 : : return;
849 : :
850 : : /*
851 : : * Program the timer hardware. We enforce the expiry for
852 : : * events which are already in the past.
853 : : */
854 : 1563 : tick_program_event(expires, 1);
855 : : }
856 : :
857 : : /*
858 : : * Clock realtime was set
859 : : *
860 : : * Change the offset of the realtime clock vs. the monotonic
861 : : * clock.
862 : : *
863 : : * We might have to reprogram the high resolution timer interrupt. On
864 : : * SMP we call the architecture specific code to retrigger _all_ high
865 : : * resolution timer interrupts. On UP we just disable interrupts and
866 : : * call the high resolution interrupt code.
867 : : */
868 : 0 : void clock_was_set(void)
869 : : {
870 : : #ifdef CONFIG_HIGH_RES_TIMERS
871 : : /* Retrigger the CPU local events everywhere */
872 : 0 : on_each_cpu(retrigger_next_event, NULL, 1);
873 : : #endif
874 : 0 : timerfd_clock_was_set();
875 : 0 : }
876 : :
877 : : /*
878 : : * During resume we might have to reprogram the high resolution timer
879 : : * interrupt on all online CPUs. However, all other CPUs will be
880 : : * stopped with IRQs interrupts disabled so the clock_was_set() call
881 : : * must be deferred.
882 : : */
883 : 0 : void hrtimers_resume(void)
884 : : {
885 : 0 : lockdep_assert_irqs_disabled();
886 : : /* Retrigger on the local CPU */
887 : 0 : retrigger_next_event(NULL);
888 : : /* And schedule a retrigger for all others */
889 : 0 : clock_was_set_delayed();
890 : 0 : }
891 : :
892 : : /*
893 : : * Counterpart to lock_hrtimer_base above:
894 : : */
895 : : static inline
896 : 11431 : void unlock_hrtimer_base(const struct hrtimer *timer, unsigned long *flags)
897 : : {
898 : 18110 : raw_spin_unlock_irqrestore(&timer->base->cpu_base->lock, *flags);
899 : : }
900 : :
901 : : /**
902 : : * hrtimer_forward - forward the timer expiry
903 : : * @timer: hrtimer to forward
904 : : * @now: forward past this time
905 : : * @interval: the interval to forward
906 : : *
907 : : * Forward the timer expiry so it will expire in the future.
908 : : * Returns the number of overruns.
909 : : *
910 : : * Can be safely called from the callback function of @timer. If
911 : : * called from other contexts @timer must neither be enqueued nor
912 : : * running the callback and the caller needs to take care of
913 : : * serialization.
914 : : *
915 : : * Note: This only updates the timer expiry value and does not requeue
916 : : * the timer.
917 : : */
918 : 90771 : u64 hrtimer_forward(struct hrtimer *timer, ktime_t now, ktime_t interval)
919 : : {
920 : 90771 : u64 orun = 1;
921 : 90771 : ktime_t delta;
922 : :
923 [ + + ]: 90771 : delta = ktime_sub(now, hrtimer_get_expires(timer));
924 : :
925 [ + + ]: 90771 : if (delta < 0)
926 : : return 0;
927 : :
928 [ - + + - ]: 90675 : if (WARN_ON(timer->state & HRTIMER_STATE_ENQUEUED))
929 : : return 0;
930 : :
931 : 90675 : if (interval < hrtimer_resolution)
932 : : interval = hrtimer_resolution;
933 : :
934 [ + + ]: 90675 : if (unlikely(delta >= interval)) {
935 [ - + ]: 2553 : s64 incr = ktime_to_ns(interval);
936 : :
937 [ - + ]: 2553 : orun = ktime_divns(delta, incr);
938 [ + - ]: 2553 : hrtimer_add_expires_ns(timer, incr * orun);
939 [ + - ]: 2553 : if (hrtimer_get_expires_tv64(timer) > now)
940 : : return orun;
941 : : /*
942 : : * This (and the ktime_add() below) is the
943 : : * correction for exact:
944 : : */
945 : 2553 : orun++;
946 : : }
947 [ + - ]: 90675 : hrtimer_add_expires(timer, interval);
948 : :
949 : 90675 : return orun;
950 : : }
951 : : EXPORT_SYMBOL_GPL(hrtimer_forward);
952 : :
953 : : /*
954 : : * enqueue_hrtimer - internal function to (re)start a timer
955 : : *
956 : : * The timer is inserted in expiry order. Insertion into the
957 : : * red black tree is O(log(n)). Must hold the base lock.
958 : : *
959 : : * Returns 1 when the new timer is the leftmost timer in the tree.
960 : : */
961 : 96698 : static int enqueue_hrtimer(struct hrtimer *timer,
962 : : struct hrtimer_clock_base *base,
963 : : enum hrtimer_mode mode)
964 : : {
965 : 96698 : debug_activate(timer, mode);
966 : :
967 : 96698 : base->cpu_base->active_bases |= 1 << base->index;
968 : :
969 : : /* Pairs with the lockless read in hrtimer_is_queued() */
970 : 96698 : WRITE_ONCE(timer->state, HRTIMER_STATE_ENQUEUED);
971 : :
972 : 96698 : return timerqueue_add(&base->active, &timer->node);
973 : : }
974 : :
975 : : /*
976 : : * __remove_hrtimer - internal function to remove a timer
977 : : *
978 : : * Caller must hold the base lock.
979 : : *
980 : : * High resolution timer mode reprograms the clock event device when the
981 : : * timer is the one which expires next. The caller can disable this by setting
982 : : * reprogram to zero. This is useful, when the context does a reprogramming
983 : : * anyway (e.g. timer interrupt)
984 : : */
985 : 96362 : static void __remove_hrtimer(struct hrtimer *timer,
986 : : struct hrtimer_clock_base *base,
987 : : u8 newstate, int reprogram)
988 : : {
989 : 96362 : struct hrtimer_cpu_base *cpu_base = base->cpu_base;
990 : 96362 : u8 state = timer->state;
991 : :
992 : : /* Pairs with the lockless read in hrtimer_is_queued() */
993 [ + - ]: 96362 : WRITE_ONCE(timer->state, newstate);
994 [ + - ]: 96362 : if (!(state & HRTIMER_STATE_ENQUEUED))
995 : : return;
996 : :
997 [ + + ]: 96362 : if (!timerqueue_del(&base->active, &timer->node))
998 : 26703 : cpu_base->active_bases &= ~(1 << base->index);
999 : :
1000 : : /*
1001 : : * Note: If reprogram is false we do not update
1002 : : * cpu_base->next_timer. This happens when we remove the first
1003 : : * timer on a remote cpu. No harm as we never dereference
1004 : : * cpu_base->next_timer. So the worst thing what can happen is
1005 : : * an superflous call to hrtimer_force_reprogram() on the
1006 : : * remote cpu later on if the same timer gets enqueued again.
1007 : : */
1008 [ + + + + ]: 96362 : if (reprogram && timer == cpu_base->next_timer)
1009 : 838 : hrtimer_force_reprogram(cpu_base, 1);
1010 : : }
1011 : :
1012 : : /*
1013 : : * remove hrtimer, called with base lock held
1014 : : */
1015 : : static inline int
1016 : 11311 : remove_hrtimer(struct hrtimer *timer, struct hrtimer_clock_base *base, bool restart)
1017 : : {
1018 : 11311 : u8 state = timer->state;
1019 : :
1020 [ + + ]: 11311 : if (state & HRTIMER_STATE_ENQUEUED) {
1021 : 5354 : int reprogram;
1022 : :
1023 : : /*
1024 : : * Remove the timer and force reprogramming when high
1025 : : * resolution mode is active and the timer is on the current
1026 : : * CPU. If we remove a timer on another CPU, reprogramming is
1027 : : * skipped. The interrupt event on this CPU is fired and
1028 : : * reprogramming happens in the interrupt handler. This is a
1029 : : * rare case and less expensive than a smp call.
1030 : : */
1031 : 5354 : debug_deactivate(timer);
1032 : 5354 : reprogram = base->cpu_base == this_cpu_ptr(&hrtimer_bases);
1033 : :
1034 [ + + ]: 5354 : if (!restart)
1035 : 4752 : state = HRTIMER_STATE_INACTIVE;
1036 : :
1037 : 5354 : __remove_hrtimer(timer, base, state, reprogram);
1038 : 5354 : return 1;
1039 : : }
1040 : : return 0;
1041 : : }
1042 : :
1043 : 6559 : static inline ktime_t hrtimer_update_lowres(struct hrtimer *timer, ktime_t tim,
1044 : : const enum hrtimer_mode mode)
1045 : : {
1046 : : #ifdef CONFIG_TIME_LOW_RES
1047 : : /*
1048 : : * CONFIG_TIME_LOW_RES indicates that the system has no way to return
1049 : : * granular time values. For relative timers we add hrtimer_resolution
1050 : : * (i.e. one jiffie) to prevent short timeouts.
1051 : : */
1052 : : timer->is_rel = mode & HRTIMER_MODE_REL;
1053 : : if (timer->is_rel)
1054 : : tim = ktime_add_safe(tim, hrtimer_resolution);
1055 : : #endif
1056 : 6559 : return tim;
1057 : : }
1058 : :
1059 : : static void
1060 : 0 : hrtimer_update_softirq_timer(struct hrtimer_cpu_base *cpu_base, bool reprogram)
1061 : : {
1062 : 0 : ktime_t expires;
1063 : :
1064 : : /*
1065 : : * Find the next SOFT expiration.
1066 : : */
1067 : 0 : expires = __hrtimer_get_next_event(cpu_base, HRTIMER_ACTIVE_SOFT);
1068 : :
1069 : : /*
1070 : : * reprogramming needs to be triggered, even if the next soft
1071 : : * hrtimer expires at the same time than the next hard
1072 : : * hrtimer. cpu_base->softirq_expires_next needs to be updated!
1073 : : */
1074 [ # # ]: 0 : if (expires == KTIME_MAX)
1075 : : return;
1076 : :
1077 : : /*
1078 : : * cpu_base->*next_timer is recomputed by __hrtimer_get_next_event()
1079 : : * cpu_base->*expires_next is only set by hrtimer_reprogram()
1080 : : */
1081 : 0 : hrtimer_reprogram(cpu_base->softirq_next_timer, reprogram);
1082 : : }
1083 : :
1084 : 6559 : static int __hrtimer_start_range_ns(struct hrtimer *timer, ktime_t tim,
1085 : : u64 delta_ns, const enum hrtimer_mode mode,
1086 : : struct hrtimer_clock_base *base)
1087 : : {
1088 : 6559 : struct hrtimer_clock_base *new_base;
1089 : :
1090 : : /* Remove an active timer from the queue: */
1091 : 6559 : remove_hrtimer(timer, base, true);
1092 : :
1093 [ + + ]: 6559 : if (mode & HRTIMER_MODE_REL)
1094 : 121 : tim = ktime_add_safe(tim, base->get_time());
1095 : :
1096 : 6559 : tim = hrtimer_update_lowres(timer, tim, mode);
1097 : :
1098 [ + - ]: 6559 : hrtimer_set_expires_range_ns(timer, tim, delta_ns);
1099 : :
1100 : : /* Switch the timer base, if necessary: */
1101 : 6559 : new_base = switch_hrtimer_base(timer, base, mode & HRTIMER_MODE_PINNED);
1102 : :
1103 : 6559 : return enqueue_hrtimer(timer, new_base, mode);
1104 : : }
1105 : :
1106 : : /**
1107 : : * hrtimer_start_range_ns - (re)start an hrtimer
1108 : : * @timer: the timer to be added
1109 : : * @tim: expiry time
1110 : : * @delta_ns: "slack" range for the timer
1111 : : * @mode: timer mode: absolute (HRTIMER_MODE_ABS) or
1112 : : * relative (HRTIMER_MODE_REL), and pinned (HRTIMER_MODE_PINNED);
1113 : : * softirq based mode is considered for debug purpose only!
1114 : : */
1115 : 6559 : void hrtimer_start_range_ns(struct hrtimer *timer, ktime_t tim,
1116 : : u64 delta_ns, const enum hrtimer_mode mode)
1117 : : {
1118 : 6559 : struct hrtimer_clock_base *base;
1119 : 6559 : unsigned long flags;
1120 : :
1121 : : /*
1122 : : * Check whether the HRTIMER_MODE_SOFT bit and hrtimer.is_soft
1123 : : * match on CONFIG_PREEMPT_RT = n. With PREEMPT_RT check the hard
1124 : : * expiry mode because unmarked timers are moved to softirq expiry.
1125 : : */
1126 : 6559 : if (!IS_ENABLED(CONFIG_PREEMPT_RT))
1127 [ - + ]: 6559 : WARN_ON_ONCE(!(mode & HRTIMER_MODE_SOFT) ^ !timer->is_soft);
1128 : : else
1129 : : WARN_ON_ONCE(!(mode & HRTIMER_MODE_HARD) ^ !timer->is_hard);
1130 : :
1131 : 6559 : base = lock_hrtimer_base(timer, &flags);
1132 : :
1133 [ + + ]: 6559 : if (__hrtimer_start_range_ns(timer, tim, delta_ns, mode, base))
1134 : 1653 : hrtimer_reprogram(timer, true);
1135 : :
1136 : 6559 : unlock_hrtimer_base(timer, &flags);
1137 : 6559 : }
1138 : : EXPORT_SYMBOL_GPL(hrtimer_start_range_ns);
1139 : :
1140 : : /**
1141 : : * hrtimer_try_to_cancel - try to deactivate a timer
1142 : : * @timer: hrtimer to stop
1143 : : *
1144 : : * Returns:
1145 : : *
1146 : : * * 0 when the timer was not active
1147 : : * * 1 when the timer was active
1148 : : * * -1 when the timer is currently executing the callback function and
1149 : : * cannot be stopped
1150 : : */
1151 : 26527 : int hrtimer_try_to_cancel(struct hrtimer *timer)
1152 : : {
1153 : 26527 : struct hrtimer_clock_base *base;
1154 : 26527 : unsigned long flags;
1155 : 26527 : int ret = -1;
1156 : :
1157 : : /*
1158 : : * Check lockless first. If the timer is not active (neither
1159 : : * enqueued nor running the callback, nothing to do here. The
1160 : : * base lock does not serialize against a concurrent enqueue,
1161 : : * so we can avoid taking it.
1162 : : */
1163 [ + + ]: 26527 : if (!hrtimer_active(timer))
1164 : : return 0;
1165 : :
1166 : 4752 : base = lock_hrtimer_base(timer, &flags);
1167 : :
1168 [ + - ]: 4752 : if (!hrtimer_callback_running(timer))
1169 : 4752 : ret = remove_hrtimer(timer, base, false);
1170 : :
1171 : 4752 : unlock_hrtimer_base(timer, &flags);
1172 : :
1173 : 4752 : return ret;
1174 : :
1175 : : }
1176 : : EXPORT_SYMBOL_GPL(hrtimer_try_to_cancel);
1177 : :
1178 : : #ifdef CONFIG_PREEMPT_RT
1179 : : static void hrtimer_cpu_base_init_expiry_lock(struct hrtimer_cpu_base *base)
1180 : : {
1181 : : spin_lock_init(&base->softirq_expiry_lock);
1182 : : }
1183 : :
1184 : : static void hrtimer_cpu_base_lock_expiry(struct hrtimer_cpu_base *base)
1185 : : {
1186 : : spin_lock(&base->softirq_expiry_lock);
1187 : : }
1188 : :
1189 : : static void hrtimer_cpu_base_unlock_expiry(struct hrtimer_cpu_base *base)
1190 : : {
1191 : : spin_unlock(&base->softirq_expiry_lock);
1192 : : }
1193 : :
1194 : : /*
1195 : : * The counterpart to hrtimer_cancel_wait_running().
1196 : : *
1197 : : * If there is a waiter for cpu_base->expiry_lock, then it was waiting for
1198 : : * the timer callback to finish. Drop expiry_lock and reaquire it. That
1199 : : * allows the waiter to acquire the lock and make progress.
1200 : : */
1201 : : static void hrtimer_sync_wait_running(struct hrtimer_cpu_base *cpu_base,
1202 : : unsigned long flags)
1203 : : {
1204 : : if (atomic_read(&cpu_base->timer_waiters)) {
1205 : : raw_spin_unlock_irqrestore(&cpu_base->lock, flags);
1206 : : spin_unlock(&cpu_base->softirq_expiry_lock);
1207 : : spin_lock(&cpu_base->softirq_expiry_lock);
1208 : : raw_spin_lock_irq(&cpu_base->lock);
1209 : : }
1210 : : }
1211 : :
1212 : : /*
1213 : : * This function is called on PREEMPT_RT kernels when the fast path
1214 : : * deletion of a timer failed because the timer callback function was
1215 : : * running.
1216 : : *
1217 : : * This prevents priority inversion: if the soft irq thread is preempted
1218 : : * in the middle of a timer callback, then calling del_timer_sync() can
1219 : : * lead to two issues:
1220 : : *
1221 : : * - If the caller is on a remote CPU then it has to spin wait for the timer
1222 : : * handler to complete. This can result in unbound priority inversion.
1223 : : *
1224 : : * - If the caller originates from the task which preempted the timer
1225 : : * handler on the same CPU, then spin waiting for the timer handler to
1226 : : * complete is never going to end.
1227 : : */
1228 : : void hrtimer_cancel_wait_running(const struct hrtimer *timer)
1229 : : {
1230 : : /* Lockless read. Prevent the compiler from reloading it below */
1231 : : struct hrtimer_clock_base *base = READ_ONCE(timer->base);
1232 : :
1233 : : /*
1234 : : * Just relax if the timer expires in hard interrupt context or if
1235 : : * it is currently on the migration base.
1236 : : */
1237 : : if (!timer->is_soft || is_migration_base(base)) {
1238 : : cpu_relax();
1239 : : return;
1240 : : }
1241 : :
1242 : : /*
1243 : : * Mark the base as contended and grab the expiry lock, which is
1244 : : * held by the softirq across the timer callback. Drop the lock
1245 : : * immediately so the softirq can expire the next timer. In theory
1246 : : * the timer could already be running again, but that's more than
1247 : : * unlikely and just causes another wait loop.
1248 : : */
1249 : : atomic_inc(&base->cpu_base->timer_waiters);
1250 : : spin_lock_bh(&base->cpu_base->softirq_expiry_lock);
1251 : : atomic_dec(&base->cpu_base->timer_waiters);
1252 : : spin_unlock_bh(&base->cpu_base->softirq_expiry_lock);
1253 : : }
1254 : : #else
1255 : : static inline void
1256 : 30 : hrtimer_cpu_base_init_expiry_lock(struct hrtimer_cpu_base *base) { }
1257 : : static inline void
1258 : 0 : hrtimer_cpu_base_lock_expiry(struct hrtimer_cpu_base *base) { }
1259 : : static inline void
1260 : 0 : hrtimer_cpu_base_unlock_expiry(struct hrtimer_cpu_base *base) { }
1261 : : static inline void hrtimer_sync_wait_running(struct hrtimer_cpu_base *base,
1262 : : unsigned long flags) { }
1263 : : #endif
1264 : :
1265 : : /**
1266 : : * hrtimer_cancel - cancel a timer and wait for the handler to finish.
1267 : : * @timer: the timer to be cancelled
1268 : : *
1269 : : * Returns:
1270 : : * 0 when the timer was not active
1271 : : * 1 when the timer was active
1272 : : */
1273 : 19454 : int hrtimer_cancel(struct hrtimer *timer)
1274 : : {
1275 : 20390 : int ret;
1276 : :
1277 : 20390 : do {
1278 : 20390 : ret = hrtimer_try_to_cancel(timer);
1279 : :
1280 [ - + - + : 20390 : if (ret < 0)
- + ]
1281 : 0 : hrtimer_cancel_wait_running(timer);
1282 [ - + - + : 20390 : } while (ret < 0);
- + ]
1283 : 20390 : return ret;
1284 : : }
1285 : : EXPORT_SYMBOL_GPL(hrtimer_cancel);
1286 : :
1287 : : /**
1288 : : * hrtimer_get_remaining - get remaining time for the timer
1289 : : * @timer: the timer to read
1290 : : * @adjust: adjust relative timers when CONFIG_TIME_LOW_RES=y
1291 : : */
1292 : 120 : ktime_t __hrtimer_get_remaining(const struct hrtimer *timer, bool adjust)
1293 : : {
1294 : 120 : unsigned long flags;
1295 : 120 : ktime_t rem;
1296 : :
1297 : 120 : lock_hrtimer_base(timer, &flags);
1298 : 120 : if (IS_ENABLED(CONFIG_TIME_LOW_RES) && adjust)
1299 : : rem = hrtimer_expires_remaining_adjusted(timer);
1300 : : else
1301 : 120 : rem = hrtimer_expires_remaining(timer);
1302 : 120 : unlock_hrtimer_base(timer, &flags);
1303 : :
1304 : 120 : return rem;
1305 : : }
1306 : : EXPORT_SYMBOL_GPL(__hrtimer_get_remaining);
1307 : :
1308 : : #ifdef CONFIG_NO_HZ_COMMON
1309 : : /**
1310 : : * hrtimer_get_next_event - get the time until next expiry event
1311 : : *
1312 : : * Returns the next expiry time or KTIME_MAX if no timer is pending.
1313 : : */
1314 : 835 : u64 hrtimer_get_next_event(void)
1315 : : {
1316 : 835 : struct hrtimer_cpu_base *cpu_base = this_cpu_ptr(&hrtimer_bases);
1317 : 835 : u64 expires = KTIME_MAX;
1318 : 835 : unsigned long flags;
1319 : :
1320 : 835 : raw_spin_lock_irqsave(&cpu_base->lock, flags);
1321 : :
1322 [ - + ]: 835 : if (!__hrtimer_hres_active(cpu_base))
1323 : 0 : expires = __hrtimer_get_next_event(cpu_base, HRTIMER_ACTIVE_ALL);
1324 : :
1325 : 835 : raw_spin_unlock_irqrestore(&cpu_base->lock, flags);
1326 : :
1327 : 835 : return expires;
1328 : : }
1329 : :
1330 : : /**
1331 : : * hrtimer_next_event_without - time until next expiry event w/o one timer
1332 : : * @exclude: timer to exclude
1333 : : *
1334 : : * Returns the next expiry time over all timers except for the @exclude one or
1335 : : * KTIME_MAX if none of them is pending.
1336 : : */
1337 : 0 : u64 hrtimer_next_event_without(const struct hrtimer *exclude)
1338 : : {
1339 : 0 : struct hrtimer_cpu_base *cpu_base = this_cpu_ptr(&hrtimer_bases);
1340 : 0 : u64 expires = KTIME_MAX;
1341 : 0 : unsigned long flags;
1342 : :
1343 : 0 : raw_spin_lock_irqsave(&cpu_base->lock, flags);
1344 : :
1345 [ # # ]: 0 : if (__hrtimer_hres_active(cpu_base)) {
1346 : 0 : unsigned int active;
1347 : :
1348 [ # # ]: 0 : if (!cpu_base->softirq_activated) {
1349 : 0 : active = cpu_base->active_bases & HRTIMER_ACTIVE_SOFT;
1350 : 0 : expires = __hrtimer_next_event_base(cpu_base, exclude,
1351 : : active, KTIME_MAX);
1352 : : }
1353 : 0 : active = cpu_base->active_bases & HRTIMER_ACTIVE_HARD;
1354 : 0 : expires = __hrtimer_next_event_base(cpu_base, exclude, active,
1355 : : expires);
1356 : : }
1357 : :
1358 : 0 : raw_spin_unlock_irqrestore(&cpu_base->lock, flags);
1359 : :
1360 : 0 : return expires;
1361 : : }
1362 : : #endif
1363 : :
1364 : 79142 : static inline int hrtimer_clockid_to_base(clockid_t clock_id)
1365 : : {
1366 [ + - ]: 79142 : if (likely(clock_id < MAX_CLOCKS)) {
1367 : 79142 : int base = hrtimer_clock_to_base_table[clock_id];
1368 : :
1369 [ - + ]: 79142 : if (likely(base != HRTIMER_MAX_CLOCK_BASES))
1370 : : return base;
1371 : : }
1372 : 0 : WARN(1, "Invalid clockid %d. Using MONOTONIC\n", clock_id);
1373 : 0 : return HRTIMER_BASE_MONOTONIC;
1374 : : }
1375 : :
1376 : 79142 : static void __hrtimer_init(struct hrtimer *timer, clockid_t clock_id,
1377 : : enum hrtimer_mode mode)
1378 : : {
1379 : 79142 : bool softtimer = !!(mode & HRTIMER_MODE_SOFT);
1380 : 79142 : struct hrtimer_cpu_base *cpu_base;
1381 : 79142 : int base;
1382 : :
1383 : : /*
1384 : : * On PREEMPT_RT enabled kernels hrtimers which are not explicitely
1385 : : * marked for hard interrupt expiry mode are moved into soft
1386 : : * interrupt context for latency reasons and because the callbacks
1387 : : * can invoke functions which might sleep on RT, e.g. spin_lock().
1388 : : */
1389 : 79142 : if (IS_ENABLED(CONFIG_PREEMPT_RT) && !(mode & HRTIMER_MODE_HARD))
1390 : : softtimer = true;
1391 : :
1392 : 79142 : memset(timer, 0, sizeof(struct hrtimer));
1393 : :
1394 : 79142 : cpu_base = raw_cpu_ptr(&hrtimer_bases);
1395 : :
1396 : : /*
1397 : : * POSIX magic: Relative CLOCK_REALTIME timers are not affected by
1398 : : * clock modifications, so they needs to become CLOCK_MONOTONIC to
1399 : : * ensure POSIX compliance.
1400 : : */
1401 [ + + ]: 79142 : if (clock_id == CLOCK_REALTIME && mode & HRTIMER_MODE_REL)
1402 : : clock_id = CLOCK_MONOTONIC;
1403 : :
1404 [ + + ]: 79142 : base = softtimer ? HRTIMER_MAX_CLOCK_BASES / 2 : 0;
1405 : 79142 : base += hrtimer_clockid_to_base(clock_id);
1406 : 79142 : timer->is_soft = softtimer;
1407 : 79142 : timer->is_hard = !softtimer;
1408 : 79142 : timer->base = &cpu_base->clock_base[base];
1409 : 79142 : timerqueue_init(&timer->node);
1410 : 79142 : }
1411 : :
1412 : : /**
1413 : : * hrtimer_init - initialize a timer to the given clock
1414 : : * @timer: the timer to be initialized
1415 : : * @clock_id: the clock to be used
1416 : : * @mode: The modes which are relevant for intitialization:
1417 : : * HRTIMER_MODE_ABS, HRTIMER_MODE_REL, HRTIMER_MODE_ABS_SOFT,
1418 : : * HRTIMER_MODE_REL_SOFT
1419 : : *
1420 : : * The PINNED variants of the above can be handed in,
1421 : : * but the PINNED bit is ignored as pinning happens
1422 : : * when the hrtimer is started
1423 : : */
1424 : 78140 : void hrtimer_init(struct hrtimer *timer, clockid_t clock_id,
1425 : : enum hrtimer_mode mode)
1426 : : {
1427 : 78140 : debug_init(timer, clock_id, mode);
1428 : 78140 : __hrtimer_init(timer, clock_id, mode);
1429 : 78140 : }
1430 : : EXPORT_SYMBOL_GPL(hrtimer_init);
1431 : :
1432 : : /*
1433 : : * A timer is active, when it is enqueued into the rbtree or the
1434 : : * callback function is running or it's in the state of being migrated
1435 : : * to another cpu.
1436 : : *
1437 : : * It is important for this function to not return a false negative.
1438 : : */
1439 : 121229 : bool hrtimer_active(const struct hrtimer *timer)
1440 : : {
1441 : 121229 : struct hrtimer_clock_base *base;
1442 : 121229 : unsigned int seq;
1443 : :
1444 : 121229 : do {
1445 : 121229 : base = READ_ONCE(timer->base);
1446 : 121229 : seq = raw_read_seqcount_begin(&base->seq);
1447 : :
1448 [ + + ]: 121229 : if (timer->state != HRTIMER_STATE_INACTIVE ||
1449 [ + - ]: 116447 : base->running == timer)
1450 : : return true;
1451 : :
1452 [ - + ]: 232894 : } while (read_seqcount_retry(&base->seq, seq) ||
1453 [ - + - + ]: 116447 : base != READ_ONCE(timer->base));
1454 : :
1455 : : return false;
1456 : : }
1457 : : EXPORT_SYMBOL_GPL(hrtimer_active);
1458 : :
1459 : : /*
1460 : : * The write_seqcount_barrier()s in __run_hrtimer() split the thing into 3
1461 : : * distinct sections:
1462 : : *
1463 : : * - queued: the timer is queued
1464 : : * - callback: the timer is being ran
1465 : : * - post: the timer is inactive or (re)queued
1466 : : *
1467 : : * On the read side we ensure we observe timer->state and cpu_base->running
1468 : : * from the same section, if anything changed while we looked at it, we retry.
1469 : : * This includes timer->base changing because sequence numbers alone are
1470 : : * insufficient for that.
1471 : : *
1472 : : * The sequence numbers are required because otherwise we could still observe
1473 : : * a false negative if the read side got smeared over multiple consequtive
1474 : : * __run_hrtimer() invocations.
1475 : : */
1476 : :
1477 : 91008 : static void __run_hrtimer(struct hrtimer_cpu_base *cpu_base,
1478 : : struct hrtimer_clock_base *base,
1479 : : struct hrtimer *timer, ktime_t *now,
1480 : : unsigned long flags) __must_hold(&cpu_base->lock)
1481 : : {
1482 : 91008 : enum hrtimer_restart (*fn)(struct hrtimer *);
1483 : 91008 : int restart;
1484 : :
1485 : 91008 : lockdep_assert_held(&cpu_base->lock);
1486 : :
1487 : 91008 : debug_deactivate(timer);
1488 : 91008 : base->running = timer;
1489 : :
1490 : : /*
1491 : : * Separate the ->running assignment from the ->state assignment.
1492 : : *
1493 : : * As with a regular write barrier, this ensures the read side in
1494 : : * hrtimer_active() cannot observe base->running == NULL &&
1495 : : * timer->state == INACTIVE.
1496 : : */
1497 : 91008 : raw_write_seqcount_barrier(&base->seq);
1498 : :
1499 : 91008 : __remove_hrtimer(timer, base, HRTIMER_STATE_INACTIVE, 0);
1500 : 91008 : fn = timer->function;
1501 : :
1502 : : /*
1503 : : * Clear the 'is relative' flag for the TIME_LOW_RES case. If the
1504 : : * timer is restarted with a period then it becomes an absolute
1505 : : * timer. If its not restarted it does not matter.
1506 : : */
1507 : 91008 : if (IS_ENABLED(CONFIG_TIME_LOW_RES))
1508 : : timer->is_rel = false;
1509 : :
1510 : : /*
1511 : : * The timer is marked as running in the CPU base, so it is
1512 : : * protected against migration to a different CPU even if the lock
1513 : : * is dropped.
1514 : : */
1515 : 91008 : raw_spin_unlock_irqrestore(&cpu_base->lock, flags);
1516 : 91008 : trace_hrtimer_expire_entry(timer, now);
1517 : 91008 : restart = fn(timer);
1518 : 91008 : trace_hrtimer_expire_exit(timer);
1519 : 91008 : raw_spin_lock_irq(&cpu_base->lock);
1520 : :
1521 : : /*
1522 : : * Note: We clear the running state after enqueue_hrtimer and
1523 : : * we do not reprogram the event hardware. Happens either in
1524 : : * hrtimer_start_range_ns() or in hrtimer_interrupt()
1525 : : *
1526 : : * Note: Because we dropped the cpu_base->lock above,
1527 : : * hrtimer_start_range_ns() can have popped in and enqueued the timer
1528 : : * for us already.
1529 : : */
1530 [ + + ]: 91008 : if (restart != HRTIMER_NORESTART &&
1531 [ + - ]: 90139 : !(timer->state & HRTIMER_STATE_ENQUEUED))
1532 : 90139 : enqueue_hrtimer(timer, base, HRTIMER_MODE_ABS);
1533 : :
1534 : : /*
1535 : : * Separate the ->running assignment from the ->state assignment.
1536 : : *
1537 : : * As with a regular write barrier, this ensures the read side in
1538 : : * hrtimer_active() cannot observe base->running.timer == NULL &&
1539 : : * timer->state == INACTIVE.
1540 : : */
1541 : 91008 : raw_write_seqcount_barrier(&base->seq);
1542 : :
1543 [ - + ]: 91008 : WARN_ON_ONCE(base->running != timer);
1544 : 91008 : base->running = NULL;
1545 : 91008 : }
1546 : :
1547 : 97906 : static void __hrtimer_run_queues(struct hrtimer_cpu_base *cpu_base, ktime_t now,
1548 : : unsigned long flags, unsigned int active_mask)
1549 : : {
1550 : 97906 : struct hrtimer_clock_base *base;
1551 : 97906 : unsigned int active = cpu_base->active_bases & active_mask;
1552 : :
1553 [ + + + - ]: 304334 : for_each_active_base(base, cpu_base, active) {
1554 : 206428 : struct timerqueue_node *node;
1555 : 206428 : ktime_t basenow;
1556 : :
1557 : 206428 : basenow = ktime_add(now, base->offset);
1558 : :
1559 [ + + ]: 297436 : while ((node = timerqueue_getnext(&base->active))) {
1560 : 297220 : struct hrtimer *timer;
1561 : :
1562 : 297220 : timer = container_of(node, struct hrtimer, node);
1563 : :
1564 : : /*
1565 : : * The immediate goal for using the softexpires is
1566 : : * minimizing wakeups, not running timers at the
1567 : : * earliest interrupt after their soft expiration.
1568 : : * This allows us to avoid using a Priority Search
1569 : : * Tree, which can answer a stabbing querry for
1570 : : * overlapping intervals and instead use the simple
1571 : : * BST we already have.
1572 : : * We don't add extra wakeups by delaying timers that
1573 : : * are right-of a not yet expired timer, because that
1574 : : * timer will have to trigger a wakeup anyway.
1575 : : */
1576 [ + + ]: 297220 : if (basenow < hrtimer_get_softexpires_tv64(timer))
1577 : : break;
1578 : :
1579 : 91008 : __run_hrtimer(cpu_base, base, timer, &basenow, flags);
1580 : 91008 : if (active_mask == HRTIMER_ACTIVE_SOFT)
1581 : : hrtimer_sync_wait_running(cpu_base, flags);
1582 : : }
1583 : : }
1584 : 97906 : }
1585 : :
1586 : 0 : static __latent_entropy void hrtimer_run_softirq(struct softirq_action *h)
1587 : : {
1588 : 0 : struct hrtimer_cpu_base *cpu_base = this_cpu_ptr(&hrtimer_bases);
1589 : 0 : unsigned long flags;
1590 : 0 : ktime_t now;
1591 : :
1592 : 0 : hrtimer_cpu_base_lock_expiry(cpu_base);
1593 : 0 : raw_spin_lock_irqsave(&cpu_base->lock, flags);
1594 : :
1595 : 0 : now = hrtimer_update_base(cpu_base);
1596 : 0 : __hrtimer_run_queues(cpu_base, now, flags, HRTIMER_ACTIVE_SOFT);
1597 : :
1598 : 0 : cpu_base->softirq_activated = 0;
1599 : 0 : hrtimer_update_softirq_timer(cpu_base, true);
1600 : :
1601 : 0 : raw_spin_unlock_irqrestore(&cpu_base->lock, flags);
1602 : 0 : hrtimer_cpu_base_unlock_expiry(cpu_base);
1603 : 0 : }
1604 : :
1605 : : #ifdef CONFIG_HIGH_RES_TIMERS
1606 : :
1607 : : /*
1608 : : * High resolution timer interrupt
1609 : : * Called with interrupts disabled
1610 : : */
1611 : 90960 : void hrtimer_interrupt(struct clock_event_device *dev)
1612 : : {
1613 : 90960 : struct hrtimer_cpu_base *cpu_base = this_cpu_ptr(&hrtimer_bases);
1614 : 90960 : ktime_t expires_next, now, entry_time, delta;
1615 : 90960 : unsigned long flags;
1616 : 90960 : int retries = 0;
1617 : :
1618 [ - + ]: 90960 : BUG_ON(!cpu_base->hres_active);
1619 : 90960 : cpu_base->nr_events++;
1620 : 90960 : dev->next_event = KTIME_MAX;
1621 : :
1622 : 90960 : raw_spin_lock_irqsave(&cpu_base->lock, flags);
1623 : 90960 : entry_time = now = hrtimer_update_base(cpu_base);
1624 : 90991 : retry:
1625 : 90991 : cpu_base->in_hrtirq = 1;
1626 : : /*
1627 : : * We set expires_next to KTIME_MAX here with cpu_base->lock
1628 : : * held to prevent that a timer is enqueued in our queue via
1629 : : * the migration code. This does not affect enqueueing of
1630 : : * timers which run their callback and need to be requeued on
1631 : : * this CPU.
1632 : : */
1633 : 90991 : cpu_base->expires_next = KTIME_MAX;
1634 : :
1635 [ - + ]: 90991 : if (!ktime_before(now, cpu_base->softirq_expires_next)) {
1636 : 0 : cpu_base->softirq_expires_next = KTIME_MAX;
1637 : 0 : cpu_base->softirq_activated = 1;
1638 : 0 : raise_softirq_irqoff(HRTIMER_SOFTIRQ);
1639 : : }
1640 : :
1641 : 90991 : __hrtimer_run_queues(cpu_base, now, flags, HRTIMER_ACTIVE_HARD);
1642 : :
1643 : : /* Reevaluate the clock bases for the next expiry */
1644 : 90991 : expires_next = __hrtimer_get_next_event(cpu_base, HRTIMER_ACTIVE_ALL);
1645 : : /*
1646 : : * Store the new expiry value so the migration code can verify
1647 : : * against it.
1648 : : */
1649 : 90991 : cpu_base->expires_next = expires_next;
1650 : 90991 : cpu_base->in_hrtirq = 0;
1651 : 90991 : raw_spin_unlock_irqrestore(&cpu_base->lock, flags);
1652 : :
1653 : : /* Reprogramming necessary ? */
1654 [ + + ]: 90991 : if (!tick_program_event(expires_next, 0)) {
1655 : 90960 : cpu_base->hang_detected = 0;
1656 : 90960 : return;
1657 : : }
1658 : :
1659 : : /*
1660 : : * The next timer was already expired due to:
1661 : : * - tracing
1662 : : * - long lasting callbacks
1663 : : * - being scheduled away when running in a VM
1664 : : *
1665 : : * We need to prevent that we loop forever in the hrtimer
1666 : : * interrupt routine. We give it 3 attempts to avoid
1667 : : * overreacting on some spurious event.
1668 : : *
1669 : : * Acquire base lock for updating the offsets and retrieving
1670 : : * the current time.
1671 : : */
1672 : 31 : raw_spin_lock_irqsave(&cpu_base->lock, flags);
1673 : 31 : now = hrtimer_update_base(cpu_base);
1674 : 31 : cpu_base->nr_retries++;
1675 [ + - ]: 31 : if (++retries < 3)
1676 : 31 : goto retry;
1677 : : /*
1678 : : * Give the system a chance to do something else than looping
1679 : : * here. We stored the entry time, so we know exactly how long
1680 : : * we spent here. We schedule the next event this amount of
1681 : : * time away.
1682 : : */
1683 : 0 : cpu_base->nr_hangs++;
1684 : 0 : cpu_base->hang_detected = 1;
1685 : 0 : raw_spin_unlock_irqrestore(&cpu_base->lock, flags);
1686 : :
1687 : 0 : delta = ktime_sub(now, entry_time);
1688 [ # # ]: 0 : if ((unsigned int)delta > cpu_base->max_hang_time)
1689 : 0 : cpu_base->max_hang_time = (unsigned int) delta;
1690 : : /*
1691 : : * Limit it to a sensible value as we enforce a longer
1692 : : * delay. Give the CPU at least 100ms to catch up.
1693 : : */
1694 [ # # ]: 0 : if (delta > 100 * NSEC_PER_MSEC)
1695 : 0 : expires_next = ktime_add_ns(now, 100 * NSEC_PER_MSEC);
1696 : : else
1697 : 0 : expires_next = ktime_add(now, delta);
1698 : 0 : tick_program_event(expires_next, 1);
1699 [ # # ]: 0 : pr_warn_once("hrtimer: interrupt took %llu ns\n", ktime_to_ns(delta));
1700 : : }
1701 : :
1702 : : /* called with interrupts disabled */
1703 : 0 : static inline void __hrtimer_peek_ahead_timers(void)
1704 : : {
1705 : 0 : struct tick_device *td;
1706 : :
1707 [ # # ]: 0 : if (!hrtimer_hres_active())
1708 : : return;
1709 : :
1710 : 0 : td = this_cpu_ptr(&tick_cpu_device);
1711 [ # # # # ]: 0 : if (td && td->evtdev)
1712 : 0 : hrtimer_interrupt(td->evtdev);
1713 : : }
1714 : :
1715 : : #else /* CONFIG_HIGH_RES_TIMERS */
1716 : :
1717 : : static inline void __hrtimer_peek_ahead_timers(void) { }
1718 : :
1719 : : #endif /* !CONFIG_HIGH_RES_TIMERS */
1720 : :
1721 : : /*
1722 : : * Called from run_local_timers in hardirq context every jiffy
1723 : : */
1724 : 97655 : void hrtimer_run_queues(void)
1725 : : {
1726 : 97655 : struct hrtimer_cpu_base *cpu_base = this_cpu_ptr(&hrtimer_bases);
1727 : 97655 : unsigned long flags;
1728 : 97655 : ktime_t now;
1729 : :
1730 [ + + ]: 97655 : if (__hrtimer_hres_active(cpu_base))
1731 : : return;
1732 : :
1733 : : /*
1734 : : * This _is_ ugly: We have to check periodically, whether we
1735 : : * can switch to highres and / or nohz mode. The clocksource
1736 : : * switch happens with xtime_lock held. Notification from
1737 : : * there only sets the check bit in the tick_oneshot code,
1738 : : * otherwise we might deadlock vs. xtime_lock.
1739 : : */
1740 [ + + ]: 6945 : if (tick_check_oneshot_change(!hrtimer_is_hres_enabled())) {
1741 : 30 : hrtimer_switch_to_hres();
1742 : 30 : return;
1743 : : }
1744 : :
1745 : 6915 : raw_spin_lock_irqsave(&cpu_base->lock, flags);
1746 : 6915 : now = hrtimer_update_base(cpu_base);
1747 : :
1748 [ - + ]: 6915 : if (!ktime_before(now, cpu_base->softirq_expires_next)) {
1749 : 0 : cpu_base->softirq_expires_next = KTIME_MAX;
1750 : 0 : cpu_base->softirq_activated = 1;
1751 : 0 : raise_softirq_irqoff(HRTIMER_SOFTIRQ);
1752 : : }
1753 : :
1754 : 6915 : __hrtimer_run_queues(cpu_base, now, flags, HRTIMER_ACTIVE_HARD);
1755 : 6915 : raw_spin_unlock_irqrestore(&cpu_base->lock, flags);
1756 : : }
1757 : :
1758 : : /*
1759 : : * Sleep related functions:
1760 : : */
1761 : 31 : static enum hrtimer_restart hrtimer_wakeup(struct hrtimer *timer)
1762 : : {
1763 : 31 : struct hrtimer_sleeper *t =
1764 : 31 : container_of(timer, struct hrtimer_sleeper, timer);
1765 : 31 : struct task_struct *task = t->task;
1766 : :
1767 : 31 : t->task = NULL;
1768 [ + - ]: 31 : if (task)
1769 : 31 : wake_up_process(task);
1770 : :
1771 : 31 : return HRTIMER_NORESTART;
1772 : : }
1773 : :
1774 : : /**
1775 : : * hrtimer_sleeper_start_expires - Start a hrtimer sleeper timer
1776 : : * @sl: sleeper to be started
1777 : : * @mode: timer mode abs/rel
1778 : : *
1779 : : * Wrapper around hrtimer_start_expires() for hrtimer_sleeper based timers
1780 : : * to allow PREEMPT_RT to tweak the delivery mode (soft/hardirq context)
1781 : : */
1782 : 1002 : void hrtimer_sleeper_start_expires(struct hrtimer_sleeper *sl,
1783 : : enum hrtimer_mode mode)
1784 : : {
1785 : : /*
1786 : : * Make the enqueue delivery mode check work on RT. If the sleeper
1787 : : * was initialized for hard interrupt delivery, force the mode bit.
1788 : : * This is a special case for hrtimer_sleepers because
1789 : : * hrtimer_init_sleeper() determines the delivery mode on RT so the
1790 : : * fiddling with this decision is avoided at the call sites.
1791 : : */
1792 : 1002 : if (IS_ENABLED(CONFIG_PREEMPT_RT) && sl->timer.is_hard)
1793 : : mode |= HRTIMER_MODE_HARD;
1794 : :
1795 : 0 : hrtimer_start_expires(&sl->timer, mode);
1796 : 0 : }
1797 : : EXPORT_SYMBOL_GPL(hrtimer_sleeper_start_expires);
1798 : :
1799 : 1002 : static void __hrtimer_init_sleeper(struct hrtimer_sleeper *sl,
1800 : : clockid_t clock_id, enum hrtimer_mode mode)
1801 : : {
1802 : : /*
1803 : : * On PREEMPT_RT enabled kernels hrtimers which are not explicitely
1804 : : * marked for hard interrupt expiry mode are moved into soft
1805 : : * interrupt context either for latency reasons or because the
1806 : : * hrtimer callback takes regular spinlocks or invokes other
1807 : : * functions which are not suitable for hard interrupt context on
1808 : : * PREEMPT_RT.
1809 : : *
1810 : : * The hrtimer_sleeper callback is RT compatible in hard interrupt
1811 : : * context, but there is a latency concern: Untrusted userspace can
1812 : : * spawn many threads which arm timers for the same expiry time on
1813 : : * the same CPU. That causes a latency spike due to the wakeup of
1814 : : * a gazillion threads.
1815 : : *
1816 : : * OTOH, priviledged real-time user space applications rely on the
1817 : : * low latency of hard interrupt wakeups. If the current task is in
1818 : : * a real-time scheduling class, mark the mode for hard interrupt
1819 : : * expiry.
1820 : : */
1821 : 1002 : if (IS_ENABLED(CONFIG_PREEMPT_RT)) {
1822 : : if (task_is_realtime(current) && !(mode & HRTIMER_MODE_SOFT))
1823 : : mode |= HRTIMER_MODE_HARD;
1824 : : }
1825 : :
1826 : 1002 : __hrtimer_init(&sl->timer, clock_id, mode);
1827 : 1002 : sl->timer.function = hrtimer_wakeup;
1828 : 1002 : sl->task = current;
1829 : : }
1830 : :
1831 : : /**
1832 : : * hrtimer_init_sleeper - initialize sleeper to the given clock
1833 : : * @sl: sleeper to be initialized
1834 : : * @clock_id: the clock to be used
1835 : : * @mode: timer mode abs/rel
1836 : : */
1837 : 1002 : void hrtimer_init_sleeper(struct hrtimer_sleeper *sl, clockid_t clock_id,
1838 : : enum hrtimer_mode mode)
1839 : : {
1840 : 1002 : debug_init(&sl->timer, clock_id, mode);
1841 : 1002 : __hrtimer_init_sleeper(sl, clock_id, mode);
1842 : :
1843 : 1002 : }
1844 : : EXPORT_SYMBOL_GPL(hrtimer_init_sleeper);
1845 : :
1846 : 0 : int nanosleep_copyout(struct restart_block *restart, struct timespec64 *ts)
1847 : : {
1848 [ # # # ]: 0 : switch(restart->nanosleep.type) {
1849 : : #ifdef CONFIG_COMPAT_32BIT_TIME
1850 : 0 : case TT_COMPAT:
1851 [ # # ]: 0 : if (put_old_timespec32(ts, restart->nanosleep.compat_rmtp))
1852 : 0 : return -EFAULT;
1853 : : break;
1854 : : #endif
1855 : 0 : case TT_NATIVE:
1856 [ # # ]: 0 : if (put_timespec64(ts, restart->nanosleep.rmtp))
1857 : 0 : return -EFAULT;
1858 : : break;
1859 : 0 : default:
1860 : 0 : BUG();
1861 : : }
1862 : : return -ERESTART_RESTARTBLOCK;
1863 : : }
1864 : :
1865 : 61 : static int __sched do_nanosleep(struct hrtimer_sleeper *t, enum hrtimer_mode mode)
1866 : : {
1867 : 61 : struct restart_block *restart;
1868 : :
1869 : 61 : do {
1870 : 61 : set_current_state(TASK_INTERRUPTIBLE);
1871 : 61 : hrtimer_sleeper_start_expires(t, mode);
1872 : :
1873 [ + - ]: 61 : if (likely(t->task))
1874 : 61 : freezable_schedule();
1875 : :
1876 : : hrtimer_cancel(&t->timer);
1877 : 31 : mode = HRTIMER_MODE_ABS;
1878 : :
1879 [ - + - - ]: 31 : } while (t->task && !signal_pending(current));
1880 : :
1881 [ - + ]: 31 : __set_current_state(TASK_RUNNING);
1882 : :
1883 [ - + ]: 31 : if (!t->task)
1884 : : return 0;
1885 : :
1886 [ # # ]: 0 : restart = ¤t->restart_block;
1887 [ # # ]: 0 : if (restart->nanosleep.type != TT_NONE) {
1888 : 0 : ktime_t rem = hrtimer_expires_remaining(&t->timer);
1889 : 0 : struct timespec64 rmt;
1890 : :
1891 [ # # ]: 0 : if (rem <= 0)
1892 : : return 0;
1893 : 0 : rmt = ktime_to_timespec64(rem);
1894 : :
1895 : 0 : return nanosleep_copyout(restart, &rmt);
1896 : : }
1897 : : return -ERESTART_RESTARTBLOCK;
1898 : : }
1899 : :
1900 : 0 : static long __sched hrtimer_nanosleep_restart(struct restart_block *restart)
1901 : : {
1902 : 0 : struct hrtimer_sleeper t;
1903 : 0 : int ret;
1904 : :
1905 : 0 : hrtimer_init_sleeper_on_stack(&t, restart->nanosleep.clockid,
1906 : : HRTIMER_MODE_ABS);
1907 : 0 : hrtimer_set_expires_tv64(&t.timer, restart->nanosleep.expires);
1908 : 0 : ret = do_nanosleep(&t, HRTIMER_MODE_ABS);
1909 : 0 : destroy_hrtimer_on_stack(&t.timer);
1910 : 0 : return ret;
1911 : : }
1912 : :
1913 : 61 : long hrtimer_nanosleep(ktime_t rqtp, const enum hrtimer_mode mode,
1914 : : const clockid_t clockid)
1915 : : {
1916 : 61 : struct restart_block *restart;
1917 : 61 : struct hrtimer_sleeper t;
1918 : 61 : int ret = 0;
1919 : 61 : u64 slack;
1920 : :
1921 [ + - ]: 61 : slack = current->timer_slack_ns;
1922 [ + - ]: 61 : if (dl_task(current) || rt_task(current))
1923 : : slack = 0;
1924 : :
1925 : 61 : hrtimer_init_sleeper_on_stack(&t, clockid, mode);
1926 [ + - ]: 61 : hrtimer_set_expires_range_ns(&t.timer, rqtp, slack);
1927 : 61 : ret = do_nanosleep(&t, mode);
1928 [ + - ]: 31 : if (ret != -ERESTART_RESTARTBLOCK)
1929 : 31 : goto out;
1930 : :
1931 : : /* Absolute timers do not update the rmtp value and restart: */
1932 [ # # ]: 0 : if (mode == HRTIMER_MODE_ABS) {
1933 : 0 : ret = -ERESTARTNOHAND;
1934 : 0 : goto out;
1935 : : }
1936 : :
1937 : 0 : restart = ¤t->restart_block;
1938 : 0 : restart->fn = hrtimer_nanosleep_restart;
1939 : 0 : restart->nanosleep.clockid = t.timer.base->clockid;
1940 : 0 : restart->nanosleep.expires = hrtimer_get_expires_tv64(&t.timer);
1941 : 31 : out:
1942 : 31 : destroy_hrtimer_on_stack(&t.timer);
1943 : 31 : return ret;
1944 : : }
1945 : :
1946 : : #ifdef CONFIG_64BIT
1947 : :
1948 : 122 : SYSCALL_DEFINE2(nanosleep, struct __kernel_timespec __user *, rqtp,
1949 : : struct __kernel_timespec __user *, rmtp)
1950 : : {
1951 : 61 : struct timespec64 tu;
1952 : :
1953 [ + - ]: 61 : if (get_timespec64(&tu, rqtp))
1954 : : return -EFAULT;
1955 : :
1956 [ + - ]: 61 : if (!timespec64_valid(&tu))
1957 : : return -EINVAL;
1958 : :
1959 [ + - ]: 61 : current->restart_block.nanosleep.type = rmtp ? TT_NATIVE : TT_NONE;
1960 : 61 : current->restart_block.nanosleep.rmtp = rmtp;
1961 [ + - ]: 122 : return hrtimer_nanosleep(timespec64_to_ktime(tu), HRTIMER_MODE_REL,
1962 : : CLOCK_MONOTONIC);
1963 : : }
1964 : :
1965 : : #endif
1966 : :
1967 : : #ifdef CONFIG_COMPAT_32BIT_TIME
1968 : :
1969 : 0 : SYSCALL_DEFINE2(nanosleep_time32, struct old_timespec32 __user *, rqtp,
1970 : : struct old_timespec32 __user *, rmtp)
1971 : : {
1972 : 0 : struct timespec64 tu;
1973 : :
1974 [ # # ]: 0 : if (get_old_timespec32(&tu, rqtp))
1975 : : return -EFAULT;
1976 : :
1977 [ # # ]: 0 : if (!timespec64_valid(&tu))
1978 : : return -EINVAL;
1979 : :
1980 [ # # ]: 0 : current->restart_block.nanosleep.type = rmtp ? TT_COMPAT : TT_NONE;
1981 : 0 : current->restart_block.nanosleep.compat_rmtp = rmtp;
1982 [ # # ]: 0 : return hrtimer_nanosleep(timespec64_to_ktime(tu), HRTIMER_MODE_REL,
1983 : : CLOCK_MONOTONIC);
1984 : : }
1985 : : #endif
1986 : :
1987 : : /*
1988 : : * Functions related to boot-time initialization:
1989 : : */
1990 : 30 : int hrtimers_prepare_cpu(unsigned int cpu)
1991 : : {
1992 : 30 : struct hrtimer_cpu_base *cpu_base = &per_cpu(hrtimer_bases, cpu);
1993 : 30 : int i;
1994 : :
1995 [ + + ]: 270 : for (i = 0; i < HRTIMER_MAX_CLOCK_BASES; i++) {
1996 : 240 : cpu_base->clock_base[i].cpu_base = cpu_base;
1997 : 240 : timerqueue_init_head(&cpu_base->clock_base[i].active);
1998 : : }
1999 : :
2000 : 30 : cpu_base->cpu = cpu;
2001 : 30 : cpu_base->active_bases = 0;
2002 : 30 : cpu_base->hres_active = 0;
2003 : 30 : cpu_base->hang_detected = 0;
2004 : 30 : cpu_base->next_timer = NULL;
2005 : 30 : cpu_base->softirq_next_timer = NULL;
2006 : 30 : cpu_base->expires_next = KTIME_MAX;
2007 : 30 : cpu_base->softirq_expires_next = KTIME_MAX;
2008 : 30 : hrtimer_cpu_base_init_expiry_lock(cpu_base);
2009 : 30 : return 0;
2010 : : }
2011 : :
2012 : : #ifdef CONFIG_HOTPLUG_CPU
2013 : :
2014 : 0 : static void migrate_hrtimer_list(struct hrtimer_clock_base *old_base,
2015 : : struct hrtimer_clock_base *new_base)
2016 : : {
2017 : 0 : struct hrtimer *timer;
2018 : 0 : struct timerqueue_node *node;
2019 : :
2020 [ # # ]: 0 : while ((node = timerqueue_getnext(&old_base->active))) {
2021 : 0 : timer = container_of(node, struct hrtimer, node);
2022 [ # # ]: 0 : BUG_ON(hrtimer_callback_running(timer));
2023 : 0 : debug_deactivate(timer);
2024 : :
2025 : : /*
2026 : : * Mark it as ENQUEUED not INACTIVE otherwise the
2027 : : * timer could be seen as !active and just vanish away
2028 : : * under us on another CPU
2029 : : */
2030 : 0 : __remove_hrtimer(timer, old_base, HRTIMER_STATE_ENQUEUED, 0);
2031 : 0 : timer->base = new_base;
2032 : : /*
2033 : : * Enqueue the timers on the new cpu. This does not
2034 : : * reprogram the event device in case the timer
2035 : : * expires before the earliest on this CPU, but we run
2036 : : * hrtimer_interrupt after we migrated everything to
2037 : : * sort out already expired timers and reprogram the
2038 : : * event device.
2039 : : */
2040 : 0 : enqueue_hrtimer(timer, new_base, HRTIMER_MODE_ABS);
2041 : : }
2042 : 0 : }
2043 : :
2044 : 0 : int hrtimers_dead_cpu(unsigned int scpu)
2045 : : {
2046 : 0 : struct hrtimer_cpu_base *old_base, *new_base;
2047 : 0 : int i;
2048 : :
2049 [ # # ]: 0 : BUG_ON(cpu_online(scpu));
2050 : 0 : tick_cancel_sched_timer(scpu);
2051 : :
2052 : : /*
2053 : : * this BH disable ensures that raise_softirq_irqoff() does
2054 : : * not wakeup ksoftirqd (and acquire the pi-lock) while
2055 : : * holding the cpu_base lock
2056 : : */
2057 : 0 : local_bh_disable();
2058 : 0 : local_irq_disable();
2059 : 0 : old_base = &per_cpu(hrtimer_bases, scpu);
2060 : 0 : new_base = this_cpu_ptr(&hrtimer_bases);
2061 : : /*
2062 : : * The caller is globally serialized and nobody else
2063 : : * takes two locks at once, deadlock is not possible.
2064 : : */
2065 : 0 : raw_spin_lock(&new_base->lock);
2066 : 0 : raw_spin_lock_nested(&old_base->lock, SINGLE_DEPTH_NESTING);
2067 : :
2068 [ # # ]: 0 : for (i = 0; i < HRTIMER_MAX_CLOCK_BASES; i++) {
2069 : 0 : migrate_hrtimer_list(&old_base->clock_base[i],
2070 : : &new_base->clock_base[i]);
2071 : : }
2072 : :
2073 : : /*
2074 : : * The migration might have changed the first expiring softirq
2075 : : * timer on this CPU. Update it.
2076 : : */
2077 : 0 : hrtimer_update_softirq_timer(new_base, false);
2078 : :
2079 : 0 : raw_spin_unlock(&old_base->lock);
2080 : 0 : raw_spin_unlock(&new_base->lock);
2081 : :
2082 : : /* Check, if we got expired work to do */
2083 : 0 : __hrtimer_peek_ahead_timers();
2084 : 0 : local_irq_enable();
2085 : 0 : local_bh_enable();
2086 : 0 : return 0;
2087 : : }
2088 : :
2089 : : #endif /* CONFIG_HOTPLUG_CPU */
2090 : :
2091 : 30 : void __init hrtimers_init(void)
2092 : : {
2093 : 30 : hrtimers_prepare_cpu(smp_processor_id());
2094 : 30 : open_softirq(HRTIMER_SOFTIRQ, hrtimer_run_softirq);
2095 : 30 : }
2096 : :
2097 : : /**
2098 : : * schedule_hrtimeout_range_clock - sleep until timeout
2099 : : * @expires: timeout value (ktime_t)
2100 : : * @delta: slack in expires timeout (ktime_t)
2101 : : * @mode: timer mode
2102 : : * @clock_id: timer clock to be used
2103 : : */
2104 : : int __sched
2105 : 25412 : schedule_hrtimeout_range_clock(ktime_t *expires, u64 delta,
2106 : : const enum hrtimer_mode mode, clockid_t clock_id)
2107 : : {
2108 : 25412 : struct hrtimer_sleeper t;
2109 : :
2110 : : /*
2111 : : * Optimize when a zero timeout value is given. It does not
2112 : : * matter whether this is an absolute or a relative time.
2113 : : */
2114 [ + + - + ]: 25412 : if (expires && *expires == 0) {
2115 : 0 : __set_current_state(TASK_RUNNING);
2116 : 0 : return 0;
2117 : : }
2118 : :
2119 : : /*
2120 : : * A NULL parameter means "infinite"
2121 : : */
2122 [ + + ]: 25412 : if (!expires) {
2123 : 24471 : schedule();
2124 : 24471 : return -EINTR;
2125 : : }
2126 : :
2127 : 941 : hrtimer_init_sleeper_on_stack(&t, clock_id, mode);
2128 [ + - ]: 941 : hrtimer_set_expires_range_ns(&t.timer, *expires, delta);
2129 : 941 : hrtimer_sleeper_start_expires(&t, mode);
2130 : :
2131 [ + - ]: 941 : if (likely(t.task))
2132 : 941 : schedule();
2133 : :
2134 : : hrtimer_cancel(&t.timer);
2135 [ + - ]: 905 : destroy_hrtimer_on_stack(&t.timer);
2136 : :
2137 [ + - ]: 905 : __set_current_state(TASK_RUNNING);
2138 : :
2139 [ + - ]: 905 : return !t.task ? 0 : -EINTR;
2140 : : }
2141 : :
2142 : : /**
2143 : : * schedule_hrtimeout_range - sleep until timeout
2144 : : * @expires: timeout value (ktime_t)
2145 : : * @delta: slack in expires timeout (ktime_t)
2146 : : * @mode: timer mode
2147 : : *
2148 : : * Make the current task sleep until the given expiry time has
2149 : : * elapsed. The routine will return immediately unless
2150 : : * the current task state has been set (see set_current_state()).
2151 : : *
2152 : : * The @delta argument gives the kernel the freedom to schedule the
2153 : : * actual wakeup to a time that is both power and performance friendly.
2154 : : * The kernel give the normal best effort behavior for "@expires+@delta",
2155 : : * but may decide to fire the timer earlier, but no earlier than @expires.
2156 : : *
2157 : : * You can set the task state as follows -
2158 : : *
2159 : : * %TASK_UNINTERRUPTIBLE - at least @timeout time is guaranteed to
2160 : : * pass before the routine returns unless the current task is explicitly
2161 : : * woken up, (e.g. by wake_up_process()).
2162 : : *
2163 : : * %TASK_INTERRUPTIBLE - the routine may return early if a signal is
2164 : : * delivered to the current task or the current task is explicitly woken
2165 : : * up.
2166 : : *
2167 : : * The current task state is guaranteed to be TASK_RUNNING when this
2168 : : * routine returns.
2169 : : *
2170 : : * Returns 0 when the timer has expired. If the task was woken before the
2171 : : * timer expired by a signal (only possible in state TASK_INTERRUPTIBLE) or
2172 : : * by an explicit wakeup, it returns -EINTR.
2173 : : */
2174 : 25412 : int __sched schedule_hrtimeout_range(ktime_t *expires, u64 delta,
2175 : : const enum hrtimer_mode mode)
2176 : : {
2177 : 25412 : return schedule_hrtimeout_range_clock(expires, delta, mode,
2178 : : CLOCK_MONOTONIC);
2179 : : }
2180 : : EXPORT_SYMBOL_GPL(schedule_hrtimeout_range);
2181 : :
2182 : : /**
2183 : : * schedule_hrtimeout - sleep until timeout
2184 : : * @expires: timeout value (ktime_t)
2185 : : * @mode: timer mode
2186 : : *
2187 : : * Make the current task sleep until the given expiry time has
2188 : : * elapsed. The routine will return immediately unless
2189 : : * the current task state has been set (see set_current_state()).
2190 : : *
2191 : : * You can set the task state as follows -
2192 : : *
2193 : : * %TASK_UNINTERRUPTIBLE - at least @timeout time is guaranteed to
2194 : : * pass before the routine returns unless the current task is explicitly
2195 : : * woken up, (e.g. by wake_up_process()).
2196 : : *
2197 : : * %TASK_INTERRUPTIBLE - the routine may return early if a signal is
2198 : : * delivered to the current task or the current task is explicitly woken
2199 : : * up.
2200 : : *
2201 : : * The current task state is guaranteed to be TASK_RUNNING when this
2202 : : * routine returns.
2203 : : *
2204 : : * Returns 0 when the timer has expired. If the task was woken before the
2205 : : * timer expired by a signal (only possible in state TASK_INTERRUPTIBLE) or
2206 : : * by an explicit wakeup, it returns -EINTR.
2207 : : */
2208 : 0 : int __sched schedule_hrtimeout(ktime_t *expires,
2209 : : const enum hrtimer_mode mode)
2210 : : {
2211 : 0 : return schedule_hrtimeout_range(expires, 0, mode);
2212 : : }
2213 : : EXPORT_SYMBOL_GPL(schedule_hrtimeout);
|