/* * Copyright (c) 2015 Wind River Systems, Inc. * * SPDX-License-Identifier: Apache-2.0 */ #ifndef ZEPHYR_KERNEL_INCLUDE_TIMEOUT_Q_H_ #define ZEPHYR_KERNEL_INCLUDE_TIMEOUT_Q_H_ /** * @file * @brief timeout queue for threads on kernel objects * * This file is meant to be included by kernel/include/wait_q.h only */ #include #include #ifdef __cplusplus extern "C" { #endif /* initialize the timeouts part of k_thread when enabled in the kernel */ static inline void _init_timeout(struct _timeout *t, _timeout_func_t func) { /* * Must be initialized here and when dequeueing a timeout so that code * not dealing with timeouts does not have to handle this, such as when * waiting forever on a semaphore. */ t->delta_ticks_from_prev = _INACTIVE; /* * Must be initialized here so that k_wakeup can * verify the thread is not on a wait queue before aborting a timeout. */ t->wait_q = NULL; /* * Must be initialized here, so the _handle_one_timeout() * routine can check if there is a thread waiting on this timeout */ t->thread = NULL; /* * Function must be initialized before being potentially called. */ t->func = func; /* * These are initialized when enqueing on the timeout queue: * * thread->timeout.node.next * thread->timeout.node.prev */ } static ALWAYS_INLINE void _init_thread_timeout(struct _thread_base *thread_base) { _init_timeout(&thread_base->timeout, NULL); } /* remove a thread timing out from kernel object's wait queue */ static inline void _unpend_thread_timing_out(struct k_thread *thread, struct _timeout *timeout_obj) { if (timeout_obj->wait_q != NULL) { _unpend_thread_no_timeout(thread); thread->base.timeout.wait_q = NULL; } } /* * Handle one timeout from the expired timeout queue. Removes it from the wait * queue it is on if waiting for an object; in this case, the return value is * kept as -EAGAIN, set previously in _Swap(). */ static inline void _handle_one_expired_timeout(struct _timeout *timeout) { struct k_thread *thread = timeout->thread; unsigned int key = irq_lock(); timeout->delta_ticks_from_prev = _INACTIVE; K_DEBUG("timeout %p\n", timeout); if (thread != NULL) { _unpend_thread_timing_out(thread, timeout); _mark_thread_as_started(thread); _ready_thread(thread); irq_unlock(key); } else { irq_unlock(key); if (timeout->func != NULL) { timeout->func(timeout); } } } /* * Loop over all expired timeouts and handle them one by one. Should be called * with interrupts unlocked: interrupts will be locked on each interation only * for the amount of time necessary. */ static inline void _handle_expired_timeouts(sys_dlist_t *expired) { struct _timeout *timeout, *next; SYS_DLIST_FOR_EACH_CONTAINER_SAFE(expired, timeout, next, node) { _handle_one_expired_timeout(timeout); } } /* returns _INACTIVE if the timer is not active */ static inline int _abort_timeout(struct _timeout *timeout) { if (timeout->delta_ticks_from_prev == _INACTIVE) { return _INACTIVE; } if (!sys_dlist_is_tail(&_timeout_q, &timeout->node)) { sys_dnode_t *next_node = sys_dlist_peek_next(&_timeout_q, &timeout->node); struct _timeout *next = (struct _timeout *)next_node; next->delta_ticks_from_prev += timeout->delta_ticks_from_prev; } sys_dlist_remove(&timeout->node); timeout->delta_ticks_from_prev = _INACTIVE; return 0; } /* returns _INACTIVE if the timer has already expired */ static inline int _abort_thread_timeout(struct k_thread *thread) { return _abort_timeout(&thread->base.timeout); } static inline void _dump_timeout(struct _timeout *timeout, int extra_tab) { #ifdef CONFIG_KERNEL_DEBUG char *tab = extra_tab ? "\t" : ""; K_DEBUG("%stimeout %p, prev: %p, next: %p\n" "%s\tthread: %p, wait_q: %p\n" "%s\tticks remaining: %d\n" "%s\tfunction: %p\n", tab, timeout, timeout->node.prev, timeout->node.next, tab, timeout->thread, timeout->wait_q, tab, timeout->delta_ticks_from_prev, tab, timeout->func); #endif } static inline void _dump_timeout_q(void) { #ifdef CONFIG_KERNEL_DEBUG struct _timeout *timeout; K_DEBUG("_timeout_q: %p, head: %p, tail: %p\n", &_timeout_q, _timeout_q.head, _timeout_q.tail); SYS_DLIST_FOR_EACH_CONTAINER(&_timeout_q, timeout, node) { _dump_timeout(timeout, 1); } #endif } /* * Add timeout to timeout queue. Record waiting thread and wait queue if any. * * Cannot handle timeout == 0 and timeout == K_FOREVER. * * If the new timeout is expiring on the same system clock tick as other * timeouts already present in the _timeout_q, it is be _prepended_ to these * timeouts. This allows exiting the loop sooner, which is good, since * interrupts are locked while trying to find the insert point. Note that the * timeouts are then processed in the _reverse order_ if they expire on the * same tick. * * This should not cause problems to applications, unless they really expect * two timeouts queued very close to one another to expire in the same order * they were queued. This could be changed at the cost of potential longer * interrupt latency. * * Must be called with interrupts locked. */ static inline void _add_timeout(struct k_thread *thread, struct _timeout *timeout, _wait_q_t *wait_q, s32_t timeout_in_ticks) { __ASSERT(timeout_in_ticks >= 0, ""); timeout->delta_ticks_from_prev = timeout_in_ticks; timeout->thread = thread; timeout->wait_q = (sys_dlist_t *)wait_q; K_DEBUG("before adding timeout %p\n", timeout); _dump_timeout(timeout, 0); _dump_timeout_q(); /* If timer is submitted to expire ASAP with * timeout_in_ticks (duration) as zero value, * then handle timeout immedately without going * through timeout queue. */ if (!timeout_in_ticks) { _handle_one_expired_timeout(timeout); return; } s32_t *delta = &timeout->delta_ticks_from_prev; struct _timeout *in_q; #ifdef CONFIG_TICKLESS_KERNEL /* * If some time has already passed since timer was last * programmed, then that time needs to be accounted when * inserting the new timeout. We account for this * by adding the already elapsed time to the new timeout. * This is like adding this timout back in history. */ u32_t adjusted_timeout; u32_t program_time = _get_program_time(); if (program_time > 0) { *delta += _get_elapsed_program_time(); } adjusted_timeout = *delta; #endif SYS_DLIST_FOR_EACH_CONTAINER(&_timeout_q, in_q, node) { if (*delta <= in_q->delta_ticks_from_prev) { in_q->delta_ticks_from_prev -= *delta; sys_dlist_insert_before(&_timeout_q, &in_q->node, &timeout->node); goto inserted; } *delta -= in_q->delta_ticks_from_prev; } sys_dlist_append(&_timeout_q, &timeout->node); inserted: K_DEBUG("after adding timeout %p\n", timeout); _dump_timeout(timeout, 0); _dump_timeout_q(); #ifdef CONFIG_TICKLESS_KERNEL if (!program_time || (adjusted_timeout < program_time)) { _set_time(adjusted_timeout); } #endif } /* * Put thread on timeout queue. Record wait queue if any. * * Cannot handle timeout == 0 and timeout == K_FOREVER. * * Must be called with interrupts locked. */ static inline void _add_thread_timeout(struct k_thread *thread, _wait_q_t *wait_q, s32_t timeout_in_ticks) { _add_timeout(thread, &thread->base.timeout, wait_q, timeout_in_ticks); } /* find the closest deadline in the timeout queue */ static inline s32_t _get_next_timeout_expiry(void) { struct _timeout *t = (struct _timeout *) sys_dlist_peek_head(&_timeout_q); return t ? t->delta_ticks_from_prev : K_FOREVER; } #ifdef __cplusplus } #endif #endif /* ZEPHYR_KERNEL_INCLUDE_TIMEOUT_Q_H_ */