294 lines
7.5 KiB
C
294 lines
7.5 KiB
C
/*
|
|
* Copyright (c) 2015 Wind River Systems, Inc.
|
|
*
|
|
* SPDX-License-Identifier: Apache-2.0
|
|
*/
|
|
|
|
#ifndef ZEPHYR_KERNEL_INCLUDE_TIMEOUT_Q_H_
|
|
#define ZEPHYR_KERNEL_INCLUDE_TIMEOUT_Q_H_
|
|
|
|
/**
|
|
* @file
|
|
* @brief timeout queue for threads on kernel objects
|
|
*
|
|
* This file is meant to be included by kernel/include/wait_q.h only
|
|
*/
|
|
|
|
#include <misc/dlist.h>
|
|
#include <drivers/system_timer.h>
|
|
|
|
#ifdef __cplusplus
|
|
extern "C" {
|
|
#endif
|
|
|
|
/* initialize the timeouts part of k_thread when enabled in the kernel */
|
|
|
|
static inline void _init_timeout(struct _timeout *t, _timeout_func_t func)
|
|
{
|
|
/*
|
|
* Must be initialized here and when dequeueing a timeout so that code
|
|
* not dealing with timeouts does not have to handle this, such as when
|
|
* waiting forever on a semaphore.
|
|
*/
|
|
t->delta_ticks_from_prev = _INACTIVE;
|
|
|
|
/*
|
|
* Must be initialized here so that k_wakeup can
|
|
* verify the thread is not on a wait queue before aborting a timeout.
|
|
*/
|
|
t->wait_q = NULL;
|
|
|
|
/*
|
|
* Must be initialized here, so the _handle_one_timeout()
|
|
* routine can check if there is a thread waiting on this timeout
|
|
*/
|
|
t->thread = NULL;
|
|
|
|
/*
|
|
* Function must be initialized before being potentially called.
|
|
*/
|
|
t->func = func;
|
|
|
|
/*
|
|
* These are initialized when enqueing on the timeout queue:
|
|
*
|
|
* thread->timeout.node.next
|
|
* thread->timeout.node.prev
|
|
*/
|
|
}
|
|
|
|
static ALWAYS_INLINE void
|
|
_init_thread_timeout(struct _thread_base *thread_base)
|
|
{
|
|
_init_timeout(&thread_base->timeout, NULL);
|
|
}
|
|
|
|
/* remove a thread timing out from kernel object's wait queue */
|
|
|
|
static inline void _unpend_thread_timing_out(struct k_thread *thread,
|
|
struct _timeout *timeout_obj)
|
|
{
|
|
if (timeout_obj->wait_q) {
|
|
_unpend_thread_no_timeout(thread);
|
|
thread->base.timeout.wait_q = NULL;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Handle one timeout from the expired timeout queue. Removes it from the wait
|
|
* queue it is on if waiting for an object; in this case, the return value is
|
|
* kept as -EAGAIN, set previously in _Swap().
|
|
*/
|
|
|
|
static inline void _handle_one_expired_timeout(struct _timeout *timeout)
|
|
{
|
|
struct k_thread *thread = timeout->thread;
|
|
unsigned int key = irq_lock();
|
|
|
|
timeout->delta_ticks_from_prev = _INACTIVE;
|
|
|
|
K_DEBUG("timeout %p\n", timeout);
|
|
if (thread) {
|
|
_unpend_thread_timing_out(thread, timeout);
|
|
_mark_thread_as_started(thread);
|
|
_ready_thread(thread);
|
|
irq_unlock(key);
|
|
} else {
|
|
irq_unlock(key);
|
|
if (timeout->func) {
|
|
timeout->func(timeout);
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Loop over all expired timeouts and handle them one by one. Should be called
|
|
* with interrupts unlocked: interrupts will be locked on each interation only
|
|
* for the amount of time necessary.
|
|
*/
|
|
|
|
static inline void _handle_expired_timeouts(sys_dlist_t *expired)
|
|
{
|
|
struct _timeout *timeout;
|
|
|
|
SYS_DLIST_FOR_EACH_CONTAINER(expired, timeout, node) {
|
|
_handle_one_expired_timeout(timeout);
|
|
}
|
|
}
|
|
|
|
/* returns _INACTIVE if the timer is not active */
|
|
static inline int _abort_timeout(struct _timeout *timeout)
|
|
{
|
|
if (timeout->delta_ticks_from_prev == _INACTIVE) {
|
|
return _INACTIVE;
|
|
}
|
|
|
|
if (!sys_dlist_is_tail(&_timeout_q, &timeout->node)) {
|
|
sys_dnode_t *next_node =
|
|
sys_dlist_peek_next(&_timeout_q, &timeout->node);
|
|
struct _timeout *next = (struct _timeout *)next_node;
|
|
|
|
next->delta_ticks_from_prev += timeout->delta_ticks_from_prev;
|
|
}
|
|
sys_dlist_remove(&timeout->node);
|
|
timeout->delta_ticks_from_prev = _INACTIVE;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* returns _INACTIVE if the timer has already expired */
|
|
static inline int _abort_thread_timeout(struct k_thread *thread)
|
|
{
|
|
return _abort_timeout(&thread->base.timeout);
|
|
}
|
|
|
|
static inline void _dump_timeout(struct _timeout *timeout, int extra_tab)
|
|
{
|
|
#ifdef CONFIG_KERNEL_DEBUG
|
|
char *tab = extra_tab ? "\t" : "";
|
|
|
|
K_DEBUG("%stimeout %p, prev: %p, next: %p\n"
|
|
"%s\tthread: %p, wait_q: %p\n"
|
|
"%s\tticks remaining: %d\n"
|
|
"%s\tfunction: %p\n",
|
|
tab, timeout, timeout->node.prev, timeout->node.next,
|
|
tab, timeout->thread, timeout->wait_q,
|
|
tab, timeout->delta_ticks_from_prev,
|
|
tab, timeout->func);
|
|
#endif
|
|
}
|
|
|
|
static inline void _dump_timeout_q(void)
|
|
{
|
|
#ifdef CONFIG_KERNEL_DEBUG
|
|
struct _timeout *timeout;
|
|
|
|
K_DEBUG("_timeout_q: %p, head: %p, tail: %p\n",
|
|
&_timeout_q, _timeout_q.head, _timeout_q.tail);
|
|
|
|
SYS_DLIST_FOR_EACH_CONTAINER(&_timeout_q, timeout, node) {
|
|
_dump_timeout(timeout, 1);
|
|
}
|
|
#endif
|
|
}
|
|
|
|
/*
|
|
* Add timeout to timeout queue. Record waiting thread and wait queue if any.
|
|
*
|
|
* Cannot handle timeout == 0 and timeout == K_FOREVER.
|
|
*
|
|
* If the new timeout is expiring on the same system clock tick as other
|
|
* timeouts already present in the _timeout_q, it is be _prepended_ to these
|
|
* timeouts. This allows exiting the loop sooner, which is good, since
|
|
* interrupts are locked while trying to find the insert point. Note that the
|
|
* timeouts are then processed in the _reverse order_ if they expire on the
|
|
* same tick.
|
|
*
|
|
* This should not cause problems to applications, unless they really expect
|
|
* two timeouts queued very close to one another to expire in the same order
|
|
* they were queued. This could be changed at the cost of potential longer
|
|
* interrupt latency.
|
|
*
|
|
* Must be called with interrupts locked.
|
|
*/
|
|
|
|
static inline void _add_timeout(struct k_thread *thread,
|
|
struct _timeout *timeout,
|
|
_wait_q_t *wait_q,
|
|
s32_t timeout_in_ticks)
|
|
{
|
|
__ASSERT(timeout_in_ticks >= 0, "");
|
|
|
|
timeout->delta_ticks_from_prev = timeout_in_ticks;
|
|
timeout->thread = thread;
|
|
timeout->wait_q = (sys_dlist_t *)wait_q;
|
|
|
|
K_DEBUG("before adding timeout %p\n", timeout);
|
|
_dump_timeout(timeout, 0);
|
|
_dump_timeout_q();
|
|
|
|
/* If timer is submitted to expire ASAP with
|
|
* timeout_in_ticks (duration) as zero value,
|
|
* then handle timeout immedately without going
|
|
* through timeout queue.
|
|
*/
|
|
if (!timeout_in_ticks) {
|
|
_handle_one_expired_timeout(timeout);
|
|
return;
|
|
}
|
|
|
|
s32_t *delta = &timeout->delta_ticks_from_prev;
|
|
struct _timeout *in_q;
|
|
|
|
#ifdef CONFIG_TICKLESS_KERNEL
|
|
/*
|
|
* If some time has already passed since timer was last
|
|
* programmed, then that time needs to be accounted when
|
|
* inserting the new timeout. We account for this
|
|
* by adding the already elapsed time to the new timeout.
|
|
* This is like adding this timout back in history.
|
|
*/
|
|
u32_t adjusted_timeout;
|
|
u32_t program_time = _get_program_time();
|
|
|
|
if (program_time > 0) {
|
|
*delta += _get_elapsed_program_time();
|
|
}
|
|
adjusted_timeout = *delta;
|
|
#endif
|
|
SYS_DLIST_FOR_EACH_CONTAINER(&_timeout_q, in_q, node) {
|
|
if (*delta <= in_q->delta_ticks_from_prev) {
|
|
in_q->delta_ticks_from_prev -= *delta;
|
|
sys_dlist_insert_before(&_timeout_q, &in_q->node,
|
|
&timeout->node);
|
|
goto inserted;
|
|
}
|
|
|
|
*delta -= in_q->delta_ticks_from_prev;
|
|
}
|
|
|
|
sys_dlist_append(&_timeout_q, &timeout->node);
|
|
|
|
inserted:
|
|
K_DEBUG("after adding timeout %p\n", timeout);
|
|
_dump_timeout(timeout, 0);
|
|
_dump_timeout_q();
|
|
|
|
#ifdef CONFIG_TICKLESS_KERNEL
|
|
if (!program_time || (adjusted_timeout < program_time)) {
|
|
_set_time(adjusted_timeout);
|
|
}
|
|
#endif
|
|
}
|
|
|
|
/*
|
|
* Put thread on timeout queue. Record wait queue if any.
|
|
*
|
|
* Cannot handle timeout == 0 and timeout == K_FOREVER.
|
|
*
|
|
* Must be called with interrupts locked.
|
|
*/
|
|
|
|
static inline void _add_thread_timeout(struct k_thread *thread,
|
|
_wait_q_t *wait_q,
|
|
s32_t timeout_in_ticks)
|
|
{
|
|
_add_timeout(thread, &thread->base.timeout, wait_q, timeout_in_ticks);
|
|
}
|
|
|
|
/* find the closest deadline in the timeout queue */
|
|
|
|
static inline s32_t _get_next_timeout_expiry(void)
|
|
{
|
|
struct _timeout *t = (struct _timeout *)
|
|
sys_dlist_peek_head(&_timeout_q);
|
|
|
|
return t ? t->delta_ticks_from_prev : K_FOREVER;
|
|
}
|
|
|
|
#ifdef __cplusplus
|
|
}
|
|
#endif
|
|
|
|
#endif /* ZEPHYR_KERNEL_INCLUDE_TIMEOUT_Q_H_ */
|