167 lines
5.5 KiB
C
167 lines
5.5 KiB
C
/*! @file
|
|
@brief timeout queue for fibers on nanokernel objects
|
|
|
|
This file is meant to be included by nanokernel/include/wait_q.h only
|
|
*/
|
|
|
|
/*
|
|
* Copyright (c) 2015 Wind River Systems, Inc.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are met:
|
|
*
|
|
* 1) Redistributions of source code must retain the above copyright notice,
|
|
* this list of conditions and the following disclaimer.
|
|
*
|
|
* 2) Redistributions in binary form must reproduce the above copyright notice,
|
|
* this list of conditions and the following disclaimer in the documentation
|
|
* and/or other materials provided with the distribution.
|
|
*
|
|
* 3) Neither the name of Wind River Systems nor the names of its contributors
|
|
* may be used to endorse or promote products derived from this software without
|
|
* specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE
|
|
* LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
* POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#ifndef _kernel_nanokernel_include_timeout_q__h_
|
|
#define _kernel_nanokernel_include_timeout_q__h_
|
|
|
|
#include <misc/dlist.h>
|
|
|
|
/* initialize the nano timeouts part of CCS when enabled in the kernel */
|
|
|
|
static inline void _nano_timeout_ccs_init(struct ccs *ccs)
|
|
{
|
|
/*
|
|
* Must be initialized here and when dequeueing a timeout so that code
|
|
* not dealing with timeouts does not have to handle this, such as when
|
|
* waiting forever on a semaphore.
|
|
*/
|
|
ccs->nano_timeout.delta_ticks_from_prev = -1;
|
|
|
|
/*
|
|
* These are initialized when enqueing on the timeout queue:
|
|
*
|
|
* ccs->nano_timeout.node.next
|
|
* ccs->nano_timeout.node.prev
|
|
* ccs->nano_timeout.wait_q
|
|
*/
|
|
}
|
|
|
|
/*
|
|
* Handle one expired timeout.
|
|
* This removes the fiber from the timeout queue head, and also removes it
|
|
* from the wait queue it is on if waiting for an object. In that case, it
|
|
* also sets the return value to 0/NULL.
|
|
*/
|
|
|
|
static inline struct _nano_timeout *_nano_timeout_handle_one_timeout(
|
|
sys_dlist_t *timeout_q)
|
|
{
|
|
struct _nano_timeout *t = (void *)sys_dlist_get(timeout_q);
|
|
struct ccs *ccs = CONTAINER_OF(t, struct ccs, nano_timeout);
|
|
|
|
if (ccs->nano_timeout.wait_q) {
|
|
_nano_timeout_remove_ccs_from_wait_q(ccs);
|
|
fiberRtnValueSet(ccs, (unsigned int)0);
|
|
}
|
|
_nano_fiber_schedule(ccs);
|
|
t->delta_ticks_from_prev = -1;
|
|
|
|
return (struct _nano_timeout *)sys_dlist_peek_head(timeout_q);
|
|
}
|
|
|
|
/* loop over all expired timeouts and handle them one by one */
|
|
static inline void _nano_timeout_handle_timeouts(void)
|
|
{
|
|
sys_dlist_t *timeout_q = &_nanokernel.timeout_q;
|
|
struct _nano_timeout *next;
|
|
|
|
next = (struct _nano_timeout *)sys_dlist_peek_head(timeout_q);
|
|
while (next && next->delta_ticks_from_prev == 0) {
|
|
next = _nano_timeout_handle_one_timeout(timeout_q);
|
|
}
|
|
}
|
|
|
|
/* abort a timeout for a specific fiber */
|
|
static inline void _nano_timeout_abort(struct ccs *ccs)
|
|
{
|
|
sys_dlist_t *timeout_q = &_nanokernel.timeout_q;
|
|
struct _nano_timeout *t = &ccs->nano_timeout;
|
|
|
|
if (-1 == t->delta_ticks_from_prev) {
|
|
return;
|
|
}
|
|
|
|
if (!sys_dlist_is_tail(timeout_q, &t->node)) {
|
|
struct _nano_timeout *next =
|
|
(struct _nano_timeout *)sys_dlist_peek_next(timeout_q, &t->node);
|
|
next->delta_ticks_from_prev += t->delta_ticks_from_prev;
|
|
}
|
|
sys_dlist_remove(&t->node);
|
|
t->delta_ticks_from_prev = -1;
|
|
}
|
|
|
|
/*
|
|
* callback for sys_dlist_insert_at():
|
|
*
|
|
* Returns 1 if the timeout to insert is lower or equal than the next timeout
|
|
* in the queue, signifying that it should be inserted before the next.
|
|
* Returns 0 if it is greater.
|
|
*
|
|
* If it is greater, the timeout to insert is decremented by the next timeout,
|
|
* since the timeout queue is a delta queue. If it lower or equal, decrement
|
|
* the timeout of the insert point to update its delta queue value, since the
|
|
* current timeout will be inserted before it.
|
|
*/
|
|
static int _nano_timeout_insert_point_test(sys_dnode_t *test, void *timeout)
|
|
{
|
|
struct _nano_timeout *t = (void *)test;
|
|
int32_t *timeout_to_insert = timeout;
|
|
|
|
if (*timeout_to_insert > t->delta_ticks_from_prev) {
|
|
*timeout_to_insert -= t->delta_ticks_from_prev;
|
|
return 0;
|
|
}
|
|
|
|
t->delta_ticks_from_prev -= *timeout_to_insert;
|
|
return 1;
|
|
}
|
|
|
|
/* put a fiber on the timeout queue and record its wait queue */
|
|
static inline void _nano_timeout_add(struct ccs *ccs,
|
|
struct _nano_queue *wait_q,
|
|
int32_t timeout)
|
|
{
|
|
sys_dlist_t *timeout_q = &_nanokernel.timeout_q;
|
|
struct _nano_timeout *t = &ccs->nano_timeout;
|
|
|
|
t->delta_ticks_from_prev = timeout;
|
|
t->wait_q = wait_q;
|
|
sys_dlist_insert_at(timeout_q, (void *)t,
|
|
_nano_timeout_insert_point_test,
|
|
&t->delta_ticks_from_prev);
|
|
}
|
|
|
|
/* find the closest deadline in the timeout queue */
|
|
static inline uint32_t _nano_get_earliest_timeouts_deadline(void)
|
|
{
|
|
sys_dlist_t *q = &_nanokernel.timeout_q;
|
|
struct _nano_timeout *t = (struct _nano_timeout *)sys_dlist_peek_head(q);
|
|
|
|
return t ? t->delta_ticks_from_prev : TICKS_UNLIMITED;
|
|
}
|
|
|
|
#endif /* _kernel_nanokernel_include_timeout_q__h_ */
|