222 lines
5.2 KiB
C
222 lines
5.2 KiB
C
/*
|
|
* Copyright (c) 2010-2016 Wind River Systems, Inc.
|
|
*
|
|
* SPDX-License-Identifier: Apache-2.0
|
|
*/
|
|
|
|
/**
|
|
* @brief fixed-size stack object
|
|
*/
|
|
|
|
#include <zephyr/sys/math_extras.h>
|
|
#include <zephyr/kernel.h>
|
|
#include <zephyr/kernel_structs.h>
|
|
|
|
#include <zephyr/toolchain.h>
|
|
#include <ksched.h>
|
|
#include <wait_q.h>
|
|
#include <zephyr/sys/check.h>
|
|
#include <zephyr/init.h>
|
|
#include <zephyr/internal/syscall_handler.h>
|
|
#include <kernel_internal.h>
|
|
|
|
#ifdef CONFIG_OBJ_CORE_STACK
|
|
static struct k_obj_type obj_type_stack;
|
|
#endif /* CONFIG_OBJ_CORE_STACK */
|
|
|
|
void k_stack_init(struct k_stack *stack, stack_data_t *buffer,
|
|
uint32_t num_entries)
|
|
{
|
|
z_waitq_init(&stack->wait_q);
|
|
stack->lock = (struct k_spinlock) {};
|
|
stack->next = stack->base = buffer;
|
|
stack->top = stack->base + num_entries;
|
|
|
|
SYS_PORT_TRACING_OBJ_INIT(k_stack, stack);
|
|
k_object_init(stack);
|
|
|
|
#ifdef CONFIG_OBJ_CORE_STACK
|
|
k_obj_core_init_and_link(K_OBJ_CORE(stack), &obj_type_stack);
|
|
#endif /* CONFIG_OBJ_CORE_STACK */
|
|
}
|
|
|
|
int32_t z_impl_k_stack_alloc_init(struct k_stack *stack, uint32_t num_entries)
|
|
{
|
|
void *buffer;
|
|
int32_t ret;
|
|
|
|
SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_stack, alloc_init, stack);
|
|
|
|
buffer = z_thread_malloc(num_entries * sizeof(stack_data_t));
|
|
if (buffer != NULL) {
|
|
k_stack_init(stack, buffer, num_entries);
|
|
stack->flags = K_STACK_FLAG_ALLOC;
|
|
ret = (int32_t)0;
|
|
} else {
|
|
ret = -ENOMEM;
|
|
}
|
|
|
|
SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_stack, alloc_init, stack, ret);
|
|
|
|
return ret;
|
|
}
|
|
|
|
#ifdef CONFIG_USERSPACE
|
|
static inline int32_t z_vrfy_k_stack_alloc_init(struct k_stack *stack,
|
|
uint32_t num_entries)
|
|
{
|
|
size_t total_size;
|
|
|
|
K_OOPS(K_SYSCALL_OBJ_NEVER_INIT(stack, K_OBJ_STACK));
|
|
K_OOPS(K_SYSCALL_VERIFY(num_entries > 0));
|
|
K_OOPS(K_SYSCALL_VERIFY(!size_mul_overflow(num_entries, sizeof(stack_data_t),
|
|
&total_size)));
|
|
return z_impl_k_stack_alloc_init(stack, num_entries);
|
|
}
|
|
#include <syscalls/k_stack_alloc_init_mrsh.c>
|
|
#endif /* CONFIG_USERSPACE */
|
|
|
|
int k_stack_cleanup(struct k_stack *stack)
|
|
{
|
|
SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_stack, cleanup, stack);
|
|
|
|
CHECKIF(z_waitq_head(&stack->wait_q) != NULL) {
|
|
SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_stack, cleanup, stack, -EAGAIN);
|
|
|
|
return -EAGAIN;
|
|
}
|
|
|
|
if ((stack->flags & K_STACK_FLAG_ALLOC) != (uint8_t)0) {
|
|
k_free(stack->base);
|
|
stack->base = NULL;
|
|
stack->flags &= ~K_STACK_FLAG_ALLOC;
|
|
}
|
|
|
|
SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_stack, cleanup, stack, 0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int z_impl_k_stack_push(struct k_stack *stack, stack_data_t data)
|
|
{
|
|
struct k_thread *first_pending_thread;
|
|
int ret = 0;
|
|
k_spinlock_key_t key = k_spin_lock(&stack->lock);
|
|
|
|
SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_stack, push, stack);
|
|
|
|
CHECKIF(stack->next == stack->top) {
|
|
ret = -ENOMEM;
|
|
goto out;
|
|
}
|
|
|
|
first_pending_thread = z_unpend_first_thread(&stack->wait_q);
|
|
|
|
if (first_pending_thread != NULL) {
|
|
z_thread_return_value_set_with_data(first_pending_thread,
|
|
0, (void *)data);
|
|
|
|
z_ready_thread(first_pending_thread);
|
|
z_reschedule(&stack->lock, key);
|
|
goto end;
|
|
} else {
|
|
*(stack->next) = data;
|
|
stack->next++;
|
|
goto out;
|
|
}
|
|
|
|
out:
|
|
k_spin_unlock(&stack->lock, key);
|
|
|
|
end:
|
|
SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_stack, push, stack, ret);
|
|
|
|
return ret;
|
|
}
|
|
|
|
#ifdef CONFIG_USERSPACE
|
|
static inline int z_vrfy_k_stack_push(struct k_stack *stack, stack_data_t data)
|
|
{
|
|
K_OOPS(K_SYSCALL_OBJ(stack, K_OBJ_STACK));
|
|
|
|
return z_impl_k_stack_push(stack, data);
|
|
}
|
|
#include <syscalls/k_stack_push_mrsh.c>
|
|
#endif /* CONFIG_USERSPACE */
|
|
|
|
int z_impl_k_stack_pop(struct k_stack *stack, stack_data_t *data,
|
|
k_timeout_t timeout)
|
|
{
|
|
k_spinlock_key_t key;
|
|
int result;
|
|
|
|
key = k_spin_lock(&stack->lock);
|
|
|
|
SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_stack, pop, stack, timeout);
|
|
|
|
if (likely(stack->next > stack->base)) {
|
|
stack->next--;
|
|
*data = *(stack->next);
|
|
k_spin_unlock(&stack->lock, key);
|
|
|
|
SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_stack, pop, stack, timeout, 0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
SYS_PORT_TRACING_OBJ_FUNC_BLOCKING(k_stack, pop, stack, timeout);
|
|
|
|
if (K_TIMEOUT_EQ(timeout, K_NO_WAIT)) {
|
|
k_spin_unlock(&stack->lock, key);
|
|
|
|
SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_stack, pop, stack, timeout, -EBUSY);
|
|
|
|
return -EBUSY;
|
|
}
|
|
|
|
result = z_pend_curr(&stack->lock, key, &stack->wait_q, timeout);
|
|
if (result == -EAGAIN) {
|
|
SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_stack, pop, stack, timeout, -EAGAIN);
|
|
|
|
return -EAGAIN;
|
|
}
|
|
|
|
*data = (stack_data_t)_current->base.swap_data;
|
|
|
|
SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_stack, pop, stack, timeout, 0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
#ifdef CONFIG_USERSPACE
|
|
static inline int z_vrfy_k_stack_pop(struct k_stack *stack,
|
|
stack_data_t *data, k_timeout_t timeout)
|
|
{
|
|
K_OOPS(K_SYSCALL_OBJ(stack, K_OBJ_STACK));
|
|
K_OOPS(K_SYSCALL_MEMORY_WRITE(data, sizeof(stack_data_t)));
|
|
return z_impl_k_stack_pop(stack, data, timeout);
|
|
}
|
|
#include <syscalls/k_stack_pop_mrsh.c>
|
|
#endif /* CONFIG_USERSPACE */
|
|
|
|
#ifdef CONFIG_OBJ_CORE_STACK
|
|
static int init_stack_obj_core_list(void)
|
|
{
|
|
/* Initialize stack object type */
|
|
|
|
z_obj_type_init(&obj_type_stack, K_OBJ_TYPE_STACK_ID,
|
|
offsetof(struct k_stack, obj_core));
|
|
|
|
/* Initialize and link statically defined stacks */
|
|
|
|
STRUCT_SECTION_FOREACH(k_stack, stack) {
|
|
k_obj_core_init_and_link(K_OBJ_CORE(stack), &obj_type_stack);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
SYS_INIT(init_stack_obj_core_list, PRE_KERNEL_1,
|
|
CONFIG_KERNEL_INIT_PRIORITY_OBJECTS);
|
|
#endif /* CONFIG_OBJ_CORE_STACK */
|