/* * Copyright (c) 2010-2016 Wind River Systems, Inc. * * SPDX-License-Identifier: Apache-2.0 */ /** * @brief fixed-size stack object */ #include #include #include #include #include #include #include #include #include #include extern struct k_stack _k_stack_list_start[]; extern struct k_stack _k_stack_list_end[]; #ifdef CONFIG_OBJECT_TRACING struct k_stack *_trace_list_k_stack; /* * Complete initialization of statically defined stacks. */ static int init_stack_module(struct device *dev) { ARG_UNUSED(dev); struct k_stack *stack; for (stack = _k_stack_list_start; stack < _k_stack_list_end; stack++) { SYS_TRACING_OBJ_INIT(k_stack, stack); } return 0; } SYS_INIT(init_stack_module, PRE_KERNEL_1, CONFIG_KERNEL_INIT_PRIORITY_OBJECTS); #endif /* CONFIG_OBJECT_TRACING */ void _impl_k_stack_init(struct k_stack *stack, u32_t *buffer, int num_entries) { sys_dlist_init(&stack->wait_q); stack->next = stack->base = buffer; stack->top = stack->base + num_entries; SYS_TRACING_OBJ_INIT(k_stack, stack); _k_object_init(stack); } #ifdef CONFIG_USERSPACE u32_t _handler_k_stack_init(u32_t stack, u32_t buffer, u32_t num_entries_p, u32_t arg4, u32_t arg5, u32_t arg6, void *ssf) { int num_entries = (int)num_entries_p; /* FIXME why is 'num_entries' signed?? */ _SYSCALL_VERIFY(num_entries > 0, ssf); _SYSCALL_OBJ_INIT(stack, K_OBJ_STACK, ssf); _SYSCALL_MEMORY_ARRAY_WRITE(buffer, num_entries, sizeof(u32_t), ssf); _impl_k_stack_init((struct k_stack *)stack, (u32_t *)buffer, num_entries); return 0; } #endif void _impl_k_stack_push(struct k_stack *stack, u32_t data) { struct k_thread *first_pending_thread; unsigned int key; __ASSERT(stack->next != stack->top, "stack is full"); key = irq_lock(); first_pending_thread = _unpend_first_thread(&stack->wait_q); if (first_pending_thread) { _abort_thread_timeout(first_pending_thread); _ready_thread(first_pending_thread); _set_thread_return_value_with_data(first_pending_thread, 0, (void *)data); if (!_is_in_isr() && _must_switch_threads()) { (void)_Swap(key); return; } } else { *(stack->next) = data; stack->next++; } irq_unlock(key); } #ifdef CONFIG_USERSPACE u32_t _handler_k_stack_push(u32_t stack_p, u32_t data, u32_t arg3, u32_t arg4, u32_t arg5, u32_t arg6, void *ssf) { struct k_stack *stack = (struct k_stack *)stack_p; _SYSCALL_ARG2; _SYSCALL_OBJ(stack, K_OBJ_STACK, ssf); _SYSCALL_VERIFY_MSG(stack->next != stack->top, ssf, "stack is full"); _impl_k_stack_push(stack, data); return 0; } #endif int _impl_k_stack_pop(struct k_stack *stack, u32_t *data, s32_t timeout) { unsigned int key; int result; key = irq_lock(); if (likely(stack->next > stack->base)) { stack->next--; *data = *(stack->next); irq_unlock(key); return 0; } if (timeout == K_NO_WAIT) { irq_unlock(key); return -EBUSY; } _pend_current_thread(&stack->wait_q, timeout); result = _Swap(key); if (result == 0) { *data = (u32_t)_current->base.swap_data; } return result; } #ifdef CONFIG_USERSPACE u32_t _handler_k_stack_pop(u32_t stack, u32_t data, u32_t timeout, u32_t arg4, u32_t arg5, u32_t arg6, void *ssf) { _SYSCALL_ARG3; _SYSCALL_OBJ(stack, K_OBJ_STACK, ssf); _SYSCALL_MEMORY_WRITE(data, sizeof(u32_t), ssf); return _impl_k_stack_pop((struct k_stack *)stack, (u32_t *)data, timeout); } #endif