/* * Copyright (c) 2010-2016 Wind River Systems, Inc. * * SPDX-License-Identifier: Apache-2.0 */ /** * @brief fixed-size stack object */ #include #include #include #include #include #include #include #include #include #include #include extern struct k_stack _k_stack_list_start[]; extern struct k_stack _k_stack_list_end[]; #ifdef CONFIG_OBJECT_TRACING struct k_stack *_trace_list_k_stack; /* * Complete initialization of statically defined stacks. */ static int init_stack_module(struct device *dev) { ARG_UNUSED(dev); struct k_stack *stack; for (stack = _k_stack_list_start; stack < _k_stack_list_end; stack++) { SYS_TRACING_OBJ_INIT(k_stack, stack); } return 0; } SYS_INIT(init_stack_module, PRE_KERNEL_1, CONFIG_KERNEL_INIT_PRIORITY_OBJECTS); #endif /* CONFIG_OBJECT_TRACING */ void k_stack_init(struct k_stack *stack, u32_t *buffer, u32_t num_entries) { z_waitq_init(&stack->wait_q); stack->lock = (struct k_spinlock) {}; stack->next = stack->base = buffer; stack->top = stack->base + num_entries; SYS_TRACING_OBJ_INIT(k_stack, stack); z_object_init(stack); } s32_t z_impl_k_stack_alloc_init(struct k_stack *stack, u32_t num_entries) { void *buffer; s32_t ret; buffer = z_thread_malloc(num_entries * sizeof(u32_t)); if (buffer != NULL) { k_stack_init(stack, buffer, num_entries); stack->flags = K_STACK_FLAG_ALLOC; ret = (s32_t)0; } else { ret = -ENOMEM; } return ret; } #ifdef CONFIG_USERSPACE Z_SYSCALL_HANDLER(k_stack_alloc_init, stack, num_entries) { Z_OOPS(Z_SYSCALL_OBJ_NEVER_INIT(stack, K_OBJ_STACK)); Z_OOPS(Z_SYSCALL_VERIFY(num_entries > 0)); return z_impl_k_stack_alloc_init((struct k_stack *)stack, num_entries); } #endif void k_stack_cleanup(struct k_stack *stack) { __ASSERT_NO_MSG(z_waitq_head(&stack->wait_q) == NULL); if ((stack->flags & K_STACK_FLAG_ALLOC) != (u8_t)0) { k_free(stack->base); stack->base = NULL; stack->flags &= ~K_STACK_FLAG_ALLOC; } } void z_impl_k_stack_push(struct k_stack *stack, u32_t data) { struct k_thread *first_pending_thread; k_spinlock_key_t key; __ASSERT(stack->next != stack->top, "stack is full"); key = k_spin_lock(&stack->lock); first_pending_thread = z_unpend_first_thread(&stack->wait_q); if (first_pending_thread != NULL) { z_ready_thread(first_pending_thread); z_set_thread_return_value_with_data(first_pending_thread, 0, (void *)data); z_reschedule(&stack->lock, key); return; } else { *(stack->next) = data; stack->next++; k_spin_unlock(&stack->lock, key); } } #ifdef CONFIG_USERSPACE Z_SYSCALL_HANDLER(k_stack_push, stack_p, data) { struct k_stack *stack = (struct k_stack *)stack_p; Z_OOPS(Z_SYSCALL_OBJ(stack, K_OBJ_STACK)); Z_OOPS(Z_SYSCALL_VERIFY_MSG(stack->next != stack->top, "stack is full")); z_impl_k_stack_push(stack, data); return 0; } #endif int z_impl_k_stack_pop(struct k_stack *stack, u32_t *data, s32_t timeout) { k_spinlock_key_t key; int result; key = k_spin_lock(&stack->lock); if (likely(stack->next > stack->base)) { stack->next--; *data = *(stack->next); k_spin_unlock(&stack->lock, key); return 0; } if (timeout == K_NO_WAIT) { k_spin_unlock(&stack->lock, key); return -EBUSY; } result = z_pend_curr(&stack->lock, key, &stack->wait_q, timeout); if (result == -EAGAIN) { return -EAGAIN; } *data = (u32_t)_current->base.swap_data; return 0; } #ifdef CONFIG_USERSPACE Z_SYSCALL_HANDLER(k_stack_pop, stack, data, timeout) { Z_OOPS(Z_SYSCALL_OBJ(stack, K_OBJ_STACK)); Z_OOPS(Z_SYSCALL_MEMORY_WRITE(data, sizeof(u32_t))); return z_impl_k_stack_pop((struct k_stack *)stack, (u32_t *)data, timeout); } #endif