/* * Copyright (c) 2016 Wind River Systems, Inc. * * SPDX-License-Identifier: Apache-2.0 */ #ifndef _kernel_structs__h_ #define _kernel_structs__h_ #include #if !defined(_ASMLANGUAGE) #include #include #include #endif /* * Bitmask definitions for the struct k_thread.thread_state field. * * Must be before kerneL_arch_data.h because it might need them to be already * defined. */ /* states: common uses low bits, arch-specific use high bits */ /* Not a real thread */ #define _THREAD_DUMMY (1 << 0) /* Thread is waiting on an object */ #define _THREAD_PENDING (1 << 1) /* Thread has not yet started */ #define _THREAD_PRESTART (1 << 2) /* Thread has terminated */ #define _THREAD_DEAD (1 << 3) /* Thread is suspended */ #define _THREAD_SUSPENDED (1 << 4) /* Thread is actively looking at events to see if they are ready */ #define _THREAD_POLLING (1 << 5) /* end - states */ #ifdef CONFIG_STACK_SENTINEL /* Magic value in lowest bytes of the stack */ #define STACK_SENTINEL 0xF0F0F0F0 #endif /* lowest value of _thread_base.preempt at which a thread is non-preemptible */ #define _NON_PREEMPT_THRESHOLD 0x0080 /* highest value of _thread_base.preempt at which a thread is preemptible */ #define _PREEMPT_THRESHOLD (_NON_PREEMPT_THRESHOLD - 1) #include #if !defined(_ASMLANGUAGE) struct _ready_q { /* always contains next thread to run: cannot be NULL */ struct k_thread *cache; /* bitmap of priorities that contain at least one ready thread */ u32_t prio_bmap[K_NUM_PRIO_BITMAPS]; /* ready queues, one per priority */ sys_dlist_t q[K_NUM_PRIORITIES]; }; typedef struct _ready_q _ready_q_t; struct _kernel { /* nested interrupt count */ u32_t nested; /* interrupt stack pointer base */ char *irq_stack; /* currently scheduled thread */ struct k_thread *current; #ifdef CONFIG_SYS_CLOCK_EXISTS /* queue of timeouts */ sys_dlist_t timeout_q; #endif #ifdef CONFIG_SYS_POWER_MANAGEMENT s32_t idle; /* Number of ticks for kernel idling */ #endif /* * ready queue: can be big, keep after small fields, since some * assembly (e.g. ARC) are limited in the encoding of the offset */ struct _ready_q ready_q; #ifdef CONFIG_FP_SHARING /* * A 'current_sse' field does not exist in addition to the 'current_fp' * field since it's not possible to divide the IA-32 non-integer * registers into 2 distinct blocks owned by differing threads. In * other words, given that the 'fxnsave/fxrstor' instructions * save/restore both the X87 FPU and XMM registers, it's not possible * for a thread to only "own" the XMM registers. */ /* thread that owns the FP regs */ struct k_thread *current_fp; #endif #if defined(CONFIG_THREAD_MONITOR) struct k_thread *threads; /* singly linked list of ALL threads */ #endif #if defined(CONFIG_USERSPACE) /* 0 bits for ids currently in use, 1 for free ids */ u8_t free_thread_ids[CONFIG_MAX_THREAD_BYTES]; #endif /* arch-specific part of _kernel */ struct _kernel_arch arch; }; typedef struct _kernel _kernel_t; extern struct _kernel _kernel; #define _current _kernel.current #define _ready_q _kernel.ready_q #define _timeout_q _kernel.timeout_q #define _threads _kernel.threads #include static ALWAYS_INLINE void _set_thread_return_value_with_data(struct k_thread *thread, unsigned int value, void *data) { _set_thread_return_value(thread, value); thread->base.swap_data = data; } extern void _init_thread_base(struct _thread_base *thread_base, int priority, u32_t initial_state, unsigned int options); static ALWAYS_INLINE void _new_thread_init(struct k_thread *thread, char *pStack, size_t stackSize, int prio, unsigned int options) { #if !defined(CONFIG_INIT_STACKS) && !defined(CONFIG_THREAD_STACK_INFO) ARG_UNUSED(pStack); ARG_UNUSED(stackSize); #endif #ifdef CONFIG_INIT_STACKS memset(pStack, 0xaa, stackSize); #endif #ifdef CONFIG_STACK_SENTINEL /* Put the stack sentinel at the lowest 4 bytes of the stack area. * We periodically check that it's still present and kill the thread * if it isn't. */ *((u32_t *)pStack) = STACK_SENTINEL; #endif /* CONFIG_STACK_SENTINEL */ /* Initialize various struct k_thread members */ _init_thread_base(&thread->base, prio, _THREAD_PRESTART, options); /* static threads overwrite it afterwards with real value */ thread->init_data = NULL; thread->fn_abort = NULL; #ifdef CONFIG_THREAD_CUSTOM_DATA /* Initialize custom data field (value is opaque to kernel) */ thread->custom_data = NULL; #endif #if defined(CONFIG_USERSPACE) thread->mem_domain_info.mem_domain = NULL; #endif /* CONFIG_USERSPACE */ #if defined(CONFIG_THREAD_STACK_INFO) thread->stack_info.start = (u32_t)pStack; thread->stack_info.size = (u32_t)stackSize; #endif /* CONFIG_THREAD_STACK_INFO */ } #if defined(CONFIG_THREAD_MONITOR) /* * Add a thread to the kernel's list of active threads. */ static ALWAYS_INLINE void thread_monitor_init(struct k_thread *thread) { unsigned int key; key = irq_lock(); thread->next_thread = _kernel.threads; _kernel.threads = thread; irq_unlock(key); } #else #define thread_monitor_init(thread) \ do {/* do nothing */ \ } while ((0)) #endif /* CONFIG_THREAD_MONITOR */ #endif /* _ASMLANGUAGE */ #endif /* _kernel_structs__h_ */