513 lines
13 KiB
C
513 lines
13 KiB
C
/*
|
|
* Copyright (c) 2010-2014 Wind River Systems, Inc.
|
|
*
|
|
* SPDX-License-Identifier: Apache-2.0
|
|
*/
|
|
|
|
/**
|
|
* @file
|
|
* @brief Kernel initialization module
|
|
*
|
|
* This module contains routines that are used to initialize the kernel.
|
|
*/
|
|
|
|
#include <zephyr.h>
|
|
#include <offsets_short.h>
|
|
#include <kernel.h>
|
|
#include <sys/printk.h>
|
|
#include <debug/stack.h>
|
|
#include <random/rand32.h>
|
|
#include <linker/sections.h>
|
|
#include <toolchain.h>
|
|
#include <kernel_structs.h>
|
|
#include <device.h>
|
|
#include <init.h>
|
|
#include <linker/linker-defs.h>
|
|
#include <ksched.h>
|
|
#include <version.h>
|
|
#include <string.h>
|
|
#include <sys/dlist.h>
|
|
#include <kernel_internal.h>
|
|
#include <kswap.h>
|
|
#include <drivers/entropy.h>
|
|
#include <logging/log_ctrl.h>
|
|
#include <tracing/tracing.h>
|
|
#include <stdbool.h>
|
|
#include <debug/gcov.h>
|
|
|
|
#define IDLE_THREAD_NAME "idle"
|
|
#define LOG_LEVEL CONFIG_KERNEL_LOG_LEVEL
|
|
#include <logging/log.h>
|
|
LOG_MODULE_REGISTER(os);
|
|
|
|
/* boot banner items */
|
|
#if defined(CONFIG_MULTITHREADING) && defined(CONFIG_BOOT_DELAY) \
|
|
&& CONFIG_BOOT_DELAY > 0
|
|
#define BOOT_DELAY_BANNER " (delayed boot " \
|
|
STRINGIFY(CONFIG_BOOT_DELAY) "ms)"
|
|
#else
|
|
#define BOOT_DELAY_BANNER ""
|
|
#endif
|
|
|
|
/* boot time measurement items */
|
|
|
|
#ifdef CONFIG_BOOT_TIME_MEASUREMENT
|
|
u32_t __noinit z_timestamp_main; /* timestamp when main task starts */
|
|
u32_t __noinit z_timestamp_idle; /* timestamp when CPU goes idle */
|
|
#endif
|
|
|
|
/* init/main and idle threads */
|
|
K_THREAD_STACK_DEFINE(z_main_stack, CONFIG_MAIN_STACK_SIZE);
|
|
struct k_thread z_main_thread;
|
|
|
|
#ifdef CONFIG_MULTITHREADING
|
|
struct k_thread z_idle_threads[CONFIG_MP_NUM_CPUS];
|
|
static K_THREAD_STACK_ARRAY_DEFINE(z_idle_stacks, CONFIG_MP_NUM_CPUS,
|
|
CONFIG_IDLE_STACK_SIZE);
|
|
#endif /* CONFIG_MULTITHREADING */
|
|
|
|
/*
|
|
* storage space for the interrupt stack
|
|
*
|
|
* Note: This area is used as the system stack during kernel initialization,
|
|
* since the kernel hasn't yet set up its own stack areas. The dual purposing
|
|
* of this area is safe since interrupts are disabled until the kernel context
|
|
* switches to the init thread.
|
|
*/
|
|
K_THREAD_STACK_ARRAY_DEFINE(z_interrupt_stacks, CONFIG_MP_NUM_CPUS,
|
|
CONFIG_ISR_STACK_SIZE);
|
|
|
|
#ifdef CONFIG_SYS_CLOCK_EXISTS
|
|
#define initialize_timeouts() do { \
|
|
sys_dlist_init(&_timeout_q); \
|
|
} while (false)
|
|
#else
|
|
#define initialize_timeouts() do { } while ((0))
|
|
#endif
|
|
|
|
extern void idle(void *unused1, void *unused2, void *unused3);
|
|
|
|
|
|
/* LCOV_EXCL_START
|
|
*
|
|
* This code is called so early in the boot process that code coverage
|
|
* doesn't work properly. In addition, not all arches call this code,
|
|
* some like x86 do this with optimized assembly
|
|
*/
|
|
|
|
/**
|
|
*
|
|
* @brief Clear BSS
|
|
*
|
|
* This routine clears the BSS region, so all bytes are 0.
|
|
*
|
|
* @return N/A
|
|
*/
|
|
void z_bss_zero(void)
|
|
{
|
|
(void)memset(__bss_start, 0, __bss_end - __bss_start);
|
|
#ifdef DT_CCM_BASE_ADDRESS
|
|
(void)memset(&__ccm_bss_start, 0,
|
|
((u32_t) &__ccm_bss_end - (u32_t) &__ccm_bss_start));
|
|
#endif
|
|
#ifdef DT_DTCM_BASE_ADDRESS
|
|
(void)memset(&__dtcm_bss_start, 0,
|
|
((u32_t) &__dtcm_bss_end - (u32_t) &__dtcm_bss_start));
|
|
#endif
|
|
#ifdef CONFIG_CODE_DATA_RELOCATION
|
|
extern void bss_zeroing_relocation(void);
|
|
|
|
bss_zeroing_relocation();
|
|
#endif /* CONFIG_CODE_DATA_RELOCATION */
|
|
#ifdef CONFIG_COVERAGE_GCOV
|
|
(void)memset(&__gcov_bss_start, 0,
|
|
((u32_t) &__gcov_bss_end - (u32_t) &__gcov_bss_start));
|
|
#endif
|
|
}
|
|
|
|
#ifdef CONFIG_STACK_CANARIES
|
|
extern volatile uintptr_t __stack_chk_guard;
|
|
#endif /* CONFIG_STACK_CANARIES */
|
|
|
|
|
|
#ifdef CONFIG_XIP
|
|
/**
|
|
*
|
|
* @brief Copy the data section from ROM to RAM
|
|
*
|
|
* This routine copies the data section from ROM to RAM.
|
|
*
|
|
* @return N/A
|
|
*/
|
|
void z_data_copy(void)
|
|
{
|
|
(void)memcpy(&__data_ram_start, &__data_rom_start,
|
|
__data_ram_end - __data_ram_start);
|
|
#ifdef CONFIG_ARCH_HAS_RAMFUNC_SUPPORT
|
|
(void)memcpy(&_ramfunc_ram_start, &_ramfunc_rom_start,
|
|
(uintptr_t) &_ramfunc_ram_size);
|
|
#endif /* CONFIG_ARCH_HAS_RAMFUNC_SUPPORT */
|
|
#ifdef DT_CCM_BASE_ADDRESS
|
|
(void)memcpy(&__ccm_data_start, &__ccm_data_rom_start,
|
|
__ccm_data_end - __ccm_data_start);
|
|
#endif
|
|
#ifdef DT_DTCM_BASE_ADDRESS
|
|
(void)memcpy(&__dtcm_data_start, &__dtcm_data_rom_start,
|
|
__dtcm_data_end - __dtcm_data_start);
|
|
#endif
|
|
#ifdef CONFIG_CODE_DATA_RELOCATION
|
|
extern void data_copy_xip_relocation(void);
|
|
|
|
data_copy_xip_relocation();
|
|
#endif /* CONFIG_CODE_DATA_RELOCATION */
|
|
#ifdef CONFIG_USERSPACE
|
|
#ifdef CONFIG_STACK_CANARIES
|
|
/* stack canary checking is active for all C functions.
|
|
* __stack_chk_guard is some uninitialized value living in the
|
|
* app shared memory sections. Preserve it, and don't make any
|
|
* function calls to perform the memory copy. The true canary
|
|
* value gets set later in z_cstart().
|
|
*/
|
|
uintptr_t guard_copy = __stack_chk_guard;
|
|
u8_t *src = (u8_t *)&_app_smem_rom_start;
|
|
u8_t *dst = (u8_t *)&_app_smem_start;
|
|
u32_t count = _app_smem_end - _app_smem_start;
|
|
|
|
guard_copy = __stack_chk_guard;
|
|
while (count > 0) {
|
|
*(dst++) = *(src++);
|
|
count--;
|
|
}
|
|
__stack_chk_guard = guard_copy;
|
|
#else
|
|
(void)memcpy(&_app_smem_start, &_app_smem_rom_start,
|
|
_app_smem_end - _app_smem_start);
|
|
#endif /* CONFIG_STACK_CANARIES */
|
|
#endif /* CONFIG_USERSPACE */
|
|
}
|
|
#endif /* CONFIG_XIP */
|
|
|
|
/* LCOV_EXCL_STOP */
|
|
|
|
bool z_sys_post_kernel;
|
|
|
|
/**
|
|
*
|
|
* @brief Mainline for kernel's background thread
|
|
*
|
|
* This routine completes kernel initialization by invoking the remaining
|
|
* init functions, then invokes application's main() routine.
|
|
*
|
|
* @return N/A
|
|
*/
|
|
static void bg_thread_main(void *unused1, void *unused2, void *unused3)
|
|
{
|
|
ARG_UNUSED(unused1);
|
|
ARG_UNUSED(unused2);
|
|
ARG_UNUSED(unused3);
|
|
|
|
#if defined(CONFIG_BOOT_DELAY) && CONFIG_BOOT_DELAY > 0
|
|
static const unsigned int boot_delay = CONFIG_BOOT_DELAY;
|
|
#else
|
|
static const unsigned int boot_delay;
|
|
#endif
|
|
|
|
z_sys_post_kernel = true;
|
|
|
|
z_sys_device_do_config_level(_SYS_INIT_LEVEL_POST_KERNEL);
|
|
#if CONFIG_STACK_POINTER_RANDOM
|
|
z_stack_adjust_initialized = 1;
|
|
#endif
|
|
if (boot_delay > 0 && IS_ENABLED(CONFIG_MULTITHREADING)) {
|
|
printk("***** delaying boot " STRINGIFY(CONFIG_BOOT_DELAY)
|
|
"ms (per build configuration) *****\n");
|
|
k_busy_wait(CONFIG_BOOT_DELAY * USEC_PER_MSEC);
|
|
}
|
|
|
|
#if defined(CONFIG_BOOT_BANNER)
|
|
#ifdef BUILD_VERSION
|
|
printk("*** Booting Zephyr OS build %s %s ***\n",
|
|
STRINGIFY(BUILD_VERSION), BOOT_DELAY_BANNER);
|
|
#else
|
|
printk("*** Booting Zephyr OS version %s %s ***\n",
|
|
KERNEL_VERSION_STRING, BOOT_DELAY_BANNER);
|
|
#endif
|
|
#endif
|
|
|
|
/* Final init level before app starts */
|
|
z_sys_device_do_config_level(_SYS_INIT_LEVEL_APPLICATION);
|
|
|
|
#ifdef CONFIG_CPLUSPLUS
|
|
/* Process the .ctors and .init_array sections */
|
|
extern void __do_global_ctors_aux(void);
|
|
extern void __do_init_array_aux(void);
|
|
__do_global_ctors_aux();
|
|
__do_init_array_aux();
|
|
#endif
|
|
|
|
z_init_static_threads();
|
|
|
|
#ifdef CONFIG_SMP
|
|
z_smp_init();
|
|
z_sys_device_do_config_level(_SYS_INIT_LEVEL_SMP);
|
|
#endif
|
|
|
|
#ifdef CONFIG_BOOT_TIME_MEASUREMENT
|
|
z_timestamp_main = k_cycle_get_32();
|
|
#endif
|
|
|
|
extern void main(void);
|
|
|
|
main();
|
|
|
|
/* Mark nonessenrial since main() has no more work to do */
|
|
z_main_thread.base.user_options &= ~K_ESSENTIAL;
|
|
|
|
#ifdef CONFIG_COVERAGE_DUMP
|
|
/* Dump coverage data once the main() has exited. */
|
|
gcov_coverage_dump();
|
|
#endif
|
|
} /* LCOV_EXCL_LINE ... because we just dumped final coverage data */
|
|
|
|
/* LCOV_EXCL_START */
|
|
|
|
void __weak main(void)
|
|
{
|
|
/* NOP default main() if the application does not provide one. */
|
|
arch_nop();
|
|
}
|
|
|
|
/* LCOV_EXCL_STOP */
|
|
|
|
#if defined(CONFIG_MULTITHREADING)
|
|
static void init_idle_thread(struct k_thread *thread, k_thread_stack_t *stack)
|
|
{
|
|
z_setup_new_thread(thread, stack,
|
|
CONFIG_IDLE_STACK_SIZE, idle, NULL, NULL, NULL,
|
|
K_LOWEST_THREAD_PRIO, K_ESSENTIAL, IDLE_THREAD_NAME);
|
|
z_mark_thread_as_started(thread);
|
|
|
|
#ifdef CONFIG_SMP
|
|
thread->base.is_idle = 1U;
|
|
#endif
|
|
}
|
|
#endif /* CONFIG_MULTITHREADING */
|
|
|
|
/**
|
|
*
|
|
* @brief Initializes kernel data structures
|
|
*
|
|
* This routine initializes various kernel data structures, including
|
|
* the init and idle threads and any architecture-specific initialization.
|
|
*
|
|
* Note that all fields of "_kernel" are set to zero on entry, which may
|
|
* be all the initialization many of them require.
|
|
*
|
|
* @return N/A
|
|
*/
|
|
#ifdef CONFIG_MULTITHREADING
|
|
static void prepare_multithreading(struct k_thread *dummy_thread)
|
|
{
|
|
#ifdef CONFIG_ARCH_HAS_CUSTOM_SWAP_TO_MAIN
|
|
ARG_UNUSED(dummy_thread);
|
|
#else
|
|
|
|
/*
|
|
* Initialize the current execution thread to permit a level of
|
|
* debugging output if an exception should happen during kernel
|
|
* initialization. However, don't waste effort initializing the
|
|
* fields of the dummy thread beyond those needed to identify it as a
|
|
* dummy thread.
|
|
*/
|
|
dummy_thread->base.user_options = K_ESSENTIAL;
|
|
dummy_thread->base.thread_state = _THREAD_DUMMY;
|
|
#ifdef CONFIG_THREAD_STACK_INFO
|
|
dummy_thread->stack_info.start = 0U;
|
|
dummy_thread->stack_info.size = 0U;
|
|
#endif
|
|
#ifdef CONFIG_USERSPACE
|
|
dummy_thread->mem_domain_info.mem_domain = 0;
|
|
#endif
|
|
#endif /* CONFIG_ARCH_HAS_CUSTOM_SWAP_TO_MAIN */
|
|
|
|
/* _kernel.ready_q is all zeroes */
|
|
z_sched_init();
|
|
|
|
#ifndef CONFIG_SMP
|
|
/*
|
|
* prime the cache with the main thread since:
|
|
*
|
|
* - the cache can never be NULL
|
|
* - the main thread will be the one to run first
|
|
* - no other thread is initialized yet and thus their priority fields
|
|
* contain garbage, which would prevent the cache loading algorithm
|
|
* to work as intended
|
|
*/
|
|
_kernel.ready_q.cache = &z_main_thread;
|
|
#endif
|
|
|
|
z_setup_new_thread(&z_main_thread, z_main_stack,
|
|
CONFIG_MAIN_STACK_SIZE, bg_thread_main,
|
|
NULL, NULL, NULL,
|
|
CONFIG_MAIN_THREAD_PRIORITY, K_ESSENTIAL, "main");
|
|
z_mark_thread_as_started(&z_main_thread);
|
|
z_ready_thread(&z_main_thread);
|
|
|
|
for (int i = 0; i < CONFIG_MP_NUM_CPUS; i++) {
|
|
init_idle_thread(&z_idle_threads[i], z_idle_stacks[i]);
|
|
_kernel.cpus[i].idle_thread = &z_idle_threads[i];
|
|
_kernel.cpus[i].id = i;
|
|
_kernel.cpus[i].irq_stack =
|
|
(Z_THREAD_STACK_BUFFER(z_interrupt_stacks[i]) +
|
|
K_THREAD_STACK_SIZEOF(z_interrupt_stacks[i]));
|
|
}
|
|
|
|
initialize_timeouts();
|
|
}
|
|
|
|
static FUNC_NORETURN void switch_to_main_thread(void)
|
|
{
|
|
#ifdef CONFIG_ARCH_HAS_CUSTOM_SWAP_TO_MAIN
|
|
arch_switch_to_main_thread(&z_main_thread, z_main_stack,
|
|
K_THREAD_STACK_SIZEOF(z_main_stack),
|
|
bg_thread_main);
|
|
#else
|
|
/*
|
|
* Context switch to main task (entry function is _main()): the
|
|
* current fake thread is not on a wait queue or ready queue, so it
|
|
* will never be rescheduled in.
|
|
*/
|
|
z_swap_unlocked();
|
|
#endif
|
|
CODE_UNREACHABLE; /* LCOV_EXCL_LINE */
|
|
}
|
|
#endif /* CONFIG_MULTITHREADING */
|
|
|
|
#if defined(CONFIG_ENTROPY_HAS_DRIVER) || defined(CONFIG_TEST_RANDOM_GENERATOR)
|
|
void z_early_boot_rand_get(u8_t *buf, size_t length)
|
|
{
|
|
int n = sizeof(u32_t);
|
|
#ifdef CONFIG_ENTROPY_HAS_DRIVER
|
|
struct device *entropy = device_get_binding(DT_CHOSEN_ZEPHYR_ENTROPY_LABEL);
|
|
int rc;
|
|
|
|
if (entropy == NULL) {
|
|
goto sys_rand_fallback;
|
|
}
|
|
|
|
/* Try to see if driver provides an ISR-specific API */
|
|
rc = entropy_get_entropy_isr(entropy, buf, length, ENTROPY_BUSYWAIT);
|
|
if (rc == -ENOTSUP) {
|
|
/* Driver does not provide an ISR-specific API, assume it can
|
|
* be called from ISR context
|
|
*/
|
|
rc = entropy_get_entropy(entropy, buf, length);
|
|
}
|
|
|
|
if (rc >= 0) {
|
|
return;
|
|
}
|
|
|
|
/* Fall through to fallback */
|
|
|
|
sys_rand_fallback:
|
|
#endif
|
|
|
|
/* FIXME: this assumes sys_rand32_get() won't use any synchronization
|
|
* primitive, like semaphores or mutexes. It's too early in the boot
|
|
* process to use any of them. Ideally, only the path where entropy
|
|
* devices are available should be built, this is only a fallback for
|
|
* those devices without a HWRNG entropy driver.
|
|
*/
|
|
|
|
while (length > 0) {
|
|
u32_t rndbits;
|
|
u8_t *p_rndbits = (u8_t *)&rndbits;
|
|
|
|
rndbits = sys_rand32_get();
|
|
|
|
if (length < sizeof(u32_t)) {
|
|
n = length;
|
|
}
|
|
|
|
for (int i = 0; i < n; i++) {
|
|
*buf = *p_rndbits;
|
|
buf++;
|
|
p_rndbits++;
|
|
}
|
|
|
|
length -= n;
|
|
}
|
|
}
|
|
/* defined(CONFIG_ENTROPY_HAS_DRIVER) || defined(CONFIG_TEST_RANDOM_GENERATOR) */
|
|
#endif
|
|
|
|
/**
|
|
*
|
|
* @brief Initialize kernel
|
|
*
|
|
* This routine is invoked when the system is ready to run C code. The
|
|
* processor must be running in 32-bit mode, and the BSS must have been
|
|
* cleared/zeroed.
|
|
*
|
|
* @return Does not return
|
|
*/
|
|
FUNC_NORETURN void z_cstart(void)
|
|
{
|
|
#ifdef CONFIG_STACK_CANARIES
|
|
uintptr_t stack_guard;
|
|
#endif /* CONFIG_STACK_CANARIES */
|
|
|
|
/* gcov hook needed to get the coverage report.*/
|
|
gcov_static_init();
|
|
|
|
LOG_CORE_INIT();
|
|
|
|
/* perform any architecture-specific initialization */
|
|
arch_kernel_init();
|
|
|
|
#ifdef CONFIG_MULTITHREADING
|
|
struct k_thread dummy_thread = {
|
|
.base.thread_state = _THREAD_DUMMY,
|
|
# ifdef CONFIG_SCHED_CPU_MASK
|
|
.base.cpu_mask = -1,
|
|
# endif
|
|
};
|
|
|
|
_current_cpu->current = &dummy_thread;
|
|
#endif
|
|
|
|
/* perform basic hardware initialization */
|
|
z_sys_device_do_config_level(_SYS_INIT_LEVEL_PRE_KERNEL_1);
|
|
z_sys_device_do_config_level(_SYS_INIT_LEVEL_PRE_KERNEL_2);
|
|
|
|
#ifdef CONFIG_STACK_CANARIES
|
|
z_early_boot_rand_get((u8_t *)&stack_guard, sizeof(stack_guard));
|
|
__stack_chk_guard = stack_guard;
|
|
__stack_chk_guard <<= 8;
|
|
#endif /* CONFIG_STACK_CANARIES */
|
|
|
|
#ifdef CONFIG_MULTITHREADING
|
|
prepare_multithreading(&dummy_thread);
|
|
switch_to_main_thread();
|
|
#else
|
|
bg_thread_main(NULL, NULL, NULL);
|
|
|
|
/* LCOV_EXCL_START
|
|
* We've already dumped coverage data at this point.
|
|
*/
|
|
irq_lock();
|
|
while (true) {
|
|
}
|
|
/* LCOV_EXCL_STOP */
|
|
#endif
|
|
|
|
/*
|
|
* Compiler can't tell that the above routines won't return and issues
|
|
* a warning unless we explicitly tell it that control never gets this
|
|
* far.
|
|
*/
|
|
|
|
CODE_UNREACHABLE; /* LCOV_EXCL_LINE */
|
|
}
|