2020-04-04 01:01:03 +08:00
|
|
|
/*
|
|
|
|
* Copyright (c) 2020 Intel Corporation
|
|
|
|
*
|
|
|
|
* SPDX-License-Identifier: Apache-2.0
|
|
|
|
*/
|
|
|
|
|
2022-05-06 17:04:23 +08:00
|
|
|
#include <zephyr/kernel.h>
|
2020-04-04 01:01:03 +08:00
|
|
|
#include <ksched.h>
|
2022-05-06 17:04:23 +08:00
|
|
|
#include <zephyr/wait_q.h>
|
|
|
|
#include <zephyr/init.h>
|
|
|
|
#include <zephyr/linker/linker-defs.h>
|
2023-05-15 21:50:28 +08:00
|
|
|
#include <zephyr/sys/iterable_sections.h>
|
2020-04-04 01:01:03 +08:00
|
|
|
|
|
|
|
void k_heap_init(struct k_heap *h, void *mem, size_t bytes)
|
|
|
|
{
|
|
|
|
z_waitq_init(&h->wait_q);
|
|
|
|
sys_heap_init(&h->heap, mem, bytes);
|
2021-03-26 20:42:25 +08:00
|
|
|
|
|
|
|
SYS_PORT_TRACING_OBJ_INIT(k_heap, h);
|
2020-04-04 01:01:03 +08:00
|
|
|
}
|
|
|
|
|
init: remove the need for a dummy device pointer in SYS_INIT functions
The init infrastructure, found in `init.h`, is currently used by:
- `SYS_INIT`: to call functions before `main`
- `DEVICE_*`: to initialize devices
They are all sorted according to an initialization level + a priority.
`SYS_INIT` calls are really orthogonal to devices, however, the required
function signature requires a `const struct device *dev` as a first
argument. The only reason for that is because the same init machinery is
used by devices, so we have something like:
```c
struct init_entry {
int (*init)(const struct device *dev);
/* only set by DEVICE_*, otherwise NULL */
const struct device *dev;
}
```
As a result, we end up with such weird/ugly pattern:
```c
static int my_init(const struct device *dev)
{
/* always NULL! add ARG_UNUSED to avoid compiler warning */
ARG_UNUSED(dev);
...
}
```
This is really a result of poor internals isolation. This patch proposes
a to make init entries more flexible so that they can accept sytem
initialization calls like this:
```c
static int my_init(void)
{
...
}
```
This is achieved using a union:
```c
union init_function {
/* for SYS_INIT, used when init_entry.dev == NULL */
int (*sys)(void);
/* for DEVICE*, used when init_entry.dev != NULL */
int (*dev)(const struct device *dev);
};
struct init_entry {
/* stores init function (either for SYS_INIT or DEVICE*)
union init_function init_fn;
/* stores device pointer for DEVICE*, NULL for SYS_INIT. Allows
* to know which union entry to call.
*/
const struct device *dev;
}
```
This solution **does not increase ROM usage**, and allows to offer clean
public APIs for both SYS_INIT and DEVICE*. Note that however, init
machinery keeps a coupling with devices.
**NOTE**: This is a breaking change! All `SYS_INIT` functions will need
to be converted to the new signature. See the script offered in the
following commit.
Signed-off-by: Gerard Marull-Paretas <gerard.marull@nordicsemi.no>
init: convert SYS_INIT functions to the new signature
Conversion scripted using scripts/utils/migrate_sys_init.py.
Signed-off-by: Gerard Marull-Paretas <gerard.marull@nordicsemi.no>
manifest: update projects for SYS_INIT changes
Update modules with updated SYS_INIT calls:
- hal_ti
- lvgl
- sof
- TraceRecorderSource
Signed-off-by: Gerard Marull-Paretas <gerard.marull@nordicsemi.no>
tests: devicetree: devices: adjust test
Adjust test according to the recently introduced SYS_INIT
infrastructure.
Signed-off-by: Gerard Marull-Paretas <gerard.marull@nordicsemi.no>
tests: kernel: threads: adjust SYS_INIT call
Adjust to the new signature: int (*init_fn)(void);
Signed-off-by: Gerard Marull-Paretas <gerard.marull@nordicsemi.no>
2022-10-19 15:33:44 +08:00
|
|
|
static int statics_init(void)
|
2020-04-04 01:01:03 +08:00
|
|
|
{
|
2021-08-05 06:05:54 +08:00
|
|
|
STRUCT_SECTION_FOREACH(k_heap, h) {
|
2021-08-04 04:59:36 +08:00
|
|
|
#if defined(CONFIG_DEMAND_PAGING) && !defined(CONFIG_LINKER_GENERIC_SECTIONS_PRESENT_AT_BOOT)
|
|
|
|
/* Some heaps may not present at boot, so we need to wait for
|
|
|
|
* paging mechanism to be initialized before we can initialize
|
|
|
|
* each heap.
|
|
|
|
*/
|
|
|
|
extern bool z_sys_post_kernel;
|
|
|
|
bool do_clear = z_sys_post_kernel;
|
|
|
|
|
|
|
|
/* During pre-kernel init, z_sys_post_kernel == false,
|
|
|
|
* initialize if within pinned region. Otherwise skip.
|
|
|
|
* In post-kernel init, z_sys_post_kernel == true, skip those in
|
|
|
|
* pinned region as they have already been initialized and
|
|
|
|
* possibly already in use. Otherwise initialize.
|
|
|
|
*/
|
|
|
|
if (lnkr_is_pinned((uint8_t *)h) &&
|
|
|
|
lnkr_is_pinned((uint8_t *)&h->wait_q) &&
|
|
|
|
lnkr_is_region_pinned((uint8_t *)h->heap.init_mem,
|
|
|
|
h->heap.init_bytes)) {
|
|
|
|
do_clear = !do_clear;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (do_clear)
|
|
|
|
#endif /* CONFIG_DEMAND_PAGING && !CONFIG_LINKER_GENERIC_SECTIONS_PRESENT_AT_BOOT */
|
|
|
|
{
|
|
|
|
k_heap_init(h, h->heap.init_mem, h->heap.init_bytes);
|
|
|
|
}
|
2020-04-04 01:01:03 +08:00
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2022-07-02 10:06:55 +08:00
|
|
|
SYS_INIT_NAMED(statics_init_pre, statics_init, PRE_KERNEL_1, CONFIG_KERNEL_INIT_PRIORITY_OBJECTS);
|
2020-04-04 01:01:03 +08:00
|
|
|
|
2021-08-04 04:59:36 +08:00
|
|
|
#if defined(CONFIG_DEMAND_PAGING) && !defined(CONFIG_LINKER_GENERIC_SECTIONS_PRESENT_AT_BOOT)
|
|
|
|
/* Need to wait for paging mechanism to be initialized before
|
|
|
|
* heaps that are not in pinned sections can be initialized.
|
|
|
|
*/
|
2022-07-02 10:06:55 +08:00
|
|
|
SYS_INIT_NAMED(statics_init_post, statics_init, POST_KERNEL, 0);
|
2021-08-04 04:59:36 +08:00
|
|
|
#endif /* CONFIG_DEMAND_PAGING && !CONFIG_LINKER_GENERIC_SECTIONS_PRESENT_AT_BOOT */
|
|
|
|
|
2020-11-13 22:12:31 +08:00
|
|
|
void *k_heap_aligned_alloc(struct k_heap *h, size_t align, size_t bytes,
|
|
|
|
k_timeout_t timeout)
|
2020-04-04 01:01:03 +08:00
|
|
|
{
|
2021-03-13 21:22:38 +08:00
|
|
|
int64_t now, end = sys_clock_timeout_end_calc(timeout);
|
2020-04-04 01:01:03 +08:00
|
|
|
void *ret = NULL;
|
2022-09-29 07:59:45 +08:00
|
|
|
|
|
|
|
end = K_TIMEOUT_EQ(timeout, K_FOREVER) ? INT64_MAX : end;
|
|
|
|
|
2020-04-04 01:01:03 +08:00
|
|
|
k_spinlock_key_t key = k_spin_lock(&h->lock);
|
|
|
|
|
2021-03-26 20:42:25 +08:00
|
|
|
SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_heap, aligned_alloc, h, timeout);
|
|
|
|
|
2020-04-04 01:01:03 +08:00
|
|
|
__ASSERT(!arch_is_in_isr() || K_TIMEOUT_EQ(timeout, K_NO_WAIT), "");
|
|
|
|
|
2021-03-26 20:42:25 +08:00
|
|
|
bool blocked_alloc = false;
|
|
|
|
|
2020-04-04 01:01:03 +08:00
|
|
|
while (ret == NULL) {
|
2020-11-13 22:12:31 +08:00
|
|
|
ret = sys_heap_aligned_alloc(&h->heap, align, bytes);
|
2020-04-04 01:01:03 +08:00
|
|
|
|
2021-03-13 21:21:21 +08:00
|
|
|
now = sys_clock_tick_get();
|
2021-04-19 16:52:34 +08:00
|
|
|
if (!IS_ENABLED(CONFIG_MULTITHREADING) ||
|
|
|
|
(ret != NULL) || ((end - now) <= 0)) {
|
2020-04-04 01:01:03 +08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2021-03-26 20:42:25 +08:00
|
|
|
if (!blocked_alloc) {
|
|
|
|
blocked_alloc = true;
|
|
|
|
|
|
|
|
SYS_PORT_TRACING_OBJ_FUNC_BLOCKING(k_heap, aligned_alloc, h, timeout);
|
|
|
|
} else {
|
|
|
|
/**
|
|
|
|
* @todo Trace attempt to avoid empty trace segments
|
|
|
|
*/
|
|
|
|
}
|
|
|
|
|
2020-04-04 01:01:03 +08:00
|
|
|
(void) z_pend_curr(&h->lock, key, &h->wait_q,
|
|
|
|
K_TICKS(end - now));
|
|
|
|
key = k_spin_lock(&h->lock);
|
|
|
|
}
|
|
|
|
|
2021-03-26 20:42:25 +08:00
|
|
|
SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_heap, aligned_alloc, h, timeout, ret);
|
|
|
|
|
2020-04-04 01:01:03 +08:00
|
|
|
k_spin_unlock(&h->lock, key);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2021-03-26 20:42:25 +08:00
|
|
|
void *k_heap_alloc(struct k_heap *h, size_t bytes, k_timeout_t timeout)
|
|
|
|
{
|
|
|
|
SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_heap, alloc, h, timeout);
|
|
|
|
|
|
|
|
void *ret = k_heap_aligned_alloc(h, sizeof(void *), bytes, timeout);
|
|
|
|
|
|
|
|
SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_heap, alloc, h, timeout, ret);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2020-04-04 01:01:03 +08:00
|
|
|
void k_heap_free(struct k_heap *h, void *mem)
|
|
|
|
{
|
|
|
|
k_spinlock_key_t key = k_spin_lock(&h->lock);
|
|
|
|
|
|
|
|
sys_heap_free(&h->heap, mem);
|
2020-04-04 06:39:25 +08:00
|
|
|
|
2021-03-26 20:42:25 +08:00
|
|
|
SYS_PORT_TRACING_OBJ_FUNC(k_heap, free, h);
|
2021-04-19 16:52:34 +08:00
|
|
|
if (IS_ENABLED(CONFIG_MULTITHREADING) && z_unpend_all(&h->wait_q) != 0) {
|
2020-04-04 06:39:25 +08:00
|
|
|
z_reschedule(&h->lock, key);
|
|
|
|
} else {
|
|
|
|
k_spin_unlock(&h->lock, key);
|
|
|
|
}
|
|
|
|
}
|