428 lines
9.8 KiB
C
428 lines
9.8 KiB
C
/*
|
|
* Copyright (c) 2017 Intel Corporation
|
|
*
|
|
* SPDX-License-Identifier: Apache-2.0
|
|
*/
|
|
|
|
#include <kernel.h>
|
|
#include <ksched.h>
|
|
#include <wait_q.h>
|
|
#include <init.h>
|
|
#include <string.h>
|
|
#include <misc/__assert.h>
|
|
|
|
/* Linker-defined symbols bound the static pool structs */
|
|
extern struct k_mem_pool _k_mem_pool_list_start[];
|
|
extern struct k_mem_pool _k_mem_pool_list_end[];
|
|
|
|
s64_t _tick_get(void);
|
|
|
|
static struct k_mem_pool *get_pool(int id)
|
|
{
|
|
return &_k_mem_pool_list_start[id];
|
|
}
|
|
|
|
static int pool_id(struct k_mem_pool *pool)
|
|
{
|
|
return pool - &_k_mem_pool_list_start[0];
|
|
}
|
|
|
|
static void *block_ptr(struct k_mem_pool *p, size_t lsz, int block)
|
|
{
|
|
return p->buf + lsz * block;
|
|
}
|
|
|
|
static int block_num(struct k_mem_pool *p, void *block, int sz)
|
|
{
|
|
return (block - p->buf) / sz;
|
|
}
|
|
|
|
static bool level_empty(struct k_mem_pool *p, int l)
|
|
{
|
|
return sys_dlist_is_empty(&p->levels[l].free_list);
|
|
}
|
|
|
|
/* Places a 32 bit output pointer in word, and an integer bit index
|
|
* within that word as the return value
|
|
*/
|
|
static int get_bit_ptr(struct k_mem_pool *p, int level, int bn, u32_t **word)
|
|
{
|
|
u32_t *bitarray = level <= p->max_inline_level ?
|
|
&p->levels[level].bits : p->levels[level].bits_p;
|
|
|
|
*word = &bitarray[bn / 32];
|
|
|
|
return bn & 0x1f;
|
|
}
|
|
|
|
static void set_free_bit(struct k_mem_pool *p, int level, int bn)
|
|
{
|
|
u32_t *word;
|
|
int bit = get_bit_ptr(p, level, bn, &word);
|
|
|
|
*word |= (1<<bit);
|
|
}
|
|
|
|
static void clear_free_bit(struct k_mem_pool *p, int level, int bn)
|
|
{
|
|
u32_t *word;
|
|
int bit = get_bit_ptr(p, level, bn, &word);
|
|
|
|
*word &= ~(1<<bit);
|
|
}
|
|
|
|
/* Returns all four of the free bits for the specified blocks
|
|
* "partners" in the bottom 4 bits of the return value
|
|
*/
|
|
static int partner_bits(struct k_mem_pool *p, int level, int bn)
|
|
{
|
|
u32_t *word;
|
|
int bit = get_bit_ptr(p, level, bn, &word);
|
|
|
|
return (*word >> (4*(bit / 4))) & 0xf;
|
|
}
|
|
|
|
static size_t buf_size(struct k_mem_pool *p)
|
|
{
|
|
return p->n_max * p->max_sz;
|
|
}
|
|
|
|
static bool block_fits(struct k_mem_pool *p, void *block, size_t bsz)
|
|
{
|
|
return (block + bsz - 1 - p->buf) < buf_size(p);
|
|
}
|
|
|
|
static void init_mem_pool(struct k_mem_pool *p)
|
|
{
|
|
int i;
|
|
size_t buflen = p->n_max * p->max_sz, sz = p->max_sz;
|
|
u32_t *bits = p->buf + buflen;
|
|
|
|
sys_dlist_init(&p->wait_q);
|
|
|
|
for (i = 0; i < p->n_levels; i++) {
|
|
int nblocks = buflen / sz;
|
|
|
|
sys_dlist_init(&p->levels[i].free_list);
|
|
|
|
if (nblocks < 32) {
|
|
p->max_inline_level = i;
|
|
} else {
|
|
p->levels[i].bits_p = bits;
|
|
bits += (nblocks + 31)/32;
|
|
}
|
|
|
|
sz = _ALIGN4(sz / 4);
|
|
}
|
|
|
|
for (i = 0; i < p->n_max; i++) {
|
|
void *block = block_ptr(p, p->max_sz, i);
|
|
|
|
sys_dlist_append(&p->levels[0].free_list, block);
|
|
set_free_bit(p, 0, i);
|
|
}
|
|
}
|
|
|
|
int init_static_pools(struct device *unused)
|
|
{
|
|
ARG_UNUSED(unused);
|
|
struct k_mem_pool *p;
|
|
|
|
for (p = _k_mem_pool_list_start; p < _k_mem_pool_list_end; p++) {
|
|
init_mem_pool(p);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
SYS_INIT(init_static_pools, PRE_KERNEL_1, CONFIG_KERNEL_INIT_PRIORITY_OBJECTS);
|
|
|
|
/* A note on synchronization: all manipulation of the actual pool data
|
|
* happens in one of alloc_block()/free_block() or break_block(). All
|
|
* of these transition between a state where the caller "holds" a
|
|
* block pointer that is marked used in the store and one where she
|
|
* doesn't (or else they will fail, e.g. if there isn't a free block).
|
|
* So that is the basic operation that needs synchronization, which we
|
|
* can do piecewise as needed in small one-block chunks to preserve
|
|
* latency. At most (in free_block) a single locked operation
|
|
* consists of four bit sets and dlist removals. If the overall
|
|
* allocation operation fails, we just free the block we have (putting
|
|
* a block back into the list cannot fail) and return failure.
|
|
*/
|
|
|
|
static void *alloc_block(struct k_mem_pool *p, int l, size_t lsz)
|
|
{
|
|
sys_dnode_t *block;
|
|
int key = irq_lock();
|
|
|
|
block = sys_dlist_get(&p->levels[l].free_list);
|
|
if (block) {
|
|
clear_free_bit(p, l, block_num(p, block, lsz));
|
|
}
|
|
irq_unlock(key);
|
|
|
|
return block;
|
|
}
|
|
|
|
static void free_block(struct k_mem_pool *p, int level, size_t *lsizes, int bn)
|
|
{
|
|
int i, key, lsz = lsizes[level];
|
|
void *block = block_ptr(p, lsz, bn);
|
|
|
|
key = irq_lock();
|
|
|
|
set_free_bit(p, level, bn);
|
|
|
|
if (level && partner_bits(p, level, bn) == 0xf) {
|
|
for (i = 0; i < 4; i++) {
|
|
int b = (bn & ~3) + i;
|
|
|
|
clear_free_bit(p, level, b);
|
|
if (b != bn &&
|
|
block_fits(p, block_ptr(p, lsz, b), lsz)) {
|
|
sys_dlist_remove(block_ptr(p, lsz, b));
|
|
}
|
|
}
|
|
|
|
irq_unlock(key);
|
|
free_block(p, level-1, lsizes, bn / 4); /* tail recursion! */
|
|
return;
|
|
}
|
|
|
|
if (block_fits(p, block, lsz)) {
|
|
sys_dlist_append(&p->levels[level].free_list, block);
|
|
}
|
|
|
|
irq_unlock(key);
|
|
}
|
|
|
|
/* Takes a block of a given level, splits it into four blocks of the
|
|
* next smaller level, puts three into the free list as in
|
|
* free_block() but without the need to check adjacent bits or
|
|
* recombine, and returns the remaining smaller block.
|
|
*/
|
|
static void *break_block(struct k_mem_pool *p, void *block,
|
|
int l, size_t *lsizes)
|
|
{
|
|
int i, bn, key;
|
|
|
|
key = irq_lock();
|
|
|
|
bn = block_num(p, block, lsizes[l]);
|
|
|
|
for (i = 1; i < 4; i++) {
|
|
int lbn = 4*bn + i;
|
|
int lsz = lsizes[l + 1];
|
|
void *block2 = (lsz * i) + (char *)block;
|
|
|
|
set_free_bit(p, l + 1, lbn);
|
|
if (block_fits(p, block2, lsz)) {
|
|
sys_dlist_append(&p->levels[l + 1].free_list, block2);
|
|
}
|
|
}
|
|
|
|
irq_unlock(key);
|
|
|
|
return block;
|
|
}
|
|
|
|
static int pool_alloc(struct k_mem_pool *p, struct k_mem_block *block,
|
|
size_t size)
|
|
{
|
|
size_t lsizes[p->n_levels];
|
|
int i, alloc_l = -1, free_l = -1, from_l;
|
|
void *blk = NULL;
|
|
|
|
/* Walk down through levels, finding the one from which we
|
|
* want to allocate and the smallest one with a free entry
|
|
* from which we can split an allocation if needed. Along the
|
|
* way, we populate an array of sizes for each level so we
|
|
* don't need to waste RAM storing it.
|
|
*/
|
|
lsizes[0] = _ALIGN4(p->max_sz);
|
|
for (i = 0; i < p->n_levels; i++) {
|
|
if (i > 0) {
|
|
lsizes[i] = _ALIGN4(lsizes[i-1] / 4);
|
|
}
|
|
|
|
if (lsizes[i] < size) {
|
|
break;
|
|
}
|
|
|
|
alloc_l = i;
|
|
if (!level_empty(p, i)) {
|
|
free_l = i;
|
|
}
|
|
}
|
|
|
|
if (alloc_l < 0 || free_l < 0) {
|
|
block->data = NULL;
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/* Iteratively break the smallest enclosing block... */
|
|
blk = alloc_block(p, free_l, lsizes[free_l]);
|
|
|
|
if (!blk) {
|
|
/* This can happen if we race with another allocator.
|
|
* It's OK, just back out and the timeout code will
|
|
* retry. Note mild overloading: -EAGAIN isn't for
|
|
* propagation to the caller, it's to tell the loop in
|
|
* k_mem_pool_alloc() to try again synchronously. But
|
|
* it means exactly what it says.
|
|
*/
|
|
return -EAGAIN;
|
|
}
|
|
|
|
for (from_l = free_l; from_l < alloc_l; from_l++) {
|
|
blk = break_block(p, blk, from_l, lsizes);
|
|
}
|
|
|
|
/* ... until we have something to return */
|
|
block->data = blk;
|
|
block->id.pool = pool_id(p);
|
|
block->id.level = alloc_l;
|
|
block->id.block = block_num(p, block->data, lsizes[alloc_l]);
|
|
return 0;
|
|
}
|
|
|
|
int k_mem_pool_alloc(struct k_mem_pool *p, struct k_mem_block *block,
|
|
size_t size, s32_t timeout)
|
|
{
|
|
int ret, key;
|
|
s64_t end = 0;
|
|
|
|
__ASSERT(!(_is_in_isr() && timeout != K_NO_WAIT), "");
|
|
|
|
if (timeout > 0) {
|
|
end = _tick_get() + _ms_to_ticks(timeout);
|
|
}
|
|
|
|
while (1) {
|
|
ret = pool_alloc(p, block, size);
|
|
|
|
if (ret == 0 || timeout == K_NO_WAIT ||
|
|
ret == -EAGAIN || (ret && ret != -ENOMEM)) {
|
|
return ret;
|
|
}
|
|
|
|
key = irq_lock();
|
|
_pend_current_thread(&p->wait_q, timeout);
|
|
_Swap(key);
|
|
|
|
if (timeout != K_FOREVER) {
|
|
timeout = end - _tick_get();
|
|
|
|
if (timeout < 0) {
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
return -EAGAIN;
|
|
}
|
|
|
|
void k_mem_pool_free(struct k_mem_block *block)
|
|
{
|
|
int i, key, need_sched = 0;
|
|
struct k_mem_pool *p = get_pool(block->id.pool);
|
|
size_t lsizes[p->n_levels];
|
|
|
|
/* As in k_mem_pool_alloc(), we build a table of level sizes
|
|
* to avoid having to store it in precious RAM bytes.
|
|
* Overhead here is somewhat higher because free_block()
|
|
* doesn't inherently need to traverse all the larger
|
|
* sublevels.
|
|
*/
|
|
lsizes[0] = _ALIGN4(p->max_sz);
|
|
for (i = 1; i <= block->id.level; i++) {
|
|
lsizes[i] = _ALIGN4(lsizes[i-1] / 4);
|
|
}
|
|
|
|
free_block(get_pool(block->id.pool), block->id.level,
|
|
lsizes, block->id.block);
|
|
|
|
/* Wake up anyone blocked on this pool and let them repeat
|
|
* their allocation attempts
|
|
*/
|
|
key = irq_lock();
|
|
|
|
while (!sys_dlist_is_empty(&p->wait_q)) {
|
|
struct k_thread *th = (void *)sys_dlist_peek_head(&p->wait_q);
|
|
|
|
_unpend_thread(th);
|
|
_abort_thread_timeout(th);
|
|
_ready_thread(th);
|
|
need_sched = 1;
|
|
}
|
|
|
|
if (need_sched && !_is_in_isr()) {
|
|
_reschedule_threads(key);
|
|
} else {
|
|
irq_unlock(key);
|
|
}
|
|
}
|
|
|
|
#if (CONFIG_HEAP_MEM_POOL_SIZE > 0)
|
|
|
|
/*
|
|
* Heap is defined using HEAP_MEM_POOL_SIZE configuration option.
|
|
*
|
|
* This module defines the heap memory pool and the _HEAP_MEM_POOL symbol
|
|
* that has the address of the associated memory pool struct.
|
|
*/
|
|
|
|
K_MEM_POOL_DEFINE(_heap_mem_pool, 64, CONFIG_HEAP_MEM_POOL_SIZE, 1, 4);
|
|
#define _HEAP_MEM_POOL (&_heap_mem_pool)
|
|
|
|
void *k_malloc(size_t size)
|
|
{
|
|
struct k_mem_block block;
|
|
|
|
/*
|
|
* get a block large enough to hold an initial (hidden) block
|
|
* descriptor, as well as the space the caller requested
|
|
*/
|
|
size += sizeof(struct k_mem_block);
|
|
if (k_mem_pool_alloc(_HEAP_MEM_POOL, &block, size, K_NO_WAIT) != 0) {
|
|
return NULL;
|
|
}
|
|
|
|
/* save the block descriptor info at the start of the actual block */
|
|
memcpy(block.data, &block, sizeof(struct k_mem_block));
|
|
|
|
/* return address of the user area part of the block to the caller */
|
|
return (char *)block.data + sizeof(struct k_mem_block);
|
|
}
|
|
|
|
|
|
void k_free(void *ptr)
|
|
{
|
|
if (ptr != NULL) {
|
|
/* point to hidden block descriptor at start of block */
|
|
ptr = (char *)ptr - sizeof(struct k_mem_block);
|
|
|
|
/* return block to the heap memory pool */
|
|
k_mem_pool_free(ptr);
|
|
}
|
|
}
|
|
|
|
void *k_calloc(size_t nmemb, size_t size)
|
|
{
|
|
void *ret;
|
|
size_t bounds;
|
|
|
|
#ifdef CONFIG_ASSERT
|
|
__ASSERT(!__builtin_mul_overflow(nmemb, size, &bounds),
|
|
"requested size overflow");
|
|
#else
|
|
bounds = nmemb * size;
|
|
#endif
|
|
ret = k_malloc(bounds);
|
|
if (ret) {
|
|
memset(ret, 0, bounds);
|
|
}
|
|
return ret;
|
|
}
|
|
#endif
|