/* * Copyright (c) 2017 Intel Corporation * * SPDX-License-Identifier: Apache-2.0 */ #include #include #include #include #include #include #include /* Linker-defined symbols bound the static pool structs */ extern struct k_mem_pool _k_mem_pool_list_start[]; extern struct k_mem_pool _k_mem_pool_list_end[]; s64_t _tick_get(void); static struct k_mem_pool *get_pool(int id) { return &_k_mem_pool_list_start[id]; } static int pool_id(struct k_mem_pool *pool) { return pool - &_k_mem_pool_list_start[0]; } static void k_mem_pool_init(struct k_mem_pool *p) { _waitq_init(&p->wait_q); _sys_mem_pool_base_init(&p->base); } int init_static_pools(struct device *unused) { ARG_UNUSED(unused); struct k_mem_pool *p; for (p = _k_mem_pool_list_start; p < _k_mem_pool_list_end; p++) { k_mem_pool_init(p); } return 0; } SYS_INIT(init_static_pools, PRE_KERNEL_1, CONFIG_KERNEL_INIT_PRIORITY_OBJECTS); int k_mem_pool_alloc(struct k_mem_pool *p, struct k_mem_block *block, size_t size, s32_t timeout) { int ret; s64_t end = 0; __ASSERT(!(_is_in_isr() && timeout != K_NO_WAIT), ""); if (timeout > 0) { end = _tick_get() + _ms_to_ticks(timeout); } while (true) { u32_t level_num, block_num; /* There is a "managed race" in alloc that can fail * (albeit in a well-defined way, see comments there) * with -EAGAIN when simultaneous allocations happen. * Retry exactly once before sleeping to resolve it. * If we're so contended that it fails twice, then we * clearly want to block. */ for (int i = 0; i < 2; i++) { ret = _sys_mem_pool_block_alloc(&p->base, size, &level_num, &block_num, &block->data); if (ret != -EAGAIN) { break; } } if (ret == -EAGAIN) { ret = -ENOMEM; } block->id.pool = pool_id(p); block->id.level = level_num; block->id.block = block_num; if (ret == 0 || timeout == K_NO_WAIT || (ret && ret != -ENOMEM)) { return ret; } (void)_pend_current_thread(irq_lock(), &p->wait_q, timeout); if (timeout != K_FOREVER) { timeout = end - _tick_get(); if (timeout < 0) { break; } } } return -EAGAIN; } void k_mem_pool_free_id(struct k_mem_block_id *id) { int key, need_sched = 0; struct k_mem_pool *p = get_pool(id->pool); _sys_mem_pool_block_free(&p->base, id->level, id->block); /* Wake up anyone blocked on this pool and let them repeat * their allocation attempts */ key = irq_lock(); need_sched = _unpend_all(&p->wait_q); if (need_sched && !_is_in_isr()) { _reschedule(key); } else { irq_unlock(key); } } void k_mem_pool_free(struct k_mem_block *block) { k_mem_pool_free_id(&block->id); } void *k_mem_pool_malloc(struct k_mem_pool *pool, size_t size) { struct k_mem_block block; /* * get a block large enough to hold an initial (hidden) block * descriptor, as well as the space the caller requested */ if (__builtin_add_overflow(size, sizeof(struct k_mem_block_id), &size)) { return NULL; } if (k_mem_pool_alloc(pool, &block, size, K_NO_WAIT) != 0) { return NULL; } /* save the block descriptor info at the start of the actual block */ (void)memcpy(block.data, &block.id, sizeof(struct k_mem_block_id)); /* return address of the user area part of the block to the caller */ return (char *)block.data + sizeof(struct k_mem_block_id); } void k_free(void *ptr) { if (ptr != NULL) { /* point to hidden block descriptor at start of block */ ptr = (char *)ptr - sizeof(struct k_mem_block_id); /* return block to the heap memory pool */ k_mem_pool_free_id(ptr); } } #if (CONFIG_HEAP_MEM_POOL_SIZE > 0) /* * Heap is defined using HEAP_MEM_POOL_SIZE configuration option. * * This module defines the heap memory pool and the _HEAP_MEM_POOL symbol * that has the address of the associated memory pool struct. */ K_MEM_POOL_DEFINE(_heap_mem_pool, 64, CONFIG_HEAP_MEM_POOL_SIZE, 1, 4); #define _HEAP_MEM_POOL (&_heap_mem_pool) void *k_malloc(size_t size) { return k_mem_pool_malloc(_HEAP_MEM_POOL, size); } void *k_calloc(size_t nmemb, size_t size) { void *ret; size_t bounds; if (__builtin_mul_overflow(nmemb, size, &bounds)) { return NULL; } ret = k_malloc(bounds); if (ret != NULL) { (void)memset(ret, 0, bounds); } return ret; } void k_thread_system_pool_assign(struct k_thread *thread) { thread->resource_pool = _HEAP_MEM_POOL; } #endif void *z_thread_malloc(size_t size) { void *ret; if (_current->resource_pool) { ret = k_mem_pool_malloc(_current->resource_pool, size); } else { ret = NULL; } return ret; }