/**************************************************************************** * mm/mm_heap/mm_memalign.c * * SPDX-License-Identifier: Apache-2.0 * * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. The * ASF licenses this file to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance with the * License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the * License for the specific language governing permissions and limitations * under the License. * ****************************************************************************/ /**************************************************************************** * Included Files ****************************************************************************/ #include #include #include #include "mm_heap/mm.h" #include "kasan/kasan.h" /**************************************************************************** * Public Functions ****************************************************************************/ /**************************************************************************** * Name: mm_memalign * * Description: * memalign requests more than enough space from malloc, finds a region * within that chunk that meets the alignment request and then frees any * leading or trailing space. * * The alignment argument must be a power of two. 16-byte alignment is * guaranteed by normal malloc calls. * ****************************************************************************/ FAR void *mm_memalign(FAR struct mm_heap_s *heap, size_t alignment, size_t size) { FAR struct mm_allocnode_s *node; uintptr_t rawchunk; uintptr_t alignedchunk; size_t mask; size_t allocsize; size_t newsize; /* Make sure that alignment is less than half max size_t */ if (alignment >= (SIZE_MAX / 2)) { return NULL; } /* Make sure that alignment is a power of 2 */ if ((alignment & -alignment) != alignment) { return NULL; } #ifdef CONFIG_MM_HEAP_MEMPOOL if (heap->mm_mpool) { node = mempool_multiple_memalign(heap->mm_mpool, alignment, size); if (node != NULL) { return node; } } #endif /* If this requested alinement's less than or equal to the natural * alignment of malloc, then just let malloc do the work. */ if (alignment <= MM_ALIGN) { FAR void *ptr = mm_malloc(heap, size); DEBUGASSERT(ptr == NULL || ((uintptr_t)ptr) % alignment == 0); return ptr; } else if (alignment < MM_MIN_CHUNK) { alignment = MM_MIN_CHUNK; } mask = alignment - 1; /* Adjust the size to account for (1) the size of the allocated node and * (2) to make sure that it is aligned with MM_ALIGN and its size is at * least MM_MIN_CHUNK. * * Notice that we increase the allocation size by twice the requested * alignment. We do this so that there will be at least two valid * alignment points within the allocated memory. * * NOTE: These are sizes given to malloc and not chunk sizes. They do * not include MM_SIZEOF_ALLOCNODE. */ if (size < MM_MIN_CHUNK - MM_ALLOCNODE_OVERHEAD) { size = MM_MIN_CHUNK - MM_ALLOCNODE_OVERHEAD; } newsize = MM_ALIGN_UP(size); /* Make multiples of our granule size */ allocsize = newsize + 2 * alignment; /* Add double full alignment size */ if (newsize < size || allocsize < newsize) { /* Integer overflow */ return NULL; } /* Then malloc that size */ rawchunk = (uintptr_t)mm_malloc(heap, allocsize); if (rawchunk == 0) { return NULL; } kasan_poison((FAR void *)rawchunk, mm_malloc_size(heap, (FAR void *)rawchunk)); /* We need to hold the MM mutex while we muck with the chunks and * nodelist. */ DEBUGVERIFY(mm_lock(heap)); /* Get the node associated with the allocation and the next node after * the allocation. */ node = (FAR struct mm_allocnode_s *)(rawchunk - MM_SIZEOF_ALLOCNODE); heap->mm_curused -= MM_SIZEOF_NODE(node); /* Find the aligned subregion */ alignedchunk = (rawchunk + mask) & ~mask; /* Check if there is free space at the beginning of the aligned chunk */ if (alignedchunk != rawchunk) { FAR struct mm_allocnode_s *newnode; FAR struct mm_allocnode_s *next; size_t precedingsize; size_t newnodesize; /* Get the node the next node after the allocation. */ next = (FAR struct mm_allocnode_s *) ((FAR char *)node + MM_SIZEOF_NODE(node)); newnode = (FAR struct mm_allocnode_s *) (alignedchunk - MM_SIZEOF_ALLOCNODE); /* Preceding size is full size of the new 'node,' including * MM_SIZEOF_ALLOCNODE */ precedingsize = (uintptr_t)newnode - (uintptr_t)node; /* If we were unlucky, then the alignedchunk can lie in such a position * that precedingsize < SIZEOF_NODE_FREENODE. We can't let that happen * because we are going to cast 'node' to struct mm_freenode_s below. * This is why we allocated memory large enough to support two * alignment points. In this case, we will simply use the second * alignment point. */ if (precedingsize < MM_MIN_CHUNK) { alignedchunk += alignment; newnode = (FAR struct mm_allocnode_s *) (alignedchunk - MM_SIZEOF_ALLOCNODE); precedingsize = (uintptr_t)newnode - (uintptr_t)node; } /* If the previous node is free, merge node and previous node, then * set up the node size. */ if (MM_PREVNODE_IS_FREE(node)) { FAR struct mm_freenode_s *prev = (FAR struct mm_freenode_s *)((FAR char *)node - node->preceding); /* Remove the node. There must be a predecessor, but there may * not be a successor node. */ DEBUGASSERT(prev->blink); prev->blink->flink = prev->flink; if (prev->flink) { prev->flink->blink = prev->blink; } precedingsize += MM_SIZEOF_NODE(prev); node = (FAR struct mm_allocnode_s *)prev; } node->size = precedingsize; /* Set up the size of the new node */ newnodesize = (uintptr_t)next - (uintptr_t)newnode; newnode->size = newnodesize | MM_ALLOC_BIT | MM_PREVFREE_BIT; newnode->preceding = precedingsize; /* Clear the previous free bit of the next node */ next->size &= ~MM_PREVFREE_BIT; /* Convert the newnode chunk size back into malloc-compatible size by * subtracting the header size MM_ALLOCNODE_OVERHEAD. */ allocsize = newnodesize - MM_ALLOCNODE_OVERHEAD; /* Add the original, newly freed node to the free nodelist */ mm_addfreechunk(heap, (FAR struct mm_freenode_s *)node); /* Replace the original node with the newlay realloaced, * aligned node */ node = newnode; } /* Check if there is free space at the end of the aligned chunk. Convert * malloc-compatible chunk size to include MM_ALLOCNODE_OVERHEAD as needed * for mm_shrinkchunk. */ size = MM_ALIGN_UP(size + MM_ALLOCNODE_OVERHEAD); if (allocsize > size) { /* Shrink the chunk by that much -- remember, mm_shrinkchunk wants * internal chunk sizes that include MM_ALLOCNODE_OVERHEAD. */ mm_shrinkchunk(heap, node, size); } /* Update heap statistics */ heap->mm_curused += MM_SIZEOF_NODE(node); if (heap->mm_curused > heap->mm_maxused) { heap->mm_maxused = heap->mm_curused; } mm_unlock(heap); MM_ADD_BACKTRACE(heap, node); kasan_unpoison((FAR void *)alignedchunk, mm_malloc_size(heap, (FAR void *)alignedchunk)); DEBUGASSERT(alignedchunk % alignment == 0); return (FAR void *)alignedchunk; }