forked from Archive/PX4-Autopilot
209 lines
7.1 KiB
C
209 lines
7.1 KiB
C
|
/****************************************************************************
|
||
|
* mm/mm_memalign.c
|
||
|
*
|
||
|
* Copyright (C) 2007, 2009, 2011 Gregory Nutt. All rights reserved.
|
||
|
* Author: Gregory Nutt <gnutt@nuttx.org>
|
||
|
*
|
||
|
* Redistribution and use in source and binary forms, with or without
|
||
|
* modification, are permitted provided that the following conditions
|
||
|
* are met:
|
||
|
*
|
||
|
* 1. Redistributions of source code must retain the above copyright
|
||
|
* notice, this list of conditions and the following disclaimer.
|
||
|
* 2. Redistributions in binary form must reproduce the above copyright
|
||
|
* notice, this list of conditions and the following disclaimer in
|
||
|
* the documentation and/or other materials provided with the
|
||
|
* distribution.
|
||
|
* 3. Neither the name NuttX nor the names of its contributors may be
|
||
|
* used to endorse or promote products derived from this software
|
||
|
* without specific prior written permission.
|
||
|
*
|
||
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
||
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
||
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
|
||
|
* FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
|
||
|
* COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
|
||
|
* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
|
||
|
* BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS
|
||
|
* OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED
|
||
|
* AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
||
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
|
||
|
* ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
||
|
* POSSIBILITY OF SUCH DAMAGE.
|
||
|
*
|
||
|
****************************************************************************/
|
||
|
|
||
|
/****************************************************************************
|
||
|
* Included Files
|
||
|
****************************************************************************/
|
||
|
|
||
|
#include <assert.h>
|
||
|
|
||
|
#include "mm_environment.h"
|
||
|
#include "mm_internal.h"
|
||
|
|
||
|
/****************************************************************************
|
||
|
* Pre-processor Definitions
|
||
|
****************************************************************************/
|
||
|
|
||
|
/****************************************************************************
|
||
|
* Global Functions
|
||
|
****************************************************************************/
|
||
|
|
||
|
/****************************************************************************
|
||
|
* Name: memalign
|
||
|
*
|
||
|
* Description:
|
||
|
* memalign requests more than enough space from malloc, finds a region
|
||
|
* within that chunk that meets the alignment request and then frees any
|
||
|
* leading or trailing space.
|
||
|
*
|
||
|
* The alignment argument must be a power of two (not checked). 8-byte
|
||
|
* alignment is guaranteed by normal malloc calls.
|
||
|
*
|
||
|
****************************************************************************/
|
||
|
|
||
|
FAR void *memalign(size_t alignment, size_t size)
|
||
|
{
|
||
|
FAR struct mm_allocnode_s *node;
|
||
|
size_t rawchunk;
|
||
|
size_t alignedchunk;
|
||
|
size_t mask = (size_t)(alignment - 1);
|
||
|
size_t allocsize;
|
||
|
|
||
|
/* If this requested alignement less than or equal to the natural alignment
|
||
|
* of malloc, then just let malloc do the work.
|
||
|
*/
|
||
|
|
||
|
if (alignment <= MM_MIN_CHUNK)
|
||
|
{
|
||
|
return malloc(size);
|
||
|
}
|
||
|
|
||
|
/* Adjust the size to account for (1) the size of the allocated node, (2)
|
||
|
* to make sure that it is an even multiple of our granule size, and to
|
||
|
* include the alignment amount.
|
||
|
*
|
||
|
* Notice that we increase the allocation size by twice the requested
|
||
|
* alignment. We do this so that there will be at least two valid
|
||
|
* alignment points within the allocated memory.
|
||
|
*
|
||
|
* NOTE: These are sizes given to malloc and not chunk sizes. They do
|
||
|
* not include SIZEOF_MM_ALLOCNODE.
|
||
|
*/
|
||
|
|
||
|
size = MM_ALIGN_UP(size); /* Make multiples of our granule size */
|
||
|
allocsize = size + 2*alignment; /* Add double full alignment size */
|
||
|
|
||
|
/* Then malloc that size */
|
||
|
|
||
|
rawchunk = (size_t)malloc(allocsize);
|
||
|
if (rawchunk == 0)
|
||
|
{
|
||
|
return NULL;
|
||
|
}
|
||
|
|
||
|
/* We need to hold the MM semaphore while we muck with the chunks and
|
||
|
* nodelist.
|
||
|
*/
|
||
|
|
||
|
mm_takesemaphore();
|
||
|
|
||
|
/* Get the node associated with the allocation and the next node after
|
||
|
* the allocation.
|
||
|
*/
|
||
|
|
||
|
node = (FAR struct mm_allocnode_s*)(rawchunk - SIZEOF_MM_ALLOCNODE);
|
||
|
|
||
|
/* Find the aligned subregion */
|
||
|
|
||
|
alignedchunk = (rawchunk + mask) & ~mask;
|
||
|
|
||
|
/* Check if there is free space at the beginning of the aligned chunk */
|
||
|
|
||
|
if (alignedchunk != rawchunk)
|
||
|
{
|
||
|
FAR struct mm_allocnode_s *newnode;
|
||
|
FAR struct mm_allocnode_s *next;
|
||
|
size_t precedingsize;
|
||
|
|
||
|
/* Get the node the next node after the allocation. */
|
||
|
|
||
|
next = (FAR struct mm_allocnode_s*)((char*)node + node->size);
|
||
|
|
||
|
/* Make sure that there is space to convert the preceding mm_allocnode_s
|
||
|
* into an mm_freenode_s. I think that this should always be true
|
||
|
*/
|
||
|
|
||
|
DEBUGASSERT(alignedchunk >= rawchunk + 8);
|
||
|
|
||
|
newnode = (FAR struct mm_allocnode_s*)(alignedchunk - SIZEOF_MM_ALLOCNODE);
|
||
|
|
||
|
/* Preceding size is full size of the new 'node,' including
|
||
|
* SIZEOF_MM_ALLOCNODE
|
||
|
*/
|
||
|
|
||
|
precedingsize = (size_t)newnode - (size_t)node;
|
||
|
|
||
|
/* If we were unlucky, then the alignedchunk can lie in such a position
|
||
|
* that precedingsize < SIZEOF_NODE_FREENODE. We can't let that happen
|
||
|
* because we are going to cast 'node' to struct mm_freenode_s below.
|
||
|
* This is why we allocated memory large enough to support two
|
||
|
* alignment points. In this case, we will simply use the second
|
||
|
* alignment point.
|
||
|
*/
|
||
|
|
||
|
if (precedingsize < SIZEOF_MM_FREENODE)
|
||
|
{
|
||
|
alignedchunk += alignment;
|
||
|
newnode = (FAR struct mm_allocnode_s*)(alignedchunk - SIZEOF_MM_ALLOCNODE);
|
||
|
precedingsize = (size_t)newnode - (size_t)node;
|
||
|
}
|
||
|
|
||
|
/* Set up the size of the new node */
|
||
|
|
||
|
newnode->size = (size_t)next - (size_t)newnode;
|
||
|
newnode->preceding = precedingsize | MM_ALLOC_BIT;
|
||
|
|
||
|
/* Reduce the size of the original chunk and mark it not allocated, */
|
||
|
|
||
|
node->size = precedingsize;
|
||
|
node->preceding &= ~MM_ALLOC_BIT;
|
||
|
|
||
|
/* Fix the preceding size of the next node */
|
||
|
|
||
|
next->preceding = newnode->size | (next->preceding & MM_ALLOC_BIT);
|
||
|
|
||
|
/* Convert the newnode chunk size back into malloc-compatible size by
|
||
|
* subtracting the header size SIZEOF_MM_ALLOCNODE.
|
||
|
*/
|
||
|
|
||
|
allocsize = newnode->size - SIZEOF_MM_ALLOCNODE;
|
||
|
|
||
|
/* Add the original, newly freed node to the free nodelist */
|
||
|
|
||
|
mm_addfreechunk((FAR struct mm_freenode_s *)node);
|
||
|
|
||
|
/* Replace the original node with the newlay realloaced,
|
||
|
* aligned node
|
||
|
*/
|
||
|
|
||
|
node = newnode;
|
||
|
}
|
||
|
|
||
|
/* Check if there is free space at the end of the aligned chunk */
|
||
|
|
||
|
if (allocsize > size)
|
||
|
{
|
||
|
/* Shrink the chunk by that much -- remember, mm_shrinkchunk wants
|
||
|
* internal chunk sizes that include SIZEOF_MM_ALLOCNODE, and not the
|
||
|
* malloc-compatible sizes that we have.
|
||
|
*/
|
||
|
|
||
|
mm_shrinkchunk(node, size + SIZEOF_MM_ALLOCNODE);
|
||
|
}
|
||
|
|
||
|
mm_givesemaphore();
|
||
|
return (FAR void*)alignedchunk;
|
||
|
}
|