mirror of
https://github.com/jerryscript-project/jerryscript.git
synced 2025-12-15 16:29:21 +00:00
1175 lines
37 KiB
C++
1175 lines
37 KiB
C++
/* Copyright 2014-2015 Samsung Electronics Co., Ltd.
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
/** \addtogroup mem Memory allocation
|
|
* @{
|
|
*
|
|
* \addtogroup heap Heap
|
|
* @{
|
|
*/
|
|
|
|
/**
|
|
* Heap implementation
|
|
*/
|
|
|
|
#include "jrt.h"
|
|
#include "jrt-libc-includes.h"
|
|
#include "mem-allocator.h"
|
|
#include "mem-config.h"
|
|
#include "mem-heap.h"
|
|
|
|
#define MEM_ALLOCATOR_INTERNAL
|
|
|
|
#include "mem-allocator-internal.h"
|
|
|
|
/*
|
|
* Valgrind-related options and headers
|
|
*/
|
|
#ifdef JERRY_VALGRIND
|
|
# include "memcheck.h"
|
|
|
|
# define VALGRIND_NOACCESS_STRUCT(s) (void)VALGRIND_MAKE_MEM_NOACCESS((s), sizeof (*(s)))
|
|
# define VALGRIND_UNDEFINED_STRUCT(s) (void)VALGRIND_MAKE_MEM_UNDEFINED((s), sizeof (*(s)))
|
|
# define VALGRIND_DEFINED_STRUCT(s) (void)VALGRIND_MAKE_MEM_DEFINED((s), sizeof (*(s)))
|
|
# define VALGRIND_NOACCESS_SPACE(p, s) (void)VALGRIND_MAKE_MEM_NOACCESS((p), (s))
|
|
# define VALGRIND_UNDEFINED_SPACE(p, s) (void)VALGRIND_MAKE_MEM_UNDEFINED((p), (s))
|
|
# define VALGRIND_DEFINED_SPACE(p, s) (void)VALGRIND_MAKE_MEM_DEFINED((p), (s))
|
|
#else /* JERRY_VALGRIND */
|
|
# define VALGRIND_NOACCESS_STRUCT(s)
|
|
# define VALGRIND_UNDEFINED_STRUCT(s)
|
|
# define VALGRIND_DEFINED_STRUCT(s)
|
|
# define VALGRIND_NOACCESS_SPACE(p, s)
|
|
# define VALGRIND_UNDEFINED_SPACE(p, s)
|
|
# define VALGRIND_DEFINED_SPACE(p, s)
|
|
#endif /* JERRY_VALGRIND */
|
|
|
|
/**
|
|
* Magic numbers for heap memory blocks
|
|
*/
|
|
typedef enum
|
|
{
|
|
MEM_MAGIC_NUM_OF_FREE_BLOCK = 0xc809,
|
|
MEM_MAGIC_NUM_OF_ALLOCATED_BLOCK = 0x5b46
|
|
} mem_magic_num_of_block_t;
|
|
|
|
/**
|
|
* State of the block to initialize (argument of mem_init_block_header)
|
|
*
|
|
* @see mem_init_block_header
|
|
*/
|
|
typedef enum
|
|
{
|
|
MEM_BLOCK_FREE, /**< initializing free block */
|
|
MEM_BLOCK_ALLOCATED /**< initializing allocated block */
|
|
} mem_block_state_t;
|
|
|
|
/**
|
|
* Linked list direction descriptors
|
|
*/
|
|
typedef enum
|
|
{
|
|
MEM_DIRECTION_PREV = 0, /**< direction from right to left */
|
|
MEM_DIRECTION_NEXT = 1, /**< direction from left to right */
|
|
MEM_DIRECTION_COUNT = 2 /**< count of possible directions */
|
|
} mem_direction_t;
|
|
|
|
/**
|
|
* Offset in the heap
|
|
*/
|
|
typedef uint16_t mem_heap_offset_t;
|
|
JERRY_STATIC_ASSERT (sizeof (mem_heap_offset_t) * JERRY_BITSINBYTE >= MEM_HEAP_OFFSET_LOG);
|
|
|
|
/**
|
|
* Description of heap memory block layout
|
|
*/
|
|
typedef struct __attribute__ ((aligned (MEM_ALIGNMENT))) mem_block_header_t
|
|
{
|
|
uint16_t magic_num; /**< magic number (mem_magic_num_of_block_t):
|
|
MEM_MAGIC_NUM_OF_ALLOCATED_BLOCK for allocated block
|
|
or MEM_MAGIC_NUM_OF_FREE_BLOCK for free block */
|
|
mem_heap_offset_t allocated_bytes; /**< allocated area size - for allocated blocks;
|
|
0 - for free blocks */
|
|
mem_heap_offset_t neighbours[ MEM_DIRECTION_COUNT ]; /**< neighbour blocks' offsets;
|
|
0 - if the block is last in specified direction */
|
|
} mem_block_header_t;
|
|
|
|
/**
|
|
* Check that block header's size is not more than 8 bytes
|
|
*/
|
|
JERRY_STATIC_ASSERT (sizeof (mem_block_header_t) <= sizeof (uint64_t));
|
|
|
|
/**
|
|
* Chunk should have enough space for block header
|
|
*/
|
|
JERRY_STATIC_ASSERT (MEM_HEAP_CHUNK_SIZE >= sizeof (mem_block_header_t));
|
|
|
|
/**
|
|
* Chunk size should satisfy the required alignment value
|
|
*/
|
|
JERRY_STATIC_ASSERT (MEM_HEAP_CHUNK_SIZE % MEM_ALIGNMENT == 0);
|
|
|
|
/**
|
|
* Description of heap state
|
|
*/
|
|
typedef struct
|
|
{
|
|
uint8_t* heap_start; /**< first address of heap space */
|
|
size_t heap_size; /**< heap space size */
|
|
mem_block_header_t* first_block_p; /**< first block of the heap */
|
|
mem_block_header_t* last_block_p; /**< last block of the heap */
|
|
} mem_heap_state_t;
|
|
|
|
/**
|
|
* Heap state
|
|
*/
|
|
mem_heap_state_t mem_heap;
|
|
|
|
static size_t mem_get_block_chunks_count (const mem_block_header_t *block_header_p);
|
|
static size_t mem_get_block_data_space_size (const mem_block_header_t *block_header_p);
|
|
static size_t mem_get_block_chunks_count_from_data_size (size_t block_allocated_size);
|
|
|
|
static void mem_init_block_header (uint8_t *first_chunk_p,
|
|
size_t size_in_chunks,
|
|
mem_block_state_t block_state,
|
|
mem_block_header_t *prev_block_p,
|
|
mem_block_header_t *next_block_p);
|
|
static void mem_check_heap (void);
|
|
|
|
#ifdef MEM_STATS
|
|
/**
|
|
* Heap's memory usage statistics
|
|
*/
|
|
static mem_heap_stats_t mem_heap_stats;
|
|
|
|
static void mem_heap_stat_init (void);
|
|
static void mem_heap_stat_alloc_block (mem_block_header_t *block_header_p);
|
|
static void mem_heap_stat_free_block (mem_block_header_t *block_header_p);
|
|
static void mem_heap_stat_free_block_split (void);
|
|
static void mem_heap_stat_free_block_merge (void);
|
|
|
|
# define MEM_HEAP_STAT_INIT() mem_heap_stat_init ()
|
|
# define MEM_HEAP_STAT_ALLOC_BLOCK(v) mem_heap_stat_alloc_block (v)
|
|
# define MEM_HEAP_STAT_FREE_BLOCK(v) mem_heap_stat_free_block (v)
|
|
# define MEM_HEAP_STAT_FREE_BLOCK_SPLIT() mem_heap_stat_free_block_split ()
|
|
# define MEM_HEAP_STAT_FREE_BLOCK_MERGE() mem_heap_stat_free_block_merge ()
|
|
#else /* !MEM_STATS */
|
|
# define MEM_HEAP_STAT_INIT()
|
|
# define MEM_HEAP_STAT_ALLOC_BLOCK(v)
|
|
# define MEM_HEAP_STAT_FREE_BLOCK(v)
|
|
# define MEM_HEAP_STAT_FREE_BLOCK_SPLIT()
|
|
# define MEM_HEAP_STAT_FREE_BLOCK_MERGE()
|
|
#endif /* !MEM_STATS */
|
|
|
|
/**
|
|
* Measure distance between blocks.
|
|
*
|
|
* Warning:
|
|
* another_block_p should be greater than block_p.
|
|
*
|
|
* @return size in bytes between beginning of two blocks
|
|
*/
|
|
static mem_heap_offset_t
|
|
mem_get_blocks_distance (const mem_block_header_t* block_p, /**< block to measure offset from */
|
|
const mem_block_header_t* another_block_p) /**< block offset is measured for */
|
|
{
|
|
JERRY_ASSERT (another_block_p >= block_p);
|
|
|
|
ssize_t distance = ((uint8_t*) another_block_p - (uint8_t*)block_p);
|
|
|
|
JERRY_ASSERT (distance == (mem_heap_offset_t) distance);
|
|
|
|
return (mem_heap_offset_t) distance;
|
|
} /* mem_get_blocks_distance */
|
|
|
|
/**
|
|
* Get value for neighbour field.
|
|
*
|
|
* Note:
|
|
* If second_block_p is next neighbour of first_block_p,
|
|
* then first_block_p->neighbours[next] = ret_val
|
|
* second_block_p->neighbours[prev] = ret_val
|
|
*
|
|
* @return offset value for neighbours field
|
|
*/
|
|
static mem_heap_offset_t
|
|
mem_get_block_neighbour_field (const mem_block_header_t* first_block_p, /**< first of the blocks
|
|
in forward direction */
|
|
const mem_block_header_t* second_block_p) /**< second of the blocks
|
|
in forward direction */
|
|
{
|
|
JERRY_ASSERT (first_block_p != NULL
|
|
|| second_block_p != NULL);
|
|
|
|
if (first_block_p == NULL
|
|
|| second_block_p == NULL)
|
|
{
|
|
return 0;
|
|
}
|
|
else
|
|
{
|
|
JERRY_ASSERT (first_block_p < second_block_p);
|
|
|
|
return mem_get_blocks_distance (first_block_p, second_block_p);
|
|
}
|
|
} /* mem_get_block_neighbour_field */
|
|
|
|
/**
|
|
* Get block located at specified offset from specified block.
|
|
*
|
|
* @return pointer to block header, located offset bytes after specified block (if dir is next),
|
|
* pointer to block header, located offset bytes before specified block (if dir is prev).
|
|
*/
|
|
static mem_block_header_t*
|
|
mem_get_block_by_offset (const mem_block_header_t* block_p, /**< block */
|
|
mem_heap_offset_t offset, /**< offset */
|
|
mem_direction_t dir) /**< direction of offset */
|
|
{
|
|
const uint8_t* uint8_block_p = (uint8_t*) block_p;
|
|
|
|
if (dir == MEM_DIRECTION_NEXT)
|
|
{
|
|
return (mem_block_header_t*) (uint8_block_p + offset);
|
|
}
|
|
else
|
|
{
|
|
return (mem_block_header_t*) (uint8_block_p - offset);
|
|
}
|
|
} /* mem_get_block_by_offset */
|
|
|
|
/**
|
|
* Get next block in specified direction.
|
|
*
|
|
* @return pointer to next block in direction specified by dir,
|
|
* or NULL - if the block is last in specified direction.
|
|
*/
|
|
static mem_block_header_t*
|
|
mem_get_next_block_by_direction (const mem_block_header_t* block_p, /**< block */
|
|
mem_direction_t dir) /**< direction */
|
|
{
|
|
mem_heap_offset_t offset = block_p->neighbours[dir];
|
|
if (offset != 0)
|
|
{
|
|
return mem_get_block_by_offset (block_p,
|
|
offset,
|
|
dir);
|
|
}
|
|
else
|
|
{
|
|
return NULL;
|
|
}
|
|
} /* mem_get_next_block_by_direction */
|
|
|
|
/**
|
|
* get chunk count, used by the block.
|
|
*
|
|
* @return chunks count
|
|
*/
|
|
static size_t
|
|
mem_get_block_chunks_count (const mem_block_header_t *block_header_p) /**< block header */
|
|
{
|
|
JERRY_ASSERT(block_header_p != NULL);
|
|
|
|
const mem_block_header_t *next_block_p = mem_get_next_block_by_direction (block_header_p, MEM_DIRECTION_NEXT);
|
|
size_t dist_till_block_end;
|
|
|
|
if (next_block_p == NULL)
|
|
{
|
|
dist_till_block_end = (size_t) (mem_heap.heap_start + mem_heap.heap_size - (uint8_t*) block_header_p);
|
|
}
|
|
else
|
|
{
|
|
dist_till_block_end = (size_t) ((uint8_t*) next_block_p - (uint8_t*) block_header_p);
|
|
}
|
|
|
|
JERRY_ASSERT(dist_till_block_end <= mem_heap.heap_size);
|
|
JERRY_ASSERT(dist_till_block_end % MEM_HEAP_CHUNK_SIZE == 0);
|
|
|
|
return dist_till_block_end / MEM_HEAP_CHUNK_SIZE;
|
|
} /* mem_get_block_chunks_count */
|
|
|
|
/**
|
|
* Calculate block's data space size
|
|
*
|
|
* @return size of block area that can be used to store data
|
|
*/
|
|
static size_t
|
|
mem_get_block_data_space_size (const mem_block_header_t *block_header_p) /**< block header */
|
|
{
|
|
return mem_get_block_chunks_count (block_header_p) * MEM_HEAP_CHUNK_SIZE - sizeof (mem_block_header_t);
|
|
} /* mem_get_block_data_space_size */
|
|
|
|
/**
|
|
* Calculate minimum chunks count needed for block with specified size of allocated data area.
|
|
*
|
|
* @return chunks count
|
|
*/
|
|
static size_t
|
|
mem_get_block_chunks_count_from_data_size (size_t block_allocated_size) /**< size of block's allocated area */
|
|
{
|
|
return JERRY_ALIGNUP(sizeof (mem_block_header_t) + block_allocated_size, MEM_HEAP_CHUNK_SIZE) / MEM_HEAP_CHUNK_SIZE;
|
|
} /* mem_get_block_chunks_count_from_data_size */
|
|
|
|
/**
|
|
* Startup initialization of heap
|
|
*/
|
|
void
|
|
mem_heap_init (uint8_t *heap_start, /**< first address of heap space */
|
|
size_t heap_size) /**< heap space size */
|
|
{
|
|
JERRY_ASSERT(heap_start != NULL);
|
|
JERRY_ASSERT(heap_size != 0);
|
|
JERRY_ASSERT(heap_size % MEM_HEAP_CHUNK_SIZE == 0);
|
|
JERRY_ASSERT((uintptr_t) heap_start % MEM_ALIGNMENT == 0);
|
|
JERRY_ASSERT(heap_size <= (1u << MEM_HEAP_OFFSET_LOG));
|
|
|
|
mem_heap.heap_start = heap_start;
|
|
mem_heap.heap_size = heap_size;
|
|
|
|
VALGRIND_NOACCESS_SPACE(heap_start, heap_size);
|
|
|
|
mem_init_block_header (mem_heap.heap_start,
|
|
0,
|
|
MEM_BLOCK_FREE,
|
|
NULL,
|
|
NULL);
|
|
|
|
mem_heap.first_block_p = (mem_block_header_t*) mem_heap.heap_start;
|
|
mem_heap.last_block_p = mem_heap.first_block_p;
|
|
|
|
MEM_HEAP_STAT_INIT ();
|
|
} /* mem_heap_init */
|
|
|
|
/**
|
|
* Finalize heap
|
|
*/
|
|
void
|
|
mem_heap_finalize (void)
|
|
{
|
|
VALGRIND_DEFINED_SPACE(mem_heap.heap_start, mem_heap.heap_size);
|
|
|
|
JERRY_ASSERT(mem_heap.first_block_p == mem_heap.last_block_p);
|
|
JERRY_ASSERT(mem_heap.first_block_p->magic_num == MEM_MAGIC_NUM_OF_FREE_BLOCK);
|
|
|
|
VALGRIND_NOACCESS_SPACE(mem_heap.heap_start, mem_heap.heap_size);
|
|
|
|
memset (&mem_heap, 0, sizeof (mem_heap));
|
|
} /* mem_heap_finalize */
|
|
|
|
/**
|
|
* Initialize block header
|
|
*/
|
|
static void
|
|
mem_init_block_header (uint8_t *first_chunk_p, /**< address of the first chunk to use for the block */
|
|
size_t allocated_bytes, /**< size of block's allocated area */
|
|
mem_block_state_t block_state, /**< state of the block (allocated or free) */
|
|
mem_block_header_t *prev_block_p, /**< previous block */
|
|
mem_block_header_t *next_block_p) /**< next block */
|
|
{
|
|
mem_block_header_t *block_header_p = (mem_block_header_t*) first_chunk_p;
|
|
|
|
VALGRIND_UNDEFINED_STRUCT(block_header_p);
|
|
|
|
if (block_state == MEM_BLOCK_FREE)
|
|
{
|
|
block_header_p->magic_num = MEM_MAGIC_NUM_OF_FREE_BLOCK;
|
|
|
|
JERRY_ASSERT(allocated_bytes == 0);
|
|
}
|
|
else
|
|
{
|
|
block_header_p->magic_num = MEM_MAGIC_NUM_OF_ALLOCATED_BLOCK;
|
|
}
|
|
|
|
block_header_p->neighbours[ MEM_DIRECTION_PREV ] = mem_get_block_neighbour_field (prev_block_p, block_header_p);
|
|
block_header_p->neighbours[ MEM_DIRECTION_NEXT ] = mem_get_block_neighbour_field (block_header_p, next_block_p);
|
|
|
|
JERRY_ASSERT (allocated_bytes == (mem_heap_offset_t) allocated_bytes);
|
|
block_header_p->allocated_bytes = (mem_heap_offset_t) allocated_bytes;
|
|
|
|
JERRY_ASSERT(allocated_bytes <= mem_get_block_data_space_size (block_header_p));
|
|
|
|
VALGRIND_NOACCESS_STRUCT(block_header_p);
|
|
} /* mem_init_block_header */
|
|
|
|
/**
|
|
* Allocation of memory region.
|
|
*
|
|
* See also:
|
|
* mem_heap_alloc_block
|
|
*
|
|
* @return pointer to allocated memory block - if allocation is successful,
|
|
* NULL - if there is not enough memory.
|
|
*/
|
|
static
|
|
void* mem_heap_alloc_block_internal (size_t size_in_bytes, /**< size of region to allocate in bytes */
|
|
mem_heap_alloc_term_t alloc_term) /**< expected allocation term */
|
|
{
|
|
mem_block_header_t *block_p;
|
|
mem_direction_t direction;
|
|
|
|
JERRY_ASSERT (size_in_bytes != 0);
|
|
|
|
mem_check_heap ();
|
|
|
|
if (alloc_term == MEM_HEAP_ALLOC_LONG_TERM)
|
|
{
|
|
block_p = mem_heap.first_block_p;
|
|
direction = MEM_DIRECTION_NEXT;
|
|
}
|
|
else
|
|
{
|
|
JERRY_ASSERT (alloc_term == MEM_HEAP_ALLOC_SHORT_TERM);
|
|
|
|
block_p = mem_heap.last_block_p;
|
|
direction = MEM_DIRECTION_PREV;
|
|
}
|
|
|
|
/* searching for appropriate block */
|
|
while (block_p != NULL)
|
|
{
|
|
VALGRIND_DEFINED_STRUCT(block_p);
|
|
|
|
if (block_p->magic_num == MEM_MAGIC_NUM_OF_FREE_BLOCK)
|
|
{
|
|
if (mem_get_block_data_space_size (block_p) >= size_in_bytes)
|
|
{
|
|
break;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
JERRY_ASSERT(block_p->magic_num == MEM_MAGIC_NUM_OF_ALLOCATED_BLOCK);
|
|
}
|
|
|
|
mem_block_header_t *next_block_p = mem_get_next_block_by_direction (block_p, direction);
|
|
|
|
VALGRIND_NOACCESS_STRUCT(block_p);
|
|
|
|
block_p = next_block_p;
|
|
}
|
|
|
|
if (block_p == NULL)
|
|
{
|
|
/* not enough free space */
|
|
return NULL;
|
|
}
|
|
|
|
/* appropriate block found, allocating space */
|
|
size_t new_block_size_in_chunks = mem_get_block_chunks_count_from_data_size (size_in_bytes);
|
|
size_t found_block_size_in_chunks = mem_get_block_chunks_count (block_p);
|
|
|
|
JERRY_ASSERT(new_block_size_in_chunks <= found_block_size_in_chunks);
|
|
|
|
mem_block_header_t *prev_block_p = mem_get_next_block_by_direction (block_p, MEM_DIRECTION_PREV);
|
|
mem_block_header_t *next_block_p = mem_get_next_block_by_direction (block_p, MEM_DIRECTION_NEXT);
|
|
|
|
if (new_block_size_in_chunks < found_block_size_in_chunks)
|
|
{
|
|
MEM_HEAP_STAT_FREE_BLOCK_SPLIT ();
|
|
|
|
if (direction == MEM_DIRECTION_PREV)
|
|
{
|
|
prev_block_p = block_p;
|
|
uint8_t *block_end_p = (uint8_t*) block_p + found_block_size_in_chunks * MEM_HEAP_CHUNK_SIZE;
|
|
block_p = (mem_block_header_t*) (block_end_p - new_block_size_in_chunks * MEM_HEAP_CHUNK_SIZE);
|
|
|
|
VALGRIND_DEFINED_STRUCT(prev_block_p);
|
|
|
|
prev_block_p->neighbours[ MEM_DIRECTION_NEXT ] = mem_get_block_neighbour_field (prev_block_p,
|
|
block_p);
|
|
|
|
VALGRIND_NOACCESS_STRUCT(prev_block_p);
|
|
|
|
if (next_block_p == NULL)
|
|
{
|
|
mem_heap.last_block_p = block_p;
|
|
}
|
|
else
|
|
{
|
|
VALGRIND_DEFINED_STRUCT(next_block_p);
|
|
|
|
next_block_p->neighbours[ MEM_DIRECTION_PREV ] = mem_get_block_neighbour_field (block_p,
|
|
next_block_p);
|
|
|
|
VALGRIND_NOACCESS_STRUCT(next_block_p);
|
|
}
|
|
}
|
|
else
|
|
{
|
|
uint8_t *new_free_block_first_chunk_p = (uint8_t*) block_p + new_block_size_in_chunks * MEM_HEAP_CHUNK_SIZE;
|
|
mem_init_block_header (new_free_block_first_chunk_p,
|
|
0,
|
|
MEM_BLOCK_FREE,
|
|
block_p,
|
|
next_block_p);
|
|
|
|
mem_block_header_t *new_free_block_p = (mem_block_header_t*) new_free_block_first_chunk_p;
|
|
|
|
if (next_block_p == NULL)
|
|
{
|
|
mem_heap.last_block_p = new_free_block_p;
|
|
}
|
|
else
|
|
{
|
|
VALGRIND_DEFINED_STRUCT(next_block_p);
|
|
|
|
const mem_block_header_t* new_free_block_p = (mem_block_header_t*) new_free_block_first_chunk_p;
|
|
next_block_p->neighbours[ MEM_DIRECTION_PREV ] = mem_get_block_neighbour_field (new_free_block_p,
|
|
next_block_p);
|
|
|
|
VALGRIND_NOACCESS_STRUCT(next_block_p);
|
|
}
|
|
|
|
next_block_p = new_free_block_p;
|
|
}
|
|
}
|
|
|
|
mem_init_block_header ((uint8_t*) block_p,
|
|
size_in_bytes,
|
|
MEM_BLOCK_ALLOCATED,
|
|
prev_block_p,
|
|
next_block_p);
|
|
|
|
VALGRIND_DEFINED_STRUCT(block_p);
|
|
|
|
MEM_HEAP_STAT_ALLOC_BLOCK (block_p);
|
|
|
|
JERRY_ASSERT(mem_get_block_data_space_size (block_p) >= size_in_bytes);
|
|
|
|
VALGRIND_NOACCESS_STRUCT(block_p);
|
|
|
|
/* return data space beginning address */
|
|
uint8_t *data_space_p = (uint8_t*) (block_p + 1);
|
|
JERRY_ASSERT((uintptr_t) data_space_p % MEM_ALIGNMENT == 0);
|
|
|
|
VALGRIND_UNDEFINED_SPACE(data_space_p, size_in_bytes);
|
|
|
|
mem_check_heap ();
|
|
|
|
return data_space_p;
|
|
} /* mem_heap_alloc_block_internal */
|
|
|
|
/**
|
|
* Allocation of memory region.
|
|
*
|
|
* To reduce heap fragmentation there are two allocation modes - short-term and long-term.
|
|
*
|
|
* If allocation is short-term then the beginning of the heap is preferred, else - the end of the heap.
|
|
*
|
|
* It is supposed, that all short-term allocation is used during relatively short discrete sessions.
|
|
* After end of the session all short-term allocated regions are supposed to be freed.
|
|
*
|
|
* @return pointer to allocated memory block - if allocation is successful,
|
|
* NULL - if requested region size is zero or if there is not enough memory.
|
|
*/
|
|
void*
|
|
mem_heap_alloc_block (size_t size_in_bytes, /**< size of region to allocate in bytes */
|
|
mem_heap_alloc_term_t alloc_term) /**< expected allocation term */
|
|
{
|
|
if (unlikely (size_in_bytes == 0))
|
|
{
|
|
return NULL;
|
|
}
|
|
else
|
|
{
|
|
void *data_space_p = mem_heap_alloc_block_internal (size_in_bytes, alloc_term);
|
|
|
|
if (likely (data_space_p != NULL))
|
|
{
|
|
return data_space_p;
|
|
}
|
|
|
|
for (mem_try_give_memory_back_severity_t severity = MEM_TRY_GIVE_MEMORY_BACK_SEVERITY_LOW;
|
|
severity <= MEM_TRY_GIVE_MEMORY_BACK_SEVERITY_CRITICAL;
|
|
severity = (mem_try_give_memory_back_severity_t) (severity + 1))
|
|
{
|
|
mem_run_try_to_give_memory_back_callbacks (severity);
|
|
|
|
data_space_p = mem_heap_alloc_block_internal (size_in_bytes, alloc_term);
|
|
|
|
if (data_space_p != NULL)
|
|
{
|
|
return data_space_p;
|
|
}
|
|
}
|
|
|
|
JERRY_ASSERT (data_space_p == NULL);
|
|
|
|
jerry_fatal (ERR_OUT_OF_MEMORY);
|
|
}
|
|
} /* mem_heap_alloc_block */
|
|
|
|
/**
|
|
* Try to resize memory region.
|
|
*
|
|
* @return true - if resize is successful,
|
|
* false - if there is not enough memory in front of the block.
|
|
*/
|
|
bool
|
|
mem_heap_try_resize_block (void *ptr, /**< pointer to beginning of data space of the block to resize */
|
|
size_t size_in_bytes) /**< new block size */
|
|
{
|
|
uint8_t *uint8_ptr = (uint8_t*) ptr;
|
|
|
|
/* checking that uint8_ptr points to the heap */
|
|
JERRY_ASSERT(uint8_ptr >= mem_heap.heap_start
|
|
&& uint8_ptr <= mem_heap.heap_start + mem_heap.heap_size);
|
|
|
|
mem_check_heap ();
|
|
|
|
mem_block_header_t *block_p = (mem_block_header_t*) uint8_ptr - 1;
|
|
|
|
VALGRIND_DEFINED_STRUCT(block_p);
|
|
|
|
JERRY_ASSERT(block_p->magic_num == MEM_MAGIC_NUM_OF_ALLOCATED_BLOCK);
|
|
|
|
/* For heap statistics unit we show what is going on as though
|
|
* the block is freed and then new block (the same or resized)
|
|
* is allocated */
|
|
MEM_HEAP_STAT_FREE_BLOCK (block_p);
|
|
|
|
size_t current_block_may_expand_up_to = mem_get_block_data_space_size (block_p);
|
|
|
|
bool is_resized = false;
|
|
|
|
if (current_block_may_expand_up_to >= size_in_bytes)
|
|
{
|
|
is_resized = true;
|
|
}
|
|
else
|
|
{
|
|
size_t need_additional_bytes = size_in_bytes - current_block_may_expand_up_to;
|
|
|
|
mem_block_header_t *next_block_p = mem_get_next_block_by_direction (block_p, MEM_DIRECTION_NEXT);
|
|
|
|
if (next_block_p != NULL)
|
|
{
|
|
VALGRIND_DEFINED_STRUCT (next_block_p);
|
|
|
|
if (next_block_p->magic_num == MEM_MAGIC_NUM_OF_FREE_BLOCK)
|
|
{
|
|
size_t next_block_data_space_size = mem_get_block_data_space_size (next_block_p);
|
|
|
|
if (next_block_data_space_size >= need_additional_bytes)
|
|
{
|
|
/* next block is free and contains enough space */
|
|
|
|
is_resized = true;
|
|
|
|
size_t new_block_chunks_count = mem_get_block_chunks_count_from_data_size (size_in_bytes);
|
|
size_t current_block_chunks_count = mem_get_block_chunks_count (block_p);
|
|
size_t next_block_chunks_count = mem_get_block_chunks_count (next_block_p);
|
|
|
|
JERRY_ASSERT (new_block_chunks_count <= current_block_chunks_count + next_block_chunks_count);
|
|
|
|
size_t diff_in_chunks = (size_t) ((current_block_chunks_count +
|
|
next_block_chunks_count) - new_block_chunks_count);
|
|
|
|
mem_block_header_t *block_after_next_p = mem_get_next_block_by_direction (next_block_p,
|
|
MEM_DIRECTION_NEXT);
|
|
mem_block_header_t *new_next_of_current_block_p;
|
|
mem_block_header_t *new_prev_of_block_after_next_p;
|
|
|
|
if (diff_in_chunks > 0)
|
|
{
|
|
mem_block_header_t *new_free_block_p = (mem_block_header_t*) ((uint8_t*) block_p +
|
|
new_block_chunks_count * MEM_HEAP_CHUNK_SIZE);
|
|
|
|
mem_init_block_header ((uint8_t*) new_free_block_p,
|
|
0,
|
|
MEM_BLOCK_FREE,
|
|
block_p,
|
|
block_after_next_p);
|
|
|
|
new_prev_of_block_after_next_p = new_free_block_p;
|
|
new_next_of_current_block_p = new_free_block_p;
|
|
}
|
|
else
|
|
{
|
|
new_prev_of_block_after_next_p = block_p;
|
|
new_next_of_current_block_p = block_after_next_p;
|
|
}
|
|
|
|
block_p->neighbours[ MEM_DIRECTION_NEXT ] = mem_get_block_neighbour_field (block_p,
|
|
new_next_of_current_block_p);
|
|
if (block_after_next_p != NULL)
|
|
{
|
|
VALGRIND_DEFINED_STRUCT (block_after_next_p);
|
|
|
|
mem_heap_offset_t offset = mem_get_block_neighbour_field (new_prev_of_block_after_next_p,
|
|
block_after_next_p);
|
|
block_after_next_p->neighbours[ MEM_DIRECTION_PREV ] = offset;
|
|
|
|
VALGRIND_NOACCESS_STRUCT (block_after_next_p);
|
|
}
|
|
else
|
|
{
|
|
mem_heap.last_block_p = new_prev_of_block_after_next_p;
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
JERRY_ASSERT (next_block_p->magic_num == MEM_MAGIC_NUM_OF_ALLOCATED_BLOCK);
|
|
}
|
|
|
|
VALGRIND_NOACCESS_STRUCT (next_block_p);
|
|
}
|
|
}
|
|
|
|
if (is_resized)
|
|
{
|
|
JERRY_ASSERT ((mem_heap_offset_t) size_in_bytes == size_in_bytes);
|
|
|
|
if (size_in_bytes >= block_p->allocated_bytes)
|
|
{
|
|
VALGRIND_UNDEFINED_SPACE (uint8_ptr + block_p->allocated_bytes, size_in_bytes - block_p->allocated_bytes);
|
|
}
|
|
|
|
block_p->allocated_bytes = (mem_heap_offset_t) size_in_bytes;
|
|
}
|
|
|
|
MEM_HEAP_STAT_ALLOC_BLOCK (block_p);
|
|
|
|
VALGRIND_NOACCESS_STRUCT(block_p);
|
|
|
|
mem_check_heap ();
|
|
|
|
return is_resized;
|
|
} /* mem_heap_try_resize_block */
|
|
|
|
/**
|
|
* Free the memory block.
|
|
*/
|
|
void
|
|
mem_heap_free_block (void *ptr) /**< pointer to beginning of data space of the block */
|
|
{
|
|
uint8_t *uint8_ptr = (uint8_t*) ptr;
|
|
|
|
/* checking that uint8_ptr points to the heap */
|
|
JERRY_ASSERT(uint8_ptr >= mem_heap.heap_start
|
|
&& uint8_ptr <= mem_heap.heap_start + mem_heap.heap_size);
|
|
|
|
mem_check_heap ();
|
|
|
|
mem_block_header_t *block_p = (mem_block_header_t*) uint8_ptr - 1;
|
|
|
|
VALGRIND_DEFINED_STRUCT(block_p);
|
|
|
|
mem_block_header_t *prev_block_p = mem_get_next_block_by_direction (block_p, MEM_DIRECTION_PREV);
|
|
mem_block_header_t *next_block_p = mem_get_next_block_by_direction (block_p, MEM_DIRECTION_NEXT);
|
|
|
|
MEM_HEAP_STAT_FREE_BLOCK (block_p);
|
|
|
|
VALGRIND_NOACCESS_SPACE(uint8_ptr, block_p->allocated_bytes);
|
|
|
|
/* checking magic nums that are neighbour to data space */
|
|
JERRY_ASSERT(block_p->magic_num == MEM_MAGIC_NUM_OF_ALLOCATED_BLOCK);
|
|
if (next_block_p != NULL)
|
|
{
|
|
VALGRIND_DEFINED_STRUCT(next_block_p);
|
|
|
|
JERRY_ASSERT(next_block_p->magic_num == MEM_MAGIC_NUM_OF_ALLOCATED_BLOCK
|
|
|| next_block_p->magic_num == MEM_MAGIC_NUM_OF_FREE_BLOCK);
|
|
|
|
VALGRIND_NOACCESS_STRUCT(next_block_p);
|
|
}
|
|
|
|
block_p->magic_num = MEM_MAGIC_NUM_OF_FREE_BLOCK;
|
|
|
|
if (next_block_p != NULL)
|
|
{
|
|
VALGRIND_DEFINED_STRUCT(next_block_p);
|
|
|
|
if (next_block_p->magic_num == MEM_MAGIC_NUM_OF_FREE_BLOCK)
|
|
{
|
|
/* merge with the next block */
|
|
MEM_HEAP_STAT_FREE_BLOCK_MERGE ();
|
|
|
|
mem_block_header_t *next_next_block_p = mem_get_next_block_by_direction (next_block_p, MEM_DIRECTION_NEXT);
|
|
|
|
VALGRIND_NOACCESS_STRUCT(next_block_p);
|
|
|
|
next_block_p = next_next_block_p;
|
|
|
|
VALGRIND_DEFINED_STRUCT(next_block_p);
|
|
|
|
block_p->neighbours[ MEM_DIRECTION_NEXT ] = mem_get_block_neighbour_field (block_p, next_block_p);
|
|
if (next_block_p != NULL)
|
|
{
|
|
next_block_p->neighbours[ MEM_DIRECTION_PREV ] = mem_get_block_neighbour_field (block_p, next_block_p);
|
|
}
|
|
else
|
|
{
|
|
mem_heap.last_block_p = block_p;
|
|
}
|
|
}
|
|
|
|
VALGRIND_NOACCESS_STRUCT(next_block_p);
|
|
}
|
|
|
|
if (prev_block_p != NULL)
|
|
{
|
|
VALGRIND_DEFINED_STRUCT(prev_block_p);
|
|
|
|
if (prev_block_p->magic_num == MEM_MAGIC_NUM_OF_FREE_BLOCK)
|
|
{
|
|
/* merge with the previous block */
|
|
MEM_HEAP_STAT_FREE_BLOCK_MERGE ();
|
|
|
|
prev_block_p->neighbours[ MEM_DIRECTION_NEXT ] = mem_get_block_neighbour_field (prev_block_p, next_block_p);
|
|
if (next_block_p != NULL)
|
|
{
|
|
VALGRIND_DEFINED_STRUCT(next_block_p);
|
|
|
|
const mem_block_header_t* prev_block_p = mem_get_next_block_by_direction (block_p, MEM_DIRECTION_PREV);
|
|
next_block_p->neighbours[ MEM_DIRECTION_PREV ] = mem_get_block_neighbour_field (prev_block_p, next_block_p);
|
|
|
|
VALGRIND_NOACCESS_STRUCT(next_block_p);
|
|
}
|
|
else
|
|
{
|
|
mem_heap.last_block_p = prev_block_p;
|
|
}
|
|
}
|
|
|
|
VALGRIND_NOACCESS_STRUCT(prev_block_p);
|
|
}
|
|
|
|
VALGRIND_NOACCESS_STRUCT(block_p);
|
|
|
|
mem_check_heap ();
|
|
} /* mem_heap_free_block */
|
|
|
|
/**
|
|
* Recommend allocation size based on chunk size.
|
|
*
|
|
* @return recommended allocation size
|
|
*/
|
|
size_t __attr_pure___
|
|
mem_heap_recommend_allocation_size (size_t minimum_allocation_size) /**< minimum allocation size */
|
|
{
|
|
size_t minimum_allocation_size_with_block_header = minimum_allocation_size + sizeof (mem_block_header_t);
|
|
size_t heap_chunk_aligned_allocation_size = JERRY_ALIGNUP(minimum_allocation_size_with_block_header,
|
|
MEM_HEAP_CHUNK_SIZE);
|
|
|
|
return heap_chunk_aligned_allocation_size - sizeof (mem_block_header_t);
|
|
} /* mem_heap_recommend_allocation_size */
|
|
|
|
/**
|
|
* Print heap
|
|
*/
|
|
void
|
|
mem_heap_print (bool dump_block_headers, /**< print block headers */
|
|
bool dump_block_data, /**< print block with data (true)
|
|
or print only block header (false) */
|
|
bool dump_stats) /**< print heap stats */
|
|
{
|
|
mem_check_heap ();
|
|
|
|
JERRY_ASSERT(!dump_block_data || dump_block_headers);
|
|
|
|
if (dump_block_headers)
|
|
{
|
|
printf ("Heap: start=%p size=%lu, first block->%p, last block->%p\n",
|
|
mem_heap.heap_start,
|
|
(unsigned long) mem_heap.heap_size,
|
|
(void*) mem_heap.first_block_p,
|
|
(void*) mem_heap.last_block_p);
|
|
|
|
for (mem_block_header_t *block_p = mem_heap.first_block_p, *next_block_p;
|
|
block_p != NULL;
|
|
block_p = next_block_p)
|
|
{
|
|
VALGRIND_DEFINED_STRUCT(block_p);
|
|
|
|
printf ("Block (%p): magic num=0x%08x, size in chunks=%lu, previous block->%p next block->%p\n",
|
|
(void*) block_p,
|
|
block_p->magic_num,
|
|
(unsigned long) mem_get_block_chunks_count (block_p),
|
|
(void*) mem_get_next_block_by_direction (block_p, MEM_DIRECTION_PREV),
|
|
(void*) mem_get_next_block_by_direction (block_p, MEM_DIRECTION_NEXT));
|
|
|
|
if (dump_block_data)
|
|
{
|
|
uint8_t *block_data_p = (uint8_t*) (block_p + 1);
|
|
for (uint32_t offset = 0;
|
|
offset < mem_get_block_data_space_size (block_p);
|
|
offset++)
|
|
{
|
|
printf ("%02x ", block_data_p[ offset ]);
|
|
}
|
|
printf ("\n");
|
|
}
|
|
|
|
next_block_p = mem_get_next_block_by_direction (block_p, MEM_DIRECTION_NEXT);
|
|
|
|
VALGRIND_NOACCESS_STRUCT(block_p);
|
|
}
|
|
}
|
|
|
|
#ifdef MEM_STATS
|
|
if (dump_stats)
|
|
{
|
|
printf ("Heap stats:\n");
|
|
printf (" Heap size = %lu bytes\n"
|
|
" Chunk size = %lu bytes\n"
|
|
" Blocks count = %lu\n"
|
|
" Allocated blocks count = %lu\n"
|
|
" Allocated chunks count = %lu\n"
|
|
" Allocated = %lu bytes\n"
|
|
" Waste = %lu bytes\n"
|
|
" Peak allocated blocks count = %lu\n"
|
|
" Peak allocated chunks count = %lu\n"
|
|
" Peak allocated= %lu bytes\n"
|
|
" Peak waste = %lu bytes\n",
|
|
mem_heap_stats.size,
|
|
(uint64_t) MEM_HEAP_CHUNK_SIZE,
|
|
mem_heap_stats.blocks,
|
|
mem_heap_stats.allocated_blocks,
|
|
mem_heap_stats.allocated_chunks,
|
|
mem_heap_stats.allocated_bytes,
|
|
mem_heap_stats.waste_bytes,
|
|
mem_heap_stats.peak_allocated_blocks,
|
|
mem_heap_stats.peak_allocated_chunks,
|
|
mem_heap_stats.peak_allocated_bytes,
|
|
mem_heap_stats.peak_waste_bytes);
|
|
}
|
|
#else /* MEM_STATS */
|
|
(void) dump_stats;
|
|
#endif /* !MEM_STATS */
|
|
|
|
printf ("\n");
|
|
} /* mem_heap_print */
|
|
|
|
/**
|
|
* Check heap consistency
|
|
*/
|
|
static void
|
|
mem_check_heap (void)
|
|
{
|
|
#ifndef JERRY_NDEBUG
|
|
JERRY_ASSERT((uint8_t*) mem_heap.first_block_p == mem_heap.heap_start);
|
|
JERRY_ASSERT(mem_heap.heap_size % MEM_HEAP_CHUNK_SIZE == 0);
|
|
|
|
bool is_last_block_was_met = false;
|
|
size_t chunk_sizes_sum = 0;
|
|
|
|
for (mem_block_header_t *block_p = mem_heap.first_block_p, *next_block_p;
|
|
block_p != NULL;
|
|
block_p = next_block_p)
|
|
{
|
|
VALGRIND_DEFINED_STRUCT(block_p);
|
|
|
|
JERRY_ASSERT(block_p->magic_num == MEM_MAGIC_NUM_OF_FREE_BLOCK
|
|
|| block_p->magic_num == MEM_MAGIC_NUM_OF_ALLOCATED_BLOCK);
|
|
chunk_sizes_sum += mem_get_block_chunks_count (block_p);
|
|
|
|
next_block_p = mem_get_next_block_by_direction (block_p, MEM_DIRECTION_NEXT);
|
|
|
|
if (block_p == mem_heap.last_block_p)
|
|
{
|
|
is_last_block_was_met = true;
|
|
|
|
JERRY_ASSERT(next_block_p == NULL);
|
|
}
|
|
else
|
|
{
|
|
JERRY_ASSERT(next_block_p != NULL);
|
|
}
|
|
|
|
VALGRIND_NOACCESS_STRUCT(block_p);
|
|
}
|
|
|
|
JERRY_ASSERT(chunk_sizes_sum * MEM_HEAP_CHUNK_SIZE == mem_heap.heap_size);
|
|
JERRY_ASSERT(is_last_block_was_met);
|
|
|
|
bool is_first_block_was_met = false;
|
|
chunk_sizes_sum = 0;
|
|
|
|
for (mem_block_header_t *block_p = mem_heap.last_block_p, *prev_block_p;
|
|
block_p != NULL;
|
|
block_p = prev_block_p)
|
|
{
|
|
VALGRIND_DEFINED_STRUCT(block_p);
|
|
|
|
JERRY_ASSERT(block_p->magic_num == MEM_MAGIC_NUM_OF_FREE_BLOCK
|
|
|| block_p->magic_num == MEM_MAGIC_NUM_OF_ALLOCATED_BLOCK);
|
|
chunk_sizes_sum += mem_get_block_chunks_count (block_p);
|
|
|
|
prev_block_p = mem_get_next_block_by_direction (block_p, MEM_DIRECTION_PREV);
|
|
|
|
if (block_p == mem_heap.first_block_p)
|
|
{
|
|
is_first_block_was_met = true;
|
|
|
|
JERRY_ASSERT(prev_block_p == NULL);
|
|
}
|
|
else
|
|
{
|
|
JERRY_ASSERT(prev_block_p != NULL);
|
|
}
|
|
|
|
VALGRIND_NOACCESS_STRUCT(block_p);
|
|
}
|
|
|
|
JERRY_ASSERT(chunk_sizes_sum * MEM_HEAP_CHUNK_SIZE == mem_heap.heap_size);
|
|
JERRY_ASSERT(is_first_block_was_met);
|
|
#endif /* !JERRY_NDEBUG */
|
|
} /* mem_check_heap */
|
|
|
|
#ifdef MEM_STATS
|
|
/**
|
|
* Get heap memory usage statistics
|
|
*/
|
|
void
|
|
mem_heap_get_stats (mem_heap_stats_t *out_heap_stats_p) /**< out: heap stats */
|
|
{
|
|
*out_heap_stats_p = mem_heap_stats;
|
|
} /* mem_heap_get_stats */
|
|
|
|
/**
|
|
* Reset peak values in memory usage statistics
|
|
*/
|
|
void
|
|
mem_heap_stats_reset_peak (void)
|
|
{
|
|
mem_heap_stats.peak_allocated_chunks = mem_heap_stats.allocated_chunks;
|
|
mem_heap_stats.peak_allocated_blocks = mem_heap_stats.allocated_blocks;
|
|
mem_heap_stats.peak_allocated_bytes = mem_heap_stats.allocated_bytes;
|
|
mem_heap_stats.peak_waste_bytes = mem_heap_stats.waste_bytes;
|
|
} /* mem_heap_stats_reset_peak */
|
|
|
|
/**
|
|
* Initalize heap memory usage statistics account structure
|
|
*/
|
|
static void
|
|
mem_heap_stat_init ()
|
|
{
|
|
memset (&mem_heap_stats, 0, sizeof (mem_heap_stats));
|
|
|
|
mem_heap_stats.size = mem_heap.heap_size;
|
|
mem_heap_stats.blocks = 1;
|
|
} /* mem_heap_stat_init */
|
|
|
|
/**
|
|
* Account block allocation
|
|
*/
|
|
static void
|
|
mem_heap_stat_alloc_block (mem_block_header_t *block_header_p) /**< allocated block */
|
|
{
|
|
JERRY_ASSERT(block_header_p->magic_num == MEM_MAGIC_NUM_OF_ALLOCATED_BLOCK);
|
|
|
|
const size_t chunks = mem_get_block_chunks_count (block_header_p);
|
|
const size_t bytes = block_header_p->allocated_bytes;
|
|
const size_t waste_bytes = chunks * MEM_HEAP_CHUNK_SIZE - bytes;
|
|
|
|
mem_heap_stats.allocated_blocks++;
|
|
mem_heap_stats.allocated_chunks += chunks;
|
|
mem_heap_stats.allocated_bytes += bytes;
|
|
mem_heap_stats.waste_bytes += waste_bytes;
|
|
|
|
if (mem_heap_stats.allocated_blocks > mem_heap_stats.peak_allocated_blocks)
|
|
{
|
|
mem_heap_stats.peak_allocated_blocks = mem_heap_stats.allocated_blocks;
|
|
}
|
|
if (mem_heap_stats.allocated_blocks > mem_heap_stats.global_peak_allocated_blocks)
|
|
{
|
|
mem_heap_stats.global_peak_allocated_blocks = mem_heap_stats.allocated_blocks;
|
|
}
|
|
|
|
if (mem_heap_stats.allocated_chunks > mem_heap_stats.peak_allocated_chunks)
|
|
{
|
|
mem_heap_stats.peak_allocated_chunks = mem_heap_stats.allocated_chunks;
|
|
}
|
|
if (mem_heap_stats.allocated_chunks > mem_heap_stats.global_peak_allocated_chunks)
|
|
{
|
|
mem_heap_stats.global_peak_allocated_chunks = mem_heap_stats.allocated_chunks;
|
|
}
|
|
|
|
if (mem_heap_stats.allocated_bytes > mem_heap_stats.peak_allocated_bytes)
|
|
{
|
|
mem_heap_stats.peak_allocated_bytes = mem_heap_stats.allocated_bytes;
|
|
}
|
|
if (mem_heap_stats.allocated_bytes > mem_heap_stats.global_peak_allocated_bytes)
|
|
{
|
|
mem_heap_stats.global_peak_allocated_bytes = mem_heap_stats.allocated_bytes;
|
|
}
|
|
|
|
if (mem_heap_stats.waste_bytes > mem_heap_stats.peak_waste_bytes)
|
|
{
|
|
mem_heap_stats.peak_waste_bytes = mem_heap_stats.waste_bytes;
|
|
}
|
|
if (mem_heap_stats.waste_bytes > mem_heap_stats.global_peak_waste_bytes)
|
|
{
|
|
mem_heap_stats.global_peak_waste_bytes = mem_heap_stats.waste_bytes;
|
|
}
|
|
|
|
JERRY_ASSERT(mem_heap_stats.allocated_blocks <= mem_heap_stats.blocks);
|
|
JERRY_ASSERT(mem_heap_stats.allocated_bytes <= mem_heap_stats.size);
|
|
JERRY_ASSERT(mem_heap_stats.allocated_chunks <= mem_heap_stats.size / MEM_HEAP_CHUNK_SIZE);
|
|
} /* mem_heap_stat_alloc_block */
|
|
|
|
/**
|
|
* Account block freeing
|
|
*/
|
|
static void
|
|
mem_heap_stat_free_block (mem_block_header_t *block_header_p) /**< block to be freed */
|
|
{
|
|
JERRY_ASSERT(block_header_p->magic_num == MEM_MAGIC_NUM_OF_ALLOCATED_BLOCK);
|
|
|
|
const size_t chunks = mem_get_block_chunks_count (block_header_p);
|
|
const size_t bytes = block_header_p->allocated_bytes;
|
|
const size_t waste_bytes = chunks * MEM_HEAP_CHUNK_SIZE - bytes;
|
|
|
|
JERRY_ASSERT(mem_heap_stats.allocated_blocks <= mem_heap_stats.blocks);
|
|
JERRY_ASSERT(mem_heap_stats.allocated_bytes <= mem_heap_stats.size);
|
|
JERRY_ASSERT(mem_heap_stats.allocated_chunks <= mem_heap_stats.size / MEM_HEAP_CHUNK_SIZE);
|
|
|
|
JERRY_ASSERT(mem_heap_stats.allocated_blocks >= 1);
|
|
JERRY_ASSERT(mem_heap_stats.allocated_chunks >= chunks);
|
|
JERRY_ASSERT(mem_heap_stats.allocated_bytes >= bytes);
|
|
JERRY_ASSERT(mem_heap_stats.waste_bytes >= waste_bytes);
|
|
|
|
mem_heap_stats.allocated_blocks--;
|
|
mem_heap_stats.allocated_chunks -= chunks;
|
|
mem_heap_stats.allocated_bytes -= bytes;
|
|
mem_heap_stats.waste_bytes -= waste_bytes;
|
|
} /* mem_heap_stat_free_block */
|
|
|
|
/**
|
|
* Account free block split
|
|
*/
|
|
static void
|
|
mem_heap_stat_free_block_split (void)
|
|
{
|
|
mem_heap_stats.blocks++;
|
|
} /* mem_heap_stat_free_block_split */
|
|
|
|
/**
|
|
* Account free block merge
|
|
*/
|
|
static void
|
|
mem_heap_stat_free_block_merge (void)
|
|
{
|
|
mem_heap_stats.blocks--;
|
|
} /* mem_heap_stat_free_block_merge */
|
|
#endif /* MEM_STATS */
|
|
|
|
/**
|
|
* @}
|
|
* @}
|
|
*/
|