1
0
Fork 0
mirror of https://github.com/git/git.git synced 2024-05-04 18:16:11 +02:00
git/mem-pool.c

146 lines
3.1 KiB
C
Raw Normal View History

/*
* Memory Pool implementation logic.
*/
#include "cache.h"
#include "mem-pool.h"
#define BLOCK_GROWTH_SIZE 1024*1024 - sizeof(struct mp_block);
/*
* Allocate a new mp_block and insert it after the block specified in
* `insert_after`. If `insert_after` is NULL, then insert block at the
* head of the linked list.
*/
static struct mp_block *mem_pool_alloc_block(struct mem_pool *mem_pool, size_t block_alloc, struct mp_block *insert_after)
{
struct mp_block *p;
mem_pool->pool_alloc += sizeof(struct mp_block) + block_alloc;
p = xmalloc(st_add(sizeof(struct mp_block), block_alloc));
p->next_free = (char *)p->space;
p->end = p->next_free + block_alloc;
if (insert_after) {
p->next_block = insert_after->next_block;
insert_after->next_block = p;
} else {
p->next_block = mem_pool->mp_block;
mem_pool->mp_block = p;
}
return p;
}
void mem_pool_init(struct mem_pool **mem_pool, size_t initial_size)
{
struct mem_pool *pool;
if (*mem_pool)
return;
pool = xcalloc(1, sizeof(*pool));
pool->block_alloc = BLOCK_GROWTH_SIZE;
if (initial_size > 0)
mem_pool_alloc_block(pool, initial_size, NULL);
*mem_pool = pool;
}
void mem_pool_discard(struct mem_pool *mem_pool, int invalidate_memory)
{
struct mp_block *block, *block_to_free;
block alloc: allocate cache entries from mem_pool When reading large indexes from disk, a portion of the time is dominated in malloc() calls. This can be mitigated by allocating a large block of memory and manage it ourselves via memory pools. This change moves the cache entry allocation to be on top of memory pools. Design: The index_state struct will gain a notion of an associated memory_pool from which cache_entries will be allocated from. When reading in the index from disk, we have information on the number of entries and their size, which can guide us in deciding how large our initial memory allocation should be. When an index is discarded, the associated memory_pool will be discarded as well - so the lifetime of a cache_entry is tied to the lifetime of the index_state that it was allocated for. In the case of a Split Index, the following rules are followed. 1st, some terminology is defined: Terminology: - 'the_index': represents the logical view of the index - 'split_index': represents the "base" cache entries. Read from the split index file. 'the_index' can reference a single split_index, as well as cache_entries from the split_index. `the_index` will be discarded before the `split_index` is. This means that when we are allocating cache_entries in the presence of a split index, we need to allocate the entries from the `split_index`'s memory pool. This allows us to follow the pattern that `the_index` can reference cache_entries from the `split_index`, and that the cache_entries will not be freed while they are still being referenced. Managing transient cache_entry structs: Cache entries are usually allocated for an index, but this is not always the case. Cache entries are sometimes allocated because this is the type that the existing checkout_entry function works with. Because of this, the existing code needs to handle cache entries associated with an index / memory pool, and those that only exist transiently. Several strategies were contemplated around how to handle this: Chosen approach: An extra field was added to the cache_entry type to track whether the cache_entry was allocated from a memory pool or not. This is currently an int field, as there are no more available bits in the existing ce_flags bit field. If / when more bits are needed, this new field can be turned into a proper bit field. Alternatives: 1) Do not include any information about how the cache_entry was allocated. Calling code would be responsible for tracking whether the cache_entry needed to be freed or not. Pro: No extra memory overhead to track this state Con: Extra complexity in callers to handle this correctly. The extra complexity and burden to not regress this behavior in the future was more than we wanted. 2) cache_entry would gain knowledge about which mem_pool allocated it Pro: Could (potentially) do extra logic to know when a mem_pool no longer had references to any cache_entry Con: cache_entry would grow heavier by a pointer, instead of int We didn't see a tangible benefit to this approach 3) Do not add any extra information to a cache_entry, but when freeing a cache entry, check if the memory exists in a region managed by existing mem_pools. Pro: No extra memory overhead to track state Con: Extra computation is performed when freeing cache entries We decided tracking and iterating over known memory pool regions was less desirable than adding an extra field to track this stae. Signed-off-by: Jameson Miller <jamill@microsoft.com> Signed-off-by: Junio C Hamano <gitster@pobox.com>
2018-07-02 21:49:37 +02:00
block = mem_pool->mp_block;
while (block)
{
block_to_free = block;
block = block->next_block;
if (invalidate_memory)
memset(block_to_free->space, 0xDD, ((char *)block_to_free->end) - ((char *)block_to_free->space));
free(block_to_free);
}
free(mem_pool);
}
void *mem_pool_alloc(struct mem_pool *mem_pool, size_t len)
{
struct mp_block *p = NULL;
void *r;
/* round up to a 'uintmax_t' alignment */
if (len & (sizeof(uintmax_t) - 1))
len += sizeof(uintmax_t) - (len & (sizeof(uintmax_t) - 1));
if (mem_pool->mp_block &&
mem_pool->mp_block->end - mem_pool->mp_block->next_free >= len)
p = mem_pool->mp_block;
if (!p) {
if (len >= (mem_pool->block_alloc / 2))
return mem_pool_alloc_block(mem_pool, len, mem_pool->mp_block);
p = mem_pool_alloc_block(mem_pool, mem_pool->block_alloc, NULL);
}
r = p->next_free;
p->next_free += len;
return r;
}
void *mem_pool_calloc(struct mem_pool *mem_pool, size_t count, size_t size)
{
size_t len = st_mult(count, size);
void *r = mem_pool_alloc(mem_pool, len);
memset(r, 0, len);
return r;
}
int mem_pool_contains(struct mem_pool *mem_pool, void *mem)
{
struct mp_block *p;
/* Check if memory is allocated in a block */
for (p = mem_pool->mp_block; p; p = p->next_block)
if ((mem >= ((void *)p->space)) &&
(mem < ((void *)p->end)))
return 1;
return 0;
}
void mem_pool_combine(struct mem_pool *dst, struct mem_pool *src)
{
struct mp_block *p;
/* Append the blocks from src to dst */
if (dst->mp_block && src->mp_block) {
/*
* src and dst have blocks, append
* blocks from src to dst.
*/
p = dst->mp_block;
while (p->next_block)
p = p->next_block;
p->next_block = src->mp_block;
} else if (src->mp_block) {
/*
* src has blocks, dst is empty.
*/
dst->mp_block = src->mp_block;
} else {
/* src is empty, nothing to do. */
}
dst->pool_alloc += src->pool_alloc;
src->pool_alloc = 0;
src->mp_block = NULL;
}