blob: f9f5ef68f3a8dfac7caece3442bc2a6db0f6c9c1 [file] [log] [blame]
/*
* Copyright 2010 Marek Olšák <maraeo@gmail.com>
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* on the rights to use, copy, modify, merge, publish, distribute, sub
* license, and/or sell copies of the Software, and to permit persons to whom
* the Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice (including the next
* paragraph) shall be included in all copies or substantial portions of the
* Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
* THE AUTHOR(S) AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM,
* DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
* OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
* USE OR OTHER DEALINGS IN THE SOFTWARE. */
#include "util/u_slab.h"
#include "util/u_math.h"
#include "util/u_memory.h"
#include "util/u_simple_list.h"
#include <stdio.h>
#define UTIL_SLAB_MAGIC 0xcafe4321
/* The block is either allocated memory or free space. */
struct util_slab_block {
/* The header. */
/* The first next free block. */
struct util_slab_block *next_free;
intptr_t magic;
/* Memory after the last member is dedicated to the block itself.
* The allocated size is always larger than this structure. */
};
static struct util_slab_block *
util_slab_get_block(struct util_slab_mempool *pool,
struct util_slab_page *page, unsigned index)
{
return (struct util_slab_block*)
((uint8_t*)page + sizeof(struct util_slab_page) +
(pool->block_size * index));
}
static void util_slab_add_new_page(struct util_slab_mempool *pool)
{
struct util_slab_page *page;
struct util_slab_block *block;
int i;
page = MALLOC(pool->page_size);
insert_at_tail(&pool->list, page);
/* Mark all blocks as free. */
for (i = 0; i < pool->num_blocks-1; i++) {
block = util_slab_get_block(pool, page, i);
block->next_free = util_slab_get_block(pool, page, i+1);
block->magic = UTIL_SLAB_MAGIC;
}
block = util_slab_get_block(pool, page, pool->num_blocks-1);
block->next_free = pool->first_free;
block->magic = UTIL_SLAB_MAGIC;
pool->first_free = util_slab_get_block(pool, page, 0);
pool->num_pages++;
#if 0
fprintf(stderr, "New page! Num of pages: %i\n", pool->num_pages);
#endif
}
static void *util_slab_alloc_st(struct util_slab_mempool *pool)
{
struct util_slab_block *block;
if (!pool->first_free)
util_slab_add_new_page(pool);
block = pool->first_free;
assert(block->magic == UTIL_SLAB_MAGIC);
pool->first_free = block->next_free;
return (uint8_t*)block + sizeof(struct util_slab_block);
}
static void util_slab_free_st(struct util_slab_mempool *pool, void *ptr)
{
struct util_slab_block *block =
(struct util_slab_block*)
((uint8_t*)ptr - sizeof(struct util_slab_block));
assert(block->magic == UTIL_SLAB_MAGIC);
block->next_free = pool->first_free;
pool->first_free = block;
}
static void *util_slab_alloc_mt(struct util_slab_mempool *pool)
{
void *mem;
pipe_mutex_lock(pool->mutex);
mem = util_slab_alloc_st(pool);
pipe_mutex_unlock(pool->mutex);
return mem;
}
static void util_slab_free_mt(struct util_slab_mempool *pool, void *ptr)
{
pipe_mutex_lock(pool->mutex);
util_slab_free_st(pool, ptr);
pipe_mutex_unlock(pool->mutex);
}
void util_slab_set_thread_safety(struct util_slab_mempool *pool,
enum util_slab_threading threading)
{
pool->threading = threading;
if (threading) {
pool->alloc = util_slab_alloc_mt;
pool->free = util_slab_free_mt;
} else {
pool->alloc = util_slab_alloc_st;
pool->free = util_slab_free_st;
}
}
void util_slab_create(struct util_slab_mempool *pool,
unsigned item_size,
unsigned num_blocks,
enum util_slab_threading threading)
{
item_size = align(item_size, sizeof(intptr_t));
pool->num_pages = 0;
pool->num_blocks = num_blocks;
pool->block_size = sizeof(struct util_slab_block) + item_size;
pool->block_size = align(pool->block_size, sizeof(intptr_t));
pool->page_size = sizeof(struct util_slab_page) +
num_blocks * pool->block_size;
pool->first_free = NULL;
make_empty_list(&pool->list);
pipe_mutex_init(pool->mutex);
util_slab_set_thread_safety(pool, threading);
}
void util_slab_destroy(struct util_slab_mempool *pool)
{
struct util_slab_page *page, *temp;
if (pool->list.next) {
foreach_s(page, temp, &pool->list) {
remove_from_list(page);
FREE(page);
}
}
pipe_mutex_destroy(pool->mutex);
}