425 lines
14 KiB
C
425 lines
14 KiB
C
/* -*- Mode: C; tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*- */
|
|
/*
|
|
* Slabs memory allocation, based on powers-of-N. Slabs are up to 1MB in size
|
|
* and are divided into chunks. The chunk sizes start off at the size of the
|
|
* "item" structure plus space for a small key and value. They increase by
|
|
* a multiplier factor from there, up to half the maximum slab size. The last
|
|
* slab size is always 1MB, since that's the maximum item size allowed by the
|
|
* memcached protocol.
|
|
*/
|
|
#include "config.h"
|
|
|
|
#include <fcntl.h>
|
|
#include <errno.h>
|
|
#include <stdlib.h>
|
|
#include <stdio.h>
|
|
#include <string.h>
|
|
#include <assert.h>
|
|
#include <pthread.h>
|
|
#include <inttypes.h>
|
|
#include <stdarg.h>
|
|
|
|
#include "default_engine.h"
|
|
|
|
/*
|
|
* Forward Declarations
|
|
*/
|
|
static int do_slabs_newslab(struct default_engine *engine, const unsigned int id);
|
|
static void *memory_allocate(struct default_engine *engine, size_t size);
|
|
|
|
#ifndef DONT_PREALLOC_SLABS
|
|
/* Preallocate as many slab pages as possible (called from slabs_init)
|
|
on start-up, so users don't get confused out-of-memory errors when
|
|
they do have free (in-slab) space, but no space to make new slabs.
|
|
if maxslabs is 18 (POWER_LARGEST - POWER_SMALLEST + 1), then all
|
|
slab types can be made. if max memory is less than 18 MB, only the
|
|
smaller ones will be made. */
|
|
static void slabs_preallocate (const unsigned int maxslabs);
|
|
#endif
|
|
|
|
/*
|
|
* Figures out which slab class (chunk size) is required to store an item of
|
|
* a given size.
|
|
*
|
|
* Given object size, return id to use when allocating/freeing memory for object
|
|
* 0 means error: can't store such a large object
|
|
*/
|
|
|
|
unsigned int slabs_clsid(struct default_engine *engine, const size_t size) {
|
|
int res = POWER_SMALLEST;
|
|
|
|
if (size == 0)
|
|
return 0;
|
|
while (size > engine->slabs.slabclass[res].size)
|
|
if (res++ == engine->slabs.power_largest) /* won't fit in the biggest slab */
|
|
return 0;
|
|
return res;
|
|
}
|
|
|
|
/**
|
|
* Determines the chunk sizes and initializes the slab class descriptors
|
|
* accordingly.
|
|
*/
|
|
ENGINE_ERROR_CODE slabs_init(struct default_engine *engine,
|
|
const size_t limit,
|
|
const double factor,
|
|
const bool prealloc) {
|
|
int i = POWER_SMALLEST - 1;
|
|
unsigned int size = sizeof(hash_item) + engine->config.chunk_size;
|
|
|
|
engine->slabs.mem_limit = limit;
|
|
|
|
if (prealloc) {
|
|
/* Allocate everything in a big chunk with malloc */
|
|
engine->slabs.mem_base = malloc(engine->slabs.mem_limit);
|
|
if (engine->slabs.mem_base != NULL) {
|
|
engine->slabs.mem_current = engine->slabs.mem_base;
|
|
engine->slabs.mem_avail = engine->slabs.mem_limit;
|
|
} else {
|
|
return ENGINE_ENOMEM;
|
|
}
|
|
}
|
|
|
|
memset(engine->slabs.slabclass, 0, sizeof(engine->slabs.slabclass));
|
|
|
|
while (++i < POWER_LARGEST && size <= engine->config.item_size_max / factor) {
|
|
/* Make sure items are always n-byte aligned */
|
|
if (size % CHUNK_ALIGN_BYTES)
|
|
size += CHUNK_ALIGN_BYTES - (size % CHUNK_ALIGN_BYTES);
|
|
|
|
engine->slabs.slabclass[i].size = size;
|
|
engine->slabs.slabclass[i].perslab = engine->config.item_size_max / engine->slabs.slabclass[i].size;
|
|
size *= factor;
|
|
if (engine->config.verbose > 1) {
|
|
EXTENSION_LOGGER_DESCRIPTOR *logger;
|
|
logger = (void*)engine->server.extension->get_extension(EXTENSION_LOGGER);
|
|
logger->log(EXTENSION_LOG_INFO, NULL,
|
|
"slab class %3d: chunk size %9u perslab %7u\n",
|
|
i, engine->slabs.slabclass[i].size,
|
|
engine->slabs.slabclass[i].perslab);
|
|
}
|
|
}
|
|
|
|
engine->slabs.power_largest = i;
|
|
engine->slabs.slabclass[engine->slabs.power_largest].size = engine->config.item_size_max;
|
|
engine->slabs.slabclass[engine->slabs.power_largest].perslab = 1;
|
|
if (engine->config.verbose > 1) {
|
|
EXTENSION_LOGGER_DESCRIPTOR *logger;
|
|
logger = (void*)engine->server.extension->get_extension(EXTENSION_LOGGER);
|
|
logger->log(EXTENSION_LOG_INFO, NULL,
|
|
"slab class %3d: chunk size %9u perslab %7u\n",
|
|
i, engine->slabs.slabclass[i].size,
|
|
engine->slabs.slabclass[i].perslab);
|
|
}
|
|
|
|
/* for the test suite: faking of how much we've already malloc'd */
|
|
{
|
|
char *t_initial_malloc = getenv("T_MEMD_INITIAL_MALLOC");
|
|
if (t_initial_malloc) {
|
|
engine->slabs.mem_malloced = (size_t)atol(t_initial_malloc);
|
|
}
|
|
|
|
}
|
|
|
|
#ifndef DONT_PREALLOC_SLABS
|
|
{
|
|
char *pre_alloc = getenv("T_MEMD_SLABS_ALLOC");
|
|
|
|
if (pre_alloc == NULL || atoi(pre_alloc) != 0) {
|
|
slabs_preallocate(power_largest);
|
|
}
|
|
}
|
|
#endif
|
|
|
|
return ENGINE_SUCCESS;
|
|
}
|
|
|
|
#ifndef DONT_PREALLOC_SLABS
|
|
static void slabs_preallocate (const unsigned int maxslabs) {
|
|
int i;
|
|
unsigned int prealloc = 0;
|
|
|
|
/* pre-allocate a 1MB slab in every size class so people don't get
|
|
confused by non-intuitive "SERVER_ERROR out of memory"
|
|
messages. this is the most common question on the mailing
|
|
list. if you really don't want this, you can rebuild without
|
|
these three lines. */
|
|
|
|
for (i = POWER_SMALLEST; i <= POWER_LARGEST; i++) {
|
|
if (++prealloc > maxslabs)
|
|
return;
|
|
do_slabs_newslab(i);
|
|
}
|
|
|
|
}
|
|
#endif
|
|
|
|
static int grow_slab_list (struct default_engine *engine, const unsigned int id) {
|
|
slabclass_t *p = &engine->slabs.slabclass[id];
|
|
if (p->slabs == p->list_size) {
|
|
size_t new_size = (p->list_size != 0) ? p->list_size * 2 : 16;
|
|
void *new_list = realloc(p->slab_list, new_size * sizeof(void *));
|
|
if (new_list == 0) return 0;
|
|
p->list_size = new_size;
|
|
p->slab_list = new_list;
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
static int do_slabs_newslab(struct default_engine *engine, const unsigned int id) {
|
|
slabclass_t *p = &engine->slabs.slabclass[id];
|
|
int len = p->size * p->perslab;
|
|
char *ptr;
|
|
|
|
if ((engine->slabs.mem_limit && engine->slabs.mem_malloced + len > engine->slabs.mem_limit && p->slabs > 0) ||
|
|
(grow_slab_list(engine, id) == 0) ||
|
|
((ptr = memory_allocate(engine, (size_t)len)) == 0)) {
|
|
|
|
MEMCACHED_SLABS_SLABCLASS_ALLOCATE_FAILED(id);
|
|
return 0;
|
|
}
|
|
|
|
memset(ptr, 0, (size_t)len);
|
|
p->end_page_ptr = ptr;
|
|
p->end_page_free = p->perslab;
|
|
|
|
p->slab_list[p->slabs++] = ptr;
|
|
engine->slabs.mem_malloced += len;
|
|
MEMCACHED_SLABS_SLABCLASS_ALLOCATE(id);
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*@null@*/
|
|
static void *do_slabs_alloc(struct default_engine *engine, const size_t size, unsigned int id) {
|
|
slabclass_t *p;
|
|
void *ret = NULL;
|
|
|
|
if (id < POWER_SMALLEST || id > engine->slabs.power_largest) {
|
|
MEMCACHED_SLABS_ALLOCATE_FAILED(size, 0);
|
|
return NULL;
|
|
}
|
|
|
|
p = &engine->slabs.slabclass[id];
|
|
|
|
#ifdef USE_SYSTEM_MALLOC
|
|
if (engine->slabs.mem_limit && engine->slabs.mem_malloced + size > engine->slabs.mem_limit) {
|
|
MEMCACHED_SLABS_ALLOCATE_FAILED(size, id);
|
|
return 0;
|
|
}
|
|
engine->slabs.mem_malloced += size;
|
|
ret = malloc(size);
|
|
MEMCACHED_SLABS_ALLOCATE(size, id, 0, ret);
|
|
return ret;
|
|
#endif
|
|
|
|
/* fail unless we have space at the end of a recently allocated page,
|
|
we have something on our freelist, or we could allocate a new page */
|
|
if (! (p->end_page_ptr != 0 || p->sl_curr != 0 ||
|
|
do_slabs_newslab(engine, id) != 0)) {
|
|
/* We don't have more memory available */
|
|
ret = NULL;
|
|
} else if (p->sl_curr != 0) {
|
|
/* return off our freelist */
|
|
ret = p->slots[--p->sl_curr];
|
|
} else {
|
|
/* if we recently allocated a whole page, return from that */
|
|
assert(p->end_page_ptr != NULL);
|
|
ret = p->end_page_ptr;
|
|
if (--p->end_page_free != 0) {
|
|
p->end_page_ptr = ((caddr_t)p->end_page_ptr) + p->size;
|
|
} else {
|
|
p->end_page_ptr = 0;
|
|
}
|
|
}
|
|
|
|
if (ret) {
|
|
p->requested += size;
|
|
MEMCACHED_SLABS_ALLOCATE(size, id, p->size, ret);
|
|
} else {
|
|
MEMCACHED_SLABS_ALLOCATE_FAILED(size, id);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void do_slabs_free(struct default_engine *engine, void *ptr, const size_t size, unsigned int id) {
|
|
slabclass_t *p;
|
|
|
|
if (id < POWER_SMALLEST || id > engine->slabs.power_largest)
|
|
return;
|
|
|
|
MEMCACHED_SLABS_FREE(size, id, ptr);
|
|
p = &engine->slabs.slabclass[id];
|
|
|
|
#ifdef USE_SYSTEM_MALLOC
|
|
engine->slabs.mem_malloced -= size;
|
|
free(ptr);
|
|
return;
|
|
#endif
|
|
|
|
if (p->sl_curr == p->sl_total) { /* need more space on the free list */
|
|
int new_size = (p->sl_total != 0) ? p->sl_total * 2 : 16; /* 16 is arbitrary */
|
|
void **new_slots = realloc(p->slots, new_size * sizeof(void *));
|
|
if (new_slots == 0)
|
|
return;
|
|
p->slots = new_slots;
|
|
p->sl_total = new_size;
|
|
}
|
|
p->slots[p->sl_curr++] = ptr;
|
|
p->requested -= size;
|
|
return;
|
|
}
|
|
|
|
void add_statistics(const void *cookie, ADD_STAT add_stats,
|
|
const char* prefix, int num, const char *key,
|
|
const char *fmt, ...) {
|
|
char name[80], val[80];
|
|
int klen = 0, vlen;
|
|
va_list ap;
|
|
|
|
assert(cookie);
|
|
assert(add_stats);
|
|
assert(key);
|
|
|
|
va_start(ap, fmt);
|
|
vlen = vsnprintf(val, sizeof(val) - 1, fmt, ap);
|
|
va_end(ap);
|
|
|
|
if (prefix != NULL) {
|
|
klen = snprintf(name, sizeof(name), "%s:", prefix);
|
|
}
|
|
|
|
if (num != -1) {
|
|
klen += snprintf(name + klen, sizeof(name) - klen, "%d:", num);
|
|
}
|
|
|
|
klen += snprintf(name + klen, sizeof(name) - klen, "%s", key);
|
|
|
|
add_stats(name, klen, val, vlen, cookie);
|
|
}
|
|
|
|
/*@null@*/
|
|
static void do_slabs_stats(struct default_engine *engine, ADD_STAT add_stats, const void *cookie) {
|
|
int i, total;
|
|
/* Get the per-thread stats which contain some interesting aggregates */
|
|
#ifdef FUTURE
|
|
struct conn *conn = (struct conn*)cookie;
|
|
struct thread_stats thread_stats;
|
|
threadlocal_stats_aggregate(c, &thread_stats);
|
|
#endif
|
|
|
|
total = 0;
|
|
for(i = POWER_SMALLEST; i <= engine->slabs.power_largest; i++) {
|
|
slabclass_t *p = &engine->slabs.slabclass[i];
|
|
if (p->slabs != 0) {
|
|
uint32_t perslab, slabs;
|
|
slabs = p->slabs;
|
|
perslab = p->perslab;
|
|
|
|
add_statistics(cookie, add_stats, NULL, i, "chunk_size", "%u",
|
|
p->size);
|
|
add_statistics(cookie, add_stats, NULL, i, "chunks_per_page", "%u",
|
|
perslab);
|
|
add_statistics(cookie, add_stats, NULL, i, "total_pages", "%u",
|
|
slabs);
|
|
add_statistics(cookie, add_stats, NULL, i, "total_chunks", "%u",
|
|
slabs * perslab);
|
|
add_statistics(cookie, add_stats, NULL, i, "used_chunks", "%u",
|
|
slabs*perslab - p->sl_curr - p->end_page_free);
|
|
add_statistics(cookie, add_stats, NULL, i, "free_chunks", "%u",
|
|
p->sl_curr);
|
|
add_statistics(cookie, add_stats, NULL, i, "free_chunks_end", "%u",
|
|
p->end_page_free);
|
|
add_statistics(cookie, add_stats, NULL, i, "mem_requested", "%zu",
|
|
p->requested);
|
|
#ifdef FUTURE
|
|
add_statistics(cookie, add_stats, NULL, i, "get_hits", "%"PRIu64,
|
|
thread_stats.slab_stats[i].get_hits);
|
|
add_statistics(cookie, add_stats, NULL, i, "cmd_set", "%"PRIu64,
|
|
thread_stats.slab_stats[i].set_cmds);
|
|
add_statistics(cookie, add_stats, NULL, i, "delete_hits", "%"PRIu64,
|
|
thread_stats.slab_stats[i].delete_hits);
|
|
add_statistics(cookie, add_stats, NULL, i, "cas_hits", "%"PRIu64,
|
|
thread_stats.slab_stats[i].cas_hits);
|
|
add_statistics(cookie, add_stats, NULL, i, "cas_badval", "%"PRIu64,
|
|
thread_stats.slab_stats[i].cas_badval);
|
|
#endif
|
|
total++;
|
|
}
|
|
}
|
|
|
|
/* add overall slab stats and append terminator */
|
|
|
|
add_statistics(cookie, add_stats, NULL, -1, "active_slabs", "%d", total);
|
|
add_statistics(cookie, add_stats, NULL, -1, "total_malloced", "%zu",
|
|
engine->slabs.mem_malloced);
|
|
}
|
|
|
|
static void *memory_allocate(struct default_engine *engine, size_t size) {
|
|
void *ret;
|
|
|
|
if (engine->slabs.mem_base == NULL) {
|
|
/* We are not using a preallocated large memory chunk */
|
|
ret = malloc(size);
|
|
} else {
|
|
ret = engine->slabs.mem_current;
|
|
|
|
if (size > engine->slabs.mem_avail) {
|
|
return NULL;
|
|
}
|
|
|
|
/* mem_current pointer _must_ be aligned!!! */
|
|
if (size % CHUNK_ALIGN_BYTES) {
|
|
size += CHUNK_ALIGN_BYTES - (size % CHUNK_ALIGN_BYTES);
|
|
}
|
|
|
|
engine->slabs.mem_current = ((char*)engine->slabs.mem_current) + size;
|
|
if (size < engine->slabs.mem_avail) {
|
|
engine->slabs.mem_avail -= size;
|
|
} else {
|
|
engine->slabs.mem_avail = 0;
|
|
}
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
void *slabs_alloc(struct default_engine *engine, size_t size, unsigned int id) {
|
|
void *ret;
|
|
|
|
pthread_mutex_lock(&engine->slabs.lock);
|
|
ret = do_slabs_alloc(engine, size, id);
|
|
pthread_mutex_unlock(&engine->slabs.lock);
|
|
return ret;
|
|
}
|
|
|
|
void slabs_free(struct default_engine *engine, void *ptr, size_t size, unsigned int id) {
|
|
pthread_mutex_lock(&engine->slabs.lock);
|
|
do_slabs_free(engine, ptr, size, id);
|
|
pthread_mutex_unlock(&engine->slabs.lock);
|
|
}
|
|
|
|
void slabs_stats(struct default_engine *engine, ADD_STAT add_stats, const void *c) {
|
|
pthread_mutex_lock(&engine->slabs.lock);
|
|
do_slabs_stats(engine, add_stats, c);
|
|
pthread_mutex_unlock(&engine->slabs.lock);
|
|
}
|
|
|
|
void slabs_adjust_mem_requested(struct default_engine *engine, unsigned int id, size_t old, size_t ntotal)
|
|
{
|
|
pthread_mutex_lock(&engine->slabs.lock);
|
|
slabclass_t *p;
|
|
if (id < POWER_SMALLEST || id > engine->slabs.power_largest) {
|
|
EXTENSION_LOGGER_DESCRIPTOR *logger;
|
|
logger = (void*)engine->server.extension->get_extension(EXTENSION_LOGGER);
|
|
logger->log(EXTENSION_LOG_WARNING, NULL,
|
|
"Internal error! Invalid slab class\n");
|
|
abort();
|
|
}
|
|
|
|
p = &engine->slabs.slabclass[id];
|
|
p->requested = p->requested - old + ntotal;
|
|
pthread_mutex_unlock(&engine->slabs.lock);
|
|
}
|