/* -*- Mode: C; tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*- */ /* * Slabs memory allocation, based on powers-of-N. Slabs are up to 1MB in size * and are divided into chunks. The chunk sizes start off at the size of the * "item" structure plus space for a small key and value. They increase by * a multiplier factor from there, up to half the maximum slab size. The last * slab size is always 1MB, since that's the maximum item size allowed by the * memcached protocol. */ #include "config.h" #include #include #include #include #include #include #include #include #include #include "default_engine.h" /* * Forward Declarations */ static int do_slabs_newslab(struct default_engine *engine, const unsigned int id); static void *memory_allocate(struct default_engine *engine, size_t size); #ifndef DONT_PREALLOC_SLABS /* Preallocate as many slab pages as possible (called from slabs_init) on start-up, so users don't get confused out-of-memory errors when they do have free (in-slab) space, but no space to make new slabs. if maxslabs is 18 (POWER_LARGEST - POWER_SMALLEST + 1), then all slab types can be made. if max memory is less than 18 MB, only the smaller ones will be made. */ static void slabs_preallocate (const unsigned int maxslabs); #endif /* * Figures out which slab class (chunk size) is required to store an item of * a given size. * * Given object size, return id to use when allocating/freeing memory for object * 0 means error: can't store such a large object */ unsigned int slabs_clsid(struct default_engine *engine, const size_t size) { int res = POWER_SMALLEST; if (size == 0) return 0; while (size > engine->slabs.slabclass[res].size) if (res++ == engine->slabs.power_largest) /* won't fit in the biggest slab */ return 0; return res; } /** * Determines the chunk sizes and initializes the slab class descriptors * accordingly. */ ENGINE_ERROR_CODE slabs_init(struct default_engine *engine, const size_t limit, const double factor, const bool prealloc) { int i = POWER_SMALLEST - 1; unsigned int size = sizeof(hash_item) + engine->config.chunk_size; engine->slabs.mem_limit = limit; if (prealloc) { /* Allocate everything in a big chunk with malloc */ engine->slabs.mem_base = malloc(engine->slabs.mem_limit); if (engine->slabs.mem_base != NULL) { engine->slabs.mem_current = engine->slabs.mem_base; engine->slabs.mem_avail = engine->slabs.mem_limit; } else { return ENGINE_ENOMEM; } } memset(engine->slabs.slabclass, 0, sizeof(engine->slabs.slabclass)); while (++i < POWER_LARGEST && size <= engine->config.item_size_max / factor) { /* Make sure items are always n-byte aligned */ if (size % CHUNK_ALIGN_BYTES) size += CHUNK_ALIGN_BYTES - (size % CHUNK_ALIGN_BYTES); engine->slabs.slabclass[i].size = size; engine->slabs.slabclass[i].perslab = engine->config.item_size_max / engine->slabs.slabclass[i].size; size *= factor; if (engine->config.verbose > 1) { EXTENSION_LOGGER_DESCRIPTOR *logger; logger = (void*)engine->server.extension->get_extension(EXTENSION_LOGGER); logger->log(EXTENSION_LOG_INFO, NULL, "slab class %3d: chunk size %9u perslab %7u\n", i, engine->slabs.slabclass[i].size, engine->slabs.slabclass[i].perslab); } } engine->slabs.power_largest = i; engine->slabs.slabclass[engine->slabs.power_largest].size = engine->config.item_size_max; engine->slabs.slabclass[engine->slabs.power_largest].perslab = 1; if (engine->config.verbose > 1) { EXTENSION_LOGGER_DESCRIPTOR *logger; logger = (void*)engine->server.extension->get_extension(EXTENSION_LOGGER); logger->log(EXTENSION_LOG_INFO, NULL, "slab class %3d: chunk size %9u perslab %7u\n", i, engine->slabs.slabclass[i].size, engine->slabs.slabclass[i].perslab); } /* for the test suite: faking of how much we've already malloc'd */ { char *t_initial_malloc = getenv("T_MEMD_INITIAL_MALLOC"); if (t_initial_malloc) { engine->slabs.mem_malloced = (size_t)atol(t_initial_malloc); } } #ifndef DONT_PREALLOC_SLABS { char *pre_alloc = getenv("T_MEMD_SLABS_ALLOC"); if (pre_alloc == NULL || atoi(pre_alloc) != 0) { slabs_preallocate(power_largest); } } #endif return ENGINE_SUCCESS; } #ifndef DONT_PREALLOC_SLABS static void slabs_preallocate (const unsigned int maxslabs) { int i; unsigned int prealloc = 0; /* pre-allocate a 1MB slab in every size class so people don't get confused by non-intuitive "SERVER_ERROR out of memory" messages. this is the most common question on the mailing list. if you really don't want this, you can rebuild without these three lines. */ for (i = POWER_SMALLEST; i <= POWER_LARGEST; i++) { if (++prealloc > maxslabs) return; do_slabs_newslab(i); } } #endif static int grow_slab_list (struct default_engine *engine, const unsigned int id) { slabclass_t *p = &engine->slabs.slabclass[id]; if (p->slabs == p->list_size) { size_t new_size = (p->list_size != 0) ? p->list_size * 2 : 16; void *new_list = realloc(p->slab_list, new_size * sizeof(void *)); if (new_list == 0) return 0; p->list_size = new_size; p->slab_list = new_list; } return 1; } static int do_slabs_newslab(struct default_engine *engine, const unsigned int id) { slabclass_t *p = &engine->slabs.slabclass[id]; int len = p->size * p->perslab; char *ptr; if ((engine->slabs.mem_limit && engine->slabs.mem_malloced + len > engine->slabs.mem_limit && p->slabs > 0) || (grow_slab_list(engine, id) == 0) || ((ptr = memory_allocate(engine, (size_t)len)) == 0)) { MEMCACHED_SLABS_SLABCLASS_ALLOCATE_FAILED(id); return 0; } memset(ptr, 0, (size_t)len); p->end_page_ptr = ptr; p->end_page_free = p->perslab; p->slab_list[p->slabs++] = ptr; engine->slabs.mem_malloced += len; MEMCACHED_SLABS_SLABCLASS_ALLOCATE(id); return 1; } /*@null@*/ static void *do_slabs_alloc(struct default_engine *engine, const size_t size, unsigned int id) { slabclass_t *p; void *ret = NULL; if (id < POWER_SMALLEST || id > engine->slabs.power_largest) { MEMCACHED_SLABS_ALLOCATE_FAILED(size, 0); return NULL; } p = &engine->slabs.slabclass[id]; #ifdef USE_SYSTEM_MALLOC if (engine->slabs.mem_limit && engine->slabs.mem_malloced + size > engine->slabs.mem_limit) { MEMCACHED_SLABS_ALLOCATE_FAILED(size, id); return 0; } engine->slabs.mem_malloced += size; ret = malloc(size); MEMCACHED_SLABS_ALLOCATE(size, id, 0, ret); return ret; #endif /* fail unless we have space at the end of a recently allocated page, we have something on our freelist, or we could allocate a new page */ if (! (p->end_page_ptr != 0 || p->sl_curr != 0 || do_slabs_newslab(engine, id) != 0)) { /* We don't have more memory available */ ret = NULL; } else if (p->sl_curr != 0) { /* return off our freelist */ ret = p->slots[--p->sl_curr]; } else { /* if we recently allocated a whole page, return from that */ assert(p->end_page_ptr != NULL); ret = p->end_page_ptr; if (--p->end_page_free != 0) { p->end_page_ptr = ((caddr_t)p->end_page_ptr) + p->size; } else { p->end_page_ptr = 0; } } if (ret) { p->requested += size; MEMCACHED_SLABS_ALLOCATE(size, id, p->size, ret); } else { MEMCACHED_SLABS_ALLOCATE_FAILED(size, id); } return ret; } static void do_slabs_free(struct default_engine *engine, void *ptr, const size_t size, unsigned int id) { slabclass_t *p; if (id < POWER_SMALLEST || id > engine->slabs.power_largest) return; MEMCACHED_SLABS_FREE(size, id, ptr); p = &engine->slabs.slabclass[id]; #ifdef USE_SYSTEM_MALLOC engine->slabs.mem_malloced -= size; free(ptr); return; #endif if (p->sl_curr == p->sl_total) { /* need more space on the free list */ int new_size = (p->sl_total != 0) ? p->sl_total * 2 : 16; /* 16 is arbitrary */ void **new_slots = realloc(p->slots, new_size * sizeof(void *)); if (new_slots == 0) return; p->slots = new_slots; p->sl_total = new_size; } p->slots[p->sl_curr++] = ptr; p->requested -= size; return; } void add_statistics(const void *cookie, ADD_STAT add_stats, const char* prefix, int num, const char *key, const char *fmt, ...) { char name[80], val[80]; int klen = 0, vlen; va_list ap; assert(cookie); assert(add_stats); assert(key); va_start(ap, fmt); vlen = vsnprintf(val, sizeof(val) - 1, fmt, ap); va_end(ap); if (prefix != NULL) { klen = snprintf(name, sizeof(name), "%s:", prefix); } if (num != -1) { klen += snprintf(name + klen, sizeof(name) - klen, "%d:", num); } klen += snprintf(name + klen, sizeof(name) - klen, "%s", key); add_stats(name, klen, val, vlen, cookie); } /*@null@*/ static void do_slabs_stats(struct default_engine *engine, ADD_STAT add_stats, const void *cookie) { int i, total; /* Get the per-thread stats which contain some interesting aggregates */ #ifdef FUTURE struct conn *conn = (struct conn*)cookie; struct thread_stats thread_stats; threadlocal_stats_aggregate(c, &thread_stats); #endif total = 0; for(i = POWER_SMALLEST; i <= engine->slabs.power_largest; i++) { slabclass_t *p = &engine->slabs.slabclass[i]; if (p->slabs != 0) { uint32_t perslab, slabs; slabs = p->slabs; perslab = p->perslab; add_statistics(cookie, add_stats, NULL, i, "chunk_size", "%u", p->size); add_statistics(cookie, add_stats, NULL, i, "chunks_per_page", "%u", perslab); add_statistics(cookie, add_stats, NULL, i, "total_pages", "%u", slabs); add_statistics(cookie, add_stats, NULL, i, "total_chunks", "%u", slabs * perslab); add_statistics(cookie, add_stats, NULL, i, "used_chunks", "%u", slabs*perslab - p->sl_curr - p->end_page_free); add_statistics(cookie, add_stats, NULL, i, "free_chunks", "%u", p->sl_curr); add_statistics(cookie, add_stats, NULL, i, "free_chunks_end", "%u", p->end_page_free); add_statistics(cookie, add_stats, NULL, i, "mem_requested", "%zu", p->requested); #ifdef FUTURE add_statistics(cookie, add_stats, NULL, i, "get_hits", "%"PRIu64, thread_stats.slab_stats[i].get_hits); add_statistics(cookie, add_stats, NULL, i, "cmd_set", "%"PRIu64, thread_stats.slab_stats[i].set_cmds); add_statistics(cookie, add_stats, NULL, i, "delete_hits", "%"PRIu64, thread_stats.slab_stats[i].delete_hits); add_statistics(cookie, add_stats, NULL, i, "cas_hits", "%"PRIu64, thread_stats.slab_stats[i].cas_hits); add_statistics(cookie, add_stats, NULL, i, "cas_badval", "%"PRIu64, thread_stats.slab_stats[i].cas_badval); #endif total++; } } /* add overall slab stats and append terminator */ add_statistics(cookie, add_stats, NULL, -1, "active_slabs", "%d", total); add_statistics(cookie, add_stats, NULL, -1, "total_malloced", "%zu", engine->slabs.mem_malloced); } static void *memory_allocate(struct default_engine *engine, size_t size) { void *ret; if (engine->slabs.mem_base == NULL) { /* We are not using a preallocated large memory chunk */ ret = malloc(size); } else { ret = engine->slabs.mem_current; if (size > engine->slabs.mem_avail) { return NULL; } /* mem_current pointer _must_ be aligned!!! */ if (size % CHUNK_ALIGN_BYTES) { size += CHUNK_ALIGN_BYTES - (size % CHUNK_ALIGN_BYTES); } engine->slabs.mem_current = ((char*)engine->slabs.mem_current) + size; if (size < engine->slabs.mem_avail) { engine->slabs.mem_avail -= size; } else { engine->slabs.mem_avail = 0; } } return ret; } void *slabs_alloc(struct default_engine *engine, size_t size, unsigned int id) { void *ret; pthread_mutex_lock(&engine->slabs.lock); ret = do_slabs_alloc(engine, size, id); pthread_mutex_unlock(&engine->slabs.lock); return ret; } void slabs_free(struct default_engine *engine, void *ptr, size_t size, unsigned int id) { pthread_mutex_lock(&engine->slabs.lock); do_slabs_free(engine, ptr, size, id); pthread_mutex_unlock(&engine->slabs.lock); } void slabs_stats(struct default_engine *engine, ADD_STAT add_stats, const void *c) { pthread_mutex_lock(&engine->slabs.lock); do_slabs_stats(engine, add_stats, c); pthread_mutex_unlock(&engine->slabs.lock); } void slabs_adjust_mem_requested(struct default_engine *engine, unsigned int id, size_t old, size_t ntotal) { pthread_mutex_lock(&engine->slabs.lock); slabclass_t *p; if (id < POWER_SMALLEST || id > engine->slabs.power_largest) { EXTENSION_LOGGER_DESCRIPTOR *logger; logger = (void*)engine->server.extension->get_extension(EXTENSION_LOGGER); logger->log(EXTENSION_LOG_WARNING, NULL, "Internal error! Invalid slab class\n"); abort(); } p = &engine->slabs.slabclass[id]; p->requested = p->requested - old + ntotal; pthread_mutex_unlock(&engine->slabs.lock); }