X-Git-Url: https://git.sur5r.net/?a=blobdiff_plain;f=servers%2Fslapd%2Fsl_malloc.c;h=3feb7739030a64c2bac5337239b56836c20ff89b;hb=01f698bce979660270207ce052f3bdbbd79365a5;hp=ad8fd5d2320832a01a3b2441c5664cf98c51703f;hpb=423c13a6d6ad4f311f65bcf9295ecbdc2d98e0c1;p=openldap diff --git a/servers/slapd/sl_malloc.c b/servers/slapd/sl_malloc.c index ad8fd5d232..3feb773903 100644 --- a/servers/slapd/sl_malloc.c +++ b/servers/slapd/sl_malloc.c @@ -2,7 +2,7 @@ /* $OpenLDAP$ */ /* This work is part of OpenLDAP Software . * - * Copyright 2003-2010 The OpenLDAP Foundation. + * Copyright 2003-2014 The OpenLDAP Foundation. * All rights reserved. * * Redistribution and use in source and binary forms, with or without @@ -21,8 +21,85 @@ #include "slap.h" +#ifdef USE_VALGRIND +/* Get debugging help from Valgrind */ +#include +#define VGMEMP_MARK(m,s) VALGRIND_MAKE_MEM_NOACCESS(m,s) +#define VGMEMP_CREATE(h,r,z) VALGRIND_CREATE_MEMPOOL(h,r,z) +#define VGMEMP_TRIM(h,a,s) VALGRIND_MEMPOOL_TRIM(h,a,s) +#define VGMEMP_ALLOC(h,a,s) VALGRIND_MEMPOOL_ALLOC(h,a,s) +#define VGMEMP_CHANGE(h,a,b,s) VALGRIND_MEMPOOL_CHANGE(h,a,b,s) +#else +#define VGMEMP_MARK(m,s) +#define VGMEMP_CREATE(h,r,z) +#define VGMEMP_TRIM(h,a,s) +#define VGMEMP_ALLOC(h,a,s) +#define VGMEMP_CHANGE(h,a,b,s) +#endif + +/* + * This allocator returns temporary memory from a slab in a given memory + * context, aligned on a 2-int boundary. It cannot be used for data + * which will outlive the task allocating it. + * + * A new memory context attaches to the creator's thread context, if any. + * Threads cannot use other threads' memory contexts; there are no locks. + * + * The caller of slap_sl_malloc, usually a thread pool task, must + * slap_sl_free the memory before finishing: New tasks reuse the context + * and normally reset it, reclaiming memory left over from last task. + * + * The allocator helps memory fragmentation, speed and memory leaks. + * It is not (yet) reliable as a garbage collector: + * + * It falls back to context NULL - plain ber_memalloc() - when the + * context's slab is full. A reset does not reclaim such memory. + * Conversely, free/realloc of data not from the given context assumes + * context NULL. The data must not belong to another memory context. + * + * Code which has lost track of the current memory context can try + * slap_sl_context() or ch_malloc.c:ch_free/ch_realloc(). + * + * Allocations cannot yet return failure. Like ch_malloc, they succeed + * or abort slapd. This will change, do fix code which assumes success. + */ + +/* + * The stack-based allocator stores (ber_len_t)sizeof(head+block) at + * allocated blocks' head - and in freed blocks also at the tail, marked + * by ORing *next* block's head with 1. Freed blocks are only reclaimed + * from the last block forward. This is fast, but when a block is never + * freed, older blocks will not be reclaimed until the slab is reset... + */ + +#ifdef SLAP_NO_SL_MALLOC /* Useful with memory debuggers like Valgrind */ +enum { No_sl_malloc = 1 }; +#else +enum { No_sl_malloc = 0 }; +#endif + +#define SLAP_SLAB_SOBLOCK 64 + +struct slab_object { + void *so_ptr; + int so_blockhead; + LDAP_LIST_ENTRY(slab_object) so_link; +}; + +struct slab_heap { + void *sh_base; + void *sh_last; + void *sh_end; + int sh_stack; + int sh_maxorder; + unsigned char **sh_map; + LDAP_LIST_HEAD(sh_freelist, slab_object) *sh_free; + LDAP_LIST_HEAD(sh_so, slab_object) sh_sopool; +}; + enum { - Align = 2 * sizeof(int), + Align = sizeof(ber_len_t) > 2*sizeof(int) + ? sizeof(ber_len_t) : 2*sizeof(int), Align_log2 = 1 + (Align>2) + (Align>4) + (Align>8) + (Align>16), order_start = Align_log2 - 1, pad = Align - 1 @@ -37,7 +114,7 @@ static void print_slheap(int level, void *ctx); #ifdef NO_THREADS static struct slab_heap *slheap; # define SET_MEMCTX(thrctx, memctx, sfree) ((void) (slheap = (memctx))) -# define GET_MEMCTX(thrctx, memctxp) (*(memctxp) = slheap)) +# define GET_MEMCTX(thrctx, memctxp) (*(memctxp) = slheap) #else # define memctx_key ((void *) slap_sl_mem_init) # define SET_MEMCTX(thrctx, memctx, kfree) \ @@ -102,20 +179,11 @@ void slap_sl_mem_init() { assert( Align == 1 << Align_log2 ); - /* Adding head+tail preserves alignment */ - assert( 2*sizeof(ber_len_t) % Align == 0 ); ber_set_option( NULL, LBER_OPT_MEMORY_FNS, &slap_sl_mfuncs ); } -/* This allocator always returns memory aligned on a 2-int boundary. - * - * The stack-based allocator stores the size as a ber_len_t at both - * the head and tail of the allocated block. When freeing a block, the - * tail length is ORed with 1 to mark it as free. Freed space can only - * be reclaimed from the tail forward. If the tail block is never freed, - * nothing else will be reclaimed until the slab is reset... - */ +/* Create, reset or just return the memory context of the current thread. */ void * slap_sl_mem_create( ber_len_t size, @@ -128,38 +196,45 @@ slap_sl_mem_create( struct slab_heap *sh; ber_len_t size_shift; struct slab_object *so; + char *base, *newptr; + enum { Base_offset = (unsigned) -sizeof(ber_len_t) % Align }; sh = GET_MEMCTX(thrctx, &memctx); if ( sh && !new ) return sh; - /* round up to doubleword boundary */ - size = (size + Align-1) & -Align; + /* Round up to doubleword boundary, then make room for initial + * padding, preserving expected available size for pool version */ + size = ((size + Align-1) & -Align) + Base_offset; if (!sh) { sh = ch_malloc(sizeof(struct slab_heap)); - sh->sh_base = ch_malloc(size); + base = ch_malloc(size); SET_MEMCTX(thrctx, sh, slap_sl_mem_destroy); + VGMEMP_MARK(base, size); + VGMEMP_CREATE(sh, 0, 0); } else { slap_sl_mem_destroy(NULL, sh); - if ( size > (char *)sh->sh_end - (char *)sh->sh_base ) { - void *newptr; - - newptr = ch_realloc( sh->sh_base, size ); + base = sh->sh_base; + if (size > (ber_len_t) ((char *) sh->sh_end - base)) { + newptr = ch_realloc(base, size); if ( newptr == NULL ) return NULL; - sh->sh_base = newptr; + VGMEMP_CHANGE(sh, base, newptr, size); + base = newptr; } + VGMEMP_TRIM(sh, base, 0); } - sh->sh_end = (char *) sh->sh_base + size; + sh->sh_base = base; + sh->sh_end = base + size; + + /* Align (base + head of first block) == first returned block */ + base += Base_offset; + size -= Base_offset; sh->sh_stack = stack; if (stack) { - /* insert dummy len */ - { - ber_len_t *i = sh->sh_base; - *i++ = 0; - sh->sh_last = i; - } + sh->sh_last = base; + } else { int i, order = -1, order_end = -1; @@ -183,7 +258,7 @@ slap_sl_mem_create( } so = LDAP_LIST_FIRST(&sh->sh_sopool); LDAP_LIST_REMOVE(so, so_link); - so->so_ptr = sh->sh_base; + so->so_ptr = base; LDAP_LIST_INSERT_HEAD(&sh->sh_free[order-1], so, so_link); @@ -199,16 +274,20 @@ slap_sl_mem_create( memset(sh->sh_map[i], 0, nummaps); } } + return sh; } +/* + * Separate memory context from thread context. Future users must + * know the context, since ch_free/slap_sl_context() cannot find it. + */ void slap_sl_mem_detach( void *thrctx, void *memctx ) { - /* separate from context */ SET_MEMCTX(thrctx, NULL, 0); } @@ -221,15 +300,8 @@ slap_sl_malloc( struct slab_heap *sh = ctx; ber_len_t *ptr, *newptr; -#ifdef SLAP_NO_SL_MALLOC - newptr = ber_memalloc_x( size, NULL ); - if ( newptr ) return newptr; - assert( 0 ); - exit( EXIT_FAILURE ); -#endif - /* ber_set_option calls us like this */ - if (!ctx) { + if (No_sl_malloc || !ctx) { newptr = ber_memalloc_x( size, NULL ); if ( newptr ) return newptr; Debug(LDAP_DEBUG_ANY, "slap_sl_malloc of %lu bytes failed\n", @@ -238,23 +310,21 @@ slap_sl_malloc( exit( EXIT_FAILURE ); } - /* round up to doubleword boundary, plus space for len at head and tail */ - size = (size + 2*sizeof(ber_len_t) + Align-1) & -Align; + /* Add room for head, ensure room for tail when freed, and + * round up to doubleword boundary. */ + size = (size + sizeof(ber_len_t) + Align-1 + !size) & -Align; if (sh->sh_stack) { - if ((char *)sh->sh_last + size >= (char *)sh->sh_end) { - size -= 2*sizeof(ber_len_t); - Debug(LDAP_DEBUG_TRACE, - "slap_sl_malloc of %lu bytes failed, using ch_malloc\n", - (unsigned long) size, 0, 0); - return ch_malloc(size); + if (size < (ber_len_t) ((char *) sh->sh_end - (char *) sh->sh_last)) { + newptr = sh->sh_last; + sh->sh_last = (char *) sh->sh_last + size; + VGMEMP_ALLOC(sh, newptr, size); + *newptr++ = size; + return( (void *)newptr ); } - newptr = sh->sh_last; - sh->sh_last = (char *) sh->sh_last + size; + size -= sizeof(ber_len_t); - *newptr++ = size; - *(ber_len_t *)((char *)sh->sh_last - sizeof(ber_len_t)) = size; - return( (void *)newptr ); + } else { struct slab_object *so_new, *so_left, *so_right; ber_len_t size_shift; @@ -314,19 +384,30 @@ slap_sl_malloc( } Debug(LDAP_DEBUG_TRACE, - "slap_sl_malloc of %lu bytes failed, using ch_malloc\n", + "sl_malloc %lu: ch_malloc\n", (unsigned long) size, 0, 0); return ch_malloc(size); } +#define LIM_SQRT(t) /* some value < sqrt(max value of unsigned type t) */ \ + ((0UL|(t)-1) >>31>>31 > 1 ? ((t)1 <<32) - 1 : \ + (0UL|(t)-1) >>31 ? 65535U : (0UL|(t)-1) >>15 ? 255U : 15U) + void * slap_sl_calloc( ber_len_t n, ber_len_t size, void *ctx ) { void *newptr; + ber_len_t total = n * size; - newptr = slap_sl_malloc( n*size, ctx ); - if ( newptr ) { + /* The sqrt test is a slight optimization: often avoids the division */ + if ((n | size) <= LIM_SQRT(ber_len_t) || n == 0 || total/n == size) { + newptr = slap_sl_malloc( total, ctx ); memset( newptr, 0, n*size ); + } else { + Debug(LDAP_DEBUG_ANY, "slap_sl_calloc(%lu,%lu) out of range\n", + (unsigned long) n, (unsigned long) size, 0); + assert(0); + exit(EXIT_FAILURE); } return newptr; } @@ -335,27 +416,20 @@ void * slap_sl_realloc(void *ptr, ber_len_t size, void *ctx) { struct slab_heap *sh = ctx; - ber_len_t oldsize, *p = (ber_len_t *) ptr; + ber_len_t oldsize, *p = (ber_len_t *) ptr, *nextp; void *newptr; if (ptr == NULL) return slap_sl_malloc(size, ctx); -#ifdef SLAP_NO_SL_MALLOC - newptr = ber_memrealloc_x( ptr, size, NULL ); - if ( newptr ) return newptr; - assert( 0 ); - exit( EXIT_FAILURE ); -#endif - /* Not our memory? */ - if (!sh || ptr < sh->sh_base || ptr >= sh->sh_end) { + if (No_sl_malloc || !sh || ptr < sh->sh_base || ptr >= sh->sh_end) { /* Like ch_realloc(), except not trying a new context */ newptr = ber_memrealloc_x(ptr, size, NULL); if (newptr) { return newptr; } - Debug(LDAP_DEBUG_ANY, "ch_realloc of %lu bytes failed\n", + Debug(LDAP_DEBUG_ANY, "slap_sl_realloc of %lu bytes failed\n", (unsigned long) size, 0, 0); assert(0); exit( EXIT_FAILURE ); @@ -369,33 +443,35 @@ slap_sl_realloc(void *ptr, ber_len_t size, void *ctx) oldsize = p[-1]; if (sh->sh_stack) { - /* Round up to doubleword boundary, add room for head */ - size = ((size + Align-1) & -Align) + sizeof( ber_len_t ); + /* Add room for head, round up to doubleword boundary */ + size = (size + sizeof(ber_len_t) + Align-1) & -Align; p--; /* Never shrink blocks */ if (size <= oldsize) { return ptr; + } + oldsize &= -2; + nextp = (ber_len_t *) ((char *) p + oldsize); + /* If reallocing the last block, try to grow it */ - } else if ((char *) ptr + oldsize == sh->sh_last) { - if (size < (char *) sh->sh_end - (char *) ptr) { - sh->sh_last = (char *) ptr + size; - p[0] = size; - p[size/sizeof(ber_len_t)] = size; + if (nextp == sh->sh_last) { + if (size < (ber_len_t) ((char *) sh->sh_end - (char *) p)) { + sh->sh_last = (char *) p + size; + p[0] = (p[0] & 1) | size; return ptr; } /* Nowhere to grow, need to alloc and copy */ } else { /* Slight optimization of the final realloc variant */ - size -= sizeof(ber_len_t); - oldsize -= sizeof(ber_len_t); - newptr = slap_sl_malloc(size, ctx); - AC_MEMCPY(newptr, ptr, oldsize); + newptr = slap_sl_malloc(size-sizeof(ber_len_t), ctx); + AC_MEMCPY(newptr, ptr, oldsize-sizeof(ber_len_t)); /* Not last block, can just mark old region as free */ - p[p[0]/sizeof(ber_len_t)] |= 1; + nextp[-1] = oldsize; + nextp[0] |= 1; return newptr; } @@ -417,31 +493,37 @@ slap_sl_free(void *ptr, void *ctx) { struct slab_heap *sh = ctx; ber_len_t size; - ber_len_t *p = (ber_len_t *)ptr, *tmpp; + ber_len_t *p = ptr, *nextp, *tmpp; if (!ptr) return; -#ifdef SLAP_NO_SL_MALLOC - ber_memfree_x( ptr, NULL ); - return; -#endif - - if (!sh || ptr < sh->sh_base || ptr >= sh->sh_end) { + if (No_sl_malloc || !sh || ptr < sh->sh_base || ptr >= sh->sh_end) { ber_memfree_x(ptr, NULL); + return; + } + + size = *(--p); - } else if (sh->sh_stack) { - size = p[-1]; - p = (ber_len_t *) ((char *) ptr + size); - /* mark it free */ - p[-1] = size |= 1; - /* reclaim free space off tail */ - if (sh->sh_last == p) { - do { - p = (ber_len_t *) ((char *) p - size + 1) - 1; - size = p[-1]; - } while (size & 1); + if (sh->sh_stack) { + size &= -2; + nextp = (ber_len_t *) ((char *) p + size); + if (sh->sh_last != nextp) { + /* Mark it free: tail = size, head of next block |= 1 */ + nextp[-1] = size; + nextp[0] |= 1; + /* We can't tell Valgrind about it yet, because we + * still need read/write access to this block for + * when we eventually get to reclaim it. + */ + } else { + /* Reclaim freed block(s) off tail */ + while (*p & 1) { + p = (ber_len_t *) ((char *) p - p[-1]); + } sh->sh_last = p; + VGMEMP_TRIM(sh, sh->sh_base, + (char *) sh->sh_last - (char *) sh->sh_base); } } else { @@ -450,7 +532,6 @@ slap_sl_free(void *ptr, void *ctx) unsigned long diff; int i, inserted = 0, order = -1; - size = *(--p); size_shift = size + sizeof(ber_len_t) - 1; do { order++; @@ -567,6 +648,10 @@ slap_sl_free(void *ptr, void *ctx) } } +/* + * Return the memory context of the current thread if the given block of + * memory belongs to it, otherwise return NULL. + */ void * slap_sl_context( void *ptr ) {