X-Git-Url: https://git.sur5r.net/?a=blobdiff_plain;f=servers%2Fslapd%2Fsl_malloc.c;h=fbc34ff659623a180cd63b63a5d0a591144682cc;hb=6f6ee3eada75d0ea8bee306c7953417f601c5bb8;hp=a295483068cefd5fecf7bb17e03cf65a96dffa77;hpb=f416cb91a84b59704a43ddb6383c5e1a2f9ebf75;p=openldap diff --git a/servers/slapd/sl_malloc.c b/servers/slapd/sl_malloc.c index a295483068..fbc34ff659 100644 --- a/servers/slapd/sl_malloc.c +++ b/servers/slapd/sl_malloc.c @@ -2,7 +2,7 @@ /* $OpenLDAP$ */ /* This work is part of OpenLDAP Software . * - * Copyright 2003-2010 The OpenLDAP Foundation. + * Copyright 2003-2013 The OpenLDAP Foundation. * All rights reserved. * * Redistribution and use in source and binary forms, with or without @@ -21,11 +21,112 @@ #include "slap.h" +#ifdef USE_VALGRIND +/* Get debugging help from Valgrind */ +#include +#define VGMEMP_MARK(m,s) VALGRIND_MAKE_MEM_NOACCESS(m,s) +#define VGMEMP_CREATE(h,r,z) VALGRIND_CREATE_MEMPOOL(h,r,z) +#define VGMEMP_TRIM(h,a,s) VALGRIND_MEMPOOL_TRIM(h,a,s) +#define VGMEMP_ALLOC(h,a,s) VALGRIND_MEMPOOL_ALLOC(h,a,s) +#define VGMEMP_CHANGE(h,a,b,s) VALGRIND_MEMPOOL_CHANGE(h,a,b,s) +#else +#define VGMEMP_MARK(m,s) +#define VGMEMP_CREATE(h,r,z) +#define VGMEMP_TRIM(h,a,s) +#define VGMEMP_ALLOC(h,a,s) +#define VGMEMP_CHANGE(h,a,b,s) +#endif + +/* + * This allocator returns temporary memory from a slab in a given memory + * context, aligned on a 2-int boundary. It cannot be used for data + * which will outlive the task allocating it. + * + * A new memory context attaches to the creator's thread context, if any. + * Threads cannot use other threads' memory contexts; there are no locks. + * + * The caller of slap_sl_malloc, usually a thread pool task, must + * slap_sl_free the memory before finishing: New tasks reuse the context + * and normally reset it, reclaiming memory left over from last task. + * + * The allocator helps memory fragmentation, speed and memory leaks. + * It is not (yet) reliable as a garbage collector: + * + * It falls back to context NULL - plain ber_memalloc() - when the + * context's slab is full. A reset does not reclaim such memory. + * Conversely, free/realloc of data not from the given context assumes + * context NULL. The data must not belong to another memory context. + * + * Code which has lost track of the current memory context can try + * slap_sl_context() or ch_malloc.c:ch_free/ch_realloc(). + * + * Allocations cannot yet return failure. Like ch_malloc, they succeed + * or abort slapd. This will change, do fix code which assumes success. + */ + +/* + * The stack-based allocator stores (ber_len_t)sizeof(head+block) at + * allocated blocks' head - and in freed blocks also at the tail, marked + * by ORing *next* block's head with 1. Freed blocks are only reclaimed + * from the last block forward. This is fast, but when a block is never + * freed, older blocks will not be reclaimed until the slab is reset... + */ + +#ifdef SLAP_NO_SL_MALLOC /* Useful with memory debuggers like Valgrind */ +enum { No_sl_malloc = 1 }; +#else +enum { No_sl_malloc = 0 }; +#endif + +#define SLAP_SLAB_SOBLOCK 64 + +struct slab_object { + void *so_ptr; + int so_blockhead; + LDAP_LIST_ENTRY(slab_object) so_link; +}; + +struct slab_heap { + void *sh_base; + void *sh_last; + void *sh_end; + int sh_stack; + int sh_maxorder; + unsigned char **sh_map; + LDAP_LIST_HEAD(sh_freelist, slab_object) *sh_free; + LDAP_LIST_HEAD(sh_so, slab_object) sh_sopool; +}; + +enum { + Align = sizeof(ber_len_t) > 2*sizeof(int) + ? sizeof(ber_len_t) : 2*sizeof(int), + Align_log2 = 1 + (Align>2) + (Align>4) + (Align>8) + (Align>16), + order_start = Align_log2 - 1, + pad = Align - 1 +}; + static struct slab_object * slap_replenish_sopool(struct slab_heap* sh); #ifdef SLAPD_UNUSED static void print_slheap(int level, void *ctx); #endif +/* Keep memory context in a thread-local var, or in a global when no threads */ +#ifdef NO_THREADS +static struct slab_heap *slheap; +# define SET_MEMCTX(thrctx, memctx, sfree) ((void) (slheap = (memctx))) +# define GET_MEMCTX(thrctx, memctxp) (*(memctxp) = slheap) +#else +# define memctx_key ((void *) slap_sl_mem_init) +# define SET_MEMCTX(thrctx, memctx, kfree) \ + ldap_pvt_thread_pool_setkey(thrctx,memctx_key, memctx,kfree, NULL,NULL) +# define GET_MEMCTX(thrctx, memctxp) \ + ((void) (*(memctxp) = NULL), \ + (void) ldap_pvt_thread_pool_getkey(thrctx,memctx_key, memctxp,NULL), \ + *(memctxp)) +#endif /* NO_THREADS */ + + +/* Destroy the context, or if key==NULL clean it up for reuse. */ void slap_sl_mem_destroy( void *key, @@ -33,19 +134,10 @@ slap_sl_mem_destroy( ) { struct slab_heap *sh = data; - int pad = 2*sizeof(int)-1, pad_shift; - int order_start = -1, i; struct slab_object *so; + int i; - if (sh->sh_stack) { - ber_memfree_x(sh->sh_base, NULL); - ber_memfree_x(sh, NULL); - } else { - pad_shift = pad - 1; - do { - order_start++; - } while (pad_shift >>= 1); - + if (!sh->sh_stack) { for (i = 0; i <= sh->sh_maxorder - order_start; i++) { so = LDAP_LIST_FIRST(&sh->sh_free[i]); while (so) { @@ -72,6 +164,9 @@ slap_sl_mem_destroy( so = LDAP_LIST_NEXT(so, so_link); ch_free(so_tmp); } + } + + if (key != NULL) { ber_memfree_x(sh->sh_base, NULL); ber_memfree_x(sh, NULL); } @@ -83,137 +178,71 @@ BerMemoryFunctions slap_sl_mfuncs = void slap_sl_mem_init() { + assert( Align == 1 << Align_log2 ); + ber_set_option( NULL, LBER_OPT_MEMORY_FNS, &slap_sl_mfuncs ); } -#ifdef NO_THREADS -static struct slab_heap *slheap; -#endif - -/* This allocator always returns memory aligned on a 2-int boundary. - * - * The stack-based allocator stores the size as a ber_len_t at both - * the head and tail of the allocated block. When freeing a block, the - * tail length is ORed with 1 to mark it as free. Freed space can only - * be reclaimed from the tail forward. If the tail block is never freed, - * nothing else will be reclaimed until the slab is reset... - */ +/* Create, reset or just return the memory context of the current thread. */ void * slap_sl_mem_create( ber_len_t size, int stack, - void *ctx, + void *thrctx, int new ) { + void *memctx; struct slab_heap *sh; ber_len_t size_shift; - int pad = 2*sizeof(int)-1, pad_shift; - int order = -1, order_start = -1, order_end = -1; - int i; struct slab_object *so; + char *base, *newptr; + enum { Base_offset = (unsigned) -sizeof(ber_len_t) % Align }; -#ifdef NO_THREADS - sh = slheap; -#else - void *sh_tmp = NULL; - ldap_pvt_thread_pool_getkey( - ctx, (void *)slap_sl_mem_init, &sh_tmp, NULL ); - sh = sh_tmp; -#endif - + sh = GET_MEMCTX(thrctx, &memctx); if ( sh && !new ) return sh; - /* round up to doubleword boundary */ - size += pad; - size &= ~pad; + /* Round up to doubleword boundary, then make room for initial + * padding, preserving expected available size for pool version */ + size = ((size + Align-1) & -Align) + Base_offset; - if (stack) { - if (!sh) { - sh = ch_malloc(sizeof(struct slab_heap)); - sh->sh_base = ch_malloc(size); -#ifdef NO_THREADS - slheap = sh; -#else - ldap_pvt_thread_pool_setkey(ctx, (void *)slap_sl_mem_init, - (void *)sh, slap_sl_mem_destroy, NULL, NULL); -#endif - } else if ( size > (char *)sh->sh_end - (char *)sh->sh_base ) { - void *newptr; - - newptr = ch_realloc( sh->sh_base, size ); + if (!sh) { + sh = ch_malloc(sizeof(struct slab_heap)); + base = ch_malloc(size); + SET_MEMCTX(thrctx, sh, slap_sl_mem_destroy); + VGMEMP_MARK(base, size); + VGMEMP_CREATE(sh, 0, 0); + } else { + slap_sl_mem_destroy(NULL, sh); + base = sh->sh_base; + if (size > (ber_len_t) ((char *) sh->sh_end - base)) { + newptr = ch_realloc(base, size); if ( newptr == NULL ) return NULL; - sh->sh_base = newptr; + VGMEMP_CHANGE(sh, base, newptr, size); + base = newptr; } - /* insert dummy len */ - { - ber_len_t *i = sh->sh_base; - *i++ = 0; - sh->sh_last = i; - } - sh->sh_end = (char *) sh->sh_base + size; - sh->sh_stack = stack; - return sh; + VGMEMP_TRIM(sh, base, 0); + } + sh->sh_base = base; + sh->sh_end = base + size; + + /* Align (base + head of first block) == first returned block */ + base += Base_offset; + size -= Base_offset; + + sh->sh_stack = stack; + if (stack) { + sh->sh_last = base; + } else { + int i, order = -1, order_end = -1; + size_shift = size - 1; do { order_end++; } while (size_shift >>= 1); - - pad_shift = pad - 1; - do { - order_start++; - } while (pad_shift >>= 1); - order = order_end - order_start + 1; - - if (!sh) { - sh = (struct slab_heap *) ch_malloc(sizeof(struct slab_heap)); - sh->sh_base = ch_malloc(size); -#ifdef NO_THREADS - slheap = sh; -#else - ldap_pvt_thread_pool_setkey(ctx, (void *)slap_sl_mem_init, - (void *)sh, slap_sl_mem_destroy, NULL, NULL); -#endif - } else { - for (i = 0; i <= sh->sh_maxorder - order_start; i++) { - so = LDAP_LIST_FIRST(&sh->sh_free[i]); - while (so) { - struct slab_object *so_tmp = so; - so = LDAP_LIST_NEXT(so, so_link); - LDAP_LIST_INSERT_HEAD(&sh->sh_sopool, so_tmp, so_link); - } - ch_free(sh->sh_map[i]); - } - ch_free(sh->sh_free); - ch_free(sh->sh_map); - - so = LDAP_LIST_FIRST(&sh->sh_sopool); - while (so) { - struct slab_object *so_tmp = so; - so = LDAP_LIST_NEXT(so, so_link); - if (!so_tmp->so_blockhead) { - LDAP_LIST_REMOVE(so_tmp, so_link); - } - } - so = LDAP_LIST_FIRST(&sh->sh_sopool); - while (so) { - struct slab_object *so_tmp = so; - so = LDAP_LIST_NEXT(so, so_link); - ch_free(so_tmp); - } - - if (size > (char *)sh->sh_end - (char *)sh->sh_base) { - void *newptr; - - newptr = ch_realloc( sh->sh_base, size ); - if ( newptr == NULL ) return NULL; - sh->sh_base = newptr; - } - } - sh->sh_end = (char *)sh->sh_base + size; sh->sh_maxorder = order_end; sh->sh_free = (struct sh_freelist *) @@ -229,7 +258,7 @@ slap_sl_mem_create( } so = LDAP_LIST_FIRST(&sh->sh_sopool); LDAP_LIST_REMOVE(so, so_link); - so->so_ptr = sh->sh_base; + so->so_ptr = base; LDAP_LIST_INSERT_HEAD(&sh->sh_free[order-1], so, so_link); @@ -244,24 +273,22 @@ slap_sl_mem_create( sh->sh_map[i] = (unsigned char *) ch_malloc(nummaps); memset(sh->sh_map[i], 0, nummaps); } - sh->sh_stack = stack; - return sh; } + + return sh; } +/* + * Separate memory context from thread context. Future users must + * know the context, since ch_free/slap_sl_context() cannot find it. + */ void slap_sl_mem_detach( - void *ctx, + void *thrctx, void *memctx ) { -#ifdef NO_THREADS - slheap = NULL; -#else - /* separate from context */ - ldap_pvt_thread_pool_setkey( ctx, (void *)slap_sl_mem_init, - NULL, 0, NULL, NULL ); -#endif + SET_MEMCTX(thrctx, NULL, 0); } void * @@ -271,57 +298,45 @@ slap_sl_malloc( ) { struct slab_heap *sh = ctx; - int pad = 2*sizeof(int)-1, pad_shift; ber_len_t *ptr, *newptr; -#ifdef SLAP_NO_SL_MALLOC - newptr = ber_memalloc_x( size, NULL ); - if ( newptr ) return newptr; - assert( 0 ); - exit( EXIT_FAILURE ); -#endif - /* ber_set_option calls us like this */ - if (!ctx) { + if (No_sl_malloc || !ctx) { newptr = ber_memalloc_x( size, NULL ); if ( newptr ) return newptr; + Debug(LDAP_DEBUG_ANY, "slap_sl_malloc of %lu bytes failed\n", + (unsigned long) size, 0, 0); assert( 0 ); exit( EXIT_FAILURE ); } - /* round up to doubleword boundary, plus space for len at head and tail */ - size += 2*sizeof(ber_len_t) + pad; - size &= ~pad; + /* Add room for head, ensure room for tail when freed, and + * round up to doubleword boundary. */ + size = (size + sizeof(ber_len_t) + Align-1 + !size) & -Align; if (sh->sh_stack) { - if ((char *)sh->sh_last + size >= (char *)sh->sh_end) { - Debug(LDAP_DEBUG_TRACE, - "slap_sl_malloc of %lu bytes failed, using ch_malloc\n", - (long)size, 0, 0); - return ch_malloc(size); + if (size < (ber_len_t) ((char *) sh->sh_end - (char *) sh->sh_last)) { + newptr = sh->sh_last; + sh->sh_last = (char *) sh->sh_last + size; + VGMEMP_ALLOC(sh, newptr, size); + *newptr++ = size; + return( (void *)newptr ); } - newptr = sh->sh_last; - sh->sh_last = (char *) sh->sh_last + size; + size -= sizeof(ber_len_t); - *newptr++ = size; - *(ber_len_t *)((char *)sh->sh_last - sizeof(ber_len_t)) = size; - return( (void *)newptr ); + } else { struct slab_object *so_new, *so_left, *so_right; ber_len_t size_shift; - int order = -1, order_start = -1; unsigned long diff; - int i, j; + int i, j, order = -1; size_shift = size - 1; do { order++; } while (size_shift >>= 1); - pad_shift = pad - 1; - do { - order_start++; - } while (pad_shift >>= 1); + size -= sizeof(ber_len_t); for (i = order; i <= sh->sh_maxorder && LDAP_LIST_EMPTY(&sh->sh_free[i-order_start]); i++); @@ -333,7 +348,7 @@ slap_sl_malloc( diff = (unsigned long)((char*)ptr - (char*)sh->sh_base) >> (order + 1); sh->sh_map[order-order_start][diff>>3] |= (1 << (diff & 0x7)); - *ptr++ = size - sizeof(ber_len_t); + *ptr++ = size; LDAP_LIST_INSERT_HEAD(&sh->sh_sopool, so_new, so_link); return((void*)ptr); } else if (i <= sh->sh_maxorder) { @@ -352,7 +367,7 @@ slap_sl_malloc( (char*)sh->sh_base) >> (order+1); sh->sh_map[order-order_start][diff>>3] |= (1 << (diff & 0x7)); - *ptr++ = size - sizeof(ber_len_t); + *ptr++ = size; LDAP_LIST_INSERT_HEAD( &sh->sh_free[j-1-order_start], so_right, so_link); LDAP_LIST_INSERT_HEAD(&sh->sh_sopool, so_left, so_link); @@ -364,26 +379,35 @@ slap_sl_malloc( &sh->sh_free[j-1-order_start], so_left, so_link); } } - } else { - Debug( LDAP_DEBUG_TRACE, - "slap_sl_malloc of %lu bytes, falling back to ch_malloc\n", - (long)size, 0, 0); - return (void*)ch_malloc(size); } + /* FIXME: missing return; guessing we failed... */ } - /* FIXME: missing return; guessing... */ - return NULL; + Debug(LDAP_DEBUG_TRACE, + "sl_malloc %lu: ch_malloc\n", + (unsigned long) size, 0, 0); + return ch_malloc(size); } +#define LIM_SQRT(t) /* some value < sqrt(max value of unsigned type t) */ \ + ((0UL|(t)-1) >>31>>31 > 1 ? ((t)1 <<32) - 1 : \ + (0UL|(t)-1) >>31 ? 65535U : (0UL|(t)-1) >>15 ? 255U : 15U) + void * slap_sl_calloc( ber_len_t n, ber_len_t size, void *ctx ) { void *newptr; + ber_len_t total = n * size; - newptr = slap_sl_malloc( n*size, ctx ); - if ( newptr ) { + /* The sqrt test is a slight optimization: often avoids the division */ + if ((n | size) <= LIM_SQRT(ber_len_t) || n == 0 || total/n == size) { + newptr = slap_sl_malloc( total, ctx ); memset( newptr, 0, n*size ); + } else { + Debug(LDAP_DEBUG_ANY, "slap_sl_calloc(%lu,%lu) out of range\n", + (unsigned long) n, (unsigned long) size, 0); + assert(0); + exit(EXIT_FAILURE); } return newptr; } @@ -392,28 +416,21 @@ void * slap_sl_realloc(void *ptr, ber_len_t size, void *ctx) { struct slab_heap *sh = ctx; - int pad = 2*sizeof(int) -1; - ber_len_t *p = (ber_len_t *)ptr, *newptr; + ber_len_t oldsize, *p = (ber_len_t *) ptr, *nextp; + void *newptr; if (ptr == NULL) return slap_sl_malloc(size, ctx); -#ifdef SLAP_NO_SL_MALLOC - newptr = ber_memrealloc_x( ptr, size, NULL ); - if ( newptr ) return newptr; - assert( 0 ); - exit( EXIT_FAILURE ); -#endif - /* Not our memory? */ - if (!sh || ptr < sh->sh_base || ptr >= sh->sh_end) { - /* duplicate of realloc behavior, oh well */ + if (No_sl_malloc || !sh || ptr < sh->sh_base || ptr >= sh->sh_end) { + /* Like ch_realloc(), except not trying a new context */ newptr = ber_memrealloc_x(ptr, size, NULL); if (newptr) { return newptr; } - Debug(LDAP_DEBUG_ANY, "ch_realloc of %lu bytes failed\n", - (long) size, 0, 0); + Debug(LDAP_DEBUG_ANY, "slap_sl_realloc of %lu bytes failed\n", + (unsigned long) size, 0, 0); assert(0); exit( EXIT_FAILURE ); } @@ -423,45 +440,52 @@ slap_sl_realloc(void *ptr, ber_len_t size, void *ctx) return NULL; } + oldsize = p[-1]; + if (sh->sh_stack) { - /* round up to doubleword boundary */ - size += pad + sizeof( ber_len_t ); - size &= ~pad; + /* Add room for head, round up to doubleword boundary */ + size = (size + sizeof(ber_len_t) + Align-1) & -Align; p--; /* Never shrink blocks */ - if (size <= p[0]) { - newptr = ptr; + if (size <= oldsize) { + return ptr; + } - /* If reallocing the last block, we can grow it */ - } else if ((char *)ptr + p[0] == sh->sh_last && - (char *)ptr + size < (char *)sh->sh_end ) { - newptr = ptr; - sh->sh_last = (char *)ptr + size; - p[0] = size; - p[size/sizeof(ber_len_t)] = size; + oldsize &= -2; + nextp = (ber_len_t *) ((char *) p + oldsize); + + /* If reallocing the last block, try to grow it */ + if (nextp == sh->sh_last) { + if (size < (ber_len_t) ((char *) sh->sh_end - (char *) p)) { + sh->sh_last = (char *) p + size; + p[0] = (p[0] & 1) | size; + return ptr; + } /* Nowhere to grow, need to alloc and copy */ } else { + /* Slight optimization of the final realloc variant */ newptr = slap_sl_malloc(size-sizeof(ber_len_t), ctx); - AC_MEMCPY(newptr, ptr, p[0]-sizeof(ber_len_t)); - /* mark old region as free */ - p[p[0]/sizeof(ber_len_t)] |= 1; + AC_MEMCPY(newptr, ptr, oldsize-sizeof(ber_len_t)); + /* Not last block, can just mark old region as free */ + nextp[-1] = oldsize; + nextp[0] |= 1; + return newptr; } - return newptr; - } else { - void *newptr2; - newptr2 = slap_sl_malloc(size, ctx); - if (size < p[-1]) { - AC_MEMCPY(newptr2, ptr, size); - } else { - AC_MEMCPY(newptr2, ptr, p[-1]); - } - slap_sl_free(ptr, ctx); - return newptr2; + size -= sizeof(ber_len_t); + oldsize -= sizeof(ber_len_t); + + } else if (oldsize > size) { + oldsize = size; } + + newptr = slap_sl_malloc(size, ctx); + AC_MEMCPY(newptr, ptr, oldsize); + slap_sl_free(ptr, ctx); + return newptr; } void @@ -469,54 +493,50 @@ slap_sl_free(void *ptr, void *ctx) { struct slab_heap *sh = ctx; ber_len_t size; - ber_len_t *p = (ber_len_t *)ptr, *tmpp; + ber_len_t *p = ptr, *nextp, *tmpp; if (!ptr) return; -#ifdef SLAP_NO_SL_MALLOC - ber_memfree_x( ptr, NULL ); - return; -#endif - - if (!sh || ptr < sh->sh_base || ptr >= sh->sh_end) { + if (No_sl_malloc || !sh || ptr < sh->sh_base || ptr >= sh->sh_end) { ber_memfree_x(ptr, NULL); - } else if (sh->sh_stack) { - tmpp = (ber_len_t *)((char *)ptr + p[-1]); - /* mark it free */ - tmpp[-1] |= 1; - /* reclaim free space off tail */ - while ( tmpp == sh->sh_last ) { - if ( tmpp[-1] & 1 ) { - size = tmpp[-1] ^ 1; - ptr = (char *)tmpp - size; - p = (ber_len_t *)ptr; - p--; - sh->sh_last = p; - tmpp = sh->sh_last; - } else { - break; + return; + } + + size = *(--p); + + if (sh->sh_stack) { + size &= -2; + nextp = (ber_len_t *) ((char *) p + size); + if (sh->sh_last != nextp) { + /* Mark it free: tail = size, head of next block |= 1 */ + nextp[-1] = size; + nextp[0] |= 1; + /* We can't tell Valgrind about it yet, because we + * still need read/write access to this block for + * when we eventually get to reclaim it. + */ + } else { + /* Reclaim freed block(s) off tail */ + while (*p & 1) { + p = (ber_len_t *) ((char *) p - p[-1]); } + sh->sh_last = p; + VGMEMP_TRIM(sh, sh->sh_base, + (char *) sh->sh_last - (char *) sh->sh_base); } + } else { int size_shift, order_size; - int pad = 2*sizeof(int)-1, pad_shift; - int order_start = -1, order = -1; struct slab_object *so; unsigned long diff; - int i, inserted = 0; + int i, inserted = 0, order = -1; - size = *(--p); size_shift = size + sizeof(ber_len_t) - 1; do { order++; } while (size_shift >>= 1); - pad_shift = pad - 1; - do { - order_start++; - } while (pad_shift >>= 1); - for (i = order, tmpp = p; i <= sh->sh_maxorder; i++) { order_size = 1 << (i+1); diff = (unsigned long)((char*)tmpp - (char*)sh->sh_base) >> (i+1); @@ -628,25 +648,19 @@ slap_sl_free(void *ptr, void *ctx) } } +/* + * Return the memory context of the current thread if the given block of + * memory belongs to it, otherwise return NULL. + */ void * slap_sl_context( void *ptr ) { + void *memctx; struct slab_heap *sh; - void *ctx, *sh_tmp; if ( slapMode & SLAP_TOOL_MODE ) return NULL; -#ifdef NO_THREADS - sh = slheap; -#else - ctx = ldap_pvt_thread_pool_context(); - - sh_tmp = NULL; - ldap_pvt_thread_pool_getkey( - ctx, (void *)slap_sl_mem_init, &sh_tmp, NULL); - sh = sh_tmp; -#endif - + sh = GET_MEMCTX(ldap_pvt_thread_pool_context(), &memctx); if (sh && ptr >= sh->sh_base && ptr <= sh->sh_end) { return sh; } @@ -683,8 +697,6 @@ static void print_slheap(int level, void *ctx) { struct slab_heap *sh = ctx; - int order_start = -1; - int pad = 2*sizeof(int)-1, pad_shift; struct slab_object *so; int i, j, once = 0; @@ -693,11 +705,6 @@ print_slheap(int level, void *ctx) return; } - pad_shift = pad - 1; - do { - order_start++; - } while (pad_shift >>= 1); - Debug(level, "sh->sh_maxorder=%d\n", sh->sh_maxorder, 0, 0); for (i = order_start; i <= sh->sh_maxorder; i++) { @@ -714,7 +721,7 @@ print_slheap(int level, void *ctx) Debug(level, "free list:\n", 0, 0, 0); so = LDAP_LIST_FIRST(&sh->sh_free[i-order_start]); while (so) { - Debug(level, "%lx\n", (unsigned long) so->so_ptr, 0, 0); + Debug(level, "%p\n", so->so_ptr, 0, 0); so = LDAP_LIST_NEXT(so, so_link); } }