X-Git-Url: http://mj.ucw.cz/gitweb/?a=blobdiff_plain;f=lib%2Fhashtable.h;h=10aee8bb2cfd22ac1281b0e5fc5344636b825f13;hb=d5713848229da7765d89ecc2db81c5f99552f8b4;hp=e12f3772d738230886d5709cf03f5ad8230e2a0a;hpb=ae6bed064b77a1ecfe72ca357152abb6970b70ca;p=libucw.git diff --git a/lib/hashtable.h b/lib/hashtable.h index e12f3772..10aee8bb 100644 --- a/lib/hashtable.h +++ b/lib/hashtable.h @@ -1,7 +1,7 @@ /* * Sherlock Library -- Universal Hash Table * - * (c) 2002 Martin Mares + * (c) 2002--2004 Martin Mares * (c) 2002 Robert Spalek * * This software may be freely distributed and used according to the terms @@ -88,6 +88,7 @@ * HASH_ATOMIC_TYPE=t Atomic values are of type `t' instead of int. * HASH_USE_POOL=pool Allocate all nodes from given mempool. * Collides with delete/remove functions. + * HASH_AUTO_POOL=size Create a pool of the given block size automatically. * * You also get a iterator macro at no extra charge: * @@ -260,21 +261,33 @@ static inline void P(init_data) (P(node) *n UNUSED) #include -#ifndef HASH_GIVE_ALLOC -#ifdef HASH_USE_POOL - -static inline void * P(alloc) (unsigned int size) -{ return mp_alloc_fast(HASH_USE_POOL, size); } +#ifdef HASH_GIVE_ALLOC +/* If the caller has requested to use his own allocation functions, do so */ +static inline void P(init_alloc) (void) { } +static inline void P(cleanup_alloc) (void) { } + +#elif defined(HASH_USE_POOL) +/* If the caller has requested to use his mempool, do so */ +#include "lib/pools.h" +static inline void * P(alloc) (unsigned int size) { return mp_alloc_fast(HASH_USE_POOL, size); } +static inline void P(init_alloc) (void) { } +static inline void P(cleanup_alloc) (void) { } + +#elif defined(HASH_AUTO_POOL) +/* Use our own pools */ +#include "lib/pools.h" +static struct mempool *P(pool); +static inline void * P(alloc) (unsigned int size) { return mp_alloc_fast(P(pool), size); } +static inline void P(init_alloc) (void) { P(pool) = mp_new(HASH_AUTO_POOL); } +static inline void P(cleanup_alloc) (void) { mp_delete(P(pool)); } #else +/* The default allocation method */ +static inline void * P(alloc) (unsigned int size) { return xmalloc(size); } +static inline void P(free) (void *x) { xfree(x); } +static inline void P(init_alloc) (void) { } +static inline void P(cleanup_alloc) (void) { } -static inline void * P(alloc) (unsigned int size) -{ return xmalloc(size); } - -static inline void P(free) (void *x) -{ xfree(x); } - -#endif #endif #ifndef HASH_DEFAULT_SIZE @@ -312,6 +325,7 @@ static void P(init) (void) T.hash_hard_max = 1 << 28; #endif P(alloc_table)(); + P(init_alloc)(); } #ifdef HASH_WANT_CLEANUP @@ -328,6 +342,7 @@ static void P(cleanup) (void) P(free)(b); } #endif + P(cleanup_alloc)(); xfree(T.ht); } #endif @@ -512,15 +527,14 @@ static void P(remove) (P(node) *n) #define HASH_FOR_ALL(h_px, h_var) \ do { \ uns h_slot; \ - struct HASH_GLUE(h_px,bucket) *h_buck; \ - for (h_slot=0; h_slot < HASH_GLUE(h_px,table).hash_size; h_slot++) \ - for (h_buck = HASH_GLUE(h_px,table).ht[h_slot]; h_buck; h_buck = h_buck->next) \ + struct GLUE_(h_px,bucket) *h_buck; \ + for (h_slot=0; h_slot < GLUE_(h_px,table).hash_size; h_slot++) \ + for (h_buck = GLUE_(h_px,table).ht[h_slot]; h_buck; h_buck = h_buck->next) \ { \ - HASH_GLUE(h_px,node) *h_var = &h_buck->n; + GLUE_(h_px,node) *h_var = &h_buck->n; #define HASH_END_FOR } } while(0) #define HASH_BREAK #define HASH_CONTINUE continue -#define HASH_GLUE(x,y) x##_##y #endif @@ -550,6 +564,7 @@ do { \ #undef HASH_NODE #undef HASH_PREFIX #undef HASH_USE_POOL +#undef HASH_AUTO_POOL #undef HASH_WANT_CLEANUP #undef HASH_WANT_DELETE #undef HASH_WANT_FIND