]> git.ozlabs.org Git - ccan/blobdiff - ccan/tdb2/tdb.c
tdb2: get rid of zones
[ccan] / ccan / tdb2 / tdb.c
index 7544d7a1e13ef1bcea9d13ece2a4a8e01be7a3f0..c0561fb9fe1f2511797aa5604d368721a3a57f61 100644 (file)
@@ -1,6 +1,5 @@
 #include "private.h"
 #include <ccan/tdb2/tdb2.h>
-#include <ccan/hash/hash.h>
 #include <ccan/build_assert/build_assert.h>
 #include <ccan/likely/likely.h>
 #include <assert.h>
@@ -11,58 +10,13 @@ struct tdb_data tdb_null = { .dptr = NULL, .dsize = 0 };
 /* all contexts, to ensure no double-opens (fcntl locks don't nest!) */
 static struct tdb_context *tdbs = NULL;
 
-PRINTF_ATTRIBUTE(4, 5) static void
+PRINTF_FMT(4, 5) static void
 null_log_fn(struct tdb_context *tdb,
            enum tdb_debug_level level, void *priv,
            const char *fmt, ...)
 {
 }
 
-/* We do a lot of work assuming our copy of the header volatile area
- * is uptodate, and usually it is.  However, once we grab a lock, we have to
- * re-check it. */
-bool header_changed(struct tdb_context *tdb)
-{
-       uint64_t gen;
-
-       if (!(tdb->flags & TDB_NOLOCK) && tdb->header_uptodate) {
-               tdb->log(tdb, TDB_DEBUG_WARNING, tdb->log_priv,
-                        "warning: header uptodate already\n");
-       }
-
-       /* We could get a partial update if we're not holding any locks. */
-       assert((tdb->flags & TDB_NOLOCK) || tdb_has_locks(tdb));
-
-       tdb->header_uptodate = true;
-       gen = tdb_read_off(tdb, offsetof(struct tdb_header, v.generation));
-       if (unlikely(gen != tdb->header.v.generation)) {
-               tdb_read_convert(tdb, offsetof(struct tdb_header, v),
-                                &tdb->header.v, sizeof(tdb->header.v));
-               return true;
-       }
-       return false;
-}
-
-int write_header(struct tdb_context *tdb)
-{
-       assert(tdb_read_off(tdb, offsetof(struct tdb_header, v.generation))
-              == tdb->header.v.generation);
-       tdb->header.v.generation++;
-       return tdb_write_convert(tdb, offsetof(struct tdb_header, v),
-                                &tdb->header.v, sizeof(tdb->header.v));
-}
-
-static uint64_t jenkins_hash(const void *key, size_t length, uint64_t seed,
-                            void *arg)
-{
-       return hash64_stable((const unsigned char *)key, length, seed);
-}
-
-uint64_t tdb_hash(struct tdb_context *tdb, const void *ptr, size_t len)
-{
-       return tdb->khash(ptr, len, tdb->header.hash_seed, tdb->hash_priv);
-}
-
 static bool tdb_already_open(dev_t device, ino_t ino)
 {
        struct tdb_context *i;
@@ -126,59 +80,49 @@ static uint64_t random_number(struct tdb_context *tdb)
 
 struct new_database {
        struct tdb_header hdr;
-       struct tdb_used_record hrec;
-       tdb_off_t hash[1ULL << INITIAL_HASH_BITS];
-       struct tdb_used_record frec;
-       tdb_off_t free[INITIAL_FREE_BUCKETS + 1]; /* One overflow bucket */
+       struct tdb_freelist flist;
 };
 
 /* initialise a new database */
-static int tdb_new_database(struct tdb_context *tdb)
+static int tdb_new_database(struct tdb_context *tdb,
+                           struct tdb_attribute_seed *seed,
+                           struct tdb_header *hdr)
 {
        /* We make it up in memory, then write it out if not internal */
        struct new_database newdb;
-       unsigned int magic_off = offsetof(struct tdb_header, magic_food);
+       unsigned int magic_len;
 
        /* Fill in the header */
        newdb.hdr.version = TDB_VERSION;
-       newdb.hdr.hash_seed = random_number(tdb);
+       if (seed)
+               newdb.hdr.hash_seed = seed->seed;
+       else
+               newdb.hdr.hash_seed = random_number(tdb);
        newdb.hdr.hash_test = TDB_HASH_MAGIC;
        newdb.hdr.hash_test = tdb->khash(&newdb.hdr.hash_test,
                                         sizeof(newdb.hdr.hash_test),
                                         newdb.hdr.hash_seed,
                                         tdb->hash_priv);
        memset(newdb.hdr.reserved, 0, sizeof(newdb.hdr.reserved));
-       newdb.hdr.v.generation = 0;
-
-       /* The initial zone must cover the initial database size! */
-       BUILD_ASSERT((1ULL << INITIAL_ZONE_BITS) >= sizeof(newdb));
-
-       /* Free array has 1 zone, 10 buckets.  All buckets empty. */
-       newdb.hdr.v.num_zones = 1;
-       newdb.hdr.v.zone_bits = INITIAL_ZONE_BITS;
-       newdb.hdr.v.free_buckets = INITIAL_FREE_BUCKETS;
-       newdb.hdr.v.free_off = offsetof(struct new_database, free);
-       set_header(tdb, &newdb.frec, 0,
-                  sizeof(newdb.free), sizeof(newdb.free), 0);
-       memset(newdb.free, 0, sizeof(newdb.free));
-
        /* Initial hashes are empty. */
-       newdb.hdr.v.hash_bits = INITIAL_HASH_BITS;
-       newdb.hdr.v.hash_off = offsetof(struct new_database, hash);
-       set_header(tdb, &newdb.hrec, 0,
-                  sizeof(newdb.hash), sizeof(newdb.hash), 0);
-       memset(newdb.hash, 0, sizeof(newdb.hash));
+       memset(newdb.hdr.hashtable, 0, sizeof(newdb.hdr.hashtable));
+
+       /* Free is empty. */
+       newdb.hdr.free_list = offsetof(struct new_database, flist);
+       memset(&newdb.flist, 0, sizeof(newdb.flist));
+       set_header(NULL, &newdb.flist.hdr, 0,
+                  sizeof(newdb.flist.buckets), sizeof(newdb.flist.buckets), 1);
 
        /* Magic food */
        memset(newdb.hdr.magic_food, 0, sizeof(newdb.hdr.magic_food));
        strcpy(newdb.hdr.magic_food, TDB_MAGIC_FOOD);
 
        /* This creates an endian-converted database, as if read from disk */
+       magic_len = sizeof(newdb.hdr.magic_food);
        tdb_convert(tdb,
-                   (char *)&newdb.hdr + magic_off,
-                   sizeof(newdb) - magic_off);
+                   (char *)&newdb.hdr + magic_len, sizeof(newdb) - magic_len);
 
-       tdb->header = newdb.hdr;
+       *hdr = newdb.hdr;
 
        if (tdb->flags & TDB_INTERNAL) {
                tdb->map_size = sizeof(newdb);
@@ -212,6 +156,8 @@ struct tdb_context *tdb_open(const char *name, int tdb_flags,
        int save_errno;
        uint64_t hash_test;
        unsigned v;
+       struct tdb_header hdr;
+       struct tdb_attribute_seed *seed = NULL;
 
        tdb = malloc(sizeof(*tdb));
        if (!tdb) {
@@ -221,18 +167,15 @@ struct tdb_context *tdb_open(const char *name, int tdb_flags,
        }
        tdb->name = NULL;
        tdb->map_ptr = NULL;
+       tdb->direct_access = 0;
        tdb->fd = -1;
-       /* map_size will be set below. */
+       tdb->map_size = sizeof(struct tdb_header);
        tdb->ecode = TDB_SUCCESS;
-       /* header will be read in below. */
-       tdb->header_uptodate = false;
        tdb->flags = tdb_flags;
        tdb->log = null_log_fn;
        tdb->log_priv = NULL;
-       tdb->khash = jenkins_hash;
-       tdb->hash_priv = NULL;
        tdb->transaction = NULL;
-       /* last_zone will be set below. */
+       tdb_hash_init(tdb);
        tdb_io_init(tdb);
        tdb_lock_init(tdb);
 
@@ -246,6 +189,9 @@ struct tdb_context *tdb_open(const char *name, int tdb_flags,
                        tdb->khash = attr->hash.hash_fn;
                        tdb->hash_priv = attr->hash.hash_private;
                        break;
+               case TDB_ATTRIBUTE_SEED:
+                       seed = &attr->seed;
+                       break;
                default:
                        tdb->log(tdb, TDB_DEBUG_ERROR, tdb->log_priv,
                                 "tdb_open: unknown attribute type %u\n",
@@ -267,21 +213,23 @@ struct tdb_context *tdb_open(const char *name, int tdb_flags,
                tdb->read_only = true;
                /* read only databases don't do locking */
                tdb->flags |= TDB_NOLOCK;
-       } else
+               tdb->mmap_flags = PROT_READ;
+       } else {
                tdb->read_only = false;
+               tdb->mmap_flags = PROT_READ | PROT_WRITE;
+       }
 
        /* internal databases don't need any of the rest. */
        if (tdb->flags & TDB_INTERNAL) {
                tdb->flags |= (TDB_NOLOCK | TDB_NOMMAP);
-               if (tdb_new_database(tdb) != 0) {
+               if (tdb_new_database(tdb, seed, &hdr) != 0) {
                        tdb->log(tdb, TDB_DEBUG_ERROR, tdb->log_priv,
                                 "tdb_open: tdb_new_database failed!");
                        goto fail;
                }
-               TEST_IT(tdb->flags & TDB_CONVERT);
-               tdb_convert(tdb, &tdb->header, sizeof(tdb->header));
-               /* Zones don't matter for internal db. */
-               tdb->last_zone = 0;
+               tdb_convert(tdb, &hdr.hash_seed, sizeof(hdr.hash_seed));
+               tdb->hash_seed = hdr.hash_seed;
+               tdb_flist_init(tdb);
                return tdb;
        }
 
@@ -304,32 +252,33 @@ struct tdb_context *tdb_open(const char *name, int tdb_flags,
                goto fail;      /* errno set by tdb_brlock */
        }
 
-       if (!tdb_pread_all(tdb->fd, &tdb->header, sizeof(tdb->header), 0)
-           || strcmp(tdb->header.magic_food, TDB_MAGIC_FOOD) != 0) {
-               if (!(open_flags & O_CREAT) || tdb_new_database(tdb) == -1) {
+       if (!tdb_pread_all(tdb->fd, &hdr, sizeof(hdr), 0)
+           || strcmp(hdr.magic_food, TDB_MAGIC_FOOD) != 0) {
+               if (!(open_flags & O_CREAT)
+                   || tdb_new_database(tdb, seed, &hdr) == -1) {
                        if (errno == 0) {
                                errno = EIO; /* ie bad format or something */
                        }
                        goto fail;
                }
-       } else if (tdb->header.version != TDB_VERSION) {
-               if (tdb->header.version == bswap_64(TDB_VERSION))
+       } else if (hdr.version != TDB_VERSION) {
+               if (hdr.version == bswap_64(TDB_VERSION))
                        tdb->flags |= TDB_CONVERT;
                else {
                        /* wrong version */
                        tdb->log(tdb, TDB_DEBUG_ERROR, tdb->log_priv,
                                 "tdb_open: %s is unknown version 0x%llx\n",
-                                name, (long long)tdb->header.version);
+                                name, (long long)hdr.version);
                        errno = EIO;
                        goto fail;
                }
        }
 
-       tdb_convert(tdb, &tdb->header, sizeof(tdb->header));
+       tdb_convert(tdb, &hdr, sizeof(hdr));
+       tdb->hash_seed = hdr.hash_seed;
        hash_test = TDB_HASH_MAGIC;
-       hash_test = tdb->khash(&hash_test, sizeof(hash_test),
-                              tdb->header.hash_seed, tdb->hash_priv);
-       if (tdb->header.hash_test != hash_test) {
+       hash_test = tdb_hash(tdb, &hash_test, sizeof(hash_test));
+       if (hdr.hash_test != hash_test) {
                /* wrong hash variant */
                tdb->log(tdb, TDB_DEBUG_ERROR, tdb->log_priv,
                         "tdb_open: %s uses a different hash function\n",
@@ -357,12 +306,15 @@ struct tdb_context *tdb_open(const char *name, int tdb_flags,
                goto fail;
        }
 
-       tdb->map_size = st.st_size;
        tdb->device = st.st_dev;
        tdb->inode = st.st_ino;
-       tdb_mmap(tdb);
        tdb_unlock_open(tdb);
-       tdb_zone_init(tdb);
+
+       /* This make sure we have current map_size and mmap. */
+       tdb->methods->oob(tdb, tdb->map_size + 1, true);
+
+       if (tdb_flist_init(tdb) == -1)
+               goto fail;
 
        tdb->next = tdbs;
        tdbs = tdb;
@@ -394,145 +346,6 @@ struct tdb_context *tdb_open(const char *name, int tdb_flags,
        return NULL;
 }
 
-tdb_off_t hash_off(struct tdb_context *tdb, uint64_t list)
-{
-       return tdb->header.v.hash_off
-               + ((list & ((1ULL << tdb->header.v.hash_bits) - 1))
-                  * sizeof(tdb_off_t));
-}
-
-/* Returns 0 if the entry is a zero (definitely not a match).
- * Returns a valid entry offset if it's a match.  Fills in rec.
- * Otherwise returns TDB_OFF_ERR: keep searching. */
-static tdb_off_t entry_matches(struct tdb_context *tdb,
-                              uint64_t list,
-                              uint64_t hash,
-                              const struct tdb_data *key,
-                              struct tdb_used_record *rec)
-{
-       tdb_off_t off;
-       uint64_t keylen;
-       const unsigned char *rkey;
-
-       list &= ((1ULL << tdb->header.v.hash_bits) - 1);
-
-       off = tdb_read_off(tdb, tdb->header.v.hash_off
-                          + list * sizeof(tdb_off_t));
-       if (off == 0 || off == TDB_OFF_ERR)
-               return off;
-
-#if 0 /* FIXME: Check other bits. */
-       unsigned int bits, bitmask, hoffextra;
-       /* Bottom three bits show how many extra hash bits. */
-       bits = (off & ((1 << TDB_EXTRA_HASHBITS_NUM) - 1)) + 1;
-       bitmask = (1 << bits)-1;
-       hoffextra = ((off >> TDB_EXTRA_HASHBITS_NUM) & bitmask);
-       uint64_t hextra = hash >> tdb->header.v.hash_bits;
-       if ((hextra & bitmask) != hoffextra) 
-               return TDB_OFF_ERR;
-       off &= ~...;
-#endif
-
-       if (tdb_read_convert(tdb, off, rec, sizeof(*rec)) == -1)
-               return TDB_OFF_ERR;
-
-       /* FIXME: check extra bits in header! */
-       keylen = rec_key_length(rec);
-       if (keylen != key->dsize)
-               return TDB_OFF_ERR;
-
-       rkey = tdb_access_read(tdb, off + sizeof(*rec), keylen, false);
-       if (!rkey)
-               return TDB_OFF_ERR;
-       if (memcmp(rkey, key->dptr, keylen) != 0)
-               off = TDB_OFF_ERR;
-       tdb_access_release(tdb, rkey);
-       return off;
-}
-
-/* FIXME: Optimize? */
-static void unlock_lists(struct tdb_context *tdb,
-                        tdb_off_t list, tdb_len_t num,
-                        int ltype)
-{
-       tdb_off_t i;
-
-       for (i = list; i < list + num; i++)
-               tdb_unlock_list(tdb, i, ltype);
-}
-
-/* FIXME: Optimize? */
-static int lock_lists(struct tdb_context *tdb,
-                     tdb_off_t list, tdb_len_t num,
-                     int ltype)
-{
-       tdb_off_t i;
-
-       for (i = list; i < list + num; i++) {
-               if (tdb_lock_list(tdb, i, ltype, TDB_LOCK_WAIT)
-                   == TDB_OFF_ERR) {
-                       unlock_lists(tdb, list, i - list, ltype);
-                       return -1;
-               }
-       }
-       return 0;
-}
-
-/* We lock hashes up to the next empty offset.  We already hold the
- * lock on the start bucket, but we may need to release and re-grab
- * it.  If we fail, we hold no locks at all! */
-static tdb_len_t relock_hash_to_zero(struct tdb_context *tdb,
-                                    tdb_off_t start, int ltype)
-{
-       tdb_len_t num, len;
-
-again:
-       num = 1ULL << tdb->header.v.hash_bits;
-       len = tdb_find_zero_off(tdb, hash_off(tdb, start), num - start);
-       if (unlikely(len == num - start)) {
-               /* We hit the end of the hash range.  Drop lock: we have
-                  to lock start of hash first. */
-               tdb_len_t pre_locks;
-
-               tdb_unlock_list(tdb, start, ltype);
-
-               /* Grab something, so header is stable. */
-               if (tdb_lock_list(tdb, 0, ltype, TDB_LOCK_WAIT))
-                       return TDB_OFF_ERR;
-               pre_locks = tdb_find_zero_off(tdb, hash_off(tdb, 0), num);
-               /* We want to lock the zero entry as well. */
-               pre_locks++;
-               if (lock_lists(tdb, 1, pre_locks - 1, ltype) == -1) {
-                       tdb_unlock_list(tdb, 0, ltype);
-                       return TDB_OFF_ERR;
-               }
-
-               /* Now lock later ones. */
-               if (unlikely(lock_lists(tdb, start, len, ltype) == -1)) {
-                       unlock_lists(tdb, 0, pre_locks, ltype);
-                       return TDB_OFF_ERR;
-               }
-               len += pre_locks;
-       } else {
-               /* We want to lock the zero entry as well. */
-               len++;
-               /* But we already have lock on start. */
-               if (unlikely(lock_lists(tdb, start+1, len-1, ltype) == -1)) {
-                       tdb_unlock_list(tdb, start, ltype);
-                       return TDB_OFF_ERR;
-               }
-       }
-
-       /* Now, did we lose the race, and it's not zero any more? */
-       if (unlikely(tdb_read_off(tdb, hash_off(tdb, start + len - 1)) != 0)) {
-               /* Leave the start locked, as expected. */
-               unlock_lists(tdb, start + 1, len - 1, ltype);
-               goto again;
-       }
-
-       return len;
-}
-
 /* FIXME: modify, don't rewrite! */
 static int update_rec_hdr(struct tdb_context *tdb,
                          tdb_off_t off,
@@ -549,214 +362,31 @@ static int update_rec_hdr(struct tdb_context *tdb,
        return tdb_write_convert(tdb, off, rec, sizeof(*rec));
 }
 
-static int hash_add(struct tdb_context *tdb,
-                   uint64_t hash, tdb_off_t off)
-{
-       tdb_off_t i, hoff, len, num;
-
-       /* Look for next space. */
-       i = (hash & ((1ULL << tdb->header.v.hash_bits) - 1));
-       len = (1ULL << tdb->header.v.hash_bits) - i;
-       num = tdb_find_zero_off(tdb, hash_off(tdb, i), len);
-
-       if (unlikely(num == len)) {
-               /* We wrapped.  Look through start of hash table. */
-               i = 0;
-               hoff = hash_off(tdb, 0);
-               len = (1ULL << tdb->header.v.hash_bits);
-               num = tdb_find_zero_off(tdb, hoff, len);
-               if (num == len) {
-                       tdb->ecode = TDB_ERR_CORRUPT;
-                       tdb->log(tdb, TDB_DEBUG_FATAL, tdb->log_priv,
-                                "hash_add: full hash table!\n");
-                       return -1;
-               }
-       }
-       if (tdb_read_off(tdb, hash_off(tdb, i + num)) != 0) {
-               tdb->ecode = TDB_ERR_CORRUPT;
-               tdb->log(tdb, TDB_DEBUG_FATAL, tdb->log_priv,
-                        "hash_add: overwriting hash table?\n");
-               return -1;
-       }
-
-       /* FIXME: Encode extra hash bits! */
-       return tdb_write_off(tdb, hash_off(tdb, i + num), off);
-}
-
-/* If we fail, others will try after us. */
-static void enlarge_hash(struct tdb_context *tdb)
-{
-       tdb_off_t newoff, oldoff, i;
-       tdb_len_t hlen;
-       uint64_t num = 1ULL << tdb->header.v.hash_bits;
-       struct tdb_used_record pad, *r;
-       unsigned int records = 0;
-
-       /* FIXME: We should do this without holding locks throughout. */
-       if (tdb_allrecord_lock(tdb, F_WRLCK, TDB_LOCK_WAIT, false) == -1)
-               return;
-
-       /* Someone else enlarged for us?  Nothing to do. */
-       if ((1ULL << tdb->header.v.hash_bits) != num)
-               goto unlock;
-
-again:
-       /* Allocate our new array. */
-       hlen = num * sizeof(tdb_off_t) * 2;
-       newoff = alloc(tdb, 0, hlen, 0, false);
-       if (unlikely(newoff == TDB_OFF_ERR))
-               goto unlock;
-       if (unlikely(newoff == 0)) {
-               if (tdb_expand(tdb, 0, hlen, false) == -1)
-                       goto unlock;
-               goto again;
-       }
-       /* Step over record header! */
-       newoff += sizeof(struct tdb_used_record);
-
-       /* Starts all zero. */
-       if (zero_out(tdb, newoff, hlen) == -1)
-               goto unlock;
-
-       /* Update header now so we can use normal routines. */
-       oldoff = tdb->header.v.hash_off;
-
-       tdb->header.v.hash_bits++;
-       tdb->header.v.hash_off = newoff;
-
-       /* FIXME: If the space before is empty, we know this is in its ideal
-        * location.  Or steal a bit from the pointer to avoid rehash. */
-       for (i = 0; i < num; i++) {
-               tdb_off_t off;
-               off = tdb_read_off(tdb, oldoff + i * sizeof(tdb_off_t));
-               if (unlikely(off == TDB_OFF_ERR))
-                       goto oldheader;
-               if (off && hash_add(tdb, hash_record(tdb, off), off) == -1)
-                       goto oldheader;
-               if (off)
-                       records++;
-       }
-
-       tdb->log(tdb, TDB_DEBUG_TRACE, tdb->log_priv,
-                "enlarge_hash: moved %u records from %llu buckets.\n",
-                records, (long long)num);
-
-       /* Free up old hash. */
-       r = tdb_get(tdb, oldoff - sizeof(*r), &pad, sizeof(*r));
-       if (!r)
-               goto oldheader;
-       add_free_record(tdb, oldoff - sizeof(*r),
-                       sizeof(*r)+rec_data_length(r)+rec_extra_padding(r));
-
-       /* Now we write the modified header. */
-       write_header(tdb);
-unlock:
-       tdb_allrecord_unlock(tdb, F_WRLCK);
-       return;
-
-oldheader:
-       tdb->header.v.hash_bits--;
-       tdb->header.v.hash_off = oldoff;
-       goto unlock;
-}
-
-
-/* This is the slow version of the routine which searches the
- * hashtable for an entry.
- * We lock every hash bucket up to and including the next zero one.
- */
-static tdb_off_t find_and_lock_slow(struct tdb_context *tdb,
-                                   struct tdb_data key,
-                                   uint64_t h,
-                                   int ltype,
-                                   tdb_off_t *start_lock,
-                                   tdb_len_t *num_locks,
-                                   tdb_off_t *bucket,
-                                   struct tdb_used_record *rec)
-{
-       /* Warning: this may drop the lock on *bucket! */
-       *num_locks = relock_hash_to_zero(tdb, *start_lock, ltype);
-       if (*num_locks == TDB_OFF_ERR)
-               return TDB_OFF_ERR;
-
-       for (*bucket = *start_lock;
-            *bucket < *start_lock + *num_locks;
-            (*bucket)++) {
-               tdb_off_t off = entry_matches(tdb, *bucket, h, &key, rec);
-               /* Empty entry or we found it? */
-               if (off == 0 || off != TDB_OFF_ERR)
-                       return off;
-       }
-
-       /* We didn't find a zero entry?  Something went badly wrong... */
-       unlock_lists(tdb, *start_lock, *start_lock + *num_locks, ltype);
-       tdb->ecode = TDB_ERR_CORRUPT;
-       tdb->log(tdb, TDB_DEBUG_FATAL, tdb->log_priv,
-                "find_and_lock: expected to find an empty hash bucket!\n");
-       return TDB_OFF_ERR;
-}
-
-/* This is the core routine which searches the hashtable for an entry.
- * On error, no locks are held and TDB_OFF_ERR is returned.
- * Otherwise, *num_locks locks of type ltype from *start_lock are held.
- * The bucket where the entry is (or would be) is in *bucket.
- * If not found, the return value is 0.
- * If found, the return value is the offset, and *rec is the record. */
-static tdb_off_t find_and_lock(struct tdb_context *tdb,
-                              struct tdb_data key,
-                              uint64_t h,
-                              int ltype,
-                              tdb_off_t *start_lock,
-                              tdb_len_t *num_locks,
-                              tdb_off_t *bucket,
-                              struct tdb_used_record *rec)
-{
-       tdb_off_t off;
-
-       /* FIXME: can we avoid locks for some fast paths? */
-       *start_lock = tdb_lock_list(tdb, h, ltype, TDB_LOCK_WAIT);
-       if (*start_lock == TDB_OFF_ERR)
-               return TDB_OFF_ERR;
-
-       /* Fast path. */
-       off = entry_matches(tdb, *start_lock, h, &key, rec);
-       if (likely(off != TDB_OFF_ERR)) {
-               *bucket = *start_lock;
-               *num_locks = 1;
-               return off;
-       }
-
-       /* Slow path, need to grab more locks and search. */
-       return find_and_lock_slow(tdb, key, h, ltype, start_lock, num_locks,
-                                 bucket, rec);
-}
-
-/* Returns -1 on error, 0 on OK, 1 on "expand and retry." */
+/* Returns -1 on error, 0 on OK */
 static int replace_data(struct tdb_context *tdb,
-                       uint64_t h, struct tdb_data key, struct tdb_data dbuf,
-                       tdb_off_t bucket,
+                       struct hash_info *h,
+                       struct tdb_data key, struct tdb_data dbuf,
                        tdb_off_t old_off, tdb_len_t old_room,
                        bool growing)
 {
        tdb_off_t new_off;
 
        /* Allocate a new record. */
-       new_off = alloc(tdb, key.dsize, dbuf.dsize, h, growing);
+       new_off = alloc(tdb, key.dsize, dbuf.dsize, h->h, growing);
        if (unlikely(new_off == TDB_OFF_ERR))
                return -1;
 
-       if (unlikely(new_off == 0))
-               return 1;
-
        /* We didn't like the existing one: remove it. */
-       if (old_off)
+       if (old_off) {
                add_free_record(tdb, old_off,
                                sizeof(struct tdb_used_record)
                                + key.dsize + old_room);
-
-       /* FIXME: Encode extra hash bits! */
-       if (tdb_write_off(tdb, hash_off(tdb, bucket), new_off) == -1)
-               return -1;
+               if (replace_in_hash(tdb, h, new_off) == -1)
+                       return -1;
+       } else {
+               if (add_to_hash(tdb, h, new_off) == -1)
+                       return -1;
+       }
 
        new_off += sizeof(struct tdb_used_record);
        if (tdb->methods->write(tdb, new_off, key.dptr, key.dsize) == -1)
@@ -773,14 +403,13 @@ static int replace_data(struct tdb_context *tdb,
 int tdb_store(struct tdb_context *tdb,
              struct tdb_data key, struct tdb_data dbuf, int flag)
 {
-       tdb_off_t off, bucket, start, num;
+       struct hash_info h;
+       tdb_off_t off;
        tdb_len_t old_room = 0;
        struct tdb_used_record rec;
-       uint64_t h;
        int ret;
 
-       h = tdb_hash(tdb, key.dptr, key.dsize);
-       off = find_and_lock(tdb, key, h, F_WRLCK, &start, &num, &bucket, &rec);
+       off = find_and_lock(tdb, key, F_WRLCK, &h, &rec, NULL);
        if (unlikely(off == TDB_OFF_ERR))
                return -1;
 
@@ -798,13 +427,14 @@ int tdb_store(struct tdb_context *tdb,
                                /* Can modify in-place.  Easy! */
                                if (update_rec_hdr(tdb, off,
                                                   key.dsize, dbuf.dsize,
-                                                  &rec, h))
+                                                  &rec, h.h))
                                        goto fail;
                                if (tdb->methods->write(tdb, off + sizeof(rec)
                                                        + key.dsize,
                                                        dbuf.dptr, dbuf.dsize))
                                        goto fail;
-                               unlock_lists(tdb, start, num, F_WRLCK);
+                               tdb_unlock_hashes(tdb, h.hlock_start,
+                                                 h.hlock_range, F_WRLCK);
                                return 0;
                        }
                        /* FIXME: See if right record is free? */
@@ -820,41 +450,27 @@ int tdb_store(struct tdb_context *tdb,
        }
 
        /* If we didn't use the old record, this implies we're growing. */
-       ret = replace_data(tdb, h, key, dbuf, bucket, off, old_room, off != 0);
-       unlock_lists(tdb, start, num, F_WRLCK);
-
-       if (unlikely(ret == 1)) {
-               /* Expand, then try again... */
-               if (tdb_expand(tdb, key.dsize, dbuf.dsize, off != 0) == -1)
-                       return -1;
-               return tdb_store(tdb, key, dbuf, flag);
-       }
-
-       /* FIXME: by simple simulation, this approximated 60% full.
-        * Check in real case! */
-       if (unlikely(num > 4 * tdb->header.v.hash_bits - 30))
-               enlarge_hash(tdb);
-
+       ret = replace_data(tdb, &h, key, dbuf, off, old_room, off != 0);
+       tdb_unlock_hashes(tdb, h.hlock_start, h.hlock_range, F_WRLCK);
        return ret;
 
 fail:
-       unlock_lists(tdb, start, num, F_WRLCK);
+       tdb_unlock_hashes(tdb, h.hlock_start, h.hlock_range, F_WRLCK);
        return -1;
 }
 
 int tdb_append(struct tdb_context *tdb,
               struct tdb_data key, struct tdb_data dbuf)
 {
-       tdb_off_t off, bucket, start, num;
+       struct hash_info h;
+       tdb_off_t off;
        struct tdb_used_record rec;
        tdb_len_t old_room = 0, old_dlen;
-       uint64_t h;
        unsigned char *newdata;
        struct tdb_data new_dbuf;
        int ret;
 
-       h = tdb_hash(tdb, key.dptr, key.dsize);
-       off = find_and_lock(tdb, key, h, F_WRLCK, &start, &num, &bucket, &rec);
+       off = find_and_lock(tdb, key, F_WRLCK, &h, &rec, NULL);
        if (unlikely(off == TDB_OFF_ERR))
                return -1;
 
@@ -865,7 +481,7 @@ int tdb_append(struct tdb_context *tdb,
                /* Fast path: can append in place. */
                if (rec_extra_padding(&rec) >= dbuf.dsize) {
                        if (update_rec_hdr(tdb, off, key.dsize,
-                                          old_dlen + dbuf.dsize, &rec, h))
+                                          old_dlen + dbuf.dsize, &rec, h.h))
                                goto fail;
 
                        off += sizeof(rec) + key.dsize + old_dlen;
@@ -874,7 +490,8 @@ int tdb_append(struct tdb_context *tdb,
                                goto fail;
 
                        /* FIXME: tdb_increment_seqnum(tdb); */
-                       unlock_lists(tdb, start, num, F_WRLCK);
+                       tdb_unlock_hashes(tdb, h.hlock_start, h.hlock_range,
+                                         F_WRLCK);
                        return 0;
                }
                /* FIXME: Check right record free? */
@@ -902,38 +519,25 @@ int tdb_append(struct tdb_context *tdb,
        }
 
        /* If they're using tdb_append(), it implies they're growing record. */
-       ret = replace_data(tdb, h, key, new_dbuf, bucket, off, old_room, true);
-       unlock_lists(tdb, start, num, F_WRLCK);
+       ret = replace_data(tdb, &h, key, new_dbuf, off, old_room, true);
+       tdb_unlock_hashes(tdb, h.hlock_start, h.hlock_range, F_WRLCK);
        free(newdata);
 
-       if (unlikely(ret == 1)) {
-               /* Expand, then try again. */
-               if (tdb_expand(tdb, key.dsize, dbuf.dsize, true) == -1)
-                       return -1;
-               return tdb_append(tdb, key, dbuf);
-       }
-
-       /* FIXME: by simple simulation, this approximated 60% full.
-        * Check in real case! */
-       if (unlikely(num > 4 * tdb->header.v.hash_bits - 30))
-               enlarge_hash(tdb);
-
        return ret;
 
 fail:
-       unlock_lists(tdb, start, num, F_WRLCK);
+       tdb_unlock_hashes(tdb, h.hlock_start, h.hlock_range, F_WRLCK);
        return -1;
 }
 
 struct tdb_data tdb_fetch(struct tdb_context *tdb, struct tdb_data key)
 {
-       tdb_off_t off, start, num, bucket;
+       tdb_off_t off;
        struct tdb_used_record rec;
-       uint64_t h;
+       struct hash_info h;
        struct tdb_data ret;
 
-       h = tdb_hash(tdb, key.dptr, key.dsize);
-       off = find_and_lock(tdb, key, h, F_RDLCK, &start, &num, &bucket, &rec);
+       off = find_and_lock(tdb, key, F_RDLCK, &h, &rec, NULL);
        if (unlikely(off == TDB_OFF_ERR))
                return tdb_null;
 
@@ -946,71 +550,29 @@ struct tdb_data tdb_fetch(struct tdb_context *tdb, struct tdb_data key)
                                          ret.dsize);
        }
 
-       unlock_lists(tdb, start, num, F_RDLCK);
+       tdb_unlock_hashes(tdb, h.hlock_start, h.hlock_range, F_RDLCK);
        return ret;
 }
 
 int tdb_delete(struct tdb_context *tdb, struct tdb_data key)
 {
-       tdb_off_t i, bucket, off, start, num;
+       tdb_off_t off;
        struct tdb_used_record rec;
-       uint64_t h;
-
-       h = tdb_hash(tdb, key.dptr, key.dsize);
-       start = tdb_lock_list(tdb, h, F_WRLCK, TDB_LOCK_WAIT);
-       if (unlikely(start == TDB_OFF_ERR))
-               return -1;
+       struct hash_info h;
 
-       /* FIXME: Fastpath: if next is zero, we can delete without lock,
-        * since this lock protects us. */
-       off = find_and_lock_slow(tdb, key, h, F_WRLCK,
-                                &start, &num, &bucket, &rec);
+       off = find_and_lock(tdb, key, F_WRLCK, &h, &rec, NULL);
        if (unlikely(off == TDB_OFF_ERR))
                return -1;
 
        if (!off) {
-               /* FIXME: We could optimize not found case if it mattered, by
-                * reading offset after first lock: if it's zero, goto here. */
-               unlock_lists(tdb, start, num, F_WRLCK);
+               tdb_unlock_hashes(tdb, h.hlock_start, h.hlock_range, F_WRLCK);
                tdb->ecode = TDB_ERR_NOEXIST;
                return -1;
        }
-       /* Since we found the entry, we must have locked it and a zero. */
-       assert(num >= 2);
 
-       /* This actually unlinks it. */
-       if (tdb_write_off(tdb, hash_off(tdb, bucket), 0) == -1)
+       if (delete_from_hash(tdb, &h) == -1)
                goto unlock_err;
 
-       /* Rehash anything following. */
-       for (i = bucket+1; i != bucket + num - 1; i++) {
-               tdb_off_t hoff, off2;
-               uint64_t h2;
-
-               hoff = hash_off(tdb, i);
-               off2 = tdb_read_off(tdb, hoff);
-               if (unlikely(off2 == TDB_OFF_ERR))
-                       goto unlock_err;
-
-               /* This can happen if we raced. */
-               if (unlikely(off2 == 0))
-                       break;
-
-               /* Maybe use a bit to indicate it is in ideal place? */
-               h2 = hash_record(tdb, off2);
-               /* Is it happy where it is? */
-               if (hash_off(tdb, h2) == hoff)
-                       continue;
-
-               /* Remove it. */
-               if (tdb_write_off(tdb, hoff, 0) == -1)
-                       goto unlock_err;
-
-               /* Rehash it. */
-               if (hash_add(tdb, h2, off2) == -1)
-                       goto unlock_err;
-       }
-
        /* Free the deleted entry. */
        if (add_free_record(tdb, off,
                            sizeof(struct tdb_used_record)
@@ -1019,11 +581,11 @@ int tdb_delete(struct tdb_context *tdb, struct tdb_data key)
                            + rec_extra_padding(&rec)) != 0)
                goto unlock_err;
 
-       unlock_lists(tdb, start, num, F_WRLCK);
+       tdb_unlock_hashes(tdb, h.hlock_start, h.hlock_range, F_WRLCK);
        return 0;
 
 unlock_err:
-       unlock_lists(tdb, start, num, F_WRLCK);
+       tdb_unlock_hashes(tdb, h.hlock_start, h.hlock_range, F_WRLCK);
        return -1;
 }
 
@@ -1072,3 +634,21 @@ enum TDB_ERROR tdb_error(struct tdb_context *tdb)
 {
        return tdb->ecode;
 }
+
+const char *tdb_errorstr(struct tdb_context *tdb)
+{
+       /* Gcc warns if you miss a case in the switch, so use that. */
+       switch (tdb->ecode) {
+       case TDB_SUCCESS: return "Success";
+       case TDB_ERR_CORRUPT: return "Corrupt database";
+       case TDB_ERR_IO: return "IO Error";
+       case TDB_ERR_LOCK: return "Locking error";
+       case TDB_ERR_OOM: return "Out of memory";
+       case TDB_ERR_EXISTS: return "Record exists";
+       case TDB_ERR_NESTING: return "Transaction already started";
+       case TDB_ERR_EINVAL: return "Invalid parameter";
+       case TDB_ERR_NOEXIST: return "Record does not exist";
+       case TDB_ERR_RDONLY: return "write not permitted";
+       }
+       return "Invalid error code";
+}