X-Git-Url: http://git.ozlabs.org/?p=ccan;a=blobdiff_plain;f=ccan%2Ftdb2%2Ftdb.c;h=1a229dc2b94c1538e5093a8df55d74b8dd8d74f8;hp=40811c2eff83609ca2c55d757815e2a0cae9f9d1;hb=426c8dc977d0fb46286042f7072d86bc053836a1;hpb=4f73f6a6dfc0d65aa9a5055683bf7baa5a7b622e diff --git a/ccan/tdb2/tdb.c b/ccan/tdb2/tdb.c index 40811c2e..1a229dc2 100644 --- a/ccan/tdb2/tdb.c +++ b/ccan/tdb2/tdb.c @@ -1,6 +1,5 @@ #include "private.h" #include -#include #include #include #include @@ -18,51 +17,6 @@ null_log_fn(struct tdb_context *tdb, { } -/* We do a lot of work assuming our copy of the header volatile area - * is uptodate, and usually it is. However, once we grab a lock, we have to - * re-check it. */ -bool header_changed(struct tdb_context *tdb) -{ - uint64_t gen; - - if (!(tdb->flags & TDB_NOLOCK) && tdb->header_uptodate) { - tdb->log(tdb, TDB_DEBUG_WARNING, tdb->log_priv, - "warning: header uptodate already\n"); - } - - /* We could get a partial update if we're not holding any locks. */ - assert((tdb->flags & TDB_NOLOCK) || tdb_has_locks(tdb)); - - tdb->header_uptodate = true; - gen = tdb_read_off(tdb, offsetof(struct tdb_header, v.generation)); - if (unlikely(gen != tdb->header.v.generation)) { - tdb_read_convert(tdb, offsetof(struct tdb_header, v), - &tdb->header.v, sizeof(tdb->header.v)); - return true; - } - return false; -} - -int write_header(struct tdb_context *tdb) -{ - assert(tdb_read_off(tdb, offsetof(struct tdb_header, v.generation)) - == tdb->header.v.generation); - tdb->header.v.generation++; - return tdb_write_convert(tdb, offsetof(struct tdb_header, v), - &tdb->header.v, sizeof(tdb->header.v)); -} - -static uint64_t jenkins_hash(const void *key, size_t length, uint64_t seed, - void *arg) -{ - return hash64_stable((const unsigned char *)key, length, seed); -} - -uint64_t tdb_hash(struct tdb_context *tdb, const void *ptr, size_t len) -{ - return tdb->khash(ptr, len, tdb->header.hash_seed, tdb->hash_priv); -} - static bool tdb_already_open(dev_t device, ino_t ino) { struct tdb_context *i; @@ -124,66 +78,57 @@ static uint64_t random_number(struct tdb_context *tdb) return ret; } -struct new_db_head { +struct new_database { struct tdb_header hdr; + /* Initial free zone. */ struct free_zone_header zhdr; tdb_off_t free[BUCKETS_FOR_ZONE(INITIAL_ZONE_BITS) + 1]; - struct tdb_used_record hrec; - tdb_off_t hash[1ULL << INITIAL_HASH_BITS]; struct tdb_free_record frec; -}; - -struct new_database { - struct new_db_head h; /* Rest up to 1 << INITIAL_ZONE_BITS is empty. */ char space[(1 << INITIAL_ZONE_BITS) - - (sizeof(struct new_db_head) - sizeof(struct tdb_header))]; + - sizeof(struct free_zone_header) + - sizeof(tdb_off_t) * (BUCKETS_FOR_ZONE(INITIAL_ZONE_BITS)+1) + - sizeof(struct tdb_free_record)]; uint8_t tailer; /* Don't count final padding! */ }; /* initialise a new database */ -static int tdb_new_database(struct tdb_context *tdb) +static int tdb_new_database(struct tdb_context *tdb, struct tdb_header *hdr) { /* We make it up in memory, then write it out if not internal */ struct new_database newdb; - unsigned int bucket, magic_off, dbsize; + unsigned int bucket, magic_len, dbsize; /* Don't want any extra padding! */ dbsize = offsetof(struct new_database, tailer) + sizeof(newdb.tailer); /* Fill in the header */ - newdb.h.hdr.version = TDB_VERSION; - newdb.h.hdr.hash_seed = random_number(tdb); - newdb.h.hdr.hash_test = TDB_HASH_MAGIC; - newdb.h.hdr.hash_test = tdb->khash(&newdb.h.hdr.hash_test, - sizeof(newdb.h.hdr.hash_test), - newdb.h.hdr.hash_seed, - tdb->hash_priv); - memset(newdb.h.hdr.reserved, 0, sizeof(newdb.h.hdr.reserved)); - newdb.h.hdr.v.generation = 0; + newdb.hdr.version = TDB_VERSION; + newdb.hdr.hash_seed = random_number(tdb); + newdb.hdr.hash_test = TDB_HASH_MAGIC; + newdb.hdr.hash_test = tdb->khash(&newdb.hdr.hash_test, + sizeof(newdb.hdr.hash_test), + newdb.hdr.hash_seed, + tdb->hash_priv); + memset(newdb.hdr.reserved, 0, sizeof(newdb.hdr.reserved)); /* Initial hashes are empty. */ - newdb.h.hdr.v.hash_bits = INITIAL_HASH_BITS; - newdb.h.hdr.v.hash_off = offsetof(struct new_database, h.hash); - set_header(tdb, &newdb.h.hrec, 0, - sizeof(newdb.h.hash), sizeof(newdb.h.hash), 0, - INITIAL_ZONE_BITS); - memset(newdb.h.hash, 0, sizeof(newdb.h.hash)); + memset(newdb.hdr.hashtable, 0, sizeof(newdb.hdr.hashtable)); + + /* Free is mostly empty... */ + newdb.zhdr.zone_bits = INITIAL_ZONE_BITS; + memset(newdb.free, 0, sizeof(newdb.free)); /* Create the single free entry. */ - newdb.h.frec.magic_and_meta = TDB_FREE_MAGIC | INITIAL_ZONE_BITS; - newdb.h.frec.data_len = (sizeof(newdb.h.frec) + newdb.frec.magic_and_meta = TDB_FREE_MAGIC | INITIAL_ZONE_BITS; + newdb.frec.data_len = (sizeof(newdb.frec) - sizeof(struct tdb_used_record) + sizeof(newdb.space)); - /* Free is mostly empty... */ - newdb.h.zhdr.zone_bits = INITIAL_ZONE_BITS; - memset(newdb.h.free, 0, sizeof(newdb.h.free)); - - /* ... except for this one bucket. */ - bucket = size_to_bucket(INITIAL_ZONE_BITS, newdb.h.frec.data_len); - newdb.h.free[bucket] = offsetof(struct new_database, h.frec); - newdb.h.frec.next = newdb.h.frec.prev = 0; + /* Add it to the correct bucket. */ + bucket = size_to_bucket(INITIAL_ZONE_BITS, newdb.frec.data_len); + newdb.free[bucket] = offsetof(struct new_database, frec); + newdb.frec.next = newdb.frec.prev = 0; /* Clear free space to keep valgrind happy, and avoid leaking stack. */ memset(newdb.space, 0, sizeof(newdb.space)); @@ -192,16 +137,16 @@ static int tdb_new_database(struct tdb_context *tdb) newdb.tailer = INITIAL_ZONE_BITS; /* Magic food */ - memset(newdb.h.hdr.magic_food, 0, sizeof(newdb.h.hdr.magic_food)); - strcpy(newdb.h.hdr.magic_food, TDB_MAGIC_FOOD); + memset(newdb.hdr.magic_food, 0, sizeof(newdb.hdr.magic_food)); + strcpy(newdb.hdr.magic_food, TDB_MAGIC_FOOD); /* This creates an endian-converted database, as if read from disk */ - magic_off = offsetof(struct tdb_header, magic_food); + magic_len = sizeof(newdb.hdr.magic_food); tdb_convert(tdb, - (char *)&newdb.h.hdr + magic_off, - dbsize - 1 - magic_off); + (char *)&newdb.hdr + magic_len, + offsetof(struct new_database, space) - magic_len); - tdb->header = newdb.h.hdr; + *hdr = newdb.hdr; if (tdb->flags & TDB_INTERNAL) { tdb->map_size = dbsize; @@ -235,6 +180,7 @@ struct tdb_context *tdb_open(const char *name, int tdb_flags, int save_errno; uint64_t hash_test; unsigned v; + struct tdb_header hdr; tdb = malloc(sizeof(*tdb)); if (!tdb) { @@ -244,17 +190,15 @@ struct tdb_context *tdb_open(const char *name, int tdb_flags, } tdb->name = NULL; tdb->map_ptr = NULL; + tdb->direct_access = 0; tdb->fd = -1; tdb->map_size = sizeof(struct tdb_header); tdb->ecode = TDB_SUCCESS; - /* header will be read in below. */ - tdb->header_uptodate = false; tdb->flags = tdb_flags; tdb->log = null_log_fn; tdb->log_priv = NULL; - tdb->khash = jenkins_hash; - tdb->hash_priv = NULL; tdb->transaction = NULL; + tdb_hash_init(tdb); /* last_zone will be set below. */ tdb_io_init(tdb); tdb_lock_init(tdb); @@ -296,13 +240,13 @@ struct tdb_context *tdb_open(const char *name, int tdb_flags, /* internal databases don't need any of the rest. */ if (tdb->flags & TDB_INTERNAL) { tdb->flags |= (TDB_NOLOCK | TDB_NOMMAP); - if (tdb_new_database(tdb) != 0) { + if (tdb_new_database(tdb, &hdr) != 0) { tdb->log(tdb, TDB_DEBUG_ERROR, tdb->log_priv, "tdb_open: tdb_new_database failed!"); goto fail; } - TEST_IT(tdb->flags & TDB_CONVERT); - tdb_convert(tdb, &tdb->header, sizeof(tdb->header)); + tdb_convert(tdb, &hdr.hash_seed, sizeof(hdr.hash_seed)); + tdb->hash_seed = hdr.hash_seed; tdb_zone_init(tdb); return tdb; } @@ -326,32 +270,32 @@ struct tdb_context *tdb_open(const char *name, int tdb_flags, goto fail; /* errno set by tdb_brlock */ } - if (!tdb_pread_all(tdb->fd, &tdb->header, sizeof(tdb->header), 0) - || strcmp(tdb->header.magic_food, TDB_MAGIC_FOOD) != 0) { - if (!(open_flags & O_CREAT) || tdb_new_database(tdb) == -1) { + if (!tdb_pread_all(tdb->fd, &hdr, sizeof(hdr), 0) + || strcmp(hdr.magic_food, TDB_MAGIC_FOOD) != 0) { + if (!(open_flags & O_CREAT) || tdb_new_database(tdb, &hdr) == -1) { if (errno == 0) { errno = EIO; /* ie bad format or something */ } goto fail; } - } else if (tdb->header.version != TDB_VERSION) { - if (tdb->header.version == bswap_64(TDB_VERSION)) + } else if (hdr.version != TDB_VERSION) { + if (hdr.version == bswap_64(TDB_VERSION)) tdb->flags |= TDB_CONVERT; else { /* wrong version */ tdb->log(tdb, TDB_DEBUG_ERROR, tdb->log_priv, "tdb_open: %s is unknown version 0x%llx\n", - name, (long long)tdb->header.version); + name, (long long)hdr.version); errno = EIO; goto fail; } } - tdb_convert(tdb, &tdb->header, sizeof(tdb->header)); + tdb_convert(tdb, &hdr, sizeof(hdr)); + tdb->hash_seed = hdr.hash_seed; hash_test = TDB_HASH_MAGIC; - hash_test = tdb->khash(&hash_test, sizeof(hash_test), - tdb->header.hash_seed, tdb->hash_priv); - if (tdb->header.hash_test != hash_test) { + hash_test = tdb_hash(tdb, &hash_test, sizeof(hash_test)); + if (hdr.hash_test != hash_test) { /* wrong hash variant */ tdb->log(tdb, TDB_DEBUG_ERROR, tdb->log_priv, "tdb_open: %s uses a different hash function\n", @@ -420,145 +364,6 @@ struct tdb_context *tdb_open(const char *name, int tdb_flags, return NULL; } -tdb_off_t hash_off(struct tdb_context *tdb, uint64_t list) -{ - return tdb->header.v.hash_off - + ((list & ((1ULL << tdb->header.v.hash_bits) - 1)) - * sizeof(tdb_off_t)); -} - -/* Returns 0 if the entry is a zero (definitely not a match). - * Returns a valid entry offset if it's a match. Fills in rec. - * Otherwise returns TDB_OFF_ERR: keep searching. */ -static tdb_off_t entry_matches(struct tdb_context *tdb, - uint64_t list, - uint64_t hash, - const struct tdb_data *key, - struct tdb_used_record *rec) -{ - tdb_off_t off; - uint64_t keylen; - const unsigned char *rkey; - - list &= ((1ULL << tdb->header.v.hash_bits) - 1); - - off = tdb_read_off(tdb, tdb->header.v.hash_off - + list * sizeof(tdb_off_t)); - if (off == 0 || off == TDB_OFF_ERR) - return off; - -#if 0 /* FIXME: Check other bits. */ - unsigned int bits, bitmask, hoffextra; - /* Bottom three bits show how many extra hash bits. */ - bits = (off & ((1 << TDB_EXTRA_HASHBITS_NUM) - 1)) + 1; - bitmask = (1 << bits)-1; - hoffextra = ((off >> TDB_EXTRA_HASHBITS_NUM) & bitmask); - uint64_t hextra = hash >> tdb->header.v.hash_bits; - if ((hextra & bitmask) != hoffextra) - return TDB_OFF_ERR; - off &= ~...; -#endif - - if (tdb_read_convert(tdb, off, rec, sizeof(*rec)) == -1) - return TDB_OFF_ERR; - - /* FIXME: check extra bits in header! */ - keylen = rec_key_length(rec); - if (keylen != key->dsize) - return TDB_OFF_ERR; - - rkey = tdb_access_read(tdb, off + sizeof(*rec), keylen, false); - if (!rkey) - return TDB_OFF_ERR; - if (memcmp(rkey, key->dptr, keylen) != 0) - off = TDB_OFF_ERR; - tdb_access_release(tdb, rkey); - return off; -} - -/* FIXME: Optimize? */ -static void unlock_lists(struct tdb_context *tdb, - tdb_off_t list, tdb_len_t num, - int ltype) -{ - tdb_off_t i; - - for (i = list; i < list + num; i++) - tdb_unlock_list(tdb, i, ltype); -} - -/* FIXME: Optimize? */ -static int lock_lists(struct tdb_context *tdb, - tdb_off_t list, tdb_len_t num, - int ltype) -{ - tdb_off_t i; - - for (i = list; i < list + num; i++) { - if (tdb_lock_list(tdb, i, ltype, TDB_LOCK_WAIT) - == TDB_OFF_ERR) { - unlock_lists(tdb, list, i - list, ltype); - return -1; - } - } - return 0; -} - -/* We lock hashes up to the next empty offset. We already hold the - * lock on the start bucket, but we may need to release and re-grab - * it. If we fail, we hold no locks at all! */ -static tdb_len_t relock_hash_to_zero(struct tdb_context *tdb, - tdb_off_t start, int ltype) -{ - tdb_len_t num, len; - -again: - num = 1ULL << tdb->header.v.hash_bits; - len = tdb_find_zero_off(tdb, hash_off(tdb, start), num - start); - if (unlikely(len == num - start)) { - /* We hit the end of the hash range. Drop lock: we have - to lock start of hash first. */ - tdb_len_t pre_locks; - - tdb_unlock_list(tdb, start, ltype); - - /* Grab something, so header is stable. */ - if (tdb_lock_list(tdb, 0, ltype, TDB_LOCK_WAIT)) - return TDB_OFF_ERR; - pre_locks = tdb_find_zero_off(tdb, hash_off(tdb, 0), num); - /* We want to lock the zero entry as well. */ - pre_locks++; - if (lock_lists(tdb, 1, pre_locks - 1, ltype) == -1) { - tdb_unlock_list(tdb, 0, ltype); - return TDB_OFF_ERR; - } - - /* Now lock later ones. */ - if (unlikely(lock_lists(tdb, start, len, ltype) == -1)) { - unlock_lists(tdb, 0, pre_locks, ltype); - return TDB_OFF_ERR; - } - len += pre_locks; - } else { - /* We want to lock the zero entry as well. */ - len++; - /* But we already have lock on start. */ - if (unlikely(lock_lists(tdb, start+1, len-1, ltype) == -1)) { - tdb_unlock_list(tdb, start, ltype); - return TDB_OFF_ERR; - } - } - - /* Now, did we lose the race, and it's not zero any more? */ - if (unlikely(tdb_read_off(tdb, hash_off(tdb, start + len - 1)) != 0)) { - /* Leave the start locked, as expected. */ - unlock_lists(tdb, start + 1, len - 1, ltype); - goto again; - } - - return len; -} - /* FIXME: modify, don't rewrite! */ static int update_rec_hdr(struct tdb_context *tdb, tdb_off_t off, @@ -576,186 +381,10 @@ static int update_rec_hdr(struct tdb_context *tdb, return tdb_write_convert(tdb, off, rec, sizeof(*rec)); } -static int hash_add(struct tdb_context *tdb, - uint64_t hash, tdb_off_t off) -{ - tdb_off_t i, hoff, len, num; - - /* Look for next space. */ - i = (hash & ((1ULL << tdb->header.v.hash_bits) - 1)); - len = (1ULL << tdb->header.v.hash_bits) - i; - num = tdb_find_zero_off(tdb, hash_off(tdb, i), len); - - if (unlikely(num == len)) { - /* We wrapped. Look through start of hash table. */ - i = 0; - hoff = hash_off(tdb, 0); - len = (1ULL << tdb->header.v.hash_bits); - num = tdb_find_zero_off(tdb, hoff, len); - if (num == len) { - tdb->ecode = TDB_ERR_CORRUPT; - tdb->log(tdb, TDB_DEBUG_FATAL, tdb->log_priv, - "hash_add: full hash table!\n"); - return -1; - } - } - if (tdb_read_off(tdb, hash_off(tdb, i + num)) != 0) { - tdb->ecode = TDB_ERR_CORRUPT; - tdb->log(tdb, TDB_DEBUG_FATAL, tdb->log_priv, - "hash_add: overwriting hash table?\n"); - return -1; - } - - /* FIXME: Encode extra hash bits! */ - return tdb_write_off(tdb, hash_off(tdb, i + num), off); -} - -/* If we fail, others will try after us. */ -static void enlarge_hash(struct tdb_context *tdb) -{ - tdb_off_t newoff, oldoff, i; - tdb_len_t hlen; - uint64_t num = 1ULL << tdb->header.v.hash_bits; - struct tdb_used_record pad, *r; - unsigned int records = 0; - - /* FIXME: We should do this without holding locks throughout. */ - if (tdb_allrecord_lock(tdb, F_WRLCK, TDB_LOCK_WAIT, false) == -1) - return; - - /* Someone else enlarged for us? Nothing to do. */ - if ((1ULL << tdb->header.v.hash_bits) != num) - goto unlock; - - /* Allocate our new array. */ - hlen = num * sizeof(tdb_off_t) * 2; - newoff = alloc(tdb, 0, hlen, 0, false); - if (unlikely(newoff == TDB_OFF_ERR)) - goto unlock; - /* Step over record header! */ - newoff += sizeof(struct tdb_used_record); - - /* Starts all zero. */ - if (zero_out(tdb, newoff, hlen) == -1) - goto unlock; - - /* Update header now so we can use normal routines. */ - oldoff = tdb->header.v.hash_off; - - tdb->header.v.hash_bits++; - tdb->header.v.hash_off = newoff; - - /* FIXME: If the space before is empty, we know this is in its ideal - * location. Or steal a bit from the pointer to avoid rehash. */ - for (i = 0; i < num; i++) { - tdb_off_t off; - off = tdb_read_off(tdb, oldoff + i * sizeof(tdb_off_t)); - if (unlikely(off == TDB_OFF_ERR)) - goto oldheader; - if (off && hash_add(tdb, hash_record(tdb, off), off) == -1) - goto oldheader; - if (off) - records++; - } - - tdb->log(tdb, TDB_DEBUG_TRACE, tdb->log_priv, - "enlarge_hash: moved %u records from %llu buckets.\n", - records, (long long)num); - - /* Free up old hash. */ - r = tdb_get(tdb, oldoff - sizeof(*r), &pad, sizeof(*r)); - if (!r) - goto oldheader; - add_free_record(tdb, rec_zone_bits(r), oldoff - sizeof(*r), - sizeof(*r)+rec_data_length(r)+rec_extra_padding(r)); - - /* Now we write the modified header. */ - write_header(tdb); -unlock: - tdb_allrecord_unlock(tdb, F_WRLCK); - return; - -oldheader: - tdb->header.v.hash_bits--; - tdb->header.v.hash_off = oldoff; - goto unlock; -} - - -/* This is the slow version of the routine which searches the - * hashtable for an entry. - * We lock every hash bucket up to and including the next zero one. - */ -static tdb_off_t find_and_lock_slow(struct tdb_context *tdb, - struct tdb_data key, - uint64_t h, - int ltype, - tdb_off_t *start_lock, - tdb_len_t *num_locks, - tdb_off_t *bucket, - struct tdb_used_record *rec) -{ - /* Warning: this may drop the lock on *bucket! */ - *num_locks = relock_hash_to_zero(tdb, *start_lock, ltype); - if (*num_locks == TDB_OFF_ERR) - return TDB_OFF_ERR; - - for (*bucket = *start_lock; - *bucket < *start_lock + *num_locks; - (*bucket)++) { - tdb_off_t off = entry_matches(tdb, *bucket, h, &key, rec); - /* Empty entry or we found it? */ - if (off == 0 || off != TDB_OFF_ERR) - return off; - } - - /* We didn't find a zero entry? Something went badly wrong... */ - unlock_lists(tdb, *start_lock, *start_lock + *num_locks, ltype); - tdb->ecode = TDB_ERR_CORRUPT; - tdb->log(tdb, TDB_DEBUG_FATAL, tdb->log_priv, - "find_and_lock: expected to find an empty hash bucket!\n"); - return TDB_OFF_ERR; -} - -/* This is the core routine which searches the hashtable for an entry. - * On error, no locks are held and TDB_OFF_ERR is returned. - * Otherwise, *num_locks locks of type ltype from *start_lock are held. - * The bucket where the entry is (or would be) is in *bucket. - * If not found, the return value is 0. - * If found, the return value is the offset, and *rec is the record. */ -static tdb_off_t find_and_lock(struct tdb_context *tdb, - struct tdb_data key, - uint64_t h, - int ltype, - tdb_off_t *start_lock, - tdb_len_t *num_locks, - tdb_off_t *bucket, - struct tdb_used_record *rec) -{ - tdb_off_t off; - - /* FIXME: can we avoid locks for some fast paths? */ - *start_lock = tdb_lock_list(tdb, h, ltype, TDB_LOCK_WAIT); - if (*start_lock == TDB_OFF_ERR) - return TDB_OFF_ERR; - - /* Fast path. */ - off = entry_matches(tdb, *start_lock, h, &key, rec); - if (likely(off != TDB_OFF_ERR)) { - *bucket = *start_lock; - *num_locks = 1; - return off; - } - - /* Slow path, need to grab more locks and search. */ - return find_and_lock_slow(tdb, key, h, ltype, start_lock, num_locks, - bucket, rec); -} - -/* Returns -1 on error, 0 on OK" */ +/* Returns -1 on error, 0 on OK */ static int replace_data(struct tdb_context *tdb, - uint64_t h, struct tdb_data key, struct tdb_data dbuf, - tdb_off_t bucket, + struct hash_info *h, + struct tdb_data key, struct tdb_data dbuf, tdb_off_t old_off, tdb_len_t old_room, unsigned old_zone, bool growing) @@ -763,19 +392,21 @@ static int replace_data(struct tdb_context *tdb, tdb_off_t new_off; /* Allocate a new record. */ - new_off = alloc(tdb, key.dsize, dbuf.dsize, h, growing); + new_off = alloc(tdb, key.dsize, dbuf.dsize, h->h, growing); if (unlikely(new_off == TDB_OFF_ERR)) return -1; /* We didn't like the existing one: remove it. */ - if (old_off) + if (old_off) { add_free_record(tdb, old_zone, old_off, sizeof(struct tdb_used_record) + key.dsize + old_room); - - /* FIXME: Encode extra hash bits! */ - if (tdb_write_off(tdb, hash_off(tdb, bucket), new_off) == -1) - return -1; + if (replace_in_hash(tdb, h, new_off) == -1) + return -1; + } else { + if (add_to_hash(tdb, h, new_off) == -1) + return -1; + } new_off += sizeof(struct tdb_used_record); if (tdb->methods->write(tdb, new_off, key.dptr, key.dsize) == -1) @@ -792,14 +423,13 @@ static int replace_data(struct tdb_context *tdb, int tdb_store(struct tdb_context *tdb, struct tdb_data key, struct tdb_data dbuf, int flag) { - tdb_off_t off, bucket, start, num; + struct hash_info h; + tdb_off_t off; tdb_len_t old_room = 0; struct tdb_used_record rec; - uint64_t h; int ret; - h = tdb_hash(tdb, key.dptr, key.dsize); - off = find_and_lock(tdb, key, h, F_WRLCK, &start, &num, &bucket, &rec); + off = find_and_lock(tdb, key, F_WRLCK, &h, &rec); if (unlikely(off == TDB_OFF_ERR)) return -1; @@ -817,13 +447,14 @@ int tdb_store(struct tdb_context *tdb, /* Can modify in-place. Easy! */ if (update_rec_hdr(tdb, off, key.dsize, dbuf.dsize, - &rec, h)) + &rec, h.h)) goto fail; if (tdb->methods->write(tdb, off + sizeof(rec) + key.dsize, dbuf.dptr, dbuf.dsize)) goto fail; - unlock_lists(tdb, start, num, F_WRLCK); + tdb_unlock_hashes(tdb, h.hlock_start, + h.hlock_range, F_WRLCK); return 0; } /* FIXME: See if right record is free? */ @@ -839,35 +470,28 @@ int tdb_store(struct tdb_context *tdb, } /* If we didn't use the old record, this implies we're growing. */ - ret = replace_data(tdb, h, key, dbuf, bucket, off, old_room, + ret = replace_data(tdb, &h, key, dbuf, off, old_room, rec_zone_bits(&rec), off != 0); - unlock_lists(tdb, start, num, F_WRLCK); - - /* FIXME: by simple simulation, this approximated 60% full. - * Check in real case! */ - if (unlikely(num > 4 * tdb->header.v.hash_bits - 30)) - enlarge_hash(tdb); - + tdb_unlock_hashes(tdb, h.hlock_start, h.hlock_range, F_WRLCK); return ret; fail: - unlock_lists(tdb, start, num, F_WRLCK); + tdb_unlock_hashes(tdb, h.hlock_start, h.hlock_range, F_WRLCK); return -1; } int tdb_append(struct tdb_context *tdb, struct tdb_data key, struct tdb_data dbuf) { - tdb_off_t off, bucket, start, num; + struct hash_info h; + tdb_off_t off; struct tdb_used_record rec; tdb_len_t old_room = 0, old_dlen; - uint64_t h; unsigned char *newdata; struct tdb_data new_dbuf; int ret; - h = tdb_hash(tdb, key.dptr, key.dsize); - off = find_and_lock(tdb, key, h, F_WRLCK, &start, &num, &bucket, &rec); + off = find_and_lock(tdb, key, F_WRLCK, &h, &rec); if (unlikely(off == TDB_OFF_ERR)) return -1; @@ -878,7 +502,7 @@ int tdb_append(struct tdb_context *tdb, /* Fast path: can append in place. */ if (rec_extra_padding(&rec) >= dbuf.dsize) { if (update_rec_hdr(tdb, off, key.dsize, - old_dlen + dbuf.dsize, &rec, h)) + old_dlen + dbuf.dsize, &rec, h.h)) goto fail; off += sizeof(rec) + key.dsize + old_dlen; @@ -887,7 +511,8 @@ int tdb_append(struct tdb_context *tdb, goto fail; /* FIXME: tdb_increment_seqnum(tdb); */ - unlock_lists(tdb, start, num, F_WRLCK); + tdb_unlock_hashes(tdb, h.hlock_start, h.hlock_range, + F_WRLCK); return 0; } /* FIXME: Check right record free? */ @@ -915,32 +540,26 @@ int tdb_append(struct tdb_context *tdb, } /* If they're using tdb_append(), it implies they're growing record. */ - ret = replace_data(tdb, h, key, new_dbuf, bucket, off, old_room, - rec_zone_bits(&rec), true); - unlock_lists(tdb, start, num, F_WRLCK); + ret = replace_data(tdb, &h, key, new_dbuf, off, + old_room, rec_zone_bits(&rec), true); + tdb_unlock_hashes(tdb, h.hlock_start, h.hlock_range, F_WRLCK); free(newdata); - /* FIXME: by simple simulation, this approximated 60% full. - * Check in real case! */ - if (unlikely(num > 4 * tdb->header.v.hash_bits - 30)) - enlarge_hash(tdb); - return ret; fail: - unlock_lists(tdb, start, num, F_WRLCK); + tdb_unlock_hashes(tdb, h.hlock_start, h.hlock_range, F_WRLCK); return -1; } struct tdb_data tdb_fetch(struct tdb_context *tdb, struct tdb_data key) { - tdb_off_t off, start, num, bucket; + tdb_off_t off; struct tdb_used_record rec; - uint64_t h; + struct hash_info h; struct tdb_data ret; - h = tdb_hash(tdb, key.dptr, key.dsize); - off = find_and_lock(tdb, key, h, F_RDLCK, &start, &num, &bucket, &rec); + off = find_and_lock(tdb, key, F_RDLCK, &h, &rec); if (unlikely(off == TDB_OFF_ERR)) return tdb_null; @@ -953,71 +572,29 @@ struct tdb_data tdb_fetch(struct tdb_context *tdb, struct tdb_data key) ret.dsize); } - unlock_lists(tdb, start, num, F_RDLCK); + tdb_unlock_hashes(tdb, h.hlock_start, h.hlock_range, F_RDLCK); return ret; } int tdb_delete(struct tdb_context *tdb, struct tdb_data key) { - tdb_off_t i, bucket, off, start, num; + tdb_off_t off; struct tdb_used_record rec; - uint64_t h; + struct hash_info h; - h = tdb_hash(tdb, key.dptr, key.dsize); - start = tdb_lock_list(tdb, h, F_WRLCK, TDB_LOCK_WAIT); - if (unlikely(start == TDB_OFF_ERR)) - return -1; - - /* FIXME: Fastpath: if next is zero, we can delete without lock, - * since this lock protects us. */ - off = find_and_lock_slow(tdb, key, h, F_WRLCK, - &start, &num, &bucket, &rec); + off = find_and_lock(tdb, key, F_WRLCK, &h, &rec); if (unlikely(off == TDB_OFF_ERR)) return -1; if (!off) { - /* FIXME: We could optimize not found case if it mattered, by - * reading offset after first lock: if it's zero, goto here. */ - unlock_lists(tdb, start, num, F_WRLCK); + tdb_unlock_hashes(tdb, h.hlock_start, h.hlock_range, F_WRLCK); tdb->ecode = TDB_ERR_NOEXIST; return -1; } - /* Since we found the entry, we must have locked it and a zero. */ - assert(num >= 2); - /* This actually unlinks it. */ - if (tdb_write_off(tdb, hash_off(tdb, bucket), 0) == -1) + if (delete_from_hash(tdb, &h) == -1) goto unlock_err; - /* Rehash anything following. */ - for (i = bucket+1; i != bucket + num - 1; i++) { - tdb_off_t hoff, off2; - uint64_t h2; - - hoff = hash_off(tdb, i); - off2 = tdb_read_off(tdb, hoff); - if (unlikely(off2 == TDB_OFF_ERR)) - goto unlock_err; - - /* This can happen if we raced. */ - if (unlikely(off2 == 0)) - break; - - /* Maybe use a bit to indicate it is in ideal place? */ - h2 = hash_record(tdb, off2); - /* Is it happy where it is? */ - if (hash_off(tdb, h2) == hoff) - continue; - - /* Remove it. */ - if (tdb_write_off(tdb, hoff, 0) == -1) - goto unlock_err; - - /* Rehash it. */ - if (hash_add(tdb, h2, off2) == -1) - goto unlock_err; - } - /* Free the deleted entry. */ if (add_free_record(tdb, rec_zone_bits(&rec), off, sizeof(struct tdb_used_record) @@ -1026,11 +603,11 @@ int tdb_delete(struct tdb_context *tdb, struct tdb_data key) + rec_extra_padding(&rec)) != 0) goto unlock_err; - unlock_lists(tdb, start, num, F_WRLCK); + tdb_unlock_hashes(tdb, h.hlock_start, h.hlock_range, F_WRLCK); return 0; unlock_err: - unlock_lists(tdb, start, num, F_WRLCK); + tdb_unlock_hashes(tdb, h.hlock_start, h.hlock_range, F_WRLCK); return -1; }