X-Git-Url: https://git.ozlabs.org/?p=ccan;a=blobdiff_plain;f=ccan%2Ftdb2%2Fhash.c;h=51874918c62ef4fb53e5581c252e8ace6301de9a;hp=0fd4774939d0399b9aa1706a3bf4d01b53c69bbd;hb=c5e3f07a30056cb7c3c380bf690f80815ca4b6b1;hpb=bee60dd0ffe0b1b5821061b4384000c4246f2051;ds=inline diff --git a/ccan/tdb2/hash.c b/ccan/tdb2/hash.c index 0fd47749..51874918 100644 --- a/ccan/tdb2/hash.c +++ b/ccan/tdb2/hash.c @@ -129,6 +129,13 @@ bool is_subhash(tdb_off_t val) return val >> (64-TDB_OFF_UPPER_STEAL) == (1<home_bucket = use_bits(h, TDB_HASH_GROUP_BITS); - /* FIXME: Guess the depth, don't over-lock! */ - h->hlock_start = (tdb_off_t)group - << (64 - (TDB_TOPLEVEL_HASH_BITS - TDB_HASH_GROUP_BITS)); - h->hlock_range = 1ULL << (64 - (TDB_TOPLEVEL_HASH_BITS - - TDB_HASH_GROUP_BITS)); + h->hlock_start = hlock_range(group, &h->hlock_range); if (tdb_lock_hashes(tdb, h->hlock_start, h->hlock_range, ltype, TDB_LOCK_WAIT)) return TDB_OFF_ERR; @@ -537,7 +540,7 @@ again: /* Return 1 if we find something, 0 if not, -1 on error. */ int next_in_hash(struct tdb_context *tdb, int ltype, struct traverse_info *tinfo, - TDB_DATA *kbuf, unsigned int *dlen) + TDB_DATA *kbuf, size_t *dlen) { const unsigned group_bits = TDB_TOPLEVEL_HASH_BITS-TDB_HASH_GROUP_BITS; tdb_off_t hlock_start, hlock_range, off; @@ -560,6 +563,14 @@ int next_in_hash(struct tdb_context *tdb, int ltype, ltype); return -1; } + if (rec_magic(&rec) != TDB_MAGIC) { + tdb->log(tdb, TDB_DEBUG_FATAL, tdb->log_priv, + "next_in_hash:" + " corrupt record at %llu\n", + (long long)off); + return -1; + } + kbuf->dsize = rec_key_length(&rec); /* They want data as well? */ @@ -591,7 +602,7 @@ int next_in_hash(struct tdb_context *tdb, int ltype, /* Return 1 if we find something, 0 if not, -1 on error. */ int first_in_hash(struct tdb_context *tdb, int ltype, struct traverse_info *tinfo, - TDB_DATA *kbuf, unsigned int *dlen) + TDB_DATA *kbuf, size_t *dlen) { tinfo->prev = 0; tinfo->toplevel_group = 0; @@ -603,40 +614,45 @@ int first_in_hash(struct tdb_context *tdb, int ltype, return next_in_hash(tdb, ltype, tinfo, kbuf, dlen); } -TDB_DATA tdb_firstkey(struct tdb_context *tdb) +/* Even if the entry isn't in this hash bucket, you'd have to lock this + * bucket to find it. */ +static int chainlock(struct tdb_context *tdb, const TDB_DATA *key, + int ltype, enum tdb_lock_flags waitflag, + const char *func) { - struct traverse_info tinfo; - struct tdb_data k; - switch (first_in_hash(tdb, F_RDLCK, &tinfo, &k, NULL)) { - case 1: - return k; - case 0: - tdb->ecode = TDB_SUCCESS; - /* Fall thru... */ - default: - return tdb_null; - } -} + int ret; + uint64_t h = tdb_hash(tdb, key->dptr, key->dsize); + tdb_off_t lockstart, locksize; + unsigned int group, gbits; + + gbits = TDB_TOPLEVEL_HASH_BITS - TDB_HASH_GROUP_BITS; + group = bits(h, 64 - gbits, gbits); + + lockstart = hlock_range(group, &locksize); -/* We lock twice, not very efficient. We could keep last key & tinfo cached. */ -TDB_DATA tdb_nextkey(struct tdb_context *tdb, TDB_DATA key) + ret = tdb_lock_hashes(tdb, lockstart, locksize, ltype, waitflag); + tdb_trace_1rec(tdb, func, *key); + return ret; +} + +/* lock/unlock one hash chain. This is meant to be used to reduce + contention - it cannot guarantee how many records will be locked */ +int tdb_chainlock(struct tdb_context *tdb, TDB_DATA key) { - struct traverse_info tinfo; - struct hash_info h; - struct tdb_used_record rec; - - tinfo.prev = find_and_lock(tdb, key, F_RDLCK, &h, &rec, &tinfo); - if (unlikely(tinfo.prev == TDB_OFF_ERR)) - return tdb_null; - tdb_unlock_hashes(tdb, h.hlock_start, h.hlock_range, F_RDLCK); - - switch (next_in_hash(tdb, F_RDLCK, &tinfo, &key, NULL)) { - case 1: - return key; - case 0: - tdb->ecode = TDB_SUCCESS; - /* Fall thru... */ - default: - return tdb_null; - } -} + return chainlock(tdb, &key, F_WRLCK, TDB_LOCK_WAIT, "tdb_chainlock"); +} + +int tdb_chainunlock(struct tdb_context *tdb, TDB_DATA key) +{ + uint64_t h = tdb_hash(tdb, key.dptr, key.dsize); + tdb_off_t lockstart, locksize; + unsigned int group, gbits; + + gbits = TDB_TOPLEVEL_HASH_BITS - TDB_HASH_GROUP_BITS; + group = bits(h, 64 - gbits, gbits); + + lockstart = hlock_range(group, &locksize); + + tdb_trace_1rec(tdb, "tdb_chainunlock", key); + return tdb_unlock_hashes(tdb, lockstart, locksize, F_WRLCK); +}