X-Git-Url: https://git.ozlabs.org/?p=ccan;a=blobdiff_plain;f=ccan%2Ftdb2%2Ftraverse.c;h=410054d7947a704407dbf3c559a569d22ea27a87;hp=d30c24cdde5cc837dd849a4eec2662dc2ef75dee;hb=d1cea3ebf96f61b5bbac1e74975700770e06add6;hpb=ee6f11b307599200a574208372bc962eff81a9b1 diff --git a/ccan/tdb2/traverse.c b/ccan/tdb2/traverse.c index d30c24cd..410054d7 100644 --- a/ccan/tdb2/traverse.c +++ b/ccan/tdb2/traverse.c @@ -21,91 +21,24 @@ static int64_t traverse(struct tdb_context *tdb, int ltype, tdb_traverse_func fn, void *p) { - uint64_t i, num, count = 0; - tdb_off_t off, prev_bucket; - struct tdb_used_record rec; + int ret; + struct traverse_info tinfo; struct tdb_data k, d; - bool finish = false; - - /* FIXME: Do we need to start at 0? */ - prev_bucket = tdb_lock_list(tdb, 0, ltype, TDB_LOCK_WAIT); - if (prev_bucket != 0) - return -1; - - num = (1ULL << tdb->header.v.hash_bits); - - for (i = tdb_find_nonzero_off(tdb, hash_off(tdb, 0), num); - i != num && !finish; - i += tdb_find_nonzero_off(tdb, hash_off(tdb, i), num - i)) { - if (tdb_lock_list(tdb, i, ltype, TDB_LOCK_WAIT) != i) - goto fail; - - off = tdb_read_off(tdb, hash_off(tdb, i)); - if (off == TDB_OFF_ERR) { - tdb_unlock_list(tdb, i, ltype); - goto fail; - } + int64_t count = 0; - /* This race can happen, but look again. */ - if (off == 0) { - tdb_unlock_list(tdb, i, ltype); - continue; - } - - /* Drop previous lock. */ - tdb_unlock_list(tdb, prev_bucket, ltype); - prev_bucket = i; - - if (tdb_read_convert(tdb, off, &rec, sizeof(rec)) != 0) - goto fail; - - k.dsize = rec_key_length(&rec); - d.dsize = rec_data_length(&rec); - if (ltype == F_RDLCK) { - /* Read traverses can keep the lock. */ - k.dptr = (void *)tdb_access_read(tdb, - off + sizeof(rec), - k.dsize + d.dsize, - false); - } else { - k.dptr = tdb_alloc_read(tdb, off + sizeof(rec), - k.dsize + d.dsize); - } - if (!k.dptr) - goto fail; + for (ret = first_in_hash(tdb, ltype, &tinfo, &k, &d.dsize); + ret == 1; + ret = next_in_hash(tdb, ltype, &tinfo, &k, &d.dsize)) { d.dptr = k.dptr + k.dsize; + count++; - - if (ltype == F_WRLCK) { - /* Drop lock before calling out. */ - tdb_unlock_list(tdb, i, ltype); - } - if (fn && fn(tdb, k, d, p)) - finish = true; - - if (ltype == F_WRLCK) { - free(k.dptr); - /* Regain lock. FIXME: Is this necessary? */ - if (tdb_lock_list(tdb, i, ltype, TDB_LOCK_WAIT) != i) - return -1; - - /* This makes deleting under ourselves a bit nicer. */ - if (tdb_read_off(tdb, hash_off(tdb, i)) == off) - i++; - } else { - tdb_access_release(tdb, k.dptr); - i++; - } + break; } - /* Drop final lock. */ - tdb_unlock_list(tdb, prev_bucket, ltype); + if (ret < 0) + return -1; return count; - -fail: - tdb_unlock_list(tdb, prev_bucket, ltype); - return -1; } int64_t tdb_traverse(struct tdb_context *tdb, tdb_traverse_func fn, void *p) @@ -123,3 +56,41 @@ int64_t tdb_traverse_read(struct tdb_context *tdb, tdb->read_only = was_ro; return ret; } + +TDB_DATA tdb_firstkey(struct tdb_context *tdb) +{ + struct traverse_info tinfo; + struct tdb_data k; + switch (first_in_hash(tdb, F_RDLCK, &tinfo, &k, NULL)) { + case 1: + return k; + case 0: + tdb->ecode = TDB_SUCCESS; + /* Fall thru... */ + default: + return tdb_null; + } +} + +/* We lock twice, not very efficient. We could keep last key & tinfo cached. */ +TDB_DATA tdb_nextkey(struct tdb_context *tdb, TDB_DATA key) +{ + struct traverse_info tinfo; + struct hash_info h; + struct tdb_used_record rec; + + tinfo.prev = find_and_lock(tdb, key, F_RDLCK, &h, &rec, &tinfo); + if (unlikely(tinfo.prev == TDB_OFF_ERR)) + return tdb_null; + tdb_unlock_hashes(tdb, h.hlock_start, h.hlock_range, F_RDLCK); + + switch (next_in_hash(tdb, F_RDLCK, &tinfo, &key, NULL)) { + case 1: + return key; + case 0: + tdb->ecode = TDB_SUCCESS; + /* Fall thru... */ + default: + return tdb_null; + } +}