/* Return 1 if we find something, 0 if not, -1 on error. */
int next_in_hash(struct tdb_context *tdb, int ltype,
struct traverse_info *tinfo,
- TDB_DATA *kbuf, unsigned int *dlen)
+ TDB_DATA *kbuf, size_t *dlen)
{
const unsigned group_bits = TDB_TOPLEVEL_HASH_BITS-TDB_HASH_GROUP_BITS;
tdb_off_t hlock_start, hlock_range, off;
/* Return 1 if we find something, 0 if not, -1 on error. */
int first_in_hash(struct tdb_context *tdb, int ltype,
struct traverse_info *tinfo,
- TDB_DATA *kbuf, unsigned int *dlen)
+ TDB_DATA *kbuf, size_t *dlen)
{
tinfo->prev = 0;
tinfo->toplevel_group = 0;
return next_in_hash(tdb, ltype, tinfo, kbuf, dlen);
}
-
-TDB_DATA tdb_firstkey(struct tdb_context *tdb)
-{
- struct traverse_info tinfo;
- struct tdb_data k;
- switch (first_in_hash(tdb, F_RDLCK, &tinfo, &k, NULL)) {
- case 1:
- return k;
- case 0:
- tdb->ecode = TDB_SUCCESS;
- /* Fall thru... */
- default:
- return tdb_null;
- }
-}
-
-/* We lock twice, not very efficient. We could keep last key & tinfo cached. */
-TDB_DATA tdb_nextkey(struct tdb_context *tdb, TDB_DATA key)
-{
- struct traverse_info tinfo;
- struct hash_info h;
- struct tdb_used_record rec;
-
- tinfo.prev = find_and_lock(tdb, key, F_RDLCK, &h, &rec, &tinfo);
- if (unlikely(tinfo.prev == TDB_OFF_ERR))
- return tdb_null;
- tdb_unlock_hashes(tdb, h.hlock_start, h.hlock_range, F_RDLCK);
-
- switch (next_in_hash(tdb, F_RDLCK, &tinfo, &key, NULL)) {
- case 1:
- return key;
- case 0:
- tdb->ecode = TDB_SUCCESS;
- /* Fall thru... */
- default:
- return tdb_null;
- }
-}