*/
#include "tdb1_private.h"
+#include <assert.h>
TDB_DATA tdb1_null;
return ret;
}
-TDB_DATA tdb1_fetch(struct tdb_context *tdb, TDB_DATA key)
+enum TDB_ERROR tdb1_fetch(struct tdb_context *tdb, TDB_DATA key, TDB_DATA *data)
{
- TDB_DATA ret = _tdb1_fetch(tdb, key);
-
- return ret;
+ *data = _tdb1_fetch(tdb, key);
+ if (data->dptr == NULL)
+ return tdb->last_error;
+ return TDB_SUCCESS;
}
-/*
- * Find an entry in the database and hand the record's data to a parsing
- * function. The parsing function is executed under the chain read lock, so it
- * should be fast and should not block on other syscalls.
- *
- * DON'T CALL OTHER TDB CALLS FROM THE PARSER, THIS MIGHT LEAD TO SEGFAULTS.
- *
- * For mmapped tdb's that do not have a transaction open it points the parsing
- * function directly at the mmap area, it avoids the malloc/memcpy in this
- * case. If a transaction is open or no mmap is available, it has to do
- * malloc/read/parse/free.
- *
- * This is interesting for all readers of potentially large data structures in
- * the tdb records, ldb indexes being one example.
- *
- * Return -1 if the record was not found.
- */
-
-int tdb1_parse_record(struct tdb_context *tdb, TDB_DATA key,
- int (*parser)(TDB_DATA key, TDB_DATA data,
- void *private_data),
- void *private_data)
+enum TDB_ERROR tdb1_parse_record(struct tdb_context *tdb, TDB_DATA key,
+ enum TDB_ERROR (*parser)(TDB_DATA key,
+ TDB_DATA data,
+ void *private_data),
+ void *private_data)
{
tdb1_off_t rec_ptr;
struct tdb1_record rec;
- int ret;
+ enum TDB_ERROR ret;
uint32_t hash;
/* find which hash bucket it is in */
if (!(rec_ptr = tdb1_find_lock_hash(tdb,key,hash,F_RDLCK,&rec))) {
/* record not found */
- tdb->last_error = TDB_ERR_NOEXIST;
- return -1;
+ return TDB_ERR_NOEXIST;
}
ret = tdb1_parse_data(tdb, key, rec_ptr + sizeof(rec) + rec.key_len,
uint32_t hash = tdb_hash(tdb, key.dptr, key.dsize);
int ret;
+ assert(tdb->flags & TDB_VERSION1);
ret = tdb1_exists_hash(tdb, key, hash);
return ret;
}
uint32_t hash = tdb_hash(tdb, key.dptr, key.dsize);
int ret;
+ assert(tdb->flags & TDB_VERSION1);
ret = tdb1_delete_hash(tdb, key, hash);
return ret;
}
uint32_t hash;
int ret;
+ assert(tdb->flags & TDB_VERSION1);
+
if ((tdb->flags & TDB_RDONLY) || tdb->tdb1.traverse_read) {
tdb->last_error = TDB_ERR_RDONLY;
return -1;
TDB_DATA dbuf;
int ret = -1;
+ assert(tdb->flags & TDB_VERSION1);
+
/* find which hash bucket it is in */
hash = tdb_hash(tdb, key.dptr, key.dsize);
if (tdb1_lock(tdb, TDB1_BUCKET(hash), F_WRLCK) == -1)
return seqnum;
}
-int tdb1_hash_size(struct tdb_context *tdb)
-{
- return tdb->tdb1.header.hash_size;
-}
-
/*
add a region of the file to the freelist. Length is the size of the region in bytes,
tdb1_off_t recovery_head;
tdb1_len_t recovery_size = 0;
- if (tdb1_lockall(tdb) != 0) {
+ if (tdb_lockall(tdb) != TDB_SUCCESS) {
return -1;
}
}
}
- if (tdb1_unlockall(tdb) != 0) {
- tdb_logerr(tdb, tdb->last_error, TDB_LOG_ERROR,
- "tdb1_wipe_all: failed to unlock");
- goto failed;
- }
-
+ tdb_unlockall(tdb);
return 0;
failed:
- tdb1_unlockall(tdb);
+ tdb_unlockall(tdb);
return -1;
}
{
struct tdb_context *tmp_db;
struct traverse_state state;
+ union tdb_attribute hsize;
+
+ hsize.base.attr = TDB_ATTRIBUTE_TDB1_HASHSIZE;
+ hsize.base.next = NULL;
+ hsize.tdb1_hashsize.hsize = tdb->tdb1.header.hash_size;
if (tdb1_transaction_start(tdb) != 0) {
tdb_logerr(tdb, tdb->last_error, TDB_LOG_ERROR,
return -1;
}
- tmp_db = tdb1_open("tmpdb", tdb1_hash_size(tdb), TDB_INTERNAL, O_RDWR|O_CREAT, 0);
+ tmp_db = tdb_open("tmpdb", TDB_INTERNAL, O_RDWR|O_CREAT, 0, &hsize);
if (tmp_db == NULL) {
tdb->last_error = tdb_logerr(tdb, TDB_ERR_OOM, TDB_LOG_ERROR,
__location__ " Failed to create tmp_db");
state.error = TDB_SUCCESS;
state.dest_db = tmp_db;
- if (tdb1_traverse_read(tdb, repack_traverse, &state) == -1) {
+ if (tdb1_traverse(tdb, repack_traverse, &state) == -1) {
tdb_logerr(tdb, tdb->last_error, TDB_LOG_ERROR,
__location__ " Failed to traverse copying out");
tdb1_transaction_cancel(tdb);
- tdb1_close(tmp_db);
+ tdb_close(tmp_db);
return -1;
}
tdb->last_error = tdb_logerr(tdb, state.error, TDB_LOG_ERROR,
__location__ " Error during traversal");
tdb1_transaction_cancel(tdb);
- tdb1_close(tmp_db);
+ tdb_close(tmp_db);
return -1;
}
tdb_logerr(tdb, tdb->last_error, TDB_LOG_ERROR,
__location__ " Failed to wipe database\n");
tdb1_transaction_cancel(tdb);
- tdb1_close(tmp_db);
+ tdb_close(tmp_db);
return -1;
}
state.error = TDB_SUCCESS;
state.dest_db = tdb;
- if (tdb1_traverse_read(tmp_db, repack_traverse, &state) == -1) {
+ if (tdb1_traverse(tmp_db, repack_traverse, &state) == -1) {
tdb_logerr(tdb, tdb->last_error, TDB_LOG_ERROR,
__location__ " Failed to traverse copying back");
tdb1_transaction_cancel(tdb);
- tdb1_close(tmp_db);
+ tdb_close(tmp_db);
return -1;
}
tdb->last_error = tdb_logerr(tdb, state.error, TDB_LOG_ERROR,
__location__ " Error during second traversal");
tdb1_transaction_cancel(tdb);
- tdb1_close(tmp_db);
+ tdb_close(tmp_db);
return -1;
}
- tdb1_close(tmp_db);
+ tdb_close(tmp_db);
if (tdb1_transaction_commit(tdb) != 0) {
tdb_logerr(tdb, tdb->last_error, TDB_LOG_ERROR,