fl.l_pid = 0;
if (waitflag)
- return fcntl(tdb->fd, F_SETLKW, &fl);
+ return fcntl(tdb->file->fd, F_SETLKW, &fl);
else
- return fcntl(tdb->fd, F_SETLK, &fl);
+ return fcntl(tdb->file->fd, F_SETLK, &fl);
}
static int fcntl_unlock(struct tdb1_context *tdb, int rw, off_t off, off_t len)
fl.l_len = len;
fl.l_pid = 0;
- return fcntl(tdb->fd, F_SETLKW, &fl);
+ return fcntl(tdb->file->fd, F_SETLKW, &fl);
}
/* list -1 is the alloc list, otherwise a hash chain. */
if (!(flags & TDB_LOCK_PROBE) && errno != EAGAIN) {
tdb_logerr(tdb, TDB_ERR_LOCK, TDB_LOG_ERROR,
"tdb1_brlock failed (fd=%d) at offset %d rw_type=%d flags=%d len=%d",
- tdb->fd, offset, rw_type, flags, (int)len);
+ tdb->file->fd, offset, rw_type, flags, (int)len);
}
return -1;
}
tdb->last_error = tdb_logerr(tdb, TDB_ERR_LOCK, TDB_LOG_ERROR,
"tdb1_brunlock failed (fd=%d) at offset"
" %d rw_type=%d len=%d",
- tdb->fd, offset, rw_type, (int)len);
+ tdb->file->fd, offset, rw_type, (int)len);
}
return ret;
}
{
int count = 1000;
- if (tdb->allrecord_lock.count != 1) {
+ if (tdb->file->allrecord_lock.count != 1) {
tdb->last_error = tdb_logerr(tdb, TDB_ERR_LOCK, TDB_LOG_ERROR,
"tdb1_allrecord_upgrade failed: "
"count %u too high",
- tdb->allrecord_lock.count);
+ tdb->file->allrecord_lock.count);
return -1;
}
- if (tdb->allrecord_lock.off != 1) {
+ if (tdb->file->allrecord_lock.off != 1) {
tdb->last_error = tdb_logerr(tdb, TDB_ERR_LOCK, TDB_LOG_ERROR,
"tdb1_allrecord_upgrade failed:"
" already upgraded?");
struct timeval tv;
if (tdb1_brlock(tdb, F_WRLCK, TDB1_FREELIST_TOP, 0,
TDB_LOCK_WAIT|TDB_LOCK_PROBE) == 0) {
- tdb->allrecord_lock.ltype = F_WRLCK;
- tdb->allrecord_lock.off = 0;
+ tdb->file->allrecord_lock.ltype = F_WRLCK;
+ tdb->file->allrecord_lock.off = 0;
return 0;
}
if (errno != EDEADLK) {
return -1;
}
-static struct tdb1_lock_type *tdb1_find_nestlock(struct tdb1_context *tdb,
- tdb1_off_t offset)
+static struct tdb_lock *tdb1_find_nestlock(struct tdb1_context *tdb,
+ tdb1_off_t offset)
{
unsigned int i;
- for (i=0; i<tdb->num_lockrecs; i++) {
- if (tdb->lockrecs[i].off == offset) {
- return &tdb->lockrecs[i];
+ for (i=0; i<tdb->file->num_lockrecs; i++) {
+ if (tdb->file->lockrecs[i].off == offset) {
+ return &tdb->file->lockrecs[i];
}
}
return NULL;
int tdb1_nest_lock(struct tdb1_context *tdb, uint32_t offset, int ltype,
enum tdb_lock_flags flags)
{
- struct tdb1_lock_type *new_lck;
+ struct tdb_lock *new_lck;
if (offset >= lock_offset(tdb->header.hash_size)) {
tdb->last_error = tdb_logerr(tdb, TDB_ERR_LOCK, TDB_LOG_ERROR,
return 0;
}
- new_lck = (struct tdb1_lock_type *)realloc(
- tdb->lockrecs,
- sizeof(*tdb->lockrecs) * (tdb->num_lockrecs+1));
+ new_lck = (struct tdb_lock *)realloc(
+ tdb->file->lockrecs,
+ sizeof(*tdb->file->lockrecs) * (tdb->file->num_lockrecs+1));
if (new_lck == NULL) {
errno = ENOMEM;
return -1;
}
- tdb->lockrecs = new_lck;
+ tdb->file->lockrecs = new_lck;
/* Since fcntl locks don't nest, we do a lock for the first one,
and simply bump the count for future ones */
return -1;
}
- tdb->lockrecs[tdb->num_lockrecs].off = offset;
- tdb->lockrecs[tdb->num_lockrecs].count = 1;
- tdb->lockrecs[tdb->num_lockrecs].ltype = ltype;
- tdb->num_lockrecs++;
+ tdb->file->lockrecs[tdb->file->num_lockrecs].off = offset;
+ tdb->file->lockrecs[tdb->file->num_lockrecs].count = 1;
+ tdb->file->lockrecs[tdb->file->num_lockrecs].ltype = ltype;
+ tdb->file->num_lockrecs++;
return 0;
}
{
unsigned int i;
- for (i = 0; i < tdb->num_lockrecs; i++) {
- if (tdb->lockrecs[i].off >= lock_offset(-1))
+ for (i = 0; i < tdb->file->num_lockrecs; i++) {
+ if (tdb->file->lockrecs[i].off >= lock_offset(-1))
return true;
}
return false;
bool check = false;
/* a allrecord lock allows us to avoid per chain locks */
- if (tdb->allrecord_lock.count &&
- (ltype == tdb->allrecord_lock.ltype || ltype == F_RDLCK)) {
+ if (tdb->file->allrecord_lock.count &&
+ (ltype == tdb->file->allrecord_lock.ltype || ltype == F_RDLCK)) {
return 0;
}
- if (tdb->allrecord_lock.count) {
+ if (tdb->file->allrecord_lock.count) {
tdb->last_error = TDB_ERR_LOCK;
ret = -1;
} else {
int tdb1_nest_unlock(struct tdb1_context *tdb, uint32_t offset, int ltype)
{
int ret = -1;
- struct tdb1_lock_type *lck;
+ struct tdb_lock *lck;
if (tdb->flags & TDB_NOLOCK)
return 0;
* Shrink the array by overwriting the element just unlocked with the
* last array element.
*/
- *lck = tdb->lockrecs[--tdb->num_lockrecs];
+ *lck = tdb->file->lockrecs[--tdb->file->num_lockrecs];
/*
* We don't bother with realloc when the array shrinks, but if we have
* a completely idle tdb we should get rid of the locked array.
*/
- if (tdb->num_lockrecs == 0) {
- SAFE_FREE(tdb->lockrecs);
+ if (tdb->file->num_lockrecs == 0) {
+ SAFE_FREE(tdb->file->lockrecs);
}
return ret;
int tdb1_unlock(struct tdb1_context *tdb, int list, int ltype)
{
/* a global lock allows us to avoid per chain locks */
- if (tdb->allrecord_lock.count &&
- (ltype == tdb->allrecord_lock.ltype || ltype == F_RDLCK)) {
+ if (tdb->file->allrecord_lock.count &&
+ (ltype == tdb->file->allrecord_lock.ltype || ltype == F_RDLCK)) {
return 0;
}
- if (tdb->allrecord_lock.count) {
+ if (tdb->file->allrecord_lock.count) {
tdb->last_error = TDB_ERR_LOCK;
return -1;
}
return -1;
}
- if (tdb->allrecord_lock.count && tdb->allrecord_lock.ltype == ltype) {
- tdb->allrecord_lock.count++;
+ if (tdb->file->allrecord_lock.count && tdb->file->allrecord_lock.ltype == ltype) {
+ tdb->file->allrecord_lock.count++;
return 0;
}
- if (tdb->allrecord_lock.count) {
+ if (tdb->file->allrecord_lock.count) {
/* a global lock of a different type exists */
tdb->last_error = TDB_ERR_LOCK;
return -1;
return -1;
}
- tdb->allrecord_lock.count = 1;
+ tdb->file->allrecord_lock.count = 1;
/* If it's upgradable, it's actually exclusive so we can treat
* it as a write lock. */
- tdb->allrecord_lock.ltype = upgradable ? F_WRLCK : ltype;
- tdb->allrecord_lock.off = upgradable;
+ tdb->file->allrecord_lock.ltype = upgradable ? F_WRLCK : ltype;
+ tdb->file->allrecord_lock.off = upgradable;
if (tdb1_needs_recovery(tdb)) {
tdb1_allrecord_unlock(tdb, ltype);
return -1;
}
- if (tdb->allrecord_lock.count == 0) {
+ if (tdb->file->allrecord_lock.count == 0) {
tdb->last_error = TDB_ERR_LOCK;
return -1;
}
/* Upgradable locks are marked as write locks. */
- if (tdb->allrecord_lock.ltype != ltype
- && (!tdb->allrecord_lock.off || ltype != F_RDLCK)) {
+ if (tdb->file->allrecord_lock.ltype != ltype
+ && (!tdb->file->allrecord_lock.off || ltype != F_RDLCK)) {
tdb->last_error = TDB_ERR_LOCK;
return -1;
}
- if (tdb->allrecord_lock.count > 1) {
- tdb->allrecord_lock.count--;
+ if (tdb->file->allrecord_lock.count > 1) {
+ tdb->file->allrecord_lock.count--;
return 0;
}
return -1;
}
- tdb->allrecord_lock.count = 0;
- tdb->allrecord_lock.ltype = 0;
+ tdb->file->allrecord_lock.count = 0;
+ tdb->file->allrecord_lock.ltype = 0;
return 0;
}
/* record lock stops delete underneath */
int tdb1_lock_record(struct tdb1_context *tdb, tdb1_off_t off)
{
- if (tdb->allrecord_lock.count) {
+ if (tdb->file->allrecord_lock.count) {
return 0;
}
return off ? tdb1_brlock(tdb, F_RDLCK, off, 1, TDB_LOCK_WAIT) : 0;
for (i = &tdb->travlocks; i; i = i->next)
if (i->off == off)
return -1;
- if (tdb->allrecord_lock.count) {
- if (tdb->allrecord_lock.ltype == F_WRLCK) {
+ if (tdb->file->allrecord_lock.count) {
+ if (tdb->file->allrecord_lock.ltype == F_WRLCK) {
return 0;
}
return -1;
int tdb1_write_unlock_record(struct tdb1_context *tdb, tdb1_off_t off)
{
- if (tdb->allrecord_lock.count) {
+ if (tdb->file->allrecord_lock.count) {
return 0;
}
return tdb1_brunlock(tdb, F_WRLCK, off, 1);
struct tdb1_traverse_lock *i;
uint32_t count = 0;
- if (tdb->allrecord_lock.count) {
+ if (tdb->file->allrecord_lock.count) {
return 0;
}
bool tdb1_have_extra_locks(struct tdb1_context *tdb)
{
- unsigned int extra = tdb->num_lockrecs;
+ unsigned int extra = tdb->file->num_lockrecs;
/* A transaction holds the lock for all records. */
- if (!tdb->transaction && tdb->allrecord_lock.count) {
+ if (!tdb->transaction && tdb->file->allrecord_lock.count) {
return true;
}
{
unsigned int i, active = 0;
- if (tdb->allrecord_lock.count != 0) {
- tdb1_brunlock(tdb, tdb->allrecord_lock.ltype, TDB1_FREELIST_TOP, 0);
- tdb->allrecord_lock.count = 0;
+ if (tdb->file->allrecord_lock.count != 0) {
+ tdb1_brunlock(tdb, tdb->file->allrecord_lock.ltype, TDB1_FREELIST_TOP, 0);
+ tdb->file->allrecord_lock.count = 0;
}
- for (i=0;i<tdb->num_lockrecs;i++) {
- struct tdb1_lock_type *lck = &tdb->lockrecs[i];
+ for (i=0;i<tdb->file->num_lockrecs;i++) {
+ struct tdb_lock *lck = &tdb->file->lockrecs[i];
/* Don't release the active lock! Copy it to first entry. */
if (lck->off == TDB1_ACTIVE_LOCK) {
- tdb->lockrecs[active++] = *lck;
+ tdb->file->lockrecs[active++] = *lck;
} else {
tdb1_brunlock(tdb, lck->ltype, lck->off, 1);
}
}
- tdb->num_lockrecs = active;
- if (tdb->num_lockrecs == 0) {
- SAFE_FREE(tdb->lockrecs);
+ tdb->file->num_lockrecs = active;
+ if (tdb->file->num_lockrecs == 0) {
+ SAFE_FREE(tdb->file->lockrecs);
}
}