- /*
+ /*
Trivial Database 2: human-readable summary code
Copyright (C) Rusty Russell 2010
-
+
This library is free software; you can redistribute it and/or
modify it under the terms of the GNU Lesser General Public
License as published by the Free Software Foundation; either
#include <assert.h>
#include <ccan/tally/tally.h>
-static void sizes_for_bucket(unsigned bucket, size_t *min, size_t *max)
-{
- if (bucket <= 8) {
- *min = *max = TDB_MIN_DATA_LEN + bucket * 8;
- } else if (bucket == 9) {
- /* FIXME: This is twisted; fix size_to_bucket. */
- *min = TDB_MIN_DATA_LEN + (1ULL << (bucket - 3)) + 8;
- *max = TDB_MIN_DATA_LEN + (1ULL << (bucket - 2)) - 8;
- } else {
- *min = TDB_MIN_DATA_LEN + (1ULL << (bucket - 3));
- *max = TDB_MIN_DATA_LEN + (1ULL << (bucket - 2)) - 8;
- }
- assert(size_to_bucket(63, *min) == bucket);
- assert(size_to_bucket(63, *max) == bucket);
- if (bucket > 8)
- assert(size_to_bucket(63, *min - 8) == bucket - 1);
- assert(size_to_bucket(63, *max + 8) == bucket + 1);
-}
-
-static int count_hash(struct tdb_context *tdb,
- tdb_off_t hash_off, unsigned bits)
+static tdb_off_t count_hash(struct tdb_context *tdb,
+ tdb_off_t hash_off, unsigned bits)
{
const tdb_off_t *h;
- unsigned int i, count = 0;
+ tdb_off_t count = 0;
+ unsigned int i;
h = tdb_access_read(tdb, hash_off, sizeof(*h) << bits, true);
- if (!h)
- return -1;
+ if (TDB_PTR_IS_ERR(h)) {
+ return TDB_PTR_ERR(h);
+ }
for (i = 0; i < (1 << bits); i++)
count += (h[i] != 0);
return count;
}
-static tdb_len_t summarize_zone(struct tdb_context *tdb, tdb_off_t zone_off,
- struct tally *zones,
+static enum TDB_ERROR summarize(struct tdb_context *tdb,
struct tally *hashes,
- struct tally *free,
+ struct tally *ftables,
+ struct tally *fr,
struct tally *keys,
struct tally *data,
struct tally *extra,
struct tally *uncoal,
- uint64_t bucketlen[],
- unsigned int *num_buckets)
+ struct tally *chains)
{
- struct free_zone_header zhdr;
tdb_off_t off;
tdb_len_t len;
- unsigned int hdrlen;
tdb_len_t unc = 0;
- if (tdb_read_convert(tdb, zone_off, &zhdr, sizeof(zhdr)) == -1)
- return TDB_OFF_ERR;
-
- tally_add(zones, 1ULL << zhdr.zone_bits);
- *num_buckets = BUCKETS_FOR_ZONE(zhdr.zone_bits);
-
- hdrlen = sizeof(zhdr)
- + (BUCKETS_FOR_ZONE(zhdr.zone_bits) + 1) * sizeof(tdb_off_t);
- for (off = zone_off + hdrlen;
- off < zone_off + (1ULL << zhdr.zone_bits);
+ for (off = sizeof(struct tdb_header);
+ off < tdb->file->map_size;
off += len) {
- union {
+ const union {
struct tdb_used_record u;
struct tdb_free_record f;
- } pad, *p;
- p = tdb_get(tdb, off, &pad, sizeof(pad));
- if (!p)
- return TDB_OFF_ERR;
- if (frec_magic(&p->f) == TDB_FREE_MAGIC) {
- len = p->f.data_len;
- tally_add(free, len);
- bucketlen[size_to_bucket(frec_zone_bits(&p->f), len)]++;
- len += sizeof(p->u);
- unc++;
- } else {
- if (unc) {
+ struct tdb_recovery_record r;
+ } *p;
+ /* We might not be able to get the whole thing. */
+ p = tdb_access_read(tdb, off, sizeof(p->f), true);
+ if (TDB_PTR_IS_ERR(p)) {
+ return TDB_PTR_ERR(p);
+ }
+ if (frec_magic(&p->f) != TDB_FREE_MAGIC) {
+ if (unc > 1) {
tally_add(uncoal, unc);
unc = 0;
}
+ }
+
+ if (p->r.magic == TDB_RECOVERY_INVALID_MAGIC
+ || p->r.magic == TDB_RECOVERY_MAGIC) {
+ len = sizeof(p->r) + p->r.max_len;
+ } else if (frec_magic(&p->f) == TDB_FREE_MAGIC) {
+ len = frec_len(&p->f);
+ tally_add(fr, len);
+ len += sizeof(p->u);
+ unc++;
+ } else if (rec_magic(&p->u) == TDB_USED_MAGIC) {
len = sizeof(p->u)
+ rec_key_length(&p->u)
+ rec_data_length(&p->u)
+ rec_extra_padding(&p->u);
- /* FIXME: Use different magic for hashes? */
- if (!rec_key_length(&p->u) && !rec_hash(&p->u)) {
- int count = count_hash(tdb, off + sizeof(p->u),
- TDB_SUBLEVEL_HASH_BITS);
- if (count == -1)
- return TDB_OFF_ERR;
- tally_add(hashes, count);
- } else {
- tally_add(keys, rec_key_length(&p->u));
- tally_add(data, rec_data_length(&p->u));
+ tally_add(keys, rec_key_length(&p->u));
+ tally_add(data, rec_data_length(&p->u));
+ tally_add(extra, rec_extra_padding(&p->u));
+ } else if (rec_magic(&p->u) == TDB_HTABLE_MAGIC) {
+ tdb_off_t count = count_hash(tdb,
+ off + sizeof(p->u),
+ TDB_SUBLEVEL_HASH_BITS);
+ if (TDB_OFF_IS_ERR(count)) {
+ return count;
}
+ tally_add(hashes, count);
+ tally_add(extra, rec_extra_padding(&p->u));
+ len = sizeof(p->u)
+ + rec_data_length(&p->u)
+ + rec_extra_padding(&p->u);
+ } else if (rec_magic(&p->u) == TDB_FTABLE_MAGIC) {
+ len = sizeof(p->u)
+ + rec_data_length(&p->u)
+ + rec_extra_padding(&p->u);
+ tally_add(ftables, rec_data_length(&p->u));
+ tally_add(extra, rec_extra_padding(&p->u));
+ } else if (rec_magic(&p->u) == TDB_CHAIN_MAGIC) {
+ len = sizeof(p->u)
+ + rec_data_length(&p->u)
+ + rec_extra_padding(&p->u);
+ tally_add(chains, 1);
tally_add(extra, rec_extra_padding(&p->u));
+ } else {
+ len = dead_space(tdb, off);
+ if (TDB_OFF_IS_ERR(len)) {
+ return len;
+ }
}
+ tdb_access_release(tdb, p);
}
if (unc)
tally_add(uncoal, unc);
- return 1ULL << zhdr.zone_bits;
+ return TDB_SUCCESS;
}
#define SUMMARY_FORMAT \
"Size of file/data: %zu/%zu\n" \
- "Number of zones: %zu\n" \
- "Smallest/average/largest zone size: %zu/%zu/%zu\n%s" \
"Number of records: %zu\n" \
"Smallest/average/largest keys: %zu/%zu/%zu\n%s" \
"Smallest/average/largest data: %zu/%zu/%zu\n%s" \
"Number of uncoalesced records: %zu\n" \
"Smallest/average/largest uncoalesced runs: %zu/%zu/%zu\n%s" \
"Toplevel hash used: %u of %u\n" \
+ "Number of chains: %zu\n" \
"Number of subhashes: %zu\n" \
"Smallest/average/largest subhash entries: %zu/%zu/%zu\n%s" \
- "Percentage keys/data/padding/free/rechdrs/zonehdrs/hashes: %.0f/%.0f/%.0f/%.0f/%.0f/%.0f/%.0f\n"
+ "Percentage keys/data/padding/free/rechdrs/freehdrs/hashes: %.0f/%.0f/%.0f/%.0f/%.0f/%.0f/%.0f\n"
#define BUCKET_SUMMARY_FORMAT_A \
"Free bucket %zu: total entries %zu.\n" \
#define HISTO_WIDTH 70
#define HISTO_HEIGHT 20
-char *tdb_summary(struct tdb_context *tdb, enum tdb_summary_flags flags)
+enum TDB_ERROR tdb_summary(struct tdb_context *tdb,
+ enum tdb_summary_flags flags,
+ char **summary)
{
- tdb_off_t off;
tdb_len_t len;
- unsigned int i, num_buckets, max_bucket = 0;
- uint64_t total_buckets = 0;
- struct tally *zones, *hashes, *freet, *keys, *data, *extra, *uncoal,
- *buckets[BUCKETS_FOR_ZONE(63)+1] = { NULL };
- char *zonesg, *hashesg, *freeg, *keysg, *datag, *extrag, *uncoalg,
- *bucketsg[BUCKETS_FOR_ZONE(63)+1] = { NULL };
- char *ret = NULL;
+ struct tally *ftables, *hashes, *freet, *keys, *data, *extra, *uncoal,
+ *chains;
+ char *hashesg, *freeg, *keysg, *datag, *extrag, *uncoalg;
+ enum TDB_ERROR ecode;
+
+ hashesg = freeg = keysg = datag = extrag = uncoalg = NULL;
- if (tdb_allrecord_lock(tdb, F_RDLCK, TDB_LOCK_WAIT, false) != 0)
- return NULL;
+ ecode = tdb_allrecord_lock(tdb, F_RDLCK, TDB_LOCK_WAIT, false);
+ if (ecode != TDB_SUCCESS) {
+ return tdb->last_error = ecode;
+ }
- if (tdb_lock_expand(tdb, F_RDLCK) != 0) {
+ ecode = tdb_lock_expand(tdb, F_RDLCK);
+ if (ecode != TDB_SUCCESS) {
tdb_allrecord_unlock(tdb, F_RDLCK);
- return NULL;
+ return tdb->last_error = ecode;
}
/* Start stats off empty. */
- zones = tally_new(HISTO_HEIGHT);
+ ftables = tally_new(HISTO_HEIGHT);
hashes = tally_new(HISTO_HEIGHT);
freet = tally_new(HISTO_HEIGHT);
keys = tally_new(HISTO_HEIGHT);
data = tally_new(HISTO_HEIGHT);
extra = tally_new(HISTO_HEIGHT);
uncoal = tally_new(HISTO_HEIGHT);
- if (!zones || !hashes || !freet || !keys || !data || !extra
- || !uncoal) {
- tdb->ecode = TDB_ERR_OOM;
+ chains = tally_new(HISTO_HEIGHT);
+ if (!ftables || !hashes || !freet || !keys || !data || !extra
+ || !uncoal || !chains) {
+ ecode = tdb_logerr(tdb, TDB_ERR_OOM, TDB_LOG_ERROR,
+ "tdb_summary: failed to allocate"
+ " tally structures");
goto unlock;
}
- for (i = 0; i < sizeof(buckets)/sizeof(buckets[0]); i++) {
- buckets[i] = tally_new(HISTO_HEIGHT);
- if (!buckets[i]) {
- tdb->ecode = TDB_ERR_OOM;
- goto unlock;
- }
- }
-
- for (off = sizeof(struct tdb_header);
- off < tdb->map_size - 1;
- off += len) {
- uint64_t bucketlen[BUCKETS_FOR_ZONE(63)+1] = { 0 };
- len = summarize_zone(tdb, off, zones, hashes, freet, keys,
- data, extra, uncoal, bucketlen,
- &num_buckets);
- if (len == TDB_OFF_ERR)
- goto unlock;
- for (i = 0; i < num_buckets; i++)
- tally_add(buckets[i], bucketlen[i]);
- if (num_buckets > max_bucket)
- max_bucket = num_buckets;
- total_buckets += num_buckets;
+ ecode = summarize(tdb, hashes, ftables, freet, keys, data, extra,
+ uncoal, chains);
+ if (ecode != TDB_SUCCESS) {
+ goto unlock;
}
if (flags & TDB_SUMMARY_HISTOGRAMS) {
- zonesg = tally_histogram(zones, HISTO_WIDTH, HISTO_HEIGHT);
hashesg = tally_histogram(hashes, HISTO_WIDTH, HISTO_HEIGHT);
freeg = tally_histogram(freet, HISTO_WIDTH, HISTO_HEIGHT);
keysg = tally_histogram(keys, HISTO_WIDTH, HISTO_HEIGHT);
datag = tally_histogram(data, HISTO_WIDTH, HISTO_HEIGHT);
extrag = tally_histogram(extra, HISTO_WIDTH, HISTO_HEIGHT);
uncoalg = tally_histogram(uncoal, HISTO_WIDTH, HISTO_HEIGHT);
- for (i = 0; i < sizeof(buckets)/sizeof(buckets[0]); i++) {
- bucketsg[i] = tally_histogram(buckets[i],
- HISTO_WIDTH,
- HISTO_HEIGHT);
- }
- } else {
- zonesg = hashesg = freeg = keysg = datag = extrag = uncoalg
- = NULL;
}
/* 20 is max length of a %llu. */
len = strlen(SUMMARY_FORMAT) + 33*20 + 1
- + (zonesg ? strlen(zonesg) : 0)
+ (hashesg ? strlen(hashesg) : 0)
+ (freeg ? strlen(freeg) : 0)
+ (keysg ? strlen(keysg) : 0)
+ (datag ? strlen(datag) : 0)
+ (extrag ? strlen(extrag) : 0)
+ (uncoalg ? strlen(uncoalg) : 0);
- for (i = 0; i < max_bucket; i++) {
- len += strlen(BUCKET_SUMMARY_FORMAT_B) + 6 * 20
- + (bucketsg[i] ? strlen(bucketsg[i]) : 0);
- }
- ret = malloc(len);
- if (!ret)
+ *summary = malloc(len);
+ if (!*summary) {
+ ecode = tdb_logerr(tdb, TDB_ERR_OOM, TDB_LOG_ERROR,
+ "tdb_summary: failed to allocate string");
goto unlock;
-
- len = sprintf(ret, SUMMARY_FORMAT,
- (size_t)tdb->map_size,
- tally_num(keys) + tally_num(data),
- tally_num(zones),
- tally_min(zones), tally_mean(zones), tally_max(zones),
- zonesg ? zonesg : "",
- tally_num(keys),
- tally_min(keys), tally_mean(keys), tally_max(keys),
- keysg ? keysg : "",
- tally_min(data), tally_mean(data), tally_max(data),
- datag ? datag : "",
- tally_min(extra), tally_mean(extra), tally_max(extra),
- extrag ? extrag : "",
- tally_num(freet),
- tally_min(freet), tally_mean(freet), tally_max(freet),
- freeg ? freeg : "",
- tally_total(uncoal, NULL),
- tally_min(uncoal), tally_mean(uncoal), tally_max(uncoal),
- uncoalg ? uncoalg : "",
- count_hash(tdb, offsetof(struct tdb_header, hashtable),
- TDB_TOPLEVEL_HASH_BITS),
- 1 << TDB_TOPLEVEL_HASH_BITS,
- tally_num(hashes),
- tally_min(hashes), tally_mean(hashes), tally_max(hashes),
- hashesg ? hashesg : "",
- tally_total(keys, NULL) * 100.0 / tdb->map_size,
- tally_total(data, NULL) * 100.0 / tdb->map_size,
- tally_total(extra, NULL) * 100.0 / tdb->map_size,
- tally_total(freet, NULL) * 100.0 / tdb->map_size,
- (tally_num(keys) + tally_num(freet) + tally_num(hashes))
- * sizeof(struct tdb_used_record) * 100.0 / tdb->map_size,
- (tally_num(zones) * sizeof(struct free_zone_header)
- + total_buckets * sizeof(tdb_off_t))
- * 100.0 / tdb->map_size,
- (tally_num(hashes)
- * (sizeof(tdb_off_t) << TDB_SUBLEVEL_HASH_BITS)
- + (sizeof(tdb_off_t) << TDB_TOPLEVEL_HASH_BITS))
- * 100.0 / tdb->map_size);
-
- for (i = 0; i < max_bucket; i++) {
- size_t min, max;
- sizes_for_bucket(i, &min, &max);
- if (min == max) {
- len += sprintf(ret + len, BUCKET_SUMMARY_FORMAT_A,
- min, tally_total(buckets[i], NULL),
- tally_min(buckets[i]),
- tally_mean(buckets[i]),
- tally_max(buckets[i]),
- bucketsg[i] ? bucketsg[i] : "");
- } else {
- len += sprintf(ret + len, BUCKET_SUMMARY_FORMAT_B,
- min, max, tally_total(buckets[i], NULL),
- tally_min(buckets[i]),
- tally_mean(buckets[i]),
- tally_max(buckets[i]),
- bucketsg[i] ? bucketsg[i] : "");
- }
}
+ sprintf(*summary, SUMMARY_FORMAT,
+ (size_t)tdb->file->map_size,
+ tally_total(keys, NULL) + tally_total(data, NULL),
+ tally_num(keys),
+ tally_min(keys), tally_mean(keys), tally_max(keys),
+ keysg ? keysg : "",
+ tally_min(data), tally_mean(data), tally_max(data),
+ datag ? datag : "",
+ tally_min(extra), tally_mean(extra), tally_max(extra),
+ extrag ? extrag : "",
+ tally_num(freet),
+ tally_min(freet), tally_mean(freet), tally_max(freet),
+ freeg ? freeg : "",
+ tally_total(uncoal, NULL),
+ tally_min(uncoal), tally_mean(uncoal), tally_max(uncoal),
+ uncoalg ? uncoalg : "",
+ (unsigned)count_hash(tdb, offsetof(struct tdb_header,
+ hashtable),
+ TDB_TOPLEVEL_HASH_BITS),
+ 1 << TDB_TOPLEVEL_HASH_BITS,
+ tally_num(chains),
+ tally_num(hashes),
+ tally_min(hashes), tally_mean(hashes), tally_max(hashes),
+ hashesg ? hashesg : "",
+ tally_total(keys, NULL) * 100.0 / tdb->file->map_size,
+ tally_total(data, NULL) * 100.0 / tdb->file->map_size,
+ tally_total(extra, NULL) * 100.0 / tdb->file->map_size,
+ tally_total(freet, NULL) * 100.0 / tdb->file->map_size,
+ (tally_num(keys) + tally_num(freet) + tally_num(hashes))
+ * sizeof(struct tdb_used_record) * 100.0 / tdb->file->map_size,
+ tally_num(ftables) * sizeof(struct tdb_freetable)
+ * 100.0 / tdb->file->map_size,
+ (tally_num(hashes)
+ * (sizeof(tdb_off_t) << TDB_SUBLEVEL_HASH_BITS)
+ + (sizeof(tdb_off_t) << TDB_TOPLEVEL_HASH_BITS)
+ + sizeof(struct tdb_chain) * tally_num(chains))
+ * 100.0 / tdb->file->map_size);
+
unlock:
- free(zonesg);
free(hashesg);
free(freeg);
free(keysg);
free(datag);
free(extrag);
free(uncoalg);
- free(zones);
free(hashes);
free(freet);
free(keys);
free(data);
free(extra);
free(uncoal);
- for (i = 0; i < sizeof(buckets)/sizeof(buckets[0]); i++) {
- free(buckets[i]);
- free(bucketsg[i]);
- }
+ free(ftables);
+ free(chains);
tdb_allrecord_unlock(tdb, F_RDLCK);
tdb_unlock_expand(tdb, F_RDLCK);
- return ret;
+ return tdb->last_error = ecode;
}