shithub: gefs

ref: f0563cbdbce1777d5672fee507acac7ab0d6dbc0
dir: /blk.c/

View raw version
#include <u.h>
#include <libc.h>
#include <fcall.h>
#include <avl.h>
#include <pool.h>

#include "dat.h"
#include "fns.h"

typedef struct Range Range;
struct Range {
	vlong off;
	vlong len;
};

static vlong	blkalloc_lk(Arena*);
static int	blkdealloc(vlong);
static void	cachedel(vlong);
static Blk	*cacheblk(Blk*);
static Blk	*lookupblk(vlong);

Blk*
readblk(vlong bp, int flg)
{
	Blk *b;
	vlong off, rem, n;

	assert(bp != -1);
	if((b = malloc(sizeof(Blk))) == nil)
		return nil;
	off = bp;
	rem = Blksz;
	while(rem != 0){
		n = pread(fs->fd, b->buf, rem, off);
		if(n <= 0){
			free(b);
			return nil;
		}
		off += n;
		rem -= n;
	}
	memset(&b->RWLock, 0, sizeof(RWLock));
	b->type = (flg&GBraw) ? Traw : GBIT16(b->buf+0);
	b->off = bp;
	b->cnext = nil;
	b->cprev = nil;
	b->hnext = nil;
	b->data = b->buf + 10;
	switch(b->type){
	default:
		if(flg&GBraw)
			break;
		fprint(2, "invalid block @%llx\n", bp);
		abort();
		break;
	case Tarena:
	case Traw:
	case Tsuper:
	case Tlog:
		break;
	case Tpivot:
		b->nval = GBIT16(b->buf+2);
		b->valsz = GBIT16(b->buf+4);
		b->nbuf = GBIT16(b->buf+6);
		b->bufsz = GBIT16(b->buf+8);
		break;
	case Tleaf:
		b->nval = GBIT16(b->buf+2);
		b->valsz = GBIT16(b->buf+4);
		break;
	}
	return b;
}

static Arena*
pickarena(vlong hint)
{
	long n;

	n = -1; /* shut up, ken */
	if(hint > 0 || hint < fs->narena)
		n = hint / fs->arenasz;
	else if(hint == -1)
		n = ainc(&fs->nextarena) % fs->narena;
	else
		abort();
	return &fs->arenas[n];
}

Arena*
getarena(vlong b)
{
	int i;

	i = b / fs->arenasz;
	if(i < 0 || i >= fs->narena){
		werrstr("out of range block %lld", b);
		abort();
//		return nil;
	}
	return &fs->arenas[i];
}

static int
freerange(Avltree *t, vlong off, vlong len)
{
	Arange *r, *s;

	assert(len % Blksz == 0);
	if((r = calloc(1, sizeof(Arange))) == nil)
		return -1;
	r->off = off;
	r->len = len;
	avlinsert(t, r);

Again:

	s = (Arange*)avlprev(r);
	if(s != nil && s->off+s->len == r->off){
		avldelete(t, r);
		s->len = s->len + r->len;
		free(r);
		r = s;
		goto Again;
	}
	s = (Arange*)avlnext(r);
	if(s != nil && r->off+r->len == s->off){
		avldelete(t, r);
		s->off = r->off;
		s->len = s->len + r->len;
		free(r);
		r = s;
		goto Again;
	}
	return 0;
}

int
grabrange(Avltree *t, vlong off, vlong len)
{
	Arange *r, *s, q;
	vlong l;

	assert(len % Blksz == 0);
	q.off = off;
	q.len = len;
	r = (Arange*)avllookup(t, &q.Avl, -1);
	if(r == nil || off + len > r->off + r->len)
		abort();

	print("\tmerge (%llx,%llx) (%llx,%llx)\n", off, len, r->off, r->len);
	if(off == r->off){
		r->off += len;
		r->len -= len;
	}else if(off + len == r->off + r->len){
		r->len -= len;
	}else if(off > r->off && off+len < r->off + r->len){
		if((s = malloc(sizeof(Arange))) == nil)
			return -1;
		l = r->len;
		s->off = off + len;
		r->len = off - r->off;
		s->len = l - r->len - len;
		avlinsert(t, s);
	}else
		abort();

	if(r->len == 0){
		avldelete(t, r);
		free(r);
	}
	return 0;
}

Blk*
logappend(Arena *a, Blk *lb, vlong off, vlong len, int op)
{
	Blk *pb;
	vlong o;
	char *p;

	assert(off % Blksz == 0);
	assert(op == LgAlloc || op == LgFree);
	if(lb == nil || lb->logsz > Logspc - 8){
		pb = lb;
		if((o = blkalloc_lk(a)) == -1)
			return nil;
		if((lb = mallocz(sizeof(Blk), 1)) == nil)
			return nil;
		lb->data = lb->buf + Hdrsz;
		lb->flag |= Bdirty;
		lb->type = Tlog;
		lb->off = o;
		lb->logsz = Loghdsz;
		p = lb->data + lb->logsz;
		PBIT64(p + 0, (uvlong)LgEnd);
		finalize(lb);
		if(syncblk(lb) == -1){
			free(lb);
			return nil;
		}

		a->logtl = lb;
		if(pb != nil){
			p = pb->data + pb->logsz;
			PBIT64(p + 0, lb->off|LgChain);
			finalize(pb);
			if(syncblk(pb) == -1)
				return nil;
		}
	}

	p = lb->data + lb->logsz;
	if(len == Blksz){
		off |= (op & ~Lg2w);
		PBIT64(p, off);
		lb->logsz += 8;
	}else{
		off |= op;
		PBIT64(p+0, off);
		PBIT64(p+8, len);
		lb->logsz += 16;
	}
	/* this gets overwritten by the next append */
	p = lb->data + lb->logsz;
	PBIT64(p, (uvlong)LgEnd);
	return lb;
}

/*
 * Logs an allocation. Must be called
 * with arena lock held. Duplicates some/c
 * of the work in allocblk to prevent
 * recursion.
 */
int
logalloc(Arena *a, vlong off, int op)
{
	Blk *b;

	if((b = logappend(a, a->logtl, off, Blksz, op)) == nil)
		return -1;
	if(a->log == -1)
		a->log = b->off;
	if(b != a->logtl)
		a->logtl = b;
	return 0;
}

int
loadlog(Arena *a)
{
	Blk *b;
	vlong bp, ent, off, len;
	uvlong bh;
	char *p, *d;
	int op, i, n;


	bp = a->log;

Nextblk:
	dprint("block: %llx\n", bp);
	if((b = readblk(bp, 0)) == nil)
		return -1;
	p = b->data;
	bh = GBIT64(p + 0);
	/* the hash covers the log and offset */
	if(bh != siphash(p+8, Blkspc-8)){
		werrstr("corrupt log");
		return -1;
	}
	for(i = Loghdsz; i < Logspc; i += n){
		d = b->data + i;
		ent = GBIT64(d);
		op = ent & 0xff;
		off = ent & ~0xff;
		n = (op & Lg2w) ? 16 : 8;
		switch(op){
		case LgEnd:
			dprint("log@%d: end\n", i);
			/*
			 * since we want the next insertion to overwrite
			 * this, don't include the size in this entry.
			 */
			b->logsz = i;
			return 0;
		case LgChain:
			bp = off & ~0xff;
			dprint("log@%d: chain %llx\n", i, bp);
			b->logsz = i+n;
			goto Nextblk;
			break;

		case LgFlush:
			dprint("log@%d: flush: %llx\n", i, off>>8);
			lock(&fs->genlk);
			fs->gen = off >> 8;
			unlock(&fs->genlk);
			break;
		case LgAlloc:
		case LgAlloc1:
			len = (op & Lg2w) ? GBIT64(d+8) : Blksz;
			dprint("log@%d alloc: %llx+%llx\n", i, off, len);
			if(grabrange(a->free, off & ~0xff, len) == -1)
				return -1;
			break;
		case LgFree:
		case LgFree1:
			len = (op & Lg2w) ? GBIT64(d+8) : Blksz;
			dprint("log@%d free: %llx+%llx\n", i, off, len);
			if(freerange(a->free, off & ~0xff, len) == -1)
				return -1;
			break;
		case LgRef:
		case LgUnref:
			fprint(2, "unimplemented ref op at log@%d: log op %d\n", i, op);
			break;
		default:
			n = 0;
			dprint("log@%d: log op %d\n", i, op);
			abort();
			break;
		}
showfree("after");
	}
	return -1;
}

int
compresslog(Arena *a)
{
	Arange *r;
	Range *log, *nlog;
	vlong v, bp, nb, graft, oldhd;
	int i, n, sz;
	Blk *hd, *ab, *b;
	char *p;

showfree("precompress");
fprint(2, "compress start\n");

	/*
	 * Sync the current log to disk, and
	 * set up a new block log tail.  While
	 * compressing the log, nothing else is
	 * using this arena, so any allocs come
	 * from the log compression, and go into
	 * this new log segment.
	 *
	 * A bit of copy paste from newblk,
	 * because otherwise we have a deadlock
	 * allocating the block.
	 */
	if((bp = blkalloc_lk(a)) == -1)
		return -1;
	if((b = mallocz(sizeof(Blk), 1)) == nil)
		return -1;
	b->type = Tlog;
	b->flag = Bdirty;
	b->off = bp;
	b->ref = 1;
	b->data = b->buf + Hdrsz;
	b->logsz = Loghdsz;

	PBIT64(b->data+b->logsz, (uvlong)LgEnd);
	finalize(b);
	if(syncblk(b) == -1){
		free(b);
		return -1;
	}

	graft = b->off;
	if(a->logtl != nil){
		finalize(a->logtl);
		if(syncblk(a->logtl) == -1){
			free(b);
			return -1;
		}
	}
	a->logtl = b;
print("\tnew log block: %llx\n", b->off);

	/*
	 * Prepare what we're writing back.
	 * Arenas must be sized so that we can
	 * keep the merged log in memory for
	 * a rewrite.
	 */
	n = 0;
	sz = 512;
	if((log = malloc(sz*sizeof(Range))) == nil)
		return -1;
	for(r = (Arange*)avlmin(a->free); r != nil; r = (Arange*)avlnext(r)){
		if(n == sz){
			sz *= 2;
			if((nlog = realloc(log, sz*sizeof(Range))) == nil){
				free(log);
				return -1;
			}
			log = nlog;
		}
		log[n].off = r->off;
		log[n].len = r->len;
		n++;
	}
	if((b = newblk(Tlog)) == nil){
		free(log);
		return -1;
	}
	hd = b;
	b->logsz = Loghdsz;
	for(i = 0; i < n; i++)
		if((b = logappend(a, b, log[i].off, log[i].len, LgFree)) == nil)
			return -1;
	p = b->data + b->logsz;
	PBIT64(p, LgChain|graft);
	free(log);
	finalize(b);
	if(syncblk(b) == -1)
		return -1;

	oldhd = a->log;
	a->log = hd->off;
	a->logh = blkhash(hd);
	ab = a->b;
	PBIT64(ab->data + 0, a->log);
	PBIT64(ab->data + 8, a->logh);
	if(syncblk(ab) == -1)
		return -1;
checkfs();
showfree("postcompress");
	if(oldhd != -1){
		for(bp = oldhd; bp != -1; bp = nb){
			nb = -1;
			if((b = readblk(bp, 0)) == nil)
				return -1;
			for(i = Loghdsz; i < Logspc; i += n){
				p = b->data + i;
				v = GBIT64(p);
				n = (v & Lg2w) ? 16 : 8;
				if((v&0xff) == LgChain){
					nb = v & ~0xff;
					break;
				}else if((v&0xff) == LgEnd){
					nb = -1;
					break;
				}
			}
			fprint(2, "\tpostscan: freeing %llx\n", bp);
			if(blkdealloc(bp) == -1)
				return -1;
		}
	}
	finalize(a->logtl);
	if(syncblk(a->logtl) == -1)
		return -1;
showfree("postreclaim");
	return 0;
}
/*
 * Allocate from an arena, with lock
 * held. May be called recursively, to
 * alloc space for the alloc log.
 */
static vlong
blkalloc_lk(Arena *a)
{
	Avltree *t;
	Arange *r;
	vlong b;

	t = a->free;
	r = (Arange*)t->root;
	if(r == nil){
		unlock(a);
		return -1;
	}

	/*
	 * A bit of sleight of hand here:
	 * while we're changing the sorting
	 * key, but we know it won't change
	 * the sort order because the tree
	 * covers disjoint ranges
	 */
	b = r->off;
	r->len -= Blksz;
	r->off += Blksz;
	if(r->len == 0){
		avldelete(t, r);
		free(r);
	}
fprint(2, "\talloc %llx\n", b);
	return b;
}

static int
blkdealloc(vlong b)
{
	Arena *a;
	int r;

	r = -1;
	a = getarena(b);
	lock(a);
	cachedel(b);
	if(freerange(a->free, b, Blksz) == -1)
		goto out;
	if(logalloc(a, b, LgFree) == -1)
		goto out;
	r = 0;
out:
	unlock(a);
	return r;
}

vlong
blkalloc(vlong hint)
{
	Arena *a;
	vlong b;
	int tries;

	tries = 0;
again:
	a = pickarena(hint);
	if(a == nil || tries == fs->narena){
		werrstr("no empty arenas");
		return -1;
	}
	lock(a);
	/*
	 * TODO: there's an extreme edge case
	 * here.
	 *
	 * If the file system has room to alloc
	 * a data block but no log block, then
	 * we end up with it in a stuck state.
	 * The fix is to reserve alloc blocks,
	 * so that we're guaranteed to be able
	 * to log an alloc if the disk is working
	 * correctly.
	 */
	tries++;
	if((b = blkalloc_lk(a)) == -1){
		unlock(a);
		goto again;
	}
	if(logalloc(a, b, LgAlloc) == -1){
		unlock(a);
		return -1;
	}
	unlock(a);
	return b;
}

Blk*
newblk(int t)
{
	Blk *b;
	vlong bp;

	if((bp = blkalloc(-1)) == -1)
		return nil;
	if((b = mallocz(sizeof(Blk), 1)) == nil)
		return nil;
	b->type = t;
	b->flag = Bdirty;
	b->off = bp;
	b->ref = 1;
	b->data = b->buf + Hdrsz;
	return cacheblk(b);
}

static Blk*
lookupblk(vlong off)
{
	Bucket *bkt;
	u32int h;
	Blk *b;

	h = ihash(off);

	bkt = &fs->cache[h % fs->cmax];
	lock(bkt);
	for(b = bkt->b; b != nil; b = b->hnext)
		if(b->off == off)
			break;
	if(b != nil)
		pinblk(b);
	unlock(bkt);
	return b;
}

static Blk*
cacheblk(Blk *b)
{
	Bucket *bkt;
	Blk *e, *c;
	u32int h;

	/* FIXME: better hash. */
	assert(b->off != 0);
	h = ihash(b->off);
//	dprint("cache %lld (h=%xm, bkt=%d) => %p\n", b->off, h%fs->cmax, h, b);
	ainc(&b->ref);
	bkt = &fs->cache[h % fs->cmax];
	lock(bkt);
	for(e = bkt->b; e != nil; e = e->hnext){
		if(b == e)
			goto found;
		assert(b->off != e->off);
	}
	bkt->b = b;
found:
	ainc(&b->ref);
	unlock(bkt);

	lock(&fs->lrulk);
	if(b == fs->chead)
		goto Cached;
	if(b == fs->ctail)
		fs->ctail = b->cprev;

	if(b->cnext != nil)
		b->cnext->cprev = b->cprev;
	if(b->cprev != nil)
		b->cprev->cnext = b->cnext;
	if(fs->ctail == nil)
		fs->ctail = b;
	if(fs->chead != nil)
		fs->chead->cprev = b;
	if(fs->ctail == nil)
		fs->ctail = b;
	b->cnext = fs->chead;
	b->cprev = nil;
	fs->chead = b;
	if((b->flag&Bcache) == 0){
		b->flag |= Bcache;
		fs->ccount++;
		ainc(&b->ref);
	}
	c=0;
	USED(c);
/*
	for(c = fs->ctail; c != nil && fs->ccount >= fs->cmax; c = fs->ctail){
		fs->ctail = c->cprev;
		fs->ccount--; 
		putblk(c);
	}
*/
Cached:
	unlock(&fs->lrulk);
	return b;
}

static void
cachedel(vlong del)
{
	Bucket *bkt;
	Blk *b, **p;
	u32int h;

	/* FIXME: better hash. */
	h = ihash(del);

	bkt = &fs->cache[h % fs->cmax];
	lock(bkt);
	p = &bkt->b;
	for(b = bkt->b; b != nil; b = b->hnext){
		if(b->off == del){
			*p = b->hnext;
			break;
		}
		p = &b->hnext;
	}
	unlock(bkt);
	if(b == nil)
		return;

	lock(&fs->lrulk);
	if(b->cnext != nil)
		b->cnext->cprev = b->cprev;
	if(b->cprev != nil)
		b->cprev->cnext = b->cnext;
	if(fs->ctail == b)
		fs->ctail = b->cprev;
	if(fs->chead == nil)
		fs->chead = b;
	unlock(&fs->lrulk);
}

int
syncblk(Blk *b)
{
	dprint("\tsyncblk: %llx+%llx\n", b->off, Blksz);
	return pwrite(fs->fd, b->buf, Blksz, b->off);
}


void
enqueue(Blk *b)
{
	if(syncblk(b) == -1){
		ainc(&fs->broken);
		fprint(2, "write: %r");
		return;
	}
	wlock(b);
	b->flag &= ~(Bqueued|Bdirty|Bfinal);
	wunlock(b);

}

void
fillsuper(Blk *b)
{
	char *p;

	assert(b->type == Tsuper);
	p = b->data;
	memcpy(p +  0, "gefs0001", 8);
	PBIT32(p +  8, 0); /* dirty */
	PBIT32(p + 12, Blksz);
	PBIT32(p + 16, Bufspc);
	PBIT32(p + 20, Hdrsz);
	PBIT32(p + 24, fs->height);
	PBIT64(p + 32, fs->rootb);
	PBIT64(p + 40, fs->rooth);
	PBIT32(p + 48, fs->narena);
	PBIT64(p + 56, fs->arenasz);
	PBIT64(p + 64, fs->gen);
	PBIT64(p + 72, fs->nextqid);
}

void
finalize(Blk *b)
{
	vlong h;

//	assert((b->flag & Bfinal) == 0);
	b->flag |= Bfinal;
	PBIT16(b->buf, b->type);
	switch(b->type){
	default:
	case Tnone:
		abort();
		break;
	case Tpivot:
		PBIT16(b->buf+2, b->nval);
		PBIT16(b->buf+4, b->valsz);
		PBIT16(b->buf+6, b->nbuf);
		PBIT16(b->buf+8, b->bufsz);
		break;
	case Tleaf:
		PBIT16(b->buf+2, b->nval);
		PBIT16(b->buf+4, b->valsz);
		break;
	case Tlog:
		h = siphash(b->data + 8, Blkspc-8);
		PBIT64(b->data, h);
	case Tsuper:
	case Tarena:
	case Traw:
		break;
	}
}

Blk*
getblk(vlong bp, uvlong bh)
{
	Blk *b;

	if((b = lookupblk(bp)) == nil){
		if((b = readblk(bp, 0)) == nil)
			return nil;
		if(siphash(b->buf, Blksz) != bh){
			werrstr("corrupt block %llx", bp);
			return nil;
		}
	}
	return cacheblk(b);
}

Blk*
dupblk(vlong bp, uvlong bh)
{
	USED(bp, bh);
	return nil;
}

Blk*
pinblk(Blk *b)
{
	ainc(&b->ref);
	return b;
}

int
refblk(Blk *b)
{
	Arena *a;
	int r;

	a = getarena(b->off);
	lock(a);
	r = logalloc(a, b->off, LgRef);
	unlock(a);
	return r;
}

int
unrefblk(Blk *b)
{
	Arena *a;
	int r;

	a = getarena(b->off);
	lock(a);
	r = logalloc(a, b->off, LgUnref);
	unlock(a);
	return r;
}

ushort
blkfill(Blk *b)
{
	switch(b->type){
	case Tpivot:
		return 2*b->nbuf + b->bufsz +  2*b->nval + b->valsz;
	case Tleaf:
		return 2*b->nval + b->valsz;
	default:
		fprint(2, "invalid block @%lld\n", b->off);
		abort();
	}
	return 0; // shut up kencc
}

void
putblk(Blk *b)
{
	if(b == nil)
		return;
	if((b->flag & (Bdirty|Bqueued)) == Bdirty)
		enqueue(b);
	if(adec(&b->ref) == 0){
		cachedel(b->off);
		free(b);
	}
}

void
freeblk(Blk *b)
{
	Arena *a;

	assert(b->ref == 1 && b->flag & (Bdirty|Bqueued) == Bdirty);
	a = getarena(b->off);
	lock(a);
	/*
	 * TODO: what to do if we fail to log a free here??
	 * This is already an error path!
	 */
	logalloc(a, b->off, LgUnref);
	blkdealloc(b->off);
	unlock(a);
	free(b);
}

int
sync(void)
{
	int i, r;
	Blk *b;

	dprint("syncing\n");
	r = 0;
	for(i = 0; i < fs->narena; i++){
		b = fs->arenas[i].logtl;
		finalize(b);
		if(syncblk(b) == -1)
			r = -1;
	}
	/* FIXME: hit it with a big hammer -- flush the whole cache */
	for(b = fs->chead; b != nil; b = b->cnext){
//		dprint("sync %p\n", b);
		if(!(b->flag & Bdirty))
			continue;
		if(syncblk(b) == -1)
			r = -1;
	}
	return r;
}