ref: 34be16ef74b0c991850ee82931e9bd21f01a0a4f
dir: /blk.c/
#include <u.h> #include <libc.h> #include <fcall.h> #include <avl.h> #include "dat.h" #include "fns.h" typedef struct Range Range; struct Range { vlong off; vlong len; }; static vlong blkalloc_lk(Arena*); static int blkdealloc_lk(vlong); static void cachedel(vlong); QLock blklock; Blk* readblk(vlong bp, int flg) { Blk *b; vlong off, rem, n; assert(bp != -1); if((b = mallocz(sizeof(Blk), 1)) == nil) return nil; off = bp; rem = Blksz; while(rem != 0){ n = pread(fs->fd, b->buf, rem, off); if(n <= 0){ free(b); return nil; } off += n; rem -= n; } memset(&b->Lock, 0, sizeof(Lock)); b->type = (flg&GBraw) ? Traw : GBIT16(b->buf+0); b->bp.addr = bp; b->bp.hash = -1; b->bp.gen = -1; b->ref = 1; b->cnext = nil; b->cprev = nil; b->hnext = nil; b->data = b->buf + 10; switch(b->type){ default: if(flg&GBraw) break; fprint(2, "invalid block @%llx\n", bp); abort(); break; case Tarena: case Traw: case Tsuper: case Tlog: break; case Tpivot: b->nval = GBIT16(b->buf+2); b->valsz = GBIT16(b->buf+4); b->nbuf = GBIT16(b->buf+6); b->bufsz = GBIT16(b->buf+8); break; case Tleaf: b->nval = GBIT16(b->buf+2); b->valsz = GBIT16(b->buf+4); break; } return b; } static Arena* pickarena(vlong hint) { long n; n = -1; /* shut up, ken */ if(hint > 0 || hint < fs->narena) n = hint / fs->arenasz; else if(hint == -1) n = ainc(&fs->nextarena) % fs->narena; else abort(); return &fs->arenas[n]; } Arena* getarena(vlong b) { int i; i = b / fs->arenasz; if(i < 0 || i >= fs->narena){ werrstr("out of range block %lld", b); abort(); // return nil; } return &fs->arenas[i]; } static int freerange(Avltree *t, vlong off, vlong len) { Arange *r, *s; assert(len % Blksz == 0); if((r = calloc(1, sizeof(Arange))) == nil) return -1; r->off = off; r->len = len; avlinsert(t, r); Again: s = (Arange*)avlprev(r); if(s != nil && s->off+s->len == r->off){ avldelete(t, r); s->len = s->len + r->len; free(r); r = s; goto Again; } s = (Arange*)avlnext(r); if(s != nil && r->off+r->len == s->off){ avldelete(t, r); s->off = r->off; s->len = s->len + r->len; free(r); r = s; goto Again; } return 0; } int grabrange(Avltree *t, vlong off, vlong len) { Arange *r, *s, q; vlong l; assert(len % Blksz == 0); q.off = off; q.len = len; r = (Arange*)avllookup(t, &q.Avl, -1); if(r == nil || off + len > r->off + r->len) abort(); if(off == r->off){ r->off += len; r->len -= len; }else if(off + len == r->off + r->len){ r->len -= len; }else if(off > r->off && off+len < r->off + r->len){ if((s = malloc(sizeof(Arange))) == nil) return -1; l = r->len; s->off = off + len; r->len = off - r->off; s->len = l - r->len - len; avlinsert(t, s); }else abort(); if(r->len == 0){ avldelete(t, r); free(r); } return 0; } Blk* logappend(Arena *a, Blk *lb, vlong off, vlong len, int op) { Blk *pb; vlong o; char *p; assert(off % Blksz == 0); assert(op == LogAlloc || op == LogFree); if(lb == nil || lb->logsz > Logspc - 8){ pb = lb; if((o = blkalloc_lk(a)) == -1) return nil; if((lb = mallocz(sizeof(Blk), 1)) == nil) return nil; lb->data = lb->buf + Hdrsz; lb->flag |= Bdirty; lb->type = Tlog; lb->bp.addr = o; lb->logsz = Loghdsz; p = lb->data + lb->logsz; PBIT64(p + 0, (uvlong)LogEnd); finalize(lb); if(syncblk(lb) == -1){ free(lb); return nil; } a->logtl = lb; if(pb != nil){ p = pb->data + pb->logsz; PBIT64(p + 0, lb->bp.addr|LogChain); finalize(pb); if(syncblk(pb) == -1) return nil; } } p = lb->data + lb->logsz; if(len == Blksz){ off |= (op & ~Log2w); PBIT64(p, off); lb->logsz += 8; }else{ off |= op; PBIT64(p+0, off); PBIT64(p+8, len); lb->logsz += 16; } /* this gets overwritten by the next append */ p = lb->data + lb->logsz; PBIT64(p, (uvlong)LogEnd); return lb; } /* * Logs an allocation. Must be called * with arena lock held. Duplicates some/c * of the work in allocblk to prevent * recursion. */ int logop(Arena *a, vlong off, int op) { Blk *b; if((b = logappend(a, a->logtl, off, Blksz, op)) == nil) return -1; if(a->log == -1) a->log = b->bp.addr; if(b != a->logtl) a->logtl = b; return 0; } int loadlog(Arena *a) { Blk *b; vlong bp, ent, off, len; uvlong bh; char *p, *d; int op, i, n; bp = a->log; Nextblk: if((b = readblk(bp, 0)) == nil) return -1; cacheblk(b); p = b->data; bh = GBIT64(p + 0); /* the hash covers the log and offset */ if(bh != siphash(p+8, Blkspc-8)){ werrstr("corrupt log"); return -1; } for(i = Loghdsz; i < Logspc; i += n){ d = b->data + i; ent = GBIT64(d); op = ent & 0xff; off = ent & ~0xff; n = (op & Log2w) ? 16 : 8; switch(op){ case LogEnd: dprint("log@%d: end\n", i); /* * since we want the next insertion to overwrite * this, don't include the size in this entry. */ b->logsz = i; return 0; case LogChain: bp = off & ~0xff; dprint("log@%d: chain %llx\n", i, bp); b->logsz = i+n; goto Nextblk; break; case LogFlush: dprint("log@%d: flush: %llx\n", i, off>>8); fs->nextgen = (off >> 8)+1; break; case LogAlloc: case LogAlloc1: len = (op & Log2w) ? GBIT64(d+8) : Blksz; dprint("log@%d alloc: %llx+%llx\n", i, off, len); if(grabrange(a->free, off & ~0xff, len) == -1) return -1; break; case LogFree: case LogFree1: len = (op & Log2w) ? GBIT64(d+8) : Blksz; dprint("log@%d free: %llx+%llx\n", i, off, len); if(freerange(a->free, off & ~0xff, len) == -1) return -1; break; default: n = 0; dprint("log@%d: log op %d\n", i, op); abort(); break; } } return -1; } int compresslog(Arena *a) { Arange *r; Range *log, *nlog; vlong v, bp, nb, graft, oldhd; int i, n, sz; Blk *hd, *ab, *b; char *p; /* * Sync the current log to disk, and * set up a new block log tail. While * compressing the log, nothing else is * using this arena, so any allocs come * from the log compression, and go into * this new log segment. * * A bit of copy paste from newblk, * because otherwise we have a deadlock * allocating the block. */ if((bp = blkalloc_lk(a)) == -1) return -1; if((b = mallocz(sizeof(Blk), 1)) == nil) return -1; b->type = Tlog; b->flag = Bdirty; b->bp.addr = bp; b->ref = 1; b->data = b->buf + Hdrsz; b->logsz = Loghdsz; PBIT64(b->data+b->logsz, (uvlong)LogEnd); finalize(b); if(syncblk(b) == -1){ free(b); return -1; } graft = b->bp.addr; if(a->logtl != nil){ finalize(a->logtl); if(syncblk(a->logtl) == -1){ free(b); return -1; } } a->logtl = b; /* * Prepare what we're writing back. * Arenas must be sized so that we can * keep the merged log in memory for * a rewrite. */ n = 0; sz = 512; if((log = malloc(sz*sizeof(Range))) == nil) return -1; for(r = (Arange*)avlmin(a->free); r != nil; r = (Arange*)avlnext(r)){ if(n == sz){ sz *= 2; if((nlog = realloc(log, sz*sizeof(Range))) == nil){ free(log); return -1; } log = nlog; } log[n].off = r->off; log[n].len = r->len; n++; } if((b = newblk(Tlog)) == nil){ free(log); return -1; } hd = b; b->logsz = Loghdsz; for(i = 0; i < n; i++) if((b = logappend(a, b, log[i].off, log[i].len, LogFree)) == nil) return -1; p = b->data + b->logsz; PBIT64(p, LogChain|graft); free(log); finalize(b); if(syncblk(b) == -1) return -1; oldhd = a->log; a->log = hd->bp.addr; a->logh = blkhash(hd); ab = a->b; PBIT64(ab->data + 0, a->log); PBIT64(ab->data + 8, a->logh); finalize(ab); if(syncblk(ab) == -1) return -1; if(oldhd != -1){ for(bp = oldhd; bp != -1; bp = nb){ nb = -1; if((b = readblk(bp, 0)) == nil) return -1; for(i = Loghdsz; i < Logspc; i += n){ p = b->data + i; v = GBIT64(p); n = (v & Log2w) ? 16 : 8; if((v&0xff) == LogChain){ nb = v & ~0xff; break; }else if((v&0xff) == LogEnd){ nb = -1; break; } } lock(a); if(blkdealloc_lk(bp) == -1){ unlock(a); return -1; } unlock(a); } } finalize(a->logtl); if(syncblk(a->logtl) == -1) return -1; return 0; } /* * Allocate from an arena, with lock * held. May be called recursively, to * alloc space for the alloc log. */ static vlong blkalloc_lk(Arena *a) { Avltree *t; Arange *r; vlong b; t = a->free; r = (Arange*)t->root; if(r == nil){ unlock(a); return -1; } /* * A bit of sleight of hand here: * while we're changing the sorting * key, but we know it won't change * the sort order because the tree * covers disjoint ranges */ b = r->off; r->len -= Blksz; r->off += Blksz; if(r->len == 0){ avldelete(t, r); free(r); } return b; } static int blkdealloc_lk(vlong b) { Arena *a; int r; r = -1; a = getarena(b); if(freerange(a->free, b, Blksz) == -1) goto out; if(logop(a, b, LogFree) == -1) goto out; r = 0; out: return r; } vlong blkalloc(vlong hint) { Arena *a; vlong b; int tries; tries = 0; Again: a = pickarena(hint); if(a == nil || tries == fs->narena){ werrstr("no empty arenas"); return -1; } lock(a); /* * TODO: there's an extreme edge case * here. * * If the file system has room to alloc * a data block but no log block, then * we end up with it in a stuck state. * The fix is to reserve alloc blocks, * so that we're guaranteed to be able * to log an alloc if the disk is working * correctly. */ tries++; if((b = blkalloc_lk(a)) == -1){ unlock(a); goto Again; } if(logop(a, b, LogAlloc) == -1){ unlock(a); return -1; } unlock(a); return b; } Blk* newblk(int t) { Blk *b; vlong bp; if((bp = blkalloc(-1)) == -1) return nil; if((b = lookupblk(bp)) == nil){ if((b = mallocz(sizeof(Blk), 1)) == nil) return nil; /* * If the block is cached, * then the cache holds a ref * to the block, so we only * want to reset the refs * on an allocation. */ b->ref = 1; b->cnext = nil; b->cprev = nil; b->hnext = nil; } b->type = t; b->bp.addr = bp; b->bp.hash = -1; b->bp.gen = fs->nextgen; b->data = b->buf + Hdrsz; b->fnext = nil; b->flag = Bdirty; b->nval = 0; b->valsz = 0; b->nbuf = 0; b->bufsz = 0; b->logsz = 0; b->lognxt = 0; dprint("new block %B from %p, flag=%x\n", b->bp, getcallerpc(&t), b->flag); return cacheblk(b); } char* fillsuper(Blk *b) { char *p; assert(b->type == Tsuper); p = b->data; lock(b); b->flag |= Bdirty; unlock(b); memcpy(p, "gefs0001", 8); p += 8; PBIT32(p, 0); p += 4; /* dirty */ PBIT32(p, Blksz); p += 4; PBIT32(p, Bufspc); p += 4; PBIT32(p, Hdrsz); p += 4; PBIT32(p, fs->snap.ht); p += 4; PBIT64(p, fs->snap.bp.addr); p += 8; PBIT64(p, fs->snap.bp.hash); p += 8; PBIT64(p, fs->nextgen); p += 8; PBIT32(p, fs->narena); p += 4; PBIT64(p, fs->arenasz); p += 8; PBIT64(p, fs->nextqid); p += 8; return p; } void finalize(Blk *b) { vlong h; // assert((b->flag & Bfinal) == 0); lock(b); b->flag |= Bfinal; if(b->type != Traw) PBIT16(b->buf, b->type); switch(b->type){ default: case Tnone: abort(); break; case Tpivot: PBIT16(b->buf+2, b->nval); PBIT16(b->buf+4, b->valsz); PBIT16(b->buf+6, b->nbuf); PBIT16(b->buf+8, b->bufsz); b->bp.hash = blkhash(b); break; case Tleaf: PBIT16(b->buf+2, b->nval); PBIT16(b->buf+4, b->valsz); b->bp.hash = blkhash(b); break; case Tlog: h = siphash(b->data + 8, Blkspc-8); PBIT64(b->data, h); case Traw: b->bp.hash = blkhash(b); break; case Tsuper: case Tarena: break; } unlock(b); } Blk* getblk(Bptr bp, int flg) { Blk *b; if((b = lookupblk(bp.addr)) != nil) return cacheblk(b); qlock(&blklock); if((b = lookupblk(bp.addr)) != nil){ cacheblk(b); qunlock(&blklock); return b; } if((b = readblk(bp.addr, flg)) == nil){ qunlock(&blklock); return nil; } if(blkhash(b) != bp.hash){ fprint(2, "corrupt block %B: %llx != %llx\n", bp, blkhash(b), bp.hash); qunlock(&blklock); abort(); return nil; } b->bp.hash = bp.hash; b->bp.gen = bp.gen; cacheblk(b); qunlock(&blklock); return b; } Blk* dupblk(vlong bp, uvlong bh) { USED(bp, bh); return nil; } Blk* refblk(Blk *b) { ainc(&b->ref); return b; } ushort blkfill(Blk *b) { switch(b->type){ case Tpivot: return 2*b->nbuf + b->bufsz + 2*b->nval + b->valsz; case Tleaf: return 2*b->nval + b->valsz; default: fprint(2, "invalid block @%lld\n", b->bp.addr); abort(); } return 0; // shut up kencc } void putblk(Blk *b) { if(b == nil || adec(&b->ref) != 0) return; assert((b->flag & Bqueued) || !(b->flag & Bdirty)); free(b); } void freeblk(Blk *b) { lock(b); assert((b->flag & Bqueued) == 0); b->flag |= Bzombie; b->freed = getcallerpc(&b); unlock(b); dprint("freeing block %B @ %ld, from 0x%p\n", b->bp, b->ref, getcallerpc(&b)); if(b->bp.gen == fs->nextgen) freebp(b->bp); // else // deadlist(b->bp); } void freebp(Bptr bp) { Bfree *f; if((f = malloc(sizeof(Bfree))) == nil) return; f->bp = bp; lock(&fs->freelk); f->next = fs->freehd; fs->freehd = f; unlock(&fs->freelk); } void reclaimblk(Bptr bp) { Arena *a; a = getarena(bp.addr); lock(a); blkdealloc_lk(bp.addr); unlock(a); }