syncthing/Godeps/_workspace/src/github.com/syndtr/goleveldb/leveldb/db_write.go

280 lines
5.9 KiB
Go
Raw Normal View History

2014-07-06 14:46:48 +02:00
// Copyright (c) 2012, Suryandaru Triandana <syndtr@gmail.com>
// All rights reserved.
//
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
package leveldb
import (
"time"
"github.com/syndtr/goleveldb/leveldb/memdb"
"github.com/syndtr/goleveldb/leveldb/opt"
"github.com/syndtr/goleveldb/leveldb/util"
)
func (d *DB) writeJournal(b *Batch) error {
w, err := d.journal.Next()
if err != nil {
return err
}
if _, err := w.Write(b.encode()); err != nil {
return err
}
if err := d.journal.Flush(); err != nil {
return err
}
if b.sync {
return d.journalWriter.Sync()
}
return nil
}
func (d *DB) jWriter() {
defer d.closeW.Done()
for {
select {
case b := <-d.journalC:
if b != nil {
d.journalAckC <- d.writeJournal(b)
}
case _, _ = <-d.closeC:
return
}
}
}
func (d *DB) rotateMem(n int) (mem *memdb.DB, err error) {
// Wait for pending memdb compaction.
err = d.compSendIdle(d.mcompCmdC)
if err != nil {
return
}
// Create new memdb and journal.
mem, err = d.newMem(n)
if err != nil {
return
}
// Schedule memdb compaction.
d.compTrigger(d.mcompTriggerC)
return
}
func (d *DB) flush(n int) (mem *memdb.DB, nn int, err error) {
s := d.s
delayed := false
flush := func() bool {
v := s.version()
defer v.release()
mem = d.getEffectiveMem()
nn = mem.Free()
switch {
case v.tLen(0) >= kL0_SlowdownWritesTrigger && !delayed:
delayed = true
time.Sleep(time.Millisecond)
case nn >= n:
return false
case v.tLen(0) >= kL0_StopWritesTrigger:
delayed = true
err = d.compSendIdle(d.tcompCmdC)
if err != nil {
return false
}
default:
// Allow memdb to grow if it has no entry.
if mem.Len() == 0 {
nn = n
return false
}
mem, err = d.rotateMem(n)
nn = mem.Free()
return false
}
return true
}
start := time.Now()
for flush() {
}
if delayed {
s.logf("db@write delayed T·%v", time.Since(start))
}
return
}
// Write apply the given batch to the DB. The batch will be applied
// sequentially.
//
// It is safe to modify the contents of the arguments after Write returns.
func (d *DB) Write(b *Batch, wo *opt.WriteOptions) (err error) {
err = d.ok()
if err != nil || b == nil || b.len() == 0 {
return
}
b.init(wo.GetSync())
// The write happen synchronously.
retry:
select {
case d.writeC <- b:
if <-d.writeMergedC {
return <-d.writeAckC
}
goto retry
case d.writeLockC <- struct{}{}:
case _, _ = <-d.closeC:
return ErrClosed
}
merged := 0
defer func() {
<-d.writeLockC
for i := 0; i < merged; i++ {
d.writeAckC <- err
}
}()
mem, memFree, err := d.flush(b.size())
if err != nil {
return
}
// Calculate maximum size of the batch.
m := 1 << 20
if x := b.size(); x <= 128<<10 {
m = x + (128 << 10)
}
m = minInt(m, memFree)
// Merge with other batch.
drain:
for b.size() < m && !b.sync {
select {
case nb := <-d.writeC:
if b.size()+nb.size() <= m {
b.append(nb)
d.writeMergedC <- true
merged++
} else {
d.writeMergedC <- false
break drain
}
default:
break drain
}
}
// Set batch first seq number relative from last seq.
b.seq = d.seq + 1
// Write journal concurrently if it is large enough.
if b.size() >= (128 << 10) {
// Push the write batch to the journal writer
select {
case _, _ = <-d.closeC:
err = ErrClosed
return
case d.journalC <- b:
// Write into memdb
b.memReplay(mem)
}
// Wait for journal writer
select {
case _, _ = <-d.closeC:
err = ErrClosed
return
case err = <-d.journalAckC:
if err != nil {
// Revert memdb if error detected
b.revertMemReplay(mem)
return
}
}
} else {
err = d.writeJournal(b)
if err != nil {
return
}
b.memReplay(mem)
}
// Set last seq number.
d.addSeq(uint64(b.len()))
if b.size() >= memFree {
d.rotateMem(0)
}
return
}
// Put sets the value for the given key. It overwrites any previous value
// for that key; a DB is not a multi-map.
//
// It is safe to modify the contents of the arguments after Put returns.
func (d *DB) Put(key, value []byte, wo *opt.WriteOptions) error {
b := new(Batch)
b.Put(key, value)
return d.Write(b, wo)
}
// Delete deletes the value for the given key. It returns ErrNotFound if
// the DB does not contain the key.
//
// It is safe to modify the contents of the arguments after Delete returns.
func (d *DB) Delete(key []byte, wo *opt.WriteOptions) error {
b := new(Batch)
b.Delete(key)
return d.Write(b, wo)
}
2014-07-06 23:13:10 +02:00
func isMemOverlaps(icmp *iComparer, mem *memdb.DB, min, max []byte) bool {
2014-07-06 14:46:48 +02:00
iter := mem.NewIterator(nil)
defer iter.Release()
2014-07-06 23:13:10 +02:00
return (max == nil || (iter.First() && icmp.uCompare(max, iKey(iter.Key()).ukey()) >= 0)) &&
(min == nil || (iter.Last() && icmp.uCompare(min, iKey(iter.Key()).ukey()) <= 0))
2014-07-06 14:46:48 +02:00
}
// CompactRange compacts the underlying DB for the given key range.
// In particular, deleted and overwritten versions are discarded,
// and the data is rearranged to reduce the cost of operations
// needed to access the data. This operation should typically only
// be invoked by users who understand the underlying implementation.
//
// A nil Range.Start is treated as a key before all keys in the DB.
// And a nil Range.Limit is treated as a key after all keys in the DB.
// Therefore if both is nil then it will compact entire DB.
func (d *DB) CompactRange(r util.Range) error {
if err := d.ok(); err != nil {
return err
}
select {
case d.writeLockC <- struct{}{}:
case _, _ = <-d.closeC:
return ErrClosed
}
// Check for overlaps in memdb.
mem := d.getEffectiveMem()
2014-07-06 23:13:10 +02:00
if isMemOverlaps(d.s.icmp, mem, r.Start, r.Limit) {
2014-07-06 14:46:48 +02:00
// Memdb compaction.
if _, err := d.rotateMem(0); err != nil {
<-d.writeLockC
return err
}
<-d.writeLockC
if err := d.compSendIdle(d.mcompCmdC); err != nil {
return err
}
} else {
<-d.writeLockC
}
// Table compaction.
return d.compSendRange(d.tcompCmdC, -1, r.Start, r.Limit)
}