etcd/mvcc/kvstore.go

711 lines
16 KiB
Go
Raw Normal View History

2016-05-13 06:50:33 +03:00
// Copyright 2015 The etcd Authors
2015-09-15 23:54:11 +03:00
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
2016-04-25 22:32:58 +03:00
package mvcc
2015-05-22 18:11:43 +03:00
import (
2016-03-30 21:37:55 +03:00
"encoding/binary"
2015-05-27 20:35:51 +03:00
"errors"
"math"
2015-05-27 20:35:51 +03:00
"math/rand"
2015-05-22 18:11:43 +03:00
"sync"
"time"
2016-01-05 21:16:50 +03:00
"github.com/coreos/etcd/lease"
2016-04-25 22:32:58 +03:00
"github.com/coreos/etcd/mvcc/backend"
"github.com/coreos/etcd/mvcc/mvccpb"
"github.com/coreos/etcd/pkg/schedule"
2016-05-21 08:30:50 +03:00
"github.com/coreos/pkg/capnslog"
2016-03-23 03:10:28 +03:00
"golang.org/x/net/context"
2015-05-22 18:11:43 +03:00
)
var (
keyBucketName = []byte("key")
metaBucketName = []byte("meta")
2015-05-27 20:35:51 +03:00
// markedRevBytesLen is the byte length of marked revision.
// The first `revBytesLen` bytes represents a normal revision. The last
// one byte is the mark.
markedRevBytesLen = revBytesLen + 1
markBytePosition = markedRevBytesLen - 1
markTombstone byte = 't'
2016-03-30 21:37:55 +03:00
consistentIndexKeyName = []byte("consistent_index")
2015-05-31 18:59:31 +03:00
scheduledCompactKeyName = []byte("scheduledCompactRev")
finishedCompactKeyName = []byte("finishedCompactRev")
2016-04-25 22:32:58 +03:00
ErrTxnIDMismatch = errors.New("mvcc: txn id mismatch")
ErrCompacted = errors.New("mvcc: required revision has been compacted")
ErrFutureRev = errors.New("mvcc: required revision is a future revision")
ErrCanceled = errors.New("mvcc: watcher is canceled")
2016-05-21 08:30:50 +03:00
plog = capnslog.NewPackageLogger("github.com/coreos/etcd", "mvcc")
2015-05-22 18:11:43 +03:00
)
2016-03-30 21:37:55 +03:00
// ConsistentIndexGetter is an interface that wraps the Get method.
// Consistent index is the offset of an entry in a consistent replicated log.
type ConsistentIndexGetter interface {
// ConsistentIndex returns the consistent index of current executing entry.
ConsistentIndex() uint64
}
2015-05-22 18:11:43 +03:00
type store struct {
mu sync.Mutex // guards the following
2015-05-22 18:11:43 +03:00
2016-03-30 21:37:55 +03:00
ig ConsistentIndexGetter
2015-05-22 18:11:43 +03:00
b backend.Backend
kvindex index
le lease.Lessor
2015-08-20 18:39:07 +03:00
currentRev revision
// the main revision of the last compaction
2015-05-31 18:59:31 +03:00
compactMainRev int64
2015-05-22 23:35:43 +03:00
tx backend.BatchTx
txnID int64 // tracks the current txnID to verify txn operations
txnModify bool
// bytesBuf8 is a byte slice of length 8
// to avoid a repetitive allocation in saveIndex.
bytesBuf8 []byte
2016-04-25 22:32:58 +03:00
changes []mvccpb.KeyValue
fifoSched schedule.Scheduler
stopc chan struct{}
2015-05-22 18:11:43 +03:00
}
// NewStore returns a new store. It is useful to create a store inside
2016-04-25 22:32:58 +03:00
// mvcc pkg. It should only be used for testing externally.
2016-03-30 21:37:55 +03:00
func NewStore(b backend.Backend, le lease.Lessor, ig ConsistentIndexGetter) *store {
2015-05-22 18:11:43 +03:00
s := &store{
b: b,
2016-03-30 21:37:55 +03:00
ig: ig,
kvindex: newTreeIndex(),
le: le,
currentRev: revision{main: 1},
2015-05-31 18:59:31 +03:00
compactMainRev: -1,
bytesBuf8: make([]byte, 8),
fifoSched: schedule.NewFIFOScheduler(),
stopc: make(chan struct{}),
2015-05-22 18:11:43 +03:00
}
if s.le != nil {
s.le.SetRangeDeleter(s)
}
2015-05-22 18:11:43 +03:00
tx := s.b.BatchTx()
tx.Lock()
tx.UnsafeCreateBucket(keyBucketName)
tx.UnsafeCreateBucket(metaBucketName)
2015-05-22 18:11:43 +03:00
tx.Unlock()
s.b.ForceCommit()
if err := s.restore(); err != nil {
// TODO: return the error instead of panic here?
panic("failed to recover store from backend")
}
2015-05-22 18:11:43 +03:00
return s
}
2015-09-15 22:11:00 +03:00
func (s *store) Rev() int64 {
s.mu.Lock()
defer s.mu.Unlock()
2015-09-15 22:11:00 +03:00
return s.currentRev.main
}
2016-01-26 05:39:08 +03:00
func (s *store) FirstRev() int64 {
s.mu.Lock()
defer s.mu.Unlock()
return s.compactMainRev
}
2016-01-05 21:16:50 +03:00
func (s *store) Put(key, value []byte, lease lease.LeaseID) int64 {
2015-07-24 18:16:27 +03:00
id := s.TxnBegin()
2016-01-05 02:12:59 +03:00
s.put(key, value, lease)
2015-08-28 20:35:30 +03:00
s.txnEnd(id)
putCounter.Inc()
2015-05-22 18:11:43 +03:00
2015-05-31 08:56:33 +03:00
return int64(s.currentRev.main)
2015-05-22 18:11:43 +03:00
}
2016-06-22 02:20:55 +03:00
func (s *store) Range(key, end []byte, ro RangeOptions) (r *RangeResult, err error) {
2015-07-24 18:16:27 +03:00
id := s.TxnBegin()
2016-06-22 02:20:55 +03:00
kvs, count, rev, err := s.rangeKeys(key, end, ro.Limit, ro.Rev, ro.Count)
2015-08-28 20:35:30 +03:00
s.txnEnd(id)
rangeCounter.Inc()
2015-05-22 23:35:43 +03:00
2016-06-22 02:20:55 +03:00
r = &RangeResult{
KVs: kvs,
Count: count,
Rev: rev,
}
return r, err
2015-05-22 23:35:43 +03:00
}
2015-05-31 08:56:33 +03:00
func (s *store) DeleteRange(key, end []byte) (n, rev int64) {
2015-07-24 18:16:27 +03:00
id := s.TxnBegin()
n = s.deleteRange(key, end)
2015-08-28 20:35:30 +03:00
s.txnEnd(id)
deleteCounter.Inc()
2015-05-22 23:35:43 +03:00
2015-05-31 08:56:33 +03:00
return n, int64(s.currentRev.main)
2015-05-22 23:35:43 +03:00
}
2015-07-24 18:16:27 +03:00
func (s *store) TxnBegin() int64 {
2015-05-22 23:35:43 +03:00
s.mu.Lock()
2015-05-31 08:56:33 +03:00
s.currentRev.sub = 0
s.tx = s.b.BatchTx()
s.tx.Lock()
2015-05-27 20:35:51 +03:00
2015-07-24 18:16:27 +03:00
s.txnID = rand.Int63()
return s.txnID
2015-05-22 23:35:43 +03:00
}
2015-05-22 18:11:43 +03:00
2015-07-24 18:16:27 +03:00
func (s *store) TxnEnd(txnID int64) error {
2015-08-28 20:35:30 +03:00
err := s.txnEnd(txnID)
if err != nil {
return err
}
txnCounter.Inc()
return nil
}
// txnEnd is used for unlocking an internal txn. It does
// not increase the txnCounter.
func (s *store) txnEnd(txnID int64) error {
2015-07-24 18:16:27 +03:00
if txnID != s.txnID {
return ErrTxnIDMismatch
2015-05-27 20:35:51 +03:00
}
// only update index if the txn modifies the mvcc state.
// read only txn might execute with one write txn concurrently,
// it should not write its index to mvcc.
if s.txnModify {
s.saveIndex()
}
s.txnModify = false
s.tx.Unlock()
2015-05-31 08:56:33 +03:00
if s.currentRev.sub != 0 {
s.currentRev.main += 1
2015-05-22 23:35:43 +03:00
}
2015-05-31 08:56:33 +03:00
s.currentRev.sub = 0
2015-10-06 02:15:44 +03:00
dbTotalSize.Set(float64(s.b.Size()))
2015-05-22 23:35:43 +03:00
s.mu.Unlock()
2015-05-27 20:35:51 +03:00
return nil
2015-05-22 23:35:43 +03:00
}
2016-06-22 02:20:55 +03:00
func (s *store) TxnRange(txnID int64, key, end []byte, ro RangeOptions) (r *RangeResult, err error) {
2015-07-24 18:16:27 +03:00
if txnID != s.txnID {
2016-06-22 02:20:55 +03:00
return nil, ErrTxnIDMismatch
}
kvs, count, rev, err := s.rangeKeys(key, end, ro.Limit, ro.Rev, ro.Count)
r = &RangeResult{
KVs: kvs,
Count: count,
Rev: rev,
2015-05-27 20:35:51 +03:00
}
2016-06-22 02:20:55 +03:00
return r, err
2015-05-27 19:58:21 +03:00
}
2016-01-05 21:16:50 +03:00
func (s *store) TxnPut(txnID int64, key, value []byte, lease lease.LeaseID) (rev int64, err error) {
2015-07-24 18:16:27 +03:00
if txnID != s.txnID {
return 0, ErrTxnIDMismatch
2015-05-27 20:35:51 +03:00
}
2016-01-05 02:12:59 +03:00
s.put(key, value, lease)
2015-05-31 08:56:33 +03:00
return int64(s.currentRev.main + 1), nil
2015-05-27 19:58:21 +03:00
}
2015-07-24 18:16:27 +03:00
func (s *store) TxnDeleteRange(txnID int64, key, end []byte) (n, rev int64, err error) {
if txnID != s.txnID {
return 0, 0, ErrTxnIDMismatch
2015-05-27 20:35:51 +03:00
}
n = s.deleteRange(key, end)
2015-05-31 08:56:33 +03:00
if n != 0 || s.currentRev.sub != 0 {
rev = int64(s.currentRev.main + 1)
} else {
rev = int64(s.currentRev.main)
2015-05-27 19:58:21 +03:00
}
2015-05-31 08:56:33 +03:00
return n, rev, nil
2015-05-27 19:58:21 +03:00
}
func (s *store) compactBarrier(ctx context.Context, ch chan struct{}) {
if ctx == nil || ctx.Err() != nil {
s.mu.Lock()
select {
case <-s.stopc:
default:
f := func(ctx context.Context) { s.compactBarrier(ctx, ch) }
s.fifoSched.Schedule(f)
}
s.mu.Unlock()
return
}
close(ch)
}
func (s *store) Compact(rev int64) (<-chan struct{}, error) {
2015-05-31 18:59:31 +03:00
s.mu.Lock()
defer s.mu.Unlock()
if rev <= s.compactMainRev {
ch := make(chan struct{})
f := func(ctx context.Context) { s.compactBarrier(ctx, ch) }
s.fifoSched.Schedule(f)
return ch, ErrCompacted
2015-05-31 18:59:31 +03:00
}
if rev > s.currentRev.main {
return nil, ErrFutureRev
}
2015-05-31 18:59:31 +03:00
2015-08-28 20:35:30 +03:00
start := time.Now()
2015-05-31 18:59:31 +03:00
s.compactMainRev = rev
rbytes := newRevBytes()
2015-08-20 18:39:07 +03:00
revToBytes(revision{main: rev}, rbytes)
2015-05-31 18:59:31 +03:00
tx := s.b.BatchTx()
tx.Lock()
tx.UnsafePut(metaBucketName, scheduledCompactKeyName, rbytes)
2015-05-31 18:59:31 +03:00
tx.Unlock()
// ensure that desired compaction is persisted
s.b.ForceCommit()
2015-05-31 18:59:31 +03:00
keep := s.kvindex.Compact(rev)
ch := make(chan struct{})
var j = func(ctx context.Context) {
if ctx.Err() != nil {
s.compactBarrier(ctx, ch)
return
}
if !s.scheduleCompaction(rev, keep) {
s.compactBarrier(nil, ch)
return
}
close(ch)
}
s.fifoSched.Schedule(j)
2015-08-28 20:35:30 +03:00
2016-04-08 08:09:25 +03:00
indexCompactionPauseDurations.Observe(float64(time.Since(start) / time.Millisecond))
return ch, nil
2015-05-31 18:59:31 +03:00
}
2016-10-25 20:07:08 +03:00
// DefaultIgnores is a map of keys to ignore in hash checking.
var DefaultIgnores map[backend.IgnoreKey]struct{}
func init() {
DefaultIgnores = map[backend.IgnoreKey]struct{}{
// consistent index might be changed due to v2 internal sync, which
// is not controllable by the user.
{Bucket: string(metaBucketName), Key: string(consistentIndexKeyName)}: {},
}
}
2016-05-03 00:37:56 +03:00
func (s *store) Hash() (uint32, int64, error) {
s.mu.Lock()
defer s.mu.Unlock()
s.b.ForceCommit()
2016-05-03 00:37:56 +03:00
2016-10-25 20:07:08 +03:00
h, err := s.b.Hash(DefaultIgnores)
2016-05-03 00:37:56 +03:00
rev := s.currentRev.main
return h, rev, err
2015-09-14 07:29:10 +03:00
}
func (s *store) Commit() {
s.mu.Lock()
defer s.mu.Unlock()
s.tx = s.b.BatchTx()
s.tx.Lock()
s.saveIndex()
s.tx.Unlock()
s.b.ForceCommit()
}
func (s *store) Restore(b backend.Backend) error {
s.mu.Lock()
defer s.mu.Unlock()
close(s.stopc)
s.fifoSched.Stop()
s.b = b
s.kvindex = newTreeIndex()
s.currentRev = revision{main: 1}
s.compactMainRev = -1
s.tx = b.BatchTx()
s.txnID = -1
s.fifoSched = schedule.NewFIFOScheduler()
s.stopc = make(chan struct{})
return s.restore()
}
func (s *store) restore() error {
min, max := newRevBytes(), newRevBytes()
revToBytes(revision{main: 1}, min)
2015-08-20 18:39:07 +03:00
revToBytes(revision{main: math.MaxInt64, sub: math.MaxInt64}, max)
keyToLease := make(map[string]lease.LeaseID)
// use an unordered map to hold the temp index data to speed up
// the initial key index recovery.
// we will convert this unordered map into the tree index later.
unordered := make(map[string]*keyIndex, 100000)
// restore index
tx := s.b.BatchTx()
tx.Lock()
_, finishedCompactBytes := tx.UnsafeRange(metaBucketName, finishedCompactKeyName, nil, 0)
if len(finishedCompactBytes) != 0 {
s.compactMainRev = bytesToRev(finishedCompactBytes[0]).main
2016-05-21 08:30:50 +03:00
plog.Printf("restore compact to %d", s.compactMainRev)
}
// TODO: limit N to reduce max memory usage
keys, vals := tx.UnsafeRange(keyBucketName, min, max, 0)
for i, key := range keys {
2016-04-25 22:32:58 +03:00
var kv mvccpb.KeyValue
if err := kv.Unmarshal(vals[i]); err != nil {
2016-05-21 08:30:50 +03:00
plog.Fatalf("cannot unmarshal event: %v", err)
}
rev := bytesToRev(key[:revBytesLen])
// restore index
switch {
case isTombstone(key):
if ki, ok := unordered[string(kv.Key)]; ok {
ki.tombstone(rev.main, rev.sub)
}
delete(keyToLease, string(kv.Key))
default:
ki, ok := unordered[string(kv.Key)]
if ok {
ki.put(rev.main, rev.sub)
} else {
ki = &keyIndex{key: kv.Key}
ki.restore(revision{kv.CreateRevision, 0}, rev, kv.Version)
unordered[string(kv.Key)] = ki
}
if lid := lease.LeaseID(kv.Lease); lid != lease.NoLease {
keyToLease[string(kv.Key)] = lid
} else {
delete(keyToLease, string(kv.Key))
}
}
2015-08-20 18:39:07 +03:00
// update revision
s.currentRev = rev
}
// restore the tree index from the unordered index.
for _, v := range unordered {
s.kvindex.Insert(v)
}
// keys in the range [compacted revision -N, compaction] might all be deleted due to compaction.
// the correct revision should be set to compaction revision in the case, not the largest revision
// we have seen.
if s.currentRev.main < s.compactMainRev {
s.currentRev.main = s.compactMainRev
}
for key, lid := range keyToLease {
if s.le == nil {
panic("no lessor to attach lease")
}
err := s.le.Attach(lid, []lease.LeaseItem{{Key: key}})
if err != nil {
plog.Errorf("unexpected Attach error: %v", err)
}
}
_, scheduledCompactBytes := tx.UnsafeRange(metaBucketName, scheduledCompactKeyName, nil, 0)
scheduledCompact := int64(0)
if len(scheduledCompactBytes) != 0 {
scheduledCompact = bytesToRev(scheduledCompactBytes[0]).main
if scheduledCompact <= s.compactMainRev {
scheduledCompact = 0
}
}
tx.Unlock()
if scheduledCompact != 0 {
s.Compact(scheduledCompact)
2016-05-21 08:30:50 +03:00
plog.Printf("resume scheduled compaction at %d", scheduledCompact)
}
return nil
}
func (s *store) Close() error {
close(s.stopc)
s.fifoSched.Stop()
return nil
}
func (a *store) Equal(b *store) bool {
if a.currentRev != b.currentRev {
return false
}
if a.compactMainRev != b.compactMainRev {
return false
}
return a.kvindex.Equal(b.kvindex)
}
2015-05-27 19:58:21 +03:00
// range is a keyword in Go, add Keys suffix.
2016-06-22 02:20:55 +03:00
func (s *store) rangeKeys(key, end []byte, limit, rangeRev int64, countOnly bool) (kvs []mvccpb.KeyValue, count int, curRev int64, err error) {
curRev = int64(s.currentRev.main)
if s.currentRev.sub > 0 {
curRev += 1
}
if rangeRev > curRev {
2016-06-22 02:20:55 +03:00
return nil, -1, s.currentRev.main, ErrFutureRev
}
var rev int64
2015-05-31 08:56:33 +03:00
if rangeRev <= 0 {
rev = curRev
2015-05-22 18:11:43 +03:00
} else {
2015-05-31 08:56:33 +03:00
rev = rangeRev
2015-05-22 18:11:43 +03:00
}
if rev < s.compactMainRev {
2016-06-22 02:20:55 +03:00
return nil, -1, 0, ErrCompacted
2015-05-31 18:59:31 +03:00
}
2015-05-22 18:11:43 +03:00
2015-05-31 18:59:31 +03:00
_, revpairs := s.kvindex.Range(key, end, int64(rev))
if len(revpairs) == 0 {
2016-06-22 02:20:55 +03:00
return nil, 0, curRev, nil
}
if countOnly {
return nil, len(revpairs), curRev, nil
2015-05-22 18:11:43 +03:00
}
2015-05-31 18:59:31 +03:00
for _, revpair := range revpairs {
start, end := revBytesRange(revpair)
2015-05-22 18:11:43 +03:00
_, vs := s.tx.UnsafeRange(keyBucketName, start, end, 0)
2015-05-31 08:56:33 +03:00
if len(vs) != 1 {
2016-05-21 08:30:50 +03:00
plog.Fatalf("range cannot find rev (%d,%d)", revpair.main, revpair.sub)
2015-05-22 18:11:43 +03:00
}
2016-04-25 22:32:58 +03:00
var kv mvccpb.KeyValue
if err := kv.Unmarshal(vs[0]); err != nil {
2016-05-21 08:30:50 +03:00
plog.Fatalf("cannot unmarshal event: %v", err)
2015-05-22 18:11:43 +03:00
}
kvs = append(kvs, kv)
if limit > 0 && len(kvs) >= int(limit) {
break
}
2015-05-22 18:11:43 +03:00
}
2016-08-10 20:50:53 +03:00
return kvs, len(revpairs), curRev, nil
2015-05-22 18:11:43 +03:00
}
func (s *store) put(key, value []byte, leaseID lease.LeaseID) {
s.txnModify = true
rev := s.currentRev.main + 1
c := rev
oldLease := lease.NoLease
// if the key exists before, use its previous created and
// get its previous leaseID
_, created, ver, err := s.kvindex.Get(key, rev)
if err == nil {
c = created.main
oldLease = s.le.GetLease(lease.LeaseItem{Key: string(key)})
}
ibytes := newRevBytes()
2015-08-20 18:39:07 +03:00
revToBytes(revision{main: rev, sub: s.currentRev.sub}, ibytes)
2015-05-22 18:11:43 +03:00
2015-06-29 22:47:17 +03:00
ver = ver + 1
2016-04-25 22:32:58 +03:00
kv := mvccpb.KeyValue{
Key: key,
Value: value,
CreateRevision: c,
ModRevision: rev,
Version: ver,
Lease: int64(leaseID),
}
d, err := kv.Marshal()
2015-05-22 18:11:43 +03:00
if err != nil {
2016-05-21 08:30:50 +03:00
plog.Fatalf("cannot marshal event: %v", err)
2015-05-22 18:11:43 +03:00
}
s.tx.UnsafeSeqPut(keyBucketName, ibytes, d)
2015-08-20 18:39:07 +03:00
s.kvindex.Put(key, revision{main: rev, sub: s.currentRev.sub})
s.changes = append(s.changes, kv)
2015-05-31 08:56:33 +03:00
s.currentRev.sub += 1
if oldLease != lease.NoLease {
if s.le == nil {
panic("no lessor to detach lease")
}
err = s.le.Detach(oldLease, []lease.LeaseItem{{Key: string(key)}})
if err != nil {
plog.Errorf("unexpected error from lease detach: %v", err)
}
}
if leaseID != lease.NoLease {
if s.le == nil {
panic("no lessor to attach lease")
}
err = s.le.Attach(leaseID, []lease.LeaseItem{{Key: string(key)}})
if err != nil {
panic("unexpected error from lease Attach")
}
}
2015-05-22 23:35:43 +03:00
}
func (s *store) deleteRange(key, end []byte) int64 {
s.txnModify = true
rrev := s.currentRev.main
2015-05-31 08:56:33 +03:00
if s.currentRev.sub > 0 {
rrev += 1
2015-05-22 23:35:43 +03:00
}
keys, revs := s.kvindex.Range(key, end, rrev)
2015-05-22 23:35:43 +03:00
2015-05-31 08:56:33 +03:00
if len(keys) == 0 {
2015-05-22 23:35:43 +03:00
return 0
}
for i, key := range keys {
s.delete(key, revs[i])
2015-05-22 23:35:43 +03:00
}
return int64(len(keys))
2015-05-22 18:11:43 +03:00
}
func (s *store) delete(key []byte, rev revision) {
mainrev := s.currentRev.main + 1
2015-05-22 18:11:43 +03:00
ibytes := newRevBytes()
2015-08-20 18:39:07 +03:00
revToBytes(revision{main: mainrev, sub: s.currentRev.sub}, ibytes)
ibytes = appendMarkTombstone(ibytes)
2015-05-22 18:11:43 +03:00
2016-04-25 22:32:58 +03:00
kv := mvccpb.KeyValue{
Key: key,
2015-05-22 18:11:43 +03:00
}
d, err := kv.Marshal()
2015-05-22 18:11:43 +03:00
if err != nil {
2016-05-21 08:30:50 +03:00
plog.Fatalf("cannot marshal event: %v", err)
2015-05-22 18:11:43 +03:00
}
s.tx.UnsafeSeqPut(keyBucketName, ibytes, d)
2015-08-20 18:39:07 +03:00
err = s.kvindex.Tombstone(key, revision{main: mainrev, sub: s.currentRev.sub})
2015-05-22 18:11:43 +03:00
if err != nil {
2016-05-21 08:30:50 +03:00
plog.Fatalf("cannot tombstone an existing key (%s): %v", string(key), err)
2015-05-22 18:11:43 +03:00
}
s.changes = append(s.changes, kv)
2015-05-31 08:56:33 +03:00
s.currentRev.sub += 1
item := lease.LeaseItem{Key: string(key)}
leaseID := s.le.GetLease(item)
if leaseID != lease.NoLease {
err = s.le.Detach(leaseID, []lease.LeaseItem{item})
if err != nil {
plog.Errorf("cannot detach %v", err)
}
}
2015-05-22 18:11:43 +03:00
}
2016-04-25 22:32:58 +03:00
func (s *store) getChanges() []mvccpb.KeyValue {
changes := s.changes
s.changes = make([]mvccpb.KeyValue, 0, 4)
return changes
}
2016-03-30 21:37:55 +03:00
func (s *store) saveIndex() {
if s.ig == nil {
return
}
tx := s.tx
bs := s.bytesBuf8
2016-03-30 21:37:55 +03:00
binary.BigEndian.PutUint64(bs, s.ig.ConsistentIndex())
// put the index into the underlying backend
// tx has been locked in TxnBegin, so there is no need to lock it again
tx.UnsafePut(metaBucketName, consistentIndexKeyName, bs)
}
func (s *store) ConsistentIndex() uint64 {
// TODO: cache index in a uint64 field?
tx := s.b.BatchTx()
tx.Lock()
defer tx.Unlock()
_, vs := tx.UnsafeRange(metaBucketName, consistentIndexKeyName, nil, 0)
if len(vs) == 0 {
return 0
}
return binary.BigEndian.Uint64(vs[0])
}
// appendMarkTombstone appends tombstone mark to normal revision bytes.
func appendMarkTombstone(b []byte) []byte {
if len(b) != revBytesLen {
2016-05-21 08:30:50 +03:00
plog.Panicf("cannot append mark to non normal revision bytes")
}
return append(b, markTombstone)
}
// isTombstone checks whether the revision bytes is a tombstone.
func isTombstone(b []byte) bool {
return len(b) == markedRevBytesLen && b[markBytePosition] == markTombstone
}
// revBytesRange returns the range of revision bytes at
// the given revision.
func revBytesRange(rev revision) (start, end []byte) {
start = newRevBytes()
revToBytes(rev, start)
end = newRevBytes()
endRev := revision{main: rev.main, sub: rev.sub + 1}
revToBytes(endRev, end)
return start, end
}