|
|
@@ -80,7 +80,7 @@ func TestKVTxnRange(t *testing.T) { testKVRange(t, txnRangeFunc) }
|
|
|
|
|
|
func testKVRange(t *testing.T, f rangeFunc) {
|
|
|
b, tmpPath := backend.NewDefaultTmpBackend()
|
|
|
- s := NewStore(b, &lease.FakeLessor{})
|
|
|
+ s := NewStore(b, &lease.FakeLessor{}, nil)
|
|
|
defer cleanup(s, b, tmpPath)
|
|
|
|
|
|
kvs := put3TestKVs(s)
|
|
|
@@ -146,7 +146,7 @@ func TestKVTxnRangeRev(t *testing.T) { testKVRangeRev(t, normalRangeFunc) }
|
|
|
|
|
|
func testKVRangeRev(t *testing.T, f rangeFunc) {
|
|
|
b, tmpPath := backend.NewDefaultTmpBackend()
|
|
|
- s := NewStore(b, &lease.FakeLessor{})
|
|
|
+ s := NewStore(b, &lease.FakeLessor{}, nil)
|
|
|
defer cleanup(s, b, tmpPath)
|
|
|
|
|
|
kvs := put3TestKVs(s)
|
|
|
@@ -182,7 +182,7 @@ func TestKVTxnRangeBadRev(t *testing.T) { testKVRangeBadRev(t, normalRangeFunc)
|
|
|
|
|
|
func testKVRangeBadRev(t *testing.T, f rangeFunc) {
|
|
|
b, tmpPath := backend.NewDefaultTmpBackend()
|
|
|
- s := NewStore(b, &lease.FakeLessor{})
|
|
|
+ s := NewStore(b, &lease.FakeLessor{}, nil)
|
|
|
defer cleanup(s, b, tmpPath)
|
|
|
|
|
|
put3TestKVs(s)
|
|
|
@@ -213,7 +213,7 @@ func TestKVTxnRangeLimit(t *testing.T) { testKVRangeLimit(t, txnRangeFunc) }
|
|
|
|
|
|
func testKVRangeLimit(t *testing.T, f rangeFunc) {
|
|
|
b, tmpPath := backend.NewDefaultTmpBackend()
|
|
|
- s := NewStore(b, &lease.FakeLessor{})
|
|
|
+ s := NewStore(b, &lease.FakeLessor{}, nil)
|
|
|
defer cleanup(s, b, tmpPath)
|
|
|
|
|
|
kvs := put3TestKVs(s)
|
|
|
@@ -251,7 +251,7 @@ func TestKVTxnPutMultipleTimes(t *testing.T) { testKVPutMultipleTimes(t, txnPutF
|
|
|
|
|
|
func testKVPutMultipleTimes(t *testing.T, f putFunc) {
|
|
|
b, tmpPath := backend.NewDefaultTmpBackend()
|
|
|
- s := NewStore(b, &lease.FakeLessor{})
|
|
|
+ s := NewStore(b, &lease.FakeLessor{}, nil)
|
|
|
defer cleanup(s, b, tmpPath)
|
|
|
|
|
|
for i := 0; i < 10; i++ {
|
|
|
@@ -313,7 +313,7 @@ func testKVDeleteRange(t *testing.T, f deleteRangeFunc) {
|
|
|
|
|
|
for i, tt := range tests {
|
|
|
b, tmpPath := backend.NewDefaultTmpBackend()
|
|
|
- s := NewStore(b, &lease.FakeLessor{})
|
|
|
+ s := NewStore(b, &lease.FakeLessor{}, nil)
|
|
|
|
|
|
s.Put([]byte("foo"), []byte("bar"), lease.NoLease)
|
|
|
s.Put([]byte("foo1"), []byte("bar1"), lease.NoLease)
|
|
|
@@ -333,7 +333,7 @@ func TestKVTxnDeleteMultipleTimes(t *testing.T) { testKVDeleteMultipleTimes(t, t
|
|
|
|
|
|
func testKVDeleteMultipleTimes(t *testing.T, f deleteRangeFunc) {
|
|
|
b, tmpPath := backend.NewDefaultTmpBackend()
|
|
|
- s := NewStore(b, &lease.FakeLessor{})
|
|
|
+ s := NewStore(b, &lease.FakeLessor{}, nil)
|
|
|
defer cleanup(s, b, tmpPath)
|
|
|
|
|
|
s.Put([]byte("foo"), []byte("bar"), lease.NoLease)
|
|
|
@@ -354,7 +354,7 @@ func testKVDeleteMultipleTimes(t *testing.T, f deleteRangeFunc) {
|
|
|
// test that range, put, delete on single key in sequence repeatedly works correctly.
|
|
|
func TestKVOperationInSequence(t *testing.T) {
|
|
|
b, tmpPath := backend.NewDefaultTmpBackend()
|
|
|
- s := NewStore(b, &lease.FakeLessor{})
|
|
|
+ s := NewStore(b, &lease.FakeLessor{}, nil)
|
|
|
defer cleanup(s, b, tmpPath)
|
|
|
|
|
|
for i := 0; i < 10; i++ {
|
|
|
@@ -401,7 +401,7 @@ func TestKVOperationInSequence(t *testing.T) {
|
|
|
|
|
|
func TestKVTxnBlockNonTxnOperations(t *testing.T) {
|
|
|
b, tmpPath := backend.NewDefaultTmpBackend()
|
|
|
- s := NewStore(b, &lease.FakeLessor{})
|
|
|
+ s := NewStore(b, &lease.FakeLessor{}, nil)
|
|
|
|
|
|
tests := []func(){
|
|
|
func() { s.Range([]byte("foo"), nil, 0, 0) },
|
|
|
@@ -435,7 +435,7 @@ func TestKVTxnBlockNonTxnOperations(t *testing.T) {
|
|
|
|
|
|
func TestKVTxnWrongID(t *testing.T) {
|
|
|
b, tmpPath := backend.NewDefaultTmpBackend()
|
|
|
- s := NewStore(b, &lease.FakeLessor{})
|
|
|
+ s := NewStore(b, &lease.FakeLessor{}, nil)
|
|
|
defer cleanup(s, b, tmpPath)
|
|
|
|
|
|
id := s.TxnBegin()
|
|
|
@@ -472,7 +472,7 @@ func TestKVTxnWrongID(t *testing.T) {
|
|
|
// test that txn range, put, delete on single key in sequence repeatedly works correctly.
|
|
|
func TestKVTxnOperationInSequence(t *testing.T) {
|
|
|
b, tmpPath := backend.NewDefaultTmpBackend()
|
|
|
- s := NewStore(b, &lease.FakeLessor{})
|
|
|
+ s := NewStore(b, &lease.FakeLessor{}, nil)
|
|
|
defer cleanup(s, b, tmpPath)
|
|
|
|
|
|
for i := 0; i < 10; i++ {
|
|
|
@@ -528,7 +528,7 @@ func TestKVTxnOperationInSequence(t *testing.T) {
|
|
|
|
|
|
func TestKVCompactReserveLastValue(t *testing.T) {
|
|
|
b, tmpPath := backend.NewDefaultTmpBackend()
|
|
|
- s := NewStore(b, &lease.FakeLessor{})
|
|
|
+ s := NewStore(b, &lease.FakeLessor{}, nil)
|
|
|
defer cleanup(s, b, tmpPath)
|
|
|
|
|
|
s.Put([]byte("foo"), []byte("bar0"), 1)
|
|
|
@@ -582,7 +582,7 @@ func TestKVCompactReserveLastValue(t *testing.T) {
|
|
|
|
|
|
func TestKVCompactBad(t *testing.T) {
|
|
|
b, tmpPath := backend.NewDefaultTmpBackend()
|
|
|
- s := NewStore(b, &lease.FakeLessor{})
|
|
|
+ s := NewStore(b, &lease.FakeLessor{}, nil)
|
|
|
defer cleanup(s, b, tmpPath)
|
|
|
|
|
|
s.Put([]byte("foo"), []byte("bar0"), lease.NoLease)
|
|
|
@@ -615,7 +615,7 @@ func TestKVHash(t *testing.T) {
|
|
|
for i := 0; i < len(hashes); i++ {
|
|
|
var err error
|
|
|
b, tmpPath := backend.NewDefaultTmpBackend()
|
|
|
- kv := NewStore(b, &lease.FakeLessor{})
|
|
|
+ kv := NewStore(b, &lease.FakeLessor{}, nil)
|
|
|
kv.Put([]byte("foo0"), []byte("bar0"), lease.NoLease)
|
|
|
kv.Put([]byte("foo1"), []byte("bar0"), lease.NoLease)
|
|
|
hashes[i], err = kv.Hash()
|
|
|
@@ -652,7 +652,7 @@ func TestKVRestore(t *testing.T) {
|
|
|
}
|
|
|
for i, tt := range tests {
|
|
|
b, tmpPath := backend.NewDefaultTmpBackend()
|
|
|
- s := NewStore(b, &lease.FakeLessor{})
|
|
|
+ s := NewStore(b, &lease.FakeLessor{}, nil)
|
|
|
tt(s)
|
|
|
var kvss [][]storagepb.KeyValue
|
|
|
for k := int64(0); k < 10; k++ {
|
|
|
@@ -662,7 +662,7 @@ func TestKVRestore(t *testing.T) {
|
|
|
s.Close()
|
|
|
|
|
|
// ns should recover the the previous state from backend.
|
|
|
- ns := NewStore(b, &lease.FakeLessor{})
|
|
|
+ ns := NewStore(b, &lease.FakeLessor{}, nil)
|
|
|
// wait for possible compaction to finish
|
|
|
testutil.WaitSchedule()
|
|
|
var nkvss [][]storagepb.KeyValue
|
|
|
@@ -680,7 +680,7 @@ func TestKVRestore(t *testing.T) {
|
|
|
|
|
|
func TestKVSnapshot(t *testing.T) {
|
|
|
b, tmpPath := backend.NewDefaultTmpBackend()
|
|
|
- s := NewStore(b, &lease.FakeLessor{})
|
|
|
+ s := NewStore(b, &lease.FakeLessor{}, nil)
|
|
|
defer cleanup(s, b, tmpPath)
|
|
|
|
|
|
wkvs := put3TestKVs(s)
|
|
|
@@ -700,7 +700,7 @@ func TestKVSnapshot(t *testing.T) {
|
|
|
}
|
|
|
f.Close()
|
|
|
|
|
|
- ns := NewStore(b, &lease.FakeLessor{})
|
|
|
+ ns := NewStore(b, &lease.FakeLessor{}, nil)
|
|
|
defer ns.Close()
|
|
|
kvs, rev, err := ns.Range([]byte("a"), []byte("z"), 0, 0)
|
|
|
if err != nil {
|
|
|
@@ -716,7 +716,7 @@ func TestKVSnapshot(t *testing.T) {
|
|
|
|
|
|
func TestWatchableKVWatch(t *testing.T) {
|
|
|
b, tmpPath := backend.NewDefaultTmpBackend()
|
|
|
- s := WatchableKV(newWatchableStore(b, &lease.FakeLessor{}))
|
|
|
+ s := WatchableKV(newWatchableStore(b, &lease.FakeLessor{}, nil))
|
|
|
defer cleanup(s, b, tmpPath)
|
|
|
|
|
|
w := s.NewWatchStream()
|