123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306 |
- // Copyright 2015 The etcd Authors
- //
- // Licensed under the Apache License, Version 2.0 (the "License");
- // you may not use this file except in compliance with the License.
- // You may obtain a copy of the License at
- //
- // http://www.apache.org/licenses/LICENSE-2.0
- //
- // Unless required by applicable law or agreed to in writing, software
- // distributed under the License is distributed on an "AS IS" BASIS,
- // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- // See the License for the specific language governing permissions and
- // limitations under the License.
- package backend
- import (
- "fmt"
- "io/ioutil"
- "os"
- "reflect"
- "testing"
- "time"
- bolt "github.com/coreos/bbolt"
- )
- func TestBackendClose(t *testing.T) {
- b, tmpPath := NewTmpBackend(time.Hour, 10000)
- defer os.Remove(tmpPath)
- // check close could work
- done := make(chan struct{})
- go func() {
- err := b.Close()
- if err != nil {
- t.Errorf("close error = %v, want nil", err)
- }
- done <- struct{}{}
- }()
- select {
- case <-done:
- case <-time.After(10 * time.Second):
- t.Errorf("failed to close database in 10s")
- }
- }
- func TestBackendSnapshot(t *testing.T) {
- b, tmpPath := NewTmpBackend(time.Hour, 10000)
- defer cleanup(b, tmpPath)
- tx := b.BatchTx()
- tx.Lock()
- tx.UnsafeCreateBucket([]byte("test"))
- tx.UnsafePut([]byte("test"), []byte("foo"), []byte("bar"))
- tx.Unlock()
- b.ForceCommit()
- // write snapshot to a new file
- f, err := ioutil.TempFile(os.TempDir(), "etcd_backend_test")
- if err != nil {
- t.Fatal(err)
- }
- snap := b.Snapshot()
- defer snap.Close()
- if _, err := snap.WriteTo(f); err != nil {
- t.Fatal(err)
- }
- f.Close()
- // bootstrap new backend from the snapshot
- bcfg := DefaultBackendConfig()
- bcfg.Path, bcfg.BatchInterval, bcfg.BatchLimit = f.Name(), time.Hour, 10000
- nb := New(bcfg)
- defer cleanup(nb, f.Name())
- newTx := b.BatchTx()
- newTx.Lock()
- ks, _ := newTx.UnsafeRange([]byte("test"), []byte("foo"), []byte("goo"), 0)
- if len(ks) != 1 {
- t.Errorf("len(kvs) = %d, want 1", len(ks))
- }
- newTx.Unlock()
- }
- func TestBackendBatchIntervalCommit(t *testing.T) {
- // start backend with super short batch interval so
- // we do not need to wait long before commit to happen.
- b, tmpPath := NewTmpBackend(time.Nanosecond, 10000)
- defer cleanup(b, tmpPath)
- pc := b.Commits()
- tx := b.BatchTx()
- tx.Lock()
- tx.UnsafeCreateBucket([]byte("test"))
- tx.UnsafePut([]byte("test"), []byte("foo"), []byte("bar"))
- tx.Unlock()
- for i := 0; i < 10; i++ {
- if b.Commits() >= pc+1 {
- break
- }
- time.Sleep(time.Duration(i*100) * time.Millisecond)
- }
- // check whether put happens via db view
- b.db.View(func(tx *bolt.Tx) error {
- bucket := tx.Bucket([]byte("test"))
- if bucket == nil {
- t.Errorf("bucket test does not exit")
- return nil
- }
- v := bucket.Get([]byte("foo"))
- if v == nil {
- t.Errorf("foo key failed to written in backend")
- }
- return nil
- })
- }
- func TestBackendDefrag(t *testing.T) {
- b, tmpPath := NewDefaultTmpBackend()
- defer cleanup(b, tmpPath)
- tx := b.BatchTx()
- tx.Lock()
- tx.UnsafeCreateBucket([]byte("test"))
- for i := 0; i < defragLimit+100; i++ {
- tx.UnsafePut([]byte("test"), []byte(fmt.Sprintf("foo_%d", i)), []byte("bar"))
- }
- tx.Unlock()
- b.ForceCommit()
- // remove some keys to ensure the disk space will be reclaimed after defrag
- tx = b.BatchTx()
- tx.Lock()
- for i := 0; i < 50; i++ {
- tx.UnsafeDelete([]byte("test"), []byte(fmt.Sprintf("foo_%d", i)))
- }
- tx.Unlock()
- b.ForceCommit()
- size := b.Size()
- // shrink and check hash
- oh, err := b.Hash(nil)
- if err != nil {
- t.Fatal(err)
- }
- err = b.Defrag()
- if err != nil {
- t.Fatal(err)
- }
- nh, err := b.Hash(nil)
- if err != nil {
- t.Fatal(err)
- }
- if oh != nh {
- t.Errorf("hash = %v, want %v", nh, oh)
- }
- nsize := b.Size()
- if nsize >= size {
- t.Errorf("new size = %v, want < %d", nsize, size)
- }
- // try put more keys after shrink.
- tx = b.BatchTx()
- tx.Lock()
- tx.UnsafeCreateBucket([]byte("test"))
- tx.UnsafePut([]byte("test"), []byte("more"), []byte("bar"))
- tx.Unlock()
- b.ForceCommit()
- }
- // TestBackendWriteback ensures writes are stored to the read txn on write txn unlock.
- func TestBackendWriteback(t *testing.T) {
- b, tmpPath := NewDefaultTmpBackend()
- defer cleanup(b, tmpPath)
- tx := b.BatchTx()
- tx.Lock()
- tx.UnsafeCreateBucket([]byte("key"))
- tx.UnsafePut([]byte("key"), []byte("abc"), []byte("bar"))
- tx.UnsafePut([]byte("key"), []byte("def"), []byte("baz"))
- tx.UnsafePut([]byte("key"), []byte("overwrite"), []byte("1"))
- tx.Unlock()
- // overwrites should be propagated too
- tx.Lock()
- tx.UnsafePut([]byte("key"), []byte("overwrite"), []byte("2"))
- tx.Unlock()
- keys := []struct {
- key []byte
- end []byte
- limit int64
- wkey [][]byte
- wval [][]byte
- }{
- {
- key: []byte("abc"),
- end: nil,
- wkey: [][]byte{[]byte("abc")},
- wval: [][]byte{[]byte("bar")},
- },
- {
- key: []byte("abc"),
- end: []byte("def"),
- wkey: [][]byte{[]byte("abc")},
- wval: [][]byte{[]byte("bar")},
- },
- {
- key: []byte("abc"),
- end: []byte("deg"),
- wkey: [][]byte{[]byte("abc"), []byte("def")},
- wval: [][]byte{[]byte("bar"), []byte("baz")},
- },
- {
- key: []byte("abc"),
- end: []byte("\xff"),
- limit: 1,
- wkey: [][]byte{[]byte("abc")},
- wval: [][]byte{[]byte("bar")},
- },
- {
- key: []byte("abc"),
- end: []byte("\xff"),
- wkey: [][]byte{[]byte("abc"), []byte("def"), []byte("overwrite")},
- wval: [][]byte{[]byte("bar"), []byte("baz"), []byte("2")},
- },
- }
- rtx := b.ReadTx()
- for i, tt := range keys {
- rtx.Lock()
- k, v := rtx.UnsafeRange([]byte("key"), tt.key, tt.end, tt.limit)
- rtx.Unlock()
- if !reflect.DeepEqual(tt.wkey, k) || !reflect.DeepEqual(tt.wval, v) {
- t.Errorf("#%d: want k=%+v, v=%+v; got k=%+v, v=%+v", i, tt.wkey, tt.wval, k, v)
- }
- }
- }
- // TestBackendWritebackForEach checks that partially written / buffered
- // data is visited in the same order as fully committed data.
- func TestBackendWritebackForEach(t *testing.T) {
- b, tmpPath := NewTmpBackend(time.Hour, 10000)
- defer cleanup(b, tmpPath)
- tx := b.BatchTx()
- tx.Lock()
- tx.UnsafeCreateBucket([]byte("key"))
- for i := 0; i < 5; i++ {
- k := []byte(fmt.Sprintf("%04d", i))
- tx.UnsafePut([]byte("key"), k, []byte("bar"))
- }
- tx.Unlock()
- // writeback
- b.ForceCommit()
- tx.Lock()
- tx.UnsafeCreateBucket([]byte("key"))
- for i := 5; i < 20; i++ {
- k := []byte(fmt.Sprintf("%04d", i))
- tx.UnsafePut([]byte("key"), k, []byte("bar"))
- }
- tx.Unlock()
- seq := ""
- getSeq := func(k, v []byte) error {
- seq += string(k)
- return nil
- }
- rtx := b.ReadTx()
- rtx.Lock()
- rtx.UnsafeForEach([]byte("key"), getSeq)
- rtx.Unlock()
- partialSeq := seq
- seq = ""
- b.ForceCommit()
- tx.Lock()
- tx.UnsafeForEach([]byte("key"), getSeq)
- tx.Unlock()
- if seq != partialSeq {
- t.Fatalf("expected %q, got %q", seq, partialSeq)
- }
- }
- func cleanup(b Backend, path string) {
- b.Close()
- os.Remove(path)
- }
|