// Copyright 2017 The etcd Authors // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package integration import ( "context" "fmt" "math/rand" "reflect" "sync" "testing" "time" "go.etcd.io/etcd/clientv3" "go.etcd.io/etcd/clientv3/concurrency" "go.etcd.io/etcd/clientv3/leasing" "go.etcd.io/etcd/integration" "go.etcd.io/etcd/pkg/testutil" ) func TestLeasingPutGet(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) lKV1, closeLKV1, err := leasing.NewKV(clus.Client(0), "foo/") testutil.AssertNil(t, err) defer closeLKV1() lKV2, closeLKV2, err := leasing.NewKV(clus.Client(1), "foo/") testutil.AssertNil(t, err) defer closeLKV2() lKV3, closeLKV3, err := leasing.NewKV(clus.Client(2), "foo/") testutil.AssertNil(t, err) defer closeLKV3() resp, err := lKV1.Get(context.TODO(), "abc") if err != nil { t.Fatal(err) } if len(resp.Kvs) != 0 { t.Errorf("expected nil, got %q", resp.Kvs[0].Key) } if _, err = lKV1.Put(context.TODO(), "abc", "def"); err != nil { t.Fatal(err) } if resp, err = lKV2.Get(context.TODO(), "abc"); err != nil { t.Fatal(err) } if string(resp.Kvs[0].Key) != "abc" { t.Errorf("expected key=%q, got key=%q", "abc", resp.Kvs[0].Key) } if string(resp.Kvs[0].Value) != "def" { t.Errorf("expected value=%q, got value=%q", "bar", resp.Kvs[0].Value) } if _, err = lKV3.Get(context.TODO(), "abc"); err != nil { t.Fatal(err) } if _, err = lKV2.Put(context.TODO(), "abc", "ghi"); err != nil { t.Fatal(err) } if resp, err = lKV3.Get(context.TODO(), "abc"); err != nil { t.Fatal(err) } if string(resp.Kvs[0].Key) != "abc" { t.Errorf("expected key=%q, got key=%q", "abc", resp.Kvs[0].Key) } if string(resp.Kvs[0].Value) != "ghi" { t.Errorf("expected value=%q, got value=%q", "bar", resp.Kvs[0].Value) } } // TestLeasingInterval checks the leasing KV fetches key intervals. func TestLeasingInterval(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() keys := []string{"abc/a", "abc/b", "abc/a/a"} for _, k := range keys { if _, err = clus.Client(0).Put(context.TODO(), k, "v"); err != nil { t.Fatal(err) } } resp, err := lkv.Get(context.TODO(), "abc/", clientv3.WithPrefix()) if err != nil { t.Fatal(err) } if len(resp.Kvs) != 3 { t.Fatalf("expected keys %+v, got response keys %+v", keys, resp.Kvs) } // load into cache if _, err = lkv.Get(context.TODO(), "abc/a"); err != nil { t.Fatal(err) } // get when prefix is also a cached key if resp, err = lkv.Get(context.TODO(), "abc/a", clientv3.WithPrefix()); err != nil { t.Fatal(err) } if len(resp.Kvs) != 2 { t.Fatalf("expected keys %+v, got response keys %+v", keys, resp.Kvs) } } // TestLeasingPutInvalidateNew checks the leasing KV updates its cache on a Put to a new key. func TestLeasingPutInvalidateNew(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() if _, err = lkv.Get(context.TODO(), "k"); err != nil { t.Fatal(err) } if _, err = lkv.Put(context.TODO(), "k", "v"); err != nil { t.Fatal(err) } lkvResp, err := lkv.Get(context.TODO(), "k") if err != nil { t.Fatal(err) } cResp, cerr := clus.Client(0).Get(context.TODO(), "k") if cerr != nil { t.Fatal(cerr) } if !reflect.DeepEqual(lkvResp, cResp) { t.Fatalf(`expected %+v, got response %+v`, cResp, lkvResp) } } // TestLeasingPutInvalidateExisting checks the leasing KV updates its cache on a Put to an existing key. func TestLeasingPutInvalidateExisting(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) if _, err := clus.Client(0).Put(context.TODO(), "k", "abc"); err != nil { t.Fatal(err) } lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() if _, err = lkv.Get(context.TODO(), "k"); err != nil { t.Fatal(err) } if _, err = lkv.Put(context.TODO(), "k", "v"); err != nil { t.Fatal(err) } lkvResp, err := lkv.Get(context.TODO(), "k") if err != nil { t.Fatal(err) } cResp, cerr := clus.Client(0).Get(context.TODO(), "k") if cerr != nil { t.Fatal(cerr) } if !reflect.DeepEqual(lkvResp, cResp) { t.Fatalf(`expected %+v, got response %+v`, cResp, lkvResp) } } // TestLeasingGetNoLeaseTTL checks a key with a TTL is not leased. func TestLeasingGetNoLeaseTTL(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() lresp, err := clus.Client(0).Grant(context.TODO(), 60) testutil.AssertNil(t, err) _, err = clus.Client(0).Put(context.TODO(), "k", "v", clientv3.WithLease(lresp.ID)) testutil.AssertNil(t, err) gresp, err := lkv.Get(context.TODO(), "k") testutil.AssertNil(t, err) testutil.AssertEqual(t, len(gresp.Kvs), 1) clus.Members[0].Stop(t) ctx, cancel := context.WithTimeout(context.TODO(), time.Second) _, err = lkv.Get(ctx, "k") cancel() testutil.AssertEqual(t, err, ctx.Err()) } // TestLeasingGetSerializable checks the leasing KV can make serialized requests // when the etcd cluster is partitioned. func TestLeasingGetSerializable(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 2}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() if _, err = clus.Client(0).Put(context.TODO(), "cached", "abc"); err != nil { t.Fatal(err) } if _, err = lkv.Get(context.TODO(), "cached"); err != nil { t.Fatal(err) } clus.Members[1].Stop(t) // don't necessarily try to acquire leasing key ownership for new key resp, err := lkv.Get(context.TODO(), "uncached", clientv3.WithSerializable()) if err != nil { t.Fatal(err) } if len(resp.Kvs) != 0 { t.Fatalf(`expected no keys, got response %+v`, resp) } clus.Members[0].Stop(t) // leasing key ownership should have "cached" locally served cachedResp, err := lkv.Get(context.TODO(), "cached", clientv3.WithSerializable()) if err != nil { t.Fatal(err) } if len(cachedResp.Kvs) != 1 || string(cachedResp.Kvs[0].Value) != "abc" { t.Fatalf(`expected "cached"->"abc", got response %+v`, cachedResp) } } // TestLeasingPrevKey checks the cache respects WithPrevKV on puts. func TestLeasingPrevKey(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 2}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() if _, err = clus.Client(0).Put(context.TODO(), "k", "abc"); err != nil { t.Fatal(err) } // acquire leasing key if _, err = lkv.Get(context.TODO(), "k"); err != nil { t.Fatal(err) } resp, err := lkv.Put(context.TODO(), "k", "def", clientv3.WithPrevKV()) if err != nil { t.Fatal(err) } if resp.PrevKv == nil || string(resp.PrevKv.Value) != "abc" { t.Fatalf(`expected PrevKV.Value="abc", got response %+v`, resp) } } // TestLeasingRevGet checks the cache respects Get by Revision. func TestLeasingRevGet(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() putResp, err := clus.Client(0).Put(context.TODO(), "k", "abc") if err != nil { t.Fatal(err) } if _, err = clus.Client(0).Put(context.TODO(), "k", "def"); err != nil { t.Fatal(err) } // check historic revision getResp, gerr := lkv.Get(context.TODO(), "k", clientv3.WithRev(putResp.Header.Revision)) if gerr != nil { t.Fatal(gerr) } if len(getResp.Kvs) != 1 || string(getResp.Kvs[0].Value) != "abc" { t.Fatalf(`expected "k"->"abc" at rev=%d, got response %+v`, putResp.Header.Revision, getResp) } // check current revision getResp, gerr = lkv.Get(context.TODO(), "k") if gerr != nil { t.Fatal(gerr) } if len(getResp.Kvs) != 1 || string(getResp.Kvs[0].Value) != "def" { t.Fatalf(`expected "k"->"abc" at rev=%d, got response %+v`, putResp.Header.Revision, getResp) } } // TestLeasingGetWithOpts checks options that can be served through the cache do not depend on the server. func TestLeasingGetWithOpts(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() if _, err = clus.Client(0).Put(context.TODO(), "k", "abc"); err != nil { t.Fatal(err) } // in cache if _, err = lkv.Get(context.TODO(), "k", clientv3.WithKeysOnly()); err != nil { t.Fatal(err) } clus.Members[0].Stop(t) opts := []clientv3.OpOption{ clientv3.WithKeysOnly(), clientv3.WithLimit(1), clientv3.WithMinCreateRev(1), clientv3.WithMinModRev(1), clientv3.WithSort(clientv3.SortByKey, clientv3.SortAscend), clientv3.WithSerializable(), } for _, opt := range opts { if _, err := lkv.Get(context.TODO(), "k", opt); err != nil { t.Fatal(err) } } getOpts := []clientv3.OpOption{} for i := 0; i < len(opts); i++ { getOpts = append(getOpts, opts[rand.Intn(len(opts))]) } getOpts = getOpts[:rand.Intn(len(opts))] if _, err := lkv.Get(context.TODO(), "k", getOpts...); err != nil { t.Fatal(err) } } // TestLeasingConcurrentPut ensures that a get after concurrent puts returns // the recently put data. func TestLeasingConcurrentPut(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() // force key into leasing key cache if _, err = lkv.Get(context.TODO(), "k"); err != nil { t.Fatal(err) } // concurrently put through leasing client numPuts := 16 putc := make(chan *clientv3.PutResponse, numPuts) for i := 0; i < numPuts; i++ { go func() { resp, perr := lkv.Put(context.TODO(), "k", "abc") if perr != nil { t.Error(perr) } putc <- resp }() } // record maximum revision from puts maxRev := int64(0) for i := 0; i < numPuts; i++ { if resp := <-putc; resp.Header.Revision > maxRev { maxRev = resp.Header.Revision } } // confirm Get gives most recently put revisions getResp, gerr := lkv.Get(context.TODO(), "k") if gerr != nil { t.Fatal(err) } if mr := getResp.Kvs[0].ModRevision; mr != maxRev { t.Errorf("expected ModRevision %d, got %d", maxRev, mr) } if ver := getResp.Kvs[0].Version; ver != int64(numPuts) { t.Errorf("expected Version %d, got %d", numPuts, ver) } } func TestLeasingDisconnectedGet(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() if _, err = clus.Client(0).Put(context.TODO(), "cached", "abc"); err != nil { t.Fatal(err) } // get key so it's cached if _, err = lkv.Get(context.TODO(), "cached"); err != nil { t.Fatal(err) } clus.Members[0].Stop(t) // leasing key ownership should have "cached" locally served cachedResp, err := lkv.Get(context.TODO(), "cached") if err != nil { t.Fatal(err) } if len(cachedResp.Kvs) != 1 || string(cachedResp.Kvs[0].Value) != "abc" { t.Fatalf(`expected "cached"->"abc", got response %+v`, cachedResp) } } func TestLeasingDeleteOwner(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() if _, err = clus.Client(0).Put(context.TODO(), "k", "abc"); err != nil { t.Fatal(err) } // get+own / delete / get if _, err = lkv.Get(context.TODO(), "k"); err != nil { t.Fatal(err) } if _, err = lkv.Delete(context.TODO(), "k"); err != nil { t.Fatal(err) } resp, err := lkv.Get(context.TODO(), "k") if err != nil { t.Fatal(err) } if len(resp.Kvs) != 0 { t.Fatalf(`expected "k" to be deleted, got response %+v`, resp) } // try to double delete if _, err = lkv.Delete(context.TODO(), "k"); err != nil { t.Fatal(err) } } func TestLeasingDeleteNonOwner(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv1, closeLKV1, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV1() lkv2, closeLKV2, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV2() if _, err = clus.Client(0).Put(context.TODO(), "k", "abc"); err != nil { t.Fatal(err) } // acquire ownership if _, err = lkv1.Get(context.TODO(), "k"); err != nil { t.Fatal(err) } // delete via non-owner if _, err = lkv2.Delete(context.TODO(), "k"); err != nil { t.Fatal(err) } // key should be removed from lkv1 resp, err := lkv1.Get(context.TODO(), "k") if err != nil { t.Fatal(err) } if len(resp.Kvs) != 0 { t.Fatalf(`expected "k" to be deleted, got response %+v`, resp) } } func TestLeasingOverwriteResponse(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() if _, err = clus.Client(0).Put(context.TODO(), "k", "abc"); err != nil { t.Fatal(err) } resp, err := lkv.Get(context.TODO(), "k") if err != nil { t.Fatal(err) } resp.Kvs[0].Key[0] = 'z' resp.Kvs[0].Value[0] = 'z' resp, err = lkv.Get(context.TODO(), "k") if err != nil { t.Fatal(err) } if string(resp.Kvs[0].Key) != "k" { t.Errorf(`expected key "k", got %q`, string(resp.Kvs[0].Key)) } if string(resp.Kvs[0].Value) != "abc" { t.Errorf(`expected value "abc", got %q`, string(resp.Kvs[0].Value)) } } func TestLeasingOwnerPutResponse(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() if _, err = clus.Client(0).Put(context.TODO(), "k", "abc"); err != nil { t.Fatal(err) } _, gerr := lkv.Get(context.TODO(), "k") if gerr != nil { t.Fatal(gerr) } presp, err := lkv.Put(context.TODO(), "k", "def") if err != nil { t.Fatal(err) } if presp == nil { t.Fatal("expected put response, got nil") } clus.Members[0].Stop(t) gresp, gerr := lkv.Get(context.TODO(), "k") if gerr != nil { t.Fatal(gerr) } if gresp.Kvs[0].ModRevision != presp.Header.Revision { t.Errorf("expected mod revision %d, got %d", presp.Header.Revision, gresp.Kvs[0].ModRevision) } if gresp.Kvs[0].Version != 2 { t.Errorf("expected version 2, got version %d", gresp.Kvs[0].Version) } } func TestLeasingTxnOwnerGetRange(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() keyCount := rand.Intn(10) + 1 for i := 0; i < keyCount; i++ { k := fmt.Sprintf("k-%d", i) if _, err := clus.Client(0).Put(context.TODO(), k, k+k); err != nil { t.Fatal(err) } } if _, err := lkv.Get(context.TODO(), "k-"); err != nil { t.Fatal(err) } tresp, terr := lkv.Txn(context.TODO()).Then(clientv3.OpGet("k-", clientv3.WithPrefix())).Commit() if terr != nil { t.Fatal(terr) } if resp := tresp.Responses[0].GetResponseRange(); len(resp.Kvs) != keyCount { t.Fatalf("expected %d keys, got response %+v", keyCount, resp.Kvs) } } func TestLeasingTxnOwnerGet(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() keyCount := rand.Intn(10) + 1 var ops []clientv3.Op presps := make([]*clientv3.PutResponse, keyCount) for i := range presps { k := fmt.Sprintf("k-%d", i) presp, err := clus.Client(0).Put(context.TODO(), k, k+k) if err != nil { t.Fatal(err) } presps[i] = presp if _, err = lkv.Get(context.TODO(), k); err != nil { t.Fatal(err) } ops = append(ops, clientv3.OpGet(k)) } ops = ops[:rand.Intn(len(ops))] // served through cache clus.Members[0].Stop(t) var thenOps, elseOps []clientv3.Op cmps, useThen := randCmps("k-", presps) if useThen { thenOps = ops elseOps = []clientv3.Op{clientv3.OpPut("k", "1")} } else { thenOps = []clientv3.Op{clientv3.OpPut("k", "1")} elseOps = ops } tresp, terr := lkv.Txn(context.TODO()). If(cmps...). Then(thenOps...). Else(elseOps...).Commit() if terr != nil { t.Fatal(terr) } if tresp.Succeeded != useThen { t.Fatalf("expected succeeded=%v, got tresp=%+v", useThen, tresp) } if len(tresp.Responses) != len(ops) { t.Fatalf("expected %d responses, got %d", len(ops), len(tresp.Responses)) } wrev := presps[len(presps)-1].Header.Revision if tresp.Header.Revision < wrev { t.Fatalf("expected header revision >= %d, got %d", wrev, tresp.Header.Revision) } for i := range ops { k := fmt.Sprintf("k-%d", i) rr := tresp.Responses[i].GetResponseRange() if rr == nil { t.Errorf("expected get response, got %+v", tresp.Responses[i]) } if string(rr.Kvs[0].Key) != k || string(rr.Kvs[0].Value) != k+k { t.Errorf(`expected key for %q, got %+v`, k, rr.Kvs) } } } func TestLeasingTxnOwnerDeleteRange(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() keyCount := rand.Intn(10) + 1 for i := 0; i < keyCount; i++ { k := fmt.Sprintf("k-%d", i) if _, perr := clus.Client(0).Put(context.TODO(), k, k+k); perr != nil { t.Fatal(perr) } } // cache in lkv resp, err := lkv.Get(context.TODO(), "k-", clientv3.WithPrefix()) if err != nil { t.Fatal(err) } if len(resp.Kvs) != keyCount { t.Fatalf("expected %d keys, got %d", keyCount, len(resp.Kvs)) } if _, terr := lkv.Txn(context.TODO()).Then(clientv3.OpDelete("k-", clientv3.WithPrefix())).Commit(); terr != nil { t.Fatal(terr) } resp, err = lkv.Get(context.TODO(), "k-", clientv3.WithPrefix()) if err != nil { t.Fatal(err) } if len(resp.Kvs) != 0 { t.Fatalf("expected no keys, got %d", len(resp.Kvs)) } } func TestLeasingTxnOwnerDelete(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() if _, err = clus.Client(0).Put(context.TODO(), "k", "abc"); err != nil { t.Fatal(err) } // cache in lkv if _, gerr := lkv.Get(context.TODO(), "k"); gerr != nil { t.Fatal(gerr) } if _, terr := lkv.Txn(context.TODO()).Then(clientv3.OpDelete("k")).Commit(); terr != nil { t.Fatal(terr) } resp, err := lkv.Get(context.TODO(), "k") if err != nil { t.Fatal(err) } if len(resp.Kvs) != 0 { t.Fatalf("expected no keys, got %d", len(resp.Kvs)) } } func TestLeasingTxnOwnerIf(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() if _, err = clus.Client(0).Put(context.TODO(), "k", "abc"); err != nil { t.Fatal(err) } if _, err = lkv.Get(context.TODO(), "k"); err != nil { t.Fatal(err) } // served through cache clus.Members[0].Stop(t) tests := []struct { cmps []clientv3.Cmp wSucceeded bool wResponses int }{ // success { cmps: []clientv3.Cmp{clientv3.Compare(clientv3.Value("k"), "=", "abc")}, wSucceeded: true, wResponses: 1, }, { cmps: []clientv3.Cmp{clientv3.Compare(clientv3.CreateRevision("k"), "=", 2)}, wSucceeded: true, wResponses: 1, }, { cmps: []clientv3.Cmp{clientv3.Compare(clientv3.ModRevision("k"), "=", 2)}, wSucceeded: true, wResponses: 1, }, { cmps: []clientv3.Cmp{clientv3.Compare(clientv3.Version("k"), "=", 1)}, wSucceeded: true, wResponses: 1, }, // failure { cmps: []clientv3.Cmp{clientv3.Compare(clientv3.Value("k"), ">", "abc")}, }, { cmps: []clientv3.Cmp{clientv3.Compare(clientv3.CreateRevision("k"), ">", 2)}, }, { cmps: []clientv3.Cmp{clientv3.Compare(clientv3.ModRevision("k"), "=", 2)}, wSucceeded: true, wResponses: 1, }, { cmps: []clientv3.Cmp{clientv3.Compare(clientv3.Version("k"), ">", 1)}, }, { cmps: []clientv3.Cmp{clientv3.Compare(clientv3.Value("k"), "<", "abc")}, }, { cmps: []clientv3.Cmp{clientv3.Compare(clientv3.CreateRevision("k"), "<", 2)}, }, { cmps: []clientv3.Cmp{clientv3.Compare(clientv3.ModRevision("k"), "<", 2)}, }, { cmps: []clientv3.Cmp{clientv3.Compare(clientv3.Version("k"), "<", 1)}, }, { cmps: []clientv3.Cmp{ clientv3.Compare(clientv3.Version("k"), "=", 1), clientv3.Compare(clientv3.Version("k"), "<", 1), }, }, } for i, tt := range tests { tresp, terr := lkv.Txn(context.TODO()).If(tt.cmps...).Then(clientv3.OpGet("k")).Commit() if terr != nil { t.Fatal(terr) } if tresp.Succeeded != tt.wSucceeded { t.Errorf("#%d: expected succeeded %v, got %v", i, tt.wSucceeded, tresp.Succeeded) } if len(tresp.Responses) != tt.wResponses { t.Errorf("#%d: expected %d responses, got %d", i, tt.wResponses, len(tresp.Responses)) } } } func TestLeasingTxnCancel(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) lkv1, closeLKV1, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV1() lkv2, closeLKV2, err := leasing.NewKV(clus.Client(1), "pfx/") testutil.AssertNil(t, err) defer closeLKV2() // acquire lease but disconnect so no revoke in time if _, err = lkv1.Get(context.TODO(), "k"); err != nil { t.Fatal(err) } clus.Members[0].Stop(t) // wait for leader election, if any if _, err = clus.Client(1).Get(context.TODO(), "abc"); err != nil { t.Fatal(err) } ctx, cancel := context.WithCancel(context.TODO()) go func() { time.Sleep(100 * time.Millisecond) cancel() }() if _, err := lkv2.Txn(ctx).Then(clientv3.OpPut("k", "v")).Commit(); err != context.Canceled { t.Fatalf("expected %v, got %v", context.Canceled, err) } } func TestLeasingTxnNonOwnerPut(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() lkv2, closeLKV2, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV2() if _, err = clus.Client(0).Put(context.TODO(), "k", "abc"); err != nil { t.Fatal(err) } if _, err = clus.Client(0).Put(context.TODO(), "k2", "123"); err != nil { t.Fatal(err) } // cache in lkv if _, err = lkv.Get(context.TODO(), "k"); err != nil { t.Fatal(err) } if _, err = lkv.Get(context.TODO(), "k2"); err != nil { t.Fatal(err) } // invalidate via lkv2 txn opArray := make([]clientv3.Op, 0) opArray = append(opArray, clientv3.OpPut("k2", "456")) tresp, terr := lkv2.Txn(context.TODO()).Then( clientv3.OpTxn(nil, opArray, nil), clientv3.OpPut("k", "def"), clientv3.OpPut("k3", "999"), // + a key not in any cache ).Commit() if terr != nil { t.Fatal(terr) } if !tresp.Succeeded || len(tresp.Responses) != 3 { t.Fatalf("expected txn success, got %+v", tresp) } // check cache was invalidated gresp, gerr := lkv.Get(context.TODO(), "k") if gerr != nil { t.Fatal(err) } if len(gresp.Kvs) != 1 || string(gresp.Kvs[0].Value) != "def" { t.Errorf(`expected value "def", got %+v`, gresp) } gresp, gerr = lkv.Get(context.TODO(), "k2") if gerr != nil { t.Fatal(gerr) } if len(gresp.Kvs) != 1 || string(gresp.Kvs[0].Value) != "456" { t.Errorf(`expected value "def", got %+v`, gresp) } // check puts were applied and are all in the same revision w := clus.Client(0).Watch( clus.Client(0).Ctx(), "k", clientv3.WithRev(tresp.Header.Revision), clientv3.WithPrefix()) wresp := <-w c := 0 evs := []clientv3.Event{} for _, ev := range wresp.Events { evs = append(evs, *ev) if ev.Kv.ModRevision == tresp.Header.Revision { c++ } } if c != 3 { t.Fatalf("expected 3 put events, got %+v", evs) } } // TestLeasingTxnRandIfThenOrElse randomly leases keys two separate clients, then // issues a random If/{Then,Else} transaction on those keys to one client. func TestLeasingTxnRandIfThenOrElse(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv1, closeLKV1, err1 := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err1) defer closeLKV1() lkv2, closeLKV2, err2 := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err2) defer closeLKV2() keyCount := 16 dat := make([]*clientv3.PutResponse, keyCount) for i := 0; i < keyCount; i++ { k, v := fmt.Sprintf("k-%d", i), fmt.Sprintf("%d", i) dat[i], err1 = clus.Client(0).Put(context.TODO(), k, v) if err1 != nil { t.Fatal(err1) } } // nondeterministically populate leasing caches var wg sync.WaitGroup getc := make(chan struct{}, keyCount) getRandom := func(kv clientv3.KV) { defer wg.Done() for i := 0; i < keyCount/2; i++ { k := fmt.Sprintf("k-%d", rand.Intn(keyCount)) if _, err := kv.Get(context.TODO(), k); err != nil { t.Error(err) } getc <- struct{}{} } } wg.Add(2) defer wg.Wait() go getRandom(lkv1) go getRandom(lkv2) // random list of comparisons, all true cmps, useThen := randCmps("k-", dat) // random list of puts/gets; unique keys ops := []clientv3.Op{} usedIdx := make(map[int]struct{}) for i := 0; i < keyCount; i++ { idx := rand.Intn(keyCount) if _, ok := usedIdx[idx]; ok { continue } usedIdx[idx] = struct{}{} k := fmt.Sprintf("k-%d", idx) switch rand.Intn(2) { case 0: ops = append(ops, clientv3.OpGet(k)) case 1: ops = append(ops, clientv3.OpPut(k, "a")) // TODO: add delete } } // random lengths ops = ops[:rand.Intn(len(ops))] // wait for some gets to populate the leasing caches before committing for i := 0; i < keyCount/2; i++ { <-getc } // randomly choose between then and else blocks var thenOps, elseOps []clientv3.Op if useThen { thenOps = ops } else { // force failure elseOps = ops } tresp, terr := lkv1.Txn(context.TODO()).If(cmps...).Then(thenOps...).Else(elseOps...).Commit() if terr != nil { t.Fatal(terr) } // cmps always succeed if tresp.Succeeded != useThen { t.Fatalf("expected succeeded=%v, got tresp=%+v", useThen, tresp) } // get should match what was put checkPuts := func(s string, kv clientv3.KV) { for _, op := range ops { if !op.IsPut() { continue } resp, rerr := kv.Get(context.TODO(), string(op.KeyBytes())) if rerr != nil { t.Fatal(rerr) } if len(resp.Kvs) != 1 || string(resp.Kvs[0].Value) != "a" { t.Fatalf(`%s: expected value="a", got %+v`, s, resp.Kvs) } } } checkPuts("client(0)", clus.Client(0)) checkPuts("lkv1", lkv1) checkPuts("lkv2", lkv2) } func TestLeasingOwnerPutError(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() if _, err = lkv.Get(context.TODO(), "k"); err != nil { t.Fatal(err) } clus.Members[0].Stop(t) ctx, cancel := context.WithTimeout(context.TODO(), 100*time.Millisecond) defer cancel() if resp, err := lkv.Put(ctx, "k", "v"); err == nil { t.Fatalf("expected error, got response %+v", resp) } } func TestLeasingOwnerDeleteError(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() if _, err = lkv.Get(context.TODO(), "k"); err != nil { t.Fatal(err) } clus.Members[0].Stop(t) ctx, cancel := context.WithTimeout(context.TODO(), 100*time.Millisecond) defer cancel() if resp, err := lkv.Delete(ctx, "k"); err == nil { t.Fatalf("expected error, got response %+v", resp) } } func TestLeasingNonOwnerPutError(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() clus.Members[0].Stop(t) ctx, cancel := context.WithTimeout(context.TODO(), 100*time.Millisecond) defer cancel() if resp, err := lkv.Put(ctx, "k", "v"); err == nil { t.Fatalf("expected error, got response %+v", resp) } } func TestLeasingOwnerDeletePrefix(t *testing.T) { testLeasingOwnerDelete(t, clientv3.OpDelete("key/", clientv3.WithPrefix())) } func TestLeasingOwnerDeleteFrom(t *testing.T) { testLeasingOwnerDelete(t, clientv3.OpDelete("kd", clientv3.WithFromKey())) } func testLeasingOwnerDelete(t *testing.T, del clientv3.Op) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "0/") testutil.AssertNil(t, err) defer closeLKV() for i := 0; i < 8; i++ { if _, err = clus.Client(0).Put(context.TODO(), fmt.Sprintf("key/%d", i), "123"); err != nil { t.Fatal(err) } } if _, err = lkv.Get(context.TODO(), "key/1"); err != nil { t.Fatal(err) } opResp, delErr := lkv.Do(context.TODO(), del) if delErr != nil { t.Fatal(delErr) } delResp := opResp.Del() // confirm keys are invalidated from cache and deleted on etcd for i := 0; i < 8; i++ { resp, err := lkv.Get(context.TODO(), fmt.Sprintf("key/%d", i)) if err != nil { t.Fatal(err) } if len(resp.Kvs) != 0 { t.Fatalf("expected no keys on key/%d, got %+v", i, resp) } } // confirm keys were deleted atomically w := clus.Client(0).Watch( clus.Client(0).Ctx(), "key/", clientv3.WithRev(delResp.Header.Revision), clientv3.WithPrefix()) if wresp := <-w; len(wresp.Events) != 8 { t.Fatalf("expected %d delete events,got %d", 8, len(wresp.Events)) } } func TestLeasingDeleteRangeBounds(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) delkv, closeDelKV, err := leasing.NewKV(clus.Client(0), "0/") testutil.AssertNil(t, err) defer closeDelKV() getkv, closeGetKv, err := leasing.NewKV(clus.Client(0), "0/") testutil.AssertNil(t, err) defer closeGetKv() for _, k := range []string{"j", "m"} { if _, err = clus.Client(0).Put(context.TODO(), k, "123"); err != nil { t.Fatal(err) } if _, err = getkv.Get(context.TODO(), k); err != nil { t.Fatal(err) } } if _, err = delkv.Delete(context.TODO(), "k", clientv3.WithPrefix()); err != nil { t.Fatal(err) } // leases still on server? for _, k := range []string{"j", "m"} { resp, geterr := clus.Client(0).Get(context.TODO(), "0/"+k, clientv3.WithPrefix()) if geterr != nil { t.Fatal(geterr) } if len(resp.Kvs) != 1 { t.Fatalf("expected leasing key, got %+v", resp) } } // j and m should still have leases registered since not under k* clus.Members[0].Stop(t) if _, err = getkv.Get(context.TODO(), "j"); err != nil { t.Fatal(err) } if _, err = getkv.Get(context.TODO(), "m"); err != nil { t.Fatal(err) } } func TestLeasingDeleteRangeContendTxn(t *testing.T) { then := []clientv3.Op{clientv3.OpDelete("key/", clientv3.WithPrefix())} testLeasingDeleteRangeContend(t, clientv3.OpTxn(nil, then, nil)) } func TestLeaseDeleteRangeContendDel(t *testing.T) { op := clientv3.OpDelete("key/", clientv3.WithPrefix()) testLeasingDeleteRangeContend(t, op) } func testLeasingDeleteRangeContend(t *testing.T, op clientv3.Op) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) delkv, closeDelKV, err := leasing.NewKV(clus.Client(0), "0/") testutil.AssertNil(t, err) defer closeDelKV() putkv, closePutKV, err := leasing.NewKV(clus.Client(0), "0/") testutil.AssertNil(t, err) defer closePutKV() for i := 0; i < 8; i++ { key := fmt.Sprintf("key/%d", i) if _, err = clus.Client(0).Put(context.TODO(), key, "123"); err != nil { t.Fatal(err) } if _, err = putkv.Get(context.TODO(), key); err != nil { t.Fatal(err) } } ctx, cancel := context.WithCancel(context.TODO()) donec := make(chan struct{}) go func() { defer close(donec) for i := 0; ctx.Err() == nil; i++ { key := fmt.Sprintf("key/%d", i%8) putkv.Put(ctx, key, "123") putkv.Get(ctx, key) } }() _, delErr := delkv.Do(context.TODO(), op) cancel() <-donec if delErr != nil { t.Fatal(delErr) } // confirm keys on non-deleter match etcd for i := 0; i < 8; i++ { key := fmt.Sprintf("key/%d", i) resp, err := putkv.Get(context.TODO(), key) if err != nil { t.Fatal(err) } servResp, err := clus.Client(0).Get(context.TODO(), key) if err != nil { t.Fatal(err) } if !reflect.DeepEqual(resp.Kvs, servResp.Kvs) { t.Errorf("#%d: expected %+v, got %+v", i, servResp.Kvs, resp.Kvs) } } } func TestLeasingPutGetDeleteConcurrent(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkvs := make([]clientv3.KV, 16) for i := range lkvs { lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") testutil.AssertNil(t, err) defer closeLKV() lkvs[i] = lkv } getdel := func(kv clientv3.KV) { if _, err := kv.Put(context.TODO(), "k", "abc"); err != nil { t.Fatal(err) } time.Sleep(time.Millisecond) if _, err := kv.Get(context.TODO(), "k"); err != nil { t.Fatal(err) } if _, err := kv.Delete(context.TODO(), "k"); err != nil { t.Fatal(err) } time.Sleep(2 * time.Millisecond) } var wg sync.WaitGroup wg.Add(16) for i := 0; i < 16; i++ { go func() { defer wg.Done() for _, kv := range lkvs { getdel(kv) } }() } wg.Wait() resp, err := lkvs[0].Get(context.TODO(), "k") if err != nil { t.Fatal(err) } if len(resp.Kvs) > 0 { t.Fatalf("expected no kvs, got %+v", resp.Kvs) } resp, err = clus.Client(0).Get(context.TODO(), "k") if err != nil { t.Fatal(err) } if len(resp.Kvs) > 0 { t.Fatalf("expected no kvs, got %+v", resp.Kvs) } } // TestLeasingReconnectOwnerRevoke checks that revocation works if // disconnected when trying to submit revoke txn. func TestLeasingReconnectOwnerRevoke(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) lkv1, closeLKV1, err1 := leasing.NewKV(clus.Client(0), "foo/") testutil.AssertNil(t, err1) defer closeLKV1() lkv2, closeLKV2, err2 := leasing.NewKV(clus.Client(1), "foo/") testutil.AssertNil(t, err2) defer closeLKV2() if _, err := lkv1.Get(context.TODO(), "k"); err != nil { t.Fatal(err) } // force leader away from member 0 clus.Members[0].Stop(t) clus.WaitLeader(t) clus.Members[0].Restart(t) cctx, cancel := context.WithCancel(context.TODO()) sdonec, pdonec := make(chan struct{}), make(chan struct{}) // make lkv1 connection choppy so Txn fails go func() { defer close(sdonec) for i := 0; i < 3 && cctx.Err() == nil; i++ { clus.Members[0].Stop(t) time.Sleep(10 * time.Millisecond) clus.Members[0].Restart(t) } }() go func() { defer close(pdonec) if _, err := lkv2.Put(cctx, "k", "v"); err != nil { t.Log(err) } // blocks until lkv1 connection comes back resp, err := lkv1.Get(cctx, "k") if err != nil { t.Error(err) } if string(resp.Kvs[0].Value) != "v" { t.Errorf(`expected "v" value, got %+v`, resp) } }() select { case <-pdonec: cancel() <-sdonec case <-time.After(15 * time.Second): cancel() <-sdonec <-pdonec t.Fatal("took too long to revoke and put") } } // TestLeasingReconnectOwnerRevokeCompact checks that revocation works if // disconnected and the watch is compacted. func TestLeasingReconnectOwnerRevokeCompact(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) lkv1, closeLKV1, err1 := leasing.NewKV(clus.Client(0), "foo/") testutil.AssertNil(t, err1) defer closeLKV1() lkv2, closeLKV2, err2 := leasing.NewKV(clus.Client(1), "foo/") testutil.AssertNil(t, err2) defer closeLKV2() if _, err := lkv1.Get(context.TODO(), "k"); err != nil { t.Fatal(err) } clus.Members[0].Stop(t) clus.WaitLeader(t) // put some more revisions for compaction _, err := clus.Client(1).Put(context.TODO(), "a", "123") if err != nil { t.Fatal(err) } presp, err := clus.Client(1).Put(context.TODO(), "a", "123") if err != nil { t.Fatal(err) } // compact while lkv1 is disconnected rev := presp.Header.Revision if _, err = clus.Client(1).Compact(context.TODO(), rev); err != nil { t.Fatal(err) } clus.Members[0].Restart(t) cctx, cancel := context.WithTimeout(context.TODO(), 5*time.Second) defer cancel() if _, err = lkv2.Put(cctx, "k", "v"); err != nil { t.Fatal(err) } resp, err := lkv1.Get(cctx, "k") if err != nil { t.Fatal(err) } if string(resp.Kvs[0].Value) != "v" { t.Fatalf(`expected "v" value, got %+v`, resp) } } // TestLeasingReconnectOwnerConsistency checks a write error on an owner will // not cause inconsistency between the server and the client. func TestLeasingReconnectOwnerConsistency(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/") defer closeLKV() testutil.AssertNil(t, err) if _, err = lkv.Put(context.TODO(), "k", "x"); err != nil { t.Fatal(err) } if _, err = lkv.Put(context.TODO(), "kk", "y"); err != nil { t.Fatal(err) } if _, err = lkv.Get(context.TODO(), "k"); err != nil { t.Fatal(err) } for i := 0; i < 10; i++ { v := fmt.Sprintf("%d", i) donec := make(chan struct{}) clus.Members[0].DropConnections() go func() { defer close(donec) for i := 0; i < 20; i++ { clus.Members[0].DropConnections() time.Sleep(time.Millisecond) } }() switch rand.Intn(7) { case 0: _, err = lkv.Put(context.TODO(), "k", v) case 1: _, err = lkv.Delete(context.TODO(), "k") case 2: txn := lkv.Txn(context.TODO()).Then( clientv3.OpGet("k"), clientv3.OpDelete("k"), ) _, err = txn.Commit() case 3: txn := lkv.Txn(context.TODO()).Then( clientv3.OpGet("k"), clientv3.OpPut("k", v), ) _, err = txn.Commit() case 4: _, err = lkv.Do(context.TODO(), clientv3.OpPut("k", v)) case 5: _, err = lkv.Do(context.TODO(), clientv3.OpDelete("k")) case 6: _, err = lkv.Delete(context.TODO(), "k", clientv3.WithPrefix()) } <-donec if err != nil { // TODO wrap input client to generate errors break } } lresp, lerr := lkv.Get(context.TODO(), "k") if lerr != nil { t.Fatal(lerr) } cresp, cerr := clus.Client(0).Get(context.TODO(), "k") if cerr != nil { t.Fatal(cerr) } if !reflect.DeepEqual(lresp.Kvs, cresp.Kvs) { t.Fatalf("expected %+v, got %+v", cresp, lresp) } } func TestLeasingTxnAtomicCache(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/") testutil.AssertNil(t, err) defer closeLKV() puts, gets := make([]clientv3.Op, 16), make([]clientv3.Op, 16) for i := range puts { k := fmt.Sprintf("k-%d", i) puts[i], gets[i] = clientv3.OpPut(k, k), clientv3.OpGet(k) } if _, err = clus.Client(0).Txn(context.TODO()).Then(puts...).Commit(); err != nil { t.Fatal(err) } for i := range gets { if _, err = lkv.Do(context.TODO(), gets[i]); err != nil { t.Fatal(err) } } numPutters, numGetters := 16, 16 var wgPutters, wgGetters sync.WaitGroup wgPutters.Add(numPutters) wgGetters.Add(numGetters) f := func() { defer wgPutters.Done() for i := 0; i < 10; i++ { if _, txnerr := lkv.Txn(context.TODO()).Then(puts...).Commit(); err != nil { t.Fatal(txnerr) } } } donec := make(chan struct{}, numPutters) g := func() { defer wgGetters.Done() for { select { case <-donec: return default: } tresp, err := lkv.Txn(context.TODO()).Then(gets...).Commit() if err != nil { t.Error(err) } revs := make([]int64, len(gets)) for i, resp := range tresp.Responses { rr := resp.GetResponseRange() revs[i] = rr.Kvs[0].ModRevision } for i := 1; i < len(revs); i++ { if revs[i] != revs[i-1] { t.Errorf("expected matching revisions, got %+v", revs) } } } } for i := 0; i < numGetters; i++ { go g() } for i := 0; i < numPutters; i++ { go f() } wgPutters.Wait() close(donec) wgGetters.Wait() } // TestLeasingReconnectTxn checks that Txn is resilient to disconnects. func TestLeasingReconnectTxn(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/") testutil.AssertNil(t, err) defer closeLKV() if _, err = lkv.Get(context.TODO(), "k"); err != nil { t.Fatal(err) } donec := make(chan struct{}) go func() { defer close(donec) clus.Members[0].DropConnections() for i := 0; i < 10; i++ { clus.Members[0].DropConnections() time.Sleep(time.Millisecond) } time.Sleep(10 * time.Millisecond) }() _, lerr := lkv.Txn(context.TODO()). If(clientv3.Compare(clientv3.Version("k"), "=", 0)). Then(clientv3.OpGet("k")). Commit() <-donec if lerr != nil { t.Fatal(lerr) } } // TestLeasingReconnectNonOwnerGet checks a get error on an owner will // not cause inconsistency between the server and the client. func TestLeasingReconnectNonOwnerGet(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/") testutil.AssertNil(t, err) defer closeLKV() // populate a few keys so some leasing gets have keys for i := 0; i < 4; i++ { k := fmt.Sprintf("k-%d", i*2) if _, err = lkv.Put(context.TODO(), k, k[2:]); err != nil { t.Fatal(err) } } n := 0 for i := 0; i < 10; i++ { donec := make(chan struct{}) clus.Members[0].DropConnections() go func() { defer close(donec) for j := 0; j < 10; j++ { clus.Members[0].DropConnections() time.Sleep(time.Millisecond) } }() _, err = lkv.Get(context.TODO(), fmt.Sprintf("k-%d", i)) <-donec n++ if err != nil { break } } for i := 0; i < n; i++ { k := fmt.Sprintf("k-%d", i) lresp, lerr := lkv.Get(context.TODO(), k) if lerr != nil { t.Fatal(lerr) } cresp, cerr := clus.Client(0).Get(context.TODO(), k) if cerr != nil { t.Fatal(cerr) } if !reflect.DeepEqual(lresp.Kvs, cresp.Kvs) { t.Fatalf("expected %+v, got %+v", cresp, lresp) } } } func TestLeasingTxnRangeCmp(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/") testutil.AssertNil(t, err) defer closeLKV() if _, err = clus.Client(0).Put(context.TODO(), "k", "a"); err != nil { t.Fatal(err) } // k2 version = 2 if _, err = clus.Client(0).Put(context.TODO(), "k2", "a"); err != nil { t.Fatal(err) } if _, err = clus.Client(0).Put(context.TODO(), "k2", "a"); err != nil { t.Fatal(err) } // cache k if _, err = lkv.Get(context.TODO(), "k"); err != nil { t.Fatal(err) } cmp := clientv3.Compare(clientv3.Version("k").WithPrefix(), "=", 1) tresp, terr := lkv.Txn(context.TODO()).If(cmp).Commit() if terr != nil { t.Fatal(err) } if tresp.Succeeded { t.Fatalf("expected Succeeded=false, got %+v", tresp) } } func TestLeasingDo(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/") testutil.AssertNil(t, err) defer closeLKV() ops := []clientv3.Op{ clientv3.OpTxn(nil, nil, nil), clientv3.OpGet("a"), clientv3.OpPut("a/abc", "v"), clientv3.OpDelete("a", clientv3.WithPrefix()), clientv3.OpTxn(nil, nil, nil), } for i, op := range ops { resp, resperr := lkv.Do(context.TODO(), op) if resperr != nil { t.Errorf("#%d: failed (%v)", i, resperr) } switch { case op.IsGet() && resp.Get() == nil: t.Errorf("#%d: get but nil get response", i) case op.IsPut() && resp.Put() == nil: t.Errorf("#%d: put op but nil get response", i) case op.IsDelete() && resp.Del() == nil: t.Errorf("#%d: delete op but nil delete response", i) case op.IsTxn() && resp.Txn() == nil: t.Errorf("#%d: txn op but nil txn response", i) } } gresp, err := clus.Client(0).Get(context.TODO(), "a", clientv3.WithPrefix()) if err != nil { t.Fatal(err) } if len(gresp.Kvs) != 0 { t.Fatalf("expected no keys, got %+v", gresp.Kvs) } } func TestLeasingTxnOwnerPutBranch(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/") testutil.AssertNil(t, err) defer closeLKV() n := 0 treeOp := makePutTreeOp("tree", &n, 4) for i := 0; i < n; i++ { k := fmt.Sprintf("tree/%d", i) if _, err = clus.Client(0).Put(context.TODO(), k, "a"); err != nil { t.Fatal(err) } if _, err = lkv.Get(context.TODO(), k); err != nil { t.Fatal(err) } } if _, err = lkv.Do(context.TODO(), treeOp); err != nil { t.Fatal(err) } // lkv shouldn't need to call out to server for updated leased keys clus.Members[0].Stop(t) for i := 0; i < n; i++ { k := fmt.Sprintf("tree/%d", i) lkvResp, err := lkv.Get(context.TODO(), k) if err != nil { t.Fatal(err) } clusResp, err := clus.Client(1).Get(context.TODO(), k) if err != nil { t.Fatal(err) } if !reflect.DeepEqual(clusResp.Kvs, lkvResp.Kvs) { t.Fatalf("expected %+v, got %+v", clusResp.Kvs, lkvResp.Kvs) } } } func makePutTreeOp(pfx string, v *int, depth int) clientv3.Op { key := fmt.Sprintf("%s/%d", pfx, *v) *v = *v + 1 if depth == 0 { return clientv3.OpPut(key, "leaf") } t, e := makePutTreeOp(pfx, v, depth-1), makePutTreeOp(pfx, v, depth-1) tPut, ePut := clientv3.OpPut(key, "then"), clientv3.OpPut(key, "else") cmps := make([]clientv3.Cmp, 1) if rand.Intn(2) == 0 { // follow then path cmps[0] = clientv3.Compare(clientv3.Version("nokey"), "=", 0) } else { // follow else path cmps[0] = clientv3.Compare(clientv3.Version("nokey"), ">", 0) } return clientv3.OpTxn(cmps, []clientv3.Op{t, tPut}, []clientv3.Op{e, ePut}) } func randCmps(pfx string, dat []*clientv3.PutResponse) (cmps []clientv3.Cmp, then bool) { for i := 0; i < len(dat); i++ { idx := rand.Intn(len(dat)) k := fmt.Sprintf("%s%d", pfx, idx) rev := dat[idx].Header.Revision var cmp clientv3.Cmp switch rand.Intn(4) { case 0: cmp = clientv3.Compare(clientv3.CreateRevision(k), ">", rev-1) case 1: cmp = clientv3.Compare(clientv3.Version(k), "=", 1) case 2: cmp = clientv3.Compare(clientv3.CreateRevision(k), "=", rev) case 3: cmp = clientv3.Compare(clientv3.CreateRevision(k), "!=", rev+1) } cmps = append(cmps, cmp) } cmps = cmps[:rand.Intn(len(dat))] if rand.Intn(2) == 0 { return cmps, true } i := rand.Intn(len(dat)) cmps = append(cmps, clientv3.Compare(clientv3.Version(fmt.Sprintf("k-%d", i)), "=", 0)) return cmps, false } func TestLeasingSessionExpire(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/", concurrency.WithTTL(1)) testutil.AssertNil(t, err) defer closeLKV() lkv2, closeLKV2, err := leasing.NewKV(clus.Client(0), "foo/") testutil.AssertNil(t, err) defer closeLKV2() // acquire lease on abc if _, err = lkv.Get(context.TODO(), "abc"); err != nil { t.Fatal(err) } // down endpoint lkv uses for keepalives clus.Members[0].Stop(t) if err = waitForLeasingExpire(clus.Client(1), "foo/abc"); err != nil { t.Fatal(err) } waitForExpireAck(t, lkv) clus.Members[0].Restart(t) integration.WaitClientV3(t, lkv2) if _, err = lkv2.Put(context.TODO(), "abc", "def"); err != nil { t.Fatal(err) } resp, err := lkv.Get(context.TODO(), "abc") if err != nil { t.Fatal(err) } if v := string(resp.Kvs[0].Value); v != "def" { t.Fatalf("expected %q, got %q", "v", v) } } func TestLeasingSessionExpireCancel(t *testing.T) { tests := []func(context.Context, clientv3.KV) error{ func(ctx context.Context, kv clientv3.KV) error { _, err := kv.Get(ctx, "abc") return err }, func(ctx context.Context, kv clientv3.KV) error { _, err := kv.Delete(ctx, "abc") return err }, func(ctx context.Context, kv clientv3.KV) error { _, err := kv.Put(ctx, "abc", "v") return err }, func(ctx context.Context, kv clientv3.KV) error { _, err := kv.Txn(ctx).Then(clientv3.OpGet("abc")).Commit() return err }, func(ctx context.Context, kv clientv3.KV) error { _, err := kv.Do(ctx, clientv3.OpPut("abc", "v")) return err }, func(ctx context.Context, kv clientv3.KV) error { _, err := kv.Do(ctx, clientv3.OpDelete("abc")) return err }, func(ctx context.Context, kv clientv3.KV) error { _, err := kv.Do(ctx, clientv3.OpGet("abc")) return err }, func(ctx context.Context, kv clientv3.KV) error { op := clientv3.OpTxn(nil, []clientv3.Op{clientv3.OpGet("abc")}, nil) _, err := kv.Do(ctx, op) return err }, } for i := range tests { t.Run(fmt.Sprintf("test %d", i), func(t *testing.T) { defer testutil.AfterTest(t) clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/", concurrency.WithTTL(1)) testutil.AssertNil(t, err) defer closeLKV() if _, err = lkv.Get(context.TODO(), "abc"); err != nil { t.Fatal(err) } // down endpoint lkv uses for keepalives clus.Members[0].Stop(t) if err := waitForLeasingExpire(clus.Client(1), "foo/abc"); err != nil { t.Fatal(err) } waitForExpireAck(t, lkv) ctx, cancel := context.WithCancel(context.TODO()) errc := make(chan error, 1) go func() { errc <- tests[i](ctx, lkv) }() // some delay to get past for ctx.Err() != nil {} loops time.Sleep(100 * time.Millisecond) cancel() select { case err := <-errc: if err != ctx.Err() { t.Errorf("#%d: expected %v of server unavailable, got %v", i, ctx.Err(), err) } case <-time.After(5 * time.Second): t.Errorf("#%d: timed out waiting for cancel", i) } clus.Members[0].Restart(t) }) } } func waitForLeasingExpire(kv clientv3.KV, lkey string) error { for { time.Sleep(1 * time.Second) resp, err := kv.Get(context.TODO(), lkey, clientv3.WithPrefix()) if err != nil { return err } if len(resp.Kvs) == 0 { // server expired the leasing key return nil } } } func waitForExpireAck(t *testing.T, kv clientv3.KV) { // wait for leasing client to acknowledge lost lease for i := 0; i < 10; i++ { ctx, cancel := context.WithTimeout(context.TODO(), time.Second) _, err := kv.Get(ctx, "abc") cancel() if err == ctx.Err() { return } else if err != nil { t.Logf("current error: %v", err) } time.Sleep(time.Second) } t.Fatalf("waited too long to acknlowedge lease expiration") }