| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181 |
- // Copyright 2017 The etcd Authors
- //
- // Licensed under the Apache License, Version 2.0 (the "License");
- // you may not use this file except in compliance with the License.
- // You may obtain a copy of the License at
- //
- // http://www.apache.org/licenses/LICENSE-2.0
- //
- // Unless required by applicable law or agreed to in writing, software
- // distributed under the License is distributed on an "AS IS" BASIS,
- // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- // See the License for the specific language governing permissions and
- // limitations under the License.
- package integration
- import (
- "bytes"
- "context"
- "testing"
- "time"
- "github.com/coreos/etcd/clientv3"
- "github.com/coreos/etcd/integration"
- "github.com/coreos/etcd/pkg/testutil"
- )
- // TestBalancerUnderServerShutdownWatch expects that watch client
- // switch its endpoints when the member of the pinned endpoint fails.
- func TestBalancerUnderServerShutdownWatch(t *testing.T) {
- defer testutil.AfterTest(t)
- clus := integration.NewClusterV3(t, &integration.ClusterConfig{
- Size: 3,
- SkipCreatingClient: true,
- })
- defer clus.Terminate(t)
- eps := []string{clus.Members[0].GRPCAddr(), clus.Members[1].GRPCAddr(), clus.Members[2].GRPCAddr()}
- lead := clus.WaitLeader(t)
- // pin eps[lead]
- watchCli, err := clientv3.New(clientv3.Config{Endpoints: []string{eps[lead]}})
- if err != nil {
- t.Fatal(err)
- }
- defer watchCli.Close()
- // wait for eps[lead] to be pinned
- waitPinReady(t, watchCli)
- // add all eps to list, so that when the original pined one fails
- // the client can switch to other available eps
- watchCli.SetEndpoints(eps...)
- key, val := "foo", "bar"
- wch := watchCli.Watch(context.Background(), key, clientv3.WithCreatedNotify())
- select {
- case <-wch:
- case <-time.After(3 * time.Second):
- t.Fatal("took too long to create watch")
- }
- donec := make(chan struct{})
- go func() {
- defer close(donec)
- // switch to others when eps[lead] is shut down
- select {
- case ev := <-wch:
- if werr := ev.Err(); werr != nil {
- t.Fatal(werr)
- }
- if len(ev.Events) != 1 {
- t.Fatalf("expected one event, got %+v", ev)
- }
- if !bytes.Equal(ev.Events[0].Kv.Value, []byte(val)) {
- t.Fatalf("expected %q, got %+v", val, ev.Events[0].Kv)
- }
- case <-time.After(7 * time.Second):
- t.Fatal("took too long to receive events")
- }
- }()
- // shut down eps[lead]
- clus.Members[lead].Terminate(t)
- // writes to eps[lead+1]
- putCli, err := clientv3.New(clientv3.Config{Endpoints: []string{eps[(lead+1)%3]}})
- if err != nil {
- t.Fatal(err)
- }
- defer putCli.Close()
- for {
- ctx, cancel := context.WithTimeout(context.Background(), 2*time.Second)
- _, err = putCli.Put(ctx, key, val)
- cancel()
- if err == nil {
- break
- }
- if err == context.DeadlineExceeded {
- continue
- }
- t.Fatal(err)
- }
- select {
- case <-donec:
- case <-time.After(5 * time.Second): // enough time for balancer switch
- t.Fatal("took too long to receive events")
- }
- }
- func TestBalancerUnderServerShutdownPut(t *testing.T) {
- testBalancerUnderServerShutdownMutable(t, func(cli *clientv3.Client, ctx context.Context) error {
- _, err := cli.Put(ctx, "foo", "bar")
- return err
- })
- }
- func TestBalancerUnderServerShutdownDelete(t *testing.T) {
- testBalancerUnderServerShutdownMutable(t, func(cli *clientv3.Client, ctx context.Context) error {
- _, err := cli.Delete(ctx, "foo")
- return err
- })
- }
- func TestBalancerUnderServerShutdownTxn(t *testing.T) {
- testBalancerUnderServerShutdownMutable(t, func(cli *clientv3.Client, ctx context.Context) error {
- _, err := cli.Txn(ctx).
- If(clientv3.Compare(clientv3.Version("foo"), "=", 0)).
- Then(clientv3.OpPut("foo", "bar")).
- Else(clientv3.OpPut("foo", "baz")).Commit()
- return err
- })
- }
- // testBalancerUnderServerShutdownMutable expects that when the member of
- // the pinned endpoint is shut down, the balancer switches its endpoints
- // and all subsequent put/delete/txn requests succeed with new endpoints.
- func testBalancerUnderServerShutdownMutable(t *testing.T, op func(*clientv3.Client, context.Context) error) {
- defer testutil.AfterTest(t)
- clus := integration.NewClusterV3(t, &integration.ClusterConfig{
- Size: 3,
- SkipCreatingClient: true,
- })
- defer clus.Terminate(t)
- eps := []string{clus.Members[0].GRPCAddr(), clus.Members[1].GRPCAddr(), clus.Members[2].GRPCAddr()}
- // pin eps[0]
- cli, err := clientv3.New(clientv3.Config{Endpoints: []string{eps[0]}})
- if err != nil {
- t.Fatal(err)
- }
- defer cli.Close()
- // wait for eps[0] to be pinned
- waitPinReady(t, cli)
- // add all eps to list, so that when the original pined one fails
- // the client can switch to other available eps
- cli.SetEndpoints(eps...)
- // shut down eps[0]
- clus.Members[0].Terminate(t)
- // switched to others when eps[0] was explicitly shut down
- // and following request should succeed
- // TODO: remove this (expose client connection state?)
- time.Sleep(time.Second)
- cctx, ccancel := context.WithTimeout(context.Background(), time.Second)
- err = op(cli, cctx)
- ccancel()
- if err != nil {
- t.Fatal(err)
- }
- }
|