12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106 |
- // Copyright 2016 The etcd Authors
- //
- // Licensed under the Apache License, Version 2.0 (the "License");
- // you may not use this file except in compliance with the License.
- // You may obtain a copy of the License at
- //
- // http://www.apache.org/licenses/LICENSE-2.0
- //
- // Unless required by applicable law or agreed to in writing, software
- // distributed under the License is distributed on an "AS IS" BASIS,
- // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- // See the License for the specific language governing permissions and
- // limitations under the License.
- package integration
- import (
- "context"
- "crypto/tls"
- "fmt"
- "io/ioutil"
- "math/rand"
- "net"
- "net/http"
- "net/http/httptest"
- "os"
- "reflect"
- "sort"
- "strings"
- "sync"
- "sync/atomic"
- "testing"
- "time"
- "github.com/coreos/etcd/client"
- "github.com/coreos/etcd/clientv3"
- "github.com/coreos/etcd/embed"
- "github.com/coreos/etcd/etcdserver"
- "github.com/coreos/etcd/etcdserver/api/etcdhttp"
- "github.com/coreos/etcd/etcdserver/api/v2http"
- "github.com/coreos/etcd/etcdserver/api/v3client"
- "github.com/coreos/etcd/etcdserver/api/v3election"
- epb "github.com/coreos/etcd/etcdserver/api/v3election/v3electionpb"
- "github.com/coreos/etcd/etcdserver/api/v3lock"
- lockpb "github.com/coreos/etcd/etcdserver/api/v3lock/v3lockpb"
- "github.com/coreos/etcd/etcdserver/api/v3rpc"
- pb "github.com/coreos/etcd/etcdserver/etcdserverpb"
- "github.com/coreos/etcd/pkg/testutil"
- "github.com/coreos/etcd/pkg/tlsutil"
- "github.com/coreos/etcd/pkg/transport"
- "github.com/coreos/etcd/pkg/types"
- "github.com/coreos/etcd/rafthttp"
- "github.com/coreos/pkg/capnslog"
- "github.com/soheilhy/cmux"
- "google.golang.org/grpc"
- "google.golang.org/grpc/grpclog"
- "google.golang.org/grpc/keepalive"
- )
- const (
- // RequestWaitTimeout is the time duration to wait for a request to go through or detect leader loss.
- RequestWaitTimeout = 3 * time.Second
- tickDuration = 10 * time.Millisecond
- requestTimeout = 20 * time.Second
- clusterName = "etcd"
- basePort = 21000
- UrlScheme = "unix"
- UrlSchemeTLS = "unixs"
- )
- var (
- electionTicks = 10
- // integration test uses unique ports, counting up, to listen for each
- // member, ensuring restarted members can listen on the same port again.
- localListenCount int64 = 0
- testTLSInfo = transport.TLSInfo{
- KeyFile: "./fixtures/server.key.insecure",
- CertFile: "./fixtures/server.crt",
- TrustedCAFile: "./fixtures/ca.crt",
- ClientCertAuth: true,
- }
- testTLSInfoExpired = transport.TLSInfo{
- KeyFile: "./fixtures-expired/server-key.pem",
- CertFile: "./fixtures-expired/server.pem",
- TrustedCAFile: "./fixtures-expired/etcd-root-ca.pem",
- ClientCertAuth: true,
- }
- plog = capnslog.NewPackageLogger("github.com/coreos/etcd", "integration")
- )
- type ClusterConfig struct {
- Size int
- PeerTLS *transport.TLSInfo
- ClientTLS *transport.TLSInfo
- DiscoveryURL string
- UseGRPC bool
- QuotaBackendBytes int64
- MaxTxnOps uint
- MaxRequestBytes uint
- GRPCKeepAliveMinTime time.Duration
- GRPCKeepAliveInterval time.Duration
- GRPCKeepAliveTimeout time.Duration
- // SkipCreatingClient to skip creating clients for each member.
- SkipCreatingClient bool
- ClientMaxCallSendMsgSize int
- ClientMaxCallRecvMsgSize int
- }
- type cluster struct {
- cfg *ClusterConfig
- Members []*member
- }
- func schemeFromTLSInfo(tls *transport.TLSInfo) string {
- if tls == nil {
- return UrlScheme
- }
- return UrlSchemeTLS
- }
- func (c *cluster) fillClusterForMembers() error {
- if c.cfg.DiscoveryURL != "" {
- // cluster will be discovered
- return nil
- }
- addrs := make([]string, 0)
- for _, m := range c.Members {
- scheme := schemeFromTLSInfo(m.PeerTLSInfo)
- for _, l := range m.PeerListeners {
- addrs = append(addrs, fmt.Sprintf("%s=%s://%s", m.Name, scheme, l.Addr().String()))
- }
- }
- clusterStr := strings.Join(addrs, ",")
- var err error
- for _, m := range c.Members {
- m.InitialPeerURLsMap, err = types.NewURLsMap(clusterStr)
- if err != nil {
- return err
- }
- }
- return nil
- }
- func newCluster(t *testing.T, cfg *ClusterConfig) *cluster {
- c := &cluster{cfg: cfg}
- ms := make([]*member, cfg.Size)
- for i := 0; i < cfg.Size; i++ {
- ms[i] = c.mustNewMember(t)
- }
- c.Members = ms
- if err := c.fillClusterForMembers(); err != nil {
- t.Fatal(err)
- }
- return c
- }
- // NewCluster returns an unlaunched cluster of the given size which has been
- // set to use static bootstrap.
- func NewCluster(t *testing.T, size int) *cluster {
- return newCluster(t, &ClusterConfig{Size: size})
- }
- // NewClusterByConfig returns an unlaunched cluster defined by a cluster configuration
- func NewClusterByConfig(t *testing.T, cfg *ClusterConfig) *cluster {
- return newCluster(t, cfg)
- }
- func (c *cluster) Launch(t *testing.T) {
- errc := make(chan error)
- for _, m := range c.Members {
- // Members are launched in separate goroutines because if they boot
- // using discovery url, they have to wait for others to register to continue.
- go func(m *member) {
- errc <- m.Launch()
- }(m)
- }
- for range c.Members {
- if err := <-errc; err != nil {
- t.Fatalf("error setting up member: %v", err)
- }
- }
- // wait cluster to be stable to receive future client requests
- c.waitMembersMatch(t, c.HTTPMembers())
- c.waitVersion()
- }
- func (c *cluster) URL(i int) string {
- return c.Members[i].ClientURLs[0].String()
- }
- // URLs returns a list of all active client URLs in the cluster
- func (c *cluster) URLs() []string {
- return getMembersURLs(c.Members)
- }
- func getMembersURLs(members []*member) []string {
- urls := make([]string, 0)
- for _, m := range members {
- select {
- case <-m.s.StopNotify():
- continue
- default:
- }
- for _, u := range m.ClientURLs {
- urls = append(urls, u.String())
- }
- }
- return urls
- }
- // HTTPMembers returns a list of all active members as client.Members
- func (c *cluster) HTTPMembers() []client.Member {
- ms := []client.Member{}
- for _, m := range c.Members {
- pScheme := schemeFromTLSInfo(m.PeerTLSInfo)
- cScheme := schemeFromTLSInfo(m.ClientTLSInfo)
- cm := client.Member{Name: m.Name}
- for _, ln := range m.PeerListeners {
- cm.PeerURLs = append(cm.PeerURLs, pScheme+"://"+ln.Addr().String())
- }
- for _, ln := range m.ClientListeners {
- cm.ClientURLs = append(cm.ClientURLs, cScheme+"://"+ln.Addr().String())
- }
- ms = append(ms, cm)
- }
- return ms
- }
- func (c *cluster) mustNewMember(t *testing.T) *member {
- m := mustNewMember(t,
- memberConfig{
- name: c.name(rand.Int()),
- peerTLS: c.cfg.PeerTLS,
- clientTLS: c.cfg.ClientTLS,
- quotaBackendBytes: c.cfg.QuotaBackendBytes,
- maxTxnOps: c.cfg.MaxTxnOps,
- maxRequestBytes: c.cfg.MaxRequestBytes,
- grpcKeepAliveMinTime: c.cfg.GRPCKeepAliveMinTime,
- grpcKeepAliveInterval: c.cfg.GRPCKeepAliveInterval,
- grpcKeepAliveTimeout: c.cfg.GRPCKeepAliveTimeout,
- clientMaxCallSendMsgSize: c.cfg.ClientMaxCallSendMsgSize,
- clientMaxCallRecvMsgSize: c.cfg.ClientMaxCallRecvMsgSize,
- })
- m.DiscoveryURL = c.cfg.DiscoveryURL
- if c.cfg.UseGRPC {
- if err := m.listenGRPC(); err != nil {
- t.Fatal(err)
- }
- }
- return m
- }
- func (c *cluster) addMember(t *testing.T) {
- m := c.mustNewMember(t)
- scheme := schemeFromTLSInfo(c.cfg.PeerTLS)
- // send add request to the cluster
- var err error
- for i := 0; i < len(c.Members); i++ {
- clientURL := c.URL(i)
- peerURL := scheme + "://" + m.PeerListeners[0].Addr().String()
- if err = c.addMemberByURL(t, clientURL, peerURL); err == nil {
- break
- }
- }
- if err != nil {
- t.Fatalf("add member failed on all members error: %v", err)
- }
- m.InitialPeerURLsMap = types.URLsMap{}
- for _, mm := range c.Members {
- m.InitialPeerURLsMap[mm.Name] = mm.PeerURLs
- }
- m.InitialPeerURLsMap[m.Name] = m.PeerURLs
- m.NewCluster = false
- if err := m.Launch(); err != nil {
- t.Fatal(err)
- }
- c.Members = append(c.Members, m)
- // wait cluster to be stable to receive future client requests
- c.waitMembersMatch(t, c.HTTPMembers())
- }
- func (c *cluster) addMemberByURL(t *testing.T, clientURL, peerURL string) error {
- cc := MustNewHTTPClient(t, []string{clientURL}, c.cfg.ClientTLS)
- ma := client.NewMembersAPI(cc)
- ctx, cancel := context.WithTimeout(context.Background(), requestTimeout)
- _, err := ma.Add(ctx, peerURL)
- cancel()
- if err != nil {
- return err
- }
- // wait for the add node entry applied in the cluster
- members := append(c.HTTPMembers(), client.Member{PeerURLs: []string{peerURL}, ClientURLs: []string{}})
- c.waitMembersMatch(t, members)
- return nil
- }
- func (c *cluster) AddMember(t *testing.T) {
- c.addMember(t)
- }
- func (c *cluster) RemoveMember(t *testing.T, id uint64) {
- if err := c.removeMember(t, id); err != nil {
- t.Fatal(err)
- }
- }
- func (c *cluster) removeMember(t *testing.T, id uint64) error {
- // send remove request to the cluster
- cc := MustNewHTTPClient(t, c.URLs(), c.cfg.ClientTLS)
- ma := client.NewMembersAPI(cc)
- ctx, cancel := context.WithTimeout(context.Background(), requestTimeout)
- err := ma.Remove(ctx, types.ID(id).String())
- cancel()
- if err != nil {
- return err
- }
- newMembers := make([]*member, 0)
- for _, m := range c.Members {
- if uint64(m.s.ID()) != id {
- newMembers = append(newMembers, m)
- } else {
- select {
- case <-m.s.StopNotify():
- m.Terminate(t)
- // 1s stop delay + election timeout + 1s disk and network delay + connection write timeout
- // TODO: remove connection write timeout by selecting on http response closeNotifier
- // blocking on https://github.com/golang/go/issues/9524
- case <-time.After(time.Second + time.Duration(electionTicks)*tickDuration + time.Second + rafthttp.ConnWriteTimeout):
- t.Fatalf("failed to remove member %s in time", m.s.ID())
- }
- }
- }
- c.Members = newMembers
- c.waitMembersMatch(t, c.HTTPMembers())
- return nil
- }
- func (c *cluster) Terminate(t *testing.T) {
- var wg sync.WaitGroup
- wg.Add(len(c.Members))
- for _, m := range c.Members {
- go func(mm *member) {
- defer wg.Done()
- mm.Terminate(t)
- }(m)
- }
- wg.Wait()
- }
- func (c *cluster) waitMembersMatch(t *testing.T, membs []client.Member) {
- for _, u := range c.URLs() {
- cc := MustNewHTTPClient(t, []string{u}, c.cfg.ClientTLS)
- ma := client.NewMembersAPI(cc)
- for {
- ctx, cancel := context.WithTimeout(context.Background(), requestTimeout)
- ms, err := ma.List(ctx)
- cancel()
- if err == nil && isMembersEqual(ms, membs) {
- break
- }
- time.Sleep(tickDuration)
- }
- }
- }
- func (c *cluster) WaitLeader(t *testing.T) int { return c.waitLeader(t, c.Members) }
- // waitLeader waits until given members agree on the same leader.
- func (c *cluster) waitLeader(t *testing.T, membs []*member) int {
- possibleLead := make(map[uint64]bool)
- var lead uint64
- for _, m := range membs {
- possibleLead[uint64(m.s.ID())] = true
- }
- cc := MustNewHTTPClient(t, getMembersURLs(membs), nil)
- kapi := client.NewKeysAPI(cc)
- // ensure leader is up via linearizable get
- for {
- ctx, cancel := context.WithTimeout(context.Background(), 10*tickDuration+time.Second)
- _, err := kapi.Get(ctx, "0", &client.GetOptions{Quorum: true})
- cancel()
- if err == nil || strings.Contains(err.Error(), "Key not found") {
- break
- }
- }
- for lead == 0 || !possibleLead[lead] {
- lead = 0
- for _, m := range membs {
- select {
- case <-m.s.StopNotify():
- continue
- default:
- }
- if lead != 0 && lead != m.s.Lead() {
- lead = 0
- time.Sleep(10 * tickDuration)
- break
- }
- lead = m.s.Lead()
- }
- }
- for i, m := range membs {
- if uint64(m.s.ID()) == lead {
- return i
- }
- }
- return -1
- }
- func (c *cluster) WaitNoLeader(t *testing.T) { c.waitNoLeader(t, c.Members) }
- // waitNoLeader waits until given members lose leader.
- func (c *cluster) waitNoLeader(t *testing.T, membs []*member) {
- noLeader := false
- for !noLeader {
- noLeader = true
- for _, m := range membs {
- select {
- case <-m.s.StopNotify():
- continue
- default:
- }
- if m.s.Lead() != 0 {
- noLeader = false
- time.Sleep(10 * tickDuration)
- break
- }
- }
- }
- }
- func (c *cluster) waitVersion() {
- for _, m := range c.Members {
- for {
- if m.s.ClusterVersion() != nil {
- break
- }
- time.Sleep(tickDuration)
- }
- }
- }
- func (c *cluster) name(i int) string {
- return fmt.Sprint(i)
- }
- // isMembersEqual checks whether two members equal except ID field.
- // The given wmembs should always set ID field to empty string.
- func isMembersEqual(membs []client.Member, wmembs []client.Member) bool {
- sort.Sort(SortableMemberSliceByPeerURLs(membs))
- sort.Sort(SortableMemberSliceByPeerURLs(wmembs))
- for i := range membs {
- membs[i].ID = ""
- }
- return reflect.DeepEqual(membs, wmembs)
- }
- func newLocalListener(t *testing.T) net.Listener {
- c := atomic.AddInt64(&localListenCount, 1)
- // Go 1.8+ allows only numbers in port
- addr := fmt.Sprintf("127.0.0.1:%05d%05d", c+basePort, os.Getpid())
- return NewListenerWithAddr(t, addr)
- }
- func NewListenerWithAddr(t *testing.T, addr string) net.Listener {
- l, err := transport.NewUnixListener(addr)
- if err != nil {
- t.Fatal(err)
- }
- return l
- }
- type member struct {
- etcdserver.ServerConfig
- PeerListeners, ClientListeners []net.Listener
- grpcListener net.Listener
- // PeerTLSInfo enables peer TLS when set
- PeerTLSInfo *transport.TLSInfo
- // ClientTLSInfo enables client TLS when set
- ClientTLSInfo *transport.TLSInfo
- raftHandler *testutil.PauseableHandler
- s *etcdserver.EtcdServer
- serverClosers []func()
- grpcServerOpts []grpc.ServerOption
- grpcServer *grpc.Server
- grpcServerPeer *grpc.Server
- grpcAddr string
- grpcBridge *bridge
- // serverClient is a clientv3 that directly calls the etcdserver.
- serverClient *clientv3.Client
- keepDataDirTerminate bool
- clientMaxCallSendMsgSize int
- clientMaxCallRecvMsgSize int
- }
- func (m *member) GRPCAddr() string { return m.grpcAddr }
- type memberConfig struct {
- name string
- peerTLS *transport.TLSInfo
- clientTLS *transport.TLSInfo
- quotaBackendBytes int64
- maxTxnOps uint
- maxRequestBytes uint
- grpcKeepAliveMinTime time.Duration
- grpcKeepAliveInterval time.Duration
- grpcKeepAliveTimeout time.Duration
- clientMaxCallSendMsgSize int
- clientMaxCallRecvMsgSize int
- }
- // mustNewMember return an inited member with the given name. If peerTLS is
- // set, it will use https scheme to communicate between peers.
- func mustNewMember(t *testing.T, mcfg memberConfig) *member {
- var err error
- m := &member{}
- peerScheme := schemeFromTLSInfo(mcfg.peerTLS)
- clientScheme := schemeFromTLSInfo(mcfg.clientTLS)
- pln := newLocalListener(t)
- m.PeerListeners = []net.Listener{pln}
- m.PeerURLs, err = types.NewURLs([]string{peerScheme + "://" + pln.Addr().String()})
- if err != nil {
- t.Fatal(err)
- }
- m.PeerTLSInfo = mcfg.peerTLS
- cln := newLocalListener(t)
- m.ClientListeners = []net.Listener{cln}
- m.ClientURLs, err = types.NewURLs([]string{clientScheme + "://" + cln.Addr().String()})
- if err != nil {
- t.Fatal(err)
- }
- m.ClientTLSInfo = mcfg.clientTLS
- m.Name = mcfg.name
- m.DataDir, err = ioutil.TempDir(os.TempDir(), "etcd")
- if err != nil {
- t.Fatal(err)
- }
- clusterStr := fmt.Sprintf("%s=%s://%s", mcfg.name, peerScheme, pln.Addr().String())
- m.InitialPeerURLsMap, err = types.NewURLsMap(clusterStr)
- if err != nil {
- t.Fatal(err)
- }
- m.InitialClusterToken = clusterName
- m.NewCluster = true
- m.BootstrapTimeout = 10 * time.Millisecond
- if m.PeerTLSInfo != nil {
- m.ServerConfig.PeerTLSInfo = *m.PeerTLSInfo
- }
- m.ElectionTicks = electionTicks
- m.InitialElectionTickAdvance = true
- m.TickMs = uint(tickDuration / time.Millisecond)
- m.QuotaBackendBytes = mcfg.quotaBackendBytes
- m.MaxTxnOps = mcfg.maxTxnOps
- if m.MaxTxnOps == 0 {
- m.MaxTxnOps = embed.DefaultMaxTxnOps
- }
- m.MaxRequestBytes = mcfg.maxRequestBytes
- if m.MaxRequestBytes == 0 {
- m.MaxRequestBytes = embed.DefaultMaxRequestBytes
- }
- m.AuthToken = "simple" // for the purpose of integration testing, simple token is enough
- m.grpcServerOpts = []grpc.ServerOption{}
- if mcfg.grpcKeepAliveMinTime > time.Duration(0) {
- m.grpcServerOpts = append(m.grpcServerOpts, grpc.KeepaliveEnforcementPolicy(keepalive.EnforcementPolicy{
- MinTime: mcfg.grpcKeepAliveMinTime,
- PermitWithoutStream: false,
- }))
- }
- if mcfg.grpcKeepAliveInterval > time.Duration(0) &&
- mcfg.grpcKeepAliveTimeout > time.Duration(0) {
- m.grpcServerOpts = append(m.grpcServerOpts, grpc.KeepaliveParams(keepalive.ServerParameters{
- Time: mcfg.grpcKeepAliveInterval,
- Timeout: mcfg.grpcKeepAliveTimeout,
- }))
- }
- m.clientMaxCallSendMsgSize = mcfg.clientMaxCallSendMsgSize
- m.clientMaxCallRecvMsgSize = mcfg.clientMaxCallRecvMsgSize
- m.InitialCorruptCheck = true
- return m
- }
- // listenGRPC starts a grpc server over a unix domain socket on the member
- func (m *member) listenGRPC() error {
- // prefix with localhost so cert has right domain
- m.grpcAddr = "localhost:" + m.Name
- l, err := transport.NewUnixListener(m.grpcAddr)
- if err != nil {
- return fmt.Errorf("listen failed on grpc socket %s (%v)", m.grpcAddr, err)
- }
- m.grpcBridge, err = newBridge(m.grpcAddr)
- if err != nil {
- l.Close()
- return err
- }
- m.grpcAddr = schemeFromTLSInfo(m.ClientTLSInfo) + "://" + m.grpcBridge.inaddr
- m.grpcListener = l
- return nil
- }
- func (m *member) ElectionTimeout() time.Duration {
- return time.Duration(m.s.Cfg.ElectionTicks*int(m.s.Cfg.TickMs)) * time.Millisecond
- }
- func (m *member) ID() types.ID { return m.s.ID() }
- func (m *member) DropConnections() { m.grpcBridge.Reset() }
- func (m *member) PauseConnections() { m.grpcBridge.Pause() }
- func (m *member) UnpauseConnections() { m.grpcBridge.Unpause() }
- func (m *member) Blackhole() { m.grpcBridge.Blackhole() }
- func (m *member) Unblackhole() { m.grpcBridge.Unblackhole() }
- // NewClientV3 creates a new grpc client connection to the member
- func NewClientV3(m *member) (*clientv3.Client, error) {
- if m.grpcAddr == "" {
- return nil, fmt.Errorf("member not configured for grpc")
- }
- cfg := clientv3.Config{
- Endpoints: []string{m.grpcAddr},
- DialTimeout: 5 * time.Second,
- DialOptions: []grpc.DialOption{grpc.WithBlock()},
- MaxCallSendMsgSize: m.clientMaxCallSendMsgSize,
- MaxCallRecvMsgSize: m.clientMaxCallRecvMsgSize,
- }
- if m.ClientTLSInfo != nil {
- tls, err := m.ClientTLSInfo.ClientConfig()
- if err != nil {
- return nil, err
- }
- cfg.TLS = tls
- }
- return newClientV3(cfg)
- }
- // Clone returns a member with the same server configuration. The returned
- // member will not set PeerListeners and ClientListeners.
- func (m *member) Clone(t *testing.T) *member {
- mm := &member{}
- mm.ServerConfig = m.ServerConfig
- var err error
- clientURLStrs := m.ClientURLs.StringSlice()
- mm.ClientURLs, err = types.NewURLs(clientURLStrs)
- if err != nil {
- // this should never fail
- panic(err)
- }
- peerURLStrs := m.PeerURLs.StringSlice()
- mm.PeerURLs, err = types.NewURLs(peerURLStrs)
- if err != nil {
- // this should never fail
- panic(err)
- }
- clusterStr := m.InitialPeerURLsMap.String()
- mm.InitialPeerURLsMap, err = types.NewURLsMap(clusterStr)
- if err != nil {
- // this should never fail
- panic(err)
- }
- mm.InitialClusterToken = m.InitialClusterToken
- mm.ElectionTicks = m.ElectionTicks
- mm.PeerTLSInfo = m.PeerTLSInfo
- mm.ClientTLSInfo = m.ClientTLSInfo
- return mm
- }
- // Launch starts a member based on ServerConfig, PeerListeners
- // and ClientListeners.
- func (m *member) Launch() error {
- plog.Printf("launching %s (%s)", m.Name, m.grpcAddr)
- var err error
- if m.s, err = etcdserver.NewServer(m.ServerConfig); err != nil {
- return fmt.Errorf("failed to initialize the etcd server: %v", err)
- }
- m.s.SyncTicker = time.NewTicker(500 * time.Millisecond)
- m.s.Start()
- var peerTLScfg *tls.Config
- if m.PeerTLSInfo != nil && !m.PeerTLSInfo.Empty() {
- if peerTLScfg, err = m.PeerTLSInfo.ServerConfig(); err != nil {
- return err
- }
- }
- if m.grpcListener != nil {
- var (
- tlscfg *tls.Config
- )
- if m.ClientTLSInfo != nil && !m.ClientTLSInfo.Empty() {
- tlscfg, err = m.ClientTLSInfo.ServerConfig()
- if err != nil {
- return err
- }
- }
- m.grpcServer = v3rpc.Server(m.s, tlscfg, m.grpcServerOpts...)
- m.grpcServerPeer = v3rpc.Server(m.s, peerTLScfg)
- m.serverClient = v3client.New(m.s)
- lockpb.RegisterLockServer(m.grpcServer, v3lock.NewLockServer(m.serverClient))
- epb.RegisterElectionServer(m.grpcServer, v3election.NewElectionServer(m.serverClient))
- go m.grpcServer.Serve(m.grpcListener)
- }
- m.raftHandler = &testutil.PauseableHandler{Next: etcdhttp.NewPeerHandler(m.s)}
- h := (http.Handler)(m.raftHandler)
- if m.grpcListener != nil {
- h = http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
- if r.ProtoMajor == 2 && strings.Contains(r.Header.Get("Content-Type"), "application/grpc") {
- m.grpcServerPeer.ServeHTTP(w, r)
- } else {
- m.raftHandler.ServeHTTP(w, r)
- }
- })
- }
- for _, ln := range m.PeerListeners {
- cm := cmux.New(ln)
- // don't hang on matcher after closing listener
- cm.SetReadTimeout(time.Second)
- if m.grpcServer != nil {
- grpcl := cm.Match(cmux.HTTP2())
- go m.grpcServerPeer.Serve(grpcl)
- }
- // serve http1/http2 rafthttp/grpc
- ll := cm.Match(cmux.Any())
- if peerTLScfg != nil {
- if ll, err = transport.NewTLSListener(ll, m.PeerTLSInfo); err != nil {
- return err
- }
- }
- hs := &httptest.Server{
- Listener: ll,
- Config: &http.Server{Handler: h, TLSConfig: peerTLScfg},
- TLS: peerTLScfg,
- }
- hs.Start()
- donec := make(chan struct{})
- go func() {
- defer close(donec)
- cm.Serve()
- }()
- closer := func() {
- ll.Close()
- hs.CloseClientConnections()
- hs.Close()
- <-donec
- }
- m.serverClosers = append(m.serverClosers, closer)
- }
- for _, ln := range m.ClientListeners {
- hs := &httptest.Server{
- Listener: ln,
- Config: &http.Server{Handler: v2http.NewClientHandler(m.s, m.ServerConfig.ReqTimeout())},
- }
- if m.ClientTLSInfo == nil {
- hs.Start()
- } else {
- info := m.ClientTLSInfo
- hs.TLS, err = info.ServerConfig()
- if err != nil {
- return err
- }
- // baseConfig is called on initial TLS handshake start.
- //
- // Previously,
- // 1. Server has non-empty (*tls.Config).Certificates on client hello
- // 2. Server calls (*tls.Config).GetCertificate iff:
- // - Server's (*tls.Config).Certificates is not empty, or
- // - Client supplies SNI; non-empty (*tls.ClientHelloInfo).ServerName
- //
- // When (*tls.Config).Certificates is always populated on initial handshake,
- // client is expected to provide a valid matching SNI to pass the TLS
- // verification, thus trigger server (*tls.Config).GetCertificate to reload
- // TLS assets. However, a cert whose SAN field does not include domain names
- // but only IP addresses, has empty (*tls.ClientHelloInfo).ServerName, thus
- // it was never able to trigger TLS reload on initial handshake; first
- // ceritifcate object was being used, never being updated.
- //
- // Now, (*tls.Config).Certificates is created empty on initial TLS client
- // handshake, in order to trigger (*tls.Config).GetCertificate and populate
- // rest of the certificates on every new TLS connection, even when client
- // SNI is empty (e.g. cert only includes IPs).
- //
- // This introduces another problem with "httptest.Server":
- // when server initial certificates are empty, certificates
- // are overwritten by Go's internal test certs, which have
- // different SAN fields (e.g. example.com). To work around,
- // re-overwrite (*tls.Config).Certificates before starting
- // test server.
- tlsCert, err := tlsutil.NewCert(info.CertFile, info.KeyFile, nil)
- if err != nil {
- return err
- }
- hs.TLS.Certificates = []tls.Certificate{*tlsCert}
- hs.StartTLS()
- }
- closer := func() {
- ln.Close()
- hs.CloseClientConnections()
- hs.Close()
- }
- m.serverClosers = append(m.serverClosers, closer)
- }
- plog.Printf("launched %s (%s)", m.Name, m.grpcAddr)
- return nil
- }
- func (m *member) WaitOK(t *testing.T) {
- cc := MustNewHTTPClient(t, []string{m.URL()}, m.ClientTLSInfo)
- kapi := client.NewKeysAPI(cc)
- for {
- ctx, cancel := context.WithTimeout(context.Background(), requestTimeout)
- _, err := kapi.Get(ctx, "/", nil)
- if err != nil {
- time.Sleep(tickDuration)
- continue
- }
- cancel()
- break
- }
- for m.s.Leader() == 0 {
- time.Sleep(tickDuration)
- }
- }
- func (m *member) URL() string { return m.ClientURLs[0].String() }
- func (m *member) Pause() {
- m.raftHandler.Pause()
- m.s.PauseSending()
- }
- func (m *member) Resume() {
- m.raftHandler.Resume()
- m.s.ResumeSending()
- }
- // Close stops the member's etcdserver and closes its connections
- func (m *member) Close() {
- if m.grpcBridge != nil {
- m.grpcBridge.Close()
- m.grpcBridge = nil
- }
- if m.serverClient != nil {
- m.serverClient.Close()
- m.serverClient = nil
- }
- if m.grpcServer != nil {
- m.grpcServer.Stop()
- m.grpcServer.GracefulStop()
- m.grpcServer = nil
- m.grpcServerPeer.Stop()
- m.grpcServerPeer.GracefulStop()
- m.grpcServerPeer = nil
- }
- m.s.HardStop()
- for _, f := range m.serverClosers {
- f()
- }
- }
- // Stop stops the member, but the data dir of the member is preserved.
- func (m *member) Stop(t *testing.T) {
- plog.Printf("stopping %s (%s)", m.Name, m.grpcAddr)
- m.Close()
- m.serverClosers = nil
- plog.Printf("stopped %s (%s)", m.Name, m.grpcAddr)
- }
- // checkLeaderTransition waits for leader transition, returning the new leader ID.
- func checkLeaderTransition(t *testing.T, m *member, oldLead uint64) uint64 {
- interval := time.Duration(m.s.Cfg.TickMs) * time.Millisecond
- for m.s.Lead() == 0 || (m.s.Lead() == oldLead) {
- time.Sleep(interval)
- }
- return m.s.Lead()
- }
- // StopNotify unblocks when a member stop completes
- func (m *member) StopNotify() <-chan struct{} {
- return m.s.StopNotify()
- }
- // Restart starts the member using the preserved data dir.
- func (m *member) Restart(t *testing.T) error {
- plog.Printf("restarting %s (%s)", m.Name, m.grpcAddr)
- newPeerListeners := make([]net.Listener, 0)
- for _, ln := range m.PeerListeners {
- newPeerListeners = append(newPeerListeners, NewListenerWithAddr(t, ln.Addr().String()))
- }
- m.PeerListeners = newPeerListeners
- newClientListeners := make([]net.Listener, 0)
- for _, ln := range m.ClientListeners {
- newClientListeners = append(newClientListeners, NewListenerWithAddr(t, ln.Addr().String()))
- }
- m.ClientListeners = newClientListeners
- if m.grpcListener != nil {
- if err := m.listenGRPC(); err != nil {
- t.Fatal(err)
- }
- }
- err := m.Launch()
- plog.Printf("restarted %s (%s)", m.Name, m.grpcAddr)
- return err
- }
- // Terminate stops the member and removes the data dir.
- func (m *member) Terminate(t *testing.T) {
- plog.Printf("terminating %s (%s)", m.Name, m.grpcAddr)
- m.Close()
- if !m.keepDataDirTerminate {
- if err := os.RemoveAll(m.ServerConfig.DataDir); err != nil {
- t.Fatal(err)
- }
- }
- plog.Printf("terminated %s (%s)", m.Name, m.grpcAddr)
- }
- // Metric gets the metric value for a member
- func (m *member) Metric(metricName string) (string, error) {
- cfgtls := transport.TLSInfo{}
- tr, err := transport.NewTimeoutTransport(cfgtls, time.Second, time.Second, time.Second)
- if err != nil {
- return "", err
- }
- cli := &http.Client{Transport: tr}
- resp, err := cli.Get(m.ClientURLs[0].String() + "/metrics")
- if err != nil {
- return "", err
- }
- defer resp.Body.Close()
- b, rerr := ioutil.ReadAll(resp.Body)
- if rerr != nil {
- return "", rerr
- }
- lines := strings.Split(string(b), "\n")
- for _, l := range lines {
- if strings.HasPrefix(l, metricName) {
- return strings.Split(l, " ")[1], nil
- }
- }
- return "", nil
- }
- // InjectPartition drops connections from m to others, vice versa.
- func (m *member) InjectPartition(t *testing.T, others ...*member) {
- for _, other := range others {
- m.s.CutPeer(other.s.ID())
- other.s.CutPeer(m.s.ID())
- }
- }
- // RecoverPartition recovers connections from m to others, vice versa.
- func (m *member) RecoverPartition(t *testing.T, others ...*member) {
- for _, other := range others {
- m.s.MendPeer(other.s.ID())
- other.s.MendPeer(m.s.ID())
- }
- }
- func MustNewHTTPClient(t *testing.T, eps []string, tls *transport.TLSInfo) client.Client {
- cfgtls := transport.TLSInfo{}
- if tls != nil {
- cfgtls = *tls
- }
- cfg := client.Config{Transport: mustNewTransport(t, cfgtls), Endpoints: eps}
- c, err := client.New(cfg)
- if err != nil {
- t.Fatal(err)
- }
- return c
- }
- func mustNewTransport(t *testing.T, tlsInfo transport.TLSInfo) *http.Transport {
- // tick in integration test is short, so 1s dial timeout could play well.
- tr, err := transport.NewTimeoutTransport(tlsInfo, time.Second, rafthttp.ConnReadTimeout, rafthttp.ConnWriteTimeout)
- if err != nil {
- t.Fatal(err)
- }
- return tr
- }
- type SortableMemberSliceByPeerURLs []client.Member
- func (p SortableMemberSliceByPeerURLs) Len() int { return len(p) }
- func (p SortableMemberSliceByPeerURLs) Less(i, j int) bool {
- return p[i].PeerURLs[0] < p[j].PeerURLs[0]
- }
- func (p SortableMemberSliceByPeerURLs) Swap(i, j int) { p[i], p[j] = p[j], p[i] }
- type ClusterV3 struct {
- *cluster
- mu sync.Mutex
- clients []*clientv3.Client
- }
- // NewClusterV3 returns a launched cluster with a grpc client connection
- // for each cluster member.
- func NewClusterV3(t *testing.T, cfg *ClusterConfig) *ClusterV3 {
- cfg.UseGRPC = true
- if os.Getenv("CLIENT_DEBUG") != "" {
- clientv3.SetLogger(grpclog.NewLoggerV2WithVerbosity(os.Stderr, os.Stderr, os.Stderr, 4))
- }
- clus := &ClusterV3{
- cluster: NewClusterByConfig(t, cfg),
- }
- clus.Launch(t)
- if !cfg.SkipCreatingClient {
- for _, m := range clus.Members {
- client, err := NewClientV3(m)
- if err != nil {
- t.Fatalf("cannot create client: %v", err)
- }
- clus.clients = append(clus.clients, client)
- }
- }
- return clus
- }
- func (c *ClusterV3) TakeClient(idx int) {
- c.mu.Lock()
- c.clients[idx] = nil
- c.mu.Unlock()
- }
- func (c *ClusterV3) Terminate(t *testing.T) {
- c.mu.Lock()
- for _, client := range c.clients {
- if client == nil {
- continue
- }
- if err := client.Close(); err != nil {
- t.Error(err)
- }
- }
- c.mu.Unlock()
- c.cluster.Terminate(t)
- }
- func (c *ClusterV3) RandClient() *clientv3.Client {
- return c.clients[rand.Intn(len(c.clients))]
- }
- func (c *ClusterV3) Client(i int) *clientv3.Client {
- return c.clients[i]
- }
- type grpcAPI struct {
- // Cluster is the cluster API for the client's connection.
- Cluster pb.ClusterClient
- // KV is the keyvalue API for the client's connection.
- KV pb.KVClient
- // Lease is the lease API for the client's connection.
- Lease pb.LeaseClient
- // Watch is the watch API for the client's connection.
- Watch pb.WatchClient
- // Maintenance is the maintenance API for the client's connection.
- Maintenance pb.MaintenanceClient
- // Auth is the authentication API for the client's connection.
- Auth pb.AuthClient
- // Lock is the lock API for the client's connection.
- Lock lockpb.LockClient
- // Election is the election API for the client's connection.
- Election epb.ElectionClient
- }
|