| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104 |
- // Copyright 2015 The etcd Authors
- //
- // Licensed under the Apache License, Version 2.0 (the "License");
- // you may not use this file except in compliance with the License.
- // You may obtain a copy of the License at
- //
- // http://www.apache.org/licenses/LICENSE-2.0
- //
- // Unless required by applicable law or agreed to in writing, software
- // distributed under the License is distributed on an "AS IS" BASIS,
- // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- // See the License for the specific language governing permissions and
- // limitations under the License.
- package main
- import (
- "flag"
- "fmt"
- "net/http"
- "strings"
- "github.com/coreos/pkg/capnslog"
- "github.com/prometheus/client_golang/prometheus"
- )
- var plog = capnslog.NewPackageLogger("github.com/coreos/etcd", "etcd-tester")
- func main() {
- endpointStr := flag.String("agent-endpoints", "localhost:9027", "HTTP RPC endpoints of agents. Do not specify the schema.")
- datadir := flag.String("data-dir", "agent.etcd", "etcd data directory location on agent machine.")
- stressKeySize := flag.Uint("stress-key-size", 100, "the size of each key written into etcd.")
- stressKeySuffixRange := flag.Uint("stress-key-count", 250000, "the count of key range written into etcd.")
- stressKeyRangeLimit := flag.Uint("stress-range-limit", 50, "maximum number of keys to range or delete.")
- limit := flag.Int("limit", -1, "the limit of rounds to run failure set (-1 to run without limits).")
- stressQPS := flag.Int("stress-qps", 10000, "maximum number of stresser requests per second.")
- schedCases := flag.String("schedule-cases", "", "test case schedule")
- consistencyCheck := flag.Bool("consistency-check", true, "true to check consistency (revision, hash)")
- isV2Only := flag.Bool("v2-only", false, "'true' to run V2 only tester.")
- flag.Parse()
- c := &cluster{
- v2Only: *isV2Only,
- datadir: *datadir,
- stressQPS: *stressQPS,
- stressKeySize: int(*stressKeySize),
- stressKeySuffixRange: int(*stressKeySuffixRange),
- stressKeyRangeLimit: int(*stressKeyRangeLimit),
- }
- if err := c.bootstrap(strings.Split(*endpointStr, ",")); err != nil {
- plog.Fatal(err)
- }
- defer c.Terminate()
- failures := []failure{
- newFailureKillAll(),
- newFailureKillMajority(),
- newFailureKillOne(),
- newFailureKillLeader(),
- newFailureKillOneForLongTime(),
- newFailureKillLeaderForLongTime(),
- newFailureIsolate(),
- newFailureIsolateAll(),
- newFailureSlowNetworkOneMember(),
- newFailureSlowNetworkLeader(),
- newFailureSlowNetworkAll(),
- }
- // ensure cluster is fully booted to know failpoints are available
- c.WaitHealth()
- fpFailures, fperr := failpointFailures(c)
- if len(fpFailures) == 0 {
- plog.Infof("no failpoints found (%v)", fperr)
- }
- failures = append(failures, fpFailures...)
- schedule := failures
- if schedCases != nil && *schedCases != "" {
- cases := strings.Split(*schedCases, " ")
- schedule = make([]failure, len(cases))
- for i := range cases {
- caseNum := 0
- n, err := fmt.Sscanf(cases[i], "%d", &caseNum)
- if n == 0 || err != nil {
- plog.Fatalf(`couldn't parse case "%s" (%v)`, cases[i], err)
- }
- schedule[i] = failures[caseNum]
- }
- }
- t := &tester{
- failures: schedule,
- cluster: c,
- limit: *limit,
- consistencyCheck: *consistencyCheck,
- }
- sh := statusHandler{status: &t.status}
- http.Handle("/status", sh)
- http.Handle("/metrics", prometheus.Handler())
- go func() { plog.Fatal(http.ListenAndServe(":9028", nil)) }()
- t.runLoop()
- }
|