| 1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303 |
- // Copyright (c) 2012 The gocql Authors. All rights reserved.
- // Use of this source code is governed by a BSD-style
- // license that can be found in the LICENSE file.
- // +build all unit
- package gocql
- import (
- "bufio"
- "bytes"
- "context"
- "crypto/tls"
- "crypto/x509"
- "fmt"
- "io"
- "io/ioutil"
- "math/rand"
- "net"
- "os"
- "strings"
- "sync"
- "sync/atomic"
- "testing"
- "time"
- "github.com/gocql/gocql/internal/streams"
- )
- const (
- defaultProto = protoVersion2
- )
- func TestApprove(t *testing.T) {
- tests := map[bool]bool{
- approve("org.apache.cassandra.auth.PasswordAuthenticator"): true,
- approve("com.instaclustr.cassandra.auth.SharedSecretAuthenticator"): true,
- approve("com.datastax.bdp.cassandra.auth.DseAuthenticator"): true,
- approve("io.aiven.cassandra.auth.AivenAuthenticator"): true,
- approve("com.amazon.helenus.auth.HelenusAuthenticator"): true,
- approve("com.apache.cassandra.auth.FakeAuthenticator"): false,
- }
- for k, v := range tests {
- if k != v {
- t.Fatalf("expected '%v', got '%v'", k, v)
- }
- }
- }
- func TestJoinHostPort(t *testing.T) {
- tests := map[string]string{
- "127.0.0.1:0": JoinHostPort("127.0.0.1", 0),
- "127.0.0.1:1": JoinHostPort("127.0.0.1:1", 9142),
- "[2001:0db8:85a3:0000:0000:8a2e:0370:7334]:0": JoinHostPort("2001:0db8:85a3:0000:0000:8a2e:0370:7334", 0),
- "[2001:0db8:85a3:0000:0000:8a2e:0370:7334]:1": JoinHostPort("[2001:0db8:85a3:0000:0000:8a2e:0370:7334]:1", 9142),
- }
- for k, v := range tests {
- if k != v {
- t.Fatalf("expected '%v', got '%v'", k, v)
- }
- }
- }
- func testCluster(proto protoVersion, addresses ...string) *ClusterConfig {
- cluster := NewCluster(addresses...)
- cluster.ProtoVersion = int(proto)
- cluster.disableControlConn = true
- return cluster
- }
- func TestSimple(t *testing.T) {
- srv := NewTestServer(t, defaultProto, context.Background())
- defer srv.Stop()
- cluster := testCluster(defaultProto, srv.Address)
- db, err := cluster.CreateSession()
- if err != nil {
- t.Fatalf("0x%x: NewCluster: %v", defaultProto, err)
- }
- if err := db.Query("void").Exec(); err != nil {
- t.Fatalf("0x%x: %v", defaultProto, err)
- }
- }
- func TestSSLSimple(t *testing.T) {
- srv := NewSSLTestServer(t, defaultProto, context.Background())
- defer srv.Stop()
- db, err := createTestSslCluster(srv.Address, defaultProto, true).CreateSession()
- if err != nil {
- t.Fatalf("0x%x: NewCluster: %v", defaultProto, err)
- }
- if err := db.Query("void").Exec(); err != nil {
- t.Fatalf("0x%x: %v", defaultProto, err)
- }
- }
- func TestSSLSimpleNoClientCert(t *testing.T) {
- srv := NewSSLTestServer(t, defaultProto, context.Background())
- defer srv.Stop()
- db, err := createTestSslCluster(srv.Address, defaultProto, false).CreateSession()
- if err != nil {
- t.Fatalf("0x%x: NewCluster: %v", defaultProto, err)
- }
- if err := db.Query("void").Exec(); err != nil {
- t.Fatalf("0x%x: %v", defaultProto, err)
- }
- }
- func createTestSslCluster(addr string, proto protoVersion, useClientCert bool) *ClusterConfig {
- cluster := testCluster(proto, addr)
- sslOpts := &SslOptions{
- CaPath: "testdata/pki/ca.crt",
- EnableHostVerification: false,
- }
- if useClientCert {
- sslOpts.CertPath = "testdata/pki/gocql.crt"
- sslOpts.KeyPath = "testdata/pki/gocql.key"
- }
- cluster.SslOpts = sslOpts
- return cluster
- }
- func TestClosed(t *testing.T) {
- t.Skip("Skipping the execution of TestClosed for now to try to concentrate on more important test failures on Travis")
- srv := NewTestServer(t, defaultProto, context.Background())
- defer srv.Stop()
- session, err := newTestSession(defaultProto, srv.Address)
- if err != nil {
- t.Fatalf("0x%x: NewCluster: %v", defaultProto, err)
- }
- session.Close()
- if err := session.Query("void").Exec(); err != ErrSessionClosed {
- t.Fatalf("0x%x: expected %#v, got %#v", defaultProto, ErrSessionClosed, err)
- }
- }
- func newTestSession(proto protoVersion, addresses ...string) (*Session, error) {
- return testCluster(proto, addresses...).CreateSession()
- }
- func TestDNSLookupConnected(t *testing.T) {
- log := &testLogger{}
- Logger = log
- defer func() {
- Logger = &defaultLogger{}
- }()
- // Override the defaul DNS resolver and restore at the end
- failDNS = true
- defer func() { failDNS = false }()
- srv := NewTestServer(t, defaultProto, context.Background())
- defer srv.Stop()
- cluster := NewCluster("cassandra1.invalid", srv.Address, "cassandra2.invalid")
- cluster.ProtoVersion = int(defaultProto)
- cluster.disableControlConn = true
- // CreateSession() should attempt to resolve the DNS name "cassandraX.invalid"
- // and fail, but continue to connect via srv.Address
- _, err := cluster.CreateSession()
- if err != nil {
- t.Fatal("CreateSession() should have connected")
- }
- if !strings.Contains(log.String(), "gocql: dns error") {
- t.Fatalf("Expected to receive dns error log message - got '%s' instead", log.String())
- }
- }
- func TestDNSLookupError(t *testing.T) {
- log := &testLogger{}
- Logger = log
- defer func() {
- Logger = &defaultLogger{}
- }()
- // Override the defaul DNS resolver and restore at the end
- failDNS = true
- defer func() { failDNS = false }()
- cluster := NewCluster("cassandra1.invalid", "cassandra2.invalid")
- cluster.ProtoVersion = int(defaultProto)
- cluster.disableControlConn = true
- // CreateSession() should attempt to resolve each DNS name "cassandraX.invalid"
- // and fail since it could not resolve any dns entries
- _, err := cluster.CreateSession()
- if err == nil {
- t.Fatal("CreateSession() should have returned an error")
- }
- if !strings.Contains(log.String(), "gocql: dns error") {
- t.Fatalf("Expected to receive dns error log message - got '%s' instead", log.String())
- }
- if err.Error() != "gocql: unable to create session: failed to resolve any of the provided hostnames" {
- t.Fatalf("Expected CreateSession() to fail with message - got '%s' instead", err.Error())
- }
- }
- func TestStartupTimeout(t *testing.T) {
- ctx, cancel := context.WithCancel(context.Background())
- log := &testLogger{}
- Logger = log
- defer func() {
- Logger = &defaultLogger{}
- }()
- srv := NewTestServer(t, defaultProto, ctx)
- defer srv.Stop()
- // Tell the server to never respond to Startup frame
- atomic.StoreInt32(&srv.TimeoutOnStartup, 1)
- startTime := time.Now()
- cluster := NewCluster(srv.Address)
- cluster.ProtoVersion = int(defaultProto)
- cluster.disableControlConn = true
- // Set very long query connection timeout
- // so we know CreateSession() is using the ConnectTimeout
- cluster.Timeout = time.Second * 5
- // Create session should timeout during connect attempt
- _, err := cluster.CreateSession()
- if err == nil {
- t.Fatal("CreateSession() should have returned a timeout error")
- }
- elapsed := time.Since(startTime)
- if elapsed > time.Second*5 {
- t.Fatal("ConnectTimeout is not respected")
- }
- if !strings.Contains(err.Error(), "no connections were made when creating the session") {
- t.Fatalf("Expected to receive no connections error - got '%s'", err)
- }
- if !strings.Contains(log.String(), "no response to connection startup within timeout") {
- t.Fatalf("Expected to receive timeout log message - got '%s'", log.String())
- }
- cancel()
- }
- func TestTimeout(t *testing.T) {
- ctx, cancel := context.WithCancel(context.Background())
- srv := NewTestServer(t, defaultProto, ctx)
- defer srv.Stop()
- db, err := newTestSession(defaultProto, srv.Address)
- if err != nil {
- t.Fatalf("NewCluster: %v", err)
- }
- defer db.Close()
- var wg sync.WaitGroup
- wg.Add(1)
- go func() {
- defer wg.Done()
- select {
- case <-time.After(5 * time.Second):
- t.Errorf("no timeout")
- case <-ctx.Done():
- }
- }()
- if err := db.Query("kill").WithContext(ctx).Exec(); err == nil {
- t.Fatal("expected error got nil")
- }
- cancel()
- wg.Wait()
- }
- func TestCancel(t *testing.T) {
- ctx, cancel := context.WithCancel(context.Background())
- defer cancel()
- srv := NewTestServer(t, defaultProto, ctx)
- defer srv.Stop()
- cluster := testCluster(defaultProto, srv.Address)
- cluster.Timeout = 1 * time.Second
- db, err := cluster.CreateSession()
- if err != nil {
- t.Fatalf("NewCluster: %v", err)
- }
- defer db.Close()
- qry := db.Query("timeout").WithContext(ctx)
- // Make sure we finish the query without leftovers
- var wg sync.WaitGroup
- wg.Add(1)
- go func() {
- if err := qry.Exec(); err != context.Canceled {
- t.Fatalf("expected to get context cancel error: '%v', got '%v'", context.Canceled, err)
- }
- wg.Done()
- }()
- // The query will timeout after about 1 seconds, so cancel it after a short pause
- time.AfterFunc(20*time.Millisecond, cancel)
- wg.Wait()
- }
- type testQueryObserver struct {
- metrics map[string]*hostMetrics
- verbose bool
- }
- func (o *testQueryObserver) ObserveQuery(ctx context.Context, q ObservedQuery) {
- host := q.Host.ConnectAddress().String()
- o.metrics[host] = q.Metrics
- if o.verbose {
- Logger.Printf("Observed query %q. Returned %v rows, took %v on host %q with %v attempts and total latency %v. Error: %q\n",
- q.Statement, q.Rows, q.End.Sub(q.Start), host, q.Metrics.Attempts, q.Metrics.TotalLatency, q.Err)
- }
- }
- func (o *testQueryObserver) GetMetrics(host *HostInfo) *hostMetrics {
- return o.metrics[host.ConnectAddress().String()]
- }
- // TestQueryRetry will test to make sure that gocql will execute
- // the exact amount of retry queries designated by the user.
- func TestQueryRetry(t *testing.T) {
- ctx, cancel := context.WithCancel(context.Background())
- defer cancel()
- srv := NewTestServer(t, defaultProto, ctx)
- defer srv.Stop()
- db, err := newTestSession(defaultProto, srv.Address)
- if err != nil {
- t.Fatalf("NewCluster: %v", err)
- }
- defer db.Close()
- go func() {
- select {
- case <-ctx.Done():
- return
- case <-time.After(5 * time.Second):
- t.Errorf("no timeout")
- }
- }()
- rt := &SimpleRetryPolicy{NumRetries: 1}
- qry := db.Query("kill").RetryPolicy(rt)
- if err := qry.Exec(); err == nil {
- t.Fatalf("expected error")
- }
- requests := atomic.LoadInt64(&srv.nKillReq)
- attempts := qry.Attempts()
- if requests != int64(attempts) {
- t.Fatalf("expected requests %v to match query attempts %v", requests, attempts)
- }
- // the query will only be attempted once, but is being retried
- if requests != int64(rt.NumRetries) {
- t.Fatalf("failed to retry the query %v time(s). Query executed %v times", rt.NumRetries, requests-1)
- }
- }
- func TestQueryMultinodeWithMetrics(t *testing.T) {
- log := &testLogger{}
- Logger = log
- defer func() {
- Logger = &defaultLogger{}
- os.Stdout.WriteString(log.String())
- }()
- // Build a 3 node cluster to test host metric mapping
- var nodes []*TestServer
- var addresses = []string{
- "127.0.0.1",
- "127.0.0.2",
- "127.0.0.3",
- }
- // Can do with 1 context for all servers
- ctx := context.Background()
- for _, ip := range addresses {
- srv := NewTestServerWithAddress(ip+":0", t, defaultProto, ctx)
- defer srv.Stop()
- nodes = append(nodes, srv)
- }
- db, err := newTestSession(defaultProto, nodes[0].Address, nodes[1].Address, nodes[2].Address)
- if err != nil {
- t.Fatalf("NewCluster: %v", err)
- }
- defer db.Close()
- // 1 retry per host
- rt := &SimpleRetryPolicy{NumRetries: 3}
- observer := &testQueryObserver{metrics: make(map[string]*hostMetrics), verbose: false}
- qry := db.Query("kill").RetryPolicy(rt).Observer(observer)
- if err := qry.Exec(); err == nil {
- t.Fatalf("expected error")
- }
- for i, ip := range addresses {
- host := &HostInfo{connectAddress: net.ParseIP(ip)}
- queryMetric := qry.metrics.hostMetrics(host)
- observedMetrics := observer.GetMetrics(host)
- requests := int(atomic.LoadInt64(&nodes[i].nKillReq))
- hostAttempts := queryMetric.Attempts
- if requests != hostAttempts {
- t.Fatalf("expected requests %v to match query attempts %v", requests, hostAttempts)
- }
- if hostAttempts != observedMetrics.Attempts {
- t.Fatalf("expected observed attempts %v to match query attempts %v on host %v", observedMetrics.Attempts, hostAttempts, ip)
- }
- hostLatency := queryMetric.TotalLatency
- observedLatency := observedMetrics.TotalLatency
- if hostLatency != observedLatency {
- t.Fatalf("expected observed latency %v to match query latency %v on host %v", observedLatency, hostLatency, ip)
- }
- }
- // the query will only be attempted once, but is being retried
- attempts := qry.Attempts()
- if attempts != rt.NumRetries {
- t.Fatalf("failed to retry the query %v time(s). Query executed %v times", rt.NumRetries, attempts)
- }
- }
- type testRetryPolicy struct {
- NumRetries int
- }
- func (t *testRetryPolicy) Attempt(qry RetryableQuery) bool {
- return qry.Attempts() <= t.NumRetries
- }
- func (t *testRetryPolicy) GetRetryType(err error) RetryType {
- return Retry
- }
- func TestSpeculativeExecution(t *testing.T) {
- log := &testLogger{}
- Logger = log
- defer func() {
- Logger = &defaultLogger{}
- os.Stdout.WriteString(log.String())
- }()
- // Build a 3 node cluster
- var nodes []*TestServer
- var addresses = []string{
- "127.0.0.1",
- "127.0.0.2",
- "127.0.0.3",
- }
- // Can do with 1 context for all servers
- ctx := context.Background()
- for _, ip := range addresses {
- srv := NewTestServerWithAddress(ip+":0", t, defaultProto, ctx)
- defer srv.Stop()
- nodes = append(nodes, srv)
- }
- db, err := newTestSession(defaultProto, nodes[0].Address, nodes[1].Address, nodes[2].Address)
- if err != nil {
- t.Fatalf("NewCluster: %v", err)
- }
- defer db.Close()
- // Create a test retry policy, 6 retries will cover 2 executions
- rt := &testRetryPolicy{NumRetries: 8}
- // test Speculative policy with 1 additional execution
- sp := &SimpleSpeculativeExecution{NumAttempts: 1, TimeoutDelay: 200 * time.Millisecond}
- // Build the query
- qry := db.Query("speculative").RetryPolicy(rt).SetSpeculativeExecutionPolicy(sp).Idempotent(true)
- // Execute the query and close, check that it doesn't error out
- if err := qry.Exec(); err != nil {
- t.Errorf("The query failed with '%v'!\n", err)
- }
- requests1 := atomic.LoadInt64(&nodes[0].nKillReq)
- requests2 := atomic.LoadInt64(&nodes[1].nKillReq)
- requests3 := atomic.LoadInt64(&nodes[2].nKillReq)
- // Spec Attempts == 1, so expecting to see only 1 regular + 1 speculative = 2 nodes attempted
- if requests1 != 0 && requests2 != 0 && requests3 != 0 {
- t.Error("error: all 3 nodes were attempted, should have been only 2")
- }
- // Only the 4th request will generate results, so
- if requests1 != 4 && requests2 != 4 && requests3 != 4 {
- t.Error("error: none of 3 nodes was attempted 4 times!")
- }
- // "speculative" query will succeed on one arbitrary node after 4 attempts, so
- // expecting to see 4 (on successful node) + not more than 2 (as cancelled on another node) == 6
- if requests1+requests2+requests3 > 6 {
- t.Errorf("error: expected to see 6 attempts, got %v\n", requests1+requests2+requests3)
- }
- }
- func TestStreams_Protocol1(t *testing.T) {
- srv := NewTestServer(t, protoVersion1, context.Background())
- defer srv.Stop()
- // TODO: these are more like session tests and should instead operate
- // on a single Conn
- cluster := testCluster(protoVersion1, srv.Address)
- cluster.NumConns = 1
- cluster.ProtoVersion = 1
- db, err := cluster.CreateSession()
- if err != nil {
- t.Fatal(err)
- }
- defer db.Close()
- var wg sync.WaitGroup
- for i := 1; i < 128; i++ {
- // here were just validating that if we send NumStream request we get
- // a response for every stream and the lengths for the queries are set
- // correctly.
- wg.Add(1)
- go func() {
- defer wg.Done()
- if err := db.Query("void").Exec(); err != nil {
- t.Error(err)
- }
- }()
- }
- wg.Wait()
- }
- func TestStreams_Protocol3(t *testing.T) {
- srv := NewTestServer(t, protoVersion3, context.Background())
- defer srv.Stop()
- // TODO: these are more like session tests and should instead operate
- // on a single Conn
- cluster := testCluster(protoVersion3, srv.Address)
- cluster.NumConns = 1
- cluster.ProtoVersion = 3
- db, err := cluster.CreateSession()
- if err != nil {
- t.Fatal(err)
- }
- defer db.Close()
- for i := 1; i < 32768; i++ {
- // the test server processes each conn synchronously
- // here were just validating that if we send NumStream request we get
- // a response for every stream and the lengths for the queries are set
- // correctly.
- if err = db.Query("void").Exec(); err != nil {
- t.Fatal(err)
- }
- }
- }
- func BenchmarkProtocolV3(b *testing.B) {
- srv := NewTestServer(b, protoVersion3, context.Background())
- defer srv.Stop()
- // TODO: these are more like session tests and should instead operate
- // on a single Conn
- cluster := NewCluster(srv.Address)
- cluster.NumConns = 1
- cluster.ProtoVersion = 3
- db, err := cluster.CreateSession()
- if err != nil {
- b.Fatal(err)
- }
- defer db.Close()
- b.ResetTimer()
- b.ReportAllocs()
- for i := 0; i < b.N; i++ {
- if err = db.Query("void").Exec(); err != nil {
- b.Fatal(err)
- }
- }
- }
- // This tests that the policy connection pool handles SSL correctly
- func TestPolicyConnPoolSSL(t *testing.T) {
- srv := NewSSLTestServer(t, defaultProto, context.Background())
- defer srv.Stop()
- cluster := createTestSslCluster(srv.Address, defaultProto, true)
- cluster.PoolConfig.HostSelectionPolicy = RoundRobinHostPolicy()
- db, err := cluster.CreateSession()
- if err != nil {
- t.Fatalf("failed to create new session: %v", err)
- }
- if err := db.Query("void").Exec(); err != nil {
- t.Fatalf("query failed due to error: %v", err)
- }
- db.Close()
- // wait for the pool to drain
- time.Sleep(100 * time.Millisecond)
- size := db.pool.Size()
- if size != 0 {
- t.Fatalf("connection pool did not drain, still contains %d connections", size)
- }
- }
- func TestQueryTimeout(t *testing.T) {
- srv := NewTestServer(t, defaultProto, context.Background())
- defer srv.Stop()
- cluster := testCluster(defaultProto, srv.Address)
- // Set the timeout arbitrarily low so that the query hits the timeout in a
- // timely manner.
- cluster.Timeout = 1 * time.Millisecond
- db, err := cluster.CreateSession()
- if err != nil {
- t.Fatalf("NewCluster: %v", err)
- }
- defer db.Close()
- ch := make(chan error, 1)
- go func() {
- err := db.Query("timeout").Exec()
- if err != nil {
- ch <- err
- return
- }
- t.Errorf("err was nil, expected to get a timeout after %v", db.cfg.Timeout)
- }()
- select {
- case err := <-ch:
- if err != ErrTimeoutNoResponse {
- t.Fatalf("expected to get %v for timeout got %v", ErrTimeoutNoResponse, err)
- }
- case <-time.After(40*time.Millisecond + db.cfg.Timeout):
- // ensure that the query goroutines have been scheduled
- t.Fatalf("query did not timeout after %v", db.cfg.Timeout)
- }
- }
- func BenchmarkSingleConn(b *testing.B) {
- srv := NewTestServer(b, 3, context.Background())
- defer srv.Stop()
- cluster := testCluster(3, srv.Address)
- // Set the timeout arbitrarily low so that the query hits the timeout in a
- // timely manner.
- cluster.Timeout = 500 * time.Millisecond
- cluster.NumConns = 1
- db, err := cluster.CreateSession()
- if err != nil {
- b.Fatalf("NewCluster: %v", err)
- }
- defer db.Close()
- b.ResetTimer()
- b.RunParallel(func(pb *testing.PB) {
- for pb.Next() {
- err := db.Query("void").Exec()
- if err != nil {
- b.Error(err)
- return
- }
- }
- })
- }
- func TestQueryTimeoutReuseStream(t *testing.T) {
- t.Skip("no longer tests anything")
- // TODO(zariel): move this to conn test, we really just want to check what
- // happens when a conn is
- srv := NewTestServer(t, defaultProto, context.Background())
- defer srv.Stop()
- cluster := testCluster(defaultProto, srv.Address)
- // Set the timeout arbitrarily low so that the query hits the timeout in a
- // timely manner.
- cluster.Timeout = 1 * time.Millisecond
- cluster.NumConns = 1
- db, err := cluster.CreateSession()
- if err != nil {
- t.Fatalf("NewCluster: %v", err)
- }
- defer db.Close()
- db.Query("slow").Exec()
- err = db.Query("void").Exec()
- if err != nil {
- t.Fatal(err)
- }
- }
- func TestQueryTimeoutClose(t *testing.T) {
- srv := NewTestServer(t, defaultProto, context.Background())
- defer srv.Stop()
- cluster := testCluster(defaultProto, srv.Address)
- // Set the timeout arbitrarily low so that the query hits the timeout in a
- // timely manner.
- cluster.Timeout = 1000 * time.Millisecond
- cluster.NumConns = 1
- db, err := cluster.CreateSession()
- if err != nil {
- t.Fatalf("NewCluster: %v", err)
- }
- ch := make(chan error)
- go func() {
- err := db.Query("timeout").Exec()
- ch <- err
- }()
- // ensure that the above goroutine gets sheduled
- time.Sleep(50 * time.Millisecond)
- db.Close()
- select {
- case err = <-ch:
- case <-time.After(1 * time.Second):
- t.Fatal("timedout waiting to get a response once cluster is closed")
- }
- if err != ErrConnectionClosed {
- t.Fatalf("expected to get %v got %v", ErrConnectionClosed, err)
- }
- }
- func TestStream0(t *testing.T) {
- // TODO: replace this with type check
- const expErr = "gocql: received unexpected frame on stream 0"
- var buf bytes.Buffer
- f := newFramer(nil, &buf, nil, protoVersion4)
- f.writeHeader(0, opResult, 0)
- f.writeInt(resultKindVoid)
- f.wbuf[0] |= 0x80
- if err := f.finishWrite(); err != nil {
- t.Fatal(err)
- }
- conn := &Conn{
- r: bufio.NewReader(&buf),
- streams: streams.New(protoVersion4),
- }
- err := conn.recv(context.Background())
- if err == nil {
- t.Fatal("expected to get an error on stream 0")
- } else if !strings.HasPrefix(err.Error(), expErr) {
- t.Fatalf("expected to get error prefix %q got %q", expErr, err.Error())
- }
- }
- func TestContext_Timeout(t *testing.T) {
- ctx, cancel := context.WithCancel(context.Background())
- defer cancel()
- srv := NewTestServer(t, defaultProto, ctx)
- defer srv.Stop()
- cluster := testCluster(defaultProto, srv.Address)
- cluster.Timeout = 5 * time.Second
- db, err := cluster.CreateSession()
- if err != nil {
- t.Fatal(err)
- }
- defer db.Close()
- ctx, cancel = context.WithCancel(ctx)
- cancel()
- err = db.Query("timeout").WithContext(ctx).Exec()
- if err != context.Canceled {
- t.Fatalf("expected to get context cancel error: %v got %v", context.Canceled, err)
- }
- }
- // tcpConnPair returns a matching set of a TCP client side and server side connection.
- func tcpConnPair() (s, c net.Conn, err error) {
- l, err := net.Listen("tcp", "localhost:0")
- if err != nil {
- // maybe ipv6 works, if ipv4 fails?
- l, err = net.Listen("tcp6", "[::1]:0")
- if err != nil {
- return nil, nil, err
- }
- }
- defer l.Close() // we only try to accept one connection, so will stop listening.
- addr := l.Addr()
- done := make(chan struct{})
- var errDial error
- go func(done chan<- struct{}) {
- c, errDial = net.Dial(addr.Network(), addr.String())
- close(done)
- }(done)
- s, err = l.Accept()
- <-done
- if err == nil {
- err = errDial
- }
- if err != nil {
- if s != nil {
- s.Close()
- }
- if c != nil {
- c.Close()
- }
- }
- return s, c, err
- }
- func TestWriteCoalescing(t *testing.T) {
- ctx, cancel := context.WithCancel(context.Background())
- defer cancel()
- server, client, err := tcpConnPair()
- if err != nil {
- t.Fatal(err)
- }
- done := make(chan struct{}, 1)
- var (
- buf bytes.Buffer
- bufMutex sync.Mutex
- )
- go func() {
- defer close(done)
- defer server.Close()
- var err error
- b := make([]byte, 256)
- var n int
- for {
- if n, err = server.Read(b); err != nil {
- break
- }
- bufMutex.Lock()
- buf.Write(b[:n])
- bufMutex.Unlock()
- }
- if err != io.EOF {
- t.Errorf("unexpected read error: %v", err)
- }
- }()
- w := &writeCoalescer{
- c: client,
- writeCh: make(chan struct{}),
- cond: sync.NewCond(&sync.Mutex{}),
- quit: ctx.Done(),
- running: true,
- }
- go func() {
- if _, err := w.Write([]byte("one")); err != nil {
- t.Error(err)
- }
- }()
- go func() {
- if _, err := w.Write([]byte("two")); err != nil {
- t.Error(err)
- }
- }()
- bufMutex.Lock()
- if buf.Len() != 0 {
- t.Fatalf("expected buffer to be empty have: %v", buf.String())
- }
- bufMutex.Unlock()
- for true {
- w.cond.L.Lock()
- if len(w.buffers) == 2 {
- w.cond.L.Unlock()
- break
- }
- w.cond.L.Unlock()
- }
- w.flush()
- client.Close()
- <-done
- if got := buf.String(); got != "onetwo" && got != "twoone" {
- t.Fatalf("expected to get %q got %q", "onetwo or twoone", got)
- }
- }
- func TestWriteCoalescing_WriteAfterClose(t *testing.T) {
- ctx, cancel := context.WithCancel(context.Background())
- defer cancel()
- var buf bytes.Buffer
- defer cancel()
- server, client, err := tcpConnPair()
- if err != nil {
- t.Fatal(err)
- }
- done := make(chan struct{}, 1)
- go func() {
- io.Copy(&buf, server)
- server.Close()
- close(done)
- }()
- w := newWriteCoalescer(client, 0, 5*time.Millisecond, ctx.Done())
- // ensure 1 write works
- if _, err := w.Write([]byte("one")); err != nil {
- t.Fatal(err)
- }
- client.Close()
- <-done
- if v := buf.String(); v != "one" {
- t.Fatalf("expected buffer to be %q got %q", "one", v)
- }
- // now close and do a write, we should error
- cancel()
- client.Close() // close client conn too, since server won't see the answer anyway.
- if _, err := w.Write([]byte("two")); err == nil {
- t.Fatal("expected to get error for write after closing")
- } else if err != io.EOF {
- t.Fatalf("expected to get EOF got %v", err)
- }
- }
- type recordingFrameHeaderObserver struct {
- t *testing.T
- mu sync.Mutex
- frames []ObservedFrameHeader
- }
- func (r *recordingFrameHeaderObserver) ObserveFrameHeader(ctx context.Context, frm ObservedFrameHeader) {
- r.mu.Lock()
- r.frames = append(r.frames, frm)
- r.mu.Unlock()
- }
- func (r *recordingFrameHeaderObserver) getFrames() []ObservedFrameHeader {
- r.mu.Lock()
- defer r.mu.Unlock()
- return r.frames
- }
- func TestFrameHeaderObserver(t *testing.T) {
- srv := NewTestServer(t, defaultProto, context.Background())
- defer srv.Stop()
- cluster := testCluster(defaultProto, srv.Address)
- cluster.NumConns = 1
- observer := &recordingFrameHeaderObserver{t: t}
- cluster.FrameHeaderObserver = observer
- db, err := cluster.CreateSession()
- if err != nil {
- t.Fatal(err)
- }
- if err := db.Query("void").Exec(); err != nil {
- t.Fatal(err)
- }
- frames := observer.getFrames()
- expFrames := []frameOp{opSupported, opReady, opResult}
- if len(frames) != len(expFrames) {
- t.Fatalf("Expected to receive %d frames, instead received %d", len(expFrames), len(frames))
- }
- for i, op := range expFrames {
- if op != frames[i].Opcode {
- t.Fatalf("expected frame %d to be %v got %v", i, op, frames[i])
- }
- }
- voidResultFrame := frames[2]
- if voidResultFrame.Length != int32(4) {
- t.Fatalf("Expected to receive frame with body length 4, instead received body length %d", voidResultFrame.Length)
- }
- }
- func NewTestServerWithAddress(addr string, t testing.TB, protocol uint8, ctx context.Context) *TestServer {
- laddr, err := net.ResolveTCPAddr("tcp", addr)
- if err != nil {
- t.Fatal(err)
- }
- listen, err := net.ListenTCP("tcp", laddr)
- if err != nil {
- t.Fatal(err)
- }
- headerSize := 8
- if protocol > protoVersion2 {
- headerSize = 9
- }
- ctx, cancel := context.WithCancel(ctx)
- srv := &TestServer{
- Address: listen.Addr().String(),
- listen: listen,
- t: t,
- protocol: protocol,
- headerSize: headerSize,
- ctx: ctx,
- cancel: cancel,
- }
- go srv.closeWatch()
- go srv.serve()
- return srv
- }
- func NewTestServer(t testing.TB, protocol uint8, ctx context.Context) *TestServer {
- return NewTestServerWithAddress("127.0.0.1:0", t, protocol, ctx)
- }
- func NewSSLTestServer(t testing.TB, protocol uint8, ctx context.Context) *TestServer {
- pem, err := ioutil.ReadFile("testdata/pki/ca.crt")
- certPool := x509.NewCertPool()
- if !certPool.AppendCertsFromPEM(pem) {
- t.Fatalf("Failed parsing or appending certs")
- }
- mycert, err := tls.LoadX509KeyPair("testdata/pki/cassandra.crt", "testdata/pki/cassandra.key")
- if err != nil {
- t.Fatalf("could not load cert")
- }
- config := &tls.Config{
- Certificates: []tls.Certificate{mycert},
- RootCAs: certPool,
- }
- listen, err := tls.Listen("tcp", "127.0.0.1:0", config)
- if err != nil {
- t.Fatal(err)
- }
- headerSize := 8
- if protocol > protoVersion2 {
- headerSize = 9
- }
- ctx, cancel := context.WithCancel(ctx)
- srv := &TestServer{
- Address: listen.Addr().String(),
- listen: listen,
- t: t,
- protocol: protocol,
- headerSize: headerSize,
- ctx: ctx,
- cancel: cancel,
- }
- go srv.closeWatch()
- go srv.serve()
- return srv
- }
- type TestServer struct {
- Address string
- TimeoutOnStartup int32
- t testing.TB
- nreq uint64
- listen net.Listener
- nKillReq int64
- compressor Compressor
- protocol byte
- headerSize int
- ctx context.Context
- cancel context.CancelFunc
- quit chan struct{}
- mu sync.Mutex
- closed bool
- }
- func (srv *TestServer) session() (*Session, error) {
- return testCluster(protoVersion(srv.protocol), srv.Address).CreateSession()
- }
- func (srv *TestServer) host() *HostInfo {
- hosts, err := hostInfo(srv.Address, 9042)
- if err != nil {
- srv.t.Fatal(err)
- }
- return hosts[0]
- }
- func (srv *TestServer) closeWatch() {
- <-srv.ctx.Done()
- srv.mu.Lock()
- defer srv.mu.Unlock()
- srv.closeLocked()
- }
- func (srv *TestServer) serve() {
- defer srv.listen.Close()
- for !srv.isClosed() {
- conn, err := srv.listen.Accept()
- if err != nil {
- break
- }
- go func(conn net.Conn) {
- defer conn.Close()
- for !srv.isClosed() {
- framer, err := srv.readFrame(conn)
- if err != nil {
- if err == io.EOF {
- return
- }
- srv.errorLocked(err)
- return
- }
- atomic.AddUint64(&srv.nreq, 1)
- go srv.process(framer)
- }
- }(conn)
- }
- }
- func (srv *TestServer) isClosed() bool {
- srv.mu.Lock()
- defer srv.mu.Unlock()
- return srv.closed
- }
- func (srv *TestServer) closeLocked() {
- if srv.closed {
- return
- }
- srv.closed = true
- srv.listen.Close()
- srv.cancel()
- }
- func (srv *TestServer) Stop() {
- srv.mu.Lock()
- defer srv.mu.Unlock()
- srv.closeLocked()
- }
- func (srv *TestServer) errorLocked(err interface{}) {
- srv.mu.Lock()
- defer srv.mu.Unlock()
- if srv.closed {
- return
- }
- srv.t.Error(err)
- }
- func (srv *TestServer) process(f *framer) {
- head := f.header
- if head == nil {
- srv.errorLocked("process frame with a nil header")
- return
- }
- switch head.op {
- case opStartup:
- if atomic.LoadInt32(&srv.TimeoutOnStartup) > 0 {
- // Do not respond to startup command
- // wait until we get a cancel signal
- select {
- case <-srv.ctx.Done():
- return
- }
- }
- f.writeHeader(0, opReady, head.stream)
- case opOptions:
- f.writeHeader(0, opSupported, head.stream)
- f.writeShort(0)
- case opQuery:
- query := f.readLongString()
- first := query
- if n := strings.Index(query, " "); n > 0 {
- first = first[:n]
- }
- switch strings.ToLower(first) {
- case "kill":
- atomic.AddInt64(&srv.nKillReq, 1)
- f.writeHeader(0, opError, head.stream)
- f.writeInt(0x1001)
- f.writeString("query killed")
- case "use":
- f.writeInt(resultKindKeyspace)
- f.writeString(strings.TrimSpace(query[3:]))
- case "void":
- f.writeHeader(0, opResult, head.stream)
- f.writeInt(resultKindVoid)
- case "timeout":
- <-srv.ctx.Done()
- return
- case "slow":
- go func() {
- f.writeHeader(0, opResult, head.stream)
- f.writeInt(resultKindVoid)
- f.wbuf[0] = srv.protocol | 0x80
- select {
- case <-srv.ctx.Done():
- return
- case <-time.After(50 * time.Millisecond):
- f.finishWrite()
- }
- }()
- return
- case "speculative":
- atomic.AddInt64(&srv.nKillReq, 1)
- if atomic.LoadInt64(&srv.nKillReq) > 3 {
- f.writeHeader(0, opResult, head.stream)
- f.writeInt(resultKindVoid)
- f.writeString("speculative query success on the node " + srv.Address)
- } else {
- f.writeHeader(0, opError, head.stream)
- f.writeInt(0x1001)
- f.writeString("speculative error")
- rand.Seed(time.Now().UnixNano())
- <-time.After(time.Millisecond * 120)
- }
- default:
- f.writeHeader(0, opResult, head.stream)
- f.writeInt(resultKindVoid)
- }
- case opError:
- f.writeHeader(0, opError, head.stream)
- f.wbuf = append(f.wbuf, f.rbuf...)
- default:
- f.writeHeader(0, opError, head.stream)
- f.writeInt(0)
- f.writeString("not supported")
- }
- f.wbuf[0] = srv.protocol | 0x80
- if err := f.finishWrite(); err != nil {
- srv.errorLocked(err)
- }
- }
- func (srv *TestServer) readFrame(conn net.Conn) (*framer, error) {
- buf := make([]byte, srv.headerSize)
- head, err := readHeader(conn, buf)
- if err != nil {
- return nil, err
- }
- framer := newFramer(conn, conn, nil, srv.protocol)
- err = framer.readFrame(&head)
- if err != nil {
- return nil, err
- }
- // should be a request frame
- if head.version.response() {
- return nil, fmt.Errorf("expected to read a request frame got version: %v", head.version)
- } else if head.version.version() != srv.protocol {
- return nil, fmt.Errorf("expected to read protocol version 0x%x got 0x%x", srv.protocol, head.version.version())
- }
- return framer, nil
- }
|