diff --git a/tests/framework/integration/cluster.go b/tests/framework/integration/cluster.go index 76c754e99..323398c7b 100644 --- a/tests/framework/integration/cluster.go +++ b/tests/framework/integration/cluster.go @@ -174,6 +174,9 @@ type Cluster struct { Cfg *ClusterConfig Members []*Member LastMemberNum int + + mu sync.Mutex + clusterClient *clientv3.Client } func SchemeFromTLSInfo(tls *transport.TLSInfo) string { @@ -207,22 +210,6 @@ func (c *Cluster) fillClusterForMembers() error { return nil } -func newClusterFromConfig(t testutil.TB, cfg *ClusterConfig) *Cluster { - testutil.SkipTestIfShortMode(t, "Cannot start etcd Cluster in --short tests") - - c := &Cluster{Cfg: cfg} - ms := make([]*Member, cfg.Size) - for i := 0; i < cfg.Size; i++ { - ms[i] = c.mustNewMember(t) - } - c.Members = ms - if err := c.fillClusterForMembers(); err != nil { - t.Fatal(err) - } - - return c -} - func (c *Cluster) Launch(t testutil.TB) { errc := make(chan error) for _, m := range c.Members { @@ -415,23 +402,6 @@ func (c *Cluster) RemoveMember(t testutil.TB, id uint64) error { return nil } -func (c *Cluster) Terminate(t testutil.TB) { - for _, m := range c.Members { - if m.Client != nil { - m.Client.Close() - } - } - var wg sync.WaitGroup - wg.Add(len(c.Members)) - for _, m := range c.Members { - go func(mm *Member) { - defer wg.Done() - mm.Terminate(t) - }(m) - } - wg.Wait() -} - func (c *Cluster) WaitMembersMatch(t testutil.TB, membs []client.Member) { for _, u := range c.URLs() { cc := MustNewHTTPClient(t, []string{u}, c.Cfg.ClientTLS) @@ -1383,35 +1353,36 @@ func (p SortableMemberSliceByPeerURLs) Less(i, j int) bool { } func (p SortableMemberSliceByPeerURLs) Swap(i, j int) { p[i], p[j] = p[j], p[i] } -type ClusterV3 struct { - *Cluster - - mu sync.Mutex - clusterClient *clientv3.Client -} - -// NewClusterV3 returns a launched Cluster with a grpc client connection +// NewCluster returns a launched Cluster with a grpc client connection // for each Cluster member. -func NewClusterV3(t testutil.TB, cfg *ClusterConfig) *ClusterV3 { +func NewCluster(t testutil.TB, cfg *ClusterConfig) *Cluster { t.Helper() assertInTestContext(t) - clus := &ClusterV3{ - Cluster: newClusterFromConfig(t, cfg), - } - clus.Launch(t) + testutil.SkipTestIfShortMode(t, "Cannot start etcd Cluster in --short tests") - return clus + c := &Cluster{Cfg: cfg} + ms := make([]*Member, cfg.Size) + for i := 0; i < cfg.Size; i++ { + ms[i] = c.mustNewMember(t) + } + c.Members = ms + if err := c.fillClusterForMembers(); err != nil { + t.Fatal(err) + } + c.Launch(t) + + return c } -func (c *ClusterV3) TakeClient(idx int) { +func (c *Cluster) TakeClient(idx int) { c.mu.Lock() c.Members[idx].Client = nil c.mu.Unlock() } -func (c *ClusterV3) Terminate(t testutil.TB) { +func (c *Cluster) Terminate(t testutil.TB) { c.mu.Lock() if c.clusterClient != nil { if err := c.clusterClient.Close(); err != nil { @@ -1419,18 +1390,31 @@ func (c *ClusterV3) Terminate(t testutil.TB) { } } c.mu.Unlock() - c.Cluster.Terminate(t) + for _, m := range c.Members { + if m.Client != nil { + m.Client.Close() + } + } + var wg sync.WaitGroup + wg.Add(len(c.Members)) + for _, m := range c.Members { + go func(mm *Member) { + defer wg.Done() + mm.Terminate(t) + }(m) + } + wg.Wait() } -func (c *ClusterV3) RandClient() *clientv3.Client { +func (c *Cluster) RandClient() *clientv3.Client { return c.Members[rand.Intn(len(c.Members))].Client } -func (c *ClusterV3) Client(i int) *clientv3.Client { +func (c *Cluster) Client(i int) *clientv3.Client { return c.Members[i].Client } -func (c *ClusterV3) ClusterClient() (client *clientv3.Client, err error) { +func (c *Cluster) ClusterClient() (client *clientv3.Client, err error) { if c.clusterClient == nil { endpoints := []string{} for _, m := range c.Members { @@ -1450,11 +1434,11 @@ func (c *ClusterV3) ClusterClient() (client *clientv3.Client, err error) { } // NewClientV3 creates a new grpc client connection to the member -func (c *ClusterV3) NewClientV3(memberIndex int) (*clientv3.Client, error) { +func (c *Cluster) NewClientV3(memberIndex int) (*clientv3.Client, error) { return NewClientV3(c.Members[memberIndex]) } -func makeClients(t testutil.TB, clus *ClusterV3, clients *[]*clientv3.Client, chooseMemberIndex func() int) func() *clientv3.Client { +func makeClients(t testutil.TB, clus *Cluster, clients *[]*clientv3.Client, chooseMemberIndex func() int) func() *clientv3.Client { var mu sync.Mutex *clients = nil return func() *clientv3.Client { @@ -1471,13 +1455,13 @@ func makeClients(t testutil.TB, clus *ClusterV3, clients *[]*clientv3.Client, ch // MakeSingleNodeClients creates factory of clients that all connect to member 0. // All the created clients are put on the 'clients' list. The factory is thread-safe. -func MakeSingleNodeClients(t testutil.TB, clus *ClusterV3, clients *[]*clientv3.Client) func() *clientv3.Client { +func MakeSingleNodeClients(t testutil.TB, clus *Cluster, clients *[]*clientv3.Client) func() *clientv3.Client { return makeClients(t, clus, clients, func() int { return 0 }) } // MakeMultiNodeClients creates factory of clients that all connect to random members. // All the created clients are put on the 'clients' list. The factory is thread-safe. -func MakeMultiNodeClients(t testutil.TB, clus *ClusterV3, clients *[]*clientv3.Client) func() *clientv3.Client { +func MakeMultiNodeClients(t testutil.TB, clus *Cluster, clients *[]*clientv3.Client) func() *clientv3.Client { return makeClients(t, clus, clients, func() int { return rand.Intn(len(clus.Members)) }) } @@ -1510,7 +1494,7 @@ type GrpcAPI struct { } // GetLearnerMembers returns the list of learner members in Cluster using MemberList API. -func (c *ClusterV3) GetLearnerMembers() ([]*pb.Member, error) { +func (c *Cluster) GetLearnerMembers() ([]*pb.Member, error) { cli := c.Client(0) resp, err := cli.MemberList(context.Background()) if err != nil { @@ -1527,7 +1511,7 @@ func (c *ClusterV3) GetLearnerMembers() ([]*pb.Member, error) { // AddAndLaunchLearnerMember creates a leaner member, adds it to Cluster // via v3 MemberAdd API, and then launches the new member. -func (c *ClusterV3) AddAndLaunchLearnerMember(t testutil.TB) { +func (c *Cluster) AddAndLaunchLearnerMember(t testutil.TB) { m := c.mustNewMember(t) m.IsLearner = true @@ -1557,7 +1541,7 @@ func (c *ClusterV3) AddAndLaunchLearnerMember(t testutil.TB) { } // getMembers returns a list of members in Cluster, in format of etcdserverpb.Member -func (c *ClusterV3) getMembers() []*pb.Member { +func (c *Cluster) getMembers() []*pb.Member { var mems []*pb.Member for _, m := range c.Members { mem := &pb.Member{ @@ -1594,7 +1578,7 @@ func (c *ClusterV3) getMembers() []*pb.Member { // indicate that the new learner member has applied the raftpb.ConfChangeAddLearnerNode entry // which was used to add the learner itself to the Cluster, and therefore it has the correct info // on learner. -func (c *ClusterV3) waitMembersMatch(t testutil.TB) { +func (c *Cluster) waitMembersMatch(t testutil.TB) { wMembers := c.getMembers() sort.Sort(SortableProtoMemberSliceByPeerURLs(wMembers)) cli := c.Client(0) @@ -1628,7 +1612,7 @@ func (p SortableProtoMemberSliceByPeerURLs) Less(i, j int) bool { func (p SortableProtoMemberSliceByPeerURLs) Swap(i, j int) { p[i], p[j] = p[j], p[i] } // MustNewMember creates a new member instance based on the response of V3 Member Add API. -func (c *ClusterV3) MustNewMember(t testutil.TB, resp *clientv3.MemberAddResponse) *Member { +func (c *Cluster) MustNewMember(t testutil.TB, resp *clientv3.MemberAddResponse) *Member { m := c.mustNewMember(t) m.IsLearner = resp.Member.IsLearner m.NewCluster = false diff --git a/tests/integration/clientv3/cluster_test.go b/tests/integration/clientv3/cluster_test.go index edf47f29b..a702e0364 100644 --- a/tests/integration/clientv3/cluster_test.go +++ b/tests/integration/clientv3/cluster_test.go @@ -30,7 +30,7 @@ import ( func TestMemberList(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) capi := clus.RandClient() @@ -48,7 +48,7 @@ func TestMemberList(t *testing.T) { func TestMemberAdd(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) capi := clus.RandClient() @@ -67,7 +67,7 @@ func TestMemberAdd(t *testing.T) { func TestMemberAddWithExistingURLs(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) capi := clus.RandClient() @@ -91,7 +91,7 @@ func TestMemberAddWithExistingURLs(t *testing.T) { func TestMemberRemove(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) capi := clus.Client(1) @@ -129,7 +129,7 @@ func TestMemberRemove(t *testing.T) { func TestMemberUpdate(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) capi := clus.RandClient() @@ -157,7 +157,7 @@ func TestMemberUpdate(t *testing.T) { func TestMemberAddUpdateWrongURLs(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) capi := clus.RandClient() @@ -190,7 +190,7 @@ func TestMemberAddUpdateWrongURLs(t *testing.T) { func TestMemberAddForLearner(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) capi := clus.RandClient() @@ -219,7 +219,7 @@ func TestMemberAddForLearner(t *testing.T) { func TestMemberPromote(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) // member promote request can be sent to any server in cluster, @@ -296,7 +296,7 @@ func TestMemberPromote(t *testing.T) { func TestMemberPromoteMemberNotLearner(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) // member promote request can be sent to any server in cluster, @@ -332,7 +332,7 @@ func TestMemberPromoteMemberNotLearner(t *testing.T) { func TestMemberPromoteMemberNotExist(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) // member promote request can be sent to any server in cluster, @@ -382,7 +382,7 @@ func TestMaxLearnerInCluster(t *testing.T) { integration2.BeforeTest(t) // 1. start with a cluster with 3 voting member and max learner 2 - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3, ExperimentalMaxLearners: 2}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3, ExperimentalMaxLearners: 2}) defer clus.Terminate(t) // 2. adding 2 learner members should succeed diff --git a/tests/integration/clientv3/concurrency/revision_test.go b/tests/integration/clientv3/concurrency/revision_test.go index c0a08c454..7e6236a7a 100644 --- a/tests/integration/clientv3/concurrency/revision_test.go +++ b/tests/integration/clientv3/concurrency/revision_test.go @@ -28,7 +28,7 @@ import ( ) func TestRevisionMonotonicWithLeaderPartitions(t *testing.T) { - testRevisionMonotonicWithFailures(t, 11*time.Second, func(clus *integration.ClusterV3) { + testRevisionMonotonicWithFailures(t, 11*time.Second, func(clus *integration.Cluster) { for i := 0; i < 5; i++ { leader := clus.WaitLeader(t) time.Sleep(time.Second) @@ -40,7 +40,7 @@ func TestRevisionMonotonicWithLeaderPartitions(t *testing.T) { } func TestRevisionMonotonicWithPartitions(t *testing.T) { - testRevisionMonotonicWithFailures(t, 11*time.Second, func(clus *integration.ClusterV3) { + testRevisionMonotonicWithFailures(t, 11*time.Second, func(clus *integration.Cluster) { for i := 0; i < 5; i++ { time.Sleep(time.Second) clus.Members[i%3].InjectPartition(t, clus.Members[(i+1)%3], clus.Members[(i+2)%3]) @@ -51,7 +51,7 @@ func TestRevisionMonotonicWithPartitions(t *testing.T) { } func TestRevisionMonotonicWithLeaderRestarts(t *testing.T) { - testRevisionMonotonicWithFailures(t, 11*time.Second, func(clus *integration.ClusterV3) { + testRevisionMonotonicWithFailures(t, 11*time.Second, func(clus *integration.Cluster) { for i := 0; i < 5; i++ { leader := clus.WaitLeader(t) time.Sleep(time.Second) @@ -63,7 +63,7 @@ func TestRevisionMonotonicWithLeaderRestarts(t *testing.T) { } func TestRevisionMonotonicWithRestarts(t *testing.T) { - testRevisionMonotonicWithFailures(t, 11*time.Second, func(clus *integration.ClusterV3) { + testRevisionMonotonicWithFailures(t, 11*time.Second, func(clus *integration.Cluster) { for i := 0; i < 5; i++ { time.Sleep(time.Second) clus.Members[i%3].Stop(t) @@ -73,9 +73,9 @@ func TestRevisionMonotonicWithRestarts(t *testing.T) { }) } -func testRevisionMonotonicWithFailures(t *testing.T, testDuration time.Duration, injectFailures func(clus *integration.ClusterV3)) { +func testRevisionMonotonicWithFailures(t *testing.T, testDuration time.Duration, injectFailures func(clus *integration.Cluster)) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) ctx, cancel := context.WithTimeout(context.Background(), testDuration) @@ -108,7 +108,7 @@ func testRevisionMonotonicWithFailures(t *testing.T, testDuration time.Duration, t.Logf("Revision %d", resp.Header.Revision) } -func putWorker(t *testing.T, ctx context.Context, clus *integration.ClusterV3) { +func putWorker(t *testing.T, ctx context.Context, clus *integration.Cluster) { for i := 0; ; i++ { kv := clus.Client(i % 3) _, err := kv.Put(ctx, "foo", fmt.Sprintf("%d", i)) @@ -121,7 +121,7 @@ func putWorker(t *testing.T, ctx context.Context, clus *integration.ClusterV3) { } } -func getWorker(t *testing.T, ctx context.Context, clus *integration.ClusterV3) { +func getWorker(t *testing.T, ctx context.Context, clus *integration.Cluster) { var prevRev int64 for i := 0; ; i++ { kv := clus.Client(i % 3) diff --git a/tests/integration/clientv3/connectivity/black_hole_test.go b/tests/integration/clientv3/connectivity/black_hole_test.go index 6519f76ab..befe70386 100644 --- a/tests/integration/clientv3/connectivity/black_hole_test.go +++ b/tests/integration/clientv3/connectivity/black_hole_test.go @@ -35,7 +35,7 @@ import ( func TestBalancerUnderBlackholeKeepAliveWatch(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{ + clus := integration2.NewCluster(t, &integration2.ClusterConfig{ Size: 2, GRPCKeepAliveMinTime: time.Millisecond, // avoid too_many_pings UseBridge: true, @@ -168,7 +168,7 @@ func TestBalancerUnderBlackholeNoKeepAliveSerializableGet(t *testing.T) { func testBalancerUnderBlackholeNoKeepAlive(t *testing.T, op func(*clientv3.Client, context.Context) error) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{ + clus := integration2.NewCluster(t, &integration2.ClusterConfig{ Size: 2, UseBridge: true, }) diff --git a/tests/integration/clientv3/connectivity/dial_test.go b/tests/integration/clientv3/connectivity/dial_test.go index 2742ad325..7f6babe04 100644 --- a/tests/integration/clientv3/connectivity/dial_test.go +++ b/tests/integration/clientv3/connectivity/dial_test.go @@ -48,7 +48,7 @@ var ( // TestDialTLSExpired tests client with expired certs fails to dial. func TestDialTLSExpired(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, PeerTLS: &testTLSInfo, ClientTLS: &testTLSInfo}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, PeerTLS: &testTLSInfo, ClientTLS: &testTLSInfo}) defer clus.Terminate(t) tls, err := testTLSInfoExpired.ClientConfig() @@ -71,7 +71,7 @@ func TestDialTLSExpired(t *testing.T) { // when TLS endpoints (https, unixs) are given but no tls config. func TestDialTLSNoConfig(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, ClientTLS: &testTLSInfo}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, ClientTLS: &testTLSInfo}) defer clus.Terminate(t) // expect "signed by unknown authority" c, err := integration2.NewClient(t, clientv3.Config{ @@ -102,7 +102,7 @@ func TestDialSetEndpointsAfterFail(t *testing.T) { // testDialSetEndpoints ensures SetEndpoints can replace unavailable endpoints with available ones. func testDialSetEndpoints(t *testing.T, setBefore bool) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) // get endpoint list @@ -145,7 +145,7 @@ func testDialSetEndpoints(t *testing.T, setBefore bool) { // with a new one that doesn't include original endpoint. func TestSwitchSetEndpoints(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) // get non partitioned members endpoints @@ -166,7 +166,7 @@ func TestSwitchSetEndpoints(t *testing.T) { func TestRejectOldCluster(t *testing.T) { integration2.BeforeTest(t) // 2 endpoints to test multi-endpoint Status - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 2}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 2}) defer clus.Terminate(t) cfg := clientv3.Config{ @@ -186,7 +186,7 @@ func TestRejectOldCluster(t *testing.T) { // with the balancer can be dialed. func TestDialForeignEndpoint(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 2}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 2}) defer clus.Terminate(t) conn, err := clus.Client(0).Dial(clus.Client(1).Endpoints()[0]) @@ -209,7 +209,7 @@ func TestDialForeignEndpoint(t *testing.T) { // to a working endpoint will always succeed. func TestSetEndpointAndPut(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 2}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 2}) defer clus.Terminate(t) clus.Client(1).SetEndpoints(clus.Members[0].GRPCURL()) diff --git a/tests/integration/clientv3/connectivity/network_partition_test.go b/tests/integration/clientv3/connectivity/network_partition_test.go index 9e3600e03..7e01773e0 100644 --- a/tests/integration/clientv3/connectivity/network_partition_test.go +++ b/tests/integration/clientv3/connectivity/network_partition_test.go @@ -105,7 +105,7 @@ func TestBalancerUnderNetworkPartitionSerializableGet(t *testing.T) { func testBalancerUnderNetworkPartition(t *testing.T, op func(*clientv3.Client, context.Context) error, timeout time.Duration) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{ + clus := integration2.NewCluster(t, &integration2.ClusterConfig{ Size: 3, }) defer clus.Terminate(t) @@ -160,7 +160,7 @@ func testBalancerUnderNetworkPartition(t *testing.T, op func(*clientv3.Client, c func TestBalancerUnderNetworkPartitionLinearizableGetLeaderElection(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{ + clus := integration2.NewCluster(t, &integration2.ClusterConfig{ Size: 3, }) defer clus.Terminate(t) @@ -214,7 +214,7 @@ func TestBalancerUnderNetworkPartitionWatchFollower(t *testing.T) { func testBalancerUnderNetworkPartitionWatch(t *testing.T, isolateLeader bool) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{ + clus := integration2.NewCluster(t, &integration2.ClusterConfig{ Size: 3, }) defer clus.Terminate(t) @@ -273,7 +273,7 @@ func testBalancerUnderNetworkPartitionWatch(t *testing.T, isolateLeader bool) { func TestDropReadUnderNetworkPartition(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{ + clus := integration2.NewCluster(t, &integration2.ClusterConfig{ Size: 3, }) defer clus.Terminate(t) diff --git a/tests/integration/clientv3/connectivity/server_shutdown_test.go b/tests/integration/clientv3/connectivity/server_shutdown_test.go index f175782a8..4cc4fb9c0 100644 --- a/tests/integration/clientv3/connectivity/server_shutdown_test.go +++ b/tests/integration/clientv3/connectivity/server_shutdown_test.go @@ -32,7 +32,7 @@ import ( func TestBalancerUnderServerShutdownWatch(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{ + clus := integration2.NewCluster(t, &integration2.ClusterConfig{ Size: 3, UseBridge: true, }) @@ -144,7 +144,7 @@ func TestBalancerUnderServerShutdownTxn(t *testing.T) { func testBalancerUnderServerShutdownMutable(t *testing.T, op func(*clientv3.Client, context.Context) error) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{ + clus := integration2.NewCluster(t, &integration2.ClusterConfig{ Size: 3, }) defer clus.Terminate(t) @@ -201,7 +201,7 @@ func TestBalancerUnderServerShutdownGetSerializable(t *testing.T) { func testBalancerUnderServerShutdownImmutable(t *testing.T, op func(*clientv3.Client, context.Context) error, timeout time.Duration) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{ + clus := integration2.NewCluster(t, &integration2.ClusterConfig{ Size: 3, }) defer clus.Terminate(t) @@ -281,7 +281,7 @@ func testBalancerUnderServerStopInflightRangeOnRestart(t *testing.T, linearizabl cfg.Size = 3 } - clus := integration2.NewClusterV3(t, cfg) + clus := integration2.NewCluster(t, cfg) defer clus.Terminate(t) eps := []string{clus.Members[0].GRPCURL(), clus.Members[1].GRPCURL()} if linearizable { diff --git a/tests/integration/clientv3/experimental/recipes/v3_barrier_test.go b/tests/integration/clientv3/experimental/recipes/v3_barrier_test.go index 5692db140..e20885614 100644 --- a/tests/integration/clientv3/experimental/recipes/v3_barrier_test.go +++ b/tests/integration/clientv3/experimental/recipes/v3_barrier_test.go @@ -25,14 +25,14 @@ import ( func TestBarrierSingleNode(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) testBarrier(t, 5, func() *clientv3.Client { return clus.Client(0) }) } func TestBarrierMultiNode(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) testBarrier(t, 5, func() *clientv3.Client { return clus.RandClient() }) } diff --git a/tests/integration/clientv3/experimental/recipes/v3_double_barrier_test.go b/tests/integration/clientv3/experimental/recipes/v3_double_barrier_test.go index e886b90b4..d9ffb9e82 100644 --- a/tests/integration/clientv3/experimental/recipes/v3_double_barrier_test.go +++ b/tests/integration/clientv3/experimental/recipes/v3_double_barrier_test.go @@ -26,7 +26,7 @@ import ( func TestDoubleBarrier(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) waiters := 10 @@ -100,7 +100,7 @@ func TestDoubleBarrier(t *testing.T) { func TestDoubleBarrierFailover(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) waiters := 10 diff --git a/tests/integration/clientv3/experimental/recipes/v3_lock_test.go b/tests/integration/clientv3/experimental/recipes/v3_lock_test.go index cc873adf2..ee130679b 100644 --- a/tests/integration/clientv3/experimental/recipes/v3_lock_test.go +++ b/tests/integration/clientv3/experimental/recipes/v3_lock_test.go @@ -30,7 +30,7 @@ import ( func TestMutexLockSingleNode(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) var clients []*clientv3.Client @@ -41,7 +41,7 @@ func TestMutexLockSingleNode(t *testing.T) { func TestMutexLockMultiNode(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) var clients []*clientv3.Client @@ -94,7 +94,7 @@ func testMutexLock(t *testing.T, waiters int, chooseClient func() *clientv3.Clie func TestMutexTryLockSingleNode(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) var clients []*clientv3.Client @@ -104,7 +104,7 @@ func TestMutexTryLockSingleNode(t *testing.T) { func TestMutexTryLockMultiNode(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) var clients []*clientv3.Client @@ -165,7 +165,7 @@ func testMutexTryLock(t *testing.T, lockers int, chooseClient func() *clientv3.C func TestMutexSessionRelock(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) session, err := concurrency.NewSession(clus.RandClient()) if err != nil { @@ -189,7 +189,7 @@ func TestMutexSessionRelock(t *testing.T) { func TestMutexWaitsOnCurrentHolder(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) cctx := context.Background() @@ -297,7 +297,7 @@ func TestMutexWaitsOnCurrentHolder(t *testing.T) { func BenchmarkMutex4Waiters(b *testing.B) { integration2.BeforeTest(b) // XXX switch tests to use TB interface - clus := integration2.NewClusterV3(nil, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(nil, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(nil) for i := 0; i < b.N; i++ { testMutexLock(nil, 4, func() *clientv3.Client { return clus.RandClient() }) @@ -306,14 +306,14 @@ func BenchmarkMutex4Waiters(b *testing.B) { func TestRWMutexSingleNode(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) testRWMutex(t, 5, func() *clientv3.Client { return clus.Client(0) }) } func TestRWMutexMultiNode(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) testRWMutex(t, 5, func() *clientv3.Client { return clus.RandClient() }) } diff --git a/tests/integration/clientv3/experimental/recipes/v3_queue_test.go b/tests/integration/clientv3/experimental/recipes/v3_queue_test.go index 112d55e1f..22d01545e 100644 --- a/tests/integration/clientv3/experimental/recipes/v3_queue_test.go +++ b/tests/integration/clientv3/experimental/recipes/v3_queue_test.go @@ -33,7 +33,7 @@ const ( func TestQueueOneReaderOneWriter(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) done := make(chan struct{}) @@ -81,7 +81,7 @@ func BenchmarkQueue(b *testing.B) { integration2.BeforeTest(b) // XXX switch tests to use TB interface - clus := integration2.NewClusterV3(nil, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(nil, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(nil) for i := 0; i < b.N; i++ { testQueueNReaderMWriter(nil, manyQueueClients, manyQueueClients) @@ -92,7 +92,7 @@ func BenchmarkQueue(b *testing.B) { func TestPrQueueOneReaderOneWriter(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) // write out five items with random priority @@ -126,7 +126,7 @@ func TestPrQueueOneReaderOneWriter(t *testing.T) { func TestPrQueueManyReaderManyWriter(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) rqs := newPriorityQueues(clus, manyQueueClients) wqs := newPriorityQueues(clus, manyQueueClients) @@ -138,7 +138,7 @@ func BenchmarkPrQueueOneReaderOneWriter(b *testing.B) { integration2.BeforeTest(b) // XXX switch tests to use TB interface - clus := integration2.NewClusterV3(nil, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(nil, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(nil) rqs := newPriorityQueues(clus, 1) wqs := newPriorityQueues(clus, 1) @@ -149,12 +149,12 @@ func BenchmarkPrQueueOneReaderOneWriter(b *testing.B) { func testQueueNReaderMWriter(t *testing.T, n int, m int) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) testReadersWriters(t, newQueues(clus, n), newQueues(clus, m)) } -func newQueues(clus *integration2.ClusterV3, n int) (qs []testQueue) { +func newQueues(clus *integration2.Cluster, n int) (qs []testQueue) { for i := 0; i < n; i++ { etcdc := clus.RandClient() qs = append(qs, recipe.NewQueue(etcdc, "q")) @@ -162,7 +162,7 @@ func newQueues(clus *integration2.ClusterV3, n int) (qs []testQueue) { return qs } -func newPriorityQueues(clus *integration2.ClusterV3, n int) (qs []testQueue) { +func newPriorityQueues(clus *integration2.Cluster, n int) (qs []testQueue) { for i := 0; i < n; i++ { etcdc := clus.RandClient() q := &flatPriorityQueue{recipe.NewPriorityQueue(etcdc, "prq")} diff --git a/tests/integration/clientv3/kv_test.go b/tests/integration/clientv3/kv_test.go index 43cd50fc9..27c7d9329 100644 --- a/tests/integration/clientv3/kv_test.go +++ b/tests/integration/clientv3/kv_test.go @@ -42,7 +42,7 @@ func TestKVPutError(t *testing.T) { maxReqBytes = 1.5 * 1024 * 1024 // hard coded max in v3_server.go quota = int64(int(maxReqBytes*1.2) + 8*os.Getpagesize()) // make sure we have enough overhead in backend quota. See discussion in #6486. ) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, QuotaBackendBytes: quota, ClientMaxCallSendMsgSize: 100 * 1024 * 1024}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, QuotaBackendBytes: quota, ClientMaxCallSendMsgSize: 100 * 1024 * 1024}) defer clus.Terminate(t) kv := clus.RandClient() @@ -74,7 +74,7 @@ func TestKVPutError(t *testing.T) { func TestKVPut(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) lapi := clus.RandClient() @@ -119,7 +119,7 @@ func TestKVPut(t *testing.T) { func TestKVPutWithIgnoreValue(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) kv := clus.RandClient() @@ -152,7 +152,7 @@ func TestKVPutWithIgnoreValue(t *testing.T) { func TestKVPutWithIgnoreLease(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) kv := clus.RandClient() @@ -191,7 +191,7 @@ func TestKVPutWithIgnoreLease(t *testing.T) { func TestKVPutWithRequireLeader(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) clus.Members[1].Stop(t) @@ -237,7 +237,7 @@ func TestKVPutWithRequireLeader(t *testing.T) { func TestKVRange(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) kv := clus.RandClient() @@ -466,7 +466,7 @@ func TestKVRange(t *testing.T) { func TestKVGetErrConnClosed(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.Client(0) @@ -495,7 +495,7 @@ func TestKVGetErrConnClosed(t *testing.T) { func TestKVNewAfterClose(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.Client(0) @@ -522,7 +522,7 @@ func TestKVNewAfterClose(t *testing.T) { func TestKVDeleteRange(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) kv := clus.RandClient() @@ -594,7 +594,7 @@ func TestKVDeleteRange(t *testing.T) { func TestKVDelete(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) kv := clus.RandClient() @@ -626,7 +626,7 @@ func TestKVDelete(t *testing.T) { func TestKVCompactError(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) kv := clus.RandClient() @@ -656,7 +656,7 @@ func TestKVCompactError(t *testing.T) { func TestKVCompact(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) kv := clus.RandClient() @@ -712,7 +712,7 @@ func TestKVGetRetry(t *testing.T) { integration2.BeforeTest(t) clusterSize := 3 - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: clusterSize, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: clusterSize, UseBridge: true}) defer clus.Terminate(t) // because killing leader and following election @@ -765,7 +765,7 @@ func TestKVGetRetry(t *testing.T) { func TestKVPutFailGetRetry(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) kv := clus.Client(0) @@ -805,7 +805,7 @@ func TestKVPutFailGetRetry(t *testing.T) { func TestKVGetCancel(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) oldconn := clus.Client(0).ActiveConnection() @@ -828,7 +828,7 @@ func TestKVGetCancel(t *testing.T) { func TestKVGetStoppedServerAndClose(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.Client(0) @@ -846,7 +846,7 @@ func TestKVGetStoppedServerAndClose(t *testing.T) { func TestKVPutStoppedServerAndClose(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.Client(0) @@ -876,7 +876,7 @@ func TestKVPutStoppedServerAndClose(t *testing.T) { // in the presence of network errors. func TestKVPutAtMostOnce(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) if _, err := clus.Client(0).Put(context.TODO(), "k", "1"); err != nil { @@ -970,7 +970,7 @@ func TestKVLargeRequests(t *testing.T) { }, } for i, test := range tests { - clus := integration2.NewClusterV3(t, + clus := integration2.NewCluster(t, &integration2.ClusterConfig{ Size: 1, MaxRequestBytes: test.maxRequestBytesServer, @@ -1005,7 +1005,7 @@ func TestKVLargeRequests(t *testing.T) { func TestKVForLearner(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) // we have to add and launch learner member after initial cluster was created, because @@ -1084,7 +1084,7 @@ func TestKVForLearner(t *testing.T) { func TestBalancerSupportLearner(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) // we have to add and launch learner member after initial cluster was created, because diff --git a/tests/integration/clientv3/lease/lease_test.go b/tests/integration/clientv3/lease/lease_test.go index d06cffe05..7dc6bbf4f 100644 --- a/tests/integration/clientv3/lease/lease_test.go +++ b/tests/integration/clientv3/lease/lease_test.go @@ -32,7 +32,7 @@ import ( func TestLeaseNotFoundError(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) kv := clus.RandClient() @@ -46,7 +46,7 @@ func TestLeaseNotFoundError(t *testing.T) { func TestLeaseGrant(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) lapi := clus.RandClient() @@ -72,7 +72,7 @@ func TestLeaseGrant(t *testing.T) { func TestLeaseRevoke(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) lapi := clus.RandClient() @@ -98,7 +98,7 @@ func TestLeaseRevoke(t *testing.T) { func TestLeaseKeepAliveOnce(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) lapi := clus.RandClient() @@ -122,7 +122,7 @@ func TestLeaseKeepAliveOnce(t *testing.T) { func TestLeaseKeepAlive(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) lapi := clus.Client(0) @@ -162,7 +162,7 @@ func TestLeaseKeepAlive(t *testing.T) { func TestLeaseKeepAliveOneSecond(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.Client(0) @@ -190,7 +190,7 @@ func TestLeaseKeepAliveHandleFailure(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) // TODO: change this line to get a cluster client @@ -245,7 +245,7 @@ type leaseCh struct { func TestLeaseKeepAliveNotFound(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.RandClient() @@ -278,7 +278,7 @@ func TestLeaseKeepAliveNotFound(t *testing.T) { func TestLeaseGrantErrConnClosed(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.Client(0) @@ -310,7 +310,7 @@ func TestLeaseGrantErrConnClosed(t *testing.T) { func TestLeaseKeepAliveFullResponseQueue(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) lapi := clus.Client(0) @@ -350,7 +350,7 @@ func TestLeaseKeepAliveFullResponseQueue(t *testing.T) { func TestLeaseGrantNewAfterClose(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.Client(0) @@ -377,7 +377,7 @@ func TestLeaseGrantNewAfterClose(t *testing.T) { func TestLeaseRevokeNewAfterClose(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.Client(0) @@ -416,7 +416,7 @@ func TestLeaseRevokeNewAfterClose(t *testing.T) { func TestLeaseKeepAliveCloseAfterDisconnectRevoke(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) cli := clus.Client(0) @@ -462,7 +462,7 @@ func TestLeaseKeepAliveCloseAfterDisconnectRevoke(t *testing.T) { func TestLeaseKeepAliveInitTimeout(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) cli := clus.Client(0) @@ -495,7 +495,7 @@ func TestLeaseKeepAliveInitTimeout(t *testing.T) { func TestLeaseKeepAliveTTLTimeout(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) cli := clus.Client(0) @@ -530,7 +530,7 @@ func TestLeaseKeepAliveTTLTimeout(t *testing.T) { func TestLeaseTimeToLive(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) c := clus.RandClient() @@ -588,7 +588,7 @@ func TestLeaseTimeToLive(t *testing.T) { func TestLeaseTimeToLiveLeaseNotFound(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.RandClient() @@ -623,7 +623,7 @@ func TestLeaseTimeToLiveLeaseNotFound(t *testing.T) { func TestLeaseLeases(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.RandClient() @@ -656,7 +656,7 @@ func TestLeaseLeases(t *testing.T) { func TestLeaseRenewLostQuorum(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) cli := clus.Client(0) @@ -704,7 +704,7 @@ func TestLeaseRenewLostQuorum(t *testing.T) { func TestLeaseKeepAliveLoopExit(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) ctx := context.Background() @@ -728,7 +728,7 @@ func TestLeaseKeepAliveLoopExit(t *testing.T) { // transient cluster failure. func TestV3LeaseFailureOverlap(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 2, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 2, UseBridge: true}) defer clus.Terminate(t) numReqs := 5 @@ -782,7 +782,7 @@ func TestV3LeaseFailureOverlap(t *testing.T) { func TestLeaseWithRequireLeader(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 2, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 2, UseBridge: true}) defer clus.Terminate(t) c := clus.Client(0) diff --git a/tests/integration/clientv3/lease/leasing_test.go b/tests/integration/clientv3/lease/leasing_test.go index 60bd02719..55a2a4bb8 100644 --- a/tests/integration/clientv3/lease/leasing_test.go +++ b/tests/integration/clientv3/lease/leasing_test.go @@ -34,7 +34,7 @@ import ( func TestLeasingPutGet(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) lKV1, closeLKV1, err := leasing.NewKV(clus.Client(0), "foo/") @@ -92,7 +92,7 @@ func TestLeasingPutGet(t *testing.T) { // TestLeasingInterval checks the leasing KV fetches key intervals. func TestLeasingInterval(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -131,7 +131,7 @@ func TestLeasingInterval(t *testing.T) { // TestLeasingPutInvalidateNew checks the leasing KV updates its cache on a Put to a new key. func TestLeasingPutInvalidateNew(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -161,7 +161,7 @@ func TestLeasingPutInvalidateNew(t *testing.T) { // TestLeasingPutInvalidateExisting checks the leasing KV updates its cache on a Put to an existing key. func TestLeasingPutInvalidateExisting(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) if _, err := clus.Client(0).Put(context.TODO(), "k", "abc"); err != nil { @@ -195,7 +195,7 @@ func TestLeasingPutInvalidateExisting(t *testing.T) { // TestLeasingGetNoLeaseTTL checks a key with a TTL is not leased. func TestLeasingGetNoLeaseTTL(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -224,7 +224,7 @@ func TestLeasingGetNoLeaseTTL(t *testing.T) { // when the etcd cluster is partitioned. func TestLeasingGetSerializable(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 2, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 2, UseBridge: true}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -264,7 +264,7 @@ func TestLeasingGetSerializable(t *testing.T) { // TestLeasingPrevKey checks the cache respects WithPrevKV on puts. func TestLeasingPrevKey(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 2}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 2}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -290,7 +290,7 @@ func TestLeasingPrevKey(t *testing.T) { // TestLeasingRevGet checks the cache respects Get by Revision. func TestLeasingRevGet(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -326,7 +326,7 @@ func TestLeasingRevGet(t *testing.T) { // TestLeasingGetWithOpts checks options that can be served through the cache do not depend on the server. func TestLeasingGetWithOpts(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -371,7 +371,7 @@ func TestLeasingGetWithOpts(t *testing.T) { // the recently put data. func TestLeasingConcurrentPut(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -418,7 +418,7 @@ func TestLeasingConcurrentPut(t *testing.T) { func TestLeasingDisconnectedGet(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -447,7 +447,7 @@ func TestLeasingDisconnectedGet(t *testing.T) { func TestLeasingDeleteOwner(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -481,7 +481,7 @@ func TestLeasingDeleteOwner(t *testing.T) { func TestLeasingDeleteNonOwner(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv1, closeLKV1, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -516,7 +516,7 @@ func TestLeasingDeleteNonOwner(t *testing.T) { func TestLeasingOverwriteResponse(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -550,7 +550,7 @@ func TestLeasingOverwriteResponse(t *testing.T) { func TestLeasingOwnerPutResponse(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -588,7 +588,7 @@ func TestLeasingOwnerPutResponse(t *testing.T) { func TestLeasingTxnOwnerGetRange(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -617,7 +617,7 @@ func TestLeasingTxnOwnerGetRange(t *testing.T) { func TestLeasingTxnOwnerGet(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) client := clus.Client(0) @@ -703,7 +703,7 @@ func TestLeasingTxnOwnerGet(t *testing.T) { func TestLeasingTxnOwnerDeleteRange(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -742,7 +742,7 @@ func TestLeasingTxnOwnerDeleteRange(t *testing.T) { func TestLeasingTxnOwnerDelete(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -773,7 +773,7 @@ func TestLeasingTxnOwnerDelete(t *testing.T) { func TestLeasingTxnOwnerIf(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -867,7 +867,7 @@ func TestLeasingTxnOwnerIf(t *testing.T) { func TestLeasingTxnCancel(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) lkv1, closeLKV1, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -901,7 +901,7 @@ func TestLeasingTxnCancel(t *testing.T) { func TestLeasingTxnNonOwnerPut(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -979,7 +979,7 @@ func TestLeasingTxnNonOwnerPut(t *testing.T) { // issues a random If/{Then,Else} transaction on those keys to one client. func TestLeasingTxnRandIfThenOrElse(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv1, closeLKV1, err1 := leasing.NewKV(clus.Client(0), "pfx/") @@ -1085,7 +1085,7 @@ func TestLeasingTxnRandIfThenOrElse(t *testing.T) { func TestLeasingOwnerPutError(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -1106,7 +1106,7 @@ func TestLeasingOwnerPutError(t *testing.T) { func TestLeasingOwnerDeleteError(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -1127,7 +1127,7 @@ func TestLeasingOwnerDeleteError(t *testing.T) { func TestLeasingNonOwnerPutError(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/") @@ -1152,7 +1152,7 @@ func TestLeasingOwnerDeleteFrom(t *testing.T) { func testLeasingOwnerDelete(t *testing.T, del clientv3.Op) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "0/") @@ -1201,7 +1201,7 @@ func testLeasingOwnerDelete(t *testing.T, del clientv3.Op) { func TestLeasingDeleteRangeBounds(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) delkv, closeDelKV, err := leasing.NewKV(clus.Client(0), "0/") @@ -1259,7 +1259,7 @@ func TestLeaseDeleteRangeContendDel(t *testing.T) { func testLeasingDeleteRangeContend(t *testing.T, op clientv3.Op) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) delkv, closeDelKV, err := leasing.NewKV(clus.Client(0), "0/") @@ -1317,7 +1317,7 @@ func testLeasingDeleteRangeContend(t *testing.T, op clientv3.Op) { func TestLeasingPutGetDeleteConcurrent(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkvs := make([]clientv3.KV, 16) @@ -1376,7 +1376,7 @@ func TestLeasingPutGetDeleteConcurrent(t *testing.T) { // disconnected when trying to submit revoke txn. func TestLeasingReconnectOwnerRevoke(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) lkv1, closeLKV1, err1 := leasing.NewKV(clus.Client(0), "foo/") @@ -1437,7 +1437,7 @@ func TestLeasingReconnectOwnerRevoke(t *testing.T) { // disconnected and the watch is compacted. func TestLeasingReconnectOwnerRevokeCompact(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) lkv1, closeLKV1, err1 := leasing.NewKV(clus.Client(0), "foo/") @@ -1490,7 +1490,7 @@ func TestLeasingReconnectOwnerRevokeCompact(t *testing.T) { // not cause inconsistency between the server and the client. func TestLeasingReconnectOwnerConsistency(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/") @@ -1564,7 +1564,7 @@ func TestLeasingReconnectOwnerConsistency(t *testing.T) { func TestLeasingTxnAtomicCache(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/") @@ -1650,7 +1650,7 @@ func TestLeasingTxnAtomicCache(t *testing.T) { // TestLeasingReconnectTxn checks that Txn is resilient to disconnects. func TestLeasingReconnectTxn(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/") @@ -1686,7 +1686,7 @@ func TestLeasingReconnectTxn(t *testing.T) { // not cause inconsistency between the server and the client. func TestLeasingReconnectNonOwnerGet(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/") @@ -1737,7 +1737,7 @@ func TestLeasingReconnectNonOwnerGet(t *testing.T) { func TestLeasingTxnRangeCmp(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/") @@ -1772,7 +1772,7 @@ func TestLeasingTxnRangeCmp(t *testing.T) { func TestLeasingDo(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/") @@ -1814,7 +1814,7 @@ func TestLeasingDo(t *testing.T) { func TestLeasingTxnOwnerPutBranch(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/") @@ -1908,7 +1908,7 @@ func randCmps(pfx string, dat []*clientv3.PutResponse) (cmps []clientv3.Cmp, the func TestLeasingSessionExpire(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/", concurrency.WithTTL(1)) @@ -1984,7 +1984,7 @@ func TestLeasingSessionExpireCancel(t *testing.T) { for i := range tests { t.Run(fmt.Sprintf("test %d", i), func(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/", concurrency.WithTTL(1)) diff --git a/tests/integration/clientv3/maintenance_test.go b/tests/integration/clientv3/maintenance_test.go index 45bbe8c5c..ce1e629c0 100644 --- a/tests/integration/clientv3/maintenance_test.go +++ b/tests/integration/clientv3/maintenance_test.go @@ -39,7 +39,7 @@ import ( func TestMaintenanceHashKV(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) for i := 0; i < 3; i++ { @@ -72,7 +72,7 @@ func TestMaintenanceHashKV(t *testing.T) { func TestMaintenanceMoveLeader(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) oldLeadIdx := clus.WaitLeader(t) @@ -103,7 +103,7 @@ func TestMaintenanceMoveLeader(t *testing.T) { func TestMaintenanceSnapshotCancel(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) // reading snapshot with canceled context should error out @@ -146,7 +146,7 @@ func TestMaintenanceSnapshotTimeout(t *testing.T) { func testMaintenanceSnapshotTimeout(t *testing.T, snapshot func(context.Context, *clientv3.Client) (io.ReadCloser, error)) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) // reading snapshot with deadline exceeded should error out @@ -191,7 +191,7 @@ func TestMaintenanceSnapshotErrorInflight(t *testing.T) { func testMaintenanceSnapshotErrorInflight(t *testing.T, snapshot func(context.Context, *clientv3.Client) (io.ReadCloser, error)) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) // take about 1-second to read snapshot @@ -249,7 +249,7 @@ func TestMaintenanceSnapshotWithVersionVersion(t *testing.T) { integration2.BeforeTest(t) // Set SnapshotCount to 1 to force raft snapshot to ensure that storage version is set - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, SnapshotCount: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, SnapshotCount: 1}) defer clus.Terminate(t) // Put some keys to ensure that wal snapshot is triggered @@ -271,7 +271,7 @@ func TestMaintenanceSnapshotWithVersionVersion(t *testing.T) { func TestMaintenanceStatus(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) clus.WaitLeader(t) diff --git a/tests/integration/clientv3/metrics_test.go b/tests/integration/clientv3/metrics_test.go index 0c3db3dca..26b24b295 100644 --- a/tests/integration/clientv3/metrics_test.go +++ b/tests/integration/clientv3/metrics_test.go @@ -70,7 +70,7 @@ func TestV3ClientMetrics(t *testing.T) { url := "unix://" + addr + "/metrics" - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) cfg := clientv3.Config{ diff --git a/tests/integration/clientv3/mirror_test.go b/tests/integration/clientv3/mirror_test.go index 36dc71dcc..f21551bbd 100644 --- a/tests/integration/clientv3/mirror_test.go +++ b/tests/integration/clientv3/mirror_test.go @@ -30,7 +30,7 @@ import ( func TestMirrorSync(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) c := clus.Client(0) @@ -74,7 +74,7 @@ func TestMirrorSync(t *testing.T) { func TestMirrorSyncBase(t *testing.T) { integration2.BeforeTest(t) - cluster := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + cluster := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer cluster.Terminate(t) cli := cluster.Client(0) diff --git a/tests/integration/clientv3/namespace_test.go b/tests/integration/clientv3/namespace_test.go index 4b76ca34e..c214a17dc 100644 --- a/tests/integration/clientv3/namespace_test.go +++ b/tests/integration/clientv3/namespace_test.go @@ -28,7 +28,7 @@ import ( func TestNamespacePutGet(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) c := clus.Client(0) @@ -57,7 +57,7 @@ func TestNamespacePutGet(t *testing.T) { func TestNamespaceWatch(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) c := clus.Client(0) diff --git a/tests/integration/clientv3/naming/endpoints_test.go b/tests/integration/clientv3/naming/endpoints_test.go index d5cbbf8cf..2225a804b 100644 --- a/tests/integration/clientv3/naming/endpoints_test.go +++ b/tests/integration/clientv3/naming/endpoints_test.go @@ -27,7 +27,7 @@ import ( func TestEndpointManager(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) em, err := endpoints.NewManager(clus.RandClient(), "foo") @@ -89,7 +89,7 @@ func TestEndpointManager(t *testing.T) { func TestEndpointManagerAtomicity(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) c := clus.RandClient() @@ -131,7 +131,7 @@ func TestEndpointManagerAtomicity(t *testing.T) { func TestEndpointManagerCRUD(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) em, err := endpoints.NewManager(clus.RandClient(), "foo") diff --git a/tests/integration/clientv3/naming/resolver_test.go b/tests/integration/clientv3/naming/resolver_test.go index 445ebca86..1f9d2a5fc 100644 --- a/tests/integration/clientv3/naming/resolver_test.go +++ b/tests/integration/clientv3/naming/resolver_test.go @@ -47,7 +47,7 @@ func TestEtcdGrpcResolver(t *testing.T) { } defer s2.Stop() - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) em, err := endpoints.NewManager(clus.Client(0), "foo") diff --git a/tests/integration/clientv3/ordering_kv_test.go b/tests/integration/clientv3/ordering_kv_test.go index f213d497c..93b409303 100644 --- a/tests/integration/clientv3/ordering_kv_test.go +++ b/tests/integration/clientv3/ordering_kv_test.go @@ -30,7 +30,7 @@ func TestDetectKvOrderViolation(t *testing.T) { var errOrderViolation = errors.New("DetectedOrderViolation") integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) cfg := clientv3.Config{ @@ -97,7 +97,7 @@ func TestDetectTxnOrderViolation(t *testing.T) { var errOrderViolation = errors.New("DetectedOrderViolation") integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) cfg := clientv3.Config{ diff --git a/tests/integration/clientv3/ordering_util_test.go b/tests/integration/clientv3/ordering_util_test.go index 2fb0c1413..9e2f5ae6d 100644 --- a/tests/integration/clientv3/ordering_util_test.go +++ b/tests/integration/clientv3/ordering_util_test.go @@ -26,7 +26,7 @@ import ( func TestEndpointSwitchResolvesViolation(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) eps := []string{ clus.Members[0].GRPCURL(), @@ -82,7 +82,7 @@ func TestEndpointSwitchResolvesViolation(t *testing.T) { func TestUnresolvableOrderViolation(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 5, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 5, UseBridge: true}) defer clus.Terminate(t) cfg := clientv3.Config{ Endpoints: []string{ diff --git a/tests/integration/clientv3/role_test.go b/tests/integration/clientv3/role_test.go index a10e6f648..38d74d90e 100644 --- a/tests/integration/clientv3/role_test.go +++ b/tests/integration/clientv3/role_test.go @@ -25,7 +25,7 @@ import ( func TestRoleError(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) authapi := clus.RandClient() diff --git a/tests/integration/clientv3/txn_test.go b/tests/integration/clientv3/txn_test.go index b23573605..b2fa2d997 100644 --- a/tests/integration/clientv3/txn_test.go +++ b/tests/integration/clientv3/txn_test.go @@ -29,7 +29,7 @@ import ( func TestTxnError(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) kv := clus.RandClient() @@ -53,7 +53,7 @@ func TestTxnError(t *testing.T) { func TestTxnWriteFail(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) kv := clus.Client(0) @@ -103,7 +103,7 @@ func TestTxnReadRetry(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) kv := clus.Client(0) @@ -142,7 +142,7 @@ func TestTxnReadRetry(t *testing.T) { func TestTxnSuccess(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) kv := clus.Client(0) @@ -165,7 +165,7 @@ func TestTxnSuccess(t *testing.T) { func TestTxnCompareRange(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) kv := clus.Client(0) @@ -192,7 +192,7 @@ func TestTxnCompareRange(t *testing.T) { func TestTxnNested(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) kv := clus.Client(0) diff --git a/tests/integration/clientv3/user_test.go b/tests/integration/clientv3/user_test.go index f0fe73a73..1cd700fde 100644 --- a/tests/integration/clientv3/user_test.go +++ b/tests/integration/clientv3/user_test.go @@ -28,7 +28,7 @@ import ( func TestUserError(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) authapi := clus.RandClient() @@ -57,7 +57,7 @@ func TestUserError(t *testing.T) { func TestUserErrorAuth(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) authapi := clus.RandClient() @@ -114,7 +114,7 @@ func authSetupRoot(t *testing.T, auth clientv3.Auth) { func TestGetTokenWithoutAuth(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 2}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 2}) defer clus.Terminate(t) authapi := clus.RandClient() diff --git a/tests/integration/clientv3/watch_fragment_test.go b/tests/integration/clientv3/watch_fragment_test.go index ab2367932..51418d82f 100644 --- a/tests/integration/clientv3/watch_fragment_test.go +++ b/tests/integration/clientv3/watch_fragment_test.go @@ -73,7 +73,7 @@ func testWatchFragment(t *testing.T, fragment, exceedRecvLimit bool) { if exceedRecvLimit { cfg.ClientMaxCallRecvMsgSize = 1.5 * 1024 * 1024 } - clus := integration2.NewClusterV3(t, cfg) + clus := integration2.NewCluster(t, cfg) defer clus.Terminate(t) cli := clus.Client(0) diff --git a/tests/integration/clientv3/watch_test.go b/tests/integration/clientv3/watch_test.go index f7ca77039..1d472ecc5 100644 --- a/tests/integration/clientv3/watch_test.go +++ b/tests/integration/clientv3/watch_test.go @@ -36,7 +36,7 @@ import ( type watcherTest func(*testing.T, *watchctx) type watchctx struct { - clus *integration2.ClusterV3 + clus *integration2.Cluster w clientv3.Watcher kv clientv3.KV wclientMember int @@ -47,7 +47,7 @@ type watchctx struct { func runWatchTest(t *testing.T, f watcherTest) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) wclientMember := rand.Intn(3) @@ -348,7 +348,7 @@ func putAndWatch(t *testing.T, wctx *watchctx, key, val string) { func TestWatchResumeInitRev(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) cli := clus.Client(0) @@ -404,7 +404,7 @@ func TestWatchResumeInitRev(t *testing.T) { func TestWatchResumeCompacted(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) // create a waiting watcher at rev 1 @@ -491,7 +491,7 @@ func TestWatchResumeCompacted(t *testing.T) { func TestWatchCompactRevision(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) // set some keys @@ -540,7 +540,7 @@ func testWatchWithProgressNotify(t *testing.T, watchOnPut bool) { pi := 3 * time.Second defer func() { v3rpc.SetProgressReportInterval(oldpi) }() - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) wc := clus.RandClient() @@ -588,7 +588,7 @@ func TestConfigurableWatchProgressNotifyInterval(t *testing.T) { integration2.BeforeTest(t) progressInterval := 200 * time.Millisecond - clus := integration2.NewClusterV3(t, + clus := integration2.NewCluster(t, &integration2.ClusterConfig{ Size: 3, WatchProgressNotifyInterval: progressInterval, @@ -629,7 +629,7 @@ func TestWatchRequestProgress(t *testing.T) { watchTimeout := 3 * time.Second - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) wc := clus.RandClient() @@ -688,7 +688,7 @@ func TestWatchRequestProgress(t *testing.T) { func TestWatchEventType(t *testing.T) { integration2.BeforeTest(t) - cluster := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + cluster := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer cluster.Terminate(t) client := cluster.RandClient() @@ -762,7 +762,7 @@ func TestWatchEventType(t *testing.T) { func TestWatchErrConnClosed(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.Client(0) @@ -792,7 +792,7 @@ func TestWatchErrConnClosed(t *testing.T) { func TestWatchAfterClose(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.Client(0) @@ -820,7 +820,7 @@ func TestWatchAfterClose(t *testing.T) { func TestWatchWithRequireLeader(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 3}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3}) defer clus.Terminate(t) // Put a key for the non-require leader watch to read as an event. @@ -894,7 +894,7 @@ func TestWatchWithRequireLeader(t *testing.T) { func TestWatchWithFilter(t *testing.T) { integration2.BeforeTest(t) - cluster := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + cluster := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer cluster.Terminate(t) client := cluster.RandClient() @@ -933,7 +933,7 @@ func TestWatchWithFilter(t *testing.T) { func TestWatchWithCreatedNotification(t *testing.T) { integration2.BeforeTest(t) - cluster := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + cluster := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer cluster.Terminate(t) client := cluster.RandClient() @@ -955,7 +955,7 @@ func TestWatchWithCreatedNotification(t *testing.T) { func TestWatchWithCreatedNotificationDropConn(t *testing.T) { integration2.BeforeTest(t) - cluster := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) + cluster := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) defer cluster.Terminate(t) client := cluster.RandClient() @@ -984,7 +984,7 @@ func TestWatchWithCreatedNotificationDropConn(t *testing.T) { func TestWatchCancelOnServer(t *testing.T) { integration2.BeforeTest(t) - cluster := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + cluster := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer cluster.Terminate(t) client := cluster.RandClient() @@ -1050,20 +1050,20 @@ func TestWatchCancelOnServer(t *testing.T) { // 4. watcher client finishes tearing down stream on "ctx" // 5. w2 comes back canceled func TestWatchOverlapContextCancel(t *testing.T) { - f := func(clus *integration2.ClusterV3) {} + f := func(clus *integration2.Cluster) {} testWatchOverlapContextCancel(t, f) } func TestWatchOverlapDropConnContextCancel(t *testing.T) { - f := func(clus *integration2.ClusterV3) { + f := func(clus *integration2.Cluster) { clus.Members[0].Bridge().DropConnections() } testWatchOverlapContextCancel(t, f) } -func testWatchOverlapContextCancel(t *testing.T, f func(*integration2.ClusterV3)) { +func testWatchOverlapContextCancel(t *testing.T, f func(*integration2.Cluster)) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) n := 100 @@ -1124,7 +1124,7 @@ func testWatchOverlapContextCancel(t *testing.T, f func(*integration2.ClusterV3) // closing the client does not return a client closing error. func TestWatchCancelAndCloseClient(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.Client(0) ctx, cancel := context.WithCancel(context.Background()) @@ -1154,7 +1154,7 @@ func TestWatchCancelAndCloseClient(t *testing.T) { // then closes the watcher interface to ensure correct clean up. func TestWatchStressResumeClose(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) cli := clus.Client(0) @@ -1176,7 +1176,7 @@ func TestWatchStressResumeClose(t *testing.T) { // its grpc stream is disconnected / reconnecting. func TestWatchCancelDisconnected(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.Client(0) ctx, cancel := context.WithCancel(context.Background()) diff --git a/tests/integration/cluster_test.go b/tests/integration/cluster_test.go index 5a8730170..9cd65b40b 100644 --- a/tests/integration/cluster_test.go +++ b/tests/integration/cluster_test.go @@ -46,14 +46,14 @@ func TestClusterOf3(t *testing.T) { testCluster(t, 3) } func testCluster(t *testing.T, size int) { integration.BeforeTest(t) - c := integration.NewClusterV3(t, &integration.ClusterConfig{Size: size}) + c := integration.NewCluster(t, &integration.ClusterConfig{Size: size}) defer c.Terminate(t) clusterMustProgress(t, c.Members) } func TestTLSClusterOf3(t *testing.T) { integration.BeforeTest(t) - c := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3, PeerTLS: &integration.TestTLSInfo}) + c := integration.NewCluster(t, &integration.ClusterConfig{Size: 3, PeerTLS: &integration.TestTLSInfo}) defer c.Terminate(t) clusterMustProgress(t, c.Members) } @@ -62,7 +62,7 @@ func TestTLSClusterOf3(t *testing.T) { // authorities that don't issue dual-usage certificates. func TestTLSClusterOf3WithSpecificUsage(t *testing.T) { integration.BeforeTest(t) - c := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3, PeerTLS: &integration.TestTLSInfoWithSpecificUsage}) + c := integration.NewCluster(t, &integration.ClusterConfig{Size: 3, PeerTLS: &integration.TestTLSInfoWithSpecificUsage}) defer c.Terminate(t) clusterMustProgress(t, c.Members) } @@ -72,7 +72,7 @@ func TestClusterOf3UsingDiscovery(t *testing.T) { testClusterUsingDiscovery(t, 3 func testClusterUsingDiscovery(t *testing.T, size int) { integration.BeforeTest(t) - dc := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1, UseIP: true}) + dc := integration.NewCluster(t, &integration.ClusterConfig{Size: 1, UseIP: true}) defer dc.Terminate(t) // init discovery token space dcc := integration.MustNewHTTPClient(t, dc.URLs(), nil) @@ -83,14 +83,14 @@ func testClusterUsingDiscovery(t *testing.T, size int) { } cancel() - c := integration.NewClusterV3(t, &integration.ClusterConfig{Size: size, DiscoveryURL: dc.URL(0) + "/v2/keys"}) + c := integration.NewCluster(t, &integration.ClusterConfig{Size: size, DiscoveryURL: dc.URL(0) + "/v2/keys"}) defer c.Terminate(t) clusterMustProgress(t, c.Members) } func TestTLSClusterOf3UsingDiscovery(t *testing.T) { integration.BeforeTest(t) - dc := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1, UseIP: true}) + dc := integration.NewCluster(t, &integration.ClusterConfig{Size: 1, UseIP: true}) defer dc.Terminate(t) // init discovery token space dcc := integration.MustNewHTTPClient(t, dc.URLs(), nil) @@ -101,7 +101,7 @@ func TestTLSClusterOf3UsingDiscovery(t *testing.T) { } cancel() - c := integration.NewClusterV3(t, &integration.ClusterConfig{ + c := integration.NewCluster(t, &integration.ClusterConfig{ Size: 3, PeerTLS: &integration.TestTLSInfo, DiscoveryURL: dc.URL(0) + "/v2/keys"}, @@ -115,7 +115,7 @@ func TestDoubleClusterSizeOf3(t *testing.T) { testDoubleClusterSize(t, 3) } func testDoubleClusterSize(t *testing.T, size int) { integration.BeforeTest(t) - c := integration.NewClusterV3(t, &integration.ClusterConfig{Size: size}) + c := integration.NewCluster(t, &integration.ClusterConfig{Size: size}) defer c.Terminate(t) for i := 0; i < size; i++ { @@ -126,7 +126,7 @@ func testDoubleClusterSize(t *testing.T, size int) { func TestDoubleTLSClusterSizeOf3(t *testing.T) { integration.BeforeTest(t) - c := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1, PeerTLS: &integration.TestTLSInfo}) + c := integration.NewCluster(t, &integration.ClusterConfig{Size: 1, PeerTLS: &integration.TestTLSInfo}) defer c.Terminate(t) for i := 0; i < 3; i++ { @@ -140,7 +140,7 @@ func TestDecreaseClusterSizeOf5(t *testing.T) { testDecreaseClusterSize(t, 5) } func testDecreaseClusterSize(t *testing.T, size int) { integration.BeforeTest(t) - c := integration.NewClusterV3(t, &integration.ClusterConfig{Size: size}) + c := integration.NewCluster(t, &integration.ClusterConfig{Size: size}) defer c.Terminate(t) // TODO: remove the last but one member @@ -162,7 +162,7 @@ func testDecreaseClusterSize(t *testing.T, size int) { func TestForceNewCluster(t *testing.T) { integration.BeforeTest(t) - c := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3, UseBridge: true}) + c := integration.NewCluster(t, &integration.ClusterConfig{Size: 3, UseBridge: true}) defer c.Terminate(t) ctx, cancel := context.WithTimeout(context.Background(), integration.RequestTimeout) @@ -218,7 +218,7 @@ func TestForceNewCluster(t *testing.T) { func TestAddMemberAfterClusterFullRotation(t *testing.T) { integration.BeforeTest(t) - c := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + c := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer c.Terminate(t) // remove all the previous three members and add in three new members. @@ -239,7 +239,7 @@ func TestAddMemberAfterClusterFullRotation(t *testing.T) { // Ensure we can remove a member then add a new one back immediately. func TestIssue2681(t *testing.T) { integration.BeforeTest(t) - c := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 5}) + c := integration.NewCluster(t, &integration.ClusterConfig{Size: 5}) defer c.Terminate(t) c.MustRemoveMember(t, uint64(c.Members[4].Server.ID())) @@ -258,7 +258,7 @@ func TestIssue2746WithThree(t *testing.T) { testIssue2746(t, 3) } func testIssue2746(t *testing.T, members int) { integration.BeforeTest(t) - c := integration.NewClusterV3(t, &integration.ClusterConfig{Size: members, SnapshotCount: 10}) + c := integration.NewCluster(t, &integration.ClusterConfig{Size: members, SnapshotCount: 10}) defer c.Terminate(t) // force a snapshot @@ -278,7 +278,7 @@ func testIssue2746(t *testing.T, members int) { func TestIssue2904(t *testing.T) { integration.BeforeTest(t) // start 1-member Cluster to ensure member 0 is the leader of the Cluster. - c := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1, UseBridge: true}) + c := integration.NewCluster(t, &integration.ClusterConfig{Size: 1, UseBridge: true}) defer c.Terminate(t) c.AddMember(t) @@ -314,7 +314,7 @@ func TestIssue2904(t *testing.T) { func TestIssue3699(t *testing.T) { // start a Cluster of 3 nodes a, b, c integration.BeforeTest(t) - c := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3, UseBridge: true}) + c := integration.NewCluster(t, &integration.ClusterConfig{Size: 3, UseBridge: true}) defer c.Terminate(t) // make node a unavailable @@ -363,7 +363,7 @@ func TestIssue3699(t *testing.T) { // TestRejectUnhealthyAdd ensures an unhealthy cluster rejects adding members. func TestRejectUnhealthyAdd(t *testing.T) { integration.BeforeTest(t) - c := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3, UseBridge: true, StrictReconfigCheck: true}) + c := integration.NewCluster(t, &integration.ClusterConfig{Size: 3, UseBridge: true, StrictReconfigCheck: true}) defer c.Terminate(t) // make Cluster unhealthy and wait for downed peer @@ -403,7 +403,7 @@ func TestRejectUnhealthyAdd(t *testing.T) { // if quorum will be lost. func TestRejectUnhealthyRemove(t *testing.T) { integration.BeforeTest(t) - c := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 5, UseBridge: true, StrictReconfigCheck: true}) + c := integration.NewCluster(t, &integration.ClusterConfig{Size: 5, UseBridge: true, StrictReconfigCheck: true}) defer c.Terminate(t) // make cluster unhealthy and wait for downed peer; (3 up, 2 down) @@ -448,7 +448,7 @@ func TestRestartRemoved(t *testing.T) { integration.BeforeTest(t) // 1. start single-member Cluster - c := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1, StrictReconfigCheck: true, UseBridge: true}) + c := integration.NewCluster(t, &integration.ClusterConfig{Size: 1, StrictReconfigCheck: true, UseBridge: true}) defer c.Terminate(t) // 2. add a new member @@ -525,7 +525,7 @@ func clusterMustProgress(t *testing.T, members []*integration.Member) { func TestSpeedyTerminate(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3, UseBridge: true}) // Stop/Restart so requests will time out on lost leaders for i := 0; i < 3; i++ { clus.Members[i].Stop(t) diff --git a/tests/integration/grpc_test.go b/tests/integration/grpc_test.go index dfa9fadcc..9e30d4c43 100644 --- a/tests/integration/grpc_test.go +++ b/tests/integration/grpc_test.go @@ -102,7 +102,7 @@ func TestAuthority(t *testing.T) { UseIP: tc.useTCP, } cfg, tlsConfig := setupTLS(t, tc.useTLS, cfg) - clus := integration.NewClusterV3(t, &cfg) + clus := integration.NewCluster(t, &cfg) defer clus.Terminate(t) kv := setupClient(t, tc.clientURLPattern, clus, tlsConfig) @@ -132,7 +132,7 @@ func setupTLS(t *testing.T, useTLS bool, cfg integration.ClusterConfig) (integra return cfg, nil } -func setupClient(t *testing.T, endpointPattern string, clus *integration.ClusterV3, tlsConfig *tls.Config) *clientv3.Client { +func setupClient(t *testing.T, endpointPattern string, clus *integration.Cluster, tlsConfig *tls.Config) *clientv3.Client { t.Helper() endpoints := templateEndpoints(t, endpointPattern, clus) kv, err := clientv3.New(clientv3.Config{ @@ -147,7 +147,7 @@ func setupClient(t *testing.T, endpointPattern string, clus *integration.Cluster return kv } -func templateEndpoints(t *testing.T, pattern string, clus *integration.ClusterV3) []string { +func templateEndpoints(t *testing.T, pattern string, clus *integration.Cluster) []string { t.Helper() endpoints := []string{} for _, m := range clus.Members { @@ -181,7 +181,7 @@ func templateAuthority(t *testing.T, pattern string, m *integration.Member) stri return authority } -func assertAuthority(t *testing.T, expectedAuthority string, clus *integration.ClusterV3) { +func assertAuthority(t *testing.T, expectedAuthority string, clus *integration.Cluster) { t.Helper() requestsFound := 0 for _, m := range clus.Members { diff --git a/tests/integration/lazy_cluster.go b/tests/integration/lazy_cluster.go index e8ac1225d..02fc759dc 100644 --- a/tests/integration/lazy_cluster.go +++ b/tests/integration/lazy_cluster.go @@ -43,7 +43,7 @@ type LazyCluster interface { EndpointsV3() []string // Cluster - calls to this method might initialize the cluster. - Cluster() *integration.ClusterV3 + Cluster() *integration.Cluster // Transport - call to this method might initialize the cluster. Transport() *http.Transport @@ -55,7 +55,7 @@ type LazyCluster interface { type lazyCluster struct { cfg integration.ClusterConfig - cluster *integration.ClusterV3 + cluster *integration.Cluster transport *http.Transport once sync.Once tb testutil.TB @@ -82,7 +82,7 @@ func (lc *lazyCluster) mustLazyInit() { if err != nil { log.Fatal(err) } - lc.cluster = integration.NewClusterV3(lc.tb, &lc.cfg) + lc.cluster = integration.NewCluster(lc.tb, &lc.cfg) }) } @@ -106,7 +106,7 @@ func (lc *lazyCluster) EndpointsV3() []string { return lc.Cluster().Client(0).Endpoints() } -func (lc *lazyCluster) Cluster() *integration.ClusterV3 { +func (lc *lazyCluster) Cluster() *integration.Cluster { lc.mustLazyInit() return lc.cluster } diff --git a/tests/integration/member_test.go b/tests/integration/member_test.go index 0ac006e50..f358e771d 100644 --- a/tests/integration/member_test.go +++ b/tests/integration/member_test.go @@ -27,7 +27,7 @@ import ( func TestPauseMember(t *testing.T) { integration.BeforeTest(t) - c := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 5}) + c := integration.NewCluster(t, &integration.ClusterConfig{Size: 5}) defer c.Terminate(t) for i := 0; i < 5; i++ { @@ -44,7 +44,7 @@ func TestPauseMember(t *testing.T) { func TestRestartMember(t *testing.T) { integration.BeforeTest(t) - c := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3, UseBridge: true}) + c := integration.NewCluster(t, &integration.ClusterConfig{Size: 3, UseBridge: true}) defer c.Terminate(t) for i := 0; i < 3; i++ { @@ -65,7 +65,7 @@ func TestRestartMember(t *testing.T) { func TestLaunchDuplicateMemberShouldFail(t *testing.T) { integration.BeforeTest(t) size := 3 - c := integration.NewClusterV3(t, &integration.ClusterConfig{Size: size}) + c := integration.NewCluster(t, &integration.ClusterConfig{Size: size}) m := c.Members[0].Clone(t) var err error m.DataDir, err = os.MkdirTemp(t.TempDir(), "etcd") diff --git a/tests/integration/metrics_test.go b/tests/integration/metrics_test.go index 00efe8b1b..59ce0d1d3 100644 --- a/tests/integration/metrics_test.go +++ b/tests/integration/metrics_test.go @@ -31,7 +31,7 @@ import ( // TestMetricDbSizeBoot checks that the db size metric is set on boot. func TestMetricDbSizeBoot(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) v, err := clus.Members[0].Metric("etcd_debugging_mvcc_db_total_size_in_bytes") @@ -51,7 +51,7 @@ func TestMetricDbSizeDefrag(t *testing.T) { // testMetricDbSizeDefrag checks that the db size metric is set after defrag. func testMetricDbSizeDefrag(t *testing.T, name string) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) kvc := integration.ToGRPC(clus.Client(0)).KV @@ -165,7 +165,7 @@ func testMetricDbSizeDefrag(t *testing.T, name string) { func TestMetricQuotaBackendBytes(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) qs, err := clus.Members[0].Metric("etcd_server_quota_backend_bytes") @@ -183,7 +183,7 @@ func TestMetricQuotaBackendBytes(t *testing.T) { func TestMetricsHealth(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) tr, err := transport.NewTransport(transport.TLSInfo{}, 5*time.Second) diff --git a/tests/integration/network_partition_test.go b/tests/integration/network_partition_test.go index 9ea4e4534..b5d73a30a 100644 --- a/tests/integration/network_partition_test.go +++ b/tests/integration/network_partition_test.go @@ -25,7 +25,7 @@ import ( func TestNetworkPartition5MembersLeaderInMinority(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 5}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 5}) defer clus.Terminate(t) leadIndex := clus.WaitLeader(t) @@ -34,8 +34,8 @@ func TestNetworkPartition5MembersLeaderInMinority(t *testing.T) { minority := []int{leadIndex, (leadIndex + 1) % 5} majority := []int{(leadIndex + 2) % 5, (leadIndex + 3) % 5, (leadIndex + 4) % 5} - minorityMembers := getMembersByIndexSlice(clus.Cluster, minority) - majorityMembers := getMembersByIndexSlice(clus.Cluster, majority) + minorityMembers := getMembersByIndexSlice(clus, minority) + majorityMembers := getMembersByIndexSlice(clus, majority) // network partition (bi-directional) injectPartition(t, minorityMembers, majorityMembers) @@ -73,7 +73,7 @@ func TestNetworkPartition5MembersLeaderInMajority(t *testing.T) { func testNetworkPartition5MembersLeaderInMajority(t *testing.T) error { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 5}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 5}) defer clus.Terminate(t) leadIndex := clus.WaitLeader(t) @@ -82,8 +82,8 @@ func testNetworkPartition5MembersLeaderInMajority(t *testing.T) error { majority := []int{leadIndex, (leadIndex + 1) % 5, (leadIndex + 2) % 5} minority := []int{(leadIndex + 3) % 5, (leadIndex + 4) % 5} - majorityMembers := getMembersByIndexSlice(clus.Cluster, majority) - minorityMembers := getMembersByIndexSlice(clus.Cluster, minority) + majorityMembers := getMembersByIndexSlice(clus, majority) + minorityMembers := getMembersByIndexSlice(clus, minority) // network partition (bi-directional) injectPartition(t, majorityMembers, minorityMembers) @@ -112,7 +112,7 @@ func testNetworkPartition5MembersLeaderInMajority(t *testing.T) error { func TestNetworkPartition4Members(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 4}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 4}) defer clus.Terminate(t) leadIndex := clus.WaitLeader(t) @@ -121,8 +121,8 @@ func TestNetworkPartition4Members(t *testing.T) { groupA := []int{leadIndex, (leadIndex + 1) % 4} groupB := []int{(leadIndex + 2) % 4, (leadIndex + 3) % 4} - leaderPartition := getMembersByIndexSlice(clus.Cluster, groupA) - followerPartition := getMembersByIndexSlice(clus.Cluster, groupB) + leaderPartition := getMembersByIndexSlice(clus, groupA) + followerPartition := getMembersByIndexSlice(clus, groupB) // network partition (bi-directional) injectPartition(t, leaderPartition, followerPartition) diff --git a/tests/integration/proxy/grpcproxy/cluster_test.go b/tests/integration/proxy/grpcproxy/cluster_test.go index 22c6970c8..e78d2db66 100644 --- a/tests/integration/proxy/grpcproxy/cluster_test.go +++ b/tests/integration/proxy/grpcproxy/cluster_test.go @@ -33,7 +33,7 @@ import ( func TestClusterProxyMemberList(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) cts := newClusterProxyServer(zaptest.NewLogger(t), []string{clus.Members[0].GRPCURL()}, t) diff --git a/tests/integration/proxy/grpcproxy/kv_test.go b/tests/integration/proxy/grpcproxy/kv_test.go index c319c54dd..a1aacd8cb 100644 --- a/tests/integration/proxy/grpcproxy/kv_test.go +++ b/tests/integration/proxy/grpcproxy/kv_test.go @@ -30,7 +30,7 @@ import ( func TestKVProxyRange(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) kvts := newKVProxyServer([]string{clus.Members[0].GRPCURL()}, t) diff --git a/tests/integration/proxy/grpcproxy/register_test.go b/tests/integration/proxy/grpcproxy/register_test.go index d93000a5e..e4f8f004d 100644 --- a/tests/integration/proxy/grpcproxy/register_test.go +++ b/tests/integration/proxy/grpcproxy/register_test.go @@ -28,7 +28,7 @@ import ( func TestRegister(t *testing.T) { integration2.BeforeTest(t) - clus := integration2.NewClusterV3(t, &integration2.ClusterConfig{Size: 1}) + clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.Client(0) paddr := clus.Members[0].GRPCURL() diff --git a/tests/integration/v3_alarm_test.go b/tests/integration/v3_alarm_test.go index 08c0b4082..89dc26abd 100644 --- a/tests/integration/v3_alarm_test.go +++ b/tests/integration/v3_alarm_test.go @@ -36,7 +36,7 @@ func TestV3StorageQuotaApply(t *testing.T) { integration.BeforeTest(t) quotasize := int64(16 * os.Getpagesize()) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 2}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 2}) defer clus.Terminate(t) kvc1 := integration.ToGRPC(clus.Client(1)).KV @@ -139,7 +139,7 @@ func TestV3StorageQuotaApply(t *testing.T) { func TestV3AlarmDeactivate(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) kvc := integration.ToGRPC(clus.RandClient()).KV mt := integration.ToGRPC(clus.RandClient()).Maintenance @@ -172,7 +172,7 @@ func TestV3AlarmDeactivate(t *testing.T) { func TestV3CorruptAlarm(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) var wg sync.WaitGroup diff --git a/tests/integration/v3_auth_test.go b/tests/integration/v3_auth_test.go index 5d915a964..4f4e53e97 100644 --- a/tests/integration/v3_auth_test.go +++ b/tests/integration/v3_auth_test.go @@ -32,7 +32,7 @@ import ( // TestV3AuthEmptyUserGet ensures that a get with an empty user will return an empty user error. func TestV3AuthEmptyUserGet(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) ctx, cancel := context.WithTimeout(context.TODO(), 30*time.Second) @@ -51,7 +51,7 @@ func TestV3AuthEmptyUserGet(t *testing.T) { // given a valid token when authentication is disabled func TestV3AuthTokenWithDisable(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) authSetupRoot(t, integration.ToGRPC(clus.Client(0)).Auth) @@ -83,7 +83,7 @@ func TestV3AuthTokenWithDisable(t *testing.T) { func TestV3AuthRevision(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) api := integration.ToGRPC(clus.Client(0)) @@ -122,7 +122,7 @@ func TestV3AuthWithLeaseRevokeWithRootJWT(t *testing.T) { func testV3AuthWithLeaseRevokeWithRoot(t *testing.T, ccfg integration.ClusterConfig) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &ccfg) + clus := integration.NewCluster(t, &ccfg) defer clus.Terminate(t) api := integration.ToGRPC(clus.Client(0)) @@ -179,7 +179,7 @@ type user struct { func TestV3AuthWithLeaseRevoke(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) users := []user{ @@ -225,7 +225,7 @@ func TestV3AuthWithLeaseRevoke(t *testing.T) { func TestV3AuthWithLeaseAttach(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) users := []user{ @@ -337,7 +337,7 @@ func authSetupRoot(t *testing.T, auth pb.AuthClient) { func TestV3AuthNonAuthorizedRPCs(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) nonAuthedKV := clus.Client(0).KV @@ -360,7 +360,7 @@ func TestV3AuthNonAuthorizedRPCs(t *testing.T) { func TestV3AuthOldRevConcurrent(t *testing.T) { t.Skip() // TODO(jingyih): re-enable the test when #10408 is fixed. integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) authSetupRoot(t, integration.ToGRPC(clus.Client(0)).Auth) diff --git a/tests/integration/v3_election_test.go b/tests/integration/v3_election_test.go index 761afd1f4..e7f2ab000 100644 --- a/tests/integration/v3_election_test.go +++ b/tests/integration/v3_election_test.go @@ -28,7 +28,7 @@ import ( // TestElectionWait tests if followers can correctly wait for elections. func TestElectionWait(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) leaders := 3 @@ -110,7 +110,7 @@ func TestElectionWait(t *testing.T) { // TestElectionFailover tests that an election will func TestElectionFailover(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) cctx, cancel := context.WithCancel(context.TODO()) @@ -178,7 +178,7 @@ func TestElectionFailover(t *testing.T) { // with the same lock will Proclaim instead of deadlocking. func TestElectionSessionRecampaign(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.RandClient() @@ -211,7 +211,7 @@ func TestElectionSessionRecampaign(t *testing.T) { // func TestElectionOnPrefixOfExistingKey(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.RandClient() @@ -238,7 +238,7 @@ func TestElectionOnPrefixOfExistingKey(t *testing.T) { // leadership. func TestElectionOnSessionRestart(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.RandClient() @@ -285,7 +285,7 @@ func TestElectionOnSessionRestart(t *testing.T) { // a leader key with a modrev less than the compaction revision. func TestElectionObserveCompacted(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.Client(0) diff --git a/tests/integration/v3_grpc_inflight_test.go b/tests/integration/v3_grpc_inflight_test.go index b96bac45a..daf5c3bbe 100644 --- a/tests/integration/v3_grpc_inflight_test.go +++ b/tests/integration/v3_grpc_inflight_test.go @@ -32,7 +32,7 @@ import ( // does not panic the mvcc backend while defragment is running. func TestV3MaintenanceDefragmentInflightRange(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.RandClient() @@ -62,7 +62,7 @@ func TestV3MaintenanceDefragmentInflightRange(t *testing.T) { // See https://github.com/etcd-io/etcd/issues/7322 for more detail. func TestV3KVInflightRangeRequests(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) cli := clus.RandClient() diff --git a/tests/integration/v3_grpc_test.go b/tests/integration/v3_grpc_test.go index 54165c845..620cc5d9c 100644 --- a/tests/integration/v3_grpc_test.go +++ b/tests/integration/v3_grpc_test.go @@ -41,7 +41,7 @@ import ( // overwrites it, then checks that the change was applied. func TestV3PutOverwrite(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) kvc := integration.ToGRPC(clus.RandClient()).KV @@ -89,7 +89,7 @@ func TestV3PutOverwrite(t *testing.T) { // TestPutRestart checks if a put after an unrelated member restart succeeds func TestV3PutRestart(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) kvIdx := rand.Intn(3) @@ -121,7 +121,7 @@ func TestV3PutRestart(t *testing.T) { // TestV3CompactCurrentRev ensures keys are present when compacting on current revision. func TestV3CompactCurrentRev(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) kvc := integration.ToGRPC(clus.RandClient()).KV @@ -155,7 +155,7 @@ func TestV3CompactCurrentRev(t *testing.T) { // TestV3HashKV ensures that multiple calls of HashKV on same node return same hash and compact rev. func TestV3HashKV(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) kvc := integration.ToGRPC(clus.RandClient()).KV @@ -204,7 +204,7 @@ func TestV3HashKV(t *testing.T) { func TestV3TxnTooManyOps(t *testing.T) { integration.BeforeTest(t) maxTxnOps := uint(128) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3, MaxTxnOps: maxTxnOps}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3, MaxTxnOps: maxTxnOps}) defer clus.Terminate(t) kvc := integration.ToGRPC(clus.RandClient()).KV @@ -279,7 +279,7 @@ func TestV3TxnTooManyOps(t *testing.T) { func TestV3TxnDuplicateKeys(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) putreq := &pb.RequestOp{Request: &pb.RequestOp_RequestPut{RequestPut: &pb.PutRequest{Key: []byte("abc"), Value: []byte("def")}}} @@ -397,7 +397,7 @@ func TestV3TxnDuplicateKeys(t *testing.T) { // Testv3TxnRevision tests that the transaction header revision is set as expected. func TestV3TxnRevision(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) kvc := integration.ToGRPC(clus.RandClient()).KV @@ -448,7 +448,7 @@ func TestV3TxnRevision(t *testing.T) { // when compared to the Succeeded field in the txn response. func TestV3TxnCmpHeaderRev(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) kvc := integration.ToGRPC(clus.RandClient()).KV @@ -504,7 +504,7 @@ func TestV3TxnCmpHeaderRev(t *testing.T) { // TestV3TxnRangeCompare tests range comparisons in txns func TestV3TxnRangeCompare(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) // put keys, named by expected revision @@ -615,7 +615,7 @@ func TestV3TxnRangeCompare(t *testing.T) { // TestV3TxnNested tests nested txns follow paths as expected. func TestV3TxnNestedPath(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) kvc := integration.ToGRPC(clus.RandClient()).KV @@ -669,7 +669,7 @@ func TestV3TxnNestedPath(t *testing.T) { func TestV3PutIgnoreValue(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) kvc := integration.ToGRPC(clus.RandClient()).KV @@ -802,7 +802,7 @@ func TestV3PutIgnoreValue(t *testing.T) { func TestV3PutIgnoreLease(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) kvc := integration.ToGRPC(clus.RandClient()).KV @@ -941,7 +941,7 @@ func TestV3PutIgnoreLease(t *testing.T) { // TestV3PutMissingLease ensures that a Put on a key with a bogus lease fails. func TestV3PutMissingLease(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) kvc := integration.ToGRPC(clus.RandClient()).KV @@ -1069,7 +1069,7 @@ func TestV3DeleteRange(t *testing.T) { for i, tt := range tests { t.Run(tt.name, func(t *testing.T) { - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) kvc := integration.ToGRPC(clus.RandClient()).KV defer clus.Terminate(t) @@ -1124,7 +1124,7 @@ func TestV3DeleteRange(t *testing.T) { // TestV3TxnInvalidRange tests that invalid ranges are rejected in txns. func TestV3TxnInvalidRange(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) kvc := integration.ToGRPC(clus.RandClient()).KV @@ -1168,7 +1168,7 @@ func TestV3TxnInvalidRange(t *testing.T) { func TestV3TooLargeRequest(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) kvc := integration.ToGRPC(clus.RandClient()).KV @@ -1186,7 +1186,7 @@ func TestV3TooLargeRequest(t *testing.T) { // TestV3Hash tests hash. func TestV3Hash(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) cli := clus.RandClient() @@ -1211,7 +1211,7 @@ func TestV3Hash(t *testing.T) { // TestV3HashRestart ensures that hash stays the same after restart. func TestV3HashRestart(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) cli := clus.RandClient() @@ -1244,7 +1244,7 @@ func TestV3StorageQuotaAPI(t *testing.T) { integration.BeforeTest(t) quotasize := int64(16 * os.Getpagesize()) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3, UseBridge: true}) // Set a quota on one node clus.Members[0].QuotaBackendBytes = quotasize @@ -1508,7 +1508,7 @@ func TestV3RangeRequest(t *testing.T) { for i, tt := range tests { t.Run(tt.name, func(t *testing.T) { - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) for _, k := range tt.putKeys { kvc := integration.ToGRPC(clus.RandClient()).KV @@ -1554,7 +1554,7 @@ func TestV3RangeRequest(t *testing.T) { func TestTLSGRPCRejectInsecureClient(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3, ClientTLS: &integration.TestTLSInfo}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3, ClientTLS: &integration.TestTLSInfo}) defer clus.Terminate(t) // nil out TLS field so client will use an insecure connection @@ -1588,7 +1588,7 @@ func TestTLSGRPCRejectInsecureClient(t *testing.T) { func TestTLSGRPCRejectSecureClient(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) clus.Members[0].ClientTLSInfo = &integration.TestTLSInfo @@ -1607,7 +1607,7 @@ func TestTLSGRPCRejectSecureClient(t *testing.T) { func TestTLSGRPCAcceptSecureAll(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3, ClientTLS: &integration.TestTLSInfo}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3, ClientTLS: &integration.TestTLSInfo}) defer clus.Terminate(t) client, err := integration.NewClientV3(clus.Members[0]) @@ -1753,7 +1753,7 @@ func testTLSReload( tlsInfo := cloneFunc() // 2. start cluster with valid certs - clus := integration.NewClusterV3(t, &integration.ClusterConfig{ + clus := integration.NewCluster(t, &integration.ClusterConfig{ Size: 1, PeerTLS: &tlsInfo, ClientTLS: &tlsInfo, @@ -1824,7 +1824,7 @@ func testTLSReload( func TestGRPCRequireLeader(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) clus.Members[1].Stop(t) @@ -1850,7 +1850,7 @@ func TestGRPCRequireLeader(t *testing.T) { func TestGRPCStreamRequireLeader(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) client, err := integration.NewClientV3(clus.Members[0]) @@ -1924,7 +1924,7 @@ func TestV3LargeRequests(t *testing.T) { } for i, test := range tests { t.Run(fmt.Sprintf("#%d", i), func(t *testing.T) { - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1, MaxRequestBytes: test.maxRequestBytes}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1, MaxRequestBytes: test.maxRequestBytes}) defer clus.Terminate(t) kvcli := integration.ToGRPC(clus.Client(0)).KV reqput := &pb.PutRequest{Key: []byte("foo"), Value: make([]byte, test.valueSize)} diff --git a/tests/integration/v3_health_test.go b/tests/integration/v3_health_test.go index 2bd03588d..98b64c63f 100644 --- a/tests/integration/v3_health_test.go +++ b/tests/integration/v3_health_test.go @@ -25,7 +25,7 @@ import ( func TestHealthCheck(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := healthpb.NewHealthClient(clus.RandClient().ActiveConnection()) diff --git a/tests/integration/v3_kv_test.go b/tests/integration/v3_kv_test.go index 851edd294..01a4570f5 100644 --- a/tests/integration/v3_kv_test.go +++ b/tests/integration/v3_kv_test.go @@ -13,7 +13,7 @@ import ( func TestKVWithEmptyValue(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) client := clus.RandClient() diff --git a/tests/integration/v3_leadership_test.go b/tests/integration/v3_leadership_test.go index 8fba02c86..84bd97d32 100644 --- a/tests/integration/v3_leadership_test.go +++ b/tests/integration/v3_leadership_test.go @@ -33,7 +33,7 @@ func TestMoveLeaderService(t *testing.T) { testMoveLeader(t, false) } func testMoveLeader(t *testing.T, auto bool) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) oldLeadIdx := clus.WaitLeader(t) @@ -101,7 +101,7 @@ func testMoveLeader(t *testing.T, auto bool) { func TestMoveLeaderError(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) oldLeadIdx := clus.WaitLeader(t) @@ -120,7 +120,7 @@ func TestMoveLeaderError(t *testing.T) { func TestMoveLeaderToLearnerError(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) // we have to add and launch learner member after initial cluster was created, because @@ -153,7 +153,7 @@ func TestMoveLeaderToLearnerError(t *testing.T) { func TestTransferLeadershipWithLearner(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) clus.AddAndLaunchLearnerMember(t) @@ -187,7 +187,7 @@ func TestFirstCommitNotification(t *testing.T) { integration.BeforeTest(t) ctx := context.Background() clusterSize := 3 - cluster := integration.NewClusterV3(t, &integration.ClusterConfig{Size: clusterSize}) + cluster := integration.NewCluster(t, &integration.ClusterConfig{Size: clusterSize}) defer cluster.Terminate(t) oldLeaderIdx := cluster.WaitLeader(t) diff --git a/tests/integration/v3_lease_test.go b/tests/integration/v3_lease_test.go index 5151264a6..412dd7899 100644 --- a/tests/integration/v3_lease_test.go +++ b/tests/integration/v3_lease_test.go @@ -37,7 +37,7 @@ import ( func TestV3LeasePromote(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3, UseBridge: true}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3, UseBridge: true}) defer clus.Terminate(t) // create lease @@ -98,7 +98,7 @@ func TestV3LeasePromote(t *testing.T) { // TestV3LeaseRevoke ensures a key is deleted once its lease is revoked. func TestV3LeaseRevoke(t *testing.T) { integration.BeforeTest(t) - testLeaseRemoveLeasedKey(t, func(clus *integration.ClusterV3, leaseID int64) error { + testLeaseRemoveLeasedKey(t, func(clus *integration.Cluster, leaseID int64) error { lc := integration.ToGRPC(clus.RandClient()).Lease _, err := lc.LeaseRevoke(context.TODO(), &pb.LeaseRevokeRequest{ID: leaseID}) return err @@ -108,7 +108,7 @@ func TestV3LeaseRevoke(t *testing.T) { // TestV3LeaseGrantById ensures leases may be created by a given id. func TestV3LeaseGrantByID(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) // create fixed lease @@ -145,7 +145,7 @@ func TestV3LeaseGrantByID(t *testing.T) { // TestV3LeaseExpire ensures a key is deleted once a key expires. func TestV3LeaseExpire(t *testing.T) { integration.BeforeTest(t) - testLeaseRemoveLeasedKey(t, func(clus *integration.ClusterV3, leaseID int64) error { + testLeaseRemoveLeasedKey(t, func(clus *integration.Cluster, leaseID int64) error { // let lease lapse; wait for deleted key ctx, cancel := context.WithCancel(context.Background()) @@ -197,7 +197,7 @@ func TestV3LeaseExpire(t *testing.T) { // TestV3LeaseKeepAlive ensures keepalive keeps the lease alive. func TestV3LeaseKeepAlive(t *testing.T) { integration.BeforeTest(t) - testLeaseRemoveLeasedKey(t, func(clus *integration.ClusterV3, leaseID int64) error { + testLeaseRemoveLeasedKey(t, func(clus *integration.Cluster, leaseID int64) error { lc := integration.ToGRPC(clus.RandClient()).Lease lreq := &pb.LeaseKeepAliveRequest{ID: leaseID} ctx, cancel := context.WithCancel(context.Background()) @@ -284,7 +284,7 @@ func TestV3LeaseCheckpoint(t *testing.T) { EnableLeaseCheckpoint: tc.checkpointingEnabled, LeaseCheckpointInterval: tc.checkpointingInterval, } - clus := integration.NewClusterV3(t, config) + clus := integration.NewCluster(t, config) defer clus.Terminate(t) // create lease @@ -339,7 +339,7 @@ func TestV3LeaseCheckpoint(t *testing.T) { // TestV3LeaseExists creates a lease on a random client and confirms it exists in the cluster. func TestV3LeaseExists(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) // create lease @@ -363,7 +363,7 @@ func TestV3LeaseExists(t *testing.T) { // TestV3LeaseLeases creates leases and confirms list RPC fetches created ones. func TestV3LeaseLeases(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) ctx0, cancel0 := context.WithCancel(context.Background()) @@ -413,7 +413,7 @@ func TestV3LeaseTimeToLiveStress(t *testing.T) { func testLeaseStress(t *testing.T, stresser func(context.Context, pb.LeaseClient) error) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) ctx, cancel := context.WithTimeout(context.Background(), 10*time.Second) @@ -484,7 +484,7 @@ func stressLeaseTimeToLive(tctx context.Context, lc pb.LeaseClient) (reterr erro func TestV3PutOnNonExistLease(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) ctx, cancel := context.WithCancel(context.Background()) @@ -502,7 +502,7 @@ func TestV3PutOnNonExistLease(t *testing.T) { // related issue https://github.com/etcd-io/etcd/issues/6537 func TestV3GetNonExistLease(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) ctx, cancel := context.WithCancel(context.Background()) @@ -540,7 +540,7 @@ func TestV3GetNonExistLease(t *testing.T) { // TestV3LeaseSwitch tests a key can be switched from one lease to another. func TestV3LeaseSwitch(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) key := "foo" @@ -603,7 +603,7 @@ func TestV3LeaseSwitch(t *testing.T) { func TestV3LeaseFailover(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) toIsolate := clus.WaitMembersForLeader(t, clus.Members) @@ -664,7 +664,7 @@ func TestV3LeaseFailover(t *testing.T) { func TestV3LeaseRequireLeader(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) lc := integration.ToGRPC(clus.Client(0)).Lease @@ -704,7 +704,7 @@ const fiveMinTTL int64 = 300 func TestV3LeaseRecoverAndRevoke(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) kvc := integration.ToGRPC(clus.Client(0)).KV @@ -755,7 +755,7 @@ func TestV3LeaseRecoverAndRevoke(t *testing.T) { func TestV3LeaseRevokeAndRecover(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) kvc := integration.ToGRPC(clus.Client(0)).KV @@ -807,7 +807,7 @@ func TestV3LeaseRevokeAndRecover(t *testing.T) { func TestV3LeaseRecoverKeyWithDetachedLease(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) kvc := integration.ToGRPC(clus.Client(0)).KV @@ -863,7 +863,7 @@ func TestV3LeaseRecoverKeyWithDetachedLease(t *testing.T) { func TestV3LeaseRecoverKeyWithMutipleLease(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) kvc := integration.ToGRPC(clus.Client(0)).KV @@ -935,7 +935,7 @@ func TestV3LeaseRecoverKeyWithMutipleLease(t *testing.T) { } // acquireLeaseAndKey creates a new lease and creates an attached key. -func acquireLeaseAndKey(clus *integration.ClusterV3, key string) (int64, error) { +func acquireLeaseAndKey(clus *integration.Cluster, key string) (int64, error) { // create lease lresp, err := integration.ToGRPC(clus.RandClient()).Lease.LeaseGrant( context.TODO(), @@ -956,8 +956,8 @@ func acquireLeaseAndKey(clus *integration.ClusterV3, key string) (int64, error) // testLeaseRemoveLeasedKey performs some action while holding a lease with an // attached key "foo", then confirms the key is gone. -func testLeaseRemoveLeasedKey(t *testing.T, act func(*integration.ClusterV3, int64) error) { - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) +func testLeaseRemoveLeasedKey(t *testing.T, act func(*integration.Cluster, int64) error) { + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) leaseID, err := acquireLeaseAndKey(clus, "foo") @@ -980,7 +980,7 @@ func testLeaseRemoveLeasedKey(t *testing.T, act func(*integration.ClusterV3, int } } -func leaseExist(t *testing.T, clus *integration.ClusterV3, leaseID int64) bool { +func leaseExist(t *testing.T, clus *integration.Cluster, leaseID int64) bool { l := integration.ToGRPC(clus.RandClient()).Lease _, err := l.LeaseGrant(context.Background(), &pb.LeaseGrantRequest{ID: leaseID, TTL: 5}) diff --git a/tests/integration/v3_stm_test.go b/tests/integration/v3_stm_test.go index ccd7eea1d..f86085bf7 100644 --- a/tests/integration/v3_stm_test.go +++ b/tests/integration/v3_stm_test.go @@ -31,7 +31,7 @@ import ( func TestSTMConflict(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) etcdc := clus.RandClient() @@ -99,7 +99,7 @@ func TestSTMConflict(t *testing.T) { func TestSTMPutNewKey(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) etcdc := clus.RandClient() @@ -126,7 +126,7 @@ func TestSTMPutNewKey(t *testing.T) { func TestSTMAbort(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) etcdc := clus.RandClient() @@ -157,7 +157,7 @@ func TestSTMAbort(t *testing.T) { func TestSTMSerialize(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) etcdc := clus.RandClient() @@ -220,7 +220,7 @@ func TestSTMSerialize(t *testing.T) { func TestSTMApplyOnConcurrentDeletion(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) etcdc := clus.RandClient() @@ -269,7 +269,7 @@ func TestSTMApplyOnConcurrentDeletion(t *testing.T) { func TestSTMSerializableSnapshotPut(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) cli := clus.Client(0) diff --git a/tests/integration/v3_tls_test.go b/tests/integration/v3_tls_test.go index e50419261..793d3d5a0 100644 --- a/tests/integration/v3_tls_test.go +++ b/tests/integration/v3_tls_test.go @@ -48,7 +48,7 @@ func testTLSCipherSuites(t *testing.T, valid bool) { srvTLS.CipherSuites, cliTLS.CipherSuites = cipherSuites[:2], cipherSuites[2:] } - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1, ClientTLS: &srvTLS}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1, ClientTLS: &srvTLS}) defer clus.Terminate(t) cc, err := cliTLS.ClientConfig() diff --git a/tests/integration/v3_watch_restore_test.go b/tests/integration/v3_watch_restore_test.go index 532871bf2..2f3271eb9 100644 --- a/tests/integration/v3_watch_restore_test.go +++ b/tests/integration/v3_watch_restore_test.go @@ -53,7 +53,7 @@ func MustFetchNotEmptyMetric(tb testing.TB, member *integration.Member, metric s func TestV3WatchRestoreSnapshotUnsync(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{ + clus := integration.NewCluster(t, &integration.ClusterConfig{ Size: 3, SnapshotCount: 10, SnapshotCatchUpEntries: 5, diff --git a/tests/integration/v3_watch_test.go b/tests/integration/v3_watch_test.go index 59433e0cc..45b62f7ac 100644 --- a/tests/integration/v3_watch_test.go +++ b/tests/integration/v3_watch_test.go @@ -207,7 +207,7 @@ func TestV3WatchFromCurrentRevision(t *testing.T) { for i, tt := range tests { t.Run(tt.name, func(t *testing.T) { - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) wAPI := integration.ToGRPC(clus.RandClient()).Watch @@ -294,7 +294,7 @@ func TestV3WatchFromCurrentRevision(t *testing.T) { func TestV3WatchFutureRevision(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) wAPI := integration.ToGRPC(clus.RandClient()).Watch @@ -355,7 +355,7 @@ func TestV3WatchFutureRevision(t *testing.T) { func TestV3WatchWrongRange(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) wAPI := integration.ToGRPC(clus.RandClient()).Watch @@ -409,7 +409,7 @@ func TestV3WatchCancelUnsynced(t *testing.T) { } func testV3WatchCancel(t *testing.T, startRev int64) { - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second) @@ -465,7 +465,7 @@ func testV3WatchCancel(t *testing.T, startRev int64) { // overlapping puts. func TestV3WatchCurrentPutOverlap(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second) @@ -550,7 +550,7 @@ func TestV3WatchCurrentPutOverlap(t *testing.T) { func TestV3WatchEmptyKey(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second) @@ -608,7 +608,7 @@ func TestV3WatchMultipleWatchersUnsynced(t *testing.T) { // that matches all watchers, and another key that matches only // one watcher to test if it receives expected events. func testV3WatchMultipleWatchers(t *testing.T, startRev int64) { - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) kvc := integration.ToGRPC(clus.RandClient()).KV @@ -710,7 +710,7 @@ func TestV3WatchMultipleEventsTxnUnsynced(t *testing.T) { // testV3WatchMultipleEventsTxn tests Watch APIs when it receives multiple events. func testV3WatchMultipleEventsTxn(t *testing.T, startRev int64) { - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second) @@ -793,7 +793,7 @@ func (evs eventsSortByKey) Less(i, j int) bool { func TestV3WatchMultipleEventsPutUnsynced(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) kvc := integration.ToGRPC(clus.RandClient()).KV @@ -882,7 +882,7 @@ func TestV3WatchMultipleStreamsUnsynced(t *testing.T) { // testV3WatchMultipleStreams tests multiple watchers on the same key on multiple streams. func testV3WatchMultipleStreams(t *testing.T, startRev int64) { - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) wAPI := integration.ToGRPC(clus.RandClient()).Watch @@ -985,7 +985,7 @@ func TestWatchWithProgressNotify(t *testing.T) { defer func() { v3rpc.SetProgressReportInterval(oldpi) }() integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 3}) defer clus.Terminate(t) ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second) @@ -1035,7 +1035,7 @@ func TestWatchWithProgressNotify(t *testing.T) { // TestV3WatcMultiOpenhClose opens many watchers concurrently on multiple streams. func TestV3WatchClose(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1, UseBridge: true}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1, UseBridge: true}) defer clus.Terminate(t) c := clus.Client(0) @@ -1071,7 +1071,7 @@ func TestV3WatchClose(t *testing.T) { func TestV3WatchWithFilter(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second) @@ -1139,7 +1139,7 @@ func TestV3WatchWithFilter(t *testing.T) { func TestV3WatchWithPrevKV(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) wctx, wcancel := context.WithCancel(context.Background()) @@ -1214,7 +1214,7 @@ func TestV3WatchWithPrevKV(t *testing.T) { func TestV3WatchCancellation(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second) @@ -1256,7 +1256,7 @@ func TestV3WatchCancellation(t *testing.T) { func TestV3WatchCloseCancelRace(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second) diff --git a/tests/integration/v3election_grpc_test.go b/tests/integration/v3election_grpc_test.go index 32127e50a..c7bf79905 100644 --- a/tests/integration/v3election_grpc_test.go +++ b/tests/integration/v3election_grpc_test.go @@ -29,7 +29,7 @@ import ( // simultaneous leadership to multiple campaigners. func TestV3ElectionCampaign(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lease1, err1 := integration.ToGRPC(clus.RandClient()).Lease.LeaseGrant(context.TODO(), &pb.LeaseGrantRequest{TTL: 30}) @@ -91,7 +91,7 @@ func TestV3ElectionCampaign(t *testing.T) { // proclamations from different leaders uninterrupted. func TestV3ElectionObserve(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lc := integration.ToGRPC(clus.Client(0)).Election diff --git a/tests/integration/v3lock_grpc_test.go b/tests/integration/v3lock_grpc_test.go index 1396fb3d4..f293bc1a5 100644 --- a/tests/integration/v3lock_grpc_test.go +++ b/tests/integration/v3lock_grpc_test.go @@ -28,7 +28,7 @@ import ( // once it is unlocked. func TestV3LockLockWaiter(t *testing.T) { integration.BeforeTest(t) - clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1}) + clus := integration.NewCluster(t, &integration.ClusterConfig{Size: 1}) defer clus.Terminate(t) lease1, err1 := integration.ToGRPC(clus.RandClient()).Lease.LeaseGrant(context.TODO(), &pb.LeaseGrantRequest{TTL: 30})