// Copyright 2017 The etcd Authors // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package connectivity_test import ( "bytes" "context" "fmt" "testing" "time" "go.etcd.io/etcd/api/v3/v3rpc/rpctypes" "go.etcd.io/etcd/client/v3" integration2 "go.etcd.io/etcd/tests/v3/framework/integration" "go.etcd.io/etcd/tests/v3/integration/clientv3" ) // TestBalancerUnderServerShutdownWatch expects that watch client // switch its endpoints when the member of the pinned endpoint fails. func TestBalancerUnderServerShutdownWatch(t *testing.T) { integration2.BeforeTest(t) clus := integration2.NewCluster(t, &integration2.ClusterConfig{ Size: 3, UseBridge: true, }) defer clus.Terminate(t) eps := []string{clus.Members[0].GRPCURL(), clus.Members[1].GRPCURL(), clus.Members[2].GRPCURL()} lead := clus.WaitLeader(t) // pin eps[lead] watchCli, err := integration2.NewClient(t, clientv3.Config{Endpoints: []string{eps[lead]}}) if err != nil { t.Fatal(err) } defer watchCli.Close() // wait for eps[lead] to be pinned clientv3test.MustWaitPinReady(t, watchCli) // add all eps to list, so that when the original pined one fails // the client can switch to other available eps watchCli.SetEndpoints(eps...) key, val := "foo", "bar" wch := watchCli.Watch(context.Background(), key, clientv3.WithCreatedNotify()) select { case <-wch: case <-time.After(integration2.RequestWaitTimeout): t.Fatal("took too long to create watch") } donec := make(chan struct{}) go func() { defer close(donec) // switch to others when eps[lead] is shut down select { case ev := <-wch: if werr := ev.Err(); werr != nil { t.Error(werr) } if len(ev.Events) != 1 { t.Errorf("expected one event, got %+v", ev) } if !bytes.Equal(ev.Events[0].Kv.Value, []byte(val)) { t.Errorf("expected %q, got %+v", val, ev.Events[0].Kv) } case <-time.After(7 * time.Second): t.Error("took too long to receive events") } }() // shut down eps[lead] clus.Members[lead].Terminate(t) // writes to eps[lead+1] putCli, err := integration2.NewClient(t, clientv3.Config{Endpoints: []string{eps[(lead+1)%3]}}) if err != nil { t.Fatal(err) } defer putCli.Close() for { ctx, cancel := context.WithTimeout(context.Background(), 2*time.Second) _, err = putCli.Put(ctx, key, val) cancel() if err == nil { break } if clientv3test.IsClientTimeout(err) || clientv3test.IsServerCtxTimeout(err) || err == rpctypes.ErrTimeout || err == rpctypes.ErrTimeoutDueToLeaderFail { continue } t.Fatal(err) } select { case <-donec: case <-time.After(5 * time.Second): // enough time for balancer switch t.Fatal("took too long to receive events") } } func TestBalancerUnderServerShutdownPut(t *testing.T) { testBalancerUnderServerShutdownMutable(t, func(cli *clientv3.Client, ctx context.Context) error { _, err := cli.Put(ctx, "foo", "bar") return err }) } func TestBalancerUnderServerShutdownDelete(t *testing.T) { testBalancerUnderServerShutdownMutable(t, func(cli *clientv3.Client, ctx context.Context) error { _, err := cli.Delete(ctx, "foo") return err }) } func TestBalancerUnderServerShutdownTxn(t *testing.T) { testBalancerUnderServerShutdownMutable(t, func(cli *clientv3.Client, ctx context.Context) error { _, err := cli.Txn(ctx). If(clientv3.Compare(clientv3.Version("foo"), "=", 0)). Then(clientv3.OpPut("foo", "bar")). Else(clientv3.OpPut("foo", "baz")).Commit() return err }) } // testBalancerUnderServerShutdownMutable expects that when the member of // the pinned endpoint is shut down, the balancer switches its endpoints // and all subsequent put/delete/txn requests succeed with new endpoints. func testBalancerUnderServerShutdownMutable(t *testing.T, op func(*clientv3.Client, context.Context) error) { integration2.BeforeTest(t) clus := integration2.NewCluster(t, &integration2.ClusterConfig{ Size: 3, }) defer clus.Terminate(t) eps := []string{clus.Members[0].GRPCURL(), clus.Members[1].GRPCURL(), clus.Members[2].GRPCURL()} // pin eps[0] cli, err := integration2.NewClient(t, clientv3.Config{Endpoints: []string{eps[0]}}) if err != nil { t.Fatal(err) } defer cli.Close() // wait for eps[0] to be pinned clientv3test.MustWaitPinReady(t, cli) // add all eps to list, so that when the original pined one fails // the client can switch to other available eps cli.SetEndpoints(eps...) // shut down eps[0] clus.Members[0].Terminate(t) // switched to others when eps[0] was explicitly shut down // and following request should succeed // TODO: remove this (expose client connection state?) time.Sleep(time.Second) cctx, ccancel := context.WithTimeout(context.Background(), time.Second) err = op(cli, cctx) ccancel() if err != nil { t.Fatal(err) } } func TestBalancerUnderServerShutdownGetLinearizable(t *testing.T) { testBalancerUnderServerShutdownImmutable(t, func(cli *clientv3.Client, ctx context.Context) error { _, err := cli.Get(ctx, "foo") return err }, 7*time.Second) // give enough time for leader election, balancer switch } func TestBalancerUnderServerShutdownGetSerializable(t *testing.T) { testBalancerUnderServerShutdownImmutable(t, func(cli *clientv3.Client, ctx context.Context) error { _, err := cli.Get(ctx, "foo", clientv3.WithSerializable()) return err }, 2*time.Second) } // testBalancerUnderServerShutdownImmutable expects that when the member of // the pinned endpoint is shut down, the balancer switches its endpoints // and all subsequent range requests succeed with new endpoints. func testBalancerUnderServerShutdownImmutable(t *testing.T, op func(*clientv3.Client, context.Context) error, timeout time.Duration) { integration2.BeforeTest(t) clus := integration2.NewCluster(t, &integration2.ClusterConfig{ Size: 3, }) defer clus.Terminate(t) eps := []string{clus.Members[0].GRPCURL(), clus.Members[1].GRPCURL(), clus.Members[2].GRPCURL()} // pin eps[0] cli, err := integration2.NewClient(t, clientv3.Config{Endpoints: []string{eps[0]}}) if err != nil { t.Errorf("failed to create client: %v", err) } defer cli.Close() // wait for eps[0] to be pinned clientv3test.MustWaitPinReady(t, cli) // add all eps to list, so that when the original pined one fails // the client can switch to other available eps cli.SetEndpoints(eps...) // shut down eps[0] clus.Members[0].Terminate(t) // switched to others when eps[0] was explicitly shut down // and following request should succeed cctx, ccancel := context.WithTimeout(context.Background(), timeout) err = op(cli, cctx) ccancel() if err != nil { t.Errorf("failed to finish range request in time %v (timeout %v)", err, timeout) } } func TestBalancerUnderServerStopInflightLinearizableGetOnRestart(t *testing.T) { tt := []pinTestOpt{ {pinLeader: true, stopPinFirst: true}, {pinLeader: true, stopPinFirst: false}, {pinLeader: false, stopPinFirst: true}, {pinLeader: false, stopPinFirst: false}, } for _, w := range tt { t.Run(fmt.Sprintf("%#v", w), func(t *testing.T) { testBalancerUnderServerStopInflightRangeOnRestart(t, true, w) }) } } func TestBalancerUnderServerStopInflightSerializableGetOnRestart(t *testing.T) { tt := []pinTestOpt{ {pinLeader: true, stopPinFirst: true}, {pinLeader: true, stopPinFirst: false}, {pinLeader: false, stopPinFirst: true}, {pinLeader: false, stopPinFirst: false}, } for _, w := range tt { t.Run(fmt.Sprintf("%#v", w), func(t *testing.T) { testBalancerUnderServerStopInflightRangeOnRestart(t, false, w) }) } } type pinTestOpt struct { pinLeader bool stopPinFirst bool } // testBalancerUnderServerStopInflightRangeOnRestart expects // inflight range request reconnects on server restart. func testBalancerUnderServerStopInflightRangeOnRestart(t *testing.T, linearizable bool, opt pinTestOpt) { integration2.BeforeTest(t) cfg := &integration2.ClusterConfig{ Size: 2, UseBridge: true, } if linearizable { cfg.Size = 3 } clus := integration2.NewCluster(t, cfg) defer clus.Terminate(t) eps := []string{clus.Members[0].GRPCURL(), clus.Members[1].GRPCURL()} if linearizable { eps = append(eps, clus.Members[2].GRPCURL()) } lead := clus.WaitLeader(t) target := lead if !opt.pinLeader { target = (target + 1) % 2 } // pin eps[target] cli, err := integration2.NewClient(t, clientv3.Config{Endpoints: []string{eps[target]}}) if err != nil { t.Errorf("failed to create client: %v", err) } defer cli.Close() // wait for eps[target] to be pinned clientv3test.MustWaitPinReady(t, cli) // add all eps to list, so that when the original pined one fails // the client can switch to other available eps cli.SetEndpoints(eps...) if opt.stopPinFirst { clus.Members[target].Stop(t) // give some time for balancer switch before stopping the other time.Sleep(time.Second) clus.Members[(target+1)%2].Stop(t) } else { clus.Members[(target+1)%2].Stop(t) // balancer cannot pin other member since it's already stopped clus.Members[target].Stop(t) } // 3-second is the minimum interval between endpoint being marked // as unhealthy and being removed from unhealthy, so possibly // takes >5-second to unpin and repin an endpoint // TODO: decrease timeout when balancer switch rewrite clientTimeout := 7 * time.Second var gops []clientv3.OpOption if !linearizable { gops = append(gops, clientv3.WithSerializable()) } donec, readyc := make(chan struct{}), make(chan struct{}, 1) go func() { defer close(donec) ctx, cancel := context.WithTimeout(context.TODO(), clientTimeout) readyc <- struct{}{} // TODO: The new grpc load balancer will not pin to an endpoint // as intended by this test. But it will round robin member within // two attempts. // Remove retry loop once the new grpc load balancer provides retry. for i := 0; i < 2; i++ { _, err = cli.Get(ctx, "abc", gops...) if err == nil { break } } cancel() if err != nil { t.Errorf("unexpected error: %v", err) } }() <-readyc clus.Members[target].Restart(t) select { case <-time.After(clientTimeout + integration2.RequestWaitTimeout): t.Fatalf("timed out waiting for Get [linearizable: %v, opt: %+v]", linearizable, opt) case <-donec: } }