mirror of
https://github.com/etcd-io/etcd.git
synced 2024-09-27 06:25:44 +00:00

Currenly Member exports GrpcURL already as a struct variable. However, when addressing the var-naming linting issues, renaming it from GrpcURL to GRPCURL, clashes with the GRPCURL() function. Given that it's already an exported variable, there's no need to define a getter function. The use of this variable is also mixed, with calls to both the exported variable (GrpcURL) and the function [GRPCURL()]. Signed-off-by: Ivan Valdes <ivan@vald.es>
445 lines
14 KiB
Go
445 lines
14 KiB
Go
// Copyright 2017 The etcd Authors
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package clientv3test
|
|
|
|
import (
|
|
"bytes"
|
|
"context"
|
|
"crypto/sha256"
|
|
"fmt"
|
|
"io"
|
|
"math"
|
|
"os"
|
|
"path/filepath"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/stretchr/testify/assert"
|
|
"github.com/stretchr/testify/require"
|
|
"go.uber.org/zap/zaptest"
|
|
"google.golang.org/grpc"
|
|
|
|
"go.etcd.io/etcd/api/v3/v3rpc/rpctypes"
|
|
"go.etcd.io/etcd/api/v3/version"
|
|
clientv3 "go.etcd.io/etcd/client/v3"
|
|
"go.etcd.io/etcd/server/v3/lease"
|
|
"go.etcd.io/etcd/server/v3/storage/backend"
|
|
"go.etcd.io/etcd/server/v3/storage/mvcc"
|
|
"go.etcd.io/etcd/server/v3/storage/mvcc/testutil"
|
|
integration2 "go.etcd.io/etcd/tests/v3/framework/integration"
|
|
)
|
|
|
|
func TestMaintenanceHashKV(t *testing.T) {
|
|
integration2.BeforeTest(t)
|
|
|
|
clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3})
|
|
defer clus.Terminate(t)
|
|
|
|
for i := 0; i < 3; i++ {
|
|
if _, err := clus.RandClient().Put(context.Background(), "foo", "bar"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
var hv uint32
|
|
for i := 0; i < 3; i++ {
|
|
cli := clus.Client(i)
|
|
// ensure writes are replicated
|
|
if _, err := cli.Get(context.TODO(), "foo"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
hresp, err := cli.HashKV(context.Background(), clus.Members[i].GRPCURL, 0)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if hv == 0 {
|
|
hv = hresp.Hash
|
|
continue
|
|
}
|
|
if hv != hresp.Hash {
|
|
t.Fatalf("#%d: hash expected %d, got %d", i, hv, hresp.Hash)
|
|
}
|
|
}
|
|
}
|
|
|
|
// TestCompactionHash tests compaction hash
|
|
// TODO: Change this to fuzz test
|
|
func TestCompactionHash(t *testing.T) {
|
|
integration2.BeforeTest(t)
|
|
|
|
clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
cc, err := clus.ClusterClient(t)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
testutil.TestCompactionHash(context.Background(), t, hashTestCase{cc, clus.Members[0].GRPCURL}, 1000)
|
|
}
|
|
|
|
type hashTestCase struct {
|
|
*clientv3.Client
|
|
url string
|
|
}
|
|
|
|
func (tc hashTestCase) Put(ctx context.Context, key, value string) error {
|
|
_, err := tc.Client.Put(ctx, key, value)
|
|
return err
|
|
}
|
|
|
|
func (tc hashTestCase) Delete(ctx context.Context, key string) error {
|
|
_, err := tc.Client.Delete(ctx, key)
|
|
return err
|
|
}
|
|
|
|
func (tc hashTestCase) HashByRev(ctx context.Context, rev int64) (testutil.KeyValueHash, error) {
|
|
resp, err := tc.Client.HashKV(ctx, tc.url, rev)
|
|
return testutil.KeyValueHash{Hash: resp.Hash, CompactRevision: resp.CompactRevision, Revision: resp.Header.Revision}, err
|
|
}
|
|
|
|
func (tc hashTestCase) Defrag(ctx context.Context) error {
|
|
_, err := tc.Client.Defragment(ctx, tc.url)
|
|
return err
|
|
}
|
|
|
|
func (tc hashTestCase) Compact(ctx context.Context, rev int64) error {
|
|
_, err := tc.Client.Compact(ctx, rev)
|
|
// Wait for compaction to be compacted
|
|
time.Sleep(50 * time.Millisecond)
|
|
return err
|
|
}
|
|
|
|
func TestMaintenanceMoveLeader(t *testing.T) {
|
|
integration2.BeforeTest(t)
|
|
|
|
clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3})
|
|
defer clus.Terminate(t)
|
|
|
|
oldLeadIdx := clus.WaitLeader(t)
|
|
targetIdx := (oldLeadIdx + 1) % 3
|
|
target := uint64(clus.Members[targetIdx].ID())
|
|
|
|
cli := clus.Client(targetIdx)
|
|
_, err := cli.MoveLeader(context.Background(), target)
|
|
if err != rpctypes.ErrNotLeader {
|
|
t.Fatalf("error expected %v, got %v", rpctypes.ErrNotLeader, err)
|
|
}
|
|
|
|
cli = clus.Client(oldLeadIdx)
|
|
_, err = cli.MoveLeader(context.Background(), target)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
leadIdx := clus.WaitLeader(t)
|
|
lead := uint64(clus.Members[leadIdx].ID())
|
|
if target != lead {
|
|
t.Fatalf("new leader expected %d, got %d", target, lead)
|
|
}
|
|
}
|
|
|
|
// TestMaintenanceSnapshotCancel ensures that context cancel
|
|
// before snapshot reading returns corresponding context errors.
|
|
func TestMaintenanceSnapshotCancel(t *testing.T) {
|
|
integration2.BeforeTest(t)
|
|
|
|
clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
// reading snapshot with canceled context should error out
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
|
|
// Since http2 spec defines the receive windows's size and max size of
|
|
// frame in the stream, the underlayer - gRPC client can pre-read data
|
|
// from server even if the application layer hasn't read it yet.
|
|
//
|
|
// And the initialized cluster has 20KiB snapshot, which can be
|
|
// pre-read by underlayer. We should increase the snapshot's size here,
|
|
// just in case that io.Copy won't return the canceled error.
|
|
populateDataIntoCluster(t, clus, 3, 1024*1024)
|
|
|
|
rc1, err := clus.RandClient().Snapshot(ctx)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
defer rc1.Close()
|
|
|
|
// read 16 bytes to ensure that server opens snapshot
|
|
buf := make([]byte, 16)
|
|
n, err := rc1.Read(buf)
|
|
assert.NoError(t, err)
|
|
assert.Equal(t, 16, n)
|
|
|
|
cancel()
|
|
_, err = io.Copy(io.Discard, rc1)
|
|
if err != context.Canceled {
|
|
t.Errorf("expected %v, got %v", context.Canceled, err)
|
|
}
|
|
}
|
|
|
|
// TestMaintenanceSnapshotWithVersionTimeout ensures that SnapshotWithVersion function
|
|
// returns corresponding context errors when context timeout happened before snapshot reading
|
|
func TestMaintenanceSnapshotWithVersionTimeout(t *testing.T) {
|
|
testMaintenanceSnapshotTimeout(t, func(ctx context.Context, client *clientv3.Client) (io.ReadCloser, error) {
|
|
resp, err := client.SnapshotWithVersion(ctx)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return resp.Snapshot, nil
|
|
})
|
|
}
|
|
|
|
// TestMaintenanceSnapshotTimeout ensures that Snapshot function
|
|
// returns corresponding context errors when context timeout happened before snapshot reading
|
|
func TestMaintenanceSnapshotTimeout(t *testing.T) {
|
|
testMaintenanceSnapshotTimeout(t, func(ctx context.Context, client *clientv3.Client) (io.ReadCloser, error) {
|
|
return client.Snapshot(ctx)
|
|
})
|
|
}
|
|
|
|
// testMaintenanceSnapshotTimeout given snapshot function ensures that it
|
|
// returns corresponding context errors when context timeout happened before snapshot reading
|
|
func testMaintenanceSnapshotTimeout(t *testing.T, snapshot func(context.Context, *clientv3.Client) (io.ReadCloser, error)) {
|
|
integration2.BeforeTest(t)
|
|
|
|
clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
// reading snapshot with deadline exceeded should error out
|
|
ctx, cancel := context.WithTimeout(context.Background(), time.Second)
|
|
defer cancel()
|
|
|
|
// Since http2 spec defines the receive windows's size and max size of
|
|
// frame in the stream, the underlayer - gRPC client can pre-read data
|
|
// from server even if the application layer hasn't read it yet.
|
|
//
|
|
// And the initialized cluster has 20KiB snapshot, which can be
|
|
// pre-read by underlayer. We should increase the snapshot's size here,
|
|
// just in case that io.Copy won't return the timeout error.
|
|
populateDataIntoCluster(t, clus, 3, 1024*1024)
|
|
|
|
rc2, err := snapshot(ctx, clus.RandClient())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
defer rc2.Close()
|
|
|
|
time.Sleep(2 * time.Second)
|
|
|
|
_, err = io.Copy(io.Discard, rc2)
|
|
if err != nil && !IsClientTimeout(err) {
|
|
t.Errorf("expected client timeout, got %v", err)
|
|
}
|
|
}
|
|
|
|
// TestMaintenanceSnapshotWithVersionErrorInflight ensures that ReaderCloser returned by SnapshotWithVersion function
|
|
// will fail to read with corresponding context errors on inflight context cancel timeout.
|
|
func TestMaintenanceSnapshotWithVersionErrorInflight(t *testing.T) {
|
|
testMaintenanceSnapshotErrorInflight(t, func(ctx context.Context, client *clientv3.Client) (io.ReadCloser, error) {
|
|
resp, err := client.SnapshotWithVersion(ctx)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return resp.Snapshot, nil
|
|
})
|
|
}
|
|
|
|
// TestMaintenanceSnapshotErrorInflight ensures that ReaderCloser returned by Snapshot function
|
|
// will fail to read with corresponding context errors on inflight context cancel timeout.
|
|
func TestMaintenanceSnapshotErrorInflight(t *testing.T) {
|
|
testMaintenanceSnapshotErrorInflight(t, func(ctx context.Context, client *clientv3.Client) (io.ReadCloser, error) {
|
|
return client.Snapshot(ctx)
|
|
})
|
|
}
|
|
|
|
// testMaintenanceSnapshotErrorInflight given snapshot function ensures that ReaderCloser returned by it
|
|
// will fail to read with corresponding context errors on inflight context cancel timeout.
|
|
func testMaintenanceSnapshotErrorInflight(t *testing.T, snapshot func(context.Context, *clientv3.Client) (io.ReadCloser, error)) {
|
|
integration2.BeforeTest(t)
|
|
lg := zaptest.NewLogger(t)
|
|
|
|
clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, UseBridge: true})
|
|
defer clus.Terminate(t)
|
|
|
|
// take about 1-second to read snapshot
|
|
clus.Members[0].Stop(t)
|
|
dpath := filepath.Join(clus.Members[0].DataDir, "member", "snap", "db")
|
|
b := backend.NewDefaultBackend(lg, dpath)
|
|
s := mvcc.NewStore(lg, b, &lease.FakeLessor{}, mvcc.StoreConfig{CompactionBatchLimit: math.MaxInt32})
|
|
rev := 100000
|
|
for i := 2; i <= rev; i++ {
|
|
s.Put([]byte(fmt.Sprintf("%10d", i)), bytes.Repeat([]byte("a"), 1024), lease.NoLease)
|
|
}
|
|
s.Close()
|
|
b.Close()
|
|
clus.Members[0].Restart(t)
|
|
|
|
// reading snapshot with canceled context should error out
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
rc1, err := snapshot(ctx, clus.RandClient())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
defer rc1.Close()
|
|
|
|
donec := make(chan struct{})
|
|
go func() {
|
|
time.Sleep(300 * time.Millisecond)
|
|
cancel()
|
|
close(donec)
|
|
}()
|
|
_, err = io.Copy(io.Discard, rc1)
|
|
if err != nil && err != context.Canceled {
|
|
t.Errorf("expected %v, got %v", context.Canceled, err)
|
|
}
|
|
<-donec
|
|
|
|
// reading snapshot with deadline exceeded should error out
|
|
ctx, cancel = context.WithTimeout(context.Background(), time.Second)
|
|
defer cancel()
|
|
rc2, err := snapshot(ctx, clus.RandClient())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
defer rc2.Close()
|
|
|
|
// 300ms left and expect timeout while snapshot reading is in progress
|
|
time.Sleep(700 * time.Millisecond)
|
|
_, err = io.Copy(io.Discard, rc2)
|
|
if err != nil && !IsClientTimeout(err) {
|
|
t.Errorf("expected client timeout, got %v", err)
|
|
}
|
|
}
|
|
|
|
// TestMaintenanceSnapshotWithVersionVersion ensures that SnapshotWithVersion returns correct version value.
|
|
func TestMaintenanceSnapshotWithVersionVersion(t *testing.T) {
|
|
integration2.BeforeTest(t)
|
|
|
|
// Set SnapshotCount to 1 to force raft snapshot to ensure that storage version is set
|
|
clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1, SnapshotCount: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
// Put some keys to ensure that wal snapshot is triggered
|
|
for i := 0; i < 10; i++ {
|
|
clus.RandClient().Put(context.Background(), fmt.Sprintf("%d", i), "1")
|
|
}
|
|
|
|
// reading snapshot with canceled context should error out
|
|
resp, err := clus.RandClient().SnapshotWithVersion(context.Background())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
defer resp.Snapshot.Close()
|
|
if resp.Version != "3.6.0" {
|
|
t.Errorf("unexpected version, expected %q, got %q", version.Version, resp.Version)
|
|
}
|
|
}
|
|
|
|
func TestMaintenanceSnapshotContentDigest(t *testing.T) {
|
|
integration2.BeforeTest(t)
|
|
|
|
clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
populateDataIntoCluster(t, clus, 3, 1024*1024)
|
|
|
|
// reading snapshot with canceled context should error out
|
|
resp, err := clus.RandClient().SnapshotWithVersion(context.Background())
|
|
require.NoError(t, err)
|
|
defer resp.Snapshot.Close()
|
|
|
|
tmpDir := t.TempDir()
|
|
snapFile, err := os.Create(filepath.Join(tmpDir, t.Name()))
|
|
require.NoError(t, err)
|
|
defer snapFile.Close()
|
|
|
|
snapSize, err := io.Copy(snapFile, resp.Snapshot)
|
|
require.NoError(t, err)
|
|
|
|
// read the checksum
|
|
checksumSize := int64(sha256.Size)
|
|
_, err = snapFile.Seek(-checksumSize, io.SeekEnd)
|
|
require.NoError(t, err)
|
|
|
|
checksumInBytes, err := io.ReadAll(snapFile)
|
|
require.NoError(t, err)
|
|
require.Equal(t, int(checksumSize), len(checksumInBytes))
|
|
|
|
// remove the checksum part and rehash
|
|
err = snapFile.Truncate(snapSize - checksumSize)
|
|
require.NoError(t, err)
|
|
|
|
_, err = snapFile.Seek(0, io.SeekStart)
|
|
require.NoError(t, err)
|
|
|
|
hashWriter := sha256.New()
|
|
_, err = io.Copy(hashWriter, snapFile)
|
|
require.NoError(t, err)
|
|
|
|
// compare the checksum
|
|
actualChecksum := hashWriter.Sum(nil)
|
|
require.Equal(t, checksumInBytes, actualChecksum)
|
|
}
|
|
|
|
func TestMaintenanceStatus(t *testing.T) {
|
|
integration2.BeforeTest(t)
|
|
|
|
clus := integration2.NewCluster(t, &integration2.ClusterConfig{Size: 3})
|
|
defer clus.Terminate(t)
|
|
|
|
t.Logf("Waiting for leader...")
|
|
clus.WaitLeader(t)
|
|
t.Logf("Leader established.")
|
|
|
|
eps := make([]string, 3)
|
|
for i := 0; i < 3; i++ {
|
|
eps[i] = clus.Members[i].GRPCURL
|
|
}
|
|
|
|
t.Logf("Creating client...")
|
|
cli, err := integration2.NewClient(t, clientv3.Config{Endpoints: eps, DialOptions: []grpc.DialOption{grpc.WithBlock()}})
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
defer cli.Close()
|
|
t.Logf("Creating client [DONE]")
|
|
|
|
prevID, leaderFound := uint64(0), false
|
|
for i := 0; i < 3; i++ {
|
|
resp, err := cli.Status(context.TODO(), eps[i])
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
t.Logf("Response from %v: %v", i, resp)
|
|
if prevID == 0 {
|
|
prevID, leaderFound = resp.Header.MemberId, resp.Header.MemberId == resp.Leader
|
|
continue
|
|
}
|
|
if prevID == resp.Header.MemberId {
|
|
t.Errorf("#%d: status returned duplicate member ID with %016x", i, prevID)
|
|
}
|
|
if leaderFound && resp.Header.MemberId == resp.Leader {
|
|
t.Errorf("#%d: leader already found, but found another %016x", i, resp.Header.MemberId)
|
|
}
|
|
if !leaderFound {
|
|
leaderFound = resp.Header.MemberId == resp.Leader
|
|
}
|
|
}
|
|
if !leaderFound {
|
|
t.Fatal("no leader found")
|
|
}
|
|
}
|