mirror of
https://github.com/etcd-io/etcd.git
synced 2024-09-27 06:25:44 +00:00
test: refactor mix_version_test.
Signed-off-by: Siyuan Zhang <sizhang@google.com>
This commit is contained in:
parent
18fd7d67a3
commit
d602b9f4a2
@ -28,49 +28,70 @@ import (
|
|||||||
"go.etcd.io/etcd/tests/v3/framework/e2e"
|
"go.etcd.io/etcd/tests/v3/framework/e2e"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
type clusterTestCase struct {
|
||||||
|
name string
|
||||||
|
config *e2e.EtcdProcessClusterConfig
|
||||||
|
}
|
||||||
|
|
||||||
|
func clusterTestCases(size int) []clusterTestCase {
|
||||||
|
tcs := []clusterTestCase{
|
||||||
|
{
|
||||||
|
name: "CurrentVersion",
|
||||||
|
config: e2e.NewConfig(e2e.WithClusterSize(size)),
|
||||||
|
},
|
||||||
|
}
|
||||||
|
if !fileutil.Exist(e2e.BinPath.EtcdLastRelease) {
|
||||||
|
return tcs
|
||||||
|
}
|
||||||
|
|
||||||
|
tcs = append(tcs,
|
||||||
|
clusterTestCase{
|
||||||
|
name: "LastVersion",
|
||||||
|
config: e2e.NewConfig(e2e.WithClusterSize(size), e2e.WithVersion(e2e.LastVersion)),
|
||||||
|
},
|
||||||
|
)
|
||||||
|
if size > 2 {
|
||||||
|
tcs = append(tcs,
|
||||||
|
clusterTestCase{
|
||||||
|
name: "MinorityLastVersion",
|
||||||
|
config: e2e.NewConfig(e2e.WithClusterSize(size), e2e.WithVersion(e2e.MinorityLastVersion)),
|
||||||
|
}, clusterTestCase{
|
||||||
|
name: "QuorumLastVersion",
|
||||||
|
config: e2e.NewConfig(e2e.WithClusterSize(size), e2e.WithVersion(e2e.QuorumLastVersion)),
|
||||||
|
},
|
||||||
|
)
|
||||||
|
}
|
||||||
|
return tcs
|
||||||
|
}
|
||||||
|
|
||||||
// TestMixVersionsSnapshotByAddingMember tests the mix version send snapshots by adding member
|
// TestMixVersionsSnapshotByAddingMember tests the mix version send snapshots by adding member
|
||||||
func TestMixVersionsSnapshotByAddingMember(t *testing.T) {
|
func TestMixVersionsSnapshotByAddingMember(t *testing.T) {
|
||||||
cases := []struct {
|
for _, tc := range clusterTestCases(1) {
|
||||||
name string
|
t.Run(tc.name+"-adding-new-member-of-current-version", func(t *testing.T) {
|
||||||
clusterVersion e2e.ClusterVersion
|
mixVersionsSnapshotTestByAddingMember(t, tc.config, e2e.CurrentVersion)
|
||||||
newInstanceVersion e2e.ClusterVersion
|
})
|
||||||
}{
|
|
||||||
// etcd doesn't support adding a new member of old version into
|
// etcd doesn't support adding a new member of old version into
|
||||||
// a cluster with higher version. For example, etcd cluster
|
// a cluster with higher version. For example, etcd cluster
|
||||||
// version is 3.6.x, then a new member of 3.5.x can't join the
|
// version is 3.6.x, then a new member of 3.5.x can't join the
|
||||||
// cluster. Please refer to link below,
|
// cluster. Please refer to link below,
|
||||||
// https://github.com/etcd-io/etcd/blob/3e903d0b12e399519a4013c52d4635ec8bdd6863/server/etcdserver/cluster_util.go#L222-L230
|
// https://github.com/etcd-io/etcd/blob/3e903d0b12e399519a4013c52d4635ec8bdd6863/server/etcdserver/cluster_util.go#L222-L230
|
||||||
/*{
|
/*t.Run(tc.name+"-adding-new-member-of-last-version", func(t *testing.T) {
|
||||||
name: "etcd instance with last version receives snapshot from the leader with current version",
|
mixVersionsSnapshotTestByAddingMember(t, tc.config, e2e.LastVersion)
|
||||||
clusterVersion: e2e.CurrentVersion,
|
})*/
|
||||||
newInstaceVersion: e2e.LastVersion,
|
|
||||||
},*/
|
|
||||||
{
|
|
||||||
name: "etcd instance with current version receives snapshot from the leader with last version",
|
|
||||||
clusterVersion: e2e.LastVersion,
|
|
||||||
newInstanceVersion: e2e.CurrentVersion,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, tc := range cases {
|
|
||||||
t.Run(tc.name, func(t *testing.T) {
|
|
||||||
mixVersionsSnapshotTestByAddingMember(t, tc.clusterVersion, tc.newInstanceVersion)
|
|
||||||
})
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func mixVersionsSnapshotTestByAddingMember(t *testing.T, clusterVersion, newInstanceVersion e2e.ClusterVersion) {
|
func mixVersionsSnapshotTestByAddingMember(t *testing.T, cfg *e2e.EtcdProcessClusterConfig, newInstanceVersion e2e.ClusterVersion) {
|
||||||
e2e.BeforeTest(t)
|
e2e.BeforeTest(t)
|
||||||
|
|
||||||
if !fileutil.Exist(e2e.BinPath.EtcdLastRelease) {
|
if !fileutil.Exist(e2e.BinPath.EtcdLastRelease) {
|
||||||
t.Skipf("%q does not exist", e2e.BinPath.EtcdLastRelease)
|
t.Skipf("%q does not exist", e2e.BinPath.EtcdLastRelease)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Create an etcd cluster with 1 member
|
t.Logf("Create an etcd cluster with %d member", cfg.ClusterSize)
|
||||||
epc, err := e2e.NewEtcdProcessCluster(context.TODO(), t,
|
epc, err := e2e.NewEtcdProcessCluster(context.TODO(), t,
|
||||||
e2e.WithClusterSize(1),
|
e2e.WithConfig(cfg),
|
||||||
e2e.WithSnapshotCount(10),
|
e2e.WithSnapshotCount(10),
|
||||||
e2e.WithVersion(clusterVersion),
|
|
||||||
)
|
)
|
||||||
require.NoError(t, err, "failed to start etcd cluster: %v", err)
|
require.NoError(t, err, "failed to start etcd cluster: %v", err)
|
||||||
defer func() {
|
defer func() {
|
||||||
@ -78,16 +99,10 @@ func mixVersionsSnapshotTestByAddingMember(t *testing.T, clusterVersion, newInst
|
|||||||
require.NoError(t, derr, "failed to close etcd cluster: %v", derr)
|
require.NoError(t, derr, "failed to close etcd cluster: %v", derr)
|
||||||
}()
|
}()
|
||||||
|
|
||||||
// Write more than SnapshotCount entries to trigger at least a snapshot.
|
t.Log("Writing 20 keys to the cluster (more than SnapshotCount entries to trigger at least a snapshot)")
|
||||||
t.Log("Writing 20 keys to the cluster")
|
writeKVs(t, epc.Etcdctl(), 0, 20)
|
||||||
for i := 0; i < 20; i++ {
|
|
||||||
key := fmt.Sprintf("key-%d", i)
|
|
||||||
value := fmt.Sprintf("value-%d", i)
|
|
||||||
err = epc.Etcdctl().Put(context.TODO(), key, value, config.PutOptions{})
|
|
||||||
require.NoError(t, err, "failed to put %q, error: %v", key, err)
|
|
||||||
}
|
|
||||||
|
|
||||||
// start a new etcd instance, which will receive a snapshot from the leader.
|
t.Log("start a new etcd instance, which will receive a snapshot from the leader.")
|
||||||
newCfg := *epc.Cfg
|
newCfg := *epc.Cfg
|
||||||
newCfg.Version = newInstanceVersion
|
newCfg.Version = newInstanceVersion
|
||||||
newCfg.ServerConfig.SnapshotCatchUpEntries = 10
|
newCfg.ServerConfig.SnapshotCatchUpEntries = 10
|
||||||
@ -96,62 +111,29 @@ func mixVersionsSnapshotTestByAddingMember(t *testing.T, clusterVersion, newInst
|
|||||||
require.NoError(t, err, "failed to start the new etcd instance: %v", err)
|
require.NoError(t, err, "failed to start the new etcd instance: %v", err)
|
||||||
defer epc.CloseProc(context.TODO(), nil)
|
defer epc.CloseProc(context.TODO(), nil)
|
||||||
|
|
||||||
// verify all nodes have exact same revision and hash
|
assertKVHash(t, epc)
|
||||||
t.Log("Verify all nodes have exact same revision and hash")
|
|
||||||
assert.Eventually(t, func() bool {
|
|
||||||
hashKvs, err := epc.Etcdctl().HashKV(context.TODO(), 0)
|
|
||||||
if err != nil {
|
|
||||||
t.Logf("failed to get HashKV: %v", err)
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
if len(hashKvs) != 2 {
|
|
||||||
t.Logf("expected 2 hashkv responses, but got: %d", len(hashKvs))
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
|
|
||||||
if hashKvs[0].Header.Revision != hashKvs[1].Header.Revision {
|
|
||||||
t.Logf("Got different revisions, [%d, %d]", hashKvs[0].Header.Revision, hashKvs[1].Header.Revision)
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
|
|
||||||
assert.Equal(t, hashKvs[0].Hash, hashKvs[1].Hash)
|
|
||||||
|
|
||||||
return true
|
|
||||||
}, 10*time.Second, 500*time.Millisecond)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestMixVersionsSnapshotByMockingPartition(t *testing.T) {
|
func TestMixVersionsSnapshotByMockingPartition(t *testing.T) {
|
||||||
cases := []struct {
|
mockPartitionNodeIndex := 2
|
||||||
name string
|
for _, tc := range clusterTestCases(3) {
|
||||||
clusterVersion e2e.ClusterVersion
|
|
||||||
mockPartitionNodeIndex int
|
|
||||||
}{
|
|
||||||
{
|
|
||||||
name: "etcd instance with last version receives snapshot from the leader with current version",
|
|
||||||
clusterVersion: e2e.MinorityLastVersion,
|
|
||||||
mockPartitionNodeIndex: 2,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, tc := range cases {
|
|
||||||
t.Run(tc.name, func(t *testing.T) {
|
t.Run(tc.name, func(t *testing.T) {
|
||||||
mixVersionsSnapshotTestByMockPartition(t, tc.clusterVersion, tc.mockPartitionNodeIndex)
|
mixVersionsSnapshotTestByMockPartition(t, tc.config, mockPartitionNodeIndex)
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func mixVersionsSnapshotTestByMockPartition(t *testing.T, clusterVersion e2e.ClusterVersion, mockPartitionNodeIndex int) {
|
func mixVersionsSnapshotTestByMockPartition(t *testing.T, cfg *e2e.EtcdProcessClusterConfig, mockPartitionNodeIndex int) {
|
||||||
e2e.BeforeTest(t)
|
e2e.BeforeTest(t)
|
||||||
|
|
||||||
if !fileutil.Exist(e2e.BinPath.EtcdLastRelease) {
|
if !fileutil.Exist(e2e.BinPath.EtcdLastRelease) {
|
||||||
t.Skipf("%q does not exist", e2e.BinPath.EtcdLastRelease)
|
t.Skipf("%q does not exist", e2e.BinPath.EtcdLastRelease)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Create an etcd cluster with 3 member of MinorityLastVersion
|
t.Logf("Create an etcd cluster with %d member", cfg.ClusterSize)
|
||||||
epc, err := e2e.NewEtcdProcessCluster(context.TODO(), t,
|
epc, err := e2e.NewEtcdProcessCluster(context.TODO(), t,
|
||||||
e2e.WithClusterSize(3),
|
e2e.WithConfig(cfg),
|
||||||
e2e.WithSnapshotCount(10),
|
e2e.WithSnapshotCount(10),
|
||||||
e2e.WithVersion(clusterVersion),
|
|
||||||
e2e.WithSnapshotCatchUpEntries(10),
|
e2e.WithSnapshotCatchUpEntries(10),
|
||||||
)
|
)
|
||||||
require.NoError(t, err, "failed to start etcd cluster: %v", err)
|
require.NoError(t, err, "failed to start etcd cluster: %v", err)
|
||||||
@ -161,28 +143,44 @@ func mixVersionsSnapshotTestByMockPartition(t *testing.T, clusterVersion e2e.Clu
|
|||||||
}()
|
}()
|
||||||
toPartitionedMember := epc.Procs[mockPartitionNodeIndex]
|
toPartitionedMember := epc.Procs[mockPartitionNodeIndex]
|
||||||
|
|
||||||
// Stop and restart the partitioned member
|
t.Log("Stop and restart the partitioned member")
|
||||||
err = toPartitionedMember.Stop()
|
err = toPartitionedMember.Stop()
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
// Write more than SnapshotCount entries to trigger at least a snapshot.
|
t.Log("Writing 20 keys to the cluster (more than SnapshotCount entries to trigger at least a snapshot)")
|
||||||
t.Log("Writing 20 keys to the cluster")
|
writeKVs(t, epc.Etcdctl(), 0, 20)
|
||||||
for i := 0; i < 20; i++ {
|
|
||||||
key := fmt.Sprintf("key-%d", i)
|
|
||||||
value := fmt.Sprintf("value-%d", i)
|
|
||||||
err = epc.Etcdctl().Put(context.TODO(), key, value, config.PutOptions{})
|
|
||||||
require.NoError(t, err, "failed to put %q, error: %v", key, err)
|
|
||||||
}
|
|
||||||
|
|
||||||
t.Log("Verify logs to check leader has saved snapshot")
|
t.Log("Verify logs to check leader has saved snapshot")
|
||||||
leaderEPC := epc.Procs[epc.WaitLeader(t)]
|
leaderEPC := epc.Procs[epc.WaitLeader(t)]
|
||||||
e2e.AssertProcessLogs(t, leaderEPC, "saved snapshot")
|
e2e.AssertProcessLogs(t, leaderEPC, "saved snapshot")
|
||||||
|
|
||||||
// Restart the partitioned member
|
t.Log("Restart the partitioned member")
|
||||||
err = toPartitionedMember.Restart(context.TODO())
|
err = toPartitionedMember.Restart(context.TODO())
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
// verify all nodes have exact same revision and hash
|
assertKVHash(t, epc)
|
||||||
|
|
||||||
|
leaderEPC = epc.Procs[epc.WaitLeader(t)]
|
||||||
|
if leaderEPC.Config().ExecPath == e2e.BinPath.Etcd {
|
||||||
|
t.Log("Verify logs to check snapshot be sent from leader to follower")
|
||||||
|
e2e.AssertProcessLogs(t, leaderEPC, "sent database snapshot")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func writeKVs(t *testing.T, etcdctl *e2e.EtcdctlV3, startIdx, endIdx int) {
|
||||||
|
for i := startIdx; i < endIdx; i++ {
|
||||||
|
key := fmt.Sprintf("key-%d", i)
|
||||||
|
value := fmt.Sprintf("value-%d", i)
|
||||||
|
err := etcdctl.Put(context.TODO(), key, value, config.PutOptions{})
|
||||||
|
require.NoError(t, err, "failed to put %q, error: %v", key, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func assertKVHash(t *testing.T, epc *e2e.EtcdProcessCluster) {
|
||||||
|
clusterSize := len(epc.Procs)
|
||||||
|
if clusterSize < 2 {
|
||||||
|
return
|
||||||
|
}
|
||||||
t.Log("Verify all nodes have exact same revision and hash")
|
t.Log("Verify all nodes have exact same revision and hash")
|
||||||
assert.Eventually(t, func() bool {
|
assert.Eventually(t, func() bool {
|
||||||
hashKvs, err := epc.Etcdctl().HashKV(context.TODO(), 0)
|
hashKvs, err := epc.Etcdctl().HashKV(context.TODO(), 0)
|
||||||
@ -190,28 +188,18 @@ func mixVersionsSnapshotTestByMockPartition(t *testing.T, clusterVersion e2e.Clu
|
|||||||
t.Logf("failed to get HashKV: %v", err)
|
t.Logf("failed to get HashKV: %v", err)
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
if len(hashKvs) != 3 {
|
if len(hashKvs) != clusterSize {
|
||||||
t.Logf("expected 3 hashkv responses, but got: %d", len(hashKvs))
|
t.Logf("expected %d hashkv responses, but got: %d", clusterSize, len(hashKvs))
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
for i := 1; i < clusterSize; i++ {
|
||||||
|
if hashKvs[0].Header.Revision != hashKvs[i].Header.Revision {
|
||||||
|
t.Logf("Got different revisions, [%d, %d]", hashKvs[0].Header.Revision, hashKvs[1].Header.Revision)
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
if hashKvs[0].Header.Revision != hashKvs[1].Header.Revision {
|
assert.Equal(t, hashKvs[0].Hash, hashKvs[i].Hash)
|
||||||
t.Logf("Got different revisions, [%d, %d]", hashKvs[0].Header.Revision, hashKvs[1].Header.Revision)
|
|
||||||
return false
|
|
||||||
}
|
}
|
||||||
if hashKvs[1].Header.Revision != hashKvs[2].Header.Revision {
|
|
||||||
t.Logf("Got different revisions, [%d, %d]", hashKvs[1].Header.Revision, hashKvs[2].Header.Revision)
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
|
|
||||||
assert.Equal(t, hashKvs[0].Hash, hashKvs[1].Hash)
|
|
||||||
assert.Equal(t, hashKvs[1].Hash, hashKvs[2].Hash)
|
|
||||||
|
|
||||||
return true
|
return true
|
||||||
}, 10*time.Second, 500*time.Millisecond)
|
}, 10*time.Second, 500*time.Millisecond)
|
||||||
|
|
||||||
// assert process logs to check snapshot be sent
|
|
||||||
t.Log("Verify logs to check snapshot be sent from leader to follower")
|
|
||||||
leaderEPC = epc.Procs[epc.WaitLeader(t)]
|
|
||||||
e2e.AssertProcessLogs(t, leaderEPC, "sent database snapshot")
|
|
||||||
}
|
}
|
||||||
|
Loading…
x
Reference in New Issue
Block a user