mirror of
https://github.com/etcd-io/etcd.git
synced 2024-09-27 06:25:44 +00:00
Merge pull request #9765 from gyuho/watch-restore
mvcc: watcherGroup chooseAll panic (skip proxy tests for now)
This commit is contained in:
commit
b4f84f046b
105
integration/v3_watch_restore_test.go
Normal file
105
integration/v3_watch_restore_test.go
Normal file
@ -0,0 +1,105 @@
|
||||
// Copyright 2018 The etcd Authors
|
||||
//
|
||||
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||
// you may not use this file except in compliance with the License.
|
||||
// You may obtain a copy of the License at
|
||||
//
|
||||
// http://www.apache.org/licenses/LICENSE-2.0
|
||||
//
|
||||
// Unless required by applicable law or agreed to in writing, software
|
||||
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
// See the License for the specific language governing permissions and
|
||||
// limitations under the License.
|
||||
|
||||
// +build !cluster_proxy
|
||||
|
||||
package integration
|
||||
|
||||
import (
|
||||
"context"
|
||||
"fmt"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
pb "github.com/coreos/etcd/etcdserver/etcdserverpb"
|
||||
)
|
||||
|
||||
// TestV3WatchRestoreSnapshotUnsync tests whether slow follower can restore
|
||||
// from leader snapshot, and still notify on watchers from an old revision
|
||||
// that were created in synced watcher group in the first place.
|
||||
// TODO: fix panic with gRPC proxy "panic: watcher current revision should not exceed current revision"
|
||||
func TestV3WatchRestoreSnapshotUnsync(t *testing.T) {
|
||||
clus := NewClusterV3(t, &ClusterConfig{
|
||||
Size: 3,
|
||||
SnapshotCount: 10,
|
||||
SnapshotCatchUpEntries: 5,
|
||||
})
|
||||
defer clus.Terminate(t)
|
||||
|
||||
// spawn a watcher before shutdown, and put it in synced watcher
|
||||
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
||||
defer cancel()
|
||||
wStream, errW := toGRPC(clus.Client(0)).Watch.Watch(ctx)
|
||||
if errW != nil {
|
||||
t.Fatal(errW)
|
||||
}
|
||||
if err := wStream.Send(&pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
||||
CreateRequest: &pb.WatchCreateRequest{Key: []byte("foo"), StartRevision: 5}}}); err != nil {
|
||||
t.Fatalf("wStream.Send error: %v", err)
|
||||
}
|
||||
wresp, errR := wStream.Recv()
|
||||
if errR != nil {
|
||||
t.Errorf("wStream.Recv error: %v", errR)
|
||||
}
|
||||
if !wresp.Created {
|
||||
t.Errorf("wresp.Created got = %v, want = true", wresp.Created)
|
||||
}
|
||||
|
||||
clus.Members[0].InjectPartition(t, clus.Members[1:]...)
|
||||
clus.waitLeader(t, clus.Members[1:])
|
||||
time.Sleep(2 * time.Second)
|
||||
|
||||
kvc := toGRPC(clus.Client(1)).KV
|
||||
|
||||
// to trigger snapshot from the leader to the stopped follower
|
||||
for i := 0; i < 15; i++ {
|
||||
_, err := kvc.Put(context.TODO(), &pb.PutRequest{Key: []byte("foo"), Value: []byte("bar")})
|
||||
if err != nil {
|
||||
t.Errorf("#%d: couldn't put key (%v)", i, err)
|
||||
}
|
||||
}
|
||||
|
||||
// trigger snapshot send from leader to this slow follower
|
||||
// which then calls watchable store Restore
|
||||
clus.Members[0].RecoverPartition(t, clus.Members[1:]...)
|
||||
clus.WaitLeader(t)
|
||||
time.Sleep(2 * time.Second)
|
||||
|
||||
// slow follower now applies leader snapshot
|
||||
// should be able to notify on old-revision watchers in unsynced
|
||||
// make sure restore watch operation correctly moves watchers
|
||||
// between synced and unsynced watchers
|
||||
errc := make(chan error)
|
||||
go func() {
|
||||
cresp, cerr := wStream.Recv()
|
||||
if cerr != nil {
|
||||
errc <- cerr
|
||||
return
|
||||
}
|
||||
// from start revision 5 to latest revision 16
|
||||
if len(cresp.Events) != 12 {
|
||||
errc <- fmt.Errorf("expected 12 events, got %+v", cresp.Events)
|
||||
return
|
||||
}
|
||||
errc <- nil
|
||||
}()
|
||||
select {
|
||||
case <-time.After(10 * time.Second):
|
||||
t.Fatal("took too long to receive events from restored watcher")
|
||||
case err := <-errc:
|
||||
if err != nil {
|
||||
t.Fatalf("wStream.Recv error: %v", err)
|
||||
}
|
||||
}
|
||||
}
|
@ -352,84 +352,6 @@ func TestV3WatchFutureRevision(t *testing.T) {
|
||||
}
|
||||
}
|
||||
|
||||
// TestV3WatchRestoreSnapshotUnsync tests whether slow follower can restore
|
||||
// from leader snapshot, and still notify on watchers from an old revision
|
||||
// that were created in synced watcher group in the first place.
|
||||
func TestV3WatchRestoreSnapshotUnsync(t *testing.T) {
|
||||
clus := NewClusterV3(t, &ClusterConfig{
|
||||
Size: 3,
|
||||
SnapshotCount: 10,
|
||||
SnapshotCatchUpEntries: 5,
|
||||
})
|
||||
defer clus.Terminate(t)
|
||||
|
||||
// spawn a watcher before shutdown, and put it in synced watcher
|
||||
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
||||
defer cancel()
|
||||
wStream, errW := toGRPC(clus.Client(0)).Watch.Watch(ctx)
|
||||
if errW != nil {
|
||||
t.Fatal(errW)
|
||||
}
|
||||
if err := wStream.Send(&pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
||||
CreateRequest: &pb.WatchCreateRequest{Key: []byte("foo"), StartRevision: 5}}}); err != nil {
|
||||
t.Fatalf("wStream.Send error: %v", err)
|
||||
}
|
||||
wresp, errR := wStream.Recv()
|
||||
if errR != nil {
|
||||
t.Errorf("wStream.Recv error: %v", errR)
|
||||
}
|
||||
if !wresp.Created {
|
||||
t.Errorf("wresp.Created got = %v, want = true", wresp.Created)
|
||||
}
|
||||
|
||||
clus.Members[0].InjectPartition(t, clus.Members[1:]...)
|
||||
clus.waitLeader(t, clus.Members[1:])
|
||||
time.Sleep(2 * time.Second)
|
||||
|
||||
kvc := toGRPC(clus.Client(1)).KV
|
||||
|
||||
// to trigger snapshot from the leader to the stopped follower
|
||||
for i := 0; i < 15; i++ {
|
||||
_, err := kvc.Put(context.TODO(), &pb.PutRequest{Key: []byte("foo"), Value: []byte("bar")})
|
||||
if err != nil {
|
||||
t.Errorf("#%d: couldn't put key (%v)", i, err)
|
||||
}
|
||||
}
|
||||
|
||||
// trigger snapshot send from leader to this slow follower
|
||||
// which then calls watchable store Restore
|
||||
clus.Members[0].RecoverPartition(t, clus.Members[1:]...)
|
||||
clus.WaitLeader(t)
|
||||
time.Sleep(2 * time.Second)
|
||||
|
||||
// slow follower now applies leader snapshot
|
||||
// should be able to notify on old-revision watchers in unsynced
|
||||
// make sure restore watch operation correctly moves watchers
|
||||
// between synced and unsynced watchers
|
||||
errc := make(chan error)
|
||||
go func() {
|
||||
cresp, cerr := wStream.Recv()
|
||||
if cerr != nil {
|
||||
errc <- cerr
|
||||
return
|
||||
}
|
||||
// from start revision 5 to latest revision 16
|
||||
if len(cresp.Events) != 12 {
|
||||
errc <- fmt.Errorf("expected 12 events, got %+v", cresp.Events)
|
||||
return
|
||||
}
|
||||
errc <- nil
|
||||
}()
|
||||
select {
|
||||
case <-time.After(10 * time.Second):
|
||||
t.Fatal("took too long to receive events from restored watcher")
|
||||
case err := <-errc:
|
||||
if err != nil {
|
||||
t.Fatalf("wStream.Recv error: %v", err)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// TestV3WatchWrongRange tests wrong range does not create watchers.
|
||||
func TestV3WatchWrongRange(t *testing.T) {
|
||||
defer testutil.AfterTest(t)
|
||||
|
@ -15,6 +15,7 @@
|
||||
package mvcc
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"math"
|
||||
|
||||
"github.com/coreos/etcd/mvcc/mvccpb"
|
||||
@ -238,7 +239,7 @@ func (wg *watcherGroup) chooseAll(curRev, compactRev int64) int64 {
|
||||
minRev := int64(math.MaxInt64)
|
||||
for w := range wg.watchers {
|
||||
if w.minRev > curRev {
|
||||
panic("watcher current revision should not exceed current revision")
|
||||
panic(fmt.Errorf("watcher minimum revision %d should not exceed current revision %d", w.minRev, curRev))
|
||||
}
|
||||
if w.minRev < compactRev {
|
||||
select {
|
||||
|
Loading…
x
Reference in New Issue
Block a user