mirror of
https://github.com/etcd-io/etcd.git
synced 2024-09-27 06:25:44 +00:00
1298 lines
35 KiB
Go
1298 lines
35 KiB
Go
// Copyright 2016 The etcd Authors
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package integration
|
|
|
|
import (
|
|
"bytes"
|
|
"context"
|
|
"fmt"
|
|
"reflect"
|
|
"sort"
|
|
"sync"
|
|
"testing"
|
|
"time"
|
|
|
|
pb "go.etcd.io/etcd/api/v3/etcdserverpb"
|
|
"go.etcd.io/etcd/api/v3/mvccpb"
|
|
"go.etcd.io/etcd/server/v3/etcdserver/api/v3rpc"
|
|
)
|
|
|
|
// TestV3WatchFromCurrentRevision tests Watch APIs from current revision.
|
|
func TestV3WatchFromCurrentRevision(t *testing.T) {
|
|
BeforeTest(t)
|
|
tests := []struct {
|
|
putKeys []string
|
|
watchRequest *pb.WatchRequest
|
|
|
|
wresps []*pb.WatchResponse
|
|
}{
|
|
// watch the key, matching
|
|
{
|
|
[]string{"foo"},
|
|
&pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: &pb.WatchCreateRequest{
|
|
Key: []byte("foo")}}},
|
|
|
|
[]*pb.WatchResponse{
|
|
{
|
|
Header: &pb.ResponseHeader{Revision: 2},
|
|
Created: false,
|
|
Events: []*mvccpb.Event{
|
|
{
|
|
Type: mvccpb.PUT,
|
|
Kv: &mvccpb.KeyValue{Key: []byte("foo"), Value: []byte("bar"), CreateRevision: 2, ModRevision: 2, Version: 1},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
// watch the key, non-matching
|
|
{
|
|
[]string{"foo"},
|
|
&pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: &pb.WatchCreateRequest{
|
|
Key: []byte("helloworld")}}},
|
|
|
|
[]*pb.WatchResponse{},
|
|
},
|
|
// watch the prefix, matching
|
|
{
|
|
[]string{"fooLong"},
|
|
&pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: &pb.WatchCreateRequest{
|
|
Key: []byte("foo"),
|
|
RangeEnd: []byte("fop")}}},
|
|
|
|
[]*pb.WatchResponse{
|
|
{
|
|
Header: &pb.ResponseHeader{Revision: 2},
|
|
Created: false,
|
|
Events: []*mvccpb.Event{
|
|
{
|
|
Type: mvccpb.PUT,
|
|
Kv: &mvccpb.KeyValue{Key: []byte("fooLong"), Value: []byte("bar"), CreateRevision: 2, ModRevision: 2, Version: 1},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
// watch the prefix, non-matching
|
|
{
|
|
[]string{"foo"},
|
|
&pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: &pb.WatchCreateRequest{
|
|
Key: []byte("helloworld"),
|
|
RangeEnd: []byte("helloworle")}}},
|
|
|
|
[]*pb.WatchResponse{},
|
|
},
|
|
// watch full range, matching
|
|
{
|
|
[]string{"fooLong"},
|
|
&pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: &pb.WatchCreateRequest{
|
|
Key: []byte(""),
|
|
RangeEnd: []byte("\x00")}}},
|
|
|
|
[]*pb.WatchResponse{
|
|
{
|
|
Header: &pb.ResponseHeader{Revision: 2},
|
|
Created: false,
|
|
Events: []*mvccpb.Event{
|
|
{
|
|
Type: mvccpb.PUT,
|
|
Kv: &mvccpb.KeyValue{Key: []byte("fooLong"), Value: []byte("bar"), CreateRevision: 2, ModRevision: 2, Version: 1},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
// multiple puts, one watcher with matching key
|
|
{
|
|
[]string{"foo", "foo", "foo"},
|
|
&pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: &pb.WatchCreateRequest{
|
|
Key: []byte("foo")}}},
|
|
|
|
[]*pb.WatchResponse{
|
|
{
|
|
Header: &pb.ResponseHeader{Revision: 2},
|
|
Created: false,
|
|
Events: []*mvccpb.Event{
|
|
{
|
|
Type: mvccpb.PUT,
|
|
Kv: &mvccpb.KeyValue{Key: []byte("foo"), Value: []byte("bar"), CreateRevision: 2, ModRevision: 2, Version: 1},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
Header: &pb.ResponseHeader{Revision: 3},
|
|
Created: false,
|
|
Events: []*mvccpb.Event{
|
|
{
|
|
Type: mvccpb.PUT,
|
|
Kv: &mvccpb.KeyValue{Key: []byte("foo"), Value: []byte("bar"), CreateRevision: 2, ModRevision: 3, Version: 2},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
Header: &pb.ResponseHeader{Revision: 4},
|
|
Created: false,
|
|
Events: []*mvccpb.Event{
|
|
{
|
|
Type: mvccpb.PUT,
|
|
Kv: &mvccpb.KeyValue{Key: []byte("foo"), Value: []byte("bar"), CreateRevision: 2, ModRevision: 4, Version: 3},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
// multiple puts, one watcher with matching prefix
|
|
{
|
|
[]string{"foo", "foo", "foo"},
|
|
&pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: &pb.WatchCreateRequest{
|
|
Key: []byte("foo"),
|
|
RangeEnd: []byte("fop")}}},
|
|
|
|
[]*pb.WatchResponse{
|
|
{
|
|
Header: &pb.ResponseHeader{Revision: 2},
|
|
Created: false,
|
|
Events: []*mvccpb.Event{
|
|
{
|
|
Type: mvccpb.PUT,
|
|
Kv: &mvccpb.KeyValue{Key: []byte("foo"), Value: []byte("bar"), CreateRevision: 2, ModRevision: 2, Version: 1},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
Header: &pb.ResponseHeader{Revision: 3},
|
|
Created: false,
|
|
Events: []*mvccpb.Event{
|
|
{
|
|
Type: mvccpb.PUT,
|
|
Kv: &mvccpb.KeyValue{Key: []byte("foo"), Value: []byte("bar"), CreateRevision: 2, ModRevision: 3, Version: 2},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
Header: &pb.ResponseHeader{Revision: 4},
|
|
Created: false,
|
|
Events: []*mvccpb.Event{
|
|
{
|
|
Type: mvccpb.PUT,
|
|
Kv: &mvccpb.KeyValue{Key: []byte("foo"), Value: []byte("bar"), CreateRevision: 2, ModRevision: 4, Version: 3},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
for i, tt := range tests {
|
|
clus := NewClusterV3(t, &ClusterConfig{Size: 3})
|
|
|
|
wAPI := toGRPC(clus.RandClient()).Watch
|
|
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
|
defer cancel()
|
|
wStream, err := wAPI.Watch(ctx)
|
|
if err != nil {
|
|
t.Fatalf("#%d: wAPI.Watch error: %v", i, err)
|
|
}
|
|
|
|
err = wStream.Send(tt.watchRequest)
|
|
if err != nil {
|
|
t.Fatalf("#%d: wStream.Send error: %v", i, err)
|
|
}
|
|
|
|
// ensure watcher request created a new watcher
|
|
cresp, err := wStream.Recv()
|
|
if err != nil {
|
|
t.Errorf("#%d: wStream.Recv error: %v", i, err)
|
|
clus.Terminate(t)
|
|
continue
|
|
}
|
|
if !cresp.Created {
|
|
t.Errorf("#%d: did not create watchid, got %+v", i, cresp)
|
|
clus.Terminate(t)
|
|
continue
|
|
}
|
|
if cresp.Canceled {
|
|
t.Errorf("#%d: canceled watcher on create %+v", i, cresp)
|
|
clus.Terminate(t)
|
|
continue
|
|
}
|
|
|
|
createdWatchId := cresp.WatchId
|
|
if cresp.Header == nil || cresp.Header.Revision != 1 {
|
|
t.Errorf("#%d: header revision got +%v, wanted revison 1", i, cresp)
|
|
clus.Terminate(t)
|
|
continue
|
|
}
|
|
|
|
// asynchronously create keys
|
|
ch := make(chan struct{}, 1)
|
|
go func() {
|
|
for _, k := range tt.putKeys {
|
|
kvc := toGRPC(clus.RandClient()).KV
|
|
req := &pb.PutRequest{Key: []byte(k), Value: []byte("bar")}
|
|
if _, err := kvc.Put(context.TODO(), req); err != nil {
|
|
t.Errorf("#%d: couldn't put key (%v)", i, err)
|
|
}
|
|
}
|
|
ch <- struct{}{}
|
|
}()
|
|
|
|
// check stream results
|
|
for j, wresp := range tt.wresps {
|
|
resp, err := wStream.Recv()
|
|
if err != nil {
|
|
t.Errorf("#%d.%d: wStream.Recv error: %v", i, j, err)
|
|
}
|
|
|
|
if resp.Header == nil {
|
|
t.Fatalf("#%d.%d: unexpected nil resp.Header", i, j)
|
|
}
|
|
if resp.Header.Revision != wresp.Header.Revision {
|
|
t.Errorf("#%d.%d: resp.Header.Revision got = %d, want = %d", i, j, resp.Header.Revision, wresp.Header.Revision)
|
|
}
|
|
|
|
if wresp.Created != resp.Created {
|
|
t.Errorf("#%d.%d: resp.Created got = %v, want = %v", i, j, resp.Created, wresp.Created)
|
|
}
|
|
if resp.WatchId != createdWatchId {
|
|
t.Errorf("#%d.%d: resp.WatchId got = %d, want = %d", i, j, resp.WatchId, createdWatchId)
|
|
}
|
|
|
|
if !reflect.DeepEqual(resp.Events, wresp.Events) {
|
|
t.Errorf("#%d.%d: resp.Events got = %+v, want = %+v", i, j, resp.Events, wresp.Events)
|
|
}
|
|
}
|
|
|
|
rok, nr := waitResponse(wStream, 1*time.Second)
|
|
if !rok {
|
|
t.Errorf("unexpected pb.WatchResponse is received %+v", nr)
|
|
}
|
|
|
|
// wait for the client to finish sending the keys before terminating the cluster
|
|
<-ch
|
|
|
|
// can't defer because tcp ports will be in use
|
|
clus.Terminate(t)
|
|
}
|
|
}
|
|
|
|
// TestV3WatchFutureRevision tests Watch APIs from a future revision.
|
|
func TestV3WatchFutureRevision(t *testing.T) {
|
|
BeforeTest(t)
|
|
|
|
clus := NewClusterV3(t, &ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
wAPI := toGRPC(clus.RandClient()).Watch
|
|
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
|
defer cancel()
|
|
wStream, err := wAPI.Watch(ctx)
|
|
if err != nil {
|
|
t.Fatalf("wAPI.Watch error: %v", err)
|
|
}
|
|
|
|
wkey := []byte("foo")
|
|
wrev := int64(10)
|
|
req := &pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: &pb.WatchCreateRequest{Key: wkey, StartRevision: wrev}}}
|
|
err = wStream.Send(req)
|
|
if err != nil {
|
|
t.Fatalf("wStream.Send error: %v", err)
|
|
}
|
|
|
|
// ensure watcher request created a new watcher
|
|
cresp, err := wStream.Recv()
|
|
if err != nil {
|
|
t.Fatalf("wStream.Recv error: %v", err)
|
|
}
|
|
if !cresp.Created {
|
|
t.Fatalf("create %v, want %v", cresp.Created, true)
|
|
}
|
|
|
|
kvc := toGRPC(clus.RandClient()).KV
|
|
for {
|
|
req := &pb.PutRequest{Key: wkey, Value: []byte("bar")}
|
|
resp, rerr := kvc.Put(context.TODO(), req)
|
|
if rerr != nil {
|
|
t.Fatalf("couldn't put key (%v)", rerr)
|
|
}
|
|
if resp.Header.Revision == wrev {
|
|
break
|
|
}
|
|
}
|
|
|
|
// ensure watcher request created a new watcher
|
|
cresp, err = wStream.Recv()
|
|
if err != nil {
|
|
t.Fatalf("wStream.Recv error: %v", err)
|
|
}
|
|
if cresp.Header.Revision != wrev {
|
|
t.Fatalf("revision = %d, want %d", cresp.Header.Revision, wrev)
|
|
}
|
|
if len(cresp.Events) != 1 {
|
|
t.Fatalf("failed to receive events")
|
|
}
|
|
if cresp.Events[0].Kv.ModRevision != wrev {
|
|
t.Errorf("mod revision = %d, want %d", cresp.Events[0].Kv.ModRevision, wrev)
|
|
}
|
|
}
|
|
|
|
// TestV3WatchWrongRange tests wrong range does not create watchers.
|
|
func TestV3WatchWrongRange(t *testing.T) {
|
|
BeforeTest(t)
|
|
|
|
clus := NewClusterV3(t, &ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
wAPI := toGRPC(clus.RandClient()).Watch
|
|
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
|
defer cancel()
|
|
wStream, err := wAPI.Watch(ctx)
|
|
if err != nil {
|
|
t.Fatalf("wAPI.Watch error: %v", err)
|
|
}
|
|
|
|
tests := []struct {
|
|
key []byte
|
|
end []byte
|
|
canceled bool
|
|
}{
|
|
{[]byte("a"), []byte("a"), true}, // wrong range end
|
|
{[]byte("b"), []byte("a"), true}, // wrong range end
|
|
{[]byte("foo"), []byte{0}, false}, // watch request with 'WithFromKey'
|
|
}
|
|
for i, tt := range tests {
|
|
if err := wStream.Send(&pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: &pb.WatchCreateRequest{Key: tt.key, RangeEnd: tt.end, StartRevision: 1}}}); err != nil {
|
|
t.Fatalf("#%d: wStream.Send error: %v", i, err)
|
|
}
|
|
cresp, err := wStream.Recv()
|
|
if err != nil {
|
|
t.Fatalf("#%d: wStream.Recv error: %v", i, err)
|
|
}
|
|
if !cresp.Created {
|
|
t.Fatalf("#%d: create %v, want %v", i, cresp.Created, true)
|
|
}
|
|
if cresp.Canceled != tt.canceled {
|
|
t.Fatalf("#%d: canceled %v, want %v", i, tt.canceled, cresp.Canceled)
|
|
}
|
|
if tt.canceled && cresp.WatchId != -1 {
|
|
t.Fatalf("#%d: canceled watch ID %d, want -1", i, cresp.WatchId)
|
|
}
|
|
}
|
|
}
|
|
|
|
// TestV3WatchCancelSynced tests Watch APIs cancellation from synced map.
|
|
func TestV3WatchCancelSynced(t *testing.T) {
|
|
BeforeTest(t)
|
|
testV3WatchCancel(t, 0)
|
|
}
|
|
|
|
// TestV3WatchCancelUnsynced tests Watch APIs cancellation from unsynced map.
|
|
func TestV3WatchCancelUnsynced(t *testing.T) {
|
|
BeforeTest(t)
|
|
testV3WatchCancel(t, 1)
|
|
}
|
|
|
|
func testV3WatchCancel(t *testing.T, startRev int64) {
|
|
clus := NewClusterV3(t, &ClusterConfig{Size: 3})
|
|
defer clus.Terminate(t)
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
|
defer cancel()
|
|
wStream, errW := toGRPC(clus.RandClient()).Watch.Watch(ctx)
|
|
if errW != nil {
|
|
t.Fatalf("wAPI.Watch error: %v", errW)
|
|
}
|
|
|
|
wreq := &pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: &pb.WatchCreateRequest{
|
|
Key: []byte("foo"), StartRevision: startRev}}}
|
|
if err := wStream.Send(wreq); err != nil {
|
|
t.Fatalf("wStream.Send error: %v", err)
|
|
}
|
|
|
|
wresp, errR := wStream.Recv()
|
|
if errR != nil {
|
|
t.Errorf("wStream.Recv error: %v", errR)
|
|
}
|
|
if !wresp.Created {
|
|
t.Errorf("wresp.Created got = %v, want = true", wresp.Created)
|
|
}
|
|
|
|
creq := &pb.WatchRequest{RequestUnion: &pb.WatchRequest_CancelRequest{
|
|
CancelRequest: &pb.WatchCancelRequest{
|
|
WatchId: wresp.WatchId}}}
|
|
if err := wStream.Send(creq); err != nil {
|
|
t.Fatalf("wStream.Send error: %v", err)
|
|
}
|
|
|
|
cresp, err := wStream.Recv()
|
|
if err != nil {
|
|
t.Errorf("wStream.Recv error: %v", err)
|
|
}
|
|
if !cresp.Canceled {
|
|
t.Errorf("cresp.Canceled got = %v, want = true", cresp.Canceled)
|
|
}
|
|
|
|
kvc := toGRPC(clus.RandClient()).KV
|
|
if _, err := kvc.Put(context.TODO(), &pb.PutRequest{Key: []byte("foo"), Value: []byte("bar")}); err != nil {
|
|
t.Errorf("couldn't put key (%v)", err)
|
|
}
|
|
|
|
// watch got canceled, so this should block
|
|
rok, nr := waitResponse(wStream, 1*time.Second)
|
|
if !rok {
|
|
t.Errorf("unexpected pb.WatchResponse is received %+v", nr)
|
|
}
|
|
}
|
|
|
|
// TestV3WatchCurrentPutOverlap ensures current watchers receive all events with
|
|
// overlapping puts.
|
|
func TestV3WatchCurrentPutOverlap(t *testing.T) {
|
|
BeforeTest(t)
|
|
clus := NewClusterV3(t, &ClusterConfig{Size: 3})
|
|
defer clus.Terminate(t)
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
|
defer cancel()
|
|
wStream, wErr := toGRPC(clus.RandClient()).Watch.Watch(ctx)
|
|
if wErr != nil {
|
|
t.Fatalf("wAPI.Watch error: %v", wErr)
|
|
}
|
|
|
|
// last mod_revision that will be observed
|
|
nrRevisions := 32
|
|
// first revision already allocated as empty revision
|
|
var wg sync.WaitGroup
|
|
for i := 1; i < nrRevisions; i++ {
|
|
wg.Add(1)
|
|
go func() {
|
|
defer wg.Done()
|
|
kvc := toGRPC(clus.RandClient()).KV
|
|
req := &pb.PutRequest{Key: []byte("foo"), Value: []byte("bar")}
|
|
if _, err := kvc.Put(context.TODO(), req); err != nil {
|
|
t.Errorf("couldn't put key (%v)", err)
|
|
}
|
|
}()
|
|
}
|
|
|
|
// maps watcher to current expected revision
|
|
progress := make(map[int64]int64)
|
|
|
|
wreq := &pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: &pb.WatchCreateRequest{Key: []byte("foo"), RangeEnd: []byte("fop")}}}
|
|
if err := wStream.Send(wreq); err != nil {
|
|
t.Fatalf("first watch request failed (%v)", err)
|
|
}
|
|
|
|
more := true
|
|
progress[-1] = 0 // watcher creation pending
|
|
for more {
|
|
resp, err := wStream.Recv()
|
|
if err != nil {
|
|
t.Fatalf("wStream.Recv error: %v", err)
|
|
}
|
|
|
|
if resp.Created {
|
|
// accept events > header revision
|
|
progress[resp.WatchId] = resp.Header.Revision + 1
|
|
if resp.Header.Revision == int64(nrRevisions) {
|
|
// covered all revisions; create no more watchers
|
|
progress[-1] = int64(nrRevisions) + 1
|
|
} else if err := wStream.Send(wreq); err != nil {
|
|
t.Fatalf("watch request failed (%v)", err)
|
|
}
|
|
} else if len(resp.Events) == 0 {
|
|
t.Fatalf("got events %v, want non-empty", resp.Events)
|
|
} else {
|
|
wRev, ok := progress[resp.WatchId]
|
|
if !ok {
|
|
t.Fatalf("got %+v, but watch id shouldn't exist ", resp)
|
|
}
|
|
if resp.Events[0].Kv.ModRevision != wRev {
|
|
t.Fatalf("got %+v, wanted first revision %d", resp, wRev)
|
|
}
|
|
lastRev := resp.Events[len(resp.Events)-1].Kv.ModRevision
|
|
progress[resp.WatchId] = lastRev + 1
|
|
}
|
|
more = false
|
|
for _, v := range progress {
|
|
if v <= int64(nrRevisions) {
|
|
more = true
|
|
break
|
|
}
|
|
}
|
|
}
|
|
|
|
if rok, nr := waitResponse(wStream, time.Second); !rok {
|
|
t.Errorf("unexpected pb.WatchResponse is received %+v", nr)
|
|
}
|
|
|
|
wg.Wait()
|
|
}
|
|
|
|
// TestV3WatchEmptyKey ensures synced watchers see empty key PUTs as PUT events
|
|
func TestV3WatchEmptyKey(t *testing.T) {
|
|
BeforeTest(t)
|
|
|
|
clus := NewClusterV3(t, &ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
|
defer cancel()
|
|
|
|
ws, werr := toGRPC(clus.RandClient()).Watch.Watch(ctx)
|
|
if werr != nil {
|
|
t.Fatal(werr)
|
|
}
|
|
req := &pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: &pb.WatchCreateRequest{
|
|
Key: []byte("foo")}}}
|
|
if err := ws.Send(req); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if _, err := ws.Recv(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// put a key with empty value
|
|
kvc := toGRPC(clus.RandClient()).KV
|
|
preq := &pb.PutRequest{Key: []byte("foo")}
|
|
if _, err := kvc.Put(context.TODO(), preq); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// check received PUT
|
|
resp, rerr := ws.Recv()
|
|
if rerr != nil {
|
|
t.Fatal(rerr)
|
|
}
|
|
wevs := []*mvccpb.Event{
|
|
{
|
|
Type: mvccpb.PUT,
|
|
Kv: &mvccpb.KeyValue{Key: []byte("foo"), CreateRevision: 2, ModRevision: 2, Version: 1},
|
|
},
|
|
}
|
|
if !reflect.DeepEqual(resp.Events, wevs) {
|
|
t.Fatalf("got %v, expected %v", resp.Events, wevs)
|
|
}
|
|
}
|
|
|
|
func TestV3WatchMultipleWatchersSynced(t *testing.T) {
|
|
BeforeTest(t)
|
|
testV3WatchMultipleWatchers(t, 0)
|
|
}
|
|
|
|
func TestV3WatchMultipleWatchersUnsynced(t *testing.T) {
|
|
BeforeTest(t)
|
|
testV3WatchMultipleWatchers(t, 1)
|
|
}
|
|
|
|
// testV3WatchMultipleWatchers tests multiple watchers on the same key
|
|
// and one watcher with matching prefix. It first puts the key
|
|
// that matches all watchers, and another key that matches only
|
|
// one watcher to test if it receives expected events.
|
|
func testV3WatchMultipleWatchers(t *testing.T, startRev int64) {
|
|
clus := NewClusterV3(t, &ClusterConfig{Size: 3})
|
|
defer clus.Terminate(t)
|
|
|
|
kvc := toGRPC(clus.RandClient()).KV
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
|
defer cancel()
|
|
wStream, errW := toGRPC(clus.RandClient()).Watch.Watch(ctx)
|
|
if errW != nil {
|
|
t.Fatalf("wAPI.Watch error: %v", errW)
|
|
}
|
|
|
|
watchKeyN := 4
|
|
for i := 0; i < watchKeyN+1; i++ {
|
|
var wreq *pb.WatchRequest
|
|
if i < watchKeyN {
|
|
wreq = &pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: &pb.WatchCreateRequest{
|
|
Key: []byte("foo"), StartRevision: startRev}}}
|
|
} else {
|
|
wreq = &pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: &pb.WatchCreateRequest{
|
|
Key: []byte("fo"), RangeEnd: []byte("fp"), StartRevision: startRev}}}
|
|
}
|
|
if err := wStream.Send(wreq); err != nil {
|
|
t.Fatalf("wStream.Send error: %v", err)
|
|
}
|
|
}
|
|
|
|
ids := make(map[int64]struct{})
|
|
for i := 0; i < watchKeyN+1; i++ {
|
|
wresp, err := wStream.Recv()
|
|
if err != nil {
|
|
t.Fatalf("wStream.Recv error: %v", err)
|
|
}
|
|
if !wresp.Created {
|
|
t.Fatalf("wresp.Created got = %v, want = true", wresp.Created)
|
|
}
|
|
ids[wresp.WatchId] = struct{}{}
|
|
}
|
|
|
|
if _, err := kvc.Put(context.TODO(), &pb.PutRequest{Key: []byte("foo"), Value: []byte("bar")}); err != nil {
|
|
t.Fatalf("couldn't put key (%v)", err)
|
|
}
|
|
|
|
for i := 0; i < watchKeyN+1; i++ {
|
|
wresp, err := wStream.Recv()
|
|
if err != nil {
|
|
t.Fatalf("wStream.Recv error: %v", err)
|
|
}
|
|
if _, ok := ids[wresp.WatchId]; !ok {
|
|
t.Errorf("watchId %d is not created!", wresp.WatchId)
|
|
} else {
|
|
delete(ids, wresp.WatchId)
|
|
}
|
|
if len(wresp.Events) == 0 {
|
|
t.Errorf("#%d: no events received", i)
|
|
}
|
|
for _, ev := range wresp.Events {
|
|
if string(ev.Kv.Key) != "foo" {
|
|
t.Errorf("ev.Kv.Key got = %s, want = foo", ev.Kv.Key)
|
|
}
|
|
if string(ev.Kv.Value) != "bar" {
|
|
t.Errorf("ev.Kv.Value got = %s, want = bar", ev.Kv.Value)
|
|
}
|
|
}
|
|
}
|
|
|
|
// now put one key that has only one matching watcher
|
|
if _, err := kvc.Put(context.TODO(), &pb.PutRequest{Key: []byte("fo"), Value: []byte("bar")}); err != nil {
|
|
t.Fatalf("couldn't put key (%v)", err)
|
|
}
|
|
wresp, err := wStream.Recv()
|
|
if err != nil {
|
|
t.Errorf("wStream.Recv error: %v", err)
|
|
}
|
|
if len(wresp.Events) != 1 {
|
|
t.Fatalf("len(wresp.Events) got = %d, want = 1", len(wresp.Events))
|
|
}
|
|
if string(wresp.Events[0].Kv.Key) != "fo" {
|
|
t.Errorf("wresp.Events[0].Kv.Key got = %s, want = fo", wresp.Events[0].Kv.Key)
|
|
}
|
|
|
|
// now Recv should block because there is no more events coming
|
|
rok, nr := waitResponse(wStream, 1*time.Second)
|
|
if !rok {
|
|
t.Errorf("unexpected pb.WatchResponse is received %+v", nr)
|
|
}
|
|
}
|
|
|
|
func TestV3WatchMultipleEventsTxnSynced(t *testing.T) {
|
|
BeforeTest(t)
|
|
testV3WatchMultipleEventsTxn(t, 0)
|
|
}
|
|
|
|
func TestV3WatchMultipleEventsTxnUnsynced(t *testing.T) {
|
|
BeforeTest(t)
|
|
testV3WatchMultipleEventsTxn(t, 1)
|
|
}
|
|
|
|
// testV3WatchMultipleEventsTxn tests Watch APIs when it receives multiple events.
|
|
func testV3WatchMultipleEventsTxn(t *testing.T, startRev int64) {
|
|
clus := NewClusterV3(t, &ClusterConfig{Size: 3})
|
|
defer clus.Terminate(t)
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
|
defer cancel()
|
|
wStream, wErr := toGRPC(clus.RandClient()).Watch.Watch(ctx)
|
|
if wErr != nil {
|
|
t.Fatalf("wAPI.Watch error: %v", wErr)
|
|
}
|
|
|
|
wreq := &pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: &pb.WatchCreateRequest{
|
|
Key: []byte("foo"), RangeEnd: []byte("fop"), StartRevision: startRev}}}
|
|
if err := wStream.Send(wreq); err != nil {
|
|
t.Fatalf("wStream.Send error: %v", err)
|
|
}
|
|
if resp, err := wStream.Recv(); err != nil || !resp.Created {
|
|
t.Fatalf("create response failed: resp=%v, err=%v", resp, err)
|
|
}
|
|
|
|
kvc := toGRPC(clus.RandClient()).KV
|
|
txn := pb.TxnRequest{}
|
|
for i := 0; i < 3; i++ {
|
|
ru := &pb.RequestOp{}
|
|
ru.Request = &pb.RequestOp_RequestPut{
|
|
RequestPut: &pb.PutRequest{
|
|
Key: []byte(fmt.Sprintf("foo%d", i)), Value: []byte("bar")}}
|
|
txn.Success = append(txn.Success, ru)
|
|
}
|
|
|
|
tresp, err := kvc.Txn(context.Background(), &txn)
|
|
if err != nil {
|
|
t.Fatalf("kvc.Txn error: %v", err)
|
|
}
|
|
if !tresp.Succeeded {
|
|
t.Fatalf("kvc.Txn failed: %+v", tresp)
|
|
}
|
|
|
|
events := []*mvccpb.Event{}
|
|
for len(events) < 3 {
|
|
resp, err := wStream.Recv()
|
|
if err != nil {
|
|
t.Errorf("wStream.Recv error: %v", err)
|
|
}
|
|
events = append(events, resp.Events...)
|
|
}
|
|
sort.Sort(eventsSortByKey(events))
|
|
|
|
wevents := []*mvccpb.Event{
|
|
{
|
|
Type: mvccpb.PUT,
|
|
Kv: &mvccpb.KeyValue{Key: []byte("foo0"), Value: []byte("bar"), CreateRevision: 2, ModRevision: 2, Version: 1},
|
|
},
|
|
{
|
|
Type: mvccpb.PUT,
|
|
Kv: &mvccpb.KeyValue{Key: []byte("foo1"), Value: []byte("bar"), CreateRevision: 2, ModRevision: 2, Version: 1},
|
|
},
|
|
{
|
|
Type: mvccpb.PUT,
|
|
Kv: &mvccpb.KeyValue{Key: []byte("foo2"), Value: []byte("bar"), CreateRevision: 2, ModRevision: 2, Version: 1},
|
|
},
|
|
}
|
|
|
|
if !reflect.DeepEqual(events, wevents) {
|
|
t.Errorf("events got = %+v, want = %+v", events, wevents)
|
|
}
|
|
|
|
rok, nr := waitResponse(wStream, 1*time.Second)
|
|
if !rok {
|
|
t.Errorf("unexpected pb.WatchResponse is received %+v", nr)
|
|
}
|
|
}
|
|
|
|
type eventsSortByKey []*mvccpb.Event
|
|
|
|
func (evs eventsSortByKey) Len() int { return len(evs) }
|
|
func (evs eventsSortByKey) Swap(i, j int) { evs[i], evs[j] = evs[j], evs[i] }
|
|
func (evs eventsSortByKey) Less(i, j int) bool {
|
|
return bytes.Compare(evs[i].Kv.Key, evs[j].Kv.Key) < 0
|
|
}
|
|
|
|
func TestV3WatchMultipleEventsPutUnsynced(t *testing.T) {
|
|
BeforeTest(t)
|
|
clus := NewClusterV3(t, &ClusterConfig{Size: 3})
|
|
defer clus.Terminate(t)
|
|
|
|
kvc := toGRPC(clus.RandClient()).KV
|
|
|
|
if _, err := kvc.Put(context.TODO(), &pb.PutRequest{Key: []byte("foo0"), Value: []byte("bar")}); err != nil {
|
|
t.Fatalf("couldn't put key (%v)", err)
|
|
}
|
|
if _, err := kvc.Put(context.TODO(), &pb.PutRequest{Key: []byte("foo1"), Value: []byte("bar")}); err != nil {
|
|
t.Fatalf("couldn't put key (%v)", err)
|
|
}
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
|
defer cancel()
|
|
wStream, wErr := toGRPC(clus.RandClient()).Watch.Watch(ctx)
|
|
if wErr != nil {
|
|
t.Fatalf("wAPI.Watch error: %v", wErr)
|
|
}
|
|
|
|
wreq := &pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: &pb.WatchCreateRequest{
|
|
Key: []byte("foo"), RangeEnd: []byte("fop"), StartRevision: 1}}}
|
|
if err := wStream.Send(wreq); err != nil {
|
|
t.Fatalf("wStream.Send error: %v", err)
|
|
}
|
|
|
|
if _, err := kvc.Put(context.TODO(), &pb.PutRequest{Key: []byte("foo0"), Value: []byte("bar")}); err != nil {
|
|
t.Fatalf("couldn't put key (%v)", err)
|
|
}
|
|
if _, err := kvc.Put(context.TODO(), &pb.PutRequest{Key: []byte("foo1"), Value: []byte("bar")}); err != nil {
|
|
t.Fatalf("couldn't put key (%v)", err)
|
|
}
|
|
|
|
allWevents := []*mvccpb.Event{
|
|
{
|
|
Type: mvccpb.PUT,
|
|
Kv: &mvccpb.KeyValue{Key: []byte("foo0"), Value: []byte("bar"), CreateRevision: 2, ModRevision: 2, Version: 1},
|
|
},
|
|
{
|
|
Type: mvccpb.PUT,
|
|
Kv: &mvccpb.KeyValue{Key: []byte("foo1"), Value: []byte("bar"), CreateRevision: 3, ModRevision: 3, Version: 1},
|
|
},
|
|
{
|
|
Type: mvccpb.PUT,
|
|
Kv: &mvccpb.KeyValue{Key: []byte("foo0"), Value: []byte("bar"), CreateRevision: 2, ModRevision: 4, Version: 2},
|
|
},
|
|
{
|
|
Type: mvccpb.PUT,
|
|
Kv: &mvccpb.KeyValue{Key: []byte("foo1"), Value: []byte("bar"), CreateRevision: 3, ModRevision: 5, Version: 2},
|
|
},
|
|
}
|
|
|
|
events := []*mvccpb.Event{}
|
|
for len(events) < 4 {
|
|
resp, err := wStream.Recv()
|
|
if err != nil {
|
|
t.Errorf("wStream.Recv error: %v", err)
|
|
}
|
|
if resp.Created {
|
|
continue
|
|
}
|
|
events = append(events, resp.Events...)
|
|
// if PUT requests are committed by now, first receive would return
|
|
// multiple events, but if not, it returns a single event. In SSD,
|
|
// it should return 4 events at once.
|
|
}
|
|
|
|
if !reflect.DeepEqual(events, allWevents) {
|
|
t.Errorf("events got = %+v, want = %+v", events, allWevents)
|
|
}
|
|
|
|
rok, nr := waitResponse(wStream, 1*time.Second)
|
|
if !rok {
|
|
t.Errorf("unexpected pb.WatchResponse is received %+v", nr)
|
|
}
|
|
}
|
|
|
|
func TestV3WatchMultipleStreamsSynced(t *testing.T) {
|
|
BeforeTest(t)
|
|
testV3WatchMultipleStreams(t, 0)
|
|
}
|
|
|
|
func TestV3WatchMultipleStreamsUnsynced(t *testing.T) {
|
|
BeforeTest(t)
|
|
testV3WatchMultipleStreams(t, 1)
|
|
}
|
|
|
|
// testV3WatchMultipleStreams tests multiple watchers on the same key on multiple streams.
|
|
func testV3WatchMultipleStreams(t *testing.T, startRev int64) {
|
|
clus := NewClusterV3(t, &ClusterConfig{Size: 3})
|
|
defer clus.Terminate(t)
|
|
|
|
wAPI := toGRPC(clus.RandClient()).Watch
|
|
kvc := toGRPC(clus.RandClient()).KV
|
|
|
|
streams := make([]pb.Watch_WatchClient, 5)
|
|
for i := range streams {
|
|
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
|
defer cancel()
|
|
wStream, errW := wAPI.Watch(ctx)
|
|
if errW != nil {
|
|
t.Fatalf("wAPI.Watch error: %v", errW)
|
|
}
|
|
wreq := &pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: &pb.WatchCreateRequest{
|
|
Key: []byte("foo"), StartRevision: startRev}}}
|
|
if err := wStream.Send(wreq); err != nil {
|
|
t.Fatalf("wStream.Send error: %v", err)
|
|
}
|
|
streams[i] = wStream
|
|
}
|
|
|
|
for _, wStream := range streams {
|
|
wresp, err := wStream.Recv()
|
|
if err != nil {
|
|
t.Fatalf("wStream.Recv error: %v", err)
|
|
}
|
|
if !wresp.Created {
|
|
t.Fatalf("wresp.Created got = %v, want = true", wresp.Created)
|
|
}
|
|
}
|
|
|
|
if _, err := kvc.Put(context.TODO(), &pb.PutRequest{Key: []byte("foo"), Value: []byte("bar")}); err != nil {
|
|
t.Fatalf("couldn't put key (%v)", err)
|
|
}
|
|
|
|
var wg sync.WaitGroup
|
|
wg.Add(len(streams))
|
|
wevents := []*mvccpb.Event{
|
|
{
|
|
Type: mvccpb.PUT,
|
|
Kv: &mvccpb.KeyValue{Key: []byte("foo"), Value: []byte("bar"), CreateRevision: 2, ModRevision: 2, Version: 1},
|
|
},
|
|
}
|
|
for i := range streams {
|
|
go func(i int) {
|
|
defer wg.Done()
|
|
wStream := streams[i]
|
|
wresp, err := wStream.Recv()
|
|
if err != nil {
|
|
t.Errorf("wStream.Recv error: %v", err)
|
|
}
|
|
if wresp.WatchId != 0 {
|
|
t.Errorf("watchId got = %d, want = 0", wresp.WatchId)
|
|
}
|
|
if !reflect.DeepEqual(wresp.Events, wevents) {
|
|
t.Errorf("wresp.Events got = %+v, want = %+v", wresp.Events, wevents)
|
|
}
|
|
// now Recv should block because there is no more events coming
|
|
rok, nr := waitResponse(wStream, 1*time.Second)
|
|
if !rok {
|
|
t.Errorf("unexpected pb.WatchResponse is received %+v", nr)
|
|
}
|
|
}(i)
|
|
}
|
|
wg.Wait()
|
|
}
|
|
|
|
// waitResponse waits on the given stream for given duration.
|
|
// If there is no more events, true and a nil response will be
|
|
// returned closing the WatchClient stream. Or the response will
|
|
// be returned.
|
|
func waitResponse(wc pb.Watch_WatchClient, timeout time.Duration) (bool, *pb.WatchResponse) {
|
|
rCh := make(chan *pb.WatchResponse, 1)
|
|
donec := make(chan struct{})
|
|
defer close(donec)
|
|
go func() {
|
|
resp, _ := wc.Recv()
|
|
select {
|
|
case rCh <- resp:
|
|
case <-donec:
|
|
}
|
|
}()
|
|
select {
|
|
case nr := <-rCh:
|
|
return false, nr
|
|
case <-time.After(timeout):
|
|
}
|
|
// didn't get response
|
|
wc.CloseSend()
|
|
return true, nil
|
|
}
|
|
|
|
func TestWatchWithProgressNotify(t *testing.T) {
|
|
// accelerate report interval so test terminates quickly
|
|
oldpi := v3rpc.GetProgressReportInterval()
|
|
// using atomics to avoid race warnings
|
|
v3rpc.SetProgressReportInterval(3 * time.Second)
|
|
testInterval := 3 * time.Second
|
|
defer func() { v3rpc.SetProgressReportInterval(oldpi) }()
|
|
|
|
BeforeTest(t)
|
|
clus := NewClusterV3(t, &ClusterConfig{Size: 3})
|
|
defer clus.Terminate(t)
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
|
defer cancel()
|
|
wStream, wErr := toGRPC(clus.RandClient()).Watch.Watch(ctx)
|
|
if wErr != nil {
|
|
t.Fatalf("wAPI.Watch error: %v", wErr)
|
|
}
|
|
|
|
// create two watchers, one with progressNotify set.
|
|
wreq := &pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: &pb.WatchCreateRequest{Key: []byte("foo"), StartRevision: 1, ProgressNotify: true}}}
|
|
if err := wStream.Send(wreq); err != nil {
|
|
t.Fatalf("watch request failed (%v)", err)
|
|
}
|
|
wreq = &pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: &pb.WatchCreateRequest{Key: []byte("foo"), StartRevision: 1}}}
|
|
if err := wStream.Send(wreq); err != nil {
|
|
t.Fatalf("watch request failed (%v)", err)
|
|
}
|
|
|
|
// two creation + one notification
|
|
for i := 0; i < 3; i++ {
|
|
rok, resp := waitResponse(wStream, testInterval+time.Second)
|
|
if resp.Created {
|
|
continue
|
|
}
|
|
|
|
if rok {
|
|
t.Errorf("failed to receive response from watch stream")
|
|
}
|
|
if resp.Header.Revision != 1 {
|
|
t.Errorf("revision = %d, want 1", resp.Header.Revision)
|
|
}
|
|
if len(resp.Events) != 0 {
|
|
t.Errorf("len(resp.Events) = %d, want 0", len(resp.Events))
|
|
}
|
|
}
|
|
|
|
// no more notification
|
|
rok, resp := waitResponse(wStream, time.Second)
|
|
if !rok {
|
|
t.Errorf("unexpected pb.WatchResponse is received %+v", resp)
|
|
}
|
|
}
|
|
|
|
// TestV3WatcMultiOpenhClose opens many watchers concurrently on multiple streams.
|
|
func TestV3WatchClose(t *testing.T) {
|
|
BeforeTest(t)
|
|
clus := NewClusterV3(t, &ClusterConfig{Size: 1, UseBridge: true})
|
|
defer clus.Terminate(t)
|
|
|
|
c := clus.Client(0)
|
|
wapi := toGRPC(c).Watch
|
|
|
|
var wg sync.WaitGroup
|
|
wg.Add(100)
|
|
for i := 0; i < 100; i++ {
|
|
go func() {
|
|
ctx, cancel := context.WithCancel(context.TODO())
|
|
defer func() {
|
|
wg.Done()
|
|
cancel()
|
|
}()
|
|
ws, err := wapi.Watch(ctx)
|
|
if err != nil {
|
|
return
|
|
}
|
|
cr := &pb.WatchCreateRequest{Key: []byte("a")}
|
|
req := &pb.WatchRequest{
|
|
RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: cr}}
|
|
ws.Send(req)
|
|
ws.Recv()
|
|
}()
|
|
}
|
|
|
|
clus.Members[0].Bridge().DropConnections()
|
|
wg.Wait()
|
|
}
|
|
|
|
// TestV3WatchWithFilter ensures watcher filters out the events correctly.
|
|
func TestV3WatchWithFilter(t *testing.T) {
|
|
BeforeTest(t)
|
|
|
|
clus := NewClusterV3(t, &ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
|
defer cancel()
|
|
|
|
ws, werr := toGRPC(clus.RandClient()).Watch.Watch(ctx)
|
|
if werr != nil {
|
|
t.Fatal(werr)
|
|
}
|
|
req := &pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: &pb.WatchCreateRequest{
|
|
Key: []byte("foo"),
|
|
Filters: []pb.WatchCreateRequest_FilterType{pb.WatchCreateRequest_NOPUT},
|
|
}}}
|
|
if err := ws.Send(req); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if _, err := ws.Recv(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
recv := make(chan *pb.WatchResponse, 1)
|
|
go func() {
|
|
// check received PUT
|
|
resp, rerr := ws.Recv()
|
|
if rerr != nil {
|
|
t.Error(rerr)
|
|
}
|
|
recv <- resp
|
|
}()
|
|
|
|
// put a key with empty value
|
|
kvc := toGRPC(clus.RandClient()).KV
|
|
preq := &pb.PutRequest{Key: []byte("foo")}
|
|
if _, err := kvc.Put(context.TODO(), preq); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
select {
|
|
case <-recv:
|
|
t.Fatal("failed to filter out put event")
|
|
case <-time.After(100 * time.Millisecond):
|
|
}
|
|
|
|
dreq := &pb.DeleteRangeRequest{Key: []byte("foo")}
|
|
if _, err := kvc.DeleteRange(context.TODO(), dreq); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
select {
|
|
case resp := <-recv:
|
|
wevs := []*mvccpb.Event{
|
|
{
|
|
Type: mvccpb.DELETE,
|
|
Kv: &mvccpb.KeyValue{Key: []byte("foo"), ModRevision: 3},
|
|
},
|
|
}
|
|
if !reflect.DeepEqual(resp.Events, wevs) {
|
|
t.Fatalf("got %v, expected %v", resp.Events, wevs)
|
|
}
|
|
case <-time.After(100 * time.Millisecond):
|
|
t.Fatal("failed to receive delete event")
|
|
}
|
|
}
|
|
|
|
func TestV3WatchWithPrevKV(t *testing.T) {
|
|
BeforeTest(t)
|
|
clus := NewClusterV3(t, &ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
wctx, wcancel := context.WithCancel(context.Background())
|
|
defer wcancel()
|
|
|
|
tests := []struct {
|
|
key string
|
|
end string
|
|
vals []string
|
|
}{{
|
|
key: "foo",
|
|
end: "fop",
|
|
vals: []string{"bar1", "bar2"},
|
|
}, {
|
|
key: "/abc",
|
|
end: "/abd",
|
|
vals: []string{"first", "second"},
|
|
}}
|
|
for i, tt := range tests {
|
|
kvc := toGRPC(clus.RandClient()).KV
|
|
if _, err := kvc.Put(context.TODO(), &pb.PutRequest{Key: []byte(tt.key), Value: []byte(tt.vals[0])}); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
ws, werr := toGRPC(clus.RandClient()).Watch.Watch(wctx)
|
|
if werr != nil {
|
|
t.Fatal(werr)
|
|
}
|
|
|
|
req := &pb.WatchRequest{RequestUnion: &pb.WatchRequest_CreateRequest{
|
|
CreateRequest: &pb.WatchCreateRequest{
|
|
Key: []byte(tt.key),
|
|
RangeEnd: []byte(tt.end),
|
|
PrevKv: true,
|
|
}}}
|
|
if err := ws.Send(req); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if _, err := ws.Recv(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if _, err := kvc.Put(context.TODO(), &pb.PutRequest{Key: []byte(tt.key), Value: []byte(tt.vals[1])}); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
recv := make(chan *pb.WatchResponse, 1)
|
|
go func() {
|
|
// check received PUT
|
|
resp, rerr := ws.Recv()
|
|
if rerr != nil {
|
|
t.Error(rerr)
|
|
}
|
|
recv <- resp
|
|
}()
|
|
|
|
select {
|
|
case resp := <-recv:
|
|
if tt.vals[1] != string(resp.Events[0].Kv.Value) {
|
|
t.Errorf("#%d: unequal value: want=%s, get=%s", i, tt.vals[1], resp.Events[0].Kv.Value)
|
|
}
|
|
if tt.vals[0] != string(resp.Events[0].PrevKv.Value) {
|
|
t.Errorf("#%d: unequal value: want=%s, get=%s", i, tt.vals[0], resp.Events[0].PrevKv.Value)
|
|
}
|
|
case <-time.After(30 * time.Second):
|
|
t.Error("timeout waiting for watch response")
|
|
}
|
|
}
|
|
}
|
|
|
|
// TestV3WatchCancellation ensures that watch cancellation frees up server resources.
|
|
func TestV3WatchCancellation(t *testing.T) {
|
|
BeforeTest(t)
|
|
|
|
clus := NewClusterV3(t, &ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
|
defer cancel()
|
|
|
|
cli := clus.RandClient()
|
|
|
|
// increment watcher total count and keep a stream open
|
|
cli.Watch(ctx, "/foo")
|
|
|
|
for i := 0; i < 1000; i++ {
|
|
ctx, cancel := context.WithCancel(ctx)
|
|
cli.Watch(ctx, "/foo")
|
|
cancel()
|
|
}
|
|
|
|
// Wait a little for cancellations to take hold
|
|
time.Sleep(3 * time.Second)
|
|
|
|
minWatches, err := clus.Members[0].Metric("etcd_debugging_mvcc_watcher_total")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
var expected string
|
|
if ThroughProxy {
|
|
// grpc proxy has additional 2 watches open
|
|
expected = "3"
|
|
} else {
|
|
expected = "1"
|
|
}
|
|
|
|
if minWatches != expected {
|
|
t.Fatalf("expected %s watch, got %s", expected, minWatches)
|
|
}
|
|
}
|
|
|
|
// TestV3WatchCloseCancelRace ensures that watch close doesn't decrement the watcher total too far.
|
|
func TestV3WatchCloseCancelRace(t *testing.T) {
|
|
BeforeTest(t)
|
|
|
|
clus := NewClusterV3(t, &ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
|
defer cancel()
|
|
|
|
cli := clus.RandClient()
|
|
|
|
for i := 0; i < 1000; i++ {
|
|
ctx, cancel := context.WithCancel(ctx)
|
|
cli.Watch(ctx, "/foo")
|
|
cancel()
|
|
}
|
|
|
|
// Wait a little for cancellations to take hold
|
|
time.Sleep(3 * time.Second)
|
|
|
|
minWatches, err := clus.Members[0].Metric("etcd_debugging_mvcc_watcher_total")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
var expected string
|
|
if ThroughProxy {
|
|
// grpc proxy has additional 2 watches open
|
|
expected = "2"
|
|
} else {
|
|
expected = "0"
|
|
}
|
|
|
|
if minWatches != expected {
|
|
t.Fatalf("expected %s watch, got %s", expected, minWatches)
|
|
}
|
|
}
|