mirror of
https://github.com/etcd-io/etcd.git
synced 2024-09-27 06:25:44 +00:00
etcdserver: Guarantee order of requested progress notifications
Progress notifications requested using ProgressRequest were sent directly using the ctrlStream, which means that they could race against watch responses in the watchStream. This would especially happen when the stream was not synced - e.g. if you requested a progress notification on a freshly created unsynced watcher, the notification would typically arrive indicating a revision for which not all watch responses had been sent. This changes the behaviour so that v3rpc always goes through the watch stream, using a new RequestProgressAll function that closely matches the behaviour of the v3rpc code - i.e. 1. Generate a message with WatchId -1, indicating the revision for *all* watchers in the stream 2. Guarantee that a response is (eventually) sent The latter might require us to defer the response until all watchers are synced, which is likely as it should be. Note that we do *not* guarantee that the number of progress notifications matches the number of requests, only that eventually at least one gets sent. Signed-off-by: Benjamin Wang <wachao@vmware.com>
This commit is contained in:
@@ -1404,3 +1404,71 @@ func TestV3WatchCloseCancelRace(t *testing.T) {
|
||||
t.Fatalf("expected %s watch, got %s", expected, minWatches)
|
||||
}
|
||||
}
|
||||
|
||||
// TestV3WatchProgressWaitsForSync checks that progress notifications
|
||||
// don't get sent until the watcher is synchronised
|
||||
func TestV3WatchProgressWaitsForSync(t *testing.T) {
|
||||
|
||||
// Disable for gRPC proxy, as it does not support requesting
|
||||
// progress notifications
|
||||
if ThroughProxy {
|
||||
t.Skip("grpc proxy currently does not support requesting progress notifications")
|
||||
}
|
||||
|
||||
BeforeTest(t)
|
||||
|
||||
clus := NewClusterV3(t, &ClusterConfig{Size: 1})
|
||||
defer clus.Terminate(t)
|
||||
|
||||
client := clus.RandClient()
|
||||
ctx, cancel := context.WithTimeout(context.Background(), 10*time.Second)
|
||||
defer cancel()
|
||||
|
||||
// Write a couple values into key to make sure there's a
|
||||
// non-trivial amount of history.
|
||||
count := 1001
|
||||
t.Logf("Writing key 'foo' %d times", count)
|
||||
for i := 0; i < count; i++ {
|
||||
_, err := client.Put(ctx, "foo", fmt.Sprintf("bar%d", i))
|
||||
require.NoError(t, err)
|
||||
}
|
||||
|
||||
// Create watch channel starting at revision 1 (i.e. it starts
|
||||
// unsynced because of the update above)
|
||||
wch := client.Watch(ctx, "foo", clientv3.WithRev(1))
|
||||
|
||||
// Immediately request a progress notification. As the client
|
||||
// is unsynchronised, the server will have to defer the
|
||||
// notification internally.
|
||||
err := client.RequestProgress(ctx)
|
||||
require.NoError(t, err)
|
||||
|
||||
// Verify that we get the watch responses first. Note that
|
||||
// events might be spread across multiple packets.
|
||||
var event_count = 0
|
||||
for event_count < count {
|
||||
wr := <-wch
|
||||
if wr.Err() != nil {
|
||||
t.Fatal(fmt.Errorf("watch error: %w", wr.Err()))
|
||||
}
|
||||
if wr.IsProgressNotify() {
|
||||
t.Fatal("Progress notification from unsynced client!")
|
||||
}
|
||||
if wr.Header.Revision != int64(count+1) {
|
||||
t.Fatal("Incomplete watch response!")
|
||||
}
|
||||
event_count += len(wr.Events)
|
||||
}
|
||||
|
||||
// ... followed by the requested progress notification
|
||||
wr2 := <-wch
|
||||
if wr2.Err() != nil {
|
||||
t.Fatal(fmt.Errorf("watch error: %w", wr2.Err()))
|
||||
}
|
||||
if !wr2.IsProgressNotify() {
|
||||
t.Fatal("Did not receive progress notification!")
|
||||
}
|
||||
if wr2.Header.Revision != int64(count+1) {
|
||||
t.Fatal("Wrong revision in progress notification!")
|
||||
}
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user