mirror of
https://github.com/etcd-io/etcd.git
synced 2024-09-27 06:25:44 +00:00

Now Restart takes longer, and leasing Get blocks on server restarts anyway, so fix CI failures in slow machines. Signed-off-by: Gyuho Lee <gyuhox@gmail.com>
2024 lines
52 KiB
Go
2024 lines
52 KiB
Go
// Copyright 2017 The etcd Authors
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package integration
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"math/rand"
|
|
"reflect"
|
|
"sync"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/coreos/etcd/clientv3"
|
|
"github.com/coreos/etcd/clientv3/concurrency"
|
|
"github.com/coreos/etcd/clientv3/leasing"
|
|
"github.com/coreos/etcd/integration"
|
|
"github.com/coreos/etcd/pkg/testutil"
|
|
)
|
|
|
|
func TestLeasingPutGet(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3})
|
|
defer clus.Terminate(t)
|
|
|
|
lKV1, closeLKV1, err := leasing.NewKV(clus.Client(0), "foo/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV1()
|
|
|
|
lKV2, closeLKV2, err := leasing.NewKV(clus.Client(1), "foo/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV2()
|
|
|
|
lKV3, closeLKV3, err := leasing.NewKV(clus.Client(2), "foo/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV3()
|
|
|
|
resp, err := lKV1.Get(context.TODO(), "abc")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if len(resp.Kvs) != 0 {
|
|
t.Errorf("expected nil, got %q", resp.Kvs[0].Key)
|
|
}
|
|
|
|
if _, err = lKV1.Put(context.TODO(), "abc", "def"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if resp, err = lKV2.Get(context.TODO(), "abc"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if string(resp.Kvs[0].Key) != "abc" {
|
|
t.Errorf("expected key=%q, got key=%q", "abc", resp.Kvs[0].Key)
|
|
}
|
|
if string(resp.Kvs[0].Value) != "def" {
|
|
t.Errorf("expected value=%q, got value=%q", "bar", resp.Kvs[0].Value)
|
|
}
|
|
|
|
if _, err = lKV3.Get(context.TODO(), "abc"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if _, err = lKV2.Put(context.TODO(), "abc", "ghi"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if resp, err = lKV3.Get(context.TODO(), "abc"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if string(resp.Kvs[0].Key) != "abc" {
|
|
t.Errorf("expected key=%q, got key=%q", "abc", resp.Kvs[0].Key)
|
|
}
|
|
|
|
if string(resp.Kvs[0].Value) != "ghi" {
|
|
t.Errorf("expected value=%q, got value=%q", "bar", resp.Kvs[0].Value)
|
|
}
|
|
}
|
|
|
|
// TestLeasingInterval checks the leasing KV fetches key intervals.
|
|
func TestLeasingInterval(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
keys := []string{"abc/a", "abc/b", "abc/a/a"}
|
|
for _, k := range keys {
|
|
if _, err = clus.Client(0).Put(context.TODO(), k, "v"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
resp, err := lkv.Get(context.TODO(), "abc/", clientv3.WithPrefix())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if len(resp.Kvs) != 3 {
|
|
t.Fatalf("expected keys %+v, got response keys %+v", keys, resp.Kvs)
|
|
}
|
|
|
|
// load into cache
|
|
if resp, err = lkv.Get(context.TODO(), "abc/a"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// get when prefix is also a cached key
|
|
if resp, err = lkv.Get(context.TODO(), "abc/a", clientv3.WithPrefix()); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if len(resp.Kvs) != 2 {
|
|
t.Fatalf("expected keys %+v, got response keys %+v", keys, resp.Kvs)
|
|
}
|
|
}
|
|
|
|
// TestLeasingPutInvalidateNew checks the leasing KV updates its cache on a Put to a new key.
|
|
func TestLeasingPutInvalidateNew(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
if _, err = lkv.Get(context.TODO(), "k"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if _, err = lkv.Put(context.TODO(), "k", "v"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
lkvResp, err := lkv.Get(context.TODO(), "k")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
cResp, cerr := clus.Client(0).Get(context.TODO(), "k")
|
|
if cerr != nil {
|
|
t.Fatal(cerr)
|
|
}
|
|
if !reflect.DeepEqual(lkvResp, cResp) {
|
|
t.Fatalf(`expected %+v, got response %+v`, cResp, lkvResp)
|
|
}
|
|
}
|
|
|
|
// TestLeasingPutInvalidateExisting checks the leasing KV updates its cache on a Put to an existing key.
|
|
func TestLeasingPutInvalidateExisting(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
if _, err := clus.Client(0).Put(context.TODO(), "k", "abc"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
if _, err = lkv.Get(context.TODO(), "k"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if _, err = lkv.Put(context.TODO(), "k", "v"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
lkvResp, err := lkv.Get(context.TODO(), "k")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
cResp, cerr := clus.Client(0).Get(context.TODO(), "k")
|
|
if cerr != nil {
|
|
t.Fatal(cerr)
|
|
}
|
|
if !reflect.DeepEqual(lkvResp, cResp) {
|
|
t.Fatalf(`expected %+v, got response %+v`, cResp, lkvResp)
|
|
}
|
|
}
|
|
|
|
// TestLeasingGetNoLeaseTTL checks a key with a TTL is not leased.
|
|
func TestLeasingGetNoLeaseTTL(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
lresp, err := clus.Client(0).Grant(context.TODO(), 60)
|
|
testutil.AssertNil(t, err)
|
|
|
|
_, err = clus.Client(0).Put(context.TODO(), "k", "v", clientv3.WithLease(lresp.ID))
|
|
testutil.AssertNil(t, err)
|
|
|
|
gresp, err := lkv.Get(context.TODO(), "k")
|
|
testutil.AssertNil(t, err)
|
|
testutil.AssertEqual(t, len(gresp.Kvs), 1)
|
|
|
|
clus.Members[0].Stop(t)
|
|
|
|
ctx, cancel := context.WithTimeout(context.TODO(), time.Second)
|
|
_, err = lkv.Get(ctx, "k")
|
|
cancel()
|
|
testutil.AssertEqual(t, err, ctx.Err())
|
|
}
|
|
|
|
// TestLeasingGetSerializable checks the leasing KV can make serialized requests
|
|
// when the etcd cluster is partitioned.
|
|
func TestLeasingGetSerializable(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 2})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
if _, err = clus.Client(0).Put(context.TODO(), "cached", "abc"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if _, err = lkv.Get(context.TODO(), "cached"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
clus.Members[1].Stop(t)
|
|
|
|
// don't necessarily try to acquire leasing key ownership for new key
|
|
resp, err := lkv.Get(context.TODO(), "uncached", clientv3.WithSerializable())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if len(resp.Kvs) != 0 {
|
|
t.Fatalf(`expected no keys, got response %+v`, resp)
|
|
}
|
|
|
|
clus.Members[0].Stop(t)
|
|
|
|
// leasing key ownership should have "cached" locally served
|
|
cachedResp, err := lkv.Get(context.TODO(), "cached", clientv3.WithSerializable())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if len(cachedResp.Kvs) != 1 || string(cachedResp.Kvs[0].Value) != "abc" {
|
|
t.Fatalf(`expected "cached"->"abc", got response %+v`, cachedResp)
|
|
}
|
|
}
|
|
|
|
// TestLeasingPrevKey checks the cache respects WithPrevKV on puts.
|
|
func TestLeasingPrevKey(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 2})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
if _, err = clus.Client(0).Put(context.TODO(), "k", "abc"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
// acquire leasing key
|
|
if _, err = lkv.Get(context.TODO(), "k"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
resp, err := lkv.Put(context.TODO(), "k", "def", clientv3.WithPrevKV())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if resp.PrevKv == nil || string(resp.PrevKv.Value) != "abc" {
|
|
t.Fatalf(`expected PrevKV.Value="abc", got response %+v`, resp)
|
|
}
|
|
}
|
|
|
|
// TestLeasingRevGet checks the cache respects Get by Revision.
|
|
func TestLeasingRevGet(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
putResp, err := clus.Client(0).Put(context.TODO(), "k", "abc")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if _, err = clus.Client(0).Put(context.TODO(), "k", "def"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// check historic revision
|
|
getResp, gerr := lkv.Get(context.TODO(), "k", clientv3.WithRev(putResp.Header.Revision))
|
|
if gerr != nil {
|
|
t.Fatal(gerr)
|
|
}
|
|
if len(getResp.Kvs) != 1 || string(getResp.Kvs[0].Value) != "abc" {
|
|
t.Fatalf(`expeted "k"->"abc" at rev=%d, got response %+v`, putResp.Header.Revision, getResp)
|
|
}
|
|
// check current revision
|
|
getResp, gerr = lkv.Get(context.TODO(), "k")
|
|
if gerr != nil {
|
|
t.Fatal(gerr)
|
|
}
|
|
if len(getResp.Kvs) != 1 || string(getResp.Kvs[0].Value) != "def" {
|
|
t.Fatalf(`expeted "k"->"abc" at rev=%d, got response %+v`, putResp.Header.Revision, getResp)
|
|
}
|
|
}
|
|
|
|
// TestLeasingGetWithOpts checks options that can be served through the cache do not depend on the server.
|
|
func TestLeasingGetWithOpts(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
if _, err = clus.Client(0).Put(context.TODO(), "k", "abc"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
// in cache
|
|
if _, err = lkv.Get(context.TODO(), "k", clientv3.WithKeysOnly()); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
clus.Members[0].Stop(t)
|
|
|
|
opts := []clientv3.OpOption{
|
|
clientv3.WithKeysOnly(),
|
|
clientv3.WithLimit(1),
|
|
clientv3.WithMinCreateRev(1),
|
|
clientv3.WithMinModRev(1),
|
|
clientv3.WithSort(clientv3.SortByKey, clientv3.SortAscend),
|
|
clientv3.WithSerializable(),
|
|
}
|
|
for _, opt := range opts {
|
|
if _, err := lkv.Get(context.TODO(), "k", opt); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
getOpts := []clientv3.OpOption{}
|
|
for i := 0; i < len(opts); i++ {
|
|
getOpts = append(getOpts, opts[rand.Intn(len(opts))])
|
|
}
|
|
getOpts = getOpts[:rand.Intn(len(opts))]
|
|
if _, err := lkv.Get(context.TODO(), "k", getOpts...); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
// TestLeasingConcurrentPut ensures that a get after concurrent puts returns
|
|
// the recently put data.
|
|
func TestLeasingConcurrentPut(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
// force key into leasing key cache
|
|
if _, err = lkv.Get(context.TODO(), "k"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// concurrently put through leasing client
|
|
numPuts := 16
|
|
putc := make(chan *clientv3.PutResponse, numPuts)
|
|
for i := 0; i < numPuts; i++ {
|
|
go func() {
|
|
resp, perr := lkv.Put(context.TODO(), "k", "abc")
|
|
if perr != nil {
|
|
t.Fatal(perr)
|
|
}
|
|
putc <- resp
|
|
}()
|
|
}
|
|
// record maximum revision from puts
|
|
maxRev := int64(0)
|
|
for i := 0; i < numPuts; i++ {
|
|
if resp := <-putc; resp.Header.Revision > maxRev {
|
|
maxRev = resp.Header.Revision
|
|
}
|
|
}
|
|
|
|
// confirm Get gives most recently put revisions
|
|
getResp, gerr := lkv.Get(context.TODO(), "k")
|
|
if gerr != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if mr := getResp.Kvs[0].ModRevision; mr != maxRev {
|
|
t.Errorf("expected ModRevision %d, got %d", maxRev, mr)
|
|
}
|
|
if ver := getResp.Kvs[0].Version; ver != int64(numPuts) {
|
|
t.Errorf("expected Version %d, got %d", numPuts, ver)
|
|
}
|
|
}
|
|
|
|
func TestLeasingDisconnectedGet(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
if _, err = clus.Client(0).Put(context.TODO(), "cached", "abc"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
// get key so it's cached
|
|
if _, err = lkv.Get(context.TODO(), "cached"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
clus.Members[0].Stop(t)
|
|
|
|
// leasing key ownership should have "cached" locally served
|
|
cachedResp, err := lkv.Get(context.TODO(), "cached")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if len(cachedResp.Kvs) != 1 || string(cachedResp.Kvs[0].Value) != "abc" {
|
|
t.Fatalf(`expected "cached"->"abc", got response %+v`, cachedResp)
|
|
}
|
|
}
|
|
|
|
func TestLeasingDeleteOwner(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
if _, err = clus.Client(0).Put(context.TODO(), "k", "abc"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// get+own / delete / get
|
|
if _, err = lkv.Get(context.TODO(), "k"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if _, err = lkv.Delete(context.TODO(), "k"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
resp, err := lkv.Get(context.TODO(), "k")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if len(resp.Kvs) != 0 {
|
|
t.Fatalf(`expected "k" to be deleted, got response %+v`, resp)
|
|
}
|
|
// try to double delete
|
|
if _, err = lkv.Delete(context.TODO(), "k"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
func TestLeasingDeleteNonOwner(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv1, closeLKV1, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV1()
|
|
|
|
lkv2, closeLKV2, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV2()
|
|
|
|
if _, err = clus.Client(0).Put(context.TODO(), "k", "abc"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
// acquire ownership
|
|
if _, err = lkv1.Get(context.TODO(), "k"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
// delete via non-owner
|
|
if _, err = lkv2.Delete(context.TODO(), "k"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// key should be removed from lkv1
|
|
resp, err := lkv1.Get(context.TODO(), "k")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if len(resp.Kvs) != 0 {
|
|
t.Fatalf(`expected "k" to be deleted, got response %+v`, resp)
|
|
}
|
|
}
|
|
|
|
func TestLeasingOverwriteResponse(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
if _, err = clus.Client(0).Put(context.TODO(), "k", "abc"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
resp, err := lkv.Get(context.TODO(), "k")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
resp.Kvs[0].Key[0] = 'z'
|
|
resp.Kvs[0].Value[0] = 'z'
|
|
|
|
resp, err = lkv.Get(context.TODO(), "k")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if string(resp.Kvs[0].Key) != "k" {
|
|
t.Errorf(`expected key "k", got %q`, string(resp.Kvs[0].Key))
|
|
}
|
|
if string(resp.Kvs[0].Value) != "abc" {
|
|
t.Errorf(`expected value "abc", got %q`, string(resp.Kvs[0].Value))
|
|
}
|
|
}
|
|
|
|
func TestLeasingOwnerPutResponse(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
if _, err = clus.Client(0).Put(context.TODO(), "k", "abc"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
gresp, gerr := lkv.Get(context.TODO(), "k")
|
|
if gerr != nil {
|
|
t.Fatal(gerr)
|
|
}
|
|
presp, err := lkv.Put(context.TODO(), "k", "def")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if presp == nil {
|
|
t.Fatal("expected put response, got nil")
|
|
}
|
|
|
|
clus.Members[0].Stop(t)
|
|
|
|
gresp, gerr = lkv.Get(context.TODO(), "k")
|
|
if gerr != nil {
|
|
t.Fatal(gerr)
|
|
}
|
|
if gresp.Kvs[0].ModRevision != presp.Header.Revision {
|
|
t.Errorf("expected mod revision %d, got %d", presp.Header.Revision, gresp.Kvs[0].ModRevision)
|
|
}
|
|
if gresp.Kvs[0].Version != 2 {
|
|
t.Errorf("expected version 2, got version %d", gresp.Kvs[0].Version)
|
|
}
|
|
}
|
|
|
|
func TestLeasingTxnOwnerGetRange(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
keyCount := rand.Intn(10) + 1
|
|
for i := 0; i < keyCount; i++ {
|
|
k := fmt.Sprintf("k-%d", i)
|
|
if _, err := clus.Client(0).Put(context.TODO(), k, k+k); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
if _, err := lkv.Get(context.TODO(), "k-"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
tresp, terr := lkv.Txn(context.TODO()).Then(clientv3.OpGet("k-", clientv3.WithPrefix())).Commit()
|
|
if terr != nil {
|
|
t.Fatal(terr)
|
|
}
|
|
if resp := tresp.Responses[0].GetResponseRange(); len(resp.Kvs) != keyCount {
|
|
t.Fatalf("expected %d keys, got response %+v", keyCount, resp.Kvs)
|
|
}
|
|
}
|
|
|
|
func TestLeasingTxnOwnerGet(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
keyCount := rand.Intn(10) + 1
|
|
var ops []clientv3.Op
|
|
presps := make([]*clientv3.PutResponse, keyCount)
|
|
for i := range presps {
|
|
k := fmt.Sprintf("k-%d", i)
|
|
presp, err := clus.Client(0).Put(context.TODO(), k, k+k)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
presps[i] = presp
|
|
|
|
if _, err = lkv.Get(context.TODO(), k); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
ops = append(ops, clientv3.OpGet(k))
|
|
}
|
|
ops = ops[:rand.Intn(len(ops))]
|
|
|
|
// served through cache
|
|
clus.Members[0].Stop(t)
|
|
|
|
var thenOps, elseOps []clientv3.Op
|
|
cmps, useThen := randCmps("k-", presps)
|
|
|
|
if useThen {
|
|
|
|
thenOps = ops
|
|
elseOps = []clientv3.Op{clientv3.OpPut("k", "1")}
|
|
} else {
|
|
thenOps = []clientv3.Op{clientv3.OpPut("k", "1")}
|
|
elseOps = ops
|
|
}
|
|
|
|
tresp, terr := lkv.Txn(context.TODO()).
|
|
If(cmps...).
|
|
Then(thenOps...).
|
|
Else(elseOps...).Commit()
|
|
|
|
if terr != nil {
|
|
t.Fatal(terr)
|
|
}
|
|
if tresp.Succeeded != useThen {
|
|
t.Fatalf("expected succeeded=%v, got tresp=%+v", useThen, tresp)
|
|
}
|
|
if len(tresp.Responses) != len(ops) {
|
|
t.Fatalf("expected %d responses, got %d", len(ops), len(tresp.Responses))
|
|
}
|
|
wrev := presps[len(presps)-1].Header.Revision
|
|
if tresp.Header.Revision < wrev {
|
|
t.Fatalf("expected header revision >= %d, got %d", wrev, tresp.Header.Revision)
|
|
}
|
|
for i := range ops {
|
|
k := fmt.Sprintf("k-%d", i)
|
|
rr := tresp.Responses[i].GetResponseRange()
|
|
if rr == nil {
|
|
t.Errorf("expected get response, got %+v", tresp.Responses[i])
|
|
}
|
|
if string(rr.Kvs[0].Key) != k || string(rr.Kvs[0].Value) != k+k {
|
|
t.Errorf(`expected key for %q, got %+v`, k, rr.Kvs)
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestLeasingTxnOwnerDeleteRange(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
keyCount := rand.Intn(10) + 1
|
|
for i := 0; i < keyCount; i++ {
|
|
k := fmt.Sprintf("k-%d", i)
|
|
if _, perr := clus.Client(0).Put(context.TODO(), k, k+k); perr != nil {
|
|
t.Fatal(perr)
|
|
}
|
|
}
|
|
|
|
// cache in lkv
|
|
resp, err := lkv.Get(context.TODO(), "k-", clientv3.WithPrefix())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if len(resp.Kvs) != keyCount {
|
|
t.Fatalf("expected %d keys, got %d", keyCount, len(resp.Kvs))
|
|
}
|
|
|
|
if _, terr := lkv.Txn(context.TODO()).Then(clientv3.OpDelete("k-", clientv3.WithPrefix())).Commit(); terr != nil {
|
|
t.Fatal(terr)
|
|
}
|
|
|
|
resp, err = lkv.Get(context.TODO(), "k-", clientv3.WithPrefix())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if len(resp.Kvs) != 0 {
|
|
t.Fatalf("expected no keys, got %d", len(resp.Kvs))
|
|
}
|
|
}
|
|
|
|
func TestLeasingTxnOwnerDelete(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
if _, err = clus.Client(0).Put(context.TODO(), "k", "abc"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// cache in lkv
|
|
if _, gerr := lkv.Get(context.TODO(), "k"); gerr != nil {
|
|
t.Fatal(gerr)
|
|
}
|
|
|
|
if _, terr := lkv.Txn(context.TODO()).Then(clientv3.OpDelete("k")).Commit(); terr != nil {
|
|
t.Fatal(terr)
|
|
}
|
|
|
|
resp, err := lkv.Get(context.TODO(), "k")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if len(resp.Kvs) != 0 {
|
|
t.Fatalf("expected no keys, got %d", len(resp.Kvs))
|
|
}
|
|
}
|
|
|
|
func TestLeasingTxnOwnerIf(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
if _, err = clus.Client(0).Put(context.TODO(), "k", "abc"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if _, err = lkv.Get(context.TODO(), "k"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// served through cache
|
|
clus.Members[0].Stop(t)
|
|
|
|
tests := []struct {
|
|
cmps []clientv3.Cmp
|
|
wSucceeded bool
|
|
wResponses int
|
|
}{
|
|
// success
|
|
{
|
|
cmps: []clientv3.Cmp{clientv3.Compare(clientv3.Value("k"), "=", "abc")},
|
|
wSucceeded: true,
|
|
wResponses: 1,
|
|
},
|
|
{
|
|
cmps: []clientv3.Cmp{clientv3.Compare(clientv3.CreateRevision("k"), "=", 2)},
|
|
wSucceeded: true,
|
|
wResponses: 1,
|
|
},
|
|
{
|
|
cmps: []clientv3.Cmp{clientv3.Compare(clientv3.ModRevision("k"), "=", 2)},
|
|
wSucceeded: true,
|
|
wResponses: 1,
|
|
},
|
|
{
|
|
cmps: []clientv3.Cmp{clientv3.Compare(clientv3.Version("k"), "=", 1)},
|
|
wSucceeded: true,
|
|
wResponses: 1,
|
|
},
|
|
// failure
|
|
{
|
|
cmps: []clientv3.Cmp{clientv3.Compare(clientv3.Value("k"), ">", "abc")},
|
|
},
|
|
{
|
|
cmps: []clientv3.Cmp{clientv3.Compare(clientv3.CreateRevision("k"), ">", 2)},
|
|
},
|
|
{
|
|
cmps: []clientv3.Cmp{clientv3.Compare(clientv3.ModRevision("k"), "=", 2)},
|
|
wSucceeded: true,
|
|
wResponses: 1,
|
|
},
|
|
{
|
|
cmps: []clientv3.Cmp{clientv3.Compare(clientv3.Version("k"), ">", 1)},
|
|
},
|
|
{
|
|
cmps: []clientv3.Cmp{clientv3.Compare(clientv3.Value("k"), "<", "abc")},
|
|
},
|
|
{
|
|
cmps: []clientv3.Cmp{clientv3.Compare(clientv3.CreateRevision("k"), "<", 2)},
|
|
},
|
|
{
|
|
cmps: []clientv3.Cmp{clientv3.Compare(clientv3.ModRevision("k"), "<", 2)},
|
|
},
|
|
{
|
|
cmps: []clientv3.Cmp{clientv3.Compare(clientv3.Version("k"), "<", 1)},
|
|
},
|
|
{
|
|
cmps: []clientv3.Cmp{
|
|
clientv3.Compare(clientv3.Version("k"), "=", 1),
|
|
clientv3.Compare(clientv3.Version("k"), "<", 1),
|
|
},
|
|
},
|
|
}
|
|
|
|
for i, tt := range tests {
|
|
tresp, terr := lkv.Txn(context.TODO()).If(tt.cmps...).Then(clientv3.OpGet("k")).Commit()
|
|
if terr != nil {
|
|
t.Fatal(terr)
|
|
}
|
|
if tresp.Succeeded != tt.wSucceeded {
|
|
t.Errorf("#%d: expected succeeded %v, got %v", i, tt.wSucceeded, tresp.Succeeded)
|
|
}
|
|
if len(tresp.Responses) != tt.wResponses {
|
|
t.Errorf("#%d: expected %d responses, got %d", i, tt.wResponses, len(tresp.Responses))
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestLeasingTxnCancel(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv1, closeLKV1, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV1()
|
|
|
|
lkv2, closeLKV2, err := leasing.NewKV(clus.Client(1), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV2()
|
|
|
|
// acquire lease but disconnect so no revoke in time
|
|
if _, err = lkv1.Get(context.TODO(), "k"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
clus.Members[0].Stop(t)
|
|
|
|
// wait for leader election, if any
|
|
if _, err = clus.Client(1).Get(context.TODO(), "abc"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
ctx, cancel := context.WithCancel(context.TODO())
|
|
go func() {
|
|
time.Sleep(100 * time.Millisecond)
|
|
cancel()
|
|
}()
|
|
if _, err := lkv2.Txn(ctx).Then(clientv3.OpPut("k", "v")).Commit(); err != context.Canceled {
|
|
t.Fatalf("expected %v, got %v", context.Canceled, err)
|
|
}
|
|
}
|
|
|
|
func TestLeasingTxnNonOwnerPut(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
lkv2, closeLKV2, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV2()
|
|
|
|
if _, err = clus.Client(0).Put(context.TODO(), "k", "abc"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if _, err = clus.Client(0).Put(context.TODO(), "k2", "123"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
// cache in lkv
|
|
if _, err = lkv.Get(context.TODO(), "k"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if _, err = lkv.Get(context.TODO(), "k2"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
// invalidate via lkv2 txn
|
|
opArray := make([]clientv3.Op, 0)
|
|
opArray = append(opArray, clientv3.OpPut("k2", "456"))
|
|
tresp, terr := lkv2.Txn(context.TODO()).Then(
|
|
clientv3.OpTxn(nil, opArray, nil),
|
|
clientv3.OpPut("k", "def"),
|
|
clientv3.OpPut("k3", "999"), // + a key not in any cache
|
|
).Commit()
|
|
if terr != nil {
|
|
t.Fatal(terr)
|
|
}
|
|
if !tresp.Succeeded || len(tresp.Responses) != 3 {
|
|
t.Fatalf("expected txn success, got %+v", tresp)
|
|
}
|
|
// check cache was invalidated
|
|
gresp, gerr := lkv.Get(context.TODO(), "k")
|
|
if gerr != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if len(gresp.Kvs) != 1 || string(gresp.Kvs[0].Value) != "def" {
|
|
t.Errorf(`expected value "def", got %+v`, gresp)
|
|
}
|
|
gresp, gerr = lkv.Get(context.TODO(), "k2")
|
|
if gerr != nil {
|
|
t.Fatal(gerr)
|
|
}
|
|
if len(gresp.Kvs) != 1 || string(gresp.Kvs[0].Value) != "456" {
|
|
t.Errorf(`expected value "def", got %+v`, gresp)
|
|
}
|
|
// check puts were applied and are all in the same revision
|
|
w := clus.Client(0).Watch(
|
|
clus.Client(0).Ctx(),
|
|
"k",
|
|
clientv3.WithRev(tresp.Header.Revision),
|
|
clientv3.WithPrefix())
|
|
wresp := <-w
|
|
c := 0
|
|
evs := []clientv3.Event{}
|
|
for _, ev := range wresp.Events {
|
|
evs = append(evs, *ev)
|
|
if ev.Kv.ModRevision == tresp.Header.Revision {
|
|
c++
|
|
}
|
|
}
|
|
if c != 3 {
|
|
t.Fatalf("expected 3 put events, got %+v", evs)
|
|
}
|
|
}
|
|
|
|
// TestLeasingTxnRandIfThenOrElse randomly leases keys two separate clients, then
|
|
// issues a random If/{Then,Else} transaction on those keys to one client.
|
|
func TestLeasingTxnRandIfThenOrElse(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv1, closeLKV1, err1 := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err1)
|
|
defer closeLKV1()
|
|
|
|
lkv2, closeLKV2, err2 := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err2)
|
|
defer closeLKV2()
|
|
|
|
keyCount := 16
|
|
dat := make([]*clientv3.PutResponse, keyCount)
|
|
for i := 0; i < keyCount; i++ {
|
|
k, v := fmt.Sprintf("k-%d", i), fmt.Sprintf("%d", i)
|
|
dat[i], err1 = clus.Client(0).Put(context.TODO(), k, v)
|
|
if err1 != nil {
|
|
t.Fatal(err1)
|
|
}
|
|
}
|
|
|
|
// nondeterministically populate leasing caches
|
|
var wg sync.WaitGroup
|
|
getc := make(chan struct{}, keyCount)
|
|
getRandom := func(kv clientv3.KV) {
|
|
defer wg.Done()
|
|
for i := 0; i < keyCount/2; i++ {
|
|
k := fmt.Sprintf("k-%d", rand.Intn(keyCount))
|
|
if _, err := kv.Get(context.TODO(), k); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
getc <- struct{}{}
|
|
}
|
|
}
|
|
wg.Add(2)
|
|
defer wg.Wait()
|
|
go getRandom(lkv1)
|
|
go getRandom(lkv2)
|
|
|
|
// random list of comparisons, all true
|
|
cmps, useThen := randCmps("k-", dat)
|
|
// random list of puts/gets; unique keys
|
|
ops := []clientv3.Op{}
|
|
usedIdx := make(map[int]struct{})
|
|
for i := 0; i < keyCount; i++ {
|
|
idx := rand.Intn(keyCount)
|
|
if _, ok := usedIdx[idx]; ok {
|
|
continue
|
|
}
|
|
usedIdx[idx] = struct{}{}
|
|
k := fmt.Sprintf("k-%d", idx)
|
|
switch rand.Intn(2) {
|
|
case 0:
|
|
ops = append(ops, clientv3.OpGet(k))
|
|
case 1:
|
|
ops = append(ops, clientv3.OpPut(k, "a"))
|
|
// TODO: add delete
|
|
}
|
|
}
|
|
// random lengths
|
|
ops = ops[:rand.Intn(len(ops))]
|
|
|
|
// wait for some gets to populate the leasing caches before committing
|
|
for i := 0; i < keyCount/2; i++ {
|
|
<-getc
|
|
}
|
|
|
|
// randomly choose between then and else blocks
|
|
var thenOps, elseOps []clientv3.Op
|
|
if useThen {
|
|
thenOps = ops
|
|
} else {
|
|
// force failure
|
|
elseOps = ops
|
|
}
|
|
|
|
tresp, terr := lkv1.Txn(context.TODO()).If(cmps...).Then(thenOps...).Else(elseOps...).Commit()
|
|
if terr != nil {
|
|
t.Fatal(terr)
|
|
}
|
|
// cmps always succeed
|
|
if tresp.Succeeded != useThen {
|
|
t.Fatalf("expected succeeded=%v, got tresp=%+v", useThen, tresp)
|
|
}
|
|
// get should match what was put
|
|
checkPuts := func(s string, kv clientv3.KV) {
|
|
for _, op := range ops {
|
|
if !op.IsPut() {
|
|
continue
|
|
}
|
|
resp, rerr := kv.Get(context.TODO(), string(op.KeyBytes()))
|
|
if rerr != nil {
|
|
t.Fatal(rerr)
|
|
}
|
|
if len(resp.Kvs) != 1 || string(resp.Kvs[0].Value) != "a" {
|
|
t.Fatalf(`%s: expected value="a", got %+v`, s, resp.Kvs)
|
|
}
|
|
}
|
|
}
|
|
checkPuts("client(0)", clus.Client(0))
|
|
checkPuts("lkv1", lkv1)
|
|
checkPuts("lkv2", lkv2)
|
|
}
|
|
|
|
func TestLeasingOwnerPutError(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
if _, err = lkv.Get(context.TODO(), "k"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
clus.Members[0].Stop(t)
|
|
ctx, cancel := context.WithTimeout(context.TODO(), 100*time.Millisecond)
|
|
defer cancel()
|
|
if resp, err := lkv.Put(ctx, "k", "v"); err == nil {
|
|
t.Fatalf("expected error, got response %+v", resp)
|
|
}
|
|
}
|
|
|
|
func TestLeasingOwnerDeleteError(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
if _, err = lkv.Get(context.TODO(), "k"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
clus.Members[0].Stop(t)
|
|
ctx, cancel := context.WithTimeout(context.TODO(), 100*time.Millisecond)
|
|
defer cancel()
|
|
if resp, err := lkv.Delete(ctx, "k"); err == nil {
|
|
t.Fatalf("expected error, got response %+v", resp)
|
|
}
|
|
}
|
|
|
|
func TestLeasingNonOwnerPutError(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
clus.Members[0].Stop(t)
|
|
ctx, cancel := context.WithTimeout(context.TODO(), 100*time.Millisecond)
|
|
defer cancel()
|
|
if resp, err := lkv.Put(ctx, "k", "v"); err == nil {
|
|
t.Fatalf("expected error, got response %+v", resp)
|
|
}
|
|
}
|
|
|
|
func TestLeasingOwnerDeletePrefix(t *testing.T) {
|
|
testLeasingOwnerDelete(t, clientv3.OpDelete("key/", clientv3.WithPrefix()))
|
|
}
|
|
|
|
func TestLeasingOwnerDeleteFrom(t *testing.T) {
|
|
testLeasingOwnerDelete(t, clientv3.OpDelete("kd", clientv3.WithFromKey()))
|
|
}
|
|
|
|
func testLeasingOwnerDelete(t *testing.T, del clientv3.Op) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "0/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
for i := 0; i < 8; i++ {
|
|
if _, err = clus.Client(0).Put(context.TODO(), fmt.Sprintf("key/%d", i), "123"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
if _, err = lkv.Get(context.TODO(), "key/1"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
opResp, delErr := lkv.Do(context.TODO(), del)
|
|
if delErr != nil {
|
|
t.Fatal(delErr)
|
|
}
|
|
delResp := opResp.Del()
|
|
|
|
// confirm keys are invalidated from cache and deleted on etcd
|
|
for i := 0; i < 8; i++ {
|
|
resp, err := lkv.Get(context.TODO(), fmt.Sprintf("key/%d", i))
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if len(resp.Kvs) != 0 {
|
|
t.Fatalf("expected no keys on key/%d, got %+v", i, resp)
|
|
}
|
|
}
|
|
|
|
// confirm keys were deleted atomically
|
|
|
|
w := clus.Client(0).Watch(
|
|
clus.Client(0).Ctx(),
|
|
"key/",
|
|
clientv3.WithRev(delResp.Header.Revision),
|
|
clientv3.WithPrefix())
|
|
|
|
if wresp := <-w; len(wresp.Events) != 8 {
|
|
t.Fatalf("expected %d delete events,got %d", 8, len(wresp.Events))
|
|
}
|
|
}
|
|
|
|
func TestLeasingDeleteRangeBounds(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
delkv, closeDelKV, err := leasing.NewKV(clus.Client(0), "0/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeDelKV()
|
|
|
|
getkv, closeGetKv, err := leasing.NewKV(clus.Client(0), "0/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeGetKv()
|
|
|
|
for _, k := range []string{"j", "m"} {
|
|
if _, err = clus.Client(0).Put(context.TODO(), k, "123"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if _, err = getkv.Get(context.TODO(), k); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
if _, err = delkv.Delete(context.TODO(), "k", clientv3.WithPrefix()); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// leases still on server?
|
|
for _, k := range []string{"j", "m"} {
|
|
resp, geterr := clus.Client(0).Get(context.TODO(), "0/"+k, clientv3.WithPrefix())
|
|
if geterr != nil {
|
|
t.Fatal(geterr)
|
|
}
|
|
if len(resp.Kvs) != 1 {
|
|
t.Fatalf("expected leasing key, got %+v", resp)
|
|
}
|
|
}
|
|
|
|
// j and m should still have leases registered since not under k*
|
|
clus.Members[0].Stop(t)
|
|
|
|
if _, err = getkv.Get(context.TODO(), "j"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if _, err = getkv.Get(context.TODO(), "m"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
func TestLeasingDeleteRangeContendTxn(t *testing.T) {
|
|
then := []clientv3.Op{clientv3.OpDelete("key/", clientv3.WithPrefix())}
|
|
testLeasingDeleteRangeContend(t, clientv3.OpTxn(nil, then, nil))
|
|
}
|
|
|
|
func TestLeaseDeleteRangeContendDel(t *testing.T) {
|
|
op := clientv3.OpDelete("key/", clientv3.WithPrefix())
|
|
testLeasingDeleteRangeContend(t, op)
|
|
}
|
|
|
|
func testLeasingDeleteRangeContend(t *testing.T, op clientv3.Op) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
delkv, closeDelKV, err := leasing.NewKV(clus.Client(0), "0/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeDelKV()
|
|
|
|
putkv, closePutKV, err := leasing.NewKV(clus.Client(0), "0/")
|
|
testutil.AssertNil(t, err)
|
|
defer closePutKV()
|
|
|
|
for i := 0; i < 8; i++ {
|
|
key := fmt.Sprintf("key/%d", i)
|
|
if _, err = clus.Client(0).Put(context.TODO(), key, "123"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if _, err = putkv.Get(context.TODO(), key); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
ctx, cancel := context.WithCancel(context.TODO())
|
|
donec := make(chan struct{})
|
|
go func() {
|
|
defer close(donec)
|
|
for i := 0; ctx.Err() == nil; i++ {
|
|
key := fmt.Sprintf("key/%d", i%8)
|
|
putkv.Put(ctx, key, "123")
|
|
putkv.Get(ctx, key)
|
|
}
|
|
}()
|
|
|
|
_, delErr := delkv.Do(context.TODO(), op)
|
|
cancel()
|
|
<-donec
|
|
if delErr != nil {
|
|
t.Fatal(delErr)
|
|
}
|
|
|
|
// confirm keys on non-deleter match etcd
|
|
for i := 0; i < 8; i++ {
|
|
key := fmt.Sprintf("key/%d", i)
|
|
resp, err := putkv.Get(context.TODO(), key)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
servResp, err := clus.Client(0).Get(context.TODO(), key)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if !reflect.DeepEqual(resp.Kvs, servResp.Kvs) {
|
|
t.Errorf("#%d: expected %+v, got %+v", i, servResp.Kvs, resp.Kvs)
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestLeasingPutGetDeleteConcurrent(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkvs := make([]clientv3.KV, 16)
|
|
for i := range lkvs {
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "pfx/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
lkvs[i] = lkv
|
|
}
|
|
|
|
getdel := func(kv clientv3.KV) {
|
|
if _, err := kv.Put(context.TODO(), "k", "abc"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
time.Sleep(time.Millisecond)
|
|
if _, err := kv.Get(context.TODO(), "k"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if _, err := kv.Delete(context.TODO(), "k"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
time.Sleep(2 * time.Millisecond)
|
|
}
|
|
|
|
var wg sync.WaitGroup
|
|
wg.Add(16)
|
|
for i := 0; i < 16; i++ {
|
|
go func() {
|
|
defer wg.Done()
|
|
for _, kv := range lkvs {
|
|
getdel(kv)
|
|
}
|
|
}()
|
|
}
|
|
wg.Wait()
|
|
|
|
resp, err := lkvs[0].Get(context.TODO(), "k")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if len(resp.Kvs) > 0 {
|
|
t.Fatalf("expected no kvs, got %+v", resp.Kvs)
|
|
}
|
|
|
|
resp, err = clus.Client(0).Get(context.TODO(), "k")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if len(resp.Kvs) > 0 {
|
|
t.Fatalf("expected no kvs, got %+v", resp.Kvs)
|
|
}
|
|
}
|
|
|
|
// TestLeasingReconnectOwnerRevoke checks that revocation works if
|
|
// disconnected when trying to submit revoke txn.
|
|
func TestLeasingReconnectOwnerRevoke(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv1, closeLKV1, err1 := leasing.NewKV(clus.Client(0), "foo/")
|
|
testutil.AssertNil(t, err1)
|
|
defer closeLKV1()
|
|
|
|
lkv2, closeLKV2, err2 := leasing.NewKV(clus.Client(1), "foo/")
|
|
testutil.AssertNil(t, err2)
|
|
defer closeLKV2()
|
|
|
|
if _, err := lkv1.Get(context.TODO(), "k"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// force leader away from member 0
|
|
clus.Members[0].Stop(t)
|
|
clus.WaitLeader(t)
|
|
clus.Members[0].Restart(t)
|
|
|
|
cctx, cancel := context.WithCancel(context.TODO())
|
|
sdonec, pdonec := make(chan struct{}), make(chan struct{})
|
|
// make lkv1 connection choppy so Txn fails
|
|
go func() {
|
|
defer close(sdonec)
|
|
for i := 0; i < 3 && cctx.Err() == nil; i++ {
|
|
clus.Members[0].Stop(t)
|
|
time.Sleep(10 * time.Millisecond)
|
|
clus.Members[0].Restart(t)
|
|
}
|
|
}()
|
|
go func() {
|
|
defer close(pdonec)
|
|
if _, err := lkv2.Put(cctx, "k", "v"); err != nil {
|
|
t.Log(err)
|
|
}
|
|
// blocks until lkv1 connection comes back
|
|
resp, err := lkv1.Get(cctx, "k")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if string(resp.Kvs[0].Value) != "v" {
|
|
t.Fatalf(`expected "v" value, got %+v`, resp)
|
|
}
|
|
}()
|
|
select {
|
|
case <-pdonec:
|
|
cancel()
|
|
<-sdonec
|
|
case <-time.After(15 * time.Second):
|
|
cancel()
|
|
<-sdonec
|
|
<-pdonec
|
|
t.Fatal("took too long to revoke and put")
|
|
}
|
|
}
|
|
|
|
// TestLeasingReconnectOwnerRevokeCompact checks that revocation works if
|
|
// disconnected and the watch is compacted.
|
|
func TestLeasingReconnectOwnerRevokeCompact(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv1, closeLKV1, err1 := leasing.NewKV(clus.Client(0), "foo/")
|
|
testutil.AssertNil(t, err1)
|
|
defer closeLKV1()
|
|
|
|
lkv2, closeLKV2, err2 := leasing.NewKV(clus.Client(1), "foo/")
|
|
testutil.AssertNil(t, err2)
|
|
defer closeLKV2()
|
|
|
|
if _, err := lkv1.Get(context.TODO(), "k"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
clus.Members[0].Stop(t)
|
|
clus.WaitLeader(t)
|
|
|
|
// put some more revisions for compaction
|
|
presp, err := clus.Client(1).Put(context.TODO(), "a", "123")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
presp, err = clus.Client(1).Put(context.TODO(), "a", "123")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
// compact while lkv1 is disconnected
|
|
rev := presp.Header.Revision
|
|
if _, err = clus.Client(1).Compact(context.TODO(), rev); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
clus.Members[0].Restart(t)
|
|
|
|
cctx, cancel := context.WithTimeout(context.TODO(), 5*time.Second)
|
|
defer cancel()
|
|
if _, err = lkv2.Put(cctx, "k", "v"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
resp, err := lkv1.Get(cctx, "k")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if string(resp.Kvs[0].Value) != "v" {
|
|
t.Fatalf(`expected "v" value, got %+v`, resp)
|
|
}
|
|
}
|
|
|
|
// TestLeasingReconnectOwnerConsistency checks a write error on an owner will
|
|
// not cause inconsistency between the server and the client.
|
|
func TestLeasingReconnectOwnerConsistency(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/")
|
|
defer closeLKV()
|
|
testutil.AssertNil(t, err)
|
|
|
|
if _, err = lkv.Put(context.TODO(), "k", "x"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if _, err = lkv.Put(context.TODO(), "kk", "y"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if _, err = lkv.Get(context.TODO(), "k"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
for i := 0; i < 10; i++ {
|
|
v := fmt.Sprintf("%d", i)
|
|
donec := make(chan struct{})
|
|
clus.Members[0].DropConnections()
|
|
go func() {
|
|
defer close(donec)
|
|
for i := 0; i < 20; i++ {
|
|
clus.Members[0].DropConnections()
|
|
time.Sleep(time.Millisecond)
|
|
}
|
|
}()
|
|
switch rand.Intn(7) {
|
|
case 0:
|
|
_, err = lkv.Put(context.TODO(), "k", v)
|
|
case 1:
|
|
_, err = lkv.Delete(context.TODO(), "k")
|
|
case 2:
|
|
txn := lkv.Txn(context.TODO()).Then(
|
|
clientv3.OpGet("k"),
|
|
clientv3.OpDelete("k"),
|
|
)
|
|
_, err = txn.Commit()
|
|
case 3:
|
|
txn := lkv.Txn(context.TODO()).Then(
|
|
clientv3.OpGet("k"),
|
|
clientv3.OpPut("k", v),
|
|
)
|
|
_, err = txn.Commit()
|
|
case 4:
|
|
_, err = lkv.Do(context.TODO(), clientv3.OpPut("k", v))
|
|
case 5:
|
|
_, err = lkv.Do(context.TODO(), clientv3.OpDelete("k"))
|
|
case 6:
|
|
_, err = lkv.Delete(context.TODO(), "k", clientv3.WithPrefix())
|
|
}
|
|
<-donec
|
|
if err != nil {
|
|
// TODO wrap input client to generate errors
|
|
break
|
|
}
|
|
}
|
|
|
|
lresp, lerr := lkv.Get(context.TODO(), "k")
|
|
if lerr != nil {
|
|
t.Fatal(lerr)
|
|
}
|
|
cresp, cerr := clus.Client(0).Get(context.TODO(), "k")
|
|
if cerr != nil {
|
|
t.Fatal(cerr)
|
|
}
|
|
if !reflect.DeepEqual(lresp.Kvs, cresp.Kvs) {
|
|
t.Fatalf("expected %+v, got %+v", cresp, lresp)
|
|
}
|
|
}
|
|
|
|
func TestLeasingTxnAtomicCache(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
puts, gets := make([]clientv3.Op, 16), make([]clientv3.Op, 16)
|
|
for i := range puts {
|
|
k := fmt.Sprintf("k-%d", i)
|
|
puts[i], gets[i] = clientv3.OpPut(k, k), clientv3.OpGet(k)
|
|
}
|
|
if _, err = clus.Client(0).Txn(context.TODO()).Then(puts...).Commit(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
for i := range gets {
|
|
if _, err = lkv.Do(context.TODO(), gets[i]); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
numPutters, numGetters := 16, 16
|
|
|
|
var wgPutters, wgGetters sync.WaitGroup
|
|
wgPutters.Add(numPutters)
|
|
wgGetters.Add(numGetters)
|
|
|
|
f := func() {
|
|
defer wgPutters.Done()
|
|
for i := 0; i < 10; i++ {
|
|
if _, txnerr := lkv.Txn(context.TODO()).Then(puts...).Commit(); err != nil {
|
|
t.Fatal(txnerr)
|
|
}
|
|
}
|
|
}
|
|
|
|
donec := make(chan struct{}, numPutters)
|
|
g := func() {
|
|
defer wgGetters.Done()
|
|
for {
|
|
select {
|
|
case <-donec:
|
|
return
|
|
default:
|
|
}
|
|
tresp, err := lkv.Txn(context.TODO()).Then(gets...).Commit()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
revs := make([]int64, len(gets))
|
|
for i, resp := range tresp.Responses {
|
|
rr := resp.GetResponseRange()
|
|
revs[i] = rr.Kvs[0].ModRevision
|
|
}
|
|
for i := 1; i < len(revs); i++ {
|
|
if revs[i] != revs[i-1] {
|
|
t.Fatalf("expected matching revisions, got %+v", revs)
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
for i := 0; i < numGetters; i++ {
|
|
go g()
|
|
}
|
|
for i := 0; i < numPutters; i++ {
|
|
go f()
|
|
}
|
|
|
|
wgPutters.Wait()
|
|
close(donec)
|
|
wgGetters.Wait()
|
|
}
|
|
|
|
// TestLeasingReconnectTxn checks that Txn is resilient to disconnects.
|
|
func TestLeasingReconnectTxn(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
if _, err = lkv.Get(context.TODO(), "k"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
donec := make(chan struct{})
|
|
go func() {
|
|
defer close(donec)
|
|
clus.Members[0].DropConnections()
|
|
for i := 0; i < 10; i++ {
|
|
clus.Members[0].DropConnections()
|
|
time.Sleep(time.Millisecond)
|
|
}
|
|
}()
|
|
|
|
_, lerr := lkv.Txn(context.TODO()).
|
|
If(clientv3.Compare(clientv3.Version("k"), "=", 0)).
|
|
Then(clientv3.OpGet("k")).
|
|
Commit()
|
|
<-donec
|
|
if lerr != nil {
|
|
t.Fatal(lerr)
|
|
}
|
|
}
|
|
|
|
// TestLeasingReconnectNonOwnerGet checks a get error on an owner will
|
|
// not cause inconsistency between the server and the client.
|
|
func TestLeasingReconnectNonOwnerGet(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
// populate a few keys so some leasing gets have keys
|
|
for i := 0; i < 4; i++ {
|
|
k := fmt.Sprintf("k-%d", i*2)
|
|
if _, err = lkv.Put(context.TODO(), k, k[2:]); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
n := 0
|
|
for i := 0; i < 10; i++ {
|
|
donec := make(chan struct{})
|
|
clus.Members[0].DropConnections()
|
|
go func() {
|
|
defer close(donec)
|
|
for j := 0; j < 10; j++ {
|
|
clus.Members[0].DropConnections()
|
|
time.Sleep(time.Millisecond)
|
|
}
|
|
}()
|
|
_, err = lkv.Get(context.TODO(), fmt.Sprintf("k-%d", i))
|
|
<-donec
|
|
n++
|
|
if err != nil {
|
|
break
|
|
}
|
|
}
|
|
for i := 0; i < n; i++ {
|
|
k := fmt.Sprintf("k-%d", i)
|
|
lresp, lerr := lkv.Get(context.TODO(), k)
|
|
if lerr != nil {
|
|
t.Fatal(lerr)
|
|
}
|
|
cresp, cerr := clus.Client(0).Get(context.TODO(), k)
|
|
if cerr != nil {
|
|
t.Fatal(cerr)
|
|
}
|
|
if !reflect.DeepEqual(lresp.Kvs, cresp.Kvs) {
|
|
t.Fatalf("expected %+v, got %+v", cresp, lresp)
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestLeasingTxnRangeCmp(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
if _, err = clus.Client(0).Put(context.TODO(), "k", "a"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
// k2 version = 2
|
|
if _, err = clus.Client(0).Put(context.TODO(), "k2", "a"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if _, err = clus.Client(0).Put(context.TODO(), "k2", "a"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// cache k
|
|
if _, err = lkv.Get(context.TODO(), "k"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
cmp := clientv3.Compare(clientv3.Version("k").WithPrefix(), "=", 1)
|
|
tresp, terr := lkv.Txn(context.TODO()).If(cmp).Commit()
|
|
if terr != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if tresp.Succeeded {
|
|
t.Fatalf("expected Succeeded=false, got %+v", tresp)
|
|
}
|
|
}
|
|
|
|
func TestLeasingDo(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 1})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
ops := []clientv3.Op{
|
|
clientv3.OpTxn(nil, nil, nil),
|
|
clientv3.OpGet("a"),
|
|
clientv3.OpPut("a/abc", "v"),
|
|
clientv3.OpDelete("a", clientv3.WithPrefix()),
|
|
clientv3.OpTxn(nil, nil, nil),
|
|
}
|
|
for i, op := range ops {
|
|
resp, resperr := lkv.Do(context.TODO(), op)
|
|
if resperr != nil {
|
|
t.Errorf("#%d: failed (%v)", i, resperr)
|
|
}
|
|
switch {
|
|
case op.IsGet() && resp.Get() == nil:
|
|
t.Errorf("#%d: get but nil get response", i)
|
|
case op.IsPut() && resp.Put() == nil:
|
|
t.Errorf("#%d: put op but nil get response", i)
|
|
case op.IsDelete() && resp.Del() == nil:
|
|
t.Errorf("#%d: delete op but nil delete response", i)
|
|
case op.IsTxn() && resp.Txn() == nil:
|
|
t.Errorf("#%d: txn op but nil txn response", i)
|
|
}
|
|
}
|
|
|
|
gresp, err := clus.Client(0).Get(context.TODO(), "a", clientv3.WithPrefix())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if len(gresp.Kvs) != 0 {
|
|
t.Fatalf("expected no keys, got %+v", gresp.Kvs)
|
|
}
|
|
}
|
|
|
|
func TestLeasingTxnOwnerPutBranch(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
n := 0
|
|
treeOp := makePutTreeOp("tree", &n, 4)
|
|
for i := 0; i < n; i++ {
|
|
k := fmt.Sprintf("tree/%d", i)
|
|
if _, err = clus.Client(0).Put(context.TODO(), k, "a"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if _, err = lkv.Get(context.TODO(), k); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
if _, err = lkv.Do(context.TODO(), treeOp); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// lkv shouldn't need to call out to server for updated leased keys
|
|
clus.Members[0].Stop(t)
|
|
|
|
for i := 0; i < n; i++ {
|
|
k := fmt.Sprintf("tree/%d", i)
|
|
lkvResp, err := lkv.Get(context.TODO(), k)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
clusResp, err := clus.Client(1).Get(context.TODO(), k)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if !reflect.DeepEqual(clusResp.Kvs, lkvResp.Kvs) {
|
|
t.Fatalf("expected %+v, got %+v", clusResp.Kvs, lkvResp.Kvs)
|
|
}
|
|
}
|
|
}
|
|
|
|
func makePutTreeOp(pfx string, v *int, depth int) clientv3.Op {
|
|
key := fmt.Sprintf("%s/%d", pfx, *v)
|
|
*v = *v + 1
|
|
if depth == 0 {
|
|
return clientv3.OpPut(key, "leaf")
|
|
}
|
|
|
|
t, e := makePutTreeOp(pfx, v, depth-1), makePutTreeOp(pfx, v, depth-1)
|
|
tPut, ePut := clientv3.OpPut(key, "then"), clientv3.OpPut(key, "else")
|
|
|
|
cmps := make([]clientv3.Cmp, 1)
|
|
if rand.Intn(2) == 0 {
|
|
// follow then path
|
|
cmps[0] = clientv3.Compare(clientv3.Version("nokey"), "=", 0)
|
|
} else {
|
|
// follow else path
|
|
cmps[0] = clientv3.Compare(clientv3.Version("nokey"), ">", 0)
|
|
}
|
|
|
|
return clientv3.OpTxn(cmps, []clientv3.Op{t, tPut}, []clientv3.Op{e, ePut})
|
|
}
|
|
|
|
func randCmps(pfx string, dat []*clientv3.PutResponse) (cmps []clientv3.Cmp, then bool) {
|
|
for i := 0; i < len(dat); i++ {
|
|
idx := rand.Intn(len(dat))
|
|
k := fmt.Sprintf("%s%d", pfx, idx)
|
|
rev := dat[idx].Header.Revision
|
|
var cmp clientv3.Cmp
|
|
switch rand.Intn(4) {
|
|
case 0:
|
|
cmp = clientv3.Compare(clientv3.CreateRevision(k), ">", rev-1)
|
|
case 1:
|
|
cmp = clientv3.Compare(clientv3.Version(k), "=", 1)
|
|
case 2:
|
|
cmp = clientv3.Compare(clientv3.CreateRevision(k), "=", rev)
|
|
case 3:
|
|
cmp = clientv3.Compare(clientv3.CreateRevision(k), "!=", rev+1)
|
|
|
|
}
|
|
cmps = append(cmps, cmp)
|
|
}
|
|
cmps = cmps[:rand.Intn(len(dat))]
|
|
if rand.Intn(2) == 0 {
|
|
return cmps, true
|
|
}
|
|
i := rand.Intn(len(dat))
|
|
cmps = append(cmps, clientv3.Compare(clientv3.Version(fmt.Sprintf("k-%d", i)), "=", 0))
|
|
return cmps, false
|
|
}
|
|
|
|
func TestLeasingSessionExpire(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3})
|
|
defer clus.Terminate(t)
|
|
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/", concurrency.WithTTL(1))
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
lkv2, closeLKV2, err := leasing.NewKV(clus.Client(0), "foo/")
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV2()
|
|
|
|
// acquire lease on abc
|
|
if _, err = lkv.Get(context.TODO(), "abc"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// down endpoint lkv uses for keepalives
|
|
clus.Members[0].Stop(t)
|
|
if err = waitForLeasingExpire(clus.Client(1), "foo/abc"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
waitForExpireAck(t, lkv)
|
|
clus.Members[0].Restart(t)
|
|
|
|
if _, err = lkv2.Put(context.TODO(), "abc", "def"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
resp, err := lkv.Get(context.TODO(), "abc")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if v := string(resp.Kvs[0].Value); v != "def" {
|
|
t.Fatalf("expected %q, got %q", "v", v)
|
|
}
|
|
}
|
|
|
|
func TestLeasingSessionExpireCancel(t *testing.T) {
|
|
defer testutil.AfterTest(t)
|
|
clus := integration.NewClusterV3(t, &integration.ClusterConfig{Size: 3})
|
|
defer clus.Terminate(t)
|
|
|
|
tests := []func(context.Context, clientv3.KV) error{
|
|
func(ctx context.Context, kv clientv3.KV) error {
|
|
_, err := kv.Get(ctx, "abc")
|
|
return err
|
|
},
|
|
func(ctx context.Context, kv clientv3.KV) error {
|
|
_, err := kv.Delete(ctx, "abc")
|
|
return err
|
|
},
|
|
func(ctx context.Context, kv clientv3.KV) error {
|
|
_, err := kv.Put(ctx, "abc", "v")
|
|
return err
|
|
},
|
|
func(ctx context.Context, kv clientv3.KV) error {
|
|
_, err := kv.Txn(ctx).Then(clientv3.OpGet("abc")).Commit()
|
|
return err
|
|
},
|
|
func(ctx context.Context, kv clientv3.KV) error {
|
|
_, err := kv.Do(ctx, clientv3.OpPut("abc", "v"))
|
|
return err
|
|
},
|
|
func(ctx context.Context, kv clientv3.KV) error {
|
|
_, err := kv.Do(ctx, clientv3.OpDelete("abc"))
|
|
return err
|
|
},
|
|
func(ctx context.Context, kv clientv3.KV) error {
|
|
_, err := kv.Do(ctx, clientv3.OpGet("abc"))
|
|
return err
|
|
},
|
|
func(ctx context.Context, kv clientv3.KV) error {
|
|
op := clientv3.OpTxn(nil, []clientv3.Op{clientv3.OpGet("abc")}, nil)
|
|
_, err := kv.Do(ctx, op)
|
|
return err
|
|
},
|
|
}
|
|
for i := range tests {
|
|
lkv, closeLKV, err := leasing.NewKV(clus.Client(0), "foo/", concurrency.WithTTL(1))
|
|
testutil.AssertNil(t, err)
|
|
defer closeLKV()
|
|
|
|
if _, err = lkv.Get(context.TODO(), "abc"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// down endpoint lkv uses for keepalives
|
|
clus.Members[0].Stop(t)
|
|
if err := waitForLeasingExpire(clus.Client(1), "foo/abc"); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
waitForExpireAck(t, lkv)
|
|
|
|
ctx, cancel := context.WithCancel(context.TODO())
|
|
errc := make(chan error, 1)
|
|
go func() { errc <- tests[i](ctx, lkv) }()
|
|
// some delay to get past for ctx.Err() != nil {} loops
|
|
time.Sleep(100 * time.Millisecond)
|
|
cancel()
|
|
|
|
select {
|
|
case err := <-errc:
|
|
if err != ctx.Err() {
|
|
t.Errorf("#%d: expected %v, got %v", i, ctx.Err(), err)
|
|
}
|
|
case <-time.After(5 * time.Second):
|
|
t.Errorf("#%d: timed out waiting for cancel", i)
|
|
}
|
|
clus.Members[0].Restart(t)
|
|
}
|
|
}
|
|
|
|
func waitForLeasingExpire(kv clientv3.KV, lkey string) error {
|
|
for {
|
|
time.Sleep(1 * time.Second)
|
|
resp, err := kv.Get(context.TODO(), lkey, clientv3.WithPrefix())
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if len(resp.Kvs) == 0 {
|
|
// server expired the leasing key
|
|
return nil
|
|
}
|
|
}
|
|
}
|
|
|
|
func waitForExpireAck(t *testing.T, kv clientv3.KV) {
|
|
// wait for leasing client to acknowledge lost lease
|
|
for i := 0; i < 10; i++ {
|
|
ctx, cancel := context.WithTimeout(context.TODO(), time.Second)
|
|
_, err := kv.Get(ctx, "abc")
|
|
cancel()
|
|
if err == ctx.Err() {
|
|
return
|
|
}
|
|
time.Sleep(time.Second)
|
|
}
|
|
t.Fatalf("waited too long to acknlowedge lease expiration")
|
|
}
|