mirror of
				https://github.com/etcd-io/etcd.git
				synced 2024-09-27 06:25:44 +00:00 
			
		
		
		
	 1caaa9ed4a
			
		
	
	
		1caaa9ed4a
		
	
	
	
	
		
			
			Update to Go 1.12.5 testing. Remove deprecated unused and gosimple pacakges, and mask staticcheck 1006. Also, fix unconvert errors related to unnecessary type conversions and following staticcheck errors: - remove redundant return statements - use for range instead of for select - use time.Since instead of time.Now().Sub - omit comparison to bool constant - replace T.Fatal and T.Fatalf in tests with T.Error and T.Fatalf respectively because the goroutine calls T.Fatal must be called in the same goroutine as the test - fix error strings that should not be capitalized - use sort.Strings(...) instead of sort.Sort(sort.StringSlice(...)) - use he status code of Canceled instead of grpc.ErrClientConnClosing which is deprecated - use use status.Errorf instead of grpc.Errorf which is deprecated Related #10528 #10438
		
			
				
	
	
		
			324 lines
		
	
	
		
			8.5 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			324 lines
		
	
	
		
			8.5 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| // Copyright 2016 The etcd Authors
 | |
| //
 | |
| // Licensed under the Apache License, Version 2.0 (the "License");
 | |
| // you may not use this file except in compliance with the License.
 | |
| // You may obtain a copy of the License at
 | |
| //
 | |
| //     http://www.apache.org/licenses/LICENSE-2.0
 | |
| //
 | |
| // Unless required by applicable law or agreed to in writing, software
 | |
| // distributed under the License is distributed on an "AS IS" BASIS,
 | |
| // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
| // See the License for the specific language governing permissions and
 | |
| // limitations under the License.
 | |
| 
 | |
| package integration
 | |
| 
 | |
| import (
 | |
| 	"context"
 | |
| 	"math/rand"
 | |
| 	"sync"
 | |
| 	"testing"
 | |
| 	"time"
 | |
| 
 | |
| 	"go.etcd.io/etcd/clientv3"
 | |
| 	"go.etcd.io/etcd/clientv3/concurrency"
 | |
| 	"go.etcd.io/etcd/contrib/recipes"
 | |
| 	"go.etcd.io/etcd/mvcc/mvccpb"
 | |
| 	"go.etcd.io/etcd/pkg/testutil"
 | |
| )
 | |
| 
 | |
| func TestMutexSingleNode(t *testing.T) {
 | |
| 	clus := NewClusterV3(t, &ClusterConfig{Size: 3})
 | |
| 	defer clus.Terminate(t)
 | |
| 
 | |
| 	var clients []*clientv3.Client
 | |
| 	testMutex(t, 5, makeSingleNodeClients(t, clus.cluster, &clients))
 | |
| 	closeClients(t, clients)
 | |
| }
 | |
| 
 | |
| func TestMutexMultiNode(t *testing.T) {
 | |
| 	clus := NewClusterV3(t, &ClusterConfig{Size: 3})
 | |
| 	defer clus.Terminate(t)
 | |
| 
 | |
| 	var clients []*clientv3.Client
 | |
| 	testMutex(t, 5, makeMultiNodeClients(t, clus.cluster, &clients))
 | |
| 	closeClients(t, clients)
 | |
| }
 | |
| 
 | |
| func testMutex(t *testing.T, waiters int, chooseClient func() *clientv3.Client) {
 | |
| 	// stream lock acquisitions
 | |
| 	lockedC := make(chan *concurrency.Mutex)
 | |
| 	for i := 0; i < waiters; i++ {
 | |
| 		go func() {
 | |
| 			session, err := concurrency.NewSession(chooseClient())
 | |
| 			if err != nil {
 | |
| 				t.Error(err)
 | |
| 			}
 | |
| 			m := concurrency.NewMutex(session, "test-mutex")
 | |
| 			if err := m.Lock(context.TODO()); err != nil {
 | |
| 				t.Errorf("could not wait on lock (%v)", err)
 | |
| 			}
 | |
| 			lockedC <- m
 | |
| 		}()
 | |
| 	}
 | |
| 	// unlock locked mutexes
 | |
| 	timerC := time.After(time.Duration(waiters) * time.Second)
 | |
| 	for i := 0; i < waiters; i++ {
 | |
| 		select {
 | |
| 		case <-timerC:
 | |
| 			t.Fatalf("timed out waiting for lock %d", i)
 | |
| 		case m := <-lockedC:
 | |
| 			// lock acquired with m
 | |
| 			select {
 | |
| 			case <-lockedC:
 | |
| 				t.Fatalf("lock %d followers did not wait", i)
 | |
| 			default:
 | |
| 			}
 | |
| 			if err := m.Unlock(context.TODO()); err != nil {
 | |
| 				t.Fatalf("could not release lock (%v)", err)
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // TestMutexSessionRelock ensures that acquiring the same lock with the same
 | |
| // session will not result in deadlock.
 | |
| func TestMutexSessionRelock(t *testing.T) {
 | |
| 	clus := NewClusterV3(t, &ClusterConfig{Size: 3})
 | |
| 	defer clus.Terminate(t)
 | |
| 	session, err := concurrency.NewSession(clus.RandClient())
 | |
| 	if err != nil {
 | |
| 		t.Error(err)
 | |
| 	}
 | |
| 
 | |
| 	m := concurrency.NewMutex(session, "test-mutex")
 | |
| 	if err := m.Lock(context.TODO()); err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 
 | |
| 	m2 := concurrency.NewMutex(session, "test-mutex")
 | |
| 	if err := m2.Lock(context.TODO()); err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // TestMutexWaitsOnCurrentHolder ensures a mutex is only acquired once all
 | |
| // waiters older than the new owner are gone by testing the case where
 | |
| // the waiter prior to the acquirer expires before the current holder.
 | |
| func TestMutexWaitsOnCurrentHolder(t *testing.T) {
 | |
| 	defer testutil.AfterTest(t)
 | |
| 
 | |
| 	clus := NewClusterV3(t, &ClusterConfig{Size: 1})
 | |
| 	defer clus.Terminate(t)
 | |
| 
 | |
| 	cctx := context.Background()
 | |
| 
 | |
| 	cli := clus.Client(0)
 | |
| 
 | |
| 	firstOwnerSession, err := concurrency.NewSession(cli)
 | |
| 	if err != nil {
 | |
| 		t.Error(err)
 | |
| 	}
 | |
| 	defer firstOwnerSession.Close()
 | |
| 	firstOwnerMutex := concurrency.NewMutex(firstOwnerSession, "test-mutex")
 | |
| 	if err = firstOwnerMutex.Lock(cctx); err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 
 | |
| 	victimSession, err := concurrency.NewSession(cli)
 | |
| 	if err != nil {
 | |
| 		t.Error(err)
 | |
| 	}
 | |
| 	defer victimSession.Close()
 | |
| 	victimDonec := make(chan struct{})
 | |
| 	go func() {
 | |
| 		defer close(victimDonec)
 | |
| 		concurrency.NewMutex(victimSession, "test-mutex").Lock(cctx)
 | |
| 	}()
 | |
| 
 | |
| 	// ensure mutexes associated with firstOwnerSession and victimSession waits before new owner
 | |
| 	wch := cli.Watch(cctx, "test-mutex", clientv3.WithPrefix(), clientv3.WithRev(1))
 | |
| 	putCounts := 0
 | |
| 	for putCounts < 2 {
 | |
| 		select {
 | |
| 		case wrp := <-wch:
 | |
| 			putCounts += len(wrp.Events)
 | |
| 		case <-time.After(time.Second):
 | |
| 			t.Fatal("failed to receive watch response")
 | |
| 		}
 | |
| 	}
 | |
| 	if putCounts != 2 {
 | |
| 		t.Fatalf("expect 2 put events, but got %v", putCounts)
 | |
| 	}
 | |
| 
 | |
| 	newOwnerSession, err := concurrency.NewSession(cli)
 | |
| 	if err != nil {
 | |
| 		t.Error(err)
 | |
| 	}
 | |
| 	defer newOwnerSession.Close()
 | |
| 	newOwnerDonec := make(chan struct{})
 | |
| 	go func() {
 | |
| 		defer close(newOwnerDonec)
 | |
| 		concurrency.NewMutex(newOwnerSession, "test-mutex").Lock(cctx)
 | |
| 	}()
 | |
| 
 | |
| 	select {
 | |
| 	case wrp := <-wch:
 | |
| 		if len(wrp.Events) != 1 {
 | |
| 			t.Fatalf("expect a event, but got %v events", len(wrp.Events))
 | |
| 		}
 | |
| 		if e := wrp.Events[0]; e.Type != mvccpb.PUT {
 | |
| 			t.Fatalf("expect a put event on prefix test-mutex, but got event type %v", e.Type)
 | |
| 		}
 | |
| 	case <-time.After(time.Second):
 | |
| 		t.Fatalf("failed to receive a watch response")
 | |
| 	}
 | |
| 
 | |
| 	// simulate losing the client that's next in line to acquire the lock
 | |
| 	victimSession.Close()
 | |
| 
 | |
| 	// ensures the deletion of victim waiter from server side.
 | |
| 	select {
 | |
| 	case wrp := <-wch:
 | |
| 		if len(wrp.Events) != 1 {
 | |
| 			t.Fatalf("expect a event, but got %v events", len(wrp.Events))
 | |
| 		}
 | |
| 		if e := wrp.Events[0]; e.Type != mvccpb.DELETE {
 | |
| 			t.Fatalf("expect a delete event on prefix test-mutex, but got event type %v", e.Type)
 | |
| 		}
 | |
| 	case <-time.After(time.Second):
 | |
| 		t.Fatal("failed to receive a watch response")
 | |
| 	}
 | |
| 
 | |
| 	select {
 | |
| 	case <-newOwnerDonec:
 | |
| 		t.Fatal("new owner obtained lock before first owner unlocked")
 | |
| 	default:
 | |
| 	}
 | |
| 
 | |
| 	if err := firstOwnerMutex.Unlock(cctx); err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 
 | |
| 	select {
 | |
| 	case <-newOwnerDonec:
 | |
| 	case <-time.After(time.Second):
 | |
| 		t.Fatal("new owner failed to obtain lock")
 | |
| 	}
 | |
| 
 | |
| 	select {
 | |
| 	case <-victimDonec:
 | |
| 	case <-time.After(time.Second):
 | |
| 		t.Fatal("victim mutex failed to exit after first owner releases lock")
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func BenchmarkMutex4Waiters(b *testing.B) {
 | |
| 	// XXX switch tests to use TB interface
 | |
| 	clus := NewClusterV3(nil, &ClusterConfig{Size: 3})
 | |
| 	defer clus.Terminate(nil)
 | |
| 	for i := 0; i < b.N; i++ {
 | |
| 		testMutex(nil, 4, func() *clientv3.Client { return clus.RandClient() })
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func TestRWMutexSingleNode(t *testing.T) {
 | |
| 	clus := NewClusterV3(t, &ClusterConfig{Size: 3})
 | |
| 	defer clus.Terminate(t)
 | |
| 	testRWMutex(t, 5, func() *clientv3.Client { return clus.clients[0] })
 | |
| }
 | |
| 
 | |
| func TestRWMutexMultiNode(t *testing.T) {
 | |
| 	clus := NewClusterV3(t, &ClusterConfig{Size: 3})
 | |
| 	defer clus.Terminate(t)
 | |
| 	testRWMutex(t, 5, func() *clientv3.Client { return clus.RandClient() })
 | |
| }
 | |
| 
 | |
| func testRWMutex(t *testing.T, waiters int, chooseClient func() *clientv3.Client) {
 | |
| 	// stream rwlock acquistions
 | |
| 	rlockedC := make(chan *recipe.RWMutex, 1)
 | |
| 	wlockedC := make(chan *recipe.RWMutex, 1)
 | |
| 	for i := 0; i < waiters; i++ {
 | |
| 		go func() {
 | |
| 			session, err := concurrency.NewSession(chooseClient())
 | |
| 			if err != nil {
 | |
| 				t.Error(err)
 | |
| 			}
 | |
| 			rwm := recipe.NewRWMutex(session, "test-rwmutex")
 | |
| 			if rand.Intn(2) == 0 {
 | |
| 				if err := rwm.RLock(); err != nil {
 | |
| 					t.Errorf("could not rlock (%v)", err)
 | |
| 				}
 | |
| 				rlockedC <- rwm
 | |
| 			} else {
 | |
| 				if err := rwm.Lock(); err != nil {
 | |
| 					t.Errorf("could not lock (%v)", err)
 | |
| 				}
 | |
| 				wlockedC <- rwm
 | |
| 			}
 | |
| 		}()
 | |
| 	}
 | |
| 	// unlock locked rwmutexes
 | |
| 	timerC := time.After(time.Duration(waiters) * time.Second)
 | |
| 	for i := 0; i < waiters; i++ {
 | |
| 		select {
 | |
| 		case <-timerC:
 | |
| 			t.Fatalf("timed out waiting for lock %d", i)
 | |
| 		case wl := <-wlockedC:
 | |
| 			select {
 | |
| 			case <-rlockedC:
 | |
| 				t.Fatalf("rlock %d readers did not wait", i)
 | |
| 			default:
 | |
| 			}
 | |
| 			if err := wl.Unlock(); err != nil {
 | |
| 				t.Fatalf("could not release lock (%v)", err)
 | |
| 			}
 | |
| 		case rl := <-rlockedC:
 | |
| 			select {
 | |
| 			case <-wlockedC:
 | |
| 				t.Fatalf("rlock %d writers did not wait", i)
 | |
| 			default:
 | |
| 			}
 | |
| 			if err := rl.RUnlock(); err != nil {
 | |
| 				t.Fatalf("could not release rlock (%v)", err)
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func makeClients(t *testing.T, clients *[]*clientv3.Client, choose func() *member) func() *clientv3.Client {
 | |
| 	var mu sync.Mutex
 | |
| 	*clients = nil
 | |
| 	return func() *clientv3.Client {
 | |
| 		cli, err := NewClientV3(choose())
 | |
| 		if err != nil {
 | |
| 			t.Fatalf("cannot create client: %v", err)
 | |
| 		}
 | |
| 		mu.Lock()
 | |
| 		*clients = append(*clients, cli)
 | |
| 		mu.Unlock()
 | |
| 		return cli
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func makeSingleNodeClients(t *testing.T, clus *cluster, clients *[]*clientv3.Client) func() *clientv3.Client {
 | |
| 	return makeClients(t, clients, func() *member {
 | |
| 		return clus.Members[0]
 | |
| 	})
 | |
| }
 | |
| 
 | |
| func makeMultiNodeClients(t *testing.T, clus *cluster, clients *[]*clientv3.Client) func() *clientv3.Client {
 | |
| 	return makeClients(t, clients, func() *member {
 | |
| 		return clus.Members[rand.Intn(len(clus.Members))]
 | |
| 	})
 | |
| }
 | |
| 
 | |
| func closeClients(t *testing.T, clients []*clientv3.Client) {
 | |
| 	for _, cli := range clients {
 | |
| 		if err := cli.Close(); err != nil {
 | |
| 			t.Fatal(err)
 | |
| 		}
 | |
| 	}
 | |
| }
 |