mirror of
				https://github.com/etcd-io/etcd.git
				synced 2024-09-27 06:25:44 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			307 lines
		
	
	
		
			6.7 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			307 lines
		
	
	
		
			6.7 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| // Copyright 2015 The etcd Authors
 | |
| //
 | |
| // Licensed under the Apache License, Version 2.0 (the "License");
 | |
| // you may not use this file except in compliance with the License.
 | |
| // You may obtain a copy of the License at
 | |
| //
 | |
| //     http://www.apache.org/licenses/LICENSE-2.0
 | |
| //
 | |
| // Unless required by applicable law or agreed to in writing, software
 | |
| // distributed under the License is distributed on an "AS IS" BASIS,
 | |
| // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
| // See the License for the specific language governing permissions and
 | |
| // limitations under the License.
 | |
| 
 | |
| package backend
 | |
| 
 | |
| import (
 | |
| 	"fmt"
 | |
| 	"io/ioutil"
 | |
| 	"os"
 | |
| 	"reflect"
 | |
| 	"testing"
 | |
| 	"time"
 | |
| 
 | |
| 	bolt "github.com/coreos/bbolt"
 | |
| )
 | |
| 
 | |
| func TestBackendClose(t *testing.T) {
 | |
| 	b, tmpPath := NewTmpBackend(time.Hour, 10000)
 | |
| 	defer os.Remove(tmpPath)
 | |
| 
 | |
| 	// check close could work
 | |
| 	done := make(chan struct{})
 | |
| 	go func() {
 | |
| 		err := b.Close()
 | |
| 		if err != nil {
 | |
| 			t.Errorf("close error = %v, want nil", err)
 | |
| 		}
 | |
| 		done <- struct{}{}
 | |
| 	}()
 | |
| 	select {
 | |
| 	case <-done:
 | |
| 	case <-time.After(10 * time.Second):
 | |
| 		t.Errorf("failed to close database in 10s")
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func TestBackendSnapshot(t *testing.T) {
 | |
| 	b, tmpPath := NewTmpBackend(time.Hour, 10000)
 | |
| 	defer cleanup(b, tmpPath)
 | |
| 
 | |
| 	tx := b.BatchTx()
 | |
| 	tx.Lock()
 | |
| 	tx.UnsafeCreateBucket([]byte("test"))
 | |
| 	tx.UnsafePut([]byte("test"), []byte("foo"), []byte("bar"))
 | |
| 	tx.Unlock()
 | |
| 	b.ForceCommit()
 | |
| 
 | |
| 	// write snapshot to a new file
 | |
| 	f, err := ioutil.TempFile(os.TempDir(), "etcd_backend_test")
 | |
| 	if err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 	snap := b.Snapshot()
 | |
| 	defer snap.Close()
 | |
| 	if _, err := snap.WriteTo(f); err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 	f.Close()
 | |
| 
 | |
| 	// bootstrap new backend from the snapshot
 | |
| 	bcfg := DefaultBackendConfig()
 | |
| 	bcfg.Path, bcfg.BatchInterval, bcfg.BatchLimit = f.Name(), time.Hour, 10000
 | |
| 	nb := New(bcfg)
 | |
| 	defer cleanup(nb, f.Name())
 | |
| 
 | |
| 	newTx := b.BatchTx()
 | |
| 	newTx.Lock()
 | |
| 	ks, _ := newTx.UnsafeRange([]byte("test"), []byte("foo"), []byte("goo"), 0)
 | |
| 	if len(ks) != 1 {
 | |
| 		t.Errorf("len(kvs) = %d, want 1", len(ks))
 | |
| 	}
 | |
| 	newTx.Unlock()
 | |
| }
 | |
| 
 | |
| func TestBackendBatchIntervalCommit(t *testing.T) {
 | |
| 	// start backend with super short batch interval so
 | |
| 	// we do not need to wait long before commit to happen.
 | |
| 	b, tmpPath := NewTmpBackend(time.Nanosecond, 10000)
 | |
| 	defer cleanup(b, tmpPath)
 | |
| 
 | |
| 	pc := b.Commits()
 | |
| 
 | |
| 	tx := b.BatchTx()
 | |
| 	tx.Lock()
 | |
| 	tx.UnsafeCreateBucket([]byte("test"))
 | |
| 	tx.UnsafePut([]byte("test"), []byte("foo"), []byte("bar"))
 | |
| 	tx.Unlock()
 | |
| 
 | |
| 	for i := 0; i < 10; i++ {
 | |
| 		if b.Commits() >= pc+1 {
 | |
| 			break
 | |
| 		}
 | |
| 		time.Sleep(time.Duration(i*100) * time.Millisecond)
 | |
| 	}
 | |
| 
 | |
| 	// check whether put happens via db view
 | |
| 	b.db.View(func(tx *bolt.Tx) error {
 | |
| 		bucket := tx.Bucket([]byte("test"))
 | |
| 		if bucket == nil {
 | |
| 			t.Errorf("bucket test does not exit")
 | |
| 			return nil
 | |
| 		}
 | |
| 		v := bucket.Get([]byte("foo"))
 | |
| 		if v == nil {
 | |
| 			t.Errorf("foo key failed to written in backend")
 | |
| 		}
 | |
| 		return nil
 | |
| 	})
 | |
| }
 | |
| 
 | |
| func TestBackendDefrag(t *testing.T) {
 | |
| 	b, tmpPath := NewDefaultTmpBackend()
 | |
| 	defer cleanup(b, tmpPath)
 | |
| 
 | |
| 	tx := b.BatchTx()
 | |
| 	tx.Lock()
 | |
| 	tx.UnsafeCreateBucket([]byte("test"))
 | |
| 	for i := 0; i < defragLimit+100; i++ {
 | |
| 		tx.UnsafePut([]byte("test"), []byte(fmt.Sprintf("foo_%d", i)), []byte("bar"))
 | |
| 	}
 | |
| 	tx.Unlock()
 | |
| 	b.ForceCommit()
 | |
| 
 | |
| 	// remove some keys to ensure the disk space will be reclaimed after defrag
 | |
| 	tx = b.BatchTx()
 | |
| 	tx.Lock()
 | |
| 	for i := 0; i < 50; i++ {
 | |
| 		tx.UnsafeDelete([]byte("test"), []byte(fmt.Sprintf("foo_%d", i)))
 | |
| 	}
 | |
| 	tx.Unlock()
 | |
| 	b.ForceCommit()
 | |
| 
 | |
| 	size := b.Size()
 | |
| 
 | |
| 	// shrink and check hash
 | |
| 	oh, err := b.Hash(nil)
 | |
| 	if err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 
 | |
| 	err = b.Defrag()
 | |
| 	if err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 
 | |
| 	nh, err := b.Hash(nil)
 | |
| 	if err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 	if oh != nh {
 | |
| 		t.Errorf("hash = %v, want %v", nh, oh)
 | |
| 	}
 | |
| 
 | |
| 	nsize := b.Size()
 | |
| 	if nsize >= size {
 | |
| 		t.Errorf("new size = %v, want < %d", nsize, size)
 | |
| 	}
 | |
| 
 | |
| 	// try put more keys after shrink.
 | |
| 	tx = b.BatchTx()
 | |
| 	tx.Lock()
 | |
| 	tx.UnsafeCreateBucket([]byte("test"))
 | |
| 	tx.UnsafePut([]byte("test"), []byte("more"), []byte("bar"))
 | |
| 	tx.Unlock()
 | |
| 	b.ForceCommit()
 | |
| }
 | |
| 
 | |
| // TestBackendWriteback ensures writes are stored to the read txn on write txn unlock.
 | |
| func TestBackendWriteback(t *testing.T) {
 | |
| 	b, tmpPath := NewDefaultTmpBackend()
 | |
| 	defer cleanup(b, tmpPath)
 | |
| 
 | |
| 	tx := b.BatchTx()
 | |
| 	tx.Lock()
 | |
| 	tx.UnsafeCreateBucket([]byte("key"))
 | |
| 	tx.UnsafePut([]byte("key"), []byte("abc"), []byte("bar"))
 | |
| 	tx.UnsafePut([]byte("key"), []byte("def"), []byte("baz"))
 | |
| 	tx.UnsafePut([]byte("key"), []byte("overwrite"), []byte("1"))
 | |
| 	tx.Unlock()
 | |
| 
 | |
| 	// overwrites should be propagated too
 | |
| 	tx.Lock()
 | |
| 	tx.UnsafePut([]byte("key"), []byte("overwrite"), []byte("2"))
 | |
| 	tx.Unlock()
 | |
| 
 | |
| 	keys := []struct {
 | |
| 		key   []byte
 | |
| 		end   []byte
 | |
| 		limit int64
 | |
| 
 | |
| 		wkey [][]byte
 | |
| 		wval [][]byte
 | |
| 	}{
 | |
| 		{
 | |
| 			key: []byte("abc"),
 | |
| 			end: nil,
 | |
| 
 | |
| 			wkey: [][]byte{[]byte("abc")},
 | |
| 			wval: [][]byte{[]byte("bar")},
 | |
| 		},
 | |
| 		{
 | |
| 			key: []byte("abc"),
 | |
| 			end: []byte("def"),
 | |
| 
 | |
| 			wkey: [][]byte{[]byte("abc")},
 | |
| 			wval: [][]byte{[]byte("bar")},
 | |
| 		},
 | |
| 		{
 | |
| 			key: []byte("abc"),
 | |
| 			end: []byte("deg"),
 | |
| 
 | |
| 			wkey: [][]byte{[]byte("abc"), []byte("def")},
 | |
| 			wval: [][]byte{[]byte("bar"), []byte("baz")},
 | |
| 		},
 | |
| 		{
 | |
| 			key:   []byte("abc"),
 | |
| 			end:   []byte("\xff"),
 | |
| 			limit: 1,
 | |
| 
 | |
| 			wkey: [][]byte{[]byte("abc")},
 | |
| 			wval: [][]byte{[]byte("bar")},
 | |
| 		},
 | |
| 		{
 | |
| 			key: []byte("abc"),
 | |
| 			end: []byte("\xff"),
 | |
| 
 | |
| 			wkey: [][]byte{[]byte("abc"), []byte("def"), []byte("overwrite")},
 | |
| 			wval: [][]byte{[]byte("bar"), []byte("baz"), []byte("2")},
 | |
| 		},
 | |
| 	}
 | |
| 	rtx := b.ReadTx()
 | |
| 	for i, tt := range keys {
 | |
| 		rtx.Lock()
 | |
| 		k, v := rtx.UnsafeRange([]byte("key"), tt.key, tt.end, tt.limit)
 | |
| 		rtx.Unlock()
 | |
| 		if !reflect.DeepEqual(tt.wkey, k) || !reflect.DeepEqual(tt.wval, v) {
 | |
| 			t.Errorf("#%d: want k=%+v, v=%+v; got k=%+v, v=%+v", i, tt.wkey, tt.wval, k, v)
 | |
| 		}
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // TestBackendWritebackForEach checks that partially written / buffered
 | |
| // data is visited in the same order as fully committed data.
 | |
| func TestBackendWritebackForEach(t *testing.T) {
 | |
| 	b, tmpPath := NewTmpBackend(time.Hour, 10000)
 | |
| 	defer cleanup(b, tmpPath)
 | |
| 
 | |
| 	tx := b.BatchTx()
 | |
| 	tx.Lock()
 | |
| 	tx.UnsafeCreateBucket([]byte("key"))
 | |
| 	for i := 0; i < 5; i++ {
 | |
| 		k := []byte(fmt.Sprintf("%04d", i))
 | |
| 		tx.UnsafePut([]byte("key"), k, []byte("bar"))
 | |
| 	}
 | |
| 	tx.Unlock()
 | |
| 
 | |
| 	// writeback
 | |
| 	b.ForceCommit()
 | |
| 
 | |
| 	tx.Lock()
 | |
| 	tx.UnsafeCreateBucket([]byte("key"))
 | |
| 	for i := 5; i < 20; i++ {
 | |
| 		k := []byte(fmt.Sprintf("%04d", i))
 | |
| 		tx.UnsafePut([]byte("key"), k, []byte("bar"))
 | |
| 	}
 | |
| 	tx.Unlock()
 | |
| 
 | |
| 	seq := ""
 | |
| 	getSeq := func(k, v []byte) error {
 | |
| 		seq += string(k)
 | |
| 		return nil
 | |
| 	}
 | |
| 	rtx := b.ReadTx()
 | |
| 	rtx.Lock()
 | |
| 	rtx.UnsafeForEach([]byte("key"), getSeq)
 | |
| 	rtx.Unlock()
 | |
| 
 | |
| 	partialSeq := seq
 | |
| 
 | |
| 	seq = ""
 | |
| 	b.ForceCommit()
 | |
| 
 | |
| 	tx.Lock()
 | |
| 	tx.UnsafeForEach([]byte("key"), getSeq)
 | |
| 	tx.Unlock()
 | |
| 
 | |
| 	if seq != partialSeq {
 | |
| 		t.Fatalf("expected %q, got %q", seq, partialSeq)
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func cleanup(b Backend, path string) {
 | |
| 	b.Close()
 | |
| 	os.Remove(path)
 | |
| }
 | 
