mirror of
https://github.com/etcd-io/etcd.git
synced 2024-09-27 06:25:44 +00:00
Record operation from failpoint injection
Signed-off-by: Marek Siarkowicz <siarkowicz@google.com>
This commit is contained in:
parent
0bd7008eea
commit
c4e3b61a1c
@ -12,7 +12,7 @@
|
|||||||
// See the License for the specific language governing permissions and
|
// See the License for the specific language governing permissions and
|
||||||
// limitations under the License.
|
// limitations under the License.
|
||||||
|
|
||||||
package traffic
|
package client
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
@ -33,7 +33,7 @@ import (
|
|||||||
// clientv3.Client) that records all the requests and responses made. Doesn't
|
// clientv3.Client) that records all the requests and responses made. Doesn't
|
||||||
// allow for concurrent requests to confirm to model.AppendableHistory requirements.
|
// allow for concurrent requests to confirm to model.AppendableHistory requirements.
|
||||||
type RecordingClient struct {
|
type RecordingClient struct {
|
||||||
id int
|
ID int
|
||||||
client clientv3.Client
|
client clientv3.Client
|
||||||
// using baseTime time-measuring operation to get monotonic clock reading
|
// using baseTime time-measuring operation to get monotonic clock reading
|
||||||
// see https://github.com/golang/go/blob/master/src/time/time.go#L17
|
// see https://github.com/golang/go/blob/master/src/time/time.go#L17
|
||||||
@ -51,7 +51,7 @@ type TimedWatchEvent struct {
|
|||||||
Time time.Duration
|
Time time.Duration
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewClient(endpoints []string, ids identity.Provider, baseTime time.Time) (*RecordingClient, error) {
|
func NewRecordingClient(endpoints []string, ids identity.Provider, baseTime time.Time) (*RecordingClient, error) {
|
||||||
cc, err := clientv3.New(clientv3.Config{
|
cc, err := clientv3.New(clientv3.Config{
|
||||||
Endpoints: endpoints,
|
Endpoints: endpoints,
|
||||||
Logger: zap.NewNop(),
|
Logger: zap.NewNop(),
|
||||||
@ -62,7 +62,7 @@ func NewClient(endpoints []string, ids identity.Provider, baseTime time.Time) (*
|
|||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
return &RecordingClient{
|
return &RecordingClient{
|
||||||
id: ids.NewClientID(),
|
ID: ids.NewClientID(),
|
||||||
client: *cc,
|
client: *cc,
|
||||||
kvOperations: model.NewAppendableHistory(ids),
|
kvOperations: model.NewAppendableHistory(ids),
|
||||||
baseTime: baseTime,
|
baseTime: baseTime,
|
||||||
@ -75,7 +75,7 @@ func (c *RecordingClient) Close() error {
|
|||||||
|
|
||||||
func (c *RecordingClient) Report() report.ClientReport {
|
func (c *RecordingClient) Report() report.ClientReport {
|
||||||
return report.ClientReport{
|
return report.ClientReport{
|
||||||
ClientID: c.id,
|
ClientID: c.ID,
|
||||||
KeyValue: c.kvOperations.History.Operations(),
|
KeyValue: c.kvOperations.History.Operations(),
|
||||||
Watch: c.watchOperations,
|
Watch: c.watchOperations,
|
||||||
}
|
}
|
||||||
@ -190,6 +190,65 @@ func (c *RecordingClient) Defragment(ctx context.Context) (*clientv3.DefragmentR
|
|||||||
return resp, err
|
return resp, err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (c *RecordingClient) Compact(ctx context.Context, rev int64) (*clientv3.CompactResponse, error) {
|
||||||
|
c.kvMux.Lock()
|
||||||
|
defer c.kvMux.Unlock()
|
||||||
|
resp, err := c.client.Compact(ctx, rev)
|
||||||
|
return resp, err
|
||||||
|
}
|
||||||
|
func (c *RecordingClient) MemberList(ctx context.Context, opts ...clientv3.OpOption) (*clientv3.MemberListResponse, error) {
|
||||||
|
c.kvMux.Lock()
|
||||||
|
defer c.kvMux.Unlock()
|
||||||
|
resp, err := c.client.MemberList(ctx, opts...)
|
||||||
|
return resp, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *RecordingClient) MemberAdd(ctx context.Context, peerAddrs []string) (*clientv3.MemberAddResponse, error) {
|
||||||
|
c.kvMux.Lock()
|
||||||
|
defer c.kvMux.Unlock()
|
||||||
|
resp, err := c.client.MemberAdd(ctx, peerAddrs)
|
||||||
|
return resp, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *RecordingClient) MemberAddAsLearner(ctx context.Context, peerAddrs []string) (*clientv3.MemberAddResponse, error) {
|
||||||
|
c.kvMux.Lock()
|
||||||
|
defer c.kvMux.Unlock()
|
||||||
|
resp, err := c.client.MemberAddAsLearner(ctx, peerAddrs)
|
||||||
|
return resp, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *RecordingClient) MemberRemove(ctx context.Context, id uint64) (*clientv3.MemberRemoveResponse, error) {
|
||||||
|
c.kvMux.Lock()
|
||||||
|
defer c.kvMux.Unlock()
|
||||||
|
resp, err := c.client.MemberRemove(ctx, id)
|
||||||
|
return resp, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *RecordingClient) MemberUpdate(ctx context.Context, id uint64, peerAddrs []string) (*clientv3.MemberUpdateResponse, error) {
|
||||||
|
c.kvMux.Lock()
|
||||||
|
defer c.kvMux.Unlock()
|
||||||
|
resp, err := c.client.MemberUpdate(ctx, id, peerAddrs)
|
||||||
|
return resp, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *RecordingClient) MemberPromote(ctx context.Context, id uint64) (*clientv3.MemberPromoteResponse, error) {
|
||||||
|
c.kvMux.Lock()
|
||||||
|
defer c.kvMux.Unlock()
|
||||||
|
resp, err := c.client.MemberPromote(ctx, id)
|
||||||
|
return resp, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *RecordingClient) Status(ctx context.Context, endpoint string) (*clientv3.StatusResponse, error) {
|
||||||
|
c.kvMux.Lock()
|
||||||
|
defer c.kvMux.Unlock()
|
||||||
|
resp, err := c.client.Status(ctx, endpoint)
|
||||||
|
return resp, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *RecordingClient) Endpoints() []string {
|
||||||
|
return c.client.Endpoints()
|
||||||
|
}
|
||||||
|
|
||||||
func (c *RecordingClient) Watch(ctx context.Context, key string, rev int64, withPrefix bool, withProgressNotify bool, withPrevKV bool) clientv3.WatchChan {
|
func (c *RecordingClient) Watch(ctx context.Context, key string, rev int64, withPrefix bool, withProgressNotify bool, withPrevKV bool) clientv3.WatchChan {
|
||||||
request := model.WatchRequest{
|
request := model.WatchRequest{
|
||||||
Key: key,
|
Key: key,
|
@ -28,6 +28,9 @@ import (
|
|||||||
clientv3 "go.etcd.io/etcd/client/v3"
|
clientv3 "go.etcd.io/etcd/client/v3"
|
||||||
"go.etcd.io/etcd/server/v3/etcdserver"
|
"go.etcd.io/etcd/server/v3/etcdserver"
|
||||||
"go.etcd.io/etcd/tests/v3/framework/e2e"
|
"go.etcd.io/etcd/tests/v3/framework/e2e"
|
||||||
|
"go.etcd.io/etcd/tests/v3/robustness/client"
|
||||||
|
"go.etcd.io/etcd/tests/v3/robustness/identity"
|
||||||
|
"go.etcd.io/etcd/tests/v3/robustness/report"
|
||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
@ -36,26 +39,21 @@ var (
|
|||||||
|
|
||||||
type memberReplace struct{}
|
type memberReplace struct{}
|
||||||
|
|
||||||
func (f memberReplace) Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster) error {
|
func (f memberReplace) Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster, baseTime time.Time, ids identity.Provider) ([]report.ClientReport, error) {
|
||||||
memberID := uint64(rand.Int() % len(clus.Procs))
|
memberID := uint64(rand.Int() % len(clus.Procs))
|
||||||
member := clus.Procs[memberID]
|
member := clus.Procs[memberID]
|
||||||
var endpoints []string
|
var endpoints []string
|
||||||
for i := 1; i < len(clus.Procs); i++ {
|
for i := 1; i < len(clus.Procs); i++ {
|
||||||
endpoints = append(endpoints, clus.Procs[(int(memberID)+i)%len(clus.Procs)].EndpointsGRPC()...)
|
endpoints = append(endpoints, clus.Procs[(int(memberID)+i)%len(clus.Procs)].EndpointsGRPC()...)
|
||||||
}
|
}
|
||||||
cc, err := clientv3.New(clientv3.Config{
|
cc, err := client.NewRecordingClient(endpoints, ids, baseTime)
|
||||||
Endpoints: endpoints,
|
|
||||||
Logger: zap.NewNop(),
|
|
||||||
DialKeepAliveTime: 50 * time.Second,
|
|
||||||
DialKeepAliveTimeout: 100 * time.Millisecond,
|
|
||||||
})
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return nil, err
|
||||||
}
|
}
|
||||||
defer cc.Close()
|
defer cc.Close()
|
||||||
memberID, found, err := getID(ctx, cc, member.Config().Name)
|
memberID, found, err := getID(ctx, cc, member.Config().Name)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return nil, err
|
||||||
}
|
}
|
||||||
if !found {
|
if !found {
|
||||||
t.Fatal("Member not found")
|
t.Fatal("Member not found")
|
||||||
@ -65,11 +63,11 @@ func (f memberReplace) Inject(ctx context.Context, t *testing.T, lg *zap.Logger,
|
|||||||
lg.Info("Removing member", zap.String("member", member.Config().Name))
|
lg.Info("Removing member", zap.String("member", member.Config().Name))
|
||||||
_, err = cc.MemberRemove(ctx, memberID)
|
_, err = cc.MemberRemove(ctx, memberID)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return nil, err
|
||||||
}
|
}
|
||||||
_, found, err = getID(ctx, cc, member.Config().Name)
|
_, found, err = getID(ctx, cc, member.Config().Name)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return nil, err
|
||||||
}
|
}
|
||||||
if found {
|
if found {
|
||||||
t.Fatal("Expected member to be removed")
|
t.Fatal("Expected member to be removed")
|
||||||
@ -83,13 +81,13 @@ func (f memberReplace) Inject(ctx context.Context, t *testing.T, lg *zap.Logger,
|
|||||||
err = member.Wait(ctx)
|
err = member.Wait(ctx)
|
||||||
if err != nil && !strings.Contains(err.Error(), "unexpected exit code") {
|
if err != nil && !strings.Contains(err.Error(), "unexpected exit code") {
|
||||||
lg.Info("Failed to kill the process", zap.Error(err))
|
lg.Info("Failed to kill the process", zap.Error(err))
|
||||||
return fmt.Errorf("failed to kill the process within %s, err: %w", triggerTimeout, err)
|
return nil, fmt.Errorf("failed to kill the process within %s, err: %w", triggerTimeout, err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
lg.Info("Removing member data", zap.String("member", member.Config().Name))
|
lg.Info("Removing member data", zap.String("member", member.Config().Name))
|
||||||
err = os.RemoveAll(member.Config().DataDirPath)
|
err = os.RemoveAll(member.Config().DataDirPath)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
lg.Info("Adding member back", zap.String("member", member.Config().Name))
|
lg.Info("Adding member back", zap.String("member", member.Config().Name))
|
||||||
@ -97,7 +95,7 @@ func (f memberReplace) Inject(ctx context.Context, t *testing.T, lg *zap.Logger,
|
|||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case <-ctx.Done():
|
case <-ctx.Done():
|
||||||
return ctx.Err()
|
return nil, ctx.Err()
|
||||||
default:
|
default:
|
||||||
}
|
}
|
||||||
reqCtx, cancel := context.WithTimeout(ctx, time.Second)
|
reqCtx, cancel := context.WithTimeout(ctx, time.Second)
|
||||||
@ -109,17 +107,17 @@ func (f memberReplace) Inject(ctx context.Context, t *testing.T, lg *zap.Logger,
|
|||||||
}
|
}
|
||||||
err = patchArgs(member.Config().Args, "initial-cluster-state", "existing")
|
err = patchArgs(member.Config().Args, "initial-cluster-state", "existing")
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return nil, err
|
||||||
}
|
}
|
||||||
lg.Info("Starting member", zap.String("member", member.Config().Name))
|
lg.Info("Starting member", zap.String("member", member.Config().Name))
|
||||||
err = member.Start(ctx)
|
err = member.Start(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return nil, err
|
||||||
}
|
}
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case <-ctx.Done():
|
case <-ctx.Done():
|
||||||
return ctx.Err()
|
return nil, ctx.Err()
|
||||||
default:
|
default:
|
||||||
}
|
}
|
||||||
_, found, err := getID(ctx, cc, member.Config().Name)
|
_, found, err := getID(ctx, cc, member.Config().Name)
|
||||||
@ -130,7 +128,7 @@ func (f memberReplace) Inject(ctx context.Context, t *testing.T, lg *zap.Logger,
|
|||||||
break
|
break
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return nil
|
return nil, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (f memberReplace) Name() string {
|
func (f memberReplace) Name() string {
|
||||||
@ -141,7 +139,7 @@ func (f memberReplace) Available(config e2e.EtcdProcessClusterConfig, _ e2e.Etcd
|
|||||||
return config.ClusterSize > 1
|
return config.ClusterSize > 1
|
||||||
}
|
}
|
||||||
|
|
||||||
func getID(ctx context.Context, cc *clientv3.Client, name string) (id uint64, found bool, err error) {
|
func getID(ctx context.Context, cc clientv3.Cluster, name string) (id uint64, found bool, err error) {
|
||||||
resp, err := cc.MemberList(ctx)
|
resp, err := cc.MemberList(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, false, err
|
return 0, false, err
|
||||||
|
@ -26,6 +26,8 @@ import (
|
|||||||
|
|
||||||
clientv3 "go.etcd.io/etcd/client/v3"
|
clientv3 "go.etcd.io/etcd/client/v3"
|
||||||
"go.etcd.io/etcd/tests/v3/framework/e2e"
|
"go.etcd.io/etcd/tests/v3/framework/e2e"
|
||||||
|
"go.etcd.io/etcd/tests/v3/robustness/identity"
|
||||||
|
"go.etcd.io/etcd/tests/v3/robustness/report"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
@ -75,7 +77,7 @@ func Validate(clus *e2e.EtcdProcessCluster, failpoint Failpoint) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster, failpoint Failpoint, baseTime time.Time) (*InjectionReport, error) {
|
func Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster, failpoint Failpoint, baseTime time.Time, ids identity.Provider) (*FailpointReport, error) {
|
||||||
ctx, cancel := context.WithTimeout(ctx, triggerTimeout)
|
ctx, cancel := context.WithTimeout(ctx, triggerTimeout)
|
||||||
defer cancel()
|
defer cancel()
|
||||||
var err error
|
var err error
|
||||||
@ -85,7 +87,7 @@ func Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdPro
|
|||||||
}
|
}
|
||||||
lg.Info("Triggering failpoint", zap.String("failpoint", failpoint.Name()))
|
lg.Info("Triggering failpoint", zap.String("failpoint", failpoint.Name()))
|
||||||
start := time.Since(baseTime)
|
start := time.Since(baseTime)
|
||||||
err = failpoint.Inject(ctx, t, lg, clus)
|
clientReport, err := failpoint.Inject(ctx, t, lg, clus, baseTime, ids)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
lg.Error("Failed to trigger failpoint", zap.String("failpoint", failpoint.Name()), zap.Error(err))
|
lg.Error("Failed to trigger failpoint", zap.String("failpoint", failpoint.Name()), zap.Error(err))
|
||||||
return nil, fmt.Errorf("failed triggering failpoint, err: %v", err)
|
return nil, fmt.Errorf("failed triggering failpoint, err: %v", err)
|
||||||
@ -96,14 +98,22 @@ func Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdPro
|
|||||||
lg.Info("Finished triggering failpoint", zap.String("failpoint", failpoint.Name()))
|
lg.Info("Finished triggering failpoint", zap.String("failpoint", failpoint.Name()))
|
||||||
end := time.Since(baseTime)
|
end := time.Since(baseTime)
|
||||||
|
|
||||||
return &InjectionReport{
|
return &FailpointReport{
|
||||||
Start: start,
|
Injection: Injection{
|
||||||
End: end,
|
Start: start,
|
||||||
Name: failpoint.Name(),
|
End: end,
|
||||||
|
Name: failpoint.Name(),
|
||||||
|
},
|
||||||
|
Client: clientReport,
|
||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
type InjectionReport struct {
|
type FailpointReport struct {
|
||||||
|
Injection
|
||||||
|
Client []report.ClientReport
|
||||||
|
}
|
||||||
|
|
||||||
|
type Injection struct {
|
||||||
Start, End time.Duration
|
Start, End time.Duration
|
||||||
Name string
|
Name string
|
||||||
}
|
}
|
||||||
@ -137,7 +147,7 @@ func verifyClusterHealth(ctx context.Context, _ *testing.T, clus *e2e.EtcdProces
|
|||||||
}
|
}
|
||||||
|
|
||||||
type Failpoint interface {
|
type Failpoint interface {
|
||||||
Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster) error
|
Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster, baseTime time.Time, ids identity.Provider) ([]report.ClientReport, error)
|
||||||
Name() string
|
Name() string
|
||||||
AvailabilityChecker
|
AvailabilityChecker
|
||||||
}
|
}
|
||||||
|
@ -25,6 +25,8 @@ import (
|
|||||||
"go.uber.org/zap"
|
"go.uber.org/zap"
|
||||||
|
|
||||||
"go.etcd.io/etcd/tests/v3/framework/e2e"
|
"go.etcd.io/etcd/tests/v3/framework/e2e"
|
||||||
|
"go.etcd.io/etcd/tests/v3/robustness/identity"
|
||||||
|
"go.etcd.io/etcd/tests/v3/robustness/report"
|
||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
@ -73,17 +75,17 @@ const (
|
|||||||
Follower failpointTarget = "Follower"
|
Follower failpointTarget = "Follower"
|
||||||
)
|
)
|
||||||
|
|
||||||
func (f goPanicFailpoint) Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster) error {
|
func (f goPanicFailpoint) Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster, baseTime time.Time, ids identity.Provider) (reports []report.ClientReport, err error) {
|
||||||
member := f.pickMember(t, clus)
|
member := f.pickMember(t, clus)
|
||||||
|
|
||||||
for member.IsRunning() {
|
for member.IsRunning() {
|
||||||
select {
|
select {
|
||||||
case <-ctx.Done():
|
case <-ctx.Done():
|
||||||
return ctx.Err()
|
return reports, ctx.Err()
|
||||||
default:
|
default:
|
||||||
}
|
}
|
||||||
lg.Info("Setting up gofailpoint", zap.String("failpoint", f.Name()))
|
lg.Info("Setting up gofailpoint", zap.String("failpoint", f.Name()))
|
||||||
err := member.Failpoints().SetupHTTP(ctx, f.failpoint, "panic")
|
err = member.Failpoints().SetupHTTP(ctx, f.failpoint, "panic")
|
||||||
if err != nil {
|
if err != nil {
|
||||||
lg.Info("goFailpoint setup failed", zap.String("failpoint", f.Name()), zap.Error(err))
|
lg.Info("goFailpoint setup failed", zap.String("failpoint", f.Name()), zap.Error(err))
|
||||||
continue
|
continue
|
||||||
@ -93,17 +95,21 @@ func (f goPanicFailpoint) Inject(ctx context.Context, t *testing.T, lg *zap.Logg
|
|||||||
break
|
break
|
||||||
}
|
}
|
||||||
if f.trigger != nil {
|
if f.trigger != nil {
|
||||||
|
var r []report.ClientReport
|
||||||
lg.Info("Triggering gofailpoint", zap.String("failpoint", f.Name()))
|
lg.Info("Triggering gofailpoint", zap.String("failpoint", f.Name()))
|
||||||
err = f.trigger.Trigger(ctx, t, member, clus)
|
r, err = f.trigger.Trigger(ctx, t, member, clus, baseTime, ids)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
lg.Info("gofailpoint trigger failed", zap.String("failpoint", f.Name()), zap.Error(err))
|
lg.Info("gofailpoint trigger failed", zap.String("failpoint", f.Name()), zap.Error(err))
|
||||||
}
|
}
|
||||||
|
if r != nil {
|
||||||
|
reports = append(reports, r...)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
lg.Info("Waiting for member to exit", zap.String("member", member.Config().Name))
|
lg.Info("Waiting for member to exit", zap.String("member", member.Config().Name))
|
||||||
err = member.Wait(ctx)
|
err = member.Wait(ctx)
|
||||||
if err != nil && !strings.Contains(err.Error(), "unexpected exit code") {
|
if err != nil && !strings.Contains(err.Error(), "unexpected exit code") {
|
||||||
lg.Info("Member didn't exit as expected", zap.String("member", member.Config().Name), zap.Error(err))
|
lg.Info("Member didn't exit as expected", zap.String("member", member.Config().Name), zap.Error(err))
|
||||||
return fmt.Errorf("member didn't exit as expected: %v", err)
|
return reports, fmt.Errorf("member didn't exit as expected: %v", err)
|
||||||
}
|
}
|
||||||
lg.Info("Member exited as expected", zap.String("member", member.Config().Name))
|
lg.Info("Member exited as expected", zap.String("member", member.Config().Name))
|
||||||
}
|
}
|
||||||
@ -112,11 +118,11 @@ func (f goPanicFailpoint) Inject(ctx context.Context, t *testing.T, lg *zap.Logg
|
|||||||
lg.Info("Removing data that was not fsynced")
|
lg.Info("Removing data that was not fsynced")
|
||||||
err := lazyfs.ClearCache(ctx)
|
err := lazyfs.ClearCache(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return reports, err
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return member.Start(ctx)
|
return reports, member.Start(ctx)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (f goPanicFailpoint) pickMember(t *testing.T, clus *e2e.EtcdProcessCluster) e2e.EtcdProcess {
|
func (f goPanicFailpoint) pickMember(t *testing.T, clus *e2e.EtcdProcessCluster) e2e.EtcdProcess {
|
||||||
@ -155,7 +161,7 @@ type killAndGofailSleep struct {
|
|||||||
time time.Duration
|
time time.Duration
|
||||||
}
|
}
|
||||||
|
|
||||||
func (f killAndGofailSleep) Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster) error {
|
func (f killAndGofailSleep) Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster, baseTime time.Time, ids identity.Provider) ([]report.ClientReport, error) {
|
||||||
member := clus.Procs[rand.Int()%len(clus.Procs)]
|
member := clus.Procs[rand.Int()%len(clus.Procs)]
|
||||||
for member.IsRunning() {
|
for member.IsRunning() {
|
||||||
err := member.Kill()
|
err := member.Kill()
|
||||||
@ -165,20 +171,20 @@ func (f killAndGofailSleep) Inject(ctx context.Context, t *testing.T, lg *zap.Lo
|
|||||||
err = member.Wait(ctx)
|
err = member.Wait(ctx)
|
||||||
if err != nil && !strings.Contains(err.Error(), "unexpected exit code") {
|
if err != nil && !strings.Contains(err.Error(), "unexpected exit code") {
|
||||||
lg.Info("Failed to kill the process", zap.Error(err))
|
lg.Info("Failed to kill the process", zap.Error(err))
|
||||||
return fmt.Errorf("failed to kill the process within %s, err: %w", triggerTimeout, err)
|
return nil, fmt.Errorf("failed to kill the process within %s, err: %w", triggerTimeout, err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
lg.Info("Setting up goFailpoint", zap.String("failpoint", f.Name()))
|
lg.Info("Setting up goFailpoint", zap.String("failpoint", f.Name()))
|
||||||
err := member.Failpoints().SetupEnv(f.failpoint, fmt.Sprintf(`sleep(%q)`, f.time))
|
err := member.Failpoints().SetupEnv(f.failpoint, fmt.Sprintf(`sleep(%q)`, f.time))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return nil, err
|
||||||
}
|
}
|
||||||
err = member.Start(ctx)
|
err = member.Start(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return nil, err
|
||||||
}
|
}
|
||||||
// TODO: Check gofail status (https://github.com/etcd-io/gofail/pull/47) and wait for sleep to beis executed at least once.
|
// TODO: Check gofail status (https://github.com/etcd-io/gofail/pull/47) and wait for sleep to beis executed at least once.
|
||||||
return nil
|
return nil, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (f killAndGofailSleep) Name() string {
|
func (f killAndGofailSleep) Name() string {
|
||||||
@ -201,22 +207,22 @@ type gofailSleepAndDeactivate struct {
|
|||||||
time time.Duration
|
time time.Duration
|
||||||
}
|
}
|
||||||
|
|
||||||
func (f gofailSleepAndDeactivate) Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster) error {
|
func (f gofailSleepAndDeactivate) Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster, baseTime time.Time, ids identity.Provider) ([]report.ClientReport, error) {
|
||||||
member := clus.Procs[rand.Int()%len(clus.Procs)]
|
member := clus.Procs[rand.Int()%len(clus.Procs)]
|
||||||
lg.Info("Setting up gofailpoint", zap.String("failpoint", f.Name()))
|
lg.Info("Setting up gofailpoint", zap.String("failpoint", f.Name()))
|
||||||
err := member.Failpoints().SetupHTTP(ctx, f.failpoint, fmt.Sprintf(`sleep(%q)`, f.time))
|
err := member.Failpoints().SetupHTTP(ctx, f.failpoint, fmt.Sprintf(`sleep(%q)`, f.time))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
lg.Info("goFailpoint setup failed", zap.String("failpoint", f.Name()), zap.Error(err))
|
lg.Info("goFailpoint setup failed", zap.String("failpoint", f.Name()), zap.Error(err))
|
||||||
return fmt.Errorf("goFailpoint %s setup failed, err:%w", f.Name(), err)
|
return nil, fmt.Errorf("goFailpoint %s setup failed, err:%w", f.Name(), err)
|
||||||
}
|
}
|
||||||
time.Sleep(f.time)
|
time.Sleep(f.time)
|
||||||
lg.Info("Deactivating gofailpoint", zap.String("failpoint", f.Name()))
|
lg.Info("Deactivating gofailpoint", zap.String("failpoint", f.Name()))
|
||||||
err = member.Failpoints().DeactivateHTTP(ctx, f.failpoint)
|
err = member.Failpoints().DeactivateHTTP(ctx, f.failpoint)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
lg.Info("goFailpoint deactivate failed", zap.String("failpoint", f.Name()), zap.Error(err))
|
lg.Info("goFailpoint deactivate failed", zap.String("failpoint", f.Name()), zap.Error(err))
|
||||||
return fmt.Errorf("goFailpoint %s deactivate failed, err: %w", f.Name(), err)
|
return nil, fmt.Errorf("goFailpoint %s deactivate failed, err: %w", f.Name(), err)
|
||||||
}
|
}
|
||||||
return nil
|
return nil, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (f gofailSleepAndDeactivate) Name() string {
|
func (f gofailSleepAndDeactivate) Name() string {
|
||||||
|
@ -20,10 +20,13 @@ import (
|
|||||||
"math/rand"
|
"math/rand"
|
||||||
"strings"
|
"strings"
|
||||||
"testing"
|
"testing"
|
||||||
|
"time"
|
||||||
|
|
||||||
"go.uber.org/zap"
|
"go.uber.org/zap"
|
||||||
|
|
||||||
"go.etcd.io/etcd/tests/v3/framework/e2e"
|
"go.etcd.io/etcd/tests/v3/framework/e2e"
|
||||||
|
"go.etcd.io/etcd/tests/v3/robustness/identity"
|
||||||
|
"go.etcd.io/etcd/tests/v3/robustness/report"
|
||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
@ -32,7 +35,7 @@ var (
|
|||||||
|
|
||||||
type killFailpoint struct{}
|
type killFailpoint struct{}
|
||||||
|
|
||||||
func (f killFailpoint) Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster) error {
|
func (f killFailpoint) Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster, baseTime time.Time, ids identity.Provider) ([]report.ClientReport, error) {
|
||||||
member := clus.Procs[rand.Int()%len(clus.Procs)]
|
member := clus.Procs[rand.Int()%len(clus.Procs)]
|
||||||
|
|
||||||
for member.IsRunning() {
|
for member.IsRunning() {
|
||||||
@ -43,21 +46,21 @@ func (f killFailpoint) Inject(ctx context.Context, t *testing.T, lg *zap.Logger,
|
|||||||
err = member.Wait(ctx)
|
err = member.Wait(ctx)
|
||||||
if err != nil && !strings.Contains(err.Error(), "unexpected exit code") {
|
if err != nil && !strings.Contains(err.Error(), "unexpected exit code") {
|
||||||
lg.Info("Failed to kill the process", zap.Error(err))
|
lg.Info("Failed to kill the process", zap.Error(err))
|
||||||
return fmt.Errorf("failed to kill the process within %s, err: %w", triggerTimeout, err)
|
return nil, fmt.Errorf("failed to kill the process within %s, err: %w", triggerTimeout, err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if lazyfs := member.LazyFS(); lazyfs != nil {
|
if lazyfs := member.LazyFS(); lazyfs != nil {
|
||||||
lg.Info("Removing data that was not fsynced")
|
lg.Info("Removing data that was not fsynced")
|
||||||
err := lazyfs.ClearCache(ctx)
|
err := lazyfs.ClearCache(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return nil, err
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
err := member.Start(ctx)
|
err := member.Start(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return nil, err
|
||||||
}
|
}
|
||||||
return nil
|
return nil, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (f killFailpoint) Name() string {
|
func (f killFailpoint) Name() string {
|
||||||
|
@ -24,6 +24,8 @@ import (
|
|||||||
|
|
||||||
clientv3 "go.etcd.io/etcd/client/v3"
|
clientv3 "go.etcd.io/etcd/client/v3"
|
||||||
"go.etcd.io/etcd/tests/v3/framework/e2e"
|
"go.etcd.io/etcd/tests/v3/framework/e2e"
|
||||||
|
"go.etcd.io/etcd/tests/v3/robustness/identity"
|
||||||
|
"go.etcd.io/etcd/tests/v3/robustness/report"
|
||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
@ -37,9 +39,9 @@ type blackholePeerNetworkFailpoint struct {
|
|||||||
triggerBlackhole
|
triggerBlackhole
|
||||||
}
|
}
|
||||||
|
|
||||||
func (f blackholePeerNetworkFailpoint) Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster) error {
|
func (f blackholePeerNetworkFailpoint) Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster, baseTime time.Time, ids identity.Provider) ([]report.ClientReport, error) {
|
||||||
member := clus.Procs[rand.Int()%len(clus.Procs)]
|
member := clus.Procs[rand.Int()%len(clus.Procs)]
|
||||||
return f.Trigger(ctx, t, member, clus)
|
return f.Trigger(ctx, t, member, clus, baseTime, ids)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (f blackholePeerNetworkFailpoint) Name() string {
|
func (f blackholePeerNetworkFailpoint) Name() string {
|
||||||
@ -50,8 +52,8 @@ type triggerBlackhole struct {
|
|||||||
waitTillSnapshot bool
|
waitTillSnapshot bool
|
||||||
}
|
}
|
||||||
|
|
||||||
func (tb triggerBlackhole) Trigger(ctx context.Context, t *testing.T, member e2e.EtcdProcess, clus *e2e.EtcdProcessCluster) error {
|
func (tb triggerBlackhole) Trigger(ctx context.Context, t *testing.T, member e2e.EtcdProcess, clus *e2e.EtcdProcessCluster, baseTime time.Time, ids identity.Provider) ([]report.ClientReport, error) {
|
||||||
return Blackhole(ctx, t, member, clus, tb.waitTillSnapshot)
|
return nil, Blackhole(ctx, t, member, clus, tb.waitTillSnapshot)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (tb triggerBlackhole) Available(config e2e.EtcdProcessClusterConfig, process e2e.EtcdProcess) bool {
|
func (tb triggerBlackhole) Available(config e2e.EtcdProcessClusterConfig, process e2e.EtcdProcess) bool {
|
||||||
@ -153,7 +155,7 @@ type delayPeerNetworkFailpoint struct {
|
|||||||
randomizedLatency time.Duration
|
randomizedLatency time.Duration
|
||||||
}
|
}
|
||||||
|
|
||||||
func (f delayPeerNetworkFailpoint) Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster) error {
|
func (f delayPeerNetworkFailpoint) Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster, baseTime time.Time, ids identity.Provider) ([]report.ClientReport, error) {
|
||||||
member := clus.Procs[rand.Int()%len(clus.Procs)]
|
member := clus.Procs[rand.Int()%len(clus.Procs)]
|
||||||
proxy := member.PeerProxy()
|
proxy := member.PeerProxy()
|
||||||
|
|
||||||
@ -164,7 +166,7 @@ func (f delayPeerNetworkFailpoint) Inject(ctx context.Context, t *testing.T, lg
|
|||||||
lg.Info("Traffic delay removed", zap.String("member", member.Config().Name))
|
lg.Info("Traffic delay removed", zap.String("member", member.Config().Name))
|
||||||
proxy.UndelayRx()
|
proxy.UndelayRx()
|
||||||
proxy.UndelayTx()
|
proxy.UndelayTx()
|
||||||
return nil
|
return nil, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (f delayPeerNetworkFailpoint) Name() string {
|
func (f delayPeerNetworkFailpoint) Name() string {
|
||||||
@ -180,7 +182,7 @@ type dropPeerNetworkFailpoint struct {
|
|||||||
dropProbabilityPercent int
|
dropProbabilityPercent int
|
||||||
}
|
}
|
||||||
|
|
||||||
func (f dropPeerNetworkFailpoint) Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster) error {
|
func (f dropPeerNetworkFailpoint) Inject(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster, baseTime time.Time, ids identity.Provider) ([]report.ClientReport, error) {
|
||||||
member := clus.Procs[rand.Int()%len(clus.Procs)]
|
member := clus.Procs[rand.Int()%len(clus.Procs)]
|
||||||
proxy := member.PeerProxy()
|
proxy := member.PeerProxy()
|
||||||
|
|
||||||
@ -191,7 +193,7 @@ func (f dropPeerNetworkFailpoint) Inject(ctx context.Context, t *testing.T, lg *
|
|||||||
lg.Info("Traffic drop removed", zap.String("member", member.Config().Name))
|
lg.Info("Traffic drop removed", zap.String("member", member.Config().Name))
|
||||||
proxy.UnmodifyRx()
|
proxy.UnmodifyRx()
|
||||||
proxy.UnmodifyTx()
|
proxy.UnmodifyTx()
|
||||||
return nil
|
return nil, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (f dropPeerNetworkFailpoint) modifyPacket(data []byte) []byte {
|
func (f dropPeerNetworkFailpoint) modifyPacket(data []byte) []byte {
|
||||||
|
@ -21,35 +21,30 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"go.uber.org/zap"
|
|
||||||
|
|
||||||
clientv3 "go.etcd.io/etcd/client/v3"
|
|
||||||
"go.etcd.io/etcd/tests/v3/framework/e2e"
|
"go.etcd.io/etcd/tests/v3/framework/e2e"
|
||||||
|
"go.etcd.io/etcd/tests/v3/robustness/client"
|
||||||
|
"go.etcd.io/etcd/tests/v3/robustness/identity"
|
||||||
|
"go.etcd.io/etcd/tests/v3/robustness/report"
|
||||||
)
|
)
|
||||||
|
|
||||||
type trigger interface {
|
type trigger interface {
|
||||||
Trigger(ctx context.Context, t *testing.T, member e2e.EtcdProcess, clus *e2e.EtcdProcessCluster) error
|
Trigger(ctx context.Context, t *testing.T, member e2e.EtcdProcess, clus *e2e.EtcdProcessCluster, baseTime time.Time, ids identity.Provider) ([]report.ClientReport, error)
|
||||||
AvailabilityChecker
|
AvailabilityChecker
|
||||||
}
|
}
|
||||||
|
|
||||||
type triggerDefrag struct{}
|
type triggerDefrag struct{}
|
||||||
|
|
||||||
func (t triggerDefrag) Trigger(ctx context.Context, _ *testing.T, member e2e.EtcdProcess, _ *e2e.EtcdProcessCluster) error {
|
func (t triggerDefrag) Trigger(ctx context.Context, _ *testing.T, member e2e.EtcdProcess, clus *e2e.EtcdProcessCluster, baseTime time.Time, ids identity.Provider) ([]report.ClientReport, error) {
|
||||||
cc, err := clientv3.New(clientv3.Config{
|
cc, err := client.NewRecordingClient(member.EndpointsGRPC(), ids, baseTime)
|
||||||
Endpoints: member.EndpointsGRPC(),
|
|
||||||
Logger: zap.NewNop(),
|
|
||||||
DialKeepAliveTime: 10 * time.Second,
|
|
||||||
DialKeepAliveTimeout: 100 * time.Millisecond,
|
|
||||||
})
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed creating client: %w", err)
|
return nil, fmt.Errorf("failed creating client: %w", err)
|
||||||
}
|
}
|
||||||
defer cc.Close()
|
defer cc.Close()
|
||||||
_, err = cc.Defragment(ctx, member.EndpointsGRPC()[0])
|
_, err = cc.Defragment(ctx)
|
||||||
if err != nil && !strings.Contains(err.Error(), "error reading from server: EOF") {
|
if err != nil && !strings.Contains(err.Error(), "error reading from server: EOF") {
|
||||||
return err
|
return nil, err
|
||||||
}
|
}
|
||||||
return nil
|
return nil, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (t triggerDefrag) Available(e2e.EtcdProcessClusterConfig, e2e.EtcdProcess) bool {
|
func (t triggerDefrag) Available(e2e.EtcdProcessClusterConfig, e2e.EtcdProcess) bool {
|
||||||
@ -60,37 +55,32 @@ type triggerCompact struct {
|
|||||||
multiBatchCompaction bool
|
multiBatchCompaction bool
|
||||||
}
|
}
|
||||||
|
|
||||||
func (t triggerCompact) Trigger(ctx context.Context, _ *testing.T, member e2e.EtcdProcess, clus *e2e.EtcdProcessCluster) error {
|
func (t triggerCompact) Trigger(ctx context.Context, _ *testing.T, member e2e.EtcdProcess, clus *e2e.EtcdProcessCluster, baseTime time.Time, ids identity.Provider) ([]report.ClientReport, error) {
|
||||||
cc, err := clientv3.New(clientv3.Config{
|
ctx, cancel := context.WithTimeout(ctx, time.Second)
|
||||||
Endpoints: member.EndpointsGRPC(),
|
defer cancel()
|
||||||
Logger: zap.NewNop(),
|
cc, err := client.NewRecordingClient(member.EndpointsGRPC(), ids, baseTime)
|
||||||
DialKeepAliveTime: 10 * time.Second,
|
|
||||||
DialKeepAliveTimeout: 100 * time.Millisecond,
|
|
||||||
})
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed creating client: %w", err)
|
return nil, fmt.Errorf("failed creating client: %w", err)
|
||||||
}
|
}
|
||||||
defer cc.Close()
|
defer cc.Close()
|
||||||
|
|
||||||
var rev int64
|
var rev int64
|
||||||
for {
|
for {
|
||||||
resp, gerr := cc.Get(ctx, "/")
|
_, rev, err = cc.Get(ctx, "/", 0)
|
||||||
if gerr != nil {
|
if err != nil {
|
||||||
return gerr
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
rev = resp.Header.Revision
|
|
||||||
if !t.multiBatchCompaction || rev > int64(clus.Cfg.ServerConfig.ExperimentalCompactionBatchLimit) {
|
if !t.multiBatchCompaction || rev > int64(clus.Cfg.ServerConfig.ExperimentalCompactionBatchLimit) {
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
time.Sleep(50 * time.Millisecond)
|
time.Sleep(50 * time.Millisecond)
|
||||||
}
|
}
|
||||||
|
|
||||||
_, err = cc.Compact(ctx, rev)
|
_, err = cc.Compact(ctx, rev)
|
||||||
if err != nil && !strings.Contains(err.Error(), "error reading from server: EOF") {
|
if err != nil && !strings.Contains(err.Error(), "error reading from server: EOF") {
|
||||||
return err
|
return nil, err
|
||||||
}
|
}
|
||||||
return nil
|
return []report.ClientReport{cc.Report()}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (t triggerCompact) Available(e2e.EtcdProcessClusterConfig, e2e.EtcdProcess) bool {
|
func (t triggerCompact) Available(e2e.EtcdProcessClusterConfig, e2e.EtcdProcess) bool {
|
||||||
|
@ -108,8 +108,8 @@ func (s testScenario) run(ctx context.Context, t *testing.T, lg *zap.Logger, clu
|
|||||||
ctx, cancel := context.WithCancel(ctx)
|
ctx, cancel := context.WithCancel(ctx)
|
||||||
defer cancel()
|
defer cancel()
|
||||||
g := errgroup.Group{}
|
g := errgroup.Group{}
|
||||||
var operationReport, watchReport []report.ClientReport
|
var operationReport, watchReport, failpointClientReport []report.ClientReport
|
||||||
failpointInjected := make(chan failpoint.InjectionReport, 1)
|
failpointInjected := make(chan failpoint.Injection, 1)
|
||||||
|
|
||||||
// using baseTime time-measuring operation to get monotonic clock reading
|
// using baseTime time-measuring operation to get monotonic clock reading
|
||||||
// see https://github.com/golang/go/blob/master/src/time/time.go#L17
|
// see https://github.com/golang/go/blob/master/src/time/time.go#L17
|
||||||
@ -119,7 +119,7 @@ func (s testScenario) run(ctx context.Context, t *testing.T, lg *zap.Logger, clu
|
|||||||
defer close(failpointInjected)
|
defer close(failpointInjected)
|
||||||
// Give some time for traffic to reach qps target before injecting failpoint.
|
// Give some time for traffic to reach qps target before injecting failpoint.
|
||||||
time.Sleep(time.Second)
|
time.Sleep(time.Second)
|
||||||
fr, err := failpoint.Inject(ctx, t, lg, clus, s.failpoint, baseTime)
|
fr, err := failpoint.Inject(ctx, t, lg, clus, s.failpoint, baseTime, ids)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Error(err)
|
t.Error(err)
|
||||||
cancel()
|
cancel()
|
||||||
@ -127,7 +127,8 @@ func (s testScenario) run(ctx context.Context, t *testing.T, lg *zap.Logger, clu
|
|||||||
// Give some time for traffic to reach qps target after injecting failpoint.
|
// Give some time for traffic to reach qps target after injecting failpoint.
|
||||||
time.Sleep(time.Second)
|
time.Sleep(time.Second)
|
||||||
if fr != nil {
|
if fr != nil {
|
||||||
failpointInjected <- *fr
|
failpointInjected <- fr.Injection
|
||||||
|
failpointClientReport = fr.Client
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
})
|
})
|
||||||
@ -145,7 +146,7 @@ func (s testScenario) run(ctx context.Context, t *testing.T, lg *zap.Logger, clu
|
|||||||
return nil
|
return nil
|
||||||
})
|
})
|
||||||
g.Wait()
|
g.Wait()
|
||||||
return append(operationReport, watchReport...)
|
return append(operationReport, append(failpointClientReport, watchReport...)...)
|
||||||
}
|
}
|
||||||
|
|
||||||
func operationsMaxRevision(reports []report.ClientReport) int64 {
|
func operationsMaxRevision(reports []report.ClientReport) int64 {
|
||||||
|
@ -23,6 +23,7 @@ import (
|
|||||||
|
|
||||||
"go.etcd.io/etcd/api/v3/mvccpb"
|
"go.etcd.io/etcd/api/v3/mvccpb"
|
||||||
clientv3 "go.etcd.io/etcd/client/v3"
|
clientv3 "go.etcd.io/etcd/client/v3"
|
||||||
|
"go.etcd.io/etcd/tests/v3/robustness/client"
|
||||||
"go.etcd.io/etcd/tests/v3/robustness/identity"
|
"go.etcd.io/etcd/tests/v3/robustness/identity"
|
||||||
"go.etcd.io/etcd/tests/v3/robustness/model"
|
"go.etcd.io/etcd/tests/v3/robustness/model"
|
||||||
)
|
)
|
||||||
@ -94,7 +95,7 @@ func (t etcdTraffic) Name() string {
|
|||||||
return "Etcd"
|
return "Etcd"
|
||||||
}
|
}
|
||||||
|
|
||||||
func (t etcdTraffic) Run(ctx context.Context, c *RecordingClient, limiter *rate.Limiter, ids identity.Provider, lm identity.LeaseIDStorage, nonUniqueWriteLimiter ConcurrencyLimiter, finish <-chan struct{}) {
|
func (t etcdTraffic) Run(ctx context.Context, c *client.RecordingClient, limiter *rate.Limiter, ids identity.Provider, lm identity.LeaseIDStorage, nonUniqueWriteLimiter ConcurrencyLimiter, finish <-chan struct{}) {
|
||||||
lastOperationSucceeded := true
|
lastOperationSucceeded := true
|
||||||
var lastRev int64
|
var lastRev int64
|
||||||
var requestType etcdRequestType
|
var requestType etcdRequestType
|
||||||
@ -151,7 +152,7 @@ func filterOutNonUniqueEtcdWrites(choices []choiceWeight[etcdRequestType]) (resp
|
|||||||
type etcdTrafficClient struct {
|
type etcdTrafficClient struct {
|
||||||
etcdTraffic
|
etcdTraffic
|
||||||
keyPrefix string
|
keyPrefix string
|
||||||
client *RecordingClient
|
client *client.RecordingClient
|
||||||
limiter *rate.Limiter
|
limiter *rate.Limiter
|
||||||
idProvider identity.Provider
|
idProvider identity.Provider
|
||||||
leaseStorage identity.LeaseIDStorage
|
leaseStorage identity.LeaseIDStorage
|
||||||
@ -222,7 +223,7 @@ func (c etcdTrafficClient) Request(ctx context.Context, request etcdRequestType,
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
case PutWithLease:
|
case PutWithLease:
|
||||||
leaseID := c.leaseStorage.LeaseID(c.client.id)
|
leaseID := c.leaseStorage.LeaseID(c.client.ID)
|
||||||
if leaseID == 0 {
|
if leaseID == 0 {
|
||||||
var resp *clientv3.LeaseGrantResponse
|
var resp *clientv3.LeaseGrantResponse
|
||||||
resp, err = c.client.LeaseGrant(opCtx, c.leaseTTL)
|
resp, err = c.client.LeaseGrant(opCtx, c.leaseTTL)
|
||||||
@ -231,7 +232,7 @@ func (c etcdTrafficClient) Request(ctx context.Context, request etcdRequestType,
|
|||||||
rev = resp.ResponseHeader.Revision
|
rev = resp.ResponseHeader.Revision
|
||||||
}
|
}
|
||||||
if err == nil {
|
if err == nil {
|
||||||
c.leaseStorage.AddLeaseID(c.client.id, leaseID)
|
c.leaseStorage.AddLeaseID(c.client.ID, leaseID)
|
||||||
c.limiter.Wait(ctx)
|
c.limiter.Wait(ctx)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -245,13 +246,13 @@ func (c etcdTrafficClient) Request(ctx context.Context, request etcdRequestType,
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
case LeaseRevoke:
|
case LeaseRevoke:
|
||||||
leaseID := c.leaseStorage.LeaseID(c.client.id)
|
leaseID := c.leaseStorage.LeaseID(c.client.ID)
|
||||||
if leaseID != 0 {
|
if leaseID != 0 {
|
||||||
var resp *clientv3.LeaseRevokeResponse
|
var resp *clientv3.LeaseRevokeResponse
|
||||||
resp, err = c.client.LeaseRevoke(opCtx, leaseID)
|
resp, err = c.client.LeaseRevoke(opCtx, leaseID)
|
||||||
//if LeaseRevoke has failed, do not remove the mapping.
|
//if LeaseRevoke has failed, do not remove the mapping.
|
||||||
if err == nil {
|
if err == nil {
|
||||||
c.leaseStorage.RemoveLeaseID(c.client.id)
|
c.leaseStorage.RemoveLeaseID(c.client.ID)
|
||||||
}
|
}
|
||||||
if resp != nil {
|
if resp != nil {
|
||||||
rev = resp.Header.Revision
|
rev = resp.Header.Revision
|
||||||
|
@ -27,6 +27,7 @@ import (
|
|||||||
"go.etcd.io/etcd/api/v3/mvccpb"
|
"go.etcd.io/etcd/api/v3/mvccpb"
|
||||||
clientv3 "go.etcd.io/etcd/client/v3"
|
clientv3 "go.etcd.io/etcd/client/v3"
|
||||||
"go.etcd.io/etcd/pkg/v3/stringutil"
|
"go.etcd.io/etcd/pkg/v3/stringutil"
|
||||||
|
"go.etcd.io/etcd/tests/v3/robustness/client"
|
||||||
"go.etcd.io/etcd/tests/v3/robustness/identity"
|
"go.etcd.io/etcd/tests/v3/robustness/identity"
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -58,7 +59,7 @@ func (t kubernetesTraffic) Name() string {
|
|||||||
return "Kubernetes"
|
return "Kubernetes"
|
||||||
}
|
}
|
||||||
|
|
||||||
func (t kubernetesTraffic) Run(ctx context.Context, c *RecordingClient, limiter *rate.Limiter, ids identity.Provider, lm identity.LeaseIDStorage, nonUniqueWriteLimiter ConcurrencyLimiter, finish <-chan struct{}) {
|
func (t kubernetesTraffic) Run(ctx context.Context, c *client.RecordingClient, limiter *rate.Limiter, ids identity.Provider, lm identity.LeaseIDStorage, nonUniqueWriteLimiter ConcurrencyLimiter, finish <-chan struct{}) {
|
||||||
kc := &kubernetesClient{client: c}
|
kc := &kubernetesClient{client: c}
|
||||||
s := newStorage()
|
s := newStorage()
|
||||||
keyPrefix := "/registry/" + t.resource + "/"
|
keyPrefix := "/registry/" + t.resource + "/"
|
||||||
@ -214,7 +215,7 @@ const (
|
|||||||
)
|
)
|
||||||
|
|
||||||
type kubernetesClient struct {
|
type kubernetesClient struct {
|
||||||
client *RecordingClient
|
client *client.RecordingClient
|
||||||
}
|
}
|
||||||
|
|
||||||
func (k kubernetesClient) List(ctx context.Context, prefix string, revision, limit int64) (*clientv3.GetResponse, error) {
|
func (k kubernetesClient) List(ctx context.Context, prefix string, revision, limit int64) (*clientv3.GetResponse, error) {
|
||||||
|
@ -24,6 +24,7 @@ import (
|
|||||||
"golang.org/x/time/rate"
|
"golang.org/x/time/rate"
|
||||||
|
|
||||||
"go.etcd.io/etcd/tests/v3/framework/e2e"
|
"go.etcd.io/etcd/tests/v3/framework/e2e"
|
||||||
|
"go.etcd.io/etcd/tests/v3/robustness/client"
|
||||||
"go.etcd.io/etcd/tests/v3/robustness/failpoint"
|
"go.etcd.io/etcd/tests/v3/robustness/failpoint"
|
||||||
"go.etcd.io/etcd/tests/v3/robustness/identity"
|
"go.etcd.io/etcd/tests/v3/robustness/identity"
|
||||||
"go.etcd.io/etcd/tests/v3/robustness/model"
|
"go.etcd.io/etcd/tests/v3/robustness/model"
|
||||||
@ -52,7 +53,7 @@ var (
|
|||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
|
||||||
func SimulateTraffic(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster, profile Profile, traffic Traffic, failpointInjected <-chan failpoint.InjectionReport, baseTime time.Time, ids identity.Provider) []report.ClientReport {
|
func SimulateTraffic(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2e.EtcdProcessCluster, profile Profile, traffic Traffic, failpointInjected <-chan failpoint.Injection, baseTime time.Time, ids identity.Provider) []report.ClientReport {
|
||||||
mux := sync.Mutex{}
|
mux := sync.Mutex{}
|
||||||
endpoints := clus.EndpointsGRPC()
|
endpoints := clus.EndpointsGRPC()
|
||||||
|
|
||||||
@ -60,7 +61,7 @@ func SimulateTraffic(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2
|
|||||||
reports := []report.ClientReport{}
|
reports := []report.ClientReport{}
|
||||||
limiter := rate.NewLimiter(rate.Limit(profile.MaximalQPS), 200)
|
limiter := rate.NewLimiter(rate.Limit(profile.MaximalQPS), 200)
|
||||||
|
|
||||||
cc, err := NewClient(endpoints, ids, baseTime)
|
cc, err := client.NewRecordingClient(endpoints, ids, baseTime)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
@ -77,11 +78,11 @@ func SimulateTraffic(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2
|
|||||||
startTime := time.Since(baseTime)
|
startTime := time.Since(baseTime)
|
||||||
for i := 0; i < profile.ClientCount; i++ {
|
for i := 0; i < profile.ClientCount; i++ {
|
||||||
wg.Add(1)
|
wg.Add(1)
|
||||||
c, nerr := NewClient([]string{endpoints[i%len(endpoints)]}, ids, baseTime)
|
c, nerr := client.NewRecordingClient([]string{endpoints[i%len(endpoints)]}, ids, baseTime)
|
||||||
if nerr != nil {
|
if nerr != nil {
|
||||||
t.Fatal(nerr)
|
t.Fatal(nerr)
|
||||||
}
|
}
|
||||||
go func(c *RecordingClient) {
|
go func(c *client.RecordingClient) {
|
||||||
defer wg.Done()
|
defer wg.Done()
|
||||||
defer c.Close()
|
defer c.Close()
|
||||||
|
|
||||||
@ -91,7 +92,7 @@ func SimulateTraffic(ctx context.Context, t *testing.T, lg *zap.Logger, clus *e2
|
|||||||
mux.Unlock()
|
mux.Unlock()
|
||||||
}(c)
|
}(c)
|
||||||
}
|
}
|
||||||
var fr *failpoint.InjectionReport
|
var fr *failpoint.Injection
|
||||||
select {
|
select {
|
||||||
case frp, ok := <-failpointInjected:
|
case frp, ok := <-failpointInjected:
|
||||||
if !ok {
|
if !ok {
|
||||||
@ -172,7 +173,7 @@ type Profile struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
type Traffic interface {
|
type Traffic interface {
|
||||||
Run(ctx context.Context, c *RecordingClient, qpsLimiter *rate.Limiter, ids identity.Provider, lm identity.LeaseIDStorage, nonUniqueWriteLimiter ConcurrencyLimiter, finish <-chan struct{})
|
Run(ctx context.Context, c *client.RecordingClient, qpsLimiter *rate.Limiter, ids identity.Provider, lm identity.LeaseIDStorage, nonUniqueWriteLimiter ConcurrencyLimiter, finish <-chan struct{})
|
||||||
ExpectUniqueRevision() bool
|
ExpectUniqueRevision() bool
|
||||||
Name() string
|
Name() string
|
||||||
}
|
}
|
||||||
|
@ -19,9 +19,9 @@ import (
|
|||||||
|
|
||||||
"github.com/anishathalye/porcupine"
|
"github.com/anishathalye/porcupine"
|
||||||
|
|
||||||
|
"go.etcd.io/etcd/tests/v3/robustness/client"
|
||||||
"go.etcd.io/etcd/tests/v3/robustness/model"
|
"go.etcd.io/etcd/tests/v3/robustness/model"
|
||||||
"go.etcd.io/etcd/tests/v3/robustness/report"
|
"go.etcd.io/etcd/tests/v3/robustness/report"
|
||||||
"go.etcd.io/etcd/tests/v3/robustness/traffic"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
func patchedOperationHistory(reports []report.ClientReport, persistedRequests []model.EtcdRequest) []porcupine.Operation {
|
func patchedOperationHistory(reports []report.ClientReport, persistedRequests []model.EtcdRequest) []porcupine.Operation {
|
||||||
@ -46,8 +46,8 @@ func relevantOperations(reports []report.ClientReport) []porcupine.Operation {
|
|||||||
return ops
|
return ops
|
||||||
}
|
}
|
||||||
|
|
||||||
func uniqueWatchEvents(reports []report.ClientReport) map[model.Event]traffic.TimedWatchEvent {
|
func uniqueWatchEvents(reports []report.ClientReport) map[model.Event]client.TimedWatchEvent {
|
||||||
persisted := map[model.Event]traffic.TimedWatchEvent{}
|
persisted := map[model.Event]client.TimedWatchEvent{}
|
||||||
for _, r := range reports {
|
for _, r := range reports {
|
||||||
for _, op := range r.Watch {
|
for _, op := range r.Watch {
|
||||||
for _, resp := range op.Responses {
|
for _, resp := range op.Responses {
|
||||||
@ -56,7 +56,7 @@ func uniqueWatchEvents(reports []report.ClientReport) map[model.Event]traffic.Ti
|
|||||||
if prev, found := persisted[event.Event]; found && prev.Time < responseTime {
|
if prev, found := persisted[event.Event]; found && prev.Time < responseTime {
|
||||||
responseTime = prev.Time
|
responseTime = prev.Time
|
||||||
}
|
}
|
||||||
persisted[event.Event] = traffic.TimedWatchEvent{Time: responseTime, WatchEvent: event}
|
persisted[event.Event] = client.TimedWatchEvent{Time: responseTime, WatchEvent: event}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -64,7 +64,7 @@ func uniqueWatchEvents(reports []report.ClientReport) map[model.Event]traffic.Ti
|
|||||||
return persisted
|
return persisted
|
||||||
}
|
}
|
||||||
|
|
||||||
func patchOperations(operations []porcupine.Operation, watchEvents map[model.Event]traffic.TimedWatchEvent, persistedOperations map[model.EtcdOperation]int64) []porcupine.Operation {
|
func patchOperations(operations []porcupine.Operation, watchEvents map[model.Event]client.TimedWatchEvent, persistedOperations map[model.EtcdOperation]int64) []porcupine.Operation {
|
||||||
newOperations := make([]porcupine.Operation, 0, len(operations))
|
newOperations := make([]porcupine.Operation, 0, len(operations))
|
||||||
lastObservedOperation := lastOperationObservedInWatch(operations, watchEvents)
|
lastObservedOperation := lastOperationObservedInWatch(operations, watchEvents)
|
||||||
|
|
||||||
@ -104,7 +104,7 @@ func patchOperations(operations []porcupine.Operation, watchEvents map[model.Eve
|
|||||||
return newOperations
|
return newOperations
|
||||||
}
|
}
|
||||||
|
|
||||||
func lastOperationObservedInWatch(operations []porcupine.Operation, watchEvents map[model.Event]traffic.TimedWatchEvent) porcupine.Operation {
|
func lastOperationObservedInWatch(operations []porcupine.Operation, watchEvents map[model.Event]client.TimedWatchEvent) porcupine.Operation {
|
||||||
var maxCallTime int64
|
var maxCallTime int64
|
||||||
var lastOperation porcupine.Operation
|
var lastOperation porcupine.Operation
|
||||||
for _, op := range operations {
|
for _, op := range operations {
|
||||||
@ -121,7 +121,7 @@ func lastOperationObservedInWatch(operations []porcupine.Operation, watchEvents
|
|||||||
return lastOperation
|
return lastOperation
|
||||||
}
|
}
|
||||||
|
|
||||||
func matchWatchEvent(request *model.TxnRequest, watchEvents map[model.Event]traffic.TimedWatchEvent) *traffic.TimedWatchEvent {
|
func matchWatchEvent(request *model.TxnRequest, watchEvents map[model.Event]client.TimedWatchEvent) *client.TimedWatchEvent {
|
||||||
for _, etcdOp := range append(request.OperationsOnSuccess, request.OperationsOnFailure...) {
|
for _, etcdOp := range append(request.OperationsOnSuccess, request.OperationsOnFailure...) {
|
||||||
if etcdOp.Type == model.PutOperation {
|
if etcdOp.Type == model.PutOperation {
|
||||||
event, ok := watchEvents[model.Event{
|
event, ok := watchEvents[model.Event{
|
||||||
|
@ -21,9 +21,9 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
|
|
||||||
"go.etcd.io/etcd/tests/v3/framework/e2e"
|
"go.etcd.io/etcd/tests/v3/framework/e2e"
|
||||||
|
"go.etcd.io/etcd/tests/v3/robustness/client"
|
||||||
"go.etcd.io/etcd/tests/v3/robustness/identity"
|
"go.etcd.io/etcd/tests/v3/robustness/identity"
|
||||||
"go.etcd.io/etcd/tests/v3/robustness/report"
|
"go.etcd.io/etcd/tests/v3/robustness/report"
|
||||||
"go.etcd.io/etcd/tests/v3/robustness/traffic"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
func collectClusterWatchEvents(ctx context.Context, t *testing.T, clus *e2e.EtcdProcessCluster, maxRevisionChan <-chan int64, cfg watchConfig, baseTime time.Time, ids identity.Provider) []report.ClientReport {
|
func collectClusterWatchEvents(ctx context.Context, t *testing.T, clus *e2e.EtcdProcessCluster, maxRevisionChan <-chan int64, cfg watchConfig, baseTime time.Time, ids identity.Provider) []report.ClientReport {
|
||||||
@ -32,14 +32,14 @@ func collectClusterWatchEvents(ctx context.Context, t *testing.T, clus *e2e.Etcd
|
|||||||
reports := make([]report.ClientReport, len(clus.Procs))
|
reports := make([]report.ClientReport, len(clus.Procs))
|
||||||
memberMaxRevisionChans := make([]chan int64, len(clus.Procs))
|
memberMaxRevisionChans := make([]chan int64, len(clus.Procs))
|
||||||
for i, member := range clus.Procs {
|
for i, member := range clus.Procs {
|
||||||
c, err := traffic.NewClient(member.EndpointsGRPC(), ids, baseTime)
|
c, err := client.NewRecordingClient(member.EndpointsGRPC(), ids, baseTime)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
memberMaxRevisionChan := make(chan int64, 1)
|
memberMaxRevisionChan := make(chan int64, 1)
|
||||||
memberMaxRevisionChans[i] = memberMaxRevisionChan
|
memberMaxRevisionChans[i] = memberMaxRevisionChan
|
||||||
wg.Add(1)
|
wg.Add(1)
|
||||||
go func(i int, c *traffic.RecordingClient) {
|
go func(i int, c *client.RecordingClient) {
|
||||||
defer wg.Done()
|
defer wg.Done()
|
||||||
defer c.Close()
|
defer c.Close()
|
||||||
watchUntilRevision(ctx, t, c, memberMaxRevisionChan, cfg)
|
watchUntilRevision(ctx, t, c, memberMaxRevisionChan, cfg)
|
||||||
@ -65,7 +65,7 @@ type watchConfig struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// watchUntilRevision watches all changes until context is cancelled, it has observed revision provided via maxRevisionChan or maxRevisionChan was closed.
|
// watchUntilRevision watches all changes until context is cancelled, it has observed revision provided via maxRevisionChan or maxRevisionChan was closed.
|
||||||
func watchUntilRevision(ctx context.Context, t *testing.T, c *traffic.RecordingClient, maxRevisionChan <-chan int64, cfg watchConfig) {
|
func watchUntilRevision(ctx context.Context, t *testing.T, c *client.RecordingClient, maxRevisionChan <-chan int64, cfg watchConfig) {
|
||||||
var maxRevision int64
|
var maxRevision int64
|
||||||
var lastRevision int64
|
var lastRevision int64
|
||||||
ctx, cancel := context.WithCancel(ctx)
|
ctx, cancel := context.WithCancel(ctx)
|
||||||
|
Loading…
x
Reference in New Issue
Block a user