mirror of
https://github.com/etcd-io/etcd.git
synced 2024-09-27 06:25:44 +00:00
ioutil: remaining errors.Is conversions
Signed-off-by: redwrasse <mail@redwrasse.io>
This commit is contained in:
@@ -16,7 +16,6 @@ package membership
|
||||
|
||||
import (
|
||||
"encoding/json"
|
||||
"errors"
|
||||
"fmt"
|
||||
"path"
|
||||
"reflect"
|
||||
@@ -459,7 +458,7 @@ func TestClusterValidateConfigurationChangeV2(t *testing.T) {
|
||||
}
|
||||
for i, tt := range tests {
|
||||
err := cl.ValidateConfigurationChange(tt.cc)
|
||||
if !errors.Is(err, tt.werr) {
|
||||
if err != tt.werr {
|
||||
t.Errorf("#%d: validateConfigurationChange error = %v, want %v", i, err, tt.werr)
|
||||
}
|
||||
}
|
||||
|
||||
@@ -16,7 +16,6 @@ package rafthttp
|
||||
|
||||
import (
|
||||
"bytes"
|
||||
"errors"
|
||||
"reflect"
|
||||
"testing"
|
||||
|
||||
@@ -78,13 +77,13 @@ func TestMessage(t *testing.T) {
|
||||
for i, tt := range tests {
|
||||
b := &bytes.Buffer{}
|
||||
enc := &messageEncoder{w: b}
|
||||
if err := enc.encode(&tt.msg); !errors.Is(err, tt.encodeErr) {
|
||||
if err := enc.encode(&tt.msg); err != tt.encodeErr {
|
||||
t.Errorf("#%d: encode message error expected %v, got %v", i, tt.encodeErr, err)
|
||||
continue
|
||||
}
|
||||
dec := &messageDecoder{r: b}
|
||||
m, err := dec.decode()
|
||||
if !errors.Is(err, tt.decodeErr) {
|
||||
if err != tt.decodeErr {
|
||||
t.Errorf("#%d: decode message error expected %v, got %v", i, tt.decodeErr, err)
|
||||
continue
|
||||
}
|
||||
|
||||
@@ -165,7 +165,7 @@ func (p *pipeline) post(data []byte) (err error) {
|
||||
p.picker.unreachable(u)
|
||||
// errMemberRemoved is a critical error since a removed member should
|
||||
// always be stopped. So we use reportCriticalError to report it to errorc.
|
||||
if errors.Is(err, errMemberRemoved) {
|
||||
if err == errMemberRemoved {
|
||||
reportCriticalError(err, p.errorc)
|
||||
}
|
||||
return err
|
||||
|
||||
@@ -17,7 +17,6 @@ package rafthttp
|
||||
import (
|
||||
"bytes"
|
||||
"context"
|
||||
"errors"
|
||||
"io"
|
||||
"net/http"
|
||||
"time"
|
||||
@@ -111,7 +110,7 @@ func (s *snapshotSender) send(merged snap.Message) {
|
||||
|
||||
// errMemberRemoved is a critical error since a removed member should
|
||||
// always be stopped. So we use reportCriticalError to report it to errorc.
|
||||
if errors.Is(err, errMemberRemoved) {
|
||||
if err == errMemberRemoved {
|
||||
reportCriticalError(err, s.errorc)
|
||||
}
|
||||
|
||||
|
||||
@@ -16,7 +16,6 @@ package rafthttp
|
||||
|
||||
import (
|
||||
"context"
|
||||
"errors"
|
||||
"fmt"
|
||||
"io"
|
||||
"net/http"
|
||||
@@ -404,7 +403,7 @@ func (cr *streamReader) run() {
|
||||
for {
|
||||
rc, err := cr.dial(t)
|
||||
if err != nil {
|
||||
if !errors.Is(err, errUnsupportedStreamType) {
|
||||
if err != errUnsupportedStreamType {
|
||||
cr.status.deactivate(failureType{source: t.String(), action: "dial"}, err.Error())
|
||||
}
|
||||
} else {
|
||||
@@ -429,7 +428,7 @@ func (cr *streamReader) run() {
|
||||
}
|
||||
switch {
|
||||
// all data is read out
|
||||
case errors.Is(err, io.EOF):
|
||||
case err == io.EOF:
|
||||
// connection is closed by the remote
|
||||
case transport.IsClosedConnError(err):
|
||||
default:
|
||||
|
||||
@@ -256,7 +256,7 @@ func TestStreamReaderDialDetectUnsupport(t *testing.T) {
|
||||
}
|
||||
|
||||
_, err := sr.dial(typ)
|
||||
if !errors.Is(err, errUnsupportedStreamType) {
|
||||
if err != errUnsupportedStreamType {
|
||||
t.Errorf("#%d: error = %v, want %v", i, err, errUnsupportedStreamType)
|
||||
}
|
||||
}
|
||||
|
||||
@@ -15,7 +15,6 @@
|
||||
package snap
|
||||
|
||||
import (
|
||||
"errors"
|
||||
"fmt"
|
||||
"hash/crc32"
|
||||
"os"
|
||||
@@ -81,7 +80,7 @@ func TestBadCRC(t *testing.T) {
|
||||
crcTable = crc32.MakeTable(crc32.Koopman)
|
||||
|
||||
_, err = Read(zaptest.NewLogger(t), filepath.Join(dir, fmt.Sprintf("%016x-%016x.snap", 1, 1)))
|
||||
if err == nil || !errors.Is(err, ErrCRCMismatch) {
|
||||
if err == nil || err != ErrCRCMismatch {
|
||||
t.Errorf("err = %v, want %v", err, ErrCRCMismatch)
|
||||
}
|
||||
}
|
||||
@@ -222,7 +221,7 @@ func TestNoSnapshot(t *testing.T) {
|
||||
defer os.RemoveAll(dir)
|
||||
ss := New(zaptest.NewLogger(t), dir)
|
||||
_, err = ss.Load()
|
||||
if !errors.Is(err, ErrNoSnapshot) {
|
||||
if err != ErrNoSnapshot {
|
||||
t.Errorf("err = %v, want %v", err, ErrNoSnapshot)
|
||||
}
|
||||
}
|
||||
@@ -241,7 +240,7 @@ func TestEmptySnapshot(t *testing.T) {
|
||||
}
|
||||
|
||||
_, err = Read(zaptest.NewLogger(t), filepath.Join(dir, "1.snap"))
|
||||
if !errors.Is(err, ErrEmptySnapshot) {
|
||||
if err != ErrEmptySnapshot {
|
||||
t.Errorf("err = %v, want %v", err, ErrEmptySnapshot)
|
||||
}
|
||||
}
|
||||
@@ -263,7 +262,7 @@ func TestAllSnapshotBroken(t *testing.T) {
|
||||
|
||||
ss := New(zaptest.NewLogger(t), dir)
|
||||
_, err = ss.Load()
|
||||
if !errors.Is(err, ErrNoSnapshot) {
|
||||
if err != ErrNoSnapshot {
|
||||
t.Errorf("err = %v, want %v", err, ErrNoSnapshot)
|
||||
}
|
||||
}
|
||||
|
||||
@@ -187,7 +187,7 @@ func (d *discovery) joinCluster(config string) (string, error) {
|
||||
func (d *discovery) getCluster() (string, error) {
|
||||
nodes, size, index, err := d.checkCluster()
|
||||
if err != nil {
|
||||
if errors.Is(err, ErrFullCluster) {
|
||||
if err == ErrFullCluster {
|
||||
return nodesToCluster(nodes, size)
|
||||
}
|
||||
return "", err
|
||||
@@ -227,7 +227,7 @@ func (d *discovery) checkCluster() ([]*client.Node, uint64, uint64, error) {
|
||||
if eerr, ok := err.(*client.Error); ok && eerr.Code == client.ErrorCodeKeyNotFound {
|
||||
return nil, 0, 0, ErrSizeNotFound
|
||||
}
|
||||
if errors.Is(err, client.ErrInvalidJSON) {
|
||||
if err == client.ErrInvalidJSON {
|
||||
return nil, 0, 0, ErrBadDiscoveryEndpoint
|
||||
}
|
||||
if ce, ok := err.(*client.ClusterError); ok {
|
||||
|
||||
@@ -212,7 +212,7 @@ func TestCheckCluster(t *testing.T) {
|
||||
}
|
||||
}()
|
||||
ns, size, index, err := d.checkCluster()
|
||||
if !errors.Is(err, tt.werr) {
|
||||
if err != tt.werr {
|
||||
t.Errorf("#%d: err = %v, want %v", i, err, tt.werr)
|
||||
}
|
||||
if reflect.DeepEqual(ns, tt.nodes) {
|
||||
@@ -336,7 +336,7 @@ func TestCreateSelf(t *testing.T) {
|
||||
|
||||
for i, tt := range tests {
|
||||
d := newTestDiscovery(t, "1000", 1, tt.c)
|
||||
if err := d.createSelf(""); !errors.Is(err, tt.werr) {
|
||||
if err := d.createSelf(""); err != tt.werr {
|
||||
t.Errorf("#%d: err = %v, want %v", i, err, nil)
|
||||
}
|
||||
}
|
||||
@@ -383,7 +383,7 @@ func TestNodesToCluster(t *testing.T) {
|
||||
|
||||
for i, tt := range tests {
|
||||
cluster, err := nodesToCluster(tt.nodes, tt.size)
|
||||
if !errors.Is(err, tt.werr) {
|
||||
if err != tt.werr {
|
||||
t.Errorf("#%d: err = %v, want %v", i, err, tt.werr)
|
||||
}
|
||||
if !reflect.DeepEqual(cluster, tt.wcluster) {
|
||||
@@ -435,7 +435,7 @@ func TestRetryFailure(t *testing.T) {
|
||||
fc.Advance(time.Second * (0x1 << i))
|
||||
}
|
||||
}()
|
||||
if _, _, _, err := d.checkCluster(); !errors.Is(err, ErrTooManyRetries) {
|
||||
if _, _, _, err := d.checkCluster(); err != ErrTooManyRetries {
|
||||
t.Errorf("err = %v, want %v", err, ErrTooManyRetries)
|
||||
}
|
||||
}
|
||||
|
||||
@@ -91,7 +91,7 @@ func TestGetClusterSize(t *testing.T) {
|
||||
clusterToken: "fakeToken",
|
||||
}
|
||||
|
||||
if cs, err := d.getClusterSize(); !errors.Is(err, tc.expectedErr) {
|
||||
if cs, err := d.getClusterSize(); err != tc.expectedErr {
|
||||
t.Errorf("Unexpected error, expected: %v got: %v", tc.expectedErr, err)
|
||||
} else {
|
||||
if err == nil && cs != tc.expectedSize {
|
||||
@@ -387,7 +387,7 @@ func TestCheckCluster(t *testing.T) {
|
||||
}
|
||||
|
||||
clsInfo, _, _, err := d.checkCluster()
|
||||
if !errors.Is(err, tc.expectedError) {
|
||||
if err != tc.expectedError {
|
||||
t.Errorf("Unexpected error, expected: %v, got: %v", tc.expectedError, err)
|
||||
}
|
||||
|
||||
@@ -724,7 +724,7 @@ func TestGetInitClusterStr(t *testing.T) {
|
||||
}
|
||||
|
||||
retStr, err := clsInfo.getInitClusterStr(tc.clusterSize)
|
||||
if !errors.Is(err, tc.expectedError) {
|
||||
if err != tc.expectedError {
|
||||
t.Errorf("Unexpected error, expected: %v, got: %v", tc.expectedError, err)
|
||||
}
|
||||
|
||||
|
||||
@@ -16,7 +16,6 @@ package v3rpc
|
||||
|
||||
import (
|
||||
"context"
|
||||
"errors"
|
||||
"io"
|
||||
|
||||
"go.uber.org/zap"
|
||||
@@ -62,10 +61,10 @@ func (ls *LeaseServer) LeaseRevoke(ctx context.Context, rr *pb.LeaseRevokeReques
|
||||
|
||||
func (ls *LeaseServer) LeaseTimeToLive(ctx context.Context, rr *pb.LeaseTimeToLiveRequest) (*pb.LeaseTimeToLiveResponse, error) {
|
||||
resp, err := ls.le.LeaseTimeToLive(ctx, rr)
|
||||
if err != nil && !errors.Is(err, lease.ErrLeaseNotFound) {
|
||||
if err != nil && err != lease.ErrLeaseNotFound {
|
||||
return nil, togRPCError(err)
|
||||
}
|
||||
if errors.Is(err, lease.ErrLeaseNotFound) {
|
||||
if err == lease.ErrLeaseNotFound {
|
||||
resp = &pb.LeaseTimeToLiveResponse{
|
||||
Header: &pb.ResponseHeader{},
|
||||
ID: rr.ID,
|
||||
@@ -81,7 +80,7 @@ func (ls *LeaseServer) LeaseLeases(ctx context.Context, rr *pb.LeaseLeasesReques
|
||||
if err != nil && err != lease.ErrLeaseNotFound {
|
||||
return nil, togRPCError(err)
|
||||
}
|
||||
if errors.Is(err, lease.ErrLeaseNotFound) {
|
||||
if err == lease.ErrLeaseNotFound {
|
||||
resp = &pb.LeaseLeasesResponse{
|
||||
Header: &pb.ResponseHeader{},
|
||||
Leases: []*pb.LeaseStatus{},
|
||||
@@ -101,7 +100,7 @@ func (ls *LeaseServer) LeaseKeepAlive(stream pb.Lease_LeaseKeepAliveServer) (err
|
||||
case <-stream.Context().Done():
|
||||
// the only server-side cancellation is noleader for now.
|
||||
err = stream.Context().Err()
|
||||
if errors.Is(err, context.Canceled) {
|
||||
if err == context.Canceled {
|
||||
err = rpctypes.ErrGRPCNoLeader
|
||||
}
|
||||
}
|
||||
@@ -111,7 +110,7 @@ func (ls *LeaseServer) LeaseKeepAlive(stream pb.Lease_LeaseKeepAliveServer) (err
|
||||
func (ls *LeaseServer) leaseKeepAlive(stream pb.Lease_LeaseKeepAliveServer) error {
|
||||
for {
|
||||
req, err := stream.Recv()
|
||||
if errors.Is(err, io.EOF) {
|
||||
if err == io.EOF {
|
||||
return nil
|
||||
}
|
||||
if err != nil {
|
||||
@@ -134,7 +133,7 @@ func (ls *LeaseServer) leaseKeepAlive(stream pb.Lease_LeaseKeepAliveServer) erro
|
||||
ls.hdr.fill(resp.Header)
|
||||
|
||||
ttl, err := ls.le.LeaseRenew(stream.Context(), lease.LeaseID(req.ID))
|
||||
if errors.Is(err, lease.ErrLeaseNotFound) {
|
||||
if err == lease.ErrLeaseNotFound {
|
||||
err = nil
|
||||
ttl = 0
|
||||
}
|
||||
|
||||
@@ -16,7 +16,6 @@ package v3rpc
|
||||
|
||||
import (
|
||||
"context"
|
||||
errorspkg "errors"
|
||||
"strings"
|
||||
|
||||
"google.golang.org/grpc/codes"
|
||||
@@ -96,7 +95,7 @@ var toGRPCErrorMap = map[error]error{
|
||||
|
||||
func togRPCError(err error) error {
|
||||
// let gRPC server convert to codes.Canceled, codes.DeadlineExceeded
|
||||
if errorspkg.Is(err, context.Canceled) || errorspkg.Is(err, context.DeadlineExceeded) {
|
||||
if err == context.Canceled || err == context.DeadlineExceeded {
|
||||
return err
|
||||
}
|
||||
grpcErr, ok := toGRPCErrorMap[err]
|
||||
|
||||
@@ -38,7 +38,7 @@ func TestGRPCError(t *testing.T) {
|
||||
{err: errors.New("foo"), exp: status.Error(codes.Unknown, "foo")},
|
||||
}
|
||||
for i := range tt {
|
||||
if err := togRPCError(tt[i].err); !errors.Is(err, tt[i].exp) {
|
||||
if err := togRPCError(tt[i].err); err != tt[i].exp {
|
||||
if _, ok := status.FromError(err); ok {
|
||||
if err.Error() == tt[i].exp.Error() {
|
||||
continue
|
||||
|
||||
@@ -16,7 +16,6 @@ package v3rpc
|
||||
|
||||
import (
|
||||
"context"
|
||||
"errors"
|
||||
"io"
|
||||
"math/rand"
|
||||
"sync"
|
||||
@@ -212,13 +211,13 @@ func (ws *watchServer) Watch(stream pb.Watch_WatchServer) (err error) {
|
||||
// revisited.
|
||||
select {
|
||||
case err = <-errc:
|
||||
if errors.Is(err, context.Canceled) {
|
||||
if err == context.Canceled {
|
||||
err = rpctypes.ErrGRPCWatchCanceled
|
||||
}
|
||||
close(sws.ctrlStream)
|
||||
case <-stream.Context().Done():
|
||||
err = stream.Context().Err()
|
||||
if errors.Is(err, context.Canceled) {
|
||||
if err == context.Canceled {
|
||||
err = rpctypes.ErrGRPCWatchCanceled
|
||||
}
|
||||
}
|
||||
@@ -242,7 +241,7 @@ func (sws *serverWatchStream) isWatchPermitted(wcr *pb.WatchCreateRequest) error
|
||||
func (sws *serverWatchStream) recvLoop() error {
|
||||
for {
|
||||
req, err := sws.gRPCStream.Recv()
|
||||
if errors.Is(err, io.EOF) {
|
||||
if err == io.EOF {
|
||||
return nil
|
||||
}
|
||||
if err != nil {
|
||||
@@ -281,7 +280,7 @@ func (sws *serverWatchStream) recvLoop() error {
|
||||
case auth.ErrUserEmpty:
|
||||
cancelReason = rpctypes.ErrGRPCUserEmpty.Error()
|
||||
default:
|
||||
if !errors.Is(err, auth.ErrPermissionDenied) {
|
||||
if err != auth.ErrPermissionDenied {
|
||||
sws.lg.Error("unexpected error code", zap.Error(err))
|
||||
}
|
||||
cancelReason = rpctypes.ErrGRPCPermissionDenied.Error()
|
||||
|
||||
@@ -16,7 +16,6 @@ package v3rpc
|
||||
|
||||
import (
|
||||
"bytes"
|
||||
"errors"
|
||||
"math"
|
||||
"testing"
|
||||
|
||||
@@ -70,7 +69,7 @@ func TestSendFragment(t *testing.T) {
|
||||
return nil
|
||||
}
|
||||
err := sendFragments(tt[i].wr, tt[i].maxRequestBytes, testSend)
|
||||
if !errors.Is(err, tt[i].werr) {
|
||||
if err != tt[i].werr {
|
||||
t.Errorf("#%d: expected error %v, got %v", i, tt[i].werr, err)
|
||||
}
|
||||
got := len(fragmentedResp)
|
||||
|
||||
Reference in New Issue
Block a user