*: clean up code format

Signed-off-by: Gyuho Lee <gyuhox@gmail.com>
release-3.4
Gyuho Lee 2018-07-21 15:49:33 -07:00
parent 6ab3cc0a2e
commit 42bef8460c
7 changed files with 22 additions and 25 deletions

View File

@ -472,7 +472,7 @@ func TestHTTPClusterClientDoDeadlineExceedContext(t *testing.T) {
type fakeCancelContext struct{}
var fakeCancelContextError = errors.New("fake context canceled")
var errFakeCancelContext = errors.New("fake context canceled")
func (f fakeCancelContext) Deadline() (time.Time, bool) { return time.Time{}, false }
func (f fakeCancelContext) Done() <-chan struct{} {
@ -480,7 +480,7 @@ func (f fakeCancelContext) Done() <-chan struct{} {
d <- struct{}{}
return d
}
func (f fakeCancelContext) Err() error { return fakeCancelContextError }
func (f fakeCancelContext) Err() error { return errFakeCancelContext }
func (f fakeCancelContext) Value(key interface{}) interface{} { return 1 }
func withTimeout(parent context.Context, timeout time.Duration) (
@ -512,8 +512,8 @@ func TestHTTPClusterClientDoCanceledContext(t *testing.T) {
select {
case err := <-errc:
if err != fakeCancelContextError {
t.Errorf("err = %+v, want %+v", err, fakeCancelContextError)
if err != errFakeCancelContext {
t.Errorf("err = %+v, want %+v", err, errFakeCancelContext)
}
case <-time.After(time.Second):
t.Fatalf("unexpected timeout when waiting for request to fake context canceled")

View File

@ -46,7 +46,7 @@ func TestV2NoRetryEOF(t *testing.T) {
conn.Close()
}
}()
eofURL := integration.UrlScheme + "://" + lEOF.Addr().String()
eofURL := integration.URLScheme + "://" + lEOF.Addr().String()
cli := integration.MustNewHTTPClient(t, []string{eofURL, eofURL}, nil)
kapi := client.NewKeysAPI(cli)
for i, f := range noRetryList(kapi) {
@ -64,16 +64,16 @@ func TestV2NoRetryEOF(t *testing.T) {
// TestV2NoRetryNoLeader tests destructive api calls won't retry if given an error code.
func TestV2NoRetryNoLeader(t *testing.T) {
defer testutil.AfterTest(t)
lHttp := integration.NewListenerWithAddr(t, fmt.Sprintf("127.0.0.1:%05d", os.Getpid()))
lHTTP := integration.NewListenerWithAddr(t, fmt.Sprintf("127.0.0.1:%05d", os.Getpid()))
eh := &errHandler{errCode: http.StatusServiceUnavailable}
srv := httptest.NewUnstartedServer(eh)
defer lHttp.Close()
defer lHTTP.Close()
defer srv.Close()
srv.Listener = lHttp
srv.Listener = lHTTP
go srv.Start()
lHttpURL := integration.UrlScheme + "://" + lHttp.Addr().String()
lHTTPURL := integration.URLScheme + "://" + lHTTP.Addr().String()
cli := integration.MustNewHTTPClient(t, []string{lHttpURL, lHttpURL}, nil)
cli := integration.MustNewHTTPClient(t, []string{lHTTPURL, lHTTPURL}, nil)
kapi := client.NewKeysAPI(cli)
// test error code
for i, f := range noRetryList(kapi) {
@ -94,7 +94,7 @@ func TestV2RetryRefuse(t *testing.T) {
cl.Launch(t)
defer cl.Terminate(t)
// test connection refused; expect no error failover
cli := integration.MustNewHTTPClient(t, []string{integration.UrlScheme + "://refuseconn:123", cl.URL(0)}, nil)
cli := integration.MustNewHTTPClient(t, []string{integration.URLScheme + "://refuseconn:123", cl.URL(0)}, nil)
kapi := client.NewKeysAPI(cli)
if _, err := kapi.Set(context.Background(), "/delkey", "def", nil); err != nil {
t.Fatal(err)

View File

@ -393,7 +393,7 @@ func (rc *raftNode) serveChannels() {
// send proposals over raft
go func() {
var confChangeCount uint64 = 0
confChangeCount := uint64(0)
for rc.proposeC != nil && rc.confChangeC != nil {
select {
@ -409,7 +409,7 @@ func (rc *raftNode) serveChannels() {
if !ok {
rc.confChangeC = nil
} else {
confChangeCount += 1
confChangeCount++
cc.ID = confChangeCount
rc.node.ProposeConfChange(context.TODO(), cc)
}

View File

@ -86,11 +86,11 @@ func (s *simplePrinter) Grant(resp v3.LeaseGrantResponse) {
fmt.Printf("lease %016x granted with TTL(%ds)\n", resp.ID, resp.TTL)
}
func (p *simplePrinter) Revoke(id v3.LeaseID, r v3.LeaseRevokeResponse) {
func (s *simplePrinter) Revoke(id v3.LeaseID, r v3.LeaseRevokeResponse) {
fmt.Printf("lease %016x revoked\n", id)
}
func (p *simplePrinter) KeepAlive(resp v3.LeaseKeepAliveResponse) {
func (s *simplePrinter) KeepAlive(resp v3.LeaseKeepAliveResponse) {
fmt.Printf("lease %016x keepalived with TTL(%d)\n", resp.ID, resp.TTL)
}

View File

@ -68,8 +68,8 @@ const (
clusterName = "etcd"
basePort = 21000
UrlScheme = "unix"
UrlSchemeTLS = "unixs"
URLScheme = "unix"
URLSchemeTLS = "unixs"
)
var (
@ -77,7 +77,7 @@ var (
// integration test uses unique ports, counting up, to listen for each
// member, ensuring restarted members can listen on the same port again.
localListenCount int64 = 0
localListenCount = int64(0)
testTLSInfo = transport.TLSInfo{
KeyFile: "./fixtures/server.key.insecure",
@ -157,9 +157,9 @@ type cluster struct {
func schemeFromTLSInfo(tls *transport.TLSInfo) string {
if tls == nil {
return UrlScheme
return URLScheme
}
return UrlSchemeTLS
return URLSchemeTLS
}
func (c *cluster) fillClusterForMembers() error {

View File

@ -34,7 +34,7 @@ func getAuthTokenFromClient(ctx context.Context) string {
return ""
}
func withClientAuthToken(ctx context.Context, ctxWithToken context.Context) context.Context {
func withClientAuthToken(ctx, ctxWithToken context.Context) context.Context {
token := getAuthTokenFromClient(ctxWithToken)
if token != "" {
ctx = context.WithValue(ctx, rpctypes.TokenFieldNameGRPC, token)

View File

@ -148,10 +148,7 @@ func leaseTestGrantLeasesList(cx ctlCtx) error {
if err != nil {
return fmt.Errorf("lease id not in returned list (%v)", err)
}
if err = proc.Close(); err != nil {
return err
}
return nil
return proc.Close()
}
func leaseTestTimeToLiveExpired(cx ctlCtx) {