Skip to content

Commit

Permalink
*: remove extra logging
Browse files Browse the repository at this point in the history
Combine with previous commit before merging

Release note: None
  • Loading branch information
petermattis committed Oct 4, 2018
1 parent 622ad9a commit 0e8fa32
Show file tree
Hide file tree
Showing 3 changed files with 1 addition and 18 deletions.
3 changes: 0 additions & 3 deletions pkg/cmd/roachtest/bank.go
Original file line number Diff line number Diff line change
Expand Up @@ -246,9 +246,6 @@ func (s *bankState) chaosMonkey(
return false
}

c.l.Printf("round %d: verifying accounts\n", curRound)
s.verifyAccounts(ctx, t)

// Sleep until at least one client is writing successfully.
c.l.Printf("round %d: monkey sleeping while cluster recovers...\n", curRound)
for !s.done(ctx) && !madeProgress() {
Expand Down
9 changes: 1 addition & 8 deletions pkg/rpc/nodedialer/nodedialer.go
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,6 @@ package nodedialer
import (
"context"
"net"
"runtime/debug"
"time"
"unsafe"

Expand Down Expand Up @@ -102,13 +101,11 @@ func (n *Dialer) Dial(ctx context.Context, nodeID roachpb.NodeID) (_ *grpc.Clien

addr, err := n.resolver(nodeID)
if err != nil {
log.Infof(ctx, "n%d: breaker opened\n%s", nodeID, debug.Stack())
breaker.Fail()
return nil, err
}
conn, err := n.rpcContext.GRPCDial(addr.String()).Connect(ctx)
if err != nil {
log.Infof(ctx, "n%d: breaker opened\n%s", nodeID, debug.Stack())
breaker.Fail()
return nil, err
}
Expand Down Expand Up @@ -153,11 +150,7 @@ func (n *Dialer) DialInternalClient(
// ConnHealth when scheduling processors, but can then see attempts to send
// RPCs fail when dial fails due to an open breaker. Reset the breaker here
// as a stop-gap before the reconciliation occurs.
breaker := n.getBreaker(nodeID)
if !breaker.Ready() {
log.Infof(ctx, "n%d: %s: but dialed anyways", nodeID, circuit.ErrBreakerOpen)
}
breaker.Reset()
n.getBreaker(nodeID).Reset()
if err := grpcutil.ConnectionReady(conn); err != nil {
return nil, nil, err
}
Expand Down
7 changes: 0 additions & 7 deletions pkg/sql/distsql_physical_planner.go
Original file line number Diff line number Diff line change
Expand Up @@ -608,8 +608,6 @@ type distSQLNodeHealth struct {
}

func (h *distSQLNodeHealth) check(ctx context.Context, nodeID roachpb.NodeID) error {
log.Infof(ctx, "n%d: checking node health", nodeID)

{
// NB: as of #22658, ConnHealth does not work as expected; see the
// comment within. We still keep this code for now because in
Expand All @@ -624,18 +622,15 @@ func (h *distSQLNodeHealth) check(ctx context.Context, nodeID roachpb.NodeID) er
// instead). Note: this can never happen for our nodeID (which
// always has its address in the nodeMap).
log.VEventf(ctx, 1, "marking n%d as unhealthy for this plan: %v", nodeID, err)
log.Infof(ctx, "n%d: node health: %v", nodeID, err)
return err
}
log.Infof(ctx, "n%d: node health: %v", nodeID, err)
}
{
live, err := h.isLive(nodeID)
if err == nil && !live {
err = errors.New("node is not live")
}
if err != nil {
log.Infof(ctx, "n%d: node health: %v", nodeID, err)
return errors.Wrapf(err, "not using n%d due to liveness", nodeID)
}
}
Expand All @@ -649,14 +644,12 @@ func (h *distSQLNodeHealth) check(ctx context.Context, nodeID roachpb.NodeID) er
// written on startup, the most likely scenario is
// that the node is ready. We therefore return no
// error.
log.Infof(ctx, "n%d: node healthy (not draining)", nodeID)
return nil
}

if drainingInfo.Draining {
errMsg := fmt.Sprintf("not using n%d because it is draining", nodeID)
log.VEvent(ctx, 1, errMsg)
log.Infof(ctx, "n%d: node health: %s", nodeID, errMsg)
return errors.New(errMsg)
}

Expand Down

0 comments on commit 0e8fa32

Please sign in to comment.