diff --git a/satellite/repair/repairer/ec.go b/satellite/repair/repairer/ec.go index eb0405337..6905b92d8 100644 --- a/satellite/repair/repairer/ec.go +++ b/satellite/repair/repairer/ec.go @@ -127,8 +127,7 @@ func (ec *ECRepairer) Get(ctx context.Context, limits []*pb.AddressedOrderLimit, NodeId: limit.GetLimit().StorageNodeId, }) } else { - ec.log.Debug("Failed to download pieces for repair", - zap.Binary("Segment", []byte(path)), + ec.log.Info("Failed to download pieces for repair", zap.Error(err)) } return @@ -318,8 +317,7 @@ func (ec *ECRepairer) Repair(ctx context.Context, limits []*pb.AddressedOrderLim infos <- info{i: i, err: err, hash: hash} }(i, addressedLimit) } - ec.log.Info("Starting a timer for repair so that the number of pieces will be closer to the success threshold", - zap.Binary("Segment", []byte(path)), + ec.log.Debug("Starting a timer for repair so that the number of pieces will be closer to the success threshold", zap.Duration("Timer", timeout), zap.Int("Node Count", nonNilCount(limits)), zap.Int("Optimal Threshold", rs.OptimalThreshold()), @@ -328,8 +326,7 @@ func (ec *ECRepairer) Repair(ctx context.Context, limits []*pb.AddressedOrderLim var successfulCount, failureCount, cancellationCount int32 timer := time.AfterFunc(timeout, func() { if ctx.Err() != context.Canceled { - ec.log.Info("Timer expired. Canceling the long tail...", - zap.Binary("Segment", []byte(path)), + ec.log.Debug("Timer expired. Canceling the long tail...", zap.Int32("Successfully repaired", atomic.LoadInt32(&successfulCount)), ) cancel() @@ -352,8 +349,7 @@ func (ec *ECRepairer) Repair(ctx context.Context, limits []*pb.AddressedOrderLim } else { cancellationCount++ } - ec.log.Debug("Repair to storage node failed", - zap.Binary("Segment", []byte(path)), + ec.log.Info("Repair to storage node failed", zap.Stringer("Node ID", limits[info.i].GetLimit().StorageNodeId), zap.Error(info.err), ) @@ -385,7 +381,6 @@ func (ec *ECRepairer) Repair(ctx context.Context, limits []*pb.AddressedOrderLim } ec.log.Info("Successfully repaired", - zap.Binary("Segment", []byte(path)), zap.Int32("Success Count", atomic.LoadInt32(&successfulCount)), ) @@ -418,7 +413,6 @@ func (ec *ECRepairer) putPiece(ctx, parent context.Context, limit *pb.AddressedO }) if err != nil { ec.log.Debug("Failed dialing for putting piece to node", - zap.Binary("Segment", []byte(path)), zap.Stringer("Piece ID", pieceID), zap.Stringer("Node ID", storageNodeID), zap.Error(err), @@ -430,7 +424,6 @@ func (ec *ECRepairer) putPiece(ctx, parent context.Context, limit *pb.AddressedO upload, err := ps.Upload(ctx, limit.GetLimit(), privateKey) if err != nil { ec.log.Debug("Failed requesting upload of pieces to node", - zap.Binary("Segment", []byte(path)), zap.Stringer("Piece ID", pieceID), zap.Stringer("Node ID", storageNodeID), zap.Error(err), @@ -453,12 +446,10 @@ func (ec *ECRepairer) putPiece(ctx, parent context.Context, limit *pb.AddressedO // to slow connection. No error logging for this case. if ctx.Err() == context.Canceled { if parent.Err() == context.Canceled { - ec.log.Info("Upload to node canceled by user", - zap.Binary("Segment", []byte(path)), + ec.log.Debug("Upload to node canceled by user", zap.Stringer("Node ID", storageNodeID)) } else { ec.log.Debug("Node cut from upload due to slow connection", - zap.Binary("Segment", []byte(path)), zap.Stringer("Node ID", storageNodeID)) } err = context.Canceled @@ -469,7 +460,6 @@ func (ec *ECRepairer) putPiece(ctx, parent context.Context, limit *pb.AddressedO } ec.log.Debug("Failed uploading piece to node", - zap.Binary("Segment", []byte(path)), zap.Stringer("Piece ID", pieceID), zap.Stringer("Node ID", storageNodeID), zap.String("Node Address", nodeAddress), diff --git a/satellite/repair/repairer/repairer.go b/satellite/repair/repairer/repairer.go index 4095aec25..c6a2c3406 100644 --- a/satellite/repair/repairer/repairer.go +++ b/satellite/repair/repairer/repairer.go @@ -133,7 +133,7 @@ func (service *Service) process(ctx context.Context) (err error) { cancel() return err } - service.log.Info("Retrieved segment from repair queue", zap.Binary("Segment", seg.GetPath())) + service.log.Debug("Retrieved segment from repair queue") // this goroutine inherits the JobLimiter semaphore acquisition and is now responsible // for releasing it. @@ -142,7 +142,7 @@ func (service *Service) process(ctx context.Context) (err error) { defer cancel() if err := service.worker(ctx, seg); err != nil { - service.log.Error("repair worker failed:", zap.Binary("Segment", seg.GetPath()), zap.Error(err)) + service.log.Error("repair worker failed:", zap.Error(err)) } }() @@ -154,16 +154,13 @@ func (service *Service) worker(ctx context.Context, seg *pb.InjuredSegment) (err workerStartTime := time.Now().UTC() - service.log.Info("Limiter running repair on segment", - zap.Binary("Segment", seg.GetPath()), - zap.String("Segment Path", string(seg.GetPath()))) + service.log.Debug("Limiter running repair on segment") // note that shouldDelete is used even in the case where err is not null shouldDelete, err := service.repairer.Repair(ctx, string(seg.GetPath())) if shouldDelete { if irreparableErr, ok := err.(*irreparableError); ok { service.log.Error("segment could not be repaired! adding to irreparableDB for more attention", - zap.Error(err), - zap.Binary("segment", seg.GetPath())) + zap.Error(err)) segmentInfo := &pb.IrreparableSegment{ Path: seg.GetPath(), SegmentDetail: irreparableErr.segmentInfo, @@ -177,11 +174,9 @@ func (service *Service) worker(ctx context.Context, seg *pb.InjuredSegment) (err } } else if err != nil { service.log.Error("unexpected error repairing segment!", - zap.Error(err), - zap.Binary("segment", seg.GetPath())) + zap.Error(err)) } else { - service.log.Info("removing repaired segment from repair queue", - zap.Binary("Segment", seg.GetPath())) + service.log.Debug("removing repaired segment from repair queue") } if shouldDelete { delErr := service.queue.Delete(ctx, seg) diff --git a/satellite/repair/repairer/segments.go b/satellite/repair/repairer/segments.go index 5d3bdaa5b..359b33afd 100644 --- a/satellite/repair/repairer/segments.go +++ b/satellite/repair/repairer/segments.go @@ -105,7 +105,7 @@ func (repairer *SegmentRepairer) Repair(ctx context.Context, path storj.Path) (s if storj.ErrObjectNotFound.Has(err) { mon.Meter("repair_unnecessary").Mark(1) //locked mon.Meter("segment_deleted_before_repair").Mark(1) //locked - repairer.log.Debug("segment was deleted", zap.Binary("Segment", []byte(path))) + repairer.log.Debug("segment was deleted") return true, nil } return false, metainfoGetError.Wrap(err)