f11413bc8e
* Added a gc package at satellite/gc, which contains the gc.Service, which runs garbage collection integrated with the metainfoloop, and the gc PieceTracker, which implements the metainfo loop Observer interface and stores all of the filters (about which pieces are good) for each node. * Added a gc config located at satellite/gc/service.go (loop disabled by default in release) * Creates bloom filters with pieces to be retained inside the metainfo loop * Sends RetainRequests (or filters with good piece ids) to all storage nodes.
148 lines
4.3 KiB
Go
148 lines
4.3 KiB
Go
// Copyright (C) 2019 Storj Labs, Inc.
|
|
// See LICENSE for copying information.
|
|
|
|
package gc
|
|
|
|
import (
|
|
"context"
|
|
"time"
|
|
|
|
"github.com/zeebo/errs"
|
|
"go.uber.org/zap"
|
|
monkit "gopkg.in/spacemonkeygo/monkit.v2"
|
|
|
|
"storj.io/storj/internal/sync2"
|
|
"storj.io/storj/pkg/bloomfilter"
|
|
"storj.io/storj/pkg/overlay"
|
|
"storj.io/storj/pkg/pb"
|
|
"storj.io/storj/pkg/storj"
|
|
"storj.io/storj/pkg/transport"
|
|
"storj.io/storj/satellite/metainfo"
|
|
"storj.io/storj/uplink/piecestore"
|
|
)
|
|
|
|
var (
|
|
// Error defines the gc service errors class
|
|
Error = errs.Class("gc service error")
|
|
mon = monkit.Package()
|
|
)
|
|
|
|
// Config contains configurable values for garbage collection
|
|
type Config struct {
|
|
Interval time.Duration `help:"the time between each send of garbage collection filters to storage nodes" releaseDefault:"168h" devDefault:"10m"`
|
|
Enabled bool `help:"set if garbage collection is enabled or not" releaseDefault:"false" devDefault:"true"`
|
|
// value for InitialPieces currently based on average pieces per node
|
|
InitialPieces int `help:"the initial number of pieces expected for a storage node to have, used for creating a filter" releaseDefault:"400000" devDefault:"10"`
|
|
FalsePositiveRate float64 `help:"the false positive rate used for creating a garbage collection bloom filter" releaseDefault:"0.1" devDefault:"0.1"`
|
|
ConcurrentSends int `help:"the number of nodes to concurrently send garbage collection bloom filters to" releaseDefault:"1" devDefault:"1"`
|
|
}
|
|
|
|
// Service implements the garbage collection service
|
|
type Service struct {
|
|
log *zap.Logger
|
|
config Config
|
|
Loop sync2.Cycle
|
|
|
|
transport transport.Client
|
|
overlay overlay.DB
|
|
metainfoLoop *metainfo.Loop
|
|
}
|
|
|
|
// RetainInfo contains info needed for a storage node to retain important data and delete garbage data
|
|
type RetainInfo struct {
|
|
Filter *bloomfilter.Filter
|
|
CreationDate time.Time
|
|
Count int
|
|
}
|
|
|
|
// NewService creates a new instance of the gc service
|
|
func NewService(log *zap.Logger, config Config, transport transport.Client, overlay overlay.DB, loop *metainfo.Loop) *Service {
|
|
return &Service{
|
|
log: log,
|
|
config: config,
|
|
Loop: *sync2.NewCycle(config.Interval),
|
|
|
|
transport: transport,
|
|
overlay: overlay,
|
|
metainfoLoop: loop,
|
|
}
|
|
}
|
|
|
|
// Run starts the gc loop service
|
|
func (service *Service) Run(ctx context.Context) (err error) {
|
|
|
|
if !service.config.Enabled {
|
|
return nil
|
|
}
|
|
|
|
// TODO retrieve piece counts from overlay (when there is a column for them)
|
|
lastPieceCounts := make(map[storj.NodeID]int)
|
|
|
|
return service.Loop.Run(ctx, func(ctx context.Context) (err error) {
|
|
defer mon.Task()(&ctx)(&err)
|
|
|
|
pieceTracker := NewPieceTracker(service.log.Named("gc observer"), service.config, lastPieceCounts)
|
|
|
|
// collect things to retain
|
|
err = service.metainfoLoop.Join(ctx, pieceTracker)
|
|
if err != nil {
|
|
service.log.Error("error joining metainfoloop", zap.Error(err))
|
|
return nil
|
|
}
|
|
|
|
// save piece counts for next iteration
|
|
for id := range lastPieceCounts {
|
|
delete(lastPieceCounts, id)
|
|
}
|
|
for id, info := range pieceTracker.retainInfos {
|
|
lastPieceCounts[id] = info.Count
|
|
}
|
|
|
|
// monitor information
|
|
for _, info := range pieceTracker.retainInfos {
|
|
mon.IntVal("node_piece_count").Observe(int64(info.Count))
|
|
mon.IntVal("retain_filter_size_bytes").Observe(info.Filter.Size())
|
|
}
|
|
|
|
// send retain requests
|
|
limiter := sync2.NewLimiter(service.config.ConcurrentSends)
|
|
for id, info := range pieceTracker.retainInfos {
|
|
id, info := id, info
|
|
limiter.Go(ctx, func() {
|
|
err := service.sendRetainRequest(ctx, id, info)
|
|
if err != nil {
|
|
service.log.Error("error sending retain info to node", zap.Stringer("node ID", id), zap.Error(err))
|
|
}
|
|
})
|
|
}
|
|
limiter.Wait()
|
|
|
|
return nil
|
|
})
|
|
}
|
|
|
|
func (service *Service) sendRetainRequest(ctx context.Context, id storj.NodeID, info *RetainInfo) (err error) {
|
|
defer mon.Task()(&ctx, id.String())(&err)
|
|
|
|
log := service.log.Named(id.String())
|
|
|
|
dossier, err := service.overlay.Get(ctx, id)
|
|
if err != nil {
|
|
return Error.Wrap(err)
|
|
}
|
|
|
|
client, err := piecestore.Dial(ctx, service.transport, &dossier.Node, log, piecestore.DefaultConfig)
|
|
if err != nil {
|
|
return Error.Wrap(err)
|
|
}
|
|
defer func() {
|
|
err = errs.Combine(err, Error.Wrap(client.Close()))
|
|
}()
|
|
|
|
err = client.Retain(ctx, &pb.RetainRequest{
|
|
CreationDate: info.CreationDate,
|
|
Filter: info.Filter.Bytes(),
|
|
})
|
|
return Error.Wrap(err)
|
|
}
|