storj/pkg/datarepair/repairer/config.go

114 lines
3.6 KiB
Go
Raw Normal View History

// Copyright (C) 2018 Storj Labs, Inc.
// See LICENSE for copying information.
package repairer
import (
"context"
"time"
"github.com/vivint/infectious"
"go.uber.org/zap"
"storj.io/storj/pkg/datarepair/queue"
"storj.io/storj/pkg/eestream"
"storj.io/storj/pkg/miniogw"
"storj.io/storj/pkg/overlay"
2018-12-11 17:30:14 +00:00
"storj.io/storj/pkg/pb"
"storj.io/storj/pkg/pointerdb/pdbclient"
"storj.io/storj/pkg/provider"
ecclient "storj.io/storj/pkg/storage/ec"
segment "storj.io/storj/pkg/storage/segments"
"storj.io/storj/storage/redis"
)
// Config contains configurable values for repairer
type Config struct {
QueueAddress string `help:"data repair queue address" default:"redis://127.0.0.1:6378?db=1&password=abc123"`
MaxRepair int `help:"maximum segments that can be repaired concurrently" default:"100"`
Interval time.Duration `help:"how frequently checker should audit segments" default:"3600s"`
OverlayAddr string `help:"Address to contact overlay server through"`
PointerDBAddr string `help:"Address to contact pointerdb server through"`
MaxBufferMem int `help:"maximum buffer memory (in bytes) to be allocated for read buffers" default:"0x400000"`
APIKey string `help:"repairer-specific pointerdb access credential"`
2018-12-11 17:30:14 +00:00
miniogw.NodeSelectionConfig
// TODO: this is a huge bug that these are required here. these values should
// all come from the pointer for each repair. these need to be removed from the
// config
MinThreshold int `help:"TODO: remove" default:"29"`
RepairThreshold int `help:"TODO: remove" default:"35"`
SuccessThreshold int `help:"TODO: remove" default:"80"`
MaxThreshold int `help:"TODO: remove" default:"95"`
ErasureShareSize int `help:"TODO: remove" default:"1024"`
// TODO: the repairer shouldn't need to worry about inlining, as it is only
// repairing non-inlined things.
MaxInlineSize int `help:"TODO: remove" default:"4096"`
}
// Run runs the repairer with configured values
func (c Config) Run(ctx context.Context, server *provider.Provider) (err error) {
redisQ, err := redis.NewQueueFrom(c.QueueAddress)
if err != nil {
return Error.Wrap(err)
}
queue := queue.NewQueue(redisQ)
ss, err := c.getSegmentStore(ctx, server.Identity())
if err != nil {
return Error.Wrap(err)
}
repairer := newRepairer(queue, ss, c.Interval, c.MaxRepair)
ctx, cancel := context.WithCancel(ctx)
// TODO(coyle): we need to figure out how to propagate the error up to cancel the service
go func() {
if err := repairer.Run(ctx); err != nil {
defer cancel()
zap.L().Error("Error running repairer", zap.Error(err))
}
}()
return server.Run(ctx)
}
// getSegmentStore creates a new segment store from storeConfig values
func (c Config) getSegmentStore(ctx context.Context, identity *provider.FullIdentity) (ss segment.Store, err error) {
defer mon.Task()(&ctx)(&err)
var oc overlay.Client
oc, err = overlay.NewOverlayClient(identity, c.OverlayAddr)
if err != nil {
return nil, err
}
pdb, err := pdbclient.NewClient(identity, c.PointerDBAddr, c.APIKey)
if err != nil {
return nil, err
}
ec := ecclient.NewClient(identity, c.MaxBufferMem)
fc, err := infectious.NewFEC(c.MinThreshold, c.MaxThreshold)
if err != nil {
return nil, err
}
rs, err := eestream.NewRedundancyStrategy(eestream.NewRSScheme(fc, c.ErasureShareSize), c.RepairThreshold, c.SuccessThreshold)
if err != nil {
return nil, err
}
2018-12-11 17:30:14 +00:00
ns := &pb.NodeStats{
UptimeRatio: c.UptimeRatio,
AuditSuccessRatio: c.AuditSuccessRatio,
UptimeCount: c.UptimeCount,
AuditCount: c.AuditCount,
}
return segment.NewSegmentStore(oc, ec, pdb, rs, c.MaxInlineSize, ns), nil
}