storj/pkg/accounting/tally/tally.go

198 lines
5.6 KiB
Go
Raw Normal View History

// Copyright (C) 2018 Storj Labs, Inc.
// See LICENSE for copying information.
package tally
import (
"context"
"time"
"github.com/gogo/protobuf/proto"
"go.uber.org/zap"
"storj.io/storj/pkg/accounting"
dbx "storj.io/storj/pkg/accounting/dbx"
"storj.io/storj/pkg/bwagreement"
"storj.io/storj/pkg/pb"
"storj.io/storj/pkg/pointerdb"
"storj.io/storj/pkg/storj"
"storj.io/storj/storage"
)
// Tally is the service for accounting for data stored on each storage node
type Tally interface {
Run(ctx context.Context) error
}
type tally struct {
pointerdb *pointerdb.Server
overlay pb.OverlayServer
limit int
logger *zap.Logger
ticker *time.Ticker
db *accounting.Database
bwAgreement bwagreement.DB // bwagreements database
}
func newTally(logger *zap.Logger, db *accounting.Database, bwAgreement bwagreement.DB, pointerdb *pointerdb.Server, overlay pb.OverlayServer, limit int, interval time.Duration) *tally {
return &tally{
pointerdb: pointerdb,
overlay: overlay,
limit: limit,
logger: logger,
ticker: time.NewTicker(interval),
db: db,
bwAgreement: bwAgreement,
}
}
// Run the tally loop
func (t *tally) Run(ctx context.Context) (err error) {
defer mon.Task()(&ctx)(&err)
for {
err = t.calculateAtRestData(ctx)
if err != nil {
t.logger.Error("calculateAtRestData failed", zap.Error(err))
}
err = t.Query(ctx)
if err != nil {
t.logger.Error("Query for bandwith failed", zap.Error(err))
}
select {
case <-t.ticker.C: // wait for the next interval to happen
case <-ctx.Done(): // or the tally is canceled via context
return ctx.Err()
}
}
}
// calculateAtRestData iterates through the pieces on pointerdb and calculates
// the amount of at-rest data stored on each respective node
func (t *tally) calculateAtRestData(ctx context.Context) (err error) {
defer mon.Task()(&ctx)(&err)
var nodeData = make(map[storj.NodeID]int64)
err = t.pointerdb.Iterate(ctx, &pb.IterateRequest{Recurse: true},
func(it storage.Iterator) error {
var item storage.ListItem
for it.Next(&item) {
pointer := &pb.Pointer{}
err = proto.Unmarshal(item.Value, pointer)
if err != nil {
return Error.Wrap(err)
}
remote := pointer.GetRemote()
if remote == nil {
continue
}
pieces := remote.GetRemotePieces()
if pieces == nil {
t.logger.Debug("no pieces on remote segment")
continue
}
segmentSize := pointer.GetSegmentSize()
redundancy := remote.GetRedundancy()
if redundancy == nil {
t.logger.Debug("no redundancy scheme present")
continue
}
minReq := redundancy.GetMinReq()
if minReq <= 0 {
t.logger.Debug("pointer minReq must be an int greater than 0")
continue
}
pieceSize := segmentSize / int64(minReq)
for _, piece := range pieces {
nodeData[piece.NodeId] += pieceSize
}
}
return nil
},
)
return t.updateRawTable(ctx, nodeData)
}
func (t *tally) updateRawTable(ctx context.Context, nodeData map[storj.NodeID]int64) error {
//TODO
return nil
}
// Query bandwidth allocation database, selecting all new contracts since the last collection run time.
// Grouping by storage node ID and adding total of bandwidth to granular data table.
func (t *tally) Query(ctx context.Context) error {
lastBwTally, err := t.db.FindLastBwTally(ctx)
if err != nil {
return err
}
var bwAgreements []bwagreement.Agreement
if lastBwTally == nil {
t.logger.Info("Tally found no existing bandwith tracking data")
bwAgreements, err = t.bwAgreement.GetAgreements(ctx)
} else {
bwAgreements, err = t.bwAgreement.GetAgreementsSince(ctx, lastBwTally.Value)
}
if len(bwAgreements) == 0 {
t.logger.Info("Tally found no new bandwidth allocations")
return nil
}
// sum totals by node id ... todo: add nodeid as SQL column so DB can do this?
bwTotals := make(map[string]int64)
var latestBwa time.Time
for _, baRow := range bwAgreements {
rbad := &pb.RenterBandwidthAllocation_Data{}
if err := proto.Unmarshal(baRow.Agreement, rbad); err != nil {
t.logger.DPanic("Could not deserialize renter bwa in tally query")
continue
}
if baRow.CreatedAt.After(latestBwa) {
latestBwa = baRow.CreatedAt
}
bwTotals[rbad.StorageNodeId.String()] += rbad.GetTotal() // todo: check for overflow?
}
//todo: consider if we actually need EndTime in granular
if lastBwTally == nil {
t.logger.Info("No previous bandwidth timestamp found in tally query")
lastBwTally = &dbx.Value_Row{Value: latestBwa} //todo: something better here?
}
//insert all records in a transaction so if we fail, we don't have partial info stored
//todo: replace with a WithTx() method per DBX docs?
tx, err := t.db.BeginTx(ctx)
if err != nil {
t.logger.DPanic("Failed to create DB txn in tally query")
return err
}
defer func() {
if err == nil {
err = tx.Commit()
} else {
t.logger.Warn("DB txn was rolled back in tally query")
err = tx.Rollback()
}
}()
//todo: switch to bulk update SQL?
for k, v := range bwTotals {
nID := dbx.Raw_NodeId(k)
end := dbx.Raw_IntervalEndTime(latestBwa)
total := dbx.Raw_DataTotal(v)
dataType := dbx.Raw_DataType(accounting.Bandwith)
_, err = tx.Create_Raw(ctx, nID, end, total, dataType)
if err != nil {
t.logger.DPanic("Create granular SQL failed in tally query")
return err //todo: retry strategy?
}
}
//todo: move this into txn when we have masterdb?
update := dbx.Timestamps_Update_Fields{Value: dbx.Timestamps_Value(latestBwa)}
_, err = tx.Update_Timestamps_By_Name(ctx, accounting.LastBandwidthTally, update)
if err != nil {
t.logger.DPanic("Failed to update bandwith timestamp in tally query")
}
return err
}