91bd4191dd
The rollup archiver chore moves bucket bandwidth rollups and storagenode rollups that are older than a given duration to two new archive tables. Change-Id: I1626a3742ad4271bc744fbcefa6355a29d49c6a5
100 lines
3.3 KiB
Go
100 lines
3.3 KiB
Go
// Copyright (C) 2020 Storj Labs, Inc.
|
|
// See LICENSE for copying information.
|
|
|
|
package rolluparchive_test
|
|
|
|
import (
|
|
"fmt"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/stretchr/testify/require"
|
|
"go.uber.org/zap"
|
|
|
|
"storj.io/common/pb"
|
|
"storj.io/common/testcontext"
|
|
"storj.io/common/testrand"
|
|
"storj.io/storj/private/testplanet"
|
|
"storj.io/storj/satellite"
|
|
)
|
|
|
|
func TestRollupArchiveChore(t *testing.T) {
|
|
testplanet.Run(t, testplanet.Config{
|
|
SatelliteCount: 1, StorageNodeCount: 10, UplinkCount: 0,
|
|
Reconfigure: testplanet.Reconfigure{
|
|
Satellite: func(log *zap.Logger, index int, config *satellite.Config) {
|
|
// ensure that orders (and rollups) aren't marked as expired and removed
|
|
config.Orders.Expiration = time.Hour * 24 * 7
|
|
},
|
|
},
|
|
},
|
|
func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
// The purpose of this test is to ensure that the archive chore deletes
|
|
// entries in the storagenode_bandwidth_rollups and bucket_bandwidth_rollups tables
|
|
// and inserts those entries into new archive tables.
|
|
|
|
satellite := planet.Satellites[0]
|
|
satellite.Accounting.Rollup.Loop.Pause()
|
|
|
|
days := 6
|
|
|
|
currentTime := time.Now().UTC()
|
|
// Set timestamp back by the number of days we want to save
|
|
timestamp := currentTime.AddDate(0, 0, -days).Truncate(time.Millisecond)
|
|
|
|
projectID := testrand.UUID()
|
|
|
|
for i := 0; i < days; i++ {
|
|
nodeID := testrand.NodeID()
|
|
var bucketName string
|
|
bwAmount := int64(1000)
|
|
|
|
// When the bucket name and intervalStart is different, a new record is created
|
|
bucketName = fmt.Sprintf("%s%d", "testbucket", i)
|
|
|
|
err := satellite.DB.Orders().UpdateBucketBandwidthSettle(ctx,
|
|
projectID, []byte(bucketName), pb.PieceAction_GET, bwAmount, timestamp,
|
|
)
|
|
require.NoError(t, err)
|
|
|
|
err = satellite.DB.Orders().UpdateStoragenodeBandwidthSettle(ctx,
|
|
nodeID, pb.PieceAction_GET, bwAmount, timestamp)
|
|
require.NoError(t, err)
|
|
|
|
// Advance time by 24 hours
|
|
timestamp = timestamp.Add(time.Hour * 24)
|
|
}
|
|
|
|
lastWeek := currentTime.AddDate(0, 0, -7).Truncate(time.Millisecond)
|
|
nodeRollups, err := satellite.DB.StoragenodeAccounting().GetRollupsSince(ctx, lastWeek)
|
|
require.NoError(t, err)
|
|
require.Len(t, nodeRollups, days)
|
|
|
|
bucketRollups, err := satellite.DB.ProjectAccounting().GetRollupsSince(ctx, lastWeek)
|
|
require.NoError(t, err)
|
|
require.Len(t, bucketRollups, days)
|
|
|
|
// We take off a millisecond so the before isn't exactly the same as one of the interval starts.
|
|
before := currentTime.AddDate(0, 0, -days/2).Add(-time.Millisecond)
|
|
batchSize := 1000
|
|
err = satellite.Accounting.RollupArchive.ArchiveRollups(ctx, before, batchSize)
|
|
require.NoError(t, err)
|
|
|
|
nodeRollups, err = satellite.DB.StoragenodeAccounting().GetRollupsSince(ctx, lastWeek)
|
|
require.NoError(t, err)
|
|
require.Len(t, nodeRollups, days/2)
|
|
|
|
bucketRollups, err = satellite.DB.ProjectAccounting().GetRollupsSince(ctx, lastWeek)
|
|
require.NoError(t, err)
|
|
require.Len(t, bucketRollups, days/2)
|
|
|
|
nodeRollups, err = satellite.DB.StoragenodeAccounting().GetArchivedRollupsSince(ctx, lastWeek)
|
|
require.NoError(t, err)
|
|
require.Len(t, nodeRollups, days/2)
|
|
|
|
bucketRollups, err = satellite.DB.ProjectAccounting().GetArchivedRollupsSince(ctx, lastWeek)
|
|
require.NoError(t, err)
|
|
require.Len(t, bucketRollups, days/2)
|
|
})
|
|
}
|