storj/satellite/metainfo/service_test.go
Kaloyan Raev 34613e4588 cmd/satellite: command for fixing old-style objects
Jira: https://storjlabs.atlassian.net/browse/PG-67

There are a number of old-style objects (without the number of segments
in their metadata) on US-Central-1, Europe-West-1, and Asia-East-1
satellites. We want to migrate their metadata to contain the number of
segments before executing the main migration to metabase. This would
simplify the main migration to metabase with one less issue to think
about.

Change-Id: I42497ae0375b5eb972aab08c700048b9a93bb18f
2020-09-21 14:46:19 +00:00

341 lines
11 KiB
Go

// Copyright (C) 2019 Storj Labs, Inc.
// See LICENSE for copying information.
package metainfo_test
import (
"context"
"fmt"
"strconv"
"testing"
"github.com/stretchr/testify/assert"
"github.com/stretchr/testify/require"
"github.com/zeebo/errs"
"storj.io/common/memory"
"storj.io/common/pb"
"storj.io/common/storj"
"storj.io/common/testcontext"
"storj.io/common/testrand"
"storj.io/storj/private/testplanet"
"storj.io/storj/satellite/metainfo"
"storj.io/storj/satellite/metainfo/metabase"
"storj.io/storj/storage"
)
const lastSegmentIndex = -1
func TestIterate(t *testing.T) {
testplanet.Run(t, testplanet.Config{
SatelliteCount: 1, StorageNodeCount: 6, UplinkCount: 1,
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
saPeer := planet.Satellites[0]
uplinkPeer := planet.Uplinks[0]
// Setup: create 2 test buckets
err := uplinkPeer.CreateBucket(ctx, saPeer, "test1")
require.NoError(t, err)
err = uplinkPeer.CreateBucket(ctx, saPeer, "test2")
require.NoError(t, err)
// Setup: upload an object in one of the buckets
expectedData := testrand.Bytes(50 * memory.KiB)
err = uplinkPeer.Upload(ctx, saPeer, "test2", "test/path", expectedData)
require.NoError(t, err)
// Test: Confirm that only the objects are in pointerDB
// and not the bucket metadata
var itemCount int
err = saPeer.Metainfo.Database.Iterate(ctx, storage.IterateOptions{Recurse: true},
func(ctx context.Context, it storage.Iterator) error {
var item storage.ListItem
for it.Next(ctx, &item) {
itemCount++
pathElements := storj.SplitPath(storj.Path(item.Key))
// there should not be any objects in pointerDB with less than 4 path
// elements. i.e buckets should not be stored in pointerDB
require.True(t, len(pathElements) > 3)
}
return nil
})
require.NoError(t, err)
// There should only be 1 item in pointerDB, the one object
require.Equal(t, 1, itemCount)
})
}
// TestGetItems_ReturnValueOrder ensures the return value
// of GetItems will always be the same order as the requested paths.
// The test does following steps:
// - Uploads test data (multi-segment objects)
// - Gather all object paths with an extra invalid path at random position
// - Retrieve pointers using above paths
// - Ensure the nil pointer and last segment paths are in the same order as their
// corresponding paths.
func TestGetItems_ReturnValueOrder(t *testing.T) {
testplanet.Run(t, testplanet.Config{
SatelliteCount: 1, StorageNodeCount: 6, UplinkCount: 1,
Reconfigure: testplanet.Reconfigure{
Satellite: testplanet.Combine(
testplanet.ReconfigureRS(2, 2, 4, 4),
testplanet.MaxSegmentSize(3*memory.KiB),
),
},
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
satellite := planet.Satellites[0]
uplinkPeer := planet.Uplinks[0]
numItems := 5
for i := 0; i < numItems; i++ {
path := fmt.Sprintf("test/path_%d", i)
err := uplinkPeer.Upload(ctx, satellite, "bucket", path, testrand.Bytes(15*memory.KiB))
require.NoError(t, err)
}
keys, err := satellite.Metainfo.Database.List(ctx, nil, numItems)
require.NoError(t, err)
var segmentKeys = make([]metabase.SegmentKey, 0, numItems+1)
var lastSegmentPathIndices []int
// Random nil pointer
nilPointerIndex := testrand.Intn(numItems + 1)
for i, key := range keys {
segmentKeys = append(segmentKeys, metabase.SegmentKey(key))
segmentIdx, err := parseSegmentPath([]byte(key.String()))
require.NoError(t, err)
if segmentIdx == lastSegmentIndex {
lastSegmentPathIndices = append(lastSegmentPathIndices, i)
}
// set a random path to be nil.
if nilPointerIndex == i {
segmentKeys[nilPointerIndex] = nil
}
}
pointers, err := satellite.Metainfo.Service.GetItems(ctx, segmentKeys)
require.NoError(t, err)
for i, p := range pointers {
if p == nil {
require.Equal(t, nilPointerIndex, i)
continue
}
meta := pb.StreamMeta{}
metaInBytes := p.GetMetadata()
err = pb.Unmarshal(metaInBytes, &meta)
require.NoError(t, err)
lastSegmentMeta := meta.GetLastSegmentMeta()
if lastSegmentMeta != nil {
require.Equal(t, lastSegmentPathIndices[i], i)
}
}
})
}
func TestUpdatePiecesCheckDuplicates(t *testing.T) {
testplanet.Run(t, testplanet.Config{
SatelliteCount: 1, StorageNodeCount: 3, UplinkCount: 1,
Reconfigure: testplanet.Reconfigure{
Satellite: testplanet.ReconfigureRS(1, 1, 3, 3),
},
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
satellite := planet.Satellites[0]
uplinkPeer := planet.Uplinks[0]
path := "test/path"
err := uplinkPeer.Upload(ctx, satellite, "test1", path, testrand.Bytes(5*memory.KiB))
require.NoError(t, err)
keys, err := satellite.Metainfo.Database.List(ctx, nil, 1)
require.NoError(t, err)
require.Equal(t, 1, len(keys))
encPath, err := metabase.ParseSegmentKey(metabase.SegmentKey(keys[0]))
require.NoError(t, err)
pointer, err := satellite.Metainfo.Service.Get(ctx, encPath.Encode())
require.NoError(t, err)
pieces := pointer.GetRemote().GetRemotePieces()
require.False(t, hasDuplicates(pointer.GetRemote().GetRemotePieces()))
// Remove second piece in the list and replace it with
// a piece on the first node.
// This way we can ensure that we use a valid piece num.
removePiece := &pb.RemotePiece{
PieceNum: pieces[1].PieceNum,
NodeId: pieces[1].NodeId,
}
addPiece := &pb.RemotePiece{
PieceNum: pieces[1].PieceNum,
NodeId: pieces[0].NodeId,
}
// test no duplicates
updPointer, err := satellite.Metainfo.Service.UpdatePiecesCheckDuplicates(ctx, encPath.Encode(), pointer, []*pb.RemotePiece{addPiece}, []*pb.RemotePiece{removePiece}, true)
require.True(t, metainfo.ErrNodeAlreadyExists.Has(err))
require.False(t, hasDuplicates(updPointer.GetRemote().GetRemotePieces()))
// test allow duplicates
updPointer, err = satellite.Metainfo.Service.UpdatePieces(ctx, encPath.Encode(), pointer, []*pb.RemotePiece{addPiece}, []*pb.RemotePiece{removePiece})
require.NoError(t, err)
require.True(t, hasDuplicates(updPointer.GetRemote().GetRemotePieces()))
})
}
func hasDuplicates(pieces []*pb.RemotePiece) bool {
nodePieceCounts := make(map[storj.NodeID]int)
for _, piece := range pieces {
nodePieceCounts[piece.NodeId]++
}
for _, count := range nodePieceCounts {
if count > 1 {
return true
}
}
return false
}
func TestCountBuckets(t *testing.T) {
testplanet.Run(t, testplanet.Config{
SatelliteCount: 1, StorageNodeCount: 1, UplinkCount: 1,
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
saPeer := planet.Satellites[0]
uplinkPeer := planet.Uplinks[0]
projectID := planet.Uplinks[0].Projects[0].ID
count, err := saPeer.Metainfo.Service.CountBuckets(ctx, projectID)
require.NoError(t, err)
require.Equal(t, 0, count)
// Setup: create 2 test buckets
err = uplinkPeer.CreateBucket(ctx, saPeer, "test1")
require.NoError(t, err)
count, err = saPeer.Metainfo.Service.CountBuckets(ctx, projectID)
require.NoError(t, err)
require.Equal(t, 1, count)
err = uplinkPeer.CreateBucket(ctx, saPeer, "test2")
require.NoError(t, err)
count, err = saPeer.Metainfo.Service.CountBuckets(ctx, projectID)
require.NoError(t, err)
require.Equal(t, 2, count)
})
}
func parseSegmentPath(segmentPath []byte) (segmentIndex int64, err error) {
elements := storj.SplitPath(string(segmentPath))
if len(elements) < 4 {
return -1, errs.New("invalid path %q", string(segmentPath))
}
// var segmentIndex int64
if elements[1] == "l" {
segmentIndex = lastSegmentIndex
} else {
segmentIndex, err = strconv.ParseInt(elements[1][1:], 10, 64)
if err != nil {
return lastSegmentIndex, errs.Wrap(err)
}
}
return segmentIndex, nil
}
func TestFixOldStyleObject(t *testing.T) {
testplanet.Run(t, testplanet.Config{
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 1,
Reconfigure: testplanet.Reconfigure{
Satellite: testplanet.MaxSegmentSize(100 * memory.B),
},
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
satellite := planet.Satellites[0]
uplink := planet.Uplinks[0]
path := "test/path"
for i, tt := range []struct {
objectSize memory.Size
dryRun bool
expectedSegments int64
}{
{objectSize: 50 * memory.B, dryRun: true, expectedSegments: 0},
{objectSize: 50 * memory.B, dryRun: false, expectedSegments: 1},
{objectSize: 110 * memory.B, dryRun: true, expectedSegments: 0},
{objectSize: 110 * memory.B, dryRun: false, expectedSegments: 2},
{objectSize: 270 * memory.B, dryRun: true, expectedSegments: 0},
{objectSize: 270 * memory.B, dryRun: false, expectedSegments: 3},
{objectSize: 330 * memory.B, dryRun: true, expectedSegments: 0},
{objectSize: 330 * memory.B, dryRun: false, expectedSegments: 4},
} {
errTag := fmt.Sprintf("%d. %+v", i, tt)
err := uplink.Upload(ctx, satellite, "test1", path, testrand.Bytes(tt.objectSize))
require.NoError(t, err, errTag)
keys, err := satellite.Metainfo.Database.List(ctx, nil, 1)
require.NoError(t, err, errTag)
require.Equal(t, 1, len(keys))
key := metabase.SegmentKey(keys[0])
location, err := metabase.ParseSegmentKey(key)
require.NoError(t, err, errTag)
// fixing non-last segment should return error
location.Index = 1
_, err = satellite.Metainfo.Service.FixOldStyleObject(ctx, location.Encode(), tt.dryRun)
require.Error(t, err, errTag)
// fixing new-style object should return no error and changed = false
changed, err := satellite.Metainfo.Service.FixOldStyleObject(ctx, key, tt.dryRun)
require.NoError(t, err, errTag)
require.False(t, changed)
pointer, err := satellite.Metainfo.Service.Get(ctx, key)
require.NoError(t, err, errTag)
// assert the number of segments is a positive number before setting it to 0
streamMeta := &pb.StreamMeta{}
err = pb.Unmarshal(pointer.Metadata, streamMeta)
require.NoError(t, err, errTag)
require.Greater(t, streamMeta.NumberOfSegments, int64(0))
// set the number of segment to 0 turning the object to old-style
streamMeta.NumberOfSegments = 0
pointer.Metadata, err = pb.Marshal(streamMeta)
require.NoError(t, err, errTag)
err = satellite.Metainfo.Service.UnsynchronizedPut(ctx, key, pointer)
require.NoError(t, err, errTag)
// fixing old-style object should return no error and changed = true
changed, err = satellite.Metainfo.Service.FixOldStyleObject(ctx, key, tt.dryRun)
require.NoError(t, err, errTag)
require.True(t, changed)
pointer, err = satellite.Metainfo.Service.Get(ctx, key)
require.NoError(t, err, errTag)
// assert that the number of segments is set correctly for the fixed object
streamMeta = &pb.StreamMeta{}
err = pb.Unmarshal(pointer.Metadata, streamMeta)
require.NoError(t, err, errTag)
require.EqualValues(t, tt.expectedSegments, streamMeta.NumberOfSegments)
// fixing non-existing object should return no error and changed = false
err = satellite.Metainfo.Service.UnsynchronizedDelete(ctx, key)
require.NoError(t, err, errTag)
changed, err = satellite.Metainfo.Service.FixOldStyleObject(ctx, key, tt.dryRun)
require.NoError(t, err, errTag)
assert.False(t, changed, errTag)
}
})
}