6834c04539
We decided that we won't use seprate table for handling pending objects. We need to remove related code. https://github.com/storj/storj/issues/6421 Change-Id: I442b0f58da75409f725e08e2cd83d29ed4f91ec6
532 lines
16 KiB
Go
532 lines
16 KiB
Go
// Copyright (C) 2020 Storj Labs, Inc.
|
|
// See LICENSE for copying information.
|
|
|
|
package metabase_test
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/stretchr/testify/require"
|
|
"golang.org/x/sync/errgroup"
|
|
|
|
"storj.io/common/storj"
|
|
"storj.io/common/testcontext"
|
|
"storj.io/common/testrand"
|
|
"storj.io/common/uuid"
|
|
"storj.io/storj/satellite/metabase"
|
|
"storj.io/storj/satellite/metabase/metabasetest"
|
|
)
|
|
|
|
func TestDeleteBucketObjects(t *testing.T) {
|
|
metabasetest.Run(t, func(ctx *testcontext.Context, t *testing.T, db *metabase.DB) {
|
|
obj1 := metabasetest.RandObjectStream()
|
|
obj2 := metabasetest.RandObjectStream()
|
|
obj3 := metabasetest.RandObjectStream()
|
|
objX := metabasetest.RandObjectStream()
|
|
objY := metabasetest.RandObjectStream()
|
|
|
|
obj2.ProjectID, obj2.BucketName = obj1.ProjectID, obj1.BucketName
|
|
obj3.ProjectID, obj3.BucketName = obj1.ProjectID, obj1.BucketName
|
|
objX.ProjectID = obj1.ProjectID
|
|
objY.BucketName = obj1.BucketName
|
|
|
|
t.Run("invalid options", func(t *testing.T) {
|
|
defer metabasetest.DeleteAll{}.Check(ctx, t, db)
|
|
|
|
metabasetest.DeleteBucketObjects{
|
|
Opts: metabase.DeleteBucketObjects{
|
|
Bucket: metabase.BucketLocation{
|
|
ProjectID: uuid.UUID{},
|
|
BucketName: "",
|
|
},
|
|
},
|
|
ErrClass: &metabase.ErrInvalidRequest,
|
|
ErrText: "ProjectID missing",
|
|
}.Check(ctx, t, db)
|
|
|
|
metabasetest.DeleteBucketObjects{
|
|
Opts: metabase.DeleteBucketObjects{
|
|
Bucket: metabase.BucketLocation{
|
|
ProjectID: uuid.UUID{1},
|
|
BucketName: "",
|
|
},
|
|
},
|
|
ErrClass: &metabase.ErrInvalidRequest,
|
|
ErrText: "BucketName missing",
|
|
}.Check(ctx, t, db)
|
|
|
|
metabasetest.Verify{}.Check(ctx, t, db)
|
|
})
|
|
|
|
t.Run("empty bucket", func(t *testing.T) {
|
|
defer metabasetest.DeleteAll{}.Check(ctx, t, db)
|
|
|
|
metabasetest.DeleteBucketObjects{
|
|
Opts: metabase.DeleteBucketObjects{
|
|
Bucket: obj1.Location().Bucket(),
|
|
},
|
|
Deleted: 0,
|
|
}.Check(ctx, t, db)
|
|
|
|
metabasetest.Verify{}.Check(ctx, t, db)
|
|
})
|
|
|
|
t.Run("one object", func(t *testing.T) {
|
|
defer metabasetest.DeleteAll{}.Check(ctx, t, db)
|
|
|
|
metabasetest.CreateObject(ctx, t, db, obj1, 2)
|
|
|
|
metabasetest.DeleteBucketObjects{
|
|
Opts: metabase.DeleteBucketObjects{
|
|
Bucket: obj1.Location().Bucket(),
|
|
},
|
|
Deleted: 1,
|
|
}.Check(ctx, t, db)
|
|
|
|
metabasetest.Verify{}.Check(ctx, t, db)
|
|
})
|
|
|
|
t.Run("empty object", func(t *testing.T) {
|
|
defer metabasetest.DeleteAll{}.Check(ctx, t, db)
|
|
|
|
metabasetest.CreateObject(ctx, t, db, obj1, 0)
|
|
|
|
metabasetest.DeleteBucketObjects{
|
|
Opts: metabase.DeleteBucketObjects{
|
|
Bucket: obj1.Location().Bucket(),
|
|
},
|
|
Deleted: 1,
|
|
}.Check(ctx, t, db)
|
|
|
|
metabasetest.Verify{}.Check(ctx, t, db)
|
|
})
|
|
|
|
t.Run("three objects", func(t *testing.T) {
|
|
defer metabasetest.DeleteAll{}.Check(ctx, t, db)
|
|
|
|
metabasetest.CreateObject(ctx, t, db, obj1, 2)
|
|
metabasetest.CreateObject(ctx, t, db, obj2, 2)
|
|
metabasetest.CreateObject(ctx, t, db, obj3, 2)
|
|
|
|
metabasetest.DeleteBucketObjects{
|
|
Opts: metabase.DeleteBucketObjects{
|
|
Bucket: obj1.Location().Bucket(),
|
|
BatchSize: 2,
|
|
},
|
|
Deleted: 3,
|
|
}.Check(ctx, t, db)
|
|
|
|
metabasetest.Verify{}.Check(ctx, t, db)
|
|
})
|
|
|
|
t.Run("don't delete non-exact match", func(t *testing.T) {
|
|
defer metabasetest.DeleteAll{}.Check(ctx, t, db)
|
|
|
|
metabasetest.CreateObject(ctx, t, db, obj1, 1)
|
|
metabasetest.CreateObject(ctx, t, db, objX, 1)
|
|
metabasetest.CreateObject(ctx, t, db, objY, 1)
|
|
now := time.Now()
|
|
|
|
metabasetest.DeleteBucketObjects{
|
|
Opts: metabase.DeleteBucketObjects{
|
|
Bucket: obj1.Location().Bucket(),
|
|
},
|
|
Deleted: 1,
|
|
}.Check(ctx, t, db)
|
|
|
|
metabasetest.Verify{
|
|
Objects: []metabase.RawObject{
|
|
{
|
|
ObjectStream: objX,
|
|
CreatedAt: now,
|
|
Status: metabase.CommittedUnversioned,
|
|
SegmentCount: 1,
|
|
|
|
TotalPlainSize: 512,
|
|
TotalEncryptedSize: 1024,
|
|
FixedSegmentSize: 512,
|
|
Encryption: metabasetest.DefaultEncryption,
|
|
},
|
|
{
|
|
ObjectStream: objY,
|
|
CreatedAt: now,
|
|
Status: metabase.CommittedUnversioned,
|
|
SegmentCount: 1,
|
|
|
|
TotalPlainSize: 512,
|
|
TotalEncryptedSize: 1024,
|
|
FixedSegmentSize: 512,
|
|
Encryption: metabasetest.DefaultEncryption,
|
|
},
|
|
},
|
|
Segments: []metabase.RawSegment{
|
|
{
|
|
StreamID: objX.StreamID,
|
|
Position: metabase.SegmentPosition{Part: 0, Index: 0},
|
|
CreatedAt: now,
|
|
|
|
RootPieceID: storj.PieceID{1},
|
|
Pieces: metabase.Pieces{{Number: 0, StorageNode: storj.NodeID{2}}},
|
|
EncryptedKey: []byte{3},
|
|
EncryptedKeyNonce: []byte{4},
|
|
EncryptedETag: []byte{5},
|
|
|
|
EncryptedSize: 1024,
|
|
PlainSize: 512,
|
|
PlainOffset: 0,
|
|
|
|
Redundancy: metabasetest.DefaultRedundancy,
|
|
},
|
|
{
|
|
StreamID: objY.StreamID,
|
|
Position: metabase.SegmentPosition{Part: 0, Index: 0},
|
|
CreatedAt: now,
|
|
|
|
RootPieceID: storj.PieceID{1},
|
|
Pieces: metabase.Pieces{{Number: 0, StorageNode: storj.NodeID{2}}},
|
|
EncryptedKey: []byte{3},
|
|
EncryptedKeyNonce: []byte{4},
|
|
EncryptedETag: []byte{5},
|
|
|
|
EncryptedSize: 1024,
|
|
PlainSize: 512,
|
|
PlainOffset: 0,
|
|
|
|
Redundancy: metabasetest.DefaultRedundancy,
|
|
},
|
|
},
|
|
}.Check(ctx, t, db)
|
|
})
|
|
|
|
t.Run("object with multiple segments", func(t *testing.T) {
|
|
defer metabasetest.DeleteAll{}.Check(ctx, t, db)
|
|
|
|
metabasetest.CreateObject(ctx, t, db, obj1, 37)
|
|
|
|
metabasetest.DeleteBucketObjects{
|
|
Opts: metabase.DeleteBucketObjects{
|
|
Bucket: obj1.Location().Bucket(),
|
|
BatchSize: 2,
|
|
},
|
|
Deleted: 1,
|
|
}.Check(ctx, t, db)
|
|
|
|
metabasetest.Verify{}.Check(ctx, t, db)
|
|
})
|
|
|
|
t.Run("multiple objects", func(t *testing.T) {
|
|
defer metabasetest.DeleteAll{}.Check(ctx, t, db)
|
|
|
|
root := metabasetest.RandObjectStream()
|
|
for i := 0; i < 5; i++ {
|
|
obj := metabasetest.RandObjectStream()
|
|
obj.ProjectID = root.ProjectID
|
|
obj.BucketName = root.BucketName
|
|
metabasetest.CreateObject(ctx, t, db, obj, 5)
|
|
}
|
|
|
|
metabasetest.DeleteBucketObjects{
|
|
Opts: metabase.DeleteBucketObjects{
|
|
Bucket: root.Location().Bucket(),
|
|
BatchSize: 1,
|
|
},
|
|
Deleted: 5,
|
|
}.Check(ctx, t, db)
|
|
|
|
metabasetest.Verify{}.Check(ctx, t, db)
|
|
})
|
|
})
|
|
}
|
|
|
|
func TestDeleteBucketObjectsParallel(t *testing.T) {
|
|
metabasetest.Run(t, func(ctx *testcontext.Context, t *testing.T, db *metabase.DB) {
|
|
defer metabasetest.DeleteAll{}.Check(ctx, t, db)
|
|
|
|
root := metabasetest.RandObjectStream()
|
|
for i := 0; i < 5; i++ {
|
|
obj := metabasetest.RandObjectStream()
|
|
obj.ProjectID = root.ProjectID
|
|
obj.BucketName = root.BucketName
|
|
metabasetest.CreateObject(ctx, t, db, obj, 50)
|
|
}
|
|
|
|
objects, err := db.TestingAllObjects(ctx)
|
|
require.NoError(t, err)
|
|
require.Equal(t, 5, len(objects))
|
|
|
|
var errgroup errgroup.Group
|
|
for i := 0; i < 3; i++ {
|
|
errgroup.Go(func() error {
|
|
_, err := db.DeleteBucketObjects(ctx, metabase.DeleteBucketObjects{
|
|
Bucket: root.Location().Bucket(),
|
|
BatchSize: 2,
|
|
})
|
|
return err
|
|
})
|
|
}
|
|
require.NoError(t, errgroup.Wait())
|
|
|
|
metabasetest.Verify{}.Check(ctx, t, db)
|
|
})
|
|
}
|
|
|
|
func TestDeleteBucketObjectsCancel(t *testing.T) {
|
|
metabasetest.Run(t, func(ctx *testcontext.Context, t *testing.T, db *metabase.DB) {
|
|
defer metabasetest.DeleteAll{}.Check(ctx, t, db)
|
|
|
|
object := metabasetest.CreateObject(ctx, t, db, metabasetest.RandObjectStream(), 1)
|
|
|
|
testCtx, cancel := context.WithCancel(ctx)
|
|
cancel()
|
|
_, err := db.DeleteBucketObjects(testCtx, metabase.DeleteBucketObjects{
|
|
Bucket: object.Location().Bucket(),
|
|
BatchSize: 2,
|
|
})
|
|
require.Error(t, err)
|
|
|
|
metabasetest.Verify{
|
|
Objects: []metabase.RawObject{metabase.RawObject(object)},
|
|
Segments: []metabase.RawSegment{
|
|
metabasetest.DefaultRawSegment(object.ObjectStream, metabase.SegmentPosition{}),
|
|
},
|
|
}.Check(ctx, t, db)
|
|
})
|
|
}
|
|
|
|
func TestDeleteBucketWithCopies(t *testing.T) {
|
|
metabasetest.Run(t, func(ctx *testcontext.Context, t *testing.T, db *metabase.DB) {
|
|
for _, numberOfSegments := range []int{0, 1, 3} {
|
|
t.Run(fmt.Sprintf("%d segments", numberOfSegments), func(t *testing.T) {
|
|
t.Run("delete bucket with copy", func(t *testing.T) {
|
|
defer metabasetest.DeleteAll{}.Check(ctx, t, db)
|
|
originalObjStream := metabasetest.RandObjectStream()
|
|
originalObjStream.BucketName = "original-bucket"
|
|
|
|
originalObj, originalSegments := metabasetest.CreateTestObject{
|
|
CommitObject: &metabase.CommitObject{
|
|
ObjectStream: originalObjStream,
|
|
EncryptedMetadata: testrand.Bytes(64),
|
|
EncryptedMetadataNonce: testrand.Nonce().Bytes(),
|
|
EncryptedMetadataEncryptedKey: testrand.Bytes(265),
|
|
},
|
|
}.Run(ctx, t, db, originalObjStream, byte(numberOfSegments))
|
|
|
|
copyObjectStream := metabasetest.RandObjectStream()
|
|
copyObjectStream.ProjectID = originalObjStream.ProjectID
|
|
copyObjectStream.BucketName = "copy-bucket"
|
|
|
|
metabasetest.CreateObjectCopy{
|
|
OriginalObject: originalObj,
|
|
CopyObjectStream: ©ObjectStream,
|
|
}.Run(ctx, t, db)
|
|
|
|
_, err := db.DeleteBucketObjects(ctx, metabase.DeleteBucketObjects{
|
|
Bucket: metabase.BucketLocation{
|
|
ProjectID: originalObjStream.ProjectID,
|
|
BucketName: "copy-bucket",
|
|
},
|
|
BatchSize: 2,
|
|
})
|
|
require.NoError(t, err)
|
|
|
|
// Verify that we are back at the original single object
|
|
metabasetest.Verify{
|
|
Objects: []metabase.RawObject{
|
|
metabase.RawObject(originalObj),
|
|
},
|
|
Segments: metabasetest.SegmentsToRaw(originalSegments),
|
|
}.Check(ctx, t, db)
|
|
})
|
|
|
|
t.Run("delete bucket with ancestor", func(t *testing.T) {
|
|
defer metabasetest.DeleteAll{}.Check(ctx, t, db)
|
|
originalObjStream := metabasetest.RandObjectStream()
|
|
originalObjStream.BucketName = "original-bucket"
|
|
|
|
originalObj, originalSegments := metabasetest.CreateTestObject{
|
|
CommitObject: &metabase.CommitObject{
|
|
ObjectStream: originalObjStream,
|
|
EncryptedMetadata: testrand.Bytes(64),
|
|
EncryptedMetadataNonce: testrand.Nonce().Bytes(),
|
|
EncryptedMetadataEncryptedKey: testrand.Bytes(265),
|
|
},
|
|
}.Run(ctx, t, db, originalObjStream, byte(numberOfSegments))
|
|
|
|
copyObjectStream := metabasetest.RandObjectStream()
|
|
copyObjectStream.ProjectID = originalObjStream.ProjectID
|
|
copyObjectStream.BucketName = "copy-bucket"
|
|
|
|
copyObj, _, copySegments := metabasetest.CreateObjectCopy{
|
|
OriginalObject: originalObj,
|
|
CopyObjectStream: ©ObjectStream,
|
|
}.Run(ctx, t, db)
|
|
|
|
_, err := db.DeleteBucketObjects(ctx, metabase.DeleteBucketObjects{
|
|
Bucket: metabase.BucketLocation{
|
|
ProjectID: originalObjStream.ProjectID,
|
|
BucketName: "original-bucket",
|
|
},
|
|
BatchSize: 2,
|
|
})
|
|
require.NoError(t, err)
|
|
|
|
for i := range copySegments {
|
|
copySegments[i].Pieces = originalSegments[i].Pieces
|
|
}
|
|
|
|
// Verify that we are back at the original single object
|
|
metabasetest.Verify{
|
|
Objects: []metabase.RawObject{
|
|
metabase.RawObject(copyObj),
|
|
},
|
|
Segments: copySegments,
|
|
}.Check(ctx, t, db)
|
|
})
|
|
|
|
t.Run("delete bucket which has one ancestor and one copy", func(t *testing.T) {
|
|
defer metabasetest.DeleteAll{}.Check(ctx, t, db)
|
|
originalObjStream1 := metabasetest.RandObjectStream()
|
|
originalObjStream1.BucketName = "bucket1"
|
|
|
|
projectID := originalObjStream1.ProjectID
|
|
|
|
originalObjStream2 := metabasetest.RandObjectStream()
|
|
originalObjStream2.ProjectID = projectID
|
|
originalObjStream2.BucketName = "bucket2"
|
|
|
|
originalObj1, originalSegments1 := metabasetest.CreateTestObject{
|
|
CommitObject: &metabase.CommitObject{
|
|
ObjectStream: originalObjStream1,
|
|
},
|
|
}.Run(ctx, t, db, originalObjStream1, byte(numberOfSegments))
|
|
|
|
originalObj2, originalSegments2 := metabasetest.CreateTestObject{
|
|
CommitObject: &metabase.CommitObject{
|
|
ObjectStream: originalObjStream2,
|
|
},
|
|
}.Run(ctx, t, db, originalObjStream2, byte(numberOfSegments))
|
|
|
|
copyObjectStream1 := metabasetest.RandObjectStream()
|
|
copyObjectStream1.ProjectID = projectID
|
|
copyObjectStream1.BucketName = "bucket2" // copy from bucket 1 to bucket 2
|
|
|
|
copyObjectStream2 := metabasetest.RandObjectStream()
|
|
copyObjectStream2.ProjectID = projectID
|
|
copyObjectStream2.BucketName = "bucket1" // copy from bucket 2 to bucket 1
|
|
|
|
metabasetest.CreateObjectCopy{
|
|
OriginalObject: originalObj1,
|
|
CopyObjectStream: ©ObjectStream1,
|
|
}.Run(ctx, t, db)
|
|
|
|
copyObj2, _, copySegments2 := metabasetest.CreateObjectCopy{
|
|
OriginalObject: originalObj2,
|
|
CopyObjectStream: ©ObjectStream2,
|
|
}.Run(ctx, t, db)
|
|
|
|
// done preparing, delete bucket 1
|
|
_, err := db.DeleteBucketObjects(ctx, metabase.DeleteBucketObjects{
|
|
Bucket: metabase.BucketLocation{
|
|
ProjectID: projectID,
|
|
BucketName: "bucket2",
|
|
},
|
|
BatchSize: 2,
|
|
})
|
|
require.NoError(t, err)
|
|
|
|
// Prepare for check.
|
|
// obj1 is the same as before, copyObj2 should now be the original
|
|
for i := range copySegments2 {
|
|
copySegments2[i].Pieces = originalSegments2[i].Pieces
|
|
}
|
|
|
|
metabasetest.Verify{
|
|
Objects: []metabase.RawObject{
|
|
metabase.RawObject(originalObj1),
|
|
metabase.RawObject(copyObj2),
|
|
},
|
|
Segments: append(copySegments2, metabasetest.SegmentsToRaw(originalSegments1)...),
|
|
}.Check(ctx, t, db)
|
|
})
|
|
|
|
t.Run("delete bucket which has one ancestor and one copy with duplicate metadata", func(t *testing.T) {
|
|
defer metabasetest.DeleteAll{}.Check(ctx, t, db)
|
|
originalObjStream1 := metabasetest.RandObjectStream()
|
|
originalObjStream1.BucketName = "bucket1"
|
|
|
|
projectID := originalObjStream1.ProjectID
|
|
|
|
originalObjStream2 := metabasetest.RandObjectStream()
|
|
originalObjStream2.ProjectID = projectID
|
|
originalObjStream2.BucketName = "bucket2"
|
|
|
|
originalObj1, originalSegments1 := metabasetest.CreateTestObject{
|
|
CommitObject: &metabase.CommitObject{
|
|
ObjectStream: originalObjStream1,
|
|
},
|
|
}.Run(ctx, t, db, originalObjStream1, byte(numberOfSegments))
|
|
|
|
originalObj2, originalSegments2 := metabasetest.CreateTestObject{
|
|
CommitObject: &metabase.CommitObject{
|
|
ObjectStream: originalObjStream2,
|
|
},
|
|
}.Run(ctx, t, db, originalObjStream2, byte(numberOfSegments))
|
|
|
|
copyObjectStream1 := metabasetest.RandObjectStream()
|
|
copyObjectStream1.ProjectID = projectID
|
|
copyObjectStream1.BucketName = "bucket2" // copy from bucket 1 to bucket 2
|
|
|
|
copyObjectStream2 := metabasetest.RandObjectStream()
|
|
copyObjectStream2.ProjectID = projectID
|
|
copyObjectStream2.BucketName = "bucket1" // copy from bucket 2 to bucket 1
|
|
|
|
metabasetest.CreateObjectCopy{
|
|
OriginalObject: originalObj1,
|
|
CopyObjectStream: ©ObjectStream1,
|
|
}.Run(ctx, t, db)
|
|
|
|
copyObj2, _, copySegments2 := metabasetest.CreateObjectCopy{
|
|
OriginalObject: originalObj2,
|
|
CopyObjectStream: ©ObjectStream2,
|
|
}.Run(ctx, t, db)
|
|
|
|
// done preparing, delete bucket 1
|
|
_, err := db.DeleteBucketObjects(ctx, metabase.DeleteBucketObjects{
|
|
Bucket: metabase.BucketLocation{
|
|
ProjectID: projectID,
|
|
BucketName: "bucket2",
|
|
},
|
|
BatchSize: 2,
|
|
})
|
|
require.NoError(t, err)
|
|
|
|
// Prepare for check.
|
|
// obj1 is the same as before, copyObj2 should now be the original
|
|
for i := range copySegments2 {
|
|
copySegments2[i].Pieces = originalSegments2[i].Pieces
|
|
}
|
|
|
|
metabasetest.Verify{
|
|
Objects: []metabase.RawObject{
|
|
metabase.RawObject(originalObj1),
|
|
metabase.RawObject(copyObj2),
|
|
},
|
|
Segments: append(copySegments2, metabasetest.SegmentsToRaw(originalSegments1)...),
|
|
}.Check(ctx, t, db)
|
|
})
|
|
|
|
// TODO: check that DeletePieces callback is called with the correct arguments
|
|
|
|
// scenario: delete original bucket with 2 copies
|
|
|
|
// scenario: delete copy bucket with 2 copies
|
|
|
|
// scenario: delete bucket with 2 internal copies
|
|
})
|
|
}
|
|
})
|
|
}
|