2019-03-30 11:21:49 +00:00
|
|
|
// Copyright (C) 2019 Storj Labs, Inc.
|
|
|
|
// See LICENSE for copying information.
|
|
|
|
|
|
|
|
package metainfo_test
|
|
|
|
|
|
|
|
import (
|
2020-03-30 10:08:02 +01:00
|
|
|
"errors"
|
2020-06-30 22:49:29 +01:00
|
|
|
"fmt"
|
2019-03-30 11:21:49 +00:00
|
|
|
"sort"
|
2019-07-16 11:39:23 +01:00
|
|
|
"strconv"
|
2019-03-30 11:21:49 +00:00
|
|
|
"testing"
|
|
|
|
"time"
|
|
|
|
|
|
|
|
"github.com/stretchr/testify/assert"
|
|
|
|
"github.com/stretchr/testify/require"
|
2020-03-16 08:55:52 +00:00
|
|
|
"github.com/zeebo/errs"
|
2019-06-21 19:15:58 +01:00
|
|
|
"go.uber.org/zap"
|
2019-03-30 11:21:49 +00:00
|
|
|
|
2019-12-27 11:48:47 +00:00
|
|
|
"storj.io/common/errs2"
|
|
|
|
"storj.io/common/identity"
|
|
|
|
"storj.io/common/macaroon"
|
|
|
|
"storj.io/common/memory"
|
|
|
|
"storj.io/common/pb"
|
|
|
|
"storj.io/common/rpc/rpcstatus"
|
|
|
|
"storj.io/common/signing"
|
|
|
|
"storj.io/common/storj"
|
|
|
|
"storj.io/common/testcontext"
|
|
|
|
"storj.io/common/testrand"
|
2019-11-14 19:46:15 +00:00
|
|
|
"storj.io/storj/private/testplanet"
|
2019-06-21 19:15:58 +01:00
|
|
|
"storj.io/storj/satellite"
|
2020-04-09 09:19:16 +01:00
|
|
|
satMetainfo "storj.io/storj/satellite/metainfo"
|
2020-03-30 10:08:02 +01:00
|
|
|
"storj.io/uplink"
|
2020-02-21 14:07:29 +00:00
|
|
|
"storj.io/uplink/private/metainfo"
|
2020-04-09 09:19:16 +01:00
|
|
|
"storj.io/uplink/private/storage/meta"
|
2020-05-26 09:05:43 +01:00
|
|
|
"storj.io/uplink/private/testuplink"
|
2019-03-30 11:21:49 +00:00
|
|
|
)
|
|
|
|
|
2020-06-30 22:49:29 +01:00
|
|
|
func TestMaxOutBuckets(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
2020-08-26 16:19:09 +01:00
|
|
|
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 1,
|
2020-06-30 22:49:29 +01:00
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
limit := planet.Satellites[0].Config.Metainfo.ProjectLimits.MaxBuckets
|
|
|
|
for i := 1; i <= limit; i++ {
|
|
|
|
name := "test" + strconv.Itoa(i)
|
|
|
|
err := planet.Uplinks[0].CreateBucket(ctx, planet.Satellites[0], name)
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
err := planet.Uplinks[0].CreateBucket(ctx, planet.Satellites[0], fmt.Sprintf("test%d", limit+1))
|
|
|
|
require.EqualError(t, err, fmt.Sprintf("uplink: bucket: metainfo error: number of allocated buckets (%d) exceeded", limit))
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2020-06-10 15:10:44 +01:00
|
|
|
func TestRevokeAccess(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
accessIssuer := planet.Uplinks[0].Access[planet.Satellites[0].ID()]
|
|
|
|
accessUser1, err := accessIssuer.Share(uplink.Permission{
|
|
|
|
AllowDownload: true,
|
|
|
|
AllowUpload: true,
|
|
|
|
AllowList: true,
|
|
|
|
AllowDelete: true,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
accessUser2, err := accessUser1.Share(uplink.Permission{
|
|
|
|
AllowDownload: true,
|
|
|
|
AllowUpload: true,
|
|
|
|
AllowList: true,
|
|
|
|
AllowDelete: true,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
projectUser2, err := uplink.OpenProject(ctx, accessUser2)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(projectUser2.Close)
|
|
|
|
|
|
|
|
// confirm that we can create a bucket
|
|
|
|
_, err = projectUser2.CreateBucket(ctx, "bob")
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// we shouldn't be allowed to revoke ourselves or our parent
|
|
|
|
err = projectUser2.RevokeAccess(ctx, accessUser2)
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
|
|
|
err = projectUser2.RevokeAccess(ctx, accessUser1)
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
|
|
|
|
|
|
|
projectIssuer, err := uplink.OpenProject(ctx, accessIssuer)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(projectIssuer.Close)
|
|
|
|
|
|
|
|
projectUser1, err := uplink.OpenProject(ctx, accessUser1)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(projectUser1.Close)
|
|
|
|
|
|
|
|
// I should be able to revoke with accessIssuer
|
|
|
|
err = projectIssuer.RevokeAccess(ctx, accessUser1)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// should no longer be able to create bucket with access 2 or 3
|
|
|
|
_, err = projectUser2.CreateBucket(ctx, "bob1")
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
|
|
|
_, err = projectUser1.CreateBucket(ctx, "bob1")
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestRevokeMacaroon(t *testing.T) {
|
2020-06-03 14:51:02 +01:00
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
|
|
|
|
// I want the api key for the single satellite in this test
|
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
|
|
|
|
client, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(client.Close)
|
|
|
|
|
|
|
|
// Sanity check: it should work before revoke
|
|
|
|
_, err = client.ListBuckets(ctx, metainfo.ListBucketsParams{
|
|
|
|
ListOpts: storj.BucketListOptions{
|
|
|
|
Cursor: "",
|
|
|
|
Direction: storj.Forward,
|
|
|
|
Limit: 10,
|
|
|
|
},
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
err = planet.Satellites[0].API.DB.Revocation().Revoke(ctx, apiKey.Tail(), []byte("apikey"))
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
_, err = client.ListBuckets(ctx, metainfo.ListBucketsParams{})
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
|
|
|
|
|
|
|
_, err = client.BeginObject(ctx, metainfo.BeginObjectParams{})
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
|
|
|
|
2020-07-21 08:34:00 +01:00
|
|
|
_, err = client.BeginDeleteObject(ctx, metainfo.BeginDeleteObjectParams{})
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
2020-06-03 14:51:02 +01:00
|
|
|
|
|
|
|
_, err = client.ListBuckets(ctx, metainfo.ListBucketsParams{})
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
|
|
|
|
|
|
|
_, _, err = client.ListObjects(ctx, metainfo.ListObjectsParams{})
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
|
|
|
|
|
|
|
_, err = client.CreateBucket(ctx, metainfo.CreateBucketParams{})
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
|
|
|
|
|
|
|
_, err = client.DeleteBucket(ctx, metainfo.DeleteBucketParams{})
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
|
|
|
|
2020-07-21 08:34:00 +01:00
|
|
|
_, err = client.BeginDeleteObject(ctx, metainfo.BeginDeleteObjectParams{})
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
2020-06-03 14:51:02 +01:00
|
|
|
|
|
|
|
_, err = client.GetBucket(ctx, metainfo.GetBucketParams{})
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
|
|
|
|
|
|
|
_, err = client.GetObject(ctx, metainfo.GetObjectParams{})
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
|
|
|
|
|
|
|
_, err = client.GetProjectInfo(ctx)
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
|
|
|
|
|
|
|
signer := signing.SignerFromFullIdentity(planet.Satellites[0].Identity)
|
|
|
|
satStreamID := &pb.SatStreamID{
|
|
|
|
CreationDate: time.Now(),
|
|
|
|
}
|
|
|
|
signedStreamID, err := signing.SignStreamID(ctx, signer, satStreamID)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
encodedStreamID, err := pb.Marshal(signedStreamID)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
err = client.CommitObject(ctx, metainfo.CommitObjectParams{StreamID: encodedStreamID})
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
|
|
|
|
|
|
|
err = client.FinishDeleteObject(ctx, metainfo.FinishDeleteObjectParams{StreamID: encodedStreamID})
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
|
|
|
|
|
|
|
_, _, _, err = client.BeginSegment(ctx, metainfo.BeginSegmentParams{StreamID: encodedStreamID})
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
|
|
|
|
|
|
|
_, _, _, err = client.BeginDeleteSegment(ctx, metainfo.BeginDeleteSegmentParams{StreamID: encodedStreamID})
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
|
|
|
|
|
|
|
err = client.MakeInlineSegment(ctx, metainfo.MakeInlineSegmentParams{StreamID: encodedStreamID})
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
|
|
|
|
|
|
|
_, _, err = client.ListSegments(ctx, metainfo.ListSegmentsParams{StreamID: encodedStreamID})
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
|
|
|
|
|
|
|
_, _, err = client.DownloadSegment(ctx, metainfo.DownloadSegmentParams{StreamID: encodedStreamID})
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
|
|
|
|
|
|
|
// these methods needs SegmentID
|
|
|
|
|
|
|
|
signedSegmentID, err := signing.SignSegmentID(ctx, signer, &pb.SatSegmentID{
|
|
|
|
StreamId: satStreamID,
|
|
|
|
CreationDate: time.Now(),
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
encodedSegmentID, err := pb.Marshal(signedSegmentID)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
segmentID, err := storj.SegmentIDFromBytes(encodedSegmentID)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
err = client.CommitSegment(ctx, metainfo.CommitSegmentParams{SegmentID: segmentID})
|
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.PermissionDenied))
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2020-01-03 09:27:10 +00:00
|
|
|
func TestInvalidAPIKey(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
throwawayKey, err := macaroon.NewAPIKey([]byte("secret"))
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
for _, invalidAPIKey := range []string{"", "invalid", "testKey"} {
|
|
|
|
client, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], throwawayKey)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(client.Close)
|
|
|
|
|
|
|
|
client.SetRawAPIKey([]byte(invalidAPIKey))
|
|
|
|
|
|
|
|
_, err = client.BeginObject(ctx, metainfo.BeginObjectParams{})
|
2020-03-10 09:58:14 +00:00
|
|
|
assertInvalidArgument(t, err, false)
|
2020-01-03 09:27:10 +00:00
|
|
|
|
2020-07-21 08:34:00 +01:00
|
|
|
_, err = client.BeginDeleteObject(ctx, metainfo.BeginDeleteObjectParams{})
|
|
|
|
assertInvalidArgument(t, err, false)
|
2020-01-03 09:27:10 +00:00
|
|
|
|
|
|
|
_, err = client.ListBuckets(ctx, metainfo.ListBucketsParams{})
|
2020-03-10 09:58:14 +00:00
|
|
|
assertInvalidArgument(t, err, false)
|
2020-01-03 09:27:10 +00:00
|
|
|
|
|
|
|
_, _, err = client.ListObjects(ctx, metainfo.ListObjectsParams{})
|
2020-03-10 09:58:14 +00:00
|
|
|
assertInvalidArgument(t, err, false)
|
2020-01-03 09:27:10 +00:00
|
|
|
|
|
|
|
_, err = client.CreateBucket(ctx, metainfo.CreateBucketParams{})
|
2020-03-10 09:58:14 +00:00
|
|
|
assertInvalidArgument(t, err, false)
|
2020-01-03 09:27:10 +00:00
|
|
|
|
2020-03-18 14:05:24 +00:00
|
|
|
_, err = client.DeleteBucket(ctx, metainfo.DeleteBucketParams{})
|
2020-03-10 09:58:14 +00:00
|
|
|
assertInvalidArgument(t, err, false)
|
2020-01-03 09:27:10 +00:00
|
|
|
|
2020-07-21 08:34:00 +01:00
|
|
|
_, err = client.BeginDeleteObject(ctx, metainfo.BeginDeleteObjectParams{})
|
|
|
|
assertInvalidArgument(t, err, false)
|
2020-01-03 09:27:10 +00:00
|
|
|
|
|
|
|
_, err = client.GetBucket(ctx, metainfo.GetBucketParams{})
|
2020-03-10 09:58:14 +00:00
|
|
|
assertInvalidArgument(t, err, false)
|
2020-01-03 09:27:10 +00:00
|
|
|
|
|
|
|
_, err = client.GetObject(ctx, metainfo.GetObjectParams{})
|
2020-03-10 09:58:14 +00:00
|
|
|
assertInvalidArgument(t, err, false)
|
2020-01-03 09:27:10 +00:00
|
|
|
|
|
|
|
_, err = client.GetProjectInfo(ctx)
|
2020-03-10 09:58:14 +00:00
|
|
|
assertInvalidArgument(t, err, false)
|
2020-01-03 09:27:10 +00:00
|
|
|
|
|
|
|
// these methods needs StreamID to do authentication
|
|
|
|
|
|
|
|
signer := signing.SignerFromFullIdentity(planet.Satellites[0].Identity)
|
|
|
|
satStreamID := &pb.SatStreamID{
|
|
|
|
CreationDate: time.Now(),
|
|
|
|
}
|
|
|
|
signedStreamID, err := signing.SignStreamID(ctx, signer, satStreamID)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2020-04-08 13:08:57 +01:00
|
|
|
encodedStreamID, err := pb.Marshal(signedStreamID)
|
2020-01-03 09:27:10 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
streamID, err := storj.StreamIDFromBytes(encodedStreamID)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2020-06-03 14:51:02 +01:00
|
|
|
err = client.CommitObject(ctx, metainfo.CommitObjectParams{StreamID: streamID})
|
|
|
|
assertInvalidArgument(t, err, false)
|
|
|
|
|
|
|
|
err = client.FinishDeleteObject(ctx, metainfo.FinishDeleteObjectParams{StreamID: streamID})
|
|
|
|
assertInvalidArgument(t, err, false)
|
|
|
|
|
2020-01-03 09:27:10 +00:00
|
|
|
_, _, _, err = client.BeginSegment(ctx, metainfo.BeginSegmentParams{StreamID: streamID})
|
2020-03-10 09:58:14 +00:00
|
|
|
assertInvalidArgument(t, err, false)
|
2020-01-03 09:27:10 +00:00
|
|
|
|
|
|
|
_, _, _, err = client.BeginDeleteSegment(ctx, metainfo.BeginDeleteSegmentParams{StreamID: streamID})
|
2020-03-10 09:58:14 +00:00
|
|
|
assertInvalidArgument(t, err, false)
|
2020-01-03 09:27:10 +00:00
|
|
|
|
|
|
|
err = client.MakeInlineSegment(ctx, metainfo.MakeInlineSegmentParams{StreamID: streamID})
|
2020-03-10 09:58:14 +00:00
|
|
|
assertInvalidArgument(t, err, false)
|
2020-01-03 09:27:10 +00:00
|
|
|
|
|
|
|
_, _, err = client.ListSegments(ctx, metainfo.ListSegmentsParams{StreamID: streamID})
|
2020-03-10 09:58:14 +00:00
|
|
|
assertInvalidArgument(t, err, false)
|
2020-01-03 09:27:10 +00:00
|
|
|
|
|
|
|
_, _, err = client.DownloadSegment(ctx, metainfo.DownloadSegmentParams{StreamID: streamID})
|
2020-03-10 09:58:14 +00:00
|
|
|
assertInvalidArgument(t, err, false)
|
2020-01-03 09:27:10 +00:00
|
|
|
|
|
|
|
// these methods needs SegmentID
|
|
|
|
|
|
|
|
signedSegmentID, err := signing.SignSegmentID(ctx, signer, &pb.SatSegmentID{
|
|
|
|
StreamId: satStreamID,
|
|
|
|
CreationDate: time.Now(),
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2020-04-08 13:08:57 +01:00
|
|
|
encodedSegmentID, err := pb.Marshal(signedSegmentID)
|
2020-01-03 09:27:10 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
segmentID, err := storj.SegmentIDFromBytes(encodedSegmentID)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
err = client.CommitSegment(ctx, metainfo.CommitSegmentParams{SegmentID: segmentID})
|
2020-03-10 09:58:14 +00:00
|
|
|
assertInvalidArgument(t, err, false)
|
2020-01-03 09:27:10 +00:00
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2020-03-10 09:58:14 +00:00
|
|
|
func assertInvalidArgument(t *testing.T, err error, allowed bool) {
|
2019-03-30 11:21:49 +00:00
|
|
|
t.Helper()
|
|
|
|
|
2019-05-24 17:51:27 +01:00
|
|
|
// If it's allowed, we allow any non-unauthenticated error because
|
|
|
|
// some calls error after authentication checks.
|
2019-09-19 05:46:39 +01:00
|
|
|
if !allowed {
|
2020-03-10 09:58:14 +00:00
|
|
|
assert.True(t, errs2.IsRPC(err, rpcstatus.InvalidArgument))
|
2019-03-30 11:21:49 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestServiceList(t *testing.T) {
|
2019-12-06 18:03:22 +00:00
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
2019-03-30 11:21:49 +00:00
|
|
|
|
2019-12-06 18:03:22 +00:00
|
|
|
items := []struct {
|
|
|
|
Key string
|
|
|
|
Value []byte
|
|
|
|
}{
|
|
|
|
{Key: "sample.😶", Value: []byte{1}},
|
|
|
|
{Key: "müsic", Value: []byte{2}},
|
|
|
|
{Key: "müsic/söng1.mp3", Value: []byte{3}},
|
|
|
|
{Key: "müsic/söng2.mp3", Value: []byte{4}},
|
|
|
|
{Key: "müsic/album/söng3.mp3", Value: []byte{5}},
|
|
|
|
{Key: "müsic/söng4.mp3", Value: []byte{6}},
|
|
|
|
{Key: "ビデオ/movie.mkv", Value: []byte{7}},
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, item := range items {
|
|
|
|
err := planet.Uplinks[0].Upload(ctx, planet.Satellites[0], "testbucket", item.Key, item.Value)
|
|
|
|
assert.NoError(t, err)
|
|
|
|
}
|
|
|
|
|
2020-05-26 09:05:43 +01:00
|
|
|
project, err := planet.Uplinks[0].GetProject(ctx, planet.Satellites[0])
|
2019-12-06 18:03:22 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(project.Close)
|
2020-05-26 09:05:43 +01:00
|
|
|
|
|
|
|
objects := project.ListObjects(ctx, "testbucket", &uplink.ListObjectsOptions{
|
|
|
|
Recursive: true,
|
|
|
|
})
|
|
|
|
|
|
|
|
listItems := make([]*uplink.Object, 0)
|
|
|
|
for objects.Next() {
|
|
|
|
listItems = append(listItems, objects.Item())
|
|
|
|
}
|
|
|
|
require.NoError(t, objects.Err())
|
2019-03-30 11:21:49 +00:00
|
|
|
|
2019-12-06 18:03:22 +00:00
|
|
|
expected := []storj.Object{
|
|
|
|
{Path: "müsic"},
|
|
|
|
{Path: "müsic/album/söng3.mp3"},
|
|
|
|
{Path: "müsic/söng1.mp3"},
|
|
|
|
{Path: "müsic/söng2.mp3"},
|
|
|
|
{Path: "müsic/söng4.mp3"},
|
|
|
|
{Path: "sample.😶"},
|
|
|
|
{Path: "ビデオ/movie.mkv"},
|
|
|
|
}
|
|
|
|
|
2020-05-26 09:05:43 +01:00
|
|
|
require.Equal(t, len(expected), len(listItems))
|
|
|
|
sort.Slice(listItems, func(i, k int) bool {
|
|
|
|
return listItems[i].Key < listItems[k].Key
|
2019-12-06 18:03:22 +00:00
|
|
|
})
|
|
|
|
for i, item := range expected {
|
2020-05-26 09:05:43 +01:00
|
|
|
require.Equal(t, item.Path, listItems[i].Key)
|
|
|
|
require.Equal(t, item.IsPrefix, listItems[i].IsPrefix)
|
2019-12-06 18:03:22 +00:00
|
|
|
}
|
|
|
|
|
2020-05-26 09:05:43 +01:00
|
|
|
objects = project.ListObjects(ctx, "testbucket", &uplink.ListObjectsOptions{
|
|
|
|
Recursive: false,
|
|
|
|
})
|
|
|
|
|
|
|
|
listItems = make([]*uplink.Object, 0)
|
|
|
|
for objects.Next() {
|
|
|
|
listItems = append(listItems, objects.Item())
|
|
|
|
}
|
|
|
|
require.NoError(t, objects.Err())
|
2019-03-30 11:21:49 +00:00
|
|
|
|
2019-12-06 18:03:22 +00:00
|
|
|
expected = []storj.Object{
|
|
|
|
{Path: "müsic"},
|
|
|
|
{Path: "müsic/", IsPrefix: true},
|
|
|
|
{Path: "sample.😶"},
|
|
|
|
{Path: "ビデオ/", IsPrefix: true},
|
|
|
|
}
|
|
|
|
|
2020-05-26 09:05:43 +01:00
|
|
|
require.Equal(t, len(expected), len(listItems))
|
|
|
|
sort.Slice(listItems, func(i, k int) bool {
|
|
|
|
return listItems[i].Key < listItems[k].Key
|
2019-12-06 18:03:22 +00:00
|
|
|
})
|
|
|
|
for i, item := range expected {
|
2020-05-26 09:05:43 +01:00
|
|
|
t.Log(item.Path, listItems[i].Key)
|
|
|
|
require.Equal(t, item.Path, listItems[i].Key)
|
|
|
|
require.Equal(t, item.IsPrefix, listItems[i].IsPrefix)
|
2019-12-06 18:03:22 +00:00
|
|
|
}
|
2019-03-30 11:21:49 +00:00
|
|
|
})
|
|
|
|
}
|
2019-04-09 14:31:19 +01:00
|
|
|
|
2020-01-03 09:27:10 +00:00
|
|
|
func TestExpirationTimeSegment(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
|
|
|
|
err := planet.Uplinks[0].CreateBucket(ctx, planet.Satellites[0], "my-bucket-name")
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
metainfoClient, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(metainfoClient.Close)
|
|
|
|
|
|
|
|
for _, r := range []struct {
|
|
|
|
expirationDate time.Time
|
|
|
|
errFlag bool
|
|
|
|
}{
|
|
|
|
{ // expiration time not set
|
|
|
|
time.Time{},
|
|
|
|
false,
|
|
|
|
},
|
|
|
|
{ // 10 days into future
|
|
|
|
time.Now().AddDate(0, 0, 10),
|
|
|
|
false,
|
|
|
|
},
|
|
|
|
{ // current time
|
|
|
|
time.Now(),
|
|
|
|
true,
|
|
|
|
},
|
|
|
|
{ // 10 days into past
|
|
|
|
time.Now().AddDate(0, 0, -10),
|
|
|
|
true,
|
|
|
|
},
|
|
|
|
} {
|
|
|
|
_, err := metainfoClient.BeginObject(ctx, metainfo.BeginObjectParams{
|
|
|
|
Bucket: []byte("my-bucket-name"),
|
|
|
|
EncryptedPath: []byte("path"),
|
|
|
|
ExpiresAt: r.expirationDate,
|
|
|
|
})
|
|
|
|
if r.errFlag {
|
|
|
|
assert.Error(t, err)
|
|
|
|
} else {
|
|
|
|
assert.NoError(t, err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2019-06-27 18:36:51 +01:00
|
|
|
func TestGetProjectInfo(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
2019-07-16 20:16:41 +01:00
|
|
|
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 2,
|
2019-06-27 18:36:51 +01:00
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
apiKey0 := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
apiKey1 := planet.Uplinks[1].APIKey[planet.Satellites[0].ID()]
|
|
|
|
|
|
|
|
metainfo0, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey0)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
metainfo1, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey1)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
info0, err := metainfo0.GetProjectInfo(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.NotNil(t, info0.ProjectSalt)
|
|
|
|
|
|
|
|
info1, err := metainfo1.GetProjectInfo(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.NotNil(t, info1.ProjectSalt)
|
|
|
|
|
|
|
|
// Different projects should have different salts
|
|
|
|
require.NotEqual(t, info0.ProjectSalt, info1.ProjectSalt)
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2019-06-24 10:52:25 +01:00
|
|
|
func TestBucketNameValidation(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
2019-07-24 12:33:23 +01:00
|
|
|
SatelliteCount: 1, StorageNodeCount: 4, UplinkCount: 1,
|
2019-06-24 10:52:25 +01:00
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
|
2020-01-03 09:27:10 +00:00
|
|
|
metainfoClient, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
2019-06-24 10:52:25 +01:00
|
|
|
require.NoError(t, err)
|
2020-01-03 09:27:10 +00:00
|
|
|
defer ctx.Check(metainfoClient.Close)
|
2019-06-24 10:52:25 +01:00
|
|
|
|
|
|
|
validNames := []string{
|
|
|
|
"tes", "testbucket",
|
|
|
|
"test-bucket", "testbucket9",
|
|
|
|
"9testbucket", "a.b",
|
|
|
|
"test.bucket", "test-one.bucket-one",
|
|
|
|
"test.bucket.one",
|
|
|
|
"testbucket-63-0123456789012345678901234567890123456789012345abc",
|
|
|
|
}
|
|
|
|
for _, name := range validNames {
|
2020-01-03 09:27:10 +00:00
|
|
|
_, err = metainfoClient.CreateBucket(ctx, metainfo.CreateBucketParams{
|
|
|
|
Name: []byte(name),
|
|
|
|
})
|
2019-06-24 10:52:25 +01:00
|
|
|
require.NoError(t, err, "bucket name: %v", name)
|
2020-06-15 12:49:09 +01:00
|
|
|
|
|
|
|
_, err = metainfoClient.BeginObject(ctx, metainfo.BeginObjectParams{
|
|
|
|
Bucket: []byte(name),
|
|
|
|
EncryptedPath: []byte("123"),
|
|
|
|
Version: 0,
|
|
|
|
ExpiresAt: time.Now().Add(16 * 24 * time.Hour),
|
|
|
|
})
|
|
|
|
require.NoError(t, err, "bucket name: %v", name)
|
2019-06-24 10:52:25 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
invalidNames := []string{
|
|
|
|
"", "t", "te", "-testbucket",
|
|
|
|
"testbucket-", "-testbucket-",
|
|
|
|
"a.b.", "test.bucket-.one",
|
|
|
|
"test.-bucket.one", "1.2.3.4",
|
|
|
|
"192.168.1.234", "testBUCKET",
|
|
|
|
"test/bucket",
|
|
|
|
"testbucket-64-0123456789012345678901234567890123456789012345abcd",
|
|
|
|
}
|
|
|
|
for _, name := range invalidNames {
|
2020-06-15 12:49:09 +01:00
|
|
|
_, err = metainfoClient.BeginObject(ctx, metainfo.BeginObjectParams{
|
|
|
|
Bucket: []byte(name),
|
|
|
|
EncryptedPath: []byte("123"),
|
|
|
|
Version: 0,
|
|
|
|
ExpiresAt: time.Now().Add(16 * 24 * time.Hour),
|
|
|
|
})
|
2020-01-03 09:27:10 +00:00
|
|
|
require.Error(t, err, "bucket name: %v", name)
|
|
|
|
|
|
|
|
_, err = metainfoClient.CreateBucket(ctx, metainfo.CreateBucketParams{
|
|
|
|
Name: []byte(name),
|
|
|
|
})
|
2019-06-24 10:52:25 +01:00
|
|
|
require.Error(t, err, "bucket name: %v", name)
|
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
2019-07-16 11:39:23 +01:00
|
|
|
|
2019-07-23 12:09:12 +01:00
|
|
|
func TestListGetObjects(t *testing.T) {
|
2019-07-16 11:39:23 +01:00
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
uplink := planet.Uplinks[0]
|
|
|
|
|
|
|
|
files := make([]string, 10)
|
|
|
|
data := testrand.Bytes(1 * memory.KiB)
|
|
|
|
for i := 0; i < len(files); i++ {
|
|
|
|
files[i] = "path" + strconv.Itoa(i)
|
|
|
|
err := uplink.Upload(ctx, planet.Satellites[0], "testbucket", files[i], data)
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
|
2019-07-24 12:33:23 +01:00
|
|
|
metainfoClient, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
2019-07-16 11:39:23 +01:00
|
|
|
require.NoError(t, err)
|
2019-07-24 12:33:23 +01:00
|
|
|
defer ctx.Check(metainfoClient.Close)
|
2019-07-16 11:39:23 +01:00
|
|
|
|
2019-07-24 12:33:23 +01:00
|
|
|
expectedBucketName := "testbucket"
|
|
|
|
items, _, err := metainfoClient.ListObjects(ctx, metainfo.ListObjectsParams{
|
|
|
|
Bucket: []byte(expectedBucketName),
|
|
|
|
})
|
2019-07-16 11:39:23 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, len(files), len(items))
|
|
|
|
for _, item := range items {
|
|
|
|
require.NotEmpty(t, item.EncryptedPath)
|
|
|
|
require.True(t, item.CreatedAt.Before(time.Now()))
|
2019-07-23 12:09:12 +01:00
|
|
|
|
2019-08-01 10:04:31 +01:00
|
|
|
object, err := metainfoClient.GetObject(ctx, metainfo.GetObjectParams{
|
2019-07-24 12:33:23 +01:00
|
|
|
Bucket: []byte(expectedBucketName),
|
|
|
|
EncryptedPath: item.EncryptedPath,
|
|
|
|
})
|
2019-07-23 12:09:12 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, item.EncryptedPath, []byte(object.Path))
|
|
|
|
|
2019-08-01 10:04:31 +01:00
|
|
|
require.NotEmpty(t, object.StreamID)
|
2019-07-16 11:39:23 +01:00
|
|
|
}
|
|
|
|
|
2019-07-24 12:33:23 +01:00
|
|
|
items, _, err = metainfoClient.ListObjects(ctx, metainfo.ListObjectsParams{
|
|
|
|
Bucket: []byte(expectedBucketName),
|
|
|
|
Limit: 3,
|
|
|
|
})
|
2019-07-16 11:39:23 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, 3, len(items))
|
|
|
|
})
|
|
|
|
}
|
2019-07-24 12:33:23 +01:00
|
|
|
|
2020-03-16 08:55:52 +00:00
|
|
|
func TestBucketExistenceCheck(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
|
|
|
|
metainfoClient, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(metainfoClient.Close)
|
|
|
|
|
|
|
|
// test object methods for bucket existence check
|
|
|
|
_, err = metainfoClient.BeginObject(ctx, metainfo.BeginObjectParams{
|
|
|
|
Bucket: []byte("non-existing-bucket"),
|
|
|
|
EncryptedPath: []byte("encrypted-path"),
|
|
|
|
})
|
|
|
|
require.Error(t, err)
|
|
|
|
require.True(t, errs2.IsRPC(err, rpcstatus.NotFound))
|
|
|
|
require.Equal(t, storj.ErrBucketNotFound.New("%s", "non-existing-bucket").Error(), errs.Unwrap(err).Error())
|
|
|
|
|
|
|
|
_, _, err = metainfoClient.ListObjects(ctx, metainfo.ListObjectsParams{
|
|
|
|
Bucket: []byte("non-existing-bucket"),
|
|
|
|
})
|
|
|
|
require.Error(t, err)
|
|
|
|
require.True(t, errs2.IsRPC(err, rpcstatus.NotFound))
|
|
|
|
require.Equal(t, storj.ErrBucketNotFound.New("%s", "non-existing-bucket").Error(), errs.Unwrap(err).Error())
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2019-07-24 12:33:23 +01:00
|
|
|
func TestBeginCommitListSegment(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 4, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
metainfoService := planet.Satellites[0].Metainfo.Service
|
|
|
|
|
|
|
|
projects, err := planet.Satellites[0].DB.Console().Projects().GetAll(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
projectID := projects[0].ID
|
|
|
|
|
|
|
|
bucket := storj.Bucket{
|
2020-05-26 09:05:43 +01:00
|
|
|
Name: "initial-bucket",
|
|
|
|
ProjectID: projectID,
|
2019-07-24 12:33:23 +01:00
|
|
|
}
|
|
|
|
_, err = metainfoService.CreateBucket(ctx, bucket)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
metainfoClient, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(metainfoClient.Close)
|
|
|
|
|
|
|
|
params := metainfo.BeginObjectParams{
|
|
|
|
Bucket: []byte(bucket.Name),
|
|
|
|
EncryptedPath: []byte("encrypted-path"),
|
|
|
|
Redundancy: storj.RedundancyScheme{
|
|
|
|
Algorithm: storj.ReedSolomon,
|
|
|
|
ShareSize: 256,
|
|
|
|
RequiredShares: 1,
|
|
|
|
RepairShares: 1,
|
|
|
|
OptimalShares: 3,
|
|
|
|
TotalShares: 4,
|
|
|
|
},
|
2019-08-01 10:04:31 +01:00
|
|
|
EncryptionParameters: storj.EncryptionParameters{},
|
2020-03-10 22:05:01 +00:00
|
|
|
ExpiresAt: time.Now().Add(24 * time.Hour),
|
2019-07-24 12:33:23 +01:00
|
|
|
}
|
2020-01-20 14:20:52 +00:00
|
|
|
beginObjectResponse, err := metainfoClient.BeginObject(ctx, params)
|
2019-07-24 12:33:23 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
segmentID, limits, _, err := metainfoClient.BeginSegment(ctx, metainfo.BeginSegmentParams{
|
2020-01-20 14:20:52 +00:00
|
|
|
StreamID: beginObjectResponse.StreamID,
|
2019-07-24 12:33:23 +01:00
|
|
|
Position: storj.SegmentPosition{
|
2019-08-01 10:04:31 +01:00
|
|
|
Index: 0,
|
2019-07-24 12:33:23 +01:00
|
|
|
},
|
2019-09-10 16:39:47 +01:00
|
|
|
MaxOrderLimit: memory.MiB.Int64(),
|
2019-07-24 12:33:23 +01:00
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2019-09-18 14:50:33 +01:00
|
|
|
fullIDMap := make(map[storj.NodeID]*identity.FullIdentity)
|
|
|
|
for _, node := range planet.StorageNodes {
|
|
|
|
fullIDMap[node.ID()] = node.Identity
|
|
|
|
}
|
|
|
|
|
2019-07-24 12:33:23 +01:00
|
|
|
makeResult := func(num int32) *pb.SegmentPieceUploadResult {
|
2019-09-18 14:50:33 +01:00
|
|
|
nodeID := limits[num].Limit.StorageNodeId
|
|
|
|
hash := &pb.PieceHash{
|
|
|
|
PieceId: limits[num].Limit.PieceId,
|
|
|
|
PieceSize: 1048832,
|
|
|
|
Timestamp: time.Now(),
|
|
|
|
}
|
|
|
|
|
|
|
|
fullID := fullIDMap[nodeID]
|
|
|
|
require.NotNil(t, fullID)
|
|
|
|
signer := signing.SignerFromFullIdentity(fullID)
|
|
|
|
signedHash, err := signing.SignPieceHash(ctx, signer, hash)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2019-07-24 12:33:23 +01:00
|
|
|
return &pb.SegmentPieceUploadResult{
|
|
|
|
PieceNum: num,
|
2019-09-18 14:50:33 +01:00
|
|
|
NodeId: nodeID,
|
|
|
|
Hash: signedHash,
|
2019-07-24 12:33:23 +01:00
|
|
|
}
|
|
|
|
}
|
2019-10-23 07:59:56 +01:00
|
|
|
err = metainfoClient.CommitSegment(ctx, metainfo.CommitSegmentParams{
|
2019-08-01 10:04:31 +01:00
|
|
|
SegmentID: segmentID,
|
|
|
|
|
2019-07-24 12:33:23 +01:00
|
|
|
SizeEncryptedData: memory.MiB.Int64(),
|
|
|
|
UploadResult: []*pb.SegmentPieceUploadResult{
|
|
|
|
makeResult(0),
|
|
|
|
makeResult(1),
|
2019-07-31 19:28:43 +01:00
|
|
|
makeResult(2),
|
2019-07-24 12:33:23 +01:00
|
|
|
},
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2020-04-08 13:08:57 +01:00
|
|
|
metadata, err := pb.Marshal(&pb.StreamMeta{
|
2019-11-24 21:14:51 +00:00
|
|
|
NumberOfSegments: 1,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
2019-08-01 10:04:31 +01:00
|
|
|
err = metainfoClient.CommitObject(ctx, metainfo.CommitObjectParams{
|
2020-01-20 14:20:52 +00:00
|
|
|
StreamID: beginObjectResponse.StreamID,
|
2019-11-24 21:14:51 +00:00
|
|
|
EncryptedMetadata: metadata,
|
2019-08-01 10:04:31 +01:00
|
|
|
})
|
2019-07-24 12:33:23 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
objects, _, err := metainfoClient.ListObjects(ctx, metainfo.ListObjectsParams{
|
|
|
|
Bucket: []byte(bucket.Name),
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, objects, 1)
|
|
|
|
|
|
|
|
require.Equal(t, params.EncryptedPath, objects[0].EncryptedPath)
|
2020-03-10 22:05:01 +00:00
|
|
|
require.True(t, params.ExpiresAt.Equal(objects[0].ExpiresAt))
|
2019-07-24 12:33:23 +01:00
|
|
|
|
2019-08-01 10:04:31 +01:00
|
|
|
object, err := metainfoClient.GetObject(ctx, metainfo.GetObjectParams{
|
2019-07-24 12:33:23 +01:00
|
|
|
Bucket: []byte(bucket.Name),
|
|
|
|
EncryptedPath: objects[0].EncryptedPath,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2019-10-23 07:59:56 +01:00
|
|
|
segments, _, err := metainfoClient.ListSegments(ctx, metainfo.ListSegmentsParams{
|
2019-08-01 10:04:31 +01:00
|
|
|
StreamID: object.StreamID,
|
2019-07-24 12:33:23 +01:00
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, segments, 1)
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2019-12-09 12:18:54 +00:00
|
|
|
func TestListSegments(t *testing.T) {
|
2019-08-30 22:30:18 +01:00
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 4, UplinkCount: 1,
|
2020-05-26 09:05:43 +01:00
|
|
|
Reconfigure: testplanet.Reconfigure{
|
|
|
|
Satellite: testplanet.MaxSegmentSize(memory.KiB),
|
|
|
|
},
|
2019-08-30 22:30:18 +01:00
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
uplink := planet.Uplinks[0]
|
|
|
|
|
|
|
|
data := testrand.Bytes(15 * memory.KiB)
|
2020-05-26 09:05:43 +01:00
|
|
|
err := uplink.Upload(ctx, planet.Satellites[0], "testbucket", "test-path", data)
|
2019-08-30 22:30:18 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// 15KiB + encryption should be uploaded into 16 segments with SegmentSize == 1KiB
|
|
|
|
numberOfSegments := 16
|
|
|
|
|
|
|
|
metainfoClient, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(metainfoClient.Close)
|
|
|
|
|
|
|
|
items, _, err := metainfoClient.ListObjects(ctx, metainfo.ListObjectsParams{
|
|
|
|
Bucket: []byte("testbucket"),
|
|
|
|
Limit: 1,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
object, err := metainfoClient.GetObject(ctx, metainfo.GetObjectParams{
|
|
|
|
Bucket: []byte("testbucket"),
|
|
|
|
EncryptedPath: items[0].EncryptedPath,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2019-12-09 12:18:54 +00:00
|
|
|
for i, test := range []struct {
|
2019-08-30 22:30:18 +01:00
|
|
|
Index int32
|
|
|
|
Limit int32
|
|
|
|
Result int
|
|
|
|
More bool
|
|
|
|
}{
|
|
|
|
{Index: 0, Result: numberOfSegments},
|
|
|
|
{Index: 0, Result: numberOfSegments, Limit: int32(numberOfSegments), More: false},
|
|
|
|
{Index: 0, Result: 5, Limit: 5, More: true},
|
|
|
|
{Index: 16, Result: 0, More: false},
|
|
|
|
{Index: 11, Result: 5, Limit: 5, More: false},
|
|
|
|
{Index: 15, Result: 1, More: false},
|
|
|
|
} {
|
2019-10-23 07:59:56 +01:00
|
|
|
segments, more, err := metainfoClient.ListSegments(ctx, metainfo.ListSegmentsParams{
|
2019-08-30 22:30:18 +01:00
|
|
|
StreamID: object.StreamID,
|
|
|
|
Limit: test.Limit,
|
|
|
|
CursorPosition: storj.SegmentPosition{
|
|
|
|
Index: test.Index,
|
|
|
|
},
|
|
|
|
})
|
2019-12-09 12:18:54 +00:00
|
|
|
require.NoErrorf(t, err, "test case: %d", i)
|
|
|
|
require.Lenf(t, segments, test.Result, "test case: %d", i)
|
|
|
|
require.Equalf(t, test.More, more, "test case: %d", i)
|
2019-08-30 22:30:18 +01:00
|
|
|
if !more && test.Result > 0 {
|
2019-12-09 12:18:54 +00:00
|
|
|
require.Equalf(t, int32(-1), segments[test.Result-1].Position.Index, "test case: %d", i)
|
2019-08-30 22:30:18 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2019-07-24 12:33:23 +01:00
|
|
|
func TestInlineSegment(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
|
|
|
|
metainfoService := planet.Satellites[0].Metainfo.Service
|
|
|
|
|
|
|
|
projects, err := planet.Satellites[0].DB.Console().Projects().GetAll(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
projectID := projects[0].ID
|
|
|
|
|
|
|
|
// TODO maybe split into separate cases
|
|
|
|
// Test:
|
|
|
|
// * create bucket
|
|
|
|
// * begin object
|
|
|
|
// * send several inline segments
|
|
|
|
// * commit object
|
|
|
|
// * list created object
|
|
|
|
// * list object segments
|
|
|
|
// * download segments
|
|
|
|
// * delete segments and object
|
|
|
|
|
|
|
|
bucket := storj.Bucket{
|
2020-05-26 09:05:43 +01:00
|
|
|
Name: "inline-segments-bucket",
|
|
|
|
ProjectID: projectID,
|
2019-07-24 12:33:23 +01:00
|
|
|
}
|
|
|
|
_, err = metainfoService.CreateBucket(ctx, bucket)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
metainfoClient, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(metainfoClient.Close)
|
|
|
|
|
|
|
|
params := metainfo.BeginObjectParams{
|
|
|
|
Bucket: []byte(bucket.Name),
|
|
|
|
EncryptedPath: []byte("encrypted-path"),
|
|
|
|
Redundancy: storj.RedundancyScheme{
|
|
|
|
Algorithm: storj.ReedSolomon,
|
|
|
|
ShareSize: 256,
|
|
|
|
RequiredShares: 1,
|
|
|
|
RepairShares: 1,
|
|
|
|
OptimalShares: 3,
|
|
|
|
TotalShares: 4,
|
|
|
|
},
|
2019-08-01 10:04:31 +01:00
|
|
|
EncryptionParameters: storj.EncryptionParameters{},
|
2020-03-10 22:05:01 +00:00
|
|
|
ExpiresAt: time.Now().Add(24 * time.Hour),
|
2019-07-24 12:33:23 +01:00
|
|
|
}
|
2020-01-20 14:20:52 +00:00
|
|
|
beginObjectResp, err := metainfoClient.BeginObject(ctx, params)
|
2019-07-24 12:33:23 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
2019-08-01 10:04:31 +01:00
|
|
|
segments := []int32{0, 1, 2, 3, 4, 5, 6}
|
2019-07-24 12:33:23 +01:00
|
|
|
segmentsData := make([][]byte, len(segments))
|
|
|
|
for i, segment := range segments {
|
|
|
|
segmentsData[i] = testrand.Bytes(memory.KiB)
|
|
|
|
err = metainfoClient.MakeInlineSegment(ctx, metainfo.MakeInlineSegmentParams{
|
2020-01-20 14:20:52 +00:00
|
|
|
StreamID: beginObjectResp.StreamID,
|
2019-07-24 12:33:23 +01:00
|
|
|
Position: storj.SegmentPosition{
|
|
|
|
Index: segment,
|
|
|
|
},
|
|
|
|
EncryptedInlineData: segmentsData[i],
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
|
2020-04-08 13:08:57 +01:00
|
|
|
metadata, err := pb.Marshal(&pb.StreamMeta{
|
2019-11-24 21:14:51 +00:00
|
|
|
NumberOfSegments: int64(len(segments)),
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
2019-08-01 10:04:31 +01:00
|
|
|
err = metainfoClient.CommitObject(ctx, metainfo.CommitObjectParams{
|
2020-01-20 14:20:52 +00:00
|
|
|
StreamID: beginObjectResp.StreamID,
|
2019-11-24 21:14:51 +00:00
|
|
|
EncryptedMetadata: metadata,
|
2019-08-01 10:04:31 +01:00
|
|
|
})
|
2019-07-24 12:33:23 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
objects, _, err := metainfoClient.ListObjects(ctx, metainfo.ListObjectsParams{
|
|
|
|
Bucket: []byte(bucket.Name),
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, objects, 1)
|
|
|
|
|
|
|
|
require.Equal(t, params.EncryptedPath, objects[0].EncryptedPath)
|
2020-03-10 22:05:01 +00:00
|
|
|
require.True(t, params.ExpiresAt.Equal(objects[0].ExpiresAt))
|
2019-07-24 12:33:23 +01:00
|
|
|
|
2019-08-01 10:04:31 +01:00
|
|
|
object, err := metainfoClient.GetObject(ctx, metainfo.GetObjectParams{
|
2019-07-24 12:33:23 +01:00
|
|
|
Bucket: params.Bucket,
|
|
|
|
EncryptedPath: params.EncryptedPath,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2020-04-09 09:19:16 +01:00
|
|
|
{ // Confirm data larger than our configured max inline segment size of 4 KiB cannot be inlined
|
2020-04-01 10:15:24 +01:00
|
|
|
beginObjectResp, err := metainfoClient.BeginObject(ctx, metainfo.BeginObjectParams{
|
|
|
|
Bucket: []byte(bucket.Name),
|
|
|
|
EncryptedPath: []byte("too-large-inline-segment"),
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2020-04-09 09:19:16 +01:00
|
|
|
data := testrand.Bytes(10 * memory.KiB)
|
2020-04-01 10:15:24 +01:00
|
|
|
err = metainfoClient.MakeInlineSegment(ctx, metainfo.MakeInlineSegmentParams{
|
|
|
|
StreamID: beginObjectResp.StreamID,
|
|
|
|
Position: storj.SegmentPosition{
|
|
|
|
Index: 0,
|
|
|
|
},
|
|
|
|
EncryptedInlineData: data,
|
|
|
|
})
|
|
|
|
require.Error(t, err)
|
|
|
|
}
|
2019-07-24 12:33:23 +01:00
|
|
|
{ // test listing inline segments
|
|
|
|
for _, test := range []struct {
|
|
|
|
Index int32
|
|
|
|
Limit int
|
|
|
|
Result int
|
|
|
|
More bool
|
|
|
|
}{
|
|
|
|
{Index: 0, Result: len(segments), More: false},
|
|
|
|
{Index: 2, Result: len(segments) - 2, More: false},
|
|
|
|
{Index: 0, Result: 3, More: true, Limit: 3},
|
|
|
|
{Index: 0, Result: len(segments), More: false, Limit: len(segments)},
|
|
|
|
{Index: 0, Result: len(segments) - 1, More: true, Limit: len(segments) - 1},
|
|
|
|
} {
|
2019-10-23 07:59:56 +01:00
|
|
|
items, more, err := metainfoClient.ListSegments(ctx, metainfo.ListSegmentsParams{
|
2019-08-01 10:04:31 +01:00
|
|
|
StreamID: object.StreamID,
|
2019-07-24 12:33:23 +01:00
|
|
|
CursorPosition: storj.SegmentPosition{
|
|
|
|
Index: test.Index,
|
|
|
|
},
|
|
|
|
Limit: int32(test.Limit),
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, test.Result, len(items))
|
|
|
|
require.Equal(t, test.More, more)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
{ // test download inline segments
|
2019-10-23 07:59:56 +01:00
|
|
|
items, _, err := metainfoClient.ListSegments(ctx, metainfo.ListSegmentsParams{
|
2019-08-01 10:04:31 +01:00
|
|
|
StreamID: object.StreamID,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, len(segments), len(items))
|
|
|
|
|
|
|
|
for i, item := range items {
|
2019-07-24 12:33:23 +01:00
|
|
|
info, limits, err := metainfoClient.DownloadSegment(ctx, metainfo.DownloadSegmentParams{
|
2019-08-01 10:04:31 +01:00
|
|
|
StreamID: object.StreamID,
|
2019-07-24 12:33:23 +01:00
|
|
|
Position: storj.SegmentPosition{
|
2019-08-01 10:04:31 +01:00
|
|
|
Index: item.Position.Index,
|
2019-07-24 12:33:23 +01:00
|
|
|
},
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Nil(t, limits)
|
|
|
|
require.Equal(t, segmentsData[i], info.EncryptedInlineData)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
{ // test deleting segments
|
2020-07-21 08:34:00 +01:00
|
|
|
_, err := metainfoClient.BeginDeleteObject(ctx, metainfo.BeginDeleteObjectParams{
|
|
|
|
Bucket: params.Bucket,
|
|
|
|
EncryptedPath: params.EncryptedPath,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
_, err = metainfoClient.GetObject(ctx, metainfo.GetObjectParams{
|
|
|
|
Bucket: params.Bucket,
|
|
|
|
EncryptedPath: params.EncryptedPath,
|
|
|
|
})
|
|
|
|
require.Error(t, err)
|
2019-07-24 12:33:23 +01:00
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestRemoteSegment(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 4, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
uplink := planet.Uplinks[0]
|
|
|
|
|
|
|
|
expectedBucketName := "remote-segments-bucket"
|
|
|
|
err := uplink.Upload(ctx, planet.Satellites[0], expectedBucketName, "file-object", testrand.Bytes(10*memory.KiB))
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
metainfoClient, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(metainfoClient.Close)
|
|
|
|
|
|
|
|
items, _, err := metainfoClient.ListObjects(ctx, metainfo.ListObjectsParams{
|
|
|
|
Bucket: []byte(expectedBucketName),
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, items, 1)
|
|
|
|
|
|
|
|
{
|
|
|
|
// Get object
|
|
|
|
// List segments
|
|
|
|
// Download segment
|
|
|
|
|
2019-08-01 10:04:31 +01:00
|
|
|
object, err := metainfoClient.GetObject(ctx, metainfo.GetObjectParams{
|
2019-07-24 12:33:23 +01:00
|
|
|
Bucket: []byte(expectedBucketName),
|
|
|
|
EncryptedPath: items[0].EncryptedPath,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2019-10-23 07:59:56 +01:00
|
|
|
segments, _, err := metainfoClient.ListSegments(ctx, metainfo.ListSegmentsParams{
|
2019-08-01 10:04:31 +01:00
|
|
|
StreamID: object.StreamID,
|
2019-07-24 12:33:23 +01:00
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, segments, 1)
|
|
|
|
|
|
|
|
_, limits, err := metainfoClient.DownloadSegment(ctx, metainfo.DownloadSegmentParams{
|
2019-08-01 10:04:31 +01:00
|
|
|
StreamID: object.StreamID,
|
2019-07-24 12:33:23 +01:00
|
|
|
Position: storj.SegmentPosition{
|
|
|
|
Index: segments[0].Position.Index,
|
|
|
|
},
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.NotEmpty(t, limits)
|
|
|
|
}
|
|
|
|
|
|
|
|
{
|
|
|
|
// Begin deleting object
|
|
|
|
// List objects
|
|
|
|
|
2020-07-21 08:34:00 +01:00
|
|
|
_, err := metainfoClient.BeginDeleteObject(ctx, metainfo.BeginDeleteObjectParams{
|
|
|
|
Bucket: []byte(expectedBucketName),
|
|
|
|
EncryptedPath: items[0].EncryptedPath,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
items, _, err = metainfoClient.ListObjects(ctx, metainfo.ListObjectsParams{
|
|
|
|
Bucket: []byte(expectedBucketName),
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, items, 0)
|
2019-07-24 12:33:23 +01:00
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestIDs(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
|
|
|
|
metainfoClient, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(metainfoClient.Close)
|
|
|
|
|
|
|
|
{
|
|
|
|
streamID := testrand.StreamID(256)
|
2019-08-01 10:04:31 +01:00
|
|
|
err = metainfoClient.CommitObject(ctx, metainfo.CommitObjectParams{
|
|
|
|
StreamID: streamID,
|
|
|
|
})
|
2019-07-24 12:33:23 +01:00
|
|
|
require.Error(t, err) // invalid streamID
|
|
|
|
|
|
|
|
segmentID := testrand.SegmentID(512)
|
2019-10-23 07:59:56 +01:00
|
|
|
err = metainfoClient.CommitSegment(ctx, metainfo.CommitSegmentParams{
|
2019-07-24 12:33:23 +01:00
|
|
|
SegmentID: segmentID,
|
|
|
|
})
|
|
|
|
require.Error(t, err) // invalid segmentID
|
|
|
|
}
|
|
|
|
|
|
|
|
satellitePeer := signing.SignerFromFullIdentity(planet.Satellites[0].Identity)
|
|
|
|
|
|
|
|
{ // streamID expired
|
|
|
|
signedStreamID, err := signing.SignStreamID(ctx, satellitePeer, &pb.SatStreamID{
|
2019-11-15 15:44:23 +00:00
|
|
|
CreationDate: time.Now().Add(-36 * time.Hour),
|
2019-07-24 12:33:23 +01:00
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2020-04-08 13:08:57 +01:00
|
|
|
encodedStreamID, err := pb.Marshal(signedStreamID)
|
2019-07-24 12:33:23 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
streamID, err := storj.StreamIDFromBytes(encodedStreamID)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2019-08-01 10:04:31 +01:00
|
|
|
err = metainfoClient.CommitObject(ctx, metainfo.CommitObjectParams{
|
|
|
|
StreamID: streamID,
|
|
|
|
})
|
2019-07-24 12:33:23 +01:00
|
|
|
require.Error(t, err)
|
|
|
|
}
|
|
|
|
|
2019-11-15 15:44:23 +00:00
|
|
|
{ // segment id missing stream id
|
|
|
|
signedSegmentID, err := signing.SignSegmentID(ctx, satellitePeer, &pb.SatSegmentID{
|
|
|
|
CreationDate: time.Now().Add(-1 * time.Hour),
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2020-04-08 13:08:57 +01:00
|
|
|
encodedSegmentID, err := pb.Marshal(signedSegmentID)
|
2019-11-15 15:44:23 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
segmentID, err := storj.SegmentIDFromBytes(encodedSegmentID)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
err = metainfoClient.CommitSegment(ctx, metainfo.CommitSegmentParams{
|
|
|
|
SegmentID: segmentID,
|
|
|
|
})
|
|
|
|
require.Error(t, err)
|
|
|
|
}
|
|
|
|
|
2019-07-24 12:33:23 +01:00
|
|
|
{ // segmentID expired
|
|
|
|
signedSegmentID, err := signing.SignSegmentID(ctx, satellitePeer, &pb.SatSegmentID{
|
2019-11-15 15:44:23 +00:00
|
|
|
CreationDate: time.Now().Add(-36 * time.Hour),
|
|
|
|
StreamId: &pb.SatStreamID{
|
|
|
|
CreationDate: time.Now(),
|
|
|
|
},
|
2019-07-24 12:33:23 +01:00
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2020-04-08 13:08:57 +01:00
|
|
|
encodedSegmentID, err := pb.Marshal(signedSegmentID)
|
2019-07-24 12:33:23 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
segmentID, err := storj.SegmentIDFromBytes(encodedSegmentID)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2019-10-23 07:59:56 +01:00
|
|
|
err = metainfoClient.CommitSegment(ctx, metainfo.CommitSegmentParams{
|
2019-07-24 12:33:23 +01:00
|
|
|
SegmentID: segmentID,
|
|
|
|
})
|
|
|
|
require.Error(t, err)
|
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
2019-08-06 15:56:23 +01:00
|
|
|
|
|
|
|
func TestBatch(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
|
|
|
|
metainfoClient, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(metainfoClient.Close)
|
|
|
|
|
|
|
|
{ // create few buckets and list them in one batch
|
|
|
|
requests := make([]metainfo.BatchItem, 0)
|
|
|
|
numOfBuckets := 5
|
|
|
|
for i := 0; i < numOfBuckets; i++ {
|
|
|
|
requests = append(requests, &metainfo.CreateBucketParams{
|
|
|
|
Name: []byte("test-bucket-" + strconv.Itoa(i)),
|
|
|
|
PathCipher: storj.EncAESGCM,
|
|
|
|
DefaultSegmentsSize: memory.MiB.Int64(),
|
|
|
|
})
|
|
|
|
}
|
|
|
|
requests = append(requests, &metainfo.ListBucketsParams{
|
|
|
|
ListOpts: storj.BucketListOptions{
|
|
|
|
Cursor: "",
|
|
|
|
Direction: storj.After,
|
|
|
|
},
|
|
|
|
})
|
|
|
|
responses, err := metainfoClient.Batch(ctx, requests...)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, numOfBuckets+1, len(responses))
|
|
|
|
|
|
|
|
for i := 0; i < numOfBuckets; i++ {
|
|
|
|
response, err := responses[i].CreateBucket()
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, "test-bucket-"+strconv.Itoa(i), response.Bucket.Name)
|
|
|
|
|
|
|
|
_, err = responses[i].GetBucket()
|
|
|
|
require.Error(t, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
bucketsListResp, err := responses[numOfBuckets].ListBuckets()
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, numOfBuckets, len(bucketsListResp.BucketList.Items))
|
|
|
|
}
|
|
|
|
|
|
|
|
{ // create bucket, object, upload inline segments in batch, download inline segments in batch
|
|
|
|
err := planet.Uplinks[0].CreateBucket(ctx, planet.Satellites[0], "second-test-bucket")
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2019-10-22 11:23:22 +01:00
|
|
|
requests := make([]metainfo.BatchItem, 0)
|
|
|
|
requests = append(requests, &metainfo.BeginObjectParams{
|
2019-08-06 15:56:23 +01:00
|
|
|
Bucket: []byte("second-test-bucket"),
|
|
|
|
EncryptedPath: []byte("encrypted-path"),
|
|
|
|
})
|
|
|
|
numOfSegments := 10
|
|
|
|
expectedData := make([][]byte, numOfSegments)
|
|
|
|
for i := 0; i < numOfSegments; i++ {
|
|
|
|
expectedData[i] = testrand.Bytes(memory.KiB)
|
|
|
|
|
|
|
|
requests = append(requests, &metainfo.MakeInlineSegmentParams{
|
|
|
|
Position: storj.SegmentPosition{
|
|
|
|
Index: int32(i),
|
|
|
|
},
|
|
|
|
EncryptedInlineData: expectedData[i],
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2020-04-08 13:08:57 +01:00
|
|
|
metadata, err := pb.Marshal(&pb.StreamMeta{
|
2019-11-24 21:14:51 +00:00
|
|
|
NumberOfSegments: int64(numOfSegments),
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
requests = append(requests, &metainfo.CommitObjectParams{
|
|
|
|
EncryptedMetadata: metadata,
|
|
|
|
})
|
2019-10-22 11:23:22 +01:00
|
|
|
requests = append(requests, &metainfo.ListSegmentsParams{})
|
2019-08-06 15:56:23 +01:00
|
|
|
|
|
|
|
responses, err := metainfoClient.Batch(ctx, requests...)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, numOfSegments+3, len(responses))
|
|
|
|
|
2019-10-22 11:23:22 +01:00
|
|
|
listResponse, err := responses[numOfSegments+2].ListSegment()
|
2019-08-06 15:56:23 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, numOfSegments, len(listResponse.Items))
|
|
|
|
|
|
|
|
requests = make([]metainfo.BatchItem, 0)
|
2019-10-22 11:23:22 +01:00
|
|
|
requests = append(requests, &metainfo.GetObjectParams{
|
|
|
|
Bucket: []byte("second-test-bucket"),
|
|
|
|
EncryptedPath: []byte("encrypted-path"),
|
|
|
|
})
|
2019-08-06 15:56:23 +01:00
|
|
|
for _, segment := range listResponse.Items {
|
|
|
|
requests = append(requests, &metainfo.DownloadSegmentParams{
|
|
|
|
Position: segment.Position,
|
|
|
|
})
|
|
|
|
}
|
|
|
|
responses, err = metainfoClient.Batch(ctx, requests...)
|
|
|
|
require.NoError(t, err)
|
2019-10-22 11:23:22 +01:00
|
|
|
require.Equal(t, len(listResponse.Items)+1, len(responses))
|
2019-08-06 15:56:23 +01:00
|
|
|
|
2019-10-22 11:23:22 +01:00
|
|
|
for i, response := range responses[1:] {
|
2019-08-06 15:56:23 +01:00
|
|
|
downloadResponse, err := response.DownloadSegment()
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
require.Equal(t, expectedData[i], downloadResponse.Info.EncryptedInlineData)
|
|
|
|
}
|
|
|
|
}
|
2019-10-22 11:23:22 +01:00
|
|
|
|
|
|
|
{ // test case when StreamID is not set automatically
|
|
|
|
err := planet.Uplinks[0].CreateBucket(ctx, planet.Satellites[0], "third-test-bucket")
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2020-01-20 14:20:52 +00:00
|
|
|
beginObjectResp, err := metainfoClient.BeginObject(ctx, metainfo.BeginObjectParams{
|
2019-10-22 11:23:22 +01:00
|
|
|
Bucket: []byte("third-test-bucket"),
|
|
|
|
EncryptedPath: []byte("encrypted-path"),
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
requests := make([]metainfo.BatchItem, 0)
|
|
|
|
numOfSegments := 10
|
|
|
|
expectedData := make([][]byte, numOfSegments)
|
|
|
|
for i := 0; i < numOfSegments; i++ {
|
|
|
|
expectedData[i] = testrand.Bytes(memory.KiB)
|
|
|
|
|
|
|
|
requests = append(requests, &metainfo.MakeInlineSegmentParams{
|
2020-01-20 14:20:52 +00:00
|
|
|
StreamID: beginObjectResp.StreamID,
|
2019-10-22 11:23:22 +01:00
|
|
|
Position: storj.SegmentPosition{
|
|
|
|
Index: int32(i),
|
|
|
|
},
|
|
|
|
EncryptedInlineData: expectedData[i],
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2020-04-08 13:08:57 +01:00
|
|
|
metadata, err := pb.Marshal(&pb.StreamMeta{
|
2019-11-24 21:14:51 +00:00
|
|
|
NumberOfSegments: int64(numOfSegments),
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
2019-10-22 11:23:22 +01:00
|
|
|
requests = append(requests, &metainfo.CommitObjectParams{
|
2020-01-20 14:20:52 +00:00
|
|
|
StreamID: beginObjectResp.StreamID,
|
2019-11-24 21:14:51 +00:00
|
|
|
EncryptedMetadata: metadata,
|
2019-10-22 11:23:22 +01:00
|
|
|
})
|
|
|
|
|
|
|
|
responses, err := metainfoClient.Batch(ctx, requests...)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, numOfSegments+1, len(responses))
|
|
|
|
}
|
2019-08-06 15:56:23 +01:00
|
|
|
})
|
|
|
|
}
|
2019-10-31 19:04:33 +00:00
|
|
|
|
2020-01-17 15:01:36 +00:00
|
|
|
func TestRateLimit(t *testing.T) {
|
2020-01-29 15:22:22 +00:00
|
|
|
rateLimit := 2
|
2020-01-17 15:01:36 +00:00
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 1,
|
|
|
|
Reconfigure: testplanet.Reconfigure{
|
|
|
|
Satellite: func(log *zap.Logger, index int, config *satellite.Config) {
|
2020-01-29 15:22:22 +00:00
|
|
|
config.Metainfo.RateLimiter.Rate = float64(rateLimit)
|
2020-05-26 09:05:43 +01:00
|
|
|
config.Metainfo.RateLimiter.CacheExpiration = 500 * time.Millisecond
|
2020-01-17 15:01:36 +00:00
|
|
|
},
|
|
|
|
},
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
ul := planet.Uplinks[0]
|
|
|
|
satellite := planet.Satellites[0]
|
|
|
|
|
2020-05-26 09:05:43 +01:00
|
|
|
// TODO find a way to reset limiter before test is executed, currently
|
|
|
|
// testplanet is doing one additional request to get access
|
|
|
|
time.Sleep(1 * time.Second)
|
|
|
|
|
2020-01-29 15:22:22 +00:00
|
|
|
var group errs2.Group
|
|
|
|
for i := 0; i <= rateLimit; i++ {
|
|
|
|
group.Go(func() error {
|
|
|
|
return ul.CreateBucket(ctx, satellite, testrand.BucketName())
|
|
|
|
})
|
|
|
|
}
|
|
|
|
groupErrs := group.Wait()
|
|
|
|
require.Len(t, groupErrs, 1)
|
2020-01-17 15:01:36 +00:00
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestRateLimit_Disabled(t *testing.T) {
|
2020-01-29 15:22:22 +00:00
|
|
|
rateLimit := 2
|
2020-01-17 15:01:36 +00:00
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 1,
|
|
|
|
Reconfigure: testplanet.Reconfigure{
|
|
|
|
Satellite: func(log *zap.Logger, index int, config *satellite.Config) {
|
|
|
|
config.Metainfo.RateLimiter.Enabled = false
|
2020-01-29 15:22:22 +00:00
|
|
|
config.Metainfo.RateLimiter.Rate = float64(rateLimit)
|
2020-01-17 15:01:36 +00:00
|
|
|
},
|
|
|
|
},
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
ul := planet.Uplinks[0]
|
|
|
|
satellite := planet.Satellites[0]
|
|
|
|
|
2020-01-29 15:22:22 +00:00
|
|
|
var group errs2.Group
|
|
|
|
for i := 0; i <= rateLimit; i++ {
|
|
|
|
group.Go(func() error {
|
|
|
|
return ul.CreateBucket(ctx, satellite, testrand.BucketName())
|
|
|
|
})
|
|
|
|
}
|
|
|
|
groupErrs := group.Wait()
|
|
|
|
require.Len(t, groupErrs, 0)
|
2020-01-17 15:01:36 +00:00
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestRateLimit_ProjectRateLimitOverride(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 1, UplinkCount: 1,
|
|
|
|
Reconfigure: testplanet.Reconfigure{
|
|
|
|
Satellite: func(log *zap.Logger, index int, config *satellite.Config) {
|
|
|
|
config.Metainfo.RateLimiter.Rate = 2
|
2020-05-26 09:05:43 +01:00
|
|
|
config.Metainfo.RateLimiter.CacheExpiration = 500 * time.Millisecond
|
2020-01-17 15:01:36 +00:00
|
|
|
},
|
|
|
|
},
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
ul := planet.Uplinks[0]
|
|
|
|
satellite := planet.Satellites[0]
|
|
|
|
|
2020-05-26 09:05:43 +01:00
|
|
|
// TODO find a way to reset limiter before test is executed, currently
|
|
|
|
// testplanet is doing one additional request to get access
|
|
|
|
time.Sleep(1 * time.Second)
|
|
|
|
|
2020-01-17 15:01:36 +00:00
|
|
|
projects, err := satellite.DB.Console().Projects().GetAll(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, projects, 1)
|
|
|
|
|
|
|
|
rateLimit := 3
|
|
|
|
projects[0].RateLimit = &rateLimit
|
|
|
|
|
|
|
|
err = satellite.DB.Console().Projects().Update(ctx, &projects[0])
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2020-01-29 15:22:22 +00:00
|
|
|
var group errs2.Group
|
|
|
|
for i := 0; i <= rateLimit; i++ {
|
|
|
|
group.Go(func() error {
|
|
|
|
return ul.CreateBucket(ctx, satellite, testrand.BucketName())
|
|
|
|
})
|
|
|
|
}
|
|
|
|
groupErrs := group.Wait()
|
|
|
|
require.Len(t, groupErrs, 1)
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestRateLimit_ProjectRateLimitOverrideCachedExpired(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
2020-04-29 12:20:02 +01:00
|
|
|
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 1,
|
2020-01-29 15:22:22 +00:00
|
|
|
Reconfigure: testplanet.Reconfigure{
|
|
|
|
Satellite: func(log *zap.Logger, index int, config *satellite.Config) {
|
|
|
|
config.Metainfo.RateLimiter.Rate = 2
|
2020-04-29 12:20:02 +01:00
|
|
|
config.Metainfo.RateLimiter.CacheExpiration = time.Second
|
2020-01-29 15:22:22 +00:00
|
|
|
},
|
|
|
|
},
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
ul := planet.Uplinks[0]
|
|
|
|
satellite := planet.Satellites[0]
|
|
|
|
|
2020-05-26 09:05:43 +01:00
|
|
|
// TODO find a way to reset limiter before test is executed, currently
|
|
|
|
// testplanet is doing one additional request to get access
|
|
|
|
time.Sleep(2 * time.Second)
|
|
|
|
|
2020-01-29 15:22:22 +00:00
|
|
|
projects, err := satellite.DB.Console().Projects().GetAll(ctx)
|
2020-01-17 15:01:36 +00:00
|
|
|
require.NoError(t, err)
|
2020-01-29 15:22:22 +00:00
|
|
|
require.Len(t, projects, 1)
|
|
|
|
|
|
|
|
rateLimit := 3
|
|
|
|
projects[0].RateLimit = &rateLimit
|
|
|
|
|
|
|
|
err = satellite.DB.Console().Projects().Update(ctx, &projects[0])
|
2020-01-17 15:01:36 +00:00
|
|
|
require.NoError(t, err)
|
2020-01-29 15:22:22 +00:00
|
|
|
|
|
|
|
var group1 errs2.Group
|
|
|
|
|
|
|
|
for i := 0; i <= rateLimit; i++ {
|
|
|
|
group1.Go(func() error {
|
|
|
|
return ul.CreateBucket(ctx, satellite, testrand.BucketName())
|
|
|
|
})
|
|
|
|
}
|
|
|
|
group1Errs := group1.Wait()
|
|
|
|
require.Len(t, group1Errs, 1)
|
|
|
|
|
|
|
|
rateLimit = 1
|
|
|
|
projects[0].RateLimit = &rateLimit
|
|
|
|
|
|
|
|
err = satellite.DB.Console().Projects().Update(ctx, &projects[0])
|
2020-01-17 15:01:36 +00:00
|
|
|
require.NoError(t, err)
|
2020-01-29 15:22:22 +00:00
|
|
|
|
2020-04-29 12:20:02 +01:00
|
|
|
time.Sleep(2 * time.Second)
|
2020-01-29 15:22:22 +00:00
|
|
|
|
|
|
|
var group2 errs2.Group
|
|
|
|
|
|
|
|
for i := 0; i <= rateLimit; i++ {
|
|
|
|
group2.Go(func() error {
|
|
|
|
return ul.CreateBucket(ctx, satellite, testrand.BucketName())
|
|
|
|
})
|
|
|
|
}
|
|
|
|
group2Errs := group2.Wait()
|
|
|
|
require.Len(t, group2Errs, 1)
|
2020-01-17 15:01:36 +00:00
|
|
|
})
|
|
|
|
}
|
2020-02-11 17:25:35 +00:00
|
|
|
|
|
|
|
func TestOverwriteZombieSegments(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 4, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
uplink := planet.Uplinks[0]
|
|
|
|
metainfoClient, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(metainfoClient.Close)
|
|
|
|
|
|
|
|
var testCases = []struct {
|
|
|
|
deletedSegments []int32
|
|
|
|
objectSize memory.Size
|
|
|
|
segmentSize memory.Size
|
|
|
|
label string
|
|
|
|
}{
|
|
|
|
{label: "inline", deletedSegments: []int32{0, -1}, objectSize: 5 * memory.KiB, segmentSize: 1 * memory.KiB},
|
|
|
|
{label: "remote", deletedSegments: []int32{0, -1}, objectSize: 23 * memory.KiB, segmentSize: 5 * memory.KiB},
|
|
|
|
}
|
|
|
|
|
|
|
|
for i, tc := range testCases {
|
|
|
|
i := i
|
|
|
|
tc := tc
|
|
|
|
t.Run(tc.label, func(t *testing.T) {
|
|
|
|
data := testrand.Bytes(tc.objectSize)
|
|
|
|
bucket := "testbucket" + strconv.Itoa(i)
|
|
|
|
objectKey := "test-path" + strconv.Itoa(i)
|
2020-05-26 09:05:43 +01:00
|
|
|
uploadCtx := testuplink.WithMaxSegmentSize(ctx, tc.segmentSize)
|
|
|
|
err := uplink.Upload(uploadCtx, planet.Satellites[0], bucket, objectKey, data)
|
2020-02-11 17:25:35 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
items, _, err := metainfoClient.ListObjects(ctx, metainfo.ListObjectsParams{
|
|
|
|
Bucket: []byte(bucket),
|
|
|
|
Limit: 1,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
object, err := metainfoClient.GetObject(ctx, metainfo.GetObjectParams{
|
|
|
|
Bucket: []byte(bucket),
|
|
|
|
EncryptedPath: items[0].EncryptedPath,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// delete some segments to leave only zombie segments
|
|
|
|
for _, segment := range tc.deletedSegments {
|
|
|
|
_, _, _, err = metainfoClient.BeginDeleteSegment(ctx, metainfo.BeginDeleteSegmentParams{
|
|
|
|
StreamID: object.StreamID,
|
|
|
|
Position: storj.SegmentPosition{
|
|
|
|
Index: segment,
|
|
|
|
},
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
|
2020-05-26 09:05:43 +01:00
|
|
|
err = uplink.Upload(uploadCtx, planet.Satellites[0], bucket, objectKey, data)
|
2020-02-11 17:25:35 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
})
|
|
|
|
|
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
2020-02-14 12:52:00 +00:00
|
|
|
|
|
|
|
func TestBucketEmptinessBeforeDelete(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
for i := 0; i < 10; i++ {
|
|
|
|
err := planet.Uplinks[0].Upload(ctx, planet.Satellites[0], "test-bucket", "object-key"+strconv.Itoa(i), testrand.Bytes(memory.KiB))
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
for i := 0; i < 10; i++ {
|
|
|
|
err := planet.Uplinks[0].DeleteBucket(ctx, planet.Satellites[0], "test-bucket")
|
|
|
|
require.Error(t, err)
|
2020-03-30 10:08:02 +01:00
|
|
|
require.True(t, errors.Is(err, uplink.ErrBucketNotEmpty))
|
2020-02-14 12:52:00 +00:00
|
|
|
|
|
|
|
err = planet.Uplinks[0].DeleteObject(ctx, planet.Satellites[0], "test-bucket", "object-key"+strconv.Itoa(i))
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
err := planet.Uplinks[0].DeleteBucket(ctx, planet.Satellites[0], "test-bucket")
|
|
|
|
require.NoError(t, err)
|
|
|
|
})
|
|
|
|
}
|
2020-04-02 08:45:51 +01:00
|
|
|
|
|
|
|
func TestDeleteBatchWithoutPermission(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
|
|
|
|
err := planet.Uplinks[0].CreateBucket(ctx, planet.Satellites[0], "test-bucket")
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
apiKey, err = apiKey.Restrict(macaroon.Caveat{
|
|
|
|
DisallowLists: true,
|
|
|
|
DisallowReads: true,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
metainfoClient, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(metainfoClient.Close)
|
|
|
|
|
|
|
|
responses, err := metainfoClient.Batch(ctx,
|
|
|
|
// this request was causing panic becase for deleting object
|
|
|
|
// its possible to return no error and empty response for
|
|
|
|
// specific set of permissions, see `apiKey.Restrict` from above
|
|
|
|
&metainfo.BeginDeleteObjectParams{
|
|
|
|
Bucket: []byte("test-bucket"),
|
|
|
|
EncryptedPath: []byte("not-existing-object"),
|
|
|
|
},
|
|
|
|
|
|
|
|
// TODO this code should be enabled then issue with read permissions in
|
|
|
|
// DeleteBucket method currently user have always permission to read bucket
|
|
|
|
// https://storjlabs.atlassian.net/browse/USR-603
|
|
|
|
// when it will be fixed commented code from bellow should replace existing DeleteBucketParams
|
|
|
|
// the same situation like above
|
|
|
|
// &metainfo.DeleteBucketParams{
|
|
|
|
// Name: []byte("not-existing-bucket"),
|
|
|
|
// },
|
|
|
|
|
|
|
|
&metainfo.DeleteBucketParams{
|
|
|
|
Name: []byte("test-bucket"),
|
|
|
|
},
|
|
|
|
)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, 2, len(responses))
|
|
|
|
})
|
|
|
|
}
|
2020-04-09 09:19:16 +01:00
|
|
|
|
|
|
|
func TestInlineSegmentThreshold(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 4, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
2020-04-16 13:12:46 +01:00
|
|
|
projectID := planet.Uplinks[0].Projects[0].ID
|
2020-04-09 09:19:16 +01:00
|
|
|
|
|
|
|
{ // limit is max inline segment size + encryption overhead
|
|
|
|
err := planet.Uplinks[0].Upload(ctx, planet.Satellites[0], "test-bucket-inline", "inline-object", testrand.Bytes(4*memory.KiB))
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// we don't know encrypted path
|
|
|
|
prefix, err := satMetainfo.CreatePath(ctx, projectID, -1, []byte("test-bucket-inline"), []byte{})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
items, _, err := planet.Satellites[0].Metainfo.Service.List(ctx, prefix, "", false, 0, meta.All)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, 1, len(items))
|
|
|
|
|
|
|
|
pointer, err := planet.Satellites[0].Metainfo.Service.Get(ctx, prefix+"/"+items[0].Path)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, pb.Pointer_INLINE, pointer.Type)
|
|
|
|
}
|
|
|
|
|
|
|
|
{ // one more byte over limit should enough to create remote segment
|
|
|
|
err := planet.Uplinks[0].Upload(ctx, planet.Satellites[0], "test-bucket-remote", "remote-object", testrand.Bytes(4*memory.KiB+1))
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// we don't know encrypted path
|
|
|
|
prefix, err := satMetainfo.CreatePath(ctx, projectID, -1, []byte("test-bucket-remote"), []byte{})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
items, _, err := planet.Satellites[0].Metainfo.Service.List(ctx, prefix, "", false, 0, meta.All)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, 1, len(items))
|
|
|
|
|
|
|
|
pointer, err := planet.Satellites[0].Metainfo.Service.Get(ctx, prefix+"/"+items[0].Path)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, pb.Pointer_REMOTE, pointer.Type)
|
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
2020-06-01 21:07:31 +01:00
|
|
|
|
|
|
|
// TestCommitObjectMetadataSize ensures that CommitObject returns an error when the metadata provided by the user is too large.
|
|
|
|
func TestCommitObjectMetadataSize(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 4, UplinkCount: 1,
|
|
|
|
Reconfigure: testplanet.Reconfigure{
|
|
|
|
Satellite: testplanet.MaxMetadataSize(2 * memory.KiB),
|
|
|
|
},
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
metainfoService := planet.Satellites[0].Metainfo.Service
|
|
|
|
|
|
|
|
projects, err := planet.Satellites[0].DB.Console().Projects().GetAll(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
projectID := projects[0].ID
|
|
|
|
|
|
|
|
bucket := storj.Bucket{
|
|
|
|
Name: "initial-bucket",
|
|
|
|
ProjectID: projectID,
|
|
|
|
}
|
|
|
|
_, err = metainfoService.CreateBucket(ctx, bucket)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
metainfoClient, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(metainfoClient.Close)
|
|
|
|
|
|
|
|
params := metainfo.BeginObjectParams{
|
|
|
|
Bucket: []byte(bucket.Name),
|
|
|
|
EncryptedPath: []byte("encrypted-path"),
|
|
|
|
Redundancy: storj.RedundancyScheme{
|
|
|
|
Algorithm: storj.ReedSolomon,
|
|
|
|
ShareSize: 256,
|
|
|
|
RequiredShares: 1,
|
|
|
|
RepairShares: 1,
|
|
|
|
OptimalShares: 3,
|
|
|
|
TotalShares: 4,
|
|
|
|
},
|
|
|
|
EncryptionParameters: storj.EncryptionParameters{},
|
|
|
|
ExpiresAt: time.Now().Add(24 * time.Hour),
|
|
|
|
}
|
|
|
|
beginObjectResponse, err := metainfoClient.BeginObject(ctx, params)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
segmentID, limits, _, err := metainfoClient.BeginSegment(ctx, metainfo.BeginSegmentParams{
|
|
|
|
StreamID: beginObjectResponse.StreamID,
|
|
|
|
Position: storj.SegmentPosition{
|
|
|
|
Index: 0,
|
|
|
|
},
|
|
|
|
MaxOrderLimit: memory.MiB.Int64(),
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
fullIDMap := make(map[storj.NodeID]*identity.FullIdentity)
|
|
|
|
for _, node := range planet.StorageNodes {
|
|
|
|
fullIDMap[node.ID()] = node.Identity
|
|
|
|
}
|
|
|
|
|
|
|
|
makeResult := func(num int32) *pb.SegmentPieceUploadResult {
|
|
|
|
nodeID := limits[num].Limit.StorageNodeId
|
|
|
|
hash := &pb.PieceHash{
|
|
|
|
PieceId: limits[num].Limit.PieceId,
|
|
|
|
PieceSize: 1048832,
|
|
|
|
Timestamp: time.Now(),
|
|
|
|
}
|
|
|
|
|
|
|
|
fullID := fullIDMap[nodeID]
|
|
|
|
require.NotNil(t, fullID)
|
|
|
|
signer := signing.SignerFromFullIdentity(fullID)
|
|
|
|
signedHash, err := signing.SignPieceHash(ctx, signer, hash)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
return &pb.SegmentPieceUploadResult{
|
|
|
|
PieceNum: num,
|
|
|
|
NodeId: nodeID,
|
|
|
|
Hash: signedHash,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
err = metainfoClient.CommitSegment(ctx, metainfo.CommitSegmentParams{
|
|
|
|
SegmentID: segmentID,
|
|
|
|
|
|
|
|
SizeEncryptedData: memory.MiB.Int64(),
|
|
|
|
UploadResult: []*pb.SegmentPieceUploadResult{
|
|
|
|
makeResult(0),
|
|
|
|
makeResult(1),
|
|
|
|
makeResult(2),
|
|
|
|
},
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// 5KiB metadata should fail because it is too large.
|
|
|
|
metadata, err := pb.Marshal(&pb.StreamMeta{
|
|
|
|
EncryptedStreamInfo: testrand.Bytes(5 * memory.KiB),
|
|
|
|
NumberOfSegments: 1,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
err = metainfoClient.CommitObject(ctx, metainfo.CommitObjectParams{
|
|
|
|
StreamID: beginObjectResponse.StreamID,
|
|
|
|
EncryptedMetadata: metadata,
|
|
|
|
})
|
|
|
|
require.Error(t, err)
|
|
|
|
assertInvalidArgument(t, err, true)
|
|
|
|
|
|
|
|
// 1KiB metadata should not fail.
|
|
|
|
metadata, err = pb.Marshal(&pb.StreamMeta{
|
|
|
|
EncryptedStreamInfo: testrand.Bytes(1 * memory.KiB),
|
|
|
|
NumberOfSegments: 1,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
err = metainfoClient.CommitObject(ctx, metainfo.CommitObjectParams{
|
|
|
|
StreamID: beginObjectResponse.StreamID,
|
|
|
|
EncryptedMetadata: metadata,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
})
|
|
|
|
|
|
|
|
}
|
2020-07-31 12:24:40 +01:00
|
|
|
|
|
|
|
func TestDeleteRightsOnUpload(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
up := planet.Uplinks[0]
|
|
|
|
|
|
|
|
err := up.CreateBucket(ctx, planet.Satellites[0], "test-bucket")
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
data := testrand.Bytes(1 * memory.KiB)
|
|
|
|
err = up.Upload(ctx, planet.Satellites[0], "test-bucket", "test-key", data)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
access := up.Access[planet.Satellites[0].ID()]
|
|
|
|
|
|
|
|
overwrite := func(allowDelete bool) error {
|
|
|
|
permission := uplink.FullPermission()
|
|
|
|
permission.AllowDelete = allowDelete
|
|
|
|
|
|
|
|
sharedAccess, err := access.Share(permission)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
project, err := uplink.OpenProject(ctx, sharedAccess)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(project.Close)
|
|
|
|
|
|
|
|
upload, err := project.UploadObject(ctx, "test-bucket", "test-key", nil)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
_, err = upload.Write([]byte("new data"))
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
return upload.Commit()
|
|
|
|
}
|
|
|
|
|
|
|
|
require.Error(t, overwrite(false))
|
|
|
|
require.NoError(t, overwrite(true))
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestImmutableUpload(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
access := planet.Uplinks[0].Access[planet.Satellites[0].ID()]
|
|
|
|
|
|
|
|
permission := uplink.Permission{AllowUpload: true} // AllowDelete: false
|
|
|
|
sharedAccess, err := access.Share(permission)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
project, err := uplink.OpenProject(ctx, sharedAccess)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(project.Close)
|
|
|
|
|
|
|
|
_, err = project.CreateBucket(ctx, "test-bucket")
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// Uploading the object for first time should be successful.
|
|
|
|
upload, err := project.UploadObject(ctx, "test-bucket", "test-key", nil)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
_, err = upload.Write(testrand.Bytes(1 * memory.KiB))
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
err = upload.Commit()
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// Overwriting the object should fail on Commit.
|
|
|
|
upload, err = project.UploadObject(ctx, "test-bucket", "test-key", nil)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
_, err = upload.Write(testrand.Bytes(1 * memory.KiB))
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
err = upload.Commit()
|
|
|
|
require.Error(t, err)
|
|
|
|
})
|
|
|
|
}
|