2022-01-26 11:10:28 +00:00
|
|
|
// Copyright (C) 2019 Storj Labs, Inc.
|
|
|
|
// See LICENSE for copying information.
|
|
|
|
|
|
|
|
package metainfo_test
|
|
|
|
|
|
|
|
import (
|
|
|
|
"strconv"
|
|
|
|
"testing"
|
|
|
|
"time"
|
|
|
|
|
|
|
|
"github.com/stretchr/testify/assert"
|
|
|
|
"github.com/stretchr/testify/require"
|
|
|
|
|
|
|
|
"storj.io/common/errs2"
|
|
|
|
"storj.io/common/memory"
|
|
|
|
"storj.io/common/pb"
|
|
|
|
"storj.io/common/rpc/rpcstatus"
|
|
|
|
"storj.io/common/signing"
|
|
|
|
"storj.io/common/storj"
|
|
|
|
"storj.io/common/testcontext"
|
|
|
|
"storj.io/common/testrand"
|
|
|
|
"storj.io/storj/private/testplanet"
|
|
|
|
"storj.io/storj/satellite/metabase"
|
|
|
|
"storj.io/uplink/private/metaclient"
|
|
|
|
)
|
|
|
|
|
|
|
|
func TestExpirationTimeSegment(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
|
|
|
|
err := planet.Uplinks[0].CreateBucket(ctx, planet.Satellites[0], "my-bucket-name")
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
metainfoClient, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(metainfoClient.Close)
|
|
|
|
|
|
|
|
for i, r := range []struct {
|
|
|
|
expirationDate time.Time
|
|
|
|
errFlag bool
|
|
|
|
}{
|
|
|
|
{ // expiration time not set
|
|
|
|
time.Time{},
|
|
|
|
false,
|
|
|
|
},
|
|
|
|
{ // 10 days into future
|
|
|
|
time.Now().AddDate(0, 0, 10),
|
|
|
|
false,
|
|
|
|
},
|
|
|
|
{ // current time
|
|
|
|
time.Now(),
|
|
|
|
true,
|
|
|
|
},
|
|
|
|
{ // 10 days into past
|
|
|
|
time.Now().AddDate(0, 0, -10),
|
|
|
|
true,
|
|
|
|
},
|
|
|
|
} {
|
|
|
|
_, err := metainfoClient.BeginObject(ctx, metaclient.BeginObjectParams{
|
|
|
|
Bucket: []byte("my-bucket-name"),
|
|
|
|
EncryptedObjectKey: []byte("path" + strconv.Itoa(i)),
|
|
|
|
ExpiresAt: r.expirationDate,
|
|
|
|
EncryptionParameters: storj.EncryptionParameters{
|
|
|
|
CipherSuite: storj.EncAESGCM,
|
|
|
|
BlockSize: 256,
|
|
|
|
},
|
|
|
|
})
|
|
|
|
if r.errFlag {
|
|
|
|
assert.Error(t, err)
|
|
|
|
} else {
|
|
|
|
assert.NoError(t, err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestInlineSegment(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 0, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
|
|
|
|
buckets := planet.Satellites[0].API.Buckets.Service
|
|
|
|
|
|
|
|
// TODO maybe split into separate cases
|
|
|
|
// Test:
|
|
|
|
// * create bucket
|
|
|
|
// * begin object
|
|
|
|
// * send several inline segments
|
|
|
|
// * commit object
|
|
|
|
// * list created object
|
|
|
|
// * list object segments
|
|
|
|
// * download segments
|
|
|
|
// * delete segments and object
|
|
|
|
|
|
|
|
bucket := storj.Bucket{
|
|
|
|
Name: "inline-segments-bucket",
|
|
|
|
ProjectID: planet.Uplinks[0].Projects[0].ID,
|
|
|
|
}
|
|
|
|
_, err := buckets.CreateBucket(ctx, bucket)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
metainfoClient, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(metainfoClient.Close)
|
|
|
|
|
|
|
|
params := metaclient.BeginObjectParams{
|
|
|
|
Bucket: []byte(bucket.Name),
|
|
|
|
EncryptedObjectKey: []byte("encrypted-path"),
|
|
|
|
Redundancy: storj.RedundancyScheme{
|
|
|
|
Algorithm: storj.ReedSolomon,
|
|
|
|
ShareSize: 256,
|
|
|
|
RequiredShares: 1,
|
|
|
|
RepairShares: 1,
|
|
|
|
OptimalShares: 3,
|
|
|
|
TotalShares: 4,
|
|
|
|
},
|
|
|
|
EncryptionParameters: storj.EncryptionParameters{
|
|
|
|
CipherSuite: storj.EncAESGCM,
|
|
|
|
BlockSize: 256,
|
|
|
|
},
|
|
|
|
|
|
|
|
ExpiresAt: time.Now().Add(24 * time.Hour),
|
|
|
|
}
|
|
|
|
beginObjectResp, err := metainfoClient.BeginObject(ctx, params)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
segments := []int32{0, 1, 2, 3, 4, 5, 6}
|
|
|
|
segmentsData := make([][]byte, len(segments))
|
|
|
|
for i, segment := range segments {
|
|
|
|
segmentsData[i] = testrand.Bytes(memory.KiB)
|
|
|
|
err = metainfoClient.MakeInlineSegment(ctx, metaclient.MakeInlineSegmentParams{
|
|
|
|
StreamID: beginObjectResp.StreamID,
|
2022-11-07 13:48:31 +00:00
|
|
|
Position: metaclient.SegmentPosition{
|
2022-01-26 11:10:28 +00:00
|
|
|
Index: segment,
|
|
|
|
},
|
|
|
|
PlainSize: 1024,
|
|
|
|
EncryptedInlineData: segmentsData[i],
|
2022-11-07 13:48:31 +00:00
|
|
|
Encryption: metaclient.SegmentEncryption{
|
2022-01-26 11:10:28 +00:00
|
|
|
EncryptedKey: testrand.Bytes(256),
|
|
|
|
},
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
metadata, err := pb.Marshal(&pb.StreamMeta{
|
|
|
|
NumberOfSegments: int64(len(segments)),
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
err = metainfoClient.CommitObject(ctx, metaclient.CommitObjectParams{
|
|
|
|
StreamID: beginObjectResp.StreamID,
|
|
|
|
EncryptedMetadata: metadata,
|
|
|
|
EncryptedMetadataNonce: testrand.Nonce(),
|
2022-04-21 15:25:16 +01:00
|
|
|
EncryptedMetadataEncryptedKey: randomEncryptedKey,
|
2022-01-26 11:10:28 +00:00
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
objects, _, err := metainfoClient.ListObjects(ctx, metaclient.ListObjectsParams{
|
|
|
|
Bucket: []byte(bucket.Name),
|
|
|
|
IncludeSystemMetadata: true,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, objects, 1)
|
|
|
|
|
|
|
|
require.Equal(t, params.EncryptedObjectKey, objects[0].EncryptedObjectKey)
|
2022-06-21 10:03:56 +01:00
|
|
|
// TODO find better way to compare (one ExpiresAt contains time zone information)
|
2022-01-26 11:10:28 +00:00
|
|
|
require.Equal(t, params.ExpiresAt.Unix(), objects[0].ExpiresAt.Unix())
|
|
|
|
|
|
|
|
object, err := metainfoClient.GetObject(ctx, metaclient.GetObjectParams{
|
|
|
|
Bucket: params.Bucket,
|
|
|
|
EncryptedObjectKey: params.EncryptedObjectKey,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
{ // Confirm data larger than our configured max inline segment size of 4 KiB cannot be inlined
|
|
|
|
beginObjectResp, err := metainfoClient.BeginObject(ctx, metaclient.BeginObjectParams{
|
|
|
|
Bucket: []byte(bucket.Name),
|
|
|
|
EncryptedObjectKey: []byte("too-large-inline-segment"),
|
|
|
|
EncryptionParameters: storj.EncryptionParameters{
|
|
|
|
CipherSuite: storj.EncAESGCM,
|
|
|
|
BlockSize: 256,
|
|
|
|
},
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
data := testrand.Bytes(10 * memory.KiB)
|
|
|
|
err = metainfoClient.MakeInlineSegment(ctx, metaclient.MakeInlineSegmentParams{
|
|
|
|
StreamID: beginObjectResp.StreamID,
|
2022-11-07 13:48:31 +00:00
|
|
|
Position: metaclient.SegmentPosition{
|
2022-01-26 11:10:28 +00:00
|
|
|
Index: 0,
|
|
|
|
},
|
|
|
|
EncryptedInlineData: data,
|
2022-11-07 13:48:31 +00:00
|
|
|
Encryption: metaclient.SegmentEncryption{
|
2022-01-26 11:10:28 +00:00
|
|
|
EncryptedKey: testrand.Bytes(256),
|
|
|
|
},
|
|
|
|
})
|
|
|
|
require.Error(t, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
{ // test download inline segments
|
|
|
|
existingSegments := []int32{0, 1, 2, 3, 4, 5, -1}
|
|
|
|
|
|
|
|
for i, index := range existingSegments {
|
|
|
|
response, err := metainfoClient.DownloadSegmentWithRS(ctx, metaclient.DownloadSegmentParams{
|
|
|
|
StreamID: object.StreamID,
|
2022-11-07 13:48:31 +00:00
|
|
|
Position: metaclient.SegmentPosition{
|
2022-01-26 11:10:28 +00:00
|
|
|
Index: index,
|
|
|
|
},
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Nil(t, response.Limits)
|
|
|
|
require.Equal(t, segmentsData[i], response.Info.EncryptedInlineData)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
{ // test deleting segments
|
|
|
|
_, err := metainfoClient.BeginDeleteObject(ctx, metaclient.BeginDeleteObjectParams{
|
|
|
|
Bucket: params.Bucket,
|
|
|
|
EncryptedObjectKey: params.EncryptedObjectKey,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
_, err = metainfoClient.GetObject(ctx, metaclient.GetObjectParams{
|
|
|
|
Bucket: params.Bucket,
|
|
|
|
EncryptedObjectKey: params.EncryptedObjectKey,
|
|
|
|
})
|
|
|
|
require.Error(t, err)
|
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestRemoteSegment(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 4, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
uplink := planet.Uplinks[0]
|
|
|
|
|
|
|
|
expectedBucketName := "remote-segments-bucket"
|
|
|
|
err := uplink.Upload(ctx, planet.Satellites[0], expectedBucketName, "file-object", testrand.Bytes(50*memory.KiB))
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
metainfoClient, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(metainfoClient.Close)
|
|
|
|
|
|
|
|
items, _, err := metainfoClient.ListObjects(ctx, metaclient.ListObjectsParams{
|
|
|
|
Bucket: []byte(expectedBucketName),
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, items, 1)
|
|
|
|
|
|
|
|
{
|
|
|
|
// Get object
|
|
|
|
// Download segment
|
|
|
|
|
|
|
|
object, err := metainfoClient.GetObject(ctx, metaclient.GetObjectParams{
|
|
|
|
Bucket: []byte(expectedBucketName),
|
|
|
|
EncryptedObjectKey: items[0].EncryptedObjectKey,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
response, err := metainfoClient.DownloadSegmentWithRS(ctx, metaclient.DownloadSegmentParams{
|
|
|
|
StreamID: object.StreamID,
|
2022-11-07 13:48:31 +00:00
|
|
|
Position: metaclient.SegmentPosition{
|
2022-01-26 11:10:28 +00:00
|
|
|
Index: -1,
|
|
|
|
},
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.NotEmpty(t, response.Limits)
|
|
|
|
}
|
|
|
|
|
|
|
|
{
|
|
|
|
// Download Object
|
|
|
|
download, err := metainfoClient.DownloadObject(ctx, metaclient.DownloadObjectParams{
|
|
|
|
Bucket: []byte(expectedBucketName),
|
|
|
|
EncryptedObjectKey: items[0].EncryptedObjectKey,
|
|
|
|
Range: metaclient.StreamRange{
|
|
|
|
Mode: metaclient.StreamRangeStartLimit,
|
|
|
|
Start: 1,
|
|
|
|
Limit: 2,
|
|
|
|
},
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, download.DownloadedSegments, 1)
|
|
|
|
require.NotEmpty(t, download.DownloadedSegments[0].Limits)
|
|
|
|
for _, limit := range download.DownloadedSegments[0].Limits {
|
|
|
|
if limit == nil {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
// requested download size is
|
|
|
|
// [1:2}
|
|
|
|
// calculating encryption input block size (7408) indices gives us:
|
|
|
|
// 0 and 1
|
|
|
|
// converting these into output block size (7424), gives us:
|
|
|
|
// [0:7424}
|
|
|
|
// this aligned to stripe size (256), gives us:
|
|
|
|
// [0:7424}
|
|
|
|
require.Equal(t, int64(7424), limit.Limit.Limit)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
{
|
|
|
|
// Begin deleting object
|
|
|
|
// List objects
|
|
|
|
|
|
|
|
_, err := metainfoClient.BeginDeleteObject(ctx, metaclient.BeginDeleteObjectParams{
|
|
|
|
Bucket: []byte(expectedBucketName),
|
|
|
|
EncryptedObjectKey: items[0].EncryptedObjectKey,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
items, _, err = metainfoClient.ListObjects(ctx, metaclient.ListObjectsParams{
|
|
|
|
Bucket: []byte(expectedBucketName),
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, items, 0)
|
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestInlineSegmentThreshold(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 4, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
{ // limit is max inline segment size + encryption overhead
|
|
|
|
err := planet.Uplinks[0].Upload(ctx, planet.Satellites[0], "test-bucket-inline", "inline-object", testrand.Bytes(4*memory.KiB))
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// we don't know encrypted path
|
|
|
|
segments, err := planet.Satellites[0].Metabase.DB.TestingAllSegments(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, segments, 1)
|
|
|
|
require.Zero(t, segments[0].Redundancy)
|
|
|
|
require.NotEmpty(t, segments[0].InlineData)
|
|
|
|
|
|
|
|
// clean up - delete the uploaded object
|
|
|
|
objects, err := planet.Satellites[0].Metabase.DB.TestingAllObjects(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, objects, 1)
|
2022-01-28 12:34:34 +00:00
|
|
|
_, err = planet.Satellites[0].Metabase.DB.DeleteObjectExactVersion(ctx, metabase.DeleteObjectExactVersion{
|
2022-01-26 11:10:28 +00:00
|
|
|
ObjectLocation: objects[0].Location(),
|
2022-01-28 12:34:34 +00:00
|
|
|
Version: metabase.DefaultVersion,
|
2022-01-26 11:10:28 +00:00
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
{ // one more byte over limit should enough to create remote segment
|
|
|
|
err := planet.Uplinks[0].Upload(ctx, planet.Satellites[0], "test-bucket-remote", "remote-object", testrand.Bytes(4*memory.KiB+1))
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// we don't know encrypted path
|
|
|
|
segments, err := planet.Satellites[0].Metabase.DB.TestingAllSegments(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, segments, 1)
|
|
|
|
require.NotZero(t, segments[0].Redundancy)
|
|
|
|
require.Empty(t, segments[0].InlineData)
|
|
|
|
|
|
|
|
// clean up - delete the uploaded object
|
|
|
|
objects, err := planet.Satellites[0].Metabase.DB.TestingAllObjects(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, objects, 1)
|
2022-01-28 12:34:34 +00:00
|
|
|
_, err = planet.Satellites[0].Metabase.DB.DeleteObjectExactVersion(ctx, metabase.DeleteObjectExactVersion{
|
2022-01-26 11:10:28 +00:00
|
|
|
ObjectLocation: objects[0].Location(),
|
2022-01-28 12:34:34 +00:00
|
|
|
Version: metabase.DefaultVersion,
|
2022-01-26 11:10:28 +00:00
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestObjectSegmentExpiresAt(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 4, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
inlineData := testrand.Bytes(1 * memory.KiB)
|
|
|
|
inlineExpiration := time.Now().Add(2 * time.Hour)
|
|
|
|
err := planet.Uplinks[0].UploadWithExpiration(ctx, planet.Satellites[0], "hohoho", "inline_object", inlineData, inlineExpiration)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
remoteData := testrand.Bytes(10 * memory.KiB)
|
|
|
|
remoteExpiration := time.Now().Add(4 * time.Hour)
|
|
|
|
err = planet.Uplinks[0].UploadWithExpiration(ctx, planet.Satellites[0], "hohoho", "remote_object", remoteData, remoteExpiration)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
segments, err := planet.Satellites[0].Metabase.DB.TestingAllSegments(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, segments, 2)
|
|
|
|
|
|
|
|
for _, segment := range segments {
|
|
|
|
if int(segment.PlainSize) == len(inlineData) {
|
|
|
|
require.Equal(t, inlineExpiration.Unix(), segment.ExpiresAt.Unix())
|
|
|
|
} else if int(segment.PlainSize) == len(remoteData) {
|
|
|
|
require.Equal(t, remoteExpiration.Unix(), segment.ExpiresAt.Unix())
|
|
|
|
} else {
|
|
|
|
t.Fail()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestCommitSegment_Validation(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 1, UplinkCount: 1,
|
|
|
|
Reconfigure: testplanet.Reconfigure{
|
|
|
|
Satellite: testplanet.Combine(
|
|
|
|
testplanet.ReconfigureRS(1, 1, 1, 1),
|
|
|
|
),
|
|
|
|
},
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
client, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(client.Close)
|
|
|
|
|
|
|
|
err = planet.Uplinks[0].CreateBucket(ctx, planet.Satellites[0], "testbucket")
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
beginObjectResponse, err := client.BeginObject(ctx, metaclient.BeginObjectParams{
|
|
|
|
Bucket: []byte("testbucket"),
|
|
|
|
EncryptedObjectKey: []byte("a/b/testobject"),
|
|
|
|
EncryptionParameters: storj.EncryptionParameters{
|
|
|
|
CipherSuite: storj.EncAESGCM,
|
|
|
|
BlockSize: 256,
|
|
|
|
},
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
response, err := client.BeginSegment(ctx, metaclient.BeginSegmentParams{
|
|
|
|
StreamID: beginObjectResponse.StreamID,
|
2022-11-07 13:48:31 +00:00
|
|
|
Position: metaclient.SegmentPosition{
|
2022-01-26 11:10:28 +00:00
|
|
|
Index: 0,
|
|
|
|
},
|
|
|
|
MaxOrderLimit: memory.MiB.Int64(),
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// the number of results of uploaded pieces (0) is below the optimal threshold (1)
|
|
|
|
err = client.CommitSegment(ctx, metaclient.CommitSegmentParams{
|
|
|
|
SegmentID: response.SegmentID,
|
|
|
|
UploadResult: []*pb.SegmentPieceUploadResult{},
|
|
|
|
})
|
|
|
|
require.Error(t, err)
|
|
|
|
require.True(t, errs2.IsRPC(err, rpcstatus.InvalidArgument))
|
|
|
|
|
|
|
|
// piece sizes are invalid: pointer verification: no remote pieces
|
|
|
|
err = client.CommitSegment(ctx, metaclient.CommitSegmentParams{
|
|
|
|
SegmentID: response.SegmentID,
|
|
|
|
UploadResult: []*pb.SegmentPieceUploadResult{
|
|
|
|
{},
|
|
|
|
},
|
|
|
|
})
|
|
|
|
require.Error(t, err)
|
|
|
|
require.True(t, errs2.IsRPC(err, rpcstatus.InvalidArgument))
|
|
|
|
|
|
|
|
// piece sizes are invalid: pointer verification: size is invalid (-1)
|
|
|
|
err = client.CommitSegment(ctx, metaclient.CommitSegmentParams{
|
|
|
|
SegmentID: response.SegmentID,
|
|
|
|
UploadResult: []*pb.SegmentPieceUploadResult{
|
|
|
|
{
|
|
|
|
Hash: &pb.PieceHash{
|
|
|
|
PieceSize: -1,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
})
|
|
|
|
require.Error(t, err)
|
|
|
|
require.True(t, errs2.IsRPC(err, rpcstatus.InvalidArgument))
|
|
|
|
|
|
|
|
// piece sizes are invalid: pointer verification: expected size is different from provided (768 != 10000)
|
|
|
|
err = client.CommitSegment(ctx, metaclient.CommitSegmentParams{
|
|
|
|
SegmentID: response.SegmentID,
|
|
|
|
SizeEncryptedData: 512,
|
|
|
|
UploadResult: []*pb.SegmentPieceUploadResult{
|
|
|
|
{
|
|
|
|
Hash: &pb.PieceHash{
|
|
|
|
PieceSize: 10000,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
})
|
|
|
|
require.Error(t, err)
|
|
|
|
require.True(t, errs2.IsRPC(err, rpcstatus.InvalidArgument))
|
|
|
|
|
|
|
|
// piece sizes are invalid: pointer verification: sizes do not match (10000 != 9000)
|
|
|
|
err = client.CommitSegment(ctx, metaclient.CommitSegmentParams{
|
|
|
|
SegmentID: response.SegmentID,
|
|
|
|
UploadResult: []*pb.SegmentPieceUploadResult{
|
|
|
|
{
|
|
|
|
Hash: &pb.PieceHash{
|
|
|
|
PieceSize: 10000,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Hash: &pb.PieceHash{
|
|
|
|
PieceSize: 9000,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
})
|
|
|
|
require.Error(t, err)
|
|
|
|
require.True(t, errs2.IsRPC(err, rpcstatus.InvalidArgument))
|
|
|
|
|
|
|
|
// pointer verification failed: pointer verification: invalid piece number
|
|
|
|
err = client.CommitSegment(ctx, metaclient.CommitSegmentParams{
|
|
|
|
SegmentID: response.SegmentID,
|
|
|
|
SizeEncryptedData: 512,
|
|
|
|
UploadResult: []*pb.SegmentPieceUploadResult{
|
|
|
|
{
|
|
|
|
PieceNum: 10,
|
|
|
|
NodeId: response.Limits[0].Limit.StorageNodeId,
|
|
|
|
Hash: &pb.PieceHash{
|
|
|
|
PieceSize: 768,
|
|
|
|
PieceId: response.Limits[0].Limit.PieceId,
|
|
|
|
Timestamp: time.Now(),
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
})
|
|
|
|
require.Error(t, err)
|
|
|
|
require.True(t, errs2.IsRPC(err, rpcstatus.InvalidArgument))
|
|
|
|
|
|
|
|
// signature verification error (no signature)
|
|
|
|
err = client.CommitSegment(ctx, metaclient.CommitSegmentParams{
|
|
|
|
SegmentID: response.SegmentID,
|
|
|
|
SizeEncryptedData: 512,
|
|
|
|
UploadResult: []*pb.SegmentPieceUploadResult{
|
|
|
|
{
|
|
|
|
PieceNum: 0,
|
|
|
|
NodeId: response.Limits[0].Limit.StorageNodeId,
|
|
|
|
Hash: &pb.PieceHash{
|
|
|
|
PieceSize: 768,
|
|
|
|
PieceId: response.Limits[0].Limit.PieceId,
|
|
|
|
Timestamp: time.Now(),
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
})
|
|
|
|
require.Error(t, err)
|
|
|
|
require.True(t, errs2.IsRPC(err, rpcstatus.InvalidArgument))
|
|
|
|
|
|
|
|
signer := signing.SignerFromFullIdentity(planet.StorageNodes[0].Identity)
|
|
|
|
signedHash, err := signing.SignPieceHash(ctx, signer, &pb.PieceHash{
|
|
|
|
PieceSize: 768,
|
|
|
|
PieceId: response.Limits[0].Limit.PieceId,
|
|
|
|
Timestamp: time.Now(),
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// pointer verification failed: pointer verification: nil identity returned
|
|
|
|
err = client.CommitSegment(ctx, metaclient.CommitSegmentParams{
|
|
|
|
SegmentID: response.SegmentID,
|
|
|
|
SizeEncryptedData: 512,
|
|
|
|
UploadResult: []*pb.SegmentPieceUploadResult{
|
|
|
|
{
|
|
|
|
PieceNum: 0,
|
|
|
|
NodeId: testrand.NodeID(), // random node ID
|
|
|
|
Hash: signedHash,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
})
|
|
|
|
require.Error(t, err)
|
|
|
|
require.True(t, errs2.IsRPC(err, rpcstatus.InvalidArgument))
|
|
|
|
|
|
|
|
// plain segment size 513 is out of range, maximum allowed is 512
|
|
|
|
err = client.CommitSegment(ctx, metaclient.CommitSegmentParams{
|
|
|
|
SegmentID: response.SegmentID,
|
|
|
|
PlainSize: 513,
|
|
|
|
SizeEncryptedData: 512,
|
|
|
|
UploadResult: []*pb.SegmentPieceUploadResult{
|
|
|
|
{
|
|
|
|
PieceNum: 0,
|
|
|
|
NodeId: response.Limits[0].Limit.StorageNodeId,
|
|
|
|
Hash: signedHash,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
})
|
|
|
|
require.Error(t, err)
|
|
|
|
require.True(t, errs2.IsRPC(err, rpcstatus.InvalidArgument))
|
|
|
|
})
|
|
|
|
}
|