2019-03-30 11:21:49 +00:00
|
|
|
// Copyright (C) 2019 Storj Labs, Inc.
|
|
|
|
// See LICENSE for copying information.
|
|
|
|
|
|
|
|
package metainfo_test
|
|
|
|
|
|
|
|
import (
|
|
|
|
"context"
|
2019-06-13 02:35:37 +01:00
|
|
|
"crypto/rand"
|
2019-03-30 11:21:49 +00:00
|
|
|
"sort"
|
|
|
|
"testing"
|
|
|
|
"time"
|
|
|
|
|
2019-06-05 17:41:02 +01:00
|
|
|
"github.com/golang/protobuf/ptypes"
|
2019-03-30 11:21:49 +00:00
|
|
|
"github.com/stretchr/testify/assert"
|
|
|
|
"github.com/stretchr/testify/require"
|
|
|
|
"github.com/zeebo/errs"
|
2019-06-05 17:41:02 +01:00
|
|
|
"google.golang.org/grpc/codes"
|
|
|
|
"google.golang.org/grpc/status"
|
2019-03-30 11:21:49 +00:00
|
|
|
|
2019-06-05 17:41:02 +01:00
|
|
|
"storj.io/storj/internal/memory"
|
2019-03-30 11:21:49 +00:00
|
|
|
"storj.io/storj/internal/testcontext"
|
|
|
|
"storj.io/storj/internal/testplanet"
|
2019-05-24 17:51:27 +01:00
|
|
|
"storj.io/storj/pkg/macaroon"
|
2019-03-30 11:21:49 +00:00
|
|
|
"storj.io/storj/pkg/pb"
|
|
|
|
"storj.io/storj/pkg/storj"
|
|
|
|
"storj.io/storj/satellite/console"
|
2019-06-05 17:41:02 +01:00
|
|
|
"storj.io/storj/uplink/metainfo"
|
2019-03-30 11:21:49 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
// mockAPIKeys is mock for api keys store of pointerdb
|
|
|
|
type mockAPIKeys struct {
|
|
|
|
info console.APIKeyInfo
|
|
|
|
err error
|
|
|
|
}
|
|
|
|
|
|
|
|
// GetByKey return api key info for given key
|
2019-05-24 17:51:27 +01:00
|
|
|
func (keys *mockAPIKeys) GetByKey(ctx context.Context, key macaroon.APIKey) (*console.APIKeyInfo, error) {
|
2019-03-30 11:21:49 +00:00
|
|
|
return &keys.info, keys.err
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestInvalidAPIKey(t *testing.T) {
|
|
|
|
ctx := testcontext.New(t)
|
|
|
|
defer ctx.Cleanup()
|
|
|
|
|
|
|
|
planet, err := testplanet.New(t, 1, 1, 1)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(planet.Shutdown)
|
|
|
|
|
|
|
|
planet.Start(ctx)
|
|
|
|
|
|
|
|
for _, invalidAPIKey := range []string{"", "invalid", "testKey"} {
|
|
|
|
client, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], invalidAPIKey)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
_, _, err = client.CreateSegment(ctx, "hello", "world", 1, &pb.RedundancyScheme{}, 123, time.Now())
|
2019-05-24 17:51:27 +01:00
|
|
|
assertUnauthenticated(t, err, false)
|
2019-03-30 11:21:49 +00:00
|
|
|
|
|
|
|
_, err = client.CommitSegment(ctx, "testbucket", "testpath", 0, &pb.Pointer{}, nil)
|
2019-05-24 17:51:27 +01:00
|
|
|
assertUnauthenticated(t, err, false)
|
2019-03-30 11:21:49 +00:00
|
|
|
|
|
|
|
_, err = client.SegmentInfo(ctx, "testbucket", "testpath", 0)
|
2019-05-24 17:51:27 +01:00
|
|
|
assertUnauthenticated(t, err, false)
|
2019-03-30 11:21:49 +00:00
|
|
|
|
|
|
|
_, _, err = client.ReadSegment(ctx, "testbucket", "testpath", 0)
|
2019-05-24 17:51:27 +01:00
|
|
|
assertUnauthenticated(t, err, false)
|
2019-03-30 11:21:49 +00:00
|
|
|
|
|
|
|
_, err = client.DeleteSegment(ctx, "testbucket", "testpath", 0)
|
2019-05-24 17:51:27 +01:00
|
|
|
assertUnauthenticated(t, err, false)
|
2019-03-30 11:21:49 +00:00
|
|
|
|
|
|
|
_, _, err = client.ListSegments(ctx, "testbucket", "", "", "", true, 1, 0)
|
2019-05-24 17:51:27 +01:00
|
|
|
assertUnauthenticated(t, err, false)
|
2019-03-30 11:21:49 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-05-24 17:51:27 +01:00
|
|
|
func TestRestrictedAPIKey(t *testing.T) {
|
|
|
|
ctx := testcontext.New(t)
|
|
|
|
defer ctx.Cleanup()
|
|
|
|
|
|
|
|
planet, err := testplanet.New(t, 1, 1, 1)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(planet.Shutdown)
|
|
|
|
|
|
|
|
planet.Start(ctx)
|
|
|
|
|
|
|
|
key, err := macaroon.ParseAPIKey(planet.Uplinks[0].APIKey[planet.Satellites[0].ID()])
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
tests := []struct {
|
|
|
|
Caveat macaroon.Caveat
|
|
|
|
CreateSegmentAllowed bool
|
|
|
|
CommitSegmentAllowed bool
|
|
|
|
SegmentInfoAllowed bool
|
|
|
|
ReadSegmentAllowed bool
|
|
|
|
DeleteSegmentAllowed bool
|
|
|
|
ListSegmentsAllowed bool
|
|
|
|
}{
|
|
|
|
{ // Everything disallowed
|
|
|
|
Caveat: macaroon.Caveat{
|
|
|
|
DisallowReads: true,
|
|
|
|
DisallowWrites: true,
|
|
|
|
DisallowLists: true,
|
|
|
|
DisallowDeletes: true,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
|
|
|
|
{ // Read only
|
|
|
|
Caveat: macaroon.Caveat{
|
|
|
|
DisallowWrites: true,
|
|
|
|
DisallowDeletes: true,
|
|
|
|
},
|
|
|
|
SegmentInfoAllowed: true,
|
|
|
|
ReadSegmentAllowed: true,
|
|
|
|
ListSegmentsAllowed: true,
|
|
|
|
},
|
|
|
|
|
|
|
|
{ // Write only
|
|
|
|
Caveat: macaroon.Caveat{
|
|
|
|
DisallowReads: true,
|
|
|
|
DisallowLists: true,
|
|
|
|
},
|
|
|
|
CreateSegmentAllowed: true,
|
|
|
|
CommitSegmentAllowed: true,
|
|
|
|
DeleteSegmentAllowed: true,
|
|
|
|
},
|
|
|
|
|
|
|
|
{ // Bucket restriction
|
|
|
|
Caveat: macaroon.Caveat{
|
|
|
|
AllowedPaths: []*macaroon.Caveat_Path{{
|
|
|
|
Bucket: []byte("otherbucket"),
|
|
|
|
}},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
|
|
|
|
{ // Path restriction
|
|
|
|
Caveat: macaroon.Caveat{
|
|
|
|
AllowedPaths: []*macaroon.Caveat_Path{{
|
|
|
|
Bucket: []byte("testbucket"),
|
|
|
|
EncryptedPathPrefix: []byte("otherpath"),
|
|
|
|
}},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
|
|
|
|
{ // Time restriction after
|
|
|
|
Caveat: macaroon.Caveat{
|
|
|
|
NotAfter: func(x time.Time) *time.Time { return &x }(time.Now()),
|
|
|
|
},
|
|
|
|
},
|
|
|
|
|
|
|
|
{ // Time restriction before
|
|
|
|
Caveat: macaroon.Caveat{
|
|
|
|
NotBefore: func(x time.Time) *time.Time { return &x }(time.Now().Add(time.Hour)),
|
|
|
|
},
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, test := range tests {
|
|
|
|
restrictedKey, err := key.Restrict(test.Caveat)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
client, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], restrictedKey.Serialize())
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
_, _, err = client.CreateSegment(ctx, "testbucket", "testpath", 1, &pb.RedundancyScheme{}, 123, time.Now())
|
|
|
|
assertUnauthenticated(t, err, test.CreateSegmentAllowed)
|
|
|
|
|
|
|
|
_, err = client.CommitSegment(ctx, "testbucket", "testpath", 0, &pb.Pointer{}, nil)
|
|
|
|
assertUnauthenticated(t, err, test.CommitSegmentAllowed)
|
|
|
|
|
|
|
|
_, err = client.SegmentInfo(ctx, "testbucket", "testpath", 0)
|
|
|
|
assertUnauthenticated(t, err, test.SegmentInfoAllowed)
|
|
|
|
|
|
|
|
_, _, err = client.ReadSegment(ctx, "testbucket", "testpath", 0)
|
|
|
|
assertUnauthenticated(t, err, test.ReadSegmentAllowed)
|
|
|
|
|
|
|
|
_, err = client.DeleteSegment(ctx, "testbucket", "testpath", 0)
|
|
|
|
assertUnauthenticated(t, err, test.DeleteSegmentAllowed)
|
|
|
|
|
|
|
|
_, _, err = client.ListSegments(ctx, "testbucket", "testpath", "", "", true, 1, 0)
|
|
|
|
assertUnauthenticated(t, err, test.ListSegmentsAllowed)
|
|
|
|
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func assertUnauthenticated(t *testing.T, err error, allowed bool) {
|
2019-03-30 11:21:49 +00:00
|
|
|
t.Helper()
|
|
|
|
|
2019-05-24 17:51:27 +01:00
|
|
|
// If it's allowed, we allow any non-unauthenticated error because
|
|
|
|
// some calls error after authentication checks.
|
2019-03-30 11:21:49 +00:00
|
|
|
if err, ok := status.FromError(errs.Unwrap(err)); ok {
|
2019-05-24 17:51:27 +01:00
|
|
|
assert.Equal(t, codes.Unauthenticated == err.Code(), !allowed)
|
|
|
|
} else if !allowed {
|
2019-03-30 11:21:49 +00:00
|
|
|
assert.Fail(t, "got unexpected error", "%T", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestServiceList(t *testing.T) {
|
|
|
|
ctx := testcontext.New(t)
|
|
|
|
defer ctx.Cleanup()
|
|
|
|
|
|
|
|
planet, err := testplanet.New(t, 1, 6, 1)
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer ctx.Check(planet.Shutdown)
|
|
|
|
|
|
|
|
planet.Start(ctx)
|
|
|
|
|
|
|
|
items := []struct {
|
|
|
|
Key string
|
|
|
|
Value []byte
|
|
|
|
}{
|
|
|
|
{Key: "sample.😶", Value: []byte{1}},
|
|
|
|
{Key: "müsic", Value: []byte{2}},
|
|
|
|
{Key: "müsic/söng1.mp3", Value: []byte{3}},
|
|
|
|
{Key: "müsic/söng2.mp3", Value: []byte{4}},
|
|
|
|
{Key: "müsic/album/söng3.mp3", Value: []byte{5}},
|
|
|
|
{Key: "müsic/söng4.mp3", Value: []byte{6}},
|
|
|
|
{Key: "ビデオ/movie.mkv", Value: []byte{7}},
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, item := range items {
|
|
|
|
err := planet.Uplinks[0].Upload(ctx, planet.Satellites[0], "testbucket", item.Key, item.Value)
|
|
|
|
assert.NoError(t, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
config := planet.Uplinks[0].GetConfig(planet.Satellites[0])
|
|
|
|
metainfo, _, err := config.GetMetainfo(ctx, planet.Uplinks[0].Identity)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
type Test struct {
|
|
|
|
Request storj.ListOptions
|
|
|
|
Expected storj.ObjectList // objects are partial
|
|
|
|
}
|
|
|
|
|
|
|
|
list, err := metainfo.ListObjects(ctx, "testbucket", storj.ListOptions{Recursive: true, Direction: storj.After})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
expected := []storj.Object{
|
|
|
|
{Path: "müsic"},
|
|
|
|
{Path: "müsic/album/söng3.mp3"},
|
|
|
|
{Path: "müsic/söng1.mp3"},
|
|
|
|
{Path: "müsic/söng2.mp3"},
|
|
|
|
{Path: "müsic/söng4.mp3"},
|
|
|
|
{Path: "sample.😶"},
|
|
|
|
{Path: "ビデオ/movie.mkv"},
|
|
|
|
}
|
|
|
|
|
|
|
|
require.Equal(t, len(expected), len(list.Items))
|
|
|
|
sort.Slice(list.Items, func(i, k int) bool {
|
|
|
|
return list.Items[i].Path < list.Items[k].Path
|
|
|
|
})
|
|
|
|
for i, item := range expected {
|
|
|
|
require.Equal(t, item.Path, list.Items[i].Path)
|
|
|
|
require.Equal(t, item.IsPrefix, list.Items[i].IsPrefix)
|
|
|
|
}
|
|
|
|
|
|
|
|
list, err = metainfo.ListObjects(ctx, "testbucket", storj.ListOptions{Recursive: false, Direction: storj.After})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
expected = []storj.Object{
|
|
|
|
{Path: "müsic"},
|
|
|
|
{Path: "müsic/", IsPrefix: true},
|
|
|
|
{Path: "sample.😶"},
|
|
|
|
{Path: "ビデオ/", IsPrefix: true},
|
|
|
|
}
|
|
|
|
|
|
|
|
require.Equal(t, len(expected), len(list.Items))
|
|
|
|
sort.Slice(list.Items, func(i, k int) bool {
|
|
|
|
return list.Items[i].Path < list.Items[k].Path
|
|
|
|
})
|
|
|
|
for i, item := range expected {
|
2019-05-24 17:51:27 +01:00
|
|
|
t.Log(item.Path, list.Items[i].Path)
|
2019-03-30 11:21:49 +00:00
|
|
|
require.Equal(t, item.Path, list.Items[i].Path)
|
|
|
|
require.Equal(t, item.IsPrefix, list.Items[i].IsPrefix)
|
|
|
|
}
|
|
|
|
}
|
2019-04-09 14:31:19 +01:00
|
|
|
|
|
|
|
func TestCommitSegment(t *testing.T) {
|
2019-04-22 10:07:50 +01:00
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 6, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
2019-05-24 17:51:27 +01:00
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
2019-04-09 14:31:19 +01:00
|
|
|
|
2019-04-22 10:07:50 +01:00
|
|
|
metainfo, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
2019-04-09 14:31:19 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
2019-04-22 10:07:50 +01:00
|
|
|
{
|
|
|
|
// error if pointer is nil
|
|
|
|
_, err = metainfo.CommitSegment(ctx, "bucket", "path", -1, nil, []*pb.OrderLimit2{})
|
|
|
|
require.Error(t, err)
|
2019-04-09 14:31:19 +01:00
|
|
|
}
|
2019-04-22 10:07:50 +01:00
|
|
|
{
|
|
|
|
// error if bucket contains slash
|
|
|
|
_, err = metainfo.CommitSegment(ctx, "bucket/storj", "path", -1, &pb.Pointer{}, []*pb.OrderLimit2{})
|
|
|
|
require.Error(t, err)
|
2019-04-09 14:31:19 +01:00
|
|
|
}
|
2019-04-22 10:07:50 +01:00
|
|
|
{
|
|
|
|
// error if number of remote pieces is lower then repair threshold
|
|
|
|
redundancy := &pb.RedundancyScheme{
|
|
|
|
MinReq: 1,
|
|
|
|
RepairThreshold: 2,
|
|
|
|
SuccessThreshold: 4,
|
|
|
|
Total: 6,
|
|
|
|
ErasureShareSize: 10,
|
|
|
|
}
|
2019-06-05 17:41:02 +01:00
|
|
|
expirationDate := time.Now()
|
|
|
|
addresedLimits, rootPieceID, err := metainfo.CreateSegment(ctx, "bucket", "path", -1, redundancy, 1000, expirationDate)
|
2019-04-22 10:07:50 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// create number of pieces below repair threshold
|
|
|
|
usedForPieces := addresedLimits[:redundancy.RepairThreshold-1]
|
|
|
|
pieces := make([]*pb.RemotePiece, len(usedForPieces))
|
|
|
|
for i, limit := range usedForPieces {
|
|
|
|
pieces[i] = &pb.RemotePiece{
|
|
|
|
PieceNum: int32(i),
|
|
|
|
NodeId: limit.Limit.StorageNodeId,
|
|
|
|
}
|
|
|
|
}
|
2019-06-05 17:41:02 +01:00
|
|
|
|
|
|
|
expirationDateProto, err := ptypes.TimestampProto(expirationDate)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2019-04-22 10:07:50 +01:00
|
|
|
pointer := &pb.Pointer{
|
|
|
|
Type: pb.Pointer_REMOTE,
|
|
|
|
Remote: &pb.RemoteSegment{
|
|
|
|
RootPieceId: rootPieceID,
|
|
|
|
Redundancy: redundancy,
|
|
|
|
RemotePieces: pieces,
|
|
|
|
},
|
2019-06-05 17:41:02 +01:00
|
|
|
ExpirationDate: expirationDateProto,
|
2019-04-22 10:07:50 +01:00
|
|
|
}
|
2019-04-09 14:31:19 +01:00
|
|
|
|
2019-04-22 10:07:50 +01:00
|
|
|
limits := make([]*pb.OrderLimit2, len(addresedLimits))
|
|
|
|
for i, addresedLimit := range addresedLimits {
|
|
|
|
limits[i] = addresedLimit.Limit
|
|
|
|
}
|
|
|
|
_, err = metainfo.CommitSegment(ctx, "bucket", "path", -1, pointer, limits)
|
|
|
|
require.Error(t, err)
|
2019-05-17 20:02:40 +01:00
|
|
|
require.Contains(t, err.Error(), "Number of valid pieces is less than or equal to the repair threshold")
|
2019-04-09 14:31:19 +01:00
|
|
|
}
|
2019-04-22 10:07:50 +01:00
|
|
|
})
|
2019-04-09 14:31:19 +01:00
|
|
|
}
|
2019-06-05 17:41:02 +01:00
|
|
|
|
|
|
|
func TestDoubleCommitSegment(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 6, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
|
|
|
|
metainfo, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
pointer, limits := runCreateSegment(ctx, t, metainfo)
|
|
|
|
|
|
|
|
_, err = metainfo.CommitSegment(ctx, "myBucketName", "file/path", -1, pointer, limits)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
_, err = metainfo.CommitSegment(ctx, "myBucketName", "file/path", -1, pointer, limits)
|
|
|
|
require.Error(t, err)
|
|
|
|
require.Contains(t, err.Error(), "missing create request or request expired")
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestCommitSegmentPointer(t *testing.T) {
|
|
|
|
// all tests needs to generate error
|
|
|
|
tests := []struct {
|
|
|
|
// defines how modify pointer before CommitSegment
|
|
|
|
Modify func(pointer *pb.Pointer)
|
|
|
|
ErrorMessage string
|
|
|
|
}{
|
|
|
|
{
|
|
|
|
Modify: func(pointer *pb.Pointer) {
|
|
|
|
pointer.ExpirationDate.Seconds += 100
|
|
|
|
},
|
|
|
|
ErrorMessage: "pointer expiration date does not match requested one",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Modify: func(pointer *pb.Pointer) {
|
|
|
|
pointer.Remote.Redundancy.MinReq += 100
|
|
|
|
},
|
|
|
|
ErrorMessage: "pointer redundancy scheme date does not match requested one",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Modify: func(pointer *pb.Pointer) {
|
|
|
|
pointer.Remote.Redundancy.RepairThreshold += 100
|
|
|
|
},
|
|
|
|
ErrorMessage: "pointer redundancy scheme date does not match requested one",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Modify: func(pointer *pb.Pointer) {
|
|
|
|
pointer.Remote.Redundancy.SuccessThreshold += 100
|
|
|
|
},
|
|
|
|
ErrorMessage: "pointer redundancy scheme date does not match requested one",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Modify: func(pointer *pb.Pointer) {
|
|
|
|
pointer.Remote.Redundancy.Total += 100
|
|
|
|
},
|
|
|
|
// this error is triggered earlier then Create/Commit RS comparison
|
|
|
|
ErrorMessage: "invalid no order limit for piece",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Modify: func(pointer *pb.Pointer) {
|
|
|
|
pointer.Remote.Redundancy.ErasureShareSize += 100
|
|
|
|
},
|
|
|
|
ErrorMessage: "pointer redundancy scheme date does not match requested one",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Modify: func(pointer *pb.Pointer) {
|
|
|
|
pointer.Remote.Redundancy.Type = 100
|
|
|
|
},
|
|
|
|
ErrorMessage: "pointer redundancy scheme date does not match requested one",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Modify: func(pointer *pb.Pointer) {
|
|
|
|
pointer.Type = pb.Pointer_INLINE
|
|
|
|
},
|
|
|
|
ErrorMessage: "pointer type is INLINE but remote segment is set",
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 6, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
|
|
|
|
metainfo, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
for _, test := range tests {
|
|
|
|
pointer, limits := runCreateSegment(ctx, t, metainfo)
|
|
|
|
test.Modify(pointer)
|
|
|
|
|
|
|
|
_, err = metainfo.CommitSegment(ctx, "myBucketName", "file/path", -1, pointer, limits)
|
|
|
|
require.Error(t, err)
|
|
|
|
require.Contains(t, err.Error(), test.ErrorMessage)
|
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2019-06-13 02:35:37 +01:00
|
|
|
func TestValueAttributeInfo(t *testing.T) {
|
|
|
|
testplanet.Run(t, testplanet.Config{
|
|
|
|
SatelliteCount: 1, StorageNodeCount: 6, UplinkCount: 1,
|
|
|
|
}, func(t *testing.T, ctx *testcontext.Context, planet *testplanet.Planet) {
|
|
|
|
apiKey := planet.Uplinks[0].APIKey[planet.Satellites[0].ID()]
|
|
|
|
uplink := planet.Uplinks[0]
|
|
|
|
config := uplink.GetConfig(planet.Satellites[0])
|
|
|
|
metainfo, _, err := config.GetMetainfo(ctx, uplink.Identity)
|
|
|
|
require.NoError(t, err)
|
|
|
|
encScheme := config.GetEncryptionScheme()
|
|
|
|
_, err = metainfo.CreateBucket(ctx, "myBucket", &storj.Bucket{PathCipher: encScheme.Cipher})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
metainfoClient, err := planet.Uplinks[0].DialMetainfo(ctx, planet.Satellites[0], apiKey)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
keyInfo := pb.ValueAttributionRequest{
|
|
|
|
PartnerId: []byte("PartnerID"),
|
|
|
|
BucketName: []byte("myBucketName"),
|
|
|
|
}
|
|
|
|
|
|
|
|
{
|
|
|
|
// bucket with no items
|
|
|
|
err = metainfoClient.ValueAttributeInfo(ctx, "myBucket", "", -1, string(keyInfo.PartnerId))
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// no bucket exists
|
|
|
|
err = metainfoClient.ValueAttributeInfo(ctx, "myBucket1", "", -1, string(keyInfo.PartnerId))
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
{
|
|
|
|
expectedData := make([]byte, 1*memory.MiB)
|
|
|
|
_, err = rand.Read(expectedData)
|
|
|
|
assert.NoError(t, err)
|
|
|
|
|
|
|
|
err = planet.Uplinks[0].Upload(ctx, planet.Satellites[0], "myBucket", "path", expectedData)
|
|
|
|
assert.NoError(t, err)
|
|
|
|
|
|
|
|
// bucket with items
|
|
|
|
err = metainfoClient.ValueAttributeInfo(ctx, "myBucket", "", -1, string(keyInfo.PartnerId))
|
|
|
|
require.Error(t, err)
|
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2019-06-05 17:41:02 +01:00
|
|
|
func runCreateSegment(ctx context.Context, t *testing.T, metainfo metainfo.Client) (*pb.Pointer, []*pb.OrderLimit2) {
|
|
|
|
pointer := createTestPointer(t)
|
|
|
|
expirationDate, err := ptypes.Timestamp(pointer.ExpirationDate)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
addressedLimits, rootPieceID, err := metainfo.CreateSegment(ctx, "myBucketName", "file/path", -1, pointer.Remote.Redundancy, memory.MiB.Int64(), expirationDate)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
pointer.Remote.RootPieceId = rootPieceID
|
|
|
|
pointer.Remote.RemotePieces[0].NodeId = addressedLimits[0].Limit.StorageNodeId
|
|
|
|
pointer.Remote.RemotePieces[1].NodeId = addressedLimits[1].Limit.StorageNodeId
|
|
|
|
|
|
|
|
limits := make([]*pb.OrderLimit2, len(addressedLimits))
|
|
|
|
for i, addressedLimit := range addressedLimits {
|
|
|
|
limits[i] = addressedLimit.Limit
|
|
|
|
}
|
|
|
|
|
|
|
|
return pointer, limits
|
|
|
|
}
|
|
|
|
|
|
|
|
func createTestPointer(t *testing.T) *pb.Pointer {
|
|
|
|
rs := &pb.RedundancyScheme{
|
|
|
|
MinReq: 1,
|
|
|
|
RepairThreshold: 1,
|
|
|
|
SuccessThreshold: 3,
|
|
|
|
Total: 4,
|
|
|
|
ErasureShareSize: 1024,
|
|
|
|
Type: pb.RedundancyScheme_RS,
|
|
|
|
}
|
|
|
|
|
|
|
|
pointer := &pb.Pointer{
|
|
|
|
Type: pb.Pointer_REMOTE,
|
|
|
|
Remote: &pb.RemoteSegment{
|
|
|
|
Redundancy: rs,
|
|
|
|
RemotePieces: []*pb.RemotePiece{
|
|
|
|
&pb.RemotePiece{
|
|
|
|
PieceNum: 0,
|
|
|
|
},
|
|
|
|
&pb.RemotePiece{
|
|
|
|
PieceNum: 1,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
ExpirationDate: ptypes.TimestampNow(),
|
|
|
|
}
|
|
|
|
return pointer
|
|
|
|
}
|