satellite/metainfo: add buckets RPC and database (#2460)
* add db interface and methods, add sa metainfo endpoints and svc * add bucket metainfo svc funcs * add sadb bucekts * bucket list gets all buckets * filter buckets list on macaroon restrictions * update pb cipher suite to be enum * add conversion funcs * updates per comments * bucket settings should say default * add direction to list buckets, add tests * fix test bucket names * lint err * only support forward direction * add comments * minor refactoring * make sure list up to limit * update test * update protolock file * fix lint * change per PR
This commit is contained in:
parent
16156e3b3d
commit
f9696d6c5e
@ -103,10 +103,10 @@ func (p *Project) CreateBucket(ctx context.Context, name string, cfg *BucketConf
|
|||||||
cfg.setDefaults()
|
cfg.setDefaults()
|
||||||
|
|
||||||
bucket = storj.Bucket{
|
bucket = storj.Bucket{
|
||||||
PathCipher: cfg.PathCipher,
|
PathCipher: cfg.PathCipher,
|
||||||
EncryptionParameters: cfg.EncryptionParameters,
|
DefaultEncryptionParameters: cfg.EncryptionParameters,
|
||||||
RedundancyScheme: cfg.Volatile.RedundancyScheme,
|
DefaultRedundancyScheme: cfg.Volatile.RedundancyScheme,
|
||||||
SegmentsSize: cfg.Volatile.SegmentsSize.Int64(),
|
DefaultSegmentsSize: cfg.Volatile.SegmentsSize.Int64(),
|
||||||
}
|
}
|
||||||
return p.project.CreateBucket(ctx, name, &bucket)
|
return p.project.CreateBucket(ctx, name, &bucket)
|
||||||
}
|
}
|
||||||
@ -139,10 +139,10 @@ func (p *Project) GetBucketInfo(ctx context.Context, bucket string) (b storj.Buc
|
|||||||
}
|
}
|
||||||
cfg := &BucketConfig{
|
cfg := &BucketConfig{
|
||||||
PathCipher: b.PathCipher,
|
PathCipher: b.PathCipher,
|
||||||
EncryptionParameters: b.EncryptionParameters,
|
EncryptionParameters: b.DefaultEncryptionParameters,
|
||||||
}
|
}
|
||||||
cfg.Volatile.RedundancyScheme = b.RedundancyScheme
|
cfg.Volatile.RedundancyScheme = b.DefaultRedundancyScheme
|
||||||
cfg.Volatile.SegmentsSize = memory.Size(b.SegmentsSize)
|
cfg.Volatile.SegmentsSize = memory.Size(b.DefaultSegmentsSize)
|
||||||
return b, cfg, nil
|
return b, cfg, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -258,11 +258,11 @@ func (p *Project) updateBucket(ctx context.Context, bucketInfo storj.Bucket) (bu
|
|||||||
defer mon.Task()(&ctx)(&err)
|
defer mon.Task()(&ctx)(&err)
|
||||||
|
|
||||||
bucket = storj.Bucket{
|
bucket = storj.Bucket{
|
||||||
Attribution: p.uplinkCfg.Volatile.PartnerID,
|
Attribution: p.uplinkCfg.Volatile.PartnerID,
|
||||||
PathCipher: bucketInfo.PathCipher,
|
PathCipher: bucketInfo.PathCipher,
|
||||||
EncryptionParameters: bucketInfo.EncryptionParameters,
|
DefaultEncryptionParameters: bucketInfo.DefaultEncryptionParameters,
|
||||||
RedundancyScheme: bucketInfo.RedundancyScheme,
|
DefaultRedundancyScheme: bucketInfo.DefaultRedundancyScheme,
|
||||||
SegmentsSize: bucketInfo.SegmentsSize,
|
DefaultSegmentsSize: bucketInfo.DefaultSegmentsSize,
|
||||||
}
|
}
|
||||||
return p.project.CreateBucket(ctx, bucketInfo.Name, &bucket)
|
return p.project.CreateBucket(ctx, bucketInfo.Name, &bucket)
|
||||||
}
|
}
|
||||||
|
@ -15,10 +15,10 @@ func newBucketInfo(bucket *storj.Bucket) C.BucketInfo {
|
|||||||
name: C.CString(bucket.Name),
|
name: C.CString(bucket.Name),
|
||||||
created: C.int64_t(bucket.Created.Unix()),
|
created: C.int64_t(bucket.Created.Unix()),
|
||||||
path_cipher: toCCipherSuite(bucket.PathCipher),
|
path_cipher: toCCipherSuite(bucket.PathCipher),
|
||||||
segment_size: C.uint64_t(bucket.SegmentsSize),
|
segment_size: C.uint64_t(bucket.DefaultSegmentsSize),
|
||||||
|
|
||||||
encryption_parameters: convertEncryptionParameters(&bucket.EncryptionParameters),
|
encryption_parameters: convertEncryptionParameters(&bucket.DefaultEncryptionParameters),
|
||||||
redundancy_scheme: convertRedundancyScheme(&bucket.RedundancyScheme),
|
redundancy_scheme: convertRedundancyScheme(&bucket.DefaultRedundancyScheme),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -57,18 +57,18 @@ func newBucketInfo(bucket storj.Bucket) *BucketInfo {
|
|||||||
Name: bucket.Name,
|
Name: bucket.Name,
|
||||||
Created: bucket.Created.UTC().UnixNano() / int64(time.Millisecond),
|
Created: bucket.Created.UTC().UnixNano() / int64(time.Millisecond),
|
||||||
PathCipher: byte(bucket.PathCipher),
|
PathCipher: byte(bucket.PathCipher),
|
||||||
SegmentsSize: bucket.SegmentsSize,
|
SegmentsSize: bucket.DefaultSegmentsSize,
|
||||||
RedundancyScheme: &RedundancyScheme{
|
RedundancyScheme: &RedundancyScheme{
|
||||||
Algorithm: byte(bucket.RedundancyScheme.Algorithm),
|
Algorithm: byte(bucket.DefaultRedundancyScheme.Algorithm),
|
||||||
ShareSize: bucket.RedundancyScheme.ShareSize,
|
ShareSize: bucket.DefaultRedundancyScheme.ShareSize,
|
||||||
RequiredShares: bucket.RedundancyScheme.RequiredShares,
|
RequiredShares: bucket.DefaultRedundancyScheme.RequiredShares,
|
||||||
RepairShares: bucket.RedundancyScheme.RepairShares,
|
RepairShares: bucket.DefaultRedundancyScheme.RepairShares,
|
||||||
OptimalShares: bucket.RedundancyScheme.OptimalShares,
|
OptimalShares: bucket.DefaultRedundancyScheme.OptimalShares,
|
||||||
TotalShares: bucket.RedundancyScheme.TotalShares,
|
TotalShares: bucket.DefaultRedundancyScheme.TotalShares,
|
||||||
},
|
},
|
||||||
EncryptionParameters: &EncryptionParameters{
|
EncryptionParameters: &EncryptionParameters{
|
||||||
CipherSuite: byte(bucket.EncryptionParameters.CipherSuite),
|
CipherSuite: byte(bucket.DefaultEncryptionParameters.CipherSuite),
|
||||||
BlockSize: bucket.EncryptionParameters.BlockSize,
|
BlockSize: bucket.DefaultEncryptionParameters.BlockSize,
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -131,6 +131,27 @@ func (a *APIKey) Check(ctx context.Context, secret []byte, action Action, revoke
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// GetAllowedBuckets returns a list of all the allowed bucket paths that match the Action operation
|
||||||
|
func (a *APIKey) GetAllowedBuckets(ctx context.Context, action Action) (allowedBuckets map[string]struct{}, err error) {
|
||||||
|
defer mon.Task()(&ctx)(&err)
|
||||||
|
|
||||||
|
caveats := a.mac.Caveats()
|
||||||
|
for _, cavbuf := range caveats {
|
||||||
|
var cav Caveat
|
||||||
|
err := proto.Unmarshal(cavbuf, &cav)
|
||||||
|
if err != nil {
|
||||||
|
return allowedBuckets, ErrFormat.New("invalid caveat format: %v", err)
|
||||||
|
}
|
||||||
|
if cav.Allows(action) {
|
||||||
|
for _, caveatPath := range cav.AllowedPaths {
|
||||||
|
allowedBuckets[string(caveatPath.Bucket)] = struct{}{}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return allowedBuckets, err
|
||||||
|
}
|
||||||
|
|
||||||
// Restrict generates a new APIKey with the provided Caveat attached.
|
// Restrict generates a new APIKey with the provided Caveat attached.
|
||||||
func (a *APIKey) Restrict(caveat Caveat) (*APIKey, error) {
|
func (a *APIKey) Restrict(caveat Caveat) (*APIKey, error) {
|
||||||
buf, err := proto.Marshal(&caveat)
|
buf, err := proto.Marshal(&caveat)
|
||||||
@ -182,6 +203,12 @@ func (c *Caveat) Allows(action Action) bool {
|
|||||||
if len(c.AllowedPaths) == 0 {
|
if len(c.AllowedPaths) == 0 {
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
if len(action.Bucket) == 0 {
|
||||||
|
// if no action.bucket name is provided, then this call is checking that
|
||||||
|
// we can list all buckets. In that case, return true here and we will
|
||||||
|
// filter out buckets that aren't allowed later with `GetAllowedBuckets()`
|
||||||
|
return true
|
||||||
|
}
|
||||||
for _, path := range c.AllowedPaths {
|
for _, path := range c.AllowedPaths {
|
||||||
if bytes.Equal(path.Bucket, action.Bucket) {
|
if bytes.Equal(path.Bucket, action.Bucket) {
|
||||||
return true
|
return true
|
||||||
|
@ -29,35 +29,35 @@ func (db *Project) CreateBucket(ctx context.Context, bucketName string, info *st
|
|||||||
if info == nil {
|
if info == nil {
|
||||||
info = &storj.Bucket{PathCipher: storj.EncAESGCM}
|
info = &storj.Bucket{PathCipher: storj.EncAESGCM}
|
||||||
}
|
}
|
||||||
if info.EncryptionParameters.CipherSuite == storj.EncUnspecified {
|
if info.DefaultEncryptionParameters.CipherSuite == storj.EncUnspecified {
|
||||||
info.EncryptionParameters.CipherSuite = storj.EncAESGCM
|
info.DefaultEncryptionParameters.CipherSuite = storj.EncAESGCM
|
||||||
}
|
}
|
||||||
if info.EncryptionParameters.BlockSize == 0 {
|
if info.DefaultEncryptionParameters.BlockSize == 0 {
|
||||||
info.EncryptionParameters.BlockSize = db.encryptedBlockSize
|
info.DefaultEncryptionParameters.BlockSize = db.encryptedBlockSize
|
||||||
}
|
}
|
||||||
if info.RedundancyScheme.Algorithm == storj.InvalidRedundancyAlgorithm {
|
if info.DefaultRedundancyScheme.Algorithm == storj.InvalidRedundancyAlgorithm {
|
||||||
info.RedundancyScheme.Algorithm = storj.ReedSolomon
|
info.DefaultRedundancyScheme.Algorithm = storj.ReedSolomon
|
||||||
}
|
}
|
||||||
if info.RedundancyScheme.RequiredShares == 0 {
|
if info.DefaultRedundancyScheme.RequiredShares == 0 {
|
||||||
info.RedundancyScheme.RequiredShares = int16(db.redundancy.RequiredCount())
|
info.DefaultRedundancyScheme.RequiredShares = int16(db.redundancy.RequiredCount())
|
||||||
}
|
}
|
||||||
if info.RedundancyScheme.RepairShares == 0 {
|
if info.DefaultRedundancyScheme.RepairShares == 0 {
|
||||||
info.RedundancyScheme.RepairShares = int16(db.redundancy.RepairThreshold())
|
info.DefaultRedundancyScheme.RepairShares = int16(db.redundancy.RepairThreshold())
|
||||||
}
|
}
|
||||||
if info.RedundancyScheme.OptimalShares == 0 {
|
if info.DefaultRedundancyScheme.OptimalShares == 0 {
|
||||||
info.RedundancyScheme.OptimalShares = int16(db.redundancy.OptimalThreshold())
|
info.DefaultRedundancyScheme.OptimalShares = int16(db.redundancy.OptimalThreshold())
|
||||||
}
|
}
|
||||||
if info.RedundancyScheme.TotalShares == 0 {
|
if info.DefaultRedundancyScheme.TotalShares == 0 {
|
||||||
info.RedundancyScheme.TotalShares = int16(db.redundancy.TotalCount())
|
info.DefaultRedundancyScheme.TotalShares = int16(db.redundancy.TotalCount())
|
||||||
}
|
}
|
||||||
if info.RedundancyScheme.ShareSize == 0 {
|
if info.DefaultRedundancyScheme.ShareSize == 0 {
|
||||||
info.RedundancyScheme.ShareSize = int32(db.redundancy.ErasureShareSize())
|
info.DefaultRedundancyScheme.ShareSize = int32(db.redundancy.ErasureShareSize())
|
||||||
}
|
}
|
||||||
if info.SegmentsSize == 0 {
|
if info.DefaultSegmentsSize == 0 {
|
||||||
info.SegmentsSize = db.segmentsSize
|
info.DefaultSegmentsSize = db.segmentsSize
|
||||||
}
|
}
|
||||||
|
|
||||||
if err := validateBlockSize(info.RedundancyScheme, info.EncryptionParameters.BlockSize); err != nil {
|
if err := validateBlockSize(info.DefaultRedundancyScheme, info.DefaultEncryptionParameters.BlockSize); err != nil {
|
||||||
return bucketInfo, err
|
return bucketInfo, err
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -69,15 +69,15 @@ func (db *Project) CreateBucket(ctx context.Context, bucketName string, info *st
|
|||||||
userMeta := map[string]string{
|
userMeta := map[string]string{
|
||||||
"attribution-to": info.Attribution,
|
"attribution-to": info.Attribution,
|
||||||
"path-enc-type": strconv.Itoa(int(info.PathCipher)),
|
"path-enc-type": strconv.Itoa(int(info.PathCipher)),
|
||||||
"default-seg-size": strconv.FormatInt(info.SegmentsSize, 10),
|
"default-seg-size": strconv.FormatInt(info.DefaultSegmentsSize, 10),
|
||||||
"default-enc-type": strconv.Itoa(int(info.EncryptionParameters.CipherSuite)),
|
"default-enc-type": strconv.Itoa(int(info.DefaultEncryptionParameters.CipherSuite)),
|
||||||
"default-enc-blksz": strconv.FormatInt(int64(info.EncryptionParameters.BlockSize), 10),
|
"default-enc-blksz": strconv.FormatInt(int64(info.DefaultEncryptionParameters.BlockSize), 10),
|
||||||
"default-rs-algo": strconv.Itoa(int(info.RedundancyScheme.Algorithm)),
|
"default-rs-algo": strconv.Itoa(int(info.DefaultRedundancyScheme.Algorithm)),
|
||||||
"default-rs-sharsz": strconv.FormatInt(int64(info.RedundancyScheme.ShareSize), 10),
|
"default-rs-sharsz": strconv.FormatInt(int64(info.DefaultRedundancyScheme.ShareSize), 10),
|
||||||
"default-rs-reqd": strconv.Itoa(int(info.RedundancyScheme.RequiredShares)),
|
"default-rs-reqd": strconv.Itoa(int(info.DefaultRedundancyScheme.RequiredShares)),
|
||||||
"default-rs-repair": strconv.Itoa(int(info.RedundancyScheme.RepairShares)),
|
"default-rs-repair": strconv.Itoa(int(info.DefaultRedundancyScheme.RepairShares)),
|
||||||
"default-rs-optim": strconv.Itoa(int(info.RedundancyScheme.OptimalShares)),
|
"default-rs-optim": strconv.Itoa(int(info.DefaultRedundancyScheme.OptimalShares)),
|
||||||
"default-rs-total": strconv.Itoa(int(info.RedundancyScheme.TotalShares)),
|
"default-rs-total": strconv.Itoa(int(info.DefaultRedundancyScheme.TotalShares)),
|
||||||
}
|
}
|
||||||
var exp time.Time
|
var exp time.Time
|
||||||
m, err := db.buckets.Put(ctx, bucketName, r, pb.SerializableMeta{UserDefined: userMeta}, exp)
|
m, err := db.buckets.Put(ctx, bucketName, r, pb.SerializableMeta{UserDefined: userMeta}, exp)
|
||||||
@ -215,12 +215,12 @@ func bucketFromMeta(ctx context.Context, bucketName string, m objects.Meta) (out
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
es := &out.EncryptionParameters
|
es := &out.DefaultEncryptionParameters
|
||||||
rs := &out.RedundancyScheme
|
rs := &out.DefaultRedundancyScheme
|
||||||
|
|
||||||
out.Attribution = m.UserDefined["attribution-to"]
|
out.Attribution = m.UserDefined["attribution-to"]
|
||||||
applySetting("path-enc-type", 16, func(v int64) { out.PathCipher = storj.CipherSuite(v) })
|
applySetting("path-enc-type", 16, func(v int64) { out.PathCipher = storj.CipherSuite(v) })
|
||||||
applySetting("default-seg-size", 64, func(v int64) { out.SegmentsSize = v })
|
applySetting("default-seg-size", 64, func(v int64) { out.DefaultSegmentsSize = v })
|
||||||
applySetting("default-enc-type", 32, func(v int64) { es.CipherSuite = storj.CipherSuite(v) })
|
applySetting("default-enc-type", 32, func(v int64) { es.CipherSuite = storj.CipherSuite(v) })
|
||||||
applySetting("default-enc-blksz", 32, func(v int64) { es.BlockSize = int32(v) })
|
applySetting("default-enc-blksz", 32, func(v int64) { es.BlockSize = int32(v) })
|
||||||
applySetting("default-rs-algo", 32, func(v int64) { rs.Algorithm = storj.RedundancyAlgorithm(v) })
|
applySetting("default-rs-algo", 32, func(v int64) { rs.Algorithm = storj.RedundancyAlgorithm(v) })
|
||||||
|
@ -20,43 +20,43 @@ var _ = math.Inf
|
|||||||
// proto package needs to be updated.
|
// proto package needs to be updated.
|
||||||
const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package
|
const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package
|
||||||
|
|
||||||
type CipherSuite_Type int32
|
type CipherSuite int32
|
||||||
|
|
||||||
const (
|
const (
|
||||||
CipherSuite_ENC_UNSPECIFIED CipherSuite_Type = 0
|
CipherSuite_ENC_UNSPECIFIED CipherSuite = 0
|
||||||
CipherSuite_ENC_NULL CipherSuite_Type = 1
|
CipherSuite_ENC_NULL CipherSuite = 1
|
||||||
CipherSuite_ENC_AESGCM CipherSuite_Type = 2
|
CipherSuite_ENC_AESGCM CipherSuite = 2
|
||||||
CipherSuite_ENC_SECRETBOX CipherSuite_Type = 3
|
CipherSuite_ENC_SECRETBOX CipherSuite = 3
|
||||||
)
|
)
|
||||||
|
|
||||||
var CipherSuite_Type_name = map[int32]string{
|
var CipherSuite_name = map[int32]string{
|
||||||
0: "ENC_UNSPECIFIED",
|
0: "ENC_UNSPECIFIED",
|
||||||
1: "ENC_NULL",
|
1: "ENC_NULL",
|
||||||
2: "ENC_AESGCM",
|
2: "ENC_AESGCM",
|
||||||
3: "ENC_SECRETBOX",
|
3: "ENC_SECRETBOX",
|
||||||
}
|
}
|
||||||
|
|
||||||
var CipherSuite_Type_value = map[string]int32{
|
var CipherSuite_value = map[string]int32{
|
||||||
"ENC_UNSPECIFIED": 0,
|
"ENC_UNSPECIFIED": 0,
|
||||||
"ENC_NULL": 1,
|
"ENC_NULL": 1,
|
||||||
"ENC_AESGCM": 2,
|
"ENC_AESGCM": 2,
|
||||||
"ENC_SECRETBOX": 3,
|
"ENC_SECRETBOX": 3,
|
||||||
}
|
}
|
||||||
|
|
||||||
func (x CipherSuite_Type) String() string {
|
func (x CipherSuite) String() string {
|
||||||
return proto.EnumName(CipherSuite_Type_name, int32(x))
|
return proto.EnumName(CipherSuite_name, int32(x))
|
||||||
}
|
}
|
||||||
|
|
||||||
func (CipherSuite_Type) EnumDescriptor() ([]byte, []int) {
|
func (CipherSuite) EnumDescriptor() ([]byte, []int) {
|
||||||
return fileDescriptor_8293a649ce9418c6, []int{1, 0}
|
return fileDescriptor_8293a649ce9418c6, []int{0}
|
||||||
}
|
}
|
||||||
|
|
||||||
type EncryptionParameters struct {
|
type EncryptionParameters struct {
|
||||||
CipherSuite *CipherSuite `protobuf:"bytes,1,opt,name=cipher_suite,json=cipherSuite,proto3" json:"cipher_suite,omitempty"`
|
CipherSuite CipherSuite `protobuf:"varint,1,opt,name=cipher_suite,json=cipherSuite,proto3,enum=encryption.CipherSuite" json:"cipher_suite,omitempty"`
|
||||||
BlockSize int64 `protobuf:"varint,2,opt,name=block_size,json=blockSize,proto3" json:"block_size,omitempty"`
|
BlockSize int64 `protobuf:"varint,2,opt,name=block_size,json=blockSize,proto3" json:"block_size,omitempty"`
|
||||||
XXX_NoUnkeyedLiteral struct{} `json:"-"`
|
XXX_NoUnkeyedLiteral struct{} `json:"-"`
|
||||||
XXX_unrecognized []byte `json:"-"`
|
XXX_unrecognized []byte `json:"-"`
|
||||||
XXX_sizecache int32 `json:"-"`
|
XXX_sizecache int32 `json:"-"`
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *EncryptionParameters) Reset() { *m = EncryptionParameters{} }
|
func (m *EncryptionParameters) Reset() { *m = EncryptionParameters{} }
|
||||||
@ -83,11 +83,11 @@ func (m *EncryptionParameters) XXX_DiscardUnknown() {
|
|||||||
|
|
||||||
var xxx_messageInfo_EncryptionParameters proto.InternalMessageInfo
|
var xxx_messageInfo_EncryptionParameters proto.InternalMessageInfo
|
||||||
|
|
||||||
func (m *EncryptionParameters) GetCipherSuite() *CipherSuite {
|
func (m *EncryptionParameters) GetCipherSuite() CipherSuite {
|
||||||
if m != nil {
|
if m != nil {
|
||||||
return m.CipherSuite
|
return m.CipherSuite
|
||||||
}
|
}
|
||||||
return nil
|
return CipherSuite_ENC_UNSPECIFIED
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *EncryptionParameters) GetBlockSize() int64 {
|
func (m *EncryptionParameters) GetBlockSize() int64 {
|
||||||
@ -97,67 +97,27 @@ func (m *EncryptionParameters) GetBlockSize() int64 {
|
|||||||
return 0
|
return 0
|
||||||
}
|
}
|
||||||
|
|
||||||
type CipherSuite struct {
|
|
||||||
Type CipherSuite_Type `protobuf:"varint,1,opt,name=type,proto3,enum=encryption.CipherSuite_Type" json:"type,omitempty"`
|
|
||||||
XXX_NoUnkeyedLiteral struct{} `json:"-"`
|
|
||||||
XXX_unrecognized []byte `json:"-"`
|
|
||||||
XXX_sizecache int32 `json:"-"`
|
|
||||||
}
|
|
||||||
|
|
||||||
func (m *CipherSuite) Reset() { *m = CipherSuite{} }
|
|
||||||
func (m *CipherSuite) String() string { return proto.CompactTextString(m) }
|
|
||||||
func (*CipherSuite) ProtoMessage() {}
|
|
||||||
func (*CipherSuite) Descriptor() ([]byte, []int) {
|
|
||||||
return fileDescriptor_8293a649ce9418c6, []int{1}
|
|
||||||
}
|
|
||||||
func (m *CipherSuite) XXX_Unmarshal(b []byte) error {
|
|
||||||
return xxx_messageInfo_CipherSuite.Unmarshal(m, b)
|
|
||||||
}
|
|
||||||
func (m *CipherSuite) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
|
|
||||||
return xxx_messageInfo_CipherSuite.Marshal(b, m, deterministic)
|
|
||||||
}
|
|
||||||
func (m *CipherSuite) XXX_Merge(src proto.Message) {
|
|
||||||
xxx_messageInfo_CipherSuite.Merge(m, src)
|
|
||||||
}
|
|
||||||
func (m *CipherSuite) XXX_Size() int {
|
|
||||||
return xxx_messageInfo_CipherSuite.Size(m)
|
|
||||||
}
|
|
||||||
func (m *CipherSuite) XXX_DiscardUnknown() {
|
|
||||||
xxx_messageInfo_CipherSuite.DiscardUnknown(m)
|
|
||||||
}
|
|
||||||
|
|
||||||
var xxx_messageInfo_CipherSuite proto.InternalMessageInfo
|
|
||||||
|
|
||||||
func (m *CipherSuite) GetType() CipherSuite_Type {
|
|
||||||
if m != nil {
|
|
||||||
return m.Type
|
|
||||||
}
|
|
||||||
return CipherSuite_ENC_UNSPECIFIED
|
|
||||||
}
|
|
||||||
|
|
||||||
func init() {
|
func init() {
|
||||||
proto.RegisterEnum("encryption.CipherSuite_Type", CipherSuite_Type_name, CipherSuite_Type_value)
|
proto.RegisterEnum("encryption.CipherSuite", CipherSuite_name, CipherSuite_value)
|
||||||
proto.RegisterType((*EncryptionParameters)(nil), "encryption.EncryptionParameters")
|
proto.RegisterType((*EncryptionParameters)(nil), "encryption.EncryptionParameters")
|
||||||
proto.RegisterType((*CipherSuite)(nil), "encryption.CipherSuite")
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func init() { proto.RegisterFile("encryption.proto", fileDescriptor_8293a649ce9418c6) }
|
func init() { proto.RegisterFile("encryption.proto", fileDescriptor_8293a649ce9418c6) }
|
||||||
|
|
||||||
var fileDescriptor_8293a649ce9418c6 = []byte{
|
var fileDescriptor_8293a649ce9418c6 = []byte{
|
||||||
// 238 bytes of a gzipped FileDescriptorProto
|
// 209 bytes of a gzipped FileDescriptorProto
|
||||||
0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xe2, 0x12, 0x48, 0xcd, 0x4b, 0x2e,
|
0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xe2, 0x12, 0x48, 0xcd, 0x4b, 0x2e,
|
||||||
0xaa, 0x2c, 0x28, 0xc9, 0xcc, 0xcf, 0xd3, 0x2b, 0x28, 0xca, 0x2f, 0xc9, 0x17, 0xe2, 0x42, 0x88,
|
0xaa, 0x2c, 0x28, 0xc9, 0xcc, 0xcf, 0xd3, 0x2b, 0x28, 0xca, 0x2f, 0xc9, 0x17, 0xe2, 0x42, 0x88,
|
||||||
0x28, 0x15, 0x72, 0x89, 0xb8, 0xc2, 0x79, 0x01, 0x89, 0x45, 0x89, 0xb9, 0xa9, 0x25, 0xa9, 0x45,
|
0x28, 0x15, 0x72, 0x89, 0xb8, 0xc2, 0x79, 0x01, 0x89, 0x45, 0x89, 0xb9, 0xa9, 0x25, 0xa9, 0x45,
|
||||||
0xc5, 0x42, 0x56, 0x5c, 0x3c, 0xc9, 0x99, 0x05, 0x19, 0xa9, 0x45, 0xf1, 0xc5, 0xa5, 0x99, 0x25,
|
0xc5, 0x42, 0x56, 0x5c, 0x3c, 0xc9, 0x99, 0x05, 0x19, 0xa9, 0x45, 0xf1, 0xc5, 0xa5, 0x99, 0x25,
|
||||||
0xa9, 0x12, 0x8c, 0x0a, 0x8c, 0x1a, 0xdc, 0x46, 0xe2, 0x7a, 0x48, 0x86, 0x39, 0x83, 0xe5, 0x83,
|
0xa9, 0x12, 0x8c, 0x0a, 0x8c, 0x1a, 0x7c, 0x46, 0xe2, 0x7a, 0x48, 0x86, 0x39, 0x83, 0xe5, 0x83,
|
||||||
0x41, 0xd2, 0x41, 0xdc, 0xc9, 0x08, 0x8e, 0x90, 0x2c, 0x17, 0x57, 0x52, 0x4e, 0x7e, 0x72, 0x76,
|
0x41, 0xd2, 0x41, 0xdc, 0xc9, 0x08, 0x8e, 0x90, 0x2c, 0x17, 0x57, 0x52, 0x4e, 0x7e, 0x72, 0x76,
|
||||||
0x7c, 0x71, 0x66, 0x55, 0xaa, 0x04, 0x93, 0x02, 0xa3, 0x06, 0x73, 0x10, 0x27, 0x58, 0x24, 0x38,
|
0x7c, 0x71, 0x66, 0x55, 0xaa, 0x04, 0x93, 0x02, 0xa3, 0x06, 0x73, 0x10, 0x27, 0x58, 0x24, 0x38,
|
||||||
0xb3, 0x2a, 0x55, 0xa9, 0x97, 0x91, 0x8b, 0x1b, 0x49, 0xaf, 0x90, 0x01, 0x17, 0x4b, 0x49, 0x65,
|
0xb3, 0x2a, 0x55, 0x2b, 0x98, 0x8b, 0x1b, 0x49, 0xab, 0x90, 0x30, 0x17, 0xbf, 0xab, 0x9f, 0x73,
|
||||||
0x01, 0xc4, 0x0a, 0x3e, 0x23, 0x19, 0x1c, 0x56, 0xe8, 0x85, 0x54, 0x16, 0xa4, 0x06, 0x81, 0x55,
|
0x7c, 0xa8, 0x5f, 0x70, 0x80, 0xab, 0xb3, 0xa7, 0x9b, 0xa7, 0xab, 0x8b, 0x00, 0x83, 0x10, 0x0f,
|
||||||
0x2a, 0xf9, 0x70, 0xb1, 0x80, 0x78, 0x42, 0xc2, 0x5c, 0xfc, 0xae, 0x7e, 0xce, 0xf1, 0xa1, 0x7e,
|
0x17, 0x07, 0x48, 0xd0, 0x2f, 0xd4, 0xc7, 0x47, 0x80, 0x51, 0x88, 0x8f, 0x8b, 0x0b, 0xc4, 0x73,
|
||||||
0xc1, 0x01, 0xae, 0xce, 0x9e, 0x6e, 0x9e, 0xae, 0x2e, 0x02, 0x0c, 0x42, 0x3c, 0x5c, 0x1c, 0x20,
|
0x74, 0x0d, 0x76, 0x77, 0xf6, 0x15, 0x60, 0x12, 0x12, 0xe4, 0xe2, 0x05, 0xf1, 0x83, 0x5d, 0x9d,
|
||||||
0x41, 0xbf, 0x50, 0x1f, 0x1f, 0x01, 0x46, 0x21, 0x3e, 0x2e, 0x2e, 0x10, 0xcf, 0xd1, 0x35, 0xd8,
|
0x83, 0x5c, 0x43, 0x9c, 0xfc, 0x23, 0x04, 0x98, 0x9d, 0x58, 0xa2, 0x98, 0x0a, 0x92, 0x92, 0xd8,
|
||||||
0xdd, 0xd9, 0x57, 0x80, 0x49, 0x48, 0x90, 0x8b, 0x17, 0xc4, 0x0f, 0x76, 0x75, 0x0e, 0x72, 0x0d,
|
0xc0, 0x1e, 0x34, 0x06, 0x04, 0x00, 0x00, 0xff, 0xff, 0xc7, 0x9d, 0x88, 0x06, 0xf4, 0x00, 0x00,
|
||||||
0x71, 0xf2, 0x8f, 0x10, 0x60, 0x76, 0x62, 0x89, 0x62, 0x2a, 0x48, 0x4a, 0x62, 0x03, 0x87, 0x8d,
|
0x00,
|
||||||
0x31, 0x20, 0x00, 0x00, 0xff, 0xff, 0x2c, 0x45, 0xc3, 0x51, 0x2f, 0x01, 0x00, 0x00,
|
|
||||||
}
|
}
|
||||||
|
@ -11,13 +11,9 @@ message EncryptionParameters {
|
|||||||
int64 block_size = 2;
|
int64 block_size = 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
message CipherSuite {
|
enum CipherSuite {
|
||||||
enum Type {
|
ENC_UNSPECIFIED = 0;
|
||||||
ENC_UNSPECIFIED = 0;
|
ENC_NULL = 1;
|
||||||
ENC_NULL = 1;
|
ENC_AESGCM = 2;
|
||||||
ENC_AESGCM = 2;
|
ENC_SECRETBOX = 3;
|
||||||
ENC_SECRETBOX = 3;
|
|
||||||
}
|
|
||||||
|
|
||||||
Type type = 1;
|
|
||||||
}
|
}
|
||||||
|
@ -28,7 +28,7 @@ const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package
|
|||||||
|
|
||||||
type Bucket struct {
|
type Bucket struct {
|
||||||
Name []byte `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
|
Name []byte `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
|
||||||
PathCipher *CipherSuite `protobuf:"bytes,2,opt,name=path_cipher,json=pathCipher,proto3" json:"path_cipher,omitempty"`
|
PathCipher CipherSuite `protobuf:"varint,2,opt,name=path_cipher,json=pathCipher,proto3,enum=encryption.CipherSuite" json:"path_cipher,omitempty"`
|
||||||
AttributionId []byte `protobuf:"bytes,3,opt,name=attribution_id,json=attributionId,proto3" json:"attribution_id,omitempty"`
|
AttributionId []byte `protobuf:"bytes,3,opt,name=attribution_id,json=attributionId,proto3" json:"attribution_id,omitempty"`
|
||||||
CreatedAt time.Time `protobuf:"bytes,4,opt,name=created_at,json=createdAt,proto3,stdtime" json:"created_at"`
|
CreatedAt time.Time `protobuf:"bytes,4,opt,name=created_at,json=createdAt,proto3,stdtime" json:"created_at"`
|
||||||
DefaultSegmentSize int64 `protobuf:"varint,5,opt,name=default_segment_size,json=defaultSegmentSize,proto3" json:"default_segment_size,omitempty"`
|
DefaultSegmentSize int64 `protobuf:"varint,5,opt,name=default_segment_size,json=defaultSegmentSize,proto3" json:"default_segment_size,omitempty"`
|
||||||
@ -70,11 +70,11 @@ func (m *Bucket) GetName() []byte {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *Bucket) GetPathCipher() *CipherSuite {
|
func (m *Bucket) GetPathCipher() CipherSuite {
|
||||||
if m != nil {
|
if m != nil {
|
||||||
return m.PathCipher
|
return m.PathCipher
|
||||||
}
|
}
|
||||||
return nil
|
return CipherSuite_ENC_UNSPECIFIED
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *Bucket) GetAttributionId() []byte {
|
func (m *Bucket) GetAttributionId() []byte {
|
||||||
@ -160,7 +160,7 @@ func (m *BucketListItem) GetCreatedAt() time.Time {
|
|||||||
|
|
||||||
type BucketCreateRequest struct {
|
type BucketCreateRequest struct {
|
||||||
Name []byte `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
|
Name []byte `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
|
||||||
PathCipher *CipherSuite `protobuf:"bytes,2,opt,name=path_cipher,json=pathCipher,proto3" json:"path_cipher,omitempty"`
|
PathCipher CipherSuite `protobuf:"varint,2,opt,name=path_cipher,json=pathCipher,proto3,enum=encryption.CipherSuite" json:"path_cipher,omitempty"`
|
||||||
AttributionId []byte `protobuf:"bytes,3,opt,name=attribution_id,json=attributionId,proto3" json:"attribution_id,omitempty"`
|
AttributionId []byte `protobuf:"bytes,3,opt,name=attribution_id,json=attributionId,proto3" json:"attribution_id,omitempty"`
|
||||||
DefaultSegmentSize int64 `protobuf:"varint,4,opt,name=default_segment_size,json=defaultSegmentSize,proto3" json:"default_segment_size,omitempty"`
|
DefaultSegmentSize int64 `protobuf:"varint,4,opt,name=default_segment_size,json=defaultSegmentSize,proto3" json:"default_segment_size,omitempty"`
|
||||||
DefaultRedundancyScheme *RedundancyScheme `protobuf:"bytes,5,opt,name=default_redundancy_scheme,json=defaultRedundancyScheme,proto3" json:"default_redundancy_scheme,omitempty"`
|
DefaultRedundancyScheme *RedundancyScheme `protobuf:"bytes,5,opt,name=default_redundancy_scheme,json=defaultRedundancyScheme,proto3" json:"default_redundancy_scheme,omitempty"`
|
||||||
@ -201,11 +201,11 @@ func (m *BucketCreateRequest) GetName() []byte {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *BucketCreateRequest) GetPathCipher() *CipherSuite {
|
func (m *BucketCreateRequest) GetPathCipher() CipherSuite {
|
||||||
if m != nil {
|
if m != nil {
|
||||||
return m.PathCipher
|
return m.PathCipher
|
||||||
}
|
}
|
||||||
return nil
|
return CipherSuite_ENC_UNSPECIFIED
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *BucketCreateRequest) GetAttributionId() []byte {
|
func (m *BucketCreateRequest) GetAttributionId() []byte {
|
||||||
@ -1507,95 +1507,95 @@ func init() {
|
|||||||
func init() { proto.RegisterFile("metainfo.proto", fileDescriptor_631e2f30a93cd64e) }
|
func init() { proto.RegisterFile("metainfo.proto", fileDescriptor_631e2f30a93cd64e) }
|
||||||
|
|
||||||
var fileDescriptor_631e2f30a93cd64e = []byte{
|
var fileDescriptor_631e2f30a93cd64e = []byte{
|
||||||
// 1402 bytes of a gzipped FileDescriptorProto
|
// 1406 bytes of a gzipped FileDescriptorProto
|
||||||
0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xc4, 0x57, 0x4b, 0x6f, 0x1b, 0xd5,
|
0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xc4, 0x57, 0x4b, 0x6f, 0xdb, 0xc6,
|
||||||
0x17, 0xff, 0x8f, 0x93, 0x38, 0xf6, 0x71, 0x9a, 0xa4, 0xb7, 0xf9, 0x27, 0xee, 0xb8, 0x69, 0xdc,
|
0x16, 0xbe, 0x94, 0x6d, 0x59, 0x3a, 0x72, 0x6c, 0x67, 0xe2, 0x6b, 0x2b, 0x54, 0x1c, 0x2b, 0xcc,
|
||||||
0x29, 0xad, 0x82, 0x84, 0x5c, 0x94, 0x6e, 0x2a, 0x0a, 0x12, 0x79, 0xf4, 0x11, 0xd4, 0x47, 0x34,
|
0x4d, 0xa0, 0x0b, 0x5c, 0x28, 0x17, 0xce, 0x26, 0x68, 0x5a, 0xa0, 0x7e, 0xe4, 0xe1, 0x22, 0x0f,
|
||||||
0x46, 0xb4, 0x54, 0xa0, 0xd1, 0xd8, 0x73, 0xec, 0x0c, 0x78, 0x1e, 0xdc, 0xb9, 0x86, 0xb4, 0x6b,
|
0x83, 0x2a, 0x9a, 0x34, 0x68, 0x41, 0x50, 0xe2, 0x91, 0xc2, 0x56, 0x7c, 0x74, 0x38, 0x6a, 0x9d,
|
||||||
0x3e, 0x00, 0x8b, 0x2e, 0xf8, 0x1e, 0x7c, 0x09, 0xc4, 0x92, 0x25, 0x48, 0x65, 0xc7, 0x97, 0x60,
|
0xac, 0xfb, 0x03, 0xba, 0xc8, 0xa2, 0xff, 0xa3, 0x7f, 0xa2, 0xe8, 0xb2, 0xcb, 0x16, 0x48, 0x77,
|
||||||
0x83, 0xee, 0xcb, 0x33, 0x63, 0x8f, 0x9d, 0xb6, 0x4a, 0xc5, 0x6e, 0xee, 0xb9, 0xbf, 0x7b, 0xee,
|
0xfd, 0x13, 0xdd, 0x14, 0xf3, 0x12, 0x49, 0x89, 0x92, 0x93, 0xc0, 0x41, 0x77, 0x9c, 0x33, 0xdf,
|
||||||
0x39, 0xbf, 0xf3, 0xba, 0x03, 0xcb, 0x01, 0x32, 0xd7, 0x0f, 0x7b, 0x51, 0x2b, 0xa6, 0x11, 0x8b,
|
0x9c, 0x39, 0xe7, 0x3b, 0xaf, 0x21, 0xac, 0x06, 0xc8, 0x5c, 0x3f, 0xec, 0x47, 0xed, 0x98, 0x46,
|
||||||
0x48, 0x45, 0xaf, 0xcd, 0x55, 0x0c, 0xbb, 0xf4, 0x79, 0xcc, 0xfc, 0x28, 0x94, 0x7b, 0x26, 0xf4,
|
0x2c, 0x22, 0x15, 0xbd, 0x36, 0xd7, 0x31, 0xec, 0xd1, 0x17, 0x31, 0xf3, 0xa3, 0x50, 0xee, 0x99,
|
||||||
0xa3, 0xbe, 0xc2, 0x99, 0x5b, 0xfd, 0x28, 0xea, 0x0f, 0xf0, 0x86, 0x58, 0x75, 0x86, 0xbd, 0x1b,
|
0x30, 0x88, 0x06, 0x0a, 0x67, 0xee, 0x0c, 0xa2, 0x68, 0x30, 0xc4, 0x1b, 0x62, 0xd5, 0x1d, 0xf5,
|
||||||
0xcc, 0x0f, 0x30, 0x61, 0x6e, 0x10, 0x6b, 0x70, 0x18, 0x79, 0xa8, 0xbe, 0x57, 0xe2, 0xc8, 0x0f,
|
0x6f, 0x30, 0x3f, 0xc0, 0x84, 0xb9, 0x41, 0xac, 0xc1, 0x61, 0xe4, 0xa1, 0xfa, 0x5e, 0x8b, 0x23,
|
||||||
0x19, 0x52, 0xaf, 0xa3, 0x04, 0x4b, 0x11, 0xf5, 0x90, 0x26, 0x72, 0x65, 0xfd, 0x32, 0x07, 0xe5,
|
0x3f, 0x64, 0x48, 0xbd, 0xae, 0x12, 0xac, 0x44, 0xd4, 0x43, 0x9a, 0xc8, 0x95, 0xf5, 0xd3, 0x02,
|
||||||
0xbd, 0x61, 0xf7, 0x5b, 0x64, 0x84, 0xc0, 0x7c, 0xe8, 0x06, 0x58, 0x37, 0x9a, 0xc6, 0xf6, 0x92,
|
0x94, 0xf7, 0x47, 0xbd, 0xaf, 0x91, 0x11, 0x02, 0x8b, 0xa1, 0x1b, 0x60, 0xdd, 0x68, 0x1a, 0xad,
|
||||||
0x2d, 0xbe, 0xc9, 0x2d, 0xa8, 0xc5, 0x2e, 0x3b, 0x76, 0xba, 0x7e, 0x7c, 0x8c, 0xb4, 0x5e, 0x6a,
|
0x15, 0x5b, 0x7c, 0x93, 0x5b, 0x50, 0x8b, 0x5d, 0xf6, 0xdc, 0xe9, 0xf9, 0xf1, 0x73, 0xa4, 0xf5,
|
||||||
0x1a, 0xdb, 0xb5, 0x9d, 0x8d, 0x56, 0xc6, 0xbc, 0x7d, 0xb1, 0xd3, 0x1e, 0xfa, 0x0c, 0x6d, 0xe0,
|
0x52, 0xd3, 0x68, 0xad, 0xee, 0x6e, 0xb5, 0x33, 0xe6, 0x1d, 0x88, 0x9d, 0xce, 0xc8, 0x67, 0x68,
|
||||||
0x58, 0x29, 0x20, 0xd7, 0x60, 0xd9, 0x65, 0x8c, 0xfa, 0x9d, 0x21, 0x87, 0x39, 0xbe, 0x57, 0x9f,
|
0x03, 0xc7, 0x4a, 0x01, 0xb9, 0x06, 0xab, 0x2e, 0x63, 0xd4, 0xef, 0x8e, 0x38, 0xcc, 0xf1, 0xbd,
|
||||||
0x13, 0x7a, 0xcf, 0x65, 0xa4, 0x87, 0x1e, 0xd9, 0x07, 0xe8, 0x52, 0x74, 0x19, 0x7a, 0x8e, 0xcb,
|
0xfa, 0x82, 0xd0, 0x7b, 0x2e, 0x23, 0x3d, 0xf2, 0xc8, 0x01, 0x40, 0x8f, 0xa2, 0xcb, 0xd0, 0x73,
|
||||||
0xea, 0xf3, 0x42, 0xbf, 0xd9, 0x92, 0x0e, 0xb6, 0xb4, 0x83, 0xad, 0xcf, 0xb5, 0x83, 0x7b, 0x95,
|
0x5c, 0x56, 0x5f, 0x6c, 0x1a, 0xad, 0xda, 0xae, 0xd9, 0x96, 0x0e, 0xb6, 0xb5, 0x83, 0xed, 0x4f,
|
||||||
0x5f, 0x5f, 0x6d, 0xfd, 0xef, 0xa7, 0xbf, 0xb6, 0x0c, 0xbb, 0xaa, 0xce, 0xed, 0x32, 0xf2, 0x21,
|
0xb5, 0x83, 0xfb, 0x95, 0x9f, 0x5f, 0xef, 0xfc, 0xeb, 0x87, 0x3f, 0x76, 0x0c, 0xbb, 0xaa, 0xce,
|
||||||
0xac, 0x79, 0xd8, 0x73, 0x87, 0x03, 0xe6, 0x24, 0xd8, 0x0f, 0x30, 0x64, 0x4e, 0xe2, 0xbf, 0xc0,
|
0xed, 0x31, 0xf2, 0x7f, 0xd8, 0xf0, 0xb0, 0xef, 0x8e, 0x86, 0xcc, 0x49, 0x70, 0x10, 0x60, 0xc8,
|
||||||
0xfa, 0x42, 0xd3, 0xd8, 0x9e, 0xb3, 0x89, 0xda, 0x6b, 0xcb, 0xad, 0xb6, 0xff, 0x02, 0xc9, 0x13,
|
0x9c, 0xc4, 0x7f, 0x89, 0xf5, 0xa5, 0xa6, 0xd1, 0x5a, 0xb0, 0x89, 0xda, 0xeb, 0xc8, 0xad, 0x8e,
|
||||||
0xb8, 0xa8, 0x4f, 0x50, 0xf4, 0x86, 0xa1, 0xe7, 0x86, 0xdd, 0xe7, 0x4e, 0xd2, 0x3d, 0xc6, 0x00,
|
0xff, 0x12, 0xc9, 0x13, 0xb8, 0xa8, 0x4f, 0x50, 0xf4, 0x46, 0xa1, 0xe7, 0x86, 0xbd, 0x17, 0x4e,
|
||||||
0xeb, 0x65, 0x61, 0x45, 0xa3, 0x95, 0x32, 0x67, 0x8f, 0x30, 0x6d, 0x01, 0xb1, 0x37, 0xd4, 0xe9,
|
0xd2, 0x7b, 0x8e, 0x01, 0xd6, 0xcb, 0xc2, 0x8a, 0x46, 0x3b, 0x65, 0xce, 0x1e, 0x63, 0x3a, 0x02,
|
||||||
0xf1, 0x0d, 0xe2, 0xc1, 0xa6, 0x56, 0x9c, 0x92, 0xe4, 0xc4, 0x2e, 0x75, 0x03, 0x64, 0x48, 0x93,
|
0x62, 0x6f, 0xa9, 0xd3, 0x93, 0x1b, 0xc4, 0x83, 0x6d, 0xad, 0x38, 0x25, 0xc9, 0x89, 0x5d, 0xea,
|
||||||
0xfa, 0xa2, 0x50, 0xde, 0xcc, 0x52, 0x78, 0x67, 0xf4, 0x79, 0x34, 0xc2, 0xd9, 0x0d, 0xa5, 0xa6,
|
0x06, 0xc8, 0x90, 0x26, 0xf5, 0x65, 0xa1, 0xbc, 0x99, 0xa5, 0xf0, 0xce, 0xf8, 0xf3, 0x78, 0x8c,
|
||||||
0x68, 0xd3, 0xf2, 0x61, 0x59, 0x06, 0xed, 0x81, 0x9f, 0xb0, 0x43, 0x86, 0x41, 0x61, 0xf0, 0xf2,
|
0xb3, 0x1b, 0x4a, 0x4d, 0xd1, 0xa6, 0xe5, 0xc3, 0xaa, 0x0c, 0xda, 0x03, 0x3f, 0x61, 0x47, 0x0c,
|
||||||
0xdc, 0x96, 0xde, 0x8a, 0x5b, 0xeb, 0x9f, 0x12, 0x5c, 0x90, 0x77, 0xed, 0x0b, 0x99, 0x8d, 0xdf,
|
0x83, 0xc2, 0xe0, 0xe5, 0xb9, 0x2d, 0xbd, 0x13, 0xb7, 0xd6, 0x5f, 0x25, 0xb8, 0x20, 0xef, 0x3a,
|
||||||
0x0d, 0x31, 0xf9, 0x8f, 0xb2, 0x65, 0x5a, 0xa0, 0xe7, 0xdf, 0x2e, 0xd0, 0x0b, 0xef, 0x32, 0xd0,
|
0x10, 0x32, 0x1b, 0xbf, 0x19, 0x61, 0xf2, 0x0f, 0x65, 0xcb, 0xac, 0x40, 0x2f, 0xbe, 0x5b, 0xa0,
|
||||||
0xe5, 0xb3, 0x08, 0xf4, 0xa7, 0xb0, 0x96, 0x27, 0x3f, 0x89, 0xa3, 0x30, 0x41, 0xb2, 0x0d, 0xe5,
|
0x97, 0xde, 0x67, 0xa0, 0xcb, 0x67, 0x11, 0xe8, 0x8f, 0x61, 0x23, 0x4f, 0x7e, 0x12, 0x47, 0x61,
|
||||||
0x8e, 0x90, 0x0b, 0xfe, 0x6b, 0x3b, 0xab, 0xad, 0x51, 0x2f, 0x91, 0x78, 0x5b, 0xed, 0x5b, 0xd7,
|
0x82, 0xa4, 0x05, 0xe5, 0xae, 0x90, 0x0b, 0xfe, 0x6b, 0xbb, 0xeb, 0xed, 0x71, 0x2f, 0x91, 0x78,
|
||||||
0x61, 0x55, 0x4a, 0xee, 0x21, 0x9b, 0x11, 0x3b, 0xeb, 0x13, 0x38, 0x9f, 0xc1, 0xbd, 0xf1, 0x35,
|
0x5b, 0xed, 0x5b, 0xd7, 0x61, 0x5d, 0x4a, 0xee, 0x21, 0x9b, 0x13, 0x3b, 0xeb, 0x23, 0x38, 0x9f,
|
||||||
0xef, 0xeb, 0x2c, 0x39, 0xc0, 0x01, 0xce, 0xcc, 0x12, 0x6b, 0x5d, 0xfb, 0xa4, 0xa1, 0xf2, 0x32,
|
0xc1, 0xbd, 0xf5, 0x35, 0xff, 0xd5, 0x59, 0x72, 0x88, 0x43, 0x9c, 0x9b, 0x25, 0xd6, 0xa6, 0xf6,
|
||||||
0x6b, 0x57, 0x5b, 0xc0, 0x93, 0x5a, 0x2b, 0x58, 0x87, 0x72, 0x77, 0x48, 0x93, 0x88, 0x2a, 0x15,
|
0x49, 0x43, 0xe5, 0x65, 0xd6, 0x9e, 0xb6, 0x80, 0x27, 0xb5, 0x56, 0xb0, 0x09, 0xe5, 0xde, 0x88,
|
||||||
0x6a, 0x45, 0xd6, 0x60, 0x61, 0xe0, 0x07, 0xbe, 0x4c, 0xeb, 0x05, 0x5b, 0x2e, 0xac, 0xa7, 0x40,
|
0x26, 0x11, 0x55, 0x2a, 0xd4, 0x8a, 0x6c, 0xc0, 0xd2, 0xd0, 0x0f, 0x7c, 0x99, 0xd6, 0x4b, 0xb6,
|
||||||
0xb2, 0x2a, 0x94, 0x17, 0x2d, 0x58, 0xf0, 0x19, 0x06, 0x49, 0xdd, 0x68, 0xce, 0x6d, 0xd7, 0x76,
|
0x5c, 0x58, 0x4f, 0x81, 0x64, 0x55, 0x28, 0x2f, 0xda, 0xb0, 0xe4, 0x33, 0x0c, 0x92, 0xba, 0xd1,
|
||||||
0xea, 0xe3, 0x4e, 0xe8, 0x22, 0xb2, 0x25, 0x8c, 0x1b, 0x1d, 0x44, 0x14, 0x85, 0xea, 0x8a, 0x2d,
|
0x5c, 0x68, 0xd5, 0x76, 0xeb, 0x93, 0x4e, 0xe8, 0x22, 0xb2, 0x25, 0x8c, 0x1b, 0x1d, 0x44, 0x14,
|
||||||
0xbe, 0xad, 0xa7, 0xd0, 0x90, 0xe0, 0x36, 0xb2, 0xdd, 0x34, 0x27, 0x67, 0x55, 0xc3, 0x64, 0x4e,
|
0x85, 0xea, 0x8a, 0x2d, 0xbe, 0xad, 0xa7, 0xd0, 0x90, 0xe0, 0x0e, 0xb2, 0xbd, 0x34, 0x27, 0xe7,
|
||||||
0x97, 0x0a, 0x72, 0xda, 0xba, 0x0c, 0x97, 0x8a, 0x35, 0x2b, 0x5a, 0x7e, 0x34, 0xe0, 0xc2, 0xae,
|
0x55, 0xc3, 0x74, 0x4e, 0x97, 0x0a, 0x72, 0xda, 0xba, 0x0c, 0x97, 0x8a, 0x35, 0x2b, 0x5a, 0xbe,
|
||||||
0xe7, 0x51, 0x4c, 0x12, 0xf4, 0x1e, 0xf3, 0xde, 0xfd, 0x80, 0xfb, 0x4a, 0xb6, 0x35, 0x03, 0x32,
|
0x37, 0xe0, 0xc2, 0x9e, 0xe7, 0x51, 0x4c, 0x12, 0xf4, 0x1e, 0xf3, 0xde, 0xfd, 0x80, 0xfb, 0x4a,
|
||||||
0x34, 0xa4, 0xa5, 0xfa, 0x7a, 0x0a, 0x51, 0xac, 0x90, 0x7d, 0x58, 0x4b, 0x58, 0x44, 0xdd, 0x3e,
|
0x5a, 0x9a, 0x01, 0x19, 0x1a, 0xd2, 0x56, 0x7d, 0x3d, 0x85, 0x28, 0x56, 0xc8, 0x01, 0x6c, 0x24,
|
||||||
0x3a, 0x7c, 0x30, 0x38, 0xae, 0xd4, 0xa6, 0xea, 0xf3, 0x7c, 0x4b, 0x4c, 0x8b, 0x47, 0x91, 0x87,
|
0x2c, 0xa2, 0xee, 0x00, 0x1d, 0x3e, 0x18, 0x1c, 0x57, 0x6a, 0x53, 0x1d, 0xe1, 0x7c, 0x5b, 0x4c,
|
||||||
0xea, 0x1a, 0x9b, 0x28, 0x78, 0x46, 0x66, 0xbd, 0x2c, 0xc1, 0xba, 0x2a, 0xac, 0x27, 0xd4, 0x1f,
|
0x8b, 0x47, 0x91, 0x87, 0xea, 0x1a, 0x9b, 0x28, 0x78, 0x46, 0x66, 0xbd, 0x2a, 0xc1, 0xa6, 0x2a,
|
||||||
0x45, 0xf8, 0xf1, 0xc0, 0xe3, 0x31, 0xca, 0x64, 0xc9, 0x92, 0xce, 0x09, 0x4e, 0x0a, 0x2f, 0x71,
|
0xac, 0x27, 0xd4, 0x1f, 0x47, 0xf8, 0xf1, 0xd0, 0xe3, 0x31, 0xca, 0x64, 0xc9, 0x8a, 0xce, 0x09,
|
||||||
0xe5, 0xb6, 0xf8, 0x26, 0x75, 0x58, 0x54, 0x95, 0x2b, 0x2a, 0x7c, 0xce, 0xd6, 0x4b, 0x72, 0x1b,
|
0x4e, 0x0a, 0x2f, 0x71, 0xe5, 0xb6, 0xf8, 0x26, 0x75, 0x58, 0x56, 0x95, 0x2b, 0x2a, 0x7c, 0xc1,
|
||||||
0x20, 0xad, 0x50, 0x35, 0x09, 0x66, 0x96, 0x66, 0x06, 0x4e, 0x6e, 0x83, 0x19, 0xb8, 0x27, 0xba,
|
0xd6, 0x4b, 0x72, 0x1b, 0x20, 0xad, 0x50, 0x35, 0x09, 0xe6, 0x96, 0x66, 0x06, 0x4e, 0x6e, 0x83,
|
||||||
0x12, 0xd1, 0x2b, 0x9a, 0x03, 0x1b, 0x81, 0x7b, 0x72, 0x47, 0x03, 0xb2, 0x3d, 0xe2, 0x23, 0x00,
|
0x19, 0xb8, 0x27, 0xba, 0x12, 0xd1, 0x2b, 0x9a, 0x03, 0x5b, 0x81, 0x7b, 0x72, 0x47, 0x03, 0xb2,
|
||||||
0x3c, 0x89, 0x7d, 0xea, 0x72, 0xde, 0x55, 0xdd, 0xce, 0xe8, 0x93, 0x76, 0x06, 0x6d, 0xfd, 0x6c,
|
0x3d, 0xe2, 0x03, 0x00, 0x3c, 0x89, 0x7d, 0xea, 0x72, 0xde, 0x55, 0xdd, 0xce, 0xe9, 0x93, 0x76,
|
||||||
0xc0, 0x46, 0x9e, 0x16, 0x19, 0x36, 0xce, 0xcb, 0x7d, 0x58, 0x75, 0x75, 0xe0, 0x1c, 0x11, 0x0a,
|
0x06, 0x6d, 0xfd, 0x68, 0xc0, 0x56, 0x9e, 0x16, 0x19, 0x36, 0xce, 0xcb, 0x7d, 0x58, 0x77, 0x75,
|
||||||
0x9d, 0x82, 0x9b, 0x69, 0x0a, 0x16, 0x84, 0xd6, 0x5e, 0x19, 0x1d, 0x13, 0xeb, 0x84, 0xdc, 0x84,
|
0xe0, 0x1c, 0x11, 0x0a, 0x9d, 0x82, 0xdb, 0x69, 0x0a, 0x16, 0x84, 0xd6, 0x5e, 0x1b, 0x1f, 0x13,
|
||||||
0x73, 0x34, 0x8a, 0x98, 0x13, 0xfb, 0xd8, 0xc5, 0x51, 0x26, 0xed, 0xad, 0xf0, 0x86, 0xfd, 0xc7,
|
0xeb, 0x84, 0xdc, 0x84, 0x73, 0x34, 0x8a, 0x98, 0x13, 0xfb, 0xd8, 0xc3, 0x71, 0x26, 0xed, 0xaf,
|
||||||
0xab, 0xad, 0xc5, 0x23, 0x2e, 0x3f, 0x3c, 0xb0, 0x6b, 0x1c, 0x25, 0x17, 0x9e, 0xf5, 0x5b, 0x6a,
|
0xf1, 0x86, 0xfd, 0xdb, 0xeb, 0x9d, 0xe5, 0x63, 0x2e, 0x3f, 0x3a, 0xb4, 0x6b, 0x1c, 0x25, 0x17,
|
||||||
0xda, 0x7e, 0x14, 0x70, 0xbd, 0x67, 0x1d, 0xb2, 0x0f, 0x60, 0x51, 0xc5, 0x47, 0xc5, 0x8b, 0x64,
|
0x9e, 0xf5, 0x4b, 0x6a, 0xda, 0x41, 0x14, 0x70, 0xbd, 0x67, 0x1d, 0xb2, 0xff, 0xc1, 0xb2, 0x8a,
|
||||||
0xe2, 0x75, 0x24, 0xbf, 0x6c, 0x0d, 0x21, 0xb7, 0x61, 0x25, 0xa2, 0x7e, 0xdf, 0x0f, 0xdd, 0x81,
|
0x8f, 0x8a, 0x17, 0xc9, 0xc4, 0xeb, 0x58, 0x7e, 0xd9, 0x1a, 0x42, 0x6e, 0xc3, 0x5a, 0x44, 0xfd,
|
||||||
0x66, 0x63, 0x41, 0xb0, 0x51, 0x94, 0xba, 0xcb, 0x1a, 0x2a, 0x19, 0xb0, 0xee, 0x43, 0x7d, 0xcc,
|
0x81, 0x1f, 0xba, 0x43, 0xcd, 0xc6, 0x92, 0x60, 0xa3, 0x28, 0x75, 0x57, 0x35, 0x54, 0x32, 0x60,
|
||||||
0x97, 0x94, 0xe7, 0x8c, 0x19, 0xc6, 0xa9, 0x66, 0x58, 0x2e, 0x5c, 0x54, 0x9a, 0x0e, 0xa2, 0x1f,
|
0xdd, 0x87, 0xfa, 0x84, 0x2f, 0x29, 0xcf, 0x19, 0x33, 0x8c, 0x53, 0xcd, 0xb0, 0x5c, 0xb8, 0xa8,
|
||||||
0xc2, 0x41, 0xe4, 0x7a, 0x67, 0xcd, 0x8b, 0xf5, 0xd2, 0x00, 0x73, 0xe2, 0x8e, 0x77, 0x91, 0x17,
|
0x34, 0x1d, 0x46, 0xdf, 0x85, 0xc3, 0xc8, 0xf5, 0xce, 0x9a, 0x17, 0xeb, 0x95, 0x01, 0xe6, 0xd4,
|
||||||
0x19, 0xcf, 0x4b, 0xa7, 0x7b, 0xfe, 0x35, 0xfc, 0x5f, 0x59, 0x75, 0x18, 0xf6, 0xa2, 0x33, 0xf7,
|
0x1d, 0xef, 0x23, 0x2f, 0x32, 0x9e, 0x97, 0x4e, 0xf7, 0xfc, 0x4b, 0xf8, 0xb7, 0xb2, 0xea, 0x28,
|
||||||
0xfa, 0xee, 0xa8, 0x41, 0x48, 0xf5, 0x85, 0x01, 0x7a, 0x0d, 0x33, 0x9d, 0x51, 0xda, 0xe6, 0x66,
|
0xec, 0x47, 0x67, 0xee, 0xf5, 0xdd, 0x71, 0x83, 0x90, 0xea, 0x0b, 0x03, 0xf4, 0x06, 0x66, 0x3a,
|
||||||
0xc9, 0xd9, 0x19, 0xea, 0x8d, 0x72, 0x29, 0x3f, 0x81, 0xce, 0x34, 0x36, 0xd6, 0x9f, 0x06, 0xac,
|
0xe3, 0xb4, 0xcd, 0xcd, 0x92, 0xb3, 0x33, 0xd4, 0x1b, 0xe7, 0x52, 0x7e, 0x02, 0x9d, 0x69, 0x6c,
|
||||||
0xf3, 0xc9, 0xa2, 0xae, 0x4a, 0x5e, 0xc3, 0x8d, 0x75, 0x28, 0xc7, 0x14, 0x7b, 0xfe, 0x89, 0x72,
|
0xac, 0xdf, 0x0d, 0xd8, 0xe4, 0x93, 0x45, 0x5d, 0x95, 0xbc, 0x81, 0x1b, 0x9b, 0x50, 0x8e, 0x29,
|
||||||
0x44, 0xad, 0xc8, 0x16, 0xd4, 0x12, 0xe6, 0x52, 0xe6, 0xb8, 0x3d, 0xce, 0xa1, 0x7c, 0x1a, 0x81,
|
0xf6, 0xfd, 0x13, 0xe5, 0x88, 0x5a, 0x91, 0x1d, 0xa8, 0x25, 0xcc, 0xa5, 0xcc, 0x71, 0xfb, 0x9c,
|
||||||
0x10, 0xed, 0x72, 0x09, 0xd9, 0x04, 0xc0, 0xd0, 0x73, 0x3a, 0xd8, 0xe3, 0x73, 0x6b, 0x5e, 0xec,
|
0x43, 0xf9, 0x34, 0x02, 0x21, 0xda, 0xe3, 0x12, 0xb2, 0x0d, 0x80, 0xa1, 0xe7, 0x74, 0xb1, 0xcf,
|
||||||
0x57, 0x31, 0xf4, 0xf6, 0x84, 0x80, 0x5c, 0x82, 0x2a, 0x45, 0x3e, 0x38, 0xfd, 0xef, 0x65, 0x33,
|
0xe7, 0xd6, 0xa2, 0xd8, 0xaf, 0x62, 0xe8, 0xed, 0x0b, 0x01, 0xb9, 0x04, 0x55, 0x8a, 0x7c, 0x70,
|
||||||
0xac, 0xd8, 0xa9, 0x20, 0x1d, 0xa5, 0xe5, 0xcc, 0x28, 0xe5, 0x2a, 0xb9, 0xbf, 0x4e, 0x6f, 0xe0,
|
0xfa, 0xdf, 0xca, 0x66, 0x58, 0xb1, 0x53, 0x41, 0x3a, 0x4a, 0xcb, 0x99, 0x51, 0xca, 0x55, 0x72,
|
||||||
0xf6, 0xe5, 0xab, 0x75, 0xd1, 0xae, 0x72, 0xc9, 0x5d, 0x2e, 0xb0, 0x7e, 0x37, 0x60, 0x23, 0xef,
|
0x7f, 0x9d, 0xfe, 0xd0, 0x1d, 0xc8, 0x57, 0xeb, 0xb2, 0x5d, 0xe5, 0x92, 0xbb, 0x5c, 0x60, 0xfd,
|
||||||
0x5d, 0xca, 0xe1, 0xc7, 0xf9, 0x79, 0x7b, 0x3d, 0x25, 0x6e, 0xca, 0x89, 0xd6, 0x29, 0xd3, 0xd7,
|
0x6a, 0xc0, 0x56, 0xde, 0xbb, 0x94, 0xc3, 0x0f, 0xf3, 0xf3, 0xf6, 0x7a, 0x4a, 0xdc, 0x8c, 0x13,
|
||||||
0x44, 0x98, 0xd7, 0xaf, 0x5c, 0x11, 0x67, 0x23, 0x13, 0xe7, 0x37, 0x4a, 0x2e, 0xd2, 0x80, 0xaa,
|
0xed, 0x53, 0xa6, 0xaf, 0x89, 0xb0, 0xa8, 0x5f, 0xb9, 0x22, 0xce, 0x46, 0x26, 0xce, 0x6f, 0x95,
|
||||||
0x9f, 0x38, 0x8a, 0xe5, 0x39, 0x71, 0x45, 0xc5, 0x4f, 0x8e, 0xc4, 0xda, 0x7a, 0xc6, 0x13, 0xa3,
|
0x5c, 0xa4, 0x01, 0x55, 0x3f, 0x71, 0x14, 0xcb, 0x0b, 0xe2, 0x8a, 0x8a, 0x9f, 0x1c, 0x8b, 0xb5,
|
||||||
0x60, 0xbc, 0x73, 0xa7, 0xb6, 0xa0, 0x26, 0xa3, 0xe4, 0x64, 0x06, 0x3d, 0x48, 0xd1, 0x23, 0x3e,
|
0xf5, 0x8c, 0x27, 0x46, 0xc1, 0x78, 0xe7, 0x4e, 0xed, 0x40, 0x4d, 0x46, 0xc9, 0xc9, 0x0c, 0x7a,
|
||||||
0xee, 0x37, 0x01, 0x62, 0x97, 0xb2, 0x10, 0x69, 0x3a, 0xea, 0xab, 0x4a, 0x72, 0xe8, 0x59, 0x0d,
|
0x90, 0xa2, 0x47, 0x7c, 0xdc, 0x6f, 0x03, 0xc4, 0x2e, 0x65, 0x21, 0xd2, 0x74, 0xd4, 0x57, 0x95,
|
||||||
0xde, 0x76, 0x8a, 0x06, 0xfc, 0xe3, 0x81, 0x67, 0xad, 0x01, 0x39, 0xa2, 0xd1, 0x37, 0xd8, 0xcd,
|
0xe4, 0xc8, 0xb3, 0x1a, 0xbc, 0xed, 0x14, 0x0d, 0xf8, 0xc7, 0x43, 0xcf, 0xda, 0x00, 0x72, 0x4c,
|
||||||
0x56, 0xa6, 0x75, 0x0b, 0x2e, 0xe4, 0xa4, 0xea, 0x39, 0x73, 0x05, 0x96, 0x62, 0x29, 0x76, 0x12,
|
0xa3, 0xaf, 0xb0, 0x97, 0xad, 0x4c, 0xeb, 0x16, 0x5c, 0xc8, 0x49, 0xd5, 0x73, 0xe6, 0x0a, 0xac,
|
||||||
0x77, 0xa0, 0x73, 0xa8, 0xa6, 0x64, 0x6d, 0x77, 0xc0, 0x76, 0xfe, 0xae, 0x40, 0xe5, 0xa1, 0x22,
|
0xc4, 0x52, 0xec, 0x24, 0xee, 0x50, 0xe7, 0x50, 0x4d, 0xc9, 0x3a, 0xee, 0x90, 0xed, 0xfe, 0x59,
|
||||||
0x9d, 0x3c, 0x84, 0x25, 0xf9, 0x7a, 0x54, 0xff, 0x79, 0x9b, 0xe3, 0xef, 0x9f, 0xdc, 0xc3, 0xde,
|
0x81, 0xca, 0x43, 0x45, 0x3a, 0x79, 0x08, 0x2b, 0xf2, 0xf5, 0xa8, 0xfe, 0xf3, 0xb6, 0x27, 0xdf,
|
||||||
0xbc, 0x3c, 0x6d, 0x5b, 0x5d, 0x7f, 0x00, 0xd5, 0x7b, 0xc8, 0x94, 0x2e, 0x73, 0x1c, 0x9c, 0xbe,
|
0x3f, 0xb9, 0x87, 0xbd, 0x79, 0x79, 0xd6, 0xb6, 0xba, 0xfe, 0x10, 0xaa, 0xf7, 0x90, 0x29, 0x5d,
|
||||||
0x32, 0xcd, 0x46, 0xe1, 0x9e, 0xd2, 0xf2, 0x10, 0x96, 0x64, 0xf1, 0x4d, 0x33, 0x2a, 0x57, 0xfb,
|
0xe6, 0x24, 0x38, 0x7d, 0x65, 0x9a, 0x8d, 0xc2, 0x3d, 0xa5, 0xe5, 0x21, 0xac, 0xc8, 0xe2, 0x9b,
|
||||||
0x93, 0x46, 0xe5, 0x2b, 0x97, 0xdc, 0x87, 0x1a, 0xcf, 0x2d, 0xb9, 0x97, 0x90, 0x46, 0xd1, 0x13,
|
0x65, 0x54, 0xae, 0xf6, 0xa7, 0x8d, 0xca, 0x57, 0x2e, 0xb9, 0x0f, 0x35, 0x9e, 0x5b, 0x72, 0x2f,
|
||||||
0x4f, 0xeb, 0xba, 0x54, 0xbc, 0xa9, 0x34, 0x21, 0xac, 0xb5, 0xb5, 0x7b, 0x99, 0x68, 0x91, 0x6b,
|
0x21, 0x8d, 0xa2, 0x27, 0x9e, 0xd6, 0x75, 0xa9, 0x78, 0x53, 0x69, 0x42, 0xd8, 0xe8, 0x68, 0xf7,
|
||||||
0xe3, 0xa7, 0x0a, 0x33, 0xc5, 0xbc, 0x7e, 0x1a, 0x4c, 0x5d, 0xf3, 0x04, 0x56, 0x25, 0xaf, 0xaa,
|
0x32, 0xd1, 0x22, 0xd7, 0x26, 0x4f, 0x15, 0x66, 0x8a, 0x79, 0xfd, 0x34, 0x98, 0xba, 0xe6, 0x09,
|
||||||
0x1c, 0x78, 0x8a, 0x35, 0xd3, 0xb3, 0xc5, 0x2f, 0x2d, 0xf3, 0xca, 0x34, 0x44, 0x5a, 0x7c, 0x5f,
|
0xac, 0x4b, 0x5e, 0x55, 0x39, 0xf0, 0x14, 0x6b, 0xa6, 0x67, 0x8b, 0x5f, 0x5a, 0xe6, 0x95, 0x59,
|
||||||
0xc2, 0xaa, 0x9c, 0x90, 0x19, 0xc5, 0x93, 0xc7, 0xc6, 0x1f, 0x04, 0xa6, 0x35, 0x15, 0x92, 0xaa,
|
0x88, 0xb4, 0xf8, 0x3e, 0x87, 0x75, 0x39, 0x21, 0x33, 0x8a, 0xa7, 0x8f, 0x4d, 0x3e, 0x08, 0x4c,
|
||||||
0x6e, 0xc3, 0x72, 0xa6, 0xc1, 0x8b, 0xa2, 0x98, 0x38, 0x95, 0x9f, 0x2c, 0x66, 0x73, 0x0a, 0x20,
|
0x6b, 0x26, 0x24, 0x55, 0xdd, 0x81, 0xd5, 0x4c, 0x83, 0x17, 0x45, 0x31, 0x75, 0x2a, 0x3f, 0x59,
|
||||||
0x55, 0xea, 0x00, 0xd1, 0x33, 0x32, 0x63, 0xf1, 0xd5, 0x89, 0x73, 0x93, 0xc3, 0xda, 0x7c, 0x6f,
|
0xcc, 0xe6, 0x0c, 0x40, 0xaa, 0xd4, 0x01, 0xa2, 0x67, 0x64, 0xc6, 0xe2, 0xab, 0x53, 0xe7, 0xa6,
|
||||||
0x06, 0x28, 0x47, 0x88, 0x4c, 0x96, 0x99, 0x84, 0x8c, 0x8f, 0x9a, 0x02, 0x42, 0x26, 0x87, 0xc5,
|
0x87, 0xb5, 0xf9, 0x9f, 0x39, 0xa0, 0x1c, 0x21, 0x32, 0x59, 0xe6, 0x12, 0x32, 0x39, 0x6a, 0x0a,
|
||||||
0x17, 0xb0, 0x92, 0xed, 0x68, 0x63, 0x31, 0x2c, 0x6e, 0xfe, 0xd9, 0x18, 0x4e, 0x6b, 0xa0, 0x5f,
|
0x08, 0x99, 0x1e, 0x16, 0x9f, 0xc1, 0x5a, 0xb6, 0xa3, 0x4d, 0xc4, 0xb0, 0xb8, 0xf9, 0x67, 0x63,
|
||||||
0xc1, 0xf9, 0x7c, 0xda, 0x70, 0x61, 0xce, 0xa0, 0xe2, 0x26, 0x65, 0x5e, 0x9d, 0x8e, 0x49, 0xb5,
|
0x38, 0xab, 0x81, 0x7e, 0x01, 0xe7, 0xf3, 0x69, 0xc3, 0x85, 0x39, 0x83, 0x8a, 0x9b, 0x94, 0x79,
|
||||||
0x7f, 0x06, 0xb5, 0x4c, 0x5b, 0x21, 0x99, 0x72, 0x98, 0xec, 0x41, 0xe6, 0xe6, 0x94, 0x5d, 0xa9,
|
0x75, 0x36, 0x26, 0xd5, 0xfe, 0x09, 0xd4, 0x32, 0x6d, 0x85, 0x64, 0xca, 0x61, 0xba, 0x07, 0x99,
|
||||||
0x6e, 0x6f, 0xfe, 0x59, 0x29, 0xee, 0x74, 0xca, 0xe2, 0x91, 0x7c, 0xf3, 0xdf, 0x00, 0x00, 0x00,
|
0xdb, 0x33, 0x76, 0xa5, 0xba, 0xfd, 0xc5, 0x67, 0xa5, 0xb8, 0xdb, 0x2d, 0x8b, 0x47, 0xf2, 0xcd,
|
||||||
0xff, 0xff, 0x7b, 0xfa, 0x2e, 0x7b, 0xd1, 0x12, 0x00, 0x00,
|
0xbf, 0x03, 0x00, 0x00, 0xff, 0xff, 0xb9, 0xbc, 0x93, 0xd5, 0xd1, 0x12, 0x00, 0x00,
|
||||||
}
|
}
|
||||||
|
|
||||||
// Reference imports to suppress errors if they are not otherwise used.
|
// Reference imports to suppress errors if they are not otherwise used.
|
||||||
|
35
pkg/storj/bucket.go
Normal file
35
pkg/storj/bucket.go
Normal file
@ -0,0 +1,35 @@
|
|||||||
|
// Copyright (C) 2019 Storj Labs, Inc.
|
||||||
|
// See LICENSE for copying information.
|
||||||
|
|
||||||
|
package storj
|
||||||
|
|
||||||
|
import (
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/skyrings/skyring-common/tools/uuid"
|
||||||
|
"github.com/zeebo/errs"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
// ErrBucket is an error class for general bucket errors
|
||||||
|
ErrBucket = errs.Class("bucket")
|
||||||
|
|
||||||
|
// ErrNoBucket is an error class for using empty bucket name
|
||||||
|
ErrNoBucket = errs.Class("no bucket specified")
|
||||||
|
|
||||||
|
// ErrBucketNotFound is an error class for non-existing bucket
|
||||||
|
ErrBucketNotFound = errs.Class("bucket not found")
|
||||||
|
)
|
||||||
|
|
||||||
|
// Bucket contains information about a specific bucket
|
||||||
|
type Bucket struct {
|
||||||
|
ID uuid.UUID
|
||||||
|
Name string
|
||||||
|
ProjectID uuid.UUID
|
||||||
|
Attribution string
|
||||||
|
Created time.Time
|
||||||
|
PathCipher CipherSuite
|
||||||
|
DefaultSegmentsSize int64
|
||||||
|
DefaultRedundancyScheme RedundancyScheme
|
||||||
|
DefaultEncryptionParameters EncryptionParameters
|
||||||
|
}
|
@ -10,30 +10,13 @@ import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
// ErrNoBucket is an error class for using empty bucket name
|
|
||||||
ErrNoBucket = errs.Class("no bucket specified")
|
|
||||||
|
|
||||||
// ErrNoPath is an error class for using empty path
|
// ErrNoPath is an error class for using empty path
|
||||||
ErrNoPath = errs.Class("no path specified")
|
ErrNoPath = errs.Class("no path specified")
|
||||||
|
|
||||||
// ErrBucketNotFound is an error class for non-existing bucket
|
|
||||||
ErrBucketNotFound = errs.Class("bucket not found")
|
|
||||||
|
|
||||||
// ErrObjectNotFound is an error class for non-existing object
|
// ErrObjectNotFound is an error class for non-existing object
|
||||||
ErrObjectNotFound = errs.Class("object not found")
|
ErrObjectNotFound = errs.Class("object not found")
|
||||||
)
|
)
|
||||||
|
|
||||||
// Bucket contains information about a specific bucket
|
|
||||||
type Bucket struct {
|
|
||||||
Name string
|
|
||||||
Attribution string
|
|
||||||
Created time.Time
|
|
||||||
PathCipher CipherSuite
|
|
||||||
SegmentsSize int64
|
|
||||||
RedundancyScheme RedundancyScheme
|
|
||||||
EncryptionParameters EncryptionParameters
|
|
||||||
}
|
|
||||||
|
|
||||||
// Object contains information about a specific object
|
// Object contains information about a specific object
|
||||||
type Object struct {
|
type Object struct {
|
||||||
Version uint32
|
Version uint32
|
||||||
|
12
proto.lock
12
proto.lock
@ -189,7 +189,7 @@
|
|||||||
"def": {
|
"def": {
|
||||||
"enums": [
|
"enums": [
|
||||||
{
|
{
|
||||||
"name": "CipherSuite.Type",
|
"name": "CipherSuite",
|
||||||
"enum_fields": [
|
"enum_fields": [
|
||||||
{
|
{
|
||||||
"name": "ENC_UNSPECIFIED"
|
"name": "ENC_UNSPECIFIED"
|
||||||
@ -224,16 +224,6 @@
|
|||||||
"type": "int64"
|
"type": "int64"
|
||||||
}
|
}
|
||||||
]
|
]
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "CipherSuite",
|
|
||||||
"fields": [
|
|
||||||
{
|
|
||||||
"id": 1,
|
|
||||||
"name": "type",
|
|
||||||
"type": "Type"
|
|
||||||
}
|
|
||||||
]
|
|
||||||
}
|
}
|
||||||
],
|
],
|
||||||
"package": {
|
"package": {
|
||||||
|
24
satellite/metainfo/db.go
Normal file
24
satellite/metainfo/db.go
Normal file
@ -0,0 +1,24 @@
|
|||||||
|
// Copyright (C) 2018 Storj Labs, Inc.
|
||||||
|
// See LICENSE for copying information.
|
||||||
|
|
||||||
|
package metainfo
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
|
||||||
|
"github.com/skyrings/skyring-common/tools/uuid"
|
||||||
|
|
||||||
|
"storj.io/storj/pkg/storj"
|
||||||
|
)
|
||||||
|
|
||||||
|
// BucketsDB is the interface for the database to interact with buckets
|
||||||
|
type BucketsDB interface {
|
||||||
|
// Create creates a new bucket
|
||||||
|
CreateBucket(ctx context.Context, bucket storj.Bucket) (_ storj.Bucket, err error)
|
||||||
|
// Get returns an existing bucket
|
||||||
|
GetBucket(ctx context.Context, bucketName []byte, projectID uuid.UUID) (bucket storj.Bucket, err error)
|
||||||
|
// Delete deletes a bucket
|
||||||
|
DeleteBucket(ctx context.Context, bucketName []byte, projectID uuid.UUID) (err error)
|
||||||
|
// List returns all buckets for a project
|
||||||
|
ListBuckets(ctx context.Context, projectID uuid.UUID, listOpts storj.BucketListOptions, allowedBuckets map[string]struct{}) (bucketList storj.BucketList, err error)
|
||||||
|
}
|
191
satellite/metainfo/db_test.go
Normal file
191
satellite/metainfo/db_test.go
Normal file
@ -0,0 +1,191 @@
|
|||||||
|
// Copyright (C) 2019 Storj Labs, Inc.
|
||||||
|
// See LICENSE for copying information.
|
||||||
|
|
||||||
|
package metainfo_test
|
||||||
|
|
||||||
|
import (
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/skyrings/skyring-common/tools/uuid"
|
||||||
|
"github.com/stretchr/testify/require"
|
||||||
|
|
||||||
|
"storj.io/storj/internal/testcontext"
|
||||||
|
"storj.io/storj/internal/testrand"
|
||||||
|
"storj.io/storj/pkg/storj"
|
||||||
|
"storj.io/storj/satellite"
|
||||||
|
"storj.io/storj/satellite/console"
|
||||||
|
"storj.io/storj/satellite/satellitedb/satellitedbtest"
|
||||||
|
)
|
||||||
|
|
||||||
|
func newTestBucket(name string, projectID uuid.UUID) storj.Bucket {
|
||||||
|
return storj.Bucket{
|
||||||
|
ID: testrand.UUID(),
|
||||||
|
Name: name,
|
||||||
|
ProjectID: projectID,
|
||||||
|
PathCipher: storj.EncAESGCM,
|
||||||
|
DefaultSegmentsSize: 65536,
|
||||||
|
DefaultRedundancyScheme: storj.RedundancyScheme{
|
||||||
|
Algorithm: storj.ReedSolomon,
|
||||||
|
ShareSize: 9,
|
||||||
|
RequiredShares: 10,
|
||||||
|
RepairShares: 11,
|
||||||
|
OptimalShares: 12,
|
||||||
|
TotalShares: 13,
|
||||||
|
},
|
||||||
|
DefaultEncryptionParameters: storj.EncryptionParameters{
|
||||||
|
CipherSuite: storj.EncAESGCM,
|
||||||
|
BlockSize: 32,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestBasicBucketOperations(t *testing.T) {
|
||||||
|
satellitedbtest.Run(t, func(t *testing.T, db satellite.DB) {
|
||||||
|
ctx := testcontext.New(t)
|
||||||
|
defer ctx.Cleanup()
|
||||||
|
consoleDB := db.Console()
|
||||||
|
project, err := consoleDB.Projects().Insert(ctx, &console.Project{Name: "testproject1"})
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
bucketsDB := db.Buckets()
|
||||||
|
expectedBucket := newTestBucket("testbucket", project.ID)
|
||||||
|
|
||||||
|
// CreateBucket
|
||||||
|
_, err = bucketsDB.CreateBucket(ctx, expectedBucket)
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
// GetBucket
|
||||||
|
bucket, err := bucketsDB.GetBucket(ctx, []byte("testbucket"), project.ID)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, expectedBucket.ID, bucket.ID)
|
||||||
|
require.Equal(t, expectedBucket.Name, bucket.Name)
|
||||||
|
require.Equal(t, expectedBucket.ProjectID, bucket.ProjectID)
|
||||||
|
require.Equal(t, expectedBucket.PathCipher, bucket.PathCipher)
|
||||||
|
require.Equal(t, expectedBucket.DefaultSegmentsSize, bucket.DefaultSegmentsSize)
|
||||||
|
require.Equal(t, expectedBucket.DefaultRedundancyScheme, bucket.DefaultRedundancyScheme)
|
||||||
|
require.Equal(t, expectedBucket.DefaultEncryptionParameters, bucket.DefaultEncryptionParameters)
|
||||||
|
|
||||||
|
// DeleteBucket
|
||||||
|
err = bucketsDB.DeleteBucket(ctx, []byte("testbucket"), project.ID)
|
||||||
|
require.NoError(t, err)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestListBucketsAllAllowed(t *testing.T) {
|
||||||
|
testCases := []struct {
|
||||||
|
name string
|
||||||
|
cursor string
|
||||||
|
limit int
|
||||||
|
expectedItems int
|
||||||
|
expectedMore bool
|
||||||
|
}{
|
||||||
|
{"empty string cursor", "", 10, 10, false},
|
||||||
|
{"last bucket cursor", "zzz", 2, 1, false},
|
||||||
|
{"non matching cursor", "ccc", 10, 5, false},
|
||||||
|
{"first bucket cursor", "0test", 10, 10, false},
|
||||||
|
{"empty string cursor, more", "", 5, 5, true},
|
||||||
|
{"non matching cursor, more", "ccc", 3, 3, true},
|
||||||
|
{"first bucket cursor, more", "0test", 5, 5, true},
|
||||||
|
}
|
||||||
|
satellitedbtest.Run(t, func(t *testing.T, db satellite.DB) {
|
||||||
|
ctx := testcontext.New(t)
|
||||||
|
defer ctx.Cleanup()
|
||||||
|
consoleDB := db.Console()
|
||||||
|
project, err := consoleDB.Projects().Insert(ctx, &console.Project{Name: "testproject1"})
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
bucketsDB := db.Buckets()
|
||||||
|
|
||||||
|
var allowedPaths = map[string]struct{}{}
|
||||||
|
{ // setup some test buckets
|
||||||
|
var testBucketNames = []string{"aaa", "bbb", "mmm", "qqq", "zzz",
|
||||||
|
"test.bucket", "123", "0test", "999", "test-bucket.thing",
|
||||||
|
}
|
||||||
|
for _, bucket := range testBucketNames {
|
||||||
|
testBucket := newTestBucket(bucket, project.ID)
|
||||||
|
_, err := bucketsDB.CreateBucket(ctx, testBucket)
|
||||||
|
allowedPaths[bucket] = struct{}{}
|
||||||
|
if err != nil {
|
||||||
|
require.NoError(t, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, tt := range testCases {
|
||||||
|
tt := tt // avoid scopelint error
|
||||||
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
|
bucketList, err := bucketsDB.ListBuckets(ctx, project.ID, storj.BucketListOptions{
|
||||||
|
Cursor: tt.cursor,
|
||||||
|
Direction: storj.Forward,
|
||||||
|
Limit: tt.limit,
|
||||||
|
},
|
||||||
|
allowedPaths,
|
||||||
|
)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, tt.expectedItems, len(bucketList.Items))
|
||||||
|
require.Equal(t, tt.expectedMore, bucketList.More)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestListBucketsNotAllowed(t *testing.T) {
|
||||||
|
testCases := []struct {
|
||||||
|
name string
|
||||||
|
cursor string
|
||||||
|
limit int
|
||||||
|
expectedItems int
|
||||||
|
expectedMore bool
|
||||||
|
allowedPaths map[string]struct{}
|
||||||
|
expectedNames []string
|
||||||
|
}{
|
||||||
|
{"empty string cursor, 2 allowed", "", 10, 1, false, map[string]struct{}{"aaa": {}, "ddd": {}}, []string{"aaa"}},
|
||||||
|
{"empty string cursor, more", "", 2, 2, true, map[string]struct{}{"aaa": {}, "bbb": {}, "zzz": {}}, []string{"aaa", "bbb"}},
|
||||||
|
{"empty string cursor, 3 allowed", "", 4, 3, false, map[string]struct{}{"aaa": {}, "bbb": {}, "zzz": {}}, []string{"aaa", "bbb", "zzz"}},
|
||||||
|
{"last bucket cursor", "zzz", 2, 1, false, map[string]struct{}{"zzz": {}}, []string{"zzz"}},
|
||||||
|
}
|
||||||
|
satellitedbtest.Run(t, func(t *testing.T, db satellite.DB) {
|
||||||
|
ctx := testcontext.New(t)
|
||||||
|
defer ctx.Cleanup()
|
||||||
|
consoleDB := db.Console()
|
||||||
|
project, err := consoleDB.Projects().Insert(ctx, &console.Project{Name: "testproject1"})
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
bucketsDB := db.Buckets()
|
||||||
|
|
||||||
|
{ // setup some test buckets
|
||||||
|
var testBucketNames = []string{"aaa", "bbb", "mmm", "qqq", "zzz",
|
||||||
|
"test.bucket", "123", "0test", "999", "test-bucket.thing",
|
||||||
|
}
|
||||||
|
for _, bucket := range testBucketNames {
|
||||||
|
testBucket := newTestBucket(bucket, project.ID)
|
||||||
|
_, err := bucketsDB.CreateBucket(ctx, testBucket)
|
||||||
|
if err != nil {
|
||||||
|
require.NoError(t, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, tt := range testCases {
|
||||||
|
tt := tt // avoid scopelint error
|
||||||
|
listOpts := storj.BucketListOptions{
|
||||||
|
Cursor: tt.cursor,
|
||||||
|
Direction: storj.Forward,
|
||||||
|
Limit: tt.limit,
|
||||||
|
}
|
||||||
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
|
bucketList, err := bucketsDB.ListBuckets(ctx, project.ID,
|
||||||
|
listOpts,
|
||||||
|
tt.allowedPaths,
|
||||||
|
)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, tt.expectedItems, len(bucketList.Items))
|
||||||
|
require.Equal(t, tt.expectedMore, bucketList.More)
|
||||||
|
for _, actualItem := range bucketList.Items {
|
||||||
|
require.Contains(t, tt.expectedNames, actualItem.Name)
|
||||||
|
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
@ -19,6 +19,7 @@ import (
|
|||||||
monkit "gopkg.in/spacemonkeygo/monkit.v2"
|
monkit "gopkg.in/spacemonkeygo/monkit.v2"
|
||||||
|
|
||||||
"storj.io/storj/pkg/accounting"
|
"storj.io/storj/pkg/accounting"
|
||||||
|
"storj.io/storj/pkg/auth"
|
||||||
"storj.io/storj/pkg/eestream"
|
"storj.io/storj/pkg/eestream"
|
||||||
"storj.io/storj/pkg/identity"
|
"storj.io/storj/pkg/identity"
|
||||||
"storj.io/storj/pkg/macaroon"
|
"storj.io/storj/pkg/macaroon"
|
||||||
@ -665,37 +666,205 @@ func (endpoint *Endpoint) ProjectInfo(ctx context.Context, req *pb.ProjectInfoRe
|
|||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// CreateBucket creates a bucket
|
// GetBucket returns a bucket
|
||||||
func (endpoint *Endpoint) CreateBucket(ctx context.Context, req *pb.BucketCreateRequest) (_ *pb.BucketCreateResponse, err error) {
|
func (endpoint *Endpoint) GetBucket(ctx context.Context, req *pb.BucketGetRequest) (resp *pb.BucketGetResponse, err error) {
|
||||||
defer mon.Task()(&ctx)(&err)
|
defer mon.Task()(&ctx)(&err)
|
||||||
// TODO: placeholder to implement pb.MetainfoServer interface.
|
|
||||||
return &pb.BucketCreateResponse{}, err
|
keyInfo, err := endpoint.validateAuth(ctx, macaroon.Action{
|
||||||
|
Op: macaroon.ActionRead,
|
||||||
|
Bucket: req.Name,
|
||||||
|
Time: time.Now(),
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
return nil, status.Errorf(codes.Unauthenticated, err.Error())
|
||||||
|
}
|
||||||
|
|
||||||
|
bucket, err := endpoint.metainfo.GetBucket(ctx, req.GetName(), keyInfo.ProjectID)
|
||||||
|
if err != nil {
|
||||||
|
return nil, status.Errorf(codes.Internal, err.Error())
|
||||||
|
}
|
||||||
|
|
||||||
|
return &pb.BucketGetResponse{
|
||||||
|
Bucket: convertBucketToProto(ctx, bucket),
|
||||||
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// GetBucket gets a bucket
|
// CreateBucket creates a new bucket
|
||||||
func (endpoint *Endpoint) GetBucket(ctx context.Context, req *pb.BucketGetRequest) (_ *pb.BucketGetResponse, err error) {
|
func (endpoint *Endpoint) CreateBucket(ctx context.Context, req *pb.BucketCreateRequest) (resp *pb.BucketCreateResponse, err error) {
|
||||||
defer mon.Task()(&ctx)(&err)
|
defer mon.Task()(&ctx)(&err)
|
||||||
// TODO: placeholder to implement pb.MetainfoServer interface.
|
|
||||||
return &pb.BucketGetResponse{}, err
|
keyInfo, err := endpoint.validateAuth(ctx, macaroon.Action{
|
||||||
|
Op: macaroon.ActionWrite,
|
||||||
|
Bucket: req.Name,
|
||||||
|
Time: time.Now(),
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
return nil, status.Errorf(codes.Unauthenticated, err.Error())
|
||||||
|
}
|
||||||
|
|
||||||
|
err = endpoint.validateBucket(ctx, req.Name)
|
||||||
|
if err != nil {
|
||||||
|
return nil, status.Errorf(codes.InvalidArgument, err.Error())
|
||||||
|
}
|
||||||
|
|
||||||
|
err = endpoint.validateRedundancy(ctx, req.GetDefaultRedundancyScheme())
|
||||||
|
if err != nil {
|
||||||
|
return nil, status.Errorf(codes.InvalidArgument, err.Error())
|
||||||
|
}
|
||||||
|
|
||||||
|
bucket, err := convertProtoToBucket(req, keyInfo.ProjectID)
|
||||||
|
if err != nil {
|
||||||
|
return nil, status.Errorf(codes.InvalidArgument, err.Error())
|
||||||
|
}
|
||||||
|
|
||||||
|
bucket, err = endpoint.metainfo.CreateBucket(ctx, bucket)
|
||||||
|
if err != nil {
|
||||||
|
return nil, Error.Wrap(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return &pb.BucketCreateResponse{
|
||||||
|
Bucket: convertBucketToProto(ctx, bucket),
|
||||||
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// DeleteBucket deletes a bucket
|
// DeleteBucket deletes a bucket
|
||||||
func (endpoint *Endpoint) DeleteBucket(ctx context.Context, req *pb.BucketDeleteRequest) (_ *pb.BucketDeleteResponse, err error) {
|
func (endpoint *Endpoint) DeleteBucket(ctx context.Context, req *pb.BucketDeleteRequest) (resp *pb.BucketDeleteResponse, err error) {
|
||||||
defer mon.Task()(&ctx)(&err)
|
defer mon.Task()(&ctx)(&err)
|
||||||
// TODO: placeholder to implement pb.MetainfoServer interface.
|
|
||||||
return &pb.BucketDeleteResponse{}, err
|
keyInfo, err := endpoint.validateAuth(ctx, macaroon.Action{
|
||||||
|
Op: macaroon.ActionDelete,
|
||||||
|
Bucket: req.Name,
|
||||||
|
Time: time.Now(),
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
return nil, status.Errorf(codes.Unauthenticated, err.Error())
|
||||||
|
}
|
||||||
|
|
||||||
|
err = endpoint.validateBucket(ctx, req.Name)
|
||||||
|
if err != nil {
|
||||||
|
return nil, status.Errorf(codes.InvalidArgument, err.Error())
|
||||||
|
}
|
||||||
|
|
||||||
|
err = endpoint.metainfo.DeleteBucket(ctx, req.Name, keyInfo.ProjectID)
|
||||||
|
if err != nil {
|
||||||
|
return nil, status.Errorf(codes.Internal, err.Error())
|
||||||
|
}
|
||||||
|
|
||||||
|
return &pb.BucketDeleteResponse{}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// ListBuckets returns a list of buckets
|
// ListBuckets returns buckets in a project where the bucket name matches the request cursor
|
||||||
func (endpoint *Endpoint) ListBuckets(ctx context.Context, req *pb.BucketListRequest) (_ *pb.BucketListResponse, err error) {
|
func (endpoint *Endpoint) ListBuckets(ctx context.Context, req *pb.BucketListRequest) (resp *pb.BucketListResponse, err error) {
|
||||||
defer mon.Task()(&ctx)(&err)
|
defer mon.Task()(&ctx)(&err)
|
||||||
// TODO: placeholder to implement pb.MetainfoServer interface.
|
action := macaroon.Action{
|
||||||
return &pb.BucketListResponse{}, err
|
Op: macaroon.ActionRead,
|
||||||
|
Time: time.Now(),
|
||||||
|
}
|
||||||
|
keyInfo, err := endpoint.validateAuth(ctx, action)
|
||||||
|
if err != nil {
|
||||||
|
return nil, status.Errorf(codes.Unauthenticated, err.Error())
|
||||||
|
}
|
||||||
|
|
||||||
|
allowedBuckets, err := getAllowedBuckets(ctx, action)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
listOpts := storj.BucketListOptions{
|
||||||
|
Cursor: string(req.Cursor),
|
||||||
|
Limit: int(req.Limit),
|
||||||
|
// We are only supporting the forward direction for listing buckets
|
||||||
|
Direction: storj.Forward,
|
||||||
|
}
|
||||||
|
bucketList, err := endpoint.metainfo.ListBuckets(ctx, keyInfo.ProjectID, listOpts, allowedBuckets)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
bucketItems := make([]*pb.BucketListItem, len(bucketList.Items))
|
||||||
|
for i, item := range bucketList.Items {
|
||||||
|
bucketItems[i] = &pb.BucketListItem{
|
||||||
|
Name: []byte(item.Name),
|
||||||
|
CreatedAt: item.Created,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return &pb.BucketListResponse{
|
||||||
|
Items: bucketItems,
|
||||||
|
More: bucketList.More,
|
||||||
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// SetBucketAttribution returns a list of buckets
|
func getAllowedBuckets(ctx context.Context, action macaroon.Action) (allowedBuckets map[string]struct{}, err error) {
|
||||||
func (endpoint *Endpoint) SetBucketAttribution(ctx context.Context, req *pb.BucketSetAttributionRequest) (_ *pb.BucketSetAttributionResponse, err error) {
|
keyData, ok := auth.GetAPIKey(ctx)
|
||||||
defer mon.Task()(&ctx)(&err)
|
if !ok {
|
||||||
// TODO: placeholder to implement pb.MetainfoServer interface.
|
return nil, status.Errorf(codes.Unauthenticated, "Invalid API credential GetAPIKey: %v", err)
|
||||||
return &pb.BucketSetAttributionResponse{}, err
|
}
|
||||||
|
key, err := macaroon.ParseAPIKey(string(keyData))
|
||||||
|
if err != nil {
|
||||||
|
return nil, status.Errorf(codes.Unauthenticated, "Invalid API credential ParseAPIKey: %v", err)
|
||||||
|
}
|
||||||
|
allowedBuckets, err = key.GetAllowedBuckets(ctx, action)
|
||||||
|
if err != nil {
|
||||||
|
return nil, status.Errorf(codes.Internal, "GetAllowedBuckets: %v", err)
|
||||||
|
}
|
||||||
|
return allowedBuckets, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// SetBucketAttribution sets the bucket attribution.
|
||||||
|
func (endpoint *Endpoint) SetBucketAttribution(context.Context, *pb.BucketSetAttributionRequest) (resp *pb.BucketSetAttributionResponse, err error) {
|
||||||
|
return resp, status.Error(codes.Unimplemented, "not implemented")
|
||||||
|
}
|
||||||
|
|
||||||
|
func convertProtoToBucket(req *pb.BucketCreateRequest, projectID uuid.UUID) (storj.Bucket, error) {
|
||||||
|
bucketID, err := uuid.New()
|
||||||
|
if err != nil {
|
||||||
|
return storj.Bucket{}, err
|
||||||
|
}
|
||||||
|
|
||||||
|
defaultRS := req.GetDefaultRedundancyScheme()
|
||||||
|
defaultEP := req.GetDefaultEncryptionParameters()
|
||||||
|
return storj.Bucket{
|
||||||
|
ID: *bucketID,
|
||||||
|
Name: string(req.GetName()),
|
||||||
|
ProjectID: projectID,
|
||||||
|
Attribution: string(req.GetAttributionId()),
|
||||||
|
PathCipher: storj.CipherSuite(req.GetPathCipher()),
|
||||||
|
DefaultSegmentsSize: req.GetDefaultSegmentSize(),
|
||||||
|
DefaultRedundancyScheme: storj.RedundancyScheme{
|
||||||
|
Algorithm: storj.RedundancyAlgorithm(defaultRS.GetType()),
|
||||||
|
ShareSize: defaultRS.GetErasureShareSize(),
|
||||||
|
RequiredShares: int16(defaultRS.GetMinReq()),
|
||||||
|
RepairShares: int16(defaultRS.GetRepairThreshold()),
|
||||||
|
OptimalShares: int16(defaultRS.GetSuccessThreshold()),
|
||||||
|
TotalShares: int16(defaultRS.GetTotal()),
|
||||||
|
},
|
||||||
|
DefaultEncryptionParameters: storj.EncryptionParameters{
|
||||||
|
CipherSuite: storj.CipherSuite(defaultEP.CipherSuite),
|
||||||
|
BlockSize: int32(defaultEP.BlockSize),
|
||||||
|
},
|
||||||
|
}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func convertBucketToProto(ctx context.Context, bucket storj.Bucket) (pbBucket *pb.Bucket) {
|
||||||
|
rs := bucket.DefaultRedundancyScheme
|
||||||
|
return &pb.Bucket{
|
||||||
|
Name: []byte(bucket.Name),
|
||||||
|
PathCipher: pb.CipherSuite(int(bucket.PathCipher)),
|
||||||
|
AttributionId: []byte(bucket.Attribution),
|
||||||
|
CreatedAt: bucket.Created,
|
||||||
|
DefaultSegmentSize: bucket.DefaultSegmentsSize,
|
||||||
|
DefaultRedundancyScheme: &pb.RedundancyScheme{
|
||||||
|
Type: pb.RedundancyScheme_RS,
|
||||||
|
MinReq: int32(rs.RequiredShares),
|
||||||
|
Total: int32(rs.TotalShares),
|
||||||
|
RepairThreshold: int32(rs.RepairShares),
|
||||||
|
SuccessThreshold: int32(rs.OptimalShares),
|
||||||
|
ErasureShareSize: rs.ShareSize,
|
||||||
|
},
|
||||||
|
DefaultEncryptionParameters: &pb.EncryptionParameters{
|
||||||
|
CipherSuite: pb.CipherSuite(int(bucket.DefaultEncryptionParameters.CipherSuite)),
|
||||||
|
BlockSize: int64(bucket.DefaultEncryptionParameters.BlockSize),
|
||||||
|
},
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
@ -8,23 +8,26 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/gogo/protobuf/proto"
|
"github.com/gogo/protobuf/proto"
|
||||||
|
"github.com/skyrings/skyring-common/tools/uuid"
|
||||||
"github.com/zeebo/errs"
|
"github.com/zeebo/errs"
|
||||||
"go.uber.org/zap"
|
"go.uber.org/zap"
|
||||||
|
|
||||||
"storj.io/storj/pkg/pb"
|
"storj.io/storj/pkg/pb"
|
||||||
"storj.io/storj/pkg/storage/meta"
|
"storj.io/storj/pkg/storage/meta"
|
||||||
|
"storj.io/storj/pkg/storj"
|
||||||
"storj.io/storj/storage"
|
"storj.io/storj/storage"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Service structure
|
// Service structure
|
||||||
type Service struct {
|
type Service struct {
|
||||||
logger *zap.Logger
|
logger *zap.Logger
|
||||||
DB storage.KeyValueStore
|
DB storage.KeyValueStore
|
||||||
|
bucketsDB BucketsDB
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewService creates new metainfo service
|
// NewService creates new metainfo service
|
||||||
func NewService(logger *zap.Logger, db storage.KeyValueStore) *Service {
|
func NewService(logger *zap.Logger, db storage.KeyValueStore, bucketsDB BucketsDB) *Service {
|
||||||
return &Service{logger: logger, DB: db}
|
return &Service{logger: logger, DB: db, bucketsDB: bucketsDB}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Put puts pointer to db under specific path
|
// Put puts pointer to db under specific path
|
||||||
@ -165,3 +168,27 @@ func (s *Service) Iterate(ctx context.Context, prefix string, first string, recu
|
|||||||
}
|
}
|
||||||
return s.DB.Iterate(ctx, opts, f)
|
return s.DB.Iterate(ctx, opts, f)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// CreateBucket creates a new bucket in the buckets db
|
||||||
|
func (s *Service) CreateBucket(ctx context.Context, bucket storj.Bucket) (_ storj.Bucket, err error) {
|
||||||
|
defer mon.Task()(&ctx)(&err)
|
||||||
|
return s.bucketsDB.CreateBucket(ctx, bucket)
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetBucket returns an existing bucket in the buckets db
|
||||||
|
func (s *Service) GetBucket(ctx context.Context, bucketName []byte, projectID uuid.UUID) (_ storj.Bucket, err error) {
|
||||||
|
defer mon.Task()(&ctx)(&err)
|
||||||
|
return s.bucketsDB.GetBucket(ctx, bucketName, projectID)
|
||||||
|
}
|
||||||
|
|
||||||
|
// DeleteBucket deletes a bucket from the bucekts db
|
||||||
|
func (s *Service) DeleteBucket(ctx context.Context, bucketName []byte, projectID uuid.UUID) (err error) {
|
||||||
|
defer mon.Task()(&ctx)(&err)
|
||||||
|
return s.bucketsDB.DeleteBucket(ctx, bucketName, projectID)
|
||||||
|
}
|
||||||
|
|
||||||
|
// ListBuckets returns a list of buckets for a project
|
||||||
|
func (s *Service) ListBuckets(ctx context.Context, projectID uuid.UUID, listOpts storj.BucketListOptions, allowedBuckets map[string]struct{}) (bucketList storj.BucketList, err error) {
|
||||||
|
defer mon.Task()(&ctx)(&err)
|
||||||
|
return s.bucketsDB.ListBuckets(ctx, projectID, listOpts, allowedBuckets)
|
||||||
|
}
|
||||||
|
@ -98,6 +98,8 @@ type DB interface {
|
|||||||
Orders() orders.DB
|
Orders() orders.DB
|
||||||
// Containment returns database for containment
|
// Containment returns database for containment
|
||||||
Containment() audit.Containment
|
Containment() audit.Containment
|
||||||
|
// Buckets returns the database to interact with buckets
|
||||||
|
Buckets() metainfo.BucketsDB
|
||||||
}
|
}
|
||||||
|
|
||||||
// Config is the global config satellite
|
// Config is the global config satellite
|
||||||
@ -401,7 +403,10 @@ func New(log *zap.Logger, full *identity.FullIdentity, db DB, config *Config, ve
|
|||||||
}
|
}
|
||||||
|
|
||||||
peer.Metainfo.Database = db // for logging: storelogger.New(peer.Log.Named("pdb"), db)
|
peer.Metainfo.Database = db // for logging: storelogger.New(peer.Log.Named("pdb"), db)
|
||||||
peer.Metainfo.Service = metainfo.NewService(peer.Log.Named("metainfo:service"), peer.Metainfo.Database)
|
peer.Metainfo.Service = metainfo.NewService(peer.Log.Named("metainfo:service"),
|
||||||
|
peer.Metainfo.Database,
|
||||||
|
peer.DB.Buckets(),
|
||||||
|
)
|
||||||
|
|
||||||
peer.Metainfo.Endpoint2 = metainfo.NewEndpoint(
|
peer.Metainfo.Endpoint2 = metainfo.NewEndpoint(
|
||||||
peer.Log.Named("metainfo:endpoint"),
|
peer.Log.Named("metainfo:endpoint"),
|
||||||
|
176
satellite/satellitedb/buckets.go
Normal file
176
satellite/satellitedb/buckets.go
Normal file
@ -0,0 +1,176 @@
|
|||||||
|
// Copyright (C) 2019 Storj Labs, Inc.
|
||||||
|
// See LICENSE for copying information.
|
||||||
|
|
||||||
|
package satellitedb
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"errors"
|
||||||
|
|
||||||
|
"github.com/skyrings/skyring-common/tools/uuid"
|
||||||
|
|
||||||
|
"storj.io/storj/pkg/storj"
|
||||||
|
"storj.io/storj/satellite/metainfo"
|
||||||
|
dbx "storj.io/storj/satellite/satellitedb/dbx"
|
||||||
|
)
|
||||||
|
|
||||||
|
type bucketsDB struct {
|
||||||
|
db dbx.Methods
|
||||||
|
}
|
||||||
|
|
||||||
|
// Buckets returns database for interacting with buckets
|
||||||
|
func (db *DB) Buckets() metainfo.BucketsDB {
|
||||||
|
return &bucketsDB{db: db.db}
|
||||||
|
}
|
||||||
|
|
||||||
|
// CreateBucket creates a new bucket
|
||||||
|
func (db *bucketsDB) CreateBucket(ctx context.Context, bucket storj.Bucket) (_ storj.Bucket, err error) {
|
||||||
|
defer mon.Task()(&ctx)(&err)
|
||||||
|
row, err := db.db.Create_BucketMetainfo(ctx,
|
||||||
|
dbx.BucketMetainfo_Id(bucket.ID[:]),
|
||||||
|
dbx.BucketMetainfo_ProjectId(bucket.ProjectID[:]),
|
||||||
|
dbx.BucketMetainfo_Name([]byte(bucket.Name)),
|
||||||
|
dbx.BucketMetainfo_PathCipher(int(bucket.PathCipher)),
|
||||||
|
dbx.BucketMetainfo_DefaultSegmentSize(int(bucket.DefaultSegmentsSize)),
|
||||||
|
dbx.BucketMetainfo_DefaultEncryptionCipherSuite(int(bucket.DefaultEncryptionParameters.CipherSuite)),
|
||||||
|
dbx.BucketMetainfo_DefaultEncryptionBlockSize(int(bucket.DefaultEncryptionParameters.BlockSize)),
|
||||||
|
dbx.BucketMetainfo_DefaultRedundancyAlgorithm(int(bucket.DefaultRedundancyScheme.Algorithm)),
|
||||||
|
dbx.BucketMetainfo_DefaultRedundancyShareSize(int(bucket.DefaultRedundancyScheme.ShareSize)),
|
||||||
|
dbx.BucketMetainfo_DefaultRedundancyRequiredShares(int(bucket.DefaultRedundancyScheme.RequiredShares)),
|
||||||
|
dbx.BucketMetainfo_DefaultRedundancyRepairShares(int(bucket.DefaultRedundancyScheme.RepairShares)),
|
||||||
|
dbx.BucketMetainfo_DefaultRedundancyOptimalShares(int(bucket.DefaultRedundancyScheme.OptimalShares)),
|
||||||
|
dbx.BucketMetainfo_DefaultRedundancyTotalShares(int(bucket.DefaultRedundancyScheme.TotalShares)),
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
return storj.Bucket{}, storj.ErrBucket.Wrap(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
bucket, err = convertDBXtoBucket(row)
|
||||||
|
if err != nil {
|
||||||
|
return storj.Bucket{}, storj.ErrBucket.Wrap(err)
|
||||||
|
}
|
||||||
|
return bucket, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetBucket returns a bucket
|
||||||
|
func (db *bucketsDB) GetBucket(ctx context.Context, bucketName []byte, projectID uuid.UUID) (bucket storj.Bucket, err error) {
|
||||||
|
defer mon.Task()(&ctx)(&err)
|
||||||
|
dbxBucket, err := db.db.Get_BucketMetainfo_By_ProjectId_And_Name(ctx,
|
||||||
|
dbx.BucketMetainfo_ProjectId(projectID[:]),
|
||||||
|
dbx.BucketMetainfo_Name(bucketName),
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
return bucket, err
|
||||||
|
}
|
||||||
|
return convertDBXtoBucket(dbxBucket)
|
||||||
|
}
|
||||||
|
|
||||||
|
// DeleteBucket deletes a bucket
|
||||||
|
func (db *bucketsDB) DeleteBucket(ctx context.Context, bucketName []byte, projectID uuid.UUID) (err error) {
|
||||||
|
defer mon.Task()(&ctx)(&err)
|
||||||
|
_, err = db.db.Delete_BucketMetainfo_By_ProjectId_And_Name(ctx,
|
||||||
|
dbx.BucketMetainfo_ProjectId(projectID[:]),
|
||||||
|
dbx.BucketMetainfo_Name(bucketName),
|
||||||
|
)
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
// ListBuckets returns a list of buckets for a project
|
||||||
|
func (db *bucketsDB) ListBuckets(ctx context.Context, projectID uuid.UUID, listOpts storj.BucketListOptions, allowedBuckets map[string]struct{}) (bucketList storj.BucketList, err error) {
|
||||||
|
defer mon.Task()(&ctx)(&err)
|
||||||
|
|
||||||
|
const defaultListLimit = 10000
|
||||||
|
if listOpts.Limit < 1 {
|
||||||
|
listOpts.Limit = defaultListLimit
|
||||||
|
}
|
||||||
|
limit := listOpts.Limit + 1 // add one to detect More
|
||||||
|
|
||||||
|
for {
|
||||||
|
var dbxBuckets []*dbx.BucketMetainfo
|
||||||
|
switch listOpts.Direction {
|
||||||
|
// for listing buckets we are only supporting the forward direction for simplicity
|
||||||
|
case storj.Forward:
|
||||||
|
dbxBuckets, err = db.db.Limited_BucketMetainfo_By_ProjectId_And_Name_GreaterOrEqual_OrderBy_Asc_Name(ctx,
|
||||||
|
dbx.BucketMetainfo_ProjectId(projectID[:]),
|
||||||
|
dbx.BucketMetainfo_Name([]byte(listOpts.Cursor)),
|
||||||
|
limit,
|
||||||
|
0,
|
||||||
|
)
|
||||||
|
default:
|
||||||
|
return bucketList, errors.New("unknown list direction")
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
return bucketList, err
|
||||||
|
}
|
||||||
|
|
||||||
|
bucketList.More = len(dbxBuckets) > listOpts.Limit
|
||||||
|
var nextCursor string
|
||||||
|
if bucketList.More {
|
||||||
|
nextCursor = string(dbxBuckets[listOpts.Limit].Name)
|
||||||
|
// If there are more buckets than listOpts.limit returned,
|
||||||
|
// then remove the extra buckets so that we do not return
|
||||||
|
// more then the limit
|
||||||
|
dbxBuckets = dbxBuckets[0:listOpts.Limit]
|
||||||
|
}
|
||||||
|
|
||||||
|
if bucketList.Items == nil {
|
||||||
|
bucketList.Items = make([]storj.Bucket, 0, len(dbxBuckets))
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, dbxBucket := range dbxBuckets {
|
||||||
|
// Check that the bucket is allowed to be viewed
|
||||||
|
if _, ok := allowedBuckets[string(dbxBucket.Name)]; ok {
|
||||||
|
item, err := convertDBXtoBucket(dbxBucket)
|
||||||
|
if err != nil {
|
||||||
|
return bucketList, err
|
||||||
|
}
|
||||||
|
bucketList.Items = append(bucketList.Items, item)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(bucketList.Items) < listOpts.Limit && bucketList.More {
|
||||||
|
// If we filtered out disallowed buckets, then get more buckets
|
||||||
|
// out of database so that we return `limit` number of buckets
|
||||||
|
listOpts = storj.BucketListOptions{
|
||||||
|
Cursor: nextCursor,
|
||||||
|
Limit: listOpts.Limit,
|
||||||
|
Direction: storj.Forward,
|
||||||
|
}
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
return bucketList, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func convertDBXtoBucket(dbxBucket *dbx.BucketMetainfo) (bucket storj.Bucket, err error) {
|
||||||
|
id, err := bytesToUUID(dbxBucket.Id)
|
||||||
|
if err != nil {
|
||||||
|
return bucket, err
|
||||||
|
}
|
||||||
|
project, err := bytesToUUID(dbxBucket.ProjectId)
|
||||||
|
if err != nil {
|
||||||
|
return bucket, err
|
||||||
|
}
|
||||||
|
return storj.Bucket{
|
||||||
|
ID: id,
|
||||||
|
Name: string(dbxBucket.Name),
|
||||||
|
ProjectID: project,
|
||||||
|
Created: dbxBucket.CreatedAt,
|
||||||
|
PathCipher: storj.CipherSuite(dbxBucket.PathCipher),
|
||||||
|
DefaultSegmentsSize: int64(dbxBucket.DefaultSegmentSize),
|
||||||
|
DefaultRedundancyScheme: storj.RedundancyScheme{
|
||||||
|
Algorithm: storj.RedundancyAlgorithm(dbxBucket.DefaultRedundancyAlgorithm),
|
||||||
|
ShareSize: int32(dbxBucket.DefaultRedundancyShareSize),
|
||||||
|
RequiredShares: int16(dbxBucket.DefaultRedundancyRequiredShares),
|
||||||
|
RepairShares: int16(dbxBucket.DefaultRedundancyRepairShares),
|
||||||
|
OptimalShares: int16(dbxBucket.DefaultRedundancyOptimalShares),
|
||||||
|
TotalShares: int16(dbxBucket.DefaultRedundancyTotalShares),
|
||||||
|
},
|
||||||
|
DefaultEncryptionParameters: storj.EncryptionParameters{
|
||||||
|
CipherSuite: storj.CipherSuite(dbxBucket.DefaultEncryptionCipherSuite),
|
||||||
|
BlockSize: int32(dbxBucket.DefaultEncryptionBlockSize),
|
||||||
|
},
|
||||||
|
}, err
|
||||||
|
}
|
@ -770,3 +770,10 @@ delete bucket_metainfo (
|
|||||||
where bucket_metainfo.project_id = ?
|
where bucket_metainfo.project_id = ?
|
||||||
where bucket_metainfo.name = ?
|
where bucket_metainfo.name = ?
|
||||||
)
|
)
|
||||||
|
|
||||||
|
read limitoffset ( // Forward
|
||||||
|
select bucket_metainfo
|
||||||
|
where bucket_metainfo.project_id = ?
|
||||||
|
where bucket_metainfo.name >= ?
|
||||||
|
orderby asc bucket_metainfo.name
|
||||||
|
)
|
||||||
|
@ -7330,6 +7330,43 @@ func (obj *postgresImpl) Get_BucketMetainfo_By_ProjectId_And_Name(ctx context.Co
|
|||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (obj *postgresImpl) Limited_BucketMetainfo_By_ProjectId_And_Name_GreaterOrEqual_OrderBy_Asc_Name(ctx context.Context,
|
||||||
|
bucket_metainfo_project_id BucketMetainfo_ProjectId_Field,
|
||||||
|
bucket_metainfo_name_greater_or_equal BucketMetainfo_Name_Field,
|
||||||
|
limit int, offset int64) (
|
||||||
|
rows []*BucketMetainfo, err error) {
|
||||||
|
|
||||||
|
var __embed_stmt = __sqlbundle_Literal("SELECT bucket_metainfos.id, bucket_metainfos.project_id, bucket_metainfos.name, bucket_metainfos.path_cipher, bucket_metainfos.created_at, bucket_metainfos.default_segment_size, bucket_metainfos.default_encryption_cipher_suite, bucket_metainfos.default_encryption_block_size, bucket_metainfos.default_redundancy_algorithm, bucket_metainfos.default_redundancy_share_size, bucket_metainfos.default_redundancy_required_shares, bucket_metainfos.default_redundancy_repair_shares, bucket_metainfos.default_redundancy_optimal_shares, bucket_metainfos.default_redundancy_total_shares FROM bucket_metainfos WHERE bucket_metainfos.project_id = ? AND bucket_metainfos.name >= ? ORDER BY bucket_metainfos.name LIMIT ? OFFSET ?")
|
||||||
|
|
||||||
|
var __values []interface{}
|
||||||
|
__values = append(__values, bucket_metainfo_project_id.value(), bucket_metainfo_name_greater_or_equal.value())
|
||||||
|
|
||||||
|
__values = append(__values, limit, offset)
|
||||||
|
|
||||||
|
var __stmt = __sqlbundle_Render(obj.dialect, __embed_stmt)
|
||||||
|
obj.logStmt(__stmt, __values...)
|
||||||
|
|
||||||
|
__rows, err := obj.driver.Query(__stmt, __values...)
|
||||||
|
if err != nil {
|
||||||
|
return nil, obj.makeErr(err)
|
||||||
|
}
|
||||||
|
defer __rows.Close()
|
||||||
|
|
||||||
|
for __rows.Next() {
|
||||||
|
bucket_metainfo := &BucketMetainfo{}
|
||||||
|
err = __rows.Scan(&bucket_metainfo.Id, &bucket_metainfo.ProjectId, &bucket_metainfo.Name, &bucket_metainfo.PathCipher, &bucket_metainfo.CreatedAt, &bucket_metainfo.DefaultSegmentSize, &bucket_metainfo.DefaultEncryptionCipherSuite, &bucket_metainfo.DefaultEncryptionBlockSize, &bucket_metainfo.DefaultRedundancyAlgorithm, &bucket_metainfo.DefaultRedundancyShareSize, &bucket_metainfo.DefaultRedundancyRequiredShares, &bucket_metainfo.DefaultRedundancyRepairShares, &bucket_metainfo.DefaultRedundancyOptimalShares, &bucket_metainfo.DefaultRedundancyTotalShares)
|
||||||
|
if err != nil {
|
||||||
|
return nil, obj.makeErr(err)
|
||||||
|
}
|
||||||
|
rows = append(rows, bucket_metainfo)
|
||||||
|
}
|
||||||
|
if err := __rows.Err(); err != nil {
|
||||||
|
return nil, obj.makeErr(err)
|
||||||
|
}
|
||||||
|
return rows, nil
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
func (obj *postgresImpl) Update_PendingAudits_By_NodeId(ctx context.Context,
|
func (obj *postgresImpl) Update_PendingAudits_By_NodeId(ctx context.Context,
|
||||||
pending_audits_node_id PendingAudits_NodeId_Field,
|
pending_audits_node_id PendingAudits_NodeId_Field,
|
||||||
update PendingAudits_Update_Fields) (
|
update PendingAudits_Update_Fields) (
|
||||||
@ -10825,6 +10862,43 @@ func (obj *sqlite3Impl) Get_BucketMetainfo_By_ProjectId_And_Name(ctx context.Con
|
|||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (obj *sqlite3Impl) Limited_BucketMetainfo_By_ProjectId_And_Name_GreaterOrEqual_OrderBy_Asc_Name(ctx context.Context,
|
||||||
|
bucket_metainfo_project_id BucketMetainfo_ProjectId_Field,
|
||||||
|
bucket_metainfo_name_greater_or_equal BucketMetainfo_Name_Field,
|
||||||
|
limit int, offset int64) (
|
||||||
|
rows []*BucketMetainfo, err error) {
|
||||||
|
|
||||||
|
var __embed_stmt = __sqlbundle_Literal("SELECT bucket_metainfos.id, bucket_metainfos.project_id, bucket_metainfos.name, bucket_metainfos.path_cipher, bucket_metainfos.created_at, bucket_metainfos.default_segment_size, bucket_metainfos.default_encryption_cipher_suite, bucket_metainfos.default_encryption_block_size, bucket_metainfos.default_redundancy_algorithm, bucket_metainfos.default_redundancy_share_size, bucket_metainfos.default_redundancy_required_shares, bucket_metainfos.default_redundancy_repair_shares, bucket_metainfos.default_redundancy_optimal_shares, bucket_metainfos.default_redundancy_total_shares FROM bucket_metainfos WHERE bucket_metainfos.project_id = ? AND bucket_metainfos.name >= ? ORDER BY bucket_metainfos.name LIMIT ? OFFSET ?")
|
||||||
|
|
||||||
|
var __values []interface{}
|
||||||
|
__values = append(__values, bucket_metainfo_project_id.value(), bucket_metainfo_name_greater_or_equal.value())
|
||||||
|
|
||||||
|
__values = append(__values, limit, offset)
|
||||||
|
|
||||||
|
var __stmt = __sqlbundle_Render(obj.dialect, __embed_stmt)
|
||||||
|
obj.logStmt(__stmt, __values...)
|
||||||
|
|
||||||
|
__rows, err := obj.driver.Query(__stmt, __values...)
|
||||||
|
if err != nil {
|
||||||
|
return nil, obj.makeErr(err)
|
||||||
|
}
|
||||||
|
defer __rows.Close()
|
||||||
|
|
||||||
|
for __rows.Next() {
|
||||||
|
bucket_metainfo := &BucketMetainfo{}
|
||||||
|
err = __rows.Scan(&bucket_metainfo.Id, &bucket_metainfo.ProjectId, &bucket_metainfo.Name, &bucket_metainfo.PathCipher, &bucket_metainfo.CreatedAt, &bucket_metainfo.DefaultSegmentSize, &bucket_metainfo.DefaultEncryptionCipherSuite, &bucket_metainfo.DefaultEncryptionBlockSize, &bucket_metainfo.DefaultRedundancyAlgorithm, &bucket_metainfo.DefaultRedundancyShareSize, &bucket_metainfo.DefaultRedundancyRequiredShares, &bucket_metainfo.DefaultRedundancyRepairShares, &bucket_metainfo.DefaultRedundancyOptimalShares, &bucket_metainfo.DefaultRedundancyTotalShares)
|
||||||
|
if err != nil {
|
||||||
|
return nil, obj.makeErr(err)
|
||||||
|
}
|
||||||
|
rows = append(rows, bucket_metainfo)
|
||||||
|
}
|
||||||
|
if err := __rows.Err(); err != nil {
|
||||||
|
return nil, obj.makeErr(err)
|
||||||
|
}
|
||||||
|
return rows, nil
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
func (obj *sqlite3Impl) Update_PendingAudits_By_NodeId(ctx context.Context,
|
func (obj *sqlite3Impl) Update_PendingAudits_By_NodeId(ctx context.Context,
|
||||||
pending_audits_node_id PendingAudits_NodeId_Field,
|
pending_audits_node_id PendingAudits_NodeId_Field,
|
||||||
update PendingAudits_Update_Fields) (
|
update PendingAudits_Update_Fields) (
|
||||||
@ -13676,6 +13750,18 @@ func (rx *Rx) Get_ValueAttribution_By_ProjectId_And_BucketName(ctx context.Conte
|
|||||||
return tx.Get_ValueAttribution_By_ProjectId_And_BucketName(ctx, value_attribution_project_id, value_attribution_bucket_name)
|
return tx.Get_ValueAttribution_By_ProjectId_And_BucketName(ctx, value_attribution_project_id, value_attribution_bucket_name)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (rx *Rx) Limited_BucketMetainfo_By_ProjectId_And_Name_GreaterOrEqual_OrderBy_Asc_Name(ctx context.Context,
|
||||||
|
bucket_metainfo_project_id BucketMetainfo_ProjectId_Field,
|
||||||
|
bucket_metainfo_name_greater_or_equal BucketMetainfo_Name_Field,
|
||||||
|
limit int, offset int64) (
|
||||||
|
rows []*BucketMetainfo, err error) {
|
||||||
|
var tx *Tx
|
||||||
|
if tx, err = rx.getTx(ctx); err != nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
return tx.Limited_BucketMetainfo_By_ProjectId_And_Name_GreaterOrEqual_OrderBy_Asc_Name(ctx, bucket_metainfo_project_id, bucket_metainfo_name_greater_or_equal, limit, offset)
|
||||||
|
}
|
||||||
|
|
||||||
func (rx *Rx) Limited_BucketUsage_By_BucketId_And_RollupEndTime_Greater_And_RollupEndTime_LessOrEqual_OrderBy_Asc_RollupEndTime(ctx context.Context,
|
func (rx *Rx) Limited_BucketUsage_By_BucketId_And_RollupEndTime_Greater_And_RollupEndTime_LessOrEqual_OrderBy_Asc_RollupEndTime(ctx context.Context,
|
||||||
bucket_usage_bucket_id BucketUsage_BucketId_Field,
|
bucket_usage_bucket_id BucketUsage_BucketId_Field,
|
||||||
bucket_usage_rollup_end_time_greater BucketUsage_RollupEndTime_Field,
|
bucket_usage_rollup_end_time_greater BucketUsage_RollupEndTime_Field,
|
||||||
@ -14302,6 +14388,12 @@ type Methods interface {
|
|||||||
value_attribution_bucket_name ValueAttribution_BucketName_Field) (
|
value_attribution_bucket_name ValueAttribution_BucketName_Field) (
|
||||||
value_attribution *ValueAttribution, err error)
|
value_attribution *ValueAttribution, err error)
|
||||||
|
|
||||||
|
Limited_BucketMetainfo_By_ProjectId_And_Name_GreaterOrEqual_OrderBy_Asc_Name(ctx context.Context,
|
||||||
|
bucket_metainfo_project_id BucketMetainfo_ProjectId_Field,
|
||||||
|
bucket_metainfo_name_greater_or_equal BucketMetainfo_Name_Field,
|
||||||
|
limit int, offset int64) (
|
||||||
|
rows []*BucketMetainfo, err error)
|
||||||
|
|
||||||
Limited_BucketUsage_By_BucketId_And_RollupEndTime_Greater_And_RollupEndTime_LessOrEqual_OrderBy_Asc_RollupEndTime(ctx context.Context,
|
Limited_BucketUsage_By_BucketId_And_RollupEndTime_Greater_And_RollupEndTime_LessOrEqual_OrderBy_Asc_RollupEndTime(ctx context.Context,
|
||||||
bucket_usage_bucket_id BucketUsage_BucketId_Field,
|
bucket_usage_bucket_id BucketUsage_BucketId_Field,
|
||||||
bucket_usage_rollup_end_time_greater BucketUsage_RollupEndTime_Field,
|
bucket_usage_rollup_end_time_greater BucketUsage_RollupEndTime_Field,
|
||||||
|
@ -25,6 +25,7 @@ import (
|
|||||||
"storj.io/storj/satellite"
|
"storj.io/storj/satellite"
|
||||||
"storj.io/storj/satellite/attribution"
|
"storj.io/storj/satellite/attribution"
|
||||||
"storj.io/storj/satellite/console"
|
"storj.io/storj/satellite/console"
|
||||||
|
"storj.io/storj/satellite/metainfo"
|
||||||
"storj.io/storj/satellite/orders"
|
"storj.io/storj/satellite/orders"
|
||||||
"storj.io/storj/satellite/rewards"
|
"storj.io/storj/satellite/rewards"
|
||||||
)
|
)
|
||||||
@ -74,6 +75,47 @@ func (m *lockedAttribution) QueryAttribution(ctx context.Context, partnerID uuid
|
|||||||
return m.db.QueryAttribution(ctx, partnerID, start, end)
|
return m.db.QueryAttribution(ctx, partnerID, start, end)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Buckets returns the database to interact with buckets
|
||||||
|
func (m *locked) Buckets() metainfo.BucketsDB {
|
||||||
|
m.Lock()
|
||||||
|
defer m.Unlock()
|
||||||
|
return &lockedBuckets{m.Locker, m.db.Buckets()}
|
||||||
|
}
|
||||||
|
|
||||||
|
// lockedBuckets implements locking wrapper for metainfo.BucketsDB
|
||||||
|
type lockedBuckets struct {
|
||||||
|
sync.Locker
|
||||||
|
db metainfo.BucketsDB
|
||||||
|
}
|
||||||
|
|
||||||
|
// Create creates a new bucket
|
||||||
|
func (m *lockedBuckets) CreateBucket(ctx context.Context, bucket storj.Bucket) (_ storj.Bucket, err error) {
|
||||||
|
m.Lock()
|
||||||
|
defer m.Unlock()
|
||||||
|
return m.db.CreateBucket(ctx, bucket)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Delete deletes a bucket
|
||||||
|
func (m *lockedBuckets) DeleteBucket(ctx context.Context, bucketName []byte, projectID uuid.UUID) (err error) {
|
||||||
|
m.Lock()
|
||||||
|
defer m.Unlock()
|
||||||
|
return m.db.DeleteBucket(ctx, bucketName, projectID)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Get returns an existing bucket
|
||||||
|
func (m *lockedBuckets) GetBucket(ctx context.Context, bucketName []byte, projectID uuid.UUID) (bucket storj.Bucket, err error) {
|
||||||
|
m.Lock()
|
||||||
|
defer m.Unlock()
|
||||||
|
return m.db.GetBucket(ctx, bucketName, projectID)
|
||||||
|
}
|
||||||
|
|
||||||
|
// List returns all buckets for a project
|
||||||
|
func (m *lockedBuckets) ListBuckets(ctx context.Context, projectID uuid.UUID, listOpts storj.BucketListOptions, allowedBuckets map[string]struct{}) (bucketList storj.BucketList, err error) {
|
||||||
|
m.Lock()
|
||||||
|
defer m.Unlock()
|
||||||
|
return m.db.ListBuckets(ctx, projectID, listOpts, allowedBuckets)
|
||||||
|
}
|
||||||
|
|
||||||
// CertDB returns database for storing uplink's public key & ID
|
// CertDB returns database for storing uplink's public key & ID
|
||||||
func (m *locked) CertDB() certdb.DB {
|
func (m *locked) CertDB() certdb.DB {
|
||||||
m.Lock()
|
m.Lock()
|
||||||
|
Loading…
Reference in New Issue
Block a user