2019-01-18 15:10:21 +00:00
|
|
|
// Copyright (C) 2019 Storj Labs, Inc.
|
|
|
|
// See LICENSE for copying information.
|
|
|
|
|
2019-04-25 09:46:32 +01:00
|
|
|
package metainfo
|
2019-01-18 15:10:21 +00:00
|
|
|
|
|
|
|
import (
|
2019-06-04 12:55:38 +01:00
|
|
|
"context"
|
2019-07-08 23:16:50 +01:00
|
|
|
"time"
|
2019-06-04 12:55:38 +01:00
|
|
|
|
2019-01-18 15:10:21 +00:00
|
|
|
"github.com/gogo/protobuf/proto"
|
2019-07-08 23:32:18 +01:00
|
|
|
"github.com/skyrings/skyring-common/tools/uuid"
|
2019-01-18 15:10:21 +00:00
|
|
|
"go.uber.org/zap"
|
|
|
|
|
2019-12-27 11:48:47 +00:00
|
|
|
"storj.io/common/macaroon"
|
|
|
|
"storj.io/common/pb"
|
|
|
|
"storj.io/common/storj"
|
2019-01-18 15:10:21 +00:00
|
|
|
"storj.io/storj/storage"
|
2020-01-08 13:40:19 +00:00
|
|
|
"storj.io/uplink/storage/meta"
|
2019-01-18 15:10:21 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
// Service structure
|
2019-09-10 14:24:16 +01:00
|
|
|
//
|
|
|
|
// architecture: Service
|
2019-01-18 15:10:21 +00:00
|
|
|
type Service struct {
|
2019-07-08 23:32:18 +01:00
|
|
|
logger *zap.Logger
|
2019-11-06 17:02:14 +00:00
|
|
|
db PointerDB
|
2019-07-08 23:32:18 +01:00
|
|
|
bucketsDB BucketsDB
|
2019-01-18 15:10:21 +00:00
|
|
|
}
|
|
|
|
|
2019-12-06 18:14:35 +00:00
|
|
|
// NewService creates new metainfo service.
|
2019-09-10 14:24:16 +01:00
|
|
|
func NewService(logger *zap.Logger, db PointerDB, bucketsDB BucketsDB) *Service {
|
2019-11-06 17:02:14 +00:00
|
|
|
return &Service{logger: logger, db: db, bucketsDB: bucketsDB}
|
2019-01-18 15:10:21 +00:00
|
|
|
}
|
|
|
|
|
2019-12-06 18:14:35 +00:00
|
|
|
// Put puts pointer to db under specific path.
|
2019-06-05 15:23:10 +01:00
|
|
|
func (s *Service) Put(ctx context.Context, path string, pointer *pb.Pointer) (err error) {
|
2019-06-04 12:55:38 +01:00
|
|
|
defer mon.Task()(&ctx)(&err)
|
|
|
|
|
2019-01-18 15:10:21 +00:00
|
|
|
// Update the pointer with the creation date
|
2019-07-08 23:16:50 +01:00
|
|
|
pointer.CreationDate = time.Now()
|
2019-01-18 15:10:21 +00:00
|
|
|
|
|
|
|
pointerBytes, err := proto.Marshal(pointer)
|
|
|
|
if err != nil {
|
2019-07-23 15:28:06 +01:00
|
|
|
return Error.Wrap(err)
|
2019-01-18 15:10:21 +00:00
|
|
|
}
|
|
|
|
|
2019-07-25 17:59:46 +01:00
|
|
|
// CompareAndSwap is used instead of Put to avoid overwriting existing pointers
|
2019-11-06 17:02:14 +00:00
|
|
|
err = s.db.CompareAndSwap(ctx, []byte(path), nil, pointerBytes)
|
2019-07-25 17:59:46 +01:00
|
|
|
return Error.Wrap(err)
|
|
|
|
}
|
|
|
|
|
2020-02-11 17:25:35 +00:00
|
|
|
// UnsynchronizedPut puts pointer to db under specific path without verifying for existing pointer under the same path.
|
|
|
|
func (s *Service) UnsynchronizedPut(ctx context.Context, path string, pointer *pb.Pointer) (err error) {
|
|
|
|
defer mon.Task()(&ctx)(&err)
|
|
|
|
|
|
|
|
// Update the pointer with the creation date
|
|
|
|
pointer.CreationDate = time.Now()
|
|
|
|
|
|
|
|
pointerBytes, err := proto.Marshal(pointer)
|
|
|
|
if err != nil {
|
|
|
|
return Error.Wrap(err)
|
|
|
|
}
|
|
|
|
|
|
|
|
err = s.db.Put(ctx, []byte(path), pointerBytes)
|
|
|
|
return Error.Wrap(err)
|
|
|
|
}
|
|
|
|
|
2019-11-05 19:13:45 +00:00
|
|
|
// UpdatePieces calls UpdatePiecesCheckDuplicates with checkDuplicates equal to false.
|
|
|
|
func (s *Service) UpdatePieces(ctx context.Context, path string, ref *pb.Pointer, toAdd, toRemove []*pb.RemotePiece) (pointer *pb.Pointer, err error) {
|
|
|
|
return s.UpdatePiecesCheckDuplicates(ctx, path, ref, toAdd, toRemove, false)
|
|
|
|
}
|
|
|
|
|
|
|
|
// UpdatePiecesCheckDuplicates atomically adds toAdd pieces and removes toRemove pieces from
|
2019-07-25 17:59:46 +01:00
|
|
|
// the pointer under path. ref is the pointer that caller received via Get
|
|
|
|
// prior to calling this method.
|
|
|
|
//
|
2019-11-05 19:13:45 +00:00
|
|
|
// It will first check if the pointer has been deleted or replaced.
|
|
|
|
// Then if checkDuplicates is true it will return an error if the nodes to be
|
|
|
|
// added are already in the pointer.
|
|
|
|
// Then it will remove the toRemove pieces and then it will add the toAdd pieces.
|
2019-07-25 17:59:46 +01:00
|
|
|
// Replacing the node ID and the hash of a piece can be done by adding the
|
|
|
|
// piece to both toAdd and toRemove.
|
2019-11-05 19:13:45 +00:00
|
|
|
func (s *Service) UpdatePiecesCheckDuplicates(ctx context.Context, path string, ref *pb.Pointer, toAdd, toRemove []*pb.RemotePiece, checkDuplicates bool) (pointer *pb.Pointer, err error) {
|
2019-07-25 17:59:46 +01:00
|
|
|
defer mon.Task()(&ctx)(&err)
|
|
|
|
|
|
|
|
for {
|
|
|
|
// read the pointer
|
2019-11-06 17:02:14 +00:00
|
|
|
oldPointerBytes, err := s.db.Get(ctx, []byte(path))
|
2019-07-25 17:59:46 +01:00
|
|
|
if err != nil {
|
2019-12-10 20:21:30 +00:00
|
|
|
if storage.ErrKeyNotFound.Has(err) {
|
|
|
|
err = storj.ErrObjectNotFound.Wrap(err)
|
|
|
|
}
|
2019-07-25 17:59:46 +01:00
|
|
|
return nil, Error.Wrap(err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// unmarshal the pointer
|
|
|
|
pointer = &pb.Pointer{}
|
|
|
|
err = proto.Unmarshal(oldPointerBytes, pointer)
|
|
|
|
if err != nil {
|
|
|
|
return nil, Error.Wrap(err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// check if pointer has been replaced
|
|
|
|
if !pointer.GetCreationDate().Equal(ref.GetCreationDate()) {
|
|
|
|
return nil, Error.New("pointer has been replaced")
|
|
|
|
}
|
|
|
|
|
|
|
|
// put all existing pieces to a map
|
|
|
|
pieceMap := make(map[int32]*pb.RemotePiece)
|
2019-11-05 19:13:45 +00:00
|
|
|
nodePieceMap := make(map[storj.NodeID]struct{})
|
2019-07-25 17:59:46 +01:00
|
|
|
for _, piece := range pointer.GetRemote().GetRemotePieces() {
|
|
|
|
pieceMap[piece.PieceNum] = piece
|
2019-11-05 19:13:45 +00:00
|
|
|
if checkDuplicates {
|
|
|
|
nodePieceMap[piece.NodeId] = struct{}{}
|
|
|
|
}
|
2019-07-25 17:59:46 +01:00
|
|
|
}
|
2019-01-18 15:10:21 +00:00
|
|
|
|
2019-11-05 19:13:45 +00:00
|
|
|
// Return an error if the pointer already has a piece for this node
|
|
|
|
if checkDuplicates {
|
|
|
|
for _, piece := range toAdd {
|
|
|
|
_, ok := nodePieceMap[piece.NodeId]
|
|
|
|
if ok {
|
|
|
|
return nil, ErrNodeAlreadyExists.New("node id already exists in pointer. Path: %s, NodeID: %s", path, piece.NodeId.String())
|
|
|
|
}
|
|
|
|
nodePieceMap[piece.NodeId] = struct{}{}
|
|
|
|
}
|
|
|
|
}
|
2019-07-25 17:59:46 +01:00
|
|
|
// remove the toRemove pieces from the map
|
|
|
|
// only if all piece number, node id and hash match
|
|
|
|
for _, piece := range toRemove {
|
|
|
|
if piece == nil {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
existing := pieceMap[piece.PieceNum]
|
2019-09-16 17:11:12 +01:00
|
|
|
if existing != nil && existing.NodeId == piece.NodeId {
|
2019-07-25 17:59:46 +01:00
|
|
|
delete(pieceMap, piece.PieceNum)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// add the toAdd pieces to the map
|
|
|
|
for _, piece := range toAdd {
|
|
|
|
if piece == nil {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
_, exists := pieceMap[piece.PieceNum]
|
|
|
|
if exists {
|
|
|
|
return nil, Error.New("piece to add already exists (piece no: %d)", piece.PieceNum)
|
|
|
|
}
|
|
|
|
pieceMap[piece.PieceNum] = piece
|
|
|
|
}
|
|
|
|
|
|
|
|
// copy the pieces from the map back to the pointer
|
|
|
|
var pieces []*pb.RemotePiece
|
|
|
|
for _, piece := range pieceMap {
|
2019-09-16 17:11:12 +01:00
|
|
|
// clear hashes so we don't store them
|
|
|
|
piece.Hash = nil
|
2019-07-25 17:59:46 +01:00
|
|
|
pieces = append(pieces, piece)
|
|
|
|
}
|
|
|
|
pointer.GetRemote().RemotePieces = pieces
|
|
|
|
|
2019-09-17 20:18:48 +01:00
|
|
|
pointer.LastRepaired = ref.LastRepaired
|
|
|
|
pointer.RepairCount = ref.RepairCount
|
|
|
|
|
2019-07-25 17:59:46 +01:00
|
|
|
// marshal the pointer
|
|
|
|
newPointerBytes, err := proto.Marshal(pointer)
|
|
|
|
if err != nil {
|
|
|
|
return nil, Error.Wrap(err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// write the pointer using compare-and-swap
|
2019-11-06 17:02:14 +00:00
|
|
|
err = s.db.CompareAndSwap(ctx, []byte(path), oldPointerBytes, newPointerBytes)
|
2019-07-25 17:59:46 +01:00
|
|
|
if storage.ErrValueChanged.Has(err) {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
if err != nil {
|
2019-12-10 20:21:30 +00:00
|
|
|
if storage.ErrKeyNotFound.Has(err) {
|
|
|
|
err = storj.ErrObjectNotFound.Wrap(err)
|
|
|
|
}
|
2019-07-25 17:59:46 +01:00
|
|
|
return nil, Error.Wrap(err)
|
|
|
|
}
|
|
|
|
return pointer, nil
|
|
|
|
}
|
2019-01-18 15:10:21 +00:00
|
|
|
}
|
|
|
|
|
2019-12-10 11:15:35 +00:00
|
|
|
// Get gets decoded pointer from DB.
|
|
|
|
func (s *Service) Get(ctx context.Context, path string) (_ *pb.Pointer, err error) {
|
2019-06-04 12:55:38 +01:00
|
|
|
defer mon.Task()(&ctx)(&err)
|
2019-12-10 11:15:35 +00:00
|
|
|
_, pointer, err := s.GetWithBytes(ctx, path)
|
2019-01-18 15:10:21 +00:00
|
|
|
if err != nil {
|
2019-12-10 11:15:35 +00:00
|
|
|
return nil, err
|
2019-01-18 15:10:21 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return pointer, nil
|
|
|
|
}
|
|
|
|
|
2019-12-10 11:15:35 +00:00
|
|
|
// GetWithBytes gets the protocol buffers encoded and decoded pointer from the DB.
|
2019-11-06 17:02:14 +00:00
|
|
|
func (s *Service) GetWithBytes(ctx context.Context, path string) (pointerBytes []byte, pointer *pb.Pointer, err error) {
|
|
|
|
defer mon.Task()(&ctx)(&err)
|
|
|
|
|
|
|
|
pointerBytes, err = s.db.Get(ctx, []byte(path))
|
|
|
|
if err != nil {
|
2019-12-10 20:21:30 +00:00
|
|
|
if storage.ErrKeyNotFound.Has(err) {
|
|
|
|
err = storj.ErrObjectNotFound.Wrap(err)
|
|
|
|
}
|
2019-11-06 17:02:14 +00:00
|
|
|
return nil, nil, Error.Wrap(err)
|
|
|
|
}
|
|
|
|
|
|
|
|
pointer = &pb.Pointer{}
|
|
|
|
err = proto.Unmarshal(pointerBytes, pointer)
|
|
|
|
if err != nil {
|
|
|
|
return nil, nil, Error.Wrap(err)
|
|
|
|
}
|
|
|
|
|
|
|
|
return pointerBytes, pointer, nil
|
|
|
|
}
|
|
|
|
|
2019-01-18 15:10:21 +00:00
|
|
|
// List returns all Path keys in the pointers bucket
|
2019-09-25 22:30:41 +01:00
|
|
|
func (s *Service) List(ctx context.Context, prefix string, startAfter string, recursive bool, limit int32,
|
2019-01-18 15:10:21 +00:00
|
|
|
metaFlags uint32) (items []*pb.ListResponse_Item, more bool, err error) {
|
2019-06-04 12:55:38 +01:00
|
|
|
defer mon.Task()(&ctx)(&err)
|
2019-01-18 15:10:21 +00:00
|
|
|
|
|
|
|
var prefixKey storage.Key
|
|
|
|
if prefix != "" {
|
|
|
|
prefixKey = storage.Key(prefix)
|
|
|
|
if prefix[len(prefix)-1] != storage.Delimiter {
|
|
|
|
prefixKey = append(prefixKey, storage.Delimiter)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-11-06 17:02:14 +00:00
|
|
|
rawItems, more, err := storage.ListV2(ctx, s.db, storage.ListOptions{
|
2019-01-18 15:10:21 +00:00
|
|
|
Prefix: prefixKey,
|
|
|
|
StartAfter: storage.Key(startAfter),
|
|
|
|
Recursive: recursive,
|
|
|
|
Limit: int(limit),
|
|
|
|
IncludeValue: metaFlags != meta.None,
|
|
|
|
})
|
|
|
|
if err != nil {
|
2019-07-25 17:59:46 +01:00
|
|
|
return nil, false, Error.Wrap(err)
|
2019-01-18 15:10:21 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
for _, rawItem := range rawItems {
|
2019-06-04 12:55:38 +01:00
|
|
|
items = append(items, s.createListItem(ctx, rawItem, metaFlags))
|
2019-01-18 15:10:21 +00:00
|
|
|
}
|
|
|
|
return items, more, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// createListItem creates a new list item with the given path. It also adds
|
|
|
|
// the metadata according to the given metaFlags.
|
2019-06-04 12:55:38 +01:00
|
|
|
func (s *Service) createListItem(ctx context.Context, rawItem storage.ListItem, metaFlags uint32) *pb.ListResponse_Item {
|
|
|
|
defer mon.Task()(&ctx)(nil)
|
2019-01-18 15:10:21 +00:00
|
|
|
item := &pb.ListResponse_Item{
|
|
|
|
Path: rawItem.Key.String(),
|
|
|
|
IsPrefix: rawItem.IsPrefix,
|
|
|
|
}
|
|
|
|
if item.IsPrefix {
|
|
|
|
return item
|
|
|
|
}
|
|
|
|
|
|
|
|
err := s.setMetadata(item, rawItem.Value, metaFlags)
|
|
|
|
if err != nil {
|
|
|
|
s.logger.Warn("err retrieving metadata", zap.Error(err))
|
|
|
|
}
|
|
|
|
return item
|
|
|
|
}
|
|
|
|
|
|
|
|
// getMetadata adds the metadata to the given item pointer according to the
|
|
|
|
// given metaFlags
|
|
|
|
func (s *Service) setMetadata(item *pb.ListResponse_Item, data []byte, metaFlags uint32) (err error) {
|
|
|
|
if metaFlags == meta.None || len(data) == 0 {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
pr := &pb.Pointer{}
|
|
|
|
err = proto.Unmarshal(data, pr)
|
|
|
|
if err != nil {
|
2019-07-25 17:59:46 +01:00
|
|
|
return Error.Wrap(err)
|
2019-01-18 15:10:21 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Start with an empty pointer to and add only what's requested in
|
|
|
|
// metaFlags to safe to transfer payload
|
|
|
|
item.Pointer = &pb.Pointer{}
|
|
|
|
if metaFlags&meta.Modified != 0 {
|
|
|
|
item.Pointer.CreationDate = pr.GetCreationDate()
|
|
|
|
}
|
|
|
|
if metaFlags&meta.Expiration != 0 {
|
|
|
|
item.Pointer.ExpirationDate = pr.GetExpirationDate()
|
|
|
|
}
|
|
|
|
if metaFlags&meta.Size != 0 {
|
|
|
|
item.Pointer.SegmentSize = pr.GetSegmentSize()
|
|
|
|
}
|
|
|
|
if metaFlags&meta.UserDefined != 0 {
|
|
|
|
item.Pointer.Metadata = pr.GetMetadata()
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2019-11-06 17:02:14 +00:00
|
|
|
// Delete deletes a pointer bytes when it matches oldPointerBytes, otherwise it'll fail.
|
|
|
|
func (s *Service) Delete(ctx context.Context, path string, oldPointerBytes []byte) (err error) {
|
|
|
|
defer mon.Task()(&ctx)(&err)
|
|
|
|
|
2019-12-10 20:21:30 +00:00
|
|
|
err = s.db.CompareAndSwap(ctx, []byte(path), oldPointerBytes, nil)
|
|
|
|
if storage.ErrKeyNotFound.Has(err) {
|
|
|
|
err = storj.ErrObjectNotFound.Wrap(err)
|
|
|
|
}
|
|
|
|
return Error.Wrap(err)
|
2019-11-06 17:02:14 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// UnsynchronizedDelete deletes from item from db without verifying whether the pointer has changed in the database.
|
|
|
|
func (s *Service) UnsynchronizedDelete(ctx context.Context, path string) (err error) {
|
2019-06-04 12:55:38 +01:00
|
|
|
defer mon.Task()(&ctx)(&err)
|
2019-12-10 20:21:30 +00:00
|
|
|
|
|
|
|
err = s.db.Delete(ctx, []byte(path))
|
|
|
|
if storage.ErrKeyNotFound.Has(err) {
|
|
|
|
err = storj.ErrObjectNotFound.Wrap(err)
|
|
|
|
}
|
|
|
|
return Error.Wrap(err)
|
2019-01-18 15:10:21 +00:00
|
|
|
}
|
|
|
|
|
2019-07-08 23:32:18 +01:00
|
|
|
// CreateBucket creates a new bucket in the buckets db
|
|
|
|
func (s *Service) CreateBucket(ctx context.Context, bucket storj.Bucket) (_ storj.Bucket, err error) {
|
|
|
|
defer mon.Task()(&ctx)(&err)
|
|
|
|
return s.bucketsDB.CreateBucket(ctx, bucket)
|
|
|
|
}
|
|
|
|
|
|
|
|
// GetBucket returns an existing bucket in the buckets db
|
|
|
|
func (s *Service) GetBucket(ctx context.Context, bucketName []byte, projectID uuid.UUID) (_ storj.Bucket, err error) {
|
|
|
|
defer mon.Task()(&ctx)(&err)
|
|
|
|
return s.bucketsDB.GetBucket(ctx, bucketName, projectID)
|
|
|
|
}
|
|
|
|
|
2019-07-19 16:17:34 +01:00
|
|
|
// UpdateBucket returns an updated bucket in the buckets db
|
|
|
|
func (s *Service) UpdateBucket(ctx context.Context, bucket storj.Bucket) (_ storj.Bucket, err error) {
|
|
|
|
defer mon.Task()(&ctx)(&err)
|
|
|
|
return s.bucketsDB.UpdateBucket(ctx, bucket)
|
|
|
|
}
|
|
|
|
|
2019-07-08 23:32:18 +01:00
|
|
|
// DeleteBucket deletes a bucket from the bucekts db
|
|
|
|
func (s *Service) DeleteBucket(ctx context.Context, bucketName []byte, projectID uuid.UUID) (err error) {
|
|
|
|
defer mon.Task()(&ctx)(&err)
|
|
|
|
return s.bucketsDB.DeleteBucket(ctx, bucketName, projectID)
|
|
|
|
}
|
|
|
|
|
|
|
|
// ListBuckets returns a list of buckets for a project
|
2019-07-12 13:57:02 +01:00
|
|
|
func (s *Service) ListBuckets(ctx context.Context, projectID uuid.UUID, listOpts storj.BucketListOptions, allowedBuckets macaroon.AllowedBuckets) (bucketList storj.BucketList, err error) {
|
2019-07-08 23:32:18 +01:00
|
|
|
defer mon.Task()(&ctx)(&err)
|
|
|
|
return s.bucketsDB.ListBuckets(ctx, projectID, listOpts, allowedBuckets)
|
|
|
|
}
|