2019-03-18 10:55:06 +00:00
|
|
|
// Copyright (C) 2019 Storj Labs, Inc.
|
|
|
|
// See LICENSE for copying information.
|
|
|
|
|
|
|
|
package storagenodedbtest_test
|
|
|
|
|
|
|
|
import (
|
2020-01-14 11:07:35 +00:00
|
|
|
"context"
|
2019-09-18 17:17:28 +01:00
|
|
|
"path/filepath"
|
2019-07-02 22:23:02 +01:00
|
|
|
"runtime"
|
|
|
|
"sync"
|
2019-03-18 10:55:06 +00:00
|
|
|
"testing"
|
2019-07-09 22:54:00 +01:00
|
|
|
"time"
|
2019-03-18 10:55:06 +00:00
|
|
|
|
2019-07-02 22:23:02 +01:00
|
|
|
"github.com/stretchr/testify/require"
|
|
|
|
"go.uber.org/zap/zaptest"
|
|
|
|
|
2020-01-14 11:07:35 +00:00
|
|
|
"storj.io/common/errs2"
|
2019-12-27 11:48:47 +00:00
|
|
|
"storj.io/common/identity/testidentity"
|
|
|
|
"storj.io/common/pb"
|
|
|
|
"storj.io/common/signing"
|
|
|
|
"storj.io/common/storj"
|
|
|
|
"storj.io/common/testcontext"
|
|
|
|
"storj.io/common/testrand"
|
2020-03-30 10:08:50 +01:00
|
|
|
"storj.io/common/uuid"
|
2020-04-14 13:39:42 +01:00
|
|
|
"storj.io/storj/storage/filestore"
|
2019-03-18 10:55:06 +00:00
|
|
|
"storj.io/storj/storagenode"
|
2019-07-02 22:23:02 +01:00
|
|
|
"storj.io/storj/storagenode/orders"
|
|
|
|
"storj.io/storj/storagenode/storagenodedb"
|
2019-03-18 10:55:06 +00:00
|
|
|
"storj.io/storj/storagenode/storagenodedb/storagenodedbtest"
|
|
|
|
)
|
|
|
|
|
|
|
|
func TestDatabase(t *testing.T) {
|
2020-01-20 14:56:12 +00:00
|
|
|
storagenodedbtest.Run(t, func(ctx *testcontext.Context, t *testing.T, db storagenode.DB) {
|
2020-01-14 11:07:35 +00:00
|
|
|
// Ensure that database implementation handles context cancellation.
|
|
|
|
canceledCtx, cancel := context.WithCancel(ctx)
|
|
|
|
cancel()
|
|
|
|
|
|
|
|
serials := db.UsedSerials()
|
|
|
|
err := serials.Add(canceledCtx, testrand.NodeID(), testrand.SerialNumber(), time.Now().Add(time.Hour))
|
|
|
|
require.True(t, errs2.IsCanceled(err), err)
|
2019-03-18 10:55:06 +00:00
|
|
|
})
|
|
|
|
}
|
2019-07-02 22:23:02 +01:00
|
|
|
|
|
|
|
func TestFileConcurrency(t *testing.T) {
|
|
|
|
ctx := testcontext.New(t)
|
|
|
|
defer ctx.Cleanup()
|
|
|
|
|
|
|
|
log := zaptest.NewLogger(t)
|
|
|
|
|
|
|
|
db, err := storagenodedb.New(log, storagenodedb.Config{
|
2019-09-19 20:56:34 +01:00
|
|
|
Pieces: ctx.Dir("storage"),
|
|
|
|
Info2: ctx.Dir("storage") + "/info.db",
|
2019-07-02 22:23:02 +01:00
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
defer ctx.Check(db.Close)
|
|
|
|
|
|
|
|
testConcurrency(t, ctx, db)
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestInMemoryConcurrency(t *testing.T) {
|
|
|
|
ctx := testcontext.New(t)
|
|
|
|
defer ctx.Cleanup()
|
|
|
|
|
|
|
|
log := zaptest.NewLogger(t)
|
|
|
|
|
2019-09-18 17:17:28 +01:00
|
|
|
storageDir := ctx.Dir("storage")
|
|
|
|
cfg := storagenodedb.Config{
|
2020-04-14 13:39:42 +01:00
|
|
|
Pieces: storageDir,
|
|
|
|
Storage: storageDir,
|
|
|
|
Info: filepath.Join(storageDir, "piecestore.db"),
|
|
|
|
Info2: filepath.Join(storageDir, "info.db"),
|
|
|
|
Filestore: filestore.DefaultConfig,
|
2019-09-18 17:17:28 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
db, err := storagenodedb.New(log, cfg)
|
2019-07-02 22:23:02 +01:00
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
defer ctx.Check(db.Close)
|
|
|
|
|
|
|
|
testConcurrency(t, ctx, db)
|
|
|
|
}
|
|
|
|
|
|
|
|
func testConcurrency(t *testing.T, ctx *testcontext.Context, db *storagenodedb.DB) {
|
|
|
|
t.Run("Sqlite", func(t *testing.T) {
|
|
|
|
runtime.GOMAXPROCS(2)
|
|
|
|
|
2020-04-30 07:36:59 +01:00
|
|
|
err := db.MigrateToLatest(ctx)
|
2019-07-02 22:23:02 +01:00
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
|
|
|
|
ordersMap := make(map[string]orders.Info)
|
|
|
|
err = createOrders(t, ctx, ordersMap, 1000)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
err = insertOrders(t, ctx, db, ordersMap)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
err = verifyOrders(t, ctx, db, ordersMap)
|
|
|
|
require.NoError(t, err)
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
func insertOrders(t *testing.T, ctx *testcontext.Context, db *storagenodedb.DB, ordersMap map[string]orders.Info) (err error) {
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
for _, order := range ordersMap {
|
|
|
|
wg.Add(1)
|
|
|
|
o := order
|
|
|
|
go insertOrder(t, ctx, db, &wg, &o)
|
|
|
|
|
|
|
|
}
|
|
|
|
wg.Wait()
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func insertOrder(t *testing.T, ctx *testcontext.Context, db *storagenodedb.DB, wg *sync.WaitGroup, order *orders.Info) {
|
|
|
|
defer wg.Done()
|
|
|
|
err := db.Orders().Enqueue(ctx, order)
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
func verifyOrders(t *testing.T, ctx *testcontext.Context, db *storagenodedb.DB, orders map[string]orders.Info) (err error) {
|
|
|
|
dbOrders, _ := db.Orders().ListUnsent(ctx, 10000)
|
|
|
|
found := 0
|
|
|
|
for _, order := range orders {
|
|
|
|
for _, dbOrder := range dbOrders {
|
|
|
|
if order.Order.SerialNumber == dbOrder.Order.SerialNumber {
|
|
|
|
//fmt.Printf("Found %v\n", order.Order.SerialNumber)
|
|
|
|
found++
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
require.Equal(t, len(orders), found, "Number found must equal the length of the test data")
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func createOrders(t *testing.T, ctx *testcontext.Context, orders map[string]orders.Info, count int) (err error) {
|
|
|
|
for i := 0; i < count; i++ {
|
|
|
|
key, err := uuid.New()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
order := createOrder(t, ctx)
|
|
|
|
orders[key.String()] = *order
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2019-07-11 21:51:40 +01:00
|
|
|
func createOrder(t *testing.T, ctx *testcontext.Context) (info *orders.Info) {
|
2019-07-02 22:23:02 +01:00
|
|
|
storageNodeIdentity := testidentity.MustPregeneratedSignedIdentity(0, storj.LatestIDVersion())
|
|
|
|
satelliteIdentity := testidentity.MustPregeneratedSignedIdentity(1, storj.LatestIDVersion())
|
|
|
|
|
2019-07-11 21:51:40 +01:00
|
|
|
piecePublicKey, piecePrivateKey, err := storj.NewPieceKey()
|
|
|
|
require.NoError(t, err)
|
2019-07-02 22:23:02 +01:00
|
|
|
|
2019-07-11 21:51:40 +01:00
|
|
|
piece := testrand.PieceID()
|
2019-07-02 22:23:02 +01:00
|
|
|
serialNumber := testrand.SerialNumber()
|
2019-07-09 22:54:00 +01:00
|
|
|
expiration := time.Now()
|
2019-07-02 22:23:02 +01:00
|
|
|
|
|
|
|
limit, err := signing.SignOrderLimit(ctx, signing.SignerFromFullIdentity(satelliteIdentity), &pb.OrderLimit{
|
|
|
|
SerialNumber: serialNumber,
|
|
|
|
SatelliteId: satelliteIdentity.ID,
|
2019-07-11 21:51:40 +01:00
|
|
|
UplinkPublicKey: piecePublicKey,
|
2019-07-02 22:23:02 +01:00
|
|
|
StorageNodeId: storageNodeIdentity.ID,
|
|
|
|
PieceId: piece,
|
|
|
|
Limit: 100,
|
|
|
|
Action: pb.PieceAction_GET,
|
2019-07-09 22:54:00 +01:00
|
|
|
PieceExpiration: expiration,
|
|
|
|
OrderExpiration: expiration,
|
2019-07-02 22:23:02 +01:00
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2019-07-11 21:51:40 +01:00
|
|
|
order, err := signing.SignUplinkOrder(ctx, piecePrivateKey, &pb.Order{
|
2019-07-02 22:23:02 +01:00
|
|
|
SerialNumber: serialNumber,
|
|
|
|
Amount: 50,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
return &orders.Info{
|
2019-07-09 22:33:45 +01:00
|
|
|
Limit: limit,
|
|
|
|
Order: order,
|
2019-07-02 22:23:02 +01:00
|
|
|
}
|
|
|
|
}
|