835 lines
22 KiB
Go
835 lines
22 KiB
Go
// Copyright (C) 2019 Storj Labs, Inc.
|
|
// See LICENSE for copying information
|
|
|
|
// Package testplanet implements the full network wiring for testing
|
|
package testplanet
|
|
|
|
import (
|
|
"context"
|
|
"errors"
|
|
"fmt"
|
|
"io"
|
|
"io/ioutil"
|
|
"net"
|
|
"os"
|
|
"path/filepath"
|
|
"runtime"
|
|
"strconv"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/zeebo/errs"
|
|
"go.uber.org/zap"
|
|
"go.uber.org/zap/zaptest"
|
|
"golang.org/x/sync/errgroup"
|
|
|
|
"storj.io/storj/bootstrap"
|
|
"storj.io/storj/bootstrap/bootstrapdb"
|
|
"storj.io/storj/bootstrap/bootstrapweb/bootstrapserver"
|
|
"storj.io/storj/internal/memory"
|
|
"storj.io/storj/internal/testidentity"
|
|
"storj.io/storj/internal/version"
|
|
"storj.io/storj/pkg/accounting/rollup"
|
|
"storj.io/storj/pkg/accounting/tally"
|
|
"storj.io/storj/pkg/audit"
|
|
"storj.io/storj/pkg/datarepair/checker"
|
|
"storj.io/storj/pkg/datarepair/repairer"
|
|
"storj.io/storj/pkg/discovery"
|
|
"storj.io/storj/pkg/identity"
|
|
"storj.io/storj/pkg/kademlia"
|
|
"storj.io/storj/pkg/overlay"
|
|
"storj.io/storj/pkg/pb"
|
|
"storj.io/storj/pkg/peertls/extensions"
|
|
"storj.io/storj/pkg/peertls/tlsopts"
|
|
"storj.io/storj/pkg/server"
|
|
"storj.io/storj/pkg/storj"
|
|
"storj.io/storj/satellite"
|
|
"storj.io/storj/satellite/console"
|
|
"storj.io/storj/satellite/console/consoleweb"
|
|
"storj.io/storj/satellite/mailservice"
|
|
"storj.io/storj/satellite/marketing/marketingweb"
|
|
"storj.io/storj/satellite/metainfo"
|
|
"storj.io/storj/satellite/satellitedb"
|
|
"storj.io/storj/satellite/vouchers"
|
|
"storj.io/storj/storagenode"
|
|
"storj.io/storj/storagenode/collector"
|
|
"storj.io/storj/storagenode/monitor"
|
|
"storj.io/storj/storagenode/orders"
|
|
"storj.io/storj/storagenode/piecestore"
|
|
"storj.io/storj/storagenode/storagenodedb"
|
|
"storj.io/storj/versioncontrol"
|
|
)
|
|
|
|
// Peer represents one of StorageNode or Satellite
|
|
type Peer interface {
|
|
ID() storj.NodeID
|
|
Addr() string
|
|
Local() overlay.NodeDossier
|
|
|
|
Run(context.Context) error
|
|
Close() error
|
|
}
|
|
|
|
// Config describes planet configuration
|
|
type Config struct {
|
|
SatelliteCount int
|
|
StorageNodeCount int
|
|
UplinkCount int
|
|
|
|
Identities *testidentity.Identities
|
|
IdentityVersion *storj.IDVersion
|
|
Reconfigure Reconfigure
|
|
}
|
|
|
|
// Planet is a full storj system setup.
|
|
type Planet struct {
|
|
log *zap.Logger
|
|
config Config
|
|
directory string // TODO: ensure that everything is in-memory to speed things up
|
|
|
|
started bool
|
|
shutdown bool
|
|
|
|
peers []closablePeer
|
|
databases []io.Closer
|
|
uplinks []*Uplink
|
|
|
|
Bootstrap *bootstrap.Peer
|
|
VersionControl *versioncontrol.Peer
|
|
Satellites []*satellite.Peer
|
|
StorageNodes []*storagenode.Peer
|
|
Uplinks []*Uplink
|
|
|
|
identities *testidentity.Identities
|
|
whitelistPath string // TODO: in-memory
|
|
|
|
run errgroup.Group
|
|
cancel func()
|
|
}
|
|
|
|
type closablePeer struct {
|
|
peer Peer
|
|
|
|
ctx context.Context
|
|
cancel func()
|
|
|
|
close sync.Once
|
|
err error
|
|
}
|
|
|
|
// Close closes safely the peer.
|
|
func (peer *closablePeer) Close() error {
|
|
peer.cancel()
|
|
peer.close.Do(func() {
|
|
peer.err = peer.peer.Close()
|
|
})
|
|
return peer.err
|
|
}
|
|
|
|
// New creates a new full system with the given number of nodes.
|
|
func New(t zaptest.TestingT, satelliteCount, storageNodeCount, uplinkCount int) (*Planet, error) {
|
|
var log *zap.Logger
|
|
if t == nil {
|
|
log = zap.NewNop()
|
|
} else {
|
|
log = zaptest.NewLogger(t)
|
|
}
|
|
|
|
return NewWithLogger(log, satelliteCount, storageNodeCount, uplinkCount)
|
|
}
|
|
|
|
// NewWithIdentityVersion creates a new full system with the given version for node identities and the given number of nodes.
|
|
func NewWithIdentityVersion(t zaptest.TestingT, identityVersion *storj.IDVersion, satelliteCount, storageNodeCount, uplinkCount int) (*Planet, error) {
|
|
var log *zap.Logger
|
|
if t == nil {
|
|
log = zap.NewNop()
|
|
} else {
|
|
log = zaptest.NewLogger(t)
|
|
}
|
|
|
|
return NewCustom(log, Config{
|
|
SatelliteCount: satelliteCount,
|
|
StorageNodeCount: storageNodeCount,
|
|
UplinkCount: uplinkCount,
|
|
IdentityVersion: identityVersion,
|
|
})
|
|
}
|
|
|
|
// NewWithLogger creates a new full system with the given number of nodes.
|
|
func NewWithLogger(log *zap.Logger, satelliteCount, storageNodeCount, uplinkCount int) (*Planet, error) {
|
|
return NewCustom(log, Config{
|
|
SatelliteCount: satelliteCount,
|
|
StorageNodeCount: storageNodeCount,
|
|
UplinkCount: uplinkCount,
|
|
})
|
|
}
|
|
|
|
// NewCustom creates a new full system with the specified configuration.
|
|
func NewCustom(log *zap.Logger, config Config) (*Planet, error) {
|
|
if config.IdentityVersion == nil {
|
|
version := storj.LatestIDVersion()
|
|
config.IdentityVersion = &version
|
|
}
|
|
if config.Identities == nil {
|
|
config.Identities = testidentity.NewPregeneratedSignedIdentities(*config.IdentityVersion)
|
|
}
|
|
|
|
planet := &Planet{
|
|
log: log,
|
|
config: config,
|
|
identities: config.Identities,
|
|
}
|
|
|
|
var err error
|
|
planet.directory, err = ioutil.TempDir("", "planet")
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
whitelistPath, err := planet.WriteWhitelist(*config.IdentityVersion)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
planet.whitelistPath = whitelistPath
|
|
|
|
planet.VersionControl, err = planet.newVersionControlServer()
|
|
if err != nil {
|
|
return nil, errs.Combine(err, planet.Shutdown())
|
|
}
|
|
|
|
planet.Bootstrap, err = planet.newBootstrap()
|
|
if err != nil {
|
|
return nil, errs.Combine(err, planet.Shutdown())
|
|
}
|
|
|
|
planet.Satellites, err = planet.newSatellites(config.SatelliteCount)
|
|
if err != nil {
|
|
return nil, errs.Combine(err, planet.Shutdown())
|
|
}
|
|
|
|
whitelistedSatellites := make([]string, len(planet.Satellites))
|
|
for _, satellite := range planet.Satellites {
|
|
whitelistedSatellites = append(whitelistedSatellites, satellite.ID().String())
|
|
}
|
|
|
|
planet.StorageNodes, err = planet.newStorageNodes(config.StorageNodeCount, whitelistedSatellites)
|
|
if err != nil {
|
|
return nil, errs.Combine(err, planet.Shutdown())
|
|
}
|
|
|
|
planet.Uplinks, err = planet.newUplinks("uplink", config.UplinkCount, config.StorageNodeCount)
|
|
if err != nil {
|
|
return nil, errs.Combine(err, planet.Shutdown())
|
|
}
|
|
|
|
// init Satellites
|
|
for _, satellite := range planet.Satellites {
|
|
if len(satellite.Kademlia.Service.GetBootstrapNodes()) == 0 {
|
|
satellite.Kademlia.Service.SetBootstrapNodes([]pb.Node{planet.Bootstrap.Local().Node})
|
|
}
|
|
}
|
|
// init storage nodes
|
|
for _, storageNode := range planet.StorageNodes {
|
|
if len(storageNode.Kademlia.Service.GetBootstrapNodes()) == 0 {
|
|
storageNode.Kademlia.Service.SetBootstrapNodes([]pb.Node{planet.Bootstrap.Local().Node})
|
|
}
|
|
}
|
|
|
|
return planet, nil
|
|
}
|
|
|
|
// Start starts all the nodes.
|
|
func (planet *Planet) Start(ctx context.Context) {
|
|
ctx, cancel := context.WithCancel(ctx)
|
|
planet.cancel = cancel
|
|
|
|
planet.run.Go(func() error {
|
|
return planet.VersionControl.Run(ctx)
|
|
})
|
|
|
|
for i := range planet.peers {
|
|
peer := &planet.peers[i]
|
|
peer.ctx, peer.cancel = context.WithCancel(ctx)
|
|
planet.run.Go(func() error {
|
|
return peer.peer.Run(peer.ctx)
|
|
})
|
|
}
|
|
|
|
planet.started = true
|
|
|
|
planet.Bootstrap.Kademlia.Service.WaitForBootstrap()
|
|
|
|
for _, peer := range planet.StorageNodes {
|
|
peer.Kademlia.Service.WaitForBootstrap()
|
|
}
|
|
|
|
for _, peer := range planet.Satellites {
|
|
peer.Kademlia.Service.WaitForBootstrap()
|
|
}
|
|
|
|
planet.Reconnect(ctx)
|
|
}
|
|
|
|
// Reconnect reconnects all nodes with each other.
|
|
func (planet *Planet) Reconnect(ctx context.Context) {
|
|
log := planet.log.Named("reconnect")
|
|
|
|
var group errgroup.Group
|
|
|
|
// TODO: instead of pinging try to use Lookups or natural discovery to ensure
|
|
// everyone finds everyone else
|
|
|
|
for _, storageNode := range planet.StorageNodes {
|
|
storageNode := storageNode
|
|
group.Go(func() error {
|
|
_, err := storageNode.Kademlia.Service.Ping(ctx, planet.Bootstrap.Local().Node)
|
|
if err != nil {
|
|
log.Error("storage node did not find bootstrap", zap.Error(err))
|
|
}
|
|
return nil
|
|
})
|
|
}
|
|
|
|
for _, satellite := range planet.Satellites {
|
|
satellite := satellite
|
|
group.Go(func() error {
|
|
for _, storageNode := range planet.StorageNodes {
|
|
_, err := satellite.Kademlia.Service.Ping(ctx, storageNode.Local().Node)
|
|
if err != nil {
|
|
log.Error("satellite did not find storage node", zap.Error(err))
|
|
}
|
|
}
|
|
return nil
|
|
})
|
|
}
|
|
|
|
_ = group.Wait() // none of the goroutines return an error
|
|
}
|
|
|
|
// StopPeer stops a single peer in the planet
|
|
func (planet *Planet) StopPeer(peer Peer) error {
|
|
for i := range planet.peers {
|
|
p := &planet.peers[i]
|
|
if p.peer == peer {
|
|
return p.Close()
|
|
}
|
|
}
|
|
return errors.New("unknown peer")
|
|
}
|
|
|
|
// Size returns number of nodes in the network
|
|
func (planet *Planet) Size() int { return len(planet.uplinks) + len(planet.peers) }
|
|
|
|
// Shutdown shuts down all the nodes and deletes temporary directories.
|
|
func (planet *Planet) Shutdown() error {
|
|
if !planet.started {
|
|
return errors.New("Start was never called")
|
|
}
|
|
if planet.shutdown {
|
|
panic("double Shutdown")
|
|
}
|
|
planet.shutdown = true
|
|
|
|
planet.cancel()
|
|
|
|
var errlist errs.Group
|
|
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
go func() {
|
|
// TODO: add diagnostics to see what hasn't been properly shut down
|
|
timer := time.NewTimer(30 * time.Second)
|
|
defer timer.Stop()
|
|
select {
|
|
case <-timer.C:
|
|
panic("planet took too long to shutdown")
|
|
case <-ctx.Done():
|
|
}
|
|
}()
|
|
errlist.Add(planet.run.Wait())
|
|
cancel()
|
|
|
|
// shutdown in reverse order
|
|
for i := len(planet.uplinks) - 1; i >= 0; i-- {
|
|
node := planet.uplinks[i]
|
|
errlist.Add(node.Shutdown())
|
|
}
|
|
for i := len(planet.peers) - 1; i >= 0; i-- {
|
|
peer := &planet.peers[i]
|
|
errlist.Add(peer.Close())
|
|
}
|
|
for _, db := range planet.databases {
|
|
errlist.Add(db.Close())
|
|
}
|
|
errlist.Add(planet.VersionControl.Close())
|
|
|
|
errlist.Add(os.RemoveAll(planet.directory))
|
|
return errlist.Err()
|
|
}
|
|
|
|
// newUplinks creates initializes uplinks, requires peer to have at least one satellite
|
|
func (planet *Planet) newUplinks(prefix string, count, storageNodeCount int) ([]*Uplink, error) {
|
|
var xs []*Uplink
|
|
for i := 0; i < count; i++ {
|
|
uplink, err := planet.newUplink(prefix+strconv.Itoa(i), storageNodeCount)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
xs = append(xs, uplink)
|
|
}
|
|
|
|
return xs, nil
|
|
}
|
|
|
|
// newSatellites initializes satellites
|
|
func (planet *Planet) newSatellites(count int) ([]*satellite.Peer, error) {
|
|
// TODO: move into separate file
|
|
var xs []*satellite.Peer
|
|
defer func() {
|
|
for _, x := range xs {
|
|
planet.peers = append(planet.peers, closablePeer{peer: x})
|
|
}
|
|
}()
|
|
|
|
for i := 0; i < count; i++ {
|
|
prefix := "satellite" + strconv.Itoa(i)
|
|
log := planet.log.Named(prefix)
|
|
|
|
storageDir := filepath.Join(planet.directory, prefix)
|
|
if err := os.MkdirAll(storageDir, 0700); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
identity, err := planet.NewIdentity()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
var db satellite.DB
|
|
if planet.config.Reconfigure.NewSatelliteDB != nil {
|
|
db, err = planet.config.Reconfigure.NewSatelliteDB(log.Named("db"), i)
|
|
} else {
|
|
db, err = satellitedb.NewInMemory(log.Named("db"))
|
|
}
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
err = db.CreateTables()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
planet.databases = append(planet.databases, db)
|
|
|
|
config := satellite.Config{
|
|
Server: server.Config{
|
|
Address: "127.0.0.1:0",
|
|
PrivateAddress: "127.0.0.1:0",
|
|
|
|
Config: tlsopts.Config{
|
|
RevocationDBURL: "bolt://" + filepath.Join(storageDir, "revocation.db"),
|
|
UsePeerCAWhitelist: true,
|
|
PeerCAWhitelistPath: planet.whitelistPath,
|
|
PeerIDVersions: "latest",
|
|
Extensions: extensions.Config{
|
|
Revocation: false,
|
|
WhitelistSignedLeaf: false,
|
|
},
|
|
},
|
|
},
|
|
Kademlia: kademlia.Config{
|
|
Alpha: 5,
|
|
BootstrapBackoffBase: 500 * time.Millisecond,
|
|
BootstrapBackoffMax: 2 * time.Second,
|
|
DBPath: storageDir, // TODO: replace with master db
|
|
Operator: kademlia.OperatorConfig{
|
|
Email: prefix + "@mail.test",
|
|
Wallet: "0x" + strings.Repeat("00", 20),
|
|
},
|
|
},
|
|
Overlay: overlay.Config{
|
|
Node: overlay.NodeSelectionConfig{
|
|
UptimeRatio: 0,
|
|
UptimeCount: 0,
|
|
AuditSuccessRatio: 0,
|
|
AuditCount: 0,
|
|
NewNodePercentage: 0,
|
|
OnlineWindow: time.Hour,
|
|
DistinctIP: false,
|
|
|
|
AuditReputationRepairWeight: 1,
|
|
AuditReputationUplinkWeight: 1,
|
|
AuditReputationAlpha0: 1,
|
|
AuditReputationBeta0: 0,
|
|
AuditReputationLambda: 1,
|
|
AuditReputationWeight: 1,
|
|
UptimeReputationRepairWeight: 1,
|
|
UptimeReputationUplinkWeight: 1,
|
|
UptimeReputationAlpha0: 1,
|
|
UptimeReputationBeta0: 0,
|
|
UptimeReputationLambda: 1,
|
|
UptimeReputationWeight: 1,
|
|
},
|
|
},
|
|
Discovery: discovery.Config{
|
|
DiscoveryInterval: 1 * time.Second,
|
|
RefreshInterval: 1 * time.Second,
|
|
RefreshLimit: 100,
|
|
},
|
|
Metainfo: metainfo.Config{
|
|
DatabaseURL: "bolt://" + filepath.Join(storageDir, "pointers.db"),
|
|
MinRemoteSegmentSize: 0, // TODO: fix tests to work with 1024
|
|
MaxInlineSegmentSize: 8000,
|
|
Overlay: true,
|
|
BwExpiration: 45,
|
|
},
|
|
Checker: checker.Config{
|
|
Interval: 30 * time.Second,
|
|
IrreparableInterval: 15 * time.Second,
|
|
},
|
|
Repairer: repairer.Config{
|
|
MaxRepair: 10,
|
|
Interval: time.Hour,
|
|
Timeout: 10 * time.Second, // Repairs can take up to 4 seconds. Leaving room for outliers
|
|
MaxBufferMem: 4 * memory.MiB,
|
|
},
|
|
Audit: audit.Config{
|
|
MaxRetriesStatDB: 0,
|
|
Interval: 30 * time.Second,
|
|
MinBytesPerSecond: 1 * memory.KB,
|
|
},
|
|
Tally: tally.Config{
|
|
Interval: 30 * time.Second,
|
|
},
|
|
Rollup: rollup.Config{
|
|
Interval: 2 * time.Minute,
|
|
MaxAlphaUsage: 25 * memory.GB,
|
|
DeleteTallies: false,
|
|
},
|
|
Mail: mailservice.Config{
|
|
SMTPServerAddress: "smtp.mail.test:587",
|
|
From: "Labs <storj@mail.test>",
|
|
AuthType: "simulate",
|
|
},
|
|
Console: consoleweb.Config{
|
|
Address: "127.0.0.1:0",
|
|
PasswordCost: console.TestPasswordCost,
|
|
AuthTokenSecret: "my-suppa-secret-key",
|
|
},
|
|
Marketing: marketingweb.Config{
|
|
Address: "127.0.0.1:0",
|
|
},
|
|
Vouchers: vouchers.Config{
|
|
Expiration: 30,
|
|
},
|
|
Version: planet.NewVersionConfig(),
|
|
}
|
|
if planet.config.Reconfigure.Satellite != nil {
|
|
planet.config.Reconfigure.Satellite(log, i, &config)
|
|
}
|
|
|
|
// TODO: find source file, to set static path
|
|
_, filename, _, ok := runtime.Caller(0)
|
|
if !ok {
|
|
return xs, errs.New("no caller information")
|
|
}
|
|
storjRoot := strings.TrimSuffix(filename, "/internal/testplanet/planet.go")
|
|
|
|
// TODO: for development only
|
|
config.Marketing.StaticDir = filepath.Join(storjRoot, "web/marketing")
|
|
config.Console.StaticDir = filepath.Join(storjRoot, "web/satellite")
|
|
config.Mail.TemplatePath = filepath.Join(storjRoot, "web/satellite/static/emails")
|
|
|
|
verInfo := planet.NewVersionInfo()
|
|
|
|
peer, err := satellite.New(log, identity, db, &config, verInfo)
|
|
if err != nil {
|
|
return xs, err
|
|
}
|
|
|
|
log.Debug("id=" + peer.ID().String() + " addr=" + peer.Addr())
|
|
xs = append(xs, peer)
|
|
}
|
|
return xs, nil
|
|
}
|
|
|
|
// newStorageNodes initializes storage nodes
|
|
func (planet *Planet) newStorageNodes(count int, whitelistedSatelliteIDs []string) ([]*storagenode.Peer, error) {
|
|
// TODO: move into separate file
|
|
var xs []*storagenode.Peer
|
|
defer func() {
|
|
for _, x := range xs {
|
|
planet.peers = append(planet.peers, closablePeer{peer: x})
|
|
}
|
|
}()
|
|
|
|
for i := 0; i < count; i++ {
|
|
prefix := "storage" + strconv.Itoa(i)
|
|
log := planet.log.Named(prefix)
|
|
storageDir := filepath.Join(planet.directory, prefix)
|
|
|
|
if err := os.MkdirAll(storageDir, 0700); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
identity, err := planet.NewIdentity()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
var db storagenode.DB
|
|
if planet.config.Reconfigure.NewStorageNodeDB != nil {
|
|
db, err = planet.config.Reconfigure.NewStorageNodeDB(i)
|
|
} else {
|
|
db, err = storagenodedb.NewInMemory(log.Named("db"), storageDir)
|
|
}
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
err = db.CreateTables()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
planet.databases = append(planet.databases, db)
|
|
|
|
config := storagenode.Config{
|
|
Server: server.Config{
|
|
Address: "127.0.0.1:0",
|
|
PrivateAddress: "127.0.0.1:0",
|
|
|
|
Config: tlsopts.Config{
|
|
RevocationDBURL: "bolt://" + filepath.Join(storageDir, "revocation.db"),
|
|
UsePeerCAWhitelist: true,
|
|
PeerCAWhitelistPath: planet.whitelistPath,
|
|
PeerIDVersions: "*",
|
|
Extensions: extensions.Config{
|
|
Revocation: false,
|
|
WhitelistSignedLeaf: false,
|
|
},
|
|
},
|
|
},
|
|
Kademlia: kademlia.Config{
|
|
BootstrapBackoffBase: 500 * time.Millisecond,
|
|
BootstrapBackoffMax: 2 * time.Second,
|
|
Alpha: 5,
|
|
DBPath: storageDir, // TODO: replace with master db
|
|
Operator: kademlia.OperatorConfig{
|
|
Email: prefix + "@mail.test",
|
|
Wallet: "0x" + strings.Repeat("00", 20),
|
|
},
|
|
},
|
|
Storage: piecestore.OldConfig{
|
|
Path: "", // TODO: this argument won't be needed with master storagenodedb
|
|
AllocatedDiskSpace: 1 * memory.GB,
|
|
AllocatedBandwidth: memory.TB,
|
|
KBucketRefreshInterval: time.Hour,
|
|
|
|
SatelliteIDRestriction: true,
|
|
WhitelistedSatelliteIDs: strings.Join(whitelistedSatelliteIDs, ","),
|
|
},
|
|
Collector: collector.Config{
|
|
Interval: time.Minute,
|
|
},
|
|
Storage2: piecestore.Config{
|
|
Sender: orders.SenderConfig{
|
|
Interval: time.Hour,
|
|
Timeout: time.Hour,
|
|
},
|
|
Monitor: monitor.Config{
|
|
MinimumBandwidth: 100 * memory.MB,
|
|
MinimumDiskSpace: 100 * memory.MB,
|
|
},
|
|
},
|
|
Version: planet.NewVersionConfig(),
|
|
}
|
|
if planet.config.Reconfigure.StorageNode != nil {
|
|
planet.config.Reconfigure.StorageNode(i, &config)
|
|
}
|
|
|
|
newIPCount := planet.config.Reconfigure.NewIPCount
|
|
if newIPCount > 0 {
|
|
if i >= count-newIPCount {
|
|
config.Server.Address = fmt.Sprintf("127.0.0.%d:0", i+1)
|
|
config.Server.PrivateAddress = fmt.Sprintf("127.0.0.%d:0", i+1)
|
|
}
|
|
}
|
|
|
|
verInfo := planet.NewVersionInfo()
|
|
|
|
peer, err := storagenode.New(log, identity, db, config, verInfo)
|
|
if err != nil {
|
|
return xs, err
|
|
}
|
|
|
|
log.Debug("id=" + peer.ID().String() + " addr=" + peer.Addr())
|
|
xs = append(xs, peer)
|
|
}
|
|
return xs, nil
|
|
}
|
|
|
|
// newBootstrap initializes the bootstrap node
|
|
func (planet *Planet) newBootstrap() (peer *bootstrap.Peer, err error) {
|
|
// TODO: move into separate file
|
|
defer func() {
|
|
planet.peers = append(planet.peers, closablePeer{peer: peer})
|
|
}()
|
|
|
|
prefix := "bootstrap"
|
|
log := planet.log.Named(prefix)
|
|
dbDir := filepath.Join(planet.directory, prefix)
|
|
|
|
if err := os.MkdirAll(dbDir, 0700); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
identity, err := planet.NewIdentity()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
var db bootstrap.DB
|
|
if planet.config.Reconfigure.NewBootstrapDB != nil {
|
|
db, err = planet.config.Reconfigure.NewBootstrapDB(0)
|
|
} else {
|
|
db, err = bootstrapdb.NewInMemory(dbDir)
|
|
}
|
|
|
|
err = db.CreateTables()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
planet.databases = append(planet.databases, db)
|
|
|
|
config := bootstrap.Config{
|
|
Server: server.Config{
|
|
Address: "127.0.0.1:0",
|
|
PrivateAddress: "127.0.0.1:0",
|
|
|
|
Config: tlsopts.Config{
|
|
RevocationDBURL: "bolt://" + filepath.Join(dbDir, "revocation.db"),
|
|
UsePeerCAWhitelist: true,
|
|
PeerCAWhitelistPath: planet.whitelistPath,
|
|
PeerIDVersions: "latest",
|
|
Extensions: extensions.Config{
|
|
Revocation: false,
|
|
WhitelistSignedLeaf: false,
|
|
},
|
|
},
|
|
},
|
|
Kademlia: kademlia.Config{
|
|
BootstrapBackoffBase: 500 * time.Millisecond,
|
|
BootstrapBackoffMax: 2 * time.Second,
|
|
Alpha: 5,
|
|
DBPath: dbDir, // TODO: replace with master db
|
|
Operator: kademlia.OperatorConfig{
|
|
Email: prefix + "@mail.test",
|
|
Wallet: "0x" + strings.Repeat("00", 20),
|
|
},
|
|
},
|
|
Web: bootstrapserver.Config{
|
|
Address: "127.0.0.1:0",
|
|
StaticDir: "./web/bootstrap", // TODO: for development only
|
|
},
|
|
Version: planet.NewVersionConfig(),
|
|
}
|
|
if planet.config.Reconfigure.Bootstrap != nil {
|
|
planet.config.Reconfigure.Bootstrap(0, &config)
|
|
}
|
|
|
|
var verInfo version.Info
|
|
verInfo = planet.NewVersionInfo()
|
|
|
|
peer, err = bootstrap.New(log, identity, db, config, verInfo)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
log.Debug("id=" + peer.ID().String() + " addr=" + peer.Addr())
|
|
|
|
return peer, nil
|
|
}
|
|
|
|
// newVersionControlServer initializes the Versioning Server
|
|
func (planet *Planet) newVersionControlServer() (peer *versioncontrol.Peer, err error) {
|
|
|
|
prefix := "versioncontrol"
|
|
log := planet.log.Named(prefix)
|
|
dbDir := filepath.Join(planet.directory, prefix)
|
|
|
|
if err := os.MkdirAll(dbDir, 0700); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
config := &versioncontrol.Config{
|
|
Address: "127.0.0.1:0",
|
|
Versions: versioncontrol.ServiceVersions{
|
|
Bootstrap: "v0.0.1",
|
|
Satellite: "v0.0.1",
|
|
Storagenode: "v0.0.1",
|
|
Uplink: "v0.0.1",
|
|
Gateway: "v0.0.1",
|
|
},
|
|
}
|
|
peer, err = versioncontrol.New(log, config)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
log.Debug(" addr= " + peer.Addr())
|
|
|
|
return peer, nil
|
|
}
|
|
|
|
// NewVersionInfo returns the Version Info for this planet with tuned metrics.
|
|
func (planet *Planet) NewVersionInfo() version.Info {
|
|
info := version.Info{
|
|
Timestamp: time.Now(),
|
|
CommitHash: "testplanet",
|
|
Version: version.SemVer{
|
|
Major: 0,
|
|
Minor: 0,
|
|
Patch: 1},
|
|
Release: false,
|
|
}
|
|
return info
|
|
}
|
|
|
|
// NewVersionConfig returns the Version Config for this planet with tuned metrics.
|
|
func (planet *Planet) NewVersionConfig() version.Config {
|
|
return version.Config{
|
|
ServerAddress: fmt.Sprintf("http://%s/", planet.VersionControl.Addr()),
|
|
RequestTimeout: time.Second * 15,
|
|
CheckInterval: time.Minute * 5,
|
|
}
|
|
}
|
|
|
|
// Identities returns the identity provider for this planet.
|
|
func (planet *Planet) Identities() *testidentity.Identities {
|
|
return planet.identities
|
|
}
|
|
|
|
// NewIdentity creates a new identity for a node
|
|
func (planet *Planet) NewIdentity() (*identity.FullIdentity, error) {
|
|
return planet.identities.NewIdentity()
|
|
}
|
|
|
|
// NewListener creates a new listener
|
|
func (planet *Planet) NewListener() (net.Listener, error) {
|
|
return net.Listen("tcp", "127.0.0.1:0")
|
|
}
|
|
|
|
// WriteWhitelist writes the pregenerated signer's CA cert to a "CA whitelist", PEM-encoded.
|
|
func (planet *Planet) WriteWhitelist(version storj.IDVersion) (string, error) {
|
|
whitelistPath := filepath.Join(planet.directory, "whitelist.pem")
|
|
signer := testidentity.NewPregeneratedSigner(version)
|
|
err := identity.PeerCAConfig{
|
|
CertPath: whitelistPath,
|
|
}.Save(signer.PeerCA())
|
|
|
|
return whitelistPath, err
|
|
}
|