// Copyright (C) 2018 Storj Labs, Inc. // See LICENSE for copying information. package overlay import ( "context" "crypto/rand" "log" "github.com/gogo/protobuf/proto" "github.com/zeebo/errs" "go.uber.org/zap" "storj.io/storj/pkg/dht" "storj.io/storj/pkg/kademlia" "storj.io/storj/pkg/pb" "storj.io/storj/storage" "storj.io/storj/storage/boltdb" "storj.io/storj/storage/redis" "storj.io/storj/storage/storelogger" ) const ( // OverlayBucket is the string representing the bucket used for a bolt-backed overlay dht cache OverlayBucket = "overlay" ) // ErrNodeNotFound error standardization var ErrNodeNotFound = errs.New("Node not found") // OverlayError creates class of errors for stack traces var OverlayError = errs.Class("Overlay Error") // Cache is used to store overlay data in Redis type Cache struct { DB storage.KeyValueStore DHT dht.DHT } // NewRedisOverlayCache returns a pointer to a new Cache instance with an initialized connection to Redis. func NewRedisOverlayCache(address, password string, db int, DHT dht.DHT) (*Cache, error) { rc, err := redis.NewClient(address, password, db) if err != nil { return nil, err } return &Cache{ DB: rc, DHT: DHT, }, nil } // NewBoltOverlayCache returns a pointer to a new Cache instance with an initialized connection to a Bolt db. func NewBoltOverlayCache(dbPath string, DHT dht.DHT) (*Cache, error) { bc, err := boltdb.New(dbPath, OverlayBucket) if err != nil { return nil, err } return &Cache{ DB: storelogger.New(zap.L(), bc), DHT: DHT, }, nil } // Get looks up the provided nodeID from the overlay cache func (o *Cache) Get(ctx context.Context, key string) (*pb.Node, error) { b, err := o.DB.Get([]byte(key)) if err != nil { return nil, err } if b.IsZero() { // TODO: log? return an error? return nil, nil } na := &pb.Node{} if err := proto.Unmarshal(b, na); err != nil { return nil, err } return na, nil } // GetAll looks up the provided nodeIDs from the overlay cache func (o *Cache) GetAll(ctx context.Context, keys []string) ([]*pb.Node, error) { if len(keys) == 0 { return nil, OverlayError.New("no keys provided") } var ks storage.Keys for _, v := range keys { ks = append(ks, storage.Key(v)) } vs, err := o.DB.GetAll(ks) if err != nil { return nil, err } var ns []*pb.Node for _, v := range vs { if v == nil { ns = append(ns, nil) continue } na := &pb.Node{} err := proto.Unmarshal(v, na) if err != nil { return nil, OverlayError.New("could not unmarshal non-nil node: %v", err) } ns = append(ns, na) } return ns, nil } // Put adds a nodeID to the redis cache with a binary representation of proto defined Node func (o *Cache) Put(nodeID string, value pb.Node) error { data, err := proto.Marshal(&value) if err != nil { return err } return o.DB.Put(kademlia.StringToNodeID(nodeID).Bytes(), data) } // Bootstrap walks the initialized network and populates the cache func (o *Cache) Bootstrap(ctx context.Context) error { nodes, err := o.DHT.GetNodes(ctx, "", 1280) if err != nil { zap.Error(OverlayError.New("Error getting nodes from DHT: %v", err)) } for _, v := range nodes { found, err := o.DHT.FindNode(ctx, kademlia.StringToNodeID(v.Id)) if err != nil { zap.Error(ErrNodeNotFound) } node, err := proto.Marshal(&found) if err != nil { return err } if err := o.DB.Put(kademlia.StringToNodeID(found.Id).Bytes(), node); err != nil { return err } } return err } // Refresh updates the cache db with the current DHT. // We currently do not penalize nodes that are unresponsive, // but should in the future. func (o *Cache) Refresh(ctx context.Context) error { log.Print("starting cache refresh") r, err := randomID() if err != nil { return err } rid := kademlia.NodeID(r) near, err := o.DHT.GetNodes(ctx, rid.String(), 128) if err != nil { return err } for _, node := range near { pinged, err := o.DHT.Ping(ctx, *node) if err != nil { return err } err = o.DB.Put([]byte(pinged.Id), []byte(pinged.Address.Address)) if err != nil { return err } } // TODO: Kademlia hooks to do this automatically rather than at interval nodes, err := o.DHT.GetNodes(ctx, "", 128) if err != nil { return err } for _, node := range nodes { pinged, err := o.DHT.Ping(ctx, *node) if err != nil { zap.Error(ErrNodeNotFound) return err } err = o.DB.Put([]byte(pinged.Id), []byte(pinged.Address.Address)) if err != nil { return err } } return err } // Walk iterates over each node in each bucket to traverse the network func (o *Cache) Walk(ctx context.Context) error { // TODO: This should walk the cache, rather than be a duplicate of refresh return nil } func randomID() ([]byte, error) { result := make([]byte, 64) _, err := rand.Read(result) return result, err }