2019-03-27 10:24:35 +00:00
|
|
|
// Copyright (C) 2019 Storj Labs, Inc.
|
|
|
|
// See LICENSE for copying information.
|
|
|
|
|
|
|
|
package satellitedb
|
|
|
|
|
|
|
|
import (
|
|
|
|
"context"
|
2019-04-01 21:14:58 +01:00
|
|
|
"database/sql"
|
2019-03-28 20:09:23 +00:00
|
|
|
"time"
|
2019-03-27 10:24:35 +00:00
|
|
|
|
2020-01-24 21:57:11 +00:00
|
|
|
"github.com/lib/pq"
|
2019-08-15 20:05:43 +01:00
|
|
|
"github.com/zeebo/errs"
|
2020-01-10 18:53:42 +00:00
|
|
|
"go.uber.org/zap"
|
2019-06-10 15:58:28 +01:00
|
|
|
|
2019-12-27 11:48:47 +00:00
|
|
|
"storj.io/common/pb"
|
|
|
|
"storj.io/common/storj"
|
2020-03-30 10:08:50 +01:00
|
|
|
"storj.io/common/uuid"
|
2020-01-16 22:58:42 +00:00
|
|
|
"storj.io/storj/private/dbutil"
|
2019-11-14 19:46:15 +00:00
|
|
|
"storj.io/storj/private/dbutil/pgutil"
|
2020-06-01 22:40:28 +01:00
|
|
|
"storj.io/storj/private/tagsql"
|
2019-04-04 15:42:01 +01:00
|
|
|
"storj.io/storj/satellite/orders"
|
2020-01-15 02:29:51 +00:00
|
|
|
"storj.io/storj/satellite/satellitedb/dbx"
|
2019-03-27 10:24:35 +00:00
|
|
|
)
|
|
|
|
|
2019-04-01 21:14:58 +01:00
|
|
|
const defaultIntervalSeconds = int(time.Hour / time.Second)
|
|
|
|
|
2019-09-06 15:49:30 +01:00
|
|
|
var (
|
|
|
|
// ErrDifferentStorageNodes is returned when ProcessOrders gets orders from different storage nodes.
|
|
|
|
ErrDifferentStorageNodes = errs.Class("different storage nodes")
|
|
|
|
)
|
|
|
|
|
2019-03-27 10:24:35 +00:00
|
|
|
type ordersDB struct {
|
2019-12-14 02:29:54 +00:00
|
|
|
db *satelliteDB
|
2020-01-15 21:45:17 +00:00
|
|
|
|
|
|
|
reportedRollupsReadBatchSize int
|
2019-03-27 10:24:35 +00:00
|
|
|
}
|
|
|
|
|
2020-02-18 12:03:23 +00:00
|
|
|
// CreateSerialInfo creates serial number entry in database.
|
2019-06-04 12:55:38 +01:00
|
|
|
func (db *ordersDB) CreateSerialInfo(ctx context.Context, serialNumber storj.SerialNumber, bucketID []byte, limitExpiration time.Time) (err error) {
|
|
|
|
defer mon.Task()(&ctx)(&err)
|
2020-02-14 00:03:41 +00:00
|
|
|
|
2019-09-12 18:31:50 +01:00
|
|
|
return db.db.CreateNoReturn_SerialNumber(
|
2019-03-28 20:09:23 +00:00
|
|
|
ctx,
|
|
|
|
dbx.SerialNumber_SerialNumber(serialNumber.Bytes()),
|
|
|
|
dbx.SerialNumber_BucketId(bucketID),
|
2020-02-14 00:03:41 +00:00
|
|
|
dbx.SerialNumber_ExpiresAt(limitExpiration.UTC()),
|
2019-03-28 20:09:23 +00:00
|
|
|
)
|
|
|
|
}
|
|
|
|
|
2019-08-27 18:12:38 +01:00
|
|
|
// DeleteExpiredSerials deletes all expired serials in serial_number and used_serials table.
|
|
|
|
func (db *ordersDB) DeleteExpiredSerials(ctx context.Context, now time.Time) (_ int, err error) {
|
|
|
|
defer mon.Task()(&ctx)(&err)
|
2020-02-14 00:03:41 +00:00
|
|
|
|
|
|
|
count, err := db.db.Delete_SerialNumber_By_ExpiresAt_LessOrEqual(ctx, dbx.SerialNumber_ExpiresAt(now.UTC()))
|
|
|
|
if err != nil {
|
|
|
|
return 0, err
|
|
|
|
}
|
|
|
|
return int(count), nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// DeleteExpiredConsumedSerials deletes all expired serials in the consumed_serials table.
|
|
|
|
func (db *ordersDB) DeleteExpiredConsumedSerials(ctx context.Context, now time.Time) (_ int, err error) {
|
|
|
|
defer mon.Task()(&ctx, now)(&err)
|
|
|
|
|
|
|
|
count, err := db.db.Delete_ConsumedSerial_By_ExpiresAt_LessOrEqual(ctx, dbx.ConsumedSerial_ExpiresAt(now))
|
2019-08-27 18:12:38 +01:00
|
|
|
if err != nil {
|
|
|
|
return 0, err
|
|
|
|
}
|
|
|
|
return int(count), nil
|
|
|
|
}
|
|
|
|
|
2020-02-18 12:03:23 +00:00
|
|
|
// UseSerialNumber creates a used serial number entry in database from an
|
|
|
|
// existing serial number.
|
|
|
|
// It returns the bucket ID associated to serialNumber.
|
2019-06-04 12:55:38 +01:00
|
|
|
func (db *ordersDB) UseSerialNumber(ctx context.Context, serialNumber storj.SerialNumber, storageNodeID storj.NodeID) (_ []byte, err error) {
|
|
|
|
defer mon.Task()(&ctx)(&err)
|
2020-02-14 00:03:41 +00:00
|
|
|
|
2019-04-01 21:14:58 +01:00
|
|
|
statement := db.db.Rebind(
|
|
|
|
`INSERT INTO used_serials (serial_number_id, storage_node_id)
|
|
|
|
SELECT id, ? FROM serial_numbers WHERE serial_number = ?`,
|
|
|
|
)
|
2019-06-04 12:55:38 +01:00
|
|
|
_, err = db.db.ExecContext(ctx, statement, storageNodeID.Bytes(), serialNumber.Bytes())
|
2019-04-01 21:14:58 +01:00
|
|
|
if err != nil {
|
2019-10-18 22:27:57 +01:00
|
|
|
if pgutil.IsConstraintError(err) {
|
2019-04-04 15:42:01 +01:00
|
|
|
return nil, orders.ErrUsingSerialNumber.New("serial number already used")
|
|
|
|
}
|
2019-04-01 21:14:58 +01:00
|
|
|
return nil, err
|
2019-03-27 10:24:35 +00:00
|
|
|
}
|
|
|
|
|
2019-04-01 21:14:58 +01:00
|
|
|
dbxSerialNumber, err := db.db.Find_SerialNumber_By_SerialNumber(
|
|
|
|
ctx,
|
|
|
|
dbx.SerialNumber_SerialNumber(serialNumber.Bytes()),
|
|
|
|
)
|
2019-03-27 10:24:35 +00:00
|
|
|
if err != nil {
|
2019-04-01 21:14:58 +01:00
|
|
|
return nil, err
|
2019-03-27 10:24:35 +00:00
|
|
|
}
|
2019-04-04 15:42:01 +01:00
|
|
|
if dbxSerialNumber == nil {
|
|
|
|
return nil, orders.ErrUsingSerialNumber.New("serial number not found")
|
|
|
|
}
|
2019-04-01 21:14:58 +01:00
|
|
|
return dbxSerialNumber.BucketId, nil
|
|
|
|
}
|
|
|
|
|
2020-02-18 12:03:23 +00:00
|
|
|
// UpdateBucketBandwidthAllocation updates 'allocated' bandwidth for given bucket.
|
2019-06-25 16:58:42 +01:00
|
|
|
func (db *ordersDB) UpdateBucketBandwidthAllocation(ctx context.Context, projectID uuid.UUID, bucketName []byte, action pb.PieceAction, amount int64, intervalStart time.Time) (err error) {
|
2019-06-04 12:55:38 +01:00
|
|
|
defer mon.Task()(&ctx)(&err)
|
2020-02-14 00:03:41 +00:00
|
|
|
|
2020-05-01 14:24:12 +01:00
|
|
|
err = db.db.WithTx(ctx, func(ctx context.Context, tx *dbx.Tx) error {
|
|
|
|
statement := tx.Rebind(
|
|
|
|
`INSERT INTO bucket_bandwidth_rollups (bucket_name, project_id, interval_start, interval_seconds, action, inline, allocated, settled)
|
|
|
|
VALUES (?, ?, ?, ?, ?, ?, ?, ?)
|
|
|
|
ON CONFLICT(bucket_name, project_id, interval_start, action)
|
|
|
|
DO UPDATE SET allocated = bucket_bandwidth_rollups.allocated + ?`,
|
|
|
|
)
|
|
|
|
_, err = tx.Tx.ExecContext(ctx, statement,
|
|
|
|
bucketName, projectID[:], intervalStart.UTC(), defaultIntervalSeconds, action, 0, uint64(amount), 0, uint64(amount),
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
if action == pb.PieceAction_GET {
|
|
|
|
projectInterval := time.Date(intervalStart.Year(), intervalStart.Month(), 1, 0, 0, 0, 0, time.UTC)
|
|
|
|
statement = tx.Rebind(
|
|
|
|
`INSERT INTO project_bandwidth_rollups (project_id, interval_month, egress_allocated)
|
|
|
|
VALUES (?, ?, ?)
|
|
|
|
ON CONFLICT(project_id, interval_month)
|
|
|
|
DO UPDATE SET egress_allocated = project_bandwidth_rollups.egress_allocated + EXCLUDED.egress_allocated::bigint`,
|
|
|
|
)
|
|
|
|
_, err = tx.Tx.ExecContext(ctx, statement, projectID[:], projectInterval, uint64(amount))
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
})
|
2019-03-27 10:24:35 +00:00
|
|
|
|
2019-04-01 21:14:58 +01:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2020-02-18 12:03:23 +00:00
|
|
|
// UpdateBucketBandwidthSettle updates 'settled' bandwidth for given bucket.
|
2019-06-25 16:58:42 +01:00
|
|
|
func (db *ordersDB) UpdateBucketBandwidthSettle(ctx context.Context, projectID uuid.UUID, bucketName []byte, action pb.PieceAction, amount int64, intervalStart time.Time) (err error) {
|
2019-06-04 12:55:38 +01:00
|
|
|
defer mon.Task()(&ctx)(&err)
|
2020-02-14 00:03:41 +00:00
|
|
|
|
2019-04-01 21:14:58 +01:00
|
|
|
statement := db.db.Rebind(
|
2019-04-02 19:21:18 +01:00
|
|
|
`INSERT INTO bucket_bandwidth_rollups (bucket_name, project_id, interval_start, interval_seconds, action, inline, allocated, settled)
|
|
|
|
VALUES (?, ?, ?, ?, ?, ?, ?, ?)
|
|
|
|
ON CONFLICT(bucket_name, project_id, interval_start, action)
|
2019-04-01 21:14:58 +01:00
|
|
|
DO UPDATE SET settled = bucket_bandwidth_rollups.settled + ?`,
|
2019-03-27 10:24:35 +00:00
|
|
|
)
|
2019-06-04 12:55:38 +01:00
|
|
|
_, err = db.db.ExecContext(ctx, statement,
|
2020-02-14 00:03:41 +00:00
|
|
|
bucketName, projectID[:], intervalStart.UTC(), defaultIntervalSeconds, action, 0, 0, uint64(amount), uint64(amount),
|
2019-04-02 19:21:18 +01:00
|
|
|
)
|
2019-03-27 10:24:35 +00:00
|
|
|
if err != nil {
|
2019-04-01 21:14:58 +01:00
|
|
|
return err
|
2019-03-27 10:24:35 +00:00
|
|
|
}
|
2019-04-01 21:14:58 +01:00
|
|
|
return nil
|
2019-03-27 10:24:35 +00:00
|
|
|
}
|
|
|
|
|
2020-02-18 12:03:23 +00:00
|
|
|
// UpdateBucketBandwidthInline updates 'inline' bandwidth for given bucket.
|
2019-06-25 16:58:42 +01:00
|
|
|
func (db *ordersDB) UpdateBucketBandwidthInline(ctx context.Context, projectID uuid.UUID, bucketName []byte, action pb.PieceAction, amount int64, intervalStart time.Time) (err error) {
|
2019-06-04 12:55:38 +01:00
|
|
|
defer mon.Task()(&ctx)(&err)
|
2020-02-14 00:03:41 +00:00
|
|
|
|
2019-04-01 21:14:58 +01:00
|
|
|
statement := db.db.Rebind(
|
2019-04-02 19:21:18 +01:00
|
|
|
`INSERT INTO bucket_bandwidth_rollups (bucket_name, project_id, interval_start, interval_seconds, action, inline, allocated, settled)
|
|
|
|
VALUES (?, ?, ?, ?, ?, ?, ?, ?)
|
|
|
|
ON CONFLICT(bucket_name, project_id, interval_start, action)
|
2019-04-01 21:14:58 +01:00
|
|
|
DO UPDATE SET inline = bucket_bandwidth_rollups.inline + ?`,
|
|
|
|
)
|
2019-06-04 12:55:38 +01:00
|
|
|
_, err = db.db.ExecContext(ctx, statement,
|
2020-02-14 00:03:41 +00:00
|
|
|
bucketName, projectID[:], intervalStart.UTC(), defaultIntervalSeconds, action, uint64(amount), 0, 0, uint64(amount),
|
2019-04-02 19:21:18 +01:00
|
|
|
)
|
2019-03-27 10:24:35 +00:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2019-04-01 21:14:58 +01:00
|
|
|
return nil
|
|
|
|
}
|
2019-03-27 10:24:35 +00:00
|
|
|
|
2020-02-18 12:03:23 +00:00
|
|
|
// UpdateStoragenodeBandwidthSettle updates 'settled' bandwidth for given storage node for the given intervalStart time.
|
2019-06-04 12:55:38 +01:00
|
|
|
func (db *ordersDB) UpdateStoragenodeBandwidthSettle(ctx context.Context, storageNode storj.NodeID, action pb.PieceAction, amount int64, intervalStart time.Time) (err error) {
|
|
|
|
defer mon.Task()(&ctx)(&err)
|
2020-02-14 00:03:41 +00:00
|
|
|
|
2019-04-01 21:14:58 +01:00
|
|
|
statement := db.db.Rebind(
|
2020-01-12 15:20:47 +00:00
|
|
|
`INSERT INTO storagenode_bandwidth_rollups (storagenode_id, interval_start, interval_seconds, action, settled)
|
|
|
|
VALUES (?, ?, ?, ?, ?)
|
2019-04-01 21:14:58 +01:00
|
|
|
ON CONFLICT(storagenode_id, interval_start, action)
|
|
|
|
DO UPDATE SET settled = storagenode_bandwidth_rollups.settled + ?`,
|
|
|
|
)
|
2019-06-04 12:55:38 +01:00
|
|
|
_, err = db.db.ExecContext(ctx, statement,
|
2020-02-14 00:03:41 +00:00
|
|
|
storageNode.Bytes(), intervalStart.UTC(), defaultIntervalSeconds, action, uint64(amount), uint64(amount),
|
2019-04-02 19:21:18 +01:00
|
|
|
)
|
2019-03-27 10:24:35 +00:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2019-04-01 21:14:58 +01:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2020-02-18 12:03:23 +00:00
|
|
|
// GetBucketBandwidth gets total bucket bandwidth from period of time.
|
2019-06-25 16:58:42 +01:00
|
|
|
func (db *ordersDB) GetBucketBandwidth(ctx context.Context, projectID uuid.UUID, bucketName []byte, from, to time.Time) (_ int64, err error) {
|
2019-06-04 12:55:38 +01:00
|
|
|
defer mon.Task()(&ctx)(&err)
|
2020-02-14 00:03:41 +00:00
|
|
|
|
2019-04-01 21:14:58 +01:00
|
|
|
var sum *int64
|
2019-04-02 19:21:18 +01:00
|
|
|
query := `SELECT SUM(settled) FROM bucket_bandwidth_rollups WHERE bucket_name = ? AND project_id = ? AND interval_start > ? AND interval_start <= ?`
|
2020-02-14 00:03:41 +00:00
|
|
|
err = db.db.QueryRow(ctx, db.db.Rebind(query), bucketName, projectID[:], from.UTC(), to.UTC()).Scan(&sum)
|
2019-04-01 21:14:58 +01:00
|
|
|
if err == sql.ErrNoRows || sum == nil {
|
|
|
|
return 0, nil
|
|
|
|
}
|
2020-01-15 21:45:17 +00:00
|
|
|
return *sum, Error.Wrap(err)
|
2019-04-01 21:14:58 +01:00
|
|
|
}
|
2019-03-27 10:24:35 +00:00
|
|
|
|
2020-02-18 12:03:23 +00:00
|
|
|
// GetStorageNodeBandwidth gets total storage node bandwidth from period of time.
|
2019-06-04 12:55:38 +01:00
|
|
|
func (db *ordersDB) GetStorageNodeBandwidth(ctx context.Context, nodeID storj.NodeID, from, to time.Time) (_ int64, err error) {
|
|
|
|
defer mon.Task()(&ctx)(&err)
|
2020-02-14 00:03:41 +00:00
|
|
|
|
2019-04-01 21:14:58 +01:00
|
|
|
var sum *int64
|
|
|
|
query := `SELECT SUM(settled) FROM storagenode_bandwidth_rollups WHERE storagenode_id = ? AND interval_start > ? AND interval_start <= ?`
|
2020-02-14 00:03:41 +00:00
|
|
|
err = db.db.QueryRow(ctx, db.db.Rebind(query), nodeID.Bytes(), from.UTC(), to.UTC()).Scan(&sum)
|
2019-04-01 21:14:58 +01:00
|
|
|
if err == sql.ErrNoRows || sum == nil {
|
|
|
|
return 0, nil
|
|
|
|
}
|
|
|
|
return *sum, err
|
|
|
|
}
|
2019-03-27 10:24:35 +00:00
|
|
|
|
2020-02-18 12:03:23 +00:00
|
|
|
// UnuseSerialNumber removes pair serial number -> storage node id from database.
|
2019-06-04 12:55:38 +01:00
|
|
|
func (db *ordersDB) UnuseSerialNumber(ctx context.Context, serialNumber storj.SerialNumber, storageNodeID storj.NodeID) (err error) {
|
|
|
|
defer mon.Task()(&ctx)(&err)
|
2020-02-14 00:03:41 +00:00
|
|
|
|
2019-04-01 21:14:58 +01:00
|
|
|
statement := `DELETE FROM used_serials WHERE storage_node_id = ? AND
|
|
|
|
serial_number_id IN (SELECT id FROM serial_numbers WHERE serial_number = ?)`
|
2019-06-04 12:55:38 +01:00
|
|
|
_, err = db.db.ExecContext(ctx, db.db.Rebind(statement), storageNodeID.Bytes(), serialNumber.Bytes())
|
2019-04-01 21:14:58 +01:00
|
|
|
return err
|
2019-03-27 10:24:35 +00:00
|
|
|
}
|
2019-08-15 20:05:43 +01:00
|
|
|
|
2020-02-14 00:03:41 +00:00
|
|
|
// ProcessOrders take a list of order requests and inserts them into the pending serials queue.
|
2019-09-06 15:49:30 +01:00
|
|
|
//
|
|
|
|
// ProcessOrders requires that all orders come from the same storage node.
|
2020-02-14 00:03:41 +00:00
|
|
|
func (db *ordersDB) ProcessOrders(ctx context.Context, requests []*orders.ProcessOrderRequest) (responses []*orders.ProcessOrderResponse, err error) {
|
2019-08-15 20:05:43 +01:00
|
|
|
defer mon.Task()(&ctx)(&err)
|
|
|
|
|
|
|
|
if len(requests) == 0 {
|
2020-01-15 21:45:17 +00:00
|
|
|
return nil, nil
|
2019-08-15 20:05:43 +01:00
|
|
|
}
|
|
|
|
|
2019-09-06 15:49:30 +01:00
|
|
|
// check that all requests are from the same storage node
|
|
|
|
storageNodeID := requests[0].OrderLimit.StorageNodeId
|
|
|
|
for _, req := range requests[1:] {
|
|
|
|
if req.OrderLimit.StorageNodeId != storageNodeID {
|
|
|
|
return nil, ErrDifferentStorageNodes.New("requests from different storage nodes %v and %v", storageNodeID, req.OrderLimit.StorageNodeId)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-01-24 21:57:11 +00:00
|
|
|
// Do a read first to get all the project id/bucket ids. We could combine this with the
|
|
|
|
// upsert below by doing a join, but there isn't really any need for special consistency
|
|
|
|
// semantics between these two queries, and it should make things easier on the database
|
|
|
|
// (particularly cockroachDB) to have the freedom to perform them separately.
|
|
|
|
//
|
|
|
|
// We don't expect the serial_number -> bucket_id relationship ever to change, as long as a
|
|
|
|
// serial_number exists. There is a possibility of a serial_number being deleted between
|
|
|
|
// this query and the next, but that is ok too (rows in reported_serials may end up having
|
|
|
|
// serial numbers that no longer exist in serial_numbers, but that shouldn't break
|
|
|
|
// anything.)
|
|
|
|
bucketIDs, err := func() (bucketIDs [][]byte, err error) {
|
|
|
|
bucketIDs = make([][]byte, len(requests))
|
|
|
|
serialNums := make([][]byte, len(requests))
|
|
|
|
for i, request := range requests {
|
|
|
|
serialNums[i] = request.Order.SerialNumber.Bytes()
|
|
|
|
}
|
|
|
|
rows, err := db.db.QueryContext(ctx, `
|
|
|
|
SELECT request.i, sn.bucket_id
|
|
|
|
FROM
|
|
|
|
serial_numbers sn,
|
|
|
|
unnest($1::bytea[]) WITH ORDINALITY AS request(serial_number, i)
|
|
|
|
WHERE request.serial_number = sn.serial_number
|
|
|
|
`, pq.ByteaArray(serialNums))
|
|
|
|
if err != nil {
|
|
|
|
return nil, Error.Wrap(err)
|
|
|
|
}
|
2020-03-04 07:39:37 +00:00
|
|
|
defer func() { err = errs.Combine(err, rows.Err(), rows.Close()) }()
|
2020-01-24 21:57:11 +00:00
|
|
|
for rows.Next() {
|
|
|
|
var index int
|
|
|
|
var bucketID []byte
|
|
|
|
err = rows.Scan(&index, &bucketID)
|
2020-01-15 21:45:17 +00:00
|
|
|
if err != nil {
|
2020-01-24 21:57:11 +00:00
|
|
|
return nil, Error.Wrap(err)
|
2020-01-15 21:45:17 +00:00
|
|
|
}
|
2020-01-24 21:57:11 +00:00
|
|
|
bucketIDs[index-1] = bucketID
|
2020-01-15 21:45:17 +00:00
|
|
|
}
|
2020-01-24 21:57:11 +00:00
|
|
|
return bucketIDs, nil
|
|
|
|
}()
|
2020-01-15 21:45:17 +00:00
|
|
|
if err != nil {
|
|
|
|
return nil, Error.Wrap(err)
|
|
|
|
}
|
|
|
|
|
2020-01-16 22:58:42 +00:00
|
|
|
// perform all of the upserts into reported serials table
|
2020-01-24 21:57:11 +00:00
|
|
|
expiresAtArray := make([]time.Time, 0, len(requests))
|
|
|
|
bucketIDArray := make([][]byte, 0, len(requests))
|
|
|
|
actionArray := make([]pb.PieceAction, 0, len(requests))
|
|
|
|
serialNumArray := make([][]byte, 0, len(requests))
|
|
|
|
settledArray := make([]int64, 0, len(requests))
|
|
|
|
|
2020-02-18 23:00:13 +00:00
|
|
|
// remove duplicate bucket_id, serial_number pairs sent in the same request.
|
|
|
|
// postgres will complain.
|
|
|
|
type requestKey struct {
|
|
|
|
BucketID string
|
|
|
|
SerialNumber storj.SerialNumber
|
|
|
|
}
|
|
|
|
seenRequests := make(map[requestKey]struct{})
|
|
|
|
|
2020-01-24 21:57:11 +00:00
|
|
|
for i, request := range requests {
|
|
|
|
if bucketIDs[i] == nil {
|
2020-01-15 21:45:17 +00:00
|
|
|
responses = append(responses, &orders.ProcessOrderResponse{
|
|
|
|
SerialNumber: request.Order.SerialNumber,
|
2020-01-24 21:57:11 +00:00
|
|
|
Status: pb.SettlementResponse_REJECTED,
|
2020-01-15 21:45:17 +00:00
|
|
|
})
|
2020-01-24 21:57:11 +00:00
|
|
|
continue
|
2019-08-15 20:05:43 +01:00
|
|
|
}
|
2020-02-18 23:00:13 +00:00
|
|
|
|
|
|
|
// Filter duplicate requests and reject them.
|
|
|
|
key := requestKey{
|
|
|
|
BucketID: string(bucketIDs[i]),
|
|
|
|
SerialNumber: request.Order.SerialNumber,
|
|
|
|
}
|
|
|
|
if _, seen := seenRequests[key]; seen {
|
|
|
|
responses = append(responses, &orders.ProcessOrderResponse{
|
|
|
|
SerialNumber: request.Order.SerialNumber,
|
|
|
|
Status: pb.SettlementResponse_REJECTED,
|
|
|
|
})
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
seenRequests[key] = struct{}{}
|
|
|
|
|
2020-02-14 00:03:41 +00:00
|
|
|
expiresAtArray = append(expiresAtArray, request.OrderLimit.OrderExpiration)
|
2020-01-24 21:57:11 +00:00
|
|
|
bucketIDArray = append(bucketIDArray, bucketIDs[i])
|
|
|
|
actionArray = append(actionArray, request.OrderLimit.Action)
|
|
|
|
serialNumCopy := request.Order.SerialNumber
|
|
|
|
serialNumArray = append(serialNumArray, serialNumCopy[:])
|
|
|
|
settledArray = append(settledArray, request.Order.Amount)
|
|
|
|
|
|
|
|
responses = append(responses, &orders.ProcessOrderResponse{
|
|
|
|
SerialNumber: request.Order.SerialNumber,
|
|
|
|
Status: pb.SettlementResponse_ACCEPTED,
|
|
|
|
})
|
|
|
|
}
|
2019-08-29 16:14:10 +01:00
|
|
|
|
2020-01-24 21:57:11 +00:00
|
|
|
var stmt string
|
|
|
|
switch db.db.implementation {
|
|
|
|
case dbutil.Postgres:
|
|
|
|
stmt = `
|
2020-02-14 00:03:41 +00:00
|
|
|
INSERT INTO pending_serial_queue (
|
|
|
|
storage_node_id, bucket_id, serial_number, action, settled, expires_at
|
2020-01-24 21:57:11 +00:00
|
|
|
)
|
2020-02-14 00:03:41 +00:00
|
|
|
SELECT
|
|
|
|
$1::bytea,
|
|
|
|
unnest($2::bytea[]),
|
|
|
|
unnest($3::bytea[]),
|
|
|
|
unnest($4::integer[]),
|
|
|
|
unnest($5::bigint[]),
|
|
|
|
unnest($6::timestamptz[])
|
|
|
|
ON CONFLICT ( storage_node_id, bucket_id, serial_number )
|
2020-01-24 21:57:11 +00:00
|
|
|
DO UPDATE SET
|
2020-02-14 00:03:41 +00:00
|
|
|
action = EXCLUDED.action,
|
2020-01-24 21:57:11 +00:00
|
|
|
settled = EXCLUDED.settled,
|
2020-02-14 00:03:41 +00:00
|
|
|
expires_at = EXCLUDED.expires_at
|
2020-01-24 21:57:11 +00:00
|
|
|
`
|
|
|
|
case dbutil.Cockroach:
|
|
|
|
stmt = `
|
2020-02-14 00:03:41 +00:00
|
|
|
UPSERT INTO pending_serial_queue (
|
|
|
|
storage_node_id, bucket_id, serial_number, action, settled, expires_at
|
2020-01-24 21:57:11 +00:00
|
|
|
)
|
2020-02-14 00:03:41 +00:00
|
|
|
SELECT
|
|
|
|
$1::bytea,
|
|
|
|
unnest($2::bytea[]),
|
|
|
|
unnest($3::bytea[]),
|
|
|
|
unnest($4::integer[]),
|
|
|
|
unnest($5::bigint[]),
|
|
|
|
unnest($6::timestamptz[])
|
2020-01-24 21:57:11 +00:00
|
|
|
`
|
|
|
|
default:
|
|
|
|
return nil, Error.New("invalid dbType: %v", db.db.driver)
|
|
|
|
}
|
|
|
|
|
|
|
|
_, err = db.db.ExecContext(ctx, stmt,
|
|
|
|
storageNodeID.Bytes(),
|
|
|
|
pq.ByteaArray(bucketIDArray),
|
|
|
|
pq.ByteaArray(serialNumArray),
|
2020-02-14 00:03:41 +00:00
|
|
|
pq.Array(actionArray),
|
2020-01-24 21:57:11 +00:00
|
|
|
pq.Array(settledArray),
|
2020-02-14 00:03:41 +00:00
|
|
|
pq.Array(expiresAtArray),
|
2020-01-24 21:57:11 +00:00
|
|
|
)
|
2020-01-15 21:45:17 +00:00
|
|
|
if err != nil {
|
|
|
|
return nil, Error.Wrap(err)
|
|
|
|
}
|
|
|
|
return responses, nil
|
|
|
|
}
|
2019-08-29 16:14:10 +01:00
|
|
|
|
2020-01-15 21:45:17 +00:00
|
|
|
//
|
|
|
|
// transaction/batch methods
|
|
|
|
//
|
2019-08-15 20:05:43 +01:00
|
|
|
|
2020-01-15 21:45:17 +00:00
|
|
|
type ordersDBTx struct {
|
|
|
|
tx *dbx.Tx
|
2020-02-14 00:03:41 +00:00
|
|
|
db *satelliteDB
|
2020-01-15 21:45:17 +00:00
|
|
|
log *zap.Logger
|
|
|
|
}
|
2019-08-15 20:05:43 +01:00
|
|
|
|
2020-01-16 18:02:15 +00:00
|
|
|
func (db *ordersDB) WithTransaction(ctx context.Context, cb func(ctx context.Context, tx orders.Transaction) error) (err error) {
|
2020-01-15 21:45:17 +00:00
|
|
|
defer mon.Task()(&ctx)(&err)
|
|
|
|
|
|
|
|
return db.db.WithTx(ctx, func(ctx context.Context, tx *dbx.Tx) error {
|
2020-02-14 00:03:41 +00:00
|
|
|
return cb(ctx, &ordersDBTx{tx: tx, db: db.db, log: db.db.log})
|
2020-01-15 21:45:17 +00:00
|
|
|
})
|
2019-08-15 20:05:43 +01:00
|
|
|
}
|
2020-01-10 18:53:42 +00:00
|
|
|
|
2020-01-15 21:45:17 +00:00
|
|
|
func (tx *ordersDBTx) UpdateBucketBandwidthBatch(ctx context.Context, intervalStart time.Time, rollups []orders.BucketBandwidthRollup) (err error) {
|
2020-01-14 00:36:12 +00:00
|
|
|
defer mon.Task()(&ctx)(&err)
|
|
|
|
|
2020-01-10 18:53:42 +00:00
|
|
|
if len(rollups) == 0 {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2020-01-15 21:45:17 +00:00
|
|
|
orders.SortBucketBandwidthRollups(rollups)
|
|
|
|
|
2020-01-10 18:53:42 +00:00
|
|
|
intervalStart = intervalStart.UTC()
|
|
|
|
intervalStart = time.Date(intervalStart.Year(), intervalStart.Month(), intervalStart.Day(), intervalStart.Hour(), 0, 0, 0, time.UTC)
|
|
|
|
|
2020-01-24 08:47:15 +00:00
|
|
|
var bucketNames [][]byte
|
|
|
|
var projectIDs [][]byte
|
|
|
|
var actionSlice []int32
|
|
|
|
var inlineSlice []int64
|
|
|
|
var allocatedSlice []int64
|
|
|
|
var settledSlice []int64
|
2020-05-01 14:24:12 +01:00
|
|
|
var projectRUMap map[string]int64 = make(map[string]int64)
|
2020-01-24 08:47:15 +00:00
|
|
|
|
|
|
|
for _, rollup := range rollups {
|
|
|
|
rollup := rollup
|
|
|
|
bucketNames = append(bucketNames, []byte(rollup.BucketName))
|
|
|
|
projectIDs = append(projectIDs, rollup.ProjectID[:])
|
|
|
|
actionSlice = append(actionSlice, int32(rollup.Action))
|
|
|
|
inlineSlice = append(inlineSlice, rollup.Inline)
|
|
|
|
allocatedSlice = append(allocatedSlice, rollup.Allocated)
|
|
|
|
settledSlice = append(settledSlice, rollup.Settled)
|
2020-05-01 14:24:12 +01:00
|
|
|
|
|
|
|
if rollup.Action == pb.PieceAction_GET {
|
|
|
|
projectRUMap[rollup.ProjectID.String()] += rollup.Allocated
|
|
|
|
}
|
2020-01-10 18:53:42 +00:00
|
|
|
}
|
2020-01-14 00:36:12 +00:00
|
|
|
|
2020-01-24 08:47:15 +00:00
|
|
|
_, err = tx.tx.Tx.ExecContext(ctx, `
|
|
|
|
INSERT INTO bucket_bandwidth_rollups (
|
|
|
|
bucket_name, project_id,
|
|
|
|
interval_start, interval_seconds,
|
|
|
|
action, inline, allocated, settled)
|
|
|
|
SELECT
|
|
|
|
unnest($1::bytea[]), unnest($2::bytea[]),
|
|
|
|
$3, $4,
|
2020-01-31 18:00:00 +00:00
|
|
|
unnest($5::bigint[]), unnest($6::bigint[]), unnest($7::bigint[]), unnest($8::bigint[])
|
2020-01-24 08:47:15 +00:00
|
|
|
ON CONFLICT(bucket_name, project_id, interval_start, action)
|
|
|
|
DO UPDATE SET
|
|
|
|
allocated = bucket_bandwidth_rollups.allocated + EXCLUDED.allocated,
|
|
|
|
inline = bucket_bandwidth_rollups.inline + EXCLUDED.inline,
|
|
|
|
settled = bucket_bandwidth_rollups.settled + EXCLUDED.settled`,
|
|
|
|
pq.ByteaArray(bucketNames), pq.ByteaArray(projectIDs),
|
|
|
|
intervalStart, defaultIntervalSeconds,
|
|
|
|
pq.Array(actionSlice), pq.Array(inlineSlice), pq.Array(allocatedSlice), pq.Array(settledSlice))
|
2020-01-10 18:53:42 +00:00
|
|
|
if err != nil {
|
2020-01-15 21:45:17 +00:00
|
|
|
tx.log.Error("Bucket bandwidth rollup batch flush failed.", zap.Error(err))
|
2020-01-10 18:53:42 +00:00
|
|
|
}
|
2020-05-01 14:24:12 +01:00
|
|
|
|
|
|
|
var projectRUIDs [][]byte
|
|
|
|
var projectRUAllocated []int64
|
|
|
|
projectInterval := time.Date(intervalStart.Year(), intervalStart.Month(), 1, intervalStart.Hour(), 0, 0, 0, time.UTC)
|
|
|
|
|
|
|
|
for k, v := range projectRUMap {
|
|
|
|
projectID, err := uuid.FromString(k)
|
|
|
|
if err != nil {
|
|
|
|
tx.log.Error("Could not parse project UUID.", zap.Error(err))
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
projectRUIDs = append(projectRUIDs, projectID[:])
|
|
|
|
projectRUAllocated = append(projectRUAllocated, v)
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(projectRUIDs) > 0 {
|
|
|
|
_, err = tx.tx.Tx.ExecContext(ctx, `
|
|
|
|
INSERT INTO project_bandwidth_rollups(project_id, interval_month, egress_allocated)
|
|
|
|
SELECT unnest($1::bytea[]), $2, unnest($3::bigint[])
|
2020-05-30 17:02:01 +01:00
|
|
|
ON CONFLICT(project_id, interval_month)
|
2020-05-01 14:24:12 +01:00
|
|
|
DO UPDATE SET egress_allocated = project_bandwidth_rollups.egress_allocated + EXCLUDED.egress_allocated::bigint;
|
|
|
|
`,
|
|
|
|
pq.ByteaArray(projectRUIDs), projectInterval, pq.Array(projectRUAllocated))
|
|
|
|
if err != nil {
|
|
|
|
tx.log.Error("Project bandwidth rollup batch flush failed.", zap.Error(err))
|
|
|
|
}
|
|
|
|
}
|
2020-01-10 18:53:42 +00:00
|
|
|
return err
|
|
|
|
}
|
2020-01-15 21:45:17 +00:00
|
|
|
|
|
|
|
func (tx *ordersDBTx) UpdateStoragenodeBandwidthBatch(ctx context.Context, intervalStart time.Time, rollups []orders.StoragenodeBandwidthRollup) (err error) {
|
|
|
|
defer mon.Task()(&ctx)(&err)
|
|
|
|
|
|
|
|
if len(rollups) == 0 {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
orders.SortStoragenodeBandwidthRollups(rollups)
|
|
|
|
|
2020-01-24 08:47:15 +00:00
|
|
|
var storageNodeIDs []storj.NodeID
|
|
|
|
var actionSlice []int32
|
|
|
|
var allocatedSlice []int64
|
|
|
|
var settledSlice []int64
|
2020-01-15 21:45:17 +00:00
|
|
|
|
|
|
|
intervalStart = intervalStart.UTC()
|
|
|
|
intervalStart = time.Date(intervalStart.Year(), intervalStart.Month(), intervalStart.Day(), intervalStart.Hour(), 0, 0, 0, time.UTC)
|
|
|
|
|
2020-01-24 08:47:15 +00:00
|
|
|
for i := range rollups {
|
|
|
|
rollup := &rollups[i]
|
|
|
|
storageNodeIDs = append(storageNodeIDs, rollup.NodeID)
|
|
|
|
actionSlice = append(actionSlice, int32(rollup.Action))
|
|
|
|
allocatedSlice = append(allocatedSlice, rollup.Allocated)
|
|
|
|
settledSlice = append(settledSlice, rollup.Settled)
|
2020-01-15 21:45:17 +00:00
|
|
|
}
|
|
|
|
|
2020-01-24 08:47:15 +00:00
|
|
|
_, err = tx.tx.Tx.ExecContext(ctx, `
|
|
|
|
INSERT INTO storagenode_bandwidth_rollups(
|
|
|
|
storagenode_id,
|
|
|
|
interval_start, interval_seconds,
|
|
|
|
action, allocated, settled)
|
|
|
|
SELECT
|
|
|
|
unnest($1::bytea[]),
|
|
|
|
$2, $3,
|
2020-01-31 18:00:00 +00:00
|
|
|
unnest($4::bigint[]), unnest($5::bigint[]), unnest($6::bigint[])
|
2020-01-24 08:47:15 +00:00
|
|
|
ON CONFLICT(storagenode_id, interval_start, action)
|
|
|
|
DO UPDATE SET
|
|
|
|
allocated = storagenode_bandwidth_rollups.allocated + EXCLUDED.allocated,
|
|
|
|
settled = storagenode_bandwidth_rollups.settled + EXCLUDED.settled`,
|
|
|
|
postgresNodeIDList(storageNodeIDs),
|
|
|
|
intervalStart, defaultIntervalSeconds,
|
|
|
|
pq.Array(actionSlice), pq.Array(allocatedSlice), pq.Array(settledSlice))
|
2020-01-15 21:45:17 +00:00
|
|
|
if err != nil {
|
|
|
|
tx.log.Error("Storagenode bandwidth rollup batch flush failed.", zap.Error(err))
|
|
|
|
}
|
2020-01-24 08:47:15 +00:00
|
|
|
|
2020-01-15 21:45:17 +00:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2020-03-02 18:47:22 +00:00
|
|
|
// CreateConsumedSerialsBatch creates a batch of consumed serial entries.
|
2020-02-14 00:03:41 +00:00
|
|
|
func (tx *ordersDBTx) CreateConsumedSerialsBatch(ctx context.Context, consumedSerials []orders.ConsumedSerial) (err error) {
|
|
|
|
defer mon.Task()(&ctx)(&err)
|
|
|
|
|
|
|
|
if len(consumedSerials) == 0 {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
var storageNodeIDSlice [][]byte
|
|
|
|
var serialNumberSlice [][]byte
|
|
|
|
var expiresAtSlice []time.Time
|
|
|
|
|
|
|
|
for _, consumedSerial := range consumedSerials {
|
|
|
|
consumedSerial := consumedSerial
|
|
|
|
storageNodeIDSlice = append(storageNodeIDSlice, consumedSerial.NodeID.Bytes())
|
|
|
|
serialNumberSlice = append(serialNumberSlice, consumedSerial.SerialNumber.Bytes())
|
|
|
|
expiresAtSlice = append(expiresAtSlice, consumedSerial.ExpiresAt)
|
|
|
|
}
|
|
|
|
|
|
|
|
var stmt string
|
|
|
|
switch tx.db.implementation {
|
|
|
|
case dbutil.Postgres:
|
|
|
|
stmt = `
|
|
|
|
INSERT INTO consumed_serials (
|
|
|
|
storage_node_id, serial_number, expires_at
|
|
|
|
)
|
|
|
|
SELECT unnest($1::bytea[]), unnest($2::bytea[]), unnest($3::timestamptz[])
|
|
|
|
ON CONFLICT ( storage_node_id, serial_number ) DO NOTHING
|
|
|
|
`
|
|
|
|
case dbutil.Cockroach:
|
|
|
|
stmt = `
|
|
|
|
UPSERT INTO consumed_serials (
|
|
|
|
storage_node_id, serial_number, expires_at
|
|
|
|
)
|
|
|
|
SELECT unnest($1::bytea[]), unnest($2::bytea[]), unnest($3::timestamptz[])
|
|
|
|
`
|
|
|
|
default:
|
|
|
|
return Error.New("invalid dbType: %v", tx.db.driver)
|
|
|
|
}
|
|
|
|
|
|
|
|
_, err = tx.tx.Tx.ExecContext(ctx, stmt,
|
|
|
|
pq.ByteaArray(storageNodeIDSlice),
|
|
|
|
pq.ByteaArray(serialNumberSlice),
|
|
|
|
pq.Array(expiresAtSlice),
|
|
|
|
)
|
|
|
|
return Error.Wrap(err)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (tx *ordersDBTx) HasConsumedSerial(ctx context.Context, nodeID storj.NodeID, serialNumber storj.SerialNumber) (exists bool, err error) {
|
|
|
|
defer mon.Task()(&ctx)(&err)
|
|
|
|
|
|
|
|
exists, err = tx.tx.Has_ConsumedSerial_By_StorageNodeId_And_SerialNumber(ctx,
|
|
|
|
dbx.ConsumedSerial_StorageNodeId(nodeID.Bytes()),
|
|
|
|
dbx.ConsumedSerial_SerialNumber(serialNumber.Bytes()))
|
|
|
|
return exists, Error.Wrap(err)
|
|
|
|
}
|
|
|
|
|
|
|
|
//
|
|
|
|
// transaction/batch methods
|
|
|
|
//
|
|
|
|
|
2020-03-02 18:47:22 +00:00
|
|
|
type rawPendingSerial struct {
|
|
|
|
nodeID []byte
|
|
|
|
bucketID []byte
|
|
|
|
serialNumber []byte
|
|
|
|
}
|
|
|
|
|
2020-02-14 00:03:41 +00:00
|
|
|
type ordersDBQueue struct {
|
2020-06-01 22:40:28 +01:00
|
|
|
impl dbutil.Implementation
|
|
|
|
log *zap.Logger
|
|
|
|
tx tagsql.Tx
|
2020-02-14 00:03:41 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
func (db *ordersDB) WithQueue(ctx context.Context, cb func(ctx context.Context, queue orders.Queue) error) (err error) {
|
2020-01-15 21:45:17 +00:00
|
|
|
defer mon.Task()(&ctx)(&err)
|
|
|
|
|
2020-06-01 22:40:28 +01:00
|
|
|
return Error.Wrap(db.db.WithTx(ctx, func(ctx context.Context, tx *dbx.Tx) error {
|
|
|
|
return cb(ctx, &ordersDBQueue{
|
|
|
|
impl: db.db.implementation,
|
|
|
|
log: db.db.log,
|
|
|
|
tx: tx.Tx,
|
|
|
|
})
|
|
|
|
}))
|
2020-01-15 21:45:17 +00:00
|
|
|
}
|
2020-02-14 00:03:41 +00:00
|
|
|
|
2020-03-02 18:47:22 +00:00
|
|
|
func (queue *ordersDBQueue) GetPendingSerialsBatch(ctx context.Context, size int) (pendingSerials []orders.PendingSerial, done bool, err error) {
|
2020-02-14 00:03:41 +00:00
|
|
|
defer mon.Task()(&ctx)(&err)
|
|
|
|
|
2020-06-01 22:40:28 +01:00
|
|
|
// TODO: no idea of this query makes sense on cockroach. it may do a terrible job with it.
|
|
|
|
// but it's blazing fast on postgres and that's where we have the problem! :D :D :D
|
2020-03-02 18:47:22 +00:00
|
|
|
|
2020-01-19 13:42:08 +00:00
|
|
|
var rows tagsql.Rows
|
2020-06-01 22:40:28 +01:00
|
|
|
switch queue.impl {
|
|
|
|
case dbutil.Postgres:
|
|
|
|
rows, err = queue.tx.Query(ctx, `
|
|
|
|
DELETE
|
|
|
|
FROM pending_serial_queue
|
2020-03-02 18:47:22 +00:00
|
|
|
WHERE
|
2020-06-01 22:40:28 +01:00
|
|
|
ctid = any (array(
|
|
|
|
SELECT
|
|
|
|
ctid
|
|
|
|
FROM pending_serial_queue
|
|
|
|
LIMIT $1
|
|
|
|
))
|
|
|
|
RETURNING storage_node_id, bucket_id, serial_number, action, settled, expires_at, (
|
|
|
|
coalesce((
|
|
|
|
SELECT 1
|
|
|
|
FROM consumed_serials
|
|
|
|
WHERE
|
|
|
|
consumed_serials.storage_node_id = pending_serial_queue.storage_node_id
|
|
|
|
AND consumed_serials.serial_number = pending_serial_queue.serial_number
|
|
|
|
), 0))
|
|
|
|
`, size)
|
|
|
|
case dbutil.Cockroach:
|
|
|
|
rows, err = queue.tx.Query(ctx, `
|
|
|
|
DELETE
|
|
|
|
FROM pending_serial_queue
|
|
|
|
WHERE
|
|
|
|
(storage_node_id, bucket_id, serial_number) = any (array(
|
|
|
|
SELECT
|
|
|
|
(storage_node_id, bucket_id, serial_number)
|
|
|
|
FROM pending_serial_queue
|
|
|
|
LIMIT $1
|
|
|
|
))
|
|
|
|
RETURNING storage_node_id, bucket_id, serial_number, action, settled, expires_at, (
|
|
|
|
coalesce((
|
|
|
|
SELECT 1
|
|
|
|
FROM consumed_serials
|
|
|
|
WHERE
|
|
|
|
consumed_serials.storage_node_id = pending_serial_queue.storage_node_id
|
|
|
|
AND consumed_serials.serial_number = pending_serial_queue.serial_number
|
|
|
|
), 0))
|
|
|
|
`, size)
|
|
|
|
default:
|
|
|
|
return nil, false, Error.New("unhandled implementation")
|
|
|
|
}
|
2020-02-14 00:03:41 +00:00
|
|
|
if err != nil {
|
2020-03-02 18:47:22 +00:00
|
|
|
return nil, false, Error.Wrap(err)
|
|
|
|
}
|
|
|
|
defer func() { err = errs.Combine(err, Error.Wrap(rows.Close())) }()
|
|
|
|
|
|
|
|
for rows.Next() {
|
|
|
|
var consumed int
|
|
|
|
var rawPending rawPendingSerial
|
|
|
|
var pendingSerial orders.PendingSerial
|
|
|
|
|
|
|
|
err := rows.Scan(
|
|
|
|
&rawPending.nodeID,
|
|
|
|
&rawPending.bucketID,
|
|
|
|
&rawPending.serialNumber,
|
|
|
|
&pendingSerial.Action,
|
|
|
|
&pendingSerial.Settled,
|
|
|
|
&pendingSerial.ExpiresAt,
|
|
|
|
&consumed,
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
return nil, false, Error.Wrap(err)
|
|
|
|
}
|
|
|
|
|
|
|
|
size--
|
2020-02-14 00:03:41 +00:00
|
|
|
|
2020-03-02 18:47:22 +00:00
|
|
|
if consumed != 0 {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
pendingSerial.NodeID, err = storj.NodeIDFromBytes(rawPending.nodeID)
|
2020-02-14 00:03:41 +00:00
|
|
|
if err != nil {
|
|
|
|
queue.log.Error("Invalid storage node id in pending serials queue",
|
2020-03-02 18:47:22 +00:00
|
|
|
zap.Binary("id", rawPending.nodeID),
|
2020-02-14 00:03:41 +00:00
|
|
|
zap.Error(errs.Wrap(err)))
|
|
|
|
continue
|
|
|
|
}
|
2020-03-02 18:47:22 +00:00
|
|
|
pendingSerial.BucketID = rawPending.bucketID
|
|
|
|
pendingSerial.SerialNumber, err = storj.SerialNumberFromBytes(rawPending.serialNumber)
|
2020-02-14 00:03:41 +00:00
|
|
|
if err != nil {
|
|
|
|
queue.log.Error("Invalid serial number in pending serials queue",
|
2020-03-02 18:47:22 +00:00
|
|
|
zap.Binary("id", rawPending.serialNumber),
|
2020-02-14 00:03:41 +00:00
|
|
|
zap.Error(errs.Wrap(err)))
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
2020-03-02 18:47:22 +00:00
|
|
|
pendingSerials = append(pendingSerials, pendingSerial)
|
|
|
|
}
|
|
|
|
if err := rows.Err(); err != nil {
|
|
|
|
return nil, false, Error.Wrap(err)
|
2020-02-14 00:03:41 +00:00
|
|
|
}
|
|
|
|
|
2020-03-02 18:47:22 +00:00
|
|
|
return pendingSerials, size > 0, nil
|
2020-02-14 00:03:41 +00:00
|
|
|
}
|