add bucket metadata table in SA masterDB (#2369)

* add bucket metadata table in SA masterDB

* fix indentation

* update db model per CR comments

* update testdata

* add missing field on sql testdata

* fix args to testdata

* unique bucket name

* fix fkey constraint for test

* fix one too many commas

*  update timestamp type

* Trigger Jenkins

* Trigger Jenkins yet again
This commit is contained in:
Jess G 2019-07-01 13:45:21 -07:00 committed by GitHub
parent e8605d312e
commit 0158b1eb5a
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
7 changed files with 1093 additions and 1 deletions

View File

@ -730,3 +730,43 @@ read count (
)
create user_credit ()
//--- metainfo buckets ---//
model bucket (
key id
unique name project_id
field id blob
field project_id project.id restrict
field name blob
field path_cipher int
field created_at timestamp ( autoinsert )
field default_segment_size int (updatable)
field default_encryption_cipher_suite int (updatable)
field default_encryption_block_size int (updatable)
field default_redundancy_algorithm int (updatable)
field default_redundancy_share_size int (updatable)
field default_redundancy_required_shares int (updatable)
field default_redundancy_repair_shares int (updatable)
field default_redundancy_optimal_shares int (updatable)
field default_redundancy_total_shares int (updatable)
)
create bucket ()
read one (
select bucket
where bucket.project_id = ?
where bucket.name = ?
)
delete bucket (
where bucket.project_id = ?
where bucket.name = ?
)

View File

@ -479,6 +479,24 @@ CREATE TABLE api_keys (
UNIQUE ( head ),
UNIQUE ( name, project_id )
);
CREATE TABLE buckets (
id bytea NOT NULL,
project_id bytea NOT NULL REFERENCES projects( id ),
name bytea NOT NULL,
path_cipher integer NOT NULL,
created_at timestamp NOT NULL,
default_segment_size integer NOT NULL,
default_encryption_cipher_suite integer NOT NULL,
default_encryption_block_size integer NOT NULL,
default_redundancy_algorithm integer NOT NULL,
default_redundancy_share_size integer NOT NULL,
default_redundancy_required_shares integer NOT NULL,
default_redundancy_repair_shares integer NOT NULL,
default_redundancy_optimal_shares integer NOT NULL,
default_redundancy_total_shares integer NOT NULL,
PRIMARY KEY ( id ),
UNIQUE ( name, project_id )
);
CREATE TABLE project_invoice_stamps (
project_id bytea NOT NULL REFERENCES projects( id ) ON DELETE CASCADE,
invoice_id bytea NOT NULL,
@ -800,6 +818,24 @@ CREATE TABLE api_keys (
UNIQUE ( head ),
UNIQUE ( name, project_id )
);
CREATE TABLE buckets (
id BLOB NOT NULL,
project_id BLOB NOT NULL REFERENCES projects( id ),
name BLOB NOT NULL,
path_cipher INTEGER NOT NULL,
created_at TIMESTAMP NOT NULL,
default_segment_size INTEGER NOT NULL,
default_encryption_cipher_suite INTEGER NOT NULL,
default_encryption_block_size INTEGER NOT NULL,
default_redundancy_algorithm INTEGER NOT NULL,
default_redundancy_share_size INTEGER NOT NULL,
default_redundancy_required_shares INTEGER NOT NULL,
default_redundancy_repair_shares INTEGER NOT NULL,
default_redundancy_optimal_shares INTEGER NOT NULL,
default_redundancy_total_shares INTEGER NOT NULL,
PRIMARY KEY ( id ),
UNIQUE ( name, project_id )
);
CREATE TABLE project_invoice_stamps (
project_id BLOB NOT NULL REFERENCES projects( id ) ON DELETE CASCADE,
invoice_id BLOB NOT NULL,
@ -4067,6 +4103,318 @@ func (f ApiKey_CreatedAt_Field) value() interface{} {
func (ApiKey_CreatedAt_Field) _Column() string { return "created_at" }
type Bucket struct {
Id []byte
ProjectId []byte
Name []byte
PathCipher int
CreatedAt time.Time
DefaultSegmentSize int
DefaultEncryptionCipherSuite int
DefaultEncryptionBlockSize int
DefaultRedundancyAlgorithm int
DefaultRedundancyShareSize int
DefaultRedundancyRequiredShares int
DefaultRedundancyRepairShares int
DefaultRedundancyOptimalShares int
DefaultRedundancyTotalShares int
}
func (Bucket) _Table() string { return "buckets" }
type Bucket_Update_Fields struct {
DefaultSegmentSize Bucket_DefaultSegmentSize_Field
DefaultEncryptionCipherSuite Bucket_DefaultEncryptionCipherSuite_Field
DefaultEncryptionBlockSize Bucket_DefaultEncryptionBlockSize_Field
DefaultRedundancyAlgorithm Bucket_DefaultRedundancyAlgorithm_Field
DefaultRedundancyShareSize Bucket_DefaultRedundancyShareSize_Field
DefaultRedundancyRequiredShares Bucket_DefaultRedundancyRequiredShares_Field
DefaultRedundancyRepairShares Bucket_DefaultRedundancyRepairShares_Field
DefaultRedundancyOptimalShares Bucket_DefaultRedundancyOptimalShares_Field
DefaultRedundancyTotalShares Bucket_DefaultRedundancyTotalShares_Field
}
type Bucket_Id_Field struct {
_set bool
_null bool
_value []byte
}
func Bucket_Id(v []byte) Bucket_Id_Field {
return Bucket_Id_Field{_set: true, _value: v}
}
func (f Bucket_Id_Field) value() interface{} {
if !f._set || f._null {
return nil
}
return f._value
}
func (Bucket_Id_Field) _Column() string { return "id" }
type Bucket_ProjectId_Field struct {
_set bool
_null bool
_value []byte
}
func Bucket_ProjectId(v []byte) Bucket_ProjectId_Field {
return Bucket_ProjectId_Field{_set: true, _value: v}
}
func (f Bucket_ProjectId_Field) value() interface{} {
if !f._set || f._null {
return nil
}
return f._value
}
func (Bucket_ProjectId_Field) _Column() string { return "project_id" }
type Bucket_Name_Field struct {
_set bool
_null bool
_value []byte
}
func Bucket_Name(v []byte) Bucket_Name_Field {
return Bucket_Name_Field{_set: true, _value: v}
}
func (f Bucket_Name_Field) value() interface{} {
if !f._set || f._null {
return nil
}
return f._value
}
func (Bucket_Name_Field) _Column() string { return "name" }
type Bucket_PathCipher_Field struct {
_set bool
_null bool
_value int
}
func Bucket_PathCipher(v int) Bucket_PathCipher_Field {
return Bucket_PathCipher_Field{_set: true, _value: v}
}
func (f Bucket_PathCipher_Field) value() interface{} {
if !f._set || f._null {
return nil
}
return f._value
}
func (Bucket_PathCipher_Field) _Column() string { return "path_cipher" }
type Bucket_CreatedAt_Field struct {
_set bool
_null bool
_value time.Time
}
func Bucket_CreatedAt(v time.Time) Bucket_CreatedAt_Field {
v = toUTC(v)
return Bucket_CreatedAt_Field{_set: true, _value: v}
}
func (f Bucket_CreatedAt_Field) value() interface{} {
if !f._set || f._null {
return nil
}
return f._value
}
func (Bucket_CreatedAt_Field) _Column() string { return "created_at" }
type Bucket_DefaultSegmentSize_Field struct {
_set bool
_null bool
_value int
}
func Bucket_DefaultSegmentSize(v int) Bucket_DefaultSegmentSize_Field {
return Bucket_DefaultSegmentSize_Field{_set: true, _value: v}
}
func (f Bucket_DefaultSegmentSize_Field) value() interface{} {
if !f._set || f._null {
return nil
}
return f._value
}
func (Bucket_DefaultSegmentSize_Field) _Column() string { return "default_segment_size" }
type Bucket_DefaultEncryptionCipherSuite_Field struct {
_set bool
_null bool
_value int
}
func Bucket_DefaultEncryptionCipherSuite(v int) Bucket_DefaultEncryptionCipherSuite_Field {
return Bucket_DefaultEncryptionCipherSuite_Field{_set: true, _value: v}
}
func (f Bucket_DefaultEncryptionCipherSuite_Field) value() interface{} {
if !f._set || f._null {
return nil
}
return f._value
}
func (Bucket_DefaultEncryptionCipherSuite_Field) _Column() string {
return "default_encryption_cipher_suite"
}
type Bucket_DefaultEncryptionBlockSize_Field struct {
_set bool
_null bool
_value int
}
func Bucket_DefaultEncryptionBlockSize(v int) Bucket_DefaultEncryptionBlockSize_Field {
return Bucket_DefaultEncryptionBlockSize_Field{_set: true, _value: v}
}
func (f Bucket_DefaultEncryptionBlockSize_Field) value() interface{} {
if !f._set || f._null {
return nil
}
return f._value
}
func (Bucket_DefaultEncryptionBlockSize_Field) _Column() string {
return "default_encryption_block_size"
}
type Bucket_DefaultRedundancyAlgorithm_Field struct {
_set bool
_null bool
_value int
}
func Bucket_DefaultRedundancyAlgorithm(v int) Bucket_DefaultRedundancyAlgorithm_Field {
return Bucket_DefaultRedundancyAlgorithm_Field{_set: true, _value: v}
}
func (f Bucket_DefaultRedundancyAlgorithm_Field) value() interface{} {
if !f._set || f._null {
return nil
}
return f._value
}
func (Bucket_DefaultRedundancyAlgorithm_Field) _Column() string { return "default_redundancy_algorithm" }
type Bucket_DefaultRedundancyShareSize_Field struct {
_set bool
_null bool
_value int
}
func Bucket_DefaultRedundancyShareSize(v int) Bucket_DefaultRedundancyShareSize_Field {
return Bucket_DefaultRedundancyShareSize_Field{_set: true, _value: v}
}
func (f Bucket_DefaultRedundancyShareSize_Field) value() interface{} {
if !f._set || f._null {
return nil
}
return f._value
}
func (Bucket_DefaultRedundancyShareSize_Field) _Column() string {
return "default_redundancy_share_size"
}
type Bucket_DefaultRedundancyRequiredShares_Field struct {
_set bool
_null bool
_value int
}
func Bucket_DefaultRedundancyRequiredShares(v int) Bucket_DefaultRedundancyRequiredShares_Field {
return Bucket_DefaultRedundancyRequiredShares_Field{_set: true, _value: v}
}
func (f Bucket_DefaultRedundancyRequiredShares_Field) value() interface{} {
if !f._set || f._null {
return nil
}
return f._value
}
func (Bucket_DefaultRedundancyRequiredShares_Field) _Column() string {
return "default_redundancy_required_shares"
}
type Bucket_DefaultRedundancyRepairShares_Field struct {
_set bool
_null bool
_value int
}
func Bucket_DefaultRedundancyRepairShares(v int) Bucket_DefaultRedundancyRepairShares_Field {
return Bucket_DefaultRedundancyRepairShares_Field{_set: true, _value: v}
}
func (f Bucket_DefaultRedundancyRepairShares_Field) value() interface{} {
if !f._set || f._null {
return nil
}
return f._value
}
func (Bucket_DefaultRedundancyRepairShares_Field) _Column() string {
return "default_redundancy_repair_shares"
}
type Bucket_DefaultRedundancyOptimalShares_Field struct {
_set bool
_null bool
_value int
}
func Bucket_DefaultRedundancyOptimalShares(v int) Bucket_DefaultRedundancyOptimalShares_Field {
return Bucket_DefaultRedundancyOptimalShares_Field{_set: true, _value: v}
}
func (f Bucket_DefaultRedundancyOptimalShares_Field) value() interface{} {
if !f._set || f._null {
return nil
}
return f._value
}
func (Bucket_DefaultRedundancyOptimalShares_Field) _Column() string {
return "default_redundancy_optimal_shares"
}
type Bucket_DefaultRedundancyTotalShares_Field struct {
_set bool
_null bool
_value int
}
func Bucket_DefaultRedundancyTotalShares(v int) Bucket_DefaultRedundancyTotalShares_Field {
return Bucket_DefaultRedundancyTotalShares_Field{_set: true, _value: v}
}
func (f Bucket_DefaultRedundancyTotalShares_Field) value() interface{} {
if !f._set || f._null {
return nil
}
return f._value
}
func (Bucket_DefaultRedundancyTotalShares_Field) _Column() string {
return "default_redundancy_total_shares"
}
type ProjectInvoiceStamp struct {
ProjectId []byte
InvoiceId []byte
@ -5548,6 +5896,52 @@ func (obj *postgresImpl) Create_UserCredit(ctx context.Context,
}
func (obj *postgresImpl) Create_Bucket(ctx context.Context,
bucket_id Bucket_Id_Field,
bucket_project_id Bucket_ProjectId_Field,
bucket_name Bucket_Name_Field,
bucket_path_cipher Bucket_PathCipher_Field,
bucket_default_segment_size Bucket_DefaultSegmentSize_Field,
bucket_default_encryption_cipher_suite Bucket_DefaultEncryptionCipherSuite_Field,
bucket_default_encryption_block_size Bucket_DefaultEncryptionBlockSize_Field,
bucket_default_redundancy_algorithm Bucket_DefaultRedundancyAlgorithm_Field,
bucket_default_redundancy_share_size Bucket_DefaultRedundancyShareSize_Field,
bucket_default_redundancy_required_shares Bucket_DefaultRedundancyRequiredShares_Field,
bucket_default_redundancy_repair_shares Bucket_DefaultRedundancyRepairShares_Field,
bucket_default_redundancy_optimal_shares Bucket_DefaultRedundancyOptimalShares_Field,
bucket_default_redundancy_total_shares Bucket_DefaultRedundancyTotalShares_Field) (
bucket *Bucket, err error) {
__now := obj.db.Hooks.Now().UTC()
__id_val := bucket_id.value()
__project_id_val := bucket_project_id.value()
__name_val := bucket_name.value()
__path_cipher_val := bucket_path_cipher.value()
__created_at_val := __now.UTC()
__default_segment_size_val := bucket_default_segment_size.value()
__default_encryption_cipher_suite_val := bucket_default_encryption_cipher_suite.value()
__default_encryption_block_size_val := bucket_default_encryption_block_size.value()
__default_redundancy_algorithm_val := bucket_default_redundancy_algorithm.value()
__default_redundancy_share_size_val := bucket_default_redundancy_share_size.value()
__default_redundancy_required_shares_val := bucket_default_redundancy_required_shares.value()
__default_redundancy_repair_shares_val := bucket_default_redundancy_repair_shares.value()
__default_redundancy_optimal_shares_val := bucket_default_redundancy_optimal_shares.value()
__default_redundancy_total_shares_val := bucket_default_redundancy_total_shares.value()
var __embed_stmt = __sqlbundle_Literal("INSERT INTO buckets ( id, project_id, name, path_cipher, created_at, default_segment_size, default_encryption_cipher_suite, default_encryption_block_size, default_redundancy_algorithm, default_redundancy_share_size, default_redundancy_required_shares, default_redundancy_repair_shares, default_redundancy_optimal_shares, default_redundancy_total_shares ) VALUES ( ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ? ) RETURNING buckets.id, buckets.project_id, buckets.name, buckets.path_cipher, buckets.created_at, buckets.default_segment_size, buckets.default_encryption_cipher_suite, buckets.default_encryption_block_size, buckets.default_redundancy_algorithm, buckets.default_redundancy_share_size, buckets.default_redundancy_required_shares, buckets.default_redundancy_repair_shares, buckets.default_redundancy_optimal_shares, buckets.default_redundancy_total_shares")
var __stmt = __sqlbundle_Render(obj.dialect, __embed_stmt)
obj.logStmt(__stmt, __id_val, __project_id_val, __name_val, __path_cipher_val, __created_at_val, __default_segment_size_val, __default_encryption_cipher_suite_val, __default_encryption_block_size_val, __default_redundancy_algorithm_val, __default_redundancy_share_size_val, __default_redundancy_required_shares_val, __default_redundancy_repair_shares_val, __default_redundancy_optimal_shares_val, __default_redundancy_total_shares_val)
bucket = &Bucket{}
err = obj.driver.QueryRow(__stmt, __id_val, __project_id_val, __name_val, __path_cipher_val, __created_at_val, __default_segment_size_val, __default_encryption_cipher_suite_val, __default_encryption_block_size_val, __default_redundancy_algorithm_val, __default_redundancy_share_size_val, __default_redundancy_required_shares_val, __default_redundancy_repair_shares_val, __default_redundancy_optimal_shares_val, __default_redundancy_total_shares_val).Scan(&bucket.Id, &bucket.ProjectId, &bucket.Name, &bucket.PathCipher, &bucket.CreatedAt, &bucket.DefaultSegmentSize, &bucket.DefaultEncryptionCipherSuite, &bucket.DefaultEncryptionBlockSize, &bucket.DefaultRedundancyAlgorithm, &bucket.DefaultRedundancyShareSize, &bucket.DefaultRedundancyRequiredShares, &bucket.DefaultRedundancyRepairShares, &bucket.DefaultRedundancyOptimalShares, &bucket.DefaultRedundancyTotalShares)
if err != nil {
return nil, obj.makeErr(err)
}
return bucket, nil
}
func (obj *postgresImpl) Get_ValueAttribution_By_ProjectId_And_BucketName(ctx context.Context,
value_attribution_project_id ValueAttribution_ProjectId_Field,
value_attribution_bucket_name ValueAttribution_BucketName_Field) (
@ -6913,6 +7307,28 @@ func (obj *postgresImpl) Count_UserCredit_By_ReferredBy(ctx context.Context,
}
func (obj *postgresImpl) Get_Bucket_By_ProjectId_And_Name(ctx context.Context,
bucket_project_id Bucket_ProjectId_Field,
bucket_name Bucket_Name_Field) (
bucket *Bucket, err error) {
var __embed_stmt = __sqlbundle_Literal("SELECT buckets.id, buckets.project_id, buckets.name, buckets.path_cipher, buckets.created_at, buckets.default_segment_size, buckets.default_encryption_cipher_suite, buckets.default_encryption_block_size, buckets.default_redundancy_algorithm, buckets.default_redundancy_share_size, buckets.default_redundancy_required_shares, buckets.default_redundancy_repair_shares, buckets.default_redundancy_optimal_shares, buckets.default_redundancy_total_shares FROM buckets WHERE buckets.project_id = ? AND buckets.name = ?")
var __values []interface{}
__values = append(__values, bucket_project_id.value(), bucket_name.value())
var __stmt = __sqlbundle_Render(obj.dialect, __embed_stmt)
obj.logStmt(__stmt, __values...)
bucket = &Bucket{}
err = obj.driver.QueryRow(__stmt, __values...).Scan(&bucket.Id, &bucket.ProjectId, &bucket.Name, &bucket.PathCipher, &bucket.CreatedAt, &bucket.DefaultSegmentSize, &bucket.DefaultEncryptionCipherSuite, &bucket.DefaultEncryptionBlockSize, &bucket.DefaultRedundancyAlgorithm, &bucket.DefaultRedundancyShareSize, &bucket.DefaultRedundancyRequiredShares, &bucket.DefaultRedundancyRepairShares, &bucket.DefaultRedundancyOptimalShares, &bucket.DefaultRedundancyTotalShares)
if err != nil {
return nil, obj.makeErr(err)
}
return bucket, nil
}
func (obj *postgresImpl) Update_PendingAudits_By_NodeId(ctx context.Context,
pending_audits_node_id PendingAudits_NodeId_Field,
update PendingAudits_Update_Fields) (
@ -7896,6 +8312,33 @@ func (obj *postgresImpl) Delete_ResetPasswordToken_By_Secret(ctx context.Context
}
func (obj *postgresImpl) Delete_Bucket_By_ProjectId_And_Name(ctx context.Context,
bucket_project_id Bucket_ProjectId_Field,
bucket_name Bucket_Name_Field) (
deleted bool, err error) {
var __embed_stmt = __sqlbundle_Literal("DELETE FROM buckets WHERE buckets.project_id = ? AND buckets.name = ?")
var __values []interface{}
__values = append(__values, bucket_project_id.value(), bucket_name.value())
var __stmt = __sqlbundle_Render(obj.dialect, __embed_stmt)
obj.logStmt(__stmt, __values...)
__res, err := obj.driver.Exec(__stmt, __values...)
if err != nil {
return false, obj.makeErr(err)
}
__count, err := __res.RowsAffected()
if err != nil {
return false, obj.makeErr(err)
}
return __count > 0, nil
}
func (impl postgresImpl) isConstraintError(err error) (
constraint string, ok bool) {
if e, ok := err.(*pq.Error); ok {
@ -7964,6 +8407,16 @@ func (obj *postgresImpl) deleteAll(ctx context.Context) (count int64, err error)
return 0, obj.makeErr(err)
}
__count, err = __res.RowsAffected()
if err != nil {
return 0, obj.makeErr(err)
}
count += __count
__res, err = obj.driver.Exec("DELETE FROM buckets;")
if err != nil {
return 0, obj.makeErr(err)
}
__count, err = __res.RowsAffected()
if err != nil {
return 0, obj.makeErr(err)
@ -8935,6 +9388,55 @@ func (obj *sqlite3Impl) Create_UserCredit(ctx context.Context,
}
func (obj *sqlite3Impl) Create_Bucket(ctx context.Context,
bucket_id Bucket_Id_Field,
bucket_project_id Bucket_ProjectId_Field,
bucket_name Bucket_Name_Field,
bucket_path_cipher Bucket_PathCipher_Field,
bucket_default_segment_size Bucket_DefaultSegmentSize_Field,
bucket_default_encryption_cipher_suite Bucket_DefaultEncryptionCipherSuite_Field,
bucket_default_encryption_block_size Bucket_DefaultEncryptionBlockSize_Field,
bucket_default_redundancy_algorithm Bucket_DefaultRedundancyAlgorithm_Field,
bucket_default_redundancy_share_size Bucket_DefaultRedundancyShareSize_Field,
bucket_default_redundancy_required_shares Bucket_DefaultRedundancyRequiredShares_Field,
bucket_default_redundancy_repair_shares Bucket_DefaultRedundancyRepairShares_Field,
bucket_default_redundancy_optimal_shares Bucket_DefaultRedundancyOptimalShares_Field,
bucket_default_redundancy_total_shares Bucket_DefaultRedundancyTotalShares_Field) (
bucket *Bucket, err error) {
__now := obj.db.Hooks.Now().UTC()
__id_val := bucket_id.value()
__project_id_val := bucket_project_id.value()
__name_val := bucket_name.value()
__path_cipher_val := bucket_path_cipher.value()
__created_at_val := __now.UTC()
__default_segment_size_val := bucket_default_segment_size.value()
__default_encryption_cipher_suite_val := bucket_default_encryption_cipher_suite.value()
__default_encryption_block_size_val := bucket_default_encryption_block_size.value()
__default_redundancy_algorithm_val := bucket_default_redundancy_algorithm.value()
__default_redundancy_share_size_val := bucket_default_redundancy_share_size.value()
__default_redundancy_required_shares_val := bucket_default_redundancy_required_shares.value()
__default_redundancy_repair_shares_val := bucket_default_redundancy_repair_shares.value()
__default_redundancy_optimal_shares_val := bucket_default_redundancy_optimal_shares.value()
__default_redundancy_total_shares_val := bucket_default_redundancy_total_shares.value()
var __embed_stmt = __sqlbundle_Literal("INSERT INTO buckets ( id, project_id, name, path_cipher, created_at, default_segment_size, default_encryption_cipher_suite, default_encryption_block_size, default_redundancy_algorithm, default_redundancy_share_size, default_redundancy_required_shares, default_redundancy_repair_shares, default_redundancy_optimal_shares, default_redundancy_total_shares ) VALUES ( ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ? )")
var __stmt = __sqlbundle_Render(obj.dialect, __embed_stmt)
obj.logStmt(__stmt, __id_val, __project_id_val, __name_val, __path_cipher_val, __created_at_val, __default_segment_size_val, __default_encryption_cipher_suite_val, __default_encryption_block_size_val, __default_redundancy_algorithm_val, __default_redundancy_share_size_val, __default_redundancy_required_shares_val, __default_redundancy_repair_shares_val, __default_redundancy_optimal_shares_val, __default_redundancy_total_shares_val)
__res, err := obj.driver.Exec(__stmt, __id_val, __project_id_val, __name_val, __path_cipher_val, __created_at_val, __default_segment_size_val, __default_encryption_cipher_suite_val, __default_encryption_block_size_val, __default_redundancy_algorithm_val, __default_redundancy_share_size_val, __default_redundancy_required_shares_val, __default_redundancy_repair_shares_val, __default_redundancy_optimal_shares_val, __default_redundancy_total_shares_val)
if err != nil {
return nil, obj.makeErr(err)
}
__pk, err := __res.LastInsertId()
if err != nil {
return nil, obj.makeErr(err)
}
return obj.getLastBucket(ctx, __pk)
}
func (obj *sqlite3Impl) Get_ValueAttribution_By_ProjectId_And_BucketName(ctx context.Context,
value_attribution_project_id ValueAttribution_ProjectId_Field,
value_attribution_bucket_name ValueAttribution_BucketName_Field) (
@ -10300,6 +10802,28 @@ func (obj *sqlite3Impl) Count_UserCredit_By_ReferredBy(ctx context.Context,
}
func (obj *sqlite3Impl) Get_Bucket_By_ProjectId_And_Name(ctx context.Context,
bucket_project_id Bucket_ProjectId_Field,
bucket_name Bucket_Name_Field) (
bucket *Bucket, err error) {
var __embed_stmt = __sqlbundle_Literal("SELECT buckets.id, buckets.project_id, buckets.name, buckets.path_cipher, buckets.created_at, buckets.default_segment_size, buckets.default_encryption_cipher_suite, buckets.default_encryption_block_size, buckets.default_redundancy_algorithm, buckets.default_redundancy_share_size, buckets.default_redundancy_required_shares, buckets.default_redundancy_repair_shares, buckets.default_redundancy_optimal_shares, buckets.default_redundancy_total_shares FROM buckets WHERE buckets.project_id = ? AND buckets.name = ?")
var __values []interface{}
__values = append(__values, bucket_project_id.value(), bucket_name.value())
var __stmt = __sqlbundle_Render(obj.dialect, __embed_stmt)
obj.logStmt(__stmt, __values...)
bucket = &Bucket{}
err = obj.driver.QueryRow(__stmt, __values...).Scan(&bucket.Id, &bucket.ProjectId, &bucket.Name, &bucket.PathCipher, &bucket.CreatedAt, &bucket.DefaultSegmentSize, &bucket.DefaultEncryptionCipherSuite, &bucket.DefaultEncryptionBlockSize, &bucket.DefaultRedundancyAlgorithm, &bucket.DefaultRedundancyShareSize, &bucket.DefaultRedundancyRequiredShares, &bucket.DefaultRedundancyRepairShares, &bucket.DefaultRedundancyOptimalShares, &bucket.DefaultRedundancyTotalShares)
if err != nil {
return nil, obj.makeErr(err)
}
return bucket, nil
}
func (obj *sqlite3Impl) Update_PendingAudits_By_NodeId(ctx context.Context,
pending_audits_node_id PendingAudits_NodeId_Field,
update PendingAudits_Update_Fields) (
@ -11383,6 +11907,33 @@ func (obj *sqlite3Impl) Delete_ResetPasswordToken_By_Secret(ctx context.Context,
}
func (obj *sqlite3Impl) Delete_Bucket_By_ProjectId_And_Name(ctx context.Context,
bucket_project_id Bucket_ProjectId_Field,
bucket_name Bucket_Name_Field) (
deleted bool, err error) {
var __embed_stmt = __sqlbundle_Literal("DELETE FROM buckets WHERE buckets.project_id = ? AND buckets.name = ?")
var __values []interface{}
__values = append(__values, bucket_project_id.value(), bucket_name.value())
var __stmt = __sqlbundle_Render(obj.dialect, __embed_stmt)
obj.logStmt(__stmt, __values...)
__res, err := obj.driver.Exec(__stmt, __values...)
if err != nil {
return false, obj.makeErr(err)
}
__count, err := __res.RowsAffected()
if err != nil {
return false, obj.makeErr(err)
}
return __count > 0, nil
}
func (obj *sqlite3Impl) getLastValueAttribution(ctx context.Context,
pk int64) (
value_attribution *ValueAttribution, err error) {
@ -11797,6 +12348,24 @@ func (obj *sqlite3Impl) getLastUserCredit(ctx context.Context,
}
func (obj *sqlite3Impl) getLastBucket(ctx context.Context,
pk int64) (
bucket *Bucket, err error) {
var __embed_stmt = __sqlbundle_Literal("SELECT buckets.id, buckets.project_id, buckets.name, buckets.path_cipher, buckets.created_at, buckets.default_segment_size, buckets.default_encryption_cipher_suite, buckets.default_encryption_block_size, buckets.default_redundancy_algorithm, buckets.default_redundancy_share_size, buckets.default_redundancy_required_shares, buckets.default_redundancy_repair_shares, buckets.default_redundancy_optimal_shares, buckets.default_redundancy_total_shares FROM buckets WHERE _rowid_ = ?")
var __stmt = __sqlbundle_Render(obj.dialect, __embed_stmt)
obj.logStmt(__stmt, pk)
bucket = &Bucket{}
err = obj.driver.QueryRow(__stmt, pk).Scan(&bucket.Id, &bucket.ProjectId, &bucket.Name, &bucket.PathCipher, &bucket.CreatedAt, &bucket.DefaultSegmentSize, &bucket.DefaultEncryptionCipherSuite, &bucket.DefaultEncryptionBlockSize, &bucket.DefaultRedundancyAlgorithm, &bucket.DefaultRedundancyShareSize, &bucket.DefaultRedundancyRequiredShares, &bucket.DefaultRedundancyRepairShares, &bucket.DefaultRedundancyOptimalShares, &bucket.DefaultRedundancyTotalShares)
if err != nil {
return nil, obj.makeErr(err)
}
return bucket, nil
}
func (impl sqlite3Impl) isConstraintError(err error) (
constraint string, ok bool) {
if e, ok := err.(sqlite3.Error); ok {
@ -11870,6 +12439,16 @@ func (obj *sqlite3Impl) deleteAll(ctx context.Context) (count int64, err error)
return 0, obj.makeErr(err)
}
__count, err = __res.RowsAffected()
if err != nil {
return 0, obj.makeErr(err)
}
count += __count
__res, err = obj.driver.Exec("DELETE FROM buckets;")
if err != nil {
return 0, obj.makeErr(err)
}
__count, err = __res.RowsAffected()
if err != nil {
return 0, obj.makeErr(err)
@ -12317,6 +12896,29 @@ func (rx *Rx) Create_ApiKey(ctx context.Context,
}
func (rx *Rx) Create_Bucket(ctx context.Context,
bucket_id Bucket_Id_Field,
bucket_project_id Bucket_ProjectId_Field,
bucket_name Bucket_Name_Field,
bucket_path_cipher Bucket_PathCipher_Field,
bucket_default_segment_size Bucket_DefaultSegmentSize_Field,
bucket_default_encryption_cipher_suite Bucket_DefaultEncryptionCipherSuite_Field,
bucket_default_encryption_block_size Bucket_DefaultEncryptionBlockSize_Field,
bucket_default_redundancy_algorithm Bucket_DefaultRedundancyAlgorithm_Field,
bucket_default_redundancy_share_size Bucket_DefaultRedundancyShareSize_Field,
bucket_default_redundancy_required_shares Bucket_DefaultRedundancyRequiredShares_Field,
bucket_default_redundancy_repair_shares Bucket_DefaultRedundancyRepairShares_Field,
bucket_default_redundancy_optimal_shares Bucket_DefaultRedundancyOptimalShares_Field,
bucket_default_redundancy_total_shares Bucket_DefaultRedundancyTotalShares_Field) (
bucket *Bucket, err error) {
var tx *Tx
if tx, err = rx.getTx(ctx); err != nil {
return
}
return tx.Create_Bucket(ctx, bucket_id, bucket_project_id, bucket_name, bucket_path_cipher, bucket_default_segment_size, bucket_default_encryption_cipher_suite, bucket_default_encryption_block_size, bucket_default_redundancy_algorithm, bucket_default_redundancy_share_size, bucket_default_redundancy_required_shares, bucket_default_redundancy_repair_shares, bucket_default_redundancy_optimal_shares, bucket_default_redundancy_total_shares)
}
func (rx *Rx) Create_BucketStorageTally(ctx context.Context,
bucket_storage_tally_bucket_name BucketStorageTally_BucketName_Field,
bucket_storage_tally_project_id BucketStorageTally_ProjectId_Field,
@ -12661,6 +13263,17 @@ func (rx *Rx) Delete_BucketUsage_By_Id(ctx context.Context,
return tx.Delete_BucketUsage_By_Id(ctx, bucket_usage_id)
}
func (rx *Rx) Delete_Bucket_By_ProjectId_And_Name(ctx context.Context,
bucket_project_id Bucket_ProjectId_Field,
bucket_name Bucket_Name_Field) (
deleted bool, err error) {
var tx *Tx
if tx, err = rx.getTx(ctx); err != nil {
return
}
return tx.Delete_Bucket_By_ProjectId_And_Name(ctx, bucket_project_id, bucket_name)
}
func (rx *Rx) Delete_CertRecord_By_Id(ctx context.Context,
certRecord_id CertRecord_Id_Field) (
deleted bool, err error) {
@ -12869,6 +13482,17 @@ func (rx *Rx) Get_BucketUsage_By_Id(ctx context.Context,
return tx.Get_BucketUsage_By_Id(ctx, bucket_usage_id)
}
func (rx *Rx) Get_Bucket_By_ProjectId_And_Name(ctx context.Context,
bucket_project_id Bucket_ProjectId_Field,
bucket_name Bucket_Name_Field) (
bucket *Bucket, err error) {
var tx *Tx
if tx, err = rx.getTx(ctx); err != nil {
return
}
return tx.Get_Bucket_By_ProjectId_And_Name(ctx, bucket_project_id, bucket_name)
}
func (rx *Rx) Get_CertRecord_By_Id(ctx context.Context,
certRecord_id CertRecord_Id_Field) (
certRecord *CertRecord, err error) {
@ -13304,6 +13928,22 @@ type Methods interface {
api_key_secret ApiKey_Secret_Field) (
api_key *ApiKey, err error)
Create_Bucket(ctx context.Context,
bucket_id Bucket_Id_Field,
bucket_project_id Bucket_ProjectId_Field,
bucket_name Bucket_Name_Field,
bucket_path_cipher Bucket_PathCipher_Field,
bucket_default_segment_size Bucket_DefaultSegmentSize_Field,
bucket_default_encryption_cipher_suite Bucket_DefaultEncryptionCipherSuite_Field,
bucket_default_encryption_block_size Bucket_DefaultEncryptionBlockSize_Field,
bucket_default_redundancy_algorithm Bucket_DefaultRedundancyAlgorithm_Field,
bucket_default_redundancy_share_size Bucket_DefaultRedundancyShareSize_Field,
bucket_default_redundancy_required_shares Bucket_DefaultRedundancyRequiredShares_Field,
bucket_default_redundancy_repair_shares Bucket_DefaultRedundancyRepairShares_Field,
bucket_default_redundancy_optimal_shares Bucket_DefaultRedundancyOptimalShares_Field,
bucket_default_redundancy_total_shares Bucket_DefaultRedundancyTotalShares_Field) (
bucket *Bucket, err error)
Create_BucketStorageTally(ctx context.Context,
bucket_storage_tally_bucket_name BucketStorageTally_BucketName_Field,
bucket_storage_tally_project_id BucketStorageTally_ProjectId_Field,
@ -13490,6 +14130,11 @@ type Methods interface {
bucket_usage_id BucketUsage_Id_Field) (
deleted bool, err error)
Delete_Bucket_By_ProjectId_And_Name(ctx context.Context,
bucket_project_id Bucket_ProjectId_Field,
bucket_name Bucket_Name_Field) (
deleted bool, err error)
Delete_CertRecord_By_Id(ctx context.Context,
certRecord_id CertRecord_Id_Field) (
deleted bool, err error)
@ -13577,6 +14222,11 @@ type Methods interface {
bucket_usage_id BucketUsage_Id_Field) (
bucket_usage *BucketUsage, err error)
Get_Bucket_By_ProjectId_And_Name(ctx context.Context,
bucket_project_id Bucket_ProjectId_Field,
bucket_name Bucket_Name_Field) (
bucket *Bucket, err error)
Get_CertRecord_By_Id(ctx context.Context,
certRecord_id CertRecord_Id_Field) (
certRecord *CertRecord, err error)

View File

@ -207,6 +207,24 @@ CREATE TABLE api_keys (
UNIQUE ( head ),
UNIQUE ( name, project_id )
);
CREATE TABLE buckets (
id bytea NOT NULL,
project_id bytea NOT NULL REFERENCES projects( id ),
name bytea NOT NULL,
path_cipher integer NOT NULL,
created_at timestamp NOT NULL,
default_segment_size integer NOT NULL,
default_encryption_cipher_suite integer NOT NULL,
default_encryption_block_size integer NOT NULL,
default_redundancy_algorithm integer NOT NULL,
default_redundancy_share_size integer NOT NULL,
default_redundancy_required_shares integer NOT NULL,
default_redundancy_repair_shares integer NOT NULL,
default_redundancy_optimal_shares integer NOT NULL,
default_redundancy_total_shares integer NOT NULL,
PRIMARY KEY ( id ),
UNIQUE ( name, project_id )
);
CREATE TABLE project_invoice_stamps (
project_id bytea NOT NULL REFERENCES projects( id ) ON DELETE CASCADE,
invoice_id bytea NOT NULL,

View File

@ -207,6 +207,24 @@ CREATE TABLE api_keys (
UNIQUE ( head ),
UNIQUE ( name, project_id )
);
CREATE TABLE buckets (
id BLOB NOT NULL,
project_id BLOB NOT NULL REFERENCES projects( id ),
name BLOB NOT NULL,
path_cipher INTEGER NOT NULL,
created_at TIMESTAMP NOT NULL,
default_segment_size INTEGER NOT NULL,
default_encryption_cipher_suite INTEGER NOT NULL,
default_encryption_block_size INTEGER NOT NULL,
default_redundancy_algorithm INTEGER NOT NULL,
default_redundancy_share_size INTEGER NOT NULL,
default_redundancy_required_shares INTEGER NOT NULL,
default_redundancy_repair_shares INTEGER NOT NULL,
default_redundancy_optimal_shares INTEGER NOT NULL,
default_redundancy_total_shares INTEGER NOT NULL,
PRIMARY KEY ( id ),
UNIQUE ( name, project_id )
);
CREATE TABLE project_invoice_stamps (
project_id BLOB NOT NULL REFERENCES projects( id ) ON DELETE CASCADE,
invoice_id BLOB NOT NULL,

View File

@ -67,7 +67,7 @@ func (m *lockedAttribution) Insert(ctx context.Context, info *attribution.Info)
return m.db.Insert(ctx, info)
}
// QueryAttribution queries partner bucket value attribution data
// QueryAttribution queries partner bucket attribution data
func (m *lockedAttribution) QueryAttribution(ctx context.Context, partnerID uuid.UUID, start time.Time, end time.Time) ([]*attribution.CSVRow, error) {
m.Lock()
defer m.Unlock()

View File

@ -874,6 +874,30 @@ func (db *DB) PostgresMigration() *migrate.Migration {
`UPDATE bucket_bandwidth_rollups SET project_id = decode(replace(encode(project_id, 'escape'), '-', ''), 'hex') WHERE length(project_id) = 36;`,
},
},
{
Description: "Add bucket metadata table",
Version: 38,
Action: migrate.SQL{
`CREATE TABLE buckets (
id bytea NOT NULL,
project_id bytea NOT NULL REFERENCES projects( id ),
name bytea NOT NULL,
path_cipher integer NOT NULL,
created_at timestamp NOT NULL,
default_segment_size integer NOT NULL,
default_encryption_cipher_suite integer NOT NULL,
default_encryption_block_size integer NOT NULL,
default_redundancy_algorithm integer NOT NULL,
default_redundancy_share_size integer NOT NULL,
default_redundancy_required_shares integer NOT NULL,
default_redundancy_repair_shares integer NOT NULL,
default_redundancy_optimal_shares integer NOT NULL,
default_redundancy_total_shares integer NOT NULL,
PRIMARY KEY ( id ),
UNIQUE ( name, project_id )
);`,
},
},
},
}
}

View File

@ -0,0 +1,342 @@
-- AUTOGENERATED BY gopkg.in/spacemonkeygo/dbx.v1
-- DO NOT EDIT
CREATE TABLE accounting_rollups (
id bigserial NOT NULL,
node_id bytea NOT NULL,
start_time timestamp with time zone NOT NULL,
put_total bigint NOT NULL,
get_total bigint NOT NULL,
get_audit_total bigint NOT NULL,
get_repair_total bigint NOT NULL,
put_repair_total bigint NOT NULL,
at_rest_total double precision NOT NULL,
PRIMARY KEY ( id )
);
CREATE TABLE accounting_timestamps (
name text NOT NULL,
value timestamp with time zone NOT NULL,
PRIMARY KEY ( name )
);
CREATE TABLE bucket_bandwidth_rollups (
bucket_name bytea NOT NULL,
project_id bytea NOT NULL,
interval_start timestamp NOT NULL,
interval_seconds integer NOT NULL,
action integer NOT NULL,
inline bigint NOT NULL,
allocated bigint NOT NULL,
settled bigint NOT NULL,
PRIMARY KEY ( bucket_name, project_id, interval_start, action )
);
CREATE TABLE bucket_storage_tallies (
bucket_name bytea NOT NULL,
project_id bytea NOT NULL,
interval_start timestamp NOT NULL,
inline bigint NOT NULL,
remote bigint NOT NULL,
remote_segments_count integer NOT NULL,
inline_segments_count integer NOT NULL,
object_count integer NOT NULL,
metadata_size bigint NOT NULL,
PRIMARY KEY ( bucket_name, project_id, interval_start )
);
CREATE TABLE bucket_usages (
id bytea NOT NULL,
bucket_id bytea NOT NULL,
rollup_end_time timestamp with time zone NOT NULL,
remote_stored_data bigint NOT NULL,
inline_stored_data bigint NOT NULL,
remote_segments integer NOT NULL,
inline_segments integer NOT NULL,
objects integer NOT NULL,
metadata_size bigint NOT NULL,
repair_egress bigint NOT NULL,
get_egress bigint NOT NULL,
audit_egress bigint NOT NULL,
PRIMARY KEY ( id )
);
CREATE TABLE certRecords (
publickey bytea NOT NULL,
id bytea NOT NULL,
update_at timestamp with time zone NOT NULL,
PRIMARY KEY ( id )
);
CREATE TABLE injuredsegments (
path text NOT NULL,
data bytea NOT NULL,
attempted timestamp,
PRIMARY KEY ( path )
);
CREATE TABLE irreparabledbs (
segmentpath bytea NOT NULL,
segmentdetail bytea NOT NULL,
pieces_lost_count bigint NOT NULL,
seg_damaged_unix_sec bigint NOT NULL,
repair_attempt_count bigint NOT NULL,
PRIMARY KEY ( segmentpath )
);
CREATE TABLE nodes (
id bytea NOT NULL,
address text NOT NULL,
last_net text NOT NULL,
protocol integer NOT NULL,
type integer NOT NULL,
email text NOT NULL,
wallet text NOT NULL,
free_bandwidth bigint NOT NULL,
free_disk bigint NOT NULL,
major bigint NOT NULL,
minor bigint NOT NULL,
patch bigint NOT NULL,
hash text NOT NULL,
timestamp timestamp with time zone NOT NULL,
release boolean NOT NULL,
latency_90 bigint NOT NULL,
audit_success_count bigint NOT NULL,
total_audit_count bigint NOT NULL,
uptime_success_count bigint NOT NULL,
total_uptime_count bigint NOT NULL,
created_at timestamp with time zone NOT NULL,
updated_at timestamp with time zone NOT NULL,
last_contact_success timestamp with time zone NOT NULL,
last_contact_failure timestamp with time zone NOT NULL,
contained boolean NOT NULL,
disqualified timestamp with time zone,
audit_reputation_alpha double precision NOT NULL,
audit_reputation_beta double precision NOT NULL,
uptime_reputation_alpha double precision NOT NULL,
uptime_reputation_beta double precision NOT NULL,
PRIMARY KEY ( id )
);
CREATE TABLE offers (
id serial NOT NULL,
name text NOT NULL,
description text NOT NULL,
award_credit_in_cents integer NOT NULL,
invitee_credit_in_cents integer NOT NULL,
award_credit_duration_days integer NOT NULL,
invitee_credit_duration_days integer NOT NULL,
redeemable_cap integer NOT NULL,
num_redeemed integer NOT NULL,
expires_at timestamp with time zone NOT NULL,
created_at timestamp with time zone NOT NULL,
status integer NOT NULL,
type integer NOT NULL,
PRIMARY KEY ( id )
);
CREATE TABLE pending_audits (
node_id bytea NOT NULL,
piece_id bytea NOT NULL,
stripe_index bigint NOT NULL,
share_size bigint NOT NULL,
expected_share_hash bytea NOT NULL,
reverify_count bigint NOT NULL,
PRIMARY KEY ( node_id )
);
CREATE TABLE projects (
id bytea NOT NULL,
name text NOT NULL,
description text NOT NULL,
usage_limit bigint NOT NULL,
created_at timestamp with time zone NOT NULL,
PRIMARY KEY ( id )
);
CREATE TABLE registration_tokens (
secret bytea NOT NULL,
owner_id bytea,
project_limit integer NOT NULL,
created_at timestamp with time zone NOT NULL,
PRIMARY KEY ( secret ),
UNIQUE ( owner_id )
);
CREATE TABLE reset_password_tokens (
secret bytea NOT NULL,
owner_id bytea NOT NULL,
created_at timestamp with time zone NOT NULL,
PRIMARY KEY ( secret ),
UNIQUE ( owner_id )
);
CREATE TABLE serial_numbers (
id serial NOT NULL,
serial_number bytea NOT NULL,
bucket_id bytea NOT NULL,
expires_at timestamp NOT NULL,
PRIMARY KEY ( id )
);
CREATE TABLE storagenode_bandwidth_rollups (
storagenode_id bytea NOT NULL,
interval_start timestamp NOT NULL,
interval_seconds integer NOT NULL,
action integer NOT NULL,
allocated bigint NOT NULL,
settled bigint NOT NULL,
PRIMARY KEY ( storagenode_id, interval_start, action )
);
CREATE TABLE storagenode_storage_tallies (
id bigserial NOT NULL,
node_id bytea NOT NULL,
interval_end_time timestamp with time zone NOT NULL,
data_total double precision NOT NULL,
PRIMARY KEY ( id )
);
CREATE TABLE users (
id bytea NOT NULL,
email text NOT NULL,
full_name text NOT NULL,
short_name text,
password_hash bytea NOT NULL,
status integer NOT NULL,
created_at timestamp with time zone NOT NULL,
PRIMARY KEY ( id )
);
CREATE TABLE value_attributions (
project_id bytea NOT NULL,
bucket_name bytea NOT NULL,
partner_id bytea NOT NULL,
last_updated timestamp NOT NULL,
PRIMARY KEY ( project_id, bucket_name )
);
CREATE TABLE api_keys (
id bytea NOT NULL,
project_id bytea NOT NULL REFERENCES projects( id ) ON DELETE CASCADE,
head bytea NOT NULL,
name text NOT NULL,
secret bytea NOT NULL,
created_at timestamp with time zone NOT NULL,
PRIMARY KEY ( id ),
UNIQUE ( head ),
UNIQUE ( name, project_id )
);
CREATE TABLE project_invoice_stamps (
project_id bytea NOT NULL REFERENCES projects( id ) ON DELETE CASCADE,
invoice_id bytea NOT NULL,
start_date timestamp with time zone NOT NULL,
end_date timestamp with time zone NOT NULL,
created_at timestamp with time zone NOT NULL,
PRIMARY KEY ( project_id, start_date, end_date ),
UNIQUE ( invoice_id )
);
CREATE TABLE project_members (
member_id bytea NOT NULL REFERENCES users( id ) ON DELETE CASCADE,
project_id bytea NOT NULL REFERENCES projects( id ) ON DELETE CASCADE,
created_at timestamp with time zone NOT NULL,
PRIMARY KEY ( member_id, project_id )
);
CREATE TABLE used_serials (
serial_number_id integer NOT NULL REFERENCES serial_numbers( id ) ON DELETE CASCADE,
storage_node_id bytea NOT NULL,
PRIMARY KEY ( serial_number_id, storage_node_id )
);
CREATE TABLE user_credits (
id serial NOT NULL,
user_id bytea NOT NULL REFERENCES users( id ),
offer_id integer NOT NULL REFERENCES offers( id ),
referred_by bytea REFERENCES users( id ),
credits_earned_in_cents integer NOT NULL,
credits_used_in_cents integer NOT NULL,
expires_at timestamp with time zone NOT NULL,
created_at timestamp with time zone NOT NULL,
PRIMARY KEY ( id )
);
CREATE TABLE user_payments (
user_id bytea NOT NULL REFERENCES users( id ) ON DELETE CASCADE,
customer_id bytea NOT NULL,
created_at timestamp with time zone NOT NULL,
PRIMARY KEY ( user_id ),
UNIQUE ( customer_id )
);
CREATE TABLE project_payments (
project_id bytea NOT NULL REFERENCES projects( id ) ON DELETE CASCADE,
payer_id bytea NOT NULL REFERENCES user_payments( user_id ) ON DELETE CASCADE,
payment_method_id bytea NOT NULL,
created_at timestamp with time zone NOT NULL,
PRIMARY KEY ( project_id )
);
CREATE INDEX bucket_name_project_id_interval_start_interval_seconds ON bucket_bandwidth_rollups ( bucket_name, project_id, interval_start, interval_seconds );
CREATE UNIQUE INDEX bucket_id_rollup ON bucket_usages ( bucket_id, rollup_end_time );
CREATE INDEX node_last_ip ON nodes ( last_net );
CREATE UNIQUE INDEX serial_number ON serial_numbers ( serial_number );
CREATE INDEX serial_numbers_expires_at_index ON serial_numbers ( expires_at );
CREATE INDEX storagenode_id_interval_start_interval_seconds ON storagenode_bandwidth_rollups ( storagenode_id, interval_start, interval_seconds );
CREATE TABLE buckets (
id bytea NOT NULL,
project_id bytea NOT NULL REFERENCES projects( id ),
name bytea NOT NULL,
path_cipher integer NOT NULL,
created_at timestamp NOT NULL,
default_segment_size integer NOT NULL,
default_encryption_cipher_suite integer NOT NULL,
default_encryption_block_size integer NOT NULL,
default_redundancy_algorithm integer NOT NULL,
default_redundancy_share_size integer NOT NULL,
default_redundancy_required_shares integer NOT NULL,
default_redundancy_repair_shares integer NOT NULL,
default_redundancy_optimal_shares integer NOT NULL,
default_redundancy_total_shares integer NOT NULL,
PRIMARY KEY ( id ),
UNIQUE ( name, project_id )
);
---
INSERT INTO "accounting_rollups"("id", "node_id", "start_time", "put_total", "get_total", "get_audit_total", "get_repair_total", "put_repair_total", "at_rest_total") VALUES (1, E'\\367M\\177\\251]t/\\022\\256\\214\\265\\025\\224\\204:\\217\\212\\0102<\\321\\374\\020&\\271Qc\\325\\261\\354\\246\\233'::bytea, '2019-02-09 00:00:00+00', 1000, 2000, 3000, 4000, 0, 5000);
INSERT INTO "accounting_timestamps" VALUES ('LastAtRestTally', '0001-01-01 00:00:00+00');
INSERT INTO "accounting_timestamps" VALUES ('LastRollup', '0001-01-01 00:00:00+00');
INSERT INTO "accounting_timestamps" VALUES ('LastBandwidthTally', '0001-01-01 00:00:00+00');
INSERT INTO "nodes"("id", "address", "last_net", "protocol", "type", "email", "wallet", "free_bandwidth", "free_disk", "major", "minor", "patch", "hash", "timestamp", "release","latency_90", "audit_success_count", "total_audit_count", "uptime_success_count", "total_uptime_count", "created_at", "updated_at", "last_contact_success", "last_contact_failure", "contained", "disqualified", "audit_reputation_alpha", "audit_reputation_beta", "uptime_reputation_alpha", "uptime_reputation_beta") VALUES (E'\\153\\313\\233\\074\\327\\177\\136\\070\\346\\001', '127.0.0.1:55516', '', 0, 4, '', '', -1, -1, 0, 1, 0, '', 'epoch', false, 0, 0, 5, 0, 5, '2019-02-14 08:07:31.028103+00', '2019-02-14 08:07:31.108963+00', 'epoch', 'epoch', false, NULL, 50, 5, 100, 5);
INSERT INTO "nodes"("id", "address", "last_net", "protocol", "type", "email", "wallet", "free_bandwidth", "free_disk", "major", "minor", "patch", "hash", "timestamp", "release","latency_90", "audit_success_count", "total_audit_count", "uptime_success_count", "total_uptime_count", "created_at", "updated_at", "last_contact_success", "last_contact_failure", "contained", "disqualified", "audit_reputation_alpha", "audit_reputation_beta", "uptime_reputation_alpha", "uptime_reputation_beta") VALUES (E'\\006\\223\\250R\\221\\005\\365\\377v>0\\266\\365\\216\\255?\\347\\244\\371?2\\264\\262\\230\\007<\\001\\262\\263\\237\\247n', '127.0.0.1:55518', '', 0, 4, '', '', -1, -1, 0, 1, 0, '', 'epoch', false, 0, 0, 0, 3, 3, '2019-02-14 08:07:31.028103+00', '2019-02-14 08:07:31.108963+00', 'epoch', 'epoch', false, NULL, 50, 0, 100, 0);
INSERT INTO "nodes"("id", "address", "last_net", "protocol", "type", "email", "wallet", "free_bandwidth", "free_disk", "major", "minor", "patch", "hash", "timestamp", "release","latency_90", "audit_success_count", "total_audit_count", "uptime_success_count", "total_uptime_count", "created_at", "updated_at", "last_contact_success", "last_contact_failure", "contained", "disqualified", "audit_reputation_alpha", "audit_reputation_beta", "uptime_reputation_alpha", "uptime_reputation_beta") VALUES (E'\\363\\342\\363\\371>+F\\256\\263\\300\\273|\\342N\\347\\014', '127.0.0.1:55517', '', 0, 4, '', '', -1, -1, 0, 1, 0, '', 'epoch', false, 0, 0, 0, 0, 0, '2019-02-14 08:07:31.028103+00', '2019-02-14 08:07:31.108963+00', 'epoch', 'epoch', false, NULL, 50, 0, 100, 0);
INSERT INTO "nodes"("id", "address", "last_net", "protocol", "type", "email", "wallet", "free_bandwidth", "free_disk", "major", "minor", "patch", "hash", "timestamp", "release","latency_90", "audit_success_count", "total_audit_count", "uptime_success_count", "total_uptime_count", "created_at", "updated_at", "last_contact_success", "last_contact_failure", "contained", "disqualified", "audit_reputation_alpha", "audit_reputation_beta", "uptime_reputation_alpha", "uptime_reputation_beta") VALUES (E'\\363\\342\\363\\371>+F\\256\\263\\300\\273|\\342N\\347\\015', '127.0.0.1:55519', '', 0, 4, '', '', -1, -1, 0, 1, 0, '', 'epoch', false, 0, 1, 2, 1, 2, '2019-02-14 08:07:31.028103+00', '2019-02-14 08:07:31.108963+00', 'epoch', 'epoch', false, NULL, 50, 1, 100, 1);
INSERT INTO "nodes"("id", "address", "last_net", "protocol", "type", "email", "wallet", "free_bandwidth", "free_disk", "major", "minor", "patch", "hash", "timestamp", "release","latency_90", "audit_success_count", "total_audit_count", "uptime_success_count", "total_uptime_count", "created_at", "updated_at", "last_contact_success", "last_contact_failure", "contained", "disqualified", "audit_reputation_alpha", "audit_reputation_beta", "uptime_reputation_alpha", "uptime_reputation_beta") VALUES (E'\\363\\342\\363\\371>+F\\256\\263\\300\\273|\\342N\\347\\016', '127.0.0.1:55520', '', 0, 4, '', '', -1, -1, 0, 1, 0, '', 'epoch', false, 0, 300, 400, 300, 400, '2019-02-14 08:07:31.028103+00', '2019-02-14 08:07:31.108963+00', 'epoch', 'epoch', false, NULL, 300, 100, 300, 100);
INSERT INTO "projects"("id", "name", "description", "usage_limit","created_at") VALUES (E'\\022\\217/\\014\\376!K\\023\\276\\031\\311}m\\236\\205\\300'::bytea, 'ProjectName', 'projects description', 0, '2019-02-14 08:28:24.254934+00');
INSERT INTO "users"("id", "full_name", "short_name", "email", "password_hash", "status", "created_at") VALUES (E'\\363\\311\\033w\\222\\303Ci\\265\\343U\\303\\312\\204",'::bytea, 'Noahson', 'William', '1email1@mail.test', E'some_readable_hash'::bytea, 1, '2019-02-14 08:28:24.614594+00');
INSERT INTO "projects"("id", "name", "description", "usage_limit", "created_at") VALUES (E'\\363\\342\\363\\371>+F\\256\\263\\300\\273|\\342N\\347\\014'::bytea, 'projName1', 'Test project 1', 0, '2019-02-14 08:28:24.636949+00');
INSERT INTO "project_members"("member_id", "project_id", "created_at") VALUES (E'\\363\\311\\033w\\222\\303Ci\\265\\343U\\303\\312\\204",'::bytea, E'\\363\\342\\363\\371>+F\\256\\263\\300\\273|\\342N\\347\\014'::bytea, '2019-02-14 08:28:24.677953+00');
INSERT INTO "irreparabledbs" ("segmentpath", "segmentdetail", "pieces_lost_count", "seg_damaged_unix_sec", "repair_attempt_count") VALUES ('\x49616d5365676d656e746b6579696e666f30', '\x49616d5365676d656e7464657461696c696e666f30', 10, 1550159554, 10);
INSERT INTO "injuredsegments" ("path", "data") VALUES ('0', '\x0a0130120100');
INSERT INTO "injuredsegments" ("path", "data") VALUES ('here''s/a/great/path', '\x0a136865726527732f612f67726561742f70617468120a0102030405060708090a');
INSERT INTO "injuredsegments" ("path", "data") VALUES ('yet/another/cool/path', '\x0a157965742f616e6f746865722f636f6f6c2f70617468120a0102030405060708090a');
INSERT INTO "injuredsegments" ("path", "data") VALUES ('so/many/iconic/paths/to/choose/from', '\x0a23736f2f6d616e792f69636f6e69632f70617468732f746f2f63686f6f73652f66726f6d120a0102030405060708090a');
INSERT INTO "certrecords" VALUES (E'0Y0\\023\\006\\007*\\206H\\316=\\002\\001\\006\\010*\\206H\\316=\\003\\001\\007\\003B\\000\\004\\360\\267\\227\\377\\253u\\222\\337Y\\324C:GQ\\010\\277v\\010\\315D\\271\\333\\337.\\203\\023=C\\343\\014T%6\\027\\362?\\214\\326\\017U\\334\\000\\260\\224\\260J\\221\\304\\331F\\304\\221\\236zF,\\325\\326l\\215\\306\\365\\200\\022', E'L\\301|\\200\\247}F|1\\320\\232\\037n\\335\\241\\206\\244\\242\\207\\204.\\253\\357\\326\\352\\033Dt\\202`\\022\\325', '2019-02-14 08:07:31.335028+00');
INSERT INTO "bucket_usages" ("id", "bucket_id", "rollup_end_time", "remote_stored_data", "inline_stored_data", "remote_segments", "inline_segments", "objects", "metadata_size", "repair_egress", "get_egress", "audit_egress") VALUES (E'\\153\\313\\233\\074\\327\\177\\136\\070\\346\\001",'::bytea, E'\\366\\146\\032\\321\\316\\161\\070\\133\\302\\271",'::bytea, '2019-03-06 08:28:24.677953+00', 10, 11, 12, 13, 14, 15, 16, 17, 18);
INSERT INTO "registration_tokens" ("secret", "owner_id", "project_limit", "created_at") VALUES (E'\\070\\127\\144\\013\\332\\344\\102\\376\\306\\056\\303\\130\\106\\132\\321\\276\\321\\274\\170\\264\\054\\333\\221\\116\\154\\221\\335\\070\\220\\146\\344\\216'::bytea, null, 1, '2019-02-14 08:28:24.677953+00');
INSERT INTO "serial_numbers" ("id", "serial_number", "bucket_id", "expires_at") VALUES (1, E'0123456701234567'::bytea, E'\\363\\342\\363\\371>+F\\256\\263\\300\\273|\\342N\\347\\014/testbucket'::bytea, '2019-03-06 08:28:24.677953+00');
INSERT INTO "used_serials" ("serial_number_id", "storage_node_id") VALUES (1, E'\\006\\223\\250R\\221\\005\\365\\377v>0\\266\\365\\216\\255?\\347\\244\\371?2\\264\\262\\230\\007<\\001\\262\\263\\237\\247n');
INSERT INTO "storagenode_bandwidth_rollups" ("storagenode_id", "interval_start", "interval_seconds", "action", "allocated", "settled") VALUES (E'\\006\\223\\250R\\221\\005\\365\\377v>0\\266\\365\\216\\255?\\347\\244\\371?2\\264\\262\\230\\007<\\001\\262\\263\\237\\247n', '2019-03-06 08:00:00.000000+00', 3600, 1, 1024, 2024);
INSERT INTO "storagenode_storage_tallies" VALUES (1, E'\\3510\\323\\225"~\\036<\\342\\330m\\0253Jhr\\246\\233K\\246#\\2303\\351\\256\\275j\\212UM\\362\\207', '2019-02-14 08:16:57.812849+00', 1000);
INSERT INTO "bucket_bandwidth_rollups" ("bucket_name", "project_id", "interval_start", "interval_seconds", "action", "inline", "allocated", "settled") VALUES (E'testbucket'::bytea, E'\\363\\342\\363\\371>+F\\256\\263\\300\\273|\\342N\\347\\014'::bytea,'2019-03-06 08:00:00.000000+00', 3600, 1, 1024, 2024, 3024);
INSERT INTO "bucket_storage_tallies" ("bucket_name", "project_id", "interval_start", "inline", "remote", "remote_segments_count", "inline_segments_count", "object_count", "metadata_size") VALUES (E'testbucket'::bytea, E'\\363\\342\\363\\371>+F\\256\\263\\300\\273|\\342N\\347\\014'::bytea,'2019-03-06 08:00:00.000000+00', 4024, 5024, 0, 0, 0, 0);
INSERT INTO "bucket_bandwidth_rollups" ("bucket_name", "project_id", "interval_start", "interval_seconds", "action", "inline", "allocated", "settled") VALUES (E'testbucket'::bytea, E'\\170\\160\\157\\370\\274\\366\\113\\364\\272\\235\\301\\243\\321\\102\\321\\136'::bytea,'2019-03-06 08:00:00.000000+00', 3600, 1, 1024, 2024, 3024);
INSERT INTO "bucket_storage_tallies" ("bucket_name", "project_id", "interval_start", "inline", "remote", "remote_segments_count", "inline_segments_count", "object_count", "metadata_size") VALUES (E'testbucket'::bytea, E'\\170\\160\\157\\370\\274\\366\\113\\364\\272\\235\\301\\243\\321\\102\\321\\136'::bytea,'2019-03-06 08:00:00.000000+00', 4024, 5024, 0, 0, 0, 0);
INSERT INTO "reset_password_tokens" ("secret", "owner_id", "created_at") VALUES (E'\\070\\127\\144\\013\\332\\344\\102\\376\\306\\056\\303\\130\\106\\132\\321\\276\\321\\274\\170\\264\\054\\333\\221\\116\\154\\221\\335\\070\\220\\146\\344\\216'::bytea, E'\\363\\311\\033w\\222\\303Ci\\265\\343U\\303\\312\\204",'::bytea, '2019-05-08 08:28:24.677953+00');
INSERT INTO "pending_audits" ("node_id", "piece_id", "stripe_index", "share_size", "expected_share_hash", "reverify_count") VALUES (E'\\153\\313\\233\\074\\327\\177\\136\\070\\346\\001'::bytea, E'\\363\\311\\033w\\222\\303Ci\\265\\343U\\303\\312\\204",'::bytea, 5, 1024, E'\\070\\127\\144\\013\\332\\344\\102\\376\\306\\056\\303\\130\\106\\132\\321\\276\\321\\274\\170\\264\\054\\333\\221\\116\\154\\221\\335\\070\\220\\146\\344\\216'::bytea, 1);
INSERT INTO "offers" ("id", "name", "description", "award_credit_in_cents", "invitee_credit_in_cents", "award_credit_duration_days", "invitee_credit_duration_days", "redeemable_cap", "expires_at", "created_at", "num_redeemed", "status", "type") VALUES (1, 'testOffer', 'Test offer 1', 0, 0, 14, 14, 50, '2019-03-14 08:28:24.636949+00', '2019-02-14 08:28:24.636949+00', 0, 0, 0);
INSERT INTO "api_keys" ("id", "project_id", "head", "name", "secret", "created_at") VALUES (E'\\334/\\302;\\225\\355O\\323\\276f\\247\\354/6\\241\\033'::bytea, E'\\022\\217/\\014\\376!K\\023\\276\\031\\311}m\\236\\205\\300'::bytea, E'\\111\\142\\147\\304\\132\\375\\070\\163\\270\\160\\251\\370\\126\\063\\351\\037\\257\\071\\143\\375\\351\\320\\253\\232\\220\\260\\075\\173\\306\\307\\115\\136'::bytea, 'key 2', E'\\254\\011\\315\\333\\273\\365\\001\\071\\024\\154\\253\\332\\301\\216\\361\\074\\221\\367\\251\\231\\274\\333\\300\\367\\001\\272\\327\\111\\315\\123\\042\\016'::bytea, '2019-02-14 08:28:24.267934+00');
INSERT INTO "user_payments" ("user_id", "customer_id", "created_at") VALUES (E'\\363\\311\\033w\\222\\303Ci\\265\\343U\\303\\312\\204",'::bytea, E'\\022\\217/\\014\\376!K\\023\\276'::bytea, '2019-06-01 08:28:24.267934+00');
INSERT INTO "project_payments" ("project_id", "payer_id", "payment_method_id", "created_at") VALUES (E'\\022\\217/\\014\\376!K\\023\\276\\031\\311}m\\236\\205\\300'::bytea, E'\\363\\311\\033w\\222\\303Ci\\265\\343U\\303\\312\\204",'::bytea, E'\\022\\217/\\014\\376!K\\023\\276'::bytea, '2019-06-01 08:28:24.267934+00');
INSERT INTO "project_invoice_stamps" ("project_id", "invoice_id", "start_date", "end_date", "created_at") VALUES (E'\\022\\217/\\014\\376!K\\023\\276\\031\\311}m\\236\\205\\300'::bytea, E'\\363\\311\\033w\\222\\303,'::bytea, '2019-06-01 08:28:24.267934+00', '2019-06-29 08:28:24.267934+00', '2019-06-01 08:28:24.267934+00');
INSERT INTO "value_attributions" ("project_id", "bucket_name", "partner_id", "last_updated") VALUES (E'\\363\\311\\033w\\222\\303Ci\\265\\343U\\303\\312\\204",'::bytea, E''::bytea, E'\\363\\342\\363\\371>+F\\256\\263\\300\\273|\\342N\\347\\014'::bytea,'2019-02-14 08:07:31.028103+00');
INSERT INTO "user_credits" ("id", "user_id", "offer_id", "referred_by", "credits_earned_in_cents", "credits_used_in_cents", "expires_at", "created_at") VALUES (1, E'\\363\\311\\033w\\222\\303Ci\\265\\343U\\303\\312\\204",'::bytea, 1, E'\\363\\311\\033w\\222\\303Ci\\265\\343U\\303\\312\\204",'::bytea, 200, 0, '2019-10-01 08:28:24.267934+00', '2019-06-01 08:28:24.267934+00');
-- NEW DATA --
INSERT INTO "buckets" ("id", "project_id", "name", "created_at", "path_cipher", "default_segment_size", "default_encryption_cipher_suite", "default_encryption_block_size", "default_redundancy_algorithm", "default_redundancy_share_size", "default_redundancy_required_shares", "default_redundancy_repair_shares", "default_redundancy_optimal_shares", "default_redundancy_total_shares") VALUES (E'\\334/\\302;\\225\\355O\\323\\276f\\247\\354/6\\241\\033'::bytea, E'\\022\\217/\\014\\376!K\\023\\276\\031\\311}m\\236\\205\\300'::bytea, E'testbucketuniquename'::bytea, '2019-06-14 08:28:24.677953+00', 1, 65536, 1, 8192, 1, 4096, 4, 6, 8, 10);