// Copyright (C) 2019 Storj Labs, Inc. // See LICENSE for copying information. package satellitedb import ( "context" "fmt" "strings" "github.com/zeebo/errs" "go.uber.org/zap" "storj.io/private/dbutil" "storj.io/private/dbutil/cockroachutil" "storj.io/private/dbutil/pgutil" "storj.io/private/tagsql" "storj.io/storj/private/migrate" ) //go:generate go run migrate_gen.go var ( // ErrMigrate is for tracking migration errors. ErrMigrate = errs.Class("migrate") // ErrMigrateMinVersion is for migration min version errors. ErrMigrateMinVersion = errs.Class("migrate min version") ) // MigrateToLatest migrates the database to the latest version. func (db *satelliteDB) MigrateToLatest(ctx context.Context) error { // First handle the idiosyncrasies of postgres and cockroach migrations. Postgres // will need to create any schemas specified in the search path, and cockroach // will need to create the database it was told to connect to. These things should // not really be here, and instead should be assumed to exist. // This is tracked in jira ticket SM-200 switch db.impl { case dbutil.Postgres: schema, err := pgutil.ParseSchemaFromConnstr(db.source) if err != nil { return errs.New("error parsing schema: %+v", err) } if schema != "" { err = pgutil.CreateSchema(ctx, db, schema) if err != nil { return errs.New("error creating schema: %+v", err) } } case dbutil.Cockroach: var dbName string if err := db.QueryRow(ctx, `SELECT current_database();`).Scan(&dbName); err != nil { return errs.New("error querying current database: %+v", err) } _, err := db.Exec(ctx, fmt.Sprintf(`CREATE DATABASE IF NOT EXISTS %s;`, pgutil.QuoteIdentifier(dbName))) if err != nil { return errs.Wrap(err) } } switch db.impl { case dbutil.Postgres, dbutil.Cockroach: migration := db.PostgresMigration() // since we merged migration steps 0-69, the current db version should never be // less than 69 unless the migration hasn't run yet const minDBVersion = 69 dbVersion, err := migration.CurrentVersion(ctx, db.log, db.DB) if err != nil { return errs.New("error current version: %+v", err) } if dbVersion > -1 && dbVersion < minDBVersion { return ErrMigrateMinVersion.New("current database version is %d, it shouldn't be less than the min version %d", dbVersion, minDBVersion, ) } return migration.Run(ctx, db.log.Named("migrate")) default: return migrate.Create(ctx, "database", db.DB) } } // TestingMigrateToLatest is a method for creating all tables for database for testing. func (db *satelliteDB) TestingMigrateToLatest(ctx context.Context) error { switch db.impl { case dbutil.Postgres: schema, err := pgutil.ParseSchemaFromConnstr(db.source) if err != nil { return ErrMigrateMinVersion.New("error parsing schema: %+v", err) } if schema != "" { err = pgutil.CreateSchema(ctx, db, schema) if err != nil { return ErrMigrateMinVersion.New("error creating schema: %+v", err) } } case dbutil.Cockroach: var dbName string if err := db.QueryRow(ctx, `SELECT current_database();`).Scan(&dbName); err != nil { return ErrMigrateMinVersion.New("error querying current database: %+v", err) } _, err := db.Exec(ctx, fmt.Sprintf(`CREATE DATABASE IF NOT EXISTS %s;`, pgutil.QuoteIdentifier(dbName))) if err != nil { return ErrMigrateMinVersion.Wrap(err) } } switch db.impl { case dbutil.Postgres, dbutil.Cockroach: migration := db.PostgresMigration() dbVersion, err := migration.CurrentVersion(ctx, db.log, db.DB) if err != nil { return ErrMigrateMinVersion.Wrap(err) } if dbVersion > -1 { return ErrMigrateMinVersion.New("the database must be empty, got version %d", dbVersion) } testMigration := db.TestPostgresMigration() return testMigration.Run(ctx, db.log.Named("migrate")) default: return migrate.Create(ctx, "database", db.DB) } } // CheckVersion confirms the database is at the desired version. func (db *satelliteDB) CheckVersion(ctx context.Context) error { switch db.impl { case dbutil.Postgres, dbutil.Cockroach: migration := db.PostgresMigration() return migration.ValidateVersions(ctx, db.log) default: return nil } } // TestPostgresMigration returns steps needed for migrating test postgres database. func (db *satelliteDB) TestPostgresMigration() *migrate.Migration { return db.testMigration() } // PostgresMigration returns steps needed for migrating postgres database. func (db *satelliteDB) PostgresMigration() *migrate.Migration { return &migrate.Migration{ Table: "versions", Steps: []*migrate.Step{ { DB: &db.migrationDB, Description: "Initial setup", Version: 103, Action: migrate.SQL{ `CREATE TABLE accounting_rollups ( id bigserial NOT NULL, node_id bytea NOT NULL, start_time timestamp with time zone NOT NULL, put_total bigint NOT NULL, get_total bigint NOT NULL, get_audit_total bigint NOT NULL, get_repair_total bigint NOT NULL, put_repair_total bigint NOT NULL, at_rest_total double precision NOT NULL, PRIMARY KEY ( id ) );`, `CREATE INDEX accounting_rollups_start_time_index ON accounting_rollups ( start_time );`, `CREATE TABLE accounting_timestamps ( name text NOT NULL, value timestamp with time zone NOT NULL, PRIMARY KEY ( name ) );`, `CREATE TABLE bucket_bandwidth_rollups ( bucket_name bytea NOT NULL, interval_start timestamp with time zone NOT NULL, interval_seconds integer NOT NULL, action integer NOT NULL, inline bigint NOT NULL, allocated bigint NOT NULL, settled bigint NOT NULL, project_id bytea NOT NULL , CONSTRAINT bucket_bandwidth_rollups_pk PRIMARY KEY (bucket_name, project_id, interval_start, action) );`, `CREATE INDEX IF NOT EXISTS bucket_bandwidth_rollups_project_id_action_interval_index ON bucket_bandwidth_rollups ( project_id, action, interval_start );`, `CREATE TABLE bucket_storage_tallies ( bucket_name bytea NOT NULL, interval_start timestamp with time zone NOT NULL, inline bigint NOT NULL, remote bigint NOT NULL, remote_segments_count integer NOT NULL, inline_segments_count integer NOT NULL, object_count integer NOT NULL, metadata_size bigint NOT NULL, project_id bytea NOT NULL, CONSTRAINT bucket_storage_tallies_pk PRIMARY KEY (bucket_name, project_id, interval_start) );`, `CREATE TABLE injuredsegments ( data bytea NOT NULL, attempted timestamp with time zone, path bytea NOT NULL, num_healthy_pieces integer DEFAULT 52 NOT NULL, CONSTRAINT injuredsegments_pk PRIMARY KEY (path) );`, `CREATE INDEX injuredsegments_attempted_index ON injuredsegments ( attempted );`, `CREATE INDEX injuredsegments_num_healthy_pieces_index ON injuredsegments ( num_healthy_pieces );`, `CREATE TABLE irreparabledbs ( segmentpath bytea NOT NULL, segmentdetail bytea NOT NULL, pieces_lost_count bigint NOT NULL, seg_damaged_unix_sec bigint NOT NULL, repair_attempt_count bigint NOT NULL, PRIMARY KEY ( segmentpath ) );`, `CREATE TABLE nodes ( id bytea NOT NULL, audit_success_count bigint NOT NULL DEFAULT 0, total_audit_count bigint NOT NULL DEFAULT 0, uptime_success_count bigint NOT NULL, total_uptime_count bigint NOT NULL, created_at timestamp with time zone NOT NULL DEFAULT current_timestamp, updated_at timestamp with time zone NOT NULL DEFAULT current_timestamp, wallet text NOT NULL, email text NOT NULL, address text NOT NULL DEFAULT '', protocol INTEGER NOT NULL DEFAULT 0, type INTEGER NOT NULL DEFAULT 0, free_disk BIGINT NOT NULL DEFAULT -1, latency_90 BIGINT NOT NULL DEFAULT 0, last_contact_success TIMESTAMP WITH TIME ZONE NOT NULL DEFAULT 'epoch', last_contact_failure TIMESTAMP WITH TIME ZONE NOT NULL DEFAULT 'epoch', major bigint NOT NULL DEFAULT 0, minor bigint NOT NULL DEFAULT 0, patch bigint NOT NULL DEFAULT 0, hash TEXT NOT NULL DEFAULT '', timestamp TIMESTAMP WITH TIME ZONE NOT NULL DEFAULT '0001-01-01 00:00:00+00', release bool NOT NULL DEFAULT FALSE, contained bool NOT NULL DEFAULT FALSE, last_net text NOT NULL, disqualified timestamp with time zone, audit_reputation_alpha double precision NOT NULL DEFAULT 1, audit_reputation_beta double precision NOT NULL DEFAULT 0, uptime_reputation_alpha double precision NOT NULL DEFAULT 1, uptime_reputation_beta double precision NOT NULL DEFAULT 0, piece_count bigint NOT NULL DEFAULT 0, exit_loop_completed_at timestamp with time zone, exit_initiated_at timestamp with time zone, exit_finished_at timestamp with time zone, exit_success boolean NOT NULL DEFAULT FALSE, unknown_audit_reputation_alpha double precision NOT NULL DEFAULT 1, unknown_audit_reputation_beta double precision NOT NULL DEFAULT 0, suspended timestamp with time zone, last_ip_port text, vetted_at timestamp with time zone, PRIMARY KEY ( id ) );`, `CREATE INDEX node_last_ip ON nodes ( last_net );`, `CREATE TABLE offers ( id serial NOT NULL, name text NOT NULL, description text NOT NULL, type integer NOT NULL, award_credit_duration_days integer, invitee_credit_duration_days integer, redeemable_cap integer, expires_at timestamp with time zone NOT NULL, created_at timestamp with time zone NOT NULL, status integer NOT NULL, award_credit_in_cents integer NOT NULL DEFAULT 0, invitee_credit_in_cents integer NOT NULL DEFAULT 0, PRIMARY KEY ( id ) );`, `CREATE TABLE peer_identities ( node_id bytea NOT NULL, leaf_serial_number bytea NOT NULL, chain bytea NOT NULL, updated_at timestamp with time zone NOT NULL, PRIMARY KEY ( node_id ) );`, `CREATE TABLE pending_audits ( node_id bytea NOT NULL, piece_id bytea NOT NULL, stripe_index bigint NOT NULL, share_size bigint NOT NULL, expected_share_hash bytea NOT NULL, reverify_count bigint NOT NULL, path bytea NOT NULL, PRIMARY KEY ( node_id ) );`, `CREATE TABLE projects ( id bytea NOT NULL, name text NOT NULL, description text NOT NULL, created_at timestamp with time zone NOT NULL, usage_limit bigint NOT NULL DEFAULT 0, partner_id bytea, owner_id bytea NOT NULL, rate_limit integer, PRIMARY KEY ( id ) );`, `CREATE TABLE registration_tokens ( secret bytea NOT NULL, owner_id bytea UNIQUE, project_limit integer NOT NULL, created_at timestamp with time zone NOT NULL, PRIMARY KEY ( secret ) );`, `CREATE TABLE reset_password_tokens ( secret bytea NOT NULL, owner_id bytea NOT NULL UNIQUE, created_at timestamp with time zone NOT NULL, PRIMARY KEY ( secret ) );`, `CREATE TABLE serial_numbers ( id serial NOT NULL, serial_number bytea NOT NULL, bucket_id bytea NOT NULL, expires_at timestamp with time zone NOT NULL, PRIMARY KEY ( id ) );`, `CREATE INDEX serial_numbers_expires_at_index ON serial_numbers ( expires_at );`, `CREATE UNIQUE INDEX serial_number_index ON serial_numbers ( serial_number )`, `CREATE TABLE storagenode_bandwidth_rollups ( storagenode_id bytea NOT NULL, interval_start timestamp with time zone NOT NULL, interval_seconds integer NOT NULL, action integer NOT NULL, allocated bigint DEFAULT 0, settled bigint NOT NULL, PRIMARY KEY ( storagenode_id, interval_start, action ) );`, `CREATE TABLE storagenode_storage_tallies ( node_id bytea NOT NULL, interval_end_time timestamp with time zone NOT NULL, data_total double precision NOT NULL, CONSTRAINT storagenode_storage_tallies_pkey PRIMARY KEY ( interval_end_time, node_id ) );`, `CREATE INDEX storagenode_storage_tallies_node_id_index ON storagenode_storage_tallies ( node_id );`, `CREATE TABLE users ( id bytea NOT NULL, full_name text NOT NULL, short_name text, email text NOT NULL, password_hash bytea NOT NULL, status integer NOT NULL, created_at timestamp with time zone NOT NULL, partner_id bytea, normalized_email text NOT NULL, PRIMARY KEY ( id ) );`, `CREATE TABLE value_attributions ( bucket_name bytea NOT NULL, partner_id bytea NOT NULL, last_updated timestamp with time zone NOT NULL, project_id bytea NOT NULL, PRIMARY KEY (project_id, bucket_name) );`, `CREATE TABLE api_keys ( id bytea NOT NULL, project_id bytea NOT NULL REFERENCES projects( id ) ON DELETE CASCADE, head bytea NOT NULL UNIQUE, name text NOT NULL, secret bytea NOT NULL, created_at timestamp with time zone NOT NULL, partner_id bytea, PRIMARY KEY ( id ), UNIQUE ( name, project_id ) );`, `CREATE TABLE bucket_metainfos ( id bytea NOT NULL, project_id bytea NOT NULL REFERENCES projects( id ), name bytea NOT NULL, path_cipher integer NOT NULL, created_at timestamp with time zone NOT NULL, default_segment_size integer NOT NULL, default_encryption_cipher_suite integer NOT NULL, default_encryption_block_size integer NOT NULL, default_redundancy_algorithm integer NOT NULL, default_redundancy_share_size integer NOT NULL, default_redundancy_required_shares integer NOT NULL, default_redundancy_repair_shares integer NOT NULL, default_redundancy_optimal_shares integer NOT NULL, default_redundancy_total_shares integer NOT NULL, partner_id bytea, PRIMARY KEY ( id ), UNIQUE ( name, project_id ) );`, `CREATE TABLE project_invoice_stamps ( project_id bytea NOT NULL REFERENCES projects( id ) ON DELETE CASCADE, invoice_id bytea NOT NULL UNIQUE, start_date timestamp with time zone NOT NULL, end_date timestamp with time zone NOT NULL, created_at timestamp with time zone NOT NULL, PRIMARY KEY ( project_id, start_date, end_date ) );`, `CREATE TABLE project_members ( member_id bytea NOT NULL REFERENCES users( id ) ON DELETE CASCADE, project_id bytea NOT NULL REFERENCES projects( id ) ON DELETE CASCADE, created_at timestamp with time zone NOT NULL, PRIMARY KEY ( member_id, project_id ) );`, `CREATE TABLE used_serials ( serial_number_id integer NOT NULL REFERENCES serial_numbers( id ) ON DELETE CASCADE, storage_node_id bytea NOT NULL, PRIMARY KEY ( serial_number_id, storage_node_id ) );`, `CREATE TABLE user_credits ( id serial NOT NULL, user_id bytea NOT NULL REFERENCES users( id ) ON DELETE CASCADE, offer_id integer NOT NULL REFERENCES offers( id ), referred_by bytea REFERENCES users( id ) ON DELETE SET NULL, credits_earned_in_cents integer NOT NULL, credits_used_in_cents integer NOT NULL, expires_at timestamp with time zone NOT NULL, created_at timestamp with time zone NOT NULL, type text NOT NULL, PRIMARY KEY ( id ), UNIQUE (id, offer_id) );`, `CREATE UNIQUE INDEX credits_earned_user_id_offer_id ON user_credits (id, offer_id);`, `INSERT INTO offers ( id, name, description, award_credit_in_cents, invitee_credit_in_cents, expires_at, created_at, status, type, award_credit_duration_days, invitee_credit_duration_days ) VALUES ( 1, 'Default referral offer', 'Is active when no other active referral offer', 300, 600, '2119-03-14 08:28:24.636949+00', '2019-07-14 08:28:24.636949+00', 1, 2, 365, 14 ), ( 2, 'Default free credit offer', 'Is active when no active free credit offer', 0, 300, '2119-03-14 08:28:24.636949+00', '2019-07-14 08:28:24.636949+00', 1, 1, NULL, 14 ) ON CONFLICT DO NOTHING;`, `CREATE TABLE graceful_exit_progress ( node_id bytea NOT NULL, bytes_transferred bigint NOT NULL, updated_at timestamp with time zone NOT NULL, pieces_transferred bigint NOT NULL DEFAULT 0, pieces_failed bigint NOT NULL DEFAULT 0, PRIMARY KEY ( node_id ) );`, `CREATE TABLE graceful_exit_transfer_queue ( node_id bytea NOT NULL, path bytea NOT NULL, piece_num integer NOT NULL, durability_ratio double precision NOT NULL, queued_at timestamp with time zone NOT NULL, requested_at timestamp with time zone, last_failed_at timestamp with time zone, last_failed_code integer, failed_count integer, finished_at timestamp with time zone, root_piece_id bytea, order_limit_send_count integer NOT NULL DEFAULT 0, PRIMARY KEY ( node_id, path, piece_num ) );`, `CREATE TABLE stripe_customers ( user_id bytea NOT NULL, customer_id text NOT NULL UNIQUE, created_at timestamp with time zone NOT NULL, PRIMARY KEY ( user_id ) );`, `CREATE TABLE stripecoinpayments_invoice_project_records ( id bytea NOT NULL, project_id bytea NOT NULL, storage double precision NOT NULL, egress bigint NOT NULL, objects bigint NOT NULL, period_start timestamp with time zone NOT NULL, period_end timestamp with time zone NOT NULL, state integer NOT NULL, created_at timestamp with time zone NOT NULL, PRIMARY KEY ( id ), UNIQUE ( project_id, period_start, period_end ) );`, `CREATE TABLE stripecoinpayments_tx_conversion_rates ( tx_id text NOT NULL, rate bytea NOT NULL, created_at timestamp with time zone NOT NULL, PRIMARY KEY ( tx_id ) );`, `CREATE TABLE coinpayments_transactions ( id text NOT NULL, user_id bytea NOT NULL, address text NOT NULL, amount bytea NOT NULL, received bytea NOT NULL, status integer NOT NULL, key text NOT NULL, timeout integer NOT NULL, created_at timestamp with time zone NOT NULL, PRIMARY KEY ( id ) );`, `CREATE TABLE stripecoinpayments_apply_balance_intents ( tx_id text NOT NULL REFERENCES coinpayments_transactions( id ) ON DELETE CASCADE, state integer NOT NULL, created_at timestamp with time zone NOT NULL, PRIMARY KEY ( tx_id ) );`, `CREATE TABLE nodes_offline_times ( node_id bytea NOT NULL, tracked_at timestamp with time zone NOT NULL, seconds integer NOT NULL, PRIMARY KEY ( node_id, tracked_at ) );`, `CREATE INDEX nodes_offline_times_node_id_index ON nodes_offline_times ( node_id );`, `CREATE TABLE coupons ( id bytea NOT NULL, project_id bytea NOT NULL, user_id bytea NOT NULL, amount bigint NOT NULL, description text NOT NULL, type integer NOT NULL, status integer NOT NULL, duration bigint NOT NULL, created_at timestamp with time zone NOT NULL, PRIMARY KEY ( id ) );`, `CREATE TABLE coupon_usages ( coupon_id bytea NOT NULL, amount bigint NOT NULL, status integer NOT NULL, period timestamp with time zone NOT NULL, PRIMARY KEY ( coupon_id, period ) );`, `CREATE TABLE reported_serials ( expires_at timestamp with time zone NOT NULL, storage_node_id bytea NOT NULL, bucket_id bytea NOT NULL, action integer NOT NULL, serial_number bytea NOT NULL, settled bigint NOT NULL, observed_at timestamp with time zone NOT NULL, PRIMARY KEY ( expires_at, storage_node_id, bucket_id, action, serial_number ) );`, `CREATE TABLE credits ( user_id bytea NOT NULL, transaction_id text NOT NULL, amount bigint NOT NULL, created_at timestamp with time zone NOT NULL, PRIMARY KEY ( transaction_id ) );`, `CREATE TABLE credits_spendings ( id bytea NOT NULL, user_id bytea NOT NULL, project_id bytea NOT NULL, amount bigint NOT NULL, status int NOT NULL, created_at timestamp with time zone NOT NULL, PRIMARY KEY ( id ) );`, `CREATE TABLE consumed_serials ( storage_node_id bytea NOT NULL, serial_number bytea NOT NULL, expires_at timestamp with time zone NOT NULL, PRIMARY KEY ( storage_node_id, serial_number ) );`, `CREATE INDEX consumed_serials_expires_at_index ON consumed_serials ( expires_at );`, `CREATE TABLE pending_serial_queue ( storage_node_id bytea NOT NULL, bucket_id bytea NOT NULL, serial_number bytea NOT NULL, action integer NOT NULL, settled bigint NOT NULL, expires_at timestamp with time zone NOT NULL, PRIMARY KEY ( storage_node_id, bucket_id, serial_number ) );`, `CREATE TABLE storagenode_payments ( id bigserial NOT NULL, created_at timestamp with time zone NOT NULL, node_id bytea NOT NULL, period text NOT NULL, amount bigint NOT NULL, receipt text, notes text, PRIMARY KEY ( id ) );`, `CREATE INDEX storagenode_payments_node_id_period_index ON storagenode_payments ( node_id, period );`, `CREATE TABLE storagenode_paystubs ( period text NOT NULL, node_id bytea NOT NULL, created_at timestamp with time zone NOT NULL, codes text NOT NULL, usage_at_rest double precision NOT NULL, usage_get bigint NOT NULL, usage_put bigint NOT NULL, usage_get_repair bigint NOT NULL, usage_put_repair bigint NOT NULL, usage_get_audit bigint NOT NULL, comp_at_rest bigint NOT NULL, comp_get bigint NOT NULL, comp_put bigint NOT NULL, comp_get_repair bigint NOT NULL, comp_put_repair bigint NOT NULL, comp_get_audit bigint NOT NULL, surge_percent bigint NOT NULL, held bigint NOT NULL, owed bigint NOT NULL, disposed bigint NOT NULL, paid bigint NOT NULL, PRIMARY KEY ( period, node_id ) );`, `CREATE INDEX storagenode_paystubs_node_id_index ON storagenode_paystubs ( node_id );`, }, }, { DB: &db.migrationDB, Description: "Add missing bucket_bandwidth_rollups_action_interval_project_id_index index", Version: 104, Action: migrate.SQL{ `CREATE INDEX IF NOT EXISTS bucket_bandwidth_rollups_action_interval_project_id_index ON bucket_bandwidth_rollups(action, interval_start, project_id );`, }, }, { DB: &db.migrationDB, Description: "Remove all nodes from suspension mode.", Version: 105, Action: migrate.SQL{ `UPDATE nodes SET suspended=NULL;`, }, }, { DB: &db.migrationDB, Description: "Add project_bandwidth_rollup table and populate with current months data", Version: 106, Action: migrate.SQL{ `CREATE TABLE IF NOT EXISTS project_bandwidth_rollups ( project_id bytea NOT NULL, interval_month date NOT NULL, egress_allocated bigint NOT NULL, PRIMARY KEY ( project_id, interval_month ) ); INSERT INTO project_bandwidth_rollups(project_id, interval_month, egress_allocated) ( SELECT project_id, date_trunc('MONTH',now())::DATE, sum(allocated)::bigint FROM bucket_bandwidth_rollups WHERE action = 2 AND interval_start >= date_trunc('MONTH',now())::timestamp group by project_id) ON CONFLICT(project_id, interval_month) DO UPDATE SET egress_allocated = EXCLUDED.egress_allocated::bigint;`, }, }, { DB: &db.migrationDB, Description: "add separate bandwidth column", Version: 107, Action: migrate.SQL{ `ALTER TABLE projects ADD COLUMN bandwidth_limit bigint NOT NULL DEFAULT 0;`, }, }, { DB: &db.migrationDB, Description: "backfill bandwidth column with previous limits", Version: 108, SeparateTx: true, Action: migrate.SQL{ `UPDATE projects SET bandwidth_limit = usage_limit;`, }, }, { DB: &db.migrationDB, Description: "add period column to the credits_spendings table (step 1)", Version: 109, SeparateTx: true, Action: migrate.SQL{ `ALTER TABLE credits_spendings ADD COLUMN period timestamp with time zone;`, }, }, { DB: &db.migrationDB, Description: "add period column to the credits_spendings table (step 2)", Version: 110, SeparateTx: true, Action: migrate.SQL{ `UPDATE credits_spendings SET period = 'epoch';`, }, }, { DB: &db.migrationDB, Description: "add period column to the credits_spendings table (step 3)", Version: 111, SeparateTx: true, Action: migrate.SQL{ `ALTER TABLE credits_spendings ALTER COLUMN period SET NOT NULL;`, }, }, { DB: &db.migrationDB, Description: "fix incorrect calculations on backported paystub data", Version: 112, Action: migrate.SQL{` UPDATE storagenode_paystubs SET comp_at_rest = ( ((owed + held - disposed)::float / GREATEST(surge_percent::float / 100, 1))::int - comp_get - comp_get_repair - comp_get_audit ) WHERE ( abs( ((owed + held - disposed)::float / GREATEST(surge_percent::float / 100, 1))::int - comp_get - comp_get_repair - comp_get_audit ) >= 10 OR comp_at_rest < 0 ) AND codes not like '%O%' AND codes not like '%D%' AND period < '2020-03' `}, }, { DB: &db.migrationDB, Description: "drop project_id column from coupon table", Version: 113, Action: migrate.SQL{ `ALTER TABLE coupons DROP COLUMN project_id;`, }, }, { DB: &db.migrationDB, Description: "add new columns for suspension to node tables", Version: 114, Action: migrate.SQL{ `ALTER TABLE nodes ADD COLUMN unknown_audit_suspended TIMESTAMP WITH TIME ZONE;`, `ALTER TABLE nodes ADD COLUMN offline_suspended TIMESTAMP WITH TIME ZONE;`, `ALTER TABLE nodes ADD COLUMN under_review TIMESTAMP WITH TIME ZONE;`, }, }, { DB: &db.migrationDB, Description: "add revocations database", Version: 115, Action: migrate.SQL{` CREATE TABLE revocations ( revoked bytea NOT NULL, api_key_id bytea NOT NULL, PRIMARY KEY ( revoked ) ); `}, }, { DB: &db.migrationDB, Description: "add audit histories database", Version: 116, Action: migrate.SQL{ `CREATE TABLE audit_histories ( node_id bytea NOT NULL, history bytea NOT NULL, PRIMARY KEY ( node_id ) );`, }, }, { DB: &db.migrationDB, Description: "add node_api_versions table", Version: 117, Action: migrate.SQL{` CREATE TABLE node_api_versions ( id bytea NOT NULL, api_version integer NOT NULL, created_at timestamp with time zone NOT NULL, updated_at timestamp with time zone NOT NULL, PRIMARY KEY ( id ) ); `}, }, { DB: &db.migrationDB, Description: "add max_buckets field to projects and an implicit index on bucket_metainfos project_id,name", SeparateTx: true, Version: 118, Action: migrate.SQL{ `ALTER TABLE projects ADD COLUMN max_buckets INTEGER NOT NULL DEFAULT 0;`, `ALTER TABLE bucket_metainfos ADD UNIQUE (project_id, name);`, }, }, { DB: &db.migrationDB, Description: "add project_limit field to users table", Version: 119, Action: migrate.SQL{ `ALTER TABLE users ADD COLUMN project_limit INTEGER NOT NULL DEFAULT 0;`, }, }, { DB: &db.migrationDB, Description: "back fill user project limits from existing registration tokens", Version: 120, SeparateTx: true, Action: migrate.SQL{ `UPDATE users SET project_limit = registration_tokens.project_limit FROM registration_tokens WHERE users.id = registration_tokens.owner_id;`, }, }, { DB: &db.migrationDB, Description: "drop tables related to credits (old deposit bonuses)", Version: 121, Action: migrate.SQL{ `DROP TABLE credits;`, `DROP TABLE credits_spendings;`, }, }, { DB: &db.migrationDB, Description: "drop project_invoice_stamps table", Version: 122, Action: migrate.SQL{ `DROP TABLE project_invoice_stamps;`, }, }, { DB: &db.migrationDB, Description: "drop project_invoice_stamps table", Version: 123, Action: migrate.SQL{ `ALTER TABLE nodes ADD COLUMN online_score double precision NOT NULL DEFAULT 1;`, }, }, { DB: &db.migrationDB, Description: "add column and index updated_at to injuredsegments", Version: 124, Action: migrate.SQL{ `ALTER TABLE injuredsegments ADD COLUMN updated_at timestamp with time zone NOT NULL DEFAULT current_timestamp;`, `CREATE INDEX injuredsegments_updated_at_index ON injuredsegments ( updated_at );`, }, }, { DB: &db.migrationDB, Description: "make limit columns nullable", Version: 125, SeparateTx: true, Action: migrate.SQL{ `ALTER TABLE projects ALTER COLUMN max_buckets DROP NOT NULL;`, `ALTER TABLE projects ALTER COLUMN max_buckets SET DEFAULT 100;`, `ALTER TABLE projects ALTER COLUMN usage_limit DROP NOT NULL;`, `ALTER TABLE projects ALTER COLUMN usage_limit SET DEFAULT 50000000000;`, `ALTER TABLE projects ALTER COLUMN bandwidth_limit DROP NOT NULL;`, `ALTER TABLE projects ALTER COLUMN bandwidth_limit SET DEFAULT 50000000000;`, }, }, { DB: &db.migrationDB, Description: "set 0 limits back to default", Version: 126, Action: migrate.SQL{ `UPDATE projects SET max_buckets = 100 WHERE max_buckets = 0;`, `UPDATE projects SET usage_limit = 50000000000 WHERE usage_limit = 0;`, `UPDATE projects SET bandwidth_limit = 50000000000 WHERE bandwidth_limit = 0;`, }, }, { DB: &db.migrationDB, Description: "enable multiple projects for existing users", Version: 127, Action: migrate.SQL{ `UPDATE users SET project_limit=0 WHERE project_limit <= 10 and project_limit > 0;`, }, }, { DB: &db.migrationDB, Description: "drop default values for project limits", Version: 128, SeparateTx: true, Action: migrate.SQL{ `ALTER TABLE projects ALTER COLUMN max_buckets DROP DEFAULT;`, `ALTER TABLE projects ALTER COLUMN usage_limit DROP DEFAULT;`, `ALTER TABLE projects ALTER COLUMN bandwidth_limit DROP DEFAULT;`, }, }, { DB: &db.migrationDB, Description: "reset everyone with default rate limits to NULL", Version: 129, SeparateTx: true, Action: migrate.SQL{ `UPDATE projects SET max_buckets = NULL WHERE max_buckets <= 100;`, `UPDATE projects SET usage_limit = NULL WHERE usage_limit <= 50000000000;`, `UPDATE projects SET bandwidth_limit = NULL WHERE bandwidth_limit <= 50000000000;`, }, }, { DB: &db.migrationDB, Description: "add index for the gracefule exit transfer queue", Version: 130, SeparateTx: true, Action: migrate.SQL{ `CREATE INDEX IF NOT EXISTS graceful_exit_transfer_queue_nid_dr_qa_fa_lfa_index ON graceful_exit_transfer_queue ( node_id, durability_ratio, queued_at, finished_at, last_failed_at );`, }, }, { DB: &db.migrationDB, Description: "create table storagenode_bandwidth_rollups_phase2", Version: 131, SeparateTx: true, Action: migrate.SQL{ `CREATE TABLE storagenode_bandwidth_rollups_phase2 ( storagenode_id bytea NOT NULL, interval_start timestamp with time zone NOT NULL, interval_seconds integer NOT NULL, action integer NOT NULL, allocated bigint DEFAULT 0, settled bigint NOT NULL, PRIMARY KEY ( storagenode_id, interval_start, action ) );`, }, }, { DB: &db.migrationDB, Description: "add injuredsegments.segment_health", Version: 132, SeparateTx: true, Action: migrate.SQL{ `ALTER TABLE injuredsegments ADD COLUMN segment_health double precision NOT NULL DEFAULT 1;`, `CREATE INDEX injuredsegments_segment_health_index ON injuredsegments ( segment_health );`, }, }, { DB: &db.migrationDB, Description: "Use node_id and start_time for accounting_rollups pkey instead of autogenerated id", Version: 133, SeparateTx: true, Action: migrate.Func(func(ctx context.Context, log *zap.Logger, db tagsql.DB, tx tagsql.Tx) error { if _, ok := db.Driver().(*cockroachutil.Driver); ok { _, err := db.Exec(ctx, `ALTER TABLE accounting_rollups RENAME TO accounting_rollups_original;`, ) if err != nil { return ErrMigrate.Wrap(err) } _, err = db.Exec(ctx, `CREATE TABLE accounting_rollups ( node_id bytea NOT NULL, start_time timestamp with time zone NOT NULL, put_total bigint NOT NULL, get_total bigint NOT NULL, get_audit_total bigint NOT NULL, get_repair_total bigint NOT NULL, put_repair_total bigint NOT NULL, at_rest_total double precision NOT NULL, PRIMARY KEY ( node_id, start_time ) ); CREATE INDEX accounting_rollups_start_time_index ON accounting_rollups ( start_time ); INSERT INTO accounting_rollups ( node_id, start_time, put_total, get_total, get_audit_total, get_repair_total, put_repair_total, at_rest_total ) SELECT node_id, start_time, SUM(put_total)::bigint, SUM(get_total)::bigint, SUM(get_audit_total)::bigint, SUM(get_repair_total)::bigint, SUM(put_repair_total)::bigint, SUM(at_rest_total) FROM accounting_rollups_original GROUP BY node_id, start_time; DROP TABLE accounting_rollups_original;`, ) if err != nil { return ErrMigrate.Wrap(err) } return nil } _, err := db.Exec(ctx, `CREATE TABLE accounting_rollups_new ( node_id bytea NOT NULL, start_time timestamp with time zone NOT NULL, put_total bigint NOT NULL, get_total bigint NOT NULL, get_audit_total bigint NOT NULL, get_repair_total bigint NOT NULL, put_repair_total bigint NOT NULL, at_rest_total double precision NOT NULL, PRIMARY KEY ( node_id, start_time ) ); DROP INDEX accounting_rollups_start_time_index; CREATE INDEX accounting_rollups_start_time_index ON accounting_rollups_new ( start_time ); INSERT INTO accounting_rollups_new ( node_id, start_time, put_total, get_total, get_audit_total, get_repair_total, put_repair_total, at_rest_total ) SELECT node_id, start_time, SUM(put_total), SUM(get_total), SUM(get_audit_total), SUM(get_repair_total), SUM(put_repair_total), SUM(at_rest_total) FROM accounting_rollups GROUP BY node_id, start_time; DROP TABLE accounting_rollups; ALTER INDEX accounting_rollups_new_pkey RENAME TO accounting_rollups_pkey; ALTER TABLE accounting_rollups_new RENAME TO accounting_rollups;`, ) if err != nil { return ErrMigrate.Wrap(err) } return nil }), }, { DB: &db.migrationDB, Description: "drop num_healthy_pieces column from injuredsegments", Version: 134, Action: migrate.SQL{ `ALTER TABLE injuredsegments DROP COLUMN num_healthy_pieces;`, }, }, { DB: &db.migrationDB, Description: "create new index on storagenode_bandwidth_rollups to improve get nodes since query.", Version: 135, SeparateTx: true, Action: migrate.SQL{ `CREATE INDEX IF NOT EXISTS storagenode_bandwidth_rollups_interval_start_index ON storagenode_bandwidth_rollups ( interval_start );`, }, }, { DB: &db.migrationDB, Description: "create new index on bucket_storage_tallies to improve get tallies by project ID.", Version: 136, SeparateTx: true, Action: migrate.SQL{ `CREATE INDEX IF NOT EXISTS bucket_storage_tallies_project_id_index ON bucket_storage_tallies (project_id);`, }, }, { DB: &db.migrationDB, Description: "create new index on nodes to improve queries using disqualified, unknown_audit_suspended, exit_finished_at, and last_contact_success.", Version: 137, SeparateTx: true, Action: migrate.SQL{ `CREATE INDEX IF NOT EXISTS nodes_dis_unk_exit_fin_last_success_index ON nodes(disqualified, unknown_audit_suspended, exit_finished_at, last_contact_success);`, }, }, { DB: &db.migrationDB, Description: "drop node_offline_times table", Version: 138, SeparateTx: true, Action: migrate.SQL{ `DROP TABLE nodes_offline_times;`, }, }, { DB: &db.migrationDB, Description: "set default on uptime count columns", Version: 139, Action: migrate.SQL{ `ALTER TABLE nodes ALTER COLUMN uptime_success_count SET DEFAULT 0;`, `ALTER TABLE nodes ALTER COLUMN total_uptime_count SET DEFAULT 0;`, }, }, { DB: &db.migrationDB, Description: "add distributed column to storagenode_paystubs table", Version: 140, Action: migrate.Func(func(ctx context.Context, log *zap.Logger, db tagsql.DB, tx tagsql.Tx) error { _, err := db.Exec(ctx, ` ALTER TABLE storagenode_paystubs ADD COLUMN distributed BIGINT; `) if err != nil { return ErrMigrate.Wrap(err) } _, err = db.Exec(ctx, ` UPDATE storagenode_paystubs ps SET distributed = coalesce(( SELECT sum(amount)::bigint FROM storagenode_payments pm WHERE pm.period = ps.period AND pm.node_id = ps.node_id ), 0); `) if err != nil { return ErrMigrate.Wrap(err) } _, err = db.Exec(ctx, ` ALTER TABLE storagenode_paystubs ALTER COLUMN distributed SET NOT NULL; `) if err != nil { return ErrMigrate.Wrap(err) } return nil }), }, { DB: &db.migrationDB, Description: "add columns for professional users", Version: 141, Action: migrate.SQL{ `ALTER TABLE users ADD COLUMN position text;`, `ALTER TABLE users ADD COLUMN company_name text;`, `ALTER TABLE users ADD COLUMN working_on text;`, `ALTER TABLE users ADD COLUMN company_size int;`, `ALTER TABLE users ADD COLUMN is_professional boolean NOT NULL DEFAULT false;`, }, }, { DB: &db.migrationDB, Description: "drop the obsolete (name, project_id) index from bucket_metainfos table.", Version: 142, Action: migrate.Func(func(ctx context.Context, log *zap.Logger, db tagsql.DB, tx tagsql.Tx) error { if _, ok := db.Driver().(*cockroachutil.Driver); ok { _, err := db.Exec(ctx, `DROP INDEX bucket_metainfos_name_project_id_key CASCADE;`, ) if err != nil { return ErrMigrate.Wrap(err) } return nil } _, err := db.Exec(ctx, `ALTER TABLE bucket_metainfos DROP CONSTRAINT bucket_metainfos_name_project_id_key;`, ) if err != nil { return ErrMigrate.Wrap(err) } return nil }), }, { DB: &db.migrationDB, Description: "add storagenode_bandwidth_rollups_archives and bucket_bandwidth_rollup_archives", Version: 143, SeparateTx: true, Action: migrate.SQL{ ` CREATE TABLE storagenode_bandwidth_rollup_archives ( storagenode_id bytea NOT NULL, interval_start timestamp with time zone NOT NULL, interval_seconds integer NOT NULL, action integer NOT NULL, allocated bigint DEFAULT 0, settled bigint NOT NULL, PRIMARY KEY ( storagenode_id, interval_start, action ) );`, `CREATE TABLE bucket_bandwidth_rollup_archives ( bucket_name bytea NOT NULL, project_id bytea NOT NULL, interval_start timestamp with time zone NOT NULL, interval_seconds integer NOT NULL, action integer NOT NULL, inline bigint NOT NULL, allocated bigint NOT NULL, settled bigint NOT NULL, PRIMARY KEY ( bucket_name, project_id, interval_start, action ) );`, `CREATE INDEX bucket_bandwidth_rollups_archive_project_id_action_interval_index ON bucket_bandwidth_rollup_archives ( project_id, action, interval_start );`, `CREATE INDEX bucket_bandwidth_rollups_archive_action_interval_project_id_index ON bucket_bandwidth_rollup_archives ( action, interval_start, project_id );`, `CREATE INDEX storagenode_bandwidth_rollup_archives_interval_start_index ON storagenode_bandwidth_rollup_archives (interval_start);`, }, }, { DB: &db.migrationDB, Description: "delete deprecated and unused serial tables", Version: 144, Action: migrate.SQL{ `DROP TABLE used_serials;`, `DROP TABLE reported_serials;`, `DROP TABLE pending_serial_queue;`, `DROP TABLE serial_numbers;`, `DROP TABLE consumed_serials;`, }, }, { DB: &db.migrationDB, Description: "create new index on bucket_storage_tallies to improve get tallies by project ID and bucket name lookups by time interval. This replaces bucket_storage_tallies_project_id_index.", Version: 145, SeparateTx: true, Action: migrate.SQL{ ` CREATE INDEX IF NOT EXISTS bucket_storage_tallies_project_id_interval_start_index ON bucket_storage_tallies ( project_id, interval_start ); DROP INDEX IF EXISTS bucket_storage_tallies_project_id_index; `, }, }, { DB: &db.migrationDB, Description: "nodes add wallet_features column", Version: 146, Action: migrate.SQL{ `ALTER TABLE nodes ADD COLUMN wallet_features text NOT NULL DEFAULT '';`, }, }, { DB: &db.migrationDB, Description: "add employee_count column on users", Version: 147, Action: migrate.SQL{ `ALTER TABLE users ADD COLUMN employee_count text;`, }, }, { DB: &db.migrationDB, Description: "empty migration to fix backwards compat test discrepancy with release tag", Version: 148, Action: migrate.SQL{}, }, { DB: &db.migrationDB, Description: "add coupon_codes table and add nullable coupon_code_name to coupons table", Version: 149, Action: migrate.SQL{ `CREATE TABLE coupon_codes ( id bytea NOT NULL, name text NOT NULL, amount bigint NOT NULL, description text NOT NULL, type integer NOT NULL, duration bigint NOT NULL, created_at timestamp with time zone NOT NULL, PRIMARY KEY ( id ), UNIQUE ( name ) );`, `ALTER TABLE coupons ADD COLUMN coupon_code_name text;`, }, }, { DB: &db.migrationDB, Description: "drop columns uptime_reputation_alpha and uptime_reputation_beta", Version: 150, Action: migrate.SQL{ `ALTER TABLE nodes DROP COLUMN uptime_reputation_alpha;`, `ALTER TABLE nodes DROP COLUMN uptime_reputation_beta;`, }, }, { DB: &db.migrationDB, Description: "set default project and usage limits on existing users", Version: 151, Action: migrate.SQL{ `UPDATE users SET project_limit = 10 WHERE project_limit = 0;`, // 500 GB = 5e11 bytes `UPDATE projects SET usage_limit = 500000000000 WHERE usage_limit IS NULL;`, `UPDATE projects SET bandwidth_limit = 500000000000 WHERE bandwidth_limit IS NULL;`, }, }, { DB: &db.migrationDB, Description: "add offline_suspended to index on nodes", Version: 152, SeparateTx: true, Action: migrate.SQL{ `CREATE INDEX IF NOT EXISTS nodes_dis_unk_off_exit_fin_last_success_index ON nodes (disqualified, unknown_audit_suspended, offline_suspended, exit_finished_at, last_contact_success);`, `DROP INDEX IF EXISTS nodes_dis_unk_exit_fin_last_success_index;`, }, }, { DB: &db.migrationDB, Description: "add nullable coupons.duration_new, migrate coupon_codes.duration to be nullable", Version: 153, Action: migrate.SQL{ `ALTER TABLE coupons ADD COLUMN billing_periods bigint;`, `ALTER TABLE coupon_codes ADD COLUMN billing_periods bigint;`, `ALTER TABLE coupon_codes DROP COLUMN duration;`, }, }, { DB: &db.migrationDB, Description: "move duration over to billing_periods", Version: 154, Action: migrate.SQL{ `UPDATE coupons SET billing_periods = duration;`, }, }, { DB: &db.migrationDB, Description: "add have_sales_contact column on users", Version: 155, Action: migrate.SQL{ `ALTER TABLE users ADD COLUMN have_sales_contact boolean NOT NULL DEFAULT false;`, }, }, { DB: &db.migrationDB, Description: "create indexes used in production", Version: 156, Action: migrate.Func(func(ctx context.Context, log *zap.Logger, _ tagsql.DB, tx tagsql.Tx) error { storingClause := func(fields ...string) string { if db.impl == dbutil.Cockroach { return fmt.Sprintf("STORING (%s)", strings.Join(fields, ", ")) } return "" } indexes := [3]string{ `CREATE INDEX IF NOT EXISTS injuredsegments_num_healthy_pieces_attempted_index ON injuredsegments (segment_health, attempted NULLS FIRST)`, `CREATE INDEX IF NOT EXISTS nodes_type_last_cont_success_free_disk_ma_mi_patch_vetted_partial_index ON nodes (type, last_contact_success, free_disk, major, minor, patch, vetted_at) ` + storingClause("last_net", "address", "last_ip_port") + ` WHERE disqualified IS NULL AND unknown_audit_suspended IS NULL AND exit_initiated_at IS NULL AND release = true AND last_net != ''`, `CREATE INDEX IF NOT EXISTS nodes_dis_unk_aud_exit_init_rel_type_last_cont_success_stored_index ON nodes (disqualified ASC, unknown_audit_suspended ASC, exit_initiated_at ASC, release ASC, type ASC, last_contact_success DESC) ` + storingClause("free_disk", "minor", "major", "patch", "vetted_at", "last_net", "address", "last_ip_port") + ` WHERE disqualified IS NULL AND unknown_audit_suspended IS NULL AND exit_initiated_at IS NULL AND release = true`, } for _, s := range indexes { _, err := tx.ExecContext(ctx, s) if err != nil { return err } } return nil }), }, { DB: &db.migrationDB, Description: "drop unused columns total_uptime_count and uptime_success_count on nodes table", Version: 157, Action: migrate.SQL{ `ALTER TABLE nodes DROP COLUMN total_uptime_count;`, `ALTER TABLE nodes DROP COLUMN uptime_success_count;`, }, }, { DB: &db.migrationDB, Description: "create new table for computing project bandwidth daily usage.", Version: 158, Action: migrate.SQL{ `CREATE TABLE project_bandwidth_daily_rollups ( project_id bytea NOT NULL, interval_day date NOT NULL, egress_allocated bigint NOT NULL, egress_settled bigint NOT NULL, PRIMARY KEY ( project_id, interval_day ) );`, }, }, { DB: &db.migrationDB, Description: "migrate non-expiring coupons to expire in 2 billing periods", Version: 159, Action: migrate.SQL{ `UPDATE coupons SET billing_periods = 2 WHERE billing_periods is NULL;`, }, }, { DB: &db.migrationDB, Description: "add column to track dead allocated bandwidth", Version: 160, Action: migrate.SQL{ `ALTER TABLE project_bandwidth_daily_rollups ADD COLUMN egress_dead bigint NOT NULL DEFAULT 0;`, }, }, { DB: &db.migrationDB, Description: "add table for node reputation", Version: 161, Action: migrate.SQL{ `CREATE TABLE reputations ( id bytea NOT NULL, audit_success_count bigint NOT NULL DEFAULT 0, total_audit_count bigint NOT NULL DEFAULT 0, vetted_at timestamp with time zone, created_at timestamp with time zone NOT NULL DEFAULT current_timestamp, updated_at timestamp with time zone NOT NULL DEFAULT current_timestamp, contained boolean NOT NULL DEFAULT false, disqualified timestamp with time zone, suspended timestamp with time zone, unknown_audit_suspended timestamp with time zone, offline_suspended timestamp with time zone, under_review timestamp with time zone, online_score double precision NOT NULL DEFAULT 1, audit_history bytea NOT NULL, audit_reputation_alpha double precision NOT NULL DEFAULT 1, audit_reputation_beta double precision NOT NULL DEFAULT 0, unknown_audit_reputation_alpha double precision NOT NULL DEFAULT 1, unknown_audit_reputation_beta double precision NOT NULL DEFAULT 0, PRIMARY KEY ( id ) );`, }, }, { DB: &db.migrationDB, Description: "add stream_id and position columns to graceful_exit_transfer_queue", Version: 162, Action: migrate.SQL{ `CREATE TABLE graceful_exit_segment_transfer_queue ( node_id bytea NOT NULL, stream_id bytea NOT NULL, position bigint NOT NULL, piece_num integer NOT NULL, root_piece_id bytea, durability_ratio double precision NOT NULL, queued_at timestamp with time zone NOT NULL, requested_at timestamp with time zone, last_failed_at timestamp with time zone, last_failed_code integer, failed_count integer, finished_at timestamp with time zone, order_limit_send_count integer NOT NULL DEFAULT 0, PRIMARY KEY ( node_id, stream_id, position, piece_num ) );`, `CREATE INDEX graceful_exit_segment_transfer_nid_dr_qa_fa_lfa_index ON graceful_exit_segment_transfer_queue ( node_id, durability_ratio, queued_at, finished_at, last_failed_at ) ;`, `ALTER TABLE graceful_exit_progress ADD COLUMN uses_segment_transfer_queue boolean NOT NULL DEFAULT false;`, }, }, { DB: &db.migrationDB, Description: "create segment_pending_audits table, replacement for pending_audits", Version: 163, Action: migrate.SQL{ `CREATE TABLE segment_pending_audits ( node_id bytea NOT NULL, stream_id bytea NOT NULL, position bigint NOT NULL, piece_id bytea NOT NULL, stripe_index bigint NOT NULL, share_size bigint NOT NULL, expected_share_hash bytea NOT NULL, reverify_count bigint NOT NULL, PRIMARY KEY ( node_id ) );`, }, }, { DB: &db.migrationDB, Description: "add paid_tier column to users table", Version: 164, Action: migrate.SQL{ `ALTER TABLE users ADD COLUMN paid_tier bool NOT NULL DEFAULT false;`, }, }, { DB: &db.migrationDB, Description: "add repair_queue table, replacement for injuredsegments table", Version: 165, Action: migrate.SQL{ `CREATE TABLE repair_queue ( stream_id bytea NOT NULL, position bigint NOT NULL, attempted_at timestamp with time zone, updated_at timestamp with time zone NOT NULL DEFAULT current_timestamp, inserted_at timestamp with time zone NOT NULL DEFAULT current_timestamp, segment_health double precision NOT NULL DEFAULT 1, PRIMARY KEY ( stream_id, position ) )`, `CREATE INDEX repair_queue_updated_at_index ON repair_queue ( updated_at )`, `CREATE INDEX repair_queue_num_healthy_pieces_attempted_at_index ON repair_queue ( segment_health, attempted_at NULLS FIRST)`, }, }, { DB: &db.migrationDB, Description: "add total_bytes table and total_segments_count for bucket_storage_tallies table", Version: 166, Action: migrate.SQL{ `ALTER TABLE bucket_storage_tallies ADD COLUMN total_bytes bigint NOT NULL DEFAULT 0;`, `ALTER TABLE bucket_storage_tallies ADD COLUMN total_segments_count integer NOT NULL DEFAULT 0;`, }, }, { DB: &db.migrationDB, Description: "add multi-factor authentication columns mfa_enabled, mfa_secret_key, and mfa_recovery_codes into users table", Version: 167, Action: migrate.SQL{ `ALTER TABLE users ADD COLUMN mfa_enabled boolean NOT NULL DEFAULT false, ADD COLUMN mfa_secret_key text, ADD COLUMN mfa_recovery_codes text;`, }, }, { DB: &db.migrationDB, Description: "migrate audit score related data from overlaycache into reputations table", Version: 168, Action: migrate.SQL{ `TRUNCATE TABLE reputations;`, `INSERT INTO reputations ( id, audit_success_count, total_audit_count, vetted_at, created_at, updated_at, contained, disqualified, suspended, unknown_audit_suspended, offline_suspended, under_review, online_score, audit_history, audit_reputation_alpha, audit_reputation_beta, unknown_audit_reputation_alpha, unknown_audit_reputation_beta ) SELECT n.id, n.audit_success_count, n.total_audit_count, n.vetted_at, n.created_at, n.updated_at, n.contained, n.disqualified, n.suspended, n.unknown_audit_suspended, n.offline_suspended, n.under_review, n.online_score, audit_histories.history, n.audit_reputation_alpha, n.audit_reputation_beta, n.unknown_audit_reputation_alpha, n.unknown_audit_reputation_beta FROM nodes as n INNER JOIN audit_histories ON n.id = audit_histories.node_id;`, }, }, { DB: &db.migrationDB, Description: "drop tables after metaloop refactoring", Version: 169, Action: migrate.SQL{ `DROP TABLE pending_audits`, `DROP TABLE irreparabledbs`, `DROP TABLE injuredsegments`, }, }, { DB: &db.migrationDB, Description: "drop audit_history table", Version: 170, Action: migrate.SQL{ `DROP TABLE audit_histories`, }, }, { DB: &db.migrationDB, Description: "drop audit and unknown audit reputation alpha and beta from nodes table", Version: 171, Action: migrate.SQL{ `ALTER TABLE nodes DROP COLUMN audit_reputation_alpha;`, `ALTER TABLE nodes DROP COLUMN audit_reputation_beta;`, `ALTER TABLE nodes DROP COLUMN unknown_audit_reputation_alpha;`, `ALTER TABLE nodes DROP COLUMN unknown_audit_reputation_beta;`, `ALTER TABLE nodes DROP COLUMN audit_success_count`, `ALTER TABLE nodes DROP COLUMN online_score`, `ALTER TABLE nodes DROP COLUMN total_audit_count`, }, }, { DB: &db.migrationDB, Description: "add burst_limit to projects table", Version: 172, Action: migrate.SQL{ `ALTER TABLE projects ADD COLUMN burst_limit int;`, }, }, { DB: &db.migrationDB, Description: "drop graceful_exit_transfer_queue table", Version: 173, Action: migrate.SQL{ `DROP TABLE graceful_exit_transfer_queue`, }, }, { DB: &db.migrationDB, Description: "add user_agent bytes to the value_attributions, users, projects, api_keys and bucket_metainfos tables", Version: 174, Action: migrate.SQL{ `ALTER TABLE value_attributions ADD COLUMN user_agent bytea;`, `ALTER TABLE users ADD COLUMN user_agent bytea;`, `ALTER TABLE projects ADD COLUMN user_agent bytea;`, `ALTER TABLE api_keys ADD COLUMN user_agent bytea;`, `ALTER TABLE bucket_metainfos ADD COLUMN user_agent bytea;`}, }, { DB: &db.migrationDB, Description: "drop column uses_segment_transfer_queue from graceful_exit_progress", Version: 175, Action: migrate.SQL{ `ALTER TABLE graceful_exit_progress DROP COLUMN uses_segment_transfer_queue;`, }, }, { DB: &db.migrationDB, Description: "add signup_promo_code column on users", Version: 176, Action: migrate.SQL{ `ALTER TABLE users ADD COLUMN signup_promo_code text;`, }, }, // NB: after updating testdata in `testdata`, run // `go generate` to update `migratez.go`. }, } }