satellite/satellitedb: clarify test migration merging

Use a field to distinguish migration steps that need to use a
different transaction from previous steps. This is clearer than
using a func.

Change-Id: I2147369d05413f3e8ddb50c71a46ab1ba3ab5114
This commit is contained in:
Egon Elbre 2020-06-25 14:23:39 +03:00
parent df8cf8f58a
commit 13a5854535
3 changed files with 94 additions and 72 deletions

View File

@ -67,6 +67,10 @@ type Step struct {
Description string Description string
Version int // Versions should start at 0 Version int // Versions should start at 0
Action Action Action Action
// SeparateTx marks a step as it should not be merged together for optimization.
// Cockroach cannot add a column and update the value in the same transaction.
SeparateTx bool
} }
// Action is something that needs to be done // Action is something that needs to be done

View File

@ -145,14 +145,33 @@ func (db *satelliteDB) CheckVersion(ctx context.Context) error {
} }
} }
// flattenMigration joins the migration sql queries from each migration step // flattenMigration joins the migration sql queries from
// to speed up the database setup // each migration step to speed up the database setup.
//
// Steps with "SeparateTx" end up as separate migration transactions.
// Cockroach requires schema changes and updates to the values of that
// schema change to be in a different transaction.
func flattenMigration(m *migrate.Migration) (*migrate.Migration, error) { func flattenMigration(m *migrate.Migration) (*migrate.Migration, error) {
var db tagsql.DB var db tagsql.DB
var version int var version int
var statements migrate.SQL var statements migrate.SQL
var steps = []*migrate.Step{} var steps = []*migrate.Step{}
pushMerged := func() {
if len(statements) == 0 {
return
}
steps = append(steps, &migrate.Step{
DB: db,
Description: "Setup",
Version: version,
Action: migrate.SQL{strings.Join(statements, ";\n")},
})
statements = nil
}
for _, step := range m.Steps { for _, step := range m.Steps {
if db == nil { if db == nil {
db = step.DB db = step.DB
@ -160,40 +179,21 @@ func flattenMigration(m *migrate.Migration) (*migrate.Migration, error) {
return nil, errs.New("multiple databases not supported") return nil, errs.New("multiple databases not supported")
} }
if sql, ok := step.Action.(migrate.SQL); ok {
if step.SeparateTx {
pushMerged()
}
version = step.Version version = step.Version
statements = append(statements, sql...)
switch action := step.Action.(type) { } else {
case migrate.SQL: pushMerged()
statements = append(statements, action...)
case migrate.Func:
// if a migrate.Func is encountered then create a step with all
// the sql in the previous migration versions
if len(statements) > 0 {
newSQLStep := migrate.Step{
DB: db,
Description: "Setup",
Version: version - 1,
Action: migrate.SQL{strings.Join(statements, ";\n")},
}
steps = append(steps, &newSQLStep)
statements = migrate.SQL{}
}
// then add the migrate.Func step
steps = append(steps, step) steps = append(steps, step)
default:
return nil, errs.New("unexpected action type %T", step.Action)
} }
} }
if len(statements) > 0 { pushMerged()
newSQLStep := migrate.Step{
DB: db,
Description: "Setup",
Version: version,
Action: migrate.SQL{strings.Join(statements, ";\n")},
}
steps = append(steps, &newSQLStep)
}
return &migrate.Migration{ return &migrate.Migration{
Table: "versions", Table: "versions",
Steps: steps, Steps: steps,
@ -868,22 +868,34 @@ func (db *satelliteDB) PostgresMigration() *migrate.Migration {
}, },
}, },
{ {
DB: db.DB, Description: "Use time zones for bucket_bandwidth_rollups", Version: 86, Action: migrate.SQL{ DB: db.DB,
Description: "Use time zones for bucket_bandwidth_rollups",
Version: 86,
Action: migrate.SQL{
`ALTER TABLE bucket_bandwidth_rollups ALTER COLUMN interval_start TYPE TIMESTAMP WITH TIME ZONE USING interval_start AT TIME ZONE current_setting('TIMEZONE');`, `ALTER TABLE bucket_bandwidth_rollups ALTER COLUMN interval_start TYPE TIMESTAMP WITH TIME ZONE USING interval_start AT TIME ZONE current_setting('TIMEZONE');`,
}, },
}, },
{ {
DB: db.DB, Description: "Use time zones for bucket_storage_tallies", Version: 87, Action: migrate.SQL{ DB: db.DB,
Description: "Use time zones for bucket_storage_tallies",
Version: 87,
Action: migrate.SQL{
`ALTER TABLE bucket_storage_tallies ALTER COLUMN interval_start TYPE TIMESTAMP WITH TIME ZONE USING interval_start AT TIME ZONE current_setting('TIMEZONE');`, `ALTER TABLE bucket_storage_tallies ALTER COLUMN interval_start TYPE TIMESTAMP WITH TIME ZONE USING interval_start AT TIME ZONE current_setting('TIMEZONE');`,
}, },
}, },
{ {
DB: db.DB, Description: "Use time zones for graceful_exit_progress", Version: 88, Action: migrate.SQL{ DB: db.DB,
Description: "Use time zones for graceful_exit_progress",
Version: 88,
Action: migrate.SQL{
`ALTER TABLE graceful_exit_progress ALTER COLUMN updated_at TYPE TIMESTAMP WITH TIME ZONE USING updated_at AT TIME ZONE 'UTC';`, `ALTER TABLE graceful_exit_progress ALTER COLUMN updated_at TYPE TIMESTAMP WITH TIME ZONE USING updated_at AT TIME ZONE 'UTC';`,
}, },
}, },
{ {
DB: db.DB, Description: "Use time zones for graceful_exit_transfer_queue", Version: 89, Action: migrate.SQL{ DB: db.DB,
Description: "Use time zones for graceful_exit_transfer_queue",
Version: 89,
Action: migrate.SQL{
`ALTER TABLE graceful_exit_transfer_queue ALTER COLUMN queued_at TYPE TIMESTAMP WITH TIME ZONE USING queued_at AT TIME ZONE 'UTC';`, `ALTER TABLE graceful_exit_transfer_queue ALTER COLUMN queued_at TYPE TIMESTAMP WITH TIME ZONE USING queued_at AT TIME ZONE 'UTC';`,
`ALTER TABLE graceful_exit_transfer_queue ALTER COLUMN requested_at TYPE TIMESTAMP WITH TIME ZONE USING requested_at AT TIME ZONE 'UTC';`, `ALTER TABLE graceful_exit_transfer_queue ALTER COLUMN requested_at TYPE TIMESTAMP WITH TIME ZONE USING requested_at AT TIME ZONE 'UTC';`,
`ALTER TABLE graceful_exit_transfer_queue ALTER COLUMN last_failed_at TYPE TIMESTAMP WITH TIME ZONE USING last_failed_at AT TIME ZONE 'UTC';`, `ALTER TABLE graceful_exit_transfer_queue ALTER COLUMN last_failed_at TYPE TIMESTAMP WITH TIME ZONE USING last_failed_at AT TIME ZONE 'UTC';`,
@ -891,29 +903,44 @@ func (db *satelliteDB) PostgresMigration() *migrate.Migration {
}, },
}, },
{ {
DB: db.DB, Description: "Use time zones for injuredsegments", Version: 90, Action: migrate.SQL{ DB: db.DB,
Description: "Use time zones for injuredsegments",
Version: 90,
Action: migrate.SQL{
`ALTER TABLE injuredsegments ALTER COLUMN attempted TYPE TIMESTAMP WITH TIME ZONE USING attempted AT TIME ZONE 'UTC';`, `ALTER TABLE injuredsegments ALTER COLUMN attempted TYPE TIMESTAMP WITH TIME ZONE USING attempted AT TIME ZONE 'UTC';`,
}, },
}, },
{ {
DB: db.DB, Description: "Use time zones for nodes", Version: 91, Action: migrate.SQL{ DB: db.DB,
Description: "Use time zones for nodes",
Version: 91,
Action: migrate.SQL{
`ALTER TABLE nodes ALTER COLUMN exit_initiated_at TYPE TIMESTAMP WITH TIME ZONE USING exit_initiated_at AT TIME ZONE 'UTC';`, `ALTER TABLE nodes ALTER COLUMN exit_initiated_at TYPE TIMESTAMP WITH TIME ZONE USING exit_initiated_at AT TIME ZONE 'UTC';`,
`ALTER TABLE nodes ALTER COLUMN exit_loop_completed_at TYPE TIMESTAMP WITH TIME ZONE USING exit_loop_completed_at AT TIME ZONE 'UTC';`, `ALTER TABLE nodes ALTER COLUMN exit_loop_completed_at TYPE TIMESTAMP WITH TIME ZONE USING exit_loop_completed_at AT TIME ZONE 'UTC';`,
`ALTER TABLE nodes ALTER COLUMN exit_finished_at TYPE TIMESTAMP WITH TIME ZONE USING exit_finished_at AT TIME ZONE 'UTC';`, `ALTER TABLE nodes ALTER COLUMN exit_finished_at TYPE TIMESTAMP WITH TIME ZONE USING exit_finished_at AT TIME ZONE 'UTC';`,
}, },
}, },
{ {
DB: db.DB, Description: "Use time zones for serial_numbers", Version: 92, Action: migrate.SQL{ DB: db.DB,
Description: "Use time zones for serial_numbers",
Version: 92,
Action: migrate.SQL{
`ALTER TABLE serial_numbers ALTER COLUMN expires_at TYPE TIMESTAMP WITH TIME ZONE USING expires_at AT TIME ZONE 'UTC';`, `ALTER TABLE serial_numbers ALTER COLUMN expires_at TYPE TIMESTAMP WITH TIME ZONE USING expires_at AT TIME ZONE 'UTC';`,
}, },
}, },
{ {
DB: db.DB, Description: "Use time zones for storagenode_bandwidth_rollups", Version: 93, Action: migrate.SQL{ DB: db.DB,
Description: "Use time zones for storagenode_bandwidth_rollups",
Version: 93,
Action: migrate.SQL{
`ALTER TABLE storagenode_bandwidth_rollups ALTER COLUMN interval_start TYPE TIMESTAMP WITH TIME ZONE USING interval_start AT TIME ZONE current_setting('TIMEZONE');`, `ALTER TABLE storagenode_bandwidth_rollups ALTER COLUMN interval_start TYPE TIMESTAMP WITH TIME ZONE USING interval_start AT TIME ZONE current_setting('TIMEZONE');`,
}, },
}, },
{ {
DB: db.DB, Description: "Use time zones for value_attributions", Version: 94, Action: migrate.SQL{ DB: db.DB,
Description: "Use time zones for value_attributions",
Version: 94,
Action: migrate.SQL{
`ALTER TABLE value_attributions ALTER COLUMN last_updated TYPE TIMESTAMP WITH TIME ZONE USING last_updated AT TIME ZONE 'UTC';`, `ALTER TABLE value_attributions ALTER COLUMN last_updated TYPE TIMESTAMP WITH TIME ZONE USING last_updated AT TIME ZONE 'UTC';`,
}, },
}, },
@ -927,7 +954,10 @@ func (db *satelliteDB) PostgresMigration() *migrate.Migration {
}, },
}, },
{ {
DB: db.DB, Description: "Add column last_ip_port to nodes table", Version: 96, Action: migrate.SQL{ DB: db.DB,
Description: "Add column last_ip_port to nodes table",
Version: 96,
Action: migrate.SQL{
`ALTER TABLE nodes ADD COLUMN last_ip_port text`, `ALTER TABLE nodes ADD COLUMN last_ip_port text`,
}, },
}, },
@ -1079,46 +1109,35 @@ func (db *satelliteDB) PostgresMigration() *migrate.Migration {
DB: db.DB, DB: db.DB,
Description: "backfill bandwidth column with previous limits", Description: "backfill bandwidth column with previous limits",
Version: 108, Version: 108,
Action: migrate.Func(func(ctx context.Context, log *zap.Logger, db tagsql.DB, tx tagsql.Tx) error { SeparateTx: true,
// This is in a separate migrate step to prevent TestingMigrateToLatest running it in the same transaction as the previous call. Action: migrate.SQL{
_, err := tx.Exec(ctx,
`UPDATE projects SET bandwidth_limit = usage_limit;`, `UPDATE projects SET bandwidth_limit = usage_limit;`,
) },
return ErrMigrate.Wrap(err)
}),
}, },
{ {
DB: db.DB, DB: db.DB,
Description: "add period column to the credits_spendings table (step 1)", Description: "add period column to the credits_spendings table (step 1)",
Version: 109, Version: 109,
Action: migrate.Func(func(ctx context.Context, log *zap.Logger, db tagsql.DB, tx tagsql.Tx) error { Action: migrate.SQL{
_, err := tx.Exec(ctx,
`ALTER TABLE credits_spendings ADD COLUMN period timestamp with time zone;`, `ALTER TABLE credits_spendings ADD COLUMN period timestamp with time zone;`,
) },
return ErrMigrate.Wrap(err)
}),
}, },
{ {
DB: db.DB, DB: db.DB,
Description: "add period column to the credits_spendings table (step 2)", Description: "add period column to the credits_spendings table (step 2)",
Version: 110, Version: 110,
Action: migrate.Func(func(ctx context.Context, log *zap.Logger, db tagsql.DB, tx tagsql.Tx) error { SeparateTx: true,
_, err := tx.Exec(ctx, Action: migrate.SQL{
`UPDATE credits_spendings SET period = 'epoch';`, `UPDATE credits_spendings SET period = 'epoch';`,
) },
return ErrMigrate.Wrap(err)
}),
}, },
{ {
DB: db.DB, DB: db.DB,
Description: "add period column to the credits_spendings table (step 3)", Description: "add period column to the credits_spendings table (step 3)",
Version: 111, Version: 111,
Action: migrate.Func(func(ctx context.Context, log *zap.Logger, db tagsql.DB, tx tagsql.Tx) error { Action: migrate.SQL{
_, err := tx.Exec(ctx,
`ALTER TABLE credits_spendings ALTER COLUMN period SET NOT NULL;`, `ALTER TABLE credits_spendings ALTER COLUMN period SET NOT NULL;`,
) },
return ErrMigrate.Wrap(err)
}),
}, },
{ {
DB: db.DB, DB: db.DB,

View File

@ -57,7 +57,8 @@ Depending on what your migration is doing, you should then do one of these:
2. Have an `UPDATE "nodes" SET vetted_at = 'fixed date' where id = 'expected id'` in the `-- NEW DATA --` section (so that it runs in both the snapshot, and after the migration has run) and update the main section's `INSERT` for that node. Future snapshot files do not need to retain the `UPDATE` in that case, and the `INSERT` statements can just use the fixed date for the future. 2. Have an `UPDATE "nodes" SET vetted_at = 'fixed date' where id = 'expected id'` in the `-- NEW DATA --` section (so that it runs in both the snapshot, and after the migration has run) and update the main section's `INSERT` for that node. Future snapshot files do not need to retain the `UPDATE` in that case, and the `INSERT` statements can just use the fixed date for the future.
See migration 99 for the specifics, where it chose option 2. See migration 99 for the specifics, where it chose option 2.
3. Cockroach does schema changes asynchronously with regards to a transaction. This means if you need to add a column and fill it with some data, then these need to have them in separate migrations steps:
3. Cockroach does schema changes asynchronously with regards to a transaction. This means if you need to add a column and fill it with some data, then these need to have them in separate migrations steps with using `SeparateTx`:
``` ```
{ {
DB: db.DB, DB: db.DB,
@ -83,12 +84,10 @@ Depending on what your migration is doing, you should then do one of these:
DB: db.DB, DB: db.DB,
Description: "backfill bandwidth column with previous limits", Description: "backfill bandwidth column with previous limits",
Version: 108, Version: 108,
Action: migrate.Func(func(ctx context.Context, log *zap.Logger, db tagsql.DB, tx tagsql.Tx) error { SeparateTx: true,
_, err := tx.Exec(ctx, Action: migrate.SQL{
`UPDATE projects SET bandwidth_limit = usage_limit;`, `UPDATE projects SET bandwidth_limit = usage_limit;`,
) },
return ErrMigrate.Wrap(err)
}),
}, },
``` ```