335 lines
13 KiB
Ruby
335 lines
13 KiB
Ruby
# frozen_string_literal: true
|
|
|
|
module Gitlab
|
|
module Database
|
|
module MigrationHelpers
|
|
module V2
|
|
include Gitlab::Database::MigrationHelpers
|
|
|
|
# Superseded by `create_table` override below
|
|
def create_table_with_constraints(*_)
|
|
raise <<~EOM
|
|
#create_table_with_constraints is not supported anymore - use #create_table instead, for example:
|
|
|
|
create_table :db_guides do |t|
|
|
t.bigint :stars, default: 0, null: false
|
|
t.text :title, limit: 128
|
|
t.text :notes, limit: 1024
|
|
|
|
t.check_constraint 'stars > 1000', name: 'so_many_stars'
|
|
end
|
|
|
|
See https://docs.gitlab.com/ee/development/database/strings_and_the_text_data_type.html
|
|
EOM
|
|
end
|
|
|
|
# Creates a new table, optionally allowing the caller to add text limit constraints to the table.
|
|
# This method only extends Rails' `create_table` method
|
|
#
|
|
# Example:
|
|
#
|
|
# create_table :db_guides do |t|
|
|
# t.bigint :stars, default: 0, null: false
|
|
# t.text :title, limit: 128
|
|
# t.text :notes, limit: 1024
|
|
#
|
|
# t.check_constraint 'stars > 1000', name: 'so_many_stars'
|
|
# end
|
|
#
|
|
# See Rails' `create_table` for more info on the available arguments.
|
|
#
|
|
# When adding foreign keys to other tables, consider wrapping the call into a with_lock_retries block
|
|
# to avoid traffic stalls.
|
|
def create_table(table_name, *args, **kwargs, &block)
|
|
helper_context = self
|
|
|
|
super do |t|
|
|
t.define_singleton_method(:text) do |column_name, **kwargs|
|
|
limit = kwargs.delete(:limit)
|
|
|
|
super(column_name, **kwargs)
|
|
|
|
if limit
|
|
# rubocop:disable GitlabSecurity/PublicSend
|
|
name = helper_context.send(:text_limit_name, table_name, column_name)
|
|
# rubocop:enable GitlabSecurity/PublicSend
|
|
|
|
column_name = helper_context.quote_column_name(column_name)
|
|
definition = "char_length(#{column_name}) <= #{limit}"
|
|
|
|
t.check_constraint(definition, name: name)
|
|
end
|
|
end
|
|
|
|
t.instance_eval(&block) unless block.nil?
|
|
end
|
|
end
|
|
|
|
# Executes the block with a retry mechanism that alters the +lock_timeout+ and +sleep_time+ between attempts.
|
|
# The timings can be controlled via the +timing_configuration+ parameter.
|
|
# If the lock was not acquired within the retry period, a last attempt is made without using +lock_timeout+.
|
|
#
|
|
# In order to retry the block, the method wraps the block into a transaction.
|
|
#
|
|
# When called inside an open transaction it will execute the block directly if lock retries are enabled
|
|
# with `enable_lock_retries!` at migration level, otherwise it will raise an error.
|
|
#
|
|
# ==== Examples
|
|
# # Invoking without parameters
|
|
# with_lock_retries do
|
|
# drop_table :my_table
|
|
# end
|
|
#
|
|
# # Invoking with custom +timing_configuration+
|
|
# t = [
|
|
# [1.second, 1.second],
|
|
# [2.seconds, 2.seconds]
|
|
# ]
|
|
#
|
|
# with_lock_retries(timing_configuration: t) do
|
|
# drop_table :my_table # this will be retried twice
|
|
# end
|
|
#
|
|
# # Disabling the retries using an environment variable
|
|
# > export DISABLE_LOCK_RETRIES=true
|
|
#
|
|
# with_lock_retries do
|
|
# drop_table :my_table # one invocation, it will not retry at all
|
|
# end
|
|
#
|
|
# ==== Parameters
|
|
# * +timing_configuration+ - [[ActiveSupport::Duration, ActiveSupport::Duration], ...] lock timeout for the block, sleep time before the next iteration, defaults to `Gitlab::Database::WithLockRetries::DEFAULT_TIMING_CONFIGURATION`
|
|
# * +logger+ - [Gitlab::JsonLogger]
|
|
# * +env+ - [Hash] custom environment hash, see the example with `DISABLE_LOCK_RETRIES`
|
|
def with_lock_retries(*args, **kwargs, &block)
|
|
if transaction_open?
|
|
if enable_lock_retries?
|
|
Gitlab::AppLogger.warn 'Lock retries already enabled, executing the block directly'
|
|
yield
|
|
else
|
|
raise <<~EOF
|
|
#{__callee__} can not be run inside an already open transaction
|
|
|
|
Use migration-level lock retries instead, see https://docs.gitlab.com/ee/development/migration_style_guide.html#retry-mechanism-when-acquiring-database-locks
|
|
EOF
|
|
end
|
|
else
|
|
super(*args, **kwargs.merge(allow_savepoints: false), &block)
|
|
end
|
|
end
|
|
|
|
# Renames a column without requiring downtime.
|
|
#
|
|
# Concurrent renames work by using database triggers to ensure both the
|
|
# old and new column are in sync. However, this method will _not_ remove
|
|
# the triggers or the old column automatically; this needs to be done
|
|
# manually in a post-deployment migration. This can be done using the
|
|
# method `cleanup_concurrent_column_rename`.
|
|
#
|
|
# table - The name of the database table containing the column.
|
|
# old_column - The old column name.
|
|
# new_column - The new column name.
|
|
# type - The type of the new column. If no type is given the old column's
|
|
# type is used.
|
|
# batch_column_name - option is for tables without primary key, in this
|
|
# case another unique integer column can be used. Example: :user_id
|
|
def rename_column_concurrently(table, old_column, new_column, type: nil, batch_column_name: :id)
|
|
Gitlab::Database::QueryAnalyzers::RestrictAllowedSchemas.require_ddl_mode!
|
|
|
|
setup_renamed_column(__callee__, table, old_column, new_column, type, batch_column_name)
|
|
|
|
with_lock_retries do
|
|
install_bidirectional_triggers(table, old_column, new_column)
|
|
end
|
|
end
|
|
|
|
# Reverses operations performed by rename_column_concurrently.
|
|
#
|
|
# This method takes care of removing previously installed triggers as well
|
|
# as removing the new column.
|
|
#
|
|
# table - The name of the database table.
|
|
# old_column - The name of the old column.
|
|
# new_column - The name of the new column.
|
|
def undo_rename_column_concurrently(table, old_column, new_column)
|
|
teardown_rename_mechanism(table, old_column, new_column, column_to_remove: new_column)
|
|
end
|
|
|
|
# Cleans up a concurrent column name.
|
|
#
|
|
# This method takes care of removing previously installed triggers as well
|
|
# as removing the old column.
|
|
#
|
|
# table - The name of the database table.
|
|
# old_column - The name of the old column.
|
|
# new_column - The name of the new column.
|
|
def cleanup_concurrent_column_rename(table, old_column, new_column)
|
|
teardown_rename_mechanism(table, old_column, new_column, column_to_remove: old_column)
|
|
end
|
|
|
|
# Reverses the operations performed by cleanup_concurrent_column_rename.
|
|
#
|
|
# This method adds back the old_column removed
|
|
# by cleanup_concurrent_column_rename.
|
|
# It also adds back the triggers that are removed
|
|
# by cleanup_concurrent_column_rename.
|
|
#
|
|
# table - The name of the database table containing the column.
|
|
# old_column - The old column name.
|
|
# new_column - The new column name.
|
|
# type - The type of the old column. If no type is given the new column's
|
|
# type is used.
|
|
# batch_column_name - option is for tables without primary key, in this
|
|
# case another unique integer column can be used. Example: :user_id
|
|
#
|
|
def undo_cleanup_concurrent_column_rename(table, old_column, new_column, type: nil, batch_column_name: :id)
|
|
Gitlab::Database::QueryAnalyzers::RestrictAllowedSchemas.require_ddl_mode!
|
|
|
|
setup_renamed_column(__callee__, table, new_column, old_column, type, batch_column_name)
|
|
|
|
with_lock_retries do
|
|
install_bidirectional_triggers(table, old_column, new_column)
|
|
end
|
|
end
|
|
|
|
private
|
|
|
|
def setup_renamed_column(calling_operation, table, old_column, new_column, type, batch_column_name)
|
|
if transaction_open?
|
|
raise "#{calling_operation} can not be run inside a transaction"
|
|
end
|
|
|
|
column = columns(table).find { |column| column.name == old_column.to_s }
|
|
|
|
unless column
|
|
raise "Column #{old_column} does not exist on #{table}"
|
|
end
|
|
|
|
if column.default
|
|
raise "#{calling_operation} does not currently support columns with default values"
|
|
end
|
|
|
|
unless column_exists?(table, batch_column_name)
|
|
raise "Column #{batch_column_name} does not exist on #{table}"
|
|
end
|
|
|
|
check_trigger_permissions!(table)
|
|
|
|
unless column_exists?(table, new_column)
|
|
create_column_from(table, old_column, new_column, type: type, batch_column_name: batch_column_name)
|
|
end
|
|
end
|
|
|
|
def teardown_rename_mechanism(table, old_column, new_column, column_to_remove:)
|
|
return unless column_exists?(table, column_to_remove)
|
|
|
|
with_lock_retries do
|
|
check_trigger_permissions!(table)
|
|
|
|
remove_bidirectional_triggers(table, old_column, new_column)
|
|
|
|
remove_column(table, column_to_remove)
|
|
end
|
|
end
|
|
|
|
def install_bidirectional_triggers(table, old_column, new_column)
|
|
insert_trigger_name, update_old_trigger_name, update_new_trigger_name =
|
|
bidirectional_trigger_names(table, old_column, new_column)
|
|
|
|
quoted_table = quote_table_name(table)
|
|
quoted_old = quote_column_name(old_column)
|
|
quoted_new = quote_column_name(new_column)
|
|
|
|
create_insert_trigger(insert_trigger_name, quoted_table, quoted_old, quoted_new)
|
|
create_update_trigger(update_old_trigger_name, quoted_table, quoted_new, quoted_old)
|
|
create_update_trigger(update_new_trigger_name, quoted_table, quoted_old, quoted_new)
|
|
end
|
|
|
|
def remove_bidirectional_triggers(table, old_column, new_column)
|
|
insert_trigger_name, update_old_trigger_name, update_new_trigger_name =
|
|
bidirectional_trigger_names(table, old_column, new_column)
|
|
|
|
quoted_table = quote_table_name(table)
|
|
|
|
drop_trigger(insert_trigger_name, quoted_table)
|
|
drop_trigger(update_old_trigger_name, quoted_table)
|
|
drop_trigger(update_new_trigger_name, quoted_table)
|
|
end
|
|
|
|
def bidirectional_trigger_names(table, old_column, new_column)
|
|
%w[insert update_old update_new].map do |operation|
|
|
'trigger_' + Digest::SHA256.hexdigest("#{table}_#{old_column}_#{new_column}_#{operation}").first(12)
|
|
end
|
|
end
|
|
|
|
def function_name_for_trigger(trigger_name)
|
|
"function_for_#{trigger_name}"
|
|
end
|
|
|
|
def create_insert_trigger(trigger_name, quoted_table, quoted_old_column, quoted_new_column)
|
|
function_name = function_name_for_trigger(trigger_name)
|
|
|
|
execute(<<~SQL)
|
|
CREATE OR REPLACE FUNCTION #{function_name}()
|
|
RETURNS trigger
|
|
LANGUAGE plpgsql
|
|
AS $$
|
|
BEGIN
|
|
IF NEW.#{quoted_old_column} IS NULL AND NEW.#{quoted_new_column} IS NOT NULL THEN
|
|
NEW.#{quoted_old_column} = NEW.#{quoted_new_column};
|
|
END IF;
|
|
|
|
IF NEW.#{quoted_new_column} IS NULL AND NEW.#{quoted_old_column} IS NOT NULL THEN
|
|
NEW.#{quoted_new_column} = NEW.#{quoted_old_column};
|
|
END IF;
|
|
|
|
RETURN NEW;
|
|
END
|
|
$$;
|
|
|
|
DROP TRIGGER IF EXISTS #{trigger_name}
|
|
ON #{quoted_table};
|
|
|
|
CREATE TRIGGER #{trigger_name}
|
|
BEFORE INSERT ON #{quoted_table}
|
|
FOR EACH ROW EXECUTE FUNCTION #{function_name}();
|
|
SQL
|
|
end
|
|
|
|
def create_update_trigger(trigger_name, quoted_table, quoted_source_column, quoted_target_column)
|
|
function_name = function_name_for_trigger(trigger_name)
|
|
|
|
execute(<<~SQL)
|
|
CREATE OR REPLACE FUNCTION #{function_name}()
|
|
RETURNS trigger
|
|
LANGUAGE plpgsql
|
|
AS $$
|
|
BEGIN
|
|
NEW.#{quoted_target_column} := NEW.#{quoted_source_column};
|
|
RETURN NEW;
|
|
END
|
|
$$;
|
|
|
|
DROP TRIGGER IF EXISTS #{trigger_name}
|
|
ON #{quoted_table};
|
|
|
|
CREATE TRIGGER #{trigger_name}
|
|
BEFORE UPDATE OF #{quoted_source_column} ON #{quoted_table}
|
|
FOR EACH ROW EXECUTE FUNCTION #{function_name}();
|
|
SQL
|
|
end
|
|
|
|
def drop_trigger(trigger_name, quoted_table)
|
|
function_name = function_name_for_trigger(trigger_name)
|
|
|
|
execute(<<~SQL)
|
|
DROP TRIGGER IF EXISTS #{trigger_name}
|
|
ON #{quoted_table};
|
|
|
|
DROP FUNCTION IF EXISTS #{function_name};
|
|
SQL
|
|
end
|
|
end
|
|
end
|
|
end
|
|
end
|