2016-02-20 22:53:20 +01:00
# This file is auto-generated from the current state of the database. Instead
# of editing this file, please use the migrations feature of Active Record to
# incrementally modify your database, and then regenerate this schema definition.
#
2021-04-12 12:37:14 +02:00
# This file is the source Rails uses to define your schema when running `bin/rails
# db:schema:load`. When creating a new database, `bin/rails db:schema:load` tends to
# be faster and is potentially less error prone than running all of your
# migrations from scratch. Old migrations may fail to apply correctly if those
# migrations use external dependencies or application code.
2016-02-20 22:53:20 +01:00
#
# It's strongly recommended that you check this file into your version control system.
2023-08-18 18:20:55 +02:00
ActiveRecord :: Schema [ 7 . 0 ] . define ( version : 2023_08_18_142253 ) do
2016-02-20 22:53:20 +01:00
# These are extensions that must be enabled in order to support this database
enable_extension " plpgsql "
2019-09-19 20:58:19 +02:00
create_table " account_aliases " , force : :cascade do | t |
t . bigint " account_id "
t . string " acct " , default : " " , null : false
t . string " uri " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2019-09-19 20:58:19 +02:00
t . index [ " account_id " ] , name : " index_account_aliases_on_account_id "
end
2018-10-07 23:44:58 +02:00
create_table " account_conversations " , force : :cascade do | t |
t . bigint " account_id "
t . bigint " conversation_id "
t . bigint " participant_account_ids " , default : [ ] , null : false , array : true
t . bigint " status_ids " , default : [ ] , null : false , array : true
t . bigint " last_status_id "
t . integer " lock_version " , default : 0 , null : false
2018-10-19 01:47:29 +02:00
t . boolean " unread " , default : false , null : false
2018-10-07 23:44:58 +02:00
t . index [ " account_id " , " conversation_id " , " participant_account_ids " ] , name : " index_unique_conversations " , unique : true
t . index [ " conversation_id " ] , name : " index_account_conversations_on_conversation_id "
end
2020-09-15 14:37:58 +02:00
create_table " account_deletion_requests " , force : :cascade do | t |
t . bigint " account_id "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2020-09-15 14:37:58 +02:00
t . index [ " account_id " ] , name : " index_account_deletion_requests_on_account_id "
end
2017-09-22 13:20:04 +02:00
create_table " account_domain_blocks " , force : :cascade do | t |
2017-06-02 16:18:54 +02:00
t . string " domain "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
t . bigint " account_id "
2017-06-02 16:18:54 +02:00
t . index [ " account_id " , " domain " ] , name : " index_account_domain_blocks_on_account_id_and_domain " , unique : true
Account domain blocks (#2381)
* Add <ostatus:conversation /> tag to Atom input/output
Only uses ref attribute (not href) because href would be
the alternate link that's always included also.
Creates new conversation for every non-reply status. Carries
over conversation for every reply. Keeps remote URIs verbatim,
generates local URIs on the fly like the rest of them.
* Conversation muting - prevents notifications that reference a conversation
(including replies, favourites, reblogs) from being created. API endpoints
/api/v1/statuses/:id/mute and /api/v1/statuses/:id/unmute
Currently no way to tell when a status/conversation is muted, so the web UI
only has a "disable notifications" button, doesn't work as a toggle
* Display "Dismiss notifications" on all statuses in notifications column, not just own
* Add "muted" as a boolean attribute on statuses JSON
For now always false on contained reblogs, since it's only relevant for
statuses returned from the notifications endpoint, which are not nested
Remove "Disable notifications" from detailed status view, since it's
only relevant in the notifications column
* Up max class length
* Remove pending test for conversation mute
* Add tests, clean up
* Rename to "mute conversation" and "unmute conversation"
* Raise validation error when trying to mute/unmute status without conversation
* Adding account domain blocks that filter notifications and public timelines
* Add tests for domain blocks in notifications, public timelines
Filter reblogs of blocked domains from home
* Add API for listing and creating account domain blocks
* API for creating/deleting domain blocks, tests for Status#ancestors
and Status#descendants, filter domain blocks from them
* Filter domains in streaming API
* Update account_domain_block_spec.rb
2017-05-19 01:14:30 +02:00
end
2019-09-19 20:58:19 +02:00
create_table " account_migrations " , force : :cascade do | t |
t . bigint " account_id "
t . string " acct " , default : " " , null : false
t . bigint " followers_count " , default : 0 , null : false
t . bigint " target_account_id "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2019-09-19 20:58:19 +02:00
t . index [ " account_id " ] , name : " index_account_migrations_on_account_id "
2022-03-12 08:12:57 +01:00
t . index [ " target_account_id " ] , name : " index_account_migrations_on_target_account_id " , where : " (target_account_id IS NOT NULL) "
2019-09-19 20:58:19 +02:00
end
2017-10-07 20:26:43 +02:00
create_table " account_moderation_notes " , force : :cascade do | t |
t . text " content " , null : false
2017-10-10 13:12:17 +02:00
t . bigint " account_id " , null : false
t . bigint " target_account_id " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2017-10-07 20:26:43 +02:00
t . index [ " account_id " ] , name : " index_account_moderation_notes_on_account_id "
t . index [ " target_account_id " ] , name : " index_account_moderation_notes_on_target_account_id "
end
2020-07-07 15:26:51 +02:00
create_table " account_notes " , force : :cascade do | t |
t . bigint " account_id "
t . bigint " target_account_id "
t . text " comment " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2020-07-07 15:26:51 +02:00
t . index [ " account_id " , " target_account_id " ] , name : " index_account_notes_on_account_id_and_target_account_id " , unique : true
t . index [ " target_account_id " ] , name : " index_account_notes_on_target_account_id "
end
2018-08-09 09:56:53 +02:00
create_table " account_pins " , force : :cascade do | t |
t . bigint " account_id "
t . bigint " target_account_id "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2018-08-09 09:56:53 +02:00
t . index [ " account_id " , " target_account_id " ] , name : " index_account_pins_on_account_id_and_target_account_id " , unique : true
t . index [ " target_account_id " ] , name : " index_account_pins_on_target_account_id "
end
2018-11-19 00:43:52 +01:00
create_table " account_stats " , force : :cascade do | t |
t . bigint " account_id " , null : false
t . bigint " statuses_count " , default : 0 , null : false
t . bigint " following_count " , default : 0 , null : false
t . bigint " followers_count " , default : 0 , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
t . datetime " last_status_at " , precision : nil
2018-11-19 00:43:52 +01:00
t . index [ " account_id " ] , name : " index_account_stats_on_account_id " , unique : true
end
2021-08-09 23:11:50 +02:00
create_table " account_statuses_cleanup_policies " , force : :cascade do | t |
t . bigint " account_id " , null : false
t . boolean " enabled " , default : true , null : false
t . integer " min_status_age " , default : 1209600 , null : false
t . boolean " keep_direct " , default : true , null : false
t . boolean " keep_pinned " , default : true , null : false
t . boolean " keep_polls " , default : false , null : false
t . boolean " keep_media " , default : false , null : false
t . boolean " keep_self_fav " , default : true , null : false
t . boolean " keep_self_bookmark " , default : true , null : false
t . integer " min_favs "
t . integer " min_reblogs "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , null : false
t . datetime " updated_at " , null : false
2021-08-09 23:11:50 +02:00
t . index [ " account_id " ] , name : " index_account_statuses_cleanup_policies_on_account_id "
end
Add moderation warnings (#9519)
* Add moderation warnings
Replace individual routes for disabling, silencing, and suspending
a user, as well as the report update route, with a unified account
action controller that allows you to select an action (none,
disable, silence, suspend) as well as whether it should generate an
e-mail notification with optional custom text. That notification,
with the optional custom text, is saved as a warning.
Additionally, there are warning presets you can configure to save
time when performing the above.
* Use Account#local_username_and_domain
2018-12-22 20:02:09 +01:00
create_table " account_warning_presets " , force : :cascade do | t |
t . text " text " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2020-03-12 17:57:59 +01:00
t . string " title " , default : " " , null : false
Add moderation warnings (#9519)
* Add moderation warnings
Replace individual routes for disabling, silencing, and suspending
a user, as well as the report update route, with a unified account
action controller that allows you to select an action (none,
disable, silence, suspend) as well as whether it should generate an
e-mail notification with optional custom text. That notification,
with the optional custom text, is saved as a warning.
Additionally, there are warning presets you can configure to save
time when performing the above.
* Use Account#local_username_and_domain
2018-12-22 20:02:09 +01:00
end
create_table " account_warnings " , force : :cascade do | t |
t . bigint " account_id "
t . bigint " target_account_id "
t . integer " action " , default : 0 , null : false
t . text " text " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2022-01-17 09:41:33 +01:00
t . bigint " report_id "
t . string " status_ids " , array : true
2023-07-25 13:40:35 +02:00
t . datetime " overruled_at " , precision : nil
Add moderation warnings (#9519)
* Add moderation warnings
Replace individual routes for disabling, silencing, and suspending
a user, as well as the report update route, with a unified account
action controller that allows you to select an action (none,
disable, silence, suspend) as well as whether it should generate an
e-mail notification with optional custom text. That notification,
with the optional custom text, is saved as a warning.
Additionally, there are warning presets you can configure to save
time when performing the above.
* Use Account#local_username_and_domain
2018-12-22 20:02:09 +01:00
t . index [ " account_id " ] , name : " index_account_warnings_on_account_id "
t . index [ " target_account_id " ] , name : " index_account_warnings_on_target_account_id "
end
2021-04-16 22:20:32 +02:00
create_table " accounts " , id : :bigint , default : - > { " timestamp_id('accounts'::text) " } , force : :cascade do | t |
2017-06-02 16:18:54 +02:00
t . string " username " , default : " " , null : false
t . string " domain "
t . text " private_key "
t . text " public_key " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2017-06-02 16:18:54 +02:00
t . text " note " , default : " " , null : false
t . string " display_name " , default : " " , null : false
t . string " uri " , default : " " , null : false
t . string " url "
t . string " avatar_file_name "
t . string " avatar_content_type "
t . integer " avatar_file_size "
2023-07-25 13:40:35 +02:00
t . datetime " avatar_updated_at " , precision : nil
2017-06-02 16:18:54 +02:00
t . string " header_file_name "
t . string " header_content_type "
t . integer " header_file_size "
2023-07-25 13:40:35 +02:00
t . datetime " header_updated_at " , precision : nil
2017-06-02 16:18:54 +02:00
t . string " avatar_remote_url "
t . boolean " locked " , default : false , null : false
t . string " header_remote_url " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " last_webfingered_at " , precision : nil
2017-07-19 17:06:46 +02:00
t . string " inbox_url " , default : " " , null : false
t . string " outbox_url " , default : " " , null : false
t . string " shared_inbox_url " , default : " " , null : false
t . string " followers_url " , default : " " , null : false
t . integer " protocol " , default : 0 , null : false
2017-11-07 19:06:44 +01:00
t . boolean " memorial " , default : false , null : false
2017-11-18 19:39:02 +01:00
t . bigint " moved_to_account_id "
2018-03-04 09:19:11 +01:00
t . string " featured_collection_url "
2018-04-14 12:41:08 +02:00
t . jsonb " fields "
2018-05-07 09:31:07 +02:00
t . string " actor_type "
2018-12-06 17:36:11 +01:00
t . boolean " discoverable "
2019-01-10 18:46:17 +01:00
t . string " also_known_as " , array : true
2023-07-25 13:40:35 +02:00
t . datetime " silenced_at " , precision : nil
t . datetime " suspended_at " , precision : nil
2020-03-09 00:10:29 +01:00
t . boolean " hide_collections "
2020-04-26 23:29:08 +02:00
t . integer " avatar_storage_schema_version "
t . integer " header_storage_schema_version "
2020-06-02 19:24:53 +02:00
t . string " devices_url "
2020-11-08 00:28:39 +01:00
t . integer " suspension_origin "
2023-07-25 13:40:35 +02:00
t . datetime " sensitized_at " , precision : nil
2022-02-25 00:34:14 +01:00
t . boolean " trendable "
2023-07-25 13:40:35 +02:00
t . datetime " reviewed_at " , precision : nil
t . datetime " requested_review_at " , precision : nil
2023-08-14 18:54:51 +02:00
t . boolean " indexable " , default : false , null : false
2017-03-17 20:47:38 +01:00
t . index " (((setweight(to_tsvector('simple'::regconfig, (display_name)::text), 'A':: \" char \" ) || setweight(to_tsvector('simple'::regconfig, (username)::text), 'B':: \" char \" )) || setweight(to_tsvector('simple'::regconfig, (COALESCE(domain, ''::character varying))::text), 'C':: \" char \" ))) " , name : " search_index " , using : :gin
2020-06-21 12:41:38 +02:00
t . index " lower((username)::text), COALESCE(lower((domain)::text), ''::text) " , name : " index_accounts_on_username_and_domain_lower " , unique : true
2023-05-25 09:27:16 +02:00
t . index [ " domain " , " id " ] , name : " index_accounts_on_domain_and_id "
2022-03-12 08:12:57 +01:00
t . index [ " moved_to_account_id " ] , name : " index_accounts_on_moved_to_account_id " , where : " (moved_to_account_id IS NOT NULL) "
2017-06-02 16:18:54 +02:00
t . index [ " uri " ] , name : " index_accounts_on_uri "
2022-03-12 08:12:57 +01:00
t . index [ " url " ] , name : " index_accounts_on_url " , opclass : :text_pattern_ops , where : " (url IS NOT NULL) "
2016-02-20 22:53:20 +01:00
end
2023-06-01 09:49:06 +02:00
create_table " accounts_tags " , primary_key : [ " tag_id " , " account_id " ] , force : :cascade do | t |
2018-12-06 17:36:11 +01:00
t . bigint " account_id " , null : false
t . bigint " tag_id " , null : false
t . index [ " account_id " , " tag_id " ] , name : " index_accounts_tags_on_account_id_and_tag_id "
end
2017-11-24 02:05:53 +01:00
create_table " admin_action_logs " , force : :cascade do | t |
t . bigint " account_id "
t . string " action " , default : " " , null : false
t . string " target_type "
t . bigint " target_id "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2022-08-25 20:39:40 +02:00
t . string " human_identifier "
t . string " route_param "
t . string " permalink "
2017-11-24 02:05:53 +01:00
t . index [ " account_id " ] , name : " index_admin_action_logs_on_account_id "
t . index [ " target_type " , " target_id " ] , name : " index_admin_action_logs_on_target_type_and_target_id "
end
2020-01-23 22:00:13 +01:00
create_table " announcement_mutes " , force : :cascade do | t |
t . bigint " account_id "
t . bigint " announcement_id "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2020-01-23 22:00:13 +01:00
t . index [ " account_id " , " announcement_id " ] , name : " index_announcement_mutes_on_account_id_and_announcement_id " , unique : true
t . index [ " announcement_id " ] , name : " index_announcement_mutes_on_announcement_id "
end
create_table " announcement_reactions " , force : :cascade do | t |
t . bigint " account_id "
t . bigint " announcement_id "
t . string " name " , default : " " , null : false
t . bigint " custom_emoji_id "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2020-01-23 22:00:13 +01:00
t . index [ " account_id " , " announcement_id " , " name " ] , name : " index_announcement_reactions_on_account_id_and_announcement_id " , unique : true
t . index [ " announcement_id " ] , name : " index_announcement_reactions_on_announcement_id "
2022-03-12 08:12:57 +01:00
t . index [ " custom_emoji_id " ] , name : " index_announcement_reactions_on_custom_emoji_id " , where : " (custom_emoji_id IS NOT NULL) "
2020-01-23 22:00:13 +01:00
end
create_table " announcements " , force : :cascade do | t |
t . text " text " , default : " " , null : false
t . boolean " published " , default : false , null : false
t . boolean " all_day " , default : false , null : false
2023-07-25 13:40:35 +02:00
t . datetime " scheduled_at " , precision : nil
t . datetime " starts_at " , precision : nil
t . datetime " ends_at " , precision : nil
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
t . datetime " published_at " , precision : nil
2020-04-05 12:51:22 +02:00
t . bigint " status_ids " , array : true
2020-01-23 22:00:13 +01:00
end
2022-02-14 21:27:53 +01:00
create_table " appeals " , force : :cascade do | t |
t . bigint " account_id " , null : false
t . bigint " account_warning_id " , null : false
t . text " text " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " approved_at " , precision : nil
2022-02-14 21:27:53 +01:00
t . bigint " approved_by_account_id "
2023-07-25 13:40:35 +02:00
t . datetime " rejected_at " , precision : nil
2022-02-14 21:27:53 +01:00
t . bigint " rejected_by_account_id "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , null : false
t . datetime " updated_at " , null : false
2022-02-14 21:27:53 +01:00
t . index [ " account_id " ] , name : " index_appeals_on_account_id "
t . index [ " account_warning_id " ] , name : " index_appeals_on_account_warning_id " , unique : true
2022-03-12 08:12:57 +01:00
t . index [ " approved_by_account_id " ] , name : " index_appeals_on_approved_by_account_id " , where : " (approved_by_account_id IS NOT NULL) "
t . index [ " rejected_by_account_id " ] , name : " index_appeals_on_rejected_by_account_id " , where : " (rejected_by_account_id IS NOT NULL) "
2022-02-14 21:27:53 +01:00
end
2018-02-21 23:21:32 +01:00
create_table " backups " , force : :cascade do | t |
t . bigint " user_id "
t . string " dump_file_name "
t . string " dump_content_type "
2023-07-25 13:40:35 +02:00
t . datetime " dump_updated_at " , precision : nil
2018-02-21 23:21:32 +01:00
t . boolean " processed " , default : false , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2020-01-23 22:00:13 +01:00
t . bigint " dump_file_size "
2023-06-30 19:09:03 +02:00
t . index [ " user_id " ] , name : " index_backups_on_user_id "
2018-02-21 23:21:32 +01:00
end
2017-09-22 13:20:04 +02:00
create_table " blocks " , force : :cascade do | t |
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
t . bigint " account_id " , null : false
t . bigint " target_account_id " , null : false
2018-05-04 21:14:34 +02:00
t . string " uri "
2017-06-02 16:18:54 +02:00
t . index [ " account_id " , " target_account_id " ] , name : " index_blocks_on_account_id_and_target_account_id " , unique : true
2018-08-21 20:11:34 +02:00
t . index [ " target_account_id " ] , name : " index_blocks_on_target_account_id "
2016-10-03 17:11:54 +02:00
end
2018-04-10 22:13:15 +02:00
create_table " bookmarks " , force : :cascade do | t |
t . bigint " account_id " , null : false
t . bigint " status_id " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2018-04-10 22:13:15 +02:00
t . index [ " account_id " , " status_id " ] , name : " index_bookmarks_on_account_id_and_status_id " , unique : true
t . index [ " status_id " ] , name : " index_bookmarks_on_status_id "
end
2023-05-02 12:08:48 +02:00
create_table " bulk_import_rows " , force : :cascade do | t |
t . bigint " bulk_import_id " , null : false
t . jsonb " data "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , null : false
t . datetime " updated_at " , null : false
2023-05-02 12:08:48 +02:00
t . index [ " bulk_import_id " ] , name : " index_bulk_import_rows_on_bulk_import_id "
end
create_table " bulk_imports " , force : :cascade do | t |
t . integer " type " , null : false
t . integer " state " , null : false
t . integer " total_items " , default : 0 , null : false
t . integer " imported_items " , default : 0 , null : false
t . integer " processed_items " , default : 0 , null : false
2023-07-25 13:40:35 +02:00
t . datetime " finished_at " , precision : nil
2023-05-02 12:08:48 +02:00
t . boolean " overwrite " , default : false , null : false
t . boolean " likely_mismatched " , default : false , null : false
t . string " original_filename " , default : " " , null : false
t . bigint " account_id " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , null : false
t . datetime " updated_at " , null : false
2023-05-02 12:08:48 +02:00
t . index [ " account_id " ] , name : " index_bulk_imports_on_account_id "
t . index [ " id " ] , name : " index_bulk_imports_unconfirmed " , where : " (state = 0) "
end
2021-04-17 03:14:25 +02:00
create_table " canonical_email_blocks " , force : :cascade do | t |
t . string " canonical_email_hash " , default : " " , null : false
2022-08-28 03:31:54 +02:00
t . bigint " reference_account_id "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , null : false
t . datetime " updated_at " , null : false
2021-04-17 03:14:25 +02:00
t . index [ " canonical_email_hash " ] , name : " index_canonical_email_blocks_on_canonical_email_hash " , unique : true
t . index [ " reference_account_id " ] , name : " index_canonical_email_blocks_on_reference_account_id "
end
2017-09-22 13:20:04 +02:00
create_table " conversation_mutes " , force : :cascade do | t |
2017-06-02 16:18:54 +02:00
t . bigint " conversation_id " , null : false
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
t . bigint " account_id " , null : false
2017-06-02 16:18:54 +02:00
t . index [ " account_id " , " conversation_id " ] , name : " index_conversation_mutes_on_account_id_and_conversation_id " , unique : true
Feature conversations muting (#3017)
* Add <ostatus:conversation /> tag to Atom input/output
Only uses ref attribute (not href) because href would be
the alternate link that's always included also.
Creates new conversation for every non-reply status. Carries
over conversation for every reply. Keeps remote URIs verbatim,
generates local URIs on the fly like the rest of them.
* Conversation muting - prevents notifications that reference a conversation
(including replies, favourites, reblogs) from being created. API endpoints
/api/v1/statuses/:id/mute and /api/v1/statuses/:id/unmute
Currently no way to tell when a status/conversation is muted, so the web UI
only has a "disable notifications" button, doesn't work as a toggle
* Display "Dismiss notifications" on all statuses in notifications column, not just own
* Add "muted" as a boolean attribute on statuses JSON
For now always false on contained reblogs, since it's only relevant for
statuses returned from the notifications endpoint, which are not nested
Remove "Disable notifications" from detailed status view, since it's
only relevant in the notifications column
* Up max class length
* Remove pending test for conversation mute
* Add tests, clean up
* Rename to "mute conversation" and "unmute conversation"
* Raise validation error when trying to mute/unmute status without conversation
2017-05-15 03:04:13 +02:00
end
2017-06-02 16:18:54 +02:00
create_table " conversations " , force : :cascade do | t |
t . string " uri "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2022-03-12 08:12:57 +01:00
t . index [ " uri " ] , name : " index_conversations_on_uri " , unique : true , opclass : :text_pattern_ops , where : " (uri IS NOT NULL) "
2017-05-12 19:09:21 +02:00
end
2019-06-28 15:54:10 +02:00
create_table " custom_emoji_categories " , force : :cascade do | t |
t . string " name "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2019-06-28 15:54:10 +02:00
t . index [ " name " ] , name : " index_custom_emoji_categories_on_name " , unique : true
end
2017-09-19 02:42:40 +02:00
create_table " custom_emojis " , force : :cascade do | t |
t . string " shortcode " , default : " " , null : false
t . string " domain "
t . string " image_file_name "
t . string " image_content_type "
t . integer " image_file_size "
2023-07-25 13:40:35 +02:00
t . datetime " image_updated_at " , precision : nil
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2017-10-05 23:42:05 +02:00
t . boolean " disabled " , default : false , null : false
2017-10-07 17:43:42 +02:00
t . string " uri "
t . string " image_remote_url "
2017-10-27 16:11:30 +02:00
t . boolean " visible_in_picker " , default : true , null : false
2019-06-28 15:54:10 +02:00
t . bigint " category_id "
2020-04-26 23:29:08 +02:00
t . integer " image_storage_schema_version "
2017-09-19 02:42:40 +02:00
t . index [ " shortcode " , " domain " ] , name : " index_custom_emojis_on_shortcode_and_domain " , unique : true
end
Revamp post filtering system (#18058)
* Add model for custom filter keywords
* Use CustomFilterKeyword internally
Does not change the API
* Fix /filters/edit and /filters/new
* Add migration tests
* Remove whole_word column from custom_filters (covered by custom_filter_keywords)
* Redesign /filters
Instead of a list, present a card that displays more information and handles
multiple keywords per filter.
* Redesign /filters/new and /filters/edit to add and remove keywords
This adds a new gem dependency: cocoon, as well as a npm dependency:
cocoon-js-vanilla. Those are used to easily populate and remove form fields
from the user interface when manipulating multiple keyword filters at once.
* Add /api/v2/filters to edit filter with multiple keywords
Entities:
- `Filter`: `id`, `title`, `filter_action` (either `hide` or `warn`), `context`
`keywords`
- `FilterKeyword`: `id`, `keyword`, `whole_word`
API endpoits:
- `GET /api/v2/filters` to list filters (including keywords)
- `POST /api/v2/filters` to create a new filter
`keywords_attributes` can also be passed to create keywords in one request
- `GET /api/v2/filters/:id` to read a particular filter
- `PUT /api/v2/filters/:id` to update a new filter
`keywords_attributes` can also be passed to edit, delete or add keywords in
one request
- `DELETE /api/v2/filters/:id` to delete a particular filter
- `GET /api/v2/filters/:id/keywords` to list keywords for a filter
- `POST /api/v2/filters/:filter_id/keywords/:id` to add a new keyword to a
filter
- `GET /api/v2/filter_keywords/:id` to read a particular keyword
- `PUT /api/v2/filter_keywords/:id` to edit a particular keyword
- `DELETE /api/v2/filter_keywords/:id` to delete a particular keyword
* Change from `irreversible` boolean to `action` enum
* Remove irrelevent `irreversible_must_be_within_context` check
* Fix /filters/new and /filters/edit with update for filter_action
* Fix Rubocop/Codeclimate complaining about task names
* Refactor FeedManager#phrase_filtered?
This moves regexp building and filter caching to the `CustomFilter` class.
This does not change the functional behavior yet, but this changes how the
cache is built, doing per-custom_filter regexps so that filters can be matched
independently, while still offering caching.
* Perform server-side filtering and output result in REST API
* Fix numerous filters_changed events being sent when editing multiple keywords at once
* Add some tests
* Use the new API in the WebUI
- use client-side logic for filters we have fetched rules for.
This is so that filter changes can be retroactively applied without
reloading the UI.
- use server-side logic for filters we haven't fetched rules for yet
(e.g. network error, or initial timeline loading)
* Minor optimizations and refactoring
* Perform server-side filtering on the streaming server
* Change the wording of filter action labels
* Fix issues pointed out by linter
* Change design of “Show anyway” link in accordence to review comments
* Drop “irreversible” filtering behavior
* Move /api/v2/filter_keywords to /api/v1/filters/keywords
* Rename `filter_results` attribute to `filtered`
* Rename REST::LegacyFilterSerializer to REST::V1::FilterSerializer
* Fix systemChannelId value in streaming server
* Simplify code by removing client-side filtering code
The simplifcation comes at a cost though: filters aren't retroactively
applied anymore.
2022-06-28 09:42:13 +02:00
create_table " custom_filter_keywords " , force : :cascade do | t |
t . bigint " custom_filter_id " , null : false
t . text " keyword " , default : " " , null : false
t . boolean " whole_word " , default : true , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , null : false
t . datetime " updated_at " , null : false
Revamp post filtering system (#18058)
* Add model for custom filter keywords
* Use CustomFilterKeyword internally
Does not change the API
* Fix /filters/edit and /filters/new
* Add migration tests
* Remove whole_word column from custom_filters (covered by custom_filter_keywords)
* Redesign /filters
Instead of a list, present a card that displays more information and handles
multiple keywords per filter.
* Redesign /filters/new and /filters/edit to add and remove keywords
This adds a new gem dependency: cocoon, as well as a npm dependency:
cocoon-js-vanilla. Those are used to easily populate and remove form fields
from the user interface when manipulating multiple keyword filters at once.
* Add /api/v2/filters to edit filter with multiple keywords
Entities:
- `Filter`: `id`, `title`, `filter_action` (either `hide` or `warn`), `context`
`keywords`
- `FilterKeyword`: `id`, `keyword`, `whole_word`
API endpoits:
- `GET /api/v2/filters` to list filters (including keywords)
- `POST /api/v2/filters` to create a new filter
`keywords_attributes` can also be passed to create keywords in one request
- `GET /api/v2/filters/:id` to read a particular filter
- `PUT /api/v2/filters/:id` to update a new filter
`keywords_attributes` can also be passed to edit, delete or add keywords in
one request
- `DELETE /api/v2/filters/:id` to delete a particular filter
- `GET /api/v2/filters/:id/keywords` to list keywords for a filter
- `POST /api/v2/filters/:filter_id/keywords/:id` to add a new keyword to a
filter
- `GET /api/v2/filter_keywords/:id` to read a particular keyword
- `PUT /api/v2/filter_keywords/:id` to edit a particular keyword
- `DELETE /api/v2/filter_keywords/:id` to delete a particular keyword
* Change from `irreversible` boolean to `action` enum
* Remove irrelevent `irreversible_must_be_within_context` check
* Fix /filters/new and /filters/edit with update for filter_action
* Fix Rubocop/Codeclimate complaining about task names
* Refactor FeedManager#phrase_filtered?
This moves regexp building and filter caching to the `CustomFilter` class.
This does not change the functional behavior yet, but this changes how the
cache is built, doing per-custom_filter regexps so that filters can be matched
independently, while still offering caching.
* Perform server-side filtering and output result in REST API
* Fix numerous filters_changed events being sent when editing multiple keywords at once
* Add some tests
* Use the new API in the WebUI
- use client-side logic for filters we have fetched rules for.
This is so that filter changes can be retroactively applied without
reloading the UI.
- use server-side logic for filters we haven't fetched rules for yet
(e.g. network error, or initial timeline loading)
* Minor optimizations and refactoring
* Perform server-side filtering on the streaming server
* Change the wording of filter action labels
* Fix issues pointed out by linter
* Change design of “Show anyway” link in accordence to review comments
* Drop “irreversible” filtering behavior
* Move /api/v2/filter_keywords to /api/v1/filters/keywords
* Rename `filter_results` attribute to `filtered`
* Rename REST::LegacyFilterSerializer to REST::V1::FilterSerializer
* Fix systemChannelId value in streaming server
* Simplify code by removing client-side filtering code
The simplifcation comes at a cost though: filters aren't retroactively
applied anymore.
2022-06-28 09:42:13 +02:00
t . index [ " custom_filter_id " ] , name : " index_custom_filter_keywords_on_custom_filter_id "
end
2022-08-25 04:27:47 +02:00
create_table " custom_filter_statuses " , force : :cascade do | t |
t . bigint " custom_filter_id " , null : false
t . bigint " status_id " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , null : false
t . datetime " updated_at " , null : false
2022-08-25 04:27:47 +02:00
t . index [ " custom_filter_id " ] , name : " index_custom_filter_statuses_on_custom_filter_id "
t . index [ " status_id " ] , name : " index_custom_filter_statuses_on_status_id "
end
2018-06-29 15:34:36 +02:00
create_table " custom_filters " , force : :cascade do | t |
t . bigint " account_id "
2023-07-25 13:40:35 +02:00
t . datetime " expires_at " , precision : nil
2018-06-29 15:34:36 +02:00
t . text " phrase " , default : " " , null : false
t . string " context " , default : [ ] , null : false , array : true
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
Revamp post filtering system (#18058)
* Add model for custom filter keywords
* Use CustomFilterKeyword internally
Does not change the API
* Fix /filters/edit and /filters/new
* Add migration tests
* Remove whole_word column from custom_filters (covered by custom_filter_keywords)
* Redesign /filters
Instead of a list, present a card that displays more information and handles
multiple keywords per filter.
* Redesign /filters/new and /filters/edit to add and remove keywords
This adds a new gem dependency: cocoon, as well as a npm dependency:
cocoon-js-vanilla. Those are used to easily populate and remove form fields
from the user interface when manipulating multiple keyword filters at once.
* Add /api/v2/filters to edit filter with multiple keywords
Entities:
- `Filter`: `id`, `title`, `filter_action` (either `hide` or `warn`), `context`
`keywords`
- `FilterKeyword`: `id`, `keyword`, `whole_word`
API endpoits:
- `GET /api/v2/filters` to list filters (including keywords)
- `POST /api/v2/filters` to create a new filter
`keywords_attributes` can also be passed to create keywords in one request
- `GET /api/v2/filters/:id` to read a particular filter
- `PUT /api/v2/filters/:id` to update a new filter
`keywords_attributes` can also be passed to edit, delete or add keywords in
one request
- `DELETE /api/v2/filters/:id` to delete a particular filter
- `GET /api/v2/filters/:id/keywords` to list keywords for a filter
- `POST /api/v2/filters/:filter_id/keywords/:id` to add a new keyword to a
filter
- `GET /api/v2/filter_keywords/:id` to read a particular keyword
- `PUT /api/v2/filter_keywords/:id` to edit a particular keyword
- `DELETE /api/v2/filter_keywords/:id` to delete a particular keyword
* Change from `irreversible` boolean to `action` enum
* Remove irrelevent `irreversible_must_be_within_context` check
* Fix /filters/new and /filters/edit with update for filter_action
* Fix Rubocop/Codeclimate complaining about task names
* Refactor FeedManager#phrase_filtered?
This moves regexp building and filter caching to the `CustomFilter` class.
This does not change the functional behavior yet, but this changes how the
cache is built, doing per-custom_filter regexps so that filters can be matched
independently, while still offering caching.
* Perform server-side filtering and output result in REST API
* Fix numerous filters_changed events being sent when editing multiple keywords at once
* Add some tests
* Use the new API in the WebUI
- use client-side logic for filters we have fetched rules for.
This is so that filter changes can be retroactively applied without
reloading the UI.
- use server-side logic for filters we haven't fetched rules for yet
(e.g. network error, or initial timeline loading)
* Minor optimizations and refactoring
* Perform server-side filtering on the streaming server
* Change the wording of filter action labels
* Fix issues pointed out by linter
* Change design of “Show anyway” link in accordence to review comments
* Drop “irreversible” filtering behavior
* Move /api/v2/filter_keywords to /api/v1/filters/keywords
* Rename `filter_results` attribute to `filtered`
* Rename REST::LegacyFilterSerializer to REST::V1::FilterSerializer
* Fix systemChannelId value in streaming server
* Simplify code by removing client-side filtering code
The simplifcation comes at a cost though: filters aren't retroactively
applied anymore.
2022-06-28 09:42:13 +02:00
t . integer " action " , default : 0 , null : false
2018-06-29 15:34:36 +02:00
t . index [ " account_id " ] , name : " index_custom_filters_on_account_id "
end
2020-06-02 19:24:53 +02:00
create_table " devices " , force : :cascade do | t |
t . bigint " access_token_id "
t . bigint " account_id "
t . string " device_id " , default : " " , null : false
t . string " name " , default : " " , null : false
t . text " fingerprint_key " , default : " " , null : false
t . text " identity_key " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2020-06-02 19:24:53 +02:00
t . index [ " access_token_id " ] , name : " index_devices_on_access_token_id "
t . index [ " account_id " ] , name : " index_devices_on_account_id "
end
2019-07-30 11:10:46 +02:00
create_table " domain_allows " , force : :cascade do | t |
t . string " domain " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2019-07-30 11:10:46 +02:00
t . index [ " domain " ] , name : " index_domain_allows_on_domain " , unique : true
end
2017-09-22 13:20:04 +02:00
create_table " domain_blocks " , force : :cascade do | t |
2017-06-02 16:18:54 +02:00
t . string " domain " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2017-06-02 16:18:54 +02:00
t . integer " severity " , default : 0
2017-07-13 03:12:25 +02:00
t . boolean " reject_media " , default : false , null : false
2018-10-20 08:02:44 +02:00
t . boolean " reject_reports " , default : false , null : false
2019-08-07 20:20:23 +02:00
t . text " private_comment "
t . text " public_comment "
2020-12-18 08:30:41 +01:00
t . boolean " obfuscate " , default : false , null : false
2017-06-02 16:18:54 +02:00
t . index [ " domain " ] , name : " index_domain_blocks_on_domain " , unique : true
2016-10-09 14:48:43 +02:00
end
2017-10-04 15:16:10 +02:00
create_table " email_domain_blocks " , force : :cascade do | t |
2017-11-14 20:37:17 +01:00
t . string " domain " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2020-03-12 22:35:20 +01:00
t . bigint " parent_id "
2017-11-14 20:37:17 +01:00
t . index [ " domain " ] , name : " index_email_domain_blocks_on_domain " , unique : true
2017-10-04 15:16:10 +02:00
end
2020-06-02 19:24:53 +02:00
create_table " encrypted_messages " , id : :bigint , default : - > { " timestamp_id('encrypted_messages'::text) " } , force : :cascade do | t |
t . bigint " device_id "
t . bigint " from_account_id "
t . string " from_device_id " , default : " " , null : false
t . integer " type " , default : 0 , null : false
t . text " body " , default : " " , null : false
t . text " digest " , default : " " , null : false
t . text " message_franking " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2020-06-02 19:24:53 +02:00
t . index [ " device_id " ] , name : " index_encrypted_messages_on_device_id "
t . index [ " from_account_id " ] , name : " index_encrypted_messages_on_from_account_id "
end
2017-09-22 13:20:04 +02:00
create_table " favourites " , force : :cascade do | t |
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
t . bigint " account_id " , null : false
t . bigint " status_id " , null : false
2017-07-26 03:35:25 +02:00
t . index [ " account_id " , " id " ] , name : " index_favourites_on_account_id_and_id "
2017-06-02 16:18:54 +02:00
t . index [ " account_id " , " status_id " ] , name : " index_favourites_on_account_id_and_status_id " , unique : true
t . index [ " status_id " ] , name : " index_favourites_on_status_id "
2016-02-23 19:17:37 +01:00
end
2019-02-04 04:25:59 +01:00
create_table " featured_tags " , force : :cascade do | t |
2022-03-09 08:51:12 +01:00
t . bigint " account_id " , null : false
t . bigint " tag_id " , null : false
2019-02-04 04:25:59 +01:00
t . bigint " statuses_count " , default : 0 , null : false
2023-07-25 13:40:35 +02:00
t . datetime " last_status_at " , precision : nil
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2022-11-04 16:08:29 +01:00
t . string " name "
2022-10-22 14:30:59 +02:00
t . index [ " account_id " , " tag_id " ] , name : " index_featured_tags_on_account_id_and_tag_id " , unique : true
2019-02-04 04:25:59 +01:00
t . index [ " tag_id " ] , name : " index_featured_tags_on_tag_id "
end
2021-04-12 12:37:14 +02:00
create_table " follow_recommendation_suppressions " , force : :cascade do | t |
t . bigint " account_id " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , null : false
t . datetime " updated_at " , null : false
2021-04-12 12:37:14 +02:00
t . index [ " account_id " ] , name : " index_follow_recommendation_suppressions_on_account_id " , unique : true
end
2017-09-22 13:20:04 +02:00
create_table " follow_requests " , force : :cascade do | t |
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
t . bigint " account_id " , null : false
t . bigint " target_account_id " , null : false
2017-11-09 15:41:10 +01:00
t . boolean " show_reblogs " , default : true , null : false
2018-05-04 21:14:34 +02:00
t . string " uri "
2020-09-18 17:26:45 +02:00
t . boolean " notify " , default : false , null : false
2022-09-20 23:51:21 +02:00
t . string " languages " , array : true
2017-06-02 16:18:54 +02:00
t . index [ " account_id " , " target_account_id " ] , name : " index_follow_requests_on_account_id_and_target_account_id " , unique : true
2016-12-22 23:03:57 +01:00
end
2017-09-22 13:20:04 +02:00
create_table " follows " , force : :cascade do | t |
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
t . bigint " account_id " , null : false
t . bigint " target_account_id " , null : false
2017-11-09 15:41:10 +01:00
t . boolean " show_reblogs " , default : true , null : false
2018-05-04 21:14:34 +02:00
t . string " uri "
2020-09-18 17:26:45 +02:00
t . boolean " notify " , default : false , null : false
2022-09-20 23:51:21 +02:00
t . string " languages " , array : true
2017-06-02 16:18:54 +02:00
t . index [ " account_id " , " target_account_id " ] , name : " index_follows_on_account_id_and_target_account_id " , unique : true
2018-08-21 20:11:34 +02:00
t . index [ " target_account_id " ] , name : " index_follows_on_target_account_id "
2016-02-22 16:00:20 +01:00
end
2018-11-27 13:56:25 +01:00
create_table " identities " , force : :cascade do | t |
2018-02-04 05:42:13 +01:00
t . string " provider " , default : " " , null : false
t . string " uid " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2018-11-27 13:56:25 +01:00
t . bigint " user_id "
2018-02-04 05:42:13 +01:00
t . index [ " user_id " ] , name : " index_identities_on_user_id "
end
2017-09-22 13:20:04 +02:00
create_table " imports " , force : :cascade do | t |
2017-06-02 16:18:54 +02:00
t . integer " type " , null : false
2017-07-13 03:12:25 +02:00
t . boolean " approved " , default : false , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2017-06-02 16:18:54 +02:00
t . string " data_file_name "
t . string " data_content_type "
t . integer " data_file_size "
2023-07-25 13:40:35 +02:00
t . datetime " data_updated_at " , precision : nil
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
t . bigint " account_id " , null : false
2019-02-03 03:59:51 +01:00
t . boolean " overwrite " , default : false , null : false
2017-03-30 19:42:33 +02:00
end
2017-11-27 16:07:59 +01:00
create_table " invites " , force : :cascade do | t |
2018-02-07 16:35:44 +01:00
t . bigint " user_id " , null : false
2017-11-27 16:07:59 +01:00
t . string " code " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " expires_at " , precision : nil
2017-11-27 16:07:59 +01:00
t . integer " max_uses "
t . integer " uses " , default : 0 , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2018-06-15 18:00:23 +02:00
t . boolean " autofollow " , default : false , null : false
2019-08-19 11:40:42 +02:00
t . text " comment "
2017-11-27 16:07:59 +01:00
t . index [ " code " ] , name : " index_invites_on_code " , unique : true
t . index [ " user_id " ] , name : " index_invites_on_user_id "
end
2020-10-12 16:33:49 +02:00
create_table " ip_blocks " , force : :cascade do | t |
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
t . datetime " expires_at " , precision : nil
2020-10-12 16:33:49 +02:00
t . inet " ip " , default : " 0.0.0.0 " , null : false
t . integer " severity " , default : 0 , null : false
t . text " comment " , default : " " , null : false
2022-10-25 21:43:44 +02:00
t . index [ " ip " ] , name : " index_ip_blocks_on_ip " , unique : true
2020-10-12 16:33:49 +02:00
end
2017-11-18 00:16:48 +01:00
create_table " list_accounts " , force : :cascade do | t |
t . bigint " list_id " , null : false
t . bigint " account_id " , null : false
2019-11-04 13:02:01 +01:00
t . bigint " follow_id "
2023-05-02 14:40:36 +02:00
t . bigint " follow_request_id "
2017-11-18 00:16:48 +01:00
t . index [ " account_id " , " list_id " ] , name : " index_list_accounts_on_account_id_and_list_id " , unique : true
2022-03-12 08:12:57 +01:00
t . index [ " follow_id " ] , name : " index_list_accounts_on_follow_id " , where : " (follow_id IS NOT NULL) "
2023-05-02 14:40:36 +02:00
t . index [ " follow_request_id " ] , name : " index_list_accounts_on_follow_request_id " , where : " (follow_request_id IS NOT NULL) "
2017-11-18 00:16:48 +01:00
t . index [ " list_id " , " account_id " ] , name : " index_list_accounts_on_list_id_and_account_id "
end
create_table " lists " , force : :cascade do | t |
2017-12-12 04:11:17 +01:00
t . bigint " account_id " , null : false
2017-11-18 00:16:48 +01:00
t . string " title " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2018-11-04 12:02:04 +01:00
t . integer " replies_policy " , default : 0 , null : false
2023-06-07 01:53:50 +02:00
t . boolean " exclusive " , default : false , null : false
2017-11-18 00:16:48 +01:00
t . index [ " account_id " ] , name : " index_lists_on_account_id "
end
2021-06-21 17:07:30 +02:00
create_table " login_activities " , force : :cascade do | t |
t . bigint " user_id " , null : false
t . string " authentication_method "
t . string " provider "
t . boolean " success "
t . string " failure_reason "
t . inet " ip "
t . string " user_agent "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil
2021-06-21 17:07:30 +02:00
t . index [ " user_id " ] , name : " index_login_activities_on_user_id "
end
2019-09-06 13:55:51 +02:00
create_table " markers " , force : :cascade do | t |
t . bigint " user_id "
t . string " timeline " , default : " " , null : false
t . bigint " last_read_id " , default : 0 , null : false
t . integer " lock_version " , default : 0 , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2019-09-06 13:55:51 +02:00
t . index [ " user_id " , " timeline " ] , name : " index_markers_on_user_id_and_timeline " , unique : true
end
2020-07-07 15:26:51 +02:00
create_table " media_attachments " , id : :bigint , default : - > { " timestamp_id('media_attachments'::text) " } , force : :cascade do | t |
2017-06-02 16:18:54 +02:00
t . bigint " status_id "
t . string " file_file_name "
t . string " file_content_type "
t . integer " file_file_size "
2023-07-25 13:40:35 +02:00
t . datetime " file_updated_at " , precision : nil
2017-06-02 16:18:54 +02:00
t . string " remote_url " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2017-06-02 16:18:54 +02:00
t . string " shortcode "
t . integer " type " , default : 0 , null : false
t . json " file_meta "
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
t . bigint " account_id "
2017-09-28 15:31:31 +02:00
t . text " description "
2019-01-05 12:43:28 +01:00
t . bigint " scheduled_status_id "
2019-04-27 03:24:09 +02:00
t . string " blurhash "
2020-03-08 23:56:18 +01:00
t . integer " processing "
2020-04-26 23:29:08 +02:00
t . integer " file_storage_schema_version "
2020-06-29 13:56:55 +02:00
t . string " thumbnail_file_name "
t . string " thumbnail_content_type "
t . integer " thumbnail_file_size "
2023-07-25 13:40:35 +02:00
t . datetime " thumbnail_updated_at " , precision : nil
2020-06-29 13:56:55 +02:00
t . string " thumbnail_remote_url "
2021-04-26 18:57:46 +02:00
t . index [ " account_id " , " status_id " ] , name : " index_media_attachments_on_account_id_and_status_id " , order : { status_id : :desc }
2022-03-12 08:12:57 +01:00
t . index [ " scheduled_status_id " ] , name : " index_media_attachments_on_scheduled_status_id " , where : " (scheduled_status_id IS NOT NULL) "
t . index [ " shortcode " ] , name : " index_media_attachments_on_shortcode " , unique : true , opclass : :text_pattern_ops , where : " (shortcode IS NOT NULL) "
2017-06-02 16:18:54 +02:00
t . index [ " status_id " ] , name : " index_media_attachments_on_status_id "
2016-09-05 17:46:36 +02:00
end
2017-09-22 13:20:04 +02:00
create_table " mentions " , force : :cascade do | t |
2017-06-02 16:18:54 +02:00
t . bigint " status_id "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
t . bigint " account_id "
2018-10-17 17:13:04 +02:00
t . boolean " silent " , default : false , null : false
2017-06-02 16:18:54 +02:00
t . index [ " account_id " , " status_id " ] , name : " index_mentions_on_account_id_and_status_id " , unique : true
t . index [ " status_id " ] , name : " index_mentions_on_status_id "
2016-02-25 00:17:01 +01:00
end
2017-09-22 13:20:04 +02:00
create_table " mutes " , force : :cascade do | t |
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2017-09-14 05:35:48 +02:00
t . boolean " hide_notifications " , default : true , null : false
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
t . bigint " account_id " , null : false
t . bigint " target_account_id " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " expires_at " , precision : nil
2017-06-02 16:18:54 +02:00
t . index [ " account_id " , " target_account_id " ] , name : " index_mutes_on_account_id_and_target_account_id " , unique : true
2018-08-21 20:11:34 +02:00
t . index [ " target_account_id " ] , name : " index_mutes_on_target_account_id "
2017-02-06 02:51:56 +01:00
end
2017-09-22 13:20:04 +02:00
create_table " notifications " , force : :cascade do | t |
2018-03-24 12:51:28 +01:00
t . bigint " activity_id " , null : false
t . string " activity_type " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2018-03-24 12:51:28 +01:00
t . bigint " account_id " , null : false
t . bigint " from_account_id " , null : false
2020-09-18 17:26:45 +02:00
t . string " type "
t . index [ " account_id " , " id " , " type " ] , name : " index_notifications_on_account_id_and_id_and_type " , order : { id : :desc }
2017-06-02 16:18:54 +02:00
t . index [ " activity_id " , " activity_type " ] , name : " index_notifications_on_activity_id_and_activity_type "
2018-08-21 20:11:34 +02:00
t . index [ " from_account_id " ] , name : " index_notifications_on_from_account_id "
2016-11-20 00:33:02 +01:00
end
2017-09-22 13:20:04 +02:00
create_table " oauth_access_grants " , force : :cascade do | t |
2017-06-02 16:18:54 +02:00
t . string " token " , null : false
t . integer " expires_in " , null : false
t . text " redirect_uri " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " revoked_at " , precision : nil
2017-06-02 16:18:54 +02:00
t . string " scopes "
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
t . bigint " application_id " , null : false
t . bigint " resource_owner_id " , null : false
2018-08-21 20:11:34 +02:00
t . index [ " resource_owner_id " ] , name : " index_oauth_access_grants_on_resource_owner_id "
2017-06-02 16:18:54 +02:00
t . index [ " token " ] , name : " index_oauth_access_grants_on_token " , unique : true
2016-03-07 12:42:33 +01:00
end
2017-09-22 13:20:04 +02:00
create_table " oauth_access_tokens " , force : :cascade do | t |
2017-06-02 16:18:54 +02:00
t . string " token " , null : false
t . string " refresh_token "
t . integer " expires_in "
2023-07-25 13:40:35 +02:00
t . datetime " revoked_at " , precision : nil
t . datetime " created_at " , precision : nil , null : false
2017-06-02 16:18:54 +02:00
t . string " scopes "
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
t . bigint " application_id "
t . bigint " resource_owner_id "
2023-07-25 13:40:35 +02:00
t . datetime " last_used_at " , precision : nil
2022-03-01 16:48:58 +01:00
t . inet " last_used_ip "
2022-03-12 08:12:57 +01:00
t . index [ " refresh_token " ] , name : " index_oauth_access_tokens_on_refresh_token " , unique : true , opclass : :text_pattern_ops , where : " (refresh_token IS NOT NULL) "
t . index [ " resource_owner_id " ] , name : " index_oauth_access_tokens_on_resource_owner_id " , where : " (resource_owner_id IS NOT NULL) "
2017-06-02 16:18:54 +02:00
t . index [ " token " ] , name : " index_oauth_access_tokens_on_token " , unique : true
2016-03-07 12:42:33 +01:00
end
2017-09-22 13:20:04 +02:00
create_table " oauth_applications " , force : :cascade do | t |
2017-06-02 16:18:54 +02:00
t . string " name " , null : false
t . string " uid " , null : false
t . string " secret " , null : false
t . text " redirect_uri " , null : false
t . string " scopes " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil
t . datetime " updated_at " , precision : nil
2017-08-23 00:59:35 +02:00
t . boolean " superapp " , default : false , null : false
t . string " website "
t . string " owner_type "
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
t . bigint " owner_id "
2018-08-14 20:24:47 +02:00
t . boolean " confidential " , default : true , null : false
2017-08-23 00:59:35 +02:00
t . index [ " owner_id " , " owner_type " ] , name : " index_oauth_applications_on_owner_id_and_owner_type "
2023-07-02 16:56:16 +02:00
t . index [ " superapp " ] , name : " index_oauth_applications_on_superapp " , where : " (superapp = true) "
2017-06-02 16:18:54 +02:00
t . index [ " uid " ] , name : " index_oauth_applications_on_uid " , unique : true
2016-03-07 12:42:33 +01:00
end
2020-06-02 19:24:53 +02:00
create_table " one_time_keys " , force : :cascade do | t |
t . bigint " device_id "
t . string " key_id " , default : " " , null : false
t . text " key " , default : " " , null : false
t . text " signature " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2020-06-02 19:24:53 +02:00
t . index [ " device_id " ] , name : " index_one_time_keys_on_device_id "
t . index [ " key_id " ] , name : " index_one_time_keys_on_key_id "
end
2018-10-09 19:35:14 +02:00
create_table " pghero_space_stats " , force : :cascade do | t |
t . text " database "
t . text " schema "
t . text " relation "
t . bigint " size "
2023-07-25 13:40:35 +02:00
t . datetime " captured_at " , precision : nil
2018-10-09 19:35:14 +02:00
t . index [ " database " , " captured_at " ] , name : " index_pghero_space_stats_on_database_and_captured_at "
end
2019-03-03 22:18:23 +01:00
create_table " poll_votes " , force : :cascade do | t |
t . bigint " account_id "
t . bigint " poll_id "
t . integer " choice " , default : 0 , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2019-03-04 22:51:23 +01:00
t . string " uri "
2019-03-03 22:18:23 +01:00
t . index [ " account_id " ] , name : " index_poll_votes_on_account_id "
t . index [ " poll_id " ] , name : " index_poll_votes_on_poll_id "
end
create_table " polls " , force : :cascade do | t |
t . bigint " account_id "
t . bigint " status_id "
2023-07-25 13:40:35 +02:00
t . datetime " expires_at " , precision : nil
2019-03-03 22:18:23 +01:00
t . string " options " , default : [ ] , null : false , array : true
t . bigint " cached_tallies " , default : [ ] , null : false , array : true
t . boolean " multiple " , default : false , null : false
t . boolean " hide_totals " , default : false , null : false
t . bigint " votes_count " , default : 0 , null : false
2023-07-25 13:40:35 +02:00
t . datetime " last_fetched_at " , precision : nil
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2019-03-06 19:53:57 +01:00
t . integer " lock_version " , default : 0 , null : false
2019-09-29 22:58:01 +02:00
t . bigint " voters_count "
2019-03-03 22:18:23 +01:00
t . index [ " account_id " ] , name : " index_polls_on_account_id "
t . index [ " status_id " ] , name : " index_polls_on_status_id "
end
2021-11-25 13:07:38 +01:00
create_table " preview_card_providers " , force : :cascade do | t |
t . string " domain " , default : " " , null : false
t . string " icon_file_name "
t . string " icon_content_type "
t . bigint " icon_file_size "
2023-07-25 13:40:35 +02:00
t . datetime " icon_updated_at " , precision : nil
2021-11-25 13:07:38 +01:00
t . boolean " trendable "
2023-07-25 13:40:35 +02:00
t . datetime " reviewed_at " , precision : nil
t . datetime " requested_review_at " , precision : nil
t . datetime " created_at " , null : false
t . datetime " updated_at " , null : false
2021-11-25 13:07:38 +01:00
t . index [ " domain " ] , name : " index_preview_card_providers_on_domain " , unique : true
end
2022-10-08 16:45:40 +02:00
create_table " preview_card_trends " , force : :cascade do | t |
t . bigint " preview_card_id " , null : false
t . float " score " , default : 0 . 0 , null : false
t . integer " rank " , default : 0 , null : false
t . boolean " allowed " , default : false , null : false
t . string " language "
t . index [ " preview_card_id " ] , name : " index_preview_card_trends_on_preview_card_id " , unique : true
end
2017-09-01 16:20:16 +02:00
create_table " preview_cards " , force : :cascade do | t |
2017-06-02 16:18:54 +02:00
t . string " url " , default : " " , null : false
2017-09-01 16:20:16 +02:00
t . string " title " , default : " " , null : false
t . string " description " , default : " " , null : false
2017-06-02 16:18:54 +02:00
t . string " image_file_name "
t . string " image_content_type "
t . integer " image_file_size "
2023-07-25 13:40:35 +02:00
t . datetime " image_updated_at " , precision : nil
2017-06-02 16:18:54 +02:00
t . integer " type " , default : 0 , null : false
t . text " html " , default : " " , null : false
t . string " author_name " , default : " " , null : false
t . string " author_url " , default : " " , null : false
t . string " provider_name " , default : " " , null : false
t . string " provider_url " , default : " " , null : false
t . integer " width " , default : 0 , null : false
t . integer " height " , default : 0 , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2017-12-07 03:37:43 +01:00
t . string " embed_url " , default : " " , null : false
2020-04-26 23:29:08 +02:00
t . integer " image_storage_schema_version "
2020-06-05 23:10:41 +02:00
t . string " blurhash "
2021-11-25 13:07:38 +01:00
t . string " language "
t . float " max_score "
2023-07-25 13:40:35 +02:00
t . datetime " max_score_at " , precision : nil
2021-11-25 13:07:38 +01:00
t . boolean " trendable "
t . integer " link_type "
2023-07-25 13:40:35 +02:00
t . datetime " published_at "
2023-08-03 15:41:51 +02:00
t . string " image_description " , default : " " , null : false
2017-09-01 16:20:16 +02:00
t . index [ " url " ] , name : " index_preview_cards_on_url " , unique : true
end
2023-08-03 11:12:52 +02:00
create_table " preview_cards_statuses " , primary_key : [ " status_id " , " preview_card_id " ] , force : :cascade do | t |
2017-09-01 16:20:16 +02:00
t . bigint " preview_card_id " , null : false
t . bigint " status_id " , null : false
2017-01-20 01:00:14 +01:00
end
2018-07-13 02:16:06 +02:00
create_table " relays " , force : :cascade do | t |
t . string " inbox_url " , default : " " , null : false
t . string " follow_activity_id "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2018-08-12 18:16:26 +02:00
t . integer " state " , default : 0 , null : false
2018-07-13 02:16:06 +02:00
end
2018-04-02 22:04:14 +02:00
create_table " report_notes " , force : :cascade do | t |
t . text " content " , null : false
t . bigint " report_id " , null : false
t . bigint " account_id " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2018-04-02 22:04:14 +02:00
t . index [ " account_id " ] , name : " index_report_notes_on_account_id "
t . index [ " report_id " ] , name : " index_report_notes_on_report_id "
end
2017-09-22 13:20:04 +02:00
create_table " reports " , force : :cascade do | t |
2017-06-02 16:18:54 +02:00
t . bigint " status_ids " , default : [ ] , null : false , array : true
t . text " comment " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
t . bigint " account_id " , null : false
2017-09-22 13:20:04 +02:00
t . bigint " action_taken_by_account_id "
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
t . bigint " target_account_id " , null : false
2018-04-02 22:04:14 +02:00
t . bigint " assigned_account_id "
2019-03-17 15:34:56 +01:00
t . string " uri "
2020-12-15 04:30:15 +01:00
t . boolean " forwarded "
2022-01-17 09:41:33 +01:00
t . integer " category " , default : 0 , null : false
2023-07-25 13:40:35 +02:00
t . datetime " action_taken_at " , precision : nil
2022-01-17 09:41:33 +01:00
t . bigint " rule_ids " , array : true
2017-06-02 16:18:54 +02:00
t . index [ " account_id " ] , name : " index_reports_on_account_id "
2022-04-28 17:15:11 +02:00
t . index [ " action_taken_by_account_id " ] , name : " index_reports_on_action_taken_by_account_id " , where : " (action_taken_by_account_id IS NOT NULL) "
t . index [ " assigned_account_id " ] , name : " index_reports_on_assigned_account_id " , where : " (assigned_account_id IS NOT NULL) "
2017-06-02 16:18:54 +02:00
t . index [ " target_account_id " ] , name : " index_reports_on_target_account_id "
2017-02-14 20:59:26 +01:00
end
2021-02-21 19:50:12 +01:00
create_table " rules " , force : :cascade do | t |
t . integer " priority " , default : 0 , null : false
2023-07-25 13:40:35 +02:00
t . datetime " deleted_at " , precision : nil
2021-02-21 19:50:12 +01:00
t . text " text " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2021-02-21 19:50:12 +01:00
end
2019-01-05 12:43:28 +01:00
create_table " scheduled_statuses " , force : :cascade do | t |
t . bigint " account_id "
2023-07-25 13:40:35 +02:00
t . datetime " scheduled_at " , precision : nil
2019-01-05 12:43:28 +01:00
t . jsonb " params "
t . index [ " account_id " ] , name : " index_scheduled_statuses_on_account_id "
t . index [ " scheduled_at " ] , name : " index_scheduled_statuses_on_scheduled_at "
end
2017-06-23 18:50:53 +02:00
create_table " session_activations " , force : :cascade do | t |
t . string " session_id " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2017-06-25 16:54:30 +02:00
t . string " user_agent " , default : " " , null : false
t . inet " ip "
2017-09-22 13:20:04 +02:00
t . bigint " access_token_id "
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
t . bigint " user_id " , null : false
2017-09-22 13:20:04 +02:00
t . bigint " web_push_subscription_id "
2018-08-21 20:11:34 +02:00
t . index [ " access_token_id " ] , name : " index_session_activations_on_access_token_id "
2017-06-23 18:50:53 +02:00
t . index [ " session_id " ] , name : " index_session_activations_on_session_id " , unique : true
t . index [ " user_id " ] , name : " index_session_activations_on_user_id "
end
2017-09-22 13:20:04 +02:00
create_table " settings " , force : :cascade do | t |
2017-06-02 16:18:54 +02:00
t . string " var " , null : false
t . text " value "
t . string " thing_type "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil
t . datetime " updated_at " , precision : nil
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
t . bigint " thing_id "
2017-06-02 16:18:54 +02:00
t . index [ " thing_type " , " thing_id " , " var " ] , name : " index_settings_on_thing_type_and_thing_id_and_var " , unique : true
2016-10-07 13:17:56 +02:00
end
2017-09-14 00:04:30 +02:00
create_table " site_uploads " , force : :cascade do | t |
t . string " var " , default : " " , null : false
t . string " file_file_name "
t . string " file_content_type "
t . integer " file_file_size "
2023-07-25 13:40:35 +02:00
t . datetime " file_updated_at " , precision : nil
2017-09-14 00:04:30 +02:00
t . json " meta "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2022-10-13 11:29:19 +02:00
t . string " blurhash "
2017-09-14 00:04:30 +02:00
t . index [ " var " ] , name : " index_site_uploads_on_var " , unique : true
end
2022-01-19 22:37:27 +01:00
create_table " status_edits " , force : :cascade do | t |
t . bigint " status_id " , null : false
t . bigint " account_id "
t . text " text " , default : " " , null : false
t . text " spoiler_text " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , null : false
t . datetime " updated_at " , null : false
2022-02-09 17:59:43 +01:00
t . string " content_type "
2022-03-09 09:06:17 +01:00
t . bigint " ordered_media_attachment_ids " , array : true
t . text " media_descriptions " , array : true
t . string " poll_options " , array : true
t . boolean " sensitive "
2022-01-19 22:37:27 +01:00
t . index [ " account_id " ] , name : " index_status_edits_on_account_id "
t . index [ " status_id " ] , name : " index_status_edits_on_status_id "
end
2017-08-25 01:41:18 +02:00
create_table " status_pins " , force : :cascade do | t |
t . bigint " account_id " , null : false
t . bigint " status_id " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , default : - > { " now() " } , null : false
t . datetime " updated_at " , precision : nil , default : - > { " now() " } , null : false
2017-08-25 01:41:18 +02:00
t . index [ " account_id " , " status_id " ] , name : " index_status_pins_on_account_id_and_status_id " , unique : true
2022-04-28 17:15:11 +02:00
t . index [ " status_id " ] , name : " index_status_pins_on_status_id "
2017-08-25 01:41:18 +02:00
end
2018-08-14 19:19:32 +02:00
create_table " status_stats " , force : :cascade do | t |
t . bigint " status_id " , null : false
t . bigint " replies_count " , default : 0 , null : false
t . bigint " reblogs_count " , default : 0 , null : false
t . bigint " favourites_count " , default : 0 , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2018-08-14 19:19:32 +02:00
t . index [ " status_id " ] , name : " index_status_stats_on_status_id " , unique : true
end
2022-10-08 16:45:40 +02:00
create_table " status_trends " , force : :cascade do | t |
t . bigint " status_id " , null : false
t . bigint " account_id " , null : false
t . float " score " , default : 0 . 0 , null : false
t . integer " rank " , default : 0 , null : false
t . boolean " allowed " , default : false , null : false
t . string " language "
t . index [ " account_id " ] , name : " index_status_trends_on_account_id "
t . index [ " status_id " ] , name : " index_status_trends_on_status_id " , unique : true
end
Non-Serial ("Snowflake") IDs (#4801)
* Use non-serial IDs
This change makes a number of nontrivial tweaks to the data model in
Mastodon:
* All IDs are now 8 byte integers (rather than mixed 4- and 8-byte)
* IDs are now assigned as:
* Top 6 bytes: millisecond-resolution time from epoch
* Bottom 2 bytes: serial (within the millisecond) sequence number
* See /lib/tasks/db.rake's `define_timestamp_id` for details, but
note that the purpose of these changes is to make it difficult to
determine the number of objects in a table from the ID of any
object.
* The Redis sorted set used for the feed will have values used to look
up toots, rather than scores. This is almost always the same as the
existing behavior, except in the case of boosted toots. This change
was made because Redis stores scores as double-precision floats,
which cannot store the new ID format exactly. Note that this doesn't
cause problems with sorting/pagination, because ZREVRANGEBYSCORE
sorts lexicographically when scores are tied. (This will still cause
sorting issues when the ID gains a new significant digit, but that's
extraordinarily uncommon.)
Note a couple of tradeoffs have been made in this commit:
* lib/tasks/db.rake is used to enforce many/most column constraints,
because this commit seems likely to take a while to bring upstream.
Enforcing a post-migrate hook is an easier way to maintain the code
in the interim.
* Boosted toots will appear in the timeline as many times as they have
been boosted. This is a tradeoff due to the way the feed is saved in
Redis at the moment, but will be handled by a future commit.
This would effectively close Mastodon's #1059, as it is a
snowflake-like system of generating IDs. However, given how involved
the changes were simply within Mastodon, it may have unexpected
interactions with some clients, if they store IDs as doubles
(or as 4-byte integers). This was a problem that Twitter ran into with
their "snowflake" transition, particularly in JavaScript clients that
treated IDs as JS integers, rather than strings. It therefore would be
useful to test these changes at least in the web interface and popular
clients before pushing them to all users.
* Fix JavaScript interface with long IDs
Somewhat predictably, the JS interface handled IDs as numbers, which in
JS are IEEE double-precision floats. This loses some precision when
working with numbers as large as those generated by the new ID scheme,
so we instead handle them here as strings. This is relatively simple,
and doesn't appear to have caused any problems, but should definitely
be tested more thoroughly than the built-in tests. Several days of use
appear to support this working properly.
BREAKING CHANGE:
The major(!) change here is that IDs are now returned as strings by the
REST endpoints, rather than as integers. In practice, relatively few
changes were required to make the existing JS UI work with this change,
but it will likely hit API clients pretty hard: it's an entirely
different type to consume. (The one API client I tested, Tusky, handles
this with no problems, however.)
Twitter ran into this issue when introducing Snowflake IDs, and decided
to instead introduce an `id_str` field in JSON responses. I have opted
to *not* do that, and instead force all IDs to 64-bit integers
represented by strings in one go. (I believe Twitter exacerbated their
problem by rolling out the changes three times: once for statuses, once
for DMs, and once for user IDs, as well as by leaving an integer ID
value in JSON. As they said, "If you’re using the `id` field with JSON
in a Javascript-related language, there is a very high likelihood that
the integers will be silently munged by Javascript interpreters. In most
cases, this will result in behavior such as being unable to load or
delete a specific direct message, because the ID you're sending to the
API is different than the actual identifier associated with the
message." [1]) However, given that this is a significant change for API
users, alternatives or a transition time may be appropriate.
1: https://blog.twitter.com/developer/en_us/a/2011/direct-messages-going-snowflake-on-sep-30-2011.html
* Restructure feed pushes/unpushes
This was necessary because the previous behavior used Redis zset scores
to identify statuses, but those are IEEE double-precision floats, so we
can't actually use them to identify all 64-bit IDs. However, it leaves
the code in a much better state for refactoring reblog handling /
coalescing.
Feed-management code has been consolidated in FeedManager, including:
* BatchedRemoveStatusService no longer directly manipulates feed zsets
* RemoveStatusService no longer directly manipulates feed zsets
* PrecomputeFeedService has moved its logic to FeedManager#populate_feed
(PrecomputeFeedService largely made lots of calls to FeedManager, but
didn't follow the normal adding-to-feed process.)
This has the effect of unifying all of the feed push/unpush logic in
FeedManager, making it much more tractable to update it in the future.
Due to some additional checks that must be made during, for example,
batch status removals, some Redis pipelining has been removed. It does
not appear that this should cause significantly increased load, but if
necessary, some optimizations are possible in batch cases. These were
omitted in the pursuit of simplicity, but a batch_push and batch_unpush
would be possible in the future.
Tests were added to verify that pushes happen under expected conditions,
and to verify reblog behavior (both on pushing and unpushing). In the
case of unpushing, this includes testing behavior that currently leads
to confusion such as Mastodon's #2817, but this codifies that the
behavior is currently expected.
* Rubocop fixes
I could swear I made these changes already, but I must have lost them
somewhere along the line.
* Address review comments
This addresses the first two comments from review of this feature:
https://github.com/tootsuite/mastodon/pull/4801#discussion_r139336735
https://github.com/tootsuite/mastodon/pull/4801#discussion_r139336931
This adds an optional argument to FeedManager#key, the subtype of feed
key to generate. It also tests to ensure that FeedManager's settings are
such that reblogs won't be tracked forever.
* Hardcode IdToBigints migration columns
This addresses a comment during review:
https://github.com/tootsuite/mastodon/pull/4801#discussion_r139337452
This means we'll need to make sure that all _id columns going forward
are bigints, but that should happen automatically in most cases.
* Additional fixes for stringified IDs in JSON
These should be the last two. These were identified using eslint to try
to identify any plain casts to JavaScript numbers. (Some such casts are
legitimate, but these were not.)
Adding the following to .eslintrc.yml will identify casts to numbers:
~~~
no-restricted-syntax:
- warn
- selector: UnaryExpression[operator='+'] > :not(Literal)
message: Avoid the use of unary +
- selector: CallExpression[callee.name='Number']
message: Casting with Number() may coerce string IDs to numbers
~~~
The remaining three casts appear legitimate: two casts to array indices,
one in a server to turn an environment variable into a number.
* Only implement timestamp IDs for Status IDs
Per discussion in #4801, this is only being merged in for Status IDs at
this point. We do this in a migration, as there is no longer use for
a post-migration hook. We keep the initialization of the timestamp_id
function as a Rake task, as it is also needed after db:schema:load (as
db/schema.rb doesn't store Postgres functions).
* Change internal streaming payloads to stringified IDs as well
This is equivalent to 591a9af356faf2d5c7e66e3ec715502796c875cd from
#5019, with an extra change for the addition to FeedManager#unpush.
* Ensure we have a status_id_seq sequence
Apparently this is not a given when specifying a custom ID function,
so now we ensure it gets created. This uses the generic version of this
function to more easily support adding additional tables with timestamp
IDs in the future, although it would be possible to cut this down to a
less generic version if necessary. It is only run during db:schema:load
or the relevant migration, so the overhead is extraordinarily minimal.
* Transition reblogs to new Redis format
This provides a one-way migration to transition old Redis reblog entries
into the new format, with a separate tracking entry for reblogs.
It is not invertible because doing so could (if timestamp IDs are used)
require a database query for each status in each users' feed, which is
likely to be a significant toll on major instances.
* Address review comments from @akihikodaki
No functional changes.
* Additional review changes
* Heredoc cleanup
* Run db:schema:load hooks for test in development
This matches the behavior in Rails'
ActiveRecord::Tasks::DatabaseTasks.each_current_configuration, which
would otherwise break `rake db:setup` in development.
It also moves some functionality out to a library, which will be a good
place to put additional related functionality in the near future.
2017-10-04 09:56:37 +02:00
create_table " statuses " , id : :bigint , default : - > { " timestamp_id('statuses'::text) " } , force : :cascade do | t |
2017-06-02 16:18:54 +02:00
t . string " uri "
t . text " text " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2017-06-02 16:18:54 +02:00
t . bigint " in_reply_to_id "
t . bigint " reblog_of_id "
t . string " url "
2017-07-13 03:12:25 +02:00
t . boolean " sensitive " , default : false , null : false
2017-06-02 16:18:54 +02:00
t . integer " visibility " , default : 0 , null : false
t . text " spoiler_text " , default : " " , null : false
2017-07-13 03:12:25 +02:00
t . boolean " reply " , default : false , null : false
2017-06-09 18:09:37 +02:00
t . string " language "
2017-06-02 16:18:54 +02:00
t . bigint " conversation_id "
2017-09-06 20:57:52 +02:00
t . boolean " local "
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
t . bigint " account_id " , null : false
t . bigint " application_id "
t . bigint " in_reply_to_account_id "
2017-12-10 23:41:25 +01:00
t . boolean " local_only "
2019-03-03 22:18:23 +01:00
t . bigint " poll_id "
2019-05-12 20:14:35 +02:00
t . string " content_type "
2023-07-25 13:40:35 +02:00
t . datetime " deleted_at " , precision : nil
t . datetime " edited_at " , precision : nil
2022-02-25 00:34:14 +01:00
t . boolean " trendable "
2022-03-09 09:06:17 +01:00
t . bigint " ordered_media_attachment_ids " , array : true
2019-08-22 21:55:56 +02:00
t . index [ " account_id " , " id " , " visibility " , " updated_at " ] , name : " index_statuses_20190820 " , order : { id : :desc } , where : " (deleted_at IS NULL) "
2022-04-28 17:15:11 +02:00
t . index [ " account_id " ] , name : " index_statuses_on_account_id "
2022-01-17 09:41:33 +01:00
t . index [ " deleted_at " ] , name : " index_statuses_on_deleted_at " , where : " (deleted_at IS NOT NULL) "
2019-08-24 04:12:27 +02:00
t . index [ " id " , " account_id " ] , name : " index_statuses_local_20190824 " , order : { id : :desc } , where : " ((local OR (uri IS NULL)) AND (deleted_at IS NULL) AND (visibility = 0) AND (reblog_of_id IS NULL) AND ((NOT reply) OR (in_reply_to_account_id = account_id))) "
2020-01-21 18:56:34 +01:00
t . index [ " id " , " account_id " ] , name : " index_statuses_public_20200119 " , order : { id : :desc } , where : " ((deleted_at IS NULL) AND (visibility = 0) AND (reblog_of_id IS NULL) AND ((NOT reply) OR (in_reply_to_account_id = account_id))) "
2022-03-12 08:12:57 +01:00
t . index [ " in_reply_to_account_id " ] , name : " index_statuses_on_in_reply_to_account_id " , where : " (in_reply_to_account_id IS NOT NULL) "
t . index [ " in_reply_to_id " ] , name : " index_statuses_on_in_reply_to_id " , where : " (in_reply_to_id IS NOT NULL) "
2017-11-27 20:22:27 +01:00
t . index [ " reblog_of_id " , " account_id " ] , name : " index_statuses_on_reblog_of_id_and_account_id "
2022-03-12 08:12:57 +01:00
t . index [ " uri " ] , name : " index_statuses_on_uri " , unique : true , opclass : :text_pattern_ops , where : " (uri IS NOT NULL) "
2016-11-05 15:20:05 +01:00
end
2023-06-01 09:49:06 +02:00
create_table " statuses_tags " , primary_key : [ " tag_id " , " status_id " ] , force : :cascade do | t |
2017-06-02 16:18:54 +02:00
t . bigint " status_id " , null : false
2017-09-22 13:20:04 +02:00
t . bigint " tag_id " , null : false
2017-06-02 16:18:54 +02:00
t . index [ " status_id " ] , name : " index_statuses_tags_on_status_id "
2016-02-20 22:53:20 +01:00
end
2020-06-02 19:24:53 +02:00
create_table " system_keys " , force : :cascade do | t |
t . binary " key "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2020-06-02 19:24:53 +02:00
end
2022-07-17 13:49:29 +02:00
create_table " tag_follows " , force : :cascade do | t |
t . bigint " tag_id " , null : false
t . bigint " account_id " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , null : false
t . datetime " updated_at " , null : false
2022-07-17 13:49:29 +02:00
t . index [ " account_id " , " tag_id " ] , name : " index_tag_follows_on_account_id_and_tag_id " , unique : true
t . index [ " tag_id " ] , name : " index_tag_follows_on_tag_id "
end
2017-09-22 13:20:04 +02:00
create_table " tags " , force : :cascade do | t |
2017-06-02 16:18:54 +02:00
t . string " name " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2019-08-05 19:54:29 +02:00
t . boolean " usable "
t . boolean " trendable "
t . boolean " listable "
2023-07-25 13:40:35 +02:00
t . datetime " reviewed_at " , precision : nil
t . datetime " requested_review_at " , precision : nil
t . datetime " last_status_at " , precision : nil
2019-09-02 18:11:13 +02:00
t . float " max_score "
2023-07-25 13:40:35 +02:00
t . datetime " max_score_at " , precision : nil
2022-07-13 15:03:28 +02:00
t . string " display_name "
2021-04-25 06:33:28 +02:00
t . index " lower((name)::text) text_pattern_ops " , name : " index_tags_on_name_lower_btree " , unique : true
2016-11-04 19:12:59 +01:00
end
2019-01-18 15:56:55 +01:00
create_table " tombstones " , force : :cascade do | t |
t . bigint " account_id "
t . string " uri " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2019-05-09 22:03:03 +02:00
t . boolean " by_moderator "
2019-01-18 15:56:55 +01:00
t . index [ " account_id " ] , name : " index_tombstones_on_account_id "
t . index [ " uri " ] , name : " index_tombstones_on_uri "
end
2020-04-15 20:33:24 +02:00
create_table " unavailable_domains " , force : :cascade do | t |
t . string " domain " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2020-04-15 20:33:24 +02:00
t . index [ " domain " ] , name : " index_unavailable_domains_on_domain " , unique : true
end
2019-04-09 16:06:30 +02:00
create_table " user_invite_requests " , force : :cascade do | t |
t . bigint " user_id "
t . text " text "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2019-04-09 16:06:30 +02:00
t . index [ " user_id " ] , name : " index_user_invite_requests_on_user_id "
end
2022-07-05 02:41:40 +02:00
create_table " user_roles " , force : :cascade do | t |
t . string " name " , default : " " , null : false
t . string " color " , default : " " , null : false
t . integer " position " , default : 0 , null : false
t . bigint " permissions " , default : 0 , null : false
t . boolean " highlighted " , default : false , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , null : false
t . datetime " updated_at " , null : false
2022-07-05 02:41:40 +02:00
end
2017-09-22 13:20:04 +02:00
create_table " users " , force : :cascade do | t |
2017-06-02 16:18:54 +02:00
t . string " email " , default : " " , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2017-06-02 16:18:54 +02:00
t . string " encrypted_password " , default : " " , null : false
t . string " reset_password_token "
2023-07-25 13:40:35 +02:00
t . datetime " reset_password_sent_at " , precision : nil
2017-06-02 16:18:54 +02:00
t . integer " sign_in_count " , default : 0 , null : false
2023-07-25 13:40:35 +02:00
t . datetime " current_sign_in_at " , precision : nil
t . datetime " last_sign_in_at " , precision : nil
2017-07-13 03:12:25 +02:00
t . boolean " admin " , default : false , null : false
2017-06-02 16:18:54 +02:00
t . string " confirmation_token "
2023-07-25 13:40:35 +02:00
t . datetime " confirmed_at " , precision : nil
t . datetime " confirmation_sent_at " , precision : nil
2017-06-02 16:18:54 +02:00
t . string " unconfirmed_email "
t . string " locale "
t . string " encrypted_otp_secret "
t . string " encrypted_otp_secret_iv "
t . string " encrypted_otp_secret_salt "
t . integer " consumed_timestep "
2017-07-13 03:12:25 +02:00
t . boolean " otp_required_for_login " , default : false , null : false
2023-07-25 13:40:35 +02:00
t . datetime " last_emailed_at " , precision : nil
2017-06-02 16:18:54 +02:00
t . string " otp_backup_codes " , array : true
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
t . bigint " account_id " , null : false
2017-11-07 19:06:44 +01:00
t . boolean " disabled " , default : false , null : false
2017-11-11 20:23:33 +01:00
t . boolean " moderator " , default : false , null : false
2017-11-27 16:07:59 +01:00
t . bigint " invite_id "
2018-06-17 13:54:02 +02:00
t . string " chosen_languages " , array : true
2018-12-24 19:12:38 +01:00
t . bigint " created_by_application_id "
2019-03-14 05:28:30 +01:00
t . boolean " approved " , default : true , null : false
2020-06-09 10:23:06 +02:00
t . string " sign_in_token "
2023-07-25 13:40:35 +02:00
t . datetime " sign_in_token_sent_at " , precision : nil
Add WebAuthn as an alternative 2FA method (#14466)
* feat: add possibility of adding WebAuthn security keys to use as 2FA
This adds a basic UI for enabling WebAuthn 2FA. We did a little refactor
to the Settings page for editing the 2FA methods – now it will list the
methods that are available to the user (TOTP and WebAuthn) and from
there they'll be able to add or remove any of them.
Also, it's worth mentioning that for enabling WebAuthn it's required to
have TOTP enabled, so the first time that you go to the 2FA Settings
page, you'll be asked to set it up.
This work was inspired by the one donde by Github in their platform, and
despite it could be approached in different ways, we decided to go with
this one given that we feel that this gives a great UX.
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
* feat: add request for WebAuthn as second factor at login if enabled
This commits adds the feature for using WebAuthn as a second factor for
login when enabled.
If users have WebAuthn enabled, now a page requesting for the use of a
WebAuthn credential for log in will appear, although a link redirecting
to the old page for logging in using a two-factor code will also be
present.
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
* feat: add possibility of deleting WebAuthn Credentials
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
* feat: disable WebAuthn when an Admin disables 2FA for a user
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
* feat: remove ability to disable TOTP leaving only WebAuthn as 2FA
Following examples form other platforms like Github, we decided to make
Webauthn 2FA secondary to 2FA with TOTP, so that we removed the
possibility of removing TOTP authentication only, leaving users with
just WEbAuthn as 2FA. Instead, users will have to click on 'Disable 2FA'
in order to remove second factor auth.
The reason for WebAuthn being secondary to TOPT is that in that way,
users will still be able to log in using their code from their phone's
application if they don't have their security keys with them – or maybe
even lost them.
* We had to change a little the flow for setting up TOTP, given that now
it's possible to setting up again if you already had TOTP, in order to
let users modify their authenticator app – given that now it's not
possible for them to disable TOTP and set it up again with another
authenticator app.
So, basically, now instead of storing the new `otp_secret` in the
user, we store it in the session until the process of set up is
finished.
This was because, as it was before, when users clicked on 'Edit' in
the new two-factor methods lists page, but then went back without
finishing the flow, their `otp_secret` had been changed therefore
invalidating their previous authenticator app, making them unable to
log in again using TOTP.
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
* refactor: fix eslint errors
The PR build was failing given that linting returning some errors.
This commit attempts to fix them.
* refactor: normalize i18n translations
The build was failing given that i18n translations files were not
normalized.
This commits fixes that.
* refactor: avoid having the webauthn gem locked to a specific version
* refactor: use symbols for routes without '/'
* refactor: avoid sending webauthn disabled email when 2FA is disabled
When an admins disable 2FA for users, we were sending two mails
to them, one notifying that 2FA was disabled and the other to notify
that WebAuthn was disabled.
As the second one is redundant since the first email includes it, we can
remove it and send just one email to users.
* refactor: avoid creating new env variable for webauthn_origin config
* refactor: improve flash error messages for webauthn pages
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
2020-08-24 16:46:27 +02:00
t . string " webauthn_id "
2020-10-12 16:33:49 +02:00
t . inet " sign_up_ip "
2021-07-08 05:31:28 +02:00
t . boolean " skip_sign_in_token "
2022-07-05 02:41:40 +02:00
t . bigint " role_id "
2023-03-30 14:44:00 +02:00
t . text " settings "
2023-06-10 03:29:37 +02:00
t . string " time_zone "
2017-06-02 16:18:54 +02:00
t . index [ " account_id " ] , name : " index_users_on_account_id "
t . index [ " confirmation_token " ] , name : " index_users_on_confirmation_token " , unique : true
2022-03-12 08:12:57 +01:00
t . index [ " created_by_application_id " ] , name : " index_users_on_created_by_application_id " , where : " (created_by_application_id IS NOT NULL) "
2017-06-02 16:18:54 +02:00
t . index [ " email " ] , name : " index_users_on_email " , unique : true
2022-03-12 08:12:57 +01:00
t . index [ " reset_password_token " ] , name : " index_users_on_reset_password_token " , unique : true , opclass : :text_pattern_ops , where : " (reset_password_token IS NOT NULL) "
2022-07-05 02:41:40 +02:00
t . index [ " role_id " ] , name : " index_users_on_role_id " , where : " (role_id IS NOT NULL) "
2023-07-02 19:41:35 +02:00
t . index [ " unconfirmed_email " ] , name : " index_users_on_unconfirmed_email " , where : " (unconfirmed_email IS NOT NULL) "
2016-02-22 16:00:20 +01:00
end
2017-07-13 22:15:32 +02:00
create_table " web_push_subscriptions " , force : :cascade do | t |
t . string " endpoint " , null : false
t . string " key_p256dh " , null : false
t . string " key_auth " , null : false
t . json " data "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2018-05-11 11:49:12 +02:00
t . bigint " access_token_id "
t . bigint " user_id "
2022-03-12 08:12:57 +01:00
t . index [ " access_token_id " ] , name : " index_web_push_subscriptions_on_access_token_id " , where : " (access_token_id IS NOT NULL) "
2018-05-11 11:49:12 +02:00
t . index [ " user_id " ] , name : " index_web_push_subscriptions_on_user_id "
2017-07-13 22:15:32 +02:00
end
2017-09-22 13:20:04 +02:00
create_table " web_settings " , force : :cascade do | t |
2017-06-02 16:18:54 +02:00
t . json " data "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
2018-02-07 16:35:44 +01:00
t . bigint " user_id " , null : false
2017-06-02 16:18:54 +02:00
t . index [ " user_id " ] , name : " index_web_settings_on_user_id " , unique : true
2017-01-09 14:00:55 +01:00
end
Add WebAuthn as an alternative 2FA method (#14466)
* feat: add possibility of adding WebAuthn security keys to use as 2FA
This adds a basic UI for enabling WebAuthn 2FA. We did a little refactor
to the Settings page for editing the 2FA methods – now it will list the
methods that are available to the user (TOTP and WebAuthn) and from
there they'll be able to add or remove any of them.
Also, it's worth mentioning that for enabling WebAuthn it's required to
have TOTP enabled, so the first time that you go to the 2FA Settings
page, you'll be asked to set it up.
This work was inspired by the one donde by Github in their platform, and
despite it could be approached in different ways, we decided to go with
this one given that we feel that this gives a great UX.
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
* feat: add request for WebAuthn as second factor at login if enabled
This commits adds the feature for using WebAuthn as a second factor for
login when enabled.
If users have WebAuthn enabled, now a page requesting for the use of a
WebAuthn credential for log in will appear, although a link redirecting
to the old page for logging in using a two-factor code will also be
present.
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
* feat: add possibility of deleting WebAuthn Credentials
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
* feat: disable WebAuthn when an Admin disables 2FA for a user
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
* feat: remove ability to disable TOTP leaving only WebAuthn as 2FA
Following examples form other platforms like Github, we decided to make
Webauthn 2FA secondary to 2FA with TOTP, so that we removed the
possibility of removing TOTP authentication only, leaving users with
just WEbAuthn as 2FA. Instead, users will have to click on 'Disable 2FA'
in order to remove second factor auth.
The reason for WebAuthn being secondary to TOPT is that in that way,
users will still be able to log in using their code from their phone's
application if they don't have their security keys with them – or maybe
even lost them.
* We had to change a little the flow for setting up TOTP, given that now
it's possible to setting up again if you already had TOTP, in order to
let users modify their authenticator app – given that now it's not
possible for them to disable TOTP and set it up again with another
authenticator app.
So, basically, now instead of storing the new `otp_secret` in the
user, we store it in the session until the process of set up is
finished.
This was because, as it was before, when users clicked on 'Edit' in
the new two-factor methods lists page, but then went back without
finishing the flow, their `otp_secret` had been changed therefore
invalidating their previous authenticator app, making them unable to
log in again using TOTP.
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
* refactor: fix eslint errors
The PR build was failing given that linting returning some errors.
This commit attempts to fix them.
* refactor: normalize i18n translations
The build was failing given that i18n translations files were not
normalized.
This commits fixes that.
* refactor: avoid having the webauthn gem locked to a specific version
* refactor: use symbols for routes without '/'
* refactor: avoid sending webauthn disabled email when 2FA is disabled
When an admins disable 2FA for users, we were sending two mails
to them, one notifying that 2FA was disabled and the other to notify
that WebAuthn was disabled.
As the second one is redundant since the first email includes it, we can
remove it and send just one email to users.
* refactor: avoid creating new env variable for webauthn_origin config
* refactor: improve flash error messages for webauthn pages
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
2020-08-24 16:46:27 +02:00
create_table " webauthn_credentials " , force : :cascade do | t |
t . string " external_id " , null : false
t . string " public_key " , null : false
t . string " nickname " , null : false
t . bigint " sign_count " , default : 0 , null : false
t . bigint " user_id "
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , precision : nil , null : false
t . datetime " updated_at " , precision : nil , null : false
Add WebAuthn as an alternative 2FA method (#14466)
* feat: add possibility of adding WebAuthn security keys to use as 2FA
This adds a basic UI for enabling WebAuthn 2FA. We did a little refactor
to the Settings page for editing the 2FA methods – now it will list the
methods that are available to the user (TOTP and WebAuthn) and from
there they'll be able to add or remove any of them.
Also, it's worth mentioning that for enabling WebAuthn it's required to
have TOTP enabled, so the first time that you go to the 2FA Settings
page, you'll be asked to set it up.
This work was inspired by the one donde by Github in their platform, and
despite it could be approached in different ways, we decided to go with
this one given that we feel that this gives a great UX.
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
* feat: add request for WebAuthn as second factor at login if enabled
This commits adds the feature for using WebAuthn as a second factor for
login when enabled.
If users have WebAuthn enabled, now a page requesting for the use of a
WebAuthn credential for log in will appear, although a link redirecting
to the old page for logging in using a two-factor code will also be
present.
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
* feat: add possibility of deleting WebAuthn Credentials
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
* feat: disable WebAuthn when an Admin disables 2FA for a user
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
* feat: remove ability to disable TOTP leaving only WebAuthn as 2FA
Following examples form other platforms like Github, we decided to make
Webauthn 2FA secondary to 2FA with TOTP, so that we removed the
possibility of removing TOTP authentication only, leaving users with
just WEbAuthn as 2FA. Instead, users will have to click on 'Disable 2FA'
in order to remove second factor auth.
The reason for WebAuthn being secondary to TOPT is that in that way,
users will still be able to log in using their code from their phone's
application if they don't have their security keys with them – or maybe
even lost them.
* We had to change a little the flow for setting up TOTP, given that now
it's possible to setting up again if you already had TOTP, in order to
let users modify their authenticator app – given that now it's not
possible for them to disable TOTP and set it up again with another
authenticator app.
So, basically, now instead of storing the new `otp_secret` in the
user, we store it in the session until the process of set up is
finished.
This was because, as it was before, when users clicked on 'Edit' in
the new two-factor methods lists page, but then went back without
finishing the flow, their `otp_secret` had been changed therefore
invalidating their previous authenticator app, making them unable to
log in again using TOTP.
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
* refactor: fix eslint errors
The PR build was failing given that linting returning some errors.
This commit attempts to fix them.
* refactor: normalize i18n translations
The build was failing given that i18n translations files were not
normalized.
This commits fixes that.
* refactor: avoid having the webauthn gem locked to a specific version
* refactor: use symbols for routes without '/'
* refactor: avoid sending webauthn disabled email when 2FA is disabled
When an admins disable 2FA for users, we were sending two mails
to them, one notifying that 2FA was disabled and the other to notify
that WebAuthn was disabled.
As the second one is redundant since the first email includes it, we can
remove it and send just one email to users.
* refactor: avoid creating new env variable for webauthn_origin config
* refactor: improve flash error messages for webauthn pages
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
2020-08-24 16:46:27 +02:00
t . index [ " external_id " ] , name : " index_webauthn_credentials_on_external_id " , unique : true
t . index [ " user_id " ] , name : " index_webauthn_credentials_on_user_id "
end
2022-06-09 21:57:36 +02:00
create_table " webhooks " , force : :cascade do | t |
t . string " url " , null : false
t . string " events " , default : [ ] , null : false , array : true
t . string " secret " , default : " " , null : false
t . boolean " enabled " , default : true , null : false
2023-07-25 13:40:35 +02:00
t . datetime " created_at " , null : false
t . datetime " updated_at " , null : false
2023-06-06 10:42:47 +02:00
t . text " template "
2022-06-09 21:57:36 +02:00
t . index [ " url " ] , name : " index_webhooks_on_url " , unique : true
end
2019-09-19 20:58:19 +02:00
add_foreign_key " account_aliases " , " accounts " , on_delete : :cascade
2018-10-07 23:44:58 +02:00
add_foreign_key " account_conversations " , " accounts " , on_delete : :cascade
add_foreign_key " account_conversations " , " conversations " , on_delete : :cascade
2020-09-15 14:37:58 +02:00
add_foreign_key " account_deletion_requests " , " accounts " , on_delete : :cascade
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
add_foreign_key " account_domain_blocks " , " accounts " , name : " fk_206c6029bd " , on_delete : :cascade
2019-09-19 20:58:19 +02:00
add_foreign_key " account_migrations " , " accounts " , column : " target_account_id " , on_delete : :nullify
add_foreign_key " account_migrations " , " accounts " , on_delete : :cascade
2017-10-10 13:12:17 +02:00
add_foreign_key " account_moderation_notes " , " accounts "
2017-10-07 20:26:43 +02:00
add_foreign_key " account_moderation_notes " , " accounts " , column : " target_account_id "
2020-07-07 15:26:51 +02:00
add_foreign_key " account_notes " , " accounts " , column : " target_account_id " , on_delete : :cascade
add_foreign_key " account_notes " , " accounts " , on_delete : :cascade
2018-08-09 09:56:53 +02:00
add_foreign_key " account_pins " , " accounts " , column : " target_account_id " , on_delete : :cascade
add_foreign_key " account_pins " , " accounts " , on_delete : :cascade
2018-11-19 00:43:52 +01:00
add_foreign_key " account_stats " , " accounts " , on_delete : :cascade
2021-08-09 23:11:50 +02:00
add_foreign_key " account_statuses_cleanup_policies " , " accounts " , on_delete : :cascade
Add moderation warnings (#9519)
* Add moderation warnings
Replace individual routes for disabling, silencing, and suspending
a user, as well as the report update route, with a unified account
action controller that allows you to select an action (none,
disable, silence, suspend) as well as whether it should generate an
e-mail notification with optional custom text. That notification,
with the optional custom text, is saved as a warning.
Additionally, there are warning presets you can configure to save
time when performing the above.
* Use Account#local_username_and_domain
2018-12-22 20:02:09 +01:00
add_foreign_key " account_warnings " , " accounts " , column : " target_account_id " , on_delete : :cascade
add_foreign_key " account_warnings " , " accounts " , on_delete : :nullify
2022-01-17 09:41:33 +01:00
add_foreign_key " account_warnings " , " reports " , on_delete : :cascade
2017-11-18 19:39:02 +01:00
add_foreign_key " accounts " , " accounts " , column : " moved_to_account_id " , on_delete : :nullify
2017-11-24 02:05:53 +01:00
add_foreign_key " admin_action_logs " , " accounts " , on_delete : :cascade
2020-01-23 22:00:13 +01:00
add_foreign_key " announcement_mutes " , " accounts " , on_delete : :cascade
add_foreign_key " announcement_mutes " , " announcements " , on_delete : :cascade
add_foreign_key " announcement_reactions " , " accounts " , on_delete : :cascade
add_foreign_key " announcement_reactions " , " announcements " , on_delete : :cascade
add_foreign_key " announcement_reactions " , " custom_emojis " , on_delete : :cascade
2022-02-14 21:27:53 +01:00
add_foreign_key " appeals " , " account_warnings " , on_delete : :cascade
add_foreign_key " appeals " , " accounts " , column : " approved_by_account_id " , on_delete : :nullify
add_foreign_key " appeals " , " accounts " , column : " rejected_by_account_id " , on_delete : :nullify
add_foreign_key " appeals " , " accounts " , on_delete : :cascade
2018-02-28 06:54:55 +01:00
add_foreign_key " backups " , " users " , on_delete : :nullify
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
add_foreign_key " blocks " , " accounts " , column : " target_account_id " , name : " fk_9571bfabc1 " , on_delete : :cascade
add_foreign_key " blocks " , " accounts " , name : " fk_4269e03e65 " , on_delete : :cascade
2018-04-10 22:13:15 +02:00
add_foreign_key " bookmarks " , " accounts " , on_delete : :cascade
add_foreign_key " bookmarks " , " statuses " , on_delete : :cascade
2023-05-02 12:08:48 +02:00
add_foreign_key " bulk_import_rows " , " bulk_imports " , on_delete : :cascade
add_foreign_key " bulk_imports " , " accounts " , on_delete : :cascade
2021-06-30 06:13:55 +02:00
add_foreign_key " canonical_email_blocks " , " accounts " , column : " reference_account_id " , on_delete : :cascade
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
add_foreign_key " conversation_mutes " , " accounts " , name : " fk_225b4212bb " , on_delete : :cascade
2017-06-05 13:24:00 +02:00
add_foreign_key " conversation_mutes " , " conversations " , on_delete : :cascade
Revamp post filtering system (#18058)
* Add model for custom filter keywords
* Use CustomFilterKeyword internally
Does not change the API
* Fix /filters/edit and /filters/new
* Add migration tests
* Remove whole_word column from custom_filters (covered by custom_filter_keywords)
* Redesign /filters
Instead of a list, present a card that displays more information and handles
multiple keywords per filter.
* Redesign /filters/new and /filters/edit to add and remove keywords
This adds a new gem dependency: cocoon, as well as a npm dependency:
cocoon-js-vanilla. Those are used to easily populate and remove form fields
from the user interface when manipulating multiple keyword filters at once.
* Add /api/v2/filters to edit filter with multiple keywords
Entities:
- `Filter`: `id`, `title`, `filter_action` (either `hide` or `warn`), `context`
`keywords`
- `FilterKeyword`: `id`, `keyword`, `whole_word`
API endpoits:
- `GET /api/v2/filters` to list filters (including keywords)
- `POST /api/v2/filters` to create a new filter
`keywords_attributes` can also be passed to create keywords in one request
- `GET /api/v2/filters/:id` to read a particular filter
- `PUT /api/v2/filters/:id` to update a new filter
`keywords_attributes` can also be passed to edit, delete or add keywords in
one request
- `DELETE /api/v2/filters/:id` to delete a particular filter
- `GET /api/v2/filters/:id/keywords` to list keywords for a filter
- `POST /api/v2/filters/:filter_id/keywords/:id` to add a new keyword to a
filter
- `GET /api/v2/filter_keywords/:id` to read a particular keyword
- `PUT /api/v2/filter_keywords/:id` to edit a particular keyword
- `DELETE /api/v2/filter_keywords/:id` to delete a particular keyword
* Change from `irreversible` boolean to `action` enum
* Remove irrelevent `irreversible_must_be_within_context` check
* Fix /filters/new and /filters/edit with update for filter_action
* Fix Rubocop/Codeclimate complaining about task names
* Refactor FeedManager#phrase_filtered?
This moves regexp building and filter caching to the `CustomFilter` class.
This does not change the functional behavior yet, but this changes how the
cache is built, doing per-custom_filter regexps so that filters can be matched
independently, while still offering caching.
* Perform server-side filtering and output result in REST API
* Fix numerous filters_changed events being sent when editing multiple keywords at once
* Add some tests
* Use the new API in the WebUI
- use client-side logic for filters we have fetched rules for.
This is so that filter changes can be retroactively applied without
reloading the UI.
- use server-side logic for filters we haven't fetched rules for yet
(e.g. network error, or initial timeline loading)
* Minor optimizations and refactoring
* Perform server-side filtering on the streaming server
* Change the wording of filter action labels
* Fix issues pointed out by linter
* Change design of “Show anyway” link in accordence to review comments
* Drop “irreversible” filtering behavior
* Move /api/v2/filter_keywords to /api/v1/filters/keywords
* Rename `filter_results` attribute to `filtered`
* Rename REST::LegacyFilterSerializer to REST::V1::FilterSerializer
* Fix systemChannelId value in streaming server
* Simplify code by removing client-side filtering code
The simplifcation comes at a cost though: filters aren't retroactively
applied anymore.
2022-06-28 09:42:13 +02:00
add_foreign_key " custom_filter_keywords " , " custom_filters " , on_delete : :cascade
2022-08-25 04:27:47 +02:00
add_foreign_key " custom_filter_statuses " , " custom_filters " , on_delete : :cascade
add_foreign_key " custom_filter_statuses " , " statuses " , on_delete : :cascade
2018-06-29 15:34:36 +02:00
add_foreign_key " custom_filters " , " accounts " , on_delete : :cascade
2020-06-02 19:24:53 +02:00
add_foreign_key " devices " , " accounts " , on_delete : :cascade
add_foreign_key " devices " , " oauth_access_tokens " , column : " access_token_id " , on_delete : :cascade
2020-03-12 22:35:20 +01:00
add_foreign_key " email_domain_blocks " , " email_domain_blocks " , column : " parent_id " , on_delete : :cascade
2020-06-02 19:24:53 +02:00
add_foreign_key " encrypted_messages " , " accounts " , column : " from_account_id " , on_delete : :cascade
add_foreign_key " encrypted_messages " , " devices " , on_delete : :cascade
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
add_foreign_key " favourites " , " accounts " , name : " fk_5eb6c2b873 " , on_delete : :cascade
add_foreign_key " favourites " , " statuses " , name : " fk_b0e856845e " , on_delete : :cascade
2019-02-04 04:25:59 +01:00
add_foreign_key " featured_tags " , " accounts " , on_delete : :cascade
add_foreign_key " featured_tags " , " tags " , on_delete : :cascade
2021-04-12 12:37:14 +02:00
add_foreign_key " follow_recommendation_suppressions " , " accounts " , on_delete : :cascade
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
add_foreign_key " follow_requests " , " accounts " , column : " target_account_id " , name : " fk_9291ec025d " , on_delete : :cascade
add_foreign_key " follow_requests " , " accounts " , name : " fk_76d644b0e7 " , on_delete : :cascade
add_foreign_key " follows " , " accounts " , column : " target_account_id " , name : " fk_745ca29eac " , on_delete : :cascade
add_foreign_key " follows " , " accounts " , name : " fk_32ed1b5560 " , on_delete : :cascade
2018-11-27 13:56:25 +01:00
add_foreign_key " identities " , " users " , name : " fk_bea040f377 " , on_delete : :cascade
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
add_foreign_key " imports " , " accounts " , name : " fk_6db1b6e408 " , on_delete : :cascade
2017-11-27 16:07:59 +01:00
add_foreign_key " invites " , " users " , on_delete : :cascade
2017-11-18 00:16:48 +01:00
add_foreign_key " list_accounts " , " accounts " , on_delete : :cascade
2023-05-02 14:40:36 +02:00
add_foreign_key " list_accounts " , " follow_requests " , on_delete : :cascade
2017-11-18 00:16:48 +01:00
add_foreign_key " list_accounts " , " follows " , on_delete : :cascade
add_foreign_key " list_accounts " , " lists " , on_delete : :cascade
add_foreign_key " lists " , " accounts " , on_delete : :cascade
2021-06-21 17:07:30 +02:00
add_foreign_key " login_activities " , " users " , on_delete : :cascade
2019-09-06 13:55:51 +02:00
add_foreign_key " markers " , " users " , on_delete : :cascade
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
add_foreign_key " media_attachments " , " accounts " , name : " fk_96dd81e81b " , on_delete : :nullify
2019-01-05 12:43:28 +01:00
add_foreign_key " media_attachments " , " scheduled_statuses " , on_delete : :nullify
2017-06-05 13:24:00 +02:00
add_foreign_key " media_attachments " , " statuses " , on_delete : :nullify
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
add_foreign_key " mentions " , " accounts " , name : " fk_970d43f9d1 " , on_delete : :cascade
2017-06-05 13:24:00 +02:00
add_foreign_key " mentions " , " statuses " , on_delete : :cascade
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
add_foreign_key " mutes " , " accounts " , column : " target_account_id " , name : " fk_eecff219ea " , on_delete : :cascade
add_foreign_key " mutes " , " accounts " , name : " fk_b8d8daf315 " , on_delete : :cascade
add_foreign_key " notifications " , " accounts " , column : " from_account_id " , name : " fk_fbd6b0bf9e " , on_delete : :cascade
add_foreign_key " notifications " , " accounts " , name : " fk_c141c8ee55 " , on_delete : :cascade
add_foreign_key " oauth_access_grants " , " oauth_applications " , column : " application_id " , name : " fk_34d54b0a33 " , on_delete : :cascade
add_foreign_key " oauth_access_grants " , " users " , column : " resource_owner_id " , name : " fk_63b044929b " , on_delete : :cascade
add_foreign_key " oauth_access_tokens " , " oauth_applications " , column : " application_id " , name : " fk_f5fc4c1ee3 " , on_delete : :cascade
add_foreign_key " oauth_access_tokens " , " users " , column : " resource_owner_id " , name : " fk_e84df68546 " , on_delete : :cascade
add_foreign_key " oauth_applications " , " users " , column : " owner_id " , name : " fk_b0988c7c0a " , on_delete : :cascade
2020-06-02 19:24:53 +02:00
add_foreign_key " one_time_keys " , " devices " , on_delete : :cascade
2019-03-03 22:18:23 +01:00
add_foreign_key " poll_votes " , " accounts " , on_delete : :cascade
add_foreign_key " poll_votes " , " polls " , on_delete : :cascade
add_foreign_key " polls " , " accounts " , on_delete : :cascade
add_foreign_key " polls " , " statuses " , on_delete : :cascade
2022-10-08 16:45:40 +02:00
add_foreign_key " preview_card_trends " , " preview_cards " , on_delete : :cascade
2018-04-02 22:04:14 +02:00
add_foreign_key " report_notes " , " accounts " , on_delete : :cascade
add_foreign_key " report_notes " , " reports " , on_delete : :cascade
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
add_foreign_key " reports " , " accounts " , column : " action_taken_by_account_id " , name : " fk_bca45b75fd " , on_delete : :nullify
2018-04-02 22:04:14 +02:00
add_foreign_key " reports " , " accounts " , column : " assigned_account_id " , on_delete : :nullify
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
add_foreign_key " reports " , " accounts " , column : " target_account_id " , name : " fk_eb37af34f0 " , on_delete : :cascade
add_foreign_key " reports " , " accounts " , name : " fk_4b81f7522c " , on_delete : :cascade
2019-01-05 12:43:28 +01:00
add_foreign_key " scheduled_statuses " , " accounts " , on_delete : :cascade
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
add_foreign_key " session_activations " , " oauth_access_tokens " , column : " access_token_id " , name : " fk_957e5bda89 " , on_delete : :cascade
add_foreign_key " session_activations " , " users " , name : " fk_e5fda67334 " , on_delete : :cascade
2022-01-19 22:37:27 +01:00
add_foreign_key " status_edits " , " accounts " , on_delete : :nullify
add_foreign_key " status_edits " , " statuses " , on_delete : :cascade
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
add_foreign_key " status_pins " , " accounts " , name : " fk_d4cb435b62 " , on_delete : :cascade
2017-08-25 01:41:18 +02:00
add_foreign_key " status_pins " , " statuses " , on_delete : :cascade
2018-08-14 19:19:32 +02:00
add_foreign_key " status_stats " , " statuses " , on_delete : :cascade
2022-10-08 16:45:40 +02:00
add_foreign_key " status_trends " , " accounts " , on_delete : :cascade
add_foreign_key " status_trends " , " statuses " , on_delete : :cascade
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
add_foreign_key " statuses " , " accounts " , column : " in_reply_to_account_id " , name : " fk_c7fa917661 " , on_delete : :nullify
add_foreign_key " statuses " , " accounts " , name : " fk_9bda1543f7 " , on_delete : :cascade
2017-06-05 13:24:00 +02:00
add_foreign_key " statuses " , " statuses " , column : " in_reply_to_id " , on_delete : :nullify
2017-02-17 02:33:10 +01:00
add_foreign_key " statuses " , " statuses " , column : " reblog_of_id " , on_delete : :cascade
2017-06-05 13:24:00 +02:00
add_foreign_key " statuses_tags " , " statuses " , on_delete : :cascade
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
add_foreign_key " statuses_tags " , " tags " , name : " fk_3081861e21 " , on_delete : :cascade
2022-07-17 13:49:29 +02:00
add_foreign_key " tag_follows " , " accounts " , on_delete : :cascade
add_foreign_key " tag_follows " , " tags " , on_delete : :cascade
2019-01-18 15:56:55 +01:00
add_foreign_key " tombstones " , " accounts " , on_delete : :cascade
2019-04-09 16:06:30 +02:00
add_foreign_key " user_invite_requests " , " users " , on_delete : :cascade
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
add_foreign_key " users " , " accounts " , name : " fk_50500f500d " , on_delete : :cascade
2017-11-27 16:07:59 +01:00
add_foreign_key " users " , " invites " , on_delete : :nullify
2018-12-24 19:12:38 +01:00
add_foreign_key " users " , " oauth_applications " , column : " created_by_application_id " , on_delete : :nullify
2022-07-05 02:41:40 +02:00
add_foreign_key " users " , " user_roles " , column : " role_id " , on_delete : :nullify
2018-05-11 11:49:12 +02:00
add_foreign_key " web_push_subscriptions " , " oauth_access_tokens " , column : " access_token_id " , on_delete : :cascade
add_foreign_key " web_push_subscriptions " , " users " , on_delete : :cascade
Make IdsToBigints (mostly!) non-blocking (#5088)
* Make IdsToBigints (mostly!) non-blocking
This pulls in GitLab's MigrationHelpers, which include code to make
column changes in ways that Postgres can do without locking. In general,
this involves creating a new column, adding an index and any foreign
keys as appropriate, adding a trigger to keep it populated alongside
the old column, and then progressively copying data over to the new
column, before removing the old column and replacing it with the new
one.
A few changes to GitLab's MigrationHelpers were necessary:
* Some changes were made to remove dependencies on other GitLab code.
* We explicitly wait for index creation before forging ahead on column
replacements.
* We use different temporary column names, to avoid running into index
name length limits.
* We rename the generated indices back to what they "should" be after
replacing columns.
* We rename the generated foreign keys to use the new column names when
we had to create them. (This allows the migration to be rolled back
without incident.)
# Big Scary Warning
There are two things here that may trip up large instances:
1. The change for tables' "id" columns is not concurrent. In
particular, the stream_entries table may be big, and does not
concurrently migrate its id column. (On the other hand, x_id type
columns are all concurrent.)
2. This migration will take a long time to run, *but it should not
lock tables during that time* (with the exception of the "id"
columns as described above). That means this should probably be run
in `screen` or some other session that can be run for a long time.
Notably, the migration will take *longer* than it would without
these changes, but the website will still be responsive during that
time.
These changes were tested on a relatively large statuses table (256k
entries), and the service remained responsive during the migration.
Migrations both forward and backward were tested.
* Rubocop fixes
* MigrationHelpers: Support ID columns in some cases
This doesn't work in cases where the ID column is referred to as a
foreign key by another table.
* MigrationHelpers: support foreign keys for ID cols
Note that this does not yet support foreign keys on non-primary-key
columns, but Mastodon also doesn't yet have any that we've needed to
migrate.
This means we can perform fully "concurrent" migrations to change ID
column types, and the IdsToBigints migration can happen with effectively
no downtime. (A few operations require a transaction, such as renaming
columns or deleting them, but these transactions should not block for
noticeable amounts of time.)
The algorithm for generating foreign key names has changed with this,
and therefore all of those changed in schema.rb.
* Provide status, allow for interruptions
The MigrationHelpers now allow restarting the rename of a column if it
was interrupted, by removing the old "new column" and re-starting the
process.
Along with this, they now provide status updates on the changes which
are happening, as well as indications about when the changes can be
safely interrupted (when there are at least 10 seconds estimated to be
left before copying data is complete).
The IdsToBigints migration now also sorts the columns it migrates by
size, starting with the largest tables. This should provide
administrators a worst-case scenario estimate for the length of
migrations: each successive change will get faster, giving admins a
chance to abort early on if they need to run the migration later. The
idea is that this does not force them to try to time interruptions
between smaller migrations.
* Fix column sorting in IdsToBigints
Not a significant change, but it impacts the order of columns in the
database and db/schema.rb.
* Actually pause before IdsToBigints
2017-10-02 21:28:59 +02:00
add_foreign_key " web_settings " , " users " , name : " fk_11910667b2 " , on_delete : :cascade
Add WebAuthn as an alternative 2FA method (#14466)
* feat: add possibility of adding WebAuthn security keys to use as 2FA
This adds a basic UI for enabling WebAuthn 2FA. We did a little refactor
to the Settings page for editing the 2FA methods – now it will list the
methods that are available to the user (TOTP and WebAuthn) and from
there they'll be able to add or remove any of them.
Also, it's worth mentioning that for enabling WebAuthn it's required to
have TOTP enabled, so the first time that you go to the 2FA Settings
page, you'll be asked to set it up.
This work was inspired by the one donde by Github in their platform, and
despite it could be approached in different ways, we decided to go with
this one given that we feel that this gives a great UX.
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
* feat: add request for WebAuthn as second factor at login if enabled
This commits adds the feature for using WebAuthn as a second factor for
login when enabled.
If users have WebAuthn enabled, now a page requesting for the use of a
WebAuthn credential for log in will appear, although a link redirecting
to the old page for logging in using a two-factor code will also be
present.
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
* feat: add possibility of deleting WebAuthn Credentials
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
* feat: disable WebAuthn when an Admin disables 2FA for a user
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
* feat: remove ability to disable TOTP leaving only WebAuthn as 2FA
Following examples form other platforms like Github, we decided to make
Webauthn 2FA secondary to 2FA with TOTP, so that we removed the
possibility of removing TOTP authentication only, leaving users with
just WEbAuthn as 2FA. Instead, users will have to click on 'Disable 2FA'
in order to remove second factor auth.
The reason for WebAuthn being secondary to TOPT is that in that way,
users will still be able to log in using their code from their phone's
application if they don't have their security keys with them – or maybe
even lost them.
* We had to change a little the flow for setting up TOTP, given that now
it's possible to setting up again if you already had TOTP, in order to
let users modify their authenticator app – given that now it's not
possible for them to disable TOTP and set it up again with another
authenticator app.
So, basically, now instead of storing the new `otp_secret` in the
user, we store it in the session until the process of set up is
finished.
This was because, as it was before, when users clicked on 'Edit' in
the new two-factor methods lists page, but then went back without
finishing the flow, their `otp_secret` had been changed therefore
invalidating their previous authenticator app, making them unable to
log in again using TOTP.
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
* refactor: fix eslint errors
The PR build was failing given that linting returning some errors.
This commit attempts to fix them.
* refactor: normalize i18n translations
The build was failing given that i18n translations files were not
normalized.
This commits fixes that.
* refactor: avoid having the webauthn gem locked to a specific version
* refactor: use symbols for routes without '/'
* refactor: avoid sending webauthn disabled email when 2FA is disabled
When an admins disable 2FA for users, we were sending two mails
to them, one notifying that 2FA was disabled and the other to notify
that WebAuthn was disabled.
As the second one is redundant since the first email includes it, we can
remove it and send just one email to users.
* refactor: avoid creating new env variable for webauthn_origin config
* refactor: improve flash error messages for webauthn pages
Co-authored-by: Facundo Padula <facundo.padula@cedarcode.com>
2020-08-24 16:46:27 +02:00
add_foreign_key " webauthn_credentials " , " users "
2020-12-14 09:06:34 +01:00
create_view " instances " , materialized : true , sql_definition : <<-SQL
WITH domain_counts ( domain , accounts_count ) AS (
SELECT accounts . domain ,
count ( * ) AS accounts_count
FROM accounts
WHERE ( accounts . domain IS NOT NULL )
GROUP BY accounts . domain
)
SELECT domain_counts . domain ,
domain_counts . accounts_count
FROM domain_counts
UNION
SELECT domain_blocks . domain ,
COALESCE ( domain_counts . accounts_count , ( 0 ) :: bigint ) AS accounts_count
FROM ( domain_blocks
LEFT JOIN domain_counts ON ( ( ( domain_counts . domain ) :: text = ( domain_blocks . domain ) :: text ) ) )
UNION
SELECT domain_allows . domain ,
COALESCE ( domain_counts . accounts_count , ( 0 ) :: bigint ) AS accounts_count
FROM ( domain_allows
LEFT JOIN domain_counts ON ( ( ( domain_counts . domain ) :: text = ( domain_allows . domain ) :: text ) ) ) ;
SQL
2023-05-25 09:27:16 +02:00
add_index " instances " , " reverse(('.'::text || (domain)::text)), domain " , name : " index_instances_on_reverse_domain "
2020-12-14 09:06:34 +01:00
add_index " instances " , [ " domain " ] , name : " index_instances_on_domain " , unique : true
2022-01-16 13:23:50 +01:00
create_view " user_ips " , sql_definition : <<-SQL
SELECT t0 . user_id ,
t0 . ip ,
max ( t0 . used_at ) AS used_at
FROM ( SELECT users . id AS user_id ,
users . sign_up_ip AS ip ,
users . created_at AS used_at
FROM users
WHERE ( users . sign_up_ip IS NOT NULL )
UNION ALL
SELECT session_activations . user_id ,
session_activations . ip ,
session_activations . updated_at
FROM session_activations
UNION ALL
SELECT login_activities . user_id ,
login_activities . ip ,
login_activities . created_at
FROM login_activities
WHERE ( login_activities . success = true ) ) t0
GROUP BY t0 . user_id , t0 . ip ;
SQL
2021-04-12 12:37:14 +02:00
create_view " account_summaries " , materialized : true , sql_definition : <<-SQL
SELECT accounts . id AS account_id ,
mode ( ) WITHIN GROUP ( ORDER BY t0 . language ) AS language ,
mode ( ) WITHIN GROUP ( ORDER BY t0 . sensitive ) AS sensitive
FROM ( accounts
CROSS JOIN LATERAL ( SELECT statuses . account_id ,
statuses . language ,
statuses . sensitive
FROM statuses
2021-12-13 23:21:14 +01:00
WHERE ( ( statuses . account_id = accounts . id ) AND ( statuses . deleted_at IS NULL ) AND ( statuses . reblog_of_id IS NULL ) )
2021-04-12 12:37:14 +02:00
ORDER BY statuses . id DESC
LIMIT 20 ) t0 )
WHERE ( ( accounts . suspended_at IS NULL ) AND ( accounts . silenced_at IS NULL ) AND ( accounts . moved_to_account_id IS NULL ) AND ( accounts . discoverable = true ) AND ( accounts . locked = false ) )
GROUP BY accounts . id ;
SQL
add_index " account_summaries " , [ " account_id " ] , name : " index_account_summaries_on_account_id " , unique : true
2023-08-18 18:20:55 +02:00
create_view " global_follow_recommendations " , materialized : true , sql_definition : <<-SQL
2021-04-12 12:37:14 +02:00
SELECT t0 . account_id ,
sum ( t0 . rank ) AS rank ,
array_agg ( t0 . reason ) AS reason
2021-05-05 22:04:52 +02:00
FROM ( SELECT account_summaries . account_id ,
2021-04-12 12:37:14 +02:00
( ( count ( follows . id ) ) :: numeric / ( 1 . 0 + ( count ( follows . id ) ) :: numeric ) ) AS rank ,
'most_followed' :: text AS reason
2023-08-18 18:20:55 +02:00
FROM ( ( follows
2021-05-05 22:04:52 +02:00
JOIN account_summaries ON ( ( account_summaries . account_id = follows . target_account_id ) ) )
2021-04-12 12:37:14 +02:00
JOIN users ON ( ( users . account_id = follows . account_id ) ) )
2023-08-18 18:20:55 +02:00
WHERE ( ( users . current_sign_in_at > = ( now ( ) - 'P30D' :: interval ) ) AND ( account_summaries . sensitive = false ) AND ( NOT ( EXISTS ( SELECT 1
FROM follow_recommendation_suppressions
WHERE ( follow_recommendation_suppressions . account_id = follows . target_account_id ) ) ) ) )
2021-05-05 22:04:52 +02:00
GROUP BY account_summaries . account_id
2021-04-12 12:37:14 +02:00
HAVING ( count ( follows . id ) > = 5 )
UNION ALL
2021-05-05 22:04:52 +02:00
SELECT account_summaries . account_id ,
2021-04-12 12:37:14 +02:00
( sum ( ( status_stats . reblogs_count + status_stats . favourites_count ) ) / ( 1 . 0 + sum ( ( status_stats . reblogs_count + status_stats . favourites_count ) ) ) ) AS rank ,
'most_interactions' :: text AS reason
2023-08-18 18:20:55 +02:00
FROM ( ( status_stats
2021-04-12 12:37:14 +02:00
JOIN statuses ON ( ( statuses . id = status_stats . status_id ) ) )
2021-05-05 22:04:52 +02:00
JOIN account_summaries ON ( ( account_summaries . account_id = statuses . account_id ) ) )
2023-08-18 18:20:55 +02:00
WHERE ( ( statuses . id > = ( ( ( date_part ( 'epoch' :: text , ( now ( ) - 'P30D' :: interval ) ) * ( 1000 ) :: double precision ) ) :: bigint << 16 ) ) AND ( account_summaries . sensitive = false ) AND ( NOT ( EXISTS ( SELECT 1
FROM follow_recommendation_suppressions
WHERE ( follow_recommendation_suppressions . account_id = statuses . account_id ) ) ) ) )
2021-05-05 22:04:52 +02:00
GROUP BY account_summaries . account_id
2021-04-12 12:37:14 +02:00
HAVING ( sum ( ( status_stats . reblogs_count + status_stats . favourites_count ) ) > = ( 5 ) :: numeric ) ) t0
GROUP BY t0 . account_id
ORDER BY ( sum ( t0 . rank ) ) DESC ;
SQL
2023-08-18 18:20:55 +02:00
add_index " global_follow_recommendations " , [ " account_id " ] , name : " index_global_follow_recommendations_on_account_id " , unique : true
2022-04-28 17:15:11 +02:00
2017-04-27 14:42:22 +02:00
end