caution
This runbook is an Experiment. For complete, production-ready documentation, see the disaster recovery documentation.

Disaster Recovery (Geo) promotion runbooks

Tier: Premium, Ultimate Offering: Self-managed

Geo planned failover for a multi-node configuration

Component Configuration
PostgreSQL Managed by the Linux package
Geo site Multi-node
Secondaries One

This runbook guides you through a planned failover of a multi-node Geo site with one secondary. The following 2000 user reference architecture is assumed:

graph TD subgraph main[Geo deployment] subgraph Primary[Primary site, multi-node] Node_1[Rails node 1] Node_2[Rails node 2] Node_3[PostgreSQL node] Node_4[Gitaly node] Node_5[Redis node] Node_6[Monitoring node] end subgraph Secondary[Secondary site, multi-node] Node_7[Rails node 1] Node_8[Rails node 2] Node_9[PostgreSQL node] Node_10[Gitaly node] Node_11[Redis node] Node_12[Monitoring node] end end

The load balancer node and optional NFS server are omitted for clarity.

This guide results in the following:

  1. An offline primary.
  2. A promoted secondary that is now the new primary.

What is not covered:

  1. Re-adding the old primary as a secondary.
  2. Adding a new secondary.

Preparation

note
Before following any of those steps, make sure you have root access to the secondary to promote it, because there isn’t provided an automated way to promote a Geo replica and perform a failover.

On the secondary site:

  1. On the left sidebar, at the bottom, select Admin Area.
  2. Select Geo > Sites to see its status. Replicated objects (shown in green) should be close to 100%, and there should be no failures (shown in red). If a large proportion of objects aren’t replicated (shown in gray), consider giving the site more time to complete.

    Replication status

If any objects are failing to replicate, this should be investigated before scheduling the maintenance window. After a planned failover, anything that failed to replicate is lost.

You can use the Geo status API to review failed objects and the reasons for failure. A common cause of replication failures is data that is missing on the primary site - you can resolve these failures by restoring the data from backup, or removing references to the missing data.

The maintenance window doesn’t end until Geo replication and verification is completely finished. To keep the window as short as possible, you should ensure these processes are close to 100% as possible during active use.

If the secondary site is still replicating data from the primary site, follow these steps to avoid unnecessary data loss:

  1. Enable maintenance mode on the primary site, and make sure to stop any background jobs.
  2. Finish replicating and verifying all data:

    caution
    Not all data is automatically replicated. Read more about what is excluded.
    1. If you are manually replicating any data not managed by Geo, trigger the final replication process now.
    2. On the primary site:
      1. On the left sidebar, at the bottom, select Admin Area.
      2. On the left sidebar, select Monitoring > Background Jobs.
      3. On the Sidekiq dashboard, select Queues, and wait for all queues except those with geo in the name to drop to 0. These queues contain work that has been submitted by your users; failing over before it is completed, causes the work to be lost.
      4. On the left sidebar, select Geo > Sites and wait for the following conditions to be true of the secondary site you are failing over to:

        • All replication meters reach 100% replicated, 0% failures.
        • All verification meters reach 100% verified, 0% failures.
        • Database replication lag is 0 ms.
        • The Geo log cursor is up to date (0 events behind).
    3. On the secondary site:
      1. On the left sidebar, at the bottom, select Admin Area.
      2. On the left sidebar, select Monitoring > Background Jobs.
      3. On the Sidekiq dashboard, select Queues, and wait for all the geo queues to drop to 0 queued and 0 running jobs.
      4. Run an integrity check to verify the integrity of CI artifacts, LFS objects, and uploads in file storage.

    At this point, your secondary site contains an up-to-date copy of everything the primary site has, meaning nothing is lost when you fail over.

  3. In this final step, you must permanently disable the primary site.

    caution
    When the primary site goes offline, there may be data saved on the primary site that has not been replicated to the secondary site. This data should be treated as lost if you proceed.
    note
    If you plan to update the primary domain DNS record, you may wish to lower the TTL now to speed up propagation.

    When performing a failover, we want to avoid a split-brain situation where writes can occur in two different GitLab instances. So to prepare for the failover, you must disable the primary site:

    • If you have SSH access to the primary site, stop and disable GitLab:

      sudo gitlab-ctl stop
      

      Prevent GitLab from starting up again if the server unexpectedly reboots:

      sudo systemctl disable gitlab-runsvdir
      
      note
      (CentOS only) In CentOS 6 or older, it is challenging to prevent GitLab from being started if the machine reboots isn’t available (see issue 3058). It may be safest to uninstall the GitLab package completely with sudo yum remove gitlab-ee.
      note
      (Ubuntu 14.04 LTS) If you are using an older version of Ubuntu or any other distribution based on the Upstart init system, you can prevent GitLab from starting if the machine reboots as root with initctl stop gitlab-runsvvdir && echo 'manual' > /etc/init/gitlab-runsvdir.override && initctl reload-configuration.
    • If you do not have SSH access to the primary site, take the machine offline and prevent it from rebooting. As there are many ways you may prefer to accomplish this, we avoid a single recommendation. You may have to:

      • Reconfigure the load balancers.
      • Change DNS records (for example, point the primary DNS record to the secondary site to stop using the primary site).
      • Stop the virtual servers.
      • Block traffic through a firewall.
      • Revoke object storage permissions from the primary site.
      • Physically disconnect a machine.

Promoting the secondary site running GitLab 14.5 and later

  1. SSH to every Sidekiq, PostgreSQL, and Gitaly node in the secondary site and run one of the following commands:

    • To promote the secondary site to primary:

      sudo gitlab-ctl geo promote
      
    • To promote the secondary site to primary without any further confirmation:

      sudo gitlab-ctl geo promote --force
      
  2. SSH into each Rails node on your secondary site and run one of the following commands:

    • To promote the secondary site to primary:

      sudo gitlab-ctl geo promote
      
    • To promote the secondary site to primary without any further confirmation:

      sudo gitlab-ctl geo promote --force
      
  3. Verify you can connect to the newly promoted primary site using the URL used previously for the secondary site.
  4. If successful, the secondary site is now promoted to the primary site.

Promoting the secondary site running GitLab 14.4 and earlier

caution
The gitlab-ctl promote-to-primary-node and gitlab-ctl promoted-db commands are deprecated in GitLab 14.5 and later, and removed in GitLab 15.0. Use gitlab-ctl geo promote instead.
note
A new secondary should not be added at this time. If you want to add a new secondary, do this after you have completed the entire process of promoting the secondary to the primary.
caution
If you encounter an ActiveRecord::RecordInvalid: Validation failed: Name has already been taken error during this process, read the troubleshooting advice.

The gitlab-ctl promote-to-primary-node command cannot be used in conjunction with multiple servers, as it can only perform changes on a secondary with only a single machine. Instead, you must do this manually.

caution
In GitLab 13.2 and 13.3, promoting a secondary site to a primary while the secondary is paused fails. Do not pause replication before promoting a secondary. If the site is paused, be sure to resume before promoting. This issue has been fixed in GitLab 13.4 and later.
caution
If the secondary site has been paused, this performs a point-in-time recovery to the last known state. Data that was created on the primary while the secondary was paused is lost.
  1. SSH in to the PostgreSQL node in the secondary and promote PostgreSQL separately:

    sudo gitlab-ctl promote-db
    
  2. Edit /etc/gitlab/gitlab.rb on every machine in the secondary to reflect its new status as primary by removing any lines that enabled the geo_secondary_role:

    ## In pre-11.5 documentation, the role was enabled as follows. Remove this line.
    geo_secondary_role['enable'] = true
    
    ## In 11.5+ documentation, the role was enabled as follows. Remove this line.
    roles ['geo_secondary_role']
    

    After making these changes, reconfigure GitLab each machine so the changes take effect.

  3. Promote the secondary to primary. SSH into a single Rails node server and execute:

    sudo gitlab-rake geo:set_secondary_as_primary
    
  4. Verify you can connect to the newly promoted primary using the URL used previously for the secondary.

  5. Success! The secondary has now been promoted to primary.

Next steps

To regain geographic redundancy as quickly as possible, you should add a new secondary site. To do that, you can re-add the old primary as a new secondary and bring it back online.