Garage is a stateful clustered application, where all nodes are communicating together and share data structures. It makes upgrade more difficult than stateless applications so you must be more careful when upgrading. On a new version release, there is 2 possibilities:
- protocols and data structures remained the same ➡️ this is a minor upgrade
- protocols or data structures changed ➡️ this is a major upgrade
You can quickly now what type of update you will have to operate by looking at the version identifier: when we require our users to do a major upgrade, we will always bump the first nonzero component of the version identifier (e.g. from v0.7.2 to v0.8.0). Conversely, for versions that only require a minor upgrade, the first nonzero component will always stay the same (e.g. from v0.8.0 to v0.8.1).
Major upgrades are designed to be run only between contiguous versions. Example: migrations from v0.7.1 to v0.8.0 and from v0.7.0 to v0.8.2 are supported but migrations from v0.6.0 to v0.8.0 are not supported.
Prometheus metric provides
an overview for which Garage versions are currently in use within a cluster.
Minor upgrades do not imply cluster downtime. Before upgrading, you should still read the changelog and ideally test your deployment on a staging cluster before.
When you are ready, start by checking the health of your cluster.
You can force some checks with
garage repair, we recommend at least running
garage repair --all-nodes --yes tables which is very quick to run (less than a minute).
You will see that the command correctly terminated in the logs of your daemon, or using
garage worker list (the repair workers should be in the
Finally, you can simply upgrade nodes one by one. For each node: stop it, install the new binary, edit the configuration if needed, restart it.
Major upgrades can be done with minimal downtime with a bit of preparation, but the simplest way is usually to put the cluster offline for the duration of the migration. Before upgrading, you must read the changelog and you must test your deployment on a staging cluster before.
We write guides for each major upgrade, they are stored under the "Working Documents" section of this documentation.
Major upgrades with full downtime
From a high level perspective, a major upgrade looks like this:
- Disable API access (for instance in your reverse proxy, or by commenting the corresponding section in your Garage configuration file and restarting Garage)
- Check that your cluster is idle
- Make sure the health of your cluster is good (see
- Stop the whole cluster
- Back up the metadata folder of all your nodes, so that you will be able to restore it if the upgrade fails (data blocks being immutable, they should not be impacted)
- Install the new binary, update the configuration
- Start the whole cluster
- If needed, run the corresponding migration from
- Make sure the health of your cluster is good
- Enable API access (reverse step 1)
- Monitor your cluster while load comes back, check that all your applications are happy with this new version
Major upgarades with minimal downtime
There is only one operation that has to be coordinated cluster-wide: the switch of one version of the internal RPC protocol to the next. This means that an upgrade with very limited downtime can simply be performed from one major version to the next by restarting all nodes simultaneously in the new version. The downtime will simply be the time required for all nodes to stop and start again, which should be less than a minute. If all nodes fail to stop and restart simultaneously, some nodes might be temporarily shut out from the cluster as nodes using different RPC protocol versions are prevented to talk to one another.
The entire procedure would look something like this:
Make sure the health of your cluster is good (see
Take each node offline individually to back up its metadata folder, bring them back online once the backup is done. You can do all of the nodes in a single zone at once as that won't impact global cluster availability. Do not try to make a backup of the metadata folder of a running node.
Prepare your binaries and configuration files for the new Garage version
Restart all nodes simultaneously in the new version
If any specific migration procedure is required, it is usually in one of the two cases:
- It can be run on online nodes after the new version has started, during regular cluster operation.
- it has to be run offline, in which case you will have to again take all nodes offline one after the other to run the repair
For this last step, please refer to the specific documentation pertaining to the version upgrade you are doing.