Backup and Restore
Once you have your node up and running, it's time to prepare for disaster recovery. Should your machine ever have a catastrophic failure due to either hardware or software issues, or even a case of natural disaster, it's best to be prepared for such a situation by making a backup.
When running, a complete node installation along with the database can grow to be multiple gigabytes in size. Having to back up and restore such a large volume of data can be expensive, complicated and time-consuming. Luckily, there is a better way.
Instead of having to back up and restore everything, we need to back up only what is essential, that is, those files that cannot be reconstructed because they are unique to your node. For AvalancheGo node, unique files are those that identify your node on the network, in other words, files that define your NodeID.
Even if your node is a validator on the network and has multiple delegations on it, you don't need to worry about backing up anything else, because the validation and delegation transactions are also stored on the blockchain and will be restored during bootstrapping, along with the rest of the blockchain data.
The installation itself can be easily recreated by installing the node on a new machine, and all the remaining gigabytes of blockchain data can be easily recreated by the process of bootstrapping, which copies the data over from other network peers. However, if you would like to speed up the process, see the Database Backup and Restore section
NodeID
If more than one running nodes share the same NodeID, the communications from other nodes in the Avalanche network to this NodeID will be random to one of these nodes. If this NodeID is of a validator, it will dramatically impact the uptime calculation of the validator which will very likely disqualify the validator from receiving the staking rewards. Please make sure only one node with the same NodeID run at one time.
NodeID is a unique identifier that differentiates your node from all the other peers on the network. It's a string formatted like NodeID-5mb46qkSBj81k9g9e4VFjGGSbaaSLFRzD
. You can look up the technical background of how the NodeID is constructed here. In essence, NodeID is defined by two files:
staker.crt
staker.key
NodePOP is this node's BLS key and proof of possession. Nodes must register a BLS key to act as a validator on the Primary Network. Your node's POP is logged on startup and is accessible over this endpoint.
publicKey
is the 48 byte hex representation of the BLS key.proofOfPossession
is the 96 byte hex representation of the BLS signature.
NodePOP is defined by the signer.key
file.
In the default installation, they can be found in the working directory, specifically in ~/.avalanchego/staking/
. All we need to do to recreate the node on another machine is to run a new installation with those same three files.
If staker.key
and staker.crt
are removed from a node, which is restarted afterwards, they will be recreated and a new node ID will be assigned.
If the signer.key
is regenerated, the node will lose its previous BLS identity, which includes its public key and proof of possession. This change means that the node's former identity on the network will no longer be recognized, affecting its ability to participate in the consensus mechanism as before. Consequently, the node may lose its established reputation and any associated staking rewards.
If you have users defined in the keystore of your node, then you need to back up and restore those as well. Keystore API has methods that can be used to export and import user keys. Note that Keystore API is used by developers only and not intended for use in production nodes. If you don't know what a keystore API is and have not used it, you don't need to worry about it.
Backup
To back up your node, we need to store staker.crt
and staker.key
files somewhere safe and private, preferably to a different computer, to your private To back up your node, we need to store staker.crt
, staker.key
and signer.key
files somewhere safe and private, preferably to a different computer.
If someone gets a hold of your staker files, they still cannot get to your funds, as they are controlled by the wallet private keys, not by the node. But, they could re-create your node somewhere else, and depending on the circumstances make you lose the staking rewards. So make sure your staker files are secure.
If someone gains access to your signer.key
, they could potentially sign transactions on behalf of your node, which might disrupt the operations and integrity of your node on the network.
Let's get the files off the machine running the node.
From Local Node
If you're running the node locally, on your desktop computer, just navigate to where the files are and copy them somewhere safe.
On a default Linux installation, the path to them will be /home/USERNAME/.avalanchego/staking/
, where USERNAME
needs to be replaced with the actual username running the node. Select and copy the files from there to a backup location. You don't need to stop the node to do that.
From Remote Node Using scp
scp
is a 'secure copy' command line program, available built-in on Linux and MacOS computers. There is also a Windows version, pscp
, as part of the PuTTY package. If using pscp
, in the following commands replace each usage of scp
with pscp -scp
.
To copy the files from the node, you will need to be able to remotely log into the machine. You can use account password, but the secure and recommended way is to use the SSH keys. The procedure for acquiring and setting up SSH keys is highly dependent on your cloud provider and machine configuration. You can refer to our Amazon Web Services and Microsoft Azure setup guides for those providers. Other providers will have similar procedures.
When you have means of remote login into the machine, you can copy the files over with the following command:
This assumes the username on the machine is ubuntu
, replace with correct username in both places if it is different. Also, replace PUBLICIP
with the actual public IP of the machine. If scp
doesn't automatically use your downloaded SSH key, you can point to it manually:
Once executed, this command will create avalanche_backup
directory and place those three files in it. You need to store them somewhere safe.
Restore
To restore your node from a backup, we need to do the reverse: restore staker.key
, staker.crt
and signer.key
from the backup to the working directory of the new node.
First, we need to do the usual installation of the node. This will create a new NodeID, a new BLS key and a new BLS signature, which we need to replace. When the node is installed correctly, log into the machine where the node is running and stop it:
We're ready to restore the node.
To Local Node
If you're running the node locally, just copy the staker.key
, staker.crt
and signer.key
files from the backup location into the working directory, which on the default Linux installation will be /home/USERNAME/.avalanchego/staking/
. Replace USERNAME
with the actual username used to run the node.
To Remote Node Using scp
Again, the process is just the reverse operation. Using scp
we need to copy the staker.key
, staker.crt
and signer.key
files from the backup location into the remote working directory. Assuming the backed up files are located in the directory where the above backup procedure placed them:
Or if you need to specify the path to the SSH key:
And again, replace ubuntu
with correct username if different, and PUBLICIP
with the actual public IP of the machine running the node, as well as the path to the SSH key if used.
Restart the Node and Verify
Once the files have been replaced, log into the machine and start the node using:
You can now check that the node is restored with the correct NodeID and NodePOP by issuing the getNodeID API call in the same console you ran the previous command:
You should see your original NodeID and NodePOP (BLS key and BLS signature). Restore process is done.
Database
Normally, when starting a new node, you can just bootstrap from scratch. However, there are situations when you may prefer to reuse an existing database (ex: preserve keystore records, reduce sync time).
This tutorial will walk you through compressing your node's DB and moving it to another computer using zip
and scp
.
Database Backup
First, make sure to stop AvalancheGo, run:
You must stop the Avalanche node before you back up the database otherwise data could become corrupted.
Once the node is stopped, you can zip
the database directory to reduce the size of the backup and speed up the transfer using scp
:
Note: It may take > 30 minutes to zip the node's DB.
Next, you can transfer the backup to another machine:
This assumes the username on the machine is ubuntu
, replace with correct username in both places if it is different. Also, replace PUBLICIP
with the actual public IP of the machine. If scp
doesn't automatically use your downloaded SSH key, you can point to it manually:
Once executed, this command will create avalanche_db_backup.zip
directory in you home directory.
Database Restore
This tutorial assumes you have already completed "Database Backup" and have a backup at ~/avalanche_db_backup.zip.
First, we need to do the usual installation of the node. When the node is installed correctly, log into the machine where the node is running and stop it:
You must stop the Avalanche node before you restore the database otherwise data could become corrupted.
We're ready to restore the database. First, let's move the DB on the existing node (you can remove this old DB later if the restore was successful):
Next, we'll unzip the backup we moved from another node (this will place the unzipped files in ~/.avalanchego/db
when the command is run in the home directory):
After the database has been restored on a new node, use this command to start the node:
Node should now be running from the database on the new instance. To check that everything is in order and that node is not bootstrapping from scratch (which would indicate a problem), use:
The node should be catching up to the network and fetching a small number of blocks before resuming normal operation (all the ones produced from the time when the node was stopped before the backup).
Once the backup has been restored and is working as expected, the zip can be deleted:
Database Direct Copy
You may be in a situation where you don't have enough disk space to create the archive containing the whole database, so you cannot complete the backup process as described previously.
In that case, you can still migrate your database to a new computer, by using a different approach: direct copy
. Instead of creating the archive, moving the archive and unpacking it, we can do all of that on the fly.
To do so, you will need ssh
access from the destination machine (where you want the database to end up) to the source machine (where the database currently is). Setting up ssh
is the same as explained for scp
earlier in the document.
Same as shown previously, you need to stop the node (on both machines):
You must stop the Avalanche node before you back up the database otherwise data could become corrupted.
Then, on the destination machine, change to a directory where you would like to the put the database files, enter the following command:
Make sure to replace the correct path to the key, and correct IP of the source machine. This will compress the database, but instead of writing it to a file it will pipe it over ssh
directly to destination machine, where it will be decompressed and written to disk. The process can take a long time, make sure it completes before continuing.
After copying is done, all you need to do now is move the database to the correct location on the destination machine. Assuming there is a default AvalancheGo node installation, we remove the old database and replace it with the new one:
You can now start the node on the destination machine:
Node should now be running from the copied database. To check that everything is in order and that node is not bootstrapping from scratch (which would indicate a problem), use:
The node should be catching up to the network and fetching a small number of blocks before resuming normal operation (all the ones produced from the time when the node was stopped before the backup).
Summary
Essential part of securing your node is the backup that enables full and painless restoration of your node. Following this tutorial you can rest easy knowing that should you ever find yourself in a situation where you need to restore your node from scratch, you can easily and quickly do so.
If you have any problems following this tutorial, comments you want to share with us or just want to chat, you can reach us on our Discord server.
Last updated on