How to set up replicators¶
Replicators sync instances across LXD cluster links. This is useful for active-passive disaster recovery, where a leader (active) cluster handles all workloads while a standby cluster remains ready to take over if the leader fails.
LXD supports this strategy using project replicators over a cluster link.
Prerequisites¶
Before setting up replicators:
Two LXD clusters must be initialized. We will call them “leader” and “standby”.
You need sufficient permissions on both clusters to establish links and manage projects.
A cluster link must be established between the two clusters.
Network connectivity must exist between the clusters.
Prepare authentication¶
Replicators communicate over cluster links, so the linked cluster identities must be granted the permissions they need on the replicated project. Configure these permissions using authentication groups and Manage permissions.
For project replication, the cluster-link identity on each cluster typically needs at least these permissions:
operatoron the replicated project, so it can perform instance replication work in that projectcan_editon the replicated project, so replica project configuration can be validated and updated as part of the workflow
For example, if the replicated project is called myproject, you can prepare an authentication group on each cluster before creating the cluster links:
lxc auth group create replicators
lxc auth group permission add replicators project myproject operator
lxc auth group permission add replicators project myproject can_edit
Then create the cluster links with that authentication group, as described in How to create cluster links.
Configure projects for replication¶
Both clusters need a project configured with the replica.cluster and replica.mode settings.
On the leader cluster, create a project and configure it to link to the standby cluster:
lxc project create <project_name> -c replica.cluster=<standby_cluster_link_name>
On the standby cluster, create a project with the same name and configure it to link to the leader cluster:
lxc project create <project_name> -c replica.cluster=<leader_cluster_link_name>
On the standby cluster, set the project’s replica mode to
standby. This prevents new instances from being created in the project directly; existing instances can still be managed until the project is promoted toleaderduring a failover.lxc project set <project_name> replica.mode=standby
On the leader cluster, set the project’s replica mode to
leader.lxc project set <project_name> replica.mode=leader
Setting the leader
Setting replica.mode=leader on the leader cluster requires the target project on the standby cluster to already have replica.mode=standby set.
This ensures that new instances are not created on a standby cluster between replicator runs.
Create a replicator¶
After configuring the projects on both clusters, create a replicator on the leader cluster. The cluster configuration key is required and must be set to the name of an existing cluster link.
lxc replicator create <replicator_name> cluster=<standby_cluster_link_name> --project <project_name>
For example:
lxc replicator create my-replicator cluster=lxd-standby --project myproject
You can also create a replicator with a schedule and snapshot options:
lxc replicator create my-replicator cluster=lxd-standby schedule="@daily" snapshot=true --project myproject
See Replicator configuration for all available configuration options.
Run a replicator¶
To manually trigger a replicator run, use the following command on the leader cluster:
lxc replicator run <replicator_name>
This syncs all instances in the source project to the standby cluster.
To schedule replication automatically, set the schedule configuration key with a cron expression:
lxc replicator set <replicator_name> schedule="0 0 * * *"
Snapshot before replication¶
When you set snapshot=true on a replicator, LXD creates a point-in-time snapshot of each source
instance before performing the incremental refresh to the standby cluster. This gives you a
consistent rollback point on the source in case anything goes wrong during replication.
Snapshot naming and expiry are controlled entirely by the instance’s own configuration (for example
snapshots.pattern and snapshots.expiry), or by the profile applied to the instance. The
replicator does not impose its own naming scheme.
If an instance already has a snapshots.schedule set at the instance or profile level, the
replicator skips creating a new snapshot and reuses the most recent existing snapshot for the
incremental refresh instead.
When snapshot is not set (or set to false), no new snapshot is created. If a previous snapshot
already exists on the instance, it is reused naturally by the incremental refresh.
Note
Snapshots created by replication accumulate over time. Use snapshots.expiry on the instance or
profile to automatically prune them, or delete them manually with lxc snapshot delete.
Next steps¶
Once replicators are running, see How to manage replicators to view, configure, or delete replicators, and How to perform disaster recovery with replicators to fail over to the standby cluster if the leader becomes unavailable.