It is time to say goodbye: This version of Elastic Cloud Enterprise has reached end-of-life (EOL) and is no longer supported.
The documentation for this version is no longer being maintained. If you are running this version, we strongly advise you to upgrade. For the latest information, see the current release documentation.
You might need to move Elasticsearch nodes, Kibana instances, and other components of the Elastic Stack between allocators from time to time for a number of reasons:
- To prepare for removing the allocator role from the first host on which you installed Elastic Cloud Enterprise.
- To avoid downtime during maintenance: You can create a new allocator, move all deployments from an existing allocator to the new one, and then deal with the allocator that needs maintenance.
- To make room on an allocator: You can move some smaller deployments to another allocator if you need additional room for a larger one on an allocator.
- To move deployments after a failure: When host failures happen, you can move all deployments from the affected allocator to a healthy allocator quickly before spending any time on fixing the failure.
When you move all nodes from an existing allocator to the new one, ECE migrates the data to new nodes. The migration can take some time, especially when deployments contain large amounts of data and have a heavy workload. Is your deployment under a heavy workload? You might need to stop routing requests first.
Before you beginedit
Before you move the nodes and instances that are part of a deployment, you need to make sure that you have sufficient capacity on another allocator. For example: If you have a deployment with a single 32 GB Elasticsearch node and a 4 GB Kibana instance, the allocator that you are moving the deployment to needs to have at least 36 GB of capacity. Note that moving nodes does not actually move the same node onto a different allocator. Under the covers, Elastic Cloud Enterprise creates a new node and then migrates the data for you.
Elastic Cloud Enterprise will adhere to the high availability configuration when moving nodes, so make sure you have the additional capacity available in the relevant availability zone. For example: If you selected to deploy your cluster accross 3 availability zones, nodes can only move to an allocator in the same availability zone as the failed allocator. This is meant to ensure that the cluster can tolerate the failure of 2 availability zones.
If you followed our recommendation and tagged your allocators to indicate what allocators you want components of the Elastic Stack to run on, the spare capacity you plan to use must be available on an allocator with the same tags. If you did not tag your allocators and edit the default instance configurations, ECE will move nodes and instances to wherever there is space.
When you move all nodes from an existing allocator to the new one, ECE migrates the data to new nodes. The migration can take some time, especially when clusters contain large amounts of data and have a heavy workload. Is your cluster under a heavy workload? You might need to stop routing requests first.
To move nodes from one allocator to another one:
- Log into the Cloud UI.
- From the Platform menu, select Allocators.
Review the list of all allocators that are part of this installation and look for allocators that are unhealthy or find the allocator that you want to free up.
Narrow the list by name, ID, or choose from several other filters. To further define the list, use a combination of filters.
- Recommended: Put the allocator into maintenance mode before continuing.
- Click the name of an unhealthy allocator and then select Move Nodes from the menu.
- Select the nodes you want, then click Move Nodes.
To customize how you would like to move the nodes, click Customize settings, select your options, then click Move nodes.
Review Customize Settings before proceeding to move nodes.
- Gracefully move data
(Default) Gracefully move the data from the instances we’re about to remove from the cluster before stopping them. Never disable this setting at the same time as enabling
override_failsafeon a non-Highly Available cluster since it can result in data loss.
- Skip snapshot
- If an allocator has failed or is otherwise unhealthy, select this option to move the nodes but disable the snapshot attempt. As this can perform potentially destructive actions on the deployment, do not use this option on a healthy allocator unless you are an advanced user.
- Restore snapshot to latest success
- Restore the cluster to the last successful snapshot. Recommended for single-node clusters hosted on unhealthy allocators. Any data indexed after the last snapshot was taken is lost.
- Extended maintenance
- Keep new instances in maintenance mode until a snapshot has been restored. If not enabled, new instances remain in maintenance mode only until they can join a cluster.
- Set target allocators
- Request that instances be moved to the specified allocators. If no allocators are specified, or those specified are unsuitable for the instances being moved, then any suitable healthy allocator can be used.
- Create new containers for all nodes in the cluster.
- Set Timeout
- On by default.
If you did not enable maintenance mode, set a target allocator under the advanced options when moving nodes to make sure the nodes do not end up on the same allocator again. By default, moving a node moves it to any allocator that has enough capacity.
- Repeat step 6 for each of the node types until no nodes remain on the allocator.
- Optionally, once the nodes have been moved, Delete Allocator.