Local gateway settingsedit

The local gateway stores the cluster state and shard data across full cluster restarts.

The following static settings, which must be set on every master node, control how long a freshly elected master should wait before it tries to recover the cluster state and the cluster’s data:

gateway.expected_nodes
[7.7.0] Deprecated in 7.7.0. This setting will be removed in 8.0. You should use gateway.expected_data_nodes instead. The number of (data or master) nodes that are expected to be in the cluster. Recovery of local shards will start as soon as the expected number of nodes have joined the cluster. Defaults to 0
gateway.expected_master_nodes
[7.7.0] Deprecated in 7.7.0. This setting will be removed in 8.0. You should use gateway.expected_data_nodes instead. The number of master nodes that are expected to be in the cluster. Recovery of local shards will start as soon as the expected number of master nodes have joined the cluster. Defaults to 0
gateway.expected_data_nodes
The number of data nodes that are expected to be in the cluster. Recovery of local shards will start as soon as the expected number of data nodes have joined the cluster. Defaults to 0
gateway.recover_after_time
If the expected number of nodes is not achieved, the recovery process waits for the configured amount of time before trying to recover regardless. Defaults to 5m if one of the expected_nodes settings is configured.

Once the recover_after_time duration has timed out, recovery will start as long as the following conditions are met:

gateway.recover_after_nodes
[7.7.0] Deprecated in 7.7.0. This setting will be removed in 8.0. You should use gateway.recover_after_data_nodes instead. Recover as long as this many data or master nodes have joined the cluster.
gateway.recover_after_master_nodes
[7.7.0] Deprecated in 7.7.0. This setting will be removed in 8.0. You should use gateway.recover_after_data_nodes instead. Recover as long as this many master nodes have joined the cluster.
gateway.recover_after_data_nodes
Recover as long as this many data nodes have joined the cluster.

These settings only take effect on a full cluster restart.

Dangling indicesedit

When a node joins the cluster, if it finds any shards stored in its local data directory that do not already exist in the cluster, it will consider those shards to belong to a "dangling" index. You can list, import or delete dangling indices using the Dangling indices API.

The API cannot offer any guarantees as to whether the imported data truly represents the latest state of the data when the index was still part of the cluster.