Elasticsearch OpenSearch Flood Stage Disk Watermark

By Opster Team

Updated: Jun 27, 2023

| 3 min read

Quick links

Overview

There are various “watermark” thresholds on your OpenSearch cluster. As the disk fills up on a node, the first threshold to be crossed will be the “low disk watermark”. The second threshold will then be the “high disk watermark threshold”. Finally, the “disk flood stage” will be reached. Once this threshold is passed, the cluster will then block writing to ALL indices that have one shard (primary or replica) on the node which has passed the watermark. Reads (searches) will still be possible.

How to resolve the issue

Passing this threshold will cause loss of data in your application and you should not delay in taking action.

How to resolve the flood stage disk watermark issue in OpenSearch?

Here are possible actions you can take to resolve this issue:
– Delete old indices
– Remove documents from existing indices
– Increase disk space on the node
– Add new data nodes to the cluster

Although you may be reluctant to delete data, in a logging system it is often better to delete old indices (which you may be able to restore from a snapshot later if available) than to lose new data.  However, this decision will depend upon the architecture of your system and the queueing mechanisms you have available.

Check the disk space on each node

You can see the space you have available on each node by running:

GET _nodes/stats/fs

Check if the cluster is rebalancing

If the high level watermark has been passed, then OpenSearch should start rebalancing to other nodes which are still below the low watermark.  You can check to see if any rebalancing is going on by calling:

GET _cluster/health/

If you think that your cluster should be rebalancing shards to other nodes but it is not, there are probably some other cluster allocation rules which are preventing this from happening.   The most likely causes are:

  • The other nodes are already above the low disk-watermark
  • There are cluster allocation rules which govern the distribution of shards between nodes and conflict with the rebalancing requirements. (eg. zone awareness allocation)
  • There are already too many rebalancing operations in progress
  • The other nodes already contain replicas of the shards that could be rebalanced

Check the cluster settings

You can see the settings you have applied with this command:

GET _cluster/settings

If they are not appropriate, you can modify them using a command such as below:

PUT _cluster/settings
{
  "transient": {
   
    "cluster.routing.allocation.disk.watermark.low": "85%",
    "cluster.routing.allocation.disk.watermark.high": "90%",
    "cluster.routing.allocation.disk.watermark.flood_stage": "95%",
    "cluster.info.update.interval": "1m"
  }
}

How to prevent reaching this stage

There are various mechanisms that allow you to automatically delete stale data.

How to automatically delete stale data in OpenSearch:

  1. Apply ILM (Index Lifecycle Management)

    Using ILM you can get OpenSearch to automatically delete an index when your current index size reaches a given age. 

  2. Use date based indices

    If your application uses date based indices, then it is easy to delete old indices using a script or a tool such as OpenSearch curator.

  3. Use snapshots to store data offline

    It may be appropriate to store snapshotted data offline and restore it in the event that the archived data needs to be reviewed or studied.

  4. Automate / simplify process to add new data nodes

    Use automation tools such as terraform to automate the addition of new nodes to the cluster.  If this is not possible, at the very least ensure you have a clearly documented process to create new nodes, add TLS certificates and configuration and bring them into the OpenSearch cluster in a short and predictable time frame.

How helpful was this guide?

We are sorry that this post was not useful for you!

Let us improve this post!

Tell us how we can improve this post?