Failed to lock all shards for index timed out after 30 seconds – How to solve related issues

Failed to lock all shards for index timed out after 30 seconds – How to solve related issues

Opster Team

Dec-19, Version: 1.7-8.0

Before you begin reading this guide, we recommend you run Elasticsearch Error Check-Up which can resolve issues that cause many errors.

This guide will help you check for common problems that cause the log ” Failed to lock all shards for index timed out after 30 seconds ” to appear. It’s important to understand the issues related to the log, so to get started, read the general overview on common issues and tips related to the Elasticsearch concepts: indices.

Advanced users might want to skip right to the common problems section in each concept or try running the Check-Up which analyses ES to pinpoint the cause of many errors and provides suitable actionable recommendations how to resolve them (free tool that requires no installation).

Log Context

Log “Failed to lock all shards for index – timed out after 30 seconds” classname is IndicesClusterStateService.java
We extracted the following from Elasticsearch source code for those seeking an in-depth context :

                             // to the master. If we can't acquire the locks here immediately there might be a shard of this index still
                            // holding on to the lock due to a "currently canceled recovery" or so. The shard will delete itself BEFORE the
                            // lock is released so it's guaranteed to be deleted by the time we get the lock
                            indicesService.processPendingDeletes(index; indexSettings; new TimeValue(30; TimeUnit.MINUTES));
                        } catch (ShardLockObtainFailedException exc) {
                            logger.warn("[{}] failed to lock all shards for index - timed out after 30 seconds"; index);
                        } catch (InterruptedException e) {
                            logger.warn("[{}] failed to lock all shards for index - interrupted"; index);
                        }
                    }
                });

 

Run the Check-Up to get customized recommendations like this:

error

The high disk watermark threshold is about to be reached in specific nodes

error-img

Description

There are various “watermark” thresholds on each Elasticsearch cluster. When the high disk watermark threshold has been exceeded, it means disk space is running out. The node will…

error-img

Recommendations

Based on your specific ES deployment you should…

1

X-PUT curl -H [a customized code snippet to resolve the issue]

Optimize Elasticsearch Performance

Run The Tool