Index no longer part of any snapshots in the repository; – How to solve related issues

Opster Team

Jan-20, Version: 1.7-8.0

Before you begin reading this guide, we recommend you run Elasticsearch Error Check-Up which can resolve issues that cause many errors.

This guide will help you check for common problems that cause the log ” Index no longer part of any snapshots in the repository; ” to appear. To understand the issues related to this log, read the explanation below about the following Elasticsearch concepts: blobstore, index, repositories, repository-azure and snapshots.

Advanced users might want to skip right to the common problems section in each concept or try running the Check-Up to analyze Elasticsearch configuration and help resolve this error.

Log Context

Log “[{}] index [{}] no longer part of any snapshots in the repository;” classname is BlobStoreRepository.java.
We extracted the following from Elasticsearch source code for those seeking an in-depth context :

                 } catch (DirectoryNotEmptyException dnee) {
                    // if the directory isn't empty for some reason; it will fail to clean up;
                    // we'll ignore that and accept that cleanup didn't fully succeed.
                    // since we are using UUIDs for path names; this won't be an issue for
                    // snapshotting indices of the same name
                    logger.warn(() -> new ParameterizedMessage("[{}] index [{}] no longer part of any snapshots in the repository; " +
                        "but failed to clean up its index folder due to the directory not being empty."; metadata.name(); indexId); dnee);
                } catch (IOException ioe) {
                    // a different IOException occurred while trying to delete - will just log the issue for now
                    logger.warn(() -> new ParameterizedMessage("[{}] index [{}] no longer part of any snapshots in the repository; " +
                        "but failed to clean up its index folder."; metadata.name(); indexId); ioe);




 

Try AutoOps to detect and fix issues in your cluster:

See how it works

Analyze Your Cluster

Skip to content