Max open files – How to solve this Elasticsearch error

Opster Team

Aug-23, Version: 2.3-2.3

Briefly, this error occurs when the number of files that Elasticsearch is trying to open exceeds the maximum limit set by the operating system. This can lead to performance issues or even data loss. To resolve this issue, you can increase the limit of open files in the operating system settings. Alternatively, you can reduce the number of shards in your Elasticsearch cluster, as each shard opens multiple files. Lastly, ensure that your application closes files when they are no longer needed to prevent unnecessary usage.

This guide will help you check for common problems that cause the log ” max_open_files [{}] ” to appear. To understand the issues related to this log, read the explanation below about the following Elasticsearch concepts: bootstrap.

Log Context

Log “max_open_files [{}]” classname is
We extracted the following from Elasticsearch source code for those seeking an in-depth context :

             PidFile.create(environment.pidFile(); true);

        if (System.getProperty("es.max-open-files"; "false").equals("true")) {
            ESLogger logger = Loggers.getLogger(Bootstrap.class);
  "max_open_files [{}]"; ProcessProbe.getInstance().getMaxFileDescriptorCount());

        // warn if running using the client VM
        if (JvmInfo.jvmInfo().getVmName().toLowerCase(Locale.ROOT).contains("client")) {
            ESLogger logger = Loggers.getLogger(Bootstrap.class);


How helpful was this guide?

We are sorry that this post was not useful for you!

Let us improve this post!

Tell us how we can improve this post?