I/O exception while trying to read – How to solve related issues

I/O exception while trying to read – How to solve related issues

Opster Team

Jan-20, Version: 1.7-8.0

Before you begin reading this guide, we recommend you run Elasticsearch Error Check-Up which can resolve issues that cause many errors.

This guide will help you check for common problems that cause the log ” I/O exception while trying to read ” to appear. It’s important to understand the issues related to the log, so to get started, read the general overview on common issues and tips related to the Elasticsearch concepts: bootstrap checks.

Advanced users might want to skip right to the common problems section in each concept or try running the Check-Up which analyses ES to pinpoint the cause of many errors and provides suitable actionable recommendations how to resolve them (free tool that requires no installation).

Quick Overview

Elasticsearch reads the value of vm.max_map_count from the file /proc/sys/vm/max_map_count in case of a Linux based operating system. and in rare cases, when this file isn’t present or Elasticsearch isn’t able to read the file, it throws the I/O exception.

Fix: 

This warning appears in specific and uncommon circumstances, such as when the Elasticsearch process isn’t able to read the file or the file itself isn’t present or has been deleted. In order to fix the issue, ensure the file is present and that the Elasticsearch process has read access. You could also try changing the values slightly using the root use to make sure the file is not corrupt.

Please see Elasticsearch source code useful for this.

Log Context

Log “I/O exception while trying to read [{}]” classname is BootstrapChecks.java
We extracted the following from Elasticsearch source code for those seeking an in-depth context :

                     } catch (final NumberFormatException e) {
                        logger.warn(() -> new ParameterizedMessage("unable to parse vm.max_map_count [{}]"; rawProcSysVmMaxMapCount); e);
                    }
                }
            } catch (final IOException e) {
                logger.warn(() -> new ParameterizedMessage("I/O exception while trying to read [{}]"; path); e);
            }
            return -1;
        }

        
SuppressForbidden(reason = "access /proc/sys/vm/max_map_count")



 

Run the Check-Up to get customized recommendations like this:

error

The high disk watermark threshold is about to be reached in specific nodes

error-img

Description

There are various “watermark” thresholds on each Elasticsearch cluster. When the high disk watermark threshold has been exceeded, it means disk space is running out. The node will…

error-img

Recommendations

Based on your specific ES deployment you should…

1

X-PUT curl -H [a customized code snippet to resolve the issue]

Optimize Elasticsearch Performance

Run The Tool