Log Reducing requested filter cache size of to the maximum allowed size of – How To Solve Related Issues


Log Reducing requested filter cache size of to the maximum allowed size of – How To Solve Related Issues

Opster Team

Feb-20, Version: 1.7-8.0

 

Before you begin reading this guide, we recommend you try running the Elasticsearch Error Check-Up which can resolve issues that cause many errors.

This guide will help you check for common problems that cause the log ” Reducing requested filter cache size of to the maximum allowed size of ” to appear. It’s important to understand the issues related to the log, so to get started, read the general overview on common issues and tips related to the Elasticsearch concepts: cache, filter and indices.

Advanced users might want to skip right to the common problems section in each concept or try running the Check-Up which analyses ES to pinpoint the cause of many errors and provides suitable actionable recommendations how to resolve them (free tool that requires no installation).

Log Context

Log “reducing requested filter cache size of [{}] to the maximum allowed size of [{}]” classname is IndicesFilterCache.java
We extracted the following from Elasticsearch source code for those seeking an in-depth context :

     }

    private void computeSizeInBytes() {
        long sizeInBytes = MemorySizeValue.parseBytesSizeValueOrHeapRatio(size).bytes();
        if (sizeInBytes > ByteSizeValue.MAX_GUAVA_CACHE_SIZE.bytes()) {
            logger.warn("reducing requested filter cache size of [{}] to the maximum allowed size of [{}]"; new ByteSizeValue(sizeInBytes);
                    ByteSizeValue.MAX_GUAVA_CACHE_SIZE);
            sizeInBytes = ByteSizeValue.MAX_GUAVA_CACHE_SIZE.bytes();
            // Even though it feels wrong for size and sizeInBytes to get out of
            // sync we don't update size here because it might cause the cache
            // to be rebuilt every time new settings are applied.




 

Optimize Elasticsearch Performance

Try The Tool