FIELDDATA数据太大 [英] FIELDDATA Data is too large

查看:471
本文介绍了FIELDDATA数据太大的处理方法,对大家解决问题具有一定的参考价值,需要的朋友们下面随着小编来一起学习吧!

问题描述

我打开kibana并进行搜索,我得到了碎片失败的错误。我查看了elasticsearch.log文件,我看到这个错误:

I open kibana and do a search and i get the error where shards failed. I looked in the elasticsearch.log file and I saw this error:

org.elasticsearch.common.breaker.CircuitBreakingException: [FIELDDATA] Data too large, data for [@timestamp] would be larger than limit of [622775500/593.9mb]



<有没有办法增加593.9mb的限制?

Is there any way to increase that limit of 593.9mb?

推荐答案

您可以尝试增加 fielddata断路器 elasticsearch.yml 中限制为75%(默认值为60% code>配置文件并重新启动您的集群:

You can try to increase the fielddata circuit breaker limit to 75% (default is 60%) in your elasticsearch.yml config file and restart your cluster:

indices.breaker.fielddata.limit: 75%


curl -XPUT localhost:9200/_cluster/settings -d '{
  "persistent" : {
    "indices.breaker.fielddata.limit" : "40%" 
  }
}'

尝试。

这篇关于FIELDDATA数据太大的文章就介绍到这了,希望我们推荐的答案对大家有所帮助,也希望大家多多支持IT屋!

查看全文
登录 关闭
扫码关注1秒登录
发送“验证码”获取 | 15天全站免登陆