HDFS恢复到DC / OS中的机器重启 [英] HDFS resiliency to machine restarts in DC/OS

查看:185
本文介绍了HDFS恢复到DC / OS中的机器重启的处理方法,对大家解决问题具有一定的参考价值,需要的朋友们下面随着小编来一起学习吧!

问题描述

我在10台Core OS机器(3个主节点,7个代理节点)的DCOS集群上从Universe安装了HDFS。我的HA HDFS配置有2个名称节点,3个日志节点和5个数据节点。现在,我的问题是。 HDFS不能恢复机器重启吗?如果我重新启动安装了数据节点的计算机,则数据节点将被重建为其他计算机的镜像(仅在从DC / OS UI重新启动HDFS服务后)。在日志节点或名称节点处重新启动的情况下,节点将被标记为丢失,并且从不重建。 解决方案

最终,在DC / OS的Universe HDFS软件包的错误版本中发现了此问题。然而,一个全新的DC / OS HDFS包将在未来几周内在Universe上发布。



https://dcos-community.slack.com / archives / data-services / p1485801481001734


I have installed HDFS from universe on my DCOS cluster of 10 Core OS machines (3 master nodes, 7 agent nodes). My HA HDFS config has 2 name nodes, 3 journal nodes and 5 data nodes. Now, my question is. Shouldn’t the HDFS be resilient to machine restarts? If I restart a machine where a data node is installed the data node gets rebuilt as a mirror of the others (only after restarting the HDFS service from the DC/OS UI). In the case of a restart where a journal node or a name node is, the nodes will be just marked as lost and never rebuilt.

解决方案

Eventually the problem was found in a buggy version of the universe HDFS package for DC/OS. However, a completely new HDFS package for DC/OS will be released on Universe in the next few weeks.

https://dcos-community.slack.com/archives/data-services/p1485717889001709

https://dcos-community.slack.com/archives/data-services/p1485801481001734

这篇关于HDFS恢复到DC / OS中的机器重启的文章就介绍到这了,希望我们推荐的答案对大家有所帮助,也希望大家多多支持IT屋!

查看全文
登录 关闭
扫码关注1秒登录
发送“验证码”获取 | 15天全站免登陆