我对 Hadoop Cluster 数据节点故障转移有一些疑问:
1: What happen the link is down between the namenode and a datanode
(or between 2 datanodes) when the hadoop cluster is processing some data?
Does Hadoop cluster have any OOTB to recover this problem?
2: What happen one datanode is down when the hadoop cluster is processing
some data?
另外,另一个问题是关于 hadoop 集群的硬件配置。假设我们将使用我们的 hadoop 集群每天处理 100GB 的日志文件,我们需要设置多少个数据节点?对于每个数据节点的硬件配置(例如 CPU、RAM、硬盘)?