-1

我正在尝试在本地计算机上配置 hadoop-0.23.8。我正在运行,但有以下例外。你们知道我做错了什么吗

我正在使用以下系统变量

export HADOOP_OPTS="-Djava.security.krb5.realm=OX.AC.UK -Djava.security.krb5.kdc=kdc0.ox.ac.uk:kdc1.ox.ac.uk"

我也在使用 java 6 并设置了以下变量 HADOOP_PREFIX、HADOOP_HDFS_HOME 和 YARN_HOME。您能否帮助正确配置它。

192.168.1.13: namenode running as process 10522. Stop it first.
hduser@192.168.1.13: starting datanode, logging to /usr/local/hadoop-0.23.8/logs/hadoop-
hduser-datanode-MacBook-Air.local.out
hduser@192.168.1.13: 2013-06-19 00:30:30.556 java[11171:1603] Unable to load realm info from SCDynamicStore
hduser@192.168.1.13: 2013-06-19 00:30:30.558 java[11171:1603] Unable to load realm info from SCDynamicStore
Starting secondary namenodes []
hduser@192.168.1.13: starting secondarynamenode, logging to /usr/local/hadoop-0.23.8/logs/hadoop-hduser-secondarynamenode-MacBook-Air.local.out
hduser@192.168.1.13: Exception in thread "main" java.lang.IllegalArgumentException: Target address cannot be null.
hduser@192.168.1.13:    at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:179)
hduser@192.168.1.13:    at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:154)
hduser@192.168.1.13:    at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:143)
hduser@192.168.1.13:    at org.apache.hadoop.hdfs.DFSUtil.getNameServiceId(DFSUtil.java:771)
hduser@192.168.1.13:    at org.apache.hadoop.hdfs.DFSUtil.getSecondaryNameServiceId(DFSUtil.java:738)
hduser@192.168.1.13:    at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.<init>(SecondaryNameNode.java:186)
hduser@192.168.1.13:    at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.main(SecondaryNameNode.java:691)
4

1 回答 1

1

所以我的一个配置文件中有一个小错字,对此感到抱歉,但我发现了一些配置 hadoop 0.23.8 的重要信息。在 etc/hadoop 文件夹中创建一个名为 hadoop-env.sh 的文件,我们将 JAVA_HOME、HADOOP_OPTS 放入其中。ETC

于 2013-06-22T16:13:28.467 回答