我正在尝试在本地计算机上配置 hadoop-0.23.8。我正在运行,但有以下例外。你们知道我做错了什么吗
我正在使用以下系统变量
export HADOOP_OPTS="-Djava.security.krb5.realm=OX.AC.UK -Djava.security.krb5.kdc=kdc0.ox.ac.uk:kdc1.ox.ac.uk"
我也在使用 java 6 并设置了以下变量 HADOOP_PREFIX、HADOOP_HDFS_HOME 和 YARN_HOME。您能否帮助正确配置它。
192.168.1.13: namenode running as process 10522. Stop it first.
hduser@192.168.1.13: starting datanode, logging to /usr/local/hadoop-0.23.8/logs/hadoop-
hduser-datanode-MacBook-Air.local.out
hduser@192.168.1.13: 2013-06-19 00:30:30.556 java[11171:1603] Unable to load realm info from SCDynamicStore
hduser@192.168.1.13: 2013-06-19 00:30:30.558 java[11171:1603] Unable to load realm info from SCDynamicStore
Starting secondary namenodes []
hduser@192.168.1.13: starting secondarynamenode, logging to /usr/local/hadoop-0.23.8/logs/hadoop-hduser-secondarynamenode-MacBook-Air.local.out
hduser@192.168.1.13: Exception in thread "main" java.lang.IllegalArgumentException: Target address cannot be null.
hduser@192.168.1.13: at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:179)
hduser@192.168.1.13: at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:154)
hduser@192.168.1.13: at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:143)
hduser@192.168.1.13: at org.apache.hadoop.hdfs.DFSUtil.getNameServiceId(DFSUtil.java:771)
hduser@192.168.1.13: at org.apache.hadoop.hdfs.DFSUtil.getSecondaryNameServiceId(DFSUtil.java:738)
hduser@192.168.1.13: at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.<init>(SecondaryNameNode.java:186)
hduser@192.168.1.13: at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.main(SecondaryNameNode.java:691)