HBase主站因“连接被拒绝”错误而停止

时间:2013-05-06 15:04:44

标签: hadoop hbase cloudera

这发生在伪分布式和分布式模式中。 当我尝试启动HBase时,最初所有3个服务 - 主服务器,区域服务器和quorumpeer启动。然而,在一分钟内,主人停止。在日志中,这是跟踪 -

2013-05-06 20:10:25,525 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: <master/master_ip>:9000. Already tried 0 time(s).
2013-05-06 20:10:26,528 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: <master/master_ip>:9000. Already tried 1 time(s).
2013-05-06 20:10:27,530 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: <master/master_ip>:9000. Already tried 2 time(s).
2013-05-06 20:10:28,533 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: <master/master_ip>:9000. Already tried 3 time(s).
2013-05-06 20:10:29,535 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: <master/master_ip>:9000. Already tried 4 time(s).
2013-05-06 20:10:30,538 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: <master/master_ip>:9000. Already tried 5 time(s).
2013-05-06 20:10:31,540 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: <master/master_ip>:9000. Already tried 6 time(s).
2013-05-06 20:10:32,543 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: <master/master_ip>:9000. Already tried 7 time(s).
2013-05-06 20:10:33,544 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: <master/master_ip>:9000. Already tried 8 time(s).
2013-05-06 20:10:34,547 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: <master/master_ip>:9000. Already tried 9 time(s).
2013-05-06 20:10:34,550 FATAL org.apache.hadoop.hbase.master.HMaster: Unhandled exception. Starting shutdown.
java.net.ConnectException: Call to <master/master_ip>:9000 failed on connection exception: java.net.ConnectException: Connection refused
        at org.apache.hadoop.ipc.Client.wrapException(Client.java:1179)
        at org.apache.hadoop.ipc.Client.call(Client.java:1155)
        at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:226)
        at $Proxy9.getProtocolVersion(Unknown Source)
        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:398)
        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:384)
        at org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:132)
        at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:259)
        at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:220)
        at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:89)
        at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1611)
        at org.apache.hadoop.fs.FileSystem.access$300(FileSystem.java:68)
        at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:1645)
        at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1627)
        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:254)
        at org.apache.hadoop.fs.Path.getFileSystem(Path.java:183)
        at org.apache.hadoop.hbase.util.FSUtils.getRootDir(FSUtils.java:363)
        at org.apache.hadoop.hbase.master.MasterFileSystem.<init>(MasterFileSystem.java:86)
        at org.apache.hadoop.hbase.master.HMaster.finishInitialization(HMaster.java:368)
        at org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:301)
Caused by: java.net.ConnectException: Connection refused
        at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
        at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:592)
        at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
        at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:519)
        at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:484)
        at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:468)
        at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:575)
        at org.apache.hadoop.ipc.Client$Connection.access$2300(Client.java:212)
        at org.apache.hadoop.ipc.Client.getConnection(Client.java:1292)
        at org.apache.hadoop.ipc.Client.call(Client.java:1121)
        ... 18 more

我已经采取措施解决这个问题但没有成功 - 从分布式模式降级到伪分布式模式。同样的问题。 - 尝试独立模式 - 没有运气 - 对hadoop和hbase使用相同的用户(hadoop)。为hadoop设置无密码ssh。 - 同样的问题。 - 编辑/ etc / hosts文件并将localhost / servername和127.0.0.1更改为引用SO和不同来源的实际IP地址。还是同样的问题。 - 重启服务器

以下是conf文件。

HBase的-site.xml中

<configuration>
<property>
  <name>hbase.rootdir</name>
  <value>hdfs://<master>:9000/hbase</value>
        <description>The directory shared by regionservers.</description>
</property>

<property>
        <name>hbase.cluster.distributed</name>
        <value>true</value>
</property>

<property>
        <name>hbase.zookeeper.quorum</name>
        <value><master></value>
</property>

<property>
        <name>hbase.master</name>
        <value><master>:60000</value>
        <description>The host and port that the HBase master runs at.</description>
</property>

<property>
        <name>dfs.replication</name>
        <value>1</value>
        <description>The replication count for HLog and HFile storage. Should not be greater than HDFS datanode count.</description>
</property>

</configuration>

/ etc / hosts文件

127.0.0.1 localhost.localdomain localhost :: 1 localhost6.localdomain6 localhost6      。

我在这里做错了什么?

Hadoop版本 - Hadoop 0.20.2-cdh3u5 HBase版本 - 版本0.90.6-cdh3u5

2 个答案:

答案 0 :(得分:5)

通过查看配置文件,我假设您在配置文件中使用实际的主机名。如果是这种情况,请将主机名和计算机的IP添加到/ etc / hosts文件中。还要确保它与Hadoop的core-site.xml中的主机名匹配。正确的名称解析对于正确的HBase功能至关重要。

如果您仍然遇到任何问题,请正确执行here提及的步骤。我试图详细解释这个程序,如果你仔细地遵循所有步骤,我希望你能够让它运行。

HTH

答案 1 :(得分:-1)

我相信你正在尝试使用伪分布式模式。我得到了同样的错误,直到我解决了3件事:

  1. 本地/ etc / hosts文件
  2. $ cat / etc / hosts

    127.0.0.1   localhost
    255.255.255.255 broadcasthost
    ::1             localhost 
    fe80::1%lo0 localhost
    172.20.x.x  my.hostname.com
    
    1. 而不是指向主机名,指向hbase-env.sh中的localhost

    2. 更正我的类路径 A.确保Hadoop在classpath中(通过hbase-env.sh)

      export JAVA_HOME =您的java home路径 export HADOOP_HOME =你的hadoop回家之路 export HBASE_HOME =你的hbase home路径

      导出HBASE_CLASSPATH =您的hbase home / conf路径:您的hadoop home / conf路径

    3. B中。运行我的程序时,我编辑了HBase的以下bash脚本:权威指南(bin / run.sh) $ grep -v#bin / run.sh

      bin=`dirname "$0"`
      bin=`cd "$bin">/dev/null; pwd`
      
        echo "usage: $(basename $0) <example-name>"
        exit 1;
      fi
      
      MVN="mvn"
      if [ "$MAVEN_HOME" != "" ]; then
        MVN=${MAVEN_HOME}/bin/mvn
      fi
      
      CLASSPATH="${HBASE_CONF_DIR}"
      
      if [ -d "${bin}/../target/classes" ]; then
        CLASSPATH=${CLASSPATH}:${bin}/../target/classes
      fi
      
      cpfile="${bin}/../target/cached_classpath.txt"
      if [ ! -f "${cpfile}" ]; then
        ${MVN} -f "${bin}/../pom.xml" dependency:build-classpath -Dmdep.outputFile="${cpfile}" &> /dev/null
      fi
      CLASSPATH=`hbase classpath`:${CLASSPATH}:`cat "${cpfile}"`
      
      JAVA_HOME=your path to java home
      JAVA=$JAVA_HOME/bin/java
      JAVA_HEAP_MAX=-Xmx512m
      
      echo "Classpath is $CLASSPATH"
      "$JAVA" $JAVA_HEAP_MAX -classpath "$CLASSPATH" "$@"
      

      值得注意的是我正在使用Mac。我相信这些说明也适用于Linux。