hadoop集羣配置datanode沒法啓動的緣由

  2013-10-15 09:52:31,351 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: sunliang/192.168.1.232:9000. Already tried 3 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)java

2013-10-15 09:52:32,352 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: sunliang/192.168.1.232:9000. Already tried 4 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)node

2013-10-15 09:52:33,353 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: sunliang/192.168.1.232:9000. Already tried 5 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)apache

2013-10-15 09:52:34,354 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: sunliang/192.168.1.232:9000. Already tried 6 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)oop

2013-10-15 09:52:35,355 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: sunliang/192.168.1.232:9000. Already tried 7 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS).net

2013-10-15 09:52:37,822 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: sunliang/192.168.1.232:9000. Already tried 8 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)日誌

2013-10-15 09:52:38,823 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: sunliang/192.168.1.232:9000. Already tried 9 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)orm

2013-10-15 09:52:38,824 WARN org.apache.hadoop.hdfs.server.datanode.DataNode: java.net.ConnectException: Call to sunliang/192.168.1.232:9000 failed on connection exception: java.net.ConnectException: Connection refusedserver

        at org.apache.hadoop.ipc.Client.wrapException(Client.java:1142)進程

        at org.apache.hadoop.ipc.Client.call(Client.java:1118)ip

        at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:229)

        at com.sun.proxy.$Proxy5.sendHeartbeat(Unknown Source)

        at org.apache.hadoop.hdfs.server.datanode.DataNode.offerService(DataNode.java:1031)

        at org.apache.hadoop.hdfs.server.datanode.DataNode.run(DataNode.java:1588)

        at java.lang.Thread.run(Thread.java:662)

Caused by: java.net.ConnectException: Connection refused

        at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)

        at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:599)

        at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)

        at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:511)

        at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:481)

        at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:457)

        at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:583)

        at org.apache.hadoop.ipc.Client$Connection.access$2200(Client.java:205)

        at org.apache.hadoop.ipc.Client.getConnection(Client.java:1249)

        at org.apache.hadoop.ipc.Client.call(Client.java:1093)

        ... 5 more

當啓動hadoop時,用jps命令查看進程是datanode沒有啓動,而其餘的都正常,查看日誌顯示如上的內容,

解決方案

     刪除所用的tmp文件夾

而後執行hadoop namenode -format 進行格式化,在從新啓動start-all.sh就都好了

還有個問題就是有防火牆,關閉防火牆

相關文章
相關標籤/搜索