When DFS cluster; The datanodes keep on dumping the logs:2012-06-19 04:05:04,842 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Adding block pool BP-1233167859-10.10.35.8-13400930059602012-06-19 04:05:04,861 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Block pool BP-1233167859-10.10.35.8-1340093005960 (storage id DS-1255092928-10.244.15.192-50010-1340093104449) service to ec2-107-22-39-220.compute-1.amazonaws.com/10.10.35.8:9000 beginning handshake with NN2012-06-19 04:05:04,979 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Block pool Block pool BP-1233167859-10.10.35.8-1340093005960 (storage id DS-1255092928-10.244.15.192-50010-1340093104449) service to ec2-107-22-39-220.compute-1.amazonaws.com/10.10.35.8:9000 successfully registered with NN2012-06-19 04:05:04,980 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: For namenode ec2-107-22-39-220.compute-1.amazonaws.com/10.10.35.8:9000 using DELETEREPORT_INTERVAL of 300000 msec BLOCKREPORT_INTERVAL of 21600000msec Initial delay: 0msec; heartBeatInterval=30002012-06-19 04:05:05,258 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Namenode Block pool BP-1233167859-10.10.35.8-1340093005960 (storage id DS-1255092928-10.244.15.192-50010-1340093104449) service to ec2-107-22-39-220.compute-1.amazonaws.com/10.10.35.8:9000 trying to claim ACTIVE state with txid=242012-06-19 04:05:05,258 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Acknowledging ACTIVE Namenode Block pool BP-1233167859-10.10.35.8-1340093005960 (storage id DS-1255092928-10.244.15.192-50010-1340093104449) service to ec2-107-22-39-220.compute-1.amazonaws.com/10.10.35.8:90002012-06-19 04:05:05,394 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: BlockReport of 0 blocks took 1 msec to generate and 135 msecs for RPC and NN processing2012-06-19 04:05:05,394 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: sent block report, processed command:org.apache.hadoop.hdfs.server.protocol.FinalizeCommand@49d318592012-06-19 04:05:05,396 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Periodic Block Verification Scanner initialized with interval 504 hours for block pool BP-1233167859-10.10.35.8-1340093005960.2012-06-19 04:05:05,400 INFO org.apache.hadoop.hdfs.server.datanode.DataBlockScanner: Added bpid=BP-1233167859-10.10.35.8-1340093005960 to blockPoolScannerMap, new size=12012-06-19 04:05:09,845 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Starting a new period : work left in prev period : 0.00%2012-06-19 04:05:14,848 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Starting a new period : work left in prev period : 0.00%2012-06-19 04:05:19,849 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Starting a new period : work left in prev period : 0.00%2012-06-19 04:05:24,850 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Starting a new period : work left in prev period : 0.00%2012-06-19 04:05:29,851 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Starting a new period : work left in prev period : 0.00%2012-06-19 04:05:34,852 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Starting a new period : work left in prev period : 0.00%2012-06-19 04:05:39,853 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Starting a new period : work left in prev period : 0.00%2012-06-19 04:05:44,854 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Starting a new period : work left in prev period : 0.00%2012-06-19 04:05:49,855 INFO org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Starting a new period : work left in prev period : 0.00%

The DFSClient operations are also not taking place properly:2012-06-19 04:04:56,653 WARN org.apache.hadoop.hdfs.DFSClient: DataStreamer Exceptionjava.io.IOException: File /hbase/hbase.version could only be replicated to 0 nodes instead of minReplication (=1). There are 1 datanode(s) running and no node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1256) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1977) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:470) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:292) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:42602) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:427) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:916) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1692) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1688) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1232)

Please let me know if there is any configuration which needs to be set…..The hadoop version being used is: 2.0.0