Details
-
Bug
-
Status: Resolved
-
Major
-
Resolution: Incomplete
-
None
-
None
-
None
-
None
Description
Not able to run randomwriter if all the datanodes of any one of the racks are killed. (replication factor : 3)
Randomwriter job gets failed and following error message is displayed in log:
java.net.ConnectException: Connection refused
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:574)
at sun.nio.ch.SocketAdaptor.connect(SocketAdaptor.java:100)
at org.apache.hadoop.dfs.DFSClient$DFSOutputStream.createBlockOutputStream(DFSClient.java:2398)
at org.apache.hadoop.dfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2354)
at org.apache.hadoop.dfs.DFSClient$DFSOutputStream.access$1800(DFSClient.java:1744)
at org.apache.hadoop.dfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:1927)
Attachments
Issue Links
- is related to
-
HADOOP-5599 Unable to run jobs when all the nodes in rack are down
- Closed