Details
-
Improvement
-
Status: Closed
-
Major
-
Resolution: Fixed
-
2.0.0
-
None
-
None
Description
When reading an HLog, we need to got to the last block to get the file size.
In HDFS 1.0.3, it leads to HDFS-3701 / HBASE-6401
In HDFS branch-2, this bug is fixed; but we have two other issues.
1) For simple cases as a single node died, we don't have the effect of HDFS-3703, and the default location order leads us to try to connect to a dead datanode while we should not. This is not analysed yet. A specific JIRA will be created later.
2) If we are redirected to a wrong node, we experience a huge delay:
The pseudo code in DFSInputStream#readBlockLength is:
for(DatanodeInfo datanode : locatedblock.getLocations()) { try { ClientDatanodeProtocol cdp = DFSUtil.createClientDatanodeProtocolProxy( datanode, dfsClient.conf, dfsClient.getConf().socketTimeout, dfsClient.getConf().connectToDnViaHostname, locatedblock); return cdp.getReplicaVisibleLength(locatedblock.getBlock()); } catch { // retry } }
However, with this code, the connection is created with a null RetryPolicy. It's then defaulted to 10 retries, with:
public static final String IPC_CLIENT_CONNECT_MAX_RETRIES_KEY = "ipc.client.connect.max.retries"; public static final int IPC_CLIENT_CONNECT_MAX_RETRIES_DEFAULT = 10;
So if the first datanode is bad, we will try 10 times before trying the second. In the context of HBASE-6738, the split task is cancelled before we're have opened the file to split.
By nature, it's likely to be a pure HDFS issue. But may be it can be solved in HBase with the right configuration on "ipc.client.connect.max.retries".
The ideal fix (in HDFS) would be to try the datanodes once each, and then loop 10 times.
Attachments
Issue Links
- is required by
-
HBASE-5843 Improve HBase MTTR - Mean Time To Recover
- Closed