Details
-
Improvement
-
Status: Resolved
-
Major
-
Resolution: Not A Problem
-
None
-
None
-
None
Description
In the following case, the data in the HDFS is lost and a client needs to put the same file again.
- A Client puts a file to HDFS
- A DataNode crashes before replicating a block of the file to other DataNodes
I propose a metric to expose the timestamp of the oldest under-replicated/corrupt block. That way client can know what file to retain for the re-try.
Attachments
Attachments
Issue Links
- is related to
-
HDFS-10341 Add a metric to expose the timeout number of pending replication blocks
- Resolved
- relates to
-
HDFS-11024 Add rate metrics for block recovery work
- Open