In the following case, the data in the HDFS is lost and a client needs to put the same file again.
- A Client puts a file to HDFS
- A DataNode crashes before replicating a block of the file to other DataNodes
I propose a metric to expose the timestamp of the oldest under-replicated/corrupt block. That way client can know what file to retain for the re-try.