Details
-
Bug
-
Status: Closed
-
Blocker
-
Resolution: Fixed
-
0.20-append
-
None
-
None
-
hbase
Description
Before 0.18, when Datanode restarts, it deletes files under data-dir/tmp directory since these files are not valid anymore. But in 0.18 it moves these files to normal directory incorrectly making them valid blocks. One of the following would work :
- remove the tmp files during upgrade, or
- if the files under /tmp are in pre-18 format (i.e. no generation), delete them.
Currently effect of this bug is that, these files end up failing block verification and eventually get deleted. But cause incorrect over-replication at the namenode before that.
Also it looks like our policy regd treating files under tmp needs to be defined better. Right now there are probably one or two more bugs with it. Dhruba, please file them if you rememeber.
Attachments
Attachments
Issue Links
- depends upon
-
HDFS-101 DFS write pipeline : DFSClient sometimes does not detect second datanode failure
- Closed
-
HDFS-793 DataNode should first receive the whole packet ack message before it constructs and sends its own ack message for the packet
- Closed
-
HDFS-988 saveNamespace race can corrupt the edits log
- Closed
-
HDFS-606 ConcurrentModificationException in invalidateCorruptReplicas()
- Closed
-
HDFS-826 Allow a mechanism for an application to detect that datanode(s) have died in the write pipeline
- Closed
- incorporates
-
HADOOP-4997 workaround for tmp file handling on DataNodes in 0.18 (HADOOP-4663)
- Closed
- relates to
-
HDFS-57 A Datanode's datadir could have lots of blocks in the top-level directory
- Resolved
-
HADOOP-4702 Failed block replication leaves an incomplete block in receiver's tmp data directory
- Closed
-
HADOOP-4810 Data lost at cluster startup time
- Closed
-
HDFS-29 In Datanode, update block may fail due to length inconsistency
- Closed