Details
-
Bug
-
Status: Closed
-
Major
-
Resolution: Fixed
-
0.7.2
-
None
-
None
Description
for a particular file [alas, the file no longer exists -- I had to progress]
$dfs -cp foo bar
and
$dfs -get foo local
failed on a checksum error. The dfs browser's download function retrieved the file, so either that function doesn't check, or more likely the download function got a different copy.
When a checksum fails on one copy of a file that is redundantly stored, I would prefer that dfs try a different copy, mark the bad one as not existing [which should induce a fresh copy being made from one of the good copies eventually], and make the call continue to work and deliver bytes.
Ideally, if all copies have checksum errors but it's possible to piece together a good copy I would like that to be done.
-dk
Attachments
Attachments
Issue Links
- relates to
-
HADOOP-855 HDFS should repair corrupted files
- Closed