Details
-
Bug
-
Status: Resolved
-
Critical
-
Resolution: Won't Fix
-
1.0.3
-
None
-
None
Description
The implementation of needsTaskCommit() assumes that the FileSystem used for writing temporary outputs is consistent. That happens not to be the case when using the S3 native filesystem in the US Standard region. It is actually quite common in larger jobs for the exists() call to return false even if the task attempt wrote output minutes earlier, which essentially cancels the commit operation with no error. That's real life data loss right there, folks.
The saddest part is that the Hadoop APIs do not seem to provide any legitimate means for the various RecordWriters to communicate with the OutputCommitter. In my projects I have created a static map of semaphores keyed by TaskAttemptID, which all my custom RecordWriters have to be aware of. That's pretty lame.
Attachments
Issue Links
- depends upon
-
HADOOP-9565 Add a Blobstore interface to add to blobstore FileSystems
- Patch Available
- relates to
-
HADOOP-4637 Unhandled failures starting jobs with S3 as backing store
- Resolved
Thanks for the report Joshua! What would you suggest as a general solution for this? Disable use of temporary output locations (i.e. no FOC) for S3 automatically since it can never be consistent (which is the real bother)?