Uploaded image for project: 'Hadoop HDFS'
  1. Hadoop HDFS
  2. HDFS-387

Corrupted blocks leading to job failures

    XMLWordPrintableJSON

Details

    • Improvement
    • Status: Resolved
    • Major
    • Resolution: Not A Problem
    • None
    • None
    • None
    • None

    Description

      On one of our clusters we ended up with 11 singly-replicated corrupted blocks (checksum errors) such that jobs were failing because of no live blocks available.

      fsck reports the system as healthy, although it is not.

      I argue that fsck should have an option to check whether under-replicated blocks are okay.

      Even better, the namenode should automatically check under-replicated blocks with repeated replication failures for corruption and list them somewhere on the GUI. And for checksum errors, there should be an option to undo the corruption and recompute the checksums.

      Question: Is it at all probable that two or more replications of a block have checksum errors? If not, then we could reduce the checking to singly-replicated blocks.

      Attachments

        Issue Links

          Activity

            People

              Unassigned Unassigned
              ckunz Christian Kunz
              Votes:
              0 Vote for this issue
              Watchers:
              2 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: