Uploaded image for project: 'Hadoop HDFS'
  1. Hadoop HDFS
  2. HDFS-6682

Add a metric to expose the timestamp of the oldest under-replicated block

Log workAgile BoardRank to TopRank to BottomAttach filesAttach ScreenshotBulk Copy AttachmentsBulk Move AttachmentsVotersWatch issueWatchersCreate sub-taskConvert to sub-taskMoveLinkCloneLabelsUpdate Comment AuthorReplace String in CommentUpdate Comment VisibilityDelete Comments
    XMLWordPrintableJSON

Details

    • Improvement
    • Status: Resolved
    • Major
    • Resolution: Not A Problem
    • None
    • None
    • None

    Description

      In the following case, the data in the HDFS is lost and a client needs to put the same file again.

      1. A Client puts a file to HDFS
      2. A DataNode crashes before replicating a block of the file to other DataNodes

      I propose a metric to expose the timestamp of the oldest under-replicated/corrupt block. That way client can know what file to retain for the re-try.

      Attachments

        1. HDFS-6682.002.patch
          10 kB
          Akira Ajisaka
        2. HDFS-6682.003.patch
          11 kB
          Akira Ajisaka
        3. HDFS-6682.004.patch
          9 kB
          Akira Ajisaka
        4. HDFS-6682.005.patch
          10 kB
          Akira Ajisaka
        5. HDFS-6682.006.patch
          10 kB
          Akira Ajisaka
        6. HDFS-6682.patch
          9 kB
          Akira Ajisaka

        Issue Links

        Activity

          This comment will be Viewable by All Users Viewable by All Users
          Cancel

          People

            aajisaka Akira Ajisaka Assign to me
            aajisaka Akira Ajisaka
            Votes:
            1 Vote for this issue
            Watchers:
            8 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved:

              Slack

                Issue deployment