Uploaded image for project: 'Hadoop HDFS'
  1. Hadoop HDFS
  2. HDFS-16316

Improve DirectoryScanner: add regular file check related block

    XMLWordPrintableJSON

Details

    Description

      Something unusual happened in the online environment.
      The DataNode is configured with 11 disks (${dfs.datanode.data.dir}). It is normal for 10 disks to calculate the used capacity, and the calculated value for the other 1 disk is much larger, which is very strange.
      This is about the live view on the NameNode:

      This is about the live view on the DataNode:

      We can look at the view on linux:

      There is a big gap here, regarding'/mnt/dfs/11/data'. This situation should be prohibited from happening.

      I found that there are some abnormal block files.
      There are wrong blk_xxxx.meta in some subdir directories, causing abnormal computing space.
      Here are some abnormal block files:

      Such files should not be used as normal blocks. They should be actively identified and filtered, which is good for cluster stability.

      Attachments

        1. screenshot-4.png
          126 kB
          JiangHua Zhu
        2. screenshot-3.png
          141 kB
          JiangHua Zhu
        3. screenshot-2.png
          310 kB
          JiangHua Zhu
        4. screenshot-1.png
          44 kB
          JiangHua Zhu

        Issue Links

          Activity

            People

              jianghuazhu JiangHua Zhu
              jianghuazhu JiangHua Zhu
              Votes:
              0 Vote for this issue
              Watchers:
              6 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved:

                Time Tracking

                  Estimated:
                  Original Estimate - Not Specified
                  Not Specified
                  Remaining:
                  Remaining Estimate - 0h
                  0h
                  Logged:
                  Time Spent - 6h 20m
                  6h 20m