Uploaded image for project: 'Hadoop HDFS'
  1. Hadoop HDFS
  2. HDFS-2938

Recursive delete of a large directory makes namenode unresponsive

    XMLWordPrintableJSON

Details

    • Bug
    • Status: Resolved
    • Major
    • Resolution: Fixed
    • 0.22.0
    • 0.23.2
    • namenode
    • None
    • Reviewed

    Description

      When deleting a large directory with millions of files, namenode holding FSNamesystem lock will make it unresponsive for other request. In this scenario HDFS-173 added a mechanism to delete blocks in smaller chunks holding the locks. With new read/write lock changes, the mechanism from HDFS-173 is lost. Need to resurrect the mechanism back. Also a good unit test/update to existing unit test is needed to catch future errors with this functionality.

      Attachments

        1. HDFS-2938.patch
          3 kB
          Hari Mankude
        2. HDFS-2938.patch
          3 kB
          Hari Mankude

        Issue Links

          Activity

            People

              harip Hari Mankude
              sureshms Suresh Srinivas
              Votes:
              0 Vote for this issue
              Watchers:
              6 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: