Uploaded image for project: 'Hadoop HDFS'
  1. Hadoop HDFS
  2. HDFS-134

premature end-of-decommission of datanodes

    XMLWordPrintableJSON

Details

    • Bug
    • Status: Resolved
    • Major
    • Resolution: Fixed
    • None
    • None
    • None
    • None

    Description

      Decommissioning requires that the nodes be listed in the dfs.hosts.excludes file. The administrator runs the "dfsadmin -refreshNodes" command. The decommissioning process starts off. Suppose that one of the datanodes that was being decommisioned has to re-register with the namenode. This can occur if the namenode restarts or if the datanode restarts while the decommissioning was in progress. Now, the namenode refuses to talk to this datanode because it is in the excludes list! This is a premature end of the decommissioning process.

      One way to fix this bug is to make the namenode always accept registration requests, even for datanodes that are in the exclude list. The namenode, however, should set the "being decommissioned" flag for these datanodes. It should then re-start the decommisioning process for these datanodes. When the decommissioning is complete, the namenode will ask the datanodes to shutdown.

      Attachments

        Issue Links

          Activity

            People

              Unassigned Unassigned
              dhruba Dhruba Borthakur
              Votes:
              1 Vote for this issue
              Watchers:
              1 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: