Hadoop HDFS
  1. Hadoop HDFS
  2. HDFS-70

Data node should shutdown when a "critical" error is returned by the name node

    Details

    • Type: Bug Bug
    • Status: Resolved
    • Priority: Minor Minor
    • Resolution: Won't Fix
    • Affects Version/s: None
    • Fix Version/s: None
    • Component/s: None
    • Labels:
      None

      Description

      Currently data node does not distinguish between critical and non critical exceptions.
      Any exception is treated as a signal to sleep and then try again. See
      org.apache.hadoop.dfs.DataNode.run()
      This is happening because RPC always throws the same RemoteException.
      In some cases (like UnregisteredDatanodeException, IncorrectVersionException) the data
      node should shutdown rather than retry.
      This logic naturally belongs to the
      org.apache.hadoop.dfs.DataNode.offerService()
      but can be reasonably implemented (without examining the RemoteException.className
      field) after HADOOP-266 (2) is fixed.

        Activity

        Hide
        Harsh J added a comment -

        HADOOP-266 was resolved as a Won't Fix and the DN currently works OK with the way it analyzes the exception classnames and determines if it has to shutdown.

        Marking this one as Won't Fix as well, following HADOOP-266

        Show
        Harsh J added a comment - HADOOP-266 was resolved as a Won't Fix and the DN currently works OK with the way it analyzes the exception classnames and determines if it has to shutdown. Marking this one as Won't Fix as well, following HADOOP-266

          People

          • Assignee:
            Sameer Paranjpye
            Reporter:
            Konstantin Shvachko
          • Votes:
            0 Vote for this issue
            Watchers:
            0 Start watching this issue

            Dates

            • Created:
              Updated:
              Resolved:

              Development