Uploaded image for project: 'Hadoop Common'
  1. Hadoop Common
  2. HADOOP-2873

Namenode fails to re-start after cluster shutdown - DFSClient: Could not obtain blocks even all datanodes were up & live

    XMLWordPrintableJSON

Details

    • Bug
    • Status: Closed
    • Major
    • Resolution: Fixed
    • 0.17.0
    • 0.17.0
    • None
    • None
    • Incompatible change

    Description

      Namenode fails to re-start with the following exception:

      2008-02-21 14:20:48,831 INFO org.apache.hadoop.dfs.NameNode: STARTUP_MSG:
      /************************************************************
      STARTUP_MSG: Starting NameNode
      STARTUP_MSG: host = se09/141.76.xxx.xxx
      STARTUP_MSG: args = []
      STARTUP_MSG: version = 2008-02-19_11-01-48
      STARTUP_MSG: build = http://svn.apache.org/repos/asf/hadoop/core/trunk -r 628999; compiled by 'hudson' on Tue Feb 19 11:09:05 UTC 2008
      ************************************************************/
      2008-02-21 14:20:49,367 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: Initializing RPC Metrics with serverName=NameNode, port=8000
      2008-02-21 14:20:49,374 INFO org.apache.hadoop.dfs.NameNode: Namenode up at: se09.inf.tu-dresden.de/141.76.xxx.xxx:8000
      2008-02-21 14:20:49,378 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: Initializing JVM Metrics with processName=NameNode, sessionId=null
      2008-02-21 14:20:49,381 INFO org.apache.hadoop.dfs.NameNodeMetrics: Initializing NameNodeMeterics using context object:org.apache.hadoop.metrics.spi.NullContext
      2008-02-21 14:20:49,501 INFO org.apache.hadoop.fs.FSNamesystem: fsOwner=amartin,students
      2008-02-21 14:20:49,501 INFO org.apache.hadoop.fs.FSNamesystem: supergroup=supergroup
      2008-02-21 14:20:49,501 INFO org.apache.hadoop.fs.FSNamesystem: isPermissionEnabled=true
      2008-02-21 14:20:49,788 INFO org.apache.hadoop.ipc.Server: Stopping server on 8000
      2008-02-21 14:20:49,790 ERROR org.apache.hadoop.dfs.NameNode: java.io.IOException: Created 13 leases but found 4
      at org.apache.hadoop.dfs.FSImage.loadFilesUnderConstruction(FSImage.java:935)
      at org.apache.hadoop.dfs.FSImage.loadFSImage(FSImage.java:749)
      at org.apache.hadoop.dfs.FSImage.loadFSImage(FSImage.java:634)
      at org.apache.hadoop.dfs.FSImage.recoverTransitionRead(FSImage.java:223)
      at org.apache.hadoop.dfs.FSDirectory.loadFSImage(FSDirectory.java:79)
      at org.apache.hadoop.dfs.FSNamesystem.initialize(FSNamesystem.java:261)
      at org.apache.hadoop.dfs.FSNamesystem.<init>(FSNamesystem.java:242)
      at org.apache.hadoop.dfs.NameNode.initialize(NameNode.java:131)
      at org.apache.hadoop.dfs.NameNode.<init>(NameNode.java:176)
      at org.apache.hadoop.dfs.NameNode.<init>(NameNode.java:162)
      at org.apache.hadoop.dfs.NameNode.createNameNode(NameNode.java:851)
      at org.apache.hadoop.dfs.NameNode.main(NameNode.java:860)

      2008-02-21 14:20:49,791 INFO org.apache.hadoop.dfs.NameNode: SHUTDOWN_MSG:
      /************************************************************
      SHUTDOWN_MSG: Shutting down NameNode at se09/141.76.xxx.xxx
      ************************************************************/

      Cluster restart was needed since the DFS client produced the following error message even all datanodes were up:

      08/02/21 14:04:35 INFO fs.DFSClient: Could not obtain block blk_-4008950704646490788 from any node: java.io.IOException: No live nodes contain current block

      Attachments

        1. leaseConstruction.patch
          2 kB
          Dhruba Borthakur
        2. leaseConstruction.patch
          5 kB
          Dhruba Borthakur
        3. leaseConstruction.patch
          5 kB
          Dhruba Borthakur
        4. leaseConstruction.patch
          5 kB
          Dhruba Borthakur

        Activity

          People

            dhruba Dhruba Borthakur
            andremartin André Martin
            Votes:
            0 Vote for this issue
            Watchers:
            0 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved: