Uploaded image for project: 'Hadoop HDFS'
  1. Hadoop HDFS
  2. HDFS-5299

DFS client hangs in updatePipeline RPC when failover happened



    • Bug
    • Status: Closed
    • Blocker
    • Resolution: Fixed
    • 2.1.0-beta, 3.0.0-alpha1
    • 2.2.0
    • namenode
    • None
    • Reviewed


      DFSClient got hanged in updatedPipeline call to namenode when the failover happened at exactly sametime.

      When we digged down, issue found to be with handling the RetryCache in updatePipeline.

      Here are the steps :
      1. Client was writing slowly.
      2. One of the datanode was down and updatePipeline was called to ANN.
      3. Call reached the ANN, while processing updatePipeline call it got shutdown.
      3. Now Client retried (Since the api marked as AtMostOnce) to another NameNode. at that time still NN was in STANDBY and got StandbyException.
      4. Now one more time client failover happened.
      5. Now SNN became Active.
      6. Client called to current ANN again for updatePipeline,

      Now client call got hanged in NN, waiting for the cached call with same callid to be over. But this cached call is already got over last time with StandbyException.

      Conclusion :
      Always whenever the new entry is added to cache we need to update the result of the call before returning the call or throwing exception.
      I can see similar issue multiple RPCs in FSNameSystem.


        1. HDFS-5299.patch
          8 kB
          Vinayakumar B
        2. HDFS-5299.000.patch
          11 kB
          Jing Zhao



            vinayakumarb Vinayakumar B
            vinayakumarb Vinayakumar B
            0 Vote for this issue
            10 Start watching this issue