Uploaded image for project: 'Hadoop HDFS'
  1. Hadoop HDFS
  2. HDFS-1567

DFSClient should retry reading from all datanodes in round robin fashion

    XMLWordPrintableJSON

Details

    • Improvement
    • Status: Open
    • Major
    • Resolution: Unresolved
    • None
    • None
    • hdfs-client
    • None

    Description

      In the current implementation, the DFSClient retries the same datanode a few times (for reading) before marking the datanode as "dead" and moving on to trying the read-request from the next replica. This introduces additional delay when the first candidate datanode was really out-of-service. Instead, it might be beneficial (for performance) to make the DFSClient retry all datanode replicas one after another (while keeping the total number of retries per datanode the same as now); if all the replicas were tried unsuccessfully once, then loop back to retrying the first one and so on.

      Attachments

        Activity

          People

            Unassigned Unassigned
            dhruba Dhruba Borthakur
            Votes:
            0 Vote for this issue
            Watchers:
            6 Start watching this issue

            Dates

              Created:
              Updated: