Uploaded image for project: 'Hadoop Common'
  1. Hadoop Common
  2. HADOOP-11890 Uber-JIRA: Hadoop should support IPv6
  3. HADOOP-12430

Fix HDFS client gets errors trying to to connect to IPv6 DataNode

VotersWatch issueWatchersLinkCloneUpdate Comment AuthorReplace String in CommentUpdate Comment VisibilityDelete Comments
    XMLWordPrintableJSON

Details

    • Reviewed

    Description

      1st exception, on put:

      15/03/23 18:43:18 WARN hdfs.DFSClient: DataStreamer Exception
      java.lang.IllegalArgumentException: Does not contain a valid host:port authority: 2401:db00:1010:70ba:face:0:8:0:50010
      at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:212)
      at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164)
      at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:153)
      at org.apache.hadoop.hdfs.DFSOutputStream.createSocketForPipeline(DFSOutputStream.java:1607)
      at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.createBlockOutputStream(DFSOutputStream.java:1408)
      at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1361)
      at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:588)

      Appears to actually stem from code in DataNodeID which assumes it's safe to append together (ipaddr + ":" + port) – which is OK for IPv4 and not OK for IPv6. NetUtils.createSocketAddr( ) assembles a Java URI object, which requires the format proto://[2401:db00:1010:70ba:face:0:8:0]:50010

      Currently using InetAddress.getByName() to validate IPv6 (guava InetAddresses.forString has been flaky) but could also use our own parsing. (From logging this, it seems like a low-enough frequency call that the extra object creation shouldn't be problematic, and for me the slight risk of passing in bad input that is not actually an IPv4 or IPv6 address and thus calling an external DNS lookup is outweighed by getting the address normalized and avoiding rewriting parsing.)

      Alternatively, sun.net.util.IPAddressUtil.isIPv6LiteralAddress()

      -------

      2nd exception (on datanode)
      15/04/13 13:18:07 ERROR datanode.DataNode: dev1903.prn1.facebook.com:50010:DataXceiver error processing unknown operation src: /2401:db00:20:7013:face:0:7:0:54152 dst: /2401:db00:11:d010:face:0:2f:0:50010
      java.io.EOFException
      at java.io.DataInputStream.readShort(DataInputStream.java:315)
      at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:58)
      at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:226)
      at java.lang.Thread.run(Thread.java:745)

      Which also comes as client error "-get: 2401 is not an IP string literal."

      This one has existing parsing logic which needs to shift to the last colon rather than the first. Should also be a tiny bit faster by using lastIndexOf rather than split. Could alternatively use the techniques above.

      Attachments

        1. HDFS-8078.9.patch
          13 kB
          Nate Edel
        2. HDFS-8078.10.patch
          13 kB
          Nate Edel
        3. HDFS-8078.11.patch
          26 kB
          Nate Edel
        4. HDFS-8078.12.patch
          19 kB
          Nate Edel
        5. HDFS-8078.13.patch
          13 kB
          Nate Edel
        6. HDFS-8078.14.patch
          22 kB
          Nate Edel
        7. dummy.patch
          0.6 kB
          Nate Edel
        8. HDFS-8078.15.patch
          22 kB
          Nate Edel
        9. HDFS-8078-HADOOP-17800.001.patch
          21 kB
          Hemanth Boyina
        10. HDFS-8078-HADOOP-17800.002.patch
          22 kB
          Hemanth Boyina

        Issue Links

        Activity

          This comment will be Viewable by All Users Viewable by All Users
          Cancel

          People

            nkedel Nate Edel
            nkedel Nate Edel
            Votes:
            1 Vote for this issue
            Watchers:
            15 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved:

              Slack

                Issue deployment