Uploaded image for project: 'Hadoop HDFS'
  1. Hadoop HDFS
  2. HDFS-7633

BlockPoolSliceScanner fails when Datanode has too many blocks

    XMLWordPrintableJSON

Details

    • Bug
    • Status: Resolved
    • Minor
    • Resolution: Not A Problem
    • 2.6.0
    • None
    • datanode
    • None

    Description

      issue:
      When Total blocks of one of my DNs reaches 33554432, It refuses to accept more blocks, this is the ERROR.
      2015-01-16 15:21:44,571 | ERROR | DataXceiver for client at /172.1.1.8:50490 [Receiving block BP-1976278848-172.1.1.2-1419846518085:blk_1221043436_147936990] | datasight-198:25009:DataXceiver error processing WRITE_BLOCK operation src: /172.1.1.8:50490 dst: /172.1.1.11:25009 | org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:250)
      java.lang.IllegalArgumentException: n must be positive
      at java.util.Random.nextInt(Random.java:300)
      at org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner.getNewBlockScanTime(BlockPoolSliceScanner.java:263)
      at org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner.addBlock(BlockPoolSliceScanner.java:276)
      at org.apache.hadoop.hdfs.server.datanode.DataBlockScanner.addBlock(DataBlockScanner.java:193)
      at org.apache.hadoop.hdfs.server.datanode.DataNode.closeBlock(DataNode.java:1733)
      at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:765)
      at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:124)
      at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:71)
      at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:232)
      at java.lang.Thread.run(Thread.java:745)

      analysis:
      in function org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner.getNewBlockScanTime()
      when blockMap.size() is too big,
      Math.max(blockMap.size(),1) * 600 is int type, and negtive
      Math.max(blockMap.size(),1) * 600 * 1000L is long type, and negtive
      (int)period is Integer.MIN_VALUE
      Math.abs((int)period) is Integer.MIN_VALUE , which is negtive
      DFSUtil.getRandom().nextInt(periodInt) will thows IllegalArgumentException

      I use Java HotSpot (build 1.7.0_05-b05)

      Attachments

        1. HDFS-7633.patch
          0.9 kB
          Walter Su

        Issue Links

          Activity

            People

              walter.k.su Walter Su
              walter.k.su Walter Su
              Votes:
              0 Vote for this issue
              Watchers:
              4 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: