Uploaded image for project: 'Hadoop Map/Reduce'
  1. Hadoop Map/Reduce
  2. MAPREDUCE-4311

Capacity scheduler.xml does not accept decimal values for capacity and maximum-capacity settings

VotersWatch issueWatchersCreate sub-taskLinkCloneUpdate Comment AuthorReplace String in CommentUpdate Comment VisibilityDelete Comments
    XMLWordPrintableJSON

    Details

    • Target Version/s:
    • Hadoop Flags:
      Incompatible change, Reviewed

      Description

      if capacity scheduler capacity or max capacity set with decimal it errors:

      • Error starting ResourceManager

      java.lang.NumberFormatException: For input string: "10.5"
      at
      java.lang.NumberFormatException.forInputString(NumberFormatException.java:48)
      at java.lang.Integer.parseInt(Integer.java:458)
      at java.lang.Integer.parseInt(Integer.java:499)
      at org.apache.hadoop.conf.Configuration.getInt(Configuration.java:713)
      at
      org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacitySchedulerConfiguration.getCapacity(CapacitySchedulerConfiguration.java:147)
      at
      org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.LeafQueue.<init>(LeafQueue.java:147)
      at
      org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler.parseQueue(CapacityScheduler.java:297)
      at

      0.20 used to take decimal and this could be an issue on large clusters that would have queues with small allocations.

        Attachments

        1. MR-4311.patch
          15 kB
          Karthik Kambatla

        Issue Links

          Activity

            People

            • Assignee:
              kasha Karthik Kambatla
              Reporter:
              tgraves Thomas Graves

              Dates

              • Created:
                Updated:
                Resolved:

                Issue deployment