Hadoop Map/Reduce
  1. Hadoop Map/Reduce
  2. MAPREDUCE-1485

CapacityScheduler should have prevent a single job taking over large parts of a cluster

    Details

    • Type: Improvement Improvement
    • Status: Resolved
    • Priority: Major Major
    • Resolution: Fixed
    • Affects Version/s: None
    • Fix Version/s: None
    • Component/s: capacity-sched
    • Labels:
      None

      Description

      The proposal is to have a per-queue limit on the number of concurrent tasks a job can run on a cluster.

      We've seen cases where a single, large, job took over a majority of the cluster - worse, it meant that any bug in it caused issues for both the NameNode and the JobTracker.

        Activity

        Hide
        Allen Wittenauer added a comment -

        user limits were implemented.

        Show
        Allen Wittenauer added a comment - user limits were implemented.
        Hide
        David Ciemiewicz added a comment -

        I would also be good if any single user was limited to as to the total number of tasks.

        This way it doesn't matter if a user has a single job with 20,000 tasks or many 20 jobs with 1,000 tasks.

        Show
        David Ciemiewicz added a comment - I would also be good if any single user was limited to as to the total number of tasks. This way it doesn't matter if a user has a single job with 20,000 tasks or many 20 jobs with 1,000 tasks.

          People

          • Assignee:
            Arun C Murthy
            Reporter:
            Arun C Murthy
          • Votes:
            1 Vote for this issue
            Watchers:
            7 Start watching this issue

            Dates

            • Created:
              Updated:
              Resolved:

              Development