Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-5349

Spark standalone should support dynamic resource scaling

    XMLWordPrintableJSON

Details

    • Improvement
    • Status: Closed
    • Major
    • Resolution: Duplicate
    • 1.2.0
    • None
    • Spark Core
    • None

    Description

      The resource requirements of an interactive shell varies heavily. Sometimes heavy commands are executed, and sometimes the user is thinking, getting coffee, interrupted etc...

      A spark shell allocates a fixed number of worker cores (at least in standalone mode). A user thus has the choice to either block other users from the cluster by allocating all cores (default behavior), or restrict him/herself to only a few cores using the option --total-executor-cores. Either way the cores allocated to the shell has low utilization, since they will be waiting for the user a lot.

      Instead the spark shell allocate resources directly required to run the driver, and request worker cores only when computation is performed on the RDDs.

      This should allow for multiple users, to use an interactive shell concurrently while stille utilizing the entire cluster, when performing heavy operations.

      Attachments

        Issue Links

          Activity

            People

              Unassigned Unassigned
              tbertelsen Tobias Bertelsen
              Votes:
              2 Vote for this issue
              Watchers:
              1 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: