Uploaded image for project: 'Hive'
  1. Hive
  2. HIVE-7292 Hive on Spark
  3. HIVE-7768

Integrate with Spark executor scaling [Spark Branch]



    • Type: Sub-task
    • Status: Resolved
    • Priority: Critical
    • Resolution: Done
    • Affects Version/s: None
    • Fix Version/s: None
    • Component/s: Spark
    • Labels:


      A user connects to Hive and runs a query on a small time. Our SC is sized for that small table. They then run a query on a much larger table. We'll need to "re-size" the SC which I don't think Spark supports today, so we need to research what is available today in Spark and how Tez works.

      More details:
      Similar to Tez, it's likely our "SparkContext" is going to be long lived and process many queries. Some queries will be large and some small. Additionally the SC might be idle for long periods of time.

      In this JIRA we will research the following:

      • How Spark decides the number of slaves for a given RDD today
      • Given a SC when you create a new RDD based on a much larger input dataset, does the SC adjust?
      • How Tez increases/decreases the size of the running YARN application (set of slaves)
      • How Tez handles scenarios when it has a running set of slaves in YARN and requests more resources for a query and fails to get additional resources
      • How Tez decides to timeout idle slaves

      This will guide requirements we'll need from Spark.


          Issue Links



              • Assignee:
                chengxiang li Chengxiang Li
                brocknoland Brock Noland
              • Votes:
                0 Vote for this issue
                6 Start watching this issue


                • Created: