Uploaded image for project: 'Beam'
  1. Beam
  2. BEAM-4783

Add bundleSize parameter to control splitting of Spark sources (useful for Dynamic Allocation)

Details

    • Improvement
    • Status: Triage Needed
    • P2
    • Resolution: Fixed
    • 2.8.0, 2.9.0
    • 2.11.0
    • runner-spark
    • None

    Description

      When the spark-runner is used along with the configuration spark.dynamicAllocation.enabled=true the SourceRDD does not detect this. It then falls back to the value calculated in this description:
      // when running on YARN/SparkDeploy it's the result of max(totalCores, 2).
      // when running on Mesos it's 8.
      // when running local it's the total number of cores (local = 1, local[N] = N,
      // local[*] = estimation of the machine's cores).
      // ** the configuration "spark.default.parallelism" takes precedence over all of the above **
      So in most cases this default is quite small. This is an issue when using a very large input file as it will only get split in half.

      I believe that when Dynamic Allocation is enable the SourceRDD should use the DEFAULT_BUNDLE_SIZE and possibly expose a SparkPipelineOptions that allows you to change this DEFAULT_BUNDLE_SIZE.

      Attachments

        Activity

          People

            winkelman.kyle Kyle Winkelman
            winkelman.kyle Kyle Winkelman
            Votes:
            0 Vote for this issue
            Watchers:
            4 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved:

              Time Tracking

                Estimated:
                Original Estimate - Not Specified
                Not Specified
                Remaining:
                Remaining Estimate - 0h
                0h
                Logged:
                Time Spent - 8h 20m
                8h 20m