Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-40238

support scaleUpFactor and initialNumPartition in pyspark rdd API

    XMLWordPrintableJSON

Details

    • Story
    • Status: In Progress
    • Major
    • Resolution: Unresolved
    • 3.4.0
    • None
    • PySpark
    • None

    Description

      This is a followup on https://issues.apache.org/jira/browse/SPARK-40211

      `scaleUpFactor` and `initialNumPartition` config are not supported yet in pyspark rdd take API

       

      (see https://github.com/apache/spark/blob/master/python/pyspark/rdd.py#L2799)

      basically it hardcoded `scaleUpFactor` as 1 and `initialNumPartition` as 4, therefore pyspark rdd take API is inconsistent with scala API.

       

      Anyone familiar with pyspark can help support this (referring to scala implementation)?

      Attachments

        Issue Links

          Activity

            People

              Unassigned Unassigned
              liuzq12 Ziqi Liu
              Votes:
              0 Vote for this issue
              Watchers:
              3 Start watching this issue

              Dates

                Created:
                Updated: