Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-25356

Add Parquet block size (row group size) option to SparkSQL configuration

    XMLWordPrintableJSON

Details

    • Improvement
    • Status: Resolved
    • Minor
    • Resolution: Invalid
    • 2.4.0
    • None
    • SQL
    • None

    Description

      I think we should configure the Parquet buffer size when using Parquet format.

      Because for HDFS, `dfs.block.size` is configurable, sometimes we hope the block size of parquet to be consistent with it.

      And  whether this parameter `spark.sql.files.maxPartitionBytes` is best consistent with the Parquet  block size when using Parquet format?

      Attachments

        Activity

          People

            Unassigned Unassigned
            10110346 liuxian
            Votes:
            0 Vote for this issue
            Watchers:
            3 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved: