Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-44718

High On-heap memory usage is detected while doing parquet-file reading with Off-Heap memory mode enabled on spark

    XMLWordPrintableJSON

Details

    • Improvement
    • Status: Resolved
    • Major
    • Resolution: Fixed
    • 3.4.1
    • 4.0.0
    • Spark Core, SQL
    • None
    • Patch

    Description

      I see the high use of on-heap memory usage while doing the parquet file reading when the off-heap memory mode is enabled. This is caused by the memory-mode for the column vector for the vectorized reader is configured by different flag, and the default value is always set to On-Heap.

      Conf to reproduce the issue:

      spark.memory.offHeap.size 1000000
      spark.memory.offHeap.enabled true

      Enabling these configurations only will not change the memory mode used for parquet-reading by the vectorized reader to Off-Heap.

       

      Proposed PR: https://github.com/apache/spark/pull/42394

      Attachments

        Activity

          People

            majdyz Zamil Majdy
            majdyz Zamil Majdy
            Votes:
            0 Vote for this issue
            Watchers:
            4 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved: