Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-18207

class "org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificUnsafeProjection" grows beyond 64 KB

    XMLWordPrintableJSON

Details

    • Bug
    • Status: Resolved
    • Major
    • Resolution: Fixed
    • 2.0.1, 2.1.0
    • 2.1.0
    • Optimizer, SQL
    • None

    Description

      I have 2 wide dataframes that contain nested data structures, when I explode one of the dataframes, it doesn't include records with an empty nested structure (outer explode not supported). So, I create a similar dataframe with null values and union them together. See SPARK-13721 for more details as to why I have to do this.

      I was hoping that SPARK-16845 was going to address my issue, but it does not. I was asked by lwlin to open this JIRA.

      I will attach a code snippet that can be pasted into spark-shell that duplicates my code and the exception. This worked just fine in Spark 1.6.x.

      org.apache.spark.SparkException: Job aborted due to stage failure: Task 35 in stage 5.0 failed 4 times, most recent failure: Lost task 35.3 in stage 5.0 (TID 812, somehost.mydomain.com, executor 8): java.util.concurrent.ExecutionException: java.lang.Exception: failed to compile: org.codehaus.janino.JaninoRuntimeException: Code of method "apply(Lorg/apache/spark/sql/catalyst/InternalRow;)Lorg/apache/spark/sql/catalyst/expressions/UnsafeRow;" of class "org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificUnsafeProjection" grows beyond 64 KB
      

      Attachments

        1. spark-18207.txt
          19 kB
          Don Drake

        Issue Links

          Activity

            People

              kiszk Kazuaki Ishizaki
              dondrake Don Drake
              Michael Armbrust Michael Armbrust
              Votes:
              0 Vote for this issue
              Watchers:
              4 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: