Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-1757

Support saving null primitives with .saveAsParquetFile()

    XMLWordPrintableJSON

Details

    • Bug
    • Status: Closed
    • Major
    • Resolution: Fixed
    • 1.0.0
    • 1.0.0
    • SQL
    • None

    Description

      See stack trace below:

      14/05/07 21:45:51 INFO analysis.Analyzer: Max iterations (2) reached for batch MultiInstanceRelations
      14/05/07 21:45:51 INFO analysis.Analyzer: Max iterations (2) reached for batch CaseInsensitiveAttributeReferences
      14/05/07 21:45:51 INFO optimizer.Optimizer$: Max iterations (2) reached for batch ConstantFolding
      14/05/07 21:45:51 INFO optimizer.Optimizer$: Max iterations (2) reached for batch Filter Pushdown
      java.lang.RuntimeException: Unsupported datatype StructType(List())
              at scala.sys.package$.error(package.scala:27)
              at org.apache.spark.sql.parquet.ParquetTypesConverter$.fromDataType(ParquetRelation.scala:201)
              at org.apache.spark.sql.parquet.ParquetTypesConverter$$anonfun$1.apply(ParquetRelation.scala:235)
              at org.apache.spark.sql.parquet.ParquetTypesConverter$$anonfun$1.apply(ParquetRelation.scala:235)
              at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
              at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
              at scala.collection.immutable.List.foreach(List.scala:318)
              at scala.collection.TraversableLike$class.map(TraversableLike.scala:244)
              at scala.collection.AbstractTraversable.map(Traversable.scala:105)
              at org.apache.spark.sql.parquet.ParquetTypesConverter$.convertFromAttributes(ParquetRelation.scala:234)
              at org.apache.spark.sql.parquet.ParquetTypesConverter$.writeMetaData(ParquetRelation.scala:267)
              at org.apache.spark.sql.parquet.ParquetRelation$.createEmpty(ParquetRelation.scala:143)
              at org.apache.spark.sql.parquet.ParquetRelation$.create(ParquetRelation.scala:122)
              at org.apache.spark.sql.execution.SparkStrategies$ParquetOperations$.apply(SparkStrategies.scala:139)
              at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
              at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
              at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
              at org.apache.spark.sql.catalyst.planning.QueryPlanner.apply(QueryPlanner.scala:59)
              at org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan$lzycompute(SQLContext.scala:264)
              at org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan(SQLContext.scala:264)
              at org.apache.spark.sql.SQLContext$QueryExecution.executedPlan$lzycompute(SQLContext.scala:265)
              at org.apache.spark.sql.SQLContext$QueryExecution.executedPlan(SQLContext.scala:265)
              at org.apache.spark.sql.SQLContext$QueryExecution.toRdd$lzycompute(SQLContext.scala:268)
              at org.apache.spark.sql.SQLContext$QueryExecution.toRdd(SQLContext.scala:268)
              at org.apache.spark.sql.SchemaRDDLike$class.saveAsParquetFile(SchemaRDDLike.scala:66)
              at org.apache.spark.sql.SchemaRDD.saveAsParquetFile(SchemaRDD.scala:96)
      

      Attachments

        Activity

          People

            Unassigned Unassigned
            aash Andrew Ash
            Votes:
            0 Vote for this issue
            Watchers:
            1 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved: