Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-4838

StackOverflowError when serialization task

    XMLWordPrintableJSON

    Details

    • Type: Bug
    • Status: Resolved
    • Priority: Major
    • Resolution: Cannot Reproduce
    • Affects Version/s: 1.1.0
    • Fix Version/s: None
    • Component/s: Spark Core, SQL
    • Labels:
      None

      Description

      When run a sql with more than 2000 partitions, each partition a HadoopRDD, it will cause java.lang.StackOverflowError at serialize task.
      Error message from spark is:Job aborted due to stage failure: Task serialization failed: java.lang.StackOverflowError
      java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1508)
      java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1431)
      java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1177)
      java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1547)
      java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1508)
      java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1431)
      java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1177)
      ......

        Attachments

          Issue Links

            Activity

              People

              • Assignee:
                Unassigned
                Reporter:
                shenhong Hong Shen
              • Votes:
                0 Vote for this issue
                Watchers:
                2 Start watching this issue

                Dates

                • Created:
                  Updated:
                  Resolved: