Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-673

PySpark should capture and re-throw Python exceptions

    XMLWordPrintableJSON

Details

    • Improvement
    • Status: Resolved
    • Major
    • Resolution: Fixed
    • None
    • 0.7.0
    • PySpark
    • None

    Description

      Right now if there is an exception inside of a PySpark worker, it causes the worker process to exit prematurely, triggering an EOF exception at the JVM worker. This means you have to go dig through worker logs to find the exception trace.

      It would be more helpful if the Python worker instead caught the exception and passed the string representation of the exception to the JVM worker, which could then wrap it in a Java exception. e.g.

      throw new PythonException(exnString).
      

      This would make it much easier to debug python tasks, since that string would show up at the driver.

      Attachments

        Activity

          People

            patrick Patrick McFadin
            patrick Patrick McFadin
            Votes:
            0 Vote for this issue
            Watchers:
            2 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved: