Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-48217

Spark stdout and stderr getting removed at end of spark job triggered from cloudera hue workflow

    XMLWordPrintableJSON

Details

    • Improvement
    • Status: Open
    • Critical
    • Resolution: Unresolved
    • 1.6.0
    • None
    • Spark Submit
    • None

    Description

      Hello,

      we are running spark job triggered from cloudera hue workflow

      and spark printing stdout and stderr logs during execution as expected:

      e.g. -

       

      But stdout and stderr logs getting cleaned when workflows finished/status succeeded

       

      following is spark-submit command workflow is triggering:

      _/usr/bin/spark-submit _
        _--master yarn-client _
        _--driver-memory 4g _
        _--executor-memory 16g _
        _--executor-cores 4 _
        _--class tst _
        _--files `ls -m *.conf | tr -d '\n '` _
        _--conf "spark.dynamicAllocation.maxExecutors=4" _
        _--conf "spark.kryoserializer.buffer.max=1024" _
        tst.jar $*

       

       

      does someone familiar with this spark job behavior or can advise ideas to fix it?

       

      Thanks in advance

      Attachments

        1. workflow_running_logs_printed.png
          78 kB
          Noam Shemesh
        2. workflow_succeeded_logs_cleaned.png
          55 kB
          Noam Shemesh

        Activity

          People

            Unassigned Unassigned
            noamsh88 Noam Shemesh
            Votes:
            0 Vote for this issue
            Watchers:
            1 Start watching this issue

            Dates

              Created:
              Updated: