Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-26020

shuffle data from spark streaming not cleaned up when External Shuffle Service is enabled

    XMLWordPrintableJSON

    Details

    • Type: Bug
    • Status: Open
    • Priority: Major
    • Resolution: Unresolved
    • Affects Version/s: 2.3.0
    • Fix Version/s: None
    • Component/s: Block Manager, Spark Core
    • Labels:
      None

      Description

      Hi, I'm running Spark Streaming on YARN and have enabled dynamic allocation + External Spark Shuffle Service. I'm noticing that during the lifetime of my spark streaming application, the nm appcache folder is building up with blockmgr directories (filled with shuffle_*.data).

      I understand why the data is not immediately cleaned up due to dynamic executor allocation, but will any cleanup of these directories be done during the lifetime of the spark streaming application ? Some of these shuffle data are generated as part of spark jobs/stages that have already completed.

      I've initially designed the application to run perpetually, but without any cleanup eventually the cluster will run out of disk and crash the application.

      https://stackoverflow.com/questions/52923386/spark-streaming-job-doesnt-delete-shuffle-files suggests a stop gap solution of cleaning up via cron.

      YARN-8991is the ticket I filed for YARN, who suggested me to file a ticket for spark. Appreciate any help.

       

        Attachments

          Issue Links

            Activity

              People

              • Assignee:
                Unassigned
                Reporter:
                teonadi Hidayat Teonadi
              • Votes:
                1 Vote for this issue
                Watchers:
                3 Start watching this issue

                Dates

                • Created:
                  Updated: