Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-30821

Executor pods with multiple containers will not be rescheduled unless all containers fail

    XMLWordPrintableJSON

    Details

    • Type: Improvement
    • Status: In Progress
    • Priority: Major
    • Resolution: Unresolved
    • Affects Version/s: 3.1.0
    • Fix Version/s: None
    • Component/s: Kubernetes, Spark Core
    • Labels:
      None

      Description

      Since the restart policy of launched pods is Never, additional handling is required for pods that may have sidecar containers. The executor should be considered failed if any containers have terminated and have a non-zero exit code, but Spark currently only checks the pod phase. The pod phase will remain "running" as long as any pods are still running. Kubernetes sidecar support in 1.18/1.19 does not address this situation, as sidecar containers are excluded from pod phase calculation.

        Attachments

          Issue Links

            Activity

              People

              • Assignee:
                apachespark Apache Spark
                Reporter:
                hogeland Kevin Hogeland
              • Votes:
                0 Vote for this issue
                Watchers:
                2 Start watching this issue

                Dates

                • Created:
                  Updated: