Description
Spark's HiveServer2 registers a shutdown hook with the JVM Runtime.addShutdownHook() which can happen in parallel with the ShutdownHookManager sequence of spark & Hadoop, which operate the shutdowns in an ordered sequence.
This has some risks
- FS shutdown before rename of logs completes,
SPARK-6933 - Delays of rename on object stores may block FS close operation, which, on clusters with timeouts hooks (
HADOOP-12950) of FileSystem.closeAll() can force a kill of that shutdown hook and other problems.
General outcome: logs aren't present.
Proposed fix:
- register hook with org.apache.spark.util.ShutdownHookManager
HADOOP-15679to make shutdown wait time configurable, so O(data) renames don't trigger timeouts.
Attachments
Issue Links
- relates to
-
HADOOP-15679 ShutdownHookManager shutdown time needs to be configurable & extended
- Resolved
-
SPARK-6014 java.io.IOException: Filesystem is thrown when ctrl+c or ctrl+d spark-sql on YARN
- Resolved
-
SPARK-6933 Thrift Server couldn't strip .inprogress suffix after being stopped
- Closed
-
HADOOP-12950 ShutdownHookManager should have a timeout for each of the Registered shutdown hook
- Resolved
- links to