Details
-
Bug
-
Status: Resolved
-
Major
-
Resolution: Fixed
-
2.4.8, 3.0.1, 3.1.0
-
None
Description
Spark uses the current timestamp to generate a MapReduce JobID.
If > 1 job attempt is generated in the same second, these can clash
Committers which expect this to be unique can conflict with the other jobs
- S3A staging committer (cluster FS staging dir and local task output dir)
- Any committer which supports parallel jobs writing to the same destination
directory and requires unique names for the attempts - Code which uses the jobID as part of its algorithm to generate unique filenames
Note: HadoopMapReduceCommitProtocol.getFilename() doesn't use this JobID for
uniqueness, it uses task attempt ID and stage ID. It probably deserves its own
audit.
Attachments
Issue Links
- relates to
-
HADOOP-17318 S3A committer to support concurrent jobs with same app attempt ID & dest dir
- Resolved
-
SPARK-33230 FileOutputWriter jobs have duplicate JobIDs if launched in same second
- Resolved
- links to