Details
-
Bug
-
Status: Resolved
-
Major
-
Resolution: Invalid
-
2.3.0
-
None
-
None
Description
Hi All,
Spark Structured Streaming doesn't allow two structured streaming jobs to write data to the same base directory which is possible with using dstreams.
As _spark_metadata directory will be created by default for one job, second job cannot use the same directory as base path as already _spark_metadata directory is created by other job, It is throwing exception.
Is there any workaround for this, other than creating separate base path's for both the jobs.
Is it possible to create the _spark_metadata directory else where or disable without any data loss.
If I had to change the base path for both the jobs, then my whole framework will get impacted, So i don't want to do that.
Attachments
Issue Links
- is fixed by
-
SPARK-43152 User-defined output metadata path (_spark_metadata)
- Open