Details
-
Improvement
-
Status: Resolved
-
Major
-
Resolution: Fixed
-
4.0.0
Description
In order to support developing spark streaming sink in python, we need to implement python stream writer interface.
Reuse PythonPartitionWriter to implement the serialization and execution of write callback in executor.
Implement python worker process to run python streaming data sink committer and communicate with JVM through socket in spark driver. For each python streaming data sink instance there will be a long live python worker process created. Inside the python process, the python write committer will receive abort or commit function call and send back result through socket.
Attachments
Issue Links
- links to