Description
Currently, Spark writes a single file out per task, sometimes leading to very large files. It would be great to have an option to limit the max number of records written per file in a task, to avoid humongous files.
This was initially suggested by simeons.
Attachments
Issue Links
- links to