Details
-
Bug
-
Status: Resolved
-
Low
-
Resolution: Duplicate
-
None
-
None
-
None
-
Low
Description
When loading large amounts of data, currently the BulkRecordWriter will write out all the sstables, then stream them. This actually caused us GC overhead issues, due to our heap sizes for reducers. We ran into a problem where the number of SSTables on disk that had to be open would cause the jvm process to die. We also wanted a way to incrementally stream them as we created them. I created support for setting this, the default behavior is wait for them to be created. But if you increase to >= 1, you can determine the batch size.
Attachments
Issue Links
- is blocked by
-
CASSANDRA-3859 Add Progress Reporting to Cassandra OutputFormats
- Resolved