Uploaded image for project: 'Cassandra'
  1. Cassandra
  2. CASSANDRA-3945

Support incremental/batch sizes for BulkRecordWriter, due to GC overhead issues

    Details

    • Type: Bug
    • Status: Resolved
    • Priority: Minor
    • Resolution: Duplicate
    • Fix Version/s: None
    • Component/s: None
    • Labels:
      None

      Description

      When loading large amounts of data, currently the BulkRecordWriter will write out all the sstables, then stream them. This actually caused us GC overhead issues, due to our heap sizes for reducers. We ran into a problem where the number of SSTables on disk that had to be open would cause the jvm process to die. We also wanted a way to incrementally stream them as we created them. I created support for setting this, the default behavior is wait for them to be created. But if you increase to >= 1, you can determine the batch size.

        Attachments

          Issue Links

            Activity

              People

              • Assignee:
                Unassigned
                Reporter:
                lenn0x Chris Goffinet
              • Votes:
                0 Vote for this issue
                Watchers:
                0 Start watching this issue

                Dates

                • Created:
                  Updated:
                  Resolved: