Uploaded image for project: 'Cassandra'
  1. Cassandra
  2. CASSANDRA-7184

improvement of SizeTieredCompaction

    XMLWordPrintableJSON

Details

    • Improvement
    • Status: Open
    • Low
    • Resolution: Unresolved
    • None
    • Local/Compaction

    Description

      1, In our usage scenario, there is no duplicated insert and no delete . The data increased all the time, and some big sstables are generated (100GB for example). we don't want these sstables to participate in the SizeTieredCompaction any more. so we add a max threshold which is set to 100GB . Sstables larger than the threshold will not be compacted. Should this strategy be added to the trunk ?

      2, In our usage scenario, maybe hundreds of sstable need to be compacted in a Major Compaction. The total size would be larger to 5TB. So during the compaction, when the size writed reach to a configed threshhold(200GB for example), it switch to write a new sstable. In this way, we avoid to generate too huge sstables. Too huge sstable have some bad infuence:
      (1) It will be larger than the capacity of a disk;
      (2) If the sstable is corrupt, lots of objects will be influenced .
      Should this strategy be added to the trunk ?

      Attachments

        Activity

          People

            Jianwei Zhang Jianwei Zhang
            Jianwei Zhang Jianwei Zhang
            Jianwei Zhang
            Jianwei Zhang Jianwei Zhang
            Votes:
            0 Vote for this issue
            Watchers:
            2 Start watching this issue

            Dates

              Created:
              Updated:

              Time Tracking

                Estimated:
                Original Estimate - 48h
                48h
                Remaining:
                Remaining Estimate - 48h
                48h
                Logged:
                Time Spent - Not Specified
                Not Specified