Details
-
Improvement
-
Status: Closed
-
Major
-
Resolution: Later
-
None
-
None
-
None
Description
jgray and I were brainstorming some ideas about this:
In a typical heavy-write scenario, many store files do not last very long. They're flushed and then within a small number of seconds a compaction runs and they get deleted. For these "short lifetime" store files, it's less likely that a failure will occur during the window in which they're valid. So, I think we can consider some optimizations like the following:
- Flush files at replication count 2. Scan once a minute for any store files in the region that are older than 2 minutes. If they're found, increase their replication to 3. (alternatively, queue them to avoid scanning)
- More dangerous: flush files at replication count 1, but don't count them when figuring log expiration. So, if they get lost, we force log splitting to recover.
The performance gain here is that we avoid the network and disk transfer of writing the third replica for a file that we're just about to delete anyway.
Attachments
Attachments
Issue Links
- is related to
-
HBASE-2375 Revisit compaction configuration parameters
- Closed