Details
-
Improvement
-
Status: Resolved
-
Major
-
Resolution: Fixed
-
None
-
None
-
Reviewed
Description
In branch-1, BucketCache just allocate new onheap bytebuffer to construct new HFileBlock when get cached blocks. This rough allocation increases the GC pressure for those "hot" blocks.
Here introduce a RAMBuffer for those "hot" blocks in BucketCache. The thought is simple. The RAMBuffer is an timeout expiring cache. When a Multi-level block is read twice, we cache it in the RAMBuffer. When the block timeout in the cache (e.g. 60s), that means the block is not being accessed in 60s, we evict it. Not like LRU, we do not cache block when the whole RAMBuffer size reaches to a threshold (to fit different workload, the threshold is dynamic). This will prevent the RAMBuffer from being churned.
I also did a YCSB performance test.
The circumstance is:
Size of BucketCache: 40 GB
Target table size: 112 GB
Properties:
The operation distribution of YCSB workload is latest.
Client Side Metrics
See the attachment ClientSideMetrics.png
Server Side GC:
The current bucket cache triggered 217 GCs, which costs 2.74 minutes in total.
With RAMBuffer, the server side had 210 times GC and 2.56 minutes in total.
As the master & branch-2 using ByteBufferAllocator to manage the bucketcache memory allocation, the RAMBuffer may not have GC improvement as much as branch-1.
Attachments
Attachments
Issue Links
- links to