Details
-
Bug
-
Status: Open
-
Major
-
Resolution: Unresolved
-
1.18.1
-
None
-
None
-
amazon-linux-2023
Description
We have a job with small and static state size (states are updated instead of added), the job is configured to use RocksDB + full checkpointng (incremental disabled) because the diff between checkpoint is larger than full checkpoint size.
After migrating to 1.18, we observed significant and steady increase in full checkpoint size with RocksDB + full checkpointing. The increase was not observed with hashmap state backend.
I managed to reproduce the issue with following code:
StaticStateSizeGenerator115.java
StaticStateSizeGenerator118.java
Result:
On Flink 1.15, RocksDB + full checkpointing, checkpoint size is constant at 250KiB.
On Flink 1.18, RocksDB + full checkpointing, max checkpoint size got up to 38MiB before dropping (presumably due to compaction?)
On Flink 1.18, Hashmap statebackend, checkpoint size is constant at 219KiB.
Notes:
One observation I have is that the issue is more pronounced with higher parallelism, the code uses 8 parallelism. The production application that we first saw the regression got up to GiB of checkpoint size, where we only expected and observed (in 1.15) at most a couple of MiB.