Uploaded image for project: 'Flink'
  1. Flink
  2. FLINK-21251

Last valid checkpoint metadata lost after job exits restart loop

    XMLWordPrintableJSON

Details

    • Bug
    • Status: Closed
    • Critical
    • Resolution: Not A Problem
    • 1.7.2
    • None
    • None

    Description

      We have a Flink job of a relatively old version, 1.7.1, that failed with no valid checkpoint to restore. The job was first affected by a Kafka network instability and fell into the restart loop with the policy of 3 restarts in 5 minutes. After the restarts exhausted, the job turned into the final state FAILED and exits. But the problem is that the last valid checkpoint 4585 that was restored multiple times during the restarts, was corrupted (no _metadata) after the job exited. 

      I've checked the checkpoint dir on HDFS and found that chk-4585 which was finished at 12:16 was modified at 12:23 when jobmanager was shutting down with lots of error logs saying the deletes of pending checkpoints somehow failed. So I'm suspecting that the checkpoint metadata was unexpectedly deleted by jobmanager.

      The jobmanager logs are attached below.

      Attachments

        1. ch-4585 content.png
          416 kB
          Paul Lin
        2. checkpoint dir.png
          295 kB
          Paul Lin
        3. jm_logs
          2.09 MB
          Paul Lin

        Activity

          People

            Unassigned Unassigned
            Paul Lin Paul Lin
            Votes:
            0 Vote for this issue
            Watchers:
            2 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved: