Details
-
Bug
-
Status: Open
-
Minor
-
Resolution: Unresolved
-
1.4.1
-
None
-
None
-
Spark 2.1
Description
Steps :
Huge data load performed. (table has 3.5 billion records)
Actual Issue :
One of the task failed during first load and restarted again in the same executor and took double the time to load first segment, rest of the 5 load ran properly.
Due to this failure and task rerun Segment_0 has got double the number of .carbondata files and even the number of records loaded twice.
Expected :
Task should not fail in 1st load. Even after failed task restarts carbondata files and records loaded should not double.