Details
-
Bug
-
Status: Closed
-
Major
-
Resolution: Fixed
-
None
Description
https://github.com/apache/hive/blob/cdd55aa319a3440963a886ebfff11cd2a240781d/ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java#L1952-L2010
compareTempOrDuplicateFiles uses a combination of attemptId and fileSize to determine which file(s) to keep.
I've seen instances where this function throws an exception due to the fact that the newer attemptId file size is less than the older attemptId (thus failing the query).
I think this assumption is faulty, due to various factors such as file compression and the order in which values are written. It may be prudent to trust that the newest attemptId is in fact the best choice.
Attachments
Attachments
Issue Links
- causes
-
HIVE-23614 Always pass HiveConfig to removeTempOrDuplicateFiles
- Closed
- links to