Details
-
New Feature
-
Status: Closed
-
Major
-
Resolution: Fixed
-
None
-
None
-
None
-
Incompatible change, Reviewed
-
Description
Traditional message digest algorithms, like MD5, SHA1, etc., require reading the entire input message sequentially in a central location. HDFS supports large files with multiple tera bytes. The overhead of reading the entire file is huge. A distributed file checksum algorithm is needed for HDFS.
Attachments
Attachments
Issue Links
- is blocked by
-
HADOOP-3941 Extend FileSystem API to return file-checksums/file-digests
- Closed
- is related to
-
HADOOP-4197 Need to update DATA_TRANSFER_VERSION
- Closed
-
HADOOP-4176 Implement getFileChecksum(Path) in HftpFileSystem
- Closed