Details
-
Sub-task
-
Status: Resolved
-
Major
-
Resolution: Fixed
-
3.3.5
Description
iceberg and hbase could benefit from being able to give a list of individual files to delete -files which may be scattered round the bucket for better read peformance.
Add some new optional interface for an object store which allows a caller to submit a list of paths to files to delete, where
the expectation is
- if a path is a file: delete
- if a path is a dir, outcome undefined
For s3 that'd let us build these into DeleteRequest objects, and submit, without any probes first.
Cherrypicking
when cherrypicking, you must include
- followup commit #6854
- https://issues.apache.org/jira/browse/HADOOP-19196
- test fixes HADOOP-19814 and
HADOOP-19188
Attachments
Issue Links
- causes
-
HADOOP-19184 TestStagingCommitter.testJobCommitFailure failing
- Resolved
-
HADOOP-19188 TestHarFileSystem and TestFilterFileSystem failing after bulk delete API added
- Resolved
-
HADOOP-19196 Bulk delete api doesn't take the path to delete as the base path
- Resolved
- is depended upon by
-
HADOOP-19131 WrappedIO to export modern filesystem/statistics APIs in a reflection friendly form
- Resolved
- is duplicated by
-
HADOOP-19191 Batch APIs for delete
- Resolved
- is related to
-
HADOOP-18722 Optimise S3A delete objects when multiObjectDelete is disabled
- Open
-
HADOOP-19254 Implement bulk delete command as hadoop fs command operation
- Open
-
HADOOP-19203 WrappedIO BulkDelete API to raise IOEs as UncheckedIOExceptions
- Resolved
- relates to
-
HADOOP-18420 Optimise S3A’s recursive delete to drop successful S3 keys on retry of S3 DeleteObjects
- Open
- links to