Details
-
Improvement
-
Status: Resolved
-
Major
-
Resolution: Fixed
-
4.0.0
Description
When having a table directory with invalid partitions such as:
table/ invalid/... part=1/... part=2/... part=3/...
a SQL query reading all of the partitions would fail with
java.lang.AssertionError: assertion failed: Conflicting directory structures detected. Suspicious paths: table table/invalid
I propose to add a data source option and Spark SQL config to ignore invalid partition paths. The config will be disabled by default to retain the current behaviour.
spark.conf.set("spark.sql.files.ignoreInvalidPartitionPaths", "true")
spark.read.format("parquet").option("ignoreInvalidPartitionPaths", "true").load(...)
Attachments
Issue Links
- links to