Details
-
New Feature
-
Status: Resolved
-
Minor
-
Resolution: Won't Fix
-
3.1.0
-
None
-
None
Description
We need a mechanism to override the datasource implementation via configuration.
For example, suppose I have a custom CSV datasource implementation called "my_csv". One way to use it is:
val df = spark.read.format("my_csv").load(...)
Since the source data is the same format (CSV), you should be able to override the default implementation.
One proposal is to do the following:
spark.conf.set("spark.sql.datasource.override.csv", "my_csv") val df = spark.read.csv(...)
This has a benefit that the user doesn't have to change the application code to try out a new datasource implementation for the same source format.