Details
-
Bug
-
Status: Resolved
-
Major
-
Resolution: Duplicate
-
1.3.0
-
None
-
None
Description
Because py4j uses the system ClassLoader instead of the contextClassLoader of the thread, the dynamically added jars in Spark Submit can't be loaded in the driver.
This causes `Py4JError: Trying to call a package` errors.
Usually `-packages` are downloaded from some remote repo before runtime, adding them explicitly to `driver-class-path` is not an option, like we can do with `jars`. One solution is to move the fetching of `-packages` to the SparkSubmitDriverBootstrapper, and add it to the driver class-path there.
A more complete solution can be achieved through SPARK-4924.
Attachments
Issue Links
- duplicates
-
SPARK-5185 pyspark --jars does not add classes to driver class path
- Resolved
- incorporates
-
SPARK-5185 pyspark --jars does not add classes to driver class path
- Resolved
- is related to
-
SPARK-4924 Factor out code to launch Spark applications into a separate library
- Resolved
- links to