Details
-
Bug
-
Status: Closed
-
Major
-
Resolution: Not A Problem
-
3.5.1
-
None
-
None
Description
When we use spark.jars.packages configuration parameter in Python SparkSession Builder (Pyspark), it appears that the filesystems are not loaded when session starts. Because of this, Spark fails to read file from Google Cloud Storage (GCS) bucket (with GCS Connector).
I tested this with different packages so it does not appear specific to a particular package. I will attach the sample code and debug logs.