Description
after patch SPARK-17387 was applied – Sparkconf object is ignored when launching SparkContext programmatically via python from spark-submit:
https://github.com/apache/spark/blob/master/python/pyspark/context.py#L128:
in case when we are running python SparkContext(conf=xxx) from spark-submit:
conf is set, conf._jconf is None ()
passed as arg conf object is ignored (and used only when we are launching java_gateway).
how to fix:
python/pyspark/context.py:132
if conf is not None and conf._jconf is not None: # conf has been initialized in JVM properly, so use conf directly. This represent the # scenario that JVM has been launched before SparkConf is created (e.g. SparkContext is # created and then stopped, and we create a new SparkConf and new SparkContext again) self._conf = conf else: self._conf = SparkConf(_jvm=SparkContext._jvm) + if conf: + for key, value in conf.getAll(): + self._conf.set(key,value) + print(key,value)
Attachments
Attachments
Issue Links
- is duplicated by
-
SPARK-19369 SparkConf not getting properly initialized in PySpark 2.1.0
- Resolved
-
SPARK-20362 spark submit not considering user defined Configs (Pyspark)
- Resolved
- is related to
-
SPARK-19301 SparkContext is ignoring SparkConf when _jvm is not initialized on spark-submit
- Resolved
- relates to
-
SPARK-17387 Creating SparkContext() from python without spark-submit ignores user conf
- Resolved
- links to