Uploaded image for project: 'Zeppelin'
  1. Zeppelin
  2. ZEPPELIN-18

Running pyspark without deploying python libraries to every yarn node

    XMLWordPrintableJSON

Details

    • Improvement
    • Status: Resolved
    • Major
    • Resolution: Fixed
    • None
    • 0.5.0
    • None
    • None

    Description

      We only use python sources under SPARK_HOME/python for PySpark. I think copying these files into /python is better idea because it doesn't need to extract - or build - spark no longer. If PySpark has changed by upgrading Spark version, It's easy to follow up by coping new python files into zeppelin. If we should support multiple version of spark, we can solve it of having multiple python files.

      Attachments

        1. Screen Shot 2015-06-19 at 3.20.18 PM.png
          18 kB
          Jongyoul Lee
        2. Screen Shot 2015-06-23 at 7.39.19 PM.png
          58 kB
          Jongyoul Lee
        3. Screen Shot 2015-06-23 at 7.39.44 PM.png
          51 kB
          Jongyoul Lee

        Issue Links

          Activity

            People

              jongyoul Jongyoul Lee
              jongyoul Jongyoul Lee
              Votes:
              0 Vote for this issue
              Watchers:
              3 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: