Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-3008

PySpark fails due to zipimport not able to load the assembly jar (/usr/bin/python: No module named pyspark)

Log workAgile BoardRank to TopRank to BottomAttach filesAttach ScreenshotVotersWatch issueWatchersCreate sub-taskConvert to sub-taskLinkCloneLabelsUpdate Comment AuthorReplace String in CommentUpdate Comment VisibilityDelete Comments
    XMLWordPrintableJSON

    Details

    • Type: Bug
    • Status: Closed
    • Priority: Major
    • Resolution: Duplicate
    • Affects Version/s: None
    • Fix Version/s: None
    • Component/s: PySpark
    • Labels:
    • Environment:
    • Target Version/s:

      Description

      PySpark is not working. It fails because zipimport not able to import assembly jar because that contain more than 65536 files.

      Email chains in this regard are below

      http://mail-archives.apache.org/mod_mbox/incubator-spark-user/201406.mbox/%3CCAMJOb8kcGk0PqiOGJu6UoKCeysWCuSW3xwd5wRs8ikpMgD2DAg@mail.gmail.com%3E

      https://mail.python.org/pipermail/python-list/2014-May/671353.html

      Is there any work around to bypass the issue ?

        Attachments

        Issue Links

          Activity

          $i18n.getText('security.level.explanation', $currentSelection) Viewable by All Users
          Cancel

            People

            • Assignee:
              Unassigned Assign to me
              Reporter:
              flukebox Jai Kumar Singh

              Dates

              • Created:
                Updated:
                Resolved:

                Issue deployment