Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-25827

Replicating a block > 2gb with encryption fails

Attach filesAttach ScreenshotVotersWatch issueWatchersCreate sub-taskLinkCloneUpdate Comment AuthorReplace String in CommentUpdate Comment VisibilityDelete Comments
    XMLWordPrintableJSON

Details

    • Bug
    • Status: Resolved
    • Major
    • Resolution: Fixed
    • 2.4.0
    • 2.4.1, 3.0.0
    • Spark Core
    • None

    Description

      There are a couple of issues with replicating & remote reads of large encrypted blocks, which try to create buffers where they shouldn't. Some of this is properly limiting the size of arrays under SPARK-25904, but there are others specific to encryption & trying to convert EncryptedBlockData into a regular ByteBuffer.

      EDIT: moved general array size stuff under SPARK-25904.

      Attachments

        Issue Links

        Activity

          This comment will be Viewable by All Users Viewable by All Users
          Cancel

          People

            irashid Imran Rashid
            irashid Imran Rashid
            Votes:
            0 Vote for this issue
            Watchers:
            3 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved:

              Slack

                Issue deployment