Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-24273

Failure while using .checkpoint method to private S3 store via S3A connector

    XMLWordPrintableJSON

Details

    • Bug
    • Status: Resolved
    • Major
    • Resolution: Incomplete
    • 2.3.0
    • None
    • Spark Shell

    Description

      We are getting following error:

      com.amazonaws.services.s3.model.AmazonS3Exception: Status Code: 416, AWS Service: Amazon S3, AWS Request ID: tx000000000000000014126-005ae9bfd9-9ed9ac2-default, AWS Error Code: InvalidRange, AWS Error Message: null, S3 Extended Request ID: 9ed9ac2-default-default"
      

      when we use checkpoint method as below.

      val streamBucketDF = streamPacketDeltaDF
       .filter('timeDelta > maxGap && 'timeDelta <= 30000)
       .withColumn("bucket", when('timeDelta <= mediumGap, "medium")
       .otherwise("large")
       )
       .checkpoint()
      

      Do you have idea how to prevent invalid range in header to be sent, or how it can be workarounded or fixed?

      Thanks.

      Attachments

        Issue Links

          Activity

            People

              Unassigned Unassigned
              jimhox Jami Malikzade
              Votes:
              0 Vote for this issue
              Watchers:
              3 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: