Uploaded image for project: 'Apache Arrow'
  1. Apache Arrow
  2. ARROW-7156

[C#] Large record batch is written with negative buffer length

    XMLWordPrintableJSON

Details

    • Bug
    • Status: Open
    • Major
    • Resolution: Unresolved
    • 0.14.1, 0.15.1
    • None
    • C#
    • None

    Description

      I have a 30 gig arrow file with 100 batches.  the largest batch in the file causes get batch to fail - All other batches load fine. in 14.11 the individual batch errors.. in 15.1.1 the batch crashes R studio when it is used

      14.1.1

      >  rbn <- data_rbfr$get_batch(x)
      Error in ipc__RecordBatchFileReader_ReadRecordBatch(self, i) : 
      Invalid: negative malloc size
        

      15.1.1

      rbn <- data_rbfr$get_batch(x)  works!
      df <- as.data.frame(rbn) - Crashes R Studio! 

       

      Update

      I put the data in the batch into a separate file.  The file size is over 2 gigs. 

      Using 15.1.1, when I try to load this entire file via read_arrow it also fails.

      ar <- arrow::read_arrow("e:\\temp\\file.arrow") 
      Error in Table__from_RecordBatchFileReader(batch_reader) :
       Invalid: negative malloc size

       

      Attachments

        1. SingleBatch_String_85000_Rows.crash.rar
          2.54 MB
          Anthony Abate
        2. SingleBatch_String_70000_Rows.ok.rar
          2.06 MB
          Anthony Abate
        3. image-2019-11-13-16-27-30-641.png
          6 kB
          Anthony Abate

        Activity

          People

            Unassigned Unassigned
            abbot Anthony Abate
            Votes:
            0 Vote for this issue
            Watchers:
            4 Start watching this issue

            Dates

              Created:
              Updated: