Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-3613

Don't record the size of each shuffle block for large jobs

    XMLWordPrintableJSON

Details

    • Improvement
    • Status: Resolved
    • Major
    • Resolution: Fixed
    • None
    • 1.2.0
    • Shuffle, Spark Core
    • None

    Description

      MapStatus saves the size of each block (1 byte per block) for a particular map task. This actually means the shuffle metadata is O(M*R), where M = num maps and R = num reduces.

      If M is greater than a certain size, we should probably just send an average size instead of a whole array.

      Attachments

        Issue Links

          Activity

            People

              rxin Reynold Xin
              rxin Reynold Xin
              Votes:
              0 Vote for this issue
              Watchers:
              3 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: