Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-24920

Spark should allow sharing netty's memory pools across all uses

    XMLWordPrintableJSON

    Details

    • Type: Improvement
    • Status: Resolved
    • Priority: Major
    • Resolution: Fixed
    • Affects Version/s: 2.4.0
    • Fix Version/s: 3.0.0
    • Component/s: Spark Core
    • Labels:

      Description

      Spark currently creates separate netty memory pools for each of the following "services":

      1) RPC Client
      2) RPC Server
      3) BlockTransfer Client
      4) BlockTransfer Server
      5) ExternalShuffle Client

      Depending on configuration and whether its an executor or driver JVM, different of these are active, but its always either 3 or 4.
      Having them independent somewhat defeats the purpose of using pools at all. In my experiments I've found each pool will grow due to a burst of activity in the related service (eg. task start / end msgs), followed another burst in a different service (eg. sending torrent broadcast blocks). Because of the way these pools work, they allocate memory in large chunks (16 MB by default) for each netty thread, so there is often a surge of 128 MB of allocated memory, even for really tiny messages. Also a lot of this memory is offheap by default, which makes it even tougher for users to manage.

      I think it would make more sense to combine all of these into a single pool. In some experiments I tried, this noticeably decreased memory usage, both onheap and offheap (no significant performance effect in my small experiments).

      As this is a pretty core change, as I first step I'd propose just exposing this as a conf, to let user experiment more broadly across a wider range of workloads

        Attachments

          Issue Links

            Activity

              People

              • Assignee:
                attilapiros Attila Zsolt Piros
                Reporter:
                irashid Imran Rashid
              • Votes:
                1 Vote for this issue
                Watchers:
                8 Start watching this issue

                Dates

                • Created:
                  Updated:
                  Resolved: