Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-24920

Spark should allow sharing netty's memory pools across all uses

Attach filesAttach ScreenshotVotersWatch issueWatchersCreate sub-taskLinkCloneUpdate Comment AuthorReplace String in CommentUpdate Comment VisibilityDelete Comments


    • Type: Improvement
    • Status: Resolved
    • Priority: Major
    • Resolution: Fixed
    • Affects Version/s: 2.4.0
    • Fix Version/s: 3.0.0
    • Component/s: Spark Core
    • Labels:


      Spark currently creates separate netty memory pools for each of the following "services":

      1) RPC Client
      2) RPC Server
      3) BlockTransfer Client
      4) BlockTransfer Server
      5) ExternalShuffle Client

      Depending on configuration and whether its an executor or driver JVM, different of these are active, but its always either 3 or 4.
      Having them independent somewhat defeats the purpose of using pools at all. In my experiments I've found each pool will grow due to a burst of activity in the related service (eg. task start / end msgs), followed another burst in a different service (eg. sending torrent broadcast blocks). Because of the way these pools work, they allocate memory in large chunks (16 MB by default) for each netty thread, so there is often a surge of 128 MB of allocated memory, even for really tiny messages. Also a lot of this memory is offheap by default, which makes it even tougher for users to manage.

      I think it would make more sense to combine all of these into a single pool. In some experiments I tried, this noticeably decreased memory usage, both onheap and offheap (no significant performance effect in my small experiments).

      As this is a pretty core change, as I first step I'd propose just exposing this as a conf, to let user experiment more broadly across a wider range of workloads




            • Assignee:
              attilapiros Attila Zsolt Piros
              irashid Imran Rashid


              • Created:

                Issue deployment