Uploaded image for project: 'TinkerPop'
  1. TinkerPop
  2. TINKERPOP-1177

Improve documentation around Spark's storage levels

    XMLWordPrintableJSON

Details

    Description

      We should add a big warning regarding the two storage level settings, especially since we have MEMORY_ONLY as the default value. What's the impact of changing any of these settings? What is kept in memory, what not? How much memory will my executors need when I use one or another storage level?

      Currently these 2 settings mean nothing to the average user (you'll probably only know how to use them, if you have a deep understanding of how Spark works). For the averaage user we should probably also consider to make DISK_ONLY (or perhaps MEMORY_AND_DISK?) the default.

      Attachments

        Activity

          People

            dkuppitz Daniel Kuppitz
            dkuppitz Daniel Kuppitz
            Votes:
            0 Vote for this issue
            Watchers:
            4 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved: