Details
-
Bug
-
Status: Resolved
-
Major
-
Resolution: Not A Problem
-
2.1.1, 2.2.0
-
None
-
None
Description
Starting with Spark 2.1.1 I observed this bug. Here's are the steps to reproduce it:
- create a DF
- persist it
- count the items in it
- create a new DF as a transformation of the previous one
- persist it
- count the items in it
- unpersist the first DF
Once you do that you will see that also the 2nd DF is gone.
The code to reproduce it is:
val x1 = Seq(1).toDF() x1.persist() x1.count() assert(x1.storageLevel.useMemory) val x11 = x1.select($"value" * 2) x11.persist() x11.count() assert(x11.storageLevel.useMemory) x1.unpersist() assert(!x1.storageLevel.useMemory) //the following assertion FAILS assert(x11.storageLevel.useMemory)
Attachments
Issue Links
- relates to
-
SPARK-24596 Non-cascading Cache Invalidation
- Resolved
- links to