Details
-
Improvement
-
Status: Closed
-
Major
-
Resolution: Fixed
-
3.0.2-incubating
-
None
Description
If a provider is using Spark, they are currently forced to have HDFS be used to store intermediate RDD data. However, if they plan on using that data in a GraphComputer "job chain," then they should be able to lookup a .cached() RDD by name.
Create a inputGraphRDD.name and outputGraphRDD.name to make it so that the configuration references SparkContext.getPersitedRDDs().