Replies: 1 comment 3 replies
-
It is not possible to share distributed objects like DataFrame between 2 different Spark applications, since an RDD/DataFrame is tied to the context that was used to create it. The only way to do this is to have the first application write your dataframe to a disk/storage account and then load it from the second application. |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I run Apache Spark as a standalone.
I have 2 Spark applications and these connected to Apache Spark with master URL:
The first application reads data from CSV and shares this data frame in some way
DataFrame dataFrame = spark.Read().Text("input.txt");
The second application needs to access this shared data frame.
How can I do this? Is it possible?
Beta Was this translation helpful? Give feedback.
All reactions