Skip to content

Microsoft.Spark.Worker longevity #852

Answered by imback82
dbeavon asked this question in Q&A
Discussion options

You must be logged in to vote

Are you running this on Windows? If so, worker needs to be respawned for each task: https://github.com/apache/spark/blob/2fa792aa64f6153af1641d895e2f996b18dfbce4/core/src/main/scala/org/apache/spark/api/python/PythonWorkerFactory.scala#L48

On non-Windows, one worker is launched per executor and shared across tasks.

Note that we are piggybacking on how Python worker is invoked, there is not much we can do to improve this process unless we modify the OSS Spark.

Replies: 1 comment 5 replies

Comment options

You must be logged in to vote
5 replies
@dbeavon
Comment options

@imback82
Comment options

@dbeavon
Comment options

@imback82
Comment options

@dbeavon
Comment options

Answer selected by dbeavon
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants