You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The main reason is that for now we have a job setup procedure which prepares the environment inside a job container (spyt-package/src/main/bash/setup-spyt-env.sh). This script uses spark and spyt archives from cypress and extracts them inside a contained during job startup. We consider refactoring it to decrease the startup time for vanilla operations.
The sources of spyt docker image can be found here: tools/release/spyt_image/Dockerfile. It contains a spyt distributive archive inside which is uploaded to cypress. The spark distributive is taken from the Internet from the official Spark site.
The release distributives is rather stable so you don't have to include it in your custom docker images.
Why do we need update spyt on yt cluster, if we use the custom image?
If spyt version not in yt cluster, I have error spyt version not in cluster
Docker image
Create a cluster with a docker image
spark-launch-yt \ --spark-cluster-version 2.3.0 \ --params '{operation_spec={tasks={history={docker_image="MY_DOCKER_IMAGE"};master={docker_image="MY_DOCKER_IMAGE"};workers={docker_image="MY_DOCKER_IMAGE"}}}}'
The text was updated successfully, but these errors were encountered: