Integration with Pyspark #148
Replies: 2 comments 1 reply
-
Hey @jacksongoode. I've managed to get blaze running locally with pyspark, as well as on a Databricks cluster. For Databricks, you'd need to do the following:
Once you do this the first time, it becomes quicker to update the jar for new changes. You probably won't be able to run (m)any queries though, as you'll encounter #174 when trying to read even plain parquet files from the cluster. For a local setup, you can copy the jar into the spark jars folder, then set the config as part of |
Beta Was this translation helpful? Give feedback.
-
@nevi-me Hi Nevi, it's been a minute! Love to know if these instructions still hold or if it's gotten easier? |
Beta Was this translation helpful? Give feedback.
-
Just wondering how one would go about integrating this project within a Pyspark environment? We are also using Databricks. What configuration would be needed? Just the jar file? Thanks!
Beta Was this translation helpful? Give feedback.
All reactions