Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
fix: Fix materialization when running on Spark cluster. (feast-dev#3166)
* Fix materialization when running on Spark cluster. When running materialization and have Spark offline store configured to use cluster (`spark.master` pointing to actual Spark master node) `self.to_spark_df().write.parquet(temp_dir, mode="overwrite")` will create parquet file in worker node but `return pq.read_table(temp_dir)` is executed on driver node and it can't read from worker. Proposed fix makes materialization work when run on Spark cluster. Signed-off-by: ckarwicki <[email protected]> Signed-off-by: ckarwicki <[email protected]> * Fix linter. Signed-off-by: ckarwicki <[email protected]> Signed-off-by: ckarwicki <[email protected]> Signed-off-by: ckarwicki <[email protected]> * Fix linter. Signed-off-by: ckarwicki <[email protected]> Signed-off-by: ckarwicki <[email protected]> Signed-off-by: ckarwicki <[email protected]> Signed-off-by: ckarwicki <[email protected]> Signed-off-by: ckarwicki <[email protected]> Signed-off-by: ckarwicki <[email protected]>
- Loading branch information