You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
What happened:
I have a call to write_deltalake that successfully writes out a delta table, but it doesn't seem to respect what is input through the writer_properties parameter. In this case, I'm wanting to specify that it used ZSTD level 3 compression, but the output does not get compressed at all (and there is no .zstd.parquet file extensions).
What you expected to happen:
The files produced should be compressed with ZSTD level 3 compression. All other parameters in the code snippet below are confirmed to be working, including partitioning.
You can either write with WriterProperties using engine='rust' or you pass the ds.ParquetFileWriteOptions to file_options when you also set engine='pyarrow' :)
@nholt01 the normal write without partitioning with rust engine I saw sometimes 3-4x faster writes.
With partitioning I saw similar speeds between pyarrow and rust engine.
I suggest to use the rust engine writer, this will be the way forward for the library and all the newer protocol versions will only supported with that one (for example constraints).
The only thing that is missing is predicate overwrite in rust but this is on its way
Environment
Delta-rs version:
Binding: python 0.15.1
Environment:
Bug
What happened:
I have a call to
write_deltalake
that successfully writes out a delta table, but it doesn't seem to respect what is input through thewriter_properties
parameter. In this case, I'm wanting to specify that it used ZSTD level 3 compression, but the output does not get compressed at all (and there is no .zstd.parquet file extensions).What you expected to happen:
The files produced should be compressed with ZSTD level 3 compression. All other parameters in the code snippet below are confirmed to be working, including partitioning.
How to reproduce it:
More details:
The text was updated successfully, but these errors were encountered: