You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
What happened: When I perform an optimize.compact() operation following an append to a table, I encounter an exception. Oddly, this optimize.compact() succeeds after the initial append to an empty table but fails after subsequent appends. This is the exception I get:
_internal.DeltaError: Data does not match the schema or partitions of the table: Unexpected Arrow schema: got: Field { name: "name", data_type: Utf8, nullable: true, dict_id: 0, dict_is_ordered: false, metadata: {} }, Field { name: "surname", data_type: Utf8, nullable: true, dict_id: 0, dict_is_ordered: false, metadata: {} }, expected: Field { name: "name", data_type: Utf8, nullable: false, dict_id: 0, dict_is_ordered: false, metadata: {} }, Field { name: "surname", data_type: Utf8, nullable: false, dict_id: 0, dict_is_ordered: false, metadata: {} }
The diff in the exception shows that there are expected non nullable values that are nullable in the Arrow schema. The issue does not manifest when using pyarrow == 7.0.0. Actually it does
What you expected to happen: The optimize.compact() operation should work.
…ed large/normal arrow (#1926)
# Description
- Fixes optimize.compact not working when a table has parquet files with
large and normal arrow types. Basically it cast the recordbatch to
normal arrow types
# Issues
- closes#1889
Environment
Delta-rs version: 0.12.0
Binding: python
Environment:
Bug
What happened: When I perform an
optimize.compact()
operation following an append to a table, I encounter an exception. Oddly, this optimize.compact() succeeds after the initial append to an empty table but fails after subsequent appends. This is the exception I get:_internal.DeltaError: Data does not match the schema or partitions of the table: Unexpected Arrow schema: got: Field { name: "name", data_type: Utf8, nullable: true, dict_id: 0, dict_is_ordered: false, metadata: {} }, Field { name: "surname", data_type: Utf8, nullable: true, dict_id: 0, dict_is_ordered: false, metadata: {} }, expected: Field { name: "name", data_type: Utf8, nullable: false, dict_id: 0, dict_is_ordered: false, metadata: {} }, Field { name: "surname", data_type: Utf8, nullable: false, dict_id: 0, dict_is_ordered: false, metadata: {} }
The diff in the exception shows that there are expected non nullable values that are nullable in the Arrow schema.
The issue does not manifest when usingActually it doespyarrow == 7.0.0
.What you expected to happen: The
optimize.compact()
operation should work.How to reproduce it:
More details: ok
The text was updated successfully, but these errors were encountered: