Skip to content
This repository has been archived by the owner on Sep 18, 2023. It is now read-only.

[NSE-1171] Throw RuntimeException when reading duplicate fields in case-insensitive mode #1173

Merged

Conversation

jackylee-ch
Copy link
Contributor

What changes were proposed in this pull request?

We didn't cover the corner case when reading duplicate fields in case-insensitive mode. If there is one more field matched, we just return the first field that matched. However in vanilla spark, it will throw RuntimeException in theses cases.

How was this patch tested?

unit tests

@github-actions
Copy link

#1171

@PHILO-HE
Copy link
Collaborator

Thanks for your patch. All internal tests have passed. The patch looks good to me.
@zhouyuan, please have a further check.

Copy link
Collaborator

@zhouyuan zhouyuan left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

👍

@zhouyuan zhouyuan merged commit 9b60057 into oap-project:main Nov 30, 2022
zhouyuan pushed a commit to zhouyuan/native-sql-engine that referenced this pull request Dec 14, 2022
…se-insensitive mode (oap-project#1173)

* throw exception if one more columns matched in case insensitive mode

* add schema check in arrow v2
zhouyuan added a commit that referenced this pull request Dec 14, 2022
* [NSE-1170] Set correct row number in batch scan w/ partition columns (#1172)

* [NSE-1171] Throw RuntimeException when reading duplicate fields in case-insensitive mode (#1173)

* throw exception if one more columns matched in case insensitive mode

* add schema check in arrow v2

* bump h2/pgsql version (#1176)

* bump h2/pgsql version

Signed-off-by: Yuan Zhou <[email protected]>

* ignore one failed test

Signed-off-by: Yuan Zhou <[email protected]>

Signed-off-by: Yuan Zhou <[email protected]>

* [NSE-956] allow to write parquet with compression (#1014)

This patch adds support for writing parquet with compression

df.coalesce(1).write.format("arrow").option("parquet.compression","zstd").save(path)

Signed-off-by: Yuan Zhou [email protected]

* [NSE-1161] Support read-write parquet conversion to read-write arrow (#1162)

* add ArrowConvertExtension

* do not convert parquet fileformat while writing to partitioned/bucketed/sorted output

* fix cache failed

* care about write codec

* disable convertor extension by default

* add some comments

* remove wrong compress type check (#1178)

Since the compresssion has been supported in #1014 . The extra compression check in ArrowConvertorExtension can be remove now.

* fix to use right arrow branch (#1179)


fix to use right arrow branch
Signed-off-by: Yuan Zhou <[email protected]>

* [NSE-1171] Support merge parquet schema and read missing schema (#1175)

* Support merge parquet schema and read missing schema

* fix error

* optimize null vectors

* optimize code

* optimize code

* change code

* add schema merge suite tests

* add test for struct type

* to use 1.5 branch arrow

Signed-off-by: Yuan Zhou <[email protected]>

Signed-off-by: Yuan Zhou <[email protected]>
Signed-off-by: Yuan Zhou [email protected]
Co-authored-by: Jacky Lee <[email protected]>
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants