-
Notifications
You must be signed in to change notification settings - Fork 42
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: add write_engine
parameter to read_FORMATNAME
methods to control how data is written to BigQuery
#371
Conversation
Blocked by googleapis/python-bigquery#1815 |
third_party/bigframes_vendored/google_cloud_bigquery/_pandas_helpers.py
Outdated
Show resolved
Hide resolved
Test failure is a real one:
|
third_party/bigframes_vendored/google_cloud_bigquery/tests/unit/test_pyarrow_hlpers.py
Outdated
Show resolved
Hide resolved
third_party/bigframes_vendored/google_cloud_bigquery/_pyarrow_helpers.py
Outdated
Show resolved
Hide resolved
third_party/bigframes_vendored/google_cloud_bigquery/_pyarrow_helpers.py
Outdated
Show resolved
Hide resolved
third_party/bigframes_vendored/google_cloud_bigquery/tests/unit/test_pyarrow_hlpers.py
Outdated
Show resolved
Hide resolved
Marking as |
I've mailed googleapis/python-bigquery-pandas#814 as a first step of this project. I plan to follow-up that PR with one that copies the |
…ntrol how data is written to BigQuery
eebdd0e
to
3fca092
Compare
@shobsi This is ready for another look. Not quite merge-ready yet, as I need to update pandas-gbq's schema detection to account for Edit: googleapis/python-bigquery-pandas#832 mailed to fix DATETIME system test. |
Looks like there are a few more tests for me to cleanup. These do look related to my change:
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Updated. Thanks for the review!
Looks like I still need a fix in pandas-gbq for empty arrays.
googleapis/python-bigquery-pandas#838 to fix the remaining system test. Tested locally and it works. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM to merge after the pandas-gbq fix and update
third_party/bigframes_vendored/google_cloud_bigquery/tests/unit/test_pyarrow_hlpers.py
Outdated
Show resolved
Hide resolved
# BigFrames doesn't distinguish between string and large_string because the | ||
# largest string (2 GB) is already larger than the largest BigQuery row. | ||
if pa.types.is_string(arrow_dtype) or pa.types.is_large_string(arrow_dtype): | ||
return STRING_DTYPE |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@GarrettWu Would this break the JSON support you're adding? It was needed for some failing CSV tests.
Thank you for opening a Pull Request! Before submitting your PR, there are a few things you can do to make sure it goes smoothly:
Fixes internal issue 323176126
🦕