-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Asset Inventory : google.api_core.exceptions.BadRequest: 400 Too many fields: 10090 #533
Comments
the update failed on:
|
@bmenasha can you take a look? |
thanks for reporting. this should be fixed by #535 , the production dataflow template has been updated with this change. let me know if it doesn't resolve the issue. |
I have relauch the pipeline with this parameter : |
Marquis, can you please try running this dataflow template gs://professional-services-tools-asset-inventory/test/import_pipeline ? |
done dataflow job launched with this parameters : |
same result During handling of the above exception, another exception occurred: Traceback (most recent call last): |
Thanks Marquis. I think this is caused by WRITE_APPEND combining the columns on the existing table putting us over the 10k column limit. The fixed submitted only kept the current snapshot under 10k. You should be able to complete a successful import with write_disposition set to WRITE_EMPTY but this will delete your old tables, saving only the most recent snapshot. To support WRITE_APPEND the import pipeline will need to read the existing schema merge it with the current schema truncating any new columns on the data to import that is over 10k. This should be doable and i'll work to complete that, but it will take a while longer. |
10,000 columns also seems like a lot, can you send these schema for one of these failing tables to me? ([email protected]). thanks.
the logs should also show the bigquery job id of the load job that failed and it too will show the schema. or you can list failed load job ids with
and for the failing job run
i'm just curious what these 10,000 columns are! |
Thanks Marquis. The latest change should resolve this but it's not backwards compatible with existing imports and will require importing to a new dateset. I deployed this fix to the test pipeline and will promote to the production one in a few days. Yes it will break any WRITE_APPEND imports but actually those just broke recently due to a cloud run api change so perfect time for it i guess. |
solved by starting a new dataset. |
Hello
Dataflow pipeline is broken since 6th of august due to quota error: Too Many fields 10090!
What's wrong? How to correct this
The text was updated successfully, but these errors were encountered: