-
Notifications
You must be signed in to change notification settings - Fork 78
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Remove metadata associated with column #590
Comments
Sure, will add an option to disable metadata generation. By default metadata will be generated though because it helps migrating data to relational databases. E.g. Just out of the curiosity, what is the purpose of the validation and why metadata is the issue here? |
This should be available at the current Use
|
Thanks for your quick response @yruslan. The purpose of the validation is, when we send a file from one program to another, on arrival we do a schema validation through spark. To do this we create a default StructType schema that we can then compare to the schema that comes in from our other program. This compare will compare the entirety of the schema, even hidden values such as metadata since we use the .diff() method in spark Scala. Since not all the data we use goes through cobrix, we can’t change the StructField we are expecting to contain metadata, and so this is where our issue stemmed from. |
Thanks for describing your use case! |
This is released in 2.6.5 |
Background [Optional]
I'm using Cobrix v2.6.2 to convert the cobol file to parquet. I noticed that recently your team has added the 'maxLength' metadata for spark schema string fields in v2.6.0. This addition of metadata is causing validation issues for us during dataframe transformations and we are having hard time in removing this metadata associated with each fields.
Question
I do not need this metadata info to be associated with the field. Is there a way to disable metadata generation? something like .option("metadata", false)
Appreciate your response.
The text was updated successfully, but these errors were encountered: