Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[BUG] test_coalesce fails with DATAGEN_SEED=1698940723 #9684

Closed
abellina opened this issue Nov 13, 2023 · 0 comments · Fixed by #9852
Closed

[BUG] test_coalesce fails with DATAGEN_SEED=1698940723 #9684

abellina opened this issue Nov 13, 2023 · 0 comments · Fixed by #9852
Assignees
Labels
bug Something isn't working

Comments

@abellina
Copy link
Collaborator

abellina commented Nov 13, 2023

 SPARK_RAPIDS_TEST_DATAGEN_SEED=1698940723 ./run_pyspark_from_build.sh -k test_coalesce\ and\ Double\ and\ not_null

CPU fails with duplicate key. Note that I tried setting mapKeyDedupPolicy=LAST_WIN and the test passes.

2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                   : java.lang.RuntimeException: Duplicate map key NaN was found, please check the input data. If you want to remove the duplicated keys, you can set spark.sql.mapKeyDedupPolicy to LAST_WIN so that the key inserted at last takes precedence.^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at org.apache.spark.sql.catalyst.util.ArrayBasedMapBuilder.put(ArrayBasedMapBuilder.scala:72)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at org.apache.spark.sql.catalyst.expressions.CreateMap.eval(complexTypeCreator.scala:229)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at org.apache.spark.sql.catalyst.optimizer.ConstantFolding$$anonfun$apply$1$$anonfun$applyOrElse$1.applyOrElse(expressions.scala:66)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at org.apache.spark.sql.catalyst.optimizer.ConstantFolding$$anonfun$apply$1$$anonfun$applyOrElse$1.applyOrElse(expressions.scala:54)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$transformDown$1(TreeNode.scala:317)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:73)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:317)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$transformDown$3(TreeNode.scala:322)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at org.apache.spark.sql.catalyst.trees.TreeNode.mapChild$2(TreeNode.scala:376)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$mapChildren$4(TreeNode.scala:437)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at scala.collection.TraversableLike.$anonfun$map$1(TraversableLike.scala:238)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at scala.collection.TraversableLike.map(TraversableLike.scala:238)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at scala.collection.TraversableLike.map$(TraversableLike.scala:231)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at scala.collection.AbstractTraversable.map(Traversable.scala:108)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$mapChildren$1(TreeNode.scala:437)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:243)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:405)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:358)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:322)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$transformDown$3(TreeNode.scala:322)^[[0m
[2023-11-02T17:07:56.059Z] ^[[1m^[[31mE                     at org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$mapChildren$1(TreeNode.scala:407)^[[0m
@abellina abellina added bug Something isn't working ? - Needs Triage Need team to review and classify labels Nov 13, 2023
@abellina abellina changed the title [BUG] test_map_scalars_supported_key_types fails with DATAGEN_SEED=1698940723 [BUG] test_coalesce fails with DATAGEN_SEED=1698940723 Nov 13, 2023
@mattahrens mattahrens removed the ? - Needs Triage Need team to review and classify label Nov 14, 2023
@thirtiseven thirtiseven self-assigned this Nov 29, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

Successfully merging a pull request may close this issue.

3 participants