Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[SPARK-48148][CORE] JSON objects should not be modified when read as STRING #46408

Closed
wants to merge 9 commits into from

Conversation

eric-maynard
Copy link

@eric-maynard eric-maynard commented May 6, 2024

What changes were proposed in this pull request?

Currently, when reading a JSON like this:

{"a": {"b": -999.99999999999999999999999999999999995}}

With the schema:

a STRING

Spark will yield a result like this:

{"b": -1000.0}

Other changes such as changes to the input string's whitespace may also occur. In some cases, we apply scientific notation to an input floating-point number when reading it as STRING.

This applies to reading JSON files (as with spark.read.json) as well as the SQL expression from_json.

Why are the changes needed?

Correctness issues may occur if a field is read as a STRING and then later parsed (e.g. with from_json) after the contents have been modified.

Does this PR introduce any user-facing change?

Yes, when reading non-string fields from a JSON object using the STRING type, we will now extract the field exactly as it appears.

How was this patch tested?

Added a test in JsonSuite.scala

Was this patch authored or co-authored using generative AI tooling?

No

@github-actions github-actions bot added the SQL label May 6, 2024

val df = spark.read.schema("data STRING").json(path.getAbsolutePath)

val expected = s"""{"v": ${granularFloat}}"""
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can you add more test cases for the following?

  • {"data": {"v": "abc"}}, expected: "{"v": "abc"}"
  • {"data":{"v": "0.999"}}, expected: "{"v": "0.999"}"
  • {"data": [1, 2, 3]}, expected: "[1, 2, 3]"
  • {"data": }, expected the object as string.

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Added more tests -- can you clarify the last example and what we expect that to do? It seems like invalid JSON

Utils.tryWithResource(factory.createGenerator(writer, JsonEncoding.UTF8)) {
generator => generator.copyCurrentStructure(parser)
val startLocation = parser.getTokenLocation
startLocation.contentReference().getRawContent match {
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Is there an existing API to get the remaining content as string? Also, would it work with multi-line JSON?

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I was not able to find such an existing API -- there is JacksonParser.getText but that appears to simply get the current value if it's a string value.

Copy link
Author

@eric-maynard eric-maynard May 8, 2024

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

wrt. multiline JSON, I have added a test to cover this. It seems that the content reference is not a byte array when using multiline mode.

@sadikovi
Copy link
Contributor

sadikovi commented May 8, 2024

cc @dongjoon-hyun @HyukjinKwon

@HyukjinKwon
Copy link
Member

SPARK-48148: values are unchanged when read as string *** FAILED *** (134 milliseconds)

seems it fails

expectedExactData = Seq(s"""{"v": ${granularFloat}}""")
)
// In multiLine, we fall back to the inexact method:
extractData(
s"""{"data": {"white":\n"space"}}""",
Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This makes \n no longer function as a newline here

@eric-maynard
Copy link
Author

Hey @HyukjinKwon, can you take another look and possible re-trigger tests? I believe multiline should be working now.

@HyukjinKwon
Copy link
Member

Merged to master.

@HyukjinKwon
Copy link
Member

btw you can trigger on your own https://github.com/eric-maynard/spark/runs/24789350525 I can't trigger :-).

JacobZheng0927 pushed a commit to JacobZheng0927/spark that referenced this pull request May 11, 2024
…STRING

### What changes were proposed in this pull request?

Currently, when reading a JSON like this:
```
{"a": {"b": -999.99999999999999999999999999999999995}}
```

With the schema:
```
a STRING
```

Spark will yield a result like this:
```
{"b": -1000.0}
```

Other changes such as changes to the input string's whitespace may also occur. In some cases, we apply scientific notation to an input floating-point number when reading it as STRING.

This applies to reading JSON files (as with `spark.read.json`) as well as the SQL expression `from_json`.

### Why are the changes needed?

Correctness issues may occur if a field is read as a STRING and then later parsed (e.g. with `from_json`) after the contents have been modified.

### Does this PR introduce _any_ user-facing change?
Yes, when reading non-string fields from a JSON object using the STRING type, we will now extract the field exactly as it appears.

### How was this patch tested?
Added a test in `JsonSuite.scala`

### Was this patch authored or co-authored using generative AI tooling?
No

Closes apache#46408 from eric-maynard/SPARK-48148.

Lead-authored-by: Eric Maynard <[email protected]>
Co-authored-by: Hyukjin Kwon <[email protected]>
Signed-off-by: Hyukjin Kwon <[email protected]>
@revans2
Copy link
Contributor

revans2 commented Jun 25, 2024

Is there a reason this was only done for scan and not from_json/get_json_object? Especially from_json where in the past it behaved almost identically to the JSON scan (everything except empty row handling and new line characters).

@HyukjinKwon
Copy link
Member

@eric-maynard would you mind checking if this is true? we should match the behaviour at least with from_json.

@revans2
Copy link
Contributor

revans2 commented Jul 11, 2024

@eric-maynard and @HyukjinKwon any update on fixing at least from_json if not also get_json_object? If you need me to try and try and debug this and put up a patch I am happy to.

@HyukjinKwon
Copy link
Member

ping @eric-maynard

@HyukjinKwon
Copy link
Member

sent offline ping too.

@sandip-db
Copy link
Contributor

I did some tests and it looks like that this PR doesn't address the same issue with from_json. Json expressions including from_json use CreateJacksonParser.utf8String to create jackson parser, which in turn converts UTF8String to an InputStreamReader. The latter doesn't allow to copy buffers from random offset. I tried to create jackson parser using the underlying byte array of UTF8String, but ran into SPARK-16548 (that was fixed by #17693).

@HyukjinKwon
Copy link
Member

Let me take a look

@HyukjinKwon
Copy link
Member

Hm, I think this PR also doesn't cover multiline cases.

Another problem is that it seems parser.currentLocation.getByteOffset doesn't report correctly too so we cannot really get the current byte offset ..

@HyukjinKwon
Copy link
Member

This is what I tried:

ff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/json/CreateJacksonParser.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/cat
alyst/json/CreateJacksonParser.scala
index ba7b54fc04e8..721d611bc414 100644
--- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/json/CreateJacksonParser.scala
+++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/json/CreateJacksonParser.scala
@@ -28,6 +28,10 @@ import org.apache.hadoop.io.Text
 import org.apache.spark.sql.catalyst.InternalRow
 import org.apache.spark.unsafe.types.UTF8String

+private[sql] trait WithByteArrayInputStream {
+  val source: ByteArrayInputStream
+}
+
 object CreateJacksonParser extends Serializable {
   def string(jsonFactory: JsonFactory, record: String): JsonParser = {
     jsonFactory.createParser(record)
@@ -40,7 +44,10 @@ object CreateJacksonParser extends Serializable {
     val bain = new ByteArrayInputStream(
       bb.array(), bb.arrayOffset() + bb.position(), bb.remaining())

-    jsonFactory.createParser(new InputStreamReader(bain, StandardCharsets.UTF_8))
+    jsonFactory.createParser(
+      new InputStreamReader(bain, StandardCharsets.UTF_8) with WithByteArrayInputStream {
+        override val source: ByteArrayInputStream = bain
+      })
   }

   def text(jsonFactory: JsonFactory, record: Text): JsonParser = {
diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/json/JacksonParser.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/
json/JacksonParser.scala
index 32a1731a93d4..d98a14302d5a 100644
--- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/json/JacksonParser.scala
+++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/json/JacksonParser.scala
@@ -17,7 +17,7 @@

 package org.apache.spark.sql.catalyst.json

-import java.io.{ByteArrayOutputStream, CharConversionException}
+import java.io.{ByteArrayOutputStream, CharConversionException, InputStreamReader}
 import java.nio.charset.MalformedInputException

 import scala.collection.mutable.ArrayBuffer
@@ -332,6 +332,14 @@ class JacksonParser(
               val buffer = new Array[Byte](size)
               positionedReadable.read(startLocation.getByteOffset, buffer, 0, size)
               UTF8String.fromBytes(buffer, 0, size)
+            case inputStream: InputStreamReader with WithByteArrayInputStream =>
+              skipAhead()
+              val endLocation = parser.currentLocation.getByteOffset
+
+              val size = endLocation.toInt - (startLocation.getByteOffset.toInt)
+              val buffer = new Array[Byte](size)
+              inputStream.source.read(buffer, startLocation.getByteOffset.toInt, size)
+              UTF8String.fromBytes(buffer, 0, size)
             case _ =>
               val writer = new ByteArrayOutputStream()
               Utils.tryWithResource(factory.createGenerator(writer, JsonEncoding.UTF8)) {
diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/JsonExpressionsSuite.scala b/sql/catalyst/src/test/scala/org/apache/spark
/sql/catalyst/expressions/JsonExpressionsSuite.scala
index a23e7f44a48d..12cc8165a39e 100644
--- a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/JsonExpressionsSuite.scala
+++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/JsonExpressionsSuite.scala
@@ -432,6 +432,16 @@ class JsonExpressionsSuite extends SparkFunSuite with ExpressionEvalHelper with
     )
   }

+  test("from_json - nested object to string") {
+    val jsonData = """{"a": {"b": -999.99999999999999999999999999999999995}}"""
+    val schema = StructType(StructField("a", StringType) :: Nil)
+    checkEvaluation(
+      JsonToStructs(schema, Map.empty, Literal(jsonData), UTC_OPT),
+      InternalRow("""{"b": -999.99999999999999999999999999999999995}""")
+    )
+  }
+
+
   test("from_json - invalid data") {
     val jsonData = """{"a" 1}"""
     val schema = StructType(StructField("a", IntegerType) :: Nil)

HyukjinKwon pushed a commit that referenced this pull request Dec 1, 2024
### What changes were proposed in this pull request?

#46408 attempts to set the feature flag `INCLUDE_SOURCE_IN_LOCATION` in the JSON parser and reverts the flag to the original value. The reverting code is incorrect and accidentally sets the `AUTO_CLOSE_SOURCE` feature to false. The reason is that `overrideStdFeatures(value, mask)` sets the feature flags selected by `mask` to `value`. `originalMask` is a value of 0/1. When it is 1, it selects `AUTO_CLOSE_SOURCE`, whose ordinal is 0 ([reference](https://github.com/FasterXML/jackson-core/blob/172369cc390ace0f68a5032701634bdc984c2af8/src/main/java/com/fasterxml/jackson/core/JsonParser.java#L112)). The old code doesn't revert `INCLUDE_SOURCE_IN_LOCATION` to the original value either. As a result, when the JSON parser is closed, the underlying input stream is not closed, which can lead to memory leak.

### Why are the changes needed?

Perform the originally intended feature, and avoid memory leak.

### Does this PR introduce _any_ user-facing change?

No.

### How was this patch tested?

New unit test. It would fail without the change in the PR.

### Was this patch authored or co-authored using generative AI tooling?

No.

Closes #49018 from chenhao-db/fix_json_parser_flag.

Authored-by: Chenhao Li <[email protected]>
Signed-off-by: Hyukjin Kwon <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants