-
Notifications
You must be signed in to change notification settings - Fork 28.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[SPARK-7311] Introduce internal Serializer API for determining if serializers support object relocation #5924
Changes from 8 commits
b9624ee
86d4dcd
450fa21
0ba75e6
2c1233a
4aa61b2
123b992
0a7ebd7
50a68ca
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -125,6 +125,13 @@ class KryoSerializer(conf: SparkConf) | |
override def newInstance(): SerializerInstance = { | ||
new KryoSerializerInstance(this) | ||
} | ||
|
||
private[spark] override lazy val supportsRelocationOfSerializedObjects: Boolean = { | ||
// If auto-flush is disabled, then Kryo may store references to duplicate occurrences of objects | ||
// in the stream rather than writing those objects' serialized bytes, breaking relocation. See | ||
// https://groups.google.com/d/msg/kryo-users/6ZUSyfjjtdo/FhGG1KHDXPgJ for more details. | ||
newInstance().asInstanceOf[KryoSerializerInstance].getAutoReset() | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I couldn't seem to find a good way to determine whether auto-reset was enabled short of actually creating a serializer instance. Fortunately, it's fairly cheap to create a KryoSerializerInstance that you don't write to because its buffers are allocated lazily. Combined with #5606, which enables serializer re-use in many circumstances, I don't think that this will carry a huge performance penalty (especially since this is a I suppose that one alternative would be to move this method from |
||
} | ||
} | ||
|
||
private[spark] | ||
|
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -23,7 +23,7 @@ import java.nio.ByteBuffer | |
import scala.reflect.ClassTag | ||
|
||
import org.apache.spark.{SparkConf, SparkEnv} | ||
import org.apache.spark.annotation.DeveloperApi | ||
import org.apache.spark.annotation.{Private, Experimental, DeveloperApi} | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. nit: alphabetize There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Argh, I really need to fix my IntelliJ settings. I switched versions and didn't port all of my import sorting settings over, so stuff like this keeps happening :( |
||
import org.apache.spark.util.{Utils, ByteBufferInputStream, NextIterator} | ||
|
||
/** | ||
|
@@ -63,6 +63,39 @@ abstract class Serializer { | |
|
||
/** Creates a new [[SerializerInstance]]. */ | ||
def newInstance(): SerializerInstance | ||
|
||
/** | ||
* :: Private :: | ||
* Returns true if this serializer supports relocation of its serialized objects and false | ||
* otherwise. This should return true if and only if reordering the bytes of serialized objects | ||
* in serialization stream output is equivalent to having re-ordered those elements prior to | ||
* serializing them. More specifically, the following should hold if a serializer supports | ||
* relocation: | ||
* | ||
* {{{ | ||
* serOut.open() | ||
* position = 0 | ||
* serOut.write(obj1) | ||
* serOut.flush() | ||
* position = # of bytes writen to stream so far | ||
* obj1Bytes = output[0:position-1] | ||
* serOut.write(obj2) | ||
* serOut.flush() | ||
* position2 = # of bytes written to stream so far | ||
* obj2Bytes = output[position:position2-1] | ||
* serIn.open([obj2bytes] concatenate [obj1bytes]) should return (obj2, obj1) | ||
* }}} | ||
* | ||
* In general, this property should hold for serializers that are stateless and that do not | ||
* write special metadata at the beginning or end of the serialization stream. | ||
* | ||
* This API is private to Spark; this method should not be overridden in third-party subclasses | ||
* or called in user code and is subject to removal in future Spark releases. | ||
* | ||
* See SPARK-7311 for more details. | ||
*/ | ||
@Private | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I'm not sure that we want to commit to this as a stable public API, which is why I've chosen to mark this as private and leave comments warning users that this API is private and subject to change. If someone can think of a better way to restrict use / implementation of this method, I'd be happy to incorporate that change. |
||
private[spark] def supportsRelocationOfSerializedObjects: Boolean = false | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I'm not set on this name, by the way; happy to change if someone thinks of a less verbose name that's not misleading. |
||
} | ||
|
||
|
||
|
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,119 @@ | ||
/* | ||
* Licensed to the Apache Software Foundation (ASF) under one or more | ||
* contributor license agreements. See the NOTICE file distributed with | ||
* this work for additional information regarding copyright ownership. | ||
* The ASF licenses this file to You under the Apache License, Version 2.0 | ||
* (the "License"); you may not use this file except in compliance with | ||
* the License. You may obtain a copy of the License at | ||
* | ||
* http://www.apache.org/licenses/LICENSE-2.0 | ||
* | ||
* Unless required by applicable law or agreed to in writing, software | ||
* distributed under the License is distributed on an "AS IS" BASIS, | ||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
* See the License for the specific language governing permissions and | ||
* limitations under the License. | ||
*/ | ||
|
||
package org.apache.spark.serializer | ||
|
||
import java.io.{ByteArrayInputStream, ByteArrayOutputStream} | ||
|
||
import scala.util.Random | ||
|
||
import org.scalatest.{Assertions, FunSuite} | ||
|
||
import org.apache.spark.SparkConf | ||
import org.apache.spark.serializer.KryoTest.RegistratorWithoutAutoReset | ||
|
||
/** | ||
* Tests to ensure that [[Serializer]] implementations obey the API contracts for methods that | ||
* describe properties of the serialized stream, such as | ||
* [[Serializer.supportsRelocationOfSerializedObjects]]. | ||
*/ | ||
class SerializerPropertiesSuite extends FunSuite { | ||
|
||
import SerializerPropertiesSuite._ | ||
|
||
test("JavaSerializer does not support relocation") { | ||
// Per a comment on the SPARK-4550 JIRA ticket, Java serialization appears to write out the | ||
// full class name the first time an object is written to an output stream, but subsequent | ||
// references to the class write a more compact identifier; this prevents relocation. | ||
val ser = new JavaSerializer(new SparkConf()) | ||
testSupportsRelocationOfSerializedObjects(ser, generateRandomItem) | ||
} | ||
|
||
test("KryoSerializer supports relocation when auto-reset is enabled") { | ||
val ser = new KryoSerializer(new SparkConf) | ||
assert(ser.newInstance().asInstanceOf[KryoSerializerInstance].getAutoReset()) | ||
testSupportsRelocationOfSerializedObjects(ser, generateRandomItem) | ||
} | ||
|
||
test("KryoSerializer does not support relocation when auto-reset is disabled") { | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I tested this manually and have verified that this test fails if you leave out the check for the auto-reset flag. |
||
val conf = new SparkConf().set("spark.kryo.registrator", | ||
classOf[RegistratorWithoutAutoReset].getName) | ||
val ser = new KryoSerializer(conf) | ||
assert(!ser.newInstance().asInstanceOf[KryoSerializerInstance].getAutoReset()) | ||
testSupportsRelocationOfSerializedObjects(ser, generateRandomItem) | ||
} | ||
|
||
} | ||
|
||
object SerializerPropertiesSuite extends Assertions { | ||
|
||
def generateRandomItem(rand: Random): Any = { | ||
val randomFunctions: Seq[() => Any] = Seq( | ||
() => rand.nextInt(), | ||
() => rand.nextString(rand.nextInt(10)), | ||
() => rand.nextDouble(), | ||
() => rand.nextBoolean(), | ||
() => (rand.nextInt(), rand.nextString(rand.nextInt(10))), | ||
() => MyCaseClass(rand.nextInt(), rand.nextString(rand.nextInt(10))), | ||
() => { | ||
val x = MyCaseClass(rand.nextInt(), rand.nextString(rand.nextInt(10))) | ||
(x, x) | ||
} | ||
) | ||
randomFunctions(rand.nextInt(randomFunctions.size)).apply() | ||
} | ||
|
||
def testSupportsRelocationOfSerializedObjects( | ||
serializer: Serializer, | ||
generateRandomItem: Random => Any): Unit = { | ||
if (!serializer.supportsRelocationOfSerializedObjects) { | ||
return | ||
} | ||
val NUM_TRIALS = 5 | ||
val rand = new Random(42) | ||
for (_ <- 1 to NUM_TRIALS) { | ||
val items = { | ||
// Make sure that we have duplicate occurrences of the same object in the stream: | ||
val randomItems = Seq.fill(10)(generateRandomItem(rand)) | ||
randomItems ++ randomItems.take(5) | ||
} | ||
val baos = new ByteArrayOutputStream() | ||
val serStream = serializer.newInstance().serializeStream(baos) | ||
def serializeItem(item: Any): Array[Byte] = { | ||
val itemStartOffset = baos.toByteArray.length | ||
serStream.writeObject(item) | ||
serStream.flush() | ||
val itemEndOffset = baos.toByteArray.length | ||
baos.toByteArray.slice(itemStartOffset, itemEndOffset).clone() | ||
} | ||
val itemsAndSerializedItems: Seq[(Any, Array[Byte])] = { | ||
val serItems = items.map { | ||
item => (item, serializeItem(item)) | ||
} | ||
serStream.close() | ||
rand.shuffle(serItems) | ||
} | ||
val reorderedSerializedData: Array[Byte] = itemsAndSerializedItems.flatMap(_._2).toArray | ||
val deserializedItemsStream = serializer.newInstance().deserializeStream( | ||
new ByteArrayInputStream(reorderedSerializedData)) | ||
assert(deserializedItemsStream.asIterator.toSeq === itemsAndSerializedItems.map(_._1)) | ||
deserializedItemsStream.close() | ||
} | ||
} | ||
} | ||
|
||
private case class MyCaseClass(foo: Int, bar: String) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Should this say "auto-reset" instead of "auto-flush"?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes; good catch.