-
Notifications
You must be signed in to change notification settings - Fork 28.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
SPARK-1093: Annotate developer and experimental API's #274
Changes from 6 commits
9d48cbf
42c1f09
5a76ed6
d86818e
cf23ec6
542a736
2f706f1
cd7a465
1ed27d2
8452309
c581dce
4a9e90c
7fb13b2
870a7ba
403ba52
d6b0afd
f390b13
e4c76b9
c318b24
f7d124f
037755c
824011b
99192ef
f3954e0
0d48908
c1bcb41
a01c076
982a473
e849f64
99b223a
8d0c873
bfe7b52
042c803
44179e4
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -17,7 +17,7 @@ | |
|
||
package org.apache.spark.metrics.sink | ||
|
||
trait Sink { | ||
private[spark] trait Sink { | ||
def start: Unit | ||
def stop: Unit | ||
} |
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -57,6 +57,7 @@ private[spark] class CoGroupPartition(idx: Int, val deps: Array[CoGroupSplitDep] | |
* @param rdds parent RDDs. | ||
* @param part partitioner used to partition the shuffle output. | ||
*/ | ||
private[spark] | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. You might want to relax this since I don't think the user cannot construct Product2 version of the CoGroupedRDD in PairRDDFunctions |
||
class CoGroupedRDD[K](@transient var rdds: Seq[RDD[_ <: Product2[K, _]]], part: Partitioner) | ||
extends RDD[(K, Seq[Seq[_]])](rdds.head.context, Nil) { | ||
|
||
|
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -24,7 +24,7 @@ import org.apache.spark.{Partition, SparkContext, TaskContext} | |
/** | ||
* An RDD that is empty, i.e. has no element in it. | ||
*/ | ||
class EmptyRDD[T: ClassTag](sc: SparkContext) extends RDD[T](sc, Nil) { | ||
private[spark] class EmptyRDD[T: ClassTag](sc: SparkContext) extends RDD[T](sc, Nil) { | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Is there a factory method for this? There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. this is only used inside of the spark package and it is always instantiated directly |
||
|
||
override def getPartitions: Array[Partition] = Array.empty | ||
|
||
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
make sure you mark RDD.mapPartitionsWithContext as well