diff --git a/project/MimaBuild.scala b/project/MimaBuild.scala
index 4bfa9bf95ad78..9cb31d70444ff 100644
--- a/project/MimaBuild.scala
+++ b/project/MimaBuild.scala
@@ -75,8 +75,8 @@ object MimaBuild {
excludeSparkClass("streaming.dstream.NetworkReceiver#BlockGenerator") ++
excludeSparkClass("streaming.dstream.NetworkReceiver#BlockGenerator#Block") ++
excludeSparkClass("streaming.dstream.ReportError") ++
- excludeSparkClass("org.apache.spark.streaming.dstream.ReportBlock") ++
- excludeSparkClass("org.apache.spark.streaming.dstream.DStream")
+ excludeSparkClass("streaming.dstream.ReportBlock") ++
+ excludeSparkClass("streaming.dstream.DStream")
case _ => Seq()
}
diff --git a/streaming/src/main/scala/org/apache/spark/streaming/ui/StreamingJobProgressListener.scala b/streaming/src/main/scala/org/apache/spark/streaming/ui/StreamingJobProgressListener.scala
index 8921b99f53a23..8b025b09ed34d 100644
--- a/streaming/src/main/scala/org/apache/spark/streaming/ui/StreamingJobProgressListener.scala
+++ b/streaming/src/main/scala/org/apache/spark/streaming/ui/StreamingJobProgressListener.scala
@@ -33,7 +33,7 @@ private[ui] class StreamingJobProgressListener(ssc: StreamingContext) extends St
private val waitingBatchInfos = new HashMap[Time, BatchInfo]
private val runningBatchInfos = new HashMap[Time, BatchInfo]
private val completedaBatchInfos = new Queue[BatchInfo]
- private val batchInfoLimit = ssc.conf.getInt("spark.steaming.ui.maxBatches", 100)
+ private val batchInfoLimit = ssc.conf.getInt("spark.streaming.ui.retainedBatches", 100)
private var totalCompletedBatches = 0L
private val receiverInfos = new HashMap[Int, ReceiverInfo]
@@ -82,7 +82,7 @@ private[ui] class StreamingJobProgressListener(ssc: StreamingContext) extends St
runningBatchInfos.values.toSeq
}
- def completedBatches: Seq[BatchInfo] = synchronized {
+ def retainedCompletedBatches: Seq[BatchInfo] = synchronized {
completedaBatchInfos.toSeq
}
@@ -99,7 +99,7 @@ private[ui] class StreamingJobProgressListener(ssc: StreamingContext) extends St
}
def receivedRecordsDistributions: Map[Int, Option[Distribution]] = synchronized {
- val latestBatchInfos = allBatches.reverse.take(batchInfoLimit)
+ val latestBatchInfos = retainedBatches.reverse.take(batchInfoLimit)
val latestBlockInfos = latestBatchInfos.map(_.receivedBlockInfo)
(0 until numNetworkReceivers).map { receiverId =>
val blockInfoOfParticularReceiver = latestBlockInfos.map { batchInfo =>
@@ -134,10 +134,10 @@ private[ui] class StreamingJobProgressListener(ssc: StreamingContext) extends St
}
def lastReceivedBatch: Option[BatchInfo] = {
- allBatches.lastOption
+ retainedBatches.lastOption
}
- private def allBatches: Seq[BatchInfo] = synchronized {
+ private def retainedBatches: Seq[BatchInfo] = synchronized {
(waitingBatchInfos.values.toSeq ++
runningBatchInfos.values.toSeq ++ completedaBatchInfos).sortBy(_.batchTime)(Time.ordering)
}
diff --git a/streaming/src/main/scala/org/apache/spark/streaming/ui/StreamingPage.scala b/streaming/src/main/scala/org/apache/spark/streaming/ui/StreamingPage.scala
index 6fdfd8d05dcbb..290efaef51481 100644
--- a/streaming/src/main/scala/org/apache/spark/streaming/ui/StreamingPage.scala
+++ b/streaming/src/main/scala/org/apache/spark/streaming/ui/StreamingPage.scala
@@ -33,13 +33,13 @@ private[ui] class StreamingPage(parent: StreamingTab)
private val listener = parent.listener
private val startTime = Calendar.getInstance().getTime()
- private val emptyCellTest = "-"
+ private val emptyCell = "-"
/** Render the page */
override def render(request: HttpServletRequest): Seq[Node] = {
val content =
generateBasicStats() ++
-