diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/ui/AllExecutionsPage.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/ui/AllExecutionsPage.scala index 41929ed474fa..f9c69864a336 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/ui/AllExecutionsPage.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/ui/AllExecutionsPage.scala @@ -20,7 +20,7 @@ package org.apache.spark.sql.execution.ui import javax.servlet.http.HttpServletRequest import scala.collection.mutable -import scala.xml.Node +import scala.xml.{Node, NodeSeq} import org.apache.commons.lang3.StringEscapeUtils @@ -38,19 +38,19 @@ private[ui] class AllExecutionsPage(parent: SQLTab) extends WebUIPage("") with L if (listener.getRunningExecutions.nonEmpty) { _content ++= new RunningExecutionTable( - parent, "Running Queries", currentTime, + parent, s"Running Queries (${listener.getRunningExecutions.size})", currentTime, listener.getRunningExecutions.sortBy(_.submissionTime).reverse).toNodeSeq } if (listener.getCompletedExecutions.nonEmpty) { _content ++= new CompletedExecutionTable( - parent, "Completed Queries", currentTime, + parent, s"Completed Queries (${listener.getCompletedExecutions.size})", currentTime, listener.getCompletedExecutions.sortBy(_.submissionTime).reverse).toNodeSeq } if (listener.getFailedExecutions.nonEmpty) { _content ++= new FailedExecutionTable( - parent, "Failed Queries", currentTime, + parent, s"Failed Queries (${listener.getFailedExecutions.size})", currentTime, listener.getFailedExecutions.sortBy(_.submissionTime).reverse).toNodeSeq } _content @@ -61,7 +61,36 @@ private[ui] class AllExecutionsPage(parent: SQLTab) extends WebUIPage("") with L details.parentNode.querySelector('.stage-details').classList.toggle('collapsed') }} - UIUtils.headerSparkPage("SQL", content, parent, Some(5000)) + val summary: NodeSeq = +