Skip to content

Commit 028b0ac

Browse files
committed
rename
1 parent fd413d4 commit 028b0ac

File tree

4 files changed

+10
-10
lines changed

4 files changed

+10
-10
lines changed

sql/core/src/main/scala/org/apache/spark/sql/execution/SparkPlanInfo.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -54,7 +54,7 @@ private[execution] object SparkPlanInfo {
5454
val children = plan match {
5555
case ReusedExchangeExec(_, child) => child :: Nil
5656
case a: AdaptiveSparkPlanExec => a.finalPlan :: Nil
57-
case fragment: QueryStageExec => fragment.plan :: Nil
57+
case stage: QueryStageExec => stage.plan :: Nil
5858
case _ => plan.children ++ plan.subqueries
5959
}
6060
val metrics = plan.metrics.toSeq.map { case (key, metric) =>

sql/core/src/main/scala/org/apache/spark/sql/execution/adaptive/QueryStageManager.scala

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -49,7 +49,7 @@ class QueryStageManager(
4949
initialPlan: SparkPlan,
5050
session: SparkSession,
5151
callback: QueryStageManagerCallback)
52-
extends EventLoop[QueryStageManagerEvent]("QueryFragmentCreator") {
52+
extends EventLoop[QueryStageManagerEvent]("QueryStageCreator") {
5353

5454
private def conf = session.sessionState.conf
5555

@@ -201,7 +201,7 @@ case class CreateStageResult(newPlan: SparkPlan, allChildStagesReady: Boolean)
201201

202202
object QueryStageManager {
203203
private val executionContext = ExecutionContext.fromExecutorService(
204-
ThreadUtils.newDaemonCachedThreadPool("QueryFragmentCreator", 16))
204+
ThreadUtils.newDaemonCachedThreadPool("QueryStageCreator", 16))
205205
}
206206

207207
trait QueryStageManagerCallback {

sql/core/src/main/scala/org/apache/spark/sql/execution/adaptive/rule/ReduceNumShufflePartitions.scala

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -55,9 +55,9 @@ case class ReduceNumShufflePartitions(conf: SQLConf) extends Rule[SparkPlan] {
5555

5656
override def apply(plan: SparkPlan): SparkPlan = {
5757
val shuffleMetrics: Seq[MapOutputStatistics] = plan.collect {
58-
case fragment: ShuffleQueryStageExec =>
59-
val metricsFuture = fragment.mapOutputStatisticsFuture
60-
assert(metricsFuture.isCompleted, "ShuffleQueryFragment should already be ready")
58+
case stage: ShuffleQueryStageExec =>
59+
val metricsFuture = stage.mapOutputStatisticsFuture
60+
assert(metricsFuture.isCompleted, "ShuffleQueryStageExec should already be ready")
6161
ThreadUtils.awaitResult(metricsFuture, Duration.Zero)
6262
}
6363

@@ -88,7 +88,7 @@ case class ReduceNumShufflePartitions(conf: SQLConf) extends Rule[SparkPlan] {
8888

8989
/**
9090
* Estimates partition start indices for post-shuffle partitions based on
91-
* mapOutputStatistics provided by all pre-shuffle fragments.
91+
* mapOutputStatistics provided by all pre-shuffle stages.
9292
*/
9393
// visible for testing.
9494
private[sql] def estimatePartitionStartIndices(
@@ -112,7 +112,7 @@ case class ReduceNumShufflePartitions(conf: SQLConf) extends Rule[SparkPlan] {
112112
s"advisoryTargetPostShuffleInputSize: $advisoryTargetPostShuffleInputSize, " +
113113
s"targetPostShuffleInputSize $targetPostShuffleInputSize.")
114114

115-
// Make sure we do get the same number of pre-shuffle partitions for those fragments.
115+
// Make sure we do get the same number of pre-shuffle partitions for those stages.
116116
val distinctNumPreShufflePartitions =
117117
mapOutputStatistics.map(stats => stats.bytesByPartitionId.length).distinct
118118
// The reason that we are expecting a single value of the number of pre-shuffle partitions
@@ -135,7 +135,7 @@ case class ReduceNumShufflePartitions(conf: SQLConf) extends Rule[SparkPlan] {
135135

136136
var i = 0
137137
while (i < numPreShufflePartitions) {
138-
// We calculate the total size of ith pre-shuffle partitions from all pre-shuffle fragments.
138+
// We calculate the total size of ith pre-shuffle partitions from all pre-shuffle stages.
139139
// Then, we add the total size to postShuffleInputSize.
140140
var nextShuffleInputSize = 0L
141141
var j = 0

sql/core/src/main/scala/org/apache/spark/sql/execution/ui/SparkPlanGraph.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -96,7 +96,7 @@ object SparkPlanGraph {
9696
case "InputAdapter" =>
9797
buildSparkPlanGraphNode(
9898
planInfo.children.head, nodeIdGenerator, nodes, edges, parent, null, exchanges)
99-
case "BroadcastQueryFragment" | "ResultQueryFragment" | "ShuffleQueryFragment" =>
99+
case "BroadcastQueryStage" | "ShuffleQueryStage" =>
100100
if (exchanges.contains(planInfo.children.head)) {
101101
// Point to the re-used exchange
102102
val node = exchanges(planInfo.children.head)

0 commit comments

Comments
 (0)