From e1553a46403a57d6980db63974ffb7dc22108240 Mon Sep 17 00:00:00 2001 From: Mikhail Bautin Date: Wed, 20 May 2015 17:40:27 -0700 Subject: [PATCH] [SPARK-6880][CORE] Use properties from ActiveJob associated with a Stage Applying fix from https://github.com/apache/spark/pull/6291 by Mark Hamstra --- .../scala/org/apache/spark/scheduler/DAGScheduler.scala | 9 +++------ 1 file changed, 3 insertions(+), 6 deletions(-) diff --git a/core/src/main/scala/org/apache/spark/scheduler/DAGScheduler.scala b/core/src/main/scala/org/apache/spark/scheduler/DAGScheduler.scala index d742de1bf8a09..e7a641ef983f1 100644 --- a/core/src/main/scala/org/apache/spark/scheduler/DAGScheduler.scala +++ b/core/src/main/scala/org/apache/spark/scheduler/DAGScheduler.scala @@ -808,12 +808,9 @@ class DAGScheduler( } } - val properties = if (jobIdToActiveJob.contains(jobId)) { - jobIdToActiveJob(stage.jobId).properties - } else { - // this stage will be assigned to "default" pool - null - } + // Use the scheduling pool, job group, description, etc. from an ActiveJob associated + // with this Stage + val properties = jobIdToActiveJob(jobId).properties runningStages += stage // SparkListenerStageSubmitted should be posted before testing whether tasks are