Skip to content

Commit

Permalink
Revert "[SPARK-25299] add more scheduling logs (#617)"
Browse files Browse the repository at this point in the history
This reverts commit 45b0d28.
  • Loading branch information
rshkv authored and jdcasale committed Jul 17, 2020
1 parent 321fee3 commit cdcaffe
Show file tree
Hide file tree
Showing 2 changed files with 0 additions and 7 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -1138,15 +1138,13 @@ private[spark] class DAGScheduler(
}
} catch {
case NonFatal(e) =>
logDebug(s"Creating new stage attempt due to exception", e)
stage.makeNewStageAttempt(partitionsToCompute.size)
listenerBus.post(SparkListenerStageSubmitted(stage.latestInfo, properties))
abortStage(stage, s"Task creation failed: $e\n${Utils.exceptionString(e)}", Some(e))
runningStages -= stage
return
}

logDebug(s"Creating new stage attempt")
stage.makeNewStageAttempt(partitionsToCompute.size, taskIdToLocations.values.toSeq)

// If there are tasks to execute, record the submission time of the stage. Otherwise,
Expand Down Expand Up @@ -1528,7 +1526,6 @@ private[spark] class DAGScheduler(
s"(attempt ${failedStage.latestInfo.attemptNumber}) running")
} else {
failedStage.failedAttemptIds.add(task.stageAttemptId)
logInfo(s"Adding failed attemptId ${task.stageAttemptId} to stage ${failedStage}")
val shouldAbortStage =
failedStage.failedAttemptIds.size >= maxConsecutiveStageAttempts ||
disallowStageRetryForTest
Expand All @@ -1550,11 +1547,8 @@ private[spark] class DAGScheduler(
// Mark all the map as broken in the map stage, to ensure retry all the tasks on
// resubmitted stage attempt.
mapOutputTracker.unregisterAllMapOutput(shuffleId)
logInfo(s"Map stage ${mapStage} is a barrier, so unregister all map outputs")
} else if (mapId != -1) {
// Mark the map whose fetch failed as broken in the map stage
logInfo(s"Unregistering MapOutput " +
s"(shuffleId=$shuffleId, mapId=$mapId, bmAddr=$bmAddress")
mapOutputTracker.unregisterMapOutput(shuffleId, mapId, bmAddress)
}

Expand Down
1 change: 0 additions & 1 deletion core/src/main/scala/org/apache/spark/scheduler/Stage.scala
Original file line number Diff line number Diff line change
Expand Up @@ -101,7 +101,6 @@ private[scheduler] abstract class Stage(
metrics.register(rdd.sparkContext)
_latestInfo = StageInfo.fromStage(
this, nextAttemptId, Some(numPartitionsToCompute), metrics, taskLocalityPreferences)
logInfo(s"Making new attemptId $nextAttemptId for stage $id")
nextAttemptId += 1
}

Expand Down

0 comments on commit cdcaffe

Please sign in to comment.