diff --git a/core/src/main/scala/org/apache/spark/TaskState.scala b/core/src/main/scala/org/apache/spark/TaskState.scala index cbace7b5f9f37c3acf0fe0dac77316bfad9d3129..596ce67d4cecf5b49b7a890156ed57d51fbc3ce7 100644 --- a/core/src/main/scala/org/apache/spark/TaskState.scala +++ b/core/src/main/scala/org/apache/spark/TaskState.scala @@ -21,7 +21,7 @@ private[spark] object TaskState extends Enumeration { val LAUNCHING, RUNNING, FINISHED, FAILED, KILLED, LOST = Value - val FINISHED_STATES = Set(FINISHED, FAILED, KILLED, LOST) + private val FINISHED_STATES = Set(FINISHED, FAILED, KILLED, LOST) type TaskState = Value diff --git a/core/src/main/scala/org/apache/spark/scheduler/Task.scala b/core/src/main/scala/org/apache/spark/scheduler/Task.scala index 1ed36bf0692f811aefc2e1ae91c05d37c31c04ff..ea9dc3988d9344f60b6a3ccbf670c69b2b214ce9 100644 --- a/core/src/main/scala/org/apache/spark/scheduler/Task.scala +++ b/core/src/main/scala/org/apache/spark/scheduler/Task.scala @@ -239,7 +239,7 @@ private[spark] object Task { * and return the task itself as a serialized ByteBuffer. The caller can then update its * ClassLoaders and deserialize the task. * - * @return (taskFiles, taskJars, taskBytes) + * @return (taskFiles, taskJars, taskProps, taskBytes) */ def deserializeWithDependencies(serializedTask: ByteBuffer) : (HashMap[String, Long], HashMap[String, Long], Properties, ByteBuffer) = {