- Dec 25, 2013
-
-
liguoqiang authored
-
Matei Zaharia authored
Typo: avaiable -> available
-
Reynold Xin authored
Fixed job name in the java streaming example.
-
- Dec 24, 2013
-
-
Andrew Ash authored
-
Patrick Wendell authored
Deduplicate Local and Cluster schedulers. The code in LocalScheduler/LocalTaskSetManager was nearly identical to the code in ClusterScheduler/ClusterTaskSetManager. The redundancy made making updating the schedulers unnecessarily painful and error- prone. This commit combines the two into a single TaskScheduler/ TaskSetManager. Unfortunately the diff makes this change look much more invasive than it is -- TaskScheduler.scala is only superficially changed (names updated, overrides removed) from the old ClusterScheduler.scala, and the same with TaskSetManager.scala. Thanks @rxin for suggesting this change!
-
Patrick Wendell authored
Clean up shuffle files once their metadata is gone Previously, we would only clean the in-memory metadata for consolidated shuffle files. Additionally, fixes a bug where the Metadata Cleaner was ignoring type-specific TTLs.
-
Kay Ousterhout authored
-
Matei Zaharia authored
Change the order of CLASSPATH. SPARK_TOOLS_JAR should be placed after CLASSPATH or at least after SPARK_CLASSPATH. If SPARK_TOOLS_JAR is placed before CLASSPATH, all assembled classes and resources in spark-tools-assembly.jar beat those in CLASSPATH or SPARK_CLASSPATH, which might be replaced by customized versions.
-
azuryyu authored
-
azuryyu authored
-
Reynold Xin authored
Show full stack trace and time taken in unit tests.
-
- Dec 23, 2013
-
-
Reynold Xin authored
-
Matei Zaharia authored
Refactored the streaming scheduler and added StreamingListener interface - Refactored the streaming scheduler for cleaner code. Specifically, the JobManager was renamed to JobScheduler, as it does the actual scheduling of Spark jobs to the SparkContext. The earlier Scheduler was renamed to JobGenerator, as it actually generates the jobs from the DStreams. The JobScheduler starts the JobGenerator. Also, moved all the scheduler related code from spark.streaming to spark.streaming.scheduler package. - Implemented the StreamingListener interface, similar to SparkListener. The streaming version of StatusReportListener prints the batch processing time statistics (for now). Added StreamingListernerSuite to test it. - Refactored streaming TestSuiteBase for deduping code in the other streaming testsuites.
-
Tathagata Das authored
-
Tathagata Das authored
-
Tathagata Das authored
-
Reynold Xin authored
Added SPARK-968 implementation for review Added SPARK-968 implementation for review
-
wangda.tan authored
-
- Dec 22, 2013
-
-
Kay Ousterhout authored
-
wangda.tan authored
-
Kay Ousterhout authored
-
- Dec 20, 2013
-
-
Tathagata Das authored
-
Kay Ousterhout authored
-
Kay Ousterhout authored
Conflicts: core/src/main/scala/org/apache/spark/scheduler/cluster/ClusterTaskSetManager.scala core/src/main/scala/org/apache/spark/scheduler/cluster/CoarseGrainedSchedulerBackend.scala core/src/test/scala/org/apache/spark/scheduler/TaskSetManagerSuite.scala
-
Patrick Wendell authored
Minor cleanup for standalone scheduler See commit messages
-
- Dec 19, 2013
-
-
Kay Ousterhout authored
Merge master into 127
-
Patrick Wendell authored
Track and report task result serialisation time. - DirectTaskResult now has a ByteBuffer valueBytes instead of a T value. - DirectTaskResult now has a member function T value() that deserialises valueBytes. - Executor serialises value into a ByteBuffer and passes it to DTR's ctor. - Executor tracks the time taken to do so and puts it in a new field in TaskMetrics. - StagePage now reports serialisation time from TaskMetrics along with the other things it reported.
-
Aaron Davidson authored
-
Aaron Davidson authored
As a lonely child with no one to care for it... we had to put it down.
-
Aaron Davidson authored
-
Aaron Davidson authored
Previously, we would only clean the in-memory metadata for consolidated shuffle files. Additionally, fixes a bug where the Metadata Cleaner was ignoring type- specific TTLs.
-
Reynold Xin authored
Add collectPartition to JavaRDD interface. This interface is useful for implementing `take` from other language frontends where the data is serialized. Also remove `takePartition` from PythonRDD and use `collectPartition` in rdd.py. Thanks @concretevitamin for the original change and tests.
-
Shivaram Venkataraman authored
-
Shivaram Venkataraman authored
Change the implementation to use runJob instead of PartitionPruningRDD. Also update the unit tests and the python take implementation to use the new interface.
-
Matei Zaharia authored
Add toString to Java RDD, and __repr__ to Python RDD Addresses [SPARK-992](https://spark-project.atlassian.net/browse/SPARK-992)
-
Nick Pentreath authored
-
Reynold Xin authored
[SPARK-959] Explicitly depend on org.eclipse.jetty.orbit jar Without this, in some cases, Ivy attempts to download the wrong file and fails, stopping the whole build. See [bug](https://spark-project.atlassian.net/browse/SPARK-959) for more details. Note that this may not be the best solution, as I do not understand the root cause of why this only happens for some people. However, it is reported to work.
-
Tathagata Das authored
-
Aaron Davidson authored
Without this, in some cases, Ivy attempts to download the wrong file and fails, stopping the whole build. See bug for more details. (This is probably also the beginning of the slow death of our recently prettified dependencies. Form follow function.)
-
Reynold Xin authored
Increase spark.akka.askTimeout default to 30 seconds In experimental clusters we've observed that a 10 second timeout was insufficient, despite having a low number of nodes and relatively small workload (16 nodes, <1.5 TB data). This would cause an entire job to fail at the beginning of the reduce phase. There is no particular reason for this value to be small as a timeout should only occur in an exceptional situation. Also centralized the reading of spark.akka.askTimeout to AkkaUtils (surely this can later be cleaned up to use Typesafe). Finally, deleted some lurking implicits. If anyone can think of a reason they should still be there, please let me know.
-