- Jan 15, 2014
-
-
Kay Ousterhout authored
Prior to this commit, if a task crashes the JVM, the task (and all other tasks running on that executor) is marked at KILLED rather than FAILED. As a result, the TaskSetManager will retry the task indefiniteily rather than failing the job after maxFailures. This commit fixes that problem by marking tasks as FAILED rather than killed when an executor is lost. The downside of this commit is that if task A fails because another task running on the same executor caused the VM to crash, the failure will incorrectly be counted as a failure of task A. This should not be an issue because we typically set maxFailures to 3, and it is unlikely that a task will be co-located with a JVM-crashing task multiple times.
-
Patrick Wendell authored
GraphX shouldn't list Spark as provided. I noticed this when building an application against GraphX to audit the released artifacts.
-
Patrick Wendell authored
-
Patrick Wendell authored
Updated Debian packaging
-
Thomas Graves authored
More yarn code refactor Try to retrive common code in yarn alpha/stable for client and workerRunnable to reduce duplicated codes. By put them into a trait in common dir and extends with them. Same works could be done for the remaining files in alpha/stable , while the remainning files have much more overlapping codes with different API call here and there within functions, and will need much more close review , aslo it might divide functions into too small trifle ones, thus might not deserve to be done in this way. So just make it run for these two files firstly.
-
Reynold Xin authored
Rename VertexID -> VertexId in GraphX
-
Patrick Wendell authored
Fixed the flaky tests by making SparkConf not serializable SparkConf was being serialized with CoGroupedRDD and Aggregator, which somehow caused OptionalJavaException while being deserialized as part of a ShuffleMapTask. SparkConf should not even be serializable (according to conversation with Matei). This change fixes that. @mateiz @pwendell
-
Patrick Wendell authored
Fixed SVDPlusPlusSuite in Maven build. This should go into 0.9.0 also.
-
Tathagata Das authored
-
Tathagata Das authored
Changed SparkConf to not be serializable. And also fixed unit-test log paths in log4j.properties of external modules.
-
Reynold Xin authored
-
Mark Hamstra authored
-
Ankur Dave authored
-
Mark Hamstra authored
-
- Jan 14, 2014
-
-
Reynold Xin authored
Additional edits for clarity in the graphx programming guide. Added an overview of the Graph and GraphOps functions and fixed numerous typos.
-
Reynold Xin authored
Describe caching and uncaching in GraphX programming guide
-
Ankur Dave authored
-
Reynold Xin authored
Don't clone records for text files
-
Reynold Xin authored
Add GraphX dependency to examples/pom.xml
-
Reynold Xin authored
Deprecate rather than remove old combineValuesByKey function
-
Ankur Dave authored
-
Patrick Wendell authored
-
Patrick Wendell authored
-
Reynold Xin authored
API doc update & make Broadcast public In #413 Broadcast was mistakenly made private[spark]. I changed it to public again. Also exposing id in public given the R frontend requires that. Copied some of the documentation from the programming guide to API Doc for Broadcast and Accumulator. This should be cherry picked into branch-0.9 as well for 0.9.0 release.
-
Patrick Wendell authored
-
Patrick Wendell authored
-
Reynold Xin authored
-
Reynold Xin authored
Maintain Serializable API compatibility by reverting back to java.io.Serializable for Broadcast and Accumulator.
-
Reynold Xin authored
-
Reynold Xin authored
-
Reynold Xin authored
-
Reynold Xin authored
Note that previously Broadcast class was accidentally marked as private[spark]. It needs to be public for broadcast variables to work. Also exposing the broadcast varaible id.
-
Joseph E. Gonzalez authored
-
Reynold Xin authored
Improving the graphx-programming-guide This PR will track a few minor improvements to the content and formatting of the graphx-programming-guide.
-
Joseph E. Gonzalez authored
-
Patrick Wendell authored
Add missing header files
-
Patrick Wendell authored
-
Patrick Wendell authored
Removed unnecessary DStream operations and updated docs Removed StreamingContext.registerInputStream and registerOutputStream - they were useless. InputDStream has been made to register itself, and just registering a DStream as output stream cause RDD objects to be created but the RDDs will not be computed at all.. Also made DStream.register() private[streaming] for the same reasons. Updated docs, specially added package documentation for streaming package. Also, changed NetworkWordCount's input storage level to use MEMORY_ONLY, replication on the local machine causes warning messages (as replication fails) which is scary for a new user trying out his/her first example.
-
Tathagata Das authored
-
Tathagata Das authored
Conflicts: streaming/src/main/scala/org/apache/spark/streaming/dstream/DStream.scala
-