- Jan 07, 2014
-
-
Mark Hamstra authored
-
Reynold Xin authored
Mllib 16 bugfix Bug fix: https://spark-project.atlassian.net/browse/MLLIB-16 Hi, I fixed the bug and added a test suite for `GradientDescent`. There are 2 checks in the test case. First, the final loss must be lower than the initial one. Second, the trend of loss sequence should be decreasing, i.e., at least 80% iterations have lower losses than their prior iterations. Thanks!
-
Reynold Xin authored
add the comments about SPARK_WORKER_DIR this env variable seems to be forgotten in many cases we need to set this variable, e.g. in EC2, we have to move the large application log files from the EBS to the ephemeral storage
-
CodingCat authored
this env variable seems to be forgotten …
-
Reynold Xin authored
Suggested small changes to Java code for slightly more standard style, encapsulation and in some cases performance Sorry if this is too abrupt or not a welcome set of changes, but thought I'd see if I could contribute a little. I'm a Java developer and just getting seriously into Spark. So I thought I'd suggest a number of small changes to the couple Java parts of the code to make it a little tighter, more standard and even a bit faster. Feel free to take all, some or none of this. Happy to explain any of it.
-
Reynold Xin authored
spark -> org.apache.spark Changed package name spark to org.apache.spark which was missing in some of the files
-
Sean Owen authored
-
Patrick Wendell authored
Conf improvements There are two new features. 1. Allow users to set arbitrary akka configurations via spark conf. 2. Allow configuration to be printed in logs for diagnosis.
-
Reynold Xin authored
Add a script to download sbt if not present on the system As per the discussion on the dev mailing list this script will use the system sbt if present or otherwise attempt to install the sbt launcher. The fall back error message in the event it fails instructs the user to install sbt. While the URLs it fetches from aren't controlled by the spark project directly, they are stable and the current authoritative sources.
-
Holden Karau authored
-
Prashant Sharma authored
-
Prashant Sharma authored
-
Prashant Sharma authored
-
Holden Karau authored
-
prabeesh authored
-
- Jan 06, 2014
-
-
Patrick Wendell authored
Update stop-slaves.sh The most recently version has changed the directory structure, but this script "sbin/stop-all.sh" doesn't change with it accordingly. This mistake makes "sbin/stop-all.sh" can't stop the slave node.
-
sproblvem authored
The most recently version has changed the directory structure, but this script "sbin/stop-all.sh" doesn't change with it accordingly. This mistake makes "sbin/stop-all.sh" can't stop the slave node.
-
Patrick Wendell authored
Fix test breaking downstream builds This wasn't detected in the pull-request-builder because it manually sets SPARK_HOME. I'm going to change that (it should't do this) to make it like the other builds.
-
Patrick Wendell authored
-
Patrick Wendell authored
Made java options to be applied during tests so that they become self explanatory.
-
Patrick Wendell authored
SPARK-1005 Ning upgrade
-
Patrick Wendell authored
Clarify spark.cores.max in docs It controls the count of cores across the cluster, not on a per-machine basis.
-
Patrick Wendell authored
Change protobuf version for yarn alpha back to 2.4.1 The maven build for yarn-alpha uses the wrong protobuf version and hence the generated assembly jar doesn't work with Hadoop 0.23. Removing the setting for the yarn-alpha profile since the default protobuf version is 2.4.1 at the top of the pom file.
-
Patrick Wendell authored
Fix handling of empty SPARK_EXAMPLES_JAR Currently if SPARK_EXAMPLES_JAR is left unset you get a null pointer exception when running the examples (atleast on spark on yarn). The null now gets turned into a string of "null" when its put into the SparkConf so addJar no longer properly ignores it. This fixes that so that it can be left unset.
-
Thomas Graves authored
-
Andrew Ash authored
It controls the count of cores across the cluster, not on a per-machine basis.
-
Sean Owen authored
-
Thomas Graves authored
-
Prashant Sharma authored
-
Prashant Sharma authored
-
Xusen Yin authored
-
Holden Karau authored
-
Patrick Wendell authored
Quiet ERROR-level Akka Logs This fixes an issue I've seen where akka logs a bunch of things at ERROR level when connecting to a standalone cluster, even in the normal case. I noticed that even when lifecycle logging was disabled, the netty code inside of akka still logged away via akka's EndpointWriter class. There are also some other log streams that I think are new in akka 2.2.1 that I've disabled. Finally, I added some better logging to the standalone client. This makes it more clear when a connection failure occurs what is going on. Previously it never explicitly said if a connection attempt had failed. The commit messages here have some more detail.
-
Holden Karau authored
-
Holden Karau authored
-
Holden Karau authored
-
- Jan 05, 2014
-
-
Patrick Wendell authored
-
Xusen Yin authored
-
Reynold Xin authored
Removing SPARK_EXAMPLES_JAR in the code This re-writes all of the examples to use the `SparkContext.jarOfClass` mechanism for loading the examples jar. This necessary for environments like YARN and the Standalone mode where example programs will be submit from inside the cluster rather than at the client using `./spark-example`. This still leaves SPARK_EXAMPLES_JAR in place in the shell scripts for setting up the classpath if `./spark-example` is run.
-
Reynold Xin authored
Fall back to zero-arg constructor for Serializer initialization if there is no constructor that accepts SparkConf. This maintains backward compatibility with older serializers implemented by users.
-