- Aug 31, 2013
-
-
Matei Zaharia authored
-
Matei Zaharia authored
-
Matei Zaharia authored
- Use "fluid" layout that can expand to wide browser windows, instead of the old one's limit of 1200 px - Remove unnecessary <hr> elements - Switch back to Bootstrap's default theme and tweak progress bar colors - Make headers more consistent between deploy and app UIs - Replace some inline CSS with stylesheets
-
Matei Zaharia authored
spark-daemon
-
Matei Zaharia authored
Pyspark sampling function
-
Matei Zaharia authored
Don't send SIGINT / ctrl-c to Py4J gateway subprocess
-
- Aug 30, 2013
-
-
Matei Zaharia authored
PySpark: implementing subtractByKey(), subtract() and keyBy()
-
Reynold Xin authored
Make HadoopRDD's configuration accessible
-
Mikhail Bautin authored
-
Mikhail Bautin authored
-
Reynold Xin authored
Fix broken build by removing addIntercept
-
Shivaram Venkataraman authored
-
Evan Sparks authored
Adding linear regression and refactoring Ridge regression to use SGD
-
Evan Sparks authored
Change SVM to use {0,1} labels
-
- Aug 29, 2013
-
-
Matei Zaharia authored
Change build and run instructions to use assemblies
-
Reynold Xin authored
Fix removed block zero size log reporting
-
Matei Zaharia authored
-
Matei Zaharia authored
-
Matei Zaharia authored
are now needed
-
Matei Zaharia authored
This includes the following changes: - The "assembly" package now builds in Maven by default, and creates an assembly containing both hadoop-client and Spark, unlike the old BigTop distribution assembly that skipped hadoop-client - There is now a bigtop-dist package to build the old BigTop assembly - The repl-bin package is no longer built by default since the scripts don't reply on it; instead it can be enabled with -Prepl-bin - Py4J is now included in the assembly/lib folder as a local Maven repo, so that the Maven package can link to it - run-example now adds the original Spark classpath as well because the Maven examples assembly lists spark-core and such as provided - The various Maven projects add a spark-yarn dependency correctly
-
Matei Zaharia authored
-
Matei Zaharia authored
-
Matei Zaharia authored
-
Matei Zaharia authored
-
Matei Zaharia authored
-
Matei Zaharia authored
This commit makes Spark invocation saner by using an assembly JAR to find all of Spark's dependencies instead of adding all the JARs in lib_managed. It also packages the examples into an assembly and uses that as SPARK_EXAMPLES_JAR. Finally, it replaces the old "run" script with two better-named scripts: "run-examples" for examples, and "spark-class" for Spark internal classes (e.g. REPL, master, etc). This is also designed to minimize the confusion people have in trying to use "run" to run their own classes; it's not meant to do that, but now at least if they look at it, they can modify run-examples to do a decent job for them. As part of this, Bagel's examples are also now properly moved to the examples package instead of bagel.
-
jerryshao authored
-
- Aug 28, 2013
-
-
Patrick Wendell authored
Expose `isLocal` in SparkContext.
-
Matei Zaharia authored
Hot fix for command runner
-
Patrick Wendell authored
-
Patrick Wendell authored
-
Patrick Wendell authored
-
Andre Schumacher authored
-
Josh Rosen authored
This addresses SPARK-885, a usability issue where PySpark's Java gateway process would be killed if the user hit ctrl-c. Note that SIGINT still won't cancel the running s This fix is based on http://stackoverflow.com/questions/5045771
-
Andre Schumacher authored
-
Matei Zaharia authored
Spark on Yarn should use yarn approved directories for spark.local.dir and tmp
-
Y.CORP.YAHOO.COM\tgraves authored
-
- Aug 27, 2013
-
-
Matei Zaharia authored
Spark on Yarn allow users to specify environment variables
-
Y.CORP.YAHOO.COM\tgraves authored
-
Y.CORP.YAHOO.COM\tgraves authored
-