- Dec 13, 2013
-
-
Prashant Sharma authored
-
- Dec 12, 2013
-
-
Prashant Sharma authored
-
- Dec 10, 2013
-
-
Prashant Sharma authored
-
- Dec 03, 2013
-
-
Harvey Feng authored
This also includes: -Change `isNewYarn` to `isNewHadoop`, since the protobuf-2.5 dependency is from Hadoop-2.2 itself. -Regexp bugix Credits to @alig for this patch.
-
- Nov 26, 2013
-
-
Harvey Feng authored
If the Hadoop used is version 2.2 or derived from it, then Spark will be compiled against protobuf-2.5 and a protobuf-2.5 version of Akka 2.0.5.
-
- Nov 15, 2013
-
-
Aaron Davidson authored
-
Aaron Davidson authored
I've diff'd this patch against my own -- since they were both created independently, this means that two sets of eyes have gone over all the merge conflicts that were created, so I'm feeling significantly more confident in the resulting PR. @rxin has looked at the changes to the repl and is resoundingly confident that they are correct.
-
- Nov 14, 2013
-
-
Raymond Liu authored
-
- Nov 12, 2013
-
-
Nathan Howell authored
-
Nathan Howell authored
sbt-assembly is setup to pick the first META-INF/services/org.apache.hadoop.security.SecurityInfo file instead of merging them. This causes Kerberos authentication to fail, this manifests itself in the "info:null" debug log statement: DEBUG SaslRpcClient: Get token info proto:interface org.apache.hadoop.yarn.api.ApplicationClientProtocolPB info:null DEBUG SaslRpcClient: Get kerberos info proto:interface org.apache.hadoop.yarn.api.ApplicationClientProtocolPB info:null ERROR UserGroupInformation: PriviledgedActionException as:foo@BAR (auth:KERBEROS) cause:org.apache.hadoop.security.AccessControlException: Client cannot authenticate via:[TOKEN, KERBEROS] DEBUG UserGroupInformation: PrivilegedAction as:foo@BAR (auth:KERBEROS) from:org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:583) WARN Client: Exception encountered while connecting to the server : org.apache.hadoop.security.AccessControlException: Client cannot authenticate via:[TOKEN, KERBEROS] ERROR UserGroupInformation: PriviledgedActionException as:foo@BAR (auth:KERBEROS) cause:java.io.IOException: org.apache.hadoop.security.AccessControlException: Client cannot authenticate via:[TOKEN, KERBEROS] This previously would just contain a single class: $ unzip -c assembly/target/scala-2.10/spark-assembly-0.9.0-incubating-SNAPSHOT-hadoop2.2.0.jar META-INF/services/org.apache.hadoop.security.SecurityInfo Archive: assembly/target/scala-2.10/spark-assembly-0.9.0-incubating-SNAPSHOT-hadoop2.2.0.jar inflating: META-INF/services/org.apache.hadoop.security.SecurityInfo org.apache.hadoop.security.AnnotatedSecurityInfo And now has the full list of classes: $ unzip -c assembly/target/scala-2.10/spark-assembly-0.9.0-incubating-SNAPSHOT-hadoop2.2.0.jar META-INF/services/org.apache.hadoop.security.SecurityInfoArchive: assembly/target/scala-2.10/spark-assembly-0.9.0-incubating-SNAPSHOT-hadoop2.2.0.jar inflating: META-INF/services/org.apache.hadoop.security.SecurityInfo org.apache.hadoop.security.AnnotatedSecurityInfo org.apache.hadoop.mapreduce.v2.app.MRClientSecurityInfo org.apache.hadoop.mapreduce.v2.security.client.ClientHSSecurityInfo org.apache.hadoop.yarn.security.client.ClientRMSecurityInfo org.apache.hadoop.yarn.security.ContainerManagerSecurityInfo org.apache.hadoop.yarn.security.SchedulerSecurityInfo org.apache.hadoop.yarn.security.admin.AdminSecurityInfo org.apache.hadoop.yarn.server.RMNMSecurityInfoClass
-
- Nov 11, 2013
-
-
tgravescs authored
-
- Nov 09, 2013
-
-
Josh Rosen authored
This allows the JavaAPICompletenessChecker to be run with Spark 0.8+.
-
- Nov 08, 2013
-
-
Russell Cardullo authored
This adds a metrics sink for graphite. The sink must be configured with the host and port of a graphite node and optionally may be configured with a prefix that will be prepended to all metrics that are sent to graphite.
-
- Oct 25, 2013
-
-
Patrick Wendell authored
Kafka uses an older version of jopt that causes bad conflicts with the version used by spark-perf. It's not easy to remove this downstream because of the way that spark-perf uses Spark (by including a spark assembly as an unmanaged jar). This fixes the problem at its source by just never including it.
-
- Oct 24, 2013
-
-
Prashant Sharma authored
-
- Oct 23, 2013
-
-
Matei Zaharia authored
-
- Oct 16, 2013
-
-
prabeesh authored
-
Shivaram Venkataraman authored
-
prabeesh authored
-
Patrick Wendell authored
- Examples assembly included a log4j.properties which clobbered Spark's - Example had an error where some classes weren't serializable - Did some other clean-up in this example
-
- Oct 12, 2013
-
-
jerryshao authored
-
- Oct 11, 2013
-
-
Shivaram Venkataraman authored
-
- Oct 09, 2013
-
-
Shivaram Venkataraman authored
-
- Oct 07, 2013
-
-
Reynold Xin authored
Resolving package conflicts with hadoop 0.23.9 Hadoop 0.23.9 is having a package conflict with easymock's dependencies. (cherry picked from commit 023e3fdf) Signed-off-by:
Reynold Xin <rxin@apache.org>
-
- Oct 05, 2013
-
-
Martin Weindel authored
using Scala 2.10.3, resolved maven-scala-plugin warning
-
- Oct 01, 2013
-
-
Du Li authored
-
- Sep 26, 2013
-
-
Aaron Davidson authored
This patch implements full distributed fault tolerance for standalone scheduler Masters. There is only one master Leader at a time, which is actively serving scheduling requests. If this Leader crashes, another master will eventually be elected, reconstruct the state from the first Master, and continue serving scheduling requests. Leader election is performed using the ZooKeeper leader election pattern. We try to minimize the use of ZooKeeper and the assumptions about ZooKeeper's behavior, so there is a layer of retries and session monitoring on top of the ZooKeeper client. Master failover follows directly from the single-node Master recovery via the file system (patch 194ba4b8), save that the Master state is stored in ZooKeeper instead. Configuration: By default, no recovery mechanism is enabled (spark.deploy.recoveryMode = NONE). By setting spark.deploy.recoveryMode to ZOOKEEPER and setting spark.deploy.zookeeper.url to an appropriate ZooKeeper URL, ZooKeeper recovery mode is enabled. By setting spark.deploy.recoveryMode to FILESYSTEM and setting spark.deploy.recoveryDirectory to an appropriate directory accessible by the Master, we will keep the behavior of from 194ba4b8. Additionally, places where a Master could be specificied by a spark:// url can now take comma-delimited lists to specify backup masters. Note that this is only used for registration of NEW Workers and application Clients. Once a Worker or Client has registered with the Master Leader, it is "in the system" and will never need to register again. Forthcoming: Documentation, tests (! - only ad hoc testing has been performed so far) I do not intend for this commit to be merged until tests are added, but this patch should still be mostly reviewable until then.
-
Reynold Xin authored
-
Prashant Sharma authored
-
- Sep 24, 2013
-
-
Patrick Wendell authored
-
- Sep 21, 2013
-
-
Prashant Sharma authored
-
- Sep 15, 2013
-
-
Patrick Wendell authored
-
Prashant Sharma authored
-
- Sep 14, 2013
-
-
Holden Karau authored
-
- Sep 11, 2013
-
-
Patrick Wendell authored
Due to this change in HDFS: https://issues.apache.org/jira/browse/HADOOP-7549 there is a bug when using the new assembly builds. The symptom is that any HDFS access results in an exception saying "No filesystem for scheme 'hdfs'". This adds a merge strategy in the assembly build which fixes the problem.
-
- Sep 10, 2013
-
-
Matei Zaharia authored
-
- Sep 08, 2013
-
-
Patrick Wendell authored
-
Patrick Wendell authored
-
- Sep 07, 2013
-
-
Patrick Wendell authored
-
- Sep 06, 2013
-
-
Jey Kottalam authored
-