Skip to content
Snippets Groups Projects
  1. Dec 13, 2013
  2. Dec 12, 2013
  3. Dec 10, 2013
  4. Dec 03, 2013
  5. Nov 26, 2013
  6. Nov 15, 2013
    • Aaron Davidson's avatar
      Use Kafka 2.10 (again) · ce1d2af7
      Aaron Davidson authored
      ce1d2af7
    • Aaron Davidson's avatar
      Various merge corrections · f629ba95
      Aaron Davidson authored
      I've diff'd this patch against my own -- since they were both created
      independently, this means that two sets of eyes have gone over all the
      merge conflicts that were created, so I'm feeling significantly more
      confident in the resulting PR.
      
      @rxin has looked at the changes to the repl and is resoundingly
      confident that they are correct.
      f629ba95
  7. Nov 14, 2013
  8. Nov 12, 2013
    • Nathan Howell's avatar
      Upgrade to sbt-assembly 0.9.2 · 48eac0bc
      Nathan Howell authored
      48eac0bc
    • Nathan Howell's avatar
      spark-assembly.jar fails to authenticate with YARN ResourceManager · 23146a67
      Nathan Howell authored
      sbt-assembly is setup to pick the first META-INF/services/org.apache.hadoop.security.SecurityInfo file instead of merging them. This causes Kerberos authentication to fail, this manifests itself in the "info:null" debug log statement:
      
          DEBUG SaslRpcClient: Get token info proto:interface org.apache.hadoop.yarn.api.ApplicationClientProtocolPB info:null
          DEBUG SaslRpcClient: Get kerberos info proto:interface org.apache.hadoop.yarn.api.ApplicationClientProtocolPB info:null
          ERROR UserGroupInformation: PriviledgedActionException as:foo@BAR (auth:KERBEROS) cause:org.apache.hadoop.security.AccessControlException: Client cannot authenticate via:[TOKEN, KERBEROS]
          DEBUG UserGroupInformation: PrivilegedAction as:foo@BAR (auth:KERBEROS) from:org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:583)
          WARN Client: Exception encountered while connecting to the server : org.apache.hadoop.security.AccessControlException: Client cannot authenticate via:[TOKEN, KERBEROS]
          ERROR UserGroupInformation: PriviledgedActionException as:foo@BAR (auth:KERBEROS) cause:java.io.IOException: org.apache.hadoop.security.AccessControlException: Client cannot authenticate via:[TOKEN, KERBEROS]
      
      This previously would just contain a single class:
      
      $ unzip -c assembly/target/scala-2.10/spark-assembly-0.9.0-incubating-SNAPSHOT-hadoop2.2.0.jar META-INF/services/org.apache.hadoop.security.SecurityInfo
      Archive:  assembly/target/scala-2.10/spark-assembly-0.9.0-incubating-SNAPSHOT-hadoop2.2.0.jar
        inflating: META-INF/services/org.apache.hadoop.security.SecurityInfo
      
          org.apache.hadoop.security.AnnotatedSecurityInfo
      
      And now has the full list of classes:
      
      $ unzip -c assembly/target/scala-2.10/spark-assembly-0.9.0-incubating-SNAPSHOT-hadoop2.2.0.jar META-INF/services/org.apache.hadoop.security.SecurityInfoArchive:  assembly/target/scala-2.10/spark-assembly-0.9.0-incubating-SNAPSHOT-hadoop2.2.0.jar
        inflating: META-INF/services/org.apache.hadoop.security.SecurityInfo
      
          org.apache.hadoop.security.AnnotatedSecurityInfo
          org.apache.hadoop.mapreduce.v2.app.MRClientSecurityInfo
          org.apache.hadoop.mapreduce.v2.security.client.ClientHSSecurityInfo
          org.apache.hadoop.yarn.security.client.ClientRMSecurityInfo
          org.apache.hadoop.yarn.security.ContainerManagerSecurityInfo
          org.apache.hadoop.yarn.security.SchedulerSecurityInfo
          org.apache.hadoop.yarn.security.admin.AdminSecurityInfo
          org.apache.hadoop.yarn.server.RMNMSecurityInfoClass
      23146a67
  9. Nov 11, 2013
  10. Nov 09, 2013
  11. Nov 08, 2013
    • Russell Cardullo's avatar
      Add graphite sink for metrics · ef85a51f
      Russell Cardullo authored
      This adds a metrics sink for graphite.  The sink must
      be configured with the host and port of a graphite node
      and optionally may be configured with a prefix that will
      be prepended to all metrics that are sent to graphite.
      ef85a51f
  12. Oct 25, 2013
    • Patrick Wendell's avatar
      Exclude jopt from kafka dependency. · af4a529f
      Patrick Wendell authored
      Kafka uses an older version of jopt that causes bad conflicts with the version
      used by spark-perf. It's not easy to remove this downstream because of the way
      that spark-perf uses Spark (by including a spark assembly as an unmanaged jar).
      This fixes the problem at its source by just never including it.
      af4a529f
  13. Oct 24, 2013
  14. Oct 23, 2013
  15. Oct 16, 2013
  16. Oct 12, 2013
  17. Oct 11, 2013
  18. Oct 09, 2013
  19. Oct 07, 2013
  20. Oct 05, 2013
  21. Oct 01, 2013
  22. Sep 26, 2013
    • Aaron Davidson's avatar
      Standalone Scheduler fault tolerance using ZooKeeper · f549ea33
      Aaron Davidson authored
      This patch implements full distributed fault tolerance for standalone scheduler Masters.
      There is only one master Leader at a time, which is actively serving scheduling
      requests. If this Leader crashes, another master will eventually be elected, reconstruct
      the state from the first Master, and continue serving scheduling requests.
      
      Leader election is performed using the ZooKeeper leader election pattern. We try to minimize
      the use of ZooKeeper and the assumptions about ZooKeeper's behavior, so there is a layer of
      retries and session monitoring on top of the ZooKeeper client.
      
      Master failover follows directly from the single-node Master recovery via the file
      system (patch 194ba4b8), save that the Master state is stored in ZooKeeper instead.
      
      Configuration:
      By default, no recovery mechanism is enabled (spark.deploy.recoveryMode = NONE).
      By setting spark.deploy.recoveryMode to ZOOKEEPER and setting spark.deploy.zookeeper.url
      to an appropriate ZooKeeper URL, ZooKeeper recovery mode is enabled.
      By setting spark.deploy.recoveryMode to FILESYSTEM and setting spark.deploy.recoveryDirectory
      to an appropriate directory accessible by the Master, we will keep the behavior of from 194ba4b8.
      
      Additionally, places where a Master could be specificied by a spark:// url can now take
      comma-delimited lists to specify backup masters. Note that this is only used for registration
      of NEW Workers and application Clients. Once a Worker or Client has registered with the
      Master Leader, it is "in the system" and will never need to register again.
      
      Forthcoming:
      Documentation, tests (! - only ad hoc testing has been performed so far)
      I do not intend for this commit to be merged until tests are added, but this patch should
      still be mostly reviewable until then.
      f549ea33
    • Reynold Xin's avatar
      Removed scala -optimize flag. · 3f283278
      Reynold Xin authored
      3f283278
    • Prashant Sharma's avatar
      fixed maven build for scala 2.10 · 7ff4c2d3
      Prashant Sharma authored
      7ff4c2d3
  23. Sep 24, 2013
  24. Sep 21, 2013
  25. Sep 15, 2013
  26. Sep 14, 2013
  27. Sep 11, 2013
  28. Sep 10, 2013
  29. Sep 08, 2013
  30. Sep 07, 2013
  31. Sep 06, 2013
Loading