Skip to content
Snippets Groups Projects
  1. Jul 05, 2017
    • Dongjoon Hyun's avatar
      [SPARK-21278][PYSPARK] Upgrade to Py4J 0.10.6 · c8d0aba1
      Dongjoon Hyun authored
      ## What changes were proposed in this pull request?
      
      This PR aims to bump Py4J in order to fix the following float/double bug.
      Py4J 0.10.5 fixes this (https://github.com/bartdag/py4j/issues/272) and the latest Py4J is 0.10.6.
      
      **BEFORE**
      ```
      >>> df = spark.range(1)
      >>> df.select(df['id'] + 17.133574204226083).show()
      +--------------------+
      |(id + 17.1335742042)|
      +--------------------+
      |       17.1335742042|
      +--------------------+
      ```
      
      **AFTER**
      ```
      >>> df = spark.range(1)
      >>> df.select(df['id'] + 17.133574204226083).show()
      +-------------------------+
      |(id + 17.133574204226083)|
      +-------------------------+
      |       17.133574204226083|
      +-------------------------+
      ```
      
      ## How was this patch tested?
      
      Manual.
      
      Author: Dongjoon Hyun <dongjoon@apache.org>
      
      Closes #18546 from dongjoon-hyun/SPARK-21278.
      c8d0aba1
  2. Jun 20, 2017
    • Xingbo Jiang's avatar
      [SPARK-20989][CORE] Fail to start multiple workers on one host if external... · ef162289
      Xingbo Jiang authored
      [SPARK-20989][CORE] Fail to start multiple workers on one host if external shuffle service is enabled in standalone mode
      
      ## What changes were proposed in this pull request?
      
      In standalone mode, if we enable external shuffle service by setting `spark.shuffle.service.enabled` to true, and then we try to start multiple workers on one host(by setting `SPARK_WORKER_INSTANCES=3` in spark-env.sh, and then run `sbin/start-slaves.sh`), we can only launch one worker on each host successfully and the rest of the workers fail to launch.
      The reason is the port of external shuffle service if configed by `spark.shuffle.service.port`, so currently we could start no more than one external shuffle service on each host. In our case, each worker tries to start a external shuffle service, and only one of them succeeded doing this.
      
      We should give explicit reason of failure instead of fail silently.
      
      ## How was this patch tested?
      Manually test by the following steps:
      1. SET `SPARK_WORKER_INSTANCES=1` in `conf/spark-env.sh`;
      2. SET `spark.shuffle.service.enabled` to `true` in `conf/spark-defaults.conf`;
      3. Run `sbin/start-all.sh`.
      
      Before the change, you will see no error in the command line, as the following:
      ```
      starting org.apache.spark.deploy.master.Master, logging to /Users/xxx/workspace/spark/logs/spark-xxx-org.apache.spark.deploy.master.Master-1-xxx.local.out
      localhost: starting org.apache.spark.deploy.worker.Worker, logging to /Users/xxx/workspace/spark/logs/spark-xxx-org.apache.spark.deploy.worker.Worker-1-xxx.local.out
      localhost: starting org.apache.spark.deploy.worker.Worker, logging to /Users/xxx/workspace/spark/logs/spark-xxx-org.apache.spark.deploy.worker.Worker-2-xxx.local.out
      localhost: starting org.apache.spark.deploy.worker.Worker, logging to /Users/xxx/workspace/spark/logs/spark-xxx-org.apache.spark.deploy.worker.Worker-3-xxx.local.out
      ```
      And you can see in the webUI that only one worker is running.
      
      After the change, you get explicit error messages in the command line:
      ```
      starting org.apache.spark.deploy.master.Master, logging to /Users/xxx/workspace/spark/logs/spark-xxx-org.apache.spark.deploy.master.Master-1-xxx.local.out
      localhost: starting org.apache.spark.deploy.worker.Worker, logging to /Users/xxx/workspace/spark/logs/spark-xxx-org.apache.spark.deploy.worker.Worker-1-xxx.local.out
      localhost: failed to launch: nice -n 0 /Users/xxx/workspace/spark/bin/spark-class org.apache.spark.deploy.worker.Worker --webui-port 8081 spark://xxx.local:7077
      localhost:   17/06/13 23:24:53 INFO SecurityManager: Changing view acls to: xxx
      localhost:   17/06/13 23:24:53 INFO SecurityManager: Changing modify acls to: xxx
      localhost:   17/06/13 23:24:53 INFO SecurityManager: Changing view acls groups to:
      localhost:   17/06/13 23:24:53 INFO SecurityManager: Changing modify acls groups to:
      localhost:   17/06/13 23:24:53 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users  with view permissions: Set(xxx); groups with view permissions: Set(); users  with modify permissions: Set(xxx); groups with modify permissions: Set()
      localhost:   17/06/13 23:24:54 INFO Utils: Successfully started service 'sparkWorker' on port 63354.
      localhost:   Exception in thread "main" java.lang.IllegalArgumentException: requirement failed: Start multiple worker on one host failed because we may launch no more than one external shuffle service on each host, please set spark.shuffle.service.enabled to false or set SPARK_WORKER_INSTANCES to 1 to resolve the conflict.
      localhost:   	at scala.Predef$.require(Predef.scala:224)
      localhost:   	at org.apache.spark.deploy.worker.Worker$.main(Worker.scala:752)
      localhost:   	at org.apache.spark.deploy.worker.Worker.main(Worker.scala)
      localhost: full log in /Users/xxx/workspace/spark/logs/spark-xxx-org.apache.spark.deploy.worker.Worker-1-xxx.local.out
      localhost: starting org.apache.spark.deploy.worker.Worker, logging to /Users/xxx/workspace/spark/logs/spark-xxx-org.apache.spark.deploy.worker.Worker-2-xxx.local.out
      localhost: failed to launch: nice -n 0 /Users/xxx/workspace/spark/bin/spark-class org.apache.spark.deploy.worker.Worker --webui-port 8082 spark://xxx.local:7077
      localhost:   17/06/13 23:24:56 INFO SecurityManager: Changing view acls to: xxx
      localhost:   17/06/13 23:24:56 INFO SecurityManager: Changing modify acls to: xxx
      localhost:   17/06/13 23:24:56 INFO SecurityManager: Changing view acls groups to:
      localhost:   17/06/13 23:24:56 INFO SecurityManager: Changing modify acls groups to:
      localhost:   17/06/13 23:24:56 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users  with view permissions: Set(xxx); groups with view permissions: Set(); users  with modify permissions: Set(xxx); groups with modify permissions: Set()
      localhost:   17/06/13 23:24:56 INFO Utils: Successfully started service 'sparkWorker' on port 63359.
      localhost:   Exception in thread "main" java.lang.IllegalArgumentException: requirement failed: Start multiple worker on one host failed because we may launch no more than one external shuffle service on each host, please set spark.shuffle.service.enabled to false or set SPARK_WORKER_INSTANCES to 1 to resolve the conflict.
      localhost:   	at scala.Predef$.require(Predef.scala:224)
      localhost:   	at org.apache.spark.deploy.worker.Worker$.main(Worker.scala:752)
      localhost:   	at org.apache.spark.deploy.worker.Worker.main(Worker.scala)
      localhost: full log in /Users/xxx/workspace/spark/logs/spark-xxx-org.apache.spark.deploy.worker.Worker-2-xxx.local.out
      localhost: starting org.apache.spark.deploy.worker.Worker, logging to /Users/xxx/workspace/spark/logs/spark-xxx-org.apache.spark.deploy.worker.Worker-3-xxx.local.out
      localhost: failed to launch: nice -n 0 /Users/xxx/workspace/spark/bin/spark-class org.apache.spark.deploy.worker.Worker --webui-port 8083 spark://xxx.local:7077
      localhost:   17/06/13 23:24:59 INFO SecurityManager: Changing view acls to: xxx
      localhost:   17/06/13 23:24:59 INFO SecurityManager: Changing modify acls to: xxx
      localhost:   17/06/13 23:24:59 INFO SecurityManager: Changing view acls groups to:
      localhost:   17/06/13 23:24:59 INFO SecurityManager: Changing modify acls groups to:
      localhost:   17/06/13 23:24:59 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users  with view permissions: Set(xxx); groups with view permissions: Set(); users  with modify permissions: Set(xxx); groups with modify permissions: Set()
      localhost:   17/06/13 23:24:59 INFO Utils: Successfully started service 'sparkWorker' on port 63360.
      localhost:   Exception in thread "main" java.lang.IllegalArgumentException: requirement failed: Start multiple worker on one host failed because we may launch no more than one external shuffle service on each host, please set spark.shuffle.service.enabled to false or set SPARK_WORKER_INSTANCES to 1 to resolve the conflict.
      localhost:   	at scala.Predef$.require(Predef.scala:224)
      localhost:   	at org.apache.spark.deploy.worker.Worker$.main(Worker.scala:752)
      localhost:   	at org.apache.spark.deploy.worker.Worker.main(Worker.scala)
      localhost: full log in /Users/xxx/workspace/spark/logs/spark-xxx-org.apache.spark.deploy.worker.Worker-3-xxx.local.out
      ```
      
      Author: Xingbo Jiang <xingbo.jiang@databricks.com>
      
      Closes #18290 from jiangxb1987/start-slave.
      ef162289
  3. Jan 06, 2017
  4. Dec 01, 2016
    • Yuming Wang's avatar
      [SPARK-18645][DEPLOY] Fix spark-daemon.sh arguments error lead to throws Unrecognized option · 2ab8551e
      Yuming Wang authored
      ## What changes were proposed in this pull request?
      
      spark-daemon.sh will lost single quotes around after #15338. as follows:
      ```
      execute_command nice -n 0 bash /opt/cloudera/parcels/SPARK-2.1.0-cdh5.4.3.d20161129-21.04.38/lib/spark/bin/spark-submit --class org.apache.spark.sql.hive.thriftserver.HiveThriftServer2 --name Thrift JDBC/ODBC Server --conf spark.driver.extraJavaOptions=-XX:+UseG1GC -XX:-HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=/tmp
      ```
      With this fix, as follows:
      ```
      execute_command nice -n 0 bash /opt/cloudera/parcels/SPARK-2.1.0-cdh5.4.3.d20161129-21.04.38/lib/spark/bin/spark-submit --class org.apache.spark.sql.hive.thriftserver.HiveThriftServer2 --name 'Thrift JDBC/ODBC Server' --conf 'spark.driver.extraJavaOptions=-XX:+UseG1GC -XX:-HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=/tmp'
      ```
      
      ## How was this patch tested?
      
      - Manual tests
      - Build the package and start-thriftserver.sh with `--conf 'spark.driver.extraJavaOptions=-XX:+UseG1GC -XX:-HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=/tmp'`
      
      Author: Yuming Wang <wgyumg@gmail.com>
      
      Closes #16079 from wangyum/SPARK-18645.
      2ab8551e
  5. Oct 22, 2016
    • Erik O'Shaughnessy's avatar
      [SPARK-17944][DEPLOY] sbin/start-* scripts use of `hostname -f` fail with Solaris · 625fddda
      Erik O'Shaughnessy authored
      ## What changes were proposed in this pull request?
      
      Modify sbin/start-master.sh, sbin/start-mesos-dispatcher.sh and sbin/start-slaves.sh to use the output of 'uname' to select which OS-specific command-line is used to determine the host's fully qualified host name.
      
      ## How was this patch tested?
      
      Tested by hand; starting on Solaris, Linux and macOS.
      
      Author: Erik O'Shaughnessy <erik.oshaughnessy@gmail.com>
      
      Closes #15557 from JnyJny/SPARK-17944.
      625fddda
  6. Oct 21, 2016
    • Jagadeesan's avatar
      [SPARK-17960][PYSPARK][UPGRADE TO PY4J 0.10.4] · 595893d3
      Jagadeesan authored
      ## What changes were proposed in this pull request?
      
      1) Upgrade the Py4J version on the Java side
      2) Update the py4j src zip file we bundle with Spark
      
      ## How was this patch tested?
      
      Existing doctests & unit tests pass
      
      Author: Jagadeesan <as2@us.ibm.com>
      
      Closes #15514 from jagadeesanas2/SPARK-17960.
      595893d3
  7. Oct 20, 2016
    • Mike Ihbe's avatar
      [SPARK-11653][DEPLOY] Allow spark-daemon.sh to run in the foreground · c2c107ab
      Mike Ihbe authored
      ## What changes were proposed in this pull request?
      
      Add a SPARK_NO_DAEMONIZE environment variable flag to spark-daemon.sh that causes the process it would run to be run in the foreground.
      
      It looks like there has been some prior work in https://github.com/apache/spark/pull/3881, but there was some talk about these being refactored. I'm not sure if that happened or not, but that PR is almost 2 years old at this point so it was worth revisiting.
      
      ## How was this patch tested?
      
      ./dev/run-tests still seems to work. It doesn't look like these scripts have tests, but if I missed them just let me know.
      
      Author: Mike Ihbe <mikejihbe@gmail.com>
      
      Closes #15338 from mikejihbe/SPARK-11653.
      c2c107ab
  8. Oct 03, 2016
  9. Sep 18, 2016
    • Sean Owen's avatar
      [SPARK-17546][DEPLOY] start-* scripts should use hostname -f · 342c0e65
      Sean Owen authored
      ## What changes were proposed in this pull request?
      
      Call `hostname -f` to get fully qualified host name
      
      ## How was this patch tested?
      
      Jenkins tests of course, but also verified output of command on OS X and Linux
      
      Author: Sean Owen <sowen@cloudera.com>
      
      Closes #15129 from srowen/SPARK-17546.
      342c0e65
  10. Sep 11, 2016
    • Bryan Cutler's avatar
      [SPARK-17336][PYSPARK] Fix appending multiple times to PYTHONPATH from spark-config.sh · c76baff0
      Bryan Cutler authored
      ## What changes were proposed in this pull request?
      During startup of Spark standalone, the script file spark-config.sh appends to the PYTHONPATH and can be sourced many times, causing duplicates in the path.  This change adds a env flag that is set when the PYTHONPATH is appended so it will happen only one time.
      
      ## How was this patch tested?
      Manually started standalone master/worker and verified PYTHONPATH has no duplicate entries.
      
      Author: Bryan Cutler <cutlerb@gmail.com>
      
      Closes #15028 from BryanCutler/fix-duplicate-pythonpath-SPARK-17336.
      c76baff0
  11. Aug 24, 2016
    • Sean Owen's avatar
      [SPARK-16781][PYSPARK] java launched by PySpark as gateway may not be the same... · 0b3a4be9
      Sean Owen authored
      [SPARK-16781][PYSPARK] java launched by PySpark as gateway may not be the same java used in the spark environment
      
      ## What changes were proposed in this pull request?
      
      Update to py4j 0.10.3 to enable JAVA_HOME support
      
      ## How was this patch tested?
      
      Pyspark tests
      
      Author: Sean Owen <sowen@cloudera.com>
      
      Closes #14748 from srowen/SPARK-16781.
      0b3a4be9
  12. Jun 12, 2016
    • bomeng's avatar
      [SPARK-15806][DOCUMENTATION] update doc for SPARK_MASTER_IP · 50248dcf
      bomeng authored
      ## What changes were proposed in this pull request?
      
      SPARK_MASTER_IP is a deprecated environment variable. It is replaced by SPARK_MASTER_HOST according to MasterArguments.scala.
      
      ## How was this patch tested?
      
      Manually verified.
      
      Author: bomeng <bmeng@us.ibm.com>
      
      Closes #13543 from bomeng/SPARK-15806.
      50248dcf
  13. May 20, 2016
    • WeichenXu's avatar
      [SPARK-15203][DEPLOY] The spark daemon shell script error, daemon process... · a3ceb875
      WeichenXu authored
      [SPARK-15203][DEPLOY] The spark daemon shell script error, daemon process start successfully but script output fail message
      
      ## What changes were proposed in this pull request?
      
      fix the bug:
      The spark daemon shell script error, daemon process start successfully but script output fail message
      
      ## How was this patch tested?
      
      existing test.
      
      Author: WeichenXu <WeichenXu123@outlook.com>
      
      Closes #13172 from WeichenXu123/fix-spark-15203.
      a3ceb875
  14. May 13, 2016
  15. Mar 14, 2016
  16. Feb 27, 2016
    • Reynold Xin's avatar
      [SPARK-13521][BUILD] Remove reference to Tachyon in cluster & release scripts · 59e3e10b
      Reynold Xin authored
      ## What changes were proposed in this pull request?
      We provide a very limited set of cluster management script in Spark for Tachyon, although Tachyon itself provides a much better version of it. Given now Spark users can simply use Tachyon as a normal file system and does not require extensive configurations, we can remove this management capabilities to simplify Spark bash scripts.
      
      Note that this also reduces coupling between a 3rd party external system and Spark's release scripts, and would eliminate possibility for failures such as Tachyon being renamed or the tar balls being relocated.
      
      ## How was this patch tested?
      N/A
      
      Author: Reynold Xin <rxin@databricks.com>
      
      Closes #11400 from rxin/release-script.
      59e3e10b
  17. Feb 20, 2016
    • Timothy Chen's avatar
      [SPARK-13414][MESOS] Allow multiple dispatchers to be launched. · a4a081d1
      Timothy Chen authored
      ## What changes were proposed in this pull request?
      
      Users might want to start multiple mesos dispatchers, as each dispatcher can potentially be part of different roles and used for multi-tenancy.
      
      To allow multiple Mesos dispatchers to be launched, we need to be able to specify a instance number when starting the dispatcher daemon.
      
      ## How was the this patch tested?
      
      Manual testing
      
      Author: Timothy Chen <tnachen@gmail.com>
      
      Closes #11281 from tnachen/multiple_cluster_dispatchers.
      a4a081d1
  18. Jan 12, 2016
  19. Nov 09, 2015
    • Charles Yeh's avatar
      [SPARK-11218][CORE] show help messages for start-slave and start-master · 9e48cdfb
      Charles Yeh authored
      Addressing https://issues.apache.org/jira/browse/SPARK-11218, mostly copied start-thriftserver.sh.
      ```
      charlesyeh-mbp:spark charlesyeh$ ./sbin/start-master.sh --help
      Usage: Master [options]
      
      Options:
        -i HOST, --ip HOST     Hostname to listen on (deprecated, please use --host or -h)
        -h HOST, --host HOST   Hostname to listen on
        -p PORT, --port PORT   Port to listen on (default: 7077)
        --webui-port PORT      Port for web UI (default: 8080)
        --properties-file FILE Path to a custom Spark properties file.
                               Default is conf/spark-defaults.conf.
      ```
      ```
      charlesyeh-mbp:spark charlesyeh$ ./sbin/start-slave.sh
      Usage: Worker [options] <master>
      
      Master must be a URL of the form spark://hostname:port
      
      Options:
        -c CORES, --cores CORES  Number of cores to use
        -m MEM, --memory MEM     Amount of memory to use (e.g. 1000M, 2G)
        -d DIR, --work-dir DIR   Directory to run apps in (default: SPARK_HOME/work)
        -i HOST, --ip IP         Hostname to listen on (deprecated, please use --host or -h)
        -h HOST, --host HOST     Hostname to listen on
        -p PORT, --port PORT     Port to listen on (default: random)
        --webui-port PORT        Port for web UI (default: 8081)
        --properties-file FILE   Path to a custom Spark properties file.
                                 Default is conf/spark-defaults.conf.
      ```
      
      Author: Charles Yeh <charlesyeh@dropbox.com>
      
      Closes #9432 from CharlesYeh/helpmsg.
      9e48cdfb
  20. Nov 04, 2015
    • jerryshao's avatar
      [SPARK-2960][DEPLOY] Support executing Spark from symlinks (reopen) · 8aff36e9
      jerryshao authored
      This PR is based on the work of roji to support running Spark scripts from symlinks. Thanks for the great work roji . Would you mind taking a look at this PR, thanks a lot.
      
      For releases like HDP and others, normally it will expose the Spark executables as symlinks and put in `PATH`, but current Spark's scripts do not support finding real path from symlink recursively, this will make spark fail to execute from symlink. This PR try to solve this issue by finding the absolute path from symlink.
      
      Instead of using `readlink -f` like what this PR (https://github.com/apache/spark/pull/2386) implemented is that `-f` is not support for Mac, so here manually seeking the path through loop.
      
      I've tested with Mac and Linux (Cent OS), looks fine.
      
      This PR did not fix the scripts under `sbin` folder, not sure if it needs to be fixed also?
      
      Please help to review, any comment is greatly appreciated.
      
      Author: jerryshao <sshao@hortonworks.com>
      Author: Shay Rojansky <roji@roji.org>
      
      Closes #8669 from jerryshao/SPARK-2960.
      8aff36e9
  21. Oct 20, 2015
  22. Oct 02, 2015
    • Joshi's avatar
      [SPARK-10317] [CORE] Compatibility between history server script and functionality · f85aa064
      Joshi authored
      Compatibility between history server script and functionality
      
      The history server has its argument parsing class in HistoryServerArguments. However, this doesn't get involved in the start-history-server.sh codepath where the $0 arg is assigned to spark.history.fs.logDirectory and all other arguments discarded (e.g --property-file.)
      This stops the other options being usable from this script
      
      Author: Joshi <rekhajoshm@gmail.com>
      Author: Rekha Joshi <rekhajoshm@gmail.com>
      
      Closes #8758 from rekhajoshm/SPARK-10317.
      f85aa064
  23. Aug 03, 2015
    • Steve Loughran's avatar
      [SPARK-8064] [SQL] Build against Hive 1.2.1 · a2409d1c
      Steve Loughran authored
      Cherry picked the parts of the initial SPARK-8064 WiP branch needed to get sql/hive to compile against hive 1.2.1. That's the ASF release packaged under org.apache.hive, not any fork.
      
      Tests not run yet: that's what the machines are for
      
      Author: Steve Loughran <stevel@hortonworks.com>
      Author: Cheng Lian <lian@databricks.com>
      Author: Michael Armbrust <michael@databricks.com>
      Author: Patrick Wendell <patrick@databricks.com>
      
      Closes #7191 from steveloughran/stevel/feature/SPARK-8064-hive-1.2-002 and squashes the following commits:
      
      7556d85 [Cheng Lian] Updates .q files and corresponding golden files
      ef4af62 [Steve Loughran] Merge commit '6a92bb09f46a04d6cd8c41bdba3ecb727ebb9030' into stevel/feature/SPARK-8064-hive-1.2-002
      6a92bb0 [Cheng Lian] Overrides HiveConf time vars
      dcbb391 [Cheng Lian] Adds com.twitter:parquet-hadoop-bundle:1.6.0 for Hive Parquet SerDe
      0bbe475 [Steve Loughran] SPARK-8064 scalastyle rejects the standard Hadoop ASF license header...
      fdf759b [Steve Loughran] SPARK-8064 classpath dependency suite to be in sync with shading in final (?) hive-exec spark
      7a6c727 [Steve Loughran] SPARK-8064 switch to second staging repo of the spark-hive artifacts. This one has the protobuf-shaded hive-exec jar
      376c003 [Steve Loughran] SPARK-8064 purge duplicate protobuf declaration
      2c74697 [Steve Loughran] SPARK-8064 switch to the protobuf shaded hive-exec jar with tests to chase it down
      cc44020 [Steve Loughran] SPARK-8064 remove hadoop.version from runtest.py, as profile will fix that automatically.
      6901fa9 [Steve Loughran] SPARK-8064 explicit protobuf import
      da310dc [Michael Armbrust] Fixes for Hive tests.
      a775a75 [Steve Loughran] SPARK-8064 cherry-pick-incomplete
      7404f34 [Patrick Wendell] Add spark-hive staging repo
      832c164 [Steve Loughran] SPARK-8064 try to supress compiler warnings on Complex.java pasted-thrift-code
      312c0d4 [Steve Loughran] SPARK-8064  maven/ivy dependency purge; calcite declaration needed
      fa5ae7b [Steve Loughran] HIVE-8064 fix up hive-thriftserver dependencies and cut back on evicted references in the hive- packages; this keeps mvn and ivy resolution compatible, as the reconciliation policy is "by hand"
      c188048 [Steve Loughran] SPARK-8064 manage the Hive depencencies to that -things that aren't needed are excluded -sql/hive built with ivy is in sync with the maven reconciliation policy, rather than latest-first
      4c8be8d [Cheng Lian] WIP: Partial fix for Thrift server and CLI tests
      314eb3c [Steve Loughran] SPARK-8064 deprecation warning  noise in one of the tests
      17b0341 [Steve Loughran] SPARK-8064 IDE-hinted cleanups of Complex.java to reduce compiler warnings. It's all autogenerated code, so still ugly.
      d029b92 [Steve Loughran] SPARK-8064 rely on unescaping to have already taken place, so go straight to map of serde options
      23eca7e [Steve Loughran] HIVE-8064 handle raw and escaped property tokens
      54d9b06 [Steve Loughran] SPARK-8064 fix compilation regression surfacing from rebase
      0b12d5f [Steve Loughran] HIVE-8064 use subset of hive complex type whose types deserialize
      fce73b6 [Steve Loughran] SPARK-8064 poms rely implicitly on the version of kryo chill provides
      fd3aa5d [Steve Loughran] SPARK-8064 version of hive to d/l from ivy is 1.2.1
      dc73ece [Steve Loughran] SPARK-8064 revert to master's determinstic pushdown strategy
      d3c1e4a [Steve Loughran] SPARK-8064 purge UnionType
      051cc21 [Steve Loughran] SPARK-8064 switch to an unshaded version of hive-exec-core, which must have been built with Kryo 2.21. This currently looks for a (locally built) version 1.2.1.spark
      6684c60 [Steve Loughran] SPARK-8064 ignore RTE raised in blocking process.exitValue() call
      e6121e5 [Steve Loughran] SPARK-8064 address review comments
      aa43dc6 [Steve Loughran] SPARK-8064  more robust teardown on JavaMetastoreDatasourcesSuite
      f2bff01 [Steve Loughran] SPARK-8064 better takeup of asynchronously caught error text
      8b1ef38 [Steve Loughran] SPARK-8064: on failures executing spark-submit in HiveSparkSubmitSuite, print command line and all logged output.
      5a9ce6b [Steve Loughran] SPARK-8064 add explicit reason for kv split failure, rather than array OOB. *does not address the issue*
      642b63a [Steve Loughran] SPARK-8064 reinstate something cut briefly during rebasing
      97194dc [Steve Loughran] SPARK-8064 add extra logging to the YarnClusterSuite classpath test. There should be no reason why this is failing on jenkins, but as it is (and presumably its CP-related), improve the logging including any exception raised.
      335357f [Steve Loughran] SPARK-8064 fail fast on thrive process spawning tests on exit codes and/or error string patterns seen in log.
      3ed872f [Steve Loughran] SPARK-8064 rename field double to  dbl
      bca55e5 [Steve Loughran] SPARK-8064 missed one of the `date` escapes
      41d6479 [Steve Loughran] SPARK-8064 wrap tests with withTable() calls to avoid table-exists exceptions
      2bc29a4 [Steve Loughran] SPARK-8064 ParquetSuites to escape `date` field name
      1ab9bc4 [Steve Loughran] SPARK-8064 TestHive to use sered2.thrift.test.Complex
      bf3a249 [Steve Loughran] SPARK-8064: more resubmit than fix; tighten startup timeout to 60s. Still no obvious reason why jersey server code in spark-assembly isn't being picked up -it hasn't been shaded
      c829b8f [Steve Loughran] SPARK-8064: reinstate yarn-rm-server dependencies to hive-exec to ensure that jersey server is on classpath on hadoop versions < 2.6
      0b0f738 [Steve Loughran] SPARK-8064: thrift server startup to fail fast on any exception in the main thread
      13abaf1 [Steve Loughran] SPARK-8064 Hive compatibilty tests sin sync with explain/show output from Hive 1.2.1
      d14d5ea [Steve Loughran] SPARK-8064: DATE is now a predicate; you can't use it as a field in select ops
      26eef1c [Steve Loughran] SPARK-8064: HIVE-9039 renamed TOK_UNION => TOK_UNIONALL while adding TOK_UNIONDISTINCT
      3d64523 [Steve Loughran] SPARK-8064 improve diagns on uknown token; fix scalastyle failure
      d0360f6 [Steve Loughran] SPARK-8064: delicate merge in of the branch vanzin/hive-1.1
      1126e5a [Steve Loughran] SPARK-8064: name of unrecognized file format wasn't appearing in error text
      8cb09c4 [Steve Loughran] SPARK-8064: test resilience/assertion improvements. Independent of the rest of the work; can be backported to earlier versions
      dec12cb [Steve Loughran] SPARK-8064: when a CLI suite test fails include the full output text in the raised exception; this ensures that the stdout/stderr is included in jenkins reports, so it becomes possible to diagnose the cause.
      463a670 [Steve Loughran] SPARK-8064 run-tests.py adds a hadoop-2.6 profile, and changes info messages to say "w/Hive 1.2.1" in console output
      2531099 [Steve Loughran] SPARK-8064 successful attempt to get rid of pentaho as a transitive dependency of hive-exec
      1d59100 [Steve Loughran] SPARK-8064 (unsuccessful) attempt to get rid of pentaho as a transitive dependency of hive-exec
      75733fc [Steve Loughran] SPARK-8064 change thrift binary startup message to "Starting ThriftBinaryCLIService on port"
      3ebc279 [Steve Loughran] SPARK-8064 move strings used to check for http/bin thrift services up into constants
      c80979d [Steve Loughran] SPARK-8064: SparkSQLCLIDriver drops remote mode support. CLISuite Tests pass instead of timing out: undetected regression?
      27e8370 [Steve Loughran] SPARK-8064 fix some style & IDE warnings
      00e50d6 [Steve Loughran] SPARK-8064 stop excluding hive shims from dependency (commented out , for now)
      cb4f142 [Steve Loughran] SPARK-8054 cut pentaho dependency from calcite
      f7aa9cb [Steve Loughran] SPARK-8064 everything compiles with some commenting and moving of classes into a hive package
      6c310b4 [Steve Loughran] SPARK-8064 subclass  Hive ServerOptionsProcessor to make it public again
      f61a675 [Steve Loughran] SPARK-8064 thrift server switched to Hive 1.2.1, though it doesn't compile everywhere
      4890b9d [Steve Loughran] SPARK-8064, build against Hive 1.2.1
      a2409d1c
    • Timothy Chen's avatar
      [SPARK-8873] [MESOS] Clean up shuffle files if external shuffle service is used · 95dccc63
      Timothy Chen authored
      This patch builds directly on #7820, which is largely written by tnachen. The only addition is one commit for cleaning up the code. There should be no functional differences between this and #7820.
      
      Author: Timothy Chen <tnachen@gmail.com>
      Author: Andrew Or <andrew@databricks.com>
      
      Closes #7881 from andrewor14/tim-cleanup-mesos-shuffle and squashes the following commits:
      
      8894f7d [Andrew Or] Clean up code
      2a5fa10 [Andrew Or] Merge branch 'mesos_shuffle_clean' of github.com:tnachen/spark into tim-cleanup-mesos-shuffle
      fadff89 [Timothy Chen] Address comments.
      e4d0f1d [Timothy Chen] Clean up external shuffle data on driver exit with Mesos.
      95dccc63
  24. May 15, 2015
    • Sean Owen's avatar
      [SPARK-5412] [DEPLOY] Cannot bind Master to a specific hostname as per the documentation · 8ab1450d
      Sean Owen authored
      Pass args to start-master.sh through to start-daemon.sh, as other scripts do, so that things like --host have effect on start-master.sh as per docs
      
      Author: Sean Owen <sowen@cloudera.com>
      
      Closes #6185 from srowen/SPARK-5412 and squashes the following commits:
      
      b3ce9da [Sean Owen] Pass args to start-master.sh through to start-daemon.sh, as other scripts do, so that things like --host have effect on start-master.sh as per docs
      8ab1450d
  25. Apr 28, 2015
    • Timothy Chen's avatar
      [SPARK-5338] [MESOS] Add cluster mode support for Mesos · 53befacc
      Timothy Chen authored
      This patch adds the support for cluster mode to run on Mesos.
      It introduces a new Mesos framework dedicated to launch new apps/drivers, and can be called with the spark-submit script and specifying --master flag to the cluster mode REST interface instead of Mesos master.
      
      Example:
      ./bin/spark-submit --deploy-mode cluster --class org.apache.spark.examples.SparkPi --master mesos://10.0.0.206:8077 --executor-memory 1G --total-executor-cores 100 examples/target/spark-examples_2.10-1.3.0-SNAPSHOT.jar 30
      
      Part of this patch is also to abstract the StandaloneRestServer so it can have different implementations of the REST endpoints.
      
      Features of the cluster mode in this PR:
      - Supports supervise mode where scheduler will keep trying to reschedule exited job.
      - Adds a new UI for the cluster mode scheduler to see all the running jobs, finished jobs, and supervise jobs waiting to be retried
      - Supports state persistence to ZK, so when the cluster scheduler fails over it can pick up all the queued and running jobs
      
      Author: Timothy Chen <tnachen@gmail.com>
      Author: Luc Bourlier <luc.bourlier@typesafe.com>
      
      Closes #5144 from tnachen/mesos_cluster_mode and squashes the following commits:
      
      069e946 [Timothy Chen] Fix rebase.
      e24b512 [Timothy Chen] Persist submitted driver.
      390c491 [Timothy Chen] Fix zk conf key for mesos zk engine.
      e324ac1 [Timothy Chen] Fix merge.
      fd5259d [Timothy Chen] Address review comments.
      1553230 [Timothy Chen] Address review comments.
      c6c6b73 [Timothy Chen] Pass spark properties to mesos cluster tasks.
      f7d8046 [Timothy Chen] Change app name to spark cluster.
      17f93a2 [Timothy Chen] Fix head of line blocking in scheduling drivers.
      6ff8e5c [Timothy Chen] Address comments and add logging.
      df355cd [Timothy Chen] Add metrics to mesos cluster scheduler.
      20f7284 [Timothy Chen] Address review comments
      7252612 [Timothy Chen] Fix tests.
      a46ad66 [Timothy Chen] Allow zk cli param override.
      920fc4b [Timothy Chen] Fix scala style issues.
      862b5b5 [Timothy Chen] Support asking driver status when it's retrying.
      7f214c2 [Timothy Chen] Fix RetryState visibility
      e0f33f7 [Timothy Chen] Add supervise support and persist retries.
      371ce65 [Timothy Chen] Handle cluster mode recovery and state persistence.
      3d4dfa1 [Luc Bourlier] Adds support to kill submissions
      febfaba [Timothy Chen] Bound the finished drivers in memory
      543a98d [Timothy Chen] Schedule multiple jobs
      6887e5e [Timothy Chen] Support looking at SPARK_EXECUTOR_URI env variable in schedulers
      8ec76bc [Timothy Chen] Fix Mesos dispatcher UI.
      d57d77d [Timothy Chen] Add documentation
      825afa0 [Luc Bourlier] Supports more spark-submit parameters
      b8e7181 [Luc Bourlier] Adds a shutdown latch to keep the deamon running
      0fa7780 [Luc Bourlier] Launch task through the mesos scheduler
      5b7a12b [Timothy Chen] WIP: Making a cluster mode a mesos framework.
      4b2f5ef [Timothy Chen] Specify user jar in command to be replaced with local.
      e775001 [Timothy Chen] Support fetching remote uris in driver runner.
      7179495 [Timothy Chen] Change Driver page output and add logging
      880bc27 [Timothy Chen] Add Mesos Cluster UI to display driver results
      9986731 [Timothy Chen] Kill drivers when shutdown
      67cbc18 [Timothy Chen] Rename StandaloneRestClient to RestClient and add sbin scripts
      e3facdd [Timothy Chen] Add Mesos Cluster dispatcher
      53befacc
    • Iulian Dragos's avatar
      [SPARK-4286] Add an external shuffle service that can be run as a daemon. · 8aab94d8
      Iulian Dragos authored
      This allows Mesos deployments to use the shuffle service (and implicitly dynamic allocation). It does so by adding a new "main" class and two corresponding scripts in `sbin`:
      
      - `sbin/start-shuffle-service.sh`
      - `sbin/stop-shuffle-service.sh`
      
      Specific options can be passed in `SPARK_SHUFFLE_OPTS`.
      
      This is picking up work from #3861 /cc tnachen
      
      Author: Iulian Dragos <jaguarul@gmail.com>
      
      Closes #4990 from dragos/feature/external-shuffle-service and squashes the following commits:
      
      6c2b148 [Iulian Dragos] Import order and wrong name fixup.
      07804ad [Iulian Dragos] Moved ExternalShuffleService to the `deploy` package + other minor tweaks.
      4dc1f91 [Iulian Dragos] Reviewer’s comments:
      8145429 [Iulian Dragos] Add an external shuffle service that can be run as a daemon.
      8aab94d8
  26. Apr 17, 2015
    • Punya Biswal's avatar
      [SPARK-6952] Handle long args when detecting PID reuse · f6a9a57a
      Punya Biswal authored
      sbin/spark-daemon.sh used
      
          ps -p "$TARGET_PID" -o args=
      
      to figure out whether the process running with the expected PID is actually a Spark
      daemon. When running with a large classpath, the output of ps gets
      truncated and the check fails spuriously.
      
      This weakens the check to see if it's a java command (which is something
      we do in other parts of the script) rather than looking for the specific
      main class name. This means that SPARK-4832 might happen under a
      slightly broader range of circumstances (a java program happened to
      reuse the same PID), but it seems worthwhile compared to failing
      consistently with a large classpath.
      
      Author: Punya Biswal <pbiswal@palantir.com>
      
      Closes #5535 from punya/feature/SPARK-6952 and squashes the following commits:
      
      7ea12d1 [Punya Biswal] Handle long args when detecting PID reuse
      f6a9a57a
  27. Apr 13, 2015
    • Nathan Kronenfeld's avatar
      [Spark-4848] Allow different Worker configurations in standalone cluster · 435b8779
      Nathan Kronenfeld authored
      This refixes #3699 with the latest code.
      This fixes SPARK-4848
      
      I've changed the stand-alone cluster scripts to allow different workers to have different numbers of instances, with both port and web-ui port following allong appropriately.
      
      I did this by moving the loop over instances from start-slaves and stop-slaves (on the master) to start-slave and stop-slave (on the worker).
      
      Wile I was at it, I changed SPARK_WORKER_PORT to work the same way as SPARK_WORKER_WEBUI_PORT, since the new methods work fine for both.
      
      Author: Nathan Kronenfeld <nkronenfeld@oculusinfo.com>
      
      Closes #5140 from nkronenfeld/feature/spark-4848 and squashes the following commits:
      
      cf5f47e [Nathan Kronenfeld] Merge remote branch 'upstream/master' into feature/spark-4848
      044ca6f [Nathan Kronenfeld] Documentation and formatting as requested by by andrewor14
      d739640 [Nathan Kronenfeld] Move looping through instances from the master to the workers, so that each worker respects its own number of instances and web-ui port
      435b8779
    • Pradeep Chanumolu's avatar
      [SPARK-6671] Add status command for spark daemons · 240ea03f
      Pradeep Chanumolu authored
      SPARK-6671
      Currently using the spark-daemon.sh script we can start and stop the spark demons. But we cannot get the status of the daemons. It will be nice to include the status command in the spark-daemon.sh script, through which we can know if the spark demon is alive or not.
      
      Author: Pradeep Chanumolu <pchanumolu@maprtech.com>
      
      Closes #5327 from pchanumolu/master and squashes the following commits:
      
      d3a1f05 [Pradeep Chanumolu] Make status command check consistent with Stop command
      5062926 [Pradeep Chanumolu] Fix indentation in spark-daemon.sh
      3e66bc8 [Pradeep Chanumolu] SPARK-6671 : Add status command to spark daemons
      1ac3918 [Pradeep Chanumolu] Add status command to spark-daemon
      240ea03f
  28. Mar 30, 2015
    • Jose Manuel Gomez's avatar
      [HOTFIX] Update start-slave.sh · 19d4c392
      Jose Manuel Gomez authored
      wihtout this change the below error happens when I execute sbin/start-all.sh
      
      localhost: /spark-1.3/sbin/start-slave.sh: line 32: unexpected EOF while looking for matching `"'
      localhost: /spark-1.3/sbin/start-slave.sh: line 33: syntax error: unexpected end of file
      
      my operating system is Linux Mint 17.1 Rebecca
      
      Author: Jose Manuel Gomez <jmgomez@stratio.com>
      
      Closes #5262 from josegom/patch-2 and squashes the following commits:
      
      453af8b [Jose Manuel Gomez] Update start-slave.sh
      2c456bd [Jose Manuel Gomez] Update start-slave.sh
      19d4c392
  29. Mar 28, 2015
  30. Mar 11, 2015
    • Marcelo Vanzin's avatar
      [SPARK-4924] Add a library for launching Spark jobs programmatically. · 517975d8
      Marcelo Vanzin authored
      This change encapsulates all the logic involved in launching a Spark job
      into a small Java library that can be easily embedded into other applications.
      
      The overall goal of this change is twofold, as described in the bug:
      
      - Provide a public API for launching Spark processes. This is a common request
        from users and currently there's no good answer for it.
      
      - Remove a lot of the duplicated code and other coupling that exists in the
        different parts of Spark that deal with launching processes.
      
      A lot of the duplication was due to different code needed to build an
      application's classpath (and the bootstrapper needed to run the driver in
      certain situations), and also different code needed to parse spark-submit
      command line options in different contexts. The change centralizes those
      as much as possible so that all code paths can rely on the library for
      handling those appropriately.
      
      Author: Marcelo Vanzin <vanzin@cloudera.com>
      
      Closes #3916 from vanzin/SPARK-4924 and squashes the following commits:
      
      18c7e4d [Marcelo Vanzin] Fix make-distribution.sh.
      2ce741f [Marcelo Vanzin] Add lots of quotes.
      3b28a75 [Marcelo Vanzin] Update new pom.
      a1b8af1 [Marcelo Vanzin] Merge branch 'master' into SPARK-4924
      897141f [Marcelo Vanzin] Review feedback.
      e2367d2 [Marcelo Vanzin] Merge branch 'master' into SPARK-4924
      28cd35e [Marcelo Vanzin] Remove stale comment.
      b1d86b0 [Marcelo Vanzin] Merge branch 'master' into SPARK-4924
      00505f9 [Marcelo Vanzin] Add blurb about new API in the programming guide.
      5f4ddcc [Marcelo Vanzin] Better usage messages.
      92a9cfb [Marcelo Vanzin] Fix Win32 launcher, usage.
      6184c07 [Marcelo Vanzin] Rename field.
      4c19196 [Marcelo Vanzin] Update comment.
      7e66c18 [Marcelo Vanzin] Fix pyspark tests.
      0031a8e [Marcelo Vanzin] Review feedback.
      c12d84b [Marcelo Vanzin] Review feedback. And fix spark-submit on Windows.
      e2d4d71 [Marcelo Vanzin] Simplify some code used to launch pyspark.
      43008a7 [Marcelo Vanzin] Don't make builder extend SparkLauncher.
      b4d6912 [Marcelo Vanzin] Use spark-submit script in SparkLauncher.
      28b1434 [Marcelo Vanzin] Add a comment.
      304333a [Marcelo Vanzin] Fix propagation of properties file arg.
      bb67b93 [Marcelo Vanzin] Remove unrelated Yarn change (that is also wrong).
      8ec0243 [Marcelo Vanzin] Add missing newline.
      95ddfa8 [Marcelo Vanzin] Fix handling of --help for spark-class command builder.
      72da7ec [Marcelo Vanzin] Rename SparkClassLauncher.
      62978e4 [Marcelo Vanzin] Minor cleanup of Windows code path.
      9cd5b44 [Marcelo Vanzin] Make all non-public APIs package-private.
      e4c80b6 [Marcelo Vanzin] Reorganize the code so that only SparkLauncher is public.
      e50dc5e [Marcelo Vanzin] Merge branch 'master' into SPARK-4924
      de81da2 [Marcelo Vanzin] Fix CommandUtils.
      86a87bf [Marcelo Vanzin] Merge branch 'master' into SPARK-4924
      2061967 [Marcelo Vanzin] Merge branch 'master' into SPARK-4924
      46d46da [Marcelo Vanzin] Clean up a test and make it more future-proof.
      b93692a [Marcelo Vanzin] Merge branch 'master' into SPARK-4924
      ad03c48 [Marcelo Vanzin] Revert "Fix a thread-safety issue in "local" mode."
      0b509d0 [Marcelo Vanzin] Merge branch 'master' into SPARK-4924
      23aa2a9 [Marcelo Vanzin] Read java-opts from conf dir, not spark home.
      7cff919 [Marcelo Vanzin] Javadoc updates.
      eae4d8e [Marcelo Vanzin] Fix new unit tests on Windows.
      e570fb5 [Marcelo Vanzin] Merge branch 'master' into SPARK-4924
      44cd5f7 [Marcelo Vanzin] Add package-info.java, clean up javadocs.
      f7cacff [Marcelo Vanzin] Remove "launch Spark in new thread" feature.
      7ed8859 [Marcelo Vanzin] Some more feedback.
      54cd4fd [Marcelo Vanzin] Merge branch 'master' into SPARK-4924
      61919df [Marcelo Vanzin] Clean leftover debug statement.
      aae5897 [Marcelo Vanzin] Use launcher classes instead of jars in non-release mode.
      e584fc3 [Marcelo Vanzin] Rework command building a little bit.
      525ef5b [Marcelo Vanzin] Rework Unix spark-class to handle argument with newlines.
      8ac4e92 [Marcelo Vanzin] Minor test cleanup.
      e946a99 [Marcelo Vanzin] Merge PySparkLauncher into SparkSubmitCliLauncher.
      c617539 [Marcelo Vanzin] Review feedback round 1.
      fc6a3e2 [Marcelo Vanzin] Merge branch 'master' into SPARK-4924
      f26556b [Marcelo Vanzin] Fix a thread-safety issue in "local" mode.
      2f4e8b4 [Marcelo Vanzin] Changes needed to make this work with SPARK-4048.
      799fc20 [Marcelo Vanzin] Merge branch 'master' into SPARK-4924
      bb5d324 [Marcelo Vanzin] Merge branch 'master' into SPARK-4924
      53faef1 [Marcelo Vanzin] Merge branch 'master' into SPARK-4924
      a7936ef [Marcelo Vanzin] Fix pyspark tests.
      656374e [Marcelo Vanzin] Mima fixes.
      4d511e7 [Marcelo Vanzin] Fix tools search code.
      7a01e4a [Marcelo Vanzin] Fix pyspark on Yarn.
      1b3f6e9 [Marcelo Vanzin] Call SparkSubmit from spark-class launcher for unknown classes.
      25c5ae6 [Marcelo Vanzin] Centralize SparkSubmit command line parsing.
      27be98a [Marcelo Vanzin] Modify Spark to use launcher lib.
      6f70eea [Marcelo Vanzin] [SPARK-4924] Add a library for launching Spark jobs programatically.
      517975d8
  31. Mar 07, 2015
    • WangTaoTheTonic's avatar
      [Minor]fix the wrong description · 729c05bd
      WangTaoTheTonic authored
      Found it by accident. I'm not gonna file jira for this as it is a very tiny fix.
      
      Author: WangTaoTheTonic <wangtao111@huawei.com>
      
      Closes #4936 from WangTaoTheTonic/wrongdesc and squashes the following commits:
      
      fb8a8ec [WangTaoTheTonic] fix the wrong description
      aca5596 [WangTaoTheTonic] fix the wrong description
      729c05bd
  32. Mar 06, 2015
    • Zhang, Liye's avatar
      [CORE, DEPLOY][minor] align arguments order with docs of worker · d8b3da9d
      Zhang, Liye authored
      The help message for starting `worker` is `Usage: Worker [options] <master>`. While in `start-slaves.sh`, the format is not align with that, it is confusing for the fist glance.
      
      Author: Zhang, Liye <liye.zhang@intel.com>
      
      Closes #4924 from liyezhang556520/startSlaves and squashes the following commits:
      
      7fd5deb [Zhang, Liye] align arguments order with docs of worker
      d8b3da9d
  33. Feb 19, 2015
    • Zhan Zhang's avatar
      [Spark-5889] Remove pid file after stopping service. · ad6b169d
      Zhan Zhang authored
      Currently the pid file is not deleted, and potentially may cause some problem after service is stopped. The fix remove the pid file after service stopped.
      
      Author: Zhan Zhang <zhazhan@gmail.com>
      
      Closes #4676 from zhzhan/spark-5889 and squashes the following commits:
      
      eb01be1 [Zhan Zhang] solve review comments
      b4c009e [Zhan Zhang] solve review comments
      018110a [Zhan Zhang] spark-5889: remove pid file after stopping service
      088d2a2 [Zhan Zhang] squash all commits
      c1f1fa5 [Zhan Zhang] test
      ad6b169d
    • Cheng Hao's avatar
      [SPARK-5825] [Spark Submit] Remove the double checking instance name when stopping the service · 94cdb05f
      Cheng Hao authored
      `spark-daemon.sh` will confirm the process id by fuzzy matching the class name while stopping the service, however, it will fail if the java process arguments is very long (greater than 4096 characters).
      This PR looses the check for the service process.
      
      Author: Cheng Hao <hao.cheng@intel.com>
      
      Closes #4611 from chenghao-intel/stopping_service and squashes the following commits:
      
      a0051f6 [Cheng Hao] loosen the process checking while stopping a service
      94cdb05f
  34. Feb 13, 2015
    • WangTaoTheTonic's avatar
      [SPARK-4832][Deploy]some other processes might take the daemon pid · 1768bd51
      WangTaoTheTonic authored
      Some other processes might use the pid saved in pid file. In that case we should ignore it and launch daemons.
      
      JIRA is down for maintenance. I will file one once it return.
      
      Author: WangTaoTheTonic <barneystinson@aliyun.com>
      Author: WangTaoTheTonic <wangtao111@huawei.com>
      
      Closes #3683 from WangTaoTheTonic/otherproc and squashes the following commits:
      
      daa86a1 [WangTaoTheTonic] some bash style fix
      8befee7 [WangTaoTheTonic] handle the mistake scenario
      cf4ecc6 [WangTaoTheTonic] remove redundant condition
      f36cfb4 [WangTaoTheTonic] some other processes might take the pid
      1768bd51
  35. Feb 01, 2015
    • Tom Panning's avatar
      [SPARK-5176] The thrift server does not support cluster mode · 1ca0a101
      Tom Panning authored
      Output an error message if the thrift server is started in cluster mode.
      
      Author: Tom Panning <tom.panning@nextcentury.com>
      
      Closes #4137 from tpanningnextcen/spark-5176-thrift-cluster-mode-error and squashes the following commits:
      
      f5c0509 [Tom Panning] [SPARK-5176] The thrift server does not support cluster mode
      1ca0a101
  36. Jan 19, 2015
    • Jongyoul Lee's avatar
      [SPARK-5088] Use spark-class for running executors directly · 4a4f9ccb
      Jongyoul Lee authored
      Author: Jongyoul Lee <jongyoul@gmail.com>
      
      Closes #3897 from jongyoul/SPARK-5088 and squashes the following commits:
      
      8232aa8 [Jongyoul Lee] [SPARK-5088] Use spark-class for running executors directly - Added a listenerBus for fixing test cases
      932289f [Jongyoul Lee] [SPARK-5088] Use spark-class for running executors directly - Rebased from master
      613cb47 [Jongyoul Lee] [SPARK-5088] Use spark-class for running executors directly - Fixed code if spark.executor.uri doesn't have any value - Added test cases
      ff57bda [Jongyoul Lee] [SPARK-5088] Use spark-class for running executors directly - Adjusted orders of import
      97e4bd4 [Jongyoul Lee] [SPARK-5088] Use spark-class for running executors directly - Changed command for using spark-class directly - Delete sbin/spark-executor and moved some codes into spark-class' case statement
      4a4f9ccb
Loading