Skip to content
Snippets Groups Projects
  1. May 26, 2017
    • Zheng RuiFeng's avatar
      [SPARK-20849][DOC][SPARKR] Document R DecisionTree · a97c4970
      Zheng RuiFeng authored
      ## What changes were proposed in this pull request?
      1, add an example for sparkr `decisionTree`
      2, document it in user guide
      
      ## How was this patch tested?
      local submit
      
      Author: Zheng RuiFeng <ruifengz@foxmail.com>
      
      Closes #18067 from zhengruifeng/dt_example.
      a97c4970
  2. May 25, 2017
    • Michael Allman's avatar
      [SPARK-20888][SQL][DOCS] Document change of default setting of... · c1e7989c
      Michael Allman authored
      [SPARK-20888][SQL][DOCS] Document change of default setting of spark.sql.hive.caseSensitiveInferenceMode
      
      (Link to Jira: https://issues.apache.org/jira/browse/SPARK-20888)
      
      ## What changes were proposed in this pull request?
      
      Document change of default setting of spark.sql.hive.caseSensitiveInferenceMode configuration key from NEVER_INFO to INFER_AND_SAVE in the Spark SQL 2.1 to 2.2 migration notes.
      
      Author: Michael Allman <michael@videoamp.com>
      
      Closes #18112 from mallman/spark-20888-document_infer_and_save.
      c1e7989c
    • jinxing's avatar
      [SPARK-19659] Fetch big blocks to disk when shuffle-read. · 3f94e64a
      jinxing authored
      ## What changes were proposed in this pull request?
      
      Currently the whole block is fetched into memory(off heap by default) when shuffle-read. A block is defined by (shuffleId, mapId, reduceId). Thus it can be large when skew situations. If OOM happens during shuffle read, job will be killed and users will be notified to "Consider boosting spark.yarn.executor.memoryOverhead". Adjusting parameter and allocating more memory can resolve the OOM. However the approach is not perfectly suitable for production environment, especially for data warehouse.
      Using Spark SQL as data engine in warehouse, users hope to have a unified parameter(e.g. memory) but less resource wasted(resource is allocated but not used). The hope is strong especially when migrating data engine to Spark from another one(e.g. Hive). Tuning the parameter for thousands of SQLs one by one is very time consuming.
      It's not always easy to predict skew situations, when happen, it make sense to fetch remote blocks to disk for shuffle-read, rather than kill the job because of OOM.
      
      In this pr, I propose to fetch big blocks to disk(which is also mentioned in SPARK-3019):
      
      1. Track average size and also the outliers(which are larger than 2*avgSize) in MapStatus;
      2. Request memory from `MemoryManager` before fetch blocks and release the memory to `MemoryManager` when `ManagedBuffer` is released.
      3. Fetch remote blocks to disk when failing acquiring memory from `MemoryManager`, otherwise fetch to memory.
      
      This is an improvement for memory control when shuffle blocks and help to avoid OOM in scenarios like below:
      1. Single huge block;
      2. Sizes of many blocks are underestimated in `MapStatus` and the actual footprint of blocks is much larger than the estimated.
      
      ## How was this patch tested?
      Added unit test in `MapStatusSuite` and `ShuffleBlockFetcherIteratorSuite`.
      
      Author: jinxing <jinxing6042@126.com>
      
      Closes #16989 from jinxing64/SPARK-19659.
      3f94e64a
  3. May 22, 2017
    • jinxing's avatar
      [SPARK-20801] Record accurate size of blocks in MapStatus when it's above threshold. · 2597674b
      jinxing authored
      ## What changes were proposed in this pull request?
      
      Currently, when number of reduces is above 2000, HighlyCompressedMapStatus is used to store size of blocks. in HighlyCompressedMapStatus, only average size is stored for non empty blocks. Which is not good for memory control when we shuffle blocks. It makes sense to store the accurate size of block when it's above threshold.
      
      ## How was this patch tested?
      
      Added test in MapStatusSuite.
      
      Author: jinxing <jinxing6042@126.com>
      
      Closes #18031 from jinxing64/SPARK-20801.
      2597674b
    • Nick Pentreath's avatar
      [SPARK-20506][DOCS] Add HTML links to highlight list in MLlib guide for 2.2 · be846db4
      Nick Pentreath authored
      Quick follow up to #17996 - forgot to add the HTML links to the relevant sections of the guide in the highlights list.
      
      ## How was this patch tested?
      
      Built docs locally and tested links.
      
      Author: Nick Pentreath <nickp@za.ibm.com>
      
      Closes #18043 from MLnick/SPARK-20506-2.2-migration-guide-2.
      be846db4
  4. May 19, 2017
  5. May 18, 2017
  6. May 17, 2017
    • Yanbo Liang's avatar
      [SPARK-20505][ML] Add docs and examples for ml.stat.Correlation and ml.stat.ChiSquareTest. · 697a5e55
      Yanbo Liang authored
      ## What changes were proposed in this pull request?
      Add docs and examples for ```ml.stat.Correlation``` and ```ml.stat.ChiSquareTest```.
      
      ## How was this patch tested?
      Generate docs and run examples manually, successfully.
      
      Author: Yanbo Liang <ybliang8@gmail.com>
      
      Closes #17994 from yanboliang/spark-20505.
      697a5e55
    • Andrew Ray's avatar
      [SPARK-20769][DOC] Incorrect documentation for using Jupyter notebook · 19954176
      Andrew Ray authored
      ## What changes were proposed in this pull request?
      
      SPARK-13973 incorrectly removed the required PYSPARK_DRIVER_PYTHON_OPTS=notebook from documentation to use pyspark with Jupyter notebook. This patch corrects the documentation error.
      
      ## How was this patch tested?
      
      Tested invocation locally with
      ```bash
      PYSPARK_DRIVER_PYTHON=jupyter PYSPARK_DRIVER_PYTHON_OPTS=notebook ./bin/pyspark
      ```
      
      Author: Andrew Ray <ray.andrew@gmail.com>
      
      Closes #18001 from aray/patch-1.
      19954176
  7. May 09, 2017
    • uncleGen's avatar
      [SPARK-20373][SQL][SS] Batch queries with 'Dataset/DataFrame.withWatermark()` does not execute · c0189abc
      uncleGen authored
      ## What changes were proposed in this pull request?
      
      Any Dataset/DataFrame batch query with the operation `withWatermark` does not execute because the batch planner does not have any rule to explicitly handle the EventTimeWatermark logical plan.
      The right solution is to simply remove the plan node, as the watermark should not affect any batch query in any way.
      
      Changes:
      - In this PR, we add a new rule `EliminateEventTimeWatermark` to check if we need to ignore the event time watermark. We will ignore watermark in any batch query.
      
      Depends upon:
      - [SPARK-20672](https://issues.apache.org/jira/browse/SPARK-20672). We can not add this rule into analyzer directly, because streaming query will be copied to `triggerLogicalPlan ` in every trigger, and the rule will be applied to `triggerLogicalPlan` mistakenly.
      
      Others:
      - A typo fix in example.
      
      ## How was this patch tested?
      
      add new unit test.
      
      Author: uncleGen <hustyugm@gmail.com>
      
      Closes #17896 from uncleGen/SPARK-20373.
      c0189abc
  8. May 08, 2017
    • jerryshao's avatar
      [SPARK-20605][CORE][YARN][MESOS] Deprecate not used AM and executor port configuration · 829cd7b8
      jerryshao authored
      ## What changes were proposed in this pull request?
      
      After SPARK-10997, client mode Netty RpcEnv doesn't require to start server, so port configurations are not used any more, here propose to remove these two configurations: "spark.executor.port" and "spark.am.port".
      
      ## How was this patch tested?
      
      Existing UTs.
      
      Author: jerryshao <sshao@hortonworks.com>
      
      Closes #17866 from jerryshao/SPARK-20605.
      829cd7b8
  9. May 07, 2017
    • Steve Loughran's avatar
      [SPARK-7481][BUILD] Add spark-hadoop-cloud module to pull in object store access. · 2cf83c47
      Steve Loughran authored
      ## What changes were proposed in this pull request?
      
      Add a new `spark-hadoop-cloud` module and maven profile to pull in object store support from `hadoop-openstack`, `hadoop-aws` and `hadoop-azure` (Hadoop 2.7+) JARs, along with their dependencies, fixing up the dependencies so that everything works, in particular Jackson.
      
      It restores `s3n://` access to S3, adds its `s3a://` replacement, OpenStack `swift://` and azure `wasb://`.
      
      There's a documentation page, `cloud_integration.md`, which covers the basic details of using Spark with object stores, referring the reader to the supplier's own documentation, with specific warnings on security and the possible mismatch between a store's behavior and that of a filesystem. In particular, users are advised be very cautious when trying to use an object store as the destination of data, and to consult the documentation of the storage supplier and the connector.
      
      (this is the successor to #12004; I can't re-open it)
      
      ## How was this patch tested?
      
      Downstream tests exist in [https://github.com/steveloughran/spark-cloud-examples/tree/master/cloud-examples](https://github.com/steveloughran/spark-cloud-examples/tree/master/cloud-examples)
      
      Those verify that the dependencies are sufficient to allow downstream applications to work with s3a, azure wasb and swift storage connectors, and perform basic IO & dataframe operations thereon. All seems well.
      
      Manually clean build & verify that assembly contains the relevant aws-* hadoop-* artifacts on Hadoop 2.6; azure on a hadoop-2.7 profile.
      
      SBT build: `build/sbt -Phadoop-cloud -Phadoop-2.7 package`
      maven build `mvn install -Phadoop-cloud -Phadoop-2.7`
      
      This PR *does not* update `dev/deps/spark-deps-hadoop-2.7` or `dev/deps/spark-deps-hadoop-2.6`, because unless the hadoop-cloud profile is enabled, no extra JARs show up in the dependency list. The dependency check in Jenkins isn't setting the property, so the new JARs aren't visible.
      
      Author: Steve Loughran <stevel@apache.org>
      Author: Steve Loughran <stevel@hortonworks.com>
      
      Closes #17834 from steveloughran/cloud/SPARK-7481-current.
      2cf83c47
  10. May 04, 2017
    • Felix Cheung's avatar
      [SPARK-20015][SPARKR][SS][DOC][EXAMPLE] Document R Structured Streaming... · b8302ccd
      Felix Cheung authored
      [SPARK-20015][SPARKR][SS][DOC][EXAMPLE] Document R Structured Streaming (experimental) in R vignettes and R & SS programming guide, R example
      
      ## What changes were proposed in this pull request?
      
      Add
      - R vignettes
      - R programming guide
      - SS programming guide
      - R example
      
      Also disable spark.als in vignettes for now since it's failing (SPARK-20402)
      
      ## How was this patch tested?
      
      manually
      
      Author: Felix Cheung <felixcheung_m@hotmail.com>
      
      Closes #17814 from felixcheung/rdocss.
      b8302ccd
  11. May 03, 2017
    • MechCoder's avatar
      [SPARK-6227][MLLIB][PYSPARK] Implement PySpark wrappers for SVD and PCA (v2) · db2fb84b
      MechCoder authored
      Add PCA and SVD to PySpark's wrappers for `RowMatrix` and `IndexedRowMatrix` (SVD only).
      
      Based on #7963, updated.
      
      ## How was this patch tested?
      
      New doc tests and unit tests. Ran all examples locally.
      
      Author: MechCoder <manojkumarsivaraj334@gmail.com>
      Author: Nick Pentreath <nickp@za.ibm.com>
      
      Closes #17621 from MLnick/SPARK-6227-pyspark-svd-pca.
      db2fb84b
  12. May 01, 2017
  13. Apr 30, 2017
    • 郭小龙 10207633's avatar
      [SPARK-20521][DOC][CORE] The default of 'spark.worker.cleanup.appDataTtl'... · 4d99b95a
      郭小龙 10207633 authored
      [SPARK-20521][DOC][CORE] The default of 'spark.worker.cleanup.appDataTtl' should be 604800 in spark-standalone.md
      
      ## What changes were proposed in this pull request?
      
      Currently, our project needs to be set to clean up the worker directory cleanup cycle is three days.
      When I follow http://spark.apache.org/docs/latest/spark-standalone.html, configure the 'spark.worker.cleanup.appDataTtl' parameter, I configured to 3 * 24 * 3600.
      When I start the spark service, the startup fails, and the worker log displays the error log as follows:
      
      2017-04-28 15:02:03,306 INFO Utils: Successfully started service 'sparkWorker' on port 48728.
      Exception in thread "main" java.lang.NumberFormatException: For input string: "3 * 24 * 3600"
      	at java.lang.NumberFormatException.forInputString(NumberFormatException.java:65)
      	at java.lang.Long.parseLong(Long.java:430)
      	at java.lang.Long.parseLong(Long.java:483)
      	at scala.collection.immutable.StringLike$class.toLong(StringLike.scala:276)
      	at scala.collection.immutable.StringOps.toLong(StringOps.scala:29)
      	at org.apache.spark.SparkConf$$anonfun$getLong$2.apply(SparkConf.scala:380)
      	at org.apache.spark.SparkConf$$anonfun$getLong$2.apply(SparkConf.scala:380)
      	at scala.Option.map(Option.scala:146)
      	at org.apache.spark.SparkConf.getLong(SparkConf.scala:380)
      	at org.apache.spark.deploy.worker.Worker.<init>(Worker.scala:100)
      	at org.apache.spark.deploy.worker.Worker$.startRpcEnvAndEndpoint(Worker.scala:730)
      	at org.apache.spark.deploy.worker.Worker$.main(Worker.scala:709)
      	at org.apache.spark.deploy.worker.Worker.main(Worker.scala)
      
      **Because we put 7 * 24 * 3600 as a string, forced to convert to the dragon type,  will lead to problems in the program.**
      
      **So I think the default value of the current configuration should be a specific long value, rather than 7 * 24 * 3600,should be 604800. Because it would mislead users for similar configurations, resulting in spark start failure.**
      
      ## How was this patch tested?
      manual tests
      
      Please review http://spark.apache.org/contributing.html before opening a pull request.
      
      Author: 郭小龙 10207633 <guo.xiaolong1@zte.com.cn>
      Author: guoxiaolong <guo.xiaolong1@zte.com.cn>
      Author: guoxiaolongzte <guo.xiaolong1@zte.com.cn>
      
      Closes #17798 from guoxiaolongzte/SPARK-20521.
      4d99b95a
  14. Apr 29, 2017
    • Yuhao Yang's avatar
      [SPARK-19791][ML] Add doc and example for fpgrowth · add9d1bb
      Yuhao Yang authored
      ## What changes were proposed in this pull request?
      
      Add a new section for fpm
      Add Example for FPGrowth in scala and Java
      
      updated: Rewrite transform to be more compact.
      
      ## How was this patch tested?
      
      local doc generation.
      
      Author: Yuhao Yang <yuhao.yang@intel.com>
      
      Closes #17130 from hhbyyh/fpmdoc.
      add9d1bb
    • wangmiao1981's avatar
      [SPARK-20477][SPARKR][DOC] Document R bisecting k-means in R programming guide · b28c3bc2
      wangmiao1981 authored
      ## What changes were proposed in this pull request?
      
      Add hyper link in the SparkR programming guide.
      
      ## How was this patch tested?
      
      Build doc and manually check the doc link.
      
      Author: wangmiao1981 <wm624@hotmail.com>
      
      Closes #17805 from wangmiao1981/doc.
      b28c3bc2
  15. Apr 28, 2017
    • wangmiao1981's avatar
      [SPARKR][DOC] Document LinearSVC in R programming guide · 7fe82497
      wangmiao1981 authored
      ## What changes were proposed in this pull request?
      
      add link to svmLinear in the SparkR programming document.
      
      ## How was this patch tested?
      
      Build doc manually and click the link to the document. It looks good.
      
      Author: wangmiao1981 <wm624@hotmail.com>
      
      Closes #17797 from wangmiao1981/doc.
      7fe82497
  16. Apr 27, 2017
  17. Apr 26, 2017
    • Mark Grover's avatar
      [SPARK-20435][CORE] More thorough redaction of sensitive information · 66636ef0
      Mark Grover authored
      This change does a more thorough redaction of sensitive information from logs and UI
      Add unit tests that ensure that no regressions happen that leak sensitive information to the logs.
      
      The motivation for this change was appearance of password like so in `SparkListenerEnvironmentUpdate` in event logs under some JVM configurations:
      `"sun.java.command":"org.apache.spark.deploy.SparkSubmit ... --conf spark.executorEnv.HADOOP_CREDSTORE_PASSWORD=secret_password ..."
      `
      Previously redaction logic was only checking if the key matched the secret regex pattern, it'd redact it's value. That worked for most cases. However, in the above case, the key (sun.java.command) doesn't tell much, so the value needs to be searched. This PR expands the check to check for values as well.
      
      ## How was this patch tested?
      
      New unit tests added that ensure that no sensitive information is present in the event logs or the yarn logs. Old unit test in UtilsSuite was modified because the test was asserting that a non-sensitive property's value won't be redacted. However, the non-sensitive value had the literal "secret" in it which was causing it to redact. Simply updating the non-sensitive property's value to another arbitrary value (that didn't have "secret" in it) fixed it.
      
      Author: Mark Grover <mark@apache.org>
      
      Closes #17725 from markgrover/spark-20435.
      66636ef0
    • anabranch's avatar
      [SPARK-20400][DOCS] Remove References to 3rd Party Vendor Tools · 7a365257
      anabranch authored
      ## What changes were proposed in this pull request?
      
      Simple documentation change to remove explicit vendor references.
      
      ## How was this patch tested?
      
      NA
      
      Please review http://spark.apache.org/contributing.html before opening a pull request.
      
      Author: anabranch <bill@databricks.com>
      
      Closes #17695 from anabranch/remove-vendor.
      7a365257
    • zero323's avatar
      [SPARK-20437][R] R wrappers for rollup and cube · df58a95a
      zero323 authored
      ## What changes were proposed in this pull request?
      
      - Add `rollup` and `cube` methods and corresponding generics.
      - Add short description to the vignette.
      
      ## How was this patch tested?
      
      - Existing unit tests.
      - Additional unit tests covering new features.
      - `check-cran.sh`.
      
      Author: zero323 <zero323@users.noreply.github.com>
      
      Closes #17728 from zero323/SPARK-20437.
      df58a95a
  18. Apr 25, 2017
    • ding's avatar
      [SPARK-5484][GRAPHX] Periodically do checkpoint in Pregel · 0a7f5f27
      ding authored
      ## What changes were proposed in this pull request?
      
      Pregel-based iterative algorithms with more than ~50 iterations begin to slow down and eventually fail with a StackOverflowError due to Spark's lack of support for long lineage chains.
      
      This PR causes Pregel to checkpoint the graph periodically if the checkpoint directory is set.
      This PR moves PeriodicGraphCheckpointer.scala from mllib to graphx, moves PeriodicRDDCheckpointer.scala, PeriodicCheckpointer.scala from mllib to core
      ## How was this patch tested?
      
      unit tests, manual tests
      (Please explain how this patch was tested. E.g. unit tests, integration tests, manual tests)
      
      (If this patch involves UI changes, please attach a screenshot; otherwise, remove this)
      
      Author: ding <ding@localhost.localdomain>
      Author: dding3 <ding.ding@intel.com>
      Author: Michael Allman <michael@videoamp.com>
      
      Closes #15125 from dding3/cp2_pregel.
      0a7f5f27
    • Armin Braun's avatar
      [SPARK-20455][DOCS] Fix Broken Docker IT Docs · c8f12195
      Armin Braun authored
      ## What changes were proposed in this pull request?
      
      Just added the Maven `test`goal.
      
      ## How was this patch tested?
      
      No test needed, just a trivial documentation fix.
      
      Author: Armin Braun <me@obrown.io>
      
      Closes #17756 from original-brownbear/SPARK-20455.
      c8f12195
  19. Apr 24, 2017
  20. Apr 21, 2017
    • 郭小龙 10207633's avatar
      [SPARK-20401][DOC] In the spark official configuration document, the... · ad290402
      郭小龙 10207633 authored
      [SPARK-20401][DOC] In the spark official configuration document, the 'spark.driver.supervise' configuration parameter specification and default values are necessary.
      
      ## What changes were proposed in this pull request?
      Use the REST interface submits the spark job.
      e.g.
      curl -X  POST http://10.43.183.120:6066/v1/submissions/create --header "Content-Type:application/json;charset=UTF-8" --data'{
          "action": "CreateSubmissionRequest",
          "appArgs": [
              "myAppArgument"
          ],
          "appResource": "/home/mr/gxl/test.jar",
          "clientSparkVersion": "2.2.0",
          "environmentVariables": {
              "SPARK_ENV_LOADED": "1"
          },
          "mainClass": "cn.zte.HdfsTest",
          "sparkProperties": {
              "spark.jars": "/home/mr/gxl/test.jar",
              **"spark.driver.supervise": "true",**
              "spark.app.name": "HdfsTest",
              "spark.eventLog.enabled": "false",
              "spark.submit.deployMode": "cluster",
              "spark.master": "spark://10.43.183.120:6066"
          }
      }'
      
      **I hope that make sure that the driver is automatically restarted if it fails with non-zero exit code.
      But I can not find the 'spark.driver.supervise' configuration parameter specification and default values from the spark official document.**
      ## How was this patch tested?
      
      manual tests
      
      Please review http://spark.apache.org/contributing.html before opening a pull request.
      
      Author: 郭小龙 10207633 <guo.xiaolong1@zte.com.cn>
      Author: guoxiaolong <guo.xiaolong1@zte.com.cn>
      Author: guoxiaolongzte <guo.xiaolong1@zte.com.cn>
      
      Closes #17696 from guoxiaolongzte/SPARK-20401.
      ad290402
    • Hervé's avatar
      Small rewording about history server use case · 34767997
      Hervé authored
      Hello
      PR #10991 removed the built-in history view from Spark Standalone, so the history server is no longer useful to Yarn or Mesos only.
      
      Author: Hervé <dud225@users.noreply.github.com>
      
      Closes #17709 from dud225/patch-1.
      34767997
  21. Apr 19, 2017
    • ymahajan's avatar
      Fixed typos in docs · bdc60569
      ymahajan authored
      ## What changes were proposed in this pull request?
      
      Typos at a couple of place in the docs.
      
      ## How was this patch tested?
      
      build including docs
      
      Please review http://spark.apache.org/contributing.html before opening a pull request.
      
      Author: ymahajan <ymahajan@snappydata.io>
      
      Closes #17690 from ymahajan/master.
      bdc60569
    • cody koeninger's avatar
      [SPARK-20036][DOC] Note incompatible dependencies on org.apache.kafka artifacts · 71a8e9df
      cody koeninger authored
      ## What changes were proposed in this pull request?
      
      Note that you shouldn't manually add dependencies on org.apache.kafka artifacts
      
      ## How was this patch tested?
      
      Doc only change, did jekyll build and looked at the page.
      
      Author: cody koeninger <cody@koeninger.org>
      
      Closes #17675 from koeninger/SPARK-20036.
      71a8e9df
  22. Apr 16, 2017
    • Ji Yan's avatar
      [SPARK-19740][MESOS] Add support in Spark to pass arbitrary parameters into... · a888fed3
      Ji Yan authored
      [SPARK-19740][MESOS] Add support in Spark to pass arbitrary parameters into docker when running on mesos with docker containerizer
      
      ## What changes were proposed in this pull request?
      
      Allow passing in arbitrary parameters into docker when launching spark executors on mesos with docker containerizer tnachen
      
      ## How was this patch tested?
      
      Manually built and tested with passed in parameter
      
      Author: Ji Yan <jiyan@Jis-MacBook-Air.local>
      
      Closes #17109 from yanji84/ji/allow_set_docker_user.
      a888fed3
  23. Apr 12, 2017
    • hyukjinkwon's avatar
      [MINOR][DOCS] JSON APIs related documentation fixes · bca4259f
      hyukjinkwon authored
      ## What changes were proposed in this pull request?
      
      This PR proposes corrections related to JSON APIs as below:
      
      - Rendering links in Python documentation
      - Replacing `RDD` to `Dataset` in programing guide
      - Adding missing description about JSON Lines consistently in `DataFrameReader.json` in Python API
      - De-duplicating little bit of `DataFrameReader.json` in Scala/Java API
      
      ## How was this patch tested?
      
      Manually build the documentation via `jekyll build`. Corresponding snapstops will be left on the codes.
      
      Note that currently there are Javadoc8 breaks in several places. These are proposed to be handled in https://github.com/apache/spark/pull/17477. So, this PR does not fix those.
      
      Author: hyukjinkwon <gurwls223@gmail.com>
      
      Closes #17602 from HyukjinKwon/minor-json-documentation.
      bca4259f
    • Lee Dongjin's avatar
      [MINOR][DOCS] Fix spacings in Structured Streaming Programming Guide · b9384382
      Lee Dongjin authored
      ## What changes were proposed in this pull request?
      
      1. Omitted space between the sentences: `... on static data.The Spark SQL engine will ...` -> `... on static data. The Spark SQL engine will ...`
      2. Omitted colon in Output Model section.
      
      ## How was this patch tested?
      
      None.
      
      Author: Lee Dongjin <dongjin@apache.org>
      
      Closes #17564 from dongjinleekr/feature/fix-programming-guide.
      b9384382
  24. Apr 11, 2017
  25. Apr 07, 2017
    • 郭小龙 10207633's avatar
      [SPARK-20218][DOC][APP-ID] applications//stages' in REST API,add description. · 9e0893b5
      郭小龙 10207633 authored
      ## What changes were proposed in this pull request?
      
      1. '/applications/[app-id]/stages' in rest api.status should add description '?status=[active|complete|pending|failed] list only stages in the state.'
      
      Now the lack of this description, resulting in the use of this api do not know the use of the status through the brush stage list.
      
      2.'/applications/[app-id]/stages/[stage-id]' in REST API,remove redundant description ‘?status=[active|complete|pending|failed] list only stages in the state.’.
      Because only one stage is determined based on stage-id.
      
      code:
        GET
        def stageList(QueryParam("status") statuses: JList[StageStatus]): Seq[StageData] = {
          val listener = ui.jobProgressListener
          val stageAndStatus = AllStagesResource.stagesAndStatus(ui)
          val adjStatuses = {
            if (statuses.isEmpty()) {
              Arrays.asList(StageStatus.values(): _*)
            } else {
              statuses
            }
          };
      
      ## How was this patch tested?
      
      manual tests
      
      Please review http://spark.apache.org/contributing.html before opening a pull request.
      
      Author: 郭小龙 10207633 <guo.xiaolong1@zte.com.cn>
      
      Closes #17534 from guoxiaolongzte/SPARK-20218.
      9e0893b5
  26. Apr 06, 2017
    • Kalvin Chau's avatar
      [SPARK-20085][MESOS] Configurable mesos labels for executors · c8fc1f3b
      Kalvin Chau authored
      ## What changes were proposed in this pull request?
      
      Add spark.mesos.task.labels configuration option to add mesos key:value labels to the executor.
      
       "k1:v1,k2:v2" as the format, colons separating key-value and commas to list out more than one.
      
      Discussion of labels with mgummelt at #17404
      
      ## How was this patch tested?
      
      Added unit tests to verify labels were added correctly, with incorrect labels being ignored and added a test to test the name of the executor.
      
      Tested with: `./build/sbt -Pmesos mesos/test`
      
      Please review http://spark.apache.org/contributing.html before opening a pull request.
      
      Author: Kalvin Chau <kalvin.chau@viasat.com>
      
      Closes #17413 from kalvinnchau/mesos-labels.
      c8fc1f3b
  27. Apr 05, 2017
Loading