Skip to content
Snippets Groups Projects
  1. Aug 16, 2016
    • Sean Zhong's avatar
      [SPARK-17034][SQL] adds expression UnresolvedOrdinal to represent the ordinals... · 7b65030e
      Sean Zhong authored
      [SPARK-17034][SQL] adds expression UnresolvedOrdinal to represent the ordinals in GROUP BY or ORDER BY
      
      ## What changes were proposed in this pull request?
      
      This PR adds expression `UnresolvedOrdinal` to represent the ordinal in GROUP BY or ORDER BY, and fixes the rules when resolving ordinals.
      
      Ordinals in GROUP BY or ORDER BY like `1` in `order by 1` or `group by 1` should be considered as unresolved before analysis. But in current code, it uses `Literal` expression to store the ordinal. This is inappropriate as `Literal` itself is a resolved expression, it gives the user a wrong message that the ordinals has already been resolved.
      
      ### Before this change
      
      Ordinal is stored as `Literal` expression
      
      ```
      scala> sc.setLogLevel("TRACE")
      scala> sql("select a from t group by 1 order by 1")
      ...
      'Sort [1 ASC], true
       +- 'Aggregate [1], ['a]
           +- 'UnresolvedRelation `t
      ```
      
      For query:
      
      ```
      scala> Seq(1).toDF("a").createOrReplaceTempView("t")
      scala> sql("select count(a), a from t group by 2 having a > 0").show
      ```
      
      During analysis, the intermediate plan before applying rule `ResolveAggregateFunctions` is:
      
      ```
      'Filter ('a > 0)
         +- Aggregate [2], [count(1) AS count(1)#83L, a#81]
              +- LocalRelation [value#7 AS a#9]
      ```
      
      Before this PR, rule `ResolveAggregateFunctions` believes all expressions of `Aggregate` have already been resolved, and tries to resolve the expressions in `Filter` directly. But this is wrong, as ordinal `2` in Aggregate is not really resolved!
      
      ### After this change
      
      Ordinals are stored as `UnresolvedOrdinal`.
      
      ```
      scala> sc.setLogLevel("TRACE")
      scala> sql("select a from t group by 1 order by 1")
      ...
      'Sort [unresolvedordinal(1) ASC], true
       +- 'Aggregate [unresolvedordinal(1)], ['a]
            +- 'UnresolvedRelation `t`
      ```
      
      ## How was this patch tested?
      
      Unit tests.
      
      Author: Sean Zhong <seanzhong@databricks.com>
      
      Closes #14616 from clockfly/spark-16955.
      7b65030e
  2. Aug 15, 2016
    • Wenchen Fan's avatar
      [SPARK-16916][SQL] serde/storage properties should not have limitations · 7de30d6e
      Wenchen Fan authored
      ## What changes were proposed in this pull request?
      
      `CatalogStorageFormat.properties` can be used in 2 ways:
      
      1. for hive tables, it stores the serde properties.
      2. for data source tables, it stores the data source options, e.g. `path`, `skipHiveMetadata`, etc.
      
      however, both of them have nothing to do with data source properties, e.g. `spark.sql.sources.provider`, so they should not have limitations about data source properties.
      
      ## How was this patch tested?
      
      existing tests
      
      Author: Wenchen Fan <wenchen@databricks.com>
      
      Closes #14506 from cloud-fan/table-prop.
      7de30d6e
    • Shixiong Zhu's avatar
      [SPARK-17065][SQL] Improve the error message when encountering an incompatible DataSourceRegister · 268b71d0
      Shixiong Zhu authored
      ## What changes were proposed in this pull request?
      
      Add an instruction to ask the user to remove or upgrade the incompatible DataSourceRegister in the error message.
      
      ## How was this patch tested?
      
      Test command:
      ```
      build/sbt -Dscala-2.10 package
      SPARK_SCALA_VERSION=2.10 bin/spark-shell --packages ai.h2o:sparkling-water-core_2.10:1.6.5
      
      scala> Seq(1).toDS().write.format("parquet").save("foo")
      ```
      
      Before:
      ```
      java.util.ServiceConfigurationError: org.apache.spark.sql.sources.DataSourceRegister: Provider org.apache.spark.h2o.DefaultSource could not be instantiated
      	at java.util.ServiceLoader.fail(ServiceLoader.java:232)
      	at java.util.ServiceLoader.access$100(ServiceLoader.java:185)
      	at java.util.ServiceLoader$LazyIterator.nextService(ServiceLoader.java:384)
      	at java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:404)
      	at java.util.ServiceLoader$1.next(ServiceLoader.java:480)
      ...
      Caused by: java.lang.NoClassDefFoundError: org/apache/spark/Logging
      	at java.lang.ClassLoader.defineClass1(Native Method)
      	at java.lang.ClassLoader.defineClass(ClassLoader.java:760)
      	at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
      	at java.net.URLClassLoader.defineClass(URLClassLoader.java:467)
      	at java.net.URLClassLoader.access$100(URLClassLoader.java:73)
      	at java.net.URLClassLoader$1.run(URLClassLoader.java:368)
      	at java.net.URLClassLoader$1.run(URLClassLoader.java:362)
      	at java.security.AccessController.doPrivileged(Native Method)
      ...
      ```
      
      After:
      
      ```
      java.lang.ClassNotFoundException: Detected an incompatible DataSourceRegister. Please remove the incompatible library from classpath or upgrade it. Error: org.apache.spark.sql.sources.DataSourceRegister: Provider org.apache.spark.h2o.DefaultSource could not be instantiated
      	at org.apache.spark.sql.execution.datasources.DataSource.lookupDataSource(DataSource.scala:178)
      	at org.apache.spark.sql.execution.datasources.DataSource.providingClass$lzycompute(DataSource.scala:79)
      	at org.apache.spark.sql.execution.datasources.DataSource.providingClass(DataSource.scala:79)
      	at org.apache.spark.sql.execution.datasources.DataSource.write(DataSource.scala:441)
      	at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:213)
      	at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:196)
      ...
      ```
      
      Author: Shixiong Zhu <shixiong@databricks.com>
      
      Closes #14651 from zsxwing/SPARK-17065.
      268b71d0
    • Davies Liu's avatar
      [SPARK-16700][PYSPARK][SQL] create DataFrame from dict/Row with schema · fffb0c0d
      Davies Liu authored
      ## What changes were proposed in this pull request?
      
      In 2.0, we verify the data type against schema for every row for safety, but with performance cost, this PR make it optional.
      
      When we verify the data type for StructType, it does not support all the types we support in infer schema (for example, dict), this PR fix that to make them consistent.
      
      For Row object which is created using named arguments, the order of fields are sorted by name, they may be not different than the order in provided schema, this PR fix that by ignore the order of fields in this case.
      
      ## How was this patch tested?
      
      Created regression tests for them.
      
      Author: Davies Liu <davies@databricks.com>
      
      Closes #14469 from davies/py_dict.
      fffb0c0d
    • Marcelo Vanzin's avatar
      [SPARK-16671][CORE][SQL] Consolidate code to do variable substitution. · 5da6c4b2
      Marcelo Vanzin authored
      Both core and sql have slightly different code that does variable substitution
      of config values. This change refactors that code and encapsulates the logic
      of reading config values and expading variables in a new helper class, which
      can be configured so that both core and sql can use it without losing existing
      functionality, and allows for easier testing and makes it easier to add more
      features in the future.
      
      Tested with existing and new unit tests, and by running spark-shell with
      some configs referencing variables and making sure it behaved as expected.
      
      Author: Marcelo Vanzin <vanzin@cloudera.com>
      
      Closes #14468 from vanzin/SPARK-16671.
      5da6c4b2
    • Junyang Qian's avatar
      [SPARK-16508][SPARKR] Split docs for arrange and orderBy methods · 564fe614
      Junyang Qian authored
      ## What changes were proposed in this pull request?
      
      This PR splits arrange and orderBy methods according to their functionality (the former for sorting sparkDataFrame and the latter for windowSpec).
      
      ## How was this patch tested?
      
      ![screen shot 2016-08-06 at 6 39 19 pm](https://cloud.githubusercontent.com/assets/15318264/17459969/51eade28-5c05-11e6-8ca1-8d8a8e344bab.png)
      ![screen shot 2016-08-06 at 6 39 29 pm](https://cloud.githubusercontent.com/assets/15318264/17459966/51e3c246-5c05-11e6-8d35-3e905ca48676.png)
      ![screen shot 2016-08-06 at 6 40 02 pm](https://cloud.githubusercontent.com/assets/15318264/17459967/51e650ec-5c05-11e6-8698-0f037f5199ff.png)
      
      Author: Junyang Qian <junyangq@databricks.com>
      
      Closes #14522 from junyangq/SPARK-16508-0.
      564fe614
    • WeichenXu's avatar
      [SPARK-16934][ML][MLLIB] Update LogisticCostAggregator serialization code to... · 3d8bfe7a
      WeichenXu authored
      [SPARK-16934][ML][MLLIB] Update LogisticCostAggregator serialization code to make it consistent with LinearRegression
      
      ## What changes were proposed in this pull request?
      
      Update LogisticCostAggregator serialization code to make it consistent with #14109
      
      ## How was this patch tested?
      MLlib 2.0:
      ![image](https://cloud.githubusercontent.com/assets/19235986/17649601/5e2a79ac-61ee-11e6-833c-3bd8b5250470.png)
      
      After this PR:
      ![image](https://cloud.githubusercontent.com/assets/19235986/17649599/52b002ae-61ee-11e6-9402-9feb3439880f.png)
      
      Author: WeichenXu <WeichenXu123@outlook.com>
      
      Closes #14520 from WeichenXu123/improve_logistic_regression_costfun.
      3d8bfe7a
    • Yanbo Liang's avatar
      [TRIVIAL][ML] Fix LogisticRegression typo in error message. · ddf0d1e3
      Yanbo Liang authored
      ## What changes were proposed in this pull request?
      Fix ```LogisticRegression``` typo in error message.
      
      ## How was this patch tested?
      Docs change, no new tests.
      
      Author: Yanbo Liang <ybliang8@gmail.com>
      
      Closes #14633 from yanboliang/lr-typo.
      ddf0d1e3
    • Stavros Kontopoulos's avatar
      [SPARK-11714][MESOS] Make Spark on Mesos honor port restrictions on coarse grain mode · 1a028bde
      Stavros Kontopoulos authored
      - Make mesos coarse grained scheduler accept port offers and pre-assign ports
      
      Previous attempt was for fine grained: https://github.com/apache/spark/pull/10808
      
      Author: Stavros Kontopoulos <stavros.kontopoulos@lightbend.com>
      Author: Stavros Kontopoulos <stavros.kontopoulos@typesafe.com>
      
      Closes #11157 from skonto/honour_ports_coarse.
      1a028bde
  3. Aug 14, 2016
    • Zhenglai Zhang's avatar
      [WIP][MINOR][TYPO] Fix several trivival typos · 2a3d286f
      Zhenglai Zhang authored
      ## What changes were proposed in this pull request?
      
      * Fixed one typo `"overriden"` as `"overridden"`, also make sure no other same typo.
      * Fixed one typo `"lowcase"` as `"lowercase"`, also make sure no other same typo.
      
      ## How was this patch tested?
      
      Since the change is very tiny, so I just make sure compilation is successful.
      I am new to the spark community,  please feel free to let me do other necessary steps.
      
      Thanks in advance!
      
      ----
      Updated: Found another typo `lowcase` later and fixed then in the same patch
      
      Author: Zhenglai Zhang <zhenglaizhang@hotmail.com>
      
      Closes #14622 from zhenglaizhang/fixtypo.
      2a3d286f
    • zero323's avatar
      [SPARK-17027][ML] Avoid integer overflow in PolynomialExpansion.getPolySize · 0ebf7c1b
      zero323 authored
      ## What changes were proposed in this pull request?
      
      Replaces custom choose function with o.a.commons.math3.CombinatoricsUtils.binomialCoefficient
      
      ## How was this patch tested?
      
      Spark unit tests
      
      Author: zero323 <zero323@users.noreply.github.com>
      
      Closes #14614 from zero323/SPARK-17027.
      0ebf7c1b
  4. Aug 13, 2016
    • Sean Owen's avatar
      [SPARK-16966][SQL][CORE] App Name is a randomUUID even when "spark.app.name" exists · cdaa562c
      Sean Owen authored
      ## What changes were proposed in this pull request?
      
      Don't override app name specified in `SparkConf` with a random app name. Only set it if the conf has no app name even after options have been applied.
      
      See also https://github.com/apache/spark/pull/14602
      This is similar to Sherry302 's original proposal in https://github.com/apache/spark/pull/14556
      
      ## How was this patch tested?
      
      Jenkins test, with new case reproducing the bug
      
      Author: Sean Owen <sowen@cloudera.com>
      
      Closes #14630 from srowen/SPARK-16966.2.
      cdaa562c
    • Luciano Resende's avatar
      [SPARK-17023][BUILD] Upgrade to Kafka 0.10.0.1 release · 67f025d9
      Luciano Resende authored
      ## What changes were proposed in this pull request?
      Update Kafka streaming connector to use Kafka 0.10.0.1 release
      
      ## How was this patch tested?
      Tested via Spark unit and integration tests
      
      Author: Luciano Resende <lresende@apache.org>
      
      Closes #14606 from lresende/kafka-upgrade.
      67f025d9
    • GraceH's avatar
      [SPARK-16968] Add additional options in jdbc when creating a new table · 8c8acdec
      GraceH authored
      ## What changes were proposed in this pull request?
      
      In the PR, we just allow the user to add additional options when create a new table in JDBC writer.
      The options can be table_options or partition_options.
      E.g., "CREATE TABLE t (name string) ENGINE=InnoDB DEFAULT CHARSET=utf8"
      
      Here is the usage example:
      ```
      df.write.option("createTableOptions", "ENGINE=InnoDB DEFAULT CHARSET=utf8").jdbc(...)
      ```
      ## How was this patch tested?
      
      (Please explain how this patch was tested. E.g. unit tests, integration tests, manual tests)
      will apply test result soon.
      
      Author: GraceH <93113783@qq.com>
      
      Closes #14559 from GraceH/jdbc_options.
      8c8acdec
    • Xin Ren's avatar
      [MINOR][CORE] fix warnings on depreciated methods in... · 7f7133bd
      Xin Ren authored
      [MINOR][CORE] fix warnings on depreciated methods in MesosClusterSchedulerSuite and DiskBlockObjectWriterSuite
      
      ## What changes were proposed in this pull request?
      
      Fixed warnings below after scanning through warnings during build:
      
      ```
      [warn] /home/jenkins/workspace/SparkPullRequestBuilder/core/src/test/scala/org/apache/spark/scheduler/cluster/mesos/MesosClusterSchedulerSuite.scala:34: imported `Utils' is permanently hidden by definition of object Utils in package mesos
      [warn] import org.apache.spark.scheduler.cluster.mesos.Utils
      [warn]                                                 ^
      ```
      
      and
      ```
      [warn] /home/jenkins/workspace/SparkPullRequestBuilder/core/src/test/scala/org/apache/spark/storage/DiskBlockObjectWriterSuite.scala:113: method shuffleBytesWritten in class ShuffleWriteMetrics is deprecated: use bytesWritten instead
      [warn]     assert(writeMetrics.shuffleBytesWritten === file.length())
      [warn]                         ^
      [warn] /home/jenkins/workspace/SparkPullRequestBuilder/core/src/test/scala/org/apache/spark/storage/DiskBlockObjectWriterSuite.scala:119: method shuffleBytesWritten in class ShuffleWriteMetrics is deprecated: use bytesWritten instead
      [warn]     assert(writeMetrics.shuffleBytesWritten === file.length())
      [warn]                         ^
      [warn] /home/jenkins/workspace/SparkPullRequestBuilder/core/src/test/scala/org/apache/spark/storage/DiskBlockObjectWriterSuite.scala:131: method shuffleBytesWritten in class ShuffleWriteMetrics is deprecated: use bytesWritten instead
      [warn]     assert(writeMetrics.shuffleBytesWritten === file.length())
      [warn]                         ^
      [warn] /home/jenkins/workspace/SparkPullRequestBuilder/core/src/test/scala/org/apache/spark/storage/DiskBlockObjectWriterSuite.scala:135: method shuffleBytesWritten in class ShuffleWriteMetrics is deprecated: use bytesWritten instead
      [warn]     assert(writeMetrics.shuffleBytesWritten === file.length())
      [warn]                         ^
      ```
      
      ## How was this patch tested?
      
      Tested manually on local laptop.
      
      Author: Xin Ren <iamshrek@126.com>
      
      Closes #14609 from keypointt/suiteWarnings.
      7f7133bd
    • Jagadeesan's avatar
      [SPARK-12370][DOCUMENTATION] Documentation should link to examples … · e46cb78b
      Jagadeesan authored
      ## What changes were proposed in this pull request?
      
      When documentation is built is should reference examples from the same build. There are times when the docs have links that point to files in the GitHub head which may not be valid on the current release. Changed that in URLs to make them point to the right tag in git using ```SPARK_VERSION_SHORT```
      
      …from its own release version] [Streaming programming guide]
      
      Author: Jagadeesan <as2@us.ibm.com>
      
      Closes #14596 from jagadeesanas2/SPARK-12370.
      e46cb78b
  5. Aug 12, 2016
    • WeichenXu's avatar
      [DOC] add config option spark.ui.enabled into document · 91f2735a
      WeichenXu authored
      ## What changes were proposed in this pull request?
      
      The configuration doc lost the config option `spark.ui.enabled` (default value is `true`)
      I think this option is important because many cases we would like to turn it off.
      so I add it.
      
      ## How was this patch tested?
      
      N/A
      
      Author: WeichenXu <WeichenXu123@outlook.com>
      
      Closes #14604 from WeichenXu123/add_doc_param_spark_ui_enabled.
      91f2735a
    • Dongjoon Hyun's avatar
      [SPARK-16771][SQL] WITH clause should not fall into infinite loop. · 2a105134
      Dongjoon Hyun authored
      ## What changes were proposed in this pull request?
      
      This PR changes the CTE resolving rule to use only **forward-declared** tables in order to prevent infinite loops. More specifically, new logic is like the following.
      
      * Resolve CTEs in `WITH` clauses first before replacing the main SQL body.
      * When resolving CTEs, only forward-declared CTEs or base tables are referenced.
        - Self-referencing is not allowed any more.
        - Cross-referencing is not allowed any more.
      
      **Reported Error Scenarios**
      ```scala
      scala> sql("WITH t AS (SELECT 1 FROM t) SELECT * FROM t")
      java.lang.StackOverflowError
      ...
      scala> sql("WITH t1 AS (SELECT * FROM t2), t2 AS (SELECT 2 FROM t1) SELECT * FROM t1, t2")
      java.lang.StackOverflowError
      ...
      ```
      Note that `t`, `t1`, and `t2` are not declared in database. Spark falls into infinite loops before resolving table names.
      
      ## How was this patch tested?
      
      Pass the Jenkins tests with new two testcases.
      
      Author: Dongjoon Hyun <dongjoon@apache.org>
      
      Closes #14397 from dongjoon-hyun/SPARK-16771-TREENODE.
      2a105134
    • Yanbo Liang's avatar
      [SPARK-17033][ML][MLLIB] GaussianMixture should use treeAggregate to improve performance · bbae20ad
      Yanbo Liang authored
      ## What changes were proposed in this pull request?
      ```GaussianMixture``` should use ```treeAggregate``` rather than ```aggregate``` to improve performance and scalability. In my test of dataset with 200 features and 1M instance, I found there is 20% increased performance.
      BTW, we should destroy broadcast variable ```compute``` at the end of each iteration.
      
      ## How was this patch tested?
      Existing tests.
      
      Author: Yanbo Liang <ybliang8@gmail.com>
      
      Closes #14621 from yanboliang/spark-17033.
      bbae20ad
    • gatorsmile's avatar
      [SPARK-16598][SQL][TEST] Added a test case for verifying the table identifier parsing · 79e2caa1
      gatorsmile authored
      #### What changes were proposed in this pull request?
      So far, the test cases of `TableIdentifierParserSuite` do not cover the quoted cases. We should add one for avoiding regression.
      
      #### How was this patch tested?
      N/A
      
      Author: gatorsmile <gatorsmile@gmail.com>
      
      Closes #14244 from gatorsmile/quotedIdentifiers.
      79e2caa1
    • hyukjinkwon's avatar
      [MINOR][DOC] Fix style in examples across documentation · f4482225
      hyukjinkwon authored
      ## What changes were proposed in this pull request?
      
      This PR fixes the documentation as below:
      
        -  Python has 4 spaces and Java and Scala has 2 spaces (See https://cwiki.apache.org/confluence/display/SPARK/Spark+Code+Style+Guide).
      
        - Avoid excessive parentheses and curly braces for anonymous functions. (See https://github.com/databricks/scala-style-guide#anonymous)
      
      ## How was this patch tested?
      
      N/A
      
      Author: hyukjinkwon <gurwls223@gmail.com>
      
      Closes #14593 from HyukjinKwon/minor-documentation.
      f4482225
    • hongshen's avatar
      [SPARK-16985] Change dataFormat from yyyyMMddHHmm to yyyyMMddHHmmss · 993923c8
      hongshen authored
      ## What changes were proposed in this pull request?
      
      In our cluster, sometimes the sql output maybe overrided. When I submit some sql, all insert into the same table, and the sql will cost less one minute, here is the detail,
      1 sql1, 11:03 insert into table.
      2 sql2, 11:04:11 insert into table.
      3 sql3, 11:04:48 insert into table.
      4 sql4, 11:05 insert into table.
      5 sql5, 11:06 insert into table.
      The sql3's output file will override the sql2's output file. here is the log:
      ```
      16/05/04 11:04:11 INFO hive.SparkHiveHadoopWriter: XXfinalPath=hdfs://tl-sng-gdt-nn-tdw.tencent-distribute.com:54310/tmp/assorz/tdw-tdwadmin/20160504/04559505496526517_-1_1204544348/10000/_tmp.p_20160428/attempt_201605041104_0001_m_000000_1
      
      16/05/04 11:04:48 INFO hive.SparkHiveHadoopWriter: XXfinalPath=hdfs://tl-sng-gdt-nn-tdw.tencent-distribute.com:54310/tmp/assorz/tdw-tdwadmin/20160504/04559505496526517_-1_212180468/10000/_tmp.p_20160428/attempt_201605041104_0001_m_000000_1
      
      ```
      
      The reason is the output file use SimpleDateFormat("yyyyMMddHHmm"), if two sql insert into the same table in the same minute, the output will be overrite. I think we should change dateFormat to "yyyyMMddHHmmss", in our cluster, we can't finished a sql in one second.
      
      ## How was this patch tested?
      
      (Please explain how this patch was tested. E.g. unit tests, integration tests, manual tests)
      
      (If this patch involves UI changes, please attach a screenshot; otherwise, remove this)
      
      Author: hongshen <shenh062326@126.com>
      
      Closes #14574 from shenh062326/SPARK-16985.
      993923c8
    • petermaxlee's avatar
      [SPARK-17013][SQL] Parse negative numeric literals · 00e103a6
      petermaxlee authored
      ## What changes were proposed in this pull request?
      This patch updates the SQL parser to parse negative numeric literals as numeric literals, instead of unary minus of positive literals.
      
      This allows the parser to parse the minimal value for each data type, e.g. "-32768S".
      
      ## How was this patch tested?
      Updated test cases.
      
      Author: petermaxlee <petermaxlee@gmail.com>
      
      Closes #14608 from petermaxlee/SPARK-17013.
      00e103a6
    • Dongjoon Hyun's avatar
      [SPARK-16975][SQL] Column-partition path starting '_' should be handled correctly · abff92bf
      Dongjoon Hyun authored
      ## What changes were proposed in this pull request?
      
      Currently, Spark ignores path names starting with underscore `_` and `.`. This causes read-failures for the column-partitioned file data sources whose partition column names starts from '_', e.g. `_col`.
      
      **Before**
      ```scala
      scala> spark.range(10).withColumn("_locality_code", $"id").write.partitionBy("_locality_code").save("/tmp/parquet")
      scala> spark.read.parquet("/tmp/parquet")
      org.apache.spark.sql.AnalysisException: Unable to infer schema for ParquetFormat at /tmp/parquet20. It must be specified manually;
      ```
      
      **After**
      ```scala
      scala> spark.range(10).withColumn("_locality_code", $"id").write.partitionBy("_locality_code").save("/tmp/parquet")
      scala> spark.read.parquet("/tmp/parquet")
      res2: org.apache.spark.sql.DataFrame = [id: bigint, _locality_code: int]
      ```
      
      ## How was this patch tested?
      
      Pass the Jenkins with a new test case.
      
      Author: Dongjoon Hyun <dongjoon@apache.org>
      
      Closes #14585 from dongjoon-hyun/SPARK-16975-PARQUET.
      abff92bf
    • Yanbo Liang's avatar
      [MINOR][ML] Rename TreeEnsembleModels to TreeEnsembleModel for PySpark · ccc6dc0f
      Yanbo Liang authored
      ## What changes were proposed in this pull request?
      Fix the typo of ```TreeEnsembleModels``` for PySpark, it should ```TreeEnsembleModel``` which will be consistent with Scala. What's more, it represents a tree ensemble model, so  ```TreeEnsembleModel``` should be more reasonable. This should not be used public, so it will not involve  breaking change.
      
      ## How was this patch tested?
      No new tests, should pass existing ones.
      
      Author: Yanbo Liang <ybliang8@gmail.com>
      
      Closes #14454 from yanboliang/TreeEnsembleModel.
      ccc6dc0f
  6. Aug 11, 2016
    • hyukjinkwon's avatar
      [SPARK-16434][SQL] Avoid per-record type dispatch in JSON when reading · ac84fb64
      hyukjinkwon authored
      ## What changes were proposed in this pull request?
      
      Currently, `JacksonParser.parse` is doing type-based dispatch for each row to convert the tokens to appropriate values for Spark.
      It might not have to be done like this because the schema is already kept.
      
      So, appropriate converters can be created first according to the schema once, and then apply them to each row.
      
      This PR corrects `JacksonParser` so that it creates all converters for the schema once and then applies them to each row rather than type dispatching for every row.
      
      Benchmark was proceeded with the codes below:
      
      #### Parser tests
      
      **Before**
      
      ```scala
      test("Benchmark for JSON converter") {
        val N = 500 << 8
        val row =
          """{"struct":{"field1": true, "field2": 92233720368547758070},
          "structWithArrayFields":{"field1":[4, 5, 6], "field2":["str1", "str2"]},
          "arrayOfString":["str1", "str2"],
          "arrayOfInteger":[1, 2147483647, -2147483648],
          "arrayOfLong":[21474836470, 9223372036854775807, -9223372036854775808],
          "arrayOfBigInteger":[922337203685477580700, -922337203685477580800],
          "arrayOfDouble":[1.2, 1.7976931348623157E308, 4.9E-324, 2.2250738585072014E-308],
          "arrayOfBoolean":[true, false, true],
          "arrayOfNull":[null, null, null, null],
          "arrayOfStruct":[{"field1": true, "field2": "str1"}, {"field1": false}, {"field3": null}],
          "arrayOfArray1":[[1, 2, 3], ["str1", "str2"]],
          "arrayOfArray2":[[1, 2, 3], [1.1, 2.1, 3.1]]
         }"""
        val data = List.fill(N)(row)
        val dummyOption = new JSONOptions(Map.empty[String, String])
        val schema =
          InferSchema.infer(spark.sparkContext.parallelize(Seq(row)), "", dummyOption)
        val factory = new JsonFactory()
      
        val benchmark = new Benchmark("JSON converter", N)
        benchmark.addCase("convert JSON file", 10) { _ =>
          data.foreach { input =>
            val parser = factory.createParser(input)
            parser.nextToken()
            JacksonParser.convertRootField(factory, parser, schema)
          }
        }
        benchmark.run()
      }
      ```
      
      ```
      JSON converter:                          Best/Avg Time(ms)    Rate(M/s)   Per Row(ns)   Relative
      ------------------------------------------------------------------------------------------------
      convert JSON file                             1697 / 1807          0.1       13256.9       1.0X
      ```
      
      **After**
      
      ```scala
      test("Benchmark for JSON converter") {
        val N = 500 << 8
        val row =
          """{"struct":{"field1": true, "field2": 92233720368547758070},
          "structWithArrayFields":{"field1":[4, 5, 6], "field2":["str1", "str2"]},
          "arrayOfString":["str1", "str2"],
          "arrayOfInteger":[1, 2147483647, -2147483648],
          "arrayOfLong":[21474836470, 9223372036854775807, -9223372036854775808],
          "arrayOfBigInteger":[922337203685477580700, -922337203685477580800],
          "arrayOfDouble":[1.2, 1.7976931348623157E308, 4.9E-324, 2.2250738585072014E-308],
          "arrayOfBoolean":[true, false, true],
          "arrayOfNull":[null, null, null, null],
          "arrayOfStruct":[{"field1": true, "field2": "str1"}, {"field1": false}, {"field3": null}],
          "arrayOfArray1":[[1, 2, 3], ["str1", "str2"]],
          "arrayOfArray2":[[1, 2, 3], [1.1, 2.1, 3.1]]
         }"""
        val data = List.fill(N)(row)
        val dummyOption = new JSONOptions(Map.empty[String, String], new SQLConf())
        val schema =
          InferSchema.infer(spark.sparkContext.parallelize(Seq(row)), dummyOption)
      
        val benchmark = new Benchmark("JSON converter", N)
        benchmark.addCase("convert JSON file", 10) { _ =>
          val parser = new JacksonParser(schema, dummyOption)
          data.foreach { input =>
            parser.parse(input)
          }
        }
        benchmark.run()
      }
      ```
      
      ```
      JSON converter:                          Best/Avg Time(ms)    Rate(M/s)   Per Row(ns)   Relative
      ------------------------------------------------------------------------------------------------
      convert JSON file                             1401 / 1461          0.1       10947.4       1.0X
      ```
      
      It seems parsing time is improved by roughly ~20%
      
      #### End-to-End test
      
      ```scala
      test("Benchmark for JSON reader") {
        val N = 500 << 8
        val row =
          """{"struct":{"field1": true, "field2": 92233720368547758070},
          "structWithArrayFields":{"field1":[4, 5, 6], "field2":["str1", "str2"]},
          "arrayOfString":["str1", "str2"],
          "arrayOfInteger":[1, 2147483647, -2147483648],
          "arrayOfLong":[21474836470, 9223372036854775807, -9223372036854775808],
          "arrayOfBigInteger":[922337203685477580700, -922337203685477580800],
          "arrayOfDouble":[1.2, 1.7976931348623157E308, 4.9E-324, 2.2250738585072014E-308],
          "arrayOfBoolean":[true, false, true],
          "arrayOfNull":[null, null, null, null],
          "arrayOfStruct":[{"field1": true, "field2": "str1"}, {"field1": false}, {"field3": null}],
          "arrayOfArray1":[[1, 2, 3], ["str1", "str2"]],
          "arrayOfArray2":[[1, 2, 3], [1.1, 2.1, 3.1]]
         }"""
        val df = spark.sqlContext.read.json(spark.sparkContext.parallelize(List.fill(N)(row)))
        withTempPath { path =>
          df.write.format("json").save(path.getCanonicalPath)
      
          val benchmark = new Benchmark("JSON reader", N)
          benchmark.addCase("reading JSON file", 10) { _ =>
            spark.read.format("json").load(path.getCanonicalPath).collect()
          }
          benchmark.run()
        }
      }
      ```
      
      **Before**
      
      ```
      JSON reader:                             Best/Avg Time(ms)    Rate(M/s)   Per Row(ns)   Relative
      ------------------------------------------------------------------------------------------------
      reading JSON file                             6485 / 6924          0.0       50665.0       1.0X
      ```
      
      **After**
      
      ```
      JSON reader:                             Best/Avg Time(ms)    Rate(M/s)   Per Row(ns)   Relative
      ------------------------------------------------------------------------------------------------
      reading JSON file                             6350 / 6529          0.0       49609.3       1.0X
      ```
      
      ## How was this patch tested?
      
      Existing test cases should cover this.
      
      Author: hyukjinkwon <gurwls223@gmail.com>
      
      Closes #14102 from HyukjinKwon/SPARK-16434.
      ac84fb64
    • Jeff Zhang's avatar
      [SPARK-13081][PYSPARK][SPARK_SUBMIT] Allow set pythonExec of driver and executor through conf… · 7a9e25c3
      Jeff Zhang authored
      Before this PR, user have to export environment variable to specify the python of driver & executor which is not so convenient for users. This PR is trying to allow user to specify python through configuration "--pyspark-driver-python" & "--pyspark-executor-python"
      
      Manually test in local & yarn mode for pyspark-shell and pyspark batch mode.
      
      Author: Jeff Zhang <zjffdu@apache.org>
      
      Closes #13146 from zjffdu/SPARK-13081.
      7a9e25c3
    • WangTaoTheTonic's avatar
      [SPARK-17022][YARN] Handle potential deadlock in driver handling messages · ea0bf91b
      WangTaoTheTonic authored
      ## What changes were proposed in this pull request?
      
      We directly send RequestExecutors to AM instead of transfer it to yarnShedulerBackend first, to avoid potential deadlock.
      
      ## How was this patch tested?
      
      manual tests
      
      Author: WangTaoTheTonic <wangtao111@huawei.com>
      
      Closes #14605 from WangTaoTheTonic/lock.
      ea0bf91b
    • huangzhaowei's avatar
      [SPARK-16868][WEB UI] Fix executor be both dead and alive on executor ui. · 4ec5c360
      huangzhaowei authored
      ## What changes were proposed in this pull request?
      In a heavy pressure of the spark application, since the executor will register it to driver block manager twice(because of heart beats), the executor will show as picture show:
      ![image](https://cloud.githubusercontent.com/assets/7404824/17467245/c1359094-5d4e-11e6-843a-f6d6347e1bf6.png)
      
      ## How was this patch tested?
      NA
      
      Details in: [SPARK-16868](https://issues.apache.org/jira/browse/SPARK-16868)
      
      Author: huangzhaowei <carlmartinmax@gmail.com>
      
      Closes #14530 from SaintBacchus/SPARK-16868.
      4ec5c360
    • Bryan Cutler's avatar
      [SPARK-13602][CORE] Add shutdown hook to DriverRunner to prevent driver process leak · 1c9a386c
      Bryan Cutler authored
      ## What changes were proposed in this pull request?
      
      Added shutdown hook to DriverRunner to kill the driver process in case the Worker JVM exits suddenly and the `WorkerWatcher` was unable to properly catch this.  Did some cleanup to consolidate driver state management and setting of finalized vars within the running thread.
      
      ## How was this patch tested?
      
      Added unit tests to verify that final state and exception variables are set accordingly for successfull, failed, and errors in the driver process.  Retrofitted existing test to verify killing of mocked process ends with the correct state and stops properly
      
      Manually tested (with deploy-mode=cluster) that the shutdown hook is called by forcibly exiting the `Worker` and various points in the code with the `WorkerWatcher` both disabled and enabled.  Also, manually killed the driver through the ui and verified that the `DriverRunner` interrupted, killed the process and exited properly.
      
      Author: Bryan Cutler <cutlerb@gmail.com>
      
      Closes #11746 from BryanCutler/DriverRunner-shutdown-hook-SPARK-13602.
      1c9a386c
    • petermaxlee's avatar
      [SPARK-17018][SQL] literals.sql for testing literal parsing · cf936782
      petermaxlee authored
      ## What changes were proposed in this pull request?
      This patch adds literals.sql for testing literal parsing end-to-end in SQL.
      
      ## How was this patch tested?
      The patch itself is only about adding test cases.
      
      Author: petermaxlee <petermaxlee@gmail.com>
      
      Closes #14598 from petermaxlee/SPARK-17018-2.
      cf936782
    • Wenchen Fan's avatar
      [SPARK-17021][SQL] simplify the constructor parameters of QuantileSummaries · acaf2a81
      Wenchen Fan authored
      ## What changes were proposed in this pull request?
      
      1. `sampled` doesn't need to be `ArrayBuffer`, we never update it, but assign new value
      2. `count` doesn't need to be `var`, we never mutate it.
      3. `headSampled` doesn't need to be in constructor, we never pass a non-empty `headSampled` to constructor
      
      ## How was this patch tested?
      
      existing tests.
      
      Author: Wenchen Fan <wenchen@databricks.com>
      
      Closes #14603 from cloud-fan/simply.
      acaf2a81
    • Davies Liu's avatar
      [SPARK-16958] [SQL] Reuse subqueries within the same query · 0f72e4f0
      Davies Liu authored
      ## What changes were proposed in this pull request?
      
      There could be multiple subqueries that generate same results, we could re-use the result instead of running it multiple times.
      
      This PR also cleanup up how we run subqueries.
      
      For SQL query
      ```sql
      select id,(select avg(id) from t) from t where id > (select avg(id) from t)
      ```
      The explain is
      ```
      == Physical Plan ==
      *Project [id#15L, Subquery subquery29 AS scalarsubquery()#35]
      :  +- Subquery subquery29
      :     +- *HashAggregate(keys=[], functions=[avg(id#15L)])
      :        +- Exchange SinglePartition
      :           +- *HashAggregate(keys=[], functions=[partial_avg(id#15L)])
      :              +- *Range (0, 1000, splits=4)
      +- *Filter (cast(id#15L as double) > Subquery subquery29)
         :  +- Subquery subquery29
         :     +- *HashAggregate(keys=[], functions=[avg(id#15L)])
         :        +- Exchange SinglePartition
         :           +- *HashAggregate(keys=[], functions=[partial_avg(id#15L)])
         :              +- *Range (0, 1000, splits=4)
         +- *Range (0, 1000, splits=4)
      ```
      The visualized plan:
      
      ![reuse-subquery](https://cloud.githubusercontent.com/assets/40902/17573229/e578d93c-5f0d-11e6-8a3c-0150d81d3aed.png)
      
      ## How was this patch tested?
      
      Existing tests.
      
      Author: Davies Liu <davies@databricks.com>
      
      Closes #14548 from davies/subq.
      0f72e4f0
    • Michael Gummelt's avatar
      [SPARK-16952] don't lookup spark home directory when executor uri is set · 4d496802
      Michael Gummelt authored
      ## What changes were proposed in this pull request?
      
      remove requirement to set spark.mesos.executor.home when spark.executor.uri is used
      
      ## How was this patch tested?
      
      unit tests
      
      Author: Michael Gummelt <mgummelt@mesosphere.io>
      
      Closes #14552 from mgummelt/fix-spark-home.
      4d496802
    • hyukjinkwon's avatar
      [SPARK-16886][EXAMPLES][DOC] Fix some examples to be consistent and indentation in documentation · 7186e8c3
      hyukjinkwon authored
      ## What changes were proposed in this pull request?
      
      Originally this PR was based on #14491 but I realised that fixing examples are more sensible rather than comments.
      
      This PR fixes three things below:
      
       - Fix two wrong examples in `structured-streaming-programming-guide.md`. Loading via `read.load(..)` without `as` will be `Dataset<Row>` not `Dataset<String>` in Java.
      
      - Fix indentation across `structured-streaming-programming-guide.md`. Python has 4 spaces and Scala and Java have double spaces. These are inconsistent across the examples.
      
      - Fix `StructuredNetworkWordCountWindowed` and  `StructuredNetworkWordCount` in Java and Scala to initially load `DataFrame` and `Dataset<Row>` to be consistent with the comments and some examples in `structured-streaming-programming-guide.md` and to match Scala and Java to Python one (Python one loads it as `DataFrame` initially).
      
      ## How was this patch tested?
      
      N/A
      
      Closes https://github.com/apache/spark/pull/14491
      
      Author: hyukjinkwon <gurwls223@gmail.com>
      Author: Ganesh Chand <ganeshchand@Ganeshs-MacBook-Pro-2.local>
      
      Closes #14564 from HyukjinKwon/SPARK-16886.
      7186e8c3
    • huangzhaowei's avatar
      [SPARK-16941] Use concurrentHashMap instead of scala Map in SparkSQLOperationManager. · a45fefd1
      huangzhaowei authored
      ## What changes were proposed in this pull request?
      ThriftServer will have some thread-safe problem in **SparkSQLOperationManager**.
      Add a SynchronizedMap trait for the maps in it to avoid this problem.
      
      Details in [SPARK-16941](https://issues.apache.org/jira/browse/SPARK-16941)
      
      ## How was this patch tested?
      NA
      
      Author: huangzhaowei <carlmartinmax@gmail.com>
      
      Closes #14534 from SaintBacchus/SPARK-16941.
      a45fefd1
    • Andrew Ash's avatar
      Correct example value for spark.ssl.YYY.XXX settings · 8a6b7037
      Andrew Ash authored
      Docs adjustment to:
      - link to other relevant section of docs
      - correct statement about the only value when actually other values are supported
      
      Author: Andrew Ash <andrew@andrewash.com>
      
      Closes #14581 from ash211/patch-10.
      8a6b7037
    • petermaxlee's avatar
      [SPARK-17015][SQL] group-by/order-by ordinal and arithmetic tests · a7b02db4
      petermaxlee authored
      ## What changes were proposed in this pull request?
      This patch adds three test files:
      1. arithmetic.sql.out
      2. order-by-ordinal.sql
      3. group-by-ordinal.sql
      
      This includes https://github.com/apache/spark/pull/14594.
      
      ## How was this patch tested?
      This is a test case change.
      
      Author: petermaxlee <petermaxlee@gmail.com>
      
      Closes #14595 from petermaxlee/SPARK-17015.
      a7b02db4
    • petermaxlee's avatar
      [SPARK-17011][SQL] Support testing exceptions in SQLQueryTestSuite · 0db373aa
      petermaxlee authored
      ## What changes were proposed in this pull request?
      This patch adds exception testing to SQLQueryTestSuite. When there is an exception in query execution, the query result contains the the exception class along with the exception message.
      
      As part of this, I moved some additional test cases for limit from SQLQuerySuite over to SQLQueryTestSuite.
      
      ## How was this patch tested?
      This is a test harness change.
      
      Author: petermaxlee <petermaxlee@gmail.com>
      
      Closes #14592 from petermaxlee/SPARK-17011.
      0db373aa
    • Tao Wang's avatar
      [SPARK-17010][MINOR][DOC] Wrong description in memory management document · 7a6a3c3f
      Tao Wang authored
      ## What changes were proposed in this pull request?
      
      change the remain percent to right one.
      
      ## How was this patch tested?
      
      Manual review
      
      Author: Tao Wang <wangtao111@huawei.com>
      
      Closes #14591 from WangTaoTheTonic/patch-1.
      7a6a3c3f
Loading