Skip to content
Snippets Groups Projects
  1. Nov 06, 2015
  2. Nov 05, 2015
    • Michael Armbrust's avatar
      [SPARK-11528] [SQL] Typed aggregations for Datasets · 363a476c
      Michael Armbrust authored
      This PR adds the ability to do typed SQL aggregations.  We will likely also want to provide an interface to allow users to do aggregations on objects, but this is deferred to another PR.
      
      ```scala
      val ds = Seq(("a", 10), ("a", 20), ("b", 1), ("b", 2), ("c", 1)).toDS()
      ds.groupBy(_._1).agg(sum("_2").as[Int]).collect()
      
      res0: Array(("a", 30), ("b", 3), ("c", 1))
      ```
      
      Author: Michael Armbrust <michael@databricks.com>
      
      Closes #9499 from marmbrus/dataset-agg.
      363a476c
    • Davies Liu's avatar
      [SPARK-7542][SQL] Support off-heap index/sort buffer · eec74ba8
      Davies Liu authored
      This brings the support of off-heap memory for array inside BytesToBytesMap and InMemorySorter, then we could allocate all the memory from off-heap for execution.
      
      Closes #8068
      
      Author: Davies Liu <davies@databricks.com>
      
      Closes #9477 from davies/unsafe_timsort.
      eec74ba8
    • Reynold Xin's avatar
      [SPARK-11540][SQL] API audit for QueryExecutionListener. · 3cc2c053
      Reynold Xin authored
      Author: Reynold Xin <rxin@databricks.com>
      
      Closes #9509 from rxin/SPARK-11540.
      3cc2c053
    • Marcelo Vanzin's avatar
      [SPARK-11538][BUILD] Force guava 14 in sbt build. · 5e31db70
      Marcelo Vanzin authored
      sbt's version resolution code always picks the most recent version, and we
      don't want that for guava.
      
      Author: Marcelo Vanzin <vanzin@cloudera.com>
      
      Closes #9508 from vanzin/SPARK-11538.
      5e31db70
    • jerryshao's avatar
      [SPARK-11457][STREAMING][YARN] Fix incorrect AM proxy filter conf recovery from checkpoint · 468ad0ae
      jerryshao authored
      Currently Yarn AM proxy filter configuration is recovered from checkpoint file when Spark Streaming application is restarted, which will lead to some unwanted behaviors:
      
      1. Wrong RM address if RM is redeployed from failure.
      2. Wrong proxyBase, since app id is updated, old app id for proxyBase is wrong.
      
      So instead of recovering from checkpoint file, these configurations should be reloaded each time when app started.
      
      This problem only exists in Yarn cluster mode, for Yarn client mode, these configurations will be updated with RPC message `AddWebUIFilter`.
      
      Please help to review tdas harishreedharan vanzin , thanks a lot.
      
      Author: jerryshao <sshao@hortonworks.com>
      
      Closes #9412 from jerryshao/SPARK-11457.
      468ad0ae
    • Yu ISHIKAWA's avatar
      [SPARK-11514][ML] Pass random seed to spark.ml DecisionTree* · 8fa8c837
      Yu ISHIKAWA authored
      cc jkbradley
      
      Author: Yu ISHIKAWA <yuu.ishikawa@gmail.com>
      
      Closes #9486 from yu-iskw/SPARK-11514.
      8fa8c837
    • Reynold Xin's avatar
      6091e91f
    • Davies Liu's avatar
      [SPARK-11537] [SQL] fix negative hours/minutes/seconds · 07414afa
      Davies Liu authored
      Currently, if the Timestamp is before epoch (1970/01/01), the hours, minutes and seconds will be negative (also rounding up).
      
      Author: Davies Liu <davies@databricks.com>
      
      Closes #9502 from davies/neg_hour.
      07414afa
    • Davies Liu's avatar
      [SPARK-11542] [SPARKR] fix glm with long fomular · 24401062
      Davies Liu authored
      Because deparse() will break the long string into multiple lines, the deserialization will fail
      
      Author: Davies Liu <davies@databricks.com>
      
      Closes #9510 from davies/fix_glm.
      24401062
    • Reynold Xin's avatar
      [SPARK-11536][SQL] Remove the internal implicit conversion from Expression to... · b6974f8f
      Reynold Xin authored
      [SPARK-11536][SQL] Remove the internal implicit conversion from Expression to Column in functions.scala
      
      Author: Reynold Xin <rxin@databricks.com>
      
      Closes #9505 from rxin/SPARK-11536.
      b6974f8f
    • Wenchen Fan's avatar
      [SPARK-10656][SQL] completely support special chars in DataFrame · d9e30c59
      Wenchen Fan authored
      the main problem is: we interpret column name with special handling of `.` for DataFrame. This enables us to write something like `df("a.b")` to get the field `b` of `a`. However, we don't need this feature in `DataFrame.apply("*")` or `DataFrame.withColumnRenamed`. In these 2 cases, the column name is the final name already, we don't need extra process to interpret it.
      
      The solution is simple, use `queryExecution.analyzed.output` to get resolved column directly, instead of using `DataFrame.resolve`.
      
      close https://github.com/apache/spark/pull/8811
      
      Author: Wenchen Fan <wenchen@databricks.com>
      
      Closes #9462 from cloud-fan/special-chars.
      d9e30c59
    • adrian555's avatar
      [SPARK-11260][SPARKR] with() function support · b9455d1f
      adrian555 authored
      Author: adrian555 <wzhuang@us.ibm.com>
      Author: Adrian Zhuang <adrian555@users.noreply.github.com>
      
      Closes #9443 from adrian555/with.
      b9455d1f
    • Reynold Xin's avatar
      [SPARK-11532][SQL] Remove implicit conversion from Expression to Column · 8a5314ef
      Reynold Xin authored
      Author: Reynold Xin <rxin@databricks.com>
      
      Closes #9500 from rxin/SPARK-11532.
      8a5314ef
    • Travis Hegner's avatar
      [SPARK-10648] Oracle dialect to handle nonspecific numeric types · 14ee0f57
      Travis Hegner authored
      This is the alternative/agreed upon solution to PR #8780.
      
      Creating an OracleDialect to handle the nonspecific numeric types that can be defined in oracle.
      
      Author: Travis Hegner <thegner@trilliumit.com>
      
      Closes #9495 from travishegner/OracleDialect.
      14ee0f57
    • Ehsan M.Kermani's avatar
      [SPARK-10265][DOCUMENTATION, ML] Fixed @Since annotation to ml.regression · f80f7b69
      Ehsan M.Kermani authored
      Here is my first commit.
      
      Author: Ehsan M.Kermani <ehsanmo1367@gmail.com>
      
      Closes #8728 from ehsanmok/SinceAnn.
      f80f7b69
    • Reynold Xin's avatar
      [SPARK-11513][SQL] Remove implicit conversion from LogicalPlan to DataFrame · 6b87acd6
      Reynold Xin authored
      This internal implicit conversion has been a source of confusion for a lot of new developers.
      
      Author: Reynold Xin <rxin@databricks.com>
      
      Closes #9479 from rxin/SPARK-11513.
      6b87acd6
    • Srinivasa Reddy Vundela's avatar
      [SPARK-11484][WEBUI] Using proxyBase set by spark AM · c76865c6
      Srinivasa Reddy Vundela authored
      Use the proxyBase set by the AM, if not found then use env. This is to fix the issue if somebody accidentally set APPLICATION_WEB_PROXY_BASE to wrong proxyBase
      
      Author: Srinivasa Reddy Vundela <vsr@cloudera.com>
      
      Closes #9448 from vundela/master.
      c76865c6
    • Yanbo Liang's avatar
      [SPARK-11473][ML] R-like summary statistics with intercept for OLS via normal equation solver · 9da7ceed
      Yanbo Liang authored
      Follow up [SPARK-9836](https://issues.apache.org/jira/browse/SPARK-9836), we should also support summary statistics for ```intercept```.
      
      Author: Yanbo Liang <ybliang8@gmail.com>
      
      Closes #9485 from yanboliang/spark-11473.
      9da7ceed
    • Huaxin Gao's avatar
      [SPARK-11474][SQL] change fetchSize to fetchsize · b072ff4d
      Huaxin Gao authored
      In DefaultDataSource.scala, it has
      override def createRelation(
      sqlContext: SQLContext,
      parameters: Map[String, String]): BaseRelation
      The parameters is CaseInsensitiveMap.
      After this line
      parameters.foreach(kv => properties.setProperty(kv._1, kv._2))
      properties is set to all lower case key/value pairs and fetchSize becomes fetchsize.
      However, in compute method in JDBCRDD, it has
      val fetchSize = properties.getProperty("fetchSize", "0").toInt
      so fetchSize value is always 0 and never gets set correctly.
      
      Author: Huaxin Gao <huaxing@oc0558782468.ibm.com>
      
      Closes #9473 from huaxingao/spark-11474.
      b072ff4d
    • Nishkam Ravi's avatar
      [SPARK-11501][CORE][YARN] Propagate spark.rpc config to executors · a4b5cefc
      Nishkam Ravi authored
      spark.rpc is supposed to be configurable but is not currently (doesn't get propagated to executors because RpcEnv.create is done before driver properties are fetched).
      
      Author: Nishkam Ravi <nishkamravi@gmail.com>
      
      Closes #9460 from nishkamravi2/master_akka.
      a4b5cefc
    • Yanbo Liang's avatar
      [SPARK-11527][ML][PYSPARK] PySpark AFTSurvivalRegressionModel should expose... · 2e86cf1b
      Yanbo Liang authored
      [SPARK-11527][ML][PYSPARK] PySpark AFTSurvivalRegressionModel should expose coefficients/intercept/scale
      
      PySpark ```AFTSurvivalRegressionModel``` should expose coefficients/intercept/scale. mengxr vectorijk
      
      Author: Yanbo Liang <ybliang8@gmail.com>
      
      Closes #9492 from yanboliang/spark-11527.
      2e86cf1b
    • Yanbo Liang's avatar
      [MINOR][ML][DOC] Rename weights to coefficients in user guide · 72634f27
      Yanbo Liang authored
      We should use ```coefficients``` rather than ```weights``` in user guide that freshman can get the right conventional name at the outset. mengxr vectorijk
      
      Author: Yanbo Liang <ybliang8@gmail.com>
      
      Closes #9493 from yanboliang/docs-coefficients.
      72634f27
    • Cheng Lian's avatar
      [MINOR][SQL] A minor log line fix · 77488fb8
      Cheng Lian authored
      `jars` in the log line is an array, so `$jars` doesn't print its content.
      
      Author: Cheng Lian <lian@databricks.com>
      
      Closes #9494 from liancheng/minor.log-fix.
      77488fb8
    • a1singh's avatar
      [SPARK-11506][MLLIB] Removed redundant operation in Online LDA implementation · a94671a0
      a1singh authored
      In file LDAOptimizer.scala:
      
      line 441: since "idx" was never used, replaced unrequired zipWithIndex.foreach with foreach.
      
      -      nonEmptyDocs.zipWithIndex.foreach { case ((_, termCounts: Vector), idx: Int) =>
      +      nonEmptyDocs.foreach { case (_, termCounts: Vector) =>
      
      Author: a1singh <a1singh@ucsd.edu>
      
      Closes #9456 from a1singh/master.
      a94671a0
    • Herman van Hovell's avatar
      [SPARK-11449][CORE] PortableDataStream should be a factory · 7bdc9219
      Herman van Hovell authored
      ```PortableDataStream``` maintains some internal state. This makes it tricky to reuse a stream (one needs to call ```close``` on both the ```PortableDataStream``` and the ```InputStream``` it produces).
      
      This PR removes all state from ```PortableDataStream``` and effectively turns it into an ```InputStream```/```Array[Byte]``` factory. This makes the user responsible for managing the ```InputStream``` it returns.
      
      cc srowen
      
      Author: Herman van Hovell <hvanhovell@questtec.nl>
      
      Closes #9417 from hvanhovell/SPARK-11449.
      7bdc9219
    • Nick Evans's avatar
      [SPARK-11378][STREAMING] make StreamingContext.awaitTerminationOrTimeout return properly · 859dff56
      Nick Evans authored
      This adds a failing test checking that `awaitTerminationOrTimeout` returns the expected value, and then fixes that failing test with the addition of a `return`.
      
      tdas zsxwing
      
      Author: Nick Evans <me@nicolasevans.org>
      
      Closes #9336 from manygrams/fix_await_termination_or_timeout.
      859dff56
    • Sean Owen's avatar
      [SPARK-11440][CORE][STREAMING][BUILD] Declare rest of @Experimental items... · 6f81eae2
      Sean Owen authored
      [SPARK-11440][CORE][STREAMING][BUILD] Declare rest of @Experimental items non-experimental if they've existed since 1.2.0
      
      Remove `Experimental` annotations in core, streaming for items that existed in 1.2.0 or before. The changes are:
      
      * SparkContext
        * binary{Files,Records} : 1.2.0
        * submitJob : 1.0.0
      * JavaSparkContext
        * binary{Files,Records} : 1.2.0
      * DoubleRDDFunctions, JavaDoubleRDD
        * {mean,sum}Approx : 1.0.0
      * PairRDDFunctions, JavaPairRDD
        * sampleByKeyExact : 1.2.0
        * countByKeyApprox : 1.0.0
      * PairRDDFunctions
        * countApproxDistinctByKey : 1.1.0
      * RDD
        * countApprox, countByValueApprox, countApproxDistinct : 1.0.0
      * JavaRDDLike
        * countApprox : 1.0.0
      * PythonHadoopUtil.Converter : 1.1.0
      * PortableDataStream : 1.2.0 (related to binaryFiles)
      * BoundedDouble : 1.0.0
      * PartialResult : 1.0.0
      * StreamingContext, JavaStreamingContext
        * binaryRecordsStream : 1.2.0
      * HiveContext
        * analyze : 1.2.0
      
      Author: Sean Owen <sowen@cloudera.com>
      
      Closes #9396 from srowen/SPARK-11440.
      6f81eae2
  3. Nov 04, 2015
    • Davies Liu's avatar
      [SPARK-11425] [SPARK-11486] Improve hybrid aggregation · 81498dd5
      Davies Liu authored
      After aggregation, the dataset could be smaller than inputs, so it's better to do hash based aggregation for all inputs, then using sort based aggregation to merge them.
      
      Author: Davies Liu <davies@databricks.com>
      
      Closes #9383 from davies/fix_switch.
      81498dd5
    • Josh Rosen's avatar
      [SPARK-11307] Reduce memory consumption of OutputCommitCoordinator · d0b56339
      Josh Rosen authored
      OutputCommitCoordinator uses a map in a place where an array would suffice, increasing its memory consumption for result stages with millions of tasks.
      
      This patch replaces that map with an array. The only tricky part of this is reasoning about the range of possible array indexes in order to make sure that we never index out of bounds.
      
      Author: Josh Rosen <joshrosen@databricks.com>
      
      Closes #9274 from JoshRosen/SPARK-11307.
      d0b56339
    • Zhenhua Wang's avatar
      [SPARK-11398] [SQL] unnecessary def dialectClassName in HiveContext, and... · a752ddad
      Zhenhua Wang authored
      [SPARK-11398] [SQL] unnecessary def dialectClassName in HiveContext, and misleading dialect conf at the start of spark-sql
      
      1. def dialectClassName in HiveContext is unnecessary.
      In HiveContext, if conf.dialect == "hiveql", getSQLDialect() will return new HiveQLDialect(this);
      else it will use super.getSQLDialect(). Then in super.getSQLDialect(), it calls dialectClassName, which is overriden in HiveContext and still return super.dialectClassName.
      So we'll never reach the code "classOf[HiveQLDialect].getCanonicalName" of def dialectClassName in HiveContext.
      
      2. When we start bin/spark-sql, the default context is HiveContext, and the corresponding dialect is hiveql.
      However, if we type "set spark.sql.dialect;", the result is "sql", which is inconsistent with the actual dialect and is misleading. For example, we can use sql like "create table" which is only allowed in hiveql, but this dialect conf shows it's "sql".
      Although this problem will not cause any execution error, it's misleading to spark sql users. Therefore I think we should fix it.
      In this pr, while procesing “set spark.sql.dialect” in SetCommand, I use "conf.dialect" instead of "getConf()" for the case of key == SQLConf.DIALECT.key, so that it will return the right dialect conf.
      
      Author: Zhenhua Wang <wangzhenhua@huawei.com>
      
      Closes #9349 from wzhfy/dialect.
      a752ddad
    • Josh Rosen's avatar
      [SPARK-11491] Update build to use Scala 2.10.5 · ce5e6a28
      Josh Rosen authored
      Spark should build against Scala 2.10.5, since that includes a fix for Scaladoc that will fix doc snapshot publishing: https://issues.scala-lang.org/browse/SI-8479
      
      Author: Josh Rosen <joshrosen@databricks.com>
      
      Closes #9450 from JoshRosen/upgrade-to-scala-2.10.5.
      ce5e6a28
    • Reynold Xin's avatar
      [SPARK-11510][SQL] Remove SQL aggregation tests for higher order statistics · b6e0a5ae
      Reynold Xin authored
      We have some aggregate function tests in both DataFrameAggregateSuite and SQLQuerySuite. The two have almost the same coverage and we should just remove the SQL one.
      
      Author: Reynold Xin <rxin@databricks.com>
      
      Closes #9475 from rxin/SPARK-11510.
      b6e0a5ae
Loading