Skip to content
Snippets Groups Projects
  1. Sep 15, 2016
    • Adam Roberts's avatar
      [SPARK-17379][BUILD] Upgrade netty-all to 4.0.41 final for bug fixes · 0ad8eeb4
      Adam Roberts authored
      ## What changes were proposed in this pull request?
      Upgrade netty-all to latest in the 4.0.x line which is 4.0.41, mentions several bug fixes and performance improvements we may find useful, see netty.io/news/2016/08/29/4-0-41-Final-4-1-5-Final.html. Initially tried to use 4.1.5 but noticed it's not backwards compatible.
      
      ## How was this patch tested?
      Existing unit tests against branch-1.6 and branch-2.0 using IBM Java 8 on Intel, Power and Z architectures
      
      Author: Adam Roberts <aroberts@uk.ibm.com>
      
      Closes #14961 from a-roberts/netty.
      0ad8eeb4
    • Tejas Patil's avatar
      [SPARK-17451][CORE] CoarseGrainedExecutorBackend should inform driver before self-kill · b4792781
      Tejas Patil authored
      ## What changes were proposed in this pull request?
      
      Jira : https://issues.apache.org/jira/browse/SPARK-17451
      
      `CoarseGrainedExecutorBackend` in some failure cases exits the JVM. While this does not have any issue, from the driver UI there is no specific reason captured for this. In this PR, I am adding functionality to `exitExecutor` to notify driver that the executor is exiting.
      
      ## How was this patch tested?
      
      Ran the change over a test env and took down shuffle service before the executor could register to it. In the driver logs, where the job failure reason is mentioned (ie. `Job aborted due to stage ...` it gives the correct reason:
      
      Before:
      `ExecutorLostFailure (executor ZZZZZZZZZ exited caused by one of the running tasks) Reason: Remote RPC client disassociated. Likely due to containers exceeding thresholds, or network issues. Check driver logs for WARN messages.`
      
      After:
      `ExecutorLostFailure (executor ZZZZZZZZZ exited caused by one of the running tasks) Reason: Unable to create executor due to java.util.concurrent.TimeoutException: Timeout waiting for task.`
      
      Author: Tejas Patil <tejasp@fb.com>
      
      Closes #15013 from tejasapatil/SPARK-17451_inform_driver.
      b4792781
    • Sean Owen's avatar
      [SPARK-17406][BUILD][HOTFIX] MiMa excludes fix · 2ad27695
      Sean Owen authored
      ## What changes were proposed in this pull request?
      
      Following https://github.com/apache/spark/pull/14969 for some reason the MiMa excludes weren't complete, but still passed the PR builder. This adds 3 more excludes from https://amplab.cs.berkeley.edu/jenkins/view/Spark%20QA%20Test%20(Dashboard)/job/spark-master-test-sbt-hadoop-2.2/1749/consoleFull
      
      It also moves the excludes to their own Seq in the build, as they probably should have been.
      Even though this is merged to 2.1.x only / master, I left the exclude in for 2.0.x in case we back port. It's a private API so is always a false positive.
      
      ## How was this patch tested?
      
      Jenkins build
      
      Author: Sean Owen <sowen@cloudera.com>
      
      Closes #15110 from srowen/SPARK-17406.2.
      2ad27695
    • John Muller's avatar
      [SPARK-17536][SQL] Minor performance improvement to JDBC batch inserts · 71a65825
      John Muller authored
      ## What changes were proposed in this pull request?
      
      Optimize a while loop during batch inserts
      
      ## How was this patch tested?
      
      Unit tests were done, specifically "mvn  test" for sql
      
      Author: John Muller <jmuller@us.imshealth.com>
      
      Closes #15098 from blue666man/SPARK-17536.
      71a65825
    • cenyuhai's avatar
      [SPARK-17406][WEB UI] limit timeline executor events · ad79fc0a
      cenyuhai authored
      ## What changes were proposed in this pull request?
      The job page will be too slow to open when there are thousands of executor events(added or removed). I found that in ExecutorsTab file, executorIdToData will not remove elements, it will increase all the time.Before this pr, it looks like [timeline1.png](https://issues.apache.org/jira/secure/attachment/12827112/timeline1.png). After this pr, it looks like [timeline2.png](https://issues.apache.org/jira/secure/attachment/12827113/timeline2.png)(we can set how many executor events will be displayed)
      
      Author: cenyuhai <cenyuhai@didichuxing.com>
      
      Closes #14969 from cenyuhai/SPARK-17406.
      ad79fc0a
    • codlife's avatar
      [SPARK-17521] Error when I use sparkContext.makeRDD(Seq()) · 647ee05e
      codlife authored
      ## What changes were proposed in this pull request?
      
       when i use sc.makeRDD below
      ```
      val data3 = sc.makeRDD(Seq())
      println(data3.partitions.length)
      ```
      I got an error:
      Exception in thread "main" java.lang.IllegalArgumentException: Positive number of slices required
      
      We can fix this bug just modify the last line ,do a check of seq.size
      ```
        def makeRDD[T: ClassTag](seq: Seq[(T, Seq[String])]): RDD[T] = withScope {
          assertNotStopped()
          val indexToPrefs = seq.zipWithIndex.map(t => (t._2, t._1._2)).toMap
          new ParallelCollectionRDD[T](this, seq.map(_._1), math.max(seq.size, defaultParallelism), indexToPrefs)
        }
      ```
      
      ## How was this patch tested?
      
       manual tests
      
      (If this patch involves UI changes, please attach a screenshot; otherwise, remove this)
      
      Author: codlife <1004910847@qq.com>
      Author: codlife <wangjianfei15@otcaix.iscas.ac.cn>
      
      Closes #15077 from codlife/master.
      647ee05e
    • Adam Roberts's avatar
      [SPARK-17524][TESTS] Use specified spark.buffer.pageSize · f893e262
      Adam Roberts authored
      ## What changes were proposed in this pull request?
      
      This PR has the appendRowUntilExceedingPageSize test in RowBasedKeyValueBatchSuite use whatever spark.buffer.pageSize value a user has specified to prevent a test failure for anyone testing Apache Spark on a box with a reduced page size. The test is currently hardcoded to use the default page size which is 64 MB so this minor PR is a test improvement
      
      ## How was this patch tested?
      Existing unit tests with 1 MB page size and with 64 MB (the default) page size
      
      Author: Adam Roberts <aroberts@uk.ibm.com>
      
      Closes #15079 from a-roberts/patch-5.
      f893e262
    • WeichenXu's avatar
      [SPARK-17507][ML][MLLIB] check weight vector size in ANN · d15b4f90
      WeichenXu authored
      ## What changes were proposed in this pull request?
      
      as the TODO described,
      check weight vector size and if wrong throw exception.
      
      ## How was this patch tested?
      
      existing tests.
      
      Author: WeichenXu <WeichenXu123@outlook.com>
      
      Closes #15060 from WeichenXu123/check_input_weight_size_of_ann.
      d15b4f90
    • gatorsmile's avatar
      [SPARK-17440][SPARK-17441] Fixed Multiple Bugs in ALTER TABLE · 6a6adb16
      gatorsmile authored
      ### What changes were proposed in this pull request?
      For the following `ALTER TABLE` DDL, we should issue an exception when the target table is a `VIEW`:
      ```SQL
       ALTER TABLE viewName SET LOCATION '/path/to/your/lovely/heart'
      
       ALTER TABLE viewName SET SERDE 'whatever'
      
       ALTER TABLE viewName SET SERDEPROPERTIES ('x' = 'y')
      
       ALTER TABLE viewName PARTITION (a=1, b=2) SET SERDEPROPERTIES ('x' = 'y')
      
       ALTER TABLE viewName ADD IF NOT EXISTS PARTITION (a='4', b='8')
      
       ALTER TABLE viewName DROP IF EXISTS PARTITION (a='2')
      
       ALTER TABLE viewName RECOVER PARTITIONS
      
       ALTER TABLE viewName PARTITION (a='1', b='q') RENAME TO PARTITION (a='100', b='p')
      ```
      
      In addition, `ALTER TABLE RENAME PARTITION` is unable to handle data source tables, just like the other `ALTER PARTITION` commands. We should issue an exception instead.
      
      ### How was this patch tested?
      Added a few test cases.
      
      Author: gatorsmile <gatorsmile@gmail.com>
      
      Closes #15004 from gatorsmile/altertable.
      6a6adb16
  2. Sep 14, 2016
    • Xing SHI's avatar
      [SPARK-17465][SPARK CORE] Inappropriate memory management in... · bb322943
      Xing SHI authored
      [SPARK-17465][SPARK CORE] Inappropriate memory management in `org.apache.spark.storage.MemoryStore` may lead to memory leak
      
      The expression like `if (memoryMap(taskAttemptId) == 0) memoryMap.remove(taskAttemptId)` in method `releaseUnrollMemoryForThisTask` and `releasePendingUnrollMemoryForThisTask` should be called after release memory operation, whatever `memoryToRelease` is > 0 or not.
      
      If the memory of a task has been set to 0 when calling a `releaseUnrollMemoryForThisTask` or a `releasePendingUnrollMemoryForThisTask` method, the key in the memory map corresponding to that task will never be removed from the hash map.
      
      See the details in [SPARK-17465](https://issues.apache.org/jira/browse/SPARK-17465).
      
      Author: Xing SHI <shi-kou@indetail.co.jp>
      
      Closes #15022 from saturday-shi/SPARK-17465.
      bb322943
    • Eric Liang's avatar
      [SPARK-17472] [PYSPARK] Better error message for serialization failures of large objects in Python · dbfc7aa4
      Eric Liang authored
      ## What changes were proposed in this pull request?
      
      For large objects, pickle does not raise useful error messages. However, we can wrap them to be slightly more user friendly:
      
      Example 1:
      ```
      def run():
        import numpy.random as nr
        b = nr.bytes(8 * 1000000000)
        sc.parallelize(range(1000), 1000).map(lambda x: len(b)).count()
      
      run()
      ```
      
      Before:
      ```
      error: 'i' format requires -2147483648 <= number <= 2147483647
      ```
      
      After:
      ```
      pickle.PicklingError: Object too large to serialize: 'i' format requires -2147483648 <= number <= 2147483647
      ```
      
      Example 2:
      ```
      def run():
        import numpy.random as nr
        b = sc.broadcast(nr.bytes(8 * 1000000000))
        sc.parallelize(range(1000), 1000).map(lambda x: len(b.value)).count()
      
      run()
      ```
      
      Before:
      ```
      SystemError: error return without exception set
      ```
      
      After:
      ```
      cPickle.PicklingError: Could not serialize broadcast: SystemError: error return without exception set
      ```
      
      ## How was this patch tested?
      
      Manually tried out these cases
      
      cc davies
      
      Author: Eric Liang <ekl@databricks.com>
      
      Closes #15026 from ericl/spark-17472.
      dbfc7aa4
    • Shixiong Zhu's avatar
      [SPARK-17463][CORE] Make CollectionAccumulator and SetAccumulator's value can be read thread-safely · e33bfaed
      Shixiong Zhu authored
      ## What changes were proposed in this pull request?
      
      Make CollectionAccumulator and SetAccumulator's value can be read thread-safely to fix the ConcurrentModificationException reported in [JIRA](https://issues.apache.org/jira/browse/SPARK-17463).
      
      ## How was this patch tested?
      
      Existing tests.
      
      Author: Shixiong Zhu <shixiong@databricks.com>
      
      Closes #15063 from zsxwing/SPARK-17463.
      e33bfaed
    • Kishor Patil's avatar
      [SPARK-17511] Yarn Dynamic Allocation: Avoid marking released container as Failed · ff6e4cbd
      Kishor Patil authored
      ## What changes were proposed in this pull request?
      
      Due to race conditions, the ` assert(numExecutorsRunning <= targetNumExecutors)` can fail causing `AssertionError`. So removed the assertion, instead moved the conditional check before launching new container:
      ```
      java.lang.AssertionError: assertion failed
              at scala.Predef$.assert(Predef.scala:156)
              at org.apache.spark.deploy.yarn.YarnAllocator$$anonfun$runAllocatedContainers$1.org$apache$spark$deploy$yarn$YarnAllocator$$anonfun$$updateInternalState$1(YarnAllocator.scala:489)
              at org.apache.spark.deploy.yarn.YarnAllocator$$anonfun$runAllocatedContainers$1$$anon$1.run(YarnAllocator.scala:519)
              at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
              at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
              at java.lang.Thread.run(Thread.java:745)
      ```
      ## How was this patch tested?
      This was manually tested using a large ForkAndJoin job with Dynamic Allocation enabled to validate the failing job succeeds, without any such exception.
      
      Author: Kishor Patil <kpatil@yahoo-inc.com>
      
      Closes #15069 from kishorvpatil/SPARK-17511.
      ff6e4cbd
    • Xin Wu's avatar
      [SPARK-10747][SQL] Support NULLS FIRST|LAST clause in ORDER BY · 040e4697
      Xin Wu authored
      ## What changes were proposed in this pull request?
      Currently, ORDER BY clause returns nulls value according to sorting order (ASC|DESC), considering null value is always smaller than non-null values.
      However, SQL2003 standard support NULLS FIRST or NULLS LAST to allow users to specify whether null values should be returned first or last, regardless of sorting order (ASC|DESC).
      
      This PR is to support this new feature.
      
      ## How was this patch tested?
      New test cases are added to test NULLS FIRST|LAST for regular select queries and windowing queries.
      
      (If this patch involves UI changes, please attach a screenshot; otherwise, remove this)
      
      Author: Xin Wu <xinwu@us.ibm.com>
      
      Closes #14842 from xwu0226/SPARK-10747.
      040e4697
    • hyukjinkwon's avatar
      [MINOR][SQL] Add missing functions for some options in SQLConf and use them where applicable · a79838bd
      hyukjinkwon authored
      ## What changes were proposed in this pull request?
      
      I first thought they are missing because they are kind of hidden options but it seems they are just missing.
      
      For example, `spark.sql.parquet.mergeSchema` is documented in [sql-programming-guide.md](https://github.com/apache/spark/blob/master/docs/sql-programming-guide.md) but this function is missing whereas many options such as `spark.sql.join.preferSortMergeJoin` are not documented but have its own function individually.
      
      So, this PR suggests making them consistent by adding the missing functions for some options in `SQLConf` and use them where applicable, in order to make them more readable.
      
      ## How was this patch tested?
      
      Existing tests should cover this.
      
      Author: hyukjinkwon <gurwls223@gmail.com>
      
      Closes #14678 from HyukjinKwon/sqlconf-cleanup.
      a79838bd
    • Josh Rosen's avatar
      [SPARK-17514] df.take(1) and df.limit(1).collect() should perform the same in Python · 6d06ff6f
      Josh Rosen authored
      ## What changes were proposed in this pull request?
      
      In PySpark, `df.take(1)` runs a single-stage job which computes only one partition of the DataFrame, while `df.limit(1).collect()` computes all partitions and runs a two-stage job. This difference in performance is confusing.
      
      The reason why `limit(1).collect()` is so much slower is that `collect()` internally maps to `df.rdd.<some-pyspark-conversions>.toLocalIterator`, which causes Spark SQL to build a query where a global limit appears in the middle of the plan; this, in turn, ends up being executed inefficiently because limits in the middle of plans are now implemented by repartitioning to a single task rather than by running a `take()` job on the driver (this was done in #7334, a patch which was a prerequisite to allowing partition-local limits to be pushed beneath unions, etc.).
      
      In order to fix this performance problem I think that we should generalize the fix from SPARK-10731 / #8876 so that `DataFrame.collect()` also delegates to the Scala implementation and shares the same performance properties. This patch modifies `DataFrame.collect()` to first collect all results to the driver and then pass them to Python, allowing this query to be planned using Spark's `CollectLimit` optimizations.
      
      ## How was this patch tested?
      
      Added a regression test in `sql/tests.py` which asserts that the expected number of jobs, stages, and tasks are run for both queries.
      
      Author: Josh Rosen <joshrosen@databricks.com>
      
      Closes #15068 from JoshRosen/pyspark-collect-limit.
      6d06ff6f
    • gatorsmile's avatar
      [SPARK-17409][SQL] Do Not Optimize Query in CTAS More Than Once · 52738d4e
      gatorsmile authored
      ### What changes were proposed in this pull request?
      As explained in https://github.com/apache/spark/pull/14797:
      >Some analyzer rules have assumptions on logical plans, optimizer may break these assumption, we should not pass an optimized query plan into QueryExecution (will be analyzed again), otherwise we may some weird bugs.
      For example, we have a rule for decimal calculation to promote the precision before binary operations, use PromotePrecision as placeholder to indicate that this rule should not apply twice. But a Optimizer rule will remove this placeholder, that break the assumption, then the rule applied twice, cause wrong result.
      
      We should not optimize the query in CTAS more than once. For example,
      ```Scala
      spark.range(99, 101).createOrReplaceTempView("tab1")
      val sqlStmt = "SELECT id, cast(id as long) * cast('1.0' as decimal(38, 18)) as num FROM tab1"
      sql(s"CREATE TABLE tab2 USING PARQUET AS $sqlStmt")
      checkAnswer(spark.table("tab2"), sql(sqlStmt))
      ```
      Before this PR, the results do not match
      ```
      == Results ==
      !== Correct Answer - 2 ==       == Spark Answer - 2 ==
      ![100,100.000000000000000000]   [100,null]
       [99,99.000000000000000000]     [99,99.000000000000000000]
      ```
      After this PR, the results match.
      ```
      +---+----------------------+
      |id |num                   |
      +---+----------------------+
      |99 |99.000000000000000000 |
      |100|100.000000000000000000|
      +---+----------------------+
      ```
      
      In this PR, we do not treat the `query` in CTAS as a child. Thus, the `query` will not be optimized when optimizing CTAS statement. However, we still need to analyze it for normalizing and verifying the CTAS in the Analyzer. Thus, we do it in the analyzer rule `PreprocessDDL`, because so far only this rule needs the analyzed plan of the `query`.
      
      ### How was this patch tested?
      Added a test
      
      Author: gatorsmile <gatorsmile@gmail.com>
      
      Closes #15048 from gatorsmile/ctasOptimized.
      52738d4e
    • Sean Owen's avatar
      [SPARK-17445][DOCS] Reference an ASF page as the main place to find third-party packages · dc0a4c91
      Sean Owen authored
      ## What changes were proposed in this pull request?
      
      Point references to spark-packages.org to https://cwiki.apache.org/confluence/display/SPARK/Third+Party+Projects
      
      This will be accompanied by a parallel change to the spark-website repo, and additional changes to this wiki.
      
      ## How was this patch tested?
      
      Jenkins tests.
      
      Author: Sean Owen <sowen@cloudera.com>
      
      Closes #15075 from srowen/SPARK-17445.
      dc0a4c91
    • Ergin Seyfe's avatar
      [SPARK-17480][SQL] Improve performance by removing or caching List.length which is O(n) · 4cea9da2
      Ergin Seyfe authored
      ## What changes were proposed in this pull request?
      Scala's List.length method is O(N) and it makes the gatherCompressibilityStats function O(N^2). Eliminate the List.length calls by writing it in Scala way.
      
      https://github.com/scala/scala/blob/2.10.x/src/library/scala/collection/LinearSeqOptimized.scala#L36
      
      As suggested. Extended the fix to HiveInspectors and AggregationIterator classes as well.
      
      ## How was this patch tested?
      Profiled a Spark job and found that CompressibleColumnBuilder is using 39% of the CPU. Out of this 39% CompressibleColumnBuilder->gatherCompressibilityStats is using 23% of it. 6.24% of the CPU is spend on List.length which is called inside gatherCompressibilityStats.
      
      After this change we started to save 6.24% of the CPU.
      
      Author: Ergin Seyfe <eseyfe@fb.com>
      
      Closes #15032 from seyfe/gatherCompressibilityStats.
      4cea9da2
    • wm624@hotmail.com's avatar
      [CORE][DOC] remove redundant comment · 18b4f035
      wm624@hotmail.com authored
      ## What changes were proposed in this pull request?
      In the comment, there is redundant `the estimated`.
      
      This PR simply remove the redundant comment and adjusts format.
      
      Author: wm624@hotmail.com <wm624@hotmail.com>
      
      Closes #15091 from wangmiao1981/comment.
      18b4f035
    • Sami Jaktholm's avatar
      [SPARK-17525][PYTHON] Remove SparkContext.clearFiles() from the PySpark API as... · b5bfcddb
      Sami Jaktholm authored
      [SPARK-17525][PYTHON] Remove SparkContext.clearFiles() from the PySpark API as it was removed from the Scala API prior to Spark 2.0.0
      
      ## What changes were proposed in this pull request?
      
      This pull request removes the SparkContext.clearFiles() method from the PySpark API as the method was removed from the Scala API in 8ce645d4. Using that method in PySpark leads to an exception as PySpark tries to call the non-existent method on the JVM side.
      
      ## How was this patch tested?
      
      Existing tests (though none of them tested this particular method).
      
      Author: Sami Jaktholm <sjakthol@outlook.com>
      
      Closes #15081 from sjakthol/pyspark-sc-clearfiles.
      b5bfcddb
    • Jagadeesan's avatar
      [SPARK-17449][DOCUMENTATION] Relation between heartbeatInterval and… · def7c265
      Jagadeesan authored
      ## What changes were proposed in this pull request?
      
      The relation between spark.network.timeout and spark.executor.heartbeatInterval should be mentioned in the document.
      
      … network timeout]
      
      Author: Jagadeesan <as2@us.ibm.com>
      
      Closes #15042 from jagadeesanas2/SPARK-17449.
      def7c265
  3. Sep 13, 2016
    • junyangq's avatar
      [SPARK-17317][SPARKR] Add SparkR vignette · a454a4d8
      junyangq authored
      ## What changes were proposed in this pull request?
      
      This PR tries to add a SparkR vignette, which works as a friendly guidance going through the functionality provided by SparkR.
      
      ## How was this patch tested?
      
      Manual test.
      
      Author: junyangq <qianjunyang@gmail.com>
      Author: Shivaram Venkataraman <shivaram@cs.berkeley.edu>
      Author: Junyang Qian <junyangq@databricks.com>
      
      Closes #14980 from junyangq/SPARKR-vignette.
      a454a4d8
    • gatorsmile's avatar
      [SPARK-17530][SQL] Add Statistics into DESCRIBE FORMATTED · 37b93f54
      gatorsmile authored
      ### What changes were proposed in this pull request?
      Statistics is missing in the output of `DESCRIBE FORMATTED`. This PR is to add it. After the PR, the output will be like:
      ```
      +----------------------------+----------------------------------------------------------------------------------------------------------------------+-------+
      |col_name                    |data_type                                                                                                             |comment|
      +----------------------------+----------------------------------------------------------------------------------------------------------------------+-------+
      |key                         |string                                                                                                                |null   |
      |value                       |string                                                                                                                |null   |
      |                            |                                                                                                                      |       |
      |# Detailed Table Information|                                                                                                                      |       |
      |Database:                   |default                                                                                                               |       |
      |Owner:                      |xiaoli                                                                                                                |       |
      |Create Time:                |Tue Sep 13 14:36:57 PDT 2016                                                                                          |       |
      |Last Access Time:           |Wed Dec 31 16:00:00 PST 1969                                                                                          |       |
      |Location:                   |file:/private/var/folders/4b/sgmfldk15js406vk7lw5llzw0000gn/T/warehouse-9982e1db-df17-4376-a140-dbbee0203d83/texttable|       |
      |Table Type:                 |MANAGED                                                                                                               |       |
      |Statistics:                 |sizeInBytes=5812, rowCount=500, isBroadcastable=false                                                                 |       |
      |Table Parameters:           |                                                                                                                      |       |
      |  rawDataSize               |-1                                                                                                                    |       |
      |  numFiles                  |1                                                                                                                     |       |
      |  transient_lastDdlTime     |1473802620                                                                                                            |       |
      |  totalSize                 |5812                                                                                                                  |       |
      |  COLUMN_STATS_ACCURATE     |false                                                                                                                 |       |
      |  numRows                   |-1                                                                                                                    |       |
      |                            |                                                                                                                      |       |
      |# Storage Information       |                                                                                                                      |       |
      |SerDe Library:              |org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe                                                                    |       |
      |InputFormat:                |org.apache.hadoop.mapred.TextInputFormat                                                                              |       |
      |OutputFormat:               |org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat                                                            |       |
      |Compressed:                 |No                                                                                                                    |       |
      |Storage Desc Parameters:    |                                                                                                                      |       |
      |  serialization.format      |1                                                                                                                     |       |
      +----------------------------+----------------------------------------------------------------------------------------------------------------------+-------+
      ```
      
      Also improve the output of statistics in `DESCRIBE EXTENDED` by removing duplicate `Statistics`. Below is the example after the PR:
      
      ```
      +----------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-------+
      |col_name                    |data_type                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        |comment|
      +----------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-------+
      |key                         |string                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           |null   |
      |value                       |string                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           |null   |
      |                            |                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                 |       |
      |# Detailed Table Information|CatalogTable(
      	Table: `default`.`texttable`
      	Owner: xiaoli
      	Created: Tue Sep 13 14:38:43 PDT 2016
      	Last Access: Wed Dec 31 16:00:00 PST 1969
      	Type: MANAGED
      	Schema: [StructField(key,StringType,true), StructField(value,StringType,true)]
      	Provider: hive
      	Properties: [rawDataSize=-1, numFiles=1, transient_lastDdlTime=1473802726, totalSize=5812, COLUMN_STATS_ACCURATE=false, numRows=-1]
      	Statistics: sizeInBytes=5812, rowCount=500, isBroadcastable=false
      	Storage(Location: file:/private/var/folders/4b/sgmfldk15js406vk7lw5llzw0000gn/T/warehouse-8ea5c5a0-5680-4778-91cb-c6334cf8a708/texttable, InputFormat: org.apache.hadoop.mapred.TextInputFormat, OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, Serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, Properties: [serialization.format=1]))|       |
      +----------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-------+
      ```
      
      ### How was this patch tested?
      Manually tested.
      
      Author: gatorsmile <gatorsmile@gmail.com>
      
      Closes #15083 from gatorsmile/descFormattedStats.
      37b93f54
    • Burak Yavuz's avatar
      [SPARK-17531] Don't initialize Hive Listeners for the Execution Client · 72edc7e9
      Burak Yavuz authored
      ## What changes were proposed in this pull request?
      
      If a user provides listeners inside the Hive Conf, the configuration for these listeners are passed to the Hive Execution Client as well. This may cause issues for two reasons:
      1. The Execution Client will actually generate garbage
      2. The listener class needs to be both in the Spark Classpath and Hive Classpath
      
      This PR empties the listener configurations in `HiveUtils.newTemporaryConfiguration` so that the execution client will not contain the listener confs, but the metadata client will.
      
      ## How was this patch tested?
      
      Unit tests
      
      Author: Burak Yavuz <brkyvz@gmail.com>
      
      Closes #15086 from brkyvz/null-listeners.
      72edc7e9
    • jiangxingbo's avatar
      [SPARK-17142][SQL] Complex query triggers binding error in HashAggregateExec · 4ba63b19
      jiangxingbo authored
      ## What changes were proposed in this pull request?
      
      In `ReorderAssociativeOperator` rule, we extract foldable expressions with Add/Multiply arithmetics, and replace with eval literal. For example, `(a + 1) + (b + 2)` is optimized to `(a + b + 3)` by this rule.
      For aggregate operator, output expressions should be derived from groupingExpressions, current implemenation of `ReorderAssociativeOperator` rule may break this promise. A instance could be:
      ```
      SELECT
        ((t1.a + 1) + (t2.a + 2)) AS out_col
      FROM
        testdata2 AS t1
      INNER JOIN
        testdata2 AS t2
      ON
        (t1.a = t2.a)
      GROUP BY (t1.a + 1), (t2.a + 2)
      ```
      `((t1.a + 1) + (t2.a + 2))` is optimized to `(t1.a + t2.a + 3)`, which could not be derived from `ExpressionSet((t1.a +1), (t2.a + 2))`.
      Maybe we should improve the rule of `ReorderAssociativeOperator` by adding a GroupingExpressionSet to keep Aggregate.groupingExpressions, and respect these expressions during the optimize stage.
      
      ## How was this patch tested?
      
      Add new test case in `ReorderAssociativeOperatorSuite`.
      
      Author: jiangxingbo <jiangxb1987@gmail.com>
      
      Closes #14917 from jiangxb1987/rao.
      4ba63b19
    • Josh Rosen's avatar
      [SPARK-17515] CollectLimit.execute() should perform per-partition limits · 3f6a2bb3
      Josh Rosen authored
      ## What changes were proposed in this pull request?
      
      CollectLimit.execute() incorrectly omits per-partition limits, leading to performance regressions in case this case is hit (which should not happen in normal operation, but can occur in some cases (see #15068 for one example).
      
      ## How was this patch tested?
      
      Regression test in SQLQuerySuite that asserts the number of records scanned from the input RDD.
      
      Author: Josh Rosen <joshrosen@databricks.com>
      
      Closes #15070 from JoshRosen/SPARK-17515.
      3f6a2bb3
    • hyukjinkwon's avatar
      [BUILD] Closing some stale PRs and ones suggested to be closed by committer(s) · 46f5c201
      hyukjinkwon authored
      ## What changes were proposed in this pull request?
      
      This PR proposes to close some stale PRs and ones suggested to be closed by committer(s)
      
      Closes #10052
      Closes #11079
      Closes #12661
      Closes #12772
      Closes #12958
      Closes #12990
      Closes #13409
      Closes #13779
      Closes #13811
      Closes #14577
      Closes #14714
      Closes #14875
      Closes #15020
      
      ## How was this patch tested?
      
      N/A
      
      Author: hyukjinkwon <gurwls223@gmail.com>
      
      Closes #15057 from HyukjinKwon/closing-stale-pr.
      46f5c201
  4. Sep 12, 2016
    • Davies Liu's avatar
      [SPARK-17474] [SQL] fix python udf in TakeOrderedAndProjectExec · a91ab705
      Davies Liu authored
      ## What changes were proposed in this pull request?
      
      When there is any Python UDF in the Project between Sort and Limit, it will be collected into TakeOrderedAndProjectExec, ExtractPythonUDFs failed to pull the Python UDFs out because QueryPlan.expressions does not include the expression inside Option[Seq[Expression]].
      
      Ideally, we should fix the `QueryPlan.expressions`, but tried with no luck (it always run into infinite loop). In PR, I changed the TakeOrderedAndProjectExec to no use Option[Seq[Expression]] to workaround it. cc JoshRosen
      
      ## How was this patch tested?
      
      Added regression test.
      
      Author: Davies Liu <davies@databricks.com>
      
      Closes #15030 from davies/all_expr.
      a91ab705
    • Josh Rosen's avatar
      [SPARK-17485] Prevent failed remote reads of cached blocks from failing entire job · f9c580f1
      Josh Rosen authored
      ## What changes were proposed in this pull request?
      
      In Spark's `RDD.getOrCompute` we first try to read a local copy of a cached RDD block, then a remote copy, and only fall back to recomputing the block if no cached copy (local or remote) can be read. This logic works correctly in the case where no remote copies of the block exist, but if there _are_ remote copies and reads of those copies fail (due to network issues or internal Spark bugs) then the BlockManager will throw a `BlockFetchException` that will fail the task (and which could possibly fail the whole job if the read failures keep occurring).
      
      In the cases of TorrentBroadcast and task result fetching we really do want to fail the entire job in case no remote blocks can be fetched, but this logic is inappropriate for reads of cached RDD blocks because those can/should be recomputed in case cached blocks are unavailable.
      
      Therefore, I think that the `BlockManager.getRemoteBytes()` method should never throw on remote fetch errors and, instead, should handle failures by returning `None`.
      
      ## How was this patch tested?
      
      Block manager changes should be covered by modified tests in `BlockManagerSuite`: the old tests expected exceptions to be thrown on failed remote reads, while the modified tests now expect `None` to be returned from the `getRemote*` method.
      
      I also manually inspected all usages of `BlockManager.getRemoteValues()`, `getRemoteBytes()`, and `get()` to verify that they correctly pattern-match on the result and handle `None`. Note that these `None` branches are already exercised because the old `getRemoteBytes` returned `None` when no remote locations for the block could be found (which could occur if an executor died and its block manager de-registered with the master).
      
      Author: Josh Rosen <joshrosen@databricks.com>
      
      Closes #15037 from JoshRosen/SPARK-17485.
      f9c580f1
    • Josh Rosen's avatar
      [SPARK-14818] Post-2.0 MiMa exclusion and build changes · 7c51b99a
      Josh Rosen authored
      This patch makes a handful of post-Spark-2.0 MiMa exclusion and build updates. It should be merged to master and a subset of it should be picked into branch-2.0 in order to test Spark 2.0.1-SNAPSHOT.
      
      - Remove the ` sketch`, `mllibLocal`, and `streamingKafka010` from the list of excluded subprojects so that MiMa checks them.
      - Remove now-unnecessary special-case handling of the Kafka 0.8 artifact in `mimaSettings`.
      - Move the exclusion added in SPARK-14743 from `v20excludes` to `v21excludes`, since that patch was only merged into master and not branch-2.0.
      - Add exclusions for an API change introduced by SPARK-17096 / #14675.
      - Add missing exclusions for the `o.a.spark.internal` and `o.a.spark.sql.internal` packages.
      
      Author: Josh Rosen <joshrosen@databricks.com>
      
      Closes #15061 from JoshRosen/post-2.0-mima-changes.
      7c51b99a
    • Josh Rosen's avatar
      [SPARK-17483] Refactoring in BlockManager status reporting and block removal · 3d40896f
      Josh Rosen authored
      This patch makes three minor refactorings to the BlockManager:
      
      - Move the `if (info.tellMaster)` check out of `reportBlockStatus`; this fixes an issue where a debug logging message would incorrectly claim to have reported a block status to the master even though no message had been sent (in case `info.tellMaster == false`). This also makes it easier to write code which unconditionally sends block statuses to the master (which is necessary in another patch of mine).
      - Split  `removeBlock()` into two methods, the existing method and an internal `removeBlockInternal()` method which is designed to be called by internal code that already holds a write lock on the block. This is also needed by a followup patch.
      - Instead of calling `getCurrentBlockStatus()` in `removeBlock()`, just pass `BlockStatus.empty`; the block status should always be empty following complete removal of a block.
      
      These changes were originally authored as part of a bug fix patch which is targeted at branch-2.0 and master; I've split them out here into their own separate PR in order to make them easier to review and so that the behavior-changing parts of my other patch can be isolated to their own PR.
      
      Author: Josh Rosen <joshrosen@databricks.com>
      
      Closes #15036 from JoshRosen/cache-failure-race-conditions-refactorings-only.
      3d40896f
    • Sean Zhong's avatar
      [SPARK-17503][CORE] Fix memory leak in Memory store when unable to cache the whole RDD in memory · 1742c3ab
      Sean Zhong authored
      ## What changes were proposed in this pull request?
      
         MemoryStore may throw OutOfMemoryError when trying to cache a super big RDD that cannot fit in memory.
         ```
         scala> sc.parallelize(1 to 1000000000, 100).map(x => new Array[Long](1000)).cache().count()
      
         java.lang.OutOfMemoryError: Java heap space
      	at $line14.$read$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$anonfun$1.apply(<console>:24)
      	at $line14.$read$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$anonfun$1.apply(<console>:23)
      	at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
      	at scala.collection.Iterator$JoinIterator.next(Iterator.scala:232)
      	at org.apache.spark.storage.memory.PartiallyUnrolledIterator.next(MemoryStore.scala:683)
      	at org.apache.spark.InterruptibleIterator.next(InterruptibleIterator.scala:43)
      	at org.apache.spark.util.Utils$.getIteratorSize(Utils.scala:1684)
      	at org.apache.spark.rdd.RDD$$anonfun$count$1.apply(RDD.scala:1134)
      	at org.apache.spark.rdd.RDD$$anonfun$count$1.apply(RDD.scala:1134)
      	at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1915)
      	at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1915)
      	at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:70)
      	at org.apache.spark.scheduler.Task.run(Task.scala:86)
      	at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:274)
      	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
      	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
      	at java.lang.Thread.run(Thread.java:745)
         ```
      
      Spark MemoryStore uses SizeTrackingVector as a temporary unrolling buffer to store all input values that it has read so far before transferring the values to storage memory cache. The problem is that when the input RDD is too big for caching in memory, the temporary unrolling memory SizeTrackingVector is not garbage collected in time. As SizeTrackingVector can occupy all available storage memory, it may cause the executor JVM to run out of memory quickly.
      
      More info can be found at https://issues.apache.org/jira/browse/SPARK-17503
      
      ## How was this patch tested?
      
      Unit test and manual test.
      
      ### Before change
      
      Heap memory consumption
      <img width="702" alt="screen shot 2016-09-12 at 4 16 15 pm" src="https://cloud.githubusercontent.com/assets/2595532/18429524/60d73a26-7906-11e6-9768-6f286f5c58c8.png">
      
      Heap dump
      <img width="1402" alt="screen shot 2016-09-12 at 4 34 19 pm" src="https://cloud.githubusercontent.com/assets/2595532/18429577/cbc1ef20-7906-11e6-847b-b5903f450b3b.png">
      
      ### After change
      
      Heap memory consumption
      <img width="706" alt="screen shot 2016-09-12 at 4 29 10 pm" src="https://cloud.githubusercontent.com/assets/2595532/18429503/4abe9342-7906-11e6-844a-b2f815072624.png">
      
      Author: Sean Zhong <seanzhong@databricks.com>
      
      Closes #15056 from clockfly/memory_store_leak.
      1742c3ab
    • WeichenXu's avatar
      [SPARK CORE][MINOR] fix "default partitioner cannot partition array keys"... · 8087ecf8
      WeichenXu authored
      [SPARK CORE][MINOR] fix "default partitioner cannot partition array keys" error message in PairRDDfunctions
      
      ## What changes were proposed in this pull request?
      
      In order to avoid confusing user,
      error message in `PairRDDfunctions`
      `Default partitioner cannot partition array keys.`
      is updated,
      the one in `partitionBy` is replaced with
      `Specified partitioner cannot partition array keys.`
      other is replaced with
      `Specified or default partitioner cannot partition array keys.`
      
      ## How was this patch tested?
      
      N/A
      
      Author: WeichenXu <WeichenXu123@outlook.com>
      
      Closes #15045 from WeichenXu123/fix_partitionBy_error_message.
      8087ecf8
    • Gaetan Semet's avatar
      [SPARK-16992][PYSPARK] use map comprehension in doc · b3c22912
      Gaetan Semet authored
      Code is equivalent, but map comprehency is most of the time faster than a map.
      
      Author: Gaetan Semet <gaetan@xeberon.net>
      
      Closes #14863 from Stibbons/map_comprehension.
      b3c22912
    • codlife's avatar
      [SPARK-17447] Performance improvement in Partitioner.defaultPartitioner without sortBy · 4efcdb7f
      codlife authored
      ## What changes were proposed in this pull request?
      
      if there are many rdds in some situations,the sort will loss he performance servely,actually we needn't sort the rdds , we can just scan the rdds one time to gain the same goal.
      
      ## How was this patch tested?
      
      manual tests
      
      Author: codlife <1004910847@qq.com>
      
      Closes #15039 from codlife/master.
      4efcdb7f
    • cenyuhai's avatar
      [SPARK-17171][WEB UI] DAG will list all partitions in the graph · cc87280f
      cenyuhai authored
      ## What changes were proposed in this pull request?
      DAG will list all partitions in the graph, it is too slow and hard to see all graph.
      Always we don't want to see all partitions,we just want to see the relations of DAG graph.
      So I just show 2 root nodes for Rdds.
      
      Before this PR, the DAG graph looks like [dag1.png](https://issues.apache.org/jira/secure/attachment/12824702/dag1.png), [dag3.png](https://issues.apache.org/jira/secure/attachment/12825456/dag3.png), after this PR, the DAG graph looks like [dag2.png](https://issues.apache.org/jira/secure/attachment/12824703/dag2.png),[dag4.png](https://issues.apache.org/jira/secure/attachment/12825457/dag4.png)
      
      Author: cenyuhai <cenyuhai@didichuxing.com>
      Author: 岑玉海 <261810726@qq.com>
      
      Closes #14737 from cenyuhai/SPARK-17171.
      cc87280f
  5. Sep 11, 2016
    • Josh Rosen's avatar
      [SPARK-17486] Remove unused TaskMetricsUIData.updatedBlockStatuses field · 72eec70b
      Josh Rosen authored
      The `TaskMetricsUIData.updatedBlockStatuses` field is assigned to but never read, increasing the memory consumption of the web UI. We should remove this field.
      
      Author: Josh Rosen <joshrosen@databricks.com>
      
      Closes #15038 from JoshRosen/remove-updated-block-statuses-from-TaskMetricsUIData.
      72eec70b
    • Sameer Agarwal's avatar
      [SPARK-17415][SQL] Better error message for driver-side broadcast join OOMs · 767d4807
      Sameer Agarwal authored
      ## What changes were proposed in this pull request?
      
      This is a trivial patch that catches all `OutOfMemoryError` while building the broadcast hash relation and rethrows it by wrapping it in a nice error message.
      
      ## How was this patch tested?
      
      Existing Tests
      
      Author: Sameer Agarwal <sameerag@cs.berkeley.edu>
      
      Closes #14979 from sameeragarwal/broadcast-join-error.
      767d4807
    • Yanbo Liang's avatar
      [SPARK-17389][FOLLOW-UP][ML] Change KMeans k-means|| default init steps from 5 to 2. · 883c7631
      Yanbo Liang authored
      ## What changes were proposed in this pull request?
      #14956 reduced default k-means|| init steps to 2 from 5 only for spark.mllib package, we should also do same change for spark.ml and PySpark.
      
      ## How was this patch tested?
      Existing tests.
      
      Author: Yanbo Liang <ybliang8@gmail.com>
      
      Closes #15050 from yanboliang/spark-17389.
      883c7631
Loading