Skip to content
Snippets Groups Projects
  1. Sep 22, 2015
  2. Sep 21, 2015
  3. Sep 20, 2015
  4. Sep 19, 2015
    • Josh Rosen's avatar
      [SPARK-10710] Remove ability to disable spilling in core and SQL · 2117eea7
      Josh Rosen authored
      It does not make much sense to set `spark.shuffle.spill` or `spark.sql.planner.externalSort` to false: I believe that these configurations were initially added as "escape hatches" to guard against bugs in the external operators, but these operators are now mature and well-tested. In addition, these configurations are not handled in a consistent way anymore: SQL's Tungsten codepath ignores these configurations and will continue to use spilling operators. Similarly, Spark Core's `tungsten-sort` shuffle manager does not respect `spark.shuffle.spill=false`.
      
      This pull request removes these configurations, adds warnings at the appropriate places, and deletes a large amount of code which was only used in code paths that did not support spilling.
      
      Author: Josh Rosen <joshrosen@databricks.com>
      
      Closes #8831 from JoshRosen/remove-ability-to-disable-spilling.
      2117eea7
    • zsxwing's avatar
      [SPARK-10155] [SQL] Change SqlParser to object to avoid memory leak · e789000b
      zsxwing authored
      Since `scala.util.parsing.combinator.Parsers` is thread-safe since Scala 2.10 (See [SI-4929](https://issues.scala-lang.org/browse/SI-4929)), we can change SqlParser to object to avoid memory leak.
      
      I didn't change other subclasses of `scala.util.parsing.combinator.Parsers` because there is only one instance in one SQLContext, which should not be an issue.
      
      Author: zsxwing <zsxwing@gmail.com>
      
      Closes #8357 from zsxwing/sql-memory-leak.
      e789000b
    • Alexis Seigneurin's avatar
      Fixed links to the API · d83b6aae
      Alexis Seigneurin authored
      Submitting this change on the master branch as requested in https://github.com/apache/spark/pull/8819#issuecomment-141505941
      
      Author: Alexis Seigneurin <alexis.seigneurin@gmail.com>
      
      Closes #8838 from aseigneurin/patch-2.
      d83b6aae
    • Kousuke Saruta's avatar
      [SPARK-10584] [SQL] [DOC] Documentation about the compatible Hive version is wrong. · d507f9c0
      Kousuke Saruta authored
      In Spark 1.5.0, Spark SQL is compatible with Hive 0.12.0 through 1.2.1 but the documentation is wrong.
      
      /CC yhuai
      
      Author: Kousuke Saruta <sarutak@oss.nttdata.co.jp>
      
      Closes #8776 from sarutak/SPARK-10584-2.
      d507f9c0
    • Andrew Or's avatar
      [SPARK-10474] [SQL] Aggregation fails to allocate memory for pointer array · 7ff8d68c
      Andrew Or authored
      When `TungstenAggregation` hits memory pressure, it switches from hash-based to sort-based aggregation in-place. However, in the process we try to allocate the pointer array for writing to the new `UnsafeExternalSorter` *before* actually freeing the memory from the hash map. This lead to the following exception:
      ```
       java.io.IOException: Could not acquire 65536 bytes of memory
              at org.apache.spark.util.collection.unsafe.sort.UnsafeExternalSorter.initializeForWriting(UnsafeExternalSorter.java:169)
              at org.apache.spark.util.collection.unsafe.sort.UnsafeExternalSorter.spill(UnsafeExternalSorter.java:220)
              at org.apache.spark.sql.execution.UnsafeKVExternalSorter.<init>(UnsafeKVExternalSorter.java:126)
              at org.apache.spark.sql.execution.UnsafeFixedWidthAggregationMap.destructAndCreateExternalSorter(UnsafeFixedWidthAggregationMap.java:257)
              at org.apache.spark.sql.execution.aggregate.TungstenAggregationIterator.switchToSortBasedAggregation(TungstenAggregationIterator.scala:435)
      ```
      
      Author: Andrew Or <andrew@databricks.com>
      
      Closes #8827 from andrewor14/allocate-pointer-array.
      7ff8d68c
  5. Sep 18, 2015
Loading