Skip to content
Snippets Groups Projects
  1. Dec 10, 2014
    • Patrick Wendell's avatar
      MAINTENANCE: Automated closing of pull requests. · 36bdb5b7
      Patrick Wendell authored
      This commit exists to close the following pull requests on Github:
      
      Closes #2883 (close requested by 'pwendell')
      Closes #3364 (close requested by 'pwendell')
      Closes #4458 (close requested by 'pwendell')
      Closes #1574 (close requested by 'andrewor14')
      Closes #2546 (close requested by 'andrewor14')
      Closes #2516 (close requested by 'andrewor14')
      Closes #154 (close requested by 'andrewor14')
      36bdb5b7
    • Andrew Or's avatar
      [SPARK-4759] Fix driver hanging from coalescing partitions · 4f93d0ca
      Andrew Or authored
      The driver hangs sometimes when we coalesce RDD partitions. See JIRA for more details and reproduction.
      
      This is because our use of empty string as default preferred location in `CoalescedRDDPartition` causes the `TaskSetManager` to schedule the corresponding task on host `""` (empty string). The intended semantics here, however, is that the partition does not have a preferred location, and the TSM should schedule the corresponding task accordingly.
      
      Author: Andrew Or <andrew@databricks.com>
      
      Closes #3633 from andrewor14/coalesce-preferred-loc and squashes the following commits:
      
      e520d6b [Andrew Or] Oops
      3ebf8bd [Andrew Or] A few comments
      f370a4e [Andrew Or] Fix tests
      2f7dfb6 [Andrew Or] Avoid using empty string as default preferred location
      4f93d0ca
    • Ilya Ganelin's avatar
      [SPARK-4569] Rename 'externalSorting' in Aggregator · 447ae2de
      Ilya Ganelin authored
      Hi all - I've renamed the unhelpfully named variable and added a comment clarifying what's actually happening.
      
      Author: Ilya Ganelin <ilya.ganelin@capitalone.com>
      
      Closes #3666 from ilganeli/SPARK-4569B and squashes the following commits:
      
      1810394 [Ilya Ganelin] [SPARK-4569] Rename 'externalSorting' in Aggregator
      e2d2092 [Ilya Ganelin] [SPARK-4569] Rename 'externalSorting' in Aggregator
      d7cefec [Ilya Ganelin] [SPARK-4569] Rename 'externalSorting' in Aggregator
      5b3f39c [Ilya Ganelin] [SPARK-4569] Rename  in Aggregator
      447ae2de
    • Daoyuan Wang's avatar
      [SPARK-4793] [Deploy] ensure .jar at end of line · e230da18
      Daoyuan Wang authored
      sometimes I switch between different version and do not want to rebuild spark, so I rename assembly.jar into .jar.bak, but still caught by `compute-classpath.sh`
      
      Author: Daoyuan Wang <daoyuan.wang@intel.com>
      
      Closes #3641 from adrian-wang/jar and squashes the following commits:
      
      45cbfd0 [Daoyuan Wang] ensure .jar at end of line
      e230da18
    • Andrew Or's avatar
      [SPARK-4215] Allow requesting / killing executors only in YARN mode · faa8fd81
      Andrew Or authored
      Currently this doesn't do anything in other modes, so we might as well just disable it rather than having the user mistakenly rely on it.
      
      Author: Andrew Or <andrew@databricks.com>
      
      Closes #3615 from andrewor14/dynamic-allocation-yarn-only and squashes the following commits:
      
      ce6487a [Andrew Or] Allow requesting / killing executors only in YARN mode
      faa8fd81
    • Andrew Or's avatar
      [SPARK-4771][Docs] Document standalone cluster supervise mode · 56212831
      Andrew Or authored
      tdas looks like streaming already refers to the supervise mode. The link from there is broken though.
      
      Author: Andrew Or <andrew@databricks.com>
      
      Closes #3627 from andrewor14/document-supervise and squashes the following commits:
      
      9ca0908 [Andrew Or] Wording changes
      2b55ed2 [Andrew Or] Document standalone cluster supervise mode
      56212831
    • Kousuke Saruta's avatar
      [SPARK-4329][WebUI] HistoryPage pagenation · 0fc637b4
      Kousuke Saruta authored
      Current HistoryPage have links only to previous page or next page.
      I suggest to add index to access history pages easily.
      
      I implemented like following pics.
      
      If there are many pages, current page +/- N pages, head page and last page are indexed.
      
      ![2014-11-10 16 13 25](https://cloud.githubusercontent.com/assets/4736016/4986246/9c7bbac4-6937-11e4-8695-8634d039d5b6.png)
      ![2014-11-10 16 03 21](https://cloud.githubusercontent.com/assets/4736016/4986210/3951bb74-6937-11e4-8b4e-9f90d266d736.png)
      ![2014-11-10 16 03 39](https://cloud.githubusercontent.com/assets/4736016/4986211/3b196ad8-6937-11e4-9f81-74bc0a6dad5b.png)
      ![2014-11-10 16 03 49](https://cloud.githubusercontent.com/assets/4736016/4986213/40686138-6937-11e4-86c0-41100f0404f6.png)
      ![2014-11-10 16 04 04](https://cloud.githubusercontent.com/assets/4736016/4986215/4326c9b4-6937-11e4-87ac-0f30c86ec6e3.png)
      
      Author: Kousuke Saruta <sarutak@oss.nttdata.co.jp>
      
      Closes #3194 from sarutak/history-page-indexing and squashes the following commits:
      
      15d3d2d [Kousuke Saruta] Simplified code
      c93932e [Kousuke Saruta] Merge branch 'master' of git://git.apache.org/spark into history-page-indexing
      1c2f605 [Kousuke Saruta] Merge branch 'master' of git://git.apache.org/spark into history-page-indexing
      76b05e3 [Kousuke Saruta] Merge branch 'master' of git://git.apache.org/spark into history-page-indexing
      b2240f8 [Kousuke Saruta] Fixed style
      ec7922e [Kousuke Saruta] Simplified code
      755a004 [Kousuke Saruta] Merge branch 'master' of git://git.apache.org/spark into history-page-indexing
      cfa242b [Kousuke Saruta] Added index to HistoryPage
      0fc637b4
    • GuoQiang Li's avatar
      [SPARK-4161]Spark shell class path is not correctly set if... · 742e7093
      GuoQiang Li authored
      [SPARK-4161]Spark shell class path is not correctly set if "spark.driver.extraClassPath" is set in defaults.conf
      
      Author: GuoQiang Li <witgo@qq.com>
      
      Closes #3050 from witgo/SPARK-4161 and squashes the following commits:
      
      abb6fa4 [GuoQiang Li] move usejavacp opt to spark-shell
      89e39e7 [GuoQiang Li] review commit
      c2a6f04 [GuoQiang Li] Spark shell class path is not correctly set if "spark.driver.extraClassPath" is set in defaults.conf
      742e7093
    • Nathan Kronenfeld's avatar
      [SPARK-4772] Clear local copies of accumulators as soon as we're done with them · 94b377f9
      Nathan Kronenfeld authored
      Accumulators keep thread-local copies of themselves.  These copies were only cleared at the beginning of a task.  This meant that (a) the memory they used was tied up until the next task ran on that thread, and (b) if a thread died, the memory it had used for accumulators was locked up forever on that worker.
      
      This PR clears the thread-local copies of accumulators at the end of each task, in the tasks finally block, to make sure they are cleaned up between tasks.  It also stores them in a ThreadLocal object, so that if, for some reason, the thread dies, any memory they are using at the time should be freed up.
      
      Author: Nathan Kronenfeld <nkronenfeld@oculusinfo.com>
      
      Closes #3570 from nkronenfeld/Accumulator-Improvements and squashes the following commits:
      
      a581f3f [Nathan Kronenfeld] Change Accumulators to private[spark] instead of adding mima exclude to get around false positive in mima tests
      b6c2180 [Nathan Kronenfeld] Include MiMa exclude as per build error instructions - this version incompatibility should be irrelevent, as it will only surface if a master is talking to a worker running a different version of spark.
      537baad [Nathan Kronenfeld] Fuller refactoring as intended, incorporating JR's suggestions for ThreadLocal localAccums, and keeping clear(), but also calling it in tasks' finally block, rather than just at the beginning of the task.
      39a82f2 [Nathan Kronenfeld] Clear local copies of accumulators as soon as we're done with them
      94b377f9
    • Josh Rosen's avatar
      [Minor] Use <sup> tag for help icon in web UI page header · f79c1cfc
      Josh Rosen authored
      This small commit makes the `(?)` web UI help link into a superscript, which should address feedback that the current design makes it look like an error occurred or like information is missing.
      
      Before:
      
      ![image](https://cloud.githubusercontent.com/assets/50748/5370611/a3ed0034-7fd9-11e4-870f-05bd9faad5b9.png)
      
      After:
      
      ![image](https://cloud.githubusercontent.com/assets/50748/5370602/6c5ca8d6-7fd9-11e4-8d1a-568d71290aa7.png)
      
      Author: Josh Rosen <joshrosen@databricks.com>
      
      Closes #3659 from JoshRosen/webui-help-sup and squashes the following commits:
      
      bd72899 [Josh Rosen] Use <sup> tag for help icon in web UI page header.
      f79c1cfc
  2. Dec 09, 2014
    • Reynold Xin's avatar
      Config updates for the new shuffle transport. · 9bd9334f
      Reynold Xin authored
      Author: Reynold Xin <rxin@databricks.com>
      
      Closes #3657 from rxin/conf-update and squashes the following commits:
      
      7370eab [Reynold Xin] Config updates for the new shuffle transport.
      9bd9334f
    • Reynold Xin's avatar
      [SPARK-4740] Create multiple concurrent connections between two peer nodes in Netty. · 2b9b7268
      Reynold Xin authored
      It's been reported that when the number of disks is large and the number of nodes is small, Netty network throughput is low compared with NIO. We suspect the problem is that only a small number of disks are utilized to serve shuffle files at any given point, due to connection reuse. This patch adds a new config parameter to specify the number of concurrent connections between two peer nodes, default to 2.
      
      Author: Reynold Xin <rxin@databricks.com>
      
      Closes #3625 from rxin/SPARK-4740 and squashes the following commits:
      
      ad4241a [Reynold Xin] Updated javadoc.
      f33c72b [Reynold Xin] Code review feedback.
      0fefabb [Reynold Xin] Use double check in synchronization.
      41dfcb2 [Reynold Xin] Added test case.
      9076b4a [Reynold Xin] Fixed two NPEs.
      3e1306c [Reynold Xin] Minor style fix.
      4f21673 [Reynold Xin] [SPARK-4740] Create multiple concurrent connections between two peer nodes in Netty.
      2b9b7268
    • Sean Owen's avatar
      SPARK-4805 [CORE] BlockTransferMessage.toByteArray() trips assertion · d8f84f26
      Sean Owen authored
      Allocate enough room for type byte as well as message, to avoid tripping assertion about capacity of the buffer
      
      Author: Sean Owen <sowen@cloudera.com>
      
      Closes #3650 from srowen/SPARK-4805 and squashes the following commits:
      
      9e1d502 [Sean Owen] Allocate enough room for type byte as well as message, to avoid tripping assertion about capacity of the buffer
      d8f84f26
    • Sandy Ryza's avatar
      SPARK-4567. Make SparkJobInfo and SparkStageInfo serializable · 5e4c06f8
      Sandy Ryza authored
      Author: Sandy Ryza <sandy@cloudera.com>
      
      Closes #3426 from sryza/sandy-spark-4567 and squashes the following commits:
      
      cb4b8d2 [Sandy Ryza] SPARK-4567. Make SparkJobInfo and SparkStageInfo serializable
      5e4c06f8
    • hushan[胡珊]'s avatar
      [SPARK-4714] BlockManager.dropFromMemory() should check whether block has been... · 30dca924
      hushan[胡珊] authored
      [SPARK-4714] BlockManager.dropFromMemory() should check whether block has been removed after synchronizing on BlockInfo instance.
      
      After synchronizing on the `info` lock in the `removeBlock`/`dropOldBlocks`/`dropFromMemory` methods in BlockManager, the block that `info` represented may have already removed.
      
      The three methods have the same logic to get the `info` lock:
      ```
         info = blockInfo.get(id)
         if (info != null) {
           info.synchronized {
             // do something
           }
         }
      ```
      
      So, there is chance that when a thread enters the `info.synchronized` block, `info` has already been removed from the `blockInfo` map by some other thread who entered `info.synchronized` first.
      
      The `removeBlock` and `dropOldBlocks` methods are idempotent, so it's safe for them to run on blocks that have already been removed.
      But in `dropFromMemory` it may be problematic since it may drop block data which already removed into the diskstore, and this calls data store operations that are not designed to handle missing blocks.
      
      This patch fixes this issue by adding a check to `dropFromMemory` to test whether blocks have been removed by a racing thread.
      
      Author: hushan[胡珊] <hushan@xiaomi.com>
      
      Closes #3574 from suyanNone/refine-block-concurrency and squashes the following commits:
      
      edb989d [hushan[胡珊]] Refine code style and comments position
      55fa4ba [hushan[胡珊]] refine code
      e57e270 [hushan[胡珊]] add check info is already remove or not while having gotten info.syn
      30dca924
    • Kay Ousterhout's avatar
      [SPARK-4765] Make GC time always shown in UI. · 1f511063
      Kay Ousterhout authored
      This commit removes the GC time for each task from the set of
      optional, additional metrics, and instead always shows it for
      each task.
      
      cc pwendell
      
      Author: Kay Ousterhout <kayousterhout@gmail.com>
      
      Closes #3622 from kayousterhout/gc_time and squashes the following commits:
      
      15ac242 [Kay Ousterhout] Make TaskDetailsClassNames private[spark]
      e71d893 [Kay Ousterhout] [SPARK-4765] Make GC time always shown in UI.
      1f511063
    • maji2014's avatar
      [SPARK-4691][shuffle] Restructure a few lines in shuffle code · b3107446
      maji2014 authored
      In HashShuffleReader.scala and HashShuffleWriter.scala, no need to judge "dep.aggregator.isEmpty" again as this is judged by "dep.aggregator.isDefined"
      
      In SortShuffleWriter.scala, "dep.aggregator.isEmpty"  is better than "!dep.aggregator.isDefined" ?
      
      Author: maji2014 <maji3@asiainfo.com>
      
      Closes #3553 from maji2014/spark-4691 and squashes the following commits:
      
      bf7b14d [maji2014] change a elegant way for SortShuffleWriter.scala
      10d0cf0 [maji2014] change a elegant way
      d8f52dc [maji2014] code optimization for judgement
      b3107446
    • jbencook's avatar
      [SPARK-874] adding a --wait flag · 61f1a702
      jbencook authored
      This PR adds a --wait flag to the `./sbin/stop-all.sh` script.
      
      Author: jbencook <jbenjamincook@gmail.com>
      
      Closes #3567 from jbencook/master and squashes the following commits:
      
      d05c5bb [jbencook] [SPARK-874] adding a --wait flag
      61f1a702
    • Sandy Ryza's avatar
      SPARK-4338. [YARN] Ditch yarn-alpha. · 912563aa
      Sandy Ryza authored
      Sorry if this is a little premature with 1.2 still not out the door, but it will make other work like SPARK-4136 and SPARK-2089 a lot easier.
      
      Author: Sandy Ryza <sandy@cloudera.com>
      
      Closes #3215 from sryza/sandy-spark-4338 and squashes the following commits:
      
      1c5ac08 [Sandy Ryza] Update building Spark docs and remove unnecessary newline
      9c1421c [Sandy Ryza] SPARK-4338. Ditch yarn-alpha.
      912563aa
    • Cheng Hao's avatar
      [SPARK-4785][SQL] Initilize Hive UDFs on the driver and serialize them with a wrapper · 383c5555
      Cheng Hao authored
      Different from Hive 0.12.0, in Hive 0.13.1 UDF/UDAF/UDTF (aka Hive function) objects should only be initialized once on the driver side and then serialized to executors. However, not all function objects are serializable (e.g. GenericUDF doesn't implement Serializable). Hive 0.13.1 solves this issue with Kryo or XML serializer. Several utility ser/de methods are provided in class o.a.h.h.q.e.Utilities for this purpose. In this PR we chose Kryo for efficiency. The Kryo serializer used here is created in Hive. Spark Kryo serializer wasn't used because there's no available SparkConf instance.
      
      Author: Cheng Hao <hao.cheng@intel.com>
      Author: Cheng Lian <lian@databricks.com>
      
      Closes #3640 from chenghao-intel/udf_serde and squashes the following commits:
      
      8e13756 [Cheng Hao] Update the comment
      74466a3 [Cheng Hao] refactor as feedbacks
      396c0e1 [Cheng Hao] avoid Simple UDF to be serialized
      e9c3212 [Cheng Hao] update the comment
      19cbd46 [Cheng Hao] support udf instance ser/de after initialization
      383c5555
    • zsxwing's avatar
      [SPARK-3154][STREAMING] Replace ConcurrentHashMap with mutable.HashMap and... · bcb5cdad
      zsxwing authored
      [SPARK-3154][STREAMING] Replace ConcurrentHashMap with mutable.HashMap and remove @volatile from 'stopped'
      
      Since `sequenceNumberToProcessor` and `stopped` are both protected by the lock `sequenceNumberToProcessor`, `ConcurrentHashMap` and `volatile` is unnecessary. So this PR updated them accordingly.
      
      Author: zsxwing <zsxwing@gmail.com>
      
      Closes #3634 from zsxwing/SPARK-3154 and squashes the following commits:
      
      0d087ac [zsxwing] Replace ConcurrentHashMap with mutable.HashMap and remove @volatile from 'stopped'
      bcb5cdad
  3. Dec 08, 2014
    • Cheng Hao's avatar
      [SPARK-4769] [SQL] CTAS does not work when reading from temporary tables · 51b1fe14
      Cheng Hao authored
      This is the code refactor and follow ups for #2570
      
      Author: Cheng Hao <hao.cheng@intel.com>
      
      Closes #3336 from chenghao-intel/createtbl and squashes the following commits:
      
      3563142 [Cheng Hao] remove the unused variable
      e215187 [Cheng Hao] eliminate the compiling warning
      4f97f14 [Cheng Hao] fix bug in unittest
      5d58812 [Cheng Hao] revert the API changes
      b85b620 [Cheng Hao] fix the regression of temp tabl not found in CTAS
      51b1fe14
    • Jacky Li's avatar
      [SQL] remove unnecessary import in spark-sql · 94438436
      Jacky Li authored
      Author: Jacky Li <jacky.likun@huawei.com>
      
      Closes #3630 from jackylk/remove and squashes the following commits:
      
      150e7e0 [Jacky Li] remove unnecessary import
      94438436
    • Sandy Ryza's avatar
      SPARK-4770. [DOC] [YARN] spark.scheduler.minRegisteredResourcesRatio doc... · cda94d15
      Sandy Ryza authored
      ...umented default is incorrect for YARN
      
      Author: Sandy Ryza <sandy@cloudera.com>
      
      Closes #3624 from sryza/sandy-spark-4770 and squashes the following commits:
      
      bd81a3a [Sandy Ryza] SPARK-4770. [DOC] [YARN] spark.scheduler.minRegisteredResourcesRatio documented default is incorrect for YARN
      cda94d15
    • Sean Owen's avatar
      SPARK-3926 [CORE] Reopened: result of JavaRDD collectAsMap() is not serializable · e829bfa1
      Sean Owen authored
      My original 'fix' didn't fix at all. Now, there's a unit test to check whether it works. Of the two options to really fix it -- copy the `Map` to a `java.util.HashMap`, or copy and modify Scala's implementation in `Wrappers.MapWrapper`, I went with the latter.
      
      Author: Sean Owen <sowen@cloudera.com>
      
      Closes #3587 from srowen/SPARK-3926 and squashes the following commits:
      
      8586bb9 [Sean Owen] Remove unneeded no-arg constructor, and add additional note about copied code in LICENSE
      7bb0e66 [Sean Owen] Make SerializableMapWrapper actually serialize, and add unit test
      e829bfa1
    • Andrew Or's avatar
      [SPARK-4750] Dynamic allocation - synchronize kills · 65f929d5
      Andrew Or authored
      Simple omission on my part.
      
      Author: Andrew Or <andrew@databricks.com>
      
      Closes #3612 from andrewor14/dynamic-allocation-synchronization and squashes the following commits:
      
      1f03b60 [Andrew Or] Synchronize kills
      65f929d5
    • Kostas Sakellis's avatar
      [SPARK-4774] [SQL] Makes HiveFromSpark more portable · d6a972b3
      Kostas Sakellis authored
      HiveFromSpark read the kv1.txt file from SPARK_HOME/examples/src/main/resources/kv1.txt which assumed
      you had a source tree checked out. Now we copy the kv1.txt file to a temporary file and delete it when
      the jvm shuts down. This allows us to run this example outside of a spark source tree.
      
      Author: Kostas Sakellis <kostas@cloudera.com>
      
      Closes #3628 from ksakellis/kostas-spark-4774 and squashes the following commits:
      
      6770f83 [Kostas Sakellis] [SPARK-4774] [SQL] Makes HiveFromSpark more portable
      d6a972b3
    • Christophe Préaud's avatar
      [SPARK-4764] Ensure that files are fetched atomically · ab2abcb5
      Christophe Préaud authored
      tempFile is created in the same directory than targetFile, so that the
      move from tempFile to targetFile is always atomic
      
      Author: Christophe Préaud <christophe.preaud@kelkoo.com>
      
      Closes #2855 from preaudc/master and squashes the following commits:
      
      9ba89ca [Christophe Préaud] Ensure that files are fetched atomically
      54419ae [Christophe Préaud] Merge remote-tracking branch 'upstream/master'
      c6a5590 [Christophe Préaud] Revert commit 8ea871f8130b2490f1bad7374a819bf56f0ccbbd
      7456a33 [Christophe Préaud] Merge remote-tracking branch 'upstream/master'
      8ea871f [Christophe Préaud] Ensure that files are fetched atomically
      ab2abcb5
  4. Dec 07, 2014
    • Takeshi Yamamuro's avatar
      [SPARK-4620] Add unpersist in Graph and GraphImpl · 8817fc7f
      Takeshi Yamamuro authored
      Add an IF to uncache both vertices and edges of Graph/GraphImpl.
      This IF is useful when iterative graph operations build a new graph in each iteration, and the vertices and edges of previous iterations are no longer needed for following iterations.
      
      Author: Takeshi Yamamuro <linguin.m.s@gmail.com>
      
      This patch had conflicts when merged, resolved by
      Committer: Ankur Dave <ankurdave@gmail.com>
      
      Closes #3476 from maropu/UnpersistInGraphSpike and squashes the following commits:
      
      77a006a [Takeshi Yamamuro] Add unpersist in Graph and GraphImpl
      8817fc7f
    • Takeshi Yamamuro's avatar
      [SPARK-4646] Replace Scala.util.Sorting.quickSort with Sorter(TimSort) in Spark · 2e6b736b
      Takeshi Yamamuro authored
      This patch just replaces a native quick sorter with Sorter(TimSort) in Spark.
      It could get performance gains by ~8% in my quick experiments.
      
      Author: Takeshi Yamamuro <linguin.m.s@gmail.com>
      
      Closes #3507 from maropu/TimSortInEdgePartitionBuilderSpike and squashes the following commits:
      
      8d4e5d2 [Takeshi Yamamuro] Remove a wildcard import
      3527e00 [Takeshi Yamamuro] Replace Scala.util.Sorting.quickSort with Sorter(TimSort) in Spark
      2e6b736b
  5. Dec 06, 2014
  6. Dec 05, 2014
    • CrazyJvm's avatar
      Streaming doc : do you mean inadvertently? · 6eb1b6f6
      CrazyJvm authored
      Author: CrazyJvm <crazyjvm@gmail.com>
      
      Closes #3620 from CrazyJvm/streaming-foreachRDD and squashes the following commits:
      
      b72886b [CrazyJvm] do you mean inadvertently?
      6eb1b6f6
    • Zhang, Liye's avatar
      [SPARK-4005][CORE] handle message replies in receive instead of in the individual private methods · 98a7d099
      Zhang, Liye authored
      In BlockManagermasterActor, when handling message type UpdateBlockInfo, the message replies is in handled in individual private methods, should handle it in receive of Akka.
      
      Author: Zhang, Liye <liye.zhang@intel.com>
      
      Closes #2853 from liyezhang556520/akkaRecv and squashes the following commits:
      
      9b06f0a [Zhang, Liye] remove the unreachable code
      bf518cd [Zhang, Liye] change the indent
      242166b [Zhang, Liye] modified accroding to the comments
      d4b929b [Zhang, Liye] [SPARK-4005][CORE] handle message replies in receive instead of in the individual private methods
      98a7d099
    • Cheng Lian's avatar
      [SPARK-4761][SQL] Enables Kryo by default in Spark SQL Thrift server · 6f61e1f9
      Cheng Lian authored
      Enables Kryo and disables reference tracking by default in Spark SQL Thrift server. Configurations explicitly defined by users in `spark-defaults.conf` are respected (the Thrift server is started by `spark-submit`, which handles configuration properties properly).
      
      <!-- Reviewable:start -->
      [<img src="https://reviewable.io/review_button.png" height=40 alt="Review on Reviewable"/>](https://reviewable.io/reviews/apache/spark/3621)
      <!-- Reviewable:end -->
      
      Author: Cheng Lian <lian@databricks.com>
      
      Closes #3621 from liancheng/kryo-by-default and squashes the following commits:
      
      70c2775 [Cheng Lian] Enables Kryo by default in Spark SQL Thrift server
      6f61e1f9
    • Michael Armbrust's avatar
      [SPARK-4753][SQL] Use catalyst for partition pruning in newParquet. · f5801e81
      Michael Armbrust authored
      Author: Michael Armbrust <michael@databricks.com>
      
      Closes #3613 from marmbrus/parquetPartitionPruning and squashes the following commits:
      
      4f138f8 [Michael Armbrust] Use catalyst for partition pruning in newParquet.
      f5801e81
  7. Dec 04, 2014
Loading