Skip to content
Snippets Groups Projects
  1. Feb 04, 2016
    • gatorsmile's avatar
      [SPARK-12850][SQL] Support Bucket Pruning (Predicate Pushdown for Bucketed Tables) · e3c75c63
      gatorsmile authored
      JIRA: https://issues.apache.org/jira/browse/SPARK-12850
      
      This PR is to support bucket pruning when the predicates are `EqualTo`, `EqualNullSafe`, `IsNull`, `In`, and `InSet`.
      
      Like HIVE, in this PR, the bucket pruning works when the bucketing key has one and only one column.
      
      So far, I do not find a way to verify how many buckets are actually scanned. However, I did verify it when doing the debug. Could you provide a suggestion how to do it properly? Thank you! cloud-fan yhuai rxin marmbrus
      
      BTW, we can add more cases to support complex predicate including `Or` and `And`. Please let me know if I should do it in this PR.
      
      Maybe we also need to add test cases to verify if bucket pruning works well for each data type.
      
      Author: gatorsmile <gatorsmile@gmail.com>
      
      Closes #10942 from gatorsmile/pruningBuckets.
      e3c75c63
    • Raafat Akkad's avatar
      [SPARK-13052] waitingApps metric doesn't show the number of apps currently in the WAITING state · 6dbfc407
      Raafat Akkad authored
      Author: Raafat Akkad <raafat.akkad@gmail.com>
      
      Closes #10959 from RaafatAkkad/master.
      6dbfc407
    • Andrew Or's avatar
      7a4b37f0
    • Shixiong Zhu's avatar
      [SPARK-13195][STREAMING] Fix NoSuchElementException when a state is not set... · 8e2f2963
      Shixiong Zhu authored
      [SPARK-13195][STREAMING] Fix NoSuchElementException when a state is not set but timeoutThreshold is defined
      
      Check the state Existence before calling get.
      
      Author: Shixiong Zhu <shixiong@databricks.com>
      
      Closes #11081 from zsxwing/SPARK-13195.
      8e2f2963
    • Andrew Or's avatar
      [SPARK-13079][SQL] InMemoryCatalog follow-ups · bd38dd6f
      Andrew Or authored
      This patch incorporates review feedback from #11069, which is already merged.
      
      Author: Andrew Or <andrew@databricks.com>
      
      Closes #11080 from andrewor14/catalog-follow-ups.
      bd38dd6f
    • Andrew Or's avatar
      [SPARK-12330][MESOS][HOTFIX] Rename timeout config · c756bda4
      Andrew Or authored
      The config already describes time and accepts a general format
      that is not restricted to ms. This commit renames the internal
      config to use a format that's consistent in Spark.
      c756bda4
    • Josh Rosen's avatar
      [SPARK-13168][SQL] Collapse adjacent repartition operators · 33212cb9
      Josh Rosen authored
      Spark SQL should collapse adjacent `Repartition` operators and only keep the last one.
      
      Author: Josh Rosen <joshrosen@databricks.com>
      
      Closes #11064 from JoshRosen/collapse-repartition.
      33212cb9
    • Andrew Or's avatar
      MAINTENANCE: Automated closing of pull requests. · 085f510a
      Andrew Or authored
      This commit exists to close the following pull requests on Github:
      
      Closes #7971 (requested by yhuai)
      Closes #8539 (requested by srowen)
      Closes #8746 (requested by yhuai)
      Closes #9288 (requested by andrewor14)
      Closes #9321 (requested by andrewor14)
      Closes #9935 (requested by JoshRosen)
      Closes #10442 (requested by andrewor14)
      Closes #10585 (requested by srowen)
      Closes #10785 (requested by srowen)
      Closes #10832 (requested by andrewor14)
      Closes #10941 (requested by marmbrus)
      Closes #11024 (requested by andrewor14)
      085f510a
    • Andrew Or's avatar
      [SPARK-13053][TEST] Unignore tests in InternalAccumulatorSuite · 15205da8
      Andrew Or authored
      These were ignored because they are incorrectly written; they don't actually trigger stage retries, which is what the tests are testing. These tests are now rewritten to induce stage retries through fetch failures.
      
      Note: there were 2 tests before and now there's only 1. What happened? It turns out that the case where we only resubmit a subset of of the original missing partitions is very difficult to simulate in tests without potentially introducing flakiness. This is because the `DAGScheduler` removes all map outputs associated with a given executor when this happens, and we will need multiple executors to trigger this case, and sometimes the scheduler still removes map outputs from all executors.
      
      Author: Andrew Or <andrew@databricks.com>
      
      Closes #10969 from andrewor14/unignore-accum-test.
      15205da8
    • Andrew Or's avatar
      [SPARK-13162] Standalone mode does not respect initial executors · 4120bcba
      Andrew Or authored
      Currently the Master would always set an application's initial executor limit to infinity. If the user specified `spark.dynamicAllocation.initialExecutors`, the config would not take effect. This is similar to #11047 but for standalone mode.
      
      Author: Andrew Or <andrew@databricks.com>
      
      Closes #11054 from andrewor14/standalone-da-initial.
      4120bcba
    • Holden Karau's avatar
      [SPARK-13164][CORE] Replace deprecated synchronized buffer in core · 62a7c283
      Holden Karau authored
      Building with scala 2.11 results in the warning trait SynchronizedBuffer in package mutable is deprecated: Synchronization via traits is deprecated as it is inherently unreliable. Consider java.util.concurrent.ConcurrentLinkedQueue as an alternative. Investigation shows we are already using ConcurrentLinkedQueue in other locations so switch our uses of SynchronizedBuffer to ConcurrentLinkedQueue.
      
      Author: Holden Karau <holden@us.ibm.com>
      
      Closes #11059 from holdenk/SPARK-13164-replace-deprecated-synchronized-buffer-in-core.
      62a7c283
    • Charles Allen's avatar
      [SPARK-12330][MESOS] Fix mesos coarse mode cleanup · 2eaeafe8
      Charles Allen authored
      In the current implementation the mesos coarse scheduler does not wait for the mesos tasks to complete before ending the driver. This causes a race where the task has to finish cleaning up before the mesos driver terminates it with a SIGINT (and SIGKILL after 3 seconds if the SIGINT doesn't work).
      
      This PR causes the mesos coarse scheduler to wait for the mesos tasks to finish (with a timeout defined by `spark.mesos.coarse.shutdown.ms`)
      
      This PR also fixes a regression caused by [SPARK-10987] whereby submitting a shutdown causes a race between the local shutdown procedure and the notification of the scheduler driver disconnection. If the scheduler driver disconnection wins the race, the coarse executor incorrectly exits with status 1 (instead of the proper status 0)
      
      With this patch the mesos coarse scheduler terminates properly, the executors clean up, and the tasks are reported as `FINISHED` in the Mesos console (as opposed to `KILLED` in < 1.6 or `FAILED` in 1.6 and later)
      
      Author: Charles Allen <charles@allen-net.com>
      
      Closes #10319 from drcrallen/SPARK-12330.
      2eaeafe8
    • Reynold Xin's avatar
      [SPARK-12828][SQL] Natural join follow-up · dee801ad
      Reynold Xin authored
      This is a small addendum to #10762 to make the code more robust again future changes.
      
      Author: Reynold Xin <rxin@databricks.com>
      
      Closes #11070 from rxin/SPARK-12828-natural-join.
      dee801ad
    • Liang-Chi Hsieh's avatar
      [SPARK-13113] [CORE] Remove unnecessary bit operation when decoding page number · d3908714
      Liang-Chi Hsieh authored
      JIRA: https://issues.apache.org/jira/browse/SPARK-13113
      
      As we shift bits right, looks like the bitwise AND operation is unnecessary.
      
      Author: Liang-Chi Hsieh <viirya@gmail.com>
      
      Closes #11002 from viirya/improve-decodepagenumber.
      d3908714
  2. Feb 03, 2016
    • Yuhao Yang's avatar
      [ML][DOC] fix wrong api link in ml onevsrest · c2c956bc
      Yuhao Yang authored
      minor fix for api link in ml onevsrest
      
      Author: Yuhao Yang <hhbyyh@gmail.com>
      
      Closes #11068 from hhbyyh/onevsrestDoc.
      c2c956bc
    • Daoyuan Wang's avatar
      [SPARK-12828][SQL] add natural join support · 0f81318a
      Daoyuan Wang authored
      Jira:
      https://issues.apache.org/jira/browse/SPARK-12828
      
      Author: Daoyuan Wang <daoyuan.wang@intel.com>
      
      Closes #10762 from adrian-wang/naturaljoin.
      0f81318a
    • Andrew Or's avatar
      [SPARK-13079][SQL] Extend and implement InMemoryCatalog · a6483112
      Andrew Or authored
      This is a step towards consolidating `SQLContext` and `HiveContext`.
      
      This patch extends the existing Catalog API added in #10982 to include methods for handling table partitions. In particular, a partition is identified by `PartitionSpec`, which is just a `Map[String, String]`. The Catalog is still not used by anything yet, but its API is now more or less complete and an implementation is fully tested.
      
      About 200 lines are test code.
      
      Author: Andrew Or <andrew@databricks.com>
      
      Closes #11069 from andrewor14/catalog.
      a6483112
    • Holden Karau's avatar
      [SPARK-13152][CORE] Fix task metrics deprecation warning · a8e2ba77
      Holden Karau authored
      Make an internal non-deprecated version of incBytesRead and incRecordsRead so we don't have unecessary deprecation warnings in our build.
      
      Right now incBytesRead and incRecordsRead are marked as deprecated and for internal use only. We should make private[spark] versions which are not deprecated and switch to those internally so as to not clutter up the warning messages when building.
      
      cc andrewor14 who did the initial deprecation
      
      Author: Holden Karau <holden@us.ibm.com>
      
      Closes #11056 from holdenk/SPARK-13152-fix-task-metrics-deprecation-warnings.
      a8e2ba77
    • Davies Liu's avatar
      [SPARK-13131] [SQL] Use best and average time in benchmark · de091452
      Davies Liu authored
      Best time is stabler than average time, also added a column for nano seconds per row (which could be used to estimate contributions of each components in a query).
      
      Having best time and average time together for more information (we can see kind of variance).
      
      rate, time per row and relative are all calculated using best time.
      
      The result looks like this:
      ```
      Intel(R) Core(TM) i7-4558U CPU  2.80GHz
      rang/filter/sum:                    Best/Avg Time(ms)    Rate(M/s)   Per Row(ns)   Relative
      -------------------------------------------------------------------------------------------
      rang/filter/sum codegen=false          14332 / 16646         36.0          27.8       1.0X
      rang/filter/sum codegen=true              845 /  940        620.0           1.6      17.0X
      ```
      
      Author: Davies Liu <davies@databricks.com>
      
      Closes #11018 from davies/gen_bench.
      de091452
    • Reynold Xin's avatar
      [SPARK-13166][SQL] Remove DataStreamReader/Writer · 915a7539
      Reynold Xin authored
      They seem redundant and we can simply use DataFrameReader/Writer. The new usage looks like:
      
      ```scala
      val df = sqlContext.read.stream("...")
      val handle = df.write.stream("...")
      handle.stop()
      ```
      
      Author: Reynold Xin <rxin@databricks.com>
      
      Closes #11062 from rxin/SPARK-13166.
      915a7539
    • Alex Bozarth's avatar
      [SPARK-3611][WEB UI] Show number of cores for each executor in application web UI · 3221eddb
      Alex Bozarth authored
      Added a Cores column in the Executors UI
      
      Author: Alex Bozarth <ajbozart@us.ibm.com>
      
      Closes #11039 from ajbozarth/spark3611.
      3221eddb
    • Herman van Hovell's avatar
      [SPARK-13157] [SQL] Support any kind of input for SQL commands. · 9dd2741e
      Herman van Hovell authored
      The ```SparkSqlLexer``` currently swallows characters which have not been defined in the grammar. This causes problems with SQL commands, such as: ```add jar file:///tmp/ab/TestUDTF.jar```. In this example the `````` is swallowed.
      
      This PR adds an extra Lexer rule to handle such input, and makes a tiny modification to the ```ASTNode```.
      
      cc davies liancheng
      
      Author: Herman van Hovell <hvanhovell@questtec.nl>
      
      Closes #11052 from hvanhovell/SPARK-13157.
      9dd2741e
    • Davies Liu's avatar
      [SPARK-12798] [SQL] generated BroadcastHashJoin · c4feec26
      Davies Liu authored
      A row from stream side could match multiple rows on build side, the loop for these matched rows should not be interrupted when emitting a row, so we buffer the output rows in a linked list, check the termination condition on producer loop (for example, Range or Aggregate).
      
      Author: Davies Liu <davies@databricks.com>
      
      Closes #10989 from davies/gen_join.
      c4feec26
    • Mario Briggs's avatar
      [SPARK-12739][STREAMING] Details of batch in Streaming tab uses two Duration columns · e9eb248e
      Mario Briggs authored
      I have clearly prefix the two 'Duration' columns in 'Details of Batch' Streaming tab as 'Output Op Duration' and 'Job Duration'
      
      Author: Mario Briggs <mario.briggs@in.ibm.com>
      Author: mariobriggs <mariobriggs@in.ibm.com>
      
      Closes #11022 from mariobriggs/spark-12739.
      e9eb248e
    • Sameer Agarwal's avatar
      [SPARK-12957][SQL] Initial support for constraint propagation in SparkSQL · 138c300f
      Sameer Agarwal authored
      Based on the semantics of a query, we can derive a number of data constraints on output of each (logical or physical) operator. For instance, if a filter defines `‘a > 10`, we know that the output data of this filter satisfies 2 constraints:
      
      1. `‘a > 10`
      2. `isNotNull(‘a)`
      
      This PR proposes a possible way of keeping track of these constraints and propagating them in the logical plan, which can then help us build more advanced optimizations (such as pruning redundant filters, optimizing joins, among others). We define constraints as a set of (implicitly conjunctive) expressions. For e.g., if a filter operator has constraints = `Set(‘a > 10, ‘b < 100)`, it’s implied that the outputs satisfy both individual constraints (i.e., `‘a > 10` AND `‘b < 100`).
      
      Design Document: https://docs.google.com/a/databricks.com/document/d/1WQRgDurUBV9Y6CWOBS75PQIqJwT-6WftVa18xzm7nCo/edit?usp=sharing
      
      Author: Sameer Agarwal <sameer@databricks.com>
      
      Closes #10844 from sameeragarwal/constraints.
      138c300f
    • Davies Liu's avatar
      [SPARK-13147] [SQL] improve readability of generated code · e86f8f63
      Davies Liu authored
      1. try to avoid the suffix (unique id)
      2. remove the comment if there is no code generated.
      3. re-arrange the order of functions
      4. trop the new line for inlined blocks.
      
      Author: Davies Liu <davies@databricks.com>
      
      Closes #11032 from davies/better_suffix.
      e86f8f63
  3. Feb 02, 2016
Loading