Skip to content
Snippets Groups Projects
  1. Jan 06, 2015
    • Sean Owen's avatar
      SPARK-4159 [CORE] Maven build doesn't run JUnit test suites · 4cba6eb4
      Sean Owen authored
      This PR:
      
      - Reenables `surefire`, and copies config from `scalatest` (which is itself an old fork of `surefire`, so similar)
      - Tells `surefire` to test only Java tests
      - Enables `surefire` and `scalatest` for all children, and in turn eliminates some duplication.
      
      For me this causes the Scala and Java tests to be run once each, it seems, as desired. It doesn't affect the SBT build but works for Maven. I still need to verify that all of the Scala tests and Java tests are being run.
      
      Author: Sean Owen <sowen@cloudera.com>
      
      Closes #3651 from srowen/SPARK-4159 and squashes the following commits:
      
      2e8a0af [Sean Owen] Remove specialized SPARK_HOME setting for REPL, YARN tests as it appears to be obsolete
      12e4558 [Sean Owen] Append to unit-test.log instead of overwriting, so that both surefire and scalatest output is preserved. Also standardize/correct comments a bit.
      e6f8601 [Sean Owen] Reenable Java tests by reenabling surefire with config cloned from scalatest; centralize test config in the parent
      4cba6eb4
    • kj-ki's avatar
      [Minor] Fix comments for GraphX 2D partitioning strategy · 5e3ec111
      kj-ki authored
      The sum of vertices on matrix (v0 to v11) is 12. And, I think one same block overlaps in this strategy.
      
      This is minor PR, so I didn't file in JIRA.
      
      Author: kj-ki <kikushima.kenji@lab.ntt.co.jp>
      
      Closes #3904 from kj-ki/fix-partitionstrategy-comments and squashes the following commits:
      
      79829d9 [kj-ki] Fix comments for 2D partitioning.
      5e3ec111
    • Josh Rosen's avatar
      [SPARK-1600] Refactor FileInputStream tests to remove Thread.sleep() calls and SystemClock usage · a6394bc2
      Josh Rosen authored
      This patch refactors Spark Streaming's FileInputStream tests to remove uses of Thread.sleep() and SystemClock, which should hopefully resolve some longstanding flakiness in these tests (see SPARK-1600).
      
      Key changes:
      
      - Modify FileInputDStream to use the scheduler's Clock instead of System.currentTimeMillis(); this allows it to be tested using ManualClock.
      - Fix a synchronization issue in ManualClock's `currentTime` method.
      - Add a StreamingTestWaiter class which allows callers to block until a certain number of batches have finished.
      - Change the FileInputStream tests so that files' modification times are manually set based off of ManualClock; this eliminates many Thread.sleep calls.
      - Update these tests to use the withStreamingContext fixture.
      
      Author: Josh Rosen <joshrosen@databricks.com>
      
      Closes #3801 from JoshRosen/SPARK-1600 and squashes the following commits:
      
      e4494f4 [Josh Rosen] Address a potential race when setting file modification times
      8340bd0 [Josh Rosen] Use set comparisons for output.
      0b9c252 [Josh Rosen] Fix some ManualClock usage problems.
      1cc689f [Josh Rosen] ConcurrentHashMap -> SynchronizedMap
      db26c3a [Josh Rosen] Use standard timeout in ScalaTest `eventually` blocks.
      3939432 [Josh Rosen] Rename StreamingTestWaiter to BatchCounter
      0b9c3a1 [Josh Rosen] Wait for checkpoint to complete
      863d71a [Josh Rosen] Remove Thread.sleep that was used to make task run slowly
      b4442c3 [Josh Rosen] batchTimeToSelectedFiles should be thread-safe
      15b48ee [Josh Rosen] Replace several TestWaiter methods w/ ScalaTest eventually.
      fffc51c [Josh Rosen] Revert "Remove last remaining sleep() call"
      dbb8247 [Josh Rosen] Remove last remaining sleep() call
      566a63f [Josh Rosen] Fix log message and comment typos
      da32f3f [Josh Rosen] Fix log message and comment typos
      3689214 [Josh Rosen] Merge remote-tracking branch 'origin/master' into SPARK-1600
      c8f06b1 [Josh Rosen] Remove Thread.sleep calls in FileInputStream CheckpointSuite test.
      d4f2d87 [Josh Rosen] Refactor file input stream tests to not rely on SystemClock.
      dda1403 [Josh Rosen] Add StreamingTestWaiter class.
      3c3efc3 [Josh Rosen] Synchronize `currentTime` in ManualClock
      a95ddc4 [Josh Rosen] Modify FileInputDStream to use Clock class.
      a6394bc2
    • Kostas Sakellis's avatar
      SPARK-4843 [YARN] Squash ExecutorRunnableUtil and ExecutorRunnable · 451546aa
      Kostas Sakellis authored
      ExecutorRunnableUtil is a parent of ExecutorRunnable because of the yarn-alpha and yarn-stable split. Now that yarn-alpha is gone, this commit squashes the unnecessary hierarchy. The methods from ExecutorRunnableUtil are added as private.
      
      Author: Kostas Sakellis <kostas@cloudera.com>
      
      Closes #3696 from ksakellis/kostas-spark-4843 and squashes the following commits:
      
      486716f [Kostas Sakellis] Moved prepareEnvironment call to after yarnConf declaration
      470e22e [Kostas Sakellis] Fixed indentation and renamed sparkConf variable
      9b1b1c9 [Kostas Sakellis] SPARK-4843 [YARN] Squash ExecutorRunnableUtil and ExecutorRunnable
      451546aa
  2. Jan 05, 2015
    • Reynold Xin's avatar
      [SPARK-5040][SQL] Support expressing unresolved attributes using $"attribute... · 04d55d8e
      Reynold Xin authored
      [SPARK-5040][SQL] Support expressing unresolved attributes using $"attribute name" notation in SQL DSL.
      
      Author: Reynold Xin <rxin@databricks.com>
      
      Closes #3862 from rxin/stringcontext-attr and squashes the following commits:
      
      9b10f57 [Reynold Xin] Rename StrongToAttributeConversionHelper
      72121af [Reynold Xin] [SPARK-5040][SQL] Support expressing unresolved attributes using $"attribute name" notation in SQL DSL.
      04d55d8e
    • Reynold Xin's avatar
      [SPARK-5093] Set spark.network.timeout to 120s consistently. · bbcba3a9
      Reynold Xin authored
      Author: Reynold Xin <rxin@databricks.com>
      
      Closes #3903 from rxin/timeout-120 and squashes the following commits:
      
      7c2138e [Reynold Xin] [SPARK-5093] Set spark.network.timeout to 120s consistently.
      bbcba3a9
    • freeman's avatar
      [SPARK-5089][PYSPARK][MLLIB] Fix vector convert · 6c6f3257
      freeman authored
      This is a small change addressing a potentially significant bug in how PySpark + MLlib handles non-float64 numpy arrays. The automatic conversion to `DenseVector` that occurs when passing RDDs to MLlib algorithms in PySpark should automatically upcast to float64s, but currently this wasn't actually happening. As a result, non-float64 would be silently parsed inappropriately during SerDe, yielding erroneous results when running, for example, KMeans.
      
      The PR includes the fix, as well as a new test for the correct conversion behavior.
      
      davies
      
      Author: freeman <the.freeman.lab@gmail.com>
      
      Closes #3902 from freeman-lab/fix-vector-convert and squashes the following commits:
      
      764db47 [freeman] Add a test for proper conversion behavior
      704f97e [freeman] Return array after changing type
      6c6f3257
    • Jongyoul Lee's avatar
      [SPARK-4465] runAsSparkUser doesn't affect TaskRunner in Mesos environme... · 1c0e7ce0
      Jongyoul Lee authored
      ...nt at all.
      
      - fixed a scope of runAsSparkUser from MesosExecutorDriver.run to MesosExecutorBackend.launchTask
      - See the Jira Issue for more details.
      
      Author: Jongyoul Lee <jongyoul@gmail.com>
      
      Closes #3741 from jongyoul/SPARK-4465 and squashes the following commits:
      
      46ad71e [Jongyoul Lee] [SPARK-4465] runAsSparkUser doesn't affect TaskRunner in Mesos environment at all. - Removed unused import
      3d6631f [Jongyoul Lee] [SPARK-4465] runAsSparkUser doesn't affect TaskRunner in Mesos environment at all. - Removed comments and adjusted indentations
      2343f13 [Jongyoul Lee] [SPARK-4465] runAsSparkUser doesn't affect TaskRunner in Mesos environment at all. - fixed a scope of runAsSparkUser from MesosExecutorDriver.run to MesosExecutorBackend.launchTask
      1c0e7ce0
    • WangTao's avatar
      [SPARK-5057] Log message in failed askWithReply attempts · ce39b344
      WangTao authored
      https://issues.apache.org/jira/browse/SPARK-5057
      
      Author: WangTao <barneystinson@aliyun.com>
      Author: WangTaoTheTonic <barneystinson@aliyun.com>
      
      Closes #3875 from WangTaoTheTonic/SPARK-5057 and squashes the following commits:
      
      1503487 [WangTao] use string interpolation
      706c8a7 [WangTaoTheTonic] log more messages
      ce39b344
    • Varun Saxena's avatar
      [SPARK-4688] Have a single shared network timeout in Spark · d3f07fd2
      Varun Saxena authored
      [SPARK-4688] Have a single shared network timeout in Spark
      
      Author: Varun Saxena <vsaxena.varun@gmail.com>
      Author: varunsaxena <vsaxena.varun@gmail.com>
      
      Closes #3562 from varunsaxena/SPARK-4688 and squashes the following commits:
      
      6e97f72 [Varun Saxena] [SPARK-4688] Single shared network timeout
      cd783a2 [Varun Saxena] SPARK-4688
      d6f8c29 [Varun Saxena] SCALA-4688
      9562b15 [Varun Saxena] SPARK-4688
      a75f014 [varunsaxena] SPARK-4688
      594226c [varunsaxena] SPARK-4688
      d3f07fd2
  3. Jan 04, 2015
    • zsxwing's avatar
      [SPARK-5074][Core] Fix a non-deterministic test failure · 5c506cec
      zsxwing authored
      Add `assert(sc.listenerBus.waitUntilEmpty(WAIT_TIMEOUT_MILLIS))` to make sure `sparkListener` receive the message.
      
      Author: zsxwing <zsxwing@gmail.com>
      
      Closes #3889 from zsxwing/SPARK-5074 and squashes the following commits:
      
      e61c198 [zsxwing] Fix a non-deterministic test failure
      5c506cec
    • zsxwing's avatar
      [SPARK-5083][Core] Fix a flaky test in TaskResultGetterSuite · 27e7f5a7
      zsxwing authored
      Because `sparkEnv.blockManager.master.removeBlock` is asynchronous, we need to make sure the block has already been removed before calling `super.enqueueSuccessfulTask`.
      
      Author: zsxwing <zsxwing@gmail.com>
      
      Closes #3894 from zsxwing/SPARK-5083 and squashes the following commits:
      
      d97c03d [zsxwing] Fix a flaky test in TaskResultGetterSuite
      27e7f5a7
    • zsxwing's avatar
      [SPARK-5069][Core] Fix the race condition of TaskSchedulerImpl.dagScheduler · 6c726a3f
      zsxwing authored
      It's not necessary to set `TaskSchedulerImpl.dagScheduler` in preStart. It's safe to set it after `initializeEventProcessActor()`.
      
      Author: zsxwing <zsxwing@gmail.com>
      
      Closes #3887 from zsxwing/SPARK-5069 and squashes the following commits:
      
      d95894f [zsxwing] Fix the race condition of TaskSchedulerImpl.dagScheduler
      6c726a3f
    • zsxwing's avatar
      [SPARK-5067][Core] Use '===' to compare well-defined case class · 72396522
      zsxwing authored
      A simple fix would be adding `assert(e1.appId == e2.appId)` for `SparkListenerApplicationStart`. But actually we can use `===` for well-defined case class directly. Therefore, instead of fixing this issue, I use `===` to compare those well-defined case classes (all fields have implemented a correct `equals` method, such as primitive types)
      
      Author: zsxwing <zsxwing@gmail.com>
      
      Closes #3886 from zsxwing/SPARK-5067 and squashes the following commits:
      
      0a51711 [zsxwing] Use '===' to compare well-defined case class
      72396522
    • Josh Rosen's avatar
      [SPARK-4835] Disable validateOutputSpecs for Spark Streaming jobs · 939ba1f8
      Josh Rosen authored
      This patch disables output spec. validation for jobs launched through Spark Streaming, since this interferes with checkpoint recovery.
      
      Hadoop OutputFormats have a `checkOutputSpecs` method which performs certain checks prior to writing output, such as checking whether the output directory already exists.  SPARK-1100 added checks for FileOutputFormat, SPARK-1677 (#947) added a SparkConf configuration to disable these checks, and SPARK-2309 (#1088) extended these checks to run for all OutputFormats, not just FileOutputFormat.
      
      In Spark Streaming, we might have to re-process a batch during checkpoint recovery, so `save` actions may be called multiple times.  In addition to `DStream`'s own save actions, users might use `transform` or `foreachRDD` and call the `RDD` and `PairRDD` save actions.  When output spec. validation is enabled, the second calls to these actions will fail due to existing output.
      
      This patch automatically disables output spec. validation for jobs submitted by the Spark Streaming scheduler.  This is done by using Scala's `DynamicVariable` to propagate the bypass setting without having to mutate SparkConf or introduce a global variable.
      
      Author: Josh Rosen <joshrosen@databricks.com>
      
      Closes #3832 from JoshRosen/SPARK-4835 and squashes the following commits:
      
      36eaf35 [Josh Rosen] Add comment explaining use of transform() in test.
      6485cf8 [Josh Rosen] Add test case in Streaming; fix bug for transform()
      7b3e06a [Josh Rosen] Remove Streaming-specific setting to undo this change; update conf. guide
      bf9094d [Josh Rosen] Revise disableOutputSpecValidation() comment to not refer to Spark Streaming.
      e581d17 [Josh Rosen] Deduplicate isOutputSpecValidationEnabled logic.
      762e473 [Josh Rosen] [SPARK-4835] Disable validateOutputSpecs for Spark Streaming jobs.
      939ba1f8
    • bilna's avatar
      [SPARK-4631] unit test for MQTT · e767d7dd
      bilna authored
      Please review the unit test for MQTT
      
      Author: bilna <bilnap@am.amrita.edu>
      Author: Bilna P <bilna.p@gmail.com>
      
      Closes #3844 from Bilna/master and squashes the following commits:
      
      acea3a3 [bilna] Adding dependency with scope test
      28681fa [bilna] Merge remote-tracking branch 'upstream/master'
      fac3904 [bilna] Correction in Indentation and coding style
      ed9db4c [bilna] Merge remote-tracking branch 'upstream/master'
      4b34ee7 [Bilna P] Update MQTTStreamSuite.scala
      04503cf [bilna] Added embedded broker service for mqtt test
      89d804e [bilna] Merge remote-tracking branch 'upstream/master'
      fc8eb28 [bilna] Merge remote-tracking branch 'upstream/master'
      4b58094 [Bilna P] Update MQTTStreamSuite.scala
      b1ac4ad [bilna] Added BeforeAndAfter
      5f6bfd2 [bilna] Added BeforeAndAfter
      e8b6623 [Bilna P] Update MQTTStreamSuite.scala
      5ca6691 [Bilna P] Update MQTTStreamSuite.scala
      8616495 [bilna] [SPARK-4631] unit test for MQTT
      e767d7dd
    • Dale's avatar
      [SPARK-4787] Stop SparkContext if a DAGScheduler init error occurs · 3fddc946
      Dale authored
      Author: Dale <tigerquoll@outlook.com>
      
      Closes #3809 from tigerquoll/SPARK-4787 and squashes the following commits:
      
      5661e01 [Dale] [SPARK-4787] Ensure that call to stop() doesn't lose the exception by using a finally block.
      2172578 [Dale] [SPARK-4787] Stop context properly if an exception occurs during DAGScheduler initialization.
      3fddc946
    • Brennon York's avatar
      [SPARK-794][Core] Remove sleep() in ClusterScheduler.stop · b96008d5
      Brennon York authored
      Removed `sleep()` from the `stop()` method of the `TaskSchedulerImpl` class which, from the JIRA ticket, is believed to be a legacy artifact slowing down testing originally introduced in the `ClusterScheduler` class.
      
      Author: Brennon York <brennon.york@capitalone.com>
      
      Closes #3851 from brennonyork/SPARK-794 and squashes the following commits:
      
      04c3e64 [Brennon York] Removed sleep() from the stop() method
      b96008d5
  4. Jan 03, 2015
    • sigmoidanalytics's avatar
      [SPARK-5058] Updated broken links · 342612b6
      sigmoidanalytics authored
      Updated the broken link pointing to the KafkaWordCount example to the correct one.
      
      Author: sigmoidanalytics <mayur@sigmoidanalytics.com>
      
      Closes #3877 from sigmoidanalytics/patch-1 and squashes the following commits:
      
      3e19b31 [sigmoidanalytics] Updated broken links
      342612b6
  5. Jan 02, 2015
    • Akhil Das's avatar
      Fixed typos in streaming-kafka-integration.md · cdccc263
      Akhil Das authored
      Changed projrect to project :)
      
      Author: Akhil Das <akhld@darktech.ca>
      
      Closes #3876 from akhld/patch-1 and squashes the following commits:
      
      e0cf9ef [Akhil Das] Fixed typos in streaming-kafka-integration.md
      cdccc263
    • Yadong Qi's avatar
      [SPARK-3325][Streaming] Add a parameter to the method print in class DStream · bd88b718
      Yadong Qi authored
      This PR is a fixed version of the original PR #3237 by watermen and scwf.
      This adds the ability to specify how many elements to print in `DStream.print`.
      
      Author: Yadong Qi <qiyadong2010@gmail.com>
      Author: q00251598 <qiyadong@huawei.com>
      Author: Tathagata Das <tathagata.das1565@gmail.com>
      Author: wangfei <wangfei1@huawei.com>
      
      Closes #3865 from tdas/print-num and squashes the following commits:
      
      cd34e9e [Tathagata Das] Fix bug
      7c09f16 [Tathagata Das] Merge remote-tracking branch 'apache-github/master' into HEAD
      bb35d1a [Yadong Qi] Update MimaExcludes.scala
      f8098ca [Yadong Qi] Update MimaExcludes.scala
      f6ac3cb [Yadong Qi] Update MimaExcludes.scala
      e4ed897 [Yadong Qi] Update MimaExcludes.scala
      3b9d5cf [wangfei] fix conflicts
      ec8a3af [q00251598] move to  Spark 1.3
      26a70c0 [q00251598] extend the Python DStream's print
      b589a4b [q00251598] add another print function
      bd88b718
  6. Jan 01, 2015
    • Josh Rosen's avatar
      [HOTFIX] Bind web UI to ephemeral port in DriverSuite · 01283980
      Josh Rosen authored
      The job launched by DriverSuite should bind the web UI to an ephemeral port, since it looks like port contention in this test has caused a large number of Jenkins failures when many builds are started simultaneously.  Our tests already disable the web UI, but this doesn't affect subprocesses launched by our tests.  In this case, I've opted to bind to an ephemeral port instead of disabling the UI because disabling features in this test may mask its ability to catch certain bugs.
      
      See also: e24d3a9a
      
      Author: Josh Rosen <joshrosen@databricks.com>
      
      Closes #3873 from JoshRosen/driversuite-webui-port and squashes the following commits:
      
      48cd05c [Josh Rosen] [HOTFIX] Bind web UI to ephemeral port in DriverSuite.
      01283980
  7. Dec 31, 2014
    • Reynold Xin's avatar
      [SPARK-5038] Add explicit return type for implicit functions. · 7749dd6c
      Reynold Xin authored
      As we learned in #3580, not explicitly typing implicit functions can lead to compiler bugs and potentially unexpected runtime behavior.
      
      This is a follow up PR for rest of Spark (outside Spark SQL). The original PR for Spark SQL can be found at https://github.com/apache/spark/pull/3859
      
      Author: Reynold Xin <rxin@databricks.com>
      
      Closes #3860 from rxin/implicit and squashes the following commits:
      
      73702f9 [Reynold Xin] [SPARK-5038] Add explicit return type for implicit functions.
      7749dd6c
    • Sean Owen's avatar
      SPARK-2757 [BUILD] [STREAMING] Add Mima test for Spark Sink after 1.10 is released · 4bb12488
      Sean Owen authored
      Re-enable MiMa for Streaming Flume Sink module, now that 1.1.0 is released, per the JIRA TO-DO. That's pretty much all there is to this.
      
      Author: Sean Owen <sowen@cloudera.com>
      
      Closes #3842 from srowen/SPARK-2757 and squashes the following commits:
      
      50ff80e [Sean Owen] Exclude apparent false positive turned up by re-enabling MiMa checks for Streaming Flume Sink
      0e5ba5c [Sean Owen] Re-enable MiMa for Streaming Flume Sink module
      4bb12488
    • Josh Rosen's avatar
      [SPARK-5035] [Streaming] ReceiverMessage trait should extend Serializable · fe6efacc
      Josh Rosen authored
      Spark Streaming's ReceiverMessage trait should extend Serializable in order to fix a subtle bug that only occurs when running on a real cluster:
      
      If you attempt to send a fire-and-forget message to a remote Akka actor and that message cannot be serialized, then this seems to lead to more-or-less silent failures. As an optimization, Akka skips message serialization for messages sent within the same JVM. As a result, Spark's unit tests will never fail due to non-serializable Akka messages, but these will cause mostly-silent failures when running on a real cluster.
      
      Before this patch, here was the code for ReceiverMessage:
      
      ```
      /** Messages sent to the NetworkReceiver. */
      private[streaming] sealed trait ReceiverMessage
      private[streaming] object StopReceiver extends ReceiverMessage
      ```
      
      Since ReceiverMessage does not extend Serializable and StopReceiver is a regular `object`, not a `case object`, StopReceiver will throw serialization errors. As a result, graceful receiver shutdown is broken on real clusters (and local-cluster mode) but works in local modes. If you want to reproduce this, try running the word count example from the Streaming Programming Guide in the Spark shell:
      
      ```
      import org.apache.spark._
      import org.apache.spark.streaming._
      import org.apache.spark.streaming.StreamingContext._
      val ssc = new StreamingContext(sc, Seconds(10))
      // Create a DStream that will connect to hostname:port, like localhost:9999
      val lines = ssc.socketTextStream("localhost", 9999)
      // Split each line into words
      val words = lines.flatMap(_.split(" "))
      import org.apache.spark.streaming.StreamingContext._
      // Count each word in each batch
      val pairs = words.map(word => (word, 1))
      val wordCounts = pairs.reduceByKey(_ + _)
      // Print the first ten elements of each RDD generated in this DStream to the console
      wordCounts.print()
      ssc.start()
      Thread.sleep(10000)
      ssc.stop(true, true)
      ```
      
      Prior to this patch, this would work correctly in local mode but fail when running against a real cluster (it would report that some receivers were not shut down).
      
      Author: Josh Rosen <joshrosen@databricks.com>
      
      Closes #3857 from JoshRosen/SPARK-5035 and squashes the following commits:
      
      71d0eae [Josh Rosen] [SPARK-5035] ReceiverMessage trait should extend Serializable.
      fe6efacc
    • Travis Galoppo's avatar
      SPARK-5020 [MLlib] GaussianMixtureModel.predictMembership() should take an RDD only · c4f0b4f3
      Travis Galoppo authored
      Removed unnecessary parameters to predictMembership()
      
      CC: jkbradley
      
      Author: Travis Galoppo <tjg2107@columbia.edu>
      
      Closes #3854 from tgaloppo/spark-5020 and squashes the following commits:
      
      1bf4669 [Travis Galoppo] renamed predictMembership() to predictSoft()
      0f1d96e [Travis Galoppo] SPARK-5020 - Removed superfluous parameters from predictMembership()
      c4f0b4f3
    • jerryshao's avatar
      [SPARK-5028][Streaming]Add total received and processed records metrics to Streaming UI · fdc2aa49
      jerryshao authored
      This is a follow-up work of [SPARK-4537](https://issues.apache.org/jira/browse/SPARK-4537). Adding total received records and processed records metrics back to UI.
      
      ![screenshot](https://dl.dropboxusercontent.com/u/19230832/screenshot.png)
      
      Author: jerryshao <saisai.shao@intel.com>
      
      Closes #3852 from jerryshao/SPARK-5028 and squashes the following commits:
      
      c8c4877 [jerryshao] Add total received and processed metrics to Streaming UI
      fdc2aa49
    • Hari Shreedharan's avatar
      [SPARK-4790][STREAMING] Fix ReceivedBlockTrackerSuite waits for old file... · 3610d3c6
      Hari Shreedharan authored
      ...s to get deleted before continuing.
      
      Since the deletes are happening asynchronously, the getFileStatus call might throw an exception in older HDFS
      versions, if the delete happens between the time listFiles is called on the directory and getFileStatus is called
      on the file in the getFileStatus method.
      
      This PR addresses this by adding an option to delete the files synchronously and then waiting for the deletion to
      complete before proceeding.
      
      Author: Hari Shreedharan <hshreedharan@apache.org>
      
      Closes #3726 from harishreedharan/spark-4790 and squashes the following commits:
      
      bbbacd1 [Hari Shreedharan] Call cleanUpOldLogs only once in the tests.
      3255f17 [Hari Shreedharan] Add test for async deletion. Remove method from ReceiverTracker that does not take waitForCompletion.
      e4c83ec [Hari Shreedharan] Making waitForCompletion a mandatory param. Remove eventually from WALSuite since the cleanup method returns only after all files are deleted.
      af00fd1 [Hari Shreedharan] [SPARK-4790][STREAMING] Fix ReceivedBlockTrackerSuite waits for old files to get deleted before continuing.
      3610d3c6
    • Reynold Xin's avatar
      [SPARK-5038][SQL] Add explicit return type for implicit functions in Spark SQL · c88a3d7f
      Reynold Xin authored
      As we learned in https://github.com/apache/spark/pull/3580, not explicitly typing implicit functions can lead to compiler bugs and potentially unexpected runtime behavior.
      
      Author: Reynold Xin <rxin@databricks.com>
      
      Closes #3859 from rxin/sql-implicits and squashes the following commits:
      
      30c2c24 [Reynold Xin] [SPARK-5038] Add explicit return type for implicit functions in Spark SQL.
      c88a3d7f
    • Josh Rosen's avatar
      [HOTFIX] Disable Spark UI in SparkSubmitSuite tests · e24d3a9a
      Josh Rosen authored
      This should fix a major cause of build breaks when running many parallel tests.
      e24d3a9a
    • Sean Owen's avatar
      SPARK-4547 [MLLIB] OOM when making bins in BinaryClassificationMetrics · 3d194cc7
      Sean Owen authored
      Now that I've implemented the basics here, I'm less convinced there is a need for this change, somehow. Callers can downsample before or after. Really the OOM is not in the ROC curve code, but in code that might `collect()` it for local analysis. Still, might be useful to down-sample since the ROC curve probably never needs millions of points.
      
      This is a first pass. Since the `(score,label)` are already grouped and sorted, I think it's sufficient to just take every Nth such pair, in order to downsample by a factor of N? this is just like retaining every Nth point on the curve, which I think is the goal. All of the data is still used to build the curve of course.
      
      What do you think about the API, and usefulness?
      
      Author: Sean Owen <sowen@cloudera.com>
      
      Closes #3702 from srowen/SPARK-4547 and squashes the following commits:
      
      1d34d05 [Sean Owen] Indent and reorganize numBins scaladoc
      692d825 [Sean Owen] Change handling of large numBins, make 2nd consturctor instead of optional param, style change
      a03610e [Sean Owen] Add downsamplingFactor to BinaryClassificationMetrics
      3d194cc7
    • Brennon York's avatar
      [SPARK-4298][Core] - The spark-submit cannot read Main-Class from Manifest. · 8e14c5eb
      Brennon York authored
      Resolves a bug where the `Main-Class` from a .jar file wasn't being read in properly. This was caused by the fact that the `primaryResource` object was a URI and needed to be normalized through a call to `.getPath` before it could be passed into the `JarFile` object.
      
      Author: Brennon York <brennon.york@capitalone.com>
      
      Closes #3561 from brennonyork/SPARK-4298 and squashes the following commits:
      
      5e0fce1 [Brennon York] Use string interpolation for error messages, moved comment line from original code to above its necessary code segment
      14daa20 [Brennon York] pushed mainClass assignment into match statement, removed spurious spaces, removed { } from case statements, removed return values
      c6dad68 [Brennon York] Set case statement to support multiple jar URI's and enabled the 'file' URI to load the main-class
      8d20936 [Brennon York] updated to reset the error message back to the default
      a043039 [Brennon York] updated to split the uri and jar vals
      8da7cbf [Brennon York] fixes SPARK-4298
      8e14c5eb
    • Liang-Chi Hsieh's avatar
      [SPARK-4797] Replace breezeSquaredDistance · 06a9aa58
      Liang-Chi Hsieh authored
      This PR replaces slow breezeSquaredDistance.
      
      Author: Liang-Chi Hsieh <viirya@gmail.com>
      
      Closes #3643 from viirya/faster_squareddistance and squashes the following commits:
      
      f28b275 [Liang-Chi Hsieh] Move the implementation to linalg.Vectors and rename as sqdist.
      0bc48ee [Liang-Chi Hsieh] Merge branch 'master' into faster_squareddistance
      ba34422 [Liang-Chi Hsieh] Fix bug.
      91849d0 [Liang-Chi Hsieh] Modified for comment.
      44a65ad [Liang-Chi Hsieh] Modified for comments.
      35db395 [Liang-Chi Hsieh] Fix bug and some modifications for comments.
      f4f5ebb [Liang-Chi Hsieh] Follow BLAS.dot pattern to replace intersect, diff with while-loop.
      a36e09f [Liang-Chi Hsieh] Use while-loop to replace foreach for better performance.
      d3e0628 [Liang-Chi Hsieh] Make the methods private.
      dd415bc [Liang-Chi Hsieh] Consider different cases of SparseVector and DenseVector.
      13669db [Liang-Chi Hsieh] Replace breezeSquaredDistance.
      06a9aa58
  8. Dec 30, 2014
    • Josh Rosen's avatar
      [SPARK-1010] Clean up uses of System.setProperty in unit tests · 352ed6bb
      Josh Rosen authored
      Several of our tests call System.setProperty (or test code which implicitly sets system properties) and don't always reset/clear the modified properties, which can create ordering dependencies between tests and cause hard-to-diagnose failures.
      
      This patch removes most uses of System.setProperty from our tests, since in most cases we can use SparkConf to set these configurations (there are a few exceptions, including the tests of SparkConf itself).
      
      For the cases where we continue to use System.setProperty, this patch introduces a `ResetSystemProperties` ScalaTest mixin class which snapshots the system properties before individual tests and to automatically restores them on test completion / failure.  See the block comment at the top of the ResetSystemProperties class for more details.
      
      Author: Josh Rosen <joshrosen@databricks.com>
      
      Closes #3739 from JoshRosen/cleanup-system-properties-in-tests and squashes the following commits:
      
      0236d66 [Josh Rosen] Replace setProperty uses in two example programs / tools
      3888fe3 [Josh Rosen] Remove setProperty use in LocalJavaStreamingContext
      4f4031d [Josh Rosen] Add note on why SparkSubmitSuite needs ResetSystemProperties
      4742a5b [Josh Rosen] Clarify ResetSystemProperties trait inheritance ordering.
      0eaf0b6 [Josh Rosen] Remove setProperty call in TaskResultGetterSuite.
      7a3d224 [Josh Rosen] Fix trait ordering
      3fdb554 [Josh Rosen] Remove setProperty call in TaskSchedulerImplSuite
      bee20df [Josh Rosen] Remove setProperty calls in SparkContextSchedulerCreationSuite
      655587c [Josh Rosen] Remove setProperty calls in JobCancellationSuite
      3f2f955 [Josh Rosen] Remove System.setProperty calls in DistributedSuite
      cfe9cce [Josh Rosen] Remove use of system properties in SparkContextSuite
      8783ab0 [Josh Rosen] Remove TestUtils.setSystemProperty, since it is subsumed by the ResetSystemProperties trait.
      633a84a [Josh Rosen] Remove use of system properties in FileServerSuite
      25bfce2 [Josh Rosen] Use ResetSystemProperties in UtilsSuite
      1d1aa5a [Josh Rosen] Use ResetSystemProperties in SizeEstimatorSuite
      dd9492b [Josh Rosen] Use ResetSystemProperties in AkkaUtilsSuite
      b0daff2 [Josh Rosen] Use ResetSystemProperties in BlockManagerSuite
      e9ded62 [Josh Rosen] Use ResetSystemProperties in TaskSchedulerImplSuite
      5b3cb54 [Josh Rosen] Use ResetSystemProperties in SparkListenerSuite
      0995c4b [Josh Rosen] Use ResetSystemProperties in SparkContextSchedulerCreationSuite
      c83ded8 [Josh Rosen] Use ResetSystemProperties in SparkConfSuite
      51aa870 [Josh Rosen] Use withSystemProperty in ShuffleSuite
      60a63a1 [Josh Rosen] Use ResetSystemProperties in JobCancellationSuite
      14a92e4 [Josh Rosen] Use withSystemProperty in FileServerSuite
      628f46c [Josh Rosen] Use ResetSystemProperties in DistributedSuite
      9e3e0dd [Josh Rosen] Add ResetSystemProperties test fixture mixin; use it in SparkSubmitSuite.
      4dcea38 [Josh Rosen] Move withSystemProperty to TestUtils class.
      352ed6bb
    • Liu Jiongzhou's avatar
      [SPARK-4998][MLlib]delete the "train" function · 035bac88
      Liu Jiongzhou authored
      To make the functions with the same in "object" effective, specially when using java reflection.
      As the "train" function defined in "class DecisionTree" will hide the functions with the same name in "object DecisionTree".
      
      JIRA[SPARK-4998]
      
      Author: Liu Jiongzhou <ljzzju@163.com>
      
      Closes #3836 from ljzzju/master and squashes the following commits:
      
      4e13133 [Liu Jiongzhou] [MLlib]delete the "train" function
      035bac88
    • zsxwing's avatar
      [SPARK-4813][Streaming] Fix the issue that ContextWaiter didn't handle 'spurious wakeup' · 6a897829
      zsxwing authored
      Used `Condition` to rewrite `ContextWaiter` because it provides a convenient API `awaitNanos` for timeout.
      
      Author: zsxwing <zsxwing@gmail.com>
      
      Closes #3661 from zsxwing/SPARK-4813 and squashes the following commits:
      
      52247f5 [zsxwing] Add explicit unit type
      be42bcf [zsxwing] Update as per review suggestion
      e06bd4f [zsxwing] Fix the issue that ContextWaiter didn't handle 'spurious wakeup'
      6a897829
    • Jakub Dubovsky's avatar
      [Spark-4995] Replace Vector.toBreeze.activeIterator with foreachActive · 0f31992c
      Jakub Dubovsky authored
      New foreachActive method of vector was introduced by SPARK-4431 as more efficient alternative to vector.toBreeze.activeIterator. There are some parts of codebase where it was not yet replaced.
      
      dbtsai
      
      Author: Jakub Dubovsky <dubovsky@avast.com>
      
      Closes #3846 from james64/SPARK-4995-foreachActive and squashes the following commits:
      
      3eb7e37 [Jakub Dubovsky] Scalastyle fix
      32fe6c6 [Jakub Dubovsky] activeIterator removed - IndexedRowMatrix.toBreeze
      47a4777 [Jakub Dubovsky] activeIterator removed in RowMatrix.toBreeze
      90a7d98 [Jakub Dubovsky] activeIterator removed in MLUtils.saveAsLibSVMFile
      0f31992c
    • Sean Owen's avatar
      SPARK-3955 part 2 [CORE] [HOTFIX] Different versions between... · b239ea1c
      Sean Owen authored
      SPARK-3955 part 2 [CORE] [HOTFIX] Different versions between jackson-mapper-asl and jackson-core-asl
      
      pwendell https://github.com/apache/spark/commit/2483c1efb6429a7d8a20c96d18ce2fec93a1aff9 didn't actually add a reference to `jackson-core-asl` as intended, but a second redundant reference to `jackson-mapper-asl`, as markhamstra picked up on (https://github.com/apache/spark/pull/3716#issuecomment-68180192)  This just rectifies the typo. I missed it as well; the original PR https://github.com/apache/spark/pull/2818 had it correct and I also didn't see the problem.
      
      Author: Sean Owen <sowen@cloudera.com>
      
      Closes #3829 from srowen/SPARK-3955 and squashes the following commits:
      
      6cfdc4e [Sean Owen] Actually refer to jackson-core-asl
      b239ea1c
    • wangxiaojing's avatar
      [SPARK-4570][SQL]add BroadcastLeftSemiJoinHash · 07fa1910
      wangxiaojing authored
      JIRA issue: [SPARK-4570](https://issues.apache.org/jira/browse/SPARK-4570)
      We are planning to create a `BroadcastLeftSemiJoinHash` to implement the broadcast join for `left semijoin`
      In left semijoin :
      If the size of data from right side is smaller than the user-settable threshold `AUTO_BROADCASTJOIN_THRESHOLD`,
      the planner would mark it as the `broadcast` relation and mark the other relation as the stream side. The broadcast table will be broadcasted to all of the executors involved in the join, as a `org.apache.spark.broadcast.Broadcast` object. It will use `joins.BroadcastLeftSemiJoinHash`.,else it will use `joins.LeftSemiJoinHash`.
      
      The benchmark suggests these  made the optimized version 4x faster  when `left semijoin`
      <pre><code>
      Original:
      left semi join : 9288 ms
      Optimized:
      left semi join : 1963 ms
      </code></pre>
      The micro benchmark load `data1/kv3.txt` into a normal Hive table.
      Benchmark code:
      <pre><code>
       def benchmark(f: => Unit) = {
          val begin = System.currentTimeMillis()
          f
          val end = System.currentTimeMillis()
          end - begin
        }
        val sc = new SparkContext(
          new SparkConf()
            .setMaster("local")
            .setAppName(getClass.getSimpleName.stripSuffix("$")))
        val hiveContext = new HiveContext(sc)
        import hiveContext._
        sql("drop table if exists left_table")
        sql("drop table if exists right_table")
        sql( """create table left_table (key int, value string)
             """.stripMargin)
        sql( s"""load data local inpath "/data1/kv3.txt" into table left_table""")
        sql( """create table right_table (key int, value string)
             """.stripMargin)
        sql(
          """
            |from left_table
            |insert overwrite table right_table
            |select left_table.key, left_table.value
          """.stripMargin)
      
        val leftSimeJoin = sql(
          """select a.key from left_table a
            |left semi join right_table b on a.key = b.key""".stripMargin)
        val leftSemiJoinDuration = benchmark(leftSimeJoin.count())
        println(s"left semi join : $leftSemiJoinDuration ms ")
      </code></pre>
      
      Author: wangxiaojing <u9jing@gmail.com>
      
      Closes #3442 from wangxiaojing/SPARK-4570 and squashes the following commits:
      
      a4a43c9 [wangxiaojing] rebase
      f103983 [wangxiaojing] change style
      fbe4887 [wangxiaojing] change style
      ff2e618 [wangxiaojing] add testsuite
      1a8da2a [wangxiaojing] add BroadcastLeftSemiJoinHash
      07fa1910
    • wangfei's avatar
      [SPARK-4935][SQL] When hive.cli.print.header configured, spark-sql aborted if... · 8f29b7ca
      wangfei authored
      [SPARK-4935][SQL] When hive.cli.print.header configured, spark-sql aborted if passed in a invalid sql
      
      If we passed in a wrong sql like ```abdcdfsfs```, the spark-sql script aborted.
      
      Author: wangfei <wangfei1@huawei.com>
      Author: Fei Wang <wangfei1@huawei.com>
      
      Closes #3761 from scwf/patch-10 and squashes the following commits:
      
      46dc344 [Fei Wang] revert console.printError(rc.getErrorMessage())
      0330e07 [wangfei] avoid to print error message repeatedly
      1614a11 [wangfei] spark-sql abort when passed in a wrong sql
      8f29b7ca
Loading