Skip to content
Snippets Groups Projects
  1. Sep 17, 2015
  2. Sep 16, 2015
  3. Sep 15, 2015
    • Joseph K. Bradley's avatar
      [SPARK-10595] [ML] [MLLIB] [DOCS] Various ML guide cleanups · b921fe4d
      Joseph K. Bradley authored
      Various ML guide cleanups.
      
      * ml-guide.md: Make it easier to access the algorithm-specific guides.
      * LDA user guide: EM often begins with useless topics, but running longer generally improves them dramatically.  E.g., 10 iterations on a Wikipedia dataset produces useless topics, but 50 iterations produces very meaningful topics.
      * mllib-feature-extraction.html#elementwiseproduct: “w” parameter should be “scalingVec”
      * Clean up Binarizer user guide a little.
      * Document in Pipeline that users should not put an instance into the Pipeline in more than 1 place.
      * spark.ml Word2Vec user guide: clean up grammar/writing
      * Chi Sq Feature Selector docs: Improve text in doc.
      
      CC: mengxr feynmanliang
      
      Author: Joseph K. Bradley <joseph@databricks.com>
      
      Closes #8752 from jkbradley/mlguide-fixes-1.5.
      b921fe4d
    • sureshthalamati's avatar
      [SPARK-9078] [SQL] Allow jdbc dialects to override the query used to check the table. · 64c29afc
      sureshthalamati authored
      Current implementation uses query with a LIMIT clause to find if table already exists. This syntax works only in some database systems. This patch changes the default query to the one that is likely to work on most databases, and adds a new method to the  JdbcDialect abstract class to allow  dialects to override the default query.
      
      I looked at using the JDBC meta data calls, it turns out there is no common way to find the current schema, catalog..etc.  There is a new method Connection.getSchema() , but that is available only starting jdk1.7 , and existing jdbc drivers may not have implemented it.  Other option was to use jdbc escape syntax clause for LIMIT, not sure on how well this supported in all the databases also. After looking at all the jdbc metadata options my conclusion was most common way is to use the simple select query with 'where 1 =0' , and allow dialects to customize as needed
      
      Author: sureshthalamati <suresh.thalamati@gmail.com>
      
      Closes #8676 from sureshthalamati/table_exists_spark-9078.
      64c29afc
    • Andrew Or's avatar
      [SPARK-10613] [SPARK-10624] [SQL] Reduce LocalNode tests dependency on SQLContext · 35a19f33
      Andrew Or authored
      Instead of relying on `DataFrames` to verify our answers, we can just use simple arrays. This significantly simplifies the test logic for `LocalNode`s and reduces a lot of code duplicated from `SparkPlanTest`.
      
      This also fixes an additional issue [SPARK-10624](https://issues.apache.org/jira/browse/SPARK-10624) where the output of `TakeOrderedAndProjectNode` is not actually ordered.
      
      Author: Andrew Or <andrew@databricks.com>
      
      Closes #8764 from andrewor14/sql-local-tests-cleanup.
      35a19f33
    • Josh Rosen's avatar
      [SPARK-10381] Fix mixup of taskAttemptNumber & attemptId in OutputCommitCoordinator · 38700ea4
      Josh Rosen authored
      When speculative execution is enabled, consider a scenario where the authorized committer of a particular output partition fails during the OutputCommitter.commitTask() call. In this case, the OutputCommitCoordinator is supposed to release that committer's exclusive lock on committing once that task fails. However, due to a unit mismatch (we used task attempt number in one place and task attempt id in another) the lock will not be released, causing Spark to go into an infinite retry loop.
      
      This bug was masked by the fact that the OutputCommitCoordinator does not have enough end-to-end tests (the current tests use many mocks). Other factors contributing to this bug are the fact that we have many similarly-named identifiers that have different semantics but the same data types (e.g. attemptNumber and taskAttemptId, with inconsistent variable naming which makes them difficult to distinguish).
      
      This patch adds a regression test and fixes this bug by always using task attempt numbers throughout this code.
      
      Author: Josh Rosen <joshrosen@databricks.com>
      
      Closes #8544 from JoshRosen/SPARK-10381.
      38700ea4
    • vinodkc's avatar
      [SPARK-10575] [SPARK CORE] Wrapped RDD.takeSample with Scope · 99ecfa59
      vinodkc authored
      Remove return statements in RDD.takeSample and wrap it withScope
      
      Author: vinodkc <vinod.kc.in@gmail.com>
      Author: vinodkc <vinodkc@users.noreply.github.com>
      Author: Vinod K C <vinod.kc@huawei.com>
      
      Closes #8730 from vinodkc/fix_takesample_return.
      99ecfa59
    • Reynold Xin's avatar
      [SPARK-10612] [SQL] Add prepare to LocalNode. · a63cdc76
      Reynold Xin authored
      The idea is that we should separate the function call that does memory reservation (i.e. prepare) from the function call that consumes the input (e.g. open()), so all operators can be a chance to reserve memory before they are all consumed.
      
      Author: Reynold Xin <rxin@databricks.com>
      
      Closes #8761 from rxin/SPARK-10612.
      a63cdc76
    • Andrew Or's avatar
      [SPARK-10548] [SPARK-10563] [SQL] Fix concurrent SQL executions · b6e99863
      Andrew Or authored
      *Note: this is for master branch only.* The fix for branch-1.5 is at #8721.
      
      The query execution ID is currently passed from a thread to its children, which is not the intended behavior. This led to `IllegalArgumentException: spark.sql.execution.id is already set` when running queries in parallel, e.g.:
      ```
      (1 to 100).par.foreach { _ =>
        sc.parallelize(1 to 5).map { i => (i, i) }.toDF("a", "b").count()
      }
      ```
      The cause is `SparkContext`'s local properties are inherited by default. This patch adds a way to exclude keys we don't want to be inherited, and makes SQL go through that code path.
      
      Author: Andrew Or <andrew@databricks.com>
      
      Closes #8710 from andrewor14/concurrent-sql-executions.
      b6e99863
    • DB Tsai's avatar
      [SPARK-7685] [ML] Apply weights to different samples in Logistic Regression · be52faa7
      DB Tsai authored
      In fraud detection dataset, almost all the samples are negative while only couple of them are positive. This type of high imbalanced data will bias the models toward negative resulting poor performance. In python-scikit, they provide a correction allowing users to Over-/undersample the samples of each class according to the given weights. In auto mode, selects weights inversely proportional to class frequencies in the training set. This can be done in a more efficient way by multiplying the weights into loss and gradient instead of doing actual over/undersampling in the training dataset which is very expensive.
      http://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html
      On the other hand, some of the training data maybe more important like the training samples from tenure users while the training samples from new users maybe less important. We should be able to provide another "weight: Double" information in the LabeledPoint to weight them differently in the learning algorithm.
      
      Author: DB Tsai <dbt@netflix.com>
      Author: DB Tsai <dbt@dbs-mac-pro.corp.netflix.com>
      
      Closes #7884 from dbtsai/SPARK-7685.
      be52faa7
    • Wenchen Fan's avatar
      [SPARK-10475] [SQL] improve column prunning for Project on Sort · 31a229aa
      Wenchen Fan authored
      Sometimes we can't push down the whole `Project` though `Sort`, but we still have a chance to push down part of it.
      
      Author: Wenchen Fan <cloud0fan@outlook.com>
      
      Closes #8644 from cloud-fan/column-prune.
      31a229aa
    • Liang-Chi Hsieh's avatar
      [SPARK-10437] [SQL] Support aggregation expressions in Order By · 841972e2
      Liang-Chi Hsieh authored
      JIRA: https://issues.apache.org/jira/browse/SPARK-10437
      
      If an expression in `SortOrder` is a resolved one, such as `count(1)`, the corresponding rule in `Analyzer` to make it work in order by will not be applied.
      
      Author: Liang-Chi Hsieh <viirya@appier.com>
      
      Closes #8599 from viirya/orderby-agg.
      841972e2
    • Marcelo Vanzin's avatar
    • Jacek Laskowski's avatar
      [DOCS] Small fixes to Spark on Yarn doc · 416003b2
      Jacek Laskowski authored
      * a follow-up to 16b6d186 as `--num-executors` flag is not suppported.
      * links + formatting
      
      Author: Jacek Laskowski <jacek.laskowski@deepsense.io>
      
      Closes #8762 from jaceklaskowski/docs-spark-on-yarn.
      416003b2
    • Xiangrui Meng's avatar
      Closes #8738 · 0d9ab016
      Xiangrui Meng authored
      Closes #8767
      Closes #2491
      Closes #6795
      Closes #2096
      Closes #7722
      0d9ab016
    • noelsmith's avatar
      [PYSPARK] [MLLIB] [DOCS] Replaced addversion with versionadded in mllib.random · 7ca30b50
      noelsmith authored
      Missed this when reviewing `pyspark.mllib.random` for SPARK-10275.
      
      Author: noelsmith <mail@noelsmith.com>
      
      Closes #8773 from noel-smith/mllib-random-versionadded-fix.
      7ca30b50
    • Marcelo Vanzin's avatar
      [SPARK-10300] [BUILD] [TESTS] Add support for test tags in run-tests.py. · 8abef21d
      Marcelo Vanzin authored
      This change does two things:
      
      - tag a few tests and adds the mechanism in the build to be able to disable those tags,
        both in maven and sbt, for both junit and scalatest suites.
      - add some logic to run-tests.py to disable some tags depending on what files have
        changed; that's used to disable expensive tests when a module hasn't explicitly
        been changed, to speed up testing for changes that don't directly affect those
        modules.
      
      Author: Marcelo Vanzin <vanzin@cloudera.com>
      
      Closes #8437 from vanzin/test-tags.
      8abef21d
    • Yuhao Yang's avatar
      [SPARK-10491] [MLLIB] move RowMatrix.dspr to BLAS · c35fdcb7
      Yuhao Yang authored
      jira: https://issues.apache.org/jira/browse/SPARK-10491
      
      We implemented dspr with sparse vector support in `RowMatrix`. This method is also used in WeightedLeastSquares and other places. It would be useful to move it to `linalg.BLAS`.
      
      Let me know if new UT needed.
      
      Author: Yuhao Yang <hhbyyh@gmail.com>
      
      Closes #8663 from hhbyyh/movedspr.
      c35fdcb7
    • Reynold Xin's avatar
      Update version to 1.6.0-SNAPSHOT. · 09b7e7c1
      Reynold Xin authored
      Author: Reynold Xin <rxin@databricks.com>
      
      Closes #8350 from rxin/1.6.
      09b7e7c1
    • Robin East's avatar
      [SPARK-10598] [DOCS] · 6503c4b5
      Robin East authored
      Comments preceding toMessage method state: "The edge partition is encoded in the lower
         * 30 bytes of the Int, and the position is encoded in the upper 2 bytes of the Int.". References to bytes should be changed to bits.
      
      This contribution is my original work and I license the work to the Spark project under it's open source license.
      
      Author: Robin East <robin.east@xense.co.uk>
      
      Closes #8756 from insidedctm/master.
      6503c4b5
    • Jacek Laskowski's avatar
      Small fixes to docs · 833be733
      Jacek Laskowski authored
      Links work now properly + consistent use of *Spark standalone cluster* (Spark uppercase + lowercase the rest -- seems agreed in the other places in the docs).
      
      Author: Jacek Laskowski <jacek.laskowski@deepsense.io>
      
      Closes #8759 from jaceklaskowski/docs-submitting-apps.
      833be733
  4. Sep 14, 2015
    • Yu ISHIKAWA's avatar
      [SPARK-10275] [MLLIB] Add @since annotation to pyspark.mllib.random · a2249359
      Yu ISHIKAWA authored
      Author: Yu ISHIKAWA <yuu.ishikawa@gmail.com>
      
      Closes #8666 from yu-iskw/SPARK-10275.
      a2249359
    • noelsmith's avatar
      [SPARK-10273] Add @since annotation to pyspark.mllib.feature · 610971ec
      noelsmith authored
      Duplicated the since decorator from pyspark.sql into pyspark (also tweaked to handle functions without docstrings).
      
      Added since to methods + "versionadded::" to classes (derived from the git file history in pyspark).
      
      Author: noelsmith <mail@noelsmith.com>
      
      Closes #8633 from noel-smith/SPARK-10273-since-mllib-feature.
      610971ec
    • Yanbo Liang's avatar
      [SPARK-9793] [MLLIB] [PYSPARK] PySpark DenseVector, SparseVector implement... · 4ae4d547
      Yanbo Liang authored
      [SPARK-9793] [MLLIB] [PYSPARK] PySpark DenseVector, SparseVector implement __eq__ and __hash__ correctly
      
      PySpark DenseVector, SparseVector ```__eq__``` method should use semantics equality, and DenseVector can compared with SparseVector.
      Implement PySpark DenseVector, SparseVector ```__hash__``` method based on the first 16 entries. That will make PySpark Vector objects can be used in collections.
      
      Author: Yanbo Liang <ybliang8@gmail.com>
      
      Closes #8166 from yanboliang/spark-9793.
      4ae4d547
Loading