Skip to content
Snippets Groups Projects
  1. Jul 24, 2015
  2. Jul 23, 2015
  3. Jul 22, 2015
    • Xiangrui Meng's avatar
      [SPARK-8364] [SPARKR] Add crosstab to SparkR DataFrames · 2f5cbd86
      Xiangrui Meng authored
      Add `crosstab` to SparkR DataFrames, which takes two column names and returns a local R data.frame. This is similar to `table` in R. However, `table` in SparkR is used for loading SQL tables as DataFrames. The return type is data.frame instead table for `crosstab` to be compatible with Scala/Python.
      
      I couldn't run R tests successfully on my local. Many unit tests failed. So let's try Jenkins.
      
      Author: Xiangrui Meng <meng@databricks.com>
      
      Closes #7318 from mengxr/SPARK-8364 and squashes the following commits:
      
      d75e894 [Xiangrui Meng] fix tests
      53f6ddd [Xiangrui Meng] fix tests
      f1348d6 [Xiangrui Meng] update test
      47cb088 [Xiangrui Meng] Merge remote-tracking branch 'apache/master' into SPARK-8364
      5621262 [Xiangrui Meng] first version without test
      2f5cbd86
  4. Jul 20, 2015
    • Eric Liang's avatar
      [SPARK-9201] [ML] Initial integration of MLlib + SparkR using RFormula · 1cbdd899
      Eric Liang authored
      This exposes the SparkR:::glm() and SparkR:::predict() APIs. It was necessary to change RFormula to silently drop the label column if it was missing from the input dataset, which is kind of a hack but necessary to integrate with the Pipeline API.
      
      The umbrella design doc for MLlib + SparkR integration can be viewed here: https://docs.google.com/document/d/10NZNSEurN2EdWM31uFYsgayIPfCFHiuIu3pCWrUmP_c/edit
      
      mengxr
      
      Author: Eric Liang <ekl@databricks.com>
      
      Closes #7483 from ericl/spark-8774 and squashes the following commits:
      
      3dfac0c [Eric Liang] update
      17ef516 [Eric Liang] more comments
      1753a0f [Eric Liang] make glm generic
      b0f50f8 [Eric Liang] equivalence test
      550d56d [Eric Liang] export methods
      c015697 [Eric Liang] second pass
      117949a [Eric Liang] comments
      5afbc67 [Eric Liang] test label columns
      6b7f15f [Eric Liang] Fri Jul 17 14:20:22 PDT 2015
      3a63ae5 [Eric Liang] Fri Jul 17 13:41:52 PDT 2015
      ce61367 [Eric Liang] Fri Jul 17 13:41:17 PDT 2015
      0299c59 [Eric Liang] Fri Jul 17 13:40:32 PDT 2015
      e37603f [Eric Liang] Fri Jul 17 12:15:03 PDT 2015
      d417d0c [Eric Liang] Merge remote-tracking branch 'upstream/master' into spark-8774
      29a2ce7 [Eric Liang] Merge branch 'spark-8774-1' into spark-8774
      d1959d2 [Eric Liang] clarify comment
      2db68aa [Eric Liang] second round of comments
      dc3c943 [Eric Liang] address comments
      5765ec6 [Eric Liang] fix style checks
      1f361b0 [Eric Liang] doc
      d33211b [Eric Liang] r support
      fb0826b [Eric Liang] [SPARK-8774] Add R model formula with basic support as a transformer
      1cbdd899
    • Yu ISHIKAWA's avatar
      [SPARK-9052] [SPARKR] Fix comments after curly braces · 2bdf9914
      Yu ISHIKAWA authored
      [[SPARK-9052] Fix comments after curly braces - ASF JIRA](https://issues.apache.org/jira/browse/SPARK-9052)
      
      This is the full result of lintr at the rivision:01155162.
      [[SPARK-9052] the result of lint-r at the revision:01155162](https://gist.github.com/yu-iskw/e7246041b173a3f29482)
      
      This is the difference of the result between before and after.
      https://gist.github.com/yu-iskw/e7246041b173a3f29482/revisions
      
      Author: Yu ISHIKAWA <yuu.ishikawa@gmail.com>
      
      Closes #7440 from yu-iskw/SPARK-9052 and squashes the following commits:
      
      015d738 [Yu ISHIKAWA] Fix the indentations and move the placement of commna
      5cc30fe [Yu ISHIKAWA] Fix the indentation in a condition
      4ead0e5 [Yu ISHIKAWA] [SPARK-9052][SparkR] Fix comments after curly braces
      2bdf9914
  5. Jul 17, 2015
  6. Jul 16, 2015
  7. Jul 15, 2015
    • Liang-Chi Hsieh's avatar
      [SPARK-8840] [SPARKR] Add float coercion on SparkR · 6f690259
      Liang-Chi Hsieh authored
      JIRA: https://issues.apache.org/jira/browse/SPARK-8840
      
      Currently the type coercion rules don't include float type. This PR simply adds it.
      
      Author: Liang-Chi Hsieh <viirya@appier.com>
      
      Closes #7280 from viirya/add_r_float_coercion and squashes the following commits:
      
      c86dc0e [Liang-Chi Hsieh] For comments.
      dbf0c1b [Liang-Chi Hsieh] Implicitly convert Double to Float based on provided schema.
      733015a [Liang-Chi Hsieh] Add test case for DataFrame with float type.
      30c2a40 [Liang-Chi Hsieh] Update test case.
      52b5294 [Liang-Chi Hsieh] Merge remote-tracking branch 'upstream/master' into add_r_float_coercion
      6f9159d [Liang-Chi Hsieh] Add another test case.
      8db3244 [Liang-Chi Hsieh] schema also needs to support float. add test case.
      0dcc992 [Liang-Chi Hsieh] Add float coercion on SparkR.
      6f690259
    • Sun Rui's avatar
      [SPARK-8808] [SPARKR] Fix assignments in SparkR. · f650a005
      Sun Rui authored
      Author: Sun Rui <rui.sun@intel.com>
      
      Closes #7395 from sun-rui/SPARK-8808 and squashes the following commits:
      
      ce603bc [Sun Rui] Use '<-' instead of '='.
      88590b1 [Sun Rui] Use '<-' instead of '='.
      f650a005
  8. Jul 13, 2015
    • Sun Rui's avatar
      [SPARK-6797] [SPARKR] Add support for YARN cluster mode. · 7f487c8b
      Sun Rui authored
      This PR enables SparkR to dynamically ship the SparkR binary package to the AM node in YARN cluster mode, thus it is no longer required that the SparkR package be installed on each worker node.
      
      This PR uses the JDK jar tool to package the SparkR package, because jar is thought to be available on both Linux/Windows platforms where JDK has been installed.
      
      This PR does not address the R worker involved in RDD API. Will address it in a separate JIRA issue.
      
      This PR does not address SBT build. SparkR installation and packaging by SBT will be addressed in a separate JIRA issue.
      
      R/install-dev.bat is not tested. shivaram , Could you help to test it?
      
      Author: Sun Rui <rui.sun@intel.com>
      
      Closes #6743 from sun-rui/SPARK-6797 and squashes the following commits:
      
      ca63c86 [Sun Rui] Adjust MimaExcludes after rebase.
      7313374 [Sun Rui] Fix unit test errors.
      72695fb [Sun Rui] Fix unit test failures.
      193882f [Sun Rui] Fix Mima test error.
      fe25a33 [Sun Rui] Fix Mima test error.
      35ecfa3 [Sun Rui] Fix comments.
      c38a005 [Sun Rui] Unzipped SparkR binary package is still required for standalone and Mesos modes.
      b05340c [Sun Rui] Fix scala style.
      2ca5048 [Sun Rui] Fix comments.
      1acefd1 [Sun Rui] Fix scala style.
      0aa1e97 [Sun Rui] Fix scala style.
      41d4f17 [Sun Rui] Add support for locating SparkR package for R workers required by RDD APIs.
      49ff948 [Sun Rui] Invoke jar.exe with full path in install-dev.bat.
      7b916c5 [Sun Rui] Use 'rem' consistently.
      3bed438 [Sun Rui] Add a comment.
      681afb0 [Sun Rui] Fix a bug that RRunner does not handle client deployment modes.
      cedfbe2 [Sun Rui] [SPARK-6797][SPARKR] Add support for YARN cluster mode.
      7f487c8b
  9. Jul 09, 2015
  10. Jul 06, 2015
    • Dirceu Semighini Filho's avatar
      Small update in the readme file · 57c72fcc
      Dirceu Semighini Filho authored
      Just change the attribute from -PsparkR to -Psparkr
      
      Author: Dirceu Semighini Filho <dirceu.semighini@gmail.com>
      
      Closes #7242 from dirceusemighini/patch-1 and squashes the following commits:
      
      fad5991 [Dirceu Semighini Filho] Small update in the readme file
      57c72fcc
  11. Jul 05, 2015
  12. Jul 02, 2015
    • Ilya Ganelin's avatar
      [SPARK-3071] Increase default driver memory · 3697232b
      Ilya Ganelin authored
      I've updated default values in comments, documentation, and in the command line builder to be 1g based on comments in the JIRA. I've also updated most usages to point at a single variable defined in the Utils.scala and JavaUtils.java files. This wasn't possible in all cases (R, shell scripts etc.) but usage in most code is now pointing at the same place.
      
      Please let me know if I've missed anything.
      
      Will the spark-shell use the value within the command line builder during instantiation?
      
      Author: Ilya Ganelin <ilya.ganelin@capitalone.com>
      
      Closes #7132 from ilganeli/SPARK-3071 and squashes the following commits:
      
      4074164 [Ilya Ganelin] String fix
      271610b [Ilya Ganelin] Merge branch 'SPARK-3071' of github.com:ilganeli/spark into SPARK-3071
      273b6e9 [Ilya Ganelin] Test fix
      fd67721 [Ilya Ganelin] Update JavaUtils.java
      26cc177 [Ilya Ganelin] test fix
      e5db35d [Ilya Ganelin] Fixed test failure
      39732a1 [Ilya Ganelin] merge fix
      a6f7deb [Ilya Ganelin] Created default value for DRIVER MEM in Utils that's now used in almost all locations instead of setting manually in each
      09ad698 [Ilya Ganelin] Update SubmitRestProtocolSuite.scala
      19b6f25 [Ilya Ganelin] Missed one doc update
      2698a3d [Ilya Ganelin] Updated default value for driver memory
      3697232b
  13. Jul 01, 2015
    • Sun Rui's avatar
      [SPARK-7714] [SPARKR] SparkR tests should use more specific expectations than expect_true · 69c5dee2
      Sun Rui authored
      1. Update the pattern 'expect_true(a == b)' to 'expect_equal(a, b)'.
      2. Update the pattern 'expect_true(inherits(a, b))' to 'expect_is(a, b)'.
      3. Update the pattern 'expect_true(identical(a, b))' to 'expect_identical(a, b)'.
      
      Author: Sun Rui <rui.sun@intel.com>
      
      Closes #7152 from sun-rui/SPARK-7714 and squashes the following commits:
      
      8ad2440 [Sun Rui] Fix test case errors.
      8fe9f0c [Sun Rui] Update the pattern 'expect_true(identical(a, b))' to 'expect_identical(a, b)'.
      f1b8005 [Sun Rui] Update the pattern 'expect_true(inherits(a, b))' to 'expect_is(a, b)'.
      f631e94 [Sun Rui] Update the pattern 'expect_true(a == b)' to 'expect_equal(a, b)'.
      69c5dee2
  14. Jun 30, 2015
  15. Jun 26, 2015
    • cafreeman's avatar
      [SPARK-8607] SparkR -- jars not being added to application classpath correctly · 9d118177
      cafreeman authored
      Add `getStaticClass` method in SparkR's `RBackendHandler`
      
      This is a fix for the problem referenced in [SPARK-5185](https://issues.apache.org/jira/browse/SPARK-5185
      
      ).
      
      cc shivaram
      
      Author: cafreeman <cfreeman@alteryx.com>
      
      Closes #7001 from cafreeman/branch-1.4 and squashes the following commits:
      
      8f81194 [cafreeman] Add missing license
      31aedcf [cafreeman] Refactor test to call an external R script
      2c22073 [cafreeman] Merge branch 'branch-1.4' of github.com:apache/spark into branch-1.4
      0bea809 [cafreeman] Fixed relative path issue and added smaller JAR
      ee25e60 [cafreeman] Merge branch 'branch-1.4' of github.com:apache/spark into branch-1.4
      9a5c362 [cafreeman] test for including JAR when launching sparkContext
      9101223 [cafreeman] Merge branch 'branch-1.4' of github.com:apache/spark into branch-1.4
      5a80844 [cafreeman] Fix style nits
      7c6bd0c [cafreeman] [SPARK-8607] SparkR
      
      (cherry picked from commit 2579948b)
      Signed-off-by: default avatarShivaram Venkataraman <shivaram@cs.berkeley.edu>
      9d118177
    • cafreeman's avatar
      [SPARK-8662] SparkR Update SparkSQL Test · a56516fc
      cafreeman authored
      Test `infer_type` using a more fine-grained approach rather than comparing environments. Since `all.equal`'s behavior has changed in R 3.2, the test became unpassable.
      
      JIRA here:
      https://issues.apache.org/jira/browse/SPARK-8662
      
      
      
      Author: cafreeman <cfreeman@alteryx.com>
      
      Closes #7045 from cafreeman/R32_Test and squashes the following commits:
      
      b97cc52 [cafreeman] Add `checkStructField` utility
      3381e5c [cafreeman] Update SparkSQL Test
      
      (cherry picked from commit 78b31a2a)
      Signed-off-by: default avatarShivaram Venkataraman <shivaram@cs.berkeley.edu>
      a56516fc
  16. Jun 25, 2015
  17. Jun 24, 2015
    • Holden Karau's avatar
      [SPARK-8506] Add pakages to R context created through init. · 43e66192
      Holden Karau authored
      Author: Holden Karau <holden@pigscanfly.ca>
      
      Closes #6928 from holdenk/SPARK-8506-sparkr-does-not-provide-an-easy-way-to-depend-on-spark-packages-when-performing-init-from-inside-of-r and squashes the following commits:
      
      b60dd63 [Holden Karau] Add an example with the spark-csv package
      fa8bc92 [Holden Karau] typo: sparm -> spark
      865a90c [Holden Karau] strip spaces for comparision
      c7a4471 [Holden Karau] Add some documentation
      c1a9233 [Holden Karau] refactor for testing
      c818556 [Holden Karau] Add pakages to R
      43e66192
  18. Jun 23, 2015
    • Alok  Singh's avatar
      [SPARK-8111] [SPARKR] SparkR shell should display Spark logo and version banner on startup. · f2fb0285
      Alok Singh authored
      spark version is taken from the environment variable SPARK_VERSION
      
      Author: Alok  Singh <singhal@Aloks-MacBook-Pro.local>
      Author: Alok  Singh <singhal@aloks-mbp.usca.ibm.com>
      
      Closes #6944 from aloknsingh/aloknsingh_spark_jiras and squashes the following commits:
      
      ed607bd [Alok  Singh] [SPARK-8111][SparkR] As per suggestion, 1) using the version from sparkContext rather than the Sys.env. 2) change "Welcome to SparkR!" to "Welcome to" followed by Spark logo and version
      acd5b85 [Alok  Singh] fix the jira SPARK-8111 to add the spark version and logo. Currently spark version is taken from the environment variable SPARK_VERSION
      f2fb0285
    • Yu ISHIKAWA's avatar
      [SPARK-8431] [SPARKR] Add in operator to DataFrame Column in SparkR · d4f63351
      Yu ISHIKAWA authored
      [[SPARK-8431] Add in operator to DataFrame Column in SparkR - ASF JIRA](https://issues.apache.org/jira/browse/SPARK-8431)
      
      Author: Yu ISHIKAWA <yuu.ishikawa@gmail.com>
      
      Closes #6941 from yu-iskw/SPARK-8431 and squashes the following commits:
      
      1f64423 [Yu ISHIKAWA] Modify the comment
      f4309a7 [Yu ISHIKAWA] Make a `setMethod` for `%in%` be independent
      6e37936 [Yu ISHIKAWA] Modify a variable name
      c196173 [Yu ISHIKAWA] [SPARK-8431][SparkR] Add in operator to DataFrame Column in SparkR
      d4f63351
  19. Jun 22, 2015
  20. Jun 20, 2015
    • Yu ISHIKAWA's avatar
      [SPARK-8495] [SPARKR] Add a `.lintr` file to validate the SparkR files and the `lint-r` script · 004f5737
      Yu ISHIKAWA authored
      Thank Shivaram Venkataraman for your support. This is a prototype script to validate the R files.
      
      Author: Yu ISHIKAWA <yuu.ishikawa@gmail.com>
      
      Closes #6922 from yu-iskw/SPARK-6813 and squashes the following commits:
      
      c1ffe6b [Yu ISHIKAWA] Modify to save result to a log file and add a rule to validate
      5520806 [Yu ISHIKAWA] Exclude the .lintr file not to check Apache lincence
      8f94680 [Yu ISHIKAWA] [SPARK-8495][SparkR] Add a `.lintr` file to validate the SparkR files and the `lint-r` script
      004f5737
  21. Jun 19, 2015
    • Hossein's avatar
      [SPARK-8452] [SPARKR] expose jobGroup API in SparkR · 1fa29c2d
      Hossein authored
      This pull request adds following methods to SparkR:
      
      ```R
      setJobGroup()
      cancelJobGroup()
      clearJobGroup()
      ```
      For each method, the spark context is passed as the first argument. There does not seem to be a good way to test these in R.
      
      cc shivaram and davies
      
      Author: Hossein <hossein@databricks.com>
      
      Closes #6889 from falaki/SPARK-8452 and squashes the following commits:
      
      9ce9f1e [Hossein] Added basic tests to verify methods can be called and won't throw errors
      c706af9 [Hossein] Added examples
      a2c19af [Hossein] taking spark context as first argument
      343ca77 [Hossein] Added setJobGroup, cancelJobGroup and clearJobGroup to SparkR
      1fa29c2d
  22. Jun 15, 2015
    • andrewor14's avatar
      [SPARK-8350] [R] Log R unit test output to "unit-tests.log" · 56d4e8a2
      andrewor14 authored
      Right now it's logged to "R-unit-tests.log". Jenkins currently only archives files named "unit-tests.log", and this is what all other modules (e.g. SQL, network, REPL) use.
      1. We should be consistent
      2. I don't want to reconfigure Jenkins to accept a different file
      
      shivaram
      
      Author: andrewor14 <andrew@databricks.com>
      Author: Andrew Or <andrew@databricks.com>
      
      Closes #6807 from andrewor14/r-logs and squashes the following commits:
      
      96005d2 [andrewor14] Nest unit-tests.log further until R
      407c46c [andrewor14] Add target to log path
      d7b68ae [Andrew Or] Log R unit test output to "unit-tests.log"
      56d4e8a2
  23. Jun 08, 2015
  24. Jun 05, 2015
    • Shivaram Venkataraman's avatar
      [SPARK-8085] [SPARKR] Support user-specified schema in read.df · 12f5eaee
      Shivaram Venkataraman authored
      cc davies sun-rui
      
      Author: Shivaram Venkataraman <shivaram@cs.berkeley.edu>
      
      Closes #6620 from shivaram/sparkr-read-schema and squashes the following commits:
      
      16a6726 [Shivaram Venkataraman] Fix loadDF to pass schema Also add a unit test
      a229877 [Shivaram Venkataraman] Use wrapper function to DataFrameReader
      ee70ba8 [Shivaram Venkataraman] Support user-specified schema in read.df
      12f5eaee
  25. Jun 04, 2015
    • Shivaram Venkataraman's avatar
      [SPARK-8027] [SPARKR] Move man pages creation to install-dev.sh · 3dc00528
      Shivaram Venkataraman authored
      This also helps us get rid of the sparkr-docs maven profile as docs are now built by just using -Psparkr when the roxygen2 package is available
      
      Related to discussion in #6567
      
      cc pwendell srowen -- Let me know if this looks better
      
      Author: Shivaram Venkataraman <shivaram@cs.berkeley.edu>
      
      Closes #6593 from shivaram/sparkr-pom-cleanup and squashes the following commits:
      
      b282241 [Shivaram Venkataraman] Remove sparkr-docs from release script as well
      8f100a5 [Shivaram Venkataraman] Move man pages creation to install-dev.sh This also helps us get rid of the sparkr-docs maven profile as docs are now built by just using -Psparkr when the roxygen2 package is available
      3dc00528
  26. Jun 03, 2015
    • Shivaram Venkataraman's avatar
      [SPARK-8084] [SPARKR] Make SparkR scripts fail on error · 0576c3c4
      Shivaram Venkataraman authored
      cc shaneknapp pwendell JoshRosen
      
      Author: Shivaram Venkataraman <shivaram@cs.berkeley.edu>
      
      Closes #6623 from shivaram/SPARK-8084 and squashes the following commits:
      
      0ec5b26 [Shivaram Venkataraman] Make SparkR scripts fail on error
      0576c3c4
    • Sun Rui's avatar
      [SPARK-8063] [SPARKR] Spark master URL conflict between MASTER env variable... · 708c63bb
      Sun Rui authored
      [SPARK-8063] [SPARKR] Spark master URL conflict between MASTER env variable and --master command line option.
      
      Author: Sun Rui <rui.sun@intel.com>
      
      Closes #6605 from sun-rui/SPARK-8063 and squashes the following commits:
      
      51ca48b [Sun Rui] [SPARK-8063][SPARKR] Spark master URL conflict between MASTER env variable and --master command line option.
      708c63bb
  27. May 31, 2015
    • Sun Rui's avatar
      [SPARK-7227] [SPARKR] Support fillna / dropna in R DataFrame. · 46576ab3
      Sun Rui authored
      Author: Sun Rui <rui.sun@intel.com>
      
      Closes #6183 from sun-rui/SPARK-7227 and squashes the following commits:
      
      dd6f5b3 [Sun Rui] Rename readEnv() back to readMap(). Add alias na.omit() for dropna().
      41cf725 [Sun Rui] [SPARK-7227][SPARKR] Support fillna / dropna in R DataFrame.
      46576ab3
  28. May 29, 2015
    • Shivaram Venkataraman's avatar
      [SPARK-7954] [SPARKR] Create SparkContext in sparkRSQL init · 5fb97dca
      Shivaram Venkataraman authored
      cc davies
      
      Author: Shivaram Venkataraman <shivaram@cs.berkeley.edu>
      
      Closes #6507 from shivaram/sparkr-init and squashes the following commits:
      
      6fdd169 [Shivaram Venkataraman] Create SparkContext in sparkRSQL init
      5fb97dca
    • Shivaram Venkataraman's avatar
      [SPARK-6806] [SPARKR] [DOCS] Add a new SparkR programming guide · 5f48e5c3
      Shivaram Venkataraman authored
      This PR adds a new SparkR programming guide at the top-level. This will be useful for R users as our APIs don't directly match the Scala/Python APIs and as we need to explain SparkR without using RDDs as examples etc.
      
      cc rxin davies pwendell
      
      cc cafreeman -- Would be great if you could also take a look at this !
      
      Author: Shivaram Venkataraman <shivaram@cs.berkeley.edu>
      
      Closes #6490 from shivaram/sparkr-guide and squashes the following commits:
      
      d5ff360 [Shivaram Venkataraman] Add a section on HiveContext, HQL queries
      408dce5 [Shivaram Venkataraman] Fix link
      dbb86e3 [Shivaram Venkataraman] Fix minor typo
      9aff5e0 [Shivaram Venkataraman] Address comments, use dplyr-like syntax in example
      d09703c [Shivaram Venkataraman] Fix default argument in read.df
      ea816a1 [Shivaram Venkataraman] Add a new SparkR programming guide Also update write.df, read.df to handle defaults better
      5f48e5c3
  29. May 23, 2015
    • Shivaram Venkataraman's avatar
      [SPARK-6811] Copy SparkR lib in make-distribution.sh · a40bca01
      Shivaram Venkataraman authored
      This change also remove native libraries from SparkR to make sure our distribution works across platforms
      
      Tested by building on Mac, running on Amazon Linux (CentOS), Windows VM and vice-versa (built on Linux run on Mac)
      
      I will also test this with YARN soon and update this PR.
      
      Author: Shivaram Venkataraman <shivaram@cs.berkeley.edu>
      
      Closes #6373 from shivaram/sparkr-binary and squashes the following commits:
      
      ae41b5c [Shivaram Venkataraman] Remove native libraries from SparkR Also include the built SparkR package in make-distribution.sh
      a40bca01
    • Davies Liu's avatar
      [SPARK-6806] [SPARKR] [DOCS] Fill in SparkR examples in programming guide · 7af3818c
      Davies Liu authored
      sqlCtx -> sqlContext
      
      You can check the docs by:
      
      ```
      $ cd docs
      $ SKIP_SCALADOC=1 jekyll serve
      ```
      cc shivaram
      
      Author: Davies Liu <davies@databricks.com>
      
      Closes #5442 from davies/r_docs and squashes the following commits:
      
      7a12ec6 [Davies Liu] remove rdd in R docs
      8496b26 [Davies Liu] remove the docs related to RDD
      e23b9d6 [Davies Liu] delete R docs for RDD API
      222e4ff [Davies Liu] Merge branch 'master' into r_docs
      89684ce [Davies Liu] Merge branch 'r_docs' of github.com:davies/spark into r_docs
      f0a10e1 [Davies Liu] address comments from @shivaram
      f61de71 [Davies Liu] Update pairRDD.R
      3ef7cf3 [Davies Liu] use + instead of function(a,b) a+b
      2f10a77 [Davies Liu] address comments from @cafreeman
      9c2a062 [Davies Liu] mention R api together with Python API
      23f751a [Davies Liu] Fill in SparkR examples in programming guide
      7af3818c
  30. May 18, 2015
    • Josh Rosen's avatar
      [SPARK-7687] [SQL] DataFrame.describe() should cast all aggregates to String · c9fa870a
      Josh Rosen authored
      In `DataFrame.describe()`, the `count` aggregate produces an integer, the `avg` and `stdev` aggregates produce doubles, and `min` and `max` aggregates can produce varying types depending on what type of column they're applied to.  As a result, we should cast all aggregate results to String so that `describe()`'s output types match its declared output schema.
      
      Author: Josh Rosen <joshrosen@databricks.com>
      
      Closes #6218 from JoshRosen/SPARK-7687 and squashes the following commits:
      
      146b615 [Josh Rosen] Fix R test.
      2974bd5 [Josh Rosen] Cast to string type instead
      f206580 [Josh Rosen] Cast to double to fix SPARK-7687
      307ecbf [Josh Rosen] Add failing regression test for SPARK-7687
      c9fa870a
  31. May 15, 2015
    • qhuang's avatar
      [SPARK-7226] [SPARKR] Support math functions in R DataFrame · 50da9e89
      qhuang authored
      Author: qhuang <qian.huang@intel.com>
      
      Closes #6170 from hqzizania/master and squashes the following commits:
      
      f20c39f [qhuang] add tests units and fixes
      2a7d121 [qhuang] use a function name more familiar to R users
      07aa72e [qhuang] Support math functions in R DataFrame
      50da9e89
  32. May 13, 2015
    • Sun Rui's avatar
      [SPARK-7482] [SPARKR] Rename some DataFrame API methods in SparkR to match... · df9b94a5
      Sun Rui authored
      [SPARK-7482] [SPARKR] Rename some DataFrame API methods in SparkR to match their counterparts in Scala.
      
      Author: Sun Rui <rui.sun@intel.com>
      
      Closes #6007 from sun-rui/SPARK-7482 and squashes the following commits:
      
      5c5cf5e [Sun Rui] Implement alias loadDF() as a new function.
      3a30c10 [Sun Rui] Rename load()/save() to read.df()/write.df(). Also add loadDF()/saveDF() as aliases.
      9f569d6 [Sun Rui] [SPARK-7482][SparkR] Rename some DataFrame API methods in SparkR to match their counterparts in Scala.
      df9b94a5
Loading