Skip to content
Snippets Groups Projects
  1. Jul 25, 2014
    • Cheng Lian's avatar
      [SPARK-2410][SQL] Merging Hive Thrift/JDBC server · 06dc0d2c
      Cheng Lian authored
      JIRA issue:
      
      - Main: [SPARK-2410](https://issues.apache.org/jira/browse/SPARK-2410)
      - Related: [SPARK-2678](https://issues.apache.org/jira/browse/SPARK-2678)
      
      Cherry picked the Hive Thrift/JDBC server from [branch-1.0-jdbc](https://github.com/apache/spark/tree/branch-1.0-jdbc).
      
      (Thanks chenghao-intel for his initial contribution of the Spark SQL CLI.)
      
      TODO
      
      - [x] Use `spark-submit` to launch the server, the CLI and beeline
      - [x] Migration guideline draft for Shark users
      
      ----
      
      Hit by a bug in `SparkSubmitArguments` while working on this PR: all application options that are recognized by `SparkSubmitArguments` are stolen as `SparkSubmit` options. For example:
      
      ```bash
      $ spark-submit --class org.apache.hive.beeline.BeeLine spark-internal --help
      ```
      
      This actually shows usage information of `SparkSubmit` rather than `BeeLine`.
      
      ~~Fixed this bug here since the `spark-internal` related stuff also touches `SparkSubmitArguments` and I'd like to avoid conflict.~~
      
      **UPDATE** The bug mentioned above is now tracked by [SPARK-2678](https://issues.apache.org/jira/browse/SPARK-2678). Decided to revert changes to this bug since it involves more subtle considerations and worth a separate PR.
      
      Author: Cheng Lian <lian.cs.zju@gmail.com>
      
      Closes #1399 from liancheng/thriftserver and squashes the following commits:
      
      090beea [Cheng Lian] Revert changes related to SPARK-2678, decided to move them to another PR
      21c6cf4 [Cheng Lian] Updated Spark SQL programming guide docs
      fe0af31 [Cheng Lian] Reordered spark-submit options in spark-shell[.cmd]
      199e3fb [Cheng Lian] Disabled MIMA for hive-thriftserver
      1083e9d [Cheng Lian] Fixed failed test suites
      7db82a1 [Cheng Lian] Fixed spark-submit application options handling logic
      9cc0f06 [Cheng Lian] Starts beeline with spark-submit
      cfcf461 [Cheng Lian] Updated documents and build scripts for the newly added hive-thriftserver profile
      061880f [Cheng Lian] Addressed all comments by @pwendell
      7755062 [Cheng Lian] Adapts test suites to spark-submit settings
      40bafef [Cheng Lian] Fixed more license header issues
      e214aab [Cheng Lian] Added missing license headers
      b8905ba [Cheng Lian] Fixed minor issues in spark-sql and start-thriftserver.sh
      f975d22 [Cheng Lian] Updated docs for Hive compatibility and Shark migration guide draft
      3ad4e75 [Cheng Lian] Starts spark-sql shell with spark-submit
      a5310d1 [Cheng Lian] Make HiveThriftServer2 play well with spark-submit
      61f39f4 [Cheng Lian] Starts Hive Thrift server via spark-submit
      2c4c539 [Cheng Lian] Cherry picked the Hive Thrift server
      06dc0d2c
  2. Jul 19, 2014
    • Patrick Wendell's avatar
      SPARK-2596 HOTFIX: Deal with non-existent JIRAs. · d39e3b96
      Patrick Wendell authored
      A small bug that was found in our JIRA sync script.
      d39e3b96
    • Patrick Wendell's avatar
      SPARK-2596 A tool for mirroring github pull requests on JIRA. · 49e47274
      Patrick Wendell authored
      For a bunch of reasons we should automatically populate a JIRA with information about new pull requests when they arrive. I've written a small python script to do this that we can run from Jenkins every 5 or 10 minutes to keep things in sync.
      
      Author: Patrick Wendell <pwendell@gmail.com>
      
      Closes #1496 from pwendell/github-integration and squashes the following commits:
      
      55ad226 [Patrick Wendell] Small fix
      afda547 [Patrick Wendell] Use sequence instead of dictiory for JIRA's
      3e18cc1 [Patrick Wendell] Small edits
      84c5606 [Patrick Wendell] SPARK-2596 A tool for mirroring github pull requests on JIRA.
      49e47274
  3. Jul 17, 2014
    • Patrick Wendell's avatar
      SPARK-2526: Simplify options in make-distribution.sh · d0ea4968
      Patrick Wendell authored
      Right now we have a bunch of parallel logic in make-distribution.sh
      that's just extra work to maintain. We should just pass through
      Maven profiles in this case and keep the script simple. See
      the JIRA for more details.
      
      Author: Patrick Wendell <pwendell@gmail.com>
      
      Closes #1445 from pwendell/make-distribution.sh and squashes the following commits:
      
      f1294ea [Patrick Wendell] Simplify options in make-distribution.sh.
      d0ea4968
  4. Jul 15, 2014
    • witgo's avatar
      SPARK-2480: Resolve sbt warnings "NOTE: SPARK_YARN is deprecated, please use -Pyarn flag" · 9dd635eb
      witgo authored
      Author: witgo <witgo@qq.com>
      
      Closes #1404 from witgo/run-tests and squashes the following commits:
      
      f703aee [witgo] fix Note: implicit method fromPairDStream is not applicable here because it comes after the application point and it lacks an explicit result type
      2944f51 [witgo] Remove "NOTE: SPARK_YARN is deprecated, please use -Pyarn flag"
      ef59c70 [witgo] fix Note: implicit method fromPairDStream is not applicable here because it comes after the application point and it lacks an explicit result type
      6cefee5 [witgo] Remove "NOTE: SPARK_YARN is deprecated, please use -Pyarn flag"
      9dd635eb
  5. Jul 10, 2014
    • Prashant Sharma's avatar
      [SPARK-1776] Have Spark's SBT build read dependencies from Maven. · 628932b8
      Prashant Sharma authored
      Patch introduces the new way of working also retaining the existing ways of doing things.
      
      For example build instruction for yarn in maven is
      `mvn -Pyarn -PHadoop2.2 clean package -DskipTests`
      in sbt it can become
      `MAVEN_PROFILES="yarn, hadoop-2.2" sbt/sbt clean assembly`
      Also supports
      `sbt/sbt -Pyarn -Phadoop-2.2 -Dhadoop.version=2.2.0 clean assembly`
      
      Author: Prashant Sharma <prashant.s@imaginea.com>
      Author: Patrick Wendell <pwendell@gmail.com>
      
      Closes #772 from ScrapCodes/sbt-maven and squashes the following commits:
      
      a8ac951 [Prashant Sharma] Updated sbt version.
      62b09bb [Prashant Sharma] Improvements.
      fa6221d [Prashant Sharma] Excluding sql from mima
      4b8875e [Prashant Sharma] Sbt assembly no longer builds tools by default.
      72651ca [Prashant Sharma] Addresses code reivew comments.
      acab73d [Prashant Sharma] Revert "Small fix to run-examples script."
      ac4312c [Prashant Sharma] Revert "minor fix"
      6af91ac [Prashant Sharma] Ported oldDeps back. + fixes issues with prev commit.
      65cf06c [Prashant Sharma] Servelet API jars mess up with the other servlet jars on the class path.
      446768e [Prashant Sharma] minor fix
      89b9777 [Prashant Sharma] Merge conflicts
      d0a02f2 [Prashant Sharma] Bumped up pom versions, Since the build now depends on pom it is better updated there. + general cleanups.
      dccc8ac [Prashant Sharma] updated mima to check against 1.0
      a49c61b [Prashant Sharma] Fix for tools jar
      a2f5ae1 [Prashant Sharma] Fixes a bug in dependencies.
      cf88758 [Prashant Sharma] cleanup
      9439ea3 [Prashant Sharma] Small fix to run-examples script.
      96cea1f [Prashant Sharma] SPARK-1776 Have Spark's SBT build read dependencies from Maven.
      36efa62 [Patrick Wendell] Set project name in pom files and added eclipse/intellij plugins.
      4973dbd [Patrick Wendell] Example build using pom reader.
      628932b8
  6. Jul 09, 2014
    • Patrick Wendell's avatar
      SPARK-2416: Allow richer reporting of unit test results · 2e0a037d
      Patrick Wendell authored
      The built-in Jenkins integration is pretty bad. It's very confusing to users whether tests have passed or failed and we can't easily customize the message.
      
      With some small scripting around the Github API we can do much better than this.
      
      Author: Patrick Wendell <pwendell@gmail.com>
      
      Closes #1340 from pwendell/better-qa-messages and squashes the following commits:
      
      fd6077d [Patrick Wendell] Better automation for unit tests.
      2e0a037d
  7. Jul 04, 2014
  8. Jun 26, 2014
    • Patrick Wendell's avatar
      Strip '@' symbols when merging pull requests. · f1f7385a
      Patrick Wendell authored
      Currently all of the commits with 'X' in them cause person X to
      receive e-mails every time someone makes a public fork of Spark.
      
      marmbrus who requested this.
      
      Author: Patrick Wendell <pwendell@gmail.com>
      
      Closes #1239 from pwendell/strip and squashes the following commits:
      
      22e5a97 [Patrick Wendell] Strip '@' symbols when merging pull requests.
      f1f7385a
  9. Jun 22, 2014
    • Patrick Wendell's avatar
      SPARK-2231: dev/run-tests should include YARN and use a recent Hadoop version · 58b32f34
      Patrick Wendell authored
      ...rsion
      
      Author: Patrick Wendell <pwendell@gmail.com>
      
      Closes #1175 from pwendell/test-hadoop-version and squashes the following commits:
      
      9210ef4 [Patrick Wendell] SPARK-2231: dev/run-tests should include YARN and use a recent Hadoop version
      58b32f34
    • Sean Owen's avatar
      SPARK-1996. Remove use of special Maven repo for Akka · 1db9cbc3
      Sean Owen authored
      Just following up Matei's suggestion to remove the Akka repo references. Builds and the audit-release script appear OK.
      
      Author: Sean Owen <sowen@cloudera.com>
      
      Closes #1170 from srowen/SPARK-1996 and squashes the following commits:
      
      5ca2930 [Sean Owen] Remove outdated Akka repository references
      1db9cbc3
  10. Jun 21, 2014
  11. Jun 11, 2014
    • Prashant Sharma's avatar
      [SPARK-2069] MIMA false positives · 5b754b45
      Prashant Sharma authored
      Fixes SPARK 2070 and 2071
      
      Author: Prashant Sharma <prashant.s@imaginea.com>
      
      Closes #1021 from ScrapCodes/SPARK-2070/package-private-methods and squashes the following commits:
      
      7979a57 [Prashant Sharma] addressed code review comments
      558546d [Prashant Sharma] A little fancy error message.
      59275ab [Prashant Sharma] SPARK-2071 Mima ignores classes and its members from previous versions too.
      0c4ff2b [Prashant Sharma] SPARK-2070 Ignore methods along with annotated classes.
      5b754b45
  12. Jun 10, 2014
    • Patrick Wendell's avatar
      HOTFIX: Fix Python tests on Jenkins. · fb499be1
      Patrick Wendell authored
      Author: Patrick Wendell <pwendell@gmail.com>
      
      Closes #1036 from pwendell/jenkins-test and squashes the following commits:
      
      9c99856 [Patrick Wendell] Better output during tests
      71e7b74 [Patrick Wendell] Removing incorrect python path
      74984db [Patrick Wendell] HOTFIX: Allow PySpark tests to run on Jenkins.
      fb499be1
  13. Jun 07, 2014
    • Patrick Wendell's avatar
      HOTFIX: Support empty body in merge script · 3ace10dc
      Patrick Wendell authored
      Discovered in #992
      
      Author: Patrick Wendell <pwendell@gmail.com>
      
      Closes #1007 from pwendell/hotfix and squashes the following commits:
      
      af90aa0 [Patrick Wendell] HOTFIX: Support empty body in merge script
      3ace10dc
  14. Jun 05, 2014
    • Patrick Wendell's avatar
      HOTFIX: Remove generated-mima-excludes file after runing MIMA. · f6143f12
      Patrick Wendell authored
      This has been causing some false failures on PR's that don't merge
      correctly.
      
      Author: Patrick Wendell <pwendell@gmail.com>
      
      Closes #971 from pwendell/mima and squashes the following commits:
      
      1dc80aa [Patrick Wendell] HOTFIX: Remove generated-mima-excludes file after runing MIMA.
      f6143f12
  15. Jun 03, 2014
    • Xiangrui Meng's avatar
      use env default python in merge_spark_pr.py · e8d93ee5
      Xiangrui Meng authored
      A minor change to use env default python instead of fixed `/usr/bin/python`.
      
      Author: Xiangrui Meng <meng@databricks.com>
      
      Closes #965 from mengxr/merge-pr-python and squashes the following commits:
      
      1ae0013 [Xiangrui Meng] use env default python in merge_spark_pr.py
      e8d93ee5
  16. Jun 01, 2014
    • Patrick Wendell's avatar
      Better explanation for how to use MIMA excludes. · d17d2214
      Patrick Wendell authored
      This patch does a few things:
      1. We have a file MimaExcludes.scala exclusively for excludes.
      2. The test runner tells users about that file if a test fails.
      3. I've added back the excludes used from 0.9->1.0. We should keep
         these in the project as an official audit trail of times where
         we decided to make exceptions.
      
      Author: Patrick Wendell <pwendell@gmail.com>
      
      Closes #937 from pwendell/mima and squashes the following commits:
      
      7ee0db2 [Patrick Wendell] Better explanation for how to use MIMA excludes.
      d17d2214
  17. May 26, 2014
  18. May 22, 2014
    • Tathagata Das's avatar
      Updated scripts for auditing releases · b2bdd0e5
      Tathagata Das authored
      - Added script to automatically generate change list CHANGES.txt
      - Added test for verifying linking against maven distributions of `spark-sql` and `spark-hive`
      - Added SBT projects for testing functionality of `spark-sql` and `spark-hive`
      - Fixed issues in existing tests that might have come up because of changes in Spark 1.0
      
      Author: Tathagata Das <tathagata.das1565@gmail.com>
      
      Closes #844 from tdas/update-dev-scripts and squashes the following commits:
      
      25090ba [Tathagata Das] Added missing license
      e2e20b3 [Tathagata Das] Updated tests for auditing releases.
      b2bdd0e5
  19. May 15, 2014
  20. May 12, 2014
  21. Apr 29, 2014
  22. Apr 28, 2014
  23. Apr 27, 2014
    • Patrick Wendell's avatar
      HOTFIX: Minor patch to merge script. · 3d9fb096
      Patrick Wendell authored
      3d9fb096
    • Patrick Wendell's avatar
      SPARK-1648 Support closing JIRA's as part of merge script. · fe65beea
      Patrick Wendell authored
      Adds an automated hook in the merge script that can close the JIRA,
      set the fix versions, and leave a comment on the JIRA indicating the
      PR in which it was resolved. This ensures that (a) we always close JIRA's
      when issues are merged and (b) there is a link to the pull request in every JIRA.
      
      This requires a python library called `jira-client`. We could look at embedding this
      library in our project, but it seemed simple enough to just gracefully disable this
      feature if it is not installed. It can be installed with `pip install jira-client`.
      
      Author: Patrick Wendell <pwendell@gmail.com>
      
      Closes #570 from pwendell/jira-pr-merge and squashes the following commits:
      
      3022b96 [Patrick Wendell] SPARK-1648 Support closing JIRA's as part of merge script.
      fe65beea
  24. Apr 24, 2014
  25. Apr 23, 2014
    • Patrick Wendell's avatar
      SPARK-1119 and other build improvements · cd4ed293
      Patrick Wendell authored
      1. Makes assembly and examples jar naming consistent in maven/sbt.
      2. Updates make-distribution.sh to use Maven and fixes some bugs.
      3. Updates the create-release script to call make-distribution script.
      
      Author: Patrick Wendell <pwendell@gmail.com>
      
      Closes #502 from pwendell/make-distribution and squashes the following commits:
      
      1a97f0d [Patrick Wendell] SPARK-1119 and other build improvements
      cd4ed293
  26. Apr 21, 2014
    • Patrick Wendell's avatar
      Dev script: include RC name in git tag · 5a5b3346
      Patrick Wendell authored
      5a5b3346
    • Patrick Wendell's avatar
      Clean up and simplify Spark configuration · fb98488f
      Patrick Wendell authored
      Over time as we've added more deployment modes, this have gotten a bit unwieldy with user-facing configuration options in Spark. Going forward we'll advise all users to run `spark-submit` to launch applications. This is a WIP patch but it makes the following improvements:
      
      1. Improved `spark-env.sh.template` which was missing a lot of things users now set in that file.
      2. Removes the shipping of SPARK_CLASSPATH, SPARK_JAVA_OPTS, and SPARK_LIBRARY_PATH to the executors on the cluster. This was an ugly hack. Instead it introduces config variables spark.executor.extraJavaOpts, spark.executor.extraLibraryPath, and spark.executor.extraClassPath.
      3. Adds ability to set these same variables for the driver using `spark-submit`.
      4. Allows you to load system properties from a `spark-defaults.conf` file when running `spark-submit`. This will allow setting both SparkConf options and other system properties utilized by `spark-submit`.
      5. Made `SPARK_LOCAL_IP` an environment variable rather than a SparkConf property. This is more consistent with it being set on each node.
      
      Author: Patrick Wendell <pwendell@gmail.com>
      
      Closes #299 from pwendell/config-cleanup and squashes the following commits:
      
      127f301 [Patrick Wendell] Improvements to testing
      a006464 [Patrick Wendell] Moving properties file template.
      b4b496c [Patrick Wendell] spark-defaults.properties -> spark-defaults.conf
      0086939 [Patrick Wendell] Minor style fixes
      af09e3e [Patrick Wendell] Mention config file in docs and clean-up docs
      b16e6a2 [Patrick Wendell] Cleanup of spark-submit script and Scala quick start guide
      af0adf7 [Patrick Wendell] Automatically add user jar
      a56b125 [Patrick Wendell] Responses to Tom's review
      d50c388 [Patrick Wendell] Merge remote-tracking branch 'apache/master' into config-cleanup
      a762901 [Patrick Wendell] Fixing test failures
      ffa00fe [Patrick Wendell] Review feedback
      fda0301 [Patrick Wendell] Note
      308f1f6 [Patrick Wendell] Properly escape quotes and other clean-up for YARN
      e83cd8f [Patrick Wendell] Changes to allow re-use of test applications
      be42f35 [Patrick Wendell] Handle case where SPARK_HOME is not set
      c2a2909 [Patrick Wendell] Test compile fixes
      4ee6f9d [Patrick Wendell] Making YARN doc changes consistent
      afc9ed8 [Patrick Wendell] Cleaning up line limits and two compile errors.
      b08893b [Patrick Wendell] Additional improvements.
      ace4ead [Patrick Wendell] Responses to review feedback.
      b72d183 [Patrick Wendell] Review feedback for spark env file
      46555c1 [Patrick Wendell] Review feedback and import clean-ups
      437aed1 [Patrick Wendell] Small fix
      761ebcd [Patrick Wendell] Library path and classpath for drivers
      7cc70e4 [Patrick Wendell] Clean up terminology inside of spark-env script
      5b0ba8e [Patrick Wendell] Don't ship executor envs
      84cc5e5 [Patrick Wendell] Small clean-up
      1f75238 [Patrick Wendell] SPARK_JAVA_OPTS --> SPARK_MASTER_OPTS for master settings
      4982331 [Patrick Wendell] Remove SPARK_LIBRARY_PATH
      6eaf7d0 [Patrick Wendell] executorJavaOpts
      0faa3b6 [Patrick Wendell] Stash of adding config options in submit script and YARN
      ac2d65e [Patrick Wendell] Change spark.local.dir -> SPARK_LOCAL_DIRS
      fb98488f
  27. Apr 17, 2014
    • Patrick Wendell's avatar
      FIX: Don't build Hive in assembly unless running Hive tests. · 6c746ba3
      Patrick Wendell authored
      This will make the tests more stable when not running SQL tests.
      
      Author: Patrick Wendell <pwendell@gmail.com>
      
      Closes #439 from pwendell/hive-tests and squashes the following commits:
      
      88a6032 [Patrick Wendell] FIX: Don't build Hive in assembly unless running Hive tests.
      6c746ba3
  28. Apr 16, 2014
    • Patrick Wendell's avatar
      Add clean to build · 987760ec
      Patrick Wendell authored
      987760ec
    • Patrick Wendell's avatar
      Minor addition to SPARK-1497 · 82349fbd
      Patrick Wendell authored
      82349fbd
    • Sean Owen's avatar
      SPARK-1497. Fix scalastyle warnings in YARN, Hive code · 77f83679
      Sean Owen authored
      (I wasn't sure how to automatically set `SPARK_YARN=true` and `SPARK_HIVE=true` when running scalastyle, but these are the errors that turn up.)
      
      Author: Sean Owen <sowen@cloudera.com>
      
      Closes #413 from srowen/SPARK-1497 and squashes the following commits:
      
      f0c9318 [Sean Owen] Fix more scalastyle warnings in yarn
      80bf4c3 [Sean Owen] Add YARN alpha / YARN profile to scalastyle check
      026319c [Sean Owen] Fix scalastyle warnings in YARN, Hive code
      77f83679
  29. Apr 15, 2014
    • Patrick Wendell's avatar
      SPARK-1455: Better isolation for unit tests. · 5aaf9836
      Patrick Wendell authored
      This is a simple first step towards avoiding running the Hive tests
      whenever possible.
      
      Author: Patrick Wendell <pwendell@gmail.com>
      
      Closes #420 from pwendell/test-isolation and squashes the following commits:
      
      350c8af [Patrick Wendell] SPARK-1455: Better isolation for unit tests.
      5aaf9836
    • Ahir Reddy's avatar
      SPARK-1374: PySpark API for SparkSQL · c99bcb7f
      Ahir Reddy authored
      An initial API that exposes SparkSQL functionality in PySpark. A PythonRDD composed of dictionaries, with string keys and primitive values (boolean, float, int, long, string) can be converted into a SchemaRDD that supports sql queries.
      
      ```
      from pyspark.context import SQLContext
      sqlCtx = SQLContext(sc)
      rdd = sc.parallelize([{"field1" : 1, "field2" : "row1"}, {"field1" : 2, "field2": "row2"}, {"field1" : 3, "field2": "row3"}])
      srdd = sqlCtx.applySchema(rdd)
      sqlCtx.registerRDDAsTable(srdd, "table1")
      srdd2 = sqlCtx.sql("SELECT field1 AS f1, field2 as f2 from table1")
      srdd2.collect()
      ```
      The last line yields ```[{"f1" : 1, "f2" : "row1"}, {"f1" : 2, "f2": "row2"}, {"f1" : 3, "f2": "row3"}]```
      
      Author: Ahir Reddy <ahirreddy@gmail.com>
      Author: Michael Armbrust <michael@databricks.com>
      
      Closes #363 from ahirreddy/pysql and squashes the following commits:
      
      0294497 [Ahir Reddy] Updated log4j properties to supress Hive Warns
      307d6e0 [Ahir Reddy] Style fix
      6f7b8f6 [Ahir Reddy] Temporary fix MIMA checker. Since we now assemble Spark jar with Hive, we don't want to check the interfaces of all of our hive dependencies
      3ef074a [Ahir Reddy] Updated documentation because classes moved to sql.py
      29245bf [Ahir Reddy] Cache underlying SchemaRDD instead of generating and caching PythonRDD
      f2312c7 [Ahir Reddy] Moved everything into sql.py
      a19afe4 [Ahir Reddy] Doc fixes
      6d658ba [Ahir Reddy] Remove the metastore directory created by the HiveContext tests in SparkSQL
      521ff6d [Ahir Reddy] Trying to get spark to build with hive
      ab95eba [Ahir Reddy] Set SPARK_HIVE=true on jenkins
      ded03e7 [Ahir Reddy] Added doc test for HiveContext
      22de1d4 [Ahir Reddy] Fixed maven pyrolite dependency
      e4da06c [Ahir Reddy] Display message if hive is not built into spark
      227a0be [Michael Armbrust] Update API links. Fix Hive example.
      58e2aa9 [Michael Armbrust] Build Docs for pyspark SQL Api.  Minor fixes.
      4285340 [Michael Armbrust] Fix building of Hive API Docs.
      38a92b0 [Michael Armbrust] Add note to future non-python developers about python docs.
      337b201 [Ahir Reddy] Changed com.clearspring.analytics stream version from 2.4.0 to 2.5.1 to match SBT build, and added pyrolite to maven build
      40491c9 [Ahir Reddy] PR Changes + Method Visibility
      1836944 [Michael Armbrust] Fix comments.
      e00980f [Michael Armbrust] First draft of python sql programming guide.
      b0192d3 [Ahir Reddy] Added Long, Double and Boolean as usable types + unit test
      f98a422 [Ahir Reddy] HiveContexts
      79621cf [Ahir Reddy] cleaning up cruft
      b406ba0 [Ahir Reddy] doctest formatting
      20936a5 [Ahir Reddy] Added tests and documentation
      e4d21b4 [Ahir Reddy] Added pyrolite dependency
      79f739d [Ahir Reddy] added more tests
      7515ba0 [Ahir Reddy] added more tests :)
      d26ec5e [Ahir Reddy] added test
      e9f5b8d [Ahir Reddy] adding tests
      906d180 [Ahir Reddy] added todo explaining cost of creating Row object in python
      251f99d [Ahir Reddy] for now only allow dictionaries as input
      09b9980 [Ahir Reddy] made jrdd explicitly lazy
      c608947 [Ahir Reddy] SchemaRDD now has all RDD operations
      725c91e [Ahir Reddy] awesome row objects
      55d1c76 [Ahir Reddy] return row objects
      4fe1319 [Ahir Reddy] output dictionaries correctly
      be079de [Ahir Reddy] returning dictionaries works
      cd5f79f [Ahir Reddy] Switched to using Scala SQLContext
      e948bd9 [Ahir Reddy] yippie
      4886052 [Ahir Reddy] even better
      c0fb1c6 [Ahir Reddy] more working
      043ca85 [Ahir Reddy] working
      5496f9f [Ahir Reddy] doesn't crash
      b8b904b [Ahir Reddy] Added schema rdd class
      67ba875 [Ahir Reddy] java to python, and python to java
      bcc0f23 [Ahir Reddy] Java to python
      ab6025d [Ahir Reddy] compiling
      c99bcb7f
Loading