diff --git a/dev/change-version-to-2.10.sh b/dev/change-version-to-2.10.sh index 0962d34c52f289bf62bb71c847f7586de2574c0c..b718d94f849dda79709a8bca76e120712e4b8735 100755 --- a/dev/change-version-to-2.10.sh +++ b/dev/change-version-to-2.10.sh @@ -17,7 +17,7 @@ # limitations under the License. # -# This script exists for backwards compability. Use change-scala-version.sh instead. +# This script exists for backwards compatibility. Use change-scala-version.sh instead. echo "This script is deprecated. Please instead run: change-scala-version.sh 2.10" $(dirname $0)/change-scala-version.sh 2.10 diff --git a/dev/change-version-to-2.11.sh b/dev/change-version-to-2.11.sh index 4ccfeef09fd04302ae966a41920ccf63d6e1ce9a..93087959a38ddab0cbaa16f842ef9de190921308 100755 --- a/dev/change-version-to-2.11.sh +++ b/dev/change-version-to-2.11.sh @@ -17,7 +17,7 @@ # limitations under the License. # -# This script exists for backwards compability. Use change-scala-version.sh instead. +# This script exists for backwards compatibility. Use change-scala-version.sh instead. echo "This script is deprecated. Please instead run: change-scala-version.sh 2.11" $(dirname $0)/change-scala-version.sh 2.11 diff --git a/python/pyspark/__init__.py b/python/pyspark/__init__.py index 14c51a306e1c21480656963c8446a56bcdeb09f4..4d142c91629cc7dd15dd7006f8c996d3d8aa6eb6 100644 --- a/python/pyspark/__init__.py +++ b/python/pyspark/__init__.py @@ -35,7 +35,7 @@ Public classes: - :class:`StorageLevel`: Finer-grained cache persistence levels. - :class:`TaskContext`: - Information about the current running task, avaialble on the workers and experimental. + Information about the current running task, available on the workers and experimental. """ diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/ExpressionSet.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/ExpressionSet.scala index f93e5736de4014357f02ce660bae11c6d3ac6bcf..ede0b1654bbd6749b50750823fbf35bad9b04a27 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/ExpressionSet.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/ExpressionSet.scala @@ -39,7 +39,7 @@ object ExpressionSet { * guaranteed to see at least one such expression. For example: * * {{{ - * val set = AttributeSet(a + 1, 1 + a) + * val set = ExpressionSet(a + 1, 1 + a) * * set.iterator => Iterator(a + 1) * set.contains(a + 1) => true diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/BatchCommitLog.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/BatchCommitLog.scala index a34938f911f765f381d1d84a32cc538a8a434dac..5e24e8fc4e3cc1759e240af84c46fe7a4439c656 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/BatchCommitLog.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/BatchCommitLog.scala @@ -33,7 +33,7 @@ import org.apache.spark.sql.SparkSession * - process batch 1 * - write batch 1 to completion log * - trigger batch 2 - * - obtain bactch 2 offsets and write to offset log + * - obtain batch 2 offsets and write to offset log * - process batch 2 * - write batch 2 to completion log * .... diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala index 8569c2d76b694d56273565268524bd07131b87a8..5db354d79bb6eaddbb591b76a5d5917c51dbc14b 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala @@ -507,7 +507,7 @@ class DataFrameAggregateSuite extends QueryTest with SharedSQLContext { Row(2.0) :: Row(2.0) :: Row(2.0) :: Nil) } - test("SQL decimal test (used for catching certain demical handling bugs in aggregates)") { + test("SQL decimal test (used for catching certain decimal handling bugs in aggregates)") { checkAnswer( decimalData.groupBy('a cast DecimalType(10, 2)).agg(avg('b cast DecimalType(10, 2))), Seq(Row(new java.math.BigDecimal(1.0), new java.math.BigDecimal(1.5)), diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/FileSourceStrategySuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/FileSourceStrategySuite.scala index 9a2dcafb5e4b3140b4f02893dd6116261dfb4def..d77f0c298ffe3c0b0af949a5d2643f1118e086c2 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/FileSourceStrategySuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/FileSourceStrategySuite.scala @@ -244,7 +244,7 @@ class FileSourceStrategySuite extends QueryTest with SharedSQLContext with Predi val df2 = table.where("(p1 + c2) = 2 AND c1 = 1") // Filter on data only are advisory so we have to reevaluate. assert(getPhysicalFilters(df2) contains resolve(df2, "c1 = 1")) - // Need to evalaute filters that are not pushed down. + // Need to evaluate filters that are not pushed down. assert(getPhysicalFilters(df2) contains resolve(df2, "(p1 + c2) = 2")) }