diff --git a/dev/change-version-to-2.10.sh b/dev/change-version-to-2.10.sh
index 0962d34c52f289bf62bb71c847f7586de2574c0c..b718d94f849dda79709a8bca76e120712e4b8735 100755
--- a/dev/change-version-to-2.10.sh
+++ b/dev/change-version-to-2.10.sh
@@ -17,7 +17,7 @@
 # limitations under the License.
 #
 
-# This script exists for backwards compability. Use change-scala-version.sh instead.
+# This script exists for backwards compatibility. Use change-scala-version.sh instead.
 echo "This script is deprecated. Please instead run: change-scala-version.sh 2.10"
 
 $(dirname $0)/change-scala-version.sh 2.10
diff --git a/dev/change-version-to-2.11.sh b/dev/change-version-to-2.11.sh
index 4ccfeef09fd04302ae966a41920ccf63d6e1ce9a..93087959a38ddab0cbaa16f842ef9de190921308 100755
--- a/dev/change-version-to-2.11.sh
+++ b/dev/change-version-to-2.11.sh
@@ -17,7 +17,7 @@
 # limitations under the License.
 #
 
-# This script exists for backwards compability. Use change-scala-version.sh instead.
+# This script exists for backwards compatibility. Use change-scala-version.sh instead.
 echo "This script is deprecated. Please instead run: change-scala-version.sh 2.11"
 
 $(dirname $0)/change-scala-version.sh 2.11
diff --git a/python/pyspark/__init__.py b/python/pyspark/__init__.py
index 14c51a306e1c21480656963c8446a56bcdeb09f4..4d142c91629cc7dd15dd7006f8c996d3d8aa6eb6 100644
--- a/python/pyspark/__init__.py
+++ b/python/pyspark/__init__.py
@@ -35,7 +35,7 @@ Public classes:
   - :class:`StorageLevel`:
       Finer-grained cache persistence levels.
   - :class:`TaskContext`:
-      Information about the current running task, avaialble on the workers and experimental.
+      Information about the current running task, available on the workers and experimental.
 
 """
 
diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/ExpressionSet.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/ExpressionSet.scala
index f93e5736de4014357f02ce660bae11c6d3ac6bcf..ede0b1654bbd6749b50750823fbf35bad9b04a27 100644
--- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/ExpressionSet.scala
+++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/ExpressionSet.scala
@@ -39,7 +39,7 @@ object ExpressionSet {
  * guaranteed to see at least one such expression.  For example:
  *
  * {{{
- *   val set = AttributeSet(a + 1, 1 + a)
+ *   val set = ExpressionSet(a + 1, 1 + a)
  *
  *   set.iterator => Iterator(a + 1)
  *   set.contains(a + 1) => true
diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/BatchCommitLog.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/BatchCommitLog.scala
index a34938f911f765f381d1d84a32cc538a8a434dac..5e24e8fc4e3cc1759e240af84c46fe7a4439c656 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/BatchCommitLog.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/BatchCommitLog.scala
@@ -33,7 +33,7 @@ import org.apache.spark.sql.SparkSession
  * - process batch 1
  * - write batch 1 to completion log
  * - trigger batch 2
- * - obtain bactch 2 offsets and write to offset log
+ * - obtain batch 2 offsets and write to offset log
  * - process batch 2
  * - write batch 2 to completion log
  * ....
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala
index 8569c2d76b694d56273565268524bd07131b87a8..5db354d79bb6eaddbb591b76a5d5917c51dbc14b 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala
@@ -507,7 +507,7 @@ class DataFrameAggregateSuite extends QueryTest with SharedSQLContext {
       Row(2.0) :: Row(2.0) :: Row(2.0) :: Nil)
   }
 
-  test("SQL decimal test (used for catching certain demical handling bugs in aggregates)") {
+  test("SQL decimal test (used for catching certain decimal handling bugs in aggregates)") {
     checkAnswer(
       decimalData.groupBy('a cast DecimalType(10, 2)).agg(avg('b cast DecimalType(10, 2))),
       Seq(Row(new java.math.BigDecimal(1.0), new java.math.BigDecimal(1.5)),
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/FileSourceStrategySuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/FileSourceStrategySuite.scala
index 9a2dcafb5e4b3140b4f02893dd6116261dfb4def..d77f0c298ffe3c0b0af949a5d2643f1118e086c2 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/FileSourceStrategySuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/FileSourceStrategySuite.scala
@@ -244,7 +244,7 @@ class FileSourceStrategySuite extends QueryTest with SharedSQLContext with Predi
     val df2 = table.where("(p1 + c2) = 2 AND c1 = 1")
     // Filter on data only are advisory so we have to reevaluate.
     assert(getPhysicalFilters(df2) contains resolve(df2, "c1 = 1"))
-    // Need to evalaute filters that are not pushed down.
+    // Need to evaluate filters that are not pushed down.
     assert(getPhysicalFilters(df2) contains resolve(df2, "(p1 + c2) = 2"))
   }