diff --git a/python/docs/index.rst b/python/docs/index.rst
index 306ffdb0e0f138ae03797b4931e33659b474cb62..421c8de86a3cc1b895bfcf5b2159d2b0bcc51658 100644
--- a/python/docs/index.rst
+++ b/python/docs/index.rst
@@ -50,4 +50,3 @@ Indices and tables
 ==================
 
 * :ref:`search`
-
diff --git a/python/docs/pyspark.sql.rst b/python/docs/pyspark.sql.rst
index 3be9533c126d2b07deeae54fc9d70e360d94ff26..09848b880194d35b225f67f337baa1cd7cb68234 100644
--- a/python/docs/pyspark.sql.rst
+++ b/python/docs/pyspark.sql.rst
@@ -8,14 +8,12 @@ Module Context
     :members:
     :undoc-members:
 
-
 pyspark.sql.types module
 ------------------------
 .. automodule:: pyspark.sql.types
     :members:
     :undoc-members:
 
-
 pyspark.sql.functions module
 ----------------------------
 .. automodule:: pyspark.sql.functions
diff --git a/python/pyspark/sql/__init__.py b/python/pyspark/sql/__init__.py
index cff73ff192e51e55f72337c2ce417f2d5d5a4832..22ec416f6c584379677f01392b97def17c2ace13 100644
--- a/python/pyspark/sql/__init__.py
+++ b/python/pyspark/sql/__init__.py
@@ -18,7 +18,7 @@
 """
 Important classes of Spark SQL and DataFrames:
 
-    - :class:`pyspark.sql.SQLContext`
+    - :class:`pyspark.sql.SparkSession`
       Main entry point for :class:`DataFrame` and SQL functionality.
     - :class:`pyspark.sql.DataFrame`
       A distributed collection of data grouped into named columns.
@@ -26,8 +26,6 @@ Important classes of Spark SQL and DataFrames:
       A column expression in a :class:`DataFrame`.
     - :class:`pyspark.sql.Row`
       A row of data in a :class:`DataFrame`.
-    - :class:`pyspark.sql.HiveContext`
-      Main entry point for accessing data stored in Apache Hive.
     - :class:`pyspark.sql.GroupedData`
       Aggregation methods, returned by :func:`DataFrame.groupBy`.
     - :class:`pyspark.sql.DataFrameNaFunctions`
@@ -45,7 +43,7 @@ from __future__ import absolute_import
 
 
 from pyspark.sql.types import Row
-from pyspark.sql.context import SQLContext, HiveContext
+from pyspark.sql.context import SQLContext, HiveContext, UDFRegistration
 from pyspark.sql.session import SparkSession
 from pyspark.sql.column import Column
 from pyspark.sql.dataframe import DataFrame, DataFrameNaFunctions, DataFrameStatFunctions
@@ -55,7 +53,8 @@ from pyspark.sql.window import Window, WindowSpec
 
 
 __all__ = [
-    'SparkSession', 'SQLContext', 'HiveContext', 'DataFrame', 'GroupedData', 'Column',
-    'Row', 'DataFrameNaFunctions', 'DataFrameStatFunctions', 'Window', 'WindowSpec',
+    'SparkSession', 'SQLContext', 'HiveContext', 'UDFRegistration',
+    'DataFrame', 'GroupedData', 'Column', 'Row',
+    'DataFrameNaFunctions', 'DataFrameStatFunctions', 'Window', 'WindowSpec',
     'DataFrameReader', 'DataFrameWriter'
 ]