diff --git a/sql/hive-thriftserver/src/main/scala/org/apache/spark/sql/hive/thriftserver/SparkSQLCLIDriver.scala b/sql/hive-thriftserver/src/main/scala/org/apache/spark/sql/hive/thriftserver/SparkSQLCLIDriver.scala
index b092f42372171b0004ec5a5bfcbe38ce8edc079e..7ba4564602ecd3b3b3208d6c3a25b93ebc4b262f 100755
--- a/sql/hive-thriftserver/src/main/scala/org/apache/spark/sql/hive/thriftserver/SparkSQLCLIDriver.scala
+++ b/sql/hive-thriftserver/src/main/scala/org/apache/spark/sql/hive/thriftserver/SparkSQLCLIDriver.scala
@@ -73,18 +73,6 @@ private[hive] object SparkSQLCLIDriver {
       System.exit(1)
     }
 
-    // NOTE: It is critical to do this here so that log4j is reinitialized
-    // before any of the other core hive classes are loaded
-    var logInitFailed = false
-    var logInitDetailMessage: String = null
-    try {
-      logInitDetailMessage = LogUtils.initHiveLog4j()
-    } catch {
-      case e: LogInitializationException =>
-        logInitFailed = true
-        logInitDetailMessage = e.getMessage
-    }
-
     val sessionState = new CliSessionState(new HiveConf(classOf[SessionState]))
 
     sessionState.in = System.in
@@ -100,11 +88,6 @@ private[hive] object SparkSQLCLIDriver {
       System.exit(2)
     }
 
-    if (!sessionState.getIsSilent) {
-      if (logInitFailed) System.err.println(logInitDetailMessage)
-      else SessionState.getConsole.printInfo(logInitDetailMessage)
-    }
-
     // Set all properties specified via command line.
     val conf: HiveConf = sessionState.getConf
     sessionState.cmdProperties.entrySet().foreach { item: java.util.Map.Entry[Object, Object] =>