diff --git a/common/network-shuffle/src/main/java/org/apache/spark/network/shuffle/ExternalShuffleBlockHandler.java b/common/network-shuffle/src/main/java/org/apache/spark/network/shuffle/ExternalShuffleBlockHandler.java
index 1cc0fb65d726f101d777f2d3905285c79b642a44..1270cef621b7dfe9fb9fbe45f0b60615109306a4 100644
--- a/common/network-shuffle/src/main/java/org/apache/spark/network/shuffle/ExternalShuffleBlockHandler.java
+++ b/common/network-shuffle/src/main/java/org/apache/spark/network/shuffle/ExternalShuffleBlockHandler.java
@@ -113,7 +113,8 @@ public class ExternalShuffleBlockHandler extends RpcHandler {
       }
 
     } else if (msgObj instanceof RegisterExecutor) {
-      final Timer.Context responseDelayContext = metrics.registerExecutorRequestLatencyMillis.time();
+      final Timer.Context responseDelayContext =
+        metrics.registerExecutorRequestLatencyMillis.time();
       try {
         RegisterExecutor msg = (RegisterExecutor) msgObj;
         checkAuth(client, msg.appId);
diff --git a/common/network-yarn/src/main/java/org/apache/spark/network/yarn/YarnShuffleService.java b/common/network-yarn/src/main/java/org/apache/spark/network/yarn/YarnShuffleService.java
index df17dacdefaafa269934d949e963e7d787f6b0a6..22e47ac1b7923282df67959bef112f35bdf932e3 100644
--- a/common/network-yarn/src/main/java/org/apache/spark/network/yarn/YarnShuffleService.java
+++ b/common/network-yarn/src/main/java/org/apache/spark/network/yarn/YarnShuffleService.java
@@ -131,7 +131,7 @@ public class YarnShuffleService extends AuxiliaryService {
 
     try {
       // In case this NM was killed while there were running spark applications, we need to restore
-      // lost state for the existing executors.  We look for an existing file in the NM's local dirs.
+      // lost state for the existing executors. We look for an existing file in the NM's local dirs.
       // If we don't find one, then we choose a file to use to save the state next time.  Even if
       // an application was stopped while the NM was down, we expect yarn to call stopApplication()
       // when it comes back
diff --git a/examples/src/main/java/org/apache/spark/examples/sql/JavaSQLDataSourceExample.java b/examples/src/main/java/org/apache/spark/examples/sql/JavaSQLDataSourceExample.java
index 2b94b9f114e2d934a0f46d10b9835dce0370a1d1..ec02c8bbb8efeb8a61b03d8baa7a2fc981f489a5 100644
--- a/examples/src/main/java/org/apache/spark/examples/sql/JavaSQLDataSourceExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/sql/JavaSQLDataSourceExample.java
@@ -110,11 +110,13 @@ public class JavaSQLDataSourceExample {
     usersDF.select("name", "favorite_color").write().save("namesAndFavColors.parquet");
     // $example off:generic_load_save_functions$
     // $example on:manual_load_options$
-    Dataset<Row> peopleDF = spark.read().format("json").load("examples/src/main/resources/people.json");
+    Dataset<Row> peopleDF =
+      spark.read().format("json").load("examples/src/main/resources/people.json");
     peopleDF.select("name", "age").write().format("parquet").save("namesAndAges.parquet");
     // $example off:manual_load_options$
     // $example on:direct_sql$
-    Dataset<Row> sqlDF = spark.sql("SELECT * FROM parquet.`examples/src/main/resources/users.parquet`");
+    Dataset<Row> sqlDF =
+      spark.sql("SELECT * FROM parquet.`examples/src/main/resources/users.parquet`");
     // $example off:direct_sql$
   }