diff --git a/core/src/main/scala/org/apache/spark/deploy/Client.scala b/core/src/main/scala/org/apache/spark/deploy/Client.scala index ee276e1b71138c301dade1f1e16052b1065a27f4..a4de3d7eaf458886134b8a06e12072f80bbe8802 100644 --- a/core/src/main/scala/org/apache/spark/deploy/Client.scala +++ b/core/src/main/scala/org/apache/spark/deploy/Client.scala @@ -221,7 +221,9 @@ object Client { val conf = new SparkConf() val driverArgs = new ClientArguments(args) - conf.set("spark.rpc.askTimeout", "10") + if (!conf.contains("spark.rpc.askTimeout")) { + conf.set("spark.rpc.askTimeout", "10s") + } Logger.getRootLogger.setLevel(driverArgs.logLevel) val rpcEnv = diff --git a/docs/configuration.md b/docs/configuration.md index e0c661349caab66b5a083907f9523d723a59565a..c2329b411fc693c44d4598cb7bde4bbb8e708d67 100644 --- a/docs/configuration.md +++ b/docs/configuration.md @@ -1175,7 +1175,7 @@ Apart from these, the following properties are also available, and may be useful </tr> <tr> <td><code>spark.rpc.askTimeout</code></td> - <td>120s</td> + <td><code>spark.network.timeout</code></td> <td> Duration for an RPC ask operation to wait before timing out. </td> @@ -1531,7 +1531,7 @@ Apart from these, the following properties are also available, and may be useful </tr> <tr> <td><code>spark.core.connection.ack.wait.timeout</code></td> - <td>60s</td> + <td><code>spark.network.timeout</code></td> <td> How long for the connection to wait for ack to occur before timing out and giving up. To avoid unwilling timeout caused by long pause like GC,