diff --git a/core/src/main/scala/org/apache/spark/deploy/SparkHadoopUtil.scala b/core/src/main/scala/org/apache/spark/deploy/SparkHadoopUtil.scala index 90c71cc6cfab79394175eb43f5af221121f9691b..671e8e4484f604d984365b3b567593ddedff64c9 100644 --- a/core/src/main/scala/org/apache/spark/deploy/SparkHadoopUtil.scala +++ b/core/src/main/scala/org/apache/spark/deploy/SparkHadoopUtil.scala @@ -25,7 +25,6 @@ import java.util.{Arrays, Comparator, Date} import scala.collection.JavaConverters._ import scala.concurrent.duration._ -import scala.language.postfixOps import scala.util.control.NonFatal import com.google.common.primitives.Longs diff --git a/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala b/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala index 2ce49ca1345f226a6857ddb06e6be020e59baeeb..dc05e764c395199517fb81cdaac95914bc70f28f 100644 --- a/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala +++ b/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala @@ -25,7 +25,6 @@ import java.util.concurrent.atomic.AtomicLong import scala.collection.mutable.ArrayBuffer import scala.collection.mutable.HashMap import scala.collection.mutable.HashSet -import scala.language.postfixOps import scala.util.Random import org.apache.spark._ diff --git a/core/src/main/scala/org/apache/spark/util/RpcUtils.scala b/core/src/main/scala/org/apache/spark/util/RpcUtils.scala index 2bb8de568e803e6e554d060da6ba3fffd16ed983..e3b588374ce1aba72fbfd12766c8c92f0d05736c 100644 --- a/core/src/main/scala/org/apache/spark/util/RpcUtils.scala +++ b/core/src/main/scala/org/apache/spark/util/RpcUtils.scala @@ -17,8 +17,6 @@ package org.apache.spark.util -import scala.language.postfixOps - import org.apache.spark.SparkConf import org.apache.spark.rpc.{RpcAddress, RpcEndpointRef, RpcEnv, RpcTimeout} diff --git a/core/src/test/scala/org/apache/spark/HeartbeatReceiverSuite.scala b/core/src/test/scala/org/apache/spark/HeartbeatReceiverSuite.scala index 5e2ba311ee7732ce08d3176e527f6d2e87d522b8..5f59c176ab78265f6819b16a0bf7b48e7a119e0a 100644 --- a/core/src/test/scala/org/apache/spark/HeartbeatReceiverSuite.scala +++ b/core/src/test/scala/org/apache/spark/HeartbeatReceiverSuite.scala @@ -22,7 +22,6 @@ import java.util.concurrent.{ExecutorService, TimeUnit} import scala.collection.Map import scala.collection.mutable import scala.concurrent.duration._ -import scala.language.postfixOps import org.mockito.Matchers import org.mockito.Matchers._ diff --git a/core/src/test/scala/org/apache/spark/deploy/history/ApplicationCacheSuite.scala b/core/src/test/scala/org/apache/spark/deploy/history/ApplicationCacheSuite.scala index 4ab000b53ad10209b1d059f9c2ceb8fc07264665..e3304be792af75c040c3169d3ab2bfa1c7d14d99 100644 --- a/core/src/test/scala/org/apache/spark/deploy/history/ApplicationCacheSuite.scala +++ b/core/src/test/scala/org/apache/spark/deploy/history/ApplicationCacheSuite.scala @@ -23,7 +23,6 @@ import javax.servlet.http.{HttpServletRequest, HttpServletResponse} import scala.collection.mutable import scala.collection.mutable.ListBuffer -import scala.language.postfixOps import com.codahale.metrics.Counter import com.google.common.cache.LoadingCache diff --git a/core/src/test/scala/org/apache/spark/rdd/PipedRDDSuite.scala b/core/src/test/scala/org/apache/spark/rdd/PipedRDDSuite.scala index 59b90974ae8a91bcd29830f6af7e60d95e9cac5d..387f3e2502c51613c70c5219c3770fded57801b9 100644 --- a/core/src/test/scala/org/apache/spark/rdd/PipedRDDSuite.scala +++ b/core/src/test/scala/org/apache/spark/rdd/PipedRDDSuite.scala @@ -21,7 +21,6 @@ import java.io.File import scala.collection.Map import scala.io.Codec -import scala.language.postfixOps import scala.sys.process._ import scala.util.Try @@ -215,7 +214,8 @@ class PipedRDDSuite extends SparkFunSuite with SharedSparkContext { } def testCommandAvailable(command: String): Boolean = { - Try(Process(command) !!).isSuccess + val attempt = Try(Process(command).run().exitValue()) + attempt.isSuccess && attempt.get == 0 } def testExportInputFile(varName: String) { diff --git a/core/src/test/scala/org/apache/spark/storage/MemoryStoreSuite.scala b/core/src/test/scala/org/apache/spark/storage/MemoryStoreSuite.scala index 145d432afe85e21a769b958f0942ab31a45f086c..c11de826677e089502e30cb64cad25bc561bf3cf 100644 --- a/core/src/test/scala/org/apache/spark/storage/MemoryStoreSuite.scala +++ b/core/src/test/scala/org/apache/spark/storage/MemoryStoreSuite.scala @@ -20,7 +20,6 @@ package org.apache.spark.storage import java.nio.ByteBuffer import scala.language.implicitConversions -import scala.language.postfixOps import scala.language.reflectiveCalls import scala.reflect.ClassTag diff --git a/external/kafka-0-10/src/main/scala/org/apache/spark/streaming/kafka010/KafkaTestUtils.scala b/external/kafka-0-10/src/main/scala/org/apache/spark/streaming/kafka010/KafkaTestUtils.scala index ecabe1c365b41efa97a20ded33efb758c624394b..e73823e89883b3841682bdee15e9a4e9ef3b268a 100644 --- a/external/kafka-0-10/src/main/scala/org/apache/spark/streaming/kafka010/KafkaTestUtils.scala +++ b/external/kafka-0-10/src/main/scala/org/apache/spark/streaming/kafka010/KafkaTestUtils.scala @@ -25,7 +25,6 @@ import java.util.concurrent.TimeoutException import scala.annotation.tailrec import scala.collection.JavaConverters._ -import scala.language.postfixOps import scala.util.control.NonFatal import kafka.admin.AdminUtils @@ -279,4 +278,3 @@ private[kafka010] class KafkaTestUtils extends Logging { } } } - diff --git a/external/kafka-0-8/src/main/scala/org/apache/spark/streaming/kafka/KafkaTestUtils.scala b/external/kafka-0-8/src/main/scala/org/apache/spark/streaming/kafka/KafkaTestUtils.scala index abfd7aad4c5c64c48cf708d8bc647f7d8571aae7..03c9ca7524e5d61d3706c9efc717e36cc496bb11 100644 --- a/external/kafka-0-8/src/main/scala/org/apache/spark/streaming/kafka/KafkaTestUtils.scala +++ b/external/kafka-0-8/src/main/scala/org/apache/spark/streaming/kafka/KafkaTestUtils.scala @@ -25,7 +25,6 @@ import java.util.concurrent.TimeoutException import scala.annotation.tailrec import scala.collection.JavaConverters._ -import scala.language.postfixOps import scala.util.control.NonFatal import kafka.admin.AdminUtils @@ -274,4 +273,3 @@ private[kafka] class KafkaTestUtils extends Logging { } } } - diff --git a/graphx/src/main/scala/org/apache/spark/graphx/lib/PageRank.scala b/graphx/src/main/scala/org/apache/spark/graphx/lib/PageRank.scala index 0a1622bca0f4b05bebce585fe49cb8bd6aa32daf..2f5bd4ed4ff6b9dd6e17044745e1c63d50f14e88 100644 --- a/graphx/src/main/scala/org/apache/spark/graphx/lib/PageRank.scala +++ b/graphx/src/main/scala/org/apache/spark/graphx/lib/PageRank.scala @@ -17,7 +17,6 @@ package org.apache.spark.graphx.lib -import scala.language.postfixOps import scala.reflect.ClassTag import org.apache.spark.graphx._ @@ -109,7 +108,7 @@ object PageRank extends Logging { require(resetProb >= 0 && resetProb <= 1, s"Random reset probability must belong" + s" to [0, 1], but got ${resetProb}") - val personalized = srcId isDefined + val personalized = srcId.isDefined val src: VertexId = srcId.getOrElse(-1L) // Initialize the PageRank graph with each edge attribute having diff --git a/mllib/src/main/scala/org/apache/spark/mllib/util/MFDataGenerator.scala b/mllib/src/main/scala/org/apache/spark/mllib/util/MFDataGenerator.scala index 898a09e51636c3b9eed47f7246b5c1747eccdd0f..42c5bcdd39f76692ac5a5d4a9b5eb872a105397e 100644 --- a/mllib/src/main/scala/org/apache/spark/mllib/util/MFDataGenerator.scala +++ b/mllib/src/main/scala/org/apache/spark/mllib/util/MFDataGenerator.scala @@ -19,7 +19,6 @@ package org.apache.spark.mllib.util import java.{util => ju} -import scala.language.postfixOps import scala.util.Random import org.apache.spark.SparkContext diff --git a/repl/src/test/scala/org/apache/spark/repl/ExecutorClassLoaderSuite.scala b/repl/src/test/scala/org/apache/spark/repl/ExecutorClassLoaderSuite.scala index 12e98565dcef32c95574f8ae638461d005a5ddb8..3d622d42f40863683316331facdb6d0b7c7e29d8 100644 --- a/repl/src/test/scala/org/apache/spark/repl/ExecutorClassLoaderSuite.scala +++ b/repl/src/test/scala/org/apache/spark/repl/ExecutorClassLoaderSuite.scala @@ -27,7 +27,6 @@ import java.util import scala.concurrent.duration._ import scala.io.Source import scala.language.implicitConversions -import scala.language.postfixOps import com.google.common.io.Files import org.mockito.Matchers.anyString diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/types/DecimalSuite.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/types/DecimalSuite.scala index e1675c95907af651eaa00bf179c5deea67c00601..a10c0e39eb6873e177c4e5c8ebb9d4baaf09f701 100644 --- a/sql/catalyst/src/test/scala/org/apache/spark/sql/types/DecimalSuite.scala +++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/types/DecimalSuite.scala @@ -17,8 +17,6 @@ package org.apache.spark.sql.types -import scala.language.postfixOps - import org.scalatest.PrivateMethodTester import org.apache.spark.SparkFunSuite diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DataFrameSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/DataFrameSuite.scala index 62cfd24041b3d5e8675a1c29bf1f0dc50f697b2b..499f3180379c2277ed3fb4ddbc626f45b5792bdb 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/DataFrameSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/DataFrameSuite.scala @@ -21,7 +21,6 @@ import java.io.File import java.nio.charset.StandardCharsets import java.util.UUID -import scala.language.postfixOps import scala.util.Random import org.scalatest.Matchers._ diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DatasetAggregatorSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/DatasetAggregatorSuite.scala index ddc4dcd2395b220bdc7c90d088d383bee5afa5c6..b117fbd0bcf97b208c2ab01388c8de2aa2c97416 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/DatasetAggregatorSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/DatasetAggregatorSuite.scala @@ -17,8 +17,6 @@ package org.apache.spark.sql -import scala.language.postfixOps - import org.apache.spark.sql.catalyst.encoders.ExpressionEncoder import org.apache.spark.sql.expressions.Aggregator import org.apache.spark.sql.expressions.scalalang.typed diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DatasetCacheSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/DatasetCacheSuite.scala index ac9f6c2f38537a5967853289c5e430716a2057a4..8d5e9645df89452e3ad469e10af5933452428134 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/DatasetCacheSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/DatasetCacheSuite.scala @@ -17,8 +17,6 @@ package org.apache.spark.sql -import scala.language.postfixOps - import org.apache.spark.sql.functions._ import org.apache.spark.sql.test.SharedSQLContext diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DatasetPrimitiveSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/DatasetPrimitiveSuite.scala index 6aa3d3fe808b29494d5c280b442e03953e8380d8..f8d4c61967f95ad646239cc93529517066f67f3e 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/DatasetPrimitiveSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/DatasetPrimitiveSuite.scala @@ -17,8 +17,6 @@ package org.apache.spark.sql -import scala.language.postfixOps - import org.apache.spark.sql.test.SharedSQLContext case class IntClass(value: Int) diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DatasetSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/DatasetSuite.scala index 8a756fd4749a3340382e330576c0a6b2dad56d58..88fb1472b668b67397b8dc036c49aa1bf33bc471 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/DatasetSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/DatasetSuite.scala @@ -20,8 +20,6 @@ package org.apache.spark.sql import java.io.{Externalizable, ObjectInput, ObjectOutput} import java.sql.{Date, Timestamp} -import scala.language.postfixOps - import org.apache.spark.sql.catalyst.encoders.{OuterScopes, RowEncoder} import org.apache.spark.sql.catalyst.util.sideBySide import org.apache.spark.sql.execution.streaming.MemoryStream diff --git a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/SQLQuerySuite.scala b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/SQLQuerySuite.scala index b659325a625927d6ba867bd3a47a9facc06633d3..e6fe47aa65f34b041bda74ffcab2b33c1e669a64 100644 --- a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/SQLQuerySuite.scala +++ b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/SQLQuerySuite.scala @@ -1790,6 +1790,7 @@ class SQLQuerySuite extends QueryTest with SQLTestUtils with TestHiveSingleton { } def testCommandAvailable(command: String): Boolean = { - Try(Process(command) !!).isSuccess + val attempt = Try(Process(command).run().exitValue()) + attempt.isSuccess && attempt.get == 0 } } diff --git a/streaming/src/test/scala/org/apache/spark/streaming/InputStreamsSuite.scala b/streaming/src/test/scala/org/apache/spark/streaming/InputStreamsSuite.scala index 00d506c2f18bbeaa4b25d8385f6589c2afae13e6..9ecfa48091a0eb6c5cdec0d4b15935689b2b6f3b 100644 --- a/streaming/src/test/scala/org/apache/spark/streaming/InputStreamsSuite.scala +++ b/streaming/src/test/scala/org/apache/spark/streaming/InputStreamsSuite.scala @@ -25,7 +25,6 @@ import java.util.concurrent.atomic.AtomicInteger import scala.collection.JavaConverters._ import scala.collection.mutable -import scala.language.postfixOps import com.google.common.io.Files import org.apache.hadoop.fs.Path diff --git a/yarn/src/main/scala/org/apache/spark/deploy/yarn/AMDelegationTokenRenewer.scala b/yarn/src/main/scala/org/apache/spark/deploy/yarn/AMDelegationTokenRenewer.scala index a6a4fec3ba9e9e33d0ee41acc2c89056048789fc..310a7a6b05e7810d8d1b8772c2dc026a7e4b64ee 100644 --- a/yarn/src/main/scala/org/apache/spark/deploy/yarn/AMDelegationTokenRenewer.scala +++ b/yarn/src/main/scala/org/apache/spark/deploy/yarn/AMDelegationTokenRenewer.scala @@ -19,8 +19,6 @@ package org.apache.spark.deploy.yarn import java.security.PrivilegedExceptionAction import java.util.concurrent.{Executors, TimeUnit} -import scala.language.postfixOps - import org.apache.hadoop.conf.Configuration import org.apache.hadoop.fs.{FileSystem, Path} import org.apache.hadoop.security.UserGroupInformation @@ -128,7 +126,7 @@ private[yarn] class AMDelegationTokenRenewer( try { val remoteFs = FileSystem.get(freshHadoopConf) val credentialsPath = new Path(credentialsFile) - val thresholdTime = System.currentTimeMillis() - (daysToKeepFiles days).toMillis + val thresholdTime = System.currentTimeMillis() - (daysToKeepFiles.days).toMillis hadoopUtil.listFilesSorted( remoteFs, credentialsPath.getParent, credentialsPath.getName, SparkHadoopUtil.SPARK_YARN_CREDS_TEMP_EXTENSION)