From f5041579ff573f988b673c2506fa4edc32f5ad84 Mon Sep 17 00:00:00 2001 From: Sean Owen <sowen@cloudera.com> Date: Sun, 4 May 2014 17:43:28 -0700 Subject: [PATCH] SPARK-1629. Addendum: Depend on commons lang3 (already used by tachyon) as it's used in ReplSuite, and return to use lang3 utility in Utils.scala For consideration. This was proposed in related discussion: https://github.com/apache/spark/pull/569 Author: Sean Owen <sowen@cloudera.com> Closes #635 from srowen/SPARK-1629.2 and squashes the following commits: a442b98 [Sean Owen] Depend on commons lang3 (already used by tachyon) as it's used in ReplSuite, and return to use lang3 utility in Utils.scala --- core/pom.xml | 4 ++++ core/src/main/scala/org/apache/spark/util/Utils.scala | 5 ++--- pom.xml | 5 +++++ project/SparkBuild.scala | 1 + 4 files changed, 12 insertions(+), 3 deletions(-) diff --git a/core/pom.xml b/core/pom.xml index 822b5b1dd7..36c71e67b5 100644 --- a/core/pom.xml +++ b/core/pom.xml @@ -69,6 +69,10 @@ <groupId>com.google.guava</groupId> <artifactId>guava</artifactId> </dependency> + <dependency> + <groupId>org.apache.commons</groupId> + <artifactId>commons-lang3</artifactId> + </dependency> <dependency> <groupId>com.google.code.findbugs</groupId> <artifactId>jsr305</artifactId> diff --git a/core/src/main/scala/org/apache/spark/util/Utils.scala b/core/src/main/scala/org/apache/spark/util/Utils.scala index acd7eef6d2..bef4dab3d7 100644 --- a/core/src/main/scala/org/apache/spark/util/Utils.scala +++ b/core/src/main/scala/org/apache/spark/util/Utils.scala @@ -32,6 +32,7 @@ import scala.util.Try import com.google.common.io.Files import com.google.common.util.concurrent.ThreadFactoryBuilder +import org.apache.commons.lang3.SystemUtils import org.apache.hadoop.fs.{FileSystem, FileUtil, Path} import org.json4s._ import tachyon.client.{TachyonFile,TachyonFS} @@ -1073,9 +1074,7 @@ private[spark] object Utils extends Logging { /** * Return true if this is Windows. */ - def isWindows = { - Option(System.getProperty("os.name")).exists(_.startsWith("Windows")) - } + def isWindows = SystemUtils.IS_OS_WINDOWS /** * Indicates whether Spark is currently running unit tests. diff --git a/pom.xml b/pom.xml index ebd359a9de..7678043484 100644 --- a/pom.xml +++ b/pom.xml @@ -221,6 +221,11 @@ <artifactId>guava</artifactId> <version>14.0.1</version> </dependency> + <dependency> + <groupId>org.apache.commons</groupId> + <artifactId>commons-lang3</artifactId> + <version>3.3.2</version> + </dependency> <dependency> <groupId>com.google.code.findbugs</groupId> <artifactId>jsr305</artifactId> diff --git a/project/SparkBuild.scala b/project/SparkBuild.scala index 51f7335111..19aa3c0607 100644 --- a/project/SparkBuild.scala +++ b/project/SparkBuild.scala @@ -326,6 +326,7 @@ object SparkBuild extends Build { name := "spark-core", libraryDependencies ++= Seq( "com.google.guava" % "guava" % "14.0.1", + "org.apache.commons" % "commons-lang3" % "3.3.2", "com.google.code.findbugs" % "jsr305" % "1.3.9", "log4j" % "log4j" % "1.2.17", "org.slf4j" % "slf4j-api" % slf4jVersion, -- GitLab