From 25a6a39e6d6db1264ab7633d1dcfe886415fbf1a Mon Sep 17 00:00:00 2001 From: Matei Zaharia <matei@eecs.berkeley.edu> Date: Sun, 19 Aug 2012 18:59:16 -0700 Subject: [PATCH] Added other SparkContext constructors to JavaSparkContext --- core/src/main/scala/spark/api/java/JavaSparkContext.scala | 6 ++++++ 1 file changed, 6 insertions(+) diff --git a/core/src/main/scala/spark/api/java/JavaSparkContext.scala b/core/src/main/scala/spark/api/java/JavaSparkContext.scala index 08c92b145e..4a7d945a8d 100644 --- a/core/src/main/scala/spark/api/java/JavaSparkContext.scala +++ b/core/src/main/scala/spark/api/java/JavaSparkContext.scala @@ -20,6 +20,12 @@ class JavaSparkContext(val sc: SparkContext) extends JavaSparkContextVarargsWork def this(master: String, frameworkName: String) = this(new SparkContext(master, frameworkName)) + def this(master: String, frameworkName: String, sparkHome: String, jarFile: String) = + this(new SparkContext(master, frameworkName, sparkHome, Seq(jarFile))) + + def this(master: String, frameworkName: String, sparkHome: String, jars: Array[String]) = + this(new SparkContext(master, frameworkName, sparkHome, jars.toSeq)) + val env = sc.env def parallelize[T](list: java.util.List[T], numSlices: Int): JavaRDD[T] = { -- GitLab