From 25a6a39e6d6db1264ab7633d1dcfe886415fbf1a Mon Sep 17 00:00:00 2001
From: Matei Zaharia <matei@eecs.berkeley.edu>
Date: Sun, 19 Aug 2012 18:59:16 -0700
Subject: [PATCH] Added other SparkContext constructors to JavaSparkContext

---
 core/src/main/scala/spark/api/java/JavaSparkContext.scala | 6 ++++++
 1 file changed, 6 insertions(+)

diff --git a/core/src/main/scala/spark/api/java/JavaSparkContext.scala b/core/src/main/scala/spark/api/java/JavaSparkContext.scala
index 08c92b145e..4a7d945a8d 100644
--- a/core/src/main/scala/spark/api/java/JavaSparkContext.scala
+++ b/core/src/main/scala/spark/api/java/JavaSparkContext.scala
@@ -20,6 +20,12 @@ class JavaSparkContext(val sc: SparkContext) extends JavaSparkContextVarargsWork
 
   def this(master: String, frameworkName: String) = this(new SparkContext(master, frameworkName))
 
+  def this(master: String, frameworkName: String, sparkHome: String, jarFile: String) =
+    this(new SparkContext(master, frameworkName, sparkHome, Seq(jarFile)))
+
+  def this(master: String, frameworkName: String, sparkHome: String, jars: Array[String]) =
+    this(new SparkContext(master, frameworkName, sparkHome, jars.toSeq))
+
   val env = sc.env
 
   def parallelize[T](list: java.util.List[T], numSlices: Int): JavaRDD[T] = {
-- 
GitLab