diff --git a/src/scala/spark/Broadcast.scala b/src/scala/spark/Broadcast.scala
index afff500bb07dc1a7400a362851a79d87b70a6196..b4e896f096a4cdc7784f689fdb1d7d5f88fb8931 100644
--- a/src/scala/spark/Broadcast.scala
+++ b/src/scala/spark/Broadcast.scala
@@ -67,7 +67,7 @@ extends Logging {
 
 @serializable
 case class SourceInfo (val hostAddress: String, val listenPort: Int, 
-  val totalBlocks: Int, val totalBytes: Int, val replicaID: Int)  
+  val totalBlocks: Int, val totalBytes: Int)  
 extends Comparable [SourceInfo] with Logging {
 
   var currentLeechers = 0
diff --git a/src/scala/spark/ChainedBroadcast.scala b/src/scala/spark/ChainedBroadcast.scala
index 3fc6a0fc9076d65a0844d898048a2483ccc60677..ddeb93a8da7a6e216ff88994836917ebcc21c133 100644
--- a/src/scala/spark/ChainedBroadcast.scala
+++ b/src/scala/spark/ChainedBroadcast.scala
@@ -1,7 +1,5 @@
 package spark
 
-import com.google.common.collect.MapMaker
-
 import java.io._
 import java.net._
 import java.util.{Comparator, PriorityQueue, Random, UUID}
@@ -84,7 +82,7 @@ extends Broadcast with Logging {
 
     pqOfSources = new PriorityQueue[SourceInfo]
     val masterSource_0 = 
-      SourceInfo (hostAddress, listenPort, totalBlocks, totalBytes, 0) 
+      SourceInfo (hostAddress, listenPort, totalBlocks, totalBytes) 
     pqOfSources.add (masterSource_0)
 
     // Register with the Tracker
@@ -288,7 +286,7 @@ extends Broadcast with Logging {
       logInfo ("Connected to Master's guiding object")
 
       // Send local source information
-      oosMaster.writeObject(SourceInfo (hostAddress, listenPort, -1, -1, 0))
+      oosMaster.writeObject(SourceInfo (hostAddress, listenPort, -1, -1))
       oosMaster.flush
 
       // Receive source information from Master        
@@ -520,7 +518,7 @@ extends Broadcast with Logging {
 
             // Add this new (if it can finish) source to the PQ of sources
             thisWorkerInfo = SourceInfo (sourceInfo.hostAddress, 
-              sourceInfo.listenPort, totalBlocks, totalBytes, 0)
+              sourceInfo.listenPort, totalBlocks, totalBytes)
             logInfo ("Adding possible new source to pqOfSources: " + thisWorkerInfo)    
             pqOfSources.add (thisWorkerInfo)
           }
@@ -713,7 +711,7 @@ extends Broadcast with Logging {
 
 private object ChainedBroadcast
 extends Logging {
-  val values = new MapMaker ().softValues ().makeMap[UUID, Any]
+  val values = Cache.newKeySpace()
 
   var valueToGuidePortMap = Map[UUID, Int] ()
   
diff --git a/src/scala/spark/DfsBroadcast.scala b/src/scala/spark/DfsBroadcast.scala
index 5be5f98e8cad561355045b5963aca2b712b1f2be..a249961fd5c560caf1ebc19c4a623bff935f0060 100644
--- a/src/scala/spark/DfsBroadcast.scala
+++ b/src/scala/spark/DfsBroadcast.scala
@@ -1,7 +1,5 @@
 package spark
 
-import com.google.common.collect.MapMaker
-
 import java.io._
 import java.net._
 import java.util.UUID
@@ -56,7 +54,7 @@ extends Broadcast with Logging {
 
 private object DfsBroadcast
 extends Logging {
-  val values = new MapMaker ().softValues ().makeMap[UUID, Any]
+  val values = Cache.newKeySpace()
 
   private var initialized = false