diff --git a/src/scala/spark/Broadcast.scala b/src/scala/spark/Broadcast.scala index afff500bb07dc1a7400a362851a79d87b70a6196..b4e896f096a4cdc7784f689fdb1d7d5f88fb8931 100644 --- a/src/scala/spark/Broadcast.scala +++ b/src/scala/spark/Broadcast.scala @@ -67,7 +67,7 @@ extends Logging { @serializable case class SourceInfo (val hostAddress: String, val listenPort: Int, - val totalBlocks: Int, val totalBytes: Int, val replicaID: Int) + val totalBlocks: Int, val totalBytes: Int) extends Comparable [SourceInfo] with Logging { var currentLeechers = 0 diff --git a/src/scala/spark/ChainedBroadcast.scala b/src/scala/spark/ChainedBroadcast.scala index 3fc6a0fc9076d65a0844d898048a2483ccc60677..ddeb93a8da7a6e216ff88994836917ebcc21c133 100644 --- a/src/scala/spark/ChainedBroadcast.scala +++ b/src/scala/spark/ChainedBroadcast.scala @@ -1,7 +1,5 @@ package spark -import com.google.common.collect.MapMaker - import java.io._ import java.net._ import java.util.{Comparator, PriorityQueue, Random, UUID} @@ -84,7 +82,7 @@ extends Broadcast with Logging { pqOfSources = new PriorityQueue[SourceInfo] val masterSource_0 = - SourceInfo (hostAddress, listenPort, totalBlocks, totalBytes, 0) + SourceInfo (hostAddress, listenPort, totalBlocks, totalBytes) pqOfSources.add (masterSource_0) // Register with the Tracker @@ -288,7 +286,7 @@ extends Broadcast with Logging { logInfo ("Connected to Master's guiding object") // Send local source information - oosMaster.writeObject(SourceInfo (hostAddress, listenPort, -1, -1, 0)) + oosMaster.writeObject(SourceInfo (hostAddress, listenPort, -1, -1)) oosMaster.flush // Receive source information from Master @@ -520,7 +518,7 @@ extends Broadcast with Logging { // Add this new (if it can finish) source to the PQ of sources thisWorkerInfo = SourceInfo (sourceInfo.hostAddress, - sourceInfo.listenPort, totalBlocks, totalBytes, 0) + sourceInfo.listenPort, totalBlocks, totalBytes) logInfo ("Adding possible new source to pqOfSources: " + thisWorkerInfo) pqOfSources.add (thisWorkerInfo) } @@ -713,7 +711,7 @@ extends Broadcast with Logging { private object ChainedBroadcast extends Logging { - val values = new MapMaker ().softValues ().makeMap[UUID, Any] + val values = Cache.newKeySpace() var valueToGuidePortMap = Map[UUID, Int] () diff --git a/src/scala/spark/DfsBroadcast.scala b/src/scala/spark/DfsBroadcast.scala index 5be5f98e8cad561355045b5963aca2b712b1f2be..a249961fd5c560caf1ebc19c4a623bff935f0060 100644 --- a/src/scala/spark/DfsBroadcast.scala +++ b/src/scala/spark/DfsBroadcast.scala @@ -1,7 +1,5 @@ package spark -import com.google.common.collect.MapMaker - import java.io._ import java.net._ import java.util.UUID @@ -56,7 +54,7 @@ extends Broadcast with Logging { private object DfsBroadcast extends Logging { - val values = new MapMaker ().softValues ().makeMap[UUID, Any] + val values = Cache.newKeySpace() private var initialized = false