diff --git a/core/src/main/scala/spark/RDD.scala b/core/src/main/scala/spark/RDD.scala
index 68e39652f01b24121b97d2122dee5321e883170e..47bdb09986f67618008c1a602e74b1761be76c26 100644
--- a/core/src/main/scala/spark/RDD.scala
+++ b/core/src/main/scala/spark/RDD.scala
@@ -283,32 +283,6 @@ extends RDD[Array[T]](prev.context) {
         (k, (vs.asInstanceOf[Seq[V]], w1s.asInstanceOf[Seq[W1]], w2s.asInstanceOf[Seq[W2]]))
     }
   }
-  
-  /*
-  def groupWith[W](other: RDD[(K, W)]): RDD[(K, (Seq[V], Seq[W]))] = {
-    if (self.partitioner != None) {
-      val part = self.partitoner.get
-      if (other.partitioner != None && other.partitioner.get == part) {
-        // Can do a partition-wise cogroup
-        return new PartitionWiseGroupedRDD(self, other)
-      }
-    }
-    
-    val vs: RDD[(K, Either[V, W])] = self.map { case (k, v) => (k, Left(v)) }
-    val ws: RDD[(K, Either[V, W])] = other.map { case (k, w) => (k, Right(w)) }
-    (vs ++ ws).groupByKey(numSplits).flatMap {
-      case (k, seq) => {
-        val vbuf = new ArrayBuffer[V]
-        val wbuf = new ArrayBuffer[W]
-        seq.foreach(_ match {
-          case Left(v) => vbuf += v
-          case Right(w) => wbuf += w
-        })
-        for (v <- vbuf; w <- wbuf) yield (k, (v, w))
-      }
-    }
-  }
-  */
 }
 
 class MappedValuesRDD[K, V, U](