From a3d5f922109caa878f8350fe0634514b8af55cbc Mon Sep 17 00:00:00 2001
From: Ethan Jewett <esjewett@gmail.com>
Date: Tue, 7 May 2013 11:43:06 -0500
Subject: [PATCH] Switch to using SparkContext method to create RDD

---
 examples/src/main/scala/spark/examples/HBaseTest.scala | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)

diff --git a/examples/src/main/scala/spark/examples/HBaseTest.scala b/examples/src/main/scala/spark/examples/HBaseTest.scala
index d94b25828d..9bad876860 100644
--- a/examples/src/main/scala/spark/examples/HBaseTest.scala
+++ b/examples/src/main/scala/spark/examples/HBaseTest.scala
@@ -24,9 +24,9 @@ object HBaseTest {
       admin.createTable(tableDesc)
     }
 
-    val hBaseRDD = new NewHadoopRDD(sc, classOf[TableInputFormat], 
+    val hBaseRDD = sc.newAPIHadoopRDD(conf, classOf[TableInputFormat], 
       classOf[org.apache.hadoop.hbase.io.ImmutableBytesWritable],
-      classOf[org.apache.hadoop.hbase.client.Result], conf)
+      classOf[org.apache.hadoop.hbase.client.Result])
 
     hBaseRDD.count()
 
-- 
GitLab