diff --git a/hbase-spark/pom.xml b/hbase-spark/pom.xml
index 251ea596e47..7c7590e22b6 100644
--- a/hbase-spark/pom.xml
+++ b/hbase-spark/pom.xml
@@ -37,7 +37,7 @@
Apache HBase - Spark
- 1.3.0
+ 1.6.0
2.10.4
2.10
true
diff --git a/hbase-spark/src/test/scala/org/apache/hadoop/hbase/spark/DefaultSourceSuite.scala b/hbase-spark/src/test/scala/org/apache/hadoop/hbase/spark/DefaultSourceSuite.scala
index 30ddfc499c0..04dd9baf7b2 100644
--- a/hbase-spark/src/test/scala/org/apache/hadoop/hbase/spark/DefaultSourceSuite.scala
+++ b/hbase-spark/src/test/scala/org/apache/hadoop/hbase/spark/DefaultSourceSuite.scala
@@ -20,7 +20,7 @@ package org.apache.hadoop.hbase.spark
import org.apache.hadoop.hbase.client.{Put, ConnectionFactory}
import org.apache.hadoop.hbase.spark.datasources.HBaseSparkConf
import org.apache.hadoop.hbase.util.Bytes
-import org.apache.hadoop.hbase.{TableNotFoundException, TableName, HBaseTestingUtility}
+import org.apache.hadoop.hbase.{TableName, HBaseTestingUtility}
import org.apache.spark.sql.{DataFrame, SQLContext}
import org.apache.spark.{SparkConf, SparkContext, Logging}
import org.scalatest.{BeforeAndAfterAll, BeforeAndAfterEach, FunSuite}
@@ -514,7 +514,7 @@ BeforeAndAfterEach with BeforeAndAfterAll with Logging {
test("Test table that doesn't exist") {
- intercept[TableNotFoundException] {
+ intercept[Exception] {
df = sqlContext.load("org.apache.hadoop.hbase.spark",
Map("hbase.columns.mapping" ->
"KEY_FIELD STRING :key, A_FIELD STRING c:a, B_FIELD STRING c:b,",