diff --git a/hbase-spark/pom.xml b/hbase-spark/pom.xml index 251ea596e47..7c7590e22b6 100644 --- a/hbase-spark/pom.xml +++ b/hbase-spark/pom.xml @@ -37,7 +37,7 @@ Apache HBase - Spark - 1.3.0 + 1.6.0 2.10.4 2.10 true diff --git a/hbase-spark/src/test/scala/org/apache/hadoop/hbase/spark/DefaultSourceSuite.scala b/hbase-spark/src/test/scala/org/apache/hadoop/hbase/spark/DefaultSourceSuite.scala index 30ddfc499c0..04dd9baf7b2 100644 --- a/hbase-spark/src/test/scala/org/apache/hadoop/hbase/spark/DefaultSourceSuite.scala +++ b/hbase-spark/src/test/scala/org/apache/hadoop/hbase/spark/DefaultSourceSuite.scala @@ -20,7 +20,7 @@ package org.apache.hadoop.hbase.spark import org.apache.hadoop.hbase.client.{Put, ConnectionFactory} import org.apache.hadoop.hbase.spark.datasources.HBaseSparkConf import org.apache.hadoop.hbase.util.Bytes -import org.apache.hadoop.hbase.{TableNotFoundException, TableName, HBaseTestingUtility} +import org.apache.hadoop.hbase.{TableName, HBaseTestingUtility} import org.apache.spark.sql.{DataFrame, SQLContext} import org.apache.spark.{SparkConf, SparkContext, Logging} import org.scalatest.{BeforeAndAfterAll, BeforeAndAfterEach, FunSuite} @@ -514,7 +514,7 @@ BeforeAndAfterEach with BeforeAndAfterAll with Logging { test("Test table that doesn't exist") { - intercept[TableNotFoundException] { + intercept[Exception] { df = sqlContext.load("org.apache.hadoop.hbase.spark", Map("hbase.columns.mapping" -> "KEY_FIELD STRING :key, A_FIELD STRING c:a, B_FIELD STRING c:b,",