HADOOP-18025. Upgrade HBase version to 1.7.1 for hbase1 profile (#3722)

This commit is contained in:
Viraj Jasani 2021-12-02 08:10:30 +05:30 committed by GitHub
parent df4197592f
commit b34dcb5b3a
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
9 changed files with 56 additions and 29 deletions

View File

@ -297,10 +297,10 @@ org.apache.curator:curator-client:5.2.0
org.apache.curator:curator-framework:5.2.0
org.apache.curator:curator-recipes:5.2.0
org.apache.geronimo.specs:geronimo-jcache_1.0_spec:1.0-alpha-1
org.apache.hbase:hbase-annotations:1.4.8
org.apache.hbase:hbase-client:1.4.8
org.apache.hbase:hbase-common:1.4.8
org.apache.hbase:hbase-protocol:1.4.8
org.apache.hbase:hbase-annotations:1.7.1
org.apache.hbase:hbase-client:1.7.1
org.apache.hbase:hbase-common:1.7.1
org.apache.hbase:hbase-protocol:1.7.1
org.apache.htrace:htrace-core:3.1.0-incubating
org.apache.htrace:htrace-core4:4.1.0-incubating
org.apache.httpcomponents:httpclient:4.5.6

View File

@ -196,7 +196,7 @@
<swagger-annotations-version>1.5.4</swagger-annotations-version>
<snakeyaml.version>1.26</snakeyaml.version>
<hbase.one.version>1.4.8</hbase.one.version>
<hbase.one.version>1.7.1</hbase.one.version>
<hbase.two.version>2.0.2</hbase.two.version>
<junit.version>4.13.2</junit.version>
<junit.jupiter.version>5.5.1</junit.jupiter.version>
@ -2393,7 +2393,7 @@
</activation>
<properties>
<hbase.version>${hbase.one.version}</hbase.version>
<hbase-compatible-hadoop.version>2.5.1</hbase-compatible-hadoop.version>
<hbase-compatible-hadoop.version>2.8.5</hbase-compatible-hadoop.version>
<hbase-compatible-guava.version>12.0.1</hbase-compatible-guava.version>
<hbase-server-artifactid>hadoop-yarn-server-timelineservice-hbase-server-1</hbase-server-artifactid>
</properties>

View File

@ -96,6 +96,10 @@
<groupId>tomcat</groupId>
<artifactId>jasper-runtime</artifactId>
</exclusion>
<exclusion>
<groupId>org.apache.hadoop</groupId>
<artifactId>hadoop-hdfs-client</artifactId>
</exclusion>
</exclusions>
</dependency>
@ -106,6 +110,12 @@
<artifactId>hadoop-auth</artifactId>
<version>${hbase-compatible-hadoop.version}</version>
<scope>test</scope>
<exclusions>
<exclusion>
<groupId>org.apache.hadoop</groupId>
<artifactId>hadoop-hdfs-client</artifactId>
</exclusion>
</exclusions>
</dependency>
<dependency>
@ -117,6 +127,10 @@
<groupId>org.apache.hadoop</groupId>
<artifactId>hadoop-common</artifactId>
</exclusion>
<exclusion>
<groupId>org.apache.hadoop</groupId>
<artifactId>hadoop-hdfs-client</artifactId>
</exclusion>
</exclusions>
</dependency>
@ -311,6 +325,12 @@
<artifactId>hadoop-hdfs</artifactId>
<version>${hbase-compatible-hadoop.version}</version>
<scope>test</scope>
<exclusions>
<exclusion>
<groupId>org.apache.hadoop</groupId>
<artifactId>hadoop-hdfs-client</artifactId>
</exclusion>
</exclusions>
</dependency>
<!-- 'mvn dependency:analyze' fails to detect use of this direct
@ -321,6 +341,19 @@
<version>${hbase-compatible-hadoop.version}</version>
<type>test-jar</type>
<scope>test</scope>
<exclusions>
<exclusion>
<groupId>org.apache.hadoop</groupId>
<artifactId>hadoop-hdfs-client</artifactId>
</exclusion>
</exclusions>
</dependency>
<dependency>
<groupId>org.apache.hadoop</groupId>
<artifactId>hadoop-hdfs-client</artifactId>
<version>${hbase-compatible-hadoop.version}</version>
<scope>test</scope>
</dependency>
<dependency>
@ -470,14 +503,6 @@
</exclusion>
</exclusions>
</dependency>
<!-- 'mvn dependency:analyze' fails to detect use of this direct
dependency -->
<dependency>
<groupId>org.apache.hadoop</groupId>
<artifactId>hadoop-hdfs-client</artifactId>
<version>${hbase-compatible-hadoop.version}</version>
<scope>test</scope>
</dependency>
<!-- 'mvn dependency:analyze' fails to detect use of this direct
dependency -->
<!-- This is needed by HBaseTestingUtility -->

View File

@ -412,7 +412,7 @@ class ApplicationEntityReader extends GenericEntityReader {
}
// set start row
scan.setStartRow(applicationRowKey.getRowKey());
scan.withStartRow(applicationRowKey.getRowKey());
// get the bytes for stop row
applicationRowKeyPrefix = new ApplicationRowKeyPrefix(
@ -420,7 +420,7 @@ class ApplicationEntityReader extends GenericEntityReader {
context.getFlowRunId());
// set stop row
scan.setStopRow(
scan.withStopRow(
HBaseTimelineStorageUtils.calculateTheClosestNextRowKeyForPrefix(
applicationRowKeyPrefix.getRowKeyPrefix()));
}

View File

@ -158,9 +158,11 @@ public final class EntityTypeReader extends AbstractTimelineStorageReader {
private ResultScanner getResult(Configuration hbaseConf, Connection conn,
FilterList filterList, byte[] startPrefix, byte[] endPrefix)
throws IOException {
Scan scan = new Scan(startPrefix, endPrefix);
scan.setFilter(filterList);
scan.setSmall(true);
Scan scan = new Scan()
.withStartRow(startPrefix)
.withStopRow(endPrefix)
.setFilter(filterList)
.setSmall(true);
return ENTITY_TABLE.getResultScanner(hbaseConf, conn, scan);
}

View File

@ -133,16 +133,16 @@ class FlowActivityEntityReader extends TimelineEntityReader {
throw new BadRequestException(
"fromid doesn't belong to clusterId=" + clusterId);
}
scan.setStartRow(key.getRowKey());
scan.setStopRow(
scan.withStartRow(key.getRowKey());
scan.withStopRow(
new FlowActivityRowKeyPrefix(clusterId,
(getFilters().getCreatedTimeBegin() <= 0 ? 0
: (getFilters().getCreatedTimeBegin() - 1)))
.getRowKeyPrefix());
} else {
scan.setStartRow(new FlowActivityRowKeyPrefix(clusterId, getFilters()
scan.withStartRow(new FlowActivityRowKeyPrefix(clusterId, getFilters()
.getCreatedTimeEnd()).getRowKeyPrefix());
scan.setStopRow(new FlowActivityRowKeyPrefix(clusterId, (getFilters()
scan.withStopRow(new FlowActivityRowKeyPrefix(clusterId, (getFilters()
.getCreatedTimeBegin() <= 0 ? 0
: (getFilters().getCreatedTimeBegin() - 1))).getRowKeyPrefix());
}

View File

@ -241,14 +241,14 @@ class FlowRunEntityReader extends TimelineEntityReader {
"fromid doesn't belong to clusterId=" + context.getClusterId());
}
// set start row
scan.setStartRow(flowRunRowKey.getRowKey());
scan.withStartRow(flowRunRowKey.getRowKey());
// get the bytes for stop row
flowRunRowKeyPrefix = new FlowRunRowKeyPrefix(context.getClusterId(),
context.getUserId(), context.getFlowName());
// set stop row
scan.setStopRow(
scan.withStopRow(
HBaseTimelineStorageUtils.calculateTheClosestNextRowKeyForPrefix(
flowRunRowKeyPrefix.getRowKeyPrefix()));
}

View File

@ -519,7 +519,7 @@ class GenericEntityReader extends TimelineEntityReader {
}
// set start row
scan.setStartRow(entityRowKey.getRowKey());
scan.withStartRow(entityRowKey.getRowKey());
// get the bytes for stop row
entityRowKeyPrefix = new EntityRowKeyPrefix(context.getClusterId(),
@ -527,7 +527,7 @@ class GenericEntityReader extends TimelineEntityReader {
context.getAppId(), context.getEntityType(), null, null);
// set stop row
scan.setStopRow(
scan.withStopRow(
HBaseTimelineStorageUtils.calculateTheClosestNextRowKeyForPrefix(
entityRowKeyPrefix.getRowKeyPrefix()));

View File

@ -372,7 +372,7 @@ class SubApplicationEntityReader extends GenericEntityReader {
}
// set start row
scan.setStartRow(entityRowKey.getRowKey());
scan.withStartRow(entityRowKey.getRowKey());
// get the bytes for stop row
subApplicationRowKeyPrefix = new SubApplicationRowKeyPrefix(
@ -380,7 +380,7 @@ class SubApplicationEntityReader extends GenericEntityReader {
context.getEntityType(), null, null, null);
// set stop row
scan.setStopRow(
scan.withStopRow(
HBaseTimelineStorageUtils.calculateTheClosestNextRowKeyForPrefix(
subApplicationRowKeyPrefix.getRowKeyPrefix()));