diff --git a/hbase-checkstyle/src/main/resources/hbase/checkstyle-suppressions.xml b/hbase-checkstyle/src/main/resources/hbase/checkstyle-suppressions.xml
index 9feb5558775..ad79163c555 100644
--- a/hbase-checkstyle/src/main/resources/hbase/checkstyle-suppressions.xml
+++ b/hbase-checkstyle/src/main/resources/hbase/checkstyle-suppressions.xml
@@ -38,4 +38,5 @@
+
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/AbstractHBaseTool.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/AbstractHBaseTool.java
index 1dd720143ad..b4548844ea4 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/AbstractHBaseTool.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/AbstractHBaseTool.java
@@ -22,7 +22,6 @@ import java.util.Comparator;
import java.util.HashMap;
import java.util.List;
-import org.apache.hadoop.conf.Configurable;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hbase.HBaseConfiguration;
import org.apache.hadoop.util.Tool;
@@ -46,9 +45,9 @@ import org.apache.hbase.thirdparty.org.apache.commons.cli.ParseException;
* command-line argument parsing.
*/
@InterfaceAudience.Private
-public abstract class AbstractHBaseTool implements Tool, Configurable {
- protected static final int EXIT_SUCCESS = 0;
- protected static final int EXIT_FAILURE = 1;
+public abstract class AbstractHBaseTool implements Tool {
+ public static final int EXIT_SUCCESS = 0;
+ public static final int EXIT_FAILURE = 1;
public static final String SHORT_HELP_OPTION = "h";
public static final String LONG_HELP_OPTION = "help";
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BulkLoadObserver.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BulkLoadObserver.java
index 25e6522018e..b69a7270371 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BulkLoadObserver.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BulkLoadObserver.java
@@ -21,13 +21,9 @@ package org.apache.hadoop.hbase.coprocessor;
import java.io.IOException;
-import org.apache.hadoop.hbase.Coprocessor;
import org.apache.hadoop.hbase.HBaseInterfaceAudience;
-import org.apache.hadoop.hbase.TableName;
import org.apache.yetus.audience.InterfaceAudience;
import org.apache.yetus.audience.InterfaceStability;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.PrepareBulkLoadRequest;
-import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CleanupBulkLoadRequest;
/**
* Coprocessors implement this interface to observe and mediate bulk load operations.
@@ -55,7 +51,7 @@ public interface BulkLoadObserver {
* It can't bypass the default action, e.g., ctx.bypass() won't have effect.
* If you need to get the region or table name, get it from the
* ctx
as follows: code>ctx.getEnvironment().getRegion()
. Use
- * getRegionInfo to fetch the encodedName and use getTabldDescriptor() to get the tableName.
+ * getRegionInfo to fetch the encodedName and use getTableDescriptor() to get the tableName.
* @param ctx the environment to interact with the framework and master
*/
default void prePrepareBulkLoad(ObserverContext ctx)
@@ -66,7 +62,7 @@ public interface BulkLoadObserver {
* It can't bypass the default action, e.g., ctx.bypass() won't have effect.
* If you need to get the region or table name, get it from the
* ctx
as follows: code>ctx.getEnvironment().getRegion()
. Use
- * getRegionInfo to fetch the encodedName and use getTabldDescriptor() to get the tableName.
+ * getRegionInfo to fetch the encodedName and use getTableDescriptor() to get the tableName.
* @param ctx the environment to interact with the framework and master
*/
default void preCleanupBulkLoad(ObserverContext ctx)
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/DataBlockEncodingValidator.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/DataBlockEncodingValidator.java
new file mode 100644
index 00000000000..e72521b1225
--- /dev/null
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/DataBlockEncodingValidator.java
@@ -0,0 +1,108 @@
+/**
+ *
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.hadoop.hbase.tool;
+
+import java.io.IOException;
+import java.util.List;
+
+import org.apache.hadoop.hbase.HBaseInterfaceAudience;
+import org.apache.hadoop.hbase.client.Admin;
+import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
+import org.apache.hadoop.hbase.client.Connection;
+import org.apache.hadoop.hbase.client.ConnectionFactory;
+import org.apache.hadoop.hbase.client.TableDescriptor;
+import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding;
+import org.apache.hadoop.hbase.util.AbstractHBaseTool;
+import org.apache.hadoop.hbase.util.Bytes;
+import org.apache.yetus.audience.InterfaceAudience;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLine;
+
+@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS)
+public class DataBlockEncodingValidator extends AbstractHBaseTool {
+
+ private static final Logger LOG = LoggerFactory.getLogger(DataBlockEncodingValidator.class);
+ private static final byte[] DATA_BLOCK_ENCODING = Bytes.toBytes("DATA_BLOCK_ENCODING");
+
+ /**
+ * Check DataBlockEncodings of column families are compatible.
+ *
+ * @return number of column families with incompatible DataBlockEncoding
+ * @throws IOException if a remote or network exception occurs
+ */
+ private int validateDBE() throws IOException {
+ int incompatibilities = 0;
+
+ LOG.info("Validating Data Block Encodings");
+
+ try (Connection connection = ConnectionFactory.createConnection(getConf());
+ Admin admin = connection.getAdmin()) {
+ List tableDescriptors = admin.listTableDescriptors();
+ String encoding = "";
+
+ for (TableDescriptor td : tableDescriptors) {
+ ColumnFamilyDescriptor[] columnFamilies = td.getColumnFamilies();
+ for (ColumnFamilyDescriptor cfd : columnFamilies) {
+ try {
+ encoding = Bytes.toString(cfd.getValue(DATA_BLOCK_ENCODING));
+ // IllegalArgumentException will be thrown if encoding is incompatible with 2.0
+ DataBlockEncoding.valueOf(encoding);
+ } catch (IllegalArgumentException e) {
+ incompatibilities++;
+ LOG.warn("Incompatible DataBlockEncoding for table: {}, cf: {}, encoding: {}",
+ td.getTableName().getNameAsString(), cfd.getNameAsString(), encoding);
+ }
+ }
+ }
+ }
+
+ if (incompatibilities > 0) {
+ LOG.warn("There are {} column families with incompatible Data Block Encodings. Do not "
+ + "upgrade until these encodings are converted to a supported one.", incompatibilities);
+ LOG.warn("Check http://hbase.apache.org/book.html#upgrade2.0.prefix-tree.removed "
+ + "for instructions.");
+ } else {
+ LOG.info("The used Data Block Encodings are compatible with HBase 2.0.");
+ }
+
+ return incompatibilities;
+ }
+
+ @Override
+ protected void printUsage() {
+ String header = "hbase " + PreUpgradeValidator.TOOL_NAME + " " +
+ PreUpgradeValidator.VALIDATE_DBE_NAME;
+ printUsage(header, null, "");
+ }
+
+ @Override
+ protected void addOptions() {
+ }
+
+ @Override
+ protected void processOptions(CommandLine cmd) {
+ }
+
+ @Override
+ protected int doWork() throws Exception {
+ return (validateDBE() == 0) ? EXIT_SUCCESS : EXIT_FAILURE;
+ }
+}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/PreUpgradeValidator.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/PreUpgradeValidator.java
index 6fe5a92b22a..a3c505ef60a 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/PreUpgradeValidator.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/PreUpgradeValidator.java
@@ -18,111 +18,99 @@
*/
package org.apache.hadoop.hbase.tool;
-import java.io.IOException;
-import java.util.List;
+import java.util.Arrays;
+import org.apache.hadoop.conf.Configuration;
+import org.apache.hadoop.hbase.HBaseConfiguration;
import org.apache.hadoop.hbase.HBaseInterfaceAudience;
-import org.apache.hadoop.hbase.client.Admin;
-import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
-import org.apache.hadoop.hbase.client.Connection;
-import org.apache.hadoop.hbase.client.ConnectionFactory;
-import org.apache.hadoop.hbase.client.TableDescriptor;
-import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding;
+import org.apache.hadoop.hbase.tool.coprocessor.CoprocessorValidator;
import org.apache.hadoop.hbase.util.AbstractHBaseTool;
-import org.apache.hadoop.hbase.util.Bytes;
+import org.apache.hadoop.util.Tool;
+import org.apache.hadoop.util.ToolRunner;
import org.apache.yetus.audience.InterfaceAudience;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
-import org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLine;
-
/**
* Tool for validating that cluster can be upgraded from HBase 1.x to 2.0
*
* Available validations:
*
- * - all: Run all pre-upgrade validations
- * - validateDBE: Check Data Block Encoding for column families
+ * - validate-cp: Validates Co-processors compatibility
+ * - validate-dbe: Check Data Block Encoding for column families
*
*
*/
@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS)
-public class PreUpgradeValidator extends AbstractHBaseTool {
+public class PreUpgradeValidator implements Tool {
+ private static final Logger LOG = LoggerFactory
+ .getLogger(PreUpgradeValidator.class);
- public static final String NAME = "pre-upgrade";
- private static final Logger LOG = LoggerFactory.getLogger(PreUpgradeValidator.class);
- private static final byte[] DATA_BLOCK_ENCODING = Bytes.toBytes("DATA_BLOCK_ENCODING");
- private boolean validateAll;
- private boolean validateDBE;
+ public static final String TOOL_NAME = "pre-upgrade";
+ public static final String VALIDATE_CP_NAME = "validate-cp";
+ public static final String VALIDATE_DBE_NAME = "validate-dbe";
- /**
- * Check DataBlockEncodings of column families are compatible.
- *
- * @return number of column families with incompatible DataBlockEncoding
- * @throws IOException if a remote or network exception occurs
- */
- private int validateDBE() throws IOException {
- int incompatibilities = 0;
+ private Configuration configuration;
- LOG.info("Validating Data Block Encodings");
-
- try (Connection connection = ConnectionFactory.createConnection(getConf());
- Admin admin = connection.getAdmin()) {
- List tableDescriptors = admin.listTableDescriptors();
- String encoding = "";
-
- for (TableDescriptor td : tableDescriptors) {
- ColumnFamilyDescriptor[] columnFamilies = td.getColumnFamilies();
- for (ColumnFamilyDescriptor cfd : columnFamilies) {
- try {
- encoding = Bytes.toString(cfd.getValue(DATA_BLOCK_ENCODING));
- // IllegalArgumentException will be thrown if encoding is incompatible with 2.0
- DataBlockEncoding.valueOf(encoding);
- } catch (IllegalArgumentException e) {
- incompatibilities++;
- LOG.warn("Incompatible DataBlockEncoding for table: {}, cf: {}, encoding: {}",
- td.getTableName().getNameAsString(), cfd.getNameAsString(), encoding);
- }
- }
- }
- }
-
- if (incompatibilities > 0) {
- LOG.warn("There are {} column families with incompatible Data Block Encodings. Do not "
- + "upgrade until these encodings are converted to a supported one.", incompatibilities);
- LOG.warn("Check http://hbase.apache.org/book.html#upgrade2.0.prefix-tree.removed "
- + "for instructions.");
- } else {
- LOG.info("The used Data Block Encodings are compatible with HBase 2.0.");
- }
- return incompatibilities;
+ @Override
+ public Configuration getConf() {
+ return configuration;
}
@Override
- protected void addOptions() {
- addOptNoArg("all", "Run all pre-upgrade validations");
- addOptNoArg("validateDBE", "Validate DataBlockEncodings are compatible");
+ public void setConf(Configuration conf) {
+ this.configuration = conf;
+ }
+
+ private void printUsage() {
+ System.out.println("usage: hbase " + TOOL_NAME + " command ...");
+ System.out.println("Available commands:");
+ System.out.printf(" %-12s Validate co-processors are compatible with HBase%n",
+ VALIDATE_CP_NAME);
+ System.out.printf(" %-12s Validate DataBlockEncoding are compatible on the cluster%n",
+ VALIDATE_DBE_NAME);
+ System.out.println("For further information, please use command -h");
}
@Override
- protected void processOptions(CommandLine cmd) {
- validateAll = cmd.hasOption("all");
- validateDBE = cmd.hasOption("validateDBE");
- }
-
- @Override
- protected int doWork() throws Exception {
- boolean validationFailed = false;
- if (validateDBE || validateAll) {
- if (validateDBE() > 0) {
- validationFailed = true;
- }
+ public int run(String[] args) throws Exception {
+ if (args.length == 0) {
+ printUsage();
+ return AbstractHBaseTool.EXIT_FAILURE;
}
- return validationFailed ? 1 : 0;
+ Tool tool;
+
+ switch (args[0]) {
+ case VALIDATE_CP_NAME:
+ tool = new CoprocessorValidator();
+ break;
+ case VALIDATE_DBE_NAME:
+ tool = new DataBlockEncodingValidator();
+ break;
+ case "-h":
+ printUsage();
+ return AbstractHBaseTool.EXIT_FAILURE;
+ default:
+ System.err.println("Unknown command: " + args[0]);
+ printUsage();
+ return AbstractHBaseTool.EXIT_FAILURE;
+ }
+
+ tool.setConf(getConf());
+ return tool.run(Arrays.copyOfRange(args, 1, args.length));
}
public static void main(String[] args) {
- new PreUpgradeValidator().doStaticMain(args);
+ int ret;
+
+ try {
+ ret = ToolRunner.run(HBaseConfiguration.create(), new PreUpgradeValidator(), args);
+ } catch (Exception e) {
+ LOG.error("Error running command-line tool", e);
+ ret = AbstractHBaseTool.EXIT_FAILURE;
+ }
+
+ System.exit(ret);
}
}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/coprocessor/Branch1CoprocessorMethods.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/coprocessor/Branch1CoprocessorMethods.java
new file mode 100644
index 00000000000..0f5d829de6b
--- /dev/null
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/coprocessor/Branch1CoprocessorMethods.java
@@ -0,0 +1,1137 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.hadoop.hbase.tool.coprocessor;
+
+import org.apache.yetus.audience.InterfaceAudience;
+
+@InterfaceAudience.Private
+public class Branch1CoprocessorMethods extends CoprocessorMethods {
+ public Branch1CoprocessorMethods() {
+ addMethods();
+ }
+
+ /*
+ * This list of methods was generated from HBase 1.4.4.
+ */
+ private void addMethods() {
+ /* BulkLoadObserver */
+
+ addMethod("prePrepareBulkLoad",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.PrepareBulkLoadRequest");
+
+ addMethod("preCleanupBulkLoad",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.CleanupBulkLoadRequest");
+
+ /* EndpointObserver */
+
+ addMethod("postEndpointInvocation",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "com.google.protobuf.Service",
+ "java.lang.String",
+ "com.google.protobuf.Message",
+ "com.google.protobuf.Message.Builder");
+
+ addMethod("preEndpointInvocation",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "com.google.protobuf.Service",
+ "java.lang.String",
+ "com.google.protobuf.Message");
+
+ /* MasterObserver */
+
+ addMethod("preCreateTable",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HTableDescriptor",
+ "org.apache.hadoop.hbase.HRegionInfo[]");
+
+ addMethod("postCreateTable",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HTableDescriptor",
+ "org.apache.hadoop.hbase.HRegionInfo[]");
+
+ addMethod("preDeleteTable",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName");
+
+ addMethod("postDeleteTable",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName");
+
+ addMethod("preDeleteTableHandler",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName");
+
+ addMethod("preMove",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HRegionInfo",
+ "org.apache.hadoop.hbase.ServerName",
+ "org.apache.hadoop.hbase.ServerName");
+
+ addMethod("preCreateTableHandler",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HTableDescriptor",
+ "org.apache.hadoop.hbase.HRegionInfo[]");
+
+ addMethod("postCreateTableHandler",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HTableDescriptor",
+ "org.apache.hadoop.hbase.HRegionInfo[]");
+
+ addMethod("postMove",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HRegionInfo",
+ "org.apache.hadoop.hbase.ServerName",
+ "org.apache.hadoop.hbase.ServerName");
+
+ addMethod("postDeleteTableHandler",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName");
+
+ addMethod("preTruncateTable",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName");
+
+ addMethod("postTruncateTable",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName");
+
+ addMethod("preTruncateTableHandler",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName");
+
+ addMethod("postTruncateTableHandler",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName");
+
+ addMethod("preModifyTable",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName",
+ "org.apache.hadoop.hbase.HTableDescriptor");
+
+ addMethod("postModifyTable",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName",
+ "org.apache.hadoop.hbase.HTableDescriptor");
+
+ addMethod("preModifyTableHandler",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName",
+ "org.apache.hadoop.hbase.HTableDescriptor");
+
+ addMethod("postModifyTableHandler",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName",
+ "org.apache.hadoop.hbase.HTableDescriptor");
+
+ addMethod("preAddColumn",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName",
+ "org.apache.hadoop.hbase.HColumnDescriptor");
+
+ addMethod("postAddColumn",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName",
+ "org.apache.hadoop.hbase.HColumnDescriptor");
+
+ addMethod("preAddColumnHandler",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName",
+ "org.apache.hadoop.hbase.HColumnDescriptor");
+
+ addMethod("postAddColumnHandler",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName",
+ "org.apache.hadoop.hbase.HColumnDescriptor");
+
+ addMethod("preModifyColumn",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName",
+ "org.apache.hadoop.hbase.HColumnDescriptor");
+
+ addMethod("postModifyColumn",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName",
+ "org.apache.hadoop.hbase.HColumnDescriptor");
+
+ addMethod("preModifyColumnHandler",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName",
+ "org.apache.hadoop.hbase.HColumnDescriptor");
+
+ addMethod("postModifyColumnHandler",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName",
+ "org.apache.hadoop.hbase.HColumnDescriptor");
+
+ addMethod("preDeleteColumn",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName",
+ "byte[]");
+
+ addMethod("postDeleteColumn",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName",
+ "byte[]");
+
+ addMethod("preDeleteColumnHandler",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName",
+ "byte[]");
+
+ addMethod("postDeleteColumnHandler",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName",
+ "byte[]");
+
+ addMethod("preEnableTable",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName");
+
+ addMethod("postEnableTable",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName");
+
+ addMethod("preEnableTableHandler",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName");
+
+ addMethod("postEnableTableHandler",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName");
+
+ addMethod("preDisableTable",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName");
+
+ addMethod("postDisableTable",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName");
+
+ addMethod("preDisableTableHandler",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName");
+
+ addMethod("postDisableTableHandler",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName");
+
+ addMethod("preAbortProcedure",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.procedure2.ProcedureExecutor",
+ "long");
+
+ addMethod("postAbortProcedure",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("preListProcedures",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("postListProcedures",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.List");
+
+ addMethod("preAssign",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HRegionInfo");
+
+ addMethod("postAssign",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HRegionInfo");
+
+ addMethod("preUnassign",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HRegionInfo",
+ "boolean");
+
+ addMethod("postUnassign",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HRegionInfo",
+ "boolean");
+
+ addMethod("preRegionOffline",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HRegionInfo");
+
+ addMethod("postRegionOffline",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HRegionInfo");
+
+ addMethod("preBalance",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("postBalance",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.List");
+
+ addMethod("preSetSplitOrMergeEnabled",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "boolean",
+ "org.apache.hadoop.hbase.client.Admin.MasterSwitchType");
+
+ addMethod("postSetSplitOrMergeEnabled",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "boolean",
+ "org.apache.hadoop.hbase.client.Admin.MasterSwitchType");
+
+ addMethod("preBalanceSwitch",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "boolean");
+
+ addMethod("postBalanceSwitch",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "boolean",
+ "boolean");
+
+ addMethod("preShutdown",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("preStopMaster",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("postStartMaster",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("preMasterInitialization",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("preSnapshot",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription",
+ "org.apache.hadoop.hbase.HTableDescriptor");
+
+ addMethod("postSnapshot",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription",
+ "org.apache.hadoop.hbase.HTableDescriptor");
+
+ addMethod("preListSnapshot",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription");
+
+ addMethod("postListSnapshot",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription");
+
+ addMethod("preCloneSnapshot",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription",
+ "org.apache.hadoop.hbase.HTableDescriptor");
+
+ addMethod("postCloneSnapshot",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription",
+ "org.apache.hadoop.hbase.HTableDescriptor");
+
+ addMethod("preRestoreSnapshot",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription",
+ "org.apache.hadoop.hbase.HTableDescriptor");
+
+ addMethod("postRestoreSnapshot",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription",
+ "org.apache.hadoop.hbase.HTableDescriptor");
+
+ addMethod("preDeleteSnapshot",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription");
+
+ addMethod("postDeleteSnapshot",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription");
+
+ addMethod("preGetTableDescriptors",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.List",
+ "java.util.List");
+
+ addMethod("preGetTableDescriptors",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.List",
+ "java.util.List",
+ "java.lang.String");
+
+ addMethod("postGetTableDescriptors",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.List",
+ "java.util.List",
+ "java.lang.String");
+
+ addMethod("postGetTableDescriptors",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.List");
+
+ addMethod("preGetTableNames",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.List",
+ "java.lang.String");
+
+ addMethod("postGetTableNames",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.List",
+ "java.lang.String");
+
+ addMethod("preCreateNamespace",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.NamespaceDescriptor");
+
+ addMethod("postCreateNamespace",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.NamespaceDescriptor");
+
+ addMethod("preDeleteNamespace",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.lang.String");
+
+ addMethod("postDeleteNamespace",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.lang.String");
+
+ addMethod("preModifyNamespace",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.NamespaceDescriptor");
+
+ addMethod("postModifyNamespace",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.NamespaceDescriptor");
+
+ addMethod("preGetNamespaceDescriptor",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.lang.String");
+
+ addMethod("postGetNamespaceDescriptor",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.NamespaceDescriptor");
+
+ addMethod("preListNamespaceDescriptors",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.List");
+
+ addMethod("postListNamespaceDescriptors",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.List");
+
+ addMethod("preTableFlush",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName");
+
+ addMethod("postTableFlush",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName");
+
+ addMethod("preSetUserQuota",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.lang.String",
+ "java.lang.String",
+ "org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas");
+
+ addMethod("preSetUserQuota",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.lang.String",
+ "org.apache.hadoop.hbase.TableName",
+ "org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas");
+
+ addMethod("preSetUserQuota",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.lang.String",
+ "org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas");
+
+ addMethod("postSetUserQuota",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.lang.String",
+ "java.lang.String",
+ "org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas");
+
+ addMethod("postSetUserQuota",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.lang.String",
+ "org.apache.hadoop.hbase.TableName",
+ "org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas");
+
+ addMethod("postSetUserQuota",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.lang.String",
+ "org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas");
+
+ addMethod("preSetTableQuota",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName",
+ "org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas");
+
+ addMethod("postSetTableQuota",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.TableName",
+ "org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas");
+
+ addMethod("preSetNamespaceQuota",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.lang.String",
+ "org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas");
+
+ addMethod("postSetNamespaceQuota",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.lang.String",
+ "org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.Quotas");
+
+ addMethod("preDispatchMerge",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HRegionInfo",
+ "org.apache.hadoop.hbase.HRegionInfo");
+
+ addMethod("postDispatchMerge",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HRegionInfo",
+ "org.apache.hadoop.hbase.HRegionInfo");
+
+ addMethod("preGetClusterStatus",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("postGetClusterStatus",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.ClusterStatus");
+
+ addMethod("preClearDeadServers",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("postClearDeadServers",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.List",
+ "java.util.List");
+
+ addMethod("preMoveServers",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.Set",
+ "java.lang.String");
+
+ addMethod("postMoveServers",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.Set",
+ "java.lang.String");
+
+ addMethod("preMoveTables",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.Set",
+ "java.lang.String");
+
+ addMethod("postMoveTables",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.Set",
+ "java.lang.String");
+
+ addMethod("preMoveServersAndTables",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.Set",
+ "java.util.Set",
+ "java.lang.String");
+
+ addMethod("postMoveServersAndTables",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.Set",
+ "java.util.Set",
+ "java.lang.String");
+
+ addMethod("preAddRSGroup",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.lang.String");
+
+ addMethod("postAddRSGroup",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.lang.String");
+
+ addMethod("preRemoveRSGroup",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.lang.String");
+
+ addMethod("postRemoveRSGroup",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.lang.String");
+
+ addMethod("preRemoveServers",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.Set");
+
+ addMethod("postRemoveServers",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.Set");
+
+ addMethod("preBalanceRSGroup",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.lang.String");
+
+ addMethod("postBalanceRSGroup",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.lang.String",
+ "boolean");
+
+ /* RegionObserver */
+
+ addMethod("preOpen",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("postOpen",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("postLogReplay",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("preFlushScannerOpen",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Store",
+ "org.apache.hadoop.hbase.regionserver.KeyValueScanner",
+ "org.apache.hadoop.hbase.regionserver.InternalScanner",
+ "long");
+
+ addMethod("preFlushScannerOpen",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Store",
+ "org.apache.hadoop.hbase.regionserver.KeyValueScanner",
+ "org.apache.hadoop.hbase.regionserver.InternalScanner");
+
+ addMethod("preFlush",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Store",
+ "org.apache.hadoop.hbase.regionserver.InternalScanner");
+
+ addMethod("preFlush",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("postFlush",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Store",
+ "org.apache.hadoop.hbase.regionserver.StoreFile");
+
+ addMethod("postFlush",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("preCompactSelection",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Store",
+ "java.util.List");
+
+ addMethod("preCompactSelection",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Store",
+ "java.util.List",
+ "org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest");
+
+ addMethod("postCompactSelection",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Store",
+ "com.google.common.collect.ImmutableList");
+
+ addMethod("postCompactSelection",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Store",
+ "com.google.common.collect.ImmutableList",
+ "org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest");
+
+ addMethod("preCompact",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Store",
+ "org.apache.hadoop.hbase.regionserver.InternalScanner",
+ "org.apache.hadoop.hbase.regionserver.ScanType");
+
+ addMethod("preCompact",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Store",
+ "org.apache.hadoop.hbase.regionserver.InternalScanner",
+ "org.apache.hadoop.hbase.regionserver.ScanType",
+ "org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest");
+
+ addMethod("preClose",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "boolean");
+
+ addMethod("preCompactScannerOpen",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Store",
+ "java.util.List",
+ "org.apache.hadoop.hbase.regionserver.ScanType",
+ "long",
+ "org.apache.hadoop.hbase.regionserver.InternalScanner");
+
+ addMethod("preCompactScannerOpen",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Store",
+ "java.util.List",
+ "org.apache.hadoop.hbase.regionserver.ScanType",
+ "long",
+ "org.apache.hadoop.hbase.regionserver.InternalScanner",
+ "org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest",
+ "long");
+
+ addMethod("preCompactScannerOpen",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Store",
+ "java.util.List",
+ "org.apache.hadoop.hbase.regionserver.ScanType",
+ "long",
+ "org.apache.hadoop.hbase.regionserver.InternalScanner",
+ "org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest");
+
+ addMethod("postCompact",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Store",
+ "org.apache.hadoop.hbase.regionserver.StoreFile");
+
+ addMethod("postCompact",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Store",
+ "org.apache.hadoop.hbase.regionserver.StoreFile",
+ "org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest");
+
+ addMethod("preSplit",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "byte[]");
+
+ addMethod("preSplit",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("postSplit",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Region",
+ "org.apache.hadoop.hbase.regionserver.Region");
+
+ addMethod("preSplitBeforePONR",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "byte[]",
+ "java.util.List");
+
+ addMethod("preSplitAfterPONR",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("preRollBackSplit",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("postRollBackSplit",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("postCompleteSplit",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("postClose",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "boolean");
+
+ addMethod("preGetClosestRowBefore",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "byte[]",
+ "byte[]",
+ "org.apache.hadoop.hbase.client.Result");
+
+ addMethod("postGetClosestRowBefore",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "byte[]",
+ "byte[]",
+ "org.apache.hadoop.hbase.client.Result");
+
+ addMethod("preGetOp",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.client.Get",
+ "java.util.List");
+
+ addMethod("postGetOp",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.client.Get",
+ "java.util.List");
+
+ addMethod("preExists",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.client.Get",
+ "boolean");
+
+ addMethod("postExists",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.client.Get",
+ "boolean");
+
+ addMethod("prePut",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.client.Put",
+ "org.apache.hadoop.hbase.regionserver.wal.WALEdit",
+ "org.apache.hadoop.hbase.client.Durability");
+
+ addMethod("postPut",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.client.Put",
+ "org.apache.hadoop.hbase.regionserver.wal.WALEdit",
+ "org.apache.hadoop.hbase.client.Durability");
+
+ addMethod("preDelete",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.client.Delete",
+ "org.apache.hadoop.hbase.regionserver.wal.WALEdit",
+ "org.apache.hadoop.hbase.client.Durability");
+
+ addMethod("prePrepareTimeStampForDeleteVersion",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.client.Mutation",
+ "org.apache.hadoop.hbase.Cell",
+ "byte[]",
+ "org.apache.hadoop.hbase.client.Get");
+
+ addMethod("postDelete",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.client.Delete",
+ "org.apache.hadoop.hbase.regionserver.wal.WALEdit",
+ "org.apache.hadoop.hbase.client.Durability");
+
+ addMethod("preBatchMutate",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress");
+
+ addMethod("postBatchMutate",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress");
+
+ addMethod("postStartRegionOperation",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Region.Operation");
+
+ addMethod("postCloseRegionOperation",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Region.Operation");
+
+ addMethod("postBatchMutateIndispensably",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress",
+ "boolean");
+
+ addMethod("preCheckAndPut",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "byte[]",
+ "byte[]",
+ "byte[]",
+ "org.apache.hadoop.hbase.filter.CompareFilter.CompareOp",
+ "org.apache.hadoop.hbase.filter.ByteArrayComparable",
+ "org.apache.hadoop.hbase.client.Put",
+ "boolean");
+
+ addMethod("preCheckAndPutAfterRowLock",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "byte[]",
+ "byte[]",
+ "byte[]",
+ "org.apache.hadoop.hbase.filter.CompareFilter.CompareOp",
+ "org.apache.hadoop.hbase.filter.ByteArrayComparable",
+ "org.apache.hadoop.hbase.client.Put",
+ "boolean");
+
+ addMethod("postCheckAndPut",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "byte[]",
+ "byte[]",
+ "byte[]",
+ "org.apache.hadoop.hbase.filter.CompareFilter.CompareOp",
+ "org.apache.hadoop.hbase.filter.ByteArrayComparable",
+ "org.apache.hadoop.hbase.client.Put",
+ "boolean");
+
+ addMethod("preCheckAndDelete",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "byte[]",
+ "byte[]",
+ "byte[]",
+ "org.apache.hadoop.hbase.filter.CompareFilter.CompareOp",
+ "org.apache.hadoop.hbase.filter.ByteArrayComparable",
+ "org.apache.hadoop.hbase.client.Delete",
+ "boolean");
+
+ addMethod("preCheckAndDeleteAfterRowLock",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "byte[]",
+ "byte[]",
+ "byte[]",
+ "org.apache.hadoop.hbase.filter.CompareFilter.CompareOp",
+ "org.apache.hadoop.hbase.filter.ByteArrayComparable",
+ "org.apache.hadoop.hbase.client.Delete",
+ "boolean");
+
+ addMethod("postCheckAndDelete",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "byte[]",
+ "byte[]",
+ "byte[]",
+ "org.apache.hadoop.hbase.filter.CompareFilter.CompareOp",
+ "org.apache.hadoop.hbase.filter.ByteArrayComparable",
+ "org.apache.hadoop.hbase.client.Delete",
+ "boolean");
+
+ addMethod("preIncrementColumnValue",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "byte[]",
+ "byte[]",
+ "byte[]",
+ "long",
+ "boolean");
+
+ addMethod("postIncrementColumnValue",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "byte[]",
+ "byte[]",
+ "byte[]",
+ "long",
+ "boolean",
+ "long");
+
+ addMethod("preAppend",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.client.Append");
+
+ addMethod("preAppendAfterRowLock",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.client.Append");
+
+ addMethod("postAppend",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.client.Append",
+ "org.apache.hadoop.hbase.client.Result");
+
+ addMethod("preIncrement",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.client.Increment");
+
+ addMethod("preIncrementAfterRowLock",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.client.Increment");
+
+ addMethod("postIncrement",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.client.Increment",
+ "org.apache.hadoop.hbase.client.Result");
+
+ addMethod("preScannerOpen",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.client.Scan",
+ "org.apache.hadoop.hbase.regionserver.RegionScanner");
+
+ addMethod("preStoreScannerOpen",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Store",
+ "org.apache.hadoop.hbase.client.Scan",
+ "java.util.NavigableSet",
+ "org.apache.hadoop.hbase.regionserver.KeyValueScanner");
+
+ addMethod("postScannerOpen",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.client.Scan",
+ "org.apache.hadoop.hbase.regionserver.RegionScanner");
+
+ addMethod("preScannerNext",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.InternalScanner",
+ "java.util.List",
+ "int",
+ "boolean");
+
+ addMethod("postScannerNext",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.InternalScanner",
+ "java.util.List",
+ "int",
+ "boolean");
+
+ addMethod("postScannerFilterRow",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.InternalScanner",
+ "byte[]",
+ "int",
+ "short",
+ "boolean");
+
+ addMethod("preScannerClose",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.InternalScanner");
+
+ addMethod("postScannerClose",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.InternalScanner");
+
+ addMethod("preWALRestore",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HRegionInfo",
+ "org.apache.hadoop.hbase.regionserver.wal.HLogKey",
+ "org.apache.hadoop.hbase.regionserver.wal.WALEdit");
+
+ addMethod("preWALRestore",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HRegionInfo",
+ "org.apache.hadoop.hbase.wal.WALKey",
+ "org.apache.hadoop.hbase.regionserver.wal.WALEdit");
+
+ addMethod("postWALRestore",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HRegionInfo",
+ "org.apache.hadoop.hbase.regionserver.wal.HLogKey",
+ "org.apache.hadoop.hbase.regionserver.wal.WALEdit");
+
+ addMethod("postWALRestore",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HRegionInfo",
+ "org.apache.hadoop.hbase.wal.WALKey",
+ "org.apache.hadoop.hbase.regionserver.wal.WALEdit");
+
+ addMethod("preBulkLoadHFile",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.List");
+
+ addMethod("preCommitStoreFile",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "byte[]",
+ "java.util.List");
+
+ addMethod("postCommitStoreFile",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "byte[]",
+ "org.apache.hadoop.fs.Path",
+ "org.apache.hadoop.fs.Path");
+
+ addMethod("postBulkLoadHFile",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.List",
+ "boolean");
+
+ addMethod("preStoreFileReaderOpen",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.fs.FileSystem",
+ "org.apache.hadoop.fs.Path",
+ "org.apache.hadoop.hbase.io.FSDataInputStreamWrapper",
+ "long",
+ "org.apache.hadoop.hbase.io.hfile.CacheConfig",
+ "org.apache.hadoop.hbase.io.Reference",
+ "org.apache.hadoop.hbase.regionserver.StoreFile.Reader");
+
+ addMethod("postStoreFileReaderOpen",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.fs.FileSystem",
+ "org.apache.hadoop.fs.Path",
+ "org.apache.hadoop.hbase.io.FSDataInputStreamWrapper",
+ "long",
+ "org.apache.hadoop.hbase.io.hfile.CacheConfig",
+ "org.apache.hadoop.hbase.io.Reference",
+ "org.apache.hadoop.hbase.regionserver.StoreFile.Reader");
+
+ addMethod("postMutationBeforeWAL",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.coprocessor.RegionObserver.MutationType",
+ "org.apache.hadoop.hbase.client.Mutation",
+ "org.apache.hadoop.hbase.Cell",
+ "org.apache.hadoop.hbase.Cell");
+
+ addMethod("postInstantiateDeleteTracker",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.DeleteTracker");
+
+ /* RegionServerObserver */
+
+ addMethod("preMerge",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Region",
+ "org.apache.hadoop.hbase.regionserver.Region");
+
+ addMethod("preStopRegionServer",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("postMerge",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Region",
+ "org.apache.hadoop.hbase.regionserver.Region",
+ "org.apache.hadoop.hbase.regionserver.Region");
+
+ addMethod("preMergeCommit",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Region",
+ "org.apache.hadoop.hbase.regionserver.Region",
+ "java.util.List");
+
+ addMethod("postMergeCommit",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Region",
+ "org.apache.hadoop.hbase.regionserver.Region",
+ "org.apache.hadoop.hbase.regionserver.Region");
+
+ addMethod("preRollBackMerge",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Region",
+ "org.apache.hadoop.hbase.regionserver.Region");
+
+ addMethod("postRollBackMerge",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.regionserver.Region",
+ "org.apache.hadoop.hbase.regionserver.Region");
+
+ addMethod("preRollWALWriterRequest",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("postRollWALWriterRequest",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext");
+
+ addMethod("postCreateReplicationEndPoint",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.replication.ReplicationEndpoint");
+
+ addMethod("preReplicateLogEntries",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.List",
+ "org.apache.hadoop.hbase.CellScanner");
+
+ addMethod("postReplicateLogEntries",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "java.util.List",
+ "org.apache.hadoop.hbase.CellScanner");
+
+ /* WALObserver */
+
+ addMethod("preWALWrite",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HRegionInfo",
+ "org.apache.hadoop.hbase.wal.WALKey",
+ "org.apache.hadoop.hbase.regionserver.wal.WALEdit");
+
+ addMethod("preWALWrite",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HRegionInfo",
+ "org.apache.hadoop.hbase.regionserver.wal.HLogKey",
+ "org.apache.hadoop.hbase.regionserver.wal.WALEdit");
+
+ addMethod("postWALWrite",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HRegionInfo",
+ "org.apache.hadoop.hbase.regionserver.wal.HLogKey",
+ "org.apache.hadoop.hbase.regionserver.wal.WALEdit");
+
+ addMethod("postWALWrite",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.hbase.HRegionInfo",
+ "org.apache.hadoop.hbase.wal.WALKey",
+ "org.apache.hadoop.hbase.regionserver.wal.WALEdit");
+
+ addMethod("preWALRoll",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.fs.Path",
+ "org.apache.hadoop.fs.Path");
+
+ addMethod("postWALRoll",
+ "org.apache.hadoop.hbase.coprocessor.ObserverContext",
+ "org.apache.hadoop.fs.Path",
+ "org.apache.hadoop.fs.Path");
+ }
+}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/coprocessor/CoprocessorMethod.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/coprocessor/CoprocessorMethod.java
new file mode 100644
index 00000000000..60e38417135
--- /dev/null
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/coprocessor/CoprocessorMethod.java
@@ -0,0 +1,73 @@
+/**
+ *
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.hadoop.hbase.tool.coprocessor;
+
+import java.util.ArrayList;
+import java.util.List;
+import java.util.Objects;
+
+import org.apache.yetus.audience.InterfaceAudience;
+
+@InterfaceAudience.Private
+public class CoprocessorMethod {
+ private final String name;
+ private final List parameters;
+
+ public CoprocessorMethod(String name) {
+ this.name = name;
+
+ parameters = new ArrayList<>();
+ }
+
+ public CoprocessorMethod withParameters(String ... parameters) {
+ for (String parameter : parameters) {
+ this.parameters.add(parameter);
+ }
+
+ return this;
+ }
+
+ public CoprocessorMethod withParameters(Class> ... parameters) {
+ for (Class> parameter : parameters) {
+ this.parameters.add(parameter.getCanonicalName());
+ }
+
+ return this;
+ }
+
+ @Override
+ public boolean equals(Object obj) {
+ if (obj == this) {
+ return true;
+ } else if (!(obj instanceof CoprocessorMethod)) {
+ return false;
+ }
+
+ CoprocessorMethod other = (CoprocessorMethod)obj;
+
+ return Objects.equals(name, other.name) &&
+ Objects.equals(parameters, other.parameters);
+ }
+
+ @Override
+ public int hashCode() {
+ return Objects.hash(name, parameters);
+ }
+}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/coprocessor/CoprocessorMethods.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/coprocessor/CoprocessorMethods.java
new file mode 100644
index 00000000000..2e0c801b8aa
--- /dev/null
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/coprocessor/CoprocessorMethods.java
@@ -0,0 +1,66 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.hadoop.hbase.tool.coprocessor;
+
+import java.lang.reflect.Method;
+import java.util.HashSet;
+import java.util.Set;
+
+import org.apache.yetus.audience.InterfaceAudience;
+
+@InterfaceAudience.Private
+public class CoprocessorMethods {
+ private final Set methods;
+
+ public CoprocessorMethods() {
+ methods = new HashSet<>();
+ }
+
+ public void addMethod(String name, String ... parameters) {
+ CoprocessorMethod cpMethod = new CoprocessorMethod(name).withParameters(parameters);
+ methods.add(cpMethod);
+ }
+
+ public void addMethod(String name, Class> ... parameters) {
+ CoprocessorMethod cpMethod = new CoprocessorMethod(name).withParameters(parameters);
+ methods.add(cpMethod);
+ }
+
+ public void addMethod(Method method) {
+ CoprocessorMethod cpMethod = new CoprocessorMethod(method.getName())
+ .withParameters(method.getParameterTypes());
+ methods.add(cpMethod);
+ }
+
+ public boolean hasMethod(String name, String ... parameters) {
+ CoprocessorMethod method = new CoprocessorMethod(name).withParameters(parameters);
+ return methods.contains(method);
+ }
+
+ public boolean hasMethod(String name, Class> ... parameters) {
+ CoprocessorMethod method = new CoprocessorMethod(name).withParameters(parameters);
+ return methods.contains(method);
+ }
+
+ public boolean hasMethod(Method method) {
+ CoprocessorMethod cpMethod = new CoprocessorMethod(method.getName())
+ .withParameters(method.getParameterTypes());
+ return methods.contains(cpMethod);
+ }
+}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/coprocessor/CoprocessorValidator.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/coprocessor/CoprocessorValidator.java
new file mode 100644
index 00000000000..c6d57236ae8
--- /dev/null
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/coprocessor/CoprocessorValidator.java
@@ -0,0 +1,247 @@
+/**
+ *
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.hadoop.hbase.tool.coprocessor;
+
+import java.io.IOException;
+import java.lang.reflect.Method;
+import java.net.URL;
+import java.net.URLClassLoader;
+import java.nio.file.Path;
+import java.nio.file.Paths;
+import java.security.AccessController;
+import java.security.PrivilegedAction;
+import java.util.ArrayList;
+import java.util.List;
+import java.util.jar.JarEntry;
+import java.util.jar.JarFile;
+import java.util.stream.Collectors;
+
+import org.apache.hadoop.hbase.Coprocessor;
+import org.apache.hadoop.hbase.HBaseInterfaceAudience;
+import org.apache.hadoop.hbase.tool.PreUpgradeValidator;
+import org.apache.hadoop.hbase.tool.coprocessor.CoprocessorViolation.Severity;
+import org.apache.hadoop.hbase.util.AbstractHBaseTool;
+import org.apache.yetus.audience.InterfaceAudience;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
+import org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLine;
+import org.apache.hbase.thirdparty.org.apache.commons.cli.ParseException;
+
+@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS)
+public class CoprocessorValidator extends AbstractHBaseTool {
+ private static final Logger LOG = LoggerFactory
+ .getLogger(CoprocessorValidator.class);
+
+ private CoprocessorMethods branch1;
+ private CoprocessorMethods current;
+
+ private boolean dieOnWarnings;
+ private boolean scan;
+ private List args;
+
+ public CoprocessorValidator() {
+ branch1 = new Branch1CoprocessorMethods();
+ current = new CurrentCoprocessorMethods();
+ }
+
+ /**
+ * This classloader implementation calls {@link #resolveClass(Class)}
+ * method for every loaded class. It means that some extra validation will
+ * take place
+ * according to JLS.
+ */
+ private static final class ResolverUrlClassLoader extends URLClassLoader {
+ private ResolverUrlClassLoader(URL[] urls) {
+ super(urls, ResolverUrlClassLoader.class.getClassLoader());
+ }
+
+ @Override
+ public Class> loadClass(String name) throws ClassNotFoundException {
+ return loadClass(name, true);
+ }
+ }
+
+ private ResolverUrlClassLoader createClassLoader(URL[] urls) {
+ return AccessController.doPrivileged(new PrivilegedAction() {
+ @Override
+ public ResolverUrlClassLoader run() {
+ return new ResolverUrlClassLoader(urls);
+ }
+ });
+ }
+
+ private void validate(ClassLoader classLoader, String className,
+ List violations) {
+ LOG.debug("Validating class '{}'.", className);
+
+ try {
+ Class> clazz = classLoader.loadClass(className);
+
+ for (Method method : clazz.getDeclaredMethods()) {
+ LOG.trace("Validating method '{}'.", method);
+
+ if (branch1.hasMethod(method) && !current.hasMethod(method)) {
+ CoprocessorViolation violation = new CoprocessorViolation(Severity.WARNING,
+ "Method '" + method + "' was removed from new coprocessor API, "
+ + "so it won't be called by HBase.");
+ violations.add(violation);
+ }
+ }
+ } catch (ClassNotFoundException e) {
+ CoprocessorViolation violation = new CoprocessorViolation(Severity.ERROR,
+ "No such class '" + className + "'.", e);
+ violations.add(violation);
+ } catch (RuntimeException | Error e) {
+ CoprocessorViolation violation = new CoprocessorViolation(Severity.ERROR,
+ "Could not validate class '" + className + "'.", e);
+ violations.add(violation);
+ }
+ }
+
+ public List validate(ClassLoader classLoader, List classNames) {
+ List violations = new ArrayList<>();
+
+ for (String className : classNames) {
+ validate(classLoader, className, violations);
+ }
+
+ return violations;
+ }
+
+ public List validate(List urls, List classNames)
+ throws IOException {
+ URL[] urlArray = new URL[urls.size()];
+ urls.toArray(urlArray);
+
+ try (ResolverUrlClassLoader classLoader = createClassLoader(urlArray)) {
+ return validate(classLoader, classNames);
+ }
+ }
+
+ @VisibleForTesting
+ protected List getJarClasses(Path path) throws IOException {
+ try (JarFile jarFile = new JarFile(path.toFile())) {
+ return jarFile.stream()
+ .map(JarEntry::getName)
+ .filter((name) -> name.endsWith(".class"))
+ .map((name) -> name.substring(0, name.length() - 6).replace('/', '.'))
+ .collect(Collectors.toList());
+ }
+ }
+
+ @VisibleForTesting
+ protected List filterObservers(ClassLoader classLoader,
+ Iterable classNames) throws ClassNotFoundException {
+ List filteredClassNames = new ArrayList<>();
+
+ for (String className : classNames) {
+ LOG.debug("Scanning class '{}'.", className);
+
+ Class> clazz = classLoader.loadClass(className);
+
+ if (Coprocessor.class.isAssignableFrom(clazz)) {
+ LOG.debug("Found coprocessor class '{}'.", className);
+ filteredClassNames.add(className);
+ }
+ }
+
+ return filteredClassNames;
+ }
+
+ @Override
+ protected void printUsage() {
+ String header = "hbase " + PreUpgradeValidator.TOOL_NAME + " " +
+ PreUpgradeValidator.VALIDATE_CP_NAME + " -scan|";
+ printUsage(header, "Options:", "");
+ }
+
+ @Override
+ protected void addOptions() {
+ addOptNoArg("e", "Treat warnings as errors.");
+ addOptNoArg("scan", "Scan jar for observers.");
+ }
+
+ @Override
+ protected void processOptions(CommandLine cmd) {
+ scan = cmd.hasOption("scan");
+ dieOnWarnings = cmd.hasOption("e");
+ args = cmd.getArgList();
+ }
+
+ @Override
+ protected int doWork() throws Exception {
+ if (args.size() < 1) {
+ System.err.println("Missing jar file.");
+ printUsage();
+ return EXIT_FAILURE;
+ }
+
+ String jar = args.get(0);
+
+ if (args.size() == 1 && !scan) {
+ throw new ParseException("Missing classes or -scan option.");
+ } else if (args.size() > 1 && scan) {
+ throw new ParseException("Can't use classes with -scan option.");
+ }
+
+ Path jarPath = Paths.get(jar);
+ URL[] urls = new URL[] { jarPath.toUri().toURL() };
+
+ List violations;
+
+ try (ResolverUrlClassLoader classLoader = createClassLoader(urls)) {
+ List classNames;
+
+ if (scan) {
+ List jarClassNames = getJarClasses(jarPath);
+ classNames = filterObservers(classLoader, jarClassNames);
+ } else {
+ classNames = args.subList(1, args.size());
+ }
+
+ violations = validate(classLoader, classNames);
+ }
+
+ boolean error = false;
+
+ for (CoprocessorViolation violation : violations) {
+ switch (violation.getSeverity()) {
+ case WARNING:
+ System.err.println("[WARNING] " + violation.getMessage());
+
+ if (dieOnWarnings) {
+ error = true;
+ }
+
+ break;
+ case ERROR:
+ System.err.println("[ERROR] " + violation.getMessage());
+ error = true;
+
+ break;
+ }
+ }
+
+ return (error) ? EXIT_FAILURE : EXIT_SUCCESS;
+ }
+}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/coprocessor/CoprocessorViolation.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/coprocessor/CoprocessorViolation.java
new file mode 100644
index 00000000000..c403c074179
--- /dev/null
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/coprocessor/CoprocessorViolation.java
@@ -0,0 +1,56 @@
+/**
+ *
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.hadoop.hbase.tool.coprocessor;
+
+import org.apache.yetus.audience.InterfaceAudience;
+
+import org.apache.hbase.thirdparty.com.google.common.base.Throwables;
+
+@InterfaceAudience.Private
+public class CoprocessorViolation {
+ public enum Severity {
+ WARNING, ERROR
+ }
+
+ private final Severity severity;
+ private final String message;
+
+ public CoprocessorViolation(Severity severity, String message) {
+ this(severity, message, null);
+ }
+
+ public CoprocessorViolation(Severity severity, String message, Throwable t) {
+ this.severity = severity;
+
+ if (t == null) {
+ this.message = message;
+ } else {
+ this.message = message + "\n" + Throwables.getStackTraceAsString(t);
+ }
+ }
+
+ public Severity getSeverity() {
+ return severity;
+ }
+
+ public String getMessage() {
+ return message;
+ }
+}
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/coprocessor/CurrentCoprocessorMethods.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/coprocessor/CurrentCoprocessorMethods.java
new file mode 100644
index 00000000000..265cf5158ee
--- /dev/null
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/coprocessor/CurrentCoprocessorMethods.java
@@ -0,0 +1,47 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.hadoop.hbase.tool.coprocessor;
+
+import java.lang.reflect.Method;
+
+import org.apache.hadoop.hbase.coprocessor.BulkLoadObserver;
+import org.apache.hadoop.hbase.coprocessor.EndpointObserver;
+import org.apache.hadoop.hbase.coprocessor.MasterObserver;
+import org.apache.hadoop.hbase.coprocessor.RegionObserver;
+import org.apache.hadoop.hbase.coprocessor.RegionServerObserver;
+import org.apache.hadoop.hbase.coprocessor.WALObserver;
+import org.apache.yetus.audience.InterfaceAudience;
+
+@InterfaceAudience.Private
+public class CurrentCoprocessorMethods extends CoprocessorMethods {
+ public CurrentCoprocessorMethods() {
+ addMethods(BulkLoadObserver.class);
+ addMethods(EndpointObserver.class);
+ addMethods(MasterObserver.class);
+ addMethods(RegionObserver.class);
+ addMethods(RegionServerObserver.class);
+ addMethods(WALObserver.class);
+ }
+
+ private void addMethods(Class> clazz) {
+ for (Method method : clazz.getDeclaredMethods()) {
+ addMethod(method);
+ }
+ }
+}
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/tool/coprocessor/CoprocessorValidatorTest.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/tool/coprocessor/CoprocessorValidatorTest.java
new file mode 100644
index 00000000000..8926ff56ee4
--- /dev/null
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/tool/coprocessor/CoprocessorValidatorTest.java
@@ -0,0 +1,177 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.hadoop.hbase.tool.coprocessor;
+
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertTrue;
+
+import java.io.IOException;
+import java.util.List;
+
+import org.apache.hadoop.hbase.Coprocessor;
+import org.apache.hadoop.hbase.CoprocessorEnvironment;
+import org.apache.hadoop.hbase.HBaseClassTestRule;
+import org.apache.hadoop.hbase.HRegionInfo;
+import org.apache.hadoop.hbase.HTableDescriptor;
+import org.apache.hadoop.hbase.coprocessor.MasterCoprocessorEnvironment;
+import org.apache.hadoop.hbase.coprocessor.ObserverContext;
+import org.apache.hadoop.hbase.testclassification.SmallTests;
+import org.apache.hadoop.hbase.tool.coprocessor.CoprocessorViolation.Severity;
+import org.junit.ClassRule;
+import org.junit.Test;
+import org.junit.experimental.categories.Category;
+
+import org.apache.hbase.thirdparty.com.google.common.collect.Lists;
+
+@Category({ SmallTests.class })
+@SuppressWarnings("deprecation")
+public class CoprocessorValidatorTest {
+ @ClassRule
+ public static final HBaseClassTestRule CLASS_RULE =
+ HBaseClassTestRule.forClass(CoprocessorValidatorTest.class);
+
+ private CoprocessorValidator validator;
+
+ public CoprocessorValidatorTest() {
+ validator = new CoprocessorValidator();
+ }
+
+ private static ClassLoader getClassLoader() {
+ return CoprocessorValidatorTest.class.getClassLoader();
+ }
+
+ private static String getFullClassName(String className) {
+ return CoprocessorValidatorTest.class.getName() + "$" + className;
+ }
+
+ @SuppressWarnings({"rawtypes", "unused"})
+ private static class TestObserver implements Coprocessor {
+ @Override
+ public void start(CoprocessorEnvironment env) throws IOException {
+ }
+
+ @Override
+ public void stop(CoprocessorEnvironment env) throws IOException {
+ }
+ }
+
+ @Test
+ public void testFilterObservers() throws Exception {
+ String filterObservers = getFullClassName("TestObserver");
+ List classNames = Lists.newArrayList(
+ filterObservers, getClass().getName());
+ List filteredClassNames = validator.filterObservers(getClassLoader(), classNames);
+
+ assertEquals(1, filteredClassNames.size());
+ assertEquals(filterObservers, filteredClassNames.get(0));
+ }
+
+ private List validate(String className) {
+ ClassLoader classLoader = getClass().getClassLoader();
+ return validate(classLoader, className);
+ }
+
+ private List validate(ClassLoader classLoader, String className) {
+ List classNames = Lists.newArrayList(getClass().getName() + "$" + className);
+ return validator.validate(classLoader, classNames);
+ }
+
+ /*
+ * In this test case, we are try to load a not-existent class.
+ */
+ @Test
+ public void testNoSuchClass() throws IOException {
+ List violations = validate("NoSuchClass");
+ assertEquals(1, violations.size());
+
+ CoprocessorViolation violation = violations.get(0);
+ assertEquals(Severity.ERROR, violation.getSeverity());
+ assertTrue(violation.getMessage().contains(
+ "java.lang.ClassNotFoundException: " +
+ "org.apache.hadoop.hbase.tool.coprocessor.CoprocessorValidatorTest$NoSuchClass"));
+ }
+
+ /*
+ * In this test case, we are validating MissingClass coprocessor, which
+ * references a missing class. With a special classloader, we prevent that
+ * class to be loaded at runtime. It simulates similar cases where a class
+ * is no more on our classpath.
+ * E.g. org.apache.hadoop.hbase.regionserver.wal.WALEdit was moved to
+ * org.apache.hadoop.hbase.wal, so class loading will fail on 2.0.
+ */
+ private static class MissingClass {
+ }
+
+ @SuppressWarnings("unused")
+ private static class MissingClassObserver {
+ public void method(MissingClass missingClass) {
+ }
+ }
+
+ private static class MissingClassClassLoader extends ClassLoader {
+ public MissingClassClassLoader() {
+ super(getClassLoader());
+ }
+
+ @Override
+ public Class> loadClass(String name) throws ClassNotFoundException {
+ if (name.equals(getFullClassName("MissingClass"))) {
+ throw new ClassNotFoundException(name);
+ }
+
+ return super.findClass(name);
+ }
+ }
+
+ @Test
+ public void testMissingClass() throws IOException {
+ MissingClassClassLoader missingClassClassLoader = new MissingClassClassLoader();
+ List violations = validate(missingClassClassLoader,
+ "MissingClassObserver");
+ assertEquals(1, violations.size());
+
+ CoprocessorViolation violation = violations.get(0);
+ assertEquals(Severity.ERROR, violation.getSeverity());
+ assertTrue(violation.getMessage().contains(
+ "java.lang.ClassNotFoundException: " +
+ "org.apache.hadoop.hbase.tool.coprocessor.CoprocessorValidatorTest$MissingClass"));
+ }
+
+ /*
+ * ObsoleteMethod coprocessor implements preCreateTable method which has
+ * HRegionInfo parameters. In our current implementation, we pass only
+ * RegionInfo parameters, so this method won't be called by HBase at all.
+ */
+ @SuppressWarnings("unused")
+ private static class ObsoleteMethodObserver /* implements MasterObserver */ {
+ public void preCreateTable(ObserverContext ctx,
+ HTableDescriptor desc, HRegionInfo[] regions) throws IOException {
+ }
+ }
+
+ @Test
+ public void testObsoleteMethod() throws IOException {
+ List violations = validate("ObsoleteMethodObserver");
+ assertEquals(1, violations.size());
+
+ CoprocessorViolation violation = violations.get(0);
+ assertEquals(Severity.WARNING, violation.getSeverity());
+ assertTrue(violation.getMessage().contains("was removed from new coprocessor API"));
+ }
+}
diff --git a/src/main/asciidoc/_chapters/ops_mgt.adoc b/src/main/asciidoc/_chapters/ops_mgt.adoc
index 6fef714c916..60b749fba97 100644
--- a/src/main/asciidoc/_chapters/ops_mgt.adoc
+++ b/src/main/asciidoc/_chapters/ops_mgt.adoc
@@ -845,11 +845,39 @@ The LoadTestTool has received many updates in recent HBase releases, including s
[[ops.pre-upgrade]]
=== Pre-Upgrade validator
Pre-Upgrade validator tool can be used to check the cluster for known incompatibilities before upgrading from HBase 1 to HBase 2.
-To run all the checks use the `-all` flag.
[source, bash]
----
-$ bin/hbase pre-upgrade -all
+$ bin/hbase pre-upgrade command ...
+----
+
+==== Coprocessor validation
+
+HBase supports co-processors for a long time, but the co-processor API can be changed between major releases. Co-processor validator tries to determine
+whether the old co-processors are still compatible with the actual HBase version.
+
+[source, bash]
+----
+$ bin/hbase pre-upgrade validate-cp -scan|
+Options:
+ -e Treat warnings as errors.
+ -scan Scan jar for observers.
+----
+
+The first parameter of the tool is the `jar` file which holds the co-processor implementation. Further parameters can be `-scan` when the tool will
+search the jar file for `Coprocessor` implementations or the `classes` can be explicitly given.
+
+The tool can report errors and warnings. Errors mean that HBase won't be able to load the coprocessor, because it is incompatible with the current version
+of HBase. Warnings mean that the co-processors can be loaded, but they won't work as expected. If `-e` option is given, then the tool will also fail
+for warnings.
+
+Please note that this tool cannot validate every aspect of jar files, it just does some static checks.
+
+For example:
+
+[source, bash]
+----
+$ bin/hbase pre-upgrade validate-cp my-coprocessor.jar MyMasterObserver MyRegionObserver
----
==== DataBlockEncoding validation
@@ -858,7 +886,7 @@ To verify that none of the column families are using incompatible Data Block Enc
[source, bash]
----
-$ bin/hbase pre-upgrade -validateDBE
+$ bin/hbase pre-upgrade validate-dbe
----
This check validates all column families and print out any incompatibilities.