HBASE-18954 Make *CoprocessorHost classes private.
Change-Id: I89fded0f74ad83c9bcc2a2b2de925f56aed4e11b
This commit is contained in:
parent
240b4b16ff
commit
202e414eb2
|
@ -53,6 +53,7 @@ import org.apache.hadoop.hbase.mapreduce.ResultSerialization;
|
||||||
import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
|
import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken;
|
import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.ExportProtos;
|
import org.apache.hadoop.hbase.protobuf.generated.ExportProtos;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.HRegion;
|
||||||
import org.apache.hadoop.hbase.regionserver.InternalScanner;
|
import org.apache.hadoop.hbase.regionserver.InternalScanner;
|
||||||
import org.apache.hadoop.hbase.regionserver.Region;
|
import org.apache.hadoop.hbase.regionserver.Region;
|
||||||
import org.apache.hadoop.hbase.regionserver.RegionScanner;
|
import org.apache.hadoop.hbase.regionserver.RegionScanner;
|
||||||
|
@ -377,10 +378,10 @@ public class Export extends ExportProtos.ExportService implements RegionCoproces
|
||||||
|
|
||||||
private static class ScanCoprocessor {
|
private static class ScanCoprocessor {
|
||||||
|
|
||||||
private final Region region;
|
private final HRegion region;
|
||||||
|
|
||||||
ScanCoprocessor(final Region region) {
|
ScanCoprocessor(final Region region) {
|
||||||
this.region = region;
|
this.region = (HRegion) region;
|
||||||
}
|
}
|
||||||
|
|
||||||
RegionScanner checkScannerOpen(final Scan scan) throws IOException {
|
RegionScanner checkScannerOpen(final Scan scan) throws IOException {
|
||||||
|
|
|
@ -84,7 +84,7 @@ public class SecureBulkLoadEndpoint extends SecureBulkLoadService implements Reg
|
||||||
try {
|
try {
|
||||||
SecureBulkLoadManager secureBulkLoadManager = this.rsServices.getSecureBulkLoadManager();
|
SecureBulkLoadManager secureBulkLoadManager = this.rsServices.getSecureBulkLoadManager();
|
||||||
|
|
||||||
String bulkToken = secureBulkLoadManager.prepareBulkLoad(this.env.getRegion(),
|
String bulkToken = secureBulkLoadManager.prepareBulkLoad((HRegion) this.env.getRegion(),
|
||||||
convert(request));
|
convert(request));
|
||||||
done.run(PrepareBulkLoadResponse.newBuilder().setBulkToken(bulkToken).build());
|
done.run(PrepareBulkLoadResponse.newBuilder().setBulkToken(bulkToken).build());
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
|
@ -110,7 +110,7 @@ public class SecureBulkLoadEndpoint extends SecureBulkLoadService implements Reg
|
||||||
RpcCallback<CleanupBulkLoadResponse> done) {
|
RpcCallback<CleanupBulkLoadResponse> done) {
|
||||||
try {
|
try {
|
||||||
SecureBulkLoadManager secureBulkLoadManager = this.rsServices.getSecureBulkLoadManager();
|
SecureBulkLoadManager secureBulkLoadManager = this.rsServices.getSecureBulkLoadManager();
|
||||||
secureBulkLoadManager.cleanupBulkLoad(this.env.getRegion(), convert(request));
|
secureBulkLoadManager.cleanupBulkLoad((HRegion) this.env.getRegion(), convert(request));
|
||||||
done.run(CleanupBulkLoadResponse.newBuilder().build());
|
done.run(CleanupBulkLoadResponse.newBuilder().build());
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
CoprocessorRpcUtils.setControllerException(controller, e);
|
CoprocessorRpcUtils.setControllerException(controller, e);
|
||||||
|
|
|
@ -21,8 +21,16 @@ package org.apache.hadoop.hbase.coprocessor;
|
||||||
import org.apache.commons.logging.Log;
|
import org.apache.commons.logging.Log;
|
||||||
import org.apache.commons.logging.LogFactory;
|
import org.apache.commons.logging.LogFactory;
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.hbase.*;
|
import org.apache.hadoop.hbase.Coprocessor;
|
||||||
|
import org.apache.hadoop.hbase.CoprocessorEnvironment;
|
||||||
|
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||||
|
import org.apache.hadoop.hbase.HColumnDescriptor;
|
||||||
|
import org.apache.hadoop.hbase.HTableDescriptor;
|
||||||
|
import org.apache.hadoop.hbase.MiniHBaseCluster;
|
||||||
|
import org.apache.hadoop.hbase.ServerName;
|
||||||
|
import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.hadoop.hbase.client.Admin;
|
import org.apache.hadoop.hbase.client.Admin;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.HRegion;
|
||||||
import org.apache.hadoop.hbase.regionserver.Region;
|
import org.apache.hadoop.hbase.regionserver.Region;
|
||||||
import org.apache.hadoop.hbase.regionserver.TestServerCustomProtocol;
|
import org.apache.hadoop.hbase.regionserver.TestServerCustomProtocol;
|
||||||
import org.apache.hadoop.hbase.testclassification.CoprocessorTests;
|
import org.apache.hadoop.hbase.testclassification.CoprocessorTests;
|
||||||
|
@ -176,7 +184,7 @@ public class TestClassLoading {
|
||||||
boolean found1 = true, found2 = true, found2_k1 = true, found2_k2 = true, found2_k3 = true;
|
boolean found1 = true, found2 = true, found2_k1 = true, found2_k2 = true, found2_k3 = true;
|
||||||
Map<Region, Set<ClassLoader>> regionsActiveClassLoaders = new HashMap<>();
|
Map<Region, Set<ClassLoader>> regionsActiveClassLoaders = new HashMap<>();
|
||||||
MiniHBaseCluster hbase = TEST_UTIL.getHBaseCluster();
|
MiniHBaseCluster hbase = TEST_UTIL.getHBaseCluster();
|
||||||
for (Region region:
|
for (HRegion region:
|
||||||
hbase.getRegionServer(0).getOnlineRegionsLocalContext()) {
|
hbase.getRegionServer(0).getOnlineRegionsLocalContext()) {
|
||||||
if (region.getRegionInfo().getRegionNameAsString().startsWith(tableName.getNameAsString())) {
|
if (region.getRegionInfo().getRegionNameAsString().startsWith(tableName.getNameAsString())) {
|
||||||
foundTableRegion = true;
|
foundTableRegion = true;
|
||||||
|
@ -245,7 +253,7 @@ public class TestClassLoading {
|
||||||
// verify that the coprocessor was loaded
|
// verify that the coprocessor was loaded
|
||||||
boolean found = false;
|
boolean found = false;
|
||||||
MiniHBaseCluster hbase = TEST_UTIL.getHBaseCluster();
|
MiniHBaseCluster hbase = TEST_UTIL.getHBaseCluster();
|
||||||
for (Region region: hbase.getRegionServer(0).getOnlineRegionsLocalContext()) {
|
for (HRegion region: hbase.getRegionServer(0).getOnlineRegionsLocalContext()) {
|
||||||
if (region.getRegionInfo().getRegionNameAsString().startsWith(cpName3)) {
|
if (region.getRegionInfo().getRegionNameAsString().startsWith(cpName3)) {
|
||||||
found = (region.getCoprocessorHost().findCoprocessor(cpName3) != null);
|
found = (region.getCoprocessorHost().findCoprocessor(cpName3) != null);
|
||||||
}
|
}
|
||||||
|
@ -270,7 +278,7 @@ public class TestClassLoading {
|
||||||
// verify that the coprocessor was loaded correctly
|
// verify that the coprocessor was loaded correctly
|
||||||
boolean found = false;
|
boolean found = false;
|
||||||
MiniHBaseCluster hbase = TEST_UTIL.getHBaseCluster();
|
MiniHBaseCluster hbase = TEST_UTIL.getHBaseCluster();
|
||||||
for (Region region: hbase.getRegionServer(0).getOnlineRegionsLocalContext()) {
|
for (HRegion region: hbase.getRegionServer(0).getOnlineRegionsLocalContext()) {
|
||||||
if (region.getRegionInfo().getRegionNameAsString().startsWith(cpName4)) {
|
if (region.getRegionInfo().getRegionNameAsString().startsWith(cpName4)) {
|
||||||
Coprocessor cp = region.getCoprocessorHost().findCoprocessor(cpName4);
|
Coprocessor cp = region.getCoprocessorHost().findCoprocessor(cpName4);
|
||||||
if (cp != null) {
|
if (cp != null) {
|
||||||
|
@ -341,7 +349,7 @@ public class TestClassLoading {
|
||||||
found6_k4 = false;
|
found6_k4 = false;
|
||||||
|
|
||||||
MiniHBaseCluster hbase = TEST_UTIL.getHBaseCluster();
|
MiniHBaseCluster hbase = TEST_UTIL.getHBaseCluster();
|
||||||
for (Region region: hbase.getRegionServer(0).getOnlineRegionsLocalContext()) {
|
for (HRegion region: hbase.getRegionServer(0).getOnlineRegionsLocalContext()) {
|
||||||
if (region.getRegionInfo().getRegionNameAsString().startsWith(tableName.getNameAsString())) {
|
if (region.getRegionInfo().getRegionNameAsString().startsWith(tableName.getNameAsString())) {
|
||||||
found_1 = found_1 ||
|
found_1 = found_1 ||
|
||||||
(region.getCoprocessorHost().findCoprocessor(cpName1) != null);
|
(region.getCoprocessorHost().findCoprocessor(cpName1) != null);
|
||||||
|
@ -429,7 +437,7 @@ public class TestClassLoading {
|
||||||
boolean found1 = false, found2 = false, found2_k1 = false,
|
boolean found1 = false, found2 = false, found2_k1 = false,
|
||||||
found2_k2 = false, found2_k3 = false;
|
found2_k2 = false, found2_k3 = false;
|
||||||
MiniHBaseCluster hbase = TEST_UTIL.getHBaseCluster();
|
MiniHBaseCluster hbase = TEST_UTIL.getHBaseCluster();
|
||||||
for (Region region: hbase.getRegionServer(0).getOnlineRegionsLocalContext()) {
|
for (HRegion region: hbase.getRegionServer(0).getOnlineRegionsLocalContext()) {
|
||||||
if (region.getRegionInfo().getRegionNameAsString().startsWith(tableName.getNameAsString())) {
|
if (region.getRegionInfo().getRegionNameAsString().startsWith(tableName.getNameAsString())) {
|
||||||
CoprocessorEnvironment env;
|
CoprocessorEnvironment env;
|
||||||
env = region.getCoprocessorHost().findCoprocessorEnvironment(cpName1);
|
env = region.getCoprocessorHost().findCoprocessorEnvironment(cpName1);
|
||||||
|
|
|
@ -58,8 +58,7 @@ import org.apache.hadoop.hbase.util.SortedList;
|
||||||
* @param <E> type of specific coprocessor environment this host requires.
|
* @param <E> type of specific coprocessor environment this host requires.
|
||||||
* provides
|
* provides
|
||||||
*/
|
*/
|
||||||
@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.COPROC)
|
@InterfaceAudience.Private
|
||||||
@InterfaceStability.Evolving
|
|
||||||
public abstract class CoprocessorHost<C extends Coprocessor, E extends CoprocessorEnvironment<C>> {
|
public abstract class CoprocessorHost<C extends Coprocessor, E extends CoprocessorEnvironment<C>> {
|
||||||
public static final String REGION_COPROCESSOR_CONF_KEY =
|
public static final String REGION_COPROCESSOR_CONF_KEY =
|
||||||
"hbase.coprocessor.region.classes";
|
"hbase.coprocessor.region.classes";
|
||||||
|
|
|
@ -7906,7 +7906,6 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi
|
||||||
}
|
}
|
||||||
|
|
||||||
/** @return the coprocessor host */
|
/** @return the coprocessor host */
|
||||||
@Override
|
|
||||||
public RegionCoprocessorHost getCoprocessorHost() {
|
public RegionCoprocessorHost getCoprocessorHost() {
|
||||||
return coprocessorHost;
|
return coprocessorHost;
|
||||||
}
|
}
|
||||||
|
|
|
@ -3086,7 +3086,7 @@ public class HRegionServer extends HasThread implements
|
||||||
protected boolean closeRegion(String encodedName, final boolean abort, final ServerName sn)
|
protected boolean closeRegion(String encodedName, final boolean abort, final ServerName sn)
|
||||||
throws NotServingRegionException {
|
throws NotServingRegionException {
|
||||||
//Check for permissions to close.
|
//Check for permissions to close.
|
||||||
Region actualRegion = this.getRegion(encodedName);
|
HRegion actualRegion = this.getRegion(encodedName);
|
||||||
// Can be null if we're calling close on a region that's not online
|
// Can be null if we're calling close on a region that's not online
|
||||||
if ((actualRegion != null) && (actualRegion.getCoprocessorHost() != null)) {
|
if ((actualRegion != null) && (actualRegion.getCoprocessorHost() != null)) {
|
||||||
try {
|
try {
|
||||||
|
@ -3205,7 +3205,7 @@ public class HRegionServer extends HasThread implements
|
||||||
* @return HRegion for the passed binary <code>regionName</code> or null if
|
* @return HRegion for the passed binary <code>regionName</code> or null if
|
||||||
* named region is not member of the online regions.
|
* named region is not member of the online regions.
|
||||||
*/
|
*/
|
||||||
public Region getOnlineRegion(final byte[] regionName) {
|
public HRegion getOnlineRegion(final byte[] regionName) {
|
||||||
String encodedRegionName = RegionInfo.encodeRegionName(regionName);
|
String encodedRegionName = RegionInfo.encodeRegionName(regionName);
|
||||||
return this.onlineRegions.get(encodedRegionName);
|
return this.onlineRegions.get(encodedRegionName);
|
||||||
}
|
}
|
||||||
|
|
|
@ -2120,7 +2120,6 @@ public class HStore implements Store, HeapSize, StoreConfigInformation, Propagat
|
||||||
return this.region;
|
return this.region;
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
|
||||||
public RegionCoprocessorHost getCoprocessorHost() {
|
public RegionCoprocessorHost getCoprocessorHost() {
|
||||||
return this.region.getCoprocessorHost();
|
return this.region.getCoprocessorHost();
|
||||||
}
|
}
|
||||||
|
|
|
@ -453,10 +453,6 @@ public interface Region extends ConfigurationObserver {
|
||||||
*/
|
*/
|
||||||
void put(Put put) throws IOException;
|
void put(Put put) throws IOException;
|
||||||
|
|
||||||
/** @return the coprocessor host */
|
|
||||||
// TODO To be removed by HBASE-18954
|
|
||||||
RegionCoprocessorHost getCoprocessorHost();
|
|
||||||
|
|
||||||
///////////////////////////////////////////////////////////////////////////
|
///////////////////////////////////////////////////////////////////////////
|
||||||
// Flushes, compactions, splits, etc.
|
// Flushes, compactions, splits, etc.
|
||||||
// Wizards only, please
|
// Wizards only, please
|
||||||
|
|
|
@ -91,8 +91,7 @@ import org.apache.yetus.audience.InterfaceStability;
|
||||||
* Implements the coprocessor environment and runtime support for coprocessors
|
* Implements the coprocessor environment and runtime support for coprocessors
|
||||||
* loaded within a {@link Region}.
|
* loaded within a {@link Region}.
|
||||||
*/
|
*/
|
||||||
@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.COPROC)
|
@InterfaceAudience.Private
|
||||||
@InterfaceStability.Evolving
|
|
||||||
public class RegionCoprocessorHost
|
public class RegionCoprocessorHost
|
||||||
extends CoprocessorHost<RegionCoprocessor, RegionCoprocessorEnvironment> {
|
extends CoprocessorHost<RegionCoprocessor, RegionCoprocessorEnvironment> {
|
||||||
|
|
||||||
|
|
|
@ -39,8 +39,7 @@ import org.apache.hadoop.hbase.security.User;
|
||||||
import org.apache.yetus.audience.InterfaceAudience;
|
import org.apache.yetus.audience.InterfaceAudience;
|
||||||
import org.apache.yetus.audience.InterfaceStability;
|
import org.apache.yetus.audience.InterfaceStability;
|
||||||
|
|
||||||
@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.COPROC)
|
@InterfaceAudience.Private
|
||||||
@InterfaceStability.Evolving
|
|
||||||
public class RegionServerCoprocessorHost extends
|
public class RegionServerCoprocessorHost extends
|
||||||
CoprocessorHost<RegionServerCoprocessor, RegionServerCoprocessorEnvironment> {
|
CoprocessorHost<RegionServerCoprocessor, RegionServerCoprocessorEnvironment> {
|
||||||
|
|
||||||
|
|
|
@ -132,7 +132,7 @@ public class SecureBulkLoadManager {
|
||||||
public void stop() throws IOException {
|
public void stop() throws IOException {
|
||||||
}
|
}
|
||||||
|
|
||||||
public String prepareBulkLoad(final Region region, final PrepareBulkLoadRequest request)
|
public String prepareBulkLoad(final HRegion region, final PrepareBulkLoadRequest request)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
User user = getActiveUser();
|
User user = getActiveUser();
|
||||||
region.getCoprocessorHost().prePrepareBulkLoad(user);
|
region.getCoprocessorHost().prePrepareBulkLoad(user);
|
||||||
|
@ -144,7 +144,7 @@ public class SecureBulkLoadManager {
|
||||||
return bulkToken;
|
return bulkToken;
|
||||||
}
|
}
|
||||||
|
|
||||||
public void cleanupBulkLoad(final Region region, final CleanupBulkLoadRequest request)
|
public void cleanupBulkLoad(final HRegion region, final CleanupBulkLoadRequest request)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
region.getCoprocessorHost().preCleanupBulkLoad(getActiveUser());
|
region.getCoprocessorHost().preCleanupBulkLoad(getActiveUser());
|
||||||
|
|
||||||
|
|
|
@ -204,8 +204,6 @@ public interface Store {
|
||||||
*/
|
*/
|
||||||
RegionInfo getRegionInfo();
|
RegionInfo getRegionInfo();
|
||||||
|
|
||||||
RegionCoprocessorHost getCoprocessorHost();
|
|
||||||
|
|
||||||
boolean areWritesEnabled();
|
boolean areWritesEnabled();
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
|
@ -44,7 +44,6 @@ import org.apache.hadoop.hbase.coprocessor.RegionObserver;
|
||||||
import org.apache.hadoop.hbase.regionserver.HRegionServer;
|
import org.apache.hadoop.hbase.regionserver.HRegionServer;
|
||||||
import org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress;
|
import org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress;
|
||||||
import org.apache.hadoop.hbase.TableName;
|
import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.hadoop.hbase.regionserver.Region;
|
|
||||||
import org.apache.hadoop.hbase.regionserver.RegionScanner;
|
import org.apache.hadoop.hbase.regionserver.RegionScanner;
|
||||||
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||||
import org.apache.hadoop.hbase.HColumnDescriptor;
|
import org.apache.hadoop.hbase.HColumnDescriptor;
|
||||||
|
@ -960,7 +959,7 @@ public class TestFromClientSide3 {
|
||||||
HRegion region = (HRegion) find(tableName);
|
HRegion region = (HRegion) find(tableName);
|
||||||
assertEquals(0, region.getLockedRows().size());
|
assertEquals(0, region.getLockedRows().size());
|
||||||
}
|
}
|
||||||
private static Region find(final TableName tableName)
|
private static HRegion find(final TableName tableName)
|
||||||
throws IOException, InterruptedException {
|
throws IOException, InterruptedException {
|
||||||
HRegionServer rs = TEST_UTIL.getRSForFirstRegionInTable(tableName);
|
HRegionServer rs = TEST_UTIL.getRSForFirstRegionInTable(tableName);
|
||||||
List<HRegion> regions = rs.getRegions(tableName);
|
List<HRegion> regions = rs.getRegions(tableName);
|
||||||
|
@ -970,7 +969,7 @@ public class TestFromClientSide3 {
|
||||||
|
|
||||||
private static <T extends RegionObserver> T find(final TableName tableName,
|
private static <T extends RegionObserver> T find(final TableName tableName,
|
||||||
Class<T> clz) throws IOException, InterruptedException {
|
Class<T> clz) throws IOException, InterruptedException {
|
||||||
Region region = find(tableName);
|
HRegion region = find(tableName);
|
||||||
Coprocessor cp = region.getCoprocessorHost().findCoprocessor(clz.getName());
|
Coprocessor cp = region.getCoprocessorHost().findCoprocessor(clz.getName());
|
||||||
assertTrue("The cp instance should be " + clz.getName()
|
assertTrue("The cp instance should be " + clz.getName()
|
||||||
+ ", current instance is " + cp.getClass().getName(), clz.isInstance(cp));
|
+ ", current instance is " + cp.getClass().getName(), clz.isInstance(cp));
|
||||||
|
|
|
@ -209,7 +209,7 @@ public class TestRegionObserverScannerOpenHook {
|
||||||
byte[][] FAMILIES = new byte[][] { A };
|
byte[][] FAMILIES = new byte[][] { A };
|
||||||
|
|
||||||
Configuration conf = HBaseConfiguration.create();
|
Configuration conf = HBaseConfiguration.create();
|
||||||
Region region = initHRegion(TABLE, getClass().getName(), conf, FAMILIES);
|
HRegion region = initHRegion(TABLE, getClass().getName(), conf, FAMILIES);
|
||||||
RegionCoprocessorHost h = region.getCoprocessorHost();
|
RegionCoprocessorHost h = region.getCoprocessorHost();
|
||||||
h.load(NoDataFromScan.class, Coprocessor.PRIORITY_HIGHEST, conf);
|
h.load(NoDataFromScan.class, Coprocessor.PRIORITY_HIGHEST, conf);
|
||||||
h.load(EmptyRegionObsever.class, Coprocessor.PRIORITY_USER, conf);
|
h.load(EmptyRegionObsever.class, Coprocessor.PRIORITY_USER, conf);
|
||||||
|
|
|
@ -60,7 +60,7 @@ import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos;
|
import org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService;
|
import org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.CheckPermissionsRequest;
|
import org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.CheckPermissionsRequest;
|
||||||
import org.apache.hadoop.hbase.regionserver.Region;
|
import org.apache.hadoop.hbase.regionserver.HRegion;
|
||||||
import org.apache.hadoop.hbase.security.AccessDeniedException;
|
import org.apache.hadoop.hbase.security.AccessDeniedException;
|
||||||
import org.apache.hadoop.hbase.security.User;
|
import org.apache.hadoop.hbase.security.User;
|
||||||
import org.apache.hadoop.hbase.security.access.Permission.Action;
|
import org.apache.hadoop.hbase.security.access.Permission.Action;
|
||||||
|
@ -305,7 +305,7 @@ public class SecureTestUtil {
|
||||||
private static List<AccessController> getAccessControllers(MiniHBaseCluster cluster) {
|
private static List<AccessController> getAccessControllers(MiniHBaseCluster cluster) {
|
||||||
List<AccessController> result = Lists.newArrayList();
|
List<AccessController> result = Lists.newArrayList();
|
||||||
for (RegionServerThread t: cluster.getLiveRegionServerThreads()) {
|
for (RegionServerThread t: cluster.getLiveRegionServerThreads()) {
|
||||||
for (Region region: t.getRegionServer().getOnlineRegionsLocalContext()) {
|
for (HRegion region: t.getRegionServer().getOnlineRegionsLocalContext()) {
|
||||||
Coprocessor cp = region.getCoprocessorHost().findCoprocessor(AccessController.class);
|
Coprocessor cp = region.getCoprocessorHost().findCoprocessor(AccessController.class);
|
||||||
if (cp != null) {
|
if (cp != null) {
|
||||||
result.add((AccessController)cp);
|
result.add((AccessController)cp);
|
||||||
|
|
|
@ -112,8 +112,8 @@ import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos;
|
import org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService;
|
import org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService;
|
||||||
import org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.CheckPermissionsRequest;
|
import org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.CheckPermissionsRequest;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.HRegion;
|
||||||
import org.apache.hadoop.hbase.regionserver.HRegionServer;
|
import org.apache.hadoop.hbase.regionserver.HRegionServer;
|
||||||
import org.apache.hadoop.hbase.regionserver.Region;
|
|
||||||
import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost;
|
import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost;
|
||||||
import org.apache.hadoop.hbase.regionserver.RegionServerCoprocessorHost;
|
import org.apache.hadoop.hbase.regionserver.RegionServerCoprocessorHost;
|
||||||
import org.apache.hadoop.hbase.regionserver.ScanType;
|
import org.apache.hadoop.hbase.regionserver.ScanType;
|
||||||
|
@ -275,7 +275,7 @@ public class TestAccessController extends SecureTestUtil {
|
||||||
htd.setOwner(USER_OWNER);
|
htd.setOwner(USER_OWNER);
|
||||||
createTable(TEST_UTIL, htd, new byte[][] { Bytes.toBytes("s") });
|
createTable(TEST_UTIL, htd, new byte[][] { Bytes.toBytes("s") });
|
||||||
|
|
||||||
Region region = TEST_UTIL.getHBaseCluster().getRegions(TEST_TABLE).get(0);
|
HRegion region = TEST_UTIL.getHBaseCluster().getRegions(TEST_TABLE).get(0);
|
||||||
RegionCoprocessorHost rcpHost = region.getCoprocessorHost();
|
RegionCoprocessorHost rcpHost = region.getCoprocessorHost();
|
||||||
RCP_ENV = rcpHost.createEnvironment(ACCESS_CONTROLLER, Coprocessor.PRIORITY_HIGHEST, 1, conf);
|
RCP_ENV = rcpHost.createEnvironment(ACCESS_CONTROLLER, Coprocessor.PRIORITY_HIGHEST, 1, conf);
|
||||||
|
|
||||||
|
@ -2605,7 +2605,7 @@ public class TestAccessController extends SecureTestUtil {
|
||||||
for (JVMClusterUtil.RegionServerThread thread:
|
for (JVMClusterUtil.RegionServerThread thread:
|
||||||
TEST_UTIL.getMiniHBaseCluster().getRegionServerThreads()) {
|
TEST_UTIL.getMiniHBaseCluster().getRegionServerThreads()) {
|
||||||
HRegionServer rs = thread.getRegionServer();
|
HRegionServer rs = thread.getRegionServer();
|
||||||
for (Region region: rs.getRegions(TEST_TABLE)) {
|
for (HRegion region: rs.getRegions(TEST_TABLE)) {
|
||||||
region.getCoprocessorHost().load(PingCoprocessor.class,
|
region.getCoprocessorHost().load(PingCoprocessor.class,
|
||||||
Coprocessor.PRIORITY_USER, conf);
|
Coprocessor.PRIORITY_USER, conf);
|
||||||
}
|
}
|
||||||
|
|
|
@ -37,8 +37,8 @@ import org.apache.hadoop.hbase.coprocessor.ObserverContextImpl;
|
||||||
import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment;
|
import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment;
|
||||||
import org.apache.hadoop.hbase.coprocessor.RegionServerCoprocessorEnvironment;
|
import org.apache.hadoop.hbase.coprocessor.RegionServerCoprocessorEnvironment;
|
||||||
import org.apache.hadoop.hbase.master.MasterCoprocessorHost;
|
import org.apache.hadoop.hbase.master.MasterCoprocessorHost;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.HRegion;
|
||||||
import org.apache.hadoop.hbase.regionserver.HRegionServer;
|
import org.apache.hadoop.hbase.regionserver.HRegionServer;
|
||||||
import org.apache.hadoop.hbase.regionserver.Region;
|
|
||||||
import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost;
|
import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost;
|
||||||
import org.apache.hadoop.hbase.regionserver.RegionServerCoprocessorHost;
|
import org.apache.hadoop.hbase.regionserver.RegionServerCoprocessorHost;
|
||||||
import org.apache.hadoop.hbase.security.User;
|
import org.apache.hadoop.hbase.security.User;
|
||||||
|
@ -214,7 +214,7 @@ public class TestAccessController3 extends SecureTestUtil {
|
||||||
htd.setOwner(USER_OWNER);
|
htd.setOwner(USER_OWNER);
|
||||||
createTable(TEST_UTIL, htd, new byte[][] { Bytes.toBytes("s") });
|
createTable(TEST_UTIL, htd, new byte[][] { Bytes.toBytes("s") });
|
||||||
|
|
||||||
Region region = TEST_UTIL.getHBaseCluster().getRegions(TEST_TABLE).get(0);
|
HRegion region = TEST_UTIL.getHBaseCluster().getRegions(TEST_TABLE).get(0);
|
||||||
RegionCoprocessorHost rcpHost = region.getCoprocessorHost();
|
RegionCoprocessorHost rcpHost = region.getCoprocessorHost();
|
||||||
RCP_ENV = rcpHost.createEnvironment(ACCESS_CONTROLLER, Coprocessor.PRIORITY_HIGHEST, 1, conf);
|
RCP_ENV = rcpHost.createEnvironment(ACCESS_CONTROLLER, Coprocessor.PRIORITY_HIGHEST, 1, conf);
|
||||||
|
|
||||||
|
|
|
@ -58,8 +58,8 @@ import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment;
|
||||||
import org.apache.hadoop.hbase.coprocessor.RegionServerCoprocessorEnvironment;
|
import org.apache.hadoop.hbase.coprocessor.RegionServerCoprocessorEnvironment;
|
||||||
import org.apache.hadoop.hbase.filter.BinaryComparator;
|
import org.apache.hadoop.hbase.filter.BinaryComparator;
|
||||||
import org.apache.hadoop.hbase.master.MasterCoprocessorHost;
|
import org.apache.hadoop.hbase.master.MasterCoprocessorHost;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.HRegion;
|
||||||
import org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress;
|
import org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress;
|
||||||
import org.apache.hadoop.hbase.regionserver.Region;
|
|
||||||
import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost;
|
import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost;
|
||||||
import org.apache.hadoop.hbase.regionserver.RegionScanner;
|
import org.apache.hadoop.hbase.regionserver.RegionScanner;
|
||||||
import org.apache.hadoop.hbase.regionserver.RegionServerCoprocessorHost;
|
import org.apache.hadoop.hbase.regionserver.RegionServerCoprocessorHost;
|
||||||
|
@ -189,7 +189,7 @@ public class TestWithDisabledAuthorization extends SecureTestUtil {
|
||||||
admin.createTable(htd, new byte[][] { Bytes.toBytes("s") });
|
admin.createTable(htd, new byte[][] { Bytes.toBytes("s") });
|
||||||
TEST_UTIL.waitUntilAllRegionsAssigned(TEST_TABLE.getTableName());
|
TEST_UTIL.waitUntilAllRegionsAssigned(TEST_TABLE.getTableName());
|
||||||
|
|
||||||
Region region = TEST_UTIL.getHBaseCluster().getRegions(TEST_TABLE.getTableName()).get(0);
|
HRegion region = TEST_UTIL.getHBaseCluster().getRegions(TEST_TABLE.getTableName()).get(0);
|
||||||
RegionCoprocessorHost rcpHost = region.getCoprocessorHost();
|
RegionCoprocessorHost rcpHost = region.getCoprocessorHost();
|
||||||
RCP_ENV = rcpHost.createEnvironment(ACCESS_CONTROLLER,
|
RCP_ENV = rcpHost.createEnvironment(ACCESS_CONTROLLER,
|
||||||
Coprocessor.PRIORITY_HIGHEST, 1, TEST_UTIL.getConfiguration());
|
Coprocessor.PRIORITY_HIGHEST, 1, TEST_UTIL.getConfiguration());
|
||||||
|
|
Loading…
Reference in New Issue